| /src/external/gpl3/gcc/dist/gcc/ |
| ipa-modref.h | 1 /* Search for references that a functions loads or stores. 31 modref_records *loads; variable 40 /* Side effects does not include memory loads and stores which are 41 expressed using loads, stores and calls_interposable fields. */ 50 In this situation the loads summary is not useful for DSE but 56 /* Total number of accesses in loads tree. */ 106 however it is possible that the other function body contains more loads
|
| ipa-modref.cc | 1 /* Search for references that a functions loads or stores. 21 /* Mod/ref pass records summary about loads and stores performed by the 280 : loads (NULL), stores (NULL), retslot_flags (0), static_chain_flags (0), 289 if (loads) 290 ggc_delete (loads); 340 if (loads && !loads->every_base) 359 modref_records_lto *loads; variable 379 : loads (NULL), stores (NULL), retslot_flags (0), static_chain_flags (0), 387 if (loads) [all...] |
| tree-vectorizer.h | 273 /* The group of nodes that contain loads of this SLP instance. */ 274 vec<slp_tree> loads; member in class:_slp_instance 296 #define SLP_INSTANCE_LOADS(S) (S)->loads 541 For example, if we implement a group of scalar loads using gather 542 loads, we'll use a separate gather load for each scalar load, and 1194 or a group of loads or stores. */ 1196 /* An access to an invariant address. This is used only for loads. */ 1229 /* The access uses gather loads or scatter stores. */ 1346 /* For loads only, the gap from the previous load. For consecutive loads, GA [all...] |
| /src/external/gpl3/gcc.old/dist/gcc/ |
| ipa-modref.h | 1 /* Search for references that a functions loads or stores. 31 modref_records *loads; variable 40 /* Side effects does not include memory loads and stores which are 41 expressed using loads, stores and calls_interposable fields. */ 50 In this situation the loads summary is not useful for DSE but 56 /* Total number of accesses in loads tree. */ 106 however it is possible that the other function body contains more loads
|
| ipa-modref.cc | 1 /* Search for references that a functions loads or stores. 21 /* Mod/ref pass records summary about loads and stores performed by the 278 : loads (NULL), stores (NULL), retslot_flags (0), static_chain_flags (0), 287 if (loads) 288 ggc_delete (loads); 338 if (loads && !loads->every_base) 357 modref_records_lto *loads; variable 377 : loads (NULL), stores (NULL), retslot_flags (0), static_chain_flags (0), 385 if (loads) [all...] |
| tree-vectorizer.h | 249 /* The group of nodes that contain loads of this SLP instance. */ 250 vec<slp_tree> loads; member in class:_slp_instance 272 #define SLP_INSTANCE_LOADS(S) (S)->loads 505 For example, if we implement a group of scalar loads using gather 506 loads, we'll use a separate gather load for each scalar load, and 1074 or a group of loads or stores. */ 1076 /* An access to an invariant address. This is used only for loads. */ 1109 /* The access uses gather loads or scatter stores. */ 1225 /* For loads only, the gap from the previous load. For consecutive loads, GA [all...] |
| /src/sys/compat/linux32/common/ |
| linux32_types.h | 106 netbsd32_u_long loads[3]; member in struct:linux32_sysinfo
|
| /src/external/apache2/llvm/dist/llvm/lib/Target/Hexagon/MCTargetDesc/ |
| HexagonShuffler.h | 147 // Number of memory operations, loads, solo loads, stores, solo stores, 150 unsigned loads; member in struct:llvm::HexagonShuffler::HexagonPacketSummary
|
| /src/external/gpl2/gmake/dist/ |
| getloadavg.c | 1008 int loads; 1011 loads = getloadavg (avg, 3); 1012 if (loads == -1) 1017 if (loads > 0) 1019 if (loads > 1) 1021 if (loads > 2) 1023 if (loads > 0) 1005 int loads; local
|
| /src/external/gpl3/gcc/dist/gcc/config/rs6000/ |
| rs6000-pcrel-opt.cc | 20 /* This file implements a RTL pass that looks for pc-relative loads of the 134 unsigned long loads; member in struct:__anon13419 257 the optimization. The float loads are all indexed so there are no problems 364 counters.loads++; 602 /* Given an insn that loads up a base register with the address of an 611 ADDR_INSN is the insn that loads the external symbol address. */ 704 /* We can only optimize loads and stores. Ignore everything else. */ 839 "\n# of loads of an address of an external symbol = %lu\n", 842 fprintf (dump_file, "# of PCREL_OPT loads = %lu (adjacent %lu)\n", 843 counters.loads, counters.adjacent_loads) [all...] |
| rs6000-string.cc | 153 is being generated, bswap loads are used. 504 LOAD_MODE is the machine mode for the loads. 562 LOAD_MODE is the machine mode for the loads. 631 OFF_REG is the register to use for the string offset for loads. 826 OFF_REG is the register to use for the string offset for loads. 2683 /* Generate loads and stores for a move of v4si mode using lvx/stvx. 2750 rtx loads[MAX_MOVE_REG]; local 2818 loads+stores. Also don't use it unless we likely already 2902 loads[num_reg] = (*gen_func.movlen) (tmp_reg, src, move_bytes); 2907 loads[num_reg] = (*gen_func.mov) (tmp_reg, src) [all...] |
| /src/external/gpl3/gcc.old/dist/gcc/config/rs6000/ |
| rs6000-pcrel-opt.cc | 20 /* This file implements a RTL pass that looks for pc-relative loads of the 134 unsigned long loads; member in struct:__anon15847 257 the optimization. The float loads are all indexed so there are no problems 364 counters.loads++; 602 /* Given an insn that loads up a base register with the address of an 611 ADDR_INSN is the insn that loads the external symbol address. */ 704 /* We can only optimize loads and stores. Ignore everything else. */ 839 "\n# of loads of an address of an external symbol = %lu\n", 842 fprintf (dump_file, "# of PCREL_OPT loads = %lu (adjacent %lu)\n", 843 counters.loads, counters.adjacent_loads) [all...] |
| rs6000-string.cc | 153 is being generated, bswap loads are used. 504 LOAD_MODE is the machine mode for the loads. 562 LOAD_MODE is the machine mode for the loads. 631 OFF_REG is the register to use for the string offset for loads. 826 OFF_REG is the register to use for the string offset for loads. 2692 /* Generate loads and stores for a move of v4si mode using lvx/stvx. 2759 rtx loads[MAX_MOVE_REG]; local 2827 loads+stores. Also don't use it unless we likely already 2911 loads[num_reg] = (*gen_func.movlen) (tmp_reg, src, move_bytes); 2916 loads[num_reg] = (*gen_func.mov) (tmp_reg, src) [all...] |
| /src/sys/arch/x86/x86/ |
| bus_dma.c | 138 BUSDMA_EVCNT_DECL(loads); variable 386 STAT_INCR(loads); 844 * never reorders loads from wp/wt/wb or uc memory, but it 845 * may execute loads from wc/wc+ memory early, e.g. with 982 * using a DMA memory buffer, with loads and stores, and is 990 * barrier is needed for prior loads. 1011 * loads and stores. 1015 * data from the device), prefetched loads are safe. x86
|
| /src/sys/compat/linux/common/ |
| linux_misc.h | 71 unsigned long loads[3]; member in struct:linux_sysinfo
|
| /src/external/apache2/llvm/dist/llvm/lib/CodeGen/ |
| MachineVerifier.cpp | 1027 // Generic loads and stores must have a single MachineMemOperand 2051 bool loads = MI->mayLoad(); local 2055 if (stores && loads) { 2065 loads = false; 2070 if (loads == stores) 2073 if (loads && !LI.liveAt(Idx.getRegSlot(true))) { 2074 report("Instruction loads from dead spill slot", MO, MONum);
|