| /src/external/bsd/jemalloc/dist/src/ |
| pa.c | 8 pa_nactive_add(pa_shard_t *shard, size_t add_pages) { 9 atomic_fetch_add_zu(&shard->nactive, add_pages, ATOMIC_RELAXED); 13 pa_nactive_sub(pa_shard_t *shard, size_t sub_pages) { 14 assert(atomic_load_zu(&shard->nactive, ATOMIC_RELAXED) >= sub_pages); 15 atomic_fetch_sub_zu(&shard->nactive, sub_pages, ATOMIC_RELAXED); 32 pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central, 39 if (edata_cache_init(&shard->edata_cache, base)) { 43 if (pac_init(tsdn, &shard->pac, base, emap, &shard->edata_cache, 49 shard->ind = ind [all...] |
| pa_extra.c | 12 pa_shard_prefork0(tsdn_t *tsdn, pa_shard_t *shard) { 13 malloc_mutex_prefork(tsdn, &shard->pac.decay_dirty.mtx); 14 malloc_mutex_prefork(tsdn, &shard->pac.decay_muzzy.mtx); 18 pa_shard_prefork2(tsdn_t *tsdn, pa_shard_t *shard) { 19 if (shard->ever_used_hpa) { 20 sec_prefork2(tsdn, &shard->hpa_sec); 25 pa_shard_prefork3(tsdn_t *tsdn, pa_shard_t *shard) { 26 malloc_mutex_prefork(tsdn, &shard->pac.grow_mtx); 27 if (shard->ever_used_hpa) { 28 hpa_shard_prefork3(tsdn, &shard->hpa_shard) [all...] |
| hpa.c | 57 hpa_do_consistency_checks(hpa_shard_t *shard) { 58 assert(shard->base != NULL); 97 * shard is exhausted. We should hold the grow_mtx on that shard. 178 hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap, 184 err = malloc_mutex_init(&shard->grow_mtx, "hpa_shard_grow", 189 err = malloc_mutex_init(&shard->mtx, "hpa_shard", 196 shard->central = central; 197 shard->base = base; 198 edata_cache_fast_init(&shard->ecf, edata_cache) 724 hpa_shard_t *shard = hpa_from_pai(self); local 851 hpa_shard_t *shard = hpa_from_pai(self); local 888 hpa_shard_t *shard = hpa_from_pai(self); local [all...] |
| sec.c | 45 sec_shard_t *shard = shard_cur; local 47 bool err = malloc_mutex_init(&shard->mtx, "sec_shard", 52 shard->enabled = true; 53 shard->bins = bin_cur; 55 sec_bin_init(&shard->bins[j]); 58 shard->bytes_cur = 0; 59 shard->to_flush_next = 0; 91 * Eventually, we should implement affinity, tracking source shard using 118 * empty cache, we'll try to fill it, which can push the shard over it's limit. 121 sec_flush_some_and_unlock(tsdn_t *tsdn, sec_t *sec, sec_shard_t *shard) { 237 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local 346 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local [all...] |
| /src/external/bsd/jemalloc.old/dist/src/ |
| pa.c | 8 pa_nactive_add(pa_shard_t *shard, size_t add_pages) { 9 atomic_fetch_add_zu(&shard->nactive, add_pages, ATOMIC_RELAXED); 13 pa_nactive_sub(pa_shard_t *shard, size_t sub_pages) { 14 assert(atomic_load_zu(&shard->nactive, ATOMIC_RELAXED) >= sub_pages); 15 atomic_fetch_sub_zu(&shard->nactive, sub_pages, ATOMIC_RELAXED); 32 pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central, 39 if (edata_cache_init(&shard->edata_cache, base)) { 43 if (pac_init(tsdn, &shard->pac, base, emap, &shard->edata_cache, 49 shard->ind = ind [all...] |
| pa_extra.c | 12 pa_shard_prefork0(tsdn_t *tsdn, pa_shard_t *shard) { 13 malloc_mutex_prefork(tsdn, &shard->pac.decay_dirty.mtx); 14 malloc_mutex_prefork(tsdn, &shard->pac.decay_muzzy.mtx); 18 pa_shard_prefork2(tsdn_t *tsdn, pa_shard_t *shard) { 19 if (shard->ever_used_hpa) { 20 sec_prefork2(tsdn, &shard->hpa_sec); 25 pa_shard_prefork3(tsdn_t *tsdn, pa_shard_t *shard) { 26 malloc_mutex_prefork(tsdn, &shard->pac.grow_mtx); 27 if (shard->ever_used_hpa) { 28 hpa_shard_prefork3(tsdn, &shard->hpa_shard) [all...] |
| hpa.c | 57 hpa_do_consistency_checks(hpa_shard_t *shard) { 58 assert(shard->base != NULL); 97 * shard is exhausted. We should hold the grow_mtx on that shard. 178 hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap, 184 err = malloc_mutex_init(&shard->grow_mtx, "hpa_shard_grow", 189 err = malloc_mutex_init(&shard->mtx, "hpa_shard", 196 shard->central = central; 197 shard->base = base; 198 edata_cache_fast_init(&shard->ecf, edata_cache) 724 hpa_shard_t *shard = hpa_from_pai(self); local 851 hpa_shard_t *shard = hpa_from_pai(self); local 888 hpa_shard_t *shard = hpa_from_pai(self); local [all...] |
| sec.c | 45 sec_shard_t *shard = shard_cur; local 47 bool err = malloc_mutex_init(&shard->mtx, "sec_shard", 52 shard->enabled = true; 53 shard->bins = bin_cur; 55 sec_bin_init(&shard->bins[j]); 58 shard->bytes_cur = 0; 59 shard->to_flush_next = 0; 91 * Eventually, we should implement affinity, tracking source shard using 118 * empty cache, we'll try to fill it, which can push the shard over it's limit. 121 sec_flush_some_and_unlock(tsdn_t *tsdn, sec_t *sec, sec_shard_t *shard) { 237 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local 346 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local [all...] |
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| pa.h | 69 /* The central PA this shard is associated with. */ 114 /* The emap this shard is tied to. */ 122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) { 123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 && 124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0; 128 pa_shard_ehooks_get(pa_shard_t *shard) { 129 return base_ehooks_get(shard->base); 137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central, 143 * This isn't exposed to users; we allow late enablement of the HPA shard so 146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard, [all...] |
| hpa.h | 89 /* Protects most of this shard's state. */ 92 * Guards the shard's access to the central allocator (preventing 93 * multiple threads operating on this shard from accessing the central 125 /* The configuration choices for this hpa shard. */ 130 * hpa shard. 135 * Those stats which are copied directly into the CTL-centric hpa shard 141 * Last time we performed purge on this shard. 153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap, 158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard, 162 * Notify the shard that we won't use it for allocations much longer. Due t [all...] |
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| pa.h | 69 /* The central PA this shard is associated with. */ 114 /* The emap this shard is tied to. */ 122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) { 123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 && 124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0; 128 pa_shard_ehooks_get(pa_shard_t *shard) { 129 return base_ehooks_get(shard->base); 137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central, 143 * This isn't exposed to users; we allow late enablement of the HPA shard so 146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard, [all...] |
| hpa.h | 89 /* Protects most of this shard's state. */ 92 * Guards the shard's access to the central allocator (preventing 93 * multiple threads operating on this shard from accessing the central 125 /* The configuration choices for this hpa shard. */ 130 * hpa shard. 135 * Those stats which are copied directly into the CTL-centric hpa shard 141 * Last time we performed purge on this shard. 153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap, 158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard, 162 * Notify the shard that we won't use it for allocations much longer. Due t [all...] |
| /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/ |
| pa.h | 69 /* The central PA this shard is associated with. */ 114 /* The emap this shard is tied to. */ 122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) { 123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 && 124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0; 128 pa_shard_ehooks_get(pa_shard_t *shard) { 129 return base_ehooks_get(shard->base); 137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central, 143 * This isn't exposed to users; we allow late enablement of the HPA shard so 146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard, [all...] |
| hpa.h | 89 /* Protects most of this shard's state. */ 92 * Guards the shard's access to the central allocator (preventing 93 * multiple threads operating on this shard from accessing the central 125 /* The configuration choices for this hpa shard. */ 130 * hpa shard. 135 * Those stats which are copied directly into the CTL-centric hpa shard 141 * Last time we performed purge on this shard. 153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap, 158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard, 162 * Notify the shard that we won't use it for allocations much longer. Due t [all...] |
| /src/external/bsd/jemalloc.old/include/jemalloc/internal/ |
| pa.h | 69 /* The central PA this shard is associated with. */ 114 /* The emap this shard is tied to. */ 122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) { 123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 && 124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0; 128 pa_shard_ehooks_get(pa_shard_t *shard) { 129 return base_ehooks_get(shard->base); 137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central, 143 * This isn't exposed to users; we allow late enablement of the HPA shard so 146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard, [all...] |
| hpa.h | 89 /* Protects most of this shard's state. */ 92 * Guards the shard's access to the central allocator (preventing 93 * multiple threads operating on this shard from accessing the central 125 /* The configuration choices for this hpa shard. */ 130 * hpa shard. 135 * Those stats which are copied directly into the CTL-centric hpa shard 141 * Last time we performed purge on this shard. 153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap, 158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard, 162 * Notify the shard that we won't use it for allocations much longer. Due t [all...] |
| /src/external/bsd/jemalloc/dist/test/unit/ |
| hpa.c | 16 hpa_shard_t shard; member in struct:test_data_s 58 err = hpa_shard_init(&test_data->shard, &test_data->central, 67 destroy_test_data(hpa_shard_t *shard) { 68 test_data_t *test_data = (test_data_t *)shard; 76 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local 84 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX, PAGE, false, false, 87 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX + PAGE, PAGE, false, 91 destroy_test_data(shard); 156 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local 191 edata_t *edata = pai_alloc(tsdn, &shard->pai 252 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local 377 hpa_shard_t *shard = create_test_data(&hooks, &opts); local [all...] |
| pa.c | 42 pa_shard_t shard; member in struct:test_data_s 73 err = pa_shard_init(TSDN_NULL, &test_data->shard, &test_data->central, 92 edata_t *edata = pa_alloc(TSDN_NULL, &test_data->shard, PAGE, 96 pa_dalloc(TSDN_NULL, &test_data->shard, edata, 99 &test_data->shard.pac.decay_dirty.mtx); 100 pac_decay_all(TSDN_NULL, &test_data->shard.pac, 101 &test_data->shard.pac.decay_dirty, 102 &test_data->shard.pac.stats->decay_dirty, 103 &test_data->shard.pac.ecache_dirty, true); 105 &test_data->shard.pac.decay_dirty.mtx) [all...] |
| /src/external/bsd/jemalloc.old/dist/test/unit/ |
| hpa.c | 16 hpa_shard_t shard; member in struct:test_data_s 58 err = hpa_shard_init(&test_data->shard, &test_data->central, 67 destroy_test_data(hpa_shard_t *shard) { 68 test_data_t *test_data = (test_data_t *)shard; 76 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local 84 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX, PAGE, false, false, 87 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX + PAGE, PAGE, false, 91 destroy_test_data(shard); 156 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local 191 edata_t *edata = pai_alloc(tsdn, &shard->pai 252 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local 377 hpa_shard_t *shard = create_test_data(&hooks, &opts); local [all...] |
| pa.c | 42 pa_shard_t shard; member in struct:test_data_s 73 err = pa_shard_init(TSDN_NULL, &test_data->shard, &test_data->central, 92 edata_t *edata = pa_alloc(TSDN_NULL, &test_data->shard, PAGE, 96 pa_dalloc(TSDN_NULL, &test_data->shard, edata, 99 &test_data->shard.pac.decay_dirty.mtx); 100 pac_decay_all(TSDN_NULL, &test_data->shard.pac, 101 &test_data->shard.pac.decay_dirty, 102 &test_data->shard.pac.stats->decay_dirty, 103 &test_data->shard.pac.ecache_dirty, true); 105 &test_data->shard.pac.decay_dirty.mtx) [all...] |
| /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/tests/ |
| sanitizer_thread_registry_test.cc | 163 uptr shard; // started from 1. member in struct:__sanitizer::__anon4147::RunThreadArgs 170 uptr shard = (uptr)arg; local 171 num_joined[shard]++; 174 uptr shard = (uptr)arg; local 175 num_started[shard]++; 178 uptr shard = (uptr)arg; local 179 num_created[shard]++; 190 args->registry->CreateThread(0, false, 0, (void*)args->shard)); 192 args->registry->StartThread(tids[i], 0, false, (void*)args->shard); 196 args->registry->JoinThread(tids[i], (void*)args->shard); [all...] |
| /src/external/gpl3/gdb/dist/gdb/dwarf2/ |
| cooked-index.c | 101 for (auto &shard : m_shards) 103 auto this_shard = shard.get (); 136 for (const auto &shard : m_shards) 138 dwarf2_per_cu *result = shard->lookup (addr); 153 for (const auto &shard : m_shards) 154 result.push_back (shard->m_addrmap); 166 for (auto &shard : m_shards) 167 result_range.push_back (shard->find (name, completing)); 191 for (const auto &shard : m_shards) 193 const cooked_index_entry *entry = shard->get_main () [all...] |
| cooked-index.h | 31 #include "dwarf2/cooked-index-shard.h" 125 for (auto &shard : m_shards) 126 result_range.push_back (shard->all_entries ());
|
| /src/external/apache2/llvm/dist/llvm/utils/lit/lit/ |
| cl_arguments.py | 180 selection_group.add_argument("--run-shard", 183 help="Run shard #N of the testsuite", 220 parser.error("--num-shards and --run-shard must be used together") 222 parser.error("--run-shard must be between 1 and --num-shards (inclusive)") 223 opts.shard = (opts.runShard, opts.numShards) 225 opts.shard = None
|
| /src/external/gpl3/gdb.old/dist/gdb/dwarf2/ |
| read-debug-names.c | 107 std::unique_ptr<cooked_index_shard> shard; member in struct:mapped_debug_names_reader 241 *result = shard->add (die_offset, (dwarf_tag) indexval.dwarf_tag, flags, 351 indexes.push_back (std::move (m_map.shard)); 768 map.shard = std::make_unique<cooked_index_shard> (); 769 map.shard->install_addrmap (&addrmap);
|