HomeSort by: relevance | last modified time | path
    Searched refs:shard (Results 1 - 25 of 55) sorted by relevancy

1 2 3

  /src/external/bsd/jemalloc/dist/src/
pa.c 8 pa_nactive_add(pa_shard_t *shard, size_t add_pages) {
9 atomic_fetch_add_zu(&shard->nactive, add_pages, ATOMIC_RELAXED);
13 pa_nactive_sub(pa_shard_t *shard, size_t sub_pages) {
14 assert(atomic_load_zu(&shard->nactive, ATOMIC_RELAXED) >= sub_pages);
15 atomic_fetch_sub_zu(&shard->nactive, sub_pages, ATOMIC_RELAXED);
32 pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central,
39 if (edata_cache_init(&shard->edata_cache, base)) {
43 if (pac_init(tsdn, &shard->pac, base, emap, &shard->edata_cache,
49 shard->ind = ind
    [all...]
pa_extra.c 12 pa_shard_prefork0(tsdn_t *tsdn, pa_shard_t *shard) {
13 malloc_mutex_prefork(tsdn, &shard->pac.decay_dirty.mtx);
14 malloc_mutex_prefork(tsdn, &shard->pac.decay_muzzy.mtx);
18 pa_shard_prefork2(tsdn_t *tsdn, pa_shard_t *shard) {
19 if (shard->ever_used_hpa) {
20 sec_prefork2(tsdn, &shard->hpa_sec);
25 pa_shard_prefork3(tsdn_t *tsdn, pa_shard_t *shard) {
26 malloc_mutex_prefork(tsdn, &shard->pac.grow_mtx);
27 if (shard->ever_used_hpa) {
28 hpa_shard_prefork3(tsdn, &shard->hpa_shard)
    [all...]
hpa.c 57 hpa_do_consistency_checks(hpa_shard_t *shard) {
58 assert(shard->base != NULL);
97 * shard is exhausted. We should hold the grow_mtx on that shard.
178 hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap,
184 err = malloc_mutex_init(&shard->grow_mtx, "hpa_shard_grow",
189 err = malloc_mutex_init(&shard->mtx, "hpa_shard",
196 shard->central = central;
197 shard->base = base;
198 edata_cache_fast_init(&shard->ecf, edata_cache)
724 hpa_shard_t *shard = hpa_from_pai(self); local
851 hpa_shard_t *shard = hpa_from_pai(self); local
888 hpa_shard_t *shard = hpa_from_pai(self); local
    [all...]
sec.c 45 sec_shard_t *shard = shard_cur; local
47 bool err = malloc_mutex_init(&shard->mtx, "sec_shard",
52 shard->enabled = true;
53 shard->bins = bin_cur;
55 sec_bin_init(&shard->bins[j]);
58 shard->bytes_cur = 0;
59 shard->to_flush_next = 0;
91 * Eventually, we should implement affinity, tracking source shard using
118 * empty cache, we'll try to fill it, which can push the shard over it's limit.
121 sec_flush_some_and_unlock(tsdn_t *tsdn, sec_t *sec, sec_shard_t *shard) {
237 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local
346 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local
    [all...]
  /src/external/bsd/jemalloc.old/dist/src/
pa.c 8 pa_nactive_add(pa_shard_t *shard, size_t add_pages) {
9 atomic_fetch_add_zu(&shard->nactive, add_pages, ATOMIC_RELAXED);
13 pa_nactive_sub(pa_shard_t *shard, size_t sub_pages) {
14 assert(atomic_load_zu(&shard->nactive, ATOMIC_RELAXED) >= sub_pages);
15 atomic_fetch_sub_zu(&shard->nactive, sub_pages, ATOMIC_RELAXED);
32 pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central,
39 if (edata_cache_init(&shard->edata_cache, base)) {
43 if (pac_init(tsdn, &shard->pac, base, emap, &shard->edata_cache,
49 shard->ind = ind
    [all...]
pa_extra.c 12 pa_shard_prefork0(tsdn_t *tsdn, pa_shard_t *shard) {
13 malloc_mutex_prefork(tsdn, &shard->pac.decay_dirty.mtx);
14 malloc_mutex_prefork(tsdn, &shard->pac.decay_muzzy.mtx);
18 pa_shard_prefork2(tsdn_t *tsdn, pa_shard_t *shard) {
19 if (shard->ever_used_hpa) {
20 sec_prefork2(tsdn, &shard->hpa_sec);
25 pa_shard_prefork3(tsdn_t *tsdn, pa_shard_t *shard) {
26 malloc_mutex_prefork(tsdn, &shard->pac.grow_mtx);
27 if (shard->ever_used_hpa) {
28 hpa_shard_prefork3(tsdn, &shard->hpa_shard)
    [all...]
hpa.c 57 hpa_do_consistency_checks(hpa_shard_t *shard) {
58 assert(shard->base != NULL);
97 * shard is exhausted. We should hold the grow_mtx on that shard.
178 hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap,
184 err = malloc_mutex_init(&shard->grow_mtx, "hpa_shard_grow",
189 err = malloc_mutex_init(&shard->mtx, "hpa_shard",
196 shard->central = central;
197 shard->base = base;
198 edata_cache_fast_init(&shard->ecf, edata_cache)
724 hpa_shard_t *shard = hpa_from_pai(self); local
851 hpa_shard_t *shard = hpa_from_pai(self); local
888 hpa_shard_t *shard = hpa_from_pai(self); local
    [all...]
sec.c 45 sec_shard_t *shard = shard_cur; local
47 bool err = malloc_mutex_init(&shard->mtx, "sec_shard",
52 shard->enabled = true;
53 shard->bins = bin_cur;
55 sec_bin_init(&shard->bins[j]);
58 shard->bytes_cur = 0;
59 shard->to_flush_next = 0;
91 * Eventually, we should implement affinity, tracking source shard using
118 * empty cache, we'll try to fill it, which can push the shard over it's limit.
121 sec_flush_some_and_unlock(tsdn_t *tsdn, sec_t *sec, sec_shard_t *shard) {
237 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local
346 sec_shard_t *shard = sec_shard_pick(tsdn, sec); local
    [all...]
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
pa.h 69 /* The central PA this shard is associated with. */
114 /* The emap this shard is tied to. */
122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) {
123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 &&
124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0;
128 pa_shard_ehooks_get(pa_shard_t *shard) {
129 return base_ehooks_get(shard->base);
137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central,
143 * This isn't exposed to users; we allow late enablement of the HPA shard so
146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard,
    [all...]
hpa.h 89 /* Protects most of this shard's state. */
92 * Guards the shard's access to the central allocator (preventing
93 * multiple threads operating on this shard from accessing the central
125 /* The configuration choices for this hpa shard. */
130 * hpa shard.
135 * Those stats which are copied directly into the CTL-centric hpa shard
141 * Last time we performed purge on this shard.
153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap,
158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard,
162 * Notify the shard that we won't use it for allocations much longer. Due t
    [all...]
  /src/external/bsd/jemalloc/include/jemalloc/internal/
pa.h 69 /* The central PA this shard is associated with. */
114 /* The emap this shard is tied to. */
122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) {
123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 &&
124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0;
128 pa_shard_ehooks_get(pa_shard_t *shard) {
129 return base_ehooks_get(shard->base);
137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central,
143 * This isn't exposed to users; we allow late enablement of the HPA shard so
146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard,
    [all...]
hpa.h 89 /* Protects most of this shard's state. */
92 * Guards the shard's access to the central allocator (preventing
93 * multiple threads operating on this shard from accessing the central
125 /* The configuration choices for this hpa shard. */
130 * hpa shard.
135 * Those stats which are copied directly into the CTL-centric hpa shard
141 * Last time we performed purge on this shard.
153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap,
158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard,
162 * Notify the shard that we won't use it for allocations much longer. Due t
    [all...]
  /src/external/bsd/jemalloc.old/dist/include/jemalloc/internal/
pa.h 69 /* The central PA this shard is associated with. */
114 /* The emap this shard is tied to. */
122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) {
123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 &&
124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0;
128 pa_shard_ehooks_get(pa_shard_t *shard) {
129 return base_ehooks_get(shard->base);
137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central,
143 * This isn't exposed to users; we allow late enablement of the HPA shard so
146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard,
    [all...]
hpa.h 89 /* Protects most of this shard's state. */
92 * Guards the shard's access to the central allocator (preventing
93 * multiple threads operating on this shard from accessing the central
125 /* The configuration choices for this hpa shard. */
130 * hpa shard.
135 * Those stats which are copied directly into the CTL-centric hpa shard
141 * Last time we performed purge on this shard.
153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap,
158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard,
162 * Notify the shard that we won't use it for allocations much longer. Due t
    [all...]
  /src/external/bsd/jemalloc.old/include/jemalloc/internal/
pa.h 69 /* The central PA this shard is associated with. */
114 /* The emap this shard is tied to. */
122 pa_shard_dont_decay_muzzy(pa_shard_t *shard) {
123 return ecache_npages_get(&shard->pac.ecache_muzzy) == 0 &&
124 pac_decay_ms_get(&shard->pac, extent_state_muzzy) <= 0;
128 pa_shard_ehooks_get(pa_shard_t *shard) {
129 return base_ehooks_get(shard->base);
137 bool pa_shard_init(tsdn_t *tsdn, pa_shard_t *shard, pa_central_t *central,
143 * This isn't exposed to users; we allow late enablement of the HPA shard so
146 bool pa_shard_enable_hpa(tsdn_t *tsdn, pa_shard_t *shard,
    [all...]
hpa.h 89 /* Protects most of this shard's state. */
92 * Guards the shard's access to the central allocator (preventing
93 * multiple threads operating on this shard from accessing the central
125 /* The configuration choices for this hpa shard. */
130 * hpa shard.
135 * Those stats which are copied directly into the CTL-centric hpa shard
141 * Last time we performed purge on this shard.
153 bool hpa_shard_init(hpa_shard_t *shard, hpa_central_t *central, emap_t *emap,
158 void hpa_shard_stats_merge(tsdn_t *tsdn, hpa_shard_t *shard,
162 * Notify the shard that we won't use it for allocations much longer. Due t
    [all...]
  /src/external/bsd/jemalloc/dist/test/unit/
hpa.c 16 hpa_shard_t shard; member in struct:test_data_s
58 err = hpa_shard_init(&test_data->shard, &test_data->central,
67 destroy_test_data(hpa_shard_t *shard) {
68 test_data_t *test_data = (test_data_t *)shard;
76 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local
84 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX, PAGE, false, false,
87 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX + PAGE, PAGE, false,
91 destroy_test_data(shard);
156 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local
191 edata_t *edata = pai_alloc(tsdn, &shard->pai
252 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local
377 hpa_shard_t *shard = create_test_data(&hooks, &opts); local
    [all...]
pa.c 42 pa_shard_t shard; member in struct:test_data_s
73 err = pa_shard_init(TSDN_NULL, &test_data->shard, &test_data->central,
92 edata_t *edata = pa_alloc(TSDN_NULL, &test_data->shard, PAGE,
96 pa_dalloc(TSDN_NULL, &test_data->shard, edata,
99 &test_data->shard.pac.decay_dirty.mtx);
100 pac_decay_all(TSDN_NULL, &test_data->shard.pac,
101 &test_data->shard.pac.decay_dirty,
102 &test_data->shard.pac.stats->decay_dirty,
103 &test_data->shard.pac.ecache_dirty, true);
105 &test_data->shard.pac.decay_dirty.mtx)
    [all...]
  /src/external/bsd/jemalloc.old/dist/test/unit/
hpa.c 16 hpa_shard_t shard; member in struct:test_data_s
58 err = hpa_shard_init(&test_data->shard, &test_data->central,
67 destroy_test_data(hpa_shard_t *shard) {
68 test_data_t *test_data = (test_data_t *)shard;
76 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local
84 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX, PAGE, false, false,
87 edata = pai_alloc(tsdn, &shard->pai, ALLOC_MAX + PAGE, PAGE, false,
91 destroy_test_data(shard);
156 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local
191 edata_t *edata = pai_alloc(tsdn, &shard->pai
252 hpa_shard_t *shard = create_test_data(&hpa_hooks_default, local
377 hpa_shard_t *shard = create_test_data(&hooks, &opts); local
    [all...]
pa.c 42 pa_shard_t shard; member in struct:test_data_s
73 err = pa_shard_init(TSDN_NULL, &test_data->shard, &test_data->central,
92 edata_t *edata = pa_alloc(TSDN_NULL, &test_data->shard, PAGE,
96 pa_dalloc(TSDN_NULL, &test_data->shard, edata,
99 &test_data->shard.pac.decay_dirty.mtx);
100 pac_decay_all(TSDN_NULL, &test_data->shard.pac,
101 &test_data->shard.pac.decay_dirty,
102 &test_data->shard.pac.stats->decay_dirty,
103 &test_data->shard.pac.ecache_dirty, true);
105 &test_data->shard.pac.decay_dirty.mtx)
    [all...]
  /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/tests/
sanitizer_thread_registry_test.cc 163 uptr shard; // started from 1. member in struct:__sanitizer::__anon4147::RunThreadArgs
170 uptr shard = (uptr)arg; local
171 num_joined[shard]++;
174 uptr shard = (uptr)arg; local
175 num_started[shard]++;
178 uptr shard = (uptr)arg; local
179 num_created[shard]++;
190 args->registry->CreateThread(0, false, 0, (void*)args->shard));
192 args->registry->StartThread(tids[i], 0, false, (void*)args->shard);
196 args->registry->JoinThread(tids[i], (void*)args->shard);
    [all...]
  /src/external/gpl3/gdb/dist/gdb/dwarf2/
cooked-index.c 101 for (auto &shard : m_shards)
103 auto this_shard = shard.get ();
136 for (const auto &shard : m_shards)
138 dwarf2_per_cu *result = shard->lookup (addr);
153 for (const auto &shard : m_shards)
154 result.push_back (shard->m_addrmap);
166 for (auto &shard : m_shards)
167 result_range.push_back (shard->find (name, completing));
191 for (const auto &shard : m_shards)
193 const cooked_index_entry *entry = shard->get_main ()
    [all...]
cooked-index.h 31 #include "dwarf2/cooked-index-shard.h"
125 for (auto &shard : m_shards)
126 result_range.push_back (shard->all_entries ());
  /src/external/apache2/llvm/dist/llvm/utils/lit/lit/
cl_arguments.py 180 selection_group.add_argument("--run-shard",
183 help="Run shard #N of the testsuite",
220 parser.error("--num-shards and --run-shard must be used together")
222 parser.error("--run-shard must be between 1 and --num-shards (inclusive)")
223 opts.shard = (opts.runShard, opts.numShards)
225 opts.shard = None
  /src/external/gpl3/gdb.old/dist/gdb/dwarf2/
read-debug-names.c 107 std::unique_ptr<cooked_index_shard> shard; member in struct:mapped_debug_names_reader
241 *result = shard->add (die_offset, (dwarf_tag) indexval.dwarf_tag, flags,
351 indexes.push_back (std::move (m_map.shard));
768 map.shard = std::make_unique<cooked_index_shard> ();
769 map.shard->install_addrmap (&addrmap);

Completed in 21 milliseconds

1 2 3