| /src/external/bsd/openldap/dist/servers/lloadd/ |
| epoch.c | 1 /* $NetBSD: epoch.c,v 1.3 2025/09/05 21:16:24 christos Exp $ */ 3 /* epoch.c - epoch based memory reclamation */ 19 /** @file epoch.c 21 * Implementation of epoch based memory reclamation, in principle 27 * Also the problems with epoch based memory reclamation are still 28 * present - a thread actively observing an epoch getting stuck will 34 __RCSID("$NetBSD: epoch.c,v 1.3 2025/09/05 21:16:24 christos Exp $"); 39 #include <epoch.h> 43 #define EPOCH_PREV(epoch) ( ( (epoch) + EPOCH_MASK - 1 ) % EPOCH_MASK 61 epoch_t epoch; local 75 epoch_t epoch; local 117 epoch_t epoch; local 249 epoch_t epoch = __atomic_load_n( ¤t_epoch, __ATOMIC_ACQUIRE ); local [all...] |
| connection.c | 78 epoch_t epoch; local 83 epoch = epoch_join(); 146 epoch_leave( epoch ); 147 epoch = epoch_join(); 165 epoch_leave( epoch ); 183 epoch_t epoch; local 199 epoch = epoch_join(); 201 epoch_leave( epoch ); 208 epoch = epoch_join(); 295 epoch_leave( epoch ); 307 epoch_t epoch; local [all...] |
| epoch.h | 1 /* $NetBSD: epoch.h,v 1.3 2025/09/05 21:16:24 christos Exp $ */ 3 /* epoch.h - epoch based memory reclamation */ 22 /** @file epoch.h 24 * Implementation of epoch based memory reclamation, in principle 51 * To simplify locking, memory is only freed when the current epoch 57 * @return The observed epoch, to be passed to #epoch_leave() 69 * @param[in] epoch Epoch identifier returned by a previous call to 72 void epoch_leave( epoch_t epoch ); [all...] |
| tier.c | 55 epoch_t epoch = epoch_join(); local 65 epoch_leave( epoch ); 75 epoch_t epoch = epoch_join(); local 79 epoch_leave( epoch );
|
| /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
| tsan_mutexset.cc | 25 void MutexSet::Add(u64 id, bool write, u64 epoch) { 30 descs_[i].epoch = epoch; 39 if (descs_[i].epoch < minepoch) { 40 minepoch = descs_[i].epoch; 50 descs_[size_].epoch = epoch;
|
| tsan_mutexset.h | 28 u64 epoch; member in struct:__tsan::MutexSet::Desc 35 void Add(u64 id, bool write, u64 epoch); 60 void MutexSet::Add(u64 id, bool write, u64 epoch) {}
|
| tsan_clock.cc | 147 if (clk_[tid] < dirty.epoch) { 148 clk_[tid] = dirty.epoch; 161 u64 epoch = src_elem.epoch; local 162 if (*dst_pos < epoch) { 163 *dst_pos = epoch; 200 if (dst->elem(tid_).epoch > last_acquire_) { 219 ce.epoch = max(ce.epoch, clk_[i]); 257 dst->dirty_[0].epoch = clk_[tid_] [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/ |
| i915_globals.c | 28 static atomic_t epoch; variable 34 int epoch; member in struct:park_work 58 park.epoch = atomic_inc_return(&epoch); 70 if (park.epoch != atomic_read(&epoch)) { 145 atomic_inc(&epoch);
|
| /src/external/bsd/ntp/dist/scripts/monitoring/ |
| timelocal.pl | 27 @epoch = localtime(0); 28 $tzmin = $epoch[2] * 60 + $epoch[1]; # minutes east of GMT 31 $tzmin -= 24 * 60 if $epoch[5] == 70; # account for the date line 65 $year += $YearFix if $year < $epoch[5];
|
| /src/external/bsd/jemalloc.old/dist/test/unit/ |
| stats.c | 30 uint64_t epoch; local 39 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), 70 uint64_t epoch; local 90 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), 146 uint64_t epoch, nmalloc, ndalloc, nrequests; local 157 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)) 191 uint64_t epoch, nmalloc, ndalloc; local 231 uint64_t epoch, nmalloc, ndalloc, nrequests, nfills, nflushes; local 319 uint64_t epoch, nmalloc, ndalloc; local [all...] |
| retained.c | 10 static atomic_u_t epoch; variable 38 uint64_t epoch = 1; local 39 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, 40 sizeof(epoch)), 0, "Unexpected mallctl() failure"); 72 /* Busy-wait for next epoch. */ 75 while ((cur_epoch = atomic_load_u(&epoch, ATOMIC_ACQUIRE)) != 79 assert_u_eq(cur_epoch, next_epoch, "Unexpected epoch"); 107 atomic_store_u(&epoch, 0, ATOMIC_RELAXED); 117 atomic_store_u(&epoch, e, ATOMIC_RELEASE) [all...] |
| /src/external/bsd/jemalloc/dist/test/unit/ |
| stats.c | 33 uint64_t epoch; local 42 expect_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), 73 uint64_t epoch; local 94 expect_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)), 150 uint64_t epoch, nmalloc, ndalloc, nrequests; local 161 expect_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)) 195 uint64_t epoch, nmalloc, ndalloc; local 235 uint64_t epoch, nmalloc, ndalloc, nrequests, nfills, nflushes; local 328 uint64_t epoch, nmalloc, ndalloc; local 369 uint64_t epoch; local [all...] |
| hpa_background_thread.c | 29 uint64_t epoch = 1; local 30 size_t sz = sizeof(epoch); 31 err = je_mallctl("epoch", (void *)&epoch, &sz, (void *)&epoch, 32 sizeof(epoch));
|
| tcache_max.c | 62 uint64_t epoch; local 63 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)),
|
| /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
| tsan_mutexset.cpp | 22 void MutexSet::Add(u64 id, bool write, u64 epoch) { 27 descs_[i].epoch = epoch; 36 if (descs_[i].epoch < minepoch) { 37 minepoch = descs_[i].epoch; 49 descs_[size_].epoch = epoch; 98 descs_[size_].epoch = 0;
|
| tsan_clock.cpp | 144 if (clk_[tid] < dirty.epoch) { 145 clk_[tid] = dirty.epoch; 157 u64 epoch = src_elem.epoch; local 158 if (*dst_pos < epoch) { 159 *dst_pos = epoch; 201 if (clk_[i] < ce.epoch) { 202 clk_[i] = ce.epoch; 205 ce.epoch = tmp; 252 ce.epoch = max(ce.epoch, clk_[i]) [all...] |
| tsan_mutexset.h | 29 u64 epoch; member in struct:__tsan::MutexSet::Desc 44 void Add(u64 id, bool write, u64 epoch); 85 void MutexSet::Add(u64 id, bool write, u64 epoch) {}
|
| tsan_shadow.h | 22 // epoch : kClkBits 25 FastState(u64 tid, u64 epoch) { 27 x_ |= epoch; 29 DCHECK_EQ(epoch, this->epoch()); 47 u64 epoch() const { function in class:__tsan::FastState 53 u64 old_epoch = epoch(); 55 DCHECK_EQ(old_epoch + 1, epoch()); 81 return epoch() & mask; 101 // epoch : kClkBit [all...] |
| /src/crypto/external/bsd/heimdal/dist/kdc/ |
| rx.h | 64 uint32_t epoch; member in struct:rx_header
|
| /src/sys/external/bsd/common/linux/ |
| linux_srcu.c | 45 * All new srcu read sections get counted in the active epoch. 47 * epoch has zero readers. When a thread calls synchronize_srcu, 50 * number of readers in the now-draining epoch, and waits for the 58 * epoch. 135 unsigned epoch = gen & 1; /* active epoch */ local 138 cpu->src_count[epoch] += delta; 232 * readers on this CPU in the inactive epoch to the global count 240 unsigned gen, epoch; local 247 epoch = 1 ^ (gen & 1); /* draining epoch * [all...] |
| /src/crypto/external/apache2/openssl/dist/util/perl/TLSProxy/ |
| Record.pm | 83 my $epoch; 91 ($content_type, $version, $epoch, 110 print " Epoch: $epoch\n"; 121 $epoch, 215 $epoch, 227 $epoch, 254 0, #epoch 271 $epoch, 285 epoch => $epoch 488 sub epoch subroutine [all...] |
| /src/sys/external/bsd/compiler_rt/dist/lib/tsan/tests/unit/ |
| tsan_shadow_test.cc | 22 EXPECT_EQ(s.epoch(), (u64)22); 31 EXPECT_EQ(s.epoch(), (u64)23); 33 EXPECT_EQ(s.epoch(), (u64)24);
|
| /src/sys/kern/ |
| subr_cprng.c | 251 /* Set the epoch uninitialized so we reseed on first use. */ 269 cprng_strong_reseed(struct cprng_strong *cprng, unsigned epoch, 280 * big deal -- worst case, we rewind the entropy epoch here and 295 (*ccp)->cc_epoch = epoch; 302 unsigned epoch; local 323 /* If the entropy epoch has changed, (re)seed. */ 324 epoch = entropy_epoch(); 325 if (__predict_false(epoch != cc->cc_epoch)) 326 cprng_strong_reseed(cprng, epoch, &cc, &s); 331 cprng_strong_reseed(cprng, epoch, &cc, &s) [all...] |
| /src/external/bsd/jemalloc/dist/src/ |
| decay.c | 14 * Generate a new deadline that is uniformly random within the next epoch after 19 nstime_copy(&decay->deadline, &decay->epoch); 39 nstime_copy(&decay->epoch, cur_time); 77 if (unlikely(!nstime_monotonic() && nstime_compare(&decay->epoch, 80 * Time went backwards. Move the epoch back in time and 84 * to clock jitter triggering premature epoch advances, but 88 nstime_copy(&decay->epoch, new_time); 92 assert(nstime_compare(&decay->epoch, new_time) <= 0); 187 nstime_subtract(&delta, &decay->epoch); 192 /* Add nadvance_u64 decay intervals to epoch. * [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
| tsan_shadow.h | 32 Epoch epoch() const { return static_cast<Epoch>(part_.epoch_); } function in class:__tsan::FastState 34 void SetEpoch(Epoch epoch) { part_.epoch_ = static_cast<u16>(epoch); } 74 DCHECK_EQ(epoch(), epoch0); 81 Epoch epoch() const { return static_cast<Epoch>(part_.epoch_); function in class:__tsan::Shadow [all...] |