HomeSort by: relevance | last modified time | path
    Searched refs:epoch (Results 1 - 25 of 147) sorted by relevancy

1 2 3 4 5 6

  /src/external/bsd/openldap/dist/servers/lloadd/
epoch.c 1 /* $NetBSD: epoch.c,v 1.3 2025/09/05 21:16:24 christos Exp $ */
3 /* epoch.c - epoch based memory reclamation */
19 /** @file epoch.c
21 * Implementation of epoch based memory reclamation, in principle
27 * Also the problems with epoch based memory reclamation are still
28 * present - a thread actively observing an epoch getting stuck will
34 __RCSID("$NetBSD: epoch.c,v 1.3 2025/09/05 21:16:24 christos Exp $");
39 #include <epoch.h>
43 #define EPOCH_PREV(epoch) ( ( (epoch) + EPOCH_MASK - 1 ) % EPOCH_MASK
61 epoch_t epoch; local
75 epoch_t epoch; local
117 epoch_t epoch; local
249 epoch_t epoch = __atomic_load_n( &current_epoch, __ATOMIC_ACQUIRE ); local
    [all...]
connection.c 78 epoch_t epoch; local
83 epoch = epoch_join();
146 epoch_leave( epoch );
147 epoch = epoch_join();
165 epoch_leave( epoch );
183 epoch_t epoch; local
199 epoch = epoch_join();
201 epoch_leave( epoch );
208 epoch = epoch_join();
295 epoch_leave( epoch );
307 epoch_t epoch; local
    [all...]
epoch.h 1 /* $NetBSD: epoch.h,v 1.3 2025/09/05 21:16:24 christos Exp $ */
3 /* epoch.h - epoch based memory reclamation */
22 /** @file epoch.h
24 * Implementation of epoch based memory reclamation, in principle
51 * To simplify locking, memory is only freed when the current epoch
57 * @return The observed epoch, to be passed to #epoch_leave()
69 * @param[in] epoch Epoch identifier returned by a previous call to
72 void epoch_leave( epoch_t epoch );
    [all...]
tier.c 55 epoch_t epoch = epoch_join(); local
65 epoch_leave( epoch );
75 epoch_t epoch = epoch_join(); local
79 epoch_leave( epoch );
  /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/
tsan_mutexset.cc 25 void MutexSet::Add(u64 id, bool write, u64 epoch) {
30 descs_[i].epoch = epoch;
39 if (descs_[i].epoch < minepoch) {
40 minepoch = descs_[i].epoch;
50 descs_[size_].epoch = epoch;
tsan_mutexset.h 28 u64 epoch; member in struct:__tsan::MutexSet::Desc
35 void Add(u64 id, bool write, u64 epoch);
60 void MutexSet::Add(u64 id, bool write, u64 epoch) {}
tsan_clock.cc 147 if (clk_[tid] < dirty.epoch) {
148 clk_[tid] = dirty.epoch;
161 u64 epoch = src_elem.epoch; local
162 if (*dst_pos < epoch) {
163 *dst_pos = epoch;
200 if (dst->elem(tid_).epoch > last_acquire_) {
219 ce.epoch = max(ce.epoch, clk_[i]);
257 dst->dirty_[0].epoch = clk_[tid_]
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/
i915_globals.c 28 static atomic_t epoch; variable
34 int epoch; member in struct:park_work
58 park.epoch = atomic_inc_return(&epoch);
70 if (park.epoch != atomic_read(&epoch)) {
145 atomic_inc(&epoch);
  /src/external/bsd/ntp/dist/scripts/monitoring/
timelocal.pl 27 @epoch = localtime(0);
28 $tzmin = $epoch[2] * 60 + $epoch[1]; # minutes east of GMT
31 $tzmin -= 24 * 60 if $epoch[5] == 70; # account for the date line
65 $year += $YearFix if $year < $epoch[5];
  /src/external/bsd/jemalloc.old/dist/test/unit/
stats.c 30 uint64_t epoch; local
39 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)),
70 uint64_t epoch; local
90 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)),
146 uint64_t epoch, nmalloc, ndalloc, nrequests; local
157 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch))
191 uint64_t epoch, nmalloc, ndalloc; local
231 uint64_t epoch, nmalloc, ndalloc, nrequests, nfills, nflushes; local
319 uint64_t epoch, nmalloc, ndalloc; local
    [all...]
retained.c 10 static atomic_u_t epoch; variable
38 uint64_t epoch = 1; local
39 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch,
40 sizeof(epoch)), 0, "Unexpected mallctl() failure");
72 /* Busy-wait for next epoch. */
75 while ((cur_epoch = atomic_load_u(&epoch, ATOMIC_ACQUIRE)) !=
79 assert_u_eq(cur_epoch, next_epoch, "Unexpected epoch");
107 atomic_store_u(&epoch, 0, ATOMIC_RELAXED);
117 atomic_store_u(&epoch, e, ATOMIC_RELEASE)
    [all...]
  /src/external/bsd/jemalloc/dist/test/unit/
stats.c 33 uint64_t epoch; local
42 expect_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)),
73 uint64_t epoch; local
94 expect_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)),
150 uint64_t epoch, nmalloc, ndalloc, nrequests; local
161 expect_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch))
195 uint64_t epoch, nmalloc, ndalloc; local
235 uint64_t epoch, nmalloc, ndalloc, nrequests, nfills, nflushes; local
328 uint64_t epoch, nmalloc, ndalloc; local
369 uint64_t epoch; local
    [all...]
hpa_background_thread.c 29 uint64_t epoch = 1; local
30 size_t sz = sizeof(epoch);
31 err = je_mallctl("epoch", (void *)&epoch, &sz, (void *)&epoch,
32 sizeof(epoch));
tcache_max.c 62 uint64_t epoch; local
63 assert_d_eq(mallctl("epoch", NULL, NULL, (void *)&epoch, sizeof(epoch)),
  /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/
tsan_mutexset.cpp 22 void MutexSet::Add(u64 id, bool write, u64 epoch) {
27 descs_[i].epoch = epoch;
36 if (descs_[i].epoch < minepoch) {
37 minepoch = descs_[i].epoch;
49 descs_[size_].epoch = epoch;
98 descs_[size_].epoch = 0;
tsan_clock.cpp 144 if (clk_[tid] < dirty.epoch) {
145 clk_[tid] = dirty.epoch;
157 u64 epoch = src_elem.epoch; local
158 if (*dst_pos < epoch) {
159 *dst_pos = epoch;
201 if (clk_[i] < ce.epoch) {
202 clk_[i] = ce.epoch;
205 ce.epoch = tmp;
252 ce.epoch = max(ce.epoch, clk_[i])
    [all...]
tsan_mutexset.h 29 u64 epoch; member in struct:__tsan::MutexSet::Desc
44 void Add(u64 id, bool write, u64 epoch);
85 void MutexSet::Add(u64 id, bool write, u64 epoch) {}
tsan_shadow.h 22 // epoch : kClkBits
25 FastState(u64 tid, u64 epoch) {
27 x_ |= epoch;
29 DCHECK_EQ(epoch, this->epoch());
47 u64 epoch() const { function in class:__tsan::FastState
53 u64 old_epoch = epoch();
55 DCHECK_EQ(old_epoch + 1, epoch());
81 return epoch() & mask;
101 // epoch : kClkBit
    [all...]
  /src/crypto/external/bsd/heimdal/dist/kdc/
rx.h 64 uint32_t epoch; member in struct:rx_header
  /src/sys/external/bsd/common/linux/
linux_srcu.c 45 * All new srcu read sections get counted in the active epoch.
47 * epoch has zero readers. When a thread calls synchronize_srcu,
50 * number of readers in the now-draining epoch, and waits for the
58 * epoch.
135 unsigned epoch = gen & 1; /* active epoch */ local
138 cpu->src_count[epoch] += delta;
232 * readers on this CPU in the inactive epoch to the global count
240 unsigned gen, epoch; local
247 epoch = 1 ^ (gen & 1); /* draining epoch *
    [all...]
  /src/crypto/external/apache2/openssl/dist/util/perl/TLSProxy/
Record.pm 83 my $epoch;
91 ($content_type, $version, $epoch,
110 print " Epoch: $epoch\n";
121 $epoch,
215 $epoch,
227 $epoch,
254 0, #epoch
271 $epoch,
285 epoch => $epoch
488 sub epoch subroutine
    [all...]
  /src/sys/external/bsd/compiler_rt/dist/lib/tsan/tests/unit/
tsan_shadow_test.cc 22 EXPECT_EQ(s.epoch(), (u64)22);
31 EXPECT_EQ(s.epoch(), (u64)23);
33 EXPECT_EQ(s.epoch(), (u64)24);
  /src/sys/kern/
subr_cprng.c 251 /* Set the epoch uninitialized so we reseed on first use. */
269 cprng_strong_reseed(struct cprng_strong *cprng, unsigned epoch,
280 * big deal -- worst case, we rewind the entropy epoch here and
295 (*ccp)->cc_epoch = epoch;
302 unsigned epoch; local
323 /* If the entropy epoch has changed, (re)seed. */
324 epoch = entropy_epoch();
325 if (__predict_false(epoch != cc->cc_epoch))
326 cprng_strong_reseed(cprng, epoch, &cc, &s);
331 cprng_strong_reseed(cprng, epoch, &cc, &s)
    [all...]
  /src/external/bsd/jemalloc/dist/src/
decay.c 14 * Generate a new deadline that is uniformly random within the next epoch after
19 nstime_copy(&decay->deadline, &decay->epoch);
39 nstime_copy(&decay->epoch, cur_time);
77 if (unlikely(!nstime_monotonic() && nstime_compare(&decay->epoch,
80 * Time went backwards. Move the epoch back in time and
84 * to clock jitter triggering premature epoch advances, but
88 nstime_copy(&decay->epoch, new_time);
92 assert(nstime_compare(&decay->epoch, new_time) <= 0);
187 nstime_subtract(&delta, &decay->epoch);
192 /* Add nadvance_u64 decay intervals to epoch. *
    [all...]
  /src/external/gpl3/gcc/dist/libsanitizer/tsan/
tsan_shadow.h 32 Epoch epoch() const { return static_cast<Epoch>(part_.epoch_); } function in class:__tsan::FastState
34 void SetEpoch(Epoch epoch) { part_.epoch_ = static_cast<u16>(epoch); }
74 DCHECK_EQ(epoch(), epoch0);
81 Epoch epoch() const { return static_cast<Epoch>(part_.epoch_); function in class:__tsan::Shadow
    [all...]

Completed in 42 milliseconds

1 2 3 4 5 6