HomeSort by: relevance | last modified time | path
    Searched defs:locked (Results 1 - 25 of 79) sorted by relevancy

1 2 3 4

  /src/external/ibm-public/postfix/dist/src/global/
mbox_open.h 34 int locked; /* what locks were set */ member in struct:__anon23117
mbox_open.c 114 int locked = 0; local
144 locked |= MBOX_DOT_LOCK;
168 if (locked & MBOX_DOT_LOCK)
186 locked |= lock_style;
189 if (locked & MBOX_DOT_LOCK)
212 if (locked & MBOX_DOT_LOCK)
220 mp->locked = locked;
237 if (mp->locked & MBOX_DOT_LOCK)
  /src/sys/external/bsd/drm2/dist/include/drm/
drm_modeset_lock.h 38 * @locked: list of held locks
60 struct list_head locked; member in struct:drm_modeset_acquire_ctx
74 * @head: used to hold its place on &drm_atomi_state.locked list when
86 * Resources that are locked as part of an atomic update are added
  /src/sys/external/isc/libsodium/dist/src/libsodium/sodium/
core.c 40 static volatile int locked; variable
103 assert(locked == 0);
104 locked = 1;
112 if (locked == 0) {
118 locked = 0;
134 assert(locked == 0);
135 locked = 1;
145 if (locked == 0) {
151 locked = 0;
  /src/external/bsd/openldap/dist/servers/slapd/back-monitor/
search.c 105 Entry *sub_nv = NULL, *sub_ch = NULL, *locked = e; local
139 monitor_cache_release( mi, locked );
140 locked = NULL;
147 if ( locked ) {
148 monitor_cache_release( mi, locked );
167 if ( locked ) {
168 monitor_cache_release( mi, locked );
  /src/lib/libpthread/
pthread_lock.c 80 int locked; local
83 locked = __SIMPLELOCK_LOCKED_P(alp);
87 return !locked;
  /src/sys/external/bsd/drm2/dist/drm/vmwgfx/
ttm_lock.c 68 bool locked = false; local
72 locked = true;
74 return locked;
93 static bool __ttm_read_trylock(struct ttm_lock *lock, bool *locked)
97 *locked = false;
103 *locked = true;
115 bool locked; local
120 __ttm_read_trylock(lock, &locked));
123 __ttm_read_trylock(lock, &locked));
127 BUG_ON(locked);
144 bool locked = false; local
188 bool locked = false; local
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/engine/disp/
nouveau_nvkm_engine_disp_vga.c 135 bool locked = !nvkm_rdvgac(device, 0, 0x1f); local
145 return locked;
  /src/external/bsd/tmux/dist/
cmd-wait-for.c 51 int locked; member in struct:wait_channel
91 wc->locked = 0;
107 if (wc->locked)
209 if (wc->locked) {
215 wc->locked = 1;
226 if (wc == NULL || !wc->locked) {
227 cmdq_error(item, "channel %s not locked", name);
236 wc->locked = 0;
261 wc->locked = 0;
  /src/external/gpl3/gcc/dist/libbacktrace/
mmap.c 119 int locked; local
133 locked = 1;
135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
137 if (locked)
199 int locked; local
227 locked = 1;
229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
231 if (locked)
  /src/external/gpl3/gcc.old/dist/libbacktrace/
mmap.c 119 int locked; local
133 locked = 1;
135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
137 if (locked)
199 int locked; local
227 locked = 1;
229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
231 if (locked)
  /src/external/gpl3/gdb/dist/libbacktrace/
mmap.c 119 int locked; local
133 locked = 1;
135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
137 if (locked)
199 int locked; local
227 locked = 1;
229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
231 if (locked)
  /src/external/gpl3/gdb.old/dist/libbacktrace/
mmap.c 119 int locked; local
133 locked = 1;
135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
137 if (locked)
199 int locked; local
227 locked = 1;
229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0;
231 if (locked)
  /src/sys/arch/hppa/hppa/
intr.c 411 bool locked = false; local
458 locked = true;
469 if (locked) {
471 locked = false;
  /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
amdgpu_mxgpu_ai.c 246 int locked; local
256 locked = mutex_trylock(&adev->lock_reset);
257 if (locked)
269 if (locked) {
amdgpu_mxgpu_nv.c 248 int locked; local
258 locked = mutex_trylock(&adev->lock_reset);
259 if (locked)
271 if (locked) {
  /src/sys/external/bsd/drm2/linux/
linux_ww_mutex.c 239 * True if anyone holds mutex locked at the moment, false if not.
243 * XXX WARNING: This returns true if it is locked by ANYONE. Does
250 int locked; local
255 locked = false;
260 locked = true;
268 return locked;
345 * With mutex locked and in the WW_CTX or WW_WANTOWN state, owned
397 * With mutex locked and in the WW_CTX or WW_WANTOWN state, owned
611 goto locked;
667 locked: KASSERT((mutex->wwm_state == WW_CTX) |
    [all...]
  /src/sys/kern/
vfs_getcwd.c 89 * On entry, *lvpp is a locked vnode reference; on exit, it is vput and NULL'ed
90 * On exit, *uvpp is either NULL or is a locked vnode reference.
117 * current directory is still locked.
138 /* At this point, lvp is locked */
321 bool locked = false; local
375 locked = true;
385 if (!locked) {
386 locked = true;
396 } else if (locked) {
  /src/external/bsd/jemalloc/dist/include/jemalloc/internal/
mutex.h 52 atomic_b_t locked; member in struct:malloc_mutex_s::__anon6753::__anon6754
152 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED);
178 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED);
218 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED);
227 atomic_store_b(&mutex->locked, false, ATOMIC_RELAXED);
  /src/external/bsd/jemalloc/include/jemalloc/internal/
mutex.h 52 atomic_b_t locked; member in struct:malloc_mutex_s::__anon6832::__anon6833
152 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED);
178 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED);
218 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED);
227 atomic_store_b(&mutex->locked, false, ATOMIC_RELAXED);
  /src/external/bsd/ppp/usr.sbin/pppd/
tdb.h 45 int *locked; /* set if we have a chain locked */ member in struct:__anon7996
  /src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/
sanitizer_mutex.cpp 55 // The table defines a static mutex type hierarchy (what mutex types can be locked
67 // Adjacency matrix of what mutexes can be locked under what mutexes.
148 LockDesc locked[kMutexTypeMax]; member in struct:__sanitizer::InternalDeadlockDetector
154 // Find the last locked mutex type.
159 if (locked[i].seq == 0)
161 CHECK_NE(locked[i].seq, max_seq);
162 if (max_seq < locked[i].seq) {
163 max_seq = locked[i].seq;
169 CHECK_EQ(locked[type].seq, max_seq);
170 CHECK(locked[type].pc)
    [all...]
  /src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/
sanitizer_mutex.cpp 55 // The table defines a static mutex type hierarchy (what mutex types can be locked
67 // Adjacency matrix of what mutexes can be locked under what mutexes.
148 LockDesc locked[kMutexTypeMax]; member in struct:__sanitizer::InternalDeadlockDetector
154 // Find the last locked mutex type.
159 if (locked[i].seq == 0)
161 CHECK_NE(locked[i].seq, max_seq);
162 if (max_seq < locked[i].seq) {
163 max_seq = locked[i].seq;
169 CHECK_EQ(locked[type].seq, max_seq);
170 CHECK(locked[type].pc)
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/
i915_gem_userptr.c 504 int locked = 0; local
512 if (!locked) {
514 locked = 1;
521 pvec + pinned, NULL, &locked);
527 if (locked)
  /src/sys/uvm/
uvm_loan.c 81 * must be locked. to modify pg->loan_count, both the owner of the page
82 * and pg->interlock must be locked. pg->flags is (as always) locked by
88 * from dying pg->interlock should be locked. this forces us to sometimes
123 * on entry the map is locked by the caller)
170 /* locked: map, amap, uobj */
181 /* locked: if (rv > 0) => map, amap, uobj [o.w. unlocked] */
210 * unlock what we locked, unlock the maps and return
285 * map now locked. now do the loanout...
334 * => called with map, amap, uobj locked
604 bool locked; local
    [all...]

Completed in 47 milliseconds

1 2 3 4