| /src/external/ibm-public/postfix/dist/src/global/ |
| mbox_open.h | 34 int locked; /* what locks were set */ member in struct:__anon23117
|
| mbox_open.c | 114 int locked = 0; local 144 locked |= MBOX_DOT_LOCK; 168 if (locked & MBOX_DOT_LOCK) 186 locked |= lock_style; 189 if (locked & MBOX_DOT_LOCK) 212 if (locked & MBOX_DOT_LOCK) 220 mp->locked = locked; 237 if (mp->locked & MBOX_DOT_LOCK)
|
| /src/sys/external/bsd/drm2/dist/include/drm/ |
| drm_modeset_lock.h | 38 * @locked: list of held locks 60 struct list_head locked; member in struct:drm_modeset_acquire_ctx 74 * @head: used to hold its place on &drm_atomi_state.locked list when 86 * Resources that are locked as part of an atomic update are added
|
| /src/sys/external/isc/libsodium/dist/src/libsodium/sodium/ |
| core.c | 40 static volatile int locked; variable 103 assert(locked == 0); 104 locked = 1; 112 if (locked == 0) { 118 locked = 0; 134 assert(locked == 0); 135 locked = 1; 145 if (locked == 0) { 151 locked = 0;
|
| /src/external/bsd/openldap/dist/servers/slapd/back-monitor/ |
| search.c | 105 Entry *sub_nv = NULL, *sub_ch = NULL, *locked = e; local 139 monitor_cache_release( mi, locked ); 140 locked = NULL; 147 if ( locked ) { 148 monitor_cache_release( mi, locked ); 167 if ( locked ) { 168 monitor_cache_release( mi, locked );
|
| /src/lib/libpthread/ |
| pthread_lock.c | 80 int locked; local 83 locked = __SIMPLELOCK_LOCKED_P(alp); 87 return !locked;
|
| /src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
| ttm_lock.c | 68 bool locked = false; local 72 locked = true; 74 return locked; 93 static bool __ttm_read_trylock(struct ttm_lock *lock, bool *locked) 97 *locked = false; 103 *locked = true; 115 bool locked; local 120 __ttm_read_trylock(lock, &locked)); 123 __ttm_read_trylock(lock, &locked)); 127 BUG_ON(locked); 144 bool locked = false; local 188 bool locked = false; local [all...] |
| /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/engine/disp/ |
| nouveau_nvkm_engine_disp_vga.c | 135 bool locked = !nvkm_rdvgac(device, 0, 0x1f); local 145 return locked;
|
| /src/external/bsd/tmux/dist/ |
| cmd-wait-for.c | 51 int locked; member in struct:wait_channel 91 wc->locked = 0; 107 if (wc->locked) 209 if (wc->locked) { 215 wc->locked = 1; 226 if (wc == NULL || !wc->locked) { 227 cmdq_error(item, "channel %s not locked", name); 236 wc->locked = 0; 261 wc->locked = 0;
|
| /src/external/gpl3/gcc/dist/libbacktrace/ |
| mmap.c | 119 int locked; local 133 locked = 1; 135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 137 if (locked) 199 int locked; local 227 locked = 1; 229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 231 if (locked)
|
| /src/external/gpl3/gcc.old/dist/libbacktrace/ |
| mmap.c | 119 int locked; local 133 locked = 1; 135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 137 if (locked) 199 int locked; local 227 locked = 1; 229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 231 if (locked)
|
| /src/external/gpl3/gdb/dist/libbacktrace/ |
| mmap.c | 119 int locked; local 133 locked = 1; 135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 137 if (locked) 199 int locked; local 227 locked = 1; 229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 231 if (locked)
|
| /src/external/gpl3/gdb.old/dist/libbacktrace/ |
| mmap.c | 119 int locked; local 133 locked = 1; 135 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 137 if (locked) 199 int locked; local 227 locked = 1; 229 locked = __sync_lock_test_and_set (&state->lock_alloc, 1) == 0; 231 if (locked)
|
| /src/sys/arch/hppa/hppa/ |
| intr.c | 411 bool locked = false; local 458 locked = true; 469 if (locked) { 471 locked = false;
|
| /src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
| amdgpu_mxgpu_ai.c | 246 int locked; local 256 locked = mutex_trylock(&adev->lock_reset); 257 if (locked) 269 if (locked) {
|
| amdgpu_mxgpu_nv.c | 248 int locked; local 258 locked = mutex_trylock(&adev->lock_reset); 259 if (locked) 271 if (locked) {
|
| /src/sys/external/bsd/drm2/linux/ |
| linux_ww_mutex.c | 239 * True if anyone holds mutex locked at the moment, false if not. 243 * XXX WARNING: This returns true if it is locked by ANYONE. Does 250 int locked; local 255 locked = false; 260 locked = true; 268 return locked; 345 * With mutex locked and in the WW_CTX or WW_WANTOWN state, owned 397 * With mutex locked and in the WW_CTX or WW_WANTOWN state, owned 611 goto locked; 667 locked: KASSERT((mutex->wwm_state == WW_CTX) | [all...] |
| /src/sys/kern/ |
| vfs_getcwd.c | 89 * On entry, *lvpp is a locked vnode reference; on exit, it is vput and NULL'ed 90 * On exit, *uvpp is either NULL or is a locked vnode reference. 117 * current directory is still locked. 138 /* At this point, lvp is locked */ 321 bool locked = false; local 375 locked = true; 385 if (!locked) { 386 locked = true; 396 } else if (locked) {
|
| /src/external/bsd/jemalloc/dist/include/jemalloc/internal/ |
| mutex.h | 52 atomic_b_t locked; member in struct:malloc_mutex_s::__anon6753::__anon6754 152 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); 178 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); 218 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); 227 atomic_store_b(&mutex->locked, false, ATOMIC_RELAXED);
|
| /src/external/bsd/jemalloc/include/jemalloc/internal/ |
| mutex.h | 52 atomic_b_t locked; member in struct:malloc_mutex_s::__anon6832::__anon6833 152 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); 178 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); 218 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); 227 atomic_store_b(&mutex->locked, false, ATOMIC_RELAXED);
|
| /src/external/bsd/ppp/usr.sbin/pppd/ |
| tdb.h | 45 int *locked; /* set if we have a chain locked */ member in struct:__anon7996
|
| /src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
| sanitizer_mutex.cpp | 55 // The table defines a static mutex type hierarchy (what mutex types can be locked 67 // Adjacency matrix of what mutexes can be locked under what mutexes. 148 LockDesc locked[kMutexTypeMax]; member in struct:__sanitizer::InternalDeadlockDetector 154 // Find the last locked mutex type. 159 if (locked[i].seq == 0) 161 CHECK_NE(locked[i].seq, max_seq); 162 if (max_seq < locked[i].seq) { 163 max_seq = locked[i].seq; 169 CHECK_EQ(locked[type].seq, max_seq); 170 CHECK(locked[type].pc) [all...] |
| /src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
| sanitizer_mutex.cpp | 55 // The table defines a static mutex type hierarchy (what mutex types can be locked 67 // Adjacency matrix of what mutexes can be locked under what mutexes. 148 LockDesc locked[kMutexTypeMax]; member in struct:__sanitizer::InternalDeadlockDetector 154 // Find the last locked mutex type. 159 if (locked[i].seq == 0) 161 CHECK_NE(locked[i].seq, max_seq); 162 if (max_seq < locked[i].seq) { 163 max_seq = locked[i].seq; 169 CHECK_EQ(locked[type].seq, max_seq); 170 CHECK(locked[type].pc) [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| i915_gem_userptr.c | 504 int locked = 0; local 512 if (!locked) { 514 locked = 1; 521 pvec + pinned, NULL, &locked); 527 if (locked)
|
| /src/sys/uvm/ |
| uvm_loan.c | 81 * must be locked. to modify pg->loan_count, both the owner of the page 82 * and pg->interlock must be locked. pg->flags is (as always) locked by 88 * from dying pg->interlock should be locked. this forces us to sometimes 123 * on entry the map is locked by the caller) 170 /* locked: map, amap, uobj */ 181 /* locked: if (rv > 0) => map, amap, uobj [o.w. unlocked] */ 210 * unlock what we locked, unlock the maps and return 285 * map now locked. now do the loanout... 334 * => called with map, amap, uobj locked 604 bool locked; local [all...] |