| /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
| tsan_interface.h | 193 typedef unsigned long long a64; // NOLINT typedef in namespace:__tsan 223 a64 __tsan_atomic64_load(const volatile a64 *a, morder mo); 236 void __tsan_atomic64_store(volatile a64 *a, a64 v, morder mo); 249 a64 __tsan_atomic64_exchange(volatile a64 *a, a64 v, morder mo); 262 a64 __tsan_atomic64_fetch_add(volatile a64 *a, a64 v, morder mo) [all...] |
| /src/external/bsd/elftoolchain/dist/libelf/ |
| libelf_align.c | 47 unsigned int a64; member in struct:align 53 .a64 = __alignof__(Elf64_##N) \ 57 .a64 = __alignof__(Elf64_##V) \ 61 .a64 = __alignof__(int64_t) \ 64 #define MALIGN(N) { .a32 = 0, .a64 = 0 } 65 #define MALIGN64(N) { .a32 = 0, .a64 = 0 } 66 #define MALIGN_WORD(N) { .a32 = 0, .a64 = 0 } 72 .a64 = 0 \ 77 [ELF_T_BYTE] = { .a32 = 1, .a64 = 1 }, 109 malign[t].a64); [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
| tsan_interface.h | 209 typedef unsigned long long a64; typedef in namespace:__tsan 240 a64 __tsan_atomic64_load(const volatile a64 *a, morder mo); 253 void __tsan_atomic64_store(volatile a64 *a, a64 v, morder mo); 266 a64 __tsan_atomic64_exchange(volatile a64 *a, a64 v, morder mo); 279 a64 __tsan_atomic64_fetch_add(volatile a64 *a, a64 v, morder mo) [all...] |
| /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
| tsan_interface.h | 197 typedef unsigned long long a64; typedef in namespace:__tsan 228 a64 __tsan_atomic64_load(const volatile a64 *a, morder mo); 241 void __tsan_atomic64_store(volatile a64 *a, a64 v, morder mo); 254 a64 __tsan_atomic64_exchange(volatile a64 *a, a64 v, morder mo); 267 a64 __tsan_atomic64_fetch_add(volatile a64 *a, a64 v, morder mo) [all...] |
| /src/sys/arch/aarch64/aarch64/ |
| aarch32_syscall.c | 54 register_t a64[EMULNAMEU(SYS_MAXSYSARGS)]; member in union:args 182 args64buf.a64[i] = args32buf.a32[i]; 186 error = trace_enter(code, callp, args64buf.a64); 192 trace_exit(code, callp, args64buf.a64, rval, error);
|
| /src/regress/lib/libc/divrem/ |
| divremtest.c | 85 intmax_t a64, b64, sr64; local 130 a64 = IM(1) << i; 135 T64S( a64 + k, b64 + l, sr64); 136 T64S( a64 + k, -b64 + l, sr64); 137 T64S(-a64 + k, b64 + l, sr64); 138 T64S(-a64 + k, -b64 + l, sr64); 139 T64U(UIM(a64) + k, UIM(b64) + l, ur64); 146 for(a64 = -(1 << KLE); a64 < 1 << KLE; ++a64) { [all...] |