| /src/crypto/external/bsd/openssh/dist/ |
| srclimit.c | 73 srclimit_mask_addr(const struct xaddr *addr, int bits, struct xaddr *masked) 79 addr_and(masked, addr, &xmask) != 0) { 176 /* Connection allowed, store masked address. */ 350 struct xaddr masked; local 398 if (srclimit_mask_addr(addr, bits, &masked) != 0) 422 penalty->addr = masked;
|
| /src/external/bsd/ntp/dist/ntpd/ |
| ntp_restrict.c | 372 struct in6_addr masked; local 379 MASK_IPV6_ADDR(&masked, addr, &res->v6.mask); 380 if (ADDR6_EQ(&masked, &res->v6.addr)
|
| /src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
| tsan_rtl_access.cpp | 182 auto masked = static_cast<RawShadow>(static_cast<u32>(old) | local 184 if (masked == cur.raw())
|
| /src/sys/kern/ |
| kern_sig.c | 951 const bool masked = sigismember(mask, signo); local 952 if (caught && !masked) { 968 * If the signal is masked or ignored, then unmask it and 973 if (masked || ignored) { 1075 * Returns true if signal is ignored or masked for the specified LWP. 1095 int rv, masked; local 1126 * SIGCONT can be masked, but if LWP is stopped, it needs restart. 1127 * Note: SIGKILL and SIGSTOP cannot be masked. 1129 masked = sigismember(&l->l_sigmask, sig); 1130 if (masked && ((prop & SA_CONT) == 0 || l->l_stat != LSSTOP)) [all...] |
| /src/sys/arch/aarch64/aarch64/ |
| pmap.c | 955 vm_prot_t masked; local 958 masked = prot & refmod; 969 switch (masked & (VM_PROT_READ | VM_PROT_WRITE)) {
|
| /src/external/gpl3/gdb/dist/gdb/ |
| frame.c | 159 bool masked; member in struct:frame_info::__anon17848 203 frame->prev_pc.masked = true; 214 return frame->next->prev_pc.masked; 503 fi->next->prev_pc.masked ? "[PAC]" : "");
|
| /src/external/gpl3/gdb.old/dist/gdb/ |
| frame.c | 159 bool masked; member in struct:frame_info::__anon20587 203 frame->prev_pc.masked = true; 214 return frame->next->prev_pc.masked; 503 fi->next->prev_pc.masked ? "[PAC]" : "");
|
| /src/external/gpl3/gcc/dist/gcc/ |
| gimple-ssa-store-merging.cc | 245 order to have a nop. The number is masked according to the size of 251 order to have a byte swap. The number is masked according to the 589 later masked off. */ 3949 unsigned int try_bytesize = try_size / BITS_PER_UNIT, masked; local 3950 for (masked = 0; masked < try_bytesize; ++masked) 3951 if (group->mask[try_pos - bytepos + masked] != (unsigned char) ~0U 3953 || group->val[try_pos - bytepos + masked] != 0)) 3955 masked *= BITS_PER_UNIT [all...] |
| tree-vect-patterns.cc | 2583 masked = container & (((1 << bitsize) - 1) << bitpos); 2584 result = masked >> bitpos; 2592 masked = container & (((1 << bitsize) - 1) << bitpos); 2593 result = masked >> bitpos; 2599 masked = container << (prec - bitsize - bitpos); 2600 result = (type_out) (masked >> (prec - bitsize)); 2759 tree masked = gimple_assign_lhs (pattern_stmt); local 2763 RSHIFT_EXPR, masked, 2828 masked = shifted & (mask << bitpos); // Mask off the non-relevant bits in 2833 written = cleared | masked; // Write bits 2905 tree masked = gimple_build (&stmts, BIT_AND_EXPR, container_type, shifted, local [all...] |
| combine.cc | 7420 rtx mask, cleared, masked; 7526 masked = simplify_gen_binary (ASHIFT, compute_mode, 7533 cleared, masked)); 9901 /* If SRC is masked by an AND that does not make a difference in 10552 /* We can also widen if the bits brought in will be masked off. This 10985 sure the result will be masked. See PR70222. */ 12261 /* The resulting comparison is always unsigned since we masked 7397 rtx mask, cleared, masked; local
|
| /src/external/gpl3/gcc.old/dist/gcc/ |
| gimple-ssa-store-merging.cc | 245 order to have a nop. The number is masked according to the size of 251 order to have a byte swap. The number is masked according to the 589 later masked off. */ 3845 unsigned int try_bytesize = try_size / BITS_PER_UNIT, masked; local 3846 for (masked = 0; masked < try_bytesize; ++masked) 3847 if (group->mask[try_pos - bytepos + masked] != (unsigned char) ~0U 3849 || group->val[try_pos - bytepos + masked] != 0)) 3851 masked *= BITS_PER_UNIT [all...] |
| combine.cc | 7380 rtx mask, cleared, masked; 7481 masked = simplify_gen_binary (ASHIFT, compute_mode, 7490 cleared, masked)); 9850 /* If SRC is masked by an AND that does not make a difference in 10496 /* We can also widen if the bits brought in will be masked off. This 10929 sure the result will be masked. See PR70222. */ 12131 /* The resulting comparison is always unsigned since we masked 7357 rtx mask, cleared, masked; local
|
| tree-vect-stmts.cc | 1837 /* Return the mask input to a masked load or store. VEC_MASK is the vectorized 1840 Return VEC_MASK if LOOP_MASK is null or if VEC_MASK is already masked, 2085 /* STMT_INFO is either a masked or unconditional store. Return the value 2316 or using masked loads with a static mask. */ 3557 "can't use a fully-masked loop because no" 6306 Similarly, if this operation is part of a reduction, a fully-masked 6324 "can't use a fully-masked loop because no" 7544 "SLP of masked stores not supported.\n"); 7625 "unsupported access type for masked store.\n"); 8184 /* Shouldn't go with length-based approach if fully masked. * 10357 bool masked = false; local [all...] |
| /src/external/gpl3/gdb/dist/sim/bfin/ |
| bfin-sim.c | 764 bu64 masked, new_val = val; local 769 masked = val & mask; 776 masked = new_val & mask; 807 masked &= 0xFFFFFFFF; 811 || (sgn != masked) 820 masked &= 0xFFFFFFFFFFull;
|
| /src/external/gpl3/gdb.old/dist/sim/bfin/ |
| bfin-sim.c | 764 bu64 masked, new_val = val; local 769 masked = val & mask; 776 masked = new_val & mask; 807 masked &= 0xFFFFFFFF; 811 || (sgn != masked) 820 masked &= 0xFFFFFFFFFFull;
|
| /src/external/gpl3/gcc/dist/gcc/config/i386/ |
| i386-expand.cc | 15430 int masked = 1; local 15443 masked = 0; 15451 masked = 0; 15459 masked = 0; 15464 masked = 0; 15520 if (! masked) 15537 /* If merge is 0 then we're about to emit z-masked variant. */ 15540 /* If merge is the same as accum then emit merge-masked variant. */ 15559 masked = 0; 15564 masked = 0 [all...] |
| /src/external/gpl3/gcc.old/dist/gcc/config/i386/ |
| i386-expand.cc | 14571 int masked = 1; local 14584 masked = 0; 14592 masked = 0; 14600 masked = 0; 14605 masked = 0; 14661 if (! masked) 14678 /* If merge is 0 then we're about to emit z-masked variant. */ 14681 /* If merge is the same as accum then emit merge-masked variant. */ 14700 masked = 0; 14705 masked = 0 [all...] |
| /src/external/gpl3/gcc/dist/gcc/fortran/ |
| gfortran.h | 909 unsigned unmaskable:1, masked:1, contained:1, mod_proc:1, abstract:1; member in struct:__anon13616
|
| /src/external/gpl3/gcc.old/dist/gcc/fortran/ |
| gfortran.h | 901 unsigned unmaskable:1, masked:1, contained:1, mod_proc:1, abstract:1; member in struct:__anon16051
|