/src/sys/external/bsd/compiler_rt/dist/lib/asan/ |
asan_fake_stack.h | 92 static uptr FlagsOffset(uptr stack_size_log, uptr class_id) { 93 uptr t = kNumberOfSizeClasses - 1 - class_id; 98 static uptr NumberOfFrames(uptr stack_size_log, uptr class_id) { 99 return ((uptr)1) << (stack_size_log - kMinStackFrameSizeLog - class_id); 103 static uptr ModuloNumberOfFrames(uptr stack_size_log, uptr class_id, uptr n) { 104 return n & (NumberOfFrames(stack_size_log, class_id) - 1); 107 // The pointer to the flags of the given class_id. 108 u8 *GetFlags(uptr stack_size_log, uptr class_id) { 110 FlagsOffset(stack_size_log, class_id); 113 // Get frame by class_id and pos [all...] |
asan_fake_stack.cc | 30 ALWAYS_INLINE void SetShadow(uptr ptr, uptr size, uptr class_id, u64 magic) { 32 if (SHADOW_SCALE == 3 && class_id <= 6) { 34 for (uptr i = 0; i < (((uptr)1) << class_id); i++) { 70 for (uptr class_id = 0; class_id < kNumberOfSizeClasses; class_id++) local in function:__asan::FakeStack::Destroy 71 str.append("%zd: %zd/%zd; ", class_id, hint_position_[class_id], 72 NumberOfFrames(stack_size_log(), class_id)); 88 FakeFrame *FakeStack::Allocate(uptr stack_size_log, uptr class_id, 121 uptr class_id = (ptr - beg) >> stack_size_log; local in function:__asan::FakeStack::AddrIsInFakeStack 144 for (uptr class_id = 0; class_id < kNumberOfSizeClasses; class_id++) { local in function:__asan::FakeStack::GC 161 for (uptr class_id = 0; class_id < kNumberOfSizeClasses; class_id++) { local in function:__asan::FakeStack::ForEachFakeFrame [all...] |
/src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
sanitizer_allocator_local_cache.h | 41 void *Allocate(SizeClassAllocator *allocator, uptr class_id) { 42 CHECK_NE(class_id, 0UL); 43 CHECK_LT(class_id, kNumClasses); 44 PerClass *c = &per_class_[class_id]; 46 if (UNLIKELY(!Refill(c, allocator, class_id))) 53 allocator->GetRegionBeginBySizeClass(class_id), chunk)); 56 void Deallocate(SizeClassAllocator *allocator, uptr class_id, void *p) { 57 CHECK_NE(class_id, 0UL); 58 CHECK_LT(class_id, kNumClasses); 61 PerClass *c = &per_class_[class_id]; [all...] |
sanitizer_allocator_primary64.h | 99 for (uptr class_id = 1; class_id < kNumClasses; class_id++) { local in function:SizeClassAllocator64::ForceReleaseToOS 100 BlockingMutexLock l(&GetRegionInfo(class_id)->mutex); 101 MaybeReleaseToOS(class_id, true /*force*/); 110 NOINLINE void ReturnToAllocator(AllocatorStats *stat, uptr class_id, 112 RegionInfo *region = GetRegionInfo(class_id); 113 uptr region_beg = GetRegionBeginBySizeClass(class_id); 125 class_id, ClassIdToSize(class_id)); 184 uptr class_id = GetSizeClass(p); local in function:SizeClassAllocator64::GetBlockBegin 206 uptr class_id = GetSizeClass(p); local in function:SizeClassAllocator64::GetMetaData 228 for (uptr class_id = 0; class_id < stats_size; class_id++) local in function:SizeClassAllocator64::FillMemoryProfile 252 for (uptr class_id = 0; class_id < kNumClasses; class_id++) local in function:SizeClassAllocator64::PrintStats 260 for (uptr class_id = 1; class_id < kNumClasses; class_id++) { local in function:SizeClassAllocator64::PrintStats 273 for (uptr class_id = 1; class_id < kNumClasses; class_id++) local in function:SizeClassAllocator64::PrintStats 294 for (uptr class_id = 1; class_id < kNumClasses; class_id++) { local in function:SizeClassAllocator64::ForEachChunk [all...] |
sanitizer_allocator_primary32.h | 107 static uptr ClassIdToSize(uptr class_id) { 108 return (class_id == SizeClassMap::kBatchClassID) ? 109 kBatchSize : SizeClassMap::Size(class_id); 160 uptr class_id) { 161 DCHECK_LT(class_id, kNumClasses); 162 SizeClassInfo *sci = GetSizeClassInfo(class_id); 165 if (UNLIKELY(!PopulateFreeList(stat, c, sci, class_id))) 174 NOINLINE void DeallocateBatch(AllocatorStats *stat, uptr class_id, 176 DCHECK_LT(class_id, kNumClasses); 178 SizeClassInfo *sci = GetSizeClassInfo(class_id); [all...] |
sanitizer_allocator_size_class_map.h | 30 // Classes 1 - 16 correspond to sizes 16 to 256 (size = class_id * 16). 146 static uptr Size(uptr class_id) { 150 if (UNLIKELY(class_id == kBatchClassID)) 152 if (class_id <= kMidClass) 153 return kMinSize * class_id; 154 class_id -= kMidClass; 155 uptr t = kMidSize << (class_id >> S); 156 return t + (t >> S) * (class_id & M);
|
/src/sys/external/bsd/compiler_rt/dist/lib/asan/tests/ |
asan_fake_stack_test.cc | 50 for (uptr class_id = 0; class_id < FakeStack::kNumberOfSizeClasses; local in function:__asan::TEST 51 class_id++) { 52 uptr frame_size = FakeStack::BytesInSizeClass(class_id); 54 EXPECT_EQ(offset, FakeStack::FlagsOffset(stack_size_log, class_id)); 55 // printf("%zd: %zd => %zd %zd\n", stack_size_log, class_id, offset, 56 // FakeStack::FlagsOffset(stack_size_log, class_id)); 136 uptr class_id = depth / 3; local in function:__asan::RecursiveFunction 137 FakeFrame *ff = fs->Allocate(fs->stack_size_log(), class_id, 0); 142 fs->Deallocate(reinterpret_cast<uptr>(ff), class_id); [all...] |
/src/usr.sbin/altq/altqstat/ |
qdisc_hfsc.c | 74 last[i].class_id = 999999; /* XXX */ 99 if (sp->class_id != lp->class_id) { 105 sp->class_id, clnames[i], sp->class_handle,
|
/src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/tests/ |
sanitizer_allocator_test.cc | 230 uptr class_id = a->GetSizeClass(x); local in function:TestSizeClassAllocator 231 CHECK_EQ(class_id, Allocator::SizeClassMapT::ClassID(size)); 739 for (int class_id = 1; class_id <= 5; class_id++) { local in function:TestSizeClassAllocatorLocalCache 743 allocated[i] = cache.Allocate(a, class_id); 746 cache.Deallocate(a, class_id, allocated[i]); 826 uptr class_id; member in struct:NewThreadParams 832 params->thread_cache->Deallocate(params->allocator, params->class_id, params); 846 uptr class_id = DefaultSizeClassMap::ClassID(sizeof(NewThreadParams)) local in function:TEST 1270 for (uptr class_id = 1; class_id <= Allocator::SizeClassMapT::kLargestClassID; local in function:TestReleaseFreeMemoryToOS [all...] |
/src/sys/altq/ |
altq_hfsc.h | 69 u_int class_id; member in struct:hfsc_classstats
|
altq_hfsc.c | 1616 sp->class_id = cl->cl_id;
|