Lines Matching defs:allocator_data
539 struct omp_allocator_data *allocator_data;
561 allocator_data = (struct omp_allocator_data *) allocator;
562 if (new_alignment < allocator_data->alignment)
563 new_alignment = allocator_data->alignment;
565 memkind = allocator_data->memkind;
570 allocator_data = NULL;
600 if (__builtin_expect (allocator_data
601 && allocator_data->pool_size < ~(uintptr_t) 0, 0))
604 if (new_size > allocator_data->pool_size)
607 used_pool_size = __atomic_load_n (&allocator_data->used_pool_size,
614 || new_pool_size > allocator_data->pool_size)
616 if (__atomic_compare_exchange_n (&allocator_data->used_pool_size,
624 gomp_mutex_lock (&allocator_data->lock);
625 if (__builtin_add_overflow (allocator_data->used_pool_size, new_size,
627 || used_pool_size > allocator_data->pool_size)
629 gomp_mutex_unlock (&allocator_data->lock);
632 allocator_data->used_pool_size = used_pool_size;
633 gomp_mutex_unlock (&allocator_data->lock);
651 ptr = MEMSPACE_ALLOC (allocator_data->memspace, new_size,
652 allocator_data->pinned);
656 __atomic_add_fetch (&allocator_data->used_pool_size, -new_size,
659 gomp_mutex_lock (&allocator_data->lock);
660 allocator_data->used_pool_size -= new_size;
661 gomp_mutex_unlock (&allocator_data->lock);
686 memspace = (allocator_data
687 ? allocator_data->memspace
690 allocator_data && allocator_data->pinned);
709 int fallback = (allocator_data
710 ? allocator_data->fallback
726 allocator = allocator_data->fb_data;
769 struct omp_allocator_data *allocator_data
771 if (allocator_data->pool_size < ~(uintptr_t) 0)
774 __atomic_add_fetch (&allocator_data->used_pool_size, -data->size,
777 gomp_mutex_lock (&allocator_data->lock);
778 allocator_data->used_pool_size -= data->size;
779 gomp_mutex_unlock (&allocator_data->lock);
783 if (allocator_data->memkind == GOMP_MEMKIND_LIBNUMA)
793 if (allocator_data->memkind)
796 void *kind = *memkind_data->kinds[allocator_data->memkind];
802 memspace = allocator_data->memspace;
803 pinned = allocator_data->pinned;
843 struct omp_allocator_data *allocator_data;
865 allocator_data = (struct omp_allocator_data *) allocator;
866 if (new_alignment < allocator_data->alignment)
867 new_alignment = allocator_data->alignment;
869 memkind = allocator_data->memkind;
874 allocator_data = NULL;
906 if (__builtin_expect (allocator_data
907 && allocator_data->pool_size < ~(uintptr_t) 0, 0))
910 if (new_size > allocator_data->pool_size)
913 used_pool_size = __atomic_load_n (&allocator_data->used_pool_size,
920 || new_pool_size > allocator_data->pool_size)
922 if (__atomic_compare_exchange_n (&allocator_data->used_pool_size,
930 gomp_mutex_lock (&allocator_data->lock);
931 if (__builtin_add_overflow (allocator_data->used_pool_size, new_size,
933 || used_pool_size > allocator_data->pool_size)
935 gomp_mutex_unlock (&allocator_data->lock);
938 allocator_data->used_pool_size = used_pool_size;
939 gomp_mutex_unlock (&allocator_data->lock);
959 ptr = MEMSPACE_CALLOC (allocator_data->memspace, new_size,
960 allocator_data->pinned);
964 __atomic_add_fetch (&allocator_data->used_pool_size, -new_size,
967 gomp_mutex_lock (&allocator_data->lock);
968 allocator_data->used_pool_size -= new_size;
969 gomp_mutex_unlock (&allocator_data->lock);
996 memspace = (allocator_data
997 ? allocator_data->memspace
1000 allocator_data && allocator_data->pinned);
1019 int fallback = (allocator_data
1020 ? allocator_data->fallback
1036 allocator = allocator_data->fb_data;
1054 struct omp_allocator_data *allocator_data, *free_allocator_data;
1081 allocator_data = (struct omp_allocator_data *) allocator;
1082 if (new_alignment < allocator_data->alignment)
1083 new_alignment = allocator_data->alignment;
1085 memkind = allocator_data->memkind;
1090 allocator_data = NULL;
1146 if (__builtin_expect (allocator_data
1147 && allocator_data->pool_size < ~(uintptr_t) 0, 0))
1156 && free_allocator_data == allocator_data
1161 && new_size - prev_size > allocator_data->pool_size)
1164 used_pool_size = __atomic_load_n (&allocator_data->used_pool_size,
1173 || new_pool_size > allocator_data->pool_size)
1178 if (__atomic_compare_exchange_n (&allocator_data->used_pool_size,
1186 gomp_mutex_lock (&allocator_data->lock);
1189 if (__builtin_add_overflow (allocator_data->used_pool_size,
1192 || used_pool_size > allocator_data->pool_size)
1194 gomp_mutex_unlock (&allocator_data->lock);
1199 used_pool_size = (allocator_data->used_pool_size
1201 allocator_data->used_pool_size = used_pool_size;
1202 gomp_mutex_unlock (&allocator_data->lock);
1231 new_ptr = MEMSPACE_REALLOC (allocator_data->memspace, data->ptr,
1235 allocator_data->pinned);
1237 new_ptr = MEMSPACE_ALLOC (allocator_data->memspace, new_size,
1238 allocator_data->pinned);
1242 __atomic_add_fetch (&allocator_data->used_pool_size,
1246 gomp_mutex_lock (&allocator_data->lock);
1247 allocator_data->used_pool_size -= new_size - prev_size;
1248 gomp_mutex_unlock (&allocator_data->lock);
1288 memspace = (allocator_data
1289 ? allocator_data->memspace
1294 allocator_data && allocator_data->pinned);
1325 memspace = (allocator_data
1326 ? allocator_data->memspace
1329 allocator_data && allocator_data->pinned);
1390 int fallback = (allocator_data
1391 ? allocator_data->fallback
1407 allocator = allocator_data->fb_data;