Lines Matching refs:parent

75 slab_get_element(struct slab_parent_pool *parent,
79 ((uint8_t*)&page[1] + (parent->element_size * index));
99 * Create a parent pool for the allocation of same-sized objects.
105 slab_create_parent(struct slab_parent_pool *parent,
109 simple_mtx_init(&parent->mutex, mtx_plain);
110 parent->element_size = ALIGN_POT(sizeof(struct slab_element_header) + item_size,
112 parent->num_elements = num_items;
116 slab_destroy_parent(struct slab_parent_pool *parent)
118 simple_mtx_destroy(&parent->mutex);
122 * Create a child pool linked to the given parent.
125 struct slab_parent_pool *parent)
127 pool->parent = parent;
141 if (!pool->parent)
144 simple_mtx_lock(&pool->parent->mutex);
149 p_atomic_set(&page->u.num_remaining, pool->parent->num_elements);
151 for (unsigned i = 0; i < pool->parent->num_elements; ++i) {
152 struct slab_element_header *elt = slab_get_element(pool->parent, page, i);
163 simple_mtx_unlock(&pool->parent->mutex);
172 pool->parent = NULL;
179 pool->parent->num_elements * pool->parent->element_size);
184 for (unsigned i = 0; i < pool->parent->num_elements; ++i) {
185 struct slab_element_header *elt = slab_get_element(pool->parent, page, i);
214 simple_mtx_lock(&pool->parent->mutex);
217 simple_mtx_unlock(&pool->parent->mutex);
239 * allocated is allowed, as long the pool belong to the same parent. No
260 if (pool->parent)
261 simple_mtx_lock(&pool->parent->mutex);
272 if (pool->parent)
273 simple_mtx_unlock(&pool->parent->mutex);
275 if (pool->parent)
276 simple_mtx_unlock(&pool->parent->mutex);
304 slab_destroy_parent(&mempool->parent);
318 slab_create_parent(&mempool->parent, item_size, num_items);
319 slab_create_child(&mempool->child, &mempool->parent);