17ec681f3Smrg/* 27ec681f3Smrg * Copyright © 2019 Intel Corporation 37ec681f3Smrg * 47ec681f3Smrg * Permission is hereby granted, free of charge, to any person obtaining a 57ec681f3Smrg * copy of this software and associated documentation files (the "Software"), 67ec681f3Smrg * to deal in the Software without restriction, including without limitation 77ec681f3Smrg * the rights to use, copy, modify, merge, publish, distribute, sublicense, 87ec681f3Smrg * and/or sell copies of the Software, and to permit persons to whom the 97ec681f3Smrg * Software is furnished to do so, subject to the following conditions: 107ec681f3Smrg * 117ec681f3Smrg * The above copyright notice and this permission notice (including the next 127ec681f3Smrg * paragraph) shall be included in all copies or substantial portions of the 137ec681f3Smrg * Software. 147ec681f3Smrg * 157ec681f3Smrg * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 167ec681f3Smrg * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 177ec681f3Smrg * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 187ec681f3Smrg * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 197ec681f3Smrg * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 207ec681f3Smrg * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS 217ec681f3Smrg * IN THE SOFTWARE. 227ec681f3Smrg */ 237ec681f3Smrg 247ec681f3Smrg#include "sparse_array.h" 257ec681f3Smrg#include "os_memory.h" 267ec681f3Smrg 277ec681f3Smrg/* Aligning our allocations to 64 has two advantages: 287ec681f3Smrg * 297ec681f3Smrg * 1. On x86 platforms, it means that they are cache-line aligned so we 307ec681f3Smrg * reduce the likelihood that one of our allocations shares a cache line 317ec681f3Smrg * with some other allocation. 327ec681f3Smrg * 337ec681f3Smrg * 2. It lets us use the bottom 6 bits of the pointer to store the tree level 347ec681f3Smrg * of the node so we can avoid some pointer indirections. 357ec681f3Smrg */ 367ec681f3Smrg#define NODE_ALLOC_ALIGN 64 377ec681f3Smrg 387ec681f3Smrgvoid 397ec681f3Smrgutil_sparse_array_init(struct util_sparse_array *arr, 407ec681f3Smrg size_t elem_size, size_t node_size) 417ec681f3Smrg{ 427ec681f3Smrg memset(arr, 0, sizeof(*arr)); 437ec681f3Smrg arr->elem_size = elem_size; 447ec681f3Smrg arr->node_size_log2 = util_logbase2_64(node_size); 457ec681f3Smrg assert(node_size >= 2 && node_size == (1ull << arr->node_size_log2)); 467ec681f3Smrg} 477ec681f3Smrg 487ec681f3Smrg#define NODE_PTR_MASK (~((uintptr_t)NODE_ALLOC_ALIGN - 1)) 497ec681f3Smrg#define NODE_LEVEL_MASK ((uintptr_t)NODE_ALLOC_ALIGN - 1) 507ec681f3Smrg#define NULL_NODE 0 517ec681f3Smrg 527ec681f3Smrgstatic inline uintptr_t 537ec681f3Smrg_util_sparse_array_node(void *data, unsigned level) 547ec681f3Smrg{ 557ec681f3Smrg assert(data != NULL); 567ec681f3Smrg assert(((uintptr_t)data & NODE_LEVEL_MASK) == 0); 577ec681f3Smrg assert((level & NODE_PTR_MASK) == 0); 587ec681f3Smrg return (uintptr_t)data | level; 597ec681f3Smrg} 607ec681f3Smrg 617ec681f3Smrgstatic inline void * 627ec681f3Smrg_util_sparse_array_node_data(uintptr_t handle) 637ec681f3Smrg{ 647ec681f3Smrg return (void *)(handle & NODE_PTR_MASK); 657ec681f3Smrg} 667ec681f3Smrg 677ec681f3Smrgstatic inline unsigned 687ec681f3Smrg_util_sparse_array_node_level(uintptr_t handle) 697ec681f3Smrg{ 707ec681f3Smrg return handle & NODE_LEVEL_MASK; 717ec681f3Smrg} 727ec681f3Smrg 737ec681f3Smrgstatic inline void 747ec681f3Smrg_util_sparse_array_node_finish(struct util_sparse_array *arr, 757ec681f3Smrg uintptr_t node) 767ec681f3Smrg{ 777ec681f3Smrg if (_util_sparse_array_node_level(node) > 0) { 787ec681f3Smrg uintptr_t *children = _util_sparse_array_node_data(node); 797ec681f3Smrg size_t node_size = 1ull << arr->node_size_log2; 807ec681f3Smrg for (size_t i = 0; i < node_size; i++) { 817ec681f3Smrg if (children[i]) 827ec681f3Smrg _util_sparse_array_node_finish(arr, children[i]); 837ec681f3Smrg } 847ec681f3Smrg } 857ec681f3Smrg 867ec681f3Smrg os_free_aligned(_util_sparse_array_node_data(node)); 877ec681f3Smrg} 887ec681f3Smrg 897ec681f3Smrgvoid 907ec681f3Smrgutil_sparse_array_finish(struct util_sparse_array *arr) 917ec681f3Smrg{ 927ec681f3Smrg if (arr->root) 937ec681f3Smrg _util_sparse_array_node_finish(arr, arr->root); 947ec681f3Smrg} 957ec681f3Smrg 967ec681f3Smrgstatic inline uintptr_t 977ec681f3Smrg_util_sparse_array_node_alloc(struct util_sparse_array *arr, 987ec681f3Smrg unsigned level) 997ec681f3Smrg{ 1007ec681f3Smrg size_t size; 1017ec681f3Smrg if (level == 0) { 1027ec681f3Smrg size = arr->elem_size << arr->node_size_log2; 1037ec681f3Smrg } else { 1047ec681f3Smrg size = sizeof(uintptr_t) << arr->node_size_log2; 1057ec681f3Smrg } 1067ec681f3Smrg 1077ec681f3Smrg void *data = os_malloc_aligned(size, NODE_ALLOC_ALIGN); 1087ec681f3Smrg memset(data, 0, size); 1097ec681f3Smrg 1107ec681f3Smrg return _util_sparse_array_node(data, level); 1117ec681f3Smrg} 1127ec681f3Smrg 1137ec681f3Smrgstatic inline uintptr_t 1147ec681f3Smrg_util_sparse_array_set_or_free_node(uintptr_t *node_ptr, 1157ec681f3Smrg uintptr_t cmp_node, 1167ec681f3Smrg uintptr_t node) 1177ec681f3Smrg{ 1187ec681f3Smrg uintptr_t prev_node = p_atomic_cmpxchg(node_ptr, cmp_node, node); 1197ec681f3Smrg 1207ec681f3Smrg if (prev_node != cmp_node) { 1217ec681f3Smrg /* We lost the race. Free this one and return the one that was already 1227ec681f3Smrg * allocated. 1237ec681f3Smrg */ 1247ec681f3Smrg os_free_aligned(_util_sparse_array_node_data(node)); 1257ec681f3Smrg return prev_node; 1267ec681f3Smrg } else { 1277ec681f3Smrg return node; 1287ec681f3Smrg } 1297ec681f3Smrg} 1307ec681f3Smrg 1317ec681f3Smrgvoid * 1327ec681f3Smrgutil_sparse_array_get(struct util_sparse_array *arr, uint64_t idx) 1337ec681f3Smrg{ 1347ec681f3Smrg const unsigned node_size_log2 = arr->node_size_log2; 1357ec681f3Smrg uintptr_t root = p_atomic_read(&arr->root); 1367ec681f3Smrg if (unlikely(!root)) { 1377ec681f3Smrg unsigned root_level = 0; 1387ec681f3Smrg uint64_t idx_iter = idx >> node_size_log2; 1397ec681f3Smrg while (idx_iter) { 1407ec681f3Smrg idx_iter >>= node_size_log2; 1417ec681f3Smrg root_level++; 1427ec681f3Smrg } 1437ec681f3Smrg uintptr_t new_root = _util_sparse_array_node_alloc(arr, root_level); 1447ec681f3Smrg root = _util_sparse_array_set_or_free_node(&arr->root, 1457ec681f3Smrg NULL_NODE, new_root); 1467ec681f3Smrg } 1477ec681f3Smrg 1487ec681f3Smrg while (1) { 1497ec681f3Smrg unsigned root_level = _util_sparse_array_node_level(root); 1507ec681f3Smrg uint64_t root_idx = idx >> (root_level * node_size_log2); 1517ec681f3Smrg if (likely(root_idx < (1ull << node_size_log2))) 1527ec681f3Smrg break; 1537ec681f3Smrg 1547ec681f3Smrg /* In this case, we have a root but its level is low enough that the 1557ec681f3Smrg * requested index is out-of-bounds. 1567ec681f3Smrg */ 1577ec681f3Smrg uintptr_t new_root = _util_sparse_array_node_alloc(arr, root_level + 1); 1587ec681f3Smrg 1597ec681f3Smrg uintptr_t *new_root_children = _util_sparse_array_node_data(new_root); 1607ec681f3Smrg new_root_children[0] = root; 1617ec681f3Smrg 1627ec681f3Smrg /* We only add one at a time instead of the whole tree because it's 1637ec681f3Smrg * easier to ensure correctness of both the tree building and the 1647ec681f3Smrg * clean-up path. Because we're only adding one node we never have to 1657ec681f3Smrg * worry about trying to free multiple things without freeing the old 1667ec681f3Smrg * things. 1677ec681f3Smrg */ 1687ec681f3Smrg root = _util_sparse_array_set_or_free_node(&arr->root, root, new_root); 1697ec681f3Smrg } 1707ec681f3Smrg 1717ec681f3Smrg void *node_data = _util_sparse_array_node_data(root); 1727ec681f3Smrg unsigned node_level = _util_sparse_array_node_level(root); 1737ec681f3Smrg while (node_level > 0) { 1747ec681f3Smrg uint64_t child_idx = (idx >> (node_level * node_size_log2)) & 1757ec681f3Smrg ((1ull << node_size_log2) - 1); 1767ec681f3Smrg 1777ec681f3Smrg uintptr_t *children = node_data; 1787ec681f3Smrg uintptr_t child = p_atomic_read(&children[child_idx]); 1797ec681f3Smrg 1807ec681f3Smrg if (unlikely(!child)) { 1817ec681f3Smrg child = _util_sparse_array_node_alloc(arr, node_level - 1); 1827ec681f3Smrg child = _util_sparse_array_set_or_free_node(&children[child_idx], 1837ec681f3Smrg NULL_NODE, child); 1847ec681f3Smrg } 1857ec681f3Smrg 1867ec681f3Smrg node_data = _util_sparse_array_node_data(child); 1877ec681f3Smrg node_level = _util_sparse_array_node_level(child); 1887ec681f3Smrg } 1897ec681f3Smrg 1907ec681f3Smrg uint64_t elem_idx = idx & ((1ull << node_size_log2) - 1); 1917ec681f3Smrg return (void *)((char *)node_data + (elem_idx * arr->elem_size)); 1927ec681f3Smrg} 1937ec681f3Smrg 1947ec681f3Smrgstatic void 1957ec681f3Smrgvalidate_node_level(struct util_sparse_array *arr, 1967ec681f3Smrg uintptr_t node, unsigned level) 1977ec681f3Smrg{ 1987ec681f3Smrg assert(_util_sparse_array_node_level(node) == level); 1997ec681f3Smrg 2007ec681f3Smrg if (_util_sparse_array_node_level(node) > 0) { 2017ec681f3Smrg uintptr_t *children = _util_sparse_array_node_data(node); 2027ec681f3Smrg size_t node_size = 1ull << arr->node_size_log2; 2037ec681f3Smrg for (size_t i = 0; i < node_size; i++) { 2047ec681f3Smrg if (children[i]) 2057ec681f3Smrg validate_node_level(arr, children[i], level - 1); 2067ec681f3Smrg } 2077ec681f3Smrg } 2087ec681f3Smrg} 2097ec681f3Smrg 2107ec681f3Smrgvoid 2117ec681f3Smrgutil_sparse_array_validate(struct util_sparse_array *arr) 2127ec681f3Smrg{ 2137ec681f3Smrg uintptr_t root = p_atomic_read(&arr->root); 2147ec681f3Smrg validate_node_level(arr, root, _util_sparse_array_node_level(root)); 2157ec681f3Smrg} 2167ec681f3Smrg 2177ec681f3Smrgvoid 2187ec681f3Smrgutil_sparse_array_free_list_init(struct util_sparse_array_free_list *fl, 2197ec681f3Smrg struct util_sparse_array *arr, 2207ec681f3Smrg uint32_t sentinel, 2217ec681f3Smrg uint32_t next_offset) 2227ec681f3Smrg{ 2237ec681f3Smrg fl->head = sentinel; 2247ec681f3Smrg fl->arr = arr; 2257ec681f3Smrg fl->sentinel = sentinel; 2267ec681f3Smrg fl->next_offset = next_offset; 2277ec681f3Smrg} 2287ec681f3Smrg 2297ec681f3Smrgstatic uint64_t 2307ec681f3Smrgfree_list_head(uint64_t old, uint32_t next) 2317ec681f3Smrg{ 2327ec681f3Smrg return ((old & 0xffffffff00000000ull) + 0x100000000ull) | next; 2337ec681f3Smrg} 2347ec681f3Smrg 2357ec681f3Smrgvoid 2367ec681f3Smrgutil_sparse_array_free_list_push(struct util_sparse_array_free_list *fl, 2377ec681f3Smrg uint32_t *items, unsigned num_items) 2387ec681f3Smrg{ 2397ec681f3Smrg assert(num_items > 0); 2407ec681f3Smrg assert(items[0] != fl->sentinel); 2417ec681f3Smrg void *last_elem = util_sparse_array_get(fl->arr, items[0]); 2427ec681f3Smrg uint32_t *last_next = (uint32_t *)((char *)last_elem + fl->next_offset); 2437ec681f3Smrg for (unsigned i = 1; i < num_items; i++) { 2447ec681f3Smrg p_atomic_set(last_next, items[i]); 2457ec681f3Smrg assert(items[i] != fl->sentinel); 2467ec681f3Smrg last_elem = util_sparse_array_get(fl->arr, items[i]); 2477ec681f3Smrg last_next = (uint32_t *)((char *)last_elem + fl->next_offset); 2487ec681f3Smrg } 2497ec681f3Smrg 2507ec681f3Smrg uint64_t current_head, old_head; 2517ec681f3Smrg old_head = p_atomic_read(&fl->head); 2527ec681f3Smrg do { 2537ec681f3Smrg current_head = old_head; 2547ec681f3Smrg p_atomic_set(last_next, (uint32_t)current_head); /* Index is the bottom 32 bits */ 2557ec681f3Smrg uint64_t new_head = free_list_head(current_head, items[0]); 2567ec681f3Smrg old_head = p_atomic_cmpxchg(&fl->head, current_head, new_head); 2577ec681f3Smrg } while (old_head != current_head); 2587ec681f3Smrg} 2597ec681f3Smrg 2607ec681f3Smrguint32_t 2617ec681f3Smrgutil_sparse_array_free_list_pop_idx(struct util_sparse_array_free_list *fl) 2627ec681f3Smrg{ 2637ec681f3Smrg uint64_t current_head; 2647ec681f3Smrg 2657ec681f3Smrg current_head = p_atomic_read(&fl->head); 2667ec681f3Smrg while (1) { 2677ec681f3Smrg if ((uint32_t)current_head == fl->sentinel) 2687ec681f3Smrg return fl->sentinel; 2697ec681f3Smrg 2707ec681f3Smrg uint32_t head_idx = current_head; /* Index is the bottom 32 bits */ 2717ec681f3Smrg void *head_elem = util_sparse_array_get(fl->arr, head_idx); 2727ec681f3Smrg uint32_t *head_next = (uint32_t *)((char *)head_elem + fl->next_offset); 2737ec681f3Smrg uint64_t new_head = free_list_head(current_head, p_atomic_read(head_next)); 2747ec681f3Smrg uint64_t old_head = p_atomic_cmpxchg(&fl->head, current_head, new_head); 2757ec681f3Smrg if (old_head == current_head) 2767ec681f3Smrg return head_idx; 2777ec681f3Smrg current_head = old_head; 2787ec681f3Smrg } 2797ec681f3Smrg} 2807ec681f3Smrg 2817ec681f3Smrgvoid * 2827ec681f3Smrgutil_sparse_array_free_list_pop_elem(struct util_sparse_array_free_list *fl) 2837ec681f3Smrg{ 2847ec681f3Smrg uint64_t current_head; 2857ec681f3Smrg 2867ec681f3Smrg current_head = p_atomic_read(&fl->head); 2877ec681f3Smrg while (1) { 2887ec681f3Smrg if ((uint32_t)current_head == fl->sentinel) 2897ec681f3Smrg return NULL; 2907ec681f3Smrg 2917ec681f3Smrg uint32_t head_idx = current_head; /* Index is the bottom 32 bits */ 2927ec681f3Smrg void *head_elem = util_sparse_array_get(fl->arr, head_idx); 2937ec681f3Smrg uint32_t *head_next = (uint32_t *)((char *)head_elem + fl->next_offset); 2947ec681f3Smrg uint64_t new_head = free_list_head(current_head, p_atomic_read(head_next)); 2957ec681f3Smrg uint64_t old_head = p_atomic_cmpxchg(&fl->head, current_head, new_head); 2967ec681f3Smrg if (old_head == current_head) 2977ec681f3Smrg return head_elem; 2987ec681f3Smrg current_head = old_head; 2997ec681f3Smrg } 3007ec681f3Smrg} 301