17ec681f3Smrg/*
27ec681f3Smrg * Copyright © 2019 Intel Corporation
37ec681f3Smrg *
47ec681f3Smrg * Permission is hereby granted, free of charge, to any person obtaining a
57ec681f3Smrg * copy of this software and associated documentation files (the "Software"),
67ec681f3Smrg * to deal in the Software without restriction, including without limitation
77ec681f3Smrg * the rights to use, copy, modify, merge, publish, distribute, sublicense,
87ec681f3Smrg * and/or sell copies of the Software, and to permit persons to whom the
97ec681f3Smrg * Software is furnished to do so, subject to the following conditions:
107ec681f3Smrg *
117ec681f3Smrg * The above copyright notice and this permission notice (including the next
127ec681f3Smrg * paragraph) shall be included in all copies or substantial portions of the
137ec681f3Smrg * Software.
147ec681f3Smrg *
157ec681f3Smrg * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
167ec681f3Smrg * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
177ec681f3Smrg * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
187ec681f3Smrg * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
197ec681f3Smrg * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
207ec681f3Smrg * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
217ec681f3Smrg * IN THE SOFTWARE.
227ec681f3Smrg */
237ec681f3Smrg
247ec681f3Smrg#include "sparse_array.h"
257ec681f3Smrg#include "os_memory.h"
267ec681f3Smrg
277ec681f3Smrg/* Aligning our allocations to 64 has two advantages:
287ec681f3Smrg *
297ec681f3Smrg *  1. On x86 platforms, it means that they are cache-line aligned so we
307ec681f3Smrg *     reduce the likelihood that one of our allocations shares a cache line
317ec681f3Smrg *     with some other allocation.
327ec681f3Smrg *
337ec681f3Smrg *  2. It lets us use the bottom 6 bits of the pointer to store the tree level
347ec681f3Smrg *     of the node so we can avoid some pointer indirections.
357ec681f3Smrg */
367ec681f3Smrg#define NODE_ALLOC_ALIGN 64
377ec681f3Smrg
387ec681f3Smrgvoid
397ec681f3Smrgutil_sparse_array_init(struct util_sparse_array *arr,
407ec681f3Smrg                       size_t elem_size, size_t node_size)
417ec681f3Smrg{
427ec681f3Smrg   memset(arr, 0, sizeof(*arr));
437ec681f3Smrg   arr->elem_size = elem_size;
447ec681f3Smrg   arr->node_size_log2 = util_logbase2_64(node_size);
457ec681f3Smrg   assert(node_size >= 2 && node_size == (1ull << arr->node_size_log2));
467ec681f3Smrg}
477ec681f3Smrg
487ec681f3Smrg#define NODE_PTR_MASK (~((uintptr_t)NODE_ALLOC_ALIGN - 1))
497ec681f3Smrg#define NODE_LEVEL_MASK ((uintptr_t)NODE_ALLOC_ALIGN - 1)
507ec681f3Smrg#define NULL_NODE 0
517ec681f3Smrg
527ec681f3Smrgstatic inline uintptr_t
537ec681f3Smrg_util_sparse_array_node(void *data, unsigned level)
547ec681f3Smrg{
557ec681f3Smrg   assert(data != NULL);
567ec681f3Smrg   assert(((uintptr_t)data & NODE_LEVEL_MASK) == 0);
577ec681f3Smrg   assert((level & NODE_PTR_MASK) == 0);
587ec681f3Smrg   return (uintptr_t)data | level;
597ec681f3Smrg}
607ec681f3Smrg
617ec681f3Smrgstatic inline void *
627ec681f3Smrg_util_sparse_array_node_data(uintptr_t handle)
637ec681f3Smrg{
647ec681f3Smrg   return (void *)(handle & NODE_PTR_MASK);
657ec681f3Smrg}
667ec681f3Smrg
677ec681f3Smrgstatic inline unsigned
687ec681f3Smrg_util_sparse_array_node_level(uintptr_t handle)
697ec681f3Smrg{
707ec681f3Smrg   return handle & NODE_LEVEL_MASK;
717ec681f3Smrg}
727ec681f3Smrg
737ec681f3Smrgstatic inline void
747ec681f3Smrg_util_sparse_array_node_finish(struct util_sparse_array *arr,
757ec681f3Smrg                               uintptr_t node)
767ec681f3Smrg{
777ec681f3Smrg   if (_util_sparse_array_node_level(node) > 0) {
787ec681f3Smrg      uintptr_t *children = _util_sparse_array_node_data(node);
797ec681f3Smrg      size_t node_size = 1ull << arr->node_size_log2;
807ec681f3Smrg      for (size_t i = 0; i < node_size; i++) {
817ec681f3Smrg         if (children[i])
827ec681f3Smrg            _util_sparse_array_node_finish(arr, children[i]);
837ec681f3Smrg      }
847ec681f3Smrg   }
857ec681f3Smrg
867ec681f3Smrg   os_free_aligned(_util_sparse_array_node_data(node));
877ec681f3Smrg}
887ec681f3Smrg
897ec681f3Smrgvoid
907ec681f3Smrgutil_sparse_array_finish(struct util_sparse_array *arr)
917ec681f3Smrg{
927ec681f3Smrg   if (arr->root)
937ec681f3Smrg      _util_sparse_array_node_finish(arr, arr->root);
947ec681f3Smrg}
957ec681f3Smrg
967ec681f3Smrgstatic inline uintptr_t
977ec681f3Smrg_util_sparse_array_node_alloc(struct util_sparse_array *arr,
987ec681f3Smrg                              unsigned level)
997ec681f3Smrg{
1007ec681f3Smrg   size_t size;
1017ec681f3Smrg   if (level == 0) {
1027ec681f3Smrg      size = arr->elem_size << arr->node_size_log2;
1037ec681f3Smrg   } else {
1047ec681f3Smrg      size = sizeof(uintptr_t) << arr->node_size_log2;
1057ec681f3Smrg   }
1067ec681f3Smrg
1077ec681f3Smrg   void *data = os_malloc_aligned(size, NODE_ALLOC_ALIGN);
1087ec681f3Smrg   memset(data, 0, size);
1097ec681f3Smrg
1107ec681f3Smrg   return _util_sparse_array_node(data, level);
1117ec681f3Smrg}
1127ec681f3Smrg
1137ec681f3Smrgstatic inline uintptr_t
1147ec681f3Smrg_util_sparse_array_set_or_free_node(uintptr_t *node_ptr,
1157ec681f3Smrg                                    uintptr_t cmp_node,
1167ec681f3Smrg                                    uintptr_t node)
1177ec681f3Smrg{
1187ec681f3Smrg   uintptr_t prev_node = p_atomic_cmpxchg(node_ptr, cmp_node, node);
1197ec681f3Smrg
1207ec681f3Smrg   if (prev_node != cmp_node) {
1217ec681f3Smrg      /* We lost the race.  Free this one and return the one that was already
1227ec681f3Smrg       * allocated.
1237ec681f3Smrg       */
1247ec681f3Smrg      os_free_aligned(_util_sparse_array_node_data(node));
1257ec681f3Smrg      return prev_node;
1267ec681f3Smrg   } else {
1277ec681f3Smrg      return node;
1287ec681f3Smrg   }
1297ec681f3Smrg}
1307ec681f3Smrg
1317ec681f3Smrgvoid *
1327ec681f3Smrgutil_sparse_array_get(struct util_sparse_array *arr, uint64_t idx)
1337ec681f3Smrg{
1347ec681f3Smrg   const unsigned node_size_log2 = arr->node_size_log2;
1357ec681f3Smrg   uintptr_t root = p_atomic_read(&arr->root);
1367ec681f3Smrg   if (unlikely(!root)) {
1377ec681f3Smrg      unsigned root_level = 0;
1387ec681f3Smrg      uint64_t idx_iter = idx >> node_size_log2;
1397ec681f3Smrg      while (idx_iter) {
1407ec681f3Smrg         idx_iter >>= node_size_log2;
1417ec681f3Smrg         root_level++;
1427ec681f3Smrg      }
1437ec681f3Smrg      uintptr_t new_root = _util_sparse_array_node_alloc(arr, root_level);
1447ec681f3Smrg      root = _util_sparse_array_set_or_free_node(&arr->root,
1457ec681f3Smrg                                                 NULL_NODE, new_root);
1467ec681f3Smrg   }
1477ec681f3Smrg
1487ec681f3Smrg   while (1) {
1497ec681f3Smrg      unsigned root_level = _util_sparse_array_node_level(root);
1507ec681f3Smrg      uint64_t root_idx = idx >> (root_level * node_size_log2);
1517ec681f3Smrg      if (likely(root_idx < (1ull << node_size_log2)))
1527ec681f3Smrg         break;
1537ec681f3Smrg
1547ec681f3Smrg      /* In this case, we have a root but its level is low enough that the
1557ec681f3Smrg       * requested index is out-of-bounds.
1567ec681f3Smrg       */
1577ec681f3Smrg      uintptr_t new_root = _util_sparse_array_node_alloc(arr, root_level + 1);
1587ec681f3Smrg
1597ec681f3Smrg      uintptr_t *new_root_children = _util_sparse_array_node_data(new_root);
1607ec681f3Smrg      new_root_children[0] = root;
1617ec681f3Smrg
1627ec681f3Smrg      /* We only add one at a time instead of the whole tree because it's
1637ec681f3Smrg       * easier to ensure correctness of both the tree building and the
1647ec681f3Smrg       * clean-up path.  Because we're only adding one node we never have to
1657ec681f3Smrg       * worry about trying to free multiple things without freeing the old
1667ec681f3Smrg       * things.
1677ec681f3Smrg       */
1687ec681f3Smrg      root = _util_sparse_array_set_or_free_node(&arr->root, root, new_root);
1697ec681f3Smrg   }
1707ec681f3Smrg
1717ec681f3Smrg   void *node_data = _util_sparse_array_node_data(root);
1727ec681f3Smrg   unsigned node_level = _util_sparse_array_node_level(root);
1737ec681f3Smrg   while (node_level > 0) {
1747ec681f3Smrg      uint64_t child_idx = (idx >> (node_level * node_size_log2)) &
1757ec681f3Smrg                           ((1ull << node_size_log2) - 1);
1767ec681f3Smrg
1777ec681f3Smrg      uintptr_t *children = node_data;
1787ec681f3Smrg      uintptr_t child = p_atomic_read(&children[child_idx]);
1797ec681f3Smrg
1807ec681f3Smrg      if (unlikely(!child)) {
1817ec681f3Smrg         child = _util_sparse_array_node_alloc(arr, node_level - 1);
1827ec681f3Smrg         child = _util_sparse_array_set_or_free_node(&children[child_idx],
1837ec681f3Smrg                                                     NULL_NODE, child);
1847ec681f3Smrg      }
1857ec681f3Smrg
1867ec681f3Smrg      node_data = _util_sparse_array_node_data(child);
1877ec681f3Smrg      node_level = _util_sparse_array_node_level(child);
1887ec681f3Smrg   }
1897ec681f3Smrg
1907ec681f3Smrg   uint64_t elem_idx = idx & ((1ull << node_size_log2) - 1);
1917ec681f3Smrg   return (void *)((char *)node_data + (elem_idx * arr->elem_size));
1927ec681f3Smrg}
1937ec681f3Smrg
1947ec681f3Smrgstatic void
1957ec681f3Smrgvalidate_node_level(struct util_sparse_array *arr,
1967ec681f3Smrg                    uintptr_t node, unsigned level)
1977ec681f3Smrg{
1987ec681f3Smrg   assert(_util_sparse_array_node_level(node) == level);
1997ec681f3Smrg
2007ec681f3Smrg   if (_util_sparse_array_node_level(node) > 0) {
2017ec681f3Smrg      uintptr_t *children = _util_sparse_array_node_data(node);
2027ec681f3Smrg      size_t node_size = 1ull << arr->node_size_log2;
2037ec681f3Smrg      for (size_t i = 0; i < node_size; i++) {
2047ec681f3Smrg         if (children[i])
2057ec681f3Smrg            validate_node_level(arr, children[i], level - 1);
2067ec681f3Smrg      }
2077ec681f3Smrg   }
2087ec681f3Smrg}
2097ec681f3Smrg
2107ec681f3Smrgvoid
2117ec681f3Smrgutil_sparse_array_validate(struct util_sparse_array *arr)
2127ec681f3Smrg{
2137ec681f3Smrg   uintptr_t root = p_atomic_read(&arr->root);
2147ec681f3Smrg   validate_node_level(arr, root, _util_sparse_array_node_level(root));
2157ec681f3Smrg}
2167ec681f3Smrg
2177ec681f3Smrgvoid
2187ec681f3Smrgutil_sparse_array_free_list_init(struct util_sparse_array_free_list *fl,
2197ec681f3Smrg                                 struct util_sparse_array *arr,
2207ec681f3Smrg                                 uint32_t sentinel,
2217ec681f3Smrg                                 uint32_t next_offset)
2227ec681f3Smrg{
2237ec681f3Smrg   fl->head = sentinel;
2247ec681f3Smrg   fl->arr = arr;
2257ec681f3Smrg   fl->sentinel = sentinel;
2267ec681f3Smrg   fl->next_offset = next_offset;
2277ec681f3Smrg}
2287ec681f3Smrg
2297ec681f3Smrgstatic uint64_t
2307ec681f3Smrgfree_list_head(uint64_t old, uint32_t next)
2317ec681f3Smrg{
2327ec681f3Smrg   return ((old & 0xffffffff00000000ull) + 0x100000000ull) | next;
2337ec681f3Smrg}
2347ec681f3Smrg
2357ec681f3Smrgvoid
2367ec681f3Smrgutil_sparse_array_free_list_push(struct util_sparse_array_free_list *fl,
2377ec681f3Smrg                                 uint32_t *items, unsigned num_items)
2387ec681f3Smrg{
2397ec681f3Smrg   assert(num_items > 0);
2407ec681f3Smrg   assert(items[0] != fl->sentinel);
2417ec681f3Smrg   void *last_elem = util_sparse_array_get(fl->arr, items[0]);
2427ec681f3Smrg   uint32_t *last_next = (uint32_t *)((char *)last_elem + fl->next_offset);
2437ec681f3Smrg   for (unsigned i = 1; i < num_items; i++) {
2447ec681f3Smrg      p_atomic_set(last_next, items[i]);
2457ec681f3Smrg      assert(items[i] != fl->sentinel);
2467ec681f3Smrg      last_elem = util_sparse_array_get(fl->arr, items[i]);
2477ec681f3Smrg      last_next = (uint32_t *)((char *)last_elem + fl->next_offset);
2487ec681f3Smrg   }
2497ec681f3Smrg
2507ec681f3Smrg   uint64_t current_head, old_head;
2517ec681f3Smrg   old_head = p_atomic_read(&fl->head);
2527ec681f3Smrg   do {
2537ec681f3Smrg      current_head = old_head;
2547ec681f3Smrg      p_atomic_set(last_next, (uint32_t)current_head); /* Index is the bottom 32 bits */
2557ec681f3Smrg      uint64_t new_head = free_list_head(current_head, items[0]);
2567ec681f3Smrg      old_head = p_atomic_cmpxchg(&fl->head, current_head, new_head);
2577ec681f3Smrg   } while (old_head != current_head);
2587ec681f3Smrg}
2597ec681f3Smrg
2607ec681f3Smrguint32_t
2617ec681f3Smrgutil_sparse_array_free_list_pop_idx(struct util_sparse_array_free_list *fl)
2627ec681f3Smrg{
2637ec681f3Smrg   uint64_t current_head;
2647ec681f3Smrg
2657ec681f3Smrg   current_head = p_atomic_read(&fl->head);
2667ec681f3Smrg   while (1) {
2677ec681f3Smrg      if ((uint32_t)current_head == fl->sentinel)
2687ec681f3Smrg         return fl->sentinel;
2697ec681f3Smrg
2707ec681f3Smrg      uint32_t head_idx = current_head; /* Index is the bottom 32 bits */
2717ec681f3Smrg      void *head_elem = util_sparse_array_get(fl->arr, head_idx);
2727ec681f3Smrg      uint32_t *head_next = (uint32_t *)((char *)head_elem + fl->next_offset);
2737ec681f3Smrg      uint64_t new_head = free_list_head(current_head, p_atomic_read(head_next));
2747ec681f3Smrg      uint64_t old_head = p_atomic_cmpxchg(&fl->head, current_head, new_head);
2757ec681f3Smrg      if (old_head == current_head)
2767ec681f3Smrg         return head_idx;
2777ec681f3Smrg      current_head = old_head;
2787ec681f3Smrg   }
2797ec681f3Smrg}
2807ec681f3Smrg
2817ec681f3Smrgvoid *
2827ec681f3Smrgutil_sparse_array_free_list_pop_elem(struct util_sparse_array_free_list *fl)
2837ec681f3Smrg{
2847ec681f3Smrg   uint64_t current_head;
2857ec681f3Smrg
2867ec681f3Smrg   current_head = p_atomic_read(&fl->head);
2877ec681f3Smrg   while (1) {
2887ec681f3Smrg      if ((uint32_t)current_head == fl->sentinel)
2897ec681f3Smrg         return NULL;
2907ec681f3Smrg
2917ec681f3Smrg      uint32_t head_idx = current_head; /* Index is the bottom 32 bits */
2927ec681f3Smrg      void *head_elem = util_sparse_array_get(fl->arr, head_idx);
2937ec681f3Smrg      uint32_t *head_next = (uint32_t *)((char *)head_elem + fl->next_offset);
2947ec681f3Smrg      uint64_t new_head = free_list_head(current_head, p_atomic_read(head_next));
2957ec681f3Smrg      uint64_t old_head = p_atomic_cmpxchg(&fl->head, current_head, new_head);
2967ec681f3Smrg      if (old_head == current_head)
2977ec681f3Smrg         return head_elem;
2987ec681f3Smrg      current_head = old_head;
2997ec681f3Smrg   }
3007ec681f3Smrg}
301