^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) // SPDX-License-Identifier: GPL-2.0-only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * multiorder.c: Multi-order radix tree entry testing
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) * Copyright (c) 2016 Intel Corporation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * Author: Ross Zwisler <ross.zwisler@linux.intel.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) * Author: Matthew Wilcox <matthew.r.wilcox@intel.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) #include <linux/radix-tree.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) #include <linux/slab.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) #include <linux/errno.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #include <pthread.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) #include "test.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) static int item_insert_order(struct xarray *xa, unsigned long index,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) unsigned order)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) XA_STATE_ORDER(xas, xa, index, order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) struct item *item = item_create(index, order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) do {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) xas_lock(&xas);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) xas_store(&xas, item);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) xas_unlock(&xas);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) } while (xas_nomem(&xas, GFP_KERNEL));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) if (!xas_error(&xas))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) free(item);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) return xas_error(&xas);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) void multiorder_iteration(struct xarray *xa)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) XA_STATE(xas, xa, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) struct item *item;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) int i, j, err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) #define NUM_ENTRIES 11
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) int index[NUM_ENTRIES] = {0, 2, 4, 8, 16, 32, 34, 36, 64, 72, 128};
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7};
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) printv(1, "Multiorder iteration test\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) for (i = 0; i < NUM_ENTRIES; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) err = item_insert_order(xa, index[i], order[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) assert(!err);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) for (j = 0; j < 256; j++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) for (i = 0; i < NUM_ENTRIES; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) if (j <= (index[i] | ((1 << order[i]) - 1)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) xas_set(&xas, j);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) xas_for_each(&xas, item, ULONG_MAX) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) int height = order[i] / XA_CHUNK_SHIFT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) int shift = height * XA_CHUNK_SHIFT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) unsigned long mask = (1UL << order[i]) - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) assert((xas.xa_index | mask) == (index[i] | mask));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) assert(xas.xa_node->shift == shift);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) assert(!radix_tree_is_internal_node(item));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) assert((item->index | mask) == (index[i] | mask));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) assert(item->order == order[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) i++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) item_kill_tree(xa);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) void multiorder_tagged_iteration(struct xarray *xa)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) XA_STATE(xas, xa, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) struct item *item;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) int i, j;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) #define MT_NUM_ENTRIES 9
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) int index[MT_NUM_ENTRIES] = {0, 2, 4, 16, 32, 40, 64, 72, 128};
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7};
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) #define TAG_ENTRIES 7
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) int tag_index[TAG_ENTRIES] = {0, 4, 16, 40, 64, 72, 128};
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) printv(1, "Multiorder tagged iteration test\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) for (i = 0; i < MT_NUM_ENTRIES; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) assert(!item_insert_order(xa, index[i], order[i]));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) assert(!xa_marked(xa, XA_MARK_1));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) for (i = 0; i < TAG_ENTRIES; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) xa_set_mark(xa, tag_index[i], XA_MARK_1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) for (j = 0; j < 256; j++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) int k;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) for (i = 0; i < TAG_ENTRIES; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) for (k = i; index[k] < tag_index[i]; k++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) ;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) if (j <= (index[k] | ((1 << order[k]) - 1)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) xas_set(&xas, j);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) xas_for_each_marked(&xas, item, ULONG_MAX, XA_MARK_1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) unsigned long mask;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) for (k = i; index[k] < tag_index[i]; k++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) ;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) mask = (1UL << order[k]) - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) assert((xas.xa_index | mask) == (tag_index[i] | mask));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) assert(!xa_is_internal(item));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) assert((item->index | mask) == (tag_index[i] | mask));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) assert(item->order == order[k]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) i++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) assert(tag_tagged_items(xa, 0, ULONG_MAX, TAG_ENTRIES, XA_MARK_1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) XA_MARK_2) == TAG_ENTRIES);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) for (j = 0; j < 256; j++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) int mask, k;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) for (i = 0; i < TAG_ENTRIES; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) for (k = i; index[k] < tag_index[i]; k++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) ;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) if (j <= (index[k] | ((1 << order[k]) - 1)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) xas_set(&xas, j);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) xas_for_each_marked(&xas, item, ULONG_MAX, XA_MARK_2) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) for (k = i; index[k] < tag_index[i]; k++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) ;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) mask = (1 << order[k]) - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) assert((xas.xa_index | mask) == (tag_index[i] | mask));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) assert(!xa_is_internal(item));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) assert((item->index | mask) == (tag_index[i] | mask));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) assert(item->order == order[k]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) i++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) assert(tag_tagged_items(xa, 1, ULONG_MAX, MT_NUM_ENTRIES * 2, XA_MARK_1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) XA_MARK_0) == TAG_ENTRIES);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) i = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) xas_set(&xas, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) xas_for_each_marked(&xas, item, ULONG_MAX, XA_MARK_0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) assert(xas.xa_index == tag_index[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) i++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) assert(i == TAG_ENTRIES);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) item_kill_tree(xa);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) bool stop_iteration = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) static void *creator_func(void *ptr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) /* 'order' is set up to ensure we have sibling entries */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) unsigned int order = RADIX_TREE_MAP_SHIFT - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) struct radix_tree_root *tree = ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) for (i = 0; i < 10000; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) item_insert_order(tree, 0, order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) item_delete_rcu(tree, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) stop_iteration = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) static void *iterator_func(void *ptr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) XA_STATE(xas, ptr, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) struct item *item;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) while (!stop_iteration) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) rcu_read_lock();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) xas_for_each(&xas, item, ULONG_MAX) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) if (xas_retry(&xas, item))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) item_sanity(item, xas.xa_index);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) rcu_read_unlock();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) static void multiorder_iteration_race(struct xarray *xa)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) const int num_threads = sysconf(_SC_NPROCESSORS_ONLN);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) pthread_t worker_thread[num_threads];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) pthread_create(&worker_thread[0], NULL, &creator_func, xa);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) for (i = 1; i < num_threads; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) pthread_create(&worker_thread[i], NULL, &iterator_func, xa);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) for (i = 0; i < num_threads; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) pthread_join(worker_thread[i], NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) item_kill_tree(xa);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) static DEFINE_XARRAY(array);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) void multiorder_checks(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) multiorder_iteration(&array);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) multiorder_tagged_iteration(&array);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220) multiorder_iteration_race(&array);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) radix_tree_cpu_dead(0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) int __weak main(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) rcu_register_thread();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) radix_tree_init();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) multiorder_checks();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) rcu_unregister_thread();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) }