1 // SPDX-License-Identifier: GPL-2.0+ 1 // SPDX-License-Identifier: GPL-2.0+ 2 /* 2 /* 3 * test_xarray.c: Test the XArray API 3 * test_xarray.c: Test the XArray API 4 * Copyright (c) 2017-2018 Microsoft Corporati 4 * Copyright (c) 2017-2018 Microsoft Corporation 5 * Copyright (c) 2019-2020 Oracle 5 * Copyright (c) 2019-2020 Oracle 6 * Author: Matthew Wilcox <willy@infradead.org 6 * Author: Matthew Wilcox <willy@infradead.org> 7 */ 7 */ 8 8 9 #include <linux/xarray.h> 9 #include <linux/xarray.h> 10 #include <linux/module.h> 10 #include <linux/module.h> 11 11 12 static unsigned int tests_run; 12 static unsigned int tests_run; 13 static unsigned int tests_passed; 13 static unsigned int tests_passed; 14 14 15 static const unsigned int order_limit = 15 static const unsigned int order_limit = 16 IS_ENABLED(CONFIG_XARRAY_MULTI 16 IS_ENABLED(CONFIG_XARRAY_MULTI) ? BITS_PER_LONG : 1; 17 17 18 #ifndef XA_DEBUG 18 #ifndef XA_DEBUG 19 # ifdef __KERNEL__ 19 # ifdef __KERNEL__ 20 void xa_dump(const struct xarray *xa) { } 20 void xa_dump(const struct xarray *xa) { } 21 # endif 21 # endif 22 #undef XA_BUG_ON 22 #undef XA_BUG_ON 23 #define XA_BUG_ON(xa, x) do { 23 #define XA_BUG_ON(xa, x) do { \ 24 tests_run++; 24 tests_run++; \ 25 if (x) { 25 if (x) { \ 26 printk("BUG at %s:%d\n", __fun 26 printk("BUG at %s:%d\n", __func__, __LINE__); \ 27 xa_dump(xa); 27 xa_dump(xa); \ 28 dump_stack(); 28 dump_stack(); \ 29 } else { 29 } else { \ 30 tests_passed++; 30 tests_passed++; \ 31 } 31 } \ 32 } while (0) 32 } while (0) 33 #endif 33 #endif 34 34 35 static void *xa_mk_index(unsigned long index) 35 static void *xa_mk_index(unsigned long index) 36 { 36 { 37 return xa_mk_value(index & LONG_MAX); 37 return xa_mk_value(index & LONG_MAX); 38 } 38 } 39 39 40 static void *xa_store_index(struct xarray *xa, 40 static void *xa_store_index(struct xarray *xa, unsigned long index, gfp_t gfp) 41 { 41 { 42 return xa_store(xa, index, xa_mk_index 42 return xa_store(xa, index, xa_mk_index(index), gfp); 43 } 43 } 44 44 45 static void xa_insert_index(struct xarray *xa, 45 static void xa_insert_index(struct xarray *xa, unsigned long index) 46 { 46 { 47 XA_BUG_ON(xa, xa_insert(xa, index, xa_ 47 XA_BUG_ON(xa, xa_insert(xa, index, xa_mk_index(index), 48 GFP_KERNEL) != 48 GFP_KERNEL) != 0); 49 } 49 } 50 50 51 static void xa_alloc_index(struct xarray *xa, 51 static void xa_alloc_index(struct xarray *xa, unsigned long index, gfp_t gfp) 52 { 52 { 53 u32 id; 53 u32 id; 54 54 55 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_ 55 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(index), xa_limit_32b, 56 gfp) != 0); 56 gfp) != 0); 57 XA_BUG_ON(xa, id != index); 57 XA_BUG_ON(xa, id != index); 58 } 58 } 59 59 60 static void xa_erase_index(struct xarray *xa, 60 static void xa_erase_index(struct xarray *xa, unsigned long index) 61 { 61 { 62 XA_BUG_ON(xa, xa_erase(xa, index) != x 62 XA_BUG_ON(xa, xa_erase(xa, index) != xa_mk_index(index)); 63 XA_BUG_ON(xa, xa_load(xa, index) != NU 63 XA_BUG_ON(xa, xa_load(xa, index) != NULL); 64 } 64 } 65 65 66 /* 66 /* 67 * If anyone needs this, please move it to xar 67 * If anyone needs this, please move it to xarray.c. We have no current 68 * users outside the test suite because all cu 68 * users outside the test suite because all current multislot users want 69 * to use the advanced API. 69 * to use the advanced API. 70 */ 70 */ 71 static void *xa_store_order(struct xarray *xa, 71 static void *xa_store_order(struct xarray *xa, unsigned long index, 72 unsigned order, void *entry, g 72 unsigned order, void *entry, gfp_t gfp) 73 { 73 { 74 XA_STATE_ORDER(xas, xa, index, order); 74 XA_STATE_ORDER(xas, xa, index, order); 75 void *curr; 75 void *curr; 76 76 77 do { 77 do { 78 xas_lock(&xas); 78 xas_lock(&xas); 79 curr = xas_store(&xas, entry); 79 curr = xas_store(&xas, entry); 80 xas_unlock(&xas); 80 xas_unlock(&xas); 81 } while (xas_nomem(&xas, gfp)); 81 } while (xas_nomem(&xas, gfp)); 82 82 83 return curr; 83 return curr; 84 } 84 } 85 85 86 static noinline void check_xa_err(struct xarra 86 static noinline void check_xa_err(struct xarray *xa) 87 { 87 { 88 XA_BUG_ON(xa, xa_err(xa_store_index(xa 88 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 0, GFP_NOWAIT)) != 0); 89 XA_BUG_ON(xa, xa_err(xa_erase(xa, 0)) 89 XA_BUG_ON(xa, xa_err(xa_erase(xa, 0)) != 0); 90 #ifndef __KERNEL__ 90 #ifndef __KERNEL__ 91 /* The kernel does not fail GFP_NOWAIT 91 /* The kernel does not fail GFP_NOWAIT allocations */ 92 XA_BUG_ON(xa, xa_err(xa_store_index(xa 92 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM); 93 XA_BUG_ON(xa, xa_err(xa_store_index(xa 93 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_NOWAIT)) != -ENOMEM); 94 #endif 94 #endif 95 XA_BUG_ON(xa, xa_err(xa_store_index(xa 95 XA_BUG_ON(xa, xa_err(xa_store_index(xa, 1, GFP_KERNEL)) != 0); 96 XA_BUG_ON(xa, xa_err(xa_store(xa, 1, x 96 XA_BUG_ON(xa, xa_err(xa_store(xa, 1, xa_mk_value(0), GFP_KERNEL)) != 0); 97 XA_BUG_ON(xa, xa_err(xa_erase(xa, 1)) 97 XA_BUG_ON(xa, xa_err(xa_erase(xa, 1)) != 0); 98 // kills the test-suite :-( 98 // kills the test-suite :-( 99 // XA_BUG_ON(xa, xa_err(xa_store(xa, 0, x 99 // XA_BUG_ON(xa, xa_err(xa_store(xa, 0, xa_mk_internal(0), 0)) != -EINVAL); 100 } 100 } 101 101 102 static noinline void check_xas_retry(struct xa 102 static noinline void check_xas_retry(struct xarray *xa) 103 { 103 { 104 XA_STATE(xas, xa, 0); 104 XA_STATE(xas, xa, 0); 105 void *entry; 105 void *entry; 106 106 107 xa_store_index(xa, 0, GFP_KERNEL); 107 xa_store_index(xa, 0, GFP_KERNEL); 108 xa_store_index(xa, 1, GFP_KERNEL); 108 xa_store_index(xa, 1, GFP_KERNEL); 109 109 110 rcu_read_lock(); 110 rcu_read_lock(); 111 XA_BUG_ON(xa, xas_find(&xas, ULONG_MAX 111 XA_BUG_ON(xa, xas_find(&xas, ULONG_MAX) != xa_mk_value(0)); 112 xa_erase_index(xa, 1); 112 xa_erase_index(xa, 1); 113 XA_BUG_ON(xa, !xa_is_retry(xas_reload( 113 XA_BUG_ON(xa, !xa_is_retry(xas_reload(&xas))); 114 XA_BUG_ON(xa, xas_retry(&xas, NULL)); 114 XA_BUG_ON(xa, xas_retry(&xas, NULL)); 115 XA_BUG_ON(xa, xas_retry(&xas, xa_mk_va 115 XA_BUG_ON(xa, xas_retry(&xas, xa_mk_value(0))); 116 xas_reset(&xas); 116 xas_reset(&xas); 117 XA_BUG_ON(xa, xas.xa_node != XAS_RESTA 117 XA_BUG_ON(xa, xas.xa_node != XAS_RESTART); 118 XA_BUG_ON(xa, xas_next_entry(&xas, ULO 118 XA_BUG_ON(xa, xas_next_entry(&xas, ULONG_MAX) != xa_mk_value(0)); 119 XA_BUG_ON(xa, xas.xa_node != NULL); 119 XA_BUG_ON(xa, xas.xa_node != NULL); 120 rcu_read_unlock(); 120 rcu_read_unlock(); 121 121 122 XA_BUG_ON(xa, xa_store_index(xa, 1, GF 122 XA_BUG_ON(xa, xa_store_index(xa, 1, GFP_KERNEL) != NULL); 123 123 124 rcu_read_lock(); 124 rcu_read_lock(); 125 XA_BUG_ON(xa, !xa_is_internal(xas_relo 125 XA_BUG_ON(xa, !xa_is_internal(xas_reload(&xas))); 126 xas.xa_node = XAS_RESTART; 126 xas.xa_node = XAS_RESTART; 127 XA_BUG_ON(xa, xas_next_entry(&xas, ULO 127 XA_BUG_ON(xa, xas_next_entry(&xas, ULONG_MAX) != xa_mk_value(0)); 128 rcu_read_unlock(); 128 rcu_read_unlock(); 129 129 130 /* Make sure we can iterate through re 130 /* Make sure we can iterate through retry entries */ 131 xas_lock(&xas); 131 xas_lock(&xas); 132 xas_set(&xas, 0); 132 xas_set(&xas, 0); 133 xas_store(&xas, XA_RETRY_ENTRY); 133 xas_store(&xas, XA_RETRY_ENTRY); 134 xas_set(&xas, 1); 134 xas_set(&xas, 1); 135 xas_store(&xas, XA_RETRY_ENTRY); 135 xas_store(&xas, XA_RETRY_ENTRY); 136 136 137 xas_set(&xas, 0); 137 xas_set(&xas, 0); 138 xas_for_each(&xas, entry, ULONG_MAX) { 138 xas_for_each(&xas, entry, ULONG_MAX) { 139 xas_store(&xas, xa_mk_index(xa 139 xas_store(&xas, xa_mk_index(xas.xa_index)); 140 } 140 } 141 xas_unlock(&xas); 141 xas_unlock(&xas); 142 142 143 xa_erase_index(xa, 0); 143 xa_erase_index(xa, 0); 144 xa_erase_index(xa, 1); 144 xa_erase_index(xa, 1); 145 } 145 } 146 146 147 static noinline void check_xa_load(struct xarr 147 static noinline void check_xa_load(struct xarray *xa) 148 { 148 { 149 unsigned long i, j; 149 unsigned long i, j; 150 150 151 for (i = 0; i < 1024; i++) { 151 for (i = 0; i < 1024; i++) { 152 for (j = 0; j < 1024; j++) { 152 for (j = 0; j < 1024; j++) { 153 void *entry = xa_load( 153 void *entry = xa_load(xa, j); 154 if (j < i) 154 if (j < i) 155 XA_BUG_ON(xa, 155 XA_BUG_ON(xa, xa_to_value(entry) != j); 156 else 156 else 157 XA_BUG_ON(xa, 157 XA_BUG_ON(xa, entry); 158 } 158 } 159 XA_BUG_ON(xa, xa_store_index(x 159 XA_BUG_ON(xa, xa_store_index(xa, i, GFP_KERNEL) != NULL); 160 } 160 } 161 161 162 for (i = 0; i < 1024; i++) { 162 for (i = 0; i < 1024; i++) { 163 for (j = 0; j < 1024; j++) { 163 for (j = 0; j < 1024; j++) { 164 void *entry = xa_load( 164 void *entry = xa_load(xa, j); 165 if (j >= i) 165 if (j >= i) 166 XA_BUG_ON(xa, 166 XA_BUG_ON(xa, xa_to_value(entry) != j); 167 else 167 else 168 XA_BUG_ON(xa, 168 XA_BUG_ON(xa, entry); 169 } 169 } 170 xa_erase_index(xa, i); 170 xa_erase_index(xa, i); 171 } 171 } 172 XA_BUG_ON(xa, !xa_empty(xa)); 172 XA_BUG_ON(xa, !xa_empty(xa)); 173 } 173 } 174 174 175 static noinline void check_xa_mark_1(struct xa 175 static noinline void check_xa_mark_1(struct xarray *xa, unsigned long index) 176 { 176 { 177 unsigned int order; 177 unsigned int order; 178 unsigned int max_order = IS_ENABLED(CO 178 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 8 : 1; 179 179 180 /* NULL elements have no marks set */ 180 /* NULL elements have no marks set */ 181 XA_BUG_ON(xa, xa_get_mark(xa, index, X 181 XA_BUG_ON(xa, xa_get_mark(xa, index, XA_MARK_0)); 182 xa_set_mark(xa, index, XA_MARK_0); 182 xa_set_mark(xa, index, XA_MARK_0); 183 XA_BUG_ON(xa, xa_get_mark(xa, index, X 183 XA_BUG_ON(xa, xa_get_mark(xa, index, XA_MARK_0)); 184 184 185 /* Storing a pointer will not make a m 185 /* Storing a pointer will not make a mark appear */ 186 XA_BUG_ON(xa, xa_store_index(xa, index 186 XA_BUG_ON(xa, xa_store_index(xa, index, GFP_KERNEL) != NULL); 187 XA_BUG_ON(xa, xa_get_mark(xa, index, X 187 XA_BUG_ON(xa, xa_get_mark(xa, index, XA_MARK_0)); 188 xa_set_mark(xa, index, XA_MARK_0); 188 xa_set_mark(xa, index, XA_MARK_0); 189 XA_BUG_ON(xa, !xa_get_mark(xa, index, 189 XA_BUG_ON(xa, !xa_get_mark(xa, index, XA_MARK_0)); 190 190 191 /* Setting one mark will not set anoth 191 /* Setting one mark will not set another mark */ 192 XA_BUG_ON(xa, xa_get_mark(xa, index + 192 XA_BUG_ON(xa, xa_get_mark(xa, index + 1, XA_MARK_0)); 193 XA_BUG_ON(xa, xa_get_mark(xa, index, X 193 XA_BUG_ON(xa, xa_get_mark(xa, index, XA_MARK_1)); 194 194 195 /* Storing NULL clears marks, and they 195 /* Storing NULL clears marks, and they can't be set again */ 196 xa_erase_index(xa, index); 196 xa_erase_index(xa, index); 197 XA_BUG_ON(xa, !xa_empty(xa)); 197 XA_BUG_ON(xa, !xa_empty(xa)); 198 XA_BUG_ON(xa, xa_get_mark(xa, index, X 198 XA_BUG_ON(xa, xa_get_mark(xa, index, XA_MARK_0)); 199 xa_set_mark(xa, index, XA_MARK_0); 199 xa_set_mark(xa, index, XA_MARK_0); 200 XA_BUG_ON(xa, xa_get_mark(xa, index, X 200 XA_BUG_ON(xa, xa_get_mark(xa, index, XA_MARK_0)); 201 201 202 /* 202 /* 203 * Storing a multi-index entry over en 203 * Storing a multi-index entry over entries with marks gives the 204 * entire entry the union of the marks 204 * entire entry the union of the marks 205 */ 205 */ 206 BUG_ON((index % 4) != 0); 206 BUG_ON((index % 4) != 0); 207 for (order = 2; order < max_order; ord 207 for (order = 2; order < max_order; order++) { 208 unsigned long base = round_dow 208 unsigned long base = round_down(index, 1UL << order); 209 unsigned long next = base + (1 209 unsigned long next = base + (1UL << order); 210 unsigned long i; 210 unsigned long i; 211 211 212 XA_BUG_ON(xa, xa_store_index(x 212 XA_BUG_ON(xa, xa_store_index(xa, index + 1, GFP_KERNEL)); 213 xa_set_mark(xa, index + 1, XA_ 213 xa_set_mark(xa, index + 1, XA_MARK_0); 214 XA_BUG_ON(xa, xa_store_index(x 214 XA_BUG_ON(xa, xa_store_index(xa, index + 2, GFP_KERNEL)); 215 xa_set_mark(xa, index + 2, XA_ 215 xa_set_mark(xa, index + 2, XA_MARK_2); 216 XA_BUG_ON(xa, xa_store_index(x 216 XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL)); 217 xa_store_order(xa, index, orde 217 xa_store_order(xa, index, order, xa_mk_index(index), 218 GFP_KERNEL); 218 GFP_KERNEL); 219 for (i = base; i < next; i++) 219 for (i = base; i < next; i++) { 220 XA_STATE(xas, xa, i); 220 XA_STATE(xas, xa, i); 221 unsigned int seen = 0; 221 unsigned int seen = 0; 222 void *entry; 222 void *entry; 223 223 224 XA_BUG_ON(xa, !xa_get_ 224 XA_BUG_ON(xa, !xa_get_mark(xa, i, XA_MARK_0)); 225 XA_BUG_ON(xa, xa_get_m 225 XA_BUG_ON(xa, xa_get_mark(xa, i, XA_MARK_1)); 226 XA_BUG_ON(xa, !xa_get_ 226 XA_BUG_ON(xa, !xa_get_mark(xa, i, XA_MARK_2)); 227 227 228 /* We should see two e 228 /* We should see two elements in the array */ 229 rcu_read_lock(); 229 rcu_read_lock(); 230 xas_for_each(&xas, ent 230 xas_for_each(&xas, entry, ULONG_MAX) 231 seen++; 231 seen++; 232 rcu_read_unlock(); 232 rcu_read_unlock(); 233 XA_BUG_ON(xa, seen != 233 XA_BUG_ON(xa, seen != 2); 234 234 235 /* One of which is mar 235 /* One of which is marked */ 236 xas_set(&xas, 0); 236 xas_set(&xas, 0); 237 seen = 0; 237 seen = 0; 238 rcu_read_lock(); 238 rcu_read_lock(); 239 xas_for_each_marked(&x 239 xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0) 240 seen++; 240 seen++; 241 rcu_read_unlock(); 241 rcu_read_unlock(); 242 XA_BUG_ON(xa, seen != 242 XA_BUG_ON(xa, seen != 1); 243 } 243 } 244 XA_BUG_ON(xa, xa_get_mark(xa, 244 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_0)); 245 XA_BUG_ON(xa, xa_get_mark(xa, 245 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_1)); 246 XA_BUG_ON(xa, xa_get_mark(xa, 246 XA_BUG_ON(xa, xa_get_mark(xa, next, XA_MARK_2)); 247 xa_erase_index(xa, index); 247 xa_erase_index(xa, index); 248 xa_erase_index(xa, next); 248 xa_erase_index(xa, next); 249 XA_BUG_ON(xa, !xa_empty(xa)); 249 XA_BUG_ON(xa, !xa_empty(xa)); 250 } 250 } 251 XA_BUG_ON(xa, !xa_empty(xa)); 251 XA_BUG_ON(xa, !xa_empty(xa)); 252 } 252 } 253 253 254 static noinline void check_xa_mark_2(struct xa 254 static noinline void check_xa_mark_2(struct xarray *xa) 255 { 255 { 256 XA_STATE(xas, xa, 0); 256 XA_STATE(xas, xa, 0); 257 unsigned long index; 257 unsigned long index; 258 unsigned int count = 0; 258 unsigned int count = 0; 259 void *entry; 259 void *entry; 260 260 261 xa_store_index(xa, 0, GFP_KERNEL); 261 xa_store_index(xa, 0, GFP_KERNEL); 262 xa_set_mark(xa, 0, XA_MARK_0); 262 xa_set_mark(xa, 0, XA_MARK_0); 263 xas_lock(&xas); 263 xas_lock(&xas); 264 xas_load(&xas); 264 xas_load(&xas); 265 xas_init_marks(&xas); 265 xas_init_marks(&xas); 266 xas_unlock(&xas); 266 xas_unlock(&xas); 267 XA_BUG_ON(xa, !xa_get_mark(xa, 0, XA_M 267 XA_BUG_ON(xa, !xa_get_mark(xa, 0, XA_MARK_0) == 0); 268 268 269 for (index = 3500; index < 4500; index 269 for (index = 3500; index < 4500; index++) { 270 xa_store_index(xa, index, GFP_ 270 xa_store_index(xa, index, GFP_KERNEL); 271 xa_set_mark(xa, index, XA_MARK 271 xa_set_mark(xa, index, XA_MARK_0); 272 } 272 } 273 273 274 xas_reset(&xas); 274 xas_reset(&xas); 275 rcu_read_lock(); 275 rcu_read_lock(); 276 xas_for_each_marked(&xas, entry, ULONG 276 xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0) 277 count++; 277 count++; 278 rcu_read_unlock(); 278 rcu_read_unlock(); 279 XA_BUG_ON(xa, count != 1000); 279 XA_BUG_ON(xa, count != 1000); 280 280 281 xas_lock(&xas); 281 xas_lock(&xas); 282 xas_for_each(&xas, entry, ULONG_MAX) { 282 xas_for_each(&xas, entry, ULONG_MAX) { 283 xas_init_marks(&xas); 283 xas_init_marks(&xas); 284 XA_BUG_ON(xa, !xa_get_mark(xa, 284 XA_BUG_ON(xa, !xa_get_mark(xa, xas.xa_index, XA_MARK_0)); 285 XA_BUG_ON(xa, !xas_get_mark(&x 285 XA_BUG_ON(xa, !xas_get_mark(&xas, XA_MARK_0)); 286 } 286 } 287 xas_unlock(&xas); 287 xas_unlock(&xas); 288 288 289 xa_destroy(xa); 289 xa_destroy(xa); 290 } 290 } 291 291 292 static noinline void check_xa_mark_3(struct xa 292 static noinline void check_xa_mark_3(struct xarray *xa) 293 { 293 { 294 #ifdef CONFIG_XARRAY_MULTI 294 #ifdef CONFIG_XARRAY_MULTI 295 XA_STATE(xas, xa, 0x41); 295 XA_STATE(xas, xa, 0x41); 296 void *entry; 296 void *entry; 297 int count = 0; 297 int count = 0; 298 298 299 xa_store_order(xa, 0x40, 2, xa_mk_inde 299 xa_store_order(xa, 0x40, 2, xa_mk_index(0x40), GFP_KERNEL); 300 xa_set_mark(xa, 0x41, XA_MARK_0); 300 xa_set_mark(xa, 0x41, XA_MARK_0); 301 301 302 rcu_read_lock(); 302 rcu_read_lock(); 303 xas_for_each_marked(&xas, entry, ULONG 303 xas_for_each_marked(&xas, entry, ULONG_MAX, XA_MARK_0) { 304 count++; 304 count++; 305 XA_BUG_ON(xa, entry != xa_mk_i 305 XA_BUG_ON(xa, entry != xa_mk_index(0x40)); 306 } 306 } 307 XA_BUG_ON(xa, count != 1); 307 XA_BUG_ON(xa, count != 1); 308 rcu_read_unlock(); 308 rcu_read_unlock(); 309 xa_destroy(xa); 309 xa_destroy(xa); 310 #endif 310 #endif 311 } 311 } 312 312 313 static noinline void check_xa_mark(struct xarr 313 static noinline void check_xa_mark(struct xarray *xa) 314 { 314 { 315 unsigned long index; 315 unsigned long index; 316 316 317 for (index = 0; index < 16384; index + 317 for (index = 0; index < 16384; index += 4) 318 check_xa_mark_1(xa, index); 318 check_xa_mark_1(xa, index); 319 319 320 check_xa_mark_2(xa); 320 check_xa_mark_2(xa); 321 check_xa_mark_3(xa); 321 check_xa_mark_3(xa); 322 } 322 } 323 323 324 static noinline void check_xa_shrink(struct xa 324 static noinline void check_xa_shrink(struct xarray *xa) 325 { 325 { 326 XA_STATE(xas, xa, 1); 326 XA_STATE(xas, xa, 1); 327 struct xa_node *node; 327 struct xa_node *node; 328 unsigned int order; 328 unsigned int order; 329 unsigned int max_order = IS_ENABLED(CO 329 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 15 : 1; 330 330 331 XA_BUG_ON(xa, !xa_empty(xa)); 331 XA_BUG_ON(xa, !xa_empty(xa)); 332 XA_BUG_ON(xa, xa_store_index(xa, 0, GF 332 XA_BUG_ON(xa, xa_store_index(xa, 0, GFP_KERNEL) != NULL); 333 XA_BUG_ON(xa, xa_store_index(xa, 1, GF 333 XA_BUG_ON(xa, xa_store_index(xa, 1, GFP_KERNEL) != NULL); 334 334 335 /* 335 /* 336 * Check that erasing the entry at 1 s 336 * Check that erasing the entry at 1 shrinks the tree and properly 337 * marks the node as being deleted. 337 * marks the node as being deleted. 338 */ 338 */ 339 xas_lock(&xas); 339 xas_lock(&xas); 340 XA_BUG_ON(xa, xas_load(&xas) != xa_mk_ 340 XA_BUG_ON(xa, xas_load(&xas) != xa_mk_value(1)); 341 node = xas.xa_node; 341 node = xas.xa_node; 342 XA_BUG_ON(xa, xa_entry_locked(xa, node 342 XA_BUG_ON(xa, xa_entry_locked(xa, node, 0) != xa_mk_value(0)); 343 XA_BUG_ON(xa, xas_store(&xas, NULL) != 343 XA_BUG_ON(xa, xas_store(&xas, NULL) != xa_mk_value(1)); 344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); 344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); 345 XA_BUG_ON(xa, xas.xa_node != XAS_BOUND 345 XA_BUG_ON(xa, xas.xa_node != XAS_BOUNDS); 346 XA_BUG_ON(xa, xa_entry_locked(xa, node 346 XA_BUG_ON(xa, xa_entry_locked(xa, node, 0) != XA_RETRY_ENTRY); 347 XA_BUG_ON(xa, xas_load(&xas) != NULL); 347 XA_BUG_ON(xa, xas_load(&xas) != NULL); 348 xas_unlock(&xas); 348 xas_unlock(&xas); 349 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_ 349 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); 350 xa_erase_index(xa, 0); 350 xa_erase_index(xa, 0); 351 XA_BUG_ON(xa, !xa_empty(xa)); 351 XA_BUG_ON(xa, !xa_empty(xa)); 352 352 353 for (order = 0; order < max_order; ord 353 for (order = 0; order < max_order; order++) { 354 unsigned long max = (1UL << or 354 unsigned long max = (1UL << order) - 1; 355 xa_store_order(xa, 0, order, x 355 xa_store_order(xa, 0, order, xa_mk_value(0), GFP_KERNEL); 356 XA_BUG_ON(xa, xa_load(xa, max) 356 XA_BUG_ON(xa, xa_load(xa, max) != xa_mk_value(0)); 357 XA_BUG_ON(xa, xa_load(xa, max 357 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); 358 rcu_read_lock(); 358 rcu_read_lock(); 359 node = xa_head(xa); 359 node = xa_head(xa); 360 rcu_read_unlock(); 360 rcu_read_unlock(); 361 XA_BUG_ON(xa, xa_store_index(x 361 XA_BUG_ON(xa, xa_store_index(xa, ULONG_MAX, GFP_KERNEL) != 362 NULL); 362 NULL); 363 rcu_read_lock(); 363 rcu_read_lock(); 364 XA_BUG_ON(xa, xa_head(xa) == n 364 XA_BUG_ON(xa, xa_head(xa) == node); 365 rcu_read_unlock(); 365 rcu_read_unlock(); 366 XA_BUG_ON(xa, xa_load(xa, max 366 XA_BUG_ON(xa, xa_load(xa, max + 1) != NULL); 367 xa_erase_index(xa, ULONG_MAX); 367 xa_erase_index(xa, ULONG_MAX); 368 XA_BUG_ON(xa, xa->xa_head != n 368 XA_BUG_ON(xa, xa->xa_head != node); 369 xa_erase_index(xa, 0); 369 xa_erase_index(xa, 0); 370 } 370 } 371 } 371 } 372 372 373 static noinline void check_insert(struct xarra 373 static noinline void check_insert(struct xarray *xa) 374 { 374 { 375 unsigned long i; 375 unsigned long i; 376 376 377 for (i = 0; i < 1024; i++) { 377 for (i = 0; i < 1024; i++) { 378 xa_insert_index(xa, i); 378 xa_insert_index(xa, i); 379 XA_BUG_ON(xa, xa_load(xa, i - 379 XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); 380 XA_BUG_ON(xa, xa_load(xa, i + 380 XA_BUG_ON(xa, xa_load(xa, i + 1) != NULL); 381 xa_erase_index(xa, i); 381 xa_erase_index(xa, i); 382 } 382 } 383 383 384 for (i = 10; i < BITS_PER_LONG; i++) { 384 for (i = 10; i < BITS_PER_LONG; i++) { 385 xa_insert_index(xa, 1UL << i); 385 xa_insert_index(xa, 1UL << i); 386 XA_BUG_ON(xa, xa_load(xa, (1UL 386 XA_BUG_ON(xa, xa_load(xa, (1UL << i) - 1) != NULL); 387 XA_BUG_ON(xa, xa_load(xa, (1UL 387 XA_BUG_ON(xa, xa_load(xa, (1UL << i) + 1) != NULL); 388 xa_erase_index(xa, 1UL << i); 388 xa_erase_index(xa, 1UL << i); 389 389 390 xa_insert_index(xa, (1UL << i) 390 xa_insert_index(xa, (1UL << i) - 1); 391 XA_BUG_ON(xa, xa_load(xa, (1UL 391 XA_BUG_ON(xa, xa_load(xa, (1UL << i) - 2) != NULL); 392 XA_BUG_ON(xa, xa_load(xa, 1UL 392 XA_BUG_ON(xa, xa_load(xa, 1UL << i) != NULL); 393 xa_erase_index(xa, (1UL << i) 393 xa_erase_index(xa, (1UL << i) - 1); 394 } 394 } 395 395 396 xa_insert_index(xa, ~0UL); 396 xa_insert_index(xa, ~0UL); 397 XA_BUG_ON(xa, xa_load(xa, 0UL) != NULL 397 XA_BUG_ON(xa, xa_load(xa, 0UL) != NULL); 398 XA_BUG_ON(xa, xa_load(xa, ~1UL) != NUL 398 XA_BUG_ON(xa, xa_load(xa, ~1UL) != NULL); 399 xa_erase_index(xa, ~0UL); 399 xa_erase_index(xa, ~0UL); 400 400 401 XA_BUG_ON(xa, !xa_empty(xa)); 401 XA_BUG_ON(xa, !xa_empty(xa)); 402 } 402 } 403 403 404 static noinline void check_cmpxchg(struct xarr 404 static noinline void check_cmpxchg(struct xarray *xa) 405 { 405 { 406 void *FIVE = xa_mk_value(5); 406 void *FIVE = xa_mk_value(5); 407 void *SIX = xa_mk_value(6); 407 void *SIX = xa_mk_value(6); 408 void *LOTS = xa_mk_value(12345678); 408 void *LOTS = xa_mk_value(12345678); 409 409 410 XA_BUG_ON(xa, !xa_empty(xa)); 410 XA_BUG_ON(xa, !xa_empty(xa)); 411 XA_BUG_ON(xa, xa_store_index(xa, 12345 411 XA_BUG_ON(xa, xa_store_index(xa, 12345678, GFP_KERNEL) != NULL); 412 XA_BUG_ON(xa, xa_insert(xa, 12345678, 412 XA_BUG_ON(xa, xa_insert(xa, 12345678, xa, GFP_KERNEL) != -EBUSY); 413 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, 413 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, SIX, FIVE, GFP_KERNEL) != LOTS); 414 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, 414 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, LOTS, FIVE, GFP_KERNEL) != LOTS); 415 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, 415 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, FIVE, LOTS, GFP_KERNEL) != FIVE); 416 XA_BUG_ON(xa, xa_cmpxchg(xa, 5, FIVE, 416 XA_BUG_ON(xa, xa_cmpxchg(xa, 5, FIVE, NULL, GFP_KERNEL) != NULL); 417 XA_BUG_ON(xa, xa_cmpxchg(xa, 5, NULL, 417 XA_BUG_ON(xa, xa_cmpxchg(xa, 5, NULL, FIVE, GFP_KERNEL) != NULL); 418 XA_BUG_ON(xa, xa_insert(xa, 5, FIVE, G 418 XA_BUG_ON(xa, xa_insert(xa, 5, FIVE, GFP_KERNEL) != -EBUSY); 419 XA_BUG_ON(xa, xa_cmpxchg(xa, 5, FIVE, 419 XA_BUG_ON(xa, xa_cmpxchg(xa, 5, FIVE, NULL, GFP_KERNEL) != FIVE); 420 XA_BUG_ON(xa, xa_insert(xa, 5, FIVE, G 420 XA_BUG_ON(xa, xa_insert(xa, 5, FIVE, GFP_KERNEL) == -EBUSY); 421 xa_erase_index(xa, 12345678); 421 xa_erase_index(xa, 12345678); 422 xa_erase_index(xa, 5); 422 xa_erase_index(xa, 5); 423 XA_BUG_ON(xa, !xa_empty(xa)); 423 XA_BUG_ON(xa, !xa_empty(xa)); 424 } 424 } 425 425 426 static noinline void check_cmpxchg_order(struc << 427 { << 428 #ifdef CONFIG_XARRAY_MULTI << 429 void *FIVE = xa_mk_value(5); << 430 unsigned int i, order = 3; << 431 << 432 XA_BUG_ON(xa, xa_store_order(xa, 0, or << 433 << 434 /* Check entry FIVE has the order save << 435 XA_BUG_ON(xa, xa_get_order(xa, xa_to_v << 436 << 437 /* Check all the tied indexes have the << 438 for (i = 0; i < (1 << order); i++) { << 439 XA_BUG_ON(xa, xa_load(xa, i) ! << 440 XA_BUG_ON(xa, xa_get_order(xa, << 441 } << 442 << 443 /* Ensure that nothing is stored at in << 444 XA_BUG_ON(xa, xa_load(xa, 1 << order) << 445 << 446 /* << 447 * Additionally, keep the node informa << 448 * '1 << order' << 449 */ << 450 XA_BUG_ON(xa, xa_store_order(xa, 1 << << 451 for (i = (1 << order); i < (1 << order << 452 XA_BUG_ON(xa, xa_load(xa, i) ! << 453 XA_BUG_ON(xa, xa_get_order(xa, << 454 } << 455 << 456 /* Conditionally replace FIVE entry at << 457 XA_BUG_ON(xa, xa_cmpxchg(xa, 0, FIVE, << 458 << 459 /* Verify the order is lost at FIVE (a << 460 XA_BUG_ON(xa, xa_get_order(xa, xa_to_v << 461 << 462 /* Verify the order and entries are lo << 463 for (i = 0; i < (1 << order); i++) { << 464 XA_BUG_ON(xa, xa_load(xa, i) ! << 465 XA_BUG_ON(xa, xa_get_order(xa, << 466 } << 467 << 468 /* Verify node and order are kept at ' << 469 for (i = (1 << order); i < (1 << order << 470 XA_BUG_ON(xa, xa_load(xa, i) ! << 471 XA_BUG_ON(xa, xa_get_order(xa, << 472 } << 473 << 474 xa_store_order(xa, 0, BITS_PER_LONG - << 475 XA_BUG_ON(xa, !xa_empty(xa)); << 476 #endif << 477 } << 478 << 479 static noinline void check_reserve(struct xarr 426 static noinline void check_reserve(struct xarray *xa) 480 { 427 { 481 void *entry; 428 void *entry; 482 unsigned long index; 429 unsigned long index; 483 int count; 430 int count; 484 431 485 /* An array with a reserved entry is n 432 /* An array with a reserved entry is not empty */ 486 XA_BUG_ON(xa, !xa_empty(xa)); 433 XA_BUG_ON(xa, !xa_empty(xa)); 487 XA_BUG_ON(xa, xa_reserve(xa, 12345678, 434 XA_BUG_ON(xa, xa_reserve(xa, 12345678, GFP_KERNEL) != 0); 488 XA_BUG_ON(xa, xa_empty(xa)); 435 XA_BUG_ON(xa, xa_empty(xa)); 489 XA_BUG_ON(xa, xa_load(xa, 12345678)); 436 XA_BUG_ON(xa, xa_load(xa, 12345678)); 490 xa_release(xa, 12345678); 437 xa_release(xa, 12345678); 491 XA_BUG_ON(xa, !xa_empty(xa)); 438 XA_BUG_ON(xa, !xa_empty(xa)); 492 439 493 /* Releasing a used entry does nothing 440 /* Releasing a used entry does nothing */ 494 XA_BUG_ON(xa, xa_reserve(xa, 12345678, 441 XA_BUG_ON(xa, xa_reserve(xa, 12345678, GFP_KERNEL) != 0); 495 XA_BUG_ON(xa, xa_store_index(xa, 12345 442 XA_BUG_ON(xa, xa_store_index(xa, 12345678, GFP_NOWAIT) != NULL); 496 xa_release(xa, 12345678); 443 xa_release(xa, 12345678); 497 xa_erase_index(xa, 12345678); 444 xa_erase_index(xa, 12345678); 498 XA_BUG_ON(xa, !xa_empty(xa)); 445 XA_BUG_ON(xa, !xa_empty(xa)); 499 446 500 /* cmpxchg sees a reserved entry as ZE 447 /* cmpxchg sees a reserved entry as ZERO */ 501 XA_BUG_ON(xa, xa_reserve(xa, 12345678, 448 XA_BUG_ON(xa, xa_reserve(xa, 12345678, GFP_KERNEL) != 0); 502 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, 449 XA_BUG_ON(xa, xa_cmpxchg(xa, 12345678, XA_ZERO_ENTRY, 503 xa_mk_value(12 450 xa_mk_value(12345678), GFP_NOWAIT) != NULL); 504 xa_release(xa, 12345678); 451 xa_release(xa, 12345678); 505 xa_erase_index(xa, 12345678); 452 xa_erase_index(xa, 12345678); 506 XA_BUG_ON(xa, !xa_empty(xa)); 453 XA_BUG_ON(xa, !xa_empty(xa)); 507 454 508 /* xa_insert treats it as busy */ 455 /* xa_insert treats it as busy */ 509 XA_BUG_ON(xa, xa_reserve(xa, 12345678, 456 XA_BUG_ON(xa, xa_reserve(xa, 12345678, GFP_KERNEL) != 0); 510 XA_BUG_ON(xa, xa_insert(xa, 12345678, 457 XA_BUG_ON(xa, xa_insert(xa, 12345678, xa_mk_value(12345678), 0) != 511 -EBUSY); 458 -EBUSY); 512 XA_BUG_ON(xa, xa_empty(xa)); 459 XA_BUG_ON(xa, xa_empty(xa)); 513 XA_BUG_ON(xa, xa_erase(xa, 12345678) ! 460 XA_BUG_ON(xa, xa_erase(xa, 12345678) != NULL); 514 XA_BUG_ON(xa, !xa_empty(xa)); 461 XA_BUG_ON(xa, !xa_empty(xa)); 515 462 516 /* Can iterate through a reserved entr 463 /* Can iterate through a reserved entry */ 517 xa_store_index(xa, 5, GFP_KERNEL); 464 xa_store_index(xa, 5, GFP_KERNEL); 518 XA_BUG_ON(xa, xa_reserve(xa, 6, GFP_KE 465 XA_BUG_ON(xa, xa_reserve(xa, 6, GFP_KERNEL) != 0); 519 xa_store_index(xa, 7, GFP_KERNEL); 466 xa_store_index(xa, 7, GFP_KERNEL); 520 467 521 count = 0; 468 count = 0; 522 xa_for_each(xa, index, entry) { 469 xa_for_each(xa, index, entry) { 523 XA_BUG_ON(xa, index != 5 && in 470 XA_BUG_ON(xa, index != 5 && index != 7); 524 count++; 471 count++; 525 } 472 } 526 XA_BUG_ON(xa, count != 2); 473 XA_BUG_ON(xa, count != 2); 527 474 528 /* If we free a reserved entry, we sho 475 /* If we free a reserved entry, we should be able to allocate it */ 529 if (xa->xa_flags & XA_FLAGS_ALLOC) { 476 if (xa->xa_flags & XA_FLAGS_ALLOC) { 530 u32 id; 477 u32 id; 531 478 532 XA_BUG_ON(xa, xa_alloc(xa, &id 479 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_value(8), 533 XA_LIM 480 XA_LIMIT(5, 10), GFP_KERNEL) != 0); 534 XA_BUG_ON(xa, id != 8); 481 XA_BUG_ON(xa, id != 8); 535 482 536 xa_release(xa, 6); 483 xa_release(xa, 6); 537 XA_BUG_ON(xa, xa_alloc(xa, &id 484 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_value(6), 538 XA_LIM 485 XA_LIMIT(5, 10), GFP_KERNEL) != 0); 539 XA_BUG_ON(xa, id != 6); 486 XA_BUG_ON(xa, id != 6); 540 } 487 } 541 488 542 xa_destroy(xa); 489 xa_destroy(xa); 543 } 490 } 544 491 545 static noinline void check_xas_erase(struct xa 492 static noinline void check_xas_erase(struct xarray *xa) 546 { 493 { 547 XA_STATE(xas, xa, 0); 494 XA_STATE(xas, xa, 0); 548 void *entry; 495 void *entry; 549 unsigned long i, j; 496 unsigned long i, j; 550 497 551 for (i = 0; i < 200; i++) { 498 for (i = 0; i < 200; i++) { 552 for (j = i; j < 2 * i + 17; j+ 499 for (j = i; j < 2 * i + 17; j++) { 553 xas_set(&xas, j); 500 xas_set(&xas, j); 554 do { 501 do { 555 xas_lock(&xas) 502 xas_lock(&xas); 556 xas_store(&xas 503 xas_store(&xas, xa_mk_index(j)); 557 xas_unlock(&xa 504 xas_unlock(&xas); 558 } while (xas_nomem(&xa 505 } while (xas_nomem(&xas, GFP_KERNEL)); 559 } 506 } 560 507 561 xas_set(&xas, ULONG_MAX); 508 xas_set(&xas, ULONG_MAX); 562 do { 509 do { 563 xas_lock(&xas); 510 xas_lock(&xas); 564 xas_store(&xas, xa_mk_ 511 xas_store(&xas, xa_mk_value(0)); 565 xas_unlock(&xas); 512 xas_unlock(&xas); 566 } while (xas_nomem(&xas, GFP_K 513 } while (xas_nomem(&xas, GFP_KERNEL)); 567 514 568 xas_lock(&xas); 515 xas_lock(&xas); 569 xas_store(&xas, NULL); 516 xas_store(&xas, NULL); 570 517 571 xas_set(&xas, 0); 518 xas_set(&xas, 0); 572 j = i; 519 j = i; 573 xas_for_each(&xas, entry, ULON 520 xas_for_each(&xas, entry, ULONG_MAX) { 574 XA_BUG_ON(xa, entry != 521 XA_BUG_ON(xa, entry != xa_mk_index(j)); 575 xas_store(&xas, NULL); 522 xas_store(&xas, NULL); 576 j++; 523 j++; 577 } 524 } 578 xas_unlock(&xas); 525 xas_unlock(&xas); 579 XA_BUG_ON(xa, !xa_empty(xa)); 526 XA_BUG_ON(xa, !xa_empty(xa)); 580 } 527 } 581 } 528 } 582 529 583 #ifdef CONFIG_XARRAY_MULTI 530 #ifdef CONFIG_XARRAY_MULTI 584 static noinline void check_multi_store_1(struc 531 static noinline void check_multi_store_1(struct xarray *xa, unsigned long index, 585 unsigned int order) 532 unsigned int order) 586 { 533 { 587 XA_STATE(xas, xa, index); 534 XA_STATE(xas, xa, index); 588 unsigned long min = index & ~((1UL << 535 unsigned long min = index & ~((1UL << order) - 1); 589 unsigned long max = min + (1UL << orde 536 unsigned long max = min + (1UL << order); 590 537 591 xa_store_order(xa, index, order, xa_mk 538 xa_store_order(xa, index, order, xa_mk_index(index), GFP_KERNEL); 592 XA_BUG_ON(xa, xa_load(xa, min) != xa_m 539 XA_BUG_ON(xa, xa_load(xa, min) != xa_mk_index(index)); 593 XA_BUG_ON(xa, xa_load(xa, max - 1) != 540 XA_BUG_ON(xa, xa_load(xa, max - 1) != xa_mk_index(index)); 594 XA_BUG_ON(xa, xa_load(xa, max) != NULL 541 XA_BUG_ON(xa, xa_load(xa, max) != NULL); 595 XA_BUG_ON(xa, xa_load(xa, min - 1) != 542 XA_BUG_ON(xa, xa_load(xa, min - 1) != NULL); 596 543 597 xas_lock(&xas); 544 xas_lock(&xas); 598 XA_BUG_ON(xa, xas_store(&xas, xa_mk_in 545 XA_BUG_ON(xa, xas_store(&xas, xa_mk_index(min)) != xa_mk_index(index)); 599 xas_unlock(&xas); 546 xas_unlock(&xas); 600 XA_BUG_ON(xa, xa_load(xa, min) != xa_m 547 XA_BUG_ON(xa, xa_load(xa, min) != xa_mk_index(min)); 601 XA_BUG_ON(xa, xa_load(xa, max - 1) != 548 XA_BUG_ON(xa, xa_load(xa, max - 1) != xa_mk_index(min)); 602 XA_BUG_ON(xa, xa_load(xa, max) != NULL 549 XA_BUG_ON(xa, xa_load(xa, max) != NULL); 603 XA_BUG_ON(xa, xa_load(xa, min - 1) != 550 XA_BUG_ON(xa, xa_load(xa, min - 1) != NULL); 604 551 605 xa_erase_index(xa, min); 552 xa_erase_index(xa, min); 606 XA_BUG_ON(xa, !xa_empty(xa)); 553 XA_BUG_ON(xa, !xa_empty(xa)); 607 } 554 } 608 555 609 static noinline void check_multi_store_2(struc 556 static noinline void check_multi_store_2(struct xarray *xa, unsigned long index, 610 unsigned int order) 557 unsigned int order) 611 { 558 { 612 XA_STATE(xas, xa, index); 559 XA_STATE(xas, xa, index); 613 xa_store_order(xa, index, order, xa_mk 560 xa_store_order(xa, index, order, xa_mk_value(0), GFP_KERNEL); 614 561 615 xas_lock(&xas); 562 xas_lock(&xas); 616 XA_BUG_ON(xa, xas_store(&xas, xa_mk_va 563 XA_BUG_ON(xa, xas_store(&xas, xa_mk_value(1)) != xa_mk_value(0)); 617 XA_BUG_ON(xa, xas.xa_index != index); 564 XA_BUG_ON(xa, xas.xa_index != index); 618 XA_BUG_ON(xa, xas_store(&xas, NULL) != 565 XA_BUG_ON(xa, xas_store(&xas, NULL) != xa_mk_value(1)); 619 xas_unlock(&xas); 566 xas_unlock(&xas); 620 XA_BUG_ON(xa, !xa_empty(xa)); 567 XA_BUG_ON(xa, !xa_empty(xa)); 621 } 568 } 622 569 623 static noinline void check_multi_store_3(struc 570 static noinline void check_multi_store_3(struct xarray *xa, unsigned long index, 624 unsigned int order) 571 unsigned int order) 625 { 572 { 626 XA_STATE(xas, xa, 0); 573 XA_STATE(xas, xa, 0); 627 void *entry; 574 void *entry; 628 int n = 0; 575 int n = 0; 629 576 630 xa_store_order(xa, index, order, xa_mk 577 xa_store_order(xa, index, order, xa_mk_index(index), GFP_KERNEL); 631 578 632 xas_lock(&xas); 579 xas_lock(&xas); 633 xas_for_each(&xas, entry, ULONG_MAX) { 580 xas_for_each(&xas, entry, ULONG_MAX) { 634 XA_BUG_ON(xa, entry != xa_mk_i 581 XA_BUG_ON(xa, entry != xa_mk_index(index)); 635 n++; 582 n++; 636 } 583 } 637 XA_BUG_ON(xa, n != 1); 584 XA_BUG_ON(xa, n != 1); 638 xas_set(&xas, index + 1); 585 xas_set(&xas, index + 1); 639 xas_for_each(&xas, entry, ULONG_MAX) { 586 xas_for_each(&xas, entry, ULONG_MAX) { 640 XA_BUG_ON(xa, entry != xa_mk_i 587 XA_BUG_ON(xa, entry != xa_mk_index(index)); 641 n++; 588 n++; 642 } 589 } 643 XA_BUG_ON(xa, n != 2); 590 XA_BUG_ON(xa, n != 2); 644 xas_unlock(&xas); 591 xas_unlock(&xas); 645 592 646 xa_destroy(xa); 593 xa_destroy(xa); 647 } 594 } 648 #endif 595 #endif 649 596 650 static noinline void check_multi_store(struct 597 static noinline void check_multi_store(struct xarray *xa) 651 { 598 { 652 #ifdef CONFIG_XARRAY_MULTI 599 #ifdef CONFIG_XARRAY_MULTI 653 unsigned long i, j, k; 600 unsigned long i, j, k; 654 unsigned int max_order = (sizeof(long) 601 unsigned int max_order = (sizeof(long) == 4) ? 30 : 60; 655 602 656 /* Loading from any position returns t 603 /* Loading from any position returns the same value */ 657 xa_store_order(xa, 0, 1, xa_mk_value(0 604 xa_store_order(xa, 0, 1, xa_mk_value(0), GFP_KERNEL); 658 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_ 605 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); 659 XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_ 606 XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_value(0)); 660 XA_BUG_ON(xa, xa_load(xa, 2) != NULL); 607 XA_BUG_ON(xa, xa_load(xa, 2) != NULL); 661 rcu_read_lock(); 608 rcu_read_lock(); 662 XA_BUG_ON(xa, xa_to_node(xa_head(xa))- 609 XA_BUG_ON(xa, xa_to_node(xa_head(xa))->count != 2); 663 XA_BUG_ON(xa, xa_to_node(xa_head(xa))- 610 XA_BUG_ON(xa, xa_to_node(xa_head(xa))->nr_values != 2); 664 rcu_read_unlock(); 611 rcu_read_unlock(); 665 612 666 /* Storing adjacent to the value does 613 /* Storing adjacent to the value does not alter the value */ 667 xa_store(xa, 3, xa, GFP_KERNEL); 614 xa_store(xa, 3, xa, GFP_KERNEL); 668 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_ 615 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(0)); 669 XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_ 616 XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_value(0)); 670 XA_BUG_ON(xa, xa_load(xa, 2) != NULL); 617 XA_BUG_ON(xa, xa_load(xa, 2) != NULL); 671 rcu_read_lock(); 618 rcu_read_lock(); 672 XA_BUG_ON(xa, xa_to_node(xa_head(xa))- 619 XA_BUG_ON(xa, xa_to_node(xa_head(xa))->count != 3); 673 XA_BUG_ON(xa, xa_to_node(xa_head(xa))- 620 XA_BUG_ON(xa, xa_to_node(xa_head(xa))->nr_values != 2); 674 rcu_read_unlock(); 621 rcu_read_unlock(); 675 622 676 /* Overwriting multiple indexes works 623 /* Overwriting multiple indexes works */ 677 xa_store_order(xa, 0, 2, xa_mk_value(1 624 xa_store_order(xa, 0, 2, xa_mk_value(1), GFP_KERNEL); 678 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_ 625 XA_BUG_ON(xa, xa_load(xa, 0) != xa_mk_value(1)); 679 XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_ 626 XA_BUG_ON(xa, xa_load(xa, 1) != xa_mk_value(1)); 680 XA_BUG_ON(xa, xa_load(xa, 2) != xa_mk_ 627 XA_BUG_ON(xa, xa_load(xa, 2) != xa_mk_value(1)); 681 XA_BUG_ON(xa, xa_load(xa, 3) != xa_mk_ 628 XA_BUG_ON(xa, xa_load(xa, 3) != xa_mk_value(1)); 682 XA_BUG_ON(xa, xa_load(xa, 4) != NULL); 629 XA_BUG_ON(xa, xa_load(xa, 4) != NULL); 683 rcu_read_lock(); 630 rcu_read_lock(); 684 XA_BUG_ON(xa, xa_to_node(xa_head(xa))- 631 XA_BUG_ON(xa, xa_to_node(xa_head(xa))->count != 4); 685 XA_BUG_ON(xa, xa_to_node(xa_head(xa))- 632 XA_BUG_ON(xa, xa_to_node(xa_head(xa))->nr_values != 4); 686 rcu_read_unlock(); 633 rcu_read_unlock(); 687 634 688 /* We can erase multiple values with a 635 /* We can erase multiple values with a single store */ 689 xa_store_order(xa, 0, BITS_PER_LONG - 636 xa_store_order(xa, 0, BITS_PER_LONG - 1, NULL, GFP_KERNEL); 690 XA_BUG_ON(xa, !xa_empty(xa)); 637 XA_BUG_ON(xa, !xa_empty(xa)); 691 638 692 /* Even when the first slot is empty b 639 /* Even when the first slot is empty but the others aren't */ 693 xa_store_index(xa, 1, GFP_KERNEL); 640 xa_store_index(xa, 1, GFP_KERNEL); 694 xa_store_index(xa, 2, GFP_KERNEL); 641 xa_store_index(xa, 2, GFP_KERNEL); 695 xa_store_order(xa, 0, 2, NULL, GFP_KER 642 xa_store_order(xa, 0, 2, NULL, GFP_KERNEL); 696 XA_BUG_ON(xa, !xa_empty(xa)); 643 XA_BUG_ON(xa, !xa_empty(xa)); 697 644 698 for (i = 0; i < max_order; i++) { 645 for (i = 0; i < max_order; i++) { 699 for (j = 0; j < max_order; j++ 646 for (j = 0; j < max_order; j++) { 700 xa_store_order(xa, 0, 647 xa_store_order(xa, 0, i, xa_mk_index(i), GFP_KERNEL); 701 xa_store_order(xa, 0, 648 xa_store_order(xa, 0, j, xa_mk_index(j), GFP_KERNEL); 702 649 703 for (k = 0; k < max_or 650 for (k = 0; k < max_order; k++) { 704 void *entry = 651 void *entry = xa_load(xa, (1UL << k) - 1); 705 if ((i < k) && 652 if ((i < k) && (j < k)) 706 XA_BUG 653 XA_BUG_ON(xa, entry != NULL); 707 else 654 else 708 XA_BUG 655 XA_BUG_ON(xa, entry != xa_mk_index(j)); 709 } 656 } 710 657 711 xa_erase(xa, 0); 658 xa_erase(xa, 0); 712 XA_BUG_ON(xa, !xa_empt 659 XA_BUG_ON(xa, !xa_empty(xa)); 713 } 660 } 714 } 661 } 715 662 716 for (i = 0; i < 20; i++) { 663 for (i = 0; i < 20; i++) { 717 check_multi_store_1(xa, 200, i 664 check_multi_store_1(xa, 200, i); 718 check_multi_store_1(xa, 0, i); 665 check_multi_store_1(xa, 0, i); 719 check_multi_store_1(xa, (1UL < 666 check_multi_store_1(xa, (1UL << i) + 1, i); 720 } 667 } 721 check_multi_store_2(xa, 4095, 9); 668 check_multi_store_2(xa, 4095, 9); 722 669 723 for (i = 1; i < 20; i++) { 670 for (i = 1; i < 20; i++) { 724 check_multi_store_3(xa, 0, i); 671 check_multi_store_3(xa, 0, i); 725 check_multi_store_3(xa, 1UL << 672 check_multi_store_3(xa, 1UL << i, i); 726 } 673 } 727 #endif 674 #endif 728 } 675 } 729 676 730 #ifdef CONFIG_XARRAY_MULTI << 731 /* mimics page cache __filemap_add_folio() */ << 732 static noinline void check_xa_multi_store_adv_ << 733 << 734 << 735 << 736 { << 737 XA_STATE(xas, xa, index); << 738 unsigned int nrpages = 1UL << order; << 739 << 740 /* users are responsible for index ali << 741 XA_BUG_ON(xa, index & (nrpages - 1)); << 742 << 743 xas_set_order(&xas, index, order); << 744 << 745 do { << 746 xas_lock_irq(&xas); << 747 xas_store(&xas, p); << 748 xas_unlock_irq(&xas); << 749 /* << 750 * In our selftest case the on << 751 * there not to be enough memo << 752 * entire page cache, so verif << 753 * into here. The xas_nomem() << 754 * that condition for us so to << 755 */ << 756 XA_BUG_ON(xa, xas_error(&xas) << 757 } while (xas_nomem(&xas, GFP_KERNEL)); << 758 << 759 XA_BUG_ON(xa, xas_error(&xas)); << 760 XA_BUG_ON(xa, xa_load(xa, index) != p) << 761 } << 762 << 763 /* mimics page_cache_delete() */ << 764 static noinline void check_xa_multi_store_adv_ << 765 << 766 << 767 { << 768 XA_STATE(xas, xa, index); << 769 << 770 xas_set_order(&xas, index, order); << 771 xas_store(&xas, NULL); << 772 xas_init_marks(&xas); << 773 } << 774 << 775 static noinline void check_xa_multi_store_adv_ << 776 << 777 << 778 { << 779 xa_lock_irq(xa); << 780 check_xa_multi_store_adv_del_entry(xa, << 781 xa_unlock_irq(xa); << 782 } << 783 << 784 /* mimics page cache filemap_get_entry() */ << 785 static noinline void *test_get_entry(struct xa << 786 { << 787 XA_STATE(xas, xa, index); << 788 void *p; << 789 static unsigned int loops = 0; << 790 << 791 rcu_read_lock(); << 792 repeat: << 793 xas_reset(&xas); << 794 p = xas_load(&xas); << 795 if (xas_retry(&xas, p)) << 796 goto repeat; << 797 rcu_read_unlock(); << 798 << 799 /* << 800 * This is not part of the page cache, << 801 * aggressive and does not want to tru << 802 * test it, and for order 20 (4 GiB bl << 803 * over a million entries which can ca << 804 * APIs won't be stupid, proper page c << 805 * order so when using a larger order << 806 */ << 807 if (++loops % XA_CHECK_SCHED == 0) << 808 schedule(); << 809 << 810 return p; << 811 } << 812 << 813 static unsigned long some_val = 0xdeadbeef; << 814 static unsigned long some_val_2 = 0xdeaddead; << 815 << 816 /* mimics the page cache usage */ << 817 static noinline void check_xa_multi_store_adv( << 818 << 819 << 820 { << 821 unsigned int nrpages = 1UL << order; << 822 unsigned long index, base, next_index, << 823 unsigned int i; << 824 << 825 index = pos >> PAGE_SHIFT; << 826 base = round_down(index, nrpages); << 827 next_index = round_down(base + nrpages << 828 next_next_index = round_down(next_inde << 829 << 830 check_xa_multi_store_adv_add(xa, base, << 831 << 832 for (i = 0; i < nrpages; i++) << 833 XA_BUG_ON(xa, test_get_entry(x << 834 << 835 XA_BUG_ON(xa, test_get_entry(xa, next_ << 836 << 837 /* Use order 0 for the next item */ << 838 check_xa_multi_store_adv_add(xa, next_ << 839 XA_BUG_ON(xa, test_get_entry(xa, next_ << 840 << 841 /* Remove the next item */ << 842 check_xa_multi_store_adv_delete(xa, ne << 843 << 844 /* Now use order for a new pointer */ << 845 check_xa_multi_store_adv_add(xa, next_ << 846 << 847 for (i = 0; i < nrpages; i++) << 848 XA_BUG_ON(xa, test_get_entry(x << 849 << 850 check_xa_multi_store_adv_delete(xa, ne << 851 check_xa_multi_store_adv_delete(xa, ba << 852 XA_BUG_ON(xa, !xa_empty(xa)); << 853 << 854 /* starting fresh again */ << 855 << 856 /* let's test some holes now */ << 857 << 858 /* hole at base and next_next */ << 859 check_xa_multi_store_adv_add(xa, next_ << 860 << 861 for (i = 0; i < nrpages; i++) << 862 XA_BUG_ON(xa, test_get_entry(x << 863 << 864 for (i = 0; i < nrpages; i++) << 865 XA_BUG_ON(xa, test_get_entry(x << 866 << 867 for (i = 0; i < nrpages; i++) << 868 XA_BUG_ON(xa, test_get_entry(x << 869 << 870 check_xa_multi_store_adv_delete(xa, ne << 871 XA_BUG_ON(xa, !xa_empty(xa)); << 872 << 873 /* hole at base and next */ << 874 << 875 check_xa_multi_store_adv_add(xa, next_ << 876 << 877 for (i = 0; i < nrpages; i++) << 878 XA_BUG_ON(xa, test_get_entry(x << 879 << 880 for (i = 0; i < nrpages; i++) << 881 XA_BUG_ON(xa, test_get_entry(x << 882 << 883 for (i = 0; i < nrpages; i++) << 884 XA_BUG_ON(xa, test_get_entry(x << 885 << 886 check_xa_multi_store_adv_delete(xa, ne << 887 XA_BUG_ON(xa, !xa_empty(xa)); << 888 } << 889 #endif << 890 << 891 static noinline void check_multi_store_advance << 892 { << 893 #ifdef CONFIG_XARRAY_MULTI << 894 unsigned int max_order = IS_ENABLED(CO << 895 unsigned long end = ULONG_MAX/2; << 896 unsigned long pos, i; << 897 << 898 /* << 899 * About 117 million tests below. << 900 */ << 901 for (pos = 7; pos < end; pos = (pos * << 902 for (i = 0; i < max_order; i++ << 903 check_xa_multi_store_a << 904 check_xa_multi_store_a << 905 } << 906 } << 907 #endif << 908 } << 909 << 910 static noinline void check_xa_alloc_1(struct x 677 static noinline void check_xa_alloc_1(struct xarray *xa, unsigned int base) 911 { 678 { 912 int i; 679 int i; 913 u32 id; 680 u32 id; 914 681 915 XA_BUG_ON(xa, !xa_empty(xa)); 682 XA_BUG_ON(xa, !xa_empty(xa)); 916 /* An empty array should assign %base 683 /* An empty array should assign %base to the first alloc */ 917 xa_alloc_index(xa, base, GFP_KERNEL); 684 xa_alloc_index(xa, base, GFP_KERNEL); 918 685 919 /* Erasing it should make the array em 686 /* Erasing it should make the array empty again */ 920 xa_erase_index(xa, base); 687 xa_erase_index(xa, base); 921 XA_BUG_ON(xa, !xa_empty(xa)); 688 XA_BUG_ON(xa, !xa_empty(xa)); 922 689 923 /* And it should assign %base again */ 690 /* And it should assign %base again */ 924 xa_alloc_index(xa, base, GFP_KERNEL); 691 xa_alloc_index(xa, base, GFP_KERNEL); 925 692 926 /* Allocating and then erasing a lot s 693 /* Allocating and then erasing a lot should not lose base */ 927 for (i = base + 1; i < 2 * XA_CHUNK_SI 694 for (i = base + 1; i < 2 * XA_CHUNK_SIZE; i++) 928 xa_alloc_index(xa, i, GFP_KERN 695 xa_alloc_index(xa, i, GFP_KERNEL); 929 for (i = base; i < 2 * XA_CHUNK_SIZE; 696 for (i = base; i < 2 * XA_CHUNK_SIZE; i++) 930 xa_erase_index(xa, i); 697 xa_erase_index(xa, i); 931 xa_alloc_index(xa, base, GFP_KERNEL); 698 xa_alloc_index(xa, base, GFP_KERNEL); 932 699 933 /* Destroying the array should do the 700 /* Destroying the array should do the same as erasing */ 934 xa_destroy(xa); 701 xa_destroy(xa); 935 702 936 /* And it should assign %base again */ 703 /* And it should assign %base again */ 937 xa_alloc_index(xa, base, GFP_KERNEL); 704 xa_alloc_index(xa, base, GFP_KERNEL); 938 705 939 /* The next assigned ID should be base 706 /* The next assigned ID should be base+1 */ 940 xa_alloc_index(xa, base + 1, GFP_KERNE 707 xa_alloc_index(xa, base + 1, GFP_KERNEL); 941 xa_erase_index(xa, base + 1); 708 xa_erase_index(xa, base + 1); 942 709 943 /* Storing a value should mark it used 710 /* Storing a value should mark it used */ 944 xa_store_index(xa, base + 1, GFP_KERNE 711 xa_store_index(xa, base + 1, GFP_KERNEL); 945 xa_alloc_index(xa, base + 2, GFP_KERNE 712 xa_alloc_index(xa, base + 2, GFP_KERNEL); 946 713 947 /* If we then erase base, it should be 714 /* If we then erase base, it should be free */ 948 xa_erase_index(xa, base); 715 xa_erase_index(xa, base); 949 xa_alloc_index(xa, base, GFP_KERNEL); 716 xa_alloc_index(xa, base, GFP_KERNEL); 950 717 951 xa_erase_index(xa, base + 1); 718 xa_erase_index(xa, base + 1); 952 xa_erase_index(xa, base + 2); 719 xa_erase_index(xa, base + 2); 953 720 954 for (i = 1; i < 5000; i++) { 721 for (i = 1; i < 5000; i++) { 955 xa_alloc_index(xa, base + i, G 722 xa_alloc_index(xa, base + i, GFP_KERNEL); 956 } 723 } 957 724 958 xa_destroy(xa); 725 xa_destroy(xa); 959 726 960 /* Check that we fail properly at the 727 /* Check that we fail properly at the limit of allocation */ 961 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_ 728 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(UINT_MAX - 1), 962 XA_LIMIT(UINT_ 729 XA_LIMIT(UINT_MAX - 1, UINT_MAX), 963 GFP_KERNEL) != 730 GFP_KERNEL) != 0); 964 XA_BUG_ON(xa, id != 0xfffffffeU); 731 XA_BUG_ON(xa, id != 0xfffffffeU); 965 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_ 732 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(UINT_MAX), 966 XA_LIMIT(UINT_ 733 XA_LIMIT(UINT_MAX - 1, UINT_MAX), 967 GFP_KERNEL) != 734 GFP_KERNEL) != 0); 968 XA_BUG_ON(xa, id != 0xffffffffU); 735 XA_BUG_ON(xa, id != 0xffffffffU); 969 id = 3; 736 id = 3; 970 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_ 737 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(0), 971 XA_LIMIT(UINT_ 738 XA_LIMIT(UINT_MAX - 1, UINT_MAX), 972 GFP_KERNEL) != 739 GFP_KERNEL) != -EBUSY); 973 XA_BUG_ON(xa, id != 3); 740 XA_BUG_ON(xa, id != 3); 974 xa_destroy(xa); 741 xa_destroy(xa); 975 742 976 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_ 743 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(10), XA_LIMIT(10, 5), 977 GFP_KERNEL) != 744 GFP_KERNEL) != -EBUSY); 978 XA_BUG_ON(xa, xa_store_index(xa, 3, GF 745 XA_BUG_ON(xa, xa_store_index(xa, 3, GFP_KERNEL) != 0); 979 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_ 746 XA_BUG_ON(xa, xa_alloc(xa, &id, xa_mk_index(10), XA_LIMIT(10, 5), 980 GFP_KERNEL) != 747 GFP_KERNEL) != -EBUSY); 981 xa_erase_index(xa, 3); 748 xa_erase_index(xa, 3); 982 XA_BUG_ON(xa, !xa_empty(xa)); 749 XA_BUG_ON(xa, !xa_empty(xa)); 983 } 750 } 984 751 985 static noinline void check_xa_alloc_2(struct x 752 static noinline void check_xa_alloc_2(struct xarray *xa, unsigned int base) 986 { 753 { 987 unsigned int i, id; 754 unsigned int i, id; 988 unsigned long index; 755 unsigned long index; 989 void *entry; 756 void *entry; 990 757 991 /* Allocate and free a NULL and check 758 /* Allocate and free a NULL and check xa_empty() behaves */ 992 XA_BUG_ON(xa, !xa_empty(xa)); 759 XA_BUG_ON(xa, !xa_empty(xa)); 993 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, 760 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0); 994 XA_BUG_ON(xa, id != base); 761 XA_BUG_ON(xa, id != base); 995 XA_BUG_ON(xa, xa_empty(xa)); 762 XA_BUG_ON(xa, xa_empty(xa)); 996 XA_BUG_ON(xa, xa_erase(xa, id) != NULL 763 XA_BUG_ON(xa, xa_erase(xa, id) != NULL); 997 XA_BUG_ON(xa, !xa_empty(xa)); 764 XA_BUG_ON(xa, !xa_empty(xa)); 998 765 999 /* Ditto, but check destroy instead of 766 /* Ditto, but check destroy instead of erase */ 1000 XA_BUG_ON(xa, !xa_empty(xa)); 767 XA_BUG_ON(xa, !xa_empty(xa)); 1001 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, 768 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0); 1002 XA_BUG_ON(xa, id != base); 769 XA_BUG_ON(xa, id != base); 1003 XA_BUG_ON(xa, xa_empty(xa)); 770 XA_BUG_ON(xa, xa_empty(xa)); 1004 xa_destroy(xa); 771 xa_destroy(xa); 1005 XA_BUG_ON(xa, !xa_empty(xa)); 772 XA_BUG_ON(xa, !xa_empty(xa)); 1006 773 1007 for (i = base; i < base + 10; i++) { 774 for (i = base; i < base + 10; i++) { 1008 XA_BUG_ON(xa, xa_alloc(xa, &i 775 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, 1009 GFP_K 776 GFP_KERNEL) != 0); 1010 XA_BUG_ON(xa, id != i); 777 XA_BUG_ON(xa, id != i); 1011 } 778 } 1012 779 1013 XA_BUG_ON(xa, xa_store(xa, 3, xa_mk_i 780 XA_BUG_ON(xa, xa_store(xa, 3, xa_mk_index(3), GFP_KERNEL) != NULL); 1014 XA_BUG_ON(xa, xa_store(xa, 4, xa_mk_i 781 XA_BUG_ON(xa, xa_store(xa, 4, xa_mk_index(4), GFP_KERNEL) != NULL); 1015 XA_BUG_ON(xa, xa_store(xa, 4, NULL, G 782 XA_BUG_ON(xa, xa_store(xa, 4, NULL, GFP_KERNEL) != xa_mk_index(4)); 1016 XA_BUG_ON(xa, xa_erase(xa, 5) != NULL 783 XA_BUG_ON(xa, xa_erase(xa, 5) != NULL); 1017 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, 784 XA_BUG_ON(xa, xa_alloc(xa, &id, NULL, xa_limit_32b, GFP_KERNEL) != 0); 1018 XA_BUG_ON(xa, id != 5); 785 XA_BUG_ON(xa, id != 5); 1019 786 1020 xa_for_each(xa, index, entry) { 787 xa_for_each(xa, index, entry) { 1021 xa_erase_index(xa, index); 788 xa_erase_index(xa, index); 1022 } 789 } 1023 790 1024 for (i = base; i < base + 9; i++) { 791 for (i = base; i < base + 9; i++) { 1025 XA_BUG_ON(xa, xa_erase(xa, i) 792 XA_BUG_ON(xa, xa_erase(xa, i) != NULL); 1026 XA_BUG_ON(xa, xa_empty(xa)); 793 XA_BUG_ON(xa, xa_empty(xa)); 1027 } 794 } 1028 XA_BUG_ON(xa, xa_erase(xa, 8) != NULL 795 XA_BUG_ON(xa, xa_erase(xa, 8) != NULL); 1029 XA_BUG_ON(xa, xa_empty(xa)); 796 XA_BUG_ON(xa, xa_empty(xa)); 1030 XA_BUG_ON(xa, xa_erase(xa, base + 9) 797 XA_BUG_ON(xa, xa_erase(xa, base + 9) != NULL); 1031 XA_BUG_ON(xa, !xa_empty(xa)); 798 XA_BUG_ON(xa, !xa_empty(xa)); 1032 799 1033 xa_destroy(xa); 800 xa_destroy(xa); 1034 } 801 } 1035 802 1036 static noinline void check_xa_alloc_3(struct 803 static noinline void check_xa_alloc_3(struct xarray *xa, unsigned int base) 1037 { 804 { 1038 struct xa_limit limit = XA_LIMIT(1, 0 805 struct xa_limit limit = XA_LIMIT(1, 0x3fff); 1039 u32 next = 0; 806 u32 next = 0; 1040 unsigned int i, id; 807 unsigned int i, id; 1041 unsigned long index; 808 unsigned long index; 1042 void *entry; 809 void *entry; 1043 810 1044 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id 811 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(1), limit, 1045 &next, GFP_KE 812 &next, GFP_KERNEL) != 0); 1046 XA_BUG_ON(xa, id != 1); 813 XA_BUG_ON(xa, id != 1); 1047 814 1048 next = 0x3ffd; 815 next = 0x3ffd; 1049 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id 816 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(0x3ffd), limit, 1050 &next, GFP_KE 817 &next, GFP_KERNEL) != 0); 1051 XA_BUG_ON(xa, id != 0x3ffd); 818 XA_BUG_ON(xa, id != 0x3ffd); 1052 xa_erase_index(xa, 0x3ffd); 819 xa_erase_index(xa, 0x3ffd); 1053 xa_erase_index(xa, 1); 820 xa_erase_index(xa, 1); 1054 XA_BUG_ON(xa, !xa_empty(xa)); 821 XA_BUG_ON(xa, !xa_empty(xa)); 1055 822 1056 for (i = 0x3ffe; i < 0x4003; i++) { 823 for (i = 0x3ffe; i < 0x4003; i++) { 1057 if (i < 0x4000) 824 if (i < 0x4000) 1058 entry = xa_mk_index(i 825 entry = xa_mk_index(i); 1059 else 826 else 1060 entry = xa_mk_index(i 827 entry = xa_mk_index(i - 0x3fff); 1061 XA_BUG_ON(xa, xa_alloc_cyclic 828 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, entry, limit, 1062 &next 829 &next, GFP_KERNEL) != (id == 1)); 1063 XA_BUG_ON(xa, xa_mk_index(id) 830 XA_BUG_ON(xa, xa_mk_index(id) != entry); 1064 } 831 } 1065 832 1066 /* Check wrap-around is handled corre 833 /* Check wrap-around is handled correctly */ 1067 if (base != 0) 834 if (base != 0) 1068 xa_erase_index(xa, base); 835 xa_erase_index(xa, base); 1069 xa_erase_index(xa, base + 1); 836 xa_erase_index(xa, base + 1); 1070 next = UINT_MAX; 837 next = UINT_MAX; 1071 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id 838 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(UINT_MAX), 1072 xa_limit_32b, 839 xa_limit_32b, &next, GFP_KERNEL) != 0); 1073 XA_BUG_ON(xa, id != UINT_MAX); 840 XA_BUG_ON(xa, id != UINT_MAX); 1074 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id 841 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(base), 1075 xa_limit_32b, 842 xa_limit_32b, &next, GFP_KERNEL) != 1); 1076 XA_BUG_ON(xa, id != base); 843 XA_BUG_ON(xa, id != base); 1077 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id 844 XA_BUG_ON(xa, xa_alloc_cyclic(xa, &id, xa_mk_index(base + 1), 1078 xa_limit_32b, 845 xa_limit_32b, &next, GFP_KERNEL) != 0); 1079 XA_BUG_ON(xa, id != base + 1); 846 XA_BUG_ON(xa, id != base + 1); 1080 847 1081 xa_for_each(xa, index, entry) 848 xa_for_each(xa, index, entry) 1082 xa_erase_index(xa, index); 849 xa_erase_index(xa, index); 1083 850 1084 XA_BUG_ON(xa, !xa_empty(xa)); 851 XA_BUG_ON(xa, !xa_empty(xa)); 1085 } 852 } 1086 853 1087 static DEFINE_XARRAY_ALLOC(xa0); 854 static DEFINE_XARRAY_ALLOC(xa0); 1088 static DEFINE_XARRAY_ALLOC1(xa1); 855 static DEFINE_XARRAY_ALLOC1(xa1); 1089 856 1090 static noinline void check_xa_alloc(void) 857 static noinline void check_xa_alloc(void) 1091 { 858 { 1092 check_xa_alloc_1(&xa0, 0); 859 check_xa_alloc_1(&xa0, 0); 1093 check_xa_alloc_1(&xa1, 1); 860 check_xa_alloc_1(&xa1, 1); 1094 check_xa_alloc_2(&xa0, 0); 861 check_xa_alloc_2(&xa0, 0); 1095 check_xa_alloc_2(&xa1, 1); 862 check_xa_alloc_2(&xa1, 1); 1096 check_xa_alloc_3(&xa0, 0); 863 check_xa_alloc_3(&xa0, 0); 1097 check_xa_alloc_3(&xa1, 1); 864 check_xa_alloc_3(&xa1, 1); 1098 } 865 } 1099 866 1100 static noinline void __check_store_iter(struc 867 static noinline void __check_store_iter(struct xarray *xa, unsigned long start, 1101 unsigned int order, u 868 unsigned int order, unsigned int present) 1102 { 869 { 1103 XA_STATE_ORDER(xas, xa, start, order) 870 XA_STATE_ORDER(xas, xa, start, order); 1104 void *entry; 871 void *entry; 1105 unsigned int count = 0; 872 unsigned int count = 0; 1106 873 1107 retry: 874 retry: 1108 xas_lock(&xas); 875 xas_lock(&xas); 1109 xas_for_each_conflict(&xas, entry) { 876 xas_for_each_conflict(&xas, entry) { 1110 XA_BUG_ON(xa, !xa_is_value(en 877 XA_BUG_ON(xa, !xa_is_value(entry)); 1111 XA_BUG_ON(xa, entry < xa_mk_i 878 XA_BUG_ON(xa, entry < xa_mk_index(start)); 1112 XA_BUG_ON(xa, entry > xa_mk_i 879 XA_BUG_ON(xa, entry > xa_mk_index(start + (1UL << order) - 1)); 1113 count++; 880 count++; 1114 } 881 } 1115 xas_store(&xas, xa_mk_index(start)); 882 xas_store(&xas, xa_mk_index(start)); 1116 xas_unlock(&xas); 883 xas_unlock(&xas); 1117 if (xas_nomem(&xas, GFP_KERNEL)) { 884 if (xas_nomem(&xas, GFP_KERNEL)) { 1118 count = 0; 885 count = 0; 1119 goto retry; 886 goto retry; 1120 } 887 } 1121 XA_BUG_ON(xa, xas_error(&xas)); 888 XA_BUG_ON(xa, xas_error(&xas)); 1122 XA_BUG_ON(xa, count != present); 889 XA_BUG_ON(xa, count != present); 1123 XA_BUG_ON(xa, xa_load(xa, start) != x 890 XA_BUG_ON(xa, xa_load(xa, start) != xa_mk_index(start)); 1124 XA_BUG_ON(xa, xa_load(xa, start + (1U 891 XA_BUG_ON(xa, xa_load(xa, start + (1UL << order) - 1) != 1125 xa_mk_index(start)); 892 xa_mk_index(start)); 1126 xa_erase_index(xa, start); 893 xa_erase_index(xa, start); 1127 } 894 } 1128 895 1129 static noinline void check_store_iter(struct 896 static noinline void check_store_iter(struct xarray *xa) 1130 { 897 { 1131 unsigned int i, j; 898 unsigned int i, j; 1132 unsigned int max_order = IS_ENABLED(C 899 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 20 : 1; 1133 900 1134 for (i = 0; i < max_order; i++) { 901 for (i = 0; i < max_order; i++) { 1135 unsigned int min = 1 << i; 902 unsigned int min = 1 << i; 1136 unsigned int max = (2 << i) - 903 unsigned int max = (2 << i) - 1; 1137 __check_store_iter(xa, 0, i, 904 __check_store_iter(xa, 0, i, 0); 1138 XA_BUG_ON(xa, !xa_empty(xa)); 905 XA_BUG_ON(xa, !xa_empty(xa)); 1139 __check_store_iter(xa, min, i 906 __check_store_iter(xa, min, i, 0); 1140 XA_BUG_ON(xa, !xa_empty(xa)); 907 XA_BUG_ON(xa, !xa_empty(xa)); 1141 908 1142 xa_store_index(xa, min, GFP_K 909 xa_store_index(xa, min, GFP_KERNEL); 1143 __check_store_iter(xa, min, i 910 __check_store_iter(xa, min, i, 1); 1144 XA_BUG_ON(xa, !xa_empty(xa)); 911 XA_BUG_ON(xa, !xa_empty(xa)); 1145 xa_store_index(xa, max, GFP_K 912 xa_store_index(xa, max, GFP_KERNEL); 1146 __check_store_iter(xa, min, i 913 __check_store_iter(xa, min, i, 1); 1147 XA_BUG_ON(xa, !xa_empty(xa)); 914 XA_BUG_ON(xa, !xa_empty(xa)); 1148 915 1149 for (j = 0; j < min; j++) 916 for (j = 0; j < min; j++) 1150 xa_store_index(xa, j, 917 xa_store_index(xa, j, GFP_KERNEL); 1151 __check_store_iter(xa, 0, i, 918 __check_store_iter(xa, 0, i, min); 1152 XA_BUG_ON(xa, !xa_empty(xa)); 919 XA_BUG_ON(xa, !xa_empty(xa)); 1153 for (j = 0; j < min; j++) 920 for (j = 0; j < min; j++) 1154 xa_store_index(xa, mi 921 xa_store_index(xa, min + j, GFP_KERNEL); 1155 __check_store_iter(xa, min, i 922 __check_store_iter(xa, min, i, min); 1156 XA_BUG_ON(xa, !xa_empty(xa)); 923 XA_BUG_ON(xa, !xa_empty(xa)); 1157 } 924 } 1158 #ifdef CONFIG_XARRAY_MULTI 925 #ifdef CONFIG_XARRAY_MULTI 1159 xa_store_index(xa, 63, GFP_KERNEL); 926 xa_store_index(xa, 63, GFP_KERNEL); 1160 xa_store_index(xa, 65, GFP_KERNEL); 927 xa_store_index(xa, 65, GFP_KERNEL); 1161 __check_store_iter(xa, 64, 2, 1); 928 __check_store_iter(xa, 64, 2, 1); 1162 xa_erase_index(xa, 63); 929 xa_erase_index(xa, 63); 1163 #endif 930 #endif 1164 XA_BUG_ON(xa, !xa_empty(xa)); 931 XA_BUG_ON(xa, !xa_empty(xa)); 1165 } 932 } 1166 933 1167 static noinline void check_multi_find_1(struc 934 static noinline void check_multi_find_1(struct xarray *xa, unsigned order) 1168 { 935 { 1169 #ifdef CONFIG_XARRAY_MULTI 936 #ifdef CONFIG_XARRAY_MULTI 1170 unsigned long multi = 3 << order; 937 unsigned long multi = 3 << order; 1171 unsigned long next = 4 << order; 938 unsigned long next = 4 << order; 1172 unsigned long index; 939 unsigned long index; 1173 940 1174 xa_store_order(xa, multi, order, xa_m 941 xa_store_order(xa, multi, order, xa_mk_value(multi), GFP_KERNEL); 1175 XA_BUG_ON(xa, xa_store_index(xa, next 942 XA_BUG_ON(xa, xa_store_index(xa, next, GFP_KERNEL) != NULL); 1176 XA_BUG_ON(xa, xa_store_index(xa, next 943 XA_BUG_ON(xa, xa_store_index(xa, next + 1, GFP_KERNEL) != NULL); 1177 944 1178 index = 0; 945 index = 0; 1179 XA_BUG_ON(xa, xa_find(xa, &index, ULO 946 XA_BUG_ON(xa, xa_find(xa, &index, ULONG_MAX, XA_PRESENT) != 1180 xa_mk_value(multi)); 947 xa_mk_value(multi)); 1181 XA_BUG_ON(xa, index != multi); 948 XA_BUG_ON(xa, index != multi); 1182 index = multi + 1; 949 index = multi + 1; 1183 XA_BUG_ON(xa, xa_find(xa, &index, ULO 950 XA_BUG_ON(xa, xa_find(xa, &index, ULONG_MAX, XA_PRESENT) != 1184 xa_mk_value(multi)); 951 xa_mk_value(multi)); 1185 XA_BUG_ON(xa, (index < multi) || (ind 952 XA_BUG_ON(xa, (index < multi) || (index >= next)); 1186 XA_BUG_ON(xa, xa_find_after(xa, &inde 953 XA_BUG_ON(xa, xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT) != 1187 xa_mk_value(next)); 954 xa_mk_value(next)); 1188 XA_BUG_ON(xa, index != next); 955 XA_BUG_ON(xa, index != next); 1189 XA_BUG_ON(xa, xa_find_after(xa, &inde 956 XA_BUG_ON(xa, xa_find_after(xa, &index, next, XA_PRESENT) != NULL); 1190 XA_BUG_ON(xa, index != next); 957 XA_BUG_ON(xa, index != next); 1191 958 1192 xa_erase_index(xa, multi); 959 xa_erase_index(xa, multi); 1193 xa_erase_index(xa, next); 960 xa_erase_index(xa, next); 1194 xa_erase_index(xa, next + 1); 961 xa_erase_index(xa, next + 1); 1195 XA_BUG_ON(xa, !xa_empty(xa)); 962 XA_BUG_ON(xa, !xa_empty(xa)); 1196 #endif 963 #endif 1197 } 964 } 1198 965 1199 static noinline void check_multi_find_2(struc 966 static noinline void check_multi_find_2(struct xarray *xa) 1200 { 967 { 1201 unsigned int max_order = IS_ENABLED(C 968 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 10 : 1; 1202 unsigned int i, j; 969 unsigned int i, j; 1203 void *entry; 970 void *entry; 1204 971 1205 for (i = 0; i < max_order; i++) { 972 for (i = 0; i < max_order; i++) { 1206 unsigned long index = 1UL << 973 unsigned long index = 1UL << i; 1207 for (j = 0; j < index; j++) { 974 for (j = 0; j < index; j++) { 1208 XA_STATE(xas, xa, j + 975 XA_STATE(xas, xa, j + index); 1209 xa_store_index(xa, in 976 xa_store_index(xa, index - 1, GFP_KERNEL); 1210 xa_store_order(xa, in 977 xa_store_order(xa, index, i, xa_mk_index(index), 1211 GFP_K 978 GFP_KERNEL); 1212 rcu_read_lock(); 979 rcu_read_lock(); 1213 xas_for_each(&xas, en 980 xas_for_each(&xas, entry, ULONG_MAX) { 1214 xa_erase_inde 981 xa_erase_index(xa, index); 1215 } 982 } 1216 rcu_read_unlock(); 983 rcu_read_unlock(); 1217 xa_erase_index(xa, in 984 xa_erase_index(xa, index - 1); 1218 XA_BUG_ON(xa, !xa_emp 985 XA_BUG_ON(xa, !xa_empty(xa)); 1219 } 986 } 1220 } 987 } 1221 } 988 } 1222 989 1223 static noinline void check_multi_find_3(struc 990 static noinline void check_multi_find_3(struct xarray *xa) 1224 { 991 { 1225 unsigned int order; 992 unsigned int order; 1226 993 1227 for (order = 5; order < order_limit; 994 for (order = 5; order < order_limit; order++) { 1228 unsigned long index = 1UL << 995 unsigned long index = 1UL << (order - 5); 1229 996 1230 XA_BUG_ON(xa, !xa_empty(xa)); 997 XA_BUG_ON(xa, !xa_empty(xa)); 1231 xa_store_order(xa, 0, order - 998 xa_store_order(xa, 0, order - 4, xa_mk_index(0), GFP_KERNEL); 1232 XA_BUG_ON(xa, xa_find_after(x 999 XA_BUG_ON(xa, xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT)); 1233 xa_erase_index(xa, 0); 1000 xa_erase_index(xa, 0); 1234 } 1001 } 1235 } 1002 } 1236 1003 1237 static noinline void check_find_1(struct xarr 1004 static noinline void check_find_1(struct xarray *xa) 1238 { 1005 { 1239 unsigned long i, j, k; 1006 unsigned long i, j, k; 1240 1007 1241 XA_BUG_ON(xa, !xa_empty(xa)); 1008 XA_BUG_ON(xa, !xa_empty(xa)); 1242 1009 1243 /* 1010 /* 1244 * Check xa_find with all pairs betwe 1011 * Check xa_find with all pairs between 0 and 99 inclusive, 1245 * starting at every index between 0 1012 * starting at every index between 0 and 99 1246 */ 1013 */ 1247 for (i = 0; i < 100; i++) { 1014 for (i = 0; i < 100; i++) { 1248 XA_BUG_ON(xa, xa_store_index( 1015 XA_BUG_ON(xa, xa_store_index(xa, i, GFP_KERNEL) != NULL); 1249 xa_set_mark(xa, i, XA_MARK_0) 1016 xa_set_mark(xa, i, XA_MARK_0); 1250 for (j = 0; j < i; j++) { 1017 for (j = 0; j < i; j++) { 1251 XA_BUG_ON(xa, xa_stor 1018 XA_BUG_ON(xa, xa_store_index(xa, j, GFP_KERNEL) != 1252 NULL) 1019 NULL); 1253 xa_set_mark(xa, j, XA 1020 xa_set_mark(xa, j, XA_MARK_0); 1254 for (k = 0; k < 100; 1021 for (k = 0; k < 100; k++) { 1255 unsigned long 1022 unsigned long index = k; 1256 void *entry = 1023 void *entry = xa_find(xa, &index, ULONG_MAX, 1257 1024 XA_PRESENT); 1258 if (k <= j) 1025 if (k <= j) 1259 XA_BU 1026 XA_BUG_ON(xa, index != j); 1260 else if (k <= 1027 else if (k <= i) 1261 XA_BU 1028 XA_BUG_ON(xa, index != i); 1262 else 1029 else 1263 XA_BU 1030 XA_BUG_ON(xa, entry != NULL); 1264 1031 1265 index = k; 1032 index = k; 1266 entry = xa_fi 1033 entry = xa_find(xa, &index, ULONG_MAX, 1267 1034 XA_MARK_0); 1268 if (k <= j) 1035 if (k <= j) 1269 XA_BU 1036 XA_BUG_ON(xa, index != j); 1270 else if (k <= 1037 else if (k <= i) 1271 XA_BU 1038 XA_BUG_ON(xa, index != i); 1272 else 1039 else 1273 XA_BU 1040 XA_BUG_ON(xa, entry != NULL); 1274 } 1041 } 1275 xa_erase_index(xa, j) 1042 xa_erase_index(xa, j); 1276 XA_BUG_ON(xa, xa_get_ 1043 XA_BUG_ON(xa, xa_get_mark(xa, j, XA_MARK_0)); 1277 XA_BUG_ON(xa, !xa_get 1044 XA_BUG_ON(xa, !xa_get_mark(xa, i, XA_MARK_0)); 1278 } 1045 } 1279 xa_erase_index(xa, i); 1046 xa_erase_index(xa, i); 1280 XA_BUG_ON(xa, xa_get_mark(xa, 1047 XA_BUG_ON(xa, xa_get_mark(xa, i, XA_MARK_0)); 1281 } 1048 } 1282 XA_BUG_ON(xa, !xa_empty(xa)); 1049 XA_BUG_ON(xa, !xa_empty(xa)); 1283 } 1050 } 1284 1051 1285 static noinline void check_find_2(struct xarr 1052 static noinline void check_find_2(struct xarray *xa) 1286 { 1053 { 1287 void *entry; 1054 void *entry; 1288 unsigned long i, j, index; 1055 unsigned long i, j, index; 1289 1056 1290 xa_for_each(xa, index, entry) { 1057 xa_for_each(xa, index, entry) { 1291 XA_BUG_ON(xa, true); 1058 XA_BUG_ON(xa, true); 1292 } 1059 } 1293 1060 1294 for (i = 0; i < 1024; i++) { 1061 for (i = 0; i < 1024; i++) { 1295 xa_store_index(xa, index, GFP 1062 xa_store_index(xa, index, GFP_KERNEL); 1296 j = 0; 1063 j = 0; 1297 xa_for_each(xa, index, entry) 1064 xa_for_each(xa, index, entry) { 1298 XA_BUG_ON(xa, xa_mk_i 1065 XA_BUG_ON(xa, xa_mk_index(index) != entry); 1299 XA_BUG_ON(xa, index ! 1066 XA_BUG_ON(xa, index != j++); 1300 } 1067 } 1301 } 1068 } 1302 1069 1303 xa_destroy(xa); 1070 xa_destroy(xa); 1304 } 1071 } 1305 1072 1306 static noinline void check_find_3(struct xarr 1073 static noinline void check_find_3(struct xarray *xa) 1307 { 1074 { 1308 XA_STATE(xas, xa, 0); 1075 XA_STATE(xas, xa, 0); 1309 unsigned long i, j, k; 1076 unsigned long i, j, k; 1310 void *entry; 1077 void *entry; 1311 1078 1312 for (i = 0; i < 100; i++) { 1079 for (i = 0; i < 100; i++) { 1313 for (j = 0; j < 100; j++) { 1080 for (j = 0; j < 100; j++) { 1314 rcu_read_lock(); 1081 rcu_read_lock(); 1315 for (k = 0; k < 100; 1082 for (k = 0; k < 100; k++) { 1316 xas_set(&xas, 1083 xas_set(&xas, j); 1317 xas_for_each_ 1084 xas_for_each_marked(&xas, entry, k, XA_MARK_0) 1318 ; 1085 ; 1319 if (j > k) 1086 if (j > k) 1320 XA_BU 1087 XA_BUG_ON(xa, 1321 1088 xas.xa_node != XAS_RESTART); 1322 } 1089 } 1323 rcu_read_unlock(); 1090 rcu_read_unlock(); 1324 } 1091 } 1325 xa_store_index(xa, i, GFP_KER 1092 xa_store_index(xa, i, GFP_KERNEL); 1326 xa_set_mark(xa, i, XA_MARK_0) 1093 xa_set_mark(xa, i, XA_MARK_0); 1327 } 1094 } 1328 xa_destroy(xa); 1095 xa_destroy(xa); 1329 } 1096 } 1330 1097 1331 static noinline void check_find_4(struct xarr 1098 static noinline void check_find_4(struct xarray *xa) 1332 { 1099 { 1333 unsigned long index = 0; 1100 unsigned long index = 0; 1334 void *entry; 1101 void *entry; 1335 1102 1336 xa_store_index(xa, ULONG_MAX, GFP_KER 1103 xa_store_index(xa, ULONG_MAX, GFP_KERNEL); 1337 1104 1338 entry = xa_find_after(xa, &index, ULO 1105 entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT); 1339 XA_BUG_ON(xa, entry != xa_mk_index(UL 1106 XA_BUG_ON(xa, entry != xa_mk_index(ULONG_MAX)); 1340 1107 1341 entry = xa_find_after(xa, &index, ULO 1108 entry = xa_find_after(xa, &index, ULONG_MAX, XA_PRESENT); 1342 XA_BUG_ON(xa, entry); 1109 XA_BUG_ON(xa, entry); 1343 1110 1344 xa_erase_index(xa, ULONG_MAX); 1111 xa_erase_index(xa, ULONG_MAX); 1345 } 1112 } 1346 1113 1347 static noinline void check_find(struct xarray 1114 static noinline void check_find(struct xarray *xa) 1348 { 1115 { 1349 unsigned i; 1116 unsigned i; 1350 1117 1351 check_find_1(xa); 1118 check_find_1(xa); 1352 check_find_2(xa); 1119 check_find_2(xa); 1353 check_find_3(xa); 1120 check_find_3(xa); 1354 check_find_4(xa); 1121 check_find_4(xa); 1355 1122 1356 for (i = 2; i < 10; i++) 1123 for (i = 2; i < 10; i++) 1357 check_multi_find_1(xa, i); 1124 check_multi_find_1(xa, i); 1358 check_multi_find_2(xa); 1125 check_multi_find_2(xa); 1359 check_multi_find_3(xa); 1126 check_multi_find_3(xa); 1360 } 1127 } 1361 1128 1362 /* See find_swap_entry() in mm/shmem.c */ 1129 /* See find_swap_entry() in mm/shmem.c */ 1363 static noinline unsigned long xa_find_entry(s 1130 static noinline unsigned long xa_find_entry(struct xarray *xa, void *item) 1364 { 1131 { 1365 XA_STATE(xas, xa, 0); 1132 XA_STATE(xas, xa, 0); 1366 unsigned int checked = 0; 1133 unsigned int checked = 0; 1367 void *entry; 1134 void *entry; 1368 1135 1369 rcu_read_lock(); 1136 rcu_read_lock(); 1370 xas_for_each(&xas, entry, ULONG_MAX) 1137 xas_for_each(&xas, entry, ULONG_MAX) { 1371 if (xas_retry(&xas, entry)) 1138 if (xas_retry(&xas, entry)) 1372 continue; 1139 continue; 1373 if (entry == item) 1140 if (entry == item) 1374 break; 1141 break; 1375 checked++; 1142 checked++; 1376 if ((checked % 4) != 0) 1143 if ((checked % 4) != 0) 1377 continue; 1144 continue; 1378 xas_pause(&xas); 1145 xas_pause(&xas); 1379 } 1146 } 1380 rcu_read_unlock(); 1147 rcu_read_unlock(); 1381 1148 1382 return entry ? xas.xa_index : -1; 1149 return entry ? xas.xa_index : -1; 1383 } 1150 } 1384 1151 1385 static noinline void check_find_entry(struct 1152 static noinline void check_find_entry(struct xarray *xa) 1386 { 1153 { 1387 #ifdef CONFIG_XARRAY_MULTI 1154 #ifdef CONFIG_XARRAY_MULTI 1388 unsigned int order; 1155 unsigned int order; 1389 unsigned long offset, index; 1156 unsigned long offset, index; 1390 1157 1391 for (order = 0; order < 20; order++) 1158 for (order = 0; order < 20; order++) { 1392 for (offset = 0; offset < (1U 1159 for (offset = 0; offset < (1UL << (order + 3)); 1393 offset += (1UL << order) 1160 offset += (1UL << order)) { 1394 for (index = 0; index 1161 for (index = 0; index < (1UL << (order + 5)); 1395 index += (1UL << 1162 index += (1UL << order)) { 1396 xa_store_orde 1163 xa_store_order(xa, index, order, 1397 1164 xa_mk_index(index), GFP_KERNEL); 1398 XA_BUG_ON(xa, 1165 XA_BUG_ON(xa, xa_load(xa, index) != 1399 1166 xa_mk_index(index)); 1400 XA_BUG_ON(xa, 1167 XA_BUG_ON(xa, xa_find_entry(xa, 1401 1168 xa_mk_index(index)) != index); 1402 } 1169 } 1403 XA_BUG_ON(xa, xa_find 1170 XA_BUG_ON(xa, xa_find_entry(xa, xa) != -1); 1404 xa_destroy(xa); 1171 xa_destroy(xa); 1405 } 1172 } 1406 } 1173 } 1407 #endif 1174 #endif 1408 1175 1409 XA_BUG_ON(xa, xa_find_entry(xa, xa) ! 1176 XA_BUG_ON(xa, xa_find_entry(xa, xa) != -1); 1410 xa_store_index(xa, ULONG_MAX, GFP_KER 1177 xa_store_index(xa, ULONG_MAX, GFP_KERNEL); 1411 XA_BUG_ON(xa, xa_find_entry(xa, xa) ! 1178 XA_BUG_ON(xa, xa_find_entry(xa, xa) != -1); 1412 XA_BUG_ON(xa, xa_find_entry(xa, xa_mk 1179 XA_BUG_ON(xa, xa_find_entry(xa, xa_mk_index(ULONG_MAX)) != -1); 1413 xa_erase_index(xa, ULONG_MAX); 1180 xa_erase_index(xa, ULONG_MAX); 1414 XA_BUG_ON(xa, !xa_empty(xa)); 1181 XA_BUG_ON(xa, !xa_empty(xa)); 1415 } 1182 } 1416 1183 1417 static noinline void check_pause(struct xarra 1184 static noinline void check_pause(struct xarray *xa) 1418 { 1185 { 1419 XA_STATE(xas, xa, 0); 1186 XA_STATE(xas, xa, 0); 1420 void *entry; 1187 void *entry; 1421 unsigned int order; 1188 unsigned int order; 1422 unsigned long index = 1; 1189 unsigned long index = 1; 1423 unsigned int count = 0; 1190 unsigned int count = 0; 1424 1191 1425 for (order = 0; order < order_limit; 1192 for (order = 0; order < order_limit; order++) { 1426 XA_BUG_ON(xa, xa_store_order( 1193 XA_BUG_ON(xa, xa_store_order(xa, index, order, 1427 xa_mk 1194 xa_mk_index(index), GFP_KERNEL)); 1428 index += 1UL << order; 1195 index += 1UL << order; 1429 } 1196 } 1430 1197 1431 rcu_read_lock(); 1198 rcu_read_lock(); 1432 xas_for_each(&xas, entry, ULONG_MAX) 1199 xas_for_each(&xas, entry, ULONG_MAX) { 1433 XA_BUG_ON(xa, entry != xa_mk_ 1200 XA_BUG_ON(xa, entry != xa_mk_index(1UL << count)); 1434 count++; 1201 count++; 1435 } 1202 } 1436 rcu_read_unlock(); 1203 rcu_read_unlock(); 1437 XA_BUG_ON(xa, count != order_limit); 1204 XA_BUG_ON(xa, count != order_limit); 1438 1205 1439 count = 0; 1206 count = 0; 1440 xas_set(&xas, 0); 1207 xas_set(&xas, 0); 1441 rcu_read_lock(); 1208 rcu_read_lock(); 1442 xas_for_each(&xas, entry, ULONG_MAX) 1209 xas_for_each(&xas, entry, ULONG_MAX) { 1443 XA_BUG_ON(xa, entry != xa_mk_ 1210 XA_BUG_ON(xa, entry != xa_mk_index(1UL << count)); 1444 count++; 1211 count++; 1445 xas_pause(&xas); 1212 xas_pause(&xas); 1446 } 1213 } 1447 rcu_read_unlock(); 1214 rcu_read_unlock(); 1448 XA_BUG_ON(xa, count != order_limit); 1215 XA_BUG_ON(xa, count != order_limit); 1449 1216 1450 xa_destroy(xa); 1217 xa_destroy(xa); 1451 } 1218 } 1452 1219 1453 static noinline void check_move_tiny(struct x 1220 static noinline void check_move_tiny(struct xarray *xa) 1454 { 1221 { 1455 XA_STATE(xas, xa, 0); 1222 XA_STATE(xas, xa, 0); 1456 1223 1457 XA_BUG_ON(xa, !xa_empty(xa)); 1224 XA_BUG_ON(xa, !xa_empty(xa)); 1458 rcu_read_lock(); 1225 rcu_read_lock(); 1459 XA_BUG_ON(xa, xas_next(&xas) != NULL) 1226 XA_BUG_ON(xa, xas_next(&xas) != NULL); 1460 XA_BUG_ON(xa, xas_next(&xas) != NULL) 1227 XA_BUG_ON(xa, xas_next(&xas) != NULL); 1461 rcu_read_unlock(); 1228 rcu_read_unlock(); 1462 xa_store_index(xa, 0, GFP_KERNEL); 1229 xa_store_index(xa, 0, GFP_KERNEL); 1463 rcu_read_lock(); 1230 rcu_read_lock(); 1464 xas_set(&xas, 0); 1231 xas_set(&xas, 0); 1465 XA_BUG_ON(xa, xas_next(&xas) != xa_mk 1232 XA_BUG_ON(xa, xas_next(&xas) != xa_mk_index(0)); 1466 XA_BUG_ON(xa, xas_next(&xas) != NULL) 1233 XA_BUG_ON(xa, xas_next(&xas) != NULL); 1467 xas_set(&xas, 0); 1234 xas_set(&xas, 0); 1468 XA_BUG_ON(xa, xas_prev(&xas) != xa_mk 1235 XA_BUG_ON(xa, xas_prev(&xas) != xa_mk_index(0)); 1469 XA_BUG_ON(xa, xas_prev(&xas) != NULL) 1236 XA_BUG_ON(xa, xas_prev(&xas) != NULL); 1470 rcu_read_unlock(); 1237 rcu_read_unlock(); 1471 xa_erase_index(xa, 0); 1238 xa_erase_index(xa, 0); 1472 XA_BUG_ON(xa, !xa_empty(xa)); 1239 XA_BUG_ON(xa, !xa_empty(xa)); 1473 } 1240 } 1474 1241 1475 static noinline void check_move_max(struct xa 1242 static noinline void check_move_max(struct xarray *xa) 1476 { 1243 { 1477 XA_STATE(xas, xa, 0); 1244 XA_STATE(xas, xa, 0); 1478 1245 1479 xa_store_index(xa, ULONG_MAX, GFP_KER 1246 xa_store_index(xa, ULONG_MAX, GFP_KERNEL); 1480 rcu_read_lock(); 1247 rcu_read_lock(); 1481 XA_BUG_ON(xa, xas_find(&xas, ULONG_MA 1248 XA_BUG_ON(xa, xas_find(&xas, ULONG_MAX) != xa_mk_index(ULONG_MAX)); 1482 XA_BUG_ON(xa, xas_find(&xas, ULONG_MA 1249 XA_BUG_ON(xa, xas_find(&xas, ULONG_MAX) != NULL); 1483 rcu_read_unlock(); 1250 rcu_read_unlock(); 1484 1251 1485 xas_set(&xas, 0); 1252 xas_set(&xas, 0); 1486 rcu_read_lock(); 1253 rcu_read_lock(); 1487 XA_BUG_ON(xa, xas_find(&xas, ULONG_MA 1254 XA_BUG_ON(xa, xas_find(&xas, ULONG_MAX) != xa_mk_index(ULONG_MAX)); 1488 xas_pause(&xas); 1255 xas_pause(&xas); 1489 XA_BUG_ON(xa, xas_find(&xas, ULONG_MA 1256 XA_BUG_ON(xa, xas_find(&xas, ULONG_MAX) != NULL); 1490 rcu_read_unlock(); 1257 rcu_read_unlock(); 1491 1258 1492 xa_erase_index(xa, ULONG_MAX); 1259 xa_erase_index(xa, ULONG_MAX); 1493 XA_BUG_ON(xa, !xa_empty(xa)); 1260 XA_BUG_ON(xa, !xa_empty(xa)); 1494 } 1261 } 1495 1262 1496 static noinline void check_move_small(struct 1263 static noinline void check_move_small(struct xarray *xa, unsigned long idx) 1497 { 1264 { 1498 XA_STATE(xas, xa, 0); 1265 XA_STATE(xas, xa, 0); 1499 unsigned long i; 1266 unsigned long i; 1500 1267 1501 xa_store_index(xa, 0, GFP_KERNEL); 1268 xa_store_index(xa, 0, GFP_KERNEL); 1502 xa_store_index(xa, idx, GFP_KERNEL); 1269 xa_store_index(xa, idx, GFP_KERNEL); 1503 1270 1504 rcu_read_lock(); 1271 rcu_read_lock(); 1505 for (i = 0; i < idx * 4; i++) { 1272 for (i = 0; i < idx * 4; i++) { 1506 void *entry = xas_next(&xas); 1273 void *entry = xas_next(&xas); 1507 if (i <= idx) 1274 if (i <= idx) 1508 XA_BUG_ON(xa, xas.xa_ 1275 XA_BUG_ON(xa, xas.xa_node == XAS_RESTART); 1509 XA_BUG_ON(xa, xas.xa_index != 1276 XA_BUG_ON(xa, xas.xa_index != i); 1510 if (i == 0 || i == idx) 1277 if (i == 0 || i == idx) 1511 XA_BUG_ON(xa, entry ! 1278 XA_BUG_ON(xa, entry != xa_mk_index(i)); 1512 else 1279 else 1513 XA_BUG_ON(xa, entry ! 1280 XA_BUG_ON(xa, entry != NULL); 1514 } 1281 } 1515 xas_next(&xas); 1282 xas_next(&xas); 1516 XA_BUG_ON(xa, xas.xa_index != i); 1283 XA_BUG_ON(xa, xas.xa_index != i); 1517 1284 1518 do { 1285 do { 1519 void *entry = xas_prev(&xas); 1286 void *entry = xas_prev(&xas); 1520 i--; 1287 i--; 1521 if (i <= idx) 1288 if (i <= idx) 1522 XA_BUG_ON(xa, xas.xa_ 1289 XA_BUG_ON(xa, xas.xa_node == XAS_RESTART); 1523 XA_BUG_ON(xa, xas.xa_index != 1290 XA_BUG_ON(xa, xas.xa_index != i); 1524 if (i == 0 || i == idx) 1291 if (i == 0 || i == idx) 1525 XA_BUG_ON(xa, entry ! 1292 XA_BUG_ON(xa, entry != xa_mk_index(i)); 1526 else 1293 else 1527 XA_BUG_ON(xa, entry ! 1294 XA_BUG_ON(xa, entry != NULL); 1528 } while (i > 0); 1295 } while (i > 0); 1529 1296 1530 xas_set(&xas, ULONG_MAX); 1297 xas_set(&xas, ULONG_MAX); 1531 XA_BUG_ON(xa, xas_next(&xas) != NULL) 1298 XA_BUG_ON(xa, xas_next(&xas) != NULL); 1532 XA_BUG_ON(xa, xas.xa_index != ULONG_M 1299 XA_BUG_ON(xa, xas.xa_index != ULONG_MAX); 1533 XA_BUG_ON(xa, xas_next(&xas) != xa_mk 1300 XA_BUG_ON(xa, xas_next(&xas) != xa_mk_value(0)); 1534 XA_BUG_ON(xa, xas.xa_index != 0); 1301 XA_BUG_ON(xa, xas.xa_index != 0); 1535 XA_BUG_ON(xa, xas_prev(&xas) != NULL) 1302 XA_BUG_ON(xa, xas_prev(&xas) != NULL); 1536 XA_BUG_ON(xa, xas.xa_index != ULONG_M 1303 XA_BUG_ON(xa, xas.xa_index != ULONG_MAX); 1537 rcu_read_unlock(); 1304 rcu_read_unlock(); 1538 1305 1539 xa_erase_index(xa, 0); 1306 xa_erase_index(xa, 0); 1540 xa_erase_index(xa, idx); 1307 xa_erase_index(xa, idx); 1541 XA_BUG_ON(xa, !xa_empty(xa)); 1308 XA_BUG_ON(xa, !xa_empty(xa)); 1542 } 1309 } 1543 1310 1544 static noinline void check_move(struct xarray 1311 static noinline void check_move(struct xarray *xa) 1545 { 1312 { 1546 XA_STATE(xas, xa, (1 << 16) - 1); 1313 XA_STATE(xas, xa, (1 << 16) - 1); 1547 unsigned long i; 1314 unsigned long i; 1548 1315 1549 for (i = 0; i < (1 << 16); i++) 1316 for (i = 0; i < (1 << 16); i++) 1550 XA_BUG_ON(xa, xa_store_index( 1317 XA_BUG_ON(xa, xa_store_index(xa, i, GFP_KERNEL) != NULL); 1551 1318 1552 rcu_read_lock(); 1319 rcu_read_lock(); 1553 do { 1320 do { 1554 void *entry = xas_prev(&xas); 1321 void *entry = xas_prev(&xas); 1555 i--; 1322 i--; 1556 XA_BUG_ON(xa, entry != xa_mk_ 1323 XA_BUG_ON(xa, entry != xa_mk_index(i)); 1557 XA_BUG_ON(xa, i != xas.xa_ind 1324 XA_BUG_ON(xa, i != xas.xa_index); 1558 } while (i != 0); 1325 } while (i != 0); 1559 1326 1560 XA_BUG_ON(xa, xas_prev(&xas) != NULL) 1327 XA_BUG_ON(xa, xas_prev(&xas) != NULL); 1561 XA_BUG_ON(xa, xas.xa_index != ULONG_M 1328 XA_BUG_ON(xa, xas.xa_index != ULONG_MAX); 1562 1329 1563 do { 1330 do { 1564 void *entry = xas_next(&xas); 1331 void *entry = xas_next(&xas); 1565 XA_BUG_ON(xa, entry != xa_mk_ 1332 XA_BUG_ON(xa, entry != xa_mk_index(i)); 1566 XA_BUG_ON(xa, i != xas.xa_ind 1333 XA_BUG_ON(xa, i != xas.xa_index); 1567 i++; 1334 i++; 1568 } while (i < (1 << 16)); 1335 } while (i < (1 << 16)); 1569 rcu_read_unlock(); 1336 rcu_read_unlock(); 1570 1337 1571 for (i = (1 << 8); i < (1 << 15); i++ 1338 for (i = (1 << 8); i < (1 << 15); i++) 1572 xa_erase_index(xa, i); 1339 xa_erase_index(xa, i); 1573 1340 1574 i = xas.xa_index; 1341 i = xas.xa_index; 1575 1342 1576 rcu_read_lock(); 1343 rcu_read_lock(); 1577 do { 1344 do { 1578 void *entry = xas_prev(&xas); 1345 void *entry = xas_prev(&xas); 1579 i--; 1346 i--; 1580 if ((i < (1 << 8)) || (i >= ( 1347 if ((i < (1 << 8)) || (i >= (1 << 15))) 1581 XA_BUG_ON(xa, entry ! 1348 XA_BUG_ON(xa, entry != xa_mk_index(i)); 1582 else 1349 else 1583 XA_BUG_ON(xa, entry ! 1350 XA_BUG_ON(xa, entry != NULL); 1584 XA_BUG_ON(xa, i != xas.xa_ind 1351 XA_BUG_ON(xa, i != xas.xa_index); 1585 } while (i != 0); 1352 } while (i != 0); 1586 1353 1587 XA_BUG_ON(xa, xas_prev(&xas) != NULL) 1354 XA_BUG_ON(xa, xas_prev(&xas) != NULL); 1588 XA_BUG_ON(xa, xas.xa_index != ULONG_M 1355 XA_BUG_ON(xa, xas.xa_index != ULONG_MAX); 1589 1356 1590 do { 1357 do { 1591 void *entry = xas_next(&xas); 1358 void *entry = xas_next(&xas); 1592 if ((i < (1 << 8)) || (i >= ( 1359 if ((i < (1 << 8)) || (i >= (1 << 15))) 1593 XA_BUG_ON(xa, entry ! 1360 XA_BUG_ON(xa, entry != xa_mk_index(i)); 1594 else 1361 else 1595 XA_BUG_ON(xa, entry ! 1362 XA_BUG_ON(xa, entry != NULL); 1596 XA_BUG_ON(xa, i != xas.xa_ind 1363 XA_BUG_ON(xa, i != xas.xa_index); 1597 i++; 1364 i++; 1598 } while (i < (1 << 16)); 1365 } while (i < (1 << 16)); 1599 rcu_read_unlock(); 1366 rcu_read_unlock(); 1600 1367 1601 xa_destroy(xa); 1368 xa_destroy(xa); 1602 1369 1603 check_move_tiny(xa); 1370 check_move_tiny(xa); 1604 check_move_max(xa); 1371 check_move_max(xa); 1605 1372 1606 for (i = 0; i < 16; i++) 1373 for (i = 0; i < 16; i++) 1607 check_move_small(xa, 1UL << i 1374 check_move_small(xa, 1UL << i); 1608 1375 1609 for (i = 2; i < 16; i++) 1376 for (i = 2; i < 16; i++) 1610 check_move_small(xa, (1UL << 1377 check_move_small(xa, (1UL << i) - 1); 1611 } 1378 } 1612 1379 1613 static noinline void xa_store_many_order(stru 1380 static noinline void xa_store_many_order(struct xarray *xa, 1614 unsigned long index, unsigned 1381 unsigned long index, unsigned order) 1615 { 1382 { 1616 XA_STATE_ORDER(xas, xa, index, order) 1383 XA_STATE_ORDER(xas, xa, index, order); 1617 unsigned int i = 0; 1384 unsigned int i = 0; 1618 1385 1619 do { 1386 do { 1620 xas_lock(&xas); 1387 xas_lock(&xas); 1621 XA_BUG_ON(xa, xas_find_confli 1388 XA_BUG_ON(xa, xas_find_conflict(&xas)); 1622 xas_create_range(&xas); 1389 xas_create_range(&xas); 1623 if (xas_error(&xas)) 1390 if (xas_error(&xas)) 1624 goto unlock; 1391 goto unlock; 1625 for (i = 0; i < (1U << order) 1392 for (i = 0; i < (1U << order); i++) { 1626 XA_BUG_ON(xa, xas_sto 1393 XA_BUG_ON(xa, xas_store(&xas, xa_mk_index(index + i))); 1627 xas_next(&xas); 1394 xas_next(&xas); 1628 } 1395 } 1629 unlock: 1396 unlock: 1630 xas_unlock(&xas); 1397 xas_unlock(&xas); 1631 } while (xas_nomem(&xas, GFP_KERNEL)) 1398 } while (xas_nomem(&xas, GFP_KERNEL)); 1632 1399 1633 XA_BUG_ON(xa, xas_error(&xas)); 1400 XA_BUG_ON(xa, xas_error(&xas)); 1634 } 1401 } 1635 1402 1636 static noinline void check_create_range_1(str 1403 static noinline void check_create_range_1(struct xarray *xa, 1637 unsigned long index, unsigned 1404 unsigned long index, unsigned order) 1638 { 1405 { 1639 unsigned long i; 1406 unsigned long i; 1640 1407 1641 xa_store_many_order(xa, index, order) 1408 xa_store_many_order(xa, index, order); 1642 for (i = index; i < index + (1UL << o 1409 for (i = index; i < index + (1UL << order); i++) 1643 xa_erase_index(xa, i); 1410 xa_erase_index(xa, i); 1644 XA_BUG_ON(xa, !xa_empty(xa)); 1411 XA_BUG_ON(xa, !xa_empty(xa)); 1645 } 1412 } 1646 1413 1647 static noinline void check_create_range_2(str 1414 static noinline void check_create_range_2(struct xarray *xa, unsigned order) 1648 { 1415 { 1649 unsigned long i; 1416 unsigned long i; 1650 unsigned long nr = 1UL << order; 1417 unsigned long nr = 1UL << order; 1651 1418 1652 for (i = 0; i < nr * nr; i += nr) 1419 for (i = 0; i < nr * nr; i += nr) 1653 xa_store_many_order(xa, i, or 1420 xa_store_many_order(xa, i, order); 1654 for (i = 0; i < nr * nr; i++) 1421 for (i = 0; i < nr * nr; i++) 1655 xa_erase_index(xa, i); 1422 xa_erase_index(xa, i); 1656 XA_BUG_ON(xa, !xa_empty(xa)); 1423 XA_BUG_ON(xa, !xa_empty(xa)); 1657 } 1424 } 1658 1425 1659 static noinline void check_create_range_3(voi 1426 static noinline void check_create_range_3(void) 1660 { 1427 { 1661 XA_STATE(xas, NULL, 0); 1428 XA_STATE(xas, NULL, 0); 1662 xas_set_err(&xas, -EEXIST); 1429 xas_set_err(&xas, -EEXIST); 1663 xas_create_range(&xas); 1430 xas_create_range(&xas); 1664 XA_BUG_ON(NULL, xas_error(&xas) != -E 1431 XA_BUG_ON(NULL, xas_error(&xas) != -EEXIST); 1665 } 1432 } 1666 1433 1667 static noinline void check_create_range_4(str 1434 static noinline void check_create_range_4(struct xarray *xa, 1668 unsigned long index, unsigned 1435 unsigned long index, unsigned order) 1669 { 1436 { 1670 XA_STATE_ORDER(xas, xa, index, order) 1437 XA_STATE_ORDER(xas, xa, index, order); 1671 unsigned long base = xas.xa_index; 1438 unsigned long base = xas.xa_index; 1672 unsigned long i = 0; 1439 unsigned long i = 0; 1673 1440 1674 xa_store_index(xa, index, GFP_KERNEL) 1441 xa_store_index(xa, index, GFP_KERNEL); 1675 do { 1442 do { 1676 xas_lock(&xas); 1443 xas_lock(&xas); 1677 xas_create_range(&xas); 1444 xas_create_range(&xas); 1678 if (xas_error(&xas)) 1445 if (xas_error(&xas)) 1679 goto unlock; 1446 goto unlock; 1680 for (i = 0; i < (1UL << order 1447 for (i = 0; i < (1UL << order); i++) { 1681 void *old = xas_store 1448 void *old = xas_store(&xas, xa_mk_index(base + i)); 1682 if (xas.xa_index == i 1449 if (xas.xa_index == index) 1683 XA_BUG_ON(xa, 1450 XA_BUG_ON(xa, old != xa_mk_index(base + i)); 1684 else 1451 else 1685 XA_BUG_ON(xa, 1452 XA_BUG_ON(xa, old != NULL); 1686 xas_next(&xas); 1453 xas_next(&xas); 1687 } 1454 } 1688 unlock: 1455 unlock: 1689 xas_unlock(&xas); 1456 xas_unlock(&xas); 1690 } while (xas_nomem(&xas, GFP_KERNEL)) 1457 } while (xas_nomem(&xas, GFP_KERNEL)); 1691 1458 1692 XA_BUG_ON(xa, xas_error(&xas)); 1459 XA_BUG_ON(xa, xas_error(&xas)); 1693 1460 1694 for (i = base; i < base + (1UL << ord 1461 for (i = base; i < base + (1UL << order); i++) 1695 xa_erase_index(xa, i); 1462 xa_erase_index(xa, i); 1696 XA_BUG_ON(xa, !xa_empty(xa)); 1463 XA_BUG_ON(xa, !xa_empty(xa)); 1697 } 1464 } 1698 1465 1699 static noinline void check_create_range_5(str 1466 static noinline void check_create_range_5(struct xarray *xa, 1700 unsigned long index, unsigned 1467 unsigned long index, unsigned int order) 1701 { 1468 { 1702 XA_STATE_ORDER(xas, xa, index, order) 1469 XA_STATE_ORDER(xas, xa, index, order); 1703 unsigned int i; 1470 unsigned int i; 1704 1471 1705 xa_store_order(xa, index, order, xa_m 1472 xa_store_order(xa, index, order, xa_mk_index(index), GFP_KERNEL); 1706 1473 1707 for (i = 0; i < order + 10; i++) { 1474 for (i = 0; i < order + 10; i++) { 1708 do { 1475 do { 1709 xas_lock(&xas); 1476 xas_lock(&xas); 1710 xas_create_range(&xas 1477 xas_create_range(&xas); 1711 xas_unlock(&xas); 1478 xas_unlock(&xas); 1712 } while (xas_nomem(&xas, GFP_ 1479 } while (xas_nomem(&xas, GFP_KERNEL)); 1713 } 1480 } 1714 1481 1715 xa_destroy(xa); 1482 xa_destroy(xa); 1716 } 1483 } 1717 1484 1718 static noinline void check_create_range(struc 1485 static noinline void check_create_range(struct xarray *xa) 1719 { 1486 { 1720 unsigned int order; 1487 unsigned int order; 1721 unsigned int max_order = IS_ENABLED(C 1488 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 12 : 1; 1722 1489 1723 for (order = 0; order < max_order; or 1490 for (order = 0; order < max_order; order++) { 1724 check_create_range_1(xa, 0, o 1491 check_create_range_1(xa, 0, order); 1725 check_create_range_1(xa, 1U < 1492 check_create_range_1(xa, 1U << order, order); 1726 check_create_range_1(xa, 2U < 1493 check_create_range_1(xa, 2U << order, order); 1727 check_create_range_1(xa, 3U < 1494 check_create_range_1(xa, 3U << order, order); 1728 check_create_range_1(xa, 1U < 1495 check_create_range_1(xa, 1U << 24, order); 1729 if (order < 10) 1496 if (order < 10) 1730 check_create_range_2( 1497 check_create_range_2(xa, order); 1731 1498 1732 check_create_range_4(xa, 0, o 1499 check_create_range_4(xa, 0, order); 1733 check_create_range_4(xa, 1U < 1500 check_create_range_4(xa, 1U << order, order); 1734 check_create_range_4(xa, 2U < 1501 check_create_range_4(xa, 2U << order, order); 1735 check_create_range_4(xa, 3U < 1502 check_create_range_4(xa, 3U << order, order); 1736 check_create_range_4(xa, 1U < 1503 check_create_range_4(xa, 1U << 24, order); 1737 1504 1738 check_create_range_4(xa, 1, o 1505 check_create_range_4(xa, 1, order); 1739 check_create_range_4(xa, (1U 1506 check_create_range_4(xa, (1U << order) + 1, order); 1740 check_create_range_4(xa, (2U 1507 check_create_range_4(xa, (2U << order) + 1, order); 1741 check_create_range_4(xa, (2U 1508 check_create_range_4(xa, (2U << order) - 1, order); 1742 check_create_range_4(xa, (3U 1509 check_create_range_4(xa, (3U << order) + 1, order); 1743 check_create_range_4(xa, (3U 1510 check_create_range_4(xa, (3U << order) - 1, order); 1744 check_create_range_4(xa, (1U 1511 check_create_range_4(xa, (1U << 24) + 1, order); 1745 1512 1746 check_create_range_5(xa, 0, o 1513 check_create_range_5(xa, 0, order); 1747 check_create_range_5(xa, (1U 1514 check_create_range_5(xa, (1U << order), order); 1748 } 1515 } 1749 1516 1750 check_create_range_3(); 1517 check_create_range_3(); 1751 } 1518 } 1752 1519 1753 static noinline void __check_store_range(stru 1520 static noinline void __check_store_range(struct xarray *xa, unsigned long first, 1754 unsigned long last) 1521 unsigned long last) 1755 { 1522 { 1756 #ifdef CONFIG_XARRAY_MULTI 1523 #ifdef CONFIG_XARRAY_MULTI 1757 xa_store_range(xa, first, last, xa_mk 1524 xa_store_range(xa, first, last, xa_mk_index(first), GFP_KERNEL); 1758 1525 1759 XA_BUG_ON(xa, xa_load(xa, first) != x 1526 XA_BUG_ON(xa, xa_load(xa, first) != xa_mk_index(first)); 1760 XA_BUG_ON(xa, xa_load(xa, last) != xa 1527 XA_BUG_ON(xa, xa_load(xa, last) != xa_mk_index(first)); 1761 XA_BUG_ON(xa, xa_load(xa, first - 1) 1528 XA_BUG_ON(xa, xa_load(xa, first - 1) != NULL); 1762 XA_BUG_ON(xa, xa_load(xa, last + 1) ! 1529 XA_BUG_ON(xa, xa_load(xa, last + 1) != NULL); 1763 1530 1764 xa_store_range(xa, first, last, NULL, 1531 xa_store_range(xa, first, last, NULL, GFP_KERNEL); 1765 #endif 1532 #endif 1766 1533 1767 XA_BUG_ON(xa, !xa_empty(xa)); 1534 XA_BUG_ON(xa, !xa_empty(xa)); 1768 } 1535 } 1769 1536 1770 static noinline void check_store_range(struct 1537 static noinline void check_store_range(struct xarray *xa) 1771 { 1538 { 1772 unsigned long i, j; 1539 unsigned long i, j; 1773 1540 1774 for (i = 0; i < 128; i++) { 1541 for (i = 0; i < 128; i++) { 1775 for (j = i; j < 128; j++) { 1542 for (j = i; j < 128; j++) { 1776 __check_store_range(x 1543 __check_store_range(xa, i, j); 1777 __check_store_range(x 1544 __check_store_range(xa, 128 + i, 128 + j); 1778 __check_store_range(x 1545 __check_store_range(xa, 4095 + i, 4095 + j); 1779 __check_store_range(x 1546 __check_store_range(xa, 4096 + i, 4096 + j); 1780 __check_store_range(x 1547 __check_store_range(xa, 123456 + i, 123456 + j); 1781 __check_store_range(x 1548 __check_store_range(xa, (1 << 24) + i, (1 << 24) + j); 1782 } 1549 } 1783 } 1550 } 1784 } 1551 } 1785 1552 1786 #ifdef CONFIG_XARRAY_MULTI 1553 #ifdef CONFIG_XARRAY_MULTI 1787 static void check_split_1(struct xarray *xa, 1554 static void check_split_1(struct xarray *xa, unsigned long index, 1788 unsigned int 1555 unsigned int order, unsigned int new_order) 1789 { 1556 { 1790 XA_STATE_ORDER(xas, xa, index, new_or 1557 XA_STATE_ORDER(xas, xa, index, new_order); 1791 unsigned int i, found; !! 1558 unsigned int i; 1792 void *entry; << 1793 1559 1794 xa_store_order(xa, index, order, xa, 1560 xa_store_order(xa, index, order, xa, GFP_KERNEL); 1795 xa_set_mark(xa, index, XA_MARK_1); << 1796 1561 1797 xas_split_alloc(&xas, xa, order, GFP_ 1562 xas_split_alloc(&xas, xa, order, GFP_KERNEL); 1798 xas_lock(&xas); 1563 xas_lock(&xas); 1799 xas_split(&xas, xa, order); 1564 xas_split(&xas, xa, order); 1800 for (i = 0; i < (1 << order); i += (1 1565 for (i = 0; i < (1 << order); i += (1 << new_order)) 1801 __xa_store(xa, index + i, xa_ 1566 __xa_store(xa, index + i, xa_mk_index(index + i), 0); 1802 xas_unlock(&xas); 1567 xas_unlock(&xas); 1803 1568 1804 for (i = 0; i < (1 << order); i++) { 1569 for (i = 0; i < (1 << order); i++) { 1805 unsigned int val = index + (i 1570 unsigned int val = index + (i & ~((1 << new_order) - 1)); 1806 XA_BUG_ON(xa, xa_load(xa, ind 1571 XA_BUG_ON(xa, xa_load(xa, index + i) != xa_mk_index(val)); 1807 } 1572 } 1808 1573 1809 xa_set_mark(xa, index, XA_MARK_0); 1574 xa_set_mark(xa, index, XA_MARK_0); 1810 XA_BUG_ON(xa, !xa_get_mark(xa, index, 1575 XA_BUG_ON(xa, !xa_get_mark(xa, index, XA_MARK_0)); 1811 1576 1812 xas_set_order(&xas, index, 0); << 1813 found = 0; << 1814 rcu_read_lock(); << 1815 xas_for_each_marked(&xas, entry, ULON << 1816 found++; << 1817 XA_BUG_ON(xa, xa_is_internal( << 1818 } << 1819 rcu_read_unlock(); << 1820 XA_BUG_ON(xa, found != 1 << (order - << 1821 << 1822 xa_destroy(xa); 1577 xa_destroy(xa); 1823 } 1578 } 1824 1579 1825 static noinline void check_split(struct xarra 1580 static noinline void check_split(struct xarray *xa) 1826 { 1581 { 1827 unsigned int order, new_order; 1582 unsigned int order, new_order; 1828 1583 1829 XA_BUG_ON(xa, !xa_empty(xa)); 1584 XA_BUG_ON(xa, !xa_empty(xa)); 1830 1585 1831 for (order = 1; order < 2 * XA_CHUNK_ 1586 for (order = 1; order < 2 * XA_CHUNK_SHIFT; order++) { 1832 for (new_order = 0; new_order 1587 for (new_order = 0; new_order < order; new_order++) { 1833 check_split_1(xa, 0, 1588 check_split_1(xa, 0, order, new_order); 1834 check_split_1(xa, 1UL 1589 check_split_1(xa, 1UL << order, order, new_order); 1835 check_split_1(xa, 3UL 1590 check_split_1(xa, 3UL << order, order, new_order); 1836 } 1591 } 1837 } 1592 } 1838 } 1593 } 1839 #else 1594 #else 1840 static void check_split(struct xarray *xa) { 1595 static void check_split(struct xarray *xa) { } 1841 #endif 1596 #endif 1842 1597 1843 static void check_align_1(struct xarray *xa, 1598 static void check_align_1(struct xarray *xa, char *name) 1844 { 1599 { 1845 int i; 1600 int i; 1846 unsigned int id; 1601 unsigned int id; 1847 unsigned long index; 1602 unsigned long index; 1848 void *entry; 1603 void *entry; 1849 1604 1850 for (i = 0; i < 8; i++) { 1605 for (i = 0; i < 8; i++) { 1851 XA_BUG_ON(xa, xa_alloc(xa, &i 1606 XA_BUG_ON(xa, xa_alloc(xa, &id, name + i, xa_limit_32b, 1852 GFP_K 1607 GFP_KERNEL) != 0); 1853 XA_BUG_ON(xa, id != i); 1608 XA_BUG_ON(xa, id != i); 1854 } 1609 } 1855 xa_for_each(xa, index, entry) 1610 xa_for_each(xa, index, entry) 1856 XA_BUG_ON(xa, xa_is_err(entry 1611 XA_BUG_ON(xa, xa_is_err(entry)); 1857 xa_destroy(xa); 1612 xa_destroy(xa); 1858 } 1613 } 1859 1614 1860 /* 1615 /* 1861 * We should always be able to store without 1616 * We should always be able to store without allocating memory after 1862 * reserving a slot. 1617 * reserving a slot. 1863 */ 1618 */ 1864 static void check_align_2(struct xarray *xa, 1619 static void check_align_2(struct xarray *xa, char *name) 1865 { 1620 { 1866 int i; 1621 int i; 1867 1622 1868 XA_BUG_ON(xa, !xa_empty(xa)); 1623 XA_BUG_ON(xa, !xa_empty(xa)); 1869 1624 1870 for (i = 0; i < 8; i++) { 1625 for (i = 0; i < 8; i++) { 1871 XA_BUG_ON(xa, xa_store(xa, 0, 1626 XA_BUG_ON(xa, xa_store(xa, 0, name + i, GFP_KERNEL) != NULL); 1872 xa_erase(xa, 0); 1627 xa_erase(xa, 0); 1873 } 1628 } 1874 1629 1875 for (i = 0; i < 8; i++) { 1630 for (i = 0; i < 8; i++) { 1876 XA_BUG_ON(xa, xa_reserve(xa, 1631 XA_BUG_ON(xa, xa_reserve(xa, 0, GFP_KERNEL) != 0); 1877 XA_BUG_ON(xa, xa_store(xa, 0, 1632 XA_BUG_ON(xa, xa_store(xa, 0, name + i, 0) != NULL); 1878 xa_erase(xa, 0); 1633 xa_erase(xa, 0); 1879 } 1634 } 1880 1635 1881 XA_BUG_ON(xa, !xa_empty(xa)); 1636 XA_BUG_ON(xa, !xa_empty(xa)); 1882 } 1637 } 1883 1638 1884 static noinline void check_align(struct xarra 1639 static noinline void check_align(struct xarray *xa) 1885 { 1640 { 1886 char name[] = "Motorola 68000"; 1641 char name[] = "Motorola 68000"; 1887 1642 1888 check_align_1(xa, name); 1643 check_align_1(xa, name); 1889 check_align_1(xa, name + 1); 1644 check_align_1(xa, name + 1); 1890 check_align_1(xa, name + 2); 1645 check_align_1(xa, name + 2); 1891 check_align_1(xa, name + 3); 1646 check_align_1(xa, name + 3); 1892 check_align_2(xa, name); 1647 check_align_2(xa, name); 1893 } 1648 } 1894 1649 1895 static LIST_HEAD(shadow_nodes); 1650 static LIST_HEAD(shadow_nodes); 1896 1651 1897 static void test_update_node(struct xa_node * 1652 static void test_update_node(struct xa_node *node) 1898 { 1653 { 1899 if (node->count && node->count == nod 1654 if (node->count && node->count == node->nr_values) { 1900 if (list_empty(&node->private 1655 if (list_empty(&node->private_list)) 1901 list_add(&shadow_node 1656 list_add(&shadow_nodes, &node->private_list); 1902 } else { 1657 } else { 1903 if (!list_empty(&node->privat 1658 if (!list_empty(&node->private_list)) 1904 list_del_init(&node-> 1659 list_del_init(&node->private_list); 1905 } 1660 } 1906 } 1661 } 1907 1662 1908 static noinline void shadow_remove(struct xar 1663 static noinline void shadow_remove(struct xarray *xa) 1909 { 1664 { 1910 struct xa_node *node; 1665 struct xa_node *node; 1911 1666 1912 xa_lock(xa); 1667 xa_lock(xa); 1913 while ((node = list_first_entry_or_nu 1668 while ((node = list_first_entry_or_null(&shadow_nodes, 1914 struc 1669 struct xa_node, private_list))) { 1915 XA_BUG_ON(xa, node->array != 1670 XA_BUG_ON(xa, node->array != xa); 1916 list_del_init(&node->private_ 1671 list_del_init(&node->private_list); 1917 xa_delete_node(node, test_upd 1672 xa_delete_node(node, test_update_node); 1918 } 1673 } 1919 xa_unlock(xa); 1674 xa_unlock(xa); 1920 } 1675 } 1921 1676 1922 static noinline void check_workingset(struct 1677 static noinline void check_workingset(struct xarray *xa, unsigned long index) 1923 { 1678 { 1924 XA_STATE(xas, xa, index); 1679 XA_STATE(xas, xa, index); 1925 xas_set_update(&xas, test_update_node 1680 xas_set_update(&xas, test_update_node); 1926 1681 1927 do { 1682 do { 1928 xas_lock(&xas); 1683 xas_lock(&xas); 1929 xas_store(&xas, xa_mk_value(0 1684 xas_store(&xas, xa_mk_value(0)); 1930 xas_next(&xas); 1685 xas_next(&xas); 1931 xas_store(&xas, xa_mk_value(1 1686 xas_store(&xas, xa_mk_value(1)); 1932 xas_unlock(&xas); 1687 xas_unlock(&xas); 1933 } while (xas_nomem(&xas, GFP_KERNEL)) 1688 } while (xas_nomem(&xas, GFP_KERNEL)); 1934 1689 1935 XA_BUG_ON(xa, list_empty(&shadow_node 1690 XA_BUG_ON(xa, list_empty(&shadow_nodes)); 1936 1691 1937 xas_lock(&xas); 1692 xas_lock(&xas); 1938 xas_next(&xas); 1693 xas_next(&xas); 1939 xas_store(&xas, &xas); 1694 xas_store(&xas, &xas); 1940 XA_BUG_ON(xa, !list_empty(&shadow_nod 1695 XA_BUG_ON(xa, !list_empty(&shadow_nodes)); 1941 1696 1942 xas_store(&xas, xa_mk_value(2)); 1697 xas_store(&xas, xa_mk_value(2)); 1943 xas_unlock(&xas); 1698 xas_unlock(&xas); 1944 XA_BUG_ON(xa, list_empty(&shadow_node 1699 XA_BUG_ON(xa, list_empty(&shadow_nodes)); 1945 1700 1946 shadow_remove(xa); 1701 shadow_remove(xa); 1947 XA_BUG_ON(xa, !list_empty(&shadow_nod 1702 XA_BUG_ON(xa, !list_empty(&shadow_nodes)); 1948 XA_BUG_ON(xa, !xa_empty(xa)); 1703 XA_BUG_ON(xa, !xa_empty(xa)); 1949 } 1704 } 1950 1705 1951 /* 1706 /* 1952 * Check that the pointer / value / sibling e 1707 * Check that the pointer / value / sibling entries are accounted the 1953 * way we expect them to be. 1708 * way we expect them to be. 1954 */ 1709 */ 1955 static noinline void check_account(struct xar 1710 static noinline void check_account(struct xarray *xa) 1956 { 1711 { 1957 #ifdef CONFIG_XARRAY_MULTI 1712 #ifdef CONFIG_XARRAY_MULTI 1958 unsigned int order; 1713 unsigned int order; 1959 1714 1960 for (order = 1; order < 12; order++) 1715 for (order = 1; order < 12; order++) { 1961 XA_STATE(xas, xa, 1 << order) 1716 XA_STATE(xas, xa, 1 << order); 1962 1717 1963 xa_store_order(xa, 0, order, 1718 xa_store_order(xa, 0, order, xa, GFP_KERNEL); 1964 rcu_read_lock(); 1719 rcu_read_lock(); 1965 xas_load(&xas); 1720 xas_load(&xas); 1966 XA_BUG_ON(xa, xas.xa_node->co 1721 XA_BUG_ON(xa, xas.xa_node->count == 0); 1967 XA_BUG_ON(xa, xas.xa_node->co 1722 XA_BUG_ON(xa, xas.xa_node->count > (1 << order)); 1968 XA_BUG_ON(xa, xas.xa_node->nr 1723 XA_BUG_ON(xa, xas.xa_node->nr_values != 0); 1969 rcu_read_unlock(); 1724 rcu_read_unlock(); 1970 1725 1971 xa_store_order(xa, 1 << order 1726 xa_store_order(xa, 1 << order, order, xa_mk_index(1UL << order), 1972 GFP_KERNEL); 1727 GFP_KERNEL); 1973 XA_BUG_ON(xa, xas.xa_node->co 1728 XA_BUG_ON(xa, xas.xa_node->count != xas.xa_node->nr_values * 2); 1974 1729 1975 xa_erase(xa, 1 << order); 1730 xa_erase(xa, 1 << order); 1976 XA_BUG_ON(xa, xas.xa_node->nr 1731 XA_BUG_ON(xa, xas.xa_node->nr_values != 0); 1977 1732 1978 xa_erase(xa, 0); 1733 xa_erase(xa, 0); 1979 XA_BUG_ON(xa, !xa_empty(xa)); 1734 XA_BUG_ON(xa, !xa_empty(xa)); 1980 } 1735 } 1981 #endif 1736 #endif 1982 } 1737 } 1983 1738 1984 static noinline void check_get_order(struct x 1739 static noinline void check_get_order(struct xarray *xa) 1985 { 1740 { 1986 unsigned int max_order = IS_ENABLED(C 1741 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 20 : 1; 1987 unsigned int order; 1742 unsigned int order; 1988 unsigned long i, j; 1743 unsigned long i, j; 1989 1744 1990 for (i = 0; i < 3; i++) 1745 for (i = 0; i < 3; i++) 1991 XA_BUG_ON(xa, xa_get_order(xa 1746 XA_BUG_ON(xa, xa_get_order(xa, i) != 0); 1992 1747 1993 for (order = 0; order < max_order; or 1748 for (order = 0; order < max_order; order++) { 1994 for (i = 0; i < 10; i++) { 1749 for (i = 0; i < 10; i++) { 1995 xa_store_order(xa, i 1750 xa_store_order(xa, i << order, order, 1996 xa_mk 1751 xa_mk_index(i << order), GFP_KERNEL); 1997 for (j = i << order; 1752 for (j = i << order; j < (i + 1) << order; j++) 1998 XA_BUG_ON(xa, 1753 XA_BUG_ON(xa, xa_get_order(xa, j) != order); 1999 xa_erase(xa, i << ord 1754 xa_erase(xa, i << order); 2000 } 1755 } 2001 } 1756 } 2002 } 1757 } 2003 1758 2004 static noinline void check_xas_get_order(stru 1759 static noinline void check_xas_get_order(struct xarray *xa) 2005 { 1760 { 2006 XA_STATE(xas, xa, 0); 1761 XA_STATE(xas, xa, 0); 2007 1762 2008 unsigned int max_order = IS_ENABLED(C 1763 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 20 : 1; 2009 unsigned int order; 1764 unsigned int order; 2010 unsigned long i, j; 1765 unsigned long i, j; 2011 1766 2012 for (order = 0; order < max_order; or 1767 for (order = 0; order < max_order; order++) { 2013 for (i = 0; i < 10; i++) { 1768 for (i = 0; i < 10; i++) { 2014 xas_set_order(&xas, i 1769 xas_set_order(&xas, i << order, order); 2015 do { 1770 do { 2016 xas_lock(&xas 1771 xas_lock(&xas); 2017 xas_store(&xa 1772 xas_store(&xas, xa_mk_value(i)); 2018 xas_unlock(&x 1773 xas_unlock(&xas); 2019 } while (xas_nomem(&x 1774 } while (xas_nomem(&xas, GFP_KERNEL)); 2020 1775 2021 for (j = i << order; 1776 for (j = i << order; j < (i + 1) << order; j++) { 2022 xas_set_order 1777 xas_set_order(&xas, j, 0); 2023 rcu_read_lock 1778 rcu_read_lock(); 2024 xas_load(&xas 1779 xas_load(&xas); 2025 XA_BUG_ON(xa, 1780 XA_BUG_ON(xa, xas_get_order(&xas) != order); 2026 rcu_read_unlo 1781 rcu_read_unlock(); 2027 } 1782 } 2028 1783 2029 xas_lock(&xas); 1784 xas_lock(&xas); 2030 xas_set_order(&xas, i 1785 xas_set_order(&xas, i << order, order); 2031 xas_store(&xas, NULL) 1786 xas_store(&xas, NULL); 2032 xas_unlock(&xas); 1787 xas_unlock(&xas); 2033 } 1788 } 2034 } 1789 } 2035 } 1790 } 2036 1791 2037 static noinline void check_xas_conflict_get_o 1792 static noinline void check_xas_conflict_get_order(struct xarray *xa) 2038 { 1793 { 2039 XA_STATE(xas, xa, 0); 1794 XA_STATE(xas, xa, 0); 2040 1795 2041 void *entry; 1796 void *entry; 2042 int only_once; 1797 int only_once; 2043 unsigned int max_order = IS_ENABLED(C 1798 unsigned int max_order = IS_ENABLED(CONFIG_XARRAY_MULTI) ? 20 : 1; 2044 unsigned int order; 1799 unsigned int order; 2045 unsigned long i, j, k; 1800 unsigned long i, j, k; 2046 1801 2047 for (order = 0; order < max_order; or 1802 for (order = 0; order < max_order; order++) { 2048 for (i = 0; i < 10; i++) { 1803 for (i = 0; i < 10; i++) { 2049 xas_set_order(&xas, i 1804 xas_set_order(&xas, i << order, order); 2050 do { 1805 do { 2051 xas_lock(&xas 1806 xas_lock(&xas); 2052 xas_store(&xa 1807 xas_store(&xas, xa_mk_value(i)); 2053 xas_unlock(&x 1808 xas_unlock(&xas); 2054 } while (xas_nomem(&x 1809 } while (xas_nomem(&xas, GFP_KERNEL)); 2055 1810 2056 /* 1811 /* 2057 * Ensure xas_get_ord 1812 * Ensure xas_get_order works with xas_for_each_conflict. 2058 */ 1813 */ 2059 j = i << order; 1814 j = i << order; 2060 for (k = 0; k < order 1815 for (k = 0; k < order; k++) { 2061 only_once = 0 1816 only_once = 0; 2062 xas_set_order 1817 xas_set_order(&xas, j + (1 << k), k); 2063 xas_lock(&xas 1818 xas_lock(&xas); 2064 xas_for_each_ 1819 xas_for_each_conflict(&xas, entry) { 2065 XA_BU 1820 XA_BUG_ON(xa, entry != xa_mk_value(i)); 2066 XA_BU 1821 XA_BUG_ON(xa, xas_get_order(&xas) != order); 2067 only_ 1822 only_once++; 2068 } 1823 } 2069 XA_BUG_ON(xa, 1824 XA_BUG_ON(xa, only_once != 1); 2070 xas_unlock(&x 1825 xas_unlock(&xas); 2071 } 1826 } 2072 1827 2073 if (order < max_order 1828 if (order < max_order - 1) { 2074 only_once = 0 1829 only_once = 0; 2075 xas_set_order 1830 xas_set_order(&xas, (i & ~1UL) << order, order + 1); 2076 xas_lock(&xas 1831 xas_lock(&xas); 2077 xas_for_each_ 1832 xas_for_each_conflict(&xas, entry) { 2078 XA_BU 1833 XA_BUG_ON(xa, entry != xa_mk_value(i)); 2079 XA_BU 1834 XA_BUG_ON(xa, xas_get_order(&xas) != order); 2080 only_ 1835 only_once++; 2081 } 1836 } 2082 XA_BUG_ON(xa, 1837 XA_BUG_ON(xa, only_once != 1); 2083 xas_unlock(&x 1838 xas_unlock(&xas); 2084 } 1839 } 2085 1840 2086 xas_set_order(&xas, i 1841 xas_set_order(&xas, i << order, order); 2087 xas_lock(&xas); 1842 xas_lock(&xas); 2088 xas_store(&xas, NULL) 1843 xas_store(&xas, NULL); 2089 xas_unlock(&xas); 1844 xas_unlock(&xas); 2090 } 1845 } 2091 } 1846 } 2092 } 1847 } 2093 1848 2094 1849 2095 static noinline void check_destroy(struct xar 1850 static noinline void check_destroy(struct xarray *xa) 2096 { 1851 { 2097 unsigned long index; 1852 unsigned long index; 2098 1853 2099 XA_BUG_ON(xa, !xa_empty(xa)); 1854 XA_BUG_ON(xa, !xa_empty(xa)); 2100 1855 2101 /* Destroying an empty array is a no- 1856 /* Destroying an empty array is a no-op */ 2102 xa_destroy(xa); 1857 xa_destroy(xa); 2103 XA_BUG_ON(xa, !xa_empty(xa)); 1858 XA_BUG_ON(xa, !xa_empty(xa)); 2104 1859 2105 /* Destroying an array with a single 1860 /* Destroying an array with a single entry */ 2106 for (index = 0; index < 1000; index++ 1861 for (index = 0; index < 1000; index++) { 2107 xa_store_index(xa, index, GFP 1862 xa_store_index(xa, index, GFP_KERNEL); 2108 XA_BUG_ON(xa, xa_empty(xa)); 1863 XA_BUG_ON(xa, xa_empty(xa)); 2109 xa_destroy(xa); 1864 xa_destroy(xa); 2110 XA_BUG_ON(xa, !xa_empty(xa)); 1865 XA_BUG_ON(xa, !xa_empty(xa)); 2111 } 1866 } 2112 1867 2113 /* Destroying an array with a single 1868 /* Destroying an array with a single entry at ULONG_MAX */ 2114 xa_store(xa, ULONG_MAX, xa, GFP_KERNE 1869 xa_store(xa, ULONG_MAX, xa, GFP_KERNEL); 2115 XA_BUG_ON(xa, xa_empty(xa)); 1870 XA_BUG_ON(xa, xa_empty(xa)); 2116 xa_destroy(xa); 1871 xa_destroy(xa); 2117 XA_BUG_ON(xa, !xa_empty(xa)); 1872 XA_BUG_ON(xa, !xa_empty(xa)); 2118 1873 2119 #ifdef CONFIG_XARRAY_MULTI 1874 #ifdef CONFIG_XARRAY_MULTI 2120 /* Destroying an array with a multi-i 1875 /* Destroying an array with a multi-index entry */ 2121 xa_store_order(xa, 1 << 11, 11, xa, G 1876 xa_store_order(xa, 1 << 11, 11, xa, GFP_KERNEL); 2122 XA_BUG_ON(xa, xa_empty(xa)); 1877 XA_BUG_ON(xa, xa_empty(xa)); 2123 xa_destroy(xa); 1878 xa_destroy(xa); 2124 XA_BUG_ON(xa, !xa_empty(xa)); 1879 XA_BUG_ON(xa, !xa_empty(xa)); 2125 #endif 1880 #endif 2126 } 1881 } 2127 1882 2128 static DEFINE_XARRAY(array); 1883 static DEFINE_XARRAY(array); 2129 1884 2130 static int xarray_checks(void) 1885 static int xarray_checks(void) 2131 { 1886 { 2132 check_xa_err(&array); 1887 check_xa_err(&array); 2133 check_xas_retry(&array); 1888 check_xas_retry(&array); 2134 check_xa_load(&array); 1889 check_xa_load(&array); 2135 check_xa_mark(&array); 1890 check_xa_mark(&array); 2136 check_xa_shrink(&array); 1891 check_xa_shrink(&array); 2137 check_xas_erase(&array); 1892 check_xas_erase(&array); 2138 check_insert(&array); 1893 check_insert(&array); 2139 check_cmpxchg(&array); 1894 check_cmpxchg(&array); 2140 check_cmpxchg_order(&array); << 2141 check_reserve(&array); 1895 check_reserve(&array); 2142 check_reserve(&xa0); 1896 check_reserve(&xa0); 2143 check_multi_store(&array); 1897 check_multi_store(&array); 2144 check_multi_store_advanced(&array); << 2145 check_get_order(&array); 1898 check_get_order(&array); 2146 check_xas_get_order(&array); 1899 check_xas_get_order(&array); 2147 check_xas_conflict_get_order(&array); 1900 check_xas_conflict_get_order(&array); 2148 check_xa_alloc(); 1901 check_xa_alloc(); 2149 check_find(&array); 1902 check_find(&array); 2150 check_find_entry(&array); 1903 check_find_entry(&array); 2151 check_pause(&array); 1904 check_pause(&array); 2152 check_account(&array); 1905 check_account(&array); 2153 check_destroy(&array); 1906 check_destroy(&array); 2154 check_move(&array); 1907 check_move(&array); 2155 check_create_range(&array); 1908 check_create_range(&array); 2156 check_store_range(&array); 1909 check_store_range(&array); 2157 check_store_iter(&array); 1910 check_store_iter(&array); 2158 check_align(&xa0); 1911 check_align(&xa0); 2159 check_split(&array); 1912 check_split(&array); 2160 1913 2161 check_workingset(&array, 0); 1914 check_workingset(&array, 0); 2162 check_workingset(&array, 64); 1915 check_workingset(&array, 64); 2163 check_workingset(&array, 4096); 1916 check_workingset(&array, 4096); 2164 1917 2165 printk("XArray: %u of %u tests passed 1918 printk("XArray: %u of %u tests passed\n", tests_passed, tests_run); 2166 return (tests_run == tests_passed) ? 1919 return (tests_run == tests_passed) ? 0 : -EINVAL; 2167 } 1920 } 2168 1921 2169 static void xarray_exit(void) 1922 static void xarray_exit(void) 2170 { 1923 { 2171 } 1924 } 2172 1925 2173 module_init(xarray_checks); 1926 module_init(xarray_checks); 2174 module_exit(xarray_exit); 1927 module_exit(xarray_exit); 2175 MODULE_AUTHOR("Matthew Wilcox <willy@infradea 1928 MODULE_AUTHOR("Matthew Wilcox <willy@infradead.org>"); 2176 MODULE_DESCRIPTION("XArray API test module"); << 2177 MODULE_LICENSE("GPL"); 1929 MODULE_LICENSE("GPL"); 2178 1930
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.