Searched refs:mas_next (Results 1 – 6 of 6) sorted by relevance
/linux-6.12.1/lib/ |
D | test_maple_tree.c | 1336 entry = mas_next(&mas, limit); in check_next_entry() 1449 ptr = mas_next(&mas, ULONG_MAX); in check_root_expand() 1467 ptr = mas_next(&mas, ULONG_MAX); in check_root_expand() 1548 mas_next(&mas, ULONG_MAX); in check_gap_combining() 1549 entry = mas_next(&mas, ULONG_MAX); in check_gap_combining() 1579 entry = mas_next(&mas, ULONG_MAX); in check_gap_combining() 1581 mas_next(&mas, ULONG_MAX); /* go to the next entry. */ in check_gap_combining() 2077 mas_next(&mas, 1000); in next_prev_test() 2107 val = mas_next(&mas, 1000); in next_prev_test() 2117 val = mas_next(&mas, 1000); in next_prev_test() [all …]
|
D | maple_tree.c | 5743 void *mas_next(struct ma_state *mas, unsigned long max) in mas_next() function 5753 EXPORT_SYMBOL_GPL(mas_next); 5796 entry = mas_next(&mas, max); in mt_next()
|
/linux-6.12.1/Documentation/core-api/ |
D | maple_tree.rst | 173 Using a maple state allows mas_next() and mas_prev() to function as if the 175 performance penalty is outweighed by cache optimization. mas_next() will
|
/linux-6.12.1/include/linux/ |
D | maple_tree.h | 532 void *mas_next(struct ma_state *mas, unsigned long max);
|
/linux-6.12.1/mm/ |
D | vma.c | 1295 vma_test = mas_next(&test, vms->vma_count - 1); in vms_gather_munmap_vmas()
|
/linux-6.12.1/tools/testing/radix-tree/ |
D | maple.c | 765 entry = mas_next(&tmp, mas_end->last); in mas_ce2_over_count() 774 entry = mas_next(&tmp, mas_end->last); in mas_ce2_over_count() 35215 MT_BUG_ON(mt, mas_next(&mas_reader, ULONG_MAX) != xa_mk_value(val)); in check_rcu_simulated() 35230 mas_next(&mas_reader, ULONG_MAX); in check_rcu_simulated()
|