/linux-6.12.1/fs/xfs/libxfs/ |
D | xfs_btree.c | 137 if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops)) in __xfs_btree_check_lblock_hdr() 142 cur->bc_ops->get_maxrecs(cur, level)) in __xfs_btree_check_lblock_hdr() 239 if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops)) in __xfs_btree_check_agblock() 244 cur->bc_ops->get_maxrecs(cur, level)) in __xfs_btree_check_agblock() 268 switch (cur->bc_ops->type) { in __xfs_btree_check_block() 283 if (cur->bc_ops->ptr_len == XFS_BTREE_SHORT_PTR_LEN) in xfs_btree_block_errtag() 322 switch (cur->bc_ops->type) { in __xfs_btree_check_ptr() 358 switch (cur->bc_ops->type) { in xfs_btree_check_ptr() 362 cur->bc_ops->name, cur->bc_flags, level, index, in xfs_btree_check_ptr() 369 cur->bc_ino.whichfork, cur->bc_ops->name, in xfs_btree_check_ptr() [all …]
|
D | xfs_btree_staging.c | 64 ASSERT(cur->bc_ops->type != XFS_BTREE_TYPE_INODE); in xfs_btree_stage_afakeroot() 131 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_INODE); in xfs_btree_stage_ifakeroot() 310 xfs_btree_init_block(cur->bc_mp, ifp->if_broot, cur->bc_ops, in xfs_btree_bload_prep_block() 442 if (level == cur->bc_nlevels - 1 && cur->bc_ops->get_dmaxrecs) in xfs_btree_bload_max_npb() 443 return cur->bc_ops->get_dmaxrecs(cur, level); in xfs_btree_bload_max_npb() 445 ret = cur->bc_ops->get_maxrecs(cur, level); in xfs_btree_bload_max_npb() 470 return max_t(unsigned int, cur->bc_ops->get_minrecs(cur, level), npb); in xfs_btree_bload_desired_npb() 498 if (cur->bc_ops->get_dmaxrecs) in xfs_btree_bload_level_geometry() 499 maxnr = cur->bc_ops->get_dmaxrecs(cur, level); in xfs_btree_bload_level_geometry() 501 maxnr = cur->bc_ops->get_maxrecs(cur, level); in xfs_btree_bload_level_geometry() [all …]
|
D | xfs_btree.h | 83 (cur)->bc_ops->statoff + __XBTS_ ## stat) 86 (cur)->bc_ops->statoff + __XBTS_ ## stat, val) 251 const struct xfs_btree_ops *bc_ops; member 534 return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) < 0; in xfs_btree_keycmp_lt() 543 return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) > 0; in xfs_btree_keycmp_gt() 552 return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) == 0; in xfs_btree_keycmp_eq() 590 return cur->bc_ops->diff_two_keys(cur, key1, key2, mask) < 0; in xfs_btree_masked_keycmp_lt() 600 return cur->bc_ops->diff_two_keys(cur, key1, key2, mask) > 0; in xfs_btree_masked_keycmp_gt() 624 if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN) in xfs_btree_islastblock() 667 cur->bc_ops = ops; in xfs_btree_alloc_cursor() [all …]
|
D | xfs_btree_mem.c | 29 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); in xfbtree_set_root() 41 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); in xfbtree_init_ptr_from_cur() 53 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); in xfbtree_dup_cursor() 55 ncur = xfs_btree_alloc_cursor(cur->bc_mp, cur->bc_tp, cur->bc_ops, in xfbtree_dup_cursor() 172 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); in xfbtree_alloc_block() 198 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_MEM); in xfbtree_free_block()
|
D | xfs_ag.c | 401 xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 0, id->agno); in xfs_btroot_init() 467 xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 0, id->agno); in xfs_bnoroot_init() 483 xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 4, id->agno); in xfs_rmaproot_init() 695 const struct xfs_btree_ops *bc_ops; member 749 .bc_ops = &xfs_bnobt_ops, in xfs_ag_init_headers() 757 .bc_ops = &xfs_cntbt_ops, in xfs_ag_init_headers() 765 .bc_ops = &xfs_inobt_ops, in xfs_ag_init_headers() 773 .bc_ops = &xfs_finobt_ops, in xfs_ag_init_headers() 781 .bc_ops = &xfs_rmapbt_ops, in xfs_ag_init_headers() 789 .bc_ops = &xfs_refcountbt_ops, in xfs_ag_init_headers() [all …]
|
D | xfs_alloc_btree.c | 54 if (xfs_btree_is_bno(cur->bc_ops)) { in xfs_allocbt_set_root() 183 if (xfs_btree_is_bno(cur->bc_ops)) in xfs_allocbt_init_ptr_from_cur() 546 if (xfs_btree_is_bno(cur->bc_ops)) { in xfs_allocbt_commit_staged_btree()
|
D | xfs_ag.h | 325 const struct xfs_btree_ops *bc_ops; /* btree ops */ member
|
D | xfs_ialloc_btree.c | 93 if (xfs_btree_is_fino(cur->bc_ops)) in xfs_inobt_mod_blockcount() 538 if (xfs_btree_is_ino(cur->bc_ops)) { in xfs_inobt_commit_staged_btree()
|
D | xfs_alloc.c | 278 cur->bc_ops->name, cur->bc_ag.pag->pag_agno, fa); in xfs_alloc_complain_bad_rec() 1022 bool isbnobt = xfs_btree_is_bno(cur->bc_ops); in xfs_alloc_cur_check() 4067 ASSERT(xfs_btree_is_bno(cur->bc_ops)); in xfs_alloc_query_range() 4081 ASSERT(xfs_btree_is_bno(cur->bc_ops)); in xfs_alloc_query_all()
|
D | xfs_ialloc.c | 145 cur->bc_ops->name, cur->bc_ag.pag->pag_agno, fa); in xfs_inobt_complain_bad_rec() 2910 ASSERT(xfs_btree_is_ino(cur->bc_ops)); in xfs_ialloc_count_inodes()
|
D | xfs_rmap.c | 272 if (xfs_btree_is_mem_rmap(cur->bc_ops)) in xfs_rmap_check_btrec() 285 if (xfs_btree_is_mem_rmap(cur->bc_ops)) in xfs_rmap_complain_bad_rec()
|
/linux-6.12.1/fs/xfs/scrub/ |
D | btree.c | 50 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE) in __xchk_btree_process_error() 94 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE) in __xchk_btree_set_corrupt() 156 !cur->bc_ops->recs_inorder(cur, &bs->lastrec, rec)) in xchk_btree_rec() 158 memcpy(&bs->lastrec, rec, cur->bc_ops->rec_len); in xchk_btree_rec() 165 cur->bc_ops->init_key_from_rec(&key, rec); in xchk_btree_rec() 171 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING)) in xchk_btree_rec() 175 cur->bc_ops->init_high_key_from_rec(&hkey, rec); in xchk_btree_rec() 204 !cur->bc_ops->keys_inorder(cur, &bs->lastkey[level - 1].key, key)) in xchk_btree_key() 206 memcpy(&bs->lastkey[level - 1].key, key, cur->bc_ops->key_len); in xchk_btree_key() 218 if (!(cur->bc_ops->geom_flags & XFS_BTGEO_OVERLAPPING)) in xchk_btree_key() [all …]
|
D | trace.c | 47 cur->bc_ops->type == XFS_BTREE_TYPE_INODE) in xchk_btree_cur_fsbno()
|
D | ialloc.c | 79 ASSERT(xfs_btree_is_fino(cur->bc_ops)); in xchk_inobt_xref_finobt() 182 ASSERT(xfs_btree_is_ino(cur->bc_ops)); in xchk_finobt_xref_inobt() 517 if (xfs_btree_is_fino(bs->cur->bc_ops)) { in xchk_iallocbt_rec_alignment()
|
D | health.c | 258 unsigned int mask = (*curp)->bc_ops->sick_mask; in xchk_ag_btree_del_cursor_if_sick()
|
D | ialloc_repair.c | 545 cur->bc_ops->init_rec_from_cur(cur, block_rec); in xrep_fibt_get_records() 572 cur->bc_ops->init_rec_from_cur(cur, block_rec); in xrep_ibt_get_records()
|
D | newbt.c | 547 if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN) in xrep_newbt_claim_block()
|
D | refcount_repair.c | 538 cur->bc_ops->init_rec_from_cur(cur, block_rec); in xrep_refc_get_records()
|
D | bmap_repair.c | 455 cur->bc_ops->init_rec_from_cur(cur, block_rec); in xrep_bmap_get_records()
|
D | alloc_repair.c | 625 cur->bc_ops->init_rec_from_cur(cur, block_rec); in xrep_abt_get_records()
|
D | trace.h | 551 __string(name, cur->bc_ops->name) 593 __string(name, cur->bc_ops->name) 636 __string(name, cur->bc_ops->name) 674 __string(name, cur->bc_ops->name) 714 __string(name, cur->bc_ops->name)
|
D | rmap_repair.c | 1181 cur->bc_ops->init_rec_from_cur(cur, block_rec); in xrep_rmap_get_records()
|
/linux-6.12.1/fs/xfs/ |
D | xfs_health.c | 530 switch (cur->bc_ops->type) { in xfs_btree_mark_sick() 535 ASSERT(cur->bc_ops->sick_mask); in xfs_btree_mark_sick() 536 xfs_ag_mark_sick(cur->bc_ag.pag, cur->bc_ops->sick_mask); in xfs_btree_mark_sick() 539 if (xfs_btree_is_bmap(cur->bc_ops)) { in xfs_btree_mark_sick()
|
D | xfs_trace.h | 1906 __string(name, cur->bc_ops->name) 2495 __string(name, cur->bc_ops->name) 2533 __string(name, cur->bc_ops->name) 2539 switch (cur->bc_ops->type) { 2556 if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN) { 2586 __string(name, cur->bc_ops->name) 2593 if (cur->bc_ops->type == XFS_BTREE_TYPE_INODE) 2840 switch (cur->bc_ops->type) { 2894 switch (cur->bc_ops->type) { 4325 __string(name, cur->bc_ops->name) [all …]
|