/linux-6.12.1/fs/xfs/ |
D | xfs_buf_mem.c | 138 loff_t pos = BBTOB(xfs_buf_daddr(bp)); in xmbuf_map_page() 145 if (BBTOB(bp->b_length) != XMBUF_BLOCKSIZE) in xmbuf_map_page() 218 pos = BBTOB(xfs_buf_daddr(bp)); in xmbuf_stale() 219 shmem_truncate_range(inode, pos, pos + BBTOB(bp->b_length) - 1); in xmbuf_stale()
|
D | xfs_buf_item.c | 242 offset += BBTOB(bp->b_maps[i].bm_len); in xfs_buf_item_size() 443 offset += BBTOB(bp->b_maps[i].bm_len); in xfs_buf_item_format() 804 bp->b_addr, BBTOB(bp->b_length), fa); in xfs_buf_item_precommit() 898 chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len), in xfs_buf_item_init() 907 BBTOB(bp->b_maps[i].bm_len)); in xfs_buf_item_init() 1029 end = start + BBTOB(bp->b_maps[i].bm_len) - 1; in xfs_buf_item_log() 1033 start += BBTOB(bp->b_maps[i].bm_len); in xfs_buf_item_log() 1050 start += BBTOB(bp->b_maps[i].bm_len); in xfs_buf_item_log()
|
D | xfs_log_priv.h | 378 min_t(int, (log)->l_logsize >> 3, BBTOB(XLOG_TOTAL_REC_SHIFT(log)) << 4) 607 return BBTOB(hi_block - lo_block); in xlog_lsn_sub() 609 return (uint64_t)log->l_logsize - BBTOB(lo_block - hi_block); in xlog_lsn_sub()
|
D | xfs_buf_item_recover.c | 475 ASSERT(BBTOB(bp->b_length) >= in xlog_recover_do_reg_buffer() 615 inodes_per_buf = BBTOB(bp->b_length) >> mp->m_sb.sb_inodelog; in xlog_recover_do_inode_buffer() 657 ASSERT((reg_buf_offset + reg_buf_bytes) <= BBTOB(bp->b_length)); in xlog_recover_do_inode_buffer() 1061 (BBTOB(bp->b_length) != M_IGEO(log->l_mp)->inode_cluster_size)) { in xlog_recover_buf_commit_pass2()
|
D | xfs_buf.h | 366 return xfs_verify_cksum(bp->b_addr, BBTOB(bp->b_length), in xfs_buf_verify_cksum() 373 xfs_update_cksum(bp->b_addr, BBTOB(bp->b_length), in xfs_buf_update_cksum()
|
D | xfs_buf.c | 338 size_t size = BBTOB(bp->b_length); in xfs_buf_alloc_kmem() 375 bp->b_page_count = DIV_ROUND_UP(BBTOB(bp->b_length), PAGE_SIZE); in xfs_buf_alloc_pages() 541 ASSERT(!(BBTOB(map->bm_len) < btp->bt_meta_sectorsize)); in xfs_buf_map_verify() 542 ASSERT(!(BBTOB(map->bm_bn) & (xfs_off_t)btp->bt_meta_sectormask)); in xfs_buf_map_verify() 645 } else if (BBTOB(new_bp->b_length) >= PAGE_SIZE || in xfs_buf_find_insert() 1551 size = min_t(int, BBTOB(bp->b_maps[map].bm_len), *count); in xfs_buf_ioapply_map() 1671 size = BBTOB(bp->b_length); in _xfs_buf_ioapply() 1815 BBTOB(bp->b_length) - boff); in xfs_buf_zero()
|
D | xfs_fsmap.c | 37 dest->fmr_physical = BBTOB(src->fmr_physical); in xfs_fsmap_from_internal() 39 dest->fmr_offset = BBTOB(src->fmr_offset); in xfs_fsmap_from_internal() 40 dest->fmr_length = BBTOB(src->fmr_length); in xfs_fsmap_from_internal()
|
D | xfs_log_recover.c | 104 return kvzalloc(BBTOB(nbblks), GFP_KERNEL | __GFP_RETRY_MAYFAIL); in xlog_alloc_buffer() 116 return BBTOB(blk_no & ((xfs_daddr_t)log->l_sectBBsize - 1)); in xlog_align() 141 BBTOB(nbblks), data, op); in xlog_do_io() 1562 buffer + BBTOB(ealign - start_block)); in xlog_write_log_records() 3133 offset + BBTOB(split_hblks)); in xlog_do_recovery_pass() 3194 offset + BBTOB(split_bblks)); in xlog_do_recovery_pass()
|
D | xfs_mount.c | 310 if ((BBTOB(mp->m_dalign) & mp->m_blockmask) || in xfs_validate_new_dalign() 311 (BBTOB(mp->m_swidth) & mp->m_blockmask)) { in xfs_validate_new_dalign()
|
D | xfs_dquot.c | 241 memset(d, 0, BBTOB(q->qi_dqchunklen)); in xfs_qm_init_dquot_blk() 277 xfs_trans_log_buf(tp, bp, 0, BBTOB(q->qi_dqchunklen) - 1); in xfs_qm_init_dquot_blk()
|
D | xfs_trans_buf.c | 541 ASSERT(first <= last && last < BBTOB(bp->b_length)); in xfs_trans_log_buf()
|
D | xfs_log.c | 1384 log->l_logsize = BBTOB(num_bblks); in xlog_alloc_log() 1738 unsigned int split_offset = BBTOB(log->l_logBBsize - bno); in xlog_split_iclog()
|
/linux-6.12.1/fs/xfs/libxfs/ |
D | xfs_attr_remote.c | 147 len = BBTOB(bp->b_length); in __xfs_attr3_rmt_read_verify() 212 len = BBTOB(bp->b_length); in xfs_attr3_rmt_write_verify() 302 unsigned int len = BBTOB(bp->b_length); in xfs_attr_rmtval_copyout() 351 unsigned int len = BBTOB(bp->b_length); in xfs_attr_rmtval_copyin()
|
D | xfs_dquot_buf.c | 25 return BBTOB(nbblks) / sizeof(struct xfs_dqblk); in xfs_calc_dquots_per_chunk()
|
D | xfs_sb.c | 1143 xfs_buf_zero(bp, 0, BBTOB(bp->b_length)); in xfs_update_secondary_sbs() 1425 BBTOB(mp->m_dalign), BBTOB(mp->m_swidth)); in xfs_validate_stripe_geometry()
|
D | xfs_symlink_remote.c | 191 ASSERT(BBTOB(bp->b_length) >= in xfs_symlink_local_to_remote()
|
D | xfs_fs.h | 1041 #define BBTOB(bbs) ((bbs) << BBSHIFT) macro
|
D | xfs_dir2_data.c | 306 bp->b_addr, BBTOB(bp->b_length), __FILE__, __LINE__, in xfs_dir3_data_check()
|
/linux-6.12.1/fs/xfs/scrub/ |
D | agheader_repair.c | 63 xfs_buf_zero(bp, 0, BBTOB(bp->b_length)); in xrep_superblock() 80 xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1); in xrep_superblock() 208 memset(agf, 0, BBTOB(agf_bp->b_length)); in xrep_agf_init_header() 325 xfs_trans_log_buf(sc->tp, agf_bp, 0, BBTOB(agf_bp->b_length) - 1); in xrep_agf_commit_new() 688 memset(agfl, 0xFF, BBTOB(agfl_bp->b_length)); in xrep_agfl_init_header() 707 xfs_trans_log_buf(sc->tp, agfl_bp, 0, BBTOB(agfl_bp->b_length) - 1); in xrep_agfl_init_header() 897 memset(agi, 0, BBTOB(agi_bp->b_length)); in xrep_agi_init_header() 1701 xfs_trans_log_buf(sc->tp, agi_bp, 0, BBTOB(agi_bp->b_length) - 1); in xrep_agi_commit_new()
|
D | quota_repair.c | 386 xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1); in xrep_quota_block()
|
D | ialloc_repair.c | 162 if (offset >= BBTOB(cluster_bp->b_length)) in xrep_ibt_check_ifree()
|
D | ialloc.c | 442 if (imap.im_boffset >= BBTOB(cluster_bp->b_length)) { in xchk_iallocbt_check_cluster()
|
D | reap.c | 756 chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len), in xreap_buf_loggable()
|
D | agheader.c | 354 BBTOB(bp->b_length) - sizeof(struct xfs_dsb))) in xchk_superblock()
|
/linux-6.12.1/Documentation/filesystems/xfs/ |
D | xfs-self-describing-metadata.rst | 232 !xfs_verify_cksum(bp->b_addr, BBTOB(bp->b_length), 321 xfs_update_cksum(bp->b_addr, BBTOB(bp->b_length), XFS_FOO_CRC_OFF);
|