/linux-6.12.1/init/ |
D | do_mounts_rd.c | 66 int nblocks = -1; in identify_ramdisk_image() local 96 nblocks = 0; in identify_ramdisk_image() 106 nblocks = (ntohl(romfsb->size)+BLOCK_SIZE-1)>>BLOCK_SIZE_BITS; in identify_ramdisk_image() 114 nblocks = (cramfsb->size + BLOCK_SIZE - 1) >> BLOCK_SIZE_BITS; in identify_ramdisk_image() 123 nblocks = (le64_to_cpu(squashfsb->bytes_used) + BLOCK_SIZE - 1) in identify_ramdisk_image() 138 nblocks = (cramfsb->size + BLOCK_SIZE - 1) >> BLOCK_SIZE_BITS; in identify_ramdisk_image() 154 nblocks = minixsb->s_nzones << minixsb->s_log_zone_size; in identify_ramdisk_image() 164 nblocks = n; in identify_ramdisk_image() 174 return nblocks; in identify_ramdisk_image() 190 int nblocks, i; in rd_load_image() local [all …]
|
/linux-6.12.1/fs/jfs/ |
D | jfs_dmap.c | 65 int nblocks); 72 static int dbAllocAny(struct bmap * bmp, s64 nblocks, int l2nb, s64 * results); 74 int nblocks); 76 int nblocks, 79 int nblocks); 80 static int dbAllocDmapLev(struct bmap * bmp, struct dmap * dp, int nblocks, 83 static int dbAllocAG(struct bmap * bmp, int agno, s64 nblocks, int l2nb, 85 static int dbAllocCtl(struct bmap * bmp, s64 nblocks, int l2nb, s64 blkno, 87 static int dbExtend(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks); 92 int nblocks); [all …]
|
D | jfs_discard.c | 32 void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks) in jfs_issue_discard() argument 37 r = sb_issue_discard(sb, blkno, nblocks, GFP_NOFS, 0); in jfs_issue_discard() 41 (unsigned long long)nblocks, r); in jfs_issue_discard() 46 (unsigned long long)nblocks, r); in jfs_issue_discard()
|
D | jfs_dmap.h | 148 __le32 nblocks; /* 4: num blks covered by this dmap */ member 284 extern int dbFree(struct inode *ipbmap, s64 blkno, s64 nblocks); 287 int free, s64 blkno, s64 nblocks, struct tblock * tblk); 291 extern int dbAlloc(struct inode *ipbmap, s64 hint, s64 nblocks, s64 * results); 294 s64 blkno, s64 nblocks, s64 addnblocks, s64 * results); 297 extern int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks); 298 extern int dbExtendFS(struct inode *ipbmap, s64 blkno, s64 nblocks);
|
D | resize.c | 62 s64 XAddress, XSize, nblocks, xoff, xaddr, t64; in jfs_extendfs() local 313 nblocks = min(t64 - mapSize, XSize); in jfs_extendfs() 322 if ((rc = dbExtendFS(ipbmap, XAddress, nblocks))) in jfs_extendfs() 332 XSize -= nblocks; in jfs_extendfs() 382 xlen = min(xlen, (int) nblocks) & ~(sbi->nbperpage - 1); in jfs_extendfs() 387 if ((rc = xtAppend(tid, ipbmap, 0, xoff, nblocks, &xlen, &xaddr, 0))) { in jfs_extendfs()
|
/linux-6.12.1/arch/arm64/crypto/ |
D | sm4-neon-glue.c | 22 unsigned int nblocks); 24 u8 *iv, unsigned int nblocks); 26 u8 *iv, unsigned int nblocks); 47 unsigned int nblocks; in sm4_ecb_do_crypt() local 49 nblocks = nbytes / SM4_BLOCK_SIZE; in sm4_ecb_do_crypt() 50 if (nblocks) { in sm4_ecb_do_crypt() 53 sm4_neon_crypt(rkey, dst, src, nblocks); in sm4_ecb_do_crypt() 125 unsigned int nblocks; in sm4_cbc_decrypt() local 127 nblocks = nbytes / SM4_BLOCK_SIZE; in sm4_cbc_decrypt() 128 if (nblocks) { in sm4_cbc_decrypt() [all …]
|
D | polyval-ce-glue.c | 45 const u8 *in, size_t nblocks, u8 *accumulator); 49 const u8 *in, size_t nblocks, u8 *accumulator) in internal_polyval_update() argument 53 pmull_polyval_update(keys, in, nblocks, accumulator); in internal_polyval_update() 57 nblocks, accumulator); in internal_polyval_update() 107 unsigned int nblocks; in polyval_arm64_update() local 127 nblocks = min(srclen, 4096U) / POLYVAL_BLOCK_SIZE; in polyval_arm64_update() 128 internal_polyval_update(tctx, src, nblocks, dctx->buffer); in polyval_arm64_update() 129 srclen -= nblocks * POLYVAL_BLOCK_SIZE; in polyval_arm64_update() 130 src += nblocks * POLYVAL_BLOCK_SIZE; in polyval_arm64_update()
|
D | sm4-ce-glue.c | 33 u8 *iv, unsigned int nblocks); 35 u8 *iv, unsigned int nblocks); 49 const u8 *src, unsigned int nblocks, 168 unsigned int nblocks; in sm4_cbc_crypt() local 170 nblocks = nbytes / SM4_BLOCK_SIZE; in sm4_cbc_crypt() 171 if (nblocks) { in sm4_cbc_crypt() 176 walk.iv, nblocks); in sm4_cbc_crypt() 179 walk.iv, nblocks); in sm4_cbc_crypt() 342 int nblocks = DIV_ROUND_UP(req->cryptlen, SM4_BLOCK_SIZE) - 2; in sm4_xts_crypt() local 351 nblocks * SM4_BLOCK_SIZE, req->iv); in sm4_xts_crypt() [all …]
|
D | sm4-ce-gcm-glue.c | 24 const u8 *src, unsigned int nblocks); 101 unsigned int nblocks; in gcm_calculate_auth_mac() local 114 nblocks = n / GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac() 115 if (nblocks) { in gcm_calculate_auth_mac() 117 ptr, nblocks); in gcm_calculate_auth_mac() 118 ptr += nblocks * GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac()
|
/linux-6.12.1/arch/x86/crypto/ |
D | sm4_aesni_avx_glue.c | 23 const u8 *src, int nblocks); 25 const u8 *src, int nblocks); 59 unsigned int nblocks = min(nbytes >> 4, 4u); in ecb_do_crypt() local 60 sm4_aesni_avx_crypt4(rkey, dst, src, nblocks); in ecb_do_crypt() 61 dst += nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt() 62 src += nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt() 63 nbytes -= nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt() 151 unsigned int nblocks = min(nbytes >> 4, 8u); in sm4_avx_cbc_decrypt() local 155 src, nblocks); in sm4_avx_cbc_decrypt() 157 src += ((int)nblocks - 2) * SM4_BLOCK_SIZE; in sm4_avx_cbc_decrypt() [all …]
|
D | blake2s-glue.c | 19 const u8 *block, const size_t nblocks, 22 const u8 *block, const size_t nblocks, 29 size_t nblocks, const u32 inc) in blake2s_compress() argument 35 blake2s_compress_generic(state, block, nblocks, inc); in blake2s_compress() 40 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress() 51 nblocks -= blocks; in blake2s_compress() 53 } while (nblocks); in blake2s_compress()
|
D | polyval-clmulni_glue.c | 49 const u8 *in, size_t nblocks, u8 *accumulator); 58 const u8 *in, size_t nblocks, u8 *accumulator) in internal_polyval_update() argument 62 clmul_polyval_update(keys, in, nblocks, accumulator); in internal_polyval_update() 66 nblocks, accumulator); in internal_polyval_update() 116 unsigned int nblocks; in polyval_x86_update() local 136 nblocks = min(srclen, 4096U) / POLYVAL_BLOCK_SIZE; in polyval_x86_update() 137 internal_polyval_update(tctx, src, nblocks, dctx->buffer); in polyval_x86_update() 138 srclen -= nblocks * POLYVAL_BLOCK_SIZE; in polyval_x86_update() 139 src += nblocks * POLYVAL_BLOCK_SIZE; in polyval_x86_update()
|
/linux-6.12.1/arch/arm/crypto/ |
D | blake2b-neon-glue.c | 19 const u8 *block, size_t nblocks, u32 inc); 22 const u8 *block, size_t nblocks, u32 inc) in blake2b_compress_arch() argument 25 blake2b_compress_generic(state, block, nblocks, inc); in blake2b_compress_arch() 30 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress_arch() 37 nblocks -= blocks; in blake2b_compress_arch() 39 } while (nblocks); in blake2b_compress_arch()
|
/linux-6.12.1/include/crypto/internal/ |
D | blake2b.h | 15 const u8 *block, size_t nblocks, u32 inc); 23 const u8 *block, size_t nblocks, u32 inc); 41 const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2B_BLOCK_SIZE); in __blake2b_update() local 43 (*compress)(state, in, nblocks - 1, BLAKE2B_BLOCK_SIZE); in __blake2b_update() 44 in += BLAKE2B_BLOCK_SIZE * (nblocks - 1); in __blake2b_update() 45 inlen -= BLAKE2B_BLOCK_SIZE * (nblocks - 1); in __blake2b_update()
|
D | blake2s.h | 14 size_t nblocks, const u32 inc); 17 size_t nblocks, const u32 inc);
|
/linux-6.12.1/lib/vdso/ |
D | getrandom.c | 73 size_t batch_len, nblocks, orig_len = len; in __cvdso_getrandom_data() local 235 nblocks = len / CHACHA_BLOCK_SIZE; in __cvdso_getrandom_data() 236 if (nblocks) { in __cvdso_getrandom_data() 237 __arch_chacha20_blocks_nostack(buffer, state->key, counter, nblocks); in __cvdso_getrandom_data() 238 buffer += nblocks * CHACHA_BLOCK_SIZE; in __cvdso_getrandom_data() 239 len -= nblocks * CHACHA_BLOCK_SIZE; in __cvdso_getrandom_data()
|
/linux-6.12.1/lib/crypto/ |
D | blake2s-generic.c | 40 size_t nblocks, const u32 inc) 44 size_t nblocks, const u32 inc) in blake2s_compress_generic() argument 51 (nblocks > 1 && inc != BLAKE2S_BLOCK_SIZE)); in blake2s_compress_generic() 53 while (nblocks > 0) { in blake2s_compress_generic() 106 --nblocks; in blake2s_compress_generic()
|
D | blake2s.c | 38 const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2S_BLOCK_SIZE); in blake2s_update() local 39 blake2s_compress(state, in, nblocks - 1, BLAKE2S_BLOCK_SIZE); in blake2s_update() 40 in += BLAKE2S_BLOCK_SIZE * (nblocks - 1); in blake2s_update() 41 inlen -= BLAKE2S_BLOCK_SIZE * (nblocks - 1); in blake2s_update()
|
/linux-6.12.1/fs/nilfs2/ |
D | the_nilfs.c | 427 u64 nsegments, nblocks; in nilfs_store_disk_layout() local 487 nblocks = sb_bdev_nr_blocks(nilfs->ns_sb); in nilfs_store_disk_layout() 488 if (nblocks) { in nilfs_store_disk_layout() 496 if (nblocks < min_block_count) { in nilfs_store_disk_layout() 500 (unsigned long long)nblocks); in nilfs_store_disk_layout() 782 sector_t start = 0, nblocks = 0; in nilfs_discard_segments() local 792 if (!nblocks) { in nilfs_discard_segments() 794 nblocks = seg_end - seg_start + 1; in nilfs_discard_segments() 795 } else if (start + nblocks == seg_start) { in nilfs_discard_segments() 796 nblocks += seg_end - seg_start + 1; in nilfs_discard_segments() [all …]
|
D | sufile.c | 567 unsigned long nblocks, time64_t modtime) in nilfs_sufile_set_segment_usage() argument 589 su->su_nblocks = cpu_to_le32(nblocks); in nilfs_sufile_set_segment_usage() 1082 sector_t start = 0, nblocks = 0; in nilfs_sufile_trim_fs() local 1138 if (!nblocks) { in nilfs_sufile_trim_fs() 1141 nblocks = seg_end - seg_start + 1; in nilfs_sufile_trim_fs() 1145 if (start + nblocks == seg_start) { in nilfs_sufile_trim_fs() 1147 nblocks += seg_end - seg_start + 1; in nilfs_sufile_trim_fs() 1153 nblocks -= start_block - start; in nilfs_sufile_trim_fs() 1157 if (nblocks >= minlen) { in nilfs_sufile_trim_fs() 1162 nblocks * sects_per_block, in nilfs_sufile_trim_fs() [all …]
|
D | segbuf.h | 34 unsigned long nblocks; member 131 return segbuf->sb_sum.nblocks == segbuf->sb_sum.nsumblk; in nilfs_segbuf_empty() 139 segbuf->sb_sum.nblocks++; in nilfs_segbuf_add_segsum_buffer() 148 segbuf->sb_sum.nblocks++; in nilfs_segbuf_add_payload_buffer()
|
/linux-6.12.1/fs/reiserfs/ |
D | xattr.h | 75 size_t nblocks = JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() local 78 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() 80 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() 83 return nblocks; in reiserfs_xattr_jcreate_nblocks()
|
/linux-6.12.1/fs/ext4/ |
D | ext4_jbd2.h | 305 #define ext4_journal_start_sb(sb, type, nblocks) \ argument 306 __ext4_journal_start_sb(NULL, (sb), __LINE__, (type), (nblocks), 0,\ 309 #define ext4_journal_start(inode, type, nblocks) \ argument 310 __ext4_journal_start((inode), __LINE__, (type), (nblocks), 0, \ 344 static inline int ext4_journal_extend(handle_t *handle, int nblocks, int revoke) in ext4_journal_extend() argument 347 return jbd2_journal_extend(handle, nblocks, revoke); in ext4_journal_extend() 351 static inline int ext4_journal_restart(handle_t *handle, int nblocks, in ext4_journal_restart() argument 355 return jbd2__journal_restart(handle, nblocks, revoke, GFP_NOFS); in ext4_journal_restart()
|
/linux-6.12.1/block/partitions/ |
D | sysv68.c | 44 __be32 nblocks; /* slice size (in blocks) */ member 83 if (be32_to_cpu(slice->nblocks)) { in sysv68_partition() 86 be32_to_cpu(slice->nblocks)); in sysv68_partition()
|
/linux-6.12.1/fs/ |
D | mpage.c | 169 unsigned nblocks; in do_mpage_readpage() local 194 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage() 197 block_in_file < (args->first_logical_block + nblocks)) { in do_mpage_readpage() 199 unsigned last = nblocks - map_offset; in do_mpage_readpage() 258 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage() 260 if (relative_block == nblocks) { in do_mpage_readpage() 304 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage() 305 if ((buffer_boundary(map_bh) && relative_block == nblocks) || in do_mpage_readpage()
|