Home
last modified time | relevance | path

Searched refs:nblocks (Results 1 – 25 of 97) sorted by relevance

1234

/linux-6.12.1/init/
Ddo_mounts_rd.c66 int nblocks = -1; in identify_ramdisk_image() local
96 nblocks = 0; in identify_ramdisk_image()
106 nblocks = (ntohl(romfsb->size)+BLOCK_SIZE-1)>>BLOCK_SIZE_BITS; in identify_ramdisk_image()
114 nblocks = (cramfsb->size + BLOCK_SIZE - 1) >> BLOCK_SIZE_BITS; in identify_ramdisk_image()
123 nblocks = (le64_to_cpu(squashfsb->bytes_used) + BLOCK_SIZE - 1) in identify_ramdisk_image()
138 nblocks = (cramfsb->size + BLOCK_SIZE - 1) >> BLOCK_SIZE_BITS; in identify_ramdisk_image()
154 nblocks = minixsb->s_nzones << minixsb->s_log_zone_size; in identify_ramdisk_image()
164 nblocks = n; in identify_ramdisk_image()
174 return nblocks; in identify_ramdisk_image()
190 int nblocks, i; in rd_load_image() local
[all …]
/linux-6.12.1/fs/jfs/
Djfs_dmap.c65 int nblocks);
72 static int dbAllocAny(struct bmap * bmp, s64 nblocks, int l2nb, s64 * results);
74 int nblocks);
76 int nblocks,
79 int nblocks);
80 static int dbAllocDmapLev(struct bmap * bmp, struct dmap * dp, int nblocks,
83 static int dbAllocAG(struct bmap * bmp, int agno, s64 nblocks, int l2nb,
85 static int dbAllocCtl(struct bmap * bmp, s64 nblocks, int l2nb, s64 blkno,
87 static int dbExtend(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks);
92 int nblocks);
[all …]
Djfs_discard.c32 void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks) in jfs_issue_discard() argument
37 r = sb_issue_discard(sb, blkno, nblocks, GFP_NOFS, 0); in jfs_issue_discard()
41 (unsigned long long)nblocks, r); in jfs_issue_discard()
46 (unsigned long long)nblocks, r); in jfs_issue_discard()
Djfs_dmap.h148 __le32 nblocks; /* 4: num blks covered by this dmap */ member
284 extern int dbFree(struct inode *ipbmap, s64 blkno, s64 nblocks);
287 int free, s64 blkno, s64 nblocks, struct tblock * tblk);
291 extern int dbAlloc(struct inode *ipbmap, s64 hint, s64 nblocks, s64 * results);
294 s64 blkno, s64 nblocks, s64 addnblocks, s64 * results);
297 extern int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks);
298 extern int dbExtendFS(struct inode *ipbmap, s64 blkno, s64 nblocks);
Dresize.c62 s64 XAddress, XSize, nblocks, xoff, xaddr, t64; in jfs_extendfs() local
313 nblocks = min(t64 - mapSize, XSize); in jfs_extendfs()
322 if ((rc = dbExtendFS(ipbmap, XAddress, nblocks))) in jfs_extendfs()
332 XSize -= nblocks; in jfs_extendfs()
382 xlen = min(xlen, (int) nblocks) & ~(sbi->nbperpage - 1); in jfs_extendfs()
387 if ((rc = xtAppend(tid, ipbmap, 0, xoff, nblocks, &xlen, &xaddr, 0))) { in jfs_extendfs()
/linux-6.12.1/arch/arm64/crypto/
Dsm4-neon-glue.c22 unsigned int nblocks);
24 u8 *iv, unsigned int nblocks);
26 u8 *iv, unsigned int nblocks);
47 unsigned int nblocks; in sm4_ecb_do_crypt() local
49 nblocks = nbytes / SM4_BLOCK_SIZE; in sm4_ecb_do_crypt()
50 if (nblocks) { in sm4_ecb_do_crypt()
53 sm4_neon_crypt(rkey, dst, src, nblocks); in sm4_ecb_do_crypt()
125 unsigned int nblocks; in sm4_cbc_decrypt() local
127 nblocks = nbytes / SM4_BLOCK_SIZE; in sm4_cbc_decrypt()
128 if (nblocks) { in sm4_cbc_decrypt()
[all …]
Dpolyval-ce-glue.c45 const u8 *in, size_t nblocks, u8 *accumulator);
49 const u8 *in, size_t nblocks, u8 *accumulator) in internal_polyval_update() argument
53 pmull_polyval_update(keys, in, nblocks, accumulator); in internal_polyval_update()
57 nblocks, accumulator); in internal_polyval_update()
107 unsigned int nblocks; in polyval_arm64_update() local
127 nblocks = min(srclen, 4096U) / POLYVAL_BLOCK_SIZE; in polyval_arm64_update()
128 internal_polyval_update(tctx, src, nblocks, dctx->buffer); in polyval_arm64_update()
129 srclen -= nblocks * POLYVAL_BLOCK_SIZE; in polyval_arm64_update()
130 src += nblocks * POLYVAL_BLOCK_SIZE; in polyval_arm64_update()
Dsm4-ce-glue.c33 u8 *iv, unsigned int nblocks);
35 u8 *iv, unsigned int nblocks);
49 const u8 *src, unsigned int nblocks,
168 unsigned int nblocks; in sm4_cbc_crypt() local
170 nblocks = nbytes / SM4_BLOCK_SIZE; in sm4_cbc_crypt()
171 if (nblocks) { in sm4_cbc_crypt()
176 walk.iv, nblocks); in sm4_cbc_crypt()
179 walk.iv, nblocks); in sm4_cbc_crypt()
342 int nblocks = DIV_ROUND_UP(req->cryptlen, SM4_BLOCK_SIZE) - 2; in sm4_xts_crypt() local
351 nblocks * SM4_BLOCK_SIZE, req->iv); in sm4_xts_crypt()
[all …]
Dsm4-ce-gcm-glue.c24 const u8 *src, unsigned int nblocks);
101 unsigned int nblocks; in gcm_calculate_auth_mac() local
114 nblocks = n / GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac()
115 if (nblocks) { in gcm_calculate_auth_mac()
117 ptr, nblocks); in gcm_calculate_auth_mac()
118 ptr += nblocks * GHASH_BLOCK_SIZE; in gcm_calculate_auth_mac()
/linux-6.12.1/arch/x86/crypto/
Dsm4_aesni_avx_glue.c23 const u8 *src, int nblocks);
25 const u8 *src, int nblocks);
59 unsigned int nblocks = min(nbytes >> 4, 4u); in ecb_do_crypt() local
60 sm4_aesni_avx_crypt4(rkey, dst, src, nblocks); in ecb_do_crypt()
61 dst += nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt()
62 src += nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt()
63 nbytes -= nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt()
151 unsigned int nblocks = min(nbytes >> 4, 8u); in sm4_avx_cbc_decrypt() local
155 src, nblocks); in sm4_avx_cbc_decrypt()
157 src += ((int)nblocks - 2) * SM4_BLOCK_SIZE; in sm4_avx_cbc_decrypt()
[all …]
Dblake2s-glue.c19 const u8 *block, const size_t nblocks,
22 const u8 *block, const size_t nblocks,
29 size_t nblocks, const u32 inc) in blake2s_compress() argument
35 blake2s_compress_generic(state, block, nblocks, inc); in blake2s_compress()
40 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress()
51 nblocks -= blocks; in blake2s_compress()
53 } while (nblocks); in blake2s_compress()
Dpolyval-clmulni_glue.c49 const u8 *in, size_t nblocks, u8 *accumulator);
58 const u8 *in, size_t nblocks, u8 *accumulator) in internal_polyval_update() argument
62 clmul_polyval_update(keys, in, nblocks, accumulator); in internal_polyval_update()
66 nblocks, accumulator); in internal_polyval_update()
116 unsigned int nblocks; in polyval_x86_update() local
136 nblocks = min(srclen, 4096U) / POLYVAL_BLOCK_SIZE; in polyval_x86_update()
137 internal_polyval_update(tctx, src, nblocks, dctx->buffer); in polyval_x86_update()
138 srclen -= nblocks * POLYVAL_BLOCK_SIZE; in polyval_x86_update()
139 src += nblocks * POLYVAL_BLOCK_SIZE; in polyval_x86_update()
/linux-6.12.1/arch/arm/crypto/
Dblake2b-neon-glue.c19 const u8 *block, size_t nblocks, u32 inc);
22 const u8 *block, size_t nblocks, u32 inc) in blake2b_compress_arch() argument
25 blake2b_compress_generic(state, block, nblocks, inc); in blake2b_compress_arch()
30 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress_arch()
37 nblocks -= blocks; in blake2b_compress_arch()
39 } while (nblocks); in blake2b_compress_arch()
/linux-6.12.1/include/crypto/internal/
Dblake2b.h15 const u8 *block, size_t nblocks, u32 inc);
23 const u8 *block, size_t nblocks, u32 inc);
41 const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2B_BLOCK_SIZE); in __blake2b_update() local
43 (*compress)(state, in, nblocks - 1, BLAKE2B_BLOCK_SIZE); in __blake2b_update()
44 in += BLAKE2B_BLOCK_SIZE * (nblocks - 1); in __blake2b_update()
45 inlen -= BLAKE2B_BLOCK_SIZE * (nblocks - 1); in __blake2b_update()
Dblake2s.h14 size_t nblocks, const u32 inc);
17 size_t nblocks, const u32 inc);
/linux-6.12.1/lib/vdso/
Dgetrandom.c73 size_t batch_len, nblocks, orig_len = len; in __cvdso_getrandom_data() local
235 nblocks = len / CHACHA_BLOCK_SIZE; in __cvdso_getrandom_data()
236 if (nblocks) { in __cvdso_getrandom_data()
237 __arch_chacha20_blocks_nostack(buffer, state->key, counter, nblocks); in __cvdso_getrandom_data()
238 buffer += nblocks * CHACHA_BLOCK_SIZE; in __cvdso_getrandom_data()
239 len -= nblocks * CHACHA_BLOCK_SIZE; in __cvdso_getrandom_data()
/linux-6.12.1/lib/crypto/
Dblake2s-generic.c40 size_t nblocks, const u32 inc)
44 size_t nblocks, const u32 inc) in blake2s_compress_generic() argument
51 (nblocks > 1 && inc != BLAKE2S_BLOCK_SIZE)); in blake2s_compress_generic()
53 while (nblocks > 0) { in blake2s_compress_generic()
106 --nblocks; in blake2s_compress_generic()
Dblake2s.c38 const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2S_BLOCK_SIZE); in blake2s_update() local
39 blake2s_compress(state, in, nblocks - 1, BLAKE2S_BLOCK_SIZE); in blake2s_update()
40 in += BLAKE2S_BLOCK_SIZE * (nblocks - 1); in blake2s_update()
41 inlen -= BLAKE2S_BLOCK_SIZE * (nblocks - 1); in blake2s_update()
/linux-6.12.1/fs/nilfs2/
Dthe_nilfs.c427 u64 nsegments, nblocks; in nilfs_store_disk_layout() local
487 nblocks = sb_bdev_nr_blocks(nilfs->ns_sb); in nilfs_store_disk_layout()
488 if (nblocks) { in nilfs_store_disk_layout()
496 if (nblocks < min_block_count) { in nilfs_store_disk_layout()
500 (unsigned long long)nblocks); in nilfs_store_disk_layout()
782 sector_t start = 0, nblocks = 0; in nilfs_discard_segments() local
792 if (!nblocks) { in nilfs_discard_segments()
794 nblocks = seg_end - seg_start + 1; in nilfs_discard_segments()
795 } else if (start + nblocks == seg_start) { in nilfs_discard_segments()
796 nblocks += seg_end - seg_start + 1; in nilfs_discard_segments()
[all …]
Dsufile.c567 unsigned long nblocks, time64_t modtime) in nilfs_sufile_set_segment_usage() argument
589 su->su_nblocks = cpu_to_le32(nblocks); in nilfs_sufile_set_segment_usage()
1082 sector_t start = 0, nblocks = 0; in nilfs_sufile_trim_fs() local
1138 if (!nblocks) { in nilfs_sufile_trim_fs()
1141 nblocks = seg_end - seg_start + 1; in nilfs_sufile_trim_fs()
1145 if (start + nblocks == seg_start) { in nilfs_sufile_trim_fs()
1147 nblocks += seg_end - seg_start + 1; in nilfs_sufile_trim_fs()
1153 nblocks -= start_block - start; in nilfs_sufile_trim_fs()
1157 if (nblocks >= minlen) { in nilfs_sufile_trim_fs()
1162 nblocks * sects_per_block, in nilfs_sufile_trim_fs()
[all …]
Dsegbuf.h34 unsigned long nblocks; member
131 return segbuf->sb_sum.nblocks == segbuf->sb_sum.nsumblk; in nilfs_segbuf_empty()
139 segbuf->sb_sum.nblocks++; in nilfs_segbuf_add_segsum_buffer()
148 segbuf->sb_sum.nblocks++; in nilfs_segbuf_add_payload_buffer()
/linux-6.12.1/fs/reiserfs/
Dxattr.h75 size_t nblocks = JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() local
78 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks()
80 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks()
83 return nblocks; in reiserfs_xattr_jcreate_nblocks()
/linux-6.12.1/fs/ext4/
Dext4_jbd2.h305 #define ext4_journal_start_sb(sb, type, nblocks) \ argument
306 __ext4_journal_start_sb(NULL, (sb), __LINE__, (type), (nblocks), 0,\
309 #define ext4_journal_start(inode, type, nblocks) \ argument
310 __ext4_journal_start((inode), __LINE__, (type), (nblocks), 0, \
344 static inline int ext4_journal_extend(handle_t *handle, int nblocks, int revoke) in ext4_journal_extend() argument
347 return jbd2_journal_extend(handle, nblocks, revoke); in ext4_journal_extend()
351 static inline int ext4_journal_restart(handle_t *handle, int nblocks, in ext4_journal_restart() argument
355 return jbd2__journal_restart(handle, nblocks, revoke, GFP_NOFS); in ext4_journal_restart()
/linux-6.12.1/block/partitions/
Dsysv68.c44 __be32 nblocks; /* slice size (in blocks) */ member
83 if (be32_to_cpu(slice->nblocks)) { in sysv68_partition()
86 be32_to_cpu(slice->nblocks)); in sysv68_partition()
/linux-6.12.1/fs/
Dmpage.c169 unsigned nblocks; in do_mpage_readpage() local
194 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage()
197 block_in_file < (args->first_logical_block + nblocks)) { in do_mpage_readpage()
199 unsigned last = nblocks - map_offset; in do_mpage_readpage()
258 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage()
260 if (relative_block == nblocks) { in do_mpage_readpage()
304 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage()
305 if ((buffer_boundary(map_bh) && relative_block == nblocks) || in do_mpage_readpage()

1234