Home
last modified time | relevance | path

Searched refs:partial (Results 1 – 25 of 241) sorted by relevance

12345678910

/linux-6.12.1/fs/ext4/
Dindirect.c252 Indirect *partial) in ext4_find_goal() argument
260 goal = ext4_find_near(inode, partial); in ext4_find_goal()
538 Indirect *partial; in ext4_ind_map_blocks() local
554 partial = ext4_get_branch(inode, depth, offsets, chain, &err); in ext4_ind_map_blocks()
557 if (!partial) { in ext4_ind_map_blocks()
586 for (i = partial - chain + 1; i < depth; i++) in ext4_ind_map_blocks()
620 ar.goal = ext4_find_goal(inode, map->m_lblk, partial); in ext4_ind_map_blocks()
623 indirect_blks = (chain + depth) - partial - 1; in ext4_ind_map_blocks()
629 ar.len = ext4_blks_to_allocate(partial, indirect_blks, in ext4_ind_map_blocks()
636 offsets + (partial - chain), partial); in ext4_ind_map_blocks()
[all …]
/linux-6.12.1/fs/minix/
Ditree_common.c158 Indirect *partial; in get_block() local
166 partial = get_branch(inode, depth, offsets, chain, &err); in get_block()
169 if (!partial) { in get_block()
173 partial = chain+depth-1; /* the whole chain */ in get_block()
180 while (partial > chain) { in get_block()
181 brelse(partial->bh); in get_block()
182 partial--; in get_block()
196 left = (chain + depth) - partial; in get_block()
197 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block()
201 if (splice_branch(inode, chain, partial, left) < 0) in get_block()
[all …]
/linux-6.12.1/include/crypto/
Dsha1_base.h41 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_base_do_update() local
45 if (unlikely((partial + len) >= SHA1_BLOCK_SIZE)) { in sha1_base_do_update()
48 if (partial) { in sha1_base_do_update()
49 int p = SHA1_BLOCK_SIZE - partial; in sha1_base_do_update()
51 memcpy(sctx->buffer + partial, data, p); in sha1_base_do_update()
65 partial = 0; in sha1_base_do_update()
68 memcpy(sctx->buffer + partial, data, len); in sha1_base_do_update()
79 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_base_do_finalize() local
81 sctx->buffer[partial++] = 0x80; in sha1_base_do_finalize()
82 if (partial > bit_offset) { in sha1_base_do_finalize()
[all …]
Dsm3_base.h44 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_base_do_update() local
48 if (unlikely((partial + len) >= SM3_BLOCK_SIZE)) { in sm3_base_do_update()
51 if (partial) { in sm3_base_do_update()
52 int p = SM3_BLOCK_SIZE - partial; in sm3_base_do_update()
54 memcpy(sctx->buffer + partial, data, p); in sm3_base_do_update()
68 partial = 0; in sm3_base_do_update()
71 memcpy(sctx->buffer + partial, data, len); in sm3_base_do_update()
82 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_base_do_finalize() local
84 sctx->buffer[partial++] = 0x80; in sm3_base_do_finalize()
85 if (partial > bit_offset) { in sm3_base_do_finalize()
[all …]
Dsha256_base.h42 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in lib_sha256_base_do_update() local
46 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { in lib_sha256_base_do_update()
49 if (partial) { in lib_sha256_base_do_update()
50 int p = SHA256_BLOCK_SIZE - partial; in lib_sha256_base_do_update()
52 memcpy(sctx->buf + partial, data, p); in lib_sha256_base_do_update()
66 partial = 0; in lib_sha256_base_do_update()
69 memcpy(sctx->buf + partial, data, len); in lib_sha256_base_do_update()
89 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in lib_sha256_base_do_finalize() local
91 sctx->buf[partial++] = 0x80; in lib_sha256_base_do_finalize()
92 if (partial > bit_offset) { in lib_sha256_base_do_finalize()
[all …]
Dsha512_base.h62 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update() local
68 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update()
71 if (partial) { in sha512_base_do_update()
72 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update()
74 memcpy(sctx->buf + partial, data, p); in sha512_base_do_update()
88 partial = 0; in sha512_base_do_update()
91 memcpy(sctx->buf + partial, data, len); in sha512_base_do_update()
102 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize() local
104 sctx->buf[partial++] = 0x80; in sha512_base_do_finalize()
105 if (partial > bit_offset) { in sha512_base_do_finalize()
[all …]
/linux-6.12.1/fs/sysv/
Ditree.c214 Indirect *partial; in get_block() local
222 partial = get_branch(inode, depth, offsets, chain, &err); in get_block()
225 if (!partial) { in get_block()
230 partial = chain+depth-1; /* the whole chain */ in get_block()
237 while (partial > chain) { in get_block()
238 brelse(partial->bh); in get_block()
239 partial--; in get_block()
253 left = (chain + depth) - partial; in get_block()
254 err = alloc_branch(inode, left, offsets+(partial-chain), partial); in get_block()
258 if (splice_branch(inode, chain, partial, left) < 0) in get_block()
[all …]
/linux-6.12.1/drivers/crypto/
Dpadlock-sha.c284 unsigned int partial, done; in padlock_sha1_update_nano() local
291 partial = sctx->count & 0x3f; in padlock_sha1_update_nano()
297 if ((partial + len) >= SHA1_BLOCK_SIZE) { in padlock_sha1_update_nano()
300 if (partial) { in padlock_sha1_update_nano()
301 done = -partial; in padlock_sha1_update_nano()
302 memcpy(sctx->buffer + partial, data, in padlock_sha1_update_nano()
321 partial = 0; in padlock_sha1_update_nano()
324 memcpy(sctx->buffer + partial, src, len - done); in padlock_sha1_update_nano()
332 unsigned int partial, padlen; in padlock_sha1_final_nano() local
339 partial = state->count & 0x3f; in padlock_sha1_final_nano()
[all …]
/linux-6.12.1/arch/arm64/crypto/
Dsha3-ce-glue.c43 if ((sctx->partial + len) >= sctx->rsiz) { in sha3_update()
46 if (sctx->partial) { in sha3_update()
47 int p = sctx->rsiz - sctx->partial; in sha3_update()
49 memcpy(sctx->buf + sctx->partial, data, p); in sha3_update()
56 sctx->partial = 0; in sha3_update()
75 memcpy(sctx->buf + sctx->partial, data, len); in sha3_update()
76 sctx->partial += len; in sha3_update()
91 sctx->buf[sctx->partial++] = 0x06; in sha3_final()
92 memset(sctx->buf + sctx->partial, 0, sctx->rsiz - sctx->partial); in sha3_final()
/linux-6.12.1/arch/powerpc/crypto/
Dsha1.c30 unsigned int partial, done; in powerpc_sha1_update() local
33 partial = sctx->count & 0x3f; in powerpc_sha1_update()
38 if ((partial + len) > 63) { in powerpc_sha1_update()
40 if (partial) { in powerpc_sha1_update()
41 done = -partial; in powerpc_sha1_update()
42 memcpy(sctx->buffer + partial, data, done + 64); in powerpc_sha1_update()
52 partial = 0; in powerpc_sha1_update()
54 memcpy(sctx->buffer + partial, src, len - done); in powerpc_sha1_update()
Daes-gcm-p10.S637 # check partial block
888 std 15, 56(7) # clear partial?
927 # check partial block
929 ld 15, 56(7) # partial?
943 # create partial block mask
955 lxvb16x 16, 15, 10 # load partial block mask
965 std 12, 56(7) # update partial?
986 lxvb16x \_mask, 0, 10 # load partial block mask
990 # Handle multiple partial blocks for encrypt and decrypt
1029 # store partial block
[all …]
/linux-6.12.1/fs/ext2/
Dinode.c331 Indirect *partial) in ext2_find_goal() argument
346 return ext2_find_near(inode, partial); in ext2_find_goal()
632 Indirect *partial; in ext2_get_blocks() local
648 partial = ext2_get_branch(inode, depth, offsets, chain, &err); in ext2_get_blocks()
650 if (!partial) { in ext2_get_blocks()
666 partial = chain + depth - 1; in ext2_get_blocks()
696 if (err == -EAGAIN || !verify_chain(chain, partial)) { in ext2_get_blocks()
697 while (partial > chain) { in ext2_get_blocks()
698 brelse(partial->bh); in ext2_get_blocks()
699 partial--; in ext2_get_blocks()
[all …]
/linux-6.12.1/arch/sparc/crypto/
Dsha1_glue.c31 unsigned int len, unsigned int partial) in __sha1_sparc64_update() argument
36 if (partial) { in __sha1_sparc64_update()
37 done = SHA1_BLOCK_SIZE - partial; in __sha1_sparc64_update()
38 memcpy(sctx->buffer + partial, data, done); in __sha1_sparc64_update()
55 unsigned int partial = sctx->count % SHA1_BLOCK_SIZE; in sha1_sparc64_update() local
58 if (partial + len < SHA1_BLOCK_SIZE) { in sha1_sparc64_update()
60 memcpy(sctx->buffer + partial, data, len); in sha1_sparc64_update()
62 __sha1_sparc64_update(sctx, data, len, partial); in sha1_sparc64_update()
Dmd5_glue.c47 unsigned int len, unsigned int partial) in __md5_sparc64_update() argument
52 if (partial) { in __md5_sparc64_update()
53 done = MD5_HMAC_BLOCK_SIZE - partial; in __md5_sparc64_update()
54 memcpy((u8 *)sctx->block + partial, data, done); in __md5_sparc64_update()
71 unsigned int partial = sctx->byte_count % MD5_HMAC_BLOCK_SIZE; in md5_sparc64_update() local
74 if (partial + len < MD5_HMAC_BLOCK_SIZE) { in md5_sparc64_update()
76 memcpy((u8 *)sctx->block + partial, data, len); in md5_sparc64_update()
78 __md5_sparc64_update(sctx, data, len, partial); in md5_sparc64_update()
Dsha512_glue.c30 unsigned int len, unsigned int partial) in __sha512_sparc64_update() argument
36 if (partial) { in __sha512_sparc64_update()
37 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update()
38 memcpy(sctx->buf + partial, data, done); in __sha512_sparc64_update()
55 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() local
58 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update()
61 memcpy(sctx->buf + partial, data, len); in sha512_sparc64_update()
63 __sha512_sparc64_update(sctx, data, len, partial); in sha512_sparc64_update()
Dsha256_glue.c31 unsigned int len, unsigned int partial) in __sha256_sparc64_update() argument
36 if (partial) { in __sha256_sparc64_update()
37 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update()
38 memcpy(sctx->buf + partial, data, done); in __sha256_sparc64_update()
55 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update() local
58 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update()
60 memcpy(sctx->buf + partial, data, len); in sha256_sparc64_update()
62 __sha256_sparc64_update(sctx, data, len, partial); in sha256_sparc64_update()
/linux-6.12.1/arch/mips/cavium-octeon/crypto/
Docteon-sha1.c78 unsigned int partial; in __octeon_sha1_update() local
82 partial = sctx->count % SHA1_BLOCK_SIZE; in __octeon_sha1_update()
87 if ((partial + len) >= SHA1_BLOCK_SIZE) { in __octeon_sha1_update()
88 if (partial) { in __octeon_sha1_update()
89 done = -partial; in __octeon_sha1_update()
90 memcpy(sctx->buffer + partial, data, in __octeon_sha1_update()
101 partial = 0; in __octeon_sha1_update()
103 memcpy(sctx->buffer + partial, src, len - done); in __octeon_sha1_update()
Docteon-sha256.c70 unsigned int partial; in __octeon_sha256_update() local
74 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update()
79 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update()
80 if (partial) { in __octeon_sha256_update()
81 done = -partial; in __octeon_sha256_update()
82 memcpy(sctx->buf + partial, data, in __octeon_sha256_update()
93 partial = 0; in __octeon_sha256_update()
95 memcpy(sctx->buf + partial, src, len - done); in __octeon_sha256_update()
/linux-6.12.1/drivers/usb/storage/
Dfreecom.c227 unsigned int partial; in freecom_transport() local
266 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport()
267 usb_stor_dbg(us, "foo Status result %d %u\n", result, partial); in freecom_transport()
271 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport()
308 FCM_STATUS_PACKET_LENGTH, &partial); in freecom_transport()
310 usb_stor_dbg(us, "bar Status result %d %u\n", result, partial); in freecom_transport()
314 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport()
317 if (partial != 4) in freecom_transport()
376 FCM_PACKET_LENGTH, &partial); in freecom_transport()
377 US_DEBUG(pdump(us, (void *)fst, partial)); in freecom_transport()
[all …]
/linux-6.12.1/net/smc/
Dsmc_rx.c154 struct partial_page *partial; in smc_rx_splice() local
167 partial = kcalloc(nr_pages, sizeof(*partial), GFP_KERNEL); in smc_rx_splice()
168 if (!partial) in smc_rx_splice()
184 partial[0].offset = src - (char *)smc->conn.rmb_desc->cpu_addr; in smc_rx_splice()
185 partial[0].len = len; in smc_rx_splice()
186 partial[0].private = (unsigned long)priv[0]; in smc_rx_splice()
197 partial[i].offset = offset; in smc_rx_splice()
198 partial[i].len = size; in smc_rx_splice()
199 partial[i].private = (unsigned long)priv[i]; in smc_rx_splice()
208 spd.partial = partial; in smc_rx_splice()
[all …]
/linux-6.12.1/arch/x86/include/asm/
Dunwind.h72 bool *partial) in unwind_get_entry_regs() argument
77 if (partial) { in unwind_get_entry_regs()
79 *partial = !state->full_regs; in unwind_get_entry_regs()
81 *partial = false; in unwind_get_entry_regs()
89 bool *partial) in unwind_get_entry_regs() argument
/linux-6.12.1/crypto/
Dsha3_generic.c168 sctx->partial = 0; in crypto_sha3_init()
185 if ((sctx->partial + len) > (sctx->rsiz - 1)) { in crypto_sha3_update()
186 if (sctx->partial) { in crypto_sha3_update()
187 done = -sctx->partial; in crypto_sha3_update()
188 memcpy(sctx->buf + sctx->partial, data, in crypto_sha3_update()
204 sctx->partial = 0; in crypto_sha3_update()
206 memcpy(sctx->buf + sctx->partial, src, len - done); in crypto_sha3_update()
207 sctx->partial += (len - done); in crypto_sha3_update()
216 unsigned int i, inlen = sctx->partial; in crypto_sha3_final()
Dsm3.c180 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_update() local
185 if ((partial + len) >= SM3_BLOCK_SIZE) { in sm3_update()
188 if (partial) { in sm3_update()
189 int p = SM3_BLOCK_SIZE - partial; in sm3_update()
191 memcpy(sctx->buffer + partial, data, p); in sm3_update()
208 partial = 0; in sm3_update()
211 memcpy(sctx->buffer + partial, data, len); in sm3_update()
220 unsigned int partial = sctx->count % SM3_BLOCK_SIZE; in sm3_final() local
224 sctx->buffer[partial++] = 0x80; in sm3_final()
225 if (partial > bit_offset) { in sm3_final()
[all …]
/linux-6.12.1/drivers/crypto/stm32/
Dstm32-crc32.c69 u32 partial; /* crc32c: partial in first 4 bytes of that struct */ member
137 ctx->partial = readl_relaxed(crc->regs + CRC_DR); in stm32_crc_init()
163 ctx->partial = crc32_le(ctx->partial, d8, length); in burst_update()
165 ctx->partial = __crc32c_le(ctx->partial, d8, length); in burst_update()
176 writel_relaxed(bitrev32(ctx->partial), crc->regs + CRC_INIT); in burst_update()
206 ctx->partial = readl_relaxed(crc->regs + CRC_DR); in burst_update()
249 ~ctx->partial : ctx->partial, out); in stm32_crc_final()
/linux-6.12.1/arch/x86/kernel/
Ddumpstack.c155 bool partial, const char *log_lvl) in show_regs_if_on_stack() argument
166 if (!partial && on_stack(info, regs, sizeof(*regs))) { in show_regs_if_on_stack()
169 } else if (partial && on_stack(info, (void *)regs + IRET_FRAME_OFFSET, in show_regs_if_on_stack()
193 bool partial = false; in show_trace_log_lvl() local
198 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl()
240 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in show_trace_log_lvl()
299 regs = unwind_get_entry_regs(&state, &partial); in show_trace_log_lvl()
301 show_regs_if_on_stack(&stack_info, regs, partial, log_lvl); in show_trace_log_lvl()

12345678910