/linux-6.12.1/lib/crypto/ |
D | chacha20poly1305.c | 57 __chacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, in __chacha20poly1305_encrypt() argument 74 chacha20_crypt(chacha_state, dst, src, src_len); in __chacha20poly1305_encrypt() 76 poly1305_update(&poly1305_state, dst, src_len); in __chacha20poly1305_encrypt() 77 if (src_len & 0xf) in __chacha20poly1305_encrypt() 78 poly1305_update(&poly1305_state, pad0, 0x10 - (src_len & 0xf)); in __chacha20poly1305_encrypt() 81 b.lens[1] = cpu_to_le64(src_len); in __chacha20poly1305_encrypt() 84 poly1305_final(&poly1305_state, dst + src_len); in __chacha20poly1305_encrypt() 90 void chacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, in chacha20poly1305_encrypt() argument 105 __chacha20poly1305_encrypt(dst, src, src_len, ad, ad_len, chacha_state); in chacha20poly1305_encrypt() 112 void xchacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, in xchacha20poly1305_encrypt() argument [all …]
|
/linux-6.12.1/lib/ |
D | decompress_unlzo.c | 105 u32 src_len, dst_len; in unlzo() local 207 src_len = get_unaligned_be32(in_buf); in unlzo() 211 if (src_len <= 0 || src_len > dst_len) { in unlzo() 217 if (fill && in_len < src_len) { in unlzo() 218 skip = fill(in_buf + in_len, src_len - in_len); in unlzo() 222 if (in_len < src_len) { in unlzo() 231 if (unlikely(dst_len == src_len)) in unlzo() 232 memcpy(out_buf, in_buf, src_len); in unlzo() 234 r = lzo1x_decompress_safe((u8 *) in_buf, src_len, in unlzo() 248 *posp += src_len + 12; in unlzo() [all …]
|
/linux-6.12.1/arch/s390/include/asm/ |
D | cpacf.h | 402 u8 *dest, const u8 *src, long src_len) in cpacf_km() argument 408 s.odd = (unsigned long)src_len; in cpacf_km() 419 return src_len - s.odd; in cpacf_km() 434 u8 *dest, const u8 *src, long src_len) in cpacf_kmc() argument 440 s.odd = (unsigned long)src_len; in cpacf_kmc() 451 return src_len - s.odd; in cpacf_kmc() 463 const u8 *src, long src_len) in cpacf_kimd() argument 468 s.odd = (unsigned long)src_len; in cpacf_kimd() 488 const u8 *src, long src_len) in cpacf_klmd() argument 493 s.odd = (unsigned long)src_len; in cpacf_klmd() [all …]
|
/linux-6.12.1/fs/bcachefs/ |
D | compress.c | 156 size_t src_len = src->bi_iter.bi_size; in __bio_uncompress() local 167 src_len, dst_len, dst_len); in __bio_uncompress() 174 .avail_in = src_len, in __bio_uncompress() 195 if (real_src_len > src_len - 4) in __bio_uncompress() 298 void *src, size_t src_len, in attempt_compress() argument 307 int len = src_len; in attempt_compress() 312 if (len < src_len) in attempt_compress() 319 src_len, dst_len, in attempt_compress() 328 .avail_in = src_len, in attempt_compress() 372 src, src_len, in attempt_compress() [all …]
|
/linux-6.12.1/include/crypto/ |
D | chacha20poly1305.h | 18 void chacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, 24 chacha20poly1305_decrypt(u8 *dst, const u8 *src, const size_t src_len, 28 void xchacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, 34 u8 *dst, const u8 *src, const size_t src_len, const u8 *ad, 38 bool chacha20poly1305_encrypt_sg_inplace(struct scatterlist *src, size_t src_len, 43 bool chacha20poly1305_decrypt_sg_inplace(struct scatterlist *src, size_t src_len,
|
D | akcipher.h | 39 unsigned int src_len; member 254 unsigned int src_len, in akcipher_request_set_crypt() argument 259 req->src_len = src_len; in akcipher_request_set_crypt()
|
/linux-6.12.1/net/smc/ |
D | smc_tx.c | 357 size_t src_off, size_t src_len, in smcr_tx_rdma_writes() argument 366 int src_len_sum = src_len, dst_len_sum = dst_len; in smcr_tx_rdma_writes() 388 sge[srcchunk].length = src_len; in smcr_tx_rdma_writes() 394 src_off += src_len; in smcr_tx_rdma_writes() 401 src_len = dst_len - src_len; /* remainder */ in smcr_tx_rdma_writes() 402 src_len_sum += src_len; in smcr_tx_rdma_writes() 413 src_len = min_t(int, dst_len, conn->sndbuf_desc->len - in smcr_tx_rdma_writes() 415 src_len_sum = src_len; in smcr_tx_rdma_writes() 422 size_t src_off, size_t src_len, in smcd_tx_rdma_writes() argument 425 int src_len_sum = src_len, dst_len_sum = dst_len; in smcd_tx_rdma_writes() [all …]
|
/linux-6.12.1/net/ipv4/ |
D | fib_rules.c | 38 u8 src_len; member 54 if (r->dst_len || r->src_len || r->dscp) in fib4_rule_matchall() 287 if (frh->src_len) in fib4_rule_configure() 304 rule4->src_len = frh->src_len; in fib4_rule_configure() 305 rule4->srcmask = inet_make_mask(rule4->src_len); in fib4_rule_configure() 344 if (frh->src_len && (rule4->src_len != frh->src_len)) in fib4_rule_compare() 368 if (frh->src_len && (rule4->src != nla_get_in_addr(tb[FRA_SRC]))) in fib4_rule_compare() 383 frh->src_len = rule4->src_len; in fib4_rule_fill() 396 (rule4->src_len && in fib4_rule_fill()
|
/linux-6.12.1/include/linux/ |
D | ccp.h | 183 u64 src_len; /* In bytes */ member 243 u64 src_len; /* In bytes */ member 292 u64 src_len; /* In bytes */ member 354 u64 src_len; /* In bytes */ member 382 u32 src_len; /* In bytes */ member 441 u64 src_len; /* In bytes */ member 470 u64 src_len; /* In bytes */ member
|
D | lzo.h | 24 int lzo1x_1_compress(const unsigned char *src, size_t src_len, 28 int lzorle1x_1_compress(const unsigned char *src, size_t src_len, 32 int lzo1x_decompress_safe(const unsigned char *src, size_t src_len,
|
/linux-6.12.1/drivers/misc/mei/ |
D | vsc-tp.c | 161 int ret, offset = 0, cpy_len, src_len, dst_len = sizeof(struct vsc_tp_packet); in vsc_tp_xfer_helper() local 178 src_len = next_xfer_len; in vsc_tp_xfer_helper() 184 src_len = next_xfer_len - (src - rx_buf); in vsc_tp_xfer_helper() 188 while (src_len > 0) { in vsc_tp_xfer_helper() 189 cpy_len = min(src_len, dst_len); in vsc_tp_xfer_helper() 193 src_len -= cpy_len; in vsc_tp_xfer_helper() 201 if (!src_len) in vsc_tp_xfer_helper() 215 if (src_len) { in vsc_tp_xfer_helper() 218 cpy_len = min(src_len, remain); in vsc_tp_xfer_helper() 222 src_len -= cpy_len; in vsc_tp_xfer_helper() [all …]
|
/linux-6.12.1/drivers/block/zram/ |
D | backend_lz4hc.c | 77 ret = LZ4_compress_HC(req->src, req->dst, req->src_len, in lz4hc_compress() 88 req->src_len, req->dst_len); in lz4hc_compress() 103 ret = LZ4_decompress_safe(req->src, req->dst, req->src_len, in lz4hc_decompress() 112 req->dst, req->src_len, in lz4hc_decompress()
|
D | backend_lz4.c | 77 ret = LZ4_compress_fast(req->src, req->dst, req->src_len, in lz4_compress() 86 req->dst, req->src_len, in lz4_compress() 102 ret = LZ4_decompress_safe(req->src, req->dst, req->src_len, in lz4_decompress() 111 req->dst, req->src_len, in lz4_decompress()
|
D | backend_zstd.c | 187 req->src, req->src_len, &zp->cprm); in zstd_compress() 191 req->src_len, in zstd_compress() 208 req->src, req->src_len); in zstd_decompress() 212 req->src_len, zp->ddict); in zstd_decompress()
|
D | zcomp.c | 127 .src_len = PAGE_SIZE, in zcomp_compress() 139 const void *src, unsigned int src_len, void *dst) in zcomp_decompress() argument 144 .src_len = src_len, in zcomp_decompress()
|
D | backend_842.c | 38 ret = sw842_compress(req->src, req->src_len, req->dst, &dlen, in compress_842() 50 return sw842_decompress(req->src, req->src_len, req->dst, &dlen); in decompress_842()
|
D | backend_lzo.c | 36 ret = lzo1x_1_compress(req->src, req->src_len, req->dst, in lzo_compress() 46 ret = lzo1x_decompress_safe(req->src, req->src_len, in lzo_decompress()
|
D | backend_lzorle.c | 36 ret = lzorle1x_1_compress(req->src, req->src_len, req->dst, in lzorle_compress() 46 ret = lzo1x_decompress_safe(req->src, req->src_len, in lzorle_decompress()
|
/linux-6.12.1/crypto/ |
D | rsa-pkcs1pad.c | 260 if (req->src_len > ctx->key_size - 11) in pkcs1pad_encrypt() 268 req_ctx->in_buf = kmalloc(ctx->key_size - 1 - req->src_len, in pkcs1pad_encrypt() 273 ps_end = ctx->key_size - req->src_len - 2; in pkcs1pad_encrypt() 280 ctx->key_size - 1 - req->src_len, req->src); in pkcs1pad_encrypt() 371 if (!ctx->key_size || req->src_len != ctx->key_size) in pkcs1pad_decrypt() 387 req_ctx->out_sg, req->src_len, in pkcs1pad_decrypt() 414 if (req->src_len + digest_info_size > ctx->key_size - 11) in pkcs1pad_sign() 422 req_ctx->in_buf = kmalloc(ctx->key_size - 1 - req->src_len, in pkcs1pad_sign() 427 ps_end = ctx->key_size - digest_info_size - req->src_len - 2; in pkcs1pad_sign() 437 ctx->key_size - 1 - req->src_len, req->src); in pkcs1pad_sign() [all …]
|
D | ecrdsa.c | 99 req->dst_len * 2 != req->src_len || in ecrdsa_verify() 100 WARN_ON(req->src_len > sizeof(sig)) || in ecrdsa_verify() 104 sg_copy_to_buffer(req->src, sg_nents_for_len(req->src, req->src_len), in ecrdsa_verify() 105 sig, req->src_len); in ecrdsa_verify() 108 req->src_len + req->dst_len), in ecrdsa_verify() 109 digest, req->dst_len, req->src_len); in ecrdsa_verify()
|
D | ecdsa.c | 144 buffer = kmalloc(req->src_len + req->dst_len, GFP_KERNEL); in ecdsa_verify() 149 sg_nents_for_len(req->src, req->src_len + req->dst_len), in ecdsa_verify() 150 buffer, req->src_len + req->dst_len, 0); in ecdsa_verify() 153 buffer, req->src_len); in ecdsa_verify() 160 ecc_digits_from_bytes(buffer + req->src_len, bufsize, in ecdsa_verify()
|
/linux-6.12.1/drivers/staging/rtl8723bs/os_dep/ |
D | osdep_service.c | 122 void rtw_buf_update(u8 **buf, u32 *buf_len, u8 *src, u32 src_len) in rtw_buf_update() argument 131 if (!src || !src_len) in rtw_buf_update() 135 dup = rtw_malloc(src_len); in rtw_buf_update() 137 dup_len = src_len; in rtw_buf_update()
|
/linux-6.12.1/kernel/bpf/ |
D | crypto.c | 270 u32 src_len, dst_len, siv_len; in bpf_crypto_crypt() local 279 src_len = __bpf_dynptr_size(src); in bpf_crypto_crypt() 281 if (!src_len || !dst_len) in bpf_crypto_crypt() 287 psrc = __bpf_dynptr_data(src, src_len); in bpf_crypto_crypt() 298 err = decrypt ? ctx->type->decrypt(ctx->tfm, psrc, pdst, src_len, piv) in bpf_crypto_crypt() 299 : ctx->type->encrypt(ctx->tfm, psrc, pdst, src_len, piv); in bpf_crypto_crypt()
|
/linux-6.12.1/tools/bpf/bpftool/ |
D | tracelog.c | 42 size_t src_len; in find_tracefs_mnt_single() local 47 src_len = strlen(mntpt); in find_tracefs_mnt_single() 48 if (src_len + 1 >= PATH_MAX) { in find_tracefs_mnt_single()
|
/linux-6.12.1/drivers/crypto/virtio/ |
D | virtio_crypto_akcipher_algs.c | 234 unsigned int src_len = verify ? req->src_len + req->dst_len : req->src_len; in __virtio_crypto_akcipher_do_req() local 241 src_buf = kcalloc_node(src_len, 1, GFP_KERNEL, node); in __virtio_crypto_akcipher_do_req() 247 sg_copy_to_buffer(req->src, sg_nents(req->src), src_buf, src_len); in __virtio_crypto_akcipher_do_req() 248 sg_init_one(&srcdata_sg, src_buf, src_len); in __virtio_crypto_akcipher_do_req() 251 sg_copy_to_buffer(req->src, sg_nents(req->src), src_buf, src_len); in __virtio_crypto_akcipher_do_req() 252 sg_init_one(&srcdata_sg, src_buf, src_len); in __virtio_crypto_akcipher_do_req() 313 akcipher_req->para.src_data_len = cpu_to_le32(req->src_len); in virtio_crypto_rsa_do_req()
|