Home
last modified time | relevance | path

Searched refs:iv_size (Results 1 – 15 of 15) sorted by relevance

/linux-6.12.1/drivers/md/
Ddm-crypt.c181 unsigned int iv_size; member
351 memset(iv, 0, cc->iv_size); in crypt_iv_plain_gen()
360 memset(iv, 0, cc->iv_size); in crypt_iv_plain64_gen()
369 memset(iv, 0, cc->iv_size); in crypt_iv_plain64be_gen()
371 *(__be64 *)&iv[cc->iv_size - sizeof(u64)] = cpu_to_be64(dmreq->iv_sector); in crypt_iv_plain64be_gen()
383 memset(iv, 0, cc->iv_size); in crypt_iv_essiv_gen()
429 memset(iv, 0, cc->iv_size - sizeof(u64)); /* rest is cleared below */ in crypt_iv_benbi_gen()
432 put_unaligned(val, (__be64 *)(iv + cc->iv_size - sizeof(u64))); in crypt_iv_benbi_gen()
440 memset(iv, 0, cc->iv_size); in crypt_iv_null_gen()
556 memcpy(iv, &md5state.hash, cc->iv_size); in crypt_iv_lmk_one()
[all …]
/linux-6.12.1/drivers/crypto/intel/keembay/
Dkeembay-ocs-aes-core.c256 int iv_size = crypto_skcipher_ivsize(tfm); in kmb_ocs_sk_validate_input() local
272 if (!req->iv || iv_size != AES_BLOCK_SIZE) in kmb_ocs_sk_validate_input()
283 if (!req->iv || iv_size != AES_BLOCK_SIZE) in kmb_ocs_sk_validate_input()
293 if (!req->iv || iv_size != AES_BLOCK_SIZE) in kmb_ocs_sk_validate_input()
403 int iv_size = crypto_skcipher_ivsize(tfm); in kmb_ocs_sk_prepare_inplace() local
415 req->cryptlen - iv_size, iv_size, 0); in kmb_ocs_sk_prepare_inplace()
520 int iv_size = crypto_skcipher_ivsize(tfm); in kmb_ocs_sk_run() local
551 req->cryptlen, req->iv, iv_size); in kmb_ocs_sk_run()
571 req->cryptlen - iv_size, in kmb_ocs_sk_run()
572 iv_size, 0); in kmb_ocs_sk_run()
[all …]
Docs-aes.c601 const u8 *iv, u32 iv_size, in ocs_aes_validate_inputs() argument
664 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs()
680 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs()
696 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs()
703 if (!iv || iv_size != GCM_AES_IV_SIZE) in ocs_aes_validate_inputs()
730 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs()
804 u32 iv_size) in ocs_aes_op() argument
809 rc = ocs_aes_validate_inputs(src_dma_list, src_size, iv, iv_size, 0, 0, in ocs_aes_op()
Docs-aes.h78 u32 iv_size);
/linux-6.12.1/net/tls/
Dtls.h316 prot->iv_size); in tls_advance_record_sn()
336 size_t pkt_len, iv_size = prot->iv_size; in tls_fill_prepend() local
341 pkt_len += iv_size; in tls_fill_prepend()
344 ctx->tx.iv + prot->salt_size, iv_size); in tls_fill_prepend()
Dtls_sw.c558 prot->iv_size + prot->salt_size); in tls_do_encryption()
1544 prot->iv_size + prot->salt_size); in tls_decrypt_sg()
1548 prot->iv_size); in tls_decrypt_sg()
2389 cipher_overhead += prot->iv_size; in tls_rx_msg_size()
2680 prot->iv_size = cipher_desc->iv; in init_prot_info()
/linux-6.12.1/drivers/s390/cio/
Dairq.c113 static inline unsigned long iv_size(unsigned long bits) in iv_size() function
137 size = iv_size(bits); in airq_iv_create()
209 cio_dma_free(iv->vector, iv_size(iv->bits)); in airq_iv_release()
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/
Damdgpu_ih.c290 uint32_t iv_size = 32; in amdgpu_ih_decode_iv_ts_helper() local
294 rptr += iv_size * offset; in amdgpu_ih_decode_iv_ts_helper()
/linux-6.12.1/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/
Dchcr_ktls.h50 u32 iv_size; member
Dchcr_ktls.c90 tx_info->iv_size = TLS_CIPHER_AES_GCM_128_IV_SIZE; in chcr_ktls_save_keys()
1183 cipher_start = TLS_HEADER_SIZE + tx_info->iv_size + 1; in chcr_ktls_xmit_wr_complete()
1362 (!tls_rec_offset ? TLS_HEADER_SIZE + tx_info->iv_size : 0); in chcr_ktls_xmit_wr_short()
1419 memcpy(pos + tx_info->salt_size, &iv_record, tx_info->iv_size); in chcr_ktls_xmit_wr_short()
1420 *(__be32 *)(pos + tx_info->salt_size + tx_info->iv_size) = in chcr_ktls_xmit_wr_short()
1422 (TLS_HEADER_SIZE + tx_info->iv_size)) / AES_BLOCK_LEN) : 0)); in chcr_ktls_xmit_wr_short()
1802 if (tls_rec_offset + data_len <= (TLS_HEADER_SIZE + tx_info->iv_size)) { in chcr_short_record_handler()
1820 if (tls_rec_offset < (TLS_HEADER_SIZE + tx_info->iv_size)) { in chcr_short_record_handler()
1827 (TLS_HEADER_SIZE + tx_info->iv_size)) in chcr_short_record_handler()
/linux-6.12.1/drivers/crypto/
Dsa2ul.c108 u8 iv_size; member
585 if (cfg->iv_size) in sa_format_cmdl_gen()
586 auth_offset += cfg->iv_size; in sa_format_cmdl_gen()
601 if (cfg->iv_size) { in sa_format_cmdl_gen()
605 upd_info->enc_iv.size = cfg->iv_size; in sa_format_cmdl_gen()
608 SA_CMDL_HEADER_SIZE_BYTES + cfg->iv_size; in sa_format_cmdl_gen()
611 (SA_CTX_ENC_AUX2_OFFSET | (cfg->iv_size >> 3)); in sa_format_cmdl_gen()
612 total += SA_CMDL_HEADER_SIZE_BYTES + cfg->iv_size; in sa_format_cmdl_gen()
903 cfg.iv_size = crypto_skcipher_ivsize(tfm); in sa_cipher_setkey()
1467 cfg.iv_size = 0; in sa_sha_setup()
[all …]
/linux-6.12.1/drivers/crypto/hisilicon/sec2/
Dsec_crypto.c1392 u32 iv_size = req->ctx->c_ctx.ivsize; in sec_update_iv() local
1412 sz = sg_pcopy_to_buffer(sgl, sg_nents(sgl), iv, iv_size, in sec_update_iv()
1413 cryptlen - iv_size); in sec_update_iv()
1414 if (unlikely(sz != iv_size)) in sec_update_iv()
1417 sz = cryptlen / iv_size; in sec_update_iv()
1418 if (cryptlen % iv_size) in sec_update_iv()
1420 ctr_iv_inc(iv, iv_size, sz); in sec_update_iv()
2147 sec_min_key_size, sec_max_key_size, blk_size, iv_size)\ argument
2166 .ivsize = iv_size,\
2387 ctx_exit, blk_size, iv_size, max_authsize)\ argument
[all …]
/linux-6.12.1/include/net/
Dtls.h211 u16 iv_size; member
/linux-6.12.1/net/tipc/
Dcrypto.c690 unsigned int iv_size, req_size; in tipc_aead_mem_alloc() local
694 iv_size = crypto_aead_ivsize(tfm); in tipc_aead_mem_alloc()
698 len += iv_size; in tipc_aead_mem_alloc()
711 *req = (struct aead_request *)PTR_ALIGN(*iv + iv_size, in tipc_aead_mem_alloc()
/linux-6.12.1/fs/smb/client/
Dsmb2ops.c4204 unsigned int iv_size = crypto_aead_ivsize(tfm); in smb2_aead_req_alloc() local
4212 len = iv_size; in smb2_aead_req_alloc()
4225 *req = (struct aead_request *)PTR_ALIGN(*iv + iv_size, in smb2_aead_req_alloc()