Searched refs:crypto_tfm_ctx_alignment (Results 1 – 25 of 25) sorted by relevance
/linux-6.12.1/include/crypto/internal/ |
D | kpp.h | 57 if (align <= crypto_tfm_ctx_alignment()) in kpp_request_ctx_dma() 72 reqsize += crypto_dma_align() & ~(crypto_tfm_ctx_alignment() - 1); in kpp_set_reqsize_dma()
|
D | akcipher.h | 40 if (align <= crypto_tfm_ctx_alignment()) in akcipher_request_ctx_dma() 55 reqsize += crypto_dma_align() & ~(crypto_tfm_ctx_alignment() - 1); in akcipher_set_reqsize_dma()
|
D | aead.h | 77 if (align <= crypto_tfm_ctx_alignment()) in aead_request_ctx_dma() 129 reqsize += crypto_dma_align() & ~(crypto_tfm_ctx_alignment() - 1); in crypto_aead_set_reqsize_dma()
|
D | hash.h | 169 reqsize += crypto_dma_align() & ~(crypto_tfm_ctx_alignment() - 1); in crypto_ahash_set_reqsize_dma() 200 if (align <= crypto_tfm_ctx_alignment()) in ahash_request_ctx_dma()
|
D | skcipher.h | 190 reqsize += crypto_dma_align() & ~(crypto_tfm_ctx_alignment() - 1); in crypto_skcipher_set_reqsize_dma() 249 if (align <= crypto_tfm_ctx_alignment()) in skcipher_request_ctx_dma()
|
/linux-6.12.1/include/crypto/ |
D | algapi.h | 189 if (align <= crypto_tfm_ctx_alignment()) in crypto_tfm_ctx_align() 202 return (crypto_dma_align() - 1) & ~(crypto_tfm_ctx_alignment() - 1); in crypto_dma_padding()
|
/linux-6.12.1/crypto/ |
D | cts.c | 303 crypto_tfm_ctx_alignment()) + in crypto_cts_init_tfm() 304 (align & ~(crypto_tfm_ctx_alignment() - 1)) + bsize; in crypto_cts_init_tfm()
|
D | gcm.c | 542 align &= ~(crypto_tfm_ctx_alignment() - 1); in crypto_gcm_init_tfm() 800 align &= ~(crypto_tfm_ctx_alignment() - 1); in crypto_rfc4106_init_tfm() 804 ALIGN(crypto_aead_reqsize(aead), crypto_tfm_ctx_alignment()) + in crypto_rfc4106_init_tfm() 1007 align &= ~(crypto_tfm_ctx_alignment() - 1); in crypto_rfc4543_init_tfm() 1011 ALIGN(crypto_aead_reqsize(aead), crypto_tfm_ctx_alignment()) + in crypto_rfc4543_init_tfm()
|
D | ccm.c | 416 align &= ~(crypto_tfm_ctx_alignment() - 1); in crypto_ccm_init_tfm() 681 align &= ~(crypto_tfm_ctx_alignment() - 1); in crypto_rfc4309_init_tfm() 685 ALIGN(crypto_aead_reqsize(aead), crypto_tfm_ctx_alignment()) + in crypto_rfc4309_init_tfm()
|
D | skcipher.c | 242 a = crypto_tfm_ctx_alignment() - 1; in skcipher_next_slow() 397 unsigned a = crypto_tfm_ctx_alignment() - 1; in skcipher_copy_iv() 737 reqsize = am & ~(crypto_tfm_ctx_alignment() - 1); in crypto_skcipher_init_tfm()
|
D | ctr.c | 234 align &= ~(crypto_tfm_ctx_alignment() - 1); in crypto_rfc3686_init_tfm()
|
D | api.c | 378 len = alg->cra_alignmask & ~(crypto_tfm_ctx_alignment() - 1); in crypto_ctxsize()
|
D | chacha20poly1305.c | 524 align &= ~(crypto_tfm_ctx_alignment() - 1); in chachapoly_init()
|
D | ahash.c | 279 reqsize = ALIGN(reqsize, crypto_tfm_ctx_alignment()); in ahash_save_req()
|
D | algapi.c | 998 (alg->cra_alignmask & ~(crypto_tfm_ctx_alignment() - 1)); in crypto_alg_extsize()
|
/linux-6.12.1/include/linux/ |
D | crypto.h | 490 static inline unsigned int crypto_tfm_ctx_alignment(void) in crypto_tfm_ctx_alignment() function
|
/linux-6.12.1/net/ipv4/ |
D | ah4.c | 32 len = ALIGN(len, crypto_tfm_ctx_alignment()); in ah_alloc_tmp() 58 crypto_tfm_ctx_alignment()); in ah_tmp_req()
|
D | esp4.c | 57 ~(crypto_tfm_ctx_alignment() - 1); in esp_alloc_tmp() 58 len = ALIGN(len, crypto_tfm_ctx_alignment()); in esp_alloc_tmp() 86 crypto_tfm_ctx_alignment()); in esp_tmp_req()
|
/linux-6.12.1/net/ipv6/ |
D | ah6.c | 56 len = ALIGN(len, crypto_tfm_ctx_alignment()); in ah_alloc_tmp() 87 crypto_tfm_ctx_alignment()); in ah_tmp_req()
|
D | esp6.c | 74 ~(crypto_tfm_ctx_alignment() - 1); in esp_alloc_tmp() 75 len = ALIGN(len, crypto_tfm_ctx_alignment()); in esp_alloc_tmp() 103 crypto_tfm_ctx_alignment()); in esp_tmp_req()
|
/linux-6.12.1/drivers/crypto/ |
D | padlock-aes.c | 92 if (align <= crypto_tfm_ctx_alignment()) in aes_ctx_common()
|
/linux-6.12.1/net/tipc/ |
D | crypto.c | 699 len += crypto_aead_alignmask(tfm) & ~(crypto_tfm_ctx_alignment() - 1); in tipc_aead_mem_alloc() 700 len = ALIGN(len, crypto_tfm_ctx_alignment()); in tipc_aead_mem_alloc() 712 crypto_tfm_ctx_alignment()); in tipc_aead_mem_alloc()
|
/linux-6.12.1/arch/x86/crypto/ |
D | aesni-intel_glue.c | 56 if (crypto_tfm_ctx_alignment() >= AESNI_ALIGN) in aes_align_addr()
|
/linux-6.12.1/fs/smb/client/ |
D | smb2ops.c | 4213 len += crypto_aead_alignmask(tfm) & ~(crypto_tfm_ctx_alignment() - 1); in smb2_aead_req_alloc() 4214 len = ALIGN(len, crypto_tfm_ctx_alignment()); in smb2_aead_req_alloc() 4226 crypto_tfm_ctx_alignment()); in smb2_aead_req_alloc()
|
/linux-6.12.1/drivers/crypto/hisilicon/hpre/ |
D | hpre_crypto.c | 160 return (hpre_align_sz() - 1) & ~(crypto_tfm_ctx_alignment() - 1); in hpre_align_pd()
|