/linux-6.12.1/crypto/ |
D | scompress.c | 209 struct crypto_alg *calg = tfm->__crt_alg; in crypto_init_scomp_ops_async() local 214 if (!crypto_mod_get(calg)) in crypto_init_scomp_ops_async() 217 scomp = crypto_create_tfm(calg, &crypto_scomp_type); in crypto_init_scomp_ops_async() 219 crypto_mod_put(calg); in crypto_init_scomp_ops_async() 282 struct crypto_alg *base = &alg->calg.base; in crypto_register_scomp() 284 comp_prepare_alg(&alg->calg); in crypto_register_scomp()
|
D | akcipher.c | 261 struct crypto_alg *calg = tfm->__crt_alg; in crypto_init_akcipher_ops_sig() local 264 if (!crypto_mod_get(calg)) in crypto_init_akcipher_ops_sig() 267 akcipher = crypto_create_tfm(calg, &crypto_akcipher_type); in crypto_init_akcipher_ops_sig() 269 crypto_mod_put(calg); in crypto_init_akcipher_ops_sig()
|
D | acompress.c | 28 return container_of(alg, struct acomp_alg, calg.base); in __crypto_acomp_alg() 165 struct crypto_alg *base = &alg->calg.base; in crypto_register_acomp() 167 comp_prepare_alg(&alg->calg); in crypto_register_acomp()
|
D | lskcipher.c | 309 struct crypto_alg *calg = tfm->__crt_alg; in crypto_init_lskcipher_ops_sg() local 312 if (!crypto_mod_get(calg)) in crypto_init_lskcipher_ops_sg() 315 skcipher = crypto_create_tfm(calg, &crypto_lskcipher_type); in crypto_init_lskcipher_ops_sg() 317 crypto_mod_put(calg); in crypto_init_lskcipher_ops_sg()
|
D | ahash.c | 115 struct crypto_alg *calg = tfm->__crt_alg; in crypto_init_ahash_using_shash() local 120 if (!crypto_mod_get(calg)) in crypto_init_ahash_using_shash() 123 shash = crypto_create_tfm(calg, &crypto_shash_type); in crypto_init_ahash_using_shash() 125 crypto_mod_put(calg); in crypto_init_ahash_using_shash()
|
/linux-6.12.1/net/xfrm/ |
D | xfrm_ipcomp.c | 333 if (!x->calg) { in ipcomp_init_state() 352 ipcd->tfms = ipcomp_alloc_tfms(x->calg->alg_name); in ipcomp_init_state() 357 calg_desc = xfrm_calg_get_byname(x->calg->alg_name, 0); in ipcomp_init_state()
|
D | xfrm_user.c | 822 if ((err = attach_one_algo(&x->calg, &x->props.calgo, in xfrm_state_construct() 1160 static int copy_to_user_calg(struct xfrm_algo *calg, struct sk_buff *skb) in copy_to_user_calg() argument 1162 struct nlattr *nla = nla_reserve(skb, XFRMA_ALG_COMP, sizeof(*calg)); in copy_to_user_calg() 1169 strscpy_pad(ap->alg_name, calg->alg_name, sizeof(ap->alg_name)); in copy_to_user_calg() 1248 if (x->calg) { in copy_to_user_state_extra() 1249 ret = copy_to_user_calg(x->calg, skb); in copy_to_user_state_extra() 3459 if (x->calg) in xfrm_sa_len() 3460 l += nla_total_size(sizeof(*x->calg)); in xfrm_sa_len()
|
D | xfrm_state.c | 531 kfree(x->calg); in ___xfrm_state_destroy() 1753 if (orig->calg) { in xfrm_state_clone() 1754 x->calg = xfrm_algo_clone(orig->calg); in xfrm_state_clone() 1755 if (!x->calg) in xfrm_state_clone()
|
/linux-6.12.1/include/crypto/internal/ |
D | acompress.h | 48 struct comp_alg_common calg; member
|
D | scompress.h | 45 struct comp_alg_common calg; member
|
/linux-6.12.1/drivers/net/netdevsim/ |
D | ipsec.c | 147 if (xs->calg) { in nsim_ipsec_add_sa()
|
/linux-6.12.1/net/key/ |
D | af_key.c | 859 BUG_ON(x->ealg && x->calg); in __pfkey_xfrm_state2msg() 866 if (x->calg) { in __pfkey_xfrm_state2msg() 867 struct xfrm_algo_desc *a = xfrm_calg_get_byname(x->calg->alg_name, 0); in __pfkey_xfrm_state2msg() 1199 x->calg = kmalloc(sizeof(*x->calg), GFP_KERNEL); in pfkey_msg2xfrm_state() 1200 if (!x->calg) { in pfkey_msg2xfrm_state() 1204 strcpy(x->calg->alg_name, a->name); in pfkey_msg2xfrm_state()
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbevf/ |
D | ipsec.c | 292 if (xs->calg) { in ixgbevf_ipsec_add_sa()
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_ipsec.c | 597 if (xs->calg) { in ixgbe_ipsec_add_sa()
|
/linux-6.12.1/drivers/crypto/hisilicon/sec2/ |
D | sec_crypto.c | 1205 #define GEN_SEC_AEAD_SETKEY_FUNC(name, aalg, calg, maclen, cmode) \ argument 1209 return sec_aead_setkey(tfm, key, keylen, aalg, calg, maclen, cmode);\
|
/linux-6.12.1/include/net/ |
D | xfrm.h | 222 struct xfrm_algo *calg; member
|