/linux-6.12.1/drivers/crypto/ccp/ |
D | ccp-crypto-rsa.c | 52 req->dst_len = rctx->cmd.u.rsa.key_size >> 3; in ccp_rsa_complete() 61 return ctx->u.rsa.n_len; in ccp_rsa_maxsize() 75 rctx->cmd.u.rsa.key_size = ctx->u.rsa.key_len; /* in bits */ in ccp_rsa_crypt() 77 rctx->cmd.u.rsa.exp = &ctx->u.rsa.e_sg; in ccp_rsa_crypt() 78 rctx->cmd.u.rsa.exp_len = ctx->u.rsa.e_len; in ccp_rsa_crypt() 80 rctx->cmd.u.rsa.exp = &ctx->u.rsa.d_sg; in ccp_rsa_crypt() 81 rctx->cmd.u.rsa.exp_len = ctx->u.rsa.d_len; in ccp_rsa_crypt() 83 rctx->cmd.u.rsa.mod = &ctx->u.rsa.n_sg; in ccp_rsa_crypt() 84 rctx->cmd.u.rsa.mod_len = ctx->u.rsa.n_len; in ccp_rsa_crypt() 85 rctx->cmd.u.rsa.src = req->src; in ccp_rsa_crypt() [all …]
|
D | ccp-ops.c | 1837 struct ccp_rsa_engine *rsa = &cmd->u.rsa; in ccp_run_rsa_cmd() local 1844 if (rsa->key_size > cmd_q->ccp->vdata->rsamax) in ccp_run_rsa_cmd() 1847 if (!rsa->exp || !rsa->mod || !rsa->src || !rsa->dst) in ccp_run_rsa_cmd() 1862 o_len = 32 * ((rsa->key_size + 255) / 256); in ccp_run_rsa_cmd() 1890 ret = ccp_reverse_set_dm_area(&exp, 0, rsa->exp, 0, rsa->exp_len); in ccp_run_rsa_cmd() 1919 ret = ccp_reverse_set_dm_area(&src, 0, rsa->mod, 0, rsa->mod_len); in ccp_run_rsa_cmd() 1922 ret = ccp_reverse_set_dm_area(&src, o_len, rsa->src, 0, rsa->src_len); in ccp_run_rsa_cmd() 1939 op.u.rsa.mod_size = rsa->key_size; in ccp_run_rsa_cmd() 1940 op.u.rsa.input_len = i_len; in ccp_run_rsa_cmd() 1942 ret = cmd_q->ccp->vdata->perform->rsa(&op); in ccp_run_rsa_cmd() [all …]
|
D | Makefile | 25 ccp-crypto-rsa.o \
|
D | ccp-dev.h | 551 struct ccp_rsa_op rsa; member 657 int (*rsa)(struct ccp_op *); member
|
D | ccp-dev-v3.c | 238 | (op->u.rsa.mod_size << REQ1_RSA_MOD_SIZE_SHIFT) in ccp_perform_rsa() 241 cr[1] = op->u.rsa.input_len - 1; in ccp_perform_rsa() 572 .rsa = ccp_perform_rsa,
|
D | ccp-dev-v5.c | 119 } rsa; member 149 #define CCP_RSA_SIZE(p) ((p)->rsa.size) 468 CCP_RSA_SIZE(&function) = (op->u.rsa.mod_size + 7) >> 3; in ccp5_perform_rsa() 471 CCP5_CMD_LEN(&desc) = op->u.rsa.input_len; in ccp5_perform_rsa() 1103 .rsa = ccp5_perform_rsa,
|
D | ccp-crypto.h | 267 struct ccp_rsa_ctx rsa; member
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_ipsec.c | 396 struct rx_sa *rsa; in ixgbe_ipsec_find_rx_state() local 400 hash_for_each_possible_rcu(ipsec->rx_sa_list, rsa, hlist, in ixgbe_ipsec_find_rx_state() 402 if (rsa->mode & IXGBE_RXTXMOD_VF) in ixgbe_ipsec_find_rx_state() 404 if (spi == rsa->xs->id.spi && in ixgbe_ipsec_find_rx_state() 405 ((ip4 && *daddr == rsa->xs->id.daddr.a4) || in ixgbe_ipsec_find_rx_state() 406 (!ip4 && !memcmp(daddr, &rsa->xs->id.daddr.a6, in ixgbe_ipsec_find_rx_state() 407 sizeof(rsa->xs->id.daddr.a6)))) && in ixgbe_ipsec_find_rx_state() 408 proto == rsa->xs->id.proto) { in ixgbe_ipsec_find_rx_state() 409 ret = rsa->xs; in ixgbe_ipsec_find_rx_state() 595 struct rx_sa rsa; in ixgbe_ipsec_add_sa() local [all …]
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbevf/ |
D | ipsec.c | 183 struct rx_sa *rsa; in ixgbevf_ipsec_find_rx_state() local 186 hash_for_each_possible_rcu(ipsec->rx_sa_list, rsa, hlist, in ixgbevf_ipsec_find_rx_state() 188 if (spi == rsa->xs->id.spi && in ixgbevf_ipsec_find_rx_state() 189 ((ip4 && *daddr == rsa->xs->id.daddr.a4) || in ixgbevf_ipsec_find_rx_state() 190 (!ip4 && !memcmp(daddr, &rsa->xs->id.daddr.a6, in ixgbevf_ipsec_find_rx_state() 191 sizeof(rsa->xs->id.daddr.a6)))) && in ixgbevf_ipsec_find_rx_state() 192 proto == rsa->xs->id.proto) { in ixgbevf_ipsec_find_rx_state() 193 ret = rsa->xs; in ixgbevf_ipsec_find_rx_state() 290 struct rx_sa rsa; in ixgbevf_ipsec_add_sa() local 305 memset(&rsa, 0, sizeof(rsa)); in ixgbevf_ipsec_add_sa() [all …]
|
/linux-6.12.1/drivers/crypto/intel/qat/qat_common/ |
D | qat_asym_algs.c | 119 struct qat_rsa_input_params rsa; member 123 struct qat_rsa_output_params rsa; member 132 struct qat_rsa_ctx *rsa; member 136 struct akcipher_request *rsa; member 569 struct akcipher_request *areq = req->areq.rsa; in qat_rsa_cb() 570 struct device *dev = &GET_DEV(req->ctx.rsa->inst->accel_dev); in qat_rsa_cb() 576 dma_unmap_single(dev, req->in.rsa.enc.m, req->ctx.rsa->key_sz, in qat_rsa_cb() 581 areq->dst_len = req->ctx.rsa->key_sz; in qat_rsa_cb() 582 dma_unmap_single(dev, req->out.rsa.enc.c, req->ctx.rsa->key_sz, in qat_rsa_cb() 730 qat_req->ctx.rsa = ctx; in qat_rsa_enc() [all …]
|
/linux-6.12.1/drivers/crypto/hisilicon/hpre/ |
D | hpre_crypto.c | 127 struct hpre_rsa_ctx rsa; member 142 struct akcipher_request *rsa; member 470 areq = req->areq.rsa; in hpre_rsa_cb() 535 h_req->areq.rsa = akreq; in hpre_msg_request_set() 810 akcipher_request_set_tfm(req, ctx->rsa.soft_tfm); in hpre_rsa_enc() 816 if (unlikely(!ctx->rsa.pubkey)) in hpre_rsa_enc() 824 msg->key = cpu_to_le64(ctx->rsa.dma_pubkey); in hpre_rsa_enc() 858 akcipher_request_set_tfm(req, ctx->rsa.soft_tfm); in hpre_rsa_dec() 864 if (unlikely(!ctx->rsa.prikey)) in hpre_rsa_dec() 872 msg->key = cpu_to_le64(ctx->rsa.dma_crt_prikey); in hpre_rsa_dec() [all …]
|
/linux-6.12.1/drivers/gpu/drm/i915/gt/uc/ |
D | intel_guc_fw.c | 54 u32 rsa[UOS_RSA_SCRATCH_COUNT]; in guc_xfer_rsa_mmio() local 58 copied = intel_uc_fw_copy_rsa(guc_fw, rsa, sizeof(rsa)); in guc_xfer_rsa_mmio() 59 if (copied < sizeof(rsa)) in guc_xfer_rsa_mmio() 63 intel_uncore_write(uncore, UOS_RSA_SCRATCH(i), rsa[i]); in guc_xfer_rsa_mmio()
|
/linux-6.12.1/crypto/ |
D | rsa.c | 385 static struct akcipher_alg rsa = { variable 405 err = crypto_register_akcipher(&rsa); in rsa_init() 411 crypto_unregister_akcipher(&rsa); in rsa_init() 421 crypto_unregister_akcipher(&rsa); in rsa_exit()
|
D | Makefile | 48 rsa_generic-y += rsa.o 50 rsa_generic-y += rsa-pkcs1pad.o
|
/linux-6.12.1/drivers/crypto/starfive/ |
D | Makefile | 4 jh7110-crypto-objs := jh7110-cryp.o jh7110-hash.o jh7110-rsa.o jh7110-aes.o
|
/linux-6.12.1/drivers/gpu/drm/xe/ |
D | xe_guc.c | 490 u32 rsa[UOS_RSA_SCRATCH_COUNT]; in guc_xfer_rsa() local 501 copied = xe_uc_fw_copy_rsa(&guc->fw, rsa, sizeof(rsa)); in guc_xfer_rsa() 502 if (copied < sizeof(rsa)) in guc_xfer_rsa() 506 xe_mmio_write32(gt, UOS_RSA_SCRATCH(i), rsa[i]); in guc_xfer_rsa()
|
/linux-6.12.1/arch/s390/include/asm/ |
D | kvm_host.h | 1032 int __sie64a(phys_addr_t sie_block_phys, struct kvm_s390_sie_block *sie_block, u64 *rsa, 1035 static inline int sie64a(struct kvm_s390_sie_block *sie_block, u64 *rsa, unsigned long gasce) in sie64a() argument 1037 return __sie64a(virt_to_phys(sie_block), sie_block, rsa, gasce); in sie64a()
|
/linux-6.12.1/include/linux/ |
D | ccp.h | 656 struct ccp_rsa_engine rsa; member
|
/linux-6.12.1/drivers/crypto/virtio/ |
D | virtio_crypto_akcipher_algs.c | 423 para.u.rsa.padding_algo = cpu_to_le32(padding_algo); in virtio_crypto_rsa_set_key() 424 para.u.rsa.hash_algo = cpu_to_le32(hash_algo); in virtio_crypto_rsa_set_key()
|
/linux-6.12.1/include/uapi/linux/ |
D | virtio_crypto.h | 230 struct virtio_crypto_rsa_session_para rsa; member
|
/linux-6.12.1/drivers/clk/meson/ |
D | a1-peripherals.c | 1856 static MESON_GATE(rsa, SYS_CLK_EN1, 8); 1920 [CLKID_RSA] = &rsa.hw, 2078 &rsa,
|
/linux-6.12.1/Documentation/arch/powerpc/ |
D | hvcs.rst | 14 Author(s): Ryan S. Arnold <rsa@us.ibm.com>
|