Lines Matching refs:inbuf

51 	struct tegra_se_datbuf inbuf;  member
581 cpuvaddr[i++] = lower_32_bits(rctx->inbuf.addr); in tegra_gmac_prep_cmd()
582 cpuvaddr[i++] = SE_ADDR_HI_MSB(upper_32_bits(rctx->inbuf.addr)) | in tegra_gmac_prep_cmd()
637 cpuvaddr[i++] = lower_32_bits(rctx->inbuf.addr); in tegra_gcm_crypt_prep_cmd()
638 cpuvaddr[i++] = SE_ADDR_HI_MSB(upper_32_bits(rctx->inbuf.addr)) | in tegra_gcm_crypt_prep_cmd()
713 scatterwalk_map_and_copy(rctx->inbuf.buf, in tegra_gcm_do_gmac()
730 scatterwalk_map_and_copy(rctx->inbuf.buf, rctx->src_sg, in tegra_gcm_do_crypt()
811 data_count = (rctx->inbuf.size / AES_BLOCK_SIZE) - 1; in tegra_cbcmac_prep_cmd()
820 cpuvaddr[i++] = lower_32_bits(rctx->inbuf.addr); in tegra_cbcmac_prep_cmd()
821 cpuvaddr[i++] = SE_ADDR_HI_MSB(upper_32_bits(rctx->inbuf.addr)) | in tegra_cbcmac_prep_cmd()
822 SE_ADDR_HI_SZ(rctx->inbuf.size); in tegra_cbcmac_prep_cmd()
852 cpuvaddr[i++] = (rctx->inbuf.size / AES_BLOCK_SIZE) - 1; in tegra_ctr_prep_cmd()
858 cpuvaddr[i++] = lower_32_bits(rctx->inbuf.addr); in tegra_ctr_prep_cmd()
859 cpuvaddr[i++] = SE_ADDR_HI_MSB(upper_32_bits(rctx->inbuf.addr)) | in tegra_ctr_prep_cmd()
860 SE_ADDR_HI_SZ(rctx->inbuf.size); in tegra_ctr_prep_cmd()
865 SE_ADDR_HI_SZ(rctx->inbuf.size); in tegra_ctr_prep_cmd()
981 memcpy(rctx->inbuf.buf, nonce, 16); in tegra_ccm_format_blocks()
986 memcpy(rctx->inbuf.buf + offset, adata, alen); in tegra_ccm_format_blocks()
989 scatterwalk_map_and_copy(rctx->inbuf.buf + offset, in tegra_ccm_format_blocks()
993 offset += tegra_ccm_add_padding(rctx->inbuf.buf + offset, in tegra_ccm_format_blocks()
1052 scatterwalk_map_and_copy(rctx->inbuf.buf + offset, in tegra_ccm_compute_auth()
1056 offset += tegra_ccm_add_padding(rctx->inbuf.buf + offset, rctx->cryptlen); in tegra_ccm_compute_auth()
1058 rctx->inbuf.size = offset; in tegra_ccm_compute_auth()
1080 memcpy(rctx->inbuf.buf, rctx->authdata, rctx->authsize); in tegra_ccm_do_ctr()
1082 scatterwalk_map_and_copy(rctx->inbuf.buf, sg, in tegra_ccm_do_ctr()
1087 offset += tegra_ccm_add_padding(rctx->inbuf.buf + offset, rctx->authsize); in tegra_ccm_do_ctr()
1091 scatterwalk_map_and_copy(rctx->inbuf.buf + offset, sg, in tegra_ccm_do_ctr()
1094 offset += tegra_ccm_add_padding(rctx->inbuf.buf + offset, rctx->cryptlen); in tegra_ccm_do_ctr()
1097 rctx->inbuf.size = offset; in tegra_ccm_do_ctr()
1149 rctx->inbuf.buf = dma_alloc_coherent(ctx->se->dev, SE_AES_BUFLEN, in tegra_ccm_do_one_req()
1150 &rctx->inbuf.addr, GFP_KERNEL); in tegra_ccm_do_one_req()
1151 if (!rctx->inbuf.buf) in tegra_ccm_do_one_req()
1154 rctx->inbuf.size = SE_AES_BUFLEN; in tegra_ccm_do_one_req()
1203 rctx->inbuf.buf, rctx->inbuf.addr); in tegra_ccm_do_one_req()
1219 rctx->inbuf.buf = dma_alloc_coherent(ctx->se->dev, SE_AES_BUFLEN, in tegra_gcm_do_one_req()
1220 &rctx->inbuf.addr, GFP_KERNEL); in tegra_gcm_do_one_req()
1221 if (!rctx->inbuf.buf) in tegra_gcm_do_one_req()
1224 rctx->inbuf.size = SE_AES_BUFLEN; in tegra_gcm_do_one_req()
1276 rctx->inbuf.buf, rctx->inbuf.addr); in tegra_gcm_do_one_req()