Home
last modified time | relevance | path

Searched refs:hw_desc (Results 1 – 18 of 18) sorted by relevance

/linux-6.12.1/drivers/dma/dw-axi-dmac/
Ddw-axi-dmac-platform.c300 desc->hw_desc = kcalloc(num, sizeof(*desc->hw_desc), GFP_NOWAIT); in axi_desc_alloc()
301 if (!desc->hw_desc) { in axi_desc_alloc()
333 struct axi_dma_hw_desc *hw_desc; in axi_desc_put() local
337 hw_desc = &desc->hw_desc[descs_put]; in axi_desc_put()
338 dma_pool_free(chan->desc_pool, hw_desc->lli, hw_desc->llp); in axi_desc_put()
341 kfree(desc->hw_desc); in axi_desc_put()
378 len = vd_to_axi_desc(vdesc)->hw_desc[0].len; in dma_chan_tx_status()
471 write_chan_llp(chan, first->hw_desc[0].llp | lms); in axi_chan_block_xfer_start()
634 static void set_desc_dest_master(struct axi_dma_hw_desc *hw_desc, in set_desc_dest_master() argument
640 val = le32_to_cpu(hw_desc->lli->ctl_lo); in set_desc_dest_master()
[all …]
Ddw-axi-dmac.h100 struct axi_dma_hw_desc *hw_desc; member
/linux-6.12.1/drivers/crypto/
Dsahara.c196 struct sahara_hw_desc *hw_desc[SAHARA_MAX_HW_DESC]; member
403 dev_dbg(dev->device, "\thdr = 0x%08x\n", dev->hw_desc[i]->hdr); in sahara_dump_descriptors()
404 dev_dbg(dev->device, "\tlen1 = %u\n", dev->hw_desc[i]->len1); in sahara_dump_descriptors()
405 dev_dbg(dev->device, "\tp1 = 0x%08x\n", dev->hw_desc[i]->p1); in sahara_dump_descriptors()
406 dev_dbg(dev->device, "\tlen2 = %u\n", dev->hw_desc[i]->len2); in sahara_dump_descriptors()
407 dev_dbg(dev->device, "\tp2 = 0x%08x\n", dev->hw_desc[i]->p2); in sahara_dump_descriptors()
409 dev->hw_desc[i]->next); in sahara_dump_descriptors()
444 dev->hw_desc[idx]->len1 = AES_BLOCK_SIZE; in sahara_hw_descriptor_create()
445 dev->hw_desc[idx]->p1 = dev->iv_phys_base; in sahara_hw_descriptor_create()
447 dev->hw_desc[idx]->len1 = 0; in sahara_hw_descriptor_create()
[all …]
/linux-6.12.1/drivers/dma/ppc4xx/
Dadma.c177 print_cb(chan, iter->hw_desc); in print_cb_list()
238 p = desc->hw_desc; in ppc440spe_desc_init_interrupt()
239 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_interrupt()
245 memset(desc->hw_desc, 0, sizeof(struct dma_cdb)); in ppc440spe_desc_init_interrupt()
262 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_null_xor()
274 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_xor() local
276 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_xor()
281 hw_desc->cbc = XOR_CBCR_TGT_BIT | src_cnt; in ppc440spe_desc_init_xor()
284 hw_desc->cbc |= XOR_CBCR_CBCE_BIT; in ppc440spe_desc_init_xor()
294 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_dma2pq() local
[all …]
Dadma.h152 void *hw_desc; member
/linux-6.12.1/drivers/dma/xilinx/
Dxilinx_dpdma.c511 struct xilinx_dpdma_hw_desc *hw_desc = &sw_desc->hw; in xilinx_dpdma_sw_desc_set_dma_addrs() local
514 hw_desc->src_addr = lower_32_bits(dma_addr[0]); in xilinx_dpdma_sw_desc_set_dma_addrs()
516 hw_desc->addr_ext |= in xilinx_dpdma_sw_desc_set_dma_addrs()
521 u32 *addr = &hw_desc->src_addr2; in xilinx_dpdma_sw_desc_set_dma_addrs()
526 u32 *addr_ext = &hw_desc->addr_ext_23; in xilinx_dpdma_sw_desc_set_dma_addrs()
600 struct xilinx_dpdma_hw_desc *hw_desc = &sw_desc->hw; in xilinx_dpdma_chan_dump_tx_desc() local
604 dev_dbg(dev, "control: 0x%08x\n", hw_desc->control); in xilinx_dpdma_chan_dump_tx_desc()
605 dev_dbg(dev, "desc_id: 0x%08x\n", hw_desc->desc_id); in xilinx_dpdma_chan_dump_tx_desc()
606 dev_dbg(dev, "xfer_size: 0x%08x\n", hw_desc->xfer_size); in xilinx_dpdma_chan_dump_tx_desc()
607 dev_dbg(dev, "hsize_stride: 0x%08x\n", hw_desc->hsize_stride); in xilinx_dpdma_chan_dump_tx_desc()
[all …]
/linux-6.12.1/include/linux/platform_data/
Ddma-iop32x.h86 void *hw_desc; member
107 container_of(addr_hw_desc, struct iop_adma_desc_slot, hw_desc)
108 #define iop_hw_desc_slot_idx(hw_desc, idx) \ argument
109 ( (void *) (((unsigned long) hw_desc) + ((idx) << 5)) )
/linux-6.12.1/drivers/dma/
Dpxa_dma.c93 struct pxad_desc_hw *hw_desc[] __counted_by(nb_desc);
520 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in set_updater_desc()
521 dma_addr_t dma = sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr; in set_updater_desc()
531 sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr = sw_desc->first; in set_updater_desc()
538 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in is_desc_completed()
551 desc1->hw_desc[desc1->nb_desc - 1]->ddadr = dma_to_chain; in pxad_desc_chain()
727 dma = sw_desc->hw_desc[i - 1]->ddadr; in pxad_free_desc()
731 sw_desc->hw_desc[i], dma); in pxad_free_desc()
745 sw_desc = kzalloc(struct_size(sw_desc, hw_desc, nb_hw_desc), in pxad_alloc_desc()
761 sw_desc->hw_desc[i] = desc; in pxad_alloc_desc()
[all …]
Dmv_xor.c52 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_init() local
54 hw_desc->status = XOR_DESC_DMA_OWNED; in mv_desc_init()
55 hw_desc->phy_next_desc = 0; in mv_desc_init()
57 hw_desc->desc_command = (flags & DMA_PREP_INTERRUPT) ? in mv_desc_init()
59 hw_desc->phy_dest_addr = addr; in mv_desc_init()
60 hw_desc->byte_count = byte_count; in mv_desc_init()
65 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_mode() local
70 hw_desc->desc_command |= XOR_DESC_OPERATION_XOR; in mv_desc_set_mode()
73 hw_desc->desc_command |= XOR_DESC_OPERATION_MEMCPY; in mv_desc_set_mode()
84 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_next_desc() local
[all …]
Dmv_xor.h145 void *hw_desc; member
188 container_of(addr_hw_desc, struct mv_xor_desc_slot, hw_desc)
190 #define mv_hw_desc_slot_idx(hw_desc, idx) \ argument
191 ((void *)(((unsigned long)hw_desc) + ((idx) << 5)))
Daltera-msgdma.c162 struct msgdma_extended_desc hw_desc; member
358 desc = &new->hw_desc; in msgdma_prep_memcpy()
423 desc = &new->hw_desc; in msgdma_prep_slave_sg()
499 void __iomem *hw_desc = mdev->desc; in msgdma_copy_one() local
518 memcpy((void __force *)hw_desc, &desc->hw_desc, in msgdma_copy_one()
519 sizeof(desc->hw_desc) - sizeof(u32)); in msgdma_copy_one()
524 iowrite32(desc->hw_desc.control, hw_desc + in msgdma_copy_one()
Dmv_xor_v2.c184 struct mv_xor_v2_descriptor hw_desc; member
323 memcpy(dest_hw_desc, &sw_desc->hw_desc, xor_dev->desc_size); in mv_xor_v2_tx_submit()
398 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_memcpy()
454 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_xor()
502 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_interrupt()
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep_vf/
Doctep_vf_main.c602 struct octep_vf_tx_desc_hw *hw_desc; in octep_vf_start_xmit() local
626 hw_desc = &iq->desc_ring[wi]; in octep_vf_start_xmit()
627 hw_desc->ih64 = 0; in octep_vf_start_xmit()
632 ih = &hw_desc->ih; in octep_vf_start_xmit()
644 hw_desc->dptr = tx_buffer->dma; in octep_vf_start_xmit()
680 hw_desc->dptr = tx_buffer->sglist_dma; in octep_vf_start_xmit()
684 hw_desc->txm.ol_flags = OCTEP_VF_TX_OFFLOAD_CKSUM; in octep_vf_start_xmit()
685 hw_desc->txm.ol_flags |= OCTEP_VF_TX_OFFLOAD_TSO; in octep_vf_start_xmit()
686 hw_desc->txm.gso_size = skb_shinfo(skb)->gso_size; in octep_vf_start_xmit()
687 hw_desc->txm.gso_segs = skb_shinfo(skb)->gso_segs; in octep_vf_start_xmit()
[all …]
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep/
Doctep_main.c848 struct octep_tx_desc_hw *hw_desc; in octep_start_xmit() local
872 hw_desc = &iq->desc_ring[wi]; in octep_start_xmit()
873 hw_desc->ih64 = 0; in octep_start_xmit()
878 ih = &hw_desc->ih; in octep_start_xmit()
889 hw_desc->dptr = tx_buffer->dma; in octep_start_xmit()
925 hw_desc->dptr = tx_buffer->sglist_dma; in octep_start_xmit()
930 hw_desc->txm.ol_flags = OCTEP_TX_OFFLOAD_CKSUM; in octep_start_xmit()
931 hw_desc->txm.ol_flags |= OCTEP_TX_OFFLOAD_TSO; in octep_start_xmit()
932 hw_desc->txm.gso_size = skb_shinfo(skb)->gso_size; in octep_start_xmit()
933 hw_desc->txm.gso_segs = skb_shinfo(skb)->gso_segs; in octep_start_xmit()
[all …]
/linux-6.12.1/drivers/crypto/caam/
Dcaamhash.c547 u32 hw_desc[DESC_JOB_IO_LEN_MAX / sizeof(u32)] ____cacheline_aligned; member
718 init_job_desc_shared(edesc->hw_desc, sh_desc_dma, desc_len(sh_desc), in ahash_edesc_alloc()
754 append_seq_in_ptr(edesc->hw_desc, src_dma, first_bytes + to_hash, in ahash_edesc_add_src()
766 u32 *desc = state->edesc->hw_desc; in ahash_do_one_req()
795 u32 *desc = edesc->hw_desc; in ahash_enqueue_req()
906 desc = edesc->hw_desc; in ahash_update_ctx()
967 desc = edesc->hw_desc; in ahash_final_ctx()
1047 desc = edesc->hw_desc; in ahash_finup_ctx()
1129 desc = edesc->hw_desc; in ahash_digest()
1166 desc = edesc->hw_desc; in ahash_final_no_ctx()
[all …]
Dcaamalg.c906 u32 hw_desc[]; member
933 u32 hw_desc[]; member
1078 u32 *desc = edesc->hw_desc; in init_aead_job()
1139 u32 *desc = edesc->hw_desc; in init_gcm_job()
1169 u32 *desc = edesc->hw_desc; in init_chachapoly_job()
1210 u32 *desc = edesc->hw_desc; in init_authenc_job()
1257 u32 *desc = edesc->hw_desc; in init_skcipher_job()
1450 u32 *desc = edesc->hw_desc; in aead_enqueue_req()
1486 desc = edesc->hw_desc; in chachapoly_crypt()
1524 DUMP_PREFIX_ADDRESS, 16, 4, edesc->hw_desc, in aead_crypt()
[all …]
Dcaampkc.h146 u32 hw_desc[]; member
Dcaampkc.c383 u32 *desc = req_ctx->edesc->hw_desc; in akcipher_do_one_req()
678 u32 *desc = edesc->hw_desc; in akcipher_enqueue_req()
743 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub); in caam_rsa_enc()
772 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1); in caam_rsa_dec_priv_f1()
801 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2); in caam_rsa_dec_priv_f2()
830 init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3); in caam_rsa_dec_priv_f3()