/linux-6.12.1/drivers/dma/dw-axi-dmac/ |
D | dw-axi-dmac-platform.c | 300 desc->hw_desc = kcalloc(num, sizeof(*desc->hw_desc), GFP_NOWAIT); in axi_desc_alloc() 301 if (!desc->hw_desc) { in axi_desc_alloc() 333 struct axi_dma_hw_desc *hw_desc; in axi_desc_put() local 337 hw_desc = &desc->hw_desc[descs_put]; in axi_desc_put() 338 dma_pool_free(chan->desc_pool, hw_desc->lli, hw_desc->llp); in axi_desc_put() 341 kfree(desc->hw_desc); in axi_desc_put() 378 len = vd_to_axi_desc(vdesc)->hw_desc[0].len; in dma_chan_tx_status() 471 write_chan_llp(chan, first->hw_desc[0].llp | lms); in axi_chan_block_xfer_start() 634 static void set_desc_dest_master(struct axi_dma_hw_desc *hw_desc, in set_desc_dest_master() argument 640 val = le32_to_cpu(hw_desc->lli->ctl_lo); in set_desc_dest_master() [all …]
|
D | dw-axi-dmac.h | 100 struct axi_dma_hw_desc *hw_desc; member
|
/linux-6.12.1/drivers/crypto/ |
D | sahara.c | 196 struct sahara_hw_desc *hw_desc[SAHARA_MAX_HW_DESC]; member 403 dev_dbg(dev->device, "\thdr = 0x%08x\n", dev->hw_desc[i]->hdr); in sahara_dump_descriptors() 404 dev_dbg(dev->device, "\tlen1 = %u\n", dev->hw_desc[i]->len1); in sahara_dump_descriptors() 405 dev_dbg(dev->device, "\tp1 = 0x%08x\n", dev->hw_desc[i]->p1); in sahara_dump_descriptors() 406 dev_dbg(dev->device, "\tlen2 = %u\n", dev->hw_desc[i]->len2); in sahara_dump_descriptors() 407 dev_dbg(dev->device, "\tp2 = 0x%08x\n", dev->hw_desc[i]->p2); in sahara_dump_descriptors() 409 dev->hw_desc[i]->next); in sahara_dump_descriptors() 444 dev->hw_desc[idx]->len1 = AES_BLOCK_SIZE; in sahara_hw_descriptor_create() 445 dev->hw_desc[idx]->p1 = dev->iv_phys_base; in sahara_hw_descriptor_create() 447 dev->hw_desc[idx]->len1 = 0; in sahara_hw_descriptor_create() [all …]
|
/linux-6.12.1/drivers/dma/ppc4xx/ |
D | adma.c | 177 print_cb(chan, iter->hw_desc); in print_cb_list() 238 p = desc->hw_desc; in ppc440spe_desc_init_interrupt() 239 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_interrupt() 245 memset(desc->hw_desc, 0, sizeof(struct dma_cdb)); in ppc440spe_desc_init_interrupt() 262 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_null_xor() 274 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_xor() local 276 memset(desc->hw_desc, 0, sizeof(struct xor_cb)); in ppc440spe_desc_init_xor() 281 hw_desc->cbc = XOR_CBCR_TGT_BIT | src_cnt; in ppc440spe_desc_init_xor() 284 hw_desc->cbc |= XOR_CBCR_CBCE_BIT; in ppc440spe_desc_init_xor() 294 struct xor_cb *hw_desc = desc->hw_desc; in ppc440spe_desc_init_dma2pq() local [all …]
|
D | adma.h | 152 void *hw_desc; member
|
/linux-6.12.1/drivers/dma/xilinx/ |
D | xilinx_dpdma.c | 511 struct xilinx_dpdma_hw_desc *hw_desc = &sw_desc->hw; in xilinx_dpdma_sw_desc_set_dma_addrs() local 514 hw_desc->src_addr = lower_32_bits(dma_addr[0]); in xilinx_dpdma_sw_desc_set_dma_addrs() 516 hw_desc->addr_ext |= in xilinx_dpdma_sw_desc_set_dma_addrs() 521 u32 *addr = &hw_desc->src_addr2; in xilinx_dpdma_sw_desc_set_dma_addrs() 526 u32 *addr_ext = &hw_desc->addr_ext_23; in xilinx_dpdma_sw_desc_set_dma_addrs() 600 struct xilinx_dpdma_hw_desc *hw_desc = &sw_desc->hw; in xilinx_dpdma_chan_dump_tx_desc() local 604 dev_dbg(dev, "control: 0x%08x\n", hw_desc->control); in xilinx_dpdma_chan_dump_tx_desc() 605 dev_dbg(dev, "desc_id: 0x%08x\n", hw_desc->desc_id); in xilinx_dpdma_chan_dump_tx_desc() 606 dev_dbg(dev, "xfer_size: 0x%08x\n", hw_desc->xfer_size); in xilinx_dpdma_chan_dump_tx_desc() 607 dev_dbg(dev, "hsize_stride: 0x%08x\n", hw_desc->hsize_stride); in xilinx_dpdma_chan_dump_tx_desc() [all …]
|
/linux-6.12.1/include/linux/platform_data/ |
D | dma-iop32x.h | 86 void *hw_desc; member 107 container_of(addr_hw_desc, struct iop_adma_desc_slot, hw_desc) 108 #define iop_hw_desc_slot_idx(hw_desc, idx) \ argument 109 ( (void *) (((unsigned long) hw_desc) + ((idx) << 5)) )
|
/linux-6.12.1/drivers/dma/ |
D | pxa_dma.c | 93 struct pxad_desc_hw *hw_desc[] __counted_by(nb_desc); 520 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in set_updater_desc() 521 dma_addr_t dma = sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr; in set_updater_desc() 531 sw_desc->hw_desc[sw_desc->nb_desc - 2]->ddadr = sw_desc->first; in set_updater_desc() 538 sw_desc->hw_desc[sw_desc->nb_desc - 1]; in is_desc_completed() 551 desc1->hw_desc[desc1->nb_desc - 1]->ddadr = dma_to_chain; in pxad_desc_chain() 727 dma = sw_desc->hw_desc[i - 1]->ddadr; in pxad_free_desc() 731 sw_desc->hw_desc[i], dma); in pxad_free_desc() 745 sw_desc = kzalloc(struct_size(sw_desc, hw_desc, nb_hw_desc), in pxad_alloc_desc() 761 sw_desc->hw_desc[i] = desc; in pxad_alloc_desc() [all …]
|
D | mv_xor.c | 52 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_init() local 54 hw_desc->status = XOR_DESC_DMA_OWNED; in mv_desc_init() 55 hw_desc->phy_next_desc = 0; in mv_desc_init() 57 hw_desc->desc_command = (flags & DMA_PREP_INTERRUPT) ? in mv_desc_init() 59 hw_desc->phy_dest_addr = addr; in mv_desc_init() 60 hw_desc->byte_count = byte_count; in mv_desc_init() 65 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_mode() local 70 hw_desc->desc_command |= XOR_DESC_OPERATION_XOR; in mv_desc_set_mode() 73 hw_desc->desc_command |= XOR_DESC_OPERATION_MEMCPY; in mv_desc_set_mode() 84 struct mv_xor_desc *hw_desc = desc->hw_desc; in mv_desc_set_next_desc() local [all …]
|
D | mv_xor.h | 145 void *hw_desc; member 188 container_of(addr_hw_desc, struct mv_xor_desc_slot, hw_desc) 190 #define mv_hw_desc_slot_idx(hw_desc, idx) \ argument 191 ((void *)(((unsigned long)hw_desc) + ((idx) << 5)))
|
D | altera-msgdma.c | 162 struct msgdma_extended_desc hw_desc; member 358 desc = &new->hw_desc; in msgdma_prep_memcpy() 423 desc = &new->hw_desc; in msgdma_prep_slave_sg() 499 void __iomem *hw_desc = mdev->desc; in msgdma_copy_one() local 518 memcpy((void __force *)hw_desc, &desc->hw_desc, in msgdma_copy_one() 519 sizeof(desc->hw_desc) - sizeof(u32)); in msgdma_copy_one() 524 iowrite32(desc->hw_desc.control, hw_desc + in msgdma_copy_one()
|
D | mv_xor_v2.c | 184 struct mv_xor_v2_descriptor hw_desc; member 323 memcpy(dest_hw_desc, &sw_desc->hw_desc, xor_dev->desc_size); in mv_xor_v2_tx_submit() 398 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_memcpy() 454 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_xor() 502 hw_descriptor = &sw_desc->hw_desc; in mv_xor_v2_prep_dma_interrupt()
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep_vf/ |
D | octep_vf_main.c | 602 struct octep_vf_tx_desc_hw *hw_desc; in octep_vf_start_xmit() local 626 hw_desc = &iq->desc_ring[wi]; in octep_vf_start_xmit() 627 hw_desc->ih64 = 0; in octep_vf_start_xmit() 632 ih = &hw_desc->ih; in octep_vf_start_xmit() 644 hw_desc->dptr = tx_buffer->dma; in octep_vf_start_xmit() 680 hw_desc->dptr = tx_buffer->sglist_dma; in octep_vf_start_xmit() 684 hw_desc->txm.ol_flags = OCTEP_VF_TX_OFFLOAD_CKSUM; in octep_vf_start_xmit() 685 hw_desc->txm.ol_flags |= OCTEP_VF_TX_OFFLOAD_TSO; in octep_vf_start_xmit() 686 hw_desc->txm.gso_size = skb_shinfo(skb)->gso_size; in octep_vf_start_xmit() 687 hw_desc->txm.gso_segs = skb_shinfo(skb)->gso_segs; in octep_vf_start_xmit() [all …]
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep/ |
D | octep_main.c | 848 struct octep_tx_desc_hw *hw_desc; in octep_start_xmit() local 872 hw_desc = &iq->desc_ring[wi]; in octep_start_xmit() 873 hw_desc->ih64 = 0; in octep_start_xmit() 878 ih = &hw_desc->ih; in octep_start_xmit() 889 hw_desc->dptr = tx_buffer->dma; in octep_start_xmit() 925 hw_desc->dptr = tx_buffer->sglist_dma; in octep_start_xmit() 930 hw_desc->txm.ol_flags = OCTEP_TX_OFFLOAD_CKSUM; in octep_start_xmit() 931 hw_desc->txm.ol_flags |= OCTEP_TX_OFFLOAD_TSO; in octep_start_xmit() 932 hw_desc->txm.gso_size = skb_shinfo(skb)->gso_size; in octep_start_xmit() 933 hw_desc->txm.gso_segs = skb_shinfo(skb)->gso_segs; in octep_start_xmit() [all …]
|
/linux-6.12.1/drivers/crypto/caam/ |
D | caamhash.c | 547 u32 hw_desc[DESC_JOB_IO_LEN_MAX / sizeof(u32)] ____cacheline_aligned; member 718 init_job_desc_shared(edesc->hw_desc, sh_desc_dma, desc_len(sh_desc), in ahash_edesc_alloc() 754 append_seq_in_ptr(edesc->hw_desc, src_dma, first_bytes + to_hash, in ahash_edesc_add_src() 766 u32 *desc = state->edesc->hw_desc; in ahash_do_one_req() 795 u32 *desc = edesc->hw_desc; in ahash_enqueue_req() 906 desc = edesc->hw_desc; in ahash_update_ctx() 967 desc = edesc->hw_desc; in ahash_final_ctx() 1047 desc = edesc->hw_desc; in ahash_finup_ctx() 1129 desc = edesc->hw_desc; in ahash_digest() 1166 desc = edesc->hw_desc; in ahash_final_no_ctx() [all …]
|
D | caamalg.c | 906 u32 hw_desc[]; member 933 u32 hw_desc[]; member 1078 u32 *desc = edesc->hw_desc; in init_aead_job() 1139 u32 *desc = edesc->hw_desc; in init_gcm_job() 1169 u32 *desc = edesc->hw_desc; in init_chachapoly_job() 1210 u32 *desc = edesc->hw_desc; in init_authenc_job() 1257 u32 *desc = edesc->hw_desc; in init_skcipher_job() 1450 u32 *desc = edesc->hw_desc; in aead_enqueue_req() 1486 desc = edesc->hw_desc; in chachapoly_crypt() 1524 DUMP_PREFIX_ADDRESS, 16, 4, edesc->hw_desc, in aead_crypt() [all …]
|
D | caampkc.h | 146 u32 hw_desc[]; member
|
D | caampkc.c | 383 u32 *desc = req_ctx->edesc->hw_desc; in akcipher_do_one_req() 678 u32 *desc = edesc->hw_desc; in akcipher_enqueue_req() 743 init_rsa_pub_desc(edesc->hw_desc, &edesc->pdb.pub); in caam_rsa_enc() 772 init_rsa_priv_f1_desc(edesc->hw_desc, &edesc->pdb.priv_f1); in caam_rsa_dec_priv_f1() 801 init_rsa_priv_f2_desc(edesc->hw_desc, &edesc->pdb.priv_f2); in caam_rsa_dec_priv_f2() 830 init_rsa_priv_f3_desc(edesc->hw_desc, &edesc->pdb.priv_f3); in caam_rsa_dec_priv_f3()
|