/linux-6.12.1/drivers/net/ethernet/fungible/funeth/ |
D | funeth_tx.c | 349 unsigned int ndesc; in fun_start_xmit() local 357 ndesc = write_pkt_desc(skb, q, tls_len); in fun_start_xmit() 358 if (unlikely(!ndesc)) { in fun_start_xmit() 363 q->prod_cnt += ndesc; in fun_start_xmit() 425 unsigned int npkts = 0, nbytes = 0, ndesc = 0; in fun_txq_reclaim() local 448 ndesc += pkt_desc; in fun_txq_reclaim() 454 q->cons_cnt += ndesc; in fun_txq_reclaim() 486 unsigned int npkts = 0, ndesc = 0, head, reclaim_idx; in fun_xdpq_clean() local 505 ndesc += pkt_desc; in fun_xdpq_clean() 510 q->cons_cnt += ndesc; in fun_xdpq_clean() [all …]
|
/linux-6.12.1/drivers/net/ethernet/altera/ |
D | altera_sgdma.c | 13 struct sgdma_descrip __iomem *ndesc, 173 struct sgdma_descrip __iomem *ndesc = &descbase[1]; in sgdma_tx_buffer() local 180 ndesc, /* next descriptor */ in sgdma_tx_buffer() 181 sgdma_txphysaddr(priv, ndesc), in sgdma_tx_buffer() 294 struct sgdma_descrip __iomem *ndesc, in sgdma_setup_descrip() argument 305 u32 ctrl = csrrd8(ndesc, sgdma_descroffs(control)); in sgdma_setup_descrip() 307 csrwr8(ctrl, ndesc, sgdma_descroffs(control)); in sgdma_setup_descrip() 342 struct sgdma_descrip __iomem *ndesc = &descbase[1]; in sgdma_async_read() local 353 ndesc, /* next descriptor */ in sgdma_async_read() 354 sgdma_rxphysaddr(priv, ndesc), in sgdma_async_read()
|
/linux-6.12.1/drivers/net/ethernet/socionext/ |
D | sni_ave.c | 233 u32 ndesc; /* number of descriptor */ member 683 u32 proc_idx, done_idx, ndesc, cmdsts; in ave_tx_complete() local 690 ndesc = priv->tx.ndesc; in ave_tx_complete() 723 done_idx = (done_idx + 1) % ndesc; in ave_tx_complete() 750 u32 ndesc, cmdsts; in ave_rx_receive() local 754 ndesc = priv->rx.ndesc; in ave_rx_receive() 755 restpkt = ((proc_idx + ndesc - 1) - done_idx) % ndesc; in ave_rx_receive() 770 proc_idx = (proc_idx + 1) % ndesc; in ave_rx_receive() 794 proc_idx = (proc_idx + 1) % ndesc; in ave_rx_receive() 809 done_idx = (done_idx + 1) % ndesc; in ave_rx_receive() [all …]
|
/linux-6.12.1/drivers/net/wireless/mediatek/mt76/ |
D | dma.c | 193 Q_WRITE(q, ring_size, MT_DMA_RRO_EN | q->ndesc); in mt76_dma_sync_idx() 195 Q_WRITE(q, ring_size, q->ndesc); in mt76_dma_sync_idx() 203 if (!q || !q->ndesc) in __mt76_dma_queue_reset() 210 for (i = 0; i < q->ndesc; i++) in __mt76_dma_queue_reset() 278 q->head = (q->head + 1) % q->ndesc; in mt76_dma_add_rx_buf() 303 next = (q->head + 1) % q->ndesc; in mt76_dma_add_buf() 387 if (!q || !q->ndesc) in mt76_dma_tx_cleanup() 495 q->tail = (q->tail + 1) % q->ndesc; in mt76_dma_dequeue() 511 if (q->queued + 1 >= q->ndesc - 1) in mt76_dma_tx_queue_skb_raw() 590 if (q->queued + (tx_info.nbuf + 1) / 2 >= q->ndesc - 1) { in mt76_dma_tx_queue_skb() [all …]
|
D | usb.c | 446 q->tail = (q->tail + 1) % q->ndesc; in mt76u_get_next_rx_entry() 581 q->head = (q->head + 1) % q->ndesc; in mt76u_complete_rx() 648 for (i = 0; i < q->ndesc; i++) { in mt76u_submit_rx_buffers() 677 q->ndesc = MT_NUM_RX_ENTRIES; in mt76u_alloc_rx_queue() 680 for (i = 0; i < q->ndesc; i++) { in mt76u_alloc_rx_queue() 700 for (i = 0; i < q->ndesc; i++) { in mt76u_free_rx_queue() 731 for (j = 0; j < q->ndesc; j++) in mt76u_stop_rx() 745 for (j = 0; j < q->ndesc; j++) in mt76u_resume_rx() 864 if (q->queued == q->ndesc) in mt76u_tx_queue_skb() 879 q->head = (q->head + 1) % q->ndesc; in mt76u_tx_queue_skb() [all …]
|
D | sdio.c | 315 q->ndesc = MT76S_NUM_RX_ENTRIES; in mt76s_alloc_rx_queue() 338 q->ndesc = MT76S_NUM_TX_ENTRIES; in mt76s_alloc_tx_queue() 374 q->tail = (q->tail + 1) % q->ndesc; in mt76s_get_next_rx_entry() 529 if (q->queued == q->ndesc) in mt76s_tx_queue_skb() 543 q->head = (q->head + 1) % q->ndesc; in mt76s_tx_queue_skb() 562 if (q->queued == q->ndesc) { in mt76s_tx_queue_skb_raw() 574 q->head = (q->head + 1) % q->ndesc; in mt76s_tx_queue_skb_raw() 622 for (j = 0; j < q->ndesc; j++) { in mt76s_deinit()
|
D | wed.c | 102 if (!q || !q->ndesc) in mt76_wed_dma_setup() 142 q->head = q->ndesc - 1; in mt76_wed_dma_setup() 149 q->head = q->ndesc - 1; in mt76_wed_dma_setup()
|
D | sdio_txrx.c | 119 int index = (q->head + i) % q->ndesc; in mt76s_rx_run_queue() 133 if (q->queued + i + 1 == q->ndesc) in mt76s_rx_run_queue() 142 q->head = (q->head + i) % q->ndesc; in mt76s_rx_run_queue() 292 q->first = (q->first + 1) % q->ndesc; in mt76s_tx_run_queue()
|
D | debugfs.c | 83 queued = mt76_is_usb(dev) ? q->ndesc - q->queued : q->queued; in mt76_rx_queues_read()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
D | chcr_ktls.c | 844 u32 len, cpl = 0, ndesc, wr_len, wr_mid = 0; in chcr_ktls_xmit_tcb_cpls() local 852 ndesc = DIV_ROUND_UP(len, 64); in chcr_ktls_xmit_tcb_cpls() 854 credits = chcr_txq_avail(&q->q) - ndesc; in chcr_ktls_xmit_tcb_cpls() 922 ndesc = DIV_ROUND_UP(len, 64); in chcr_ktls_xmit_tcb_cpls() 923 chcr_txq_advance(&q->q, ndesc); in chcr_ktls_xmit_tcb_cpls() 924 cxgb4_ring_tx_db(tx_info->adap, &q->q, ndesc); in chcr_ktls_xmit_tcb_cpls() 995 unsigned int ndesc; in chcr_ktls_write_tcp_options() local 1016 ndesc = DIV_ROUND_UP(len16, 4); in chcr_ktls_write_tcp_options() 1018 credits = chcr_txq_avail(&q->q) - ndesc; in chcr_ktls_write_tcp_options() 1079 chcr_txq_advance(&q->q, ndesc); in chcr_ktls_write_tcp_options() [all …]
|
/linux-6.12.1/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
D | chcr_ipsec.c | 576 u16 ndesc; in ch_ipsec_crypto_wreq() local 585 ndesc = DIV_ROUND_UP(flits, 2); in ch_ipsec_crypto_wreq() 601 wr_mid = FW_CRYPTO_LOOKASIDE_WR_LEN16_V(ndesc); in ch_ipsec_crypto_wreq() 614 wr->req.ulptx.len = htonl(ndesc - 1); in ch_ipsec_crypto_wreq() 711 unsigned int last_desc, ndesc, flits = 0; in ch_ipsec_xmit() local 741 ndesc = flits_to_desc(flits); in ch_ipsec_xmit() 742 credits = txq_avail(&q->q) - ndesc; in ch_ipsec_xmit() 748 dev->name, qidx, credits, ndesc, txq_avail(&q->q), in ch_ipsec_xmit() 753 last_desc = q->q.pidx + ndesc - 1; in ch_ipsec_xmit() 791 txq_advance(&q->q, ndesc); in ch_ipsec_xmit() [all …]
|
/linux-6.12.1/drivers/net/ethernet/ti/ |
D | netcp_core.c | 108 static void get_pkt_info(dma_addr_t *buff, u32 *buff_len, dma_addr_t *ndesc, in get_pkt_info() argument 113 *ndesc = le32_to_cpu(desc->next_desc); in get_pkt_info() 150 static void set_pkt_info(dma_addr_t buff, u32 buff_len, u32 ndesc, in set_pkt_info() argument 155 desc->next_desc = cpu_to_le32(ndesc); in set_pkt_info() 580 struct knav_dma_desc *ndesc; in netcp_free_rx_desc_chain() local 582 unsigned int buf_len, dma_sz = sizeof(*ndesc); in netcp_free_rx_desc_chain() 589 ndesc = knav_pool_desc_unmap(netcp->rx_pool, dma_desc, dma_sz); in netcp_free_rx_desc_chain() 590 if (unlikely(!ndesc)) { in netcp_free_rx_desc_chain() 594 get_pkt_info(&dma_buf, &tmp, &dma_desc, ndesc); in netcp_free_rx_desc_chain() 598 buf_ptr = (void *)GET_SW_DATA0(ndesc); in netcp_free_rx_desc_chain() [all …]
|
/linux-6.12.1/drivers/usb/mon/ |
D | mon_bin.c | 113 unsigned int ndesc; /* Actual number of ISO descriptors */ member 451 struct urb *urb, unsigned int ndesc) in mon_bin_collate_isodesc() argument 458 while (ndesc-- != 0) { in mon_bin_collate_isodesc() 469 unsigned int offset, struct urb *urb, char ev_type, unsigned int ndesc) in mon_bin_get_isodesc() argument 475 while (ndesc-- != 0) { in mon_bin_get_isodesc() 498 unsigned int ndesc, lendesc; in mon_bin_event() local 516 ndesc = 0; in mon_bin_event() 518 ndesc = ISODESC_MAX; in mon_bin_event() 520 ndesc = urb->number_of_packets; in mon_bin_event() 523 length = mon_bin_collate_isodesc(rp, urb, ndesc); in mon_bin_event() [all …]
|
D | mon_text.c | 200 int i, ndesc; in mon_text_event() local 233 if ((ndesc = urb->number_of_packets) > ISODESC_MAX) in mon_text_event() 234 ndesc = ISODESC_MAX; in mon_text_event() 237 for (i = 0; i < ndesc; i++) { in mon_text_event() 591 int ndesc; /* Display this many */ in mon_text_read_isodesc() local 597 ndesc = ep->numdesc; in mon_text_read_isodesc() 598 if (ndesc > ISODESC_MAX) in mon_text_read_isodesc() 599 ndesc = ISODESC_MAX; in mon_text_read_isodesc() 600 if (ndesc < 0) in mon_text_read_isodesc() 601 ndesc = 0; in mon_text_read_isodesc() [all …]
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 1491 unsigned int last_desc, flits, ndesc; in cxgb4_eth_xmit() local 1549 ndesc = flits_to_desc(flits); in cxgb4_eth_xmit() 1550 credits = txq_avail(&q->q) - ndesc; in cxgb4_eth_xmit() 1566 last_desc = q->q.pidx + ndesc - 1; in cxgb4_eth_xmit() 1719 txq_advance(&q->q, ndesc); in cxgb4_eth_xmit() 1721 cxgb4_ring_tx_db(adap, &q->q, ndesc); in cxgb4_eth_xmit() 1812 unsigned int last_desc, flits, ndesc; in cxgb4_vf_eth_xmit() local 1856 ndesc = flits_to_desc(flits); in cxgb4_vf_eth_xmit() 1857 credits = txq_avail(&txq->q) - ndesc; in cxgb4_vf_eth_xmit() 1872 last_desc = txq->q.pidx + ndesc - 1; in cxgb4_vf_eth_xmit() [all …]
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 1087 static void write_wr_hdr_sgl(unsigned int ndesc, struct sk_buff *skb, in write_wr_hdr_sgl() argument 1105 if (likely(ndesc == 1)) { in write_wr_hdr_sgl() 1128 ndesc--; in write_wr_hdr_sgl() 1158 WARN_ON(ndesc != 0); in write_wr_hdr_sgl() 1179 struct sge_txq *q, unsigned int ndesc, in write_tx_pkt_wr() argument 1237 sgp = ndesc == 1 ? (struct sg_ent *)&d->flit[flits] : sgl; in write_tx_pkt_wr() 1240 write_wr_hdr_sgl(ndesc, skb, d, pidx, q, sgl, flits, sgl_flits, gen, in write_tx_pkt_wr() 1263 unsigned int ndesc, pidx, credits, gen, compl; in t3_eth_xmit() local 1288 ndesc = calc_tx_descs(skb); in t3_eth_xmit() 1290 if (unlikely(credits < ndesc)) { in t3_eth_xmit() [all …]
|
/linux-6.12.1/drivers/net/ethernet/mediatek/ |
D | airoha_eth.c | 743 int ndesc; member 1411 while (q->queued < q->ndesc - 1) { in airoha_qdma_fill_rx_queue() 1423 q->head = (q->head + 1) % q->ndesc; in airoha_qdma_fill_rx_queue() 1498 q->tail = (q->tail + 1) % q->ndesc; in airoha_qdma_rx_process() 1554 struct airoha_qdma *qdma, int ndesc) in airoha_qdma_init_rx_queue() argument 1571 q->ndesc = ndesc; in airoha_qdma_init_rx_queue() 1574 q->entry = devm_kzalloc(eth->dev, q->ndesc * sizeof(*q->entry), in airoha_qdma_init_rx_queue() 1587 q->desc = dmam_alloc_coherent(eth->dev, q->ndesc * sizeof(*q->desc), in airoha_qdma_init_rx_queue() 1597 FIELD_PREP(RX_RING_SIZE_MASK, ndesc)); in airoha_qdma_init_rx_queue() 1599 thr = clamp(ndesc >> 3, 1, 32); in airoha_qdma_init_rx_queue() [all …]
|
/linux-6.12.1/drivers/dma/ |
D | idma64.c | 195 static struct idma64_desc *idma64_alloc_desc(unsigned int ndesc) in idma64_alloc_desc() argument 203 desc->hw = kcalloc(ndesc, sizeof(*desc->hw), GFP_NOWAIT); in idma64_alloc_desc() 217 if (desc->ndesc) { in idma64_desc_free() 218 unsigned int i = desc->ndesc; in idma64_desc_free() 280 unsigned int i = desc->ndesc; in idma64_desc_fill() 320 desc->ndesc = i; in idma64_prep_slave_sg() 329 desc->ndesc = sg_len; in idma64_prep_slave_sg() 362 } while (++i < desc->ndesc); in idma64_active_desc_size()
|
D | idma64.h | 117 unsigned int ndesc; member
|
/linux-6.12.1/drivers/net/wireless/ath/ath9k/ |
D | init.c | 285 int nbuf, int ndesc, bool is_tx) in ath_descdma_setup() argument 292 name, nbuf, ndesc); in ath_descdma_setup() 308 dd->dd_desc_len = desc_len * nbuf * ndesc; in ath_descdma_setup() 348 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) { in ath_descdma_setup() 364 ds += (desc_len * ndesc); in ath_descdma_setup() 379 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) { in ath_descdma_setup() 395 ds += (desc_len * ndesc); in ath_descdma_setup()
|
/linux-6.12.1/drivers/infiniband/hw/qib/ |
D | qib_verbs.c | 143 u32 ndesc = 1; /* count the header */ in qib_count_sge() local 150 ndesc = 0; in qib_count_sge() 153 ndesc++; in qib_count_sge() 173 return ndesc; in qib_count_sge() 782 u32 ndesc; in qib_verbs_send_dma() local 817 ndesc = qib_count_sge(ss, len); in qib_verbs_send_dma() 818 if (ndesc >= ppd->sdma_descq_cnt) in qib_verbs_send_dma() 819 ndesc = 0; in qib_verbs_send_dma() 821 ndesc = 1; in qib_verbs_send_dma() 822 if (ndesc) { in qib_verbs_send_dma() [all …]
|
D | qib_user_sdma.c | 803 int *maxpkts, int *ndesc) in qib_user_sdma_queue_pkts() argument 1023 *ndesc += pkt->naddr; in qib_user_sdma_queue_pkts() 1409 int ndesc = 0; in qib_user_sdma_writev() local 1412 iov, dim, &list, &mxp, &ndesc); in qib_user_sdma_writev() 1425 if (qib_sdma_descq_freecnt(ppd) < ndesc) { in qib_user_sdma_writev()
|
/linux-6.12.1/drivers/infiniband/ulp/srp/ |
D | ib_srp.h | 347 unsigned int ndesc; member
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1162 unsigned int flits, ndesc; in t4vf_eth_xmit() local 1213 ndesc = flits_to_desc(flits); in t4vf_eth_xmit() 1214 credits = txq_avail(&txq->q) - ndesc; in t4vf_eth_xmit() 1364 ndesc, credits, txq->q.pidx, skb->len, ssi->nr_frags); in t4vf_eth_xmit() 1435 last_desc = tq->pidx + ndesc - 1; in t4vf_eth_xmit() 1446 txq_advance(&txq->q, ndesc); in t4vf_eth_xmit() 1448 ring_tx_db(adapter, &txq->q, ndesc); in t4vf_eth_xmit()
|
/linux-6.12.1/drivers/net/ethernet/atheros/ |
D | ag71xx.c | 1447 int i, ring_mask, ndesc, split; in ag71xx_fill_dma_desc() local 1451 ndesc = 0; in ag71xx_fill_dma_desc() 1460 i = (ring->curr + ndesc) & ring_mask; in ag71xx_fill_dma_desc() 1484 if (!ndesc) in ag71xx_fill_dma_desc() 1488 ndesc++; in ag71xx_fill_dma_desc() 1491 return ndesc; in ag71xx_fill_dma_desc()
|