Searched refs:tx_next (Results 1 – 9 of 9) sorted by relevance
/linux-6.12.1/drivers/net/ethernet/aeroflex/ |
D | greth.c | 324 greth->tx_next = 0; in greth_init_rings() 425 bdp = greth->tx_bd_base + greth->tx_next; in greth_start_xmit() 433 greth->tx_bufs_length[greth->tx_next] = skb->len & GRETH_BD_LEN; in greth_start_xmit() 436 if (greth->tx_next == GRETH_TXBD_NUM_MASK) { in greth_start_xmit() 440 greth->tx_next = NEXT_TX(greth->tx_next); in greth_start_xmit() 454 static inline u16 greth_num_free_bds(u16 tx_last, u16 tx_next) in greth_num_free_bds() argument 456 if (tx_next < tx_last) in greth_num_free_bds() 457 return (tx_last - tx_next) - 1; in greth_num_free_bds() 459 return GRETH_TXBD_NUM - (tx_next - tx_last) - 1; in greth_num_free_bds() 476 if (greth_num_free_bds(tx_last, greth->tx_next) < nr_frags + 1) { in greth_start_xmit_gbit() [all …]
|
D | greth.h | 109 u16 tx_next; member
|
/linux-6.12.1/drivers/net/wwan/t7xx/ |
D | t7xx_hif_cldma.h | 88 struct cldma_request *tx_next; member
|
D | t7xx_hif_cldma.c | 68 queue->tx_next = NULL; in md_cd_queue_struct_reset() 310 req = list_prev_entry_circular(queue->tx_next, &queue->tr_ring->gpd_ring, entry); in t7xx_cldma_txq_empty_hndl() 515 queue->tx_next = req; in t7xx_cldma_q_reset() 948 tx_req = queue->tx_next; in t7xx_cldma_send_skb() 954 queue->tx_next = list_next_entry_circular(tx_req, gpd_ring, entry); in t7xx_cldma_send_skb() 1140 t7xx_cldma_hw_set_start_addr(hw_info, qno_t, md_ctrl->txq[qno_t].tx_next->gpd_addr, in t7xx_cldma_resume_early()
|
/linux-6.12.1/drivers/net/ethernet/actions/ |
D | owl-emac.c | 699 unsigned int tx_next; in owl_emac_tx_complete() local 722 tx_next = ring->tail; in owl_emac_tx_complete() 724 while ((tx_next = owl_emac_ring_get_next(ring, tx_next)) != ring->head) { in owl_emac_tx_complete() 725 status = READ_ONCE(ring->descs[tx_next].status); in owl_emac_tx_complete()
|
/linux-6.12.1/drivers/net/ethernet/cavium/octeon/ |
D | octeon_mgmt.c | 127 unsigned int tx_next; member 988 p->tx_next = 0; in octeon_mgmt_open() 1309 p->tx_ring[p->tx_next] = re.d64; in octeon_mgmt_xmit() 1310 p->tx_next = (p->tx_next + 1) % OCTEON_MGMT_TX_RING_SIZE; in octeon_mgmt_xmit()
|
/linux-6.12.1/drivers/infiniband/hw/hfi1/ |
D | sdma.c | 2435 struct sdma_txreq *tx, *tx_next; in sdma_send_txlist() local 2443 list_for_each_entry_safe(tx, tx_next, tx_list, list) { in sdma_send_txlist() 2476 list_for_each_entry_safe(tx, tx_next, tx_list, list) { in sdma_send_txlist()
|
/linux-6.12.1/drivers/net/ethernet/marvell/ |
D | sky2.h | 2228 u16 tx_next; /* debug only */ member
|
D | sky2.c | 2025 sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); in sky2_tx_complete() 4429 for (idx = sky2->tx_next; idx != sky2->tx_prod && idx < sky2->tx_ring_size; in sky2_debug_show()
|