/linux-6.12.1/drivers/net/ethernet/intel/igc/ |
D | igc_dump.c | 144 dma_unmap_len(buffer_info, len), in igc_rings_dump() 195 dma_unmap_len(buffer_info, len), in igc_rings_dump() 204 dma_unmap_len(buffer_info, len), in igc_rings_dump()
|
/linux-6.12.1/include/net/libeth/ |
D | tx.h | 106 dma_unmap_len(sqe, len), DMA_TO_DEVICE); in libeth_tx_complete()
|
/linux-6.12.1/drivers/net/ethernet/google/gve/ |
D | gve_tx_dqo.c | 175 dma_unmap_len(cur_state, len[j]), in gve_tx_clean_pending_packets() 180 dma_unmap_len(cur_state, len[j]), in gve_tx_clean_pending_packets() 683 dma_unmap_len(pkt, len[i]), in gve_tx_add_skb_no_copy_dqo() 688 dma_unmap_len(pkt, len[i]), in gve_tx_add_skb_no_copy_dqo() 1046 dma_unmap_len(pkt, len[0]), DMA_TO_DEVICE); in gve_unmap_packet() 1049 dma_unmap_len(pkt, len[i]), DMA_TO_DEVICE); in gve_unmap_packet()
|
D | gve_tx.c | 436 dma_unmap_len(info, len), in gve_tx_unmap_buf() 441 dma_unmap_len(info, len), in gve_tx_unmap_buf()
|
/linux-6.12.1/drivers/net/ethernet/intel/iavf/ |
D | iavf_txrx.c | 37 if (dma_unmap_len(tx_buffer, len)) in iavf_unmap_and_free_tx_resource() 40 dma_unmap_len(tx_buffer, len), in iavf_unmap_and_free_tx_resource() 42 } else if (dma_unmap_len(tx_buffer, len)) { in iavf_unmap_and_free_tx_resource() 45 dma_unmap_len(tx_buffer, len), in iavf_unmap_and_free_tx_resource() 255 dma_unmap_len(tx_buf, len), in iavf_clean_tx_irq() 277 if (dma_unmap_len(tx_buf, len)) { in iavf_clean_tx_irq() 280 dma_unmap_len(tx_buf, len), in iavf_clean_tx_irq()
|
/linux-6.12.1/drivers/net/ethernet/arc/ |
D | emac_main.c | 145 dma_unmap_len(tx_buff, len), DMA_TO_DEVICE); in arc_emac_tx_clean() 243 dma_unmap_len(rx_buff, len), DMA_FROM_DEVICE); in arc_emac_rx() 564 dma_unmap_len(tx_buff, len), in arc_free_tx_queue() 596 dma_unmap_len(rx_buff, len), in arc_free_rx_queue()
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_txrx_lib.c | 248 dma_unmap_len(tx_buf, len), DMA_TO_DEVICE); in ice_clean_xdp_tx_buf() 434 dma_unmap_len(tx_buf, len), DMA_TO_DEVICE); in __ice_xmit_xdp_ring()
|
D | ice_txrx.c | 115 if (dma_unmap_len(tx_buf, len)) in ice_unmap_and_free_tx_buf() 118 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf() 266 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq() 286 if (dma_unmap_len(tx_buf, len)) { in ice_clean_tx_irq() 289 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq() 2503 if (dma_unmap_len(tx_buf, len)) in ice_clean_ctrl_tx_irq() 2506 dma_unmap_len(tx_buf, len), in ice_clean_ctrl_tx_irq()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 507 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in free_freelQ_buffers() 621 if (likely(dma_unmap_len(ce, dma_len))) { in free_cmdQ_buffers() 624 dma_unmap_len(ce, dma_len), in free_cmdQ_buffers() 1055 dma_unmap_len(ce, dma_len), in get_packet() 1060 dma_unmap_len(ce, dma_len), in get_packet() 1073 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in get_packet() 1097 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in unexpected_offload()
|
/linux-6.12.1/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_netdev.c | 149 if (dma_unmap_len(tx_buffer, len)) in fm10k_unmap_and_free_tx_resource() 152 dma_unmap_len(tx_buffer, len), in fm10k_unmap_and_free_tx_resource() 154 } else if (dma_unmap_len(tx_buffer, len)) { in fm10k_unmap_and_free_tx_resource() 157 dma_unmap_len(tx_buffer, len), in fm10k_unmap_and_free_tx_resource()
|
D | fm10k_main.c | 1222 dma_unmap_len(tx_buffer, len), in fm10k_clean_tx_irq() 1241 if (dma_unmap_len(tx_buffer, len)) { in fm10k_clean_tx_irq() 1244 dma_unmap_len(tx_buffer, len), in fm10k_clean_tx_irq()
|
/linux-6.12.1/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 760 if (dma_unmap_len(tx_buffer, len)) in i40e_unmap_and_free_tx_resource() 763 dma_unmap_len(tx_buffer, len), in i40e_unmap_and_free_tx_resource() 765 } else if (dma_unmap_len(tx_buffer, len)) { in i40e_unmap_and_free_tx_resource() 768 dma_unmap_len(tx_buffer, len), in i40e_unmap_and_free_tx_resource() 974 dma_unmap_len(tx_buf, len), in i40e_clean_tx_irq() 996 if (dma_unmap_len(tx_buf, len)) { in i40e_clean_tx_irq() 999 dma_unmap_len(tx_buf, len), in i40e_clean_tx_irq() 3846 if (dma_unmap_len(tx_bi, len)) in i40e_xmit_xdp_ring() 3849 dma_unmap_len(tx_bi, len), in i40e_xmit_xdp_ring()
|
D | i40e_xsk.c | 641 dma_unmap_len(tx_bi, len), DMA_TO_DEVICE); in i40e_clean_xdp_tx_buffer()
|
/linux-6.12.1/drivers/net/ethernet/wangxun/libwx/ |
D | wx_lib.c | 746 dma_unmap_len(tx_buffer, len), in wx_clean_tx_irq() 764 if (dma_unmap_len(tx_buffer, len)) { in wx_clean_tx_irq() 767 dma_unmap_len(tx_buffer, len), in wx_clean_tx_irq() 1048 if (dma_unmap_len(tx_buffer, len)) in wx_tx_map() 1051 dma_unmap_len(tx_buffer, len), in wx_tx_map() 2323 dma_unmap_len(tx_buffer, len), in wx_clean_tx_ring() 2342 if (dma_unmap_len(tx_buffer, len)) in wx_clean_tx_ring() 2345 dma_unmap_len(tx_buffer, len), in wx_clean_tx_ring()
|
/linux-6.12.1/drivers/net/ethernet/qlogic/ |
D | qla3xxx.c | 1948 dma_unmap_len(&tx_cb->map[0], maplen), DMA_TO_DEVICE); in ql_process_mac_tx_intr() 1954 dma_unmap_len(&tx_cb->map[i], maplen), in ql_process_mac_tx_intr() 2024 dma_unmap_len(lrg_buf_cb2, maplen), DMA_FROM_DEVICE); in ql_process_mac_rx_intr() 2069 dma_unmap_len(lrg_buf_cb2, maplen), DMA_FROM_DEVICE); in ql_process_macip_rx_intr() 2424 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map() 2432 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map() 2717 dma_unmap_len(lrg_buf_cb, maplen), in ql_free_large_buffers() 3635 dma_unmap_len(&tx_cb->map[0], maplen), in ql_reset_work() 3640 dma_unmap_len(&tx_cb->map[j], maplen), in ql_reset_work()
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf_main.c | 312 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq() 330 if (dma_unmap_len(tx_buffer, len)) { in ixgbevf_clean_tx_irq() 333 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq() 2422 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring() 2441 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_clean_tx_ring() 2444 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring() 4075 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map() 4078 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map() 4087 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map() 4090 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map()
|
/linux-6.12.1/include/linux/ |
D | dma-mapping.h | 631 #define dma_unmap_len(PTR, LEN_NAME) ((PTR)->LEN_NAME) macro 638 #define dma_unmap_len(PTR, LEN_NAME) (0) macro
|
/linux-6.12.1/drivers/net/ethernet/alacritech/ |
D | slicoss.c | 369 dma_unmap_len(buff, map_len), DMA_TO_DEVICE); in slic_xmit_complete() 588 dma_unmap_len(buff, map_len), in slic_handle_receive() 908 dma_unmap_len(buff, map_len), DMA_TO_DEVICE); in slic_free_tx_queue() 950 dma_unmap_len(buff, map_len), in slic_free_rx_queue()
|
/linux-6.12.1/drivers/net/ethernet/atheros/alx/ |
D | main.c | 57 if (dma_unmap_len(txb, size)) { in alx_free_txbuf() 60 dma_unmap_len(txb, size), in alx_free_txbuf() 244 dma_unmap_len(rxb, size), in alx_clean_rx_irq() 519 dma_unmap_len(cur_buf, size), in alx_free_rxring_buf()
|
/linux-6.12.1/drivers/net/ethernet/intel/igb/ |
D | igb_main.c | 368 dma_unmap_len(buffer_info, len), in igb_dump() 417 dma_unmap_len(buffer_info, len), in igb_dump() 426 dma_unmap_len(buffer_info, len), in igb_dump() 4904 dma_unmap_len(tx_buffer, len), in igb_clean_tx_ring() 4923 if (dma_unmap_len(tx_buffer, len)) in igb_clean_tx_ring() 4926 dma_unmap_len(tx_buffer, len), in igb_clean_tx_ring() 6305 if (dma_unmap_len(tx_buffer, len)) in igb_tx_map() 6308 dma_unmap_len(tx_buffer, len), in igb_tx_map() 6317 if (dma_unmap_len(tx_buffer, len)) in igb_tx_map() 6320 dma_unmap_len(tx_buffer, len), in igb_tx_map() [all …]
|
/linux-6.12.1/drivers/net/ethernet/marvell/ |
D | skge.c | 1002 dma_unmap_len(e, maplen), in skge_rx_clean() 2844 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_xmit_frame() 2848 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_xmit_frame() 2866 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_tx_unmap() 2869 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_tx_unmap() 3082 dma_unmap_len(e, maplen), in skge_rx_get() 3087 dma_unmap_len(e, maplen), in skge_rx_get() 3110 dma_unmap_len(&ee, maplen), DMA_FROM_DEVICE); in skge_rx_get()
|
/linux-6.12.1/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_xsk.c | 452 dma_unmap_len(tx_bi, len), DMA_TO_DEVICE); in ixgbe_clean_xdp_tx_buffer()
|
D | ixgbe_main.c | 559 dma_unmap_len(tx_buffer, len), in ixgbe_print_buffer() 677 if (dma_unmap_len(tx_buffer, len) > 0) { in ixgbe_dump() 694 dma_unmap_len(tx_buffer, len), in ixgbe_dump() 705 dma_unmap_len(tx_buffer, len), in ixgbe_dump() 1201 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_irq() 1219 if (dma_unmap_len(tx_buffer, len)) { in ixgbe_clean_tx_irq() 1222 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_irq() 6080 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_ring() 6099 if (dma_unmap_len(tx_buffer, len)) in ixgbe_clean_tx_ring() 6102 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_ring() [all …]
|
/linux-6.12.1/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_xdp.c | 143 dma_unmap_len(tx_buf, len), in bnxt_tx_int_xdp()
|
/linux-6.12.1/drivers/net/ethernet/alteon/ |
D | acenic.c | 2039 if (dma_unmap_len(info, maplen)) { in ace_tx_int() 2042 dma_unmap_len(info, maplen), in ace_tx_int() 2321 if (dma_unmap_len(info, maplen)) { in ace_close() 2334 dma_unmap_len(info, maplen), in ace_close()
|