/linux-6.12.1/net/tls/ |
D | tls_strp.c | 58 skb_frag_size(frag))); in tls_strp_skb_copy() 59 offset += skb_frag_size(frag); in tls_strp_skb_copy() 224 chunk = min_t(size_t, len, PAGE_SIZE - skb_frag_size(frag)); in tls_strp_copyin_frag() 227 skb_frag_size(frag), in tls_strp_copyin_frag() 262 chunk = min_t(size_t, chunk, PAGE_SIZE - skb_frag_size(frag)); in tls_strp_copyin_frag() 265 skb_frag_size(frag), in tls_strp_copyin_frag()
|
D | tls_device_fallback.c | 283 skb_frag_size(frag), skb_frag_off(frag)); in fill_sg_in() 285 remaining -= skb_frag_size(frag); in fill_sg_in()
|
/linux-6.12.1/net/core/ |
D | tso.c | 57 tso->size = skb_frag_size(frag); in tso_build_data() 83 tso->size = skb_frag_size(frag); in tso_start()
|
D | skbuff.c | 1376 len -= skb_frag_size(frag); in skb_dump() 1383 skb_frag_size(frag), p, p_off, p_len, in skb_dump() 2012 skb_frag_foreach_page(f, skb_frag_off(f), skb_frag_size(f), in skb_copy_ubufs() 2718 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim() 2860 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() 2928 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() 3004 end = start + skb_frag_size(f); in skb_copy_bits() 3196 skb_frag_off(f), skb_frag_size(f), in __skb_splice_bits() 3311 if (offset < skb_frag_size(frag)) in __skb_send_sock() 3314 offset -= skb_frag_size(frag); in __skb_send_sock() [all …]
|
D | datagram.c | 420 end = start + skb_frag_size(frag); in __skb_datagram_iter() 578 end = start + skb_frag_size(frag); in skb_copy_datagram_from_iter() 671 start == skb_frag_off(last) + skb_frag_size(last)) { in zerocopy_fill_skb_from_iter()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | xdp.c | 139 len = skb_frag_size(frag); in mlx5e_xmit_xdp_buff() 475 tmp.len = skb_frag_size(frag); in mlx5e_xmit_xdp_frame_mpwqe() 632 dseg->byte_count = cpu_to_be32(skb_frag_size(frag)); in mlx5e_xmit_xdp_frame() 697 skb_frag_size(frag), DMA_TO_DEVICE); in mlx5e_free_xdpsq_desc() 896 skb_frag_size(frag), DMA_TO_DEVICE); in mlx5e_xdp_xmit() 903 skb_frag_size(&xdptxdf.sinfo->frags[j]), in mlx5e_xdp_xmit() 920 skb_frag_size(&xdptxdf.sinfo->frags[j]), in mlx5e_xdp_xmit()
|
/linux-6.12.1/drivers/net/ethernet/sfc/ |
D | tx_tso.c | 212 skb_frag_size(frag), DMA_TO_DEVICE); in tso_get_fragment() 214 st->unmap_len = skb_frag_size(frag); in tso_get_fragment() 215 st->in_len = skb_frag_size(frag); in tso_get_fragment()
|
/linux-6.12.1/drivers/net/ethernet/netronome/nfp/nfdk/ |
D | rings.c | 44 size = skb_frag_size(frag); in nfp_nfdk_tx_ring_reset() 46 skb_frag_size(frag), DMA_TO_DEVICE); in nfp_nfdk_tx_ring_reset()
|
D | dp.c | 135 n_descs += DIV_ROUND_UP(skb_frag_size(frag), in nfp_nfdk_tx_maybe_close_block() 373 dma_len = skb_frag_size(frag); in nfp_nfdk_tx() 450 skb_frag_size(frag), DMA_TO_DEVICE); in nfp_nfdk_tx() 521 size = skb_frag_size(frag); in nfp_nfdk_tx_complete() 525 skb_frag_size(frag), DMA_TO_DEVICE); in nfp_nfdk_tx_complete()
|
/linux-6.12.1/drivers/net/ethernet/intel/iavf/ |
D | iavf_txrx.c | 1837 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 1838 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 1839 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 1840 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 1841 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 1847 int stale_size = skb_frag_size(stale); in __iavf_chk_linearize() 1849 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 1989 size = skb_frag_size(frag); in iavf_tx_map()
|
D | iavf_txrx.h | 353 size = skb_frag_size(frag++); in iavf_xmit_descriptor_count()
|
/linux-6.12.1/net/ipv4/ |
D | tcp_sigpool.c | 351 sg_set_page(&sg, page, skb_frag_size(f), offset_in_page(offset)); in tcp_sigpool_hash_skb_data() 352 ahash_request_set_crypt(req, &sg, NULL, skb_frag_size(f)); in tcp_sigpool_hash_skb_data()
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_txrx.c | 1685 size = skb_frag_size(frag); in ice_tx_map() 2140 size = skb_frag_size(frag++); in ice_xmit_desc_count() 2184 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2185 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2186 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2187 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2188 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2194 int stale_size = skb_frag_size(stale); in __ice_chk_linearize() 2196 sum += skb_frag_size(frag++); in __ice_chk_linearize()
|
D | ice_xsk.c | 591 memcpy(addr, skb_frag_page(frag), skb_frag_size(frag)); in ice_construct_skb_zc() 594 addr, 0, skb_frag_size(frag)); in ice_construct_skb_zc() 732 size = skb_frag_size(&sinfo->frags[frag]); in ice_xmit_xdp_tx_zc()
|
/linux-6.12.1/drivers/net/ethernet/pensando/ionic/ |
D | ionic_txrx.c | 355 skb_frag_size(frag), in ionic_xdp_post_frame() 359 skb_frag_size(frag)); in ionic_xdp_post_frame() 366 bi->len = skb_frag_size(frag); in ionic_xdp_post_frame() 1128 dma_addr = ionic_tx_map_frag(q, frag, 0, skb_frag_size(frag)); in ionic_tx_map_skb() 1132 buf_info->len = skb_frag_size(frag); in ionic_tx_map_skb() 1743 frag_rem = skb_frag_size(frag); in ionic_tx_descs_needed()
|
/linux-6.12.1/drivers/net/ethernet/aeroflex/ |
D | greth.c | 113 skb_frag_size(&skb_shinfo(skb)->frags[i]), true); in greth_print_tx_packet() 203 skb_frag_size(frag), in greth_clean_rings() 526 status |= skb_frag_size(frag) & GRETH_BD_LEN; in greth_start_xmit_gbit() 540 dma_addr = skb_frag_dma_map(greth->dev, frag, 0, skb_frag_size(frag), in greth_start_xmit_gbit() 721 skb_frag_size(frag), in greth_clean_tx_gbit()
|
/linux-6.12.1/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 3514 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3515 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3516 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3517 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3518 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3524 int stale_size = skb_frag_size(stale); in __i40e_chk_linearize() 3526 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3648 size = skb_frag_size(frag); in i40e_tx_map() 3824 size = skb_frag_size(&sinfo->frags[i]); in i40e_xmit_xdp_ring()
|
D | i40e_xsk.c | 332 memcpy(addr, skb_frag_page(frag), skb_frag_size(frag)); in i40e_construct_skb_zc() 335 addr, 0, skb_frag_size(frag)); in i40e_construct_skb_zc()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 290 *++addr = skb_frag_dma_map(dev, fp, 0, skb_frag_size(fp), in map_skb() 299 dma_unmap_page(dev, *--addr, skb_frag_size(fp), DMA_TO_DEVICE); in map_skb() 917 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl() 933 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 934 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl() 939 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ktls_tx.c | 664 remaining -= skb_frag_size(frag); in tx_sync_info_get() 738 fsz = skb_frag_size(frag); in tx_post_resync_dump() 795 orig_fsz = skb_frag_size(f); in mlx5e_ktls_tx_handle_ooo()
|
/linux-6.12.1/drivers/net/ethernet/netronome/nfp/nfd3/ |
D | rings.c | 66 skb_frag_size(frag), DMA_TO_DEVICE); in nfp_nfd3_tx_ring_reset()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
D | chcr_ktls.c | 41 frag_size = skb_frag_size(frag); in chcr_get_nfrags_to_send() 46 frag_size = skb_frag_size(frag); in chcr_get_nfrags_to_send() 48 frag_size = min(len, skb_frag_size(frag) - start); in chcr_get_nfrags_to_send() 54 frag_size = min(len, skb_frag_size(&si->frags[frag_idx])); in chcr_get_nfrags_to_send() 1842 frag_size = skb_frag_size(&record->frags[i]); in chcr_short_record_handler() 1850 frag_delta = skb_frag_size(f) - remaining; in chcr_short_record_handler()
|
/linux-6.12.1/drivers/infiniband/hw/hfi1/ |
D | vnic_sdma.c | 70 skb_frag_size(frag), in build_vnic_ulp_payload()
|
/linux-6.12.1/drivers/net/ethernet/intel/idpf/ |
D | idpf_txrx.c | 2233 size = skb_frag_size(&shinfo->frags[i]); in idpf_tx_desc_count_required() 2485 size = skb_frag_size(frag); in idpf_tx_splitq_map() 2625 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2626 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2627 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2628 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2629 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2635 int stale_size = skb_frag_size(stale); in __idpf_chk_linearize() 2637 sum += skb_frag_size(frag++); in __idpf_chk_linearize()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 257 *++addr = skb_frag_dma_map(dev, fp, 0, skb_frag_size(fp), in cxgb4_map_skb() 266 dma_unmap_page(dev, *--addr, skb_frag_size(fp), DMA_TO_DEVICE); in cxgb4_map_skb() 285 dma_unmap_page(dev, *addr++, skb_frag_size(fp), DMA_TO_DEVICE); in unmap_skb() 839 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in cxgb4_write_sgl() 855 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl() 856 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in cxgb4_write_sgl() 861 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl() 914 frag_size = skb_frag_size(frag); in cxgb4_write_partial_sgl() 920 frag_size = skb_frag_size(frag); in cxgb4_write_partial_sgl() 923 frag_size = min(len, skb_frag_size(frag) - start); in cxgb4_write_partial_sgl() [all …]
|