/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/ |
D | alloc.c | 82 buf->frags = kcalloc(buf->npages, sizeof(struct mlx5_buf_list), in mlx5_frag_buf_alloc_node() 84 if (!buf->frags) in mlx5_frag_buf_alloc_node() 88 struct mlx5_buf_list *frag = &buf->frags[i]; in mlx5_frag_buf_alloc_node() 97 buf->frags[i].buf, buf->frags[i].map); in mlx5_frag_buf_alloc_node() 109 dma_free_coherent(mlx5_core_dma_dev(dev), PAGE_SIZE, buf->frags[i].buf, in mlx5_frag_buf_alloc_node() 110 buf->frags[i].map); in mlx5_frag_buf_alloc_node() 111 kfree(buf->frags); in mlx5_frag_buf_alloc_node() 125 dma_free_coherent(mlx5_core_dma_dev(dev), frag_sz, buf->frags[i].buf, in mlx5_frag_buf_free() 126 buf->frags[i].map); in mlx5_frag_buf_free() 129 kfree(buf->frags); in mlx5_frag_buf_free() [all …]
|
D | wq.c | 61 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_cyc_create() 128 mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc); in mlx5_wq_qp_create() 136 mlx5_init_fbc_offset(wq_ctrl->buf.frags, in mlx5_wq_qp_create() 142 mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages, in mlx5_wq_qp_create() 185 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc); in mlx5_cqwq_create() 234 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_ll_create()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 78 struct mlx4_en_rx_alloc *frags, in mlx4_en_alloc_frags() argument 83 for (i = 0; i < priv->num_frags; i++, frags++) { in mlx4_en_alloc_frags() 84 if (!frags->page) { in mlx4_en_alloc_frags() 85 if (mlx4_alloc_page(priv, frags, gfp)) { in mlx4_en_alloc_frags() 91 rx_desc->data[i].addr = cpu_to_be64(frags->dma + in mlx4_en_alloc_frags() 92 frags->page_offset); in mlx4_en_alloc_frags() 142 struct mlx4_en_rx_alloc *frags = ring->rx_info + in mlx4_en_prepare_rx_desc() local 146 if (!frags->page) { in mlx4_en_prepare_rx_desc() 148 frags->page = ring->page_cache.buf[ring->page_cache.index].page; in mlx4_en_prepare_rx_desc() 149 frags->dma = ring->page_cache.buf[ring->page_cache.index].dma; in mlx4_en_prepare_rx_desc() [all …]
|
/linux-6.12.1/tools/testing/selftests/bpf/ |
D | generate_udp_fragments.py | 43 def print_frags(f, frags, v6): argument 44 for idx, frag in enumerate(frags): 77 frags = [f.build() for f in pkt.fragment(24)] 81 print_frags(f, frags, False)
|
/linux-6.12.1/drivers/net/wireless/intel/iwlwifi/ |
D | iwl-dbg-tlv.c | 403 struct iwl_dram_data *frag = &fw_mon->frags[i]; in iwl_dbg_tlv_fragments_free() 413 kfree(fw_mon->frags); in iwl_dbg_tlv_fragments_free() 414 fw_mon->frags = NULL; in iwl_dbg_tlv_fragments_free() 596 memset(fw_mon->frags[i].block, 0, in iwl_dbg_tlv_alloc_fragments() 597 fw_mon->frags[i].size); in iwl_dbg_tlv_alloc_fragments() 621 fw_mon->frags = kcalloc(num_frags, sizeof(*fw_mon->frags), GFP_KERNEL); in iwl_dbg_tlv_alloc_fragments() 622 if (!fw_mon->frags) in iwl_dbg_tlv_alloc_fragments() 632 pages = iwl_dbg_tlv_alloc_fragment(fwrt, &fw_mon->frags[i], in iwl_dbg_tlv_alloc_fragments() 707 struct iwl_buf_alloc_frag *frag = &data.frags[j]; in iwl_dbg_tlv_apply_buffer() 709 &fw_mon->frags[fw_mon_idx++]; in iwl_dbg_tlv_apply_buffer() [all …]
|
/linux-6.12.1/net/core/ |
D | gro.c | 143 frag = pinfo->frags + nr_frags; in skb_gro_receive() 144 frag2 = skbinfo->frags + i; in skb_gro_receive() 164 skb_frag_t *frag = pinfo->frags + nr_frags; in skb_gro_receive() 180 memcpy(frag + 1, skbinfo->frags, sizeof(*frag) * skbinfo->nr_frags); in skb_gro_receive() 198 skb_frag_off_add(&skbinfo->frags[0], eat); in skb_gro_receive() 199 skb_frag_size_sub(&skbinfo->frags[0], eat); in skb_gro_receive() 404 frag0 = &pinfo->frags[0]; in skb_gro_reset_offset() 427 skb_frag_off_add(&pinfo->frags[0], grow); in gro_pull_from_frag0() 428 skb_frag_size_sub(&pinfo->frags[0], grow); in gro_pull_from_frag0() 430 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { in gro_pull_from_frag0() [all …]
|
D | net_test.c | 46 const unsigned int *frags; member 75 .frags = (const unsigned int[]) { GSO_TEST_SIZE, 1 }, 83 .frags = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 }, 92 .frags = (const unsigned int[]) { GSO_TEST_SIZE, 3 }, 174 skb_fill_page_desc(skb, i, page, pg_off, tcase->frags[i]); in gso_test_func() 175 pg_off += tcase->frags[i]; in gso_test_func()
|
D | skbuff.c | 894 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag() 1072 head_netmem = netmem_compound_head(shinfo->frags[i].netmem); in skb_pp_frag_ref() 1119 __skb_frag_unref(&shinfo->frags[i], skb->pp_recycle); in skb_release_data() 1369 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump() 2007 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in skb_copy_ubufs() 2224 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; in __pskb_copy_fclone() 2287 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); in pskb_expand_head() 2405 offsetof(struct skb_shared_info, frags[shinfo->nr_frags])); in __skb_unclone_keeptruesize() 2718 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim() 2725 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); in ___pskb_trim() [all …]
|
D | tso.c | 54 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 80 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
/linux-6.12.1/drivers/net/ethernet/fungible/funeth/ |
D | funeth_rx.c | 141 static void *fun_run_xdp(struct funeth_rxq *q, skb_frag_t *frags, void *buf_va, in fun_run_xdp() argument 150 xdp_init_buff(&xdp, ALIGN(skb_frag_size(frags), FUN_EPRQ_PKT_ALIGN), in fun_run_xdp() 152 xdp_prepare_buff(&xdp, buf_va, FUN_XDP_HEADROOM, skb_frag_size(frags) - in fun_run_xdp() 161 skb_frag_size_set(frags, xdp.data_end - xdp.data); in fun_run_xdp() 162 skb_frag_off_add(frags, xdp.data - xdp.data_hard_start); in fun_run_xdp() 297 skb_frag_t *frags) in fun_gather_pkt() argument 326 skb_frag_fill_page_desc(frags++, buf->page, q->buf_offset, in fun_gather_pkt() 370 skb_frag_t frags[RX_MAX_FRAGS]; in fun_handle_cqe_pkt() local 392 ref_ok = fun_gather_pkt(q, tot_len, frags); in fun_handle_cqe_pkt() 393 va = skb_frag_address(frags); in fun_handle_cqe_pkt() [all …]
|
/linux-6.12.1/drivers/net/wireless/ath/ath10k/ |
D | htt_tx.c | 1393 struct htt_data_tx_desc_frag *frags; in ath10k_htt_tx_32() local 1456 frags = (struct htt_data_tx_desc_frag *) in ath10k_htt_tx_32() 1457 &ext_desc_t[msdu_id].frags; in ath10k_htt_tx_32() 1459 frags[0].tword_addr.paddr_lo = in ath10k_htt_tx_32() 1461 frags[0].tword_addr.paddr_hi = 0; in ath10k_htt_tx_32() 1462 frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len); in ath10k_htt_tx_32() 1467 frags = txbuf->frags; in ath10k_htt_tx_32() 1468 frags[0].dword_addr.paddr = in ath10k_htt_tx_32() 1470 frags[0].dword_addr.len = __cpu_to_le32(msdu->len); in ath10k_htt_tx_32() 1471 frags[1].dword_addr.paddr = 0; in ath10k_htt_tx_32() [all …]
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep/ |
D | octep_tx.c | 40 u8 frags, i; in octep_iq_process_completions() local 67 frags = shinfo->nr_frags; in octep_iq_process_completions() 74 while (frags--) { in octep_iq_process_completions() 109 u8 frags, i; in octep_iq_free_pending() local 128 frags = shinfo->nr_frags; in octep_iq_free_pending() 136 while (frags--) { in octep_iq_free_pending()
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep_vf/ |
D | octep_vf_tx.c | 41 u8 frags, i; in octep_vf_iq_process_completions() local 68 frags = shinfo->nr_frags; in octep_vf_iq_process_completions() 75 while (frags--) { in octep_vf_iq_process_completions() 108 u8 frags, i; in octep_vf_iq_free_pending() local 127 frags = shinfo->nr_frags; in octep_vf_iq_free_pending() 135 while (frags--) { in octep_vf_iq_free_pending()
|
/linux-6.12.1/drivers/net/wireless/ath/wil6210/ |
D | trace.h | 212 TP_PROTO(u8 vring, u16 index, unsigned int len, u8 frags), 213 TP_ARGS(vring, index, len, frags), 216 __field(u8, frags) 222 __entry->frags = frags; 227 __entry->vring, __entry->index, __entry->len, __entry->frags)
|
/linux-6.12.1/include/linux/ |
D | skbuff_ref.h | 32 __skb_frag_ref(&skb_shinfo(skb)->frags[f]); in skb_frag_ref() 71 __skb_frag_unref(&shinfo->frags[f], skb->pp_recycle); in skb_frag_unref()
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_txrx_lib.c | 275 u32 frags, xdp_tx = 0; in ice_clean_xdp_irq() local 303 frags = tx_buf->nr_frags; in ice_clean_xdp_irq() 306 ready_frames -= frags + 1; in ice_clean_xdp_irq() 313 for (int i = 0; i < frags; i++) { in ice_clean_xdp_irq() 407 data = skb_frag_address(&sinfo->frags[frag]); in __ice_xmit_xdp_ring() 408 size = skb_frag_size(&sinfo->frags[frag]); in __ice_xmit_xdp_ring()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 288 end = &si->frags[si->nr_frags]; in map_skb() 289 for (fp = si->frags; fp < end; fp++) { in map_skb() 298 while (fp-- > si->frags) in map_skb() 917 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl() 933 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 934 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl() 939 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 1476 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags() 1477 gl->frags[0].offset + offset, in copy_frags() 1478 gl->frags[0].size - offset); in copy_frags() [all …]
|
/linux-6.12.1/drivers/net/xen-netback/ |
D | netback.c | 387 skb_frag_t *frags = shinfo->frags; in xenvif_get_requests() local 473 frag_set_pending_idx(&frags[shinfo->nr_frags], pending_idx); in xenvif_get_requests() 486 frags = shinfo->frags; in xenvif_get_requests() 499 frag_set_pending_idx(&frags[shinfo->nr_frags], in xenvif_get_requests() 567 frag_get_pending_idx(&shinfo->frags[0]) == in xenvif_tx_check_gop() 610 pending_idx = frag_get_pending_idx(&shinfo->frags[i]); in xenvif_tx_check_gop() 653 pending_idx = frag_get_pending_idx(&shinfo->frags[j]); in xenvif_tx_check_gop() 664 pending_idx = frag_get_pending_idx(&first_shinfo->frags[j]); in xenvif_tx_check_gop() 695 skb_frag_t *frag = shinfo->frags + i; in xenvif_fill_frags() 1114 skb_frag_t frags[MAX_SKB_FRAGS]; in xenvif_handle_frag_list() local [all …]
|
/linux-6.12.1/block/partitions/ |
D | ldm.c | 1216 static bool ldm_frag_add (const u8 *data, int size, struct list_head *frags) in ldm_frag_add() argument 1222 BUG_ON (!data || !frags); in ldm_frag_add() 1241 list_for_each (item, frags) { in ldm_frag_add() 1258 list_add_tail (&f->list, frags); in ldm_frag_add() 1307 static bool ldm_frag_commit (struct list_head *frags, struct ldmdb *ldb) in ldm_frag_commit() argument 1312 BUG_ON (!frags || !ldb); in ldm_frag_commit() 1314 list_for_each (item, frags) { in ldm_frag_commit() 1348 LIST_HEAD (frags); in ldm_get_vblks() 1375 if (!ldm_frag_add (data, size, &frags)) in ldm_get_vblks() 1384 result = ldm_frag_commit (&frags, ldb); /* Failures, already logged */ in ldm_get_vblks() [all …]
|
/linux-6.12.1/drivers/target/iscsi/cxgbit/ |
D | cxgbit_main.c | 226 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags() 227 gl->frags[0].offset + offset, in cxgbit_copy_frags() 228 gl->frags[0].size - offset); in cxgbit_copy_frags() 231 gl->frags[i].page, in cxgbit_copy_frags() 232 gl->frags[i].offset, in cxgbit_copy_frags() 233 gl->frags[i].size); in cxgbit_copy_frags() 238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags() 308 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl()
|
D | cxgbit_target.c | 882 skb_frag_t *dfrag = &ssi->frags[pdu_cb->dfrag_idx]; in cxgbit_handle_immediate_data() 1401 pdu_cb->ddigest, pdu_cb->frags); in cxgbit_lro_skb_dump() 1404 skb, i, skb_frag_off(&ssi->frags[i]), in cxgbit_lro_skb_dump() 1405 skb_frag_size(&ssi->frags[i])); in cxgbit_lro_skb_dump() 1416 put_page(skb_frag_page(&ssi->frags[i])); in cxgbit_lro_hskb_reset() 1441 memcpy(&hssi->frags[hfrag_idx], &ssi->frags[pdu_cb->hfrag_idx], in cxgbit_lro_skb_merge() 1444 get_page(skb_frag_page(&hssi->frags[hfrag_idx])); in cxgbit_lro_skb_merge() 1446 hpdu_cb->frags++; in cxgbit_lro_skb_merge() 1449 len = skb_frag_size(&hssi->frags[hfrag_idx]); in cxgbit_lro_skb_merge() 1463 memcpy(&hssi->frags[dfrag_idx], in cxgbit_lro_skb_merge() [all …]
|
/linux-6.12.1/drivers/misc/mei/ |
D | vsc-fw-loader.c | 235 struct vsc_img_frag frags[VSC_IMG_FRAG_MAX]; member 428 frag = &fw_loader->frags[csi_image_map[i].frag_index]; in vsc_identify_csi_image() 505 frag = &fw_loader->frags[frag_index]; in vsc_identify_ace_image() 513 last_frag = &fw_loader->frags[frag_index - 1]; in vsc_identify_ace_image() 534 struct vsc_img_frag *frag = &fw_loader->frags[VSC_IMG_SKU_CFG_FRAG]; in vsc_identify_cfg_image() 576 struct vsc_img_frag *frag = &fw_loader->frags[VSC_IMG_BOOTLOADER_FRAG]; in vsc_download_bootloader() 644 struct vsc_img_frag *frag = &fw_loader->frags[i]; in vsc_download_firmware() 658 struct vsc_img_frag *frag = &fw_loader->frags[i]; in vsc_download_firmware()
|
/linux-6.12.1/Documentation/networking/ |
D | devmem.rst | 214 Freeing frags 228 The user must pass no more than 128 tokens, with no more than 1024 total frags 230 than 1024 frags, the kernel will free up to 1024 frags and return early. 232 The kernel returns the number of actual frags freed. The number of frags freed 236 (b) the user passed more than 1024 frags.
|
/linux-6.12.1/net/sunrpc/auth_gss/ |
D | gss_krb5_crypto.c | 477 struct scatterlist frags[4]; member 494 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor() 505 sg_mark_end(&desc->frags[desc->fragno - 1]); in decryptor() 507 skcipher_request_set_crypt(desc->req, desc->frags, desc->frags, in decryptor() 514 sg_init_table(desc->frags, 4); in decryptor() 517 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor() 547 sg_init_table(desc.frags, 4); in gss_decrypt_xdr_buf() 767 sg_init_table(desc.frags, 4); in krb5_cbc_cts_decrypt()
|
/linux-6.12.1/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_xdp.c | 60 skb_frag_t *frag = &sinfo->frags[i]; in bnxt_xmit_bd() 130 int j, frags; in bnxt_tx_int_xdp() local 153 frags = tx_buf->nr_frags; in bnxt_tx_int_xdp() 154 for (j = 0; j < frags; j++) { in bnxt_tx_int_xdp() 213 struct page *page = skb_frag_page(&shinfo->frags[i]); in bnxt_xdp_buff_frags_free()
|