Home
last modified time | relevance | path

Searched refs:frags (Results 1 – 25 of 297) sorted by relevance

12345678910>>...12

/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/
Dalloc.c82 buf->frags = kcalloc(buf->npages, sizeof(struct mlx5_buf_list), in mlx5_frag_buf_alloc_node()
84 if (!buf->frags) in mlx5_frag_buf_alloc_node()
88 struct mlx5_buf_list *frag = &buf->frags[i]; in mlx5_frag_buf_alloc_node()
97 buf->frags[i].buf, buf->frags[i].map); in mlx5_frag_buf_alloc_node()
109 dma_free_coherent(mlx5_core_dma_dev(dev), PAGE_SIZE, buf->frags[i].buf, in mlx5_frag_buf_alloc_node()
110 buf->frags[i].map); in mlx5_frag_buf_alloc_node()
111 kfree(buf->frags); in mlx5_frag_buf_alloc_node()
125 dma_free_coherent(mlx5_core_dma_dev(dev), frag_sz, buf->frags[i].buf, in mlx5_frag_buf_free()
126 buf->frags[i].map); in mlx5_frag_buf_free()
129 kfree(buf->frags); in mlx5_frag_buf_free()
[all …]
Dwq.c61 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_cyc_create()
128 mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc); in mlx5_wq_qp_create()
136 mlx5_init_fbc_offset(wq_ctrl->buf.frags, in mlx5_wq_qp_create()
142 mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages, in mlx5_wq_qp_create()
185 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc); in mlx5_cqwq_create()
234 mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc); in mlx5_wq_ll_create()
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c78 struct mlx4_en_rx_alloc *frags, in mlx4_en_alloc_frags() argument
83 for (i = 0; i < priv->num_frags; i++, frags++) { in mlx4_en_alloc_frags()
84 if (!frags->page) { in mlx4_en_alloc_frags()
85 if (mlx4_alloc_page(priv, frags, gfp)) { in mlx4_en_alloc_frags()
91 rx_desc->data[i].addr = cpu_to_be64(frags->dma + in mlx4_en_alloc_frags()
92 frags->page_offset); in mlx4_en_alloc_frags()
142 struct mlx4_en_rx_alloc *frags = ring->rx_info + in mlx4_en_prepare_rx_desc() local
146 if (!frags->page) { in mlx4_en_prepare_rx_desc()
148 frags->page = ring->page_cache.buf[ring->page_cache.index].page; in mlx4_en_prepare_rx_desc()
149 frags->dma = ring->page_cache.buf[ring->page_cache.index].dma; in mlx4_en_prepare_rx_desc()
[all …]
/linux-6.12.1/tools/testing/selftests/bpf/
Dgenerate_udp_fragments.py43 def print_frags(f, frags, v6): argument
44 for idx, frag in enumerate(frags):
77 frags = [f.build() for f in pkt.fragment(24)]
81 print_frags(f, frags, False)
/linux-6.12.1/drivers/net/wireless/intel/iwlwifi/
Diwl-dbg-tlv.c403 struct iwl_dram_data *frag = &fw_mon->frags[i]; in iwl_dbg_tlv_fragments_free()
413 kfree(fw_mon->frags); in iwl_dbg_tlv_fragments_free()
414 fw_mon->frags = NULL; in iwl_dbg_tlv_fragments_free()
596 memset(fw_mon->frags[i].block, 0, in iwl_dbg_tlv_alloc_fragments()
597 fw_mon->frags[i].size); in iwl_dbg_tlv_alloc_fragments()
621 fw_mon->frags = kcalloc(num_frags, sizeof(*fw_mon->frags), GFP_KERNEL); in iwl_dbg_tlv_alloc_fragments()
622 if (!fw_mon->frags) in iwl_dbg_tlv_alloc_fragments()
632 pages = iwl_dbg_tlv_alloc_fragment(fwrt, &fw_mon->frags[i], in iwl_dbg_tlv_alloc_fragments()
707 struct iwl_buf_alloc_frag *frag = &data.frags[j]; in iwl_dbg_tlv_apply_buffer()
709 &fw_mon->frags[fw_mon_idx++]; in iwl_dbg_tlv_apply_buffer()
[all …]
/linux-6.12.1/net/core/
Dgro.c143 frag = pinfo->frags + nr_frags; in skb_gro_receive()
144 frag2 = skbinfo->frags + i; in skb_gro_receive()
164 skb_frag_t *frag = pinfo->frags + nr_frags; in skb_gro_receive()
180 memcpy(frag + 1, skbinfo->frags, sizeof(*frag) * skbinfo->nr_frags); in skb_gro_receive()
198 skb_frag_off_add(&skbinfo->frags[0], eat); in skb_gro_receive()
199 skb_frag_size_sub(&skbinfo->frags[0], eat); in skb_gro_receive()
404 frag0 = &pinfo->frags[0]; in skb_gro_reset_offset()
427 skb_frag_off_add(&pinfo->frags[0], grow); in gro_pull_from_frag0()
428 skb_frag_size_sub(&pinfo->frags[0], grow); in gro_pull_from_frag0()
430 if (unlikely(!skb_frag_size(&pinfo->frags[0]))) { in gro_pull_from_frag0()
[all …]
Dnet_test.c46 const unsigned int *frags; member
75 .frags = (const unsigned int[]) { GSO_TEST_SIZE, 1 },
83 .frags = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 },
92 .frags = (const unsigned int[]) { GSO_TEST_SIZE, 3 },
174 skb_fill_page_desc(skb, i, page, pg_off, tcase->frags[i]); in gso_test_func()
175 pg_off += tcase->frags[i]; in gso_test_func()
Dskbuff.c894 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag()
1072 head_netmem = netmem_compound_head(shinfo->frags[i].netmem); in skb_pp_frag_ref()
1119 __skb_frag_unref(&shinfo->frags[i], skb->pp_recycle); in skb_release_data()
1369 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump()
2007 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in skb_copy_ubufs()
2224 skb_shinfo(n)->frags[i] = skb_shinfo(skb)->frags[i]; in __pskb_copy_fclone()
2287 offsetof(struct skb_shared_info, frags[skb_shinfo(skb)->nr_frags])); in pskb_expand_head()
2405 offsetof(struct skb_shared_info, frags[shinfo->nr_frags])); in __skb_unclone_keeptruesize()
2718 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim()
2725 skb_frag_size_set(&skb_shinfo(skb)->frags[i++], len - offset); in ___pskb_trim()
[all …]
Dtso.c54 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data()
80 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
/linux-6.12.1/drivers/net/ethernet/fungible/funeth/
Dfuneth_rx.c141 static void *fun_run_xdp(struct funeth_rxq *q, skb_frag_t *frags, void *buf_va, in fun_run_xdp() argument
150 xdp_init_buff(&xdp, ALIGN(skb_frag_size(frags), FUN_EPRQ_PKT_ALIGN), in fun_run_xdp()
152 xdp_prepare_buff(&xdp, buf_va, FUN_XDP_HEADROOM, skb_frag_size(frags) - in fun_run_xdp()
161 skb_frag_size_set(frags, xdp.data_end - xdp.data); in fun_run_xdp()
162 skb_frag_off_add(frags, xdp.data - xdp.data_hard_start); in fun_run_xdp()
297 skb_frag_t *frags) in fun_gather_pkt() argument
326 skb_frag_fill_page_desc(frags++, buf->page, q->buf_offset, in fun_gather_pkt()
370 skb_frag_t frags[RX_MAX_FRAGS]; in fun_handle_cqe_pkt() local
392 ref_ok = fun_gather_pkt(q, tot_len, frags); in fun_handle_cqe_pkt()
393 va = skb_frag_address(frags); in fun_handle_cqe_pkt()
[all …]
/linux-6.12.1/drivers/net/wireless/ath/ath10k/
Dhtt_tx.c1393 struct htt_data_tx_desc_frag *frags; in ath10k_htt_tx_32() local
1456 frags = (struct htt_data_tx_desc_frag *) in ath10k_htt_tx_32()
1457 &ext_desc_t[msdu_id].frags; in ath10k_htt_tx_32()
1459 frags[0].tword_addr.paddr_lo = in ath10k_htt_tx_32()
1461 frags[0].tword_addr.paddr_hi = 0; in ath10k_htt_tx_32()
1462 frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len); in ath10k_htt_tx_32()
1467 frags = txbuf->frags; in ath10k_htt_tx_32()
1468 frags[0].dword_addr.paddr = in ath10k_htt_tx_32()
1470 frags[0].dword_addr.len = __cpu_to_le32(msdu->len); in ath10k_htt_tx_32()
1471 frags[1].dword_addr.paddr = 0; in ath10k_htt_tx_32()
[all …]
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep/
Doctep_tx.c40 u8 frags, i; in octep_iq_process_completions() local
67 frags = shinfo->nr_frags; in octep_iq_process_completions()
74 while (frags--) { in octep_iq_process_completions()
109 u8 frags, i; in octep_iq_free_pending() local
128 frags = shinfo->nr_frags; in octep_iq_free_pending()
136 while (frags--) { in octep_iq_free_pending()
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep_vf/
Doctep_vf_tx.c41 u8 frags, i; in octep_vf_iq_process_completions() local
68 frags = shinfo->nr_frags; in octep_vf_iq_process_completions()
75 while (frags--) { in octep_vf_iq_process_completions()
108 u8 frags, i; in octep_vf_iq_free_pending() local
127 frags = shinfo->nr_frags; in octep_vf_iq_free_pending()
135 while (frags--) { in octep_vf_iq_free_pending()
/linux-6.12.1/drivers/net/wireless/ath/wil6210/
Dtrace.h212 TP_PROTO(u8 vring, u16 index, unsigned int len, u8 frags),
213 TP_ARGS(vring, index, len, frags),
216 __field(u8, frags)
222 __entry->frags = frags;
227 __entry->vring, __entry->index, __entry->len, __entry->frags)
/linux-6.12.1/include/linux/
Dskbuff_ref.h32 __skb_frag_ref(&skb_shinfo(skb)->frags[f]); in skb_frag_ref()
71 __skb_frag_unref(&shinfo->frags[f], skb->pp_recycle); in skb_frag_unref()
/linux-6.12.1/drivers/net/ethernet/intel/ice/
Dice_txrx_lib.c275 u32 frags, xdp_tx = 0; in ice_clean_xdp_irq() local
303 frags = tx_buf->nr_frags; in ice_clean_xdp_irq()
306 ready_frames -= frags + 1; in ice_clean_xdp_irq()
313 for (int i = 0; i < frags; i++) { in ice_clean_xdp_irq()
407 data = skb_frag_address(&sinfo->frags[frag]); in __ice_xmit_xdp_ring()
408 size = skb_frag_size(&sinfo->frags[frag]); in __ice_xmit_xdp_ring()
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c288 end = &si->frags[si->nr_frags]; in map_skb()
289 for (fp = si->frags; fp < end; fp++) { in map_skb()
298 while (fp-- > si->frags) in map_skb()
917 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl()
933 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
934 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl()
939 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
1476 __skb_fill_page_desc(skb, 0, gl->frags[0].page, in copy_frags()
1477 gl->frags[0].offset + offset, in copy_frags()
1478 gl->frags[0].size - offset); in copy_frags()
[all …]
/linux-6.12.1/drivers/net/xen-netback/
Dnetback.c387 skb_frag_t *frags = shinfo->frags; in xenvif_get_requests() local
473 frag_set_pending_idx(&frags[shinfo->nr_frags], pending_idx); in xenvif_get_requests()
486 frags = shinfo->frags; in xenvif_get_requests()
499 frag_set_pending_idx(&frags[shinfo->nr_frags], in xenvif_get_requests()
567 frag_get_pending_idx(&shinfo->frags[0]) == in xenvif_tx_check_gop()
610 pending_idx = frag_get_pending_idx(&shinfo->frags[i]); in xenvif_tx_check_gop()
653 pending_idx = frag_get_pending_idx(&shinfo->frags[j]); in xenvif_tx_check_gop()
664 pending_idx = frag_get_pending_idx(&first_shinfo->frags[j]); in xenvif_tx_check_gop()
695 skb_frag_t *frag = shinfo->frags + i; in xenvif_fill_frags()
1114 skb_frag_t frags[MAX_SKB_FRAGS]; in xenvif_handle_frag_list() local
[all …]
/linux-6.12.1/block/partitions/
Dldm.c1216 static bool ldm_frag_add (const u8 *data, int size, struct list_head *frags) in ldm_frag_add() argument
1222 BUG_ON (!data || !frags); in ldm_frag_add()
1241 list_for_each (item, frags) { in ldm_frag_add()
1258 list_add_tail (&f->list, frags); in ldm_frag_add()
1307 static bool ldm_frag_commit (struct list_head *frags, struct ldmdb *ldb) in ldm_frag_commit() argument
1312 BUG_ON (!frags || !ldb); in ldm_frag_commit()
1314 list_for_each (item, frags) { in ldm_frag_commit()
1348 LIST_HEAD (frags); in ldm_get_vblks()
1375 if (!ldm_frag_add (data, size, &frags)) in ldm_get_vblks()
1384 result = ldm_frag_commit (&frags, ldb); /* Failures, already logged */ in ldm_get_vblks()
[all …]
/linux-6.12.1/drivers/target/iscsi/cxgbit/
Dcxgbit_main.c226 __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, in cxgbit_copy_frags()
227 gl->frags[0].offset + offset, in cxgbit_copy_frags()
228 gl->frags[0].size - offset); in cxgbit_copy_frags()
231 gl->frags[i].page, in cxgbit_copy_frags()
232 gl->frags[i].offset, in cxgbit_copy_frags()
233 gl->frags[i].size); in cxgbit_copy_frags()
238 get_page(gl->frags[gl->nfrags - 1].page); in cxgbit_copy_frags()
308 pdu_cb->frags += gl->nfrags; in cxgbit_lro_add_packet_gl()
Dcxgbit_target.c882 skb_frag_t *dfrag = &ssi->frags[pdu_cb->dfrag_idx]; in cxgbit_handle_immediate_data()
1401 pdu_cb->ddigest, pdu_cb->frags); in cxgbit_lro_skb_dump()
1404 skb, i, skb_frag_off(&ssi->frags[i]), in cxgbit_lro_skb_dump()
1405 skb_frag_size(&ssi->frags[i])); in cxgbit_lro_skb_dump()
1416 put_page(skb_frag_page(&ssi->frags[i])); in cxgbit_lro_hskb_reset()
1441 memcpy(&hssi->frags[hfrag_idx], &ssi->frags[pdu_cb->hfrag_idx], in cxgbit_lro_skb_merge()
1444 get_page(skb_frag_page(&hssi->frags[hfrag_idx])); in cxgbit_lro_skb_merge()
1446 hpdu_cb->frags++; in cxgbit_lro_skb_merge()
1449 len = skb_frag_size(&hssi->frags[hfrag_idx]); in cxgbit_lro_skb_merge()
1463 memcpy(&hssi->frags[dfrag_idx], in cxgbit_lro_skb_merge()
[all …]
/linux-6.12.1/drivers/misc/mei/
Dvsc-fw-loader.c235 struct vsc_img_frag frags[VSC_IMG_FRAG_MAX]; member
428 frag = &fw_loader->frags[csi_image_map[i].frag_index]; in vsc_identify_csi_image()
505 frag = &fw_loader->frags[frag_index]; in vsc_identify_ace_image()
513 last_frag = &fw_loader->frags[frag_index - 1]; in vsc_identify_ace_image()
534 struct vsc_img_frag *frag = &fw_loader->frags[VSC_IMG_SKU_CFG_FRAG]; in vsc_identify_cfg_image()
576 struct vsc_img_frag *frag = &fw_loader->frags[VSC_IMG_BOOTLOADER_FRAG]; in vsc_download_bootloader()
644 struct vsc_img_frag *frag = &fw_loader->frags[i]; in vsc_download_firmware()
658 struct vsc_img_frag *frag = &fw_loader->frags[i]; in vsc_download_firmware()
/linux-6.12.1/Documentation/networking/
Ddevmem.rst214 Freeing frags
228 The user must pass no more than 128 tokens, with no more than 1024 total frags
230 than 1024 frags, the kernel will free up to 1024 frags and return early.
232 The kernel returns the number of actual frags freed. The number of frags freed
236 (b) the user passed more than 1024 frags.
/linux-6.12.1/net/sunrpc/auth_gss/
Dgss_krb5_crypto.c477 struct scatterlist frags[4]; member
494 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor()
505 sg_mark_end(&desc->frags[desc->fragno - 1]); in decryptor()
507 skcipher_request_set_crypt(desc->req, desc->frags, desc->frags, in decryptor()
514 sg_init_table(desc->frags, 4); in decryptor()
517 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor()
547 sg_init_table(desc.frags, 4); in gss_decrypt_xdr_buf()
767 sg_init_table(desc.frags, 4); in krb5_cbc_cts_decrypt()
/linux-6.12.1/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_xdp.c60 skb_frag_t *frag = &sinfo->frags[i]; in bnxt_xmit_bd()
130 int j, frags; in bnxt_tx_int_xdp() local
153 frags = tx_buf->nr_frags; in bnxt_tx_int_xdp()
154 for (j = 0; j < frags; j++) { in bnxt_tx_int_xdp()
213 struct page *page = skb_frag_page(&shinfo->frags[i]); in bnxt_xdp_buff_frags_free()

12345678910>>...12