Home
last modified time | relevance | path

Searched refs:batch_size (Results 1 – 25 of 30) sorted by relevance

12

/linux-6.12.1/tools/testing/selftests/bpf/
Dtest_lru_map.c244 unsigned int batch_size; in test_lru_sanity1() local
257 batch_size = tgt_free / 2; in test_lru_sanity1()
258 assert(batch_size * 2 == tgt_free); in test_lru_sanity1()
260 map_size = tgt_free + batch_size; in test_lru_sanity1()
276 end_key = 1 + batch_size; in test_lru_sanity1()
321 unsigned int batch_size; in test_lru_sanity2() local
334 batch_size = tgt_free / 2; in test_lru_sanity2()
335 assert(batch_size * 2 == tgt_free); in test_lru_sanity2()
337 map_size = tgt_free + batch_size; in test_lru_sanity2()
376 end_key = 1 + batch_size; in test_lru_sanity2()
[all …]
Dxskxceiver.c254 sock_opt = xsk->batch_size; in enable_busy_poll()
504 ifobj->xsk_arr[j].batch_size = DEFAULT_BATCH_SIZE; in __test_spec_init()
1091 static int complete_pkts(struct xsk_socket_info *xsk, int batch_size) in complete_pkts() argument
1103 rcvd = xsk_ring_cons__peek(&xsk->umem->cq, batch_size, &idx); in complete_pkts()
1157 rcvd = xsk_ring_cons__peek(&xsk->rx, xsk->batch_size, &idx_rx); in __receive_pkts()
1309 if (pkts_in_flight >= (int)((umem_size(umem) - xsk->batch_size * buffer_len) / in __send_pkts()
1320 while (xsk_ring_prod__reserve(&xsk->tx, xsk->batch_size, &idx) < xsk->batch_size) { in __send_pkts()
1340 complete_pkts(xsk, xsk->batch_size); in __send_pkts()
1343 for (i = 0; i < xsk->batch_size; i++) { in __send_pkts()
1351 if (nb_frags > xsk->batch_size - i) { in __send_pkts()
[all …]
Dxskxceiver.h95 u32 batch_size; member
/linux-6.12.1/tools/testing/selftests/bpf/map_tests/
Dmap_in_map_batch_ops.c123 __u32 batch_size, bool delete_entries) in fetch_and_validate() argument
137 for (step_size = batch_size; in fetch_and_validate()
139 step_size += batch_size) { in fetch_and_validate()
180 delete_entries ? "LOOKUP_AND_DELETE" : "LOOKUP", batch_size); in fetch_and_validate()
193 int batch_size[2] = {5, 10}; in _map_in_map_batch_ops() local
222 batch_size[op_index], false); in _map_in_map_batch_ops()
/linux-6.12.1/net/bpf/
Dtest_run.c119 u32 batch_size; member
168 .pool_size = xdp->batch_size, in xdp_test_run_setup()
174 xdp->frames = kvmalloc_array(xdp->batch_size, sizeof(void *), GFP_KERNEL); in xdp_test_run_setup()
178 xdp->skbs = kvmalloc_array(xdp->batch_size, sizeof(void *), GFP_KERNEL); in xdp_test_run_setup()
298 batch_sz = min_t(u32, repeat, xdp->batch_size); in xdp_test_run_batch()
373 u32 repeat, u32 batch_size, u32 *time) in bpf_test_run_xdp_live() argument
376 struct xdp_test_data xdp = { .batch_size = batch_size }; in bpf_test_run_xdp_live()
691 if (kattr->test.flags || kattr->test.cpu || kattr->test.batch_size) in bpf_prog_test_run_tracing()
766 kattr->test.repeat || kattr->test.batch_size) in bpf_prog_test_run_raw_tp()
998 kattr->test.cpu || kattr->test.batch_size) in bpf_prog_test_run_skb()
[all …]
/linux-6.12.1/drivers/gpu/drm/xe/
Dxe_migrate.c809 u32 batch_size = 2; /* arb_clear() + MI_BATCH_BUFFER_END */ in xe_migrate_copy() local
831 batch_size += pte_update_size(m, pte_flags, src, &src_it, &src_L0, in xe_migrate_copy()
836 batch_size += pte_update_size(m, pte_flags, dst, &dst_it, &src_L0, in xe_migrate_copy()
842 batch_size += pte_update_size(m, 0, NULL, &ccs_it, &ccs_size, in xe_migrate_copy()
850 batch_size += ((copy_only_ccs) ? 0 : EMIT_COPY_DW) + in xe_migrate_copy()
853 bb = xe_bb_new(gt, batch_size, usm); in xe_migrate_copy()
1086 u32 batch_size, update_idx; in xe_migrate_clear() local
1096 batch_size = 2 + in xe_migrate_clear()
1103 batch_size += EMIT_COPY_CCS_DW; in xe_migrate_clear()
1110 bb = xe_bb_new(gt, batch_size, usm); in xe_migrate_clear()
[all …]
/linux-6.12.1/drivers/gpu/drm/i915/gt/
Dintel_renderstate.c90 so->batch_size = rodata->batch_items * sizeof(u32); in render_state_setup()
223 so->batch_offset, so->batch_size, in intel_renderstate_emit()
Dintel_renderstate.h40 u32 batch_size; member
/linux-6.12.1/drivers/gpu/drm/xe/tests/
Dxe_migrate.c396 u32 batch_size = 2; /* arb_clear() + MI_BATCH_BUFFER_END */ in blt_copy() local
411 batch_size += pte_update_size(m, pte_flags, src, &src_it, &src_L0, in blt_copy()
417 batch_size += pte_update_size(m, pte_flags, dst, &dst_it, &src_L0, in blt_copy()
422 batch_size += ((copy_only_ccs) ? 0 : EMIT_COPY_DW) + in blt_copy()
425 bb = xe_bb_new(gt, batch_size, xe->info.has_usm); in blt_copy()
/linux-6.12.1/drivers/net/ethernet/sfc/falcon/
Drx.c334 unsigned int fill_level, batch_size; in ef4_fast_push_rx_descriptors() local
352 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in ef4_fast_push_rx_descriptors()
354 EF4_BUG_ON_PARANOID(space < batch_size); in ef4_fast_push_rx_descriptors()
371 } while ((space -= batch_size) >= batch_size); in ef4_fast_push_rx_descriptors()
/linux-6.12.1/drivers/net/ethernet/sfc/siena/
Drx_common.c461 unsigned int fill_level, batch_size; in efx_siena_fast_push_rx_descriptors() local
479 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in efx_siena_fast_push_rx_descriptors()
481 EFX_WARN_ON_ONCE_PARANOID(space < batch_size); in efx_siena_fast_push_rx_descriptors()
496 } while ((space -= batch_size) >= batch_size); in efx_siena_fast_push_rx_descriptors()
/linux-6.12.1/drivers/net/ethernet/sfc/
Drx_common.c460 unsigned int fill_level, batch_size; in efx_fast_push_rx_descriptors() local
478 batch_size = efx->rx_pages_per_batch * efx->rx_bufs_per_page; in efx_fast_push_rx_descriptors()
480 EFX_WARN_ON_ONCE_PARANOID(space < batch_size); in efx_fast_push_rx_descriptors()
495 } while ((space -= batch_size) >= batch_size); in efx_fast_push_rx_descriptors()
/linux-6.12.1/mm/
Dshrinker.c380 long batch_size = shrinker->batch ? shrinker->batch in do_shrink_slab() local
430 while (total_scan >= batch_size || in do_shrink_slab()
433 unsigned long nr_to_scan = min(batch_size, total_scan); in do_shrink_slab()
Dmemory.c2126 const int batch_size = min_t(int, pages_to_write_in_pmd, 8); in insert_pages() local
2133 for (pte = start_pte; pte_idx < batch_size; ++pte, ++pte_idx) { in insert_pages()
2146 pages_to_write_in_pmd -= batch_size; in insert_pages()
2147 remaining_pages_total -= batch_size; in insert_pages()
/linux-6.12.1/tools/lib/bpf/
Dbpf.h672 __u32 batch_size; member
674 #define bpf_test_run_opts__last_field batch_size
Dbpf.c989 attr.test.batch_size = OPTS_GET(opts, batch_size, 0); in bpf_prog_test_run_opts()
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/
Dxdp_do_redirect.c104 .batch_size = 64, in test_xdp_do_redirect()
/linux-6.12.1/drivers/dma/idxd/
Dsysfs.c1049 u64 batch_size; in wq_max_batch_size_store() local
1058 rc = __get_sysfs_u64(buf, &batch_size); in wq_max_batch_size_store()
1062 if (batch_size > idxd->max_batch_size) in wq_max_batch_size_store()
1065 idxd_wq_set_max_batch_size(idxd->data->type, wq, (u32)batch_size); in wq_max_batch_size_store()
/linux-6.12.1/drivers/md/
Draid5.c6650 int i, batch_size = 0, hash; in handle_active_stripes() local
6653 while (batch_size < MAX_STRIPE_BATCH && in handle_active_stripes()
6655 batch[batch_size++] = sh; in handle_active_stripes()
6657 if (batch_size == 0) { in handle_active_stripes()
6665 return batch_size; in handle_active_stripes()
6680 for (i = 0; i < batch_size; i++) in handle_active_stripes()
6687 for (i = 0; i < batch_size; i++) { in handle_active_stripes()
6691 return batch_size; in handle_active_stripes()
6710 int batch_size, released; in raid5_do_work() local
6714 batch_size = handle_active_stripes(conf, group_id, worker, in raid5_do_work()
[all …]
/linux-6.12.1/Documentation/bpf/
Dbpf_prog_run.rst94 the ``batch_size`` parameter, up to a maximum of 256 packets. For each batch,
/linux-6.12.1/drivers/media/pci/tw5864/
Dtw5864-video.c1323 int batch_size = H264_BUF_CNT; in tw5864_handle_frame_work() local
1326 while (dev->h264_buf_r_index != dev->h264_buf_w_index && batch_size--) { in tw5864_handle_frame_work()
/linux-6.12.1/Documentation/networking/
Daf_xdp.rst679 int batch_size)
683 xsk_ring_prod__reserve(&xsk->tx, batch_size, &idx);
685 for (i = 0; i < batch_size;) {
706 if (i == batch_size) {
/linux-6.12.1/drivers/net/wireless/mediatek/mt76/
Dmt76_connac_mcu.c2083 int i, n_chan, batch_size, idx = 0, tx_power, last_ch, err = 0; in mt76_connac_mcu_rate_txpower_band() local
2107 batch_size = DIV_ROUND_UP(n_chan, batch_len); in mt76_connac_mcu_rate_txpower_band()
2116 for (i = 0; i < batch_size; i++) { in mt76_connac_mcu_rate_txpower_band()
2121 num_ch = i == batch_size - 1 ? n_chan - i * batch_len : batch_len; in mt76_connac_mcu_rate_txpower_band()
/linux-6.12.1/drivers/net/wireless/mediatek/mt76/mt7925/
Dmcu.c3313 int i, sku_len, batch_size, batch_len = 3; in mt7925_mcu_rate_txpower_band() local
3368 batch_size = DIV_ROUND_UP(n_chan, batch_len); in mt7925_mcu_rate_txpower_band()
3380 for (i = 0; i < batch_size; i++) { in mt7925_mcu_rate_txpower_band()
3385 num_ch = i == batch_size - 1 ? n_chan % batch_len : batch_len; in mt7925_mcu_rate_txpower_band()
/linux-6.12.1/fs/btrfs/
Dtree-log.c3725 int batch_size = 0; in process_dir_items_leaf() local
3806 if (batch_size == 0) in process_dir_items_leaf()
3808 batch_size++; in process_dir_items_leaf()
3811 if (batch_size > 0) { in process_dir_items_leaf()
3815 batch_start, batch_size); in process_dir_items_leaf()

12