/linux-6.12.1/tools/perf/util/ |
D | bpf_lock_contention.c | 36 bpf_map__set_max_entries(skel->maps.lock_stat, con->map_nr_entries); in lock_contention_prepare() 37 bpf_map__set_max_entries(skel->maps.tstamp, con->map_nr_entries); in lock_contention_prepare() 40 bpf_map__set_max_entries(skel->maps.task_data, con->map_nr_entries); in lock_contention_prepare() 42 bpf_map__set_max_entries(skel->maps.task_data, 1); in lock_contention_prepare() 45 bpf_map__set_max_entries(skel->maps.stacks, con->map_nr_entries); in lock_contention_prepare() 47 bpf_map__set_max_entries(skel->maps.stacks, 1); in lock_contention_prepare() 96 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in lock_contention_prepare() 97 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in lock_contention_prepare() 98 bpf_map__set_max_entries(skel->maps.type_filter, ntypes); in lock_contention_prepare() 99 bpf_map__set_max_entries(skel->maps.addr_filter, naddrs); in lock_contention_prepare() [all …]
|
D | bpf_counter_cgroup.c | 71 bpf_map__set_max_entries(skel->maps.events, map_size); in bperf_load_program() 72 bpf_map__set_max_entries(skel->maps.cgrp_idx, nr_cgroups); in bperf_load_program() 75 bpf_map__set_max_entries(skel->maps.prev_readings, map_size); in bperf_load_program() 78 bpf_map__set_max_entries(skel->maps.cgrp_readings, map_size); in bperf_load_program()
|
D | bpf_off_cpu.c | 148 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in off_cpu_prepare() 174 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in off_cpu_prepare() 179 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in off_cpu_prepare() 182 bpf_map__set_max_entries(skel->maps.task_filter, MAX_PROC); in off_cpu_prepare() 189 bpf_map__set_max_entries(skel->maps.cgroup_filter, ncgrps); in off_cpu_prepare()
|
D | bpf_counter.c | 130 bpf_map__set_max_entries(skel->maps.events, evsel__nr_cpus(evsel)); in bpf_program_profiler_load_one() 131 bpf_map__set_max_entries(skel->maps.fentry_readings, 1); in bpf_program_profiler_load_one() 132 bpf_map__set_max_entries(skel->maps.accum_readings, 1); in bpf_program_profiler_load_one() 410 bpf_map__set_max_entries(skel->maps.events, libbpf_num_possible_cpus()); in bperf_reload_leader_program() 530 bpf_map__set_max_entries(evsel->follower_skel->maps.accum_readings, in bperf__load() 533 bpf_map__set_max_entries(evsel->follower_skel->maps.filter, in bperf__load()
|
D | bpf_ftrace.c | 42 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in perf_ftrace__latency_prepare_bpf() 48 bpf_map__set_max_entries(skel->maps.task_filter, ntasks); in perf_ftrace__latency_prepare_bpf()
|
D | bpf-filter.c | 651 bpf_map__set_max_entries(skel->maps.filters, MAX_FILTERS); in perf_bpf_filter__pin() 652 bpf_map__set_max_entries(skel->maps.event_hash, MAX_EVT_HASH); in perf_bpf_filter__pin() 653 bpf_map__set_max_entries(skel->maps.idx_hash, MAX_IDX_HASH); in perf_bpf_filter__pin() 654 bpf_map__set_max_entries(skel->maps.dropped, MAX_FILTERS); in perf_bpf_filter__pin()
|
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/ |
D | ringbuf_multi.c | 57 ASSERT_OK(bpf_map__set_max_entries(skel->maps.ringbuf1, page_size + 1), "rb1_resize"); in test_ringbuf_multi() 59 ASSERT_OK(bpf_map__set_max_entries(skel->maps.ringbuf1, page_size), "rb1_reset"); in test_ringbuf_multi() 78 …if (!ASSERT_ERR(bpf_map__set_max_entries(skel->maps.ringbuf1, 3 * page_size), "rb1_resize_after_lo… in test_ringbuf_multi()
|
D | mmap.c | 36 err = bpf_map__set_max_entries(skel->maps.rdonly_map, page_size); in test_mmap() 41 err = bpf_map__set_max_entries(skel->maps.data_map, in test_mmap()
|
D | map_init.c | 52 err = bpf_map__set_max_entries(skel->maps.hashmap1, map_sz); in setup()
|
D | user_ringbuf.c | 80 err = bpf_map__set_max_entries(skel->maps.user_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel() 84 err = bpf_map__set_max_entries(skel->maps.kernel_ringbuf, c_ringbuf_size); in open_load_ringbuf_skel()
|
D | lookup_and_delete.c | 58 err = bpf_map__set_max_entries(skel->maps.hash_map, MAX_ENTRIES); in setup_prog()
|
/linux-6.12.1/tools/testing/selftests/bpf/benchs/ |
D | bench_bloom_filter_map.c | 272 bpf_map__set_max_entries(skel->maps.hashmap, args.nr_entries); in setup_skeleton() 274 bpf_map__set_max_entries(skel->maps.array_map, args.nr_entries); in setup_skeleton() 276 bpf_map__set_max_entries(skel->maps.bloom_map, args.nr_entries); in setup_skeleton()
|
D | bench_bpf_hashmap_lookup.c | 166 bpf_map__set_max_entries(ctx.skel->maps.hash_map_bench, args.max_entries); in setup()
|
D | bench_htab_mem.c | 183 bpf_map__set_max_entries(map, MAX(8192, 64 * env.nr_cpus)); in htab_mem_setup()
|
D | bench_ringbufs.c | 154 bpf_map__set_max_entries(skel->maps.ringbuf, args.ringbuf_sz); in ringbuf_setup_skeleton()
|
/linux-6.12.1/tools/lib/bpf/ |
D | libbpf.map | 188 bpf_map__set_max_entries;
|
D | libbpf.h | 1021 LIBBPF_API int bpf_map__set_max_entries(struct bpf_map *map, __u32 max_entries);
|
D | libbpf.c | 4908 int bpf_map__set_max_entries(struct bpf_map *map, __u32 max_entries) in bpf_map__set_max_entries() function
|
/linux-6.12.1/Documentation/bpf/ |
D | map_cpumap.rst | 166 if (bpf_map__set_max_entries(cpu_map, libbpf_num_possible_cpus()) < 0) {
|
/linux-6.12.1/tools/bpf/bpftool/ |
D | prog.c | 2400 bpf_map__set_max_entries(profile_obj->maps.events, num_metric * num_cpu); in do_profile() 2401 bpf_map__set_max_entries(profile_obj->maps.fentry_readings, num_metric); in do_profile() 2402 bpf_map__set_max_entries(profile_obj->maps.accum_readings, num_metric); in do_profile() 2403 bpf_map__set_max_entries(profile_obj->maps.counts, 1); in do_profile()
|
/linux-6.12.1/samples/bpf/ |
D | map_perf_test_user.c | 415 bpf_map__set_max_entries(map, num_map_entries); in fixup_map()
|
D | xdp_sample_user.c | 1220 if (bpf_map__set_max_entries(sample_map[i], sample_map_count[i]) < 0) in sample_setup_maps()
|
/linux-6.12.1/tools/testing/selftests/bpf/ |
D | veristat.c | 929 bpf_map__set_max_entries(map, 1); in fixup_obj()
|