Home
last modified time | relevance | path

Searched refs:arena (Results 1 – 25 of 29) sorted by relevance

12

/linux-6.12.1/drivers/nvdimm/
Dbtt.c29 static struct device *to_dev(struct arena_info *arena) in to_dev() argument
31 return &arena->nd_btt->dev; in to_dev()
39 static int arena_read_bytes(struct arena_info *arena, resource_size_t offset, in arena_read_bytes() argument
42 struct nd_btt *nd_btt = arena->nd_btt; in arena_read_bytes()
50 static int arena_write_bytes(struct arena_info *arena, resource_size_t offset, in arena_write_bytes() argument
53 struct nd_btt *nd_btt = arena->nd_btt; in arena_write_bytes()
61 static int btt_info_write(struct arena_info *arena, struct btt_sb *super) in btt_info_write() argument
70 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->infooff, 512), in btt_info_write()
71 "arena->infooff: %#llx is unaligned\n", arena->infooff); in btt_info_write()
72 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->info2off, 512), in btt_info_write()
[all …]
/linux-6.12.1/kernel/bpf/
Darena.c53 u64 bpf_arena_get_kern_vm_start(struct bpf_arena *arena) in bpf_arena_get_kern_vm_start() argument
55 return arena ? (u64) (long) arena->kern_vm->addr + GUARD_SZ / 2 : 0; in bpf_arena_get_kern_vm_start()
58 u64 bpf_arena_get_user_vm_start(struct bpf_arena *arena) in bpf_arena_get_user_vm_start() argument
60 return arena ? arena->user_vm_start : 0; in bpf_arena_get_user_vm_start()
88 static long compute_pgoff(struct bpf_arena *arena, long uaddr) in compute_pgoff() argument
90 return (u32)(uaddr - (u32)arena->user_vm_start) >> PAGE_SHIFT; in compute_pgoff()
97 struct bpf_arena *arena; in arena_map_alloc() local
124 arena = bpf_map_area_alloc(sizeof(*arena), numa_node); in arena_map_alloc()
125 if (!arena) in arena_map_alloc()
128 arena->kern_vm = kern_vm; in arena_map_alloc()
[all …]
DMakefile19 obj-$(CONFIG_BPF_SYSCALL) += arena.o
/linux-6.12.1/arch/alpha/kernel/
Dpci_iommu.c63 struct pci_iommu_arena *arena; in iommu_arena_new_node() local
74 arena = memblock_alloc(sizeof(*arena), SMP_CACHE_BYTES); in iommu_arena_new_node()
75 if (!arena) in iommu_arena_new_node()
77 sizeof(*arena)); in iommu_arena_new_node()
78 arena->ptes = memblock_alloc(mem_size, align); in iommu_arena_new_node()
79 if (!arena->ptes) in iommu_arena_new_node()
83 spin_lock_init(&arena->lock); in iommu_arena_new_node()
84 arena->hose = hose; in iommu_arena_new_node()
85 arena->dma_base = base; in iommu_arena_new_node()
86 arena->size = window_size; in iommu_arena_new_node()
[all …]
Dcore_titan.c584 struct pci_iommu_arena *arena; member
601 aper->arena = agp->hose->sg_pci; in titan_agp_setup()
603 aper->pg_start = iommu_reserve(aper->arena, aper->pg_count, in titan_agp_setup()
612 aper->arena->dma_base + aper->pg_start * PAGE_SIZE; in titan_agp_setup()
625 status = iommu_release(aper->arena, aper->pg_start, aper->pg_count); in titan_agp_cleanup()
629 iommu_unbind(aper->arena, aper->pg_start, aper->pg_count); in titan_agp_cleanup()
630 status = iommu_release(aper->arena, aper->pg_start, in titan_agp_cleanup()
686 return iommu_bind(aper->arena, aper->pg_start + pg_start, in titan_agp_bind_memory()
694 return iommu_unbind(aper->arena, aper->pg_start + pg_start, in titan_agp_unbind_memory()
702 unsigned long baddr = addr - aper->arena->dma_base; in titan_agp_translate()
[all …]
Dcore_marvel.c854 struct pci_iommu_arena *arena; member
870 aper->arena = agp->hose->sg_pci; in marvel_agp_setup()
872 aper->pg_start = iommu_reserve(aper->arena, aper->pg_count, in marvel_agp_setup()
882 aper->arena->dma_base + aper->pg_start * PAGE_SIZE; in marvel_agp_setup()
895 status = iommu_release(aper->arena, aper->pg_start, aper->pg_count); in marvel_agp_cleanup()
899 iommu_unbind(aper->arena, aper->pg_start, aper->pg_count); in marvel_agp_cleanup()
900 status = iommu_release(aper->arena, aper->pg_start, in marvel_agp_cleanup()
978 return iommu_bind(aper->arena, aper->pg_start + pg_start, in marvel_agp_bind_memory()
986 return iommu_unbind(aper->arena, aper->pg_start + pg_start, in marvel_agp_unbind_memory()
994 unsigned long baddr = addr - aper->arena->dma_base; in marvel_agp_translate()
[all …]
Dcore_cia.c356 struct pci_iommu_arena *arena = pci_isa_hose->sg_isa; in verify_tb_operation() local
375 addr0 = arena->dma_base; in verify_tb_operation()
449 cia_pci_tbi(arena->hose, 0, -1); in verify_tb_operation()
464 arena->ptes[4] = pte0; in verify_tb_operation()
488 arena->ptes[5] = pte0; in verify_tb_operation()
501 arena->align_entry = 4; in verify_tb_operation()
524 arena->ptes[4] = 0; in verify_tb_operation()
525 arena->ptes[5] = 0; in verify_tb_operation()
539 alpha_mv.mv_pci_tbi(arena->hose, 0, -1); in verify_tb_operation()
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/
Darena_atomics.c19 ASSERT_EQ(skel->arena->add64_value, 3, "add64_value"); in test_add()
20 ASSERT_EQ(skel->arena->add64_result, 1, "add64_result"); in test_add()
22 ASSERT_EQ(skel->arena->add32_value, 3, "add32_value"); in test_add()
23 ASSERT_EQ(skel->arena->add32_result, 1, "add32_result"); in test_add()
25 ASSERT_EQ(skel->arena->add_stack_value_copy, 3, "add_stack_value"); in test_add()
26 ASSERT_EQ(skel->arena->add_stack_result, 1, "add_stack_result"); in test_add()
28 ASSERT_EQ(skel->arena->add_noreturn_value, 3, "add_noreturn_value"); in test_add()
44 ASSERT_EQ(skel->arena->sub64_value, -1, "sub64_value"); in test_sub()
45 ASSERT_EQ(skel->arena->sub64_result, 1, "sub64_result"); in test_sub()
47 ASSERT_EQ(skel->arena->sub32_value, -1, "sub32_value"); in test_sub()
[all …]
Darena_list.c52 ASSERT_EQ(skel->arena->arena_sum, expected_sum, "__arena sum of elems"); in test_arena_list_add_del()
53 ASSERT_EQ(skel->arena->test_val, cnt + 1, "num of elems"); in test_arena_list_add_del()
60 ASSERT_EQ(skel->arena->arena_sum, expected_sum, "__arena sum of elems"); in test_arena_list_add_del()
Darena_htab.c46 area = bpf_map__initial_value(skel->maps.arena, &arena_sz); in test_arena_htab_llvm()
/linux-6.12.1/tools/testing/selftests/bpf/progs/
Dverifier_arena.c21 } arena SEC(".maps");
30 page1 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
34 page2 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
38 no_page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
45 bpf_arena_free_pages(&arena, (void __arena *)page2, 1); in basic_alloc1()
50 page3 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1()
69 page1 = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in basic_alloc2()
87 bpf_arena_free_pages(&arena, (void __arena *)page1, 2); in basic_alloc2()
108 struct bpf_arena___l *ar = (struct bpf_arena___l *)&arena; in basic_alloc3()
Dverifier_arena_large.c18 } arena SEC(".maps");
28 page1 = base = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in big_alloc1()
32 page2 = bpf_arena_alloc_pages(&arena, base + ARENA_SIZE - PAGE_SIZE, in big_alloc1()
37 no_page = bpf_arena_alloc_pages(&arena, base + ARENA_SIZE, in big_alloc1()
45 bpf_arena_free_pages(&arena, (void __arena *)page1, 1); in big_alloc1()
50 page3 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in big_alloc1()
Darena_atomics.c19 } arena SEC(".maps");
217 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in uaf()
218 bpf_arena_free_pages(&arena, page, 1); in uaf()
Darena_htab.c14 } arena SEC(".maps");
Darena_list.c19 } arena SEC(".maps");
/linux-6.12.1/arch/powerpc/platforms/pseries/
Drtas-work-area.c42 char *arena; member
137 const phys_addr_t pa_start = __pa(rwa_state.arena); in rtas_work_area_allocator_init()
144 if (!rwa_state.arena) in rtas_work_area_allocator_init()
157 err = gen_pool_add(pool, (unsigned long)rwa_state.arena, in rtas_work_area_allocator_init()
209 rwa_state.arena = memblock_alloc_try_nid(size, align, min, limit, nid); in rtas_work_area_reserve_arena()
/linux-6.12.1/Documentation/driver-api/nvdimm/
Dbtt.rst35 Each arena follows the same layout for its metadata, and all references in an
36 arena are internal to it (with the exception of one field that points to the
37 next arena). The following depicts the "On-disk" metadata layout::
105 ABA Arena Block Address - Block offset/number within an arena
106 Premap ABA The block offset into an arena, which was decided upon by range
112 arena.
117 the external LBA at 768G. This falls into the second arena, and of the 512G
118 worth of blocks that this arena contributes, this block is at 256G. Thus, the
157 While 'nfree' describes the number of concurrent IOs an arena can process
183 The RTT is a simple, per arena table with 'nfree' entries. Every reader inserts
[all …]
/linux-6.12.1/tools/testing/selftests/bpf/
Dbpf_arena_alloc.h33 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in bpf_alloc()
62 bpf_arena_free_pages(&arena, addr, 1); in bpf_free()
Dbpf_arena_common.h57 __weak char arena[1]; variable
Dbpf_arena_htab.h95 void __arena *buckets = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in htab_init()
/linux-6.12.1/arch/riscv/net/
Dbpf_jit_core.c83 ctx->arena_vm_start = bpf_arena_get_kern_vm_start(prog->aux->arena); in bpf_int_jit_compile()
84 ctx->user_vm_start = bpf_arena_get_user_vm_start(prog->aux->arena); in bpf_int_jit_compile()
/linux-6.12.1/arch/arm64/net/
Dbpf_jit_comp.c647 const bool arena = BPF_MODE(code) == BPF_PROBE_ATOMIC; in emit_lse_atomic() local
651 if (off || arena) { in emit_lse_atomic()
657 if (arena) { in emit_lse_atomic()
1801 ctx.user_vm_start = bpf_arena_get_user_vm_start(prog->aux->arena); in bpf_int_jit_compile()
1802 ctx.arena_vm_start = bpf_arena_get_kern_vm_start(prog->aux->arena); in bpf_int_jit_compile()
/linux-6.12.1/arch/x86/net/
Dbpf_jit_comp.c746 if (bpf_arena_get_kern_vm_start(bpf_prog->aux->arena)) in emit_bpf_tail_call_indirect()
813 if (bpf_arena_get_kern_vm_start(bpf_prog->aux->arena)) in emit_bpf_tail_call_direct()
1430 arena_vm_start = bpf_arena_get_kern_vm_start(bpf_prog->aux->arena); in do_jit()
1431 user_vm_start = bpf_arena_get_user_vm_start(bpf_prog->aux->arena); in do_jit()
/linux-6.12.1/tools/bpf/bpftool/Documentation/
Dbpftool-map.rst58 | | **task_storage** | **bloom_filter** | **user_ringbuf** | **cgrp_storage** | **arena** }
/linux-6.12.1/include/linux/
Dbpf.h543 u64 bpf_arena_get_kern_vm_start(struct bpf_arena *arena);
544 u64 bpf_arena_get_user_vm_start(struct bpf_arena *arena);
1494 struct bpf_arena *arena; member

12