/linux-6.12.1/drivers/nvdimm/ |
D | btt.c | 29 static struct device *to_dev(struct arena_info *arena) in to_dev() argument 31 return &arena->nd_btt->dev; in to_dev() 39 static int arena_read_bytes(struct arena_info *arena, resource_size_t offset, in arena_read_bytes() argument 42 struct nd_btt *nd_btt = arena->nd_btt; in arena_read_bytes() 50 static int arena_write_bytes(struct arena_info *arena, resource_size_t offset, in arena_write_bytes() argument 53 struct nd_btt *nd_btt = arena->nd_btt; in arena_write_bytes() 61 static int btt_info_write(struct arena_info *arena, struct btt_sb *super) in btt_info_write() argument 70 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->infooff, 512), in btt_info_write() 71 "arena->infooff: %#llx is unaligned\n", arena->infooff); in btt_info_write() 72 dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->info2off, 512), in btt_info_write() [all …]
|
/linux-6.12.1/kernel/bpf/ |
D | arena.c | 53 u64 bpf_arena_get_kern_vm_start(struct bpf_arena *arena) in bpf_arena_get_kern_vm_start() argument 55 return arena ? (u64) (long) arena->kern_vm->addr + GUARD_SZ / 2 : 0; in bpf_arena_get_kern_vm_start() 58 u64 bpf_arena_get_user_vm_start(struct bpf_arena *arena) in bpf_arena_get_user_vm_start() argument 60 return arena ? arena->user_vm_start : 0; in bpf_arena_get_user_vm_start() 88 static long compute_pgoff(struct bpf_arena *arena, long uaddr) in compute_pgoff() argument 90 return (u32)(uaddr - (u32)arena->user_vm_start) >> PAGE_SHIFT; in compute_pgoff() 97 struct bpf_arena *arena; in arena_map_alloc() local 124 arena = bpf_map_area_alloc(sizeof(*arena), numa_node); in arena_map_alloc() 125 if (!arena) in arena_map_alloc() 128 arena->kern_vm = kern_vm; in arena_map_alloc() [all …]
|
D | Makefile | 19 obj-$(CONFIG_BPF_SYSCALL) += arena.o
|
/linux-6.12.1/arch/alpha/kernel/ |
D | pci_iommu.c | 63 struct pci_iommu_arena *arena; in iommu_arena_new_node() local 74 arena = memblock_alloc(sizeof(*arena), SMP_CACHE_BYTES); in iommu_arena_new_node() 75 if (!arena) in iommu_arena_new_node() 77 sizeof(*arena)); in iommu_arena_new_node() 78 arena->ptes = memblock_alloc(mem_size, align); in iommu_arena_new_node() 79 if (!arena->ptes) in iommu_arena_new_node() 83 spin_lock_init(&arena->lock); in iommu_arena_new_node() 84 arena->hose = hose; in iommu_arena_new_node() 85 arena->dma_base = base; in iommu_arena_new_node() 86 arena->size = window_size; in iommu_arena_new_node() [all …]
|
D | core_titan.c | 584 struct pci_iommu_arena *arena; member 601 aper->arena = agp->hose->sg_pci; in titan_agp_setup() 603 aper->pg_start = iommu_reserve(aper->arena, aper->pg_count, in titan_agp_setup() 612 aper->arena->dma_base + aper->pg_start * PAGE_SIZE; in titan_agp_setup() 625 status = iommu_release(aper->arena, aper->pg_start, aper->pg_count); in titan_agp_cleanup() 629 iommu_unbind(aper->arena, aper->pg_start, aper->pg_count); in titan_agp_cleanup() 630 status = iommu_release(aper->arena, aper->pg_start, in titan_agp_cleanup() 686 return iommu_bind(aper->arena, aper->pg_start + pg_start, in titan_agp_bind_memory() 694 return iommu_unbind(aper->arena, aper->pg_start + pg_start, in titan_agp_unbind_memory() 702 unsigned long baddr = addr - aper->arena->dma_base; in titan_agp_translate() [all …]
|
D | core_marvel.c | 854 struct pci_iommu_arena *arena; member 870 aper->arena = agp->hose->sg_pci; in marvel_agp_setup() 872 aper->pg_start = iommu_reserve(aper->arena, aper->pg_count, in marvel_agp_setup() 882 aper->arena->dma_base + aper->pg_start * PAGE_SIZE; in marvel_agp_setup() 895 status = iommu_release(aper->arena, aper->pg_start, aper->pg_count); in marvel_agp_cleanup() 899 iommu_unbind(aper->arena, aper->pg_start, aper->pg_count); in marvel_agp_cleanup() 900 status = iommu_release(aper->arena, aper->pg_start, in marvel_agp_cleanup() 978 return iommu_bind(aper->arena, aper->pg_start + pg_start, in marvel_agp_bind_memory() 986 return iommu_unbind(aper->arena, aper->pg_start + pg_start, in marvel_agp_unbind_memory() 994 unsigned long baddr = addr - aper->arena->dma_base; in marvel_agp_translate() [all …]
|
D | core_cia.c | 356 struct pci_iommu_arena *arena = pci_isa_hose->sg_isa; in verify_tb_operation() local 375 addr0 = arena->dma_base; in verify_tb_operation() 449 cia_pci_tbi(arena->hose, 0, -1); in verify_tb_operation() 464 arena->ptes[4] = pte0; in verify_tb_operation() 488 arena->ptes[5] = pte0; in verify_tb_operation() 501 arena->align_entry = 4; in verify_tb_operation() 524 arena->ptes[4] = 0; in verify_tb_operation() 525 arena->ptes[5] = 0; in verify_tb_operation() 539 alpha_mv.mv_pci_tbi(arena->hose, 0, -1); in verify_tb_operation()
|
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/ |
D | arena_atomics.c | 19 ASSERT_EQ(skel->arena->add64_value, 3, "add64_value"); in test_add() 20 ASSERT_EQ(skel->arena->add64_result, 1, "add64_result"); in test_add() 22 ASSERT_EQ(skel->arena->add32_value, 3, "add32_value"); in test_add() 23 ASSERT_EQ(skel->arena->add32_result, 1, "add32_result"); in test_add() 25 ASSERT_EQ(skel->arena->add_stack_value_copy, 3, "add_stack_value"); in test_add() 26 ASSERT_EQ(skel->arena->add_stack_result, 1, "add_stack_result"); in test_add() 28 ASSERT_EQ(skel->arena->add_noreturn_value, 3, "add_noreturn_value"); in test_add() 44 ASSERT_EQ(skel->arena->sub64_value, -1, "sub64_value"); in test_sub() 45 ASSERT_EQ(skel->arena->sub64_result, 1, "sub64_result"); in test_sub() 47 ASSERT_EQ(skel->arena->sub32_value, -1, "sub32_value"); in test_sub() [all …]
|
D | arena_list.c | 52 ASSERT_EQ(skel->arena->arena_sum, expected_sum, "__arena sum of elems"); in test_arena_list_add_del() 53 ASSERT_EQ(skel->arena->test_val, cnt + 1, "num of elems"); in test_arena_list_add_del() 60 ASSERT_EQ(skel->arena->arena_sum, expected_sum, "__arena sum of elems"); in test_arena_list_add_del()
|
D | arena_htab.c | 46 area = bpf_map__initial_value(skel->maps.arena, &arena_sz); in test_arena_htab_llvm()
|
/linux-6.12.1/tools/testing/selftests/bpf/progs/ |
D | verifier_arena.c | 21 } arena SEC(".maps"); 30 page1 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1() 34 page2 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1() 38 no_page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1() 45 bpf_arena_free_pages(&arena, (void __arena *)page2, 1); in basic_alloc1() 50 page3 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in basic_alloc1() 69 page1 = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in basic_alloc2() 87 bpf_arena_free_pages(&arena, (void __arena *)page1, 2); in basic_alloc2() 108 struct bpf_arena___l *ar = (struct bpf_arena___l *)&arena; in basic_alloc3()
|
D | verifier_arena_large.c | 18 } arena SEC(".maps"); 28 page1 = base = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in big_alloc1() 32 page2 = bpf_arena_alloc_pages(&arena, base + ARENA_SIZE - PAGE_SIZE, in big_alloc1() 37 no_page = bpf_arena_alloc_pages(&arena, base + ARENA_SIZE, in big_alloc1() 45 bpf_arena_free_pages(&arena, (void __arena *)page1, 1); in big_alloc1() 50 page3 = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in big_alloc1()
|
D | arena_atomics.c | 19 } arena SEC(".maps"); 217 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in uaf() 218 bpf_arena_free_pages(&arena, page, 1); in uaf()
|
D | arena_htab.c | 14 } arena SEC(".maps");
|
D | arena_list.c | 19 } arena SEC(".maps");
|
/linux-6.12.1/arch/powerpc/platforms/pseries/ |
D | rtas-work-area.c | 42 char *arena; member 137 const phys_addr_t pa_start = __pa(rwa_state.arena); in rtas_work_area_allocator_init() 144 if (!rwa_state.arena) in rtas_work_area_allocator_init() 157 err = gen_pool_add(pool, (unsigned long)rwa_state.arena, in rtas_work_area_allocator_init() 209 rwa_state.arena = memblock_alloc_try_nid(size, align, min, limit, nid); in rtas_work_area_reserve_arena()
|
/linux-6.12.1/Documentation/driver-api/nvdimm/ |
D | btt.rst | 35 Each arena follows the same layout for its metadata, and all references in an 36 arena are internal to it (with the exception of one field that points to the 37 next arena). The following depicts the "On-disk" metadata layout:: 105 ABA Arena Block Address - Block offset/number within an arena 106 Premap ABA The block offset into an arena, which was decided upon by range 112 arena. 117 the external LBA at 768G. This falls into the second arena, and of the 512G 118 worth of blocks that this arena contributes, this block is at 256G. Thus, the 157 While 'nfree' describes the number of concurrent IOs an arena can process 183 The RTT is a simple, per arena table with 'nfree' entries. Every reader inserts [all …]
|
/linux-6.12.1/tools/testing/selftests/bpf/ |
D | bpf_arena_alloc.h | 33 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in bpf_alloc() 62 bpf_arena_free_pages(&arena, addr, 1); in bpf_free()
|
D | bpf_arena_common.h | 57 __weak char arena[1]; variable
|
D | bpf_arena_htab.h | 95 void __arena *buckets = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in htab_init()
|
/linux-6.12.1/arch/riscv/net/ |
D | bpf_jit_core.c | 83 ctx->arena_vm_start = bpf_arena_get_kern_vm_start(prog->aux->arena); in bpf_int_jit_compile() 84 ctx->user_vm_start = bpf_arena_get_user_vm_start(prog->aux->arena); in bpf_int_jit_compile()
|
/linux-6.12.1/arch/arm64/net/ |
D | bpf_jit_comp.c | 647 const bool arena = BPF_MODE(code) == BPF_PROBE_ATOMIC; in emit_lse_atomic() local 651 if (off || arena) { in emit_lse_atomic() 657 if (arena) { in emit_lse_atomic() 1801 ctx.user_vm_start = bpf_arena_get_user_vm_start(prog->aux->arena); in bpf_int_jit_compile() 1802 ctx.arena_vm_start = bpf_arena_get_kern_vm_start(prog->aux->arena); in bpf_int_jit_compile()
|
/linux-6.12.1/arch/x86/net/ |
D | bpf_jit_comp.c | 746 if (bpf_arena_get_kern_vm_start(bpf_prog->aux->arena)) in emit_bpf_tail_call_indirect() 813 if (bpf_arena_get_kern_vm_start(bpf_prog->aux->arena)) in emit_bpf_tail_call_direct() 1430 arena_vm_start = bpf_arena_get_kern_vm_start(bpf_prog->aux->arena); in do_jit() 1431 user_vm_start = bpf_arena_get_user_vm_start(bpf_prog->aux->arena); in do_jit()
|
/linux-6.12.1/tools/bpf/bpftool/Documentation/ |
D | bpftool-map.rst | 58 | | **task_storage** | **bloom_filter** | **user_ringbuf** | **cgrp_storage** | **arena** }
|
/linux-6.12.1/include/linux/ |
D | bpf.h | 543 u64 bpf_arena_get_kern_vm_start(struct bpf_arena *arena); 544 u64 bpf_arena_get_user_vm_start(struct bpf_arena *arena); 1494 struct bpf_arena *arena; member
|