/linux-6.12.1/arch/powerpc/mm/ |
D | mem.c | 339 #ifdef MODULES_VADDR in mem_init() 341 MODULES_VADDR, MODULES_END); in mem_init() 421 for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE) in prealloc_execmem_pgtable() 439 #ifdef MODULES_VADDR in execmem_arch_setup() 442 BUILD_BUG_ON(TASK_SIZE > MODULES_VADDR); in execmem_arch_setup() 445 if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) { in execmem_arch_setup() 447 fallback_start = MODULES_VADDR; in execmem_arch_setup() 450 start = MODULES_VADDR; in execmem_arch_setup()
|
/linux-6.12.1/arch/arm/include/asm/ |
D | memory.h | 60 #define MODULES_VADDR (PAGE_OFFSET - SZ_16M) macro 63 #define MODULES_VADDR (PAGE_OFFSET - SZ_8M) macro 66 #if TASK_SIZE > MODULES_VADDR 84 #define XIP_VIRT_ADDR(physaddr) (MODULES_VADDR + ((physaddr) & 0x000fffff)) 127 #define MODULES_VADDR PAGE_OFFSET macro
|
/linux-6.12.1/arch/arm/mm/ |
D | pgd.c | 57 new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR), in pgd_alloc() 58 MODULES_VADDR); in pgd_alloc() 62 new_pud = pud_alloc(mm, new_p4d, MODULES_VADDR); in pgd_alloc() 79 (pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE)) in pgd_alloc()
|
D | init.c | 295 BUILD_BUG_ON(TASK_SIZE > MODULES_VADDR); in mem_init() 296 BUG_ON(TASK_SIZE > MODULES_VADDR); in mem_init() 500 #undef MODULES_VADDR 501 #define MODULES_VADDR (((unsigned long)_exiprom + ~PMD_MASK) & PMD_MASK) macro 519 .start = MODULES_VADDR, in execmem_arch_setup()
|
D | pageattr.c | 66 if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) && in change_memory_common()
|
/linux-6.12.1/arch/x86/include/asm/ |
D | pgtable_32_areas.h | 47 #define MODULES_VADDR VMALLOC_START macro 49 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
|
D | pgtable_64_types.h | 194 #define MODULES_VADDR (__START_KERNEL_map + KERNEL_IMAGE_SIZE) macro 201 #define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
/linux-6.12.1/arch/loongarch/include/asm/ |
D | kasan.h | 53 #define XKVRANGE_VC_START MODULES_VADDR 54 #define XKVRANGE_VC_SIZE round_up(KFENCE_AREA_END - MODULES_VADDR + 1, PGDIR_SIZE)
|
/linux-6.12.1/arch/x86/kernel/ |
D | head64.c | 431 BUILD_BUG_ON(MODULES_VADDR < __START_KERNEL_map); in x86_64_start_kernel() 432 BUILD_BUG_ON(MODULES_VADDR - __START_KERNEL_map < KERNEL_IMAGE_SIZE); in x86_64_start_kernel() 435 BUILD_BUG_ON((MODULES_VADDR & ~PMD_MASK) != 0); in x86_64_start_kernel() 436 BUILD_BUG_ON(!(MODULES_VADDR > __START_KERNEL)); in x86_64_start_kernel()
|
/linux-6.12.1/arch/powerpc/mm/ptdump/ |
D | ptdump.c | 77 #ifdef MODULES_VADDR 266 #ifdef MODULES_VADDR in populate_markers() 267 address_markers[i++].start_address = MODULES_VADDR; in populate_markers()
|
/linux-6.12.1/Documentation/arch/arm/ |
D | memory.rst | 76 MODULES_VADDR MODULES_END-1 Kernel module space 80 TASK_SIZE MODULES_VADDR-1 KASAn shadow memory when KASan is in use. 81 The range from MODULES_VADDR to the top
|
/linux-6.12.1/arch/sparc/mm/ |
D | execmem.c | 12 .start = MODULES_VADDR, in execmem_arch_setup()
|
/linux-6.12.1/arch/powerpc/mm/kasan/ |
D | init_32.c | 119 #ifdef MODULES_VADDR in kasan_unmap_early_shadow_vmalloc() 120 k_start = (unsigned long)kasan_mem_to_shadow((void *)MODULES_VADDR); in kasan_unmap_early_shadow_vmalloc()
|
/linux-6.12.1/arch/riscv/kernel/ |
D | vmcore_info.c | 17 vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR); in arch_crash_save_vmcoreinfo()
|
/linux-6.12.1/arch/arm64/kernel/ |
D | vmcore_info.c | 24 vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR); in arch_crash_save_vmcoreinfo()
|
/linux-6.12.1/arch/s390/boot/ |
D | startup.c | 30 unsigned long __bootdata_preserved(MODULES_VADDR); 303 MODULES_VADDR = MODULES_END - MODULES_LEN; in setup_kernel_memory_layout() 304 VMALLOC_END = MODULES_VADDR; in setup_kernel_memory_layout()
|
/linux-6.12.1/arch/mips/mm/ |
D | fault.c | 86 #ifdef MODULES_VADDR in __do_page_fault() 87 if (unlikely(address >= MODULES_VADDR && address < MODULES_END)) in __do_page_fault()
|
/linux-6.12.1/fs/proc/ |
D | kcore.c | 703 #if defined(CONFIG_MODULES) && defined(MODULES_VADDR) 710 if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) { in add_modules_range() 711 kclist_add(&kcore_modules, (void *)MODULES_VADDR, in add_modules_range() 712 MODULES_END - MODULES_VADDR, KCORE_VMALLOC); in add_modules_range()
|
/linux-6.12.1/arch/arm64/include/asm/ |
D | memory.h | 47 #define MODULES_END (MODULES_VADDR + MODULES_VSIZE) 48 #define MODULES_VADDR (_PAGE_END(VA_BITS_MIN)) macro
|
/linux-6.12.1/arch/powerpc/mm/book3s32/ |
D | mmu.c | 189 if (addr < ALIGN_DOWN(MODULES_VADDR, SZ_256M)) in is_module_segment() 226 BUILD_BUG_ON(ALIGN_DOWN(MODULES_VADDR, SZ_256M) < TASK_SIZE); in mmu_mark_initmem_nx()
|
/linux-6.12.1/arch/um/include/asm/ |
D | pgtable.h | 52 #define MODULES_VADDR VMALLOC_START macro 54 #define MODULES_LEN (MODULES_VADDR - MODULES_END)
|
/linux-6.12.1/arch/nios2/mm/ |
D | init.c | 160 .start = MODULES_VADDR, in execmem_arch_setup()
|
/linux-6.12.1/mm/kmsan/ |
D | kmsan.h | 180 return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END); in kmsan_internal_is_module_addr()
|
/linux-6.12.1/arch/powerpc/include/asm/nohash/32/ |
D | mmu-8xx.h | 175 #define MODULES_VADDR (MODULES_END - MODULES_SIZE) macro
|
/linux-6.12.1/Documentation/arch/s390/ |
D | mm.rst | 91 +- MODULES_VADDR --+ modules area start
|