Searched refs:PAGE_END (Results 1 – 9 of 9) sorted by relevance
/linux-6.12.1/arch/arm64/include/asm/ |
D | memory.h | 106 #define PAGE_END KASAN_SHADOW_START macro 110 #define PAGE_END (_PAGE_END(VA_BITS_MIN)) macro 113 #define PHYSMEM_END __pa(PAGE_END - 1) 333 #define __is_lm_address(addr) (((u64)(addr) - PAGE_OFFSET) < (PAGE_END - PAGE_OFFSET))
|
/linux-6.12.1/scripts/gdb/linux/ |
D | mm.py | 119 …self.PAGE_END = self.KASAN_SHADOW_END - (1 << (self.vabits_actual - self.KASAN_SHADOW_SCALE_SHIFT)) 121 self.PAGE_END = self._PAGE_END(self.VA_BITS_MIN) 195 if (addr - self.PAGE_OFFSET) < (self.PAGE_END - self.PAGE_OFFSET):
|
/linux-6.12.1/arch/arm64/kernel/ |
D | machine_kexec.c | 135 rc = trans_pgd_create_copy(&info, &trans_pgd, PAGE_OFFSET, PAGE_END); in machine_kexec_post_load()
|
D | hibernate.c | 419 PAGE_END); in swsusp_arch_resume()
|
/linux-6.12.1/arch/arm64/mm/ |
D | ptdump.c | 334 { PAGE_END, "Linear Mapping end" }, in ptdump_init()
|
D | init.c | 189 s64 linear_region_size = PAGE_END - _PAGE_OFFSET(vabits_actual); in arm64_memblock_init()
|
D | mmu.c | 1300 WARN_ON((start < PAGE_OFFSET) || (end > PAGE_END)); in __remove_pgd_mapping() 1303 free_empty_tables(start, end, PAGE_OFFSET, PAGE_END); in __remove_pgd_mapping() 1310 u64 end_linear_pa = __pa(PAGE_END - 1); in arch_get_mappable_range()
|
D | kasan_init.c | 345 kasan_populate_early_shadow(kasan_mem_to_shadow((void *)PAGE_END), in kasan_init_shadow()
|
/linux-6.12.1/arch/riscv/mm/ |
D | init.c | 1497 #define PAGE_END KASAN_SHADOW_START macro 1506 preallocate_pgd_pages_range(PAGE_OFFSET, PAGE_END, "direct map"); in pgtable_cache_init() 1778 mhp_range.end = __pa(PAGE_END - 1); in arch_get_mappable_range()
|