/linux-6.12.1/arch/loongarch/mm/ |
D | cache.c | 165 [VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 168 [VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID | 171 [VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 174 [VM_EXEC] = __pgprot(_CACHE_CC | _PAGE_VALID | 176 [VM_EXEC | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 178 [VM_EXEC | VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID | 180 [VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 185 [VM_SHARED | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 188 [VM_SHARED | VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID | 191 [VM_SHARED | VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | [all …]
|
D | tlbex.S | 98 ori t0, t0, _PAGE_VALID 135 ori t0, ra, _PAGE_VALID 138 ori t0, ra, _PAGE_VALID 140 ori t0, ra, _PAGE_VALID 255 ori t0, t0, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED) 293 ori t0, ra, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED) 296 ori t0, ra, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED) 298 ori t0, ra, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED) 413 ori t0, t0, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED) 450 ori t0, ra, (_PAGE_VALID | _PAGE_DIRTY | _PAGE_MODIFIED) [all …]
|
D | pageattr.c | 217 return __set_memory(addr, 1, __pgprot(0), __pgprot(_PAGE_PRESENT | _PAGE_VALID)); in set_direct_map_invalid_noflush()
|
/linux-6.12.1/arch/alpha/include/asm/ |
D | pgtable.h | 62 #define _PAGE_VALID 0x0001 macro 95 #define _PAGE_TABLE (_PAGE_VALID | __DIRTY_BITS | __ACCESS_BITS) 102 #define PAGE_NONE __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOR | _PAGE_FOW | _PAGE_FOE) 103 #define PAGE_SHARED __pgprot(_PAGE_VALID | __ACCESS_BITS) 104 #define PAGE_COPY __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW) 105 #define PAGE_READONLY __pgprot(_PAGE_VALID | __ACCESS_BITS | _PAGE_FOW) 106 #define PAGE_KERNEL __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE | _PAGE_KWE) 108 #define _PAGE_NORMAL(x) __pgprot(_PAGE_VALID | __ACCESS_BITS | (x)) 232 extern inline int pte_present(pte_t pte) { return pte_val(pte) & _PAGE_VALID; } in pte_present() 240 extern inline int pmd_present(pmd_t pmd) { return pmd_val(pmd) & _PAGE_VALID; } in pmd_present() [all …]
|
/linux-6.12.1/arch/loongarch/include/asm/ |
D | kfence.h | 60 set_pte(pte, __pte(pte_val(ptep_get(pte)) & ~(_PAGE_VALID | _PAGE_PRESENT))); in kfence_protect_page() 62 set_pte(pte, __pte(pte_val(ptep_get(pte)) | (_PAGE_VALID | _PAGE_PRESENT))); in kfence_protect_page()
|
D | pgtable-bits.h | 45 #define _PAGE_VALID (_ULCAST_(1) << _PAGE_VALID_SHIFT) macro 76 #define __READABLE (_PAGE_VALID)
|
D | pgtable.h | 576 pmd_val(pmd) &= ~(_PAGE_PRESENT | _PAGE_VALID | _PAGE_DIRTY | _PAGE_PROTNONE); in pmd_mkinvalid()
|
/linux-6.12.1/arch/csky/include/asm/ |
D | pgtable.h | 62 #define PAGE_KERNEL __pgprot(_PAGE_BASE | _PAGE_READ | _PAGE_VALID | \ 67 #define _PAGE_IOREMAP (_PAGE_BASE | _PAGE_READ | _PAGE_VALID | \ 175 pte_val(pte) &= ~(_PAGE_ACCESSED|_PAGE_VALID); in pte_mkold() 199 pte_val(pte) |= _PAGE_VALID; in pte_mkyoung()
|
/linux-6.12.1/arch/sparc/mm/ |
D | init_64.c | 2230 kern_linear_pte_xor[1] = (_PAGE_VALID | _PAGE_SZ256MB_4V) ^ in sun4v_linear_pte_xor_finalize() 2239 kern_linear_pte_xor[2] = (_PAGE_VALID | _PAGE_SZ2GB_4V) ^ in sun4v_linear_pte_xor_finalize() 2248 kern_linear_pte_xor[3] = (_PAGE_VALID | _PAGE_SZ16GB_4V) ^ in sun4v_linear_pte_xor_finalize() 2593 pte_base = (_PAGE_VALID | _PAGE_SZ4MB_4U | in vmemmap_populate() 2597 pte_base = (_PAGE_VALID | _PAGE_SZ4MB_4V | in vmemmap_populate() 2624 if (!(pte & _PAGE_VALID)) { in vmemmap_populate() 2674 PAGE_KERNEL = __pgprot (_PAGE_PRESENT_4U | _PAGE_VALID | in sun4u_pgprot_init() 2678 PAGE_KERNEL_LOCKED = __pgprot (_PAGE_PRESENT_4U | _PAGE_VALID | in sun4u_pgprot_init() 2687 pg_iobits = (_PAGE_VALID | _PAGE_PRESENT_4U | __DIRTY_BITS_4U | in sun4u_pgprot_init() 2691 kern_linear_pte_xor[0] = _PAGE_VALID ^ PAGE_OFFSET; in sun4u_pgprot_init() [all …]
|
D | tlb.c | 158 if (pte_val(*pte) & _PAGE_VALID) { in tlb_batch_pmd_scan() 253 entry = __pmd(pmd_val(*pmdp) & ~_PAGE_VALID); in pmdp_invalidate()
|
/linux-6.12.1/arch/mips/include/asm/ |
D | pgtable-bits.h | 180 #define _PAGE_VALID (1 << _PAGE_VALID_SHIFT) macro 195 #define _PAGE_SILENT_READ _PAGE_VALID
|
D | pgtable.h | 734 pmd_val(pmd) &= ~(_PAGE_PRESENT | _PAGE_VALID | _PAGE_DIRTY);
|
/linux-6.12.1/arch/csky/abiv2/inc/abi/ |
D | pgtable-bits.h | 18 #define _PAGE_VALID (1<<1) macro
|
/linux-6.12.1/arch/csky/abiv1/inc/abi/ |
D | pgtable-bits.h | 18 #define _PAGE_VALID (1<<7) macro
|
/linux-6.12.1/arch/mips/mm/ |
D | tlbex.c | 1638 unsigned int hwmode = mode & (_PAGE_VALID | _PAGE_DIRTY); in iPTE_SW() 1744 unsigned int mode = _PAGE_VALID | _PAGE_ACCESSED; in build_make_valid() 1784 unsigned int mode = (_PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_VALID in build_make_write() 2090 uasm_il_bbit0(&p, &r, wr.r1, ilog2(_PAGE_VALID), in build_r4000_tlb_load_handler() 2093 uasm_i_andi(&p, wr.r3, wr.r1, _PAGE_VALID); in build_r4000_tlb_load_handler() 2156 uasm_il_bbit0(&p, &r, wr.r1, ilog2(_PAGE_VALID), in build_r4000_tlb_load_handler() 2159 uasm_i_andi(&p, wr.r3, wr.r1, _PAGE_VALID); in build_r4000_tlb_load_handler() 2208 uasm_i_ori(&p, wr.r1, wr.r1, (_PAGE_ACCESSED | _PAGE_VALID)); in build_r4000_tlb_load_handler() 2264 _PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_VALID | _PAGE_DIRTY); in build_r4000_tlb_store_handler() 2321 _PAGE_ACCESSED | _PAGE_MODIFIED | _PAGE_VALID | _PAGE_DIRTY); in build_r4000_tlb_modify_handler()
|
D | c-r4k.c | 568 if ((mm == current->active_mm) && (pte_val(*ptep) & _PAGE_VALID)) in local_r4k_flush_cache_page()
|
/linux-6.12.1/arch/alpha/kernel/ |
D | proto.h | 200 prot = __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE in __alpha_remap_area_pages()
|
/linux-6.12.1/arch/hexagon/include/asm/ |
D | pgtable.h | 54 #define _PAGE_VALID _PAGE_PRESENT macro
|
/linux-6.12.1/arch/sparc/include/asm/ |
D | pgtable_64.h | 112 #define _PAGE_VALID _AC(0x8000000000000000,UL) /* Valid TTE */ macro 675 return pte_val(a) & _PAGE_VALID; in pte_accessible()
|