Lines Matching refs:PMD_SIZE
102 if (map_page_size == PMD_SIZE) { in early_map_kernel_page()
165 if (map_page_size == PMD_SIZE) { in __map_kernel_page()
329 } else if (IS_ALIGNED(addr, PMD_SIZE) && gap >= PMD_SIZE && in create_physical_mapping()
331 mapping_size = PMD_SIZE; in create_physical_mapping()
752 unsigned long start = ALIGN_DOWN(addr, PMD_SIZE); in vmemmap_pmd_is_unused()
754 return !vmemmap_populated(start, PMD_SIZE); in vmemmap_pmd_is_unused()
847 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
848 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
850 free_vmemmap_pages(pmd_page(*pmd), altmap, get_order(PMD_SIZE)); in remove_pmd_table()
856 free_vmemmap_pages(pmd_page(*pmd), altmap, get_order(PMD_SIZE)); in remove_pmd_table()
1017 VM_BUG_ON(!IS_ALIGNED(addr, PMD_SIZE)); in vmemmap_set_pmd()
1160 if (altmap && (!IS_ALIGNED(addr, PMD_SIZE) || in radix__vmemmap_populate()
1161 altmap_cross_boundary(altmap, addr, PMD_SIZE))) { in radix__vmemmap_populate()
1169 p = vmemmap_alloc_block_buf(PMD_SIZE, node, altmap); in radix__vmemmap_populate()
1328 addr_pfn += (PMD_SIZE >> PAGE_SHIFT); in vmemmap_populate_compound_pages()
1681 flush_tlb_kernel_range(addr, addr + PMD_SIZE); in pmd_free_pte_page()