/linux-6.12.1/arch/x86/mm/ |
D | init.c | 325 unsigned long start_pfn, unsigned long end_pfn, in save_mr() argument 328 if (start_pfn < end_pfn) { in save_mr() 332 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 403 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 419 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 421 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 423 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 425 if (end_pfn > limit_pfn) in split_mem_range() 426 end_pfn = limit_pfn; in split_mem_range() 427 if (start_pfn < end_pfn) { in split_mem_range() [all …]
|
D | init_32.c | 256 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local 267 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init() 296 if (pfn >= end_pfn) in kernel_physical_mapping_init() 304 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 344 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init() 399 unsigned long start_pfn, unsigned long end_pfn) in add_highpages_with_active_regions() argument 406 start_pfn, end_pfn); in add_highpages_with_active_regions() 408 start_pfn, end_pfn); in add_highpages_with_active_regions()
|
/linux-6.12.1/mm/ |
D | mm_init.c | 301 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local 304 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages() 305 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages() 463 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local 481 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes() 485 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes() 491 kernel_pages = min(end_pfn, usable_startpfn) in find_zone_movable_pfns_for_nodes() 500 if (end_pfn <= usable_startpfn) { in find_zone_movable_pfns_for_nodes() 508 zone_movable_pfn[nid] = end_pfn; in find_zone_movable_pfns_for_nodes() 519 size_pages = end_pfn - start_pfn; in find_zone_movable_pfns_for_nodes() [all …]
|
D | page_isolation.c | 33 static struct page *has_unmovable_pages(unsigned long start_pfn, unsigned long end_pfn, in has_unmovable_pages() argument 41 pageblock_start_pfn(end_pfn - 1)); in has_unmovable_pages() 55 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in has_unmovable_pages() 148 unsigned long start_pfn, unsigned long end_pfn) in set_migratetype_isolate() argument 179 end_pfn); in set_migratetype_isolate() 480 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 487 unsigned long isolate_end = pageblock_align(end_pfn); in start_isolate_page_range() 514 start_pfn, end_pfn)) { in start_isolate_page_range() 534 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 540 unsigned long isolate_end = pageblock_align(end_pfn); in undo_isolate_page_range() [all …]
|
D | page_idle.c | 120 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 130 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 131 if (end_pfn > max_pfn) in page_idle_bitmap_read() 132 end_pfn = max_pfn; in page_idle_bitmap_read() 134 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 165 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 175 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 176 if (end_pfn > max_pfn) in page_idle_bitmap_write() 177 end_pfn = max_pfn; in page_idle_bitmap_write() 179 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_write()
|
D | memory_hotplug.c | 372 const unsigned long end_pfn = pfn + nr_pages; in __add_pages() local 399 for (; pfn < end_pfn; pfn += cur_nr_pages) { in __add_pages() 401 cur_nr_pages = min(end_pfn - pfn, in __add_pages() 416 unsigned long end_pfn) in find_smallest_section_pfn() argument 418 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 437 unsigned long end_pfn) in find_biggest_section_pfn() argument 442 pfn = end_pfn - 1; in find_biggest_section_pfn() 460 unsigned long end_pfn) in shrink_zone_span() argument 472 pfn = find_smallest_section_pfn(nid, zone, end_pfn, in shrink_zone_span() 481 } else if (zone_end_pfn(zone) == end_pfn) { in shrink_zone_span() [all …]
|
D | compaction.c | 590 unsigned long end_pfn, in isolate_freepages_block() argument 609 for (; blockpfn < end_pfn; blockpfn += stride, page += stride) { in isolate_freepages_block() 633 if (blockpfn + (1UL << order) <= end_pfn) { in isolate_freepages_block() 688 if (unlikely(blockpfn > end_pfn)) in isolate_freepages_block() 689 blockpfn = end_pfn; in isolate_freepages_block() 702 if (strict && blockpfn < end_pfn) in isolate_freepages_block() 727 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument 741 for (; pfn < end_pfn; pfn += isolated, in isolate_freepages_range() 757 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range() 781 if (pfn < end_pfn) { in isolate_freepages_range() [all …]
|
D | sparse.c | 130 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument 141 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 144 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 145 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 148 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 150 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 594 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument 598 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 612 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in offline_mem_sections() argument 616 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections()
|
D | bootmem_info.c | 104 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 115 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 118 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node()
|
/linux-6.12.1/include/trace/events/ |
D | page_isolation.h | 14 unsigned long end_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 21 __field(unsigned long, end_pfn) 27 __entry->end_pfn = end_pfn; 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, 33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
|
D | compaction.h | 18 unsigned long end_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 26 __field(unsigned long, end_pfn) 33 __entry->end_pfn = end_pfn; 40 __entry->end_pfn, 49 unsigned long end_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 60 unsigned long end_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 71 unsigned long end_pfn, [all …]
|
/linux-6.12.1/arch/sh/mm/ |
D | numa.c | 25 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 31 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 38 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 49 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
D | init.c | 210 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 212 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 219 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 224 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 228 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) in do_init_bootmem() 229 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem()
|
/linux-6.12.1/arch/sparc/mm/ |
D | init_32.c | 65 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 67 if (end_pfn <= max_low_pfn) in calc_highpages() 73 nr += end_pfn - start_pfn; in calc_highpages() 235 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 240 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region() 243 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 284 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local 286 if (end_pfn <= highstart_pfn) in mem_init() 292 map_high_region(start_pfn, end_pfn); in mem_init()
|
/linux-6.12.1/arch/mips/loongson64/ |
D | numa.c | 85 unsigned long start_pfn, end_pfn; in node_mem_init() local 91 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 93 node, start_pfn, end_pfn); in node_mem_init() 98 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; in node_mem_init() 108 max_low_pfn = end_pfn; in node_mem_init()
|
/linux-6.12.1/include/linux/ |
D | node.h | 108 unsigned long end_pfn, 112 unsigned long end_pfn, in register_memory_blocks_under_node() argument 132 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node() local 137 register_memory_blocks_under_node(nid, start_pfn, end_pfn, in register_one_node()
|
D | page-isolation.h | 41 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 44 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 47 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
/linux-6.12.1/drivers/base/ |
D | arch_numa.c | 195 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 197 if (start_pfn >= end_pfn) in setup_node_data() 204 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_node_data() 213 unsigned long start_pfn, end_pfn; in numa_register_nodes() local 215 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes() 216 setup_node_data(nid, start_pfn, end_pfn); in numa_register_nodes()
|
/linux-6.12.1/arch/loongarch/kernel/ |
D | numa.c | 192 unsigned long start_pfn, end_pfn; in node_mem_init() local 199 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 201 node, start_pfn, end_pfn); in node_mem_init() 376 unsigned long start_pfn, end_pfn; in paging_init() local 378 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in paging_init() 380 if (end_pfn > max_low_pfn) in paging_init() 381 max_low_pfn = end_pfn; in paging_init()
|
/linux-6.12.1/arch/x86/xen/ |
D | setup.c | 218 unsigned long end_pfn) in xen_set_identity_and_release_chunk() argument 223 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk() 226 end = min(end_pfn, ini_nr_pages); in xen_set_identity_and_release_chunk() 245 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk() 347 unsigned long start_pfn, unsigned long end_pfn, unsigned long remap_pfn) in xen_set_identity_and_remap_chunk() argument 351 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk() 393 for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) in xen_set_identity_and_remap_chunk() 402 unsigned long start_pfn, unsigned long end_pfn, in xen_count_remap_pages() argument 408 return remap_pages + min(end_pfn, ini_nr_pages) - start_pfn; in xen_count_remap_pages() 412 unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, in xen_foreach_remap_area() argument [all …]
|
/linux-6.12.1/drivers/gpu/drm/imagination/ |
D | pvr_vm_mips.c | 157 s32 end_pfn; in pvr_vm_mips_map() local 173 end_pfn = (end & fw_dev->fw_heap_info.offset_mask) >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K; in pvr_vm_mips_map() 182 for (pfn = start_pfn; pfn <= end_pfn; pfn++) { in pvr_vm_mips_map() 230 const u32 end_pfn = (end & fw_dev->fw_heap_info.offset_mask) >> in pvr_vm_mips_unmap() local 233 for (u32 pfn = start_pfn; pfn < end_pfn; pfn++) in pvr_vm_mips_unmap()
|
/linux-6.12.1/arch/sh/kernel/ |
D | swsusp.c | 22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 24 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
D | setup.c | 194 unsigned long end_pfn) in __add_active_range() argument 202 end = end_pfn << PAGE_SHIFT; in __add_active_range() 211 start_pfn, end_pfn); in __add_active_range() 235 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
|
/linux-6.12.1/arch/x86/virt/vmx/tdx/ |
D | tdx.c | 185 unsigned long end_pfn, int nid) in add_tdx_memblock() argument 195 tmb->end_pfn = end_pfn; in add_tdx_memblock() 223 unsigned long start_pfn, end_pfn; in build_tdx_memlist() local 226 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in build_tdx_memlist() 234 if (start_pfn >= end_pfn) in build_tdx_memlist() 242 ret = add_tdx_memblock(tmb_list, start_pfn, end_pfn, nid); in build_tdx_memlist() 426 end = TDMR_ALIGN_UP(PFN_PHYS(tmb->end_pfn)); in fill_out_tdmrs() 525 if (tmb->end_pfn > PHYS_PFN(tdmr->base)) in tdmr_get_nid() 772 end = PFN_PHYS(tmb->end_pfn); in tdmr_populate_rsvd_holes() 1371 static bool is_tdx_memory(unsigned long start_pfn, unsigned long end_pfn) in is_tdx_memory() argument [all …]
|
/linux-6.12.1/arch/x86/platform/efi/ |
D | efi_32.c | 38 u64 start_pfn, end_pfn, end; in efi_map_region() local 45 end_pfn = PFN_UP(end); in efi_map_region() 47 if (pfn_range_is_mapped(start_pfn, end_pfn)) { in efi_map_region()
|