/linux-6.12.1/mm/ |
D | memory-tiers.c | 370 for_each_node_state(node, N_MEMORY) { in disable_all_demotion_targets() 393 for_each_node_state(node, N_MEMORY) { in dump_demotion_targets() 429 for_each_node_state(node, N_MEMORY) { in establish_demotion_targets() 447 nodes_andnot(tier_nodes, node_states[N_MEMORY], tier_nodes); in establish_demotion_targets() 495 lower_tier = node_states[N_MEMORY]; in establish_demotion_targets() 542 if (!node_state(node, N_MEMORY)) in set_node_memory_tier() 706 for_each_node_state(nid, N_MEMORY) { in memory_tier_late_init() 929 nodes_and(default_dram_nodes, node_states[N_MEMORY], in memory_tier_init()
|
D | mm_init.c | 309 node_set_state(nid, N_MEMORY); in early_calculate_totalpages() 347 nodemask_t saved_node_state = node_states[N_MEMORY]; in find_zone_movable_pfns_for_nodes() 349 int usable_nodes = nodes_weight(node_states[N_MEMORY]); in find_zone_movable_pfns_for_nodes() 462 for_each_node_state(nid, N_MEMORY) { in find_zone_movable_pfns_for_nodes() 562 node_states[N_MEMORY] = saved_node_state; in find_zone_movable_pfns_for_nodes() 646 if (num_node_state(N_MEMORY) == 1) in fixup_hashdist() 1853 node_set_state(nid, N_MEMORY); in free_area_init() 2281 atomic_set(&pgdat_init_n_undone, num_node_state(N_MEMORY)); in page_alloc_init_late() 2282 for_each_node_state(nid, N_MEMORY) { in page_alloc_init_late() 2306 for_each_node_state(nid, N_MEMORY) in page_alloc_init_late()
|
D | hugetlb_cgroup.c | 481 for_each_node_state(nid, N_MEMORY) in hugetlb_cgroup_read_numa_stat() 486 for_each_node_state(nid, N_MEMORY) in hugetlb_cgroup_read_numa_stat() 504 for_each_node_state(nid, N_MEMORY) { in hugetlb_cgroup_read_numa_stat()
|
D | memory_hotplug.c | 684 if (!node_state(nid, N_MEMORY)) in node_states_check_changes_online() 696 node_set_state(node, N_MEMORY); in node_states_set_node() 1830 nodemask_t nmask = node_states[N_MEMORY]; in do_migrate_range() 1920 node_clear_state(node, N_MEMORY); in node_states_clear_node()
|
D | page_ext.c | 467 for_each_node_state(nid, N_MEMORY) { in page_ext_init()
|
D | hugetlb.c | 2586 folio = remove_pool_hugetlb_folio(h, &node_states[N_MEMORY], 1); in return_unused_surplus_pages() 3133 for_each_node_mask_to_alloc(&h->next_nid_to_alloc, nr_nodes, node, &node_states[N_MEMORY]) { in __alloc_bootmem_huge_page() 3290 .size = num_node_state(N_MEMORY), in gather_bootmem_prealloc() 3293 .max_threads = num_node_state(N_MEMORY), in gather_bootmem_prealloc() 3314 &node_states[N_MEMORY]); in hugetlb_hstate_alloc_pages_onenode() 3370 struct folio *folio = alloc_pool_huge_folio(h, &node_states[N_MEMORY], in hugetlb_pages_alloc_boot_node() 3426 job.max_threads = num_node_state(N_MEMORY) * 2; in hugetlb_pages_alloc_boot() 3427 job.min_chunk = h->max_huge_pages / num_node_state(N_MEMORY) / 2; in hugetlb_pages_alloc_boot() 3948 n_mask = &node_states[N_MEMORY]; in __nr_hugepages_store_common() 4107 n_mask = &node_states[N_MEMORY]; in demote_store()
|
D | mempolicy.c | 277 cpuset_current_mems_allowed, node_states[N_MEMORY]); in mpol_set_nodemask() 3118 for_each_node_state(nid, N_MEMORY) { in numa_policy_init() 3190 if (!nodes_subset(nodes, node_states[N_MEMORY])) in mpol_parse_str() 3222 nodes = node_states[N_MEMORY]; in mpol_parse_str()
|
D | vmstat.c | 1686 if (!node_state(pgdat->node_id, N_MEMORY)) in pagetypeinfo_show() 2261 if (!node_state(pgdat->node_id, N_MEMORY)) in unusable_show()
|
D | oom_kill.c | 287 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc()
|
D | compaction.c | 3267 for_each_node_state(nid, N_MEMORY) { in kcompactd_cpu_online() 3350 for_each_node_state(nid, N_MEMORY) in kcompactd_init()
|
D | memcontrol-v1.c | 2694 for_each_node_state(nid, N_MEMORY) in memcg_numa_stat_show() 2706 for_each_node_state(nid, N_MEMORY) in memcg_numa_stat_show()
|
D | vmscan.c | 2824 for_each_node_state(nid, N_MEMORY) { in lru_gen_add_mm() 5213 for_each_node_state(nid, N_MEMORY) { in lru_gen_seq_start() 5424 if (nid < 0 || nid >= MAX_NUMNODES || !node_state(nid, N_MEMORY)) in run_cmd() 7378 for_each_node_state(nid, N_MEMORY) in kswapd_init()
|
/linux-6.12.1/Documentation/translations/zh_CN/core-api/ |
D | memory-hotplug.rst | 76 - status_change_nid是当nodemask的N_MEMORY被(将)设置/清除时设置的节点id。这
|
/linux-6.12.1/include/linux/ |
D | nodemask.h | 407 N_MEMORY, /* The node has memory(regular, high, movable) */ enumerator 445 #define first_memory_node first_node(node_states[N_MEMORY]) 452 return next_node(nid, node_states[N_MEMORY]); in next_memory_node()
|
D | cpuset.h | 216 #define cpuset_current_mems_allowed (node_states[N_MEMORY])
|
/linux-6.12.1/drivers/base/ |
D | node.c | 931 [N_MEMORY] = _NODE_ATTR(has_memory, N_MEMORY), 944 &node_state_attr[N_MEMORY].attr.attr,
|
/linux-6.12.1/Documentation/core-api/ |
D | memory-hotplug.rst | 67 - status_change_nid is set node id when N_MEMORY of nodemask is (will be)
|
/linux-6.12.1/Documentation/mm/ |
D | physical_memory.rst | 170 ``N_MEMORY`` 184 node_states[N_MEMORY]
|
/linux-6.12.1/kernel/cgroup/ |
D | cpuset.c | 378 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY])) in guarantee_online_mems() 380 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]); in guarantee_online_mems() 3831 new_mems = node_states[N_MEMORY]; in cpuset_handle_hotplug() 3945 top_cpuset.effective_mems = node_states[N_MEMORY]; in cpuset_init_smp()
|
/linux-6.12.1/fs/proc/ |
D | kcore.c | 255 for_each_node_state(nid, N_MEMORY) { in kcore_ram_list()
|
D | task_mmu.c | 2899 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats() 2924 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats_pmd() 3073 for_each_node_state(nid, N_MEMORY) in show_numa_map()
|
/linux-6.12.1/arch/x86/kernel/cpu/sgx/ |
D | main.c | 853 node_state(nid, N_MEMORY) && node_state(nid, N_CPU)) in sgx_page_cache_init()
|
/linux-6.12.1/init/ |
D | main.c | 1558 set_mems_allowed(node_states[N_MEMORY]); in kernel_init_freeable()
|
/linux-6.12.1/kernel/ |
D | kthread.c | 747 set_mems_allowed(node_states[N_MEMORY]); in kthreadd()
|
/linux-6.12.1/arch/x86/mm/ |
D | init_64.c | 832 node_clear_state(0, N_MEMORY); in paging_init()
|