/linux-6.12.1/include/linux/ |
D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
D | mmzone.h | 1246 struct zonelist { struct 1311 struct zonelist node_zonelists[MAX_ZONELISTS]; 1686 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1690 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist() 1733 struct zonelist *zonelist; in movable_only_nodes() local 1746 zonelist = &NODE_DATA(nid)->node_zonelists[ZONELIST_FALLBACK]; in movable_only_nodes() 1747 z = first_zones_zonelist(zonelist, ZONE_NORMAL, nodes); in movable_only_nodes()
|
D | swap.h | 408 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
D | gfp.h | 193 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
/linux-6.12.1/Documentation/translations/zh_CN/mm/ |
D | numa.rst | 57 中的一个或多个]构建了一个有序的“区列表”。zonelist指定了当一个选定的区/节点不能满足分配请求 63 代表了相对稀缺的资源。Linux选择了一个默认的Node ordered zonelist。这意味着在使用按NUMA距
|
/linux-6.12.1/Documentation/mm/ |
D | numa.rst | 74 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 84 a default Node ordered zonelist. This means it tries to fallback to other zones 89 Linux will attempt to allocate from the first node in the appropriate zonelist 92 nodes' zones in the selected zonelist looking for the first zone in the list 120 zonelist--will not be the node itself. Rather, it will be the node that the
|
/linux-6.12.1/mm/ |
D | page_alloc.c | 2087 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 2095 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx, in unreserve_highatomic_pageblock() 3577 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 3819 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 3927 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 3981 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx, in wake_all_kswapds() 4122 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4261 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4272 struct zoneref *z = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4362 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() [all …]
|
D | vmscan.c | 6122 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 6143 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 6245 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 6263 shrink_zones(zonelist, sc); in do_try_to_free_pages() 6280 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 6398 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 6436 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 6478 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 6507 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 6513 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
D | oom_kill.c | 271 if (!oc->zonelist) in constrained_alloc() 295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
D | mm_init.c | 54 struct zonelist *zonelist; in mminit_verify_zonelist() local 62 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 73 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
D | internal.h | 513 struct zonelist *zonelist; member
|
D | mempolicy.c | 1951 struct zonelist *zonelist; in mempolicy_slab_node() local 1953 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1954 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
D | compaction.c | 2462 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2822 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
D | hugetlb.c | 1354 struct zonelist *zonelist; in dequeue_hugetlb_folio_nodemask() local 1363 zonelist = node_zonelist(nid, gfp_mask); in dequeue_hugetlb_folio_nodemask() 1367 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_hugetlb_folio_nodemask()
|
D | slub.c | 2871 struct zonelist *zonelist; in get_any_partial() local 2902 zonelist = node_zonelist(mempolicy_slab_node(), pc->flags); in get_any_partial() 2903 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
D | memcontrol.c | 1585 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
/linux-6.12.1/drivers/tty/ |
D | sysrq.c | 389 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
/linux-6.12.1/Documentation/admin-guide/sysctl/ |
D | vm.rst | 702 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 708 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
/linux-6.12.1/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 234 node zonelist.
|
/linux-6.12.1/Documentation/admin-guide/ |
D | kernel-parameters.txt | 4191 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|