Lines Matching refs:vmemmap_pages

40 	struct list_head	*vmemmap_pages;  member
232 list_add(&page->lru, walk->vmemmap_pages); in vmemmap_remap_pte()
264 page = list_first_entry(walk->vmemmap_pages, struct page, lru); in vmemmap_restore_pte()
320 struct list_head *vmemmap_pages, in vmemmap_remap_free() argument
327 .vmemmap_pages = vmemmap_pages, in vmemmap_remap_free()
345 list_add(&walk.reuse_page->lru, vmemmap_pages); in vmemmap_remap_free()
376 .vmemmap_pages = vmemmap_pages, in vmemmap_remap_free()
426 LIST_HEAD(vmemmap_pages); in vmemmap_remap_alloc()
430 .vmemmap_pages = &vmemmap_pages, in vmemmap_remap_alloc()
437 if (alloc_vmemmap_page_list(start, end, &vmemmap_pages)) in vmemmap_remap_alloc()
561 struct list_head *vmemmap_pages, in __hugetlb_vmemmap_optimize_folio() argument
602 vmemmap_pages, flags); in __hugetlb_vmemmap_optimize_folio()
623 LIST_HEAD(vmemmap_pages); in hugetlb_vmemmap_optimize_folio()
625 __hugetlb_vmemmap_optimize_folio(h, folio, &vmemmap_pages, VMEMMAP_SYNCHRONIZE_RCU); in hugetlb_vmemmap_optimize_folio()
626 free_vmemmap_page_list(&vmemmap_pages); in hugetlb_vmemmap_optimize_folio()
651 LIST_HEAD(vmemmap_pages); in hugetlb_vmemmap_optimize_folios()
672 ret = __hugetlb_vmemmap_optimize_folio(h, folio, &vmemmap_pages, flags); in hugetlb_vmemmap_optimize_folios()
684 if (ret == -ENOMEM && !list_empty(&vmemmap_pages)) { in hugetlb_vmemmap_optimize_folios()
686 free_vmemmap_page_list(&vmemmap_pages); in hugetlb_vmemmap_optimize_folios()
687 INIT_LIST_HEAD(&vmemmap_pages); in hugetlb_vmemmap_optimize_folios()
688 __hugetlb_vmemmap_optimize_folio(h, folio, &vmemmap_pages, flags); in hugetlb_vmemmap_optimize_folios()
693 free_vmemmap_page_list(&vmemmap_pages); in hugetlb_vmemmap_optimize_folios()