/linux-6.12.1/mm/ |
D | swap_state.c | 438 struct folio *new_folio = NULL; in __read_swap_cache_async() local 476 if (!new_folio) { in __read_swap_cache_async() 477 new_folio = folio_alloc_mpol(gfp_mask, 0, mpol, ilx, numa_node_id()); in __read_swap_cache_async() 478 if (!new_folio) in __read_swap_cache_async() 515 __folio_set_locked(new_folio); in __read_swap_cache_async() 516 __folio_set_swapbacked(new_folio); in __read_swap_cache_async() 518 if (mem_cgroup_swapin_charge_folio(new_folio, NULL, gfp_mask, entry)) in __read_swap_cache_async() 522 if (add_to_swap_cache(new_folio, entry, gfp_mask & GFP_RECLAIM_MASK, &shadow)) in __read_swap_cache_async() 528 workingset_refault(new_folio, shadow); in __read_swap_cache_async() 531 folio_add_lru(new_folio); in __read_swap_cache_async() [all …]
|
D | khugepaged.c | 1811 struct folio *folio, *tmp, *new_folio; in collapse_file() local 1821 result = alloc_charge_folio(&new_folio, mm, cc); in collapse_file() 1825 __folio_set_locked(new_folio); in collapse_file() 1827 __folio_set_swapbacked(new_folio); in collapse_file() 1828 new_folio->index = start; in collapse_file() 1829 new_folio->mapping = mapping; in collapse_file() 2058 dst = folio_page(new_folio, 0); in collapse_file() 2144 __lruvec_stat_mod_folio(new_folio, NR_SHMEM_THPS, HPAGE_PMD_NR); in collapse_file() 2146 __lruvec_stat_mod_folio(new_folio, NR_FILE_THPS, HPAGE_PMD_NR); in collapse_file() 2149 __lruvec_stat_mod_folio(new_folio, NR_FILE_PAGES, nr_none); in collapse_file() [all …]
|
D | hugetlb.c | 2852 struct folio *new_folio = NULL; in alloc_and_dissolve_hugetlb_folio() local 2884 if (!new_folio) { in alloc_and_dissolve_hugetlb_folio() 2886 new_folio = alloc_buddy_hugetlb_folio(h, gfp_mask, nid, in alloc_and_dissolve_hugetlb_folio() 2888 if (!new_folio) in alloc_and_dissolve_hugetlb_folio() 2890 __prep_new_hugetlb_folio(h, new_folio); in alloc_and_dissolve_hugetlb_folio() 2908 enqueue_hugetlb_folio(h, new_folio); in alloc_and_dissolve_hugetlb_folio() 2921 if (new_folio) in alloc_and_dissolve_hugetlb_folio() 2922 update_and_free_hugetlb_folio(h, new_folio, false); in alloc_and_dissolve_hugetlb_folio() 5193 struct folio *new_folio, pte_t old, unsigned long sz) in hugetlb_install_folio() argument 5195 pte_t newpte = make_huge_pte(vma, &new_folio->page, 1); in hugetlb_install_folio() [all …]
|
D | memory.c | 918 struct folio *new_folio; in copy_present_page() local 921 new_folio = *prealloc; in copy_present_page() 922 if (!new_folio) in copy_present_page() 930 if (copy_mc_user_highpage(&new_folio->page, page, addr, src_vma)) in copy_present_page() 934 __folio_mark_uptodate(new_folio); in copy_present_page() 935 folio_add_new_anon_rmap(new_folio, dst_vma, addr, RMAP_EXCLUSIVE); in copy_present_page() 936 folio_add_lru_vma(new_folio, dst_vma); in copy_present_page() 940 pte = mk_pte(&new_folio->page, dst_vma->vm_page_prot); in copy_present_page() 1059 struct folio *new_folio; in folio_prealloc() local 1062 new_folio = vma_alloc_zeroed_movable_folio(vma, addr); in folio_prealloc() [all …]
|
D | ksm.c | 2953 struct folio *new_folio; in ksm_might_need_to_copy() local 2973 new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr, false); in ksm_might_need_to_copy() 2974 if (new_folio && in ksm_might_need_to_copy() 2975 mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL)) { in ksm_might_need_to_copy() 2976 folio_put(new_folio); in ksm_might_need_to_copy() 2977 new_folio = NULL; in ksm_might_need_to_copy() 2979 if (new_folio) { in ksm_might_need_to_copy() 2980 if (copy_mc_user_highpage(folio_page(new_folio, 0), page, in ksm_might_need_to_copy() 2982 folio_put(new_folio); in ksm_might_need_to_copy() 2985 folio_set_dirty(new_folio); in ksm_might_need_to_copy() [all …]
|
D | hugetlb_cgroup.c | 900 void hugetlb_cgroup_migrate(struct folio *old_folio, struct folio *new_folio) in hugetlb_cgroup_migrate() argument 916 set_hugetlb_cgroup(new_folio, h_cg); in hugetlb_cgroup_migrate() 917 set_hugetlb_cgroup_rsvd(new_folio, h_cg_rsvd); in hugetlb_cgroup_migrate() 918 list_move(&new_folio->lru, &h->hugepage_activelist); in hugetlb_cgroup_migrate()
|
D | huge_memory.c | 3087 struct folio *new_folio = (struct folio *)page_tail; in __split_huge_page_tail() local 3139 new_folio->swap.val = folio->swap.val + tail; in __split_huge_page_tail() 3153 folio_set_large_rmappable(new_folio); in __split_huge_page_tail() 3159 folio_nr_pages(new_folio) : 0)); in __split_huge_page_tail() 3162 folio_set_young(new_folio); in __split_huge_page_tail() 3164 folio_set_idle(new_folio); in __split_huge_page_tail() 3166 folio_xchg_last_cpupid(new_folio, folio_last_cpupid(folio)); in __split_huge_page_tail() 3228 struct folio *new_folio = (struct folio *)head; in __split_huge_page() local 3230 folio_set_order(new_folio, new_order); in __split_huge_page() 3268 struct folio *new_folio = page_folio(subpage); in __split_huge_page() local [all …]
|
/linux-6.12.1/fs/sysv/ |
D | namei.c | 218 struct folio *new_folio; in sysv_rename() local 226 new_de = sysv_find_entry(new_dentry, &new_folio); in sysv_rename() 229 err = sysv_set_link(new_de, new_folio, old_inode); in sysv_rename() 230 folio_release_kmap(new_folio, new_de); in sysv_rename()
|
/linux-6.12.1/fs/minix/ |
D | namei.c | 204 struct folio *new_folio; in minix_rename() local 212 new_de = minix_find_entry(new_dentry, &new_folio); in minix_rename() 215 err = minix_set_link(new_de, new_folio, old_inode); in minix_rename() 216 folio_release_kmap(new_folio, new_de); in minix_rename()
|
/linux-6.12.1/include/trace/events/ |
D | huge_memory.h | 210 TP_PROTO(struct mm_struct *mm, struct folio *new_folio, pgoff_t index, 213 TP_ARGS(mm, new_folio, index, addr, is_shmem, file, nr, result), 227 __entry->hpfn = new_folio ? folio_pfn(new_folio) : -1;
|
/linux-6.12.1/fs/ext2/ |
D | namei.c | 354 struct folio *new_folio; in ext2_rename() local 362 &new_folio); in ext2_rename() 367 err = ext2_set_link(new_dir, new_de, new_folio, old_inode, true); in ext2_rename() 368 folio_release_kmap(new_folio, new_de); in ext2_rename()
|
/linux-6.12.1/fs/ufs/ |
D | namei.c | 273 struct folio *new_folio; in ufs_rename() local 281 new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_folio); in ufs_rename() 284 ufs_set_link(new_dir, new_de, new_folio, old_inode, 1); in ufs_rename()
|
/linux-6.12.1/include/linux/ |
D | hugetlb_cgroup.h | 157 struct folio *new_folio); 266 struct folio *new_folio) in hugetlb_cgroup_migrate() argument
|
D | hugetlb.h | 161 void move_hugetlb_state(struct folio *old_folio, struct folio *new_folio, int reason); 438 struct folio *new_folio, int reason) in move_hugetlb_state() argument
|
/linux-6.12.1/fs/nilfs2/ |
D | namei.c | 391 struct folio *new_folio; in nilfs_rename() local 399 &new_folio); in nilfs_rename() 404 nilfs_set_link(new_dir, new_de, new_folio, old_inode); in nilfs_rename() 405 folio_release_kmap(new_folio, new_de); in nilfs_rename()
|
/linux-6.12.1/fs/jbd2/ |
D | journal.c | 333 struct folio *new_folio; in jbd2_journal_write_metadata_buffer() local 360 new_folio = virt_to_folio(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 361 new_offset = offset_in_folio(new_folio, jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 369 new_folio = bh_in->b_folio; in jbd2_journal_write_metadata_buffer() 370 new_offset = offset_in_folio(new_folio, bh_in->b_data); in jbd2_journal_write_metadata_buffer() 371 mapped_data = kmap_local_folio(new_folio, new_offset); in jbd2_journal_write_metadata_buffer() 402 memcpy_from_folio(tmp, new_folio, new_offset, bh_in->b_size); in jbd2_journal_write_metadata_buffer() 411 new_folio = virt_to_folio(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 412 new_offset = offset_in_folio(new_folio, jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 417 folio_set_bh(new_bh, new_folio, new_offset); in jbd2_journal_write_metadata_buffer()
|
/linux-6.12.1/kernel/events/ |
D | uprobes.c | 165 struct folio *new_folio; in __replace_page() local 175 new_folio = page_folio(new_page); in __replace_page() 176 err = mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL); in __replace_page() 191 folio_get(new_folio); in __replace_page() 192 folio_add_new_anon_rmap(new_folio, vma, addr, RMAP_EXCLUSIVE); in __replace_page() 193 folio_add_lru_vma(new_folio, vma); in __replace_page()
|