/linux-6.12.1/fs/hfsplus/ |
D | bnode.c | 124 struct page **src_page, **dst_page; in hfs_bnode_copy() local 132 src_page = src_node->page + (src >> PAGE_SHIFT); in hfs_bnode_copy() 139 memcpy_page(*dst_page, src, *src_page, src, l); in hfs_bnode_copy() 144 memcpy_page(*++dst_page, 0, *++src_page, 0, l); in hfs_bnode_copy() 152 src_ptr = kmap_local_page(*src_page) + src; in hfs_bnode_copy() 170 src_page++; in hfs_bnode_copy() 177 struct page **src_page, **dst_page; in hfs_bnode_move() local 188 src_page = node->page + (src >> PAGE_SHIFT); in hfs_bnode_move() 197 src_ptr = kmap_local_page(*src_page); in hfs_bnode_move() 204 src_page--; in hfs_bnode_move() [all …]
|
/linux-6.12.1/arch/arm/lib/ |
D | uaccess_with_memcpy.c | 230 struct page *src_page, *dst_page; 236 src_page = alloc_page(GFP_KERNEL); 237 if (!src_page) 242 kernel_ptr = page_address(src_page); 275 put_page(src_page);
|
/linux-6.12.1/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_blit.c | 369 u32 src_page = src_offset >> PAGE_SHIFT; in vmw_bo_cpu_blit_line() local 373 bool unmap_src = d->src_addr && (src_page != d->mapped_src || in vmw_bo_cpu_blit_line() 403 if (WARN_ON_ONCE(src_page >= d->src_num_pages)) in vmw_bo_cpu_blit_line() 407 kmap_atomic_prot(d->src_pages[src_page], in vmw_bo_cpu_blit_line() 412 d->mapped_src = src_page; in vmw_bo_cpu_blit_line()
|
/linux-6.12.1/fs/crypto/ |
D | crypto.c | 110 struct page *src_page, struct page *dest_page, in fscrypt_crypt_data_unit() argument 139 sg_set_page(&src, src_page, len, offs); in fscrypt_crypt_data_unit()
|
D | fscrypt_private.h | 303 struct page *src_page, struct page *dest_page,
|
/linux-6.12.1/fs/hfs/ |
D | bnode.c | 111 struct page *src_page, *dst_page; in hfs_bnode_copy() local 118 src_page = src_node->page[0]; in hfs_bnode_copy() 121 memcpy_page(dst_page, dst, src_page, src, len); in hfs_bnode_copy()
|
/linux-6.12.1/arch/x86/kernel/cpu/sgx/ |
D | ioctl.c | 206 struct page *src_page; in __sgx_encl_add_page() local 217 ret = get_user_pages(src, 1, 0, &src_page); in __sgx_encl_add_page() 224 pginfo.contents = (unsigned long)kmap_local_page(src_page); in __sgx_encl_add_page() 229 put_page(src_page); in __sgx_encl_add_page()
|
/linux-6.12.1/include/linux/ |
D | highmem.h | 395 struct page *src_page, size_t src_off, in memcpy_page() argument 399 char *src = kmap_local_page(src_page); in memcpy_page()
|
/linux-6.12.1/mm/ |
D | khugepaged.c | 731 struct page *src_page = pte_page(pteval); in __collapse_huge_page_copy_succeeded() local 733 src = page_folio(src_page); in __collapse_huge_page_copy_succeeded() 743 folio_remove_rmap_pte(src, src_page, vma); in __collapse_huge_page_copy_succeeded() 745 free_page_and_swap_cache(src_page); in __collapse_huge_page_copy_succeeded() 813 struct page *src_page; in __collapse_huge_page_copy() local 819 src_page = pte_page(pteval); in __collapse_huge_page_copy() 820 if (copy_mc_user_highpage(page, src_page, src_addr, vma) > 0) { in __collapse_huge_page_copy()
|
D | huge_memory.c | 1571 struct page *src_page; in copy_huge_pmd() local 1655 src_page = pmd_page(pmd); in copy_huge_pmd() 1656 VM_BUG_ON_PAGE(!PageHead(src_page), src_page); in copy_huge_pmd() 1657 src_folio = page_folio(src_page); in copy_huge_pmd() 1660 if (unlikely(folio_try_dup_anon_rmap_pmd(src_folio, src_page, src_vma))) { in copy_huge_pmd() 2382 struct page *src_page; in move_pages_huge_pmd() local 2413 src_page = pmd_page(src_pmdval); in move_pages_huge_pmd() 2416 if (unlikely(!PageAnonExclusive(src_page))) { in move_pages_huge_pmd() 2421 src_folio = page_folio(src_page); in move_pages_huge_pmd() 2486 _dst_pmd = mk_huge_pmd(src_page, dst_vma->vm_page_prot); in move_pages_huge_pmd()
|
D | memory.c | 6825 struct page *src_page; in copy_user_gigantic_page() local 6829 src_page = folio_page(src, i); in copy_user_gigantic_page() 6832 if (copy_mc_user_highpage(dst_page, src_page, in copy_user_gigantic_page()
|
/linux-6.12.1/fs/f2fs/ |
D | node.c | 140 struct page *src_page; in get_next_nat_page() local 150 src_page = get_current_nat_page(sbi, nid); in get_next_nat_page() 151 if (IS_ERR(src_page)) in get_next_nat_page() 152 return src_page; in get_next_nat_page() 154 f2fs_bug_on(sbi, PageDirty(src_page)); in get_next_nat_page() 156 src_addr = page_address(src_page); in get_next_nat_page() 160 f2fs_put_page(src_page, 1); in get_next_nat_page()
|
/linux-6.12.1/fs/ecryptfs/ |
D | crypto.c | 352 struct page *src_page, in crypt_extent() argument 355 pgoff_t page_index = op == ENCRYPT ? src_page->index : dst_page->index; in crypt_extent() 375 sg_set_page(&src_sg, src_page, extent_size, in crypt_extent()
|
/linux-6.12.1/arch/x86/crypto/ |
D | aesni-intel_glue.c | 604 struct page *src_page = sg_page(src); in xts_crypt() local 606 void *src_virt = kmap_local_page(src_page) + src->offset; in xts_crypt()
|