/linux-6.12.1/arch/sh/mm/ |
D | ioremap_fixed.c | 52 unsigned int nrpages; in ioremap_fixed() local 79 nrpages = size >> PAGE_SHIFT; in ioremap_fixed() 80 if (nrpages > FIX_N_IOREMAPS) in ioremap_fixed() 88 while (nrpages > 0) { in ioremap_fixed() 93 --nrpages; in ioremap_fixed() 104 unsigned int nrpages; in iounmap_fixed() local 122 nrpages = map->size >> PAGE_SHIFT; in iounmap_fixed() 124 idx = FIX_IOREMAP_BEGIN + slot + nrpages - 1; in iounmap_fixed() 125 while (nrpages > 0) { in iounmap_fixed() 128 --nrpages; in iounmap_fixed()
|
/linux-6.12.1/fs/erofs/ |
D | zutil.c | 12 unsigned int nrpages; member 45 if (requiredpages > gbuf->nrpages) { in z_erofs_get_gbuf() 65 int z_erofs_gbuf_growsize(unsigned int nrpages) in z_erofs_gbuf_growsize() argument 75 if (nrpages <= z_erofs_gbuf_nrpages) { in z_erofs_gbuf_growsize() 82 tmp_pages = kcalloc(nrpages, sizeof(*tmp_pages), GFP_KERNEL); in z_erofs_gbuf_growsize() 86 for (j = 0; j < gbuf->nrpages; ++j) in z_erofs_gbuf_growsize() 90 j = alloc_pages_bulk_array(GFP_KERNEL, nrpages, in z_erofs_gbuf_growsize() 94 } while (j != nrpages); in z_erofs_gbuf_growsize() 96 ptr = vmap(tmp_pages, nrpages, VM_MAP, PAGE_KERNEL); in z_erofs_gbuf_growsize() 105 gbuf->nrpages = nrpages; in z_erofs_gbuf_growsize() [all …]
|
D | internal.h | 477 int z_erofs_gbuf_growsize(unsigned int nrpages);
|
/linux-6.12.1/mm/ |
D | early_ioremap.c | 104 unsigned int nrpages; in __early_ioremap() local 138 nrpages = size >> PAGE_SHIFT; in __early_ioremap() 139 if (WARN_ON(nrpages > NR_FIX_BTMAPS)) in __early_ioremap() 146 while (nrpages > 0) { in __early_ioremap() 153 --nrpages; in __early_ioremap() 166 unsigned int nrpages; in early_iounmap() local 195 nrpages = PAGE_ALIGN(offset + size) >> PAGE_SHIFT; in early_iounmap() 198 while (nrpages > 0) { in early_iounmap() 204 --nrpages; in early_iounmap()
|
D | fadvise.c | 39 unsigned long nrpages; in generic_fadvise() local 102 nrpages = end_index - start_index + 1; in generic_fadvise() 103 if (!nrpages) in generic_fadvise() 104 nrpages = ~0UL; in generic_fadvise() 106 force_page_cache_readahead(mapping, file, start_index, nrpages); in generic_fadvise()
|
D | mprotect.c | 609 long nrpages = (end - start) >> PAGE_SHIFT; in mprotect_fixup() local 649 if (!may_expand_vm(mm, newflags, nrpages) && in mprotect_fixup() 650 may_expand_vm(mm, oldflags, nrpages)) in mprotect_fixup() 654 charged = nrpages; in mprotect_fixup() 685 vm_unacct_memory(nrpages); in mprotect_fixup() 696 vm_stat_account(mm, oldflags, -nrpages); in mprotect_fixup() 697 vm_stat_account(mm, newflags, nrpages); in mprotect_fixup()
|
D | vma.c | 1222 long nrpages; in vms_gather_munmap_vmas() local 1241 nrpages = vma_pages(next); in vms_gather_munmap_vmas() 1243 vms->nr_pages += nrpages; in vms_gather_munmap_vmas() 1245 vms->locked_vm += nrpages; in vms_gather_munmap_vmas() 1248 vms->nr_accounted += nrpages; in vms_gather_munmap_vmas() 1251 vms->exec_vm += nrpages; in vms_gather_munmap_vmas() 1253 vms->stack_vm += nrpages; in vms_gather_munmap_vmas() 1255 vms->data_vm += nrpages; in vms_gather_munmap_vmas()
|
D | swap_state.c | 124 address_space->nrpages += nr; in add_to_swap_cache() 165 address_space->nrpages -= nr; in __delete_from_swap_cache()
|
D | filemap.c | 155 mapping->nrpages -= nr; in page_cache_delete() 325 mapping->nrpages -= total_pages; in page_cache_delete_batch() 635 return mapping->nrpages; in mapping_needs_writeback() 920 mapping->nrpages += nr; in __filemap_add_folio() 3952 if (mapping->nrpages && in kiocb_invalidate_post_direct_write() 4244 if (!mapping || !mapping->nrpages || end < start) in filemap_invalidate_inode() 4250 if (!mapping->nrpages) in filemap_invalidate_inode()
|
/linux-6.12.1/arch/x86/mm/ |
D | ioremap.c | 48 unsigned long nrpages = size >> PAGE_SHIFT; in ioremap_change_attr() local 54 err = _set_memory_uc(vaddr, nrpages); in ioremap_change_attr() 57 err = _set_memory_wc(vaddr, nrpages); in ioremap_change_attr() 60 err = _set_memory_wt(vaddr, nrpages); in ioremap_change_attr() 63 err = _set_memory_wb(vaddr, nrpages); in ioremap_change_attr()
|
/linux-6.12.1/lib/ |
D | test_xarray.c | 738 unsigned int nrpages = 1UL << order; in check_xa_multi_store_adv_add() local 741 XA_BUG_ON(xa, index & (nrpages - 1)); in check_xa_multi_store_adv_add() 821 unsigned int nrpages = 1UL << order; in check_xa_multi_store_adv() local 826 base = round_down(index, nrpages); in check_xa_multi_store_adv() 827 next_index = round_down(base + nrpages, nrpages); in check_xa_multi_store_adv() 828 next_next_index = round_down(next_index + nrpages, nrpages); in check_xa_multi_store_adv() 832 for (i = 0; i < nrpages; i++) in check_xa_multi_store_adv() 847 for (i = 0; i < nrpages; i++) in check_xa_multi_store_adv() 861 for (i = 0; i < nrpages; i++) in check_xa_multi_store_adv() 864 for (i = 0; i < nrpages; i++) in check_xa_multi_store_adv() [all …]
|
/linux-6.12.1/fs/f2fs/ |
D | debug.c | 161 si->node_pages = NODE_MAPPING(sbi)->nrpages; in update_general_status() 163 si->meta_pages = META_MAPPING(sbi)->nrpages; in update_general_status() 166 si->compress_pages = COMPRESS_MAPPING(sbi)->nrpages; in update_general_status() 323 unsigned long npages = NODE_MAPPING(sbi)->nrpages; in update_mem_info() 328 unsigned long npages = META_MAPPING(sbi)->nrpages; in update_mem_info() 334 unsigned long npages = COMPRESS_MAPPING(sbi)->nrpages; in update_mem_info()
|
D | node.c | 111 (COMPRESS_MAPPING(sbi)->nrpages < in f2fs_available_free_memory() 2861 int i, idx, last_offset, nrpages; in f2fs_restore_node_summary() local 2868 for (i = 0; i < last_offset; i += nrpages, addr += nrpages) { in f2fs_restore_node_summary() 2869 nrpages = bio_max_segs(last_offset - i); in f2fs_restore_node_summary() 2872 f2fs_ra_meta_pages(sbi, addr, nrpages, META_POR, true); in f2fs_restore_node_summary() 2874 for (idx = addr; idx < addr + nrpages; idx++) { in f2fs_restore_node_summary() 2889 addr + nrpages); in f2fs_restore_node_summary()
|
/linux-6.12.1/fs/nilfs2/ |
D | page.c | 332 smap->nrpages--; in nilfs_copy_back_pages() 343 dmap->nrpages++; in nilfs_copy_back_pages() 493 if (inode->i_mapping->nrpages == 0) in nilfs_find_uncommitted_extent()
|
/linux-6.12.1/fs/netfs/ |
D | locking.c | 139 if (inode->i_mapping->nrpages != 0) { in netfs_block_buffered()
|
/linux-6.12.1/block/ |
D | bdev.c | 99 if (mapping->nrpages) { in invalidate_bdev() 482 ret += inode->i_mapping->nrpages; in nr_blockdev_pages() 1228 mapping->nrpages == 0) { in sync_bdevs()
|
/linux-6.12.1/fs/nfs/ |
D | inode.c | 148 if (mapping->nrpages != 0) { in nfs_sync_mapping() 209 if (inode->i_mapping->nrpages == 0) in nfs_set_cache_invalid() 215 if (inode->i_mapping->nrpages == 0 || in nfs_set_cache_invalid() 259 if (mapping->nrpages != 0) { in nfs_zap_mapping() 1102 if (inode->i_mapping->nrpages == 0) in nfs_close_context() 1378 if (mapping->nrpages != 0) { in nfs_invalidate_mapping()
|
/linux-6.12.1/fs/bcachefs/ |
D | fs-io-direct.c | 190 if (unlikely(mapping->nrpages)) { in bch2_read_iter() 647 if (unlikely(mapping->nrpages)) { in bch2_direct_write()
|
D | fs-io-pagecache.c | 58 if (!mapping->nrpages) in bch2_write_invalidate_inode_pages_range() 65 if (!mapping->nrpages) in bch2_write_invalidate_inode_pages_range()
|
/linux-6.12.1/fs/jffs2/ |
D | dir.c | 191 inode->i_mapping->nrpages = 0; in jffs2_create() 214 f->inocache->pino_nlink, inode->i_mapping->nrpages); in jffs2_create()
|
/linux-6.12.1/fs/orangefs/ |
D | file.c | 420 #define mapping_nrpages(idata) ((idata)->nrpages)
|
/linux-6.12.1/ipc/ |
D | shm.c | 939 *rss_add += pages_per_huge_page(h) * mapping->nrpages; in shm_add_rss_swap() 945 *rss_add += inode->i_mapping->nrpages; in shm_add_rss_swap() 949 *rss_add += inode->i_mapping->nrpages; in shm_add_rss_swap()
|
/linux-6.12.1/fs/gfs2/ |
D | glock.c | 2406 unsigned long nrpages = 0; in gfs2_dump_glock() local 2411 nrpages = mapping->nrpages; in gfs2_dump_glock() 2430 (int)gl->gl_lockref.count, gl->gl_hold_time, nrpages); in gfs2_dump_glock()
|
D | log.c | 722 if (ip->i_inode.i_mapping->nrpages == 0) { in gfs2_ordered_write() 743 if (ip->i_inode.i_mapping->nrpages == 0) in gfs2_ordered_wait()
|
/linux-6.12.1/fs/ |
D | dax.c | 626 mapping->nrpages -= PG_PMD_NR; in grab_mapping_entry() 642 mapping->nrpages += 1UL << order; in grab_mapping_entry() 762 mapping->nrpages -= 1UL << dax_entry_order(entry); in __dax_invalidate_entry()
|