/linux-6.12.1/mm/ |
D | mempolicy.c | 480 nodemask_t *nmask; member 500 return node_isset(nid, *qp->nmask) == !(flags & MPOL_MF_INVERT); in queue_folio_required() 768 .nmask = nodes, in queue_pages_range() 927 static long do_get_mempolicy(int *policy, nodemask_t *nmask, in do_get_mempolicy() argument 944 *nmask = cpuset_current_mems_allowed; in do_get_mempolicy() 1009 if (nmask) { in do_get_mempolicy() 1011 *nmask = pol->w.user_nodemask; in do_get_mempolicy() 1014 get_policy_nodemask(pol, nmask); in do_get_mempolicy() 1065 nodemask_t nmask; in migrate_to_node() local 1076 nodes_clear(nmask); in migrate_to_node() [all …]
|
D | memory-tiers.c | 173 nodemask_t nmask; in nodelist_show() local 176 nmask = get_memtier_nodemask(to_memory_tier(dev)); in nodelist_show() 177 ret = sysfs_emit(buf, "%*pbl\n", nodemask_pr_args(&nmask)); in nodelist_show()
|
D | memory_hotplug.c | 1830 nodemask_t nmask = node_states[N_MEMORY]; in do_migrate_range() local 1832 .nmask = &nmask, in do_migrate_range() 1849 node_clear(mtc.nid, nmask); in do_migrate_range() 1850 if (nodes_empty(nmask)) in do_migrate_range() 1851 node_set(mtc.nid, nmask); in do_migrate_range()
|
D | hugetlb.c | 1351 int nid, nodemask_t *nmask) in dequeue_hugetlb_folio_nodemask() argument 1367 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_hugetlb_folio_nodemask() 2002 gfp_t gfp_mask, int nid, nodemask_t *nmask, in alloc_buddy_hugetlb_folio() argument 2024 folio = __folio_alloc(gfp_mask, order, nid, nmask); in alloc_buddy_hugetlb_folio() 2066 gfp_t gfp_mask, int nid, nodemask_t *nmask, in only_alloc_fresh_hugetlb_folio() argument 2072 folio = alloc_gigantic_folio(h, gfp_mask, nid, nmask); in only_alloc_fresh_hugetlb_folio() 2074 folio = alloc_buddy_hugetlb_folio(h, gfp_mask, nid, nmask, node_alloc_noretry); in only_alloc_fresh_hugetlb_folio() 2088 gfp_t gfp_mask, int nid, nodemask_t *nmask) in alloc_fresh_hugetlb_folio() argument 2093 folio = alloc_gigantic_folio(h, gfp_mask, nid, nmask); in alloc_fresh_hugetlb_folio() 2095 folio = alloc_buddy_hugetlb_folio(h, gfp_mask, nid, nmask, NULL); in alloc_fresh_hugetlb_folio() [all …]
|
D | migrate.c | 2145 mtc->nmask, gfp_mask, in alloc_migration_target() 2162 return __folio_alloc(gfp_mask, order, nid, mtc->nmask); in alloc_migration_target()
|
D | internal.h | 1219 nodemask_t *nmask; member
|
D | vmscan.c | 961 allowed_mask = mtc->nmask; in alloc_migrate_folio() 971 mtc->nmask = NULL; in alloc_migrate_folio() 978 mtc->nmask = allowed_mask; in alloc_migrate_folio() 1003 .nmask = &allowed_mask, in demote_folio_list()
|
/linux-6.12.1/drivers/clk/ |
D | clk-fractional-divider.c | 74 u32 mmask, nmask; in clk_fd_get_div() local 90 nmask = GENMASK(fd->nwidth - 1, 0) << fd->nshift; in clk_fd_get_div() 93 n = (val & nmask) >> fd->nshift; in clk_fd_get_div() 181 u32 mmask, nmask; in clk_fd_set_rate() local 199 nmask = GENMASK(fd->nwidth - 1, 0) << fd->nshift; in clk_fd_set_rate() 207 val &= ~(mmask | nmask); in clk_fd_set_rate()
|
/linux-6.12.1/tools/testing/selftests/kvm/include/ |
D | numaif.h | 20 long get_mempolicy(int *policy, const unsigned long *nmask, in get_mempolicy() argument 23 return syscall(__NR_get_mempolicy, policy, nmask, in get_mempolicy()
|
/linux-6.12.1/drivers/ntb/test/ |
D | ntb_pingpong.c | 105 u64 nmask; member 124 if (link & pp->nmask) in pp_find_next_peer() 125 pidx = __ffs64(link & pp->nmask); in pp_find_next_peer() 307 pp->nmask = GENMASK_ULL(pcnt - 1, pidx); in pp_init_flds() 310 pp->in_db, pp->pmask, pp->nmask); in pp_init_flds()
|
/linux-6.12.1/drivers/net/ethernet/microchip/vcap/ |
D | vcap_api_debugfs.c | 92 u8 nvalue[16], nmask[16]; in vcap_debugfs_show_rule_keyfield() local 96 vcap_netbytes_copy(nmask, data->u128.mask, in vcap_debugfs_show_rule_keyfield() 97 sizeof(nmask)); in vcap_debugfs_show_rule_keyfield() 98 out->prf(out->dst, "%pI6/%pI6", nvalue, nmask); in vcap_debugfs_show_rule_keyfield()
|
/linux-6.12.1/include/linux/ |
D | hugetlb.h | 693 nodemask_t *nmask, gfp_t gfp_mask, 696 nodemask_t *nmask, gfp_t gfp_mask); 1067 nodemask_t *nmask, gfp_t gfp_mask) 1074 nodemask_t *nmask, gfp_t gfp_mask,
|
D | syscalls.h | 827 const unsigned long __user *nmask, 831 unsigned long __user *nmask, 834 asmlinkage long sys_set_mempolicy(int mode, const unsigned long __user *nmask,
|
/linux-6.12.1/drivers/scsi/cxlflash/ |
D | vlun.c | 1038 rhte->nmask = MC_RHT_NMASK; in cxlflash_disk_virtual_open() 1240 if (ctxi_dst->rht_start[i].nmask != 0) { in cxlflash_disk_clone() 1291 if (ctxi_src->rht_start[i].nmask == 0) in cxlflash_disk_clone() 1296 ctxi_dst->rht_start[i].nmask = ctxi_src->rht_start[i].nmask; in cxlflash_disk_clone()
|
D | sislite.h | 513 u8 nmask; member
|
D | superpipe.c | 462 if (unlikely(rhte->nmask == 0)) { in get_rhte() 490 if (ctxi->rht_start[i].nmask == 0) { in rhte_checkout() 513 rhte->nmask = 0; in rhte_checkin()
|
/linux-6.12.1/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 431 long set_mempolicy(int mode, const unsigned long *nmask, 436 'nmask'. 'nmask' points to a bit mask of node ids containing at least 447 const unsigned long *nmask, unsigned long maxnode, 460 const unsigned long *nmask, unsigned long maxnode, 463 mbind() installs the policy specified by (mode, nmask, maxnodes) as a
|
/linux-6.12.1/mm/damon/ |
D | paddr.c | 346 .nmask = &allowed_mask in __damon_pa_migrate_folio_list()
|
/linux-6.12.1/net/sched/ |
D | sch_api.c | 689 unsigned int nsize, nmask, osize; in qdisc_class_hash_grow() local 696 nmask = nsize - 1; in qdisc_class_hash_grow() 707 h = qdisc_class_hash(cl->classid, nmask); in qdisc_class_hash_grow() 713 clhash->hashmask = nmask; in qdisc_class_hash_grow()
|
/linux-6.12.1/drivers/scsi/ |
D | pmcraid.c | 432 u32 nmask = gmask | GLOBAL_INTERRUPT_MASK; in pmcraid_disable_interrupts() local 435 iowrite32(nmask, pinstance->int_regs.global_interrupt_mask_reg); in pmcraid_disable_interrupts() 459 u32 nmask = gmask & (~GLOBAL_INTERRUPT_MASK); in pmcraid_enable_interrupts() local 461 iowrite32(nmask, pinstance->int_regs.global_interrupt_mask_reg); in pmcraid_enable_interrupts()
|