/linux-6.12.1/arch/x86/kernel/ |
D | smpboot.c | 1048 zalloc_cpumask_var_node(&per_cpu(cpu_sibling_map, cpu), GFP_KERNEL, node); in smp_prepare_cpus_common() 1049 zalloc_cpumask_var_node(&per_cpu(cpu_core_map, cpu), GFP_KERNEL, node); in smp_prepare_cpus_common() 1050 zalloc_cpumask_var_node(&per_cpu(cpu_die_map, cpu), GFP_KERNEL, node); in smp_prepare_cpus_common() 1051 zalloc_cpumask_var_node(&per_cpu(cpu_llc_shared_map, cpu), GFP_KERNEL, node); in smp_prepare_cpus_common() 1052 zalloc_cpumask_var_node(&per_cpu(cpu_l2c_shared_map, cpu), GFP_KERNEL, node); in smp_prepare_cpus_common()
|
D | kvm.c | 690 zalloc_cpumask_var_node(per_cpu_ptr(&__pv_cpu_mask, cpu), in kvm_alloc_cpumask()
|
/linux-6.12.1/arch/powerpc/kernel/ |
D | smp.c | 926 zalloc_cpumask_var_node(mask, GFP_KERNEL, cpu_to_node(cpu)); in update_mask_from_threadgroup() 1071 zalloc_cpumask_var_node(&per_cpu(cpu_smallcore_map, cpu), in init_big_cores() 1109 zalloc_cpumask_var_node(&per_cpu(cpu_sibling_map, cpu), in smp_prepare_cpus() 1111 zalloc_cpumask_var_node(&per_cpu(cpu_l2_cache_map, cpu), in smp_prepare_cpus() 1113 zalloc_cpumask_var_node(&per_cpu(cpu_core_map, cpu), in smp_prepare_cpus() 1116 zalloc_cpumask_var_node(&per_cpu(cpu_coregroup_map, cpu), in smp_prepare_cpus()
|
/linux-6.12.1/kernel/irq/ |
D | irqdesc.c | 56 if (!zalloc_cpumask_var_node(&desc->irq_common_data.affinity, in alloc_masks() 61 if (!zalloc_cpumask_var_node(&desc->irq_common_data.effective_affinity, in alloc_masks() 69 if (!zalloc_cpumask_var_node(&desc->pending_mask, GFP_KERNEL, node)) { in alloc_masks()
|
/linux-6.12.1/arch/x86/kernel/apic/ |
D | x2apic_cluster.c | 188 if (!zalloc_cpumask_var_node(&per_cpu(ipi_mask, cpu), GFP_KERNEL, node)) in x2apic_prepare_cpu()
|
/linux-6.12.1/kernel/ |
D | smp.c | 58 if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL, in smpcfd_prepare_cpu() 61 if (!zalloc_cpumask_var_node(&cfd->cpumask_ipi, GFP_KERNEL, in smpcfd_prepare_cpu()
|
D | workqueue.c | 7748 BUG_ON(!zalloc_cpumask_var_node(&pt->pod_cpus[0], GFP_KERNEL, NUMA_NO_NODE)); in workqueue_init_early()
|
/linux-6.12.1/include/linux/ |
D | cpumask.h | 936 bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node) in zalloc_cpumask_var_node() function 996 static __always_inline bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, in zalloc_cpumask_var_node() function
|
/linux-6.12.1/drivers/xen/ |
D | xen-acpi-processor.c | 536 if (!zalloc_cpumask_var_node( in xen_acpi_processor_init()
|
/linux-6.12.1/drivers/cpufreq/ |
D | acpi-cpufreq.c | 566 if (!zalloc_cpumask_var_node( in acpi_cpufreq_early_init()
|
/linux-6.12.1/kernel/sched/ |
D | rt.c | 2456 zalloc_cpumask_var_node(&per_cpu(local_cpu_mask, i), in init_sched_rt_class()
|
D | deadline.c | 2943 zalloc_cpumask_var_node(&per_cpu(local_cpu_mask_dl, i), in init_sched_dl_class()
|
D | fair.c | 13663 zalloc_cpumask_var_node(&per_cpu(load_balance_mask, i), GFP_KERNEL, cpu_to_node(i)); in init_sched_fair_class() 13664 zalloc_cpumask_var_node(&per_cpu(select_rq_mask, i), GFP_KERNEL, cpu_to_node(i)); in init_sched_fair_class() 13665 zalloc_cpumask_var_node(&per_cpu(should_we_balance_tmpmask, i), in init_sched_fair_class()
|
D | core.c | 8523 zalloc_cpumask_var_node(&rq->scratch_mask, GFP_KERNEL, cpu_to_node(i)); in sched_init()
|
/linux-6.12.1/block/ |
D | blk-mq.c | 3756 if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node)) in blk_mq_alloc_hctx()
|
/linux-6.12.1/drivers/net/ethernet/pensando/ionic/ |
D | ionic_lif.c | 3197 if (!zalloc_cpumask_var_node(&affinity_masks[i], GFP_KERNEL, in ionic_affinity_masks_alloc()
|