/linux-6.12.1/arch/sparc/include/asm/ |
D | smp_32.h | 68 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, 0, 0, 0, 0); in xc0() 73 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, arg1, 0, 0, 0); in xc1() 77 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, arg1, arg2, 0, 0); in xc2() 83 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, in xc3() 90 sparc32_ipi_ops->cross_call(func, *cpu_online_mask, in xc4()
|
/linux-6.12.1/kernel/irq/ |
D | cpuhotplug.c | 40 !cpumask_intersects(m, cpu_online_mask)) { in irq_needs_fixup() 113 if (!cpumask_intersects(affinity, cpu_online_mask)) { in migrate_one_irq() 123 affinity = cpu_online_mask; in migrate_one_irq() 139 if (err == -ENOSPC && !irqd_affinity_is_managed(d) && affinity != cpu_online_mask) { in migrate_one_irq() 143 affinity = cpu_online_mask; in migrate_one_irq()
|
D | migration.c | 29 if (!cpumask_intersects(desc->pending_mask, cpu_online_mask)) { in irq_fixup_move_pending() 77 if (cpumask_intersects(desc->pending_mask, cpu_online_mask)) { in irq_move_masked_irq()
|
/linux-6.12.1/drivers/net/wireguard/ |
D | queueing.h | 110 !cpumask_test_cpu(cpu, cpu_online_mask))) { in wg_cpumask_choose_online() 111 cpu_index = id % cpumask_weight(cpu_online_mask); in wg_cpumask_choose_online() 112 cpu = cpumask_first(cpu_online_mask); in wg_cpumask_choose_online() 114 cpu = cpumask_next(cpu, cpu_online_mask); in wg_cpumask_choose_online() 127 int cpu = cpumask_next(READ_ONCE(*last_cpu), cpu_online_mask); in wg_cpumask_next_online() 129 cpu = cpumask_first(cpu_online_mask); in wg_cpumask_next_online()
|
/linux-6.12.1/arch/riscv/kernel/ |
D | unaligned_access_speed.c | 151 cpumask_and(&fast_except_me, &fast_misaligned_access, cpu_online_mask); in set_unaligned_access_static_branches_except_cpu() 170 cpumask_and(&fast_and_online, &fast_misaligned_access, cpu_online_mask); in set_unaligned_access_static_branches() 232 for_each_cpu(cpu, cpu_online_mask) { in check_unaligned_access_speed_all_cpus() 254 for_each_cpu(cpu, cpu_online_mask) { in check_unaligned_access_speed_all_cpus()
|
D | sys_hwprobe.c | 277 cpumask_copy(&cpus, cpu_online_mask); in hwprobe_get_values() 290 cpumask_and(&cpus, &cpus, cpu_online_mask); in hwprobe_get_values() 338 cpumask_copy(&cpus, cpu_online_mask); in hwprobe_get_cpus() 340 cpumask_and(&cpus, &cpus, cpu_online_mask); in hwprobe_get_cpus() 416 hwprobe_one_pair(&pair, cpu_online_mask); in init_hwprobe_vdso_data()
|
D | smp.c | 258 cpumask_copy(&mask, cpu_online_mask); in smp_send_stop() 273 cpumask_pr_args(cpu_online_mask)); in smp_send_stop() 310 cpumask_copy(&mask, cpu_online_mask); in crash_smp_send_stop()
|
/linux-6.12.1/include/asm-generic/ |
D | topology.h | 49 #define cpumask_of_node(node) ((node) == 0 ? cpu_online_mask : cpu_none_mask) 51 #define cpumask_of_node(node) ((void)(node), cpu_online_mask)
|
/linux-6.12.1/arch/x86/xen/ |
D | smp.c | 149 for_each_cpu_and(cpu, mask, cpu_online_mask) in __xen_send_IPI_mask() 220 __xen_send_IPI_mask(cpu_online_mask, xen_vector); in xen_send_IPI_all() 241 for_each_cpu_and(cpu, mask, cpu_online_mask) { in xen_send_IPI_mask_allbutself() 251 xen_send_IPI_mask_allbutself(cpu_online_mask, vector); in xen_send_IPI_allbutself()
|
/linux-6.12.1/arch/x86/kernel/apic/ |
D | bigsmp_32.c | 23 default_send_IPI_mask_allbutself_phys(cpu_online_mask, vector); in bigsmp_send_IPI_allbutself() 28 default_send_IPI_mask_sequence_phys(cpu_online_mask, vector); in bigsmp_send_IPI_all()
|
D | ipi.c | 59 __apic_send_IPI_mask_allbutself(cpu_online_mask, vector); in apic_send_IPI_allbutself() 86 if (!cpumask_or_equal(mask, cpumask_of(cpu), cpu_online_mask)) in native_send_call_func_ipi() 286 WARN_ON(mask & ~cpumask_bits(cpu_online_mask)[0]); in default_send_IPI_mask_logical()
|
/linux-6.12.1/arch/powerpc/kernel/ |
D | irq.c | 364 if (cpumask_equal(mask, cpu_online_mask)) { in irq_choose_cpu() 373 irq_rover = cpumask_next(irq_rover, cpu_online_mask); in irq_choose_cpu() 375 irq_rover = cpumask_first(cpu_online_mask); in irq_choose_cpu() 381 cpuid = cpumask_first_and(mask, cpu_online_mask); in irq_choose_cpu()
|
D | rtasd.c | 435 cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in rtas_event_scan() 437 cpu = cpumask_first(cpu_online_mask); in rtas_event_scan() 490 schedule_delayed_work_on(cpumask_first(cpu_online_mask), in start_event_scan()
|
/linux-6.12.1/arch/arm/mach-omap2/ |
D | cpuidle44xx.c | 133 if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) { in omap_enter_idle_coupled() 191 if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) { in omap_enter_idle_coupled() 329 return cpuidle_register(idle_driver, cpu_online_mask); in omap4_idle_init()
|
/linux-6.12.1/include/linux/ |
D | nmi.h | 162 arch_trigger_cpumask_backtrace(cpu_online_mask, -1); in trigger_all_cpu_backtrace() 168 arch_trigger_cpumask_backtrace(cpu_online_mask, exclude_cpu); in trigger_allbutcpu_cpu_backtrace()
|
D | topology.h | 254 return cpumask_nth_and(cpu, cpus, cpu_online_mask); in sched_numa_find_nth_cpu() 278 cpu_online_mask, \
|
D | smp.h | 71 on_each_cpu_cond_mask(NULL, func, info, wait, cpu_online_mask); in on_each_cpu() 105 on_each_cpu_cond_mask(cond_func, func, info, wait, cpu_online_mask); in on_each_cpu_cond()
|
/linux-6.12.1/Documentation/translations/zh_TW/admin-guide/ |
D | cputopology.rst | 64 導致未上線的CPU。[~cpu_online_mask + cpus >= NR_CPUS] 66 online: 在線的CPU,可供調度使用。[cpu_online_mask]
|
/linux-6.12.1/Documentation/translations/zh_CN/admin-guide/ |
D | cputopology.rst | 64 导致未上线的CPU。[~cpu_online_mask + cpus >= NR_CPUS] 66 online: 在线的CPU,可供调度使用。[cpu_online_mask]
|
/linux-6.12.1/drivers/powercap/ |
D | dtpm_cpu.c | 50 nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(pd->cpus)); in set_pd_power_limit() 82 for_each_cpu_and(cpu, pd_mask, cpu_online_mask) in scale_pd_power_uw() 126 nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(em->cpus)); in update_pd_power_uw()
|
/linux-6.12.1/arch/x86/kernel/ |
D | tsc_sync.c | 101 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in tsc_sync_check_timer_fn() 103 next_cpu = cpumask_first(cpu_online_mask); in tsc_sync_check_timer_fn() 455 smp_call_function_single(cpumask_first(cpu_online_mask), check_tsc_sync_source, in check_tsc_sync_target()
|
/linux-6.12.1/kernel/time/ |
D | clocksource.c | 307 cpumask_copy(&cpus_chosen, cpu_online_mask); in clocksource_verify_choose_cpus() 318 cpu = cpumask_first(cpu_online_mask); in clocksource_verify_choose_cpus() 320 cpu = cpumask_next(cpu, cpu_online_mask); in clocksource_verify_choose_cpus() 338 cpu = cpumask_next(cpu - 1, cpu_online_mask); in clocksource_verify_choose_cpus() 340 cpu = cpumask_first(cpu_online_mask); in clocksource_verify_choose_cpus() 581 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in clocksource_watchdog() 583 next_cpu = cpumask_first(cpu_online_mask); in clocksource_watchdog() 603 add_timer_on(&watchdog_timer, cpumask_first(cpu_online_mask)); in clocksource_start_watchdog()
|
/linux-6.12.1/arch/s390/kernel/ |
D | processor.c | 339 unsigned long first = cpumask_first(cpu_online_mask); in show_cpuinfo() 353 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_update() 355 *pos = cpumask_first(cpu_online_mask); in c_update()
|
/linux-6.12.1/drivers/cpufreq/ |
D | speedstep-ich.c | 263 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); in speedstep_target() 298 policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask); in speedstep_cpu_init()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/ |
D | irq_affinity.c | 25 for_each_cpu_and(cpu, req_mask, cpu_online_mask) { in cpu_get_least_loaded() 40 best_cpu = cpumask_first(cpu_online_mask); in cpu_get_least_loaded()
|