/linux-6.12.1/kernel/ |
D | panic.c | 198 int old_cpu, this_cpu; in nmi_panic() local 200 old_cpu = PANIC_CPU_INVALID; in nmi_panic() 204 if (atomic_try_cmpxchg(&panic_cpu, &old_cpu, this_cpu)) in nmi_panic() 206 else if (old_cpu != this_cpu) in nmi_panic() 293 int old_cpu, this_cpu; in panic() local 330 old_cpu = PANIC_CPU_INVALID; in panic() 334 if (atomic_try_cmpxchg(&panic_cpu, &old_cpu, this_cpu)) { in panic() 336 } else if (old_cpu != this_cpu) in panic()
|
D | crash_core.c | 131 int old_cpu, this_cpu; in crash_kexec() local 138 old_cpu = PANIC_CPU_INVALID; in crash_kexec() 141 if (atomic_try_cmpxchg(&panic_cpu, &old_cpu, this_cpu)) { in crash_kexec()
|
/linux-6.12.1/drivers/xen/events/ |
D | events_internal.h | 19 unsigned int old_cpu); 65 unsigned int old_cpu) in xen_evtchn_port_bind_to_cpu() argument 67 evtchn_ops->bind_to_cpu(evtchn, cpu, old_cpu); in xen_evtchn_port_bind_to_cpu()
|
D | events_2l.c | 56 unsigned int old_cpu) in evtchn_2l_bind_to_cpu() argument 58 clear_bit(evtchn, BM(per_cpu(cpu_evtchn_mask, old_cpu))); in evtchn_2l_bind_to_cpu()
|
D | events_fifo.c | 189 unsigned int old_cpu) in evtchn_fifo_bind_to_cpu() argument
|
/linux-6.12.1/arch/x86/kernel/ |
D | smp.c | 151 unsigned int old_cpu, this_cpu; in native_stop_other_cpus() local 158 old_cpu = -1; in native_stop_other_cpus() 160 if (!atomic_try_cmpxchg(&stopping_cpu, &old_cpu, this_cpu)) in native_stop_other_cpus()
|
/linux-6.12.1/arch/mips/sibyte/sb1250/ |
D | irq.c | 73 int i = 0, old_cpu, cpu, int_on; in sb1250_set_affinity() local 87 old_cpu = sb1250_irq_owner[irq]; in sb1250_set_affinity() 88 cur_ints = ____raw_readq(IOADDR(A_IMR_MAPPER(old_cpu) + in sb1250_set_affinity() 94 ____raw_writeq(cur_ints, IOADDR(A_IMR_MAPPER(old_cpu) + in sb1250_set_affinity()
|
/linux-6.12.1/arch/mips/sibyte/bcm1480/ |
D | irq.c | 81 int i = 0, old_cpu, cpu, int_on, k; in bcm1480_set_affinity() local 94 old_cpu = bcm1480_irq_owner[irq]; in bcm1480_set_affinity() 101 …cur_ints = ____raw_readq(IOADDR(A_BCM1480_IMR_MAPPER(old_cpu) + R_BCM1480_IMR_INTERRUPT_MASK_H + (… in bcm1480_set_affinity() 106 …____raw_writeq(cur_ints, IOADDR(A_BCM1480_IMR_MAPPER(old_cpu) + R_BCM1480_IMR_INTERRUPT_MASK_H + (… in bcm1480_set_affinity()
|
/linux-6.12.1/kernel/debug/kdb/ |
D | kdb_io.c | 623 int this_cpu, old_cpu; in vkdb_printf() local 635 old_cpu = cmpxchg(&kdb_printf_cpu, -1, this_cpu); in vkdb_printf() 636 if (old_cpu == -1 || old_cpu == this_cpu) in vkdb_printf() 886 smp_store_release(&kdb_printf_cpu, old_cpu); in vkdb_printf()
|
/linux-6.12.1/kernel/locking/ |
D | osq_lock.c | 56 int old_cpu) in osq_wait_next() argument 62 atomic_cmpxchg_acquire(&lock->tail, curr, old_cpu) == curr) { in osq_wait_next()
|
/linux-6.12.1/tools/perf/scripts/python/ |
D | sched-migration.py | 191 def migrate(self, ts_list, new, old_cpu, new_cpu): argument 192 if old_cpu == new_cpu: 194 old_rq = self.prev.rqs[old_cpu] 196 self.rqs[old_cpu] = out_rq 207 self.event_cpus.append(old_cpu)
|
/linux-6.12.1/drivers/irqchip/ |
D | irq-bcm6345-l1.c | 193 unsigned int old_cpu = cpu_for_irq(intc, d); in bcm6345_l1_set_affinity() local 205 if (old_cpu != new_cpu) { in bcm6345_l1_set_affinity() 206 enabled = intc->cpus[old_cpu]->enable_cache[word] & mask; in bcm6345_l1_set_affinity()
|
/linux-6.12.1/arch/x86/kernel/cpu/ |
D | mshyperv.c | 266 unsigned int old_cpu, this_cpu; in hv_nmi_unknown() local 271 old_cpu = -1; in hv_nmi_unknown() 273 if (!atomic_try_cmpxchg(&nmi_cpu, &old_cpu, this_cpu)) in hv_nmi_unknown()
|
/linux-6.12.1/drivers/hv/ |
D | hyperv_vmbus.h | 453 static inline void hv_update_allocated_cpus(unsigned int old_cpu, in hv_update_allocated_cpus() argument 457 hv_clear_allocated_cpu(old_cpu); in hv_update_allocated_cpus()
|
/linux-6.12.1/drivers/infiniband/hw/hfi1/ |
D | affinity.c | 745 int i, old_cpu; in hfi1_update_sdma_affinity() local 755 old_cpu = sde->cpu; in hfi1_update_sdma_affinity() 778 if (cpumask_test_cpu(old_cpu, &other_msix->mask)) in hfi1_update_sdma_affinity() 781 cpumask_clear_cpu(old_cpu, &set->mask); in hfi1_update_sdma_affinity() 782 cpumask_clear_cpu(old_cpu, &set->used); in hfi1_update_sdma_affinity()
|
/linux-6.12.1/arch/x86/events/intel/ |
D | uncore.c | 1471 static void uncore_change_type_ctx(struct intel_uncore_type *type, int old_cpu, in uncore_change_type_ctx() argument 1478 die = topology_logical_die_id(old_cpu < 0 ? new_cpu : old_cpu); in uncore_change_type_ctx() 1484 if (old_cpu < 0) { in uncore_change_type_ctx() 1493 WARN_ON_ONCE(box->cpu != -1 && box->cpu != old_cpu); in uncore_change_type_ctx() 1495 cpumask_clear_cpu(old_cpu, &pmu->cpu_mask); in uncore_change_type_ctx() 1502 perf_pmu_migrate_context(&pmu->pmu, old_cpu, new_cpu); in uncore_change_type_ctx() 1509 int old_cpu, int new_cpu) in uncore_change_context() argument 1512 uncore_change_type_ctx(*uncores, old_cpu, new_cpu); in uncore_change_context()
|
/linux-6.12.1/arch/powerpc/kernel/ |
D | fadump.c | 687 int old_cpu, this_cpu; in crash_fadump() local 702 old_cpu = cmpxchg(&crashing_cpu, -1, this_cpu); in crash_fadump() 704 if (old_cpu != -1) { in crash_fadump()
|
/linux-6.12.1/arch/powerpc/perf/ |
D | imc-pmu.c | 344 static void nest_change_cpu_context(int old_cpu, int new_cpu) in nest_change_cpu_context() argument 348 if (old_cpu < 0 || new_cpu < 0) in nest_change_cpu_context() 352 perf_pmu_migrate_context(&(*pn)->pmu, old_cpu, new_cpu); in nest_change_cpu_context()
|