Home
last modified time | relevance | path

Searched refs:smp_mb (Results 1 – 25 of 367) sorted by relevance

12345678910>>...15

/linux-6.12.1/arch/arc/include/asm/
Dspinlock.h40 smp_mb(); in arch_spin_lock()
62 smp_mb(); in arch_spin_trylock()
69 smp_mb(); in arch_spin_unlock()
105 smp_mb(); in arch_read_lock()
129 smp_mb(); in arch_read_trylock()
163 smp_mb(); in arch_write_lock()
188 smp_mb(); in arch_write_trylock()
197 smp_mb(); in arch_read_unlock()
215 smp_mb(); in arch_write_unlock()
231 smp_mb(); in arch_spin_lock()
[all …]
Dfutex.h20 smp_mb(); \
42 smp_mb() \
48 smp_mb(); \
69 smp_mb() \
133 smp_mb(); in futex_atomic_cmpxchg_inatomic()
160 smp_mb(); in futex_atomic_cmpxchg_inatomic()
Datomic64-arcv2.h145 smp_mb(); in ATOMIC64_OPS()
158 smp_mb(); in ATOMIC64_OPS()
168 smp_mb(); in arch_atomic64_xchg()
179 smp_mb(); in arch_atomic64_xchg()
189 smp_mb(); in arch_atomic64_dec_if_positive()
203 smp_mb(); in arch_atomic64_dec_if_positive()
213 smp_mb(); in arch_atomic64_fetch_add_unless()
229 smp_mb(); in arch_atomic64_fetch_add_unless()
/linux-6.12.1/arch/arm/include/asm/
Dspinlock.h78 smp_mb(); in arch_spin_lock()
100 smp_mb(); in arch_spin_trylock()
109 smp_mb(); in arch_spin_unlock()
155 smp_mb(); in arch_write_lock()
175 smp_mb(); in arch_write_trylock()
184 smp_mb(); in arch_write_unlock()
224 smp_mb(); in arch_read_lock()
231 smp_mb(); in arch_read_unlock()
266 smp_mb(); in arch_read_trylock()
Dfutex.h28 smp_mb(); \
56 smp_mb(); in futex_atomic_cmpxchg_inatomic()
73 smp_mb(); in futex_atomic_cmpxchg_inatomic()
Datomic.h133 smp_mb(); in arch_atomic_fetch_add_unless()
150 smp_mb(); in arch_atomic_fetch_add_unless()
459 smp_mb(); in arch_atomic64_dec_if_positive()
476 smp_mb(); in arch_atomic64_dec_if_positive()
487 smp_mb(); in arch_atomic64_fetch_add_unless()
506 smp_mb(); in arch_atomic64_fetch_add_unless()
/linux-6.12.1/arch/alpha/include/asm/
Datomic.h72 smp_mb(); \
90 smp_mb(); \
126 smp_mb(); \
145 smp_mb(); \
206 smp_mb(); in ATOMIC_OPS()
221 smp_mb(); in ATOMIC_OPS()
229 smp_mb(); in arch_atomic64_fetch_add_unless()
244 smp_mb(); in arch_atomic64_fetch_add_unless()
252 smp_mb(); in arch_atomic64_dec_if_positive()
266 smp_mb(); in arch_atomic64_dec_if_positive()
Dcmpxchg.h258 smp_mb(); \
261 smp_mb(); \
270 smp_mb(); \
273 smp_mb(); \
/linux-6.12.1/kernel/sched/
Dmembarrier.c170 smp_mb(); /* IPIs should be serializing but paranoid. */ in ipi_mb()
185 smp_mb(); /* IPIs should be serializing but paranoid. */ in ipi_sync_core()
199 smp_mb(); in ipi_rseq()
217 smp_mb(); in ipi_sync_rq_state()
227 smp_mb(); in membarrier_exec_mmap()
260 smp_mb(); /* system call entry is not a mb. */ in membarrier_global_expedited()
310 smp_mb(); /* exit from system call is not a mb */ in membarrier_global_expedited()
354 smp_mb(); /* system call entry is not a mb. */ in membarrier_private_expedited()
431 smp_mb(); /* exit from system call is not a mb */ in membarrier_private_expedited()
452 smp_mb(); in sync_runqueues_membarrier_state()
/linux-6.12.1/tools/include/asm/
Dbarrier.h46 #ifndef smp_mb
47 # define smp_mb() mb() macro
53 smp_mb(); \
62 smp_mb(); \
/linux-6.12.1/tools/memory-model/litmus-tests/
DIRIW+fencembonceonces+OnceOnce.litmus6 * Test of independent reads from independent writes with smp_mb()
7 * between each pairs of reads. In other words, is smp_mb() sufficient to
26 smp_mb();
41 smp_mb();
DR+fencembonceonces.litmus6 * This is the fully ordered (via smp_mb()) version of one of the classic
17 smp_mb();
26 smp_mb();
DSB+fencembonceonces.litmus19 smp_mb();
28 smp_mb();
DREADME24 Test of independent reads from independent writes with smp_mb()
25 between each pairs of reads. In other words, is smp_mb()
41 separated by smp_mb(). This addition of an external process to
53 Does a control dependency and an smp_mb() suffice for the
109 This is the fully ordered (via smp_mb()) version of one of
114 As above, but without the smp_mb() invocations.
117 This is the fully ordered (again, via smp_mb() version of store
122 As above, but without the smp_mb() invocations.
/linux-6.12.1/include/asm-generic/
Dbarrier.h98 #ifndef smp_mb
99 #define smp_mb() do { kcsan_mb(); __smp_mb(); } while (0) macro
112 #ifndef smp_mb
113 #define smp_mb() barrier() macro
302 # define smp_mb__after_switch_mm() smp_mb()
/linux-6.12.1/arch/sh/kernel/
Dftrace.c137 smp_mb(); in arch_ftrace_nmi_enter()
143 smp_mb(); in arch_ftrace_nmi_exit()
174 smp_mb(); in do_ftrace_mod_code()
179 smp_mb(); in do_ftrace_mod_code()
184 smp_mb(); in do_ftrace_mod_code()
/linux-6.12.1/tools/virtio/ringtest/
Dmain.h117 #define smp_mb() asm volatile("lock; addl $0,-132(%%rsp)" ::: "memory", "cc") macro
119 #define smp_mb() asm volatile("dmb ish" ::: "memory") macro
125 #define smp_mb() __sync_synchronize() macro
188 smp_mb(); /* Enforce dependency ordering from x */ \
Dring.c183 smp_mb(); in enable_call()
193 smp_mb(); in kick_available()
215 smp_mb(); in enable_kick()
260 smp_mb(); in call_used()
Dvirtio_ring_0_9.c222 smp_mb(); in enable_call()
232 smp_mb(); in kick_available()
254 smp_mb(); in enable_kick()
325 smp_mb(); in call_used()
/linux-6.12.1/arch/arm64/include/asm/vdso/
Dcompat_barrier.h27 #undef smp_mb
31 #define smp_mb() aarch32_smp_mb() macro
/linux-6.12.1/drivers/comedi/drivers/
Ddyna_pci10xx.c80 smp_mb(); in dyna_pci10xx_insn_read_ai()
111 smp_mb(); in dyna_pci10xx_insn_write_ao()
130 smp_mb(); in dyna_pci10xx_di_insn_bits()
150 smp_mb(); in dyna_pci10xx_do_insn_bits()
/linux-6.12.1/arch/hexagon/include/asm/
Dspinlock.h111 smp_mb(); in arch_write_unlock()
132 smp_mb(); in arch_spin_unlock()
/linux-6.12.1/arch/riscv/kernel/
Dcpu_ops_sbi.c73 smp_mb(); in sbi_cpu_start()
77 smp_mb(); in sbi_cpu_start()
/linux-6.12.1/tools/memory-model/Documentation/
Drecipes.txt149 smp_mb();
183 smp_mb();
337 * smp_wmb() (B) smp_mb() (D)
344 Of course, given that smp_mb() is strictly stronger than either smp_wmb()
346 smp_wmb() would also work with smp_mb() replacing either or both of the
374 smp_mb();
392 * smp_wmb() (B) smp_mb() (D)
457 smp_mb();
483 smp_mb();
490 smp_mb();
[all …]
/linux-6.12.1/fs/bcachefs/
Dsix.c158 smp_mb(); in __do_six_trylock()
167 smp_mb(); in __do_six_trylock()
182 smp_mb(); in __do_six_trylock()
195 smp_mb(); in __do_six_trylock()
577 smp_mb(); /* unlock barrier */ in do_six_unlock_type()
579 smp_mb(); /* between unlocking and checking for waiters */ in do_six_unlock_type()

12345678910>>...15