Home
last modified time | relevance | path

Searched refs:locked (Results 1 – 25 of 430) sorted by relevance

12345678910>>...18

/linux-6.12.1/Documentation/i2c/
Di2c-topology.rst40 mux-locked or parent-locked muxes.
43 Mux-locked muxes
46 Mux-locked muxes does not lock the entire parent adapter during the
48 adapter are locked. Mux-locked muxes are mostly interesting if the
50 their tasks. Since the parent adapter is not fully locked during the
55 Mux-locked Example
62 | root |--+--| locked | '--------'
87 Mux-locked caveats
90 When using a mux-locked mux, be aware of the following restrictions:
93 If you build a topology with a mux-locked mux being the parent
[all …]
/linux-6.12.1/kernel/locking/
Dqspinlock_paravirt.h92 try_cmpxchg_acquire(&lock->locked, &old, _Q_LOCKED_VAL)) { in pv_hybrid_queued_unfair_trylock()
124 return !READ_ONCE(lock->locked) && in trylock_clear_pending()
299 if (READ_ONCE(node->locked)) in pv_wait_node()
319 if (!READ_ONCE(node->locked)) { in pv_wait_node()
340 !READ_ONCE(node->locked)); in pv_wait_node()
387 WRITE_ONCE(lock->locked, _Q_SLOW_VAL); in pv_kick_node()
452 if (xchg(&lock->locked, _Q_SLOW_VAL) == 0) { in pv_wait_head_or_lock()
458 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in pv_wait_head_or_lock()
466 pv_wait(&lock->locked, _Q_SLOW_VAL); in pv_wait_head_or_lock()
499 __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked) in __pv_queued_spin_unlock_slowpath() argument
[all …]
Dosq_lock.c17 int locked; /* 1 if lock acquired */ member
100 node->locked = 0; in osq_lock()
146 if (smp_cond_load_relaxed(&node->locked, VAL || need_resched() || in osq_lock()
173 if (smp_load_acquire(&node->locked)) in osq_lock()
228 WRITE_ONCE(next->locked, 1); in osq_unlock()
234 WRITE_ONCE(next->locked, 1); in osq_unlock()
Dmcs_spinlock.h20 int locked; /* 1 if lock acquired */ member
70 node->locked = 0; in mcs_spin_lock()
94 arch_mcs_spin_lock_contended(&node->locked); in mcs_spin_lock()
118 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
Dqspinlock.c261 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in set_locked()
380 smp_cond_load_acquire(&lock->locked, !VAL); in queued_spin_lock_slowpath()
434 node->locked = 0; in queued_spin_lock_slowpath()
474 arch_mcs_spin_lock_contended(&node->locked); in queued_spin_lock_slowpath()
509 goto locked; in queued_spin_lock_slowpath()
513 locked: in queued_spin_lock_slowpath()
553 arch_mcs_spin_unlock_contended(&next->locked); in queued_spin_lock_slowpath()
/linux-6.12.1/tools/testing/selftests/net/forwarding/
Dbridge_locked_port.sh102 bridge link set dev $swp1 locked on
112 bridge link set dev $swp1 locked off
133 bridge link set dev $swp1 locked on
142 bridge link set dev $swp1 locked off
161 bridge link set dev $swp1 locked on
170 bridge link set dev $swp1 locked off
187 bridge link set dev $swp1 learning on locked on
195 bridge link set dev $swp1 learning on locked on mab on
212 bridge link set dev $swp1 learning off locked off mab off
226 bridge link set dev $swp1 learning on locked on mab on
[all …]
/linux-6.12.1/scripts/coccinelle/locks/
Ddouble_lock.cocci17 @locked@
42 position p1 != locked.p1;
43 position locked.p;
45 expression x <= locked.E1;
46 expression E,locked.E1;
65 expression x <= locked.E1;
66 expression locked.E1;
69 position locked.p,p1,p2;
Dcall_kern.cocci39 @locked exists@
74 @depends on locked && patch@
81 @depends on locked && !patch@
90 p1 << locked.p1;
91 p2 << locked.p2;
101 p1 << locked.p1;
102 p2 << locked.p2;
/linux-6.12.1/drivers/net/ethernet/intel/ice/
Dice_dcb_lib.h18 int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked);
24 ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked);
26 void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked);
28 int ice_init_pf_dcb(struct ice_pf *pf, bool locked);
93 ice_init_pf_dcb(struct ice_pf *pf, bool __always_unused locked) in ice_init_pf_dcb() argument
102 bool __always_unused locked) in ice_pf_dcb_cfg() argument
131 static inline void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked) { } in ice_pf_dcb_recfg() argument
Dice_dcb_lib.c283 static void ice_dcb_ena_dis_vsi(struct ice_pf *pf, bool ena, bool locked) in ice_dcb_ena_dis_vsi() argument
297 ice_ena_vsi(vsi, locked); in ice_dcb_ena_dis_vsi()
299 ice_dis_vsi(vsi, locked); in ice_dcb_ena_dis_vsi()
350 int ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked) in ice_pf_dcb_cfg() argument
415 if (!locked) in ice_pf_dcb_cfg()
449 if (!locked) in ice_pf_dcb_cfg()
605 static int ice_dcb_init_cfg(struct ice_pf *pf, bool locked) in ice_dcb_init_cfg() argument
620 if (ice_pf_dcb_cfg(pf, newcfg, locked)) in ice_dcb_init_cfg()
634 int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked) in ice_dcb_sw_dflt_cfg() argument
667 ret = ice_pf_dcb_cfg(pf, dcbcfg, locked); in ice_dcb_sw_dflt_cfg()
[all …]
/linux-6.12.1/drivers/media/dvb-frontends/
Dlgs8gxx.c295 static int lgs8gxx_is_locked(struct lgs8gxx_state *priv, u8 *locked) in lgs8gxx_is_locked() argument
308 *locked = ((t & 0x80) == 0x80) ? 1 : 0; in lgs8gxx_is_locked()
310 *locked = ((t & 0xC0) == 0xC0) ? 1 : 0; in lgs8gxx_is_locked()
315 static int lgs8gxx_wait_ca_lock(struct lgs8gxx_state *priv, u8 *locked) in lgs8gxx_wait_ca_lock() argument
331 *locked = (ret == 0) ? 1 : 0; in lgs8gxx_wait_ca_lock()
359 u8 *locked) in lgs8gxx_autolock_gi() argument
384 err = lgs8gxx_wait_ca_lock(priv, locked); in lgs8gxx_autolock_gi()
385 if (err || !(*locked)) in lgs8gxx_autolock_gi()
393 *locked = 0; in lgs8gxx_autolock_gi()
403 u8 locked = 0, tmp_gi; in lgs8gxx_auto_detect() local
[all …]
Datbm8830.c147 static int is_locked(struct atbm_state *priv, u8 *locked) in is_locked() argument
153 if (locked != NULL) in is_locked()
154 *locked = (status == 1); in is_locked()
261 u8 locked = 0; in atbm8830_set_fe() local
277 is_locked(priv, &locked); in atbm8830_set_fe()
278 if (locked != 0) { in atbm8830_set_fe()
329 u8 locked = 0; in atbm8830_read_status() local
335 is_locked(priv, &locked); in atbm8830_read_status()
336 if (locked) { in atbm8830_read_status()
/linux-6.12.1/mm/
Dgup.c1150 int *locked) in faultin_page() argument
1195 *locked = 0; in faultin_page()
1217 *locked = 0; in faultin_page()
1418 int *locked) in __get_user_pages() argument
1486 PTR_ERR(page) == -EMLINK, locked); in __get_user_pages()
1712 int *locked, in __get_user_pages_locked() argument
1725 if (!*locked) { in __get_user_pages_locked()
1729 *locked = 1; in __get_user_pages_locked()
1752 locked); in __get_user_pages_locked()
1760 if (!*locked) { in __get_user_pages_locked()
[all …]
Dmlock.c614 unsigned long locked; in do_mlock() local
628 locked = len >> PAGE_SHIFT; in do_mlock()
633 locked += current->mm->locked_vm; in do_mlock()
634 if ((locked > lock_limit) && (!capable(CAP_IPC_LOCK))) { in do_mlock()
641 locked -= count_mm_mlocked_page_nr(current->mm, in do_mlock()
646 if ((locked <= lock_limit) || capable(CAP_IPC_LOCK)) in do_mlock()
793 unsigned long lock_limit, locked; in user_shm_lock() local
797 locked = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in user_shm_lock()
802 memlock = inc_rlimit_ucounts(ucounts, UCOUNT_RLIMIT_MEMLOCK, locked); in user_shm_lock()
805 dec_rlimit_ucounts(ucounts, UCOUNT_RLIMIT_MEMLOCK, locked); in user_shm_lock()
[all …]
Dmemcontrol-v1.h99 bool memcg1_oom_prepare(struct mem_cgroup *memcg, bool *locked);
100 void memcg1_oom_finish(struct mem_cgroup *memcg, bool locked);
136 static inline bool memcg1_oom_prepare(struct mem_cgroup *memcg, bool *locked) { return true; } in memcg1_oom_prepare() argument
137 static inline void memcg1_oom_finish(struct mem_cgroup *memcg, bool locked) {} in memcg1_oom_finish() argument
Dcompaction.c566 unsigned long flags, bool *locked, struct compact_control *cc) in compact_unlock_should_abort() argument
568 if (*locked) { in compact_unlock_should_abort()
570 *locked = false; in compact_unlock_should_abort()
598 bool locked = false; in isolate_freepages_block() local
619 &locked, cc)) in isolate_freepages_block()
646 if (!locked) { in isolate_freepages_block()
647 locked = compact_lock_irqsave(&cc->zone->lock, in isolate_freepages_block()
682 if (locked) in isolate_freepages_block()
875 struct lruvec *locked = NULL; in isolate_migratepages_block() local
946 if (locked) { in isolate_migratepages_block()
[all …]
/linux-6.12.1/drivers/mtd/spi-nor/
Dotp.c253 int ret, locked; in spi_nor_mtd_otp_info() local
266 locked = ops->is_locked(nor, i); in spi_nor_mtd_otp_info()
267 if (locked < 0) { in spi_nor_mtd_otp_info()
268 ret = locked; in spi_nor_mtd_otp_info()
272 buf->locked = !!locked; in spi_nor_mtd_otp_info()
289 int locked; in spi_nor_mtd_otp_range_is_locked() local
298 locked = ops->is_locked(nor, region); in spi_nor_mtd_otp_range_is_locked()
300 if (locked) in spi_nor_mtd_otp_range_is_locked()
301 return locked; in spi_nor_mtd_otp_range_is_locked()
/linux-6.12.1/drivers/thermal/intel/
Dintel_quark_dts_thermal.c100 bool locked; member
129 if (!aux_entry->locked) { in soc_dts_enable()
157 if (!aux_entry->locked) { in soc_dts_disable()
206 if (aux_entry->locked) { in update_trip_temp()
316 if (!aux_entry->locked) { in free_soc_dts()
351 aux_entry->locked = !!(out & QRK_DTS_LOCK_BIT); in alloc_soc_dts()
354 if (!aux_entry->locked) { in alloc_soc_dts()
/linux-6.12.1/sound/core/seq/
Dseq_queue.c88 static struct snd_seq_queue *queue_new(int owner, int locked) in queue_new() argument
114 q->locked = locked; in queue_new()
160 struct snd_seq_queue *snd_seq_queue_alloc(int client, int locked, unsigned int info_flags) in snd_seq_queue_alloc() argument
164 q = queue_new(client, locked); in snd_seq_queue_alloc()
342 return (q->owner == client) || (!q->locked && !q->klocked); in check_access()
385 int snd_seq_queue_set_owner(int queueid, int client, int locked) in snd_seq_queue_set_owner() argument
398 q->locked = locked ? 1 : 0; in snd_seq_queue_set_owner()
718 bool locked; in snd_seq_info_queues_read() local
733 locked = q->locked; in snd_seq_info_queues_read()
739 snd_iprintf(buffer, "lock status : %s\n", locked ? "Locked" : "Free"); in snd_seq_info_queues_read()
Dseq_queue.h29 bool locked; /* timer is only accesibble by owner if set */ member
57 struct snd_seq_queue *snd_seq_queue_alloc(int client, int locked, unsigned int flags);
86 int snd_seq_queue_set_owner(int queueid, int client, int locked);
/linux-6.12.1/drivers/gpu/drm/xe/
Dxe_pcode.c124 u32 *status, bool atomic, int timeout_us, bool locked) in pcode_try_request() argument
131 if (locked) in pcode_try_request()
270 int xe_pcode_ready(struct xe_device *xe, bool locked) in xe_pcode_ready() argument
283 if (locked) in xe_pcode_ready()
289 &status, false, timeout_us, locked); in xe_pcode_ready()
291 if (locked) in xe_pcode_ready()
/linux-6.12.1/drivers/staging/vme_user/
Dvme_bridge.h19 int locked; member
32 int locked; member
65 int locked; member
76 int locked; member
/linux-6.12.1/drivers/tty/
Dtty_ldsem.c234 int locked = 0; in down_write_failed() local
265 locked = writer_trylock(sem); in down_write_failed()
266 if (locked) in down_write_failed()
270 if (!locked) in down_write_failed()
280 if (!locked && list_empty(&sem->write_wait)) in down_write_failed()
288 if (!locked) in down_write_failed()
/linux-6.12.1/Documentation/ABI/testing/
Dsysfs-platform-intel-pmc15 * bit 31 - global reset is locked
20 in case the register is not locked.
21 The "global reset bit" should be locked on a production
/linux-6.12.1/drivers/infiniband/hw/qib/
Dqib_user_pages.c97 unsigned long locked, lock_limit; in qib_get_user_pages() local
102 locked = atomic64_add_return(num_pages, &current->mm->pinned_vm); in qib_get_user_pages()
104 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) { in qib_get_user_pages()

12345678910>>...18