/* * Copyright (c) 2014-2021 The Linux Foundation. All rights reserved. * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved. * * Permission to use, copy, modify, and/or distribute this software for * any purpose with or without fee is hereby granted, provided that the * above copyright notice and this permission notice appear in all * copies. * * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR * PERFORMANCE OF THIS SOFTWARE. */ /** * @file qdf_lock.h * This file abstracts locking operations. */ #ifndef _QDF_LOCK_H #define _QDF_LOCK_H #include #include #include #include #ifndef QDF_LOCK_STATS #define QDF_LOCK_STATS 0 #endif #ifndef QDF_LOCK_STATS_DESTROY_PRINT #define QDF_LOCK_STATS_DESTROY_PRINT 0 #endif #ifndef QDF_LOCK_STATS_BUG_ON #define QDF_LOCK_STATS_BUG_ON 0 #endif #ifndef QDF_LOCK_STATS_LIST #define QDF_LOCK_STATS_LIST 0 #endif /* Max hold time in micro seconds, 0 to disable detection*/ #ifdef VCPU_TIMESTOLEN #define QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_IRQ 400000 #else #define QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_IRQ 10000 #endif #define QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK 0 #if QDF_LOCK_STATS #define QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_BH 2000000 #else #define QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_BH 1000000 #endif #if !QDF_LOCK_STATS struct lock_stats {}; #define BEFORE_LOCK(x...) do {} while (0) #define AFTER_LOCK(x...) do {} while (0) #define BEFORE_TRYLOCK(x...) do {} while (0) #define AFTER_TRYLOCK(x...) do {} while (0) #define BEFORE_UNLOCK(x...) do {} while (0) #define qdf_lock_stats_create(x...) do {} while (0) #define qdf_lock_stats_destroy(x...) do {} while (0) #define qdf_lock_stats_init(x...) do {} while (0) #define qdf_lock_stats_deinit(x...) do {} while (0) #else void qdf_lock_stats_init(void); void qdf_lock_stats_deinit(void); struct qdf_lock_cookie; struct lock_stats { const char *initialization_fn; const char *acquired_by; int line; int acquired; int contended; uint64_t contention_time; uint64_t non_contention_time; uint64_t held_time; uint64_t last_acquired; uint64_t max_contention_wait; uint64_t max_held_time; int num_large_contentions; int num_large_holds; struct qdf_lock_cookie *cookie; }; #define LARGE_CONTENTION QDF_LOG_TIMESTAMP_CYCLES_PER_10_US #define BEFORE_LOCK(lock, was_locked) \ do { \ uint64_t BEFORE_LOCK_time; \ uint64_t AFTER_LOCK_time; \ bool BEFORE_LOCK_is_locked = was_locked; \ BEFORE_LOCK_time = qdf_get_log_timestamp_lightweight(); \ do {} while (0) #define AFTER_LOCK(lock, func) \ lock->stats.acquired_by = func; \ AFTER_LOCK_time = qdf_get_log_timestamp_lightweight(); \ lock->stats.acquired++; \ lock->stats.last_acquired = AFTER_LOCK_time; \ if (BEFORE_LOCK_is_locked) { \ lock->stats.contended++; \ lock->stats.contention_time += \ (AFTER_LOCK_time - BEFORE_LOCK_time); \ } else { \ lock->stats.non_contention_time += \ (AFTER_LOCK_time - BEFORE_LOCK_time); \ } \ \ if (AFTER_LOCK_time - BEFORE_LOCK_time > LARGE_CONTENTION) \ lock->stats.num_large_contentions++; \ \ if (AFTER_LOCK_time - BEFORE_LOCK_time > \ lock->stats.max_contention_wait) \ lock->stats.max_contention_wait = \ AFTER_LOCK_time - BEFORE_LOCK_time; \ } while (0) #define BEFORE_TRYLOCK(lock) \ do { \ uint64_t BEFORE_LOCK_time; \ uint64_t AFTER_LOCK_time; \ BEFORE_LOCK_time = qdf_get_log_timestamp_lightweight(); \ do {} while (0) #define AFTER_TRYLOCK(lock, trylock_return, func) \ AFTER_LOCK_time = qdf_get_log_timestamp_lightweight(); \ if (trylock_return) { \ lock->stats.acquired++; \ lock->stats.last_acquired = AFTER_LOCK_time; \ lock->stats.non_contention_time += \ (AFTER_LOCK_time - BEFORE_LOCK_time); \ lock->stats.acquired_by = func; \ } \ } while (0) /* max_hold_time in US */ #define BEFORE_UNLOCK(lock, max_hold_time) \ do {\ uint64_t BEFORE_UNLOCK_time; \ uint64_t held_time; \ BEFORE_UNLOCK_time = qdf_get_log_timestamp_lightweight(); \ \ if (unlikely(BEFORE_UNLOCK_time < lock->stats.last_acquired)) \ held_time = 0; \ else \ held_time = BEFORE_UNLOCK_time - lock->stats.last_acquired; \ \ lock->stats.held_time += held_time; \ \ if (held_time > lock->stats.max_held_time) \ lock->stats.max_held_time = held_time; \ \ if (held_time > LARGE_CONTENTION) \ lock->stats.num_large_holds++; \ if (QDF_LOCK_STATS_BUG_ON && max_hold_time && \ held_time > qdf_usecs_to_log_timestamp(max_hold_time)) { \ QDF_TRACE(QDF_MODULE_ID_QDF, QDF_TRACE_LEVEL_ERROR, \ "BEFORE_UNLOCK: lock held too long (%lluus)", \ qdf_log_timestamp_to_usecs(held_time)); \ QDF_BUG(0); \ } \ lock->stats.acquired_by = NULL; \ } while (0) void qdf_lock_stats_cookie_destroy(struct lock_stats *stats); void qdf_lock_stats_cookie_create(struct lock_stats *stats, const char *func, int line); static inline void qdf_lock_stats_destroy(struct lock_stats *stats) { if (QDF_LOCK_STATS_DESTROY_PRINT) { QDF_TRACE(QDF_MODULE_ID_QDF, QDF_TRACE_LEVEL_DEBUG, "%s: lock: %s %d \t" "acquired:\t%d\tcontended:\t%d\t" "contention_time\t%llu\tmax_contention_wait:\t%llu\t" "non_contention_time\t%llu\t" "held_time\t%llu\tmax_held:\t%llu" , __func__, stats->initialization_fn, stats->line, stats->acquired, stats->contended, qdf_log_timestamp_to_usecs(stats->contention_time), qdf_log_timestamp_to_usecs(stats->max_contention_wait), qdf_log_timestamp_to_usecs(stats->non_contention_time), qdf_log_timestamp_to_usecs(stats->held_time), qdf_log_timestamp_to_usecs(stats->max_held_time)); } if (QDF_LOCK_STATS_LIST) qdf_lock_stats_cookie_destroy(stats); } #ifndef MEMORY_DEBUG #define qdf_mem_malloc_debug(x, y, z) qdf_mem_malloc(x) #endif /* qdf_lock_stats_create() - initialize the lock stats structure * */ static inline void qdf_lock_stats_create(struct lock_stats *stats, const char *func, int line) { qdf_mem_zero(stats, sizeof(*stats)); stats->initialization_fn = func; stats->line = line; if (QDF_LOCK_STATS_LIST) qdf_lock_stats_cookie_create(stats, func, line); } #endif #include #define WIFI_POWER_EVENT_DEFAULT_WAKELOCK_TIMEOUT 0 #define WIFI_POWER_EVENT_WAKELOCK_TAKEN 0 #define WIFI_POWER_EVENT_WAKELOCK_RELEASED 1 /** * qdf_semaphore_acquire_timeout() - Take the semaphore before timeout * @m: semaphore to take * @timeout: maximum time to try to take the semaphore * Return: int */ static inline int qdf_semaphore_acquire_timeout(struct semaphore *m, unsigned long timeout) { return __qdf_semaphore_acquire_timeout(m, timeout); } struct qdf_spinlock { __qdf_spinlock_t lock; struct lock_stats stats; }; /** * @brief Platform spinlock object */ typedef struct qdf_spinlock qdf_spinlock_t; /** * @brief Platform mutex object */ typedef __qdf_semaphore_t qdf_semaphore_t; typedef __qdf_mutex_t qdf_mutex_t; /* function Declaration */ QDF_STATUS qdf_mutex_create(qdf_mutex_t *m, const char *func, int line); #define qdf_mutex_create(m) qdf_mutex_create(m, __func__, __LINE__) QDF_STATUS qdf_mutex_acquire(qdf_mutex_t *m); QDF_STATUS qdf_mutex_release(qdf_mutex_t *m); QDF_STATUS qdf_mutex_destroy(qdf_mutex_t *lock); /** * qdf_spinlock_create - Initialize a spinlock * @lock: spinlock object pointer * Return: none */ static inline void qdf_spinlock_create(qdf_spinlock_t *lock, const char *func, int line) { __qdf_spinlock_create(&lock->lock); /* spinlock stats create relies on the spinlock working allread */ qdf_lock_stats_create(&lock->stats, func, line); } #define qdf_spinlock_create(x) qdf_spinlock_create(x, __func__, __LINE__) /** * qdf_spinlock_destroy - Delete a spinlock * @lock: spinlock object pointer * Return: none */ static inline void qdf_spinlock_destroy(qdf_spinlock_t *lock) { qdf_lock_stats_destroy(&lock->stats); __qdf_spinlock_destroy(&lock->lock); } /** * qdf_spin_is_locked() - check if the spinlock is locked * @lock: spinlock object * * Return: nonzero if lock is held. */ static inline int qdf_spin_is_locked(qdf_spinlock_t *lock) { return __qdf_spin_is_locked(&lock->lock); } /** * qdf_spin_trylock_bh() - spin trylock bottomhalf * @lock: spinlock object * * Return: nonzero if lock is acquired */ static inline int qdf_spin_trylock_bh(qdf_spinlock_t *lock, const char *func) { int trylock_return; BEFORE_TRYLOCK(lock); trylock_return = __qdf_spin_trylock_bh(&lock->lock); AFTER_TRYLOCK(lock, trylock_return, func); return trylock_return; } #define qdf_spin_trylock_bh(lock) qdf_spin_trylock_bh(lock, __func__) /** * qdf_spin_trylock() - spin trylock * @lock: spinlock object * Return: int */ static inline int qdf_spin_trylock(qdf_spinlock_t *lock, const char *func) { int result = 0; BEFORE_LOCK(lock, qdf_spin_is_locked(lock)); result = __qdf_spin_trylock(&lock->lock); AFTER_LOCK(lock, func); return result; } #define qdf_spin_trylock(lock) qdf_spin_trylock(lock, __func__) /** * qdf_spin_lock_bh() - locks the spinlock mutex in soft irq context * @lock: spinlock object pointer * Return: none */ static inline void qdf_spin_lock_bh(qdf_spinlock_t *lock, const char *func) { BEFORE_LOCK(lock, qdf_spin_is_locked(lock)); __qdf_spin_lock_bh(&lock->lock); AFTER_LOCK(lock, func); } #define qdf_spin_lock_bh(lock) qdf_spin_lock_bh(lock, __func__) /** * qdf_spin_unlock_bh() - unlocks the spinlock mutex in soft irq context * @lock: spinlock object pointer * Return: none */ static inline void qdf_spin_unlock_bh(qdf_spinlock_t *lock) { BEFORE_UNLOCK(lock, QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_BH); __qdf_spin_unlock_bh(&lock->lock); } /** * qdf_spinlock_irq_exec - Execute the input function with spinlock held * and interrupt disabled. * @hdl: OS handle * @lock: spinlock to be held for the critical region * @func: critical region function that to be executed * @context: context of the critical region function * Return: Boolean status returned by the critical region function */ static inline bool qdf_spinlock_irq_exec(qdf_handle_t hdl, qdf_spinlock_t *lock, qdf_irqlocked_func_t func, void *arg) { return __qdf_spinlock_irq_exec(hdl, &lock->lock, func, arg); } /** * qdf_spin_lock() - Acquire a Spinlock(SMP) & disable Preemption (Preemptive) * @lock: Lock object * * Return: none */ static inline void qdf_spin_lock(qdf_spinlock_t *lock, const char *func) { BEFORE_LOCK(lock, qdf_spin_is_locked(lock)); __qdf_spin_lock(&lock->lock); AFTER_LOCK(lock, func); } #define qdf_spin_lock(lock) qdf_spin_lock(lock, __func__) /** * qdf_spin_unlock() - Unlock the spinlock and enables the Preemption * @lock: Lock object * * Return: none */ static inline void qdf_spin_unlock(qdf_spinlock_t *lock) { BEFORE_UNLOCK(lock, QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK); __qdf_spin_unlock(&lock->lock); } /** * qdf_spin_lock_irq() - Acquire a Spinlock(SMP) & save the irq state * @lock: Lock object * @flags: flags * * Return: none */ static inline void qdf_spin_lock_irq(qdf_spinlock_t *lock, unsigned long flags, const char *func) { BEFORE_LOCK(lock, qdf_spin_is_locked(lock)); __qdf_spin_lock_irq(&lock->lock.spinlock, flags); AFTER_LOCK(lock, func); } #define qdf_spin_lock_irq(lock, flags) qdf_spin_lock_irq(lock, flags, __func__) /** * qdf_spin_lock_irqsave() - Acquire a Spinlock (SMP) & disable Preemption * (Preemptive) and disable IRQs * @lock: Lock object * * Return: none */ static inline void qdf_spin_lock_irqsave(qdf_spinlock_t *lock, const char *func) { BEFORE_LOCK(lock, qdf_spin_is_locked(lock)); __qdf_spin_lock_irqsave(&lock->lock); AFTER_LOCK(lock, func); } #define qdf_spin_lock_irqsave(lock) qdf_spin_lock_irqsave(lock, __func__) /** * qdf_spin_unlock_irqrestore() - Unlock the spinlock and enables the * Preemption and enable IRQ * @lock: Lock object * * Return: none */ static inline void qdf_spin_unlock_irqrestore(qdf_spinlock_t *lock) { BEFORE_UNLOCK(lock, QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_IRQ); __qdf_spin_unlock_irqrestore(&lock->lock); } /** * qdf_spin_unlock_irq() - Unlock a Spinlock(SMP) & save the restore state * @lock: Lock object * @flags: flags * * Return: none */ static inline void qdf_spin_unlock_irq(qdf_spinlock_t *lock, unsigned long flags) { BEFORE_UNLOCK(lock, QDF_MAX_HOLD_TIME_ALOWED_SPINLOCK_IRQ); __qdf_spin_unlock_irq(&lock->lock.spinlock, flags); } /** * qdf_semaphore_init() - initialize a semaphore * @m: Semaphore to initialize * Return: None */ static inline void qdf_semaphore_init(qdf_semaphore_t *m) { __qdf_semaphore_init(m); } /** * qdf_semaphore_acquire() - take the semaphore * @m: Semaphore to take * Return: int */ static inline int qdf_semaphore_acquire(qdf_semaphore_t *m) { return __qdf_semaphore_acquire(m); } /** * qdf_semaphore_release() - give the semaphore * @m: Semaphore to give * Return: None */ static inline void qdf_semaphore_release(qdf_semaphore_t *m) { __qdf_semaphore_release(m); } /** * qdf_semaphore_acquire_intr - Take the semaphore, interruptible version * @osdev: OS Device * @m: mutex to take * Return: int */ static inline int qdf_semaphore_acquire_intr(qdf_semaphore_t *m) { return __qdf_semaphore_acquire_intr(m); } #ifdef WLAN_WAKE_LOCK_DEBUG /** * qdf_wake_lock_check_for_leaks() - assert no wake lock leaks * * Return: None */ void qdf_wake_lock_check_for_leaks(void); /** * qdf_wake_lock_feature_init() - global init logic for wake lock * * Return: None */ void qdf_wake_lock_feature_init(void); /** * qdf_wake_lock_feature_deinit() - global de-init logic for wake lock * * Return: None */ void qdf_wake_lock_feature_deinit(void); #else static inline void qdf_wake_lock_check_for_leaks(void) { } static inline void qdf_wake_lock_feature_init(void) { } static inline void qdf_wake_lock_feature_deinit(void) { } #endif /* WLAN_WAKE_LOCK_DEBUG */ /** * __qdf_wake_lock_create() - initialize a wake lock * @lock: The wake lock to initialize * @name: Name of wake lock * @func: caller function * @line: caller line * Return: * QDF status success: if wake lock is initialized * QDF status failure: if wake lock was not initialized */ QDF_STATUS __qdf_wake_lock_create(qdf_wake_lock_t *lock, const char *name, const char *func, uint32_t line); /** * qdf_wake_lock_create() - initialized a wakeup source lock * @lock: the wakeup source lock to initialize * @name: the name of wakeup source lock * * Return: QDF_STATUS */ #define qdf_wake_lock_create(lock, name) \ __qdf_wake_lock_create(lock, name, __func__, __LINE__) QDF_STATUS qdf_wake_lock_acquire(qdf_wake_lock_t *lock, uint32_t reason); const char *qdf_wake_lock_name(qdf_wake_lock_t *lock); QDF_STATUS qdf_wake_lock_timeout_acquire(qdf_wake_lock_t *lock, uint32_t msec); QDF_STATUS qdf_wake_lock_release(qdf_wake_lock_t *lock, uint32_t reason); /** * __qdf_wake_lock_destroy() - destroy a wake lock * @lock: The wake lock to destroy * @func: caller function * @line: caller line * * Return: None */ void __qdf_wake_lock_destroy(qdf_wake_lock_t *lock, const char *func, uint32_t line); /** * qdf_wake_lock_destroy() - deinitialize a wakeup source lock * @lock: the wakeup source lock to de-initialize * * Return: None */ #define qdf_wake_lock_destroy(lock) \ __qdf_wake_lock_destroy(lock, __func__, __LINE__) void qdf_pm_system_wakeup(void); QDF_STATUS qdf_spinlock_acquire(qdf_spinlock_t *lock); QDF_STATUS qdf_spinlock_release(qdf_spinlock_t *lock); /** * enum qdf_rtpm_type - Get and Put calls types * @QDF_RTPM_GET: Increment usage count and when system is suspended * schedule resume process, return depends on pm state. * @QDF_RTPM_GET_FORCE: Increment usage count and when system is suspended * shedule resume process, returns success irrespective of * pm_state. * @QDF_RTPM_GET_SYNC: Increment usage count and when system is suspended, * wait till process is resumed. * @QDF_RTPM_GET_NORESUME: Only increments usage count. * @QDF_RTPM_PUT: Decrements usage count and puts system in idle state. * @QDF_RTPM_PUT_SYNC_SUSPEND: Decrements usage count and puts system in * suspended state. * @QDF_RTPM_PUT_NOIDLE: Decrements usage count. */ enum qdf_rtpm_call_type { QDF_RTPM_GET, QDF_RTPM_GET_FORCE, QDF_RTPM_GET_SYNC, QDF_RTPM_GET_NORESUME, QDF_RTPM_PUT, QDF_RTPM_PUT_SYNC_SUSPEND, QDF_RTPM_PUT_NOIDLE, }; /** * enum qdf_rtpm_client_id - modules registered with runtime pm module * @QDF_RTPM_ID_RESERVED: Reserved ID * @QDF_RTPM_ID_PM_QOS_NOTIFY: PM QOS context * @QDF_RTPM_ID_BUS_SUSPEND: APSS Bus suspend context * @QDF_RTPM_ID_MAX: Max id */ enum qdf_rtpm_client_id { QDF_RTPM_ID_RESERVED, QDF_RTPM_ID_PM_QOS_NOTIFY, QDF_RTPM_ID_WIPHY_SUSPEND, QDF_RTPM_ID_MAX }; #define qdf_runtime_lock_init(lock) __qdf_runtime_lock_init(lock, #lock) #ifdef FEATURE_RUNTIME_PM /** * qdf_rtpm_register() - QDF wrapper to register a module with runtime PM. * @id: ID of the module which needs to be registered * @hif_rpm_cbk: callback to be called when get was called in suspended state. * @prevent_multiple_get: not allow simultaneous get calls or put calls * * Return: success status if registered */ QDF_STATUS qdf_rtpm_register(uint32_t id, void (*hif_rpm_cbk)(void)); /** * qdf_rtpm_deregister() - QDF wrapper to deregister the module * @id: ID of the module which needs to be de-registered * * Return: success status if successfully de-registered */ QDF_STATUS qdf_rtpm_deregister(uint32_t id); /** * qdf_runtime_lock_init() - initialize runtime lock * @name: name of the runtime lock * * Initialize a runtime pm lock. This lock can be used * to prevent the runtime pm system from putting the bus * to sleep. * * Return: Success if lock initialized */ QDF_STATUS __qdf_runtime_lock_init(qdf_runtime_lock_t *lock, const char *name); /** * qdf_runtime_lock_deinit() - deinitialize runtime pm lock * @lock: the lock to deinitialize * * Ensures the lock is released. Frees the runtime lock. * * Return: void */ void qdf_runtime_lock_deinit(qdf_runtime_lock_t *lock); /** * qdf_rtpm_get() - Incremeant usage_count on the device to avoid suspend. * @type: get call types from hif_rpm_type * @id: ID of the module calling get() * * Return: success if a get has been issued, else error code. */ QDF_STATUS qdf_rtpm_get(uint8_t type, uint32_t id); /** * qdf_rtpm_put() - Decremeant usage_count on the device to avoid suspend. * @type: put call types from hif_rpm_type * @id: ID of the module calling put() * * Return: success if a put has been issued, else error code. */ QDF_STATUS qdf_rtpm_put(uint8_t type, uint32_t id); /** * qdf_runtime_pm_prevent_suspend() - Prevent Runtime suspend * @data: runtime PM lock * * This function will prevent runtime suspend, by incrementing * device's usage count. * * Return: status */ QDF_STATUS qdf_runtime_pm_prevent_suspend(qdf_runtime_lock_t *lock); /** * qdf_runtime_pm_prevent_suspend_sync() - Synchronized Prevent Runtime suspend * @data: runtime PM lock * * This function will prevent runtime suspend, by incrementing * device's usage count and waits till system is in resumed state. * * Return: status */ QDF_STATUS qdf_runtime_pm_prevent_suspend_sync(qdf_runtime_lock_t *lock); /** * qdf_runtime_pm_allow_suspend() - Allow Runtime suspend * @data: runtime PM lock * * This function will allow runtime suspend, by decrementing * device's usage count. * * Return: status */ QDF_STATUS qdf_runtime_pm_allow_suspend(qdf_runtime_lock_t *lock); /** * qdf_pm_runtime_sync_resume() - Invoke synchronous runtime resume. * * This function will invoke synchronous runtime resume. * * Return: Success if state is ON */ QDF_STATUS qdf_rtpm_sync_resume(void); #else static inline QDF_STATUS qdf_rtpm_register(uint32_t id, void (*hif_rpm_cbk)(void)) { return 0; } static inline QDF_STATUS qdf_rtpm_deregister(uint32_t id) { return QDF_STATUS_SUCCESS; } static inline QDF_STATUS __qdf_runtime_lock_init(qdf_runtime_lock_t *lock, const char *name) { return QDF_STATUS_SUCCESS; } static inline void qdf_runtime_lock_deinit(qdf_runtime_lock_t *lock) { } static inline QDF_STATUS qdf_rtpm_get(uint8_t type, uint32_t id) { return QDF_STATUS_SUCCESS; } static inline QDF_STATUS qdf_rtpm_put(uint8_t type, uint32_t id) { return QDF_STATUS_SUCCESS; } static inline QDF_STATUS qdf_runtime_pm_prevent_suspend(qdf_runtime_lock_t *lock) { return QDF_STATUS_SUCCESS; } static inline QDF_STATUS qdf_runtime_pm_prevent_suspend_sync(qdf_runtime_lock_t *lock) { return QDF_STATUS_SUCCESS; } static inline QDF_STATUS qdf_runtime_pm_allow_suspend(qdf_runtime_lock_t *lock) { return QDF_STATUS_SUCCESS; } static inline QDF_STATUS qdf_rtpm_sync_resume(void) { return QDF_STATUS_SUCCESS; } #endif /* FEATURE_RUNTIME_PM */ #endif /* _QDF_LOCK_H */