123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280 |
- // SPDX-License-Identifier: GPL-2.0-only
- /*
- * PREEMPT_RT substitution for spin/rw_locks
- *
- * spinlocks and rwlocks on RT are based on rtmutexes, with a few twists to
- * resemble the non RT semantics:
- *
- * - Contrary to plain rtmutexes, spinlocks and rwlocks are state
- * preserving. The task state is saved before blocking on the underlying
- * rtmutex, and restored when the lock has been acquired. Regular wakeups
- * during that time are redirected to the saved state so no wake up is
- * missed.
- *
- * - Non RT spin/rwlocks disable preemption and eventually interrupts.
- * Disabling preemption has the side effect of disabling migration and
- * preventing RCU grace periods.
- *
- * The RT substitutions explicitly disable migration and take
- * rcu_read_lock() across the lock held section.
- */
- #include <linux/spinlock.h>
- #include <linux/export.h>
- #define RT_MUTEX_BUILD_SPINLOCKS
- #include "rtmutex.c"
- /*
- * __might_resched() skips the state check as rtlocks are state
- * preserving. Take RCU nesting into account as spin/read/write_lock() can
- * legitimately nest into an RCU read side critical section.
- */
- #define RTLOCK_RESCHED_OFFSETS \
- (rcu_preempt_depth() << MIGHT_RESCHED_RCU_SHIFT)
- #define rtlock_might_resched() \
- __might_resched(__FILE__, __LINE__, RTLOCK_RESCHED_OFFSETS)
- static __always_inline void rtlock_lock(struct rt_mutex_base *rtm)
- {
- if (unlikely(!rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
- rtlock_slowlock(rtm);
- }
- static __always_inline void __rt_spin_lock(spinlock_t *lock)
- {
- rtlock_might_resched();
- rtlock_lock(&lock->lock);
- rcu_read_lock();
- migrate_disable();
- }
- void __sched rt_spin_lock(spinlock_t *lock)
- {
- spin_acquire(&lock->dep_map, 0, 0, _RET_IP_);
- __rt_spin_lock(lock);
- }
- EXPORT_SYMBOL(rt_spin_lock);
- #ifdef CONFIG_DEBUG_LOCK_ALLOC
- void __sched rt_spin_lock_nested(spinlock_t *lock, int subclass)
- {
- spin_acquire(&lock->dep_map, subclass, 0, _RET_IP_);
- __rt_spin_lock(lock);
- }
- EXPORT_SYMBOL(rt_spin_lock_nested);
- void __sched rt_spin_lock_nest_lock(spinlock_t *lock,
- struct lockdep_map *nest_lock)
- {
- spin_acquire_nest(&lock->dep_map, 0, 0, nest_lock, _RET_IP_);
- __rt_spin_lock(lock);
- }
- EXPORT_SYMBOL(rt_spin_lock_nest_lock);
- #endif
- void __sched rt_spin_unlock(spinlock_t *lock)
- {
- spin_release(&lock->dep_map, _RET_IP_);
- migrate_enable();
- rcu_read_unlock();
- if (unlikely(!rt_mutex_cmpxchg_release(&lock->lock, current, NULL)))
- rt_mutex_slowunlock(&lock->lock);
- }
- EXPORT_SYMBOL(rt_spin_unlock);
- /*
- * Wait for the lock to get unlocked: instead of polling for an unlock
- * (like raw spinlocks do), lock and unlock, to force the kernel to
- * schedule if there's contention:
- */
- void __sched rt_spin_lock_unlock(spinlock_t *lock)
- {
- spin_lock(lock);
- spin_unlock(lock);
- }
- EXPORT_SYMBOL(rt_spin_lock_unlock);
- static __always_inline int __rt_spin_trylock(spinlock_t *lock)
- {
- int ret = 1;
- if (unlikely(!rt_mutex_cmpxchg_acquire(&lock->lock, NULL, current)))
- ret = rt_mutex_slowtrylock(&lock->lock);
- if (ret) {
- spin_acquire(&lock->dep_map, 0, 1, _RET_IP_);
- rcu_read_lock();
- migrate_disable();
- }
- return ret;
- }
- int __sched rt_spin_trylock(spinlock_t *lock)
- {
- return __rt_spin_trylock(lock);
- }
- EXPORT_SYMBOL(rt_spin_trylock);
- int __sched rt_spin_trylock_bh(spinlock_t *lock)
- {
- int ret;
- local_bh_disable();
- ret = __rt_spin_trylock(lock);
- if (!ret)
- local_bh_enable();
- return ret;
- }
- EXPORT_SYMBOL(rt_spin_trylock_bh);
- #ifdef CONFIG_DEBUG_LOCK_ALLOC
- void __rt_spin_lock_init(spinlock_t *lock, const char *name,
- struct lock_class_key *key, bool percpu)
- {
- u8 type = percpu ? LD_LOCK_PERCPU : LD_LOCK_NORMAL;
- debug_check_no_locks_freed((void *)lock, sizeof(*lock));
- lockdep_init_map_type(&lock->dep_map, name, key, 0, LD_WAIT_CONFIG,
- LD_WAIT_INV, type);
- }
- EXPORT_SYMBOL(__rt_spin_lock_init);
- #endif
- /*
- * RT-specific reader/writer locks
- */
- #define rwbase_set_and_save_current_state(state) \
- current_save_and_set_rtlock_wait_state()
- #define rwbase_restore_current_state() \
- current_restore_rtlock_saved_state()
- static __always_inline int
- rwbase_rtmutex_lock_state(struct rt_mutex_base *rtm, unsigned int state)
- {
- if (unlikely(!rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
- rtlock_slowlock(rtm);
- return 0;
- }
- static __always_inline int
- rwbase_rtmutex_slowlock_locked(struct rt_mutex_base *rtm, unsigned int state)
- {
- rtlock_slowlock_locked(rtm);
- return 0;
- }
- static __always_inline void rwbase_rtmutex_unlock(struct rt_mutex_base *rtm)
- {
- if (likely(rt_mutex_cmpxchg_acquire(rtm, current, NULL)))
- return;
- rt_mutex_slowunlock(rtm);
- }
- static __always_inline int rwbase_rtmutex_trylock(struct rt_mutex_base *rtm)
- {
- if (likely(rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
- return 1;
- return rt_mutex_slowtrylock(rtm);
- }
- #define rwbase_signal_pending_state(state, current) (0)
- #define rwbase_schedule() \
- schedule_rtlock()
- #include "rwbase_rt.c"
- /*
- * The common functions which get wrapped into the rwlock API.
- */
- int __sched rt_read_trylock(rwlock_t *rwlock)
- {
- int ret;
- ret = rwbase_read_trylock(&rwlock->rwbase);
- if (ret) {
- rwlock_acquire_read(&rwlock->dep_map, 0, 1, _RET_IP_);
- rcu_read_lock();
- migrate_disable();
- }
- return ret;
- }
- EXPORT_SYMBOL(rt_read_trylock);
- int __sched rt_write_trylock(rwlock_t *rwlock)
- {
- int ret;
- ret = rwbase_write_trylock(&rwlock->rwbase);
- if (ret) {
- rwlock_acquire(&rwlock->dep_map, 0, 1, _RET_IP_);
- rcu_read_lock();
- migrate_disable();
- }
- return ret;
- }
- EXPORT_SYMBOL(rt_write_trylock);
- void __sched rt_read_lock(rwlock_t *rwlock)
- {
- rtlock_might_resched();
- rwlock_acquire_read(&rwlock->dep_map, 0, 0, _RET_IP_);
- rwbase_read_lock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
- rcu_read_lock();
- migrate_disable();
- }
- EXPORT_SYMBOL(rt_read_lock);
- void __sched rt_write_lock(rwlock_t *rwlock)
- {
- rtlock_might_resched();
- rwlock_acquire(&rwlock->dep_map, 0, 0, _RET_IP_);
- rwbase_write_lock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
- rcu_read_lock();
- migrate_disable();
- }
- EXPORT_SYMBOL(rt_write_lock);
- #ifdef CONFIG_DEBUG_LOCK_ALLOC
- void __sched rt_write_lock_nested(rwlock_t *rwlock, int subclass)
- {
- rtlock_might_resched();
- rwlock_acquire(&rwlock->dep_map, subclass, 0, _RET_IP_);
- rwbase_write_lock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
- rcu_read_lock();
- migrate_disable();
- }
- EXPORT_SYMBOL(rt_write_lock_nested);
- #endif
- void __sched rt_read_unlock(rwlock_t *rwlock)
- {
- rwlock_release(&rwlock->dep_map, _RET_IP_);
- migrate_enable();
- rcu_read_unlock();
- rwbase_read_unlock(&rwlock->rwbase, TASK_RTLOCK_WAIT);
- }
- EXPORT_SYMBOL(rt_read_unlock);
- void __sched rt_write_unlock(rwlock_t *rwlock)
- {
- rwlock_release(&rwlock->dep_map, _RET_IP_);
- rcu_read_unlock();
- migrate_enable();
- rwbase_write_unlock(&rwlock->rwbase);
- }
- EXPORT_SYMBOL(rt_write_unlock);
- #ifdef CONFIG_DEBUG_LOCK_ALLOC
- void __rt_rwlock_init(rwlock_t *rwlock, const char *name,
- struct lock_class_key *key)
- {
- debug_check_no_locks_freed((void *)rwlock, sizeof(*rwlock));
- lockdep_init_map_wait(&rwlock->dep_map, name, key, 0, LD_WAIT_CONFIG);
- }
- EXPORT_SYMBOL(__rt_rwlock_init);
- #endif
|