-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
locking/spinlock: Provide RT variant header: <linux/spinlock_rt.h>
Provide the necessary wrappers around the actual rtmutex based spinlock implementation. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Signed-off-by: Ingo Molnar <mingo@kernel.org> Link: https://lore.kernel.org/r/20210815211303.712897671@linutronix.de
- Loading branch information
Thomas Gleixner
authored and
Ingo Molnar
committed
Aug 17, 2021
1 parent
051790e
commit 342a932
Showing
3 changed files
with
162 additions
and
1 deletion.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,149 @@ | ||
// SPDX-License-Identifier: GPL-2.0-only | ||
#ifndef __LINUX_SPINLOCK_RT_H | ||
#define __LINUX_SPINLOCK_RT_H | ||
|
||
#ifndef __LINUX_SPINLOCK_H | ||
#error Do not include directly. Use spinlock.h | ||
#endif | ||
|
||
#ifdef CONFIG_DEBUG_LOCK_ALLOC | ||
extern void __rt_spin_lock_init(spinlock_t *lock, const char *name, | ||
struct lock_class_key *key); | ||
#else | ||
static inline void __rt_spin_lock_init(spinlock_t *lock, const char *name, | ||
struct lock_class_key *key) | ||
{ | ||
} | ||
#endif | ||
|
||
#define spin_lock_init(slock) \ | ||
do { \ | ||
static struct lock_class_key __key; \ | ||
\ | ||
rt_mutex_base_init(&(slock)->lock); \ | ||
__rt_spin_lock_init(slock, #slock, &__key); \ | ||
} while (0) | ||
|
||
extern void rt_spin_lock(spinlock_t *lock); | ||
extern void rt_spin_lock_nested(spinlock_t *lock, int subclass); | ||
extern void rt_spin_lock_nest_lock(spinlock_t *lock, struct lockdep_map *nest_lock); | ||
extern void rt_spin_unlock(spinlock_t *lock); | ||
extern void rt_spin_lock_unlock(spinlock_t *lock); | ||
extern int rt_spin_trylock_bh(spinlock_t *lock); | ||
extern int rt_spin_trylock(spinlock_t *lock); | ||
|
||
static __always_inline void spin_lock(spinlock_t *lock) | ||
{ | ||
rt_spin_lock(lock); | ||
} | ||
|
||
#ifdef CONFIG_LOCKDEP | ||
# define __spin_lock_nested(lock, subclass) \ | ||
rt_spin_lock_nested(lock, subclass) | ||
|
||
# define __spin_lock_nest_lock(lock, nest_lock) \ | ||
do { \ | ||
typecheck(struct lockdep_map *, &(nest_lock)->dep_map); \ | ||
rt_spin_lock_nest_lock(lock, &(nest_lock)->dep_map); \ | ||
} while (0) | ||
# define __spin_lock_irqsave_nested(lock, flags, subclass) \ | ||
do { \ | ||
typecheck(unsigned long, flags); \ | ||
flags = 0; \ | ||
__spin_lock_nested(lock, subclass); \ | ||
} while (0) | ||
|
||
#else | ||
/* | ||
* Always evaluate the 'subclass' argument to avoid that the compiler | ||
* warns about set-but-not-used variables when building with | ||
* CONFIG_DEBUG_LOCK_ALLOC=n and with W=1. | ||
*/ | ||
# define __spin_lock_nested(lock, subclass) spin_lock(((void)(subclass), (lock))) | ||
# define __spin_lock_nest_lock(lock, subclass) spin_lock(((void)(subclass), (lock))) | ||
# define __spin_lock_irqsave_nested(lock, flags, subclass) \ | ||
spin_lock_irqsave(((void)(subclass), (lock)), flags) | ||
#endif | ||
|
||
#define spin_lock_nested(lock, subclass) \ | ||
__spin_lock_nested(lock, subclass) | ||
|
||
#define spin_lock_nest_lock(lock, nest_lock) \ | ||
__spin_lock_nest_lock(lock, nest_lock) | ||
|
||
#define spin_lock_irqsave_nested(lock, flags, subclass) \ | ||
__spin_lock_irqsave_nested(lock, flags, subclass) | ||
|
||
static __always_inline void spin_lock_bh(spinlock_t *lock) | ||
{ | ||
/* Investigate: Drop bh when blocking ? */ | ||
local_bh_disable(); | ||
rt_spin_lock(lock); | ||
} | ||
|
||
static __always_inline void spin_lock_irq(spinlock_t *lock) | ||
{ | ||
rt_spin_lock(lock); | ||
} | ||
|
||
#define spin_lock_irqsave(lock, flags) \ | ||
do { \ | ||
typecheck(unsigned long, flags); \ | ||
flags = 0; \ | ||
spin_lock(lock); \ | ||
} while (0) | ||
|
||
static __always_inline void spin_unlock(spinlock_t *lock) | ||
{ | ||
rt_spin_unlock(lock); | ||
} | ||
|
||
static __always_inline void spin_unlock_bh(spinlock_t *lock) | ||
{ | ||
rt_spin_unlock(lock); | ||
local_bh_enable(); | ||
} | ||
|
||
static __always_inline void spin_unlock_irq(spinlock_t *lock) | ||
{ | ||
rt_spin_unlock(lock); | ||
} | ||
|
||
static __always_inline void spin_unlock_irqrestore(spinlock_t *lock, | ||
unsigned long flags) | ||
{ | ||
rt_spin_unlock(lock); | ||
} | ||
|
||
#define spin_trylock(lock) \ | ||
__cond_lock(lock, rt_spin_trylock(lock)) | ||
|
||
#define spin_trylock_bh(lock) \ | ||
__cond_lock(lock, rt_spin_trylock_bh(lock)) | ||
|
||
#define spin_trylock_irq(lock) \ | ||
__cond_lock(lock, rt_spin_trylock(lock)) | ||
|
||
#define __spin_trylock_irqsave(lock, flags) \ | ||
({ \ | ||
int __locked; \ | ||
\ | ||
typecheck(unsigned long, flags); \ | ||
flags = 0; \ | ||
__locked = spin_trylock(lock); \ | ||
__locked; \ | ||
}) | ||
|
||
#define spin_trylock_irqsave(lock, flags) \ | ||
__cond_lock(lock, __spin_trylock_irqsave(lock, flags)) | ||
|
||
#define spin_is_contended(lock) (((void)(lock), 0)) | ||
|
||
static inline int spin_is_locked(spinlock_t *lock) | ||
{ | ||
return rt_mutex_base_is_locked(&lock->lock); | ||
} | ||
|
||
#define assert_spin_locked(lock) BUG_ON(!spin_is_locked(lock)) | ||
|
||
#endif |