/openbmc/linux/kernel/locking/ |
H A D | qspinlock.c | 149 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending() 162 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked() 177 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 195 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending() 206 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked() 221 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 250 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire() 262 static __always_inline void set_locked(struct qspinlock *lock) in set_locked() 276 static __always_inline void __pv_kick_node(struct qspinlock *lock, in __pv_kick_node() 278 static __always_inline u32 __pv_wait_head_or_lock(struct qspinlock *lock, in __pv_wait_head_or_lock() [all …]
|
H A D | qspinlock_paravirt.h | 81 static inline bool pv_hybrid_queued_unfair_trylock(struct qspinlock *lock) in pv_hybrid_queued_unfair_trylock() 109 static __always_inline void set_pending(struct qspinlock *lock) in set_pending() 119 static __always_inline int trylock_clear_pending(struct qspinlock *lock) in trylock_clear_pending() 126 static __always_inline void set_pending(struct qspinlock *lock) in set_pending() 131 static __always_inline int trylock_clear_pending(struct qspinlock *lock) in trylock_clear_pending() 172 struct qspinlock *lock; member 239 static struct pv_node *pv_unhash(struct qspinlock *lock) in pv_unhash() 360 static void pv_kick_node(struct qspinlock *lock, struct mcs_spinlock *node) in pv_kick_node() 403 pv_wait_head_or_lock(struct qspinlock *lock, struct mcs_spinlock *node) in pv_wait_head_or_lock() 503 __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked) in __pv_queued_spin_unlock_slowpath() [all …]
|
H A D | spinlock_rt.c | 44 static __always_inline void __rt_spin_lock(spinlock_t *lock) in __rt_spin_lock() 52 void __sched rt_spin_lock(spinlock_t *lock) in rt_spin_lock() 60 void __sched rt_spin_lock_nested(spinlock_t *lock, int subclass) in rt_spin_lock_nested() 67 void __sched rt_spin_lock_nest_lock(spinlock_t *lock, in rt_spin_lock_nest_lock() 76 void __sched rt_spin_unlock(spinlock_t *lock) in rt_spin_unlock() 92 void __sched rt_spin_lock_unlock(spinlock_t *lock) in rt_spin_lock_unlock() 99 static __always_inline int __rt_spin_trylock(spinlock_t *lock) in __rt_spin_trylock() 114 int __sched rt_spin_trylock(spinlock_t *lock) in rt_spin_trylock() 120 int __sched rt_spin_trylock_bh(spinlock_t *lock) in rt_spin_trylock_bh() 133 void __rt_spin_lock_init(spinlock_t *lock, const char *name, in __rt_spin_lock_init()
|
H A D | rtmutex_common.h | 56 struct rt_mutex_base *lock; member 110 static inline int rt_mutex_has_waiters(struct rt_mutex_base *lock) in rt_mutex_has_waiters() 120 static inline bool rt_mutex_waiter_is_top_waiter(struct rt_mutex_base *lock, in rt_mutex_waiter_is_top_waiter() 128 static inline struct rt_mutex_waiter *rt_mutex_top_waiter(struct rt_mutex_base *lock) in rt_mutex_top_waiter() 157 static inline struct task_struct *rt_mutex_owner(struct rt_mutex_base *lock) in rt_mutex_owner() 179 static inline void __rt_mutex_base_init(struct rt_mutex_base *lock) in __rt_mutex_base_init() 187 static inline void debug_rt_mutex_unlock(struct rt_mutex_base *lock) in debug_rt_mutex_unlock() 193 static inline void debug_rt_mutex_proxy_unlock(struct rt_mutex_base *lock) in debug_rt_mutex_proxy_unlock() 228 static inline struct task_struct *rt_mutex_owner(struct rt_mutex_base *lock) in rt_mutex_owner()
|
H A D | mutex.h | 38 # define debug_mutex_lock_common(lock, waiter) do { } while (0) argument 39 # define debug_mutex_wake_waiter(lock, waiter) do { } while (0) argument 41 # define debug_mutex_add_waiter(lock, waiter, ti) do { } while (0) argument 42 # define debug_mutex_remove_waiter(lock, waiter, ti) do { } while (0) argument 43 # define debug_mutex_unlock(lock) do { } while (0) argument 44 # define debug_mutex_init(lock, name, key) do { } while (0) argument
|
H A D | ww_rt_mutex.c | 12 int ww_mutex_trylock(struct ww_mutex *lock, struct ww_acquire_ctx *ww_ctx) in ww_mutex_trylock() 38 __ww_rt_mutex_lock(struct ww_mutex *lock, struct ww_acquire_ctx *ww_ctx, in __ww_rt_mutex_lock() 79 ww_mutex_lock(struct ww_mutex *lock, struct ww_acquire_ctx *ctx) in ww_mutex_lock() 86 ww_mutex_lock_interruptible(struct ww_mutex *lock, struct ww_acquire_ctx *ctx) in ww_mutex_lock_interruptible() 92 void __sched ww_mutex_unlock(struct ww_mutex *lock) in ww_mutex_unlock()
|
/openbmc/linux/include/asm-generic/ |
H A D | spinlock.h | 33 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() 53 static __always_inline bool arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock() 63 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() 71 static __always_inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() 78 static __always_inline int arch_spin_is_contended(arch_spinlock_t *lock) in arch_spin_is_contended() 85 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked()
|
H A D | qrwlock.h | 44 static inline int queued_read_trylock(struct qrwlock *lock) in queued_read_trylock() 63 static inline int queued_write_trylock(struct qrwlock *lock) in queued_write_trylock() 78 static inline void queued_read_lock(struct qrwlock *lock) in queued_read_lock() 94 static inline void queued_write_lock(struct qrwlock *lock) in queued_write_lock() 108 static inline void queued_read_unlock(struct qrwlock *lock) in queued_read_unlock() 120 static inline void queued_write_unlock(struct qrwlock *lock) in queued_write_unlock() 130 static inline int queued_rwlock_is_contended(struct qrwlock *lock) in queued_rwlock_is_contended()
|
/openbmc/linux/fs/ocfs2/dlm/ |
H A D | dlmast.c | 47 static int dlm_should_cancel_bast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in dlm_should_cancel_bast() 74 void __dlm_queue_ast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in __dlm_queue_ast() 129 void dlm_queue_ast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in dlm_queue_ast() 140 void __dlm_queue_bast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in __dlm_queue_bast() 167 struct dlm_lock *lock) in dlm_update_lvb() 197 struct dlm_lock *lock) in dlm_do_local_ast() 215 struct dlm_lock *lock) in dlm_do_remote_ast() 239 struct dlm_lock *lock, int blocked_type) in dlm_do_local_bast() 263 struct dlm_lock *lock = NULL; in dlm_proxy_ast_handler() local 422 struct dlm_lock *lock, int msg_type, in dlm_send_proxy_ast_msg()
|
H A D | dlmlock.c | 74 struct dlm_lock *lock) in dlm_can_grant_new_lock() 103 struct dlm_lock *lock, int flags) in dlmlock_master() 182 struct dlm_lock *lock) in dlm_revert_pending_lock() 199 struct dlm_lock *lock, int flags) in dlmlock_remote() 285 struct dlm_lock *lock, int flags) in dlm_send_remote_lock_request() 324 void dlm_lock_get(struct dlm_lock *lock) in dlm_lock_get() 329 void dlm_lock_put(struct dlm_lock *lock) in dlm_lock_put() 336 struct dlm_lock *lock; in dlm_lock_release() local 356 void dlm_lock_attach_lockres(struct dlm_lock *lock, in dlm_lock_attach_lockres() 408 struct dlm_lock *lock; in dlm_new_lock() local [all …]
|
H A D | dlmunlock.c | 83 struct dlm_lock *lock, in dlmunlock_common() 266 struct dlm_lock *lock) in dlm_commit_pending_unlock() 274 struct dlm_lock *lock) in dlm_commit_pending_cancel() 283 struct dlm_lock *lock, in dlmunlock_master() 293 struct dlm_lock *lock, in dlmunlock_remote() 309 struct dlm_lock *lock, in dlm_send_remote_unlock_request() 395 struct dlm_lock *lock = NULL, *iter; in dlm_unlock_lock_handler() local 527 struct dlm_lock *lock, in dlm_get_cancel_actions() 559 struct dlm_lock *lock, in dlm_get_unlock_actions() 589 struct dlm_lock *lock = NULL; in dlmunlock() local
|
/openbmc/linux/arch/powerpc/include/asm/ |
H A D | qspinlock.h | 71 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked() 76 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked() 81 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended() 92 static __always_inline int __queued_spin_trylock_nosteal(struct qspinlock *lock) in __queued_spin_trylock_nosteal() 114 static __always_inline int __queued_spin_trylock_steal(struct qspinlock *lock) in __queued_spin_trylock_steal() 138 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock() 148 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock() 154 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
|
/openbmc/qemu/util/ |
H A D | qemu-coroutine-lock.c | 41 void coroutine_fn qemu_co_queue_wait_impl(CoQueue *queue, QemuLockable *lock, in qemu_co_queue_wait_impl() 74 bool qemu_co_enter_next_impl(CoQueue *queue, QemuLockable *lock) in qemu_co_enter_next_impl() 100 void qemu_co_enter_all_impl(CoQueue *queue, QemuLockable *lock) in qemu_co_enter_all_impl() 338 void qemu_co_rwlock_init(CoRwlock *lock) in qemu_co_rwlock_init() 346 static void coroutine_fn qemu_co_rwlock_maybe_wake_one(CoRwlock *lock) in qemu_co_rwlock_maybe_wake_one() 379 void coroutine_fn qemu_co_rwlock_rdlock(CoRwlock *lock) in qemu_co_rwlock_rdlock() 404 void coroutine_fn qemu_co_rwlock_unlock(CoRwlock *lock) in qemu_co_rwlock_unlock() 422 void coroutine_fn qemu_co_rwlock_downgrade(CoRwlock *lock) in qemu_co_rwlock_downgrade() 432 void coroutine_fn qemu_co_rwlock_wrlock(CoRwlock *lock) in qemu_co_rwlock_wrlock() 452 void coroutine_fn qemu_co_rwlock_upgrade(CoRwlock *lock) in qemu_co_rwlock_upgrade()
|
/openbmc/linux/arch/x86/include/asm/ |
H A D | qspinlock.h | 14 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire() 44 static inline void native_queued_spin_unlock(struct qspinlock *lock) in native_queued_spin_unlock() 49 static inline void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath() 54 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock() 88 static inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
|
/openbmc/linux/include/linux/ |
H A D | lockdep.h | 214 #define lockdep_set_class(lock, key) \ argument 232 #define lockdep_set_subclass(lock, sub) \ argument 238 #define lockdep_set_novalidate_class(lock) \ argument 388 # define lockdep_init_map(lock, name, key, sub) \ argument 479 #define LOCK_CONTENDED(_lock, try, lock) \ argument 505 #define LOCK_CONTENDED(_lock, try, lock) \ argument 587 # define might_lock(lock) \ argument 593 # define might_lock_read(lock) \ argument 599 # define might_lock_nested(lock, subclass) \ argument 661 # define might_lock(lock) do { } while (0) argument [all …]
|
H A D | rwlock_rt.h | 51 #define read_lock_irqsave(lock, flags) \ argument 58 #define read_trylock(lock) __cond_lock(lock, rt_read_trylock(lock)) argument 93 #define write_lock_nested(lock, subclass) rt_write_lock(((void)(subclass), (lock))) argument 107 #define write_lock_irqsave(lock, flags) \ argument 114 #define write_trylock(lock) __cond_lock(lock, rt_write_trylock(lock)) argument 116 #define write_trylock_irqsave(lock, flags) \ argument 148 #define rwlock_is_contended(lock) (((void)(lock), 0)) argument
|
H A D | seqlock.h | 291 #define SEQCNT_SPINLOCK_ZERO(name, lock) SEQCOUNT_LOCKNAME_ZERO(name, lock) argument 292 #define SEQCNT_RWLOCK_ZERO(name, lock) SEQCOUNT_LOCKNAME_ZERO(name, lock) argument 293 #define SEQCNT_MUTEX_ZERO(name, lock) SEQCOUNT_LOCKNAME_ZERO(name, lock) argument 803 spinlock_t lock; member 976 #define write_seqlock_irqsave(lock, flags) \ argument 1092 #define read_seqlock_excl_irqsave(lock, flags) \ argument 1138 static inline void read_seqbegin_or_lock(seqlock_t *lock, int *seq) in read_seqbegin_or_lock() 1153 static inline int need_seqretry(seqlock_t *lock, int seq) in need_seqretry() 1166 static inline void done_seqretry(seqlock_t *lock, int seq) in done_seqretry() 1193 read_seqbegin_or_lock_irqsave(seqlock_t *lock, int *seq) in read_seqbegin_or_lock_irqsave() [all …]
|
/openbmc/linux/tools/virtio/ringtest/ |
H A D | ptr_ring.c | 64 static void spin_lock_init(spinlock_t *lock) in spin_lock_init() 70 static void spin_lock(spinlock_t *lock) in spin_lock() 76 static void spin_unlock(spinlock_t *lock) in spin_unlock() 82 static void spin_lock_bh(spinlock_t *lock) in spin_lock_bh() 87 static void spin_unlock_bh(spinlock_t *lock) in spin_unlock_bh() 92 static void spin_lock_irq(spinlock_t *lock) in spin_lock_irq() 97 static void spin_unlock_irq(spinlock_t *lock) in spin_unlock_irq() 102 static void spin_lock_irqsave(spinlock_t *lock, unsigned long f) in spin_lock_irqsave() 107 static void spin_unlock_irqrestore(spinlock_t *lock, unsigned long f) in spin_unlock_irqrestore()
|
/openbmc/linux/lib/ |
H A D | atomic64.c | 28 raw_spinlock_t lock; member 48 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_read() local 61 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_set() local 134 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_dec_if_positive() local 149 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_cmpxchg() local 164 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_xchg() local 178 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_fetch_add_unless() local
|
H A D | dec_and_lock.c | 21 int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock) in _atomic_dec_and_lock() 37 int _atomic_dec_and_lock_irqsave(atomic_t *atomic, spinlock_t *lock, in _atomic_dec_and_lock_irqsave() 53 int _atomic_dec_and_raw_lock(atomic_t *atomic, raw_spinlock_t *lock) in _atomic_dec_and_raw_lock() 68 int _atomic_dec_and_raw_lock_irqsave(atomic_t *atomic, raw_spinlock_t *lock, in _atomic_dec_and_raw_lock_irqsave()
|
/openbmc/linux/arch/mips/include/asm/octeon/ |
H A D | cvmx-spinlock.h | 63 static inline void cvmx_spinlock_init(cvmx_spinlock_t *lock) in cvmx_spinlock_init() 74 static inline int cvmx_spinlock_locked(cvmx_spinlock_t *lock) in cvmx_spinlock_locked() 84 static inline void cvmx_spinlock_unlock(cvmx_spinlock_t *lock) in cvmx_spinlock_unlock() 103 static inline unsigned int cvmx_spinlock_trylock(cvmx_spinlock_t *lock) in cvmx_spinlock_trylock() 128 static inline void cvmx_spinlock_lock(cvmx_spinlock_t *lock) in cvmx_spinlock_lock()
|
/openbmc/linux/arch/arm64/kvm/hyp/include/nvhe/ |
H A D | spinlock.h | 44 static inline void hyp_spin_lock(hyp_spinlock_t *lock) in hyp_spin_lock() 82 static inline void hyp_spin_unlock(hyp_spinlock_t *lock) in hyp_spin_unlock() 101 static inline bool hyp_spin_is_locked(hyp_spinlock_t *lock) in hyp_spin_is_locked() 109 static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) in hyp_assert_lock_held() 122 static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) { } in hyp_assert_lock_held()
|
/openbmc/linux/drivers/md/persistent-data/ |
H A D | dm-block-manager.c | 45 spinlock_t lock; member 61 static unsigned int __find_holder(struct block_lock *lock, in __find_holder() 100 static int __check_holder(struct block_lock *lock) in __check_holder() 150 static void __wake_many(struct block_lock *lock) in __wake_many() 175 static void bl_init(struct block_lock *lock) in bl_init() 186 static int __available_for_read(struct block_lock *lock) in __available_for_read() 193 static int bl_down_read(struct block_lock *lock) in bl_down_read() 245 static void bl_up_read(struct block_lock *lock) in bl_up_read() 256 static int bl_down_write(struct block_lock *lock) in bl_down_write() 292 static void bl_up_write(struct block_lock *lock) in bl_up_write() [all …]
|
/openbmc/linux/arch/arm/include/asm/ |
H A D | spinlock.h | 56 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() 81 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock() 107 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() 114 static inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked() 119 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() 124 static inline int arch_spin_is_contended(arch_spinlock_t *lock) in arch_spin_is_contended()
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_lock.c | 63 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_lock_take() local 112 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_lock_transfer() local 127 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_legacy_lock_free() local 168 struct drm_lock *lock = data; in drm_legacy_lock() local 259 struct drm_lock *lock = data; in drm_legacy_unlock() local 312 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_legacy_idlelock_release() local
|