Home
last modified time | relevance | path

Searched defs:lock (Results 26 – 50 of 2667) sorted by relevance

12345678910>>...107

/openbmc/linux/kernel/locking/
H A Dqspinlock.c149 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending()
162 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked()
177 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail()
195 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending()
206 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked()
221 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail()
250 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire()
262 static __always_inline void set_locked(struct qspinlock *lock) in set_locked()
276 static __always_inline void __pv_kick_node(struct qspinlock *lock, in __pv_kick_node()
278 static __always_inline u32 __pv_wait_head_or_lock(struct qspinlock *lock, in __pv_wait_head_or_lock()
[all …]
H A Dqspinlock_paravirt.h81 static inline bool pv_hybrid_queued_unfair_trylock(struct qspinlock *lock) in pv_hybrid_queued_unfair_trylock()
109 static __always_inline void set_pending(struct qspinlock *lock) in set_pending()
119 static __always_inline int trylock_clear_pending(struct qspinlock *lock) in trylock_clear_pending()
126 static __always_inline void set_pending(struct qspinlock *lock) in set_pending()
131 static __always_inline int trylock_clear_pending(struct qspinlock *lock) in trylock_clear_pending()
172 struct qspinlock *lock; member
239 static struct pv_node *pv_unhash(struct qspinlock *lock) in pv_unhash()
360 static void pv_kick_node(struct qspinlock *lock, struct mcs_spinlock *node) in pv_kick_node()
403 pv_wait_head_or_lock(struct qspinlock *lock, struct mcs_spinlock *node) in pv_wait_head_or_lock()
503 __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked) in __pv_queued_spin_unlock_slowpath()
[all …]
H A Dspinlock_rt.c44 static __always_inline void __rt_spin_lock(spinlock_t *lock) in __rt_spin_lock()
52 void __sched rt_spin_lock(spinlock_t *lock) in rt_spin_lock()
60 void __sched rt_spin_lock_nested(spinlock_t *lock, int subclass) in rt_spin_lock_nested()
67 void __sched rt_spin_lock_nest_lock(spinlock_t *lock, in rt_spin_lock_nest_lock()
76 void __sched rt_spin_unlock(spinlock_t *lock) in rt_spin_unlock()
92 void __sched rt_spin_lock_unlock(spinlock_t *lock) in rt_spin_lock_unlock()
99 static __always_inline int __rt_spin_trylock(spinlock_t *lock) in __rt_spin_trylock()
114 int __sched rt_spin_trylock(spinlock_t *lock) in rt_spin_trylock()
120 int __sched rt_spin_trylock_bh(spinlock_t *lock) in rt_spin_trylock_bh()
133 void __rt_spin_lock_init(spinlock_t *lock, const char *name, in __rt_spin_lock_init()
H A Drtmutex_common.h56 struct rt_mutex_base *lock; member
110 static inline int rt_mutex_has_waiters(struct rt_mutex_base *lock) in rt_mutex_has_waiters()
120 static inline bool rt_mutex_waiter_is_top_waiter(struct rt_mutex_base *lock, in rt_mutex_waiter_is_top_waiter()
128 static inline struct rt_mutex_waiter *rt_mutex_top_waiter(struct rt_mutex_base *lock) in rt_mutex_top_waiter()
157 static inline struct task_struct *rt_mutex_owner(struct rt_mutex_base *lock) in rt_mutex_owner()
179 static inline void __rt_mutex_base_init(struct rt_mutex_base *lock) in __rt_mutex_base_init()
187 static inline void debug_rt_mutex_unlock(struct rt_mutex_base *lock) in debug_rt_mutex_unlock()
193 static inline void debug_rt_mutex_proxy_unlock(struct rt_mutex_base *lock) in debug_rt_mutex_proxy_unlock()
228 static inline struct task_struct *rt_mutex_owner(struct rt_mutex_base *lock) in rt_mutex_owner()
H A Dmutex.h38 # define debug_mutex_lock_common(lock, waiter) do { } while (0) argument
39 # define debug_mutex_wake_waiter(lock, waiter) do { } while (0) argument
41 # define debug_mutex_add_waiter(lock, waiter, ti) do { } while (0) argument
42 # define debug_mutex_remove_waiter(lock, waiter, ti) do { } while (0) argument
43 # define debug_mutex_unlock(lock) do { } while (0) argument
44 # define debug_mutex_init(lock, name, key) do { } while (0) argument
H A Dww_rt_mutex.c12 int ww_mutex_trylock(struct ww_mutex *lock, struct ww_acquire_ctx *ww_ctx) in ww_mutex_trylock()
38 __ww_rt_mutex_lock(struct ww_mutex *lock, struct ww_acquire_ctx *ww_ctx, in __ww_rt_mutex_lock()
79 ww_mutex_lock(struct ww_mutex *lock, struct ww_acquire_ctx *ctx) in ww_mutex_lock()
86 ww_mutex_lock_interruptible(struct ww_mutex *lock, struct ww_acquire_ctx *ctx) in ww_mutex_lock_interruptible()
92 void __sched ww_mutex_unlock(struct ww_mutex *lock) in ww_mutex_unlock()
/openbmc/linux/include/asm-generic/
H A Dspinlock.h33 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
53 static __always_inline bool arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
63 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
71 static __always_inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked()
78 static __always_inline int arch_spin_is_contended(arch_spinlock_t *lock) in arch_spin_is_contended()
85 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked()
H A Dqrwlock.h44 static inline int queued_read_trylock(struct qrwlock *lock) in queued_read_trylock()
63 static inline int queued_write_trylock(struct qrwlock *lock) in queued_write_trylock()
78 static inline void queued_read_lock(struct qrwlock *lock) in queued_read_lock()
94 static inline void queued_write_lock(struct qrwlock *lock) in queued_write_lock()
108 static inline void queued_read_unlock(struct qrwlock *lock) in queued_read_unlock()
120 static inline void queued_write_unlock(struct qrwlock *lock) in queued_write_unlock()
130 static inline int queued_rwlock_is_contended(struct qrwlock *lock) in queued_rwlock_is_contended()
/openbmc/linux/fs/ocfs2/dlm/
H A Ddlmast.c47 static int dlm_should_cancel_bast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in dlm_should_cancel_bast()
74 void __dlm_queue_ast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in __dlm_queue_ast()
129 void dlm_queue_ast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in dlm_queue_ast()
140 void __dlm_queue_bast(struct dlm_ctxt *dlm, struct dlm_lock *lock) in __dlm_queue_bast()
167 struct dlm_lock *lock) in dlm_update_lvb()
197 struct dlm_lock *lock) in dlm_do_local_ast()
215 struct dlm_lock *lock) in dlm_do_remote_ast()
239 struct dlm_lock *lock, int blocked_type) in dlm_do_local_bast()
263 struct dlm_lock *lock = NULL; in dlm_proxy_ast_handler() local
422 struct dlm_lock *lock, int msg_type, in dlm_send_proxy_ast_msg()
H A Ddlmlock.c74 struct dlm_lock *lock) in dlm_can_grant_new_lock()
103 struct dlm_lock *lock, int flags) in dlmlock_master()
182 struct dlm_lock *lock) in dlm_revert_pending_lock()
199 struct dlm_lock *lock, int flags) in dlmlock_remote()
285 struct dlm_lock *lock, int flags) in dlm_send_remote_lock_request()
324 void dlm_lock_get(struct dlm_lock *lock) in dlm_lock_get()
329 void dlm_lock_put(struct dlm_lock *lock) in dlm_lock_put()
336 struct dlm_lock *lock; in dlm_lock_release() local
356 void dlm_lock_attach_lockres(struct dlm_lock *lock, in dlm_lock_attach_lockres()
408 struct dlm_lock *lock; in dlm_new_lock() local
[all …]
H A Ddlmunlock.c83 struct dlm_lock *lock, in dlmunlock_common()
266 struct dlm_lock *lock) in dlm_commit_pending_unlock()
274 struct dlm_lock *lock) in dlm_commit_pending_cancel()
283 struct dlm_lock *lock, in dlmunlock_master()
293 struct dlm_lock *lock, in dlmunlock_remote()
309 struct dlm_lock *lock, in dlm_send_remote_unlock_request()
395 struct dlm_lock *lock = NULL, *iter; in dlm_unlock_lock_handler() local
527 struct dlm_lock *lock, in dlm_get_cancel_actions()
559 struct dlm_lock *lock, in dlm_get_unlock_actions()
589 struct dlm_lock *lock = NULL; in dlmunlock() local
/openbmc/linux/arch/powerpc/include/asm/
H A Dqspinlock.h71 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked()
76 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked()
81 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended()
92 static __always_inline int __queued_spin_trylock_nosteal(struct qspinlock *lock) in __queued_spin_trylock_nosteal()
114 static __always_inline int __queued_spin_trylock_steal(struct qspinlock *lock) in __queued_spin_trylock_steal()
138 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock()
148 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock()
154 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
/openbmc/qemu/util/
H A Dqemu-coroutine-lock.c41 void coroutine_fn qemu_co_queue_wait_impl(CoQueue *queue, QemuLockable *lock, in qemu_co_queue_wait_impl()
74 bool qemu_co_enter_next_impl(CoQueue *queue, QemuLockable *lock) in qemu_co_enter_next_impl()
100 void qemu_co_enter_all_impl(CoQueue *queue, QemuLockable *lock) in qemu_co_enter_all_impl()
338 void qemu_co_rwlock_init(CoRwlock *lock) in qemu_co_rwlock_init()
346 static void coroutine_fn qemu_co_rwlock_maybe_wake_one(CoRwlock *lock) in qemu_co_rwlock_maybe_wake_one()
379 void coroutine_fn qemu_co_rwlock_rdlock(CoRwlock *lock) in qemu_co_rwlock_rdlock()
404 void coroutine_fn qemu_co_rwlock_unlock(CoRwlock *lock) in qemu_co_rwlock_unlock()
422 void coroutine_fn qemu_co_rwlock_downgrade(CoRwlock *lock) in qemu_co_rwlock_downgrade()
432 void coroutine_fn qemu_co_rwlock_wrlock(CoRwlock *lock) in qemu_co_rwlock_wrlock()
452 void coroutine_fn qemu_co_rwlock_upgrade(CoRwlock *lock) in qemu_co_rwlock_upgrade()
/openbmc/linux/arch/x86/include/asm/
H A Dqspinlock.h14 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire()
44 static inline void native_queued_spin_unlock(struct qspinlock *lock) in native_queued_spin_unlock()
49 static inline void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath()
54 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
88 static inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
/openbmc/linux/include/linux/
H A Dlockdep.h214 #define lockdep_set_class(lock, key) \ argument
232 #define lockdep_set_subclass(lock, sub) \ argument
238 #define lockdep_set_novalidate_class(lock) \ argument
388 # define lockdep_init_map(lock, name, key, sub) \ argument
479 #define LOCK_CONTENDED(_lock, try, lock) \ argument
505 #define LOCK_CONTENDED(_lock, try, lock) \ argument
587 # define might_lock(lock) \ argument
593 # define might_lock_read(lock) \ argument
599 # define might_lock_nested(lock, subclass) \ argument
661 # define might_lock(lock) do { } while (0) argument
[all …]
H A Drwlock_rt.h51 #define read_lock_irqsave(lock, flags) \ argument
58 #define read_trylock(lock) __cond_lock(lock, rt_read_trylock(lock)) argument
93 #define write_lock_nested(lock, subclass) rt_write_lock(((void)(subclass), (lock))) argument
107 #define write_lock_irqsave(lock, flags) \ argument
114 #define write_trylock(lock) __cond_lock(lock, rt_write_trylock(lock)) argument
116 #define write_trylock_irqsave(lock, flags) \ argument
148 #define rwlock_is_contended(lock) (((void)(lock), 0)) argument
H A Dseqlock.h291 #define SEQCNT_SPINLOCK_ZERO(name, lock) SEQCOUNT_LOCKNAME_ZERO(name, lock) argument
292 #define SEQCNT_RWLOCK_ZERO(name, lock) SEQCOUNT_LOCKNAME_ZERO(name, lock) argument
293 #define SEQCNT_MUTEX_ZERO(name, lock) SEQCOUNT_LOCKNAME_ZERO(name, lock) argument
803 spinlock_t lock; member
976 #define write_seqlock_irqsave(lock, flags) \ argument
1092 #define read_seqlock_excl_irqsave(lock, flags) \ argument
1138 static inline void read_seqbegin_or_lock(seqlock_t *lock, int *seq) in read_seqbegin_or_lock()
1153 static inline int need_seqretry(seqlock_t *lock, int seq) in need_seqretry()
1166 static inline void done_seqretry(seqlock_t *lock, int seq) in done_seqretry()
1193 read_seqbegin_or_lock_irqsave(seqlock_t *lock, int *seq) in read_seqbegin_or_lock_irqsave()
[all …]
/openbmc/linux/tools/virtio/ringtest/
H A Dptr_ring.c64 static void spin_lock_init(spinlock_t *lock) in spin_lock_init()
70 static void spin_lock(spinlock_t *lock) in spin_lock()
76 static void spin_unlock(spinlock_t *lock) in spin_unlock()
82 static void spin_lock_bh(spinlock_t *lock) in spin_lock_bh()
87 static void spin_unlock_bh(spinlock_t *lock) in spin_unlock_bh()
92 static void spin_lock_irq(spinlock_t *lock) in spin_lock_irq()
97 static void spin_unlock_irq(spinlock_t *lock) in spin_unlock_irq()
102 static void spin_lock_irqsave(spinlock_t *lock, unsigned long f) in spin_lock_irqsave()
107 static void spin_unlock_irqrestore(spinlock_t *lock, unsigned long f) in spin_unlock_irqrestore()
/openbmc/linux/lib/
H A Datomic64.c28 raw_spinlock_t lock; member
48 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_read() local
61 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_set() local
134 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_dec_if_positive() local
149 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_cmpxchg() local
164 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_xchg() local
178 raw_spinlock_t *lock = lock_addr(v); in generic_atomic64_fetch_add_unless() local
H A Ddec_and_lock.c21 int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock) in _atomic_dec_and_lock()
37 int _atomic_dec_and_lock_irqsave(atomic_t *atomic, spinlock_t *lock, in _atomic_dec_and_lock_irqsave()
53 int _atomic_dec_and_raw_lock(atomic_t *atomic, raw_spinlock_t *lock) in _atomic_dec_and_raw_lock()
68 int _atomic_dec_and_raw_lock_irqsave(atomic_t *atomic, raw_spinlock_t *lock, in _atomic_dec_and_raw_lock_irqsave()
/openbmc/linux/arch/mips/include/asm/octeon/
H A Dcvmx-spinlock.h63 static inline void cvmx_spinlock_init(cvmx_spinlock_t *lock) in cvmx_spinlock_init()
74 static inline int cvmx_spinlock_locked(cvmx_spinlock_t *lock) in cvmx_spinlock_locked()
84 static inline void cvmx_spinlock_unlock(cvmx_spinlock_t *lock) in cvmx_spinlock_unlock()
103 static inline unsigned int cvmx_spinlock_trylock(cvmx_spinlock_t *lock) in cvmx_spinlock_trylock()
128 static inline void cvmx_spinlock_lock(cvmx_spinlock_t *lock) in cvmx_spinlock_lock()
/openbmc/linux/arch/arm64/kvm/hyp/include/nvhe/
H A Dspinlock.h44 static inline void hyp_spin_lock(hyp_spinlock_t *lock) in hyp_spin_lock()
82 static inline void hyp_spin_unlock(hyp_spinlock_t *lock) in hyp_spin_unlock()
101 static inline bool hyp_spin_is_locked(hyp_spinlock_t *lock) in hyp_spin_is_locked()
109 static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) in hyp_assert_lock_held()
122 static inline void hyp_assert_lock_held(hyp_spinlock_t *lock) { } in hyp_assert_lock_held()
/openbmc/linux/drivers/md/persistent-data/
H A Ddm-block-manager.c45 spinlock_t lock; member
61 static unsigned int __find_holder(struct block_lock *lock, in __find_holder()
100 static int __check_holder(struct block_lock *lock) in __check_holder()
150 static void __wake_many(struct block_lock *lock) in __wake_many()
175 static void bl_init(struct block_lock *lock) in bl_init()
186 static int __available_for_read(struct block_lock *lock) in __available_for_read()
193 static int bl_down_read(struct block_lock *lock) in bl_down_read()
245 static void bl_up_read(struct block_lock *lock) in bl_up_read()
256 static int bl_down_write(struct block_lock *lock) in bl_down_write()
292 static void bl_up_write(struct block_lock *lock) in bl_up_write()
[all …]
/openbmc/linux/arch/arm/include/asm/
H A Dspinlock.h56 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock()
81 static inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock()
107 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock()
114 static inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked()
119 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked()
124 static inline int arch_spin_is_contended(arch_spinlock_t *lock) in arch_spin_is_contended()
/openbmc/linux/drivers/gpu/drm/
H A Ddrm_lock.c63 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_lock_take() local
112 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_lock_transfer() local
127 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_legacy_lock_free() local
168 struct drm_lock *lock = data; in drm_legacy_lock() local
259 struct drm_lock *lock = data; in drm_legacy_unlock() local
312 volatile unsigned int *lock = &lock_data->hw_lock->lock; in drm_legacy_idlelock_release() local

12345678910>>...107