Lines Matching refs:lock
36 extern void queued_read_lock_slowpath(struct qrwlock *lock);
37 extern void queued_write_lock_slowpath(struct qrwlock *lock);
44 static inline int queued_read_trylock(struct qrwlock *lock) in queued_read_trylock() argument
48 cnts = atomic_read(&lock->cnts); in queued_read_trylock()
50 cnts = (u32)atomic_add_return_acquire(_QR_BIAS, &lock->cnts); in queued_read_trylock()
53 atomic_sub(_QR_BIAS, &lock->cnts); in queued_read_trylock()
63 static inline int queued_write_trylock(struct qrwlock *lock) in queued_write_trylock() argument
67 cnts = atomic_read(&lock->cnts); in queued_write_trylock()
71 return likely(atomic_try_cmpxchg_acquire(&lock->cnts, &cnts, in queued_write_trylock()
78 static inline void queued_read_lock(struct qrwlock *lock) in queued_read_lock() argument
82 cnts = atomic_add_return_acquire(_QR_BIAS, &lock->cnts); in queued_read_lock()
87 queued_read_lock_slowpath(lock); in queued_read_lock()
94 static inline void queued_write_lock(struct qrwlock *lock) in queued_write_lock() argument
98 if (likely(atomic_try_cmpxchg_acquire(&lock->cnts, &cnts, _QW_LOCKED))) in queued_write_lock()
101 queued_write_lock_slowpath(lock); in queued_write_lock()
108 static inline void queued_read_unlock(struct qrwlock *lock) in queued_read_unlock() argument
113 (void)atomic_sub_return_release(_QR_BIAS, &lock->cnts); in queued_read_unlock()
120 static inline void queued_write_unlock(struct qrwlock *lock) in queued_write_unlock() argument
122 smp_store_release(&lock->wlocked, 0); in queued_write_unlock()
130 static inline int queued_rwlock_is_contended(struct qrwlock *lock) in queued_rwlock_is_contended() argument
132 return arch_spin_is_locked(&lock->wait_lock); in queued_rwlock_is_contended()