1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef _ASM_POWERPC_QSPINLOCK_H 3 #define _ASM_POWERPC_QSPINLOCK_H 4 5 #include <linux/compiler.h> 6 #include <asm/qspinlock_types.h> 7 #include <asm/paravirt.h> 8 9 #ifdef CONFIG_PPC64 10 /* 11 * Use the EH=1 hint for accesses that result in the lock being acquired. 12 * The hardware is supposed to optimise this pattern by holding the lock 13 * cacheline longer, and releasing when a store to the same memory (the 14 * unlock) is performed. 15 */ 16 #define _Q_SPIN_EH_HINT 1 17 #else 18 #define _Q_SPIN_EH_HINT 0 19 #endif 20 21 /* 22 * The trylock itself may steal. This makes trylocks slightly stronger, and 23 * makes locks slightly more efficient when stealing. 24 * 25 * This is compile-time, so if true then there may always be stealers, so the 26 * nosteal paths become unused. 27 */ 28 #define _Q_SPIN_TRY_LOCK_STEAL 1 29 30 /* 31 * Put a speculation barrier after testing the lock/node and finding it 32 * busy. Try to prevent pointless speculation in slow paths. 33 * 34 * Slows down the lockstorm microbenchmark with no stealing, where locking 35 * is purely FIFO through the queue. May have more benefit in real workload 36 * where speculating into the wrong place could have a greater cost. 37 */ 38 #define _Q_SPIN_SPEC_BARRIER 0 39 40 #ifdef CONFIG_PPC64 41 /* 42 * Execute a miso instruction after passing the MCS lock ownership to the 43 * queue head. Miso is intended to make stores visible to other CPUs sooner. 44 * 45 * This seems to make the lockstorm microbenchmark nospin test go slightly 46 * faster on POWER10, but disable for now. 47 */ 48 #define _Q_SPIN_MISO 0 49 #else 50 #define _Q_SPIN_MISO 0 51 #endif 52 53 #ifdef CONFIG_PPC64 54 /* 55 * This executes miso after an unlock of the lock word, having ownership 56 * pass to the next CPU sooner. This will slow the uncontended path to some 57 * degree. Not evidence it helps yet. 58 */ 59 #define _Q_SPIN_MISO_UNLOCK 0 60 #else 61 #define _Q_SPIN_MISO_UNLOCK 0 62 #endif 63 64 /* 65 * Seems to slow down lockstorm microbenchmark, suspect queue node just 66 * has to become shared again right afterwards when its waiter spins on 67 * the lock field. 68 */ 69 #define _Q_SPIN_PREFETCH_NEXT 0 70 71 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) 72 { 73 return READ_ONCE(lock->val); 74 } 75 76 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) 77 { 78 return !lock.val; 79 } 80 81 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) 82 { 83 return !!(READ_ONCE(lock->val) & _Q_TAIL_CPU_MASK); 84 } 85 86 static __always_inline u32 queued_spin_encode_locked_val(void) 87 { 88 /* XXX: make this use lock value in paca like simple spinlocks? */ 89 return _Q_LOCKED_VAL | (smp_processor_id() << _Q_OWNER_CPU_OFFSET); 90 } 91 92 static __always_inline int __queued_spin_trylock_nosteal(struct qspinlock *lock) 93 { 94 u32 new = queued_spin_encode_locked_val(); 95 u32 prev; 96 97 /* Trylock succeeds only when unlocked and no queued nodes */ 98 asm volatile( 99 "1: lwarx %0,0,%1,%3 # __queued_spin_trylock_nosteal \n" 100 " cmpwi 0,%0,0 \n" 101 " bne- 2f \n" 102 " stwcx. %2,0,%1 \n" 103 " bne- 1b \n" 104 "\t" PPC_ACQUIRE_BARRIER " \n" 105 "2: \n" 106 : "=&r" (prev) 107 : "r" (&lock->val), "r" (new), 108 "i" (_Q_SPIN_EH_HINT) 109 : "cr0", "memory"); 110 111 return likely(prev == 0); 112 } 113 114 static __always_inline int __queued_spin_trylock_steal(struct qspinlock *lock) 115 { 116 u32 new = queued_spin_encode_locked_val(); 117 u32 prev, tmp; 118 119 /* Trylock may get ahead of queued nodes if it finds unlocked */ 120 asm volatile( 121 "1: lwarx %0,0,%2,%5 # __queued_spin_trylock_steal \n" 122 " andc. %1,%0,%4 \n" 123 " bne- 2f \n" 124 " and %1,%0,%4 \n" 125 " or %1,%1,%3 \n" 126 " stwcx. %1,0,%2 \n" 127 " bne- 1b \n" 128 "\t" PPC_ACQUIRE_BARRIER " \n" 129 "2: \n" 130 : "=&r" (prev), "=&r" (tmp) 131 : "r" (&lock->val), "r" (new), "r" (_Q_TAIL_CPU_MASK), 132 "i" (_Q_SPIN_EH_HINT) 133 : "cr0", "memory"); 134 135 return likely(!(prev & ~_Q_TAIL_CPU_MASK)); 136 } 137 138 static __always_inline int queued_spin_trylock(struct qspinlock *lock) 139 { 140 if (!_Q_SPIN_TRY_LOCK_STEAL) 141 return __queued_spin_trylock_nosteal(lock); 142 else 143 return __queued_spin_trylock_steal(lock); 144 } 145 146 void queued_spin_lock_slowpath(struct qspinlock *lock); 147 148 static __always_inline void queued_spin_lock(struct qspinlock *lock) 149 { 150 if (!queued_spin_trylock(lock)) 151 queued_spin_lock_slowpath(lock); 152 } 153 154 static inline void queued_spin_unlock(struct qspinlock *lock) 155 { 156 smp_store_release(&lock->locked, 0); 157 if (_Q_SPIN_MISO_UNLOCK) 158 asm volatile("miso" ::: "memory"); 159 } 160 161 #define arch_spin_is_locked(l) queued_spin_is_locked(l) 162 #define arch_spin_is_contended(l) queued_spin_is_contended(l) 163 #define arch_spin_value_unlocked(l) queued_spin_value_unlocked(l) 164 #define arch_spin_lock(l) queued_spin_lock(l) 165 #define arch_spin_trylock(l) queued_spin_trylock(l) 166 #define arch_spin_unlock(l) queued_spin_unlock(l) 167 168 #ifdef CONFIG_PARAVIRT_SPINLOCKS 169 void pv_spinlocks_init(void); 170 #else 171 static inline void pv_spinlocks_init(void) { } 172 #endif 173 174 #endif /* _ASM_POWERPC_QSPINLOCK_H */ 175