11c27b644SPaul E. McKenneyC MP+porevlocks 21c27b644SPaul E. McKenney 38f32543bSPaul E. McKenney(* 48f32543bSPaul E. McKenney * Result: Never 58f32543bSPaul E. McKenney * 68f32543bSPaul E. McKenney * This litmus test demonstrates how lock acquisitions and releases can 78f32543bSPaul E. McKenney * stand in for smp_load_acquire() and smp_store_release(), respectively. 88f32543bSPaul E. McKenney * In other words, when holding a given lock (or indeed after releasing a 98f32543bSPaul E. McKenney * given lock), a CPU is not only guaranteed to see the accesses that other 108f32543bSPaul E. McKenney * CPUs made while previously holding that lock, it is also guaranteed to 118f32543bSPaul E. McKenney * see all prior accesses by those other CPUs. 128f32543bSPaul E. McKenney *) 138f32543bSPaul E. McKenney 14*5c587f9bSAkira Yokosawa{} 151c27b644SPaul E. McKenney 16b6ff3084SPaul E. McKenneyP0(int *buf, int *flag, spinlock_t *mylock) // Consumer 171c27b644SPaul E. McKenney{ 181c27b644SPaul E. McKenney int r0; 191c27b644SPaul E. McKenney int r1; 201c27b644SPaul E. McKenney 21acc4bdc5SPaul E. McKenney r0 = READ_ONCE(*flag); 221c27b644SPaul E. McKenney spin_lock(mylock); 23acc4bdc5SPaul E. McKenney r1 = READ_ONCE(*buf); 241c27b644SPaul E. McKenney spin_unlock(mylock); 251c27b644SPaul E. McKenney} 261c27b644SPaul E. McKenney 27b6ff3084SPaul E. McKenneyP1(int *buf, int *flag, spinlock_t *mylock) // Producer 281c27b644SPaul E. McKenney{ 291c27b644SPaul E. McKenney spin_lock(mylock); 30acc4bdc5SPaul E. McKenney WRITE_ONCE(*buf, 1); 311c27b644SPaul E. McKenney spin_unlock(mylock); 32acc4bdc5SPaul E. McKenney WRITE_ONCE(*flag, 1); 331c27b644SPaul E. McKenney} 341c27b644SPaul E. McKenney 35b6ff3084SPaul E. McKenneyexists (0:r0=1 /\ 0:r1=0) (* Bad outcome. *) 36