Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0+ |
| 2 | // |
Andrea Parri | 1a00b455 | 2018-05-14 16:33:56 -0700 | [diff] [blame] | 3 | // An earlier version of this file appeared in the companion webpage for |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 4 | // "Frightening small children and disconcerting grown-ups: Concurrency |
| 5 | // in the Linux kernel" by Alglave, Maranget, McKenney, Parri, and Stern, |
Andrea Parri | 1a00b455 | 2018-05-14 16:33:56 -0700 | [diff] [blame] | 6 | // which appeared in ASPLOS 2018. |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 7 | |
| 8 | // ONCE |
| 9 | READ_ONCE(X) __load{once}(X) |
| 10 | WRITE_ONCE(X,V) { __store{once}(X,V); } |
| 11 | |
| 12 | // Release Acquire and friends |
| 13 | smp_store_release(X,V) { __store{release}(*X,V); } |
| 14 | smp_load_acquire(X) __load{acquire}(*X) |
| 15 | rcu_assign_pointer(X,V) { __store{release}(X,V); } |
Alan Stern | bd5c0ba | 2018-03-07 09:27:40 -0800 | [diff] [blame] | 16 | rcu_dereference(X) __load{once}(X) |
Andrea Parri | bf8c6d9 | 2018-05-14 16:33:45 -0700 | [diff] [blame] | 17 | smp_store_mb(X,V) { __store{once}(X,V); __fence{mb}; } |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 18 | |
| 19 | // Fences |
Andrea Parri | d17013e | 2018-05-14 16:33:46 -0700 | [diff] [blame] | 20 | smp_mb() { __fence{mb}; } |
| 21 | smp_rmb() { __fence{rmb}; } |
| 22 | smp_wmb() { __fence{wmb}; } |
| 23 | smp_mb__before_atomic() { __fence{before-atomic}; } |
| 24 | smp_mb__after_atomic() { __fence{after-atomic}; } |
| 25 | smp_mb__after_spinlock() { __fence{after-spinlock}; } |
Andrea Parri | 5b735eb | 2018-12-03 15:04:49 -0800 | [diff] [blame] | 26 | smp_mb__after_unlock_lock() { __fence{after-unlock-lock}; } |
Alan Stern | 0031e38 | 2019-04-22 12:18:09 -0400 | [diff] [blame] | 27 | barrier() { __fence{barrier}; } |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 28 | |
| 29 | // Exchange |
| 30 | xchg(X,V) __xchg{mb}(X,V) |
| 31 | xchg_relaxed(X,V) __xchg{once}(X,V) |
| 32 | xchg_release(X,V) __xchg{release}(X,V) |
| 33 | xchg_acquire(X,V) __xchg{acquire}(X,V) |
| 34 | cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) |
| 35 | cmpxchg_relaxed(X,V,W) __cmpxchg{once}(X,V,W) |
| 36 | cmpxchg_acquire(X,V,W) __cmpxchg{acquire}(X,V,W) |
| 37 | cmpxchg_release(X,V,W) __cmpxchg{release}(X,V,W) |
| 38 | |
| 39 | // Spinlocks |
Andrea Parri | d17013e | 2018-05-14 16:33:46 -0700 | [diff] [blame] | 40 | spin_lock(X) { __lock(X); } |
| 41 | spin_unlock(X) { __unlock(X); } |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 42 | spin_trylock(X) __trylock(X) |
Luc Maranget | 15553dc | 2018-05-14 16:33:48 -0700 | [diff] [blame] | 43 | spin_is_locked(X) __islocked(X) |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 44 | |
| 45 | // RCU |
| 46 | rcu_read_lock() { __fence{rcu-lock}; } |
Andrea Parri | d17013e | 2018-05-14 16:33:46 -0700 | [diff] [blame] | 47 | rcu_read_unlock() { __fence{rcu-unlock}; } |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 48 | synchronize_rcu() { __fence{sync-rcu}; } |
| 49 | synchronize_rcu_expedited() { __fence{sync-rcu}; } |
| 50 | |
Alan Stern | a3f600d | 2018-11-15 11:20:37 -0500 | [diff] [blame] | 51 | // SRCU |
| 52 | srcu_read_lock(X) __srcu{srcu-lock}(X) |
Luc Maranget | 9393998 | 2018-12-27 16:27:12 +0100 | [diff] [blame] | 53 | srcu_read_unlock(X,Y) { __srcu{srcu-unlock}(X,Y); } |
Alan Stern | a3f600d | 2018-11-15 11:20:37 -0500 | [diff] [blame] | 54 | synchronize_srcu(X) { __srcu{sync-srcu}(X); } |
Paul E. McKenney | a5220e7 | 2019-03-19 13:25:03 -0700 | [diff] [blame] | 55 | synchronize_srcu_expedited(X) { __srcu{sync-srcu}(X); } |
Alan Stern | a3f600d | 2018-11-15 11:20:37 -0500 | [diff] [blame] | 56 | |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 57 | // Atomic |
| 58 | atomic_read(X) READ_ONCE(*X) |
Andrea Parri | d17013e | 2018-05-14 16:33:46 -0700 | [diff] [blame] | 59 | atomic_set(X,V) { WRITE_ONCE(*X,V); } |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 60 | atomic_read_acquire(X) smp_load_acquire(X) |
| 61 | atomic_set_release(X,V) { smp_store_release(X,V); } |
| 62 | |
Andrea Parri | d17013e | 2018-05-14 16:33:46 -0700 | [diff] [blame] | 63 | atomic_add(V,X) { __atomic_op(X,+,V); } |
| 64 | atomic_sub(V,X) { __atomic_op(X,-,V); } |
| 65 | atomic_inc(X) { __atomic_op(X,+,1); } |
| 66 | atomic_dec(X) { __atomic_op(X,-,1); } |
Paul E. McKenney | 1c27b64 | 2018-01-18 19:58:55 -0800 | [diff] [blame] | 67 | |
| 68 | atomic_add_return(V,X) __atomic_op_return{mb}(X,+,V) |
| 69 | atomic_add_return_relaxed(V,X) __atomic_op_return{once}(X,+,V) |
| 70 | atomic_add_return_acquire(V,X) __atomic_op_return{acquire}(X,+,V) |
| 71 | atomic_add_return_release(V,X) __atomic_op_return{release}(X,+,V) |
| 72 | atomic_fetch_add(V,X) __atomic_fetch_op{mb}(X,+,V) |
| 73 | atomic_fetch_add_relaxed(V,X) __atomic_fetch_op{once}(X,+,V) |
| 74 | atomic_fetch_add_acquire(V,X) __atomic_fetch_op{acquire}(X,+,V) |
| 75 | atomic_fetch_add_release(V,X) __atomic_fetch_op{release}(X,+,V) |
| 76 | |
| 77 | atomic_inc_return(X) __atomic_op_return{mb}(X,+,1) |
| 78 | atomic_inc_return_relaxed(X) __atomic_op_return{once}(X,+,1) |
| 79 | atomic_inc_return_acquire(X) __atomic_op_return{acquire}(X,+,1) |
| 80 | atomic_inc_return_release(X) __atomic_op_return{release}(X,+,1) |
| 81 | atomic_fetch_inc(X) __atomic_fetch_op{mb}(X,+,1) |
| 82 | atomic_fetch_inc_relaxed(X) __atomic_fetch_op{once}(X,+,1) |
| 83 | atomic_fetch_inc_acquire(X) __atomic_fetch_op{acquire}(X,+,1) |
| 84 | atomic_fetch_inc_release(X) __atomic_fetch_op{release}(X,+,1) |
| 85 | |
| 86 | atomic_sub_return(V,X) __atomic_op_return{mb}(X,-,V) |
| 87 | atomic_sub_return_relaxed(V,X) __atomic_op_return{once}(X,-,V) |
| 88 | atomic_sub_return_acquire(V,X) __atomic_op_return{acquire}(X,-,V) |
| 89 | atomic_sub_return_release(V,X) __atomic_op_return{release}(X,-,V) |
| 90 | atomic_fetch_sub(V,X) __atomic_fetch_op{mb}(X,-,V) |
| 91 | atomic_fetch_sub_relaxed(V,X) __atomic_fetch_op{once}(X,-,V) |
| 92 | atomic_fetch_sub_acquire(V,X) __atomic_fetch_op{acquire}(X,-,V) |
| 93 | atomic_fetch_sub_release(V,X) __atomic_fetch_op{release}(X,-,V) |
| 94 | |
| 95 | atomic_dec_return(X) __atomic_op_return{mb}(X,-,1) |
| 96 | atomic_dec_return_relaxed(X) __atomic_op_return{once}(X,-,1) |
| 97 | atomic_dec_return_acquire(X) __atomic_op_return{acquire}(X,-,1) |
| 98 | atomic_dec_return_release(X) __atomic_op_return{release}(X,-,1) |
| 99 | atomic_fetch_dec(X) __atomic_fetch_op{mb}(X,-,1) |
| 100 | atomic_fetch_dec_relaxed(X) __atomic_fetch_op{once}(X,-,1) |
| 101 | atomic_fetch_dec_acquire(X) __atomic_fetch_op{acquire}(X,-,1) |
| 102 | atomic_fetch_dec_release(X) __atomic_fetch_op{release}(X,-,1) |
| 103 | |
| 104 | atomic_xchg(X,V) __xchg{mb}(X,V) |
| 105 | atomic_xchg_relaxed(X,V) __xchg{once}(X,V) |
| 106 | atomic_xchg_release(X,V) __xchg{release}(X,V) |
| 107 | atomic_xchg_acquire(X,V) __xchg{acquire}(X,V) |
| 108 | atomic_cmpxchg(X,V,W) __cmpxchg{mb}(X,V,W) |
| 109 | atomic_cmpxchg_relaxed(X,V,W) __cmpxchg{once}(X,V,W) |
| 110 | atomic_cmpxchg_acquire(X,V,W) __cmpxchg{acquire}(X,V,W) |
| 111 | atomic_cmpxchg_release(X,V,W) __cmpxchg{release}(X,V,W) |
| 112 | |
| 113 | atomic_sub_and_test(V,X) __atomic_op_return{mb}(X,-,V) == 0 |
| 114 | atomic_dec_and_test(X) __atomic_op_return{mb}(X,-,1) == 0 |
| 115 | atomic_inc_and_test(X) __atomic_op_return{mb}(X,+,1) == 0 |
| 116 | atomic_add_negative(V,X) __atomic_op_return{mb}(X,+,V) < 0 |