David Sterba | c1d7c51 | 2018-04-03 19:23:33 +0200 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0 |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 2 | /* |
| 3 | * Copyright (C) 2008 Oracle. All rights reserved. |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 4 | */ |
David Sterba | c1d7c51 | 2018-04-03 19:23:33 +0200 | [diff] [blame] | 5 | |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 6 | #include <linux/sched.h> |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 7 | #include <linux/pagemap.h> |
| 8 | #include <linux/spinlock.h> |
| 9 | #include <linux/page-flags.h> |
Chris Mason | 4881ee5 | 2008-07-24 09:51:08 -0400 | [diff] [blame] | 10 | #include <asm/bug.h> |
David Sterba | 602cbe9 | 2019-08-21 18:48:25 +0200 | [diff] [blame] | 11 | #include "misc.h" |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 12 | #include "ctree.h" |
| 13 | #include "extent_io.h" |
| 14 | #include "locking.h" |
| 15 | |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 16 | /* |
| 17 | * Extent buffer locking |
| 18 | * ===================== |
| 19 | * |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 20 | * We use a rw_semaphore for tree locking, and the semantics are exactly the |
| 21 | * same: |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 22 | * |
| 23 | * - reader/writer exclusion |
| 24 | * - writer/writer exclusion |
| 25 | * - reader/reader sharing |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 26 | * - try-lock semantics for readers and writers |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 27 | * |
Josef Bacik | 4048dae | 2020-11-06 16:27:32 -0500 | [diff] [blame] | 28 | * The rwsem implementation does opportunistic spinning which reduces number of |
| 29 | * times the locking task needs to sleep. |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 30 | */ |
| 31 | |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 32 | /* |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 33 | * __btrfs_tree_read_lock - lock extent buffer for read |
| 34 | * @eb: the eb to be locked |
| 35 | * @nest: the nesting level to be used for lockdep |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 36 | * |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 37 | * This takes the read lock on the extent buffer, using the specified nesting |
| 38 | * level for lockdep purposes. |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 39 | */ |
Josef Bacik | 0ecae6f | 2020-11-06 16:27:35 -0500 | [diff] [blame] | 40 | void __btrfs_tree_read_lock(struct extent_buffer *eb, enum btrfs_lock_nesting nest) |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 41 | { |
Qu Wenruo | 34e73cc | 2019-04-15 21:15:24 +0800 | [diff] [blame] | 42 | u64 start_ns = 0; |
| 43 | |
| 44 | if (trace_btrfs_tree_read_lock_enabled()) |
| 45 | start_ns = ktime_get_ns(); |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 46 | |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 47 | down_read_nested(&eb->lock, nest); |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 48 | eb->lock_owner = current->pid; |
Qu Wenruo | 34e73cc | 2019-04-15 21:15:24 +0800 | [diff] [blame] | 49 | trace_btrfs_tree_read_lock(eb, start_ns); |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 50 | } |
| 51 | |
Josef Bacik | 5189941 | 2020-08-20 11:46:01 -0400 | [diff] [blame] | 52 | void btrfs_tree_read_lock(struct extent_buffer *eb) |
| 53 | { |
Josef Bacik | 0ecae6f | 2020-11-06 16:27:35 -0500 | [diff] [blame] | 54 | __btrfs_tree_read_lock(eb, BTRFS_NESTING_NORMAL); |
Josef Bacik | 5189941 | 2020-08-20 11:46:01 -0400 | [diff] [blame] | 55 | } |
| 56 | |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 57 | /* |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 58 | * Try-lock for read. |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 59 | * |
David Sterba | 1a9fd41 | 2021-05-21 17:42:23 +0200 | [diff] [blame] | 60 | * Return 1 if the rwlock has been taken, 0 otherwise |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 61 | */ |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 62 | int btrfs_try_tree_read_lock(struct extent_buffer *eb) |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 63 | { |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 64 | if (down_read_trylock(&eb->lock)) { |
| 65 | eb->lock_owner = current->pid; |
| 66 | trace_btrfs_try_tree_read_lock(eb); |
| 67 | return 1; |
Chris Mason | f9efa9c | 2008-06-25 16:14:04 -0400 | [diff] [blame] | 68 | } |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 69 | return 0; |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 70 | } |
| 71 | |
| 72 | /* |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 73 | * Try-lock for write. |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 74 | * |
David Sterba | 1a9fd41 | 2021-05-21 17:42:23 +0200 | [diff] [blame] | 75 | * Return 1 if the rwlock has been taken, 0 otherwise |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 76 | */ |
| 77 | int btrfs_try_tree_write_lock(struct extent_buffer *eb) |
| 78 | { |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 79 | if (down_write_trylock(&eb->lock)) { |
| 80 | eb->lock_owner = current->pid; |
| 81 | trace_btrfs_try_tree_write_lock(eb); |
| 82 | return 1; |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 83 | } |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 84 | return 0; |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 85 | } |
| 86 | |
| 87 | /* |
Josef Bacik | 4048dae | 2020-11-06 16:27:32 -0500 | [diff] [blame] | 88 | * Release read lock. |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 89 | */ |
| 90 | void btrfs_tree_read_unlock(struct extent_buffer *eb) |
| 91 | { |
Qu Wenruo | 31aab40 | 2019-04-15 21:15:25 +0800 | [diff] [blame] | 92 | trace_btrfs_tree_read_unlock(eb); |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 93 | eb->lock_owner = 0; |
| 94 | up_read(&eb->lock); |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 95 | } |
| 96 | |
| 97 | /* |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 98 | * __btrfs_tree_lock - lock eb for write |
| 99 | * @eb: the eb to lock |
| 100 | * @nest: the nesting to use for the lock |
David Sterba | d4e253b | 2019-10-16 18:29:10 +0200 | [diff] [blame] | 101 | * |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 102 | * Returns with the eb->lock write locked. |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 103 | */ |
Josef Bacik | fd7ba1c | 2020-08-20 11:46:02 -0400 | [diff] [blame] | 104 | void __btrfs_tree_lock(struct extent_buffer *eb, enum btrfs_lock_nesting nest) |
Jules Irenge | 78d933c | 2020-03-31 21:46:42 +0100 | [diff] [blame] | 105 | __acquires(&eb->lock) |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 106 | { |
Qu Wenruo | 34e73cc | 2019-04-15 21:15:24 +0800 | [diff] [blame] | 107 | u64 start_ns = 0; |
| 108 | |
| 109 | if (trace_btrfs_tree_lock_enabled()) |
| 110 | start_ns = ktime_get_ns(); |
| 111 | |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 112 | down_write_nested(&eb->lock, nest); |
Arne Jansen | 5b25f70 | 2011-09-13 10:55:48 +0200 | [diff] [blame] | 113 | eb->lock_owner = current->pid; |
Qu Wenruo | 34e73cc | 2019-04-15 21:15:24 +0800 | [diff] [blame] | 114 | trace_btrfs_tree_lock(eb, start_ns); |
Chris Mason | b4ce94d | 2009-02-04 09:25:08 -0500 | [diff] [blame] | 115 | } |
| 116 | |
Josef Bacik | fd7ba1c | 2020-08-20 11:46:02 -0400 | [diff] [blame] | 117 | void btrfs_tree_lock(struct extent_buffer *eb) |
| 118 | { |
| 119 | __btrfs_tree_lock(eb, BTRFS_NESTING_NORMAL); |
| 120 | } |
| 121 | |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 122 | /* |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 123 | * Release the write lock. |
Chris Mason | bd68151 | 2011-07-16 15:23:14 -0400 | [diff] [blame] | 124 | */ |
Jeff Mahoney | 143bede | 2012-03-01 14:56:26 +0100 | [diff] [blame] | 125 | void btrfs_tree_unlock(struct extent_buffer *eb) |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 126 | { |
Qu Wenruo | 31aab40 | 2019-04-15 21:15:25 +0800 | [diff] [blame] | 127 | trace_btrfs_tree_unlock(eb); |
Chris Mason | ea4ebde | 2014-06-19 14:16:52 -0700 | [diff] [blame] | 128 | eb->lock_owner = 0; |
Josef Bacik | 196d59a | 2020-08-20 11:46:09 -0400 | [diff] [blame] | 129 | up_write(&eb->lock); |
Chris Mason | 925baed | 2008-06-25 16:01:30 -0400 | [diff] [blame] | 130 | } |
David Sterba | ed2b1d3 | 2019-09-24 19:17:17 +0200 | [diff] [blame] | 131 | |
| 132 | /* |
David Sterba | 1f95ec0 | 2019-09-24 19:17:17 +0200 | [diff] [blame] | 133 | * This releases any locks held in the path starting at level and going all the |
| 134 | * way up to the root. |
| 135 | * |
| 136 | * btrfs_search_slot will keep the lock held on higher nodes in a few corner |
| 137 | * cases, such as COW of the block at slot zero in the node. This ignores |
| 138 | * those rules, and it should only be called when there are no more updates to |
| 139 | * be done higher up in the tree. |
| 140 | */ |
| 141 | void btrfs_unlock_up_safe(struct btrfs_path *path, int level) |
| 142 | { |
| 143 | int i; |
| 144 | |
| 145 | if (path->keep_locks) |
| 146 | return; |
| 147 | |
| 148 | for (i = level; i < BTRFS_MAX_LEVEL; i++) { |
| 149 | if (!path->nodes[i]) |
| 150 | continue; |
| 151 | if (!path->locks[i]) |
| 152 | continue; |
| 153 | btrfs_tree_unlock_rw(path->nodes[i], path->locks[i]); |
| 154 | path->locks[i] = 0; |
| 155 | } |
| 156 | } |
David Sterba | b908c33 | 2020-02-05 17:26:51 +0100 | [diff] [blame] | 157 | |
| 158 | /* |
| 159 | * Loop around taking references on and locking the root node of the tree until |
| 160 | * we end up with a lock on the root node. |
| 161 | * |
| 162 | * Return: root extent buffer with write lock held |
| 163 | */ |
| 164 | struct extent_buffer *btrfs_lock_root_node(struct btrfs_root *root) |
| 165 | { |
| 166 | struct extent_buffer *eb; |
| 167 | |
| 168 | while (1) { |
| 169 | eb = btrfs_root_node(root); |
| 170 | btrfs_tree_lock(eb); |
| 171 | if (eb == root->node) |
| 172 | break; |
| 173 | btrfs_tree_unlock(eb); |
| 174 | free_extent_buffer(eb); |
| 175 | } |
| 176 | return eb; |
| 177 | } |
| 178 | |
| 179 | /* |
| 180 | * Loop around taking references on and locking the root node of the tree until |
| 181 | * we end up with a lock on the root node. |
| 182 | * |
| 183 | * Return: root extent buffer with read lock held |
| 184 | */ |
Josef Bacik | 1bb9659 | 2020-11-06 16:27:33 -0500 | [diff] [blame] | 185 | struct extent_buffer *btrfs_read_lock_root_node(struct btrfs_root *root) |
David Sterba | b908c33 | 2020-02-05 17:26:51 +0100 | [diff] [blame] | 186 | { |
| 187 | struct extent_buffer *eb; |
| 188 | |
| 189 | while (1) { |
| 190 | eb = btrfs_root_node(root); |
Josef Bacik | 1bb9659 | 2020-11-06 16:27:33 -0500 | [diff] [blame] | 191 | btrfs_tree_read_lock(eb); |
David Sterba | b908c33 | 2020-02-05 17:26:51 +0100 | [diff] [blame] | 192 | if (eb == root->node) |
| 193 | break; |
| 194 | btrfs_tree_read_unlock(eb); |
| 195 | free_extent_buffer(eb); |
| 196 | } |
| 197 | return eb; |
| 198 | } |
Nikolay Borisov | 2992df7 | 2020-01-30 14:59:44 +0200 | [diff] [blame] | 199 | |
| 200 | /* |
| 201 | * DREW locks |
| 202 | * ========== |
| 203 | * |
| 204 | * DREW stands for double-reader-writer-exclusion lock. It's used in situation |
| 205 | * where you want to provide A-B exclusion but not AA or BB. |
| 206 | * |
| 207 | * Currently implementation gives more priority to reader. If a reader and a |
| 208 | * writer both race to acquire their respective sides of the lock the writer |
| 209 | * would yield its lock as soon as it detects a concurrent reader. Additionally |
| 210 | * if there are pending readers no new writers would be allowed to come in and |
| 211 | * acquire the lock. |
| 212 | */ |
| 213 | |
| 214 | int btrfs_drew_lock_init(struct btrfs_drew_lock *lock) |
| 215 | { |
| 216 | int ret; |
| 217 | |
| 218 | ret = percpu_counter_init(&lock->writers, 0, GFP_KERNEL); |
| 219 | if (ret) |
| 220 | return ret; |
| 221 | |
| 222 | atomic_set(&lock->readers, 0); |
| 223 | init_waitqueue_head(&lock->pending_readers); |
| 224 | init_waitqueue_head(&lock->pending_writers); |
| 225 | |
| 226 | return 0; |
| 227 | } |
| 228 | |
| 229 | void btrfs_drew_lock_destroy(struct btrfs_drew_lock *lock) |
| 230 | { |
| 231 | percpu_counter_destroy(&lock->writers); |
| 232 | } |
| 233 | |
| 234 | /* Return true if acquisition is successful, false otherwise */ |
| 235 | bool btrfs_drew_try_write_lock(struct btrfs_drew_lock *lock) |
| 236 | { |
| 237 | if (atomic_read(&lock->readers)) |
| 238 | return false; |
| 239 | |
| 240 | percpu_counter_inc(&lock->writers); |
| 241 | |
| 242 | /* Ensure writers count is updated before we check for pending readers */ |
| 243 | smp_mb(); |
| 244 | if (atomic_read(&lock->readers)) { |
| 245 | btrfs_drew_write_unlock(lock); |
| 246 | return false; |
| 247 | } |
| 248 | |
| 249 | return true; |
| 250 | } |
| 251 | |
| 252 | void btrfs_drew_write_lock(struct btrfs_drew_lock *lock) |
| 253 | { |
| 254 | while (true) { |
| 255 | if (btrfs_drew_try_write_lock(lock)) |
| 256 | return; |
| 257 | wait_event(lock->pending_writers, !atomic_read(&lock->readers)); |
| 258 | } |
| 259 | } |
| 260 | |
| 261 | void btrfs_drew_write_unlock(struct btrfs_drew_lock *lock) |
| 262 | { |
| 263 | percpu_counter_dec(&lock->writers); |
| 264 | cond_wake_up(&lock->pending_readers); |
| 265 | } |
| 266 | |
| 267 | void btrfs_drew_read_lock(struct btrfs_drew_lock *lock) |
| 268 | { |
| 269 | atomic_inc(&lock->readers); |
| 270 | |
| 271 | /* |
| 272 | * Ensure the pending reader count is perceieved BEFORE this reader |
| 273 | * goes to sleep in case of active writers. This guarantees new writers |
| 274 | * won't be allowed and that the current reader will be woken up when |
| 275 | * the last active writer finishes its jobs. |
| 276 | */ |
| 277 | smp_mb__after_atomic(); |
| 278 | |
| 279 | wait_event(lock->pending_readers, |
| 280 | percpu_counter_sum(&lock->writers) == 0); |
| 281 | } |
| 282 | |
| 283 | void btrfs_drew_read_unlock(struct btrfs_drew_lock *lock) |
| 284 | { |
| 285 | /* |
| 286 | * atomic_dec_and_test implies a full barrier, so woken up writers |
| 287 | * are guaranteed to see the decrement |
| 288 | */ |
| 289 | if (atomic_dec_and_test(&lock->readers)) |
| 290 | wake_up(&lock->pending_writers); |
| 291 | } |