blob: e7758778abefa4659bbcf6e9be7909a790938a24 [file] [log] [blame]
Thomas Gleixner328970d2019-05-24 12:04:05 +02001// SPDX-License-Identifier: GPL-2.0-or-later
Masahiro Yamadafa60ce22021-05-06 18:06:44 -07002/*
Mark Fashehccd979b2005-12-15 14:31:24 -08003 * io.c
4 *
5 * Buffer cache handling
6 *
7 * Copyright (C) 2002, 2004 Oracle. All rights reserved.
Mark Fashehccd979b2005-12-15 14:31:24 -08008 */
9
10#include <linux/fs.h>
11#include <linux/types.h>
Mark Fashehccd979b2005-12-15 14:31:24 -080012#include <linux/highmem.h>
Christoph Hellwig2f8b5442016-11-01 07:40:13 -060013#include <linux/bio.h>
Mark Fashehccd979b2005-12-15 14:31:24 -080014
15#include <cluster/masklog.h>
16
17#include "ocfs2.h"
18
19#include "alloc.h"
20#include "inode.h"
21#include "journal.h"
22#include "uptodate.h"
Mark Fashehccd979b2005-12-15 14:31:24 -080023#include "buffer_head_io.h"
Tao Ma15057e92011-02-24 16:09:38 +080024#include "ocfs2_trace.h"
Mark Fashehccd979b2005-12-15 14:31:24 -080025
Joel Becker970e4932008-11-13 14:49:19 -080026/*
27 * Bits on bh->b_state used by ocfs2.
28 *
Mark Fashehb86c86fa2008-11-18 17:16:47 -080029 * These MUST be after the JBD2 bits. Hence, we use BH_JBDPrivateStart.
Joel Becker970e4932008-11-13 14:49:19 -080030 */
31enum ocfs2_state_bits {
Mark Fashehb86c86fa2008-11-18 17:16:47 -080032 BH_NeedsValidate = BH_JBDPrivateStart,
Joel Becker970e4932008-11-13 14:49:19 -080033};
34
35/* Expand the magic b_state functions */
36BUFFER_FNS(NeedsValidate, needs_validate);
37
Mark Fashehccd979b2005-12-15 14:31:24 -080038int ocfs2_write_block(struct ocfs2_super *osb, struct buffer_head *bh,
Joel Becker8cb471e2009-02-10 20:00:41 -080039 struct ocfs2_caching_info *ci)
Mark Fashehccd979b2005-12-15 14:31:24 -080040{
41 int ret = 0;
42
Tao Ma15057e92011-02-24 16:09:38 +080043 trace_ocfs2_write_block((unsigned long long)bh->b_blocknr, ci);
Mark Fashehccd979b2005-12-15 14:31:24 -080044
45 BUG_ON(bh->b_blocknr < OCFS2_SUPER_BLOCK_BLKNO);
46 BUG_ON(buffer_jbd(bh));
47
48 /* No need to check for a soft readonly file system here. non
49 * journalled writes are only ever done on system files which
50 * can get modified during recovery even if read-only. */
51 if (ocfs2_is_hard_readonly(osb)) {
52 ret = -EROFS;
Tao Mac1e8d352011-03-07 16:43:21 +080053 mlog_errno(ret);
Mark Fashehccd979b2005-12-15 14:31:24 -080054 goto out;
55 }
56
Joel Becker8cb471e2009-02-10 20:00:41 -080057 ocfs2_metadata_cache_io_lock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -080058
59 lock_buffer(bh);
60 set_buffer_uptodate(bh);
61
62 /* remove from dirty list before I/O. */
63 clear_buffer_dirty(bh);
64
Joel Beckerda1e9092008-10-09 17:20:29 -070065 get_bh(bh); /* for end_buffer_write_sync() */
Mark Fashehccd979b2005-12-15 14:31:24 -080066 bh->b_end_io = end_buffer_write_sync;
Mike Christie2a222ca2016-06-05 14:31:43 -050067 submit_bh(REQ_OP_WRITE, 0, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -080068
69 wait_on_buffer(bh);
70
71 if (buffer_uptodate(bh)) {
Joel Becker8cb471e2009-02-10 20:00:41 -080072 ocfs2_set_buffer_uptodate(ci, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -080073 } else {
74 /* We don't need to remove the clustered uptodate
75 * information for this bh as it's not marked locally
76 * uptodate. */
77 ret = -EIO;
Tao Mac1e8d352011-03-07 16:43:21 +080078 mlog_errno(ret);
Mark Fashehccd979b2005-12-15 14:31:24 -080079 }
80
Joel Becker8cb471e2009-02-10 20:00:41 -080081 ocfs2_metadata_cache_io_unlock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -080082out:
Mark Fashehccd979b2005-12-15 14:31:24 -080083 return ret;
84}
85
Changwei Gecf76c782018-11-02 15:48:19 -070086/* Caller must provide a bhs[] with all NULL or non-NULL entries, so it
87 * will be easier to handle read failure.
88 */
Joel Beckerda1e9092008-10-09 17:20:29 -070089int ocfs2_read_blocks_sync(struct ocfs2_super *osb, u64 block,
90 unsigned int nr, struct buffer_head *bhs[])
91{
92 int status = 0;
93 unsigned int i;
94 struct buffer_head *bh;
Changwei Gecf76c782018-11-02 15:48:19 -070095 int new_bh = 0;
Joel Beckerda1e9092008-10-09 17:20:29 -070096
Tao Ma15057e92011-02-24 16:09:38 +080097 trace_ocfs2_read_blocks_sync((unsigned long long)block, nr);
98
99 if (!nr)
Joel Beckerda1e9092008-10-09 17:20:29 -0700100 goto bail;
Joel Beckerda1e9092008-10-09 17:20:29 -0700101
Changwei Gecf76c782018-11-02 15:48:19 -0700102 /* Don't put buffer head and re-assign it to NULL if it is allocated
103 * outside since the caller can't be aware of this alternation!
104 */
105 new_bh = (bhs[0] == NULL);
106
Joel Beckerda1e9092008-10-09 17:20:29 -0700107 for (i = 0 ; i < nr ; i++) {
108 if (bhs[i] == NULL) {
109 bhs[i] = sb_getblk(osb->sb, block++);
110 if (bhs[i] == NULL) {
Rui Xiang7391a292013-11-12 15:06:54 -0800111 status = -ENOMEM;
Joel Beckerda1e9092008-10-09 17:20:29 -0700112 mlog_errno(status);
Changwei Gecf76c782018-11-02 15:48:19 -0700113 break;
Joel Beckerda1e9092008-10-09 17:20:29 -0700114 }
115 }
116 bh = bhs[i];
117
118 if (buffer_jbd(bh)) {
Tao Ma15057e92011-02-24 16:09:38 +0800119 trace_ocfs2_read_blocks_sync_jbd(
120 (unsigned long long)bh->b_blocknr);
Joel Beckerda1e9092008-10-09 17:20:29 -0700121 continue;
122 }
123
124 if (buffer_dirty(bh)) {
125 /* This should probably be a BUG, or
126 * at least return an error. */
127 mlog(ML_ERROR,
128 "trying to sync read a dirty "
129 "buffer! (blocknr = %llu), skipping\n",
130 (unsigned long long)bh->b_blocknr);
131 continue;
132 }
133
134 lock_buffer(bh);
135 if (buffer_jbd(bh)) {
Gang He7186ee02016-06-24 14:50:13 -0700136#ifdef CATCH_BH_JBD_RACES
Joel Beckerda1e9092008-10-09 17:20:29 -0700137 mlog(ML_ERROR,
138 "block %llu had the JBD bit set "
139 "while I was in lock_buffer!",
140 (unsigned long long)bh->b_blocknr);
141 BUG();
Gang He7186ee02016-06-24 14:50:13 -0700142#else
143 unlock_buffer(bh);
144 continue;
145#endif
Joel Beckerda1e9092008-10-09 17:20:29 -0700146 }
147
Joel Beckerda1e9092008-10-09 17:20:29 -0700148 get_bh(bh); /* for end_buffer_read_sync() */
149 bh->b_end_io = end_buffer_read_sync;
Mike Christie2a222ca2016-06-05 14:31:43 -0500150 submit_bh(REQ_OP_READ, 0, bh);
Joel Beckerda1e9092008-10-09 17:20:29 -0700151 }
152
Changwei Gecf76c782018-11-02 15:48:19 -0700153read_failure:
Joel Beckerda1e9092008-10-09 17:20:29 -0700154 for (i = nr; i > 0; i--) {
155 bh = bhs[i - 1];
156
Changwei Gecf76c782018-11-02 15:48:19 -0700157 if (unlikely(status)) {
158 if (new_bh && bh) {
159 /* If middle bh fails, let previous bh
160 * finish its read and then put it to
161 * aovoid bh leak
162 */
163 if (!buffer_jbd(bh))
164 wait_on_buffer(bh);
165 put_bh(bh);
166 bhs[i - 1] = NULL;
167 } else if (bh && buffer_uptodate(bh)) {
168 clear_buffer_uptodate(bh);
169 }
170 continue;
171 }
172
Mark Fashehd6b58f82008-11-21 14:06:55 -0800173 /* No need to wait on the buffer if it's managed by JBD. */
174 if (!buffer_jbd(bh))
175 wait_on_buffer(bh);
Joel Beckerda1e9092008-10-09 17:20:29 -0700176
Joel Beckerda1e9092008-10-09 17:20:29 -0700177 if (!buffer_uptodate(bh)) {
178 /* Status won't be cleared from here on out,
179 * so we can safely record this and loop back
180 * to cleanup the other buffers. */
181 status = -EIO;
Changwei Gecf76c782018-11-02 15:48:19 -0700182 goto read_failure;
Joel Beckerda1e9092008-10-09 17:20:29 -0700183 }
184 }
185
186bail:
187 return status;
188}
189
Changwei Gecf76c782018-11-02 15:48:19 -0700190/* Caller must provide a bhs[] with all NULL or non-NULL entries, so it
191 * will be easier to handle read failure.
192 */
Joel Becker8cb471e2009-02-10 20:00:41 -0800193int ocfs2_read_blocks(struct ocfs2_caching_info *ci, u64 block, int nr,
Joel Becker970e4932008-11-13 14:49:19 -0800194 struct buffer_head *bhs[], int flags,
195 int (*validate)(struct super_block *sb,
196 struct buffer_head *bh))
Mark Fashehccd979b2005-12-15 14:31:24 -0800197{
198 int status = 0;
Mark Fashehccd979b2005-12-15 14:31:24 -0800199 int i, ignore_cache = 0;
200 struct buffer_head *bh;
Joel Becker8cb471e2009-02-10 20:00:41 -0800201 struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
Changwei Gecf76c782018-11-02 15:48:19 -0700202 int new_bh = 0;
Mark Fashehccd979b2005-12-15 14:31:24 -0800203
Tao Ma15057e92011-02-24 16:09:38 +0800204 trace_ocfs2_read_blocks_begin(ci, (unsigned long long)block, nr, flags);
Mark Fashehccd979b2005-12-15 14:31:24 -0800205
Joel Becker8cb471e2009-02-10 20:00:41 -0800206 BUG_ON(!ci);
Joel Beckerd4a8c932008-10-09 17:20:34 -0700207 BUG_ON((flags & OCFS2_BH_READAHEAD) &&
208 (flags & OCFS2_BH_IGNORE_CACHE));
Mark Fashehaa958872006-04-21 13:49:02 -0700209
Joel Becker31d33072008-10-09 17:20:30 -0700210 if (bhs == NULL) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800211 status = -EINVAL;
212 mlog_errno(status);
213 goto bail;
214 }
215
216 if (nr < 0) {
217 mlog(ML_ERROR, "asked to read %d blocks!\n", nr);
218 status = -EINVAL;
219 mlog_errno(status);
220 goto bail;
221 }
222
223 if (nr == 0) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800224 status = 0;
225 goto bail;
226 }
227
Changwei Gecf76c782018-11-02 15:48:19 -0700228 /* Don't put buffer head and re-assign it to NULL if it is allocated
229 * outside since the caller can't be aware of this alternation!
230 */
231 new_bh = (bhs[0] == NULL);
232
Joel Becker8cb471e2009-02-10 20:00:41 -0800233 ocfs2_metadata_cache_io_lock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -0800234 for (i = 0 ; i < nr ; i++) {
235 if (bhs[i] == NULL) {
Joel Becker8cb471e2009-02-10 20:00:41 -0800236 bhs[i] = sb_getblk(sb, block++);
Mark Fashehccd979b2005-12-15 14:31:24 -0800237 if (bhs[i] == NULL) {
Joel Becker8cb471e2009-02-10 20:00:41 -0800238 ocfs2_metadata_cache_io_unlock(ci);
Rui Xiang7391a292013-11-12 15:06:54 -0800239 status = -ENOMEM;
Mark Fashehccd979b2005-12-15 14:31:24 -0800240 mlog_errno(status);
Changwei Gecf76c782018-11-02 15:48:19 -0700241 /* Don't forget to put previous bh! */
242 break;
Mark Fashehccd979b2005-12-15 14:31:24 -0800243 }
244 }
245 bh = bhs[i];
Joel Beckerd4a8c932008-10-09 17:20:34 -0700246 ignore_cache = (flags & OCFS2_BH_IGNORE_CACHE);
Mark Fashehccd979b2005-12-15 14:31:24 -0800247
Mark Fashehaa958872006-04-21 13:49:02 -0700248 /* There are three read-ahead cases here which we need to
249 * be concerned with. All three assume a buffer has
250 * previously been submitted with OCFS2_BH_READAHEAD
251 * and it hasn't yet completed I/O.
252 *
253 * 1) The current request is sync to disk. This rarely
254 * happens these days, and never when performance
255 * matters - the code can just wait on the buffer
256 * lock and re-submit.
257 *
258 * 2) The current request is cached, but not
259 * readahead. ocfs2_buffer_uptodate() will return
260 * false anyway, so we'll wind up waiting on the
261 * buffer lock to do I/O. We re-check the request
262 * with after getting the lock to avoid a re-submit.
263 *
264 * 3) The current request is readahead (and so must
265 * also be a caching one). We short circuit if the
266 * buffer is locked (under I/O) and if it's in the
267 * uptodate cache. The re-check from #2 catches the
268 * case that the previous read-ahead completes just
269 * before our is-it-in-flight check.
270 */
271
Joel Becker8cb471e2009-02-10 20:00:41 -0800272 if (!ignore_cache && !ocfs2_buffer_uptodate(ci, bh)) {
Tao Mad7014852011-02-24 16:22:20 +0800273 trace_ocfs2_read_blocks_from_disk(
Mark Fashehccd979b2005-12-15 14:31:24 -0800274 (unsigned long long)bh->b_blocknr,
Joel Becker8cb471e2009-02-10 20:00:41 -0800275 (unsigned long long)ocfs2_metadata_cache_owner(ci));
Joel Beckerd4a8c932008-10-09 17:20:34 -0700276 /* We're using ignore_cache here to say
277 * "go to disk" */
Mark Fashehccd979b2005-12-15 14:31:24 -0800278 ignore_cache = 1;
279 }
280
Tao Ma15057e92011-02-24 16:09:38 +0800281 trace_ocfs2_read_blocks_bh((unsigned long long)bh->b_blocknr,
282 ignore_cache, buffer_jbd(bh), buffer_dirty(bh));
283
Mark Fashehccd979b2005-12-15 14:31:24 -0800284 if (buffer_jbd(bh)) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800285 continue;
286 }
287
Joel Beckerd4a8c932008-10-09 17:20:34 -0700288 if (ignore_cache) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800289 if (buffer_dirty(bh)) {
290 /* This should probably be a BUG, or
291 * at least return an error. */
Mark Fashehccd979b2005-12-15 14:31:24 -0800292 continue;
293 }
294
Mark Fashehaa958872006-04-21 13:49:02 -0700295 /* A read-ahead request was made - if the
296 * buffer is already under read-ahead from a
297 * previously submitted request than we are
298 * done here. */
299 if ((flags & OCFS2_BH_READAHEAD)
Joel Becker8cb471e2009-02-10 20:00:41 -0800300 && ocfs2_buffer_read_ahead(ci, bh))
Mark Fashehaa958872006-04-21 13:49:02 -0700301 continue;
302
Mark Fashehccd979b2005-12-15 14:31:24 -0800303 lock_buffer(bh);
304 if (buffer_jbd(bh)) {
305#ifdef CATCH_BH_JBD_RACES
306 mlog(ML_ERROR, "block %llu had the JBD bit set "
307 "while I was in lock_buffer!",
308 (unsigned long long)bh->b_blocknr);
309 BUG();
310#else
311 unlock_buffer(bh);
312 continue;
313#endif
314 }
Mark Fashehaa958872006-04-21 13:49:02 -0700315
316 /* Re-check ocfs2_buffer_uptodate() as a
317 * previously read-ahead buffer may have
318 * completed I/O while we were waiting for the
319 * buffer lock. */
Joel Beckerd4a8c932008-10-09 17:20:34 -0700320 if (!(flags & OCFS2_BH_IGNORE_CACHE)
Mark Fashehaa958872006-04-21 13:49:02 -0700321 && !(flags & OCFS2_BH_READAHEAD)
Joel Becker8cb471e2009-02-10 20:00:41 -0800322 && ocfs2_buffer_uptodate(ci, bh)) {
Mark Fashehaa958872006-04-21 13:49:02 -0700323 unlock_buffer(bh);
324 continue;
325 }
326
Mark Fashehccd979b2005-12-15 14:31:24 -0800327 get_bh(bh); /* for end_buffer_read_sync() */
Joel Becker970e4932008-11-13 14:49:19 -0800328 if (validate)
329 set_buffer_needs_validate(bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800330 bh->b_end_io = end_buffer_read_sync;
Mike Christie2a222ca2016-06-05 14:31:43 -0500331 submit_bh(REQ_OP_READ, 0, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800332 continue;
333 }
334 }
335
Changwei Gecf76c782018-11-02 15:48:19 -0700336read_failure:
Mark Fashehccd979b2005-12-15 14:31:24 -0800337 for (i = (nr - 1); i >= 0; i--) {
338 bh = bhs[i];
339
Mark Fashehaa958872006-04-21 13:49:02 -0700340 if (!(flags & OCFS2_BH_READAHEAD)) {
Changwei Gecf76c782018-11-02 15:48:19 -0700341 if (unlikely(status)) {
342 /* Clear the buffers on error including those
343 * ever succeeded in reading
344 */
345 if (new_bh && bh) {
346 /* If middle bh fails, let previous bh
347 * finish its read and then put it to
348 * aovoid bh leak
349 */
350 if (!buffer_jbd(bh))
351 wait_on_buffer(bh);
352 put_bh(bh);
353 bhs[i] = NULL;
354 } else if (bh && buffer_uptodate(bh)) {
355 clear_buffer_uptodate(bh);
356 }
Goldwyn Rodrigues34237682015-09-04 15:44:20 -0700357 continue;
358 }
Mark Fashehaa958872006-04-21 13:49:02 -0700359 /* We know this can't have changed as we hold the
Joel Becker8cb471e2009-02-10 20:00:41 -0800360 * owner sem. Avoid doing any work on the bh if the
Mark Fashehaa958872006-04-21 13:49:02 -0700361 * journal has it. */
362 if (!buffer_jbd(bh))
363 wait_on_buffer(bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800364
Mark Fashehaa958872006-04-21 13:49:02 -0700365 if (!buffer_uptodate(bh)) {
366 /* Status won't be cleared from here on out,
367 * so we can safely record this and loop back
368 * to cleanup the other buffers. Don't need to
369 * remove the clustered uptodate information
370 * for this bh as it's not marked locally
371 * uptodate. */
372 status = -EIO;
Junxiao Bi234b69e2018-09-20 12:22:51 -0700373 clear_buffer_needs_validate(bh);
Changwei Gecf76c782018-11-02 15:48:19 -0700374 goto read_failure;
Mark Fashehaa958872006-04-21 13:49:02 -0700375 }
Joel Becker970e4932008-11-13 14:49:19 -0800376
377 if (buffer_needs_validate(bh)) {
378 /* We never set NeedsValidate if the
379 * buffer was held by the journal, so
380 * that better not have changed */
381 BUG_ON(buffer_jbd(bh));
382 clear_buffer_needs_validate(bh);
Joel Becker8cb471e2009-02-10 20:00:41 -0800383 status = validate(sb, bh);
Changwei Gecf76c782018-11-02 15:48:19 -0700384 if (status)
385 goto read_failure;
Joel Becker970e4932008-11-13 14:49:19 -0800386 }
Mark Fashehccd979b2005-12-15 14:31:24 -0800387 }
388
Mark Fashehaa958872006-04-21 13:49:02 -0700389 /* Always set the buffer in the cache, even if it was
390 * a forced read, or read-ahead which hasn't yet
391 * completed. */
Joel Becker8cb471e2009-02-10 20:00:41 -0800392 ocfs2_set_buffer_uptodate(ci, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800393 }
Joel Becker8cb471e2009-02-10 20:00:41 -0800394 ocfs2_metadata_cache_io_unlock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -0800395
Tao Ma15057e92011-02-24 16:09:38 +0800396 trace_ocfs2_read_blocks_end((unsigned long long)block, nr,
397 flags, ignore_cache);
Mark Fashehccd979b2005-12-15 14:31:24 -0800398
399bail:
400
Mark Fashehccd979b2005-12-15 14:31:24 -0800401 return status;
402}
Tao Mad6590722007-12-18 15:47:03 +0800403
404/* Check whether the blkno is the super block or one of the backups. */
405static void ocfs2_check_super_or_backup(struct super_block *sb,
406 sector_t blkno)
407{
408 int i;
409 u64 backup_blkno;
410
411 if (blkno == OCFS2_SUPER_BLOCK_BLKNO)
412 return;
413
414 for (i = 0; i < OCFS2_MAX_BACKUP_SUPERBLOCKS; i++) {
415 backup_blkno = ocfs2_backup_super_blkno(sb, i);
416 if (backup_blkno == blkno)
417 return;
418 }
419
420 BUG();
421}
422
423/*
424 * Write super block and backups doesn't need to collaborate with journal,
Joel Becker8cb471e2009-02-10 20:00:41 -0800425 * so we don't need to lock ip_io_mutex and ci doesn't need to bea passed
Tao Mad6590722007-12-18 15:47:03 +0800426 * into this function.
427 */
428int ocfs2_write_super_or_backup(struct ocfs2_super *osb,
429 struct buffer_head *bh)
430{
431 int ret = 0;
Joel Beckera42ab8e2010-03-31 18:25:44 -0700432 struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data;
Tao Mad6590722007-12-18 15:47:03 +0800433
Tao Mad6590722007-12-18 15:47:03 +0800434 BUG_ON(buffer_jbd(bh));
435 ocfs2_check_super_or_backup(osb->sb, bh->b_blocknr);
436
437 if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) {
438 ret = -EROFS;
Tao Mac1e8d352011-03-07 16:43:21 +0800439 mlog_errno(ret);
Tao Mad6590722007-12-18 15:47:03 +0800440 goto out;
441 }
442
443 lock_buffer(bh);
444 set_buffer_uptodate(bh);
445
446 /* remove from dirty list before I/O. */
447 clear_buffer_dirty(bh);
448
449 get_bh(bh); /* for end_buffer_write_sync() */
450 bh->b_end_io = end_buffer_write_sync;
Joel Beckera42ab8e2010-03-31 18:25:44 -0700451 ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &di->i_check);
Mike Christie2a222ca2016-06-05 14:31:43 -0500452 submit_bh(REQ_OP_WRITE, 0, bh);
Tao Mad6590722007-12-18 15:47:03 +0800453
454 wait_on_buffer(bh);
455
456 if (!buffer_uptodate(bh)) {
457 ret = -EIO;
Tao Mac1e8d352011-03-07 16:43:21 +0800458 mlog_errno(ret);
Tao Mad6590722007-12-18 15:47:03 +0800459 }
460
461out:
Tao Mad6590722007-12-18 15:47:03 +0800462 return ret;
463}