blob: e855d8260433a2c890f0cc30b10980717c1c2858 [file] [log] [blame]
Theodore Ts'of5166762017-12-17 22:00:59 -05001// SPDX-License-Identifier: GPL-2.0+
Dave Kleikamp470decc2006-10-11 01:20:57 -07002/*
Mingming Caof7f4bcc2006-10-11 01:20:59 -07003 * linux/fs/jbd2/commit.c
Dave Kleikamp470decc2006-10-11 01:20:57 -07004 *
5 * Written by Stephen C. Tweedie <sct@redhat.com>, 1998
6 *
7 * Copyright 1998 Red Hat corp --- All Rights Reserved
8 *
Dave Kleikamp470decc2006-10-11 01:20:57 -07009 * Journal commit routines for the generic filesystem journaling code;
10 * part of the ext2fs journaling system.
11 */
12
13#include <linux/time.h>
14#include <linux/fs.h>
Mingming Caof7f4bcc2006-10-11 01:20:59 -070015#include <linux/jbd2.h>
Dave Kleikamp470decc2006-10-11 01:20:57 -070016#include <linux/errno.h>
17#include <linux/slab.h>
18#include <linux/mm.h>
19#include <linux/pagemap.h>
Johann Lombardi8e85fb32008-01-28 23:58:27 -050020#include <linux/jiffies.h>
Girish Shilamkar818d2762008-01-28 23:58:27 -050021#include <linux/crc32.h>
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -040022#include <linux/writeback.h>
23#include <linux/backing-dev.h>
Theodore Ts'ofd984962009-01-05 21:34:13 -050024#include <linux/bio.h>
Theodore Ts'o0e3d2a62009-09-11 09:30:12 -040025#include <linux/blkdev.h>
Brian King39e3ac22010-10-27 21:25:12 -040026#include <linux/bitops.h>
Theodore Ts'o879c5e62009-06-17 11:47:48 -040027#include <trace/events/jbd2.h>
Dave Kleikamp470decc2006-10-11 01:20:57 -070028
29/*
Jan Karab34090e2013-06-04 12:08:56 -040030 * IO end handler for temporary buffer_heads handling writes to the journal.
Dave Kleikamp470decc2006-10-11 01:20:57 -070031 */
32static void journal_end_buffer_io_sync(struct buffer_head *bh, int uptodate)
33{
Jan Karab34090e2013-06-04 12:08:56 -040034 struct buffer_head *orig_bh = bh->b_private;
35
Dave Kleikamp470decc2006-10-11 01:20:57 -070036 BUFFER_TRACE(bh, "");
37 if (uptodate)
38 set_buffer_uptodate(bh);
39 else
40 clear_buffer_uptodate(bh);
Jan Karab34090e2013-06-04 12:08:56 -040041 if (orig_bh) {
42 clear_bit_unlock(BH_Shadow, &orig_bh->b_state);
Peter Zijlstra4e857c52014-03-17 18:06:10 +010043 smp_mb__after_atomic();
Jan Karab34090e2013-06-04 12:08:56 -040044 wake_up_bit(&orig_bh->b_state, BH_Shadow);
45 }
Dave Kleikamp470decc2006-10-11 01:20:57 -070046 unlock_buffer(bh);
47}
48
49/*
Jan Kara87c89c22008-07-11 19:27:31 -040050 * When an ext4 file is truncated, it is possible that some pages are not
51 * successfully freed, because they are attached to a committing transaction.
Dave Kleikamp470decc2006-10-11 01:20:57 -070052 * After the transaction commits, these pages are left on the LRU, with no
53 * ->mapping, and with attached buffers. These pages are trivially reclaimable
54 * by the VM, but their apparent absence upsets the VM accounting, and it makes
55 * the numbers in /proc/meminfo look odd.
56 *
57 * So here, we have a buffer which has just come off the forget list. Look to
58 * see if we can strip all buffers from the backing page.
59 *
60 * Called under lock_journal(), and possibly under journal_datalist_lock. The
61 * caller provided us with a ref against the buffer, and we drop that here.
62 */
63static void release_buffer_page(struct buffer_head *bh)
64{
65 struct page *page;
66
67 if (buffer_dirty(bh))
68 goto nope;
69 if (atomic_read(&bh->b_count) != 1)
70 goto nope;
71 page = bh->b_page;
72 if (!page)
73 goto nope;
74 if (page->mapping)
75 goto nope;
76
77 /* OK, it's a truncated page */
Nick Piggin529ae9a2008-08-02 12:01:03 +020078 if (!trylock_page(page))
Dave Kleikamp470decc2006-10-11 01:20:57 -070079 goto nope;
80
Kirill A. Shutemov09cbfea2016-04-01 15:29:47 +030081 get_page(page);
Dave Kleikamp470decc2006-10-11 01:20:57 -070082 __brelse(bh);
83 try_to_free_buffers(page);
84 unlock_page(page);
Kirill A. Shutemov09cbfea2016-04-01 15:29:47 +030085 put_page(page);
Dave Kleikamp470decc2006-10-11 01:20:57 -070086 return;
87
88nope:
89 __brelse(bh);
90}
91
Jan Karae5a120a2013-06-04 12:06:01 -040092static void jbd2_commit_block_csum_set(journal_t *j, struct buffer_head *bh)
Darrick J. Wong1f56c582012-05-27 08:10:25 -040093{
94 struct commit_header *h;
95 __u32 csum;
96
Darrick J. Wongdb9ee222014-08-27 18:40:07 -040097 if (!jbd2_journal_has_csum_v2or3(j))
Darrick J. Wong1f56c582012-05-27 08:10:25 -040098 return;
99
Jan Karae5a120a2013-06-04 12:06:01 -0400100 h = (struct commit_header *)(bh->b_data);
Darrick J. Wong1f56c582012-05-27 08:10:25 -0400101 h->h_chksum_type = 0;
102 h->h_chksum_size = 0;
103 h->h_chksum[0] = 0;
Jan Karae5a120a2013-06-04 12:06:01 -0400104 csum = jbd2_chksum(j, j->j_csum_seed, bh->b_data, j->j_blocksize);
Darrick J. Wong1f56c582012-05-27 08:10:25 -0400105 h->h_chksum[0] = cpu_to_be32(csum);
106}
107
Dave Kleikamp470decc2006-10-11 01:20:57 -0700108/*
Girish Shilamkar818d2762008-01-28 23:58:27 -0500109 * Done it all: now submit the commit record. We should have
Dave Kleikamp470decc2006-10-11 01:20:57 -0700110 * cleaned up our previous buffers by now, so if we are in abort
111 * mode we can now just skip the rest of the journal write
112 * entirely.
113 *
114 * Returns 1 if the journal needs to be aborted or 0 on success
115 */
Girish Shilamkar818d2762008-01-28 23:58:27 -0500116static int journal_submit_commit_record(journal_t *journal,
117 transaction_t *commit_transaction,
118 struct buffer_head **cbh,
119 __u32 crc32_sum)
Dave Kleikamp470decc2006-10-11 01:20:57 -0700120{
Girish Shilamkar818d2762008-01-28 23:58:27 -0500121 struct commit_header *tmp;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700122 struct buffer_head *bh;
Girish Shilamkar818d2762008-01-28 23:58:27 -0500123 int ret;
Arnd Bergmannb42d1d62018-07-29 15:51:47 -0400124 struct timespec64 now;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700125
Zhang Huan6cba6112011-04-05 19:16:20 -0400126 *cbh = NULL;
127
Dave Kleikamp470decc2006-10-11 01:20:57 -0700128 if (is_journal_aborted(journal))
129 return 0;
130
Jan Kara32ab6712016-02-22 23:17:15 -0500131 bh = jbd2_journal_get_descriptor_buffer(commit_transaction,
132 JBD2_COMMIT_BLOCK);
Jan Karae5a120a2013-06-04 12:06:01 -0400133 if (!bh)
Dave Kleikamp470decc2006-10-11 01:20:57 -0700134 return 1;
135
Girish Shilamkar818d2762008-01-28 23:58:27 -0500136 tmp = (struct commit_header *)bh->b_data;
Arnd Bergmannb42d1d62018-07-29 15:51:47 -0400137 ktime_get_coarse_real_ts64(&now);
Theodore Ts'o736603a2008-07-11 19:27:31 -0400138 tmp->h_commit_sec = cpu_to_be64(now.tv_sec);
139 tmp->h_commit_nsec = cpu_to_be32(now.tv_nsec);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500140
Darrick J. Wong56316a02015-10-17 16:18:45 -0400141 if (jbd2_has_feature_checksum(journal)) {
Girish Shilamkar818d2762008-01-28 23:58:27 -0500142 tmp->h_chksum_type = JBD2_CRC32_CHKSUM;
143 tmp->h_chksum_size = JBD2_CRC32_CHKSUM_SIZE;
144 tmp->h_chksum[0] = cpu_to_be32(crc32_sum);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700145 }
Jan Karae5a120a2013-06-04 12:06:01 -0400146 jbd2_commit_block_csum_set(journal, bh);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700147
Jan Karae5a120a2013-06-04 12:06:01 -0400148 BUFFER_TRACE(bh, "submit commit block");
Girish Shilamkar818d2762008-01-28 23:58:27 -0500149 lock_buffer(bh);
Theodore Ts'o45a90bf2008-10-06 12:04:02 -0400150 clear_buffer_dirty(bh);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500151 set_buffer_uptodate(bh);
152 bh->b_end_io = journal_end_buffer_io_sync;
153
154 if (journal->j_flags & JBD2_BARRIER &&
Darrick J. Wong56316a02015-10-17 16:18:45 -0400155 !jbd2_has_feature_async_commit(journal))
Christoph Hellwig70fd7612016-11-01 07:40:10 -0600156 ret = submit_bh(REQ_OP_WRITE,
157 REQ_SYNC | REQ_PREFLUSH | REQ_FUA, bh);
Christoph Hellwig9c355752010-08-18 05:29:17 -0400158 else
Christoph Hellwig70fd7612016-11-01 07:40:10 -0600159 ret = submit_bh(REQ_OP_WRITE, REQ_SYNC, bh);
Christoph Hellwig9c355752010-08-18 05:29:17 -0400160
Girish Shilamkar818d2762008-01-28 23:58:27 -0500161 *cbh = bh;
162 return ret;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700163}
164
Girish Shilamkar818d2762008-01-28 23:58:27 -0500165/*
166 * This function along with journal_submit_commit_record
167 * allows to write the commit record asynchronously.
168 */
Theodore Ts'ofd984962009-01-05 21:34:13 -0500169static int journal_wait_on_commit_record(journal_t *journal,
170 struct buffer_head *bh)
Girish Shilamkar818d2762008-01-28 23:58:27 -0500171{
172 int ret = 0;
173
174 clear_buffer_dirty(bh);
175 wait_on_buffer(bh);
176
177 if (unlikely(!buffer_uptodate(bh)))
178 ret = -EIO;
179 put_bh(bh); /* One for getblk() */
Girish Shilamkar818d2762008-01-28 23:58:27 -0500180
181 return ret;
182}
183
184/*
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400185 * write the filemap data using writepage() address_space_operations.
186 * We don't do block allocation here even for delalloc. We don't
Liu Songa4977302019-05-30 15:15:57 -0400187 * use writepages() because with delayed allocation we may be doing
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400188 * block allocation in writepages().
189 */
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400190static int journal_submit_inode_data_buffers(struct address_space *mapping,
191 loff_t dirty_start, loff_t dirty_end)
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400192{
193 int ret;
194 struct writeback_control wbc = {
195 .sync_mode = WB_SYNC_ALL,
196 .nr_to_write = mapping->nrpages * 2,
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400197 .range_start = dirty_start,
198 .range_end = dirty_end,
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400199 };
200
201 ret = generic_writepages(mapping, &wbc);
202 return ret;
203}
204
205/*
Jan Karac851ed52008-07-11 19:27:31 -0400206 * Submit all the data buffers of inode associated with the transaction to
207 * disk.
208 *
209 * We are in a committing transaction. Therefore no new inode can be added to
210 * our inode list. We use JI_COMMIT_RUNNING flag to protect inode we currently
211 * operate on from being released while we write out pages.
212 */
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400213static int journal_submit_data_buffers(journal_t *journal,
Jan Karac851ed52008-07-11 19:27:31 -0400214 transaction_t *commit_transaction)
215{
216 struct jbd2_inode *jinode;
217 int err, ret = 0;
218 struct address_space *mapping;
219
220 spin_lock(&journal->j_list_lock);
221 list_for_each_entry(jinode, &commit_transaction->t_inode_list, i_list) {
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400222 loff_t dirty_start = jinode->i_dirty_start;
223 loff_t dirty_end = jinode->i_dirty_end;
224
Jan Kara41617e12016-04-24 00:56:07 -0400225 if (!(jinode->i_flags & JI_WRITE_DATA))
226 continue;
Jan Karac851ed52008-07-11 19:27:31 -0400227 mapping = jinode->i_vfs_inode->i_mapping;
Jan Karacb0d9d42016-02-22 23:20:30 -0500228 jinode->i_flags |= JI_COMMIT_RUNNING;
Jan Karac851ed52008-07-11 19:27:31 -0400229 spin_unlock(&journal->j_list_lock);
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400230 /*
231 * submit the inode data buffers. We use writepage
232 * instead of writepages. Because writepages can do
233 * block allocation with delalloc. We need to write
234 * only allocated blocks here.
235 */
Theodore Ts'o879c5e62009-06-17 11:47:48 -0400236 trace_jbd2_submit_inode_data(jinode->i_vfs_inode);
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400237 err = journal_submit_inode_data_buffers(mapping, dirty_start,
238 dirty_end);
Jan Karac851ed52008-07-11 19:27:31 -0400239 if (!ret)
240 ret = err;
241 spin_lock(&journal->j_list_lock);
242 J_ASSERT(jinode->i_transaction == commit_transaction);
Jan Karacb0d9d42016-02-22 23:20:30 -0500243 jinode->i_flags &= ~JI_COMMIT_RUNNING;
244 smp_mb();
Jan Karac851ed52008-07-11 19:27:31 -0400245 wake_up_bit(&jinode->i_flags, __JI_COMMIT_RUNNING);
246 }
247 spin_unlock(&journal->j_list_lock);
248 return ret;
249}
250
251/*
252 * Wait for data submitted for writeout, refile inodes to proper
253 * transaction if needed.
254 *
255 */
256static int journal_finish_inode_data_buffers(journal_t *journal,
257 transaction_t *commit_transaction)
258{
259 struct jbd2_inode *jinode, *next_i;
260 int err, ret = 0;
261
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400262 /* For locking, see the comment in journal_submit_data_buffers() */
Jan Karac851ed52008-07-11 19:27:31 -0400263 spin_lock(&journal->j_list_lock);
264 list_for_each_entry(jinode, &commit_transaction->t_inode_list, i_list) {
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400265 loff_t dirty_start = jinode->i_dirty_start;
266 loff_t dirty_end = jinode->i_dirty_end;
267
Jan Kara41617e12016-04-24 00:56:07 -0400268 if (!(jinode->i_flags & JI_WAIT_DATA))
269 continue;
Jan Karacb0d9d42016-02-22 23:20:30 -0500270 jinode->i_flags |= JI_COMMIT_RUNNING;
Jan Karac851ed52008-07-11 19:27:31 -0400271 spin_unlock(&journal->j_list_lock);
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400272 err = filemap_fdatawait_range_keep_errors(
273 jinode->i_vfs_inode->i_mapping, dirty_start,
274 dirty_end);
Jeff Layton76341ca2017-07-06 07:02:22 -0400275 if (!ret)
276 ret = err;
Jan Karac851ed52008-07-11 19:27:31 -0400277 spin_lock(&journal->j_list_lock);
Jan Karacb0d9d42016-02-22 23:20:30 -0500278 jinode->i_flags &= ~JI_COMMIT_RUNNING;
279 smp_mb();
Jan Karac851ed52008-07-11 19:27:31 -0400280 wake_up_bit(&jinode->i_flags, __JI_COMMIT_RUNNING);
281 }
282
283 /* Now refile inode to proper lists */
284 list_for_each_entry_safe(jinode, next_i,
285 &commit_transaction->t_inode_list, i_list) {
286 list_del(&jinode->i_list);
287 if (jinode->i_next_transaction) {
288 jinode->i_transaction = jinode->i_next_transaction;
289 jinode->i_next_transaction = NULL;
290 list_add(&jinode->i_list,
291 &jinode->i_transaction->t_inode_list);
292 } else {
293 jinode->i_transaction = NULL;
Ross Zwisler6ba0e7d2019-06-20 17:24:56 -0400294 jinode->i_dirty_start = 0;
295 jinode->i_dirty_end = 0;
Jan Karac851ed52008-07-11 19:27:31 -0400296 }
297 }
298 spin_unlock(&journal->j_list_lock);
299
300 return ret;
301}
302
Girish Shilamkar818d2762008-01-28 23:58:27 -0500303static __u32 jbd2_checksum_data(__u32 crc32_sum, struct buffer_head *bh)
304{
305 struct page *page = bh->b_page;
306 char *addr;
307 __u32 checksum;
308
Cong Wang303a8f22011-11-25 23:14:31 +0800309 addr = kmap_atomic(page);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500310 checksum = crc32_be(crc32_sum,
311 (void *)(addr + offset_in_page(bh->b_data)), bh->b_size);
Cong Wang303a8f22011-11-25 23:14:31 +0800312 kunmap_atomic(addr);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500313
314 return checksum;
315}
316
Darrick J. Wongdb9ee222014-08-27 18:40:07 -0400317static void write_tag_block(journal_t *j, journal_block_tag_t *tag,
Mingming Cao18eba7a2006-10-11 01:21:13 -0700318 unsigned long long block)
Zach Brownb517bea2006-10-11 01:21:08 -0700319{
320 tag->t_blocknr = cpu_to_be32(block & (u32)~0);
Darrick J. Wong56316a02015-10-17 16:18:45 -0400321 if (jbd2_has_feature_64bit(j))
Zach Brownb517bea2006-10-11 01:21:08 -0700322 tag->t_blocknr_high = cpu_to_be32((block >> 31) >> 1);
323}
324
Darrick J. Wongc3900872012-05-27 08:12:12 -0400325static void jbd2_block_tag_csum_set(journal_t *j, journal_block_tag_t *tag,
326 struct buffer_head *bh, __u32 sequence)
327{
Darrick J. Wongdb9ee222014-08-27 18:40:07 -0400328 journal_block_tag3_t *tag3 = (journal_block_tag3_t *)tag;
Darrick J. Wongc3900872012-05-27 08:12:12 -0400329 struct page *page = bh->b_page;
330 __u8 *addr;
Darrick J. Wongeee06c52013-05-28 07:31:59 -0400331 __u32 csum32;
Darrick J. Wong18a6ea12013-08-28 14:59:58 -0400332 __be32 seq;
Darrick J. Wongc3900872012-05-27 08:12:12 -0400333
Darrick J. Wongdb9ee222014-08-27 18:40:07 -0400334 if (!jbd2_journal_has_csum_v2or3(j))
Darrick J. Wongc3900872012-05-27 08:12:12 -0400335 return;
336
Darrick J. Wong18a6ea12013-08-28 14:59:58 -0400337 seq = cpu_to_be32(sequence);
Cong Wang906adea2012-06-23 11:24:48 +0800338 addr = kmap_atomic(page);
Darrick J. Wong18a6ea12013-08-28 14:59:58 -0400339 csum32 = jbd2_chksum(j, j->j_csum_seed, (__u8 *)&seq, sizeof(seq));
Darrick J. Wongeee06c52013-05-28 07:31:59 -0400340 csum32 = jbd2_chksum(j, csum32, addr + offset_in_page(bh->b_data),
341 bh->b_size);
Cong Wang906adea2012-06-23 11:24:48 +0800342 kunmap_atomic(addr);
Darrick J. Wongc3900872012-05-27 08:12:12 -0400343
Darrick J. Wong56316a02015-10-17 16:18:45 -0400344 if (jbd2_has_feature_csum3(j))
Darrick J. Wongdb9ee222014-08-27 18:40:07 -0400345 tag3->t_checksum = cpu_to_be32(csum32);
346 else
347 tag->t_checksum = cpu_to_be16(csum32);
Darrick J. Wongc3900872012-05-27 08:12:12 -0400348}
Dave Kleikamp470decc2006-10-11 01:20:57 -0700349/*
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700350 * jbd2_journal_commit_transaction
Dave Kleikamp470decc2006-10-11 01:20:57 -0700351 *
352 * The primary function for committing a transaction to the log. This
353 * function is called by the journal thread to begin a complete commit.
354 */
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700355void jbd2_journal_commit_transaction(journal_t *journal)
Dave Kleikamp470decc2006-10-11 01:20:57 -0700356{
Johann Lombardi8e85fb32008-01-28 23:58:27 -0500357 struct transaction_stats_s stats;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700358 transaction_t *commit_transaction;
Jan Karae5a120a2013-06-04 12:06:01 -0400359 struct journal_head *jh;
360 struct buffer_head *descriptor;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700361 struct buffer_head **wbuf = journal->j_wbuf;
362 int bufs;
363 int flags;
364 int err;
Mingming Cao18eba7a2006-10-11 01:21:13 -0700365 unsigned long long blocknr;
Josef Bacike07f7182008-11-26 01:14:26 -0500366 ktime_t start_time;
367 u64 commit_time;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700368 char *tagp = NULL;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700369 journal_block_tag_t *tag = NULL;
370 int space_left = 0;
371 int first_tag = 0;
372 int tag_flag;
Dmitry Monakhov794446c2013-04-03 22:06:52 -0400373 int i;
Zach Brownb517bea2006-10-11 01:21:08 -0700374 int tag_bytes = journal_tag_bytes(journal);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500375 struct buffer_head *cbh = NULL; /* For transactional checksums */
376 __u32 crc32_sum = ~0;
Jens Axboe82f04ab2011-03-17 11:01:52 +0100377 struct blk_plug plug;
Jan Kara33395782012-03-13 22:45:38 -0400378 /* Tail of the journal */
379 unsigned long first_block;
380 tid_t first_tid;
381 int update_tail;
Darrick J. Wong3caa4872012-05-27 08:10:22 -0400382 int csum_size = 0;
Jan Karaf5113ef2013-06-04 12:01:45 -0400383 LIST_HEAD(io_bufs);
Jan Karae5a120a2013-06-04 12:06:01 -0400384 LIST_HEAD(log_bufs);
Darrick J. Wong3caa4872012-05-27 08:10:22 -0400385
Darrick J. Wongdb9ee222014-08-27 18:40:07 -0400386 if (jbd2_journal_has_csum_v2or3(journal))
Darrick J. Wong3caa4872012-05-27 08:10:22 -0400387 csum_size = sizeof(struct jbd2_journal_block_tail);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700388
389 /*
390 * First job: lock down the current transaction and wait for
391 * all outstanding updates to complete.
392 */
393
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700394 /* Do we need to erase the effects of a prior jbd2_journal_flush? */
395 if (journal->j_flags & JBD2_FLUSHED) {
Dave Kleikamp470decc2006-10-11 01:20:57 -0700396 jbd_debug(3, "super block updated\n");
Tejun Heo6fa7aa52016-10-28 12:58:12 -0400397 mutex_lock_io(&journal->j_checkpoint_mutex);
Jan Kara79feb522012-03-13 22:22:54 -0400398 /*
399 * We hold j_checkpoint_mutex so tail cannot change under us.
400 * We don't need any special data guarantees for writing sb
401 * since journal is empty and it is ok for write to be
402 * flushed only with transaction commit.
403 */
404 jbd2_journal_update_sb_log_tail(journal,
405 journal->j_tail_sequence,
406 journal->j_tail,
Christoph Hellwig70fd7612016-11-01 07:40:10 -0600407 REQ_SYNC);
Jan Karaa78bb112012-03-13 15:43:04 -0400408 mutex_unlock(&journal->j_checkpoint_mutex);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700409 } else {
410 jbd_debug(3, "superblock not updated\n");
411 }
412
413 J_ASSERT(journal->j_running_transaction != NULL);
414 J_ASSERT(journal->j_committing_transaction == NULL);
415
416 commit_transaction = journal->j_running_transaction;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700417
Theodore Ts'o879c5e62009-06-17 11:47:48 -0400418 trace_jbd2_start_commit(journal, commit_transaction);
Eryu Guanf2a44522011-11-01 19:09:18 -0400419 jbd_debug(1, "JBD2: starting commit of transaction %d\n",
Dave Kleikamp470decc2006-10-11 01:20:57 -0700420 commit_transaction->t_tid);
421
Theodore Ts'oa931da62010-08-03 21:35:12 -0400422 write_lock(&journal->j_state_lock);
Paul Gortmaker3ca841c2013-06-12 22:46:35 -0400423 J_ASSERT(commit_transaction->t_state == T_RUNNING);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700424 commit_transaction->t_state = T_LOCKED;
425
Theodore Ts'o879c5e62009-06-17 11:47:48 -0400426 trace_jbd2_commit_locking(journal, commit_transaction);
Theodore Ts'obf699322009-09-30 00:32:06 -0400427 stats.run.rs_wait = commit_transaction->t_max_wait;
Theodore Ts'o9fff24a2013-02-06 22:30:23 -0500428 stats.run.rs_request_delay = 0;
Theodore Ts'obf699322009-09-30 00:32:06 -0400429 stats.run.rs_locked = jiffies;
Theodore Ts'o9fff24a2013-02-06 22:30:23 -0500430 if (commit_transaction->t_requested)
431 stats.run.rs_request_delay =
432 jbd2_time_diff(commit_transaction->t_requested,
433 stats.run.rs_locked);
Theodore Ts'obf699322009-09-30 00:32:06 -0400434 stats.run.rs_running = jbd2_time_diff(commit_transaction->t_start,
435 stats.run.rs_locked);
Johann Lombardi8e85fb32008-01-28 23:58:27 -0500436
Dave Kleikamp470decc2006-10-11 01:20:57 -0700437 spin_lock(&commit_transaction->t_handle_lock);
Theodore Ts'oa51dca92010-08-02 08:43:25 -0400438 while (atomic_read(&commit_transaction->t_updates)) {
Dave Kleikamp470decc2006-10-11 01:20:57 -0700439 DEFINE_WAIT(wait);
440
441 prepare_to_wait(&journal->j_wait_updates, &wait,
442 TASK_UNINTERRUPTIBLE);
Theodore Ts'oa51dca92010-08-02 08:43:25 -0400443 if (atomic_read(&commit_transaction->t_updates)) {
Dave Kleikamp470decc2006-10-11 01:20:57 -0700444 spin_unlock(&commit_transaction->t_handle_lock);
Theodore Ts'oa931da62010-08-03 21:35:12 -0400445 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700446 schedule();
Theodore Ts'oa931da62010-08-03 21:35:12 -0400447 write_lock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700448 spin_lock(&commit_transaction->t_handle_lock);
449 }
450 finish_wait(&journal->j_wait_updates, &wait);
451 }
452 spin_unlock(&commit_transaction->t_handle_lock);
Jan Kara96f1e092018-12-03 23:16:07 -0500453 commit_transaction->t_state = T_SWITCH;
454 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700455
Theodore Ts'oa51dca92010-08-02 08:43:25 -0400456 J_ASSERT (atomic_read(&commit_transaction->t_outstanding_credits) <=
Dave Kleikamp470decc2006-10-11 01:20:57 -0700457 journal->j_max_transaction_buffers);
458
459 /*
460 * First thing we are allowed to do is to discard any remaining
461 * BJ_Reserved buffers. Note, it is _not_ permissible to assume
462 * that there are no such buffers: if a large filesystem
463 * operation like a truncate needs to split itself over multiple
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700464 * transactions, then it may try to do a jbd2_journal_restart() while
Dave Kleikamp470decc2006-10-11 01:20:57 -0700465 * there are still BJ_Reserved buffers outstanding. These must
466 * be released cleanly from the current transaction.
467 *
468 * In this case, the filesystem must still reserve write access
469 * again before modifying the buffer in the new transaction, but
470 * we do not require it to remember exactly which old buffers it
471 * has reserved. This is consistent with the existing behaviour
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700472 * that multiple jbd2_journal_get_write_access() calls to the same
Lucas De Marchi25985ed2011-03-30 22:57:33 -0300473 * buffer are perfectly permissible.
Dave Kleikamp470decc2006-10-11 01:20:57 -0700474 */
475 while (commit_transaction->t_reserved_list) {
476 jh = commit_transaction->t_reserved_list;
477 JBUFFER_TRACE(jh, "reserved, unused: refile");
478 /*
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700479 * A jbd2_journal_get_undo_access()+jbd2_journal_release_buffer() may
Dave Kleikamp470decc2006-10-11 01:20:57 -0700480 * leave undo-committed data.
481 */
482 if (jh->b_committed_data) {
483 struct buffer_head *bh = jh2bh(jh);
484
Thomas Gleixner46417062019-08-09 14:42:32 +0200485 spin_lock(&jh->b_state_lock);
Mingming Caoaf1e76d2007-10-16 18:38:25 -0400486 jbd2_free(jh->b_committed_data, bh->b_size);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700487 jh->b_committed_data = NULL;
Thomas Gleixner46417062019-08-09 14:42:32 +0200488 spin_unlock(&jh->b_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700489 }
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700490 jbd2_journal_refile_buffer(journal, jh);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700491 }
492
493 /*
494 * Now try to drop any written-back buffers from the journal's
495 * checkpoint lists. We do this *before* commit because it potentially
496 * frees some memory
497 */
498 spin_lock(&journal->j_list_lock);
Jan Kara841df7df2015-07-28 14:57:14 -0400499 __jbd2_journal_clean_checkpoint_list(journal, false);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700500 spin_unlock(&journal->j_list_lock);
501
Eryu Guanf2a44522011-11-01 19:09:18 -0400502 jbd_debug(3, "JBD2: commit phase 1\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700503
504 /*
Yongqiang Yang1ba37262011-12-28 17:46:46 -0500505 * Clear revoked flag to reflect there is no revoked buffers
506 * in the next transaction which is going to be started.
507 */
508 jbd2_clear_buffer_revoked_flags(journal);
509
510 /*
Dave Kleikamp470decc2006-10-11 01:20:57 -0700511 * Switch to a new revoke table.
512 */
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700513 jbd2_journal_switch_revoke_table(journal);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700514
Jan Kara8f7d89f2013-06-04 12:35:11 -0400515 /*
516 * Reserved credits cannot be claimed anymore, free them
517 */
518 atomic_sub(atomic_read(&journal->j_reserved_credits),
519 &commit_transaction->t_outstanding_credits);
520
Jan Kara96f1e092018-12-03 23:16:07 -0500521 write_lock(&journal->j_state_lock);
Theodore Ts'o879c5e62009-06-17 11:47:48 -0400522 trace_jbd2_commit_flushing(journal, commit_transaction);
Theodore Ts'obf699322009-09-30 00:32:06 -0400523 stats.run.rs_flushing = jiffies;
524 stats.run.rs_locked = jbd2_time_diff(stats.run.rs_locked,
525 stats.run.rs_flushing);
Johann Lombardi8e85fb32008-01-28 23:58:27 -0500526
Dave Kleikamp470decc2006-10-11 01:20:57 -0700527 commit_transaction->t_state = T_FLUSH;
528 journal->j_committing_transaction = commit_transaction;
529 journal->j_running_transaction = NULL;
Josef Bacike07f7182008-11-26 01:14:26 -0500530 start_time = ktime_get();
Dave Kleikamp470decc2006-10-11 01:20:57 -0700531 commit_transaction->t_log_start = journal->j_head;
532 wake_up(&journal->j_wait_transaction_locked);
Theodore Ts'oa931da62010-08-03 21:35:12 -0400533 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700534
Paul Gortmakercfc7bc82013-06-12 22:56:35 -0400535 jbd_debug(3, "JBD2: commit phase 2a\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700536
537 /*
Dave Kleikamp470decc2006-10-11 01:20:57 -0700538 * Now start flushing things to disk, in the order they appear
539 * on the transaction lists. Data blocks go first.
540 */
Aneesh Kumar K.Vcd1aac32008-07-11 19:27:31 -0400541 err = journal_submit_data_buffers(journal, commit_transaction);
Jan Karac851ed52008-07-11 19:27:31 -0400542 if (err)
543 jbd2_journal_abort(journal, err);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700544
Jens Axboe82f04ab2011-03-17 11:01:52 +0100545 blk_start_plug(&plug);
Jan Kara9bcf9762016-02-22 23:07:30 -0500546 jbd2_journal_write_revoke_records(commit_transaction, &log_bufs);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700547
Paul Gortmakercfc7bc82013-06-12 22:56:35 -0400548 jbd_debug(3, "JBD2: commit phase 2b\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700549
550 /*
Dave Kleikamp470decc2006-10-11 01:20:57 -0700551 * Way to go: we have now written out all of the data for a
552 * transaction! Now comes the tricky part: we need to write out
553 * metadata. Loop over the transaction's entire buffer list:
554 */
Theodore Ts'oa931da62010-08-03 21:35:12 -0400555 write_lock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700556 commit_transaction->t_state = T_COMMIT;
Theodore Ts'oa931da62010-08-03 21:35:12 -0400557 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700558
Theodore Ts'o879c5e62009-06-17 11:47:48 -0400559 trace_jbd2_commit_logging(journal, commit_transaction);
Theodore Ts'obf699322009-09-30 00:32:06 -0400560 stats.run.rs_logging = jiffies;
561 stats.run.rs_flushing = jbd2_time_diff(stats.run.rs_flushing,
562 stats.run.rs_logging);
Jan Kara9f356e52019-11-05 17:44:24 +0100563 stats.run.rs_blocks = commit_transaction->t_nr_buffers;
Theodore Ts'obf699322009-09-30 00:32:06 -0400564 stats.run.rs_blocks_logged = 0;
Johann Lombardi8e85fb32008-01-28 23:58:27 -0500565
Josef Bacik1dfc3222008-04-17 10:38:59 -0400566 J_ASSERT(commit_transaction->t_nr_buffers <=
Theodore Ts'oa51dca92010-08-02 08:43:25 -0400567 atomic_read(&commit_transaction->t_outstanding_credits));
Josef Bacik1dfc3222008-04-17 10:38:59 -0400568
Jan Kara87c89c22008-07-11 19:27:31 -0400569 err = 0;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700570 bufs = 0;
Jan Karae5a120a2013-06-04 12:06:01 -0400571 descriptor = NULL;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700572 while (commit_transaction->t_buffers) {
573
574 /* Find the next buffer to be journaled... */
575
576 jh = commit_transaction->t_buffers;
577
578 /* If we're in abort mode, we just un-journal the buffer and
Hidehiro Kawai7ad74452008-10-10 20:29:31 -0400579 release it. */
Dave Kleikamp470decc2006-10-11 01:20:57 -0700580
581 if (is_journal_aborted(journal)) {
Hidehiro Kawai7ad74452008-10-10 20:29:31 -0400582 clear_buffer_jbddirty(jh2bh(jh));
Dave Kleikamp470decc2006-10-11 01:20:57 -0700583 JBUFFER_TRACE(jh, "journal is aborting: refile");
Joel Beckere06c8222008-09-11 15:35:47 -0700584 jbd2_buffer_abort_trigger(jh,
585 jh->b_frozen_data ?
586 jh->b_frozen_triggers :
587 jh->b_triggers);
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700588 jbd2_journal_refile_buffer(journal, jh);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700589 /* If that was the last one, we need to clean up
590 * any descriptor buffers which may have been
591 * already allocated, even if we are now
592 * aborting. */
593 if (!commit_transaction->t_buffers)
594 goto start_journal_io;
595 continue;
596 }
597
598 /* Make sure we have a descriptor block in which to
599 record the metadata buffer. */
600
601 if (!descriptor) {
Dave Kleikamp470decc2006-10-11 01:20:57 -0700602 J_ASSERT (bufs == 0);
603
Eryu Guanf2a44522011-11-01 19:09:18 -0400604 jbd_debug(4, "JBD2: get descriptor\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700605
Jan Kara32ab6712016-02-22 23:17:15 -0500606 descriptor = jbd2_journal_get_descriptor_buffer(
607 commit_transaction,
608 JBD2_DESCRIPTOR_BLOCK);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700609 if (!descriptor) {
Jan Karaa7fa2ba2007-10-16 18:38:25 -0400610 jbd2_journal_abort(journal, -EIO);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700611 continue;
612 }
613
Eryu Guanf2a44522011-11-01 19:09:18 -0400614 jbd_debug(4, "JBD2: got buffer %llu (%p)\n",
Jan Karae5a120a2013-06-04 12:06:01 -0400615 (unsigned long long)descriptor->b_blocknr,
616 descriptor->b_data);
Jan Karae5a120a2013-06-04 12:06:01 -0400617 tagp = &descriptor->b_data[sizeof(journal_header_t)];
618 space_left = descriptor->b_size -
619 sizeof(journal_header_t);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700620 first_tag = 1;
Jan Karae5a120a2013-06-04 12:06:01 -0400621 set_buffer_jwrite(descriptor);
622 set_buffer_dirty(descriptor);
623 wbuf[bufs++] = descriptor;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700624
625 /* Record it so that we can wait for IO
626 completion later */
Jan Karae5a120a2013-06-04 12:06:01 -0400627 BUFFER_TRACE(descriptor, "ph3: file as descriptor");
628 jbd2_file_log_bh(&log_bufs, descriptor);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700629 }
630
631 /* Where is the buffer to be written? */
632
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700633 err = jbd2_journal_next_log_block(journal, &blocknr);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700634 /* If the block mapping failed, just abandon the buffer
635 and repeat this loop: we'll fall into the
636 refile-on-abort condition above. */
637 if (err) {
Jan Karaa7fa2ba2007-10-16 18:38:25 -0400638 jbd2_journal_abort(journal, err);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700639 continue;
640 }
641
642 /*
643 * start_this_handle() uses t_outstanding_credits to determine
Jan Kara0db45882019-11-05 17:44:08 +0100644 * the free space in the log.
Dave Kleikamp470decc2006-10-11 01:20:57 -0700645 */
Theodore Ts'oa51dca92010-08-02 08:43:25 -0400646 atomic_dec(&commit_transaction->t_outstanding_credits);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700647
648 /* Bump b_count to prevent truncate from stumbling over
649 the shadowed buffer! @@@ This can go if we ever get
Jan Karaf5113ef2013-06-04 12:01:45 -0400650 rid of the shadow pairing of buffers. */
Dave Kleikamp470decc2006-10-11 01:20:57 -0700651 atomic_inc(&jh2bh(jh)->b_count);
652
Dave Kleikamp470decc2006-10-11 01:20:57 -0700653 /*
Jan Karaf5113ef2013-06-04 12:01:45 -0400654 * Make a temporary IO buffer with which to write it out
655 * (this will requeue the metadata buffer to BJ_Shadow).
Dave Kleikamp470decc2006-10-11 01:20:57 -0700656 */
Jan Karaf5113ef2013-06-04 12:01:45 -0400657 set_bit(BH_JWrite, &jh2bh(jh)->b_state);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700658 JBUFFER_TRACE(jh, "ph3: write metadata");
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700659 flags = jbd2_journal_write_metadata_buffer(commit_transaction,
Jan Karaf5113ef2013-06-04 12:01:45 -0400660 jh, &wbuf[bufs], blocknr);
Theodore Ts'oe6ec1162009-12-01 09:04:42 -0500661 if (flags < 0) {
662 jbd2_journal_abort(journal, flags);
663 continue;
664 }
Jan Karaf5113ef2013-06-04 12:01:45 -0400665 jbd2_file_log_bh(&io_bufs, wbuf[bufs]);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700666
667 /* Record the new block's tag in the current descriptor
668 buffer */
669
670 tag_flag = 0;
671 if (flags & 1)
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700672 tag_flag |= JBD2_FLAG_ESCAPE;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700673 if (!first_tag)
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700674 tag_flag |= JBD2_FLAG_SAME_UUID;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700675
676 tag = (journal_block_tag_t *) tagp;
Darrick J. Wongdb9ee222014-08-27 18:40:07 -0400677 write_tag_block(journal, tag, jh2bh(jh)->b_blocknr);
Darrick J. Wong8f888ef2012-05-22 22:43:41 -0400678 tag->t_flags = cpu_to_be16(tag_flag);
Jan Karaf5113ef2013-06-04 12:01:45 -0400679 jbd2_block_tag_csum_set(journal, tag, wbuf[bufs],
Darrick J. Wongc3900872012-05-27 08:12:12 -0400680 commit_transaction->t_tid);
Zach Brownb517bea2006-10-11 01:21:08 -0700681 tagp += tag_bytes;
682 space_left -= tag_bytes;
Jan Karaf5113ef2013-06-04 12:01:45 -0400683 bufs++;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700684
685 if (first_tag) {
686 memcpy (tagp, journal->j_uuid, 16);
687 tagp += 16;
688 space_left -= 16;
689 first_tag = 0;
690 }
691
692 /* If there's no more to do, or if the descriptor is full,
693 let the IO rip! */
694
695 if (bufs == journal->j_wbufsize ||
696 commit_transaction->t_buffers == NULL ||
Darrick J. Wong3caa4872012-05-27 08:10:22 -0400697 space_left < tag_bytes + 16 + csum_size) {
Dave Kleikamp470decc2006-10-11 01:20:57 -0700698
Eryu Guanf2a44522011-11-01 19:09:18 -0400699 jbd_debug(4, "JBD2: Submit %d IOs\n", bufs);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700700
701 /* Write an end-of-descriptor marker before
702 submitting the IOs. "tag" still points to
703 the last tag we set up. */
704
Darrick J. Wong8f888ef2012-05-22 22:43:41 -0400705 tag->t_flags |= cpu_to_be16(JBD2_FLAG_LAST_TAG);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700706start_journal_io:
luojiajun6e876c32019-03-01 00:30:00 -0500707 if (descriptor)
708 jbd2_descriptor_block_csum_set(journal,
709 descriptor);
710
Dave Kleikamp470decc2006-10-11 01:20:57 -0700711 for (i = 0; i < bufs; i++) {
712 struct buffer_head *bh = wbuf[i];
Girish Shilamkar818d2762008-01-28 23:58:27 -0500713 /*
714 * Compute checksum.
715 */
Darrick J. Wong56316a02015-10-17 16:18:45 -0400716 if (jbd2_has_feature_checksum(journal)) {
Girish Shilamkar818d2762008-01-28 23:58:27 -0500717 crc32_sum =
718 jbd2_checksum_data(crc32_sum, bh);
719 }
720
Dave Kleikamp470decc2006-10-11 01:20:57 -0700721 lock_buffer(bh);
722 clear_buffer_dirty(bh);
723 set_buffer_uptodate(bh);
724 bh->b_end_io = journal_end_buffer_io_sync;
Christoph Hellwig70fd7612016-11-01 07:40:10 -0600725 submit_bh(REQ_OP_WRITE, REQ_SYNC, bh);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700726 }
727 cond_resched();
728
729 /* Force a new descriptor to be generated next
730 time round the loop. */
731 descriptor = NULL;
732 bufs = 0;
733 }
734 }
735
Jan Karac851ed52008-07-11 19:27:31 -0400736 err = journal_finish_inode_data_buffers(journal, commit_transaction);
Hidehiro Kawaie9e34f42008-07-31 22:26:04 -0400737 if (err) {
Hidehiro Kawaie9e34f42008-07-31 22:26:04 -0400738 printk(KERN_WARNING
739 "JBD2: Detected IO errors while flushing file data "
Theodore Ts'o05496762008-09-16 14:36:17 -0400740 "on %s\n", journal->j_devname);
Hidehiro Kawai5bf56832008-10-10 22:12:43 -0400741 if (journal->j_flags & JBD2_ABORT_ON_SYNCDATA_ERR)
742 jbd2_journal_abort(journal, err);
Hidehiro Kawaie9e34f42008-07-31 22:26:04 -0400743 err = 0;
744 }
Jan Karac851ed52008-07-11 19:27:31 -0400745
Jan Kara33395782012-03-13 22:45:38 -0400746 /*
747 * Get current oldest transaction in the log before we issue flush
748 * to the filesystem device. After the flush we can be sure that
749 * blocks of all older transactions are checkpointed to persistent
750 * storage and we will be safe to update journal start in the
751 * superblock with the numbers we get here.
752 */
753 update_tail =
754 jbd2_journal_get_log_tail(journal, &first_tid, &first_block);
755
Jan Karabbd2be32011-05-24 11:59:18 -0400756 write_lock(&journal->j_state_lock);
Jan Kara33395782012-03-13 22:45:38 -0400757 if (update_tail) {
758 long freed = first_block - journal->j_tail;
759
760 if (first_block < journal->j_tail)
761 freed += journal->j_last - journal->j_first;
762 /* Update tail only if we free significant amount of space */
763 if (freed < journal->j_maxlen / 4)
764 update_tail = 0;
765 }
Jan Karabbd2be32011-05-24 11:59:18 -0400766 J_ASSERT(commit_transaction->t_state == T_COMMIT);
767 commit_transaction->t_state = T_COMMIT_DFLUSH;
768 write_unlock(&journal->j_state_lock);
Jan Kara33395782012-03-13 22:45:38 -0400769
Girish Shilamkar818d2762008-01-28 23:58:27 -0500770 /*
771 * If the journal is not located on the file system device,
772 * then we must flush the file system device before we issue
773 * the commit record
774 */
Jan Kara81be12c2011-05-24 11:52:40 -0400775 if (commit_transaction->t_need_data_flush &&
Girish Shilamkar818d2762008-01-28 23:58:27 -0500776 (journal->j_fs_dev != journal->j_dev) &&
777 (journal->j_flags & JBD2_BARRIER))
Shaohua Li99aa7842012-04-13 10:27:35 +0800778 blkdev_issue_flush(journal->j_fs_dev, GFP_NOFS, NULL);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500779
780 /* Done it all: now write the commit record asynchronously. */
Darrick J. Wong56316a02015-10-17 16:18:45 -0400781 if (jbd2_has_feature_async_commit(journal)) {
Girish Shilamkar818d2762008-01-28 23:58:27 -0500782 err = journal_submit_commit_record(journal, commit_transaction,
Dave Kleikamp470decc2006-10-11 01:20:57 -0700783 &cbh, crc32_sum);
784 if (err)
zhangyi (F)d0a186e2019-12-04 20:46:11 +0800785 jbd2_journal_abort(journal, err);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700786 }
787
Jens Axboe82f04ab2011-03-17 11:01:52 +0100788 blk_finish_plug(&plug);
789
Dave Kleikamp470decc2006-10-11 01:20:57 -0700790 /* Lo and behold: we have just managed to send a transaction to
791 the log. Before we can commit it, wait for the IO so far to
792 complete. Control buffers being written are on the
793 transaction's t_log_list queue, and metadata buffers are on
Jan Karaf5113ef2013-06-04 12:01:45 -0400794 the io_bufs list.
Dave Kleikamp470decc2006-10-11 01:20:57 -0700795
796 Wait for the buffers in reverse order. That way we are
797 less likely to be woken up until all IOs have completed, and
798 so we incur less scheduling load.
799 */
800
Eryu Guanf2a44522011-11-01 19:09:18 -0400801 jbd_debug(3, "JBD2: commit phase 3\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700802
Jan Karaf5113ef2013-06-04 12:01:45 -0400803 while (!list_empty(&io_bufs)) {
804 struct buffer_head *bh = list_entry(io_bufs.prev,
805 struct buffer_head,
806 b_assoc_buffers);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700807
Jan Karaf5113ef2013-06-04 12:01:45 -0400808 wait_on_buffer(bh);
809 cond_resched();
Dave Kleikamp470decc2006-10-11 01:20:57 -0700810
811 if (unlikely(!buffer_uptodate(bh)))
812 err = -EIO;
Jan Karaf5113ef2013-06-04 12:01:45 -0400813 jbd2_unfile_log_bh(bh);
Jan Kara015c6032019-11-05 17:44:19 +0100814 stats.run.rs_blocks_logged++;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700815
816 /*
Jan Karaf5113ef2013-06-04 12:01:45 -0400817 * The list contains temporary buffer heads created by
818 * jbd2_journal_write_metadata_buffer().
Dave Kleikamp470decc2006-10-11 01:20:57 -0700819 */
820 BUFFER_TRACE(bh, "dumping temporary bh");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700821 __brelse(bh);
822 J_ASSERT_BH(bh, atomic_read(&bh->b_count) == 0);
823 free_buffer_head(bh);
824
Jan Karaf5113ef2013-06-04 12:01:45 -0400825 /* We also have to refile the corresponding shadowed buffer */
Dave Kleikamp470decc2006-10-11 01:20:57 -0700826 jh = commit_transaction->t_shadow_list->b_tprev;
827 bh = jh2bh(jh);
Jan Karaf5113ef2013-06-04 12:01:45 -0400828 clear_buffer_jwrite(bh);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700829 J_ASSERT_BH(bh, buffer_jbddirty(bh));
Jan Karab34090e2013-06-04 12:08:56 -0400830 J_ASSERT_BH(bh, !buffer_shadow(bh));
Dave Kleikamp470decc2006-10-11 01:20:57 -0700831
832 /* The metadata is now released for reuse, but we need
833 to remember it against this transaction so that when
834 we finally commit, we can do any checkpointing
835 required. */
836 JBUFFER_TRACE(jh, "file as BJ_Forget");
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700837 jbd2_journal_file_buffer(jh, commit_transaction, BJ_Forget);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700838 JBUFFER_TRACE(jh, "brelse shadowed buffer");
839 __brelse(bh);
840 }
841
842 J_ASSERT (commit_transaction->t_shadow_list == NULL);
843
Eryu Guanf2a44522011-11-01 19:09:18 -0400844 jbd_debug(3, "JBD2: commit phase 4\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700845
846 /* Here we wait for the revoke record and descriptor record buffers */
Jan Karae5a120a2013-06-04 12:06:01 -0400847 while (!list_empty(&log_bufs)) {
Dave Kleikamp470decc2006-10-11 01:20:57 -0700848 struct buffer_head *bh;
849
Jan Karae5a120a2013-06-04 12:06:01 -0400850 bh = list_entry(log_bufs.prev, struct buffer_head, b_assoc_buffers);
851 wait_on_buffer(bh);
852 cond_resched();
Dave Kleikamp470decc2006-10-11 01:20:57 -0700853
854 if (unlikely(!buffer_uptodate(bh)))
855 err = -EIO;
856
857 BUFFER_TRACE(bh, "ph5: control buffer writeout done: unfile");
858 clear_buffer_jwrite(bh);
Jan Karae5a120a2013-06-04 12:06:01 -0400859 jbd2_unfile_log_bh(bh);
Jan Kara015c6032019-11-05 17:44:19 +0100860 stats.run.rs_blocks_logged++;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700861 __brelse(bh); /* One for getblk */
862 /* AKPM: bforget here */
863 }
864
Hidehiro Kawai77e841d2008-10-12 16:39:16 -0400865 if (err)
866 jbd2_journal_abort(journal, err);
867
Eryu Guanf2a44522011-11-01 19:09:18 -0400868 jbd_debug(3, "JBD2: commit phase 5\n");
Jan Karabbd2be32011-05-24 11:59:18 -0400869 write_lock(&journal->j_state_lock);
870 J_ASSERT(commit_transaction->t_state == T_COMMIT_DFLUSH);
871 commit_transaction->t_state = T_COMMIT_JFLUSH;
872 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700873
Darrick J. Wong56316a02015-10-17 16:18:45 -0400874 if (!jbd2_has_feature_async_commit(journal)) {
Girish Shilamkar818d2762008-01-28 23:58:27 -0500875 err = journal_submit_commit_record(journal, commit_transaction,
876 &cbh, crc32_sum);
877 if (err)
zhangyi (F)d0a186e2019-12-04 20:46:11 +0800878 jbd2_journal_abort(journal, err);
Girish Shilamkar818d2762008-01-28 23:58:27 -0500879 }
Zhang Huan6cba6112011-04-05 19:16:20 -0400880 if (cbh)
Theodore Ts'ofd984962009-01-05 21:34:13 -0500881 err = journal_wait_on_commit_record(journal, cbh);
Jan Kara015c6032019-11-05 17:44:19 +0100882 stats.run.rs_blocks_logged++;
Darrick J. Wong56316a02015-10-17 16:18:45 -0400883 if (jbd2_has_feature_async_commit(journal) &&
Jan Karaf73bee42010-08-18 15:56:56 +0200884 journal->j_flags & JBD2_BARRIER) {
Shaohua Li99aa7842012-04-13 10:27:35 +0800885 blkdev_issue_flush(journal->j_dev, GFP_NOFS, NULL);
Jan Karaf73bee42010-08-18 15:56:56 +0200886 }
Dave Kleikamp470decc2006-10-11 01:20:57 -0700887
888 if (err)
Jan Karaa7fa2ba2007-10-16 18:38:25 -0400889 jbd2_journal_abort(journal, err);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700890
Jan Kara9f356e52019-11-05 17:44:24 +0100891 WARN_ON_ONCE(
892 atomic_read(&commit_transaction->t_outstanding_credits) < 0);
893
Jan Kara33395782012-03-13 22:45:38 -0400894 /*
895 * Now disk caches for filesystem device are flushed so we are safe to
896 * erase checkpointed transactions from the log by updating journal
897 * superblock.
898 */
899 if (update_tail)
900 jbd2_update_log_tail(journal, first_tid, first_block);
901
Dave Kleikamp470decc2006-10-11 01:20:57 -0700902 /* End of a transaction! Finally, we can do checkpoint
903 processing: any buffers committed as a result of this
904 transaction can be removed from any checkpoint list it was on
905 before. */
906
Eryu Guanf2a44522011-11-01 19:09:18 -0400907 jbd_debug(3, "JBD2: commit phase 6\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -0700908
Jan Karac851ed52008-07-11 19:27:31 -0400909 J_ASSERT(list_empty(&commit_transaction->t_inode_list));
Dave Kleikamp470decc2006-10-11 01:20:57 -0700910 J_ASSERT(commit_transaction->t_buffers == NULL);
911 J_ASSERT(commit_transaction->t_checkpoint_list == NULL);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700912 J_ASSERT(commit_transaction->t_shadow_list == NULL);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700913
914restart_loop:
915 /*
916 * As there are other places (journal_unmap_buffer()) adding buffers
917 * to this list we have to be careful and hold the j_list_lock.
918 */
919 spin_lock(&journal->j_list_lock);
920 while (commit_transaction->t_forget) {
921 transaction_t *cp_transaction;
922 struct buffer_head *bh;
Jan Karade1b7942011-06-13 15:38:22 -0400923 int try_to_free = 0;
Jan Kara93108eb2019-08-09 14:42:29 +0200924 bool drop_ref;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700925
926 jh = commit_transaction->t_forget;
927 spin_unlock(&journal->j_list_lock);
928 bh = jh2bh(jh);
Jan Karade1b7942011-06-13 15:38:22 -0400929 /*
930 * Get a reference so that bh cannot be freed before we are
931 * done with it.
932 */
933 get_bh(bh);
Thomas Gleixner46417062019-08-09 14:42:32 +0200934 spin_lock(&jh->b_state_lock);
dingdinghua23e2af32010-02-24 12:11:20 -0500935 J_ASSERT_JH(jh, jh->b_transaction == commit_transaction);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700936
937 /*
938 * If there is undo-protected committed data against
939 * this buffer, then we can remove it now. If it is a
940 * buffer needing such protection, the old frozen_data
941 * field now points to a committed version of the
942 * buffer, so rotate that field to the new committed
943 * data.
944 *
945 * Otherwise, we can just throw away the frozen data now.
Joel Beckere06c8222008-09-11 15:35:47 -0700946 *
947 * We also know that the frozen data has already fired
948 * its triggers if they exist, so we can clear that too.
Dave Kleikamp470decc2006-10-11 01:20:57 -0700949 */
950 if (jh->b_committed_data) {
Mingming Caoaf1e76d2007-10-16 18:38:25 -0400951 jbd2_free(jh->b_committed_data, bh->b_size);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700952 jh->b_committed_data = NULL;
953 if (jh->b_frozen_data) {
954 jh->b_committed_data = jh->b_frozen_data;
955 jh->b_frozen_data = NULL;
Joel Beckere06c8222008-09-11 15:35:47 -0700956 jh->b_frozen_triggers = NULL;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700957 }
958 } else if (jh->b_frozen_data) {
Mingming Caoaf1e76d2007-10-16 18:38:25 -0400959 jbd2_free(jh->b_frozen_data, bh->b_size);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700960 jh->b_frozen_data = NULL;
Joel Beckere06c8222008-09-11 15:35:47 -0700961 jh->b_frozen_triggers = NULL;
Dave Kleikamp470decc2006-10-11 01:20:57 -0700962 }
963
964 spin_lock(&journal->j_list_lock);
965 cp_transaction = jh->b_cp_transaction;
966 if (cp_transaction) {
967 JBUFFER_TRACE(jh, "remove from old cp transaction");
Johann Lombardi8e85fb32008-01-28 23:58:27 -0500968 cp_transaction->t_chp_stats.cs_dropped++;
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700969 __jbd2_journal_remove_checkpoint(jh);
Dave Kleikamp470decc2006-10-11 01:20:57 -0700970 }
971
972 /* Only re-checkpoint the buffer_head if it is marked
973 * dirty. If the buffer was added to the BJ_Forget list
Mingming Caof7f4bcc2006-10-11 01:20:59 -0700974 * by jbd2_journal_forget, it may no longer be dirty and
Dave Kleikamp470decc2006-10-11 01:20:57 -0700975 * there's no point in keeping a checkpoint record for
976 * it. */
977
Jan Karab794e7a2012-09-26 23:11:13 -0400978 /*
zhangyi (F)6a66a7d2020-02-13 14:38:20 +0800979 * A buffer which has been freed while still being journaled
980 * by a previous transaction, refile the buffer to BJ_Forget of
981 * the running transaction. If the just committed transaction
982 * contains "add to orphan" operation, we can completely
983 * invalidate the buffer now. We are rather through in that
984 * since the buffer may be still accessible when blocksize <
985 * pagesize and it is attached to the last partial page.
986 */
987 if (buffer_freed(bh) && !jh->b_next_transaction) {
zhangyi (F)c96dcee2020-02-13 14:38:21 +0800988 struct address_space *mapping;
989
zhangyi (F)6a66a7d2020-02-13 14:38:20 +0800990 clear_buffer_freed(bh);
991 clear_buffer_jbddirty(bh);
zhangyi (F)c96dcee2020-02-13 14:38:21 +0800992
993 /*
994 * Block device buffers need to stay mapped all the
995 * time, so it is enough to clear buffer_jbddirty and
996 * buffer_freed bits. For the file mapping buffers (i.e.
997 * journalled data) we need to unmap buffer and clear
998 * more bits. We also need to be careful about the check
999 * because the data page mapping can get cleared under
zhangyi (F)780f66e2020-02-17 19:27:06 +08001000 * our hands. Note that if mapping == NULL, we don't
1001 * need to make buffer unmapped because the page is
1002 * already detached from the mapping and buffers cannot
1003 * get reused.
zhangyi (F)c96dcee2020-02-13 14:38:21 +08001004 */
1005 mapping = READ_ONCE(bh->b_page->mapping);
1006 if (mapping && !sb_is_blkdev_sb(mapping->host->i_sb)) {
1007 clear_buffer_mapped(bh);
1008 clear_buffer_new(bh);
1009 clear_buffer_req(bh);
1010 bh->b_bdev = NULL;
1011 }
Dave Kleikamp470decc2006-10-11 01:20:57 -07001012 }
1013
1014 if (buffer_jbddirty(bh)) {
1015 JBUFFER_TRACE(jh, "add to new checkpointing trans");
Mingming Caof7f4bcc2006-10-11 01:20:59 -07001016 __jbd2_journal_insert_checkpoint(jh, commit_transaction);
Hidehiro Kawai7ad74452008-10-10 20:29:31 -04001017 if (is_journal_aborted(journal))
1018 clear_buffer_jbddirty(bh);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001019 } else {
1020 J_ASSERT_BH(bh, !buffer_dirty(bh));
Jan Karade1b7942011-06-13 15:38:22 -04001021 /*
1022 * The buffer on BJ_Forget list and not jbddirty means
Dave Kleikamp470decc2006-10-11 01:20:57 -07001023 * it has been freed by this transaction and hence it
1024 * could not have been reallocated until this
1025 * transaction has committed. *BUT* it could be
1026 * reallocated once we have written all the data to
1027 * disk and before we process the buffer on BJ_Forget
Jan Karade1b7942011-06-13 15:38:22 -04001028 * list.
1029 */
1030 if (!jh->b_next_transaction)
1031 try_to_free = 1;
Dave Kleikamp470decc2006-10-11 01:20:57 -07001032 }
Jan Karade1b7942011-06-13 15:38:22 -04001033 JBUFFER_TRACE(jh, "refile or unfile buffer");
Jan Kara93108eb2019-08-09 14:42:29 +02001034 drop_ref = __jbd2_journal_refile_buffer(jh);
Thomas Gleixner46417062019-08-09 14:42:32 +02001035 spin_unlock(&jh->b_state_lock);
Jan Kara93108eb2019-08-09 14:42:29 +02001036 if (drop_ref)
1037 jbd2_journal_put_journal_head(jh);
Jan Karade1b7942011-06-13 15:38:22 -04001038 if (try_to_free)
1039 release_buffer_page(bh); /* Drops bh reference */
1040 else
1041 __brelse(bh);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001042 cond_resched_lock(&journal->j_list_lock);
1043 }
1044 spin_unlock(&journal->j_list_lock);
1045 /*
Jan Karaf5a7a6b2008-01-28 23:58:27 -05001046 * This is a bit sleazy. We use j_list_lock to protect transition
1047 * of a transaction into T_FINISHED state and calling
1048 * __jbd2_journal_drop_transaction(). Otherwise we could race with
1049 * other checkpointing code processing the transaction...
Dave Kleikamp470decc2006-10-11 01:20:57 -07001050 */
Theodore Ts'oa931da62010-08-03 21:35:12 -04001051 write_lock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001052 spin_lock(&journal->j_list_lock);
1053 /*
1054 * Now recheck if some buffers did not get attached to the transaction
1055 * while the lock was dropped...
1056 */
1057 if (commit_transaction->t_forget) {
1058 spin_unlock(&journal->j_list_lock);
Theodore Ts'oa931da62010-08-03 21:35:12 -04001059 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001060 goto restart_loop;
1061 }
1062
Theodore Ts'od4e839d2014-03-08 22:34:10 -05001063 /* Add the transaction to the checkpoint list
1064 * __journal_remove_checkpoint() can not destroy transaction
1065 * under us because it is not marked as T_FINISHED yet */
1066 if (journal->j_checkpoint_transactions == NULL) {
1067 journal->j_checkpoint_transactions = commit_transaction;
1068 commit_transaction->t_cpnext = commit_transaction;
1069 commit_transaction->t_cpprev = commit_transaction;
1070 } else {
1071 commit_transaction->t_cpnext =
1072 journal->j_checkpoint_transactions;
1073 commit_transaction->t_cpprev =
1074 commit_transaction->t_cpnext->t_cpprev;
1075 commit_transaction->t_cpnext->t_cpprev =
1076 commit_transaction;
1077 commit_transaction->t_cpprev->t_cpnext =
1078 commit_transaction;
1079 }
1080 spin_unlock(&journal->j_list_lock);
1081
Dave Kleikamp470decc2006-10-11 01:20:57 -07001082 /* Done with this transaction! */
1083
Eryu Guanf2a44522011-11-01 19:09:18 -04001084 jbd_debug(3, "JBD2: commit phase 7\n");
Dave Kleikamp470decc2006-10-11 01:20:57 -07001085
Jan Karabbd2be32011-05-24 11:59:18 -04001086 J_ASSERT(commit_transaction->t_state == T_COMMIT_JFLUSH);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001087
Johann Lombardi8e85fb32008-01-28 23:58:27 -05001088 commit_transaction->t_start = jiffies;
Theodore Ts'obf699322009-09-30 00:32:06 -04001089 stats.run.rs_logging = jbd2_time_diff(stats.run.rs_logging,
1090 commit_transaction->t_start);
Johann Lombardi8e85fb32008-01-28 23:58:27 -05001091
1092 /*
Theodore Ts'obf699322009-09-30 00:32:06 -04001093 * File the transaction statistics
Johann Lombardi8e85fb32008-01-28 23:58:27 -05001094 */
Johann Lombardi8e85fb32008-01-28 23:58:27 -05001095 stats.ts_tid = commit_transaction->t_tid;
Theodore Ts'o8dd42042010-08-03 21:38:29 -04001096 stats.run.rs_handle_count =
1097 atomic_read(&commit_transaction->t_handle_count);
Theodore Ts'obf699322009-09-30 00:32:06 -04001098 trace_jbd2_run_stats(journal->j_fs_dev->bd_dev,
1099 commit_transaction->t_tid, &stats.run);
Theodore Ts'o42cf3452014-03-08 19:51:16 -05001100 stats.ts_requested = (commit_transaction->t_requested) ? 1 : 0;
Johann Lombardi8e85fb32008-01-28 23:58:27 -05001101
Dmitry Monakhov794446c2013-04-03 22:06:52 -04001102 commit_transaction->t_state = T_COMMIT_CALLBACK;
Dave Kleikamp470decc2006-10-11 01:20:57 -07001103 J_ASSERT(commit_transaction == journal->j_committing_transaction);
1104 journal->j_commit_sequence = commit_transaction->t_tid;
1105 journal->j_committing_transaction = NULL;
Josef Bacike07f7182008-11-26 01:14:26 -05001106 commit_time = ktime_to_ns(ktime_sub(ktime_get(), start_time));
Dave Kleikamp470decc2006-10-11 01:20:57 -07001107
Josef Bacike07f7182008-11-26 01:14:26 -05001108 /*
1109 * weight the commit time higher than the average time so we don't
1110 * react too strongly to vast changes in the commit time
1111 */
1112 if (likely(journal->j_average_commit_time))
1113 journal->j_average_commit_time = (commit_time +
1114 journal->j_average_commit_time*3) / 4;
1115 else
1116 journal->j_average_commit_time = commit_time;
Dmitry Monakhov794446c2013-04-03 22:06:52 -04001117
Theodore Ts'oa931da62010-08-03 21:35:12 -04001118 write_unlock(&journal->j_state_lock);
Theodore Ts'o6c20ec82008-10-28 21:08:20 -04001119
Aneesh Kumar K.Vfb684072008-11-06 17:50:21 -05001120 if (journal->j_commit_callback)
1121 journal->j_commit_callback(journal, commit_transaction);
1122
Theodore Ts'o879c5e62009-06-17 11:47:48 -04001123 trace_jbd2_end_commit(journal, commit_transaction);
Eryu Guanf2a44522011-11-01 19:09:18 -04001124 jbd_debug(1, "JBD2: commit %d complete, head %d\n",
Dave Kleikamp470decc2006-10-11 01:20:57 -07001125 journal->j_commit_sequence, journal->j_tail_sequence);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001126
Dmitry Monakhov794446c2013-04-03 22:06:52 -04001127 write_lock(&journal->j_state_lock);
1128 spin_lock(&journal->j_list_lock);
1129 commit_transaction->t_state = T_FINISHED;
Theodore Ts'od4e839d2014-03-08 22:34:10 -05001130 /* Check if the transaction can be dropped now that we are finished */
Dmitry Monakhov794446c2013-04-03 22:06:52 -04001131 if (commit_transaction->t_checkpoint_list == NULL &&
1132 commit_transaction->t_checkpoint_io_list == NULL) {
1133 __jbd2_journal_drop_transaction(journal, commit_transaction);
1134 jbd2_journal_free_transaction(commit_transaction);
1135 }
1136 spin_unlock(&journal->j_list_lock);
1137 write_unlock(&journal->j_state_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001138 wake_up(&journal->j_wait_done_commit);
Theodore Ts'o42cf3452014-03-08 19:51:16 -05001139
1140 /*
1141 * Calculate overall stats
1142 */
1143 spin_lock(&journal->j_history_lock);
1144 journal->j_stats.ts_tid++;
1145 journal->j_stats.ts_requested += stats.ts_requested;
1146 journal->j_stats.run.rs_wait += stats.run.rs_wait;
1147 journal->j_stats.run.rs_request_delay += stats.run.rs_request_delay;
1148 journal->j_stats.run.rs_running += stats.run.rs_running;
1149 journal->j_stats.run.rs_locked += stats.run.rs_locked;
1150 journal->j_stats.run.rs_flushing += stats.run.rs_flushing;
1151 journal->j_stats.run.rs_logging += stats.run.rs_logging;
1152 journal->j_stats.run.rs_handle_count += stats.run.rs_handle_count;
1153 journal->j_stats.run.rs_blocks += stats.run.rs_blocks;
1154 journal->j_stats.run.rs_blocks_logged += stats.run.rs_blocks_logged;
1155 spin_unlock(&journal->j_history_lock);
Dave Kleikamp470decc2006-10-11 01:20:57 -07001156}