xref: /openbmc/linux/fs/gfs2/trans.c (revision 79f382b9)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
4  * Copyright (C) 2004-2006 Red Hat, Inc.  All rights reserved.
5  */
6 
7 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
8 
9 #include <linux/sched.h>
10 #include <linux/slab.h>
11 #include <linux/spinlock.h>
12 #include <linux/completion.h>
13 #include <linux/buffer_head.h>
14 #include <linux/kallsyms.h>
15 #include <linux/gfs2_ondisk.h>
16 
17 #include "gfs2.h"
18 #include "incore.h"
19 #include "glock.h"
20 #include "inode.h"
21 #include "log.h"
22 #include "lops.h"
23 #include "meta_io.h"
24 #include "trans.h"
25 #include "util.h"
26 #include "trace_gfs2.h"
27 
28 static void gfs2_print_trans(struct gfs2_sbd *sdp, const struct gfs2_trans *tr)
29 {
30 	fs_warn(sdp, "Transaction created at: %pSR\n", (void *)tr->tr_ip);
31 	fs_warn(sdp, "blocks=%u revokes=%u reserved=%u touched=%u\n",
32 		tr->tr_blocks, tr->tr_revokes, tr->tr_reserved,
33 		test_bit(TR_TOUCHED, &tr->tr_flags));
34 	fs_warn(sdp, "Buf %u/%u Databuf %u/%u Revoke %u\n",
35 		tr->tr_num_buf_new, tr->tr_num_buf_rm,
36 		tr->tr_num_databuf_new, tr->tr_num_databuf_rm,
37 		tr->tr_num_revoke);
38 }
39 
40 int __gfs2_trans_begin(struct gfs2_trans *tr, struct gfs2_sbd *sdp,
41 		       unsigned int blocks, unsigned int revokes,
42 		       unsigned long ip)
43 {
44 	unsigned int extra_revokes;
45 
46 	if (current->journal_info) {
47 		gfs2_print_trans(sdp, current->journal_info);
48 		BUG();
49 	}
50 	BUG_ON(blocks == 0 && revokes == 0);
51 
52 	if (!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags))
53 		return -EROFS;
54 
55 	tr->tr_ip = ip;
56 	tr->tr_blocks = blocks;
57 	tr->tr_revokes = revokes;
58 	tr->tr_reserved = GFS2_LOG_FLUSH_MIN_BLOCKS;
59 	if (blocks) {
60 		/*
61 		 * The reserved blocks are either used for data or metadata.
62 		 * We can have mixed data and metadata, each with its own log
63 		 * descriptor block; see calc_reserved().
64 		 */
65 		tr->tr_reserved += blocks + 1 + DIV_ROUND_UP(blocks - 1, databuf_limit(sdp));
66 	}
67 	INIT_LIST_HEAD(&tr->tr_databuf);
68 	INIT_LIST_HEAD(&tr->tr_buf);
69 	INIT_LIST_HEAD(&tr->tr_list);
70 	INIT_LIST_HEAD(&tr->tr_ail1_list);
71 	INIT_LIST_HEAD(&tr->tr_ail2_list);
72 
73 	if (gfs2_assert_warn(sdp, tr->tr_reserved <= sdp->sd_jdesc->jd_blocks))
74 		return -EINVAL;
75 
76 	sb_start_intwrite(sdp->sd_vfs);
77 
78 	/*
79 	 * Try the reservations under sd_log_flush_lock to prevent log flushes
80 	 * from creating inconsistencies between the number of allocated and
81 	 * reserved revokes.  If that fails, do a full-block allocation outside
82 	 * of the lock to avoid stalling log flushes.  Then, allot the
83 	 * appropriate number of blocks to revokes, use as many revokes locally
84 	 * as needed, and "release" the surplus into the revokes pool.
85 	 */
86 
87 	down_read(&sdp->sd_log_flush_lock);
88 	if (gfs2_log_try_reserve(sdp, tr, &extra_revokes))
89 		goto reserved;
90 	up_read(&sdp->sd_log_flush_lock);
91 	gfs2_log_reserve(sdp, tr, &extra_revokes);
92 	down_read(&sdp->sd_log_flush_lock);
93 
94 reserved:
95 	gfs2_log_release_revokes(sdp, extra_revokes);
96 	if (unlikely(!test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags))) {
97 		gfs2_log_release_revokes(sdp, tr->tr_revokes);
98 		up_read(&sdp->sd_log_flush_lock);
99 		gfs2_log_release(sdp, tr->tr_reserved);
100 		sb_end_intwrite(sdp->sd_vfs);
101 		return -EROFS;
102 	}
103 
104 	current->journal_info = tr;
105 
106 	return 0;
107 }
108 
109 int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks,
110 		     unsigned int revokes)
111 {
112 	struct gfs2_trans *tr;
113 	int error;
114 
115 	tr = kmem_cache_zalloc(gfs2_trans_cachep, GFP_NOFS);
116 	if (!tr)
117 		return -ENOMEM;
118 	error = __gfs2_trans_begin(tr, sdp, blocks, revokes, _RET_IP_);
119 	if (error)
120 		kmem_cache_free(gfs2_trans_cachep, tr);
121 	return error;
122 }
123 
124 void gfs2_trans_end(struct gfs2_sbd *sdp)
125 {
126 	struct gfs2_trans *tr = current->journal_info;
127 	s64 nbuf;
128 
129 	current->journal_info = NULL;
130 
131 	if (!test_bit(TR_TOUCHED, &tr->tr_flags)) {
132 		gfs2_log_release_revokes(sdp, tr->tr_revokes);
133 		up_read(&sdp->sd_log_flush_lock);
134 		gfs2_log_release(sdp, tr->tr_reserved);
135 		if (!test_bit(TR_ONSTACK, &tr->tr_flags))
136 			gfs2_trans_free(sdp, tr);
137 		sb_end_intwrite(sdp->sd_vfs);
138 		return;
139 	}
140 
141 	gfs2_log_release_revokes(sdp, tr->tr_revokes - tr->tr_num_revoke);
142 
143 	nbuf = tr->tr_num_buf_new + tr->tr_num_databuf_new;
144 	nbuf -= tr->tr_num_buf_rm;
145 	nbuf -= tr->tr_num_databuf_rm;
146 
147 	if (gfs2_assert_withdraw(sdp, nbuf <= tr->tr_blocks) ||
148 	    gfs2_assert_withdraw(sdp, tr->tr_num_revoke <= tr->tr_revokes))
149 		gfs2_print_trans(sdp, tr);
150 
151 	gfs2_log_commit(sdp, tr);
152 	if (!test_bit(TR_ONSTACK, &tr->tr_flags) &&
153 	    !test_bit(TR_ATTACHED, &tr->tr_flags))
154 		gfs2_trans_free(sdp, tr);
155 	up_read(&sdp->sd_log_flush_lock);
156 
157 	if (sdp->sd_vfs->s_flags & SB_SYNCHRONOUS)
158 		gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL |
159 			       GFS2_LFC_TRANS_END);
160 	sb_end_intwrite(sdp->sd_vfs);
161 }
162 
163 static struct gfs2_bufdata *gfs2_alloc_bufdata(struct gfs2_glock *gl,
164 					       struct buffer_head *bh)
165 {
166 	struct gfs2_bufdata *bd;
167 
168 	bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL);
169 	bd->bd_bh = bh;
170 	bd->bd_gl = gl;
171 	INIT_LIST_HEAD(&bd->bd_list);
172 	bh->b_private = bd;
173 	return bd;
174 }
175 
176 /**
177  * gfs2_trans_add_data - Add a databuf to the transaction.
178  * @gl: The inode glock associated with the buffer
179  * @bh: The buffer to add
180  *
181  * This is used in journaled data mode.
182  * We need to journal the data block in the same way as metadata in
183  * the functions above. The difference is that here we have a tag
184  * which is two __be64's being the block number (as per meta data)
185  * and a flag which says whether the data block needs escaping or
186  * not. This means we need a new log entry for each 251 or so data
187  * blocks, which isn't an enormous overhead but twice as much as
188  * for normal metadata blocks.
189  */
190 void gfs2_trans_add_data(struct gfs2_glock *gl, struct buffer_head *bh)
191 {
192 	struct gfs2_trans *tr = current->journal_info;
193 	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
194 	struct gfs2_bufdata *bd;
195 
196 	lock_buffer(bh);
197 	if (buffer_pinned(bh)) {
198 		set_bit(TR_TOUCHED, &tr->tr_flags);
199 		goto out;
200 	}
201 	gfs2_log_lock(sdp);
202 	bd = bh->b_private;
203 	if (bd == NULL) {
204 		gfs2_log_unlock(sdp);
205 		unlock_buffer(bh);
206 		if (bh->b_private == NULL)
207 			bd = gfs2_alloc_bufdata(gl, bh);
208 		else
209 			bd = bh->b_private;
210 		lock_buffer(bh);
211 		gfs2_log_lock(sdp);
212 	}
213 	gfs2_assert(sdp, bd->bd_gl == gl);
214 	set_bit(TR_TOUCHED, &tr->tr_flags);
215 	if (list_empty(&bd->bd_list)) {
216 		set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags);
217 		set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags);
218 		gfs2_pin(sdp, bd->bd_bh);
219 		tr->tr_num_databuf_new++;
220 		list_add_tail(&bd->bd_list, &tr->tr_databuf);
221 	}
222 	gfs2_log_unlock(sdp);
223 out:
224 	unlock_buffer(bh);
225 }
226 
227 void gfs2_trans_add_meta(struct gfs2_glock *gl, struct buffer_head *bh)
228 {
229 
230 	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
231 	struct gfs2_bufdata *bd;
232 	struct gfs2_meta_header *mh;
233 	struct gfs2_trans *tr = current->journal_info;
234 	enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state);
235 
236 	lock_buffer(bh);
237 	if (buffer_pinned(bh)) {
238 		set_bit(TR_TOUCHED, &tr->tr_flags);
239 		goto out;
240 	}
241 	gfs2_log_lock(sdp);
242 	bd = bh->b_private;
243 	if (bd == NULL) {
244 		gfs2_log_unlock(sdp);
245 		unlock_buffer(bh);
246 		lock_page(bh->b_page);
247 		if (bh->b_private == NULL)
248 			bd = gfs2_alloc_bufdata(gl, bh);
249 		else
250 			bd = bh->b_private;
251 		unlock_page(bh->b_page);
252 		lock_buffer(bh);
253 		gfs2_log_lock(sdp);
254 	}
255 	gfs2_assert(sdp, bd->bd_gl == gl);
256 	set_bit(TR_TOUCHED, &tr->tr_flags);
257 	if (!list_empty(&bd->bd_list))
258 		goto out_unlock;
259 	set_bit(GLF_LFLUSH, &bd->bd_gl->gl_flags);
260 	set_bit(GLF_DIRTY, &bd->bd_gl->gl_flags);
261 	mh = (struct gfs2_meta_header *)bd->bd_bh->b_data;
262 	if (unlikely(mh->mh_magic != cpu_to_be32(GFS2_MAGIC))) {
263 		fs_err(sdp, "Attempting to add uninitialised block to "
264 		       "journal (inplace block=%lld)\n",
265 		       (unsigned long long)bd->bd_bh->b_blocknr);
266 		BUG();
267 	}
268 	if (unlikely(state == SFS_FROZEN)) {
269 		fs_info(sdp, "GFS2:adding buf while frozen\n");
270 		gfs2_assert_withdraw(sdp, 0);
271 	}
272 	if (unlikely(gfs2_withdrawn(sdp))) {
273 		fs_info(sdp, "GFS2:adding buf while withdrawn! 0x%llx\n",
274 			(unsigned long long)bd->bd_bh->b_blocknr);
275 	}
276 	gfs2_pin(sdp, bd->bd_bh);
277 	mh->__pad0 = cpu_to_be64(0);
278 	mh->mh_jid = cpu_to_be32(sdp->sd_jdesc->jd_jid);
279 	list_add(&bd->bd_list, &tr->tr_buf);
280 	tr->tr_num_buf_new++;
281 out_unlock:
282 	gfs2_log_unlock(sdp);
283 out:
284 	unlock_buffer(bh);
285 }
286 
287 void gfs2_trans_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd)
288 {
289 	struct gfs2_trans *tr = current->journal_info;
290 
291 	BUG_ON(!list_empty(&bd->bd_list));
292 	gfs2_add_revoke(sdp, bd);
293 	set_bit(TR_TOUCHED, &tr->tr_flags);
294 	tr->tr_num_revoke++;
295 }
296 
297 void gfs2_trans_remove_revoke(struct gfs2_sbd *sdp, u64 blkno, unsigned int len)
298 {
299 	struct gfs2_bufdata *bd, *tmp;
300 	unsigned int n = len;
301 
302 	gfs2_log_lock(sdp);
303 	list_for_each_entry_safe(bd, tmp, &sdp->sd_log_revokes, bd_list) {
304 		if ((bd->bd_blkno >= blkno) && (bd->bd_blkno < (blkno + len))) {
305 			list_del_init(&bd->bd_list);
306 			gfs2_assert_withdraw(sdp, sdp->sd_log_num_revoke);
307 			sdp->sd_log_num_revoke--;
308 			if (bd->bd_gl)
309 				gfs2_glock_remove_revoke(bd->bd_gl);
310 			kmem_cache_free(gfs2_bufdata_cachep, bd);
311 			gfs2_log_release_revokes(sdp, 1);
312 			if (--n == 0)
313 				break;
314 		}
315 	}
316 	gfs2_log_unlock(sdp);
317 }
318 
319 void gfs2_trans_free(struct gfs2_sbd *sdp, struct gfs2_trans *tr)
320 {
321 	if (tr == NULL)
322 		return;
323 
324 	gfs2_assert_warn(sdp, list_empty(&tr->tr_ail1_list));
325 	gfs2_assert_warn(sdp, list_empty(&tr->tr_ail2_list));
326 	gfs2_assert_warn(sdp, list_empty(&tr->tr_databuf));
327 	gfs2_assert_warn(sdp, list_empty(&tr->tr_buf));
328 	kmem_cache_free(gfs2_trans_cachep, tr);
329 }
330