Commit | Line | Data |
---|---|---|
7336d0e6 | 1 | // SPDX-License-Identifier: GPL-2.0-only |
b3b94faa DT |
2 | /* |
3 | * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. | |
cf45b752 | 4 | * Copyright (C) 2004-2008 Red Hat, Inc. All rights reserved. |
b3b94faa DT |
5 | */ |
6 | ||
b3b94faa DT |
7 | #include <linux/spinlock.h> |
8 | #include <linux/completion.h> | |
9 | #include <linux/buffer_head.h> | |
5c676f6d | 10 | #include <linux/gfs2_ondisk.h> |
6802e340 | 11 | #include <linux/bio.h> |
c65f7fb5 | 12 | #include <linux/posix_acl.h> |
f39814f6 | 13 | #include <linux/security.h> |
b3b94faa DT |
14 | |
15 | #include "gfs2.h" | |
5c676f6d | 16 | #include "incore.h" |
b3b94faa DT |
17 | #include "bmap.h" |
18 | #include "glock.h" | |
19 | #include "glops.h" | |
20 | #include "inode.h" | |
21 | #include "log.h" | |
22 | #include "meta_io.h" | |
b3b94faa DT |
23 | #include "recovery.h" |
24 | #include "rgrp.h" | |
5c676f6d | 25 | #include "util.h" |
ddacfaf7 | 26 | #include "trans.h" |
17d539f0 | 27 | #include "dir.h" |
f4686c26 | 28 | #include "lops.h" |
b3b94faa | 29 | |
2e60d768 BM |
30 | struct workqueue_struct *gfs2_freeze_wq; |
31 | ||
601ef0d5 BP |
32 | extern struct workqueue_struct *gfs2_control_wq; |
33 | ||
75549186 SW |
34 | static void gfs2_ail_error(struct gfs2_glock *gl, const struct buffer_head *bh) |
35 | { | |
69a61144 BP |
36 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
37 | ||
38 | fs_err(sdp, | |
15562c43 BP |
39 | "AIL buffer %p: blocknr %llu state 0x%08lx mapping %p page " |
40 | "state 0x%lx\n", | |
75549186 | 41 | bh, (unsigned long long)bh->b_blocknr, bh->b_state, |
11551cf1 | 42 | bh->b_folio->mapping, bh->b_folio->flags); |
69a61144 | 43 | fs_err(sdp, "AIL glock %u:%llu mapping %p\n", |
75549186 SW |
44 | gl->gl_name.ln_type, gl->gl_name.ln_number, |
45 | gfs2_glock2aspace(gl)); | |
69a61144 | 46 | gfs2_lm(sdp, "AIL error\n"); |
fffe9bee | 47 | gfs2_withdraw_delayed(sdp); |
75549186 SW |
48 | } |
49 | ||
ddacfaf7 | 50 | /** |
dba898b0 | 51 | * __gfs2_ail_flush - remove all buffers for a given lock from the AIL |
ddacfaf7 | 52 | * @gl: the glock |
b5b24d7a | 53 | * @fsync: set when called from fsync (not all buffers will be clean) |
c551f66c | 54 | * @nr_revokes: Number of buffers to revoke |
ddacfaf7 SW |
55 | * |
56 | * None of the buffers should be dirty, locked, or pinned. | |
57 | */ | |
58 | ||
1bc333f4 BM |
59 | static void __gfs2_ail_flush(struct gfs2_glock *gl, bool fsync, |
60 | unsigned int nr_revokes) | |
ddacfaf7 | 61 | { |
15562c43 | 62 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
ddacfaf7 | 63 | struct list_head *head = &gl->gl_ail_list; |
b5b24d7a | 64 | struct gfs2_bufdata *bd, *tmp; |
ddacfaf7 | 65 | struct buffer_head *bh; |
b5b24d7a | 66 | const unsigned long b_state = (1UL << BH_Dirty)|(1UL << BH_Pinned)|(1UL << BH_Lock); |
d8348de0 | 67 | |
b5b24d7a | 68 | gfs2_log_lock(sdp); |
d6a079e8 | 69 | spin_lock(&sdp->sd_ail_lock); |
1bc333f4 BM |
70 | list_for_each_entry_safe_reverse(bd, tmp, head, bd_ail_gl_list) { |
71 | if (nr_revokes == 0) | |
72 | break; | |
ddacfaf7 | 73 | bh = bd->bd_bh; |
b5b24d7a SW |
74 | if (bh->b_state & b_state) { |
75 | if (fsync) | |
76 | continue; | |
75549186 | 77 | gfs2_ail_error(gl, bh); |
b5b24d7a | 78 | } |
1ad38c43 | 79 | gfs2_trans_add_revoke(sdp, bd); |
1bc333f4 | 80 | nr_revokes--; |
ddacfaf7 | 81 | } |
8eae1ca0 | 82 | GLOCK_BUG_ON(gl, !fsync && atomic_read(&gl->gl_ail_count)); |
d6a079e8 | 83 | spin_unlock(&sdp->sd_ail_lock); |
b5b24d7a | 84 | gfs2_log_unlock(sdp); |
b0118933 AG |
85 | |
86 | if (gfs2_withdrawing(sdp)) | |
87 | gfs2_withdraw(sdp); | |
dba898b0 SW |
88 | } |
89 | ||
90 | ||
1c634f94 | 91 | static int gfs2_ail_empty_gl(struct gfs2_glock *gl) |
dba898b0 | 92 | { |
15562c43 | 93 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
dba898b0 | 94 | struct gfs2_trans tr; |
c968f578 | 95 | unsigned int revokes; |
24ab1582 | 96 | int ret = 0; |
dba898b0 | 97 | |
c968f578 | 98 | revokes = atomic_read(&gl->gl_ail_count); |
dba898b0 | 99 | |
c968f578 | 100 | if (!revokes) { |
9ff78289 BP |
101 | bool have_revokes; |
102 | bool log_in_flight; | |
103 | ||
104 | /* | |
105 | * We have nothing on the ail, but there could be revokes on | |
106 | * the sdp revoke queue, in which case, we still want to flush | |
107 | * the log and wait for it to finish. | |
108 | * | |
109 | * If the sdp revoke list is empty too, we might still have an | |
110 | * io outstanding for writing revokes, so we should wait for | |
111 | * it before returning. | |
112 | * | |
113 | * If none of these conditions are true, our revokes are all | |
114 | * flushed and we can return. | |
115 | */ | |
116 | gfs2_log_lock(sdp); | |
117 | have_revokes = !list_empty(&sdp->sd_log_revokes); | |
118 | log_in_flight = atomic_read(&sdp->sd_log_in_flight); | |
119 | gfs2_log_unlock(sdp); | |
120 | if (have_revokes) | |
121 | goto flush; | |
122 | if (log_in_flight) | |
123 | log_flush_wait(sdp); | |
1c634f94 | 124 | return 0; |
9ff78289 | 125 | } |
dba898b0 | 126 | |
c968f578 AG |
127 | memset(&tr, 0, sizeof(tr)); |
128 | set_bit(TR_ONSTACK, &tr.tr_flags); | |
129 | ret = __gfs2_trans_begin(&tr, sdp, 0, revokes, _RET_IP_); | |
b97e583c BP |
130 | if (ret) { |
131 | fs_err(sdp, "Transaction error %d: Unable to write revokes.", ret); | |
c968f578 | 132 | goto flush; |
b97e583c | 133 | } |
c968f578 | 134 | __gfs2_ail_flush(gl, 0, revokes); |
dba898b0 | 135 | gfs2_trans_end(sdp); |
c968f578 | 136 | |
9ff78289 | 137 | flush: |
644f6bf7 BP |
138 | if (!ret) |
139 | gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL | | |
140 | GFS2_LFC_AIL_EMPTY_GL); | |
24ab1582 | 141 | return ret; |
dba898b0 | 142 | } |
ddacfaf7 | 143 | |
b5b24d7a | 144 | void gfs2_ail_flush(struct gfs2_glock *gl, bool fsync) |
dba898b0 | 145 | { |
15562c43 | 146 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
dba898b0 SW |
147 | unsigned int revokes = atomic_read(&gl->gl_ail_count); |
148 | int ret; | |
149 | ||
150 | if (!revokes) | |
151 | return; | |
152 | ||
2129b428 | 153 | ret = gfs2_trans_begin(sdp, 0, revokes); |
dba898b0 SW |
154 | if (ret) |
155 | return; | |
2129b428 | 156 | __gfs2_ail_flush(gl, fsync, revokes); |
ddacfaf7 | 157 | gfs2_trans_end(sdp); |
805c0907 BP |
158 | gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL | |
159 | GFS2_LFC_AIL_FLUSH); | |
ddacfaf7 | 160 | } |
ba7f7290 | 161 | |
4a55752a BP |
162 | /** |
163 | * gfs2_rgrp_metasync - sync out the metadata of a resource group | |
164 | * @gl: the glock protecting the resource group | |
165 | * | |
166 | */ | |
167 | ||
168 | static int gfs2_rgrp_metasync(struct gfs2_glock *gl) | |
169 | { | |
170 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; | |
171 | struct address_space *metamapping = &sdp->sd_aspace; | |
172 | struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl); | |
173 | const unsigned bsize = sdp->sd_sb.sb_bsize; | |
174 | loff_t start = (rgd->rd_addr * bsize) & PAGE_MASK; | |
175 | loff_t end = PAGE_ALIGN((rgd->rd_addr + rgd->rd_length) * bsize) - 1; | |
176 | int error; | |
177 | ||
178 | filemap_fdatawrite_range(metamapping, start, end); | |
179 | error = filemap_fdatawait_range(metamapping, start, end); | |
4d927b03 | 180 | WARN_ON_ONCE(error && !gfs2_withdrawing_or_withdrawn(sdp)); |
4a55752a BP |
181 | mapping_set_error(metamapping, error); |
182 | if (error) | |
183 | gfs2_io_error(sdp); | |
184 | return error; | |
185 | } | |
186 | ||
ba7f7290 | 187 | /** |
6bac243f | 188 | * rgrp_go_sync - sync out the metadata for this glock |
b3b94faa | 189 | * @gl: the glock |
b3b94faa DT |
190 | * |
191 | * Called when demoting or unlocking an EX glock. We must flush | |
192 | * to disk all dirty buffers/pages relating to this glock, and must not | |
6f6597ba | 193 | * return to caller to demote/unlock the glock until I/O is complete. |
b3b94faa DT |
194 | */ |
195 | ||
1c634f94 | 196 | static int rgrp_go_sync(struct gfs2_glock *gl) |
b3b94faa | 197 | { |
15562c43 | 198 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
b3422cac | 199 | struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl); |
6bac243f SW |
200 | int error; |
201 | ||
fd5f446f | 202 | if (!rgd || !test_and_clear_bit(GLF_DIRTY, &gl->gl_flags)) |
1c634f94 | 203 | return 0; |
8eae1ca0 | 204 | GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE); |
b5d32bea | 205 | |
805c0907 BP |
206 | gfs2_log_flush(sdp, gl, GFS2_LOG_HEAD_FLUSH_NORMAL | |
207 | GFS2_LFC_RGRP_GO_SYNC); | |
4a55752a | 208 | error = gfs2_rgrp_metasync(gl); |
1c634f94 BP |
209 | if (!error) |
210 | error = gfs2_ail_empty_gl(gl); | |
23cfb0c3 | 211 | gfs2_free_clones(rgd); |
1c634f94 | 212 | return error; |
b3b94faa DT |
213 | } |
214 | ||
215 | /** | |
6bac243f | 216 | * rgrp_go_inval - invalidate the metadata for this glock |
b3b94faa DT |
217 | * @gl: the glock |
218 | * @flags: | |
219 | * | |
6bac243f SW |
220 | * We never used LM_ST_DEFERRED with resource groups, so that we |
221 | * should always see the metadata flag set here. | |
222 | * | |
b3b94faa DT |
223 | */ |
224 | ||
6bac243f | 225 | static void rgrp_go_inval(struct gfs2_glock *gl, int flags) |
b3b94faa | 226 | { |
15562c43 | 227 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
70d4ee94 | 228 | struct address_space *mapping = &sdp->sd_aspace; |
6f6597ba | 229 | struct gfs2_rgrpd *rgd = gfs2_glock2rgrp(gl); |
23cfb0c3 | 230 | const unsigned bsize = sdp->sd_sb.sb_bsize; |
fd5f446f | 231 | loff_t start, end; |
39b0f1e9 | 232 | |
fd5f446f BP |
233 | if (!rgd) |
234 | return; | |
235 | start = (rgd->rd_addr * bsize) & PAGE_MASK; | |
236 | end = PAGE_ALIGN((rgd->rd_addr + rgd->rd_length) * bsize) - 1; | |
23cfb0c3 | 237 | gfs2_rgrp_brelse(rgd); |
8eae1ca0 | 238 | WARN_ON_ONCE(!(flags & DIO_METADATA)); |
23cfb0c3 | 239 | truncate_inode_pages_range(mapping, start, end); |
b3b94faa DT |
240 | } |
241 | ||
f246dd4b | 242 | static void gfs2_rgrp_go_dump(struct seq_file *seq, const struct gfs2_glock *gl, |
0e539ca1 AP |
243 | const char *fs_id_buf) |
244 | { | |
16e6281b | 245 | struct gfs2_rgrpd *rgd = gl->gl_object; |
0e539ca1 AP |
246 | |
247 | if (rgd) | |
248 | gfs2_rgrp_dump(seq, rgd, fs_id_buf); | |
249 | } | |
250 | ||
4fd1a579 AG |
251 | static struct gfs2_inode *gfs2_glock2inode(struct gfs2_glock *gl) |
252 | { | |
253 | struct gfs2_inode *ip; | |
254 | ||
255 | spin_lock(&gl->gl_lockref.lock); | |
256 | ip = gl->gl_object; | |
257 | if (ip) | |
258 | set_bit(GIF_GLOP_PENDING, &ip->i_flags); | |
259 | spin_unlock(&gl->gl_lockref.lock); | |
260 | return ip; | |
261 | } | |
262 | ||
6f6597ba AG |
263 | struct gfs2_rgrpd *gfs2_glock2rgrp(struct gfs2_glock *gl) |
264 | { | |
265 | struct gfs2_rgrpd *rgd; | |
266 | ||
267 | spin_lock(&gl->gl_lockref.lock); | |
268 | rgd = gl->gl_object; | |
269 | spin_unlock(&gl->gl_lockref.lock); | |
270 | ||
271 | return rgd; | |
272 | } | |
273 | ||
4fd1a579 AG |
274 | static void gfs2_clear_glop_pending(struct gfs2_inode *ip) |
275 | { | |
276 | if (!ip) | |
277 | return; | |
278 | ||
279 | clear_bit_unlock(GIF_GLOP_PENDING, &ip->i_flags); | |
280 | wake_up_bit(&ip->i_flags, GIF_GLOP_PENDING); | |
281 | } | |
282 | ||
b5d32bea | 283 | /** |
4a55752a BP |
284 | * gfs2_inode_metasync - sync out the metadata of an inode |
285 | * @gl: the glock protecting the inode | |
286 | * | |
287 | */ | |
288 | int gfs2_inode_metasync(struct gfs2_glock *gl) | |
289 | { | |
290 | struct address_space *metamapping = gfs2_glock2aspace(gl); | |
291 | int error; | |
292 | ||
293 | filemap_fdatawrite(metamapping); | |
294 | error = filemap_fdatawait(metamapping); | |
295 | if (error) | |
296 | gfs2_io_error(gl->gl_name.ln_sbd); | |
297 | return error; | |
298 | } | |
299 | ||
300 | /** | |
301 | * inode_go_sync - Sync the dirty metadata of an inode | |
b5d32bea SW |
302 | * @gl: the glock protecting the inode |
303 | * | |
304 | */ | |
305 | ||
1c634f94 | 306 | static int inode_go_sync(struct gfs2_glock *gl) |
b5d32bea | 307 | { |
4fd1a579 AG |
308 | struct gfs2_inode *ip = gfs2_glock2inode(gl); |
309 | int isreg = ip && S_ISREG(ip->i_inode.i_mode); | |
009d8518 | 310 | struct address_space *metamapping = gfs2_glock2aspace(gl); |
bbae10fa | 311 | int error = 0, ret; |
3042a2cc | 312 | |
4fd1a579 | 313 | if (isreg) { |
582d2f7a SW |
314 | if (test_and_clear_bit(GIF_SW_PAGED, &ip->i_flags)) |
315 | unmap_shared_mapping_range(ip->i_inode.i_mapping, 0, 0); | |
316 | inode_dio_wait(&ip->i_inode); | |
317 | } | |
6bac243f | 318 | if (!test_and_clear_bit(GLF_DIRTY, &gl->gl_flags)) |
4fd1a579 | 319 | goto out; |
b5d32bea | 320 | |
8eae1ca0 | 321 | GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_EXCLUSIVE); |
6bac243f | 322 | |
805c0907 BP |
323 | gfs2_log_flush(gl->gl_name.ln_sbd, gl, GFS2_LOG_HEAD_FLUSH_NORMAL | |
324 | GFS2_LFC_INODE_GO_SYNC); | |
6bac243f | 325 | filemap_fdatawrite(metamapping); |
4fd1a579 | 326 | if (isreg) { |
6bac243f SW |
327 | struct address_space *mapping = ip->i_inode.i_mapping; |
328 | filemap_fdatawrite(mapping); | |
329 | error = filemap_fdatawait(mapping); | |
330 | mapping_set_error(mapping, error); | |
b5d32bea | 331 | } |
4a55752a | 332 | ret = gfs2_inode_metasync(gl); |
bbae10fa BP |
333 | if (!error) |
334 | error = ret; | |
24ab1582 BP |
335 | ret = gfs2_ail_empty_gl(gl); |
336 | if (!error) | |
337 | error = ret; | |
52fcd11c SW |
338 | /* |
339 | * Writeback of the data mapping may cause the dirty flag to be set | |
340 | * so we have to clear it again here. | |
341 | */ | |
4e857c58 | 342 | smp_mb__before_atomic(); |
52fcd11c | 343 | clear_bit(GLF_DIRTY, &gl->gl_flags); |
4fd1a579 AG |
344 | |
345 | out: | |
346 | gfs2_clear_glop_pending(ip); | |
1c634f94 | 347 | return error; |
b5d32bea SW |
348 | } |
349 | ||
b3b94faa DT |
350 | /** |
351 | * inode_go_inval - prepare a inode glock to be released | |
352 | * @gl: the glock | |
353 | * @flags: | |
6b49d1d9 GU |
354 | * |
355 | * Normally we invalidate everything, but if we are moving into | |
6bac243f SW |
356 | * LM_ST_DEFERRED from LM_ST_SHARED or LM_ST_EXCLUSIVE then we |
357 | * can keep hold of the metadata, since it won't have changed. | |
b3b94faa DT |
358 | * |
359 | */ | |
360 | ||
361 | static void inode_go_inval(struct gfs2_glock *gl, int flags) | |
362 | { | |
4fd1a579 | 363 | struct gfs2_inode *ip = gfs2_glock2inode(gl); |
b3b94faa | 364 | |
6bac243f | 365 | if (flags & DIO_METADATA) { |
009d8518 | 366 | struct address_space *mapping = gfs2_glock2aspace(gl); |
6bac243f | 367 | truncate_inode_pages(mapping, 0); |
c65f7fb5 | 368 | if (ip) { |
f2e70d8f | 369 | set_bit(GLF_INSTANTIATE_NEEDED, &gl->gl_flags); |
c65f7fb5 | 370 | forget_all_cached_acls(&ip->i_inode); |
f39814f6 | 371 | security_inode_invalidate_secctx(&ip->i_inode); |
17d539f0 | 372 | gfs2_dir_hash_inval(ip); |
c65f7fb5 | 373 | } |
b004157a SW |
374 | } |
375 | ||
15562c43 | 376 | if (ip == GFS2_I(gl->gl_name.ln_sbd->sd_rindex)) { |
c1696fb8 | 377 | gfs2_log_flush(gl->gl_name.ln_sbd, NULL, |
805c0907 BP |
378 | GFS2_LOG_HEAD_FLUSH_NORMAL | |
379 | GFS2_LFC_INODE_GO_INVAL); | |
15562c43 | 380 | gl->gl_name.ln_sbd->sd_rindex_uptodate = 0; |
1ce53368 | 381 | } |
3cc3f710 | 382 | if (ip && S_ISREG(ip->i_inode.i_mode)) |
b004157a | 383 | truncate_inode_pages(ip->i_inode.i_mapping, 0); |
4fd1a579 AG |
384 | |
385 | gfs2_clear_glop_pending(ip); | |
b3b94faa DT |
386 | } |
387 | ||
388 | /** | |
389 | * inode_go_demote_ok - Check to see if it's ok to unlock an inode glock | |
390 | * @gl: the glock | |
391 | * | |
392 | * Returns: 1 if it's ok | |
393 | */ | |
394 | ||
97cc1025 | 395 | static int inode_go_demote_ok(const struct gfs2_glock *gl) |
b3b94faa | 396 | { |
15562c43 | 397 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
bc015cb8 | 398 | |
97cc1025 SW |
399 | if (sdp->sd_jindex == gl->gl_object || sdp->sd_rindex == gl->gl_object) |
400 | return 0; | |
bc015cb8 | 401 | |
97cc1025 | 402 | return 1; |
b3b94faa DT |
403 | } |
404 | ||
d4b2cf1b SW |
405 | static int gfs2_dinode_in(struct gfs2_inode *ip, const void *buf) |
406 | { | |
cfcdb5ba | 407 | struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); |
d4b2cf1b | 408 | const struct gfs2_dinode *str = buf; |
580f721b | 409 | struct timespec64 atime, iatime; |
d4b2cf1b | 410 | u16 height, depth; |
4a378d8a | 411 | umode_t mode = be32_to_cpu(str->di_mode); |
7db35444 AG |
412 | struct inode *inode = &ip->i_inode; |
413 | bool is_new = inode->i_state & I_NEW; | |
d4b2cf1b | 414 | |
10398ef5 AP |
415 | if (unlikely(ip->i_no_addr != be64_to_cpu(str->di_num.no_addr))) { |
416 | gfs2_consist_inode(ip); | |
417 | return -EIO; | |
418 | } | |
419 | if (unlikely(!is_new && inode_wrong_type(inode, mode))) { | |
420 | gfs2_consist_inode(ip); | |
421 | return -EIO; | |
422 | } | |
d4b2cf1b | 423 | ip->i_no_formal_ino = be64_to_cpu(str->di_num.no_formal_ino); |
7db35444 | 424 | inode->i_mode = mode; |
4a378d8a | 425 | if (is_new) { |
7db35444 | 426 | inode->i_rdev = 0; |
4a378d8a AV |
427 | switch (mode & S_IFMT) { |
428 | case S_IFBLK: | |
429 | case S_IFCHR: | |
7db35444 AG |
430 | inode->i_rdev = MKDEV(be32_to_cpu(str->di_major), |
431 | be32_to_cpu(str->di_minor)); | |
4a378d8a AV |
432 | break; |
433 | } | |
098b9c14 | 434 | } |
d4b2cf1b | 435 | |
7db35444 AG |
436 | i_uid_write(inode, be32_to_cpu(str->di_uid)); |
437 | i_gid_write(inode, be32_to_cpu(str->di_gid)); | |
438 | set_nlink(inode, be32_to_cpu(str->di_nlink)); | |
439 | i_size_write(inode, be64_to_cpu(str->di_size)); | |
440 | gfs2_set_inode_blocks(inode, be64_to_cpu(str->di_blocks)); | |
d4b2cf1b SW |
441 | atime.tv_sec = be64_to_cpu(str->di_atime); |
442 | atime.tv_nsec = be32_to_cpu(str->di_atime_nsec); | |
580f721b JL |
443 | iatime = inode_get_atime(inode); |
444 | if (timespec64_compare(&iatime, &atime) < 0) | |
445 | inode_set_atime_to_ts(inode, atime); | |
446 | inode_set_mtime(inode, be64_to_cpu(str->di_mtime), | |
447 | be32_to_cpu(str->di_mtime_nsec)); | |
8a8b8d91 JL |
448 | inode_set_ctime(inode, be64_to_cpu(str->di_ctime), |
449 | be32_to_cpu(str->di_ctime_nsec)); | |
d4b2cf1b SW |
450 | |
451 | ip->i_goal = be64_to_cpu(str->di_goal_meta); | |
452 | ip->i_generation = be64_to_cpu(str->di_generation); | |
453 | ||
454 | ip->i_diskflags = be32_to_cpu(str->di_flags); | |
9964afbb SW |
455 | ip->i_eattr = be64_to_cpu(str->di_eattr); |
456 | /* i_diskflags and i_eattr must be set before gfs2_set_inode_flags() */ | |
7db35444 | 457 | gfs2_set_inode_flags(inode); |
d4b2cf1b | 458 | height = be16_to_cpu(str->di_height); |
10398ef5 AP |
459 | if (unlikely(height > sdp->sd_max_height)) { |
460 | gfs2_consist_inode(ip); | |
461 | return -EIO; | |
462 | } | |
d4b2cf1b SW |
463 | ip->i_height = (u8)height; |
464 | ||
465 | depth = be16_to_cpu(str->di_depth); | |
10398ef5 AP |
466 | if (unlikely(depth > GFS2_DIR_MAX_DEPTH)) { |
467 | gfs2_consist_inode(ip); | |
468 | return -EIO; | |
469 | } | |
d4b2cf1b SW |
470 | ip->i_depth = (u8)depth; |
471 | ip->i_entries = be32_to_cpu(str->di_entries); | |
472 | ||
10398ef5 AP |
473 | if (gfs2_is_stuffed(ip) && inode->i_size > gfs2_max_stuffed_size(ip)) { |
474 | gfs2_consist_inode(ip); | |
475 | return -EIO; | |
476 | } | |
7db35444 AG |
477 | if (S_ISREG(inode->i_mode)) |
478 | gfs2_set_aops(inode); | |
d4b2cf1b SW |
479 | |
480 | return 0; | |
d4b2cf1b SW |
481 | } |
482 | ||
483 | /** | |
484 | * gfs2_inode_refresh - Refresh the incore copy of the dinode | |
485 | * @ip: The GFS2 inode | |
486 | * | |
487 | * Returns: errno | |
488 | */ | |
489 | ||
490 | int gfs2_inode_refresh(struct gfs2_inode *ip) | |
491 | { | |
492 | struct buffer_head *dibh; | |
493 | int error; | |
494 | ||
495 | error = gfs2_meta_inode_buffer(ip, &dibh); | |
496 | if (error) | |
497 | return error; | |
498 | ||
d4b2cf1b SW |
499 | error = gfs2_dinode_in(ip, dibh->b_data); |
500 | brelse(dibh); | |
d4b2cf1b SW |
501 | return error; |
502 | } | |
503 | ||
b3b94faa | 504 | /** |
3278b977 | 505 | * inode_go_instantiate - read in an inode if necessary |
1181f2d9 | 506 | * @gl: The glock |
b3b94faa DT |
507 | * |
508 | * Returns: errno | |
509 | */ | |
510 | ||
5f38a4d3 | 511 | static int inode_go_instantiate(struct gfs2_glock *gl) |
b3b94faa | 512 | { |
5c676f6d | 513 | struct gfs2_inode *ip = gl->gl_object; |
b3b94faa | 514 | |
f2e70d8f | 515 | if (!ip) /* no inode to populate - read it in later */ |
86c30a01 | 516 | return 0; |
b3b94faa | 517 | |
86c30a01 AG |
518 | return gfs2_inode_refresh(ip); |
519 | } | |
520 | ||
521 | static int inode_go_held(struct gfs2_holder *gh) | |
522 | { | |
523 | struct gfs2_glock *gl = gh->gh_gl; | |
524 | struct gfs2_inode *ip = gl->gl_object; | |
525 | int error = 0; | |
526 | ||
527 | if (!ip) /* no inode to populate - read it in later */ | |
528 | return 0; | |
b3b94faa | 529 | |
582d2f7a SW |
530 | if (gh->gh_state != LM_ST_DEFERRED) |
531 | inode_dio_wait(&ip->i_inode); | |
532 | ||
383f01fb | 533 | if ((ip->i_diskflags & GFS2_DIF_TRUNC_IN_PROG) && |
b3b94faa | 534 | (gl->gl_state == LM_ST_EXCLUSIVE) && |
de3f906f AG |
535 | (gh->gh_state == LM_ST_EXCLUSIVE)) |
536 | error = gfs2_truncatei_resume(ip); | |
b3b94faa DT |
537 | |
538 | return error; | |
539 | } | |
540 | ||
6802e340 SW |
541 | /** |
542 | * inode_go_dump - print information about an inode | |
543 | * @seq: The iterator | |
c551f66c | 544 | * @gl: The glock |
3792ce97 | 545 | * @fs_id_buf: file system id (may be empty) |
6802e340 | 546 | * |
6802e340 SW |
547 | */ |
548 | ||
f246dd4b | 549 | static void inode_go_dump(struct seq_file *seq, const struct gfs2_glock *gl, |
3792ce97 | 550 | const char *fs_id_buf) |
6802e340 | 551 | { |
27a2660f | 552 | struct gfs2_inode *ip = gl->gl_object; |
f246dd4b | 553 | const struct inode *inode = &ip->i_inode; |
27a2660f | 554 | |
6802e340 | 555 | if (ip == NULL) |
ac3beb6a | 556 | return; |
27a2660f | 557 | |
3792ce97 BP |
558 | gfs2_print_dbg(seq, "%s I: n:%llu/%llu t:%u f:0x%02lx d:0x%08x s:%llu " |
559 | "p:%lu\n", fs_id_buf, | |
6802e340 SW |
560 | (unsigned long long)ip->i_no_formal_ino, |
561 | (unsigned long long)ip->i_no_addr, | |
f246dd4b | 562 | IF2DT(inode->i_mode), ip->i_flags, |
fa75cedc | 563 | (unsigned int)ip->i_diskflags, |
f246dd4b AG |
564 | (unsigned long long)i_size_read(inode), |
565 | inode->i_data.nrpages); | |
6802e340 SW |
566 | } |
567 | ||
b3b94faa | 568 | /** |
b77b4a48 | 569 | * freeze_go_callback - A cluster node is requesting a freeze |
b3b94faa | 570 | * @gl: the glock |
b77b4a48 | 571 | * @remote: true if this came from a different cluster node |
b3b94faa DT |
572 | */ |
573 | ||
b77b4a48 | 574 | static void freeze_go_callback(struct gfs2_glock *gl, bool remote) |
b3b94faa | 575 | { |
15562c43 | 576 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
b77b4a48 AG |
577 | struct super_block *sb = sdp->sd_vfs; |
578 | ||
579 | if (!remote || | |
52954b75 AG |
580 | (gl->gl_state != LM_ST_SHARED && |
581 | gl->gl_state != LM_ST_UNLOCKED) || | |
b77b4a48 AG |
582 | gl->gl_demote_state != LM_ST_UNLOCKED) |
583 | return; | |
b3b94faa | 584 | |
20b32912 | 585 | /* |
b77b4a48 | 586 | * Try to get an active super block reference to prevent racing with |
52954b75 AG |
587 | * unmount (see super_trylock_shared()). But note that unmount isn't |
588 | * the only place where a write lock on s_umount is taken, and we can | |
589 | * fail here because of things like remount as well. | |
20b32912 | 590 | */ |
b77b4a48 AG |
591 | if (down_read_trylock(&sb->s_umount)) { |
592 | atomic_inc(&sb->s_active); | |
593 | up_read(&sb->s_umount); | |
594 | if (!queue_work(gfs2_freeze_wq, &sdp->sd_freeze_work)) | |
595 | deactivate_super(sb); | |
b3b94faa | 596 | } |
b3b94faa DT |
597 | } |
598 | ||
599 | /** | |
24972557 | 600 | * freeze_go_xmote_bh - After promoting/demoting the freeze glock |
b3b94faa | 601 | * @gl: the glock |
b3b94faa | 602 | */ |
f68effb3 | 603 | static int freeze_go_xmote_bh(struct gfs2_glock *gl) |
b3b94faa | 604 | { |
15562c43 | 605 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
feaa7bba | 606 | struct gfs2_inode *ip = GFS2_I(sdp->sd_jdesc->jd_inode); |
5c676f6d | 607 | struct gfs2_glock *j_gl = ip->i_gl; |
55167622 | 608 | struct gfs2_log_header_host head; |
b3b94faa DT |
609 | int error; |
610 | ||
6802e340 | 611 | if (test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) { |
1a14d3a6 | 612 | j_gl->gl_ops->go_inval(j_gl, DIO_METADATA); |
b3b94faa | 613 | |
f4686c26 | 614 | error = gfs2_find_jhead(sdp->sd_jdesc, &head, false); |
9d9b1605 BP |
615 | if (gfs2_assert_withdraw_delayed(sdp, !error)) |
616 | return error; | |
617 | if (gfs2_assert_withdraw_delayed(sdp, head.lh_flags & | |
618 | GFS2_LOG_HEAD_UNMOUNT)) | |
619 | return -EIO; | |
620 | sdp->sd_log_sequence = head.lh_sequence + 1; | |
621 | gfs2_log_pointers_init(sdp, head.lh_blkno); | |
b3b94faa | 622 | } |
6802e340 | 623 | return 0; |
b3b94faa DT |
624 | } |
625 | ||
b94a170e BM |
626 | /** |
627 | * iopen_go_callback - schedule the dcache entry for the inode to be deleted | |
628 | * @gl: the glock | |
c551f66c | 629 | * @remote: true if this came from a different cluster node |
b94a170e | 630 | * |
f3dd1649 | 631 | * gl_lockref.lock lock is held while calling this |
b94a170e | 632 | */ |
81ffbf65 | 633 | static void iopen_go_callback(struct gfs2_glock *gl, bool remote) |
b94a170e | 634 | { |
6f6597ba | 635 | struct gfs2_inode *ip = gl->gl_object; |
15562c43 | 636 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; |
001e8e8d | 637 | |
6c0246a9 | 638 | if (!remote || sb_rdonly(sdp->sd_vfs) || |
e7beb8b6 | 639 | test_bit(SDF_KILL, &sdp->sd_flags)) |
001e8e8d | 640 | return; |
b94a170e BM |
641 | |
642 | if (gl->gl_demote_state == LM_ST_UNLOCKED && | |
009d8518 | 643 | gl->gl_state == LM_ST_SHARED && ip) { |
e66cf161 | 644 | gl->gl_lockref.count++; |
f0e56edc | 645 | if (!gfs2_queue_try_to_evict(gl)) |
e66cf161 | 646 | gl->gl_lockref.count--; |
b94a170e BM |
647 | } |
648 | } | |
649 | ||
601ef0d5 BP |
650 | /** |
651 | * inode_go_free - wake up anyone waiting for dlm's unlock ast to free it | |
652 | * @gl: glock being freed | |
653 | * | |
654 | * For now, this is only used for the journal inode glock. In withdraw | |
655 | * situations, we need to wait for the glock to be freed so that we know | |
656 | * other nodes may proceed with recovery / journal replay. | |
657 | */ | |
658 | static void inode_go_free(struct gfs2_glock *gl) | |
659 | { | |
660 | /* Note that we cannot reference gl_object because it's already set | |
661 | * to NULL by this point in its lifecycle. */ | |
662 | if (!test_bit(GLF_FREEING, &gl->gl_flags)) | |
663 | return; | |
664 | clear_bit_unlock(GLF_FREEING, &gl->gl_flags); | |
665 | wake_up_bit(&gl->gl_flags, GLF_FREEING); | |
666 | } | |
667 | ||
668 | /** | |
669 | * nondisk_go_callback - used to signal when a node did a withdraw | |
670 | * @gl: the nondisk glock | |
671 | * @remote: true if this came from a different cluster node | |
672 | * | |
673 | */ | |
674 | static void nondisk_go_callback(struct gfs2_glock *gl, bool remote) | |
675 | { | |
676 | struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; | |
677 | ||
678 | /* Ignore the callback unless it's from another node, and it's the | |
679 | live lock. */ | |
680 | if (!remote || gl->gl_name.ln_number != GFS2_LIVE_LOCK) | |
681 | return; | |
682 | ||
683 | /* First order of business is to cancel the demote request. We don't | |
684 | * really want to demote a nondisk glock. At best it's just to inform | |
685 | * us of another node's withdraw. We'll keep it in SH mode. */ | |
686 | clear_bit(GLF_DEMOTE, &gl->gl_flags); | |
687 | clear_bit(GLF_PENDING_DEMOTE, &gl->gl_flags); | |
688 | ||
689 | /* Ignore the unlock if we're withdrawn, unmounting, or in recovery. */ | |
690 | if (test_bit(SDF_NORECOVERY, &sdp->sd_flags) || | |
691 | test_bit(SDF_WITHDRAWN, &sdp->sd_flags) || | |
692 | test_bit(SDF_REMOTE_WITHDRAW, &sdp->sd_flags)) | |
693 | return; | |
694 | ||
695 | /* We only care when a node wants us to unlock, because that means | |
696 | * they want a journal recovered. */ | |
697 | if (gl->gl_demote_state != LM_ST_UNLOCKED) | |
698 | return; | |
699 | ||
700 | if (sdp->sd_args.ar_spectator) { | |
701 | fs_warn(sdp, "Spectator node cannot recover journals.\n"); | |
702 | return; | |
703 | } | |
704 | ||
705 | fs_warn(sdp, "Some node has withdrawn; checking for recovery.\n"); | |
706 | set_bit(SDF_REMOTE_WITHDRAW, &sdp->sd_flags); | |
707 | /* | |
708 | * We can't call remote_withdraw directly here or gfs2_recover_journal | |
709 | * because this is called from the glock unlock function and the | |
710 | * remote_withdraw needs to enqueue and dequeue the same "live" glock | |
711 | * we were called from. So we queue it to the control work queue in | |
712 | * lock_dlm. | |
713 | */ | |
714 | queue_delayed_work(gfs2_control_wq, &sdp->sd_control_work, 0); | |
715 | } | |
716 | ||
8fb4b536 | 717 | const struct gfs2_glock_operations gfs2_meta_glops = { |
ea67eedb | 718 | .go_type = LM_TYPE_META, |
a72d2401 | 719 | .go_flags = GLOF_NONDISK, |
b3b94faa DT |
720 | }; |
721 | ||
8fb4b536 | 722 | const struct gfs2_glock_operations gfs2_inode_glops = { |
06dfc306 | 723 | .go_sync = inode_go_sync, |
b3b94faa DT |
724 | .go_inval = inode_go_inval, |
725 | .go_demote_ok = inode_go_demote_ok, | |
3278b977 | 726 | .go_instantiate = inode_go_instantiate, |
86c30a01 | 727 | .go_held = inode_go_held, |
6802e340 | 728 | .go_dump = inode_go_dump, |
ea67eedb | 729 | .go_type = LM_TYPE_INODE, |
f286d627 | 730 | .go_flags = GLOF_ASPACE | GLOF_LRU | GLOF_LVB, |
601ef0d5 | 731 | .go_free = inode_go_free, |
b3b94faa DT |
732 | }; |
733 | ||
8fb4b536 | 734 | const struct gfs2_glock_operations gfs2_rgrp_glops = { |
06dfc306 | 735 | .go_sync = rgrp_go_sync, |
6bac243f | 736 | .go_inval = rgrp_go_inval, |
3278b977 | 737 | .go_instantiate = gfs2_rgrp_go_instantiate, |
0e539ca1 | 738 | .go_dump = gfs2_rgrp_go_dump, |
ea67eedb | 739 | .go_type = LM_TYPE_RGRP, |
70d4ee94 | 740 | .go_flags = GLOF_LVB, |
b3b94faa DT |
741 | }; |
742 | ||
24972557 | 743 | const struct gfs2_glock_operations gfs2_freeze_glops = { |
24972557 | 744 | .go_xmote_bh = freeze_go_xmote_bh, |
b77b4a48 | 745 | .go_callback = freeze_go_callback, |
ea67eedb | 746 | .go_type = LM_TYPE_NONDISK, |
a72d2401 | 747 | .go_flags = GLOF_NONDISK, |
b3b94faa DT |
748 | }; |
749 | ||
8fb4b536 | 750 | const struct gfs2_glock_operations gfs2_iopen_glops = { |
ea67eedb | 751 | .go_type = LM_TYPE_IOPEN, |
b94a170e | 752 | .go_callback = iopen_go_callback, |
74382e27 | 753 | .go_dump = inode_go_dump, |
a72d2401 | 754 | .go_flags = GLOF_LRU | GLOF_NONDISK, |
515b269d | 755 | .go_subclass = 1, |
b3b94faa DT |
756 | }; |
757 | ||
8fb4b536 | 758 | const struct gfs2_glock_operations gfs2_flock_glops = { |
ea67eedb | 759 | .go_type = LM_TYPE_FLOCK, |
a72d2401 | 760 | .go_flags = GLOF_LRU | GLOF_NONDISK, |
b3b94faa DT |
761 | }; |
762 | ||
8fb4b536 | 763 | const struct gfs2_glock_operations gfs2_nondisk_glops = { |
ea67eedb | 764 | .go_type = LM_TYPE_NONDISK, |
a72d2401 | 765 | .go_flags = GLOF_NONDISK, |
601ef0d5 | 766 | .go_callback = nondisk_go_callback, |
b3b94faa DT |
767 | }; |
768 | ||
8fb4b536 | 769 | const struct gfs2_glock_operations gfs2_quota_glops = { |
ea67eedb | 770 | .go_type = LM_TYPE_QUOTA, |
a72d2401 | 771 | .go_flags = GLOF_LVB | GLOF_LRU | GLOF_NONDISK, |
b3b94faa DT |
772 | }; |
773 | ||
8fb4b536 | 774 | const struct gfs2_glock_operations gfs2_journal_glops = { |
ea67eedb | 775 | .go_type = LM_TYPE_JOURNAL, |
a72d2401 | 776 | .go_flags = GLOF_NONDISK, |
b3b94faa DT |
777 | }; |
778 | ||
64d576ba SW |
779 | const struct gfs2_glock_operations *gfs2_glops_list[] = { |
780 | [LM_TYPE_META] = &gfs2_meta_glops, | |
781 | [LM_TYPE_INODE] = &gfs2_inode_glops, | |
782 | [LM_TYPE_RGRP] = &gfs2_rgrp_glops, | |
64d576ba SW |
783 | [LM_TYPE_IOPEN] = &gfs2_iopen_glops, |
784 | [LM_TYPE_FLOCK] = &gfs2_flock_glops, | |
785 | [LM_TYPE_NONDISK] = &gfs2_nondisk_glops, | |
786 | [LM_TYPE_QUOTA] = &gfs2_quota_glops, | |
787 | [LM_TYPE_JOURNAL] = &gfs2_journal_glops, | |
788 | }; | |
789 |