ASoC: Merge up v6.6-rc7
[linux-block.git] / fs / xfs / scrub / stats.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Copyright (C) 2023 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <djwong@kernel.org>
5  */
6 #include "xfs.h"
7 #include "xfs_fs.h"
8 #include "xfs_shared.h"
9 #include "xfs_format.h"
10 #include "xfs_trans_resv.h"
11 #include "xfs_mount.h"
12 #include "xfs_sysfs.h"
13 #include "xfs_btree.h"
14 #include "xfs_super.h"
15 #include "scrub/scrub.h"
16 #include "scrub/stats.h"
17 #include "scrub/trace.h"
18
19 struct xchk_scrub_stats {
20         /* all 32-bit counters here */
21
22         /* checking stats */
23         uint32_t                invocations;
24         uint32_t                clean;
25         uint32_t                corrupt;
26         uint32_t                preen;
27         uint32_t                xfail;
28         uint32_t                xcorrupt;
29         uint32_t                incomplete;
30         uint32_t                warning;
31         uint32_t                retries;
32
33         /* repair stats */
34         uint32_t                repair_invocations;
35         uint32_t                repair_success;
36
37         /* all 64-bit items here */
38
39         /* runtimes */
40         uint64_t                checktime_us;
41         uint64_t                repairtime_us;
42
43         /* non-counter state must go at the end for clearall */
44         spinlock_t              css_lock;
45 };
46
47 struct xchk_stats {
48         struct dentry           *cs_debugfs;
49         struct xchk_scrub_stats cs_stats[XFS_SCRUB_TYPE_NR];
50 };
51
52
53 static struct xchk_stats        global_stats;
54
55 static const char *name_map[XFS_SCRUB_TYPE_NR] = {
56         [XFS_SCRUB_TYPE_SB]             = "sb",
57         [XFS_SCRUB_TYPE_AGF]            = "agf",
58         [XFS_SCRUB_TYPE_AGFL]           = "agfl",
59         [XFS_SCRUB_TYPE_AGI]            = "agi",
60         [XFS_SCRUB_TYPE_BNOBT]          = "bnobt",
61         [XFS_SCRUB_TYPE_CNTBT]          = "cntbt",
62         [XFS_SCRUB_TYPE_INOBT]          = "inobt",
63         [XFS_SCRUB_TYPE_FINOBT]         = "finobt",
64         [XFS_SCRUB_TYPE_RMAPBT]         = "rmapbt",
65         [XFS_SCRUB_TYPE_REFCNTBT]       = "refcountbt",
66         [XFS_SCRUB_TYPE_INODE]          = "inode",
67         [XFS_SCRUB_TYPE_BMBTD]          = "bmapbtd",
68         [XFS_SCRUB_TYPE_BMBTA]          = "bmapbta",
69         [XFS_SCRUB_TYPE_BMBTC]          = "bmapbtc",
70         [XFS_SCRUB_TYPE_DIR]            = "directory",
71         [XFS_SCRUB_TYPE_XATTR]          = "xattr",
72         [XFS_SCRUB_TYPE_SYMLINK]        = "symlink",
73         [XFS_SCRUB_TYPE_PARENT]         = "parent",
74         [XFS_SCRUB_TYPE_RTBITMAP]       = "rtbitmap",
75         [XFS_SCRUB_TYPE_RTSUM]          = "rtsummary",
76         [XFS_SCRUB_TYPE_UQUOTA]         = "usrquota",
77         [XFS_SCRUB_TYPE_GQUOTA]         = "grpquota",
78         [XFS_SCRUB_TYPE_PQUOTA]         = "prjquota",
79         [XFS_SCRUB_TYPE_FSCOUNTERS]     = "fscounters",
80 };
81
82 /* Format the scrub stats into a text buffer, similar to pcp style. */
83 STATIC ssize_t
84 xchk_stats_format(
85         struct xchk_stats       *cs,
86         char                    *buf,
87         size_t                  remaining)
88 {
89         struct xchk_scrub_stats *css = &cs->cs_stats[0];
90         unsigned int            i;
91         ssize_t                 copied = 0;
92         int                     ret = 0;
93
94         for (i = 0; i < XFS_SCRUB_TYPE_NR; i++, css++) {
95                 if (!name_map[i])
96                         continue;
97
98                 ret = scnprintf(buf, remaining,
99  "%s %u %u %u %u %u %u %u %u %u %llu %u %u %llu\n",
100                                 name_map[i],
101                                 (unsigned int)css->invocations,
102                                 (unsigned int)css->clean,
103                                 (unsigned int)css->corrupt,
104                                 (unsigned int)css->preen,
105                                 (unsigned int)css->xfail,
106                                 (unsigned int)css->xcorrupt,
107                                 (unsigned int)css->incomplete,
108                                 (unsigned int)css->warning,
109                                 (unsigned int)css->retries,
110                                 (unsigned long long)css->checktime_us,
111                                 (unsigned int)css->repair_invocations,
112                                 (unsigned int)css->repair_success,
113                                 (unsigned long long)css->repairtime_us);
114                 if (ret <= 0)
115                         break;
116
117                 remaining -= ret;
118                 copied += ret;
119                 buf +=  ret;
120         }
121
122         return copied > 0 ? copied : ret;
123 }
124
125 /* Estimate the worst case buffer size required to hold the whole report. */
126 STATIC size_t
127 xchk_stats_estimate_bufsize(
128         struct xchk_stats       *cs)
129 {
130         struct xchk_scrub_stats *css = &cs->cs_stats[0];
131         unsigned int            i;
132         size_t                  field_width;
133         size_t                  ret = 0;
134
135         /* 4294967296 plus one space for each u32 field */
136         field_width = 11 * (offsetof(struct xchk_scrub_stats, checktime_us) /
137                             sizeof(uint32_t));
138
139         /* 18446744073709551615 plus one space for each u64 field */
140         field_width += 21 * ((offsetof(struct xchk_scrub_stats, css_lock) -
141                               offsetof(struct xchk_scrub_stats, checktime_us)) /
142                              sizeof(uint64_t));
143
144         for (i = 0; i < XFS_SCRUB_TYPE_NR; i++, css++) {
145                 if (!name_map[i])
146                         continue;
147
148                 /* name plus one space */
149                 ret += 1 + strlen(name_map[i]);
150
151                 /* all fields, plus newline */
152                 ret += field_width + 1;
153         }
154
155         return ret;
156 }
157
158 /* Clear all counters. */
159 STATIC void
160 xchk_stats_clearall(
161         struct xchk_stats       *cs)
162 {
163         struct xchk_scrub_stats *css = &cs->cs_stats[0];
164         unsigned int            i;
165
166         for (i = 0; i < XFS_SCRUB_TYPE_NR; i++, css++) {
167                 spin_lock(&css->css_lock);
168                 memset(css, 0, offsetof(struct xchk_scrub_stats, css_lock));
169                 spin_unlock(&css->css_lock);
170         }
171 }
172
173 #define XFS_SCRUB_OFLAG_UNCLEAN (XFS_SCRUB_OFLAG_CORRUPT | \
174                                  XFS_SCRUB_OFLAG_PREEN | \
175                                  XFS_SCRUB_OFLAG_XFAIL | \
176                                  XFS_SCRUB_OFLAG_XCORRUPT | \
177                                  XFS_SCRUB_OFLAG_INCOMPLETE | \
178                                  XFS_SCRUB_OFLAG_WARNING)
179
180 STATIC void
181 xchk_stats_merge_one(
182         struct xchk_stats               *cs,
183         const struct xfs_scrub_metadata *sm,
184         const struct xchk_stats_run     *run)
185 {
186         struct xchk_scrub_stats         *css;
187
188         if (sm->sm_type >= XFS_SCRUB_TYPE_NR) {
189                 ASSERT(sm->sm_type < XFS_SCRUB_TYPE_NR);
190                 return;
191         }
192
193         css = &cs->cs_stats[sm->sm_type];
194         spin_lock(&css->css_lock);
195         css->invocations++;
196         if (!(sm->sm_flags & XFS_SCRUB_OFLAG_UNCLEAN))
197                 css->clean++;
198         if (sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
199                 css->corrupt++;
200         if (sm->sm_flags & XFS_SCRUB_OFLAG_PREEN)
201                 css->preen++;
202         if (sm->sm_flags & XFS_SCRUB_OFLAG_XFAIL)
203                 css->xfail++;
204         if (sm->sm_flags & XFS_SCRUB_OFLAG_XCORRUPT)
205                 css->xcorrupt++;
206         if (sm->sm_flags & XFS_SCRUB_OFLAG_INCOMPLETE)
207                 css->incomplete++;
208         if (sm->sm_flags & XFS_SCRUB_OFLAG_WARNING)
209                 css->warning++;
210         css->retries += run->retries;
211         css->checktime_us += howmany_64(run->scrub_ns, NSEC_PER_USEC);
212
213         if (run->repair_attempted)
214                 css->repair_invocations++;
215         if (run->repair_succeeded)
216                 css->repair_success++;
217         css->repairtime_us += howmany_64(run->repair_ns, NSEC_PER_USEC);
218         spin_unlock(&css->css_lock);
219 }
220
221 /* Merge these scrub-run stats into the global and mount stat data. */
222 void
223 xchk_stats_merge(
224         struct xfs_mount                *mp,
225         const struct xfs_scrub_metadata *sm,
226         const struct xchk_stats_run     *run)
227 {
228         xchk_stats_merge_one(&global_stats, sm, run);
229         xchk_stats_merge_one(mp->m_scrub_stats, sm, run);
230 }
231
232 /* debugfs boilerplate */
233
234 static ssize_t
235 xchk_scrub_stats_read(
236         struct file             *file,
237         char __user             *ubuf,
238         size_t                  count,
239         loff_t                  *ppos)
240 {
241         struct xchk_stats       *cs = file->private_data;
242         char                    *buf;
243         size_t                  bufsize;
244         ssize_t                 avail, ret;
245
246         /*
247          * This generates stringly snapshot of all the scrub counters, so we
248          * do not want userspace to receive garbled text from multiple calls.
249          * If the file position is greater than 0, return a short read.
250          */
251         if (*ppos > 0)
252                 return 0;
253
254         bufsize = xchk_stats_estimate_bufsize(cs);
255
256         buf = kvmalloc(bufsize, XCHK_GFP_FLAGS);
257         if (!buf)
258                 return -ENOMEM;
259
260         avail = xchk_stats_format(cs, buf, bufsize);
261         if (avail < 0) {
262                 ret = avail;
263                 goto out;
264         }
265
266         ret = simple_read_from_buffer(ubuf, count, ppos, buf, avail);
267 out:
268         kvfree(buf);
269         return ret;
270 }
271
272 static const struct file_operations scrub_stats_fops = {
273         .open                   = simple_open,
274         .read                   = xchk_scrub_stats_read,
275 };
276
277 static ssize_t
278 xchk_clear_scrub_stats_write(
279         struct file             *file,
280         const char __user       *ubuf,
281         size_t                  count,
282         loff_t                  *ppos)
283 {
284         struct xchk_stats       *cs = file->private_data;
285         unsigned int            val;
286         int                     ret;
287
288         ret = kstrtouint_from_user(ubuf, count, 0, &val);
289         if (ret)
290                 return ret;
291
292         if (val != 1)
293                 return -EINVAL;
294
295         xchk_stats_clearall(cs);
296         return count;
297 }
298
299 static const struct file_operations clear_scrub_stats_fops = {
300         .open                   = simple_open,
301         .write                  = xchk_clear_scrub_stats_write,
302 };
303
304 /* Initialize the stats object. */
305 STATIC int
306 xchk_stats_init(
307         struct xchk_stats       *cs,
308         struct xfs_mount        *mp)
309 {
310         struct xchk_scrub_stats *css = &cs->cs_stats[0];
311         unsigned int            i;
312
313         for (i = 0; i < XFS_SCRUB_TYPE_NR; i++, css++)
314                 spin_lock_init(&css->css_lock);
315
316         return 0;
317 }
318
319 /* Connect the stats object to debugfs. */
320 void
321 xchk_stats_register(
322         struct xchk_stats       *cs,
323         struct dentry           *parent)
324 {
325         if (!parent)
326                 return;
327
328         cs->cs_debugfs = xfs_debugfs_mkdir("scrub", parent);
329         if (!cs->cs_debugfs)
330                 return;
331
332         debugfs_create_file("stats", 0644, cs->cs_debugfs, cs,
333                         &scrub_stats_fops);
334         debugfs_create_file("clear_stats", 0400, cs->cs_debugfs, cs,
335                         &clear_scrub_stats_fops);
336 }
337
338 /* Free all resources related to the stats object. */
339 STATIC int
340 xchk_stats_teardown(
341         struct xchk_stats       *cs)
342 {
343         return 0;
344 }
345
346 /* Disconnect the stats object from debugfs. */
347 void
348 xchk_stats_unregister(
349         struct xchk_stats       *cs)
350 {
351         debugfs_remove(cs->cs_debugfs);
352 }
353
354 /* Initialize global stats and register them */
355 int __init
356 xchk_global_stats_setup(
357         struct dentry           *parent)
358 {
359         int                     error;
360
361         error = xchk_stats_init(&global_stats, NULL);
362         if (error)
363                 return error;
364
365         xchk_stats_register(&global_stats, parent);
366         return 0;
367 }
368
369 /* Unregister global stats and tear them down */
370 void
371 xchk_global_stats_teardown(void)
372 {
373         xchk_stats_unregister(&global_stats);
374         xchk_stats_teardown(&global_stats);
375 }
376
377 /* Allocate per-mount stats */
378 int
379 xchk_mount_stats_alloc(
380         struct xfs_mount        *mp)
381 {
382         struct xchk_stats       *cs;
383         int                     error;
384
385         cs = kvzalloc(sizeof(struct xchk_stats), GFP_KERNEL);
386         if (!cs)
387                 return -ENOMEM;
388
389         error = xchk_stats_init(cs, mp);
390         if (error)
391                 goto out_free;
392
393         mp->m_scrub_stats = cs;
394         return 0;
395 out_free:
396         kvfree(cs);
397         return error;
398 }
399
400 /* Free per-mount stats */
401 void
402 xchk_mount_stats_free(
403         struct xfs_mount        *mp)
404 {
405         xchk_stats_teardown(mp->m_scrub_stats);
406         kvfree(mp->m_scrub_stats);
407         mp->m_scrub_stats = NULL;
408 }