1 | // SPDX-License-Identifier: GPL-2.0 |
2 | /* |
3 | * f2fs debugging statistics |
4 | * |
5 | * Copyright (c) 2012 Samsung Electronics Co., Ltd. |
6 | * http://www.samsung.com/ |
7 | * Copyright (c) 2012 Linux Foundation |
8 | * Copyright (c) 2012 Greg Kroah-Hartman <gregkh@linuxfoundation.org> |
9 | */ |
10 | |
11 | #include <linux/fs.h> |
12 | #include <linux/backing-dev.h> |
13 | #include <linux/f2fs_fs.h> |
14 | #include <linux/blkdev.h> |
15 | #include <linux/debugfs.h> |
16 | #include <linux/seq_file.h> |
17 | |
18 | #include "f2fs.h" |
19 | #include "node.h" |
20 | #include "segment.h" |
21 | #include "gc.h" |
22 | |
23 | static LIST_HEAD(f2fs_stat_list); |
24 | static DEFINE_RAW_SPINLOCK(f2fs_stat_lock); |
25 | #ifdef CONFIG_DEBUG_FS |
26 | static struct dentry *f2fs_debugfs_root; |
27 | #endif |
28 | |
29 | /* |
30 | * This function calculates BDF of every segments |
31 | */ |
32 | void f2fs_update_sit_info(struct f2fs_sb_info *sbi) |
33 | { |
34 | struct f2fs_stat_info *si = F2FS_STAT(sbi); |
35 | unsigned long long blks_per_sec, hblks_per_sec, total_vblocks; |
36 | unsigned long long bimodal, dist; |
37 | unsigned int segno, vblocks; |
38 | int ndirty = 0; |
39 | |
40 | bimodal = 0; |
41 | total_vblocks = 0; |
42 | blks_per_sec = CAP_BLKS_PER_SEC(sbi); |
43 | hblks_per_sec = blks_per_sec / 2; |
44 | for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { |
45 | vblocks = get_valid_blocks(sbi, segno, use_section: true); |
46 | dist = abs(vblocks - hblks_per_sec); |
47 | bimodal += dist * dist; |
48 | |
49 | if (vblocks > 0 && vblocks < blks_per_sec) { |
50 | total_vblocks += vblocks; |
51 | ndirty++; |
52 | } |
53 | } |
54 | dist = div_u64(MAIN_SECS(sbi) * hblks_per_sec * hblks_per_sec, divisor: 100); |
55 | si->bimodal = div64_u64(dividend: bimodal, divisor: dist); |
56 | if (si->dirty_count) |
57 | si->avg_vblocks = div_u64(dividend: total_vblocks, divisor: ndirty); |
58 | else |
59 | si->avg_vblocks = 0; |
60 | } |
61 | |
62 | #ifdef CONFIG_DEBUG_FS |
63 | static void update_general_status(struct f2fs_sb_info *sbi) |
64 | { |
65 | struct f2fs_stat_info *si = F2FS_STAT(sbi); |
66 | struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi); |
67 | int i; |
68 | |
69 | /* these will be changed if online resize is done */ |
70 | si->main_area_segs = le32_to_cpu(raw_super->segment_count_main); |
71 | si->main_area_sections = le32_to_cpu(raw_super->section_count); |
72 | si->main_area_zones = si->main_area_sections / |
73 | le32_to_cpu(raw_super->secs_per_zone); |
74 | |
75 | /* general extent cache stats */ |
76 | for (i = 0; i < NR_EXTENT_CACHES; i++) { |
77 | struct extent_tree_info *eti = &sbi->extent_tree[i]; |
78 | |
79 | si->hit_cached[i] = atomic64_read(v: &sbi->read_hit_cached[i]); |
80 | si->hit_rbtree[i] = atomic64_read(v: &sbi->read_hit_rbtree[i]); |
81 | si->total_ext[i] = atomic64_read(v: &sbi->total_hit_ext[i]); |
82 | si->hit_total[i] = si->hit_cached[i] + si->hit_rbtree[i]; |
83 | si->ext_tree[i] = atomic_read(v: &eti->total_ext_tree); |
84 | si->zombie_tree[i] = atomic_read(v: &eti->total_zombie_tree); |
85 | si->ext_node[i] = atomic_read(v: &eti->total_ext_node); |
86 | } |
87 | /* read extent_cache only */ |
88 | si->hit_largest = atomic64_read(v: &sbi->read_hit_largest); |
89 | si->hit_total[EX_READ] += si->hit_largest; |
90 | |
91 | /* block age extent_cache only */ |
92 | si->allocated_data_blocks = atomic64_read(v: &sbi->allocated_data_blocks); |
93 | |
94 | /* validation check of the segment numbers */ |
95 | si->ndirty_node = get_pages(sbi, count_type: F2FS_DIRTY_NODES); |
96 | si->ndirty_dent = get_pages(sbi, count_type: F2FS_DIRTY_DENTS); |
97 | si->ndirty_meta = get_pages(sbi, count_type: F2FS_DIRTY_META); |
98 | si->ndirty_data = get_pages(sbi, count_type: F2FS_DIRTY_DATA); |
99 | si->ndirty_qdata = get_pages(sbi, count_type: F2FS_DIRTY_QDATA); |
100 | si->ndirty_imeta = get_pages(sbi, count_type: F2FS_DIRTY_IMETA); |
101 | si->ndirty_dirs = sbi->ndirty_inode[DIR_INODE]; |
102 | si->ndirty_files = sbi->ndirty_inode[FILE_INODE]; |
103 | si->nquota_files = sbi->nquota_files; |
104 | si->ndirty_all = sbi->ndirty_inode[DIRTY_META]; |
105 | si->aw_cnt = atomic_read(v: &sbi->atomic_files); |
106 | si->max_aw_cnt = atomic_read(v: &sbi->max_aw_cnt); |
107 | si->nr_dio_read = get_pages(sbi, count_type: F2FS_DIO_READ); |
108 | si->nr_dio_write = get_pages(sbi, count_type: F2FS_DIO_WRITE); |
109 | si->nr_wb_cp_data = get_pages(sbi, count_type: F2FS_WB_CP_DATA); |
110 | si->nr_wb_data = get_pages(sbi, count_type: F2FS_WB_DATA); |
111 | si->nr_rd_data = get_pages(sbi, count_type: F2FS_RD_DATA); |
112 | si->nr_rd_node = get_pages(sbi, count_type: F2FS_RD_NODE); |
113 | si->nr_rd_meta = get_pages(sbi, count_type: F2FS_RD_META); |
114 | if (SM_I(sbi)->fcc_info) { |
115 | si->nr_flushed = |
116 | atomic_read(v: &SM_I(sbi)->fcc_info->issued_flush); |
117 | si->nr_flushing = |
118 | atomic_read(v: &SM_I(sbi)->fcc_info->queued_flush); |
119 | si->flush_list_empty = |
120 | llist_empty(head: &SM_I(sbi)->fcc_info->issue_list); |
121 | } |
122 | if (SM_I(sbi)->dcc_info) { |
123 | si->nr_discarded = |
124 | atomic_read(v: &SM_I(sbi)->dcc_info->issued_discard); |
125 | si->nr_discarding = |
126 | atomic_read(v: &SM_I(sbi)->dcc_info->queued_discard); |
127 | si->nr_discard_cmd = |
128 | atomic_read(v: &SM_I(sbi)->dcc_info->discard_cmd_cnt); |
129 | si->undiscard_blks = SM_I(sbi)->dcc_info->undiscard_blks; |
130 | } |
131 | si->nr_issued_ckpt = atomic_read(v: &sbi->cprc_info.issued_ckpt); |
132 | si->nr_total_ckpt = atomic_read(v: &sbi->cprc_info.total_ckpt); |
133 | si->nr_queued_ckpt = atomic_read(v: &sbi->cprc_info.queued_ckpt); |
134 | spin_lock(lock: &sbi->cprc_info.stat_lock); |
135 | si->cur_ckpt_time = sbi->cprc_info.cur_time; |
136 | si->peak_ckpt_time = sbi->cprc_info.peak_time; |
137 | spin_unlock(lock: &sbi->cprc_info.stat_lock); |
138 | si->total_count = BLKS_TO_SEGS(sbi, (int)sbi->user_block_count); |
139 | si->rsvd_segs = reserved_segments(sbi); |
140 | si->overp_segs = overprovision_segments(sbi); |
141 | si->valid_count = valid_user_blocks(sbi); |
142 | si->discard_blks = discard_blocks(sbi); |
143 | si->valid_node_count = valid_node_count(sbi); |
144 | si->valid_inode_count = valid_inode_count(sbi); |
145 | si->inline_xattr = atomic_read(v: &sbi->inline_xattr); |
146 | si->inline_inode = atomic_read(v: &sbi->inline_inode); |
147 | si->inline_dir = atomic_read(v: &sbi->inline_dir); |
148 | si->compr_inode = atomic_read(v: &sbi->compr_inode); |
149 | si->swapfile_inode = atomic_read(v: &sbi->swapfile_inode); |
150 | si->compr_blocks = atomic64_read(v: &sbi->compr_blocks); |
151 | si->append = sbi->im[APPEND_INO].ino_num; |
152 | si->update = sbi->im[UPDATE_INO].ino_num; |
153 | si->orphans = sbi->im[ORPHAN_INO].ino_num; |
154 | si->utilization = utilization(sbi); |
155 | |
156 | si->free_segs = free_segments(sbi); |
157 | si->free_secs = free_sections(sbi); |
158 | si->prefree_count = prefree_segments(sbi); |
159 | si->dirty_count = dirty_segments(sbi); |
160 | if (sbi->node_inode) |
161 | si->node_pages = NODE_MAPPING(sbi)->nrpages; |
162 | if (sbi->meta_inode) |
163 | si->meta_pages = META_MAPPING(sbi)->nrpages; |
164 | #ifdef CONFIG_F2FS_FS_COMPRESSION |
165 | if (sbi->compress_inode) { |
166 | si->compress_pages = COMPRESS_MAPPING(sbi)->nrpages; |
167 | si->compress_page_hit = atomic_read(v: &sbi->compress_page_hit); |
168 | } |
169 | #endif |
170 | si->nats = NM_I(sbi)->nat_cnt[TOTAL_NAT]; |
171 | si->dirty_nats = NM_I(sbi)->nat_cnt[DIRTY_NAT]; |
172 | si->sits = MAIN_SEGS(sbi); |
173 | si->dirty_sits = SIT_I(sbi)->dirty_sentries; |
174 | si->free_nids = NM_I(sbi)->nid_cnt[FREE_NID]; |
175 | si->avail_nids = NM_I(sbi)->available_nids; |
176 | si->alloc_nids = NM_I(sbi)->nid_cnt[PREALLOC_NID]; |
177 | si->io_skip_bggc = sbi->io_skip_bggc; |
178 | si->other_skip_bggc = sbi->other_skip_bggc; |
179 | si->util_free = (int)(BLKS_TO_SEGS(sbi, free_user_blocks(sbi))) |
180 | * 100 / (int)(sbi->user_block_count >> sbi->log_blocks_per_seg) |
181 | / 2; |
182 | si->util_valid = (int)(BLKS_TO_SEGS(sbi, written_block_count(sbi))) |
183 | * 100 / (int)(sbi->user_block_count >> sbi->log_blocks_per_seg) |
184 | / 2; |
185 | si->util_invalid = 50 - si->util_free - si->util_valid; |
186 | for (i = CURSEG_HOT_DATA; i < NO_CHECK_TYPE; i++) { |
187 | struct curseg_info *curseg = CURSEG_I(sbi, type: i); |
188 | |
189 | si->curseg[i] = curseg->segno; |
190 | si->cursec[i] = GET_SEC_FROM_SEG(sbi, curseg->segno); |
191 | si->curzone[i] = GET_ZONE_FROM_SEC(sbi, si->cursec[i]); |
192 | } |
193 | |
194 | for (i = META_CP; i < META_MAX; i++) |
195 | si->meta_count[i] = atomic_read(v: &sbi->meta_count[i]); |
196 | |
197 | for (i = 0; i < NO_CHECK_TYPE; i++) { |
198 | si->dirty_seg[i] = 0; |
199 | si->full_seg[i] = 0; |
200 | si->valid_blks[i] = 0; |
201 | } |
202 | |
203 | for (i = 0; i < MAIN_SEGS(sbi); i++) { |
204 | int blks = get_seg_entry(sbi, segno: i)->valid_blocks; |
205 | int type = get_seg_entry(sbi, segno: i)->type; |
206 | |
207 | if (!blks) |
208 | continue; |
209 | |
210 | if (blks == BLKS_PER_SEG(sbi)) |
211 | si->full_seg[type]++; |
212 | else |
213 | si->dirty_seg[type]++; |
214 | si->valid_blks[type] += blks; |
215 | } |
216 | |
217 | for (i = 0; i < MAX_CALL_TYPE; i++) |
218 | si->cp_call_count[i] = atomic_read(v: &sbi->cp_call_count[i]); |
219 | |
220 | for (i = 0; i < 2; i++) { |
221 | si->segment_count[i] = sbi->segment_count[i]; |
222 | si->block_count[i] = sbi->block_count[i]; |
223 | } |
224 | |
225 | si->inplace_count = atomic_read(v: &sbi->inplace_count); |
226 | } |
227 | |
228 | /* |
229 | * This function calculates memory footprint. |
230 | */ |
231 | static void update_mem_info(struct f2fs_sb_info *sbi) |
232 | { |
233 | struct f2fs_stat_info *si = F2FS_STAT(sbi); |
234 | int i; |
235 | |
236 | if (si->base_mem) |
237 | goto get_cache; |
238 | |
239 | /* build stat */ |
240 | si->base_mem = sizeof(struct f2fs_stat_info); |
241 | |
242 | /* build superblock */ |
243 | si->base_mem += sizeof(struct f2fs_sb_info) + sbi->sb->s_blocksize; |
244 | si->base_mem += 2 * sizeof(struct f2fs_inode_info); |
245 | si->base_mem += sizeof(*sbi->ckpt); |
246 | |
247 | /* build sm */ |
248 | si->base_mem += sizeof(struct f2fs_sm_info); |
249 | |
250 | /* build sit */ |
251 | si->base_mem += sizeof(struct sit_info); |
252 | si->base_mem += MAIN_SEGS(sbi) * sizeof(struct seg_entry); |
253 | si->base_mem += f2fs_bitmap_size(MAIN_SEGS(sbi)); |
254 | si->base_mem += 2 * SIT_VBLOCK_MAP_SIZE * MAIN_SEGS(sbi); |
255 | si->base_mem += SIT_VBLOCK_MAP_SIZE * MAIN_SEGS(sbi); |
256 | si->base_mem += SIT_VBLOCK_MAP_SIZE; |
257 | if (__is_large_section(sbi)) |
258 | si->base_mem += MAIN_SECS(sbi) * sizeof(struct sec_entry); |
259 | si->base_mem += __bitmap_size(sbi, flag: SIT_BITMAP); |
260 | |
261 | /* build free segmap */ |
262 | si->base_mem += sizeof(struct free_segmap_info); |
263 | si->base_mem += f2fs_bitmap_size(MAIN_SEGS(sbi)); |
264 | si->base_mem += f2fs_bitmap_size(MAIN_SECS(sbi)); |
265 | |
266 | /* build curseg */ |
267 | si->base_mem += sizeof(struct curseg_info) * NR_CURSEG_TYPE; |
268 | si->base_mem += PAGE_SIZE * NR_CURSEG_TYPE; |
269 | |
270 | /* build dirty segmap */ |
271 | si->base_mem += sizeof(struct dirty_seglist_info); |
272 | si->base_mem += NR_DIRTY_TYPE * f2fs_bitmap_size(MAIN_SEGS(sbi)); |
273 | si->base_mem += f2fs_bitmap_size(MAIN_SECS(sbi)); |
274 | |
275 | /* build nm */ |
276 | si->base_mem += sizeof(struct f2fs_nm_info); |
277 | si->base_mem += __bitmap_size(sbi, flag: NAT_BITMAP); |
278 | si->base_mem += (NM_I(sbi)->nat_bits_blocks << F2FS_BLKSIZE_BITS); |
279 | si->base_mem += NM_I(sbi)->nat_blocks * |
280 | f2fs_bitmap_size(NAT_ENTRY_PER_BLOCK); |
281 | si->base_mem += NM_I(sbi)->nat_blocks / 8; |
282 | si->base_mem += NM_I(sbi)->nat_blocks * sizeof(unsigned short); |
283 | |
284 | get_cache: |
285 | si->cache_mem = 0; |
286 | |
287 | /* build gc */ |
288 | if (sbi->gc_thread) |
289 | si->cache_mem += sizeof(struct f2fs_gc_kthread); |
290 | |
291 | /* build merge flush thread */ |
292 | if (SM_I(sbi)->fcc_info) |
293 | si->cache_mem += sizeof(struct flush_cmd_control); |
294 | if (SM_I(sbi)->dcc_info) { |
295 | si->cache_mem += sizeof(struct discard_cmd_control); |
296 | si->cache_mem += sizeof(struct discard_cmd) * |
297 | atomic_read(v: &SM_I(sbi)->dcc_info->discard_cmd_cnt); |
298 | } |
299 | |
300 | /* free nids */ |
301 | si->cache_mem += (NM_I(sbi)->nid_cnt[FREE_NID] + |
302 | NM_I(sbi)->nid_cnt[PREALLOC_NID]) * |
303 | sizeof(struct free_nid); |
304 | si->cache_mem += NM_I(sbi)->nat_cnt[TOTAL_NAT] * |
305 | sizeof(struct nat_entry); |
306 | si->cache_mem += NM_I(sbi)->nat_cnt[DIRTY_NAT] * |
307 | sizeof(struct nat_entry_set); |
308 | for (i = 0; i < MAX_INO_ENTRY; i++) |
309 | si->cache_mem += sbi->im[i].ino_num * sizeof(struct ino_entry); |
310 | |
311 | for (i = 0; i < NR_EXTENT_CACHES; i++) { |
312 | struct extent_tree_info *eti = &sbi->extent_tree[i]; |
313 | |
314 | si->ext_mem[i] = atomic_read(v: &eti->total_ext_tree) * |
315 | sizeof(struct extent_tree); |
316 | si->ext_mem[i] += atomic_read(v: &eti->total_ext_node) * |
317 | sizeof(struct extent_node); |
318 | si->cache_mem += si->ext_mem[i]; |
319 | } |
320 | |
321 | si->page_mem = 0; |
322 | if (sbi->node_inode) { |
323 | unsigned long npages = NODE_MAPPING(sbi)->nrpages; |
324 | |
325 | si->page_mem += (unsigned long long)npages << PAGE_SHIFT; |
326 | } |
327 | if (sbi->meta_inode) { |
328 | unsigned long npages = META_MAPPING(sbi)->nrpages; |
329 | |
330 | si->page_mem += (unsigned long long)npages << PAGE_SHIFT; |
331 | } |
332 | #ifdef CONFIG_F2FS_FS_COMPRESSION |
333 | if (sbi->compress_inode) { |
334 | unsigned long npages = COMPRESS_MAPPING(sbi)->nrpages; |
335 | |
336 | si->page_mem += (unsigned long long)npages << PAGE_SHIFT; |
337 | } |
338 | #endif |
339 | } |
340 | |
341 | static const char *s_flag[MAX_SBI_FLAG] = { |
342 | [SBI_IS_DIRTY] = "fs_dirty" , |
343 | [SBI_IS_CLOSE] = "closing" , |
344 | [SBI_NEED_FSCK] = "need_fsck" , |
345 | [SBI_POR_DOING] = "recovering" , |
346 | [SBI_NEED_SB_WRITE] = "sb_dirty" , |
347 | [SBI_NEED_CP] = "need_cp" , |
348 | [SBI_IS_SHUTDOWN] = "shutdown" , |
349 | [SBI_IS_RECOVERED] = "recovered" , |
350 | [SBI_CP_DISABLED] = "cp_disabled" , |
351 | [SBI_CP_DISABLED_QUICK] = "cp_disabled_quick" , |
352 | [SBI_QUOTA_NEED_FLUSH] = "quota_need_flush" , |
353 | [SBI_QUOTA_SKIP_FLUSH] = "quota_skip_flush" , |
354 | [SBI_QUOTA_NEED_REPAIR] = "quota_need_repair" , |
355 | [SBI_IS_RESIZEFS] = "resizefs" , |
356 | [SBI_IS_FREEZING] = "freezefs" , |
357 | [SBI_IS_WRITABLE] = "writable" , |
358 | }; |
359 | |
360 | static const char *ipu_mode_names[F2FS_IPU_MAX] = { |
361 | [F2FS_IPU_FORCE] = "FORCE" , |
362 | [F2FS_IPU_SSR] = "SSR" , |
363 | [F2FS_IPU_UTIL] = "UTIL" , |
364 | [F2FS_IPU_SSR_UTIL] = "SSR_UTIL" , |
365 | [F2FS_IPU_FSYNC] = "FSYNC" , |
366 | [F2FS_IPU_ASYNC] = "ASYNC" , |
367 | [F2FS_IPU_NOCACHE] = "NOCACHE" , |
368 | [F2FS_IPU_HONOR_OPU_WRITE] = "HONOR_OPU_WRITE" , |
369 | }; |
370 | |
371 | static int stat_show(struct seq_file *s, void *v) |
372 | { |
373 | struct f2fs_stat_info *si; |
374 | int i = 0, j = 0; |
375 | unsigned long flags; |
376 | |
377 | raw_spin_lock_irqsave(&f2fs_stat_lock, flags); |
378 | list_for_each_entry(si, &f2fs_stat_list, stat_list) { |
379 | struct f2fs_sb_info *sbi = si->sbi; |
380 | |
381 | update_general_status(sbi); |
382 | |
383 | seq_printf(m: s, fmt: "\n=====[ partition info(%pg). #%d, %s, CP: %s]=====\n" , |
384 | sbi->sb->s_bdev, i++, |
385 | f2fs_readonly(sb: sbi->sb) ? "RO" : "RW" , |
386 | is_set_ckpt_flags(sbi, CP_DISABLED_FLAG) ? |
387 | "Disabled" : (f2fs_cp_error(sbi) ? "Error" : "Good" )); |
388 | if (sbi->s_flag) { |
389 | seq_puts(m: s, s: "[SBI:" ); |
390 | for_each_set_bit(j, &sbi->s_flag, MAX_SBI_FLAG) |
391 | seq_printf(m: s, fmt: " %s" , s_flag[j]); |
392 | seq_puts(m: s, s: "]\n" ); |
393 | } |
394 | seq_printf(m: s, fmt: "[SB: 1] [CP: 2] [SIT: %d] [NAT: %d] " , |
395 | si->sit_area_segs, si->nat_area_segs); |
396 | seq_printf(m: s, fmt: "[SSA: %d] [MAIN: %d" , |
397 | si->ssa_area_segs, si->main_area_segs); |
398 | seq_printf(m: s, fmt: "(OverProv:%d Resv:%d)]\n\n" , |
399 | si->overp_segs, si->rsvd_segs); |
400 | seq_printf(m: s, fmt: "Current Time Sec: %llu / Mounted Time Sec: %llu\n\n" , |
401 | ktime_get_boottime_seconds(), |
402 | SIT_I(sbi)->mounted_time); |
403 | |
404 | seq_puts(m: s, s: "Policy:\n" ); |
405 | seq_puts(m: s, s: " - IPU: [" ); |
406 | if (IS_F2FS_IPU_DISABLE(sbi)) { |
407 | seq_puts(m: s, s: " DISABLE" ); |
408 | } else { |
409 | unsigned long policy = SM_I(sbi)->ipu_policy; |
410 | |
411 | for_each_set_bit(j, &policy, F2FS_IPU_MAX) |
412 | seq_printf(m: s, fmt: " %s" , ipu_mode_names[j]); |
413 | } |
414 | seq_puts(m: s, s: " ]\n\n" ); |
415 | |
416 | if (test_opt(sbi, DISCARD)) |
417 | seq_printf(m: s, fmt: "Utilization: %u%% (%u valid blocks, %u discard blocks)\n" , |
418 | si->utilization, si->valid_count, si->discard_blks); |
419 | else |
420 | seq_printf(m: s, fmt: "Utilization: %u%% (%u valid blocks)\n" , |
421 | si->utilization, si->valid_count); |
422 | |
423 | seq_printf(m: s, fmt: " - Node: %u (Inode: %u, " , |
424 | si->valid_node_count, si->valid_inode_count); |
425 | seq_printf(m: s, fmt: "Other: %u)\n - Data: %u\n" , |
426 | si->valid_node_count - si->valid_inode_count, |
427 | si->valid_count - si->valid_node_count); |
428 | seq_printf(m: s, fmt: " - Inline_xattr Inode: %u\n" , |
429 | si->inline_xattr); |
430 | seq_printf(m: s, fmt: " - Inline_data Inode: %u\n" , |
431 | si->inline_inode); |
432 | seq_printf(m: s, fmt: " - Inline_dentry Inode: %u\n" , |
433 | si->inline_dir); |
434 | seq_printf(m: s, fmt: " - Compressed Inode: %u, Blocks: %llu\n" , |
435 | si->compr_inode, si->compr_blocks); |
436 | seq_printf(m: s, fmt: " - Swapfile Inode: %u\n" , |
437 | si->swapfile_inode); |
438 | seq_printf(m: s, fmt: " - Orphan/Append/Update Inode: %u, %u, %u\n" , |
439 | si->orphans, si->append, si->update); |
440 | seq_printf(m: s, fmt: "\nMain area: %d segs, %d secs %d zones\n" , |
441 | si->main_area_segs, si->main_area_sections, |
442 | si->main_area_zones); |
443 | seq_printf(m: s, fmt: " TYPE %8s %8s %8s %10s %10s %10s\n" , |
444 | "segno" , "secno" , "zoneno" , "dirty_seg" , "full_seg" , "valid_blk" ); |
445 | seq_printf(m: s, fmt: " - COLD data: %8d %8d %8d %10u %10u %10u\n" , |
446 | si->curseg[CURSEG_COLD_DATA], |
447 | si->cursec[CURSEG_COLD_DATA], |
448 | si->curzone[CURSEG_COLD_DATA], |
449 | si->dirty_seg[CURSEG_COLD_DATA], |
450 | si->full_seg[CURSEG_COLD_DATA], |
451 | si->valid_blks[CURSEG_COLD_DATA]); |
452 | seq_printf(m: s, fmt: " - WARM data: %8d %8d %8d %10u %10u %10u\n" , |
453 | si->curseg[CURSEG_WARM_DATA], |
454 | si->cursec[CURSEG_WARM_DATA], |
455 | si->curzone[CURSEG_WARM_DATA], |
456 | si->dirty_seg[CURSEG_WARM_DATA], |
457 | si->full_seg[CURSEG_WARM_DATA], |
458 | si->valid_blks[CURSEG_WARM_DATA]); |
459 | seq_printf(m: s, fmt: " - HOT data: %8d %8d %8d %10u %10u %10u\n" , |
460 | si->curseg[CURSEG_HOT_DATA], |
461 | si->cursec[CURSEG_HOT_DATA], |
462 | si->curzone[CURSEG_HOT_DATA], |
463 | si->dirty_seg[CURSEG_HOT_DATA], |
464 | si->full_seg[CURSEG_HOT_DATA], |
465 | si->valid_blks[CURSEG_HOT_DATA]); |
466 | seq_printf(m: s, fmt: " - Dir dnode: %8d %8d %8d %10u %10u %10u\n" , |
467 | si->curseg[CURSEG_HOT_NODE], |
468 | si->cursec[CURSEG_HOT_NODE], |
469 | si->curzone[CURSEG_HOT_NODE], |
470 | si->dirty_seg[CURSEG_HOT_NODE], |
471 | si->full_seg[CURSEG_HOT_NODE], |
472 | si->valid_blks[CURSEG_HOT_NODE]); |
473 | seq_printf(m: s, fmt: " - File dnode: %8d %8d %8d %10u %10u %10u\n" , |
474 | si->curseg[CURSEG_WARM_NODE], |
475 | si->cursec[CURSEG_WARM_NODE], |
476 | si->curzone[CURSEG_WARM_NODE], |
477 | si->dirty_seg[CURSEG_WARM_NODE], |
478 | si->full_seg[CURSEG_WARM_NODE], |
479 | si->valid_blks[CURSEG_WARM_NODE]); |
480 | seq_printf(m: s, fmt: " - Indir nodes: %8d %8d %8d %10u %10u %10u\n" , |
481 | si->curseg[CURSEG_COLD_NODE], |
482 | si->cursec[CURSEG_COLD_NODE], |
483 | si->curzone[CURSEG_COLD_NODE], |
484 | si->dirty_seg[CURSEG_COLD_NODE], |
485 | si->full_seg[CURSEG_COLD_NODE], |
486 | si->valid_blks[CURSEG_COLD_NODE]); |
487 | seq_printf(m: s, fmt: " - Pinned file: %8d %8d %8d\n" , |
488 | si->curseg[CURSEG_COLD_DATA_PINNED], |
489 | si->cursec[CURSEG_COLD_DATA_PINNED], |
490 | si->curzone[CURSEG_COLD_DATA_PINNED]); |
491 | seq_printf(m: s, fmt: " - ATGC data: %8d %8d %8d\n" , |
492 | si->curseg[CURSEG_ALL_DATA_ATGC], |
493 | si->cursec[CURSEG_ALL_DATA_ATGC], |
494 | si->curzone[CURSEG_ALL_DATA_ATGC]); |
495 | seq_printf(m: s, fmt: "\n - Valid: %d\n - Dirty: %d\n" , |
496 | si->main_area_segs - si->dirty_count - |
497 | si->prefree_count - si->free_segs, |
498 | si->dirty_count); |
499 | seq_printf(m: s, fmt: " - Prefree: %d\n - Free: %d (%d)\n\n" , |
500 | si->prefree_count, si->free_segs, si->free_secs); |
501 | seq_printf(m: s, fmt: "CP calls: %d (BG: %d)\n" , |
502 | si->cp_call_count[TOTAL_CALL], |
503 | si->cp_call_count[BACKGROUND]); |
504 | seq_printf(m: s, fmt: "CP count: %d\n" , si->cp_count); |
505 | seq_printf(m: s, fmt: " - cp blocks : %u\n" , si->meta_count[META_CP]); |
506 | seq_printf(m: s, fmt: " - sit blocks : %u\n" , |
507 | si->meta_count[META_SIT]); |
508 | seq_printf(m: s, fmt: " - nat blocks : %u\n" , |
509 | si->meta_count[META_NAT]); |
510 | seq_printf(m: s, fmt: " - ssa blocks : %u\n" , |
511 | si->meta_count[META_SSA]); |
512 | seq_puts(m: s, s: "CP merge:\n" ); |
513 | seq_printf(m: s, fmt: " - Queued : %4d\n" , si->nr_queued_ckpt); |
514 | seq_printf(m: s, fmt: " - Issued : %4d\n" , si->nr_issued_ckpt); |
515 | seq_printf(m: s, fmt: " - Total : %4d\n" , si->nr_total_ckpt); |
516 | seq_printf(m: s, fmt: " - Cur time : %4d(ms)\n" , si->cur_ckpt_time); |
517 | seq_printf(m: s, fmt: " - Peak time : %4d(ms)\n" , si->peak_ckpt_time); |
518 | seq_printf(m: s, fmt: "GC calls: %d (gc_thread: %d)\n" , |
519 | si->gc_call_count[BACKGROUND] + |
520 | si->gc_call_count[FOREGROUND], |
521 | si->gc_call_count[BACKGROUND]); |
522 | if (__is_large_section(sbi)) { |
523 | seq_printf(m: s, fmt: " - data sections : %d (BG: %d)\n" , |
524 | si->gc_secs[DATA][BG_GC] + si->gc_secs[DATA][FG_GC], |
525 | si->gc_secs[DATA][BG_GC]); |
526 | seq_printf(m: s, fmt: " - node sections : %d (BG: %d)\n" , |
527 | si->gc_secs[NODE][BG_GC] + si->gc_secs[NODE][FG_GC], |
528 | si->gc_secs[NODE][BG_GC]); |
529 | } |
530 | seq_printf(m: s, fmt: " - data segments : %d (BG: %d)\n" , |
531 | si->gc_segs[DATA][BG_GC] + si->gc_segs[DATA][FG_GC], |
532 | si->gc_segs[DATA][BG_GC]); |
533 | seq_printf(m: s, fmt: " - node segments : %d (BG: %d)\n" , |
534 | si->gc_segs[NODE][BG_GC] + si->gc_segs[NODE][FG_GC], |
535 | si->gc_segs[NODE][BG_GC]); |
536 | seq_puts(m: s, s: " - Reclaimed segs :\n" ); |
537 | seq_printf(m: s, fmt: " - Normal : %d\n" , sbi->gc_reclaimed_segs[GC_NORMAL]); |
538 | seq_printf(m: s, fmt: " - Idle CB : %d\n" , sbi->gc_reclaimed_segs[GC_IDLE_CB]); |
539 | seq_printf(m: s, fmt: " - Idle Greedy : %d\n" , |
540 | sbi->gc_reclaimed_segs[GC_IDLE_GREEDY]); |
541 | seq_printf(m: s, fmt: " - Idle AT : %d\n" , sbi->gc_reclaimed_segs[GC_IDLE_AT]); |
542 | seq_printf(m: s, fmt: " - Urgent High : %d\n" , |
543 | sbi->gc_reclaimed_segs[GC_URGENT_HIGH]); |
544 | seq_printf(m: s, fmt: " - Urgent Mid : %d\n" , sbi->gc_reclaimed_segs[GC_URGENT_MID]); |
545 | seq_printf(m: s, fmt: " - Urgent Low : %d\n" , sbi->gc_reclaimed_segs[GC_URGENT_LOW]); |
546 | seq_printf(m: s, fmt: "Try to move %d blocks (BG: %d)\n" , si->tot_blks, |
547 | si->bg_data_blks + si->bg_node_blks); |
548 | seq_printf(m: s, fmt: " - data blocks : %d (%d)\n" , si->data_blks, |
549 | si->bg_data_blks); |
550 | seq_printf(m: s, fmt: " - node blocks : %d (%d)\n" , si->node_blks, |
551 | si->bg_node_blks); |
552 | seq_printf(m: s, fmt: "BG skip : IO: %u, Other: %u\n" , |
553 | si->io_skip_bggc, si->other_skip_bggc); |
554 | seq_puts(m: s, s: "\nExtent Cache (Read):\n" ); |
555 | seq_printf(m: s, fmt: " - Hit Count: L1-1:%llu L1-2:%llu L2:%llu\n" , |
556 | si->hit_largest, si->hit_cached[EX_READ], |
557 | si->hit_rbtree[EX_READ]); |
558 | seq_printf(m: s, fmt: " - Hit Ratio: %llu%% (%llu / %llu)\n" , |
559 | !si->total_ext[EX_READ] ? 0 : |
560 | div64_u64(dividend: si->hit_total[EX_READ] * 100, |
561 | divisor: si->total_ext[EX_READ]), |
562 | si->hit_total[EX_READ], si->total_ext[EX_READ]); |
563 | seq_printf(m: s, fmt: " - Inner Struct Count: tree: %d(%d), node: %d\n" , |
564 | si->ext_tree[EX_READ], si->zombie_tree[EX_READ], |
565 | si->ext_node[EX_READ]); |
566 | seq_puts(m: s, s: "\nExtent Cache (Block Age):\n" ); |
567 | seq_printf(m: s, fmt: " - Allocated Data Blocks: %llu\n" , |
568 | si->allocated_data_blocks); |
569 | seq_printf(m: s, fmt: " - Hit Count: L1:%llu L2:%llu\n" , |
570 | si->hit_cached[EX_BLOCK_AGE], |
571 | si->hit_rbtree[EX_BLOCK_AGE]); |
572 | seq_printf(m: s, fmt: " - Hit Ratio: %llu%% (%llu / %llu)\n" , |
573 | !si->total_ext[EX_BLOCK_AGE] ? 0 : |
574 | div64_u64(dividend: si->hit_total[EX_BLOCK_AGE] * 100, |
575 | divisor: si->total_ext[EX_BLOCK_AGE]), |
576 | si->hit_total[EX_BLOCK_AGE], |
577 | si->total_ext[EX_BLOCK_AGE]); |
578 | seq_printf(m: s, fmt: " - Inner Struct Count: tree: %d(%d), node: %d\n" , |
579 | si->ext_tree[EX_BLOCK_AGE], |
580 | si->zombie_tree[EX_BLOCK_AGE], |
581 | si->ext_node[EX_BLOCK_AGE]); |
582 | seq_puts(m: s, s: "\nBalancing F2FS Async:\n" ); |
583 | seq_printf(m: s, fmt: " - DIO (R: %4d, W: %4d)\n" , |
584 | si->nr_dio_read, si->nr_dio_write); |
585 | seq_printf(m: s, fmt: " - IO_R (Data: %4d, Node: %4d, Meta: %4d\n" , |
586 | si->nr_rd_data, si->nr_rd_node, si->nr_rd_meta); |
587 | seq_printf(m: s, fmt: " - IO_W (CP: %4d, Data: %4d, Flush: (%4d %4d %4d), " , |
588 | si->nr_wb_cp_data, si->nr_wb_data, |
589 | si->nr_flushing, si->nr_flushed, |
590 | si->flush_list_empty); |
591 | seq_printf(m: s, fmt: "Discard: (%4d %4d)) cmd: %4d undiscard:%4u\n" , |
592 | si->nr_discarding, si->nr_discarded, |
593 | si->nr_discard_cmd, si->undiscard_blks); |
594 | seq_printf(m: s, fmt: " - atomic IO: %4d (Max. %4d)\n" , |
595 | si->aw_cnt, si->max_aw_cnt); |
596 | seq_printf(m: s, fmt: " - compress: %4d, hit:%8d\n" , si->compress_pages, si->compress_page_hit); |
597 | seq_printf(m: s, fmt: " - nodes: %4d in %4d\n" , |
598 | si->ndirty_node, si->node_pages); |
599 | seq_printf(m: s, fmt: " - dents: %4d in dirs:%4d (%4d)\n" , |
600 | si->ndirty_dent, si->ndirty_dirs, si->ndirty_all); |
601 | seq_printf(m: s, fmt: " - datas: %4d in files:%4d\n" , |
602 | si->ndirty_data, si->ndirty_files); |
603 | seq_printf(m: s, fmt: " - quota datas: %4d in quota files:%4d\n" , |
604 | si->ndirty_qdata, si->nquota_files); |
605 | seq_printf(m: s, fmt: " - meta: %4d in %4d\n" , |
606 | si->ndirty_meta, si->meta_pages); |
607 | seq_printf(m: s, fmt: " - imeta: %4d\n" , |
608 | si->ndirty_imeta); |
609 | seq_printf(m: s, fmt: " - fsync mark: %4lld\n" , |
610 | percpu_counter_sum_positive( |
611 | fbc: &sbi->rf_node_block_count)); |
612 | seq_printf(m: s, fmt: " - NATs: %9d/%9d\n - SITs: %9d/%9d\n" , |
613 | si->dirty_nats, si->nats, si->dirty_sits, si->sits); |
614 | seq_printf(m: s, fmt: " - free_nids: %9d/%9d\n - alloc_nids: %9d\n" , |
615 | si->free_nids, si->avail_nids, si->alloc_nids); |
616 | seq_puts(m: s, s: "\nDistribution of User Blocks:" ); |
617 | seq_puts(m: s, s: " [ valid | invalid | free ]\n" ); |
618 | seq_puts(m: s, s: " [" ); |
619 | |
620 | for (j = 0; j < si->util_valid; j++) |
621 | seq_putc(m: s, c: '-'); |
622 | seq_putc(m: s, c: '|'); |
623 | |
624 | for (j = 0; j < si->util_invalid; j++) |
625 | seq_putc(m: s, c: '-'); |
626 | seq_putc(m: s, c: '|'); |
627 | |
628 | for (j = 0; j < si->util_free; j++) |
629 | seq_putc(m: s, c: '-'); |
630 | seq_puts(m: s, s: "]\n\n" ); |
631 | seq_printf(m: s, fmt: "IPU: %u blocks\n" , si->inplace_count); |
632 | seq_printf(m: s, fmt: "SSR: %u blocks in %u segments\n" , |
633 | si->block_count[SSR], si->segment_count[SSR]); |
634 | seq_printf(m: s, fmt: "LFS: %u blocks in %u segments\n" , |
635 | si->block_count[LFS], si->segment_count[LFS]); |
636 | |
637 | /* segment usage info */ |
638 | f2fs_update_sit_info(sbi); |
639 | seq_printf(m: s, fmt: "\nBDF: %u, avg. vblocks: %u\n" , |
640 | si->bimodal, si->avg_vblocks); |
641 | |
642 | /* memory footprint */ |
643 | update_mem_info(sbi); |
644 | seq_printf(m: s, fmt: "\nMemory: %llu KB\n" , |
645 | (si->base_mem + si->cache_mem + si->page_mem) >> 10); |
646 | seq_printf(m: s, fmt: " - static: %llu KB\n" , |
647 | si->base_mem >> 10); |
648 | seq_printf(m: s, fmt: " - cached all: %llu KB\n" , |
649 | si->cache_mem >> 10); |
650 | seq_printf(m: s, fmt: " - read extent cache: %llu KB\n" , |
651 | si->ext_mem[EX_READ] >> 10); |
652 | seq_printf(m: s, fmt: " - block age extent cache: %llu KB\n" , |
653 | si->ext_mem[EX_BLOCK_AGE] >> 10); |
654 | seq_printf(m: s, fmt: " - paged : %llu KB\n" , |
655 | si->page_mem >> 10); |
656 | } |
657 | raw_spin_unlock_irqrestore(&f2fs_stat_lock, flags); |
658 | return 0; |
659 | } |
660 | |
661 | DEFINE_SHOW_ATTRIBUTE(stat); |
662 | #endif |
663 | |
664 | int f2fs_build_stats(struct f2fs_sb_info *sbi) |
665 | { |
666 | struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi); |
667 | struct f2fs_stat_info *si; |
668 | unsigned long flags; |
669 | int i; |
670 | |
671 | si = f2fs_kzalloc(sbi, size: sizeof(struct f2fs_stat_info), GFP_KERNEL); |
672 | if (!si) |
673 | return -ENOMEM; |
674 | |
675 | si->all_area_segs = le32_to_cpu(raw_super->segment_count); |
676 | si->sit_area_segs = le32_to_cpu(raw_super->segment_count_sit); |
677 | si->nat_area_segs = le32_to_cpu(raw_super->segment_count_nat); |
678 | si->ssa_area_segs = le32_to_cpu(raw_super->segment_count_ssa); |
679 | si->main_area_segs = le32_to_cpu(raw_super->segment_count_main); |
680 | si->main_area_sections = le32_to_cpu(raw_super->section_count); |
681 | si->main_area_zones = si->main_area_sections / |
682 | le32_to_cpu(raw_super->secs_per_zone); |
683 | si->sbi = sbi; |
684 | sbi->stat_info = si; |
685 | |
686 | /* general extent cache stats */ |
687 | for (i = 0; i < NR_EXTENT_CACHES; i++) { |
688 | atomic64_set(v: &sbi->total_hit_ext[i], i: 0); |
689 | atomic64_set(v: &sbi->read_hit_rbtree[i], i: 0); |
690 | atomic64_set(v: &sbi->read_hit_cached[i], i: 0); |
691 | } |
692 | |
693 | /* read extent_cache only */ |
694 | atomic64_set(v: &sbi->read_hit_largest, i: 0); |
695 | |
696 | atomic_set(v: &sbi->inline_xattr, i: 0); |
697 | atomic_set(v: &sbi->inline_inode, i: 0); |
698 | atomic_set(v: &sbi->inline_dir, i: 0); |
699 | atomic_set(v: &sbi->compr_inode, i: 0); |
700 | atomic64_set(v: &sbi->compr_blocks, i: 0); |
701 | atomic_set(v: &sbi->swapfile_inode, i: 0); |
702 | atomic_set(v: &sbi->atomic_files, i: 0); |
703 | atomic_set(v: &sbi->inplace_count, i: 0); |
704 | for (i = META_CP; i < META_MAX; i++) |
705 | atomic_set(v: &sbi->meta_count[i], i: 0); |
706 | for (i = 0; i < MAX_CALL_TYPE; i++) |
707 | atomic_set(v: &sbi->cp_call_count[i], i: 0); |
708 | |
709 | atomic_set(v: &sbi->max_aw_cnt, i: 0); |
710 | |
711 | raw_spin_lock_irqsave(&f2fs_stat_lock, flags); |
712 | list_add_tail(new: &si->stat_list, head: &f2fs_stat_list); |
713 | raw_spin_unlock_irqrestore(&f2fs_stat_lock, flags); |
714 | |
715 | return 0; |
716 | } |
717 | |
718 | void f2fs_destroy_stats(struct f2fs_sb_info *sbi) |
719 | { |
720 | struct f2fs_stat_info *si = F2FS_STAT(sbi); |
721 | unsigned long flags; |
722 | |
723 | raw_spin_lock_irqsave(&f2fs_stat_lock, flags); |
724 | list_del(entry: &si->stat_list); |
725 | raw_spin_unlock_irqrestore(&f2fs_stat_lock, flags); |
726 | |
727 | kfree(objp: si); |
728 | } |
729 | |
730 | void __init f2fs_create_root_stats(void) |
731 | { |
732 | #ifdef CONFIG_DEBUG_FS |
733 | f2fs_debugfs_root = debugfs_create_dir(name: "f2fs" , NULL); |
734 | |
735 | debugfs_create_file(name: "status" , mode: 0444, parent: f2fs_debugfs_root, NULL, |
736 | fops: &stat_fops); |
737 | #endif |
738 | } |
739 | |
740 | void f2fs_destroy_root_stats(void) |
741 | { |
742 | #ifdef CONFIG_DEBUG_FS |
743 | debugfs_remove_recursive(dentry: f2fs_debugfs_root); |
744 | f2fs_debugfs_root = NULL; |
745 | #endif |
746 | } |
747 | |