1 | // SPDX-License-Identifier: GPL-2.0-or-later |
2 | /* |
3 | * resize.c |
4 | * |
5 | * volume resize. |
6 | * Inspired by ext3/resize.c. |
7 | * |
8 | * Copyright (C) 2007 Oracle. All rights reserved. |
9 | */ |
10 | |
11 | #include <linux/fs.h> |
12 | #include <linux/types.h> |
13 | |
14 | #include <cluster/masklog.h> |
15 | |
16 | #include "ocfs2.h" |
17 | |
18 | #include "alloc.h" |
19 | #include "dlmglue.h" |
20 | #include "inode.h" |
21 | #include "journal.h" |
22 | #include "super.h" |
23 | #include "sysfile.h" |
24 | #include "uptodate.h" |
25 | #include "ocfs2_trace.h" |
26 | |
27 | #include "buffer_head_io.h" |
28 | #include "suballoc.h" |
29 | #include "resize.h" |
30 | |
31 | /* |
32 | * Check whether there are new backup superblocks exist |
33 | * in the last group. If there are some, mark them or clear |
34 | * them in the bitmap. |
35 | * |
36 | * Return how many backups we find in the last group. |
37 | */ |
38 | static u16 ocfs2_calc_new_backup_super(struct inode *inode, |
39 | struct ocfs2_group_desc *gd, |
40 | u16 cl_cpg, |
41 | u16 old_bg_clusters, |
42 | int set) |
43 | { |
44 | int i; |
45 | u16 backups = 0; |
46 | u32 cluster, lgd_cluster; |
47 | u64 blkno, gd_blkno, lgd_blkno = le64_to_cpu(gd->bg_blkno); |
48 | |
49 | for (i = 0; i < OCFS2_MAX_BACKUP_SUPERBLOCKS; i++) { |
50 | blkno = ocfs2_backup_super_blkno(sb: inode->i_sb, index: i); |
51 | cluster = ocfs2_blocks_to_clusters(sb: inode->i_sb, blocks: blkno); |
52 | |
53 | gd_blkno = ocfs2_which_cluster_group(inode, cluster); |
54 | if (gd_blkno < lgd_blkno) |
55 | continue; |
56 | else if (gd_blkno > lgd_blkno) |
57 | break; |
58 | |
59 | /* check if already done backup super */ |
60 | lgd_cluster = ocfs2_blocks_to_clusters(sb: inode->i_sb, blocks: lgd_blkno); |
61 | lgd_cluster += old_bg_clusters; |
62 | if (lgd_cluster >= cluster) |
63 | continue; |
64 | |
65 | if (set) |
66 | ocfs2_set_bit(cluster % cl_cpg, |
67 | (unsigned long *)gd->bg_bitmap); |
68 | else |
69 | ocfs2_clear_bit(cluster % cl_cpg, |
70 | (unsigned long *)gd->bg_bitmap); |
71 | backups++; |
72 | } |
73 | |
74 | return backups; |
75 | } |
76 | |
77 | static int ocfs2_update_last_group_and_inode(handle_t *handle, |
78 | struct inode *bm_inode, |
79 | struct buffer_head *bm_bh, |
80 | struct buffer_head *group_bh, |
81 | u32 first_new_cluster, |
82 | int new_clusters) |
83 | { |
84 | int ret = 0; |
85 | struct ocfs2_super *osb = OCFS2_SB(bm_inode->i_sb); |
86 | struct ocfs2_dinode *fe = (struct ocfs2_dinode *) bm_bh->b_data; |
87 | struct ocfs2_chain_list *cl = &fe->id2.i_chain; |
88 | struct ocfs2_chain_rec *cr; |
89 | struct ocfs2_group_desc *group; |
90 | u16 chain, num_bits, backups = 0; |
91 | u16 cl_bpc = le16_to_cpu(cl->cl_bpc); |
92 | u16 cl_cpg = le16_to_cpu(cl->cl_cpg); |
93 | u16 old_bg_clusters; |
94 | |
95 | trace_ocfs2_update_last_group_and_inode(val1: new_clusters, |
96 | val2: first_new_cluster); |
97 | |
98 | ret = ocfs2_journal_access_gd(handle, ci: INODE_CACHE(inode: bm_inode), |
99 | bh: group_bh, OCFS2_JOURNAL_ACCESS_WRITE); |
100 | if (ret < 0) { |
101 | mlog_errno(ret); |
102 | goto out; |
103 | } |
104 | |
105 | group = (struct ocfs2_group_desc *)group_bh->b_data; |
106 | |
107 | old_bg_clusters = le16_to_cpu(group->bg_bits) / cl_bpc; |
108 | /* update the group first. */ |
109 | num_bits = new_clusters * cl_bpc; |
110 | le16_add_cpu(var: &group->bg_bits, val: num_bits); |
111 | le16_add_cpu(var: &group->bg_free_bits_count, val: num_bits); |
112 | |
113 | /* |
114 | * check whether there are some new backup superblocks exist in |
115 | * this group and update the group bitmap accordingly. |
116 | */ |
117 | if (OCFS2_HAS_COMPAT_FEATURE(osb->sb, |
118 | OCFS2_FEATURE_COMPAT_BACKUP_SB)) { |
119 | backups = ocfs2_calc_new_backup_super(inode: bm_inode, |
120 | gd: group, |
121 | cl_cpg, old_bg_clusters, set: 1); |
122 | le16_add_cpu(var: &group->bg_free_bits_count, val: -1 * backups); |
123 | } |
124 | |
125 | ocfs2_journal_dirty(handle, bh: group_bh); |
126 | |
127 | /* update the inode accordingly. */ |
128 | ret = ocfs2_journal_access_di(handle, ci: INODE_CACHE(inode: bm_inode), bh: bm_bh, |
129 | OCFS2_JOURNAL_ACCESS_WRITE); |
130 | if (ret < 0) { |
131 | mlog_errno(ret); |
132 | goto out_rollback; |
133 | } |
134 | |
135 | chain = le16_to_cpu(group->bg_chain); |
136 | cr = (&cl->cl_recs[chain]); |
137 | le32_add_cpu(var: &cr->c_total, val: num_bits); |
138 | le32_add_cpu(var: &cr->c_free, val: num_bits); |
139 | le32_add_cpu(var: &fe->id1.bitmap1.i_total, val: num_bits); |
140 | le32_add_cpu(var: &fe->i_clusters, val: new_clusters); |
141 | |
142 | if (backups) { |
143 | le32_add_cpu(var: &cr->c_free, val: -1 * backups); |
144 | le32_add_cpu(var: &fe->id1.bitmap1.i_used, val: backups); |
145 | } |
146 | |
147 | spin_lock(lock: &OCFS2_I(inode: bm_inode)->ip_lock); |
148 | OCFS2_I(inode: bm_inode)->ip_clusters = le32_to_cpu(fe->i_clusters); |
149 | le64_add_cpu(var: &fe->i_size, val: (u64)new_clusters << osb->s_clustersize_bits); |
150 | spin_unlock(lock: &OCFS2_I(inode: bm_inode)->ip_lock); |
151 | i_size_write(inode: bm_inode, le64_to_cpu(fe->i_size)); |
152 | |
153 | ocfs2_journal_dirty(handle, bh: bm_bh); |
154 | |
155 | out_rollback: |
156 | if (ret < 0) { |
157 | ocfs2_calc_new_backup_super(inode: bm_inode, |
158 | gd: group, |
159 | cl_cpg, old_bg_clusters, set: 0); |
160 | le16_add_cpu(var: &group->bg_free_bits_count, val: backups); |
161 | le16_add_cpu(var: &group->bg_bits, val: -1 * num_bits); |
162 | le16_add_cpu(var: &group->bg_free_bits_count, val: -1 * num_bits); |
163 | } |
164 | out: |
165 | if (ret) |
166 | mlog_errno(ret); |
167 | return ret; |
168 | } |
169 | |
170 | static int update_backups(struct inode * inode, u32 clusters, char *data) |
171 | { |
172 | int i, ret = 0; |
173 | u32 cluster; |
174 | u64 blkno; |
175 | struct buffer_head *backup = NULL; |
176 | struct ocfs2_dinode *backup_di = NULL; |
177 | struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); |
178 | |
179 | /* calculate the real backups we need to update. */ |
180 | for (i = 0; i < OCFS2_MAX_BACKUP_SUPERBLOCKS; i++) { |
181 | blkno = ocfs2_backup_super_blkno(sb: inode->i_sb, index: i); |
182 | cluster = ocfs2_blocks_to_clusters(sb: inode->i_sb, blocks: blkno); |
183 | if (cluster >= clusters) |
184 | break; |
185 | |
186 | ret = ocfs2_read_blocks_sync(osb, block: blkno, nr: 1, bhs: &backup); |
187 | if (ret < 0) { |
188 | mlog_errno(ret); |
189 | break; |
190 | } |
191 | |
192 | memcpy(backup->b_data, data, inode->i_sb->s_blocksize); |
193 | |
194 | backup_di = (struct ocfs2_dinode *)backup->b_data; |
195 | backup_di->i_blkno = cpu_to_le64(blkno); |
196 | |
197 | ret = ocfs2_write_super_or_backup(osb, bh: backup); |
198 | brelse(bh: backup); |
199 | backup = NULL; |
200 | if (ret < 0) { |
201 | mlog_errno(ret); |
202 | break; |
203 | } |
204 | } |
205 | |
206 | return ret; |
207 | } |
208 | |
209 | static void ocfs2_update_super_and_backups(struct inode *inode, |
210 | int new_clusters) |
211 | { |
212 | int ret; |
213 | u32 clusters = 0; |
214 | struct buffer_head *super_bh = NULL; |
215 | struct ocfs2_dinode *super_di = NULL; |
216 | struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); |
217 | |
218 | /* |
219 | * update the superblock last. |
220 | * It doesn't matter if the write failed. |
221 | */ |
222 | ret = ocfs2_read_blocks_sync(osb, OCFS2_SUPER_BLOCK_BLKNO, nr: 1, |
223 | bhs: &super_bh); |
224 | if (ret < 0) { |
225 | mlog_errno(ret); |
226 | goto out; |
227 | } |
228 | |
229 | super_di = (struct ocfs2_dinode *)super_bh->b_data; |
230 | le32_add_cpu(var: &super_di->i_clusters, val: new_clusters); |
231 | clusters = le32_to_cpu(super_di->i_clusters); |
232 | |
233 | ret = ocfs2_write_super_or_backup(osb, bh: super_bh); |
234 | if (ret < 0) { |
235 | mlog_errno(ret); |
236 | goto out; |
237 | } |
238 | |
239 | if (OCFS2_HAS_COMPAT_FEATURE(osb->sb, OCFS2_FEATURE_COMPAT_BACKUP_SB)) |
240 | ret = update_backups(inode, clusters, data: super_bh->b_data); |
241 | |
242 | out: |
243 | brelse(bh: super_bh); |
244 | if (ret) |
245 | printk(KERN_WARNING "ocfs2: Failed to update super blocks on %s" |
246 | " during fs resize. This condition is not fatal," |
247 | " but fsck.ocfs2 should be run to fix it\n" , |
248 | osb->dev_str); |
249 | return; |
250 | } |
251 | |
252 | /* |
253 | * Extend the filesystem to the new number of clusters specified. This entry |
254 | * point is only used to extend the current filesystem to the end of the last |
255 | * existing group. |
256 | */ |
257 | int ocfs2_group_extend(struct inode * inode, int new_clusters) |
258 | { |
259 | int ret; |
260 | handle_t *handle; |
261 | struct buffer_head *main_bm_bh = NULL; |
262 | struct buffer_head *group_bh = NULL; |
263 | struct inode *main_bm_inode = NULL; |
264 | struct ocfs2_dinode *fe = NULL; |
265 | struct ocfs2_group_desc *group = NULL; |
266 | struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); |
267 | u16 cl_bpc; |
268 | u32 first_new_cluster; |
269 | u64 lgd_blkno; |
270 | |
271 | if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) |
272 | return -EROFS; |
273 | |
274 | if (new_clusters < 0) |
275 | return -EINVAL; |
276 | else if (new_clusters == 0) |
277 | return 0; |
278 | |
279 | main_bm_inode = ocfs2_get_system_file_inode(osb, |
280 | type: GLOBAL_BITMAP_SYSTEM_INODE, |
281 | OCFS2_INVALID_SLOT); |
282 | if (!main_bm_inode) { |
283 | ret = -EINVAL; |
284 | mlog_errno(ret); |
285 | goto out; |
286 | } |
287 | |
288 | inode_lock(inode: main_bm_inode); |
289 | |
290 | ret = ocfs2_inode_lock(main_bm_inode, &main_bm_bh, 1); |
291 | if (ret < 0) { |
292 | mlog_errno(ret); |
293 | goto out_mutex; |
294 | } |
295 | |
296 | fe = (struct ocfs2_dinode *)main_bm_bh->b_data; |
297 | |
298 | /* main_bm_bh is validated by inode read inside ocfs2_inode_lock(), |
299 | * so any corruption is a code bug. */ |
300 | BUG_ON(!OCFS2_IS_VALID_DINODE(fe)); |
301 | |
302 | if (le16_to_cpu(fe->id2.i_chain.cl_cpg) != |
303 | ocfs2_group_bitmap_size(sb: osb->sb, suballocator: 0, |
304 | feature_incompat: osb->s_feature_incompat) * 8) { |
305 | mlog(ML_ERROR, "The disk is too old and small. " |
306 | "Force to do offline resize." ); |
307 | ret = -EINVAL; |
308 | goto out_unlock; |
309 | } |
310 | |
311 | first_new_cluster = le32_to_cpu(fe->i_clusters); |
312 | lgd_blkno = ocfs2_which_cluster_group(inode: main_bm_inode, |
313 | cluster: first_new_cluster - 1); |
314 | |
315 | ret = ocfs2_read_group_descriptor(inode: main_bm_inode, di: fe, gd_blkno: lgd_blkno, |
316 | bh: &group_bh); |
317 | if (ret < 0) { |
318 | mlog_errno(ret); |
319 | goto out_unlock; |
320 | } |
321 | group = (struct ocfs2_group_desc *)group_bh->b_data; |
322 | |
323 | cl_bpc = le16_to_cpu(fe->id2.i_chain.cl_bpc); |
324 | if (le16_to_cpu(group->bg_bits) / cl_bpc + new_clusters > |
325 | le16_to_cpu(fe->id2.i_chain.cl_cpg)) { |
326 | ret = -EINVAL; |
327 | goto out_unlock; |
328 | } |
329 | |
330 | |
331 | trace_ocfs2_group_extend( |
332 | val1: (unsigned long long)le64_to_cpu(group->bg_blkno), val2: new_clusters); |
333 | |
334 | handle = ocfs2_start_trans(osb, OCFS2_GROUP_EXTEND_CREDITS); |
335 | if (IS_ERR(ptr: handle)) { |
336 | mlog_errno(PTR_ERR(handle)); |
337 | ret = -EINVAL; |
338 | goto out_unlock; |
339 | } |
340 | |
341 | /* update the last group descriptor and inode. */ |
342 | ret = ocfs2_update_last_group_and_inode(handle, bm_inode: main_bm_inode, |
343 | bm_bh: main_bm_bh, group_bh, |
344 | first_new_cluster, |
345 | new_clusters); |
346 | if (ret) { |
347 | mlog_errno(ret); |
348 | goto out_commit; |
349 | } |
350 | |
351 | ocfs2_update_super_and_backups(inode: main_bm_inode, new_clusters); |
352 | |
353 | out_commit: |
354 | ocfs2_commit_trans(osb, handle); |
355 | out_unlock: |
356 | brelse(bh: group_bh); |
357 | brelse(bh: main_bm_bh); |
358 | |
359 | ocfs2_inode_unlock(inode: main_bm_inode, ex: 1); |
360 | |
361 | out_mutex: |
362 | inode_unlock(inode: main_bm_inode); |
363 | iput(main_bm_inode); |
364 | |
365 | out: |
366 | return ret; |
367 | } |
368 | |
369 | static int ocfs2_check_new_group(struct inode *inode, |
370 | struct ocfs2_dinode *di, |
371 | struct ocfs2_new_group_input *input, |
372 | struct buffer_head *group_bh) |
373 | { |
374 | int ret; |
375 | struct ocfs2_group_desc *gd = |
376 | (struct ocfs2_group_desc *)group_bh->b_data; |
377 | u16 cl_bpc = le16_to_cpu(di->id2.i_chain.cl_bpc); |
378 | |
379 | ret = ocfs2_check_group_descriptor(sb: inode->i_sb, di, bh: group_bh); |
380 | if (ret) |
381 | goto out; |
382 | |
383 | ret = -EINVAL; |
384 | if (le16_to_cpu(gd->bg_chain) != input->chain) |
385 | mlog(ML_ERROR, "Group descriptor # %llu has bad chain %u " |
386 | "while input has %u set.\n" , |
387 | (unsigned long long)le64_to_cpu(gd->bg_blkno), |
388 | le16_to_cpu(gd->bg_chain), input->chain); |
389 | else if (le16_to_cpu(gd->bg_bits) != input->clusters * cl_bpc) |
390 | mlog(ML_ERROR, "Group descriptor # %llu has bit count %u but " |
391 | "input has %u clusters set\n" , |
392 | (unsigned long long)le64_to_cpu(gd->bg_blkno), |
393 | le16_to_cpu(gd->bg_bits), input->clusters); |
394 | else if (le16_to_cpu(gd->bg_free_bits_count) != input->frees * cl_bpc) |
395 | mlog(ML_ERROR, "Group descriptor # %llu has free bit count %u " |
396 | "but it should have %u set\n" , |
397 | (unsigned long long)le64_to_cpu(gd->bg_blkno), |
398 | le16_to_cpu(gd->bg_bits), |
399 | input->frees * cl_bpc); |
400 | else |
401 | ret = 0; |
402 | |
403 | out: |
404 | return ret; |
405 | } |
406 | |
407 | static int ocfs2_verify_group_and_input(struct inode *inode, |
408 | struct ocfs2_dinode *di, |
409 | struct ocfs2_new_group_input *input, |
410 | struct buffer_head *group_bh) |
411 | { |
412 | u16 cl_count = le16_to_cpu(di->id2.i_chain.cl_count); |
413 | u16 cl_cpg = le16_to_cpu(di->id2.i_chain.cl_cpg); |
414 | u16 next_free = le16_to_cpu(di->id2.i_chain.cl_next_free_rec); |
415 | u32 cluster = ocfs2_blocks_to_clusters(sb: inode->i_sb, blocks: input->group); |
416 | u32 total_clusters = le32_to_cpu(di->i_clusters); |
417 | int ret = -EINVAL; |
418 | |
419 | if (cluster < total_clusters) |
420 | mlog(ML_ERROR, "add a group which is in the current volume.\n" ); |
421 | else if (input->chain >= cl_count) |
422 | mlog(ML_ERROR, "input chain exceeds the limit.\n" ); |
423 | else if (next_free != cl_count && next_free != input->chain) |
424 | mlog(ML_ERROR, |
425 | "the add group should be in chain %u\n" , next_free); |
426 | else if (total_clusters + input->clusters < total_clusters) |
427 | mlog(ML_ERROR, "add group's clusters overflow.\n" ); |
428 | else if (input->clusters > cl_cpg) |
429 | mlog(ML_ERROR, "the cluster exceeds the maximum of a group\n" ); |
430 | else if (input->frees > input->clusters) |
431 | mlog(ML_ERROR, "the free cluster exceeds the total clusters\n" ); |
432 | else if (total_clusters % cl_cpg != 0) |
433 | mlog(ML_ERROR, |
434 | "the last group isn't full. Use group extend first.\n" ); |
435 | else if (input->group != ocfs2_which_cluster_group(inode, cluster)) |
436 | mlog(ML_ERROR, "group blkno is invalid\n" ); |
437 | else if ((ret = ocfs2_check_new_group(inode, di, input, group_bh))) |
438 | mlog(ML_ERROR, "group descriptor check failed.\n" ); |
439 | else |
440 | ret = 0; |
441 | |
442 | return ret; |
443 | } |
444 | |
445 | /* Add a new group descriptor to global_bitmap. */ |
446 | int ocfs2_group_add(struct inode *inode, struct ocfs2_new_group_input *input) |
447 | { |
448 | int ret; |
449 | handle_t *handle; |
450 | struct buffer_head *main_bm_bh = NULL; |
451 | struct inode *main_bm_inode = NULL; |
452 | struct ocfs2_dinode *fe = NULL; |
453 | struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); |
454 | struct buffer_head *group_bh = NULL; |
455 | struct ocfs2_group_desc *group = NULL; |
456 | struct ocfs2_chain_list *cl; |
457 | struct ocfs2_chain_rec *cr; |
458 | u16 cl_bpc; |
459 | u64 bg_ptr; |
460 | |
461 | if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) |
462 | return -EROFS; |
463 | |
464 | main_bm_inode = ocfs2_get_system_file_inode(osb, |
465 | type: GLOBAL_BITMAP_SYSTEM_INODE, |
466 | OCFS2_INVALID_SLOT); |
467 | if (!main_bm_inode) { |
468 | ret = -EINVAL; |
469 | mlog_errno(ret); |
470 | goto out; |
471 | } |
472 | |
473 | inode_lock(inode: main_bm_inode); |
474 | |
475 | ret = ocfs2_inode_lock(main_bm_inode, &main_bm_bh, 1); |
476 | if (ret < 0) { |
477 | mlog_errno(ret); |
478 | goto out_mutex; |
479 | } |
480 | |
481 | fe = (struct ocfs2_dinode *)main_bm_bh->b_data; |
482 | |
483 | if (le16_to_cpu(fe->id2.i_chain.cl_cpg) != |
484 | ocfs2_group_bitmap_size(sb: osb->sb, suballocator: 0, |
485 | feature_incompat: osb->s_feature_incompat) * 8) { |
486 | mlog(ML_ERROR, "The disk is too old and small." |
487 | " Force to do offline resize." ); |
488 | ret = -EINVAL; |
489 | goto out_unlock; |
490 | } |
491 | |
492 | ret = ocfs2_read_blocks_sync(osb, block: input->group, nr: 1, bhs: &group_bh); |
493 | if (ret < 0) { |
494 | mlog(ML_ERROR, "Can't read the group descriptor # %llu " |
495 | "from the device." , (unsigned long long)input->group); |
496 | goto out_unlock; |
497 | } |
498 | |
499 | ocfs2_set_new_buffer_uptodate(ci: INODE_CACHE(inode), bh: group_bh); |
500 | |
501 | ret = ocfs2_verify_group_and_input(inode: main_bm_inode, di: fe, input, group_bh); |
502 | if (ret) { |
503 | mlog_errno(ret); |
504 | goto out_free_group_bh; |
505 | } |
506 | |
507 | trace_ocfs2_group_add(ull: (unsigned long long)input->group, |
508 | value1: input->chain, value2: input->clusters, value3: input->frees); |
509 | |
510 | handle = ocfs2_start_trans(osb, OCFS2_GROUP_ADD_CREDITS); |
511 | if (IS_ERR(ptr: handle)) { |
512 | mlog_errno(PTR_ERR(handle)); |
513 | ret = -EINVAL; |
514 | goto out_free_group_bh; |
515 | } |
516 | |
517 | cl_bpc = le16_to_cpu(fe->id2.i_chain.cl_bpc); |
518 | cl = &fe->id2.i_chain; |
519 | cr = &cl->cl_recs[input->chain]; |
520 | |
521 | ret = ocfs2_journal_access_gd(handle, ci: INODE_CACHE(inode: main_bm_inode), |
522 | bh: group_bh, OCFS2_JOURNAL_ACCESS_WRITE); |
523 | if (ret < 0) { |
524 | mlog_errno(ret); |
525 | goto out_commit; |
526 | } |
527 | |
528 | group = (struct ocfs2_group_desc *)group_bh->b_data; |
529 | bg_ptr = le64_to_cpu(group->bg_next_group); |
530 | group->bg_next_group = cr->c_blkno; |
531 | ocfs2_journal_dirty(handle, bh: group_bh); |
532 | |
533 | ret = ocfs2_journal_access_di(handle, ci: INODE_CACHE(inode: main_bm_inode), |
534 | bh: main_bm_bh, OCFS2_JOURNAL_ACCESS_WRITE); |
535 | if (ret < 0) { |
536 | group->bg_next_group = cpu_to_le64(bg_ptr); |
537 | mlog_errno(ret); |
538 | goto out_commit; |
539 | } |
540 | |
541 | if (input->chain == le16_to_cpu(cl->cl_next_free_rec)) { |
542 | le16_add_cpu(var: &cl->cl_next_free_rec, val: 1); |
543 | memset(cr, 0, sizeof(struct ocfs2_chain_rec)); |
544 | } |
545 | |
546 | cr->c_blkno = cpu_to_le64(input->group); |
547 | le32_add_cpu(var: &cr->c_total, val: input->clusters * cl_bpc); |
548 | le32_add_cpu(var: &cr->c_free, val: input->frees * cl_bpc); |
549 | |
550 | le32_add_cpu(var: &fe->id1.bitmap1.i_total, val: input->clusters *cl_bpc); |
551 | le32_add_cpu(var: &fe->id1.bitmap1.i_used, |
552 | val: (input->clusters - input->frees) * cl_bpc); |
553 | le32_add_cpu(var: &fe->i_clusters, val: input->clusters); |
554 | |
555 | ocfs2_journal_dirty(handle, bh: main_bm_bh); |
556 | |
557 | spin_lock(lock: &OCFS2_I(inode: main_bm_inode)->ip_lock); |
558 | OCFS2_I(inode: main_bm_inode)->ip_clusters = le32_to_cpu(fe->i_clusters); |
559 | le64_add_cpu(var: &fe->i_size, val: (u64)input->clusters << osb->s_clustersize_bits); |
560 | spin_unlock(lock: &OCFS2_I(inode: main_bm_inode)->ip_lock); |
561 | i_size_write(inode: main_bm_inode, le64_to_cpu(fe->i_size)); |
562 | |
563 | ocfs2_update_super_and_backups(inode: main_bm_inode, new_clusters: input->clusters); |
564 | |
565 | out_commit: |
566 | ocfs2_commit_trans(osb, handle); |
567 | |
568 | out_free_group_bh: |
569 | brelse(bh: group_bh); |
570 | |
571 | out_unlock: |
572 | brelse(bh: main_bm_bh); |
573 | |
574 | ocfs2_inode_unlock(inode: main_bm_inode, ex: 1); |
575 | |
576 | out_mutex: |
577 | inode_unlock(inode: main_bm_inode); |
578 | iput(main_bm_inode); |
579 | |
580 | out: |
581 | return ret; |
582 | } |
583 | |