1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23#include <linux/kernel.h>
24#include <linux/fs.h>
25#include <linux/string.h>
26#include <linux/buffer_head.h>
27#include <linux/errno.h>
28#include <linux/nilfs2_fs.h>
29#include "mdt.h"
30#include "cpfile.h"
31
32
33static inline unsigned long
34nilfs_cpfile_checkpoints_per_block(const struct inode *cpfile)
35{
36 return NILFS_MDT(cpfile)->mi_entries_per_block;
37}
38
39
40static unsigned long
41nilfs_cpfile_get_blkoff(const struct inode *cpfile, __u64 cno)
42{
43 __u64 tcno = cno + NILFS_MDT(cpfile)->mi_first_entry_offset - 1;
44 do_div(tcno, nilfs_cpfile_checkpoints_per_block(cpfile));
45 return (unsigned long)tcno;
46}
47
48
49static unsigned long
50nilfs_cpfile_get_offset(const struct inode *cpfile, __u64 cno)
51{
52 __u64 tcno = cno + NILFS_MDT(cpfile)->mi_first_entry_offset - 1;
53 return do_div(tcno, nilfs_cpfile_checkpoints_per_block(cpfile));
54}
55
56static unsigned long
57nilfs_cpfile_checkpoints_in_block(const struct inode *cpfile,
58 __u64 curr,
59 __u64 max)
60{
61 return min_t(__u64,
62 nilfs_cpfile_checkpoints_per_block(cpfile) -
63 nilfs_cpfile_get_offset(cpfile, curr),
64 max - curr);
65}
66
67static inline int nilfs_cpfile_is_in_first(const struct inode *cpfile,
68 __u64 cno)
69{
70 return nilfs_cpfile_get_blkoff(cpfile, cno) == 0;
71}
72
73static unsigned int
74nilfs_cpfile_block_add_valid_checkpoints(const struct inode *cpfile,
75 struct buffer_head *bh,
76 void *kaddr,
77 unsigned int n)
78{
79 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh);
80 unsigned int count;
81
82 count = le32_to_cpu(cp->cp_checkpoints_count) + n;
83 cp->cp_checkpoints_count = cpu_to_le32(count);
84 return count;
85}
86
87static unsigned int
88nilfs_cpfile_block_sub_valid_checkpoints(const struct inode *cpfile,
89 struct buffer_head *bh,
90 void *kaddr,
91 unsigned int n)
92{
93 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh);
94 unsigned int count;
95
96 WARN_ON(le32_to_cpu(cp->cp_checkpoints_count) < n);
97 count = le32_to_cpu(cp->cp_checkpoints_count) - n;
98 cp->cp_checkpoints_count = cpu_to_le32(count);
99 return count;
100}
101
102static inline struct nilfs_cpfile_header *
103nilfs_cpfile_block_get_header(const struct inode *cpfile,
104 struct buffer_head *bh,
105 void *kaddr)
106{
107 return kaddr + bh_offset(bh);
108}
109
110static struct nilfs_checkpoint *
111nilfs_cpfile_block_get_checkpoint(const struct inode *cpfile, __u64 cno,
112 struct buffer_head *bh,
113 void *kaddr)
114{
115 return kaddr + bh_offset(bh) + nilfs_cpfile_get_offset(cpfile, cno) *
116 NILFS_MDT(cpfile)->mi_entry_size;
117}
118
119static void nilfs_cpfile_block_init(struct inode *cpfile,
120 struct buffer_head *bh,
121 void *kaddr)
122{
123 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh);
124 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size;
125 int n = nilfs_cpfile_checkpoints_per_block(cpfile);
126
127 while (n-- > 0) {
128 nilfs_checkpoint_set_invalid(cp);
129 cp = (void *)cp + cpsz;
130 }
131}
132
133static inline int nilfs_cpfile_get_header_block(struct inode *cpfile,
134 struct buffer_head **bhp)
135{
136 return nilfs_mdt_get_block(cpfile, 0, 0, NULL, bhp);
137}
138
139static inline int nilfs_cpfile_get_checkpoint_block(struct inode *cpfile,
140 __u64 cno,
141 int create,
142 struct buffer_head **bhp)
143{
144 return nilfs_mdt_get_block(cpfile,
145 nilfs_cpfile_get_blkoff(cpfile, cno),
146 create, nilfs_cpfile_block_init, bhp);
147}
148
149static inline int nilfs_cpfile_delete_checkpoint_block(struct inode *cpfile,
150 __u64 cno)
151{
152 return nilfs_mdt_delete_block(cpfile,
153 nilfs_cpfile_get_blkoff(cpfile, cno));
154}
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181int nilfs_cpfile_get_checkpoint(struct inode *cpfile,
182 __u64 cno,
183 int create,
184 struct nilfs_checkpoint **cpp,
185 struct buffer_head **bhp)
186{
187 struct buffer_head *header_bh, *cp_bh;
188 struct nilfs_cpfile_header *header;
189 struct nilfs_checkpoint *cp;
190 void *kaddr;
191 int ret;
192
193 if (unlikely(cno < 1 || cno > nilfs_mdt_cno(cpfile) ||
194 (cno < nilfs_mdt_cno(cpfile) && create)))
195 return -EINVAL;
196
197 down_write(&NILFS_MDT(cpfile)->mi_sem);
198
199 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
200 if (ret < 0)
201 goto out_sem;
202 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, create, &cp_bh);
203 if (ret < 0)
204 goto out_header;
205 kaddr = kmap(cp_bh->b_page);
206 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
207 if (nilfs_checkpoint_invalid(cp)) {
208 if (!create) {
209 kunmap(cp_bh->b_page);
210 brelse(cp_bh);
211 ret = -ENOENT;
212 goto out_header;
213 }
214
215 nilfs_checkpoint_clear_invalid(cp);
216 if (!nilfs_cpfile_is_in_first(cpfile, cno))
217 nilfs_cpfile_block_add_valid_checkpoints(cpfile, cp_bh,
218 kaddr, 1);
219 mark_buffer_dirty(cp_bh);
220
221 kaddr = kmap_atomic(header_bh->b_page);
222 header = nilfs_cpfile_block_get_header(cpfile, header_bh,
223 kaddr);
224 le64_add_cpu(&header->ch_ncheckpoints, 1);
225 kunmap_atomic(kaddr);
226 mark_buffer_dirty(header_bh);
227 nilfs_mdt_mark_dirty(cpfile);
228 }
229
230 if (cpp != NULL)
231 *cpp = cp;
232 *bhp = cp_bh;
233
234 out_header:
235 brelse(header_bh);
236
237 out_sem:
238 up_write(&NILFS_MDT(cpfile)->mi_sem);
239 return ret;
240}
241
242
243
244
245
246
247
248
249
250
251
252void nilfs_cpfile_put_checkpoint(struct inode *cpfile, __u64 cno,
253 struct buffer_head *bh)
254{
255 kunmap(bh->b_page);
256 brelse(bh);
257}
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278int nilfs_cpfile_delete_checkpoints(struct inode *cpfile,
279 __u64 start,
280 __u64 end)
281{
282 struct buffer_head *header_bh, *cp_bh;
283 struct nilfs_cpfile_header *header;
284 struct nilfs_checkpoint *cp;
285 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size;
286 __u64 cno;
287 void *kaddr;
288 unsigned long tnicps;
289 int ret, ncps, nicps, count, i;
290
291 if (unlikely(start == 0 || start > end)) {
292 printk(KERN_ERR "%s: invalid range of checkpoint numbers: "
293 "[%llu, %llu)\n", __func__,
294 (unsigned long long)start, (unsigned long long)end);
295 return -EINVAL;
296 }
297
298 down_write(&NILFS_MDT(cpfile)->mi_sem);
299
300 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
301 if (ret < 0)
302 goto out_sem;
303 tnicps = 0;
304
305 for (cno = start; cno < end; cno += ncps) {
306 ncps = nilfs_cpfile_checkpoints_in_block(cpfile, cno, end);
307 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh);
308 if (ret < 0) {
309 if (ret != -ENOENT)
310 break;
311
312 ret = 0;
313 continue;
314 }
315
316 kaddr = kmap_atomic(cp_bh->b_page);
317 cp = nilfs_cpfile_block_get_checkpoint(
318 cpfile, cno, cp_bh, kaddr);
319 nicps = 0;
320 for (i = 0; i < ncps; i++, cp = (void *)cp + cpsz) {
321 WARN_ON(nilfs_checkpoint_snapshot(cp));
322 if (!nilfs_checkpoint_invalid(cp)) {
323 nilfs_checkpoint_set_invalid(cp);
324 nicps++;
325 }
326 }
327 if (nicps > 0) {
328 tnicps += nicps;
329 mark_buffer_dirty(cp_bh);
330 nilfs_mdt_mark_dirty(cpfile);
331 if (!nilfs_cpfile_is_in_first(cpfile, cno)) {
332 count =
333 nilfs_cpfile_block_sub_valid_checkpoints(
334 cpfile, cp_bh, kaddr, nicps);
335 if (count == 0) {
336
337 kunmap_atomic(kaddr);
338 brelse(cp_bh);
339 ret =
340 nilfs_cpfile_delete_checkpoint_block(
341 cpfile, cno);
342 if (ret == 0)
343 continue;
344 printk(KERN_ERR
345 "%s: cannot delete block\n",
346 __func__);
347 break;
348 }
349 }
350 }
351
352 kunmap_atomic(kaddr);
353 brelse(cp_bh);
354 }
355
356 if (tnicps > 0) {
357 kaddr = kmap_atomic(header_bh->b_page);
358 header = nilfs_cpfile_block_get_header(cpfile, header_bh,
359 kaddr);
360 le64_add_cpu(&header->ch_ncheckpoints, -(u64)tnicps);
361 mark_buffer_dirty(header_bh);
362 nilfs_mdt_mark_dirty(cpfile);
363 kunmap_atomic(kaddr);
364 }
365
366 brelse(header_bh);
367
368 out_sem:
369 up_write(&NILFS_MDT(cpfile)->mi_sem);
370 return ret;
371}
372
373static void nilfs_cpfile_checkpoint_to_cpinfo(struct inode *cpfile,
374 struct nilfs_checkpoint *cp,
375 struct nilfs_cpinfo *ci)
376{
377 ci->ci_flags = le32_to_cpu(cp->cp_flags);
378 ci->ci_cno = le64_to_cpu(cp->cp_cno);
379 ci->ci_create = le64_to_cpu(cp->cp_create);
380 ci->ci_nblk_inc = le64_to_cpu(cp->cp_nblk_inc);
381 ci->ci_inodes_count = le64_to_cpu(cp->cp_inodes_count);
382 ci->ci_blocks_count = le64_to_cpu(cp->cp_blocks_count);
383 ci->ci_next = le64_to_cpu(cp->cp_snapshot_list.ssl_next);
384}
385
386static ssize_t nilfs_cpfile_do_get_cpinfo(struct inode *cpfile, __u64 *cnop,
387 void *buf, unsigned cisz, size_t nci)
388{
389 struct nilfs_checkpoint *cp;
390 struct nilfs_cpinfo *ci = buf;
391 struct buffer_head *bh;
392 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size;
393 __u64 cur_cno = nilfs_mdt_cno(cpfile), cno = *cnop;
394 void *kaddr;
395 int n, ret;
396 int ncps, i;
397
398 if (cno == 0)
399 return -ENOENT;
400 down_read(&NILFS_MDT(cpfile)->mi_sem);
401
402 for (n = 0; cno < cur_cno && n < nci; cno += ncps) {
403 ncps = nilfs_cpfile_checkpoints_in_block(cpfile, cno, cur_cno);
404 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh);
405 if (ret < 0) {
406 if (ret != -ENOENT)
407 goto out;
408 continue;
409 }
410
411 kaddr = kmap_atomic(bh->b_page);
412 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr);
413 for (i = 0; i < ncps && n < nci; i++, cp = (void *)cp + cpsz) {
414 if (!nilfs_checkpoint_invalid(cp)) {
415 nilfs_cpfile_checkpoint_to_cpinfo(cpfile, cp,
416 ci);
417 ci = (void *)ci + cisz;
418 n++;
419 }
420 }
421 kunmap_atomic(kaddr);
422 brelse(bh);
423 }
424
425 ret = n;
426 if (n > 0) {
427 ci = (void *)ci - cisz;
428 *cnop = ci->ci_cno + 1;
429 }
430
431 out:
432 up_read(&NILFS_MDT(cpfile)->mi_sem);
433 return ret;
434}
435
436static ssize_t nilfs_cpfile_do_get_ssinfo(struct inode *cpfile, __u64 *cnop,
437 void *buf, unsigned cisz, size_t nci)
438{
439 struct buffer_head *bh;
440 struct nilfs_cpfile_header *header;
441 struct nilfs_checkpoint *cp;
442 struct nilfs_cpinfo *ci = buf;
443 __u64 curr = *cnop, next;
444 unsigned long curr_blkoff, next_blkoff;
445 void *kaddr;
446 int n = 0, ret;
447
448 down_read(&NILFS_MDT(cpfile)->mi_sem);
449
450 if (curr == 0) {
451 ret = nilfs_cpfile_get_header_block(cpfile, &bh);
452 if (ret < 0)
453 goto out;
454 kaddr = kmap_atomic(bh->b_page);
455 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr);
456 curr = le64_to_cpu(header->ch_snapshot_list.ssl_next);
457 kunmap_atomic(kaddr);
458 brelse(bh);
459 if (curr == 0) {
460 ret = 0;
461 goto out;
462 }
463 } else if (unlikely(curr == ~(__u64)0)) {
464 ret = 0;
465 goto out;
466 }
467
468 curr_blkoff = nilfs_cpfile_get_blkoff(cpfile, curr);
469 ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr, 0, &bh);
470 if (unlikely(ret < 0)) {
471 if (ret == -ENOENT)
472 ret = 0;
473 goto out;
474 }
475 kaddr = kmap_atomic(bh->b_page);
476 while (n < nci) {
477 cp = nilfs_cpfile_block_get_checkpoint(cpfile, curr, bh, kaddr);
478 curr = ~(__u64)0;
479 if (unlikely(nilfs_checkpoint_invalid(cp) ||
480 !nilfs_checkpoint_snapshot(cp)))
481 break;
482 nilfs_cpfile_checkpoint_to_cpinfo(cpfile, cp, ci);
483 ci = (void *)ci + cisz;
484 n++;
485 next = le64_to_cpu(cp->cp_snapshot_list.ssl_next);
486 if (next == 0)
487 break;
488
489 next_blkoff = nilfs_cpfile_get_blkoff(cpfile, next);
490 if (curr_blkoff != next_blkoff) {
491 kunmap_atomic(kaddr);
492 brelse(bh);
493 ret = nilfs_cpfile_get_checkpoint_block(cpfile, next,
494 0, &bh);
495 if (unlikely(ret < 0)) {
496 WARN_ON(ret == -ENOENT);
497 goto out;
498 }
499 kaddr = kmap_atomic(bh->b_page);
500 }
501 curr = next;
502 curr_blkoff = next_blkoff;
503 }
504 kunmap_atomic(kaddr);
505 brelse(bh);
506 *cnop = curr;
507 ret = n;
508
509 out:
510 up_read(&NILFS_MDT(cpfile)->mi_sem);
511 return ret;
512}
513
514
515
516
517
518
519
520
521
522ssize_t nilfs_cpfile_get_cpinfo(struct inode *cpfile, __u64 *cnop, int mode,
523 void *buf, unsigned cisz, size_t nci)
524{
525 switch (mode) {
526 case NILFS_CHECKPOINT:
527 return nilfs_cpfile_do_get_cpinfo(cpfile, cnop, buf, cisz, nci);
528 case NILFS_SNAPSHOT:
529 return nilfs_cpfile_do_get_ssinfo(cpfile, cnop, buf, cisz, nci);
530 default:
531 return -EINVAL;
532 }
533}
534
535
536
537
538
539
540int nilfs_cpfile_delete_checkpoint(struct inode *cpfile, __u64 cno)
541{
542 struct nilfs_cpinfo ci;
543 __u64 tcno = cno;
544 ssize_t nci;
545
546 nci = nilfs_cpfile_do_get_cpinfo(cpfile, &tcno, &ci, sizeof(ci), 1);
547 if (nci < 0)
548 return nci;
549 else if (nci == 0 || ci.ci_cno != cno)
550 return -ENOENT;
551 else if (nilfs_cpinfo_snapshot(&ci))
552 return -EBUSY;
553
554 return nilfs_cpfile_delete_checkpoints(cpfile, cno, cno + 1);
555}
556
557static struct nilfs_snapshot_list *
558nilfs_cpfile_block_get_snapshot_list(const struct inode *cpfile,
559 __u64 cno,
560 struct buffer_head *bh,
561 void *kaddr)
562{
563 struct nilfs_cpfile_header *header;
564 struct nilfs_checkpoint *cp;
565 struct nilfs_snapshot_list *list;
566
567 if (cno != 0) {
568 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr);
569 list = &cp->cp_snapshot_list;
570 } else {
571 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr);
572 list = &header->ch_snapshot_list;
573 }
574 return list;
575}
576
577static int nilfs_cpfile_set_snapshot(struct inode *cpfile, __u64 cno)
578{
579 struct buffer_head *header_bh, *curr_bh, *prev_bh, *cp_bh;
580 struct nilfs_cpfile_header *header;
581 struct nilfs_checkpoint *cp;
582 struct nilfs_snapshot_list *list;
583 __u64 curr, prev;
584 unsigned long curr_blkoff, prev_blkoff;
585 void *kaddr;
586 int ret;
587
588 if (cno == 0)
589 return -ENOENT;
590 down_write(&NILFS_MDT(cpfile)->mi_sem);
591
592 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh);
593 if (ret < 0)
594 goto out_sem;
595 kaddr = kmap_atomic(cp_bh->b_page);
596 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
597 if (nilfs_checkpoint_invalid(cp)) {
598 ret = -ENOENT;
599 kunmap_atomic(kaddr);
600 goto out_cp;
601 }
602 if (nilfs_checkpoint_snapshot(cp)) {
603 ret = 0;
604 kunmap_atomic(kaddr);
605 goto out_cp;
606 }
607 kunmap_atomic(kaddr);
608
609 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
610 if (ret < 0)
611 goto out_cp;
612 kaddr = kmap_atomic(header_bh->b_page);
613 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr);
614 list = &header->ch_snapshot_list;
615 curr_bh = header_bh;
616 get_bh(curr_bh);
617 curr = 0;
618 curr_blkoff = 0;
619 prev = le64_to_cpu(list->ssl_prev);
620 while (prev > cno) {
621 prev_blkoff = nilfs_cpfile_get_blkoff(cpfile, prev);
622 curr = prev;
623 if (curr_blkoff != prev_blkoff) {
624 kunmap_atomic(kaddr);
625 brelse(curr_bh);
626 ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr,
627 0, &curr_bh);
628 if (ret < 0)
629 goto out_header;
630 kaddr = kmap_atomic(curr_bh->b_page);
631 }
632 curr_blkoff = prev_blkoff;
633 cp = nilfs_cpfile_block_get_checkpoint(
634 cpfile, curr, curr_bh, kaddr);
635 list = &cp->cp_snapshot_list;
636 prev = le64_to_cpu(list->ssl_prev);
637 }
638 kunmap_atomic(kaddr);
639
640 if (prev != 0) {
641 ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0,
642 &prev_bh);
643 if (ret < 0)
644 goto out_curr;
645 } else {
646 prev_bh = header_bh;
647 get_bh(prev_bh);
648 }
649
650 kaddr = kmap_atomic(curr_bh->b_page);
651 list = nilfs_cpfile_block_get_snapshot_list(
652 cpfile, curr, curr_bh, kaddr);
653 list->ssl_prev = cpu_to_le64(cno);
654 kunmap_atomic(kaddr);
655
656 kaddr = kmap_atomic(cp_bh->b_page);
657 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
658 cp->cp_snapshot_list.ssl_next = cpu_to_le64(curr);
659 cp->cp_snapshot_list.ssl_prev = cpu_to_le64(prev);
660 nilfs_checkpoint_set_snapshot(cp);
661 kunmap_atomic(kaddr);
662
663 kaddr = kmap_atomic(prev_bh->b_page);
664 list = nilfs_cpfile_block_get_snapshot_list(
665 cpfile, prev, prev_bh, kaddr);
666 list->ssl_next = cpu_to_le64(cno);
667 kunmap_atomic(kaddr);
668
669 kaddr = kmap_atomic(header_bh->b_page);
670 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr);
671 le64_add_cpu(&header->ch_nsnapshots, 1);
672 kunmap_atomic(kaddr);
673
674 mark_buffer_dirty(prev_bh);
675 mark_buffer_dirty(curr_bh);
676 mark_buffer_dirty(cp_bh);
677 mark_buffer_dirty(header_bh);
678 nilfs_mdt_mark_dirty(cpfile);
679
680 brelse(prev_bh);
681
682 out_curr:
683 brelse(curr_bh);
684
685 out_header:
686 brelse(header_bh);
687
688 out_cp:
689 brelse(cp_bh);
690
691 out_sem:
692 up_write(&NILFS_MDT(cpfile)->mi_sem);
693 return ret;
694}
695
696static int nilfs_cpfile_clear_snapshot(struct inode *cpfile, __u64 cno)
697{
698 struct buffer_head *header_bh, *next_bh, *prev_bh, *cp_bh;
699 struct nilfs_cpfile_header *header;
700 struct nilfs_checkpoint *cp;
701 struct nilfs_snapshot_list *list;
702 __u64 next, prev;
703 void *kaddr;
704 int ret;
705
706 if (cno == 0)
707 return -ENOENT;
708 down_write(&NILFS_MDT(cpfile)->mi_sem);
709
710 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh);
711 if (ret < 0)
712 goto out_sem;
713 kaddr = kmap_atomic(cp_bh->b_page);
714 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
715 if (nilfs_checkpoint_invalid(cp)) {
716 ret = -ENOENT;
717 kunmap_atomic(kaddr);
718 goto out_cp;
719 }
720 if (!nilfs_checkpoint_snapshot(cp)) {
721 ret = 0;
722 kunmap_atomic(kaddr);
723 goto out_cp;
724 }
725
726 list = &cp->cp_snapshot_list;
727 next = le64_to_cpu(list->ssl_next);
728 prev = le64_to_cpu(list->ssl_prev);
729 kunmap_atomic(kaddr);
730
731 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
732 if (ret < 0)
733 goto out_cp;
734 if (next != 0) {
735 ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 0,
736 &next_bh);
737 if (ret < 0)
738 goto out_header;
739 } else {
740 next_bh = header_bh;
741 get_bh(next_bh);
742 }
743 if (prev != 0) {
744 ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0,
745 &prev_bh);
746 if (ret < 0)
747 goto out_next;
748 } else {
749 prev_bh = header_bh;
750 get_bh(prev_bh);
751 }
752
753 kaddr = kmap_atomic(next_bh->b_page);
754 list = nilfs_cpfile_block_get_snapshot_list(
755 cpfile, next, next_bh, kaddr);
756 list->ssl_prev = cpu_to_le64(prev);
757 kunmap_atomic(kaddr);
758
759 kaddr = kmap_atomic(prev_bh->b_page);
760 list = nilfs_cpfile_block_get_snapshot_list(
761 cpfile, prev, prev_bh, kaddr);
762 list->ssl_next = cpu_to_le64(next);
763 kunmap_atomic(kaddr);
764
765 kaddr = kmap_atomic(cp_bh->b_page);
766 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
767 cp->cp_snapshot_list.ssl_next = cpu_to_le64(0);
768 cp->cp_snapshot_list.ssl_prev = cpu_to_le64(0);
769 nilfs_checkpoint_clear_snapshot(cp);
770 kunmap_atomic(kaddr);
771
772 kaddr = kmap_atomic(header_bh->b_page);
773 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr);
774 le64_add_cpu(&header->ch_nsnapshots, -1);
775 kunmap_atomic(kaddr);
776
777 mark_buffer_dirty(next_bh);
778 mark_buffer_dirty(prev_bh);
779 mark_buffer_dirty(cp_bh);
780 mark_buffer_dirty(header_bh);
781 nilfs_mdt_mark_dirty(cpfile);
782
783 brelse(prev_bh);
784
785 out_next:
786 brelse(next_bh);
787
788 out_header:
789 brelse(header_bh);
790
791 out_cp:
792 brelse(cp_bh);
793
794 out_sem:
795 up_write(&NILFS_MDT(cpfile)->mi_sem);
796 return ret;
797}
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816int nilfs_cpfile_is_snapshot(struct inode *cpfile, __u64 cno)
817{
818 struct buffer_head *bh;
819 struct nilfs_checkpoint *cp;
820 void *kaddr;
821 int ret;
822
823
824
825 if (cno == 0 || cno >= nilfs_mdt_cno(cpfile))
826 return -ENOENT;
827 down_read(&NILFS_MDT(cpfile)->mi_sem);
828
829 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh);
830 if (ret < 0)
831 goto out;
832 kaddr = kmap_atomic(bh->b_page);
833 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr);
834 if (nilfs_checkpoint_invalid(cp))
835 ret = -ENOENT;
836 else
837 ret = nilfs_checkpoint_snapshot(cp);
838 kunmap_atomic(kaddr);
839 brelse(bh);
840
841 out:
842 up_read(&NILFS_MDT(cpfile)->mi_sem);
843 return ret;
844}
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864int nilfs_cpfile_change_cpmode(struct inode *cpfile, __u64 cno, int mode)
865{
866 int ret;
867
868 switch (mode) {
869 case NILFS_CHECKPOINT:
870 if (nilfs_checkpoint_is_mounted(cpfile->i_sb, cno))
871
872
873
874
875
876
877 ret = -EBUSY;
878 else
879 ret = nilfs_cpfile_clear_snapshot(cpfile, cno);
880 return ret;
881 case NILFS_SNAPSHOT:
882 return nilfs_cpfile_set_snapshot(cpfile, cno);
883 default:
884 return -EINVAL;
885 }
886}
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903int nilfs_cpfile_get_stat(struct inode *cpfile, struct nilfs_cpstat *cpstat)
904{
905 struct buffer_head *bh;
906 struct nilfs_cpfile_header *header;
907 void *kaddr;
908 int ret;
909
910 down_read(&NILFS_MDT(cpfile)->mi_sem);
911
912 ret = nilfs_cpfile_get_header_block(cpfile, &bh);
913 if (ret < 0)
914 goto out_sem;
915 kaddr = kmap_atomic(bh->b_page);
916 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr);
917 cpstat->cs_cno = nilfs_mdt_cno(cpfile);
918 cpstat->cs_ncps = le64_to_cpu(header->ch_ncheckpoints);
919 cpstat->cs_nsss = le64_to_cpu(header->ch_nsnapshots);
920 kunmap_atomic(kaddr);
921 brelse(bh);
922
923 out_sem:
924 up_read(&NILFS_MDT(cpfile)->mi_sem);
925 return ret;
926}
927
928
929
930
931
932
933
934
935int nilfs_cpfile_read(struct super_block *sb, size_t cpsize,
936 struct nilfs_inode *raw_inode, struct inode **inodep)
937{
938 struct inode *cpfile;
939 int err;
940
941 cpfile = nilfs_iget_locked(sb, NULL, NILFS_CPFILE_INO);
942 if (unlikely(!cpfile))
943 return -ENOMEM;
944 if (!(cpfile->i_state & I_NEW))
945 goto out;
946
947 err = nilfs_mdt_init(cpfile, NILFS_MDT_GFP, 0);
948 if (err)
949 goto failed;
950
951 nilfs_mdt_set_entry_size(cpfile, cpsize,
952 sizeof(struct nilfs_cpfile_header));
953
954 err = nilfs_read_inode_common(cpfile, raw_inode);
955 if (err)
956 goto failed;
957
958 unlock_new_inode(cpfile);
959 out:
960 *inodep = cpfile;
961 return 0;
962 failed:
963 iget_failed(cpfile);
964 return err;
965}
966