1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25#include <linux/random.h>
26#include "ubifs.h"
27
28
29
30
31
32
33
34
35
36
37static int make_idx_node(struct ubifs_info *c, struct ubifs_idx_node *idx,
38 struct ubifs_znode *znode, int lnum, int offs, int len)
39{
40 struct ubifs_znode *zp;
41 int i, err;
42
43
44 idx->ch.node_type = UBIFS_IDX_NODE;
45 idx->child_cnt = cpu_to_le16(znode->child_cnt);
46 idx->level = cpu_to_le16(znode->level);
47 for (i = 0; i < znode->child_cnt; i++) {
48 struct ubifs_branch *br = ubifs_idx_branch(c, idx, i);
49 struct ubifs_zbranch *zbr = &znode->zbranch[i];
50
51 key_write_idx(c, &zbr->key, &br->key);
52 br->lnum = cpu_to_le32(zbr->lnum);
53 br->offs = cpu_to_le32(zbr->offs);
54 br->len = cpu_to_le32(zbr->len);
55 if (!zbr->lnum || !zbr->len) {
56 ubifs_err("bad ref in znode");
57 ubifs_dump_znode(c, znode);
58 if (zbr->znode)
59 ubifs_dump_znode(c, zbr->znode);
60 }
61 }
62 ubifs_prepare_node(c, idx, len, 0);
63
64 znode->lnum = lnum;
65 znode->offs = offs;
66 znode->len = len;
67
68 err = insert_old_idx_znode(c, znode);
69
70
71 zp = znode->parent;
72 if (zp) {
73 struct ubifs_zbranch *zbr;
74
75 zbr = &zp->zbranch[znode->iip];
76 zbr->lnum = lnum;
77 zbr->offs = offs;
78 zbr->len = len;
79 } else {
80 c->zroot.lnum = lnum;
81 c->zroot.offs = offs;
82 c->zroot.len = len;
83 }
84 c->calc_idx_sz += ALIGN(len, 8);
85
86 atomic_long_dec(&c->dirty_zn_cnt);
87
88 ubifs_assert(ubifs_zn_dirty(znode));
89 ubifs_assert(ubifs_zn_cow(znode));
90
91
92
93
94
95 __clear_bit(DIRTY_ZNODE, &znode->flags);
96 __clear_bit(COW_ZNODE, &znode->flags);
97
98 return err;
99}
100
101
102
103
104
105
106
107
108
109
110
111static int fill_gap(struct ubifs_info *c, int lnum, int gap_start, int gap_end,
112 int *dirt)
113{
114 int len, gap_remains, gap_pos, written, pad_len;
115
116 ubifs_assert((gap_start & 7) == 0);
117 ubifs_assert((gap_end & 7) == 0);
118 ubifs_assert(gap_end >= gap_start);
119
120 gap_remains = gap_end - gap_start;
121 if (!gap_remains)
122 return 0;
123 gap_pos = gap_start;
124 written = 0;
125 while (c->enext) {
126 len = ubifs_idx_node_sz(c, c->enext->child_cnt);
127 if (len < gap_remains) {
128 struct ubifs_znode *znode = c->enext;
129 const int alen = ALIGN(len, 8);
130 int err;
131
132 ubifs_assert(alen <= gap_remains);
133 err = make_idx_node(c, c->ileb_buf + gap_pos, znode,
134 lnum, gap_pos, len);
135 if (err)
136 return err;
137 gap_remains -= alen;
138 gap_pos += alen;
139 c->enext = znode->cnext;
140 if (c->enext == c->cnext)
141 c->enext = NULL;
142 written += 1;
143 } else
144 break;
145 }
146 if (gap_end == c->leb_size) {
147 c->ileb_len = ALIGN(gap_pos, c->min_io_size);
148
149 pad_len = c->ileb_len - gap_pos;
150 } else
151
152 pad_len = gap_remains;
153 dbg_gc("LEB %d:%d to %d len %d nodes written %d wasted bytes %d",
154 lnum, gap_start, gap_end, gap_end - gap_start, written, pad_len);
155 ubifs_pad(c, c->ileb_buf + gap_pos, pad_len);
156 *dirt += pad_len;
157 return written;
158}
159
160
161
162
163
164
165
166
167
168static int find_old_idx(struct ubifs_info *c, int lnum, int offs)
169{
170 struct ubifs_old_idx *o;
171 struct rb_node *p;
172
173 p = c->old_idx.rb_node;
174 while (p) {
175 o = rb_entry(p, struct ubifs_old_idx, rb);
176 if (lnum < o->lnum)
177 p = p->rb_left;
178 else if (lnum > o->lnum)
179 p = p->rb_right;
180 else if (offs < o->offs)
181 p = p->rb_left;
182 else if (offs > o->offs)
183 p = p->rb_right;
184 else
185 return 1;
186 }
187 return 0;
188}
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203static int is_idx_node_in_use(struct ubifs_info *c, union ubifs_key *key,
204 int level, int lnum, int offs)
205{
206 int ret;
207
208 ret = is_idx_node_in_tnc(c, key, level, lnum, offs);
209 if (ret < 0)
210 return ret;
211 if (ret == 0)
212 if (find_old_idx(c, lnum, offs))
213 return 1;
214 return ret;
215}
216
217
218
219
220
221
222
223
224
225
226
227
228
229static int layout_leb_in_gaps(struct ubifs_info *c, int *p)
230{
231 struct ubifs_scan_leb *sleb;
232 struct ubifs_scan_node *snod;
233 int lnum, dirt = 0, gap_start, gap_end, err, written, tot_written;
234
235 tot_written = 0;
236
237 lnum = ubifs_find_dirty_idx_leb(c);
238 if (lnum < 0)
239
240
241
242
243 return lnum;
244 *p = lnum;
245 dbg_gc("LEB %d", lnum);
246
247
248
249
250
251 sleb = ubifs_scan(c, lnum, 0, c->ileb_buf, 0);
252 c->ileb_len = 0;
253 if (IS_ERR(sleb))
254 return PTR_ERR(sleb);
255 gap_start = 0;
256 list_for_each_entry(snod, &sleb->nodes, list) {
257 struct ubifs_idx_node *idx;
258 int in_use, level;
259
260 ubifs_assert(snod->type == UBIFS_IDX_NODE);
261 idx = snod->node;
262 key_read(c, ubifs_idx_key(c, idx), &snod->key);
263 level = le16_to_cpu(idx->level);
264
265 in_use = is_idx_node_in_use(c, &snod->key, level, lnum,
266 snod->offs);
267 if (in_use < 0) {
268 ubifs_scan_destroy(sleb);
269 return in_use;
270 }
271 if (in_use) {
272 if (in_use == 1)
273 dirt += ALIGN(snod->len, 8);
274
275
276
277
278
279
280 gap_end = snod->offs;
281
282 written = fill_gap(c, lnum, gap_start, gap_end, &dirt);
283 if (written < 0) {
284 ubifs_scan_destroy(sleb);
285 return written;
286 }
287 tot_written += written;
288 gap_start = ALIGN(snod->offs + snod->len, 8);
289 }
290 }
291 ubifs_scan_destroy(sleb);
292 c->ileb_len = c->leb_size;
293 gap_end = c->leb_size;
294
295 written = fill_gap(c, lnum, gap_start, gap_end, &dirt);
296 if (written < 0)
297 return written;
298 tot_written += written;
299 if (tot_written == 0) {
300 struct ubifs_lprops lp;
301
302 dbg_gc("LEB %d wrote %d index nodes", lnum, tot_written);
303 err = ubifs_read_one_lp(c, lnum, &lp);
304 if (err)
305 return err;
306 if (lp.free == c->leb_size) {
307
308
309
310
311 err = ubifs_change_one_lp(c, lnum,
312 c->leb_size - c->ileb_len,
313 dirt, 0, 0, 0);
314 if (err)
315 return err;
316 }
317 return 0;
318 }
319 err = ubifs_change_one_lp(c, lnum, c->leb_size - c->ileb_len, dirt,
320 0, 0, 0);
321 if (err)
322 return err;
323 err = ubifs_leb_change(c, lnum, c->ileb_buf, c->ileb_len);
324 if (err)
325 return err;
326 dbg_gc("LEB %d wrote %d index nodes", lnum, tot_written);
327 return tot_written;
328}
329
330
331
332
333
334
335
336
337
338
339static int get_leb_cnt(struct ubifs_info *c, int cnt)
340{
341 int d;
342
343
344 cnt -= (c->leb_size - c->ihead_offs) / c->max_idx_node_sz;
345 if (cnt < 0)
346 cnt = 0;
347 d = c->leb_size / c->max_idx_node_sz;
348 return DIV_ROUND_UP(cnt, d);
349}
350
351
352
353
354
355
356
357
358
359
360
361static int layout_in_gaps(struct ubifs_info *c, int cnt)
362{
363 int err, leb_needed_cnt, written, *p;
364
365 dbg_gc("%d znodes to write", cnt);
366
367 c->gap_lebs = kmalloc(sizeof(int) * (c->lst.idx_lebs + 1), GFP_NOFS);
368 if (!c->gap_lebs)
369 return -ENOMEM;
370
371 p = c->gap_lebs;
372 do {
373 ubifs_assert(p < c->gap_lebs + sizeof(int) * c->lst.idx_lebs);
374 written = layout_leb_in_gaps(c, p);
375 if (written < 0) {
376 err = written;
377 if (err != -ENOSPC) {
378 kfree(c->gap_lebs);
379 c->gap_lebs = NULL;
380 return err;
381 }
382 if (!dbg_is_chk_index(c)) {
383
384
385
386
387 ubifs_warn("out of space");
388 ubifs_dump_budg(c, &c->bi);
389 ubifs_dump_lprops(c);
390 }
391
392 err = 0;
393 break;
394 }
395 p++;
396 cnt -= written;
397 leb_needed_cnt = get_leb_cnt(c, cnt);
398 dbg_gc("%d znodes remaining, need %d LEBs, have %d", cnt,
399 leb_needed_cnt, c->ileb_cnt);
400 } while (leb_needed_cnt > c->ileb_cnt);
401
402 *p = -1;
403 return 0;
404}
405
406
407
408
409
410
411
412
413
414static int layout_in_empty_space(struct ubifs_info *c)
415{
416 struct ubifs_znode *znode, *cnext, *zp;
417 int lnum, offs, len, next_len, buf_len, buf_offs, used, avail;
418 int wlen, blen, err;
419
420 cnext = c->enext;
421 if (!cnext)
422 return 0;
423
424 lnum = c->ihead_lnum;
425 buf_offs = c->ihead_offs;
426
427 buf_len = ubifs_idx_node_sz(c, c->fanout);
428 buf_len = ALIGN(buf_len, c->min_io_size);
429 used = 0;
430 avail = buf_len;
431
432
433 next_len = ubifs_idx_node_sz(c, cnext->child_cnt);
434 if (buf_offs + next_len > c->leb_size)
435 lnum = -1;
436
437 while (1) {
438 znode = cnext;
439
440 len = ubifs_idx_node_sz(c, znode->child_cnt);
441
442
443 if (lnum == -1) {
444 if (c->ileb_nxt >= c->ileb_cnt) {
445 ubifs_err("out of space");
446 return -ENOSPC;
447 }
448 lnum = c->ilebs[c->ileb_nxt++];
449 buf_offs = 0;
450 used = 0;
451 avail = buf_len;
452 }
453
454 offs = buf_offs + used;
455
456 znode->lnum = lnum;
457 znode->offs = offs;
458 znode->len = len;
459
460
461 zp = znode->parent;
462 if (zp) {
463 struct ubifs_zbranch *zbr;
464 int i;
465
466 i = znode->iip;
467 zbr = &zp->zbranch[i];
468 zbr->lnum = lnum;
469 zbr->offs = offs;
470 zbr->len = len;
471 } else {
472 c->zroot.lnum = lnum;
473 c->zroot.offs = offs;
474 c->zroot.len = len;
475 }
476 c->calc_idx_sz += ALIGN(len, 8);
477
478
479
480
481
482 atomic_long_dec(&c->dirty_zn_cnt);
483
484
485
486
487
488 cnext = znode->cnext;
489 if (cnext == c->cnext)
490 next_len = 0;
491 else
492 next_len = ubifs_idx_node_sz(c, cnext->child_cnt);
493
494
495 wlen = used + len;
496 used += ALIGN(len, 8);
497 avail -= ALIGN(len, 8);
498
499 if (next_len != 0 &&
500 buf_offs + used + next_len <= c->leb_size &&
501 avail > 0)
502 continue;
503
504 if (avail <= 0 && next_len &&
505 buf_offs + used + next_len <= c->leb_size)
506 blen = buf_len;
507 else
508 blen = ALIGN(wlen, c->min_io_size);
509
510
511 buf_offs += blen;
512 if (next_len) {
513 if (buf_offs + next_len > c->leb_size) {
514 err = ubifs_update_one_lp(c, lnum,
515 c->leb_size - buf_offs, blen - used,
516 0, 0);
517 if (err)
518 return err;
519 lnum = -1;
520 }
521 used -= blen;
522 if (used < 0)
523 used = 0;
524 avail = buf_len - used;
525 continue;
526 }
527 err = ubifs_update_one_lp(c, lnum, c->leb_size - buf_offs,
528 blen - used, 0, 0);
529 if (err)
530 return err;
531 break;
532 }
533
534 c->dbg->new_ihead_lnum = lnum;
535 c->dbg->new_ihead_offs = buf_offs;
536
537 return 0;
538}
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553static int layout_commit(struct ubifs_info *c, int no_space, int cnt)
554{
555 int err;
556
557 if (no_space) {
558 err = layout_in_gaps(c, cnt);
559 if (err)
560 return err;
561 }
562 err = layout_in_empty_space(c);
563 return err;
564}
565
566
567
568
569
570static struct ubifs_znode *find_first_dirty(struct ubifs_znode *znode)
571{
572 int i, cont;
573
574 if (!znode)
575 return NULL;
576
577 while (1) {
578 if (znode->level == 0) {
579 if (ubifs_zn_dirty(znode))
580 return znode;
581 return NULL;
582 }
583 cont = 0;
584 for (i = 0; i < znode->child_cnt; i++) {
585 struct ubifs_zbranch *zbr = &znode->zbranch[i];
586
587 if (zbr->znode && ubifs_zn_dirty(zbr->znode)) {
588 znode = zbr->znode;
589 cont = 1;
590 break;
591 }
592 }
593 if (!cont) {
594 if (ubifs_zn_dirty(znode))
595 return znode;
596 return NULL;
597 }
598 }
599}
600
601
602
603
604
605static struct ubifs_znode *find_next_dirty(struct ubifs_znode *znode)
606{
607 int n = znode->iip + 1;
608
609 znode = znode->parent;
610 if (!znode)
611 return NULL;
612 for (; n < znode->child_cnt; n++) {
613 struct ubifs_zbranch *zbr = &znode->zbranch[n];
614
615 if (zbr->znode && ubifs_zn_dirty(zbr->znode))
616 return find_first_dirty(zbr->znode);
617 }
618 return znode;
619}
620
621
622
623
624
625
626
627static int get_znodes_to_commit(struct ubifs_info *c)
628{
629 struct ubifs_znode *znode, *cnext;
630 int cnt = 0;
631
632 c->cnext = find_first_dirty(c->zroot.znode);
633 znode = c->enext = c->cnext;
634 if (!znode) {
635 dbg_cmt("no znodes to commit");
636 return 0;
637 }
638 cnt += 1;
639 while (1) {
640 ubifs_assert(!ubifs_zn_cow(znode));
641 __set_bit(COW_ZNODE, &znode->flags);
642 znode->alt = 0;
643 cnext = find_next_dirty(znode);
644 if (!cnext) {
645 znode->cnext = c->cnext;
646 break;
647 }
648 znode->cnext = cnext;
649 znode = cnext;
650 cnt += 1;
651 }
652 dbg_cmt("committing %d znodes", cnt);
653 ubifs_assert(cnt == atomic_long_read(&c->dirty_zn_cnt));
654 return cnt;
655}
656
657
658
659
660
661
662
663
664
665
666static int alloc_idx_lebs(struct ubifs_info *c, int cnt)
667{
668 int i, leb_cnt, lnum;
669
670 c->ileb_cnt = 0;
671 c->ileb_nxt = 0;
672 leb_cnt = get_leb_cnt(c, cnt);
673 dbg_cmt("need about %d empty LEBS for TNC commit", leb_cnt);
674 if (!leb_cnt)
675 return 0;
676 c->ilebs = kmalloc(leb_cnt * sizeof(int), GFP_NOFS);
677 if (!c->ilebs)
678 return -ENOMEM;
679 for (i = 0; i < leb_cnt; i++) {
680 lnum = ubifs_find_free_leb_for_idx(c);
681 if (lnum < 0)
682 return lnum;
683 c->ilebs[c->ileb_cnt++] = lnum;
684 dbg_cmt("LEB %d", lnum);
685 }
686 if (dbg_is_chk_index(c) && !(prandom_u32() & 7))
687 return -ENOSPC;
688 return 0;
689}
690
691
692
693
694
695
696
697
698
699
700static int free_unused_idx_lebs(struct ubifs_info *c)
701{
702 int i, err = 0, lnum, er;
703
704 for (i = c->ileb_nxt; i < c->ileb_cnt; i++) {
705 lnum = c->ilebs[i];
706 dbg_cmt("LEB %d", lnum);
707 er = ubifs_change_one_lp(c, lnum, LPROPS_NC, LPROPS_NC, 0,
708 LPROPS_INDEX | LPROPS_TAKEN, 0);
709 if (!err)
710 err = er;
711 }
712 return err;
713}
714
715
716
717
718
719
720
721static int free_idx_lebs(struct ubifs_info *c)
722{
723 int err;
724
725 err = free_unused_idx_lebs(c);
726 kfree(c->ilebs);
727 c->ilebs = NULL;
728 return err;
729}
730
731
732
733
734
735
736
737
738
739
740
741int ubifs_tnc_start_commit(struct ubifs_info *c, struct ubifs_zbranch *zroot)
742{
743 int err = 0, cnt;
744
745 mutex_lock(&c->tnc_mutex);
746 err = dbg_check_tnc(c, 1);
747 if (err)
748 goto out;
749 cnt = get_znodes_to_commit(c);
750 if (cnt != 0) {
751 int no_space = 0;
752
753 err = alloc_idx_lebs(c, cnt);
754 if (err == -ENOSPC)
755 no_space = 1;
756 else if (err)
757 goto out_free;
758 err = layout_commit(c, no_space, cnt);
759 if (err)
760 goto out_free;
761 ubifs_assert(atomic_long_read(&c->dirty_zn_cnt) == 0);
762 err = free_unused_idx_lebs(c);
763 if (err)
764 goto out;
765 }
766 destroy_old_idx(c);
767 memcpy(zroot, &c->zroot, sizeof(struct ubifs_zbranch));
768
769 err = ubifs_save_dirty_idx_lnums(c);
770 if (err)
771 goto out;
772
773 spin_lock(&c->space_lock);
774
775
776
777
778
779
780
781
782 ubifs_assert(c->bi.min_idx_lebs == ubifs_calc_min_idx_lebs(c));
783 c->bi.old_idx_sz = c->calc_idx_sz;
784 c->bi.uncommitted_idx = 0;
785 c->bi.min_idx_lebs = ubifs_calc_min_idx_lebs(c);
786 spin_unlock(&c->space_lock);
787 mutex_unlock(&c->tnc_mutex);
788
789 dbg_cmt("number of index LEBs %d", c->lst.idx_lebs);
790 dbg_cmt("size of index %llu", c->calc_idx_sz);
791 return err;
792
793out_free:
794 free_idx_lebs(c);
795out:
796 mutex_unlock(&c->tnc_mutex);
797 return err;
798}
799
800
801
802
803
804
805
806
807static int write_index(struct ubifs_info *c)
808{
809 struct ubifs_idx_node *idx;
810 struct ubifs_znode *znode, *cnext;
811 int i, lnum, offs, len, next_len, buf_len, buf_offs, used;
812 int avail, wlen, err, lnum_pos = 0, blen, nxt_offs;
813
814 cnext = c->enext;
815 if (!cnext)
816 return 0;
817
818
819
820
821
822 lnum = c->ihead_lnum;
823 buf_offs = c->ihead_offs;
824
825
826 buf_len = ALIGN(c->max_idx_node_sz, c->min_io_size);
827 used = 0;
828 avail = buf_len;
829
830
831 next_len = ubifs_idx_node_sz(c, cnext->child_cnt);
832 if (buf_offs + next_len > c->leb_size) {
833 err = ubifs_update_one_lp(c, lnum, LPROPS_NC, 0, 0,
834 LPROPS_TAKEN);
835 if (err)
836 return err;
837 lnum = -1;
838 }
839
840 while (1) {
841 cond_resched();
842
843 znode = cnext;
844 idx = c->cbuf + used;
845
846
847 idx->ch.node_type = UBIFS_IDX_NODE;
848 idx->child_cnt = cpu_to_le16(znode->child_cnt);
849 idx->level = cpu_to_le16(znode->level);
850 for (i = 0; i < znode->child_cnt; i++) {
851 struct ubifs_branch *br = ubifs_idx_branch(c, idx, i);
852 struct ubifs_zbranch *zbr = &znode->zbranch[i];
853
854 key_write_idx(c, &zbr->key, &br->key);
855 br->lnum = cpu_to_le32(zbr->lnum);
856 br->offs = cpu_to_le32(zbr->offs);
857 br->len = cpu_to_le32(zbr->len);
858 if (!zbr->lnum || !zbr->len) {
859 ubifs_err("bad ref in znode");
860 ubifs_dump_znode(c, znode);
861 if (zbr->znode)
862 ubifs_dump_znode(c, zbr->znode);
863 }
864 }
865 len = ubifs_idx_node_sz(c, znode->child_cnt);
866 ubifs_prepare_node(c, idx, len, 0);
867
868
869 if (lnum == -1) {
870 lnum = c->ilebs[lnum_pos++];
871 buf_offs = 0;
872 used = 0;
873 avail = buf_len;
874 }
875 offs = buf_offs + used;
876
877 if (lnum != znode->lnum || offs != znode->offs ||
878 len != znode->len) {
879 ubifs_err("inconsistent znode posn");
880 return -EINVAL;
881 }
882
883
884 cnext = znode->cnext;
885
886 ubifs_assert(ubifs_zn_dirty(znode));
887 ubifs_assert(ubifs_zn_cow(znode));
888
889
890
891
892
893
894
895
896
897 clear_bit(DIRTY_ZNODE, &znode->flags);
898 smp_mb__before_clear_bit();
899 clear_bit(COW_ZNODE, &znode->flags);
900 smp_mb__after_clear_bit();
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927 wlen = used + len;
928 used += ALIGN(len, 8);
929 avail -= ALIGN(len, 8);
930
931
932
933
934
935 if (cnext == c->cnext)
936 next_len = 0;
937 else
938 next_len = ubifs_idx_node_sz(c, cnext->child_cnt);
939
940 nxt_offs = buf_offs + used + next_len;
941 if (next_len && nxt_offs <= c->leb_size) {
942 if (avail > 0)
943 continue;
944 else
945 blen = buf_len;
946 } else {
947 wlen = ALIGN(wlen, 8);
948 blen = ALIGN(wlen, c->min_io_size);
949 ubifs_pad(c, c->cbuf + wlen, blen - wlen);
950 }
951
952
953 err = ubifs_leb_write(c, lnum, c->cbuf, buf_offs, blen);
954 if (err)
955 return err;
956 buf_offs += blen;
957 if (next_len) {
958 if (nxt_offs > c->leb_size) {
959 err = ubifs_update_one_lp(c, lnum, LPROPS_NC, 0,
960 0, LPROPS_TAKEN);
961 if (err)
962 return err;
963 lnum = -1;
964 }
965 used -= blen;
966 if (used < 0)
967 used = 0;
968 avail = buf_len - used;
969 memmove(c->cbuf, c->cbuf + blen, used);
970 continue;
971 }
972 break;
973 }
974
975 if (lnum != c->dbg->new_ihead_lnum ||
976 buf_offs != c->dbg->new_ihead_offs) {
977 ubifs_err("inconsistent ihead");
978 return -EINVAL;
979 }
980
981 c->ihead_lnum = lnum;
982 c->ihead_offs = buf_offs;
983
984 return 0;
985}
986
987
988
989
990
991
992
993static void free_obsolete_znodes(struct ubifs_info *c)
994{
995 struct ubifs_znode *znode, *cnext;
996
997 cnext = c->cnext;
998 do {
999 znode = cnext;
1000 cnext = znode->cnext;
1001 if (ubifs_zn_obsolete(znode))
1002 kfree(znode);
1003 else {
1004 znode->cnext = NULL;
1005 atomic_long_inc(&c->clean_zn_cnt);
1006 atomic_long_inc(&ubifs_clean_zn_cnt);
1007 }
1008 } while (cnext != c->cnext);
1009}
1010
1011
1012
1013
1014
1015
1016
1017
1018static int return_gap_lebs(struct ubifs_info *c)
1019{
1020 int *p, err;
1021
1022 if (!c->gap_lebs)
1023 return 0;
1024
1025 dbg_cmt("");
1026 for (p = c->gap_lebs; *p != -1; p++) {
1027 err = ubifs_change_one_lp(c, *p, LPROPS_NC, LPROPS_NC, 0,
1028 LPROPS_TAKEN, 0);
1029 if (err)
1030 return err;
1031 }
1032
1033 kfree(c->gap_lebs);
1034 c->gap_lebs = NULL;
1035 return 0;
1036}
1037
1038
1039
1040
1041
1042
1043
1044int ubifs_tnc_end_commit(struct ubifs_info *c)
1045{
1046 int err;
1047
1048 if (!c->cnext)
1049 return 0;
1050
1051 err = return_gap_lebs(c);
1052 if (err)
1053 return err;
1054
1055 err = write_index(c);
1056 if (err)
1057 return err;
1058
1059 mutex_lock(&c->tnc_mutex);
1060
1061 dbg_cmt("TNC height is %d", c->zroot.znode->level + 1);
1062
1063 free_obsolete_znodes(c);
1064
1065 c->cnext = NULL;
1066 kfree(c->ilebs);
1067 c->ilebs = NULL;
1068
1069 mutex_unlock(&c->tnc_mutex);
1070
1071 return 0;
1072}
1073