1
2
3
4
5#include <linux/time.h>
6#include <linux/fs.h>
7#include "reiserfs.h"
8#include "acl.h"
9#include "xattr.h"
10#include <linux/exportfs.h>
11#include <linux/pagemap.h>
12#include <linux/highmem.h>
13#include <linux/slab.h>
14#include <linux/uaccess.h>
15#include <asm/unaligned.h>
16#include <linux/buffer_head.h>
17#include <linux/mpage.h>
18#include <linux/writeback.h>
19#include <linux/quotaops.h>
20#include <linux/swap.h>
21#include <linux/uio.h>
22#include <linux/bio.h>
23
24int reiserfs_commit_write(struct file *f, struct page *page,
25 unsigned from, unsigned to);
26
27void reiserfs_evict_inode(struct inode *inode)
28{
29
30
31
32
33 int jbegin_count =
34 JOURNAL_PER_BALANCE_CNT * 2 +
35 2 * REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb);
36 struct reiserfs_transaction_handle th;
37 int err;
38
39 if (!inode->i_nlink && !is_bad_inode(inode))
40 dquot_initialize(inode);
41
42 truncate_inode_pages_final(&inode->i_data);
43 if (inode->i_nlink)
44 goto no_delete;
45
46
47
48
49
50
51 if (!(inode->i_state & I_NEW) && INODE_PKEY(inode)->k_objectid != 0) {
52
53 reiserfs_delete_xattrs(inode);
54
55 reiserfs_write_lock(inode->i_sb);
56
57 if (journal_begin(&th, inode->i_sb, jbegin_count))
58 goto out;
59 reiserfs_update_inode_transaction(inode);
60
61 reiserfs_discard_prealloc(&th, inode);
62
63 err = reiserfs_delete_object(&th, inode);
64
65
66
67
68
69
70 if (!err) {
71 int depth = reiserfs_write_unlock_nested(inode->i_sb);
72 dquot_free_inode(inode);
73 reiserfs_write_lock_nested(inode->i_sb, depth);
74 }
75
76 if (journal_end(&th))
77 goto out;
78
79
80
81
82
83 if (err)
84 goto out;
85
86
87
88
89
90
91 remove_save_link(inode, 0 );
92out:
93 reiserfs_write_unlock(inode->i_sb);
94 } else {
95
96 ;
97 }
98
99
100 clear_inode(inode);
101
102 dquot_drop(inode);
103 inode->i_blocks = 0;
104 return;
105
106no_delete:
107 clear_inode(inode);
108 dquot_drop(inode);
109}
110
111static void _make_cpu_key(struct cpu_key *key, int version, __u32 dirid,
112 __u32 objectid, loff_t offset, int type, int length)
113{
114 key->version = version;
115
116 key->on_disk_key.k_dir_id = dirid;
117 key->on_disk_key.k_objectid = objectid;
118 set_cpu_key_k_offset(key, offset);
119 set_cpu_key_k_type(key, type);
120 key->key_length = length;
121}
122
123
124
125
126
127void make_cpu_key(struct cpu_key *key, struct inode *inode, loff_t offset,
128 int type, int length)
129{
130 _make_cpu_key(key, get_inode_item_key_version(inode),
131 le32_to_cpu(INODE_PKEY(inode)->k_dir_id),
132 le32_to_cpu(INODE_PKEY(inode)->k_objectid), offset, type,
133 length);
134}
135
136
137inline void make_le_item_head(struct item_head *ih, const struct cpu_key *key,
138 int version,
139 loff_t offset, int type, int length,
140 int entry_count )
141{
142 if (key) {
143 ih->ih_key.k_dir_id = cpu_to_le32(key->on_disk_key.k_dir_id);
144 ih->ih_key.k_objectid =
145 cpu_to_le32(key->on_disk_key.k_objectid);
146 }
147 put_ih_version(ih, version);
148 set_le_ih_k_offset(ih, offset);
149 set_le_ih_k_type(ih, type);
150 put_ih_item_len(ih, length);
151
152
153
154
155
156 put_ih_entry_count(ih, entry_count);
157}
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193static inline void fix_tail_page_for_writing(struct page *page)
194{
195 struct buffer_head *head, *next, *bh;
196
197 if (page && page_has_buffers(page)) {
198 head = page_buffers(page);
199 bh = head;
200 do {
201 next = bh->b_this_page;
202 if (buffer_mapped(bh) && bh->b_blocknr == 0) {
203 reiserfs_unmap_buffer(bh);
204 }
205 bh = next;
206 } while (bh != head);
207 }
208}
209
210
211
212
213
214static inline int allocation_needed(int retval, b_blocknr_t allocated,
215 struct item_head *ih,
216 __le32 * item, int pos_in_item)
217{
218 if (allocated)
219 return 0;
220 if (retval == POSITION_FOUND && is_indirect_le_ih(ih) &&
221 get_block_num(item, pos_in_item))
222 return 0;
223 return 1;
224}
225
226static inline int indirect_item_found(int retval, struct item_head *ih)
227{
228 return (retval == POSITION_FOUND) && is_indirect_le_ih(ih);
229}
230
231static inline void set_block_dev_mapped(struct buffer_head *bh,
232 b_blocknr_t block, struct inode *inode)
233{
234 map_bh(bh, inode->i_sb, block);
235}
236
237
238
239
240
241static int file_capable(struct inode *inode, sector_t block)
242{
243
244 if (get_inode_item_key_version(inode) != KEY_FORMAT_3_5 ||
245
246 block < (1 << (31 - inode->i_sb->s_blocksize_bits)))
247 return 1;
248
249 return 0;
250}
251
252static int restart_transaction(struct reiserfs_transaction_handle *th,
253 struct inode *inode, struct treepath *path)
254{
255 struct super_block *s = th->t_super;
256 int err;
257
258 BUG_ON(!th->t_trans_id);
259 BUG_ON(!th->t_refcount);
260
261 pathrelse(path);
262
263
264 if (th->t_refcount > 1) {
265 return 0;
266 }
267 reiserfs_update_sd(th, inode);
268 err = journal_end(th);
269 if (!err) {
270 err = journal_begin(th, s, JOURNAL_PER_BALANCE_CNT * 6);
271 if (!err)
272 reiserfs_update_inode_transaction(inode);
273 }
274 return err;
275}
276
277
278
279
280
281
282
283
284
285static int _get_block_create_0(struct inode *inode, sector_t block,
286 struct buffer_head *bh_result, int args)
287{
288 INITIALIZE_PATH(path);
289 struct cpu_key key;
290 struct buffer_head *bh;
291 struct item_head *ih, tmp_ih;
292 b_blocknr_t blocknr;
293 char *p = NULL;
294 int chars;
295 int ret;
296 int result;
297 int done = 0;
298 unsigned long offset;
299
300
301 make_cpu_key(&key, inode,
302 (loff_t) block * inode->i_sb->s_blocksize + 1, TYPE_ANY,
303 3);
304
305 result = search_for_position_by_key(inode->i_sb, &key, &path);
306 if (result != POSITION_FOUND) {
307 pathrelse(&path);
308 if (p)
309 kunmap(bh_result->b_page);
310 if (result == IO_ERROR)
311 return -EIO;
312
313
314
315
316
317 if ((args & GET_BLOCK_NO_HOLE)
318 && !PageUptodate(bh_result->b_page)) {
319 return -ENOENT;
320 }
321 return 0;
322 }
323
324 bh = get_last_bh(&path);
325 ih = tp_item_head(&path);
326 if (is_indirect_le_ih(ih)) {
327 __le32 *ind_item = (__le32 *) ih_item_body(bh, ih);
328
329
330
331
332
333
334 blocknr = get_block_num(ind_item, path.pos_in_item);
335 ret = 0;
336 if (blocknr) {
337 map_bh(bh_result, inode->i_sb, blocknr);
338 if (path.pos_in_item ==
339 ((ih_item_len(ih) / UNFM_P_SIZE) - 1)) {
340 set_buffer_boundary(bh_result);
341 }
342 } else
343
344
345
346
347
348
349 if ((args & GET_BLOCK_NO_HOLE)
350 && !PageUptodate(bh_result->b_page)) {
351 ret = -ENOENT;
352 }
353
354 pathrelse(&path);
355 if (p)
356 kunmap(bh_result->b_page);
357 return ret;
358 }
359
360 if (!(args & GET_BLOCK_READ_DIRECT)) {
361
362
363
364
365 pathrelse(&path);
366 if (p)
367 kunmap(bh_result->b_page);
368 return -ENOENT;
369 }
370
371
372
373
374
375
376 if (buffer_uptodate(bh_result)) {
377 goto finished;
378 } else
379
380
381
382
383
384
385 if (!bh_result->b_page || PageUptodate(bh_result->b_page)) {
386 set_buffer_uptodate(bh_result);
387 goto finished;
388 }
389
390 offset = (cpu_key_k_offset(&key) - 1) & (PAGE_SIZE - 1);
391 copy_item_head(&tmp_ih, ih);
392
393
394
395
396
397
398
399 if (!p)
400 p = (char *)kmap(bh_result->b_page);
401
402 p += offset;
403 memset(p, 0, inode->i_sb->s_blocksize);
404 do {
405 if (!is_direct_le_ih(ih)) {
406 BUG();
407 }
408
409
410
411
412
413
414 if ((le_ih_k_offset(ih) + path.pos_in_item) > inode->i_size)
415 break;
416 if ((le_ih_k_offset(ih) - 1 + ih_item_len(ih)) > inode->i_size) {
417 chars =
418 inode->i_size - (le_ih_k_offset(ih) - 1) -
419 path.pos_in_item;
420 done = 1;
421 } else {
422 chars = ih_item_len(ih) - path.pos_in_item;
423 }
424 memcpy(p, ih_item_body(bh, ih) + path.pos_in_item, chars);
425
426 if (done)
427 break;
428
429 p += chars;
430
431
432
433
434
435
436
437 if (PATH_LAST_POSITION(&path) != (B_NR_ITEMS(bh) - 1))
438 break;
439
440
441 set_cpu_key_k_offset(&key, cpu_key_k_offset(&key) + chars);
442 result = search_for_position_by_key(inode->i_sb, &key, &path);
443 if (result != POSITION_FOUND)
444
445 break;
446 bh = get_last_bh(&path);
447 ih = tp_item_head(&path);
448 } while (1);
449
450 flush_dcache_page(bh_result->b_page);
451 kunmap(bh_result->b_page);
452
453finished:
454 pathrelse(&path);
455
456 if (result == IO_ERROR)
457 return -EIO;
458
459
460
461
462
463 map_bh(bh_result, inode->i_sb, 0);
464 set_buffer_uptodate(bh_result);
465 return 0;
466}
467
468
469
470
471
472static int reiserfs_bmap(struct inode *inode, sector_t block,
473 struct buffer_head *bh_result, int create)
474{
475 if (!file_capable(inode, block))
476 return -EFBIG;
477
478 reiserfs_write_lock(inode->i_sb);
479
480 _get_block_create_0(inode, block, bh_result, 0);
481 reiserfs_write_unlock(inode->i_sb);
482 return 0;
483}
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503static int reiserfs_get_block_create_0(struct inode *inode, sector_t block,
504 struct buffer_head *bh_result,
505 int create)
506{
507 return reiserfs_get_block(inode, block, bh_result, GET_BLOCK_NO_HOLE);
508}
509
510
511
512
513
514static int reiserfs_get_blocks_direct_io(struct inode *inode,
515 sector_t iblock,
516 struct buffer_head *bh_result,
517 int create)
518{
519 int ret;
520
521 bh_result->b_page = NULL;
522
523
524
525
526
527
528 bh_result->b_size = i_blocksize(inode);
529
530 ret = reiserfs_get_block(inode, iblock, bh_result,
531 create | GET_BLOCK_NO_DANGLE);
532 if (ret)
533 goto out;
534
535
536 if (buffer_mapped(bh_result) && bh_result->b_blocknr == 0) {
537
538
539
540
541 clear_buffer_mapped(bh_result);
542 ret = -EINVAL;
543 }
544
545
546
547
548
549 if (REISERFS_I(inode)->i_flags & i_pack_on_close_mask) {
550 int err;
551
552 reiserfs_write_lock(inode->i_sb);
553
554 err = reiserfs_commit_for_inode(inode);
555 REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask;
556
557 reiserfs_write_unlock(inode->i_sb);
558
559 if (err < 0)
560 ret = err;
561 }
562out:
563 return ret;
564}
565
566
567
568
569
570
571
572
573
574
575
576static int convert_tail_for_hole(struct inode *inode,
577 struct buffer_head *bh_result,
578 loff_t tail_offset)
579{
580 unsigned long index;
581 unsigned long tail_end;
582 unsigned long tail_start;
583 struct page *tail_page;
584 struct page *hole_page = bh_result->b_page;
585 int retval = 0;
586
587 if ((tail_offset & (bh_result->b_size - 1)) != 1)
588 return -EIO;
589
590
591 tail_start = tail_offset & (PAGE_SIZE - 1);
592 tail_end = (tail_start | (bh_result->b_size - 1)) + 1;
593
594 index = tail_offset >> PAGE_SHIFT;
595
596
597
598
599 if (!hole_page || index != hole_page->index) {
600 tail_page = grab_cache_page(inode->i_mapping, index);
601 retval = -ENOMEM;
602 if (!tail_page) {
603 goto out;
604 }
605 } else {
606 tail_page = hole_page;
607 }
608
609
610
611
612
613
614
615
616
617
618
619 fix_tail_page_for_writing(tail_page);
620 retval = __reiserfs_write_begin(tail_page, tail_start,
621 tail_end - tail_start);
622 if (retval)
623 goto unlock;
624
625
626 flush_dcache_page(tail_page);
627
628 retval = reiserfs_commit_write(NULL, tail_page, tail_start, tail_end);
629
630unlock:
631 if (tail_page != hole_page) {
632 unlock_page(tail_page);
633 put_page(tail_page);
634 }
635out:
636 return retval;
637}
638
639static inline int _allocate_block(struct reiserfs_transaction_handle *th,
640 sector_t block,
641 struct inode *inode,
642 b_blocknr_t * allocated_block_nr,
643 struct treepath *path, int flags)
644{
645 BUG_ON(!th->t_trans_id);
646
647#ifdef REISERFS_PREALLOCATE
648 if (!(flags & GET_BLOCK_NO_IMUX)) {
649 return reiserfs_new_unf_blocknrs2(th, inode, allocated_block_nr,
650 path, block);
651 }
652#endif
653 return reiserfs_new_unf_blocknrs(th, inode, allocated_block_nr, path,
654 block);
655}
656
657int reiserfs_get_block(struct inode *inode, sector_t block,
658 struct buffer_head *bh_result, int create)
659{
660 int repeat, retval = 0;
661
662 b_blocknr_t allocated_block_nr = 0;
663 INITIALIZE_PATH(path);
664 int pos_in_item;
665 struct cpu_key key;
666 struct buffer_head *bh, *unbh = NULL;
667 struct item_head *ih, tmp_ih;
668 __le32 *item;
669 int done;
670 int fs_gen;
671 struct reiserfs_transaction_handle *th = NULL;
672
673
674
675
676
677
678
679
680 int jbegin_count =
681 JOURNAL_PER_BALANCE_CNT * 3 + 1 +
682 2 * REISERFS_QUOTA_TRANS_BLOCKS(inode->i_sb);
683 int version;
684 int dangle = 1;
685 loff_t new_offset =
686 (((loff_t) block) << inode->i_sb->s_blocksize_bits) + 1;
687
688 reiserfs_write_lock(inode->i_sb);
689 version = get_inode_item_key_version(inode);
690
691 if (!file_capable(inode, block)) {
692 reiserfs_write_unlock(inode->i_sb);
693 return -EFBIG;
694 }
695
696
697
698
699
700 if (!(create & GET_BLOCK_CREATE)) {
701 int ret;
702
703 ret = _get_block_create_0(inode, block, bh_result,
704 create | GET_BLOCK_READ_DIRECT);
705 reiserfs_write_unlock(inode->i_sb);
706 return ret;
707 }
708
709
710
711
712
713 if ((create & GET_BLOCK_NO_DANGLE) ||
714 reiserfs_transaction_running(inode->i_sb))
715 dangle = 0;
716
717
718
719
720
721
722 if ((have_large_tails(inode->i_sb)
723 && inode->i_size < i_block_size(inode) * 4)
724 || (have_small_tails(inode->i_sb)
725 && inode->i_size < i_block_size(inode)))
726 REISERFS_I(inode)->i_flags |= i_pack_on_close_mask;
727
728
729 make_cpu_key(&key, inode, new_offset, TYPE_ANY, 3 );
730 if ((new_offset + inode->i_sb->s_blocksize - 1) > inode->i_size) {
731start_trans:
732 th = reiserfs_persistent_transaction(inode->i_sb, jbegin_count);
733 if (!th) {
734 retval = -ENOMEM;
735 goto failure;
736 }
737 reiserfs_update_inode_transaction(inode);
738 }
739research:
740
741 retval = search_for_position_by_key(inode->i_sb, &key, &path);
742 if (retval == IO_ERROR) {
743 retval = -EIO;
744 goto failure;
745 }
746
747 bh = get_last_bh(&path);
748 ih = tp_item_head(&path);
749 item = tp_item_body(&path);
750 pos_in_item = path.pos_in_item;
751
752 fs_gen = get_generation(inode->i_sb);
753 copy_item_head(&tmp_ih, ih);
754
755 if (allocation_needed
756 (retval, allocated_block_nr, ih, item, pos_in_item)) {
757
758 if (!th) {
759 pathrelse(&path);
760 goto start_trans;
761 }
762
763 repeat =
764 _allocate_block(th, block, inode, &allocated_block_nr,
765 &path, create);
766
767
768
769
770
771
772 if (repeat == NO_DISK_SPACE || repeat == QUOTA_EXCEEDED) {
773 SB_JOURNAL(inode->i_sb)->j_next_async_flush = 1;
774 retval = restart_transaction(th, inode, &path);
775 if (retval)
776 goto failure;
777 repeat =
778 _allocate_block(th, block, inode,
779 &allocated_block_nr, NULL, create);
780
781 if (repeat != NO_DISK_SPACE && repeat != QUOTA_EXCEEDED) {
782 goto research;
783 }
784 if (repeat == QUOTA_EXCEEDED)
785 retval = -EDQUOT;
786 else
787 retval = -ENOSPC;
788 goto failure;
789 }
790
791 if (fs_changed(fs_gen, inode->i_sb)
792 && item_moved(&tmp_ih, &path)) {
793 goto research;
794 }
795 }
796
797 if (indirect_item_found(retval, ih)) {
798 b_blocknr_t unfm_ptr;
799
800
801
802
803
804 unfm_ptr = get_block_num(item, pos_in_item);
805 if (unfm_ptr == 0) {
806
807 reiserfs_prepare_for_journal(inode->i_sb, bh, 1);
808 if (fs_changed(fs_gen, inode->i_sb)
809 && item_moved(&tmp_ih, &path)) {
810 reiserfs_restore_prepared_buffer(inode->i_sb,
811 bh);
812 goto research;
813 }
814 set_buffer_new(bh_result);
815 if (buffer_dirty(bh_result)
816 && reiserfs_data_ordered(inode->i_sb))
817 reiserfs_add_ordered_list(inode, bh_result);
818 put_block_num(item, pos_in_item, allocated_block_nr);
819 unfm_ptr = allocated_block_nr;
820 journal_mark_dirty(th, bh);
821 reiserfs_update_sd(th, inode);
822 }
823 set_block_dev_mapped(bh_result, unfm_ptr, inode);
824 pathrelse(&path);
825 retval = 0;
826 if (!dangle && th)
827 retval = reiserfs_end_persistent_transaction(th);
828
829 reiserfs_write_unlock(inode->i_sb);
830
831
832
833
834
835
836 return retval;
837 }
838
839 if (!th) {
840 pathrelse(&path);
841 goto start_trans;
842 }
843
844
845
846
847
848
849 done = 0;
850 do {
851 if (is_statdata_le_ih(ih)) {
852 __le32 unp = 0;
853 struct cpu_key tmp_key;
854
855
856 make_le_item_head(&tmp_ih, &key, version, 1,
857 TYPE_INDIRECT, UNFM_P_SIZE,
858 0 );
859
860
861
862
863
864 if (cpu_key_k_offset(&key) == 1) {
865 unp = cpu_to_le32(allocated_block_nr);
866 set_block_dev_mapped(bh_result,
867 allocated_block_nr, inode);
868 set_buffer_new(bh_result);
869 done = 1;
870 }
871 tmp_key = key;
872 set_cpu_key_k_offset(&tmp_key, 1);
873 PATH_LAST_POSITION(&path)++;
874
875 retval =
876 reiserfs_insert_item(th, &path, &tmp_key, &tmp_ih,
877 inode, (char *)&unp);
878 if (retval) {
879 reiserfs_free_block(th, inode,
880 allocated_block_nr, 1);
881
882
883
884
885 goto failure;
886 }
887 } else if (is_direct_le_ih(ih)) {
888
889 loff_t tail_offset;
890
891 tail_offset =
892 ((le_ih_k_offset(ih) -
893 1) & ~(inode->i_sb->s_blocksize - 1)) + 1;
894
895
896
897
898
899
900 if (tail_offset == cpu_key_k_offset(&key)) {
901 set_block_dev_mapped(bh_result,
902 allocated_block_nr, inode);
903 unbh = bh_result;
904 done = 1;
905 } else {
906
907
908
909
910
911
912
913 pathrelse(&path);
914
915
916
917
918 BUG_ON(!th->t_refcount);
919 if (th->t_refcount == 1) {
920 retval =
921 reiserfs_end_persistent_transaction
922 (th);
923 th = NULL;
924 if (retval)
925 goto failure;
926 }
927
928 retval =
929 convert_tail_for_hole(inode, bh_result,
930 tail_offset);
931 if (retval) {
932 if (retval != -ENOSPC)
933 reiserfs_error(inode->i_sb,
934 "clm-6004",
935 "convert tail failed "
936 "inode %lu, error %d",
937 inode->i_ino,
938 retval);
939 if (allocated_block_nr) {
940
941
942
943
944 if (!th)
945 th = reiserfs_persistent_transaction(inode->i_sb, 3);
946 if (th)
947 reiserfs_free_block(th,
948 inode,
949 allocated_block_nr,
950 1);
951 }
952 goto failure;
953 }
954 goto research;
955 }
956 retval =
957 direct2indirect(th, inode, &path, unbh,
958 tail_offset);
959 if (retval) {
960 reiserfs_unmap_buffer(unbh);
961 reiserfs_free_block(th, inode,
962 allocated_block_nr, 1);
963 goto failure;
964 }
965
966
967
968
969
970
971
972
973
974
975 set_buffer_uptodate(unbh);
976
977
978
979
980
981
982 if (unbh->b_page) {
983
984
985
986
987 reiserfs_add_tail_list(inode, unbh);
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002 mark_buffer_dirty(unbh);
1003 }
1004 } else {
1005
1006
1007
1008
1009
1010 struct cpu_key tmp_key;
1011
1012
1013
1014
1015 unp_t unf_single = 0;
1016 unp_t *un;
1017 __u64 max_to_insert =
1018 MAX_ITEM_LEN(inode->i_sb->s_blocksize) /
1019 UNFM_P_SIZE;
1020 __u64 blocks_needed;
1021
1022 RFALSE(pos_in_item != ih_item_len(ih) / UNFM_P_SIZE,
1023 "vs-804: invalid position for append");
1024
1025
1026
1027
1028
1029 make_cpu_key(&tmp_key, inode,
1030 le_key_k_offset(version,
1031 &ih->ih_key) +
1032 op_bytes_number(ih,
1033 inode->i_sb->s_blocksize),
1034 TYPE_INDIRECT, 3);
1035
1036 RFALSE(cpu_key_k_offset(&tmp_key) > cpu_key_k_offset(&key),
1037 "green-805: invalid offset");
1038 blocks_needed =
1039 1 +
1040 ((cpu_key_k_offset(&key) -
1041 cpu_key_k_offset(&tmp_key)) >> inode->i_sb->
1042 s_blocksize_bits);
1043
1044 if (blocks_needed == 1) {
1045 un = &unf_single;
1046 } else {
1047 un = kcalloc(min(blocks_needed, max_to_insert),
1048 UNFM_P_SIZE, GFP_NOFS);
1049 if (!un) {
1050 un = &unf_single;
1051 blocks_needed = 1;
1052 max_to_insert = 0;
1053 }
1054 }
1055 if (blocks_needed <= max_to_insert) {
1056
1057
1058
1059
1060 un[blocks_needed - 1] =
1061 cpu_to_le32(allocated_block_nr);
1062 set_block_dev_mapped(bh_result,
1063 allocated_block_nr, inode);
1064 set_buffer_new(bh_result);
1065 done = 1;
1066 } else {
1067
1068
1069
1070
1071
1072
1073 blocks_needed =
1074 max_to_insert ? max_to_insert : 1;
1075 }
1076 retval =
1077 reiserfs_paste_into_item(th, &path, &tmp_key, inode,
1078 (char *)un,
1079 UNFM_P_SIZE *
1080 blocks_needed);
1081
1082 if (blocks_needed != 1)
1083 kfree(un);
1084
1085 if (retval) {
1086 reiserfs_free_block(th, inode,
1087 allocated_block_nr, 1);
1088 goto failure;
1089 }
1090 if (!done) {
1091
1092
1093
1094
1095
1096
1097 inode->i_size +=
1098 inode->i_sb->s_blocksize * blocks_needed;
1099 }
1100 }
1101
1102 if (done == 1)
1103 break;
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113 if (journal_transaction_should_end(th, th->t_blocks_allocated)) {
1114 retval = restart_transaction(th, inode, &path);
1115 if (retval)
1116 goto failure;
1117 }
1118
1119
1120
1121
1122
1123 reiserfs_cond_resched(inode->i_sb);
1124
1125 retval = search_for_position_by_key(inode->i_sb, &key, &path);
1126 if (retval == IO_ERROR) {
1127 retval = -EIO;
1128 goto failure;
1129 }
1130 if (retval == POSITION_FOUND) {
1131 reiserfs_warning(inode->i_sb, "vs-825",
1132 "%K should not be found", &key);
1133 retval = -EEXIST;
1134 if (allocated_block_nr)
1135 reiserfs_free_block(th, inode,
1136 allocated_block_nr, 1);
1137 pathrelse(&path);
1138 goto failure;
1139 }
1140 bh = get_last_bh(&path);
1141 ih = tp_item_head(&path);
1142 item = tp_item_body(&path);
1143 pos_in_item = path.pos_in_item;
1144 } while (1);
1145
1146 retval = 0;
1147
1148failure:
1149 if (th && (!dangle || (retval && !th->t_trans_id))) {
1150 int err;
1151 if (th->t_trans_id)
1152 reiserfs_update_sd(th, inode);
1153 err = reiserfs_end_persistent_transaction(th);
1154 if (err)
1155 retval = err;
1156 }
1157
1158 reiserfs_write_unlock(inode->i_sb);
1159 reiserfs_check_path(&path);
1160 return retval;
1161}
1162
1163static int
1164reiserfs_readpages(struct file *file, struct address_space *mapping,
1165 struct list_head *pages, unsigned nr_pages)
1166{
1167 return mpage_readpages(mapping, pages, nr_pages, reiserfs_get_block);
1168}
1169
1170
1171
1172
1173
1174
1175static int real_space_diff(struct inode *inode, int sd_size)
1176{
1177 int bytes;
1178 loff_t blocksize = inode->i_sb->s_blocksize;
1179
1180 if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode))
1181 return sd_size;
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192 bytes =
1193 ((inode->i_size +
1194 (blocksize - 1)) >> inode->i_sb->s_blocksize_bits) * UNFM_P_SIZE +
1195 sd_size;
1196 return bytes;
1197}
1198
1199static inline loff_t to_real_used_space(struct inode *inode, ulong blocks,
1200 int sd_size)
1201{
1202 if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) {
1203 return inode->i_size +
1204 (loff_t) (real_space_diff(inode, sd_size));
1205 }
1206 return ((loff_t) real_space_diff(inode, sd_size)) +
1207 (((loff_t) blocks) << 9);
1208}
1209
1210
1211static inline ulong to_fake_used_blocks(struct inode *inode, int sd_size)
1212{
1213 loff_t bytes = inode_get_bytes(inode);
1214 loff_t real_space = real_space_diff(inode, sd_size);
1215
1216
1217 if (S_ISLNK(inode->i_mode) || S_ISDIR(inode->i_mode)) {
1218 bytes += (loff_t) 511;
1219 }
1220
1221
1222
1223
1224
1225
1226 if (bytes < real_space)
1227 return 0;
1228 return (bytes - real_space) >> 9;
1229}
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239static void init_inode(struct inode *inode, struct treepath *path)
1240{
1241 struct buffer_head *bh;
1242 struct item_head *ih;
1243 __u32 rdev;
1244
1245 bh = PATH_PLAST_BUFFER(path);
1246 ih = tp_item_head(path);
1247
1248 copy_key(INODE_PKEY(inode), &ih->ih_key);
1249
1250 INIT_LIST_HEAD(&REISERFS_I(inode)->i_prealloc_list);
1251 REISERFS_I(inode)->i_flags = 0;
1252 REISERFS_I(inode)->i_prealloc_block = 0;
1253 REISERFS_I(inode)->i_prealloc_count = 0;
1254 REISERFS_I(inode)->i_trans_id = 0;
1255 REISERFS_I(inode)->i_jl = NULL;
1256 reiserfs_init_xattr_rwsem(inode);
1257
1258 if (stat_data_v1(ih)) {
1259 struct stat_data_v1 *sd =
1260 (struct stat_data_v1 *)ih_item_body(bh, ih);
1261 unsigned long blocks;
1262
1263 set_inode_item_key_version(inode, KEY_FORMAT_3_5);
1264 set_inode_sd_version(inode, STAT_DATA_V1);
1265 inode->i_mode = sd_v1_mode(sd);
1266 set_nlink(inode, sd_v1_nlink(sd));
1267 i_uid_write(inode, sd_v1_uid(sd));
1268 i_gid_write(inode, sd_v1_gid(sd));
1269 inode->i_size = sd_v1_size(sd);
1270 inode->i_atime.tv_sec = sd_v1_atime(sd);
1271 inode->i_mtime.tv_sec = sd_v1_mtime(sd);
1272 inode->i_ctime.tv_sec = sd_v1_ctime(sd);
1273 inode->i_atime.tv_nsec = 0;
1274 inode->i_ctime.tv_nsec = 0;
1275 inode->i_mtime.tv_nsec = 0;
1276
1277 inode->i_blocks = sd_v1_blocks(sd);
1278 inode->i_generation = le32_to_cpu(INODE_PKEY(inode)->k_dir_id);
1279 blocks = (inode->i_size + 511) >> 9;
1280 blocks = _ROUND_UP(blocks, inode->i_sb->s_blocksize >> 9);
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290 if (inode->i_blocks > blocks) {
1291 inode->i_blocks = blocks;
1292 }
1293
1294 rdev = sd_v1_rdev(sd);
1295 REISERFS_I(inode)->i_first_direct_byte =
1296 sd_v1_first_direct_byte(sd);
1297
1298
1299
1300
1301
1302 if (inode->i_blocks & 1) {
1303 inode->i_blocks++;
1304 }
1305 inode_set_bytes(inode,
1306 to_real_used_space(inode, inode->i_blocks,
1307 SD_V1_SIZE));
1308
1309
1310
1311
1312 REISERFS_I(inode)->i_flags &= ~i_nopack_mask;
1313 } else {
1314
1315
1316
1317
1318 struct stat_data *sd = (struct stat_data *)ih_item_body(bh, ih);
1319
1320 inode->i_mode = sd_v2_mode(sd);
1321 set_nlink(inode, sd_v2_nlink(sd));
1322 i_uid_write(inode, sd_v2_uid(sd));
1323 inode->i_size = sd_v2_size(sd);
1324 i_gid_write(inode, sd_v2_gid(sd));
1325 inode->i_mtime.tv_sec = sd_v2_mtime(sd);
1326 inode->i_atime.tv_sec = sd_v2_atime(sd);
1327 inode->i_ctime.tv_sec = sd_v2_ctime(sd);
1328 inode->i_ctime.tv_nsec = 0;
1329 inode->i_mtime.tv_nsec = 0;
1330 inode->i_atime.tv_nsec = 0;
1331 inode->i_blocks = sd_v2_blocks(sd);
1332 rdev = sd_v2_rdev(sd);
1333 if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode))
1334 inode->i_generation =
1335 le32_to_cpu(INODE_PKEY(inode)->k_dir_id);
1336 else
1337 inode->i_generation = sd_v2_generation(sd);
1338
1339 if (S_ISDIR(inode->i_mode) || S_ISLNK(inode->i_mode))
1340 set_inode_item_key_version(inode, KEY_FORMAT_3_5);
1341 else
1342 set_inode_item_key_version(inode, KEY_FORMAT_3_6);
1343 REISERFS_I(inode)->i_first_direct_byte = 0;
1344 set_inode_sd_version(inode, STAT_DATA_V2);
1345 inode_set_bytes(inode,
1346 to_real_used_space(inode, inode->i_blocks,
1347 SD_V2_SIZE));
1348
1349
1350
1351
1352 REISERFS_I(inode)->i_attrs = sd_v2_attrs(sd);
1353 sd_attrs_to_i_attrs(sd_v2_attrs(sd), inode);
1354 }
1355
1356 pathrelse(path);
1357 if (S_ISREG(inode->i_mode)) {
1358 inode->i_op = &reiserfs_file_inode_operations;
1359 inode->i_fop = &reiserfs_file_operations;
1360 inode->i_mapping->a_ops = &reiserfs_address_space_operations;
1361 } else if (S_ISDIR(inode->i_mode)) {
1362 inode->i_op = &reiserfs_dir_inode_operations;
1363 inode->i_fop = &reiserfs_dir_operations;
1364 } else if (S_ISLNK(inode->i_mode)) {
1365 inode->i_op = &reiserfs_symlink_inode_operations;
1366 inode_nohighmem(inode);
1367 inode->i_mapping->a_ops = &reiserfs_address_space_operations;
1368 } else {
1369 inode->i_blocks = 0;
1370 inode->i_op = &reiserfs_special_inode_operations;
1371 init_special_inode(inode, inode->i_mode, new_decode_dev(rdev));
1372 }
1373}
1374
1375
1376static void inode2sd(void *sd, struct inode *inode, loff_t size)
1377{
1378 struct stat_data *sd_v2 = (struct stat_data *)sd;
1379
1380 set_sd_v2_mode(sd_v2, inode->i_mode);
1381 set_sd_v2_nlink(sd_v2, inode->i_nlink);
1382 set_sd_v2_uid(sd_v2, i_uid_read(inode));
1383 set_sd_v2_size(sd_v2, size);
1384 set_sd_v2_gid(sd_v2, i_gid_read(inode));
1385 set_sd_v2_mtime(sd_v2, inode->i_mtime.tv_sec);
1386 set_sd_v2_atime(sd_v2, inode->i_atime.tv_sec);
1387 set_sd_v2_ctime(sd_v2, inode->i_ctime.tv_sec);
1388 set_sd_v2_blocks(sd_v2, to_fake_used_blocks(inode, SD_V2_SIZE));
1389 if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode))
1390 set_sd_v2_rdev(sd_v2, new_encode_dev(inode->i_rdev));
1391 else
1392 set_sd_v2_generation(sd_v2, inode->i_generation);
1393 set_sd_v2_attrs(sd_v2, REISERFS_I(inode)->i_attrs);
1394}
1395
1396
1397static void inode2sd_v1(void *sd, struct inode *inode, loff_t size)
1398{
1399 struct stat_data_v1 *sd_v1 = (struct stat_data_v1 *)sd;
1400
1401 set_sd_v1_mode(sd_v1, inode->i_mode);
1402 set_sd_v1_uid(sd_v1, i_uid_read(inode));
1403 set_sd_v1_gid(sd_v1, i_gid_read(inode));
1404 set_sd_v1_nlink(sd_v1, inode->i_nlink);
1405 set_sd_v1_size(sd_v1, size);
1406 set_sd_v1_atime(sd_v1, inode->i_atime.tv_sec);
1407 set_sd_v1_ctime(sd_v1, inode->i_ctime.tv_sec);
1408 set_sd_v1_mtime(sd_v1, inode->i_mtime.tv_sec);
1409
1410 if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode))
1411 set_sd_v1_rdev(sd_v1, new_encode_dev(inode->i_rdev));
1412 else
1413 set_sd_v1_blocks(sd_v1, to_fake_used_blocks(inode, SD_V1_SIZE));
1414
1415
1416 set_sd_v1_first_direct_byte(sd_v1,
1417 REISERFS_I(inode)->i_first_direct_byte);
1418}
1419
1420
1421
1422
1423
1424static void update_stat_data(struct treepath *path, struct inode *inode,
1425 loff_t size)
1426{
1427 struct buffer_head *bh;
1428 struct item_head *ih;
1429
1430 bh = PATH_PLAST_BUFFER(path);
1431 ih = tp_item_head(path);
1432
1433 if (!is_statdata_le_ih(ih))
1434 reiserfs_panic(inode->i_sb, "vs-13065", "key %k, found item %h",
1435 INODE_PKEY(inode), ih);
1436
1437
1438 if (stat_data_v1(ih)) {
1439 inode2sd_v1(ih_item_body(bh, ih), inode, size);
1440 } else {
1441 inode2sd(ih_item_body(bh, ih), inode, size);
1442 }
1443
1444 return;
1445}
1446
1447void reiserfs_update_sd_size(struct reiserfs_transaction_handle *th,
1448 struct inode *inode, loff_t size)
1449{
1450 struct cpu_key key;
1451 INITIALIZE_PATH(path);
1452 struct buffer_head *bh;
1453 int fs_gen;
1454 struct item_head *ih, tmp_ih;
1455 int retval;
1456
1457 BUG_ON(!th->t_trans_id);
1458
1459
1460 make_cpu_key(&key, inode, SD_OFFSET, TYPE_STAT_DATA, 3);
1461
1462 for (;;) {
1463 int pos;
1464
1465 retval = search_item(inode->i_sb, &key, &path);
1466 if (retval == IO_ERROR) {
1467 reiserfs_error(inode->i_sb, "vs-13050",
1468 "i/o failure occurred trying to "
1469 "update %K stat data", &key);
1470 return;
1471 }
1472 if (retval == ITEM_NOT_FOUND) {
1473 pos = PATH_LAST_POSITION(&path);
1474 pathrelse(&path);
1475 if (inode->i_nlink == 0) {
1476
1477 return;
1478 }
1479 reiserfs_warning(inode->i_sb, "vs-13060",
1480 "stat data of object %k (nlink == %d) "
1481 "not found (pos %d)",
1482 INODE_PKEY(inode), inode->i_nlink,
1483 pos);
1484 reiserfs_check_path(&path);
1485 return;
1486 }
1487
1488
1489
1490
1491
1492
1493 bh = get_last_bh(&path);
1494 ih = tp_item_head(&path);
1495 copy_item_head(&tmp_ih, ih);
1496 fs_gen = get_generation(inode->i_sb);
1497 reiserfs_prepare_for_journal(inode->i_sb, bh, 1);
1498
1499
1500 if (fs_changed(fs_gen, inode->i_sb)
1501 && item_moved(&tmp_ih, &path)) {
1502 reiserfs_restore_prepared_buffer(inode->i_sb, bh);
1503 continue;
1504 }
1505 break;
1506 }
1507 update_stat_data(&path, inode, size);
1508 journal_mark_dirty(th, bh);
1509 pathrelse(&path);
1510 return;
1511}
1512
1513
1514
1515
1516
1517
1518
1519
1520static void reiserfs_make_bad_inode(struct inode *inode)
1521{
1522 memset(INODE_PKEY(inode), 0, KEY_SIZE);
1523 make_bad_inode(inode);
1524}
1525
1526
1527
1528
1529
1530int reiserfs_init_locked_inode(struct inode *inode, void *p)
1531{
1532 struct reiserfs_iget_args *args = (struct reiserfs_iget_args *)p;
1533 inode->i_ino = args->objectid;
1534 INODE_PKEY(inode)->k_dir_id = cpu_to_le32(args->dirid);
1535 return 0;
1536}
1537
1538
1539
1540
1541
1542void reiserfs_read_locked_inode(struct inode *inode,
1543 struct reiserfs_iget_args *args)
1544{
1545 INITIALIZE_PATH(path_to_sd);
1546 struct cpu_key key;
1547 unsigned long dirino;
1548 int retval;
1549
1550 dirino = args->dirid;
1551
1552
1553
1554
1555
1556 key.version = KEY_FORMAT_3_5;
1557 key.on_disk_key.k_dir_id = dirino;
1558 key.on_disk_key.k_objectid = inode->i_ino;
1559 key.on_disk_key.k_offset = 0;
1560 key.on_disk_key.k_type = 0;
1561
1562
1563 retval = search_item(inode->i_sb, &key, &path_to_sd);
1564 if (retval == IO_ERROR) {
1565 reiserfs_error(inode->i_sb, "vs-13070",
1566 "i/o failure occurred trying to find "
1567 "stat data of %K", &key);
1568 reiserfs_make_bad_inode(inode);
1569 return;
1570 }
1571
1572
1573 if (retval != ITEM_FOUND) {
1574 pathrelse(&path_to_sd);
1575 reiserfs_make_bad_inode(inode);
1576 clear_nlink(inode);
1577 return;
1578 }
1579
1580 init_inode(inode, &path_to_sd);
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601 if ((inode->i_nlink == 0) &&
1602 !REISERFS_SB(inode->i_sb)->s_is_unlinked_ok) {
1603 reiserfs_warning(inode->i_sb, "vs-13075",
1604 "dead inode read from disk %K. "
1605 "This is likely to be race with knfsd. Ignore",
1606 &key);
1607 reiserfs_make_bad_inode(inode);
1608 }
1609
1610
1611 reiserfs_check_path(&path_to_sd);
1612
1613
1614
1615
1616 if (get_inode_sd_version(inode) == STAT_DATA_V1)
1617 cache_no_acl(inode);
1618}
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632int reiserfs_find_actor(struct inode *inode, void *opaque)
1633{
1634 struct reiserfs_iget_args *args;
1635
1636 args = opaque;
1637
1638 return (inode->i_ino == args->objectid) &&
1639 (le32_to_cpu(INODE_PKEY(inode)->k_dir_id) == args->dirid);
1640}
1641
1642struct inode *reiserfs_iget(struct super_block *s, const struct cpu_key *key)
1643{
1644 struct inode *inode;
1645 struct reiserfs_iget_args args;
1646 int depth;
1647
1648 args.objectid = key->on_disk_key.k_objectid;
1649 args.dirid = key->on_disk_key.k_dir_id;
1650 depth = reiserfs_write_unlock_nested(s);
1651 inode = iget5_locked(s, key->on_disk_key.k_objectid,
1652 reiserfs_find_actor, reiserfs_init_locked_inode,
1653 (void *)(&args));
1654 reiserfs_write_lock_nested(s, depth);
1655 if (!inode)
1656 return ERR_PTR(-ENOMEM);
1657
1658 if (inode->i_state & I_NEW) {
1659 reiserfs_read_locked_inode(inode, &args);
1660 unlock_new_inode(inode);
1661 }
1662
1663 if (comp_short_keys(INODE_PKEY(inode), key) || is_bad_inode(inode)) {
1664
1665 iput(inode);
1666 inode = NULL;
1667 }
1668 return inode;
1669}
1670
1671static struct dentry *reiserfs_get_dentry(struct super_block *sb,
1672 u32 objectid, u32 dir_id, u32 generation)
1673
1674{
1675 struct cpu_key key;
1676 struct inode *inode;
1677
1678 key.on_disk_key.k_objectid = objectid;
1679 key.on_disk_key.k_dir_id = dir_id;
1680 reiserfs_write_lock(sb);
1681 inode = reiserfs_iget(sb, &key);
1682 if (inode && !IS_ERR(inode) && generation != 0 &&
1683 generation != inode->i_generation) {
1684 iput(inode);
1685 inode = NULL;
1686 }
1687 reiserfs_write_unlock(sb);
1688
1689 return d_obtain_alias(inode);
1690}
1691
1692struct dentry *reiserfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
1693 int fh_len, int fh_type)
1694{
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708 if (fh_type > fh_len) {
1709 if (fh_type != 6 || fh_len != 5)
1710 reiserfs_warning(sb, "reiserfs-13077",
1711 "nfsd/reiserfs, fhtype=%d, len=%d - odd",
1712 fh_type, fh_len);
1713 fh_type = fh_len;
1714 }
1715 if (fh_len < 2)
1716 return NULL;
1717
1718 return reiserfs_get_dentry(sb, fid->raw[0], fid->raw[1],
1719 (fh_type == 3 || fh_type >= 5) ? fid->raw[2] : 0);
1720}
1721
1722struct dentry *reiserfs_fh_to_parent(struct super_block *sb, struct fid *fid,
1723 int fh_len, int fh_type)
1724{
1725 if (fh_type > fh_len)
1726 fh_type = fh_len;
1727 if (fh_type < 4)
1728 return NULL;
1729
1730 return reiserfs_get_dentry(sb,
1731 (fh_type >= 5) ? fid->raw[3] : fid->raw[2],
1732 (fh_type >= 5) ? fid->raw[4] : fid->raw[3],
1733 (fh_type == 6) ? fid->raw[5] : 0);
1734}
1735
1736int reiserfs_encode_fh(struct inode *inode, __u32 * data, int *lenp,
1737 struct inode *parent)
1738{
1739 int maxlen = *lenp;
1740
1741 if (parent && (maxlen < 5)) {
1742 *lenp = 5;
1743 return FILEID_INVALID;
1744 } else if (maxlen < 3) {
1745 *lenp = 3;
1746 return FILEID_INVALID;
1747 }
1748
1749 data[0] = inode->i_ino;
1750 data[1] = le32_to_cpu(INODE_PKEY(inode)->k_dir_id);
1751 data[2] = inode->i_generation;
1752 *lenp = 3;
1753 if (parent) {
1754 data[3] = parent->i_ino;
1755 data[4] = le32_to_cpu(INODE_PKEY(parent)->k_dir_id);
1756 *lenp = 5;
1757 if (maxlen >= 6) {
1758 data[5] = parent->i_generation;
1759 *lenp = 6;
1760 }
1761 }
1762 return *lenp;
1763}
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775int reiserfs_write_inode(struct inode *inode, struct writeback_control *wbc)
1776{
1777 struct reiserfs_transaction_handle th;
1778 int jbegin_count = 1;
1779
1780 if (sb_rdonly(inode->i_sb))
1781 return -EROFS;
1782
1783
1784
1785
1786
1787
1788
1789 if (wbc->sync_mode == WB_SYNC_ALL && !(current->flags & PF_MEMALLOC)) {
1790 reiserfs_write_lock(inode->i_sb);
1791 if (!journal_begin(&th, inode->i_sb, jbegin_count)) {
1792 reiserfs_update_sd(&th, inode);
1793 journal_end_sync(&th);
1794 }
1795 reiserfs_write_unlock(inode->i_sb);
1796 }
1797 return 0;
1798}
1799
1800
1801
1802
1803
1804static int reiserfs_new_directory(struct reiserfs_transaction_handle *th,
1805 struct inode *inode,
1806 struct item_head *ih, struct treepath *path,
1807 struct inode *dir)
1808{
1809 struct super_block *sb = th->t_super;
1810 char empty_dir[EMPTY_DIR_SIZE];
1811 char *body = empty_dir;
1812 struct cpu_key key;
1813 int retval;
1814
1815 BUG_ON(!th->t_trans_id);
1816
1817 _make_cpu_key(&key, KEY_FORMAT_3_5, le32_to_cpu(ih->ih_key.k_dir_id),
1818 le32_to_cpu(ih->ih_key.k_objectid), DOT_OFFSET,
1819 TYPE_DIRENTRY, 3 );
1820
1821
1822
1823
1824
1825
1826 if (old_format_only(sb)) {
1827 make_le_item_head(ih, NULL, KEY_FORMAT_3_5, DOT_OFFSET,
1828 TYPE_DIRENTRY, EMPTY_DIR_SIZE_V1, 2);
1829
1830 make_empty_dir_item_v1(body, ih->ih_key.k_dir_id,
1831 ih->ih_key.k_objectid,
1832 INODE_PKEY(dir)->k_dir_id,
1833 INODE_PKEY(dir)->k_objectid);
1834 } else {
1835 make_le_item_head(ih, NULL, KEY_FORMAT_3_5, DOT_OFFSET,
1836 TYPE_DIRENTRY, EMPTY_DIR_SIZE, 2);
1837
1838 make_empty_dir_item(body, ih->ih_key.k_dir_id,
1839 ih->ih_key.k_objectid,
1840 INODE_PKEY(dir)->k_dir_id,
1841 INODE_PKEY(dir)->k_objectid);
1842 }
1843
1844
1845 retval = search_item(sb, &key, path);
1846 if (retval == IO_ERROR) {
1847 reiserfs_error(sb, "vs-13080",
1848 "i/o failure occurred creating new directory");
1849 return -EIO;
1850 }
1851 if (retval == ITEM_FOUND) {
1852 pathrelse(path);
1853 reiserfs_warning(sb, "vs-13070",
1854 "object with this key exists (%k)",
1855 &(ih->ih_key));
1856 return -EEXIST;
1857 }
1858
1859
1860 return reiserfs_insert_item(th, path, &key, ih, inode, body);
1861}
1862
1863
1864
1865
1866
1867static int reiserfs_new_symlink(struct reiserfs_transaction_handle *th,
1868 struct inode *inode,
1869 struct item_head *ih,
1870 struct treepath *path, const char *symname,
1871 int item_len)
1872{
1873 struct super_block *sb = th->t_super;
1874 struct cpu_key key;
1875 int retval;
1876
1877 BUG_ON(!th->t_trans_id);
1878
1879 _make_cpu_key(&key, KEY_FORMAT_3_5,
1880 le32_to_cpu(ih->ih_key.k_dir_id),
1881 le32_to_cpu(ih->ih_key.k_objectid),
1882 1, TYPE_DIRECT, 3 );
1883
1884 make_le_item_head(ih, NULL, KEY_FORMAT_3_5, 1, TYPE_DIRECT, item_len,
1885 0 );
1886
1887
1888 retval = search_item(sb, &key, path);
1889 if (retval == IO_ERROR) {
1890 reiserfs_error(sb, "vs-13080",
1891 "i/o failure occurred creating new symlink");
1892 return -EIO;
1893 }
1894 if (retval == ITEM_FOUND) {
1895 pathrelse(path);
1896 reiserfs_warning(sb, "vs-13080",
1897 "object with this key exists (%k)",
1898 &(ih->ih_key));
1899 return -EEXIST;
1900 }
1901
1902
1903 return reiserfs_insert_item(th, path, &key, ih, inode, symname);
1904}
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926int reiserfs_new_inode(struct reiserfs_transaction_handle *th,
1927 struct inode *dir, umode_t mode, const char *symname,
1928
1929
1930 loff_t i_size, struct dentry *dentry,
1931 struct inode *inode,
1932 struct reiserfs_security_handle *security)
1933{
1934 struct super_block *sb = dir->i_sb;
1935 struct reiserfs_iget_args args;
1936 INITIALIZE_PATH(path_to_key);
1937 struct cpu_key key;
1938 struct item_head ih;
1939 struct stat_data sd;
1940 int retval;
1941 int err;
1942 int depth;
1943
1944 BUG_ON(!th->t_trans_id);
1945
1946 depth = reiserfs_write_unlock_nested(sb);
1947 err = dquot_alloc_inode(inode);
1948 reiserfs_write_lock_nested(sb, depth);
1949 if (err)
1950 goto out_end_trans;
1951 if (!dir->i_nlink) {
1952 err = -EPERM;
1953 goto out_bad_inode;
1954 }
1955
1956
1957 ih.ih_key.k_dir_id = reiserfs_choose_packing(dir);
1958 ih.ih_key.k_objectid = cpu_to_le32(reiserfs_get_unused_objectid(th));
1959 if (!ih.ih_key.k_objectid) {
1960 err = -ENOMEM;
1961 goto out_bad_inode;
1962 }
1963 args.objectid = inode->i_ino = le32_to_cpu(ih.ih_key.k_objectid);
1964 if (old_format_only(sb))
1965 make_le_item_head(&ih, NULL, KEY_FORMAT_3_5, SD_OFFSET,
1966 TYPE_STAT_DATA, SD_V1_SIZE, MAX_US_INT);
1967 else
1968 make_le_item_head(&ih, NULL, KEY_FORMAT_3_6, SD_OFFSET,
1969 TYPE_STAT_DATA, SD_SIZE, MAX_US_INT);
1970 memcpy(INODE_PKEY(inode), &ih.ih_key, KEY_SIZE);
1971 args.dirid = le32_to_cpu(ih.ih_key.k_dir_id);
1972
1973 depth = reiserfs_write_unlock_nested(inode->i_sb);
1974 err = insert_inode_locked4(inode, args.objectid,
1975 reiserfs_find_actor, &args);
1976 reiserfs_write_lock_nested(inode->i_sb, depth);
1977 if (err) {
1978 err = -EINVAL;
1979 goto out_bad_inode;
1980 }
1981
1982 if (old_format_only(sb))
1983
1984
1985
1986
1987
1988
1989 inode->i_generation = le32_to_cpu(INODE_PKEY(dir)->k_objectid);
1990 else
1991#if defined( USE_INODE_GENERATION_COUNTER )
1992 inode->i_generation =
1993 le32_to_cpu(REISERFS_SB(sb)->s_rs->s_inode_generation);
1994#else
1995 inode->i_generation = ++event;
1996#endif
1997
1998
1999 set_nlink(inode, (S_ISDIR(mode) ? 2 : 1));
2000
2001
2002
2003 inode->i_mtime = inode->i_atime = inode->i_ctime = current_time(inode);
2004 inode->i_size = i_size;
2005 inode->i_blocks = 0;
2006 inode->i_bytes = 0;
2007 REISERFS_I(inode)->i_first_direct_byte = S_ISLNK(mode) ? 1 :
2008 U32_MAX ;
2009
2010 INIT_LIST_HEAD(&REISERFS_I(inode)->i_prealloc_list);
2011 REISERFS_I(inode)->i_flags = 0;
2012 REISERFS_I(inode)->i_prealloc_block = 0;
2013 REISERFS_I(inode)->i_prealloc_count = 0;
2014 REISERFS_I(inode)->i_trans_id = 0;
2015 REISERFS_I(inode)->i_jl = NULL;
2016 REISERFS_I(inode)->i_attrs =
2017 REISERFS_I(dir)->i_attrs & REISERFS_INHERIT_MASK;
2018 sd_attrs_to_i_attrs(REISERFS_I(inode)->i_attrs, inode);
2019 reiserfs_init_xattr_rwsem(inode);
2020
2021
2022 _make_cpu_key(&key, KEY_FORMAT_3_6, le32_to_cpu(ih.ih_key.k_dir_id),
2023 le32_to_cpu(ih.ih_key.k_objectid), SD_OFFSET,
2024 TYPE_STAT_DATA, 3 );
2025
2026
2027 retval = search_item(sb, &key, &path_to_key);
2028 if (retval == IO_ERROR) {
2029 err = -EIO;
2030 goto out_bad_inode;
2031 }
2032 if (retval == ITEM_FOUND) {
2033 pathrelse(&path_to_key);
2034 err = -EEXIST;
2035 goto out_bad_inode;
2036 }
2037 if (old_format_only(sb)) {
2038
2039 if (i_uid_read(inode) & ~0xffff || i_gid_read(inode) & ~0xffff) {
2040 pathrelse(&path_to_key);
2041 err = -EINVAL;
2042 goto out_bad_inode;
2043 }
2044 inode2sd_v1(&sd, inode, inode->i_size);
2045 } else {
2046 inode2sd(&sd, inode, inode->i_size);
2047 }
2048
2049
2050
2051
2052
2053 if (old_format_only(sb) || S_ISDIR(mode) || S_ISLNK(mode))
2054 set_inode_item_key_version(inode, KEY_FORMAT_3_5);
2055 else
2056 set_inode_item_key_version(inode, KEY_FORMAT_3_6);
2057 if (old_format_only(sb))
2058 set_inode_sd_version(inode, STAT_DATA_V1);
2059 else
2060 set_inode_sd_version(inode, STAT_DATA_V2);
2061
2062
2063#ifdef DISPLACE_NEW_PACKING_LOCALITIES
2064 if (REISERFS_I(dir)->new_packing_locality)
2065 th->displace_new_blocks = 1;
2066#endif
2067 retval =
2068 reiserfs_insert_item(th, &path_to_key, &key, &ih, inode,
2069 (char *)(&sd));
2070 if (retval) {
2071 err = retval;
2072 reiserfs_check_path(&path_to_key);
2073 goto out_bad_inode;
2074 }
2075#ifdef DISPLACE_NEW_PACKING_LOCALITIES
2076 if (!th->displace_new_blocks)
2077 REISERFS_I(dir)->new_packing_locality = 0;
2078#endif
2079 if (S_ISDIR(mode)) {
2080
2081 retval =
2082 reiserfs_new_directory(th, inode, &ih, &path_to_key, dir);
2083 }
2084
2085 if (S_ISLNK(mode)) {
2086
2087 if (!old_format_only(sb))
2088 i_size = ROUND_UP(i_size);
2089 retval =
2090 reiserfs_new_symlink(th, inode, &ih, &path_to_key, symname,
2091 i_size);
2092 }
2093 if (retval) {
2094 err = retval;
2095 reiserfs_check_path(&path_to_key);
2096 journal_end(th);
2097 goto out_inserted_sd;
2098 }
2099
2100 if (reiserfs_posixacl(inode->i_sb)) {
2101 reiserfs_write_unlock(inode->i_sb);
2102 retval = reiserfs_inherit_default_acl(th, dir, dentry, inode);
2103 reiserfs_write_lock(inode->i_sb);
2104 if (retval) {
2105 err = retval;
2106 reiserfs_check_path(&path_to_key);
2107 journal_end(th);
2108 goto out_inserted_sd;
2109 }
2110 } else if (inode->i_sb->s_flags & SB_POSIXACL) {
2111 reiserfs_warning(inode->i_sb, "jdm-13090",
2112 "ACLs aren't enabled in the fs, "
2113 "but vfs thinks they are!");
2114 } else if (IS_PRIVATE(dir))
2115 inode->i_flags |= S_PRIVATE;
2116
2117 if (security->name) {
2118 reiserfs_write_unlock(inode->i_sb);
2119 retval = reiserfs_security_write(th, inode, security);
2120 reiserfs_write_lock(inode->i_sb);
2121 if (retval) {
2122 err = retval;
2123 reiserfs_check_path(&path_to_key);
2124 retval = journal_end(th);
2125 if (retval)
2126 err = retval;
2127 goto out_inserted_sd;
2128 }
2129 }
2130
2131 reiserfs_update_sd(th, inode);
2132 reiserfs_check_path(&path_to_key);
2133
2134 return 0;
2135
2136out_bad_inode:
2137
2138 INODE_PKEY(inode)->k_objectid = 0;
2139
2140
2141 depth = reiserfs_write_unlock_nested(inode->i_sb);
2142 dquot_free_inode(inode);
2143 reiserfs_write_lock_nested(inode->i_sb, depth);
2144
2145out_end_trans:
2146 journal_end(th);
2147
2148
2149
2150
2151 depth = reiserfs_write_unlock_nested(inode->i_sb);
2152 dquot_drop(inode);
2153 reiserfs_write_lock_nested(inode->i_sb, depth);
2154 inode->i_flags |= S_NOQUOTA;
2155 make_bad_inode(inode);
2156
2157out_inserted_sd:
2158 clear_nlink(inode);
2159 th->t_trans_id = 0;
2160 unlock_new_inode(inode);
2161 iput(inode);
2162 return err;
2163}
2164
2165
2166
2167
2168
2169
2170
2171
2172
2173
2174
2175
2176
2177
2178static int grab_tail_page(struct inode *inode,
2179 struct page **page_result,
2180 struct buffer_head **bh_result)
2181{
2182
2183
2184
2185
2186
2187 unsigned long index = (inode->i_size - 1) >> PAGE_SHIFT;
2188 unsigned long pos = 0;
2189 unsigned long start = 0;
2190 unsigned long blocksize = inode->i_sb->s_blocksize;
2191 unsigned long offset = (inode->i_size) & (PAGE_SIZE - 1);
2192 struct buffer_head *bh;
2193 struct buffer_head *head;
2194 struct page *page;
2195 int error;
2196
2197
2198
2199
2200
2201
2202
2203 if ((offset & (blocksize - 1)) == 0) {
2204 return -ENOENT;
2205 }
2206 page = grab_cache_page(inode->i_mapping, index);
2207 error = -ENOMEM;
2208 if (!page) {
2209 goto out;
2210 }
2211
2212 start = (offset / blocksize) * blocksize;
2213
2214 error = __block_write_begin(page, start, offset - start,
2215 reiserfs_get_block_create_0);
2216 if (error)
2217 goto unlock;
2218
2219 head = page_buffers(page);
2220 bh = head;
2221 do {
2222 if (pos >= start) {
2223 break;
2224 }
2225 bh = bh->b_this_page;
2226 pos += blocksize;
2227 } while (bh != head);
2228
2229 if (!buffer_uptodate(bh)) {
2230
2231
2232
2233
2234
2235
2236 reiserfs_error(inode->i_sb, "clm-6000",
2237 "error reading block %lu", bh->b_blocknr);
2238 error = -EIO;
2239 goto unlock;
2240 }
2241 *bh_result = bh;
2242 *page_result = page;
2243
2244out:
2245 return error;
2246
2247unlock:
2248 unlock_page(page);
2249 put_page(page);
2250 return error;
2251}
2252
2253
2254
2255
2256
2257
2258
2259int reiserfs_truncate_file(struct inode *inode, int update_timestamps)
2260{
2261 struct reiserfs_transaction_handle th;
2262
2263 unsigned long offset = inode->i_size & (PAGE_SIZE - 1);
2264 unsigned blocksize = inode->i_sb->s_blocksize;
2265 unsigned length;
2266 struct page *page = NULL;
2267 int error;
2268 struct buffer_head *bh = NULL;
2269 int err2;
2270
2271 reiserfs_write_lock(inode->i_sb);
2272
2273 if (inode->i_size > 0) {
2274 error = grab_tail_page(inode, &page, &bh);
2275 if (error) {
2276
2277
2278
2279
2280
2281 if (error != -ENOENT)
2282 reiserfs_error(inode->i_sb, "clm-6001",
2283 "grab_tail_page failed %d",
2284 error);
2285 page = NULL;
2286 bh = NULL;
2287 }
2288 }
2289
2290
2291
2292
2293
2294
2295
2296
2297
2298
2299
2300
2301
2302 error = journal_begin(&th, inode->i_sb,
2303 JOURNAL_PER_BALANCE_CNT * 2 + 1);
2304 if (error)
2305 goto out;
2306 reiserfs_update_inode_transaction(inode);
2307 if (update_timestamps)
2308
2309
2310
2311
2312
2313
2314 add_save_link(&th, inode, 1);
2315 err2 = reiserfs_do_truncate(&th, inode, page, update_timestamps);
2316 error = journal_end(&th);
2317 if (error)
2318 goto out;
2319
2320
2321 if (err2) {
2322 error = err2;
2323 goto out;
2324 }
2325
2326 if (update_timestamps) {
2327 error = remove_save_link(inode, 1 );
2328 if (error)
2329 goto out;
2330 }
2331
2332 if (page) {
2333 length = offset & (blocksize - 1);
2334
2335 if (length) {
2336 length = blocksize - length;
2337 zero_user(page, offset, length);
2338 if (buffer_mapped(bh) && bh->b_blocknr != 0) {
2339 mark_buffer_dirty(bh);
2340 }
2341 }
2342 unlock_page(page);
2343 put_page(page);
2344 }
2345
2346 reiserfs_write_unlock(inode->i_sb);
2347
2348 return 0;
2349out:
2350 if (page) {
2351 unlock_page(page);
2352 put_page(page);
2353 }
2354
2355 reiserfs_write_unlock(inode->i_sb);
2356
2357 return error;
2358}
2359
2360static int map_block_for_writepage(struct inode *inode,
2361 struct buffer_head *bh_result,
2362 unsigned long block)
2363{
2364 struct reiserfs_transaction_handle th;
2365 int fs_gen;
2366 struct item_head tmp_ih;
2367 struct item_head *ih;
2368 struct buffer_head *bh;
2369 __le32 *item;
2370 struct cpu_key key;
2371 INITIALIZE_PATH(path);
2372 int pos_in_item;
2373 int jbegin_count = JOURNAL_PER_BALANCE_CNT;
2374 loff_t byte_offset = ((loff_t)block << inode->i_sb->s_blocksize_bits)+1;
2375 int retval;
2376 int use_get_block = 0;
2377 int bytes_copied = 0;
2378 int copy_size;
2379 int trans_running = 0;
2380
2381
2382
2383
2384
2385 th.t_trans_id = 0;
2386
2387 if (!buffer_uptodate(bh_result)) {
2388 return -EIO;
2389 }
2390
2391 kmap(bh_result->b_page);
2392start_over:
2393 reiserfs_write_lock(inode->i_sb);
2394 make_cpu_key(&key, inode, byte_offset, TYPE_ANY, 3);
2395
2396research:
2397 retval = search_for_position_by_key(inode->i_sb, &key, &path);
2398 if (retval != POSITION_FOUND) {
2399 use_get_block = 1;
2400 goto out;
2401 }
2402
2403 bh = get_last_bh(&path);
2404 ih = tp_item_head(&path);
2405 item = tp_item_body(&path);
2406 pos_in_item = path.pos_in_item;
2407
2408
2409 if (indirect_item_found(retval, ih)) {
2410 if (bytes_copied > 0) {
2411 reiserfs_warning(inode->i_sb, "clm-6002",
2412 "bytes_copied %d", bytes_copied);
2413 }
2414 if (!get_block_num(item, pos_in_item)) {
2415
2416 use_get_block = 1;
2417 goto out;
2418 }
2419 set_block_dev_mapped(bh_result,
2420 get_block_num(item, pos_in_item), inode);
2421 } else if (is_direct_le_ih(ih)) {
2422 char *p;
2423 p = page_address(bh_result->b_page);
2424 p += (byte_offset - 1) & (PAGE_SIZE - 1);
2425 copy_size = ih_item_len(ih) - pos_in_item;
2426
2427 fs_gen = get_generation(inode->i_sb);
2428 copy_item_head(&tmp_ih, ih);
2429
2430 if (!trans_running) {
2431
2432 retval = journal_begin(&th, inode->i_sb, jbegin_count);
2433 if (retval)
2434 goto out;
2435 reiserfs_update_inode_transaction(inode);
2436 trans_running = 1;
2437 if (fs_changed(fs_gen, inode->i_sb)
2438 && item_moved(&tmp_ih, &path)) {
2439 reiserfs_restore_prepared_buffer(inode->i_sb,
2440 bh);
2441 goto research;
2442 }
2443 }
2444
2445 reiserfs_prepare_for_journal(inode->i_sb, bh, 1);
2446
2447 if (fs_changed(fs_gen, inode->i_sb)
2448 && item_moved(&tmp_ih, &path)) {
2449 reiserfs_restore_prepared_buffer(inode->i_sb, bh);
2450 goto research;
2451 }
2452
2453 memcpy(ih_item_body(bh, ih) + pos_in_item, p + bytes_copied,
2454 copy_size);
2455
2456 journal_mark_dirty(&th, bh);
2457 bytes_copied += copy_size;
2458 set_block_dev_mapped(bh_result, 0, inode);
2459
2460
2461 if (bytes_copied < bh_result->b_size &&
2462 (byte_offset + bytes_copied) < inode->i_size) {
2463 set_cpu_key_k_offset(&key,
2464 cpu_key_k_offset(&key) +
2465 copy_size);
2466 goto research;
2467 }
2468 } else {
2469 reiserfs_warning(inode->i_sb, "clm-6003",
2470 "bad item inode %lu", inode->i_ino);
2471 retval = -EIO;
2472 goto out;
2473 }
2474 retval = 0;
2475
2476out:
2477 pathrelse(&path);
2478 if (trans_running) {
2479 int err = journal_end(&th);
2480 if (err)
2481 retval = err;
2482 trans_running = 0;
2483 }
2484 reiserfs_write_unlock(inode->i_sb);
2485
2486
2487 if (use_get_block) {
2488 retval = reiserfs_get_block(inode, block, bh_result,
2489 GET_BLOCK_CREATE | GET_BLOCK_NO_IMUX
2490 | GET_BLOCK_NO_DANGLE);
2491 if (!retval) {
2492 if (!buffer_mapped(bh_result)
2493 || bh_result->b_blocknr == 0) {
2494
2495 use_get_block = 0;
2496 goto start_over;
2497 }
2498 }
2499 }
2500 kunmap(bh_result->b_page);
2501
2502 if (!retval && buffer_mapped(bh_result) && bh_result->b_blocknr == 0) {
2503
2504
2505
2506
2507 lock_buffer(bh_result);
2508 clear_buffer_dirty(bh_result);
2509 unlock_buffer(bh_result);
2510 }
2511 return retval;
2512}
2513
2514
2515
2516
2517
2518
2519static int reiserfs_write_full_page(struct page *page,
2520 struct writeback_control *wbc)
2521{
2522 struct inode *inode = page->mapping->host;
2523 unsigned long end_index = inode->i_size >> PAGE_SHIFT;
2524 int error = 0;
2525 unsigned long block;
2526 sector_t last_block;
2527 struct buffer_head *head, *bh;
2528 int partial = 0;
2529 int nr = 0;
2530 int checked = PageChecked(page);
2531 struct reiserfs_transaction_handle th;
2532 struct super_block *s = inode->i_sb;
2533 int bh_per_page = PAGE_SIZE / s->s_blocksize;
2534 th.t_trans_id = 0;
2535
2536
2537 if (checked && (current->flags & PF_MEMALLOC)) {
2538 redirty_page_for_writepage(wbc, page);
2539 unlock_page(page);
2540 return 0;
2541 }
2542
2543
2544
2545
2546
2547
2548
2549 if (!page_has_buffers(page)) {
2550 create_empty_buffers(page, s->s_blocksize,
2551 (1 << BH_Dirty) | (1 << BH_Uptodate));
2552 }
2553 head = page_buffers(page);
2554
2555
2556
2557
2558
2559 if (page->index >= end_index) {
2560 unsigned last_offset;
2561
2562 last_offset = inode->i_size & (PAGE_SIZE - 1);
2563
2564 if (page->index >= end_index + 1 || !last_offset) {
2565 unlock_page(page);
2566 return 0;
2567 }
2568 zero_user_segment(page, last_offset, PAGE_SIZE);
2569 }
2570 bh = head;
2571 block = page->index << (PAGE_SHIFT - s->s_blocksize_bits);
2572 last_block = (i_size_read(inode) - 1) >> inode->i_blkbits;
2573
2574 do {
2575 if (block > last_block) {
2576
2577
2578
2579
2580
2581 clear_buffer_dirty(bh);
2582 set_buffer_uptodate(bh);
2583 } else if ((checked || buffer_dirty(bh)) &&
2584 (!buffer_mapped(bh) || (buffer_mapped(bh)
2585 && bh->b_blocknr ==
2586 0))) {
2587
2588
2589
2590
2591
2592 if ((error = map_block_for_writepage(inode, bh, block))) {
2593 goto fail;
2594 }
2595 }
2596 bh = bh->b_this_page;
2597 block++;
2598 } while (bh != head);
2599
2600
2601
2602
2603
2604
2605
2606 if (checked) {
2607 ClearPageChecked(page);
2608 reiserfs_write_lock(s);
2609 error = journal_begin(&th, s, bh_per_page + 1);
2610 if (error) {
2611 reiserfs_write_unlock(s);
2612 goto fail;
2613 }
2614 reiserfs_update_inode_transaction(inode);
2615 }
2616
2617 do {
2618 get_bh(bh);
2619 if (!buffer_mapped(bh))
2620 continue;
2621 if (buffer_mapped(bh) && bh->b_blocknr == 0)
2622 continue;
2623
2624 if (checked) {
2625 reiserfs_prepare_for_journal(s, bh, 1);
2626 journal_mark_dirty(&th, bh);
2627 continue;
2628 }
2629
2630
2631
2632
2633 if (wbc->sync_mode != WB_SYNC_NONE) {
2634 lock_buffer(bh);
2635 } else {
2636 if (!trylock_buffer(bh)) {
2637 redirty_page_for_writepage(wbc, page);
2638 continue;
2639 }
2640 }
2641 if (test_clear_buffer_dirty(bh)) {
2642 mark_buffer_async_write(bh);
2643 } else {
2644 unlock_buffer(bh);
2645 }
2646 } while ((bh = bh->b_this_page) != head);
2647
2648 if (checked) {
2649 error = journal_end(&th);
2650 reiserfs_write_unlock(s);
2651 if (error)
2652 goto fail;
2653 }
2654 BUG_ON(PageWriteback(page));
2655 set_page_writeback(page);
2656 unlock_page(page);
2657
2658
2659
2660
2661
2662
2663 do {
2664 struct buffer_head *next = bh->b_this_page;
2665 if (buffer_async_write(bh)) {
2666 submit_bh(REQ_OP_WRITE, 0, bh);
2667 nr++;
2668 }
2669 put_bh(bh);
2670 bh = next;
2671 } while (bh != head);
2672
2673 error = 0;
2674done:
2675 if (nr == 0) {
2676
2677
2678
2679
2680
2681
2682 bh = head;
2683 do {
2684 if (!buffer_uptodate(bh)) {
2685 partial = 1;
2686 break;
2687 }
2688 bh = bh->b_this_page;
2689 } while (bh != head);
2690 if (!partial)
2691 SetPageUptodate(page);
2692 end_page_writeback(page);
2693 }
2694 return error;
2695
2696fail:
2697
2698
2699
2700
2701
2702 ClearPageUptodate(page);
2703 bh = head;
2704 do {
2705 get_bh(bh);
2706 if (buffer_mapped(bh) && buffer_dirty(bh) && bh->b_blocknr) {
2707 lock_buffer(bh);
2708 mark_buffer_async_write(bh);
2709 } else {
2710
2711
2712
2713
2714 clear_buffer_dirty(bh);
2715 }
2716 bh = bh->b_this_page;
2717 } while (bh != head);
2718 SetPageError(page);
2719 BUG_ON(PageWriteback(page));
2720 set_page_writeback(page);
2721 unlock_page(page);
2722 do {
2723 struct buffer_head *next = bh->b_this_page;
2724 if (buffer_async_write(bh)) {
2725 clear_buffer_dirty(bh);
2726 submit_bh(REQ_OP_WRITE, 0, bh);
2727 nr++;
2728 }
2729 put_bh(bh);
2730 bh = next;
2731 } while (bh != head);
2732 goto done;
2733}
2734
2735static int reiserfs_readpage(struct file *f, struct page *page)
2736{
2737 return block_read_full_page(page, reiserfs_get_block);
2738}
2739
2740static int reiserfs_writepage(struct page *page, struct writeback_control *wbc)
2741{
2742 struct inode *inode = page->mapping->host;
2743 reiserfs_wait_on_write_block(inode->i_sb);
2744 return reiserfs_write_full_page(page, wbc);
2745}
2746
2747static void reiserfs_truncate_failed_write(struct inode *inode)
2748{
2749 truncate_inode_pages(inode->i_mapping, inode->i_size);
2750 reiserfs_truncate_file(inode, 0);
2751}
2752
2753static int reiserfs_write_begin(struct file *file,
2754 struct address_space *mapping,
2755 loff_t pos, unsigned len, unsigned flags,
2756 struct page **pagep, void **fsdata)
2757{
2758 struct inode *inode;
2759 struct page *page;
2760 pgoff_t index;
2761 int ret;
2762 int old_ref = 0;
2763
2764 inode = mapping->host;
2765 *fsdata = NULL;
2766 if (flags & AOP_FLAG_CONT_EXPAND &&
2767 (pos & (inode->i_sb->s_blocksize - 1)) == 0) {
2768 pos ++;
2769 *fsdata = (void *)(unsigned long)flags;
2770 }
2771
2772 index = pos >> PAGE_SHIFT;
2773 page = grab_cache_page_write_begin(mapping, index, flags);
2774 if (!page)
2775 return -ENOMEM;
2776 *pagep = page;
2777
2778 reiserfs_wait_on_write_block(inode->i_sb);
2779 fix_tail_page_for_writing(page);
2780 if (reiserfs_transaction_running(inode->i_sb)) {
2781 struct reiserfs_transaction_handle *th;
2782 th = (struct reiserfs_transaction_handle *)current->
2783 journal_info;
2784 BUG_ON(!th->t_refcount);
2785 BUG_ON(!th->t_trans_id);
2786 old_ref = th->t_refcount;
2787 th->t_refcount++;
2788 }
2789 ret = __block_write_begin(page, pos, len, reiserfs_get_block);
2790 if (ret && reiserfs_transaction_running(inode->i_sb)) {
2791 struct reiserfs_transaction_handle *th = current->journal_info;
2792
2793
2794
2795
2796
2797
2798
2799
2800
2801
2802
2803
2804
2805 if (th->t_refcount > old_ref) {
2806 if (old_ref)
2807 th->t_refcount--;
2808 else {
2809 int err;
2810 reiserfs_write_lock(inode->i_sb);
2811 err = reiserfs_end_persistent_transaction(th);
2812 reiserfs_write_unlock(inode->i_sb);
2813 if (err)
2814 ret = err;
2815 }
2816 }
2817 }
2818 if (ret) {
2819 unlock_page(page);
2820 put_page(page);
2821
2822 reiserfs_truncate_failed_write(inode);
2823 }
2824 return ret;
2825}
2826
2827int __reiserfs_write_begin(struct page *page, unsigned from, unsigned len)
2828{
2829 struct inode *inode = page->mapping->host;
2830 int ret;
2831 int old_ref = 0;
2832 int depth;
2833
2834 depth = reiserfs_write_unlock_nested(inode->i_sb);
2835 reiserfs_wait_on_write_block(inode->i_sb);
2836 reiserfs_write_lock_nested(inode->i_sb, depth);
2837
2838 fix_tail_page_for_writing(page);
2839 if (reiserfs_transaction_running(inode->i_sb)) {
2840 struct reiserfs_transaction_handle *th;
2841 th = (struct reiserfs_transaction_handle *)current->
2842 journal_info;
2843 BUG_ON(!th->t_refcount);
2844 BUG_ON(!th->t_trans_id);
2845 old_ref = th->t_refcount;
2846 th->t_refcount++;
2847 }
2848
2849 ret = __block_write_begin(page, from, len, reiserfs_get_block);
2850 if (ret && reiserfs_transaction_running(inode->i_sb)) {
2851 struct reiserfs_transaction_handle *th = current->journal_info;
2852
2853
2854
2855
2856
2857
2858
2859
2860
2861
2862
2863
2864
2865 if (th->t_refcount > old_ref) {
2866 if (old_ref)
2867 th->t_refcount--;
2868 else {
2869 int err;
2870 reiserfs_write_lock(inode->i_sb);
2871 err = reiserfs_end_persistent_transaction(th);
2872 reiserfs_write_unlock(inode->i_sb);
2873 if (err)
2874 ret = err;
2875 }
2876 }
2877 }
2878 return ret;
2879
2880}
2881
2882static sector_t reiserfs_aop_bmap(struct address_space *as, sector_t block)
2883{
2884 return generic_block_bmap(as, block, reiserfs_bmap);
2885}
2886
2887static int reiserfs_write_end(struct file *file, struct address_space *mapping,
2888 loff_t pos, unsigned len, unsigned copied,
2889 struct page *page, void *fsdata)
2890{
2891 struct inode *inode = page->mapping->host;
2892 int ret = 0;
2893 int update_sd = 0;
2894 struct reiserfs_transaction_handle *th;
2895 unsigned start;
2896 bool locked = false;
2897
2898 if ((unsigned long)fsdata & AOP_FLAG_CONT_EXPAND)
2899 pos ++;
2900
2901 reiserfs_wait_on_write_block(inode->i_sb);
2902 if (reiserfs_transaction_running(inode->i_sb))
2903 th = current->journal_info;
2904 else
2905 th = NULL;
2906
2907 start = pos & (PAGE_SIZE - 1);
2908 if (unlikely(copied < len)) {
2909 if (!PageUptodate(page))
2910 copied = 0;
2911
2912 page_zero_new_buffers(page, start + copied, start + len);
2913 }
2914 flush_dcache_page(page);
2915
2916 reiserfs_commit_page(inode, page, start, start + copied);
2917
2918
2919
2920
2921
2922
2923 if (pos + copied > inode->i_size) {
2924 struct reiserfs_transaction_handle myth;
2925 reiserfs_write_lock(inode->i_sb);
2926 locked = true;
2927
2928
2929
2930
2931
2932 if ((have_large_tails(inode->i_sb)
2933 && inode->i_size > i_block_size(inode) * 4)
2934 || (have_small_tails(inode->i_sb)
2935 && inode->i_size > i_block_size(inode)))
2936 REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask;
2937
2938 ret = journal_begin(&myth, inode->i_sb, 1);
2939 if (ret)
2940 goto journal_error;
2941
2942 reiserfs_update_inode_transaction(inode);
2943 inode->i_size = pos + copied;
2944
2945
2946
2947
2948
2949 mark_inode_dirty(inode);
2950 reiserfs_update_sd(&myth, inode);
2951 update_sd = 1;
2952 ret = journal_end(&myth);
2953 if (ret)
2954 goto journal_error;
2955 }
2956 if (th) {
2957 if (!locked) {
2958 reiserfs_write_lock(inode->i_sb);
2959 locked = true;
2960 }
2961 if (!update_sd)
2962 mark_inode_dirty(inode);
2963 ret = reiserfs_end_persistent_transaction(th);
2964 if (ret)
2965 goto out;
2966 }
2967
2968out:
2969 if (locked)
2970 reiserfs_write_unlock(inode->i_sb);
2971 unlock_page(page);
2972 put_page(page);
2973
2974 if (pos + len > inode->i_size)
2975 reiserfs_truncate_failed_write(inode);
2976
2977 return ret == 0 ? copied : ret;
2978
2979journal_error:
2980 reiserfs_write_unlock(inode->i_sb);
2981 locked = false;
2982 if (th) {
2983 if (!update_sd)
2984 reiserfs_update_sd(th, inode);
2985 ret = reiserfs_end_persistent_transaction(th);
2986 }
2987 goto out;
2988}
2989
2990int reiserfs_commit_write(struct file *f, struct page *page,
2991 unsigned from, unsigned to)
2992{
2993 struct inode *inode = page->mapping->host;
2994 loff_t pos = ((loff_t) page->index << PAGE_SHIFT) + to;
2995 int ret = 0;
2996 int update_sd = 0;
2997 struct reiserfs_transaction_handle *th = NULL;
2998 int depth;
2999
3000 depth = reiserfs_write_unlock_nested(inode->i_sb);
3001 reiserfs_wait_on_write_block(inode->i_sb);
3002 reiserfs_write_lock_nested(inode->i_sb, depth);
3003
3004 if (reiserfs_transaction_running(inode->i_sb)) {
3005 th = current->journal_info;
3006 }
3007 reiserfs_commit_page(inode, page, from, to);
3008
3009
3010
3011
3012
3013
3014 if (pos > inode->i_size) {
3015 struct reiserfs_transaction_handle myth;
3016
3017
3018
3019
3020
3021 if ((have_large_tails(inode->i_sb)
3022 && inode->i_size > i_block_size(inode) * 4)
3023 || (have_small_tails(inode->i_sb)
3024 && inode->i_size > i_block_size(inode)))
3025 REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask;
3026
3027 ret = journal_begin(&myth, inode->i_sb, 1);
3028 if (ret)
3029 goto journal_error;
3030
3031 reiserfs_update_inode_transaction(inode);
3032 inode->i_size = pos;
3033
3034
3035
3036
3037
3038 mark_inode_dirty(inode);
3039 reiserfs_update_sd(&myth, inode);
3040 update_sd = 1;
3041 ret = journal_end(&myth);
3042 if (ret)
3043 goto journal_error;
3044 }
3045 if (th) {
3046 if (!update_sd)
3047 mark_inode_dirty(inode);
3048 ret = reiserfs_end_persistent_transaction(th);
3049 if (ret)
3050 goto out;
3051 }
3052
3053out:
3054 return ret;
3055
3056journal_error:
3057 if (th) {
3058 if (!update_sd)
3059 reiserfs_update_sd(th, inode);
3060 ret = reiserfs_end_persistent_transaction(th);
3061 }
3062
3063 return ret;
3064}
3065
3066void sd_attrs_to_i_attrs(__u16 sd_attrs, struct inode *inode)
3067{
3068 if (reiserfs_attrs(inode->i_sb)) {
3069 if (sd_attrs & REISERFS_SYNC_FL)
3070 inode->i_flags |= S_SYNC;
3071 else
3072 inode->i_flags &= ~S_SYNC;
3073 if (sd_attrs & REISERFS_IMMUTABLE_FL)
3074 inode->i_flags |= S_IMMUTABLE;
3075 else
3076 inode->i_flags &= ~S_IMMUTABLE;
3077 if (sd_attrs & REISERFS_APPEND_FL)
3078 inode->i_flags |= S_APPEND;
3079 else
3080 inode->i_flags &= ~S_APPEND;
3081 if (sd_attrs & REISERFS_NOATIME_FL)
3082 inode->i_flags |= S_NOATIME;
3083 else
3084 inode->i_flags &= ~S_NOATIME;
3085 if (sd_attrs & REISERFS_NOTAIL_FL)
3086 REISERFS_I(inode)->i_flags |= i_nopack_mask;
3087 else
3088 REISERFS_I(inode)->i_flags &= ~i_nopack_mask;
3089 }
3090}
3091
3092
3093
3094
3095
3096static int invalidatepage_can_drop(struct inode *inode, struct buffer_head *bh)
3097{
3098 int ret = 1;
3099 struct reiserfs_journal *j = SB_JOURNAL(inode->i_sb);
3100
3101 lock_buffer(bh);
3102 spin_lock(&j->j_dirty_buffers_lock);
3103 if (!buffer_mapped(bh)) {
3104 goto free_jh;
3105 }
3106
3107
3108
3109
3110 if (reiserfs_file_data_log(inode)) {
3111
3112
3113
3114
3115 if (buffer_journaled(bh) || buffer_journal_dirty(bh)) {
3116 ret = 0;
3117 }
3118 } else if (buffer_dirty(bh)) {
3119 struct reiserfs_journal_list *jl;
3120 struct reiserfs_jh *jh = bh->b_private;
3121
3122
3123
3124
3125
3126
3127
3128
3129
3130
3131
3132
3133
3134
3135
3136 if (jh && (jl = jh->jl)
3137 && jl != SB_JOURNAL(inode->i_sb)->j_current_jl)
3138 ret = 0;
3139 }
3140free_jh:
3141 if (ret && bh->b_private) {
3142 reiserfs_free_jh(bh);
3143 }
3144 spin_unlock(&j->j_dirty_buffers_lock);
3145 unlock_buffer(bh);
3146 return ret;
3147}
3148
3149
3150static void reiserfs_invalidatepage(struct page *page, unsigned int offset,
3151 unsigned int length)
3152{
3153 struct buffer_head *head, *bh, *next;
3154 struct inode *inode = page->mapping->host;
3155 unsigned int curr_off = 0;
3156 unsigned int stop = offset + length;
3157 int partial_page = (offset || length < PAGE_SIZE);
3158 int ret = 1;
3159
3160 BUG_ON(!PageLocked(page));
3161
3162 if (!partial_page)
3163 ClearPageChecked(page);
3164
3165 if (!page_has_buffers(page))
3166 goto out;
3167
3168 head = page_buffers(page);
3169 bh = head;
3170 do {
3171 unsigned int next_off = curr_off + bh->b_size;
3172 next = bh->b_this_page;
3173
3174 if (next_off > stop)
3175 goto out;
3176
3177
3178
3179
3180 if (offset <= curr_off) {
3181 if (invalidatepage_can_drop(inode, bh))
3182 reiserfs_unmap_buffer(bh);
3183 else
3184 ret = 0;
3185 }
3186 curr_off = next_off;
3187 bh = next;
3188 } while (bh != head);
3189
3190
3191
3192
3193
3194
3195 if (!partial_page && ret) {
3196 ret = try_to_release_page(page, 0);
3197
3198 }
3199out:
3200 return;
3201}
3202
3203static int reiserfs_set_page_dirty(struct page *page)
3204{
3205 struct inode *inode = page->mapping->host;
3206 if (reiserfs_file_data_log(inode)) {
3207 SetPageChecked(page);
3208 return __set_page_dirty_nobuffers(page);
3209 }
3210 return __set_page_dirty_buffers(page);
3211}
3212
3213
3214
3215
3216
3217
3218
3219
3220
3221
3222static int reiserfs_releasepage(struct page *page, gfp_t unused_gfp_flags)
3223{
3224 struct inode *inode = page->mapping->host;
3225 struct reiserfs_journal *j = SB_JOURNAL(inode->i_sb);
3226 struct buffer_head *head;
3227 struct buffer_head *bh;
3228 int ret = 1;
3229
3230 WARN_ON(PageChecked(page));
3231 spin_lock(&j->j_dirty_buffers_lock);
3232 head = page_buffers(page);
3233 bh = head;
3234 do {
3235 if (bh->b_private) {
3236 if (!buffer_dirty(bh) && !buffer_locked(bh)) {
3237 reiserfs_free_jh(bh);
3238 } else {
3239 ret = 0;
3240 break;
3241 }
3242 }
3243 bh = bh->b_this_page;
3244 } while (bh != head);
3245 if (ret)
3246 ret = try_to_free_buffers(page);
3247 spin_unlock(&j->j_dirty_buffers_lock);
3248 return ret;
3249}
3250
3251
3252
3253
3254
3255static ssize_t reiserfs_direct_IO(struct kiocb *iocb, struct iov_iter *iter)
3256{
3257 struct file *file = iocb->ki_filp;
3258 struct inode *inode = file->f_mapping->host;
3259 size_t count = iov_iter_count(iter);
3260 ssize_t ret;
3261
3262 ret = blockdev_direct_IO(iocb, inode, iter,
3263 reiserfs_get_blocks_direct_io);
3264
3265
3266
3267
3268
3269 if (unlikely(iov_iter_rw(iter) == WRITE && ret < 0)) {
3270 loff_t isize = i_size_read(inode);
3271 loff_t end = iocb->ki_pos + count;
3272
3273 if ((end > isize) && inode_newsize_ok(inode, isize) == 0) {
3274 truncate_setsize(inode, isize);
3275 reiserfs_vfs_truncate_file(inode);
3276 }
3277 }
3278
3279 return ret;
3280}
3281
3282int reiserfs_setattr(struct dentry *dentry, struct iattr *attr)
3283{
3284 struct inode *inode = d_inode(dentry);
3285 unsigned int ia_valid;
3286 int error;
3287
3288 error = setattr_prepare(dentry, attr);
3289 if (error)
3290 return error;
3291
3292
3293 ia_valid = attr->ia_valid &= ~(ATTR_KILL_SUID|ATTR_KILL_SGID);
3294
3295 if (is_quota_modification(inode, attr)) {
3296 error = dquot_initialize(inode);
3297 if (error)
3298 return error;
3299 }
3300 reiserfs_write_lock(inode->i_sb);
3301 if (attr->ia_valid & ATTR_SIZE) {
3302
3303
3304
3305
3306 if (get_inode_item_key_version(inode) == KEY_FORMAT_3_5 &&
3307 attr->ia_size > MAX_NON_LFS) {
3308 reiserfs_write_unlock(inode->i_sb);
3309 error = -EFBIG;
3310 goto out;
3311 }
3312
3313 inode_dio_wait(inode);
3314
3315
3316 if (attr->ia_size > inode->i_size) {
3317 error = generic_cont_expand_simple(inode, attr->ia_size);
3318 if (REISERFS_I(inode)->i_prealloc_count > 0) {
3319 int err;
3320 struct reiserfs_transaction_handle th;
3321
3322 err = journal_begin(&th, inode->i_sb, 4);
3323 if (!err) {
3324 reiserfs_discard_prealloc(&th, inode);
3325 err = journal_end(&th);
3326 }
3327 if (err)
3328 error = err;
3329 }
3330 if (error) {
3331 reiserfs_write_unlock(inode->i_sb);
3332 goto out;
3333 }
3334
3335
3336
3337
3338 attr->ia_valid |= (ATTR_MTIME | ATTR_CTIME);
3339 }
3340 }
3341 reiserfs_write_unlock(inode->i_sb);
3342
3343 if ((((attr->ia_valid & ATTR_UID) && (from_kuid(&init_user_ns, attr->ia_uid) & ~0xffff)) ||
3344 ((attr->ia_valid & ATTR_GID) && (from_kgid(&init_user_ns, attr->ia_gid) & ~0xffff))) &&
3345 (get_inode_sd_version(inode) == STAT_DATA_V1)) {
3346
3347 error = -EINVAL;
3348 goto out;
3349 }
3350
3351 if ((ia_valid & ATTR_UID && !uid_eq(attr->ia_uid, inode->i_uid)) ||
3352 (ia_valid & ATTR_GID && !gid_eq(attr->ia_gid, inode->i_gid))) {
3353 struct reiserfs_transaction_handle th;
3354 int jbegin_count =
3355 2 *
3356 (REISERFS_QUOTA_INIT_BLOCKS(inode->i_sb) +
3357 REISERFS_QUOTA_DEL_BLOCKS(inode->i_sb)) +
3358 2;
3359
3360 error = reiserfs_chown_xattrs(inode, attr);
3361
3362 if (error)
3363 return error;
3364
3365
3366
3367
3368
3369 reiserfs_write_lock(inode->i_sb);
3370 error = journal_begin(&th, inode->i_sb, jbegin_count);
3371 reiserfs_write_unlock(inode->i_sb);
3372 if (error)
3373 goto out;
3374 error = dquot_transfer(inode, attr);
3375 reiserfs_write_lock(inode->i_sb);
3376 if (error) {
3377 journal_end(&th);
3378 reiserfs_write_unlock(inode->i_sb);
3379 goto out;
3380 }
3381
3382
3383
3384
3385
3386 if (attr->ia_valid & ATTR_UID)
3387 inode->i_uid = attr->ia_uid;
3388 if (attr->ia_valid & ATTR_GID)
3389 inode->i_gid = attr->ia_gid;
3390 mark_inode_dirty(inode);
3391 error = journal_end(&th);
3392 reiserfs_write_unlock(inode->i_sb);
3393 if (error)
3394 goto out;
3395 }
3396
3397 if ((attr->ia_valid & ATTR_SIZE) &&
3398 attr->ia_size != i_size_read(inode)) {
3399 error = inode_newsize_ok(inode, attr->ia_size);
3400 if (!error) {
3401
3402
3403
3404
3405 mutex_lock(&REISERFS_I(inode)->tailpack);
3406 truncate_setsize(inode, attr->ia_size);
3407 reiserfs_truncate_file(inode, 1);
3408 mutex_unlock(&REISERFS_I(inode)->tailpack);
3409 }
3410 }
3411
3412 if (!error) {
3413 setattr_copy(inode, attr);
3414 mark_inode_dirty(inode);
3415 }
3416
3417 if (!error && reiserfs_posixacl(inode->i_sb)) {
3418 if (attr->ia_valid & ATTR_MODE)
3419 error = reiserfs_acl_chmod(inode);
3420 }
3421
3422out:
3423 return error;
3424}
3425
3426const struct address_space_operations reiserfs_address_space_operations = {
3427 .writepage = reiserfs_writepage,
3428 .readpage = reiserfs_readpage,
3429 .readpages = reiserfs_readpages,
3430 .releasepage = reiserfs_releasepage,
3431 .invalidatepage = reiserfs_invalidatepage,
3432 .write_begin = reiserfs_write_begin,
3433 .write_end = reiserfs_write_end,
3434 .bmap = reiserfs_aop_bmap,
3435 .direct_IO = reiserfs_direct_IO,
3436 .set_page_dirty = reiserfs_set_page_dirty,
3437};
3438