1
2
3
4
5
6
7
8
9
10#include <linux/kernel.h>
11#include <linux/backing-dev.h>
12#include <linux/dax.h>
13#include <linux/gfp.h>
14#include <linux/mm.h>
15#include <linux/swap.h>
16#include <linux/export.h>
17#include <linux/pagemap.h>
18#include <linux/highmem.h>
19#include <linux/pagevec.h>
20#include <linux/task_io_accounting_ops.h>
21#include <linux/buffer_head.h>
22
23#include <linux/shmem_fs.h>
24#include <linux/cleancache.h>
25#include <linux/rmap.h>
26#include "internal.h"
27
28
29
30
31
32
33static inline void __clear_shadow_entry(struct address_space *mapping,
34 pgoff_t index, void *entry)
35{
36 XA_STATE(xas, &mapping->i_pages, index);
37
38 xas_set_update(&xas, workingset_update_node);
39 if (xas_load(&xas) != entry)
40 return;
41 xas_store(&xas, NULL);
42 mapping->nrexceptional--;
43}
44
45static void clear_shadow_entry(struct address_space *mapping, pgoff_t index,
46 void *entry)
47{
48 xa_lock_irq(&mapping->i_pages);
49 __clear_shadow_entry(mapping, index, entry);
50 xa_unlock_irq(&mapping->i_pages);
51}
52
53
54
55
56
57
58static void truncate_exceptional_pvec_entries(struct address_space *mapping,
59 struct pagevec *pvec, pgoff_t *indices,
60 pgoff_t end)
61{
62 int i, j;
63 bool dax, lock;
64
65
66 if (shmem_mapping(mapping))
67 return;
68
69 for (j = 0; j < pagevec_count(pvec); j++)
70 if (xa_is_value(pvec->pages[j]))
71 break;
72
73 if (j == pagevec_count(pvec))
74 return;
75
76 dax = dax_mapping(mapping);
77 lock = !dax && indices[j] < end;
78 if (lock)
79 xa_lock_irq(&mapping->i_pages);
80
81 for (i = j; i < pagevec_count(pvec); i++) {
82 struct page *page = pvec->pages[i];
83 pgoff_t index = indices[i];
84
85 if (!xa_is_value(page)) {
86 pvec->pages[j++] = page;
87 continue;
88 }
89
90 if (index >= end)
91 continue;
92
93 if (unlikely(dax)) {
94 dax_delete_mapping_entry(mapping, index);
95 continue;
96 }
97
98 __clear_shadow_entry(mapping, index, page);
99 }
100
101 if (lock)
102 xa_unlock_irq(&mapping->i_pages);
103 pvec->nr = j;
104}
105
106
107
108
109
110static int invalidate_exceptional_entry(struct address_space *mapping,
111 pgoff_t index, void *entry)
112{
113
114 if (shmem_mapping(mapping) || dax_mapping(mapping))
115 return 1;
116 clear_shadow_entry(mapping, index, entry);
117 return 1;
118}
119
120
121
122
123
124static int invalidate_exceptional_entry2(struct address_space *mapping,
125 pgoff_t index, void *entry)
126{
127
128 if (shmem_mapping(mapping))
129 return 1;
130 if (dax_mapping(mapping))
131 return dax_invalidate_mapping_entry_sync(mapping, index);
132 clear_shadow_entry(mapping, index, entry);
133 return 1;
134}
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151void do_invalidatepage(struct page *page, unsigned int offset,
152 unsigned int length)
153{
154 void (*invalidatepage)(struct page *, unsigned int, unsigned int);
155
156 invalidatepage = page->mapping->a_ops->invalidatepage;
157#ifdef CONFIG_BLOCK
158 if (!invalidatepage)
159 invalidatepage = block_invalidatepage;
160#endif
161 if (invalidatepage)
162 (*invalidatepage)(page, offset, length);
163}
164
165
166
167
168
169
170
171
172
173
174
175static void
176truncate_cleanup_page(struct address_space *mapping, struct page *page)
177{
178 if (page_mapped(page)) {
179 pgoff_t nr = PageTransHuge(page) ? HPAGE_PMD_NR : 1;
180 unmap_mapping_pages(mapping, page->index, nr, false);
181 }
182
183 if (page_has_private(page))
184 do_invalidatepage(page, 0, PAGE_SIZE);
185
186
187
188
189
190
191 cancel_dirty_page(page);
192 ClearPageMappedToDisk(page);
193}
194
195
196
197
198
199
200
201
202
203static int
204invalidate_complete_page(struct address_space *mapping, struct page *page)
205{
206 int ret;
207
208 if (page->mapping != mapping)
209 return 0;
210
211 if (page_has_private(page) && !try_to_release_page(page, 0))
212 return 0;
213
214 ret = remove_mapping(mapping, page);
215
216 return ret;
217}
218
219int truncate_inode_page(struct address_space *mapping, struct page *page)
220{
221 VM_BUG_ON_PAGE(PageTail(page), page);
222
223 if (page->mapping != mapping)
224 return -EIO;
225
226 truncate_cleanup_page(mapping, page);
227 delete_from_page_cache(page);
228 return 0;
229}
230
231
232
233
234int generic_error_remove_page(struct address_space *mapping, struct page *page)
235{
236 if (!mapping)
237 return -EINVAL;
238
239
240
241
242 if (!S_ISREG(mapping->host->i_mode))
243 return -EIO;
244 return truncate_inode_page(mapping, page);
245}
246EXPORT_SYMBOL(generic_error_remove_page);
247
248
249
250
251
252
253
254int invalidate_inode_page(struct page *page)
255{
256 struct address_space *mapping = page_mapping(page);
257 if (!mapping)
258 return 0;
259 if (PageDirty(page) || PageWriteback(page))
260 return 0;
261 if (page_mapped(page))
262 return 0;
263 return invalidate_complete_page(mapping, page);
264}
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290void truncate_inode_pages_range(struct address_space *mapping,
291 loff_t lstart, loff_t lend)
292{
293 pgoff_t start;
294 pgoff_t end;
295 unsigned int partial_start;
296 unsigned int partial_end;
297 struct pagevec pvec;
298 pgoff_t indices[PAGEVEC_SIZE];
299 pgoff_t index;
300 int i;
301
302 if (mapping->nrpages == 0 && mapping->nrexceptional == 0)
303 goto out;
304
305
306 partial_start = lstart & (PAGE_SIZE - 1);
307 partial_end = (lend + 1) & (PAGE_SIZE - 1);
308
309
310
311
312
313
314
315 start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT;
316 if (lend == -1)
317
318
319
320
321
322 end = -1;
323 else
324 end = (lend + 1) >> PAGE_SHIFT;
325
326 pagevec_init(&pvec);
327 index = start;
328 while (index < end && pagevec_lookup_entries(&pvec, mapping, index,
329 min(end - index, (pgoff_t)PAGEVEC_SIZE),
330 indices)) {
331
332
333
334
335
336 struct pagevec locked_pvec;
337
338 pagevec_init(&locked_pvec);
339 for (i = 0; i < pagevec_count(&pvec); i++) {
340 struct page *page = pvec.pages[i];
341
342
343 index = indices[i];
344 if (index >= end)
345 break;
346
347 if (xa_is_value(page))
348 continue;
349
350 if (!trylock_page(page))
351 continue;
352 WARN_ON(page_to_index(page) != index);
353 if (PageWriteback(page)) {
354 unlock_page(page);
355 continue;
356 }
357 if (page->mapping != mapping) {
358 unlock_page(page);
359 continue;
360 }
361 pagevec_add(&locked_pvec, page);
362 }
363 for (i = 0; i < pagevec_count(&locked_pvec); i++)
364 truncate_cleanup_page(mapping, locked_pvec.pages[i]);
365 delete_from_page_cache_batch(mapping, &locked_pvec);
366 for (i = 0; i < pagevec_count(&locked_pvec); i++)
367 unlock_page(locked_pvec.pages[i]);
368 truncate_exceptional_pvec_entries(mapping, &pvec, indices, end);
369 pagevec_release(&pvec);
370 cond_resched();
371 index++;
372 }
373 if (partial_start) {
374 struct page *page = find_lock_page(mapping, start - 1);
375 if (page) {
376 unsigned int top = PAGE_SIZE;
377 if (start > end) {
378
379 top = partial_end;
380 partial_end = 0;
381 }
382 wait_on_page_writeback(page);
383 zero_user_segment(page, partial_start, top);
384 cleancache_invalidate_page(mapping, page);
385 if (page_has_private(page))
386 do_invalidatepage(page, partial_start,
387 top - partial_start);
388 unlock_page(page);
389 put_page(page);
390 }
391 }
392 if (partial_end) {
393 struct page *page = find_lock_page(mapping, end);
394 if (page) {
395 wait_on_page_writeback(page);
396 zero_user_segment(page, 0, partial_end);
397 cleancache_invalidate_page(mapping, page);
398 if (page_has_private(page))
399 do_invalidatepage(page, 0,
400 partial_end);
401 unlock_page(page);
402 put_page(page);
403 }
404 }
405
406
407
408
409 if (start >= end)
410 goto out;
411
412 index = start;
413 for ( ; ; ) {
414 cond_resched();
415 if (!pagevec_lookup_entries(&pvec, mapping, index,
416 min(end - index, (pgoff_t)PAGEVEC_SIZE), indices)) {
417
418 if (index == start)
419 break;
420
421 index = start;
422 continue;
423 }
424 if (index == start && indices[0] >= end) {
425
426 pagevec_remove_exceptionals(&pvec);
427 pagevec_release(&pvec);
428 break;
429 }
430
431 for (i = 0; i < pagevec_count(&pvec); i++) {
432 struct page *page = pvec.pages[i];
433
434
435 index = indices[i];
436 if (index >= end) {
437
438 index = start - 1;
439 break;
440 }
441
442 if (xa_is_value(page))
443 continue;
444
445 lock_page(page);
446 WARN_ON(page_to_index(page) != index);
447 wait_on_page_writeback(page);
448 truncate_inode_page(mapping, page);
449 unlock_page(page);
450 }
451 truncate_exceptional_pvec_entries(mapping, &pvec, indices, end);
452 pagevec_release(&pvec);
453 index++;
454 }
455
456out:
457 cleancache_invalidate_inode(mapping);
458}
459EXPORT_SYMBOL(truncate_inode_pages_range);
460
461
462
463
464
465
466
467
468
469
470
471
472
473void truncate_inode_pages(struct address_space *mapping, loff_t lstart)
474{
475 truncate_inode_pages_range(mapping, lstart, (loff_t)-1);
476}
477EXPORT_SYMBOL(truncate_inode_pages);
478
479
480
481
482
483
484
485
486
487
488void truncate_inode_pages_final(struct address_space *mapping)
489{
490 unsigned long nrexceptional;
491 unsigned long nrpages;
492
493
494
495
496
497
498
499
500 mapping_set_exiting(mapping);
501
502
503
504
505
506
507 nrpages = mapping->nrpages;
508 smp_rmb();
509 nrexceptional = mapping->nrexceptional;
510
511 if (nrpages || nrexceptional) {
512
513
514
515
516
517
518 xa_lock_irq(&mapping->i_pages);
519 xa_unlock_irq(&mapping->i_pages);
520 }
521
522
523
524
525
526 truncate_inode_pages(mapping, 0);
527}
528EXPORT_SYMBOL(truncate_inode_pages_final);
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543unsigned long invalidate_mapping_pages(struct address_space *mapping,
544 pgoff_t start, pgoff_t end)
545{
546 pgoff_t indices[PAGEVEC_SIZE];
547 struct pagevec pvec;
548 pgoff_t index = start;
549 unsigned long ret;
550 unsigned long count = 0;
551 int i;
552
553 pagevec_init(&pvec);
554 while (index <= end && pagevec_lookup_entries(&pvec, mapping, index,
555 min(end - index, (pgoff_t)PAGEVEC_SIZE - 1) + 1,
556 indices)) {
557 for (i = 0; i < pagevec_count(&pvec); i++) {
558 struct page *page = pvec.pages[i];
559
560
561 index = indices[i];
562 if (index > end)
563 break;
564
565 if (xa_is_value(page)) {
566 invalidate_exceptional_entry(mapping, index,
567 page);
568 continue;
569 }
570
571 if (!trylock_page(page))
572 continue;
573
574 WARN_ON(page_to_index(page) != index);
575
576
577 if (PageTransTail(page)) {
578 unlock_page(page);
579 continue;
580 } else if (PageTransHuge(page)) {
581 index += HPAGE_PMD_NR - 1;
582 i += HPAGE_PMD_NR - 1;
583
584
585
586
587
588 if (index > end) {
589 unlock_page(page);
590 continue;
591 }
592 }
593
594 ret = invalidate_inode_page(page);
595 unlock_page(page);
596
597
598
599
600 if (!ret)
601 deactivate_file_page(page);
602 count += ret;
603 }
604 pagevec_remove_exceptionals(&pvec);
605 pagevec_release(&pvec);
606 cond_resched();
607 index++;
608 }
609 return count;
610}
611EXPORT_SYMBOL(invalidate_mapping_pages);
612
613
614
615
616
617
618
619
620static int
621invalidate_complete_page2(struct address_space *mapping, struct page *page)
622{
623 unsigned long flags;
624
625 if (page->mapping != mapping)
626 return 0;
627
628 if (page_has_private(page) && !try_to_release_page(page, GFP_KERNEL))
629 return 0;
630
631 xa_lock_irqsave(&mapping->i_pages, flags);
632 if (PageDirty(page))
633 goto failed;
634
635 BUG_ON(page_has_private(page));
636 __delete_from_page_cache(page, NULL);
637 xa_unlock_irqrestore(&mapping->i_pages, flags);
638
639 if (mapping->a_ops->freepage)
640 mapping->a_ops->freepage(page);
641
642 put_page(page);
643 return 1;
644failed:
645 xa_unlock_irqrestore(&mapping->i_pages, flags);
646 return 0;
647}
648
649static int do_launder_page(struct address_space *mapping, struct page *page)
650{
651 if (!PageDirty(page))
652 return 0;
653 if (page->mapping != mapping || mapping->a_ops->launder_page == NULL)
654 return 0;
655 return mapping->a_ops->launder_page(page);
656}
657
658
659
660
661
662
663
664
665
666
667
668
669int invalidate_inode_pages2_range(struct address_space *mapping,
670 pgoff_t start, pgoff_t end)
671{
672 pgoff_t indices[PAGEVEC_SIZE];
673 struct pagevec pvec;
674 pgoff_t index;
675 int i;
676 int ret = 0;
677 int ret2 = 0;
678 int did_range_unmap = 0;
679
680 if (mapping->nrpages == 0 && mapping->nrexceptional == 0)
681 goto out;
682
683 pagevec_init(&pvec);
684 index = start;
685 while (index <= end && pagevec_lookup_entries(&pvec, mapping, index,
686 min(end - index, (pgoff_t)PAGEVEC_SIZE - 1) + 1,
687 indices)) {
688 for (i = 0; i < pagevec_count(&pvec); i++) {
689 struct page *page = pvec.pages[i];
690
691
692 index = indices[i];
693 if (index > end)
694 break;
695
696 if (xa_is_value(page)) {
697 if (!invalidate_exceptional_entry2(mapping,
698 index, page))
699 ret = -EBUSY;
700 continue;
701 }
702
703 lock_page(page);
704 WARN_ON(page_to_index(page) != index);
705 if (page->mapping != mapping) {
706 unlock_page(page);
707 continue;
708 }
709 wait_on_page_writeback(page);
710 if (page_mapped(page)) {
711 if (!did_range_unmap) {
712
713
714
715 unmap_mapping_pages(mapping, index,
716 (1 + end - index), false);
717 did_range_unmap = 1;
718 } else {
719
720
721
722 unmap_mapping_pages(mapping, index,
723 1, false);
724 }
725 }
726 BUG_ON(page_mapped(page));
727 ret2 = do_launder_page(mapping, page);
728 if (ret2 == 0) {
729 if (!invalidate_complete_page2(mapping, page))
730 ret2 = -EBUSY;
731 }
732 if (ret2 < 0)
733 ret = ret2;
734 unlock_page(page);
735 }
736 pagevec_remove_exceptionals(&pvec);
737 pagevec_release(&pvec);
738 cond_resched();
739 index++;
740 }
741
742
743
744
745
746
747
748 if (dax_mapping(mapping)) {
749 unmap_mapping_pages(mapping, start, end - start + 1, false);
750 }
751out:
752 cleancache_invalidate_inode(mapping);
753 return ret;
754}
755EXPORT_SYMBOL_GPL(invalidate_inode_pages2_range);
756
757
758
759
760
761
762
763
764
765
766int invalidate_inode_pages2(struct address_space *mapping)
767{
768 return invalidate_inode_pages2_range(mapping, 0, -1);
769}
770EXPORT_SYMBOL_GPL(invalidate_inode_pages2);
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787void truncate_pagecache(struct inode *inode, loff_t newsize)
788{
789 struct address_space *mapping = inode->i_mapping;
790 loff_t holebegin = round_up(newsize, PAGE_SIZE);
791
792
793
794
795
796
797
798
799
800
801 unmap_mapping_range(mapping, holebegin, 0, 1);
802 truncate_inode_pages(mapping, newsize);
803 unmap_mapping_range(mapping, holebegin, 0, 1);
804}
805EXPORT_SYMBOL(truncate_pagecache);
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820void truncate_setsize(struct inode *inode, loff_t newsize)
821{
822 loff_t oldsize = inode->i_size;
823
824 i_size_write(inode, newsize);
825 if (newsize > oldsize)
826 pagecache_isize_extended(inode, oldsize, newsize);
827 truncate_pagecache(inode, newsize);
828}
829EXPORT_SYMBOL(truncate_setsize);
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850void pagecache_isize_extended(struct inode *inode, loff_t from, loff_t to)
851{
852 int bsize = i_blocksize(inode);
853 loff_t rounded_from;
854 struct page *page;
855 pgoff_t index;
856
857 WARN_ON(to > inode->i_size);
858
859 if (from >= to || bsize == PAGE_SIZE)
860 return;
861
862 rounded_from = round_up(from, bsize);
863 if (to <= rounded_from || !(rounded_from & (PAGE_SIZE - 1)))
864 return;
865
866 index = from >> PAGE_SHIFT;
867 page = find_lock_page(inode->i_mapping, index);
868
869 if (!page)
870 return;
871
872
873
874
875 if (page_mkclean(page))
876 set_page_dirty(page);
877 unlock_page(page);
878 put_page(page);
879}
880EXPORT_SYMBOL(pagecache_isize_extended);
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895void truncate_pagecache_range(struct inode *inode, loff_t lstart, loff_t lend)
896{
897 struct address_space *mapping = inode->i_mapping;
898 loff_t unmap_start = round_up(lstart, PAGE_SIZE);
899 loff_t unmap_end = round_down(1 + lend, PAGE_SIZE) - 1;
900
901
902
903
904
905
906
907
908
909
910
911
912
913 if ((u64)unmap_end > (u64)unmap_start)
914 unmap_mapping_range(mapping, unmap_start,
915 1 + unmap_end - unmap_start, 0);
916 truncate_inode_pages_range(mapping, lstart, lend);
917}
918EXPORT_SYMBOL(truncate_pagecache_range);
919