1 | /* |
2 | * Compatibility functions which bloat the callers too much to make inline. |
3 | * All of the callers of these functions should be converted to use folios |
4 | * eventually. |
5 | */ |
6 | |
7 | #include <linux/migrate.h> |
8 | #include <linux/pagemap.h> |
9 | #include <linux/rmap.h> |
10 | #include <linux/swap.h> |
11 | #include "internal.h" |
12 | |
13 | struct address_space *page_mapping(struct page *page) |
14 | { |
15 | return folio_mapping(page_folio(page)); |
16 | } |
17 | EXPORT_SYMBOL(page_mapping); |
18 | |
19 | void unlock_page(struct page *page) |
20 | { |
21 | return folio_unlock(page_folio(page)); |
22 | } |
23 | EXPORT_SYMBOL(unlock_page); |
24 | |
25 | void end_page_writeback(struct page *page) |
26 | { |
27 | return folio_end_writeback(page_folio(page)); |
28 | } |
29 | EXPORT_SYMBOL(end_page_writeback); |
30 | |
31 | void wait_on_page_writeback(struct page *page) |
32 | { |
33 | return folio_wait_writeback(page_folio(page)); |
34 | } |
35 | EXPORT_SYMBOL_GPL(wait_on_page_writeback); |
36 | |
37 | void wait_for_stable_page(struct page *page) |
38 | { |
39 | return folio_wait_stable(page_folio(page)); |
40 | } |
41 | EXPORT_SYMBOL_GPL(wait_for_stable_page); |
42 | |
43 | void mark_page_accessed(struct page *page) |
44 | { |
45 | folio_mark_accessed(page_folio(page)); |
46 | } |
47 | EXPORT_SYMBOL(mark_page_accessed); |
48 | |
49 | bool set_page_writeback(struct page *page) |
50 | { |
51 | return folio_start_writeback(page_folio(page)); |
52 | } |
53 | EXPORT_SYMBOL(set_page_writeback); |
54 | |
55 | bool set_page_dirty(struct page *page) |
56 | { |
57 | return folio_mark_dirty(page_folio(page)); |
58 | } |
59 | EXPORT_SYMBOL(set_page_dirty); |
60 | |
61 | int __set_page_dirty_nobuffers(struct page *page) |
62 | { |
63 | return filemap_dirty_folio(mapping: page_mapping(page), page_folio(page)); |
64 | } |
65 | EXPORT_SYMBOL(__set_page_dirty_nobuffers); |
66 | |
67 | bool clear_page_dirty_for_io(struct page *page) |
68 | { |
69 | return folio_clear_dirty_for_io(page_folio(page)); |
70 | } |
71 | EXPORT_SYMBOL(clear_page_dirty_for_io); |
72 | |
73 | bool redirty_page_for_writepage(struct writeback_control *wbc, |
74 | struct page *page) |
75 | { |
76 | return folio_redirty_for_writepage(wbc, page_folio(page)); |
77 | } |
78 | EXPORT_SYMBOL(redirty_page_for_writepage); |
79 | |
80 | void lru_cache_add_inactive_or_unevictable(struct page *page, |
81 | struct vm_area_struct *vma) |
82 | { |
83 | folio_add_lru_vma(page_folio(page), vma); |
84 | } |
85 | |
86 | int add_to_page_cache_lru(struct page *page, struct address_space *mapping, |
87 | pgoff_t index, gfp_t gfp) |
88 | { |
89 | return filemap_add_folio(mapping, page_folio(page), index, gfp); |
90 | } |
91 | EXPORT_SYMBOL(add_to_page_cache_lru); |
92 | |
93 | noinline |
94 | struct page *pagecache_get_page(struct address_space *mapping, pgoff_t index, |
95 | fgf_t fgp_flags, gfp_t gfp) |
96 | { |
97 | struct folio *folio; |
98 | |
99 | folio = __filemap_get_folio(mapping, index, fgp_flags, gfp); |
100 | if (IS_ERR(ptr: folio)) |
101 | return NULL; |
102 | return folio_file_page(folio, index); |
103 | } |
104 | EXPORT_SYMBOL(pagecache_get_page); |
105 | |
106 | struct page *grab_cache_page_write_begin(struct address_space *mapping, |
107 | pgoff_t index) |
108 | { |
109 | return pagecache_get_page(mapping, index, FGP_WRITEBEGIN, |
110 | mapping_gfp_mask(mapping)); |
111 | } |
112 | EXPORT_SYMBOL(grab_cache_page_write_begin); |
113 | |
114 | bool isolate_lru_page(struct page *page) |
115 | { |
116 | if (WARN_RATELIMIT(PageTail(page), "trying to isolate tail page" )) |
117 | return false; |
118 | return folio_isolate_lru(folio: (struct folio *)page); |
119 | } |
120 | |
121 | void putback_lru_page(struct page *page) |
122 | { |
123 | folio_putback_lru(page_folio(page)); |
124 | } |
125 | |
126 | #ifdef CONFIG_MMU |
127 | void page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, |
128 | unsigned long address) |
129 | { |
130 | VM_BUG_ON_PAGE(PageTail(page), page); |
131 | |
132 | return folio_add_new_anon_rmap((struct folio *)page, vma, address); |
133 | } |
134 | #endif |
135 | |