1
2#include <test_progs.h>
3#include <sys/mman.h>
4#include "test_mmap.skel.h"
5
6struct map_data {
7 __u64 val[512 * 4];
8};
9
10static size_t roundup_page(size_t sz)
11{
12 long page_size = sysconf(_SC_PAGE_SIZE);
13 return (sz + page_size - 1) / page_size * page_size;
14}
15
16void test_mmap(void)
17{
18 const size_t bss_sz = roundup_page(sizeof(struct test_mmap__bss));
19 const size_t map_sz = roundup_page(sizeof(struct map_data));
20 const int zero = 0, one = 1, two = 2, far = 1500;
21 const long page_size = sysconf(_SC_PAGE_SIZE);
22 int err, duration = 0, i, data_map_fd, data_map_id, tmp_fd, rdmap_fd;
23 struct bpf_map *data_map, *bss_map;
24 void *bss_mmaped = NULL, *map_mmaped = NULL, *tmp0, *tmp1, *tmp2;
25 struct test_mmap__bss *bss_data;
26 struct bpf_map_info map_info;
27 __u32 map_info_sz = sizeof(map_info);
28 struct map_data *map_data;
29 struct test_mmap *skel;
30 __u64 val = 0;
31
32 skel = test_mmap__open();
33 if (CHECK(!skel, "skel_open", "skeleton open failed\n"))
34 return;
35
36 err = bpf_map__set_max_entries(skel->maps.rdonly_map, page_size);
37 if (CHECK(err != 0, "bpf_map__set_max_entries", "bpf_map__set_max_entries failed\n"))
38 goto cleanup;
39
40
41 err = bpf_map__set_max_entries(skel->maps.data_map,
42 4 * (page_size / sizeof(u64)));
43 if (CHECK(err != 0, "bpf_map__set_max_entries", "bpf_map__set_max_entries failed\n"))
44 goto cleanup;
45
46 err = test_mmap__load(skel);
47 if (CHECK(err != 0, "skel_load", "skeleton load failed\n"))
48 goto cleanup;
49
50 bss_map = skel->maps.bss;
51 data_map = skel->maps.data_map;
52 data_map_fd = bpf_map__fd(data_map);
53
54 rdmap_fd = bpf_map__fd(skel->maps.rdonly_map);
55 tmp1 = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, rdmap_fd, 0);
56 if (CHECK(tmp1 != MAP_FAILED, "rdonly_write_mmap", "unexpected success\n")) {
57 munmap(tmp1, page_size);
58 goto cleanup;
59 }
60
61 tmp1 = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rdmap_fd, 0);
62 if (CHECK(tmp1 == MAP_FAILED, "rdonly_read_mmap", "failed: %d\n", errno))
63 goto cleanup;
64
65
66 memset(&map_info, 0, map_info_sz);
67 err = bpf_obj_get_info_by_fd(data_map_fd, &map_info, &map_info_sz);
68 if (CHECK(err, "map_get_info", "failed %d\n", errno))
69 goto cleanup;
70 data_map_id = map_info.id;
71
72
73 bss_mmaped = mmap(NULL, bss_sz, PROT_READ | PROT_WRITE, MAP_SHARED,
74 bpf_map__fd(bss_map), 0);
75 if (CHECK(bss_mmaped == MAP_FAILED, "bss_mmap",
76 ".bss mmap failed: %d\n", errno)) {
77 bss_mmaped = NULL;
78 goto cleanup;
79 }
80
81 map_mmaped = mmap(NULL, map_sz, PROT_READ | PROT_WRITE, MAP_SHARED,
82 data_map_fd, 0);
83 if (CHECK(map_mmaped == MAP_FAILED, "data_mmap",
84 "data_map mmap failed: %d\n", errno)) {
85 map_mmaped = NULL;
86 goto cleanup;
87 }
88
89 bss_data = bss_mmaped;
90 map_data = map_mmaped;
91
92 CHECK_FAIL(bss_data->in_val);
93 CHECK_FAIL(bss_data->out_val);
94 CHECK_FAIL(skel->bss->in_val);
95 CHECK_FAIL(skel->bss->out_val);
96 CHECK_FAIL(map_data->val[0]);
97 CHECK_FAIL(map_data->val[1]);
98 CHECK_FAIL(map_data->val[2]);
99 CHECK_FAIL(map_data->val[far]);
100
101 err = test_mmap__attach(skel);
102 if (CHECK(err, "attach_raw_tp", "err %d\n", err))
103 goto cleanup;
104
105 bss_data->in_val = 123;
106 val = 111;
107 CHECK_FAIL(bpf_map_update_elem(data_map_fd, &zero, &val, 0));
108
109 usleep(1);
110
111 CHECK_FAIL(bss_data->in_val != 123);
112 CHECK_FAIL(bss_data->out_val != 123);
113 CHECK_FAIL(skel->bss->in_val != 123);
114 CHECK_FAIL(skel->bss->out_val != 123);
115 CHECK_FAIL(map_data->val[0] != 111);
116 CHECK_FAIL(map_data->val[1] != 222);
117 CHECK_FAIL(map_data->val[2] != 123);
118 CHECK_FAIL(map_data->val[far] != 3 * 123);
119
120 CHECK_FAIL(bpf_map_lookup_elem(data_map_fd, &zero, &val));
121 CHECK_FAIL(val != 111);
122 CHECK_FAIL(bpf_map_lookup_elem(data_map_fd, &one, &val));
123 CHECK_FAIL(val != 222);
124 CHECK_FAIL(bpf_map_lookup_elem(data_map_fd, &two, &val));
125 CHECK_FAIL(val != 123);
126 CHECK_FAIL(bpf_map_lookup_elem(data_map_fd, &far, &val));
127 CHECK_FAIL(val != 3 * 123);
128
129
130 err = bpf_map_freeze(data_map_fd);
131 if (CHECK(!err || errno != EBUSY, "no_freeze",
132 "data_map freeze succeeded: err=%d, errno=%d\n", err, errno))
133 goto cleanup;
134
135 err = mprotect(map_mmaped, map_sz, PROT_READ);
136 if (CHECK(err, "mprotect_ro", "mprotect to r/o failed %d\n", errno))
137 goto cleanup;
138
139
140 err = munmap(map_mmaped, map_sz);
141 map_mmaped = NULL;
142 if (CHECK(err, "data_map_munmap", "data_map munmap failed: %d\n", errno))
143 goto cleanup;
144
145
146 map_mmaped = mmap(NULL, map_sz, PROT_READ, MAP_SHARED, data_map_fd, 0);
147 if (CHECK(map_mmaped == MAP_FAILED, "data_mmap",
148 "data_map R/O mmap failed: %d\n", errno)) {
149 map_mmaped = NULL;
150 goto cleanup;
151 }
152 err = mprotect(map_mmaped, map_sz, PROT_WRITE);
153 if (CHECK(!err, "mprotect_wr", "mprotect() succeeded unexpectedly!\n"))
154 goto cleanup;
155 err = mprotect(map_mmaped, map_sz, PROT_EXEC);
156 if (CHECK(!err, "mprotect_ex", "mprotect() succeeded unexpectedly!\n"))
157 goto cleanup;
158 map_data = map_mmaped;
159
160
161 for (i = 0; i < 10; i++) {
162 int flags = i % 2 ? PROT_READ : PROT_WRITE;
163 void *p;
164
165 p = mmap(NULL, map_sz, flags, MAP_SHARED, data_map_fd, 0);
166 if (CHECK_FAIL(p == MAP_FAILED))
167 goto cleanup;
168 err = munmap(p, map_sz);
169 if (CHECK_FAIL(err))
170 goto cleanup;
171 }
172
173
174 err = bpf_map_freeze(data_map_fd);
175 if (CHECK(err, "freeze", "data_map freeze failed: err=%d, errno=%d\n",
176 err, errno))
177 goto cleanup;
178
179
180 tmp1 = mmap(NULL, map_sz, PROT_READ | PROT_WRITE, MAP_SHARED,
181 data_map_fd, 0);
182 if (CHECK(tmp1 != MAP_FAILED, "data_mmap", "mmap succeeded\n")) {
183 munmap(tmp1, map_sz);
184 goto cleanup;
185 }
186
187 bss_data->in_val = 321;
188 usleep(1);
189 CHECK_FAIL(bss_data->in_val != 321);
190 CHECK_FAIL(bss_data->out_val != 321);
191 CHECK_FAIL(skel->bss->in_val != 321);
192 CHECK_FAIL(skel->bss->out_val != 321);
193 CHECK_FAIL(map_data->val[0] != 111);
194 CHECK_FAIL(map_data->val[1] != 222);
195 CHECK_FAIL(map_data->val[2] != 321);
196 CHECK_FAIL(map_data->val[far] != 3 * 321);
197
198
199
200 tmp0 = mmap(NULL, 4 * page_size, PROT_READ, MAP_SHARED | MAP_ANONYMOUS,
201 -1, 0);
202 if (CHECK(tmp0 == MAP_FAILED, "adv_mmap0", "errno %d\n", errno))
203 goto cleanup;
204
205
206 tmp1 = mmap(tmp0, 3 * page_size, PROT_READ, MAP_SHARED | MAP_FIXED,
207 data_map_fd, 0);
208 if (CHECK(tmp0 != tmp1, "adv_mmap1", "tmp0: %p, tmp1: %p\n", tmp0, tmp1)) {
209 munmap(tmp0, 4 * page_size);
210 goto cleanup;
211 }
212
213
214 err = munmap(tmp1 + page_size, page_size);
215 if (CHECK(err, "adv_mmap2", "errno %d\n", errno)) {
216 munmap(tmp1, 4 * page_size);
217 goto cleanup;
218 }
219
220
221 tmp2 = mmap(tmp1 + page_size, page_size, PROT_READ,
222 MAP_SHARED | MAP_FIXED, data_map_fd, 0);
223 if (CHECK(tmp2 == MAP_FAILED, "adv_mmap3", "errno %d\n", errno)) {
224 munmap(tmp1, page_size);
225 munmap(tmp1 + 2*page_size, 2 * page_size);
226 goto cleanup;
227 }
228 CHECK(tmp1 + page_size != tmp2, "adv_mmap4",
229 "tmp1: %p, tmp2: %p\n", tmp1, tmp2);
230
231
232 tmp2 = mmap(tmp1, 4 * page_size, PROT_READ, MAP_SHARED | MAP_FIXED,
233 data_map_fd, 0);
234 if (CHECK(tmp2 == MAP_FAILED, "adv_mmap5", "errno %d\n", errno)) {
235 munmap(tmp1, 4 * page_size);
236 goto cleanup;
237 }
238 CHECK(tmp1 != tmp2, "adv_mmap6", "tmp1: %p, tmp2: %p\n", tmp1, tmp2);
239
240 map_data = tmp2;
241 CHECK_FAIL(bss_data->in_val != 321);
242 CHECK_FAIL(bss_data->out_val != 321);
243 CHECK_FAIL(skel->bss->in_val != 321);
244 CHECK_FAIL(skel->bss->out_val != 321);
245 CHECK_FAIL(map_data->val[0] != 111);
246 CHECK_FAIL(map_data->val[1] != 222);
247 CHECK_FAIL(map_data->val[2] != 321);
248 CHECK_FAIL(map_data->val[far] != 3 * 321);
249
250 munmap(tmp2, 4 * page_size);
251
252
253 tmp1 = mmap(NULL, 4 * page_size, PROT_READ, MAP_SHARED | MAP_FIXED,
254 data_map_fd, page_size );
255 if (CHECK(tmp1 != MAP_FAILED, "adv_mmap7", "unexpected success")) {
256 munmap(tmp1, 4 * page_size);
257 goto cleanup;
258 }
259
260 tmp1 = mmap(NULL, map_sz, PROT_READ, MAP_SHARED, data_map_fd, 0);
261 if (CHECK(tmp1 == MAP_FAILED, "last_mmap", "failed %d\n", errno))
262 goto cleanup;
263
264 test_mmap__destroy(skel);
265 skel = NULL;
266 CHECK_FAIL(munmap(bss_mmaped, bss_sz));
267 bss_mmaped = NULL;
268 CHECK_FAIL(munmap(map_mmaped, map_sz));
269 map_mmaped = NULL;
270
271
272 tmp_fd = bpf_map_get_fd_by_id(data_map_id);
273 if (CHECK(tmp_fd < 0, "get_map_by_id", "failed %d\n", errno)) {
274 munmap(tmp1, map_sz);
275 goto cleanup;
276 }
277 close(tmp_fd);
278
279
280 munmap(tmp1, map_sz);
281
282
283 for (i = 0; i < 10000; i++) {
284 __u32 id = data_map_id - 1;
285 if (bpf_map_get_next_id(id, &id) || id > data_map_id)
286 break;
287 usleep(1);
288 }
289
290
291 tmp_fd = bpf_map_get_fd_by_id(data_map_id);
292 if (CHECK(tmp_fd >= 0, "get_map_by_id_after",
293 "unexpectedly succeeded %d\n", tmp_fd)) {
294 close(tmp_fd);
295 goto cleanup;
296 }
297
298cleanup:
299 if (bss_mmaped)
300 CHECK_FAIL(munmap(bss_mmaped, bss_sz));
301 if (map_mmaped)
302 CHECK_FAIL(munmap(map_mmaped, map_sz));
303 test_mmap__destroy(skel);
304}
305