1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22#include "uvmm.h"
23#include "umem.h"
24#include "ummu.h"
25
26#include <core/client.h>
27#include <core/memory.h>
28
29#include <nvif/if000c.h>
30#include <nvif/unpack.h>
31
32static const struct nvkm_object_func nvkm_uvmm;
33struct nvkm_vmm *
34nvkm_uvmm_search(struct nvkm_client *client, u64 handle)
35{
36 struct nvkm_object *object;
37
38 object = nvkm_object_search(client, handle, &nvkm_uvmm);
39 if (IS_ERR(object))
40 return (void *)object;
41
42 return nvkm_uvmm(object)->vmm;
43}
44
45static int
46nvkm_uvmm_mthd_unmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
47{
48 struct nvkm_client *client = uvmm->object.client;
49 union {
50 struct nvif_vmm_unmap_v0 v0;
51 } *args = argv;
52 struct nvkm_vmm *vmm = uvmm->vmm;
53 struct nvkm_vma *vma;
54 int ret = -ENOSYS;
55 u64 addr;
56
57 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
58 addr = args->v0.addr;
59 } else
60 return ret;
61
62 mutex_lock(&vmm->mutex);
63 vma = nvkm_vmm_node_search(vmm, addr);
64 if (ret = -ENOENT, !vma || vma->addr != addr) {
65 VMM_DEBUG(vmm, "lookup %016llx: %016llx",
66 addr, vma ? vma->addr : ~0ULL);
67 goto done;
68 }
69
70 if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) {
71 VMM_DEBUG(vmm, "denied %016llx: %d %d %d", addr,
72 vma->user, !client->super, vma->busy);
73 goto done;
74 }
75
76 if (ret = -EINVAL, !vma->memory) {
77 VMM_DEBUG(vmm, "unmapped");
78 goto done;
79 }
80
81 nvkm_vmm_unmap_locked(vmm, vma);
82 ret = 0;
83done:
84 mutex_unlock(&vmm->mutex);
85 return ret;
86}
87
88static int
89nvkm_uvmm_mthd_map(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
90{
91 struct nvkm_client *client = uvmm->object.client;
92 union {
93 struct nvif_vmm_map_v0 v0;
94 } *args = argv;
95 u64 addr, size, handle, offset;
96 struct nvkm_vmm *vmm = uvmm->vmm;
97 struct nvkm_vma *vma;
98 struct nvkm_memory *memory;
99 int ret = -ENOSYS;
100
101 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) {
102 addr = args->v0.addr;
103 size = args->v0.size;
104 handle = args->v0.memory;
105 offset = args->v0.offset;
106 } else
107 return ret;
108
109 memory = nvkm_umem_search(client, handle);
110 if (IS_ERR(memory)) {
111 VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory));
112 return PTR_ERR(memory);
113 }
114
115 mutex_lock(&vmm->mutex);
116 if (ret = -ENOENT, !(vma = nvkm_vmm_node_search(vmm, addr))) {
117 VMM_DEBUG(vmm, "lookup %016llx", addr);
118 goto fail;
119 }
120
121 if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) {
122 VMM_DEBUG(vmm, "denied %016llx: %d %d %d", addr,
123 vma->user, !client->super, vma->busy);
124 goto fail;
125 }
126
127 if (ret = -EINVAL, vma->addr != addr || vma->size != size) {
128 if (addr + size > vma->addr + vma->size || vma->memory ||
129 (vma->refd == NVKM_VMA_PAGE_NONE && !vma->mapref)) {
130 VMM_DEBUG(vmm, "split %d %d %d "
131 "%016llx %016llx %016llx %016llx",
132 !!vma->memory, vma->refd, vma->mapref,
133 addr, size, vma->addr, (u64)vma->size);
134 goto fail;
135 }
136
137 if (vma->addr != addr) {
138 const u64 tail = vma->size + vma->addr - addr;
139 if (ret = -ENOMEM, !(vma = nvkm_vma_tail(vma, tail)))
140 goto fail;
141 vma->part = true;
142 nvkm_vmm_node_insert(vmm, vma);
143 }
144
145 if (vma->size != size) {
146 const u64 tail = vma->size - size;
147 struct nvkm_vma *tmp;
148 if (ret = -ENOMEM, !(tmp = nvkm_vma_tail(vma, tail))) {
149 nvkm_vmm_unmap_region(vmm, vma);
150 goto fail;
151 }
152 tmp->part = true;
153 nvkm_vmm_node_insert(vmm, tmp);
154 }
155 }
156 vma->busy = true;
157 mutex_unlock(&vmm->mutex);
158
159 ret = nvkm_memory_map(memory, offset, vmm, vma, argv, argc);
160 if (ret == 0) {
161
162 nvkm_memory_unref(&memory);
163 return 0;
164 }
165
166 mutex_lock(&vmm->mutex);
167 vma->busy = false;
168 nvkm_vmm_unmap_region(vmm, vma);
169fail:
170 mutex_unlock(&vmm->mutex);
171 nvkm_memory_unref(&memory);
172 return ret;
173}
174
175static int
176nvkm_uvmm_mthd_put(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
177{
178 struct nvkm_client *client = uvmm->object.client;
179 union {
180 struct nvif_vmm_put_v0 v0;
181 } *args = argv;
182 struct nvkm_vmm *vmm = uvmm->vmm;
183 struct nvkm_vma *vma;
184 int ret = -ENOSYS;
185 u64 addr;
186
187 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
188 addr = args->v0.addr;
189 } else
190 return ret;
191
192 mutex_lock(&vmm->mutex);
193 vma = nvkm_vmm_node_search(vmm, args->v0.addr);
194 if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) {
195 VMM_DEBUG(vmm, "lookup %016llx: %016llx %d", addr,
196 vma ? vma->addr : ~0ULL, vma ? vma->part : 0);
197 goto done;
198 }
199
200 if (ret = -ENOENT, (!vma->user && !client->super) || vma->busy) {
201 VMM_DEBUG(vmm, "denied %016llx: %d %d %d", addr,
202 vma->user, !client->super, vma->busy);
203 goto done;
204 }
205
206 nvkm_vmm_put_locked(vmm, vma);
207 ret = 0;
208done:
209 mutex_unlock(&vmm->mutex);
210 return ret;
211}
212
213static int
214nvkm_uvmm_mthd_get(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
215{
216 struct nvkm_client *client = uvmm->object.client;
217 union {
218 struct nvif_vmm_get_v0 v0;
219 } *args = argv;
220 struct nvkm_vmm *vmm = uvmm->vmm;
221 struct nvkm_vma *vma;
222 int ret = -ENOSYS;
223 bool getref, mapref, sparse;
224 u8 page, align;
225 u64 size;
226
227 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
228 getref = args->v0.type == NVIF_VMM_GET_V0_PTES;
229 mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR;
230 sparse = args->v0.sparse;
231 page = args->v0.page;
232 align = args->v0.align;
233 size = args->v0.size;
234 } else
235 return ret;
236
237 mutex_lock(&vmm->mutex);
238 ret = nvkm_vmm_get_locked(vmm, getref, mapref, sparse,
239 page, align, size, &vma);
240 mutex_unlock(&vmm->mutex);
241 if (ret)
242 return ret;
243
244 args->v0.addr = vma->addr;
245 vma->user = !client->super;
246 return ret;
247}
248
249static int
250nvkm_uvmm_mthd_page(struct nvkm_uvmm *uvmm, void *argv, u32 argc)
251{
252 union {
253 struct nvif_vmm_page_v0 v0;
254 } *args = argv;
255 const struct nvkm_vmm_page *page;
256 int ret = -ENOSYS;
257 u8 type, index, nr;
258
259 page = uvmm->vmm->func->page;
260 for (nr = 0; page[nr].shift; nr++);
261
262 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
263 if ((index = args->v0.index) >= nr)
264 return -EINVAL;
265 type = page[index].type;
266 args->v0.shift = page[index].shift;
267 args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE);
268 args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM);
269 args->v0.host = !!(type & NVKM_VMM_PAGE_HOST);
270 args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP);
271 } else
272 return -ENOSYS;
273
274 return 0;
275}
276
277static int
278nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc)
279{
280 struct nvkm_uvmm *uvmm = nvkm_uvmm(object);
281 switch (mthd) {
282 case NVIF_VMM_V0_PAGE : return nvkm_uvmm_mthd_page (uvmm, argv, argc);
283 case NVIF_VMM_V0_GET : return nvkm_uvmm_mthd_get (uvmm, argv, argc);
284 case NVIF_VMM_V0_PUT : return nvkm_uvmm_mthd_put (uvmm, argv, argc);
285 case NVIF_VMM_V0_MAP : return nvkm_uvmm_mthd_map (uvmm, argv, argc);
286 case NVIF_VMM_V0_UNMAP : return nvkm_uvmm_mthd_unmap (uvmm, argv, argc);
287 default:
288 break;
289 }
290 return -EINVAL;
291}
292
293static void *
294nvkm_uvmm_dtor(struct nvkm_object *object)
295{
296 struct nvkm_uvmm *uvmm = nvkm_uvmm(object);
297 nvkm_vmm_unref(&uvmm->vmm);
298 return uvmm;
299}
300
301static const struct nvkm_object_func
302nvkm_uvmm = {
303 .dtor = nvkm_uvmm_dtor,
304 .mthd = nvkm_uvmm_mthd,
305};
306
307int
308nvkm_uvmm_new(const struct nvkm_oclass *oclass, void *argv, u32 argc,
309 struct nvkm_object **pobject)
310{
311 struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu;
312 const bool more = oclass->base.maxver >= 0;
313 union {
314 struct nvif_vmm_v0 v0;
315 } *args = argv;
316 const struct nvkm_vmm_page *page;
317 struct nvkm_uvmm *uvmm;
318 int ret = -ENOSYS;
319 u64 addr, size;
320
321 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) {
322 addr = args->v0.addr;
323 size = args->v0.size;
324 } else
325 return ret;
326
327 if (!(uvmm = kzalloc(sizeof(*uvmm), GFP_KERNEL)))
328 return -ENOMEM;
329 nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object);
330 *pobject = &uvmm->object;
331
332 if (!mmu->vmm) {
333 ret = mmu->func->vmm.ctor(mmu, addr, size, argv, argc,
334 NULL, "user", &uvmm->vmm);
335 if (ret)
336 return ret;
337
338 uvmm->vmm->debug = max(uvmm->vmm->debug, oclass->client->debug);
339 } else {
340 if (size)
341 return -EINVAL;
342
343 uvmm->vmm = nvkm_vmm_ref(mmu->vmm);
344 }
345
346 page = uvmm->vmm->func->page;
347 args->v0.page_nr = 0;
348 while (page && (page++)->shift)
349 args->v0.page_nr++;
350 args->v0.addr = uvmm->vmm->start;
351 args->v0.size = uvmm->vmm->limit;
352 return 0;
353}
354