1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22#include "nv50.h"
23#include "head.h"
24#include "ior.h"
25#include "channv50.h"
26#include "rootnv50.h"
27
28#include <core/gpuobj.h>
29#include <subdev/timer.h>
30
31int
32gv100_disp_wndw_cnt(struct nvkm_disp *disp, unsigned long *pmask)
33{
34 struct nvkm_device *device = disp->engine.subdev.device;
35 *pmask = nvkm_rd32(device, 0x610064);
36 return (nvkm_rd32(device, 0x610074) & 0x03f00000) >> 20;
37}
38
39void
40gv100_disp_super(struct work_struct *work)
41{
42 struct nv50_disp *disp =
43 container_of(work, struct nv50_disp, supervisor);
44 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
45 struct nvkm_device *device = subdev->device;
46 struct nvkm_head *head;
47 u32 stat = nvkm_rd32(device, 0x6107a8);
48 u32 mask[4];
49
50 nvkm_debug(subdev, "supervisor %d: %08x\n", ffs(disp->super), stat);
51 list_for_each_entry(head, &disp->base.head, head) {
52 mask[head->id] = nvkm_rd32(device, 0x6107ac + (head->id * 4));
53 HEAD_DBG(head, "%08x", mask[head->id]);
54 }
55
56 if (disp->super & 0x00000001) {
57 nv50_disp_chan_mthd(disp->chan[0], NV_DBG_DEBUG);
58 nv50_disp_super_1(disp);
59 list_for_each_entry(head, &disp->base.head, head) {
60 if (!(mask[head->id] & 0x00001000))
61 continue;
62 nv50_disp_super_1_0(disp, head);
63 }
64 } else
65 if (disp->super & 0x00000002) {
66 list_for_each_entry(head, &disp->base.head, head) {
67 if (!(mask[head->id] & 0x00001000))
68 continue;
69 nv50_disp_super_2_0(disp, head);
70 }
71 nvkm_outp_route(&disp->base);
72 list_for_each_entry(head, &disp->base.head, head) {
73 if (!(mask[head->id] & 0x00010000))
74 continue;
75 nv50_disp_super_2_1(disp, head);
76 }
77 list_for_each_entry(head, &disp->base.head, head) {
78 if (!(mask[head->id] & 0x00001000))
79 continue;
80 nv50_disp_super_2_2(disp, head);
81 }
82 } else
83 if (disp->super & 0x00000004) {
84 list_for_each_entry(head, &disp->base.head, head) {
85 if (!(mask[head->id] & 0x00001000))
86 continue;
87 nv50_disp_super_3_0(disp, head);
88 }
89 }
90
91 list_for_each_entry(head, &disp->base.head, head)
92 nvkm_wr32(device, 0x6107ac + (head->id * 4), 0x00000000);
93 nvkm_wr32(device, 0x6107a8, 0x80000000);
94}
95
96static void
97gv100_disp_exception(struct nv50_disp *disp, int chid)
98{
99 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
100 struct nvkm_device *device = subdev->device;
101 u32 stat = nvkm_rd32(device, 0x611020 + (chid * 12));
102 u32 type = (stat & 0x00007000) >> 12;
103 u32 mthd = (stat & 0x00000fff) << 2;
104 const struct nvkm_enum *reason =
105 nvkm_enum_find(nv50_disp_intr_error_type, type);
106
107
108
109
110
111
112 if (chid <= 32) {
113 u32 data = nvkm_rd32(device, 0x611024 + (chid * 12));
114 u32 code = nvkm_rd32(device, 0x611028 + (chid * 12));
115 nvkm_error(subdev, "chid %d stat %08x reason %d [%s] "
116 "mthd %04x data %08x code %08x\n",
117 chid, stat, type, reason ? reason->name : "",
118 mthd, data, code);
119 } else {
120 nvkm_error(subdev, "chid %d stat %08x reason %d [%s] "
121 "mthd %04x\n",
122 chid, stat, type, reason ? reason->name : "", mthd);
123 }
124
125 if (chid < ARRAY_SIZE(disp->chan) && disp->chan[chid]) {
126 switch (mthd) {
127 case 0x0200:
128 nv50_disp_chan_mthd(disp->chan[chid], NV_DBG_ERROR);
129 break;
130 default:
131 break;
132 }
133 }
134
135 nvkm_wr32(device, 0x611020 + (chid * 12), 0x90000000);
136}
137
138static void
139gv100_disp_intr_ctrl_disp(struct nv50_disp *disp)
140{
141 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
142 struct nvkm_device *device = subdev->device;
143 u32 stat = nvkm_rd32(device, 0x611c30);
144
145 if (stat & 0x00000007) {
146 disp->super = (stat & 0x00000007);
147 queue_work(disp->wq, &disp->supervisor);
148 nvkm_wr32(device, 0x611860, disp->super);
149 stat &= ~0x00000007;
150 }
151
152
153
154
155 if (stat & 0x00000008)
156 stat &= ~0x00000008;
157
158 if (stat & 0x00000080) {
159 u32 error = nvkm_mask(device, 0x611848, 0x00000000, 0x00000000);
160 nvkm_warn(subdev, "error %08x\n", error);
161 stat &= ~0x00000080;
162 }
163
164 if (stat & 0x00000100) {
165 unsigned long wndws = nvkm_rd32(device, 0x611858);
166 unsigned long other = nvkm_rd32(device, 0x61185c);
167 int wndw;
168
169 nvkm_wr32(device, 0x611858, wndws);
170 nvkm_wr32(device, 0x61185c, other);
171
172
173 if (other & 0x00000001)
174 nv50_disp_chan_uevent_send(disp, 0);
175
176
177 for_each_set_bit(wndw, &wndws, disp->wndw.nr) {
178 nv50_disp_chan_uevent_send(disp, 1 + wndw);
179 }
180 }
181
182 if (stat)
183 nvkm_warn(subdev, "ctrl %08x\n", stat);
184}
185
186static void
187gv100_disp_intr_exc_other(struct nv50_disp *disp)
188{
189 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
190 struct nvkm_device *device = subdev->device;
191 u32 stat = nvkm_rd32(device, 0x611854);
192 unsigned long mask;
193 int head;
194
195 if (stat & 0x00000001) {
196 nvkm_wr32(device, 0x611854, 0x00000001);
197 gv100_disp_exception(disp, 0);
198 stat &= ~0x00000001;
199 }
200
201 if ((mask = (stat & 0x00ff0000) >> 16)) {
202 for_each_set_bit(head, &mask, disp->wndw.nr) {
203 nvkm_wr32(device, 0x611854, 0x00010000 << head);
204 gv100_disp_exception(disp, 73 + head);
205 stat &= ~(0x00010000 << head);
206 }
207 }
208
209 if (stat) {
210 nvkm_warn(subdev, "exception %08x\n", stat);
211 nvkm_wr32(device, 0x611854, stat);
212 }
213}
214
215static void
216gv100_disp_intr_exc_winim(struct nv50_disp *disp)
217{
218 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
219 struct nvkm_device *device = subdev->device;
220 unsigned long stat = nvkm_rd32(device, 0x611850);
221 int wndw;
222
223 for_each_set_bit(wndw, &stat, disp->wndw.nr) {
224 nvkm_wr32(device, 0x611850, BIT(wndw));
225 gv100_disp_exception(disp, 33 + wndw);
226 stat &= ~BIT(wndw);
227 }
228
229 if (stat) {
230 nvkm_warn(subdev, "wimm %08x\n", (u32)stat);
231 nvkm_wr32(device, 0x611850, stat);
232 }
233}
234
235static void
236gv100_disp_intr_exc_win(struct nv50_disp *disp)
237{
238 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
239 struct nvkm_device *device = subdev->device;
240 unsigned long stat = nvkm_rd32(device, 0x61184c);
241 int wndw;
242
243 for_each_set_bit(wndw, &stat, disp->wndw.nr) {
244 nvkm_wr32(device, 0x61184c, BIT(wndw));
245 gv100_disp_exception(disp, 1 + wndw);
246 stat &= ~BIT(wndw);
247 }
248
249 if (stat) {
250 nvkm_warn(subdev, "wndw %08x\n", (u32)stat);
251 nvkm_wr32(device, 0x61184c, stat);
252 }
253}
254
255static void
256gv100_disp_intr_head_timing(struct nv50_disp *disp, int head)
257{
258 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
259 struct nvkm_device *device = subdev->device;
260 u32 stat = nvkm_rd32(device, 0x611800 + (head * 0x04));
261
262
263 if (stat & 0x00000003) {
264 nvkm_wr32(device, 0x611800 + (head * 0x04), stat & 0x00000003);
265 stat &= ~0x00000003;
266 }
267
268 if (stat & 0x00000004) {
269 nvkm_disp_vblank(&disp->base, head);
270 nvkm_wr32(device, 0x611800 + (head * 0x04), 0x00000004);
271 stat &= ~0x00000004;
272 }
273
274 if (stat) {
275 nvkm_warn(subdev, "head %08x\n", stat);
276 nvkm_wr32(device, 0x611800 + (head * 0x04), stat);
277 }
278}
279
280void
281gv100_disp_intr(struct nv50_disp *disp)
282{
283 struct nvkm_subdev *subdev = &disp->base.engine.subdev;
284 struct nvkm_device *device = subdev->device;
285 u32 stat = nvkm_rd32(device, 0x611ec0);
286 unsigned long mask;
287 int head;
288
289 if ((mask = (stat & 0x000000ff))) {
290 for_each_set_bit(head, &mask, 8) {
291 gv100_disp_intr_head_timing(disp, head);
292 stat &= ~BIT(head);
293 }
294 }
295
296 if (stat & 0x00000200) {
297 gv100_disp_intr_exc_win(disp);
298 stat &= ~0x00000200;
299 }
300
301 if (stat & 0x00000400) {
302 gv100_disp_intr_exc_winim(disp);
303 stat &= ~0x00000400;
304 }
305
306 if (stat & 0x00000800) {
307 gv100_disp_intr_exc_other(disp);
308 stat &= ~0x00000800;
309 }
310
311 if (stat & 0x00001000) {
312 gv100_disp_intr_ctrl_disp(disp);
313 stat &= ~0x00001000;
314 }
315
316 if (stat)
317 nvkm_warn(subdev, "intr %08x\n", stat);
318}
319
320void
321gv100_disp_fini(struct nv50_disp *disp)
322{
323 struct nvkm_device *device = disp->base.engine.subdev.device;
324 nvkm_wr32(device, 0x611db0, 0x00000000);
325}
326
327static int
328gv100_disp_init(struct nv50_disp *disp)
329{
330 struct nvkm_device *device = disp->base.engine.subdev.device;
331 struct nvkm_head *head;
332 int i, j;
333 u32 tmp;
334
335
336 if (nvkm_rd32(device, 0x6254e8) & 0x00000002) {
337 nvkm_mask(device, 0x6254e8, 0x00000001, 0x00000000);
338 if (nvkm_msec(device, 2000,
339 if (!(nvkm_rd32(device, 0x6254e8) & 0x00000002))
340 break;
341 ) < 0)
342 return -EBUSY;
343 }
344
345
346 tmp = nvkm_rd32(device, 0x610068);
347 nvkm_wr32(device, 0x640008, tmp);
348
349
350 for (i = 0; i < disp->sor.nr; i++) {
351 tmp = nvkm_rd32(device, 0x61c000 + (i * 0x800));
352 nvkm_mask(device, 0x640000, 0x00000100 << i, 0x00000100 << i);
353 nvkm_wr32(device, 0x640144 + (i * 0x08), tmp);
354 }
355
356
357 list_for_each_entry(head, &disp->base.head, head) {
358 const int id = head->id;
359
360
361 tmp = nvkm_rd32(device, 0x616300 + (id * 0x800));
362 nvkm_wr32(device, 0x640048 + (id * 0x020), tmp);
363
364
365 for (j = 0; j < 6 * 4; j += 4) {
366 tmp = nvkm_rd32(device, 0x616100 + (id * 0x800) + j);
367 nvkm_wr32(device, 0x640030 + (id * 0x20) + j, tmp);
368 }
369 }
370
371
372 for (i = 0; i < disp->wndw.nr; i++) {
373 nvkm_mask(device, 0x640004, 1 << i, 1 << i);
374 for (j = 0; j < 6 * 4; j += 4) {
375 tmp = nvkm_rd32(device, 0x630050 + (i * 0x800) + j);
376 nvkm_wr32(device, 0x6401e4 + (i * 0x20) + j, tmp);
377 }
378 }
379
380
381 for (i = 0; i < 4; i++) {
382 tmp = nvkm_rd32(device, 0x62e000 + (i * 0x04));
383 nvkm_wr32(device, 0x640010 + (i * 0x04), tmp);
384 }
385
386 nvkm_mask(device, 0x610078, 0x00000001, 0x00000001);
387
388
389 switch (nvkm_memory_target(disp->inst->memory)) {
390 case NVKM_MEM_TARGET_VRAM: tmp = 0x00000001; break;
391 case NVKM_MEM_TARGET_NCOH: tmp = 0x00000002; break;
392 case NVKM_MEM_TARGET_HOST: tmp = 0x00000003; break;
393 default:
394 break;
395 }
396 nvkm_wr32(device, 0x610010, 0x00000008 | tmp);
397 nvkm_wr32(device, 0x610014, disp->inst->addr >> 16);
398
399
400 nvkm_wr32(device, 0x611cf0, 0x00000187);
401 nvkm_wr32(device, 0x611db0, 0x00000187);
402
403
404 nvkm_wr32(device, 0x611cec, disp->head.mask << 16 |
405 0x00000001);
406 nvkm_wr32(device, 0x611dac, 0x00000000);
407
408
409 nvkm_wr32(device, 0x611ce8, disp->wndw.mask);
410 nvkm_wr32(device, 0x611da8, 0x00000000);
411
412
413 nvkm_wr32(device, 0x611ce4, disp->wndw.mask);
414 nvkm_wr32(device, 0x611da4, 0x00000000);
415
416
417 list_for_each_entry(head, &disp->base.head, head) {
418 const u32 hoff = head->id * 4;
419 nvkm_wr32(device, 0x611cc0 + hoff, 0x00000004);
420 nvkm_wr32(device, 0x611d80 + hoff, 0x00000000);
421 }
422
423
424 nvkm_wr32(device, 0x611cf4, 0x00000000);
425 nvkm_wr32(device, 0x611db4, 0x00000000);
426 return 0;
427}
428
429static const struct nv50_disp_func
430gv100_disp = {
431 .init = gv100_disp_init,
432 .fini = gv100_disp_fini,
433 .intr = gv100_disp_intr,
434 .uevent = &gv100_disp_chan_uevent,
435 .super = gv100_disp_super,
436 .root = &gv100_disp_root_oclass,
437 .wndw = { .cnt = gv100_disp_wndw_cnt },
438 .head = { .cnt = gv100_head_cnt, .new = gv100_head_new },
439 .sor = { .cnt = gv100_sor_cnt, .new = gv100_sor_new },
440 .ramht_size = 0x2000,
441};
442
443int
444gv100_disp_new(struct nvkm_device *device, int index, struct nvkm_disp **pdisp)
445{
446 return nv50_disp_new_(&gv100_disp, device, index, pdisp);
447}
448