1
2
3
4
5
6
7#include <linux/clk.h>
8#include <linux/component.h>
9#include <linux/module.h>
10#include <linux/of_device.h>
11#include <linux/platform_device.h>
12
13#include <drm/drm_atomic.h>
14#include <drm/drm_atomic_helper.h>
15#include <drm/drm_plane_helper.h>
16#include <drm/drm_probe_helper.h>
17#include <drm/drm_vblank.h>
18
19#include "armada_crtc.h"
20#include "armada_drm.h"
21#include "armada_fb.h"
22#include "armada_gem.h"
23#include "armada_hw.h"
24#include "armada_plane.h"
25#include "armada_trace.h"
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81void
82armada_drm_crtc_update_regs(struct armada_crtc *dcrtc, struct armada_regs *regs)
83{
84 while (regs->offset != ~0) {
85 void __iomem *reg = dcrtc->base + regs->offset;
86 uint32_t val;
87
88 val = regs->mask;
89 if (val != 0)
90 val &= readl_relaxed(reg);
91 writel_relaxed(val | regs->val, reg);
92 ++regs;
93 }
94}
95
96static void armada_drm_crtc_update(struct armada_crtc *dcrtc, bool enable)
97{
98 uint32_t dumb_ctrl;
99
100 dumb_ctrl = dcrtc->cfg_dumb_ctrl;
101
102 if (enable)
103 dumb_ctrl |= CFG_DUMB_ENA;
104
105
106
107
108
109
110
111 if (!enable && (dumb_ctrl & DUMB_MASK) == DUMB24_RGB888_0) {
112 dumb_ctrl &= ~DUMB_MASK;
113 dumb_ctrl |= DUMB_BLANK;
114 }
115
116 armada_updatel(dumb_ctrl,
117 ~(CFG_INV_CSYNC | CFG_INV_HSYNC | CFG_INV_VSYNC),
118 dcrtc->base + LCD_SPU_DUMB_CTRL);
119}
120
121static void armada_drm_crtc_queue_state_event(struct drm_crtc *crtc)
122{
123 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
124 struct drm_pending_vblank_event *event;
125
126
127 event = xchg(&crtc->state->event, NULL);
128 if (event) {
129 WARN_ON(drm_crtc_vblank_get(crtc) != 0);
130 dcrtc->event = event;
131 }
132}
133
134static void armada_drm_update_gamma(struct drm_crtc *crtc)
135{
136 struct drm_property_blob *blob = crtc->state->gamma_lut;
137 void __iomem *base = drm_to_armada_crtc(crtc)->base;
138 int i;
139
140 if (blob) {
141 struct drm_color_lut *lut = blob->data;
142
143 armada_updatel(CFG_CSB_256x8, CFG_CSB_256x8 | CFG_PDWN256x8,
144 base + LCD_SPU_SRAM_PARA1);
145
146 for (i = 0; i < 256; i++) {
147 writel_relaxed(drm_color_lut_extract(lut[i].red, 8),
148 base + LCD_SPU_SRAM_WRDAT);
149 writel_relaxed(i | SRAM_WRITE | SRAM_GAMMA_YR,
150 base + LCD_SPU_SRAM_CTRL);
151 readl_relaxed(base + LCD_SPU_HWC_OVSA_HPXL_VLN);
152 writel_relaxed(drm_color_lut_extract(lut[i].green, 8),
153 base + LCD_SPU_SRAM_WRDAT);
154 writel_relaxed(i | SRAM_WRITE | SRAM_GAMMA_UG,
155 base + LCD_SPU_SRAM_CTRL);
156 readl_relaxed(base + LCD_SPU_HWC_OVSA_HPXL_VLN);
157 writel_relaxed(drm_color_lut_extract(lut[i].blue, 8),
158 base + LCD_SPU_SRAM_WRDAT);
159 writel_relaxed(i | SRAM_WRITE | SRAM_GAMMA_VB,
160 base + LCD_SPU_SRAM_CTRL);
161 readl_relaxed(base + LCD_SPU_HWC_OVSA_HPXL_VLN);
162 }
163 armada_updatel(CFG_GAMMA_ENA, CFG_GAMMA_ENA,
164 base + LCD_SPU_DMA_CTRL0);
165 } else {
166 armada_updatel(0, CFG_GAMMA_ENA, base + LCD_SPU_DMA_CTRL0);
167 armada_updatel(CFG_PDWN256x8, CFG_CSB_256x8 | CFG_PDWN256x8,
168 base + LCD_SPU_SRAM_PARA1);
169 }
170}
171
172static enum drm_mode_status armada_drm_crtc_mode_valid(struct drm_crtc *crtc,
173 const struct drm_display_mode *mode)
174{
175 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
176
177 if (mode->vscan > 1)
178 return MODE_NO_VSCAN;
179
180 if (mode->flags & DRM_MODE_FLAG_DBLSCAN)
181 return MODE_NO_DBLESCAN;
182
183 if (mode->flags & DRM_MODE_FLAG_HSKEW)
184 return MODE_H_ILLEGAL;
185
186
187 if (!dcrtc->variant->has_spu_adv_reg &&
188 mode->flags & DRM_MODE_FLAG_INTERLACE)
189 return MODE_NO_INTERLACE;
190
191 if (mode->flags & (DRM_MODE_FLAG_BCAST | DRM_MODE_FLAG_PIXMUX |
192 DRM_MODE_FLAG_CLKDIV2))
193 return MODE_BAD;
194
195 return MODE_OK;
196}
197
198
199static bool armada_drm_crtc_mode_fixup(struct drm_crtc *crtc,
200 const struct drm_display_mode *mode, struct drm_display_mode *adj)
201{
202 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
203 int ret;
204
205
206
207
208
209
210 drm_mode_set_crtcinfo(adj, CRTC_INTERLACE_HALVE_V);
211
212
213
214
215
216 if (armada_drm_crtc_mode_valid(crtc, adj) != MODE_OK)
217 return false;
218
219
220 ret = dcrtc->variant->compute_clock(dcrtc, adj, NULL);
221 if (ret)
222 return false;
223
224 return true;
225}
226
227
228static void armada_drm_crtc_disable_irq(struct armada_crtc *dcrtc, u32 mask)
229{
230 if (dcrtc->irq_ena & mask) {
231 dcrtc->irq_ena &= ~mask;
232 writel(dcrtc->irq_ena, dcrtc->base + LCD_SPU_IRQ_ENA);
233 }
234}
235
236static void armada_drm_crtc_enable_irq(struct armada_crtc *dcrtc, u32 mask)
237{
238 if ((dcrtc->irq_ena & mask) != mask) {
239 dcrtc->irq_ena |= mask;
240 writel(dcrtc->irq_ena, dcrtc->base + LCD_SPU_IRQ_ENA);
241 if (readl_relaxed(dcrtc->base + LCD_SPU_IRQ_ISR) & mask)
242 writel(0, dcrtc->base + LCD_SPU_IRQ_ISR);
243 }
244}
245
246static void armada_drm_crtc_irq(struct armada_crtc *dcrtc, u32 stat)
247{
248 struct drm_pending_vblank_event *event;
249 void __iomem *base = dcrtc->base;
250
251 if (stat & DMA_FF_UNDERFLOW)
252 DRM_ERROR("video underflow on crtc %u\n", dcrtc->num);
253 if (stat & GRA_FF_UNDERFLOW)
254 DRM_ERROR("graphics underflow on crtc %u\n", dcrtc->num);
255
256 if (stat & VSYNC_IRQ)
257 drm_crtc_handle_vblank(&dcrtc->crtc);
258
259 spin_lock(&dcrtc->irq_lock);
260 if (stat & GRA_FRAME_IRQ && dcrtc->interlaced) {
261 int i = stat & GRA_FRAME_IRQ0 ? 0 : 1;
262 uint32_t val;
263
264 writel_relaxed(dcrtc->v[i].spu_v_porch, base + LCD_SPU_V_PORCH);
265 writel_relaxed(dcrtc->v[i].spu_v_h_total,
266 base + LCD_SPUT_V_H_TOTAL);
267
268 val = readl_relaxed(base + LCD_SPU_ADV_REG);
269 val &= ~(ADV_VSYNC_L_OFF | ADV_VSYNC_H_OFF | ADV_VSYNCOFFEN);
270 val |= dcrtc->v[i].spu_adv_reg;
271 writel_relaxed(val, base + LCD_SPU_ADV_REG);
272 }
273
274 if (stat & dcrtc->irq_ena & DUMB_FRAMEDONE) {
275 if (dcrtc->update_pending) {
276 armada_drm_crtc_update_regs(dcrtc, dcrtc->regs);
277 dcrtc->update_pending = false;
278 }
279 if (dcrtc->cursor_update) {
280 writel_relaxed(dcrtc->cursor_hw_pos,
281 base + LCD_SPU_HWC_OVSA_HPXL_VLN);
282 writel_relaxed(dcrtc->cursor_hw_sz,
283 base + LCD_SPU_HWC_HPXL_VLN);
284 armada_updatel(CFG_HWC_ENA,
285 CFG_HWC_ENA | CFG_HWC_1BITMOD |
286 CFG_HWC_1BITENA,
287 base + LCD_SPU_DMA_CTRL0);
288 dcrtc->cursor_update = false;
289 }
290 armada_drm_crtc_disable_irq(dcrtc, DUMB_FRAMEDONE_ENA);
291 }
292 spin_unlock(&dcrtc->irq_lock);
293
294 if (stat & VSYNC_IRQ && !dcrtc->update_pending) {
295 event = xchg(&dcrtc->event, NULL);
296 if (event) {
297 spin_lock(&dcrtc->crtc.dev->event_lock);
298 drm_crtc_send_vblank_event(&dcrtc->crtc, event);
299 spin_unlock(&dcrtc->crtc.dev->event_lock);
300 drm_crtc_vblank_put(&dcrtc->crtc);
301 }
302 }
303}
304
305static irqreturn_t armada_drm_irq(int irq, void *arg)
306{
307 struct armada_crtc *dcrtc = arg;
308 u32 v, stat = readl_relaxed(dcrtc->base + LCD_SPU_IRQ_ISR);
309
310
311
312
313
314
315 writel_relaxed(0, dcrtc->base + LCD_SPU_IRQ_ISR);
316
317 trace_armada_drm_irq(&dcrtc->crtc, stat);
318
319
320 v = stat & dcrtc->irq_ena;
321
322 if (v & (VSYNC_IRQ|GRA_FRAME_IRQ|DUMB_FRAMEDONE)) {
323 armada_drm_crtc_irq(dcrtc, stat);
324 return IRQ_HANDLED;
325 }
326 return IRQ_NONE;
327}
328
329
330static void armada_drm_crtc_mode_set_nofb(struct drm_crtc *crtc)
331{
332 struct drm_display_mode *adj = &crtc->state->adjusted_mode;
333 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
334 struct armada_regs regs[17];
335 uint32_t lm, rm, tm, bm, val, sclk;
336 unsigned long flags;
337 unsigned i;
338 bool interlaced = !!(adj->flags & DRM_MODE_FLAG_INTERLACE);
339
340 i = 0;
341 rm = adj->crtc_hsync_start - adj->crtc_hdisplay;
342 lm = adj->crtc_htotal - adj->crtc_hsync_end;
343 bm = adj->crtc_vsync_start - adj->crtc_vdisplay;
344 tm = adj->crtc_vtotal - adj->crtc_vsync_end;
345
346 DRM_DEBUG_KMS("[CRTC:%d:%s] mode " DRM_MODE_FMT "\n",
347 crtc->base.id, crtc->name, DRM_MODE_ARG(adj));
348 DRM_DEBUG_KMS("lm %d rm %d tm %d bm %d\n", lm, rm, tm, bm);
349
350
351 dcrtc->variant->compute_clock(dcrtc, adj, &sclk);
352
353 armada_reg_queue_set(regs, i, sclk, LCD_CFG_SCLK_DIV);
354
355 spin_lock_irqsave(&dcrtc->irq_lock, flags);
356
357 dcrtc->interlaced = interlaced;
358
359 dcrtc->v[1].spu_v_h_total = adj->crtc_vtotal << 16 |
360 adj->crtc_htotal;
361 dcrtc->v[1].spu_v_porch = tm << 16 | bm;
362 val = adj->crtc_hsync_start;
363 dcrtc->v[1].spu_adv_reg = val << 20 | val | ADV_VSYNCOFFEN;
364
365 if (interlaced) {
366
367 val -= adj->crtc_htotal / 2;
368 dcrtc->v[0].spu_adv_reg = val << 20 | val | ADV_VSYNCOFFEN;
369 dcrtc->v[0].spu_v_h_total = dcrtc->v[1].spu_v_h_total +
370 (1 << 16);
371 dcrtc->v[0].spu_v_porch = dcrtc->v[1].spu_v_porch + 1;
372 } else {
373 dcrtc->v[0] = dcrtc->v[1];
374 }
375
376 val = adj->crtc_vdisplay << 16 | adj->crtc_hdisplay;
377
378 armada_reg_queue_set(regs, i, val, LCD_SPU_V_H_ACTIVE);
379 armada_reg_queue_set(regs, i, (lm << 16) | rm, LCD_SPU_H_PORCH);
380 armada_reg_queue_set(regs, i, dcrtc->v[0].spu_v_porch, LCD_SPU_V_PORCH);
381 armada_reg_queue_set(regs, i, dcrtc->v[0].spu_v_h_total,
382 LCD_SPUT_V_H_TOTAL);
383
384 if (dcrtc->variant->has_spu_adv_reg)
385 armada_reg_queue_mod(regs, i, dcrtc->v[0].spu_adv_reg,
386 ADV_VSYNC_L_OFF | ADV_VSYNC_H_OFF |
387 ADV_VSYNCOFFEN, LCD_SPU_ADV_REG);
388
389 val = adj->flags & DRM_MODE_FLAG_NVSYNC ? CFG_VSYNC_INV : 0;
390 armada_reg_queue_mod(regs, i, val, CFG_VSYNC_INV, LCD_SPU_DMA_CTRL1);
391
392
393
394
395
396
397
398
399
400 val = 0;
401 if (adj->flags & DRM_MODE_FLAG_NCSYNC)
402 val |= CFG_INV_CSYNC;
403 if (adj->flags & DRM_MODE_FLAG_NHSYNC)
404 val |= CFG_INV_HSYNC;
405 if (adj->flags & DRM_MODE_FLAG_NVSYNC)
406 val |= CFG_INV_VSYNC;
407 armada_reg_queue_mod(regs, i, val, CFG_INV_CSYNC | CFG_INV_HSYNC |
408 CFG_INV_VSYNC, LCD_SPU_DUMB_CTRL);
409 armada_reg_queue_end(regs, i);
410
411 armada_drm_crtc_update_regs(dcrtc, regs);
412 spin_unlock_irqrestore(&dcrtc->irq_lock, flags);
413}
414
415static int armada_drm_crtc_atomic_check(struct drm_crtc *crtc,
416 struct drm_atomic_state *state)
417{
418 struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state,
419 crtc);
420 DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
421
422 if (crtc_state->gamma_lut && drm_color_lut_size(crtc_state->gamma_lut) != 256)
423 return -EINVAL;
424
425 if (crtc_state->color_mgmt_changed)
426 crtc_state->planes_changed = true;
427
428 return 0;
429}
430
431static void armada_drm_crtc_atomic_begin(struct drm_crtc *crtc,
432 struct drm_atomic_state *state)
433{
434 struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state,
435 crtc);
436 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
437
438 DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
439
440 if (crtc_state->color_mgmt_changed)
441 armada_drm_update_gamma(crtc);
442
443 dcrtc->regs_idx = 0;
444 dcrtc->regs = dcrtc->atomic_regs;
445}
446
447static void armada_drm_crtc_atomic_flush(struct drm_crtc *crtc,
448 struct drm_atomic_state *state)
449{
450 struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state,
451 crtc);
452 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
453
454 DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
455
456 armada_reg_queue_end(dcrtc->regs, dcrtc->regs_idx);
457
458
459
460
461
462 if (!drm_atomic_crtc_needs_modeset(crtc_state)) {
463 dcrtc->update_pending = true;
464 armada_drm_crtc_queue_state_event(crtc);
465 spin_lock_irq(&dcrtc->irq_lock);
466 armada_drm_crtc_enable_irq(dcrtc, DUMB_FRAMEDONE_ENA);
467 spin_unlock_irq(&dcrtc->irq_lock);
468 } else {
469 spin_lock_irq(&dcrtc->irq_lock);
470 armada_drm_crtc_update_regs(dcrtc, dcrtc->regs);
471 spin_unlock_irq(&dcrtc->irq_lock);
472 }
473}
474
475static void armada_drm_crtc_atomic_disable(struct drm_crtc *crtc,
476 struct drm_atomic_state *state)
477{
478 struct drm_crtc_state *old_state = drm_atomic_get_old_crtc_state(state,
479 crtc);
480 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
481 struct drm_pending_vblank_event *event;
482
483 DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
484
485 if (old_state->adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE)
486 drm_crtc_vblank_put(crtc);
487
488 drm_crtc_vblank_off(crtc);
489 armada_drm_crtc_update(dcrtc, false);
490
491 if (!crtc->state->active) {
492
493
494
495
496 if (dcrtc->variant->disable)
497 dcrtc->variant->disable(dcrtc);
498
499
500
501
502
503 event = crtc->state->event;
504 crtc->state->event = NULL;
505 if (event) {
506 spin_lock_irq(&crtc->dev->event_lock);
507 drm_crtc_send_vblank_event(crtc, event);
508 spin_unlock_irq(&crtc->dev->event_lock);
509 }
510 }
511}
512
513static void armada_drm_crtc_atomic_enable(struct drm_crtc *crtc,
514 struct drm_atomic_state *state)
515{
516 struct drm_crtc_state *old_state = drm_atomic_get_old_crtc_state(state,
517 crtc);
518 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
519
520 DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name);
521
522 if (!old_state->active) {
523
524
525
526
527
528 if (dcrtc->variant->enable)
529 dcrtc->variant->enable(dcrtc, &crtc->state->adjusted_mode);
530 }
531 armada_drm_crtc_update(dcrtc, true);
532 drm_crtc_vblank_on(crtc);
533
534 if (crtc->state->adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE)
535 WARN_ON(drm_crtc_vblank_get(crtc));
536
537 armada_drm_crtc_queue_state_event(crtc);
538}
539
540static const struct drm_crtc_helper_funcs armada_crtc_helper_funcs = {
541 .mode_valid = armada_drm_crtc_mode_valid,
542 .mode_fixup = armada_drm_crtc_mode_fixup,
543 .mode_set_nofb = armada_drm_crtc_mode_set_nofb,
544 .atomic_check = armada_drm_crtc_atomic_check,
545 .atomic_begin = armada_drm_crtc_atomic_begin,
546 .atomic_flush = armada_drm_crtc_atomic_flush,
547 .atomic_disable = armada_drm_crtc_atomic_disable,
548 .atomic_enable = armada_drm_crtc_atomic_enable,
549};
550
551static void armada_load_cursor_argb(void __iomem *base, uint32_t *pix,
552 unsigned stride, unsigned width, unsigned height)
553{
554 uint32_t addr;
555 unsigned y;
556
557 addr = SRAM_HWC32_RAM1;
558 for (y = 0; y < height; y++) {
559 uint32_t *p = &pix[y * stride];
560 unsigned x;
561
562 for (x = 0; x < width; x++, p++) {
563 uint32_t val = *p;
564
565
566
567
568
569
570
571
572 val = (val & 0xff00ff00) |
573 (val & 0x000000ff) << 16 |
574 (val & 0x00ff0000) >> 16;
575
576 writel_relaxed(val,
577 base + LCD_SPU_SRAM_WRDAT);
578 writel_relaxed(addr | SRAM_WRITE,
579 base + LCD_SPU_SRAM_CTRL);
580 readl_relaxed(base + LCD_SPU_HWC_OVSA_HPXL_VLN);
581 addr += 1;
582 if ((addr & 0x00ff) == 0)
583 addr += 0xf00;
584 if ((addr & 0x30ff) == 0)
585 addr = SRAM_HWC32_RAM2;
586 }
587 }
588}
589
590static void armada_drm_crtc_cursor_tran(void __iomem *base)
591{
592 unsigned addr;
593
594 for (addr = 0; addr < 256; addr++) {
595
596 writel_relaxed(0x55555555, base + LCD_SPU_SRAM_WRDAT);
597 writel_relaxed(addr | SRAM_WRITE | SRAM_HWC32_TRAN,
598 base + LCD_SPU_SRAM_CTRL);
599 }
600}
601
602static int armada_drm_crtc_cursor_update(struct armada_crtc *dcrtc, bool reload)
603{
604 uint32_t xoff, xscr, w = dcrtc->cursor_w, s;
605 uint32_t yoff, yscr, h = dcrtc->cursor_h;
606 uint32_t para1;
607
608
609
610
611
612 if (dcrtc->cursor_x < 0) {
613 xoff = -dcrtc->cursor_x;
614 xscr = 0;
615 w -= min(xoff, w);
616 } else if (dcrtc->cursor_x + w > dcrtc->crtc.mode.hdisplay) {
617 xoff = 0;
618 xscr = dcrtc->cursor_x;
619 w = max_t(int, dcrtc->crtc.mode.hdisplay - dcrtc->cursor_x, 0);
620 } else {
621 xoff = 0;
622 xscr = dcrtc->cursor_x;
623 }
624
625 if (dcrtc->cursor_y < 0) {
626 yoff = -dcrtc->cursor_y;
627 yscr = 0;
628 h -= min(yoff, h);
629 } else if (dcrtc->cursor_y + h > dcrtc->crtc.mode.vdisplay) {
630 yoff = 0;
631 yscr = dcrtc->cursor_y;
632 h = max_t(int, dcrtc->crtc.mode.vdisplay - dcrtc->cursor_y, 0);
633 } else {
634 yoff = 0;
635 yscr = dcrtc->cursor_y;
636 }
637
638
639 s = dcrtc->cursor_w;
640 if (dcrtc->interlaced) {
641 s *= 2;
642 yscr /= 2;
643 h /= 2;
644 }
645
646 if (!dcrtc->cursor_obj || !h || !w) {
647 spin_lock_irq(&dcrtc->irq_lock);
648 dcrtc->cursor_update = false;
649 armada_updatel(0, CFG_HWC_ENA, dcrtc->base + LCD_SPU_DMA_CTRL0);
650 spin_unlock_irq(&dcrtc->irq_lock);
651 return 0;
652 }
653
654 spin_lock_irq(&dcrtc->irq_lock);
655 para1 = readl_relaxed(dcrtc->base + LCD_SPU_SRAM_PARA1);
656 armada_updatel(CFG_CSB_256x32, CFG_CSB_256x32 | CFG_PDWN256x32,
657 dcrtc->base + LCD_SPU_SRAM_PARA1);
658 spin_unlock_irq(&dcrtc->irq_lock);
659
660
661
662
663
664 if (!(para1 & CFG_CSB_256x32)) {
665 armada_drm_crtc_cursor_tran(dcrtc->base);
666 reload = true;
667 }
668
669 if (dcrtc->cursor_hw_sz != (h << 16 | w)) {
670 spin_lock_irq(&dcrtc->irq_lock);
671 dcrtc->cursor_update = false;
672 armada_updatel(0, CFG_HWC_ENA, dcrtc->base + LCD_SPU_DMA_CTRL0);
673 spin_unlock_irq(&dcrtc->irq_lock);
674 reload = true;
675 }
676 if (reload) {
677 struct armada_gem_object *obj = dcrtc->cursor_obj;
678 uint32_t *pix;
679
680 pix = obj->addr;
681 pix += yoff * s + xoff;
682 armada_load_cursor_argb(dcrtc->base, pix, s, w, h);
683 }
684
685
686 spin_lock_irq(&dcrtc->irq_lock);
687 dcrtc->cursor_hw_pos = yscr << 16 | xscr;
688 dcrtc->cursor_hw_sz = h << 16 | w;
689 dcrtc->cursor_update = true;
690 armada_drm_crtc_enable_irq(dcrtc, DUMB_FRAMEDONE_ENA);
691 spin_unlock_irq(&dcrtc->irq_lock);
692
693 return 0;
694}
695
696static void cursor_update(void *data)
697{
698 armada_drm_crtc_cursor_update(data, true);
699}
700
701static int armada_drm_crtc_cursor_set(struct drm_crtc *crtc,
702 struct drm_file *file, uint32_t handle, uint32_t w, uint32_t h)
703{
704 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
705 struct armada_gem_object *obj = NULL;
706 int ret;
707
708
709 if (!dcrtc->variant->has_spu_adv_reg)
710 return -ENXIO;
711
712 if (handle && w > 0 && h > 0) {
713
714 if (w > 64 || h > 64 || (w > 32 && h > 32))
715 return -ENOMEM;
716
717 obj = armada_gem_object_lookup(file, handle);
718 if (!obj)
719 return -ENOENT;
720
721
722 if (!obj->addr) {
723 drm_gem_object_put(&obj->obj);
724 return -EINVAL;
725 }
726
727 if (obj->obj.size < w * h * 4) {
728 DRM_ERROR("buffer is too small\n");
729 drm_gem_object_put(&obj->obj);
730 return -ENOMEM;
731 }
732 }
733
734 if (dcrtc->cursor_obj) {
735 dcrtc->cursor_obj->update = NULL;
736 dcrtc->cursor_obj->update_data = NULL;
737 drm_gem_object_put(&dcrtc->cursor_obj->obj);
738 }
739 dcrtc->cursor_obj = obj;
740 dcrtc->cursor_w = w;
741 dcrtc->cursor_h = h;
742 ret = armada_drm_crtc_cursor_update(dcrtc, true);
743 if (obj) {
744 obj->update_data = dcrtc;
745 obj->update = cursor_update;
746 }
747
748 return ret;
749}
750
751static int armada_drm_crtc_cursor_move(struct drm_crtc *crtc, int x, int y)
752{
753 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
754 int ret;
755
756
757 if (!dcrtc->variant->has_spu_adv_reg)
758 return -EFAULT;
759
760 dcrtc->cursor_x = x;
761 dcrtc->cursor_y = y;
762 ret = armada_drm_crtc_cursor_update(dcrtc, false);
763
764 return ret;
765}
766
767static void armada_drm_crtc_destroy(struct drm_crtc *crtc)
768{
769 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
770 struct armada_private *priv = drm_to_armada_dev(crtc->dev);
771
772 if (dcrtc->cursor_obj)
773 drm_gem_object_put(&dcrtc->cursor_obj->obj);
774
775 priv->dcrtc[dcrtc->num] = NULL;
776 drm_crtc_cleanup(&dcrtc->crtc);
777
778 if (dcrtc->variant->disable)
779 dcrtc->variant->disable(dcrtc);
780
781 writel_relaxed(0, dcrtc->base + LCD_SPU_IRQ_ENA);
782
783 of_node_put(dcrtc->crtc.port);
784
785 kfree(dcrtc);
786}
787
788static int armada_drm_crtc_late_register(struct drm_crtc *crtc)
789{
790 if (IS_ENABLED(CONFIG_DEBUG_FS))
791 armada_drm_crtc_debugfs_init(drm_to_armada_crtc(crtc));
792
793 return 0;
794}
795
796
797static int armada_drm_crtc_enable_vblank(struct drm_crtc *crtc)
798{
799 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
800 unsigned long flags;
801
802 spin_lock_irqsave(&dcrtc->irq_lock, flags);
803 armada_drm_crtc_enable_irq(dcrtc, VSYNC_IRQ_ENA);
804 spin_unlock_irqrestore(&dcrtc->irq_lock, flags);
805 return 0;
806}
807
808static void armada_drm_crtc_disable_vblank(struct drm_crtc *crtc)
809{
810 struct armada_crtc *dcrtc = drm_to_armada_crtc(crtc);
811 unsigned long flags;
812
813 spin_lock_irqsave(&dcrtc->irq_lock, flags);
814 armada_drm_crtc_disable_irq(dcrtc, VSYNC_IRQ_ENA);
815 spin_unlock_irqrestore(&dcrtc->irq_lock, flags);
816}
817
818static const struct drm_crtc_funcs armada_crtc_funcs = {
819 .reset = drm_atomic_helper_crtc_reset,
820 .cursor_set = armada_drm_crtc_cursor_set,
821 .cursor_move = armada_drm_crtc_cursor_move,
822 .destroy = armada_drm_crtc_destroy,
823 .set_config = drm_atomic_helper_set_config,
824 .page_flip = drm_atomic_helper_page_flip,
825 .atomic_duplicate_state = drm_atomic_helper_crtc_duplicate_state,
826 .atomic_destroy_state = drm_atomic_helper_crtc_destroy_state,
827 .late_register = armada_drm_crtc_late_register,
828 .enable_vblank = armada_drm_crtc_enable_vblank,
829 .disable_vblank = armada_drm_crtc_disable_vblank,
830};
831
832int armada_crtc_select_clock(struct armada_crtc *dcrtc,
833 struct armada_clk_result *res,
834 const struct armada_clocking_params *params,
835 struct clk *clks[], size_t num_clks,
836 unsigned long desired_khz)
837{
838 unsigned long desired_hz = desired_khz * 1000;
839 unsigned long desired_clk_hz;
840 unsigned long real_clk_hz;
841 unsigned long real_hz;
842 unsigned long permillage;
843 struct clk *clk;
844 u32 div;
845 int i;
846
847 DRM_DEBUG_KMS("[CRTC:%u:%s] desired clock=%luHz\n",
848 dcrtc->crtc.base.id, dcrtc->crtc.name, desired_hz);
849
850 for (i = 0; i < num_clks; i++) {
851 clk = clks[i];
852 if (!clk)
853 continue;
854
855 if (params->settable & BIT(i)) {
856 real_clk_hz = clk_round_rate(clk, desired_hz);
857 desired_clk_hz = desired_hz;
858 } else {
859 real_clk_hz = clk_get_rate(clk);
860 desired_clk_hz = real_clk_hz;
861 }
862
863
864 if (real_clk_hz == desired_hz) {
865 real_hz = real_clk_hz;
866 div = 1;
867 goto found;
868 }
869
870
871 div = DIV_ROUND_CLOSEST(real_clk_hz, desired_hz);
872 if (div == 0 || div > params->div_max)
873 continue;
874
875
876 real_hz = DIV_ROUND_CLOSEST(real_clk_hz, div);
877
878 DRM_DEBUG_KMS("[CRTC:%u:%s] clk=%u %luHz div=%u real=%luHz\n",
879 dcrtc->crtc.base.id, dcrtc->crtc.name,
880 i, real_clk_hz, div, real_hz);
881
882
883 if (real_hz < desired_hz) {
884 permillage = real_hz / desired_khz;
885 if (permillage < params->permillage_min)
886 continue;
887 } else {
888 permillage = DIV_ROUND_UP(real_hz, desired_khz);
889 if (permillage > params->permillage_max)
890 continue;
891 }
892 goto found;
893 }
894
895 return -ERANGE;
896
897found:
898 DRM_DEBUG_KMS("[CRTC:%u:%s] selected clk=%u %luHz div=%u real=%luHz\n",
899 dcrtc->crtc.base.id, dcrtc->crtc.name,
900 i, real_clk_hz, div, real_hz);
901
902 res->desired_clk_hz = desired_clk_hz;
903 res->clk = clk;
904 res->div = div;
905
906 return i;
907}
908
909static int armada_drm_crtc_create(struct drm_device *drm, struct device *dev,
910 struct resource *res, int irq, const struct armada_variant *variant,
911 struct device_node *port)
912{
913 struct armada_private *priv = drm_to_armada_dev(drm);
914 struct armada_crtc *dcrtc;
915 struct drm_plane *primary;
916 void __iomem *base;
917 int ret;
918
919 base = devm_ioremap_resource(dev, res);
920 if (IS_ERR(base))
921 return PTR_ERR(base);
922
923 dcrtc = kzalloc(sizeof(*dcrtc), GFP_KERNEL);
924 if (!dcrtc) {
925 DRM_ERROR("failed to allocate Armada crtc\n");
926 return -ENOMEM;
927 }
928
929 if (dev != drm->dev)
930 dev_set_drvdata(dev, dcrtc);
931
932 dcrtc->variant = variant;
933 dcrtc->base = base;
934 dcrtc->num = drm->mode_config.num_crtc;
935 dcrtc->cfg_dumb_ctrl = DUMB24_RGB888_0;
936 dcrtc->spu_iopad_ctrl = CFG_VSCALE_LN_EN | CFG_IOPAD_DUMB24;
937 spin_lock_init(&dcrtc->irq_lock);
938 dcrtc->irq_ena = CLEAN_SPU_IRQ_ISR;
939
940
941 writel_relaxed(0x00000001, dcrtc->base + LCD_CFG_SCLK_DIV);
942 writel_relaxed(0x00000000, dcrtc->base + LCD_SPU_BLANKCOLOR);
943 writel_relaxed(dcrtc->spu_iopad_ctrl,
944 dcrtc->base + LCD_SPU_IOPAD_CONTROL);
945 writel_relaxed(0x00000000, dcrtc->base + LCD_SPU_SRAM_PARA0);
946 writel_relaxed(CFG_PDWN256x32 | CFG_PDWN256x24 | CFG_PDWN256x8 |
947 CFG_PDWN32x32 | CFG_PDWN16x66 | CFG_PDWN32x66 |
948 CFG_PDWN64x66, dcrtc->base + LCD_SPU_SRAM_PARA1);
949 writel_relaxed(0x2032ff81, dcrtc->base + LCD_SPU_DMA_CTRL1);
950 writel_relaxed(dcrtc->irq_ena, dcrtc->base + LCD_SPU_IRQ_ENA);
951 readl_relaxed(dcrtc->base + LCD_SPU_IRQ_ISR);
952 writel_relaxed(0, dcrtc->base + LCD_SPU_IRQ_ISR);
953
954 ret = devm_request_irq(dev, irq, armada_drm_irq, 0, "armada_drm_crtc",
955 dcrtc);
956 if (ret < 0)
957 goto err_crtc;
958
959 if (dcrtc->variant->init) {
960 ret = dcrtc->variant->init(dcrtc, dev);
961 if (ret)
962 goto err_crtc;
963 }
964
965
966 armada_updatel(CFG_ARBFAST_ENA, 0, dcrtc->base + LCD_SPU_DMA_CTRL0);
967
968 priv->dcrtc[dcrtc->num] = dcrtc;
969
970 dcrtc->crtc.port = port;
971
972 primary = kzalloc(sizeof(*primary), GFP_KERNEL);
973 if (!primary) {
974 ret = -ENOMEM;
975 goto err_crtc;
976 }
977
978 ret = armada_drm_primary_plane_init(drm, primary);
979 if (ret) {
980 kfree(primary);
981 goto err_crtc;
982 }
983
984 ret = drm_crtc_init_with_planes(drm, &dcrtc->crtc, primary, NULL,
985 &armada_crtc_funcs, NULL);
986 if (ret)
987 goto err_crtc_init;
988
989 drm_crtc_helper_add(&dcrtc->crtc, &armada_crtc_helper_funcs);
990
991 ret = drm_mode_crtc_set_gamma_size(&dcrtc->crtc, 256);
992 if (ret)
993 return ret;
994
995 drm_crtc_enable_color_mgmt(&dcrtc->crtc, 0, false, 256);
996
997 return armada_overlay_plane_create(drm, 1 << dcrtc->num);
998
999err_crtc_init:
1000 primary->funcs->destroy(primary);
1001err_crtc:
1002 kfree(dcrtc);
1003
1004 return ret;
1005}
1006
1007static int
1008armada_lcd_bind(struct device *dev, struct device *master, void *data)
1009{
1010 struct platform_device *pdev = to_platform_device(dev);
1011 struct drm_device *drm = data;
1012 struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1013 int irq = platform_get_irq(pdev, 0);
1014 const struct armada_variant *variant;
1015 struct device_node *port = NULL;
1016
1017 if (irq < 0)
1018 return irq;
1019
1020 if (!dev->of_node) {
1021 const struct platform_device_id *id;
1022
1023 id = platform_get_device_id(pdev);
1024 if (!id)
1025 return -ENXIO;
1026
1027 variant = (const struct armada_variant *)id->driver_data;
1028 } else {
1029 const struct of_device_id *match;
1030 struct device_node *np, *parent = dev->of_node;
1031
1032 match = of_match_device(dev->driver->of_match_table, dev);
1033 if (!match)
1034 return -ENXIO;
1035
1036 np = of_get_child_by_name(parent, "ports");
1037 if (np)
1038 parent = np;
1039 port = of_get_child_by_name(parent, "port");
1040 of_node_put(np);
1041 if (!port) {
1042 dev_err(dev, "no port node found in %pOF\n", parent);
1043 return -ENXIO;
1044 }
1045
1046 variant = match->data;
1047 }
1048
1049 return armada_drm_crtc_create(drm, dev, res, irq, variant, port);
1050}
1051
1052static void
1053armada_lcd_unbind(struct device *dev, struct device *master, void *data)
1054{
1055 struct armada_crtc *dcrtc = dev_get_drvdata(dev);
1056
1057 armada_drm_crtc_destroy(&dcrtc->crtc);
1058}
1059
1060static const struct component_ops armada_lcd_ops = {
1061 .bind = armada_lcd_bind,
1062 .unbind = armada_lcd_unbind,
1063};
1064
1065static int armada_lcd_probe(struct platform_device *pdev)
1066{
1067 return component_add(&pdev->dev, &armada_lcd_ops);
1068}
1069
1070static int armada_lcd_remove(struct platform_device *pdev)
1071{
1072 component_del(&pdev->dev, &armada_lcd_ops);
1073 return 0;
1074}
1075
1076static const struct of_device_id armada_lcd_of_match[] = {
1077 {
1078 .compatible = "marvell,dove-lcd",
1079 .data = &armada510_ops,
1080 },
1081 {}
1082};
1083MODULE_DEVICE_TABLE(of, armada_lcd_of_match);
1084
1085static const struct platform_device_id armada_lcd_platform_ids[] = {
1086 {
1087 .name = "armada-lcd",
1088 .driver_data = (unsigned long)&armada510_ops,
1089 }, {
1090 .name = "armada-510-lcd",
1091 .driver_data = (unsigned long)&armada510_ops,
1092 },
1093 { },
1094};
1095MODULE_DEVICE_TABLE(platform, armada_lcd_platform_ids);
1096
1097struct platform_driver armada_lcd_platform_driver = {
1098 .probe = armada_lcd_probe,
1099 .remove = armada_lcd_remove,
1100 .driver = {
1101 .name = "armada-lcd",
1102 .owner = THIS_MODULE,
1103 .of_match_table = armada_lcd_of_match,
1104 },
1105 .id_table = armada_lcd_platform_ids,
1106};
1107