1
2
3
4
5
6
7#include <linux/delay.h>
8
9#include <drm/drm_vblank.h>
10
11#include "msm_drv.h"
12#include "msm_gem.h"
13#include "msm_mmu.h"
14#include "mdp4_kms.h"
15
16static struct mdp4_platform_config *mdp4_get_config(struct platform_device *dev);
17
18static int mdp4_hw_init(struct msm_kms *kms)
19{
20 struct mdp4_kms *mdp4_kms = to_mdp4_kms(to_mdp_kms(kms));
21 struct drm_device *dev = mdp4_kms->dev;
22 uint32_t version, major, minor, dmap_cfg, vg_cfg;
23 unsigned long clk;
24 int ret = 0;
25
26 pm_runtime_get_sync(dev->dev);
27
28 mdp4_enable(mdp4_kms);
29 version = mdp4_read(mdp4_kms, REG_MDP4_VERSION);
30 mdp4_disable(mdp4_kms);
31
32 major = FIELD(version, MDP4_VERSION_MAJOR);
33 minor = FIELD(version, MDP4_VERSION_MINOR);
34
35 DBG("found MDP4 version v%d.%d", major, minor);
36
37 if (major != 4) {
38 DRM_DEV_ERROR(dev->dev, "unexpected MDP version: v%d.%d\n",
39 major, minor);
40 ret = -ENXIO;
41 goto out;
42 }
43
44 mdp4_kms->rev = minor;
45
46 if (mdp4_kms->rev > 1) {
47 mdp4_write(mdp4_kms, REG_MDP4_CS_CONTROLLER0, 0x0707ffff);
48 mdp4_write(mdp4_kms, REG_MDP4_CS_CONTROLLER1, 0x03073f3f);
49 }
50
51 mdp4_write(mdp4_kms, REG_MDP4_PORTMAP_MODE, 0x3);
52
53
54 mdp4_write(mdp4_kms, REG_MDP4_READ_CNFG, 0x02222);
55
56 clk = clk_get_rate(mdp4_kms->clk);
57
58 if ((mdp4_kms->rev >= 1) || (clk >= 90000000)) {
59 dmap_cfg = 0x47;
60 vg_cfg = 0x47;
61 } else {
62 dmap_cfg = 0x27;
63 vg_cfg = 0x43;
64 }
65
66 DBG("fetch config: dmap=%02x, vg=%02x", dmap_cfg, vg_cfg);
67
68 mdp4_write(mdp4_kms, REG_MDP4_DMA_FETCH_CONFIG(DMA_P), dmap_cfg);
69 mdp4_write(mdp4_kms, REG_MDP4_DMA_FETCH_CONFIG(DMA_E), dmap_cfg);
70
71 mdp4_write(mdp4_kms, REG_MDP4_PIPE_FETCH_CONFIG(VG1), vg_cfg);
72 mdp4_write(mdp4_kms, REG_MDP4_PIPE_FETCH_CONFIG(VG2), vg_cfg);
73 mdp4_write(mdp4_kms, REG_MDP4_PIPE_FETCH_CONFIG(RGB1), vg_cfg);
74 mdp4_write(mdp4_kms, REG_MDP4_PIPE_FETCH_CONFIG(RGB2), vg_cfg);
75
76 if (mdp4_kms->rev >= 2)
77 mdp4_write(mdp4_kms, REG_MDP4_LAYERMIXER_IN_CFG_UPDATE_METHOD, 1);
78 mdp4_write(mdp4_kms, REG_MDP4_LAYERMIXER_IN_CFG, 0);
79
80
81 mdp4_write(mdp4_kms, REG_MDP4_PIPE_OP_MODE(VG1), 0);
82 mdp4_write(mdp4_kms, REG_MDP4_PIPE_OP_MODE(VG2), 0);
83 mdp4_write(mdp4_kms, REG_MDP4_DMA_P_OP_MODE, 0);
84 mdp4_write(mdp4_kms, REG_MDP4_DMA_S_OP_MODE, 0);
85 mdp4_write(mdp4_kms, REG_MDP4_OVLP_CSC_CONFIG(1), 0);
86 mdp4_write(mdp4_kms, REG_MDP4_OVLP_CSC_CONFIG(2), 0);
87
88 if (mdp4_kms->rev > 1)
89 mdp4_write(mdp4_kms, REG_MDP4_RESET_STATUS, 1);
90
91 dev->mode_config.allow_fb_modifiers = true;
92
93out:
94 pm_runtime_put_sync(dev->dev);
95
96 return ret;
97}
98
99static void mdp4_enable_commit(struct msm_kms *kms)
100{
101 struct mdp4_kms *mdp4_kms = to_mdp4_kms(to_mdp_kms(kms));
102 mdp4_enable(mdp4_kms);
103}
104
105static void mdp4_disable_commit(struct msm_kms *kms)
106{
107 struct mdp4_kms *mdp4_kms = to_mdp4_kms(to_mdp_kms(kms));
108 mdp4_disable(mdp4_kms);
109}
110
111static void mdp4_prepare_commit(struct msm_kms *kms, struct drm_atomic_state *state)
112{
113 int i;
114 struct drm_crtc *crtc;
115 struct drm_crtc_state *crtc_state;
116
117
118 for_each_new_crtc_in_state(state, crtc, crtc_state, i)
119 drm_crtc_vblank_get(crtc);
120}
121
122static void mdp4_flush_commit(struct msm_kms *kms, unsigned crtc_mask)
123{
124
125}
126
127static void mdp4_wait_flush(struct msm_kms *kms, unsigned crtc_mask)
128{
129 struct mdp4_kms *mdp4_kms = to_mdp4_kms(to_mdp_kms(kms));
130 struct drm_crtc *crtc;
131
132 for_each_crtc_mask(mdp4_kms->dev, crtc, crtc_mask)
133 mdp4_crtc_wait_for_commit_done(crtc);
134}
135
136static void mdp4_complete_commit(struct msm_kms *kms, unsigned crtc_mask)
137{
138 struct mdp4_kms *mdp4_kms = to_mdp4_kms(to_mdp_kms(kms));
139 struct drm_crtc *crtc;
140
141
142 for_each_crtc_mask(mdp4_kms->dev, crtc, crtc_mask)
143 drm_crtc_vblank_put(crtc);
144}
145
146static long mdp4_round_pixclk(struct msm_kms *kms, unsigned long rate,
147 struct drm_encoder *encoder)
148{
149
150 switch (encoder->encoder_type) {
151 case DRM_MODE_ENCODER_TMDS:
152 return mdp4_dtv_round_pixclk(encoder, rate);
153 case DRM_MODE_ENCODER_LVDS:
154 case DRM_MODE_ENCODER_DSI:
155 default:
156 return rate;
157 }
158}
159
160static void mdp4_destroy(struct msm_kms *kms)
161{
162 struct mdp4_kms *mdp4_kms = to_mdp4_kms(to_mdp_kms(kms));
163 struct device *dev = mdp4_kms->dev->dev;
164 struct msm_gem_address_space *aspace = kms->aspace;
165
166 if (mdp4_kms->blank_cursor_iova)
167 msm_gem_unpin_iova(mdp4_kms->blank_cursor_bo, kms->aspace);
168 drm_gem_object_put(mdp4_kms->blank_cursor_bo);
169
170 if (aspace) {
171 aspace->mmu->funcs->detach(aspace->mmu);
172 msm_gem_address_space_put(aspace);
173 }
174
175 if (mdp4_kms->rpm_enabled)
176 pm_runtime_disable(dev);
177
178 kfree(mdp4_kms);
179}
180
181static const struct mdp_kms_funcs kms_funcs = {
182 .base = {
183 .hw_init = mdp4_hw_init,
184 .irq_preinstall = mdp4_irq_preinstall,
185 .irq_postinstall = mdp4_irq_postinstall,
186 .irq_uninstall = mdp4_irq_uninstall,
187 .irq = mdp4_irq,
188 .enable_vblank = mdp4_enable_vblank,
189 .disable_vblank = mdp4_disable_vblank,
190 .enable_commit = mdp4_enable_commit,
191 .disable_commit = mdp4_disable_commit,
192 .prepare_commit = mdp4_prepare_commit,
193 .flush_commit = mdp4_flush_commit,
194 .wait_flush = mdp4_wait_flush,
195 .complete_commit = mdp4_complete_commit,
196 .get_format = mdp_get_format,
197 .round_pixclk = mdp4_round_pixclk,
198 .destroy = mdp4_destroy,
199 },
200 .set_irqmask = mdp4_set_irqmask,
201};
202
203int mdp4_disable(struct mdp4_kms *mdp4_kms)
204{
205 DBG("");
206
207 clk_disable_unprepare(mdp4_kms->clk);
208 if (mdp4_kms->pclk)
209 clk_disable_unprepare(mdp4_kms->pclk);
210 if (mdp4_kms->lut_clk)
211 clk_disable_unprepare(mdp4_kms->lut_clk);
212 if (mdp4_kms->axi_clk)
213 clk_disable_unprepare(mdp4_kms->axi_clk);
214
215 return 0;
216}
217
218int mdp4_enable(struct mdp4_kms *mdp4_kms)
219{
220 DBG("");
221
222 clk_prepare_enable(mdp4_kms->clk);
223 if (mdp4_kms->pclk)
224 clk_prepare_enable(mdp4_kms->pclk);
225 if (mdp4_kms->lut_clk)
226 clk_prepare_enable(mdp4_kms->lut_clk);
227 if (mdp4_kms->axi_clk)
228 clk_prepare_enable(mdp4_kms->axi_clk);
229
230 return 0;
231}
232
233
234static int mdp4_modeset_init_intf(struct mdp4_kms *mdp4_kms,
235 int intf_type)
236{
237 struct drm_device *dev = mdp4_kms->dev;
238 struct msm_drm_private *priv = dev->dev_private;
239 struct drm_encoder *encoder;
240 struct drm_connector *connector;
241 struct device_node *panel_node;
242 int dsi_id;
243 int ret;
244
245 switch (intf_type) {
246 case DRM_MODE_ENCODER_LVDS:
247
248
249
250
251 panel_node = of_graph_get_remote_node(dev->dev->of_node, 0, 0);
252 if (!panel_node)
253 return 0;
254
255 encoder = mdp4_lcdc_encoder_init(dev, panel_node);
256 if (IS_ERR(encoder)) {
257 DRM_DEV_ERROR(dev->dev, "failed to construct LCDC encoder\n");
258 return PTR_ERR(encoder);
259 }
260
261
262 encoder->possible_crtcs = 1 << DMA_P;
263
264 connector = mdp4_lvds_connector_init(dev, panel_node, encoder);
265 if (IS_ERR(connector)) {
266 DRM_DEV_ERROR(dev->dev, "failed to initialize LVDS connector\n");
267 return PTR_ERR(connector);
268 }
269
270 priv->encoders[priv->num_encoders++] = encoder;
271 priv->connectors[priv->num_connectors++] = connector;
272
273 break;
274 case DRM_MODE_ENCODER_TMDS:
275 encoder = mdp4_dtv_encoder_init(dev);
276 if (IS_ERR(encoder)) {
277 DRM_DEV_ERROR(dev->dev, "failed to construct DTV encoder\n");
278 return PTR_ERR(encoder);
279 }
280
281
282 encoder->possible_crtcs = 1 << 1;
283
284 if (priv->hdmi) {
285
286 ret = msm_hdmi_modeset_init(priv->hdmi, dev, encoder);
287 if (ret) {
288 DRM_DEV_ERROR(dev->dev, "failed to initialize HDMI: %d\n", ret);
289 return ret;
290 }
291 }
292
293 priv->encoders[priv->num_encoders++] = encoder;
294
295 break;
296 case DRM_MODE_ENCODER_DSI:
297
298 dsi_id = 0;
299
300 if (!priv->dsi[dsi_id])
301 break;
302
303 encoder = mdp4_dsi_encoder_init(dev);
304 if (IS_ERR(encoder)) {
305 ret = PTR_ERR(encoder);
306 DRM_DEV_ERROR(dev->dev,
307 "failed to construct DSI encoder: %d\n", ret);
308 return ret;
309 }
310
311
312 encoder->possible_crtcs = 1 << DMA_P;
313 priv->encoders[priv->num_encoders++] = encoder;
314
315 ret = msm_dsi_modeset_init(priv->dsi[dsi_id], dev, encoder);
316 if (ret) {
317 DRM_DEV_ERROR(dev->dev, "failed to initialize DSI: %d\n",
318 ret);
319 return ret;
320 }
321
322 break;
323 default:
324 DRM_DEV_ERROR(dev->dev, "Invalid or unsupported interface\n");
325 return -EINVAL;
326 }
327
328 return 0;
329}
330
331static int modeset_init(struct mdp4_kms *mdp4_kms)
332{
333 struct drm_device *dev = mdp4_kms->dev;
334 struct msm_drm_private *priv = dev->dev_private;
335 struct drm_plane *plane;
336 struct drm_crtc *crtc;
337 int i, ret;
338 static const enum mdp4_pipe rgb_planes[] = {
339 RGB1, RGB2,
340 };
341 static const enum mdp4_pipe vg_planes[] = {
342 VG1, VG2,
343 };
344 static const enum mdp4_dma mdp4_crtcs[] = {
345 DMA_P, DMA_E,
346 };
347 static const char * const mdp4_crtc_names[] = {
348 "DMA_P", "DMA_E",
349 };
350 static const int mdp4_intfs[] = {
351 DRM_MODE_ENCODER_LVDS,
352 DRM_MODE_ENCODER_DSI,
353 DRM_MODE_ENCODER_TMDS,
354 };
355
356
357 for (i = 0; i < ARRAY_SIZE(vg_planes); i++) {
358 plane = mdp4_plane_init(dev, vg_planes[i], false);
359 if (IS_ERR(plane)) {
360 DRM_DEV_ERROR(dev->dev,
361 "failed to construct plane for VG%d\n", i + 1);
362 ret = PTR_ERR(plane);
363 goto fail;
364 }
365 priv->planes[priv->num_planes++] = plane;
366 }
367
368 for (i = 0; i < ARRAY_SIZE(mdp4_crtcs); i++) {
369 plane = mdp4_plane_init(dev, rgb_planes[i], true);
370 if (IS_ERR(plane)) {
371 DRM_DEV_ERROR(dev->dev,
372 "failed to construct plane for RGB%d\n", i + 1);
373 ret = PTR_ERR(plane);
374 goto fail;
375 }
376
377 crtc = mdp4_crtc_init(dev, plane, priv->num_crtcs, i,
378 mdp4_crtcs[i]);
379 if (IS_ERR(crtc)) {
380 DRM_DEV_ERROR(dev->dev, "failed to construct crtc for %s\n",
381 mdp4_crtc_names[i]);
382 ret = PTR_ERR(crtc);
383 goto fail;
384 }
385
386 priv->crtcs[priv->num_crtcs++] = crtc;
387 }
388
389
390
391
392
393
394
395
396
397
398
399 for (i = 0; i < ARRAY_SIZE(mdp4_intfs); i++) {
400 ret = mdp4_modeset_init_intf(mdp4_kms, mdp4_intfs[i]);
401 if (ret) {
402 DRM_DEV_ERROR(dev->dev, "failed to initialize intf: %d, %d\n",
403 i, ret);
404 goto fail;
405 }
406 }
407
408 return 0;
409
410fail:
411 return ret;
412}
413
414struct msm_kms *mdp4_kms_init(struct drm_device *dev)
415{
416 struct platform_device *pdev = to_platform_device(dev->dev);
417 struct mdp4_platform_config *config = mdp4_get_config(pdev);
418 struct mdp4_kms *mdp4_kms;
419 struct msm_kms *kms = NULL;
420 struct msm_gem_address_space *aspace;
421 int irq, ret;
422
423 mdp4_kms = kzalloc(sizeof(*mdp4_kms), GFP_KERNEL);
424 if (!mdp4_kms) {
425 DRM_DEV_ERROR(dev->dev, "failed to allocate kms\n");
426 ret = -ENOMEM;
427 goto fail;
428 }
429
430 mdp_kms_init(&mdp4_kms->base, &kms_funcs);
431
432 kms = &mdp4_kms->base.base;
433
434 mdp4_kms->dev = dev;
435
436 mdp4_kms->mmio = msm_ioremap(pdev, NULL, "MDP4");
437 if (IS_ERR(mdp4_kms->mmio)) {
438 ret = PTR_ERR(mdp4_kms->mmio);
439 goto fail;
440 }
441
442 irq = platform_get_irq(pdev, 0);
443 if (irq < 0) {
444 ret = irq;
445 DRM_DEV_ERROR(dev->dev, "failed to get irq: %d\n", ret);
446 goto fail;
447 }
448
449 kms->irq = irq;
450
451
452
453
454
455 mdp4_kms->vdd = devm_regulator_get_exclusive(&pdev->dev, "vdd");
456 if (IS_ERR(mdp4_kms->vdd))
457 mdp4_kms->vdd = NULL;
458
459 if (mdp4_kms->vdd) {
460 ret = regulator_enable(mdp4_kms->vdd);
461 if (ret) {
462 DRM_DEV_ERROR(dev->dev, "failed to enable regulator vdd: %d\n", ret);
463 goto fail;
464 }
465 }
466
467 mdp4_kms->clk = devm_clk_get(&pdev->dev, "core_clk");
468 if (IS_ERR(mdp4_kms->clk)) {
469 DRM_DEV_ERROR(dev->dev, "failed to get core_clk\n");
470 ret = PTR_ERR(mdp4_kms->clk);
471 goto fail;
472 }
473
474 mdp4_kms->pclk = devm_clk_get(&pdev->dev, "iface_clk");
475 if (IS_ERR(mdp4_kms->pclk))
476 mdp4_kms->pclk = NULL;
477
478 if (mdp4_kms->rev >= 2) {
479 mdp4_kms->lut_clk = devm_clk_get(&pdev->dev, "lut_clk");
480 if (IS_ERR(mdp4_kms->lut_clk)) {
481 DRM_DEV_ERROR(dev->dev, "failed to get lut_clk\n");
482 ret = PTR_ERR(mdp4_kms->lut_clk);
483 goto fail;
484 }
485 }
486
487 mdp4_kms->axi_clk = devm_clk_get(&pdev->dev, "bus_clk");
488 if (IS_ERR(mdp4_kms->axi_clk)) {
489 DRM_DEV_ERROR(dev->dev, "failed to get axi_clk\n");
490 ret = PTR_ERR(mdp4_kms->axi_clk);
491 goto fail;
492 }
493
494 clk_set_rate(mdp4_kms->clk, config->max_clk);
495 if (mdp4_kms->lut_clk)
496 clk_set_rate(mdp4_kms->lut_clk, config->max_clk);
497
498 pm_runtime_enable(dev->dev);
499 mdp4_kms->rpm_enabled = true;
500
501
502
503
504
505 mdp4_enable(mdp4_kms);
506 mdp4_write(mdp4_kms, REG_MDP4_DTV_ENABLE, 0);
507 mdp4_write(mdp4_kms, REG_MDP4_LCDC_ENABLE, 0);
508 mdp4_write(mdp4_kms, REG_MDP4_DSI_ENABLE, 0);
509 mdp4_disable(mdp4_kms);
510 mdelay(16);
511
512 if (config->iommu) {
513 struct msm_mmu *mmu = msm_iommu_new(&pdev->dev,
514 config->iommu);
515
516 aspace = msm_gem_address_space_create(mmu,
517 "mdp4", 0x1000, 0x100000000 - 0x1000);
518
519 if (IS_ERR(aspace)) {
520 if (!IS_ERR(mmu))
521 mmu->funcs->destroy(mmu);
522 ret = PTR_ERR(aspace);
523 goto fail;
524 }
525
526 kms->aspace = aspace;
527 } else {
528 DRM_DEV_INFO(dev->dev, "no iommu, fallback to phys "
529 "contig buffers for scanout\n");
530 aspace = NULL;
531 }
532
533 ret = modeset_init(mdp4_kms);
534 if (ret) {
535 DRM_DEV_ERROR(dev->dev, "modeset_init failed: %d\n", ret);
536 goto fail;
537 }
538
539 mdp4_kms->blank_cursor_bo = msm_gem_new(dev, SZ_16K, MSM_BO_WC | MSM_BO_SCANOUT);
540 if (IS_ERR(mdp4_kms->blank_cursor_bo)) {
541 ret = PTR_ERR(mdp4_kms->blank_cursor_bo);
542 DRM_DEV_ERROR(dev->dev, "could not allocate blank-cursor bo: %d\n", ret);
543 mdp4_kms->blank_cursor_bo = NULL;
544 goto fail;
545 }
546
547 ret = msm_gem_get_and_pin_iova(mdp4_kms->blank_cursor_bo, kms->aspace,
548 &mdp4_kms->blank_cursor_iova);
549 if (ret) {
550 DRM_DEV_ERROR(dev->dev, "could not pin blank-cursor bo: %d\n", ret);
551 goto fail;
552 }
553
554 dev->mode_config.min_width = 0;
555 dev->mode_config.min_height = 0;
556 dev->mode_config.max_width = 2048;
557 dev->mode_config.max_height = 2048;
558
559 return kms;
560
561fail:
562 if (kms)
563 mdp4_destroy(kms);
564 return ERR_PTR(ret);
565}
566
567static struct mdp4_platform_config *mdp4_get_config(struct platform_device *dev)
568{
569 static struct mdp4_platform_config config = {};
570
571
572 config.max_clk = 266667000;
573 config.iommu = iommu_domain_alloc(&platform_bus_type);
574
575 return &config;
576}
577