linux/drivers/video/fbdev/mmp/hw/mmp_ctrl.c
<<
>>
Prefs
   1/*
   2 * linux/drivers/video/mmp/hw/mmp_ctrl.c
   3 * Marvell MMP series Display Controller support
   4 *
   5 * Copyright (C) 2012 Marvell Technology Group Ltd.
   6 * Authors:  Guoqing Li <ligq@marvell.com>
   7 *          Lisa Du <cldu@marvell.com>
   8 *          Zhou Zhu <zzhu3@marvell.com>
   9 *
  10 * This program is free software; you can redistribute it and/or modify it
  11 * under the terms of the GNU General Public License as published by the
  12 * Free Software Foundation; either version 2 of the License, or (at your
  13 * option) any later version.
  14 *
  15 * This program is distributed in the hope that it will be useful, but WITHOUT
  16 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  17 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
  18 * more details.
  19 *
  20 * You should have received a copy of the GNU General Public License along with
  21 * this program.  If not, see <http://www.gnu.org/licenses/>.
  22 *
  23 */
  24#include <linux/module.h>
  25#include <linux/moduleparam.h>
  26#include <linux/kernel.h>
  27#include <linux/errno.h>
  28#include <linux/string.h>
  29#include <linux/interrupt.h>
  30#include <linux/slab.h>
  31#include <linux/delay.h>
  32#include <linux/platform_device.h>
  33#include <linux/dma-mapping.h>
  34#include <linux/clk.h>
  35#include <linux/err.h>
  36#include <linux/vmalloc.h>
  37#include <linux/uaccess.h>
  38#include <linux/kthread.h>
  39#include <linux/io.h>
  40
  41#include "mmp_ctrl.h"
  42
  43static irqreturn_t ctrl_handle_irq(int irq, void *dev_id)
  44{
  45        struct mmphw_ctrl *ctrl = (struct mmphw_ctrl *)dev_id;
  46        u32 isr, imask, tmp;
  47
  48        isr = readl_relaxed(ctrl->reg_base + SPU_IRQ_ISR);
  49        imask = readl_relaxed(ctrl->reg_base + SPU_IRQ_ENA);
  50
  51        do {
  52                /* clear clock only */
  53                tmp = readl_relaxed(ctrl->reg_base + SPU_IRQ_ISR);
  54                if (tmp & isr)
  55                        writel_relaxed(~isr, ctrl->reg_base + SPU_IRQ_ISR);
  56        } while ((isr = readl_relaxed(ctrl->reg_base + SPU_IRQ_ISR)) & imask);
  57
  58        return IRQ_HANDLED;
  59}
  60
  61static u32 fmt_to_reg(struct mmp_overlay *overlay, int pix_fmt)
  62{
  63        u32 rbswap = 0, uvswap = 0, yuvswap = 0,
  64                csc_en = 0, val = 0,
  65                vid = overlay_is_vid(overlay);
  66
  67        switch (pix_fmt) {
  68        case PIXFMT_RGB565:
  69        case PIXFMT_RGB1555:
  70        case PIXFMT_RGB888PACK:
  71        case PIXFMT_RGB888UNPACK:
  72        case PIXFMT_RGBA888:
  73                rbswap = 1;
  74                break;
  75        case PIXFMT_VYUY:
  76        case PIXFMT_YVU422P:
  77        case PIXFMT_YVU420P:
  78                uvswap = 1;
  79                break;
  80        case PIXFMT_YUYV:
  81                yuvswap = 1;
  82                break;
  83        default:
  84                break;
  85        }
  86
  87        switch (pix_fmt) {
  88        case PIXFMT_RGB565:
  89        case PIXFMT_BGR565:
  90                break;
  91        case PIXFMT_RGB1555:
  92        case PIXFMT_BGR1555:
  93                val = 0x1;
  94                break;
  95        case PIXFMT_RGB888PACK:
  96        case PIXFMT_BGR888PACK:
  97                val = 0x2;
  98                break;
  99        case PIXFMT_RGB888UNPACK:
 100        case PIXFMT_BGR888UNPACK:
 101                val = 0x3;
 102                break;
 103        case PIXFMT_RGBA888:
 104        case PIXFMT_BGRA888:
 105                val = 0x4;
 106                break;
 107        case PIXFMT_UYVY:
 108        case PIXFMT_VYUY:
 109        case PIXFMT_YUYV:
 110                val = 0x5;
 111                csc_en = 1;
 112                break;
 113        case PIXFMT_YUV422P:
 114        case PIXFMT_YVU422P:
 115                val = 0x6;
 116                csc_en = 1;
 117                break;
 118        case PIXFMT_YUV420P:
 119        case PIXFMT_YVU420P:
 120                val = 0x7;
 121                csc_en = 1;
 122                break;
 123        default:
 124                break;
 125        }
 126
 127        return (dma_palette(0) | dma_fmt(vid, val) |
 128                dma_swaprb(vid, rbswap) | dma_swapuv(vid, uvswap) |
 129                dma_swapyuv(vid, yuvswap) | dma_csc(vid, csc_en));
 130}
 131
 132static void dmafetch_set_fmt(struct mmp_overlay *overlay)
 133{
 134        u32 tmp;
 135        struct mmp_path *path = overlay->path;
 136        tmp = readl_relaxed(ctrl_regs(path) + dma_ctrl(0, path->id));
 137        tmp &= ~dma_mask(overlay_is_vid(overlay));
 138        tmp |= fmt_to_reg(overlay, overlay->win.pix_fmt);
 139        writel_relaxed(tmp, ctrl_regs(path) + dma_ctrl(0, path->id));
 140}
 141
 142static void overlay_set_win(struct mmp_overlay *overlay, struct mmp_win *win)
 143{
 144        struct lcd_regs *regs = path_regs(overlay->path);
 145
 146        /* assert win supported */
 147        memcpy(&overlay->win, win, sizeof(struct mmp_win));
 148
 149        mutex_lock(&overlay->access_ok);
 150
 151        if (overlay_is_vid(overlay)) {
 152                writel_relaxed(win->pitch[0], &regs->v_pitch_yc);
 153                writel_relaxed(win->pitch[2] << 16 |
 154                                win->pitch[1], &regs->v_pitch_uv);
 155
 156                writel_relaxed((win->ysrc << 16) | win->xsrc, &regs->v_size);
 157                writel_relaxed((win->ydst << 16) | win->xdst, &regs->v_size_z);
 158                writel_relaxed(win->ypos << 16 | win->xpos, &regs->v_start);
 159        } else {
 160                writel_relaxed(win->pitch[0], &regs->g_pitch);
 161
 162                writel_relaxed((win->ysrc << 16) | win->xsrc, &regs->g_size);
 163                writel_relaxed((win->ydst << 16) | win->xdst, &regs->g_size_z);
 164                writel_relaxed(win->ypos << 16 | win->xpos, &regs->g_start);
 165        }
 166
 167        dmafetch_set_fmt(overlay);
 168        mutex_unlock(&overlay->access_ok);
 169}
 170
 171static void dmafetch_onoff(struct mmp_overlay *overlay, int on)
 172{
 173        u32 mask = overlay_is_vid(overlay) ? CFG_DMA_ENA_MASK :
 174                   CFG_GRA_ENA_MASK;
 175        u32 enable = overlay_is_vid(overlay) ? CFG_DMA_ENA(1) : CFG_GRA_ENA(1);
 176        u32 tmp;
 177        struct mmp_path *path = overlay->path;
 178
 179        mutex_lock(&overlay->access_ok);
 180        tmp = readl_relaxed(ctrl_regs(path) + dma_ctrl(0, path->id));
 181        tmp &= ~mask;
 182        tmp |= (on ? enable : 0);
 183        writel(tmp, ctrl_regs(path) + dma_ctrl(0, path->id));
 184        mutex_unlock(&overlay->access_ok);
 185}
 186
 187static void path_enabledisable(struct mmp_path *path, int on)
 188{
 189        u32 tmp;
 190        mutex_lock(&path->access_ok);
 191        tmp = readl_relaxed(ctrl_regs(path) + LCD_SCLK(path));
 192        if (on)
 193                tmp &= ~SCLK_DISABLE;
 194        else
 195                tmp |= SCLK_DISABLE;
 196        writel_relaxed(tmp, ctrl_regs(path) + LCD_SCLK(path));
 197        mutex_unlock(&path->access_ok);
 198}
 199
 200static void path_onoff(struct mmp_path *path, int on)
 201{
 202        if (path->status == on) {
 203                dev_info(path->dev, "path %s is already %s\n",
 204                                path->name, stat_name(path->status));
 205                return;
 206        }
 207
 208        if (on) {
 209                path_enabledisable(path, 1);
 210
 211                if (path->panel && path->panel->set_onoff)
 212                        path->panel->set_onoff(path->panel, 1);
 213        } else {
 214                if (path->panel && path->panel->set_onoff)
 215                        path->panel->set_onoff(path->panel, 0);
 216
 217                path_enabledisable(path, 0);
 218        }
 219        path->status = on;
 220}
 221
 222static void overlay_set_onoff(struct mmp_overlay *overlay, int on)
 223{
 224        if (overlay->status == on) {
 225                dev_info(overlay_to_ctrl(overlay)->dev, "overlay %s is already %s\n",
 226                        overlay->path->name, stat_name(overlay->status));
 227                return;
 228        }
 229        overlay->status = on;
 230        dmafetch_onoff(overlay, on);
 231        if (overlay->path->ops.check_status(overlay->path)
 232                        != overlay->path->status)
 233                path_onoff(overlay->path, on);
 234}
 235
 236static void overlay_set_fetch(struct mmp_overlay *overlay, int fetch_id)
 237{
 238        overlay->dmafetch_id = fetch_id;
 239}
 240
 241static int overlay_set_addr(struct mmp_overlay *overlay, struct mmp_addr *addr)
 242{
 243        struct lcd_regs *regs = path_regs(overlay->path);
 244
 245        /* FIXME: assert addr supported */
 246        memcpy(&overlay->addr, addr, sizeof(struct mmp_addr));
 247
 248        if (overlay_is_vid(overlay)) {
 249                writel_relaxed(addr->phys[0], &regs->v_y0);
 250                writel_relaxed(addr->phys[1], &regs->v_u0);
 251                writel_relaxed(addr->phys[2], &regs->v_v0);
 252        } else
 253                writel_relaxed(addr->phys[0], &regs->g_0);
 254
 255        return overlay->addr.phys[0];
 256}
 257
 258static void path_set_mode(struct mmp_path *path, struct mmp_mode *mode)
 259{
 260        struct lcd_regs *regs = path_regs(path);
 261        u32 total_x, total_y, vsync_ctrl, tmp, sclk_src, sclk_div,
 262                link_config = path_to_path_plat(path)->link_config,
 263                dsi_rbswap = path_to_path_plat(path)->link_config;
 264
 265        /* FIXME: assert videomode supported */
 266        memcpy(&path->mode, mode, sizeof(struct mmp_mode));
 267
 268        mutex_lock(&path->access_ok);
 269
 270        /* polarity of timing signals */
 271        tmp = readl_relaxed(ctrl_regs(path) + intf_ctrl(path->id)) & 0x1;
 272        tmp |= mode->vsync_invert ? 0 : 0x8;
 273        tmp |= mode->hsync_invert ? 0 : 0x4;
 274        tmp |= link_config & CFG_DUMBMODE_MASK;
 275        tmp |= CFG_DUMB_ENA(1);
 276        writel_relaxed(tmp, ctrl_regs(path) + intf_ctrl(path->id));
 277
 278        /* interface rb_swap setting */
 279        tmp = readl_relaxed(ctrl_regs(path) + intf_rbswap_ctrl(path->id)) &
 280                (~(CFG_INTFRBSWAP_MASK));
 281        tmp |= dsi_rbswap & CFG_INTFRBSWAP_MASK;
 282        writel_relaxed(tmp, ctrl_regs(path) + intf_rbswap_ctrl(path->id));
 283
 284        writel_relaxed((mode->yres << 16) | mode->xres, &regs->screen_active);
 285        writel_relaxed((mode->left_margin << 16) | mode->right_margin,
 286                &regs->screen_h_porch);
 287        writel_relaxed((mode->upper_margin << 16) | mode->lower_margin,
 288                &regs->screen_v_porch);
 289        total_x = mode->xres + mode->left_margin + mode->right_margin +
 290                mode->hsync_len;
 291        total_y = mode->yres + mode->upper_margin + mode->lower_margin +
 292                mode->vsync_len;
 293        writel_relaxed((total_y << 16) | total_x, &regs->screen_size);
 294
 295        /* vsync ctrl */
 296        if (path->output_type == PATH_OUT_DSI)
 297                vsync_ctrl = 0x01330133;
 298        else
 299                vsync_ctrl = ((mode->xres + mode->right_margin) << 16)
 300                                        | (mode->xres + mode->right_margin);
 301        writel_relaxed(vsync_ctrl, &regs->vsync_ctrl);
 302
 303        /* set pixclock div */
 304        sclk_src = clk_get_rate(path_to_ctrl(path)->clk);
 305        sclk_div = sclk_src / mode->pixclock_freq;
 306        if (sclk_div * mode->pixclock_freq < sclk_src)
 307                sclk_div++;
 308
 309        dev_info(path->dev, "%s sclk_src %d sclk_div 0x%x pclk %d\n",
 310                        __func__, sclk_src, sclk_div, mode->pixclock_freq);
 311
 312        tmp = readl_relaxed(ctrl_regs(path) + LCD_SCLK(path));
 313        tmp &= ~CLK_INT_DIV_MASK;
 314        tmp |= sclk_div;
 315        writel_relaxed(tmp, ctrl_regs(path) + LCD_SCLK(path));
 316
 317        mutex_unlock(&path->access_ok);
 318}
 319
 320static struct mmp_overlay_ops mmphw_overlay_ops = {
 321        .set_fetch = overlay_set_fetch,
 322        .set_onoff = overlay_set_onoff,
 323        .set_win = overlay_set_win,
 324        .set_addr = overlay_set_addr,
 325};
 326
 327static void ctrl_set_default(struct mmphw_ctrl *ctrl)
 328{
 329        u32 tmp, irq_mask;
 330
 331        /*
 332         * LCD Global control(LCD_TOP_CTRL) should be configed before
 333         * any other LCD registers read/write, or there maybe issues.
 334         */
 335        tmp = readl_relaxed(ctrl->reg_base + LCD_TOP_CTRL);
 336        tmp |= 0xfff0;
 337        writel_relaxed(tmp, ctrl->reg_base + LCD_TOP_CTRL);
 338
 339
 340        /* disable all interrupts */
 341        irq_mask = path_imasks(0) | err_imask(0) |
 342                   path_imasks(1) | err_imask(1);
 343        tmp = readl_relaxed(ctrl->reg_base + SPU_IRQ_ENA);
 344        tmp &= ~irq_mask;
 345        tmp |= irq_mask;
 346        writel_relaxed(tmp, ctrl->reg_base + SPU_IRQ_ENA);
 347}
 348
 349static void path_set_default(struct mmp_path *path)
 350{
 351        struct lcd_regs *regs = path_regs(path);
 352        u32 dma_ctrl1, mask, tmp, path_config;
 353
 354        path_config = path_to_path_plat(path)->path_config;
 355
 356        /* Configure IOPAD: should be parallel only */
 357        if (PATH_OUT_PARALLEL == path->output_type) {
 358                mask = CFG_IOPADMODE_MASK | CFG_BURST_MASK | CFG_BOUNDARY_MASK;
 359                tmp = readl_relaxed(ctrl_regs(path) + SPU_IOPAD_CONTROL);
 360                tmp &= ~mask;
 361                tmp |= path_config;
 362                writel_relaxed(tmp, ctrl_regs(path) + SPU_IOPAD_CONTROL);
 363        }
 364
 365        /* Select path clock source */
 366        tmp = readl_relaxed(ctrl_regs(path) + LCD_SCLK(path));
 367        tmp &= ~SCLK_SRC_SEL_MASK;
 368        tmp |= path_config;
 369        writel_relaxed(tmp, ctrl_regs(path) + LCD_SCLK(path));
 370
 371        /*
 372         * Configure default bits: vsync triggers DMA,
 373         * power save enable, configure alpha registers to
 374         * display 100% graphics, and set pixel command.
 375         */
 376        dma_ctrl1 = 0x2032ff81;
 377
 378        dma_ctrl1 |= CFG_VSYNC_INV_MASK;
 379        writel_relaxed(dma_ctrl1, ctrl_regs(path) + dma_ctrl(1, path->id));
 380
 381        /* Configure default register values */
 382        writel_relaxed(0x00000000, &regs->blank_color);
 383        writel_relaxed(0x00000000, &regs->g_1);
 384        writel_relaxed(0x00000000, &regs->g_start);
 385
 386        /*
 387         * 1.enable multiple burst request in DMA AXI
 388         * bus arbiter for faster read if not tv path;
 389         * 2.enable horizontal smooth filter;
 390         */
 391        mask = CFG_GRA_HSMOOTH_MASK | CFG_DMA_HSMOOTH_MASK | CFG_ARBFAST_ENA(1);
 392        tmp = readl_relaxed(ctrl_regs(path) + dma_ctrl(0, path->id));
 393        tmp |= mask;
 394        if (PATH_TV == path->id)
 395                tmp &= ~CFG_ARBFAST_ENA(1);
 396        writel_relaxed(tmp, ctrl_regs(path) + dma_ctrl(0, path->id));
 397}
 398
 399static int path_init(struct mmphw_path_plat *path_plat,
 400                struct mmp_mach_path_config *config)
 401{
 402        struct mmphw_ctrl *ctrl = path_plat->ctrl;
 403        struct mmp_path_info *path_info;
 404        struct mmp_path *path = NULL;
 405
 406        dev_info(ctrl->dev, "%s: %s\n", __func__, config->name);
 407
 408        /* init driver data */
 409        path_info = kzalloc(sizeof(struct mmp_path_info), GFP_KERNEL);
 410        if (!path_info) {
 411                dev_err(ctrl->dev, "%s: unable to alloc path_info for %s\n",
 412                                __func__, config->name);
 413                return 0;
 414        }
 415        path_info->name = config->name;
 416        path_info->id = path_plat->id;
 417        path_info->dev = ctrl->dev;
 418        path_info->overlay_num = config->overlay_num;
 419        path_info->overlay_ops = &mmphw_overlay_ops;
 420        path_info->set_mode = path_set_mode;
 421        path_info->plat_data = path_plat;
 422
 423        /* create/register platform device */
 424        path = mmp_register_path(path_info);
 425        if (!path) {
 426                kfree(path_info);
 427                return 0;
 428        }
 429        path_plat->path = path;
 430        path_plat->path_config = config->path_config;
 431        path_plat->link_config = config->link_config;
 432        path_plat->dsi_rbswap = config->dsi_rbswap;
 433        path_set_default(path);
 434
 435        kfree(path_info);
 436        return 1;
 437}
 438
 439static void path_deinit(struct mmphw_path_plat *path_plat)
 440{
 441        if (!path_plat)
 442                return;
 443
 444        mmp_unregister_path(path_plat->path);
 445}
 446
 447static int mmphw_probe(struct platform_device *pdev)
 448{
 449        struct mmp_mach_plat_info *mi;
 450        struct resource *res;
 451        int ret, i, size, irq;
 452        struct mmphw_path_plat *path_plat;
 453        struct mmphw_ctrl *ctrl = NULL;
 454
 455        /* get resources from platform data */
 456        res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
 457        if (res == NULL) {
 458                dev_err(&pdev->dev, "%s: no IO memory defined\n", __func__);
 459                ret = -ENOENT;
 460                goto failed;
 461        }
 462
 463        irq = platform_get_irq(pdev, 0);
 464        if (irq < 0) {
 465                dev_err(&pdev->dev, "%s: no IRQ defined\n", __func__);
 466                ret = -ENOENT;
 467                goto failed;
 468        }
 469
 470        /* get configs from platform data */
 471        mi = pdev->dev.platform_data;
 472        if (mi == NULL || !mi->path_num || !mi->paths) {
 473                dev_err(&pdev->dev, "%s: no platform data defined\n", __func__);
 474                ret = -EINVAL;
 475                goto failed;
 476        }
 477
 478        /* allocate */
 479        size = sizeof(struct mmphw_ctrl) + sizeof(struct mmphw_path_plat) *
 480               mi->path_num;
 481        ctrl = devm_kzalloc(&pdev->dev, size, GFP_KERNEL);
 482        if (!ctrl) {
 483                ret = -ENOMEM;
 484                goto failed;
 485        }
 486
 487        ctrl->name = mi->name;
 488        ctrl->path_num = mi->path_num;
 489        ctrl->dev = &pdev->dev;
 490        ctrl->irq = irq;
 491        platform_set_drvdata(pdev, ctrl);
 492        mutex_init(&ctrl->access_ok);
 493
 494        /* map registers.*/
 495        if (!devm_request_mem_region(ctrl->dev, res->start,
 496                        resource_size(res), ctrl->name)) {
 497                dev_err(ctrl->dev,
 498                        "can't request region for resource %pR\n", res);
 499                ret = -EINVAL;
 500                goto failed;
 501        }
 502
 503        ctrl->reg_base = devm_ioremap_nocache(ctrl->dev,
 504                        res->start, resource_size(res));
 505        if (ctrl->reg_base == NULL) {
 506                dev_err(ctrl->dev, "%s: res %pR map failed\n", __func__, res);
 507                ret = -ENOMEM;
 508                goto failed;
 509        }
 510
 511        /* request irq */
 512        ret = devm_request_irq(ctrl->dev, ctrl->irq, ctrl_handle_irq,
 513                IRQF_SHARED, "lcd_controller", ctrl);
 514        if (ret < 0) {
 515                dev_err(ctrl->dev, "%s unable to request IRQ %d\n",
 516                                __func__, ctrl->irq);
 517                ret = -ENXIO;
 518                goto failed;
 519        }
 520
 521        /* get clock */
 522        ctrl->clk = devm_clk_get(ctrl->dev, mi->clk_name);
 523        if (IS_ERR(ctrl->clk)) {
 524                dev_err(ctrl->dev, "unable to get clk %s\n", mi->clk_name);
 525                ret = -ENOENT;
 526                goto failed;
 527        }
 528        clk_prepare_enable(ctrl->clk);
 529
 530        /* init global regs */
 531        ctrl_set_default(ctrl);
 532
 533        /* init pathes from machine info and register them */
 534        for (i = 0; i < ctrl->path_num; i++) {
 535                /* get from config and machine info */
 536                path_plat = &ctrl->path_plats[i];
 537                path_plat->id = i;
 538                path_plat->ctrl = ctrl;
 539
 540                /* path init */
 541                if (!path_init(path_plat, &mi->paths[i])) {
 542                        ret = -EINVAL;
 543                        goto failed_path_init;
 544                }
 545        }
 546
 547#ifdef CONFIG_MMP_DISP_SPI
 548        ret = lcd_spi_register(ctrl);
 549        if (ret < 0)
 550                goto failed_path_init;
 551#endif
 552
 553        dev_info(ctrl->dev, "device init done\n");
 554
 555        return 0;
 556
 557failed_path_init:
 558        for (i = 0; i < ctrl->path_num; i++) {
 559                path_plat = &ctrl->path_plats[i];
 560                path_deinit(path_plat);
 561        }
 562
 563        clk_disable_unprepare(ctrl->clk);
 564failed:
 565        dev_err(&pdev->dev, "device init failed\n");
 566
 567        return ret;
 568}
 569
 570static struct platform_driver mmphw_driver = {
 571        .driver         = {
 572                .name   = "mmp-disp",
 573        },
 574        .probe          = mmphw_probe,
 575};
 576
 577static int mmphw_init(void)
 578{
 579        return platform_driver_register(&mmphw_driver);
 580}
 581module_init(mmphw_init);
 582
 583MODULE_AUTHOR("Li Guoqing<ligq@marvell.com>");
 584MODULE_DESCRIPTION("Framebuffer driver for mmp");
 585MODULE_LICENSE("GPL");
 586