linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/gv100.c
<<
>>
Prefs
   1/*
   2 * Copyright 2018 Red Hat Inc.
   3 *
   4 * Permission is hereby granted, free of charge, to any person obtaining a
   5 * copy of this software and associated documentation files (the "Software"),
   6 * to deal in the Software without restriction, including without limitation
   7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
   8 * and/or sell copies of the Software, and to permit persons to whom the
   9 * Software is furnished to do so, subject to the following conditions:
  10 *
  11 * The above copyright notice and this permission notice shall be included in
  12 * all copies or substantial portions of the Software.
  13 *
  14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
  17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
  18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
  19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
  20 * OTHER DEALINGS IN THE SOFTWARE.
  21 */
  22#include "nv50.h"
  23#include "head.h"
  24#include "ior.h"
  25#include "channv50.h"
  26#include "rootnv50.h"
  27
  28#include <core/gpuobj.h>
  29#include <subdev/timer.h>
  30
  31int
  32gv100_disp_wndw_cnt(struct nvkm_disp *disp, unsigned long *pmask)
  33{
  34        struct nvkm_device *device = disp->engine.subdev.device;
  35        *pmask = nvkm_rd32(device, 0x610064);
  36        return (nvkm_rd32(device, 0x610074) & 0x03f00000) >> 20;
  37}
  38
  39void
  40gv100_disp_super(struct work_struct *work)
  41{
  42        struct nv50_disp *disp =
  43                container_of(work, struct nv50_disp, supervisor);
  44        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
  45        struct nvkm_device *device = subdev->device;
  46        struct nvkm_head *head;
  47        u32 stat = nvkm_rd32(device, 0x6107a8);
  48        u32 mask[4];
  49
  50        nvkm_debug(subdev, "supervisor %d: %08x\n", ffs(disp->super), stat);
  51        list_for_each_entry(head, &disp->base.head, head) {
  52                mask[head->id] = nvkm_rd32(device, 0x6107ac + (head->id * 4));
  53                HEAD_DBG(head, "%08x", mask[head->id]);
  54        }
  55
  56        if (disp->super & 0x00000001) {
  57                nv50_disp_chan_mthd(disp->chan[0], NV_DBG_DEBUG);
  58                nv50_disp_super_1(disp);
  59                list_for_each_entry(head, &disp->base.head, head) {
  60                        if (!(mask[head->id] & 0x00001000))
  61                                continue;
  62                        nv50_disp_super_1_0(disp, head);
  63                }
  64        } else
  65        if (disp->super & 0x00000002) {
  66                list_for_each_entry(head, &disp->base.head, head) {
  67                        if (!(mask[head->id] & 0x00001000))
  68                                continue;
  69                        nv50_disp_super_2_0(disp, head);
  70                }
  71                nvkm_outp_route(&disp->base);
  72                list_for_each_entry(head, &disp->base.head, head) {
  73                        if (!(mask[head->id] & 0x00010000))
  74                                continue;
  75                        nv50_disp_super_2_1(disp, head);
  76                }
  77                list_for_each_entry(head, &disp->base.head, head) {
  78                        if (!(mask[head->id] & 0x00001000))
  79                                continue;
  80                        nv50_disp_super_2_2(disp, head);
  81                }
  82        } else
  83        if (disp->super & 0x00000004) {
  84                list_for_each_entry(head, &disp->base.head, head) {
  85                        if (!(mask[head->id] & 0x00001000))
  86                                continue;
  87                        nv50_disp_super_3_0(disp, head);
  88                }
  89        }
  90
  91        list_for_each_entry(head, &disp->base.head, head)
  92                nvkm_wr32(device, 0x6107ac + (head->id * 4), 0x00000000);
  93        nvkm_wr32(device, 0x6107a8, 0x80000000);
  94}
  95
  96static void
  97gv100_disp_exception(struct nv50_disp *disp, int chid)
  98{
  99        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 100        struct nvkm_device *device = subdev->device;
 101        u32 stat = nvkm_rd32(device, 0x611020 + (chid * 12));
 102        u32 type = (stat & 0x00007000) >> 12;
 103        u32 mthd = (stat & 0x00000fff) << 2;
 104        const struct nvkm_enum *reason =
 105                nvkm_enum_find(nv50_disp_intr_error_type, type);
 106
 107        /*TODO: Suspect 33->41 are for WRBK channel exceptions, but we
 108         *      don't support those currently.
 109         *
 110         *      CORE+WIN CHIDs map directly to the FE_EXCEPT() slots.
 111         */
 112        if (chid <= 32) {
 113                u32 data = nvkm_rd32(device, 0x611024 + (chid * 12));
 114                u32 code = nvkm_rd32(device, 0x611028 + (chid * 12));
 115                nvkm_error(subdev, "chid %d stat %08x reason %d [%s] "
 116                                   "mthd %04x data %08x code %08x\n",
 117                           chid, stat, type, reason ? reason->name : "",
 118                           mthd, data, code);
 119        } else {
 120                nvkm_error(subdev, "chid %d stat %08x reason %d [%s] "
 121                                   "mthd %04x\n",
 122                           chid, stat, type, reason ? reason->name : "", mthd);
 123        }
 124
 125        if (chid < ARRAY_SIZE(disp->chan) && disp->chan[chid]) {
 126                switch (mthd) {
 127                case 0x0200:
 128                        nv50_disp_chan_mthd(disp->chan[chid], NV_DBG_ERROR);
 129                        break;
 130                default:
 131                        break;
 132                }
 133        }
 134
 135        nvkm_wr32(device, 0x611020 + (chid * 12), 0x90000000);
 136}
 137
 138static void
 139gv100_disp_intr_ctrl_disp(struct nv50_disp *disp)
 140{
 141        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 142        struct nvkm_device *device = subdev->device;
 143        u32 stat = nvkm_rd32(device, 0x611c30);
 144
 145        if (stat & 0x00000007) {
 146                disp->super = (stat & 0x00000007);
 147                queue_work(disp->wq, &disp->supervisor);
 148                nvkm_wr32(device, 0x611860, disp->super);
 149                stat &= ~0x00000007;
 150        }
 151
 152        /*TODO: I would guess this is VBIOS_RELEASE, however, NFI how to
 153         *      ACK it, nor does RM appear to bother.
 154         */
 155        if (stat & 0x00000008)
 156                stat &= ~0x00000008;
 157
 158        if (stat & 0x00000080) {
 159                u32 error = nvkm_mask(device, 0x611848, 0x00000000, 0x00000000);
 160                nvkm_warn(subdev, "error %08x\n", error);
 161                stat &= ~0x00000080;
 162        }
 163
 164        if (stat & 0x00000100) {
 165                unsigned long wndws = nvkm_rd32(device, 0x611858);
 166                unsigned long other = nvkm_rd32(device, 0x61185c);
 167                int wndw;
 168
 169                nvkm_wr32(device, 0x611858, wndws);
 170                nvkm_wr32(device, 0x61185c, other);
 171
 172                /* AWAKEN_OTHER_CORE. */
 173                if (other & 0x00000001)
 174                        nv50_disp_chan_uevent_send(disp, 0);
 175
 176                /* AWAKEN_WIN_CH(n). */
 177                for_each_set_bit(wndw, &wndws, disp->wndw.nr) {
 178                        nv50_disp_chan_uevent_send(disp, 1 + wndw);
 179                }
 180        }
 181
 182        if (stat)
 183                nvkm_warn(subdev, "ctrl %08x\n", stat);
 184}
 185
 186static void
 187gv100_disp_intr_exc_other(struct nv50_disp *disp)
 188{
 189        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 190        struct nvkm_device *device = subdev->device;
 191        u32 stat = nvkm_rd32(device, 0x611854);
 192        unsigned long mask;
 193        int head;
 194
 195        if (stat & 0x00000001) {
 196                nvkm_wr32(device, 0x611854, 0x00000001);
 197                gv100_disp_exception(disp, 0);
 198                stat &= ~0x00000001;
 199        }
 200
 201        if ((mask = (stat & 0x00ff0000) >> 16)) {
 202                for_each_set_bit(head, &mask, disp->wndw.nr) {
 203                        nvkm_wr32(device, 0x611854, 0x00010000 << head);
 204                        gv100_disp_exception(disp, 73 + head);
 205                        stat &= ~(0x00010000 << head);
 206                }
 207        }
 208
 209        if (stat) {
 210                nvkm_warn(subdev, "exception %08x\n", stat);
 211                nvkm_wr32(device, 0x611854, stat);
 212        }
 213}
 214
 215static void
 216gv100_disp_intr_exc_winim(struct nv50_disp *disp)
 217{
 218        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 219        struct nvkm_device *device = subdev->device;
 220        unsigned long stat = nvkm_rd32(device, 0x611850);
 221        int wndw;
 222
 223        for_each_set_bit(wndw, &stat, disp->wndw.nr) {
 224                nvkm_wr32(device, 0x611850, BIT(wndw));
 225                gv100_disp_exception(disp, 33 + wndw);
 226                stat &= ~BIT(wndw);
 227        }
 228
 229        if (stat) {
 230                nvkm_warn(subdev, "wimm %08x\n", (u32)stat);
 231                nvkm_wr32(device, 0x611850, stat);
 232        }
 233}
 234
 235static void
 236gv100_disp_intr_exc_win(struct nv50_disp *disp)
 237{
 238        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 239        struct nvkm_device *device = subdev->device;
 240        unsigned long stat = nvkm_rd32(device, 0x61184c);
 241        int wndw;
 242
 243        for_each_set_bit(wndw, &stat, disp->wndw.nr) {
 244                nvkm_wr32(device, 0x61184c, BIT(wndw));
 245                gv100_disp_exception(disp, 1 + wndw);
 246                stat &= ~BIT(wndw);
 247        }
 248
 249        if (stat) {
 250                nvkm_warn(subdev, "wndw %08x\n", (u32)stat);
 251                nvkm_wr32(device, 0x61184c, stat);
 252        }
 253}
 254
 255static void
 256gv100_disp_intr_head_timing(struct nv50_disp *disp, int head)
 257{
 258        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 259        struct nvkm_device *device = subdev->device;
 260        u32 stat = nvkm_rd32(device, 0x611800 + (head * 0x04));
 261
 262        /* LAST_DATA, LOADV. */
 263        if (stat & 0x00000003) {
 264                nvkm_wr32(device, 0x611800 + (head * 0x04), stat & 0x00000003);
 265                stat &= ~0x00000003;
 266        }
 267
 268        if (stat & 0x00000004) {
 269                nvkm_disp_vblank(&disp->base, head);
 270                nvkm_wr32(device, 0x611800 + (head * 0x04), 0x00000004);
 271                stat &= ~0x00000004;
 272        }
 273
 274        if (stat) {
 275                nvkm_warn(subdev, "head %08x\n", stat);
 276                nvkm_wr32(device, 0x611800 + (head * 0x04), stat);
 277        }
 278}
 279
 280void
 281gv100_disp_intr(struct nv50_disp *disp)
 282{
 283        struct nvkm_subdev *subdev = &disp->base.engine.subdev;
 284        struct nvkm_device *device = subdev->device;
 285        u32 stat = nvkm_rd32(device, 0x611ec0);
 286        unsigned long mask;
 287        int head;
 288
 289        if ((mask = (stat & 0x000000ff))) {
 290                for_each_set_bit(head, &mask, 8) {
 291                        gv100_disp_intr_head_timing(disp, head);
 292                        stat &= ~BIT(head);
 293                }
 294        }
 295
 296        if (stat & 0x00000200) {
 297                gv100_disp_intr_exc_win(disp);
 298                stat &= ~0x00000200;
 299        }
 300
 301        if (stat & 0x00000400) {
 302                gv100_disp_intr_exc_winim(disp);
 303                stat &= ~0x00000400;
 304        }
 305
 306        if (stat & 0x00000800) {
 307                gv100_disp_intr_exc_other(disp);
 308                stat &= ~0x00000800;
 309        }
 310
 311        if (stat & 0x00001000) {
 312                gv100_disp_intr_ctrl_disp(disp);
 313                stat &= ~0x00001000;
 314        }
 315
 316        if (stat)
 317                nvkm_warn(subdev, "intr %08x\n", stat);
 318}
 319
 320void
 321gv100_disp_fini(struct nv50_disp *disp)
 322{
 323        struct nvkm_device *device = disp->base.engine.subdev.device;
 324        nvkm_wr32(device, 0x611db0, 0x00000000);
 325}
 326
 327static int
 328gv100_disp_init(struct nv50_disp *disp)
 329{
 330        struct nvkm_device *device = disp->base.engine.subdev.device;
 331        struct nvkm_head *head;
 332        int i, j;
 333        u32 tmp;
 334
 335        /* Claim ownership of display. */
 336        if (nvkm_rd32(device, 0x6254e8) & 0x00000002) {
 337                nvkm_mask(device, 0x6254e8, 0x00000001, 0x00000000);
 338                if (nvkm_msec(device, 2000,
 339                        if (!(nvkm_rd32(device, 0x6254e8) & 0x00000002))
 340                                break;
 341                ) < 0)
 342                        return -EBUSY;
 343        }
 344
 345        /* Lock pin capabilities. */
 346        tmp = nvkm_rd32(device, 0x610068);
 347        nvkm_wr32(device, 0x640008, tmp);
 348
 349        /* SOR capabilities. */
 350        for (i = 0; i < disp->sor.nr; i++) {
 351                tmp = nvkm_rd32(device, 0x61c000 + (i * 0x800));
 352                nvkm_mask(device, 0x640000, 0x00000100 << i, 0x00000100 << i);
 353                nvkm_wr32(device, 0x640144 + (i * 0x08), tmp);
 354        }
 355
 356        /* Head capabilities. */
 357        list_for_each_entry(head, &disp->base.head, head) {
 358                const int id = head->id;
 359
 360                /* RG. */
 361                tmp = nvkm_rd32(device, 0x616300 + (id * 0x800));
 362                nvkm_wr32(device, 0x640048 + (id * 0x020), tmp);
 363
 364                /* POSTCOMP. */
 365                for (j = 0; j < 6 * 4; j += 4) {
 366                        tmp = nvkm_rd32(device, 0x616100 + (id * 0x800) + j);
 367                        nvkm_wr32(device, 0x640030 + (id * 0x20) + j, tmp);
 368                }
 369        }
 370
 371        /* Window capabilities. */
 372        for (i = 0; i < disp->wndw.nr; i++) {
 373                nvkm_mask(device, 0x640004, 1 << i, 1 << i);
 374                for (j = 0; j < 6 * 4; j += 4) {
 375                        tmp = nvkm_rd32(device, 0x630050 + (i * 0x800) + j);
 376                        nvkm_wr32(device, 0x6401e4 + (i * 0x20) + j, tmp);
 377                }
 378        }
 379
 380        /* IHUB capabilities. */
 381        for (i = 0; i < 4; i++) {
 382                tmp = nvkm_rd32(device, 0x62e000 + (i * 0x04));
 383                nvkm_wr32(device, 0x640010 + (i * 0x04), tmp);
 384        }
 385
 386        nvkm_mask(device, 0x610078, 0x00000001, 0x00000001);
 387
 388        /* Setup instance memory. */
 389        switch (nvkm_memory_target(disp->inst->memory)) {
 390        case NVKM_MEM_TARGET_VRAM: tmp = 0x00000001; break;
 391        case NVKM_MEM_TARGET_NCOH: tmp = 0x00000002; break;
 392        case NVKM_MEM_TARGET_HOST: tmp = 0x00000003; break;
 393        default:
 394                break;
 395        }
 396        nvkm_wr32(device, 0x610010, 0x00000008 | tmp);
 397        nvkm_wr32(device, 0x610014, disp->inst->addr >> 16);
 398
 399        /* CTRL_DISP: AWAKEN, ERROR, SUPERVISOR[1-3]. */
 400        nvkm_wr32(device, 0x611cf0, 0x00000187); /* MSK. */
 401        nvkm_wr32(device, 0x611db0, 0x00000187); /* EN. */
 402
 403        /* EXC_OTHER: CURSn, CORE. */
 404        nvkm_wr32(device, 0x611cec, disp->head.mask << 16 |
 405                                    0x00000001); /* MSK. */
 406        nvkm_wr32(device, 0x611dac, 0x00000000); /* EN. */
 407
 408        /* EXC_WINIM. */
 409        nvkm_wr32(device, 0x611ce8, disp->wndw.mask); /* MSK. */
 410        nvkm_wr32(device, 0x611da8, 0x00000000); /* EN. */
 411
 412        /* EXC_WIN. */
 413        nvkm_wr32(device, 0x611ce4, disp->wndw.mask); /* MSK. */
 414        nvkm_wr32(device, 0x611da4, 0x00000000); /* EN. */
 415
 416        /* HEAD_TIMING(n): VBLANK. */
 417        list_for_each_entry(head, &disp->base.head, head) {
 418                const u32 hoff = head->id * 4;
 419                nvkm_wr32(device, 0x611cc0 + hoff, 0x00000004); /* MSK. */
 420                nvkm_wr32(device, 0x611d80 + hoff, 0x00000000); /* EN. */
 421        }
 422
 423        /* OR. */
 424        nvkm_wr32(device, 0x611cf4, 0x00000000); /* MSK. */
 425        nvkm_wr32(device, 0x611db4, 0x00000000); /* EN. */
 426        return 0;
 427}
 428
 429static const struct nv50_disp_func
 430gv100_disp = {
 431        .init = gv100_disp_init,
 432        .fini = gv100_disp_fini,
 433        .intr = gv100_disp_intr,
 434        .uevent = &gv100_disp_chan_uevent,
 435        .super = gv100_disp_super,
 436        .root = &gv100_disp_root_oclass,
 437        .wndw = { .cnt = gv100_disp_wndw_cnt },
 438        .head = { .cnt = gv100_head_cnt, .new = gv100_head_new },
 439        .sor = { .cnt = gv100_sor_cnt, .new = gv100_sor_new },
 440        .ramht_size = 0x2000,
 441};
 442
 443int
 444gv100_disp_new(struct nvkm_device *device, int index, struct nvkm_disp **pdisp)
 445{
 446        return nv50_disp_new_(&gv100_disp, device, index, pdisp);
 447}
 448