linux/drivers/memory/tegra/tegra210-emc-cc-r21021.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2014-2020, NVIDIA CORPORATION.  All rights reserved.
   4 */
   5
   6#include <linux/kernel.h>
   7#include <linux/io.h>
   8#include <linux/clk.h>
   9#include <linux/delay.h>
  10#include <linux/of.h>
  11
  12#include <soc/tegra/mc.h>
  13
  14#include "tegra210-emc.h"
  15#include "tegra210-mc.h"
  16
  17/*
  18 * Enable flags for specifying verbosity.
  19 */
  20#define INFO            (1 << 0)
  21#define STEPS           (1 << 1)
  22#define SUB_STEPS       (1 << 2)
  23#define PRELOCK         (1 << 3)
  24#define PRELOCK_STEPS   (1 << 4)
  25#define ACTIVE_EN       (1 << 5)
  26#define PRAMP_UP        (1 << 6)
  27#define PRAMP_DN        (1 << 7)
  28#define EMA_WRITES      (1 << 10)
  29#define EMA_UPDATES     (1 << 11)
  30#define PER_TRAIN       (1 << 16)
  31#define CC_PRINT        (1 << 17)
  32#define CCFIFO          (1 << 29)
  33#define REGS            (1 << 30)
  34#define REG_LISTS       (1 << 31)
  35
  36#define emc_dbg(emc, flags, ...) dev_dbg(emc->dev, __VA_ARGS__)
  37
  38#define DVFS_CLOCK_CHANGE_VERSION       21021
  39#define EMC_PRELOCK_VERSION             2101
  40
  41enum {
  42        DVFS_SEQUENCE = 1,
  43        WRITE_TRAINING_SEQUENCE = 2,
  44        PERIODIC_TRAINING_SEQUENCE = 3,
  45        DVFS_PT1 = 10,
  46        DVFS_UPDATE = 11,
  47        TRAINING_PT1 = 12,
  48        TRAINING_UPDATE = 13,
  49        PERIODIC_TRAINING_UPDATE = 14
  50};
  51
  52/*
  53 * PTFV defines - basically just indexes into the per table PTFV array.
  54 */
  55#define PTFV_DQSOSC_MOVAVG_C0D0U0_INDEX         0
  56#define PTFV_DQSOSC_MOVAVG_C0D0U1_INDEX         1
  57#define PTFV_DQSOSC_MOVAVG_C0D1U0_INDEX         2
  58#define PTFV_DQSOSC_MOVAVG_C0D1U1_INDEX         3
  59#define PTFV_DQSOSC_MOVAVG_C1D0U0_INDEX         4
  60#define PTFV_DQSOSC_MOVAVG_C1D0U1_INDEX         5
  61#define PTFV_DQSOSC_MOVAVG_C1D1U0_INDEX         6
  62#define PTFV_DQSOSC_MOVAVG_C1D1U1_INDEX         7
  63#define PTFV_DVFS_SAMPLES_INDEX                 9
  64#define PTFV_MOVAVG_WEIGHT_INDEX                10
  65#define PTFV_CONFIG_CTRL_INDEX                  11
  66
  67#define PTFV_CONFIG_CTRL_USE_PREVIOUS_EMA       (1 << 0)
  68
  69/*
  70 * Do arithmetic in fixed point.
  71 */
  72#define MOVAVG_PRECISION_FACTOR         100
  73
  74/*
  75 * The division portion of the average operation.
  76 */
  77#define __AVERAGE_PTFV(dev)                                             \
  78        ({ next->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] =      \
  79           next->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] /      \
  80           next->ptfv_list[PTFV_DVFS_SAMPLES_INDEX]; })
  81
  82/*
  83 * Convert val to fixed point and add it to the temporary average.
  84 */
  85#define __INCREMENT_PTFV(dev, val)                                      \
  86        ({ next->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] +=     \
  87           ((val) * MOVAVG_PRECISION_FACTOR); })
  88
  89/*
  90 * Convert a moving average back to integral form and return the value.
  91 */
  92#define __MOVAVG_AC(timing, dev)                                        \
  93        ((timing)->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX] /    \
  94         MOVAVG_PRECISION_FACTOR)
  95
  96/* Weighted update. */
  97#define __WEIGHTED_UPDATE_PTFV(dev, nval)                               \
  98        do {                                                            \
  99                int w = PTFV_MOVAVG_WEIGHT_INDEX;                       \
 100                int dqs = PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX;         \
 101                                                                        \
 102                next->ptfv_list[dqs] =                                  \
 103                        ((nval * MOVAVG_PRECISION_FACTOR) +             \
 104                         (next->ptfv_list[dqs] *                        \
 105                          next->ptfv_list[w])) /                        \
 106                        (next->ptfv_list[w] + 1);                       \
 107                                                                        \
 108                emc_dbg(emc, EMA_UPDATES, "%s: (s=%lu) EMA: %u\n",      \
 109                        __stringify(dev), nval, next->ptfv_list[dqs]);  \
 110        } while (0)
 111
 112/* Access a particular average. */
 113#define __MOVAVG(timing, dev)                      \
 114        ((timing)->ptfv_list[PTFV_DQSOSC_MOVAVG_ ## dev ## _INDEX])
 115
 116static u32 update_clock_tree_delay(struct tegra210_emc *emc, int type)
 117{
 118        bool periodic_training_update = type == PERIODIC_TRAINING_UPDATE;
 119        struct tegra210_emc_timing *last = emc->last;
 120        struct tegra210_emc_timing *next = emc->next;
 121        u32 last_timing_rate_mhz = last->rate / 1000;
 122        u32 next_timing_rate_mhz = next->rate / 1000;
 123        bool dvfs_update = type == DVFS_UPDATE;
 124        s32 tdel = 0, tmdel = 0, adel = 0;
 125        bool dvfs_pt1 = type == DVFS_PT1;
 126        unsigned long cval = 0;
 127        u32 temp[2][2], value;
 128        unsigned int i;
 129
 130        /*
 131         * Dev0 MSB.
 132         */
 133        if (dvfs_pt1 || periodic_training_update) {
 134                value = tegra210_emc_mrr_read(emc, 2, 19);
 135
 136                for (i = 0; i < emc->num_channels; i++) {
 137                        temp[i][0] = (value & 0x00ff) << 8;
 138                        temp[i][1] = (value & 0xff00) << 0;
 139                        value >>= 16;
 140                }
 141
 142                /*
 143                 * Dev0 LSB.
 144                 */
 145                value = tegra210_emc_mrr_read(emc, 2, 18);
 146
 147                for (i = 0; i < emc->num_channels; i++) {
 148                        temp[i][0] |= (value & 0x00ff) >> 0;
 149                        temp[i][1] |= (value & 0xff00) >> 8;
 150                        value >>= 16;
 151                }
 152        }
 153
 154        if (dvfs_pt1 || periodic_training_update) {
 155                cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 156                cval *= 1000000;
 157                cval /= last_timing_rate_mhz * 2 * temp[0][0];
 158        }
 159
 160        if (dvfs_pt1)
 161                __INCREMENT_PTFV(C0D0U0, cval);
 162        else if (dvfs_update)
 163                __AVERAGE_PTFV(C0D0U0);
 164        else if (periodic_training_update)
 165                __WEIGHTED_UPDATE_PTFV(C0D0U0, cval);
 166
 167        if (dvfs_update || periodic_training_update) {
 168                tdel = next->current_dram_clktree[C0D0U0] -
 169                                __MOVAVG_AC(next, C0D0U0);
 170                tmdel = (tdel < 0) ? -1 * tdel : tdel;
 171                adel = tmdel;
 172
 173                if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 174                    next->tree_margin)
 175                        next->current_dram_clktree[C0D0U0] =
 176                                __MOVAVG_AC(next, C0D0U0);
 177        }
 178
 179        if (dvfs_pt1 || periodic_training_update) {
 180                cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 181                cval *= 1000000;
 182                cval /= last_timing_rate_mhz * 2 * temp[0][1];
 183        }
 184
 185        if (dvfs_pt1)
 186                __INCREMENT_PTFV(C0D0U1, cval);
 187        else if (dvfs_update)
 188                __AVERAGE_PTFV(C0D0U1);
 189        else if (periodic_training_update)
 190                __WEIGHTED_UPDATE_PTFV(C0D0U1, cval);
 191
 192        if (dvfs_update || periodic_training_update) {
 193                tdel = next->current_dram_clktree[C0D0U1] -
 194                                __MOVAVG_AC(next, C0D0U1);
 195                tmdel = (tdel < 0) ? -1 * tdel : tdel;
 196
 197                if (tmdel > adel)
 198                        adel = tmdel;
 199
 200                if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 201                    next->tree_margin)
 202                        next->current_dram_clktree[C0D0U1] =
 203                                __MOVAVG_AC(next, C0D0U1);
 204        }
 205
 206        if (emc->num_channels > 1) {
 207                if (dvfs_pt1 || periodic_training_update) {
 208                        cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 209                        cval *= 1000000;
 210                        cval /= last_timing_rate_mhz * 2 * temp[1][0];
 211                }
 212
 213                if (dvfs_pt1)
 214                        __INCREMENT_PTFV(C1D0U0, cval);
 215                else if (dvfs_update)
 216                        __AVERAGE_PTFV(C1D0U0);
 217                else if (periodic_training_update)
 218                        __WEIGHTED_UPDATE_PTFV(C1D0U0, cval);
 219
 220                if (dvfs_update || periodic_training_update) {
 221                        tdel = next->current_dram_clktree[C1D0U0] -
 222                                        __MOVAVG_AC(next, C1D0U0);
 223                        tmdel = (tdel < 0) ? -1 * tdel : tdel;
 224
 225                        if (tmdel > adel)
 226                                adel = tmdel;
 227
 228                        if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 229                            next->tree_margin)
 230                                next->current_dram_clktree[C1D0U0] =
 231                                        __MOVAVG_AC(next, C1D0U0);
 232                }
 233
 234                if (dvfs_pt1 || periodic_training_update) {
 235                        cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 236                        cval *= 1000000;
 237                        cval /= last_timing_rate_mhz * 2 * temp[1][1];
 238                }
 239
 240                if (dvfs_pt1)
 241                        __INCREMENT_PTFV(C1D0U1, cval);
 242                else if (dvfs_update)
 243                        __AVERAGE_PTFV(C1D0U1);
 244                else if (periodic_training_update)
 245                        __WEIGHTED_UPDATE_PTFV(C1D0U1, cval);
 246
 247                if (dvfs_update || periodic_training_update) {
 248                        tdel = next->current_dram_clktree[C1D0U1] -
 249                                        __MOVAVG_AC(next, C1D0U1);
 250                        tmdel = (tdel < 0) ? -1 * tdel : tdel;
 251
 252                        if (tmdel > adel)
 253                                adel = tmdel;
 254
 255                        if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 256                            next->tree_margin)
 257                                next->current_dram_clktree[C1D0U1] =
 258                                        __MOVAVG_AC(next, C1D0U1);
 259                }
 260        }
 261
 262        if (emc->num_devices < 2)
 263                goto done;
 264
 265        /*
 266         * Dev1 MSB.
 267         */
 268        if (dvfs_pt1 || periodic_training_update) {
 269                value = tegra210_emc_mrr_read(emc, 1, 19);
 270
 271                for (i = 0; i < emc->num_channels; i++) {
 272                        temp[i][0] = (value & 0x00ff) << 8;
 273                        temp[i][1] = (value & 0xff00) << 0;
 274                        value >>= 16;
 275                }
 276
 277                /*
 278                 * Dev1 LSB.
 279                 */
 280                value = tegra210_emc_mrr_read(emc, 2, 18);
 281
 282                for (i = 0; i < emc->num_channels; i++) {
 283                        temp[i][0] |= (value & 0x00ff) >> 0;
 284                        temp[i][1] |= (value & 0xff00) >> 8;
 285                        value >>= 16;
 286                }
 287        }
 288
 289        if (dvfs_pt1 || periodic_training_update) {
 290                cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 291                cval *= 1000000;
 292                cval /= last_timing_rate_mhz * 2 * temp[0][0];
 293        }
 294
 295        if (dvfs_pt1)
 296                __INCREMENT_PTFV(C0D1U0, cval);
 297        else if (dvfs_update)
 298                __AVERAGE_PTFV(C0D1U0);
 299        else if (periodic_training_update)
 300                __WEIGHTED_UPDATE_PTFV(C0D1U0, cval);
 301
 302        if (dvfs_update || periodic_training_update) {
 303                tdel = next->current_dram_clktree[C0D1U0] -
 304                                __MOVAVG_AC(next, C0D1U0);
 305                tmdel = (tdel < 0) ? -1 * tdel : tdel;
 306
 307                if (tmdel > adel)
 308                        adel = tmdel;
 309
 310                if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 311                    next->tree_margin)
 312                        next->current_dram_clktree[C0D1U0] =
 313                                __MOVAVG_AC(next, C0D1U0);
 314        }
 315
 316        if (dvfs_pt1 || periodic_training_update) {
 317                cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 318                cval *= 1000000;
 319                cval /= last_timing_rate_mhz * 2 * temp[0][1];
 320        }
 321
 322        if (dvfs_pt1)
 323                __INCREMENT_PTFV(C0D1U1, cval);
 324        else if (dvfs_update)
 325                __AVERAGE_PTFV(C0D1U1);
 326        else if (periodic_training_update)
 327                __WEIGHTED_UPDATE_PTFV(C0D1U1, cval);
 328
 329        if (dvfs_update || periodic_training_update) {
 330                tdel = next->current_dram_clktree[C0D1U1] -
 331                                __MOVAVG_AC(next, C0D1U1);
 332                tmdel = (tdel < 0) ? -1 * tdel : tdel;
 333
 334                if (tmdel > adel)
 335                        adel = tmdel;
 336
 337                if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 338                    next->tree_margin)
 339                        next->current_dram_clktree[C0D1U1] =
 340                                __MOVAVG_AC(next, C0D1U1);
 341        }
 342
 343        if (emc->num_channels > 1) {
 344                if (dvfs_pt1 || periodic_training_update) {
 345                        cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 346                        cval *= 1000000;
 347                        cval /= last_timing_rate_mhz * 2 * temp[1][0];
 348                }
 349
 350                if (dvfs_pt1)
 351                        __INCREMENT_PTFV(C1D1U0, cval);
 352                else if (dvfs_update)
 353                        __AVERAGE_PTFV(C1D1U0);
 354                else if (periodic_training_update)
 355                        __WEIGHTED_UPDATE_PTFV(C1D1U0, cval);
 356
 357                if (dvfs_update || periodic_training_update) {
 358                        tdel = next->current_dram_clktree[C1D1U0] -
 359                                        __MOVAVG_AC(next, C1D1U0);
 360                        tmdel = (tdel < 0) ? -1 * tdel : tdel;
 361
 362                        if (tmdel > adel)
 363                                adel = tmdel;
 364
 365                        if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 366                            next->tree_margin)
 367                                next->current_dram_clktree[C1D1U0] =
 368                                        __MOVAVG_AC(next, C1D1U0);
 369                }
 370
 371                if (dvfs_pt1 || periodic_training_update) {
 372                        cval = tegra210_emc_actual_osc_clocks(last->run_clocks);
 373                        cval *= 1000000;
 374                        cval /= last_timing_rate_mhz * 2 * temp[1][1];
 375                }
 376
 377                if (dvfs_pt1)
 378                        __INCREMENT_PTFV(C1D1U1, cval);
 379                else if (dvfs_update)
 380                        __AVERAGE_PTFV(C1D1U1);
 381                else if (periodic_training_update)
 382                        __WEIGHTED_UPDATE_PTFV(C1D1U1, cval);
 383
 384                if (dvfs_update || periodic_training_update) {
 385                        tdel = next->current_dram_clktree[C1D1U1] -
 386                                        __MOVAVG_AC(next, C1D1U1);
 387                        tmdel = (tdel < 0) ? -1 * tdel : tdel;
 388
 389                        if (tmdel > adel)
 390                                adel = tmdel;
 391
 392                        if (tmdel * 128 * next_timing_rate_mhz / 1000000 >
 393                            next->tree_margin)
 394                                next->current_dram_clktree[C1D1U1] =
 395                                        __MOVAVG_AC(next, C1D1U1);
 396                }
 397        }
 398
 399done:
 400        return adel;
 401}
 402
 403static u32 periodic_compensation_handler(struct tegra210_emc *emc, u32 type,
 404                                         struct tegra210_emc_timing *last,
 405                                         struct tegra210_emc_timing *next)
 406{
 407#define __COPY_EMA(nt, lt, dev)                                         \
 408        ({ __MOVAVG(nt, dev) = __MOVAVG(lt, dev) *                      \
 409           (nt)->ptfv_list[PTFV_DVFS_SAMPLES_INDEX]; })
 410
 411        u32 i, adel = 0, samples = next->ptfv_list[PTFV_DVFS_SAMPLES_INDEX];
 412        u32 delay;
 413
 414        delay = tegra210_emc_actual_osc_clocks(last->run_clocks);
 415        delay *= 1000;
 416        delay = 2 + (delay / last->rate);
 417
 418        if (!next->periodic_training)
 419                return 0;
 420
 421        if (type == DVFS_SEQUENCE) {
 422                if (last->periodic_training &&
 423                    (next->ptfv_list[PTFV_CONFIG_CTRL_INDEX] &
 424                     PTFV_CONFIG_CTRL_USE_PREVIOUS_EMA)) {
 425                        /*
 426                         * If the previous frequency was using periodic
 427                         * calibration then we can reuse the previous
 428                         * frequencies EMA data.
 429                         */
 430                        __COPY_EMA(next, last, C0D0U0);
 431                        __COPY_EMA(next, last, C0D0U1);
 432                        __COPY_EMA(next, last, C1D0U0);
 433                        __COPY_EMA(next, last, C1D0U1);
 434                        __COPY_EMA(next, last, C0D1U0);
 435                        __COPY_EMA(next, last, C0D1U1);
 436                        __COPY_EMA(next, last, C1D1U0);
 437                        __COPY_EMA(next, last, C1D1U1);
 438                } else {
 439                        /* Reset the EMA.*/
 440                        __MOVAVG(next, C0D0U0) = 0;
 441                        __MOVAVG(next, C0D0U1) = 0;
 442                        __MOVAVG(next, C1D0U0) = 0;
 443                        __MOVAVG(next, C1D0U1) = 0;
 444                        __MOVAVG(next, C0D1U0) = 0;
 445                        __MOVAVG(next, C0D1U1) = 0;
 446                        __MOVAVG(next, C1D1U0) = 0;
 447                        __MOVAVG(next, C1D1U1) = 0;
 448
 449                        for (i = 0; i < samples; i++) {
 450                                tegra210_emc_start_periodic_compensation(emc);
 451                                udelay(delay);
 452
 453                                /*
 454                                 * Generate next sample of data.
 455                                 */
 456                                adel = update_clock_tree_delay(emc, DVFS_PT1);
 457                        }
 458                }
 459
 460                /*
 461                 * Seems like it should be part of the
 462                 * 'if (last_timing->periodic_training)' conditional
 463                 * since is already done for the else clause.
 464                 */
 465                adel = update_clock_tree_delay(emc, DVFS_UPDATE);
 466        }
 467
 468        if (type == PERIODIC_TRAINING_SEQUENCE) {
 469                tegra210_emc_start_periodic_compensation(emc);
 470                udelay(delay);
 471
 472                adel = update_clock_tree_delay(emc, PERIODIC_TRAINING_UPDATE);
 473        }
 474
 475        return adel;
 476}
 477
 478static u32 tegra210_emc_r21021_periodic_compensation(struct tegra210_emc *emc)
 479{
 480        u32 emc_cfg, emc_cfg_o, emc_cfg_update, del, value;
 481        u32 list[] = {
 482                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_0,
 483                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_1,
 484                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_2,
 485                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_3,
 486                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_0,
 487                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_1,
 488                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_2,
 489                EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_3,
 490                EMC_DATA_BRLSHFT_0,
 491                EMC_DATA_BRLSHFT_1
 492        };
 493        struct tegra210_emc_timing *last = emc->last;
 494        unsigned int items = ARRAY_SIZE(list), i;
 495        unsigned long delay;
 496
 497        if (last->periodic_training) {
 498                emc_dbg(emc, PER_TRAIN, "Periodic training starting\n");
 499
 500                value = emc_readl(emc, EMC_DBG);
 501                emc_cfg_o = emc_readl(emc, EMC_CFG);
 502                emc_cfg = emc_cfg_o & ~(EMC_CFG_DYN_SELF_REF |
 503                                        EMC_CFG_DRAM_ACPD |
 504                                        EMC_CFG_DRAM_CLKSTOP_PD);
 505
 506
 507                /*
 508                 * 1. Power optimizations should be off.
 509                 */
 510                emc_writel(emc, emc_cfg, EMC_CFG);
 511
 512                /* Does emc_timing_update() for above changes. */
 513                tegra210_emc_dll_disable(emc);
 514
 515                for (i = 0; i < emc->num_channels; i++)
 516                        tegra210_emc_wait_for_update(emc, i, EMC_EMC_STATUS,
 517                                                     EMC_EMC_STATUS_DRAM_IN_POWERDOWN_MASK,
 518                                                     0);
 519
 520                for (i = 0; i < emc->num_channels; i++)
 521                        tegra210_emc_wait_for_update(emc, i, EMC_EMC_STATUS,
 522                                                     EMC_EMC_STATUS_DRAM_IN_SELF_REFRESH_MASK,
 523                                                     0);
 524
 525                emc_cfg_update = value = emc_readl(emc, EMC_CFG_UPDATE);
 526                value &= ~EMC_CFG_UPDATE_UPDATE_DLL_IN_UPDATE_MASK;
 527                value |= (2 << EMC_CFG_UPDATE_UPDATE_DLL_IN_UPDATE_SHIFT);
 528                emc_writel(emc, value, EMC_CFG_UPDATE);
 529
 530                /*
 531                 * 2. osc kick off - this assumes training and dvfs have set
 532                 *    correct MR23.
 533                 */
 534                tegra210_emc_start_periodic_compensation(emc);
 535
 536                /*
 537                 * 3. Let dram capture its clock tree delays.
 538                 */
 539                delay = tegra210_emc_actual_osc_clocks(last->run_clocks);
 540                delay *= 1000;
 541                delay /= last->rate + 1;
 542                udelay(delay);
 543
 544                /*
 545                 * 4. Check delta wrt previous values (save value if margin
 546                 *    exceeds what is set in table).
 547                 */
 548                del = periodic_compensation_handler(emc,
 549                                                    PERIODIC_TRAINING_SEQUENCE,
 550                                                    last, last);
 551
 552                /*
 553                 * 5. Apply compensation w.r.t. trained values (if clock tree
 554                 *    has drifted more than the set margin).
 555                 */
 556                if (last->tree_margin < ((del * 128 * (last->rate / 1000)) / 1000000)) {
 557                        for (i = 0; i < items; i++) {
 558                                value = tegra210_emc_compensate(last, list[i]);
 559                                emc_dbg(emc, EMA_WRITES, "0x%08x <= 0x%08x\n",
 560                                        list[i], value);
 561                                emc_writel(emc, value, list[i]);
 562                        }
 563                }
 564
 565                emc_writel(emc, emc_cfg_o, EMC_CFG);
 566
 567                /*
 568                 * 6. Timing update actally applies the new trimmers.
 569                 */
 570                tegra210_emc_timing_update(emc);
 571
 572                /* 6.1. Restore the UPDATE_DLL_IN_UPDATE field. */
 573                emc_writel(emc, emc_cfg_update, EMC_CFG_UPDATE);
 574
 575                /* 6.2. Restore the DLL. */
 576                tegra210_emc_dll_enable(emc);
 577        }
 578
 579        return 0;
 580}
 581
 582/*
 583 * Do the clock change sequence.
 584 */
 585static void tegra210_emc_r21021_set_clock(struct tegra210_emc *emc, u32 clksrc)
 586{
 587        /* state variables */
 588        static bool fsp_for_next_freq;
 589        /* constant configuration parameters */
 590        const bool save_restore_clkstop_pd = true;
 591        const u32 zqcal_before_cc_cutoff = 2400;
 592        const bool cya_allow_ref_cc = false;
 593        const bool cya_issue_pc_ref = false;
 594        const bool opt_cc_short_zcal = true;
 595        const bool ref_b4_sref_en = false;
 596        const u32 tZQCAL_lpddr4 = 1000000;
 597        const bool opt_short_zcal = true;
 598        const bool opt_do_sw_qrst = true;
 599        const u32 opt_dvfs_mode = MAN_SR;
 600        /*
 601         * This is the timing table for the source frequency. It does _not_
 602         * necessarily correspond to the actual timing values in the EMC at the
 603         * moment. If the boot BCT differs from the table then this can happen.
 604         * However, we need it for accessing the dram_timings (which are not
 605         * really registers) array for the current frequency.
 606         */
 607        struct tegra210_emc_timing *fake, *last = emc->last, *next = emc->next;
 608        u32 tRTM, RP_war, R2P_war, TRPab_war, deltaTWATM, W2P_war, tRPST;
 609        u32 mr13_flip_fspwr, mr13_flip_fspop, ramp_up_wait, ramp_down_wait;
 610        u32 zq_wait_long, zq_latch_dvfs_wait_time, tZQCAL_lpddr4_fc_adj;
 611        u32 emc_auto_cal_config, auto_cal_en, emc_cfg, emc_sel_dpd_ctrl;
 612        u32 tFC_lpddr4 = 1000 * next->dram_timings[T_FC_LPDDR4];
 613        u32 bg_reg_mode_change, enable_bglp_reg, enable_bg_reg;
 614        bool opt_zcal_en_cc = false, is_lpddr3 = false;
 615        bool compensate_trimmer_applicable = false;
 616        u32 emc_dbg, emc_cfg_pipe_clk, emc_pin;
 617        u32 src_clk_period, dst_clk_period; /* in picoseconds */
 618        bool shared_zq_resistor = false;
 619        u32 value, dram_type;
 620        u32 opt_dll_mode = 0;
 621        unsigned long delay;
 622        unsigned int i;
 623
 624        emc_dbg(emc, INFO, "Running clock change.\n");
 625
 626        /* XXX fake == last */
 627        fake = tegra210_emc_find_timing(emc, last->rate * 1000UL);
 628        fsp_for_next_freq = !fsp_for_next_freq;
 629
 630        value = emc_readl(emc, EMC_FBIO_CFG5) & EMC_FBIO_CFG5_DRAM_TYPE_MASK;
 631        dram_type = value >> EMC_FBIO_CFG5_DRAM_TYPE_SHIFT;
 632
 633        if (last->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX] & BIT(31))
 634                shared_zq_resistor = true;
 635
 636        if ((next->burst_regs[EMC_ZCAL_INTERVAL_INDEX] != 0 &&
 637             last->burst_regs[EMC_ZCAL_INTERVAL_INDEX] == 0) ||
 638            dram_type == DRAM_TYPE_LPDDR4)
 639                opt_zcal_en_cc = true;
 640
 641        if (dram_type == DRAM_TYPE_DDR3)
 642                opt_dll_mode = tegra210_emc_get_dll_state(next);
 643
 644        if ((next->burst_regs[EMC_FBIO_CFG5_INDEX] & BIT(25)) &&
 645            (dram_type == DRAM_TYPE_LPDDR2))
 646                is_lpddr3 = true;
 647
 648        emc_readl(emc, EMC_CFG);
 649        emc_readl(emc, EMC_AUTO_CAL_CONFIG);
 650
 651        src_clk_period = 1000000000 / last->rate;
 652        dst_clk_period = 1000000000 / next->rate;
 653
 654        if (dst_clk_period <= zqcal_before_cc_cutoff)
 655                tZQCAL_lpddr4_fc_adj = tZQCAL_lpddr4 - tFC_lpddr4;
 656        else
 657                tZQCAL_lpddr4_fc_adj = tZQCAL_lpddr4;
 658
 659        tZQCAL_lpddr4_fc_adj /= dst_clk_period;
 660
 661        emc_dbg = emc_readl(emc, EMC_DBG);
 662        emc_pin = emc_readl(emc, EMC_PIN);
 663        emc_cfg_pipe_clk = emc_readl(emc, EMC_CFG_PIPE_CLK);
 664
 665        emc_cfg = next->burst_regs[EMC_CFG_INDEX];
 666        emc_cfg &= ~(EMC_CFG_DYN_SELF_REF | EMC_CFG_DRAM_ACPD |
 667                     EMC_CFG_DRAM_CLKSTOP_SR | EMC_CFG_DRAM_CLKSTOP_PD);
 668        emc_sel_dpd_ctrl = next->emc_sel_dpd_ctrl;
 669        emc_sel_dpd_ctrl &= ~(EMC_SEL_DPD_CTRL_CLK_SEL_DPD_EN |
 670                              EMC_SEL_DPD_CTRL_CA_SEL_DPD_EN |
 671                              EMC_SEL_DPD_CTRL_RESET_SEL_DPD_EN |
 672                              EMC_SEL_DPD_CTRL_ODT_SEL_DPD_EN |
 673                              EMC_SEL_DPD_CTRL_DATA_SEL_DPD_EN);
 674
 675        emc_dbg(emc, INFO, "Clock change version: %d\n",
 676                DVFS_CLOCK_CHANGE_VERSION);
 677        emc_dbg(emc, INFO, "DRAM type = %d\n", dram_type);
 678        emc_dbg(emc, INFO, "DRAM dev #: %u\n", emc->num_devices);
 679        emc_dbg(emc, INFO, "Next EMC clksrc: 0x%08x\n", clksrc);
 680        emc_dbg(emc, INFO, "DLL clksrc:      0x%08x\n", next->dll_clk_src);
 681        emc_dbg(emc, INFO, "last rate: %u, next rate %u\n", last->rate,
 682                next->rate);
 683        emc_dbg(emc, INFO, "last period: %u, next period: %u\n",
 684                src_clk_period, dst_clk_period);
 685        emc_dbg(emc, INFO, "  shared_zq_resistor: %d\n", !!shared_zq_resistor);
 686        emc_dbg(emc, INFO, "  num_channels: %u\n", emc->num_channels);
 687        emc_dbg(emc, INFO, "  opt_dll_mode: %d\n", opt_dll_mode);
 688
 689        /*
 690         * Step 1:
 691         *   Pre DVFS SW sequence.
 692         */
 693        emc_dbg(emc, STEPS, "Step 1\n");
 694        emc_dbg(emc, STEPS, "Step 1.1: Disable DLL temporarily.\n");
 695
 696        value = emc_readl(emc, EMC_CFG_DIG_DLL);
 697        value &= ~EMC_CFG_DIG_DLL_CFG_DLL_EN;
 698        emc_writel(emc, value, EMC_CFG_DIG_DLL);
 699
 700        tegra210_emc_timing_update(emc);
 701
 702        for (i = 0; i < emc->num_channels; i++)
 703                tegra210_emc_wait_for_update(emc, i, EMC_CFG_DIG_DLL,
 704                                             EMC_CFG_DIG_DLL_CFG_DLL_EN, 0);
 705
 706        emc_dbg(emc, STEPS, "Step 1.2: Disable AUTOCAL temporarily.\n");
 707
 708        emc_auto_cal_config = next->emc_auto_cal_config;
 709        auto_cal_en = emc_auto_cal_config & EMC_AUTO_CAL_CONFIG_AUTO_CAL_ENABLE;
 710        emc_auto_cal_config &= ~EMC_AUTO_CAL_CONFIG_AUTO_CAL_START;
 711        emc_auto_cal_config |= EMC_AUTO_CAL_CONFIG_AUTO_CAL_MEASURE_STALL;
 712        emc_auto_cal_config |= EMC_AUTO_CAL_CONFIG_AUTO_CAL_UPDATE_STALL;
 713        emc_auto_cal_config |= auto_cal_en;
 714        emc_writel(emc, emc_auto_cal_config, EMC_AUTO_CAL_CONFIG);
 715        emc_readl(emc, EMC_AUTO_CAL_CONFIG); /* Flush write. */
 716
 717        emc_dbg(emc, STEPS, "Step 1.3: Disable other power features.\n");
 718
 719        tegra210_emc_set_shadow_bypass(emc, ACTIVE);
 720        emc_writel(emc, emc_cfg, EMC_CFG);
 721        emc_writel(emc, emc_sel_dpd_ctrl, EMC_SEL_DPD_CTRL);
 722        tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
 723
 724        if (next->periodic_training) {
 725                tegra210_emc_reset_dram_clktree_values(next);
 726
 727                for (i = 0; i < emc->num_channels; i++)
 728                        tegra210_emc_wait_for_update(emc, i, EMC_EMC_STATUS,
 729                                                     EMC_EMC_STATUS_DRAM_IN_POWERDOWN_MASK,
 730                                                     0);
 731
 732                for (i = 0; i < emc->num_channels; i++)
 733                        tegra210_emc_wait_for_update(emc, i, EMC_EMC_STATUS,
 734                                                     EMC_EMC_STATUS_DRAM_IN_SELF_REFRESH_MASK,
 735                                                     0);
 736
 737                tegra210_emc_start_periodic_compensation(emc);
 738
 739                delay = 1000 * tegra210_emc_actual_osc_clocks(last->run_clocks);
 740                udelay((delay / last->rate) + 2);
 741
 742                value = periodic_compensation_handler(emc, DVFS_SEQUENCE, fake,
 743                                                      next);
 744                value = (value * 128 * next->rate / 1000) / 1000000;
 745
 746                if (next->periodic_training && value > next->tree_margin)
 747                        compensate_trimmer_applicable = true;
 748        }
 749
 750        emc_writel(emc, EMC_INTSTATUS_CLKCHANGE_COMPLETE, EMC_INTSTATUS);
 751        tegra210_emc_set_shadow_bypass(emc, ACTIVE);
 752        emc_writel(emc, emc_cfg, EMC_CFG);
 753        emc_writel(emc, emc_sel_dpd_ctrl, EMC_SEL_DPD_CTRL);
 754        emc_writel(emc, emc_cfg_pipe_clk | EMC_CFG_PIPE_CLK_CLK_ALWAYS_ON,
 755                   EMC_CFG_PIPE_CLK);
 756        emc_writel(emc, next->emc_fdpd_ctrl_cmd_no_ramp &
 757                        ~EMC_FDPD_CTRL_CMD_NO_RAMP_CMD_DPD_NO_RAMP_ENABLE,
 758                   EMC_FDPD_CTRL_CMD_NO_RAMP);
 759
 760        bg_reg_mode_change =
 761                ((next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 762                  EMC_PMACRO_BG_BIAS_CTRL_0_BGLP_E_PWRD) ^
 763                 (last->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 764                  EMC_PMACRO_BG_BIAS_CTRL_0_BGLP_E_PWRD)) ||
 765                ((next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 766                  EMC_PMACRO_BG_BIAS_CTRL_0_BG_E_PWRD) ^
 767                 (last->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 768                  EMC_PMACRO_BG_BIAS_CTRL_0_BG_E_PWRD));
 769        enable_bglp_reg =
 770                (next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 771                 EMC_PMACRO_BG_BIAS_CTRL_0_BGLP_E_PWRD) == 0;
 772        enable_bg_reg =
 773                (next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 774                 EMC_PMACRO_BG_BIAS_CTRL_0_BG_E_PWRD) == 0;
 775
 776        if (bg_reg_mode_change) {
 777                if (enable_bg_reg)
 778                        emc_writel(emc, last->burst_regs
 779                                   [EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 780                                   ~EMC_PMACRO_BG_BIAS_CTRL_0_BG_E_PWRD,
 781                                   EMC_PMACRO_BG_BIAS_CTRL_0);
 782
 783                if (enable_bglp_reg)
 784                        emc_writel(emc, last->burst_regs
 785                                   [EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
 786                                   ~EMC_PMACRO_BG_BIAS_CTRL_0_BGLP_E_PWRD,
 787                                   EMC_PMACRO_BG_BIAS_CTRL_0);
 788        }
 789
 790        /* Check if we need to turn on VREF generator. */
 791        if ((((last->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX] &
 792               EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQ_E_IVREF) == 0) &&
 793             ((next->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX] &
 794               EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQ_E_IVREF) == 1)) ||
 795            (((last->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX] &
 796               EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQS_E_IVREF) == 0) &&
 797             ((next->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX] &
 798               EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQS_E_IVREF) != 0))) {
 799                u32 pad_tx_ctrl =
 800                    next->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX];
 801                u32 last_pad_tx_ctrl =
 802                    last->burst_regs[EMC_PMACRO_DATA_PAD_TX_CTRL_INDEX];
 803                u32 next_dq_e_ivref, next_dqs_e_ivref;
 804
 805                next_dqs_e_ivref = pad_tx_ctrl &
 806                                   EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQS_E_IVREF;
 807                next_dq_e_ivref = pad_tx_ctrl &
 808                                  EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQ_E_IVREF;
 809                value = (last_pad_tx_ctrl &
 810                                ~EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQ_E_IVREF &
 811                                ~EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQS_E_IVREF) |
 812                        next_dq_e_ivref | next_dqs_e_ivref;
 813                emc_writel(emc, value, EMC_PMACRO_DATA_PAD_TX_CTRL);
 814                udelay(1);
 815        } else if (bg_reg_mode_change) {
 816                udelay(1);
 817        }
 818
 819        tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
 820
 821        /*
 822         * Step 2:
 823         *   Prelock the DLL.
 824         */
 825        emc_dbg(emc, STEPS, "Step 2\n");
 826
 827        if (next->burst_regs[EMC_CFG_DIG_DLL_INDEX] &
 828            EMC_CFG_DIG_DLL_CFG_DLL_EN) {
 829                emc_dbg(emc, INFO, "Prelock enabled for target frequency.\n");
 830                value = tegra210_emc_dll_prelock(emc, clksrc);
 831                emc_dbg(emc, INFO, "DLL out: 0x%03x\n", value);
 832        } else {
 833                emc_dbg(emc, INFO, "Disabling DLL for target frequency.\n");
 834                tegra210_emc_dll_disable(emc);
 835        }
 836
 837        /*
 838         * Step 3:
 839         *   Prepare autocal for the clock change.
 840         */
 841        emc_dbg(emc, STEPS, "Step 3\n");
 842
 843        tegra210_emc_set_shadow_bypass(emc, ACTIVE);
 844        emc_writel(emc, next->emc_auto_cal_config2, EMC_AUTO_CAL_CONFIG2);
 845        emc_writel(emc, next->emc_auto_cal_config3, EMC_AUTO_CAL_CONFIG3);
 846        emc_writel(emc, next->emc_auto_cal_config4, EMC_AUTO_CAL_CONFIG4);
 847        emc_writel(emc, next->emc_auto_cal_config5, EMC_AUTO_CAL_CONFIG5);
 848        emc_writel(emc, next->emc_auto_cal_config6, EMC_AUTO_CAL_CONFIG6);
 849        emc_writel(emc, next->emc_auto_cal_config7, EMC_AUTO_CAL_CONFIG7);
 850        emc_writel(emc, next->emc_auto_cal_config8, EMC_AUTO_CAL_CONFIG8);
 851        tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
 852
 853        emc_auto_cal_config |= (EMC_AUTO_CAL_CONFIG_AUTO_CAL_COMPUTE_START |
 854                                auto_cal_en);
 855        emc_writel(emc, emc_auto_cal_config, EMC_AUTO_CAL_CONFIG);
 856
 857        /*
 858         * Step 4:
 859         *   Update EMC_CFG. (??)
 860         */
 861        emc_dbg(emc, STEPS, "Step 4\n");
 862
 863        if (src_clk_period > 50000 && dram_type == DRAM_TYPE_LPDDR4)
 864                ccfifo_writel(emc, 1, EMC_SELF_REF, 0);
 865        else
 866                emc_writel(emc, next->emc_cfg_2, EMC_CFG_2);
 867
 868        /*
 869         * Step 5:
 870         *   Prepare reference variables for ZQCAL regs.
 871         */
 872        emc_dbg(emc, STEPS, "Step 5\n");
 873
 874        if (dram_type == DRAM_TYPE_LPDDR4)
 875                zq_wait_long = max((u32)1, div_o3(1000000, dst_clk_period));
 876        else if (dram_type == DRAM_TYPE_LPDDR2 || is_lpddr3)
 877                zq_wait_long = max(next->min_mrs_wait,
 878                                   div_o3(360000, dst_clk_period)) + 4;
 879        else if (dram_type == DRAM_TYPE_DDR3)
 880                zq_wait_long = max((u32)256,
 881                                   div_o3(320000, dst_clk_period) + 2);
 882        else
 883                zq_wait_long = 0;
 884
 885        /*
 886         * Step 6:
 887         *   Training code - removed.
 888         */
 889        emc_dbg(emc, STEPS, "Step 6\n");
 890
 891        /*
 892         * Step 7:
 893         *   Program FSP reference registers and send MRWs to new FSPWR.
 894         */
 895        emc_dbg(emc, STEPS, "Step 7\n");
 896        emc_dbg(emc, SUB_STEPS, "Step 7.1: Bug 200024907 - Patch RP R2P");
 897
 898        /* WAR 200024907 */
 899        if (dram_type == DRAM_TYPE_LPDDR4) {
 900                u32 nRTP = 16;
 901
 902                if (src_clk_period >= 1000000 / 1866) /* 535.91 ps */
 903                        nRTP = 14;
 904
 905                if (src_clk_period >= 1000000 / 1600) /* 625.00 ps */
 906                        nRTP = 12;
 907
 908                if (src_clk_period >= 1000000 / 1333) /* 750.19 ps */
 909                        nRTP = 10;
 910
 911                if (src_clk_period >= 1000000 / 1066) /* 938.09 ps */
 912                        nRTP = 8;
 913
 914                deltaTWATM = max_t(u32, div_o3(7500, src_clk_period), 8);
 915
 916                /*
 917                 * Originally there was a + .5 in the tRPST calculation.
 918                 * However since we can't do FP in the kernel and the tRTM
 919                 * computation was in a floating point ceiling function, adding
 920                 * one to tRTP should be ok. There is no other source of non
 921                 * integer values, so the result was always going to be
 922                 * something for the form: f_ceil(N + .5) = N + 1;
 923                 */
 924                tRPST = (last->emc_mrw & 0x80) >> 7;
 925                tRTM = fake->dram_timings[RL] + div_o3(3600, src_clk_period) +
 926                        max_t(u32, div_o3(7500, src_clk_period), 8) + tRPST +
 927                        1 + nRTP;
 928
 929                emc_dbg(emc, INFO, "tRTM = %u, EMC_RP = %u\n", tRTM,
 930                        next->burst_regs[EMC_RP_INDEX]);
 931
 932                if (last->burst_regs[EMC_RP_INDEX] < tRTM) {
 933                        if (tRTM > (last->burst_regs[EMC_R2P_INDEX] +
 934                                    last->burst_regs[EMC_RP_INDEX])) {
 935                                R2P_war = tRTM - last->burst_regs[EMC_RP_INDEX];
 936                                RP_war = last->burst_regs[EMC_RP_INDEX];
 937                                TRPab_war = last->burst_regs[EMC_TRPAB_INDEX];
 938
 939                                if (R2P_war > 63) {
 940                                        RP_war = R2P_war +
 941                                                 last->burst_regs[EMC_RP_INDEX] - 63;
 942
 943                                        if (TRPab_war < RP_war)
 944                                                TRPab_war = RP_war;
 945
 946                                        R2P_war = 63;
 947                                }
 948                        } else {
 949                                R2P_war = last->burst_regs[EMC_R2P_INDEX];
 950                                RP_war = last->burst_regs[EMC_RP_INDEX];
 951                                TRPab_war = last->burst_regs[EMC_TRPAB_INDEX];
 952                        }
 953
 954                        if (RP_war < deltaTWATM) {
 955                                W2P_war = last->burst_regs[EMC_W2P_INDEX]
 956                                          + deltaTWATM - RP_war;
 957                                if (W2P_war > 63) {
 958                                        RP_war = RP_war + W2P_war - 63;
 959                                        if (TRPab_war < RP_war)
 960                                                TRPab_war = RP_war;
 961                                        W2P_war = 63;
 962                                }
 963                        } else {
 964                                W2P_war = last->burst_regs[
 965                                          EMC_W2P_INDEX];
 966                        }
 967
 968                        if ((last->burst_regs[EMC_W2P_INDEX] ^ W2P_war) ||
 969                            (last->burst_regs[EMC_R2P_INDEX] ^ R2P_war) ||
 970                            (last->burst_regs[EMC_RP_INDEX] ^ RP_war) ||
 971                            (last->burst_regs[EMC_TRPAB_INDEX] ^ TRPab_war)) {
 972                                emc_writel(emc, RP_war, EMC_RP);
 973                                emc_writel(emc, R2P_war, EMC_R2P);
 974                                emc_writel(emc, W2P_war, EMC_W2P);
 975                                emc_writel(emc, TRPab_war, EMC_TRPAB);
 976                        }
 977
 978                        tegra210_emc_timing_update(emc);
 979                } else {
 980                        emc_dbg(emc, INFO, "Skipped WAR\n");
 981                }
 982        }
 983
 984        if (!fsp_for_next_freq) {
 985                mr13_flip_fspwr = (next->emc_mrw3 & 0xffffff3f) | 0x80;
 986                mr13_flip_fspop = (next->emc_mrw3 & 0xffffff3f) | 0x00;
 987        } else {
 988                mr13_flip_fspwr = (next->emc_mrw3 & 0xffffff3f) | 0x40;
 989                mr13_flip_fspop = (next->emc_mrw3 & 0xffffff3f) | 0xc0;
 990        }
 991
 992        if (dram_type == DRAM_TYPE_LPDDR4) {
 993                emc_writel(emc, mr13_flip_fspwr, EMC_MRW3);
 994                emc_writel(emc, next->emc_mrw, EMC_MRW);
 995                emc_writel(emc, next->emc_mrw2, EMC_MRW2);
 996        }
 997
 998        /*
 999         * Step 8:
1000         *   Program the shadow registers.
1001         */
1002        emc_dbg(emc, STEPS, "Step 8\n");
1003        emc_dbg(emc, SUB_STEPS, "Writing burst_regs\n");
1004
1005        for (i = 0; i < next->num_burst; i++) {
1006                const u16 *offsets = emc->offsets->burst;
1007                u16 offset;
1008
1009                if (!offsets[i])
1010                        continue;
1011
1012                value = next->burst_regs[i];
1013                offset = offsets[i];
1014
1015                if (dram_type != DRAM_TYPE_LPDDR4 &&
1016                    (offset == EMC_MRW6 || offset == EMC_MRW7 ||
1017                     offset == EMC_MRW8 || offset == EMC_MRW9 ||
1018                     offset == EMC_MRW10 || offset == EMC_MRW11 ||
1019                     offset == EMC_MRW12 || offset == EMC_MRW13 ||
1020                     offset == EMC_MRW14 || offset == EMC_MRW15 ||
1021                     offset == EMC_TRAINING_CTRL))
1022                        continue;
1023
1024                /* Pain... And suffering. */
1025                if (offset == EMC_CFG) {
1026                        value &= ~EMC_CFG_DRAM_ACPD;
1027                        value &= ~EMC_CFG_DYN_SELF_REF;
1028
1029                        if (dram_type == DRAM_TYPE_LPDDR4) {
1030                                value &= ~EMC_CFG_DRAM_CLKSTOP_SR;
1031                                value &= ~EMC_CFG_DRAM_CLKSTOP_PD;
1032                        }
1033                } else if (offset == EMC_MRS_WAIT_CNT &&
1034                           dram_type == DRAM_TYPE_LPDDR2 &&
1035                           opt_zcal_en_cc && !opt_cc_short_zcal &&
1036                           opt_short_zcal) {
1037                        value = (value & ~(EMC_MRS_WAIT_CNT_SHORT_WAIT_MASK <<
1038                                           EMC_MRS_WAIT_CNT_SHORT_WAIT_SHIFT)) |
1039                                ((zq_wait_long & EMC_MRS_WAIT_CNT_SHORT_WAIT_MASK) <<
1040                                                 EMC_MRS_WAIT_CNT_SHORT_WAIT_SHIFT);
1041                } else if (offset == EMC_ZCAL_WAIT_CNT &&
1042                           dram_type == DRAM_TYPE_DDR3 && opt_zcal_en_cc &&
1043                           !opt_cc_short_zcal && opt_short_zcal) {
1044                        value = (value & ~(EMC_ZCAL_WAIT_CNT_ZCAL_WAIT_CNT_MASK <<
1045                                           EMC_ZCAL_WAIT_CNT_ZCAL_WAIT_CNT_SHIFT)) |
1046                                ((zq_wait_long & EMC_ZCAL_WAIT_CNT_ZCAL_WAIT_CNT_MASK) <<
1047                                                 EMC_MRS_WAIT_CNT_SHORT_WAIT_SHIFT);
1048                } else if (offset == EMC_ZCAL_INTERVAL && opt_zcal_en_cc) {
1049                        value = 0; /* EMC_ZCAL_INTERVAL reset value. */
1050                } else if (offset == EMC_PMACRO_AUTOCAL_CFG_COMMON) {
1051                        value |= EMC_PMACRO_AUTOCAL_CFG_COMMON_E_CAL_BYPASS_DVFS;
1052                } else if (offset == EMC_PMACRO_DATA_PAD_TX_CTRL) {
1053                        value &= ~(EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQSP_TX_E_DCC |
1054                                   EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQSN_TX_E_DCC |
1055                                   EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_DQ_TX_E_DCC |
1056                                   EMC_PMACRO_DATA_PAD_TX_CTRL_DATA_CMD_TX_E_DCC);
1057                } else if (offset == EMC_PMACRO_CMD_PAD_TX_CTRL) {
1058                        value |= EMC_PMACRO_CMD_PAD_TX_CTRL_CMD_DQ_TX_DRVFORCEON;
1059                        value &= ~(EMC_PMACRO_CMD_PAD_TX_CTRL_CMD_DQSP_TX_E_DCC |
1060                                   EMC_PMACRO_CMD_PAD_TX_CTRL_CMD_DQSN_TX_E_DCC |
1061                                   EMC_PMACRO_CMD_PAD_TX_CTRL_CMD_DQ_TX_E_DCC |
1062                                   EMC_PMACRO_CMD_PAD_TX_CTRL_CMD_CMD_TX_E_DCC);
1063                } else if (offset == EMC_PMACRO_BRICK_CTRL_RFU1) {
1064                        value &= 0xf800f800;
1065                } else if (offset == EMC_PMACRO_COMMON_PAD_TX_CTRL) {
1066                        value &= 0xfffffff0;
1067                }
1068
1069                emc_writel(emc, value, offset);
1070        }
1071
1072        /* SW addition: do EMC refresh adjustment here. */
1073        tegra210_emc_adjust_timing(emc, next);
1074
1075        if (dram_type == DRAM_TYPE_LPDDR4) {
1076                value = (23 << EMC_MRW_MRW_MA_SHIFT) |
1077                        (next->run_clocks & EMC_MRW_MRW_OP_MASK);
1078                emc_writel(emc, value, EMC_MRW);
1079        }
1080
1081        /* Per channel burst registers. */
1082        emc_dbg(emc, SUB_STEPS, "Writing burst_regs_per_ch\n");
1083
1084        for (i = 0; i < next->num_burst_per_ch; i++) {
1085                const struct tegra210_emc_per_channel_regs *burst =
1086                                emc->offsets->burst_per_channel;
1087
1088                if (!burst[i].offset)
1089                        continue;
1090
1091                if (dram_type != DRAM_TYPE_LPDDR4 &&
1092                    (burst[i].offset == EMC_MRW6 ||
1093                     burst[i].offset == EMC_MRW7 ||
1094                     burst[i].offset == EMC_MRW8 ||
1095                     burst[i].offset == EMC_MRW9 ||
1096                     burst[i].offset == EMC_MRW10 ||
1097                     burst[i].offset == EMC_MRW11 ||
1098                     burst[i].offset == EMC_MRW12 ||
1099                     burst[i].offset == EMC_MRW13 ||
1100                     burst[i].offset == EMC_MRW14 ||
1101                     burst[i].offset == EMC_MRW15))
1102                        continue;
1103
1104                /* Filter out second channel if not in DUAL_CHANNEL mode. */
1105                if (emc->num_channels < 2 && burst[i].bank >= 1)
1106                        continue;
1107
1108                emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1109                        next->burst_reg_per_ch[i], burst[i].offset);
1110                emc_channel_writel(emc, burst[i].bank,
1111                                   next->burst_reg_per_ch[i],
1112                                   burst[i].offset);
1113        }
1114
1115        /* Vref regs. */
1116        emc_dbg(emc, SUB_STEPS, "Writing vref_regs\n");
1117
1118        for (i = 0; i < next->vref_num; i++) {
1119                const struct tegra210_emc_per_channel_regs *vref =
1120                                        emc->offsets->vref_per_channel;
1121
1122                if (!vref[i].offset)
1123                        continue;
1124
1125                if (emc->num_channels < 2 && vref[i].bank >= 1)
1126                        continue;
1127
1128                emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1129                        next->vref_perch_regs[i], vref[i].offset);
1130                emc_channel_writel(emc, vref[i].bank, next->vref_perch_regs[i],
1131                                   vref[i].offset);
1132        }
1133
1134        /* Trimmers. */
1135        emc_dbg(emc, SUB_STEPS, "Writing trim_regs\n");
1136
1137        for (i = 0; i < next->num_trim; i++) {
1138                const u16 *offsets = emc->offsets->trim;
1139
1140                if (!offsets[i])
1141                        continue;
1142
1143                if (compensate_trimmer_applicable &&
1144                    (offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_0 ||
1145                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_1 ||
1146                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_2 ||
1147                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_3 ||
1148                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_0 ||
1149                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_1 ||
1150                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_2 ||
1151                     offsets[i] == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_3 ||
1152                     offsets[i] == EMC_DATA_BRLSHFT_0 ||
1153                     offsets[i] == EMC_DATA_BRLSHFT_1)) {
1154                        value = tegra210_emc_compensate(next, offsets[i]);
1155                        emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1156                                value, offsets[i]);
1157                        emc_dbg(emc, EMA_WRITES, "0x%08x <= 0x%08x\n",
1158                                (u32)(u64)offsets[i], value);
1159                        emc_writel(emc, value, offsets[i]);
1160                } else {
1161                        emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1162                                next->trim_regs[i], offsets[i]);
1163                        emc_writel(emc, next->trim_regs[i], offsets[i]);
1164                }
1165        }
1166
1167        /* Per channel trimmers. */
1168        emc_dbg(emc, SUB_STEPS, "Writing trim_regs_per_ch\n");
1169
1170        for (i = 0; i < next->num_trim_per_ch; i++) {
1171                const struct tegra210_emc_per_channel_regs *trim =
1172                                &emc->offsets->trim_per_channel[0];
1173                unsigned int offset;
1174
1175                if (!trim[i].offset)
1176                        continue;
1177
1178                if (emc->num_channels < 2 && trim[i].bank >= 1)
1179                        continue;
1180
1181                offset = trim[i].offset;
1182
1183                if (compensate_trimmer_applicable &&
1184                    (offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_0 ||
1185                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_1 ||
1186                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_2 ||
1187                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK0_3 ||
1188                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_0 ||
1189                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_1 ||
1190                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_2 ||
1191                     offset == EMC_PMACRO_OB_DDLL_LONG_DQ_RANK1_3 ||
1192                     offset == EMC_DATA_BRLSHFT_0 ||
1193                     offset == EMC_DATA_BRLSHFT_1)) {
1194                        value = tegra210_emc_compensate(next, offset);
1195                        emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1196                                value, offset);
1197                        emc_dbg(emc, EMA_WRITES, "0x%08x <= 0x%08x\n", offset,
1198                                value);
1199                        emc_channel_writel(emc, trim[i].bank, value, offset);
1200                } else {
1201                        emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1202                                next->trim_perch_regs[i], offset);
1203                        emc_channel_writel(emc, trim[i].bank,
1204                                           next->trim_perch_regs[i], offset);
1205                }
1206        }
1207
1208        emc_dbg(emc, SUB_STEPS, "Writing burst_mc_regs\n");
1209
1210        for (i = 0; i < next->num_mc_regs; i++) {
1211                const u16 *offsets = emc->offsets->burst_mc;
1212                u32 *values = next->burst_mc_regs;
1213
1214                emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1215                        values[i], offsets[i]);
1216                mc_writel(emc->mc, values[i], offsets[i]);
1217        }
1218
1219        /* Registers to be programmed on the faster clock. */
1220        if (next->rate < last->rate) {
1221                const u16 *la = emc->offsets->la_scale;
1222
1223                emc_dbg(emc, SUB_STEPS, "Writing la_scale_regs\n");
1224
1225                for (i = 0; i < next->num_up_down; i++) {
1226                        emc_dbg(emc, REG_LISTS, "(%u) 0x%08x => 0x%08x\n", i,
1227                                next->la_scale_regs[i], la[i]);
1228                        mc_writel(emc->mc, next->la_scale_regs[i], la[i]);
1229                }
1230        }
1231
1232        /* Flush all the burst register writes. */
1233        mc_readl(emc->mc, MC_EMEM_ADR_CFG);
1234
1235        /*
1236         * Step 9:
1237         *   LPDDR4 section A.
1238         */
1239        emc_dbg(emc, STEPS, "Step 9\n");
1240
1241        value = next->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX];
1242        value &= ~EMC_ZCAL_WAIT_CNT_ZCAL_WAIT_CNT_MASK;
1243
1244        if (dram_type == DRAM_TYPE_LPDDR4) {
1245                emc_writel(emc, 0, EMC_ZCAL_INTERVAL);
1246                emc_writel(emc, value, EMC_ZCAL_WAIT_CNT);
1247
1248                value = emc_dbg | (EMC_DBG_WRITE_MUX_ACTIVE |
1249                                   EMC_DBG_WRITE_ACTIVE_ONLY);
1250
1251                emc_writel(emc, value, EMC_DBG);
1252                emc_writel(emc, 0, EMC_ZCAL_INTERVAL);
1253                emc_writel(emc, emc_dbg, EMC_DBG);
1254        }
1255
1256        /*
1257         * Step 10:
1258         *   LPDDR4 and DDR3 common section.
1259         */
1260        emc_dbg(emc, STEPS, "Step 10\n");
1261
1262        if (opt_dvfs_mode == MAN_SR || dram_type == DRAM_TYPE_LPDDR4) {
1263                if (dram_type == DRAM_TYPE_LPDDR4)
1264                        ccfifo_writel(emc, 0x101, EMC_SELF_REF, 0);
1265                else
1266                        ccfifo_writel(emc, 0x1, EMC_SELF_REF, 0);
1267
1268                if (dram_type == DRAM_TYPE_LPDDR4 &&
1269                    dst_clk_period <= zqcal_before_cc_cutoff) {
1270                        ccfifo_writel(emc, mr13_flip_fspwr ^ 0x40, EMC_MRW3, 0);
1271                        ccfifo_writel(emc, (next->burst_regs[EMC_MRW6_INDEX] &
1272                                                0xFFFF3F3F) |
1273                                           (last->burst_regs[EMC_MRW6_INDEX] &
1274                                                0x0000C0C0), EMC_MRW6, 0);
1275                        ccfifo_writel(emc, (next->burst_regs[EMC_MRW14_INDEX] &
1276                                                0xFFFF0707) |
1277                                           (last->burst_regs[EMC_MRW14_INDEX] &
1278                                                0x00003838), EMC_MRW14, 0);
1279
1280                        if (emc->num_devices > 1) {
1281                                ccfifo_writel(emc,
1282                                      (next->burst_regs[EMC_MRW7_INDEX] &
1283                                       0xFFFF3F3F) |
1284                                      (last->burst_regs[EMC_MRW7_INDEX] &
1285                                       0x0000C0C0), EMC_MRW7, 0);
1286                                ccfifo_writel(emc,
1287                                     (next->burst_regs[EMC_MRW15_INDEX] &
1288                                      0xFFFF0707) |
1289                                     (last->burst_regs[EMC_MRW15_INDEX] &
1290                                      0x00003838), EMC_MRW15, 0);
1291                        }
1292
1293                        if (opt_zcal_en_cc) {
1294                                if (emc->num_devices < 2)
1295                                        ccfifo_writel(emc,
1296                                                2UL << EMC_ZQ_CAL_DEV_SEL_SHIFT
1297                                                | EMC_ZQ_CAL_ZQ_CAL_CMD,
1298                                                EMC_ZQ_CAL, 0);
1299                                else if (shared_zq_resistor)
1300                                        ccfifo_writel(emc,
1301                                                2UL << EMC_ZQ_CAL_DEV_SEL_SHIFT
1302                                                | EMC_ZQ_CAL_ZQ_CAL_CMD,
1303                                                EMC_ZQ_CAL, 0);
1304                                else
1305                                        ccfifo_writel(emc,
1306                                                      EMC_ZQ_CAL_ZQ_CAL_CMD,
1307                                                      EMC_ZQ_CAL, 0);
1308                        }
1309                }
1310        }
1311
1312        if (dram_type == DRAM_TYPE_LPDDR4) {
1313                value = (1000 * fake->dram_timings[T_RP]) / src_clk_period;
1314                ccfifo_writel(emc, mr13_flip_fspop | 0x8, EMC_MRW3, value);
1315                ccfifo_writel(emc, 0, 0, tFC_lpddr4 / src_clk_period);
1316        }
1317
1318        if (dram_type == DRAM_TYPE_LPDDR4 || opt_dvfs_mode != MAN_SR) {
1319                delay = 30;
1320
1321                if (cya_allow_ref_cc) {
1322                        delay += (1000 * fake->dram_timings[T_RP]) /
1323                                        src_clk_period;
1324                        delay += 4000 * fake->dram_timings[T_RFC];
1325                }
1326
1327                ccfifo_writel(emc, emc_pin & ~(EMC_PIN_PIN_CKE_PER_DEV |
1328                                               EMC_PIN_PIN_CKEB |
1329                                               EMC_PIN_PIN_CKE),
1330                              EMC_PIN, delay);
1331        }
1332
1333        /* calculate reference delay multiplier */
1334        value = 1;
1335
1336        if (ref_b4_sref_en)
1337                value++;
1338
1339        if (cya_allow_ref_cc)
1340                value++;
1341
1342        if (cya_issue_pc_ref)
1343                value++;
1344
1345        if (dram_type != DRAM_TYPE_LPDDR4) {
1346                delay = ((1000 * fake->dram_timings[T_RP] / src_clk_period) +
1347                         (1000 * fake->dram_timings[T_RFC] / src_clk_period));
1348                delay = value * delay + 20;
1349        } else {
1350                delay = 0;
1351        }
1352
1353        /*
1354         * Step 11:
1355         *   Ramp down.
1356         */
1357        emc_dbg(emc, STEPS, "Step 11\n");
1358
1359        ccfifo_writel(emc, 0x0, EMC_CFG_SYNC, delay);
1360
1361        value = emc_dbg | EMC_DBG_WRITE_MUX_ACTIVE | EMC_DBG_WRITE_ACTIVE_ONLY;
1362        ccfifo_writel(emc, value, EMC_DBG, 0);
1363
1364        ramp_down_wait = tegra210_emc_dvfs_power_ramp_down(emc, src_clk_period,
1365                                                           0);
1366
1367        /*
1368         * Step 12:
1369         *   And finally - trigger the clock change.
1370         */
1371        emc_dbg(emc, STEPS, "Step 12\n");
1372
1373        ccfifo_writel(emc, 1, EMC_STALL_THEN_EXE_AFTER_CLKCHANGE, 0);
1374        value &= ~EMC_DBG_WRITE_ACTIVE_ONLY;
1375        ccfifo_writel(emc, value, EMC_DBG, 0);
1376
1377        /*
1378         * Step 13:
1379         *   Ramp up.
1380         */
1381        emc_dbg(emc, STEPS, "Step 13\n");
1382
1383        ramp_up_wait = tegra210_emc_dvfs_power_ramp_up(emc, dst_clk_period, 0);
1384        ccfifo_writel(emc, emc_dbg, EMC_DBG, 0);
1385
1386        /*
1387         * Step 14:
1388         *   Bringup CKE pins.
1389         */
1390        emc_dbg(emc, STEPS, "Step 14\n");
1391
1392        if (dram_type == DRAM_TYPE_LPDDR4) {
1393                value = emc_pin | EMC_PIN_PIN_CKE;
1394
1395                if (emc->num_devices <= 1)
1396                        value &= ~(EMC_PIN_PIN_CKEB | EMC_PIN_PIN_CKE_PER_DEV);
1397                else
1398                        value |= EMC_PIN_PIN_CKEB | EMC_PIN_PIN_CKE_PER_DEV;
1399
1400                ccfifo_writel(emc, value, EMC_PIN, 0);
1401        }
1402
1403        /*
1404         * Step 15: (two step 15s ??)
1405         *   Calculate zqlatch wait time; has dependency on ramping times.
1406         */
1407        emc_dbg(emc, STEPS, "Step 15\n");
1408
1409        if (dst_clk_period <= zqcal_before_cc_cutoff) {
1410                s32 t = (s32)(ramp_up_wait + ramp_down_wait) /
1411                        (s32)dst_clk_period;
1412                zq_latch_dvfs_wait_time = (s32)tZQCAL_lpddr4_fc_adj - t;
1413        } else {
1414                zq_latch_dvfs_wait_time = tZQCAL_lpddr4_fc_adj -
1415                        div_o3(1000 * next->dram_timings[T_PDEX],
1416                               dst_clk_period);
1417        }
1418
1419        emc_dbg(emc, INFO, "tZQCAL_lpddr4_fc_adj = %u\n", tZQCAL_lpddr4_fc_adj);
1420        emc_dbg(emc, INFO, "dst_clk_period = %u\n",
1421                dst_clk_period);
1422        emc_dbg(emc, INFO, "next->dram_timings[T_PDEX] = %u\n",
1423                next->dram_timings[T_PDEX]);
1424        emc_dbg(emc, INFO, "zq_latch_dvfs_wait_time = %d\n",
1425                max_t(s32, 0, zq_latch_dvfs_wait_time));
1426
1427        if (dram_type == DRAM_TYPE_LPDDR4 && opt_zcal_en_cc) {
1428                delay = div_o3(1000 * next->dram_timings[T_PDEX],
1429                               dst_clk_period);
1430
1431                if (emc->num_devices < 2) {
1432                        if (dst_clk_period > zqcal_before_cc_cutoff)
1433                                ccfifo_writel(emc,
1434                                              2UL << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1435                                              EMC_ZQ_CAL_ZQ_CAL_CMD, EMC_ZQ_CAL,
1436                                              delay);
1437
1438                        value = (mr13_flip_fspop & 0xfffffff7) | 0x0c000000;
1439                        ccfifo_writel(emc, value, EMC_MRW3, delay);
1440                        ccfifo_writel(emc, 0, EMC_SELF_REF, 0);
1441                        ccfifo_writel(emc, 0, EMC_REF, 0);
1442                        ccfifo_writel(emc, 2UL << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1443                                      EMC_ZQ_CAL_ZQ_LATCH_CMD,
1444                                      EMC_ZQ_CAL,
1445                                      max_t(s32, 0, zq_latch_dvfs_wait_time));
1446                } else if (shared_zq_resistor) {
1447                        if (dst_clk_period > zqcal_before_cc_cutoff)
1448                                ccfifo_writel(emc,
1449                                              2UL << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1450                                              EMC_ZQ_CAL_ZQ_CAL_CMD, EMC_ZQ_CAL,
1451                                              delay);
1452
1453                        ccfifo_writel(emc, 2UL << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1454                                      EMC_ZQ_CAL_ZQ_LATCH_CMD, EMC_ZQ_CAL,
1455                                      max_t(s32, 0, zq_latch_dvfs_wait_time) +
1456                                        delay);
1457                        ccfifo_writel(emc, 1UL << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1458                                      EMC_ZQ_CAL_ZQ_LATCH_CMD,
1459                                      EMC_ZQ_CAL, 0);
1460
1461                        value = (mr13_flip_fspop & 0xfffffff7) | 0x0c000000;
1462                        ccfifo_writel(emc, value, EMC_MRW3, 0);
1463                        ccfifo_writel(emc, 0, EMC_SELF_REF, 0);
1464                        ccfifo_writel(emc, 0, EMC_REF, 0);
1465
1466                        ccfifo_writel(emc, 1UL << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1467                                      EMC_ZQ_CAL_ZQ_LATCH_CMD, EMC_ZQ_CAL,
1468                                      tZQCAL_lpddr4 / dst_clk_period);
1469                } else {
1470                        if (dst_clk_period > zqcal_before_cc_cutoff)
1471                                ccfifo_writel(emc, EMC_ZQ_CAL_ZQ_CAL_CMD,
1472                                              EMC_ZQ_CAL, delay);
1473
1474                        value = (mr13_flip_fspop & 0xfffffff7) | 0x0c000000;
1475                        ccfifo_writel(emc, value, EMC_MRW3, delay);
1476                        ccfifo_writel(emc, 0, EMC_SELF_REF, 0);
1477                        ccfifo_writel(emc, 0, EMC_REF, 0);
1478
1479                        ccfifo_writel(emc, EMC_ZQ_CAL_ZQ_LATCH_CMD, EMC_ZQ_CAL,
1480                                      max_t(s32, 0, zq_latch_dvfs_wait_time));
1481                }
1482        }
1483
1484        /* WAR: delay for zqlatch */
1485        ccfifo_writel(emc, 0, 0, 10);
1486
1487        /*
1488         * Step 16:
1489         *   LPDDR4 Conditional Training Kickoff. Removed.
1490         */
1491
1492        /*
1493         * Step 17:
1494         *   MANSR exit self refresh.
1495         */
1496        emc_dbg(emc, STEPS, "Step 17\n");
1497
1498        if (opt_dvfs_mode == MAN_SR && dram_type != DRAM_TYPE_LPDDR4)
1499                ccfifo_writel(emc, 0, EMC_SELF_REF, 0);
1500
1501        /*
1502         * Step 18:
1503         *   Send MRWs to LPDDR3/DDR3.
1504         */
1505        emc_dbg(emc, STEPS, "Step 18\n");
1506
1507        if (dram_type == DRAM_TYPE_LPDDR2) {
1508                ccfifo_writel(emc, next->emc_mrw2, EMC_MRW2, 0);
1509                ccfifo_writel(emc, next->emc_mrw,  EMC_MRW,  0);
1510                if (is_lpddr3)
1511                        ccfifo_writel(emc, next->emc_mrw4, EMC_MRW4, 0);
1512        } else if (dram_type == DRAM_TYPE_DDR3) {
1513                if (opt_dll_mode)
1514                        ccfifo_writel(emc, next->emc_emrs &
1515                                      ~EMC_EMRS_USE_EMRS_LONG_CNT, EMC_EMRS, 0);
1516                ccfifo_writel(emc, next->emc_emrs2 &
1517                              ~EMC_EMRS2_USE_EMRS2_LONG_CNT, EMC_EMRS2, 0);
1518                ccfifo_writel(emc, next->emc_mrs |
1519                              EMC_EMRS_USE_EMRS_LONG_CNT, EMC_MRS, 0);
1520        }
1521
1522        /*
1523         * Step 19:
1524         *   ZQCAL for LPDDR3/DDR3
1525         */
1526        emc_dbg(emc, STEPS, "Step 19\n");
1527
1528        if (opt_zcal_en_cc) {
1529                if (dram_type == DRAM_TYPE_LPDDR2) {
1530                        value = opt_cc_short_zcal ? 90000 : 360000;
1531                        value = div_o3(value, dst_clk_period);
1532                        value = value <<
1533                                EMC_MRS_WAIT_CNT2_MRS_EXT2_WAIT_CNT_SHIFT |
1534                                value <<
1535                                EMC_MRS_WAIT_CNT2_MRS_EXT1_WAIT_CNT_SHIFT;
1536                        ccfifo_writel(emc, value, EMC_MRS_WAIT_CNT2, 0);
1537
1538                        value = opt_cc_short_zcal ? 0x56 : 0xab;
1539                        ccfifo_writel(emc, 2 << EMC_MRW_MRW_DEV_SELECTN_SHIFT |
1540                                           EMC_MRW_USE_MRW_EXT_CNT |
1541                                           10 << EMC_MRW_MRW_MA_SHIFT |
1542                                           value << EMC_MRW_MRW_OP_SHIFT,
1543                                      EMC_MRW, 0);
1544
1545                        if (emc->num_devices > 1) {
1546                                value = 1 << EMC_MRW_MRW_DEV_SELECTN_SHIFT |
1547                                        EMC_MRW_USE_MRW_EXT_CNT |
1548                                        10 << EMC_MRW_MRW_MA_SHIFT |
1549                                        value << EMC_MRW_MRW_OP_SHIFT;
1550                                ccfifo_writel(emc, value, EMC_MRW, 0);
1551                        }
1552                } else if (dram_type == DRAM_TYPE_DDR3) {
1553                        value = opt_cc_short_zcal ? 0 : EMC_ZQ_CAL_LONG;
1554
1555                        ccfifo_writel(emc, value |
1556                                           2 << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1557                                           EMC_ZQ_CAL_ZQ_CAL_CMD, EMC_ZQ_CAL,
1558                                           0);
1559
1560                        if (emc->num_devices > 1) {
1561                                value = value | 1 << EMC_ZQ_CAL_DEV_SEL_SHIFT |
1562                                                EMC_ZQ_CAL_ZQ_CAL_CMD;
1563                                ccfifo_writel(emc, value, EMC_ZQ_CAL, 0);
1564                        }
1565                }
1566        }
1567
1568        if (bg_reg_mode_change) {
1569                tegra210_emc_set_shadow_bypass(emc, ACTIVE);
1570
1571                if (ramp_up_wait <= 1250000)
1572                        delay = (1250000 - ramp_up_wait) / dst_clk_period;
1573                else
1574                        delay = 0;
1575
1576                ccfifo_writel(emc,
1577                              next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX],
1578                              EMC_PMACRO_BG_BIAS_CTRL_0, delay);
1579                tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
1580        }
1581
1582        /*
1583         * Step 20:
1584         *   Issue ref and optional QRST.
1585         */
1586        emc_dbg(emc, STEPS, "Step 20\n");
1587
1588        if (dram_type != DRAM_TYPE_LPDDR4)
1589                ccfifo_writel(emc, 0, EMC_REF, 0);
1590
1591        if (opt_do_sw_qrst) {
1592                ccfifo_writel(emc, 1, EMC_ISSUE_QRST, 0);
1593                ccfifo_writel(emc, 0, EMC_ISSUE_QRST, 2);
1594        }
1595
1596        /*
1597         * Step 21:
1598         *   Restore ZCAL and ZCAL interval.
1599         */
1600        emc_dbg(emc, STEPS, "Step 21\n");
1601
1602        if (save_restore_clkstop_pd || opt_zcal_en_cc) {
1603                ccfifo_writel(emc, emc_dbg | EMC_DBG_WRITE_MUX_ACTIVE,
1604                              EMC_DBG, 0);
1605                if (opt_zcal_en_cc && dram_type != DRAM_TYPE_LPDDR4)
1606                        ccfifo_writel(emc, next->burst_regs[EMC_ZCAL_INTERVAL_INDEX],
1607                                      EMC_ZCAL_INTERVAL, 0);
1608
1609                if (save_restore_clkstop_pd)
1610                        ccfifo_writel(emc, next->burst_regs[EMC_CFG_INDEX] &
1611                                                ~EMC_CFG_DYN_SELF_REF,
1612                                      EMC_CFG, 0);
1613                ccfifo_writel(emc, emc_dbg, EMC_DBG, 0);
1614        }
1615
1616        /*
1617         * Step 22:
1618         *   Restore EMC_CFG_PIPE_CLK.
1619         */
1620        emc_dbg(emc, STEPS, "Step 22\n");
1621
1622        ccfifo_writel(emc, emc_cfg_pipe_clk, EMC_CFG_PIPE_CLK, 0);
1623
1624        if (bg_reg_mode_change) {
1625                if (enable_bg_reg)
1626                        emc_writel(emc,
1627                                   next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
1628                                        ~EMC_PMACRO_BG_BIAS_CTRL_0_BGLP_E_PWRD,
1629                                   EMC_PMACRO_BG_BIAS_CTRL_0);
1630                else
1631                        emc_writel(emc,
1632                                   next->burst_regs[EMC_PMACRO_BG_BIAS_CTRL_0_INDEX] &
1633                                        ~EMC_PMACRO_BG_BIAS_CTRL_0_BG_E_PWRD,
1634                                   EMC_PMACRO_BG_BIAS_CTRL_0);
1635        }
1636
1637        /*
1638         * Step 23:
1639         */
1640        emc_dbg(emc, STEPS, "Step 23\n");
1641
1642        value = emc_readl(emc, EMC_CFG_DIG_DLL);
1643        value |= EMC_CFG_DIG_DLL_CFG_DLL_STALL_ALL_TRAFFIC;
1644        value &= ~EMC_CFG_DIG_DLL_CFG_DLL_STALL_RW_UNTIL_LOCK;
1645        value &= ~EMC_CFG_DIG_DLL_CFG_DLL_STALL_ALL_UNTIL_LOCK;
1646        value &= ~EMC_CFG_DIG_DLL_CFG_DLL_EN;
1647        value = (value & ~EMC_CFG_DIG_DLL_CFG_DLL_MODE_MASK) |
1648                (2 << EMC_CFG_DIG_DLL_CFG_DLL_MODE_SHIFT);
1649        emc_writel(emc, value, EMC_CFG_DIG_DLL);
1650
1651        tegra210_emc_do_clock_change(emc, clksrc);
1652
1653        /*
1654         * Step 24:
1655         *   Save training results. Removed.
1656         */
1657
1658        /*
1659         * Step 25:
1660         *   Program MC updown registers.
1661         */
1662        emc_dbg(emc, STEPS, "Step 25\n");
1663
1664        if (next->rate > last->rate) {
1665                for (i = 0; i < next->num_up_down; i++)
1666                        mc_writel(emc->mc, next->la_scale_regs[i],
1667                                  emc->offsets->la_scale[i]);
1668
1669                tegra210_emc_timing_update(emc);
1670        }
1671
1672        /*
1673         * Step 26:
1674         *   Restore ZCAL registers.
1675         */
1676        emc_dbg(emc, STEPS, "Step 26\n");
1677
1678        if (dram_type == DRAM_TYPE_LPDDR4) {
1679                tegra210_emc_set_shadow_bypass(emc, ACTIVE);
1680                emc_writel(emc, next->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX],
1681                           EMC_ZCAL_WAIT_CNT);
1682                emc_writel(emc, next->burst_regs[EMC_ZCAL_INTERVAL_INDEX],
1683                           EMC_ZCAL_INTERVAL);
1684                tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
1685        }
1686
1687        if (dram_type != DRAM_TYPE_LPDDR4 && opt_zcal_en_cc &&
1688            !opt_short_zcal && opt_cc_short_zcal) {
1689                udelay(2);
1690
1691                tegra210_emc_set_shadow_bypass(emc, ACTIVE);
1692                if (dram_type == DRAM_TYPE_LPDDR2)
1693                        emc_writel(emc, next->burst_regs[EMC_MRS_WAIT_CNT_INDEX],
1694                                   EMC_MRS_WAIT_CNT);
1695                else if (dram_type == DRAM_TYPE_DDR3)
1696                        emc_writel(emc, next->burst_regs[EMC_ZCAL_WAIT_CNT_INDEX],
1697                                   EMC_ZCAL_WAIT_CNT);
1698                tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
1699        }
1700
1701        /*
1702         * Step 27:
1703         *   Restore EMC_CFG, FDPD registers.
1704         */
1705        emc_dbg(emc, STEPS, "Step 27\n");
1706
1707        tegra210_emc_set_shadow_bypass(emc, ACTIVE);
1708        emc_writel(emc, next->burst_regs[EMC_CFG_INDEX], EMC_CFG);
1709        tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
1710        emc_writel(emc, next->emc_fdpd_ctrl_cmd_no_ramp,
1711                   EMC_FDPD_CTRL_CMD_NO_RAMP);
1712        emc_writel(emc, next->emc_sel_dpd_ctrl, EMC_SEL_DPD_CTRL);
1713
1714        /*
1715         * Step 28:
1716         *   Training recover. Removed.
1717         */
1718        emc_dbg(emc, STEPS, "Step 28\n");
1719
1720        tegra210_emc_set_shadow_bypass(emc, ACTIVE);
1721        emc_writel(emc,
1722                   next->burst_regs[EMC_PMACRO_AUTOCAL_CFG_COMMON_INDEX],
1723                   EMC_PMACRO_AUTOCAL_CFG_COMMON);
1724        tegra210_emc_set_shadow_bypass(emc, ASSEMBLY);
1725
1726        /*
1727         * Step 29:
1728         *   Power fix WAR.
1729         */
1730        emc_dbg(emc, STEPS, "Step 29\n");
1731
1732        emc_writel(emc, EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE0 |
1733                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE1 |
1734                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE2 |
1735                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE3 |
1736                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE4 |
1737                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE5 |
1738                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE6 |
1739                   EMC_PMACRO_CFG_PM_GLOBAL_0_DISABLE_CFG_BYTE7,
1740                   EMC_PMACRO_CFG_PM_GLOBAL_0);
1741        emc_writel(emc, EMC_PMACRO_TRAINING_CTRL_0_CH0_TRAINING_E_WRPTR,
1742                   EMC_PMACRO_TRAINING_CTRL_0);
1743        emc_writel(emc, EMC_PMACRO_TRAINING_CTRL_1_CH1_TRAINING_E_WRPTR,
1744                   EMC_PMACRO_TRAINING_CTRL_1);
1745        emc_writel(emc, 0, EMC_PMACRO_CFG_PM_GLOBAL_0);
1746
1747        /*
1748         * Step 30:
1749         *   Re-enable autocal.
1750         */
1751        emc_dbg(emc, STEPS, "Step 30: Re-enable DLL and AUTOCAL\n");
1752
1753        if (next->burst_regs[EMC_CFG_DIG_DLL_INDEX] & EMC_CFG_DIG_DLL_CFG_DLL_EN) {
1754                value = emc_readl(emc, EMC_CFG_DIG_DLL);
1755                value |=  EMC_CFG_DIG_DLL_CFG_DLL_STALL_ALL_TRAFFIC;
1756                value |=  EMC_CFG_DIG_DLL_CFG_DLL_EN;
1757                value &= ~EMC_CFG_DIG_DLL_CFG_DLL_STALL_RW_UNTIL_LOCK;
1758                value &= ~EMC_CFG_DIG_DLL_CFG_DLL_STALL_ALL_UNTIL_LOCK;
1759                value = (value & ~EMC_CFG_DIG_DLL_CFG_DLL_MODE_MASK) |
1760                        (2 << EMC_CFG_DIG_DLL_CFG_DLL_MODE_SHIFT);
1761                emc_writel(emc, value, EMC_CFG_DIG_DLL);
1762                tegra210_emc_timing_update(emc);
1763        }
1764
1765        emc_writel(emc, next->emc_auto_cal_config, EMC_AUTO_CAL_CONFIG);
1766
1767        /* Done! Yay. */
1768}
1769
1770const struct tegra210_emc_sequence tegra210_emc_r21021 = {
1771        .revision = 0x7,
1772        .set_clock = tegra210_emc_r21021_set_clock,
1773        .periodic_compensation = tegra210_emc_r21021_periodic_compensation,
1774};
1775