linux/drivers/bus/mhi/core/boot.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2018-2020, The Linux Foundation. All rights reserved.
   4 *
   5 */
   6
   7#include <linux/delay.h>
   8#include <linux/device.h>
   9#include <linux/dma-direction.h>
  10#include <linux/dma-mapping.h>
  11#include <linux/firmware.h>
  12#include <linux/interrupt.h>
  13#include <linux/list.h>
  14#include <linux/mhi.h>
  15#include <linux/module.h>
  16#include <linux/random.h>
  17#include <linux/slab.h>
  18#include <linux/wait.h>
  19#include "internal.h"
  20
  21/* Setup RDDM vector table for RDDM transfer and program RXVEC */
  22void mhi_rddm_prepare(struct mhi_controller *mhi_cntrl,
  23                      struct image_info *img_info)
  24{
  25        struct mhi_buf *mhi_buf = img_info->mhi_buf;
  26        struct bhi_vec_entry *bhi_vec = img_info->bhi_vec;
  27        void __iomem *base = mhi_cntrl->bhie;
  28        struct device *dev = &mhi_cntrl->mhi_dev->dev;
  29        u32 sequence_id;
  30        unsigned int i;
  31
  32        for (i = 0; i < img_info->entries - 1; i++, mhi_buf++, bhi_vec++) {
  33                bhi_vec->dma_addr = mhi_buf->dma_addr;
  34                bhi_vec->size = mhi_buf->len;
  35        }
  36
  37        dev_dbg(dev, "BHIe programming for RDDM\n");
  38
  39        mhi_write_reg(mhi_cntrl, base, BHIE_RXVECADDR_HIGH_OFFS,
  40                      upper_32_bits(mhi_buf->dma_addr));
  41
  42        mhi_write_reg(mhi_cntrl, base, BHIE_RXVECADDR_LOW_OFFS,
  43                      lower_32_bits(mhi_buf->dma_addr));
  44
  45        mhi_write_reg(mhi_cntrl, base, BHIE_RXVECSIZE_OFFS, mhi_buf->len);
  46        sequence_id = MHI_RANDOM_U32_NONZERO(BHIE_RXVECSTATUS_SEQNUM_BMSK);
  47
  48        mhi_write_reg_field(mhi_cntrl, base, BHIE_RXVECDB_OFFS,
  49                            BHIE_RXVECDB_SEQNUM_BMSK, BHIE_RXVECDB_SEQNUM_SHFT,
  50                            sequence_id);
  51
  52        dev_dbg(dev, "Address: %p and len: 0x%zx sequence: %u\n",
  53                &mhi_buf->dma_addr, mhi_buf->len, sequence_id);
  54}
  55
  56/* Collect RDDM buffer during kernel panic */
  57static int __mhi_download_rddm_in_panic(struct mhi_controller *mhi_cntrl)
  58{
  59        int ret;
  60        u32 rx_status;
  61        enum mhi_ee_type ee;
  62        const u32 delayus = 2000;
  63        u32 retry = (mhi_cntrl->timeout_ms * 1000) / delayus;
  64        const u32 rddm_timeout_us = 200000;
  65        int rddm_retry = rddm_timeout_us / delayus;
  66        void __iomem *base = mhi_cntrl->bhie;
  67        struct device *dev = &mhi_cntrl->mhi_dev->dev;
  68
  69        dev_dbg(dev, "Entered with pm_state:%s dev_state:%s ee:%s\n",
  70                to_mhi_pm_state_str(mhi_cntrl->pm_state),
  71                TO_MHI_STATE_STR(mhi_cntrl->dev_state),
  72                TO_MHI_EXEC_STR(mhi_cntrl->ee));
  73
  74        /*
  75         * This should only be executing during a kernel panic, we expect all
  76         * other cores to shutdown while we're collecting RDDM buffer. After
  77         * returning from this function, we expect the device to reset.
  78         *
  79         * Normaly, we read/write pm_state only after grabbing the
  80         * pm_lock, since we're in a panic, skipping it. Also there is no
  81         * gurantee that this state change would take effect since
  82         * we're setting it w/o grabbing pm_lock
  83         */
  84        mhi_cntrl->pm_state = MHI_PM_LD_ERR_FATAL_DETECT;
  85        /* update should take the effect immediately */
  86        smp_wmb();
  87
  88        /*
  89         * Make sure device is not already in RDDM. In case the device asserts
  90         * and a kernel panic follows, device will already be in RDDM.
  91         * Do not trigger SYS ERR again and proceed with waiting for
  92         * image download completion.
  93         */
  94        ee = mhi_get_exec_env(mhi_cntrl);
  95        if (ee == MHI_EE_MAX)
  96                goto error_exit_rddm;
  97
  98        if (ee != MHI_EE_RDDM) {
  99                dev_dbg(dev, "Trigger device into RDDM mode using SYS ERR\n");
 100                mhi_set_mhi_state(mhi_cntrl, MHI_STATE_SYS_ERR);
 101
 102                dev_dbg(dev, "Waiting for device to enter RDDM\n");
 103                while (rddm_retry--) {
 104                        ee = mhi_get_exec_env(mhi_cntrl);
 105                        if (ee == MHI_EE_RDDM)
 106                                break;
 107
 108                        udelay(delayus);
 109                }
 110
 111                if (rddm_retry <= 0) {
 112                        /* Hardware reset so force device to enter RDDM */
 113                        dev_dbg(dev,
 114                                "Did not enter RDDM, do a host req reset\n");
 115                        mhi_write_reg(mhi_cntrl, mhi_cntrl->regs,
 116                                      MHI_SOC_RESET_REQ_OFFSET,
 117                                      MHI_SOC_RESET_REQ);
 118                        udelay(delayus);
 119                }
 120
 121                ee = mhi_get_exec_env(mhi_cntrl);
 122        }
 123
 124        dev_dbg(dev,
 125                "Waiting for RDDM image download via BHIe, current EE:%s\n",
 126                TO_MHI_EXEC_STR(ee));
 127
 128        while (retry--) {
 129                ret = mhi_read_reg_field(mhi_cntrl, base, BHIE_RXVECSTATUS_OFFS,
 130                                         BHIE_RXVECSTATUS_STATUS_BMSK,
 131                                         BHIE_RXVECSTATUS_STATUS_SHFT,
 132                                         &rx_status);
 133                if (ret)
 134                        return -EIO;
 135
 136                if (rx_status == BHIE_RXVECSTATUS_STATUS_XFER_COMPL)
 137                        return 0;
 138
 139                udelay(delayus);
 140        }
 141
 142        ee = mhi_get_exec_env(mhi_cntrl);
 143        ret = mhi_read_reg(mhi_cntrl, base, BHIE_RXVECSTATUS_OFFS, &rx_status);
 144
 145        dev_err(dev, "RXVEC_STATUS: 0x%x\n", rx_status);
 146
 147error_exit_rddm:
 148        dev_err(dev, "RDDM transfer failed. Current EE: %s\n",
 149                TO_MHI_EXEC_STR(ee));
 150
 151        return -EIO;
 152}
 153
 154/* Download RDDM image from device */
 155int mhi_download_rddm_image(struct mhi_controller *mhi_cntrl, bool in_panic)
 156{
 157        void __iomem *base = mhi_cntrl->bhie;
 158        struct device *dev = &mhi_cntrl->mhi_dev->dev;
 159        u32 rx_status;
 160
 161        if (in_panic)
 162                return __mhi_download_rddm_in_panic(mhi_cntrl);
 163
 164        dev_dbg(dev, "Waiting for RDDM image download via BHIe\n");
 165
 166        /* Wait for the image download to complete */
 167        wait_event_timeout(mhi_cntrl->state_event,
 168                           mhi_read_reg_field(mhi_cntrl, base,
 169                                              BHIE_RXVECSTATUS_OFFS,
 170                                              BHIE_RXVECSTATUS_STATUS_BMSK,
 171                                              BHIE_RXVECSTATUS_STATUS_SHFT,
 172                                              &rx_status) || rx_status,
 173                           msecs_to_jiffies(mhi_cntrl->timeout_ms));
 174
 175        return (rx_status == BHIE_RXVECSTATUS_STATUS_XFER_COMPL) ? 0 : -EIO;
 176}
 177EXPORT_SYMBOL_GPL(mhi_download_rddm_image);
 178
 179static int mhi_fw_load_bhie(struct mhi_controller *mhi_cntrl,
 180                            const struct mhi_buf *mhi_buf)
 181{
 182        void __iomem *base = mhi_cntrl->bhie;
 183        struct device *dev = &mhi_cntrl->mhi_dev->dev;
 184        rwlock_t *pm_lock = &mhi_cntrl->pm_lock;
 185        u32 tx_status, sequence_id;
 186        int ret;
 187
 188        read_lock_bh(pm_lock);
 189        if (!MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
 190                read_unlock_bh(pm_lock);
 191                return -EIO;
 192        }
 193
 194        sequence_id = MHI_RANDOM_U32_NONZERO(BHIE_TXVECSTATUS_SEQNUM_BMSK);
 195        dev_dbg(dev, "Starting image download via BHIe. Sequence ID: %u\n",
 196                sequence_id);
 197        mhi_write_reg(mhi_cntrl, base, BHIE_TXVECADDR_HIGH_OFFS,
 198                      upper_32_bits(mhi_buf->dma_addr));
 199
 200        mhi_write_reg(mhi_cntrl, base, BHIE_TXVECADDR_LOW_OFFS,
 201                      lower_32_bits(mhi_buf->dma_addr));
 202
 203        mhi_write_reg(mhi_cntrl, base, BHIE_TXVECSIZE_OFFS, mhi_buf->len);
 204
 205        mhi_write_reg_field(mhi_cntrl, base, BHIE_TXVECDB_OFFS,
 206                            BHIE_TXVECDB_SEQNUM_BMSK, BHIE_TXVECDB_SEQNUM_SHFT,
 207                            sequence_id);
 208        read_unlock_bh(pm_lock);
 209
 210        /* Wait for the image download to complete */
 211        ret = wait_event_timeout(mhi_cntrl->state_event,
 212                                 MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
 213                                 mhi_read_reg_field(mhi_cntrl, base,
 214                                                   BHIE_TXVECSTATUS_OFFS,
 215                                                   BHIE_TXVECSTATUS_STATUS_BMSK,
 216                                                   BHIE_TXVECSTATUS_STATUS_SHFT,
 217                                                   &tx_status) || tx_status,
 218                                 msecs_to_jiffies(mhi_cntrl->timeout_ms));
 219        if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
 220            tx_status != BHIE_TXVECSTATUS_STATUS_XFER_COMPL)
 221                return -EIO;
 222
 223        return (!ret) ? -ETIMEDOUT : 0;
 224}
 225
 226static int mhi_fw_load_bhi(struct mhi_controller *mhi_cntrl,
 227                           dma_addr_t dma_addr,
 228                           size_t size)
 229{
 230        u32 tx_status, val, session_id;
 231        int i, ret;
 232        void __iomem *base = mhi_cntrl->bhi;
 233        rwlock_t *pm_lock = &mhi_cntrl->pm_lock;
 234        struct device *dev = &mhi_cntrl->mhi_dev->dev;
 235        struct {
 236                char *name;
 237                u32 offset;
 238        } error_reg[] = {
 239                { "ERROR_CODE", BHI_ERRCODE },
 240                { "ERROR_DBG1", BHI_ERRDBG1 },
 241                { "ERROR_DBG2", BHI_ERRDBG2 },
 242                { "ERROR_DBG3", BHI_ERRDBG3 },
 243                { NULL },
 244        };
 245
 246        read_lock_bh(pm_lock);
 247        if (!MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
 248                read_unlock_bh(pm_lock);
 249                goto invalid_pm_state;
 250        }
 251
 252        session_id = MHI_RANDOM_U32_NONZERO(BHI_TXDB_SEQNUM_BMSK);
 253        dev_dbg(dev, "Starting image download via BHI. Session ID: %u\n",
 254                session_id);
 255        mhi_write_reg(mhi_cntrl, base, BHI_STATUS, 0);
 256        mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_HIGH,
 257                      upper_32_bits(dma_addr));
 258        mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_LOW,
 259                      lower_32_bits(dma_addr));
 260        mhi_write_reg(mhi_cntrl, base, BHI_IMGSIZE, size);
 261        mhi_write_reg(mhi_cntrl, base, BHI_IMGTXDB, session_id);
 262        read_unlock_bh(pm_lock);
 263
 264        /* Wait for the image download to complete */
 265        ret = wait_event_timeout(mhi_cntrl->state_event,
 266                           MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
 267                           mhi_read_reg_field(mhi_cntrl, base, BHI_STATUS,
 268                                              BHI_STATUS_MASK, BHI_STATUS_SHIFT,
 269                                              &tx_status) || tx_status,
 270                           msecs_to_jiffies(mhi_cntrl->timeout_ms));
 271        if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state))
 272                goto invalid_pm_state;
 273
 274        if (tx_status == BHI_STATUS_ERROR) {
 275                dev_err(dev, "Image transfer failed\n");
 276                read_lock_bh(pm_lock);
 277                if (MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
 278                        for (i = 0; error_reg[i].name; i++) {
 279                                ret = mhi_read_reg(mhi_cntrl, base,
 280                                                   error_reg[i].offset, &val);
 281                                if (ret)
 282                                        break;
 283                                dev_err(dev, "Reg: %s value: 0x%x\n",
 284                                        error_reg[i].name, val);
 285                        }
 286                }
 287                read_unlock_bh(pm_lock);
 288                goto invalid_pm_state;
 289        }
 290
 291        return (!ret) ? -ETIMEDOUT : 0;
 292
 293invalid_pm_state:
 294
 295        return -EIO;
 296}
 297
 298void mhi_free_bhie_table(struct mhi_controller *mhi_cntrl,
 299                         struct image_info *image_info)
 300{
 301        int i;
 302        struct mhi_buf *mhi_buf = image_info->mhi_buf;
 303
 304        for (i = 0; i < image_info->entries; i++, mhi_buf++)
 305                dma_free_coherent(mhi_cntrl->cntrl_dev, mhi_buf->len,
 306                                  mhi_buf->buf, mhi_buf->dma_addr);
 307
 308        kfree(image_info->mhi_buf);
 309        kfree(image_info);
 310}
 311
 312int mhi_alloc_bhie_table(struct mhi_controller *mhi_cntrl,
 313                         struct image_info **image_info,
 314                         size_t alloc_size)
 315{
 316        size_t seg_size = mhi_cntrl->seg_len;
 317        int segments = DIV_ROUND_UP(alloc_size, seg_size) + 1;
 318        int i;
 319        struct image_info *img_info;
 320        struct mhi_buf *mhi_buf;
 321
 322        img_info = kzalloc(sizeof(*img_info), GFP_KERNEL);
 323        if (!img_info)
 324                return -ENOMEM;
 325
 326        /* Allocate memory for entries */
 327        img_info->mhi_buf = kcalloc(segments, sizeof(*img_info->mhi_buf),
 328                                    GFP_KERNEL);
 329        if (!img_info->mhi_buf)
 330                goto error_alloc_mhi_buf;
 331
 332        /* Allocate and populate vector table */
 333        mhi_buf = img_info->mhi_buf;
 334        for (i = 0; i < segments; i++, mhi_buf++) {
 335                size_t vec_size = seg_size;
 336
 337                /* Vector table is the last entry */
 338                if (i == segments - 1)
 339                        vec_size = sizeof(struct bhi_vec_entry) * i;
 340
 341                mhi_buf->len = vec_size;
 342                mhi_buf->buf = dma_alloc_coherent(mhi_cntrl->cntrl_dev,
 343                                                  vec_size, &mhi_buf->dma_addr,
 344                                                  GFP_KERNEL);
 345                if (!mhi_buf->buf)
 346                        goto error_alloc_segment;
 347        }
 348
 349        img_info->bhi_vec = img_info->mhi_buf[segments - 1].buf;
 350        img_info->entries = segments;
 351        *image_info = img_info;
 352
 353        return 0;
 354
 355error_alloc_segment:
 356        for (--i, --mhi_buf; i >= 0; i--, mhi_buf--)
 357                dma_free_coherent(mhi_cntrl->cntrl_dev, mhi_buf->len,
 358                                  mhi_buf->buf, mhi_buf->dma_addr);
 359
 360error_alloc_mhi_buf:
 361        kfree(img_info);
 362
 363        return -ENOMEM;
 364}
 365
 366static void mhi_firmware_copy(struct mhi_controller *mhi_cntrl,
 367                              const struct firmware *firmware,
 368                              struct image_info *img_info)
 369{
 370        size_t remainder = firmware->size;
 371        size_t to_cpy;
 372        const u8 *buf = firmware->data;
 373        struct mhi_buf *mhi_buf = img_info->mhi_buf;
 374        struct bhi_vec_entry *bhi_vec = img_info->bhi_vec;
 375
 376        while (remainder) {
 377                to_cpy = min(remainder, mhi_buf->len);
 378                memcpy(mhi_buf->buf, buf, to_cpy);
 379                bhi_vec->dma_addr = mhi_buf->dma_addr;
 380                bhi_vec->size = to_cpy;
 381
 382                buf += to_cpy;
 383                remainder -= to_cpy;
 384                bhi_vec++;
 385                mhi_buf++;
 386        }
 387}
 388
 389void mhi_fw_load_handler(struct mhi_controller *mhi_cntrl)
 390{
 391        const struct firmware *firmware = NULL;
 392        struct device *dev = &mhi_cntrl->mhi_dev->dev;
 393        const char *fw_name;
 394        void *buf;
 395        dma_addr_t dma_addr;
 396        size_t size;
 397        int i, ret;
 398
 399        if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state)) {
 400                dev_err(dev, "Device MHI is not in valid state\n");
 401                return;
 402        }
 403
 404        /* save hardware info from BHI */
 405        ret = mhi_read_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_SERIALNU,
 406                           &mhi_cntrl->serial_number);
 407        if (ret)
 408                dev_err(dev, "Could not capture serial number via BHI\n");
 409
 410        for (i = 0; i < ARRAY_SIZE(mhi_cntrl->oem_pk_hash); i++) {
 411                ret = mhi_read_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_OEMPKHASH(i),
 412                                   &mhi_cntrl->oem_pk_hash[i]);
 413                if (ret) {
 414                        dev_err(dev, "Could not capture OEM PK HASH via BHI\n");
 415                        break;
 416                }
 417        }
 418
 419        /* wait for ready on pass through or any other execution environment */
 420        if (mhi_cntrl->ee != MHI_EE_EDL && mhi_cntrl->ee != MHI_EE_PBL)
 421                goto fw_load_ready_state;
 422
 423        fw_name = (mhi_cntrl->ee == MHI_EE_EDL) ?
 424                mhi_cntrl->edl_image : mhi_cntrl->fw_image;
 425
 426        if (!fw_name || (mhi_cntrl->fbc_download && (!mhi_cntrl->sbl_size ||
 427                                                     !mhi_cntrl->seg_len))) {
 428                dev_err(dev,
 429                        "No firmware image defined or !sbl_size || !seg_len\n");
 430                goto error_fw_load;
 431        }
 432
 433        ret = request_firmware(&firmware, fw_name, dev);
 434        if (ret) {
 435                dev_err(dev, "Error loading firmware: %d\n", ret);
 436                goto error_fw_load;
 437        }
 438
 439        size = (mhi_cntrl->fbc_download) ? mhi_cntrl->sbl_size : firmware->size;
 440
 441        /* SBL size provided is maximum size, not necessarily the image size */
 442        if (size > firmware->size)
 443                size = firmware->size;
 444
 445        buf = dma_alloc_coherent(mhi_cntrl->cntrl_dev, size, &dma_addr,
 446                                 GFP_KERNEL);
 447        if (!buf) {
 448                release_firmware(firmware);
 449                goto error_fw_load;
 450        }
 451
 452        /* Download image using BHI */
 453        memcpy(buf, firmware->data, size);
 454        ret = mhi_fw_load_bhi(mhi_cntrl, dma_addr, size);
 455        dma_free_coherent(mhi_cntrl->cntrl_dev, size, buf, dma_addr);
 456
 457        /* Error or in EDL mode, we're done */
 458        if (ret) {
 459                dev_err(dev, "MHI did not load image over BHI, ret: %d\n", ret);
 460                release_firmware(firmware);
 461                goto error_fw_load;
 462        }
 463
 464        /* Wait for ready since EDL image was loaded */
 465        if (fw_name == mhi_cntrl->edl_image) {
 466                release_firmware(firmware);
 467                goto fw_load_ready_state;
 468        }
 469
 470        write_lock_irq(&mhi_cntrl->pm_lock);
 471        mhi_cntrl->dev_state = MHI_STATE_RESET;
 472        write_unlock_irq(&mhi_cntrl->pm_lock);
 473
 474        /*
 475         * If we're doing fbc, populate vector tables while
 476         * device transitioning into MHI READY state
 477         */
 478        if (mhi_cntrl->fbc_download) {
 479                ret = mhi_alloc_bhie_table(mhi_cntrl, &mhi_cntrl->fbc_image,
 480                                           firmware->size);
 481                if (ret) {
 482                        release_firmware(firmware);
 483                        goto error_fw_load;
 484                }
 485
 486                /* Load the firmware into BHIE vec table */
 487                mhi_firmware_copy(mhi_cntrl, firmware, mhi_cntrl->fbc_image);
 488        }
 489
 490        release_firmware(firmware);
 491
 492fw_load_ready_state:
 493        /* Transitioning into MHI RESET->READY state */
 494        ret = mhi_ready_state_transition(mhi_cntrl);
 495        if (ret) {
 496                dev_err(dev, "MHI did not enter READY state\n");
 497                goto error_ready_state;
 498        }
 499
 500        dev_info(dev, "Wait for device to enter SBL or Mission mode\n");
 501        return;
 502
 503error_ready_state:
 504        if (mhi_cntrl->fbc_download) {
 505                mhi_free_bhie_table(mhi_cntrl, mhi_cntrl->fbc_image);
 506                mhi_cntrl->fbc_image = NULL;
 507        }
 508
 509error_fw_load:
 510        mhi_cntrl->pm_state = MHI_PM_FW_DL_ERR;
 511        wake_up_all(&mhi_cntrl->state_event);
 512}
 513
 514int mhi_download_amss_image(struct mhi_controller *mhi_cntrl)
 515{
 516        struct image_info *image_info = mhi_cntrl->fbc_image;
 517        struct device *dev = &mhi_cntrl->mhi_dev->dev;
 518        int ret;
 519
 520        if (!image_info)
 521                return -EIO;
 522
 523        ret = mhi_fw_load_bhie(mhi_cntrl,
 524                               /* Vector table is the last entry */
 525                               &image_info->mhi_buf[image_info->entries - 1]);
 526        if (ret) {
 527                dev_err(dev, "MHI did not load AMSS, ret:%d\n", ret);
 528                mhi_cntrl->pm_state = MHI_PM_FW_DL_ERR;
 529                wake_up_all(&mhi_cntrl->state_event);
 530        }
 531
 532        return ret;
 533}
 534