linux/drivers/edac/cpc925_edac.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-only
   2/*
   3 * cpc925_edac.c, EDAC driver for IBM CPC925 Bridge and Memory Controller.
   4 *
   5 * Copyright (c) 2008 Wind River Systems, Inc.
   6 *
   7 * Authors:     Cao Qingtao <qingtao.cao@windriver.com>
   8 */
   9
  10#include <linux/module.h>
  11#include <linux/init.h>
  12#include <linux/io.h>
  13#include <linux/edac.h>
  14#include <linux/of.h>
  15#include <linux/platform_device.h>
  16#include <linux/gfp.h>
  17
  18#include "edac_module.h"
  19
  20#define CPC925_EDAC_REVISION    " Ver: 1.0.0"
  21#define CPC925_EDAC_MOD_STR     "cpc925_edac"
  22
  23#define cpc925_printk(level, fmt, arg...) \
  24        edac_printk(level, "CPC925", fmt, ##arg)
  25
  26#define cpc925_mc_printk(mci, level, fmt, arg...) \
  27        edac_mc_chipset_printk(mci, level, "CPC925", fmt, ##arg)
  28
  29/*
  30 * CPC925 registers are of 32 bits with bit0 defined at the
  31 * most significant bit and bit31 at that of least significant.
  32 */
  33#define CPC925_BITS_PER_REG     32
  34#define CPC925_BIT(nr)          (1UL << (CPC925_BITS_PER_REG - 1 - nr))
  35
  36/*
  37 * EDAC device names for the error detections of
  38 * CPU Interface and Hypertransport Link.
  39 */
  40#define CPC925_CPU_ERR_DEV      "cpu"
  41#define CPC925_HT_LINK_DEV      "htlink"
  42
  43/* Suppose DDR Refresh cycle is 15.6 microsecond */
  44#define CPC925_REF_FREQ         0xFA69
  45#define CPC925_SCRUB_BLOCK_SIZE 64      /* bytes */
  46#define CPC925_NR_CSROWS        8
  47
  48/*
  49 * All registers and bits definitions are taken from
  50 * "CPC925 Bridge and Memory Controller User Manual, SA14-2761-02".
  51 */
  52
  53/*
  54 * CPU and Memory Controller Registers
  55 */
  56/************************************************************
  57 *      Processor Interface Exception Mask Register (APIMASK)
  58 ************************************************************/
  59#define REG_APIMASK_OFFSET      0x30070
  60enum apimask_bits {
  61        APIMASK_DART    = CPC925_BIT(0), /* DART Exception */
  62        APIMASK_ADI0    = CPC925_BIT(1), /* Handshake Error on PI0_ADI */
  63        APIMASK_ADI1    = CPC925_BIT(2), /* Handshake Error on PI1_ADI */
  64        APIMASK_STAT    = CPC925_BIT(3), /* Status Exception */
  65        APIMASK_DERR    = CPC925_BIT(4), /* Data Error Exception */
  66        APIMASK_ADRS0   = CPC925_BIT(5), /* Addressing Exception on PI0 */
  67        APIMASK_ADRS1   = CPC925_BIT(6), /* Addressing Exception on PI1 */
  68                                         /* BIT(7) Reserved */
  69        APIMASK_ECC_UE_H = CPC925_BIT(8), /* UECC upper */
  70        APIMASK_ECC_CE_H = CPC925_BIT(9), /* CECC upper */
  71        APIMASK_ECC_UE_L = CPC925_BIT(10), /* UECC lower */
  72        APIMASK_ECC_CE_L = CPC925_BIT(11), /* CECC lower */
  73
  74        CPU_MASK_ENABLE = (APIMASK_DART | APIMASK_ADI0 | APIMASK_ADI1 |
  75                           APIMASK_STAT | APIMASK_DERR | APIMASK_ADRS0 |
  76                           APIMASK_ADRS1),
  77        ECC_MASK_ENABLE = (APIMASK_ECC_UE_H | APIMASK_ECC_CE_H |
  78                           APIMASK_ECC_UE_L | APIMASK_ECC_CE_L),
  79};
  80#define APIMASK_ADI(n)          CPC925_BIT(((n)+1))
  81
  82/************************************************************
  83 *      Processor Interface Exception Register (APIEXCP)
  84 ************************************************************/
  85#define REG_APIEXCP_OFFSET      0x30060
  86enum apiexcp_bits {
  87        APIEXCP_DART    = CPC925_BIT(0), /* DART Exception */
  88        APIEXCP_ADI0    = CPC925_BIT(1), /* Handshake Error on PI0_ADI */
  89        APIEXCP_ADI1    = CPC925_BIT(2), /* Handshake Error on PI1_ADI */
  90        APIEXCP_STAT    = CPC925_BIT(3), /* Status Exception */
  91        APIEXCP_DERR    = CPC925_BIT(4), /* Data Error Exception */
  92        APIEXCP_ADRS0   = CPC925_BIT(5), /* Addressing Exception on PI0 */
  93        APIEXCP_ADRS1   = CPC925_BIT(6), /* Addressing Exception on PI1 */
  94                                         /* BIT(7) Reserved */
  95        APIEXCP_ECC_UE_H = CPC925_BIT(8), /* UECC upper */
  96        APIEXCP_ECC_CE_H = CPC925_BIT(9), /* CECC upper */
  97        APIEXCP_ECC_UE_L = CPC925_BIT(10), /* UECC lower */
  98        APIEXCP_ECC_CE_L = CPC925_BIT(11), /* CECC lower */
  99
 100        CPU_EXCP_DETECTED = (APIEXCP_DART | APIEXCP_ADI0 | APIEXCP_ADI1 |
 101                             APIEXCP_STAT | APIEXCP_DERR | APIEXCP_ADRS0 |
 102                             APIEXCP_ADRS1),
 103        UECC_EXCP_DETECTED = (APIEXCP_ECC_UE_H | APIEXCP_ECC_UE_L),
 104        CECC_EXCP_DETECTED = (APIEXCP_ECC_CE_H | APIEXCP_ECC_CE_L),
 105        ECC_EXCP_DETECTED = (UECC_EXCP_DETECTED | CECC_EXCP_DETECTED),
 106};
 107
 108/************************************************************
 109 *      Memory Bus Configuration Register (MBCR)
 110************************************************************/
 111#define REG_MBCR_OFFSET         0x2190
 112#define MBCR_64BITCFG_SHIFT     23
 113#define MBCR_64BITCFG_MASK      (1UL << MBCR_64BITCFG_SHIFT)
 114#define MBCR_64BITBUS_SHIFT     22
 115#define MBCR_64BITBUS_MASK      (1UL << MBCR_64BITBUS_SHIFT)
 116
 117/************************************************************
 118 *      Memory Bank Mode Register (MBMR)
 119************************************************************/
 120#define REG_MBMR_OFFSET         0x21C0
 121#define MBMR_MODE_MAX_VALUE     0xF
 122#define MBMR_MODE_SHIFT         25
 123#define MBMR_MODE_MASK          (MBMR_MODE_MAX_VALUE << MBMR_MODE_SHIFT)
 124#define MBMR_BBA_SHIFT          24
 125#define MBMR_BBA_MASK           (1UL << MBMR_BBA_SHIFT)
 126
 127/************************************************************
 128 *      Memory Bank Boundary Address Register (MBBAR)
 129 ************************************************************/
 130#define REG_MBBAR_OFFSET        0x21D0
 131#define MBBAR_BBA_MAX_VALUE     0xFF
 132#define MBBAR_BBA_SHIFT         24
 133#define MBBAR_BBA_MASK          (MBBAR_BBA_MAX_VALUE << MBBAR_BBA_SHIFT)
 134
 135/************************************************************
 136 *      Memory Scrub Control Register (MSCR)
 137 ************************************************************/
 138#define REG_MSCR_OFFSET         0x2400
 139#define MSCR_SCRUB_MOD_MASK     0xC0000000 /* scrub_mod - bit0:1*/
 140#define MSCR_BACKGR_SCRUB       0x40000000 /* 01 */
 141#define MSCR_SI_SHIFT           16      /* si - bit8:15*/
 142#define MSCR_SI_MAX_VALUE       0xFF
 143#define MSCR_SI_MASK            (MSCR_SI_MAX_VALUE << MSCR_SI_SHIFT)
 144
 145/************************************************************
 146 *      Memory Scrub Range Start Register (MSRSR)
 147 ************************************************************/
 148#define REG_MSRSR_OFFSET        0x2410
 149
 150/************************************************************
 151 *      Memory Scrub Range End Register (MSRER)
 152 ************************************************************/
 153#define REG_MSRER_OFFSET        0x2420
 154
 155/************************************************************
 156 *      Memory Scrub Pattern Register (MSPR)
 157 ************************************************************/
 158#define REG_MSPR_OFFSET         0x2430
 159
 160/************************************************************
 161 *      Memory Check Control Register (MCCR)
 162 ************************************************************/
 163#define REG_MCCR_OFFSET         0x2440
 164enum mccr_bits {
 165        MCCR_ECC_EN     = CPC925_BIT(0), /* ECC high and low check */
 166};
 167
 168/************************************************************
 169 *      Memory Check Range End Register (MCRER)
 170 ************************************************************/
 171#define REG_MCRER_OFFSET        0x2450
 172
 173/************************************************************
 174 *      Memory Error Address Register (MEAR)
 175 ************************************************************/
 176#define REG_MEAR_OFFSET         0x2460
 177#define MEAR_BCNT_MAX_VALUE     0x3
 178#define MEAR_BCNT_SHIFT         30
 179#define MEAR_BCNT_MASK          (MEAR_BCNT_MAX_VALUE << MEAR_BCNT_SHIFT)
 180#define MEAR_RANK_MAX_VALUE     0x7
 181#define MEAR_RANK_SHIFT         27
 182#define MEAR_RANK_MASK          (MEAR_RANK_MAX_VALUE << MEAR_RANK_SHIFT)
 183#define MEAR_COL_MAX_VALUE      0x7FF
 184#define MEAR_COL_SHIFT          16
 185#define MEAR_COL_MASK           (MEAR_COL_MAX_VALUE << MEAR_COL_SHIFT)
 186#define MEAR_BANK_MAX_VALUE     0x3
 187#define MEAR_BANK_SHIFT         14
 188#define MEAR_BANK_MASK          (MEAR_BANK_MAX_VALUE << MEAR_BANK_SHIFT)
 189#define MEAR_ROW_MASK           0x00003FFF
 190
 191/************************************************************
 192 *      Memory Error Syndrome Register (MESR)
 193 ************************************************************/
 194#define REG_MESR_OFFSET         0x2470
 195#define MESR_ECC_SYN_H_MASK     0xFF00
 196#define MESR_ECC_SYN_L_MASK     0x00FF
 197
 198/************************************************************
 199 *      Memory Mode Control Register (MMCR)
 200 ************************************************************/
 201#define REG_MMCR_OFFSET         0x2500
 202enum mmcr_bits {
 203        MMCR_REG_DIMM_MODE = CPC925_BIT(3),
 204};
 205
 206/*
 207 * HyperTransport Link Registers
 208 */
 209/************************************************************
 210 *  Error Handling/Enumeration Scratch Pad Register (ERRCTRL)
 211 ************************************************************/
 212#define REG_ERRCTRL_OFFSET      0x70140
 213enum errctrl_bits {                      /* nonfatal interrupts for */
 214        ERRCTRL_SERR_NF = CPC925_BIT(0), /* system error */
 215        ERRCTRL_CRC_NF  = CPC925_BIT(1), /* CRC error */
 216        ERRCTRL_RSP_NF  = CPC925_BIT(2), /* Response error */
 217        ERRCTRL_EOC_NF  = CPC925_BIT(3), /* End-Of-Chain error */
 218        ERRCTRL_OVF_NF  = CPC925_BIT(4), /* Overflow error */
 219        ERRCTRL_PROT_NF = CPC925_BIT(5), /* Protocol error */
 220
 221        ERRCTRL_RSP_ERR = CPC925_BIT(6), /* Response error received */
 222        ERRCTRL_CHN_FAL = CPC925_BIT(7), /* Sync flooding detected */
 223
 224        HT_ERRCTRL_ENABLE = (ERRCTRL_SERR_NF | ERRCTRL_CRC_NF |
 225                             ERRCTRL_RSP_NF | ERRCTRL_EOC_NF |
 226                             ERRCTRL_OVF_NF | ERRCTRL_PROT_NF),
 227        HT_ERRCTRL_DETECTED = (ERRCTRL_RSP_ERR | ERRCTRL_CHN_FAL),
 228};
 229
 230/************************************************************
 231 *  Link Configuration and Link Control Register (LINKCTRL)
 232 ************************************************************/
 233#define REG_LINKCTRL_OFFSET     0x70110
 234enum linkctrl_bits {
 235        LINKCTRL_CRC_ERR        = (CPC925_BIT(22) | CPC925_BIT(23)),
 236        LINKCTRL_LINK_FAIL      = CPC925_BIT(27),
 237
 238        HT_LINKCTRL_DETECTED    = (LINKCTRL_CRC_ERR | LINKCTRL_LINK_FAIL),
 239};
 240
 241/************************************************************
 242 *  Link FreqCap/Error/Freq/Revision ID Register (LINKERR)
 243 ************************************************************/
 244#define REG_LINKERR_OFFSET      0x70120
 245enum linkerr_bits {
 246        LINKERR_EOC_ERR         = CPC925_BIT(17), /* End-Of-Chain error */
 247        LINKERR_OVF_ERR         = CPC925_BIT(18), /* Receive Buffer Overflow */
 248        LINKERR_PROT_ERR        = CPC925_BIT(19), /* Protocol error */
 249
 250        HT_LINKERR_DETECTED     = (LINKERR_EOC_ERR | LINKERR_OVF_ERR |
 251                                   LINKERR_PROT_ERR),
 252};
 253
 254/************************************************************
 255 *      Bridge Control Register (BRGCTRL)
 256 ************************************************************/
 257#define REG_BRGCTRL_OFFSET      0x70300
 258enum brgctrl_bits {
 259        BRGCTRL_DETSERR = CPC925_BIT(0), /* SERR on Secondary Bus */
 260        BRGCTRL_SECBUSRESET = CPC925_BIT(9), /* Secondary Bus Reset */
 261};
 262
 263/* Private structure for edac memory controller */
 264struct cpc925_mc_pdata {
 265        void __iomem *vbase;
 266        unsigned long total_mem;
 267        const char *name;
 268        int edac_idx;
 269};
 270
 271/* Private structure for common edac device */
 272struct cpc925_dev_info {
 273        void __iomem *vbase;
 274        struct platform_device *pdev;
 275        char *ctl_name;
 276        int edac_idx;
 277        struct edac_device_ctl_info *edac_dev;
 278        void (*init)(struct cpc925_dev_info *dev_info);
 279        void (*exit)(struct cpc925_dev_info *dev_info);
 280        void (*check)(struct edac_device_ctl_info *edac_dev);
 281};
 282
 283/* Get total memory size from Open Firmware DTB */
 284static void get_total_mem(struct cpc925_mc_pdata *pdata)
 285{
 286        struct device_node *np = NULL;
 287        const unsigned int *reg, *reg_end;
 288        int len, sw, aw;
 289        unsigned long start, size;
 290
 291        np = of_find_node_by_type(NULL, "memory");
 292        if (!np)
 293                return;
 294
 295        aw = of_n_addr_cells(np);
 296        sw = of_n_size_cells(np);
 297        reg = (const unsigned int *)of_get_property(np, "reg", &len);
 298        reg_end = reg + len/4;
 299
 300        pdata->total_mem = 0;
 301        do {
 302                start = of_read_number(reg, aw);
 303                reg += aw;
 304                size = of_read_number(reg, sw);
 305                reg += sw;
 306                edac_dbg(1, "start 0x%lx, size 0x%lx\n", start, size);
 307                pdata->total_mem += size;
 308        } while (reg < reg_end);
 309
 310        of_node_put(np);
 311        edac_dbg(0, "total_mem 0x%lx\n", pdata->total_mem);
 312}
 313
 314static void cpc925_init_csrows(struct mem_ctl_info *mci)
 315{
 316        struct cpc925_mc_pdata *pdata = mci->pvt_info;
 317        struct csrow_info *csrow;
 318        struct dimm_info *dimm;
 319        enum dev_type dtype;
 320        int index, j;
 321        u32 mbmr, mbbar, bba, grain;
 322        unsigned long row_size, nr_pages, last_nr_pages = 0;
 323
 324        get_total_mem(pdata);
 325
 326        for (index = 0; index < mci->nr_csrows; index++) {
 327                mbmr = __raw_readl(pdata->vbase + REG_MBMR_OFFSET +
 328                                   0x20 * index);
 329                mbbar = __raw_readl(pdata->vbase + REG_MBBAR_OFFSET +
 330                                   0x20 + index);
 331                bba = (((mbmr & MBMR_BBA_MASK) >> MBMR_BBA_SHIFT) << 8) |
 332                       ((mbbar & MBBAR_BBA_MASK) >> MBBAR_BBA_SHIFT);
 333
 334                if (bba == 0)
 335                        continue; /* not populated */
 336
 337                csrow = mci->csrows[index];
 338
 339                row_size = bba * (1UL << 28);   /* 256M */
 340                csrow->first_page = last_nr_pages;
 341                nr_pages = row_size >> PAGE_SHIFT;
 342                csrow->last_page = csrow->first_page + nr_pages - 1;
 343                last_nr_pages = csrow->last_page + 1;
 344
 345                switch (csrow->nr_channels) {
 346                case 1: /* Single channel */
 347                        grain = 32; /* four-beat burst of 32 bytes */
 348                        break;
 349                case 2: /* Dual channel */
 350                default:
 351                        grain = 64; /* four-beat burst of 64 bytes */
 352                        break;
 353                }
 354                switch ((mbmr & MBMR_MODE_MASK) >> MBMR_MODE_SHIFT) {
 355                case 6: /* 0110, no way to differentiate X8 VS X16 */
 356                case 5: /* 0101 */
 357                case 8: /* 1000 */
 358                        dtype = DEV_X16;
 359                        break;
 360                case 7: /* 0111 */
 361                case 9: /* 1001 */
 362                        dtype = DEV_X8;
 363                        break;
 364                default:
 365                        dtype = DEV_UNKNOWN;
 366                break;
 367                }
 368                for (j = 0; j < csrow->nr_channels; j++) {
 369                        dimm = csrow->channels[j]->dimm;
 370                        dimm->nr_pages = nr_pages / csrow->nr_channels;
 371                        dimm->mtype = MEM_RDDR;
 372                        dimm->edac_mode = EDAC_SECDED;
 373                        dimm->grain = grain;
 374                        dimm->dtype = dtype;
 375                }
 376        }
 377}
 378
 379/* Enable memory controller ECC detection */
 380static void cpc925_mc_init(struct mem_ctl_info *mci)
 381{
 382        struct cpc925_mc_pdata *pdata = mci->pvt_info;
 383        u32 apimask;
 384        u32 mccr;
 385
 386        /* Enable various ECC error exceptions */
 387        apimask = __raw_readl(pdata->vbase + REG_APIMASK_OFFSET);
 388        if ((apimask & ECC_MASK_ENABLE) == 0) {
 389                apimask |= ECC_MASK_ENABLE;
 390                __raw_writel(apimask, pdata->vbase + REG_APIMASK_OFFSET);
 391        }
 392
 393        /* Enable ECC detection */
 394        mccr = __raw_readl(pdata->vbase + REG_MCCR_OFFSET);
 395        if ((mccr & MCCR_ECC_EN) == 0) {
 396                mccr |= MCCR_ECC_EN;
 397                __raw_writel(mccr, pdata->vbase + REG_MCCR_OFFSET);
 398        }
 399}
 400
 401/* Disable memory controller ECC detection */
 402static void cpc925_mc_exit(struct mem_ctl_info *mci)
 403{
 404        /*
 405         * WARNING:
 406         * We are supposed to clear the ECC error detection bits,
 407         * and it will be no problem to do so. However, once they
 408         * are cleared here if we want to re-install CPC925 EDAC
 409         * module later, setting them up in cpc925_mc_init() will
 410         * trigger machine check exception.
 411         * Also, it's ok to leave ECC error detection bits enabled,
 412         * since they are reset to 1 by default or by boot loader.
 413         */
 414
 415        return;
 416}
 417
 418/*
 419 * Revert DDR column/row/bank addresses into page frame number and
 420 * offset in page.
 421 *
 422 * Suppose memory mode is 0x0111(128-bit mode, identical DIMM pairs),
 423 * physical address(PA) bits to column address(CA) bits mappings are:
 424 * CA   0   1   2   3   4   5   6   7   8   9   10
 425 * PA   59  58  57  56  55  54  53  52  51  50  49
 426 *
 427 * physical address(PA) bits to bank address(BA) bits mappings are:
 428 * BA   0   1
 429 * PA   43  44
 430 *
 431 * physical address(PA) bits to row address(RA) bits mappings are:
 432 * RA   0   1   2   3   4   5   6   7   8   9   10   11   12
 433 * PA   36  35  34  48  47  46  45  40  41  42  39   38   37
 434 */
 435static void cpc925_mc_get_pfn(struct mem_ctl_info *mci, u32 mear,
 436                unsigned long *pfn, unsigned long *offset, int *csrow)
 437{
 438        u32 bcnt, rank, col, bank, row;
 439        u32 c;
 440        unsigned long pa;
 441        int i;
 442
 443        bcnt = (mear & MEAR_BCNT_MASK) >> MEAR_BCNT_SHIFT;
 444        rank = (mear & MEAR_RANK_MASK) >> MEAR_RANK_SHIFT;
 445        col = (mear & MEAR_COL_MASK) >> MEAR_COL_SHIFT;
 446        bank = (mear & MEAR_BANK_MASK) >> MEAR_BANK_SHIFT;
 447        row = mear & MEAR_ROW_MASK;
 448
 449        *csrow = rank;
 450
 451#ifdef CONFIG_EDAC_DEBUG
 452        if (mci->csrows[rank]->first_page == 0) {
 453                cpc925_mc_printk(mci, KERN_ERR, "ECC occurs in a "
 454                        "non-populated csrow, broken hardware?\n");
 455                return;
 456        }
 457#endif
 458
 459        /* Revert csrow number */
 460        pa = mci->csrows[rank]->first_page << PAGE_SHIFT;
 461
 462        /* Revert column address */
 463        col += bcnt;
 464        for (i = 0; i < 11; i++) {
 465                c = col & 0x1;
 466                col >>= 1;
 467                pa |= c << (14 - i);
 468        }
 469
 470        /* Revert bank address */
 471        pa |= bank << 19;
 472
 473        /* Revert row address, in 4 steps */
 474        for (i = 0; i < 3; i++) {
 475                c = row & 0x1;
 476                row >>= 1;
 477                pa |= c << (26 - i);
 478        }
 479
 480        for (i = 0; i < 3; i++) {
 481                c = row & 0x1;
 482                row >>= 1;
 483                pa |= c << (21 + i);
 484        }
 485
 486        for (i = 0; i < 4; i++) {
 487                c = row & 0x1;
 488                row >>= 1;
 489                pa |= c << (18 - i);
 490        }
 491
 492        for (i = 0; i < 3; i++) {
 493                c = row & 0x1;
 494                row >>= 1;
 495                pa |= c << (29 - i);
 496        }
 497
 498        *offset = pa & (PAGE_SIZE - 1);
 499        *pfn = pa >> PAGE_SHIFT;
 500
 501        edac_dbg(0, "ECC physical address 0x%lx\n", pa);
 502}
 503
 504static int cpc925_mc_find_channel(struct mem_ctl_info *mci, u16 syndrome)
 505{
 506        if ((syndrome & MESR_ECC_SYN_H_MASK) == 0)
 507                return 0;
 508
 509        if ((syndrome & MESR_ECC_SYN_L_MASK) == 0)
 510                return 1;
 511
 512        cpc925_mc_printk(mci, KERN_INFO, "Unexpected syndrome value: 0x%x\n",
 513                         syndrome);
 514        return 1;
 515}
 516
 517/* Check memory controller registers for ECC errors */
 518static void cpc925_mc_check(struct mem_ctl_info *mci)
 519{
 520        struct cpc925_mc_pdata *pdata = mci->pvt_info;
 521        u32 apiexcp;
 522        u32 mear;
 523        u32 mesr;
 524        u16 syndrome;
 525        unsigned long pfn = 0, offset = 0;
 526        int csrow = 0, channel = 0;
 527
 528        /* APIEXCP is cleared when read */
 529        apiexcp = __raw_readl(pdata->vbase + REG_APIEXCP_OFFSET);
 530        if ((apiexcp & ECC_EXCP_DETECTED) == 0)
 531                return;
 532
 533        mesr = __raw_readl(pdata->vbase + REG_MESR_OFFSET);
 534        syndrome = mesr | (MESR_ECC_SYN_H_MASK | MESR_ECC_SYN_L_MASK);
 535
 536        mear = __raw_readl(pdata->vbase + REG_MEAR_OFFSET);
 537
 538        /* Revert column/row addresses into page frame number, etc */
 539        cpc925_mc_get_pfn(mci, mear, &pfn, &offset, &csrow);
 540
 541        if (apiexcp & CECC_EXCP_DETECTED) {
 542                cpc925_mc_printk(mci, KERN_INFO, "DRAM CECC Fault\n");
 543                channel = cpc925_mc_find_channel(mci, syndrome);
 544                edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1,
 545                                     pfn, offset, syndrome,
 546                                     csrow, channel, -1,
 547                                     mci->ctl_name, "");
 548        }
 549
 550        if (apiexcp & UECC_EXCP_DETECTED) {
 551                cpc925_mc_printk(mci, KERN_INFO, "DRAM UECC Fault\n");
 552                edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
 553                                     pfn, offset, 0,
 554                                     csrow, -1, -1,
 555                                     mci->ctl_name, "");
 556        }
 557
 558        cpc925_mc_printk(mci, KERN_INFO, "Dump registers:\n");
 559        cpc925_mc_printk(mci, KERN_INFO, "APIMASK               0x%08x\n",
 560                __raw_readl(pdata->vbase + REG_APIMASK_OFFSET));
 561        cpc925_mc_printk(mci, KERN_INFO, "APIEXCP               0x%08x\n",
 562                apiexcp);
 563        cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Ctrl        0x%08x\n",
 564                __raw_readl(pdata->vbase + REG_MSCR_OFFSET));
 565        cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Rge Start   0x%08x\n",
 566                __raw_readl(pdata->vbase + REG_MSRSR_OFFSET));
 567        cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Rge End     0x%08x\n",
 568                __raw_readl(pdata->vbase + REG_MSRER_OFFSET));
 569        cpc925_mc_printk(mci, KERN_INFO, "Mem Scrub Pattern     0x%08x\n",
 570                __raw_readl(pdata->vbase + REG_MSPR_OFFSET));
 571        cpc925_mc_printk(mci, KERN_INFO, "Mem Chk Ctrl          0x%08x\n",
 572                __raw_readl(pdata->vbase + REG_MCCR_OFFSET));
 573        cpc925_mc_printk(mci, KERN_INFO, "Mem Chk Rge End       0x%08x\n",
 574                __raw_readl(pdata->vbase + REG_MCRER_OFFSET));
 575        cpc925_mc_printk(mci, KERN_INFO, "Mem Err Address       0x%08x\n",
 576                mesr);
 577        cpc925_mc_printk(mci, KERN_INFO, "Mem Err Syndrome      0x%08x\n",
 578                syndrome);
 579}
 580
 581/******************** CPU err device********************************/
 582static u32 cpc925_cpu_mask_disabled(void)
 583{
 584        struct device_node *cpunode;
 585        static u32 mask = 0;
 586
 587        /* use cached value if available */
 588        if (mask != 0)
 589                return mask;
 590
 591        mask = APIMASK_ADI0 | APIMASK_ADI1;
 592
 593        for_each_of_cpu_node(cpunode) {
 594                const u32 *reg = of_get_property(cpunode, "reg", NULL);
 595                if (reg == NULL || *reg > 2) {
 596                        cpc925_printk(KERN_ERR, "Bad reg value at %pOF\n", cpunode);
 597                        continue;
 598                }
 599
 600                mask &= ~APIMASK_ADI(*reg);
 601        }
 602
 603        if (mask != (APIMASK_ADI0 | APIMASK_ADI1)) {
 604                /* We assume that each CPU sits on it's own PI and that
 605                 * for present CPUs the reg property equals to the PI
 606                 * interface id */
 607                cpc925_printk(KERN_WARNING,
 608                                "Assuming PI id is equal to CPU MPIC id!\n");
 609        }
 610
 611        return mask;
 612}
 613
 614/* Enable CPU Errors detection */
 615static void cpc925_cpu_init(struct cpc925_dev_info *dev_info)
 616{
 617        u32 apimask;
 618        u32 cpumask;
 619
 620        apimask = __raw_readl(dev_info->vbase + REG_APIMASK_OFFSET);
 621
 622        cpumask = cpc925_cpu_mask_disabled();
 623        if (apimask & cpumask) {
 624                cpc925_printk(KERN_WARNING, "CPU(s) not present, "
 625                                "but enabled in APIMASK, disabling\n");
 626                apimask &= ~cpumask;
 627        }
 628
 629        if ((apimask & CPU_MASK_ENABLE) == 0)
 630                apimask |= CPU_MASK_ENABLE;
 631
 632        __raw_writel(apimask, dev_info->vbase + REG_APIMASK_OFFSET);
 633}
 634
 635/* Disable CPU Errors detection */
 636static void cpc925_cpu_exit(struct cpc925_dev_info *dev_info)
 637{
 638        /*
 639         * WARNING:
 640         * We are supposed to clear the CPU error detection bits,
 641         * and it will be no problem to do so. However, once they
 642         * are cleared here if we want to re-install CPC925 EDAC
 643         * module later, setting them up in cpc925_cpu_init() will
 644         * trigger machine check exception.
 645         * Also, it's ok to leave CPU error detection bits enabled,
 646         * since they are reset to 1 by default.
 647         */
 648
 649        return;
 650}
 651
 652/* Check for CPU Errors */
 653static void cpc925_cpu_check(struct edac_device_ctl_info *edac_dev)
 654{
 655        struct cpc925_dev_info *dev_info = edac_dev->pvt_info;
 656        u32 apiexcp;
 657        u32 apimask;
 658
 659        /* APIEXCP is cleared when read */
 660        apiexcp = __raw_readl(dev_info->vbase + REG_APIEXCP_OFFSET);
 661        if ((apiexcp & CPU_EXCP_DETECTED) == 0)
 662                return;
 663
 664        if ((apiexcp & ~cpc925_cpu_mask_disabled()) == 0)
 665                return;
 666
 667        apimask = __raw_readl(dev_info->vbase + REG_APIMASK_OFFSET);
 668        cpc925_printk(KERN_INFO, "Processor Interface Fault\n"
 669                                 "Processor Interface register dump:\n");
 670        cpc925_printk(KERN_INFO, "APIMASK               0x%08x\n", apimask);
 671        cpc925_printk(KERN_INFO, "APIEXCP               0x%08x\n", apiexcp);
 672
 673        edac_device_handle_ue(edac_dev, 0, 0, edac_dev->ctl_name);
 674}
 675
 676/******************** HT Link err device****************************/
 677/* Enable HyperTransport Link Error detection */
 678static void cpc925_htlink_init(struct cpc925_dev_info *dev_info)
 679{
 680        u32 ht_errctrl;
 681
 682        ht_errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
 683        if ((ht_errctrl & HT_ERRCTRL_ENABLE) == 0) {
 684                ht_errctrl |= HT_ERRCTRL_ENABLE;
 685                __raw_writel(ht_errctrl, dev_info->vbase + REG_ERRCTRL_OFFSET);
 686        }
 687}
 688
 689/* Disable HyperTransport Link Error detection */
 690static void cpc925_htlink_exit(struct cpc925_dev_info *dev_info)
 691{
 692        u32 ht_errctrl;
 693
 694        ht_errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
 695        ht_errctrl &= ~HT_ERRCTRL_ENABLE;
 696        __raw_writel(ht_errctrl, dev_info->vbase + REG_ERRCTRL_OFFSET);
 697}
 698
 699/* Check for HyperTransport Link errors */
 700static void cpc925_htlink_check(struct edac_device_ctl_info *edac_dev)
 701{
 702        struct cpc925_dev_info *dev_info = edac_dev->pvt_info;
 703        u32 brgctrl = __raw_readl(dev_info->vbase + REG_BRGCTRL_OFFSET);
 704        u32 linkctrl = __raw_readl(dev_info->vbase + REG_LINKCTRL_OFFSET);
 705        u32 errctrl = __raw_readl(dev_info->vbase + REG_ERRCTRL_OFFSET);
 706        u32 linkerr = __raw_readl(dev_info->vbase + REG_LINKERR_OFFSET);
 707
 708        if (!((brgctrl & BRGCTRL_DETSERR) ||
 709              (linkctrl & HT_LINKCTRL_DETECTED) ||
 710              (errctrl & HT_ERRCTRL_DETECTED) ||
 711              (linkerr & HT_LINKERR_DETECTED)))
 712                return;
 713
 714        cpc925_printk(KERN_INFO, "HT Link Fault\n"
 715                                 "HT register dump:\n");
 716        cpc925_printk(KERN_INFO, "Bridge Ctrl                   0x%08x\n",
 717                      brgctrl);
 718        cpc925_printk(KERN_INFO, "Link Config Ctrl              0x%08x\n",
 719                      linkctrl);
 720        cpc925_printk(KERN_INFO, "Error Enum and Ctrl           0x%08x\n",
 721                      errctrl);
 722        cpc925_printk(KERN_INFO, "Link Error                    0x%08x\n",
 723                      linkerr);
 724
 725        /* Clear by write 1 */
 726        if (brgctrl & BRGCTRL_DETSERR)
 727                __raw_writel(BRGCTRL_DETSERR,
 728                                dev_info->vbase + REG_BRGCTRL_OFFSET);
 729
 730        if (linkctrl & HT_LINKCTRL_DETECTED)
 731                __raw_writel(HT_LINKCTRL_DETECTED,
 732                                dev_info->vbase + REG_LINKCTRL_OFFSET);
 733
 734        /* Initiate Secondary Bus Reset to clear the chain failure */
 735        if (errctrl & ERRCTRL_CHN_FAL)
 736                __raw_writel(BRGCTRL_SECBUSRESET,
 737                                dev_info->vbase + REG_BRGCTRL_OFFSET);
 738
 739        if (errctrl & ERRCTRL_RSP_ERR)
 740                __raw_writel(ERRCTRL_RSP_ERR,
 741                                dev_info->vbase + REG_ERRCTRL_OFFSET);
 742
 743        if (linkerr & HT_LINKERR_DETECTED)
 744                __raw_writel(HT_LINKERR_DETECTED,
 745                                dev_info->vbase + REG_LINKERR_OFFSET);
 746
 747        edac_device_handle_ce(edac_dev, 0, 0, edac_dev->ctl_name);
 748}
 749
 750static struct cpc925_dev_info cpc925_devs[] = {
 751        {
 752        .ctl_name = CPC925_CPU_ERR_DEV,
 753        .init = cpc925_cpu_init,
 754        .exit = cpc925_cpu_exit,
 755        .check = cpc925_cpu_check,
 756        },
 757        {
 758        .ctl_name = CPC925_HT_LINK_DEV,
 759        .init = cpc925_htlink_init,
 760        .exit = cpc925_htlink_exit,
 761        .check = cpc925_htlink_check,
 762        },
 763        { }
 764};
 765
 766/*
 767 * Add CPU Err detection and HyperTransport Link Err detection
 768 * as common "edac_device", they have no corresponding device
 769 * nodes in the Open Firmware DTB and we have to add platform
 770 * devices for them. Also, they will share the MMIO with that
 771 * of memory controller.
 772 */
 773static void cpc925_add_edac_devices(void __iomem *vbase)
 774{
 775        struct cpc925_dev_info *dev_info;
 776
 777        if (!vbase) {
 778                cpc925_printk(KERN_ERR, "MMIO not established yet\n");
 779                return;
 780        }
 781
 782        for (dev_info = &cpc925_devs[0]; dev_info->init; dev_info++) {
 783                dev_info->vbase = vbase;
 784                dev_info->pdev = platform_device_register_simple(
 785                                        dev_info->ctl_name, 0, NULL, 0);
 786                if (IS_ERR(dev_info->pdev)) {
 787                        cpc925_printk(KERN_ERR,
 788                                "Can't register platform device for %s\n",
 789                                dev_info->ctl_name);
 790                        continue;
 791                }
 792
 793                /*
 794                 * Don't have to allocate private structure but
 795                 * make use of cpc925_devs[] instead.
 796                 */
 797                dev_info->edac_idx = edac_device_alloc_index();
 798                dev_info->edac_dev =
 799                        edac_device_alloc_ctl_info(0, dev_info->ctl_name,
 800                                1, NULL, 0, 0, NULL, 0, dev_info->edac_idx);
 801                if (!dev_info->edac_dev) {
 802                        cpc925_printk(KERN_ERR, "No memory for edac device\n");
 803                        goto err1;
 804                }
 805
 806                dev_info->edac_dev->pvt_info = dev_info;
 807                dev_info->edac_dev->dev = &dev_info->pdev->dev;
 808                dev_info->edac_dev->ctl_name = dev_info->ctl_name;
 809                dev_info->edac_dev->mod_name = CPC925_EDAC_MOD_STR;
 810                dev_info->edac_dev->dev_name = dev_name(&dev_info->pdev->dev);
 811
 812                if (edac_op_state == EDAC_OPSTATE_POLL)
 813                        dev_info->edac_dev->edac_check = dev_info->check;
 814
 815                if (dev_info->init)
 816                        dev_info->init(dev_info);
 817
 818                if (edac_device_add_device(dev_info->edac_dev) > 0) {
 819                        cpc925_printk(KERN_ERR,
 820                                "Unable to add edac device for %s\n",
 821                                dev_info->ctl_name);
 822                        goto err2;
 823                }
 824
 825                edac_dbg(0, "Successfully added edac device for %s\n",
 826                         dev_info->ctl_name);
 827
 828                continue;
 829
 830err2:
 831                if (dev_info->exit)
 832                        dev_info->exit(dev_info);
 833                edac_device_free_ctl_info(dev_info->edac_dev);
 834err1:
 835                platform_device_unregister(dev_info->pdev);
 836        }
 837}
 838
 839/*
 840 * Delete the common "edac_device" for CPU Err Detection
 841 * and HyperTransport Link Err Detection
 842 */
 843static void cpc925_del_edac_devices(void)
 844{
 845        struct cpc925_dev_info *dev_info;
 846
 847        for (dev_info = &cpc925_devs[0]; dev_info->init; dev_info++) {
 848                if (dev_info->edac_dev) {
 849                        edac_device_del_device(dev_info->edac_dev->dev);
 850                        edac_device_free_ctl_info(dev_info->edac_dev);
 851                        platform_device_unregister(dev_info->pdev);
 852                }
 853
 854                if (dev_info->exit)
 855                        dev_info->exit(dev_info);
 856
 857                edac_dbg(0, "Successfully deleted edac device for %s\n",
 858                         dev_info->ctl_name);
 859        }
 860}
 861
 862/* Convert current back-ground scrub rate into byte/sec bandwidth */
 863static int cpc925_get_sdram_scrub_rate(struct mem_ctl_info *mci)
 864{
 865        struct cpc925_mc_pdata *pdata = mci->pvt_info;
 866        int bw;
 867        u32 mscr;
 868        u8 si;
 869
 870        mscr = __raw_readl(pdata->vbase + REG_MSCR_OFFSET);
 871        si = (mscr & MSCR_SI_MASK) >> MSCR_SI_SHIFT;
 872
 873        edac_dbg(0, "Mem Scrub Ctrl Register 0x%x\n", mscr);
 874
 875        if (((mscr & MSCR_SCRUB_MOD_MASK) != MSCR_BACKGR_SCRUB) ||
 876            (si == 0)) {
 877                cpc925_mc_printk(mci, KERN_INFO, "Scrub mode not enabled\n");
 878                bw = 0;
 879        } else
 880                bw = CPC925_SCRUB_BLOCK_SIZE * 0xFA67 / si;
 881
 882        return bw;
 883}
 884
 885/* Return 0 for single channel; 1 for dual channel */
 886static int cpc925_mc_get_channels(void __iomem *vbase)
 887{
 888        int dual = 0;
 889        u32 mbcr;
 890
 891        mbcr = __raw_readl(vbase + REG_MBCR_OFFSET);
 892
 893        /*
 894         * Dual channel only when 128-bit wide physical bus
 895         * and 128-bit configuration.
 896         */
 897        if (((mbcr & MBCR_64BITCFG_MASK) == 0) &&
 898            ((mbcr & MBCR_64BITBUS_MASK) == 0))
 899                dual = 1;
 900
 901        edac_dbg(0, "%s channel\n", (dual > 0) ? "Dual" : "Single");
 902
 903        return dual;
 904}
 905
 906static int cpc925_probe(struct platform_device *pdev)
 907{
 908        static int edac_mc_idx;
 909        struct mem_ctl_info *mci;
 910        struct edac_mc_layer layers[2];
 911        void __iomem *vbase;
 912        struct cpc925_mc_pdata *pdata;
 913        struct resource *r;
 914        int res = 0, nr_channels;
 915
 916        edac_dbg(0, "%s platform device found!\n", pdev->name);
 917
 918        if (!devres_open_group(&pdev->dev, cpc925_probe, GFP_KERNEL)) {
 919                res = -ENOMEM;
 920                goto out;
 921        }
 922
 923        r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
 924        if (!r) {
 925                cpc925_printk(KERN_ERR, "Unable to get resource\n");
 926                res = -ENOENT;
 927                goto err1;
 928        }
 929
 930        if (!devm_request_mem_region(&pdev->dev,
 931                                     r->start,
 932                                     resource_size(r),
 933                                     pdev->name)) {
 934                cpc925_printk(KERN_ERR, "Unable to request mem region\n");
 935                res = -EBUSY;
 936                goto err1;
 937        }
 938
 939        vbase = devm_ioremap(&pdev->dev, r->start, resource_size(r));
 940        if (!vbase) {
 941                cpc925_printk(KERN_ERR, "Unable to ioremap device\n");
 942                res = -ENOMEM;
 943                goto err2;
 944        }
 945
 946        nr_channels = cpc925_mc_get_channels(vbase) + 1;
 947
 948        layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
 949        layers[0].size = CPC925_NR_CSROWS;
 950        layers[0].is_virt_csrow = true;
 951        layers[1].type = EDAC_MC_LAYER_CHANNEL;
 952        layers[1].size = nr_channels;
 953        layers[1].is_virt_csrow = false;
 954        mci = edac_mc_alloc(edac_mc_idx, ARRAY_SIZE(layers), layers,
 955                            sizeof(struct cpc925_mc_pdata));
 956        if (!mci) {
 957                cpc925_printk(KERN_ERR, "No memory for mem_ctl_info\n");
 958                res = -ENOMEM;
 959                goto err2;
 960        }
 961
 962        pdata = mci->pvt_info;
 963        pdata->vbase = vbase;
 964        pdata->edac_idx = edac_mc_idx++;
 965        pdata->name = pdev->name;
 966
 967        mci->pdev = &pdev->dev;
 968        platform_set_drvdata(pdev, mci);
 969        mci->dev_name = dev_name(&pdev->dev);
 970        mci->mtype_cap = MEM_FLAG_RDDR | MEM_FLAG_DDR;
 971        mci->edac_ctl_cap = EDAC_FLAG_NONE | EDAC_FLAG_SECDED;
 972        mci->edac_cap = EDAC_FLAG_SECDED;
 973        mci->mod_name = CPC925_EDAC_MOD_STR;
 974        mci->ctl_name = pdev->name;
 975
 976        if (edac_op_state == EDAC_OPSTATE_POLL)
 977                mci->edac_check = cpc925_mc_check;
 978
 979        mci->ctl_page_to_phys = NULL;
 980        mci->scrub_mode = SCRUB_SW_SRC;
 981        mci->set_sdram_scrub_rate = NULL;
 982        mci->get_sdram_scrub_rate = cpc925_get_sdram_scrub_rate;
 983
 984        cpc925_init_csrows(mci);
 985
 986        /* Setup memory controller registers */
 987        cpc925_mc_init(mci);
 988
 989        if (edac_mc_add_mc(mci) > 0) {
 990                cpc925_mc_printk(mci, KERN_ERR, "Failed edac_mc_add_mc()\n");
 991                goto err3;
 992        }
 993
 994        cpc925_add_edac_devices(vbase);
 995
 996        /* get this far and it's successful */
 997        edac_dbg(0, "success\n");
 998
 999        res = 0;
1000        goto out;
1001
1002err3:
1003        cpc925_mc_exit(mci);
1004        edac_mc_free(mci);
1005err2:
1006        devm_release_mem_region(&pdev->dev, r->start, resource_size(r));
1007err1:
1008        devres_release_group(&pdev->dev, cpc925_probe);
1009out:
1010        return res;
1011}
1012
1013static int cpc925_remove(struct platform_device *pdev)
1014{
1015        struct mem_ctl_info *mci = platform_get_drvdata(pdev);
1016
1017        /*
1018         * Delete common edac devices before edac mc, because
1019         * the former share the MMIO of the latter.
1020         */
1021        cpc925_del_edac_devices();
1022        cpc925_mc_exit(mci);
1023
1024        edac_mc_del_mc(&pdev->dev);
1025        edac_mc_free(mci);
1026
1027        return 0;
1028}
1029
1030static struct platform_driver cpc925_edac_driver = {
1031        .probe = cpc925_probe,
1032        .remove = cpc925_remove,
1033        .driver = {
1034                   .name = "cpc925_edac",
1035        }
1036};
1037
1038static int __init cpc925_edac_init(void)
1039{
1040        int ret = 0;
1041
1042        printk(KERN_INFO "IBM CPC925 EDAC driver " CPC925_EDAC_REVISION "\n");
1043        printk(KERN_INFO "\t(c) 2008 Wind River Systems, Inc\n");
1044
1045        /* Only support POLL mode so far */
1046        edac_op_state = EDAC_OPSTATE_POLL;
1047
1048        ret = platform_driver_register(&cpc925_edac_driver);
1049        if (ret) {
1050                printk(KERN_WARNING "Failed to register %s\n",
1051                        CPC925_EDAC_MOD_STR);
1052        }
1053
1054        return ret;
1055}
1056
1057static void __exit cpc925_edac_exit(void)
1058{
1059        platform_driver_unregister(&cpc925_edac_driver);
1060}
1061
1062module_init(cpc925_edac_init);
1063module_exit(cpc925_edac_exit);
1064
1065MODULE_LICENSE("GPL");
1066MODULE_AUTHOR("Cao Qingtao <qingtao.cao@windriver.com>");
1067MODULE_DESCRIPTION("IBM CPC925 Bridge and MC EDAC kernel module");
1068