linux/drivers/ata/sata_sil.c
<<
>>
Prefs
   1/*
   2 *  sata_sil.c - Silicon Image SATA
   3 *
   4 *  Maintained by:  Tejun Heo <tj@kernel.org>
   5 *                  Please ALWAYS copy linux-ide@vger.kernel.org
   6 *                  on emails.
   7 *
   8 *  Copyright 2003-2005 Red Hat, Inc.
   9 *  Copyright 2003 Benjamin Herrenschmidt
  10 *
  11 *
  12 *  This program is free software; you can redistribute it and/or modify
  13 *  it under the terms of the GNU General Public License as published by
  14 *  the Free Software Foundation; either version 2, or (at your option)
  15 *  any later version.
  16 *
  17 *  This program is distributed in the hope that it will be useful,
  18 *  but WITHOUT ANY WARRANTY; without even the implied warranty of
  19 *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
  20 *  GNU General Public License for more details.
  21 *
  22 *  You should have received a copy of the GNU General Public License
  23 *  along with this program; see the file COPYING.  If not, write to
  24 *  the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
  25 *
  26 *
  27 *  libata documentation is available via 'make {ps|pdf}docs',
  28 *  as Documentation/DocBook/libata.*
  29 *
  30 *  Documentation for SiI 3112:
  31 *  http://gkernel.sourceforge.net/specs/sii/3112A_SiI-DS-0095-B2.pdf.bz2
  32 *
  33 *  Other errata and documentation available under NDA.
  34 *
  35 */
  36
  37#include <linux/kernel.h>
  38#include <linux/module.h>
  39#include <linux/pci.h>
  40#include <linux/blkdev.h>
  41#include <linux/delay.h>
  42#include <linux/interrupt.h>
  43#include <linux/device.h>
  44#include <scsi/scsi_host.h>
  45#include <linux/libata.h>
  46#include <linux/dmi.h>
  47
  48#define DRV_NAME        "sata_sil"
  49#define DRV_VERSION     "2.4"
  50
  51#define SIL_DMA_BOUNDARY        0x7fffffffUL
  52
  53enum {
  54        SIL_MMIO_BAR            = 5,
  55
  56        /*
  57         * host flags
  58         */
  59        SIL_FLAG_NO_SATA_IRQ    = (1 << 28),
  60        SIL_FLAG_RERR_ON_DMA_ACT = (1 << 29),
  61        SIL_FLAG_MOD15WRITE     = (1 << 30),
  62
  63        SIL_DFL_PORT_FLAGS      = ATA_FLAG_SATA,
  64
  65        /*
  66         * Controller IDs
  67         */
  68        sil_3112                = 0,
  69        sil_3112_no_sata_irq    = 1,
  70        sil_3512                = 2,
  71        sil_3114                = 3,
  72
  73        /*
  74         * Register offsets
  75         */
  76        SIL_SYSCFG              = 0x48,
  77
  78        /*
  79         * Register bits
  80         */
  81        /* SYSCFG */
  82        SIL_MASK_IDE0_INT       = (1 << 22),
  83        SIL_MASK_IDE1_INT       = (1 << 23),
  84        SIL_MASK_IDE2_INT       = (1 << 24),
  85        SIL_MASK_IDE3_INT       = (1 << 25),
  86        SIL_MASK_2PORT          = SIL_MASK_IDE0_INT | SIL_MASK_IDE1_INT,
  87        SIL_MASK_4PORT          = SIL_MASK_2PORT |
  88                                  SIL_MASK_IDE2_INT | SIL_MASK_IDE3_INT,
  89
  90        /* BMDMA/BMDMA2 */
  91        SIL_INTR_STEERING       = (1 << 1),
  92
  93        SIL_DMA_ENABLE          = (1 << 0),  /* DMA run switch */
  94        SIL_DMA_RDWR            = (1 << 3),  /* DMA Rd-Wr */
  95        SIL_DMA_SATA_IRQ        = (1 << 4),  /* OR of all SATA IRQs */
  96        SIL_DMA_ACTIVE          = (1 << 16), /* DMA running */
  97        SIL_DMA_ERROR           = (1 << 17), /* PCI bus error */
  98        SIL_DMA_COMPLETE        = (1 << 18), /* cmd complete / IRQ pending */
  99        SIL_DMA_N_SATA_IRQ      = (1 << 6),  /* SATA_IRQ for the next channel */
 100        SIL_DMA_N_ACTIVE        = (1 << 24), /* ACTIVE for the next channel */
 101        SIL_DMA_N_ERROR         = (1 << 25), /* ERROR for the next channel */
 102        SIL_DMA_N_COMPLETE      = (1 << 26), /* COMPLETE for the next channel */
 103
 104        /* SIEN */
 105        SIL_SIEN_N              = (1 << 16), /* triggered by SError.N */
 106
 107        /*
 108         * Others
 109         */
 110        SIL_QUIRK_MOD15WRITE    = (1 << 0),
 111        SIL_QUIRK_UDMA5MAX      = (1 << 1),
 112};
 113
 114static int sil_init_one(struct pci_dev *pdev, const struct pci_device_id *ent);
 115#ifdef CONFIG_PM_SLEEP
 116static int sil_pci_device_resume(struct pci_dev *pdev);
 117#endif
 118static void sil_dev_config(struct ata_device *dev);
 119static int sil_scr_read(struct ata_link *link, unsigned int sc_reg, u32 *val);
 120static int sil_scr_write(struct ata_link *link, unsigned int sc_reg, u32 val);
 121static int sil_set_mode(struct ata_link *link, struct ata_device **r_failed);
 122static void sil_qc_prep(struct ata_queued_cmd *qc);
 123static void sil_bmdma_setup(struct ata_queued_cmd *qc);
 124static void sil_bmdma_start(struct ata_queued_cmd *qc);
 125static void sil_bmdma_stop(struct ata_queued_cmd *qc);
 126static void sil_freeze(struct ata_port *ap);
 127static void sil_thaw(struct ata_port *ap);
 128
 129
 130static const struct pci_device_id sil_pci_tbl[] = {
 131        { PCI_VDEVICE(CMD, 0x3112), sil_3112 },
 132        { PCI_VDEVICE(CMD, 0x0240), sil_3112 },
 133        { PCI_VDEVICE(CMD, 0x3512), sil_3512 },
 134        { PCI_VDEVICE(CMD, 0x3114), sil_3114 },
 135        { PCI_VDEVICE(ATI, 0x436e), sil_3112 },
 136        { PCI_VDEVICE(ATI, 0x4379), sil_3112_no_sata_irq },
 137        { PCI_VDEVICE(ATI, 0x437a), sil_3112_no_sata_irq },
 138
 139        { }     /* terminate list */
 140};
 141
 142
 143/* TODO firmware versions should be added - eric */
 144static const struct sil_drivelist {
 145        const char *product;
 146        unsigned int quirk;
 147} sil_blacklist [] = {
 148        { "ST320012AS",         SIL_QUIRK_MOD15WRITE },
 149        { "ST330013AS",         SIL_QUIRK_MOD15WRITE },
 150        { "ST340017AS",         SIL_QUIRK_MOD15WRITE },
 151        { "ST360015AS",         SIL_QUIRK_MOD15WRITE },
 152        { "ST380023AS",         SIL_QUIRK_MOD15WRITE },
 153        { "ST3120023AS",        SIL_QUIRK_MOD15WRITE },
 154        { "ST340014ASL",        SIL_QUIRK_MOD15WRITE },
 155        { "ST360014ASL",        SIL_QUIRK_MOD15WRITE },
 156        { "ST380011ASL",        SIL_QUIRK_MOD15WRITE },
 157        { "ST3120022ASL",       SIL_QUIRK_MOD15WRITE },
 158        { "ST3160021ASL",       SIL_QUIRK_MOD15WRITE },
 159        { "TOSHIBA MK2561GSYN", SIL_QUIRK_MOD15WRITE },
 160        { "Maxtor 4D060H3",     SIL_QUIRK_UDMA5MAX },
 161        { }
 162};
 163
 164static struct pci_driver sil_pci_driver = {
 165        .name                   = DRV_NAME,
 166        .id_table               = sil_pci_tbl,
 167        .probe                  = sil_init_one,
 168        .remove                 = ata_pci_remove_one,
 169#ifdef CONFIG_PM_SLEEP
 170        .suspend                = ata_pci_device_suspend,
 171        .resume                 = sil_pci_device_resume,
 172#endif
 173};
 174
 175static struct scsi_host_template sil_sht = {
 176        ATA_BASE_SHT(DRV_NAME),
 177        /** These controllers support Large Block Transfer which allows
 178            transfer chunks up to 2GB and which cross 64KB boundaries,
 179            therefore the DMA limits are more relaxed than standard ATA SFF. */
 180        .dma_boundary           = SIL_DMA_BOUNDARY,
 181        .sg_tablesize           = ATA_MAX_PRD
 182};
 183
 184static struct ata_port_operations sil_ops = {
 185        .inherits               = &ata_bmdma32_port_ops,
 186        .dev_config             = sil_dev_config,
 187        .set_mode               = sil_set_mode,
 188        .bmdma_setup            = sil_bmdma_setup,
 189        .bmdma_start            = sil_bmdma_start,
 190        .bmdma_stop             = sil_bmdma_stop,
 191        .qc_prep                = sil_qc_prep,
 192        .freeze                 = sil_freeze,
 193        .thaw                   = sil_thaw,
 194        .scr_read               = sil_scr_read,
 195        .scr_write              = sil_scr_write,
 196};
 197
 198static const struct ata_port_info sil_port_info[] = {
 199        /* sil_3112 */
 200        {
 201                .flags          = SIL_DFL_PORT_FLAGS | SIL_FLAG_MOD15WRITE,
 202                .pio_mask       = ATA_PIO4,
 203                .mwdma_mask     = ATA_MWDMA2,
 204                .udma_mask      = ATA_UDMA5,
 205                .port_ops       = &sil_ops,
 206        },
 207        /* sil_3112_no_sata_irq */
 208        {
 209                .flags          = SIL_DFL_PORT_FLAGS | SIL_FLAG_MOD15WRITE |
 210                                  SIL_FLAG_NO_SATA_IRQ,
 211                .pio_mask       = ATA_PIO4,
 212                .mwdma_mask     = ATA_MWDMA2,
 213                .udma_mask      = ATA_UDMA5,
 214                .port_ops       = &sil_ops,
 215        },
 216        /* sil_3512 */
 217        {
 218                .flags          = SIL_DFL_PORT_FLAGS | SIL_FLAG_RERR_ON_DMA_ACT,
 219                .pio_mask       = ATA_PIO4,
 220                .mwdma_mask     = ATA_MWDMA2,
 221                .udma_mask      = ATA_UDMA5,
 222                .port_ops       = &sil_ops,
 223        },
 224        /* sil_3114 */
 225        {
 226                .flags          = SIL_DFL_PORT_FLAGS | SIL_FLAG_RERR_ON_DMA_ACT,
 227                .pio_mask       = ATA_PIO4,
 228                .mwdma_mask     = ATA_MWDMA2,
 229                .udma_mask      = ATA_UDMA5,
 230                .port_ops       = &sil_ops,
 231        },
 232};
 233
 234/* per-port register offsets */
 235/* TODO: we can probably calculate rather than use a table */
 236static const struct {
 237        unsigned long tf;       /* ATA taskfile register block */
 238        unsigned long ctl;      /* ATA control/altstatus register block */
 239        unsigned long bmdma;    /* DMA register block */
 240        unsigned long bmdma2;   /* DMA register block #2 */
 241        unsigned long fifo_cfg; /* FIFO Valid Byte Count and Control */
 242        unsigned long scr;      /* SATA control register block */
 243        unsigned long sien;     /* SATA Interrupt Enable register */
 244        unsigned long xfer_mode;/* data transfer mode register */
 245        unsigned long sfis_cfg; /* SATA FIS reception config register */
 246} sil_port[] = {
 247        /* port 0 ... */
 248        /*   tf    ctl  bmdma  bmdma2  fifo    scr   sien   mode   sfis */
 249        {  0x80,  0x8A,   0x0,  0x10,  0x40, 0x100, 0x148,  0xb4, 0x14c },
 250        {  0xC0,  0xCA,   0x8,  0x18,  0x44, 0x180, 0x1c8,  0xf4, 0x1cc },
 251        { 0x280, 0x28A, 0x200, 0x210, 0x240, 0x300, 0x348, 0x2b4, 0x34c },
 252        { 0x2C0, 0x2CA, 0x208, 0x218, 0x244, 0x380, 0x3c8, 0x2f4, 0x3cc },
 253        /* ... port 3 */
 254};
 255
 256MODULE_AUTHOR("Jeff Garzik");
 257MODULE_DESCRIPTION("low-level driver for Silicon Image SATA controller");
 258MODULE_LICENSE("GPL");
 259MODULE_DEVICE_TABLE(pci, sil_pci_tbl);
 260MODULE_VERSION(DRV_VERSION);
 261
 262static int slow_down;
 263module_param(slow_down, int, 0444);
 264MODULE_PARM_DESC(slow_down, "Sledgehammer used to work around random problems, by limiting commands to 15 sectors (0=off, 1=on)");
 265
 266
 267static void sil_bmdma_stop(struct ata_queued_cmd *qc)
 268{
 269        struct ata_port *ap = qc->ap;
 270        void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
 271        void __iomem *bmdma2 = mmio_base + sil_port[ap->port_no].bmdma2;
 272
 273        /* clear start/stop bit - can safely always write 0 */
 274        iowrite8(0, bmdma2);
 275
 276        /* one-PIO-cycle guaranteed wait, per spec, for HDMA1:0 transition */
 277        ata_sff_dma_pause(ap);
 278}
 279
 280static void sil_bmdma_setup(struct ata_queued_cmd *qc)
 281{
 282        struct ata_port *ap = qc->ap;
 283        void __iomem *bmdma = ap->ioaddr.bmdma_addr;
 284
 285        /* load PRD table addr. */
 286        iowrite32(ap->bmdma_prd_dma, bmdma + ATA_DMA_TABLE_OFS);
 287
 288        /* issue r/w command */
 289        ap->ops->sff_exec_command(ap, &qc->tf);
 290}
 291
 292static void sil_bmdma_start(struct ata_queued_cmd *qc)
 293{
 294        unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
 295        struct ata_port *ap = qc->ap;
 296        void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
 297        void __iomem *bmdma2 = mmio_base + sil_port[ap->port_no].bmdma2;
 298        u8 dmactl = ATA_DMA_START;
 299
 300        /* set transfer direction, start host DMA transaction
 301           Note: For Large Block Transfer to work, the DMA must be started
 302           using the bmdma2 register. */
 303        if (!rw)
 304                dmactl |= ATA_DMA_WR;
 305        iowrite8(dmactl, bmdma2);
 306}
 307
 308/* The way God intended PCI IDE scatter/gather lists to look and behave... */
 309static void sil_fill_sg(struct ata_queued_cmd *qc)
 310{
 311        struct scatterlist *sg;
 312        struct ata_port *ap = qc->ap;
 313        struct ata_bmdma_prd *prd, *last_prd = NULL;
 314        unsigned int si;
 315
 316        prd = &ap->bmdma_prd[0];
 317        for_each_sg(qc->sg, sg, qc->n_elem, si) {
 318                /* Note h/w doesn't support 64-bit, so we unconditionally
 319                 * truncate dma_addr_t to u32.
 320                 */
 321                u32 addr = (u32) sg_dma_address(sg);
 322                u32 sg_len = sg_dma_len(sg);
 323
 324                prd->addr = cpu_to_le32(addr);
 325                prd->flags_len = cpu_to_le32(sg_len);
 326                VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", si, addr, sg_len);
 327
 328                last_prd = prd;
 329                prd++;
 330        }
 331
 332        if (likely(last_prd))
 333                last_prd->flags_len |= cpu_to_le32(ATA_PRD_EOT);
 334}
 335
 336static void sil_qc_prep(struct ata_queued_cmd *qc)
 337{
 338        if (!(qc->flags & ATA_QCFLAG_DMAMAP))
 339                return;
 340
 341        sil_fill_sg(qc);
 342}
 343
 344static unsigned char sil_get_device_cache_line(struct pci_dev *pdev)
 345{
 346        u8 cache_line = 0;
 347        pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, &cache_line);
 348        return cache_line;
 349}
 350
 351/**
 352 *      sil_set_mode            -       wrap set_mode functions
 353 *      @link: link to set up
 354 *      @r_failed: returned device when we fail
 355 *
 356 *      Wrap the libata method for device setup as after the setup we need
 357 *      to inspect the results and do some configuration work
 358 */
 359
 360static int sil_set_mode(struct ata_link *link, struct ata_device **r_failed)
 361{
 362        struct ata_port *ap = link->ap;
 363        void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
 364        void __iomem *addr = mmio_base + sil_port[ap->port_no].xfer_mode;
 365        struct ata_device *dev;
 366        u32 tmp, dev_mode[2] = { };
 367        int rc;
 368
 369        rc = ata_do_set_mode(link, r_failed);
 370        if (rc)
 371                return rc;
 372
 373        ata_for_each_dev(dev, link, ALL) {
 374                if (!ata_dev_enabled(dev))
 375                        dev_mode[dev->devno] = 0;       /* PIO0/1/2 */
 376                else if (dev->flags & ATA_DFLAG_PIO)
 377                        dev_mode[dev->devno] = 1;       /* PIO3/4 */
 378                else
 379                        dev_mode[dev->devno] = 3;       /* UDMA */
 380                /* value 2 indicates MDMA */
 381        }
 382
 383        tmp = readl(addr);
 384        tmp &= ~((1<<5) | (1<<4) | (1<<1) | (1<<0));
 385        tmp |= dev_mode[0];
 386        tmp |= (dev_mode[1] << 4);
 387        writel(tmp, addr);
 388        readl(addr);    /* flush */
 389        return 0;
 390}
 391
 392static inline void __iomem *sil_scr_addr(struct ata_port *ap,
 393                                         unsigned int sc_reg)
 394{
 395        void __iomem *offset = ap->ioaddr.scr_addr;
 396
 397        switch (sc_reg) {
 398        case SCR_STATUS:
 399                return offset + 4;
 400        case SCR_ERROR:
 401                return offset + 8;
 402        case SCR_CONTROL:
 403                return offset;
 404        default:
 405                /* do nothing */
 406                break;
 407        }
 408
 409        return NULL;
 410}
 411
 412static int sil_scr_read(struct ata_link *link, unsigned int sc_reg, u32 *val)
 413{
 414        void __iomem *mmio = sil_scr_addr(link->ap, sc_reg);
 415
 416        if (mmio) {
 417                *val = readl(mmio);
 418                return 0;
 419        }
 420        return -EINVAL;
 421}
 422
 423static int sil_scr_write(struct ata_link *link, unsigned int sc_reg, u32 val)
 424{
 425        void __iomem *mmio = sil_scr_addr(link->ap, sc_reg);
 426
 427        if (mmio) {
 428                writel(val, mmio);
 429                return 0;
 430        }
 431        return -EINVAL;
 432}
 433
 434static void sil_host_intr(struct ata_port *ap, u32 bmdma2)
 435{
 436        struct ata_eh_info *ehi = &ap->link.eh_info;
 437        struct ata_queued_cmd *qc = ata_qc_from_tag(ap, ap->link.active_tag);
 438        u8 status;
 439
 440        if (unlikely(bmdma2 & SIL_DMA_SATA_IRQ)) {
 441                u32 serror = 0xffffffff;
 442
 443                /* SIEN doesn't mask SATA IRQs on some 3112s.  Those
 444                 * controllers continue to assert IRQ as long as
 445                 * SError bits are pending.  Clear SError immediately.
 446                 */
 447                sil_scr_read(&ap->link, SCR_ERROR, &serror);
 448                sil_scr_write(&ap->link, SCR_ERROR, serror);
 449
 450                /* Sometimes spurious interrupts occur, double check
 451                 * it's PHYRDY CHG.
 452                 */
 453                if (serror & SERR_PHYRDY_CHG) {
 454                        ap->link.eh_info.serror |= serror;
 455                        goto freeze;
 456                }
 457
 458                if (!(bmdma2 & SIL_DMA_COMPLETE))
 459                        return;
 460        }
 461
 462        if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) {
 463                /* this sometimes happens, just clear IRQ */
 464                ap->ops->sff_check_status(ap);
 465                return;
 466        }
 467
 468        /* Check whether we are expecting interrupt in this state */
 469        switch (ap->hsm_task_state) {
 470        case HSM_ST_FIRST:
 471                /* Some pre-ATAPI-4 devices assert INTRQ
 472                 * at this state when ready to receive CDB.
 473                 */
 474
 475                /* Check the ATA_DFLAG_CDB_INTR flag is enough here.
 476                 * The flag was turned on only for atapi devices.  No
 477                 * need to check ata_is_atapi(qc->tf.protocol) again.
 478                 */
 479                if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR))
 480                        goto err_hsm;
 481                break;
 482        case HSM_ST_LAST:
 483                if (ata_is_dma(qc->tf.protocol)) {
 484                        /* clear DMA-Start bit */
 485                        ap->ops->bmdma_stop(qc);
 486
 487                        if (bmdma2 & SIL_DMA_ERROR) {
 488                                qc->err_mask |= AC_ERR_HOST_BUS;
 489                                ap->hsm_task_state = HSM_ST_ERR;
 490                        }
 491                }
 492                break;
 493        case HSM_ST:
 494                break;
 495        default:
 496                goto err_hsm;
 497        }
 498
 499        /* check main status, clearing INTRQ */
 500        status = ap->ops->sff_check_status(ap);
 501        if (unlikely(status & ATA_BUSY))
 502                goto err_hsm;
 503
 504        /* ack bmdma irq events */
 505        ata_bmdma_irq_clear(ap);
 506
 507        /* kick HSM in the ass */
 508        ata_sff_hsm_move(ap, qc, status, 0);
 509
 510        if (unlikely(qc->err_mask) && ata_is_dma(qc->tf.protocol))
 511                ata_ehi_push_desc(ehi, "BMDMA2 stat 0x%x", bmdma2);
 512
 513        return;
 514
 515 err_hsm:
 516        qc->err_mask |= AC_ERR_HSM;
 517 freeze:
 518        ata_port_freeze(ap);
 519}
 520
 521static irqreturn_t sil_interrupt(int irq, void *dev_instance)
 522{
 523        struct ata_host *host = dev_instance;
 524        void __iomem *mmio_base = host->iomap[SIL_MMIO_BAR];
 525        int handled = 0;
 526        int i;
 527
 528        spin_lock(&host->lock);
 529
 530        for (i = 0; i < host->n_ports; i++) {
 531                struct ata_port *ap = host->ports[i];
 532                u32 bmdma2 = readl(mmio_base + sil_port[ap->port_no].bmdma2);
 533
 534                /* turn off SATA_IRQ if not supported */
 535                if (ap->flags & SIL_FLAG_NO_SATA_IRQ)
 536                        bmdma2 &= ~SIL_DMA_SATA_IRQ;
 537
 538                if (bmdma2 == 0xffffffff ||
 539                    !(bmdma2 & (SIL_DMA_COMPLETE | SIL_DMA_SATA_IRQ)))
 540                        continue;
 541
 542                sil_host_intr(ap, bmdma2);
 543                handled = 1;
 544        }
 545
 546        spin_unlock(&host->lock);
 547
 548        return IRQ_RETVAL(handled);
 549}
 550
 551static void sil_freeze(struct ata_port *ap)
 552{
 553        void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
 554        u32 tmp;
 555
 556        /* global IRQ mask doesn't block SATA IRQ, turn off explicitly */
 557        writel(0, mmio_base + sil_port[ap->port_no].sien);
 558
 559        /* plug IRQ */
 560        tmp = readl(mmio_base + SIL_SYSCFG);
 561        tmp |= SIL_MASK_IDE0_INT << ap->port_no;
 562        writel(tmp, mmio_base + SIL_SYSCFG);
 563        readl(mmio_base + SIL_SYSCFG);  /* flush */
 564
 565        /* Ensure DMA_ENABLE is off.
 566         *
 567         * This is because the controller will not give us access to the
 568         * taskfile registers while a DMA is in progress
 569         */
 570        iowrite8(ioread8(ap->ioaddr.bmdma_addr) & ~SIL_DMA_ENABLE,
 571                 ap->ioaddr.bmdma_addr);
 572
 573        /* According to ata_bmdma_stop, an HDMA transition requires
 574         * on PIO cycle. But we can't read a taskfile register.
 575         */
 576        ioread8(ap->ioaddr.bmdma_addr);
 577}
 578
 579static void sil_thaw(struct ata_port *ap)
 580{
 581        void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
 582        u32 tmp;
 583
 584        /* clear IRQ */
 585        ap->ops->sff_check_status(ap);
 586        ata_bmdma_irq_clear(ap);
 587
 588        /* turn on SATA IRQ if supported */
 589        if (!(ap->flags & SIL_FLAG_NO_SATA_IRQ))
 590                writel(SIL_SIEN_N, mmio_base + sil_port[ap->port_no].sien);
 591
 592        /* turn on IRQ */
 593        tmp = readl(mmio_base + SIL_SYSCFG);
 594        tmp &= ~(SIL_MASK_IDE0_INT << ap->port_no);
 595        writel(tmp, mmio_base + SIL_SYSCFG);
 596}
 597
 598/**
 599 *      sil_dev_config - Apply device/host-specific errata fixups
 600 *      @dev: Device to be examined
 601 *
 602 *      After the IDENTIFY [PACKET] DEVICE step is complete, and a
 603 *      device is known to be present, this function is called.
 604 *      We apply two errata fixups which are specific to Silicon Image,
 605 *      a Seagate and a Maxtor fixup.
 606 *
 607 *      For certain Seagate devices, we must limit the maximum sectors
 608 *      to under 8K.
 609 *
 610 *      For certain Maxtor devices, we must not program the drive
 611 *      beyond udma5.
 612 *
 613 *      Both fixups are unfairly pessimistic.  As soon as I get more
 614 *      information on these errata, I will create a more exhaustive
 615 *      list, and apply the fixups to only the specific
 616 *      devices/hosts/firmwares that need it.
 617 *
 618 *      20040111 - Seagate drives affected by the Mod15Write bug are blacklisted
 619 *      The Maxtor quirk is in the blacklist, but I'm keeping the original
 620 *      pessimistic fix for the following reasons...
 621 *      - There seems to be less info on it, only one device gleaned off the
 622 *      Windows driver, maybe only one is affected.  More info would be greatly
 623 *      appreciated.
 624 *      - But then again UDMA5 is hardly anything to complain about
 625 */
 626static void sil_dev_config(struct ata_device *dev)
 627{
 628        struct ata_port *ap = dev->link->ap;
 629        int print_info = ap->link.eh_context.i.flags & ATA_EHI_PRINTINFO;
 630        unsigned int n, quirks = 0;
 631        unsigned char model_num[ATA_ID_PROD_LEN + 1];
 632
 633        ata_id_c_string(dev->id, model_num, ATA_ID_PROD, sizeof(model_num));
 634
 635        for (n = 0; sil_blacklist[n].product; n++)
 636                if (!strcmp(sil_blacklist[n].product, model_num)) {
 637                        quirks = sil_blacklist[n].quirk;
 638                        break;
 639                }
 640
 641        /* limit requests to 15 sectors */
 642        if (slow_down ||
 643            ((ap->flags & SIL_FLAG_MOD15WRITE) &&
 644             (quirks & SIL_QUIRK_MOD15WRITE))) {
 645                if (print_info)
 646                        ata_dev_info(dev,
 647                "applying Seagate errata fix (mod15write workaround)\n");
 648                dev->max_sectors = 15;
 649                return;
 650        }
 651
 652        /* limit to udma5 */
 653        if (quirks & SIL_QUIRK_UDMA5MAX) {
 654                if (print_info)
 655                        ata_dev_info(dev, "applying Maxtor errata fix %s\n",
 656                                     model_num);
 657                dev->udma_mask &= ATA_UDMA5;
 658                return;
 659        }
 660}
 661
 662static void sil_init_controller(struct ata_host *host)
 663{
 664        struct pci_dev *pdev = to_pci_dev(host->dev);
 665        void __iomem *mmio_base = host->iomap[SIL_MMIO_BAR];
 666        u8 cls;
 667        u32 tmp;
 668        int i;
 669
 670        /* Initialize FIFO PCI bus arbitration */
 671        cls = sil_get_device_cache_line(pdev);
 672        if (cls) {
 673                cls >>= 3;
 674                cls++;  /* cls = (line_size/8)+1 */
 675                for (i = 0; i < host->n_ports; i++)
 676                        writew(cls << 8 | cls,
 677                               mmio_base + sil_port[i].fifo_cfg);
 678        } else
 679                dev_warn(&pdev->dev,
 680                         "cache line size not set.  Driver may not function\n");
 681
 682        /* Apply R_ERR on DMA activate FIS errata workaround */
 683        if (host->ports[0]->flags & SIL_FLAG_RERR_ON_DMA_ACT) {
 684                int cnt;
 685
 686                for (i = 0, cnt = 0; i < host->n_ports; i++) {
 687                        tmp = readl(mmio_base + sil_port[i].sfis_cfg);
 688                        if ((tmp & 0x3) != 0x01)
 689                                continue;
 690                        if (!cnt)
 691                                dev_info(&pdev->dev,
 692                                         "Applying R_ERR on DMA activate FIS errata fix\n");
 693                        writel(tmp & ~0x3, mmio_base + sil_port[i].sfis_cfg);
 694                        cnt++;
 695                }
 696        }
 697
 698        if (host->n_ports == 4) {
 699                /* flip the magic "make 4 ports work" bit */
 700                tmp = readl(mmio_base + sil_port[2].bmdma);
 701                if ((tmp & SIL_INTR_STEERING) == 0)
 702                        writel(tmp | SIL_INTR_STEERING,
 703                               mmio_base + sil_port[2].bmdma);
 704        }
 705}
 706
 707static bool sil_broken_system_poweroff(struct pci_dev *pdev)
 708{
 709        static const struct dmi_system_id broken_systems[] = {
 710                {
 711                        .ident = "HP Compaq nx6325",
 712                        .matches = {
 713                                DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
 714                                DMI_MATCH(DMI_PRODUCT_NAME, "HP Compaq nx6325"),
 715                        },
 716                        /* PCI slot number of the controller */
 717                        .driver_data = (void *)0x12UL,
 718                },
 719
 720                { }     /* terminate list */
 721        };
 722        const struct dmi_system_id *dmi = dmi_first_match(broken_systems);
 723
 724        if (dmi) {
 725                unsigned long slot = (unsigned long)dmi->driver_data;
 726                /* apply the quirk only to on-board controllers */
 727                return slot == PCI_SLOT(pdev->devfn);
 728        }
 729
 730        return false;
 731}
 732
 733static int sil_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
 734{
 735        int board_id = ent->driver_data;
 736        struct ata_port_info pi = sil_port_info[board_id];
 737        const struct ata_port_info *ppi[] = { &pi, NULL };
 738        struct ata_host *host;
 739        void __iomem *mmio_base;
 740        int n_ports, rc;
 741        unsigned int i;
 742
 743        ata_print_version_once(&pdev->dev, DRV_VERSION);
 744
 745        /* allocate host */
 746        n_ports = 2;
 747        if (board_id == sil_3114)
 748                n_ports = 4;
 749
 750        if (sil_broken_system_poweroff(pdev)) {
 751                pi.flags |= ATA_FLAG_NO_POWEROFF_SPINDOWN |
 752                                        ATA_FLAG_NO_HIBERNATE_SPINDOWN;
 753                dev_info(&pdev->dev, "quirky BIOS, skipping spindown "
 754                                "on poweroff and hibernation\n");
 755        }
 756
 757        host = ata_host_alloc_pinfo(&pdev->dev, ppi, n_ports);
 758        if (!host)
 759                return -ENOMEM;
 760
 761        /* acquire resources and fill host */
 762        rc = pcim_enable_device(pdev);
 763        if (rc)
 764                return rc;
 765
 766        rc = pcim_iomap_regions(pdev, 1 << SIL_MMIO_BAR, DRV_NAME);
 767        if (rc == -EBUSY)
 768                pcim_pin_device(pdev);
 769        if (rc)
 770                return rc;
 771        host->iomap = pcim_iomap_table(pdev);
 772
 773        rc = dma_set_mask(&pdev->dev, ATA_DMA_MASK);
 774        if (rc)
 775                return rc;
 776        rc = dma_set_coherent_mask(&pdev->dev, ATA_DMA_MASK);
 777        if (rc)
 778                return rc;
 779
 780        mmio_base = host->iomap[SIL_MMIO_BAR];
 781
 782        for (i = 0; i < host->n_ports; i++) {
 783                struct ata_port *ap = host->ports[i];
 784                struct ata_ioports *ioaddr = &ap->ioaddr;
 785
 786                ioaddr->cmd_addr = mmio_base + sil_port[i].tf;
 787                ioaddr->altstatus_addr =
 788                ioaddr->ctl_addr = mmio_base + sil_port[i].ctl;
 789                ioaddr->bmdma_addr = mmio_base + sil_port[i].bmdma;
 790                ioaddr->scr_addr = mmio_base + sil_port[i].scr;
 791                ata_sff_std_ports(ioaddr);
 792
 793                ata_port_pbar_desc(ap, SIL_MMIO_BAR, -1, "mmio");
 794                ata_port_pbar_desc(ap, SIL_MMIO_BAR, sil_port[i].tf, "tf");
 795        }
 796
 797        /* initialize and activate */
 798        sil_init_controller(host);
 799
 800        pci_set_master(pdev);
 801        return ata_host_activate(host, pdev->irq, sil_interrupt, IRQF_SHARED,
 802                                 &sil_sht);
 803}
 804
 805#ifdef CONFIG_PM_SLEEP
 806static int sil_pci_device_resume(struct pci_dev *pdev)
 807{
 808        struct ata_host *host = pci_get_drvdata(pdev);
 809        int rc;
 810
 811        rc = ata_pci_device_do_resume(pdev);
 812        if (rc)
 813                return rc;
 814
 815        sil_init_controller(host);
 816        ata_host_resume(host);
 817
 818        return 0;
 819}
 820#endif
 821
 822module_pci_driver(sil_pci_driver);
 823