linux/drivers/net/ethernet/aquantia/atlantic/aq_pci_func.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-only
   2/* Atlantic Network Driver
   3 *
   4 * Copyright (C) 2014-2019 aQuantia Corporation
   5 * Copyright (C) 2019-2020 Marvell International Ltd.
   6 */
   7
   8/* File aq_pci_func.c: Definition of PCI functions. */
   9
  10#include <linux/interrupt.h>
  11#include <linux/module.h>
  12
  13#include "aq_main.h"
  14#include "aq_nic.h"
  15#include "aq_vec.h"
  16#include "aq_hw.h"
  17#include "aq_pci_func.h"
  18#include "hw_atl/hw_atl_a0.h"
  19#include "hw_atl/hw_atl_b0.h"
  20#include "hw_atl2/hw_atl2.h"
  21#include "aq_filters.h"
  22#include "aq_drvinfo.h"
  23#include "aq_macsec.h"
  24
  25static const struct pci_device_id aq_pci_tbl[] = {
  26        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_0001), },
  27        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D100), },
  28        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D107), },
  29        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D108), },
  30        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_D109), },
  31
  32        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC100), },
  33        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC107), },
  34        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC108), },
  35        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC109), },
  36        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC111), },
  37        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC112), },
  38
  39        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC100S), },
  40        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC107S), },
  41        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC108S), },
  42        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC109S), },
  43        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC111S), },
  44        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC112S), },
  45
  46        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113DEV), },
  47        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113CS), },
  48        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC114CS), },
  49        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113), },
  50        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC113C), },
  51        { PCI_VDEVICE(AQUANTIA, AQ_DEVICE_ID_AQC115C), },
  52
  53        {}
  54};
  55
  56static const struct aq_board_revision_s hw_atl_boards[] = {
  57        { AQ_DEVICE_ID_0001,    AQ_HWREV_1,     &hw_atl_ops_a0, &hw_atl_a0_caps_aqc107, },
  58        { AQ_DEVICE_ID_D100,    AQ_HWREV_1,     &hw_atl_ops_a0, &hw_atl_a0_caps_aqc100, },
  59        { AQ_DEVICE_ID_D107,    AQ_HWREV_1,     &hw_atl_ops_a0, &hw_atl_a0_caps_aqc107, },
  60        { AQ_DEVICE_ID_D108,    AQ_HWREV_1,     &hw_atl_ops_a0, &hw_atl_a0_caps_aqc108, },
  61        { AQ_DEVICE_ID_D109,    AQ_HWREV_1,     &hw_atl_ops_a0, &hw_atl_a0_caps_aqc109, },
  62
  63        { AQ_DEVICE_ID_0001,    AQ_HWREV_2,     &hw_atl_ops_b0, &hw_atl_b0_caps_aqc107, },
  64        { AQ_DEVICE_ID_D100,    AQ_HWREV_2,     &hw_atl_ops_b0, &hw_atl_b0_caps_aqc100, },
  65        { AQ_DEVICE_ID_D107,    AQ_HWREV_2,     &hw_atl_ops_b0, &hw_atl_b0_caps_aqc107, },
  66        { AQ_DEVICE_ID_D108,    AQ_HWREV_2,     &hw_atl_ops_b0, &hw_atl_b0_caps_aqc108, },
  67        { AQ_DEVICE_ID_D109,    AQ_HWREV_2,     &hw_atl_ops_b0, &hw_atl_b0_caps_aqc109, },
  68
  69        { AQ_DEVICE_ID_AQC100,  AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc100, },
  70        { AQ_DEVICE_ID_AQC107,  AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc107, },
  71        { AQ_DEVICE_ID_AQC108,  AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc108, },
  72        { AQ_DEVICE_ID_AQC109,  AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc109, },
  73        { AQ_DEVICE_ID_AQC111,  AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc111, },
  74        { AQ_DEVICE_ID_AQC112,  AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc112, },
  75
  76        { AQ_DEVICE_ID_AQC100S, AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc100s, },
  77        { AQ_DEVICE_ID_AQC107S, AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc107s, },
  78        { AQ_DEVICE_ID_AQC108S, AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc108s, },
  79        { AQ_DEVICE_ID_AQC109S, AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc109s, },
  80        { AQ_DEVICE_ID_AQC111S, AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc111s, },
  81        { AQ_DEVICE_ID_AQC112S, AQ_HWREV_ANY,   &hw_atl_ops_b1, &hw_atl_b0_caps_aqc112s, },
  82
  83        { AQ_DEVICE_ID_AQC113DEV,       AQ_HWREV_ANY,   &hw_atl2_ops, &hw_atl2_caps_aqc113, },
  84        { AQ_DEVICE_ID_AQC113,          AQ_HWREV_ANY,   &hw_atl2_ops, &hw_atl2_caps_aqc113, },
  85        { AQ_DEVICE_ID_AQC113CS,        AQ_HWREV_ANY,   &hw_atl2_ops, &hw_atl2_caps_aqc113, },
  86        { AQ_DEVICE_ID_AQC114CS,        AQ_HWREV_ANY,   &hw_atl2_ops, &hw_atl2_caps_aqc113, },
  87        { AQ_DEVICE_ID_AQC113C,         AQ_HWREV_ANY,   &hw_atl2_ops, &hw_atl2_caps_aqc113, },
  88        { AQ_DEVICE_ID_AQC115C,         AQ_HWREV_ANY,   &hw_atl2_ops, &hw_atl2_caps_aqc113, },
  89};
  90
  91MODULE_DEVICE_TABLE(pci, aq_pci_tbl);
  92
  93static int aq_pci_probe_get_hw_by_id(struct pci_dev *pdev,
  94                                     const struct aq_hw_ops **ops,
  95                                     const struct aq_hw_caps_s **caps)
  96{
  97        int i;
  98
  99        if (pdev->vendor != PCI_VENDOR_ID_AQUANTIA)
 100                return -EINVAL;
 101
 102        for (i = 0; i < ARRAY_SIZE(hw_atl_boards); i++) {
 103                if (hw_atl_boards[i].devid == pdev->device &&
 104                    (hw_atl_boards[i].revision == AQ_HWREV_ANY ||
 105                     hw_atl_boards[i].revision == pdev->revision)) {
 106                        *ops = hw_atl_boards[i].ops;
 107                        *caps = hw_atl_boards[i].caps;
 108                        break;
 109                }
 110        }
 111
 112        if (i == ARRAY_SIZE(hw_atl_boards))
 113                return -EINVAL;
 114
 115        return 0;
 116}
 117
 118static int aq_pci_func_init(struct pci_dev *pdev)
 119{
 120        int err;
 121
 122        err = pci_set_dma_mask(pdev, DMA_BIT_MASK(64));
 123        if (!err)
 124                err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64));
 125        if (err) {
 126                err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
 127                if (!err)
 128                        err = pci_set_consistent_dma_mask(pdev,
 129                                                          DMA_BIT_MASK(32));
 130        }
 131        if (err != 0) {
 132                err = -ENOSR;
 133                goto err_exit;
 134        }
 135
 136        err = pci_request_regions(pdev, AQ_CFG_DRV_NAME "_mmio");
 137        if (err < 0)
 138                goto err_exit;
 139
 140        pci_set_master(pdev);
 141
 142        return 0;
 143
 144err_exit:
 145        return err;
 146}
 147
 148int aq_pci_func_alloc_irq(struct aq_nic_s *self, unsigned int i,
 149                          char *name, irq_handler_t irq_handler,
 150                          void *irq_arg, cpumask_t *affinity_mask)
 151{
 152        struct pci_dev *pdev = self->pdev;
 153        int err;
 154
 155        if (pdev->msix_enabled || pdev->msi_enabled)
 156                err = request_irq(pci_irq_vector(pdev, i), irq_handler, 0,
 157                                  name, irq_arg);
 158        else
 159                err = request_irq(pci_irq_vector(pdev, i), aq_vec_isr_legacy,
 160                                  IRQF_SHARED, name, irq_arg);
 161
 162        if (err >= 0) {
 163                self->msix_entry_mask |= (1 << i);
 164
 165                if (pdev->msix_enabled && affinity_mask)
 166                        irq_set_affinity_hint(pci_irq_vector(pdev, i),
 167                                              affinity_mask);
 168        }
 169
 170        return err;
 171}
 172
 173void aq_pci_func_free_irqs(struct aq_nic_s *self)
 174{
 175        struct pci_dev *pdev = self->pdev;
 176        unsigned int i;
 177        void *irq_data;
 178
 179        for (i = 32U; i--;) {
 180                if (!((1U << i) & self->msix_entry_mask))
 181                        continue;
 182                if (self->aq_nic_cfg.link_irq_vec &&
 183                    i == self->aq_nic_cfg.link_irq_vec)
 184                        irq_data = self;
 185                else if (i < AQ_CFG_VECS_MAX)
 186                        irq_data = self->aq_vec[i];
 187                else
 188                        continue;
 189
 190                if (pdev->msix_enabled)
 191                        irq_set_affinity_hint(pci_irq_vector(pdev, i), NULL);
 192                free_irq(pci_irq_vector(pdev, i), irq_data);
 193                self->msix_entry_mask &= ~(1U << i);
 194        }
 195}
 196
 197unsigned int aq_pci_func_get_irq_type(struct aq_nic_s *self)
 198{
 199        if (self->pdev->msix_enabled)
 200                return AQ_HW_IRQ_MSIX;
 201        if (self->pdev->msi_enabled)
 202                return AQ_HW_IRQ_MSI;
 203
 204        return AQ_HW_IRQ_LEGACY;
 205}
 206
 207static void aq_pci_free_irq_vectors(struct aq_nic_s *self)
 208{
 209        pci_free_irq_vectors(self->pdev);
 210}
 211
 212static int aq_pci_probe(struct pci_dev *pdev,
 213                        const struct pci_device_id *pci_id)
 214{
 215        struct net_device *ndev;
 216        resource_size_t mmio_pa;
 217        struct aq_nic_s *self;
 218        u32 numvecs;
 219        u32 bar;
 220        int err;
 221
 222        err = pci_enable_device(pdev);
 223        if (err)
 224                return err;
 225
 226        err = aq_pci_func_init(pdev);
 227        if (err)
 228                goto err_pci_func;
 229
 230        ndev = aq_ndev_alloc();
 231        if (!ndev) {
 232                err = -ENOMEM;
 233                goto err_ndev;
 234        }
 235
 236        self = netdev_priv(ndev);
 237        self->pdev = pdev;
 238        SET_NETDEV_DEV(ndev, &pdev->dev);
 239        pci_set_drvdata(pdev, self);
 240
 241        mutex_init(&self->fwreq_mutex);
 242
 243        err = aq_pci_probe_get_hw_by_id(pdev, &self->aq_hw_ops,
 244                                        &aq_nic_get_cfg(self)->aq_hw_caps);
 245        if (err)
 246                goto err_ioremap;
 247
 248        self->aq_hw = kzalloc(sizeof(*self->aq_hw), GFP_KERNEL);
 249        if (!self->aq_hw) {
 250                err = -ENOMEM;
 251                goto err_ioremap;
 252        }
 253        self->aq_hw->aq_nic_cfg = aq_nic_get_cfg(self);
 254        if (self->aq_hw->aq_nic_cfg->aq_hw_caps->priv_data_len) {
 255                int len = self->aq_hw->aq_nic_cfg->aq_hw_caps->priv_data_len;
 256
 257                self->aq_hw->priv = kzalloc(len, GFP_KERNEL);
 258                if (!self->aq_hw->priv) {
 259                        err = -ENOMEM;
 260                        goto err_free_aq_hw;
 261                }
 262        }
 263
 264        for (bar = 0; bar < 4; ++bar) {
 265                if (IORESOURCE_MEM & pci_resource_flags(pdev, bar)) {
 266                        resource_size_t reg_sz;
 267
 268                        mmio_pa = pci_resource_start(pdev, bar);
 269                        if (mmio_pa == 0U) {
 270                                err = -EIO;
 271                                goto err_free_aq_hw_priv;
 272                        }
 273
 274                        reg_sz = pci_resource_len(pdev, bar);
 275                        if ((reg_sz <= 24 /*ATL_REGS_SIZE*/)) {
 276                                err = -EIO;
 277                                goto err_free_aq_hw_priv;
 278                        }
 279
 280                        self->aq_hw->mmio = ioremap(mmio_pa, reg_sz);
 281                        if (!self->aq_hw->mmio) {
 282                                err = -EIO;
 283                                goto err_free_aq_hw_priv;
 284                        }
 285                        break;
 286                }
 287        }
 288
 289        if (bar == 4) {
 290                err = -EIO;
 291                goto err_free_aq_hw_priv;
 292        }
 293
 294        numvecs = min((u8)AQ_CFG_VECS_DEF,
 295                      aq_nic_get_cfg(self)->aq_hw_caps->msix_irqs);
 296        numvecs = min(numvecs, num_online_cpus());
 297        /* Request IRQ vector for PTP */
 298        numvecs += 1;
 299
 300        numvecs += AQ_HW_SERVICE_IRQS;
 301        /*enable interrupts */
 302#if !AQ_CFG_FORCE_LEGACY_INT
 303        err = pci_alloc_irq_vectors(self->pdev, 1, numvecs,
 304                                    PCI_IRQ_MSIX | PCI_IRQ_MSI |
 305                                    PCI_IRQ_LEGACY);
 306
 307        if (err < 0)
 308                goto err_hwinit;
 309        numvecs = err;
 310#endif
 311        self->irqvecs = numvecs;
 312
 313        /* net device init */
 314        aq_nic_cfg_start(self);
 315
 316        aq_nic_ndev_init(self);
 317
 318        err = aq_nic_ndev_register(self);
 319        if (err < 0)
 320                goto err_register;
 321
 322        aq_drvinfo_init(ndev);
 323
 324        return 0;
 325
 326err_register:
 327        aq_nic_free_vectors(self);
 328        aq_pci_free_irq_vectors(self);
 329err_hwinit:
 330        iounmap(self->aq_hw->mmio);
 331err_free_aq_hw_priv:
 332        kfree(self->aq_hw->priv);
 333err_free_aq_hw:
 334        kfree(self->aq_hw);
 335err_ioremap:
 336        free_netdev(ndev);
 337err_ndev:
 338        pci_release_regions(pdev);
 339err_pci_func:
 340        pci_disable_device(pdev);
 341
 342        return err;
 343}
 344
 345static void aq_pci_remove(struct pci_dev *pdev)
 346{
 347        struct aq_nic_s *self = pci_get_drvdata(pdev);
 348
 349        if (self->ndev) {
 350                aq_clear_rxnfc_all_rules(self);
 351                if (self->ndev->reg_state == NETREG_REGISTERED)
 352                        unregister_netdev(self->ndev);
 353
 354#if IS_ENABLED(CONFIG_MACSEC)
 355                aq_macsec_free(self);
 356#endif
 357                aq_nic_free_vectors(self);
 358                aq_pci_free_irq_vectors(self);
 359                iounmap(self->aq_hw->mmio);
 360                kfree(self->aq_hw->priv);
 361                kfree(self->aq_hw);
 362                pci_release_regions(pdev);
 363                free_netdev(self->ndev);
 364        }
 365
 366        pci_disable_device(pdev);
 367}
 368
 369static void aq_pci_shutdown(struct pci_dev *pdev)
 370{
 371        struct aq_nic_s *self = pci_get_drvdata(pdev);
 372
 373        aq_nic_shutdown(self);
 374
 375        pci_disable_device(pdev);
 376
 377        if (system_state == SYSTEM_POWER_OFF) {
 378                pci_wake_from_d3(pdev, false);
 379                pci_set_power_state(pdev, PCI_D3hot);
 380        }
 381}
 382
 383static int aq_suspend_common(struct device *dev, bool deep)
 384{
 385        struct aq_nic_s *nic = pci_get_drvdata(to_pci_dev(dev));
 386
 387        rtnl_lock();
 388
 389        nic->power_state = AQ_HW_POWER_STATE_D3;
 390        netif_device_detach(nic->ndev);
 391        netif_tx_stop_all_queues(nic->ndev);
 392
 393        if (netif_running(nic->ndev))
 394                aq_nic_stop(nic);
 395
 396        if (deep) {
 397                aq_nic_deinit(nic, !nic->aq_hw->aq_nic_cfg->wol);
 398                aq_nic_set_power(nic);
 399        }
 400
 401        rtnl_unlock();
 402
 403        return 0;
 404}
 405
 406static int atl_resume_common(struct device *dev, bool deep)
 407{
 408        struct pci_dev *pdev = to_pci_dev(dev);
 409        struct aq_nic_s *nic;
 410        int ret = 0;
 411
 412        nic = pci_get_drvdata(pdev);
 413
 414        rtnl_lock();
 415
 416        pci_set_power_state(pdev, PCI_D0);
 417        pci_restore_state(pdev);
 418
 419        if (deep) {
 420                ret = aq_nic_init(nic);
 421                if (ret)
 422                        goto err_exit;
 423        }
 424
 425        if (netif_running(nic->ndev)) {
 426                ret = aq_nic_start(nic);
 427                if (ret)
 428                        goto err_exit;
 429        }
 430
 431        netif_device_attach(nic->ndev);
 432        netif_tx_start_all_queues(nic->ndev);
 433
 434err_exit:
 435        if (ret < 0)
 436                aq_nic_deinit(nic, true);
 437
 438        rtnl_unlock();
 439
 440        return ret;
 441}
 442
 443static int aq_pm_freeze(struct device *dev)
 444{
 445        return aq_suspend_common(dev, false);
 446}
 447
 448static int aq_pm_suspend_poweroff(struct device *dev)
 449{
 450        return aq_suspend_common(dev, true);
 451}
 452
 453static int aq_pm_thaw(struct device *dev)
 454{
 455        return atl_resume_common(dev, false);
 456}
 457
 458static int aq_pm_resume_restore(struct device *dev)
 459{
 460        return atl_resume_common(dev, true);
 461}
 462
 463static const struct dev_pm_ops aq_pm_ops = {
 464        .suspend = aq_pm_suspend_poweroff,
 465        .poweroff = aq_pm_suspend_poweroff,
 466        .freeze = aq_pm_freeze,
 467        .resume = aq_pm_resume_restore,
 468        .restore = aq_pm_resume_restore,
 469        .thaw = aq_pm_thaw,
 470};
 471
 472static struct pci_driver aq_pci_ops = {
 473        .name = AQ_CFG_DRV_NAME,
 474        .id_table = aq_pci_tbl,
 475        .probe = aq_pci_probe,
 476        .remove = aq_pci_remove,
 477        .shutdown = aq_pci_shutdown,
 478#ifdef CONFIG_PM
 479        .driver.pm = &aq_pm_ops,
 480#endif
 481};
 482
 483int aq_pci_func_register_driver(void)
 484{
 485        return pci_register_driver(&aq_pci_ops);
 486}
 487
 488void aq_pci_func_unregister_driver(void)
 489{
 490        pci_unregister_driver(&aq_pci_ops);
 491}
 492
 493