linux/drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-only
   2/* 10G controller driver for Samsung SoCs
   3 *
   4 * Copyright (C) 2013 Samsung Electronics Co., Ltd.
   5 *              http://www.samsung.com
   6 *
   7 * Author: Siva Reddy Kallam <siva.kallam@samsung.com>
   8 */
   9#include <linux/delay.h>
  10#include <linux/export.h>
  11#include <linux/io.h>
  12#include <linux/netdevice.h>
  13#include <linux/phy.h>
  14
  15#include "sxgbe_common.h"
  16#include "sxgbe_dma.h"
  17#include "sxgbe_reg.h"
  18#include "sxgbe_desc.h"
  19
  20/* DMA core initialization */
  21static int sxgbe_dma_init(void __iomem *ioaddr, int fix_burst, int burst_map)
  22{
  23        u32 reg_val;
  24
  25        reg_val = readl(ioaddr + SXGBE_DMA_SYSBUS_MODE_REG);
  26
  27        /* if fix_burst = 0, Set UNDEF = 1 of DMA_Sys_Mode Register.
  28         * if fix_burst = 1, Set UNDEF = 0 of DMA_Sys_Mode Register.
  29         * burst_map is bitmap for  BLEN[4, 8, 16, 32, 64, 128 and 256].
  30         * Set burst_map irrespective of fix_burst value.
  31         */
  32        if (!fix_burst)
  33                reg_val |= SXGBE_DMA_AXI_UNDEF_BURST;
  34
  35        /* write burst len map */
  36        reg_val |= (burst_map << SXGBE_DMA_BLENMAP_LSHIFT);
  37
  38        writel(reg_val, ioaddr + SXGBE_DMA_SYSBUS_MODE_REG);
  39
  40        return 0;
  41}
  42
  43static void sxgbe_dma_channel_init(void __iomem *ioaddr, int cha_num,
  44                                   int fix_burst, int pbl, dma_addr_t dma_tx,
  45                                   dma_addr_t dma_rx, int t_rsize, int r_rsize)
  46{
  47        u32 reg_val;
  48        dma_addr_t dma_addr;
  49
  50        reg_val = readl(ioaddr + SXGBE_DMA_CHA_CTL_REG(cha_num));
  51        /* set the pbl */
  52        if (fix_burst) {
  53                reg_val |= SXGBE_DMA_PBL_X8MODE;
  54                writel(reg_val, ioaddr + SXGBE_DMA_CHA_CTL_REG(cha_num));
  55                /* program the TX pbl */
  56                reg_val = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cha_num));
  57                reg_val |= (pbl << SXGBE_DMA_TXPBL_LSHIFT);
  58                writel(reg_val, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cha_num));
  59                /* program the RX pbl */
  60                reg_val = readl(ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cha_num));
  61                reg_val |= (pbl << SXGBE_DMA_RXPBL_LSHIFT);
  62                writel(reg_val, ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cha_num));
  63        }
  64
  65        /* program desc registers */
  66        writel(upper_32_bits(dma_tx),
  67               ioaddr + SXGBE_DMA_CHA_TXDESC_HADD_REG(cha_num));
  68        writel(lower_32_bits(dma_tx),
  69               ioaddr + SXGBE_DMA_CHA_TXDESC_LADD_REG(cha_num));
  70
  71        writel(upper_32_bits(dma_rx),
  72               ioaddr + SXGBE_DMA_CHA_RXDESC_HADD_REG(cha_num));
  73        writel(lower_32_bits(dma_rx),
  74               ioaddr + SXGBE_DMA_CHA_RXDESC_LADD_REG(cha_num));
  75
  76        /* program tail pointers */
  77        /* assumption: upper 32 bits are constant and
  78         * same as TX/RX desc list
  79         */
  80        dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
  81        writel(lower_32_bits(dma_addr),
  82               ioaddr + SXGBE_DMA_CHA_TXDESC_TAILPTR_REG(cha_num));
  83
  84        dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
  85        writel(lower_32_bits(dma_addr),
  86               ioaddr + SXGBE_DMA_CHA_RXDESC_LADD_REG(cha_num));
  87        /* program the ring sizes */
  88        writel(t_rsize - 1, ioaddr + SXGBE_DMA_CHA_TXDESC_RINGLEN_REG(cha_num));
  89        writel(r_rsize - 1, ioaddr + SXGBE_DMA_CHA_RXDESC_RINGLEN_REG(cha_num));
  90
  91        /* Enable TX/RX interrupts */
  92        writel(SXGBE_DMA_ENA_INT,
  93               ioaddr + SXGBE_DMA_CHA_INT_ENABLE_REG(cha_num));
  94}
  95
  96static void sxgbe_enable_dma_transmission(void __iomem *ioaddr, int cha_num)
  97{
  98        u32 tx_config;
  99
 100        tx_config = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cha_num));
 101        tx_config |= SXGBE_TX_START_DMA;
 102        writel(tx_config, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cha_num));
 103}
 104
 105static void sxgbe_enable_dma_irq(void __iomem *ioaddr, int dma_cnum)
 106{
 107        /* Enable TX/RX interrupts */
 108        writel(SXGBE_DMA_ENA_INT,
 109               ioaddr + SXGBE_DMA_CHA_INT_ENABLE_REG(dma_cnum));
 110}
 111
 112static void sxgbe_disable_dma_irq(void __iomem *ioaddr, int dma_cnum)
 113{
 114        /* Disable TX/RX interrupts */
 115        writel(0, ioaddr + SXGBE_DMA_CHA_INT_ENABLE_REG(dma_cnum));
 116}
 117
 118static void sxgbe_dma_start_tx(void __iomem *ioaddr, int tchannels)
 119{
 120        int cnum;
 121        u32 tx_ctl_reg;
 122
 123        for (cnum = 0; cnum < tchannels; cnum++) {
 124                tx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum));
 125                tx_ctl_reg |= SXGBE_TX_ENABLE;
 126                writel(tx_ctl_reg,
 127                       ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum));
 128        }
 129}
 130
 131static void sxgbe_dma_start_tx_queue(void __iomem *ioaddr, int dma_cnum)
 132{
 133        u32 tx_ctl_reg;
 134
 135        tx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(dma_cnum));
 136        tx_ctl_reg |= SXGBE_TX_ENABLE;
 137        writel(tx_ctl_reg, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(dma_cnum));
 138}
 139
 140static void sxgbe_dma_stop_tx_queue(void __iomem *ioaddr, int dma_cnum)
 141{
 142        u32 tx_ctl_reg;
 143
 144        tx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(dma_cnum));
 145        tx_ctl_reg &= ~(SXGBE_TX_ENABLE);
 146        writel(tx_ctl_reg, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(dma_cnum));
 147}
 148
 149static void sxgbe_dma_stop_tx(void __iomem *ioaddr, int tchannels)
 150{
 151        int cnum;
 152        u32 tx_ctl_reg;
 153
 154        for (cnum = 0; cnum < tchannels; cnum++) {
 155                tx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum));
 156                tx_ctl_reg &= ~(SXGBE_TX_ENABLE);
 157                writel(tx_ctl_reg, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(cnum));
 158        }
 159}
 160
 161static void sxgbe_dma_start_rx(void __iomem *ioaddr, int rchannels)
 162{
 163        int cnum;
 164        u32 rx_ctl_reg;
 165
 166        for (cnum = 0; cnum < rchannels; cnum++) {
 167                rx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum));
 168                rx_ctl_reg |= SXGBE_RX_ENABLE;
 169                writel(rx_ctl_reg,
 170                       ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum));
 171        }
 172}
 173
 174static void sxgbe_dma_stop_rx(void __iomem *ioaddr, int rchannels)
 175{
 176        int cnum;
 177        u32 rx_ctl_reg;
 178
 179        for (cnum = 0; cnum < rchannels; cnum++) {
 180                rx_ctl_reg = readl(ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum));
 181                rx_ctl_reg &= ~(SXGBE_RX_ENABLE);
 182                writel(rx_ctl_reg, ioaddr + SXGBE_DMA_CHA_RXCTL_REG(cnum));
 183        }
 184}
 185
 186static int sxgbe_tx_dma_int_status(void __iomem *ioaddr, int channel_no,
 187                                   struct sxgbe_extra_stats *x)
 188{
 189        u32 int_status = readl(ioaddr + SXGBE_DMA_CHA_STATUS_REG(channel_no));
 190        u32 clear_val = 0;
 191        u32 ret_val = 0;
 192
 193        /* TX Normal Interrupt Summary */
 194        if (likely(int_status & SXGBE_DMA_INT_STATUS_NIS)) {
 195                x->normal_irq_n++;
 196                if (int_status & SXGBE_DMA_INT_STATUS_TI) {
 197                        ret_val |= handle_tx;
 198                        x->tx_normal_irq_n++;
 199                        clear_val |= SXGBE_DMA_INT_STATUS_TI;
 200                }
 201
 202                if (int_status & SXGBE_DMA_INT_STATUS_TBU) {
 203                        x->tx_underflow_irq++;
 204                        ret_val |= tx_bump_tc;
 205                        clear_val |= SXGBE_DMA_INT_STATUS_TBU;
 206                }
 207        } else if (unlikely(int_status & SXGBE_DMA_INT_STATUS_AIS)) {
 208                /* TX Abnormal Interrupt Summary */
 209                if (int_status & SXGBE_DMA_INT_STATUS_TPS) {
 210                        ret_val |= tx_hard_error;
 211                        clear_val |= SXGBE_DMA_INT_STATUS_TPS;
 212                        x->tx_process_stopped_irq++;
 213                }
 214
 215                if (int_status & SXGBE_DMA_INT_STATUS_FBE) {
 216                        ret_val |= tx_hard_error;
 217                        x->fatal_bus_error_irq++;
 218
 219                        /* Assumption: FBE bit is the combination of
 220                         * all the bus access erros and cleared when
 221                         * the respective error bits cleared
 222                         */
 223
 224                        /* check for actual cause */
 225                        if (int_status & SXGBE_DMA_INT_STATUS_TEB0) {
 226                                x->tx_read_transfer_err++;
 227                                clear_val |= SXGBE_DMA_INT_STATUS_TEB0;
 228                        } else {
 229                                x->tx_write_transfer_err++;
 230                        }
 231
 232                        if (int_status & SXGBE_DMA_INT_STATUS_TEB1) {
 233                                x->tx_desc_access_err++;
 234                                clear_val |= SXGBE_DMA_INT_STATUS_TEB1;
 235                        } else {
 236                                x->tx_buffer_access_err++;
 237                        }
 238
 239                        if (int_status & SXGBE_DMA_INT_STATUS_TEB2) {
 240                                x->tx_data_transfer_err++;
 241                                clear_val |= SXGBE_DMA_INT_STATUS_TEB2;
 242                        }
 243                }
 244
 245                /* context descriptor error */
 246                if (int_status & SXGBE_DMA_INT_STATUS_CTXTERR) {
 247                        x->tx_ctxt_desc_err++;
 248                        clear_val |= SXGBE_DMA_INT_STATUS_CTXTERR;
 249                }
 250        }
 251
 252        /* clear the served bits */
 253        writel(clear_val, ioaddr + SXGBE_DMA_CHA_STATUS_REG(channel_no));
 254
 255        return ret_val;
 256}
 257
 258static int sxgbe_rx_dma_int_status(void __iomem *ioaddr, int channel_no,
 259                                   struct sxgbe_extra_stats *x)
 260{
 261        u32 int_status = readl(ioaddr + SXGBE_DMA_CHA_STATUS_REG(channel_no));
 262        u32 clear_val = 0;
 263        u32 ret_val = 0;
 264
 265        /* RX Normal Interrupt Summary */
 266        if (likely(int_status & SXGBE_DMA_INT_STATUS_NIS)) {
 267                x->normal_irq_n++;
 268                if (int_status & SXGBE_DMA_INT_STATUS_RI) {
 269                        ret_val |= handle_rx;
 270                        x->rx_normal_irq_n++;
 271                        clear_val |= SXGBE_DMA_INT_STATUS_RI;
 272                }
 273        } else if (unlikely(int_status & SXGBE_DMA_INT_STATUS_AIS)) {
 274                /* RX Abnormal Interrupt Summary */
 275                if (int_status & SXGBE_DMA_INT_STATUS_RBU) {
 276                        ret_val |= rx_bump_tc;
 277                        clear_val |= SXGBE_DMA_INT_STATUS_RBU;
 278                        x->rx_underflow_irq++;
 279                }
 280
 281                if (int_status & SXGBE_DMA_INT_STATUS_RPS) {
 282                        ret_val |= rx_hard_error;
 283                        clear_val |= SXGBE_DMA_INT_STATUS_RPS;
 284                        x->rx_process_stopped_irq++;
 285                }
 286
 287                if (int_status & SXGBE_DMA_INT_STATUS_FBE) {
 288                        ret_val |= rx_hard_error;
 289                        x->fatal_bus_error_irq++;
 290
 291                        /* Assumption: FBE bit is the combination of
 292                         * all the bus access erros and cleared when
 293                         * the respective error bits cleared
 294                         */
 295
 296                        /* check for actual cause */
 297                        if (int_status & SXGBE_DMA_INT_STATUS_REB0) {
 298                                x->rx_read_transfer_err++;
 299                                clear_val |= SXGBE_DMA_INT_STATUS_REB0;
 300                        } else {
 301                                x->rx_write_transfer_err++;
 302                        }
 303
 304                        if (int_status & SXGBE_DMA_INT_STATUS_REB1) {
 305                                x->rx_desc_access_err++;
 306                                clear_val |= SXGBE_DMA_INT_STATUS_REB1;
 307                        } else {
 308                                x->rx_buffer_access_err++;
 309                        }
 310
 311                        if (int_status & SXGBE_DMA_INT_STATUS_REB2) {
 312                                x->rx_data_transfer_err++;
 313                                clear_val |= SXGBE_DMA_INT_STATUS_REB2;
 314                        }
 315                }
 316        }
 317
 318        /* clear the served bits */
 319        writel(clear_val, ioaddr + SXGBE_DMA_CHA_STATUS_REG(channel_no));
 320
 321        return ret_val;
 322}
 323
 324/* Program the HW RX Watchdog */
 325static void sxgbe_dma_rx_watchdog(void __iomem *ioaddr, u32 riwt)
 326{
 327        u32 que_num;
 328
 329        SXGBE_FOR_EACH_QUEUE(SXGBE_RX_QUEUES, que_num) {
 330                writel(riwt,
 331                       ioaddr + SXGBE_DMA_CHA_INT_RXWATCHTMR_REG(que_num));
 332        }
 333}
 334
 335static void sxgbe_enable_tso(void __iomem *ioaddr, u8 chan_num)
 336{
 337        u32 ctrl;
 338
 339        ctrl = readl(ioaddr + SXGBE_DMA_CHA_TXCTL_REG(chan_num));
 340        ctrl |= SXGBE_DMA_CHA_TXCTL_TSE_ENABLE;
 341        writel(ctrl, ioaddr + SXGBE_DMA_CHA_TXCTL_REG(chan_num));
 342}
 343
 344static const struct sxgbe_dma_ops sxgbe_dma_ops = {
 345        .init                           = sxgbe_dma_init,
 346        .cha_init                       = sxgbe_dma_channel_init,
 347        .enable_dma_transmission        = sxgbe_enable_dma_transmission,
 348        .enable_dma_irq                 = sxgbe_enable_dma_irq,
 349        .disable_dma_irq                = sxgbe_disable_dma_irq,
 350        .start_tx                       = sxgbe_dma_start_tx,
 351        .start_tx_queue                 = sxgbe_dma_start_tx_queue,
 352        .stop_tx                        = sxgbe_dma_stop_tx,
 353        .stop_tx_queue                  = sxgbe_dma_stop_tx_queue,
 354        .start_rx                       = sxgbe_dma_start_rx,
 355        .stop_rx                        = sxgbe_dma_stop_rx,
 356        .tx_dma_int_status              = sxgbe_tx_dma_int_status,
 357        .rx_dma_int_status              = sxgbe_rx_dma_int_status,
 358        .rx_watchdog                    = sxgbe_dma_rx_watchdog,
 359        .enable_tso                     = sxgbe_enable_tso,
 360};
 361
 362const struct sxgbe_dma_ops *sxgbe_get_dma_ops(void)
 363{
 364        return &sxgbe_dma_ops;
 365}
 366