linux/drivers/net/phy/nxp-c45-tja11xx.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0
   2/* NXP C45 PHY driver
   3 * Copyright (C) 2021 NXP
   4 * Author: Radu Pirea <radu-nicolae.pirea@oss.nxp.com>
   5 */
   6
   7#include <linux/delay.h>
   8#include <linux/ethtool.h>
   9#include <linux/ethtool_netlink.h>
  10#include <linux/kernel.h>
  11#include <linux/mii.h>
  12#include <linux/module.h>
  13#include <linux/phy.h>
  14#include <linux/processor.h>
  15#include <linux/property.h>
  16#include <linux/ptp_classify.h>
  17#include <linux/ptp_clock_kernel.h>
  18#include <linux/net_tstamp.h>
  19
  20#define PHY_ID_TJA_1103                 0x001BB010
  21
  22#define PMAPMD_B100T1_PMAPMD_CTL        0x0834
  23#define B100T1_PMAPMD_CONFIG_EN         BIT(15)
  24#define B100T1_PMAPMD_MASTER            BIT(14)
  25#define MASTER_MODE                     (B100T1_PMAPMD_CONFIG_EN | \
  26                                         B100T1_PMAPMD_MASTER)
  27#define SLAVE_MODE                      (B100T1_PMAPMD_CONFIG_EN)
  28
  29#define VEND1_DEVICE_CONTROL            0x0040
  30#define DEVICE_CONTROL_RESET            BIT(15)
  31#define DEVICE_CONTROL_CONFIG_GLOBAL_EN BIT(14)
  32#define DEVICE_CONTROL_CONFIG_ALL_EN    BIT(13)
  33
  34#define VEND1_PHY_IRQ_ACK               0x80A0
  35#define VEND1_PHY_IRQ_EN                0x80A1
  36#define VEND1_PHY_IRQ_STATUS            0x80A2
  37#define PHY_IRQ_LINK_EVENT              BIT(1)
  38
  39#define VEND1_PHY_CONTROL               0x8100
  40#define PHY_CONFIG_EN                   BIT(14)
  41#define PHY_START_OP                    BIT(0)
  42
  43#define VEND1_PHY_CONFIG                0x8108
  44#define PHY_CONFIG_AUTO                 BIT(0)
  45
  46#define VEND1_SIGNAL_QUALITY            0x8320
  47#define SQI_VALID                       BIT(14)
  48#define SQI_MASK                        GENMASK(2, 0)
  49#define MAX_SQI                         SQI_MASK
  50
  51#define VEND1_CABLE_TEST                0x8330
  52#define CABLE_TEST_ENABLE               BIT(15)
  53#define CABLE_TEST_START                BIT(14)
  54#define CABLE_TEST_VALID                BIT(13)
  55#define CABLE_TEST_OK                   0x00
  56#define CABLE_TEST_SHORTED              0x01
  57#define CABLE_TEST_OPEN                 0x02
  58#define CABLE_TEST_UNKNOWN              0x07
  59
  60#define VEND1_PORT_CONTROL              0x8040
  61#define PORT_CONTROL_EN                 BIT(14)
  62
  63#define VEND1_PORT_ABILITIES            0x8046
  64#define PTP_ABILITY                     BIT(3)
  65
  66#define VEND1_PORT_INFRA_CONTROL        0xAC00
  67#define PORT_INFRA_CONTROL_EN           BIT(14)
  68
  69#define VEND1_RXID                      0xAFCC
  70#define VEND1_TXID                      0xAFCD
  71#define ID_ENABLE                       BIT(15)
  72
  73#define VEND1_ABILITIES                 0xAFC4
  74#define RGMII_ID_ABILITY                BIT(15)
  75#define RGMII_ABILITY                   BIT(14)
  76#define RMII_ABILITY                    BIT(10)
  77#define REVMII_ABILITY                  BIT(9)
  78#define MII_ABILITY                     BIT(8)
  79#define SGMII_ABILITY                   BIT(0)
  80
  81#define VEND1_MII_BASIC_CONFIG          0xAFC6
  82#define MII_BASIC_CONFIG_REV            BIT(8)
  83#define MII_BASIC_CONFIG_SGMII          0x9
  84#define MII_BASIC_CONFIG_RGMII          0x7
  85#define MII_BASIC_CONFIG_RMII           0x5
  86#define MII_BASIC_CONFIG_MII            0x4
  87
  88#define VEND1_SYMBOL_ERROR_COUNTER      0x8350
  89#define VEND1_LINK_DROP_COUNTER         0x8352
  90#define VEND1_LINK_LOSSES_AND_FAILURES  0x8353
  91#define VEND1_R_GOOD_FRAME_CNT          0xA950
  92#define VEND1_R_BAD_FRAME_CNT           0xA952
  93#define VEND1_R_RXER_FRAME_CNT          0xA954
  94#define VEND1_RX_PREAMBLE_COUNT         0xAFCE
  95#define VEND1_TX_PREAMBLE_COUNT         0xAFCF
  96#define VEND1_RX_IPG_LENGTH             0xAFD0
  97#define VEND1_TX_IPG_LENGTH             0xAFD1
  98#define COUNTER_EN                      BIT(15)
  99
 100#define VEND1_LTC_LOAD_CTRL             0x1105
 101#define READ_LTC                        BIT(2)
 102#define LOAD_LTC                        BIT(0)
 103
 104#define VEND1_LTC_WR_NSEC_0             0x1106
 105#define VEND1_LTC_WR_NSEC_1             0x1107
 106#define VEND1_LTC_WR_SEC_0              0x1108
 107#define VEND1_LTC_WR_SEC_1              0x1109
 108
 109#define VEND1_LTC_RD_NSEC_0             0x110A
 110#define VEND1_LTC_RD_NSEC_1             0x110B
 111#define VEND1_LTC_RD_SEC_0              0x110C
 112#define VEND1_LTC_RD_SEC_1              0x110D
 113
 114#define VEND1_RATE_ADJ_SUBNS_0          0x110F
 115#define VEND1_RATE_ADJ_SUBNS_1          0x1110
 116#define CLK_RATE_ADJ_LD                 BIT(15)
 117#define CLK_RATE_ADJ_DIR                BIT(14)
 118
 119#define VEND1_HW_LTC_LOCK_CTRL          0x1115
 120#define HW_LTC_LOCK_EN                  BIT(0)
 121
 122#define VEND1_PTP_IRQ_EN                0x1131
 123#define VEND1_PTP_IRQ_STATUS            0x1132
 124#define PTP_IRQ_EGR_TS                  BIT(0)
 125
 126#define VEND1_RX_TS_INSRT_CTRL          0x114D
 127#define RX_TS_INSRT_MODE2               0x02
 128
 129#define VEND1_EGR_RING_DATA_0           0x114E
 130#define VEND1_EGR_RING_DATA_1_SEQ_ID    0x114F
 131#define VEND1_EGR_RING_DATA_2_NSEC_15_0 0x1150
 132#define VEND1_EGR_RING_DATA_3           0x1151
 133#define VEND1_EGR_RING_CTRL             0x1154
 134
 135#define RING_DATA_0_DOMAIN_NUMBER       GENMASK(7, 0)
 136#define RING_DATA_0_MSG_TYPE            GENMASK(11, 8)
 137#define RING_DATA_0_SEC_4_2             GENMASK(14, 2)
 138#define RING_DATA_0_TS_VALID            BIT(15)
 139
 140#define RING_DATA_3_NSEC_29_16          GENMASK(13, 0)
 141#define RING_DATA_3_SEC_1_0             GENMASK(15, 14)
 142#define RING_DATA_5_SEC_16_5            GENMASK(15, 4)
 143#define RING_DONE                       BIT(0)
 144
 145#define TS_SEC_MASK                     GENMASK(1, 0)
 146
 147#define VEND1_PORT_FUNC_ENABLES         0x8048
 148#define PTP_ENABLE                      BIT(3)
 149
 150#define VEND1_PORT_PTP_CONTROL          0x9000
 151#define PORT_PTP_CONTROL_BYPASS         BIT(11)
 152
 153#define VEND1_PTP_CLK_PERIOD            0x1104
 154#define PTP_CLK_PERIOD_100BT1           15ULL
 155
 156#define VEND1_EVENT_MSG_FILT            0x1148
 157#define EVENT_MSG_FILT_ALL              0x0F
 158#define EVENT_MSG_FILT_NONE             0x00
 159
 160#define VEND1_TX_PIPE_DLY_NS            0x1149
 161#define VEND1_TX_PIPEDLY_SUBNS          0x114A
 162#define VEND1_RX_PIPE_DLY_NS            0x114B
 163#define VEND1_RX_PIPEDLY_SUBNS          0x114C
 164
 165#define RGMII_PERIOD_PS                 8000U
 166#define PS_PER_DEGREE                   div_u64(RGMII_PERIOD_PS, 360)
 167#define MIN_ID_PS                       1644U
 168#define MAX_ID_PS                       2260U
 169#define DEFAULT_ID_PS                   2000U
 170
 171#define PPM_TO_SUBNS_INC(ppb)   div_u64(GENMASK(31, 0) * (ppb) * \
 172                                        PTP_CLK_PERIOD_100BT1, NSEC_PER_SEC)
 173
 174#define NXP_C45_SKB_CB(skb)     ((struct nxp_c45_skb_cb *)(skb)->cb)
 175
 176struct nxp_c45_skb_cb {
 177        struct ptp_header *header;
 178        unsigned int type;
 179};
 180
 181struct nxp_c45_hwts {
 182        u32     nsec;
 183        u32     sec;
 184        u8      domain_number;
 185        u16     sequence_id;
 186        u8      msg_type;
 187};
 188
 189struct nxp_c45_phy {
 190        struct phy_device *phydev;
 191        struct mii_timestamper mii_ts;
 192        struct ptp_clock *ptp_clock;
 193        struct ptp_clock_info caps;
 194        struct sk_buff_head tx_queue;
 195        struct sk_buff_head rx_queue;
 196        /* used to access the PTP registers atomic */
 197        struct mutex ptp_lock;
 198        int hwts_tx;
 199        int hwts_rx;
 200        u32 tx_delay;
 201        u32 rx_delay;
 202};
 203
 204struct nxp_c45_phy_stats {
 205        const char      *name;
 206        u8              mmd;
 207        u16             reg;
 208        u8              off;
 209        u16             mask;
 210};
 211
 212static bool nxp_c45_poll_txts(struct phy_device *phydev)
 213{
 214        return phydev->irq <= 0;
 215}
 216
 217static int _nxp_c45_ptp_gettimex64(struct ptp_clock_info *ptp,
 218                                   struct timespec64 *ts,
 219                                   struct ptp_system_timestamp *sts)
 220{
 221        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 222
 223        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_LOAD_CTRL,
 224                      READ_LTC);
 225        ts->tv_nsec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
 226                                   VEND1_LTC_RD_NSEC_0);
 227        ts->tv_nsec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
 228                                    VEND1_LTC_RD_NSEC_1) << 16;
 229        ts->tv_sec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
 230                                  VEND1_LTC_RD_SEC_0);
 231        ts->tv_sec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
 232                                   VEND1_LTC_RD_SEC_1) << 16;
 233
 234        return 0;
 235}
 236
 237static int nxp_c45_ptp_gettimex64(struct ptp_clock_info *ptp,
 238                                  struct timespec64 *ts,
 239                                  struct ptp_system_timestamp *sts)
 240{
 241        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 242
 243        mutex_lock(&priv->ptp_lock);
 244        _nxp_c45_ptp_gettimex64(ptp, ts, sts);
 245        mutex_unlock(&priv->ptp_lock);
 246
 247        return 0;
 248}
 249
 250static int _nxp_c45_ptp_settime64(struct ptp_clock_info *ptp,
 251                                  const struct timespec64 *ts)
 252{
 253        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 254
 255        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_NSEC_0,
 256                      ts->tv_nsec);
 257        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_NSEC_1,
 258                      ts->tv_nsec >> 16);
 259        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_SEC_0,
 260                      ts->tv_sec);
 261        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_SEC_1,
 262                      ts->tv_sec >> 16);
 263        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_LOAD_CTRL,
 264                      LOAD_LTC);
 265
 266        return 0;
 267}
 268
 269static int nxp_c45_ptp_settime64(struct ptp_clock_info *ptp,
 270                                 const struct timespec64 *ts)
 271{
 272        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 273
 274        mutex_lock(&priv->ptp_lock);
 275        _nxp_c45_ptp_settime64(ptp, ts);
 276        mutex_unlock(&priv->ptp_lock);
 277
 278        return 0;
 279}
 280
 281static int nxp_c45_ptp_adjfine(struct ptp_clock_info *ptp, long scaled_ppm)
 282{
 283        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 284        s32 ppb = scaled_ppm_to_ppb(scaled_ppm);
 285        u64 subns_inc_val;
 286        bool inc;
 287
 288        mutex_lock(&priv->ptp_lock);
 289        inc = ppb >= 0;
 290        ppb = abs(ppb);
 291
 292        subns_inc_val = PPM_TO_SUBNS_INC(ppb);
 293
 294        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_RATE_ADJ_SUBNS_0,
 295                      subns_inc_val);
 296        subns_inc_val >>= 16;
 297        subns_inc_val |= CLK_RATE_ADJ_LD;
 298        if (inc)
 299                subns_inc_val |= CLK_RATE_ADJ_DIR;
 300
 301        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_RATE_ADJ_SUBNS_1,
 302                      subns_inc_val);
 303        mutex_unlock(&priv->ptp_lock);
 304
 305        return 0;
 306}
 307
 308static int nxp_c45_ptp_adjtime(struct ptp_clock_info *ptp, s64 delta)
 309{
 310        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 311        struct timespec64 now, then;
 312
 313        mutex_lock(&priv->ptp_lock);
 314        then = ns_to_timespec64(delta);
 315        _nxp_c45_ptp_gettimex64(ptp, &now, NULL);
 316        now = timespec64_add(now, then);
 317        _nxp_c45_ptp_settime64(ptp, &now);
 318        mutex_unlock(&priv->ptp_lock);
 319
 320        return 0;
 321}
 322
 323static void nxp_c45_reconstruct_ts(struct timespec64 *ts,
 324                                   struct nxp_c45_hwts *hwts)
 325{
 326        ts->tv_nsec = hwts->nsec;
 327        if ((ts->tv_sec & TS_SEC_MASK) < (hwts->sec & TS_SEC_MASK))
 328                ts->tv_sec -= TS_SEC_MASK + 1;
 329        ts->tv_sec &= ~TS_SEC_MASK;
 330        ts->tv_sec |= hwts->sec & TS_SEC_MASK;
 331}
 332
 333static bool nxp_c45_match_ts(struct ptp_header *header,
 334                             struct nxp_c45_hwts *hwts,
 335                             unsigned int type)
 336{
 337        return ntohs(header->sequence_id) == hwts->sequence_id &&
 338               ptp_get_msgtype(header, type) == hwts->msg_type &&
 339               header->domain_number  == hwts->domain_number;
 340}
 341
 342static bool nxp_c45_get_hwtxts(struct nxp_c45_phy *priv,
 343                               struct nxp_c45_hwts *hwts)
 344{
 345        bool valid;
 346        u16 reg;
 347
 348        mutex_lock(&priv->ptp_lock);
 349        phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_CTRL,
 350                      RING_DONE);
 351        reg = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_DATA_0);
 352        valid = !!(reg & RING_DATA_0_TS_VALID);
 353        if (!valid)
 354                goto nxp_c45_get_hwtxts_out;
 355
 356        hwts->domain_number = reg;
 357        hwts->msg_type = (reg & RING_DATA_0_MSG_TYPE) >> 8;
 358        hwts->sec = (reg & RING_DATA_0_SEC_4_2) >> 10;
 359        hwts->sequence_id = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
 360                                         VEND1_EGR_RING_DATA_1_SEQ_ID);
 361        hwts->nsec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
 362                                  VEND1_EGR_RING_DATA_2_NSEC_15_0);
 363        reg = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_DATA_3);
 364        hwts->nsec |= (reg & RING_DATA_3_NSEC_29_16) << 16;
 365        hwts->sec |= (reg & RING_DATA_3_SEC_1_0) >> 14;
 366
 367nxp_c45_get_hwtxts_out:
 368        mutex_unlock(&priv->ptp_lock);
 369        return valid;
 370}
 371
 372static void nxp_c45_process_txts(struct nxp_c45_phy *priv,
 373                                 struct nxp_c45_hwts *txts)
 374{
 375        struct sk_buff *skb, *tmp, *skb_match = NULL;
 376        struct skb_shared_hwtstamps shhwtstamps;
 377        struct timespec64 ts;
 378        unsigned long flags;
 379        bool ts_match;
 380        s64 ts_ns;
 381
 382        spin_lock_irqsave(&priv->tx_queue.lock, flags);
 383        skb_queue_walk_safe(&priv->tx_queue, skb, tmp) {
 384                ts_match = nxp_c45_match_ts(NXP_C45_SKB_CB(skb)->header, txts,
 385                                            NXP_C45_SKB_CB(skb)->type);
 386                if (!ts_match)
 387                        continue;
 388                skb_match = skb;
 389                __skb_unlink(skb, &priv->tx_queue);
 390                break;
 391        }
 392        spin_unlock_irqrestore(&priv->tx_queue.lock, flags);
 393
 394        if (skb_match) {
 395                nxp_c45_ptp_gettimex64(&priv->caps, &ts, NULL);
 396                nxp_c45_reconstruct_ts(&ts, txts);
 397                memset(&shhwtstamps, 0, sizeof(shhwtstamps));
 398                ts_ns = timespec64_to_ns(&ts);
 399                shhwtstamps.hwtstamp = ns_to_ktime(ts_ns);
 400                skb_complete_tx_timestamp(skb_match, &shhwtstamps);
 401        } else {
 402                phydev_warn(priv->phydev,
 403                            "the tx timestamp doesn't match with any skb\n");
 404        }
 405}
 406
 407static long nxp_c45_do_aux_work(struct ptp_clock_info *ptp)
 408{
 409        struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
 410        bool poll_txts = nxp_c45_poll_txts(priv->phydev);
 411        struct skb_shared_hwtstamps *shhwtstamps_rx;
 412        struct nxp_c45_hwts hwts;
 413        bool reschedule = false;
 414        struct timespec64 ts;
 415        struct sk_buff *skb;
 416        bool txts_valid;
 417        u32 ts_raw;
 418
 419        while (!skb_queue_empty_lockless(&priv->tx_queue) && poll_txts) {
 420                txts_valid = nxp_c45_get_hwtxts(priv, &hwts);
 421                if (unlikely(!txts_valid)) {
 422                        /* Still more skbs in the queue */
 423                        reschedule = true;
 424                        break;
 425                }
 426
 427                nxp_c45_process_txts(priv, &hwts);
 428        }
 429
 430        while ((skb = skb_dequeue(&priv->rx_queue)) != NULL) {
 431                nxp_c45_ptp_gettimex64(&priv->caps, &ts, NULL);
 432                ts_raw = __be32_to_cpu(NXP_C45_SKB_CB(skb)->header->reserved2);
 433                hwts.sec = ts_raw >> 30;
 434                hwts.nsec = ts_raw & GENMASK(29, 0);
 435                nxp_c45_reconstruct_ts(&ts, &hwts);
 436                shhwtstamps_rx = skb_hwtstamps(skb);
 437                shhwtstamps_rx->hwtstamp = ns_to_ktime(timespec64_to_ns(&ts));
 438                NXP_C45_SKB_CB(skb)->header->reserved2 = 0;
 439                netif_rx_ni(skb);
 440        }
 441
 442        return reschedule ? 1 : -1;
 443}
 444
 445static int nxp_c45_init_ptp_clock(struct nxp_c45_phy *priv)
 446{
 447        priv->caps = (struct ptp_clock_info) {
 448                .owner          = THIS_MODULE,
 449                .name           = "NXP C45 PHC",
 450                .max_adj        = 16666666,
 451                .adjfine        = nxp_c45_ptp_adjfine,
 452                .adjtime        = nxp_c45_ptp_adjtime,
 453                .gettimex64     = nxp_c45_ptp_gettimex64,
 454                .settime64      = nxp_c45_ptp_settime64,
 455                .do_aux_work    = nxp_c45_do_aux_work,
 456        };
 457
 458        priv->ptp_clock = ptp_clock_register(&priv->caps,
 459                                             &priv->phydev->mdio.dev);
 460
 461        if (IS_ERR(priv->ptp_clock))
 462                return PTR_ERR(priv->ptp_clock);
 463
 464        if (!priv->ptp_clock)
 465                return -ENOMEM;
 466
 467        return 0;
 468}
 469
 470static void nxp_c45_txtstamp(struct mii_timestamper *mii_ts,
 471                             struct sk_buff *skb, int type)
 472{
 473        struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
 474                                                mii_ts);
 475
 476        switch (priv->hwts_tx) {
 477        case HWTSTAMP_TX_ON:
 478                NXP_C45_SKB_CB(skb)->type = type;
 479                NXP_C45_SKB_CB(skb)->header = ptp_parse_header(skb, type);
 480                skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
 481                skb_queue_tail(&priv->tx_queue, skb);
 482                if (nxp_c45_poll_txts(priv->phydev))
 483                        ptp_schedule_worker(priv->ptp_clock, 0);
 484                break;
 485        case HWTSTAMP_TX_OFF:
 486        default:
 487                kfree_skb(skb);
 488                break;
 489        }
 490}
 491
 492static bool nxp_c45_rxtstamp(struct mii_timestamper *mii_ts,
 493                             struct sk_buff *skb, int type)
 494{
 495        struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
 496                                                mii_ts);
 497        struct ptp_header *header = ptp_parse_header(skb, type);
 498
 499        if (!header)
 500                return false;
 501
 502        if (!priv->hwts_rx)
 503                return false;
 504
 505        NXP_C45_SKB_CB(skb)->header = header;
 506        skb_queue_tail(&priv->rx_queue, skb);
 507        ptp_schedule_worker(priv->ptp_clock, 0);
 508
 509        return true;
 510}
 511
 512static int nxp_c45_hwtstamp(struct mii_timestamper *mii_ts,
 513                            struct ifreq *ifreq)
 514{
 515        struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
 516                                                mii_ts);
 517        struct phy_device *phydev = priv->phydev;
 518        struct hwtstamp_config cfg;
 519
 520        if (copy_from_user(&cfg, ifreq->ifr_data, sizeof(cfg)))
 521                return -EFAULT;
 522
 523        if (cfg.tx_type < 0 || cfg.tx_type > HWTSTAMP_TX_ON)
 524                return -ERANGE;
 525
 526        priv->hwts_tx = cfg.tx_type;
 527
 528        switch (cfg.rx_filter) {
 529        case HWTSTAMP_FILTER_NONE:
 530                priv->hwts_rx = 0;
 531                break;
 532        case HWTSTAMP_FILTER_PTP_V2_L2_EVENT:
 533        case HWTSTAMP_FILTER_PTP_V2_L2_SYNC:
 534        case HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ:
 535                priv->hwts_rx = 1;
 536                cfg.rx_filter = HWTSTAMP_FILTER_PTP_V2_L2_EVENT;
 537                break;
 538        default:
 539                return -ERANGE;
 540        }
 541
 542        if (priv->hwts_rx || priv->hwts_tx) {
 543                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_EVENT_MSG_FILT,
 544                              EVENT_MSG_FILT_ALL);
 545                phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
 546                                   VEND1_PORT_PTP_CONTROL,
 547                                   PORT_PTP_CONTROL_BYPASS);
 548        } else {
 549                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_EVENT_MSG_FILT,
 550                              EVENT_MSG_FILT_NONE);
 551                phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_PTP_CONTROL,
 552                                 PORT_PTP_CONTROL_BYPASS);
 553        }
 554
 555        if (nxp_c45_poll_txts(priv->phydev))
 556                goto nxp_c45_no_ptp_irq;
 557
 558        if (priv->hwts_tx)
 559                phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
 560                                 VEND1_PTP_IRQ_EN, PTP_IRQ_EGR_TS);
 561        else
 562                phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
 563                                   VEND1_PTP_IRQ_EN, PTP_IRQ_EGR_TS);
 564
 565nxp_c45_no_ptp_irq:
 566        return copy_to_user(ifreq->ifr_data, &cfg, sizeof(cfg)) ? -EFAULT : 0;
 567}
 568
 569static int nxp_c45_ts_info(struct mii_timestamper *mii_ts,
 570                           struct ethtool_ts_info *ts_info)
 571{
 572        struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
 573                                                mii_ts);
 574
 575        ts_info->so_timestamping = SOF_TIMESTAMPING_TX_HARDWARE |
 576                        SOF_TIMESTAMPING_RX_HARDWARE |
 577                        SOF_TIMESTAMPING_RAW_HARDWARE;
 578        ts_info->phc_index = ptp_clock_index(priv->ptp_clock);
 579        ts_info->tx_types = (1 << HWTSTAMP_TX_OFF) | (1 << HWTSTAMP_TX_ON);
 580        ts_info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) |
 581                        (1 << HWTSTAMP_FILTER_PTP_V2_L2_SYNC) |
 582                        (1 << HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ) |
 583                        (1 << HWTSTAMP_FILTER_PTP_V2_L2_EVENT);
 584
 585        return 0;
 586}
 587
 588static const struct nxp_c45_phy_stats nxp_c45_hw_stats[] = {
 589        { "phy_symbol_error_cnt", MDIO_MMD_VEND1,
 590                VEND1_SYMBOL_ERROR_COUNTER, 0, GENMASK(15, 0) },
 591        { "phy_link_status_drop_cnt", MDIO_MMD_VEND1,
 592                VEND1_LINK_DROP_COUNTER, 8, GENMASK(13, 8) },
 593        { "phy_link_availability_drop_cnt", MDIO_MMD_VEND1,
 594                VEND1_LINK_DROP_COUNTER, 0, GENMASK(5, 0) },
 595        { "phy_link_loss_cnt", MDIO_MMD_VEND1,
 596                VEND1_LINK_LOSSES_AND_FAILURES, 10, GENMASK(15, 10) },
 597        { "phy_link_failure_cnt", MDIO_MMD_VEND1,
 598                VEND1_LINK_LOSSES_AND_FAILURES, 0, GENMASK(9, 0) },
 599        { "r_good_frame_cnt", MDIO_MMD_VEND1,
 600                VEND1_R_GOOD_FRAME_CNT, 0, GENMASK(15, 0) },
 601        { "r_bad_frame_cnt", MDIO_MMD_VEND1,
 602                VEND1_R_BAD_FRAME_CNT, 0, GENMASK(15, 0) },
 603        { "r_rxer_frame_cnt", MDIO_MMD_VEND1,
 604                VEND1_R_RXER_FRAME_CNT, 0, GENMASK(15, 0) },
 605        { "rx_preamble_count", MDIO_MMD_VEND1,
 606                VEND1_RX_PREAMBLE_COUNT, 0, GENMASK(5, 0) },
 607        { "tx_preamble_count", MDIO_MMD_VEND1,
 608                VEND1_TX_PREAMBLE_COUNT, 0, GENMASK(5, 0) },
 609        { "rx_ipg_length", MDIO_MMD_VEND1,
 610                VEND1_RX_IPG_LENGTH, 0, GENMASK(8, 0) },
 611        { "tx_ipg_length", MDIO_MMD_VEND1,
 612                VEND1_TX_IPG_LENGTH, 0, GENMASK(8, 0) },
 613};
 614
 615static int nxp_c45_get_sset_count(struct phy_device *phydev)
 616{
 617        return ARRAY_SIZE(nxp_c45_hw_stats);
 618}
 619
 620static void nxp_c45_get_strings(struct phy_device *phydev, u8 *data)
 621{
 622        size_t i;
 623
 624        for (i = 0; i < ARRAY_SIZE(nxp_c45_hw_stats); i++) {
 625                strncpy(data + i * ETH_GSTRING_LEN,
 626                        nxp_c45_hw_stats[i].name, ETH_GSTRING_LEN);
 627        }
 628}
 629
 630static void nxp_c45_get_stats(struct phy_device *phydev,
 631                              struct ethtool_stats *stats, u64 *data)
 632{
 633        size_t i;
 634        int ret;
 635
 636        for (i = 0; i < ARRAY_SIZE(nxp_c45_hw_stats); i++) {
 637                ret = phy_read_mmd(phydev, nxp_c45_hw_stats[i].mmd,
 638                                   nxp_c45_hw_stats[i].reg);
 639                if (ret < 0) {
 640                        data[i] = U64_MAX;
 641                } else {
 642                        data[i] = ret & nxp_c45_hw_stats[i].mask;
 643                        data[i] >>= nxp_c45_hw_stats[i].off;
 644                }
 645        }
 646}
 647
 648static int nxp_c45_config_enable(struct phy_device *phydev)
 649{
 650        phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONTROL,
 651                      DEVICE_CONTROL_CONFIG_GLOBAL_EN |
 652                      DEVICE_CONTROL_CONFIG_ALL_EN);
 653        usleep_range(400, 450);
 654
 655        phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_CONTROL,
 656                      PORT_CONTROL_EN);
 657        phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONTROL,
 658                      PHY_CONFIG_EN);
 659        phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_INFRA_CONTROL,
 660                      PORT_INFRA_CONTROL_EN);
 661
 662        return 0;
 663}
 664
 665static int nxp_c45_start_op(struct phy_device *phydev)
 666{
 667        return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONTROL,
 668                                PHY_START_OP);
 669}
 670
 671static int nxp_c45_config_intr(struct phy_device *phydev)
 672{
 673        if (phydev->interrupts == PHY_INTERRUPT_ENABLED)
 674                return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
 675                                        VEND1_PHY_IRQ_EN, PHY_IRQ_LINK_EVENT);
 676        else
 677                return phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
 678                                          VEND1_PHY_IRQ_EN, PHY_IRQ_LINK_EVENT);
 679}
 680
 681static irqreturn_t nxp_c45_handle_interrupt(struct phy_device *phydev)
 682{
 683        struct nxp_c45_phy *priv = phydev->priv;
 684        irqreturn_t ret = IRQ_NONE;
 685        struct nxp_c45_hwts hwts;
 686        int irq;
 687
 688        irq = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_IRQ_STATUS);
 689        if (irq & PHY_IRQ_LINK_EVENT) {
 690                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_IRQ_ACK,
 691                              PHY_IRQ_LINK_EVENT);
 692                phy_trigger_machine(phydev);
 693                ret = IRQ_HANDLED;
 694        }
 695
 696        /* There is no need for ACK.
 697         * The irq signal will be asserted until the EGR TS FIFO will be
 698         * emptied.
 699         */
 700        irq = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_PTP_IRQ_STATUS);
 701        if (irq & PTP_IRQ_EGR_TS) {
 702                while (nxp_c45_get_hwtxts(priv, &hwts))
 703                        nxp_c45_process_txts(priv, &hwts);
 704
 705                ret = IRQ_HANDLED;
 706        }
 707
 708        return ret;
 709}
 710
 711static int nxp_c45_soft_reset(struct phy_device *phydev)
 712{
 713        int ret;
 714
 715        ret = phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONTROL,
 716                            DEVICE_CONTROL_RESET);
 717        if (ret)
 718                return ret;
 719
 720        return phy_read_mmd_poll_timeout(phydev, MDIO_MMD_VEND1,
 721                                         VEND1_DEVICE_CONTROL, ret,
 722                                         !(ret & DEVICE_CONTROL_RESET), 20000,
 723                                         240000, false);
 724}
 725
 726static int nxp_c45_cable_test_start(struct phy_device *phydev)
 727{
 728        return phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_CABLE_TEST,
 729                             CABLE_TEST_ENABLE | CABLE_TEST_START);
 730}
 731
 732static int nxp_c45_cable_test_get_status(struct phy_device *phydev,
 733                                         bool *finished)
 734{
 735        int ret;
 736        u8 cable_test_result;
 737
 738        ret = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_CABLE_TEST);
 739        if (!(ret & CABLE_TEST_VALID)) {
 740                *finished = false;
 741                return 0;
 742        }
 743
 744        *finished = true;
 745        cable_test_result = ret & GENMASK(2, 0);
 746
 747        switch (cable_test_result) {
 748        case CABLE_TEST_OK:
 749                ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
 750                                        ETHTOOL_A_CABLE_RESULT_CODE_OK);
 751                break;
 752        case CABLE_TEST_SHORTED:
 753                ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
 754                                        ETHTOOL_A_CABLE_RESULT_CODE_SAME_SHORT);
 755                break;
 756        case CABLE_TEST_OPEN:
 757                ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
 758                                        ETHTOOL_A_CABLE_RESULT_CODE_OPEN);
 759                break;
 760        default:
 761                ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
 762                                        ETHTOOL_A_CABLE_RESULT_CODE_UNSPEC);
 763        }
 764
 765        phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_CABLE_TEST,
 766                           CABLE_TEST_ENABLE);
 767
 768        return nxp_c45_start_op(phydev);
 769}
 770
 771static int nxp_c45_setup_master_slave(struct phy_device *phydev)
 772{
 773        switch (phydev->master_slave_set) {
 774        case MASTER_SLAVE_CFG_MASTER_FORCE:
 775        case MASTER_SLAVE_CFG_MASTER_PREFERRED:
 776                phy_write_mmd(phydev, MDIO_MMD_PMAPMD, PMAPMD_B100T1_PMAPMD_CTL,
 777                              MASTER_MODE);
 778                break;
 779        case MASTER_SLAVE_CFG_SLAVE_PREFERRED:
 780        case MASTER_SLAVE_CFG_SLAVE_FORCE:
 781                phy_write_mmd(phydev, MDIO_MMD_PMAPMD, PMAPMD_B100T1_PMAPMD_CTL,
 782                              SLAVE_MODE);
 783                break;
 784        case MASTER_SLAVE_CFG_UNKNOWN:
 785        case MASTER_SLAVE_CFG_UNSUPPORTED:
 786                return 0;
 787        default:
 788                phydev_warn(phydev, "Unsupported Master/Slave mode\n");
 789                return -EOPNOTSUPP;
 790        }
 791
 792        return 0;
 793}
 794
 795static int nxp_c45_read_master_slave(struct phy_device *phydev)
 796{
 797        int reg;
 798
 799        phydev->master_slave_get = MASTER_SLAVE_CFG_UNKNOWN;
 800        phydev->master_slave_state = MASTER_SLAVE_STATE_UNKNOWN;
 801
 802        reg = phy_read_mmd(phydev, MDIO_MMD_PMAPMD, PMAPMD_B100T1_PMAPMD_CTL);
 803        if (reg < 0)
 804                return reg;
 805
 806        if (reg & B100T1_PMAPMD_MASTER) {
 807                phydev->master_slave_get = MASTER_SLAVE_CFG_MASTER_FORCE;
 808                phydev->master_slave_state = MASTER_SLAVE_STATE_MASTER;
 809        } else {
 810                phydev->master_slave_get = MASTER_SLAVE_CFG_SLAVE_FORCE;
 811                phydev->master_slave_state = MASTER_SLAVE_STATE_SLAVE;
 812        }
 813
 814        return 0;
 815}
 816
 817static int nxp_c45_config_aneg(struct phy_device *phydev)
 818{
 819        return nxp_c45_setup_master_slave(phydev);
 820}
 821
 822static int nxp_c45_read_status(struct phy_device *phydev)
 823{
 824        int ret;
 825
 826        ret = genphy_c45_read_status(phydev);
 827        if (ret)
 828                return ret;
 829
 830        ret = nxp_c45_read_master_slave(phydev);
 831        if (ret)
 832                return ret;
 833
 834        return 0;
 835}
 836
 837static int nxp_c45_get_sqi(struct phy_device *phydev)
 838{
 839        int reg;
 840
 841        reg = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_SIGNAL_QUALITY);
 842        if (!(reg & SQI_VALID))
 843                return -EINVAL;
 844
 845        reg &= SQI_MASK;
 846
 847        return reg;
 848}
 849
 850static int nxp_c45_get_sqi_max(struct phy_device *phydev)
 851{
 852        return MAX_SQI;
 853}
 854
 855static int nxp_c45_check_delay(struct phy_device *phydev, u32 delay)
 856{
 857        if (delay < MIN_ID_PS) {
 858                phydev_err(phydev, "delay value smaller than %u\n", MIN_ID_PS);
 859                return -EINVAL;
 860        }
 861
 862        if (delay > MAX_ID_PS) {
 863                phydev_err(phydev, "delay value higher than %u\n", MAX_ID_PS);
 864                return -EINVAL;
 865        }
 866
 867        return 0;
 868}
 869
 870static u64 nxp_c45_get_phase_shift(u64 phase_offset_raw)
 871{
 872        /* The delay in degree phase is 73.8 + phase_offset_raw * 0.9.
 873         * To avoid floating point operations we'll multiply by 10
 874         * and get 1 decimal point precision.
 875         */
 876        phase_offset_raw *= 10;
 877        phase_offset_raw -= 738;
 878        return div_u64(phase_offset_raw, 9);
 879}
 880
 881static void nxp_c45_disable_delays(struct phy_device *phydev)
 882{
 883        phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID, ID_ENABLE);
 884        phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID, ID_ENABLE);
 885}
 886
 887static void nxp_c45_set_delays(struct phy_device *phydev)
 888{
 889        struct nxp_c45_phy *priv = phydev->priv;
 890        u64 tx_delay = priv->tx_delay;
 891        u64 rx_delay = priv->rx_delay;
 892        u64 degree;
 893
 894        if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
 895            phydev->interface == PHY_INTERFACE_MODE_RGMII_TXID) {
 896                degree = div_u64(tx_delay, PS_PER_DEGREE);
 897                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID,
 898                              ID_ENABLE | nxp_c45_get_phase_shift(degree));
 899        } else {
 900                phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID,
 901                                   ID_ENABLE);
 902        }
 903
 904        if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
 905            phydev->interface == PHY_INTERFACE_MODE_RGMII_RXID) {
 906                degree = div_u64(rx_delay, PS_PER_DEGREE);
 907                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID,
 908                              ID_ENABLE | nxp_c45_get_phase_shift(degree));
 909        } else {
 910                phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID,
 911                                   ID_ENABLE);
 912        }
 913}
 914
 915static int nxp_c45_get_delays(struct phy_device *phydev)
 916{
 917        struct nxp_c45_phy *priv = phydev->priv;
 918        int ret;
 919
 920        if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
 921            phydev->interface == PHY_INTERFACE_MODE_RGMII_TXID) {
 922                ret = device_property_read_u32(&phydev->mdio.dev,
 923                                               "tx-internal-delay-ps",
 924                                               &priv->tx_delay);
 925                if (ret)
 926                        priv->tx_delay = DEFAULT_ID_PS;
 927
 928                ret = nxp_c45_check_delay(phydev, priv->tx_delay);
 929                if (ret) {
 930                        phydev_err(phydev,
 931                                   "tx-internal-delay-ps invalid value\n");
 932                        return ret;
 933                }
 934        }
 935
 936        if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
 937            phydev->interface == PHY_INTERFACE_MODE_RGMII_RXID) {
 938                ret = device_property_read_u32(&phydev->mdio.dev,
 939                                               "rx-internal-delay-ps",
 940                                               &priv->rx_delay);
 941                if (ret)
 942                        priv->rx_delay = DEFAULT_ID_PS;
 943
 944                ret = nxp_c45_check_delay(phydev, priv->rx_delay);
 945                if (ret) {
 946                        phydev_err(phydev,
 947                                   "rx-internal-delay-ps invalid value\n");
 948                        return ret;
 949                }
 950        }
 951
 952        return 0;
 953}
 954
 955static int nxp_c45_set_phy_mode(struct phy_device *phydev)
 956{
 957        int ret;
 958
 959        ret = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_ABILITIES);
 960        phydev_dbg(phydev, "Clause 45 managed PHY abilities 0x%x\n", ret);
 961
 962        switch (phydev->interface) {
 963        case PHY_INTERFACE_MODE_RGMII:
 964                if (!(ret & RGMII_ABILITY)) {
 965                        phydev_err(phydev, "rgmii mode not supported\n");
 966                        return -EINVAL;
 967                }
 968                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
 969                              MII_BASIC_CONFIG_RGMII);
 970                nxp_c45_disable_delays(phydev);
 971                break;
 972        case PHY_INTERFACE_MODE_RGMII_ID:
 973        case PHY_INTERFACE_MODE_RGMII_TXID:
 974        case PHY_INTERFACE_MODE_RGMII_RXID:
 975                if (!(ret & RGMII_ID_ABILITY)) {
 976                        phydev_err(phydev, "rgmii-id, rgmii-txid, rgmii-rxid modes are not supported\n");
 977                        return -EINVAL;
 978                }
 979                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
 980                              MII_BASIC_CONFIG_RGMII);
 981                ret = nxp_c45_get_delays(phydev);
 982                if (ret)
 983                        return ret;
 984
 985                nxp_c45_set_delays(phydev);
 986                break;
 987        case PHY_INTERFACE_MODE_MII:
 988                if (!(ret & MII_ABILITY)) {
 989                        phydev_err(phydev, "mii mode not supported\n");
 990                        return -EINVAL;
 991                }
 992                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
 993                              MII_BASIC_CONFIG_MII);
 994                break;
 995        case PHY_INTERFACE_MODE_REVMII:
 996                if (!(ret & REVMII_ABILITY)) {
 997                        phydev_err(phydev, "rev-mii mode not supported\n");
 998                        return -EINVAL;
 999                }
1000                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
1001                              MII_BASIC_CONFIG_MII | MII_BASIC_CONFIG_REV);
1002                break;
1003        case PHY_INTERFACE_MODE_RMII:
1004                if (!(ret & RMII_ABILITY)) {
1005                        phydev_err(phydev, "rmii mode not supported\n");
1006                        return -EINVAL;
1007                }
1008                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
1009                              MII_BASIC_CONFIG_RMII);
1010                break;
1011        case PHY_INTERFACE_MODE_SGMII:
1012                if (!(ret & SGMII_ABILITY)) {
1013                        phydev_err(phydev, "sgmii mode not supported\n");
1014                        return -EINVAL;
1015                }
1016                phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
1017                              MII_BASIC_CONFIG_SGMII);
1018                break;
1019        case PHY_INTERFACE_MODE_INTERNAL:
1020                break;
1021        default:
1022                return -EINVAL;
1023        }
1024
1025        return 0;
1026}
1027
1028static int nxp_c45_config_init(struct phy_device *phydev)
1029{
1030        int ret;
1031
1032        ret = nxp_c45_config_enable(phydev);
1033        if (ret) {
1034                phydev_err(phydev, "Failed to enable config\n");
1035                return ret;
1036        }
1037
1038        /* Bug workaround for SJA1110 rev B: enable write access
1039         * to MDIO_MMD_PMAPMD
1040         */
1041        phy_write_mmd(phydev, MDIO_MMD_VEND1, 0x01F8, 1);
1042        phy_write_mmd(phydev, MDIO_MMD_VEND1, 0x01F9, 2);
1043
1044        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONFIG,
1045                         PHY_CONFIG_AUTO);
1046
1047        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_LINK_DROP_COUNTER,
1048                         COUNTER_EN);
1049        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_PREAMBLE_COUNT,
1050                         COUNTER_EN);
1051        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TX_PREAMBLE_COUNT,
1052                         COUNTER_EN);
1053        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_IPG_LENGTH,
1054                         COUNTER_EN);
1055        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TX_IPG_LENGTH,
1056                         COUNTER_EN);
1057
1058        ret = nxp_c45_set_phy_mode(phydev);
1059        if (ret)
1060                return ret;
1061
1062        phydev->autoneg = AUTONEG_DISABLE;
1063
1064        phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PTP_CLK_PERIOD,
1065                      PTP_CLK_PERIOD_100BT1);
1066        phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_HW_LTC_LOCK_CTRL,
1067                           HW_LTC_LOCK_EN);
1068        phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_TS_INSRT_CTRL,
1069                      RX_TS_INSRT_MODE2);
1070        phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_FUNC_ENABLES,
1071                         PTP_ENABLE);
1072
1073        return nxp_c45_start_op(phydev);
1074}
1075
1076static int nxp_c45_probe(struct phy_device *phydev)
1077{
1078        struct nxp_c45_phy *priv;
1079        int ptp_ability;
1080        int ret = 0;
1081
1082        priv = devm_kzalloc(&phydev->mdio.dev, sizeof(*priv), GFP_KERNEL);
1083        if (!priv)
1084                return -ENOMEM;
1085
1086        skb_queue_head_init(&priv->tx_queue);
1087        skb_queue_head_init(&priv->rx_queue);
1088
1089        priv->phydev = phydev;
1090
1091        phydev->priv = priv;
1092
1093        mutex_init(&priv->ptp_lock);
1094
1095        ptp_ability = phy_read_mmd(phydev, MDIO_MMD_VEND1,
1096                                   VEND1_PORT_ABILITIES);
1097        ptp_ability = !!(ptp_ability & PTP_ABILITY);
1098        if (!ptp_ability) {
1099                phydev_dbg(phydev, "the phy does not support PTP");
1100                goto no_ptp_support;
1101        }
1102
1103        if (IS_ENABLED(CONFIG_PTP_1588_CLOCK) &&
1104            IS_ENABLED(CONFIG_NETWORK_PHY_TIMESTAMPING)) {
1105                priv->mii_ts.rxtstamp = nxp_c45_rxtstamp;
1106                priv->mii_ts.txtstamp = nxp_c45_txtstamp;
1107                priv->mii_ts.hwtstamp = nxp_c45_hwtstamp;
1108                priv->mii_ts.ts_info = nxp_c45_ts_info;
1109                phydev->mii_ts = &priv->mii_ts;
1110                ret = nxp_c45_init_ptp_clock(priv);
1111        } else {
1112                phydev_dbg(phydev, "PTP support not enabled even if the phy supports it");
1113        }
1114
1115no_ptp_support:
1116
1117        return ret;
1118}
1119
1120static struct phy_driver nxp_c45_driver[] = {
1121        {
1122                PHY_ID_MATCH_MODEL(PHY_ID_TJA_1103),
1123                .name                   = "NXP C45 TJA1103",
1124                .features               = PHY_BASIC_T1_FEATURES,
1125                .probe                  = nxp_c45_probe,
1126                .soft_reset             = nxp_c45_soft_reset,
1127                .config_aneg            = nxp_c45_config_aneg,
1128                .config_init            = nxp_c45_config_init,
1129                .config_intr            = nxp_c45_config_intr,
1130                .handle_interrupt       = nxp_c45_handle_interrupt,
1131                .read_status            = nxp_c45_read_status,
1132                .suspend                = genphy_c45_pma_suspend,
1133                .resume                 = genphy_c45_pma_resume,
1134                .get_sset_count         = nxp_c45_get_sset_count,
1135                .get_strings            = nxp_c45_get_strings,
1136                .get_stats              = nxp_c45_get_stats,
1137                .cable_test_start       = nxp_c45_cable_test_start,
1138                .cable_test_get_status  = nxp_c45_cable_test_get_status,
1139                .set_loopback           = genphy_c45_loopback,
1140                .get_sqi                = nxp_c45_get_sqi,
1141                .get_sqi_max            = nxp_c45_get_sqi_max,
1142        },
1143};
1144
1145module_phy_driver(nxp_c45_driver);
1146
1147static struct mdio_device_id __maybe_unused nxp_c45_tbl[] = {
1148        { PHY_ID_MATCH_MODEL(PHY_ID_TJA_1103) },
1149        { /*sentinel*/ },
1150};
1151
1152MODULE_DEVICE_TABLE(mdio, nxp_c45_tbl);
1153
1154MODULE_AUTHOR("Radu Pirea <radu-nicolae.pirea@oss.nxp.com>");
1155MODULE_DESCRIPTION("NXP C45 PHY driver");
1156MODULE_LICENSE("GPL v2");
1157