mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 12:11:40 +00:00
2111375b85
In prevision to add new UAPI for hwtstamp we will be limited to the struct ethtool_ts_info that is currently passed in fixed binary format through the ETHTOOL_GET_TS_INFO ethtool ioctl. It would be good if new kernel code already started operating on an extensible kernel variant of that structure, similar in concept to struct kernel_hwtstamp_config vs struct hwtstamp_config. Since struct ethtool_ts_info is in include/uapi/linux/ethtool.h, here we introduce the kernel-only structure in include/linux/ethtool.h. The manual copy is then made in the function called by ETHTOOL_GET_TS_INFO. Acked-by: Shannon Nelson <shannon.nelson@amd.com> Acked-by: Alexandra Winter <wintera@linux.ibm.com> Signed-off-by: Kory Maincent <kory.maincent@bootlin.com> Link: https://patch.msgid.link/20240709-feature_ptp_netnext-v17-6-b5317f50df2a@bootlin.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
1996 lines
54 KiB
C
1996 lines
54 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* NXP C45 PHY driver
|
|
* Copyright 2021-2023 NXP
|
|
* Author: Radu Pirea <radu-nicolae.pirea@oss.nxp.com>
|
|
*/
|
|
|
|
#include <linux/delay.h>
|
|
#include <linux/ethtool.h>
|
|
#include <linux/ethtool_netlink.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mii.h>
|
|
#include <linux/module.h>
|
|
#include <linux/phy.h>
|
|
#include <linux/processor.h>
|
|
#include <linux/property.h>
|
|
#include <linux/ptp_classify.h>
|
|
#include <linux/net_tstamp.h>
|
|
|
|
#include "nxp-c45-tja11xx.h"
|
|
|
|
#define PHY_ID_TJA_1103 0x001BB010
|
|
#define PHY_ID_TJA_1120 0x001BB031
|
|
|
|
#define VEND1_DEVICE_CONTROL 0x0040
|
|
#define DEVICE_CONTROL_RESET BIT(15)
|
|
#define DEVICE_CONTROL_CONFIG_GLOBAL_EN BIT(14)
|
|
#define DEVICE_CONTROL_CONFIG_ALL_EN BIT(13)
|
|
|
|
#define VEND1_DEVICE_CONFIG 0x0048
|
|
|
|
#define TJA1120_VEND1_EXT_TS_MODE 0x1012
|
|
|
|
#define TJA1120_GLOBAL_INFRA_IRQ_ACK 0x2C08
|
|
#define TJA1120_GLOBAL_INFRA_IRQ_EN 0x2C0A
|
|
#define TJA1120_GLOBAL_INFRA_IRQ_STATUS 0x2C0C
|
|
#define TJA1120_DEV_BOOT_DONE BIT(1)
|
|
|
|
#define TJA1120_VEND1_PTP_TRIG_DATA_S 0x1070
|
|
|
|
#define TJA1120_EGRESS_TS_DATA_S 0x9060
|
|
#define TJA1120_EGRESS_TS_END 0x9067
|
|
#define TJA1120_TS_VALID BIT(0)
|
|
#define TJA1120_MORE_TS BIT(15)
|
|
|
|
#define VEND1_PHY_IRQ_ACK 0x80A0
|
|
#define VEND1_PHY_IRQ_EN 0x80A1
|
|
#define VEND1_PHY_IRQ_STATUS 0x80A2
|
|
#define PHY_IRQ_LINK_EVENT BIT(1)
|
|
|
|
#define VEND1_ALWAYS_ACCESSIBLE 0x801F
|
|
#define FUSA_PASS BIT(4)
|
|
|
|
#define VEND1_PHY_CONTROL 0x8100
|
|
#define PHY_CONFIG_EN BIT(14)
|
|
#define PHY_START_OP BIT(0)
|
|
|
|
#define VEND1_PHY_CONFIG 0x8108
|
|
#define PHY_CONFIG_AUTO BIT(0)
|
|
|
|
#define TJA1120_EPHY_RESETS 0x810A
|
|
#define EPHY_PCS_RESET BIT(3)
|
|
|
|
#define VEND1_SIGNAL_QUALITY 0x8320
|
|
#define SQI_VALID BIT(14)
|
|
#define SQI_MASK GENMASK(2, 0)
|
|
#define MAX_SQI SQI_MASK
|
|
|
|
#define CABLE_TEST_ENABLE BIT(15)
|
|
#define CABLE_TEST_START BIT(14)
|
|
#define CABLE_TEST_OK 0x00
|
|
#define CABLE_TEST_SHORTED 0x01
|
|
#define CABLE_TEST_OPEN 0x02
|
|
#define CABLE_TEST_UNKNOWN 0x07
|
|
|
|
#define VEND1_PORT_CONTROL 0x8040
|
|
#define PORT_CONTROL_EN BIT(14)
|
|
|
|
#define VEND1_PORT_ABILITIES 0x8046
|
|
#define MACSEC_ABILITY BIT(5)
|
|
#define PTP_ABILITY BIT(3)
|
|
|
|
#define VEND1_PORT_FUNC_IRQ_EN 0x807A
|
|
#define MACSEC_IRQS BIT(5)
|
|
#define PTP_IRQS BIT(3)
|
|
|
|
#define VEND1_PTP_IRQ_ACK 0x9008
|
|
#define EGR_TS_IRQ BIT(1)
|
|
|
|
#define VEND1_PORT_INFRA_CONTROL 0xAC00
|
|
#define PORT_INFRA_CONTROL_EN BIT(14)
|
|
|
|
#define VEND1_RXID 0xAFCC
|
|
#define VEND1_TXID 0xAFCD
|
|
#define ID_ENABLE BIT(15)
|
|
|
|
#define VEND1_ABILITIES 0xAFC4
|
|
#define RGMII_ID_ABILITY BIT(15)
|
|
#define RGMII_ABILITY BIT(14)
|
|
#define RMII_ABILITY BIT(10)
|
|
#define REVMII_ABILITY BIT(9)
|
|
#define MII_ABILITY BIT(8)
|
|
#define SGMII_ABILITY BIT(0)
|
|
|
|
#define VEND1_MII_BASIC_CONFIG 0xAFC6
|
|
#define MII_BASIC_CONFIG_REV BIT(4)
|
|
#define MII_BASIC_CONFIG_SGMII 0x9
|
|
#define MII_BASIC_CONFIG_RGMII 0x7
|
|
#define MII_BASIC_CONFIG_RMII 0x5
|
|
#define MII_BASIC_CONFIG_MII 0x4
|
|
|
|
#define VEND1_SYMBOL_ERROR_CNT_XTD 0x8351
|
|
#define EXTENDED_CNT_EN BIT(15)
|
|
#define VEND1_MONITOR_STATUS 0xAC80
|
|
#define MONITOR_RESET BIT(15)
|
|
#define VEND1_MONITOR_CONFIG 0xAC86
|
|
#define LOST_FRAMES_CNT_EN BIT(9)
|
|
#define ALL_FRAMES_CNT_EN BIT(8)
|
|
|
|
#define VEND1_SYMBOL_ERROR_COUNTER 0x8350
|
|
#define VEND1_LINK_DROP_COUNTER 0x8352
|
|
#define VEND1_LINK_LOSSES_AND_FAILURES 0x8353
|
|
#define VEND1_RX_PREAMBLE_COUNT 0xAFCE
|
|
#define VEND1_TX_PREAMBLE_COUNT 0xAFCF
|
|
#define VEND1_RX_IPG_LENGTH 0xAFD0
|
|
#define VEND1_TX_IPG_LENGTH 0xAFD1
|
|
#define COUNTER_EN BIT(15)
|
|
|
|
#define VEND1_PTP_CONFIG 0x1102
|
|
#define EXT_TRG_EDGE BIT(1)
|
|
|
|
#define TJA1120_SYNC_TRIG_FILTER 0x1010
|
|
#define PTP_TRIG_RISE_TS BIT(3)
|
|
#define PTP_TRIG_FALLING_TS BIT(2)
|
|
|
|
#define CLK_RATE_ADJ_LD BIT(15)
|
|
#define CLK_RATE_ADJ_DIR BIT(14)
|
|
|
|
#define VEND1_RX_TS_INSRT_CTRL 0x114D
|
|
#define TJA1103_RX_TS_INSRT_MODE2 0x02
|
|
|
|
#define TJA1120_RX_TS_INSRT_CTRL 0x9012
|
|
#define TJA1120_RX_TS_INSRT_EN BIT(15)
|
|
#define TJA1120_TS_INSRT_MODE BIT(4)
|
|
|
|
#define VEND1_EGR_RING_DATA_0 0x114E
|
|
#define VEND1_EGR_RING_CTRL 0x1154
|
|
|
|
#define RING_DATA_0_TS_VALID BIT(15)
|
|
|
|
#define RING_DONE BIT(0)
|
|
|
|
#define TS_SEC_MASK GENMASK(1, 0)
|
|
|
|
#define PTP_ENABLE BIT(3)
|
|
#define PHY_TEST_ENABLE BIT(0)
|
|
|
|
#define VEND1_PORT_PTP_CONTROL 0x9000
|
|
#define PORT_PTP_CONTROL_BYPASS BIT(11)
|
|
|
|
#define PTP_CLK_PERIOD_100BT1 15ULL
|
|
#define PTP_CLK_PERIOD_1000BT1 8ULL
|
|
|
|
#define EVENT_MSG_FILT_ALL 0x0F
|
|
#define EVENT_MSG_FILT_NONE 0x00
|
|
|
|
#define VEND1_GPIO_FUNC_CONFIG_BASE 0x2C40
|
|
#define GPIO_FUNC_EN BIT(15)
|
|
#define GPIO_FUNC_PTP BIT(6)
|
|
#define GPIO_SIGNAL_PTP_TRIGGER 0x01
|
|
#define GPIO_SIGNAL_PPS_OUT 0x12
|
|
#define GPIO_DISABLE 0
|
|
#define GPIO_PPS_OUT_CFG (GPIO_FUNC_EN | GPIO_FUNC_PTP | \
|
|
GPIO_SIGNAL_PPS_OUT)
|
|
#define GPIO_EXTTS_OUT_CFG (GPIO_FUNC_EN | GPIO_FUNC_PTP | \
|
|
GPIO_SIGNAL_PTP_TRIGGER)
|
|
|
|
#define RGMII_PERIOD_PS 8000U
|
|
#define PS_PER_DEGREE div_u64(RGMII_PERIOD_PS, 360)
|
|
#define MIN_ID_PS 1644U
|
|
#define MAX_ID_PS 2260U
|
|
#define DEFAULT_ID_PS 2000U
|
|
|
|
#define PPM_TO_SUBNS_INC(ppb, ptp_clk_period) div_u64(GENMASK_ULL(31, 0) * \
|
|
(ppb) * (ptp_clk_period), NSEC_PER_SEC)
|
|
|
|
#define NXP_C45_SKB_CB(skb) ((struct nxp_c45_skb_cb *)(skb)->cb)
|
|
|
|
struct nxp_c45_phy;
|
|
|
|
struct nxp_c45_skb_cb {
|
|
struct ptp_header *header;
|
|
unsigned int type;
|
|
};
|
|
|
|
#define NXP_C45_REG_FIELD(_reg, _devad, _offset, _size) \
|
|
((struct nxp_c45_reg_field) { \
|
|
.reg = _reg, \
|
|
.devad = _devad, \
|
|
.offset = _offset, \
|
|
.size = _size, \
|
|
})
|
|
|
|
struct nxp_c45_reg_field {
|
|
u16 reg;
|
|
u8 devad;
|
|
u8 offset;
|
|
u8 size;
|
|
};
|
|
|
|
struct nxp_c45_hwts {
|
|
u32 nsec;
|
|
u32 sec;
|
|
u8 domain_number;
|
|
u16 sequence_id;
|
|
u8 msg_type;
|
|
};
|
|
|
|
struct nxp_c45_regmap {
|
|
/* PTP config regs. */
|
|
u16 vend1_ptp_clk_period;
|
|
u16 vend1_event_msg_filt;
|
|
|
|
/* LTC bits and regs. */
|
|
struct nxp_c45_reg_field ltc_read;
|
|
struct nxp_c45_reg_field ltc_write;
|
|
struct nxp_c45_reg_field ltc_lock_ctrl;
|
|
u16 vend1_ltc_wr_nsec_0;
|
|
u16 vend1_ltc_wr_nsec_1;
|
|
u16 vend1_ltc_wr_sec_0;
|
|
u16 vend1_ltc_wr_sec_1;
|
|
u16 vend1_ltc_rd_nsec_0;
|
|
u16 vend1_ltc_rd_nsec_1;
|
|
u16 vend1_ltc_rd_sec_0;
|
|
u16 vend1_ltc_rd_sec_1;
|
|
u16 vend1_rate_adj_subns_0;
|
|
u16 vend1_rate_adj_subns_1;
|
|
|
|
/* External trigger reg fields. */
|
|
struct nxp_c45_reg_field irq_egr_ts_en;
|
|
struct nxp_c45_reg_field irq_egr_ts_status;
|
|
struct nxp_c45_reg_field domain_number;
|
|
struct nxp_c45_reg_field msg_type;
|
|
struct nxp_c45_reg_field sequence_id;
|
|
struct nxp_c45_reg_field sec_1_0;
|
|
struct nxp_c45_reg_field sec_4_2;
|
|
struct nxp_c45_reg_field nsec_15_0;
|
|
struct nxp_c45_reg_field nsec_29_16;
|
|
|
|
/* PPS and EXT Trigger bits and regs. */
|
|
struct nxp_c45_reg_field pps_enable;
|
|
struct nxp_c45_reg_field pps_polarity;
|
|
u16 vend1_ext_trg_data_0;
|
|
u16 vend1_ext_trg_data_1;
|
|
u16 vend1_ext_trg_data_2;
|
|
u16 vend1_ext_trg_data_3;
|
|
u16 vend1_ext_trg_ctrl;
|
|
|
|
/* Cable test reg fields. */
|
|
u16 cable_test;
|
|
struct nxp_c45_reg_field cable_test_valid;
|
|
struct nxp_c45_reg_field cable_test_result;
|
|
};
|
|
|
|
struct nxp_c45_phy_stats {
|
|
const char *name;
|
|
const struct nxp_c45_reg_field counter;
|
|
};
|
|
|
|
struct nxp_c45_phy_data {
|
|
const struct nxp_c45_regmap *regmap;
|
|
const struct nxp_c45_phy_stats *stats;
|
|
int n_stats;
|
|
u8 ptp_clk_period;
|
|
bool ext_ts_both_edges;
|
|
bool ack_ptp_irq;
|
|
void (*counters_enable)(struct phy_device *phydev);
|
|
bool (*get_egressts)(struct nxp_c45_phy *priv,
|
|
struct nxp_c45_hwts *hwts);
|
|
bool (*get_extts)(struct nxp_c45_phy *priv, struct timespec64 *extts);
|
|
void (*ptp_init)(struct phy_device *phydev);
|
|
void (*ptp_enable)(struct phy_device *phydev, bool enable);
|
|
void (*nmi_handler)(struct phy_device *phydev,
|
|
irqreturn_t *irq_status);
|
|
};
|
|
|
|
static const
|
|
struct nxp_c45_phy_data *nxp_c45_get_data(struct phy_device *phydev)
|
|
{
|
|
return phydev->drv->driver_data;
|
|
}
|
|
|
|
static const
|
|
struct nxp_c45_regmap *nxp_c45_get_regmap(struct phy_device *phydev)
|
|
{
|
|
const struct nxp_c45_phy_data *phy_data = nxp_c45_get_data(phydev);
|
|
|
|
return phy_data->regmap;
|
|
}
|
|
|
|
static int nxp_c45_read_reg_field(struct phy_device *phydev,
|
|
const struct nxp_c45_reg_field *reg_field)
|
|
{
|
|
u16 mask;
|
|
int ret;
|
|
|
|
if (reg_field->size == 0) {
|
|
phydev_err(phydev, "Trying to read a reg field of size 0.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = phy_read_mmd(phydev, reg_field->devad, reg_field->reg);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
mask = reg_field->size == 1 ? BIT(reg_field->offset) :
|
|
GENMASK(reg_field->offset + reg_field->size - 1,
|
|
reg_field->offset);
|
|
ret &= mask;
|
|
ret >>= reg_field->offset;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int nxp_c45_write_reg_field(struct phy_device *phydev,
|
|
const struct nxp_c45_reg_field *reg_field,
|
|
u16 val)
|
|
{
|
|
u16 mask;
|
|
u16 set;
|
|
|
|
if (reg_field->size == 0) {
|
|
phydev_err(phydev, "Trying to write a reg field of size 0.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
mask = reg_field->size == 1 ? BIT(reg_field->offset) :
|
|
GENMASK(reg_field->offset + reg_field->size - 1,
|
|
reg_field->offset);
|
|
set = val << reg_field->offset;
|
|
|
|
return phy_modify_mmd_changed(phydev, reg_field->devad,
|
|
reg_field->reg, mask, set);
|
|
}
|
|
|
|
static int nxp_c45_set_reg_field(struct phy_device *phydev,
|
|
const struct nxp_c45_reg_field *reg_field)
|
|
{
|
|
if (reg_field->size != 1) {
|
|
phydev_err(phydev, "Trying to set a reg field of size different than 1.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return nxp_c45_write_reg_field(phydev, reg_field, 1);
|
|
}
|
|
|
|
static int nxp_c45_clear_reg_field(struct phy_device *phydev,
|
|
const struct nxp_c45_reg_field *reg_field)
|
|
{
|
|
if (reg_field->size != 1) {
|
|
phydev_err(phydev, "Trying to set a reg field of size different than 1.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return nxp_c45_write_reg_field(phydev, reg_field, 0);
|
|
}
|
|
|
|
static bool nxp_c45_poll_txts(struct phy_device *phydev)
|
|
{
|
|
return phydev->irq <= 0;
|
|
}
|
|
|
|
static int _nxp_c45_ptp_gettimex64(struct ptp_clock_info *ptp,
|
|
struct timespec64 *ts,
|
|
struct ptp_system_timestamp *sts)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(priv->phydev);
|
|
|
|
nxp_c45_set_reg_field(priv->phydev, ®map->ltc_read);
|
|
ts->tv_nsec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ltc_rd_nsec_0);
|
|
ts->tv_nsec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ltc_rd_nsec_1) << 16;
|
|
ts->tv_sec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ltc_rd_sec_0);
|
|
ts->tv_sec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ltc_rd_sec_1) << 16;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_ptp_gettimex64(struct ptp_clock_info *ptp,
|
|
struct timespec64 *ts,
|
|
struct ptp_system_timestamp *sts)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
|
|
mutex_lock(&priv->ptp_lock);
|
|
_nxp_c45_ptp_gettimex64(ptp, ts, sts);
|
|
mutex_unlock(&priv->ptp_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int _nxp_c45_ptp_settime64(struct ptp_clock_info *ptp,
|
|
const struct timespec64 *ts)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(priv->phydev);
|
|
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, regmap->vend1_ltc_wr_nsec_0,
|
|
ts->tv_nsec);
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, regmap->vend1_ltc_wr_nsec_1,
|
|
ts->tv_nsec >> 16);
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, regmap->vend1_ltc_wr_sec_0,
|
|
ts->tv_sec);
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, regmap->vend1_ltc_wr_sec_1,
|
|
ts->tv_sec >> 16);
|
|
nxp_c45_set_reg_field(priv->phydev, ®map->ltc_write);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_ptp_settime64(struct ptp_clock_info *ptp,
|
|
const struct timespec64 *ts)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
|
|
mutex_lock(&priv->ptp_lock);
|
|
_nxp_c45_ptp_settime64(ptp, ts);
|
|
mutex_unlock(&priv->ptp_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_ptp_adjfine(struct ptp_clock_info *ptp, long scaled_ppm)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
const struct nxp_c45_phy_data *data = nxp_c45_get_data(priv->phydev);
|
|
const struct nxp_c45_regmap *regmap = data->regmap;
|
|
s32 ppb = scaled_ppm_to_ppb(scaled_ppm);
|
|
u64 subns_inc_val;
|
|
bool inc;
|
|
|
|
mutex_lock(&priv->ptp_lock);
|
|
inc = ppb >= 0;
|
|
ppb = abs(ppb);
|
|
|
|
subns_inc_val = PPM_TO_SUBNS_INC(ppb, data->ptp_clk_period);
|
|
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_rate_adj_subns_0,
|
|
subns_inc_val);
|
|
subns_inc_val >>= 16;
|
|
subns_inc_val |= CLK_RATE_ADJ_LD;
|
|
if (inc)
|
|
subns_inc_val |= CLK_RATE_ADJ_DIR;
|
|
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_rate_adj_subns_1,
|
|
subns_inc_val);
|
|
mutex_unlock(&priv->ptp_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_ptp_adjtime(struct ptp_clock_info *ptp, s64 delta)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
struct timespec64 now, then;
|
|
|
|
mutex_lock(&priv->ptp_lock);
|
|
then = ns_to_timespec64(delta);
|
|
_nxp_c45_ptp_gettimex64(ptp, &now, NULL);
|
|
now = timespec64_add(now, then);
|
|
_nxp_c45_ptp_settime64(ptp, &now);
|
|
mutex_unlock(&priv->ptp_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void nxp_c45_reconstruct_ts(struct timespec64 *ts,
|
|
struct nxp_c45_hwts *hwts)
|
|
{
|
|
ts->tv_nsec = hwts->nsec;
|
|
if ((ts->tv_sec & TS_SEC_MASK) < (hwts->sec & TS_SEC_MASK))
|
|
ts->tv_sec -= TS_SEC_MASK + 1;
|
|
ts->tv_sec &= ~TS_SEC_MASK;
|
|
ts->tv_sec |= hwts->sec & TS_SEC_MASK;
|
|
}
|
|
|
|
static bool nxp_c45_match_ts(struct ptp_header *header,
|
|
struct nxp_c45_hwts *hwts,
|
|
unsigned int type)
|
|
{
|
|
return ntohs(header->sequence_id) == hwts->sequence_id &&
|
|
ptp_get_msgtype(header, type) == hwts->msg_type &&
|
|
header->domain_number == hwts->domain_number;
|
|
}
|
|
|
|
static bool nxp_c45_get_extts(struct nxp_c45_phy *priv,
|
|
struct timespec64 *extts)
|
|
{
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(priv->phydev);
|
|
|
|
extts->tv_nsec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_data_0);
|
|
extts->tv_nsec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_data_1) << 16;
|
|
extts->tv_sec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_data_2);
|
|
extts->tv_sec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_data_3) << 16;
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_ctrl, RING_DONE);
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool tja1120_extts_is_valid(struct phy_device *phydev)
|
|
{
|
|
bool valid;
|
|
int reg;
|
|
|
|
reg = phy_read_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_VEND1_PTP_TRIG_DATA_S);
|
|
valid = !!(reg & TJA1120_TS_VALID);
|
|
|
|
return valid;
|
|
}
|
|
|
|
static bool tja1120_get_extts(struct nxp_c45_phy *priv,
|
|
struct timespec64 *extts)
|
|
{
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(priv->phydev);
|
|
struct phy_device *phydev = priv->phydev;
|
|
bool more_ts;
|
|
bool valid;
|
|
u16 reg;
|
|
|
|
reg = phy_read_mmd(phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_ctrl);
|
|
more_ts = !!(reg & TJA1120_MORE_TS);
|
|
|
|
valid = tja1120_extts_is_valid(phydev);
|
|
if (!valid) {
|
|
if (!more_ts)
|
|
goto tja1120_get_extts_out;
|
|
|
|
/* Bug workaround for TJA1120 engineering samples: move the new
|
|
* timestamp from the FIFO to the buffer.
|
|
*/
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
regmap->vend1_ext_trg_ctrl, RING_DONE);
|
|
valid = tja1120_extts_is_valid(phydev);
|
|
if (!valid)
|
|
goto tja1120_get_extts_out;
|
|
}
|
|
|
|
nxp_c45_get_extts(priv, extts);
|
|
tja1120_get_extts_out:
|
|
return valid;
|
|
}
|
|
|
|
static void nxp_c45_read_egress_ts(struct nxp_c45_phy *priv,
|
|
struct nxp_c45_hwts *hwts)
|
|
{
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(priv->phydev);
|
|
struct phy_device *phydev = priv->phydev;
|
|
|
|
hwts->domain_number =
|
|
nxp_c45_read_reg_field(phydev, ®map->domain_number);
|
|
hwts->msg_type =
|
|
nxp_c45_read_reg_field(phydev, ®map->msg_type);
|
|
hwts->sequence_id =
|
|
nxp_c45_read_reg_field(phydev, ®map->sequence_id);
|
|
hwts->nsec =
|
|
nxp_c45_read_reg_field(phydev, ®map->nsec_15_0);
|
|
hwts->nsec |=
|
|
nxp_c45_read_reg_field(phydev, ®map->nsec_29_16) << 16;
|
|
hwts->sec = nxp_c45_read_reg_field(phydev, ®map->sec_1_0);
|
|
hwts->sec |= nxp_c45_read_reg_field(phydev, ®map->sec_4_2) << 2;
|
|
}
|
|
|
|
static bool nxp_c45_get_hwtxts(struct nxp_c45_phy *priv,
|
|
struct nxp_c45_hwts *hwts)
|
|
{
|
|
bool valid;
|
|
u16 reg;
|
|
|
|
mutex_lock(&priv->ptp_lock);
|
|
phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_CTRL,
|
|
RING_DONE);
|
|
reg = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_DATA_0);
|
|
valid = !!(reg & RING_DATA_0_TS_VALID);
|
|
if (!valid)
|
|
goto nxp_c45_get_hwtxts_out;
|
|
|
|
nxp_c45_read_egress_ts(priv, hwts);
|
|
nxp_c45_get_hwtxts_out:
|
|
mutex_unlock(&priv->ptp_lock);
|
|
return valid;
|
|
}
|
|
|
|
static bool tja1120_egress_ts_is_valid(struct phy_device *phydev)
|
|
{
|
|
bool valid;
|
|
u16 reg;
|
|
|
|
reg = phy_read_mmd(phydev, MDIO_MMD_VEND1, TJA1120_EGRESS_TS_DATA_S);
|
|
valid = !!(reg & TJA1120_TS_VALID);
|
|
|
|
return valid;
|
|
}
|
|
|
|
static bool tja1120_get_hwtxts(struct nxp_c45_phy *priv,
|
|
struct nxp_c45_hwts *hwts)
|
|
{
|
|
struct phy_device *phydev = priv->phydev;
|
|
bool more_ts;
|
|
bool valid;
|
|
u16 reg;
|
|
|
|
mutex_lock(&priv->ptp_lock);
|
|
reg = phy_read_mmd(phydev, MDIO_MMD_VEND1, TJA1120_EGRESS_TS_END);
|
|
more_ts = !!(reg & TJA1120_MORE_TS);
|
|
valid = tja1120_egress_ts_is_valid(phydev);
|
|
if (!valid) {
|
|
if (!more_ts)
|
|
goto tja1120_get_hwtxts_out;
|
|
|
|
/* Bug workaround for TJA1120 engineering samples: move the
|
|
* new timestamp from the FIFO to the buffer.
|
|
*/
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_EGRESS_TS_END, TJA1120_TS_VALID);
|
|
valid = tja1120_egress_ts_is_valid(phydev);
|
|
if (!valid)
|
|
goto tja1120_get_hwtxts_out;
|
|
}
|
|
nxp_c45_read_egress_ts(priv, hwts);
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, TJA1120_EGRESS_TS_DATA_S,
|
|
TJA1120_TS_VALID);
|
|
tja1120_get_hwtxts_out:
|
|
mutex_unlock(&priv->ptp_lock);
|
|
return valid;
|
|
}
|
|
|
|
static void nxp_c45_process_txts(struct nxp_c45_phy *priv,
|
|
struct nxp_c45_hwts *txts)
|
|
{
|
|
struct sk_buff *skb, *tmp, *skb_match = NULL;
|
|
struct skb_shared_hwtstamps shhwtstamps;
|
|
struct timespec64 ts;
|
|
unsigned long flags;
|
|
bool ts_match;
|
|
s64 ts_ns;
|
|
|
|
spin_lock_irqsave(&priv->tx_queue.lock, flags);
|
|
skb_queue_walk_safe(&priv->tx_queue, skb, tmp) {
|
|
ts_match = nxp_c45_match_ts(NXP_C45_SKB_CB(skb)->header, txts,
|
|
NXP_C45_SKB_CB(skb)->type);
|
|
if (!ts_match)
|
|
continue;
|
|
skb_match = skb;
|
|
__skb_unlink(skb, &priv->tx_queue);
|
|
break;
|
|
}
|
|
spin_unlock_irqrestore(&priv->tx_queue.lock, flags);
|
|
|
|
if (skb_match) {
|
|
nxp_c45_ptp_gettimex64(&priv->caps, &ts, NULL);
|
|
nxp_c45_reconstruct_ts(&ts, txts);
|
|
memset(&shhwtstamps, 0, sizeof(shhwtstamps));
|
|
ts_ns = timespec64_to_ns(&ts);
|
|
shhwtstamps.hwtstamp = ns_to_ktime(ts_ns);
|
|
skb_complete_tx_timestamp(skb_match, &shhwtstamps);
|
|
} else {
|
|
phydev_warn(priv->phydev,
|
|
"the tx timestamp doesn't match with any skb\n");
|
|
}
|
|
}
|
|
|
|
static long nxp_c45_do_aux_work(struct ptp_clock_info *ptp)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
const struct nxp_c45_phy_data *data = nxp_c45_get_data(priv->phydev);
|
|
bool poll_txts = nxp_c45_poll_txts(priv->phydev);
|
|
struct skb_shared_hwtstamps *shhwtstamps_rx;
|
|
struct ptp_clock_event event;
|
|
struct nxp_c45_hwts hwts;
|
|
bool reschedule = false;
|
|
struct timespec64 ts;
|
|
struct sk_buff *skb;
|
|
bool ts_valid;
|
|
u32 ts_raw;
|
|
|
|
while (!skb_queue_empty_lockless(&priv->tx_queue) && poll_txts) {
|
|
ts_valid = data->get_egressts(priv, &hwts);
|
|
if (unlikely(!ts_valid)) {
|
|
/* Still more skbs in the queue */
|
|
reschedule = true;
|
|
break;
|
|
}
|
|
|
|
nxp_c45_process_txts(priv, &hwts);
|
|
}
|
|
|
|
while ((skb = skb_dequeue(&priv->rx_queue)) != NULL) {
|
|
nxp_c45_ptp_gettimex64(&priv->caps, &ts, NULL);
|
|
ts_raw = __be32_to_cpu(NXP_C45_SKB_CB(skb)->header->reserved2);
|
|
hwts.sec = ts_raw >> 30;
|
|
hwts.nsec = ts_raw & GENMASK(29, 0);
|
|
nxp_c45_reconstruct_ts(&ts, &hwts);
|
|
shhwtstamps_rx = skb_hwtstamps(skb);
|
|
shhwtstamps_rx->hwtstamp = ns_to_ktime(timespec64_to_ns(&ts));
|
|
NXP_C45_SKB_CB(skb)->header->reserved2 = 0;
|
|
netif_rx(skb);
|
|
}
|
|
|
|
if (priv->extts) {
|
|
ts_valid = data->get_extts(priv, &ts);
|
|
if (ts_valid && timespec64_compare(&ts, &priv->extts_ts) != 0) {
|
|
priv->extts_ts = ts;
|
|
event.index = priv->extts_index;
|
|
event.type = PTP_CLOCK_EXTTS;
|
|
event.timestamp = ns_to_ktime(timespec64_to_ns(&ts));
|
|
ptp_clock_event(priv->ptp_clock, &event);
|
|
}
|
|
reschedule = true;
|
|
}
|
|
|
|
return reschedule ? 1 : -1;
|
|
}
|
|
|
|
static void nxp_c45_gpio_config(struct nxp_c45_phy *priv,
|
|
int pin, u16 pin_cfg)
|
|
{
|
|
struct phy_device *phydev = priv->phydev;
|
|
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_GPIO_FUNC_CONFIG_BASE + pin, pin_cfg);
|
|
}
|
|
|
|
static int nxp_c45_perout_enable(struct nxp_c45_phy *priv,
|
|
struct ptp_perout_request *perout, int on)
|
|
{
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(priv->phydev);
|
|
struct phy_device *phydev = priv->phydev;
|
|
int pin;
|
|
|
|
if (perout->flags & ~PTP_PEROUT_PHASE)
|
|
return -EOPNOTSUPP;
|
|
|
|
pin = ptp_find_pin(priv->ptp_clock, PTP_PF_PEROUT, perout->index);
|
|
if (pin < 0)
|
|
return pin;
|
|
|
|
if (!on) {
|
|
nxp_c45_clear_reg_field(priv->phydev,
|
|
®map->pps_enable);
|
|
nxp_c45_clear_reg_field(priv->phydev,
|
|
®map->pps_polarity);
|
|
|
|
nxp_c45_gpio_config(priv, pin, GPIO_DISABLE);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* The PPS signal is fixed to 1 second and is always generated when the
|
|
* seconds counter is incremented. The start time is not configurable.
|
|
* If the clock is adjusted, the PPS signal is automatically readjusted.
|
|
*/
|
|
if (perout->period.sec != 1 || perout->period.nsec != 0) {
|
|
phydev_warn(phydev, "The period can be set only to 1 second.");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!(perout->flags & PTP_PEROUT_PHASE)) {
|
|
if (perout->start.sec != 0 || perout->start.nsec != 0) {
|
|
phydev_warn(phydev, "The start time is not configurable. Should be set to 0 seconds and 0 nanoseconds.");
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
if (perout->phase.nsec != 0 &&
|
|
perout->phase.nsec != (NSEC_PER_SEC >> 1)) {
|
|
phydev_warn(phydev, "The phase can be set only to 0 or 500000000 nanoseconds.");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (perout->phase.nsec == 0)
|
|
nxp_c45_clear_reg_field(priv->phydev,
|
|
®map->pps_polarity);
|
|
else
|
|
nxp_c45_set_reg_field(priv->phydev,
|
|
®map->pps_polarity);
|
|
}
|
|
|
|
nxp_c45_gpio_config(priv, pin, GPIO_PPS_OUT_CFG);
|
|
|
|
nxp_c45_set_reg_field(priv->phydev, ®map->pps_enable);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void nxp_c45_set_rising_or_falling(struct phy_device *phydev,
|
|
struct ptp_extts_request *extts)
|
|
{
|
|
if (extts->flags & PTP_RISING_EDGE)
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PTP_CONFIG, EXT_TRG_EDGE);
|
|
|
|
if (extts->flags & PTP_FALLING_EDGE)
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PTP_CONFIG, EXT_TRG_EDGE);
|
|
}
|
|
|
|
static void nxp_c45_set_rising_and_falling(struct phy_device *phydev,
|
|
struct ptp_extts_request *extts)
|
|
{
|
|
/* PTP_EXTTS_REQUEST may have only the PTP_ENABLE_FEATURE flag set. In
|
|
* this case external ts will be enabled on rising edge.
|
|
*/
|
|
if (extts->flags & PTP_RISING_EDGE ||
|
|
extts->flags == PTP_ENABLE_FEATURE)
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_SYNC_TRIG_FILTER,
|
|
PTP_TRIG_RISE_TS);
|
|
else
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_SYNC_TRIG_FILTER,
|
|
PTP_TRIG_RISE_TS);
|
|
|
|
if (extts->flags & PTP_FALLING_EDGE)
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_SYNC_TRIG_FILTER,
|
|
PTP_TRIG_FALLING_TS);
|
|
else
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_SYNC_TRIG_FILTER,
|
|
PTP_TRIG_FALLING_TS);
|
|
}
|
|
|
|
static int nxp_c45_extts_enable(struct nxp_c45_phy *priv,
|
|
struct ptp_extts_request *extts, int on)
|
|
{
|
|
const struct nxp_c45_phy_data *data = nxp_c45_get_data(priv->phydev);
|
|
int pin;
|
|
|
|
if (extts->flags & ~(PTP_ENABLE_FEATURE |
|
|
PTP_RISING_EDGE |
|
|
PTP_FALLING_EDGE |
|
|
PTP_STRICT_FLAGS))
|
|
return -EOPNOTSUPP;
|
|
|
|
/* Sampling on both edges is not supported */
|
|
if ((extts->flags & PTP_RISING_EDGE) &&
|
|
(extts->flags & PTP_FALLING_EDGE) &&
|
|
!data->ext_ts_both_edges)
|
|
return -EOPNOTSUPP;
|
|
|
|
pin = ptp_find_pin(priv->ptp_clock, PTP_PF_EXTTS, extts->index);
|
|
if (pin < 0)
|
|
return pin;
|
|
|
|
if (!on) {
|
|
nxp_c45_gpio_config(priv, pin, GPIO_DISABLE);
|
|
priv->extts = false;
|
|
|
|
return 0;
|
|
}
|
|
|
|
if (data->ext_ts_both_edges)
|
|
nxp_c45_set_rising_and_falling(priv->phydev, extts);
|
|
else
|
|
nxp_c45_set_rising_or_falling(priv->phydev, extts);
|
|
|
|
nxp_c45_gpio_config(priv, pin, GPIO_EXTTS_OUT_CFG);
|
|
priv->extts = true;
|
|
priv->extts_index = extts->index;
|
|
ptp_schedule_worker(priv->ptp_clock, 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_ptp_enable(struct ptp_clock_info *ptp,
|
|
struct ptp_clock_request *req, int on)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
|
|
|
|
switch (req->type) {
|
|
case PTP_CLK_REQ_EXTTS:
|
|
return nxp_c45_extts_enable(priv, &req->extts, on);
|
|
case PTP_CLK_REQ_PEROUT:
|
|
return nxp_c45_perout_enable(priv, &req->perout, on);
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
}
|
|
|
|
static struct ptp_pin_desc nxp_c45_ptp_pins[] = {
|
|
{ "nxp_c45_gpio0", 0, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio1", 1, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio2", 2, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio3", 3, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio4", 4, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio5", 5, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio6", 6, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio7", 7, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio8", 8, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio9", 9, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio10", 10, PTP_PF_NONE},
|
|
{ "nxp_c45_gpio11", 11, PTP_PF_NONE},
|
|
};
|
|
|
|
static int nxp_c45_ptp_verify_pin(struct ptp_clock_info *ptp, unsigned int pin,
|
|
enum ptp_pin_function func, unsigned int chan)
|
|
{
|
|
if (pin >= ARRAY_SIZE(nxp_c45_ptp_pins))
|
|
return -EINVAL;
|
|
|
|
switch (func) {
|
|
case PTP_PF_NONE:
|
|
case PTP_PF_PEROUT:
|
|
case PTP_PF_EXTTS:
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_init_ptp_clock(struct nxp_c45_phy *priv)
|
|
{
|
|
priv->caps = (struct ptp_clock_info) {
|
|
.owner = THIS_MODULE,
|
|
.name = "NXP C45 PHC",
|
|
.max_adj = 16666666,
|
|
.adjfine = nxp_c45_ptp_adjfine,
|
|
.adjtime = nxp_c45_ptp_adjtime,
|
|
.gettimex64 = nxp_c45_ptp_gettimex64,
|
|
.settime64 = nxp_c45_ptp_settime64,
|
|
.enable = nxp_c45_ptp_enable,
|
|
.verify = nxp_c45_ptp_verify_pin,
|
|
.do_aux_work = nxp_c45_do_aux_work,
|
|
.pin_config = nxp_c45_ptp_pins,
|
|
.n_pins = ARRAY_SIZE(nxp_c45_ptp_pins),
|
|
.n_ext_ts = 1,
|
|
.n_per_out = 1,
|
|
};
|
|
|
|
priv->ptp_clock = ptp_clock_register(&priv->caps,
|
|
&priv->phydev->mdio.dev);
|
|
|
|
if (IS_ERR(priv->ptp_clock))
|
|
return PTR_ERR(priv->ptp_clock);
|
|
|
|
if (!priv->ptp_clock)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void nxp_c45_txtstamp(struct mii_timestamper *mii_ts,
|
|
struct sk_buff *skb, int type)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
|
|
mii_ts);
|
|
|
|
switch (priv->hwts_tx) {
|
|
case HWTSTAMP_TX_ON:
|
|
NXP_C45_SKB_CB(skb)->type = type;
|
|
NXP_C45_SKB_CB(skb)->header = ptp_parse_header(skb, type);
|
|
skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
|
|
skb_queue_tail(&priv->tx_queue, skb);
|
|
if (nxp_c45_poll_txts(priv->phydev))
|
|
ptp_schedule_worker(priv->ptp_clock, 0);
|
|
break;
|
|
case HWTSTAMP_TX_OFF:
|
|
default:
|
|
kfree_skb(skb);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static bool nxp_c45_rxtstamp(struct mii_timestamper *mii_ts,
|
|
struct sk_buff *skb, int type)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
|
|
mii_ts);
|
|
struct ptp_header *header = ptp_parse_header(skb, type);
|
|
|
|
if (!header)
|
|
return false;
|
|
|
|
if (!priv->hwts_rx)
|
|
return false;
|
|
|
|
NXP_C45_SKB_CB(skb)->header = header;
|
|
skb_queue_tail(&priv->rx_queue, skb);
|
|
ptp_schedule_worker(priv->ptp_clock, 0);
|
|
|
|
return true;
|
|
}
|
|
|
|
static int nxp_c45_hwtstamp(struct mii_timestamper *mii_ts,
|
|
struct kernel_hwtstamp_config *cfg,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
|
|
mii_ts);
|
|
struct phy_device *phydev = priv->phydev;
|
|
const struct nxp_c45_phy_data *data;
|
|
|
|
if (cfg->tx_type < 0 || cfg->tx_type > HWTSTAMP_TX_ON)
|
|
return -ERANGE;
|
|
|
|
data = nxp_c45_get_data(phydev);
|
|
priv->hwts_tx = cfg->tx_type;
|
|
|
|
switch (cfg->rx_filter) {
|
|
case HWTSTAMP_FILTER_NONE:
|
|
priv->hwts_rx = 0;
|
|
break;
|
|
case HWTSTAMP_FILTER_PTP_V2_L2_EVENT:
|
|
case HWTSTAMP_FILTER_PTP_V2_L2_SYNC:
|
|
case HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ:
|
|
priv->hwts_rx = 1;
|
|
cfg->rx_filter = HWTSTAMP_FILTER_PTP_V2_L2_EVENT;
|
|
break;
|
|
default:
|
|
return -ERANGE;
|
|
}
|
|
|
|
if (priv->hwts_rx || priv->hwts_tx) {
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
data->regmap->vend1_event_msg_filt,
|
|
EVENT_MSG_FILT_ALL);
|
|
data->ptp_enable(phydev, true);
|
|
} else {
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
data->regmap->vend1_event_msg_filt,
|
|
EVENT_MSG_FILT_NONE);
|
|
data->ptp_enable(phydev, false);
|
|
}
|
|
|
|
if (nxp_c45_poll_txts(priv->phydev))
|
|
goto nxp_c45_no_ptp_irq;
|
|
|
|
if (priv->hwts_tx)
|
|
nxp_c45_set_reg_field(phydev, &data->regmap->irq_egr_ts_en);
|
|
else
|
|
nxp_c45_clear_reg_field(phydev, &data->regmap->irq_egr_ts_en);
|
|
|
|
nxp_c45_no_ptp_irq:
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_ts_info(struct mii_timestamper *mii_ts,
|
|
struct kernel_ethtool_ts_info *ts_info)
|
|
{
|
|
struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
|
|
mii_ts);
|
|
|
|
ts_info->so_timestamping = SOF_TIMESTAMPING_TX_HARDWARE |
|
|
SOF_TIMESTAMPING_RX_HARDWARE |
|
|
SOF_TIMESTAMPING_RAW_HARDWARE;
|
|
ts_info->phc_index = ptp_clock_index(priv->ptp_clock);
|
|
ts_info->tx_types = (1 << HWTSTAMP_TX_OFF) | (1 << HWTSTAMP_TX_ON);
|
|
ts_info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) |
|
|
(1 << HWTSTAMP_FILTER_PTP_V2_L2_SYNC) |
|
|
(1 << HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ) |
|
|
(1 << HWTSTAMP_FILTER_PTP_V2_L2_EVENT);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct nxp_c45_phy_stats common_hw_stats[] = {
|
|
{ "phy_link_status_drop_cnt",
|
|
NXP_C45_REG_FIELD(0x8352, MDIO_MMD_VEND1, 8, 6), },
|
|
{ "phy_link_availability_drop_cnt",
|
|
NXP_C45_REG_FIELD(0x8352, MDIO_MMD_VEND1, 0, 6), },
|
|
{ "phy_link_loss_cnt",
|
|
NXP_C45_REG_FIELD(0x8353, MDIO_MMD_VEND1, 10, 6), },
|
|
{ "phy_link_failure_cnt",
|
|
NXP_C45_REG_FIELD(0x8353, MDIO_MMD_VEND1, 0, 10), },
|
|
{ "phy_symbol_error_cnt",
|
|
NXP_C45_REG_FIELD(0x8350, MDIO_MMD_VEND1, 0, 16) },
|
|
};
|
|
|
|
static const struct nxp_c45_phy_stats tja1103_hw_stats[] = {
|
|
{ "rx_preamble_count",
|
|
NXP_C45_REG_FIELD(0xAFCE, MDIO_MMD_VEND1, 0, 6), },
|
|
{ "tx_preamble_count",
|
|
NXP_C45_REG_FIELD(0xAFCF, MDIO_MMD_VEND1, 0, 6), },
|
|
{ "rx_ipg_length",
|
|
NXP_C45_REG_FIELD(0xAFD0, MDIO_MMD_VEND1, 0, 9), },
|
|
{ "tx_ipg_length",
|
|
NXP_C45_REG_FIELD(0xAFD1, MDIO_MMD_VEND1, 0, 9), },
|
|
};
|
|
|
|
static const struct nxp_c45_phy_stats tja1120_hw_stats[] = {
|
|
{ "phy_symbol_error_cnt_ext",
|
|
NXP_C45_REG_FIELD(0x8351, MDIO_MMD_VEND1, 0, 14) },
|
|
{ "tx_frames_xtd",
|
|
NXP_C45_REG_FIELD(0xACA1, MDIO_MMD_VEND1, 0, 8), },
|
|
{ "tx_frames",
|
|
NXP_C45_REG_FIELD(0xACA0, MDIO_MMD_VEND1, 0, 16), },
|
|
{ "rx_frames_xtd",
|
|
NXP_C45_REG_FIELD(0xACA3, MDIO_MMD_VEND1, 0, 8), },
|
|
{ "rx_frames",
|
|
NXP_C45_REG_FIELD(0xACA2, MDIO_MMD_VEND1, 0, 16), },
|
|
{ "tx_lost_frames_xtd",
|
|
NXP_C45_REG_FIELD(0xACA5, MDIO_MMD_VEND1, 0, 8), },
|
|
{ "tx_lost_frames",
|
|
NXP_C45_REG_FIELD(0xACA4, MDIO_MMD_VEND1, 0, 16), },
|
|
{ "rx_lost_frames_xtd",
|
|
NXP_C45_REG_FIELD(0xACA7, MDIO_MMD_VEND1, 0, 8), },
|
|
{ "rx_lost_frames",
|
|
NXP_C45_REG_FIELD(0xACA6, MDIO_MMD_VEND1, 0, 16), },
|
|
};
|
|
|
|
static int nxp_c45_get_sset_count(struct phy_device *phydev)
|
|
{
|
|
const struct nxp_c45_phy_data *phy_data = nxp_c45_get_data(phydev);
|
|
|
|
return ARRAY_SIZE(common_hw_stats) + (phy_data ? phy_data->n_stats : 0);
|
|
}
|
|
|
|
static void nxp_c45_get_strings(struct phy_device *phydev, u8 *data)
|
|
{
|
|
const struct nxp_c45_phy_data *phy_data = nxp_c45_get_data(phydev);
|
|
size_t count = nxp_c45_get_sset_count(phydev);
|
|
size_t idx;
|
|
size_t i;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
if (i < ARRAY_SIZE(common_hw_stats)) {
|
|
strscpy(data + i * ETH_GSTRING_LEN,
|
|
common_hw_stats[i].name, ETH_GSTRING_LEN);
|
|
continue;
|
|
}
|
|
idx = i - ARRAY_SIZE(common_hw_stats);
|
|
strscpy(data + i * ETH_GSTRING_LEN,
|
|
phy_data->stats[idx].name, ETH_GSTRING_LEN);
|
|
}
|
|
}
|
|
|
|
static void nxp_c45_get_stats(struct phy_device *phydev,
|
|
struct ethtool_stats *stats, u64 *data)
|
|
{
|
|
const struct nxp_c45_phy_data *phy_data = nxp_c45_get_data(phydev);
|
|
size_t count = nxp_c45_get_sset_count(phydev);
|
|
const struct nxp_c45_reg_field *reg_field;
|
|
size_t idx;
|
|
size_t i;
|
|
int ret;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
if (i < ARRAY_SIZE(common_hw_stats)) {
|
|
reg_field = &common_hw_stats[i].counter;
|
|
} else {
|
|
idx = i - ARRAY_SIZE(common_hw_stats);
|
|
reg_field = &phy_data->stats[idx].counter;
|
|
}
|
|
|
|
ret = nxp_c45_read_reg_field(phydev, reg_field);
|
|
if (ret < 0)
|
|
data[i] = U64_MAX;
|
|
else
|
|
data[i] = ret;
|
|
}
|
|
}
|
|
|
|
static int nxp_c45_config_enable(struct phy_device *phydev)
|
|
{
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONTROL,
|
|
DEVICE_CONTROL_CONFIG_GLOBAL_EN |
|
|
DEVICE_CONTROL_CONFIG_ALL_EN);
|
|
usleep_range(400, 450);
|
|
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_CONTROL,
|
|
PORT_CONTROL_EN);
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONTROL,
|
|
PHY_CONFIG_EN);
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_INFRA_CONTROL,
|
|
PORT_INFRA_CONTROL_EN);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_start_op(struct phy_device *phydev)
|
|
{
|
|
return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONTROL,
|
|
PHY_START_OP);
|
|
}
|
|
|
|
static int nxp_c45_config_intr(struct phy_device *phydev)
|
|
{
|
|
int ret;
|
|
|
|
if (phydev->interrupts == PHY_INTERRUPT_ENABLED) {
|
|
ret = phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_FUNC_IRQ_EN, MACSEC_IRQS);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PHY_IRQ_EN, PHY_IRQ_LINK_EVENT);
|
|
}
|
|
|
|
ret = phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_FUNC_IRQ_EN, MACSEC_IRQS);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PHY_IRQ_EN, PHY_IRQ_LINK_EVENT);
|
|
}
|
|
|
|
static int tja1103_config_intr(struct phy_device *phydev)
|
|
{
|
|
int ret;
|
|
|
|
/* We can't disable the FUSA IRQ for TJA1103, but we can clean it up. */
|
|
ret = phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_ALWAYS_ACCESSIBLE,
|
|
FUSA_PASS);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return nxp_c45_config_intr(phydev);
|
|
}
|
|
|
|
static int tja1120_config_intr(struct phy_device *phydev)
|
|
{
|
|
int ret;
|
|
|
|
if (phydev->interrupts == PHY_INTERRUPT_ENABLED)
|
|
ret = phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_GLOBAL_INFRA_IRQ_EN,
|
|
TJA1120_DEV_BOOT_DONE);
|
|
else
|
|
ret = phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_GLOBAL_INFRA_IRQ_EN,
|
|
TJA1120_DEV_BOOT_DONE);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return nxp_c45_config_intr(phydev);
|
|
}
|
|
|
|
static irqreturn_t nxp_c45_handle_interrupt(struct phy_device *phydev)
|
|
{
|
|
const struct nxp_c45_phy_data *data = nxp_c45_get_data(phydev);
|
|
struct nxp_c45_phy *priv = phydev->priv;
|
|
irqreturn_t ret = IRQ_NONE;
|
|
struct nxp_c45_hwts hwts;
|
|
int irq;
|
|
|
|
irq = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_IRQ_STATUS);
|
|
if (irq & PHY_IRQ_LINK_EVENT) {
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_IRQ_ACK,
|
|
PHY_IRQ_LINK_EVENT);
|
|
phy_trigger_machine(phydev);
|
|
ret = IRQ_HANDLED;
|
|
}
|
|
|
|
irq = nxp_c45_read_reg_field(phydev, &data->regmap->irq_egr_ts_status);
|
|
if (irq) {
|
|
/* If ack_ptp_irq is false, the IRQ bit is self-clear and will
|
|
* be cleared when the EGR TS FIFO is empty. Otherwise, the
|
|
* IRQ bit should be cleared before reading the timestamp,
|
|
*/
|
|
if (data->ack_ptp_irq)
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PTP_IRQ_ACK, EGR_TS_IRQ);
|
|
while (data->get_egressts(priv, &hwts))
|
|
nxp_c45_process_txts(priv, &hwts);
|
|
|
|
ret = IRQ_HANDLED;
|
|
}
|
|
|
|
data->nmi_handler(phydev, &ret);
|
|
nxp_c45_handle_macsec_interrupt(phydev, &ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int nxp_c45_soft_reset(struct phy_device *phydev)
|
|
{
|
|
int ret;
|
|
|
|
ret = phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONTROL,
|
|
DEVICE_CONTROL_RESET);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return phy_read_mmd_poll_timeout(phydev, MDIO_MMD_VEND1,
|
|
VEND1_DEVICE_CONTROL, ret,
|
|
!(ret & DEVICE_CONTROL_RESET), 20000,
|
|
240000, false);
|
|
}
|
|
|
|
static int nxp_c45_cable_test_start(struct phy_device *phydev)
|
|
{
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(phydev);
|
|
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_FUNC_ENABLES, PHY_TEST_ENABLE);
|
|
return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, regmap->cable_test,
|
|
CABLE_TEST_ENABLE | CABLE_TEST_START);
|
|
}
|
|
|
|
static int nxp_c45_cable_test_get_status(struct phy_device *phydev,
|
|
bool *finished)
|
|
{
|
|
const struct nxp_c45_regmap *regmap = nxp_c45_get_regmap(phydev);
|
|
int ret;
|
|
u8 cable_test_result;
|
|
|
|
ret = nxp_c45_read_reg_field(phydev, ®map->cable_test_valid);
|
|
if (!ret) {
|
|
*finished = false;
|
|
return 0;
|
|
}
|
|
|
|
*finished = true;
|
|
cable_test_result = nxp_c45_read_reg_field(phydev,
|
|
®map->cable_test_result);
|
|
|
|
switch (cable_test_result) {
|
|
case CABLE_TEST_OK:
|
|
ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
|
|
ETHTOOL_A_CABLE_RESULT_CODE_OK);
|
|
break;
|
|
case CABLE_TEST_SHORTED:
|
|
ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
|
|
ETHTOOL_A_CABLE_RESULT_CODE_SAME_SHORT);
|
|
break;
|
|
case CABLE_TEST_OPEN:
|
|
ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
|
|
ETHTOOL_A_CABLE_RESULT_CODE_OPEN);
|
|
break;
|
|
default:
|
|
ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
|
|
ETHTOOL_A_CABLE_RESULT_CODE_UNSPEC);
|
|
}
|
|
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, regmap->cable_test,
|
|
CABLE_TEST_ENABLE);
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_FUNC_ENABLES, PHY_TEST_ENABLE);
|
|
|
|
return nxp_c45_start_op(phydev);
|
|
}
|
|
|
|
static int nxp_c45_get_sqi(struct phy_device *phydev)
|
|
{
|
|
int reg;
|
|
|
|
reg = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_SIGNAL_QUALITY);
|
|
if (!(reg & SQI_VALID))
|
|
return -EINVAL;
|
|
|
|
reg &= SQI_MASK;
|
|
|
|
return reg;
|
|
}
|
|
|
|
static void tja1120_link_change_notify(struct phy_device *phydev)
|
|
{
|
|
/* Bug workaround for TJA1120 enegineering samples: fix egress
|
|
* timestamps lost after link recovery.
|
|
*/
|
|
if (phydev->state == PHY_NOLINK) {
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_EPHY_RESETS, EPHY_PCS_RESET);
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_EPHY_RESETS, EPHY_PCS_RESET);
|
|
}
|
|
}
|
|
|
|
static int nxp_c45_get_sqi_max(struct phy_device *phydev)
|
|
{
|
|
return MAX_SQI;
|
|
}
|
|
|
|
static int nxp_c45_check_delay(struct phy_device *phydev, u32 delay)
|
|
{
|
|
if (delay < MIN_ID_PS) {
|
|
phydev_err(phydev, "delay value smaller than %u\n", MIN_ID_PS);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (delay > MAX_ID_PS) {
|
|
phydev_err(phydev, "delay value higher than %u\n", MAX_ID_PS);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void nxp_c45_counters_enable(struct phy_device *phydev)
|
|
{
|
|
const struct nxp_c45_phy_data *data = nxp_c45_get_data(phydev);
|
|
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_LINK_DROP_COUNTER,
|
|
COUNTER_EN);
|
|
|
|
data->counters_enable(phydev);
|
|
}
|
|
|
|
static void nxp_c45_ptp_init(struct phy_device *phydev)
|
|
{
|
|
const struct nxp_c45_phy_data *data = nxp_c45_get_data(phydev);
|
|
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
data->regmap->vend1_ptp_clk_period,
|
|
data->ptp_clk_period);
|
|
nxp_c45_clear_reg_field(phydev, &data->regmap->ltc_lock_ctrl);
|
|
|
|
data->ptp_init(phydev);
|
|
}
|
|
|
|
static u64 nxp_c45_get_phase_shift(u64 phase_offset_raw)
|
|
{
|
|
/* The delay in degree phase is 73.8 + phase_offset_raw * 0.9.
|
|
* To avoid floating point operations we'll multiply by 10
|
|
* and get 1 decimal point precision.
|
|
*/
|
|
phase_offset_raw *= 10;
|
|
phase_offset_raw -= 738;
|
|
return div_u64(phase_offset_raw, 9);
|
|
}
|
|
|
|
static void nxp_c45_disable_delays(struct phy_device *phydev)
|
|
{
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID, ID_ENABLE);
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID, ID_ENABLE);
|
|
}
|
|
|
|
static void nxp_c45_set_delays(struct phy_device *phydev)
|
|
{
|
|
struct nxp_c45_phy *priv = phydev->priv;
|
|
u64 tx_delay = priv->tx_delay;
|
|
u64 rx_delay = priv->rx_delay;
|
|
u64 degree;
|
|
|
|
if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
|
|
phydev->interface == PHY_INTERFACE_MODE_RGMII_TXID) {
|
|
degree = div_u64(tx_delay, PS_PER_DEGREE);
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID,
|
|
ID_ENABLE | nxp_c45_get_phase_shift(degree));
|
|
} else {
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID,
|
|
ID_ENABLE);
|
|
}
|
|
|
|
if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
|
|
phydev->interface == PHY_INTERFACE_MODE_RGMII_RXID) {
|
|
degree = div_u64(rx_delay, PS_PER_DEGREE);
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID,
|
|
ID_ENABLE | nxp_c45_get_phase_shift(degree));
|
|
} else {
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID,
|
|
ID_ENABLE);
|
|
}
|
|
}
|
|
|
|
static int nxp_c45_get_delays(struct phy_device *phydev)
|
|
{
|
|
struct nxp_c45_phy *priv = phydev->priv;
|
|
int ret;
|
|
|
|
if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
|
|
phydev->interface == PHY_INTERFACE_MODE_RGMII_TXID) {
|
|
ret = device_property_read_u32(&phydev->mdio.dev,
|
|
"tx-internal-delay-ps",
|
|
&priv->tx_delay);
|
|
if (ret)
|
|
priv->tx_delay = DEFAULT_ID_PS;
|
|
|
|
ret = nxp_c45_check_delay(phydev, priv->tx_delay);
|
|
if (ret) {
|
|
phydev_err(phydev,
|
|
"tx-internal-delay-ps invalid value\n");
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
|
|
phydev->interface == PHY_INTERFACE_MODE_RGMII_RXID) {
|
|
ret = device_property_read_u32(&phydev->mdio.dev,
|
|
"rx-internal-delay-ps",
|
|
&priv->rx_delay);
|
|
if (ret)
|
|
priv->rx_delay = DEFAULT_ID_PS;
|
|
|
|
ret = nxp_c45_check_delay(phydev, priv->rx_delay);
|
|
if (ret) {
|
|
phydev_err(phydev,
|
|
"rx-internal-delay-ps invalid value\n");
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_set_phy_mode(struct phy_device *phydev)
|
|
{
|
|
int ret;
|
|
|
|
ret = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_ABILITIES);
|
|
phydev_dbg(phydev, "Clause 45 managed PHY abilities 0x%x\n", ret);
|
|
|
|
switch (phydev->interface) {
|
|
case PHY_INTERFACE_MODE_RGMII:
|
|
if (!(ret & RGMII_ABILITY)) {
|
|
phydev_err(phydev, "rgmii mode not supported\n");
|
|
return -EINVAL;
|
|
}
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
|
|
MII_BASIC_CONFIG_RGMII);
|
|
nxp_c45_disable_delays(phydev);
|
|
break;
|
|
case PHY_INTERFACE_MODE_RGMII_ID:
|
|
case PHY_INTERFACE_MODE_RGMII_TXID:
|
|
case PHY_INTERFACE_MODE_RGMII_RXID:
|
|
if (!(ret & RGMII_ID_ABILITY)) {
|
|
phydev_err(phydev, "rgmii-id, rgmii-txid, rgmii-rxid modes are not supported\n");
|
|
return -EINVAL;
|
|
}
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
|
|
MII_BASIC_CONFIG_RGMII);
|
|
ret = nxp_c45_get_delays(phydev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
nxp_c45_set_delays(phydev);
|
|
break;
|
|
case PHY_INTERFACE_MODE_MII:
|
|
if (!(ret & MII_ABILITY)) {
|
|
phydev_err(phydev, "mii mode not supported\n");
|
|
return -EINVAL;
|
|
}
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
|
|
MII_BASIC_CONFIG_MII);
|
|
break;
|
|
case PHY_INTERFACE_MODE_REVMII:
|
|
if (!(ret & REVMII_ABILITY)) {
|
|
phydev_err(phydev, "rev-mii mode not supported\n");
|
|
return -EINVAL;
|
|
}
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
|
|
MII_BASIC_CONFIG_MII | MII_BASIC_CONFIG_REV);
|
|
break;
|
|
case PHY_INTERFACE_MODE_RMII:
|
|
if (!(ret & RMII_ABILITY)) {
|
|
phydev_err(phydev, "rmii mode not supported\n");
|
|
return -EINVAL;
|
|
}
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
|
|
MII_BASIC_CONFIG_RMII);
|
|
break;
|
|
case PHY_INTERFACE_MODE_SGMII:
|
|
if (!(ret & SGMII_ABILITY)) {
|
|
phydev_err(phydev, "sgmii mode not supported\n");
|
|
return -EINVAL;
|
|
}
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
|
|
MII_BASIC_CONFIG_SGMII);
|
|
break;
|
|
case PHY_INTERFACE_MODE_INTERNAL:
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nxp_c45_config_init(struct phy_device *phydev)
|
|
{
|
|
int ret;
|
|
|
|
ret = nxp_c45_config_enable(phydev);
|
|
if (ret) {
|
|
phydev_err(phydev, "Failed to enable config\n");
|
|
return ret;
|
|
}
|
|
|
|
/* Bug workaround for SJA1110 rev B: enable write access
|
|
* to MDIO_MMD_PMAPMD
|
|
*/
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, 0x01F8, 1);
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, 0x01F9, 2);
|
|
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONFIG,
|
|
PHY_CONFIG_AUTO);
|
|
|
|
ret = nxp_c45_set_phy_mode(phydev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
phydev->autoneg = AUTONEG_DISABLE;
|
|
|
|
nxp_c45_counters_enable(phydev);
|
|
nxp_c45_ptp_init(phydev);
|
|
ret = nxp_c45_macsec_config_init(phydev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return nxp_c45_start_op(phydev);
|
|
}
|
|
|
|
static int nxp_c45_get_features(struct phy_device *phydev)
|
|
{
|
|
linkmode_set_bit(ETHTOOL_LINK_MODE_TP_BIT, phydev->supported);
|
|
linkmode_set_bit(ETHTOOL_LINK_MODE_MII_BIT, phydev->supported);
|
|
|
|
return genphy_c45_pma_read_abilities(phydev);
|
|
}
|
|
|
|
static int nxp_c45_probe(struct phy_device *phydev)
|
|
{
|
|
struct nxp_c45_phy *priv;
|
|
bool macsec_ability;
|
|
int phy_abilities;
|
|
bool ptp_ability;
|
|
int ret = 0;
|
|
|
|
priv = devm_kzalloc(&phydev->mdio.dev, sizeof(*priv), GFP_KERNEL);
|
|
if (!priv)
|
|
return -ENOMEM;
|
|
|
|
skb_queue_head_init(&priv->tx_queue);
|
|
skb_queue_head_init(&priv->rx_queue);
|
|
|
|
priv->phydev = phydev;
|
|
|
|
phydev->priv = priv;
|
|
|
|
mutex_init(&priv->ptp_lock);
|
|
|
|
phy_abilities = phy_read_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_ABILITIES);
|
|
ptp_ability = !!(phy_abilities & PTP_ABILITY);
|
|
if (!ptp_ability) {
|
|
phydev_dbg(phydev, "the phy does not support PTP");
|
|
goto no_ptp_support;
|
|
}
|
|
|
|
if (IS_ENABLED(CONFIG_PTP_1588_CLOCK) &&
|
|
IS_ENABLED(CONFIG_NETWORK_PHY_TIMESTAMPING)) {
|
|
priv->mii_ts.rxtstamp = nxp_c45_rxtstamp;
|
|
priv->mii_ts.txtstamp = nxp_c45_txtstamp;
|
|
priv->mii_ts.hwtstamp = nxp_c45_hwtstamp;
|
|
priv->mii_ts.ts_info = nxp_c45_ts_info;
|
|
phydev->mii_ts = &priv->mii_ts;
|
|
ret = nxp_c45_init_ptp_clock(priv);
|
|
|
|
/* Timestamp selected by default to keep legacy API */
|
|
phydev->default_timestamp = true;
|
|
} else {
|
|
phydev_dbg(phydev, "PTP support not enabled even if the phy supports it");
|
|
}
|
|
|
|
no_ptp_support:
|
|
macsec_ability = !!(phy_abilities & MACSEC_ABILITY);
|
|
if (!macsec_ability) {
|
|
phydev_info(phydev, "the phy does not support MACsec\n");
|
|
goto no_macsec_support;
|
|
}
|
|
|
|
if (IS_ENABLED(CONFIG_MACSEC)) {
|
|
ret = nxp_c45_macsec_probe(phydev);
|
|
phydev_dbg(phydev, "MACsec support enabled.");
|
|
} else {
|
|
phydev_dbg(phydev, "MACsec support not enabled even if the phy supports it");
|
|
}
|
|
|
|
no_macsec_support:
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void nxp_c45_remove(struct phy_device *phydev)
|
|
{
|
|
struct nxp_c45_phy *priv = phydev->priv;
|
|
|
|
if (priv->ptp_clock)
|
|
ptp_clock_unregister(priv->ptp_clock);
|
|
|
|
skb_queue_purge(&priv->tx_queue);
|
|
skb_queue_purge(&priv->rx_queue);
|
|
nxp_c45_macsec_remove(phydev);
|
|
}
|
|
|
|
static void tja1103_counters_enable(struct phy_device *phydev)
|
|
{
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_PREAMBLE_COUNT,
|
|
COUNTER_EN);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TX_PREAMBLE_COUNT,
|
|
COUNTER_EN);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_IPG_LENGTH,
|
|
COUNTER_EN);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TX_IPG_LENGTH,
|
|
COUNTER_EN);
|
|
}
|
|
|
|
static void tja1103_ptp_init(struct phy_device *phydev)
|
|
{
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_TS_INSRT_CTRL,
|
|
TJA1103_RX_TS_INSRT_MODE2);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_FUNC_ENABLES,
|
|
PTP_ENABLE);
|
|
}
|
|
|
|
static void tja1103_ptp_enable(struct phy_device *phydev, bool enable)
|
|
{
|
|
if (enable)
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_PTP_CONTROL,
|
|
PORT_PTP_CONTROL_BYPASS);
|
|
else
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_PTP_CONTROL,
|
|
PORT_PTP_CONTROL_BYPASS);
|
|
}
|
|
|
|
static void tja1103_nmi_handler(struct phy_device *phydev,
|
|
irqreturn_t *irq_status)
|
|
{
|
|
int ret;
|
|
|
|
ret = phy_read_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_ALWAYS_ACCESSIBLE);
|
|
if (ret & FUSA_PASS) {
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_ALWAYS_ACCESSIBLE,
|
|
FUSA_PASS);
|
|
*irq_status = IRQ_HANDLED;
|
|
}
|
|
}
|
|
|
|
static const struct nxp_c45_regmap tja1103_regmap = {
|
|
.vend1_ptp_clk_period = 0x1104,
|
|
.vend1_event_msg_filt = 0x1148,
|
|
.pps_enable =
|
|
NXP_C45_REG_FIELD(0x1102, MDIO_MMD_VEND1, 3, 1),
|
|
.pps_polarity =
|
|
NXP_C45_REG_FIELD(0x1102, MDIO_MMD_VEND1, 2, 1),
|
|
.ltc_lock_ctrl =
|
|
NXP_C45_REG_FIELD(0x1115, MDIO_MMD_VEND1, 0, 1),
|
|
.ltc_read =
|
|
NXP_C45_REG_FIELD(0x1105, MDIO_MMD_VEND1, 2, 1),
|
|
.ltc_write =
|
|
NXP_C45_REG_FIELD(0x1105, MDIO_MMD_VEND1, 0, 1),
|
|
.vend1_ltc_wr_nsec_0 = 0x1106,
|
|
.vend1_ltc_wr_nsec_1 = 0x1107,
|
|
.vend1_ltc_wr_sec_0 = 0x1108,
|
|
.vend1_ltc_wr_sec_1 = 0x1109,
|
|
.vend1_ltc_rd_nsec_0 = 0x110A,
|
|
.vend1_ltc_rd_nsec_1 = 0x110B,
|
|
.vend1_ltc_rd_sec_0 = 0x110C,
|
|
.vend1_ltc_rd_sec_1 = 0x110D,
|
|
.vend1_rate_adj_subns_0 = 0x110F,
|
|
.vend1_rate_adj_subns_1 = 0x1110,
|
|
.irq_egr_ts_en =
|
|
NXP_C45_REG_FIELD(0x1131, MDIO_MMD_VEND1, 0, 1),
|
|
.irq_egr_ts_status =
|
|
NXP_C45_REG_FIELD(0x1132, MDIO_MMD_VEND1, 0, 1),
|
|
.domain_number =
|
|
NXP_C45_REG_FIELD(0x114E, MDIO_MMD_VEND1, 0, 8),
|
|
.msg_type =
|
|
NXP_C45_REG_FIELD(0x114E, MDIO_MMD_VEND1, 8, 4),
|
|
.sequence_id =
|
|
NXP_C45_REG_FIELD(0x114F, MDIO_MMD_VEND1, 0, 16),
|
|
.sec_1_0 =
|
|
NXP_C45_REG_FIELD(0x1151, MDIO_MMD_VEND1, 14, 2),
|
|
.sec_4_2 =
|
|
NXP_C45_REG_FIELD(0x114E, MDIO_MMD_VEND1, 12, 3),
|
|
.nsec_15_0 =
|
|
NXP_C45_REG_FIELD(0x1150, MDIO_MMD_VEND1, 0, 16),
|
|
.nsec_29_16 =
|
|
NXP_C45_REG_FIELD(0x1151, MDIO_MMD_VEND1, 0, 14),
|
|
.vend1_ext_trg_data_0 = 0x1121,
|
|
.vend1_ext_trg_data_1 = 0x1122,
|
|
.vend1_ext_trg_data_2 = 0x1123,
|
|
.vend1_ext_trg_data_3 = 0x1124,
|
|
.vend1_ext_trg_ctrl = 0x1126,
|
|
.cable_test = 0x8330,
|
|
.cable_test_valid =
|
|
NXP_C45_REG_FIELD(0x8330, MDIO_MMD_VEND1, 13, 1),
|
|
.cable_test_result =
|
|
NXP_C45_REG_FIELD(0x8330, MDIO_MMD_VEND1, 0, 3),
|
|
};
|
|
|
|
static const struct nxp_c45_phy_data tja1103_phy_data = {
|
|
.regmap = &tja1103_regmap,
|
|
.stats = tja1103_hw_stats,
|
|
.n_stats = ARRAY_SIZE(tja1103_hw_stats),
|
|
.ptp_clk_period = PTP_CLK_PERIOD_100BT1,
|
|
.ext_ts_both_edges = false,
|
|
.ack_ptp_irq = false,
|
|
.counters_enable = tja1103_counters_enable,
|
|
.get_egressts = nxp_c45_get_hwtxts,
|
|
.get_extts = nxp_c45_get_extts,
|
|
.ptp_init = tja1103_ptp_init,
|
|
.ptp_enable = tja1103_ptp_enable,
|
|
.nmi_handler = tja1103_nmi_handler,
|
|
};
|
|
|
|
static void tja1120_counters_enable(struct phy_device *phydev)
|
|
{
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_SYMBOL_ERROR_CNT_XTD,
|
|
EXTENDED_CNT_EN);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_MONITOR_STATUS,
|
|
MONITOR_RESET);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_MONITOR_CONFIG,
|
|
ALL_FRAMES_CNT_EN | LOST_FRAMES_CNT_EN);
|
|
}
|
|
|
|
static void tja1120_ptp_init(struct phy_device *phydev)
|
|
{
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, TJA1120_RX_TS_INSRT_CTRL,
|
|
TJA1120_RX_TS_INSRT_EN | TJA1120_TS_INSRT_MODE);
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1, TJA1120_VEND1_EXT_TS_MODE,
|
|
TJA1120_TS_INSRT_MODE);
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONFIG,
|
|
PTP_ENABLE);
|
|
}
|
|
|
|
static void tja1120_ptp_enable(struct phy_device *phydev, bool enable)
|
|
{
|
|
if (enable)
|
|
phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_FUNC_ENABLES,
|
|
PTP_ENABLE);
|
|
else
|
|
phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
|
|
VEND1_PORT_FUNC_ENABLES,
|
|
PTP_ENABLE);
|
|
}
|
|
|
|
static void tja1120_nmi_handler(struct phy_device *phydev,
|
|
irqreturn_t *irq_status)
|
|
{
|
|
int ret;
|
|
|
|
ret = phy_read_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_GLOBAL_INFRA_IRQ_STATUS);
|
|
if (ret & TJA1120_DEV_BOOT_DONE) {
|
|
phy_write_mmd(phydev, MDIO_MMD_VEND1,
|
|
TJA1120_GLOBAL_INFRA_IRQ_ACK,
|
|
TJA1120_DEV_BOOT_DONE);
|
|
*irq_status = IRQ_HANDLED;
|
|
}
|
|
}
|
|
|
|
static const struct nxp_c45_regmap tja1120_regmap = {
|
|
.vend1_ptp_clk_period = 0x1020,
|
|
.vend1_event_msg_filt = 0x9010,
|
|
.pps_enable =
|
|
NXP_C45_REG_FIELD(0x1006, MDIO_MMD_VEND1, 4, 1),
|
|
.pps_polarity =
|
|
NXP_C45_REG_FIELD(0x1006, MDIO_MMD_VEND1, 5, 1),
|
|
.ltc_lock_ctrl =
|
|
NXP_C45_REG_FIELD(0x1006, MDIO_MMD_VEND1, 2, 1),
|
|
.ltc_read =
|
|
NXP_C45_REG_FIELD(0x1000, MDIO_MMD_VEND1, 1, 1),
|
|
.ltc_write =
|
|
NXP_C45_REG_FIELD(0x1000, MDIO_MMD_VEND1, 2, 1),
|
|
.vend1_ltc_wr_nsec_0 = 0x1040,
|
|
.vend1_ltc_wr_nsec_1 = 0x1041,
|
|
.vend1_ltc_wr_sec_0 = 0x1042,
|
|
.vend1_ltc_wr_sec_1 = 0x1043,
|
|
.vend1_ltc_rd_nsec_0 = 0x1048,
|
|
.vend1_ltc_rd_nsec_1 = 0x1049,
|
|
.vend1_ltc_rd_sec_0 = 0x104A,
|
|
.vend1_ltc_rd_sec_1 = 0x104B,
|
|
.vend1_rate_adj_subns_0 = 0x1030,
|
|
.vend1_rate_adj_subns_1 = 0x1031,
|
|
.irq_egr_ts_en =
|
|
NXP_C45_REG_FIELD(0x900A, MDIO_MMD_VEND1, 1, 1),
|
|
.irq_egr_ts_status =
|
|
NXP_C45_REG_FIELD(0x900C, MDIO_MMD_VEND1, 1, 1),
|
|
.domain_number =
|
|
NXP_C45_REG_FIELD(0x9061, MDIO_MMD_VEND1, 8, 8),
|
|
.msg_type =
|
|
NXP_C45_REG_FIELD(0x9061, MDIO_MMD_VEND1, 4, 4),
|
|
.sequence_id =
|
|
NXP_C45_REG_FIELD(0x9062, MDIO_MMD_VEND1, 0, 16),
|
|
.sec_1_0 =
|
|
NXP_C45_REG_FIELD(0x9065, MDIO_MMD_VEND1, 0, 2),
|
|
.sec_4_2 =
|
|
NXP_C45_REG_FIELD(0x9065, MDIO_MMD_VEND1, 2, 3),
|
|
.nsec_15_0 =
|
|
NXP_C45_REG_FIELD(0x9063, MDIO_MMD_VEND1, 0, 16),
|
|
.nsec_29_16 =
|
|
NXP_C45_REG_FIELD(0x9064, MDIO_MMD_VEND1, 0, 14),
|
|
.vend1_ext_trg_data_0 = 0x1071,
|
|
.vend1_ext_trg_data_1 = 0x1072,
|
|
.vend1_ext_trg_data_2 = 0x1073,
|
|
.vend1_ext_trg_data_3 = 0x1074,
|
|
.vend1_ext_trg_ctrl = 0x1075,
|
|
.cable_test = 0x8360,
|
|
.cable_test_valid =
|
|
NXP_C45_REG_FIELD(0x8361, MDIO_MMD_VEND1, 15, 1),
|
|
.cable_test_result =
|
|
NXP_C45_REG_FIELD(0x8361, MDIO_MMD_VEND1, 0, 3),
|
|
};
|
|
|
|
static const struct nxp_c45_phy_data tja1120_phy_data = {
|
|
.regmap = &tja1120_regmap,
|
|
.stats = tja1120_hw_stats,
|
|
.n_stats = ARRAY_SIZE(tja1120_hw_stats),
|
|
.ptp_clk_period = PTP_CLK_PERIOD_1000BT1,
|
|
.ext_ts_both_edges = true,
|
|
.ack_ptp_irq = true,
|
|
.counters_enable = tja1120_counters_enable,
|
|
.get_egressts = tja1120_get_hwtxts,
|
|
.get_extts = tja1120_get_extts,
|
|
.ptp_init = tja1120_ptp_init,
|
|
.ptp_enable = tja1120_ptp_enable,
|
|
.nmi_handler = tja1120_nmi_handler,
|
|
};
|
|
|
|
static struct phy_driver nxp_c45_driver[] = {
|
|
{
|
|
PHY_ID_MATCH_MODEL(PHY_ID_TJA_1103),
|
|
.name = "NXP C45 TJA1103",
|
|
.get_features = nxp_c45_get_features,
|
|
.driver_data = &tja1103_phy_data,
|
|
.probe = nxp_c45_probe,
|
|
.soft_reset = nxp_c45_soft_reset,
|
|
.config_aneg = genphy_c45_config_aneg,
|
|
.config_init = nxp_c45_config_init,
|
|
.config_intr = tja1103_config_intr,
|
|
.handle_interrupt = nxp_c45_handle_interrupt,
|
|
.read_status = genphy_c45_read_status,
|
|
.suspend = genphy_c45_pma_suspend,
|
|
.resume = genphy_c45_pma_resume,
|
|
.get_sset_count = nxp_c45_get_sset_count,
|
|
.get_strings = nxp_c45_get_strings,
|
|
.get_stats = nxp_c45_get_stats,
|
|
.cable_test_start = nxp_c45_cable_test_start,
|
|
.cable_test_get_status = nxp_c45_cable_test_get_status,
|
|
.set_loopback = genphy_c45_loopback,
|
|
.get_sqi = nxp_c45_get_sqi,
|
|
.get_sqi_max = nxp_c45_get_sqi_max,
|
|
.remove = nxp_c45_remove,
|
|
},
|
|
{
|
|
PHY_ID_MATCH_MODEL(PHY_ID_TJA_1120),
|
|
.name = "NXP C45 TJA1120",
|
|
.get_features = nxp_c45_get_features,
|
|
.driver_data = &tja1120_phy_data,
|
|
.probe = nxp_c45_probe,
|
|
.soft_reset = nxp_c45_soft_reset,
|
|
.config_aneg = genphy_c45_config_aneg,
|
|
.config_init = nxp_c45_config_init,
|
|
.config_intr = tja1120_config_intr,
|
|
.handle_interrupt = nxp_c45_handle_interrupt,
|
|
.read_status = genphy_c45_read_status,
|
|
.link_change_notify = tja1120_link_change_notify,
|
|
.suspend = genphy_c45_pma_suspend,
|
|
.resume = genphy_c45_pma_resume,
|
|
.get_sset_count = nxp_c45_get_sset_count,
|
|
.get_strings = nxp_c45_get_strings,
|
|
.get_stats = nxp_c45_get_stats,
|
|
.cable_test_start = nxp_c45_cable_test_start,
|
|
.cable_test_get_status = nxp_c45_cable_test_get_status,
|
|
.set_loopback = genphy_c45_loopback,
|
|
.get_sqi = nxp_c45_get_sqi,
|
|
.get_sqi_max = nxp_c45_get_sqi_max,
|
|
.remove = nxp_c45_remove,
|
|
},
|
|
};
|
|
|
|
module_phy_driver(nxp_c45_driver);
|
|
|
|
static struct mdio_device_id __maybe_unused nxp_c45_tbl[] = {
|
|
{ PHY_ID_MATCH_MODEL(PHY_ID_TJA_1103) },
|
|
{ PHY_ID_MATCH_MODEL(PHY_ID_TJA_1120) },
|
|
{ /*sentinel*/ },
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(mdio, nxp_c45_tbl);
|
|
|
|
MODULE_AUTHOR("Radu Pirea <radu-nicolae.pirea@oss.nxp.com>");
|
|
MODULE_DESCRIPTION("NXP C45 PHY driver");
|
|
MODULE_LICENSE("GPL v2");
|