forked from Minki/linux
873e65bc09
Based on 1 normalized pattern(s): this program is free software you can redistribute it and or modify it under the terms of the gnu general public license as published by the free software foundation version 2 of the license this program is distributed in the hope that it will be useful but without any warranty without even the implied warranty of merchantability or fitness for a particular purpose see the gnu general public license for more details you should have received a copy of the gnu general public license along with this program if not write to the free software foundation inc 59 temple place suite 330 boston ma 02111 1307 usa extracted by the scancode license scanner the SPDX license identifier GPL-2.0-only has been chosen to replace the boilerplate/reference in 83 file(s). Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Richard Fontana <rfontana@redhat.com> Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org> Reviewed-by: Allison Randal <allison@lohutok.net> Cc: linux-spdx@vger.kernel.org Link: https://lkml.kernel.org/r/20190527070034.021731668@linutronix.de Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
815 lines
19 KiB
C
815 lines
19 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Marvell 88SE64xx hardware specific
|
|
*
|
|
* Copyright 2007 Red Hat, Inc.
|
|
* Copyright 2008 Marvell. <kewei@marvell.com>
|
|
* Copyright 2009-2011 Marvell. <yuxiangl@marvell.com>
|
|
*/
|
|
|
|
#include "mv_sas.h"
|
|
#include "mv_64xx.h"
|
|
#include "mv_chips.h"
|
|
|
|
static void mvs_64xx_detect_porttype(struct mvs_info *mvi, int i)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 reg;
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
|
|
reg = mr32(MVS_GBL_PORT_TYPE);
|
|
phy->phy_type &= ~(PORT_TYPE_SAS | PORT_TYPE_SATA);
|
|
if (reg & MODE_SAS_SATA & (1 << i))
|
|
phy->phy_type |= PORT_TYPE_SAS;
|
|
else
|
|
phy->phy_type |= PORT_TYPE_SATA;
|
|
}
|
|
|
|
static void mvs_64xx_enable_xmt(struct mvs_info *mvi, int phy_id)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
tmp = mr32(MVS_PCS);
|
|
if (mvi->chip->n_phy <= MVS_SOC_PORTS)
|
|
tmp |= 1 << (phy_id + PCS_EN_PORT_XMT_SHIFT);
|
|
else
|
|
tmp |= 1 << (phy_id + PCS_EN_PORT_XMT_SHIFT2);
|
|
mw32(MVS_PCS, tmp);
|
|
}
|
|
|
|
static void mvs_64xx_phy_hacks(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
int i;
|
|
|
|
mvs_phy_hacks(mvi);
|
|
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
for (i = 0; i < MVS_SOC_PORTS; i++) {
|
|
mvs_write_port_vsr_addr(mvi, i, VSR_PHY_MODE8);
|
|
mvs_write_port_vsr_data(mvi, i, 0x2F0);
|
|
}
|
|
} else {
|
|
/* disable auto port detection */
|
|
mw32(MVS_GBL_PORT_TYPE, 0);
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
mvs_write_port_vsr_addr(mvi, i, VSR_PHY_MODE7);
|
|
mvs_write_port_vsr_data(mvi, i, 0x90000000);
|
|
mvs_write_port_vsr_addr(mvi, i, VSR_PHY_MODE9);
|
|
mvs_write_port_vsr_data(mvi, i, 0x50f2);
|
|
mvs_write_port_vsr_addr(mvi, i, VSR_PHY_MODE11);
|
|
mvs_write_port_vsr_data(mvi, i, 0x0e);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void mvs_64xx_stp_reset(struct mvs_info *mvi, u32 phy_id)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 reg, tmp;
|
|
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
if (phy_id < MVS_SOC_PORTS)
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL, ®);
|
|
else
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL2, ®);
|
|
|
|
} else
|
|
reg = mr32(MVS_PHY_CTL);
|
|
|
|
tmp = reg;
|
|
if (phy_id < MVS_SOC_PORTS)
|
|
tmp |= (1U << phy_id) << PCTL_LINK_OFFS;
|
|
else
|
|
tmp |= (1U << (phy_id - MVS_SOC_PORTS)) << PCTL_LINK_OFFS;
|
|
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
if (phy_id < MVS_SOC_PORTS) {
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL, tmp);
|
|
mdelay(10);
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL, reg);
|
|
} else {
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL2, tmp);
|
|
mdelay(10);
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL2, reg);
|
|
}
|
|
} else {
|
|
mw32(MVS_PHY_CTL, tmp);
|
|
mdelay(10);
|
|
mw32(MVS_PHY_CTL, reg);
|
|
}
|
|
}
|
|
|
|
static void mvs_64xx_phy_reset(struct mvs_info *mvi, u32 phy_id, int hard)
|
|
{
|
|
u32 tmp;
|
|
tmp = mvs_read_port_irq_stat(mvi, phy_id);
|
|
tmp &= ~PHYEV_RDY_CH;
|
|
mvs_write_port_irq_stat(mvi, phy_id, tmp);
|
|
tmp = mvs_read_phy_ctl(mvi, phy_id);
|
|
if (hard == MVS_HARD_RESET)
|
|
tmp |= PHY_RST_HARD;
|
|
else if (hard == MVS_SOFT_RESET)
|
|
tmp |= PHY_RST;
|
|
mvs_write_phy_ctl(mvi, phy_id, tmp);
|
|
if (hard) {
|
|
do {
|
|
tmp = mvs_read_phy_ctl(mvi, phy_id);
|
|
} while (tmp & PHY_RST_HARD);
|
|
}
|
|
}
|
|
|
|
static void
|
|
mvs_64xx_clear_srs_irq(struct mvs_info *mvi, u8 reg_set, u8 clear_all)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
if (clear_all) {
|
|
tmp = mr32(MVS_INT_STAT_SRS_0);
|
|
if (tmp) {
|
|
printk(KERN_DEBUG "check SRS 0 %08X.\n", tmp);
|
|
mw32(MVS_INT_STAT_SRS_0, tmp);
|
|
}
|
|
} else {
|
|
tmp = mr32(MVS_INT_STAT_SRS_0);
|
|
if (tmp & (1 << (reg_set % 32))) {
|
|
printk(KERN_DEBUG "register set 0x%x was stopped.\n",
|
|
reg_set);
|
|
mw32(MVS_INT_STAT_SRS_0, 1 << (reg_set % 32));
|
|
}
|
|
}
|
|
}
|
|
|
|
static int mvs_64xx_chip_reset(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
int i;
|
|
|
|
/* make sure interrupts are masked immediately (paranoia) */
|
|
mw32(MVS_GBL_CTL, 0);
|
|
tmp = mr32(MVS_GBL_CTL);
|
|
|
|
/* Reset Controller */
|
|
if (!(tmp & HBA_RST)) {
|
|
if (mvi->flags & MVF_PHY_PWR_FIX) {
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL, &tmp);
|
|
tmp &= ~PCTL_PWR_OFF;
|
|
tmp |= PCTL_PHY_DSBL;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL, tmp);
|
|
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL2, &tmp);
|
|
tmp &= ~PCTL_PWR_OFF;
|
|
tmp |= PCTL_PHY_DSBL;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL2, tmp);
|
|
}
|
|
}
|
|
|
|
/* make sure interrupts are masked immediately (paranoia) */
|
|
mw32(MVS_GBL_CTL, 0);
|
|
tmp = mr32(MVS_GBL_CTL);
|
|
|
|
/* Reset Controller */
|
|
if (!(tmp & HBA_RST)) {
|
|
/* global reset, incl. COMRESET/H_RESET_N (self-clearing) */
|
|
mw32_f(MVS_GBL_CTL, HBA_RST);
|
|
}
|
|
|
|
/* wait for reset to finish; timeout is just a guess */
|
|
i = 1000;
|
|
while (i-- > 0) {
|
|
msleep(10);
|
|
|
|
if (!(mr32(MVS_GBL_CTL) & HBA_RST))
|
|
break;
|
|
}
|
|
if (mr32(MVS_GBL_CTL) & HBA_RST) {
|
|
dev_printk(KERN_ERR, mvi->dev, "HBA reset failed\n");
|
|
return -EBUSY;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void mvs_64xx_phy_disable(struct mvs_info *mvi, u32 phy_id)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
u32 offs;
|
|
if (phy_id < 4)
|
|
offs = PCR_PHY_CTL;
|
|
else {
|
|
offs = PCR_PHY_CTL2;
|
|
phy_id -= 4;
|
|
}
|
|
pci_read_config_dword(mvi->pdev, offs, &tmp);
|
|
tmp |= 1U << (PCTL_PHY_DSBL_OFFS + phy_id);
|
|
pci_write_config_dword(mvi->pdev, offs, tmp);
|
|
} else {
|
|
tmp = mr32(MVS_PHY_CTL);
|
|
tmp |= 1U << (PCTL_PHY_DSBL_OFFS + phy_id);
|
|
mw32(MVS_PHY_CTL, tmp);
|
|
}
|
|
}
|
|
|
|
static void mvs_64xx_phy_enable(struct mvs_info *mvi, u32 phy_id)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
u32 offs;
|
|
if (phy_id < 4)
|
|
offs = PCR_PHY_CTL;
|
|
else {
|
|
offs = PCR_PHY_CTL2;
|
|
phy_id -= 4;
|
|
}
|
|
pci_read_config_dword(mvi->pdev, offs, &tmp);
|
|
tmp &= ~(1U << (PCTL_PHY_DSBL_OFFS + phy_id));
|
|
pci_write_config_dword(mvi->pdev, offs, tmp);
|
|
} else {
|
|
tmp = mr32(MVS_PHY_CTL);
|
|
tmp &= ~(1U << (PCTL_PHY_DSBL_OFFS + phy_id));
|
|
mw32(MVS_PHY_CTL, tmp);
|
|
}
|
|
}
|
|
|
|
static int mvs_64xx_init(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
int i;
|
|
u32 tmp, cctl;
|
|
|
|
if (mvi->pdev && mvi->pdev->revision == 0)
|
|
mvi->flags |= MVF_PHY_PWR_FIX;
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
mvs_show_pcie_usage(mvi);
|
|
tmp = mvs_64xx_chip_reset(mvi);
|
|
if (tmp)
|
|
return tmp;
|
|
} else {
|
|
tmp = mr32(MVS_PHY_CTL);
|
|
tmp &= ~PCTL_PWR_OFF;
|
|
tmp |= PCTL_PHY_DSBL;
|
|
mw32(MVS_PHY_CTL, tmp);
|
|
}
|
|
|
|
/* Init Chip */
|
|
/* make sure RST is set; HBA_RST /should/ have done that for us */
|
|
cctl = mr32(MVS_CTL) & 0xFFFF;
|
|
if (cctl & CCTL_RST)
|
|
cctl &= ~CCTL_RST;
|
|
else
|
|
mw32_f(MVS_CTL, cctl | CCTL_RST);
|
|
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
/* write to device control _AND_ device status register */
|
|
pci_read_config_dword(mvi->pdev, PCR_DEV_CTRL, &tmp);
|
|
tmp &= ~PRD_REQ_MASK;
|
|
tmp |= PRD_REQ_SIZE;
|
|
pci_write_config_dword(mvi->pdev, PCR_DEV_CTRL, tmp);
|
|
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL, &tmp);
|
|
tmp &= ~PCTL_PWR_OFF;
|
|
tmp &= ~PCTL_PHY_DSBL;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL, tmp);
|
|
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL2, &tmp);
|
|
tmp &= PCTL_PWR_OFF;
|
|
tmp &= ~PCTL_PHY_DSBL;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL2, tmp);
|
|
} else {
|
|
tmp = mr32(MVS_PHY_CTL);
|
|
tmp &= ~PCTL_PWR_OFF;
|
|
tmp |= PCTL_COM_ON;
|
|
tmp &= ~PCTL_PHY_DSBL;
|
|
tmp |= PCTL_LINK_RST;
|
|
mw32(MVS_PHY_CTL, tmp);
|
|
msleep(100);
|
|
tmp &= ~PCTL_LINK_RST;
|
|
mw32(MVS_PHY_CTL, tmp);
|
|
msleep(100);
|
|
}
|
|
|
|
/* reset control */
|
|
mw32(MVS_PCS, 0); /* MVS_PCS */
|
|
/* init phys */
|
|
mvs_64xx_phy_hacks(mvi);
|
|
|
|
tmp = mvs_cr32(mvi, CMD_PHY_MODE_21);
|
|
tmp &= 0x0000ffff;
|
|
tmp |= 0x00fa0000;
|
|
mvs_cw32(mvi, CMD_PHY_MODE_21, tmp);
|
|
|
|
/* enable auto port detection */
|
|
mw32(MVS_GBL_PORT_TYPE, MODE_AUTO_DET_EN);
|
|
|
|
mw32(MVS_CMD_LIST_LO, mvi->slot_dma);
|
|
mw32(MVS_CMD_LIST_HI, (mvi->slot_dma >> 16) >> 16);
|
|
|
|
mw32(MVS_RX_FIS_LO, mvi->rx_fis_dma);
|
|
mw32(MVS_RX_FIS_HI, (mvi->rx_fis_dma >> 16) >> 16);
|
|
|
|
mw32(MVS_TX_CFG, MVS_CHIP_SLOT_SZ);
|
|
mw32(MVS_TX_LO, mvi->tx_dma);
|
|
mw32(MVS_TX_HI, (mvi->tx_dma >> 16) >> 16);
|
|
|
|
mw32(MVS_RX_CFG, MVS_RX_RING_SZ);
|
|
mw32(MVS_RX_LO, mvi->rx_dma);
|
|
mw32(MVS_RX_HI, (mvi->rx_dma >> 16) >> 16);
|
|
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
/* set phy local SAS address */
|
|
/* should set little endian SAS address to 64xx chip */
|
|
mvs_set_sas_addr(mvi, i, PHYR_ADDR_LO, PHYR_ADDR_HI,
|
|
cpu_to_be64(mvi->phy[i].dev_sas_addr));
|
|
|
|
mvs_64xx_enable_xmt(mvi, i);
|
|
|
|
mvs_64xx_phy_reset(mvi, i, MVS_HARD_RESET);
|
|
msleep(500);
|
|
mvs_64xx_detect_porttype(mvi, i);
|
|
}
|
|
if (mvi->flags & MVF_FLAG_SOC) {
|
|
/* set select registers */
|
|
writel(0x0E008000, regs + 0x000);
|
|
writel(0x59000008, regs + 0x004);
|
|
writel(0x20, regs + 0x008);
|
|
writel(0x20, regs + 0x00c);
|
|
writel(0x20, regs + 0x010);
|
|
writel(0x20, regs + 0x014);
|
|
writel(0x20, regs + 0x018);
|
|
writel(0x20, regs + 0x01c);
|
|
}
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
/* clear phy int status */
|
|
tmp = mvs_read_port_irq_stat(mvi, i);
|
|
tmp &= ~PHYEV_SIG_FIS;
|
|
mvs_write_port_irq_stat(mvi, i, tmp);
|
|
|
|
/* set phy int mask */
|
|
tmp = PHYEV_RDY_CH | PHYEV_BROAD_CH | PHYEV_UNASSOC_FIS |
|
|
PHYEV_ID_DONE | PHYEV_DCDR_ERR | PHYEV_CRC_ERR |
|
|
PHYEV_DEC_ERR;
|
|
mvs_write_port_irq_mask(mvi, i, tmp);
|
|
|
|
msleep(100);
|
|
mvs_update_phyinfo(mvi, i, 1);
|
|
}
|
|
|
|
/* little endian for open address and command table, etc. */
|
|
cctl = mr32(MVS_CTL);
|
|
cctl |= CCTL_ENDIAN_CMD;
|
|
cctl |= CCTL_ENDIAN_DATA;
|
|
cctl &= ~CCTL_ENDIAN_OPEN;
|
|
cctl |= CCTL_ENDIAN_RSP;
|
|
mw32_f(MVS_CTL, cctl);
|
|
|
|
/* reset CMD queue */
|
|
tmp = mr32(MVS_PCS);
|
|
tmp |= PCS_CMD_RST;
|
|
tmp &= ~PCS_SELF_CLEAR;
|
|
mw32(MVS_PCS, tmp);
|
|
/*
|
|
* the max count is 0x1ff, while our max slot is 0x200,
|
|
* it will make count 0.
|
|
*/
|
|
tmp = 0;
|
|
if (MVS_CHIP_SLOT_SZ > 0x1ff)
|
|
mw32(MVS_INT_COAL, 0x1ff | COAL_EN);
|
|
else
|
|
mw32(MVS_INT_COAL, MVS_CHIP_SLOT_SZ | COAL_EN);
|
|
|
|
tmp = 0x10000 | interrupt_coalescing;
|
|
mw32(MVS_INT_COAL_TMOUT, tmp);
|
|
|
|
/* ladies and gentlemen, start your engines */
|
|
mw32(MVS_TX_CFG, 0);
|
|
mw32(MVS_TX_CFG, MVS_CHIP_SLOT_SZ | TX_EN);
|
|
mw32(MVS_RX_CFG, MVS_RX_RING_SZ | RX_EN);
|
|
/* enable CMD/CMPL_Q/RESP mode */
|
|
mw32(MVS_PCS, PCS_SATA_RETRY | PCS_FIS_RX_EN |
|
|
PCS_CMD_EN | PCS_CMD_STOP_ERR);
|
|
|
|
/* enable completion queue interrupt */
|
|
tmp = (CINT_PORT_MASK | CINT_DONE | CINT_MEM | CINT_SRS | CINT_CI_STOP |
|
|
CINT_DMA_PCIE);
|
|
|
|
mw32(MVS_INT_MASK, tmp);
|
|
|
|
/* Enable SRS interrupt */
|
|
mw32(MVS_INT_MASK_SRS_0, 0xFFFF);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvs_64xx_ioremap(struct mvs_info *mvi)
|
|
{
|
|
if (!mvs_ioremap(mvi, 4, 2))
|
|
return 0;
|
|
return -1;
|
|
}
|
|
|
|
static void mvs_64xx_iounmap(struct mvs_info *mvi)
|
|
{
|
|
mvs_iounmap(mvi->regs);
|
|
mvs_iounmap(mvi->regs_ex);
|
|
}
|
|
|
|
static void mvs_64xx_interrupt_enable(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
tmp = mr32(MVS_GBL_CTL);
|
|
mw32(MVS_GBL_CTL, tmp | INT_EN);
|
|
}
|
|
|
|
static void mvs_64xx_interrupt_disable(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
tmp = mr32(MVS_GBL_CTL);
|
|
mw32(MVS_GBL_CTL, tmp & ~INT_EN);
|
|
}
|
|
|
|
static u32 mvs_64xx_isr_status(struct mvs_info *mvi, int irq)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 stat;
|
|
|
|
if (!(mvi->flags & MVF_FLAG_SOC)) {
|
|
stat = mr32(MVS_GBL_INT_STAT);
|
|
|
|
if (stat == 0 || stat == 0xffffffff)
|
|
return 0;
|
|
} else
|
|
stat = 1;
|
|
return stat;
|
|
}
|
|
|
|
static irqreturn_t mvs_64xx_isr(struct mvs_info *mvi, int irq, u32 stat)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
|
|
/* clear CMD_CMPLT ASAP */
|
|
mw32_f(MVS_INT_STAT, CINT_DONE);
|
|
|
|
spin_lock(&mvi->lock);
|
|
mvs_int_full(mvi);
|
|
spin_unlock(&mvi->lock);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void mvs_64xx_command_active(struct mvs_info *mvi, u32 slot_idx)
|
|
{
|
|
u32 tmp;
|
|
mvs_cw32(mvi, 0x40 + (slot_idx >> 3), 1 << (slot_idx % 32));
|
|
mvs_cw32(mvi, 0x00 + (slot_idx >> 3), 1 << (slot_idx % 32));
|
|
do {
|
|
tmp = mvs_cr32(mvi, 0x00 + (slot_idx >> 3));
|
|
} while (tmp & 1 << (slot_idx % 32));
|
|
do {
|
|
tmp = mvs_cr32(mvi, 0x40 + (slot_idx >> 3));
|
|
} while (tmp & 1 << (slot_idx % 32));
|
|
}
|
|
|
|
static void mvs_64xx_issue_stop(struct mvs_info *mvi, enum mvs_port_type type,
|
|
u32 tfs)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
if (type == PORT_TYPE_SATA) {
|
|
tmp = mr32(MVS_INT_STAT_SRS_0) | (1U << tfs);
|
|
mw32(MVS_INT_STAT_SRS_0, tmp);
|
|
}
|
|
mw32(MVS_INT_STAT, CINT_CI_STOP);
|
|
tmp = mr32(MVS_PCS) | 0xFF00;
|
|
mw32(MVS_PCS, tmp);
|
|
}
|
|
|
|
static void mvs_64xx_free_reg_set(struct mvs_info *mvi, u8 *tfs)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp, offs;
|
|
|
|
if (*tfs == MVS_ID_NOT_MAPPED)
|
|
return;
|
|
|
|
offs = 1U << ((*tfs & 0x0f) + PCS_EN_SATA_REG_SHIFT);
|
|
if (*tfs < 16) {
|
|
tmp = mr32(MVS_PCS);
|
|
mw32(MVS_PCS, tmp & ~offs);
|
|
} else {
|
|
tmp = mr32(MVS_CTL);
|
|
mw32(MVS_CTL, tmp & ~offs);
|
|
}
|
|
|
|
tmp = mr32(MVS_INT_STAT_SRS_0) & (1U << *tfs);
|
|
if (tmp)
|
|
mw32(MVS_INT_STAT_SRS_0, tmp);
|
|
|
|
*tfs = MVS_ID_NOT_MAPPED;
|
|
return;
|
|
}
|
|
|
|
static u8 mvs_64xx_assign_reg_set(struct mvs_info *mvi, u8 *tfs)
|
|
{
|
|
int i;
|
|
u32 tmp, offs;
|
|
void __iomem *regs = mvi->regs;
|
|
|
|
if (*tfs != MVS_ID_NOT_MAPPED)
|
|
return 0;
|
|
|
|
tmp = mr32(MVS_PCS);
|
|
|
|
for (i = 0; i < mvi->chip->srs_sz; i++) {
|
|
if (i == 16)
|
|
tmp = mr32(MVS_CTL);
|
|
offs = 1U << ((i & 0x0f) + PCS_EN_SATA_REG_SHIFT);
|
|
if (!(tmp & offs)) {
|
|
*tfs = i;
|
|
|
|
if (i < 16)
|
|
mw32(MVS_PCS, tmp | offs);
|
|
else
|
|
mw32(MVS_CTL, tmp | offs);
|
|
tmp = mr32(MVS_INT_STAT_SRS_0) & (1U << i);
|
|
if (tmp)
|
|
mw32(MVS_INT_STAT_SRS_0, tmp);
|
|
return 0;
|
|
}
|
|
}
|
|
return MVS_ID_NOT_MAPPED;
|
|
}
|
|
|
|
static void mvs_64xx_make_prd(struct scatterlist *scatter, int nr, void *prd)
|
|
{
|
|
int i;
|
|
struct scatterlist *sg;
|
|
struct mvs_prd *buf_prd = prd;
|
|
for_each_sg(scatter, sg, nr, i) {
|
|
buf_prd->addr = cpu_to_le64(sg_dma_address(sg));
|
|
buf_prd->len = cpu_to_le32(sg_dma_len(sg));
|
|
buf_prd++;
|
|
}
|
|
}
|
|
|
|
static int mvs_64xx_oob_done(struct mvs_info *mvi, int i)
|
|
{
|
|
u32 phy_st;
|
|
mvs_write_port_cfg_addr(mvi, i,
|
|
PHYR_PHY_STAT);
|
|
phy_st = mvs_read_port_cfg_data(mvi, i);
|
|
if (phy_st & PHY_OOB_DTCTD)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
static void mvs_64xx_fix_phy_info(struct mvs_info *mvi, int i,
|
|
struct sas_identify_frame *id)
|
|
|
|
{
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
sas_phy->linkrate =
|
|
(phy->phy_status & PHY_NEG_SPP_PHYS_LINK_RATE_MASK) >>
|
|
PHY_NEG_SPP_PHYS_LINK_RATE_MASK_OFFSET;
|
|
|
|
phy->minimum_linkrate =
|
|
(phy->phy_status &
|
|
PHY_MIN_SPP_PHYS_LINK_RATE_MASK) >> 8;
|
|
phy->maximum_linkrate =
|
|
(phy->phy_status &
|
|
PHY_MAX_SPP_PHYS_LINK_RATE_MASK) >> 12;
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_IDENTIFY);
|
|
phy->dev_info = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ATT_DEV_INFO);
|
|
phy->att_dev_info = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ATT_ADDR_HI);
|
|
phy->att_dev_sas_addr =
|
|
(u64) mvs_read_port_cfg_data(mvi, i) << 32;
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ATT_ADDR_LO);
|
|
phy->att_dev_sas_addr |= mvs_read_port_cfg_data(mvi, i);
|
|
phy->att_dev_sas_addr = SAS_ADDR(&phy->att_dev_sas_addr);
|
|
}
|
|
|
|
static void mvs_64xx_phy_work_around(struct mvs_info *mvi, int i)
|
|
{
|
|
u32 tmp;
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
mvs_write_port_vsr_addr(mvi, i, VSR_PHY_MODE6);
|
|
tmp = mvs_read_port_vsr_data(mvi, i);
|
|
if (((phy->phy_status & PHY_NEG_SPP_PHYS_LINK_RATE_MASK) >>
|
|
PHY_NEG_SPP_PHYS_LINK_RATE_MASK_OFFSET) ==
|
|
SAS_LINK_RATE_1_5_GBPS)
|
|
tmp &= ~PHY_MODE6_LATECLK;
|
|
else
|
|
tmp |= PHY_MODE6_LATECLK;
|
|
mvs_write_port_vsr_data(mvi, i, tmp);
|
|
}
|
|
|
|
static void mvs_64xx_phy_set_link_rate(struct mvs_info *mvi, u32 phy_id,
|
|
struct sas_phy_linkrates *rates)
|
|
{
|
|
u32 lrmin = 0, lrmax = 0;
|
|
u32 tmp;
|
|
|
|
tmp = mvs_read_phy_ctl(mvi, phy_id);
|
|
lrmin = (rates->minimum_linkrate << 8);
|
|
lrmax = (rates->maximum_linkrate << 12);
|
|
|
|
if (lrmin) {
|
|
tmp &= ~(0xf << 8);
|
|
tmp |= lrmin;
|
|
}
|
|
if (lrmax) {
|
|
tmp &= ~(0xf << 12);
|
|
tmp |= lrmax;
|
|
}
|
|
mvs_write_phy_ctl(mvi, phy_id, tmp);
|
|
mvs_64xx_phy_reset(mvi, phy_id, MVS_HARD_RESET);
|
|
}
|
|
|
|
static void mvs_64xx_clear_active_cmds(struct mvs_info *mvi)
|
|
{
|
|
u32 tmp;
|
|
void __iomem *regs = mvi->regs;
|
|
tmp = mr32(MVS_PCS);
|
|
mw32(MVS_PCS, tmp & 0xFFFF);
|
|
mw32(MVS_PCS, tmp);
|
|
tmp = mr32(MVS_CTL);
|
|
mw32(MVS_CTL, tmp & 0xFFFF);
|
|
mw32(MVS_CTL, tmp);
|
|
}
|
|
|
|
|
|
static u32 mvs_64xx_spi_read_data(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs_ex;
|
|
return ior32(SPI_DATA_REG_64XX);
|
|
}
|
|
|
|
static void mvs_64xx_spi_write_data(struct mvs_info *mvi, u32 data)
|
|
{
|
|
void __iomem *regs = mvi->regs_ex;
|
|
|
|
iow32(SPI_DATA_REG_64XX, data);
|
|
}
|
|
|
|
|
|
static int mvs_64xx_spi_buildcmd(struct mvs_info *mvi,
|
|
u32 *dwCmd,
|
|
u8 cmd,
|
|
u8 read,
|
|
u8 length,
|
|
u32 addr
|
|
)
|
|
{
|
|
u32 dwTmp;
|
|
|
|
dwTmp = ((u32)cmd << 24) | ((u32)length << 19);
|
|
if (read)
|
|
dwTmp |= 1U<<23;
|
|
|
|
if (addr != MV_MAX_U32) {
|
|
dwTmp |= 1U<<22;
|
|
dwTmp |= (addr & 0x0003FFFF);
|
|
}
|
|
|
|
*dwCmd = dwTmp;
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int mvs_64xx_spi_issuecmd(struct mvs_info *mvi, u32 cmd)
|
|
{
|
|
void __iomem *regs = mvi->regs_ex;
|
|
int retry;
|
|
|
|
for (retry = 0; retry < 1; retry++) {
|
|
iow32(SPI_CTRL_REG_64XX, SPI_CTRL_VENDOR_ENABLE);
|
|
iow32(SPI_CMD_REG_64XX, cmd);
|
|
iow32(SPI_CTRL_REG_64XX,
|
|
SPI_CTRL_VENDOR_ENABLE | SPI_CTRL_SPISTART);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvs_64xx_spi_waitdataready(struct mvs_info *mvi, u32 timeout)
|
|
{
|
|
void __iomem *regs = mvi->regs_ex;
|
|
u32 i, dwTmp;
|
|
|
|
for (i = 0; i < timeout; i++) {
|
|
dwTmp = ior32(SPI_CTRL_REG_64XX);
|
|
if (!(dwTmp & SPI_CTRL_SPISTART))
|
|
return 0;
|
|
msleep(10);
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
static void mvs_64xx_fix_dma(struct mvs_info *mvi, u32 phy_mask,
|
|
int buf_len, int from, void *prd)
|
|
{
|
|
int i;
|
|
struct mvs_prd *buf_prd = prd;
|
|
dma_addr_t buf_dma = mvi->bulk_buffer_dma;
|
|
|
|
buf_prd += from;
|
|
for (i = 0; i < MAX_SG_ENTRY - from; i++) {
|
|
buf_prd->addr = cpu_to_le64(buf_dma);
|
|
buf_prd->len = cpu_to_le32(buf_len);
|
|
++buf_prd;
|
|
}
|
|
}
|
|
|
|
static void mvs_64xx_tune_interrupt(struct mvs_info *mvi, u32 time)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp = 0;
|
|
/*
|
|
* the max count is 0x1ff, while our max slot is 0x200,
|
|
* it will make count 0.
|
|
*/
|
|
if (time == 0) {
|
|
mw32(MVS_INT_COAL, 0);
|
|
mw32(MVS_INT_COAL_TMOUT, 0x10000);
|
|
} else {
|
|
if (MVS_CHIP_SLOT_SZ > 0x1ff)
|
|
mw32(MVS_INT_COAL, 0x1ff|COAL_EN);
|
|
else
|
|
mw32(MVS_INT_COAL, MVS_CHIP_SLOT_SZ|COAL_EN);
|
|
|
|
tmp = 0x10000 | time;
|
|
mw32(MVS_INT_COAL_TMOUT, tmp);
|
|
}
|
|
}
|
|
|
|
const struct mvs_dispatch mvs_64xx_dispatch = {
|
|
"mv64xx",
|
|
mvs_64xx_init,
|
|
NULL,
|
|
mvs_64xx_ioremap,
|
|
mvs_64xx_iounmap,
|
|
mvs_64xx_isr,
|
|
mvs_64xx_isr_status,
|
|
mvs_64xx_interrupt_enable,
|
|
mvs_64xx_interrupt_disable,
|
|
mvs_read_phy_ctl,
|
|
mvs_write_phy_ctl,
|
|
mvs_read_port_cfg_data,
|
|
mvs_write_port_cfg_data,
|
|
mvs_write_port_cfg_addr,
|
|
mvs_read_port_vsr_data,
|
|
mvs_write_port_vsr_data,
|
|
mvs_write_port_vsr_addr,
|
|
mvs_read_port_irq_stat,
|
|
mvs_write_port_irq_stat,
|
|
mvs_read_port_irq_mask,
|
|
mvs_write_port_irq_mask,
|
|
mvs_64xx_command_active,
|
|
mvs_64xx_clear_srs_irq,
|
|
mvs_64xx_issue_stop,
|
|
mvs_start_delivery,
|
|
mvs_rx_update,
|
|
mvs_int_full,
|
|
mvs_64xx_assign_reg_set,
|
|
mvs_64xx_free_reg_set,
|
|
mvs_get_prd_size,
|
|
mvs_get_prd_count,
|
|
mvs_64xx_make_prd,
|
|
mvs_64xx_detect_porttype,
|
|
mvs_64xx_oob_done,
|
|
mvs_64xx_fix_phy_info,
|
|
mvs_64xx_phy_work_around,
|
|
mvs_64xx_phy_set_link_rate,
|
|
mvs_hw_max_link_rate,
|
|
mvs_64xx_phy_disable,
|
|
mvs_64xx_phy_enable,
|
|
mvs_64xx_phy_reset,
|
|
mvs_64xx_stp_reset,
|
|
mvs_64xx_clear_active_cmds,
|
|
mvs_64xx_spi_read_data,
|
|
mvs_64xx_spi_write_data,
|
|
mvs_64xx_spi_buildcmd,
|
|
mvs_64xx_spi_issuecmd,
|
|
mvs_64xx_spi_waitdataready,
|
|
mvs_64xx_fix_dma,
|
|
mvs_64xx_tune_interrupt,
|
|
NULL,
|
|
};
|
|
|