2019-05-27 06:55:01 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
2017-06-14 15:33:19 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2017 Hisilicon Limited.
|
|
|
|
*/
|
|
|
|
|
2021-09-20 12:33:24 +00:00
|
|
|
#include <linux/sched/clock.h>
|
2017-06-14 15:33:19 +00:00
|
|
|
#include "hisi_sas.h"
|
|
|
|
#define DRV_NAME "hisi_sas_v3_hw"
|
|
|
|
|
2019-02-06 10:52:56 +00:00
|
|
|
/* global registers need init */
|
2017-06-14 15:33:21 +00:00
|
|
|
#define DLVRY_QUEUE_ENABLE 0x0
|
|
|
|
#define IOST_BASE_ADDR_LO 0x8
|
|
|
|
#define IOST_BASE_ADDR_HI 0xc
|
|
|
|
#define ITCT_BASE_ADDR_LO 0x10
|
|
|
|
#define ITCT_BASE_ADDR_HI 0x14
|
|
|
|
#define IO_BROKEN_MSG_ADDR_LO 0x18
|
|
|
|
#define IO_BROKEN_MSG_ADDR_HI 0x1c
|
2017-06-14 15:33:22 +00:00
|
|
|
#define PHY_CONTEXT 0x20
|
|
|
|
#define PHY_STATE 0x24
|
|
|
|
#define PHY_PORT_NUM_MA 0x28
|
|
|
|
#define PHY_CONN_RATE 0x30
|
2017-06-14 15:33:28 +00:00
|
|
|
#define ITCT_CLR 0x44
|
|
|
|
#define ITCT_CLR_EN_OFF 16
|
|
|
|
#define ITCT_CLR_EN_MSK (0x1 << ITCT_CLR_EN_OFF)
|
|
|
|
#define ITCT_DEV_OFF 0
|
|
|
|
#define ITCT_DEV_MSK (0x7ff << ITCT_DEV_OFF)
|
2019-05-29 09:58:47 +00:00
|
|
|
#define SAS_AXI_USER3 0x50
|
2017-06-14 15:33:21 +00:00
|
|
|
#define IO_SATA_BROKEN_MSG_ADDR_LO 0x58
|
|
|
|
#define IO_SATA_BROKEN_MSG_ADDR_HI 0x5c
|
|
|
|
#define SATA_INITI_D2H_STORE_ADDR_LO 0x60
|
|
|
|
#define SATA_INITI_D2H_STORE_ADDR_HI 0x64
|
|
|
|
#define CFG_MAX_TAG 0x68
|
|
|
|
#define HGC_SAS_TX_OPEN_FAIL_RETRY_CTRL 0x84
|
|
|
|
#define HGC_SAS_TXFAIL_RETRY_CTRL 0x88
|
|
|
|
#define HGC_GET_ITV_TIME 0x90
|
|
|
|
#define DEVICE_MSG_WORK_MODE 0x94
|
|
|
|
#define OPENA_WT_CONTI_TIME 0x9c
|
|
|
|
#define I_T_NEXUS_LOSS_TIME 0xa0
|
|
|
|
#define MAX_CON_TIME_LIMIT_TIME 0xa4
|
|
|
|
#define BUS_INACTIVE_LIMIT_TIME 0xa8
|
|
|
|
#define REJECT_TO_OPEN_LIMIT_TIME 0xac
|
2018-11-09 14:06:33 +00:00
|
|
|
#define CQ_INT_CONVERGE_EN 0xb0
|
2017-06-14 15:33:21 +00:00
|
|
|
#define CFG_AGING_TIME 0xbc
|
|
|
|
#define HGC_DFX_CFG2 0xc0
|
|
|
|
#define CFG_ABT_SET_QUERY_IPTT 0xd4
|
|
|
|
#define CFG_SET_ABORTED_IPTT_OFF 0
|
|
|
|
#define CFG_SET_ABORTED_IPTT_MSK (0xfff << CFG_SET_ABORTED_IPTT_OFF)
|
2017-06-14 15:33:32 +00:00
|
|
|
#define CFG_SET_ABORTED_EN_OFF 12
|
|
|
|
#define CFG_ABT_SET_IPTT_DONE 0xd8
|
|
|
|
#define CFG_ABT_SET_IPTT_DONE_OFF 0
|
|
|
|
#define HGC_IOMB_PROC1_STATUS 0x104
|
2019-04-11 12:46:41 +00:00
|
|
|
#define HGC_LM_DFX_STATUS2 0x128
|
|
|
|
#define HGC_LM_DFX_STATUS2_IOSTLIST_OFF 0
|
|
|
|
#define HGC_LM_DFX_STATUS2_IOSTLIST_MSK (0xfff << \
|
|
|
|
HGC_LM_DFX_STATUS2_IOSTLIST_OFF)
|
|
|
|
#define HGC_LM_DFX_STATUS2_ITCTLIST_OFF 12
|
|
|
|
#define HGC_LM_DFX_STATUS2_ITCTLIST_MSK (0x7ff << \
|
|
|
|
HGC_LM_DFX_STATUS2_ITCTLIST_OFF)
|
|
|
|
#define HGC_CQE_ECC_ADDR 0x13c
|
|
|
|
#define HGC_CQE_ECC_1B_ADDR_OFF 0
|
|
|
|
#define HGC_CQE_ECC_1B_ADDR_MSK (0x3f << HGC_CQE_ECC_1B_ADDR_OFF)
|
|
|
|
#define HGC_CQE_ECC_MB_ADDR_OFF 8
|
|
|
|
#define HGC_CQE_ECC_MB_ADDR_MSK (0x3f << HGC_CQE_ECC_MB_ADDR_OFF)
|
|
|
|
#define HGC_IOST_ECC_ADDR 0x140
|
|
|
|
#define HGC_IOST_ECC_1B_ADDR_OFF 0
|
|
|
|
#define HGC_IOST_ECC_1B_ADDR_MSK (0x3ff << HGC_IOST_ECC_1B_ADDR_OFF)
|
|
|
|
#define HGC_IOST_ECC_MB_ADDR_OFF 16
|
|
|
|
#define HGC_IOST_ECC_MB_ADDR_MSK (0x3ff << HGC_IOST_ECC_MB_ADDR_OFF)
|
|
|
|
#define HGC_DQE_ECC_ADDR 0x144
|
|
|
|
#define HGC_DQE_ECC_1B_ADDR_OFF 0
|
|
|
|
#define HGC_DQE_ECC_1B_ADDR_MSK (0xfff << HGC_DQE_ECC_1B_ADDR_OFF)
|
|
|
|
#define HGC_DQE_ECC_MB_ADDR_OFF 16
|
|
|
|
#define HGC_DQE_ECC_MB_ADDR_MSK (0xfff << HGC_DQE_ECC_MB_ADDR_OFF)
|
2017-06-14 15:33:22 +00:00
|
|
|
#define CHNL_INT_STATUS 0x148
|
2019-08-05 13:48:01 +00:00
|
|
|
#define TAB_DFX 0x14c
|
2019-04-11 12:46:41 +00:00
|
|
|
#define HGC_ITCT_ECC_ADDR 0x150
|
|
|
|
#define HGC_ITCT_ECC_1B_ADDR_OFF 0
|
|
|
|
#define HGC_ITCT_ECC_1B_ADDR_MSK (0x3ff << \
|
|
|
|
HGC_ITCT_ECC_1B_ADDR_OFF)
|
|
|
|
#define HGC_ITCT_ECC_MB_ADDR_OFF 16
|
|
|
|
#define HGC_ITCT_ECC_MB_ADDR_MSK (0x3ff << \
|
|
|
|
HGC_ITCT_ECC_MB_ADDR_OFF)
|
2017-10-24 15:51:48 +00:00
|
|
|
#define HGC_AXI_FIFO_ERR_INFO 0x154
|
|
|
|
#define AXI_ERR_INFO_OFF 0
|
|
|
|
#define AXI_ERR_INFO_MSK (0xff << AXI_ERR_INFO_OFF)
|
|
|
|
#define FIFO_ERR_INFO_OFF 8
|
|
|
|
#define FIFO_ERR_INFO_MSK (0xff << FIFO_ERR_INFO_OFF)
|
2019-08-05 13:48:01 +00:00
|
|
|
#define TAB_RD_TYPE 0x15c
|
2017-06-14 15:33:21 +00:00
|
|
|
#define INT_COAL_EN 0x19c
|
|
|
|
#define OQ_INT_COAL_TIME 0x1a0
|
|
|
|
#define OQ_INT_COAL_CNT 0x1a4
|
|
|
|
#define ENT_INT_COAL_TIME 0x1a8
|
|
|
|
#define ENT_INT_COAL_CNT 0x1ac
|
|
|
|
#define OQ_INT_SRC 0x1b0
|
|
|
|
#define OQ_INT_SRC_MSK 0x1b4
|
|
|
|
#define ENT_INT_SRC1 0x1b8
|
|
|
|
#define ENT_INT_SRC1_D2H_FIS_CH0_OFF 0
|
|
|
|
#define ENT_INT_SRC1_D2H_FIS_CH0_MSK (0x1 << ENT_INT_SRC1_D2H_FIS_CH0_OFF)
|
|
|
|
#define ENT_INT_SRC1_D2H_FIS_CH1_OFF 8
|
|
|
|
#define ENT_INT_SRC1_D2H_FIS_CH1_MSK (0x1 << ENT_INT_SRC1_D2H_FIS_CH1_OFF)
|
|
|
|
#define ENT_INT_SRC2 0x1bc
|
|
|
|
#define ENT_INT_SRC3 0x1c0
|
|
|
|
#define ENT_INT_SRC3_WP_DEPTH_OFF 8
|
|
|
|
#define ENT_INT_SRC3_IPTT_SLOT_NOMATCH_OFF 9
|
|
|
|
#define ENT_INT_SRC3_RP_DEPTH_OFF 10
|
|
|
|
#define ENT_INT_SRC3_AXI_OFF 11
|
|
|
|
#define ENT_INT_SRC3_FIFO_OFF 12
|
|
|
|
#define ENT_INT_SRC3_LM_OFF 14
|
|
|
|
#define ENT_INT_SRC3_ITC_INT_OFF 15
|
|
|
|
#define ENT_INT_SRC3_ITC_INT_MSK (0x1 << ENT_INT_SRC3_ITC_INT_OFF)
|
|
|
|
#define ENT_INT_SRC3_ABT_OFF 16
|
2019-04-11 12:46:41 +00:00
|
|
|
#define ENT_INT_SRC3_DQE_POISON_OFF 18
|
|
|
|
#define ENT_INT_SRC3_IOST_POISON_OFF 19
|
|
|
|
#define ENT_INT_SRC3_ITCT_POISON_OFF 20
|
|
|
|
#define ENT_INT_SRC3_ITCT_NCQ_POISON_OFF 21
|
2017-06-14 15:33:21 +00:00
|
|
|
#define ENT_INT_SRC_MSK1 0x1c4
|
|
|
|
#define ENT_INT_SRC_MSK2 0x1c8
|
|
|
|
#define ENT_INT_SRC_MSK3 0x1cc
|
2017-06-14 15:33:22 +00:00
|
|
|
#define ENT_INT_SRC_MSK3_ENT95_MSK_OFF 31
|
2017-06-14 15:33:21 +00:00
|
|
|
#define CHNL_PHYUPDOWN_INT_MSK 0x1d0
|
|
|
|
#define CHNL_ENT_INT_MSK 0x1d4
|
|
|
|
#define HGC_COM_INT_MSK 0x1d8
|
2017-06-14 15:33:22 +00:00
|
|
|
#define ENT_INT_SRC_MSK3_ENT95_MSK_MSK (0x1 << ENT_INT_SRC_MSK3_ENT95_MSK_OFF)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define SAS_ECC_INTR 0x1e8
|
2019-04-11 12:46:41 +00:00
|
|
|
#define SAS_ECC_INTR_DQE_ECC_1B_OFF 0
|
|
|
|
#define SAS_ECC_INTR_DQE_ECC_MB_OFF 1
|
|
|
|
#define SAS_ECC_INTR_IOST_ECC_1B_OFF 2
|
|
|
|
#define SAS_ECC_INTR_IOST_ECC_MB_OFF 3
|
|
|
|
#define SAS_ECC_INTR_ITCT_ECC_1B_OFF 4
|
|
|
|
#define SAS_ECC_INTR_ITCT_ECC_MB_OFF 5
|
|
|
|
#define SAS_ECC_INTR_ITCTLIST_ECC_1B_OFF 6
|
|
|
|
#define SAS_ECC_INTR_ITCTLIST_ECC_MB_OFF 7
|
|
|
|
#define SAS_ECC_INTR_IOSTLIST_ECC_1B_OFF 8
|
|
|
|
#define SAS_ECC_INTR_IOSTLIST_ECC_MB_OFF 9
|
|
|
|
#define SAS_ECC_INTR_CQE_ECC_1B_OFF 10
|
|
|
|
#define SAS_ECC_INTR_CQE_ECC_MB_OFF 11
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM0_ECC_1B_OFF 12
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM0_ECC_MB_OFF 13
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM1_ECC_1B_OFF 14
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM1_ECC_MB_OFF 15
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM2_ECC_1B_OFF 16
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM2_ECC_MB_OFF 17
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM3_ECC_1B_OFF 18
|
|
|
|
#define SAS_ECC_INTR_NCQ_MEM3_ECC_MB_OFF 19
|
|
|
|
#define SAS_ECC_INTR_OOO_RAM_ECC_1B_OFF 20
|
|
|
|
#define SAS_ECC_INTR_OOO_RAM_ECC_MB_OFF 21
|
2017-06-14 15:33:21 +00:00
|
|
|
#define SAS_ECC_INTR_MSK 0x1ec
|
|
|
|
#define HGC_ERR_STAT_EN 0x238
|
2018-05-21 10:09:20 +00:00
|
|
|
#define CQE_SEND_CNT 0x248
|
2017-06-14 15:33:21 +00:00
|
|
|
#define DLVRY_Q_0_BASE_ADDR_LO 0x260
|
|
|
|
#define DLVRY_Q_0_BASE_ADDR_HI 0x264
|
|
|
|
#define DLVRY_Q_0_DEPTH 0x268
|
|
|
|
#define DLVRY_Q_0_WR_PTR 0x26c
|
|
|
|
#define DLVRY_Q_0_RD_PTR 0x270
|
|
|
|
#define HYPER_STREAM_ID_EN_CFG 0xc80
|
|
|
|
#define OQ0_INT_SRC_MSK 0xc90
|
|
|
|
#define COMPL_Q_0_BASE_ADDR_LO 0x4e0
|
|
|
|
#define COMPL_Q_0_BASE_ADDR_HI 0x4e4
|
|
|
|
#define COMPL_Q_0_DEPTH 0x4e8
|
|
|
|
#define COMPL_Q_0_WR_PTR 0x4ec
|
|
|
|
#define COMPL_Q_0_RD_PTR 0x4f0
|
2019-04-11 12:46:41 +00:00
|
|
|
#define HGC_RXM_DFX_STATUS14 0xae8
|
|
|
|
#define HGC_RXM_DFX_STATUS14_MEM0_OFF 0
|
|
|
|
#define HGC_RXM_DFX_STATUS14_MEM0_MSK (0x1ff << \
|
|
|
|
HGC_RXM_DFX_STATUS14_MEM0_OFF)
|
|
|
|
#define HGC_RXM_DFX_STATUS14_MEM1_OFF 9
|
|
|
|
#define HGC_RXM_DFX_STATUS14_MEM1_MSK (0x1ff << \
|
|
|
|
HGC_RXM_DFX_STATUS14_MEM1_OFF)
|
|
|
|
#define HGC_RXM_DFX_STATUS14_MEM2_OFF 18
|
|
|
|
#define HGC_RXM_DFX_STATUS14_MEM2_MSK (0x1ff << \
|
|
|
|
HGC_RXM_DFX_STATUS14_MEM2_OFF)
|
|
|
|
#define HGC_RXM_DFX_STATUS15 0xaec
|
|
|
|
#define HGC_RXM_DFX_STATUS15_MEM3_OFF 0
|
|
|
|
#define HGC_RXM_DFX_STATUS15_MEM3_MSK (0x1ff << \
|
|
|
|
HGC_RXM_DFX_STATUS15_MEM3_OFF)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define AWQOS_AWCACHE_CFG 0xc84
|
|
|
|
#define ARQOS_ARCACHE_CFG 0xc88
|
2018-05-02 15:56:34 +00:00
|
|
|
#define HILINK_ERR_DFX 0xe04
|
2018-05-21 10:09:16 +00:00
|
|
|
#define SAS_GPIO_CFG_0 0x1000
|
|
|
|
#define SAS_GPIO_CFG_1 0x1004
|
|
|
|
#define SAS_GPIO_TX_0_1 0x1040
|
|
|
|
#define SAS_CFG_DRIVE_VLD 0x1070
|
2017-06-14 15:33:21 +00:00
|
|
|
|
|
|
|
/* phy registers requiring init */
|
|
|
|
#define PORT_BASE (0x2000)
|
2017-06-14 15:33:22 +00:00
|
|
|
#define PHY_CFG (PORT_BASE + 0x0)
|
|
|
|
#define HARD_PHY_LINKRATE (PORT_BASE + 0x4)
|
|
|
|
#define PHY_CFG_ENA_OFF 0
|
|
|
|
#define PHY_CFG_ENA_MSK (0x1 << PHY_CFG_ENA_OFF)
|
|
|
|
#define PHY_CFG_DC_OPT_OFF 2
|
|
|
|
#define PHY_CFG_DC_OPT_MSK (0x1 << PHY_CFG_DC_OPT_OFF)
|
2018-07-18 14:14:27 +00:00
|
|
|
#define PHY_CFG_PHY_RST_OFF 3
|
|
|
|
#define PHY_CFG_PHY_RST_MSK (0x1 << PHY_CFG_PHY_RST_OFF)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define PROG_PHY_LINK_RATE (PORT_BASE + 0x8)
|
2020-09-01 11:13:05 +00:00
|
|
|
#define CFG_PROG_PHY_LINK_RATE_OFF 0
|
|
|
|
#define CFG_PROG_PHY_LINK_RATE_MSK (0xff << CFG_PROG_PHY_LINK_RATE_OFF)
|
2020-09-01 11:13:04 +00:00
|
|
|
#define CFG_PROG_OOB_PHY_LINK_RATE_OFF 8
|
|
|
|
#define CFG_PROG_OOB_PHY_LINK_RATE_MSK (0xf << CFG_PROG_OOB_PHY_LINK_RATE_OFF)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define PHY_CTRL (PORT_BASE + 0x14)
|
|
|
|
#define PHY_CTRL_RESET_OFF 0
|
|
|
|
#define PHY_CTRL_RESET_MSK (0x1 << PHY_CTRL_RESET_OFF)
|
2018-12-17 14:40:07 +00:00
|
|
|
#define CMD_HDR_PIR_OFF 8
|
|
|
|
#define CMD_HDR_PIR_MSK (0x1 << CMD_HDR_PIR_OFF)
|
2019-02-28 14:51:02 +00:00
|
|
|
#define SERDES_CFG (PORT_BASE + 0x1c)
|
2019-09-06 12:55:36 +00:00
|
|
|
#define CFG_ALOS_CHK_DISABLE_OFF 9
|
|
|
|
#define CFG_ALOS_CHK_DISABLE_MSK (0x1 << CFG_ALOS_CHK_DISABLE_OFF)
|
|
|
|
#define SAS_PHY_BIST_CTRL (PORT_BASE + 0x2c)
|
|
|
|
#define CFG_BIST_MODE_SEL_OFF 0
|
|
|
|
#define CFG_BIST_MODE_SEL_MSK (0xf << CFG_BIST_MODE_SEL_OFF)
|
|
|
|
#define CFG_LOOP_TEST_MODE_OFF 14
|
|
|
|
#define CFG_LOOP_TEST_MODE_MSK (0x3 << CFG_LOOP_TEST_MODE_OFF)
|
|
|
|
#define CFG_RX_BIST_EN_OFF 16
|
|
|
|
#define CFG_RX_BIST_EN_MSK (0x1 << CFG_RX_BIST_EN_OFF)
|
|
|
|
#define CFG_TX_BIST_EN_OFF 17
|
|
|
|
#define CFG_TX_BIST_EN_MSK (0x1 << CFG_TX_BIST_EN_OFF)
|
|
|
|
#define CFG_BIST_TEST_OFF 18
|
|
|
|
#define CFG_BIST_TEST_MSK (0x1 << CFG_BIST_TEST_OFF)
|
|
|
|
#define SAS_PHY_BIST_CODE (PORT_BASE + 0x30)
|
|
|
|
#define SAS_PHY_BIST_CODE1 (PORT_BASE + 0x34)
|
|
|
|
#define SAS_BIST_ERR_CNT (PORT_BASE + 0x38)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define SL_CFG (PORT_BASE + 0x84)
|
2018-09-24 15:06:34 +00:00
|
|
|
#define AIP_LIMIT (PORT_BASE + 0x90)
|
2017-06-14 15:33:22 +00:00
|
|
|
#define SL_CONTROL (PORT_BASE + 0x94)
|
|
|
|
#define SL_CONTROL_NOTIFY_EN_OFF 0
|
|
|
|
#define SL_CONTROL_NOTIFY_EN_MSK (0x1 << SL_CONTROL_NOTIFY_EN_OFF)
|
|
|
|
#define SL_CTA_OFF 17
|
|
|
|
#define SL_CTA_MSK (0x1 << SL_CTA_OFF)
|
2018-05-31 12:50:43 +00:00
|
|
|
#define RX_PRIMS_STATUS (PORT_BASE + 0x98)
|
|
|
|
#define RX_BCAST_CHG_OFF 1
|
|
|
|
#define RX_BCAST_CHG_MSK (0x1 << RX_BCAST_CHG_OFF)
|
2017-06-14 15:33:22 +00:00
|
|
|
#define TX_ID_DWORD0 (PORT_BASE + 0x9c)
|
|
|
|
#define TX_ID_DWORD1 (PORT_BASE + 0xa0)
|
|
|
|
#define TX_ID_DWORD2 (PORT_BASE + 0xa4)
|
|
|
|
#define TX_ID_DWORD3 (PORT_BASE + 0xa8)
|
|
|
|
#define TX_ID_DWORD4 (PORT_BASE + 0xaC)
|
|
|
|
#define TX_ID_DWORD5 (PORT_BASE + 0xb0)
|
|
|
|
#define TX_ID_DWORD6 (PORT_BASE + 0xb4)
|
|
|
|
#define TXID_AUTO (PORT_BASE + 0xb8)
|
|
|
|
#define CT3_OFF 1
|
|
|
|
#define CT3_MSK (0x1 << CT3_OFF)
|
2017-06-14 15:33:31 +00:00
|
|
|
#define TX_HARDRST_OFF 2
|
|
|
|
#define TX_HARDRST_MSK (0x1 << TX_HARDRST_OFF)
|
2017-06-14 15:33:22 +00:00
|
|
|
#define RX_IDAF_DWORD0 (PORT_BASE + 0xc4)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define RXOP_CHECK_CFG_H (PORT_BASE + 0xfc)
|
2017-08-10 16:09:36 +00:00
|
|
|
#define STP_LINK_TIMER (PORT_BASE + 0x120)
|
2017-12-08 17:16:45 +00:00
|
|
|
#define STP_LINK_TIMEOUT_STATE (PORT_BASE + 0x124)
|
2017-10-24 15:51:43 +00:00
|
|
|
#define CON_CFG_DRIVER (PORT_BASE + 0x130)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define SAS_SSP_CON_TIMER_CFG (PORT_BASE + 0x134)
|
|
|
|
#define SAS_SMP_CON_TIMER_CFG (PORT_BASE + 0x138)
|
|
|
|
#define SAS_STP_CON_TIMER_CFG (PORT_BASE + 0x13c)
|
|
|
|
#define CHL_INT0 (PORT_BASE + 0x1b4)
|
|
|
|
#define CHL_INT0_HOTPLUG_TOUT_OFF 0
|
|
|
|
#define CHL_INT0_HOTPLUG_TOUT_MSK (0x1 << CHL_INT0_HOTPLUG_TOUT_OFF)
|
|
|
|
#define CHL_INT0_SL_RX_BCST_ACK_OFF 1
|
|
|
|
#define CHL_INT0_SL_RX_BCST_ACK_MSK (0x1 << CHL_INT0_SL_RX_BCST_ACK_OFF)
|
|
|
|
#define CHL_INT0_SL_PHY_ENABLE_OFF 2
|
|
|
|
#define CHL_INT0_SL_PHY_ENABLE_MSK (0x1 << CHL_INT0_SL_PHY_ENABLE_OFF)
|
|
|
|
#define CHL_INT0_NOT_RDY_OFF 4
|
|
|
|
#define CHL_INT0_NOT_RDY_MSK (0x1 << CHL_INT0_NOT_RDY_OFF)
|
|
|
|
#define CHL_INT0_PHY_RDY_OFF 5
|
|
|
|
#define CHL_INT0_PHY_RDY_MSK (0x1 << CHL_INT0_PHY_RDY_OFF)
|
|
|
|
#define CHL_INT1 (PORT_BASE + 0x1b8)
|
2019-04-11 12:46:41 +00:00
|
|
|
#define CHL_INT1_DMAC_TX_ECC_MB_ERR_OFF 15
|
|
|
|
#define CHL_INT1_DMAC_TX_ECC_1B_ERR_OFF 16
|
|
|
|
#define CHL_INT1_DMAC_RX_ECC_MB_ERR_OFF 17
|
|
|
|
#define CHL_INT1_DMAC_RX_ECC_1B_ERR_OFF 18
|
2017-10-24 15:51:49 +00:00
|
|
|
#define CHL_INT1_DMAC_TX_AXI_WR_ERR_OFF 19
|
|
|
|
#define CHL_INT1_DMAC_TX_AXI_RD_ERR_OFF 20
|
|
|
|
#define CHL_INT1_DMAC_RX_AXI_WR_ERR_OFF 21
|
|
|
|
#define CHL_INT1_DMAC_RX_AXI_RD_ERR_OFF 22
|
2019-04-11 12:46:41 +00:00
|
|
|
#define CHL_INT1_DMAC_TX_FIFO_ERR_OFF 23
|
|
|
|
#define CHL_INT1_DMAC_RX_FIFO_ERR_OFF 24
|
|
|
|
#define CHL_INT1_DMAC_TX_AXI_RUSER_ERR_OFF 26
|
|
|
|
#define CHL_INT1_DMAC_RX_AXI_RUSER_ERR_OFF 27
|
2017-06-14 15:33:21 +00:00
|
|
|
#define CHL_INT2 (PORT_BASE + 0x1bc)
|
2017-12-08 17:16:45 +00:00
|
|
|
#define CHL_INT2_SL_IDAF_TOUT_CONF_OFF 0
|
2019-02-28 14:50:59 +00:00
|
|
|
#define CHL_INT2_RX_DISP_ERR_OFF 28
|
|
|
|
#define CHL_INT2_RX_CODE_ERR_OFF 29
|
2018-05-02 15:56:34 +00:00
|
|
|
#define CHL_INT2_RX_INVLD_DW_OFF 30
|
2017-12-08 17:16:45 +00:00
|
|
|
#define CHL_INT2_STP_LINK_TIMEOUT_OFF 31
|
2017-06-14 15:33:21 +00:00
|
|
|
#define CHL_INT0_MSK (PORT_BASE + 0x1c0)
|
|
|
|
#define CHL_INT1_MSK (PORT_BASE + 0x1c4)
|
|
|
|
#define CHL_INT2_MSK (PORT_BASE + 0x1c8)
|
2018-12-19 15:56:43 +00:00
|
|
|
#define SAS_EC_INT_COAL_TIME (PORT_BASE + 0x1cc)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define CHL_INT_COAL_EN (PORT_BASE + 0x1d0)
|
2018-03-07 12:25:13 +00:00
|
|
|
#define SAS_RX_TRAIN_TIMER (PORT_BASE + 0x2a4)
|
2017-06-14 15:33:21 +00:00
|
|
|
#define PHY_CTRL_RDY_MSK (PORT_BASE + 0x2b0)
|
|
|
|
#define PHYCTRL_NOT_RDY_MSK (PORT_BASE + 0x2b4)
|
|
|
|
#define PHYCTRL_DWS_RESET_MSK (PORT_BASE + 0x2b8)
|
|
|
|
#define PHYCTRL_PHY_ENA_MSK (PORT_BASE + 0x2bc)
|
|
|
|
#define SL_RX_BCAST_CHK_MSK (PORT_BASE + 0x2c0)
|
|
|
|
#define PHYCTRL_OOB_RESTART_MSK (PORT_BASE + 0x2c4)
|
2017-08-10 16:09:40 +00:00
|
|
|
#define DMA_TX_STATUS (PORT_BASE + 0x2d0)
|
|
|
|
#define DMA_TX_STATUS_BUSY_OFF 0
|
|
|
|
#define DMA_TX_STATUS_BUSY_MSK (0x1 << DMA_TX_STATUS_BUSY_OFF)
|
|
|
|
#define DMA_RX_STATUS (PORT_BASE + 0x2e8)
|
|
|
|
#define DMA_RX_STATUS_BUSY_OFF 0
|
|
|
|
#define DMA_RX_STATUS_BUSY_MSK (0x1 << DMA_RX_STATUS_BUSY_OFF)
|
2018-03-07 12:25:13 +00:00
|
|
|
|
|
|
|
#define COARSETUNE_TIME (PORT_BASE + 0x304)
|
2020-09-01 11:13:07 +00:00
|
|
|
#define TXDEEMPH_G1 (PORT_BASE + 0x350)
|
2017-10-24 15:51:44 +00:00
|
|
|
#define ERR_CNT_DWS_LOST (PORT_BASE + 0x380)
|
|
|
|
#define ERR_CNT_RESET_PROB (PORT_BASE + 0x384)
|
|
|
|
#define ERR_CNT_INVLD_DW (PORT_BASE + 0x390)
|
2018-12-19 15:56:43 +00:00
|
|
|
#define ERR_CNT_CODE_ERR (PORT_BASE + 0x394)
|
2017-10-24 15:51:44 +00:00
|
|
|
#define ERR_CNT_DISP_ERR (PORT_BASE + 0x398)
|
2021-01-26 11:04:28 +00:00
|
|
|
#define DFX_FIFO_CTRL (PORT_BASE + 0x3a0)
|
|
|
|
#define DFX_FIFO_CTRL_TRIGGER_MODE_OFF 0
|
|
|
|
#define DFX_FIFO_CTRL_TRIGGER_MODE_MSK (0x7 << DFX_FIFO_CTRL_TRIGGER_MODE_OFF)
|
|
|
|
#define DFX_FIFO_CTRL_DUMP_MODE_OFF 3
|
|
|
|
#define DFX_FIFO_CTRL_DUMP_MODE_MSK (0x7 << DFX_FIFO_CTRL_DUMP_MODE_OFF)
|
|
|
|
#define DFX_FIFO_CTRL_SIGNAL_SEL_OFF 6
|
|
|
|
#define DFX_FIFO_CTRL_SIGNAL_SEL_MSK (0xF << DFX_FIFO_CTRL_SIGNAL_SEL_OFF)
|
|
|
|
#define DFX_FIFO_CTRL_DUMP_DISABLE_OFF 10
|
|
|
|
#define DFX_FIFO_CTRL_DUMP_DISABLE_MSK (0x1 << DFX_FIFO_CTRL_DUMP_DISABLE_OFF)
|
|
|
|
#define DFX_FIFO_TRIGGER (PORT_BASE + 0x3a4)
|
|
|
|
#define DFX_FIFO_TRIGGER_MSK (PORT_BASE + 0x3a8)
|
|
|
|
#define DFX_FIFO_DUMP_MSK (PORT_BASE + 0x3aC)
|
|
|
|
#define DFX_FIFO_RD_DATA (PORT_BASE + 0x3b0)
|
2017-08-10 16:09:40 +00:00
|
|
|
|
|
|
|
#define DEFAULT_ITCT_HW 2048 /* reset value, not reprogrammed */
|
|
|
|
#if (HISI_SAS_MAX_DEVICES > DEFAULT_ITCT_HW)
|
|
|
|
#error Max ITCT exceeded
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define AXI_MASTER_CFG_BASE (0x5000)
|
|
|
|
#define AM_CTRL_GLOBAL (0x0)
|
2018-07-18 14:14:29 +00:00
|
|
|
#define AM_CTRL_SHUTDOWN_REQ_OFF 0
|
|
|
|
#define AM_CTRL_SHUTDOWN_REQ_MSK (0x1 << AM_CTRL_SHUTDOWN_REQ_OFF)
|
2017-08-10 16:09:40 +00:00
|
|
|
#define AM_CURR_TRANS_RETURN (0x150)
|
|
|
|
|
|
|
|
#define AM_CFG_MAX_TRANS (0x5010)
|
|
|
|
#define AM_CFG_SINGLE_PORT_MAX_TRANS (0x5014)
|
|
|
|
#define AXI_CFG (0x5100)
|
|
|
|
#define AM_ROB_ECC_ERR_ADDR (0x510c)
|
2019-04-11 12:46:41 +00:00
|
|
|
#define AM_ROB_ECC_ERR_ADDR_OFF 0
|
|
|
|
#define AM_ROB_ECC_ERR_ADDR_MSK 0xffffffff
|
2017-06-14 15:33:21 +00:00
|
|
|
|
2017-12-08 17:16:40 +00:00
|
|
|
/* RAS registers need init */
|
|
|
|
#define RAS_BASE (0x6000)
|
|
|
|
#define SAS_RAS_INTR0 (RAS_BASE)
|
|
|
|
#define SAS_RAS_INTR1 (RAS_BASE + 0x04)
|
|
|
|
#define SAS_RAS_INTR0_MASK (RAS_BASE + 0x08)
|
|
|
|
#define SAS_RAS_INTR1_MASK (RAS_BASE + 0x0c)
|
2018-03-23 16:05:09 +00:00
|
|
|
#define CFG_SAS_RAS_INTR_MASK (RAS_BASE + 0x1c)
|
|
|
|
#define SAS_RAS_INTR2 (RAS_BASE + 0x20)
|
|
|
|
#define SAS_RAS_INTR2_MASK (RAS_BASE + 0x24)
|
2017-12-08 17:16:40 +00:00
|
|
|
|
2017-06-14 15:33:25 +00:00
|
|
|
/* HW dma structures */
|
|
|
|
/* Delivery queue header */
|
|
|
|
/* dw0 */
|
2017-06-14 15:33:29 +00:00
|
|
|
#define CMD_HDR_ABORT_FLAG_OFF 0
|
|
|
|
#define CMD_HDR_ABORT_FLAG_MSK (0x3 << CMD_HDR_ABORT_FLAG_OFF)
|
|
|
|
#define CMD_HDR_ABORT_DEVICE_TYPE_OFF 2
|
|
|
|
#define CMD_HDR_ABORT_DEVICE_TYPE_MSK (0x1 << CMD_HDR_ABORT_DEVICE_TYPE_OFF)
|
2017-06-14 15:33:25 +00:00
|
|
|
#define CMD_HDR_RESP_REPORT_OFF 5
|
|
|
|
#define CMD_HDR_RESP_REPORT_MSK (0x1 << CMD_HDR_RESP_REPORT_OFF)
|
|
|
|
#define CMD_HDR_TLR_CTRL_OFF 6
|
|
|
|
#define CMD_HDR_TLR_CTRL_MSK (0x3 << CMD_HDR_TLR_CTRL_OFF)
|
|
|
|
#define CMD_HDR_PORT_OFF 18
|
|
|
|
#define CMD_HDR_PORT_MSK (0xf << CMD_HDR_PORT_OFF)
|
|
|
|
#define CMD_HDR_PRIORITY_OFF 27
|
|
|
|
#define CMD_HDR_PRIORITY_MSK (0x1 << CMD_HDR_PRIORITY_OFF)
|
|
|
|
#define CMD_HDR_CMD_OFF 29
|
|
|
|
#define CMD_HDR_CMD_MSK (0x7 << CMD_HDR_CMD_OFF)
|
|
|
|
/* dw1 */
|
2017-06-14 15:33:27 +00:00
|
|
|
#define CMD_HDR_UNCON_CMD_OFF 3
|
2017-06-14 15:33:25 +00:00
|
|
|
#define CMD_HDR_DIR_OFF 5
|
|
|
|
#define CMD_HDR_DIR_MSK (0x3 << CMD_HDR_DIR_OFF)
|
2017-06-14 15:33:27 +00:00
|
|
|
#define CMD_HDR_RESET_OFF 7
|
|
|
|
#define CMD_HDR_RESET_MSK (0x1 << CMD_HDR_RESET_OFF)
|
2017-06-14 15:33:25 +00:00
|
|
|
#define CMD_HDR_VDTL_OFF 10
|
|
|
|
#define CMD_HDR_VDTL_MSK (0x1 << CMD_HDR_VDTL_OFF)
|
|
|
|
#define CMD_HDR_FRAME_TYPE_OFF 11
|
|
|
|
#define CMD_HDR_FRAME_TYPE_MSK (0x1f << CMD_HDR_FRAME_TYPE_OFF)
|
|
|
|
#define CMD_HDR_DEV_ID_OFF 16
|
|
|
|
#define CMD_HDR_DEV_ID_MSK (0xffff << CMD_HDR_DEV_ID_OFF)
|
|
|
|
/* dw2 */
|
|
|
|
#define CMD_HDR_CFL_OFF 0
|
|
|
|
#define CMD_HDR_CFL_MSK (0x1ff << CMD_HDR_CFL_OFF)
|
2017-06-14 15:33:27 +00:00
|
|
|
#define CMD_HDR_NCQ_TAG_OFF 10
|
|
|
|
#define CMD_HDR_NCQ_TAG_MSK (0x1f << CMD_HDR_NCQ_TAG_OFF)
|
2017-06-14 15:33:25 +00:00
|
|
|
#define CMD_HDR_MRFL_OFF 15
|
|
|
|
#define CMD_HDR_MRFL_MSK (0x1ff << CMD_HDR_MRFL_OFF)
|
|
|
|
#define CMD_HDR_SG_MOD_OFF 24
|
|
|
|
#define CMD_HDR_SG_MOD_MSK (0x3 << CMD_HDR_SG_MOD_OFF)
|
2017-06-14 15:33:26 +00:00
|
|
|
/* dw3 */
|
|
|
|
#define CMD_HDR_IPTT_OFF 0
|
|
|
|
#define CMD_HDR_IPTT_MSK (0xffff << CMD_HDR_IPTT_OFF)
|
2017-06-14 15:33:25 +00:00
|
|
|
/* dw6 */
|
|
|
|
#define CMD_HDR_DIF_SGL_LEN_OFF 0
|
|
|
|
#define CMD_HDR_DIF_SGL_LEN_MSK (0xffff << CMD_HDR_DIF_SGL_LEN_OFF)
|
|
|
|
#define CMD_HDR_DATA_SGL_LEN_OFF 16
|
|
|
|
#define CMD_HDR_DATA_SGL_LEN_MSK (0xffff << CMD_HDR_DATA_SGL_LEN_OFF)
|
2017-06-14 15:33:29 +00:00
|
|
|
/* dw7 */
|
|
|
|
#define CMD_HDR_ADDR_MODE_SEL_OFF 15
|
|
|
|
#define CMD_HDR_ADDR_MODE_SEL_MSK (1 << CMD_HDR_ADDR_MODE_SEL_OFF)
|
|
|
|
#define CMD_HDR_ABORT_IPTT_OFF 16
|
|
|
|
#define CMD_HDR_ABORT_IPTT_MSK (0xffff << CMD_HDR_ABORT_IPTT_OFF)
|
2017-06-14 15:33:25 +00:00
|
|
|
|
2017-06-14 15:33:24 +00:00
|
|
|
/* Completion header */
|
|
|
|
/* dw0 */
|
|
|
|
#define CMPLT_HDR_CMPLT_OFF 0
|
|
|
|
#define CMPLT_HDR_CMPLT_MSK (0x3 << CMPLT_HDR_CMPLT_OFF)
|
|
|
|
#define CMPLT_HDR_ERROR_PHASE_OFF 2
|
|
|
|
#define CMPLT_HDR_ERROR_PHASE_MSK (0xff << CMPLT_HDR_ERROR_PHASE_OFF)
|
|
|
|
#define CMPLT_HDR_RSPNS_XFRD_OFF 10
|
|
|
|
#define CMPLT_HDR_RSPNS_XFRD_MSK (0x1 << CMPLT_HDR_RSPNS_XFRD_OFF)
|
|
|
|
#define CMPLT_HDR_ERX_OFF 12
|
|
|
|
#define CMPLT_HDR_ERX_MSK (0x1 << CMPLT_HDR_ERX_OFF)
|
|
|
|
#define CMPLT_HDR_ABORT_STAT_OFF 13
|
|
|
|
#define CMPLT_HDR_ABORT_STAT_MSK (0x7 << CMPLT_HDR_ABORT_STAT_OFF)
|
|
|
|
/* abort_stat */
|
|
|
|
#define STAT_IO_NOT_VALID 0x1
|
|
|
|
#define STAT_IO_NO_DEVICE 0x2
|
|
|
|
#define STAT_IO_COMPLETE 0x3
|
|
|
|
#define STAT_IO_ABORTED 0x4
|
|
|
|
/* dw1 */
|
|
|
|
#define CMPLT_HDR_IPTT_OFF 0
|
|
|
|
#define CMPLT_HDR_IPTT_MSK (0xffff << CMPLT_HDR_IPTT_OFF)
|
|
|
|
#define CMPLT_HDR_DEV_ID_OFF 16
|
|
|
|
#define CMPLT_HDR_DEV_ID_MSK (0xffff << CMPLT_HDR_DEV_ID_OFF)
|
|
|
|
/* dw3 */
|
|
|
|
#define CMPLT_HDR_IO_IN_TARGET_OFF 17
|
|
|
|
#define CMPLT_HDR_IO_IN_TARGET_MSK (0x1 << CMPLT_HDR_IO_IN_TARGET_OFF)
|
|
|
|
|
2017-06-14 15:33:28 +00:00
|
|
|
/* ITCT header */
|
|
|
|
/* qw0 */
|
|
|
|
#define ITCT_HDR_DEV_TYPE_OFF 0
|
|
|
|
#define ITCT_HDR_DEV_TYPE_MSK (0x3 << ITCT_HDR_DEV_TYPE_OFF)
|
|
|
|
#define ITCT_HDR_VALID_OFF 2
|
|
|
|
#define ITCT_HDR_VALID_MSK (0x1 << ITCT_HDR_VALID_OFF)
|
|
|
|
#define ITCT_HDR_MCR_OFF 5
|
|
|
|
#define ITCT_HDR_MCR_MSK (0xf << ITCT_HDR_MCR_OFF)
|
|
|
|
#define ITCT_HDR_VLN_OFF 9
|
|
|
|
#define ITCT_HDR_VLN_MSK (0xf << ITCT_HDR_VLN_OFF)
|
|
|
|
#define ITCT_HDR_SMP_TIMEOUT_OFF 16
|
|
|
|
#define ITCT_HDR_AWT_CONTINUE_OFF 25
|
|
|
|
#define ITCT_HDR_PORT_ID_OFF 28
|
|
|
|
#define ITCT_HDR_PORT_ID_MSK (0xf << ITCT_HDR_PORT_ID_OFF)
|
|
|
|
/* qw2 */
|
|
|
|
#define ITCT_HDR_INLT_OFF 0
|
|
|
|
#define ITCT_HDR_INLT_MSK (0xffffULL << ITCT_HDR_INLT_OFF)
|
|
|
|
#define ITCT_HDR_RTOLT_OFF 48
|
|
|
|
#define ITCT_HDR_RTOLT_MSK (0xffffULL << ITCT_HDR_RTOLT_OFF)
|
|
|
|
|
2018-12-17 14:40:07 +00:00
|
|
|
struct hisi_sas_protect_iu_v3_hw {
|
|
|
|
u32 dw0;
|
|
|
|
u32 lbrtcv;
|
|
|
|
u32 lbrtgv;
|
|
|
|
u32 dw3;
|
|
|
|
u32 dw4;
|
|
|
|
u32 dw5;
|
|
|
|
u32 rsv;
|
|
|
|
};
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
struct hisi_sas_complete_v3_hdr {
|
|
|
|
__le32 dw0;
|
|
|
|
__le32 dw1;
|
|
|
|
__le32 act;
|
|
|
|
__le32 dw3;
|
|
|
|
};
|
|
|
|
|
2017-06-14 15:33:24 +00:00
|
|
|
struct hisi_sas_err_record_v3 {
|
|
|
|
/* dw0 */
|
|
|
|
__le32 trans_tx_fail_type;
|
|
|
|
|
|
|
|
/* dw1 */
|
|
|
|
__le32 trans_rx_fail_type;
|
|
|
|
|
|
|
|
/* dw2 */
|
|
|
|
__le16 dma_tx_err_type;
|
|
|
|
__le16 sipc_rx_err_type;
|
|
|
|
|
|
|
|
/* dw3 */
|
|
|
|
__le32 dma_rx_err_type;
|
|
|
|
};
|
|
|
|
|
|
|
|
#define RX_DATA_LEN_UNDERFLOW_OFF 6
|
|
|
|
#define RX_DATA_LEN_UNDERFLOW_MSK (1 << RX_DATA_LEN_UNDERFLOW_OFF)
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
#define HISI_SAS_COMMAND_ENTRIES_V3_HW 4096
|
2017-06-14 15:33:22 +00:00
|
|
|
#define HISI_SAS_MSI_COUNT_V3_HW 32
|
|
|
|
|
2017-06-14 15:33:25 +00:00
|
|
|
#define DIR_NO_DATA 0
|
|
|
|
#define DIR_TO_INI 1
|
|
|
|
#define DIR_TO_DEVICE 2
|
|
|
|
#define DIR_RESERVED 3
|
|
|
|
|
2018-05-02 15:56:31 +00:00
|
|
|
#define FIS_CMD_IS_UNCONSTRAINED(fis) \
|
|
|
|
((fis.command == ATA_CMD_READ_LOG_EXT) || \
|
|
|
|
(fis.command == ATA_CMD_READ_LOG_DMA_EXT) || \
|
|
|
|
((fis.command == ATA_CMD_DEV_RESET) && \
|
|
|
|
((fis.control & ATA_SRST) != 0)))
|
2017-06-14 15:33:27 +00:00
|
|
|
|
2018-12-17 14:40:07 +00:00
|
|
|
#define T10_INSRT_EN_OFF 0
|
|
|
|
#define T10_INSRT_EN_MSK (1 << T10_INSRT_EN_OFF)
|
|
|
|
#define T10_RMV_EN_OFF 1
|
|
|
|
#define T10_RMV_EN_MSK (1 << T10_RMV_EN_OFF)
|
|
|
|
#define T10_RPLC_EN_OFF 2
|
|
|
|
#define T10_RPLC_EN_MSK (1 << T10_RPLC_EN_OFF)
|
|
|
|
#define T10_CHK_EN_OFF 3
|
|
|
|
#define T10_CHK_EN_MSK (1 << T10_CHK_EN_OFF)
|
|
|
|
#define INCR_LBRT_OFF 5
|
|
|
|
#define INCR_LBRT_MSK (1 << INCR_LBRT_OFF)
|
|
|
|
#define USR_DATA_BLOCK_SZ_OFF 20
|
|
|
|
#define USR_DATA_BLOCK_SZ_MSK (0x3 << USR_DATA_BLOCK_SZ_OFF)
|
|
|
|
#define T10_CHK_MSK_OFF 16
|
2019-02-06 10:52:51 +00:00
|
|
|
#define T10_CHK_REF_TAG_MSK (0xf0 << T10_CHK_MSK_OFF)
|
|
|
|
#define T10_CHK_APP_TAG_MSK (0xc << T10_CHK_MSK_OFF)
|
2018-12-17 14:40:07 +00:00
|
|
|
|
scsi: hisi_sas: Issue internal abort on all relevant queues
To support queue mapped to a CPU, it needs to be ensured that issuing an
internal abort is safe, in that it is guaranteed that an internal abort is
processed for a single IO or a device after all the relevant command(s)
which it is attempting to abort have been processed by the controller.
Currently we only deliver commands for any device on a single queue to
solve this problem, as we know that commands issued on the same queue will
be processed in order, and we will not have a scenario where the internal
abort is racing against a command(s) which it is trying to abort.
To enqueue commands on queue mapped to a CPU, choosing a queue for an
command is based on the associated queue for the current CPU, so this is
not safe for internal abort since it would definitely not be guaranteed
that commands for the command devices are issued on the same queue.
To solve this issue, we take a bludgeoning approach, and issue a separate
internal abort on any queue(s) relevant to the command or device, in that
we will be guaranteed that at least one of these internal aborts will be
received last in the controller.
So, for aborting a single command, we can just force the internal abort to
be issued on the same queue as the command which we are trying to abort.
For aborting all commands associated with a device, we issue a separate
internal abort on all relevant queues. Issuing multiple internal aborts in
this fashion would have not side affect.
Signed-off-by: John Garry <john.garry@huawei.com>
Signed-off-by: Xiang Chen <chenxiang66@hisilicon.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2019-02-06 10:52:54 +00:00
|
|
|
#define BASE_VECTORS_V3_HW 16
|
2019-02-06 10:52:55 +00:00
|
|
|
#define MIN_AFFINE_VECTORS_V3_HW (BASE_VECTORS_V3_HW + 1)
|
2018-12-17 14:40:07 +00:00
|
|
|
|
2020-01-20 12:22:33 +00:00
|
|
|
#define CHNL_INT_STS_MSK 0xeeeeeeee
|
|
|
|
#define CHNL_INT_STS_PHY_MSK 0xe
|
|
|
|
#define CHNL_INT_STS_INT0_MSK BIT(1)
|
|
|
|
#define CHNL_INT_STS_INT1_MSK BIT(2)
|
|
|
|
#define CHNL_INT_STS_INT2_MSK BIT(3)
|
|
|
|
#define CHNL_WIDTH 4
|
|
|
|
|
2021-08-24 10:00:56 +00:00
|
|
|
#define BAR_NO_V3_HW 5
|
|
|
|
|
2019-04-11 12:46:41 +00:00
|
|
|
enum {
|
|
|
|
DSM_FUNC_ERR_HANDLE_MSI = 0,
|
|
|
|
};
|
|
|
|
|
2018-11-09 14:06:33 +00:00
|
|
|
static bool hisi_sas_intr_conv;
|
|
|
|
MODULE_PARM_DESC(intr_conv, "interrupt converge enable (0-1)");
|
|
|
|
|
2018-12-17 14:40:07 +00:00
|
|
|
/* permit overriding the host protection capabilities mask (EEDP/T10 PI) */
|
|
|
|
static int prot_mask;
|
|
|
|
module_param(prot_mask, int, 0);
|
|
|
|
MODULE_PARM_DESC(prot_mask, " host protection capabilities mask, def=0x0 ");
|
|
|
|
|
2020-11-24 08:46:34 +00:00
|
|
|
static void debugfs_work_handler_v3_hw(struct work_struct *work);
|
2021-04-06 11:48:29 +00:00
|
|
|
static void debugfs_snapshot_regs_v3_hw(struct hisi_hba *hisi_hba);
|
2020-11-24 08:46:34 +00:00
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
static u32 hisi_sas_read32(struct hisi_hba *hisi_hba, u32 off)
|
|
|
|
{
|
|
|
|
void __iomem *regs = hisi_hba->regs + off;
|
|
|
|
|
|
|
|
return readl(regs);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
static void hisi_sas_write32(struct hisi_hba *hisi_hba, u32 off, u32 val)
|
|
|
|
{
|
|
|
|
void __iomem *regs = hisi_hba->regs + off;
|
|
|
|
|
|
|
|
writel(val, regs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hisi_sas_phy_write32(struct hisi_hba *hisi_hba, int phy_no,
|
|
|
|
u32 off, u32 val)
|
|
|
|
{
|
|
|
|
void __iomem *regs = hisi_hba->regs + (0x400 * phy_no) + off;
|
|
|
|
|
|
|
|
writel(val, regs);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:22 +00:00
|
|
|
static u32 hisi_sas_phy_read32(struct hisi_hba *hisi_hba,
|
|
|
|
int phy_no, u32 off)
|
|
|
|
{
|
|
|
|
void __iomem *regs = hisi_hba->regs + (0x400 * phy_no) + off;
|
|
|
|
|
|
|
|
return readl(regs);
|
|
|
|
}
|
|
|
|
|
2018-05-02 15:56:33 +00:00
|
|
|
#define hisi_sas_read32_poll_timeout(off, val, cond, delay_us, \
|
|
|
|
timeout_us) \
|
|
|
|
({ \
|
|
|
|
void __iomem *regs = hisi_hba->regs + off; \
|
|
|
|
readl_poll_timeout(regs, val, cond, delay_us, timeout_us); \
|
|
|
|
})
|
|
|
|
|
|
|
|
#define hisi_sas_read32_poll_timeout_atomic(off, val, cond, delay_us, \
|
|
|
|
timeout_us) \
|
|
|
|
({ \
|
|
|
|
void __iomem *regs = hisi_hba->regs + off; \
|
|
|
|
readl_poll_timeout_atomic(regs, val, cond, delay_us, timeout_us);\
|
|
|
|
})
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
static void init_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
2020-09-01 11:13:07 +00:00
|
|
|
int i, j;
|
2017-06-14 15:33:21 +00:00
|
|
|
|
|
|
|
/* Global registers init */
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_QUEUE_ENABLE,
|
|
|
|
(u32)((1ULL << hisi_hba->queue_count) - 1));
|
2019-05-29 09:58:47 +00:00
|
|
|
hisi_sas_write32(hisi_hba, SAS_AXI_USER3, 0);
|
2017-10-24 15:51:35 +00:00
|
|
|
hisi_sas_write32(hisi_hba, CFG_MAX_TAG, 0xfff0400);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_write32(hisi_hba, HGC_SAS_TXFAIL_RETRY_CTRL, 0x108);
|
2018-09-24 15:06:34 +00:00
|
|
|
hisi_sas_write32(hisi_hba, CFG_AGING_TIME, 0x1);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_write32(hisi_hba, INT_COAL_EN, 0x1);
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_COAL_TIME, 0x1);
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_COAL_CNT, 0x1);
|
2018-11-09 14:06:33 +00:00
|
|
|
hisi_sas_write32(hisi_hba, CQ_INT_CONVERGE_EN,
|
|
|
|
hisi_sas_intr_conv);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_SRC, 0xffff);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC1, 0xffffffff);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC2, 0xffffffff);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC3, 0xffffffff);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK1, 0xfefefefe);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK2, 0xfefefefe);
|
2019-04-11 12:46:41 +00:00
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK3, 0xffc220ff);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_write32(hisi_hba, CHNL_PHYUPDOWN_INT_MSK, 0x0);
|
|
|
|
hisi_sas_write32(hisi_hba, CHNL_ENT_INT_MSK, 0x0);
|
|
|
|
hisi_sas_write32(hisi_hba, HGC_COM_INT_MSK, 0x0);
|
2019-04-11 12:46:41 +00:00
|
|
|
hisi_sas_write32(hisi_hba, SAS_ECC_INTR_MSK, 0x155555);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_write32(hisi_hba, AWQOS_AWCACHE_CFG, 0xf0f0);
|
|
|
|
hisi_sas_write32(hisi_hba, ARQOS_ARCACHE_CFG, 0xf0f0);
|
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++)
|
2020-09-01 11:13:10 +00:00
|
|
|
hisi_sas_write32(hisi_hba, OQ0_INT_SRC_MSK + 0x4 * i, 0);
|
2017-06-14 15:33:21 +00:00
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, HYPER_STREAM_ID_EN_CFG, 1);
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->n_phy; i++) {
|
2020-09-01 11:13:05 +00:00
|
|
|
enum sas_linkrate max;
|
2018-05-02 15:56:30 +00:00
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[i];
|
|
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
2020-09-01 11:13:05 +00:00
|
|
|
u32 prog_phy_link_rate = hisi_sas_phy_read32(hisi_hba, i,
|
|
|
|
PROG_PHY_LINK_RATE);
|
2018-05-02 15:56:30 +00:00
|
|
|
|
2020-09-01 11:13:05 +00:00
|
|
|
prog_phy_link_rate &= ~CFG_PROG_PHY_LINK_RATE_MSK;
|
2018-05-02 15:56:30 +00:00
|
|
|
if (!sas_phy->phy || (sas_phy->phy->maximum_linkrate <
|
2020-09-01 11:13:05 +00:00
|
|
|
SAS_LINK_RATE_1_5_GBPS))
|
|
|
|
max = SAS_LINK_RATE_12_0_GBPS;
|
|
|
|
else
|
|
|
|
max = sas_phy->phy->maximum_linkrate;
|
|
|
|
prog_phy_link_rate |= hisi_sas_get_prog_phy_linkrate_mask(max);
|
2018-05-02 15:56:30 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PROG_PHY_LINK_RATE,
|
|
|
|
prog_phy_link_rate);
|
2019-02-28 14:51:02 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, SERDES_CFG, 0xffc00);
|
2018-03-07 12:25:13 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, SAS_RX_TRAIN_TIMER, 0x13e80);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT0, 0xffffffff);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT1, 0xffffffff);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT2, 0xffffffff);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, RXOP_CHECK_CFG_H, 0x1000);
|
2019-04-11 12:46:41 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT1_MSK, 0xf2057fff);
|
2017-12-08 17:16:45 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT2_MSK, 0xffffbfe);
|
2017-06-14 15:33:21 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHY_CTRL_RDY_MSK, 0x0);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHYCTRL_NOT_RDY_MSK, 0x0);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHYCTRL_DWS_RESET_MSK, 0x0);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHYCTRL_PHY_ENA_MSK, 0x0);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, SL_RX_BCAST_CHK_MSK, 0x0);
|
2018-03-07 12:25:13 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHYCTRL_OOB_RESTART_MSK, 0x1);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, STP_LINK_TIMER, 0x7f7a120);
|
2018-05-31 12:50:50 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CON_CFG_DRIVER, 0x2a0a01);
|
2018-11-09 14:06:36 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, SAS_SSP_CON_TIMER_CFG, 0x32);
|
2019-02-28 14:50:59 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, SAS_EC_INT_COAL_TIME,
|
|
|
|
0x30f4240);
|
2018-03-07 12:25:13 +00:00
|
|
|
/* used for 12G negotiate */
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, COARSETUNE_TIME, 0x1e);
|
2018-09-24 15:06:34 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, i, AIP_LIMIT, 0x2ffff);
|
2020-09-01 11:13:07 +00:00
|
|
|
|
|
|
|
/* get default FFE configuration for BIST */
|
|
|
|
for (j = 0; j < FFE_CFG_MAX; j++) {
|
|
|
|
u32 val = hisi_sas_phy_read32(hisi_hba, i,
|
|
|
|
TXDEEMPH_G1 + (j * 0x4));
|
|
|
|
hisi_hba->debugfs_bist_ffe[i][j] = val;
|
|
|
|
}
|
2017-06-14 15:33:21 +00:00
|
|
|
}
|
2018-03-07 12:25:13 +00:00
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++) {
|
|
|
|
/* Delivery queue */
|
|
|
|
hisi_sas_write32(hisi_hba,
|
|
|
|
DLVRY_Q_0_BASE_ADDR_HI + (i * 0x14),
|
|
|
|
upper_32_bits(hisi_hba->cmd_hdr_dma[i]));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_Q_0_BASE_ADDR_LO + (i * 0x14),
|
|
|
|
lower_32_bits(hisi_hba->cmd_hdr_dma[i]));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_Q_0_DEPTH + (i * 0x14),
|
|
|
|
HISI_SAS_QUEUE_SLOTS);
|
|
|
|
|
|
|
|
/* Completion queue */
|
|
|
|
hisi_sas_write32(hisi_hba, COMPL_Q_0_BASE_ADDR_HI + (i * 0x14),
|
|
|
|
upper_32_bits(hisi_hba->complete_hdr_dma[i]));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, COMPL_Q_0_BASE_ADDR_LO + (i * 0x14),
|
|
|
|
lower_32_bits(hisi_hba->complete_hdr_dma[i]));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, COMPL_Q_0_DEPTH + (i * 0x14),
|
|
|
|
HISI_SAS_QUEUE_SLOTS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* itct */
|
|
|
|
hisi_sas_write32(hisi_hba, ITCT_BASE_ADDR_LO,
|
|
|
|
lower_32_bits(hisi_hba->itct_dma));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, ITCT_BASE_ADDR_HI,
|
|
|
|
upper_32_bits(hisi_hba->itct_dma));
|
|
|
|
|
|
|
|
/* iost */
|
|
|
|
hisi_sas_write32(hisi_hba, IOST_BASE_ADDR_LO,
|
|
|
|
lower_32_bits(hisi_hba->iost_dma));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, IOST_BASE_ADDR_HI,
|
|
|
|
upper_32_bits(hisi_hba->iost_dma));
|
|
|
|
|
|
|
|
/* breakpoint */
|
|
|
|
hisi_sas_write32(hisi_hba, IO_BROKEN_MSG_ADDR_LO,
|
|
|
|
lower_32_bits(hisi_hba->breakpoint_dma));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, IO_BROKEN_MSG_ADDR_HI,
|
|
|
|
upper_32_bits(hisi_hba->breakpoint_dma));
|
|
|
|
|
|
|
|
/* SATA broken msg */
|
|
|
|
hisi_sas_write32(hisi_hba, IO_SATA_BROKEN_MSG_ADDR_LO,
|
|
|
|
lower_32_bits(hisi_hba->sata_breakpoint_dma));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, IO_SATA_BROKEN_MSG_ADDR_HI,
|
|
|
|
upper_32_bits(hisi_hba->sata_breakpoint_dma));
|
|
|
|
|
|
|
|
/* SATA initial fis */
|
|
|
|
hisi_sas_write32(hisi_hba, SATA_INITI_D2H_STORE_ADDR_LO,
|
|
|
|
lower_32_bits(hisi_hba->initial_fis_dma));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, SATA_INITI_D2H_STORE_ADDR_HI,
|
|
|
|
upper_32_bits(hisi_hba->initial_fis_dma));
|
2017-12-08 17:16:40 +00:00
|
|
|
|
|
|
|
/* RAS registers init */
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_RAS_INTR0_MASK, 0x0);
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_RAS_INTR1_MASK, 0x0);
|
2018-03-23 16:05:09 +00:00
|
|
|
hisi_sas_write32(hisi_hba, SAS_RAS_INTR2_MASK, 0x0);
|
|
|
|
hisi_sas_write32(hisi_hba, CFG_SAS_RAS_INTR_MASK, 0x0);
|
2018-05-21 10:09:16 +00:00
|
|
|
|
|
|
|
/* LED registers init */
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_CFG_DRIVE_VLD, 0x80000ff);
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_GPIO_TX_0_1, 0x80808080);
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_GPIO_TX_0_1 + 0x4, 0x80808080);
|
|
|
|
/* Configure blink generator rate A to 1Hz and B to 4Hz */
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_GPIO_CFG_1, 0x121700);
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_GPIO_CFG_0, 0x800000);
|
2017-06-14 15:33:21 +00:00
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:22 +00:00
|
|
|
static void config_phy_opt_mode_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
u32 cfg = hisi_sas_phy_read32(hisi_hba, phy_no, PHY_CFG);
|
|
|
|
|
|
|
|
cfg &= ~PHY_CFG_DC_OPT_MSK;
|
|
|
|
cfg |= 1 << PHY_CFG_DC_OPT_OFF;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHY_CFG, cfg);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void config_id_frame_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
struct sas_identify_frame identify_frame;
|
|
|
|
u32 *identify_buffer;
|
|
|
|
|
|
|
|
memset(&identify_frame, 0, sizeof(identify_frame));
|
|
|
|
identify_frame.dev_type = SAS_END_DEVICE;
|
|
|
|
identify_frame.frame_type = 0;
|
|
|
|
identify_frame._un1 = 1;
|
|
|
|
identify_frame.initiator_bits = SAS_PROTOCOL_ALL;
|
|
|
|
identify_frame.target_bits = SAS_PROTOCOL_NONE;
|
|
|
|
memcpy(&identify_frame._un4_11[0], hisi_hba->sas_addr, SAS_ADDR_SIZE);
|
|
|
|
memcpy(&identify_frame.sas_addr[0], hisi_hba->sas_addr, SAS_ADDR_SIZE);
|
|
|
|
identify_frame.phy_id = phy_no;
|
|
|
|
identify_buffer = (u32 *)(&identify_frame);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TX_ID_DWORD0,
|
|
|
|
__swab32(identify_buffer[0]));
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TX_ID_DWORD1,
|
|
|
|
__swab32(identify_buffer[1]));
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TX_ID_DWORD2,
|
|
|
|
__swab32(identify_buffer[2]));
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TX_ID_DWORD3,
|
|
|
|
__swab32(identify_buffer[3]));
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TX_ID_DWORD4,
|
|
|
|
__swab32(identify_buffer[4]));
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TX_ID_DWORD5,
|
|
|
|
__swab32(identify_buffer[5]));
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:28 +00:00
|
|
|
static void setup_itct_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
struct hisi_sas_device *sas_dev)
|
|
|
|
{
|
|
|
|
struct domain_device *device = sas_dev->sas_device;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
u64 qw0, device_id = sas_dev->device_id;
|
|
|
|
struct hisi_sas_itct *itct = &hisi_hba->itct[device_id];
|
|
|
|
struct domain_device *parent_dev = device->parent;
|
|
|
|
struct asd_sas_port *sas_port = device->port;
|
|
|
|
struct hisi_sas_port *port = to_hisi_sas_port(sas_port);
|
2018-12-06 13:34:40 +00:00
|
|
|
u64 sas_addr;
|
2017-06-14 15:33:28 +00:00
|
|
|
|
|
|
|
memset(itct, 0, sizeof(*itct));
|
|
|
|
|
|
|
|
/* qw0 */
|
|
|
|
qw0 = 0;
|
|
|
|
switch (sas_dev->dev_type) {
|
|
|
|
case SAS_END_DEVICE:
|
|
|
|
case SAS_EDGE_EXPANDER_DEVICE:
|
|
|
|
case SAS_FANOUT_EXPANDER_DEVICE:
|
|
|
|
qw0 = HISI_SAS_DEV_TYPE_SSP << ITCT_HDR_DEV_TYPE_OFF;
|
|
|
|
break;
|
|
|
|
case SAS_SATA_DEV:
|
|
|
|
case SAS_SATA_PENDING:
|
2019-06-10 12:41:41 +00:00
|
|
|
if (parent_dev && dev_is_expander(parent_dev->dev_type))
|
2017-06-14 15:33:28 +00:00
|
|
|
qw0 = HISI_SAS_DEV_TYPE_STP << ITCT_HDR_DEV_TYPE_OFF;
|
|
|
|
else
|
|
|
|
qw0 = HISI_SAS_DEV_TYPE_SATA << ITCT_HDR_DEV_TYPE_OFF;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_warn(dev, "setup itct: unsupported dev type (%d)\n",
|
|
|
|
sas_dev->dev_type);
|
|
|
|
}
|
|
|
|
|
|
|
|
qw0 |= ((1 << ITCT_HDR_VALID_OFF) |
|
|
|
|
(device->linkrate << ITCT_HDR_MCR_OFF) |
|
|
|
|
(1 << ITCT_HDR_VLN_OFF) |
|
|
|
|
(0xfa << ITCT_HDR_SMP_TIMEOUT_OFF) |
|
|
|
|
(1 << ITCT_HDR_AWT_CONTINUE_OFF) |
|
|
|
|
(port->id << ITCT_HDR_PORT_ID_OFF));
|
|
|
|
itct->qw0 = cpu_to_le64(qw0);
|
|
|
|
|
|
|
|
/* qw1 */
|
2018-12-06 13:34:40 +00:00
|
|
|
memcpy(&sas_addr, device->sas_addr, SAS_ADDR_SIZE);
|
|
|
|
itct->sas_addr = cpu_to_le64(__swab64(sas_addr));
|
2017-06-14 15:33:28 +00:00
|
|
|
|
|
|
|
/* qw2 */
|
|
|
|
if (!dev_is_sata(device))
|
|
|
|
itct->qw2 = cpu_to_le64((5000ULL << ITCT_HDR_INLT_OFF) |
|
|
|
|
(0x1ULL << ITCT_HDR_RTOLT_OFF));
|
|
|
|
}
|
|
|
|
|
2019-10-24 14:08:10 +00:00
|
|
|
static int clear_itct_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
struct hisi_sas_device *sas_dev)
|
2017-06-14 15:33:28 +00:00
|
|
|
{
|
2017-10-24 15:51:46 +00:00
|
|
|
DECLARE_COMPLETION_ONSTACK(completion);
|
2017-06-14 15:33:28 +00:00
|
|
|
u64 dev_id = sas_dev->device_id;
|
|
|
|
struct hisi_sas_itct *itct = &hisi_hba->itct[dev_id];
|
|
|
|
u32 reg_val = hisi_sas_read32(hisi_hba, ENT_INT_SRC3);
|
2019-10-24 14:08:10 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
2017-06-14 15:33:28 +00:00
|
|
|
|
2017-10-24 15:51:46 +00:00
|
|
|
sas_dev->completion = &completion;
|
|
|
|
|
2017-06-14 15:33:28 +00:00
|
|
|
/* clear the itct interrupt state */
|
|
|
|
if (ENT_INT_SRC3_ITC_INT_MSK & reg_val)
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC3,
|
|
|
|
ENT_INT_SRC3_ITC_INT_MSK);
|
|
|
|
|
2019-02-06 10:52:56 +00:00
|
|
|
/* clear the itct table */
|
2017-10-24 15:51:46 +00:00
|
|
|
reg_val = ITCT_CLR_EN_MSK | (dev_id & ITCT_DEV_MSK);
|
2017-06-14 15:33:28 +00:00
|
|
|
hisi_sas_write32(hisi_hba, ITCT_CLR, reg_val);
|
|
|
|
|
2019-10-24 14:08:10 +00:00
|
|
|
if (!wait_for_completion_timeout(sas_dev->completion,
|
2021-06-07 09:29:37 +00:00
|
|
|
HISI_SAS_CLEAR_ITCT_TIMEOUT)) {
|
2019-10-24 14:08:10 +00:00
|
|
|
dev_warn(dev, "failed to clear ITCT\n");
|
|
|
|
return -ETIMEDOUT;
|
|
|
|
}
|
|
|
|
|
2017-10-24 15:51:46 +00:00
|
|
|
memset(itct, 0, sizeof(struct hisi_sas_itct));
|
2019-10-24 14:08:10 +00:00
|
|
|
return 0;
|
2017-06-14 15:33:28 +00:00
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:32 +00:00
|
|
|
static void dereg_device_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
struct domain_device *device)
|
|
|
|
{
|
|
|
|
struct hisi_sas_slot *slot, *slot2;
|
|
|
|
struct hisi_sas_device *sas_dev = device->lldd_dev;
|
|
|
|
u32 cfg_abt_set_query_iptt;
|
|
|
|
|
|
|
|
cfg_abt_set_query_iptt = hisi_sas_read32(hisi_hba,
|
|
|
|
CFG_ABT_SET_QUERY_IPTT);
|
|
|
|
list_for_each_entry_safe(slot, slot2, &sas_dev->list, entry) {
|
|
|
|
cfg_abt_set_query_iptt &= ~CFG_SET_ABORTED_IPTT_MSK;
|
|
|
|
cfg_abt_set_query_iptt |= (1 << CFG_SET_ABORTED_EN_OFF) |
|
|
|
|
(slot->idx << CFG_SET_ABORTED_IPTT_OFF);
|
|
|
|
hisi_sas_write32(hisi_hba, CFG_ABT_SET_QUERY_IPTT,
|
|
|
|
cfg_abt_set_query_iptt);
|
|
|
|
}
|
|
|
|
cfg_abt_set_query_iptt &= ~(1 << CFG_SET_ABORTED_EN_OFF);
|
|
|
|
hisi_sas_write32(hisi_hba, CFG_ABT_SET_QUERY_IPTT,
|
|
|
|
cfg_abt_set_query_iptt);
|
|
|
|
hisi_sas_write32(hisi_hba, CFG_ABT_SET_IPTT_DONE,
|
|
|
|
1 << CFG_ABT_SET_IPTT_DONE_OFF);
|
|
|
|
}
|
|
|
|
|
2017-08-10 16:09:40 +00:00
|
|
|
static int reset_hw_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int ret;
|
|
|
|
u32 val;
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_QUEUE_ENABLE, 0);
|
|
|
|
|
|
|
|
/* Disable all of the PHYs */
|
|
|
|
hisi_sas_stop_phys(hisi_hba);
|
|
|
|
udelay(50);
|
|
|
|
|
|
|
|
/* Ensure axi bus idle */
|
2018-05-02 15:56:33 +00:00
|
|
|
ret = hisi_sas_read32_poll_timeout(AXI_CFG, val, !val,
|
|
|
|
20000, 1000000);
|
2017-08-10 16:09:40 +00:00
|
|
|
if (ret) {
|
|
|
|
dev_err(dev, "axi bus is not idle, ret = %d!\n", ret);
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ACPI_HANDLE(dev)) {
|
|
|
|
acpi_status s;
|
|
|
|
|
|
|
|
s = acpi_evaluate_object(ACPI_HANDLE(dev), "_RST", NULL, NULL);
|
|
|
|
if (ACPI_FAILURE(s)) {
|
|
|
|
dev_err(dev, "Reset failed\n");
|
|
|
|
return -EIO;
|
|
|
|
}
|
2018-03-07 12:25:12 +00:00
|
|
|
} else {
|
2017-08-10 16:09:40 +00:00
|
|
|
dev_err(dev, "no reset method!\n");
|
2018-03-07 12:25:12 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
2017-08-10 16:09:40 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
static int hw_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
2017-08-10 16:09:40 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
2020-10-02 14:30:35 +00:00
|
|
|
struct acpi_device *acpi_dev;
|
2019-04-11 12:46:41 +00:00
|
|
|
union acpi_object *obj;
|
|
|
|
guid_t guid;
|
2017-08-10 16:09:40 +00:00
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = reset_hw_v3_hw(hisi_hba);
|
|
|
|
if (rc) {
|
2020-09-01 11:13:09 +00:00
|
|
|
dev_err(dev, "hisi_sas_reset_hw failed, rc=%d\n", rc);
|
2017-08-10 16:09:40 +00:00
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
msleep(100);
|
2017-06-14 15:33:21 +00:00
|
|
|
init_reg_v3_hw(hisi_hba);
|
|
|
|
|
2019-04-11 12:46:41 +00:00
|
|
|
if (guid_parse("D5918B4B-37AE-4E10-A99F-E5E8A6EF4C1F", &guid)) {
|
|
|
|
dev_err(dev, "Parse GUID failed\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2020-05-15 14:13:43 +00:00
|
|
|
/*
|
|
|
|
* This DSM handles some hardware-related configurations:
|
|
|
|
* 1. Switch over to MSI error handling in kernel
|
|
|
|
* 2. BIOS *may* reset some register values through this method
|
|
|
|
*/
|
2019-04-11 12:46:41 +00:00
|
|
|
obj = acpi_evaluate_dsm(ACPI_HANDLE(dev), &guid, 0,
|
|
|
|
DSM_FUNC_ERR_HANDLE_MSI, NULL);
|
|
|
|
if (!obj)
|
2020-05-15 14:13:43 +00:00
|
|
|
dev_warn(dev, "can not find DSM method, ignore\n");
|
2019-04-11 12:46:41 +00:00
|
|
|
else
|
|
|
|
ACPI_FREE(obj);
|
|
|
|
|
2020-10-02 14:30:35 +00:00
|
|
|
acpi_dev = ACPI_COMPANION(dev);
|
|
|
|
if (!acpi_device_power_manageable(acpi_dev))
|
|
|
|
dev_notice(dev, "neither _PS0 nor _PR0 is defined\n");
|
2017-06-14 15:33:21 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:22 +00:00
|
|
|
static void enable_phy_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
u32 cfg = hisi_sas_phy_read32(hisi_hba, phy_no, PHY_CFG);
|
|
|
|
|
|
|
|
cfg |= PHY_CFG_ENA_MSK;
|
2018-07-18 14:14:27 +00:00
|
|
|
cfg &= ~PHY_CFG_PHY_RST_MSK;
|
2017-06-14 15:33:22 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHY_CFG, cfg);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:31 +00:00
|
|
|
static void disable_phy_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
u32 cfg = hisi_sas_phy_read32(hisi_hba, phy_no, PHY_CFG);
|
2019-05-29 09:58:46 +00:00
|
|
|
u32 irq_msk = hisi_sas_phy_read32(hisi_hba, phy_no, CHL_INT2_MSK);
|
|
|
|
static const u32 msk = BIT(CHL_INT2_RX_DISP_ERR_OFF) |
|
|
|
|
BIT(CHL_INT2_RX_CODE_ERR_OFF) |
|
|
|
|
BIT(CHL_INT2_RX_INVLD_DW_OFF);
|
2018-07-18 14:14:27 +00:00
|
|
|
u32 state;
|
2017-06-14 15:33:31 +00:00
|
|
|
|
2019-05-29 09:58:46 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT2_MSK, msk | irq_msk);
|
|
|
|
|
2017-06-14 15:33:31 +00:00
|
|
|
cfg &= ~PHY_CFG_ENA_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHY_CFG, cfg);
|
2018-07-18 14:14:27 +00:00
|
|
|
|
|
|
|
mdelay(50);
|
|
|
|
|
|
|
|
state = hisi_sas_read32(hisi_hba, PHY_STATE);
|
|
|
|
if (state & BIT(phy_no)) {
|
|
|
|
cfg |= PHY_CFG_PHY_RST_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHY_CFG, cfg);
|
|
|
|
}
|
2019-05-29 09:58:46 +00:00
|
|
|
|
|
|
|
udelay(1);
|
|
|
|
|
|
|
|
hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_INVLD_DW);
|
|
|
|
hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_DISP_ERR);
|
|
|
|
hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_CODE_ERR);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT2, msk);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT2_MSK, irq_msk);
|
2017-06-14 15:33:31 +00:00
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:22 +00:00
|
|
|
static void start_phy_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
config_id_frame_v3_hw(hisi_hba, phy_no);
|
|
|
|
config_phy_opt_mode_v3_hw(hisi_hba, phy_no);
|
|
|
|
enable_phy_v3_hw(hisi_hba, phy_no);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:31 +00:00
|
|
|
static void phy_hard_reset_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
|
|
|
u32 txid_auto;
|
|
|
|
|
2019-04-11 12:46:38 +00:00
|
|
|
hisi_sas_phy_enable(hisi_hba, phy_no, 0);
|
2017-06-14 15:33:31 +00:00
|
|
|
if (phy->identify.device_type == SAS_END_DEVICE) {
|
|
|
|
txid_auto = hisi_sas_phy_read32(hisi_hba, phy_no, TXID_AUTO);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TXID_AUTO,
|
|
|
|
txid_auto | TX_HARDRST_MSK);
|
|
|
|
}
|
|
|
|
msleep(100);
|
2019-04-11 12:46:38 +00:00
|
|
|
hisi_sas_phy_enable(hisi_hba, phy_no, 1);
|
2017-06-14 15:33:31 +00:00
|
|
|
}
|
|
|
|
|
2018-03-07 12:25:12 +00:00
|
|
|
static enum sas_linkrate phy_get_max_linkrate_v3_hw(void)
|
2017-06-14 15:33:31 +00:00
|
|
|
{
|
|
|
|
return SAS_LINK_RATE_12_0_GBPS;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:22 +00:00
|
|
|
static void phys_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
2017-08-10 16:09:40 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->n_phy; i++) {
|
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[i];
|
|
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
|
|
|
|
if (!sas_phy->phy->enabled)
|
|
|
|
continue;
|
|
|
|
|
2019-04-11 12:46:38 +00:00
|
|
|
hisi_sas_phy_enable(hisi_hba, i, 1);
|
2017-08-10 16:09:40 +00:00
|
|
|
}
|
2017-06-14 15:33:22 +00:00
|
|
|
}
|
|
|
|
|
2019-01-25 14:22:30 +00:00
|
|
|
static void sl_notify_ssp_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
2017-06-14 15:33:22 +00:00
|
|
|
{
|
|
|
|
u32 sl_control;
|
|
|
|
|
|
|
|
sl_control = hisi_sas_phy_read32(hisi_hba, phy_no, SL_CONTROL);
|
|
|
|
sl_control |= SL_CONTROL_NOTIFY_EN_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SL_CONTROL, sl_control);
|
|
|
|
msleep(1);
|
|
|
|
sl_control = hisi_sas_phy_read32(hisi_hba, phy_no, SL_CONTROL);
|
|
|
|
sl_control &= ~SL_CONTROL_NOTIFY_EN_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SL_CONTROL, sl_control);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:30 +00:00
|
|
|
static int get_wideport_bitmap_v3_hw(struct hisi_hba *hisi_hba, int port_id)
|
|
|
|
{
|
|
|
|
int i, bitmap = 0;
|
|
|
|
u32 phy_port_num_ma = hisi_sas_read32(hisi_hba, PHY_PORT_NUM_MA);
|
2017-10-24 15:51:42 +00:00
|
|
|
u32 phy_state = hisi_sas_read32(hisi_hba, PHY_STATE);
|
2017-06-14 15:33:30 +00:00
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->n_phy; i++)
|
2017-10-24 15:51:42 +00:00
|
|
|
if (phy_state & BIT(i))
|
|
|
|
if (((phy_port_num_ma >> (i * 4)) & 0xf) == port_id)
|
|
|
|
bitmap |= BIT(i);
|
2017-06-14 15:33:30 +00:00
|
|
|
|
|
|
|
return bitmap;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:25 +00:00
|
|
|
static void start_delivery_v3_hw(struct hisi_sas_dq *dq)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = dq->hisi_hba;
|
2018-07-18 14:14:32 +00:00
|
|
|
struct hisi_sas_slot *s, *s1, *s2 = NULL;
|
2018-05-09 15:10:48 +00:00
|
|
|
int dlvry_queue = dq->id;
|
2018-07-18 14:14:32 +00:00
|
|
|
int wp;
|
2018-05-09 15:10:48 +00:00
|
|
|
|
|
|
|
list_for_each_entry_safe(s, s1, &dq->list, delivery) {
|
|
|
|
if (!s->ready)
|
|
|
|
break;
|
2018-07-18 14:14:32 +00:00
|
|
|
s2 = s;
|
2018-05-09 15:10:48 +00:00
|
|
|
list_del(&s->delivery);
|
|
|
|
}
|
|
|
|
|
2018-07-18 14:14:32 +00:00
|
|
|
if (!s2)
|
2018-05-09 15:10:48 +00:00
|
|
|
return;
|
2017-06-14 15:33:25 +00:00
|
|
|
|
2018-07-18 14:14:32 +00:00
|
|
|
/*
|
|
|
|
* Ensure that memories for slots built on other CPUs is observed.
|
|
|
|
*/
|
|
|
|
smp_rmb();
|
|
|
|
wp = (s2->dlvry_queue_slot + 1) % HISI_SAS_QUEUE_SLOTS;
|
|
|
|
|
2018-05-09 15:10:48 +00:00
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_Q_0_WR_PTR + (dlvry_queue * 0x14), wp);
|
2017-06-14 15:33:25 +00:00
|
|
|
}
|
|
|
|
|
2018-05-09 15:10:46 +00:00
|
|
|
static void prep_prd_sge_v3_hw(struct hisi_hba *hisi_hba,
|
2017-06-14 15:33:25 +00:00
|
|
|
struct hisi_sas_slot *slot,
|
|
|
|
struct hisi_sas_cmd_hdr *hdr,
|
|
|
|
struct scatterlist *scatter,
|
|
|
|
int n_elem)
|
|
|
|
{
|
2017-06-29 13:02:14 +00:00
|
|
|
struct hisi_sas_sge_page *sge_page = hisi_sas_sge_addr_mem(slot);
|
2017-06-14 15:33:25 +00:00
|
|
|
struct scatterlist *sg;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for_each_sg(scatter, sg, n_elem, i) {
|
2017-06-29 13:02:14 +00:00
|
|
|
struct hisi_sas_sge *entry = &sge_page->sge[i];
|
2017-06-14 15:33:25 +00:00
|
|
|
|
|
|
|
entry->addr = cpu_to_le64(sg_dma_address(sg));
|
|
|
|
entry->page_ctrl_0 = entry->page_ctrl_1 = 0;
|
|
|
|
entry->data_len = cpu_to_le32(sg_dma_len(sg));
|
|
|
|
entry->data_off = 0;
|
|
|
|
}
|
|
|
|
|
2017-06-29 13:02:14 +00:00
|
|
|
hdr->prd_table_addr = cpu_to_le64(hisi_sas_sge_addr_dma(slot));
|
|
|
|
|
2019-02-06 10:52:51 +00:00
|
|
|
hdr->sg_len |= cpu_to_le32(n_elem << CMD_HDR_DATA_SGL_LEN_OFF);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void prep_prd_sge_dif_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
struct hisi_sas_slot *slot,
|
|
|
|
struct hisi_sas_cmd_hdr *hdr,
|
|
|
|
struct scatterlist *scatter,
|
|
|
|
int n_elem)
|
|
|
|
{
|
|
|
|
struct hisi_sas_sge_dif_page *sge_dif_page;
|
|
|
|
struct scatterlist *sg;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
sge_dif_page = hisi_sas_sge_dif_addr_mem(slot);
|
|
|
|
|
|
|
|
for_each_sg(scatter, sg, n_elem, i) {
|
|
|
|
struct hisi_sas_sge *entry = &sge_dif_page->sge[i];
|
|
|
|
|
|
|
|
entry->addr = cpu_to_le64(sg_dma_address(sg));
|
|
|
|
entry->page_ctrl_0 = 0;
|
|
|
|
entry->page_ctrl_1 = 0;
|
|
|
|
entry->data_len = cpu_to_le32(sg_dma_len(sg));
|
|
|
|
entry->data_off = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
hdr->dif_prd_table_addr =
|
|
|
|
cpu_to_le64(hisi_sas_sge_dif_addr_dma(slot));
|
|
|
|
|
|
|
|
hdr->sg_len |= cpu_to_le32(n_elem << CMD_HDR_DIF_SGL_LEN_OFF);
|
2017-06-14 15:33:25 +00:00
|
|
|
}
|
|
|
|
|
2018-12-17 14:40:07 +00:00
|
|
|
static u32 get_prot_chk_msk_v3_hw(struct scsi_cmnd *scsi_cmnd)
|
|
|
|
{
|
|
|
|
unsigned char prot_flags = scsi_cmnd->prot_flags;
|
|
|
|
|
2019-02-06 10:52:51 +00:00
|
|
|
if (prot_flags & SCSI_PROT_REF_CHECK)
|
|
|
|
return T10_CHK_APP_TAG_MSK;
|
|
|
|
return T10_CHK_REF_TAG_MSK | T10_CHK_APP_TAG_MSK;
|
2018-12-17 14:40:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void fill_prot_v3_hw(struct scsi_cmnd *scsi_cmnd,
|
|
|
|
struct hisi_sas_protect_iu_v3_hw *prot)
|
|
|
|
{
|
|
|
|
unsigned char prot_op = scsi_get_prot_op(scsi_cmnd);
|
|
|
|
unsigned int interval = scsi_prot_interval(scsi_cmnd);
|
2021-08-09 23:03:24 +00:00
|
|
|
u32 lbrt_chk_val = t10_pi_ref_tag(scsi_cmd_to_rq(scsi_cmnd));
|
2018-12-17 14:40:07 +00:00
|
|
|
|
|
|
|
switch (prot_op) {
|
2019-02-06 10:52:51 +00:00
|
|
|
case SCSI_PROT_READ_INSERT:
|
|
|
|
prot->dw0 |= T10_INSRT_EN_MSK;
|
|
|
|
prot->lbrtgv = lbrt_chk_val;
|
|
|
|
break;
|
2018-12-17 14:40:07 +00:00
|
|
|
case SCSI_PROT_READ_STRIP:
|
|
|
|
prot->dw0 |= (T10_RMV_EN_MSK | T10_CHK_EN_MSK);
|
|
|
|
prot->lbrtcv = lbrt_chk_val;
|
|
|
|
prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd);
|
|
|
|
break;
|
2019-02-06 10:52:51 +00:00
|
|
|
case SCSI_PROT_READ_PASS:
|
|
|
|
prot->dw0 |= T10_CHK_EN_MSK;
|
|
|
|
prot->lbrtcv = lbrt_chk_val;
|
|
|
|
prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd);
|
|
|
|
break;
|
2018-12-17 14:40:07 +00:00
|
|
|
case SCSI_PROT_WRITE_INSERT:
|
|
|
|
prot->dw0 |= T10_INSRT_EN_MSK;
|
|
|
|
prot->lbrtgv = lbrt_chk_val;
|
|
|
|
break;
|
2019-02-06 10:52:51 +00:00
|
|
|
case SCSI_PROT_WRITE_STRIP:
|
|
|
|
prot->dw0 |= (T10_RMV_EN_MSK | T10_CHK_EN_MSK);
|
|
|
|
prot->lbrtcv = lbrt_chk_val;
|
|
|
|
break;
|
|
|
|
case SCSI_PROT_WRITE_PASS:
|
|
|
|
prot->dw0 |= T10_CHK_EN_MSK;
|
|
|
|
prot->lbrtcv = lbrt_chk_val;
|
|
|
|
prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd);
|
|
|
|
break;
|
2018-12-17 14:40:07 +00:00
|
|
|
default:
|
|
|
|
WARN(1, "prot_op(0x%x) is not valid\n", prot_op);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (interval) {
|
|
|
|
case 512:
|
|
|
|
break;
|
|
|
|
case 4096:
|
|
|
|
prot->dw0 |= (0x1 << USR_DATA_BLOCK_SZ_OFF);
|
|
|
|
break;
|
|
|
|
case 520:
|
|
|
|
prot->dw0 |= (0x2 << USR_DATA_BLOCK_SZ_OFF);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
WARN(1, "protection interval (0x%x) invalid\n",
|
|
|
|
interval);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
prot->dw0 |= INCR_LBRT_MSK;
|
|
|
|
}
|
|
|
|
|
2018-05-09 15:10:46 +00:00
|
|
|
static void prep_ssp_v3_hw(struct hisi_hba *hisi_hba,
|
2018-05-21 10:09:21 +00:00
|
|
|
struct hisi_sas_slot *slot)
|
2017-06-14 15:33:25 +00:00
|
|
|
{
|
|
|
|
struct sas_task *task = slot->task;
|
|
|
|
struct hisi_sas_cmd_hdr *hdr = slot->cmd_hdr;
|
|
|
|
struct domain_device *device = task->dev;
|
|
|
|
struct hisi_sas_device *sas_dev = device->lldd_dev;
|
|
|
|
struct hisi_sas_port *port = slot->port;
|
|
|
|
struct sas_ssp_task *ssp_task = &task->ssp_task;
|
|
|
|
struct scsi_cmnd *scsi_cmnd = ssp_task->cmd;
|
2022-02-17 15:42:35 +00:00
|
|
|
struct sas_tmf_task *tmf = slot->tmf;
|
2018-05-21 10:09:21 +00:00
|
|
|
int has_data = 0, priority = !!tmf;
|
2019-01-25 14:22:38 +00:00
|
|
|
unsigned char prot_op;
|
2017-06-14 15:33:25 +00:00
|
|
|
u8 *buf_cmd;
|
2018-12-17 14:40:07 +00:00
|
|
|
u32 dw1 = 0, dw2 = 0, len = 0;
|
2017-06-14 15:33:25 +00:00
|
|
|
|
|
|
|
hdr->dw0 = cpu_to_le32((1 << CMD_HDR_RESP_REPORT_OFF) |
|
|
|
|
(2 << CMD_HDR_TLR_CTRL_OFF) |
|
|
|
|
(port->id << CMD_HDR_PORT_OFF) |
|
|
|
|
(priority << CMD_HDR_PRIORITY_OFF) |
|
|
|
|
(1 << CMD_HDR_CMD_OFF)); /* ssp */
|
|
|
|
|
|
|
|
dw1 = 1 << CMD_HDR_VDTL_OFF;
|
2018-05-21 10:09:21 +00:00
|
|
|
if (tmf) {
|
2017-06-14 15:33:25 +00:00
|
|
|
dw1 |= 2 << CMD_HDR_FRAME_TYPE_OFF;
|
|
|
|
dw1 |= DIR_NO_DATA << CMD_HDR_DIR_OFF;
|
|
|
|
} else {
|
2019-01-25 14:22:38 +00:00
|
|
|
prot_op = scsi_get_prot_op(scsi_cmnd);
|
2017-06-14 15:33:25 +00:00
|
|
|
dw1 |= 1 << CMD_HDR_FRAME_TYPE_OFF;
|
|
|
|
switch (scsi_cmnd->sc_data_direction) {
|
|
|
|
case DMA_TO_DEVICE:
|
|
|
|
has_data = 1;
|
|
|
|
dw1 |= DIR_TO_DEVICE << CMD_HDR_DIR_OFF;
|
|
|
|
break;
|
|
|
|
case DMA_FROM_DEVICE:
|
|
|
|
has_data = 1;
|
|
|
|
dw1 |= DIR_TO_INI << CMD_HDR_DIR_OFF;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dw1 &= ~CMD_HDR_DIR_MSK;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* map itct entry */
|
|
|
|
dw1 |= sas_dev->device_id << CMD_HDR_DEV_ID_OFF;
|
|
|
|
|
|
|
|
dw2 = (((sizeof(struct ssp_command_iu) + sizeof(struct ssp_frame_hdr)
|
|
|
|
+ 3) / 4) << CMD_HDR_CFL_OFF) |
|
|
|
|
((HISI_SAS_MAX_SSP_RESP_SZ / 4) << CMD_HDR_MRFL_OFF) |
|
|
|
|
(2 << CMD_HDR_SG_MOD_OFF);
|
|
|
|
hdr->dw2 = cpu_to_le32(dw2);
|
|
|
|
hdr->transfer_tags = cpu_to_le32(slot->idx);
|
|
|
|
|
2019-02-06 10:52:51 +00:00
|
|
|
if (has_data) {
|
2018-05-09 15:10:46 +00:00
|
|
|
prep_prd_sge_v3_hw(hisi_hba, slot, hdr, task->scatter,
|
2019-02-06 10:52:51 +00:00
|
|
|
slot->n_elem);
|
|
|
|
|
|
|
|
if (scsi_prot_sg_count(scsi_cmnd))
|
|
|
|
prep_prd_sge_dif_v3_hw(hisi_hba, slot, hdr,
|
|
|
|
scsi_prot_sglist(scsi_cmnd),
|
|
|
|
slot->n_elem_dif);
|
|
|
|
}
|
2017-06-14 15:33:25 +00:00
|
|
|
|
2017-06-29 13:02:14 +00:00
|
|
|
hdr->cmd_table_addr = cpu_to_le64(hisi_sas_cmd_hdr_addr_dma(slot));
|
|
|
|
hdr->sts_buffer_addr = cpu_to_le64(hisi_sas_status_buf_addr_dma(slot));
|
2017-06-14 15:33:25 +00:00
|
|
|
|
2017-06-29 13:02:14 +00:00
|
|
|
buf_cmd = hisi_sas_cmd_hdr_addr_mem(slot) +
|
|
|
|
sizeof(struct ssp_frame_hdr);
|
2017-06-14 15:33:25 +00:00
|
|
|
|
2017-06-29 13:02:14 +00:00
|
|
|
memcpy(buf_cmd, &task->ssp_task.LUN, 8);
|
2018-05-21 10:09:21 +00:00
|
|
|
if (!tmf) {
|
2017-06-14 15:33:25 +00:00
|
|
|
buf_cmd[9] = ssp_task->task_attr | (ssp_task->task_prio << 3);
|
|
|
|
memcpy(buf_cmd + 12, scsi_cmnd->cmnd, scsi_cmnd->cmd_len);
|
|
|
|
} else {
|
|
|
|
buf_cmd[10] = tmf->tmf;
|
|
|
|
switch (tmf->tmf) {
|
|
|
|
case TMF_ABORT_TASK:
|
|
|
|
case TMF_QUERY_TASK:
|
|
|
|
buf_cmd[12] =
|
|
|
|
(tmf->tag_of_task_to_be_managed >> 8) & 0xff;
|
|
|
|
buf_cmd[13] =
|
|
|
|
tmf->tag_of_task_to_be_managed & 0xff;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2018-12-17 14:40:07 +00:00
|
|
|
|
|
|
|
if (has_data && (prot_op != SCSI_PROT_NORMAL)) {
|
|
|
|
struct hisi_sas_protect_iu_v3_hw prot;
|
|
|
|
u8 *buf_cmd_prot;
|
|
|
|
|
|
|
|
hdr->dw7 |= cpu_to_le32(1 << CMD_HDR_ADDR_MODE_SEL_OFF);
|
|
|
|
dw1 |= CMD_HDR_PIR_MSK;
|
|
|
|
buf_cmd_prot = hisi_sas_cmd_hdr_addr_mem(slot) +
|
|
|
|
sizeof(struct ssp_frame_hdr) +
|
|
|
|
sizeof(struct ssp_command_iu);
|
|
|
|
|
|
|
|
memset(&prot, 0, sizeof(struct hisi_sas_protect_iu_v3_hw));
|
|
|
|
fill_prot_v3_hw(scsi_cmnd, &prot);
|
|
|
|
memcpy(buf_cmd_prot, &prot,
|
|
|
|
sizeof(struct hisi_sas_protect_iu_v3_hw));
|
|
|
|
/*
|
|
|
|
* For READ, we need length of info read to memory, while for
|
|
|
|
* WRITE we need length of data written to the disk.
|
|
|
|
*/
|
2019-02-06 10:52:51 +00:00
|
|
|
if (prot_op == SCSI_PROT_WRITE_INSERT ||
|
|
|
|
prot_op == SCSI_PROT_READ_INSERT ||
|
|
|
|
prot_op == SCSI_PROT_WRITE_PASS ||
|
|
|
|
prot_op == SCSI_PROT_READ_PASS) {
|
2018-12-17 14:40:07 +00:00
|
|
|
unsigned int interval = scsi_prot_interval(scsi_cmnd);
|
|
|
|
unsigned int ilog2_interval = ilog2(interval);
|
|
|
|
|
|
|
|
len = (task->total_xfer_len >> ilog2_interval) * 8;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
hdr->dw1 = cpu_to_le32(dw1);
|
|
|
|
|
|
|
|
hdr->data_transfer_len = cpu_to_le32(task->total_xfer_len + len);
|
2017-06-14 15:33:25 +00:00
|
|
|
}
|
|
|
|
|
2018-05-09 15:10:46 +00:00
|
|
|
static void prep_smp_v3_hw(struct hisi_hba *hisi_hba,
|
2017-06-14 15:33:26 +00:00
|
|
|
struct hisi_sas_slot *slot)
|
|
|
|
{
|
|
|
|
struct sas_task *task = slot->task;
|
|
|
|
struct hisi_sas_cmd_hdr *hdr = slot->cmd_hdr;
|
|
|
|
struct domain_device *device = task->dev;
|
|
|
|
struct hisi_sas_port *port = slot->port;
|
2018-05-09 15:10:45 +00:00
|
|
|
struct scatterlist *sg_req;
|
2017-06-14 15:33:26 +00:00
|
|
|
struct hisi_sas_device *sas_dev = device->lldd_dev;
|
|
|
|
dma_addr_t req_dma_addr;
|
2018-05-09 15:10:45 +00:00
|
|
|
unsigned int req_len;
|
2017-06-14 15:33:26 +00:00
|
|
|
|
|
|
|
/* req */
|
|
|
|
sg_req = &task->smp_task.smp_req;
|
|
|
|
req_len = sg_dma_len(sg_req);
|
|
|
|
req_dma_addr = sg_dma_address(sg_req);
|
|
|
|
|
|
|
|
/* create header */
|
|
|
|
/* dw0 */
|
|
|
|
hdr->dw0 = cpu_to_le32((port->id << CMD_HDR_PORT_OFF) |
|
|
|
|
(1 << CMD_HDR_PRIORITY_OFF) | /* high pri */
|
|
|
|
(2 << CMD_HDR_CMD_OFF)); /* smp */
|
|
|
|
|
|
|
|
/* map itct entry */
|
|
|
|
hdr->dw1 = cpu_to_le32((sas_dev->device_id << CMD_HDR_DEV_ID_OFF) |
|
|
|
|
(1 << CMD_HDR_FRAME_TYPE_OFF) |
|
|
|
|
(DIR_NO_DATA << CMD_HDR_DIR_OFF));
|
|
|
|
|
|
|
|
/* dw2 */
|
|
|
|
hdr->dw2 = cpu_to_le32((((req_len - 4) / 4) << CMD_HDR_CFL_OFF) |
|
|
|
|
(HISI_SAS_MAX_SMP_RESP_SZ / 4 <<
|
|
|
|
CMD_HDR_MRFL_OFF));
|
|
|
|
|
|
|
|
hdr->transfer_tags = cpu_to_le32(slot->idx << CMD_HDR_IPTT_OFF);
|
|
|
|
|
|
|
|
hdr->cmd_table_addr = cpu_to_le64(req_dma_addr);
|
2017-06-29 13:02:14 +00:00
|
|
|
hdr->sts_buffer_addr = cpu_to_le64(hisi_sas_status_buf_addr_dma(slot));
|
2017-06-14 15:33:26 +00:00
|
|
|
}
|
|
|
|
|
2018-05-09 15:10:46 +00:00
|
|
|
static void prep_ata_v3_hw(struct hisi_hba *hisi_hba,
|
2017-06-14 15:33:27 +00:00
|
|
|
struct hisi_sas_slot *slot)
|
|
|
|
{
|
|
|
|
struct sas_task *task = slot->task;
|
|
|
|
struct domain_device *device = task->dev;
|
|
|
|
struct domain_device *parent_dev = device->parent;
|
|
|
|
struct hisi_sas_device *sas_dev = device->lldd_dev;
|
|
|
|
struct hisi_sas_cmd_hdr *hdr = slot->cmd_hdr;
|
|
|
|
struct asd_sas_port *sas_port = device->port;
|
|
|
|
struct hisi_sas_port *port = to_hisi_sas_port(sas_port);
|
|
|
|
u8 *buf_cmd;
|
2018-05-09 15:10:46 +00:00
|
|
|
int has_data = 0, hdr_tag = 0;
|
2017-06-14 15:33:27 +00:00
|
|
|
u32 dw1 = 0, dw2 = 0;
|
|
|
|
|
|
|
|
hdr->dw0 = cpu_to_le32(port->id << CMD_HDR_PORT_OFF);
|
2019-06-10 12:41:41 +00:00
|
|
|
if (parent_dev && dev_is_expander(parent_dev->dev_type))
|
2017-06-14 15:33:27 +00:00
|
|
|
hdr->dw0 |= cpu_to_le32(3 << CMD_HDR_CMD_OFF);
|
|
|
|
else
|
2019-05-29 09:58:45 +00:00
|
|
|
hdr->dw0 |= cpu_to_le32(4U << CMD_HDR_CMD_OFF);
|
2017-06-14 15:33:27 +00:00
|
|
|
|
|
|
|
switch (task->data_dir) {
|
|
|
|
case DMA_TO_DEVICE:
|
|
|
|
has_data = 1;
|
|
|
|
dw1 |= DIR_TO_DEVICE << CMD_HDR_DIR_OFF;
|
|
|
|
break;
|
|
|
|
case DMA_FROM_DEVICE:
|
|
|
|
has_data = 1;
|
|
|
|
dw1 |= DIR_TO_INI << CMD_HDR_DIR_OFF;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dw1 &= ~CMD_HDR_DIR_MSK;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((task->ata_task.fis.command == ATA_CMD_DEV_RESET) &&
|
|
|
|
(task->ata_task.fis.control & ATA_SRST))
|
|
|
|
dw1 |= 1 << CMD_HDR_RESET_OFF;
|
|
|
|
|
|
|
|
dw1 |= (hisi_sas_get_ata_protocol(
|
2017-12-28 10:20:47 +00:00
|
|
|
&task->ata_task.fis, task->data_dir))
|
2017-06-14 15:33:27 +00:00
|
|
|
<< CMD_HDR_FRAME_TYPE_OFF;
|
|
|
|
dw1 |= sas_dev->device_id << CMD_HDR_DEV_ID_OFF;
|
|
|
|
|
2018-05-02 15:56:31 +00:00
|
|
|
if (FIS_CMD_IS_UNCONSTRAINED(task->ata_task.fis))
|
2017-06-14 15:33:27 +00:00
|
|
|
dw1 |= 1 << CMD_HDR_UNCON_CMD_OFF;
|
|
|
|
|
|
|
|
hdr->dw1 = cpu_to_le32(dw1);
|
|
|
|
|
|
|
|
/* dw2 */
|
2019-09-06 12:55:31 +00:00
|
|
|
if (task->ata_task.use_ncq) {
|
|
|
|
struct ata_queued_cmd *qc = task->uldd_task;
|
|
|
|
|
|
|
|
hdr_tag = qc->tag;
|
2017-06-14 15:33:27 +00:00
|
|
|
task->ata_task.fis.sector_count |= (u8) (hdr_tag << 3);
|
|
|
|
dw2 |= hdr_tag << CMD_HDR_NCQ_TAG_OFF;
|
|
|
|
}
|
|
|
|
|
|
|
|
dw2 |= (HISI_SAS_MAX_STP_RESP_SZ / 4) << CMD_HDR_CFL_OFF |
|
|
|
|
2 << CMD_HDR_SG_MOD_OFF;
|
|
|
|
hdr->dw2 = cpu_to_le32(dw2);
|
|
|
|
|
|
|
|
/* dw3 */
|
|
|
|
hdr->transfer_tags = cpu_to_le32(slot->idx);
|
|
|
|
|
2018-05-09 15:10:46 +00:00
|
|
|
if (has_data)
|
|
|
|
prep_prd_sge_v3_hw(hisi_hba, slot, hdr, task->scatter,
|
2017-06-14 15:33:27 +00:00
|
|
|
slot->n_elem);
|
|
|
|
|
|
|
|
hdr->data_transfer_len = cpu_to_le32(task->total_xfer_len);
|
2017-06-29 13:02:14 +00:00
|
|
|
hdr->cmd_table_addr = cpu_to_le64(hisi_sas_cmd_hdr_addr_dma(slot));
|
|
|
|
hdr->sts_buffer_addr = cpu_to_le64(hisi_sas_status_buf_addr_dma(slot));
|
2017-06-14 15:33:27 +00:00
|
|
|
|
2017-06-29 13:02:14 +00:00
|
|
|
buf_cmd = hisi_sas_cmd_hdr_addr_mem(slot);
|
2017-06-14 15:33:27 +00:00
|
|
|
|
|
|
|
if (likely(!task->ata_task.device_control_reg_update))
|
|
|
|
task->ata_task.fis.flags |= 0x80; /* C=1: update ATA cmd reg */
|
|
|
|
/* fill in command FIS */
|
|
|
|
memcpy(buf_cmd, &task->ata_task.fis, sizeof(struct host_to_dev_fis));
|
|
|
|
}
|
|
|
|
|
2018-05-09 15:10:46 +00:00
|
|
|
static void prep_abort_v3_hw(struct hisi_hba *hisi_hba,
|
2017-06-14 15:33:29 +00:00
|
|
|
struct hisi_sas_slot *slot,
|
|
|
|
int device_id, int abort_flag, int tag_to_abort)
|
|
|
|
{
|
|
|
|
struct sas_task *task = slot->task;
|
|
|
|
struct domain_device *dev = task->dev;
|
|
|
|
struct hisi_sas_cmd_hdr *hdr = slot->cmd_hdr;
|
|
|
|
struct hisi_sas_port *port = slot->port;
|
|
|
|
|
|
|
|
/* dw0 */
|
2019-05-29 09:58:45 +00:00
|
|
|
hdr->dw0 = cpu_to_le32((5U << CMD_HDR_CMD_OFF) | /*abort*/
|
2017-06-14 15:33:29 +00:00
|
|
|
(port->id << CMD_HDR_PORT_OFF) |
|
2018-03-07 12:25:12 +00:00
|
|
|
(dev_is_sata(dev)
|
2017-06-14 15:33:29 +00:00
|
|
|
<< CMD_HDR_ABORT_DEVICE_TYPE_OFF) |
|
|
|
|
(abort_flag
|
|
|
|
<< CMD_HDR_ABORT_FLAG_OFF));
|
|
|
|
|
|
|
|
/* dw1 */
|
|
|
|
hdr->dw1 = cpu_to_le32(device_id
|
|
|
|
<< CMD_HDR_DEV_ID_OFF);
|
|
|
|
|
|
|
|
/* dw7 */
|
|
|
|
hdr->dw7 = cpu_to_le32(tag_to_abort << CMD_HDR_ABORT_IPTT_OFF);
|
|
|
|
hdr->transfer_tags = cpu_to_le32(slot->idx);
|
|
|
|
}
|
|
|
|
|
2018-03-07 12:25:12 +00:00
|
|
|
static irqreturn_t phy_up_v3_hw(int phy_no, struct hisi_hba *hisi_hba)
|
2017-06-14 15:33:23 +00:00
|
|
|
{
|
2019-02-28 14:50:57 +00:00
|
|
|
int i;
|
|
|
|
irqreturn_t res;
|
2018-03-07 12:25:07 +00:00
|
|
|
u32 context, port_id, link_rate;
|
2017-06-14 15:33:23 +00:00
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
|
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHYCTRL_PHY_ENA_MSK, 1);
|
|
|
|
|
|
|
|
port_id = hisi_sas_read32(hisi_hba, PHY_PORT_NUM_MA);
|
|
|
|
port_id = (port_id >> (4 * phy_no)) & 0xf;
|
|
|
|
link_rate = hisi_sas_read32(hisi_hba, PHY_CONN_RATE);
|
|
|
|
link_rate = (link_rate >> (phy_no * 4)) & 0xf;
|
|
|
|
|
|
|
|
if (port_id == 0xf) {
|
|
|
|
dev_err(dev, "phyup: phy%d invalid portid\n", phy_no);
|
|
|
|
res = IRQ_NONE;
|
|
|
|
goto end;
|
|
|
|
}
|
|
|
|
sas_phy->linkrate = link_rate;
|
|
|
|
phy->phy_type &= ~(PORT_TYPE_SAS | PORT_TYPE_SATA);
|
|
|
|
|
|
|
|
/* Check for SATA dev */
|
|
|
|
context = hisi_sas_read32(hisi_hba, PHY_CONTEXT);
|
|
|
|
if (context & (1 << phy_no)) {
|
|
|
|
struct hisi_sas_initial_fis *initial_fis;
|
|
|
|
struct dev_to_host_fis *fis;
|
|
|
|
u8 attached_sas_addr[SAS_ADDR_SIZE] = {0};
|
2019-04-11 12:46:40 +00:00
|
|
|
struct Scsi_Host *shost = hisi_hba->shost;
|
2017-06-14 15:33:23 +00:00
|
|
|
|
2017-12-08 17:16:41 +00:00
|
|
|
dev_info(dev, "phyup: phy%d link_rate=%d(sata)\n", phy_no, link_rate);
|
2017-06-14 15:33:23 +00:00
|
|
|
initial_fis = &hisi_hba->initial_fis[phy_no];
|
|
|
|
fis = &initial_fis->fis;
|
2018-07-18 14:14:33 +00:00
|
|
|
|
|
|
|
/* check ERR bit of Status Register */
|
|
|
|
if (fis->status & ATA_ERR) {
|
|
|
|
dev_warn(dev, "sata int: phy%d FIS status: 0x%x\n",
|
|
|
|
phy_no, fis->status);
|
|
|
|
hisi_sas_notify_phy_event(phy, HISI_PHYE_LINK_RESET);
|
|
|
|
res = IRQ_NONE;
|
|
|
|
goto end;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
sas_phy->oob_mode = SATA_OOB_MODE;
|
|
|
|
attached_sas_addr[0] = 0x50;
|
2019-04-11 12:46:40 +00:00
|
|
|
attached_sas_addr[6] = shost->host_no;
|
2017-06-14 15:33:23 +00:00
|
|
|
attached_sas_addr[7] = phy_no;
|
|
|
|
memcpy(sas_phy->attached_sas_addr,
|
|
|
|
attached_sas_addr,
|
|
|
|
SAS_ADDR_SIZE);
|
|
|
|
memcpy(sas_phy->frame_rcvd, fis,
|
|
|
|
sizeof(struct dev_to_host_fis));
|
|
|
|
phy->phy_type |= PORT_TYPE_SATA;
|
|
|
|
phy->identify.device_type = SAS_SATA_DEV;
|
|
|
|
phy->frame_rcvd_size = sizeof(struct dev_to_host_fis);
|
|
|
|
phy->identify.target_port_protocols = SAS_PROTOCOL_SATA;
|
|
|
|
} else {
|
|
|
|
u32 *frame_rcvd = (u32 *)sas_phy->frame_rcvd;
|
|
|
|
struct sas_identify_frame *id =
|
|
|
|
(struct sas_identify_frame *)frame_rcvd;
|
|
|
|
|
|
|
|
dev_info(dev, "phyup: phy%d link_rate=%d\n", phy_no, link_rate);
|
|
|
|
for (i = 0; i < 6; i++) {
|
|
|
|
u32 idaf = hisi_sas_phy_read32(hisi_hba, phy_no,
|
|
|
|
RX_IDAF_DWORD0 + (i * 4));
|
|
|
|
frame_rcvd[i] = __swab32(idaf);
|
|
|
|
}
|
|
|
|
sas_phy->oob_mode = SAS_OOB_MODE;
|
|
|
|
memcpy(sas_phy->attached_sas_addr,
|
|
|
|
&id->sas_addr,
|
|
|
|
SAS_ADDR_SIZE);
|
|
|
|
phy->phy_type |= PORT_TYPE_SAS;
|
|
|
|
phy->identify.device_type = id->dev_type;
|
|
|
|
phy->frame_rcvd_size = sizeof(struct sas_identify_frame);
|
|
|
|
if (phy->identify.device_type == SAS_END_DEVICE)
|
|
|
|
phy->identify.target_port_protocols =
|
|
|
|
SAS_PROTOCOL_SSP;
|
|
|
|
else if (phy->identify.device_type != SAS_PHY_UNUSED)
|
|
|
|
phy->identify.target_port_protocols =
|
|
|
|
SAS_PROTOCOL_SMP;
|
|
|
|
}
|
|
|
|
|
|
|
|
phy->port_id = port_id;
|
2021-12-15 14:37:40 +00:00
|
|
|
|
2021-12-20 11:21:36 +00:00
|
|
|
/* Call pm_runtime_put_sync() with pairs in hisi_sas_phyup_pm_work() */
|
|
|
|
pm_runtime_get_noresume(dev);
|
|
|
|
hisi_sas_notify_phy_event(phy, HISI_PHYE_PHY_UP_PM);
|
2021-12-15 14:37:40 +00:00
|
|
|
|
2018-03-07 12:25:12 +00:00
|
|
|
res = IRQ_HANDLED;
|
2021-12-15 14:37:40 +00:00
|
|
|
|
|
|
|
spin_lock(&phy->lock);
|
|
|
|
/* Delete timer and set phy_attached atomically */
|
|
|
|
del_timer(&phy->timer);
|
|
|
|
phy->phy_attached = 1;
|
|
|
|
spin_unlock(&phy->lock);
|
2017-06-14 15:33:23 +00:00
|
|
|
end:
|
2021-10-12 12:26:26 +00:00
|
|
|
if (phy->reset_completion)
|
|
|
|
complete(phy->reset_completion);
|
2017-06-14 15:33:23 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT0,
|
|
|
|
CHL_INT0_SL_PHY_ENABLE_MSK);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHYCTRL_PHY_ENA_MSK, 0);
|
|
|
|
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2018-03-07 12:25:12 +00:00
|
|
|
static irqreturn_t phy_down_v3_hw(int phy_no, struct hisi_hba *hisi_hba)
|
2017-06-14 15:33:23 +00:00
|
|
|
{
|
2019-01-25 14:22:35 +00:00
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
2017-06-14 15:33:23 +00:00
|
|
|
u32 phy_state, sl_ctrl, txid_auto;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
|
2019-10-24 14:08:25 +00:00
|
|
|
atomic_inc(&phy->down_cnt);
|
|
|
|
|
2019-01-25 14:22:35 +00:00
|
|
|
del_timer(&phy->timer);
|
2017-06-14 15:33:23 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHYCTRL_NOT_RDY_MSK, 1);
|
|
|
|
|
|
|
|
phy_state = hisi_sas_read32(hisi_hba, PHY_STATE);
|
|
|
|
dev_info(dev, "phydown: phy%d phy_state=0x%x\n", phy_no, phy_state);
|
2021-01-18 10:09:47 +00:00
|
|
|
hisi_sas_phy_down(hisi_hba, phy_no, (phy_state & 1 << phy_no) ? 1 : 0,
|
|
|
|
GFP_ATOMIC);
|
2017-06-14 15:33:23 +00:00
|
|
|
|
|
|
|
sl_ctrl = hisi_sas_phy_read32(hisi_hba, phy_no, SL_CONTROL);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SL_CONTROL,
|
|
|
|
sl_ctrl&(~SL_CTA_MSK));
|
|
|
|
|
|
|
|
txid_auto = hisi_sas_phy_read32(hisi_hba, phy_no, TXID_AUTO);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TXID_AUTO,
|
|
|
|
txid_auto | CT3_MSK);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT0, CHL_INT0_NOT_RDY_MSK);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PHYCTRL_NOT_RDY_MSK, 0);
|
|
|
|
|
2018-03-07 12:25:12 +00:00
|
|
|
return IRQ_HANDLED;
|
2017-06-14 15:33:23 +00:00
|
|
|
}
|
|
|
|
|
2018-03-07 12:25:12 +00:00
|
|
|
static irqreturn_t phy_bcast_v3_hw(int phy_no, struct hisi_hba *hisi_hba)
|
2017-06-14 15:33:23 +00:00
|
|
|
{
|
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
|
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
2018-05-31 12:50:43 +00:00
|
|
|
u32 bcast_status;
|
2017-06-14 15:33:23 +00:00
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SL_RX_BCAST_CHK_MSK, 1);
|
2018-05-31 12:50:43 +00:00
|
|
|
bcast_status = hisi_sas_phy_read32(hisi_hba, phy_no, RX_PRIMS_STATUS);
|
2018-05-31 12:50:46 +00:00
|
|
|
if ((bcast_status & RX_BCAST_CHG_MSK) &&
|
2021-08-24 10:00:58 +00:00
|
|
|
!test_bit(HISI_SAS_RESETTING_BIT, &hisi_hba->flags))
|
2021-01-18 10:09:49 +00:00
|
|
|
sas_notify_port_event(sas_phy, PORTE_BROADCAST_RCVD,
|
|
|
|
GFP_ATOMIC);
|
2017-06-14 15:33:23 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT0,
|
|
|
|
CHL_INT0_SL_RX_BCST_ACK_MSK);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SL_RX_BCAST_CHK_MSK, 0);
|
2018-03-07 12:25:12 +00:00
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
2017-06-14 15:33:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static irqreturn_t int_phy_up_down_bcast_v3_hw(int irq_no, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = p;
|
|
|
|
u32 irq_msk;
|
|
|
|
int phy_no = 0;
|
|
|
|
irqreturn_t res = IRQ_NONE;
|
|
|
|
|
|
|
|
irq_msk = hisi_sas_read32(hisi_hba, CHNL_INT_STATUS)
|
|
|
|
& 0x11111111;
|
|
|
|
while (irq_msk) {
|
|
|
|
if (irq_msk & 1) {
|
|
|
|
u32 irq_value = hisi_sas_phy_read32(hisi_hba, phy_no,
|
|
|
|
CHL_INT0);
|
|
|
|
u32 phy_state = hisi_sas_read32(hisi_hba, PHY_STATE);
|
|
|
|
int rdy = phy_state & (1 << phy_no);
|
|
|
|
|
|
|
|
if (rdy) {
|
|
|
|
if (irq_value & CHL_INT0_SL_PHY_ENABLE_MSK)
|
|
|
|
/* phy up */
|
|
|
|
if (phy_up_v3_hw(phy_no, hisi_hba)
|
|
|
|
== IRQ_HANDLED)
|
|
|
|
res = IRQ_HANDLED;
|
|
|
|
if (irq_value & CHL_INT0_SL_RX_BCST_ACK_MSK)
|
|
|
|
/* phy bcast */
|
2018-03-07 12:25:12 +00:00
|
|
|
if (phy_bcast_v3_hw(phy_no, hisi_hba)
|
|
|
|
== IRQ_HANDLED)
|
|
|
|
res = IRQ_HANDLED;
|
2017-06-14 15:33:23 +00:00
|
|
|
} else {
|
|
|
|
if (irq_value & CHL_INT0_NOT_RDY_MSK)
|
|
|
|
/* phy down */
|
|
|
|
if (phy_down_v3_hw(phy_no, hisi_hba)
|
|
|
|
== IRQ_HANDLED)
|
|
|
|
res = IRQ_HANDLED;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
irq_msk >>= 4;
|
|
|
|
phy_no++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2017-10-24 15:51:49 +00:00
|
|
|
static const struct hisi_sas_hw_error port_axi_error[] = {
|
2019-04-11 12:46:41 +00:00
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_TX_ECC_MB_ERR_OFF),
|
|
|
|
.msg = "dmac_tx_ecc_bad_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_RX_ECC_MB_ERR_OFF),
|
|
|
|
.msg = "dmac_rx_ecc_bad_err",
|
|
|
|
},
|
2017-10-24 15:51:49 +00:00
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_TX_AXI_WR_ERR_OFF),
|
|
|
|
.msg = "dma_tx_axi_wr_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_TX_AXI_RD_ERR_OFF),
|
|
|
|
.msg = "dma_tx_axi_rd_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_RX_AXI_WR_ERR_OFF),
|
|
|
|
.msg = "dma_rx_axi_wr_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_RX_AXI_RD_ERR_OFF),
|
|
|
|
.msg = "dma_rx_axi_rd_err",
|
|
|
|
},
|
2019-04-11 12:46:41 +00:00
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_TX_FIFO_ERR_OFF),
|
|
|
|
.msg = "dma_tx_fifo_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_RX_FIFO_ERR_OFF),
|
|
|
|
.msg = "dma_rx_fifo_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_TX_AXI_RUSER_ERR_OFF),
|
|
|
|
.msg = "dma_tx_axi_ruser_err",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(CHL_INT1_DMAC_RX_AXI_RUSER_ERR_OFF),
|
|
|
|
.msg = "dma_rx_axi_ruser_err",
|
|
|
|
},
|
2017-10-24 15:51:49 +00:00
|
|
|
};
|
|
|
|
|
2018-07-18 14:14:26 +00:00
|
|
|
static void handle_chl_int1_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
2017-06-14 15:33:23 +00:00
|
|
|
{
|
2018-07-18 14:14:26 +00:00
|
|
|
u32 irq_value = hisi_sas_phy_read32(hisi_hba, phy_no, CHL_INT1);
|
|
|
|
u32 irq_msk = hisi_sas_phy_read32(hisi_hba, phy_no, CHL_INT1_MSK);
|
2017-06-14 15:33:23 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
2018-07-18 14:14:26 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
irq_value &= ~irq_msk;
|
2021-04-06 11:48:30 +00:00
|
|
|
if (!irq_value) {
|
|
|
|
dev_warn(dev, "phy%d channel int 1 received with status bits cleared\n",
|
|
|
|
phy_no);
|
2018-07-18 14:14:26 +00:00
|
|
|
return;
|
2021-04-06 11:48:30 +00:00
|
|
|
}
|
2018-07-18 14:14:26 +00:00
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(port_axi_error); i++) {
|
|
|
|
const struct hisi_sas_hw_error *error = &port_axi_error[i];
|
|
|
|
|
|
|
|
if (!(irq_value & error->irq_msk))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
dev_err(dev, "%s error (phy%d 0x%x) found!\n",
|
|
|
|
error->msg, phy_no, irq_value);
|
|
|
|
queue_work(hisi_hba->wq, &hisi_hba->rst_work);
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT1, irq_value);
|
|
|
|
}
|
|
|
|
|
2019-02-28 14:50:59 +00:00
|
|
|
static void phy_get_events_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
|
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
struct sas_phy *sphy = sas_phy->phy;
|
|
|
|
unsigned long flags;
|
|
|
|
u32 reg_value;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&phy->lock, flags);
|
|
|
|
|
|
|
|
/* loss dword sync */
|
|
|
|
reg_value = hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_DWS_LOST);
|
|
|
|
sphy->loss_of_dword_sync_count += reg_value;
|
|
|
|
|
|
|
|
/* phy reset problem */
|
|
|
|
reg_value = hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_RESET_PROB);
|
|
|
|
sphy->phy_reset_problem_count += reg_value;
|
|
|
|
|
|
|
|
/* invalid dword */
|
|
|
|
reg_value = hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_INVLD_DW);
|
|
|
|
sphy->invalid_dword_count += reg_value;
|
|
|
|
|
|
|
|
/* disparity err */
|
|
|
|
reg_value = hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_DISP_ERR);
|
|
|
|
sphy->running_disparity_error_count += reg_value;
|
|
|
|
|
|
|
|
/* code violation error */
|
|
|
|
reg_value = hisi_sas_phy_read32(hisi_hba, phy_no, ERR_CNT_CODE_ERR);
|
|
|
|
phy->code_violation_err_count += reg_value;
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&phy->lock, flags);
|
|
|
|
}
|
|
|
|
|
2018-07-18 14:14:26 +00:00
|
|
|
static void handle_chl_int2_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
u32 irq_msk = hisi_sas_phy_read32(hisi_hba, phy_no, CHL_INT2_MSK);
|
|
|
|
u32 irq_value = hisi_sas_phy_read32(hisi_hba, phy_no, CHL_INT2);
|
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
2018-05-02 15:56:34 +00:00
|
|
|
struct pci_dev *pci_dev = hisi_hba->pci_dev;
|
2018-07-18 14:14:26 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
2019-02-28 14:50:59 +00:00
|
|
|
static const u32 msk = BIT(CHL_INT2_RX_DISP_ERR_OFF) |
|
|
|
|
BIT(CHL_INT2_RX_CODE_ERR_OFF) |
|
|
|
|
BIT(CHL_INT2_RX_INVLD_DW_OFF);
|
2018-07-18 14:14:26 +00:00
|
|
|
|
|
|
|
irq_value &= ~irq_msk;
|
2021-04-06 11:48:30 +00:00
|
|
|
if (!irq_value) {
|
|
|
|
dev_warn(dev, "phy%d channel int 2 received with status bits cleared\n",
|
|
|
|
phy_no);
|
2018-07-18 14:14:26 +00:00
|
|
|
return;
|
2021-04-06 11:48:30 +00:00
|
|
|
}
|
2018-07-18 14:14:26 +00:00
|
|
|
|
|
|
|
if (irq_value & BIT(CHL_INT2_SL_IDAF_TOUT_CONF_OFF)) {
|
|
|
|
dev_warn(dev, "phy%d identify timeout\n", phy_no);
|
|
|
|
hisi_sas_notify_phy_event(phy, HISI_PHYE_LINK_RESET);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (irq_value & BIT(CHL_INT2_STP_LINK_TIMEOUT_OFF)) {
|
|
|
|
u32 reg_value = hisi_sas_phy_read32(hisi_hba, phy_no,
|
|
|
|
STP_LINK_TIMEOUT_STATE);
|
|
|
|
|
|
|
|
dev_warn(dev, "phy%d stp link timeout (0x%x)\n",
|
|
|
|
phy_no, reg_value);
|
|
|
|
if (reg_value & BIT(4))
|
|
|
|
hisi_sas_notify_phy_event(phy, HISI_PHYE_LINK_RESET);
|
|
|
|
}
|
|
|
|
|
2019-02-28 14:50:59 +00:00
|
|
|
if (pci_dev->revision > 0x20 && (irq_value & msk)) {
|
|
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
struct sas_phy *sphy = sas_phy->phy;
|
|
|
|
|
|
|
|
phy_get_events_v3_hw(hisi_hba, phy_no);
|
|
|
|
|
|
|
|
if (irq_value & BIT(CHL_INT2_RX_INVLD_DW_OFF))
|
|
|
|
dev_info(dev, "phy%d invalid dword cnt: %u\n", phy_no,
|
|
|
|
sphy->invalid_dword_count);
|
|
|
|
|
|
|
|
if (irq_value & BIT(CHL_INT2_RX_CODE_ERR_OFF))
|
|
|
|
dev_info(dev, "phy%d code violation cnt: %u\n", phy_no,
|
|
|
|
phy->code_violation_err_count);
|
|
|
|
|
|
|
|
if (irq_value & BIT(CHL_INT2_RX_DISP_ERR_OFF))
|
|
|
|
dev_info(dev, "phy%d disparity error cnt: %u\n", phy_no,
|
|
|
|
sphy->running_disparity_error_count);
|
|
|
|
}
|
|
|
|
|
2018-07-18 14:14:26 +00:00
|
|
|
if ((irq_value & BIT(CHL_INT2_RX_INVLD_DW_OFF)) &&
|
|
|
|
(pci_dev->revision == 0x20)) {
|
|
|
|
u32 reg_value;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = hisi_sas_read32_poll_timeout_atomic(
|
|
|
|
HILINK_ERR_DFX, reg_value,
|
|
|
|
!((reg_value >> 8) & BIT(phy_no)),
|
|
|
|
1000, 10000);
|
2018-07-18 14:14:27 +00:00
|
|
|
if (rc)
|
|
|
|
hisi_sas_notify_phy_event(phy, HISI_PHYE_LINK_RESET);
|
2018-07-18 14:14:26 +00:00
|
|
|
}
|
2018-07-18 14:14:27 +00:00
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT2, irq_value);
|
2018-07-18 14:14:26 +00:00
|
|
|
}
|
|
|
|
|
2019-01-25 14:22:35 +00:00
|
|
|
static void handle_chl_int0_v3_hw(struct hisi_hba *hisi_hba, int phy_no)
|
|
|
|
{
|
|
|
|
u32 irq_value0 = hisi_sas_phy_read32(hisi_hba, phy_no, CHL_INT0);
|
|
|
|
|
|
|
|
if (irq_value0 & CHL_INT0_PHY_RDY_MSK)
|
|
|
|
hisi_sas_phy_oob_ready(hisi_hba, phy_no);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, CHL_INT0,
|
|
|
|
irq_value0 & (~CHL_INT0_SL_RX_BCST_ACK_MSK)
|
|
|
|
& (~CHL_INT0_SL_PHY_ENABLE_MSK)
|
|
|
|
& (~CHL_INT0_NOT_RDY_MSK));
|
|
|
|
}
|
|
|
|
|
2018-07-18 14:14:26 +00:00
|
|
|
static irqreturn_t int_chnl_int_v3_hw(int irq_no, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = p;
|
2018-05-02 15:56:32 +00:00
|
|
|
u32 irq_msk;
|
2017-06-14 15:33:23 +00:00
|
|
|
int phy_no = 0;
|
|
|
|
|
|
|
|
irq_msk = hisi_sas_read32(hisi_hba, CHNL_INT_STATUS)
|
2020-01-20 12:22:33 +00:00
|
|
|
& CHNL_INT_STS_MSK;
|
2017-06-14 15:33:23 +00:00
|
|
|
|
|
|
|
while (irq_msk) {
|
2020-01-20 12:22:33 +00:00
|
|
|
if (irq_msk & (CHNL_INT_STS_INT0_MSK << (phy_no * CHNL_WIDTH)))
|
2019-01-25 14:22:35 +00:00
|
|
|
handle_chl_int0_v3_hw(hisi_hba, phy_no);
|
2017-12-08 17:16:45 +00:00
|
|
|
|
2020-01-20 12:22:33 +00:00
|
|
|
if (irq_msk & (CHNL_INT_STS_INT1_MSK << (phy_no * CHNL_WIDTH)))
|
2018-07-18 14:14:26 +00:00
|
|
|
handle_chl_int1_v3_hw(hisi_hba, phy_no);
|
2017-12-08 17:16:45 +00:00
|
|
|
|
2020-01-20 12:22:33 +00:00
|
|
|
if (irq_msk & (CHNL_INT_STS_INT2_MSK << (phy_no * CHNL_WIDTH)))
|
2018-07-18 14:14:26 +00:00
|
|
|
handle_chl_int2_v3_hw(hisi_hba, phy_no);
|
2017-06-14 15:33:23 +00:00
|
|
|
|
2020-01-20 12:22:33 +00:00
|
|
|
irq_msk &= ~(CHNL_INT_STS_PHY_MSK << (phy_no * CHNL_WIDTH));
|
2017-06-14 15:33:23 +00:00
|
|
|
phy_no++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
|
|
|
}
|
|
|
|
|
2019-04-11 12:46:41 +00:00
|
|
|
static const struct hisi_sas_hw_error multi_bit_ecc_errors[] = {
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_DQE_ECC_MB_OFF),
|
|
|
|
.msk = HGC_DQE_ECC_MB_ADDR_MSK,
|
|
|
|
.shift = HGC_DQE_ECC_MB_ADDR_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "hgc_dqe_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_DQE_ECC_ADDR,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_IOST_ECC_MB_OFF),
|
|
|
|
.msk = HGC_IOST_ECC_MB_ADDR_MSK,
|
|
|
|
.shift = HGC_IOST_ECC_MB_ADDR_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "hgc_iost_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_IOST_ECC_ADDR,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_ITCT_ECC_MB_OFF),
|
|
|
|
.msk = HGC_ITCT_ECC_MB_ADDR_MSK,
|
|
|
|
.shift = HGC_ITCT_ECC_MB_ADDR_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "hgc_itct_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_ITCT_ECC_ADDR,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_IOSTLIST_ECC_MB_OFF),
|
|
|
|
.msk = HGC_LM_DFX_STATUS2_IOSTLIST_MSK,
|
|
|
|
.shift = HGC_LM_DFX_STATUS2_IOSTLIST_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "hgc_iostl_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_LM_DFX_STATUS2,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_ITCTLIST_ECC_MB_OFF),
|
|
|
|
.msk = HGC_LM_DFX_STATUS2_ITCTLIST_MSK,
|
|
|
|
.shift = HGC_LM_DFX_STATUS2_ITCTLIST_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "hgc_itctl_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_LM_DFX_STATUS2,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_CQE_ECC_MB_OFF),
|
|
|
|
.msk = HGC_CQE_ECC_MB_ADDR_MSK,
|
|
|
|
.shift = HGC_CQE_ECC_MB_ADDR_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "hgc_cqe_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_CQE_ECC_ADDR,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_NCQ_MEM0_ECC_MB_OFF),
|
|
|
|
.msk = HGC_RXM_DFX_STATUS14_MEM0_MSK,
|
|
|
|
.shift = HGC_RXM_DFX_STATUS14_MEM0_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "rxm_mem0_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_RXM_DFX_STATUS14,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_NCQ_MEM1_ECC_MB_OFF),
|
|
|
|
.msk = HGC_RXM_DFX_STATUS14_MEM1_MSK,
|
|
|
|
.shift = HGC_RXM_DFX_STATUS14_MEM1_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "rxm_mem1_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_RXM_DFX_STATUS14,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_NCQ_MEM2_ECC_MB_OFF),
|
|
|
|
.msk = HGC_RXM_DFX_STATUS14_MEM2_MSK,
|
|
|
|
.shift = HGC_RXM_DFX_STATUS14_MEM2_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "rxm_mem2_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_RXM_DFX_STATUS14,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_NCQ_MEM3_ECC_MB_OFF),
|
|
|
|
.msk = HGC_RXM_DFX_STATUS15_MEM3_MSK,
|
|
|
|
.shift = HGC_RXM_DFX_STATUS15_MEM3_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "rxm_mem3_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = HGC_RXM_DFX_STATUS15,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(SAS_ECC_INTR_OOO_RAM_ECC_MB_OFF),
|
|
|
|
.msk = AM_ROB_ECC_ERR_ADDR_MSK,
|
|
|
|
.shift = AM_ROB_ECC_ERR_ADDR_OFF,
|
2019-05-29 09:58:43 +00:00
|
|
|
.msg = "ooo_ram_eccbad_intr",
|
2019-04-11 12:46:41 +00:00
|
|
|
.reg = AM_ROB_ECC_ERR_ADDR,
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
static void multi_bit_ecc_error_process_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
u32 irq_value)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
const struct hisi_sas_hw_error *ecc_error;
|
|
|
|
u32 val;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(multi_bit_ecc_errors); i++) {
|
|
|
|
ecc_error = &multi_bit_ecc_errors[i];
|
|
|
|
if (irq_value & ecc_error->irq_msk) {
|
|
|
|
val = hisi_sas_read32(hisi_hba, ecc_error->reg);
|
|
|
|
val &= ecc_error->msk;
|
|
|
|
val >>= ecc_error->shift;
|
2019-05-29 09:58:43 +00:00
|
|
|
dev_err(dev, "%s (0x%x) found: mem addr is 0x%08X\n",
|
|
|
|
ecc_error->msg, irq_value, val);
|
2019-04-11 12:46:41 +00:00
|
|
|
queue_work(hisi_hba->wq, &hisi_hba->rst_work);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void fatal_ecc_int_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
u32 irq_value, irq_msk;
|
|
|
|
|
|
|
|
irq_msk = hisi_sas_read32(hisi_hba, SAS_ECC_INTR_MSK);
|
2019-08-05 13:48:10 +00:00
|
|
|
hisi_sas_write32(hisi_hba, SAS_ECC_INTR_MSK, 0xffffffff);
|
2019-04-11 12:46:41 +00:00
|
|
|
|
|
|
|
irq_value = hisi_sas_read32(hisi_hba, SAS_ECC_INTR);
|
|
|
|
if (irq_value)
|
|
|
|
multi_bit_ecc_error_process_v3_hw(hisi_hba, irq_value);
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_ECC_INTR, irq_value);
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_ECC_INTR_MSK, irq_msk);
|
|
|
|
}
|
|
|
|
|
2017-10-24 15:51:48 +00:00
|
|
|
static const struct hisi_sas_hw_error axi_error[] = {
|
|
|
|
{ .msk = BIT(0), .msg = "IOST_AXI_W_ERR" },
|
|
|
|
{ .msk = BIT(1), .msg = "IOST_AXI_R_ERR" },
|
|
|
|
{ .msk = BIT(2), .msg = "ITCT_AXI_W_ERR" },
|
|
|
|
{ .msk = BIT(3), .msg = "ITCT_AXI_R_ERR" },
|
|
|
|
{ .msk = BIT(4), .msg = "SATA_AXI_W_ERR" },
|
|
|
|
{ .msk = BIT(5), .msg = "SATA_AXI_R_ERR" },
|
|
|
|
{ .msk = BIT(6), .msg = "DQE_AXI_R_ERR" },
|
|
|
|
{ .msk = BIT(7), .msg = "CQE_AXI_W_ERR" },
|
2019-04-11 12:46:44 +00:00
|
|
|
{}
|
2017-10-24 15:51:48 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static const struct hisi_sas_hw_error fifo_error[] = {
|
|
|
|
{ .msk = BIT(8), .msg = "CQE_WINFO_FIFO" },
|
|
|
|
{ .msk = BIT(9), .msg = "CQE_MSG_FIFIO" },
|
|
|
|
{ .msk = BIT(10), .msg = "GETDQE_FIFO" },
|
|
|
|
{ .msk = BIT(11), .msg = "CMDP_FIFO" },
|
|
|
|
{ .msk = BIT(12), .msg = "AWTCTRL_FIFO" },
|
2019-04-11 12:46:44 +00:00
|
|
|
{}
|
2017-10-24 15:51:48 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static const struct hisi_sas_hw_error fatal_axi_error[] = {
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_WP_DEPTH_OFF),
|
|
|
|
.msg = "write pointer and depth",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_IPTT_SLOT_NOMATCH_OFF),
|
|
|
|
.msg = "iptt no match slot",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_RP_DEPTH_OFF),
|
|
|
|
.msg = "read pointer and depth",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_AXI_OFF),
|
|
|
|
.reg = HGC_AXI_FIFO_ERR_INFO,
|
|
|
|
.sub = axi_error,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_FIFO_OFF),
|
|
|
|
.reg = HGC_AXI_FIFO_ERR_INFO,
|
|
|
|
.sub = fifo_error,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_LM_OFF),
|
|
|
|
.msg = "LM add/fetch list",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_ABT_OFF),
|
|
|
|
.msg = "SAS_HGC_ABT fetch LM list",
|
|
|
|
},
|
2019-04-11 12:46:41 +00:00
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_DQE_POISON_OFF),
|
|
|
|
.msg = "read dqe poison",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_IOST_POISON_OFF),
|
|
|
|
.msg = "read iost poison",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_ITCT_POISON_OFF),
|
|
|
|
.msg = "read itct poison",
|
|
|
|
},
|
|
|
|
{
|
|
|
|
.irq_msk = BIT(ENT_INT_SRC3_ITCT_NCQ_POISON_OFF),
|
|
|
|
.msg = "read itct ncq poison",
|
|
|
|
},
|
|
|
|
|
2017-10-24 15:51:48 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static irqreturn_t fatal_axi_int_v3_hw(int irq_no, void *p)
|
|
|
|
{
|
|
|
|
u32 irq_value, irq_msk;
|
|
|
|
struct hisi_hba *hisi_hba = p;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
2019-01-25 14:22:31 +00:00
|
|
|
struct pci_dev *pdev = hisi_hba->pci_dev;
|
2017-10-24 15:51:48 +00:00
|
|
|
int i;
|
|
|
|
|
|
|
|
irq_msk = hisi_sas_read32(hisi_hba, ENT_INT_SRC_MSK3);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK3, irq_msk | 0x1df00);
|
|
|
|
|
|
|
|
irq_value = hisi_sas_read32(hisi_hba, ENT_INT_SRC3);
|
2018-03-23 16:05:09 +00:00
|
|
|
irq_value &= ~irq_msk;
|
2017-10-24 15:51:48 +00:00
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(fatal_axi_error); i++) {
|
|
|
|
const struct hisi_sas_hw_error *error = &fatal_axi_error[i];
|
|
|
|
|
|
|
|
if (!(irq_value & error->irq_msk))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (error->sub) {
|
|
|
|
const struct hisi_sas_hw_error *sub = error->sub;
|
|
|
|
u32 err_value = hisi_sas_read32(hisi_hba, error->reg);
|
|
|
|
|
|
|
|
for (; sub->msk || sub->msg; sub++) {
|
|
|
|
if (!(err_value & sub->msk))
|
|
|
|
continue;
|
|
|
|
|
2017-12-08 17:16:41 +00:00
|
|
|
dev_err(dev, "%s error (0x%x) found!\n",
|
2017-10-24 15:51:48 +00:00
|
|
|
sub->msg, irq_value);
|
|
|
|
queue_work(hisi_hba->wq, &hisi_hba->rst_work);
|
|
|
|
}
|
|
|
|
} else {
|
2017-12-08 17:16:41 +00:00
|
|
|
dev_err(dev, "%s error (0x%x) found!\n",
|
2017-10-24 15:51:48 +00:00
|
|
|
error->msg, irq_value);
|
|
|
|
queue_work(hisi_hba->wq, &hisi_hba->rst_work);
|
|
|
|
}
|
2019-01-25 14:22:31 +00:00
|
|
|
|
|
|
|
if (pdev->revision < 0x21) {
|
|
|
|
u32 reg_val;
|
|
|
|
|
|
|
|
reg_val = hisi_sas_read32(hisi_hba,
|
|
|
|
AXI_MASTER_CFG_BASE +
|
|
|
|
AM_CTRL_GLOBAL);
|
|
|
|
reg_val |= AM_CTRL_SHUTDOWN_REQ_MSK;
|
|
|
|
hisi_sas_write32(hisi_hba, AXI_MASTER_CFG_BASE +
|
|
|
|
AM_CTRL_GLOBAL, reg_val);
|
|
|
|
}
|
2017-10-24 15:51:48 +00:00
|
|
|
}
|
|
|
|
|
2019-04-11 12:46:41 +00:00
|
|
|
fatal_ecc_int_v3_hw(hisi_hba);
|
|
|
|
|
2017-10-24 15:51:48 +00:00
|
|
|
if (irq_value & BIT(ENT_INT_SRC3_ITC_INT_OFF)) {
|
|
|
|
u32 reg_val = hisi_sas_read32(hisi_hba, ITCT_CLR);
|
|
|
|
u32 dev_id = reg_val & ITCT_DEV_MSK;
|
|
|
|
struct hisi_sas_device *sas_dev =
|
|
|
|
&hisi_hba->devices[dev_id];
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, ITCT_CLR, 0);
|
|
|
|
dev_dbg(dev, "clear ITCT ok\n");
|
|
|
|
complete(sas_dev->completion);
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC3, irq_value & 0x1df00);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK3, irq_msk);
|
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:24 +00:00
|
|
|
static void
|
|
|
|
slot_err_v3_hw(struct hisi_hba *hisi_hba, struct sas_task *task,
|
|
|
|
struct hisi_sas_slot *slot)
|
|
|
|
{
|
|
|
|
struct task_status_struct *ts = &task->task_status;
|
|
|
|
struct hisi_sas_complete_v3_hdr *complete_queue =
|
|
|
|
hisi_hba->complete_hdr[slot->cmplt_queue];
|
|
|
|
struct hisi_sas_complete_v3_hdr *complete_hdr =
|
|
|
|
&complete_queue[slot->cmplt_queue_slot];
|
2017-06-29 13:02:14 +00:00
|
|
|
struct hisi_sas_err_record_v3 *record =
|
|
|
|
hisi_sas_status_buf_addr_mem(slot);
|
2018-12-06 13:34:40 +00:00
|
|
|
u32 dma_rx_err_type = le32_to_cpu(record->dma_rx_err_type);
|
|
|
|
u32 trans_tx_fail_type = le32_to_cpu(record->trans_tx_fail_type);
|
|
|
|
u32 dw3 = le32_to_cpu(complete_hdr->dw3);
|
2017-06-14 15:33:24 +00:00
|
|
|
|
|
|
|
switch (task->task_proto) {
|
|
|
|
case SAS_PROTOCOL_SSP:
|
|
|
|
if (dma_rx_err_type & RX_DATA_LEN_UNDERFLOW_MSK) {
|
|
|
|
ts->residual = trans_tx_fail_type;
|
|
|
|
ts->stat = SAS_DATA_UNDERRUN;
|
2018-12-06 13:34:40 +00:00
|
|
|
} else if (dw3 & CMPLT_HDR_IO_IN_TARGET_MSK) {
|
2017-06-14 15:33:24 +00:00
|
|
|
ts->stat = SAS_QUEUE_FULL;
|
|
|
|
slot->abort = 1;
|
|
|
|
} else {
|
|
|
|
ts->stat = SAS_OPEN_REJECT;
|
|
|
|
ts->open_rej_reason = SAS_OREJ_RSVD_RETRY;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case SAS_PROTOCOL_SATA:
|
|
|
|
case SAS_PROTOCOL_STP:
|
|
|
|
case SAS_PROTOCOL_SATA | SAS_PROTOCOL_STP:
|
|
|
|
if (dma_rx_err_type & RX_DATA_LEN_UNDERFLOW_MSK) {
|
|
|
|
ts->residual = trans_tx_fail_type;
|
|
|
|
ts->stat = SAS_DATA_UNDERRUN;
|
2018-12-06 13:34:40 +00:00
|
|
|
} else if (dw3 & CMPLT_HDR_IO_IN_TARGET_MSK) {
|
2017-06-14 15:33:24 +00:00
|
|
|
ts->stat = SAS_PHY_DOWN;
|
|
|
|
slot->abort = 1;
|
|
|
|
} else {
|
|
|
|
ts->stat = SAS_OPEN_REJECT;
|
|
|
|
ts->open_rej_reason = SAS_OREJ_RSVD_RETRY;
|
|
|
|
}
|
|
|
|
hisi_sas_sata_done(task, slot);
|
|
|
|
break;
|
|
|
|
case SAS_PROTOCOL_SMP:
|
2021-05-24 02:54:55 +00:00
|
|
|
ts->stat = SAS_SAM_STAT_CHECK_CONDITION;
|
2017-06-14 15:33:24 +00:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-05-15 14:13:45 +00:00
|
|
|
static void slot_complete_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
struct hisi_sas_slot *slot)
|
2017-06-14 15:33:24 +00:00
|
|
|
{
|
|
|
|
struct sas_task *task = slot->task;
|
|
|
|
struct hisi_sas_device *sas_dev;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
struct task_status_struct *ts;
|
|
|
|
struct domain_device *device;
|
2018-05-02 15:56:26 +00:00
|
|
|
struct sas_ha_struct *ha;
|
2017-06-14 15:33:24 +00:00
|
|
|
struct hisi_sas_complete_v3_hdr *complete_queue =
|
|
|
|
hisi_hba->complete_hdr[slot->cmplt_queue];
|
|
|
|
struct hisi_sas_complete_v3_hdr *complete_hdr =
|
|
|
|
&complete_queue[slot->cmplt_queue_slot];
|
|
|
|
unsigned long flags;
|
2018-05-02 15:56:26 +00:00
|
|
|
bool is_internal = slot->is_internal;
|
2018-12-06 13:34:40 +00:00
|
|
|
u32 dw0, dw1, dw3;
|
2017-06-14 15:33:24 +00:00
|
|
|
|
|
|
|
if (unlikely(!task || !task->lldd_task || !task->dev))
|
2020-05-15 14:13:45 +00:00
|
|
|
return;
|
2017-06-14 15:33:24 +00:00
|
|
|
|
|
|
|
ts = &task->task_status;
|
|
|
|
device = task->dev;
|
2018-05-02 15:56:26 +00:00
|
|
|
ha = device->port->ha;
|
2017-06-14 15:33:24 +00:00
|
|
|
sas_dev = device->lldd_dev;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&task->task_state_lock, flags);
|
2022-02-10 10:43:24 +00:00
|
|
|
task->task_state_flags &= ~SAS_TASK_STATE_PENDING;
|
2017-06-14 15:33:24 +00:00
|
|
|
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
|
|
|
|
|
|
|
memset(ts, 0, sizeof(*ts));
|
|
|
|
ts->resp = SAS_TASK_COMPLETE;
|
|
|
|
|
|
|
|
if (unlikely(!sas_dev)) {
|
|
|
|
dev_dbg(dev, "slot complete: port has not device\n");
|
|
|
|
ts->stat = SAS_PHY_DOWN;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2018-12-06 13:34:40 +00:00
|
|
|
dw0 = le32_to_cpu(complete_hdr->dw0);
|
|
|
|
dw1 = le32_to_cpu(complete_hdr->dw1);
|
|
|
|
dw3 = le32_to_cpu(complete_hdr->dw3);
|
|
|
|
|
2017-06-14 15:33:24 +00:00
|
|
|
/*
|
|
|
|
* Use SAS+TMF status codes
|
|
|
|
*/
|
2018-12-06 13:34:40 +00:00
|
|
|
switch ((dw0 & CMPLT_HDR_ABORT_STAT_MSK) >> CMPLT_HDR_ABORT_STAT_OFF) {
|
2017-06-14 15:33:24 +00:00
|
|
|
case STAT_IO_ABORTED:
|
|
|
|
/* this IO has been aborted by abort command */
|
|
|
|
ts->stat = SAS_ABORTED_TASK;
|
|
|
|
goto out;
|
|
|
|
case STAT_IO_COMPLETE:
|
|
|
|
/* internal abort command complete */
|
|
|
|
ts->stat = TMF_RESP_FUNC_SUCC;
|
|
|
|
goto out;
|
|
|
|
case STAT_IO_NO_DEVICE:
|
|
|
|
ts->stat = TMF_RESP_FUNC_COMPLETE;
|
|
|
|
goto out;
|
|
|
|
case STAT_IO_NOT_VALID:
|
|
|
|
/*
|
|
|
|
* abort single IO, the controller can't find the IO
|
|
|
|
*/
|
|
|
|
ts->stat = TMF_RESP_FUNC_FAILED;
|
|
|
|
goto out;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* check for erroneous completion */
|
2018-12-06 13:34:40 +00:00
|
|
|
if ((dw0 & CMPLT_HDR_CMPLT_MSK) == 0x3) {
|
2017-12-08 17:16:41 +00:00
|
|
|
u32 *error_info = hisi_sas_status_buf_addr_mem(slot);
|
|
|
|
|
2017-06-14 15:33:24 +00:00
|
|
|
slot_err_v3_hw(hisi_hba, task, slot);
|
2017-12-08 17:16:41 +00:00
|
|
|
if (ts->stat != SAS_DATA_UNDERRUN)
|
2021-04-06 11:48:27 +00:00
|
|
|
dev_info(dev, "erroneous completion iptt=%d task=%pK dev id=%d addr=%016llx CQ hdr: 0x%x 0x%x 0x%x 0x%x Error info: 0x%x 0x%x 0x%x 0x%x\n",
|
2019-04-11 12:46:44 +00:00
|
|
|
slot->idx, task, sas_dev->device_id,
|
2021-04-06 11:48:27 +00:00
|
|
|
SAS_ADDR(device->sas_addr),
|
2019-04-11 12:46:44 +00:00
|
|
|
dw0, dw1, complete_hdr->act, dw3,
|
|
|
|
error_info[0], error_info[1],
|
|
|
|
error_info[2], error_info[3]);
|
2020-07-13 08:04:30 +00:00
|
|
|
if (unlikely(slot->abort)) {
|
|
|
|
sas_task_abort(task);
|
2020-05-15 14:13:45 +00:00
|
|
|
return;
|
2020-07-13 08:04:30 +00:00
|
|
|
}
|
2017-06-14 15:33:24 +00:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (task->task_proto) {
|
|
|
|
case SAS_PROTOCOL_SSP: {
|
2017-06-29 13:02:14 +00:00
|
|
|
struct ssp_response_iu *iu =
|
|
|
|
hisi_sas_status_buf_addr_mem(slot) +
|
2017-06-14 15:33:24 +00:00
|
|
|
sizeof(struct hisi_sas_err_record);
|
|
|
|
|
|
|
|
sas_ssp_task_response(dev, task, iu);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case SAS_PROTOCOL_SMP: {
|
|
|
|
struct scatterlist *sg_resp = &task->smp_task.smp_resp;
|
2019-08-05 13:48:06 +00:00
|
|
|
void *to = page_address(sg_page(sg_resp));
|
2017-06-14 15:33:24 +00:00
|
|
|
|
2021-05-24 02:54:55 +00:00
|
|
|
ts->stat = SAS_SAM_STAT_GOOD;
|
2017-06-14 15:33:24 +00:00
|
|
|
|
|
|
|
dma_unmap_sg(dev, &task->smp_task.smp_req, 1,
|
|
|
|
DMA_TO_DEVICE);
|
|
|
|
memcpy(to + sg_resp->offset,
|
2017-06-29 13:02:14 +00:00
|
|
|
hisi_sas_status_buf_addr_mem(slot) +
|
2017-06-14 15:33:24 +00:00
|
|
|
sizeof(struct hisi_sas_err_record),
|
2019-08-05 13:48:07 +00:00
|
|
|
sg_resp->length);
|
2017-06-14 15:33:24 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
case SAS_PROTOCOL_SATA:
|
|
|
|
case SAS_PROTOCOL_STP:
|
|
|
|
case SAS_PROTOCOL_SATA | SAS_PROTOCOL_STP:
|
2021-05-24 02:54:55 +00:00
|
|
|
ts->stat = SAS_SAM_STAT_GOOD;
|
2017-06-14 15:33:24 +00:00
|
|
|
hisi_sas_sata_done(task, slot);
|
|
|
|
break;
|
|
|
|
default:
|
2021-05-24 02:54:55 +00:00
|
|
|
ts->stat = SAS_SAM_STAT_CHECK_CONDITION;
|
2017-06-14 15:33:24 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!slot->port->port_attached) {
|
2017-12-08 17:16:41 +00:00
|
|
|
dev_warn(dev, "slot complete: port %d has removed\n",
|
2017-06-14 15:33:24 +00:00
|
|
|
slot->port->sas_port.id);
|
|
|
|
ts->stat = SAS_PHY_DOWN;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_lock_irqsave(&task->task_state_lock, flags);
|
2018-05-02 15:56:25 +00:00
|
|
|
if (task->task_state_flags & SAS_TASK_STATE_ABORTED) {
|
|
|
|
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
2019-08-05 13:48:11 +00:00
|
|
|
dev_info(dev, "slot complete: task(%pK) aborted\n", task);
|
2020-05-15 14:13:45 +00:00
|
|
|
return;
|
2018-05-02 15:56:25 +00:00
|
|
|
}
|
2017-06-14 15:33:24 +00:00
|
|
|
task->task_state_flags |= SAS_TASK_STATE_DONE;
|
|
|
|
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
2018-09-24 15:06:31 +00:00
|
|
|
hisi_sas_slot_task_free(hisi_hba, task, slot);
|
2017-06-14 15:33:24 +00:00
|
|
|
|
2018-05-02 15:56:26 +00:00
|
|
|
if (!is_internal && (task->task_proto != SAS_PROTOCOL_SMP)) {
|
|
|
|
spin_lock_irqsave(&device->done_lock, flags);
|
|
|
|
if (test_bit(SAS_HA_FROZEN, &ha->state)) {
|
|
|
|
spin_unlock_irqrestore(&device->done_lock, flags);
|
2019-08-05 13:48:11 +00:00
|
|
|
dev_info(dev, "slot complete: task(%pK) ignored\n ",
|
2018-05-02 15:56:26 +00:00
|
|
|
task);
|
2020-05-15 14:13:45 +00:00
|
|
|
return;
|
2018-05-02 15:56:26 +00:00
|
|
|
}
|
|
|
|
spin_unlock_irqrestore(&device->done_lock, flags);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:24 +00:00
|
|
|
if (task->task_done)
|
|
|
|
task->task_done(task);
|
|
|
|
}
|
|
|
|
|
2020-01-20 12:22:31 +00:00
|
|
|
static irqreturn_t cq_thread_v3_hw(int irq_no, void *p)
|
2017-06-14 15:33:24 +00:00
|
|
|
{
|
2020-01-20 12:22:31 +00:00
|
|
|
struct hisi_sas_cq *cq = p;
|
2017-06-14 15:33:24 +00:00
|
|
|
struct hisi_hba *hisi_hba = cq->hisi_hba;
|
|
|
|
struct hisi_sas_slot *slot;
|
|
|
|
struct hisi_sas_complete_v3_hdr *complete_queue;
|
2017-12-08 17:16:39 +00:00
|
|
|
u32 rd_point = cq->rd_point, wr_point;
|
2017-06-14 15:33:24 +00:00
|
|
|
int queue = cq->id;
|
|
|
|
|
|
|
|
complete_queue = hisi_hba->complete_hdr[queue];
|
|
|
|
|
|
|
|
wr_point = hisi_sas_read32(hisi_hba, COMPL_Q_0_WR_PTR +
|
|
|
|
(0x14 * queue));
|
|
|
|
|
|
|
|
while (rd_point != wr_point) {
|
|
|
|
struct hisi_sas_complete_v3_hdr *complete_hdr;
|
2018-03-23 16:05:13 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
2018-12-06 13:34:40 +00:00
|
|
|
u32 dw1;
|
2017-06-14 15:33:24 +00:00
|
|
|
int iptt;
|
|
|
|
|
|
|
|
complete_hdr = &complete_queue[rd_point];
|
2018-12-06 13:34:40 +00:00
|
|
|
dw1 = le32_to_cpu(complete_hdr->dw1);
|
2017-06-14 15:33:24 +00:00
|
|
|
|
2018-12-06 13:34:40 +00:00
|
|
|
iptt = dw1 & CMPLT_HDR_IPTT_MSK;
|
2018-03-23 16:05:13 +00:00
|
|
|
if (likely(iptt < HISI_SAS_COMMAND_ENTRIES_V3_HW)) {
|
|
|
|
slot = &hisi_hba->slot_info[iptt];
|
|
|
|
slot->cmplt_queue_slot = rd_point;
|
|
|
|
slot->cmplt_queue = queue;
|
|
|
|
slot_complete_v3_hw(hisi_hba, slot);
|
|
|
|
} else
|
|
|
|
dev_err(dev, "IPTT %d is invalid, discard it.\n", iptt);
|
2017-06-14 15:33:24 +00:00
|
|
|
|
|
|
|
if (++rd_point >= HISI_SAS_QUEUE_SLOTS)
|
|
|
|
rd_point = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* update rd_point */
|
|
|
|
cq->rd_point = rd_point;
|
|
|
|
hisi_sas_write32(hisi_hba, COMPL_Q_0_RD_PTR + (0x14 * queue), rd_point);
|
2020-01-20 12:22:31 +00:00
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
2017-06-14 15:33:24 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static irqreturn_t cq_interrupt_v3_hw(int irq_no, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_cq *cq = p;
|
|
|
|
struct hisi_hba *hisi_hba = cq->hisi_hba;
|
|
|
|
int queue = cq->id;
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_SRC, 1 << queue);
|
|
|
|
|
2020-01-20 12:22:31 +00:00
|
|
|
return IRQ_WAKE_THREAD;
|
2017-06-14 15:33:24 +00:00
|
|
|
}
|
|
|
|
|
2020-08-19 15:20:32 +00:00
|
|
|
static int interrupt_preinit_v3_hw(struct hisi_hba *hisi_hba)
|
2019-02-06 10:52:55 +00:00
|
|
|
{
|
2020-08-19 15:20:32 +00:00
|
|
|
int vectors;
|
|
|
|
int max_msi = HISI_SAS_MSI_COUNT_V3_HW, min_msi;
|
|
|
|
struct Scsi_Host *shost = hisi_hba->shost;
|
|
|
|
struct irq_affinity desc = {
|
|
|
|
.pre_vectors = BASE_VECTORS_V3_HW,
|
|
|
|
};
|
2019-02-06 10:52:55 +00:00
|
|
|
|
2020-08-19 15:20:32 +00:00
|
|
|
min_msi = MIN_AFFINE_VECTORS_V3_HW;
|
|
|
|
vectors = pci_alloc_irq_vectors_affinity(hisi_hba->pci_dev,
|
|
|
|
min_msi, max_msi,
|
|
|
|
PCI_IRQ_MSI |
|
|
|
|
PCI_IRQ_AFFINITY,
|
|
|
|
&desc);
|
|
|
|
if (vectors < 0)
|
|
|
|
return -ENOENT;
|
2019-02-06 10:52:55 +00:00
|
|
|
|
|
|
|
|
2020-08-19 15:20:32 +00:00
|
|
|
hisi_hba->cq_nvecs = vectors - BASE_VECTORS_V3_HW;
|
|
|
|
shost->nr_hw_queues = hisi_hba->cq_nvecs;
|
|
|
|
|
|
|
|
return 0;
|
2019-02-06 10:52:55 +00:00
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
static int interrupt_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
struct pci_dev *pdev = hisi_hba->pci_dev;
|
2020-08-19 15:20:32 +00:00
|
|
|
int rc, i;
|
scsi: hisi_sas: Issue internal abort on all relevant queues
To support queue mapped to a CPU, it needs to be ensured that issuing an
internal abort is safe, in that it is guaranteed that an internal abort is
processed for a single IO or a device after all the relevant command(s)
which it is attempting to abort have been processed by the controller.
Currently we only deliver commands for any device on a single queue to
solve this problem, as we know that commands issued on the same queue will
be processed in order, and we will not have a scenario where the internal
abort is racing against a command(s) which it is trying to abort.
To enqueue commands on queue mapped to a CPU, choosing a queue for an
command is based on the associated queue for the current CPU, so this is
not safe for internal abort since it would definitely not be guaranteed
that commands for the command devices are issued on the same queue.
To solve this issue, we take a bludgeoning approach, and issue a separate
internal abort on any queue(s) relevant to the command or device, in that
we will be guaranteed that at least one of these internal aborts will be
received last in the controller.
So, for aborting a single command, we can just force the internal abort to
be issued on the same queue as the command which we are trying to abort.
For aborting all commands associated with a device, we issue a separate
internal abort on all relevant queues. Issuing multiple internal aborts in
this fashion would have not side affect.
Signed-off-by: John Garry <john.garry@huawei.com>
Signed-off-by: Xiang Chen <chenxiang66@hisilicon.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2019-02-06 10:52:54 +00:00
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
rc = devm_request_irq(dev, pci_irq_vector(pdev, 1),
|
|
|
|
int_phy_up_down_bcast_v3_hw, 0,
|
|
|
|
DRV_NAME " phy", hisi_hba);
|
|
|
|
if (rc) {
|
|
|
|
dev_err(dev, "could not request phy interrupt, rc=%d\n", rc);
|
2020-11-24 08:46:33 +00:00
|
|
|
return -ENOENT;
|
2017-06-14 15:33:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
rc = devm_request_irq(dev, pci_irq_vector(pdev, 2),
|
|
|
|
int_chnl_int_v3_hw, 0,
|
|
|
|
DRV_NAME " channel", hisi_hba);
|
|
|
|
if (rc) {
|
|
|
|
dev_err(dev, "could not request chnl interrupt, rc=%d\n", rc);
|
2020-11-24 08:46:33 +00:00
|
|
|
return -ENOENT;
|
2017-06-14 15:33:23 +00:00
|
|
|
}
|
|
|
|
|
2017-10-24 15:51:48 +00:00
|
|
|
rc = devm_request_irq(dev, pci_irq_vector(pdev, 11),
|
|
|
|
fatal_axi_int_v3_hw, 0,
|
|
|
|
DRV_NAME " fatal", hisi_hba);
|
|
|
|
if (rc) {
|
|
|
|
dev_err(dev, "could not request fatal interrupt, rc=%d\n", rc);
|
2020-11-24 08:46:33 +00:00
|
|
|
return -ENOENT;
|
2017-10-24 15:51:48 +00:00
|
|
|
}
|
|
|
|
|
2020-01-20 12:22:35 +00:00
|
|
|
if (hisi_sas_intr_conv)
|
|
|
|
dev_info(dev, "Enable interrupt converge\n");
|
|
|
|
|
scsi: hisi_sas: Issue internal abort on all relevant queues
To support queue mapped to a CPU, it needs to be ensured that issuing an
internal abort is safe, in that it is guaranteed that an internal abort is
processed for a single IO or a device after all the relevant command(s)
which it is attempting to abort have been processed by the controller.
Currently we only deliver commands for any device on a single queue to
solve this problem, as we know that commands issued on the same queue will
be processed in order, and we will not have a scenario where the internal
abort is racing against a command(s) which it is trying to abort.
To enqueue commands on queue mapped to a CPU, choosing a queue for an
command is based on the associated queue for the current CPU, so this is
not safe for internal abort since it would definitely not be guaranteed
that commands for the command devices are issued on the same queue.
To solve this issue, we take a bludgeoning approach, and issue a separate
internal abort on any queue(s) relevant to the command or device, in that
we will be guaranteed that at least one of these internal aborts will be
received last in the controller.
So, for aborting a single command, we can just force the internal abort to
be issued on the same queue as the command which we are trying to abort.
For aborting all commands associated with a device, we issue a separate
internal abort on all relevant queues. Issuing multiple internal aborts in
this fashion would have not side affect.
Signed-off-by: John Garry <john.garry@huawei.com>
Signed-off-by: Xiang Chen <chenxiang66@hisilicon.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2019-02-06 10:52:54 +00:00
|
|
|
for (i = 0; i < hisi_hba->cq_nvecs; i++) {
|
2017-06-14 15:33:24 +00:00
|
|
|
struct hisi_sas_cq *cq = &hisi_hba->cq[i];
|
2018-11-09 14:06:33 +00:00
|
|
|
int nr = hisi_sas_intr_conv ? 16 : 16 + i;
|
2020-01-20 12:22:31 +00:00
|
|
|
unsigned long irqflags = hisi_sas_intr_conv ? IRQF_SHARED :
|
|
|
|
IRQF_ONESHOT;
|
|
|
|
|
|
|
|
cq->irq_no = pci_irq_vector(pdev, nr);
|
|
|
|
rc = devm_request_threaded_irq(dev, cq->irq_no,
|
|
|
|
cq_interrupt_v3_hw,
|
|
|
|
cq_thread_v3_hw,
|
|
|
|
irqflags,
|
2018-11-09 14:06:33 +00:00
|
|
|
DRV_NAME " cq", cq);
|
2017-06-14 15:33:24 +00:00
|
|
|
if (rc) {
|
2019-04-11 12:46:44 +00:00
|
|
|
dev_err(dev, "could not request cq%d interrupt, rc=%d\n",
|
2017-06-14 15:33:24 +00:00
|
|
|
i, rc);
|
2020-11-24 08:46:33 +00:00
|
|
|
return -ENOENT;
|
2017-06-14 15:33:24 +00:00
|
|
|
}
|
2020-12-07 13:30:55 +00:00
|
|
|
cq->irq_mask = pci_irq_get_affinity(pdev, i + BASE_VECTORS_V3_HW);
|
|
|
|
if (!cq->irq_mask) {
|
|
|
|
dev_err(dev, "could not get cq%d irq affinity!\n", i);
|
|
|
|
return -ENOENT;
|
|
|
|
}
|
2017-06-14 15:33:24 +00:00
|
|
|
}
|
2017-06-14 15:33:23 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
static int hisi_sas_v3_init(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = hw_init_v3_hw(hisi_hba);
|
|
|
|
if (rc)
|
|
|
|
return rc;
|
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
rc = interrupt_init_v3_hw(hisi_hba);
|
|
|
|
if (rc)
|
|
|
|
return rc;
|
|
|
|
|
2017-06-14 15:33:21 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-08-10 16:09:41 +00:00
|
|
|
static void phy_set_linkrate_v3_hw(struct hisi_hba *hisi_hba, int phy_no,
|
|
|
|
struct sas_phy_linkrates *r)
|
|
|
|
{
|
2018-05-21 10:09:13 +00:00
|
|
|
enum sas_linkrate max = r->maximum_linkrate;
|
2020-09-01 11:13:05 +00:00
|
|
|
u32 prog_phy_link_rate = hisi_sas_phy_read32(hisi_hba, phy_no,
|
|
|
|
PROG_PHY_LINK_RATE);
|
2017-08-10 16:09:41 +00:00
|
|
|
|
2020-09-01 11:13:05 +00:00
|
|
|
prog_phy_link_rate &= ~CFG_PROG_PHY_LINK_RATE_MSK;
|
2018-05-02 15:56:30 +00:00
|
|
|
prog_phy_link_rate |= hisi_sas_get_prog_phy_linkrate_mask(max);
|
2017-08-10 16:09:41 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PROG_PHY_LINK_RATE,
|
2018-05-21 10:09:13 +00:00
|
|
|
prog_phy_link_rate);
|
2017-08-10 16:09:41 +00:00
|
|
|
}
|
|
|
|
|
2017-08-10 16:09:40 +00:00
|
|
|
static void interrupt_disable_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct pci_dev *pdev = hisi_hba->pci_dev;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
synchronize_irq(pci_irq_vector(pdev, 1));
|
|
|
|
synchronize_irq(pci_irq_vector(pdev, 2));
|
|
|
|
synchronize_irq(pci_irq_vector(pdev, 11));
|
2020-10-02 14:30:32 +00:00
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++)
|
2017-08-10 16:09:40 +00:00
|
|
|
hisi_sas_write32(hisi_hba, OQ0_INT_SRC_MSK + 0x4 * i, 0x1);
|
2020-10-02 14:30:32 +00:00
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->cq_nvecs; i++)
|
2017-08-10 16:09:40 +00:00
|
|
|
synchronize_irq(pci_irq_vector(pdev, i + 16));
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK1, 0xffffffff);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK2, 0xffffffff);
|
|
|
|
hisi_sas_write32(hisi_hba, ENT_INT_SRC_MSK3, 0xffffffff);
|
|
|
|
hisi_sas_write32(hisi_hba, SAS_ECC_INTR_MSK, 0xffffffff);
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->n_phy; i++) {
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT1_MSK, 0xffffffff);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, CHL_INT2_MSK, 0xffffffff);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHYCTRL_NOT_RDY_MSK, 0x1);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, PHYCTRL_PHY_ENA_MSK, 0x1);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, i, SL_RX_BCAST_CHK_MSK, 0x1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 get_phys_state_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
return hisi_sas_read32(hisi_hba, PHY_STATE);
|
|
|
|
}
|
|
|
|
|
2018-07-18 14:14:29 +00:00
|
|
|
static int disable_host_v3_hw(struct hisi_hba *hisi_hba)
|
2017-08-10 16:09:40 +00:00
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
2018-07-18 14:14:29 +00:00
|
|
|
u32 status, reg_val;
|
2017-08-10 16:09:40 +00:00
|
|
|
int rc;
|
|
|
|
|
|
|
|
interrupt_disable_v3_hw(hisi_hba);
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_QUEUE_ENABLE, 0x0);
|
|
|
|
|
|
|
|
hisi_sas_stop_phys(hisi_hba);
|
|
|
|
|
|
|
|
mdelay(10);
|
|
|
|
|
2018-07-18 14:14:29 +00:00
|
|
|
reg_val = hisi_sas_read32(hisi_hba, AXI_MASTER_CFG_BASE +
|
|
|
|
AM_CTRL_GLOBAL);
|
|
|
|
reg_val |= AM_CTRL_SHUTDOWN_REQ_MSK;
|
|
|
|
hisi_sas_write32(hisi_hba, AXI_MASTER_CFG_BASE +
|
|
|
|
AM_CTRL_GLOBAL, reg_val);
|
2017-08-10 16:09:40 +00:00
|
|
|
|
|
|
|
/* wait until bus idle */
|
2018-05-02 15:56:33 +00:00
|
|
|
rc = hisi_sas_read32_poll_timeout(AXI_MASTER_CFG_BASE +
|
|
|
|
AM_CURR_TRANS_RETURN, status,
|
|
|
|
status == 0x3, 10, 100);
|
2017-08-10 16:09:40 +00:00
|
|
|
if (rc) {
|
2018-07-18 14:14:29 +00:00
|
|
|
dev_err(dev, "axi bus is not idle, rc=%d\n", rc);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int soft_reset_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = disable_host_v3_hw(hisi_hba);
|
|
|
|
if (rc) {
|
|
|
|
dev_err(dev, "soft reset: disable host failed rc=%d\n", rc);
|
2017-08-10 16:09:40 +00:00
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_sas_init_mem(hisi_hba);
|
|
|
|
|
|
|
|
return hw_init_v3_hw(hisi_hba);
|
|
|
|
}
|
|
|
|
|
2018-05-21 10:09:16 +00:00
|
|
|
static int write_gpio_v3_hw(struct hisi_hba *hisi_hba, u8 reg_type,
|
|
|
|
u8 reg_index, u8 reg_count, u8 *write_data)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
u32 *data = (u32 *)write_data;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
switch (reg_type) {
|
|
|
|
case SAS_GPIO_REG_TX:
|
|
|
|
if ((reg_index + reg_count) > ((hisi_hba->n_phy + 3) / 4)) {
|
|
|
|
dev_err(dev, "write gpio: invalid reg range[%d, %d]\n",
|
|
|
|
reg_index, reg_index + reg_count - 1);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < reg_count; i++)
|
|
|
|
hisi_sas_write32(hisi_hba,
|
|
|
|
SAS_GPIO_TX_0_1 + (reg_index + i) * 4,
|
|
|
|
data[i]);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_err(dev, "write gpio: unsupported or bad reg type %d\n",
|
2019-04-11 12:46:44 +00:00
|
|
|
reg_type);
|
2018-05-21 10:09:16 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-09-06 12:55:34 +00:00
|
|
|
static void wait_cmds_complete_timeout_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
int delay_ms, int timeout_ms)
|
2018-05-21 10:09:20 +00:00
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int entries, entries_old = 0, time;
|
|
|
|
|
|
|
|
for (time = 0; time < timeout_ms; time += delay_ms) {
|
|
|
|
entries = hisi_sas_read32(hisi_hba, CQE_SEND_CNT);
|
|
|
|
if (entries == entries_old)
|
|
|
|
break;
|
|
|
|
|
|
|
|
entries_old = entries;
|
|
|
|
msleep(delay_ms);
|
|
|
|
}
|
|
|
|
|
2019-09-06 12:55:34 +00:00
|
|
|
if (time >= timeout_ms) {
|
|
|
|
dev_dbg(dev, "Wait commands complete timeout!\n");
|
|
|
|
return;
|
|
|
|
}
|
2018-12-19 15:56:41 +00:00
|
|
|
|
2018-05-21 10:09:20 +00:00
|
|
|
dev_dbg(dev, "wait commands complete %dms\n", time);
|
|
|
|
}
|
|
|
|
|
2018-11-09 14:06:33 +00:00
|
|
|
static ssize_t intr_conv_v3_hw_show(struct device *dev,
|
|
|
|
struct device_attribute *attr, char *buf)
|
|
|
|
{
|
|
|
|
return scnprintf(buf, PAGE_SIZE, "%u\n", hisi_sas_intr_conv);
|
|
|
|
}
|
|
|
|
static DEVICE_ATTR_RO(intr_conv_v3_hw);
|
|
|
|
|
2018-11-09 14:06:34 +00:00
|
|
|
static void config_intr_coal_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
/* config those registers between enable and disable PHYs */
|
|
|
|
hisi_sas_stop_phys(hisi_hba);
|
|
|
|
|
|
|
|
if (hisi_hba->intr_coal_ticks == 0 ||
|
|
|
|
hisi_hba->intr_coal_count == 0) {
|
|
|
|
hisi_sas_write32(hisi_hba, INT_COAL_EN, 0x1);
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_COAL_TIME, 0x1);
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_COAL_CNT, 0x1);
|
|
|
|
} else {
|
|
|
|
hisi_sas_write32(hisi_hba, INT_COAL_EN, 0x3);
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_COAL_TIME,
|
|
|
|
hisi_hba->intr_coal_ticks);
|
|
|
|
hisi_sas_write32(hisi_hba, OQ_INT_COAL_CNT,
|
|
|
|
hisi_hba->intr_coal_count);
|
|
|
|
}
|
|
|
|
phys_init_v3_hw(hisi_hba);
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t intr_coal_ticks_v3_hw_show(struct device *dev,
|
|
|
|
struct device_attribute *attr,
|
|
|
|
char *buf)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
|
|
struct hisi_hba *hisi_hba = shost_priv(shost);
|
|
|
|
|
|
|
|
return scnprintf(buf, PAGE_SIZE, "%u\n",
|
|
|
|
hisi_hba->intr_coal_ticks);
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t intr_coal_ticks_v3_hw_store(struct device *dev,
|
|
|
|
struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
|
|
struct hisi_hba *hisi_hba = shost_priv(shost);
|
|
|
|
u32 intr_coal_ticks;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = kstrtou32(buf, 10, &intr_coal_ticks);
|
|
|
|
if (ret) {
|
|
|
|
dev_err(dev, "Input data of interrupt coalesce unmatch\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (intr_coal_ticks >= BIT(24)) {
|
|
|
|
dev_err(dev, "intr_coal_ticks must be less than 2^24!\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_hba->intr_coal_ticks = intr_coal_ticks;
|
|
|
|
|
|
|
|
config_intr_coal_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
static DEVICE_ATTR_RW(intr_coal_ticks_v3_hw);
|
|
|
|
|
|
|
|
static ssize_t intr_coal_count_v3_hw_show(struct device *dev,
|
|
|
|
struct device_attribute
|
|
|
|
*attr, char *buf)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
|
|
struct hisi_hba *hisi_hba = shost_priv(shost);
|
|
|
|
|
|
|
|
return scnprintf(buf, PAGE_SIZE, "%u\n",
|
|
|
|
hisi_hba->intr_coal_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t intr_coal_count_v3_hw_store(struct device *dev,
|
|
|
|
struct device_attribute
|
|
|
|
*attr, const char *buf, size_t count)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost = class_to_shost(dev);
|
|
|
|
struct hisi_hba *hisi_hba = shost_priv(shost);
|
|
|
|
u32 intr_coal_count;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = kstrtou32(buf, 10, &intr_coal_count);
|
|
|
|
if (ret) {
|
|
|
|
dev_err(dev, "Input data of interrupt coalesce unmatch\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (intr_coal_count >= BIT(8)) {
|
|
|
|
dev_err(dev, "intr_coal_count must be less than 2^8!\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_hba->intr_coal_count = intr_coal_count;
|
|
|
|
|
|
|
|
config_intr_coal_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
static DEVICE_ATTR_RW(intr_coal_count_v3_hw);
|
|
|
|
|
2020-10-02 14:30:36 +00:00
|
|
|
static int slave_configure_v3_hw(struct scsi_device *sdev)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost = dev_to_shost(&sdev->sdev_gendev);
|
|
|
|
struct domain_device *ddev = sdev_to_domain_dev(sdev);
|
|
|
|
struct hisi_hba *hisi_hba = shost_priv(shost);
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int ret = sas_slave_configure(sdev);
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
if (!dev_is_sata(ddev))
|
|
|
|
sas_change_queue_depth(sdev, 64);
|
|
|
|
|
|
|
|
if (sdev->type == TYPE_ENCLOSURE)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!device_link_add(&sdev->sdev_gendev, dev,
|
|
|
|
DL_FLAG_PM_RUNTIME | DL_FLAG_RPM_ACTIVE)) {
|
|
|
|
if (pm_runtime_enabled(dev)) {
|
|
|
|
dev_info(dev, "add device link failed, disable runtime PM for the host\n");
|
|
|
|
pm_runtime_disable(dev);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-10-12 23:35:32 +00:00
|
|
|
static struct attribute *host_v3_hw_attrs[] = {
|
|
|
|
&dev_attr_phy_event_threshold.attr,
|
|
|
|
&dev_attr_intr_conv_v3_hw.attr,
|
|
|
|
&dev_attr_intr_coal_ticks_v3_hw.attr,
|
|
|
|
&dev_attr_intr_coal_count_v3_hw.attr,
|
2018-11-09 14:06:32 +00:00
|
|
|
NULL
|
|
|
|
};
|
|
|
|
|
2021-10-12 23:35:32 +00:00
|
|
|
ATTRIBUTE_GROUPS(host_v3_hw);
|
|
|
|
|
2020-11-24 08:46:34 +00:00
|
|
|
#define HISI_SAS_DEBUGFS_REG(x) {#x, x}
|
|
|
|
|
|
|
|
struct hisi_sas_debugfs_reg_lu {
|
|
|
|
char *name;
|
|
|
|
int off;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct hisi_sas_debugfs_reg {
|
|
|
|
const struct hisi_sas_debugfs_reg_lu *lu;
|
|
|
|
int count;
|
|
|
|
int base_off;
|
|
|
|
};
|
|
|
|
|
2018-12-19 15:56:43 +00:00
|
|
|
static const struct hisi_sas_debugfs_reg_lu debugfs_port_reg_lu[] = {
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HARD_PHY_LINKRATE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PROG_PHY_LINK_RATE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_CTRL),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SL_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(AIP_LIMIT),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SL_CONTROL),
|
|
|
|
HISI_SAS_DEBUGFS_REG(RX_PRIMS_STATUS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD0),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD2),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD3),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD4),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD5),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TX_ID_DWORD6),
|
|
|
|
HISI_SAS_DEBUGFS_REG(TXID_AUTO),
|
|
|
|
HISI_SAS_DEBUGFS_REG(RX_IDAF_DWORD0),
|
|
|
|
HISI_SAS_DEBUGFS_REG(RXOP_CHECK_CFG_H),
|
|
|
|
HISI_SAS_DEBUGFS_REG(STP_LINK_TIMER),
|
|
|
|
HISI_SAS_DEBUGFS_REG(STP_LINK_TIMEOUT_STATE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CON_CFG_DRIVER),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_SSP_CON_TIMER_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_SMP_CON_TIMER_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_STP_CON_TIMER_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT0),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT2),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT0_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT1_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT2_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_EC_INT_COAL_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHL_INT_COAL_EN),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RX_TRAIN_TIMER),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_CTRL_RDY_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHYCTRL_NOT_RDY_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHYCTRL_DWS_RESET_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHYCTRL_PHY_ENA_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SL_RX_BCAST_CHK_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHYCTRL_OOB_RESTART_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(DMA_TX_STATUS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(DMA_RX_STATUS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(COARSETUNE_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ERR_CNT_DWS_LOST),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ERR_CNT_RESET_PROB),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ERR_CNT_INVLD_DW),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ERR_CNT_CODE_ERR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ERR_CNT_DISP_ERR),
|
|
|
|
{}
|
|
|
|
};
|
|
|
|
|
2018-12-19 15:56:40 +00:00
|
|
|
static const struct hisi_sas_debugfs_reg debugfs_port_reg = {
|
2018-12-19 15:56:43 +00:00
|
|
|
.lu = debugfs_port_reg_lu,
|
|
|
|
.count = 0x100,
|
2018-12-19 15:56:41 +00:00
|
|
|
.base_off = PORT_BASE,
|
2018-12-19 15:56:40 +00:00
|
|
|
};
|
|
|
|
|
2018-12-19 15:56:42 +00:00
|
|
|
static const struct hisi_sas_debugfs_reg_lu debugfs_global_reg_lu[] = {
|
|
|
|
HISI_SAS_DEBUGFS_REG(DLVRY_QUEUE_ENABLE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_CONTEXT),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_STATE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_PORT_NUM_MA),
|
|
|
|
HISI_SAS_DEBUGFS_REG(PHY_CONN_RATE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ITCT_CLR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(IO_SATA_BROKEN_MSG_ADDR_LO),
|
|
|
|
HISI_SAS_DEBUGFS_REG(IO_SATA_BROKEN_MSG_ADDR_HI),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SATA_INITI_D2H_STORE_ADDR_LO),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SATA_INITI_D2H_STORE_ADDR_HI),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CFG_MAX_TAG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_SAS_TX_OPEN_FAIL_RETRY_CTRL),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_SAS_TXFAIL_RETRY_CTRL),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_GET_ITV_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(DEVICE_MSG_WORK_MODE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(OPENA_WT_CONTI_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(I_T_NEXUS_LOSS_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(MAX_CON_TIME_LIMIT_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(BUS_INACTIVE_LIMIT_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(REJECT_TO_OPEN_LIMIT_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CQ_INT_CONVERGE_EN),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CFG_AGING_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_DFX_CFG2),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CFG_ABT_SET_QUERY_IPTT),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CFG_ABT_SET_IPTT_DONE),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_IOMB_PROC1_STATUS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHNL_INT_STATUS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_AXI_FIFO_ERR_INFO),
|
|
|
|
HISI_SAS_DEBUGFS_REG(INT_COAL_EN),
|
|
|
|
HISI_SAS_DEBUGFS_REG(OQ_INT_COAL_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(OQ_INT_COAL_CNT),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_COAL_TIME),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_COAL_CNT),
|
|
|
|
HISI_SAS_DEBUGFS_REG(OQ_INT_SRC),
|
|
|
|
HISI_SAS_DEBUGFS_REG(OQ_INT_SRC_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_SRC1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_SRC2),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_SRC3),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_SRC_MSK1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_SRC_MSK2),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ENT_INT_SRC_MSK3),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHNL_PHYUPDOWN_INT_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CHNL_ENT_INT_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_COM_INT_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_ECC_INTR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_ECC_INTR_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HGC_ERR_STAT_EN),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CQE_SEND_CNT),
|
|
|
|
HISI_SAS_DEBUGFS_REG(DLVRY_Q_0_DEPTH),
|
|
|
|
HISI_SAS_DEBUGFS_REG(DLVRY_Q_0_WR_PTR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(DLVRY_Q_0_RD_PTR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HYPER_STREAM_ID_EN_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(OQ0_INT_SRC_MSK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(COMPL_Q_0_DEPTH),
|
|
|
|
HISI_SAS_DEBUGFS_REG(COMPL_Q_0_WR_PTR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(COMPL_Q_0_RD_PTR),
|
|
|
|
HISI_SAS_DEBUGFS_REG(AWQOS_AWCACHE_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(ARQOS_ARCACHE_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(HILINK_ERR_DFX),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_GPIO_CFG_0),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_GPIO_CFG_1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_GPIO_TX_0_1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_CFG_DRIVE_VLD),
|
|
|
|
{}
|
|
|
|
};
|
|
|
|
|
2018-12-19 15:56:40 +00:00
|
|
|
static const struct hisi_sas_debugfs_reg debugfs_global_reg = {
|
2018-12-19 15:56:42 +00:00
|
|
|
.lu = debugfs_global_reg_lu,
|
|
|
|
.count = 0x800,
|
2018-12-19 15:56:40 +00:00
|
|
|
};
|
|
|
|
|
2019-08-05 13:48:02 +00:00
|
|
|
static const struct hisi_sas_debugfs_reg_lu debugfs_axi_reg_lu[] = {
|
|
|
|
HISI_SAS_DEBUGFS_REG(AM_CFG_MAX_TRANS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(AM_CFG_SINGLE_PORT_MAX_TRANS),
|
|
|
|
HISI_SAS_DEBUGFS_REG(AXI_CFG),
|
|
|
|
HISI_SAS_DEBUGFS_REG(AM_ROB_ECC_ERR_ADDR),
|
|
|
|
{}
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct hisi_sas_debugfs_reg debugfs_axi_reg = {
|
|
|
|
.lu = debugfs_axi_reg_lu,
|
|
|
|
.count = 0x61,
|
|
|
|
.base_off = AXI_MASTER_CFG_BASE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct hisi_sas_debugfs_reg_lu debugfs_ras_reg_lu[] = {
|
2020-05-15 14:13:44 +00:00
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RAS_INTR0),
|
2019-08-05 13:48:02 +00:00
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RAS_INTR1),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RAS_INTR0_MASK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RAS_INTR1_MASK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(CFG_SAS_RAS_INTR_MASK),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RAS_INTR2),
|
|
|
|
HISI_SAS_DEBUGFS_REG(SAS_RAS_INTR2_MASK),
|
|
|
|
{}
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct hisi_sas_debugfs_reg debugfs_ras_reg = {
|
|
|
|
.lu = debugfs_ras_reg_lu,
|
|
|
|
.count = 0x10,
|
|
|
|
.base_off = RAS_BASE,
|
|
|
|
};
|
|
|
|
|
2018-12-19 15:56:41 +00:00
|
|
|
static void debugfs_snapshot_prepare_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
set_bit(HISI_SAS_REJECT_CMD_BIT, &hisi_hba->flags);
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_QUEUE_ENABLE, 0);
|
|
|
|
|
2019-09-06 12:55:34 +00:00
|
|
|
wait_cmds_complete_timeout_v3_hw(hisi_hba, 100, 5000);
|
2018-12-19 15:56:41 +00:00
|
|
|
|
2020-01-20 12:22:31 +00:00
|
|
|
hisi_sas_sync_irqs(hisi_hba);
|
2018-12-19 15:56:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_restore_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
hisi_sas_write32(hisi_hba, DLVRY_QUEUE_ENABLE,
|
|
|
|
(u32)((1ULL << hisi_hba->queue_count) - 1));
|
|
|
|
|
|
|
|
clear_bit(HISI_SAS_REJECT_CMD_BIT, &hisi_hba->flags);
|
|
|
|
}
|
|
|
|
|
2019-08-05 13:48:01 +00:00
|
|
|
static void read_iost_itct_cache_v3_hw(struct hisi_hba *hisi_hba,
|
|
|
|
enum hisi_sas_debugfs_cache_type type,
|
|
|
|
u32 *cache)
|
|
|
|
{
|
|
|
|
u32 cache_dw_size = HISI_SAS_IOST_ITCT_CACHE_DW_SZ *
|
|
|
|
HISI_SAS_IOST_ITCT_CACHE_NUM;
|
2020-03-11 15:22:24 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
2019-08-05 13:48:01 +00:00
|
|
|
u32 *buf = cache;
|
|
|
|
u32 i, val;
|
|
|
|
|
|
|
|
hisi_sas_write32(hisi_hba, TAB_RD_TYPE, type);
|
|
|
|
|
|
|
|
for (i = 0; i < HISI_SAS_IOST_ITCT_CACHE_DW_SZ; i++) {
|
|
|
|
val = hisi_sas_read32(hisi_hba, TAB_DFX);
|
|
|
|
if (val == 0xffffffff)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (val != 0xffffffff) {
|
2020-03-11 15:22:24 +00:00
|
|
|
dev_err(dev, "Issue occurred in reading IOST/ITCT cache!\n");
|
2019-08-05 13:48:01 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
memset(buf, 0, cache_dw_size * 4);
|
|
|
|
buf[0] = val;
|
|
|
|
|
|
|
|
for (i = 1; i < cache_dw_size; i++)
|
|
|
|
buf[i] = hisi_sas_read32(hisi_hba, TAB_DFX);
|
|
|
|
}
|
|
|
|
|
2019-09-06 12:55:36 +00:00
|
|
|
static void hisi_sas_bist_test_prep_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
u32 reg_val;
|
2020-09-01 11:13:06 +00:00
|
|
|
int phy_no = hisi_hba->debugfs_bist_phy_no;
|
2020-09-01 11:13:07 +00:00
|
|
|
int i;
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* disable PHY */
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_enable(hisi_hba, phy_no, 0);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
2020-09-01 11:13:07 +00:00
|
|
|
/* update FFE */
|
|
|
|
for (i = 0; i < FFE_CFG_MAX; i++)
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, TXDEEMPH_G1 + (i * 0x4),
|
|
|
|
hisi_hba->debugfs_bist_ffe[phy_no][i]);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* disable ALOS */
|
2020-09-01 11:13:06 +00:00
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no, SERDES_CFG);
|
2019-09-06 12:55:36 +00:00
|
|
|
reg_val |= CFG_ALOS_CHK_DISABLE_MSK;
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SERDES_CFG, reg_val);
|
2019-09-06 12:55:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void hisi_sas_bist_test_restore_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
u32 reg_val;
|
2020-09-01 11:13:06 +00:00
|
|
|
int phy_no = hisi_hba->debugfs_bist_phy_no;
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* disable loopback */
|
2020-09-01 11:13:06 +00:00
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no, SAS_PHY_BIST_CTRL);
|
2019-09-06 12:55:36 +00:00
|
|
|
reg_val &= ~(CFG_RX_BIST_EN_MSK | CFG_TX_BIST_EN_MSK |
|
|
|
|
CFG_BIST_TEST_MSK);
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SAS_PHY_BIST_CTRL, reg_val);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* enable ALOS */
|
2020-09-01 11:13:06 +00:00
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no, SERDES_CFG);
|
2019-09-06 12:55:36 +00:00
|
|
|
reg_val &= ~CFG_ALOS_CHK_DISABLE_MSK;
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SERDES_CFG, reg_val);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* restore the linkrate */
|
2020-09-01 11:13:06 +00:00
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no, PROG_PHY_LINK_RATE);
|
2019-09-06 12:55:36 +00:00
|
|
|
/* init OOB link rate as 1.5 Gbits */
|
2020-09-01 11:13:04 +00:00
|
|
|
reg_val &= ~CFG_PROG_OOB_PHY_LINK_RATE_MSK;
|
|
|
|
reg_val |= (0x8 << CFG_PROG_OOB_PHY_LINK_RATE_OFF);
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PROG_PHY_LINK_RATE, reg_val);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* enable PHY */
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_enable(hisi_hba, phy_no, 1);
|
2019-09-06 12:55:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#define SAS_PHY_BIST_CODE_INIT 0x1
|
|
|
|
#define SAS_PHY_BIST_CODE1_INIT 0X80
|
|
|
|
static int debugfs_set_bist_v3_hw(struct hisi_hba *hisi_hba, bool enable)
|
|
|
|
{
|
|
|
|
u32 reg_val, mode_tmp;
|
|
|
|
u32 linkrate = hisi_hba->debugfs_bist_linkrate;
|
2020-09-01 11:13:06 +00:00
|
|
|
u32 phy_no = hisi_hba->debugfs_bist_phy_no;
|
2020-09-01 11:13:07 +00:00
|
|
|
u32 *ffe = hisi_hba->debugfs_bist_ffe[phy_no];
|
2019-09-06 12:55:36 +00:00
|
|
|
u32 code_mode = hisi_hba->debugfs_bist_code_mode;
|
|
|
|
u32 path_mode = hisi_hba->debugfs_bist_mode;
|
2020-09-01 11:13:08 +00:00
|
|
|
u32 *fix_code = &hisi_hba->debugfs_bist_fixed_code[0];
|
2019-09-06 12:55:36 +00:00
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
|
2020-09-01 11:13:08 +00:00
|
|
|
dev_info(dev, "BIST info:phy%d link_rate=%d code_mode=%d path_mode=%d ffe={0x%x, 0x%x, 0x%x, 0x%x, 0x%x, 0x%x, 0x%x} fixed_code={0x%x, 0x%x}\n",
|
2020-09-01 11:13:07 +00:00
|
|
|
phy_no, linkrate, code_mode, path_mode,
|
|
|
|
ffe[FFE_SAS_1_5_GBPS], ffe[FFE_SAS_3_0_GBPS],
|
|
|
|
ffe[FFE_SAS_6_0_GBPS], ffe[FFE_SAS_12_0_GBPS],
|
|
|
|
ffe[FFE_SATA_1_5_GBPS], ffe[FFE_SATA_3_0_GBPS],
|
2020-09-01 11:13:08 +00:00
|
|
|
ffe[FFE_SATA_6_0_GBPS], fix_code[FIXED_CODE],
|
|
|
|
fix_code[FIXED_CODE_1]);
|
2019-09-06 12:55:36 +00:00
|
|
|
mode_tmp = path_mode ? 2 : 1;
|
|
|
|
if (enable) {
|
|
|
|
/* some preparations before bist test */
|
|
|
|
hisi_sas_bist_test_prep_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
/* set linkrate of bit test*/
|
2020-09-01 11:13:06 +00:00
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no,
|
2019-09-06 12:55:36 +00:00
|
|
|
PROG_PHY_LINK_RATE);
|
2020-09-01 11:13:04 +00:00
|
|
|
reg_val &= ~CFG_PROG_OOB_PHY_LINK_RATE_MSK;
|
|
|
|
reg_val |= (linkrate << CFG_PROG_OOB_PHY_LINK_RATE_OFF);
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, PROG_PHY_LINK_RATE,
|
|
|
|
reg_val);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* set code mode of bit test */
|
2020-09-01 11:13:06 +00:00
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no,
|
2019-09-06 12:55:36 +00:00
|
|
|
SAS_PHY_BIST_CTRL);
|
2020-09-01 11:13:08 +00:00
|
|
|
reg_val &= ~(CFG_BIST_MODE_SEL_MSK | CFG_LOOP_TEST_MODE_MSK |
|
|
|
|
CFG_RX_BIST_EN_MSK | CFG_TX_BIST_EN_MSK |
|
|
|
|
CFG_BIST_TEST_MSK);
|
2019-09-06 12:55:36 +00:00
|
|
|
reg_val |= ((code_mode << CFG_BIST_MODE_SEL_OFF) |
|
|
|
|
(mode_tmp << CFG_LOOP_TEST_MODE_OFF) |
|
|
|
|
CFG_BIST_TEST_MSK);
|
2020-09-01 11:13:08 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SAS_PHY_BIST_CTRL,
|
|
|
|
reg_val);
|
2019-09-06 12:55:36 +00:00
|
|
|
|
|
|
|
/* set the bist init value */
|
2020-09-01 11:13:08 +00:00
|
|
|
if (code_mode == HISI_SAS_BIST_CODE_MODE_FIXED_DATA) {
|
|
|
|
reg_val = hisi_hba->debugfs_bist_fixed_code[0];
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no,
|
|
|
|
SAS_PHY_BIST_CODE, reg_val);
|
|
|
|
|
|
|
|
reg_val = hisi_hba->debugfs_bist_fixed_code[1];
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no,
|
|
|
|
SAS_PHY_BIST_CODE1, reg_val);
|
|
|
|
} else {
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no,
|
|
|
|
SAS_PHY_BIST_CODE,
|
|
|
|
SAS_PHY_BIST_CODE_INIT);
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no,
|
|
|
|
SAS_PHY_BIST_CODE1,
|
|
|
|
SAS_PHY_BIST_CODE1_INIT);
|
|
|
|
}
|
2019-09-06 12:55:36 +00:00
|
|
|
|
2019-10-24 14:08:09 +00:00
|
|
|
mdelay(100);
|
|
|
|
reg_val |= (CFG_RX_BIST_EN_MSK | CFG_TX_BIST_EN_MSK);
|
2020-09-01 11:13:08 +00:00
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, SAS_PHY_BIST_CTRL,
|
|
|
|
reg_val);
|
2019-10-24 14:08:09 +00:00
|
|
|
|
2019-09-06 12:55:36 +00:00
|
|
|
/* clear error bit */
|
|
|
|
mdelay(100);
|
2020-09-01 11:13:06 +00:00
|
|
|
hisi_sas_phy_read32(hisi_hba, phy_no, SAS_BIST_ERR_CNT);
|
2019-09-06 12:55:36 +00:00
|
|
|
} else {
|
|
|
|
/* disable bist test and recover it */
|
|
|
|
hisi_hba->debugfs_bist_cnt += hisi_sas_phy_read32(hisi_hba,
|
2020-09-01 11:13:06 +00:00
|
|
|
phy_no, SAS_BIST_ERR_CNT);
|
2019-09-06 12:55:36 +00:00
|
|
|
hisi_sas_bist_test_restore_v3_hw(hisi_hba);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-08-19 15:20:32 +00:00
|
|
|
static int hisi_sas_map_queues(struct Scsi_Host *shost)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = shost_priv(shost);
|
|
|
|
struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
|
|
|
|
|
|
|
|
return blk_mq_pci_map_queues(qmap, hisi_hba->pci_dev,
|
|
|
|
BASE_VECTORS_V3_HW);
|
|
|
|
}
|
|
|
|
|
2018-05-21 10:09:18 +00:00
|
|
|
static struct scsi_host_template sht_v3_hw = {
|
|
|
|
.name = DRV_NAME,
|
2020-05-12 11:32:58 +00:00
|
|
|
.proc_name = DRV_NAME,
|
2018-05-21 10:09:18 +00:00
|
|
|
.module = THIS_MODULE,
|
|
|
|
.queuecommand = sas_queuecommand,
|
2020-06-15 06:46:24 +00:00
|
|
|
.dma_need_drain = ata_scsi_dma_need_drain,
|
2018-05-21 10:09:18 +00:00
|
|
|
.target_alloc = sas_target_alloc,
|
2020-10-02 14:30:36 +00:00
|
|
|
.slave_configure = slave_configure_v3_hw,
|
2018-05-21 10:09:18 +00:00
|
|
|
.scan_finished = hisi_sas_scan_finished,
|
|
|
|
.scan_start = hisi_sas_scan_start,
|
2020-08-19 15:20:32 +00:00
|
|
|
.map_queues = hisi_sas_map_queues,
|
2018-05-21 10:09:18 +00:00
|
|
|
.change_queue_depth = sas_change_queue_depth,
|
|
|
|
.bios_param = sas_bios_param,
|
|
|
|
.this_id = -1,
|
2018-12-06 13:34:42 +00:00
|
|
|
.sg_tablesize = HISI_SAS_SGE_PAGE_CNT,
|
2019-02-06 10:52:51 +00:00
|
|
|
.sg_prot_tablesize = HISI_SAS_SGE_PAGE_CNT,
|
2018-05-21 10:09:18 +00:00
|
|
|
.max_sectors = SCSI_DEFAULT_MAX_SECTORS,
|
|
|
|
.eh_device_reset_handler = sas_eh_device_reset_handler,
|
|
|
|
.eh_target_reset_handler = sas_eh_target_reset_handler,
|
2021-10-12 12:26:25 +00:00
|
|
|
.slave_alloc = hisi_sas_slave_alloc,
|
2018-05-21 10:09:18 +00:00
|
|
|
.target_destroy = sas_target_destroy,
|
|
|
|
.ioctl = sas_ioctl,
|
2019-11-30 19:28:12 +00:00
|
|
|
#ifdef CONFIG_COMPAT
|
|
|
|
.compat_ioctl = sas_ioctl,
|
|
|
|
#endif
|
2021-10-12 23:35:32 +00:00
|
|
|
.shost_groups = host_v3_hw_groups,
|
2018-09-24 15:06:33 +00:00
|
|
|
.tag_alloc_policy = BLK_TAG_ALLOC_RR,
|
2019-04-11 12:46:36 +00:00
|
|
|
.host_reset = hisi_sas_host_reset,
|
2020-08-19 15:20:32 +00:00
|
|
|
.host_tagset = 1,
|
2018-05-21 10:09:18 +00:00
|
|
|
};
|
|
|
|
|
2017-06-14 15:33:20 +00:00
|
|
|
static const struct hisi_sas_hw hisi_sas_v3_hw = {
|
2017-06-14 15:33:28 +00:00
|
|
|
.setup_itct = setup_itct_v3_hw,
|
2017-06-14 15:33:30 +00:00
|
|
|
.get_wideport_bitmap = get_wideport_bitmap_v3_hw,
|
2017-06-14 15:33:21 +00:00
|
|
|
.complete_hdr_size = sizeof(struct hisi_sas_complete_v3_hdr),
|
2017-12-08 17:16:34 +00:00
|
|
|
.clear_itct = clear_itct_v3_hw,
|
2019-01-25 14:22:30 +00:00
|
|
|
.sl_notify_ssp = sl_notify_ssp_v3_hw,
|
2017-06-14 15:33:25 +00:00
|
|
|
.prep_ssp = prep_ssp_v3_hw,
|
2017-06-14 15:33:26 +00:00
|
|
|
.prep_smp = prep_smp_v3_hw,
|
2017-06-14 15:33:27 +00:00
|
|
|
.prep_stp = prep_ata_v3_hw,
|
2017-06-14 15:33:29 +00:00
|
|
|
.prep_abort = prep_abort_v3_hw,
|
2017-06-14 15:33:25 +00:00
|
|
|
.start_delivery = start_delivery_v3_hw,
|
2017-06-14 15:33:22 +00:00
|
|
|
.phys_init = phys_init_v3_hw,
|
2017-10-24 15:51:36 +00:00
|
|
|
.phy_start = start_phy_v3_hw,
|
2017-06-14 15:33:31 +00:00
|
|
|
.phy_disable = disable_phy_v3_hw,
|
|
|
|
.phy_hard_reset = phy_hard_reset_v3_hw,
|
|
|
|
.phy_get_max_linkrate = phy_get_max_linkrate_v3_hw,
|
2017-08-10 16:09:41 +00:00
|
|
|
.phy_set_linkrate = phy_set_linkrate_v3_hw,
|
2017-06-14 15:33:32 +00:00
|
|
|
.dereg_device = dereg_device_v3_hw,
|
2017-08-10 16:09:40 +00:00
|
|
|
.soft_reset = soft_reset_v3_hw,
|
|
|
|
.get_phys_state = get_phys_state_v3_hw,
|
2017-10-24 15:51:44 +00:00
|
|
|
.get_events = phy_get_events_v3_hw,
|
2018-05-21 10:09:16 +00:00
|
|
|
.write_gpio = write_gpio_v3_hw,
|
2018-05-21 10:09:20 +00:00
|
|
|
.wait_cmds_complete_timeout = wait_cmds_complete_timeout_v3_hw,
|
2021-04-06 11:48:29 +00:00
|
|
|
.debugfs_snapshot_regs = debugfs_snapshot_regs_v3_hw,
|
2017-06-14 15:33:20 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct Scsi_Host *
|
|
|
|
hisi_sas_shost_alloc_pci(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost;
|
|
|
|
struct hisi_hba *hisi_hba;
|
|
|
|
struct device *dev = &pdev->dev;
|
|
|
|
|
2018-05-21 10:09:18 +00:00
|
|
|
shost = scsi_host_alloc(&sht_v3_hw, sizeof(*hisi_hba));
|
2017-08-10 16:09:43 +00:00
|
|
|
if (!shost) {
|
|
|
|
dev_err(dev, "shost alloc failed\n");
|
|
|
|
return NULL;
|
|
|
|
}
|
2017-06-14 15:33:20 +00:00
|
|
|
hisi_hba = shost_priv(shost);
|
|
|
|
|
2017-10-24 15:51:45 +00:00
|
|
|
INIT_WORK(&hisi_hba->rst_work, hisi_sas_rst_work_handler);
|
2020-11-24 08:46:34 +00:00
|
|
|
INIT_WORK(&hisi_hba->debugfs_work, debugfs_work_handler_v3_hw);
|
2017-06-14 15:33:20 +00:00
|
|
|
hisi_hba->hw = &hisi_sas_v3_hw;
|
|
|
|
hisi_hba->pci_dev = pdev;
|
|
|
|
hisi_hba->dev = dev;
|
|
|
|
hisi_hba->shost = shost;
|
|
|
|
SHOST_TO_SAS_HA(shost) = &hisi_hba->sha;
|
|
|
|
|
2018-12-17 14:40:07 +00:00
|
|
|
if (prot_mask & ~HISI_SAS_PROT_MASK)
|
|
|
|
dev_err(dev, "unsupported protection mask 0x%x, using default (0x0)\n",
|
|
|
|
prot_mask);
|
|
|
|
else
|
|
|
|
hisi_hba->prot_mask = prot_mask;
|
|
|
|
|
2017-06-14 15:33:20 +00:00
|
|
|
if (hisi_sas_get_fw_info(hisi_hba) < 0)
|
|
|
|
goto err_out;
|
|
|
|
|
2019-01-25 14:22:33 +00:00
|
|
|
if (hisi_sas_alloc(hisi_hba)) {
|
2017-06-14 15:33:20 +00:00
|
|
|
hisi_sas_free(hisi_hba);
|
|
|
|
goto err_out;
|
|
|
|
}
|
|
|
|
|
|
|
|
return shost;
|
|
|
|
err_out:
|
2017-08-10 16:09:43 +00:00
|
|
|
scsi_host_put(shost);
|
2017-06-14 15:33:20 +00:00
|
|
|
dev_err(dev, "shost alloc failed\n");
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2020-11-24 08:46:34 +00:00
|
|
|
static void debugfs_snapshot_cq_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int queue_entry_size = hisi_hba->hw->complete_hdr_size;
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++)
|
|
|
|
memcpy(hisi_hba->debugfs_cq[dump_index][i].complete_hdr,
|
|
|
|
hisi_hba->complete_hdr[i],
|
|
|
|
HISI_SAS_QUEUE_SLOTS * queue_entry_size);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_dq_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int queue_entry_size = sizeof(struct hisi_sas_cmd_hdr);
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++) {
|
|
|
|
struct hisi_sas_cmd_hdr *debugfs_cmd_hdr, *cmd_hdr;
|
|
|
|
int j;
|
|
|
|
|
|
|
|
debugfs_cmd_hdr = hisi_hba->debugfs_dq[dump_index][i].hdr;
|
|
|
|
cmd_hdr = hisi_hba->cmd_hdr[i];
|
|
|
|
|
|
|
|
for (j = 0; j < HISI_SAS_QUEUE_SLOTS; j++)
|
|
|
|
memcpy(&debugfs_cmd_hdr[j], &cmd_hdr[j],
|
|
|
|
queue_entry_size);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_port_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
const struct hisi_sas_debugfs_reg *port = &debugfs_port_reg;
|
|
|
|
int i, phy_cnt;
|
|
|
|
u32 offset;
|
|
|
|
u32 *databuf;
|
|
|
|
|
|
|
|
for (phy_cnt = 0; phy_cnt < hisi_hba->n_phy; phy_cnt++) {
|
|
|
|
databuf = hisi_hba->debugfs_port_reg[dump_index][phy_cnt].data;
|
|
|
|
for (i = 0; i < port->count; i++, databuf++) {
|
|
|
|
offset = port->base_off + 4 * i;
|
|
|
|
*databuf = hisi_sas_phy_read32(hisi_hba, phy_cnt,
|
|
|
|
offset);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_global_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
u32 *databuf = hisi_hba->debugfs_regs[dump_index][DEBUGFS_GLOBAL].data;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < debugfs_axi_reg.count; i++, databuf++)
|
|
|
|
*databuf = hisi_sas_read32(hisi_hba, 4 * i);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_axi_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
u32 *databuf = hisi_hba->debugfs_regs[dump_index][DEBUGFS_AXI].data;
|
|
|
|
const struct hisi_sas_debugfs_reg *axi = &debugfs_axi_reg;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < axi->count; i++, databuf++)
|
|
|
|
*databuf = hisi_sas_read32(hisi_hba, 4 * i + axi->base_off);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_ras_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
u32 *databuf = hisi_hba->debugfs_regs[dump_index][DEBUGFS_RAS].data;
|
|
|
|
const struct hisi_sas_debugfs_reg *ras = &debugfs_ras_reg;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ras->count; i++, databuf++)
|
|
|
|
*databuf = hisi_sas_read32(hisi_hba, 4 * i + ras->base_off);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_itct_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
void *cachebuf = hisi_hba->debugfs_itct_cache[dump_index].cache;
|
|
|
|
void *databuf = hisi_hba->debugfs_itct[dump_index].itct;
|
|
|
|
struct hisi_sas_itct *itct;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
read_iost_itct_cache_v3_hw(hisi_hba, HISI_SAS_ITCT_CACHE, cachebuf);
|
|
|
|
|
|
|
|
itct = hisi_hba->itct;
|
|
|
|
|
|
|
|
for (i = 0; i < HISI_SAS_MAX_ITCT_ENTRIES; i++, itct++) {
|
|
|
|
memcpy(databuf, itct, sizeof(struct hisi_sas_itct));
|
|
|
|
databuf += sizeof(struct hisi_sas_itct);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_iost_reg_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
int max_command_entries = HISI_SAS_MAX_COMMANDS;
|
|
|
|
void *cachebuf = hisi_hba->debugfs_iost_cache[dump_index].cache;
|
|
|
|
void *databuf = hisi_hba->debugfs_iost[dump_index].iost;
|
|
|
|
struct hisi_sas_iost *iost;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
read_iost_itct_cache_v3_hw(hisi_hba, HISI_SAS_IOST_CACHE, cachebuf);
|
|
|
|
|
|
|
|
iost = hisi_hba->iost;
|
|
|
|
|
|
|
|
for (i = 0; i < max_command_entries; i++, iost++) {
|
|
|
|
memcpy(databuf, iost, sizeof(struct hisi_sas_iost));
|
|
|
|
databuf += sizeof(struct hisi_sas_iost);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static const char *
|
|
|
|
debugfs_to_reg_name_v3_hw(int off, int base_off,
|
|
|
|
const struct hisi_sas_debugfs_reg_lu *lu)
|
|
|
|
{
|
|
|
|
for (; lu->name; lu++) {
|
|
|
|
if (off == lu->off - base_off)
|
|
|
|
return lu->name;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_print_reg_v3_hw(u32 *regs_val, struct seq_file *s,
|
|
|
|
const struct hisi_sas_debugfs_reg *reg)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < reg->count; i++) {
|
|
|
|
int off = i * 4;
|
|
|
|
const char *name;
|
|
|
|
|
|
|
|
name = debugfs_to_reg_name_v3_hw(off, reg->base_off,
|
|
|
|
reg->lu);
|
|
|
|
|
|
|
|
if (name)
|
|
|
|
seq_printf(s, "0x%08x 0x%08x %s\n", off,
|
|
|
|
regs_val[i], name);
|
|
|
|
else
|
|
|
|
seq_printf(s, "0x%08x 0x%08x\n", off,
|
|
|
|
regs_val[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_global_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_regs *global = s->private;
|
|
|
|
|
|
|
|
debugfs_print_reg_v3_hw(global->data, s,
|
|
|
|
&debugfs_global_reg);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_global_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_axi_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_regs *axi = s->private;
|
|
|
|
|
|
|
|
debugfs_print_reg_v3_hw(axi->data, s,
|
|
|
|
&debugfs_axi_reg);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_axi_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_ras_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_regs *ras = s->private;
|
|
|
|
|
|
|
|
debugfs_print_reg_v3_hw(ras->data, s,
|
|
|
|
&debugfs_ras_reg);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_ras_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_port_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_port *port = s->private;
|
|
|
|
const struct hisi_sas_debugfs_reg *reg_port = &debugfs_port_reg;
|
|
|
|
|
|
|
|
debugfs_print_reg_v3_hw(port->data, s, reg_port);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_port_v3_hw);
|
|
|
|
|
|
|
|
static void debugfs_show_row_64_v3_hw(struct seq_file *s, int index,
|
|
|
|
int sz, __le64 *ptr)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
/* completion header size not fixed per HW version */
|
|
|
|
seq_printf(s, "index %04d:\n\t", index);
|
|
|
|
for (i = 1; i <= sz / 8; i++, ptr++) {
|
|
|
|
seq_printf(s, " 0x%016llx", le64_to_cpu(*ptr));
|
|
|
|
if (!(i % 2))
|
|
|
|
seq_puts(s, "\n\t");
|
|
|
|
}
|
|
|
|
|
|
|
|
seq_puts(s, "\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_show_row_32_v3_hw(struct seq_file *s, int index,
|
|
|
|
int sz, __le32 *ptr)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
/* completion header size not fixed per HW version */
|
|
|
|
seq_printf(s, "index %04d:\n\t", index);
|
|
|
|
for (i = 1; i <= sz / 4; i++, ptr++) {
|
|
|
|
seq_printf(s, " 0x%08x", le32_to_cpu(*ptr));
|
|
|
|
if (!(i % 4))
|
|
|
|
seq_puts(s, "\n\t");
|
|
|
|
}
|
|
|
|
seq_puts(s, "\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_cq_show_slot_v3_hw(struct seq_file *s, int slot,
|
|
|
|
struct hisi_sas_debugfs_cq *debugfs_cq)
|
|
|
|
{
|
|
|
|
struct hisi_sas_cq *cq = debugfs_cq->cq;
|
|
|
|
struct hisi_hba *hisi_hba = cq->hisi_hba;
|
|
|
|
__le32 *complete_hdr = debugfs_cq->complete_hdr +
|
|
|
|
(hisi_hba->hw->complete_hdr_size * slot);
|
|
|
|
|
|
|
|
debugfs_show_row_32_v3_hw(s, slot,
|
|
|
|
hisi_hba->hw->complete_hdr_size,
|
|
|
|
complete_hdr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_cq_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_cq *debugfs_cq = s->private;
|
|
|
|
int slot;
|
|
|
|
|
|
|
|
for (slot = 0; slot < HISI_SAS_QUEUE_SLOTS; slot++)
|
|
|
|
debugfs_cq_show_slot_v3_hw(s, slot, debugfs_cq);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_cq_v3_hw);
|
|
|
|
|
|
|
|
static void debugfs_dq_show_slot_v3_hw(struct seq_file *s, int slot,
|
|
|
|
void *dq_ptr)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_dq *debugfs_dq = dq_ptr;
|
|
|
|
void *cmd_queue = debugfs_dq->hdr;
|
|
|
|
__le32 *cmd_hdr = cmd_queue +
|
|
|
|
sizeof(struct hisi_sas_cmd_hdr) * slot;
|
|
|
|
|
|
|
|
debugfs_show_row_32_v3_hw(s, slot, sizeof(struct hisi_sas_cmd_hdr),
|
|
|
|
cmd_hdr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_dq_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
int slot;
|
|
|
|
|
|
|
|
for (slot = 0; slot < HISI_SAS_QUEUE_SLOTS; slot++)
|
|
|
|
debugfs_dq_show_slot_v3_hw(s, slot, s->private);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_dq_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_iost_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_iost *debugfs_iost = s->private;
|
|
|
|
struct hisi_sas_iost *iost = debugfs_iost->iost;
|
|
|
|
int i, max_command_entries = HISI_SAS_MAX_COMMANDS;
|
|
|
|
|
|
|
|
for (i = 0; i < max_command_entries; i++, iost++) {
|
|
|
|
__le64 *data = &iost->qw0;
|
|
|
|
|
|
|
|
debugfs_show_row_64_v3_hw(s, i, sizeof(*iost), data);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_iost_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_iost_cache_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_iost_cache *debugfs_iost_cache = s->private;
|
|
|
|
struct hisi_sas_iost_itct_cache *iost_cache =
|
|
|
|
debugfs_iost_cache->cache;
|
|
|
|
u32 cache_size = HISI_SAS_IOST_ITCT_CACHE_DW_SZ * 4;
|
|
|
|
int i, tab_idx;
|
|
|
|
__le64 *iost;
|
|
|
|
|
|
|
|
for (i = 0; i < HISI_SAS_IOST_ITCT_CACHE_NUM; i++, iost_cache++) {
|
|
|
|
/*
|
|
|
|
* Data struct of IOST cache:
|
|
|
|
* Data[1]: BIT0~15: Table index
|
|
|
|
* Bit16: Valid mask
|
|
|
|
* Data[2]~[9]: IOST table
|
|
|
|
*/
|
|
|
|
tab_idx = (iost_cache->data[1] & 0xffff);
|
|
|
|
iost = (__le64 *)iost_cache;
|
|
|
|
|
|
|
|
debugfs_show_row_64_v3_hw(s, tab_idx, cache_size, iost);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_iost_cache_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_itct_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
struct hisi_sas_debugfs_itct *debugfs_itct = s->private;
|
|
|
|
struct hisi_sas_itct *itct = debugfs_itct->itct;
|
|
|
|
|
|
|
|
for (i = 0; i < HISI_SAS_MAX_ITCT_ENTRIES; i++, itct++) {
|
|
|
|
__le64 *data = &itct->qw0;
|
|
|
|
|
|
|
|
debugfs_show_row_64_v3_hw(s, i, sizeof(*itct), data);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_itct_v3_hw);
|
|
|
|
|
|
|
|
static int debugfs_itct_cache_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_debugfs_itct_cache *debugfs_itct_cache = s->private;
|
|
|
|
struct hisi_sas_iost_itct_cache *itct_cache =
|
|
|
|
debugfs_itct_cache->cache;
|
|
|
|
u32 cache_size = HISI_SAS_IOST_ITCT_CACHE_DW_SZ * 4;
|
|
|
|
int i, tab_idx;
|
|
|
|
__le64 *itct;
|
|
|
|
|
|
|
|
for (i = 0; i < HISI_SAS_IOST_ITCT_CACHE_NUM; i++, itct_cache++) {
|
|
|
|
/*
|
|
|
|
* Data struct of ITCT cache:
|
|
|
|
* Data[1]: BIT0~15: Table index
|
|
|
|
* Bit16: Valid mask
|
|
|
|
* Data[2]~[9]: ITCT table
|
|
|
|
*/
|
|
|
|
tab_idx = itct_cache->data[1] & 0xffff;
|
|
|
|
itct = (__le64 *)itct_cache;
|
|
|
|
|
|
|
|
debugfs_show_row_64_v3_hw(s, tab_idx, cache_size, itct);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_itct_cache_v3_hw);
|
|
|
|
|
|
|
|
static void debugfs_create_files_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
u64 *debugfs_timestamp;
|
|
|
|
int dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
struct dentry *dump_dentry;
|
|
|
|
struct dentry *dentry;
|
|
|
|
char name[256];
|
|
|
|
int p;
|
|
|
|
int c;
|
|
|
|
int d;
|
|
|
|
|
|
|
|
snprintf(name, 256, "%d", dump_index);
|
|
|
|
|
|
|
|
dump_dentry = debugfs_create_dir(name, hisi_hba->debugfs_dump_dentry);
|
|
|
|
|
|
|
|
debugfs_timestamp = &hisi_hba->debugfs_timestamp[dump_index];
|
|
|
|
|
|
|
|
debugfs_create_u64("timestamp", 0400, dump_dentry,
|
|
|
|
debugfs_timestamp);
|
|
|
|
|
|
|
|
debugfs_create_file("global", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_regs[dump_index][DEBUGFS_GLOBAL],
|
|
|
|
&debugfs_global_v3_hw_fops);
|
|
|
|
|
|
|
|
/* Create port dir and files */
|
|
|
|
dentry = debugfs_create_dir("port", dump_dentry);
|
|
|
|
for (p = 0; p < hisi_hba->n_phy; p++) {
|
|
|
|
snprintf(name, 256, "%d", p);
|
|
|
|
|
|
|
|
debugfs_create_file(name, 0400, dentry,
|
|
|
|
&hisi_hba->debugfs_port_reg[dump_index][p],
|
|
|
|
&debugfs_port_v3_hw_fops);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Create CQ dir and files */
|
|
|
|
dentry = debugfs_create_dir("cq", dump_dentry);
|
|
|
|
for (c = 0; c < hisi_hba->queue_count; c++) {
|
|
|
|
snprintf(name, 256, "%d", c);
|
|
|
|
|
|
|
|
debugfs_create_file(name, 0400, dentry,
|
|
|
|
&hisi_hba->debugfs_cq[dump_index][c],
|
|
|
|
&debugfs_cq_v3_hw_fops);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Create DQ dir and files */
|
|
|
|
dentry = debugfs_create_dir("dq", dump_dentry);
|
|
|
|
for (d = 0; d < hisi_hba->queue_count; d++) {
|
|
|
|
snprintf(name, 256, "%d", d);
|
|
|
|
|
|
|
|
debugfs_create_file(name, 0400, dentry,
|
|
|
|
&hisi_hba->debugfs_dq[dump_index][d],
|
|
|
|
&debugfs_dq_v3_hw_fops);
|
|
|
|
}
|
|
|
|
|
|
|
|
debugfs_create_file("iost", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_iost[dump_index],
|
|
|
|
&debugfs_iost_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("iost_cache", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_iost_cache[dump_index],
|
|
|
|
&debugfs_iost_cache_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("itct", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_itct[dump_index],
|
|
|
|
&debugfs_itct_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("itct_cache", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_itct_cache[dump_index],
|
|
|
|
&debugfs_itct_cache_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("axi", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_regs[dump_index][DEBUGFS_AXI],
|
|
|
|
&debugfs_axi_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("ras", 0400, dump_dentry,
|
|
|
|
&hisi_hba->debugfs_regs[dump_index][DEBUGFS_RAS],
|
|
|
|
&debugfs_ras_v3_hw_fops);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_snapshot_regs_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
2021-04-06 11:48:29 +00:00
|
|
|
int debugfs_dump_index = hisi_hba->debugfs_dump_index;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
u64 timestamp = local_clock();
|
|
|
|
|
|
|
|
if (debugfs_dump_index >= hisi_sas_debugfs_dump_count) {
|
|
|
|
dev_warn(dev, "dump count exceeded!\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
do_div(timestamp, NSEC_PER_MSEC);
|
|
|
|
hisi_hba->debugfs_timestamp[debugfs_dump_index] = timestamp;
|
|
|
|
|
2020-11-24 08:46:34 +00:00
|
|
|
debugfs_snapshot_prepare_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
debugfs_snapshot_global_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_port_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_axi_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_ras_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_cq_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_dq_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_itct_reg_v3_hw(hisi_hba);
|
|
|
|
debugfs_snapshot_iost_reg_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
debugfs_create_files_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
debugfs_snapshot_restore_v3_hw(hisi_hba);
|
2021-08-24 10:01:00 +00:00
|
|
|
hisi_hba->debugfs_dump_index++;
|
2020-11-24 08:46:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t debugfs_trigger_dump_v3_hw_write(struct file *file,
|
|
|
|
const char __user *user_buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = file->f_inode->i_private;
|
|
|
|
char buf[8];
|
|
|
|
|
|
|
|
if (hisi_hba->debugfs_dump_index >= hisi_sas_debugfs_dump_count)
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
if (count > 8)
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
if (copy_from_user(buf, user_buf, count))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
if (buf[0] != '1')
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
queue_work(hisi_hba->wq, &hisi_hba->debugfs_work);
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_trigger_dump_v3_hw_fops = {
|
|
|
|
.write = &debugfs_trigger_dump_v3_hw_write,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
enum {
|
|
|
|
HISI_SAS_BIST_LOOPBACK_MODE_DIGITAL = 0,
|
|
|
|
HISI_SAS_BIST_LOOPBACK_MODE_SERDES,
|
|
|
|
HISI_SAS_BIST_LOOPBACK_MODE_REMOTE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct {
|
|
|
|
int value;
|
|
|
|
char *name;
|
|
|
|
} debugfs_loop_linkrate_v3_hw[] = {
|
|
|
|
{ SAS_LINK_RATE_1_5_GBPS, "1.5 Gbit" },
|
|
|
|
{ SAS_LINK_RATE_3_0_GBPS, "3.0 Gbit" },
|
|
|
|
{ SAS_LINK_RATE_6_0_GBPS, "6.0 Gbit" },
|
|
|
|
{ SAS_LINK_RATE_12_0_GBPS, "12.0 Gbit" },
|
|
|
|
};
|
|
|
|
|
|
|
|
static int debugfs_bist_linkrate_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = s->private;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(debugfs_loop_linkrate_v3_hw); i++) {
|
|
|
|
int match = (hisi_hba->debugfs_bist_linkrate ==
|
|
|
|
debugfs_loop_linkrate_v3_hw[i].value);
|
|
|
|
|
|
|
|
seq_printf(s, "%s%s%s ", match ? "[" : "",
|
|
|
|
debugfs_loop_linkrate_v3_hw[i].name,
|
|
|
|
match ? "]" : "");
|
|
|
|
}
|
|
|
|
seq_puts(s, "\n");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t debugfs_bist_linkrate_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *m = filp->private_data;
|
|
|
|
struct hisi_hba *hisi_hba = m->private;
|
|
|
|
char kbuf[16] = {}, *pkbuf;
|
|
|
|
bool found = false;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (hisi_hba->debugfs_bist_enable)
|
|
|
|
return -EPERM;
|
|
|
|
|
|
|
|
if (count >= sizeof(kbuf))
|
|
|
|
return -EOVERFLOW;
|
|
|
|
|
|
|
|
if (copy_from_user(kbuf, buf, count))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
pkbuf = strstrip(kbuf);
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(debugfs_loop_linkrate_v3_hw); i++) {
|
|
|
|
if (!strncmp(debugfs_loop_linkrate_v3_hw[i].name,
|
|
|
|
pkbuf, 16)) {
|
|
|
|
hisi_hba->debugfs_bist_linkrate =
|
|
|
|
debugfs_loop_linkrate_v3_hw[i].value;
|
|
|
|
found = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!found)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_linkrate_v3_hw_open(struct inode *inode,
|
|
|
|
struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_bist_linkrate_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_bist_linkrate_v3_hw_fops = {
|
|
|
|
.open = debugfs_bist_linkrate_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_bist_linkrate_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct {
|
|
|
|
int value;
|
|
|
|
char *name;
|
|
|
|
} debugfs_loop_code_mode_v3_hw[] = {
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_PRBS7, "PRBS7" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_PRBS23, "PRBS23" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_PRBS31, "PRBS31" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_JTPAT, "JTPAT" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_CJTPAT, "CJTPAT" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_SCRAMBED_0, "SCRAMBED_0" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_TRAIN, "TRAIN" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_TRAIN_DONE, "TRAIN_DONE" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_HFTP, "HFTP" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_MFTP, "MFTP" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_LFTP, "LFTP" },
|
|
|
|
{ HISI_SAS_BIST_CODE_MODE_FIXED_DATA, "FIXED_DATA" },
|
|
|
|
};
|
|
|
|
|
|
|
|
static int debugfs_bist_code_mode_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = s->private;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(debugfs_loop_code_mode_v3_hw); i++) {
|
|
|
|
int match = (hisi_hba->debugfs_bist_code_mode ==
|
|
|
|
debugfs_loop_code_mode_v3_hw[i].value);
|
|
|
|
|
|
|
|
seq_printf(s, "%s%s%s ", match ? "[" : "",
|
|
|
|
debugfs_loop_code_mode_v3_hw[i].name,
|
|
|
|
match ? "]" : "");
|
|
|
|
}
|
|
|
|
seq_puts(s, "\n");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t debugfs_bist_code_mode_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count,
|
|
|
|
loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *m = filp->private_data;
|
|
|
|
struct hisi_hba *hisi_hba = m->private;
|
|
|
|
char kbuf[16] = {}, *pkbuf;
|
|
|
|
bool found = false;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (hisi_hba->debugfs_bist_enable)
|
|
|
|
return -EPERM;
|
|
|
|
|
|
|
|
if (count >= sizeof(kbuf))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (copy_from_user(kbuf, buf, count))
|
|
|
|
return -EOVERFLOW;
|
|
|
|
|
|
|
|
pkbuf = strstrip(kbuf);
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(debugfs_loop_code_mode_v3_hw); i++) {
|
|
|
|
if (!strncmp(debugfs_loop_code_mode_v3_hw[i].name,
|
|
|
|
pkbuf, 16)) {
|
|
|
|
hisi_hba->debugfs_bist_code_mode =
|
|
|
|
debugfs_loop_code_mode_v3_hw[i].value;
|
|
|
|
found = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!found)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_code_mode_v3_hw_open(struct inode *inode,
|
|
|
|
struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_bist_code_mode_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_bist_code_mode_v3_hw_fops = {
|
|
|
|
.open = debugfs_bist_code_mode_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_bist_code_mode_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static ssize_t debugfs_bist_phy_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *m = filp->private_data;
|
|
|
|
struct hisi_hba *hisi_hba = m->private;
|
|
|
|
unsigned int phy_no;
|
|
|
|
int val;
|
|
|
|
|
|
|
|
if (hisi_hba->debugfs_bist_enable)
|
|
|
|
return -EPERM;
|
|
|
|
|
|
|
|
val = kstrtouint_from_user(buf, count, 0, &phy_no);
|
|
|
|
if (val)
|
|
|
|
return val;
|
|
|
|
|
|
|
|
if (phy_no >= hisi_hba->n_phy)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
hisi_hba->debugfs_bist_phy_no = phy_no;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_phy_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = s->private;
|
|
|
|
|
|
|
|
seq_printf(s, "%d\n", hisi_hba->debugfs_bist_phy_no);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_phy_v3_hw_open(struct inode *inode,
|
|
|
|
struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_bist_phy_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_bist_phy_v3_hw_fops = {
|
|
|
|
.open = debugfs_bist_phy_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_bist_phy_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct {
|
|
|
|
int value;
|
|
|
|
char *name;
|
|
|
|
} debugfs_loop_modes_v3_hw[] = {
|
|
|
|
{ HISI_SAS_BIST_LOOPBACK_MODE_DIGITAL, "digital" },
|
|
|
|
{ HISI_SAS_BIST_LOOPBACK_MODE_SERDES, "serdes" },
|
|
|
|
{ HISI_SAS_BIST_LOOPBACK_MODE_REMOTE, "remote" },
|
|
|
|
};
|
|
|
|
|
|
|
|
static int debugfs_bist_mode_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = s->private;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(debugfs_loop_modes_v3_hw); i++) {
|
|
|
|
int match = (hisi_hba->debugfs_bist_mode ==
|
|
|
|
debugfs_loop_modes_v3_hw[i].value);
|
|
|
|
|
|
|
|
seq_printf(s, "%s%s%s ", match ? "[" : "",
|
|
|
|
debugfs_loop_modes_v3_hw[i].name,
|
|
|
|
match ? "]" : "");
|
|
|
|
}
|
|
|
|
seq_puts(s, "\n");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t debugfs_bist_mode_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *m = filp->private_data;
|
|
|
|
struct hisi_hba *hisi_hba = m->private;
|
|
|
|
char kbuf[16] = {}, *pkbuf;
|
|
|
|
bool found = false;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (hisi_hba->debugfs_bist_enable)
|
|
|
|
return -EPERM;
|
|
|
|
|
|
|
|
if (count >= sizeof(kbuf))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (copy_from_user(kbuf, buf, count))
|
|
|
|
return -EOVERFLOW;
|
|
|
|
|
|
|
|
pkbuf = strstrip(kbuf);
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(debugfs_loop_modes_v3_hw); i++) {
|
|
|
|
if (!strncmp(debugfs_loop_modes_v3_hw[i].name, pkbuf, 16)) {
|
|
|
|
hisi_hba->debugfs_bist_mode =
|
|
|
|
debugfs_loop_modes_v3_hw[i].value;
|
|
|
|
found = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!found)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_mode_v3_hw_open(struct inode *inode,
|
|
|
|
struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_bist_mode_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_bist_mode_v3_hw_fops = {
|
|
|
|
.open = debugfs_bist_mode_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_bist_mode_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static ssize_t debugfs_bist_enable_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *m = filp->private_data;
|
|
|
|
struct hisi_hba *hisi_hba = m->private;
|
|
|
|
unsigned int enable;
|
|
|
|
int val;
|
|
|
|
|
|
|
|
val = kstrtouint_from_user(buf, count, 0, &enable);
|
|
|
|
if (val)
|
|
|
|
return val;
|
|
|
|
|
|
|
|
if (enable > 1)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (enable == hisi_hba->debugfs_bist_enable)
|
|
|
|
return count;
|
|
|
|
|
|
|
|
val = debugfs_set_bist_v3_hw(hisi_hba, enable);
|
|
|
|
if (val < 0)
|
|
|
|
return val;
|
|
|
|
|
|
|
|
hisi_hba->debugfs_bist_enable = enable;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_enable_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = s->private;
|
|
|
|
|
|
|
|
seq_printf(s, "%d\n", hisi_hba->debugfs_bist_enable);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_bist_enable_v3_hw_open(struct inode *inode,
|
|
|
|
struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_bist_enable_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_bist_enable_v3_hw_fops = {
|
|
|
|
.open = debugfs_bist_enable_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_bist_enable_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct {
|
|
|
|
char *name;
|
|
|
|
} debugfs_ffe_name_v3_hw[FFE_CFG_MAX] = {
|
|
|
|
{ "SAS_1_5_GBPS" },
|
|
|
|
{ "SAS_3_0_GBPS" },
|
|
|
|
{ "SAS_6_0_GBPS" },
|
|
|
|
{ "SAS_12_0_GBPS" },
|
|
|
|
{ "FFE_RESV" },
|
|
|
|
{ "SATA_1_5_GBPS" },
|
|
|
|
{ "SATA_3_0_GBPS" },
|
|
|
|
{ "SATA_6_0_GBPS" },
|
|
|
|
};
|
|
|
|
|
|
|
|
static ssize_t debugfs_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *m = filp->private_data;
|
|
|
|
u32 *val = m->private;
|
|
|
|
int res;
|
|
|
|
|
|
|
|
res = kstrtouint_from_user(buf, count, 0, val);
|
|
|
|
if (res)
|
|
|
|
return res;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
u32 *val = s->private;
|
|
|
|
|
|
|
|
seq_printf(s, "0x%x\n", *val);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_v3_hw_open(struct inode *inode, struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_v3_hw_fops = {
|
|
|
|
.open = debugfs_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static ssize_t debugfs_phy_down_cnt_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct seq_file *s = filp->private_data;
|
|
|
|
struct hisi_sas_phy *phy = s->private;
|
|
|
|
unsigned int set_val;
|
|
|
|
int res;
|
|
|
|
|
|
|
|
res = kstrtouint_from_user(buf, count, 0, &set_val);
|
|
|
|
if (res)
|
|
|
|
return res;
|
|
|
|
|
|
|
|
if (set_val > 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
atomic_set(&phy->down_cnt, 0);
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_phy_down_cnt_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_phy *phy = s->private;
|
|
|
|
|
|
|
|
seq_printf(s, "%d\n", atomic_read(&phy->down_cnt));
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_phy_down_cnt_v3_hw_open(struct inode *inode,
|
|
|
|
struct file *filp)
|
|
|
|
{
|
|
|
|
return single_open(filp, debugfs_phy_down_cnt_v3_hw_show,
|
|
|
|
inode->i_private);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_phy_down_cnt_v3_hw_fops = {
|
|
|
|
.open = debugfs_phy_down_cnt_v3_hw_open,
|
|
|
|
.read = seq_read,
|
|
|
|
.write = debugfs_phy_down_cnt_v3_hw_write,
|
|
|
|
.llseek = seq_lseek,
|
|
|
|
.release = single_release,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
2021-01-26 11:04:28 +00:00
|
|
|
enum fifo_dump_mode_v3_hw {
|
|
|
|
FIFO_DUMP_FORVER = (1U << 0),
|
|
|
|
FIFO_DUMP_AFTER_TRIGGER = (1U << 1),
|
|
|
|
FIFO_DUMP_UNTILL_TRIGGER = (1U << 2),
|
|
|
|
};
|
|
|
|
|
|
|
|
enum fifo_trigger_mode_v3_hw {
|
|
|
|
FIFO_TRIGGER_EDGE = (1U << 0),
|
|
|
|
FIFO_TRIGGER_SAME_LEVEL = (1U << 1),
|
|
|
|
FIFO_TRIGGER_DIFF_LEVEL = (1U << 2),
|
|
|
|
};
|
|
|
|
|
|
|
|
static int debugfs_is_fifo_config_valid_v3_hw(struct hisi_sas_phy *phy)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = phy->hisi_hba;
|
|
|
|
|
|
|
|
if (phy->fifo.signal_sel > 0xf) {
|
|
|
|
dev_info(hisi_hba->dev, "Invalid signal select: %u\n",
|
|
|
|
phy->fifo.signal_sel);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (phy->fifo.dump_mode) {
|
|
|
|
case FIFO_DUMP_FORVER:
|
|
|
|
case FIFO_DUMP_AFTER_TRIGGER:
|
|
|
|
case FIFO_DUMP_UNTILL_TRIGGER:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_info(hisi_hba->dev, "Invalid dump mode: %u\n",
|
|
|
|
phy->fifo.dump_mode);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* when FIFO_DUMP_FORVER, no need to check trigger_mode */
|
|
|
|
if (phy->fifo.dump_mode == FIFO_DUMP_FORVER)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
switch (phy->fifo.trigger_mode) {
|
|
|
|
case FIFO_TRIGGER_EDGE:
|
|
|
|
case FIFO_TRIGGER_SAME_LEVEL:
|
|
|
|
case FIFO_TRIGGER_DIFF_LEVEL:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
dev_info(hisi_hba->dev, "Invalid trigger mode: %u\n",
|
|
|
|
phy->fifo.trigger_mode);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_update_fifo_config_v3_hw(struct hisi_sas_phy *phy)
|
|
|
|
{
|
|
|
|
u32 trigger_mode = phy->fifo.trigger_mode;
|
|
|
|
u32 signal_sel = phy->fifo.signal_sel;
|
|
|
|
u32 dump_mode = phy->fifo.dump_mode;
|
|
|
|
struct hisi_hba *hisi_hba = phy->hisi_hba;
|
|
|
|
int phy_no = phy->sas_phy.id;
|
|
|
|
u32 reg_val;
|
|
|
|
int res;
|
|
|
|
|
|
|
|
/* Check the validity of trace FIFO configuration */
|
|
|
|
res = debugfs_is_fifo_config_valid_v3_hw(phy);
|
|
|
|
if (res)
|
|
|
|
return res;
|
|
|
|
|
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
/* Disable trace FIFO before update configuration */
|
|
|
|
reg_val |= DFX_FIFO_CTRL_DUMP_DISABLE_MSK;
|
|
|
|
|
|
|
|
/* Update trace FIFO configuration */
|
|
|
|
reg_val &= ~(DFX_FIFO_CTRL_DUMP_MODE_MSK |
|
|
|
|
DFX_FIFO_CTRL_SIGNAL_SEL_MSK |
|
|
|
|
DFX_FIFO_CTRL_TRIGGER_MODE_MSK);
|
|
|
|
|
|
|
|
reg_val |= ((trigger_mode << DFX_FIFO_CTRL_TRIGGER_MODE_OFF) |
|
|
|
|
(dump_mode << DFX_FIFO_CTRL_DUMP_MODE_OFF) |
|
|
|
|
(signal_sel << DFX_FIFO_CTRL_SIGNAL_SEL_OFF));
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_CTRL, reg_val);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_DUMP_MSK,
|
|
|
|
phy->fifo.dump_msk);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_TRIGGER,
|
|
|
|
phy->fifo.trigger);
|
|
|
|
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_TRIGGER_MSK,
|
|
|
|
phy->fifo.trigger_msk);
|
|
|
|
|
|
|
|
/* Enable trace FIFO after updated configuration */
|
|
|
|
reg_val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
reg_val &= ~DFX_FIFO_CTRL_DUMP_DISABLE_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_CTRL, reg_val);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t debugfs_fifo_update_cfg_v3_hw_write(struct file *filp,
|
|
|
|
const char __user *buf,
|
|
|
|
size_t count, loff_t *ppos)
|
|
|
|
{
|
|
|
|
struct hisi_sas_phy *phy = filp->private_data;
|
|
|
|
bool update;
|
|
|
|
int val;
|
|
|
|
|
|
|
|
val = kstrtobool_from_user(buf, count, &update);
|
|
|
|
if (val)
|
|
|
|
return val;
|
|
|
|
|
|
|
|
if (update != 1)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
val = debugfs_update_fifo_config_v3_hw(phy);
|
|
|
|
if (val)
|
|
|
|
return val;
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct file_operations debugfs_fifo_update_cfg_v3_hw_fops = {
|
|
|
|
.open = simple_open,
|
|
|
|
.write = debugfs_fifo_update_cfg_v3_hw_write,
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
};
|
|
|
|
|
|
|
|
static void debugfs_read_fifo_data_v3_hw(struct hisi_sas_phy *phy)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba = phy->hisi_hba;
|
|
|
|
u32 *buf = phy->fifo.rd_data;
|
|
|
|
int phy_no = phy->sas_phy.id;
|
|
|
|
u32 val;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
memset(buf, 0, sizeof(phy->fifo.rd_data));
|
|
|
|
|
|
|
|
/* Disable trace FIFO before read data */
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
val |= DFX_FIFO_CTRL_DUMP_DISABLE_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_CTRL, val);
|
|
|
|
|
|
|
|
for (i = 0; i < HISI_SAS_FIFO_DATA_DW_SIZE; i++) {
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no,
|
|
|
|
DFX_FIFO_RD_DATA);
|
|
|
|
buf[i] = val;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Enable trace FIFO after read data */
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
val &= ~DFX_FIFO_CTRL_DUMP_DISABLE_MSK;
|
|
|
|
hisi_sas_phy_write32(hisi_hba, phy_no, DFX_FIFO_CTRL, val);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int debugfs_fifo_data_v3_hw_show(struct seq_file *s, void *p)
|
|
|
|
{
|
|
|
|
struct hisi_sas_phy *phy = s->private;
|
|
|
|
|
|
|
|
debugfs_read_fifo_data_v3_hw(phy);
|
|
|
|
|
|
|
|
debugfs_show_row_32_v3_hw(s, 0, HISI_SAS_FIFO_DATA_DW_SIZE * 4,
|
|
|
|
phy->fifo.rd_data);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
DEFINE_SHOW_ATTRIBUTE(debugfs_fifo_data_v3_hw);
|
|
|
|
|
|
|
|
static void debugfs_fifo_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
int phy_no;
|
|
|
|
|
|
|
|
hisi_hba->debugfs_fifo_dentry =
|
|
|
|
debugfs_create_dir("fifo", hisi_hba->debugfs_dir);
|
|
|
|
|
|
|
|
for (phy_no = 0; phy_no < hisi_hba->n_phy; phy_no++) {
|
|
|
|
struct hisi_sas_phy *phy = &hisi_hba->phy[phy_no];
|
|
|
|
struct dentry *port_dentry;
|
|
|
|
char name[256];
|
|
|
|
u32 val;
|
|
|
|
|
|
|
|
/* get default configuration for trace FIFO */
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
val &= DFX_FIFO_CTRL_DUMP_MODE_MSK;
|
|
|
|
val >>= DFX_FIFO_CTRL_DUMP_MODE_OFF;
|
|
|
|
phy->fifo.dump_mode = val;
|
|
|
|
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
val &= DFX_FIFO_CTRL_TRIGGER_MODE_MSK;
|
|
|
|
val >>= DFX_FIFO_CTRL_TRIGGER_MODE_OFF;
|
|
|
|
phy->fifo.trigger_mode = val;
|
|
|
|
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_CTRL);
|
|
|
|
val &= DFX_FIFO_CTRL_SIGNAL_SEL_MSK;
|
|
|
|
val >>= DFX_FIFO_CTRL_SIGNAL_SEL_OFF;
|
|
|
|
phy->fifo.signal_sel = val;
|
|
|
|
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_DUMP_MSK);
|
|
|
|
phy->fifo.dump_msk = val;
|
|
|
|
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_TRIGGER);
|
|
|
|
phy->fifo.trigger = val;
|
|
|
|
val = hisi_sas_phy_read32(hisi_hba, phy_no, DFX_FIFO_TRIGGER_MSK);
|
|
|
|
phy->fifo.trigger_msk = val;
|
|
|
|
|
|
|
|
snprintf(name, 256, "%d", phy_no);
|
|
|
|
port_dentry = debugfs_create_dir(name,
|
|
|
|
hisi_hba->debugfs_fifo_dentry);
|
|
|
|
|
|
|
|
debugfs_create_file("update_config", 0200, port_dentry, phy,
|
|
|
|
&debugfs_fifo_update_cfg_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("signal_sel", 0600, port_dentry,
|
|
|
|
&phy->fifo.signal_sel,
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("dump_msk", 0600, port_dentry,
|
|
|
|
&phy->fifo.dump_msk,
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("dump_mode", 0600, port_dentry,
|
|
|
|
&phy->fifo.dump_mode,
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("trigger_mode", 0600, port_dentry,
|
|
|
|
&phy->fifo.trigger_mode,
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("trigger", 0600, port_dentry,
|
|
|
|
&phy->fifo.trigger,
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("trigger_msk", 0600, port_dentry,
|
|
|
|
&phy->fifo.trigger_msk,
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("fifo_data", 0400, port_dentry, phy,
|
|
|
|
&debugfs_fifo_data_v3_hw_fops);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-11-24 08:46:34 +00:00
|
|
|
static void debugfs_work_handler_v3_hw(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct hisi_hba *hisi_hba =
|
|
|
|
container_of(work, struct hisi_hba, debugfs_work);
|
|
|
|
|
|
|
|
debugfs_snapshot_regs_v3_hw(hisi_hba);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_release_v3_hw(struct hisi_hba *hisi_hba, int dump_index)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_iost_cache[dump_index].cache);
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_itct_cache[dump_index].cache);
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_iost[dump_index].iost);
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_itct[dump_index].itct);
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++)
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_dq[dump_index][i].hdr);
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->queue_count; i++)
|
|
|
|
devm_kfree(dev,
|
|
|
|
hisi_hba->debugfs_cq[dump_index][i].complete_hdr);
|
|
|
|
|
|
|
|
for (i = 0; i < DEBUGFS_REGS_NUM; i++)
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_regs[dump_index][i].data);
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->n_phy; i++)
|
|
|
|
devm_kfree(dev, hisi_hba->debugfs_port_reg[dump_index][i].data);
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct hisi_sas_debugfs_reg *debugfs_reg_array_v3_hw[DEBUGFS_REGS_NUM] = {
|
|
|
|
[DEBUGFS_GLOBAL] = &debugfs_global_reg,
|
|
|
|
[DEBUGFS_AXI] = &debugfs_axi_reg,
|
|
|
|
[DEBUGFS_RAS] = &debugfs_ras_reg,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int debugfs_alloc_v3_hw(struct hisi_hba *hisi_hba, int dump_index)
|
|
|
|
{
|
|
|
|
const struct hisi_sas_hw *hw = hisi_hba->hw;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int p, c, d, r, i;
|
|
|
|
size_t sz;
|
|
|
|
|
|
|
|
for (r = 0; r < DEBUGFS_REGS_NUM; r++) {
|
|
|
|
struct hisi_sas_debugfs_regs *regs =
|
|
|
|
&hisi_hba->debugfs_regs[dump_index][r];
|
|
|
|
|
|
|
|
sz = debugfs_reg_array_v3_hw[r]->count * 4;
|
|
|
|
regs->data = devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!regs->data)
|
|
|
|
goto fail;
|
|
|
|
regs->hisi_hba = hisi_hba;
|
|
|
|
}
|
|
|
|
|
|
|
|
sz = debugfs_port_reg.count * 4;
|
|
|
|
for (p = 0; p < hisi_hba->n_phy; p++) {
|
|
|
|
struct hisi_sas_debugfs_port *port =
|
|
|
|
&hisi_hba->debugfs_port_reg[dump_index][p];
|
|
|
|
|
|
|
|
port->data = devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!port->data)
|
|
|
|
goto fail;
|
|
|
|
port->phy = &hisi_hba->phy[p];
|
|
|
|
}
|
|
|
|
|
|
|
|
sz = hw->complete_hdr_size * HISI_SAS_QUEUE_SLOTS;
|
|
|
|
for (c = 0; c < hisi_hba->queue_count; c++) {
|
|
|
|
struct hisi_sas_debugfs_cq *cq =
|
|
|
|
&hisi_hba->debugfs_cq[dump_index][c];
|
|
|
|
|
|
|
|
cq->complete_hdr = devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!cq->complete_hdr)
|
|
|
|
goto fail;
|
|
|
|
cq->cq = &hisi_hba->cq[c];
|
|
|
|
}
|
|
|
|
|
|
|
|
sz = sizeof(struct hisi_sas_cmd_hdr) * HISI_SAS_QUEUE_SLOTS;
|
|
|
|
for (d = 0; d < hisi_hba->queue_count; d++) {
|
|
|
|
struct hisi_sas_debugfs_dq *dq =
|
|
|
|
&hisi_hba->debugfs_dq[dump_index][d];
|
|
|
|
|
|
|
|
dq->hdr = devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!dq->hdr)
|
|
|
|
goto fail;
|
|
|
|
dq->dq = &hisi_hba->dq[d];
|
|
|
|
}
|
|
|
|
|
|
|
|
sz = HISI_SAS_MAX_COMMANDS * sizeof(struct hisi_sas_iost);
|
|
|
|
|
|
|
|
hisi_hba->debugfs_iost[dump_index].iost =
|
|
|
|
devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!hisi_hba->debugfs_iost[dump_index].iost)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
sz = HISI_SAS_IOST_ITCT_CACHE_NUM *
|
|
|
|
sizeof(struct hisi_sas_iost_itct_cache);
|
|
|
|
|
|
|
|
hisi_hba->debugfs_iost_cache[dump_index].cache =
|
|
|
|
devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!hisi_hba->debugfs_iost_cache[dump_index].cache)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
sz = HISI_SAS_IOST_ITCT_CACHE_NUM *
|
|
|
|
sizeof(struct hisi_sas_iost_itct_cache);
|
|
|
|
|
|
|
|
hisi_hba->debugfs_itct_cache[dump_index].cache =
|
|
|
|
devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!hisi_hba->debugfs_itct_cache[dump_index].cache)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
/* New memory allocation must be locate before itct */
|
|
|
|
sz = HISI_SAS_MAX_ITCT_ENTRIES * sizeof(struct hisi_sas_itct);
|
|
|
|
|
|
|
|
hisi_hba->debugfs_itct[dump_index].itct =
|
|
|
|
devm_kmalloc(dev, sz, GFP_KERNEL);
|
|
|
|
if (!hisi_hba->debugfs_itct[dump_index].itct)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
|
|
for (i = 0; i < hisi_sas_debugfs_dump_count; i++)
|
|
|
|
debugfs_release_v3_hw(hisi_hba, i);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_phy_down_cnt_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct dentry *dir = debugfs_create_dir("phy_down_cnt",
|
|
|
|
hisi_hba->debugfs_dir);
|
|
|
|
char name[16];
|
|
|
|
int phy_no;
|
|
|
|
|
|
|
|
for (phy_no = 0; phy_no < hisi_hba->n_phy; phy_no++) {
|
|
|
|
snprintf(name, 16, "%d", phy_no);
|
|
|
|
debugfs_create_file(name, 0600, dir,
|
|
|
|
&hisi_hba->phy[phy_no],
|
|
|
|
&debugfs_phy_down_cnt_v3_hw_fops);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_bist_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct dentry *ports_dentry;
|
|
|
|
int phy_no;
|
|
|
|
|
|
|
|
hisi_hba->debugfs_bist_dentry =
|
|
|
|
debugfs_create_dir("bist", hisi_hba->debugfs_dir);
|
|
|
|
debugfs_create_file("link_rate", 0600,
|
|
|
|
hisi_hba->debugfs_bist_dentry, hisi_hba,
|
|
|
|
&debugfs_bist_linkrate_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("code_mode", 0600,
|
|
|
|
hisi_hba->debugfs_bist_dentry, hisi_hba,
|
|
|
|
&debugfs_bist_code_mode_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("fixed_code", 0600,
|
|
|
|
hisi_hba->debugfs_bist_dentry,
|
|
|
|
&hisi_hba->debugfs_bist_fixed_code[0],
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("fixed_code_1", 0600,
|
|
|
|
hisi_hba->debugfs_bist_dentry,
|
|
|
|
&hisi_hba->debugfs_bist_fixed_code[1],
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("phy_id", 0600, hisi_hba->debugfs_bist_dentry,
|
|
|
|
hisi_hba, &debugfs_bist_phy_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_u32("cnt", 0600, hisi_hba->debugfs_bist_dentry,
|
|
|
|
&hisi_hba->debugfs_bist_cnt);
|
|
|
|
|
|
|
|
debugfs_create_file("loopback_mode", 0600,
|
|
|
|
hisi_hba->debugfs_bist_dentry,
|
|
|
|
hisi_hba, &debugfs_bist_mode_v3_hw_fops);
|
|
|
|
|
|
|
|
debugfs_create_file("enable", 0600, hisi_hba->debugfs_bist_dentry,
|
|
|
|
hisi_hba, &debugfs_bist_enable_v3_hw_fops);
|
|
|
|
|
|
|
|
ports_dentry = debugfs_create_dir("port", hisi_hba->debugfs_bist_dentry);
|
|
|
|
|
|
|
|
for (phy_no = 0; phy_no < hisi_hba->n_phy; phy_no++) {
|
|
|
|
struct dentry *port_dentry;
|
|
|
|
struct dentry *ffe_dentry;
|
|
|
|
char name[256];
|
|
|
|
int i;
|
|
|
|
|
|
|
|
snprintf(name, 256, "%d", phy_no);
|
|
|
|
port_dentry = debugfs_create_dir(name, ports_dentry);
|
|
|
|
ffe_dentry = debugfs_create_dir("ffe", port_dentry);
|
|
|
|
for (i = 0; i < FFE_CFG_MAX; i++) {
|
|
|
|
if (i == FFE_RESV)
|
|
|
|
continue;
|
|
|
|
debugfs_create_file(debugfs_ffe_name_v3_hw[i].name,
|
|
|
|
0600, ffe_dentry,
|
|
|
|
&hisi_hba->debugfs_bist_ffe[phy_no][i],
|
|
|
|
&debugfs_v3_hw_fops);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_hba->debugfs_bist_linkrate = SAS_LINK_RATE_1_5_GBPS;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_init_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
hisi_hba->debugfs_dir = debugfs_create_dir(dev_name(dev),
|
|
|
|
hisi_sas_debugfs_dir);
|
|
|
|
debugfs_create_file("trigger_dump", 0200,
|
|
|
|
hisi_hba->debugfs_dir,
|
|
|
|
hisi_hba,
|
|
|
|
&debugfs_trigger_dump_v3_hw_fops);
|
|
|
|
|
|
|
|
/* create bist structures */
|
|
|
|
debugfs_bist_init_v3_hw(hisi_hba);
|
|
|
|
|
|
|
|
hisi_hba->debugfs_dump_dentry =
|
|
|
|
debugfs_create_dir("dump", hisi_hba->debugfs_dir);
|
|
|
|
|
|
|
|
debugfs_phy_down_cnt_init_v3_hw(hisi_hba);
|
2021-01-26 11:04:28 +00:00
|
|
|
debugfs_fifo_init_v3_hw(hisi_hba);
|
2020-11-24 08:46:34 +00:00
|
|
|
|
|
|
|
for (i = 0; i < hisi_sas_debugfs_dump_count; i++) {
|
|
|
|
if (debugfs_alloc_v3_hw(hisi_hba, i)) {
|
|
|
|
debugfs_remove_recursive(hisi_hba->debugfs_dir);
|
|
|
|
dev_dbg(dev, "failed to init debugfs!\n");
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void debugfs_exit_v3_hw(struct hisi_hba *hisi_hba)
|
|
|
|
{
|
|
|
|
debugfs_remove_recursive(hisi_hba->debugfs_dir);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:19 +00:00
|
|
|
static int
|
|
|
|
hisi_sas_v3_probe(struct pci_dev *pdev, const struct pci_device_id *id)
|
|
|
|
{
|
2017-06-14 15:33:20 +00:00
|
|
|
struct Scsi_Host *shost;
|
|
|
|
struct hisi_hba *hisi_hba;
|
|
|
|
struct device *dev = &pdev->dev;
|
|
|
|
struct asd_sas_phy **arr_phy;
|
|
|
|
struct asd_sas_port **arr_port;
|
|
|
|
struct sas_ha_struct *sha;
|
|
|
|
int rc, phy_nr, port_nr, i;
|
|
|
|
|
2021-08-24 10:00:56 +00:00
|
|
|
rc = pcim_enable_device(pdev);
|
2017-06-14 15:33:20 +00:00
|
|
|
if (rc)
|
|
|
|
goto err_out;
|
|
|
|
|
|
|
|
pci_set_master(pdev);
|
|
|
|
|
2021-08-24 10:00:56 +00:00
|
|
|
rc = pcim_iomap_regions(pdev, 1 << BAR_NO_V3_HW, DRV_NAME);
|
2017-06-14 15:33:20 +00:00
|
|
|
if (rc)
|
2021-08-24 10:00:56 +00:00
|
|
|
goto err_out;
|
2017-06-14 15:33:20 +00:00
|
|
|
|
2019-02-18 07:34:25 +00:00
|
|
|
rc = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
|
|
|
|
if (rc) {
|
2018-10-18 13:10:17 +00:00
|
|
|
dev_err(dev, "No usable DMA addressing method\n");
|
2019-02-18 07:34:25 +00:00
|
|
|
rc = -ENODEV;
|
2021-08-24 10:00:56 +00:00
|
|
|
goto err_out;
|
2017-06-14 15:33:20 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
shost = hisi_sas_shost_alloc_pci(pdev);
|
|
|
|
if (!shost) {
|
|
|
|
rc = -ENOMEM;
|
2021-08-24 10:00:56 +00:00
|
|
|
goto err_out;
|
2017-06-14 15:33:20 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
sha = SHOST_TO_SAS_HA(shost);
|
|
|
|
hisi_hba = shost_priv(shost);
|
|
|
|
dev_set_drvdata(dev, sha);
|
|
|
|
|
2021-08-24 10:00:56 +00:00
|
|
|
hisi_hba->regs = pcim_iomap_table(pdev)[BAR_NO_V3_HW];
|
2017-06-14 15:33:20 +00:00
|
|
|
if (!hisi_hba->regs) {
|
2019-04-11 12:46:44 +00:00
|
|
|
dev_err(dev, "cannot map register\n");
|
2017-06-14 15:33:20 +00:00
|
|
|
rc = -ENOMEM;
|
|
|
|
goto err_out_ha;
|
|
|
|
}
|
|
|
|
|
|
|
|
phy_nr = port_nr = hisi_hba->n_phy;
|
|
|
|
|
|
|
|
arr_phy = devm_kcalloc(dev, phy_nr, sizeof(void *), GFP_KERNEL);
|
|
|
|
arr_port = devm_kcalloc(dev, port_nr, sizeof(void *), GFP_KERNEL);
|
|
|
|
if (!arr_phy || !arr_port) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto err_out_ha;
|
|
|
|
}
|
|
|
|
|
|
|
|
sha->sas_phy = arr_phy;
|
|
|
|
sha->sas_port = arr_port;
|
|
|
|
sha->core.shost = shost;
|
|
|
|
sha->lldd_ha = hisi_hba;
|
|
|
|
|
|
|
|
shost->transportt = hisi_sas_stt;
|
|
|
|
shost->max_id = HISI_SAS_MAX_DEVICES;
|
|
|
|
shost->max_lun = ~0;
|
|
|
|
shost->max_channel = 1;
|
|
|
|
shost->max_cmd_len = 16;
|
2019-08-05 13:47:58 +00:00
|
|
|
shost->can_queue = HISI_SAS_UNRESERVED_IPTT;
|
|
|
|
shost->cmd_per_lun = HISI_SAS_UNRESERVED_IPTT;
|
2017-06-14 15:33:20 +00:00
|
|
|
|
|
|
|
sha->sas_ha_name = DRV_NAME;
|
|
|
|
sha->dev = dev;
|
|
|
|
sha->lldd_module = THIS_MODULE;
|
|
|
|
sha->sas_addr = &hisi_hba->sas_addr[0];
|
|
|
|
sha->num_phys = hisi_hba->n_phy;
|
|
|
|
|
|
|
|
for (i = 0; i < hisi_hba->n_phy; i++) {
|
|
|
|
sha->sas_phy[i] = &hisi_hba->phy[i].sas_phy;
|
|
|
|
sha->sas_port[i] = &hisi_hba->port[i].sas_port;
|
|
|
|
}
|
|
|
|
|
2019-01-10 13:32:41 +00:00
|
|
|
if (hisi_hba->prot_mask) {
|
|
|
|
dev_info(dev, "Registering for DIF/DIX prot_mask=0x%x\n",
|
|
|
|
prot_mask);
|
|
|
|
scsi_host_set_prot(hisi_hba->shost, prot_mask);
|
2019-03-10 00:53:47 +00:00
|
|
|
if (hisi_hba->prot_mask & HISI_SAS_DIX_PROT_MASK)
|
|
|
|
scsi_host_set_guard(hisi_hba->shost,
|
|
|
|
SHOST_DIX_GUARD_CRC);
|
2019-01-10 13:32:41 +00:00
|
|
|
}
|
|
|
|
|
2018-12-19 15:56:39 +00:00
|
|
|
if (hisi_sas_debugfs_enable)
|
2020-11-24 08:46:34 +00:00
|
|
|
debugfs_init_v3_hw(hisi_hba);
|
2018-12-19 15:56:39 +00:00
|
|
|
|
2020-08-19 15:20:32 +00:00
|
|
|
rc = interrupt_preinit_v3_hw(hisi_hba);
|
|
|
|
if (rc)
|
2020-11-24 08:46:33 +00:00
|
|
|
goto err_out_debugfs;
|
2021-08-24 10:00:57 +00:00
|
|
|
|
2017-06-14 15:33:20 +00:00
|
|
|
rc = scsi_add_host(shost, dev);
|
|
|
|
if (rc)
|
2020-11-24 08:46:33 +00:00
|
|
|
goto err_out_free_irq_vectors;
|
2017-06-14 15:33:20 +00:00
|
|
|
|
|
|
|
rc = sas_register_ha(sha);
|
|
|
|
if (rc)
|
|
|
|
goto err_out_register_ha;
|
|
|
|
|
2020-11-24 08:46:32 +00:00
|
|
|
rc = hisi_sas_v3_init(hisi_hba);
|
2017-06-14 15:33:20 +00:00
|
|
|
if (rc)
|
2021-04-06 11:48:28 +00:00
|
|
|
goto err_out_hw_init;
|
2017-06-14 15:33:20 +00:00
|
|
|
|
|
|
|
scsi_scan_host(shost);
|
|
|
|
|
2021-12-20 11:21:38 +00:00
|
|
|
pm_runtime_set_autosuspend_delay(dev, 5000);
|
|
|
|
pm_runtime_use_autosuspend(dev);
|
2020-10-02 14:30:34 +00:00
|
|
|
/*
|
|
|
|
* For the situation that there are ATA disks connected with SAS
|
|
|
|
* controller, it additionally creates ata_port which will affect the
|
|
|
|
* child_count of hisi_hba->dev. Even if suspended all the disks,
|
|
|
|
* ata_port is still and the child_count of hisi_hba->dev is not 0.
|
|
|
|
* So use pm_suspend_ignore_children() to ignore the effect to
|
|
|
|
* hisi_hba->dev.
|
|
|
|
*/
|
|
|
|
pm_suspend_ignore_children(dev, true);
|
|
|
|
pm_runtime_put_noidle(&pdev->dev);
|
|
|
|
|
2017-06-14 15:33:19 +00:00
|
|
|
return 0;
|
2017-06-14 15:33:20 +00:00
|
|
|
|
2021-04-06 11:48:28 +00:00
|
|
|
err_out_hw_init:
|
|
|
|
sas_unregister_ha(sha);
|
2017-06-14 15:33:20 +00:00
|
|
|
err_out_register_ha:
|
|
|
|
scsi_remove_host(shost);
|
2020-11-24 08:46:33 +00:00
|
|
|
err_out_free_irq_vectors:
|
|
|
|
pci_free_irq_vectors(pdev);
|
|
|
|
err_out_debugfs:
|
2020-11-24 08:46:34 +00:00
|
|
|
debugfs_exit_v3_hw(hisi_hba);
|
2017-06-14 15:33:20 +00:00
|
|
|
err_out_ha:
|
2020-11-24 08:46:33 +00:00
|
|
|
hisi_sas_free(hisi_hba);
|
2017-08-10 16:09:43 +00:00
|
|
|
scsi_host_put(shost);
|
2017-06-14 15:33:20 +00:00
|
|
|
err_out:
|
|
|
|
return rc;
|
2017-06-14 15:33:19 +00:00
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
static void
|
|
|
|
hisi_sas_v3_destroy_irqs(struct pci_dev *pdev, struct hisi_hba *hisi_hba)
|
|
|
|
{
|
2017-06-14 15:33:24 +00:00
|
|
|
int i;
|
|
|
|
|
2021-05-19 13:05:19 +00:00
|
|
|
devm_free_irq(&pdev->dev, pci_irq_vector(pdev, 1), hisi_hba);
|
|
|
|
devm_free_irq(&pdev->dev, pci_irq_vector(pdev, 2), hisi_hba);
|
|
|
|
devm_free_irq(&pdev->dev, pci_irq_vector(pdev, 11), hisi_hba);
|
scsi: hisi_sas: Issue internal abort on all relevant queues
To support queue mapped to a CPU, it needs to be ensured that issuing an
internal abort is safe, in that it is guaranteed that an internal abort is
processed for a single IO or a device after all the relevant command(s)
which it is attempting to abort have been processed by the controller.
Currently we only deliver commands for any device on a single queue to
solve this problem, as we know that commands issued on the same queue will
be processed in order, and we will not have a scenario where the internal
abort is racing against a command(s) which it is trying to abort.
To enqueue commands on queue mapped to a CPU, choosing a queue for an
command is based on the associated queue for the current CPU, so this is
not safe for internal abort since it would definitely not be guaranteed
that commands for the command devices are issued on the same queue.
To solve this issue, we take a bludgeoning approach, and issue a separate
internal abort on any queue(s) relevant to the command or device, in that
we will be guaranteed that at least one of these internal aborts will be
received last in the controller.
So, for aborting a single command, we can just force the internal abort to
be issued on the same queue as the command which we are trying to abort.
For aborting all commands associated with a device, we issue a separate
internal abort on all relevant queues. Issuing multiple internal aborts in
this fashion would have not side affect.
Signed-off-by: John Garry <john.garry@huawei.com>
Signed-off-by: Xiang Chen <chenxiang66@hisilicon.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2019-02-06 10:52:54 +00:00
|
|
|
for (i = 0; i < hisi_hba->cq_nvecs; i++) {
|
2017-06-14 15:33:24 +00:00
|
|
|
struct hisi_sas_cq *cq = &hisi_hba->cq[i];
|
2018-11-09 14:06:33 +00:00
|
|
|
int nr = hisi_sas_intr_conv ? 16 : 16 + i;
|
2017-06-14 15:33:24 +00:00
|
|
|
|
2021-05-19 13:05:19 +00:00
|
|
|
devm_free_irq(&pdev->dev, pci_irq_vector(pdev, nr), cq);
|
2017-06-14 15:33:24 +00:00
|
|
|
}
|
2017-06-14 15:33:23 +00:00
|
|
|
pci_free_irq_vectors(pdev);
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:19 +00:00
|
|
|
static void hisi_sas_v3_remove(struct pci_dev *pdev)
|
|
|
|
{
|
2017-06-14 15:33:20 +00:00
|
|
|
struct device *dev = &pdev->dev;
|
|
|
|
struct sas_ha_struct *sha = dev_get_drvdata(dev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
2017-08-10 16:09:43 +00:00
|
|
|
struct Scsi_Host *shost = sha->core.shost;
|
2017-06-14 15:33:20 +00:00
|
|
|
|
2020-10-02 14:30:34 +00:00
|
|
|
pm_runtime_get_noresume(dev);
|
2021-08-24 10:00:59 +00:00
|
|
|
del_timer_sync(&hisi_hba->timer);
|
2018-03-23 16:05:10 +00:00
|
|
|
|
2017-06-14 15:33:20 +00:00
|
|
|
sas_unregister_ha(sha);
|
2021-01-26 11:04:27 +00:00
|
|
|
flush_workqueue(hisi_hba->wq);
|
2017-06-14 15:33:20 +00:00
|
|
|
sas_remove_host(sha->core.shost);
|
|
|
|
|
2017-06-14 15:33:23 +00:00
|
|
|
hisi_sas_v3_destroy_irqs(pdev, hisi_hba);
|
2017-08-10 16:09:43 +00:00
|
|
|
hisi_sas_free(hisi_hba);
|
2020-11-24 08:46:34 +00:00
|
|
|
debugfs_exit_v3_hw(hisi_hba);
|
2017-08-10 16:09:43 +00:00
|
|
|
scsi_host_put(shost);
|
2017-06-14 15:33:19 +00:00
|
|
|
}
|
|
|
|
|
2018-07-18 14:14:30 +00:00
|
|
|
static void hisi_sas_reset_prepare_v3_hw(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
dev_info(dev, "FLR prepare\n");
|
2021-12-15 14:37:39 +00:00
|
|
|
down(&hisi_hba->sem);
|
2021-08-24 10:00:58 +00:00
|
|
|
set_bit(HISI_SAS_RESETTING_BIT, &hisi_hba->flags);
|
2018-07-18 14:14:30 +00:00
|
|
|
hisi_sas_controller_reset_prepare(hisi_hba);
|
|
|
|
|
|
|
|
rc = disable_host_v3_hw(hisi_hba);
|
|
|
|
if (rc)
|
|
|
|
dev_err(dev, "FLR: disable host failed rc=%d\n", rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hisi_sas_reset_done_v3_hw(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
hisi_sas_init_mem(hisi_hba);
|
|
|
|
|
|
|
|
rc = hw_init_v3_hw(hisi_hba);
|
|
|
|
if (rc) {
|
|
|
|
dev_err(dev, "FLR: hw init failed rc=%d\n", rc);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_sas_controller_reset_done(hisi_hba);
|
|
|
|
dev_info(dev, "FLR done\n");
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:19 +00:00
|
|
|
enum {
|
|
|
|
/* instances of the controller */
|
|
|
|
hip08,
|
|
|
|
};
|
|
|
|
|
2020-10-02 14:30:34 +00:00
|
|
|
static int _suspend_v3_hw(struct device *device)
|
2017-12-08 17:16:50 +00:00
|
|
|
{
|
2020-10-02 14:30:33 +00:00
|
|
|
struct pci_dev *pdev = to_pci_dev(device);
|
2017-12-08 17:16:50 +00:00
|
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
struct Scsi_Host *shost = hisi_hba->shost;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
if (!pdev->pm_cap) {
|
|
|
|
dev_err(dev, "PCI PM not supported\n");
|
|
|
|
return -ENODEV;
|
|
|
|
}
|
|
|
|
|
2021-08-24 10:00:58 +00:00
|
|
|
if (test_and_set_bit(HISI_SAS_RESETTING_BIT, &hisi_hba->flags))
|
2018-05-31 12:50:45 +00:00
|
|
|
return -1;
|
|
|
|
|
2021-12-20 11:21:31 +00:00
|
|
|
dev_warn(dev, "entering suspend state\n");
|
|
|
|
|
2017-12-08 17:16:50 +00:00
|
|
|
scsi_block_requests(shost);
|
|
|
|
set_bit(HISI_SAS_REJECT_CMD_BIT, &hisi_hba->flags);
|
|
|
|
flush_workqueue(hisi_hba->wq);
|
|
|
|
|
2018-07-18 14:14:29 +00:00
|
|
|
rc = disable_host_v3_hw(hisi_hba);
|
2017-12-08 17:16:50 +00:00
|
|
|
if (rc) {
|
2018-07-18 14:14:29 +00:00
|
|
|
dev_err(dev, "PM suspend: disable host failed rc=%d\n", rc);
|
2017-12-08 17:16:50 +00:00
|
|
|
clear_bit(HISI_SAS_REJECT_CMD_BIT, &hisi_hba->flags);
|
2021-08-24 10:00:58 +00:00
|
|
|
clear_bit(HISI_SAS_RESETTING_BIT, &hisi_hba->flags);
|
2017-12-08 17:16:50 +00:00
|
|
|
scsi_unblock_requests(shost);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
hisi_sas_init_mem(hisi_hba);
|
|
|
|
|
|
|
|
hisi_sas_release_tasks(hisi_hba);
|
|
|
|
|
|
|
|
sas_suspend_ha(sha);
|
2021-12-20 11:21:31 +00:00
|
|
|
|
|
|
|
dev_warn(dev, "end of suspending controller\n");
|
2017-12-08 17:16:50 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-10-02 14:30:34 +00:00
|
|
|
static int _resume_v3_hw(struct device *device)
|
2017-12-08 17:16:50 +00:00
|
|
|
{
|
2020-10-02 14:30:33 +00:00
|
|
|
struct pci_dev *pdev = to_pci_dev(device);
|
2017-12-08 17:16:50 +00:00
|
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
|
|
|
struct Scsi_Host *shost = hisi_hba->shost;
|
|
|
|
struct device *dev = hisi_hba->dev;
|
|
|
|
unsigned int rc;
|
2018-12-06 13:34:40 +00:00
|
|
|
pci_power_t device_state = pdev->current_state;
|
2017-12-08 17:16:50 +00:00
|
|
|
|
|
|
|
dev_warn(dev, "resuming from operating state [D%d]\n",
|
2019-04-11 12:46:44 +00:00
|
|
|
device_state);
|
2017-12-08 17:16:50 +00:00
|
|
|
|
|
|
|
scsi_unblock_requests(shost);
|
|
|
|
clear_bit(HISI_SAS_REJECT_CMD_BIT, &hisi_hba->flags);
|
|
|
|
|
|
|
|
sas_prep_resume_ha(sha);
|
2019-09-06 12:55:30 +00:00
|
|
|
rc = hw_init_v3_hw(hisi_hba);
|
|
|
|
if (rc) {
|
|
|
|
scsi_remove_host(shost);
|
2019-11-12 09:30:57 +00:00
|
|
|
return rc;
|
2019-09-06 12:55:30 +00:00
|
|
|
}
|
2020-11-24 08:46:32 +00:00
|
|
|
phys_init_v3_hw(hisi_hba);
|
2021-12-20 11:21:24 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If a directly-attached disk is removed during suspend, a deadlock
|
|
|
|
* may occur, as the PHYE_RESUME_TIMEOUT processing will require the
|
|
|
|
* hisi_hba->device to be active, which can only happen when resume
|
|
|
|
* completes. So don't wait for the HA event workqueue to drain upon
|
|
|
|
* resume.
|
|
|
|
*/
|
|
|
|
sas_resume_ha_no_sync(sha);
|
2021-08-24 10:00:58 +00:00
|
|
|
clear_bit(HISI_SAS_RESETTING_BIT, &hisi_hba->flags);
|
2017-12-08 17:16:50 +00:00
|
|
|
|
2021-12-20 11:21:31 +00:00
|
|
|
dev_warn(dev, "end of resuming controller\n");
|
|
|
|
|
2017-12-08 17:16:50 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-11-02 16:47:15 +00:00
|
|
|
static int __maybe_unused suspend_v3_hw(struct device *device)
|
2020-10-02 14:30:34 +00:00
|
|
|
{
|
|
|
|
struct pci_dev *pdev = to_pci_dev(device);
|
|
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
set_bit(HISI_SAS_PM_BIT, &hisi_hba->flags);
|
|
|
|
|
|
|
|
rc = _suspend_v3_hw(device);
|
|
|
|
if (rc)
|
|
|
|
clear_bit(HISI_SAS_PM_BIT, &hisi_hba->flags);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2020-11-02 16:47:15 +00:00
|
|
|
static int __maybe_unused resume_v3_hw(struct device *device)
|
2020-10-02 14:30:34 +00:00
|
|
|
{
|
|
|
|
struct pci_dev *pdev = to_pci_dev(device);
|
|
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
|
|
struct hisi_hba *hisi_hba = sha->lldd_ha;
|
|
|
|
int rc = _resume_v3_hw(device);
|
|
|
|
|
|
|
|
clear_bit(HISI_SAS_PM_BIT, &hisi_hba->flags);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2017-06-14 15:33:19 +00:00
|
|
|
static const struct pci_device_id sas_v3_pci_table[] = {
|
|
|
|
{ PCI_VDEVICE(HUAWEI, 0xa230), hip08 },
|
|
|
|
{}
|
|
|
|
};
|
2018-03-07 12:25:14 +00:00
|
|
|
MODULE_DEVICE_TABLE(pci, sas_v3_pci_table);
|
2017-06-14 15:33:19 +00:00
|
|
|
|
2017-12-08 17:16:40 +00:00
|
|
|
static const struct pci_error_handlers hisi_sas_err_handler = {
|
2018-07-18 14:14:30 +00:00
|
|
|
.reset_prepare = hisi_sas_reset_prepare_v3_hw,
|
|
|
|
.reset_done = hisi_sas_reset_done_v3_hw,
|
2017-12-08 17:16:40 +00:00
|
|
|
};
|
|
|
|
|
2020-11-02 16:47:15 +00:00
|
|
|
static UNIVERSAL_DEV_PM_OPS(hisi_sas_v3_pm_ops,
|
|
|
|
suspend_v3_hw,
|
|
|
|
resume_v3_hw,
|
|
|
|
NULL);
|
2020-10-02 14:30:33 +00:00
|
|
|
|
2017-06-14 15:33:19 +00:00
|
|
|
static struct pci_driver sas_v3_pci_driver = {
|
|
|
|
.name = DRV_NAME,
|
|
|
|
.id_table = sas_v3_pci_table,
|
|
|
|
.probe = hisi_sas_v3_probe,
|
|
|
|
.remove = hisi_sas_v3_remove,
|
2017-12-08 17:16:40 +00:00
|
|
|
.err_handler = &hisi_sas_err_handler,
|
2020-10-02 14:30:33 +00:00
|
|
|
.driver.pm = &hisi_sas_v3_pm_ops,
|
2017-06-14 15:33:19 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
module_pci_driver(sas_v3_pci_driver);
|
2018-11-09 14:06:33 +00:00
|
|
|
module_param_named(intr_conv, hisi_sas_intr_conv, bool, 0444);
|
2017-06-14 15:33:19 +00:00
|
|
|
|
|
|
|
MODULE_LICENSE("GPL");
|
|
|
|
MODULE_AUTHOR("John Garry <john.garry@huawei.com>");
|
|
|
|
MODULE_DESCRIPTION("HISILICON SAS controller v3 hw driver based on pci device");
|
2018-03-07 12:25:14 +00:00
|
|
|
MODULE_ALIAS("pci:" DRV_NAME);
|