linux/drivers/scsi/qedi/qedi_fw.c

2159 lines
65 KiB
C
Raw Normal View History

// SPDX-License-Identifier: GPL-2.0-only
/*
* QLogic iSCSI Offload Driver
* Copyright (c) 2016 Cavium Inc.
*/
#include <linux/blkdev.h>
#include <scsi/scsi_tcq.h>
#include <linux/delay.h>
#include "qedi.h"
#include "qedi_iscsi.h"
#include "qedi_gbl.h"
#include "qedi_fw_iscsi.h"
#include "qedi_fw_scsi.h"
static int send_iscsi_tmf(struct qedi_conn *qedi_conn,
struct iscsi_task *mtask, struct iscsi_task *ctask);
void qedi_iscsi_unmap_sg_list(struct qedi_cmd *cmd)
{
struct scsi_cmnd *sc = cmd->scsi_cmd;
if (cmd->io_tbl.sge_valid && sc) {
cmd->io_tbl.sge_valid = 0;
scsi_dma_unmap(sc);
}
}
static void qedi_process_logout_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_logout_rsp *resp_hdr;
struct iscsi_session *session = conn->session;
struct iscsi_logout_response_hdr *cqe_logout_response;
struct qedi_cmd *cmd;
cmd = (struct qedi_cmd *)task->dd_data;
cqe_logout_response = &cqe->cqe_common.iscsi_hdr.logout_response;
spin_lock(&session->back_lock);
resp_hdr = (struct iscsi_logout_rsp *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr, 0, sizeof(struct iscsi_hdr));
resp_hdr->opcode = cqe_logout_response->opcode;
resp_hdr->flags = cqe_logout_response->flags;
resp_hdr->hlength = 0;
resp_hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
resp_hdr->statsn = cpu_to_be32(cqe_logout_response->stat_sn);
resp_hdr->exp_cmdsn = cpu_to_be32(cqe_logout_response->exp_cmd_sn);
resp_hdr->max_cmdsn = cpu_to_be32(cqe_logout_response->max_cmd_sn);
resp_hdr->t2wait = cpu_to_be32(cqe_logout_response->time_2_wait);
resp_hdr->t2retain = cpu_to_be32(cqe_logout_response->time_2_retain);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
} else {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Active cmd list node already deleted, tid=0x%x, cid=0x%x, io_cmd_node=%p\n",
cmd->task_id, qedi_conn->iscsi_conn_id,
&cmd->io_cmd);
}
spin_unlock(&qedi_conn->list_lock);
cmd->state = RESPONSE_RECEIVED;
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr, NULL, 0);
spin_unlock(&session->back_lock);
}
static void qedi_process_text_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_task_context *task_ctx;
struct iscsi_text_rsp *resp_hdr_ptr;
struct iscsi_text_response_hdr *cqe_text_response;
struct qedi_cmd *cmd;
int pld_len;
cmd = (struct qedi_cmd *)task->dd_data;
task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
cqe_text_response = &cqe->cqe_common.iscsi_hdr.text_response;
spin_lock(&session->back_lock);
resp_hdr_ptr = (struct iscsi_text_rsp *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr_ptr, 0, sizeof(struct iscsi_hdr));
resp_hdr_ptr->opcode = cqe_text_response->opcode;
resp_hdr_ptr->flags = cqe_text_response->flags;
resp_hdr_ptr->hlength = 0;
hton24(resp_hdr_ptr->dlength,
(cqe_text_response->hdr_second_dword &
ISCSI_TEXT_RESPONSE_HDR_DATA_SEG_LEN_MASK));
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
resp_hdr_ptr->ttt = cqe_text_response->ttt;
resp_hdr_ptr->statsn = cpu_to_be32(cqe_text_response->stat_sn);
resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_text_response->exp_cmd_sn);
resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_text_response->max_cmd_sn);
pld_len = cqe_text_response->hdr_second_dword &
ISCSI_TEXT_RESPONSE_HDR_DATA_SEG_LEN_MASK;
qedi_conn->gen_pdu.resp_wr_ptr = qedi_conn->gen_pdu.resp_buf + pld_len;
memset(task_ctx, '\0', sizeof(*task_ctx));
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
} else {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Active cmd list node already deleted, tid=0x%x, cid=0x%x, io_cmd_node=%p\n",
cmd->task_id, qedi_conn->iscsi_conn_id,
&cmd->io_cmd);
}
spin_unlock(&qedi_conn->list_lock);
cmd->state = RESPONSE_RECEIVED;
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr,
qedi_conn->gen_pdu.resp_buf,
(qedi_conn->gen_pdu.resp_wr_ptr -
qedi_conn->gen_pdu.resp_buf));
spin_unlock(&session->back_lock);
}
static void qedi_tmf_resp_work(struct work_struct *work)
{
struct qedi_cmd *qedi_cmd =
container_of(work, struct qedi_cmd, tmf_work);
struct qedi_conn *qedi_conn = qedi_cmd->conn;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_tm_rsp *resp_hdr_ptr;
int rval = 0;
resp_hdr_ptr = (struct iscsi_tm_rsp *)qedi_cmd->tmf_resp_buf;
rval = qedi_cleanup_all_io(qedi, qedi_conn, qedi_cmd->task, true);
if (rval)
goto exit_tmf_resp;
spin_lock(&session->back_lock);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr, NULL, 0);
spin_unlock(&session->back_lock);
exit_tmf_resp:
kfree(resp_hdr_ptr);
scsi: qedi: Complete TMF works before disconnect We need to make sure that abort and reset completion work has completed before ep_disconnect returns. After ep_disconnect we can't manipulate cmds because libiscsi will call conn_stop and take onwership. We are trying to make sure abort work and reset completion work has completed before we do the cmd clean up in ep_disconnect. The problem is that: 1. the work function sets the QEDI_CONN_FW_CLEANUP bit, so if the work was still pending we would not see the bit set. We need to do this before the work is queued. 2. If we had multiple works queued then we could break from the loop in qedi_ep_disconnect early because when abort work 1 completes it could clear QEDI_CONN_FW_CLEANUP. qedi_ep_disconnect could then see that before work 2 has run. 3. A TMF reset completion work could run after ep_disconnect starts cleaning up cmds via qedi_clearsq. ep_disconnect's call to qedi_clearsq -> qedi_cleanup_all_io would might think it's done cleaning up cmds, but the reset completion work could still be running. We then return from ep_disconnect while still doing cleanup. This replaces the bit with a counter to track the number of queued TMF works, and adds a bool to prevent new works from starting from the completion path once a ep_disconnect starts. Link: https://lore.kernel.org/r/20210525181821.7617-28-michael.christie@oracle.com Reviewed-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-05-25 18:18:20 +00:00
spin_lock(&qedi_conn->tmf_work_lock);
qedi_conn->fw_cleanup_works--;
spin_unlock(&qedi_conn->tmf_work_lock);
}
static void qedi_process_tmf_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_tmf_response_hdr *cqe_tmp_response;
struct iscsi_tm_rsp *resp_hdr_ptr;
struct iscsi_tm *tmf_hdr;
struct qedi_cmd *qedi_cmd = NULL;
cqe_tmp_response = &cqe->cqe_common.iscsi_hdr.tmf_response;
qedi_cmd = task->dd_data;
qedi_cmd->tmf_resp_buf = kzalloc(sizeof(*resp_hdr_ptr), GFP_ATOMIC);
if (!qedi_cmd->tmf_resp_buf) {
QEDI_ERR(&qedi->dbg_ctx,
"Failed to allocate resp buf, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return;
}
spin_lock(&session->back_lock);
resp_hdr_ptr = (struct iscsi_tm_rsp *)qedi_cmd->tmf_resp_buf;
memset(resp_hdr_ptr, 0, sizeof(struct iscsi_tm_rsp));
/* Fill up the header */
resp_hdr_ptr->opcode = cqe_tmp_response->opcode;
resp_hdr_ptr->flags = cqe_tmp_response->hdr_flags;
resp_hdr_ptr->response = cqe_tmp_response->hdr_response;
resp_hdr_ptr->hlength = 0;
hton24(resp_hdr_ptr->dlength,
(cqe_tmp_response->hdr_second_dword &
ISCSI_TMF_RESPONSE_HDR_DATA_SEG_LEN_MASK));
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
resp_hdr_ptr->statsn = cpu_to_be32(cqe_tmp_response->stat_sn);
resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_tmp_response->exp_cmd_sn);
resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_tmp_response->max_cmd_sn);
tmf_hdr = (struct iscsi_tm *)qedi_cmd->task->hdr;
spin_lock(&qedi_conn->list_lock);
if (likely(qedi_cmd->io_cmd_in_list)) {
qedi_cmd->io_cmd_in_list = false;
list_del_init(&qedi_cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
scsi: qedi: Complete TMF works before disconnect We need to make sure that abort and reset completion work has completed before ep_disconnect returns. After ep_disconnect we can't manipulate cmds because libiscsi will call conn_stop and take onwership. We are trying to make sure abort work and reset completion work has completed before we do the cmd clean up in ep_disconnect. The problem is that: 1. the work function sets the QEDI_CONN_FW_CLEANUP bit, so if the work was still pending we would not see the bit set. We need to do this before the work is queued. 2. If we had multiple works queued then we could break from the loop in qedi_ep_disconnect early because when abort work 1 completes it could clear QEDI_CONN_FW_CLEANUP. qedi_ep_disconnect could then see that before work 2 has run. 3. A TMF reset completion work could run after ep_disconnect starts cleaning up cmds via qedi_clearsq. ep_disconnect's call to qedi_clearsq -> qedi_cleanup_all_io would might think it's done cleaning up cmds, but the reset completion work could still be running. We then return from ep_disconnect while still doing cleanup. This replaces the bit with a counter to track the number of queued TMF works, and adds a bool to prevent new works from starting from the completion path once a ep_disconnect starts. Link: https://lore.kernel.org/r/20210525181821.7617-28-michael.christie@oracle.com Reviewed-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-05-25 18:18:20 +00:00
spin_lock(&qedi_conn->tmf_work_lock);
switch (tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) {
case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET:
case ISCSI_TM_FUNC_TARGET_WARM_RESET:
case ISCSI_TM_FUNC_TARGET_COLD_RESET:
if (qedi_conn->ep_disconnect_starting) {
/* Session is down so ep_disconnect will clean up */
spin_unlock(&qedi_conn->tmf_work_lock);
goto unblock_sess;
}
qedi_conn->fw_cleanup_works++;
spin_unlock(&qedi_conn->tmf_work_lock);
INIT_WORK(&qedi_cmd->tmf_work, qedi_tmf_resp_work);
queue_work(qedi->tmf_thread, &qedi_cmd->tmf_work);
goto unblock_sess;
}
scsi: qedi: Complete TMF works before disconnect We need to make sure that abort and reset completion work has completed before ep_disconnect returns. After ep_disconnect we can't manipulate cmds because libiscsi will call conn_stop and take onwership. We are trying to make sure abort work and reset completion work has completed before we do the cmd clean up in ep_disconnect. The problem is that: 1. the work function sets the QEDI_CONN_FW_CLEANUP bit, so if the work was still pending we would not see the bit set. We need to do this before the work is queued. 2. If we had multiple works queued then we could break from the loop in qedi_ep_disconnect early because when abort work 1 completes it could clear QEDI_CONN_FW_CLEANUP. qedi_ep_disconnect could then see that before work 2 has run. 3. A TMF reset completion work could run after ep_disconnect starts cleaning up cmds via qedi_clearsq. ep_disconnect's call to qedi_clearsq -> qedi_cleanup_all_io would might think it's done cleaning up cmds, but the reset completion work could still be running. We then return from ep_disconnect while still doing cleanup. This replaces the bit with a counter to track the number of queued TMF works, and adds a bool to prevent new works from starting from the completion path once a ep_disconnect starts. Link: https://lore.kernel.org/r/20210525181821.7617-28-michael.christie@oracle.com Reviewed-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-05-25 18:18:20 +00:00
spin_unlock(&qedi_conn->tmf_work_lock);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr, NULL, 0);
kfree(resp_hdr_ptr);
unblock_sess:
spin_unlock(&session->back_lock);
}
static void qedi_process_login_resp(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_task_context *task_ctx;
struct iscsi_login_rsp *resp_hdr_ptr;
struct iscsi_login_response_hdr *cqe_login_response;
struct qedi_cmd *cmd;
int pld_len;
cmd = (struct qedi_cmd *)task->dd_data;
cqe_login_response = &cqe->cqe_common.iscsi_hdr.login_response;
task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
spin_lock(&session->back_lock);
resp_hdr_ptr = (struct iscsi_login_rsp *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr_ptr, 0, sizeof(struct iscsi_login_rsp));
resp_hdr_ptr->opcode = cqe_login_response->opcode;
resp_hdr_ptr->flags = cqe_login_response->flags_attr;
resp_hdr_ptr->hlength = 0;
hton24(resp_hdr_ptr->dlength,
(cqe_login_response->hdr_second_dword &
ISCSI_LOGIN_RESPONSE_HDR_DATA_SEG_LEN_MASK));
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
resp_hdr_ptr->tsih = cqe_login_response->tsih;
resp_hdr_ptr->statsn = cpu_to_be32(cqe_login_response->stat_sn);
resp_hdr_ptr->exp_cmdsn = cpu_to_be32(cqe_login_response->exp_cmd_sn);
resp_hdr_ptr->max_cmdsn = cpu_to_be32(cqe_login_response->max_cmd_sn);
resp_hdr_ptr->status_class = cqe_login_response->status_class;
resp_hdr_ptr->status_detail = cqe_login_response->status_detail;
pld_len = cqe_login_response->hdr_second_dword &
ISCSI_LOGIN_RESPONSE_HDR_DATA_SEG_LEN_MASK;
qedi_conn->gen_pdu.resp_wr_ptr = qedi_conn->gen_pdu.resp_buf + pld_len;
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
memset(task_ctx, '\0', sizeof(*task_ctx));
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr_ptr,
qedi_conn->gen_pdu.resp_buf,
(qedi_conn->gen_pdu.resp_wr_ptr -
qedi_conn->gen_pdu.resp_buf));
spin_unlock(&session->back_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
cmd->state = RESPONSE_RECEIVED;
}
static void qedi_get_rq_bdq_buf(struct qedi_ctx *qedi,
struct iscsi_cqe_unsolicited *cqe,
char *ptr, int len)
{
u16 idx = 0;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"pld_len [%d], bdq_prod_idx [%d], idx [%d]\n",
len, qedi->bdq_prod_idx,
(qedi->bdq_prod_idx % qedi->rq_num_entries));
/* Obtain buffer address from rqe_opaque */
idx = cqe->rqe_opaque;
if (idx > (QEDI_BDQ_NUM - 1)) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"wrong idx %d returned by FW, dropping the unsolicited pkt\n",
idx);
return;
}
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"rqe_opaque [0x%p], idx [%d]\n", cqe->rqe_opaque, idx);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"unsol_cqe_type = %d\n", cqe->unsol_cqe_type);
switch (cqe->unsol_cqe_type) {
case ISCSI_CQE_UNSOLICITED_SINGLE:
case ISCSI_CQE_UNSOLICITED_FIRST:
if (len)
memcpy(ptr, (void *)qedi->bdq[idx].buf_addr, len);
break;
case ISCSI_CQE_UNSOLICITED_MIDDLE:
case ISCSI_CQE_UNSOLICITED_LAST:
break;
default:
break;
}
}
static void qedi_put_rq_bdq_buf(struct qedi_ctx *qedi,
struct iscsi_cqe_unsolicited *cqe,
int count)
{
u16 idx = 0;
struct scsi_bd *pbl;
/* Obtain buffer address from rqe_opaque */
idx = cqe->rqe_opaque;
if (idx > (QEDI_BDQ_NUM - 1)) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"wrong idx %d returned by FW, dropping the unsolicited pkt\n",
idx);
return;
}
pbl = (struct scsi_bd *)qedi->bdq_pbl;
pbl += (qedi->bdq_prod_idx % qedi->rq_num_entries);
pbl->address.hi = cpu_to_le32(QEDI_U64_HI(qedi->bdq[idx].buf_dma));
pbl->address.lo = cpu_to_le32(QEDI_U64_LO(qedi->bdq[idx].buf_dma));
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"pbl [0x%p] pbl->address hi [0x%llx] lo [0x%llx] idx [%d]\n",
pbl, pbl->address.hi, pbl->address.lo, idx);
pbl->opaque.iscsi_opaque.reserved_zero[0] = 0;
pbl->opaque.iscsi_opaque.reserved_zero[1] = 0;
pbl->opaque.iscsi_opaque.reserved_zero[2] = 0;
pbl->opaque.iscsi_opaque.opaque = cpu_to_le32(idx);
/* Increment producer to let f/w know we've handled the frame */
qedi->bdq_prod_idx += count;
writew(qedi->bdq_prod_idx, qedi->bdq_primary_prod);
readw(qedi->bdq_primary_prod);
writew(qedi->bdq_prod_idx, qedi->bdq_secondary_prod);
readw(qedi->bdq_secondary_prod);
}
static void qedi_unsol_pdu_adjust_bdq(struct qedi_ctx *qedi,
struct iscsi_cqe_unsolicited *cqe,
u32 pdu_len, u32 num_bdqs,
char *bdq_data)
{
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"num_bdqs [%d]\n", num_bdqs);
qedi_get_rq_bdq_buf(qedi, cqe, bdq_data, pdu_len);
qedi_put_rq_bdq_buf(qedi, cqe, (num_bdqs + 1));
}
static int qedi_process_nopin_mesg(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn, u16 que_idx)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_nop_in_hdr *cqe_nop_in;
struct iscsi_nopin *hdr;
struct qedi_cmd *cmd;
int tgt_async_nop = 0;
u32 lun[2];
u32 pdu_len, num_bdqs;
char bdq_data[QEDI_BDQ_BUF_SIZE];
unsigned long flags;
spin_lock_bh(&session->back_lock);
cqe_nop_in = &cqe->cqe_common.iscsi_hdr.nop_in;
pdu_len = cqe_nop_in->hdr_second_dword &
ISCSI_NOP_IN_HDR_DATA_SEG_LEN_MASK;
num_bdqs = pdu_len / QEDI_BDQ_BUF_SIZE;
hdr = (struct iscsi_nopin *)&qedi_conn->gen_pdu.resp_hdr;
memset(hdr, 0, sizeof(struct iscsi_hdr));
hdr->opcode = cqe_nop_in->opcode;
hdr->max_cmdsn = cpu_to_be32(cqe_nop_in->max_cmd_sn);
hdr->exp_cmdsn = cpu_to_be32(cqe_nop_in->exp_cmd_sn);
hdr->statsn = cpu_to_be32(cqe_nop_in->stat_sn);
hdr->ttt = cpu_to_be32(cqe_nop_in->ttt);
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
spin_lock_irqsave(&qedi->hba_lock, flags);
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
pdu_len, num_bdqs, bdq_data);
hdr->itt = RESERVED_ITT;
tgt_async_nop = 1;
spin_unlock_irqrestore(&qedi->hba_lock, flags);
goto done;
}
/* Response to one of our nop-outs */
if (task) {
cmd = task->dd_data;
hdr->flags = ISCSI_FLAG_CMD_FINAL;
hdr->itt = build_itt(cqe->cqe_solicited.itid,
conn->session->age);
lun[0] = 0xffffffff;
lun[1] = 0xffffffff;
memcpy(&hdr->lun, lun, sizeof(struct scsi_lun));
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
cmd->state = RESPONSE_RECEIVED;
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
}
done:
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr, bdq_data, pdu_len);
spin_unlock_bh(&session->back_lock);
return tgt_async_nop;
}
static void qedi_process_async_mesg(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn,
u16 que_idx)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_async_msg_hdr *cqe_async_msg;
struct iscsi_async *resp_hdr;
u32 lun[2];
u32 pdu_len, num_bdqs;
char bdq_data[QEDI_BDQ_BUF_SIZE];
unsigned long flags;
spin_lock_bh(&session->back_lock);
cqe_async_msg = &cqe->cqe_common.iscsi_hdr.async_msg;
pdu_len = cqe_async_msg->hdr_second_dword &
ISCSI_ASYNC_MSG_HDR_DATA_SEG_LEN_MASK;
num_bdqs = pdu_len / QEDI_BDQ_BUF_SIZE;
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
spin_lock_irqsave(&qedi->hba_lock, flags);
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
pdu_len, num_bdqs, bdq_data);
spin_unlock_irqrestore(&qedi->hba_lock, flags);
}
resp_hdr = (struct iscsi_async *)&qedi_conn->gen_pdu.resp_hdr;
memset(resp_hdr, 0, sizeof(struct iscsi_hdr));
resp_hdr->opcode = cqe_async_msg->opcode;
resp_hdr->flags = 0x80;
lun[0] = cpu_to_be32(cqe_async_msg->lun.lo);
lun[1] = cpu_to_be32(cqe_async_msg->lun.hi);
memcpy(&resp_hdr->lun, lun, sizeof(struct scsi_lun));
resp_hdr->exp_cmdsn = cpu_to_be32(cqe_async_msg->exp_cmd_sn);
resp_hdr->max_cmdsn = cpu_to_be32(cqe_async_msg->max_cmd_sn);
resp_hdr->statsn = cpu_to_be32(cqe_async_msg->stat_sn);
resp_hdr->async_event = cqe_async_msg->async_event;
resp_hdr->async_vcode = cqe_async_msg->async_vcode;
resp_hdr->param1 = cpu_to_be16(cqe_async_msg->param1_rsrv);
resp_hdr->param2 = cpu_to_be16(cqe_async_msg->param2_rsrv);
resp_hdr->param3 = cpu_to_be16(cqe_async_msg->param3_rsrv);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)resp_hdr, bdq_data,
pdu_len);
spin_unlock_bh(&session->back_lock);
}
static void qedi_process_reject_mesg(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn,
uint16_t que_idx)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_reject_hdr *cqe_reject;
struct iscsi_reject *hdr;
u32 pld_len, num_bdqs;
unsigned long flags;
spin_lock_bh(&session->back_lock);
cqe_reject = &cqe->cqe_common.iscsi_hdr.reject;
pld_len = cqe_reject->hdr_second_dword &
ISCSI_REJECT_HDR_DATA_SEG_LEN_MASK;
num_bdqs = pld_len / QEDI_BDQ_BUF_SIZE;
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
spin_lock_irqsave(&qedi->hba_lock, flags);
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
pld_len, num_bdqs, conn->data);
spin_unlock_irqrestore(&qedi->hba_lock, flags);
}
hdr = (struct iscsi_reject *)&qedi_conn->gen_pdu.resp_hdr;
memset(hdr, 0, sizeof(struct iscsi_hdr));
hdr->opcode = cqe_reject->opcode;
hdr->reason = cqe_reject->hdr_reason;
hdr->flags = cqe_reject->hdr_flags;
hton24(hdr->dlength, (cqe_reject->hdr_second_dword &
ISCSI_REJECT_HDR_DATA_SEG_LEN_MASK));
hdr->max_cmdsn = cpu_to_be32(cqe_reject->max_cmd_sn);
hdr->exp_cmdsn = cpu_to_be32(cqe_reject->exp_cmd_sn);
hdr->statsn = cpu_to_be32(cqe_reject->stat_sn);
hdr->ffffffff = cpu_to_be32(0xffffffff);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr,
conn->data, pld_len);
spin_unlock_bh(&session->back_lock);
}
static void qedi_scsi_completion(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct iscsi_conn *conn)
{
struct scsi_cmnd *sc_cmd;
struct qedi_cmd *cmd = task->dd_data;
struct iscsi_session *session = conn->session;
struct iscsi_scsi_rsp *hdr;
struct iscsi_data_in_hdr *cqe_data_in;
int datalen = 0;
struct qedi_conn *qedi_conn;
u32 iscsi_cid;
u8 cqe_err_bits = 0;
iscsi_cid = cqe->cqe_common.conn_id;
qedi_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
cqe_data_in = &cqe->cqe_common.iscsi_hdr.data_in;
cqe_err_bits =
cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
spin_lock_bh(&session->back_lock);
/* get the scsi command */
sc_cmd = cmd->scsi_cmd;
if (!sc_cmd) {
QEDI_WARN(&qedi->dbg_ctx, "sc_cmd is NULL!\n");
goto error;
}
if (!iscsi_cmd(sc_cmd)->task) {
QEDI_WARN(&qedi->dbg_ctx,
"NULL task pointer, returned in another context.\n");
goto error;
}
if (!scsi_cmd_to_rq(sc_cmd)->q) {
QEDI_WARN(&qedi->dbg_ctx,
"request->q is NULL so request is not valid, sc_cmd=%p.\n",
sc_cmd);
goto error;
}
qedi_iscsi_unmap_sg_list(cmd);
hdr = (struct iscsi_scsi_rsp *)task->hdr;
hdr->opcode = cqe_data_in->opcode;
hdr->max_cmdsn = cpu_to_be32(cqe_data_in->max_cmd_sn);
hdr->exp_cmdsn = cpu_to_be32(cqe_data_in->exp_cmd_sn);
hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
hdr->response = cqe_data_in->reserved1;
hdr->cmd_status = cqe_data_in->status_rsvd;
hdr->flags = cqe_data_in->flags;
hdr->residual_count = cpu_to_be32(cqe_data_in->residual_count);
if (hdr->cmd_status == SAM_STAT_CHECK_CONDITION) {
datalen = cqe_data_in->reserved2 &
ISCSI_COMMON_HDR_DATA_SEG_LEN_MASK;
memcpy((char *)conn->data, (char *)cmd->sense_buffer, datalen);
}
/* If f/w reports data underrun err then set residual to IO transfer
* length, set Underrun flag and clear Overrun flag explicitly
*/
if (unlikely(cqe_err_bits &&
GET_FIELD(cqe_err_bits, CQE_ERROR_BITMAP_UNDER_RUN_ERR))) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Under flow itt=0x%x proto flags=0x%x tid=0x%x cid 0x%x fw resid 0x%x sc dlen 0x%x\n",
hdr->itt, cqe_data_in->flags, cmd->task_id,
qedi_conn->iscsi_conn_id, hdr->residual_count,
scsi_bufflen(sc_cmd));
hdr->residual_count = cpu_to_be32(scsi_bufflen(sc_cmd));
hdr->flags |= ISCSI_FLAG_CMD_UNDERFLOW;
hdr->flags &= (~ISCSI_FLAG_CMD_OVERFLOW);
}
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_TID,
"Freeing tid=0x%x for cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
cmd->state = RESPONSE_RECEIVED;
if (qedi_io_tracing)
qedi_trace_io(qedi, task, cmd->task_id, QEDI_IO_TRACE_RSP);
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr,
conn->data, datalen);
error:
spin_unlock_bh(&session->back_lock);
}
static void qedi_mtask_completion(struct qedi_ctx *qedi,
union iscsi_cqe *cqe,
struct iscsi_task *task,
struct qedi_conn *conn, uint16_t que_idx)
{
struct iscsi_conn *iscsi_conn;
u32 hdr_opcode;
hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
iscsi_conn = conn->cls_conn->dd_data;
switch (hdr_opcode) {
case ISCSI_OPCODE_SCSI_RESPONSE:
case ISCSI_OPCODE_DATA_IN:
qedi_scsi_completion(qedi, cqe, task, iscsi_conn);
break;
case ISCSI_OPCODE_LOGIN_RESPONSE:
qedi_process_login_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_TMF_RESPONSE:
qedi_process_tmf_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_TEXT_RESPONSE:
qedi_process_text_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_LOGOUT_RESPONSE:
qedi_process_logout_resp(qedi, cqe, task, conn);
break;
case ISCSI_OPCODE_NOP_IN:
qedi_process_nopin_mesg(qedi, cqe, task, conn, que_idx);
break;
default:
QEDI_ERR(&qedi->dbg_ctx, "unknown opcode\n");
}
}
static void qedi_process_nopin_local_cmpl(struct qedi_ctx *qedi,
struct iscsi_cqe_solicited *cqe,
struct iscsi_task *task,
struct qedi_conn *qedi_conn)
{
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
struct qedi_cmd *cmd = task->dd_data;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_UNSOL,
"itid=0x%x, cmd task id=0x%x\n",
cqe->itid, cmd->task_id);
cmd->state = RESPONSE_RECEIVED;
spin_lock_bh(&session->back_lock);
__iscsi_put_task(task);
spin_unlock_bh(&session->back_lock);
}
static void qedi_process_cmd_cleanup_resp(struct qedi_ctx *qedi,
struct iscsi_cqe_solicited *cqe,
struct iscsi_conn *conn)
{
struct qedi_work_map *work, *work_tmp;
u32 proto_itt = cqe->itid;
int found = 0;
struct qedi_cmd *qedi_cmd = NULL;
u32 iscsi_cid;
struct qedi_conn *qedi_conn;
struct qedi_cmd *dbg_cmd;
struct iscsi_task *mtask, *task;
struct iscsi_tm *tmf_hdr = NULL;
iscsi_cid = cqe->conn_id;
qedi_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
scsi: qedi: Fix kernel crash during port toggle BUG: unable to handle kernel NULL pointer dereference at 0000000000000100 [ 985.596918] IP: _raw_spin_lock_bh+0x17/0x30 [ 985.601581] PGD 0 P4D 0 [ 985.604405] Oops: 0002 [#1] SMP : [ 985.704533] CPU: 16 PID: 1156 Comm: qedi_thread/16 Not tainted 4.16.0-rc2 #1 [ 985.712397] Hardware name: Dell Inc. PowerEdge R730/0599V5, BIOS 2.4.3 01/17/2017 [ 985.720747] RIP: 0010:_raw_spin_lock_bh+0x17/0x30 [ 985.725996] RSP: 0018:ffffa4b1c43d3e10 EFLAGS: 00010246 [ 985.731823] RAX: 0000000000000000 RBX: ffff94a31bd03000 RCX: 0000000000000000 [ 985.739783] RDX: 0000000000000001 RSI: ffff94a32fa16938 RDI: 0000000000000100 [ 985.747744] RBP: 0000000000000004 R08: 0000000000000000 R09: 0000000000000a33 [ 985.755703] R10: 0000000000000000 R11: ffffa4b1c43d3af0 R12: 0000000000000000 [ 985.763662] R13: ffff94a301f40818 R14: 0000000000000000 R15: 000000000000000c [ 985.771622] FS: 0000000000000000(0000) GS:ffff94a32fa00000(0000) knlGS:0000000000000000 [ 985.780649] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 985.787057] CR2: 0000000000000100 CR3: 000000067a009006 CR4: 00000000001606e0 [ 985.795017] Call Trace: [ 985.797747] qedi_fp_process_cqes+0x258/0x980 [qedi] [ 985.803294] qedi_percpu_io_thread+0x10f/0x1b0 [qedi] [ 985.808931] kthread+0xf5/0x130 [ 985.812434] ? qedi_free_uio+0xd0/0xd0 [qedi] [ 985.817298] ? kthread_bind+0x10/0x10 [ 985.821372] ? do_syscall_64+0x6e/0x1a0 Signed-off-by: Manish Rangankar <manish.rangankar@cavium.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2018-02-26 09:01:17 +00:00
if (!qedi_conn) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"icid not found 0x%x\n", cqe->conn_id);
return;
}
/* Based on this itt get the corresponding qedi_cmd */
spin_lock_bh(&qedi_conn->tmf_work_lock);
list_for_each_entry_safe(work, work_tmp, &qedi_conn->tmf_work_list,
list) {
if (work->rtid == proto_itt) {
/* We found the command */
qedi_cmd = work->qedi_cmd;
if (!qedi_cmd->list_tmf_work) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"TMF work not found, cqe->tid=0x%x, cid=0x%x\n",
proto_itt, qedi_conn->iscsi_conn_id);
WARN_ON(1);
}
found = 1;
mtask = qedi_cmd->task;
task = work->ctask;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
list_del_init(&work->list);
kfree(work);
qedi_cmd->list_tmf_work = NULL;
}
}
spin_unlock_bh(&qedi_conn->tmf_work_lock);
if (!found)
goto check_cleanup_reqs;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"TMF work, cqe->tid=0x%x, tmf flags=0x%x, cid=0x%x\n",
proto_itt, tmf_hdr->flags, qedi_conn->iscsi_conn_id);
spin_lock_bh(&conn->session->back_lock);
if (iscsi_task_is_completed(task)) {
QEDI_NOTICE(&qedi->dbg_ctx,
"IO task completed, tmf rtt=0x%x, cid=0x%x\n",
get_itt(tmf_hdr->rtt), qedi_conn->iscsi_conn_id);
goto unlock;
}
dbg_cmd = task->dd_data;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Abort tmf rtt=0x%x, i/o itt=0x%x, i/o tid=0x%x, cid=0x%x\n",
get_itt(tmf_hdr->rtt), get_itt(task->itt), dbg_cmd->task_id,
qedi_conn->iscsi_conn_id);
spin_lock(&qedi_conn->list_lock);
if (likely(dbg_cmd->io_cmd_in_list)) {
dbg_cmd->io_cmd_in_list = false;
list_del_init(&dbg_cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
qedi_cmd->state = CLEANUP_RECV;
unlock:
spin_unlock_bh(&conn->session->back_lock);
wake_up_interruptible(&qedi_conn->wait_queue);
return;
check_cleanup_reqs:
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
if (atomic_inc_return(&qedi_conn->cmd_cleanup_cmpl) ==
qedi_conn->cmd_cleanup_req) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Freeing tid=0x%x for cid=0x%x\n",
cqe->itid, qedi_conn->iscsi_conn_id);
wake_up(&qedi_conn->wait_queue);
}
}
void qedi_fp_process_cqes(struct qedi_work *work)
{
struct qedi_ctx *qedi = work->qedi;
union iscsi_cqe *cqe = &work->cqe;
struct iscsi_task *task = NULL;
struct iscsi_nopout *nopout_hdr;
struct qedi_conn *q_conn;
struct iscsi_conn *conn;
struct qedi_cmd *qedi_cmd;
u32 comp_type;
u32 iscsi_cid;
u32 hdr_opcode;
u16 que_idx = work->que_idx;
u8 cqe_err_bits = 0;
comp_type = cqe->cqe_common.cqe_type;
hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
cqe_err_bits =
cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_CONN,
"fw_cid=0x%x, cqe type=0x%x, opcode=0x%x\n",
cqe->cqe_common.conn_id, comp_type, hdr_opcode);
if (comp_type >= MAX_ISCSI_CQES_TYPE) {
QEDI_WARN(&qedi->dbg_ctx, "Invalid CqE type\n");
return;
}
iscsi_cid = cqe->cqe_common.conn_id;
q_conn = qedi->cid_que.conn_cid_tbl[iscsi_cid];
if (!q_conn) {
QEDI_WARN(&qedi->dbg_ctx,
"Session no longer exists for cid=0x%x!!\n",
iscsi_cid);
return;
}
conn = q_conn->cls_conn->dd_data;
if (unlikely(cqe_err_bits &&
GET_FIELD(cqe_err_bits,
CQE_ERROR_BITMAP_DATA_DIGEST_ERR))) {
iscsi_conn_failure(conn, ISCSI_ERR_DATA_DGST);
return;
}
switch (comp_type) {
case ISCSI_CQE_TYPE_SOLICITED:
case ISCSI_CQE_TYPE_SOLICITED_WITH_SENSE:
qedi_cmd = container_of(work, struct qedi_cmd, cqe_work);
task = qedi_cmd->task;
if (!task) {
QEDI_WARN(&qedi->dbg_ctx, "task is NULL\n");
return;
}
/* Process NOPIN local completion */
nopout_hdr = (struct iscsi_nopout *)task->hdr;
if ((nopout_hdr->itt == RESERVED_ITT) &&
(cqe->cqe_solicited.itid != (u16)RESERVED_ITT)) {
qedi_process_nopin_local_cmpl(qedi, &cqe->cqe_solicited,
task, q_conn);
} else {
cqe->cqe_solicited.itid =
qedi_get_itt(cqe->cqe_solicited);
/* Process other solicited responses */
qedi_mtask_completion(qedi, cqe, task, q_conn, que_idx);
}
break;
case ISCSI_CQE_TYPE_UNSOLICITED:
switch (hdr_opcode) {
case ISCSI_OPCODE_NOP_IN:
qedi_process_nopin_mesg(qedi, cqe, task, q_conn,
que_idx);
break;
case ISCSI_OPCODE_ASYNC_MSG:
qedi_process_async_mesg(qedi, cqe, task, q_conn,
que_idx);
break;
case ISCSI_OPCODE_REJECT:
qedi_process_reject_mesg(qedi, cqe, task, q_conn,
que_idx);
break;
}
goto exit_fp_process;
case ISCSI_CQE_TYPE_DUMMY:
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM, "Dummy CqE\n");
goto exit_fp_process;
case ISCSI_CQE_TYPE_TASK_CLEANUP:
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM, "CleanUp CqE\n");
qedi_process_cmd_cleanup_resp(qedi, &cqe->cqe_solicited, conn);
goto exit_fp_process;
default:
QEDI_ERR(&qedi->dbg_ctx, "Error cqe.\n");
break;
}
exit_fp_process:
return;
}
static void qedi_ring_doorbell(struct qedi_conn *qedi_conn)
{
qedi_conn->ep->db_data.sq_prod = qedi_conn->ep->fw_sq_prod_idx;
/* wmb - Make sure fw idx is coherent */
wmb();
writel(*(u32 *)&qedi_conn->ep->db_data, qedi_conn->ep->p_doorbell);
/* Make sure fw write idx is coherent, and include both memory barriers
* as a failsafe as for some architectures the call is the same but on
* others they are two different assembly operations.
*/
wmb();
QEDI_INFO(&qedi_conn->qedi->dbg_ctx, QEDI_LOG_MP_REQ,
"prod_idx=0x%x, fw_prod_idx=0x%x, cid=0x%x\n",
qedi_conn->ep->sq_prod_idx, qedi_conn->ep->fw_sq_prod_idx,
qedi_conn->iscsi_conn_id);
}
static u16 qedi_get_wqe_idx(struct qedi_conn *qedi_conn)
{
struct qedi_endpoint *ep;
u16 rval;
ep = qedi_conn->ep;
rval = ep->sq_prod_idx;
/* Increament SQ index */
ep->sq_prod_idx++;
ep->fw_sq_prod_idx++;
if (ep->sq_prod_idx == QEDI_SQ_SIZE)
ep->sq_prod_idx = 0;
return rval;
}
int qedi_send_iscsi_login(struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct iscsi_login_req_hdr login_req_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct iscsi_task_context *fw_task_ctx;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_login_req *login_hdr;
struct scsi_sge *resp_sge = NULL;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
ep = qedi_conn->ep;
login_hdr = (struct iscsi_login_req *)task->hdr;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&login_req_pdu_header, 0, sizeof(login_req_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
login_req_pdu_header.opcode = login_hdr->opcode;
login_req_pdu_header.version_min = login_hdr->min_version;
login_req_pdu_header.version_max = login_hdr->max_version;
login_req_pdu_header.flags_attr = login_hdr->flags;
login_req_pdu_header.isid_tabc = swab32p((u32 *)login_hdr->isid);
login_req_pdu_header.isid_d = swab16p((u16 *)&login_hdr->isid[4]);
login_req_pdu_header.tsih = login_hdr->tsih;
login_req_pdu_header.hdr_second_dword = ntoh24(login_hdr->dlength);
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
login_req_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
login_req_pdu_header.cid = qedi_conn->iscsi_conn_id;
login_req_pdu_header.cmd_sn = be32_to_cpu(login_hdr->cmdsn);
login_req_pdu_header.exp_stat_sn = be32_to_cpu(login_hdr->exp_statsn);
login_req_pdu_header.exp_stat_sn = 0;
/* Fill tx AHS and rx buffer */
tx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.req_dma_addr);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
tx_sgl_task_params.total_buffer_size = ntoh24(login_hdr->dlength);
tx_sgl_task_params.num_sges = 1;
rx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.resp_dma_addr);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
rx_sgl_task_params.num_sges = 1;
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = ntoh24(login_hdr->dlength);
task_params.rx_io_size = resp_sge->sge_len;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_login_request_task(&task_params,
&login_req_pdu_header,
&tx_sgl_task_params,
&rx_sgl_task_params);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_send_iscsi_logout(struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct iscsi_logout_req_hdr logout_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct iscsi_task_context *fw_task_ctx;
struct iscsi_logout *logout_hdr = NULL;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
logout_hdr = (struct iscsi_logout *)task->hdr;
ep = qedi_conn->ep;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&logout_pdu_header, 0, sizeof(logout_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
logout_pdu_header.opcode = logout_hdr->opcode;
logout_pdu_header.reason_code = 0x80 | logout_hdr->flags;
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
logout_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
logout_pdu_header.exp_stat_sn = be32_to_cpu(logout_hdr->exp_statsn);
logout_pdu_header.cmd_sn = be32_to_cpu(logout_hdr->cmdsn);
logout_pdu_header.cid = qedi_conn->iscsi_conn_id;
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = 0;
task_params.rx_io_size = 0;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_logout_request_task(&task_params,
&logout_pdu_header,
NULL, NULL);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_cleanup_all_io(struct qedi_ctx *qedi, struct qedi_conn *qedi_conn,
struct iscsi_task *task, bool in_recovery)
{
int rval;
struct iscsi_task *ctask;
struct qedi_cmd *cmd, *cmd_tmp;
struct iscsi_tm *tmf_hdr;
unsigned int lun = 0;
bool lun_reset = false;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct iscsi_session *session = conn->session;
/* From recovery, task is NULL or from tmf resp valid task */
if (task) {
tmf_hdr = (struct iscsi_tm *)task->hdr;
if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_LOGICAL_UNIT_RESET) {
lun_reset = true;
lun = scsilun_to_int(&tmf_hdr->lun);
}
}
qedi_conn->cmd_cleanup_req = 0;
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
atomic_set(&qedi_conn->cmd_cleanup_cmpl, 0);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"active_cmd_count=%d, cid=0x%x, in_recovery=%d, lun_reset=%d\n",
qedi_conn->active_cmd_count, qedi_conn->iscsi_conn_id,
in_recovery, lun_reset);
if (lun_reset)
spin_lock_bh(&session->back_lock);
spin_lock(&qedi_conn->list_lock);
list_for_each_entry_safe(cmd, cmd_tmp, &qedi_conn->active_cmd_list,
io_cmd) {
ctask = cmd->task;
if (ctask == task)
continue;
if (lun_reset) {
if (cmd->scsi_cmd && cmd->scsi_cmd->device) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"tid=0x%x itt=0x%x scsi_cmd_ptr=%p device=%p task_state=%d cmd_state=0%x cid=0x%x\n",
cmd->task_id, get_itt(ctask->itt),
cmd->scsi_cmd, cmd->scsi_cmd->device,
ctask->state, cmd->state,
qedi_conn->iscsi_conn_id);
if (cmd->scsi_cmd->device->lun != lun)
continue;
}
}
qedi_conn->cmd_cleanup_req++;
qedi_iscsi_cleanup_task(ctask, true);
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
QEDI_WARN(&qedi->dbg_ctx,
"Deleted active cmd list node io_cmd=%p, cid=0x%x\n",
&cmd->io_cmd, qedi_conn->iscsi_conn_id);
}
spin_unlock(&qedi_conn->list_lock);
if (lun_reset)
spin_unlock_bh(&session->back_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"cmd_cleanup_req=%d, cid=0x%x\n",
qedi_conn->cmd_cleanup_req,
qedi_conn->iscsi_conn_id);
rval = wait_event_interruptible_timeout(qedi_conn->wait_queue,
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
(qedi_conn->cmd_cleanup_req ==
atomic_read(&qedi_conn->cmd_cleanup_cmpl)) ||
test_bit(QEDI_IN_RECOVERY, &qedi->flags),
5 * HZ);
if (rval) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"i/o cmd_cleanup_req=%d, equal to cmd_cleanup_cmpl=%d, cid=0x%x\n",
qedi_conn->cmd_cleanup_req,
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
atomic_read(&qedi_conn->cmd_cleanup_cmpl),
qedi_conn->iscsi_conn_id);
return 0;
}
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"i/o cmd_cleanup_req=%d, not equal to cmd_cleanup_cmpl=%d, cid=0x%x\n",
qedi_conn->cmd_cleanup_req,
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
atomic_read(&qedi_conn->cmd_cleanup_cmpl),
qedi_conn->iscsi_conn_id);
iscsi_host_for_each_session(qedi->shost,
qedi_mark_device_missing);
qedi_ops->common->drain(qedi->cdev);
/* Enable IOs for all other sessions except current.*/
if (!wait_event_interruptible_timeout(qedi_conn->wait_queue,
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
(qedi_conn->cmd_cleanup_req ==
atomic_read(&qedi_conn->cmd_cleanup_cmpl)) ||
test_bit(QEDI_IN_RECOVERY, &qedi->flags),
5 * HZ)) {
iscsi_host_for_each_session(qedi->shost,
qedi_mark_device_available);
return -1;
}
iscsi_host_for_each_session(qedi->shost,
qedi_mark_device_available);
return 0;
}
void qedi_clearsq(struct qedi_ctx *qedi, struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct qedi_endpoint *qedi_ep;
int rval;
qedi_ep = qedi_conn->ep;
qedi_conn->cmd_cleanup_req = 0;
scsi: qedi: Fix cmd_cleanup_cmpl counter mismatch issue When issued LUN reset under heavy I/O we hit the qedi WARN_ON because of a mismatch in firmware I/O cmd cleanup request count and I/O cmd cleanup response count received. The mismatch is because of a race caused by the postfix increment of cmd_cleanup_cmpl. [qedi_clearsq:1295]:18: fatal error, need hard reset, cid=0x0 WARNING: CPU: 48 PID: 110963 at drivers/scsi/qedi/qedi_fw.c:1296 qedi_clearsq+0xa5/0xd0 [qedi] CPU: 48 PID: 110963 Comm: kworker/u130:0 Kdump: loaded Tainted: G W Hardware name: HPE ProLiant DL385 Gen10/ProLiant DL385 Gen10, BIOS A40 04/15/2020 Workqueue: iscsi_conn_cleanup iscsi_cleanup_conn_work_fn [scsi_transport_iscsi] RIP: 0010:qedi_clearsq+0xa5/0xd0 [qedi] RSP: 0018:ffffac2162c7fd98 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff975213c40ab8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9761bf816858 RDI: ffff9761bf816858 RBP: ffff975247018628 R08: 000000000000522c R09: 000000000000005b R10: 0000000000000000 R11: ffffac2162c7fbd8 R12: ffff97522e1b2be8 R13: 0000000000000000 R14: ffff97522e1b2800 R15: 0000000000000001 FS: 0000000000000000(0000) GS:ffff9761bf800000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f1a34e3e1a0 CR3: 0000000108bb2000 CR4: 0000000000350ee0 Call Trace: qedi_ep_disconnect+0x533/0x550 [qedi] ? iscsi_dbg_trace+0x63/0x80 [scsi_transport_iscsi] ? _cond_resched+0x15/0x30 ? iscsi_suspend_queue+0x19/0x40 [libiscsi] iscsi_ep_disconnect+0xb0/0x130 [scsi_transport_iscsi] iscsi_cleanup_conn_work_fn+0x82/0x130 [scsi_transport_iscsi] process_one_work+0x1a7/0x360 ? create_worker+0x1a0/0x1a0 worker_thread+0x30/0x390 ? create_worker+0x1a0/0x1a0 kthread+0x116/0x130 ? kthread_flush_work_fn+0x10/0x10 ret_from_fork+0x22/0x40 ---[ end trace 5f1441f59082235c ]--- Link: https://lore.kernel.org/r/20211203095218.5477-1-mrangankar@marvell.com Reviewed-by: Lee Duncan <lduncan@suse.com> Reviewed-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-12-03 09:52:18 +00:00
atomic_set(&qedi_conn->cmd_cleanup_cmpl, 0);
if (!qedi_ep) {
QEDI_WARN(&qedi->dbg_ctx,
"Cannot proceed, ep already disconnected, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return;
}
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Clearing SQ for cid=0x%x, conn=%p, ep=%p\n",
qedi_conn->iscsi_conn_id, qedi_conn, qedi_ep);
qedi_ops->clear_sq(qedi->cdev, qedi_ep->handle);
rval = qedi_cleanup_all_io(qedi, qedi_conn, task, true);
if (rval) {
QEDI_ERR(&qedi->dbg_ctx,
"fatal error, need hard reset, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
WARN_ON(1);
}
}
static int qedi_wait_for_cleanup_request(struct qedi_ctx *qedi,
struct qedi_conn *qedi_conn,
struct iscsi_task *task,
struct qedi_cmd *qedi_cmd,
struct qedi_work_map *list_work)
{
struct qedi_cmd *cmd = (struct qedi_cmd *)task->dd_data;
int wait;
wait = wait_event_interruptible_timeout(qedi_conn->wait_queue,
((qedi_cmd->state ==
CLEANUP_RECV) ||
((qedi_cmd->type == TYPEIO) &&
(cmd->state ==
RESPONSE_RECEIVED))),
5 * HZ);
if (!wait) {
qedi_cmd->state = CLEANUP_WAIT_FAILED;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Cleanup timedout tid=0x%x, issue connection recovery, cid=0x%x\n",
cmd->task_id, qedi_conn->iscsi_conn_id);
return -1;
}
return 0;
}
static void qedi_abort_work(struct work_struct *work)
{
struct qedi_cmd *qedi_cmd =
container_of(work, struct qedi_cmd, tmf_work);
struct qedi_conn *qedi_conn = qedi_cmd->conn;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_conn *conn = qedi_conn->cls_conn->dd_data;
struct qedi_work_map *list_work = NULL;
struct iscsi_task *mtask;
struct qedi_cmd *cmd;
struct iscsi_task *ctask;
struct iscsi_tm *tmf_hdr;
s16 rval = 0;
mtask = qedi_cmd->task;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
spin_lock_bh(&conn->session->back_lock);
ctask = iscsi_itt_to_ctask(conn, tmf_hdr->rtt);
if (!ctask) {
spin_unlock_bh(&conn->session->back_lock);
QEDI_ERR(&qedi->dbg_ctx, "Invalid RTT. Letting abort timeout.\n");
goto clear_cleanup;
}
if (iscsi_task_is_completed(ctask)) {
spin_unlock_bh(&conn->session->back_lock);
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Task already completed\n");
/*
* We have to still send the TMF because libiscsi needs the
* response to avoid a timeout.
*/
goto send_tmf;
}
spin_unlock_bh(&conn->session->back_lock);
cmd = (struct qedi_cmd *)ctask->dd_data;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"Abort tmf rtt=0x%x, cmd itt=0x%x, cmd tid=0x%x, cid=0x%x\n",
get_itt(tmf_hdr->rtt), get_itt(ctask->itt), cmd->task_id,
qedi_conn->iscsi_conn_id);
if (qedi_do_not_recover) {
QEDI_ERR(&qedi->dbg_ctx, "DONT SEND CLEANUP/ABORT %d\n",
qedi_do_not_recover);
goto clear_cleanup;
}
list_work = kzalloc(sizeof(*list_work), GFP_NOIO);
if (!list_work) {
QEDI_ERR(&qedi->dbg_ctx, "Memory allocation failed\n");
goto clear_cleanup;
}
qedi_cmd->type = TYPEIO;
qedi_cmd->state = CLEANUP_WAIT;
list_work->qedi_cmd = qedi_cmd;
list_work->rtid = cmd->task_id;
list_work->state = QEDI_WORK_SCHEDULED;
list_work->ctask = ctask;
qedi_cmd->list_tmf_work = list_work;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"Queue tmf work=%p, list node=%p, cid=0x%x, tmf flags=0x%x\n",
list_work->ptr_tmf_work, list_work, qedi_conn->iscsi_conn_id,
tmf_hdr->flags);
spin_lock_bh(&qedi_conn->tmf_work_lock);
list_add_tail(&list_work->list, &qedi_conn->tmf_work_list);
spin_unlock_bh(&qedi_conn->tmf_work_lock);
qedi_iscsi_cleanup_task(ctask, false);
rval = qedi_wait_for_cleanup_request(qedi, qedi_conn, ctask, qedi_cmd,
list_work);
if (rval == -1) {
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_INFO,
"FW cleanup got escalated, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
goto ldel_exit;
}
send_tmf:
send_iscsi_tmf(qedi_conn, qedi_cmd->task, ctask);
scsi: qedi: Complete TMF works before disconnect We need to make sure that abort and reset completion work has completed before ep_disconnect returns. After ep_disconnect we can't manipulate cmds because libiscsi will call conn_stop and take onwership. We are trying to make sure abort work and reset completion work has completed before we do the cmd clean up in ep_disconnect. The problem is that: 1. the work function sets the QEDI_CONN_FW_CLEANUP bit, so if the work was still pending we would not see the bit set. We need to do this before the work is queued. 2. If we had multiple works queued then we could break from the loop in qedi_ep_disconnect early because when abort work 1 completes it could clear QEDI_CONN_FW_CLEANUP. qedi_ep_disconnect could then see that before work 2 has run. 3. A TMF reset completion work could run after ep_disconnect starts cleaning up cmds via qedi_clearsq. ep_disconnect's call to qedi_clearsq -> qedi_cleanup_all_io would might think it's done cleaning up cmds, but the reset completion work could still be running. We then return from ep_disconnect while still doing cleanup. This replaces the bit with a counter to track the number of queued TMF works, and adds a bool to prevent new works from starting from the completion path once a ep_disconnect starts. Link: https://lore.kernel.org/r/20210525181821.7617-28-michael.christie@oracle.com Reviewed-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-05-25 18:18:20 +00:00
goto clear_cleanup;
ldel_exit:
spin_lock_bh(&qedi_conn->tmf_work_lock);
if (qedi_cmd->list_tmf_work) {
list_del_init(&list_work->list);
qedi_cmd->list_tmf_work = NULL;
kfree(list_work);
}
spin_unlock_bh(&qedi_conn->tmf_work_lock);
spin_lock(&qedi_conn->list_lock);
if (likely(cmd->io_cmd_in_list)) {
cmd->io_cmd_in_list = false;
list_del_init(&cmd->io_cmd);
qedi_conn->active_cmd_count--;
}
spin_unlock(&qedi_conn->list_lock);
scsi: qedi: Complete TMF works before disconnect We need to make sure that abort and reset completion work has completed before ep_disconnect returns. After ep_disconnect we can't manipulate cmds because libiscsi will call conn_stop and take onwership. We are trying to make sure abort work and reset completion work has completed before we do the cmd clean up in ep_disconnect. The problem is that: 1. the work function sets the QEDI_CONN_FW_CLEANUP bit, so if the work was still pending we would not see the bit set. We need to do this before the work is queued. 2. If we had multiple works queued then we could break from the loop in qedi_ep_disconnect early because when abort work 1 completes it could clear QEDI_CONN_FW_CLEANUP. qedi_ep_disconnect could then see that before work 2 has run. 3. A TMF reset completion work could run after ep_disconnect starts cleaning up cmds via qedi_clearsq. ep_disconnect's call to qedi_clearsq -> qedi_cleanup_all_io would might think it's done cleaning up cmds, but the reset completion work could still be running. We then return from ep_disconnect while still doing cleanup. This replaces the bit with a counter to track the number of queued TMF works, and adds a bool to prevent new works from starting from the completion path once a ep_disconnect starts. Link: https://lore.kernel.org/r/20210525181821.7617-28-michael.christie@oracle.com Reviewed-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-05-25 18:18:20 +00:00
clear_cleanup:
spin_lock(&qedi_conn->tmf_work_lock);
qedi_conn->fw_cleanup_works--;
spin_unlock(&qedi_conn->tmf_work_lock);
}
static int send_iscsi_tmf(struct qedi_conn *qedi_conn, struct iscsi_task *mtask,
struct iscsi_task *ctask)
{
struct iscsi_tmf_request_hdr tmf_pdu_header;
struct iscsi_task_params task_params;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_task_context *fw_task_ctx;
struct iscsi_tm *tmf_hdr;
struct qedi_cmd *qedi_cmd;
struct qedi_cmd *cmd;
struct qedi_endpoint *ep;
u32 scsi_lun[2];
s16 tid = 0;
u16 sq_idx = 0;
tmf_hdr = (struct iscsi_tm *)mtask->hdr;
qedi_cmd = (struct qedi_cmd *)mtask->dd_data;
ep = qedi_conn->ep;
if (!ep)
return -ENODEV;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&tmf_pdu_header, 0, sizeof(tmf_pdu_header));
/* Update header info */
qedi_update_itt_map(qedi, tid, mtask->itt, qedi_cmd);
tmf_pdu_header.itt = qedi_set_itt(tid, get_itt(mtask->itt));
tmf_pdu_header.cmd_sn = be32_to_cpu(tmf_hdr->cmdsn);
memcpy(scsi_lun, &tmf_hdr->lun, sizeof(struct scsi_lun));
tmf_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
tmf_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
if ((tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) ==
ISCSI_TM_FUNC_ABORT_TASK) {
cmd = (struct qedi_cmd *)ctask->dd_data;
tmf_pdu_header.rtt =
qedi_set_itt(cmd->task_id,
get_itt(tmf_hdr->rtt));
} else {
tmf_pdu_header.rtt = ISCSI_RESERVED_TAG;
}
tmf_pdu_header.opcode = tmf_hdr->opcode;
tmf_pdu_header.function = tmf_hdr->flags;
tmf_pdu_header.hdr_second_dword = ntoh24(tmf_hdr->dlength);
tmf_pdu_header.ref_cmd_sn = be32_to_cpu(tmf_hdr->refcmdsn);
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = 0;
task_params.rx_io_size = 0;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
init_initiator_tmf_request_task(&task_params, &tmf_pdu_header);
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_send_iscsi_tmf(struct qedi_conn *qedi_conn, struct iscsi_task *mtask)
{
struct iscsi_tm *tmf_hdr = (struct iscsi_tm *)mtask->hdr;
struct qedi_cmd *qedi_cmd = mtask->dd_data;
struct qedi_ctx *qedi = qedi_conn->qedi;
int rc = 0;
switch (tmf_hdr->flags & ISCSI_FLAG_TM_FUNC_MASK) {
case ISCSI_TM_FUNC_ABORT_TASK:
scsi: qedi: Complete TMF works before disconnect We need to make sure that abort and reset completion work has completed before ep_disconnect returns. After ep_disconnect we can't manipulate cmds because libiscsi will call conn_stop and take onwership. We are trying to make sure abort work and reset completion work has completed before we do the cmd clean up in ep_disconnect. The problem is that: 1. the work function sets the QEDI_CONN_FW_CLEANUP bit, so if the work was still pending we would not see the bit set. We need to do this before the work is queued. 2. If we had multiple works queued then we could break from the loop in qedi_ep_disconnect early because when abort work 1 completes it could clear QEDI_CONN_FW_CLEANUP. qedi_ep_disconnect could then see that before work 2 has run. 3. A TMF reset completion work could run after ep_disconnect starts cleaning up cmds via qedi_clearsq. ep_disconnect's call to qedi_clearsq -> qedi_cleanup_all_io would might think it's done cleaning up cmds, but the reset completion work could still be running. We then return from ep_disconnect while still doing cleanup. This replaces the bit with a counter to track the number of queued TMF works, and adds a bool to prevent new works from starting from the completion path once a ep_disconnect starts. Link: https://lore.kernel.org/r/20210525181821.7617-28-michael.christie@oracle.com Reviewed-by: Manish Rangankar <mrangankar@marvell.com> Signed-off-by: Mike Christie <michael.christie@oracle.com> Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2021-05-25 18:18:20 +00:00
spin_lock(&qedi_conn->tmf_work_lock);
qedi_conn->fw_cleanup_works++;
spin_unlock(&qedi_conn->tmf_work_lock);
INIT_WORK(&qedi_cmd->tmf_work, qedi_abort_work);
queue_work(qedi->tmf_thread, &qedi_cmd->tmf_work);
break;
case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET:
case ISCSI_TM_FUNC_TARGET_WARM_RESET:
case ISCSI_TM_FUNC_TARGET_COLD_RESET:
rc = send_iscsi_tmf(qedi_conn, mtask, NULL);
break;
default:
QEDI_ERR(&qedi->dbg_ctx, "Invalid tmf, cid=0x%x\n",
qedi_conn->iscsi_conn_id);
return -EINVAL;
}
return rc;
}
int qedi_send_iscsi_text(struct qedi_conn *qedi_conn,
struct iscsi_task *task)
{
struct iscsi_text_request_hdr text_request_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct iscsi_task_context *fw_task_ctx;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_text *text_hdr;
struct scsi_sge *req_sge = NULL;
struct scsi_sge *resp_sge = NULL;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
req_sge = (struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
text_hdr = (struct iscsi_text *)task->hdr;
ep = qedi_conn->ep;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&text_request_pdu_header, 0, sizeof(text_request_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
text_request_pdu_header.opcode = text_hdr->opcode;
text_request_pdu_header.flags_attr = text_hdr->flags;
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
text_request_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
text_request_pdu_header.ttt = text_hdr->ttt;
text_request_pdu_header.cmd_sn = be32_to_cpu(text_hdr->cmdsn);
text_request_pdu_header.exp_stat_sn = be32_to_cpu(text_hdr->exp_statsn);
text_request_pdu_header.hdr_second_dword = ntoh24(text_hdr->dlength);
/* Fill tx AHS and rx buffer */
tx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.req_dma_addr);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
tx_sgl_task_params.total_buffer_size = req_sge->sge_len;
tx_sgl_task_params.num_sges = 1;
rx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.resp_dma_addr);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
rx_sgl_task_params.num_sges = 1;
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = ntoh24(text_hdr->dlength);
task_params.rx_io_size = resp_sge->sge_len;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_text_request_task(&task_params,
&text_request_pdu_header,
&tx_sgl_task_params,
&rx_sgl_task_params);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_send_iscsi_nopout(struct qedi_conn *qedi_conn,
struct iscsi_task *task,
char *datap, int data_len, int unsol)
{
struct iscsi_nop_out_hdr nop_out_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct iscsi_task_params task_params;
struct qedi_ctx *qedi = qedi_conn->qedi;
struct iscsi_task_context *fw_task_ctx;
struct iscsi_nopout *nopout_hdr;
struct scsi_sge *resp_sge = NULL;
struct qedi_cmd *qedi_cmd;
struct qedi_endpoint *ep;
u32 scsi_lun[2];
s16 tid = 0;
u16 sq_idx = 0;
int rval = 0;
resp_sge = (struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
qedi_cmd = (struct qedi_cmd *)task->dd_data;
nopout_hdr = (struct iscsi_nopout *)task->hdr;
ep = qedi_conn->ep;
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
qedi_cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&nop_out_pdu_header, 0, sizeof(nop_out_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
/* Update header info */
nop_out_pdu_header.opcode = nopout_hdr->opcode;
SET_FIELD(nop_out_pdu_header.flags_attr, ISCSI_NOP_OUT_HDR_CONST1, 1);
SET_FIELD(nop_out_pdu_header.flags_attr, ISCSI_NOP_OUT_HDR_RSRV, 0);
memcpy(scsi_lun, &nopout_hdr->lun, sizeof(struct scsi_lun));
nop_out_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
nop_out_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
nop_out_pdu_header.cmd_sn = be32_to_cpu(nopout_hdr->cmdsn);
nop_out_pdu_header.exp_stat_sn = be32_to_cpu(nopout_hdr->exp_statsn);
qedi_update_itt_map(qedi, tid, task->itt, qedi_cmd);
if (nopout_hdr->ttt != ISCSI_TTT_ALL_ONES) {
nop_out_pdu_header.itt = be32_to_cpu(nopout_hdr->itt);
nop_out_pdu_header.ttt = be32_to_cpu(nopout_hdr->ttt);
} else {
nop_out_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
nop_out_pdu_header.ttt = ISCSI_TTT_ALL_ONES;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&qedi_cmd->io_cmd, &qedi_conn->active_cmd_list);
qedi_cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
}
/* Fill tx AHS and rx buffer */
if (data_len) {
tx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.req_bd_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.req_dma_addr);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.req_dma_addr >> 32);
tx_sgl_task_params.total_buffer_size = data_len;
tx_sgl_task_params.num_sges = 1;
rx_sgl_task_params.sgl =
(struct scsi_sge *)qedi_conn->gen_pdu.resp_bd_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(qedi_conn->gen_pdu.resp_dma_addr);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)qedi_conn->gen_pdu.resp_dma_addr >> 32);
rx_sgl_task_params.total_buffer_size = resp_sge->sge_len;
rx_sgl_task_params.num_sges = 1;
}
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = 0;
task_params.tx_io_size = data_len;
task_params.rx_io_size = resp_sge->sge_len;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
rval = init_initiator_nop_out_task(&task_params,
&nop_out_pdu_header,
&tx_sgl_task_params,
&rx_sgl_task_params);
if (rval)
return -1;
qedi_ring_doorbell(qedi_conn);
return 0;
}
static int qedi_split_bd(struct qedi_cmd *cmd, u64 addr, int sg_len,
int bd_index)
{
struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
int frag_size, sg_frags;
sg_frags = 0;
while (sg_len) {
if (addr % QEDI_PAGE_SIZE)
frag_size =
(QEDI_PAGE_SIZE - (addr % QEDI_PAGE_SIZE));
else
frag_size = (sg_len > QEDI_BD_SPLIT_SZ) ? 0 :
(sg_len % QEDI_BD_SPLIT_SZ);
if (frag_size == 0)
frag_size = QEDI_BD_SPLIT_SZ;
bd[bd_index + sg_frags].sge_addr.lo = (addr & 0xffffffff);
bd[bd_index + sg_frags].sge_addr.hi = (addr >> 32);
bd[bd_index + sg_frags].sge_len = (u16)frag_size;
QEDI_INFO(&cmd->conn->qedi->dbg_ctx, QEDI_LOG_IO,
"split sge %d: addr=%llx, len=%x",
(bd_index + sg_frags), addr, frag_size);
addr += (u64)frag_size;
sg_frags++;
sg_len -= frag_size;
}
return sg_frags;
}
static int qedi_map_scsi_sg(struct qedi_ctx *qedi, struct qedi_cmd *cmd)
{
struct scsi_cmnd *sc = cmd->scsi_cmd;
struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
struct scatterlist *sg;
int byte_count = 0;
int bd_count = 0;
int sg_count;
int sg_len;
int sg_frags;
u64 addr, end_addr;
int i;
WARN_ON(scsi_sg_count(sc) > QEDI_ISCSI_MAX_BDS_PER_CMD);
sg_count = dma_map_sg(&qedi->pdev->dev, scsi_sglist(sc),
scsi_sg_count(sc), sc->sc_data_direction);
/*
* New condition to send single SGE as cached-SGL.
* Single SGE with length less than 64K.
*/
sg = scsi_sglist(sc);
if ((sg_count == 1) && (sg_dma_len(sg) <= MAX_SGLEN_FOR_CACHESGL)) {
sg_len = sg_dma_len(sg);
addr = (u64)sg_dma_address(sg);
bd[bd_count].sge_addr.lo = (addr & 0xffffffff);
bd[bd_count].sge_addr.hi = (addr >> 32);
bd[bd_count].sge_len = (u16)sg_len;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO,
"single-cached-sgl: bd_count:%d addr=%llx, len=%x",
sg_count, addr, sg_len);
return ++bd_count;
}
scsi_for_each_sg(sc, sg, sg_count, i) {
sg_len = sg_dma_len(sg);
addr = (u64)sg_dma_address(sg);
end_addr = (addr + sg_len);
/*
* first sg elem in the 'list',
* check if end addr is page-aligned.
*/
if ((i == 0) && (sg_count > 1) && (end_addr % QEDI_PAGE_SIZE))
cmd->use_slowpath = true;
/*
* last sg elem in the 'list',
* check if start addr is page-aligned.
*/
else if ((i == (sg_count - 1)) &&
(sg_count > 1) && (addr % QEDI_PAGE_SIZE))
cmd->use_slowpath = true;
/*
* middle sg elements in list,
* check if start and end addr is page-aligned
*/
else if ((i != 0) && (i != (sg_count - 1)) &&
((addr % QEDI_PAGE_SIZE) ||
(end_addr % QEDI_PAGE_SIZE)))
cmd->use_slowpath = true;
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO, "sg[%d] size=0x%x",
i, sg_len);
if (sg_len > QEDI_BD_SPLIT_SZ) {
sg_frags = qedi_split_bd(cmd, addr, sg_len, bd_count);
} else {
sg_frags = 1;
bd[bd_count].sge_addr.lo = addr & 0xffffffff;
bd[bd_count].sge_addr.hi = addr >> 32;
bd[bd_count].sge_len = sg_len;
}
byte_count += sg_len;
bd_count += sg_frags;
}
if (byte_count != scsi_bufflen(sc))
QEDI_ERR(&qedi->dbg_ctx,
"byte_count = %d != scsi_bufflen = %d\n", byte_count,
scsi_bufflen(sc));
else
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO, "byte_count = %d\n",
byte_count);
WARN_ON(byte_count != scsi_bufflen(sc));
return bd_count;
}
static void qedi_iscsi_map_sg_list(struct qedi_cmd *cmd)
{
int bd_count;
struct scsi_cmnd *sc = cmd->scsi_cmd;
if (scsi_sg_count(sc)) {
bd_count = qedi_map_scsi_sg(cmd->conn->qedi, cmd);
if (bd_count == 0)
return;
} else {
struct scsi_sge *bd = cmd->io_tbl.sge_tbl;
bd[0].sge_addr.lo = 0;
bd[0].sge_addr.hi = 0;
bd[0].sge_len = 0;
bd_count = 0;
}
cmd->io_tbl.sge_valid = bd_count;
}
static void qedi_cpy_scsi_cdb(struct scsi_cmnd *sc, u32 *dstp)
{
u32 dword;
int lpcnt;
u8 *srcp;
lpcnt = sc->cmd_len / sizeof(dword);
srcp = (u8 *)sc->cmnd;
while (lpcnt--) {
memcpy(&dword, (const void *)srcp, 4);
*dstp = cpu_to_be32(dword);
srcp += 4;
dstp++;
}
if (sc->cmd_len & 0x3) {
dword = (u32)srcp[0] | ((u32)srcp[1] << 8);
*dstp = cpu_to_be32(dword);
}
}
void qedi_trace_io(struct qedi_ctx *qedi, struct iscsi_task *task,
u16 tid, int8_t direction)
{
struct qedi_io_log *io_log;
struct iscsi_conn *conn = task->conn;
struct qedi_conn *qedi_conn = conn->dd_data;
struct scsi_cmnd *sc_cmd = task->sc;
unsigned long flags;
spin_lock_irqsave(&qedi->io_trace_lock, flags);
io_log = &qedi->io_trace_buf[qedi->io_trace_idx];
io_log->direction = direction;
io_log->task_id = tid;
io_log->cid = qedi_conn->iscsi_conn_id;
io_log->lun = sc_cmd->device->lun;
io_log->op = sc_cmd->cmnd[0];
io_log->lba[0] = sc_cmd->cmnd[2];
io_log->lba[1] = sc_cmd->cmnd[3];
io_log->lba[2] = sc_cmd->cmnd[4];
io_log->lba[3] = sc_cmd->cmnd[5];
io_log->bufflen = scsi_bufflen(sc_cmd);
io_log->sg_count = scsi_sg_count(sc_cmd);
io_log->fast_sgs = qedi->fast_sgls;
io_log->cached_sgs = qedi->cached_sgls;
io_log->slow_sgs = qedi->slow_sgls;
io_log->cached_sge = qedi->use_cached_sge;
io_log->slow_sge = qedi->use_slow_sge;
io_log->fast_sge = qedi->use_fast_sge;
io_log->result = sc_cmd->result;
io_log->jiffies = jiffies;
io_log->blk_req_cpu = smp_processor_id();
if (direction == QEDI_IO_TRACE_REQ) {
/* For requests we only care about the submission CPU */
io_log->req_cpu = smp_processor_id() % qedi->num_queues;
io_log->intr_cpu = 0;
io_log->blk_rsp_cpu = 0;
} else if (direction == QEDI_IO_TRACE_RSP) {
io_log->req_cpu = smp_processor_id() % qedi->num_queues;
io_log->intr_cpu = qedi->intr_cpu;
io_log->blk_rsp_cpu = smp_processor_id();
}
qedi->io_trace_idx++;
if (qedi->io_trace_idx == QEDI_IO_TRACE_SIZE)
qedi->io_trace_idx = 0;
qedi->use_cached_sge = false;
qedi->use_slow_sge = false;
qedi->use_fast_sge = false;
spin_unlock_irqrestore(&qedi->io_trace_lock, flags);
}
int qedi_iscsi_send_ioreq(struct iscsi_task *task)
{
struct iscsi_conn *conn = task->conn;
struct iscsi_session *session = conn->session;
struct Scsi_Host *shost = iscsi_session_to_shost(session->cls_session);
struct qedi_ctx *qedi = iscsi_host_priv(shost);
struct qedi_conn *qedi_conn = conn->dd_data;
struct qedi_cmd *cmd = task->dd_data;
struct scsi_cmnd *sc = task->sc;
struct iscsi_cmd_hdr cmd_pdu_header;
struct scsi_sgl_task_params tx_sgl_task_params;
struct scsi_sgl_task_params rx_sgl_task_params;
struct scsi_sgl_task_params *prx_sgl = NULL;
struct scsi_sgl_task_params *ptx_sgl = NULL;
struct iscsi_task_params task_params;
struct iscsi_conn_params conn_params;
struct scsi_initiator_cmd_params cmd_params;
struct iscsi_task_context *fw_task_ctx;
struct iscsi_cls_conn *cls_conn;
struct iscsi_scsi_req *hdr = (struct iscsi_scsi_req *)task->hdr;
enum iscsi_task_type task_type = MAX_ISCSI_TASK_TYPE;
struct qedi_endpoint *ep;
u32 scsi_lun[2];
s16 tid = 0;
u16 sq_idx = 0;
u16 cq_idx;
int rval = 0;
ep = qedi_conn->ep;
cls_conn = qedi_conn->cls_conn;
conn = cls_conn->dd_data;
qedi_iscsi_map_sg_list(cmd);
int_to_scsilun(sc->device->lun, (struct scsi_lun *)scsi_lun);
tid = qedi_get_task_idx(qedi);
if (tid == -1)
return -ENOMEM;
fw_task_ctx =
(struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
tid);
memset(fw_task_ctx, 0, sizeof(struct iscsi_task_context));
cmd->task_id = tid;
memset(&task_params, 0, sizeof(task_params));
memset(&cmd_pdu_header, 0, sizeof(cmd_pdu_header));
memset(&tx_sgl_task_params, 0, sizeof(tx_sgl_task_params));
memset(&rx_sgl_task_params, 0, sizeof(rx_sgl_task_params));
memset(&conn_params, 0, sizeof(conn_params));
memset(&cmd_params, 0, sizeof(cmd_params));
cq_idx = smp_processor_id() % qedi->num_queues;
/* Update header info */
SET_FIELD(cmd_pdu_header.flags_attr, ISCSI_CMD_HDR_ATTR,
ISCSI_ATTR_SIMPLE);
if (hdr->cdb[0] != TEST_UNIT_READY) {
if (sc->sc_data_direction == DMA_TO_DEVICE) {
SET_FIELD(cmd_pdu_header.flags_attr,
ISCSI_CMD_HDR_WRITE, 1);
task_type = ISCSI_TASK_TYPE_INITIATOR_WRITE;
} else {
SET_FIELD(cmd_pdu_header.flags_attr,
ISCSI_CMD_HDR_READ, 1);
task_type = ISCSI_TASK_TYPE_INITIATOR_READ;
}
}
cmd_pdu_header.lun.lo = be32_to_cpu(scsi_lun[0]);
cmd_pdu_header.lun.hi = be32_to_cpu(scsi_lun[1]);
qedi_update_itt_map(qedi, tid, task->itt, cmd);
cmd_pdu_header.itt = qedi_set_itt(tid, get_itt(task->itt));
cmd_pdu_header.expected_transfer_length = cpu_to_be32(hdr->data_length);
cmd_pdu_header.hdr_second_dword = ntoh24(hdr->dlength);
cmd_pdu_header.cmd_sn = be32_to_cpu(hdr->cmdsn);
cmd_pdu_header.hdr_first_byte = hdr->opcode;
qedi_cpy_scsi_cdb(sc, (u32 *)cmd_pdu_header.cdb);
/* Fill tx AHS and rx buffer */
if (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE) {
tx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
tx_sgl_task_params.sgl_phys_addr.lo =
(u32)(cmd->io_tbl.sge_tbl_dma);
tx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32);
tx_sgl_task_params.total_buffer_size = scsi_bufflen(sc);
tx_sgl_task_params.num_sges = cmd->io_tbl.sge_valid;
if (cmd->use_slowpath)
tx_sgl_task_params.small_mid_sge = true;
} else if (task_type == ISCSI_TASK_TYPE_INITIATOR_READ) {
rx_sgl_task_params.sgl = cmd->io_tbl.sge_tbl;
rx_sgl_task_params.sgl_phys_addr.lo =
(u32)(cmd->io_tbl.sge_tbl_dma);
rx_sgl_task_params.sgl_phys_addr.hi =
(u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32);
rx_sgl_task_params.total_buffer_size = scsi_bufflen(sc);
rx_sgl_task_params.num_sges = cmd->io_tbl.sge_valid;
}
/* Add conn param */
conn_params.first_burst_length = conn->session->first_burst;
conn_params.max_send_pdu_length = conn->max_xmit_dlength;
conn_params.max_burst_length = conn->session->max_burst;
if (conn->session->initial_r2t_en)
conn_params.initial_r2t = true;
if (conn->session->imm_data_en)
conn_params.immediate_data = true;
/* Add cmd params */
cmd_params.sense_data_buffer_phys_addr.lo = (u32)cmd->sense_buffer_dma;
cmd_params.sense_data_buffer_phys_addr.hi =
(u32)((u64)cmd->sense_buffer_dma >> 32);
/* Fill fw input params */
task_params.context = fw_task_ctx;
task_params.conn_icid = (u16)qedi_conn->iscsi_conn_id;
task_params.itid = tid;
task_params.cq_rss_number = cq_idx;
if (task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE)
task_params.tx_io_size = scsi_bufflen(sc);
else if (task_type == ISCSI_TASK_TYPE_INITIATOR_READ)
task_params.rx_io_size = scsi_bufflen(sc);
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
QEDI_INFO(&qedi->dbg_ctx, QEDI_LOG_IO,
"%s: %s-SGL: sg_len=0x%x num_sges=0x%x first-sge-lo=0x%x first-sge-hi=0x%x\n",
(task_type == ISCSI_TASK_TYPE_INITIATOR_WRITE) ?
"Write " : "Read ", (cmd->io_tbl.sge_valid == 1) ?
"Single" : (cmd->use_slowpath ? "SLOW" : "FAST"),
(u16)cmd->io_tbl.sge_valid, scsi_bufflen(sc),
(u32)(cmd->io_tbl.sge_tbl_dma),
(u32)((u64)cmd->io_tbl.sge_tbl_dma >> 32));
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
if (task_params.tx_io_size != 0)
ptx_sgl = &tx_sgl_task_params;
if (task_params.rx_io_size != 0)
prx_sgl = &rx_sgl_task_params;
rval = init_initiator_rw_iscsi_task(&task_params, &conn_params,
&cmd_params, &cmd_pdu_header,
ptx_sgl, prx_sgl,
NULL);
if (rval)
return -1;
spin_lock(&qedi_conn->list_lock);
list_add_tail(&cmd->io_cmd, &qedi_conn->active_cmd_list);
cmd->io_cmd_in_list = true;
qedi_conn->active_cmd_count++;
spin_unlock(&qedi_conn->list_lock);
qedi_ring_doorbell(qedi_conn);
return 0;
}
int qedi_iscsi_cleanup_task(struct iscsi_task *task, bool mark_cmd_node_deleted)
{
struct iscsi_task_params task_params;
struct qedi_endpoint *ep;
struct iscsi_conn *conn = task->conn;
struct qedi_conn *qedi_conn = conn->dd_data;
struct qedi_cmd *cmd = task->dd_data;
u16 sq_idx = 0;
int rval = 0;
QEDI_INFO(&qedi_conn->qedi->dbg_ctx, QEDI_LOG_SCSI_TM,
"issue cleanup tid=0x%x itt=0x%x task_state=%d cmd_state=0%x cid=0x%x\n",
cmd->task_id, get_itt(task->itt), task->state,
cmd->state, qedi_conn->iscsi_conn_id);
memset(&task_params, 0, sizeof(task_params));
ep = qedi_conn->ep;
sq_idx = qedi_get_wqe_idx(qedi_conn);
task_params.sqe = &ep->sq[sq_idx];
memset(task_params.sqe, 0, sizeof(struct iscsi_wqe));
task_params.itid = cmd->task_id;
rval = init_cleanup_task(&task_params);
if (rval)
return rval;
qedi_ring_doorbell(qedi_conn);
return 0;
}