2017-02-12 21:52:37 +00:00
|
|
|
/*******************************************************************
|
|
|
|
* This file is part of the Emulex Linux Device Driver for *
|
2017-02-12 21:52:39 +00:00
|
|
|
* Fibre Channsel Host Bus Adapters. *
|
|
|
|
* Copyright (C) 2017 Broadcom. All Rights Reserved. The term *
|
|
|
|
* “Broadcom” refers to Broadcom Limited and/or its subsidiaries. *
|
2017-02-12 21:52:37 +00:00
|
|
|
* Copyright (C) 2004-2016 Emulex. All rights reserved. *
|
|
|
|
* EMULEX and SLI are trademarks of Emulex. *
|
2017-02-12 21:52:39 +00:00
|
|
|
* www.broadcom.com *
|
2017-02-12 21:52:37 +00:00
|
|
|
* Portions Copyright (C) 2004-2005 Christoph Hellwig *
|
|
|
|
* *
|
|
|
|
* This program is free software; you can redistribute it and/or *
|
|
|
|
* modify it under the terms of version 2 of the GNU General *
|
|
|
|
* Public License as published by the Free Software Foundation. *
|
|
|
|
* This program is distributed in the hope that it will be useful. *
|
|
|
|
* ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND *
|
|
|
|
* WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, *
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE *
|
|
|
|
* DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD *
|
|
|
|
* TO BE LEGALLY INVALID. See the GNU General Public License for *
|
|
|
|
* more details, a copy of which can be found in the file COPYING *
|
|
|
|
* included with this package. *
|
|
|
|
********************************************************************/
|
|
|
|
#include <linux/pci.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/delay.h>
|
|
|
|
#include <asm/unaligned.h>
|
|
|
|
#include <linux/crc-t10dif.h>
|
|
|
|
#include <net/checksum.h>
|
|
|
|
|
|
|
|
#include <scsi/scsi.h>
|
|
|
|
#include <scsi/scsi_device.h>
|
|
|
|
#include <scsi/scsi_eh.h>
|
|
|
|
#include <scsi/scsi_host.h>
|
|
|
|
#include <scsi/scsi_tcq.h>
|
|
|
|
#include <scsi/scsi_transport_fc.h>
|
|
|
|
#include <scsi/fc/fc_fs.h>
|
|
|
|
|
|
|
|
#include <../drivers/nvme/host/nvme.h>
|
|
|
|
#include <linux/nvme-fc-driver.h>
|
|
|
|
|
|
|
|
#include "lpfc_version.h"
|
|
|
|
#include "lpfc_hw4.h"
|
|
|
|
#include "lpfc_hw.h"
|
|
|
|
#include "lpfc_sli.h"
|
|
|
|
#include "lpfc_sli4.h"
|
|
|
|
#include "lpfc_nl.h"
|
|
|
|
#include "lpfc_disc.h"
|
|
|
|
#include "lpfc.h"
|
|
|
|
#include "lpfc_scsi.h"
|
|
|
|
#include "lpfc_nvme.h"
|
|
|
|
#include "lpfc_nvmet.h"
|
|
|
|
#include "lpfc_logmsg.h"
|
|
|
|
#include "lpfc_crtn.h"
|
|
|
|
#include "lpfc_vport.h"
|
2017-02-12 21:52:38 +00:00
|
|
|
#include "lpfc_debugfs.h"
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
static struct lpfc_iocbq *lpfc_nvmet_prep_ls_wqe(struct lpfc_hba *,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *,
|
|
|
|
dma_addr_t rspbuf,
|
|
|
|
uint16_t rspsize);
|
|
|
|
static struct lpfc_iocbq *lpfc_nvmet_prep_fcp_wqe(struct lpfc_hba *,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *);
|
|
|
|
static int lpfc_nvmet_sol_fcp_issue_abort(struct lpfc_hba *,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *,
|
|
|
|
uint32_t, uint16_t);
|
|
|
|
static int lpfc_nvmet_unsol_fcp_issue_abort(struct lpfc_hba *,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *,
|
|
|
|
uint32_t, uint16_t);
|
|
|
|
static int lpfc_nvmet_unsol_ls_issue_abort(struct lpfc_hba *,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *,
|
|
|
|
uint32_t, uint16_t);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_xmt_ls_rsp_cmp - Completion handler for LS Response
|
|
|
|
* @phba: Pointer to HBA context object.
|
|
|
|
* @cmdwqe: Pointer to driver command WQE object.
|
|
|
|
* @wcqe: Pointer to driver response CQE object.
|
|
|
|
*
|
|
|
|
* The function is called from SLI ring event handler with no
|
|
|
|
* lock held. This function is the completion handler for NVME LS commands
|
|
|
|
* The function frees memory resources used for the NVME commands.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_xmt_ls_rsp_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe,
|
|
|
|
struct lpfc_wcqe_complete *wcqe)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct nvmefc_tgt_ls_req *rsp;
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
uint32_t status, result;
|
|
|
|
|
|
|
|
status = bf_get(lpfc_wcqe_c_status, wcqe);
|
|
|
|
result = wcqe->parameter;
|
|
|
|
if (!phba->targetport)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
|
|
|
|
if (status)
|
|
|
|
atomic_inc(&tgtp->xmt_ls_rsp_error);
|
|
|
|
else
|
|
|
|
atomic_inc(&tgtp->xmt_ls_rsp_cmpl);
|
|
|
|
|
|
|
|
out:
|
|
|
|
ctxp = cmdwqe->context2;
|
|
|
|
rsp = &ctxp->ctx.ls_req;
|
|
|
|
|
2017-02-12 21:52:38 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET LS CMPL: xri x%x stat x%x result x%x\n",
|
|
|
|
ctxp->oxid, status, result);
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC,
|
|
|
|
"6038 %s: Entrypoint: ctx %p status %x/%x\n", __func__,
|
|
|
|
ctxp, status, result);
|
|
|
|
|
|
|
|
lpfc_nlp_put(cmdwqe->context1);
|
|
|
|
cmdwqe->context2 = NULL;
|
|
|
|
cmdwqe->context3 = NULL;
|
|
|
|
lpfc_sli_release_iocbq(phba, cmdwqe);
|
|
|
|
rsp->done(rsp);
|
|
|
|
kfree(ctxp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_rq_post - Repost a NVMET RQ DMA buffer and clean up context
|
|
|
|
* @phba: HBA buffer is associated with
|
|
|
|
* @ctxp: context to clean up
|
|
|
|
* @mp: Buffer to free
|
|
|
|
*
|
|
|
|
* Description: Frees the given DMA buffer in the appropriate way given by
|
|
|
|
* reposting it to its associated RQ so it can be reused.
|
|
|
|
*
|
|
|
|
* Notes: Takes phba->hbalock. Can be called with or without other locks held.
|
|
|
|
*
|
|
|
|
* Returns: None
|
|
|
|
**/
|
|
|
|
void
|
|
|
|
lpfc_nvmet_rq_post(struct lpfc_hba *phba, struct lpfc_nvmet_rcv_ctx *ctxp,
|
|
|
|
struct lpfc_dmabuf *mp)
|
|
|
|
{
|
|
|
|
if (ctxp) {
|
|
|
|
if (ctxp->txrdy) {
|
|
|
|
pci_pool_free(phba->txrdy_payload_pool, ctxp->txrdy,
|
|
|
|
ctxp->txrdy_phys);
|
|
|
|
ctxp->txrdy = NULL;
|
|
|
|
ctxp->txrdy_phys = 0;
|
|
|
|
}
|
|
|
|
ctxp->state = LPFC_NVMET_STE_FREE;
|
|
|
|
}
|
|
|
|
lpfc_rq_buf_free(phba, mp);
|
|
|
|
}
|
|
|
|
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_ktime(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp)
|
|
|
|
{
|
|
|
|
uint64_t seg1, seg2, seg3, seg4, seg5;
|
|
|
|
uint64_t seg6, seg7, seg8, seg9, seg10;
|
|
|
|
|
|
|
|
if (!phba->ktime_on)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (!ctxp->ts_isr_cmd || !ctxp->ts_cmd_nvme ||
|
|
|
|
!ctxp->ts_nvme_data || !ctxp->ts_data_wqput ||
|
|
|
|
!ctxp->ts_isr_data || !ctxp->ts_data_nvme ||
|
|
|
|
!ctxp->ts_nvme_status || !ctxp->ts_status_wqput ||
|
|
|
|
!ctxp->ts_isr_status || !ctxp->ts_status_nvme)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (ctxp->ts_isr_cmd > ctxp->ts_cmd_nvme)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_cmd_nvme > ctxp->ts_nvme_data)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_nvme_data > ctxp->ts_data_wqput)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_data_wqput > ctxp->ts_isr_data)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_isr_data > ctxp->ts_data_nvme)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_data_nvme > ctxp->ts_nvme_status)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_nvme_status > ctxp->ts_status_wqput)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_status_wqput > ctxp->ts_isr_status)
|
|
|
|
return;
|
|
|
|
if (ctxp->ts_isr_status > ctxp->ts_status_nvme)
|
|
|
|
return;
|
|
|
|
/*
|
|
|
|
* Segment 1 - Time from FCP command received by MSI-X ISR
|
|
|
|
* to FCP command is passed to NVME Layer.
|
|
|
|
* Segment 2 - Time from FCP command payload handed
|
|
|
|
* off to NVME Layer to Driver receives a Command op
|
|
|
|
* from NVME Layer.
|
|
|
|
* Segment 3 - Time from Driver receives a Command op
|
|
|
|
* from NVME Layer to Command is put on WQ.
|
|
|
|
* Segment 4 - Time from Driver WQ put is done
|
|
|
|
* to MSI-X ISR for Command cmpl.
|
|
|
|
* Segment 5 - Time from MSI-X ISR for Command cmpl to
|
|
|
|
* Command cmpl is passed to NVME Layer.
|
|
|
|
* Segment 6 - Time from Command cmpl is passed to NVME
|
|
|
|
* Layer to Driver receives a RSP op from NVME Layer.
|
|
|
|
* Segment 7 - Time from Driver receives a RSP op from
|
|
|
|
* NVME Layer to WQ put is done on TRSP FCP Status.
|
|
|
|
* Segment 8 - Time from Driver WQ put is done on TRSP
|
|
|
|
* FCP Status to MSI-X ISR for TRSP cmpl.
|
|
|
|
* Segment 9 - Time from MSI-X ISR for TRSP cmpl to
|
|
|
|
* TRSP cmpl is passed to NVME Layer.
|
|
|
|
* Segment 10 - Time from FCP command received by
|
|
|
|
* MSI-X ISR to command is completed on wire.
|
|
|
|
* (Segments 1 thru 8) for READDATA / WRITEDATA
|
|
|
|
* (Segments 1 thru 4) for READDATA_RSP
|
|
|
|
*/
|
|
|
|
seg1 = ctxp->ts_cmd_nvme - ctxp->ts_isr_cmd;
|
|
|
|
seg2 = (ctxp->ts_nvme_data - ctxp->ts_isr_cmd) - seg1;
|
|
|
|
seg3 = (ctxp->ts_data_wqput - ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2;
|
|
|
|
seg4 = (ctxp->ts_isr_data - ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2 - seg3;
|
|
|
|
seg5 = (ctxp->ts_data_nvme - ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2 - seg3 - seg4;
|
|
|
|
|
|
|
|
/* For auto rsp commands seg6 thru seg10 will be 0 */
|
|
|
|
if (ctxp->ts_nvme_status > ctxp->ts_data_nvme) {
|
|
|
|
seg6 = (ctxp->ts_nvme_status -
|
|
|
|
ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2 - seg3 - seg4 - seg5;
|
|
|
|
seg7 = (ctxp->ts_status_wqput -
|
|
|
|
ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2 - seg3 -
|
|
|
|
seg4 - seg5 - seg6;
|
|
|
|
seg8 = (ctxp->ts_isr_status -
|
|
|
|
ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2 - seg3 - seg4 -
|
|
|
|
seg5 - seg6 - seg7;
|
|
|
|
seg9 = (ctxp->ts_status_nvme -
|
|
|
|
ctxp->ts_isr_cmd) -
|
|
|
|
seg1 - seg2 - seg3 - seg4 -
|
|
|
|
seg5 - seg6 - seg7 - seg8;
|
|
|
|
seg10 = (ctxp->ts_isr_status -
|
|
|
|
ctxp->ts_isr_cmd);
|
|
|
|
} else {
|
|
|
|
seg6 = 0;
|
|
|
|
seg7 = 0;
|
|
|
|
seg8 = 0;
|
|
|
|
seg9 = 0;
|
|
|
|
seg10 = (ctxp->ts_isr_data - ctxp->ts_isr_cmd);
|
|
|
|
}
|
|
|
|
|
|
|
|
phba->ktime_seg1_total += seg1;
|
|
|
|
if (seg1 < phba->ktime_seg1_min)
|
|
|
|
phba->ktime_seg1_min = seg1;
|
|
|
|
else if (seg1 > phba->ktime_seg1_max)
|
|
|
|
phba->ktime_seg1_max = seg1;
|
|
|
|
|
|
|
|
phba->ktime_seg2_total += seg2;
|
|
|
|
if (seg2 < phba->ktime_seg2_min)
|
|
|
|
phba->ktime_seg2_min = seg2;
|
|
|
|
else if (seg2 > phba->ktime_seg2_max)
|
|
|
|
phba->ktime_seg2_max = seg2;
|
|
|
|
|
|
|
|
phba->ktime_seg3_total += seg3;
|
|
|
|
if (seg3 < phba->ktime_seg3_min)
|
|
|
|
phba->ktime_seg3_min = seg3;
|
|
|
|
else if (seg3 > phba->ktime_seg3_max)
|
|
|
|
phba->ktime_seg3_max = seg3;
|
|
|
|
|
|
|
|
phba->ktime_seg4_total += seg4;
|
|
|
|
if (seg4 < phba->ktime_seg4_min)
|
|
|
|
phba->ktime_seg4_min = seg4;
|
|
|
|
else if (seg4 > phba->ktime_seg4_max)
|
|
|
|
phba->ktime_seg4_max = seg4;
|
|
|
|
|
|
|
|
phba->ktime_seg5_total += seg5;
|
|
|
|
if (seg5 < phba->ktime_seg5_min)
|
|
|
|
phba->ktime_seg5_min = seg5;
|
|
|
|
else if (seg5 > phba->ktime_seg5_max)
|
|
|
|
phba->ktime_seg5_max = seg5;
|
|
|
|
|
|
|
|
phba->ktime_data_samples++;
|
|
|
|
if (!seg6)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
phba->ktime_seg6_total += seg6;
|
|
|
|
if (seg6 < phba->ktime_seg6_min)
|
|
|
|
phba->ktime_seg6_min = seg6;
|
|
|
|
else if (seg6 > phba->ktime_seg6_max)
|
|
|
|
phba->ktime_seg6_max = seg6;
|
|
|
|
|
|
|
|
phba->ktime_seg7_total += seg7;
|
|
|
|
if (seg7 < phba->ktime_seg7_min)
|
|
|
|
phba->ktime_seg7_min = seg7;
|
|
|
|
else if (seg7 > phba->ktime_seg7_max)
|
|
|
|
phba->ktime_seg7_max = seg7;
|
|
|
|
|
|
|
|
phba->ktime_seg8_total += seg8;
|
|
|
|
if (seg8 < phba->ktime_seg8_min)
|
|
|
|
phba->ktime_seg8_min = seg8;
|
|
|
|
else if (seg8 > phba->ktime_seg8_max)
|
|
|
|
phba->ktime_seg8_max = seg8;
|
|
|
|
|
|
|
|
phba->ktime_seg9_total += seg9;
|
|
|
|
if (seg9 < phba->ktime_seg9_min)
|
|
|
|
phba->ktime_seg9_min = seg9;
|
|
|
|
else if (seg9 > phba->ktime_seg9_max)
|
|
|
|
phba->ktime_seg9_max = seg9;
|
|
|
|
out:
|
|
|
|
phba->ktime_seg10_total += seg10;
|
|
|
|
if (seg10 < phba->ktime_seg10_min)
|
|
|
|
phba->ktime_seg10_min = seg10;
|
|
|
|
else if (seg10 > phba->ktime_seg10_max)
|
|
|
|
phba->ktime_seg10_max = seg10;
|
|
|
|
phba->ktime_status_samples++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
/**
|
|
|
|
* lpfc_nvmet_xmt_fcp_op_cmp - Completion handler for FCP Response
|
|
|
|
* @phba: Pointer to HBA context object.
|
|
|
|
* @cmdwqe: Pointer to driver command WQE object.
|
|
|
|
* @wcqe: Pointer to driver response CQE object.
|
|
|
|
*
|
|
|
|
* The function is called from SLI ring event handler with no
|
|
|
|
* lock held. This function is the completion handler for NVME FCP commands
|
|
|
|
* The function frees memory resources used for the NVME commands.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_xmt_fcp_op_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe,
|
|
|
|
struct lpfc_wcqe_complete *wcqe)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct nvmefc_tgt_fcp_req *rsp;
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
uint32_t status, result, op, start_clean;
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
uint32_t id;
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
ctxp = cmdwqe->context2;
|
|
|
|
rsp = &ctxp->ctx.fcp_req;
|
|
|
|
op = rsp->op;
|
|
|
|
ctxp->flag &= ~LPFC_NVMET_IO_INP;
|
|
|
|
|
|
|
|
status = bf_get(lpfc_wcqe_c_status, wcqe);
|
|
|
|
result = wcqe->parameter;
|
|
|
|
|
|
|
|
if (!phba->targetport)
|
|
|
|
goto out;
|
|
|
|
|
2017-02-12 21:52:38 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET FCP CMPL: xri x%x op x%x status x%x\n",
|
|
|
|
ctxp->oxid, op, status);
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
if (status) {
|
|
|
|
rsp->fcp_error = NVME_SC_DATA_XFER_ERROR;
|
|
|
|
rsp->transferred_length = 0;
|
|
|
|
atomic_inc(&tgtp->xmt_fcp_rsp_error);
|
|
|
|
} else {
|
|
|
|
rsp->fcp_error = NVME_SC_SUCCESS;
|
|
|
|
if (op == NVMET_FCOP_RSP)
|
|
|
|
rsp->transferred_length = rsp->rsplen;
|
|
|
|
else
|
|
|
|
rsp->transferred_length = rsp->transfer_length;
|
|
|
|
atomic_inc(&tgtp->xmt_fcp_rsp_cmpl);
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
if ((op == NVMET_FCOP_READDATA_RSP) ||
|
|
|
|
(op == NVMET_FCOP_RSP)) {
|
|
|
|
/* Sanity check */
|
|
|
|
ctxp->state = LPFC_NVMET_STE_DONE;
|
|
|
|
ctxp->entry_cnt++;
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
if (phba->ktime_on) {
|
|
|
|
if (rsp->op == NVMET_FCOP_READDATA_RSP) {
|
|
|
|
ctxp->ts_isr_data =
|
|
|
|
cmdwqe->isr_timestamp;
|
|
|
|
ctxp->ts_data_nvme =
|
|
|
|
ktime_get_ns();
|
|
|
|
ctxp->ts_nvme_status =
|
|
|
|
ctxp->ts_data_nvme;
|
|
|
|
ctxp->ts_status_wqput =
|
|
|
|
ctxp->ts_data_nvme;
|
|
|
|
ctxp->ts_isr_status =
|
|
|
|
ctxp->ts_data_nvme;
|
|
|
|
ctxp->ts_status_nvme =
|
|
|
|
ctxp->ts_data_nvme;
|
|
|
|
} else {
|
|
|
|
ctxp->ts_isr_status =
|
|
|
|
cmdwqe->isr_timestamp;
|
|
|
|
ctxp->ts_status_nvme =
|
|
|
|
ktime_get_ns();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (phba->cpucheck_on & LPFC_CHECK_NVMET_IO) {
|
|
|
|
id = smp_processor_id();
|
|
|
|
if (ctxp->cpu != id)
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6703 CPU Check cmpl: "
|
|
|
|
"cpu %d expect %d\n",
|
|
|
|
id, ctxp->cpu);
|
|
|
|
if (ctxp->cpu < LPFC_CHECK_CPU_CNT)
|
|
|
|
phba->cpucheck_cmpl_io[id]++;
|
|
|
|
}
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
rsp->done(rsp);
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
if (phba->ktime_on)
|
|
|
|
lpfc_nvmet_ktime(phba, ctxp);
|
|
|
|
#endif
|
2017-04-11 18:32:29 +00:00
|
|
|
/* lpfc_nvmet_xmt_fcp_release() will recycle the context */
|
2017-02-12 21:52:37 +00:00
|
|
|
} else {
|
|
|
|
ctxp->entry_cnt++;
|
|
|
|
start_clean = offsetof(struct lpfc_iocbq, wqe);
|
|
|
|
memset(((char *)cmdwqe) + start_clean, 0,
|
|
|
|
(sizeof(struct lpfc_iocbq) - start_clean));
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
if (phba->ktime_on) {
|
|
|
|
ctxp->ts_isr_data = cmdwqe->isr_timestamp;
|
|
|
|
ctxp->ts_data_nvme = ktime_get_ns();
|
|
|
|
}
|
|
|
|
if (phba->cpucheck_on & LPFC_CHECK_NVMET_IO) {
|
|
|
|
id = smp_processor_id();
|
|
|
|
if (ctxp->cpu != id)
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6704 CPU Check cmdcmpl: "
|
|
|
|
"cpu %d expect %d\n",
|
|
|
|
id, ctxp->cpu);
|
|
|
|
if (ctxp->cpu < LPFC_CHECK_CPU_CNT)
|
|
|
|
phba->cpucheck_ccmpl_io[id]++;
|
|
|
|
}
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
rsp->done(rsp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
lpfc_nvmet_xmt_ls_rsp(struct nvmet_fc_target_port *tgtport,
|
|
|
|
struct nvmefc_tgt_ls_req *rsp)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp =
|
|
|
|
container_of(rsp, struct lpfc_nvmet_rcv_ctx, ctx.ls_req);
|
|
|
|
struct lpfc_hba *phba = ctxp->phba;
|
|
|
|
struct hbq_dmabuf *nvmebuf =
|
|
|
|
(struct hbq_dmabuf *)ctxp->rqb_buffer;
|
|
|
|
struct lpfc_iocbq *nvmewqeq;
|
|
|
|
struct lpfc_nvmet_tgtport *nvmep = tgtport->private;
|
|
|
|
struct lpfc_dmabuf dmabuf;
|
|
|
|
struct ulp_bde64 bpl;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC,
|
|
|
|
"6023 %s: Entrypoint ctx %p %p\n", __func__,
|
|
|
|
ctxp, tgtport);
|
|
|
|
|
|
|
|
nvmewqeq = lpfc_nvmet_prep_ls_wqe(phba, ctxp, rsp->rspdma,
|
|
|
|
rsp->rsplen);
|
|
|
|
if (nvmewqeq == NULL) {
|
|
|
|
atomic_inc(&nvmep->xmt_ls_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6150 LS Drop IO x%x: Prep\n",
|
|
|
|
ctxp->oxid);
|
|
|
|
lpfc_in_buf_free(phba, &nvmebuf->dbuf);
|
|
|
|
lpfc_nvmet_unsol_ls_issue_abort(phba, ctxp,
|
|
|
|
ctxp->sid, ctxp->oxid);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Save numBdes for bpl2sgl */
|
|
|
|
nvmewqeq->rsvd2 = 1;
|
|
|
|
nvmewqeq->hba_wqidx = 0;
|
|
|
|
nvmewqeq->context3 = &dmabuf;
|
|
|
|
dmabuf.virt = &bpl;
|
|
|
|
bpl.addrLow = nvmewqeq->wqe.xmit_sequence.bde.addrLow;
|
|
|
|
bpl.addrHigh = nvmewqeq->wqe.xmit_sequence.bde.addrHigh;
|
|
|
|
bpl.tus.f.bdeSize = rsp->rsplen;
|
|
|
|
bpl.tus.f.bdeFlags = 0;
|
|
|
|
bpl.tus.w = le32_to_cpu(bpl.tus.w);
|
|
|
|
|
|
|
|
nvmewqeq->wqe_cmpl = lpfc_nvmet_xmt_ls_rsp_cmp;
|
|
|
|
nvmewqeq->iocb_cmpl = NULL;
|
|
|
|
nvmewqeq->context2 = ctxp;
|
|
|
|
|
2017-02-12 21:52:38 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET LS RESP: xri x%x wqidx x%x len x%x\n",
|
|
|
|
ctxp->oxid, nvmewqeq->hba_wqidx, rsp->rsplen);
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
rc = lpfc_sli4_issue_wqe(phba, LPFC_ELS_RING, nvmewqeq);
|
|
|
|
if (rc == WQE_SUCCESS) {
|
|
|
|
/*
|
|
|
|
* Okay to repost buffer here, but wait till cmpl
|
|
|
|
* before freeing ctxp and iocbq.
|
|
|
|
*/
|
|
|
|
lpfc_in_buf_free(phba, &nvmebuf->dbuf);
|
|
|
|
ctxp->rqb_buffer = 0;
|
|
|
|
atomic_inc(&nvmep->xmt_ls_rsp);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/* Give back resources */
|
|
|
|
atomic_inc(&nvmep->xmt_ls_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6151 LS Drop IO x%x: Issue %d\n",
|
|
|
|
ctxp->oxid, rc);
|
|
|
|
|
|
|
|
lpfc_nlp_put(nvmewqeq->context1);
|
|
|
|
|
|
|
|
lpfc_in_buf_free(phba, &nvmebuf->dbuf);
|
|
|
|
lpfc_nvmet_unsol_ls_issue_abort(phba, ctxp, ctxp->sid, ctxp->oxid);
|
|
|
|
return -ENXIO;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
lpfc_nvmet_xmt_fcp_op(struct nvmet_fc_target_port *tgtport,
|
|
|
|
struct nvmefc_tgt_fcp_req *rsp)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private;
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp =
|
|
|
|
container_of(rsp, struct lpfc_nvmet_rcv_ctx, ctx.fcp_req);
|
|
|
|
struct lpfc_hba *phba = ctxp->phba;
|
|
|
|
struct lpfc_iocbq *nvmewqeq;
|
2017-02-12 21:52:38 +00:00
|
|
|
int rc, id;
|
|
|
|
|
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
if (phba->ktime_on) {
|
|
|
|
if (rsp->op == NVMET_FCOP_RSP)
|
|
|
|
ctxp->ts_nvme_status = ktime_get_ns();
|
|
|
|
else
|
|
|
|
ctxp->ts_nvme_data = ktime_get_ns();
|
|
|
|
}
|
|
|
|
if (phba->cpucheck_on & LPFC_CHECK_NVMET_IO) {
|
|
|
|
id = smp_processor_id();
|
|
|
|
ctxp->cpu = id;
|
|
|
|
if (id < LPFC_CHECK_CPU_CNT)
|
|
|
|
phba->cpucheck_xmt_io[id]++;
|
|
|
|
if (rsp->hwqid != id) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6705 CPU Check OP: "
|
|
|
|
"cpu %d expect %d\n",
|
|
|
|
id, rsp->hwqid);
|
|
|
|
ctxp->cpu = rsp->hwqid;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
/* Sanity check */
|
|
|
|
if (ctxp->state == LPFC_NVMET_STE_ABORT) {
|
|
|
|
atomic_inc(&lpfc_nvmep->xmt_fcp_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6102 Bad state IO x%x aborted\n",
|
|
|
|
ctxp->oxid);
|
2017-03-04 17:30:28 +00:00
|
|
|
rc = -ENXIO;
|
2017-02-12 21:52:37 +00:00
|
|
|
goto aerr;
|
|
|
|
}
|
|
|
|
|
|
|
|
nvmewqeq = lpfc_nvmet_prep_fcp_wqe(phba, ctxp);
|
|
|
|
if (nvmewqeq == NULL) {
|
|
|
|
atomic_inc(&lpfc_nvmep->xmt_fcp_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6152 FCP Drop IO x%x: Prep\n",
|
|
|
|
ctxp->oxid);
|
2017-03-04 17:30:28 +00:00
|
|
|
rc = -ENXIO;
|
2017-02-12 21:52:37 +00:00
|
|
|
goto aerr;
|
|
|
|
}
|
|
|
|
|
|
|
|
nvmewqeq->wqe_cmpl = lpfc_nvmet_xmt_fcp_op_cmp;
|
|
|
|
nvmewqeq->iocb_cmpl = NULL;
|
|
|
|
nvmewqeq->context2 = ctxp;
|
|
|
|
nvmewqeq->iocb_flag |= LPFC_IO_NVMET;
|
|
|
|
ctxp->wqeq->hba_wqidx = rsp->hwqid;
|
|
|
|
|
2017-02-12 21:52:38 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET FCP CMND: xri x%x op x%x len x%x\n",
|
|
|
|
ctxp->oxid, rsp->op, rsp->rsplen);
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, nvmewqeq);
|
|
|
|
if (rc == WQE_SUCCESS) {
|
|
|
|
ctxp->flag |= LPFC_NVMET_IO_INP;
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
if (!phba->ktime_on)
|
|
|
|
return 0;
|
|
|
|
if (rsp->op == NVMET_FCOP_RSP)
|
|
|
|
ctxp->ts_status_wqput = ktime_get_ns();
|
|
|
|
else
|
|
|
|
ctxp->ts_data_wqput = ktime_get_ns();
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Give back resources */
|
|
|
|
atomic_inc(&lpfc_nvmep->xmt_fcp_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6153 FCP Drop IO x%x: Issue: %d\n",
|
|
|
|
ctxp->oxid, rc);
|
|
|
|
|
|
|
|
ctxp->wqeq->hba_wqidx = 0;
|
|
|
|
nvmewqeq->context2 = NULL;
|
|
|
|
nvmewqeq->context3 = NULL;
|
2017-03-04 17:30:28 +00:00
|
|
|
rc = -EBUSY;
|
2017-02-12 21:52:37 +00:00
|
|
|
aerr:
|
2017-03-04 17:30:28 +00:00
|
|
|
return rc;
|
2017-02-12 21:52:37 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_targetport_delete(struct nvmet_fc_target_port *targetport)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tport = targetport->private;
|
|
|
|
|
|
|
|
/* release any threads waiting for the unreg to complete */
|
|
|
|
complete(&tport->tport_unreg_done);
|
|
|
|
}
|
|
|
|
|
nvmet_fc: Rework target side abort handling
target transport:
----------------------
There are cases when there is a need to abort in-progress target
operations (writedata) so that controller termination or errors can
clean up. That can't happen currently as the abort is another target
op type, so it can't be used till the running one finishes (and it may
not). Solve by removing the abort op type and creating a separate
downcall from the transport to the lldd to request an io to be aborted.
The transport will abort ios on queue teardown or io errors. In general
the transport tries to call the lldd abort only when the io state is
idle. Meaning: ops that transmit data (readdata or rsp) will always
finish their transmit (or the lldd will see a state on the
link or initiator port that fails the transmit) and the done call for
the operation will occur. The transport will wait for the op done
upcall before calling the abort function, and as the io is idle, the
io can be cleaned up immediately after the abort call; Similarly, ios
that are not waiting for data or transmitting data must be in the nvmet
layer being processed. The transport will wait for the nvmet layer
completion before calling the abort function, and as the io is idle,
the io can be cleaned up immediately after the abort call; As for ops
that are waiting for data (writedata), they may be outstanding
indefinitely if the lldd doesn't see a condition where the initiatior
port or link is bad. In those cases, the transport will call the abort
function and wait for the lldd's op done upcall for the operation, where
it will then clean up the io.
Additionally, if a lldd receives an ABTS and matches it to an outstanding
request in the transport, A new new transport upcall was created to abort
the outstanding request in the transport. The transport expects any
outstanding op call (readdata or writedata) will completed by the lldd and
the operation upcall made. The transport doesn't act on the reported
abort (e.g. clean up the io) until an op done upcall occurs, a new op is
attempted, or the nvmet layer completes the io processing.
fcloop:
----------------------
Updated to support the new target apis.
On fcp io aborts from the initiator, the loopback context is updated to
NULL out the half that has completed. The initiator side is immediately
called after the abort request with an io completion (abort status).
On fcp io aborts from the target, the io is stopped and the initiator side
sees it as an aborted io. Target side ops, perhaps in progress while the
initiator side is done, continue but noop the data movement as there's no
structure on the initiator side to reference.
patch also contains:
----------------------
Revised lpfc to support the new abort api
commonized rsp buffer syncing and nulling of private data based on
calling paths.
errors in op done calls don't take action on the fod. They're bad
operations which implies the fod may be bad.
Signed-off-by: James Smart <james.smart@broadcom.com>
Signed-off-by: Sagi Grimberg <sagi@grimberg.me>
2017-04-11 18:32:31 +00:00
|
|
|
static void
|
|
|
|
lpfc_nvmet_xmt_fcp_abort(struct nvmet_fc_target_port *tgtport,
|
|
|
|
struct nvmefc_tgt_fcp_req *req)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private;
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp =
|
|
|
|
container_of(req, struct lpfc_nvmet_rcv_ctx, ctx.fcp_req);
|
|
|
|
struct lpfc_hba *phba = ctxp->phba;
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS,
|
|
|
|
"6103 Abort op: oxri x%x %d cnt %d\n",
|
|
|
|
ctxp->oxid, ctxp->state, ctxp->entry_cnt);
|
|
|
|
|
|
|
|
lpfc_nvmeio_data(phba, "NVMET FCP ABRT: xri x%x state x%x cnt x%x\n",
|
|
|
|
ctxp->oxid, ctxp->state, ctxp->entry_cnt);
|
|
|
|
|
|
|
|
atomic_inc(&lpfc_nvmep->xmt_fcp_abort);
|
|
|
|
ctxp->entry_cnt++;
|
|
|
|
ctxp->flag |= LPFC_NVMET_ABORT_OP;
|
|
|
|
if (ctxp->flag & LPFC_NVMET_IO_INP)
|
|
|
|
lpfc_nvmet_sol_fcp_issue_abort(phba, ctxp, ctxp->sid,
|
|
|
|
ctxp->oxid);
|
|
|
|
else
|
|
|
|
lpfc_nvmet_unsol_fcp_issue_abort(phba, ctxp, ctxp->sid,
|
|
|
|
ctxp->oxid);
|
|
|
|
}
|
|
|
|
|
2017-04-11 18:32:29 +00:00
|
|
|
static void
|
|
|
|
lpfc_nvmet_xmt_fcp_release(struct nvmet_fc_target_port *tgtport,
|
|
|
|
struct nvmefc_tgt_fcp_req *rsp)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *lpfc_nvmep = tgtport->private;
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp =
|
|
|
|
container_of(rsp, struct lpfc_nvmet_rcv_ctx, ctx.fcp_req);
|
|
|
|
struct lpfc_hba *phba = ctxp->phba;
|
|
|
|
unsigned long flags;
|
|
|
|
bool aborting = false;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&ctxp->ctxlock, flags);
|
|
|
|
if (ctxp->flag & LPFC_NVMET_ABORT_OP) {
|
|
|
|
aborting = true;
|
|
|
|
ctxp->flag |= LPFC_NVMET_CTX_RLS;
|
|
|
|
}
|
|
|
|
spin_unlock_irqrestore(&ctxp->ctxlock, flags);
|
|
|
|
|
|
|
|
if (aborting)
|
|
|
|
/* let the abort path do the real release */
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* Sanity check */
|
|
|
|
if (ctxp->state != LPFC_NVMET_STE_DONE) {
|
|
|
|
atomic_inc(&lpfc_nvmep->xmt_fcp_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6117 Bad state IO x%x aborted\n",
|
|
|
|
ctxp->oxid);
|
|
|
|
}
|
|
|
|
|
|
|
|
lpfc_nvmeio_data(phba, "NVMET FCP FREE: xri x%x ste %d\n", ctxp->oxid,
|
|
|
|
ctxp->state, 0);
|
|
|
|
|
|
|
|
lpfc_nvmet_rq_post(phba, ctxp, &ctxp->rqb_buffer->hbuf);
|
|
|
|
}
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
static struct nvmet_fc_target_template lpfc_tgttemplate = {
|
|
|
|
.targetport_delete = lpfc_nvmet_targetport_delete,
|
|
|
|
.xmt_ls_rsp = lpfc_nvmet_xmt_ls_rsp,
|
|
|
|
.fcp_op = lpfc_nvmet_xmt_fcp_op,
|
nvmet_fc: Rework target side abort handling
target transport:
----------------------
There are cases when there is a need to abort in-progress target
operations (writedata) so that controller termination or errors can
clean up. That can't happen currently as the abort is another target
op type, so it can't be used till the running one finishes (and it may
not). Solve by removing the abort op type and creating a separate
downcall from the transport to the lldd to request an io to be aborted.
The transport will abort ios on queue teardown or io errors. In general
the transport tries to call the lldd abort only when the io state is
idle. Meaning: ops that transmit data (readdata or rsp) will always
finish their transmit (or the lldd will see a state on the
link or initiator port that fails the transmit) and the done call for
the operation will occur. The transport will wait for the op done
upcall before calling the abort function, and as the io is idle, the
io can be cleaned up immediately after the abort call; Similarly, ios
that are not waiting for data or transmitting data must be in the nvmet
layer being processed. The transport will wait for the nvmet layer
completion before calling the abort function, and as the io is idle,
the io can be cleaned up immediately after the abort call; As for ops
that are waiting for data (writedata), they may be outstanding
indefinitely if the lldd doesn't see a condition where the initiatior
port or link is bad. In those cases, the transport will call the abort
function and wait for the lldd's op done upcall for the operation, where
it will then clean up the io.
Additionally, if a lldd receives an ABTS and matches it to an outstanding
request in the transport, A new new transport upcall was created to abort
the outstanding request in the transport. The transport expects any
outstanding op call (readdata or writedata) will completed by the lldd and
the operation upcall made. The transport doesn't act on the reported
abort (e.g. clean up the io) until an op done upcall occurs, a new op is
attempted, or the nvmet layer completes the io processing.
fcloop:
----------------------
Updated to support the new target apis.
On fcp io aborts from the initiator, the loopback context is updated to
NULL out the half that has completed. The initiator side is immediately
called after the abort request with an io completion (abort status).
On fcp io aborts from the target, the io is stopped and the initiator side
sees it as an aborted io. Target side ops, perhaps in progress while the
initiator side is done, continue but noop the data movement as there's no
structure on the initiator side to reference.
patch also contains:
----------------------
Revised lpfc to support the new abort api
commonized rsp buffer syncing and nulling of private data based on
calling paths.
errors in op done calls don't take action on the fod. They're bad
operations which implies the fod may be bad.
Signed-off-by: James Smart <james.smart@broadcom.com>
Signed-off-by: Sagi Grimberg <sagi@grimberg.me>
2017-04-11 18:32:31 +00:00
|
|
|
.fcp_abort = lpfc_nvmet_xmt_fcp_abort,
|
2017-04-11 18:32:29 +00:00
|
|
|
.fcp_req_release = lpfc_nvmet_xmt_fcp_release,
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
.max_hw_queues = 1,
|
|
|
|
.max_sgl_segments = LPFC_NVMET_DEFAULT_SEGS,
|
|
|
|
.max_dif_sgl_segments = LPFC_NVMET_DEFAULT_SEGS,
|
|
|
|
.dma_boundary = 0xFFFFFFFF,
|
|
|
|
|
|
|
|
/* optional features */
|
|
|
|
.target_features = 0,
|
|
|
|
/* sizes of additional private data for data structures */
|
|
|
|
.target_priv_sz = sizeof(struct lpfc_nvmet_tgtport),
|
|
|
|
};
|
|
|
|
|
|
|
|
int
|
|
|
|
lpfc_nvmet_create_targetport(struct lpfc_hba *phba)
|
|
|
|
{
|
|
|
|
struct lpfc_vport *vport = phba->pport;
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct nvmet_fc_port_info pinfo;
|
|
|
|
int error = 0;
|
|
|
|
|
|
|
|
if (phba->targetport)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
memset(&pinfo, 0, sizeof(struct nvmet_fc_port_info));
|
|
|
|
pinfo.node_name = wwn_to_u64(vport->fc_nodename.u.wwn);
|
|
|
|
pinfo.port_name = wwn_to_u64(vport->fc_portname.u.wwn);
|
|
|
|
pinfo.port_id = vport->fc_myDID;
|
|
|
|
|
2017-04-21 23:05:01 +00:00
|
|
|
/* Limit to LPFC_MAX_NVME_SEG_CNT.
|
|
|
|
* For now need + 1 to get around NVME transport logic.
|
|
|
|
*/
|
|
|
|
if (phba->cfg_sg_seg_cnt > LPFC_MAX_NVME_SEG_CNT) {
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME | LOG_INIT,
|
|
|
|
"6400 Reducing sg segment cnt to %d\n",
|
|
|
|
LPFC_MAX_NVME_SEG_CNT);
|
|
|
|
phba->cfg_nvme_seg_cnt = LPFC_MAX_NVME_SEG_CNT;
|
|
|
|
} else {
|
|
|
|
phba->cfg_nvme_seg_cnt = phba->cfg_sg_seg_cnt;
|
|
|
|
}
|
|
|
|
lpfc_tgttemplate.max_sgl_segments = phba->cfg_nvme_seg_cnt + 1;
|
2017-02-12 21:52:37 +00:00
|
|
|
lpfc_tgttemplate.max_hw_queues = phba->cfg_nvme_io_channel;
|
|
|
|
lpfc_tgttemplate.target_features = NVMET_FCTGTFEAT_READDATA_RSP |
|
nvmet_fc: add target feature flags for upcall isr contexts
Two new feature flags were added to control whether upcalls to the
transport result in context switches or stay in the calling context.
NVMET_FCTGTFEAT_CMD_IN_ISR:
By default, if the flag is not set, the transport assumes the
lldd is in a non-isr context and in the cpu context it should be
for the io queue. As such, the cmd handler is called directly in the
calling context.
If the flag is set, indicating the upcall is an isr context, the
transport mandates a transition to a workqueue. The workqueue assigned
to the queue is used for the context.
NVMET_FCTGTFEAT_OPDONE_IN_ISR
By default, if the flag is not set, the transport assumes the
lldd is in a non-isr context and in the cpu context it should be
for the io queue. As such, the fcp operation done callback is called
directly in the calling context.
If the flag is set, indicating the upcall is an isr context, the
transport mandates a transition to a workqueue. The workqueue assigned
to the queue is used for the context.
Updated lpfc for flags
Signed-off-by: James Smart <james.smart@broadcom.com>
Signed-off-by: Sagi Grimberg <sagi@grimberg.me>
2017-04-11 18:32:28 +00:00
|
|
|
NVMET_FCTGTFEAT_NEEDS_CMD_CPUSCHED |
|
|
|
|
NVMET_FCTGTFEAT_CMD_IN_ISR |
|
|
|
|
NVMET_FCTGTFEAT_OPDONE_IN_ISR;
|
2017-02-12 21:52:37 +00:00
|
|
|
|
2017-03-08 22:36:01 +00:00
|
|
|
#if (IS_ENABLED(CONFIG_NVME_TARGET_FC))
|
2017-02-12 21:52:37 +00:00
|
|
|
error = nvmet_fc_register_targetport(&pinfo, &lpfc_tgttemplate,
|
|
|
|
&phba->pcidev->dev,
|
|
|
|
&phba->targetport);
|
2017-03-04 17:30:33 +00:00
|
|
|
#else
|
|
|
|
error = -ENOMEM;
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
if (error) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC,
|
|
|
|
"6025 Cannot register NVME targetport "
|
|
|
|
"x%x\n", error);
|
|
|
|
phba->targetport = NULL;
|
|
|
|
} else {
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)
|
|
|
|
phba->targetport->private;
|
|
|
|
tgtp->phba = phba;
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC,
|
|
|
|
"6026 Registered NVME "
|
|
|
|
"targetport: %p, private %p "
|
|
|
|
"portnm %llx nodenm %llx\n",
|
|
|
|
phba->targetport, tgtp,
|
|
|
|
pinfo.port_name, pinfo.node_name);
|
|
|
|
|
|
|
|
atomic_set(&tgtp->rcv_ls_req_in, 0);
|
|
|
|
atomic_set(&tgtp->rcv_ls_req_out, 0);
|
|
|
|
atomic_set(&tgtp->rcv_ls_req_drop, 0);
|
|
|
|
atomic_set(&tgtp->xmt_ls_abort, 0);
|
|
|
|
atomic_set(&tgtp->xmt_ls_rsp, 0);
|
|
|
|
atomic_set(&tgtp->xmt_ls_drop, 0);
|
|
|
|
atomic_set(&tgtp->xmt_ls_rsp_error, 0);
|
|
|
|
atomic_set(&tgtp->xmt_ls_rsp_cmpl, 0);
|
|
|
|
atomic_set(&tgtp->rcv_fcp_cmd_in, 0);
|
|
|
|
atomic_set(&tgtp->rcv_fcp_cmd_out, 0);
|
|
|
|
atomic_set(&tgtp->rcv_fcp_cmd_drop, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_abort, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_drop, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_read_rsp, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_read, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_write, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_rsp, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_rsp_cmpl, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_rsp_error, 0);
|
|
|
|
atomic_set(&tgtp->xmt_fcp_rsp_drop, 0);
|
|
|
|
atomic_set(&tgtp->xmt_abort_rsp, 0);
|
|
|
|
atomic_set(&tgtp->xmt_abort_rsp_error, 0);
|
|
|
|
atomic_set(&tgtp->xmt_abort_cmpl, 0);
|
|
|
|
}
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
lpfc_nvmet_update_targetport(struct lpfc_hba *phba)
|
|
|
|
{
|
|
|
|
struct lpfc_vport *vport = phba->pport;
|
|
|
|
|
|
|
|
if (!phba->targetport)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
lpfc_printf_vlog(vport, KERN_INFO, LOG_NVME,
|
|
|
|
"6007 Update NVMET port %p did x%x\n",
|
|
|
|
phba->targetport, vport->fc_myDID);
|
|
|
|
|
|
|
|
phba->targetport->port_id = vport->fc_myDID;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-03-04 17:30:30 +00:00
|
|
|
/**
|
|
|
|
* lpfc_sli4_nvmet_xri_aborted - Fast-path process of nvmet xri abort
|
|
|
|
* @phba: pointer to lpfc hba data structure.
|
|
|
|
* @axri: pointer to the nvmet xri abort wcqe structure.
|
|
|
|
*
|
|
|
|
* This routine is invoked by the worker thread to process a SLI4 fast-path
|
|
|
|
* NVMET aborted xri.
|
|
|
|
**/
|
|
|
|
void
|
|
|
|
lpfc_sli4_nvmet_xri_aborted(struct lpfc_hba *phba,
|
|
|
|
struct sli4_wcqe_xri_aborted *axri)
|
|
|
|
{
|
|
|
|
/* TODO: work in progress */
|
|
|
|
}
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
void
|
|
|
|
lpfc_nvmet_destroy_targetport(struct lpfc_hba *phba)
|
|
|
|
{
|
2017-03-08 22:36:01 +00:00
|
|
|
#if (IS_ENABLED(CONFIG_NVME_TARGET_FC))
|
2017-02-12 21:52:37 +00:00
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
|
|
|
|
if (phba->nvmet_support == 0)
|
|
|
|
return;
|
|
|
|
if (phba->targetport) {
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
init_completion(&tgtp->tport_unreg_done);
|
|
|
|
nvmet_fc_unregister_targetport(phba->targetport);
|
|
|
|
wait_for_completion_timeout(&tgtp->tport_unreg_done, 5);
|
|
|
|
}
|
|
|
|
phba->targetport = NULL;
|
2017-03-04 17:30:33 +00:00
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_unsol_ls_buffer - Process an unsolicited event data buffer
|
|
|
|
* @phba: pointer to lpfc hba data structure.
|
|
|
|
* @pring: pointer to a SLI ring.
|
|
|
|
* @nvmebuf: pointer to lpfc nvme command HBQ data structure.
|
|
|
|
*
|
|
|
|
* This routine is used for processing the WQE associated with a unsolicited
|
|
|
|
* event. It first determines whether there is an existing ndlp that matches
|
|
|
|
* the DID from the unsolicited WQE. If not, it will create a new one with
|
|
|
|
* the DID from the unsolicited WQE. The ELS command from the unsolicited
|
|
|
|
* WQE is then used to invoke the proper routine and to set up proper state
|
|
|
|
* of the discovery state machine.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_unsol_ls_buffer(struct lpfc_hba *phba, struct lpfc_sli_ring *pring,
|
|
|
|
struct hbq_dmabuf *nvmebuf)
|
|
|
|
{
|
2017-03-08 22:36:01 +00:00
|
|
|
#if (IS_ENABLED(CONFIG_NVME_TARGET_FC))
|
2017-02-12 21:52:37 +00:00
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct fc_frame_header *fc_hdr;
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
uint32_t *payload;
|
|
|
|
uint32_t size, oxid, sid, rc;
|
|
|
|
|
|
|
|
if (!nvmebuf || !phba->targetport) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6154 LS Drop IO\n");
|
2017-02-12 21:52:38 +00:00
|
|
|
oxid = 0;
|
|
|
|
size = 0;
|
|
|
|
sid = 0;
|
2017-02-12 21:52:37 +00:00
|
|
|
goto dropit;
|
|
|
|
}
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
payload = (uint32_t *)(nvmebuf->dbuf.virt);
|
|
|
|
fc_hdr = (struct fc_frame_header *)(nvmebuf->hbuf.virt);
|
|
|
|
size = bf_get(lpfc_rcqe_length, &nvmebuf->cq_event.cqe.rcqe_cmpl);
|
|
|
|
oxid = be16_to_cpu(fc_hdr->fh_ox_id);
|
|
|
|
sid = sli4_sid_from_fc_hdr(fc_hdr);
|
|
|
|
|
|
|
|
ctxp = kzalloc(sizeof(struct lpfc_nvmet_rcv_ctx), GFP_ATOMIC);
|
|
|
|
if (ctxp == NULL) {
|
|
|
|
atomic_inc(&tgtp->rcv_ls_req_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6155 LS Drop IO x%x: Alloc\n",
|
|
|
|
oxid);
|
|
|
|
dropit:
|
2017-02-12 21:52:38 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET LS DROP: "
|
|
|
|
"xri x%x sz %d from %06x\n",
|
|
|
|
oxid, size, sid);
|
2017-02-12 21:52:37 +00:00
|
|
|
if (nvmebuf)
|
|
|
|
lpfc_in_buf_free(phba, &nvmebuf->dbuf);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
ctxp->phba = phba;
|
|
|
|
ctxp->size = size;
|
|
|
|
ctxp->oxid = oxid;
|
|
|
|
ctxp->sid = sid;
|
|
|
|
ctxp->wqeq = NULL;
|
|
|
|
ctxp->state = LPFC_NVMET_STE_RCV;
|
|
|
|
ctxp->rqb_buffer = (void *)nvmebuf;
|
2017-02-12 21:52:38 +00:00
|
|
|
|
|
|
|
lpfc_nvmeio_data(phba, "NVMET LS RCV: xri x%x sz %d from %06x\n",
|
|
|
|
oxid, size, sid);
|
2017-02-12 21:52:37 +00:00
|
|
|
/*
|
|
|
|
* The calling sequence should be:
|
|
|
|
* nvmet_fc_rcv_ls_req -> lpfc_nvmet_xmt_ls_rsp/cmp ->_req->done
|
|
|
|
* lpfc_nvmet_xmt_ls_rsp_cmp should free the allocated ctxp.
|
|
|
|
*/
|
|
|
|
atomic_inc(&tgtp->rcv_ls_req_in);
|
|
|
|
rc = nvmet_fc_rcv_ls_req(phba->targetport, &ctxp->ctx.ls_req,
|
|
|
|
payload, size);
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC,
|
|
|
|
"6037 %s: ctx %p sz %d rc %d: %08x %08x %08x "
|
|
|
|
"%08x %08x %08x\n", __func__, ctxp, size, rc,
|
|
|
|
*payload, *(payload+1), *(payload+2),
|
|
|
|
*(payload+3), *(payload+4), *(payload+5));
|
2017-02-12 21:52:38 +00:00
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
if (rc == 0) {
|
|
|
|
atomic_inc(&tgtp->rcv_ls_req_out);
|
|
|
|
return;
|
|
|
|
}
|
2017-02-12 21:52:38 +00:00
|
|
|
|
|
|
|
lpfc_nvmeio_data(phba, "NVMET LS DROP: xri x%x sz %d from %06x\n",
|
|
|
|
oxid, size, sid);
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
atomic_inc(&tgtp->rcv_ls_req_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6156 LS Drop IO x%x: nvmet_fc_rcv_ls_req %d\n",
|
|
|
|
ctxp->oxid, rc);
|
|
|
|
|
|
|
|
/* We assume a rcv'ed cmd ALWAYs fits into 1 buffer */
|
|
|
|
if (nvmebuf)
|
|
|
|
lpfc_in_buf_free(phba, &nvmebuf->dbuf);
|
|
|
|
|
|
|
|
atomic_inc(&tgtp->xmt_ls_abort);
|
|
|
|
lpfc_nvmet_unsol_ls_issue_abort(phba, ctxp, sid, oxid);
|
2017-03-04 17:30:33 +00:00
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_unsol_fcp_buffer - Process an unsolicited event data buffer
|
|
|
|
* @phba: pointer to lpfc hba data structure.
|
|
|
|
* @pring: pointer to a SLI ring.
|
|
|
|
* @nvmebuf: pointer to lpfc nvme command HBQ data structure.
|
|
|
|
*
|
|
|
|
* This routine is used for processing the WQE associated with a unsolicited
|
|
|
|
* event. It first determines whether there is an existing ndlp that matches
|
|
|
|
* the DID from the unsolicited WQE. If not, it will create a new one with
|
|
|
|
* the DID from the unsolicited WQE. The ELS command from the unsolicited
|
|
|
|
* WQE is then used to invoke the proper routine and to set up proper state
|
|
|
|
* of the discovery state machine.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_unsol_fcp_buffer(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_sli_ring *pring,
|
|
|
|
struct rqb_dmabuf *nvmebuf,
|
|
|
|
uint64_t isr_timestamp)
|
|
|
|
{
|
2017-03-08 22:36:01 +00:00
|
|
|
#if (IS_ENABLED(CONFIG_NVME_TARGET_FC))
|
2017-02-12 21:52:37 +00:00
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct fc_frame_header *fc_hdr;
|
|
|
|
uint32_t *payload;
|
|
|
|
uint32_t size, oxid, sid, rc;
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
uint32_t id;
|
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
if (!nvmebuf || !phba->targetport) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6157 FCP Drop IO\n");
|
2017-02-12 21:52:38 +00:00
|
|
|
oxid = 0;
|
|
|
|
size = 0;
|
|
|
|
sid = 0;
|
2017-02-12 21:52:37 +00:00
|
|
|
goto dropit;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
payload = (uint32_t *)(nvmebuf->dbuf.virt);
|
|
|
|
fc_hdr = (struct fc_frame_header *)(nvmebuf->hbuf.virt);
|
|
|
|
size = nvmebuf->bytes_recv;
|
|
|
|
oxid = be16_to_cpu(fc_hdr->fh_ox_id);
|
|
|
|
sid = sli4_sid_from_fc_hdr(fc_hdr);
|
|
|
|
|
|
|
|
ctxp = (struct lpfc_nvmet_rcv_ctx *)nvmebuf->context;
|
|
|
|
if (ctxp == NULL) {
|
|
|
|
atomic_inc(&tgtp->rcv_fcp_cmd_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6158 FCP Drop IO x%x: Alloc\n",
|
|
|
|
oxid);
|
|
|
|
lpfc_nvmet_rq_post(phba, NULL, &nvmebuf->hbuf);
|
|
|
|
/* Cannot send ABTS without context */
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
memset(ctxp, 0, sizeof(ctxp->ctx));
|
|
|
|
ctxp->wqeq = NULL;
|
|
|
|
ctxp->txrdy = NULL;
|
|
|
|
ctxp->offset = 0;
|
|
|
|
ctxp->phba = phba;
|
|
|
|
ctxp->size = size;
|
|
|
|
ctxp->oxid = oxid;
|
|
|
|
ctxp->sid = sid;
|
|
|
|
ctxp->state = LPFC_NVMET_STE_RCV;
|
|
|
|
ctxp->rqb_buffer = nvmebuf;
|
|
|
|
ctxp->entry_cnt = 1;
|
|
|
|
ctxp->flag = 0;
|
2017-04-21 23:04:59 +00:00
|
|
|
spin_lock_init(&ctxp->ctxlock);
|
2017-02-12 21:52:37 +00:00
|
|
|
|
2017-02-12 21:52:38 +00:00
|
|
|
#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
|
|
|
|
if (phba->ktime_on) {
|
|
|
|
ctxp->ts_isr_cmd = isr_timestamp;
|
|
|
|
ctxp->ts_cmd_nvme = ktime_get_ns();
|
|
|
|
ctxp->ts_nvme_data = 0;
|
|
|
|
ctxp->ts_data_wqput = 0;
|
|
|
|
ctxp->ts_isr_data = 0;
|
|
|
|
ctxp->ts_data_nvme = 0;
|
|
|
|
ctxp->ts_nvme_status = 0;
|
|
|
|
ctxp->ts_status_wqput = 0;
|
|
|
|
ctxp->ts_isr_status = 0;
|
|
|
|
ctxp->ts_status_nvme = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (phba->cpucheck_on & LPFC_CHECK_NVMET_RCV) {
|
|
|
|
id = smp_processor_id();
|
|
|
|
if (id < LPFC_CHECK_CPU_CNT)
|
|
|
|
phba->cpucheck_rcv_io[id]++;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2017-04-21 23:04:59 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET FCP RCV: xri x%x sz %d CPU %02x\n",
|
|
|
|
oxid, size, smp_processor_id());
|
2017-02-12 21:52:38 +00:00
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
atomic_inc(&tgtp->rcv_fcp_cmd_in);
|
|
|
|
/*
|
|
|
|
* The calling sequence should be:
|
|
|
|
* nvmet_fc_rcv_fcp_req -> lpfc_nvmet_xmt_fcp_op/cmp -> req->done
|
|
|
|
* lpfc_nvmet_xmt_fcp_op_cmp should free the allocated ctxp.
|
|
|
|
*/
|
|
|
|
rc = nvmet_fc_rcv_fcp_req(phba->targetport, &ctxp->ctx.fcp_req,
|
|
|
|
payload, size);
|
|
|
|
|
|
|
|
/* Process FCP command */
|
|
|
|
if (rc == 0) {
|
|
|
|
atomic_inc(&tgtp->rcv_fcp_cmd_out);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
atomic_inc(&tgtp->rcv_fcp_cmd_drop);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
2017-03-04 17:30:30 +00:00
|
|
|
"6159 FCP Drop IO x%x: err x%x\n",
|
2017-02-12 21:52:37 +00:00
|
|
|
ctxp->oxid, rc);
|
|
|
|
dropit:
|
2017-02-12 21:52:38 +00:00
|
|
|
lpfc_nvmeio_data(phba, "NVMET FCP DROP: xri x%x sz %d from %06x\n",
|
|
|
|
oxid, size, sid);
|
2017-02-12 21:52:37 +00:00
|
|
|
if (oxid) {
|
|
|
|
lpfc_nvmet_unsol_fcp_issue_abort(phba, ctxp, sid, oxid);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (nvmebuf) {
|
|
|
|
nvmebuf->iocbq->hba_wqidx = 0;
|
|
|
|
/* We assume a rcv'ed cmd ALWAYs fits into 1 buffer */
|
|
|
|
lpfc_nvmet_rq_post(phba, NULL, &nvmebuf->hbuf);
|
|
|
|
}
|
2017-03-04 17:30:33 +00:00
|
|
|
#endif
|
2017-02-12 21:52:37 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_unsol_ls_event - Process an unsolicited event from an nvme nport
|
|
|
|
* @phba: pointer to lpfc hba data structure.
|
|
|
|
* @pring: pointer to a SLI ring.
|
|
|
|
* @nvmebuf: pointer to received nvme data structure.
|
|
|
|
*
|
|
|
|
* This routine is used to process an unsolicited event received from a SLI
|
|
|
|
* (Service Level Interface) ring. The actual processing of the data buffer
|
|
|
|
* associated with the unsolicited event is done by invoking the routine
|
|
|
|
* lpfc_nvmet_unsol_ls_buffer() after properly set up the buffer from the
|
|
|
|
* SLI RQ on which the unsolicited event was received.
|
|
|
|
**/
|
|
|
|
void
|
|
|
|
lpfc_nvmet_unsol_ls_event(struct lpfc_hba *phba, struct lpfc_sli_ring *pring,
|
|
|
|
struct lpfc_iocbq *piocb)
|
|
|
|
{
|
|
|
|
struct lpfc_dmabuf *d_buf;
|
|
|
|
struct hbq_dmabuf *nvmebuf;
|
|
|
|
|
|
|
|
d_buf = piocb->context2;
|
|
|
|
nvmebuf = container_of(d_buf, struct hbq_dmabuf, dbuf);
|
|
|
|
|
|
|
|
if (phba->nvmet_support == 0) {
|
|
|
|
lpfc_in_buf_free(phba, &nvmebuf->dbuf);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
lpfc_nvmet_unsol_ls_buffer(phba, pring, nvmebuf);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_unsol_fcp_event - Process an unsolicited event from an nvme nport
|
|
|
|
* @phba: pointer to lpfc hba data structure.
|
|
|
|
* @pring: pointer to a SLI ring.
|
|
|
|
* @nvmebuf: pointer to received nvme data structure.
|
|
|
|
*
|
|
|
|
* This routine is used to process an unsolicited event received from a SLI
|
|
|
|
* (Service Level Interface) ring. The actual processing of the data buffer
|
|
|
|
* associated with the unsolicited event is done by invoking the routine
|
|
|
|
* lpfc_nvmet_unsol_fcp_buffer() after properly set up the buffer from the
|
|
|
|
* SLI RQ on which the unsolicited event was received.
|
|
|
|
**/
|
|
|
|
void
|
|
|
|
lpfc_nvmet_unsol_fcp_event(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_sli_ring *pring,
|
|
|
|
struct rqb_dmabuf *nvmebuf,
|
|
|
|
uint64_t isr_timestamp)
|
|
|
|
{
|
|
|
|
if (phba->nvmet_support == 0) {
|
|
|
|
lpfc_nvmet_rq_post(phba, NULL, &nvmebuf->hbuf);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
lpfc_nvmet_unsol_fcp_buffer(phba, pring, nvmebuf,
|
|
|
|
isr_timestamp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_prep_ls_wqe - Allocate and prepare a lpfc wqe data structure
|
|
|
|
* @phba: pointer to a host N_Port data structure.
|
|
|
|
* @ctxp: Context info for NVME LS Request
|
|
|
|
* @rspbuf: DMA buffer of NVME command.
|
|
|
|
* @rspsize: size of the NVME command.
|
|
|
|
*
|
|
|
|
* This routine is used for allocating a lpfc-WQE data structure from
|
|
|
|
* the driver lpfc-WQE free-list and prepare the WQE with the parameters
|
|
|
|
* passed into the routine for discovery state machine to issue an Extended
|
|
|
|
* Link Service (NVME) commands. It is a generic lpfc-WQE allocation
|
|
|
|
* and preparation routine that is used by all the discovery state machine
|
|
|
|
* routines and the NVME command-specific fields will be later set up by
|
|
|
|
* the individual discovery machine routines after calling this routine
|
|
|
|
* allocating and preparing a generic WQE data structure. It fills in the
|
|
|
|
* Buffer Descriptor Entries (BDEs), allocates buffers for both command
|
|
|
|
* payload and response payload (if expected). The reference count on the
|
|
|
|
* ndlp is incremented by 1 and the reference to the ndlp is put into
|
|
|
|
* context1 of the WQE data structure for this WQE to hold the ndlp
|
|
|
|
* reference for the command's callback function to access later.
|
|
|
|
*
|
|
|
|
* Return code
|
|
|
|
* Pointer to the newly allocated/prepared nvme wqe data structure
|
|
|
|
* NULL - when nvme wqe data structure allocation/preparation failed
|
|
|
|
**/
|
|
|
|
static struct lpfc_iocbq *
|
|
|
|
lpfc_nvmet_prep_ls_wqe(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp,
|
|
|
|
dma_addr_t rspbuf, uint16_t rspsize)
|
|
|
|
{
|
|
|
|
struct lpfc_nodelist *ndlp;
|
|
|
|
struct lpfc_iocbq *nvmewqe;
|
|
|
|
union lpfc_wqe *wqe;
|
|
|
|
|
|
|
|
if (!lpfc_is_link_up(phba)) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC,
|
|
|
|
"6104 lpfc_nvmet_prep_ls_wqe: link err: "
|
|
|
|
"NPORT x%x oxid:x%x\n",
|
|
|
|
ctxp->sid, ctxp->oxid);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Allocate buffer for command wqe */
|
|
|
|
nvmewqe = lpfc_sli_get_iocbq(phba);
|
|
|
|
if (nvmewqe == NULL) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC,
|
|
|
|
"6105 lpfc_nvmet_prep_ls_wqe: No WQE: "
|
|
|
|
"NPORT x%x oxid:x%x\n",
|
|
|
|
ctxp->sid, ctxp->oxid);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
ndlp = lpfc_findnode_did(phba->pport, ctxp->sid);
|
|
|
|
if (!ndlp || !NLP_CHK_NODE_ACT(ndlp) ||
|
|
|
|
((ndlp->nlp_state != NLP_STE_UNMAPPED_NODE) &&
|
|
|
|
(ndlp->nlp_state != NLP_STE_MAPPED_NODE))) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_DISC,
|
|
|
|
"6106 lpfc_nvmet_prep_ls_wqe: No ndlp: "
|
|
|
|
"NPORT x%x oxid:x%x\n",
|
|
|
|
ctxp->sid, ctxp->oxid);
|
|
|
|
goto nvme_wqe_free_wqeq_exit;
|
|
|
|
}
|
|
|
|
ctxp->wqeq = nvmewqe;
|
|
|
|
|
|
|
|
/* prevent preparing wqe with NULL ndlp reference */
|
|
|
|
nvmewqe->context1 = lpfc_nlp_get(ndlp);
|
|
|
|
if (nvmewqe->context1 == NULL)
|
|
|
|
goto nvme_wqe_free_wqeq_exit;
|
|
|
|
nvmewqe->context2 = ctxp;
|
|
|
|
|
|
|
|
wqe = &nvmewqe->wqe;
|
|
|
|
memset(wqe, 0, sizeof(union lpfc_wqe));
|
|
|
|
|
|
|
|
/* Words 0 - 2 */
|
|
|
|
wqe->xmit_sequence.bde.tus.f.bdeFlags = BUFF_TYPE_BDE_64;
|
|
|
|
wqe->xmit_sequence.bde.tus.f.bdeSize = rspsize;
|
|
|
|
wqe->xmit_sequence.bde.addrLow = le32_to_cpu(putPaddrLow(rspbuf));
|
|
|
|
wqe->xmit_sequence.bde.addrHigh = le32_to_cpu(putPaddrHigh(rspbuf));
|
|
|
|
|
|
|
|
/* Word 3 */
|
|
|
|
|
|
|
|
/* Word 4 */
|
|
|
|
|
|
|
|
/* Word 5 */
|
|
|
|
bf_set(wqe_dfctl, &wqe->xmit_sequence.wge_ctl, 0);
|
|
|
|
bf_set(wqe_ls, &wqe->xmit_sequence.wge_ctl, 1);
|
|
|
|
bf_set(wqe_la, &wqe->xmit_sequence.wge_ctl, 0);
|
2017-03-04 17:30:26 +00:00
|
|
|
bf_set(wqe_rctl, &wqe->xmit_sequence.wge_ctl, FC_RCTL_ELS4_REP);
|
2017-02-12 21:52:37 +00:00
|
|
|
bf_set(wqe_type, &wqe->xmit_sequence.wge_ctl, FC_TYPE_NVME);
|
|
|
|
|
|
|
|
/* Word 6 */
|
|
|
|
bf_set(wqe_ctxt_tag, &wqe->xmit_sequence.wqe_com,
|
|
|
|
phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]);
|
|
|
|
bf_set(wqe_xri_tag, &wqe->xmit_sequence.wqe_com, nvmewqe->sli4_xritag);
|
|
|
|
|
|
|
|
/* Word 7 */
|
|
|
|
bf_set(wqe_cmnd, &wqe->xmit_sequence.wqe_com,
|
|
|
|
CMD_XMIT_SEQUENCE64_WQE);
|
|
|
|
bf_set(wqe_ct, &wqe->xmit_sequence.wqe_com, SLI4_CT_RPI);
|
|
|
|
bf_set(wqe_class, &wqe->xmit_sequence.wqe_com, CLASS3);
|
|
|
|
bf_set(wqe_pu, &wqe->xmit_sequence.wqe_com, 0);
|
|
|
|
|
|
|
|
/* Word 8 */
|
|
|
|
wqe->xmit_sequence.wqe_com.abort_tag = nvmewqe->iotag;
|
|
|
|
|
|
|
|
/* Word 9 */
|
|
|
|
bf_set(wqe_reqtag, &wqe->xmit_sequence.wqe_com, nvmewqe->iotag);
|
|
|
|
/* Needs to be set by caller */
|
|
|
|
bf_set(wqe_rcvoxid, &wqe->xmit_sequence.wqe_com, ctxp->oxid);
|
|
|
|
|
|
|
|
/* Word 10 */
|
|
|
|
bf_set(wqe_dbde, &wqe->xmit_sequence.wqe_com, 1);
|
|
|
|
bf_set(wqe_iod, &wqe->xmit_sequence.wqe_com, LPFC_WQE_IOD_WRITE);
|
|
|
|
bf_set(wqe_lenloc, &wqe->xmit_sequence.wqe_com,
|
|
|
|
LPFC_WQE_LENLOC_WORD12);
|
|
|
|
bf_set(wqe_ebde_cnt, &wqe->xmit_sequence.wqe_com, 0);
|
|
|
|
|
|
|
|
/* Word 11 */
|
|
|
|
bf_set(wqe_cqid, &wqe->xmit_sequence.wqe_com,
|
|
|
|
LPFC_WQE_CQ_ID_DEFAULT);
|
|
|
|
bf_set(wqe_cmd_type, &wqe->xmit_sequence.wqe_com,
|
|
|
|
OTHER_COMMAND);
|
|
|
|
|
|
|
|
/* Word 12 */
|
|
|
|
wqe->xmit_sequence.xmit_len = rspsize;
|
|
|
|
|
|
|
|
nvmewqe->retry = 1;
|
|
|
|
nvmewqe->vport = phba->pport;
|
|
|
|
nvmewqe->drvrTimeout = (phba->fc_ratov * 3) + LPFC_DRVR_TIMEOUT;
|
|
|
|
nvmewqe->iocb_flag |= LPFC_IO_NVME_LS;
|
|
|
|
|
|
|
|
/* Xmit NVME response to remote NPORT <did> */
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC,
|
|
|
|
"6039 Xmit NVME LS response to remote "
|
|
|
|
"NPORT x%x iotag:x%x oxid:x%x size:x%x\n",
|
|
|
|
ndlp->nlp_DID, nvmewqe->iotag, ctxp->oxid,
|
|
|
|
rspsize);
|
|
|
|
return nvmewqe;
|
|
|
|
|
|
|
|
nvme_wqe_free_wqeq_exit:
|
|
|
|
nvmewqe->context2 = NULL;
|
|
|
|
nvmewqe->context3 = NULL;
|
|
|
|
lpfc_sli_release_iocbq(phba, nvmewqe);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static struct lpfc_iocbq *
|
|
|
|
lpfc_nvmet_prep_fcp_wqe(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp)
|
|
|
|
{
|
|
|
|
struct nvmefc_tgt_fcp_req *rsp = &ctxp->ctx.fcp_req;
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct sli4_sge *sgl;
|
|
|
|
struct lpfc_nodelist *ndlp;
|
|
|
|
struct lpfc_iocbq *nvmewqe;
|
|
|
|
struct scatterlist *sgel;
|
|
|
|
union lpfc_wqe128 *wqe;
|
|
|
|
uint32_t *txrdy;
|
|
|
|
dma_addr_t physaddr;
|
|
|
|
int i, cnt;
|
|
|
|
int xc = 1;
|
|
|
|
|
|
|
|
if (!lpfc_is_link_up(phba)) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6107 lpfc_nvmet_prep_fcp_wqe: link err:"
|
|
|
|
"NPORT x%x oxid:x%x\n", ctxp->sid,
|
|
|
|
ctxp->oxid);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
ndlp = lpfc_findnode_did(phba->pport, ctxp->sid);
|
|
|
|
if (!ndlp || !NLP_CHK_NODE_ACT(ndlp) ||
|
|
|
|
((ndlp->nlp_state != NLP_STE_UNMAPPED_NODE) &&
|
|
|
|
(ndlp->nlp_state != NLP_STE_MAPPED_NODE))) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6108 lpfc_nvmet_prep_fcp_wqe: no ndlp: "
|
|
|
|
"NPORT x%x oxid:x%x\n",
|
|
|
|
ctxp->sid, ctxp->oxid);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2017-04-21 23:05:01 +00:00
|
|
|
if (rsp->sg_cnt > phba->cfg_nvme_seg_cnt) {
|
2017-02-12 21:52:37 +00:00
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6109 lpfc_nvmet_prep_fcp_wqe: seg cnt err: "
|
2017-04-21 23:05:01 +00:00
|
|
|
"NPORT x%x oxid:x%x cnt %d\n",
|
|
|
|
ctxp->sid, ctxp->oxid, phba->cfg_nvme_seg_cnt);
|
2017-02-12 21:52:37 +00:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
nvmewqe = ctxp->wqeq;
|
|
|
|
if (nvmewqe == NULL) {
|
|
|
|
/* Allocate buffer for command wqe */
|
|
|
|
nvmewqe = ctxp->rqb_buffer->iocbq;
|
|
|
|
if (nvmewqe == NULL) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6110 lpfc_nvmet_prep_fcp_wqe: No "
|
|
|
|
"WQE: NPORT x%x oxid:x%x\n",
|
|
|
|
ctxp->sid, ctxp->oxid);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
ctxp->wqeq = nvmewqe;
|
|
|
|
xc = 0; /* create new XRI */
|
|
|
|
nvmewqe->sli4_lxritag = NO_XRI;
|
|
|
|
nvmewqe->sli4_xritag = NO_XRI;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Sanity check */
|
|
|
|
if (((ctxp->state == LPFC_NVMET_STE_RCV) &&
|
|
|
|
(ctxp->entry_cnt == 1)) ||
|
|
|
|
((ctxp->state == LPFC_NVMET_STE_DATA) &&
|
|
|
|
(ctxp->entry_cnt > 1))) {
|
|
|
|
wqe = (union lpfc_wqe128 *)&nvmewqe->wqe;
|
|
|
|
} else {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6111 Wrong state %s: %d cnt %d\n",
|
|
|
|
__func__, ctxp->state, ctxp->entry_cnt);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
sgl = (struct sli4_sge *)ctxp->rqb_buffer->sglq->sgl;
|
|
|
|
switch (rsp->op) {
|
|
|
|
case NVMET_FCOP_READDATA:
|
|
|
|
case NVMET_FCOP_READDATA_RSP:
|
|
|
|
/* Words 0 - 2 : The first sg segment */
|
|
|
|
sgel = &rsp->sg[0];
|
|
|
|
physaddr = sg_dma_address(sgel);
|
|
|
|
wqe->fcp_tsend.bde.tus.f.bdeFlags = BUFF_TYPE_BDE_64;
|
|
|
|
wqe->fcp_tsend.bde.tus.f.bdeSize = sg_dma_len(sgel);
|
|
|
|
wqe->fcp_tsend.bde.addrLow = cpu_to_le32(putPaddrLow(physaddr));
|
|
|
|
wqe->fcp_tsend.bde.addrHigh =
|
|
|
|
cpu_to_le32(putPaddrHigh(physaddr));
|
|
|
|
|
|
|
|
/* Word 3 */
|
|
|
|
wqe->fcp_tsend.payload_offset_len = 0;
|
|
|
|
|
|
|
|
/* Word 4 */
|
|
|
|
wqe->fcp_tsend.relative_offset = ctxp->offset;
|
|
|
|
|
|
|
|
/* Word 5 */
|
|
|
|
|
|
|
|
/* Word 6 */
|
|
|
|
bf_set(wqe_ctxt_tag, &wqe->fcp_tsend.wqe_com,
|
|
|
|
phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]);
|
|
|
|
bf_set(wqe_xri_tag, &wqe->fcp_tsend.wqe_com,
|
|
|
|
nvmewqe->sli4_xritag);
|
|
|
|
|
|
|
|
/* Word 7 */
|
|
|
|
bf_set(wqe_cmnd, &wqe->fcp_tsend.wqe_com, CMD_FCP_TSEND64_WQE);
|
|
|
|
|
|
|
|
/* Word 8 */
|
|
|
|
wqe->fcp_tsend.wqe_com.abort_tag = nvmewqe->iotag;
|
|
|
|
|
|
|
|
/* Word 9 */
|
|
|
|
bf_set(wqe_reqtag, &wqe->fcp_tsend.wqe_com, nvmewqe->iotag);
|
|
|
|
bf_set(wqe_rcvoxid, &wqe->fcp_tsend.wqe_com, ctxp->oxid);
|
|
|
|
|
|
|
|
/* Word 10 */
|
|
|
|
bf_set(wqe_nvme, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
bf_set(wqe_dbde, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
bf_set(wqe_iod, &wqe->fcp_tsend.wqe_com, LPFC_WQE_IOD_WRITE);
|
|
|
|
bf_set(wqe_lenloc, &wqe->fcp_tsend.wqe_com,
|
|
|
|
LPFC_WQE_LENLOC_WORD12);
|
|
|
|
bf_set(wqe_ebde_cnt, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_xc, &wqe->fcp_tsend.wqe_com, xc);
|
|
|
|
bf_set(wqe_nvme, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
if (phba->cfg_nvme_oas)
|
|
|
|
bf_set(wqe_oas, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
|
|
|
|
/* Word 11 */
|
|
|
|
bf_set(wqe_cqid, &wqe->fcp_tsend.wqe_com,
|
|
|
|
LPFC_WQE_CQ_ID_DEFAULT);
|
|
|
|
bf_set(wqe_cmd_type, &wqe->fcp_tsend.wqe_com,
|
|
|
|
FCP_COMMAND_TSEND);
|
|
|
|
|
|
|
|
/* Word 12 */
|
|
|
|
wqe->fcp_tsend.fcp_data_len = rsp->transfer_length;
|
|
|
|
|
|
|
|
/* Setup 2 SKIP SGEs */
|
|
|
|
sgl->addr_hi = 0;
|
|
|
|
sgl->addr_lo = 0;
|
|
|
|
sgl->word2 = 0;
|
|
|
|
bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
|
|
|
|
sgl->word2 = cpu_to_le32(sgl->word2);
|
|
|
|
sgl->sge_len = 0;
|
|
|
|
sgl++;
|
|
|
|
sgl->addr_hi = 0;
|
|
|
|
sgl->addr_lo = 0;
|
|
|
|
sgl->word2 = 0;
|
|
|
|
bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
|
|
|
|
sgl->word2 = cpu_to_le32(sgl->word2);
|
|
|
|
sgl->sge_len = 0;
|
|
|
|
sgl++;
|
|
|
|
if (rsp->op == NVMET_FCOP_READDATA_RSP) {
|
|
|
|
atomic_inc(&tgtp->xmt_fcp_read_rsp);
|
|
|
|
bf_set(wqe_ar, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
if ((ndlp->nlp_flag & NLP_SUPPRESS_RSP) &&
|
|
|
|
(rsp->rsplen == 12)) {
|
|
|
|
bf_set(wqe_sup, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
bf_set(wqe_wqes, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsp, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsplen, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
} else {
|
|
|
|
bf_set(wqe_sup, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_wqes, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
bf_set(wqe_irsp, &wqe->fcp_tsend.wqe_com, 1);
|
|
|
|
bf_set(wqe_irsplen, &wqe->fcp_tsend.wqe_com,
|
|
|
|
((rsp->rsplen >> 2) - 1));
|
|
|
|
memcpy(&wqe->words[16], rsp->rspaddr,
|
|
|
|
rsp->rsplen);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
atomic_inc(&tgtp->xmt_fcp_read);
|
|
|
|
|
|
|
|
bf_set(wqe_sup, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_wqes, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsp, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_ar, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsplen, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
}
|
|
|
|
ctxp->state = LPFC_NVMET_STE_DATA;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case NVMET_FCOP_WRITEDATA:
|
|
|
|
/* Words 0 - 2 : The first sg segment */
|
|
|
|
txrdy = pci_pool_alloc(phba->txrdy_payload_pool,
|
|
|
|
GFP_KERNEL, &physaddr);
|
|
|
|
if (!txrdy) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_IOERR,
|
|
|
|
"6041 Bad txrdy buffer: oxid x%x\n",
|
|
|
|
ctxp->oxid);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
ctxp->txrdy = txrdy;
|
|
|
|
ctxp->txrdy_phys = physaddr;
|
|
|
|
wqe->fcp_treceive.bde.tus.f.bdeFlags = BUFF_TYPE_BDE_64;
|
|
|
|
wqe->fcp_treceive.bde.tus.f.bdeSize = TXRDY_PAYLOAD_LEN;
|
|
|
|
wqe->fcp_treceive.bde.addrLow =
|
|
|
|
cpu_to_le32(putPaddrLow(physaddr));
|
|
|
|
wqe->fcp_treceive.bde.addrHigh =
|
|
|
|
cpu_to_le32(putPaddrHigh(physaddr));
|
|
|
|
|
|
|
|
/* Word 3 */
|
|
|
|
wqe->fcp_treceive.payload_offset_len = TXRDY_PAYLOAD_LEN;
|
|
|
|
|
|
|
|
/* Word 4 */
|
|
|
|
wqe->fcp_treceive.relative_offset = ctxp->offset;
|
|
|
|
|
|
|
|
/* Word 5 */
|
|
|
|
|
|
|
|
/* Word 6 */
|
|
|
|
bf_set(wqe_ctxt_tag, &wqe->fcp_treceive.wqe_com,
|
|
|
|
phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]);
|
|
|
|
bf_set(wqe_xri_tag, &wqe->fcp_treceive.wqe_com,
|
|
|
|
nvmewqe->sli4_xritag);
|
|
|
|
|
|
|
|
/* Word 7 */
|
|
|
|
bf_set(wqe_ar, &wqe->fcp_treceive.wqe_com, 0);
|
|
|
|
bf_set(wqe_cmnd, &wqe->fcp_treceive.wqe_com,
|
|
|
|
CMD_FCP_TRECEIVE64_WQE);
|
|
|
|
|
|
|
|
/* Word 8 */
|
|
|
|
wqe->fcp_treceive.wqe_com.abort_tag = nvmewqe->iotag;
|
|
|
|
|
|
|
|
/* Word 9 */
|
|
|
|
bf_set(wqe_reqtag, &wqe->fcp_treceive.wqe_com, nvmewqe->iotag);
|
|
|
|
bf_set(wqe_rcvoxid, &wqe->fcp_treceive.wqe_com, ctxp->oxid);
|
|
|
|
|
|
|
|
/* Word 10 */
|
|
|
|
bf_set(wqe_nvme, &wqe->fcp_treceive.wqe_com, 1);
|
|
|
|
bf_set(wqe_dbde, &wqe->fcp_treceive.wqe_com, 1);
|
|
|
|
bf_set(wqe_iod, &wqe->fcp_treceive.wqe_com, LPFC_WQE_IOD_READ);
|
|
|
|
bf_set(wqe_lenloc, &wqe->fcp_treceive.wqe_com,
|
|
|
|
LPFC_WQE_LENLOC_WORD12);
|
|
|
|
bf_set(wqe_xc, &wqe->fcp_treceive.wqe_com, xc);
|
|
|
|
bf_set(wqe_wqes, &wqe->fcp_treceive.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsp, &wqe->fcp_treceive.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsplen, &wqe->fcp_treceive.wqe_com, 0);
|
|
|
|
bf_set(wqe_nvme, &wqe->fcp_treceive.wqe_com, 1);
|
|
|
|
if (phba->cfg_nvme_oas)
|
|
|
|
bf_set(wqe_oas, &wqe->fcp_treceive.wqe_com, 1);
|
|
|
|
|
|
|
|
/* Word 11 */
|
|
|
|
bf_set(wqe_cqid, &wqe->fcp_treceive.wqe_com,
|
|
|
|
LPFC_WQE_CQ_ID_DEFAULT);
|
|
|
|
bf_set(wqe_cmd_type, &wqe->fcp_treceive.wqe_com,
|
|
|
|
FCP_COMMAND_TRECEIVE);
|
|
|
|
bf_set(wqe_sup, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
|
|
|
|
/* Word 12 */
|
|
|
|
wqe->fcp_tsend.fcp_data_len = rsp->transfer_length;
|
|
|
|
|
|
|
|
/* Setup 1 TXRDY and 1 SKIP SGE */
|
|
|
|
txrdy[0] = 0;
|
|
|
|
txrdy[1] = cpu_to_be32(rsp->transfer_length);
|
|
|
|
txrdy[2] = 0;
|
|
|
|
|
|
|
|
sgl->addr_hi = putPaddrHigh(physaddr);
|
|
|
|
sgl->addr_lo = putPaddrLow(physaddr);
|
|
|
|
sgl->word2 = 0;
|
|
|
|
bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA);
|
|
|
|
sgl->word2 = cpu_to_le32(sgl->word2);
|
|
|
|
sgl->sge_len = cpu_to_le32(TXRDY_PAYLOAD_LEN);
|
|
|
|
sgl++;
|
|
|
|
sgl->addr_hi = 0;
|
|
|
|
sgl->addr_lo = 0;
|
|
|
|
sgl->word2 = 0;
|
|
|
|
bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_SKIP);
|
|
|
|
sgl->word2 = cpu_to_le32(sgl->word2);
|
|
|
|
sgl->sge_len = 0;
|
|
|
|
sgl++;
|
|
|
|
ctxp->state = LPFC_NVMET_STE_DATA;
|
|
|
|
atomic_inc(&tgtp->xmt_fcp_write);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case NVMET_FCOP_RSP:
|
|
|
|
/* Words 0 - 2 */
|
|
|
|
physaddr = rsp->rspdma;
|
|
|
|
wqe->fcp_trsp.bde.tus.f.bdeFlags = BUFF_TYPE_BDE_64;
|
|
|
|
wqe->fcp_trsp.bde.tus.f.bdeSize = rsp->rsplen;
|
|
|
|
wqe->fcp_trsp.bde.addrLow =
|
|
|
|
cpu_to_le32(putPaddrLow(physaddr));
|
|
|
|
wqe->fcp_trsp.bde.addrHigh =
|
|
|
|
cpu_to_le32(putPaddrHigh(physaddr));
|
|
|
|
|
|
|
|
/* Word 3 */
|
|
|
|
wqe->fcp_trsp.response_len = rsp->rsplen;
|
|
|
|
|
|
|
|
/* Word 4 */
|
|
|
|
wqe->fcp_trsp.rsvd_4_5[0] = 0;
|
|
|
|
|
|
|
|
|
|
|
|
/* Word 5 */
|
|
|
|
|
|
|
|
/* Word 6 */
|
|
|
|
bf_set(wqe_ctxt_tag, &wqe->fcp_trsp.wqe_com,
|
|
|
|
phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]);
|
|
|
|
bf_set(wqe_xri_tag, &wqe->fcp_trsp.wqe_com,
|
|
|
|
nvmewqe->sli4_xritag);
|
|
|
|
|
|
|
|
/* Word 7 */
|
|
|
|
bf_set(wqe_ag, &wqe->fcp_trsp.wqe_com, 1);
|
|
|
|
bf_set(wqe_cmnd, &wqe->fcp_trsp.wqe_com, CMD_FCP_TRSP64_WQE);
|
|
|
|
|
|
|
|
/* Word 8 */
|
|
|
|
wqe->fcp_trsp.wqe_com.abort_tag = nvmewqe->iotag;
|
|
|
|
|
|
|
|
/* Word 9 */
|
|
|
|
bf_set(wqe_reqtag, &wqe->fcp_trsp.wqe_com, nvmewqe->iotag);
|
|
|
|
bf_set(wqe_rcvoxid, &wqe->fcp_trsp.wqe_com, ctxp->oxid);
|
|
|
|
|
|
|
|
/* Word 10 */
|
|
|
|
bf_set(wqe_nvme, &wqe->fcp_trsp.wqe_com, 1);
|
|
|
|
bf_set(wqe_dbde, &wqe->fcp_trsp.wqe_com, 0);
|
|
|
|
bf_set(wqe_iod, &wqe->fcp_trsp.wqe_com, LPFC_WQE_IOD_WRITE);
|
|
|
|
bf_set(wqe_lenloc, &wqe->fcp_trsp.wqe_com,
|
|
|
|
LPFC_WQE_LENLOC_WORD3);
|
|
|
|
bf_set(wqe_xc, &wqe->fcp_trsp.wqe_com, xc);
|
|
|
|
bf_set(wqe_nvme, &wqe->fcp_trsp.wqe_com, 1);
|
|
|
|
if (phba->cfg_nvme_oas)
|
|
|
|
bf_set(wqe_oas, &wqe->fcp_trsp.wqe_com, 1);
|
|
|
|
|
|
|
|
/* Word 11 */
|
|
|
|
bf_set(wqe_cqid, &wqe->fcp_trsp.wqe_com,
|
|
|
|
LPFC_WQE_CQ_ID_DEFAULT);
|
|
|
|
bf_set(wqe_cmd_type, &wqe->fcp_trsp.wqe_com,
|
|
|
|
FCP_COMMAND_TRSP);
|
|
|
|
bf_set(wqe_sup, &wqe->fcp_tsend.wqe_com, 0);
|
|
|
|
ctxp->state = LPFC_NVMET_STE_RSP;
|
|
|
|
|
|
|
|
if (rsp->rsplen == LPFC_NVMET_SUCCESS_LEN) {
|
|
|
|
/* Good response - all zero's on wire */
|
|
|
|
bf_set(wqe_wqes, &wqe->fcp_trsp.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsp, &wqe->fcp_trsp.wqe_com, 0);
|
|
|
|
bf_set(wqe_irsplen, &wqe->fcp_trsp.wqe_com, 0);
|
|
|
|
} else {
|
|
|
|
bf_set(wqe_wqes, &wqe->fcp_trsp.wqe_com, 1);
|
|
|
|
bf_set(wqe_irsp, &wqe->fcp_trsp.wqe_com, 1);
|
|
|
|
bf_set(wqe_irsplen, &wqe->fcp_trsp.wqe_com,
|
|
|
|
((rsp->rsplen >> 2) - 1));
|
|
|
|
memcpy(&wqe->words[16], rsp->rspaddr, rsp->rsplen);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Use rspbuf, NOT sg list */
|
|
|
|
rsp->sg_cnt = 0;
|
|
|
|
sgl->word2 = 0;
|
|
|
|
atomic_inc(&tgtp->xmt_fcp_rsp);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_IOERR,
|
|
|
|
"6064 Unknown Rsp Op %d\n",
|
|
|
|
rsp->op);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
nvmewqe->retry = 1;
|
|
|
|
nvmewqe->vport = phba->pport;
|
|
|
|
nvmewqe->drvrTimeout = (phba->fc_ratov * 3) + LPFC_DRVR_TIMEOUT;
|
|
|
|
nvmewqe->context1 = ndlp;
|
|
|
|
|
|
|
|
for (i = 0; i < rsp->sg_cnt; i++) {
|
|
|
|
sgel = &rsp->sg[i];
|
|
|
|
physaddr = sg_dma_address(sgel);
|
|
|
|
cnt = sg_dma_len(sgel);
|
|
|
|
sgl->addr_hi = putPaddrHigh(physaddr);
|
|
|
|
sgl->addr_lo = putPaddrLow(physaddr);
|
|
|
|
sgl->word2 = 0;
|
|
|
|
bf_set(lpfc_sli4_sge_type, sgl, LPFC_SGE_TYPE_DATA);
|
|
|
|
bf_set(lpfc_sli4_sge_offset, sgl, ctxp->offset);
|
|
|
|
if ((i+1) == rsp->sg_cnt)
|
|
|
|
bf_set(lpfc_sli4_sge_last, sgl, 1);
|
|
|
|
sgl->word2 = cpu_to_le32(sgl->word2);
|
|
|
|
sgl->sge_len = cpu_to_le32(cnt);
|
|
|
|
sgl++;
|
|
|
|
ctxp->offset += cnt;
|
|
|
|
}
|
|
|
|
return nvmewqe;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_sol_fcp_abort_cmp - Completion handler for ABTS
|
|
|
|
* @phba: Pointer to HBA context object.
|
|
|
|
* @cmdwqe: Pointer to driver command WQE object.
|
|
|
|
* @wcqe: Pointer to driver response CQE object.
|
|
|
|
*
|
|
|
|
* The function is called from SLI ring event handler with no
|
|
|
|
* lock held. This function is the completion handler for NVME ABTS for FCP cmds
|
|
|
|
* The function frees memory resources used for the NVME commands.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_sol_fcp_abort_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe,
|
|
|
|
struct lpfc_wcqe_complete *wcqe)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
uint32_t status, result;
|
2017-04-11 18:32:29 +00:00
|
|
|
unsigned long flags;
|
|
|
|
bool released = false;
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
ctxp = cmdwqe->context2;
|
|
|
|
status = bf_get(lpfc_wcqe_c_status, wcqe);
|
|
|
|
result = wcqe->parameter;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
atomic_inc(&tgtp->xmt_abort_cmpl);
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS,
|
|
|
|
"6165 Abort cmpl: xri x%x WCQE: %08x %08x %08x %08x\n",
|
|
|
|
ctxp->oxid, wcqe->word0, wcqe->total_data_placed,
|
|
|
|
result, wcqe->word3);
|
|
|
|
|
|
|
|
ctxp->state = LPFC_NVMET_STE_DONE;
|
2017-04-11 18:32:29 +00:00
|
|
|
spin_lock_irqsave(&ctxp->ctxlock, flags);
|
|
|
|
if (ctxp->flag & LPFC_NVMET_CTX_RLS)
|
|
|
|
released = true;
|
|
|
|
ctxp->flag &= ~LPFC_NVMET_ABORT_OP;
|
|
|
|
spin_unlock_irqrestore(&ctxp->ctxlock, flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* if transport has released ctx, then can reuse it. Otherwise,
|
|
|
|
* will be recycled by transport release call.
|
|
|
|
*/
|
|
|
|
if (released)
|
|
|
|
lpfc_nvmet_rq_post(phba, ctxp, &ctxp->rqb_buffer->hbuf);
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
cmdwqe->context2 = NULL;
|
|
|
|
cmdwqe->context3 = NULL;
|
|
|
|
lpfc_sli_release_iocbq(phba, cmdwqe);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_xmt_fcp_abort_cmp - Completion handler for ABTS
|
|
|
|
* @phba: Pointer to HBA context object.
|
|
|
|
* @cmdwqe: Pointer to driver command WQE object.
|
|
|
|
* @wcqe: Pointer to driver response CQE object.
|
|
|
|
*
|
|
|
|
* The function is called from SLI ring event handler with no
|
|
|
|
* lock held. This function is the completion handler for NVME ABTS for FCP cmds
|
|
|
|
* The function frees memory resources used for the NVME commands.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_xmt_fcp_abort_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe,
|
|
|
|
struct lpfc_wcqe_complete *wcqe)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
2017-04-11 18:32:29 +00:00
|
|
|
unsigned long flags;
|
2017-02-12 21:52:37 +00:00
|
|
|
uint32_t status, result;
|
2017-04-11 18:32:29 +00:00
|
|
|
bool released = false;
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
ctxp = cmdwqe->context2;
|
|
|
|
status = bf_get(lpfc_wcqe_c_status, wcqe);
|
|
|
|
result = wcqe->parameter;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
atomic_inc(&tgtp->xmt_abort_cmpl);
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS,
|
|
|
|
"6070 Abort cmpl: ctx %p WCQE: %08x %08x %08x %08x\n",
|
|
|
|
ctxp, wcqe->word0, wcqe->total_data_placed,
|
|
|
|
result, wcqe->word3);
|
|
|
|
|
|
|
|
if (ctxp) {
|
|
|
|
/* Sanity check */
|
|
|
|
if (ctxp->state != LPFC_NVMET_STE_ABORT) {
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS,
|
|
|
|
"6112 ABORT Wrong state:%d oxid x%x\n",
|
|
|
|
ctxp->state, ctxp->oxid);
|
|
|
|
}
|
|
|
|
ctxp->state = LPFC_NVMET_STE_DONE;
|
2017-04-11 18:32:29 +00:00
|
|
|
spin_lock_irqsave(&ctxp->ctxlock, flags);
|
|
|
|
if (ctxp->flag & LPFC_NVMET_CTX_RLS)
|
|
|
|
released = true;
|
|
|
|
ctxp->flag &= ~LPFC_NVMET_ABORT_OP;
|
|
|
|
spin_unlock_irqrestore(&ctxp->ctxlock, flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* if transport has released ctx, then can reuse it. Otherwise,
|
|
|
|
* will be recycled by transport release call.
|
|
|
|
*/
|
|
|
|
if (released)
|
|
|
|
lpfc_nvmet_rq_post(phba, ctxp, &ctxp->rqb_buffer->hbuf);
|
|
|
|
|
2017-02-12 21:52:37 +00:00
|
|
|
cmdwqe->context2 = NULL;
|
|
|
|
cmdwqe->context3 = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* lpfc_nvmet_xmt_ls_abort_cmp - Completion handler for ABTS
|
|
|
|
* @phba: Pointer to HBA context object.
|
|
|
|
* @cmdwqe: Pointer to driver command WQE object.
|
|
|
|
* @wcqe: Pointer to driver response CQE object.
|
|
|
|
*
|
|
|
|
* The function is called from SLI ring event handler with no
|
|
|
|
* lock held. This function is the completion handler for NVME ABTS for LS cmds
|
|
|
|
* The function frees memory resources used for the NVME commands.
|
|
|
|
**/
|
|
|
|
static void
|
|
|
|
lpfc_nvmet_xmt_ls_abort_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe,
|
|
|
|
struct lpfc_wcqe_complete *wcqe)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp;
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
uint32_t status, result;
|
|
|
|
|
|
|
|
ctxp = cmdwqe->context2;
|
|
|
|
status = bf_get(lpfc_wcqe_c_status, wcqe);
|
|
|
|
result = wcqe->parameter;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
atomic_inc(&tgtp->xmt_abort_cmpl);
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS,
|
|
|
|
"6083 Abort cmpl: ctx %p WCQE: %08x %08x %08x %08x\n",
|
|
|
|
ctxp, wcqe->word0, wcqe->total_data_placed,
|
|
|
|
result, wcqe->word3);
|
|
|
|
|
|
|
|
if (ctxp) {
|
|
|
|
cmdwqe->context2 = NULL;
|
|
|
|
cmdwqe->context3 = NULL;
|
|
|
|
lpfc_sli_release_iocbq(phba, cmdwqe);
|
|
|
|
kfree(ctxp);
|
|
|
|
} else
|
|
|
|
lpfc_sli_release_iocbq(phba, cmdwqe);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
lpfc_nvmet_unsol_issue_abort(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp,
|
|
|
|
uint32_t sid, uint16_t xri)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct lpfc_iocbq *abts_wqeq;
|
|
|
|
union lpfc_wqe *wqe_abts;
|
|
|
|
struct lpfc_nodelist *ndlp;
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS,
|
2017-03-04 17:30:30 +00:00
|
|
|
"6067 Abort: sid %x xri x%x/x%x\n",
|
|
|
|
sid, xri, ctxp->wqeq->sli4_xritag);
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
|
|
|
|
ndlp = lpfc_findnode_did(phba->pport, sid);
|
|
|
|
if (!ndlp || !NLP_CHK_NODE_ACT(ndlp) ||
|
|
|
|
((ndlp->nlp_state != NLP_STE_UNMAPPED_NODE) &&
|
|
|
|
(ndlp->nlp_state != NLP_STE_MAPPED_NODE))) {
|
|
|
|
atomic_inc(&tgtp->xmt_abort_rsp_error);
|
|
|
|
lpfc_printf_log(phba, KERN_WARNING, LOG_NVME_ABTS,
|
|
|
|
"6134 Drop ABTS - wrong NDLP state x%x.\n",
|
2017-03-04 17:30:39 +00:00
|
|
|
(ndlp) ? ndlp->nlp_state : NLP_STE_MAX_STATE);
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
/* No failure to an ABTS request. */
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
abts_wqeq = ctxp->wqeq;
|
|
|
|
wqe_abts = &abts_wqeq->wqe;
|
|
|
|
ctxp->state = LPFC_NVMET_STE_ABORT;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Since we zero the whole WQE, we need to ensure we set the WQE fields
|
|
|
|
* that were initialized in lpfc_sli4_nvmet_alloc.
|
|
|
|
*/
|
|
|
|
memset(wqe_abts, 0, sizeof(union lpfc_wqe));
|
|
|
|
|
|
|
|
/* Word 5 */
|
|
|
|
bf_set(wqe_dfctl, &wqe_abts->xmit_sequence.wge_ctl, 0);
|
|
|
|
bf_set(wqe_ls, &wqe_abts->xmit_sequence.wge_ctl, 1);
|
|
|
|
bf_set(wqe_la, &wqe_abts->xmit_sequence.wge_ctl, 0);
|
|
|
|
bf_set(wqe_rctl, &wqe_abts->xmit_sequence.wge_ctl, FC_RCTL_BA_ABTS);
|
|
|
|
bf_set(wqe_type, &wqe_abts->xmit_sequence.wge_ctl, FC_TYPE_BLS);
|
|
|
|
|
|
|
|
/* Word 6 */
|
|
|
|
bf_set(wqe_ctxt_tag, &wqe_abts->xmit_sequence.wqe_com,
|
|
|
|
phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]);
|
|
|
|
bf_set(wqe_xri_tag, &wqe_abts->xmit_sequence.wqe_com,
|
|
|
|
abts_wqeq->sli4_xritag);
|
|
|
|
|
|
|
|
/* Word 7 */
|
|
|
|
bf_set(wqe_cmnd, &wqe_abts->xmit_sequence.wqe_com,
|
|
|
|
CMD_XMIT_SEQUENCE64_WQE);
|
|
|
|
bf_set(wqe_ct, &wqe_abts->xmit_sequence.wqe_com, SLI4_CT_RPI);
|
|
|
|
bf_set(wqe_class, &wqe_abts->xmit_sequence.wqe_com, CLASS3);
|
|
|
|
bf_set(wqe_pu, &wqe_abts->xmit_sequence.wqe_com, 0);
|
|
|
|
|
|
|
|
/* Word 8 */
|
|
|
|
wqe_abts->xmit_sequence.wqe_com.abort_tag = abts_wqeq->iotag;
|
|
|
|
|
|
|
|
/* Word 9 */
|
|
|
|
bf_set(wqe_reqtag, &wqe_abts->xmit_sequence.wqe_com, abts_wqeq->iotag);
|
|
|
|
/* Needs to be set by caller */
|
|
|
|
bf_set(wqe_rcvoxid, &wqe_abts->xmit_sequence.wqe_com, xri);
|
|
|
|
|
|
|
|
/* Word 10 */
|
|
|
|
bf_set(wqe_dbde, &wqe_abts->xmit_sequence.wqe_com, 1);
|
|
|
|
bf_set(wqe_iod, &wqe_abts->xmit_sequence.wqe_com, LPFC_WQE_IOD_WRITE);
|
|
|
|
bf_set(wqe_lenloc, &wqe_abts->xmit_sequence.wqe_com,
|
|
|
|
LPFC_WQE_LENLOC_WORD12);
|
|
|
|
bf_set(wqe_ebde_cnt, &wqe_abts->xmit_sequence.wqe_com, 0);
|
|
|
|
bf_set(wqe_qosd, &wqe_abts->xmit_sequence.wqe_com, 0);
|
|
|
|
|
|
|
|
/* Word 11 */
|
|
|
|
bf_set(wqe_cqid, &wqe_abts->xmit_sequence.wqe_com,
|
|
|
|
LPFC_WQE_CQ_ID_DEFAULT);
|
|
|
|
bf_set(wqe_cmd_type, &wqe_abts->xmit_sequence.wqe_com,
|
|
|
|
OTHER_COMMAND);
|
|
|
|
|
|
|
|
abts_wqeq->vport = phba->pport;
|
|
|
|
abts_wqeq->context1 = ndlp;
|
|
|
|
abts_wqeq->context2 = ctxp;
|
|
|
|
abts_wqeq->context3 = NULL;
|
|
|
|
abts_wqeq->rsvd2 = 0;
|
|
|
|
/* hba_wqidx should already be setup from command we are aborting */
|
|
|
|
abts_wqeq->iocb.ulpCommand = CMD_XMIT_SEQUENCE64_CR;
|
|
|
|
abts_wqeq->iocb.ulpLe = 1;
|
|
|
|
|
|
|
|
lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS,
|
|
|
|
"6069 Issue ABTS to xri x%x reqtag x%x\n",
|
|
|
|
xri, abts_wqeq->iotag);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
lpfc_nvmet_sol_fcp_issue_abort(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp,
|
|
|
|
uint32_t sid, uint16_t xri)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct lpfc_iocbq *abts_wqeq;
|
|
|
|
union lpfc_wqe *abts_wqe;
|
|
|
|
struct lpfc_nodelist *ndlp;
|
|
|
|
unsigned long flags;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
if (!ctxp->wqeq) {
|
|
|
|
ctxp->wqeq = ctxp->rqb_buffer->iocbq;
|
|
|
|
ctxp->wqeq->hba_wqidx = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
ndlp = lpfc_findnode_did(phba->pport, sid);
|
|
|
|
if (!ndlp || !NLP_CHK_NODE_ACT(ndlp) ||
|
|
|
|
((ndlp->nlp_state != NLP_STE_UNMAPPED_NODE) &&
|
|
|
|
(ndlp->nlp_state != NLP_STE_MAPPED_NODE))) {
|
|
|
|
atomic_inc(&tgtp->xmt_abort_rsp_error);
|
|
|
|
lpfc_printf_log(phba, KERN_WARNING, LOG_NVME_ABTS,
|
|
|
|
"6160 Drop ABTS - wrong NDLP state x%x.\n",
|
2017-03-04 17:30:39 +00:00
|
|
|
(ndlp) ? ndlp->nlp_state : NLP_STE_MAX_STATE);
|
2017-02-12 21:52:37 +00:00
|
|
|
|
|
|
|
/* No failure to an ABTS request. */
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Issue ABTS for this WQE based on iotag */
|
|
|
|
ctxp->abort_wqeq = lpfc_sli_get_iocbq(phba);
|
|
|
|
if (!ctxp->abort_wqeq) {
|
|
|
|
lpfc_printf_log(phba, KERN_WARNING, LOG_NVME_ABTS,
|
|
|
|
"6161 Abort failed: No wqeqs: "
|
|
|
|
"xri: x%x\n", ctxp->oxid);
|
|
|
|
/* No failure to an ABTS request. */
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
abts_wqeq = ctxp->abort_wqeq;
|
|
|
|
abts_wqe = &abts_wqeq->wqe;
|
|
|
|
ctxp->state = LPFC_NVMET_STE_ABORT;
|
|
|
|
|
|
|
|
/* Announce entry to new IO submit field. */
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME_ABTS,
|
|
|
|
"6162 Abort Request to rport DID x%06x "
|
|
|
|
"for xri x%x x%x\n",
|
|
|
|
ctxp->sid, ctxp->oxid, ctxp->wqeq->sli4_xritag);
|
|
|
|
|
|
|
|
/* If the hba is getting reset, this flag is set. It is
|
|
|
|
* cleared when the reset is complete and rings reestablished.
|
|
|
|
*/
|
|
|
|
spin_lock_irqsave(&phba->hbalock, flags);
|
|
|
|
/* driver queued commands are in process of being flushed */
|
|
|
|
if (phba->hba_flag & HBA_NVME_IOQ_FLUSH) {
|
|
|
|
spin_unlock_irqrestore(&phba->hbalock, flags);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME,
|
|
|
|
"6163 Driver in reset cleanup - flushing "
|
|
|
|
"NVME Req now. hba_flag x%x oxid x%x\n",
|
|
|
|
phba->hba_flag, ctxp->oxid);
|
|
|
|
lpfc_sli_release_iocbq(phba, abts_wqeq);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Outstanding abort is in progress */
|
|
|
|
if (abts_wqeq->iocb_flag & LPFC_DRIVER_ABORTED) {
|
|
|
|
spin_unlock_irqrestore(&phba->hbalock, flags);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME,
|
|
|
|
"6164 Outstanding NVME I/O Abort Request "
|
|
|
|
"still pending on oxid x%x\n",
|
|
|
|
ctxp->oxid);
|
|
|
|
lpfc_sli_release_iocbq(phba, abts_wqeq);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Ready - mark outstanding as aborted by driver. */
|
|
|
|
abts_wqeq->iocb_flag |= LPFC_DRIVER_ABORTED;
|
|
|
|
|
|
|
|
/* WQEs are reused. Clear stale data and set key fields to
|
|
|
|
* zero like ia, iaab, iaar, xri_tag, and ctxt_tag.
|
|
|
|
*/
|
|
|
|
memset(abts_wqe, 0, sizeof(union lpfc_wqe));
|
|
|
|
|
|
|
|
/* word 3 */
|
|
|
|
bf_set(abort_cmd_criteria, &abts_wqe->abort_cmd, T_XRI_TAG);
|
|
|
|
|
|
|
|
/* word 7 */
|
|
|
|
bf_set(wqe_ct, &abts_wqe->abort_cmd.wqe_com, 0);
|
|
|
|
bf_set(wqe_cmnd, &abts_wqe->abort_cmd.wqe_com, CMD_ABORT_XRI_CX);
|
|
|
|
|
|
|
|
/* word 8 - tell the FW to abort the IO associated with this
|
|
|
|
* outstanding exchange ID.
|
|
|
|
*/
|
|
|
|
abts_wqe->abort_cmd.wqe_com.abort_tag = ctxp->wqeq->sli4_xritag;
|
|
|
|
|
|
|
|
/* word 9 - this is the iotag for the abts_wqe completion. */
|
|
|
|
bf_set(wqe_reqtag, &abts_wqe->abort_cmd.wqe_com,
|
|
|
|
abts_wqeq->iotag);
|
|
|
|
|
|
|
|
/* word 10 */
|
|
|
|
bf_set(wqe_qosd, &abts_wqe->abort_cmd.wqe_com, 1);
|
|
|
|
bf_set(wqe_lenloc, &abts_wqe->abort_cmd.wqe_com, LPFC_WQE_LENLOC_NONE);
|
|
|
|
|
|
|
|
/* word 11 */
|
|
|
|
bf_set(wqe_cmd_type, &abts_wqe->abort_cmd.wqe_com, OTHER_COMMAND);
|
|
|
|
bf_set(wqe_wqec, &abts_wqe->abort_cmd.wqe_com, 1);
|
|
|
|
bf_set(wqe_cqid, &abts_wqe->abort_cmd.wqe_com, LPFC_WQE_CQ_ID_DEFAULT);
|
|
|
|
|
|
|
|
/* ABTS WQE must go to the same WQ as the WQE to be aborted */
|
|
|
|
abts_wqeq->hba_wqidx = ctxp->wqeq->hba_wqidx;
|
|
|
|
abts_wqeq->wqe_cmpl = lpfc_nvmet_sol_fcp_abort_cmp;
|
|
|
|
abts_wqeq->iocb_cmpl = 0;
|
|
|
|
abts_wqeq->iocb_flag |= LPFC_IO_NVME;
|
|
|
|
abts_wqeq->context2 = ctxp;
|
|
|
|
rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, abts_wqeq);
|
|
|
|
spin_unlock_irqrestore(&phba->hbalock, flags);
|
|
|
|
if (rc == WQE_SUCCESS)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
lpfc_sli_release_iocbq(phba, abts_wqeq);
|
|
|
|
lpfc_printf_log(phba, KERN_ERR, LOG_NVME,
|
|
|
|
"6166 Failed abts issue_wqe with status x%x "
|
|
|
|
"for oxid x%x.\n",
|
|
|
|
rc, ctxp->oxid);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int
|
|
|
|
lpfc_nvmet_unsol_fcp_issue_abort(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp,
|
|
|
|
uint32_t sid, uint16_t xri)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct lpfc_iocbq *abts_wqeq;
|
|
|
|
unsigned long flags;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
if (!ctxp->wqeq) {
|
|
|
|
ctxp->wqeq = ctxp->rqb_buffer->iocbq;
|
|
|
|
ctxp->wqeq->hba_wqidx = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
rc = lpfc_nvmet_unsol_issue_abort(phba, ctxp, sid, xri);
|
|
|
|
if (rc == 0)
|
|
|
|
goto aerr;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&phba->hbalock, flags);
|
|
|
|
abts_wqeq = ctxp->wqeq;
|
|
|
|
abts_wqeq->wqe_cmpl = lpfc_nvmet_xmt_fcp_abort_cmp;
|
|
|
|
abts_wqeq->iocb_cmpl = 0;
|
|
|
|
abts_wqeq->iocb_flag |= LPFC_IO_NVMET;
|
|
|
|
rc = lpfc_sli4_issue_wqe(phba, LPFC_FCP_RING, abts_wqeq);
|
|
|
|
spin_unlock_irqrestore(&phba->hbalock, flags);
|
|
|
|
if (rc == WQE_SUCCESS) {
|
|
|
|
atomic_inc(&tgtp->xmt_abort_rsp);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
aerr:
|
|
|
|
lpfc_nvmet_rq_post(phba, ctxp, &ctxp->rqb_buffer->hbuf);
|
|
|
|
atomic_inc(&tgtp->xmt_abort_rsp_error);
|
|
|
|
lpfc_printf_log(phba, KERN_WARNING, LOG_NVME_ABTS,
|
|
|
|
"6135 Failed to Issue ABTS for oxid x%x. Status x%x\n",
|
|
|
|
ctxp->oxid, rc);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
lpfc_nvmet_unsol_ls_issue_abort(struct lpfc_hba *phba,
|
|
|
|
struct lpfc_nvmet_rcv_ctx *ctxp,
|
|
|
|
uint32_t sid, uint16_t xri)
|
|
|
|
{
|
|
|
|
struct lpfc_nvmet_tgtport *tgtp;
|
|
|
|
struct lpfc_iocbq *abts_wqeq;
|
|
|
|
union lpfc_wqe *wqe_abts;
|
|
|
|
unsigned long flags;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
tgtp = (struct lpfc_nvmet_tgtport *)phba->targetport->private;
|
|
|
|
if (!ctxp->wqeq) {
|
|
|
|
/* Issue ABTS for this WQE based on iotag */
|
|
|
|
ctxp->wqeq = lpfc_sli_get_iocbq(phba);
|
|
|
|
if (!ctxp->wqeq) {
|
|
|
|
lpfc_printf_log(phba, KERN_WARNING, LOG_NVME_ABTS,
|
|
|
|
"6068 Abort failed: No wqeqs: "
|
|
|
|
"xri: x%x\n", xri);
|
|
|
|
/* No failure to an ABTS request. */
|
|
|
|
kfree(ctxp);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
abts_wqeq = ctxp->wqeq;
|
|
|
|
wqe_abts = &abts_wqeq->wqe;
|
|
|
|
lpfc_nvmet_unsol_issue_abort(phba, ctxp, sid, xri);
|
|
|
|
|
|
|
|
spin_lock_irqsave(&phba->hbalock, flags);
|
|
|
|
abts_wqeq->wqe_cmpl = lpfc_nvmet_xmt_ls_abort_cmp;
|
|
|
|
abts_wqeq->iocb_cmpl = 0;
|
|
|
|
abts_wqeq->iocb_flag |= LPFC_IO_NVME_LS;
|
|
|
|
rc = lpfc_sli4_issue_wqe(phba, LPFC_ELS_RING, abts_wqeq);
|
|
|
|
spin_unlock_irqrestore(&phba->hbalock, flags);
|
|
|
|
if (rc == WQE_SUCCESS) {
|
|
|
|
atomic_inc(&tgtp->xmt_abort_rsp);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
atomic_inc(&tgtp->xmt_abort_rsp_error);
|
|
|
|
abts_wqeq->context2 = NULL;
|
|
|
|
abts_wqeq->context3 = NULL;
|
|
|
|
lpfc_sli_release_iocbq(phba, abts_wqeq);
|
|
|
|
kfree(ctxp);
|
|
|
|
lpfc_printf_log(phba, KERN_WARNING, LOG_NVME_ABTS,
|
|
|
|
"6056 Failed to Issue ABTS. Status x%x\n", rc);
|
|
|
|
return 0;
|
|
|
|
}
|