Enabled the PF->VF Mailbox support. Mailbox message are interpreted
as {type, opcode, data}. Supported message types are REQ, ACK and NACK.
Signed-off-by: Srikanth Jampala <Jampala.Srikanth@cavium.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
		
	
			
		
			
				
	
	
		
			454 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			454 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| #include <linux/pci.h>
 | |
| #include <linux/printk.h>
 | |
| #include <linux/slab.h>
 | |
| 
 | |
| #include "nitrox_dev.h"
 | |
| #include "nitrox_csr.h"
 | |
| #include "nitrox_common.h"
 | |
| #include "nitrox_hal.h"
 | |
| #include "nitrox_mbx.h"
 | |
| 
 | |
| /**
 | |
|  * One vector for each type of ring
 | |
|  *  - NPS packet ring, AQMQ ring and ZQMQ ring
 | |
|  */
 | |
| #define NR_RING_VECTORS 3
 | |
| #define NR_NON_RING_VECTORS 1
 | |
| /* base entry for packet ring/port */
 | |
| #define PKT_RING_MSIX_BASE 0
 | |
| #define NON_RING_MSIX_BASE 192
 | |
| 
 | |
| /**
 | |
|  * nps_pkt_slc_isr - IRQ handler for NPS solicit port
 | |
|  * @irq: irq number
 | |
|  * @data: argument
 | |
|  */
 | |
| static irqreturn_t nps_pkt_slc_isr(int irq, void *data)
 | |
| {
 | |
| 	struct nitrox_q_vector *qvec = data;
 | |
| 	union nps_pkt_slc_cnts slc_cnts;
 | |
| 	struct nitrox_cmdq *cmdq = qvec->cmdq;
 | |
| 
 | |
| 	slc_cnts.value = readq(cmdq->compl_cnt_csr_addr);
 | |
| 	/* New packet on SLC output port */
 | |
| 	if (slc_cnts.s.slc_int)
 | |
| 		tasklet_hi_schedule(&qvec->resp_tasklet);
 | |
| 
 | |
| 	return IRQ_HANDLED;
 | |
| }
 | |
| 
 | |
| static void clear_nps_core_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	u64 value;
 | |
| 
 | |
| 	/* Write 1 to clear */
 | |
| 	value = nitrox_read_csr(ndev, NPS_CORE_INT);
 | |
| 	nitrox_write_csr(ndev, NPS_CORE_INT, value);
 | |
| 
 | |
| 	dev_err_ratelimited(DEV(ndev), "NSP_CORE_INT  0x%016llx\n", value);
 | |
| }
 | |
| 
 | |
| static void clear_nps_pkt_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	union nps_pkt_int pkt_int;
 | |
| 	unsigned long value, offset;
 | |
| 	int i;
 | |
| 
 | |
| 	pkt_int.value = nitrox_read_csr(ndev, NPS_PKT_INT);
 | |
| 	dev_err_ratelimited(DEV(ndev), "NPS_PKT_INT  0x%016llx\n",
 | |
| 			    pkt_int.value);
 | |
| 
 | |
| 	if (pkt_int.s.slc_err) {
 | |
| 		offset = NPS_PKT_SLC_ERR_TYPE;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		dev_err_ratelimited(DEV(ndev),
 | |
| 				    "NPS_PKT_SLC_ERR_TYPE  0x%016lx\n", value);
 | |
| 
 | |
| 		offset = NPS_PKT_SLC_RERR_LO;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		/* enable the solicit ports */
 | |
| 		for_each_set_bit(i, &value, BITS_PER_LONG)
 | |
| 			enable_pkt_solicit_port(ndev, i);
 | |
| 
 | |
| 		dev_err_ratelimited(DEV(ndev),
 | |
| 				    "NPS_PKT_SLC_RERR_LO  0x%016lx\n", value);
 | |
| 
 | |
| 		offset = NPS_PKT_SLC_RERR_HI;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		dev_err_ratelimited(DEV(ndev),
 | |
| 				    "NPS_PKT_SLC_RERR_HI  0x%016lx\n", value);
 | |
| 	}
 | |
| 
 | |
| 	if (pkt_int.s.in_err) {
 | |
| 		offset = NPS_PKT_IN_ERR_TYPE;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		dev_err_ratelimited(DEV(ndev),
 | |
| 				    "NPS_PKT_IN_ERR_TYPE  0x%016lx\n", value);
 | |
| 		offset = NPS_PKT_IN_RERR_LO;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		/* enable the input ring */
 | |
| 		for_each_set_bit(i, &value, BITS_PER_LONG)
 | |
| 			enable_pkt_input_ring(ndev, i);
 | |
| 
 | |
| 		dev_err_ratelimited(DEV(ndev),
 | |
| 				    "NPS_PKT_IN_RERR_LO  0x%016lx\n", value);
 | |
| 
 | |
| 		offset = NPS_PKT_IN_RERR_HI;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		dev_err_ratelimited(DEV(ndev),
 | |
| 				    "NPS_PKT_IN_RERR_HI  0x%016lx\n", value);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void clear_pom_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	u64 value;
 | |
| 
 | |
| 	value = nitrox_read_csr(ndev, POM_INT);
 | |
| 	nitrox_write_csr(ndev, POM_INT, value);
 | |
| 	dev_err_ratelimited(DEV(ndev), "POM_INT  0x%016llx\n", value);
 | |
| }
 | |
| 
 | |
| static void clear_pem_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	u64 value;
 | |
| 
 | |
| 	value = nitrox_read_csr(ndev, PEM0_INT);
 | |
| 	nitrox_write_csr(ndev, PEM0_INT, value);
 | |
| 	dev_err_ratelimited(DEV(ndev), "PEM(0)_INT  0x%016llx\n", value);
 | |
| }
 | |
| 
 | |
| static void clear_lbc_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	union lbc_int lbc_int;
 | |
| 	u64 value, offset;
 | |
| 	int i;
 | |
| 
 | |
| 	lbc_int.value = nitrox_read_csr(ndev, LBC_INT);
 | |
| 	dev_err_ratelimited(DEV(ndev), "LBC_INT  0x%016llx\n", lbc_int.value);
 | |
| 
 | |
| 	if (lbc_int.s.dma_rd_err) {
 | |
| 		for (i = 0; i < NR_CLUSTERS; i++) {
 | |
| 			offset = EFL_CORE_VF_ERR_INT0X(i);
 | |
| 			value = nitrox_read_csr(ndev, offset);
 | |
| 			nitrox_write_csr(ndev, offset, value);
 | |
| 			offset = EFL_CORE_VF_ERR_INT1X(i);
 | |
| 			value = nitrox_read_csr(ndev, offset);
 | |
| 			nitrox_write_csr(ndev, offset, value);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (lbc_int.s.cam_soft_err) {
 | |
| 		dev_err_ratelimited(DEV(ndev), "CAM_SOFT_ERR, invalidating LBC\n");
 | |
| 		invalidate_lbc(ndev);
 | |
| 	}
 | |
| 
 | |
| 	if (lbc_int.s.pref_dat_len_mismatch_err) {
 | |
| 		offset = LBC_PLM_VF1_64_INT;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		offset = LBC_PLM_VF65_128_INT;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 	}
 | |
| 
 | |
| 	if (lbc_int.s.rd_dat_len_mismatch_err) {
 | |
| 		offset = LBC_ELM_VF1_64_INT;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 		offset = LBC_ELM_VF65_128_INT;
 | |
| 		value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, value);
 | |
| 	}
 | |
| 	nitrox_write_csr(ndev, LBC_INT, lbc_int.value);
 | |
| }
 | |
| 
 | |
| static void clear_efl_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < NR_CLUSTERS; i++) {
 | |
| 		union efl_core_int core_int;
 | |
| 		u64 value, offset;
 | |
| 
 | |
| 		offset = EFL_CORE_INTX(i);
 | |
| 		core_int.value = nitrox_read_csr(ndev, offset);
 | |
| 		nitrox_write_csr(ndev, offset, core_int.value);
 | |
| 		dev_err_ratelimited(DEV(ndev), "ELF_CORE(%d)_INT  0x%016llx\n",
 | |
| 				    i, core_int.value);
 | |
| 		if (core_int.s.se_err) {
 | |
| 			offset = EFL_CORE_SE_ERR_INTX(i);
 | |
| 			value = nitrox_read_csr(ndev, offset);
 | |
| 			nitrox_write_csr(ndev, offset, value);
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void clear_bmi_err_intr(struct nitrox_device *ndev)
 | |
| {
 | |
| 	u64 value;
 | |
| 
 | |
| 	value = nitrox_read_csr(ndev, BMI_INT);
 | |
| 	nitrox_write_csr(ndev, BMI_INT, value);
 | |
| 	dev_err_ratelimited(DEV(ndev), "BMI_INT  0x%016llx\n", value);
 | |
| }
 | |
| 
 | |
| static void nps_core_int_tasklet(unsigned long data)
 | |
| {
 | |
| 	struct nitrox_q_vector *qvec = (void *)(uintptr_t)(data);
 | |
| 	struct nitrox_device *ndev = qvec->ndev;
 | |
| 
 | |
| 	/* if pf mode do queue recovery */
 | |
| 	if (ndev->mode == __NDEV_MODE_PF) {
 | |
| 	} else {
 | |
| 		/**
 | |
| 		 * if VF(s) enabled communicate the error information
 | |
| 		 * to VF(s)
 | |
| 		 */
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * nps_core_int_isr - interrupt handler for NITROX errors and
 | |
|  *   mailbox communication
 | |
|  */
 | |
| static irqreturn_t nps_core_int_isr(int irq, void *data)
 | |
| {
 | |
| 	struct nitrox_q_vector *qvec = data;
 | |
| 	struct nitrox_device *ndev = qvec->ndev;
 | |
| 	union nps_core_int_active core_int;
 | |
| 
 | |
| 	core_int.value = nitrox_read_csr(ndev, NPS_CORE_INT_ACTIVE);
 | |
| 
 | |
| 	if (core_int.s.nps_core)
 | |
| 		clear_nps_core_err_intr(ndev);
 | |
| 
 | |
| 	if (core_int.s.nps_pkt)
 | |
| 		clear_nps_pkt_err_intr(ndev);
 | |
| 
 | |
| 	if (core_int.s.pom)
 | |
| 		clear_pom_err_intr(ndev);
 | |
| 
 | |
| 	if (core_int.s.pem)
 | |
| 		clear_pem_err_intr(ndev);
 | |
| 
 | |
| 	if (core_int.s.lbc)
 | |
| 		clear_lbc_err_intr(ndev);
 | |
| 
 | |
| 	if (core_int.s.efl)
 | |
| 		clear_efl_err_intr(ndev);
 | |
| 
 | |
| 	if (core_int.s.bmi)
 | |
| 		clear_bmi_err_intr(ndev);
 | |
| 
 | |
| 	/* Mailbox interrupt */
 | |
| 	if (core_int.s.mbox)
 | |
| 		nitrox_pf2vf_mbox_handler(ndev);
 | |
| 
 | |
| 	/* If more work callback the ISR, set resend */
 | |
| 	core_int.s.resend = 1;
 | |
| 	nitrox_write_csr(ndev, NPS_CORE_INT_ACTIVE, core_int.value);
 | |
| 
 | |
| 	return IRQ_HANDLED;
 | |
| }
 | |
| 
 | |
| void nitrox_unregister_interrupts(struct nitrox_device *ndev)
 | |
| {
 | |
| 	struct pci_dev *pdev = ndev->pdev;
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < ndev->num_vecs; i++) {
 | |
| 		struct nitrox_q_vector *qvec;
 | |
| 		int vec;
 | |
| 
 | |
| 		qvec = ndev->qvec + i;
 | |
| 		if (!qvec->valid)
 | |
| 			continue;
 | |
| 
 | |
| 		/* get the vector number */
 | |
| 		vec = pci_irq_vector(pdev, i);
 | |
| 		irq_set_affinity_hint(vec, NULL);
 | |
| 		free_irq(vec, qvec);
 | |
| 
 | |
| 		tasklet_disable(&qvec->resp_tasklet);
 | |
| 		tasklet_kill(&qvec->resp_tasklet);
 | |
| 		qvec->valid = false;
 | |
| 	}
 | |
| 	kfree(ndev->qvec);
 | |
| 	ndev->qvec = NULL;
 | |
| 	pci_free_irq_vectors(pdev);
 | |
| }
 | |
| 
 | |
| int nitrox_register_interrupts(struct nitrox_device *ndev)
 | |
| {
 | |
| 	struct pci_dev *pdev = ndev->pdev;
 | |
| 	struct nitrox_q_vector *qvec;
 | |
| 	int nr_vecs, vec, cpu;
 | |
| 	int ret, i;
 | |
| 
 | |
| 	/*
 | |
| 	 * PF MSI-X vectors
 | |
| 	 *
 | |
| 	 * Entry 0: NPS PKT ring 0
 | |
| 	 * Entry 1: AQMQ ring 0
 | |
| 	 * Entry 2: ZQM ring 0
 | |
| 	 * Entry 3: NPS PKT ring 1
 | |
| 	 * Entry 4: AQMQ ring 1
 | |
| 	 * Entry 5: ZQM ring 1
 | |
| 	 * ....
 | |
| 	 * Entry 192: NPS_CORE_INT_ACTIVE
 | |
| 	 */
 | |
| 	nr_vecs = pci_msix_vec_count(pdev);
 | |
| 
 | |
| 	/* Enable MSI-X */
 | |
| 	ret = pci_alloc_irq_vectors(pdev, nr_vecs, nr_vecs, PCI_IRQ_MSIX);
 | |
| 	if (ret < 0) {
 | |
| 		dev_err(DEV(ndev), "msix vectors %d alloc failed\n", nr_vecs);
 | |
| 		return ret;
 | |
| 	}
 | |
| 	ndev->num_vecs = nr_vecs;
 | |
| 
 | |
| 	ndev->qvec = kcalloc(nr_vecs, sizeof(*qvec), GFP_KERNEL);
 | |
| 	if (!ndev->qvec) {
 | |
| 		pci_free_irq_vectors(pdev);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	/* request irqs for packet rings/ports */
 | |
| 	for (i = PKT_RING_MSIX_BASE; i < (nr_vecs - 1); i += NR_RING_VECTORS) {
 | |
| 		qvec = &ndev->qvec[i];
 | |
| 
 | |
| 		qvec->ring = i / NR_RING_VECTORS;
 | |
| 		if (qvec->ring >= ndev->nr_queues)
 | |
| 			break;
 | |
| 
 | |
| 		qvec->cmdq = &ndev->pkt_inq[qvec->ring];
 | |
| 		snprintf(qvec->name, IRQ_NAMESZ, "nitrox-pkt%d", qvec->ring);
 | |
| 		/* get the vector number */
 | |
| 		vec = pci_irq_vector(pdev, i);
 | |
| 		ret = request_irq(vec, nps_pkt_slc_isr, 0, qvec->name, qvec);
 | |
| 		if (ret) {
 | |
| 			dev_err(DEV(ndev), "irq failed for pkt ring/port%d\n",
 | |
| 				qvec->ring);
 | |
| 			goto irq_fail;
 | |
| 		}
 | |
| 		cpu = qvec->ring % num_online_cpus();
 | |
| 		irq_set_affinity_hint(vec, get_cpu_mask(cpu));
 | |
| 
 | |
| 		tasklet_init(&qvec->resp_tasklet, pkt_slc_resp_tasklet,
 | |
| 			     (unsigned long)qvec);
 | |
| 		qvec->valid = true;
 | |
| 	}
 | |
| 
 | |
| 	/* request irqs for non ring vectors */
 | |
| 	i = NON_RING_MSIX_BASE;
 | |
| 	qvec = &ndev->qvec[i];
 | |
| 	qvec->ndev = ndev;
 | |
| 
 | |
| 	snprintf(qvec->name, IRQ_NAMESZ, "nitrox-core-int%d", i);
 | |
| 	/* get the vector number */
 | |
| 	vec = pci_irq_vector(pdev, i);
 | |
| 	ret = request_irq(vec, nps_core_int_isr, 0, qvec->name, qvec);
 | |
| 	if (ret) {
 | |
| 		dev_err(DEV(ndev), "irq failed for nitrox-core-int%d\n", i);
 | |
| 		goto irq_fail;
 | |
| 	}
 | |
| 	cpu = num_online_cpus();
 | |
| 	irq_set_affinity_hint(vec, get_cpu_mask(cpu));
 | |
| 
 | |
| 	tasklet_init(&qvec->resp_tasklet, nps_core_int_tasklet,
 | |
| 		     (unsigned long)qvec);
 | |
| 	qvec->valid = true;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| irq_fail:
 | |
| 	nitrox_unregister_interrupts(ndev);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void nitrox_sriov_unregister_interrupts(struct nitrox_device *ndev)
 | |
| {
 | |
| 	struct pci_dev *pdev = ndev->pdev;
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < ndev->num_vecs; i++) {
 | |
| 		struct nitrox_q_vector *qvec;
 | |
| 		int vec;
 | |
| 
 | |
| 		qvec = ndev->qvec + i;
 | |
| 		if (!qvec->valid)
 | |
| 			continue;
 | |
| 
 | |
| 		vec = ndev->iov.msix.vector;
 | |
| 		irq_set_affinity_hint(vec, NULL);
 | |
| 		free_irq(vec, qvec);
 | |
| 
 | |
| 		tasklet_disable(&qvec->resp_tasklet);
 | |
| 		tasklet_kill(&qvec->resp_tasklet);
 | |
| 		qvec->valid = false;
 | |
| 	}
 | |
| 	kfree(ndev->qvec);
 | |
| 	ndev->qvec = NULL;
 | |
| 	pci_disable_msix(pdev);
 | |
| }
 | |
| 
 | |
| int nitrox_sriov_register_interupts(struct nitrox_device *ndev)
 | |
| {
 | |
| 	struct pci_dev *pdev = ndev->pdev;
 | |
| 	struct nitrox_q_vector *qvec;
 | |
| 	int vec, cpu;
 | |
| 	int ret;
 | |
| 
 | |
| 	/**
 | |
| 	 * only non ring vectors i.e Entry 192 is available
 | |
| 	 * for PF in SR-IOV mode.
 | |
| 	 */
 | |
| 	ndev->iov.msix.entry = NON_RING_MSIX_BASE;
 | |
| 	ret = pci_enable_msix_exact(pdev, &ndev->iov.msix, NR_NON_RING_VECTORS);
 | |
| 	if (ret) {
 | |
| 		dev_err(DEV(ndev), "failed to allocate nps-core-int%d\n",
 | |
| 			NON_RING_MSIX_BASE);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	qvec = kcalloc(NR_NON_RING_VECTORS, sizeof(*qvec), GFP_KERNEL);
 | |
| 	if (!qvec) {
 | |
| 		pci_disable_msix(pdev);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 	qvec->ndev = ndev;
 | |
| 
 | |
| 	ndev->qvec = qvec;
 | |
| 	ndev->num_vecs = NR_NON_RING_VECTORS;
 | |
| 	snprintf(qvec->name, IRQ_NAMESZ, "nitrox-core-int%d",
 | |
| 		 NON_RING_MSIX_BASE);
 | |
| 
 | |
| 	vec = ndev->iov.msix.vector;
 | |
| 	ret = request_irq(vec, nps_core_int_isr, 0, qvec->name, qvec);
 | |
| 	if (ret) {
 | |
| 		dev_err(DEV(ndev), "irq failed for nitrox-core-int%d\n",
 | |
| 			NON_RING_MSIX_BASE);
 | |
| 		goto iov_irq_fail;
 | |
| 	}
 | |
| 	cpu = num_online_cpus();
 | |
| 	irq_set_affinity_hint(vec, get_cpu_mask(cpu));
 | |
| 
 | |
| 	tasklet_init(&qvec->resp_tasklet, nps_core_int_tasklet,
 | |
| 		     (unsigned long)qvec);
 | |
| 	qvec->valid = true;
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| iov_irq_fail:
 | |
| 	nitrox_sriov_unregister_interrupts(ndev);
 | |
| 	return ret;
 | |
| }
 |