forked from Minki/linux
19fd08b85b
- Fix RDMA uapi headers to actually compile in userspace and be more complete - Three shared with netdev pull requests from Mellanox: * 7 patches, mostly to net with 1 IB related one at the back). This series addresses an IRQ performance issue (patch 1), cleanups related to the fix for the IRQ performance problem (patches 2-6), and then extends the fragmented completion queue support that already exists in the net side of the driver to the ib side of the driver (patch 7). * Mostly IB, with 5 patches to net that are needed to support the remaining 10 patches to the IB subsystem. This series extends the current 'representor' framework when the mlx5 driver is in switchdev mode from being a netdev only construct to being a netdev/IB dev construct. The IB dev is limited to raw Eth queue pairs only, but by having an IB dev of this type attached to the representor for a switchdev port, it enables DPDK to work on the switchdev device. * All net related, but needed as infrastructure for the rdma driver - Updates for the hns, i40iw, bnxt_re, cxgb3, cxgb4, hns drivers - SRP performance updates - IB uverbs write path cleanup patch series from Leon - Add RDMA_CM support to ib_srpt. This is disabled by default. Users need to set the port for ib_srpt to listen on in configfs in order for it to be enabled (/sys/kernel/config/target/srpt/discovery_auth/rdma_cm_port) - TSO and Scatter FCS support in mlx4 - Refactor of modify_qp routine to resolve problems seen while working on new code that is forthcoming - More refactoring and updates of RDMA CM for containers support from Parav - mlx5 'fine grained packet pacing', 'ipsec offload' and 'device memory' user API features - Infrastructure updates for the new IOCTL interface, based on increased usage - ABI compatibility bug fixes to fully support 32 bit userspace on 64 bit kernel as was originally intended. See the commit messages for extensive details - Syzkaller bugs and code cleanups motivated by them -----BEGIN PGP SIGNATURE----- Version: GnuPG v2 iQIcBAABCgAGBQJax5Z0AAoJEDht9xV+IJsacCwQAJBIgmLCvVp5fBu2kJcXMMVI y3l2YNzAUJvDDKv1r5yTC9ugBXEkDtgzi/W/C2/5es2yUG/QeT/zzQ3YPrtsnN68 5FkiXQ35Tt7+PBHMr0cacGRmF4M3Td3MeW0X5aJaBKhqlNKwA+aF18pjGWBmpVYx URYCwLb5BZBKVh4+1Leebsk4i0/7jSauAqE5M+9notuAUfBCoY1/Eve3DipEIBBp EyrEnMDIdujYRsg4KHlxFKKJ1EFGItknLQbNL1+SEa0Oe0SnEl5Bd53Yxfz7ekNP oOWQe5csTcs3Yr4Ob0TC+69CzI71zKbz6qPDILTwXmsPFZJ9ipJs4S8D6F7ra8tb D5aT1EdRzh/vAORPC9T3DQ3VsHdvhwpUMG7knnKrVT9X/g7E+gSji1BqaQaTr/xs i40GepHT7lM/TWEuee/6LRpqdhuOhud7vfaRFwn2JGRX9suqTcvwhkBkPUDGV5XX 5RkHcWOb/7KvmpG7S1gaRGK5kO208LgmAZi7REaJFoZB74FqSneMR6NHIH07ha41 Zou7rnxV68CT2bgu27m+72EsprgmBkVDeEzXgKxVI/+PZ1oadUFpgcZ3pRLOPWVx rEqjHu65rlA/YPog4iXQaMfSwt/oRD3cVJS/n8EdJKXi4Qt2RDDGdyOmt74w4prM QuLEdvJIFmwrND1KDoqn =Ku8g -----END PGP SIGNATURE----- Merge tag 'for-linus-unmerged' of git://git.kernel.org/pub/scm/linux/kernel/git/rdma/rdma Pull rdma updates from Jason Gunthorpe: "Doug and I are at a conference next week so if another PR is sent I expect it to only be bug fixes. Parav noted yesterday that there are some fringe case behavior changes in his work that he would like to fix, and I see that Intel has a number of rc looking patches for HFI1 they posted yesterday. Parav is again the biggest contributor by patch count with his ongoing work to enable container support in the RDMA stack, followed by Leon doing syzkaller inspired cleanups, though most of the actual fixing went to RC. There is one uncomfortable series here fixing the user ABI to actually work as intended in 32 bit mode. There are lots of notes in the commit messages, but the basic summary is we don't think there is an actual 32 bit kernel user of drivers/infiniband for several good reasons. However we are seeing people want to use a 32 bit user space with 64 bit kernel, which didn't completely work today. So in fixing it we required a 32 bit rxe user to upgrade their userspace. rxe users are still already quite rare and we think a 32 bit one is non-existing. - Fix RDMA uapi headers to actually compile in userspace and be more complete - Three shared with netdev pull requests from Mellanox: * 7 patches, mostly to net with 1 IB related one at the back). This series addresses an IRQ performance issue (patch 1), cleanups related to the fix for the IRQ performance problem (patches 2-6), and then extends the fragmented completion queue support that already exists in the net side of the driver to the ib side of the driver (patch 7). * Mostly IB, with 5 patches to net that are needed to support the remaining 10 patches to the IB subsystem. This series extends the current 'representor' framework when the mlx5 driver is in switchdev mode from being a netdev only construct to being a netdev/IB dev construct. The IB dev is limited to raw Eth queue pairs only, but by having an IB dev of this type attached to the representor for a switchdev port, it enables DPDK to work on the switchdev device. * All net related, but needed as infrastructure for the rdma driver - Updates for the hns, i40iw, bnxt_re, cxgb3, cxgb4, hns drivers - SRP performance updates - IB uverbs write path cleanup patch series from Leon - Add RDMA_CM support to ib_srpt. This is disabled by default. Users need to set the port for ib_srpt to listen on in configfs in order for it to be enabled (/sys/kernel/config/target/srpt/discovery_auth/rdma_cm_port) - TSO and Scatter FCS support in mlx4 - Refactor of modify_qp routine to resolve problems seen while working on new code that is forthcoming - More refactoring and updates of RDMA CM for containers support from Parav - mlx5 'fine grained packet pacing', 'ipsec offload' and 'device memory' user API features - Infrastructure updates for the new IOCTL interface, based on increased usage - ABI compatibility bug fixes to fully support 32 bit userspace on 64 bit kernel as was originally intended. See the commit messages for extensive details - Syzkaller bugs and code cleanups motivated by them" * tag 'for-linus-unmerged' of git://git.kernel.org/pub/scm/linux/kernel/git/rdma/rdma: (199 commits) IB/rxe: Fix for oops in rxe_register_device on ppc64le arch IB/mlx5: Device memory mr registration support net/mlx5: Mkey creation command adjustments IB/mlx5: Device memory support in mlx5_ib net/mlx5: Query device memory capabilities IB/uverbs: Add device memory registration ioctl support IB/uverbs: Add alloc/free dm uverbs ioctl support IB/uverbs: Add device memory capabilities reporting IB/uverbs: Expose device memory capabilities to user RDMA/qedr: Fix wmb usage in qedr IB/rxe: Removed GID add/del dummy routines RDMA/qedr: Zero stack memory before copying to user space IB/mlx5: Add ability to hash by IPSEC_SPI when creating a TIR IB/mlx5: Add information for querying IPsec capabilities IB/mlx5: Add IPsec support for egress and ingress {net,IB}/mlx5: Add ipsec helper IB/mlx5: Add modify_flow_action_esp verb IB/mlx5: Add implementation for create and destroy action_xfrm IB/uverbs: Introduce ESP steering match filter IB/uverbs: Add modify ESP flow_action ...
716 lines
20 KiB
C
716 lines
20 KiB
C
/*
|
|
* Copyright (c) 2013, Cisco Systems, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*
|
|
* Author: Upinder Malhi <umalhi@cisco.com>
|
|
* Author: Anant Deepak <anadeepa@cisco.com>
|
|
* Author: Cesare Cantu' <cantuc@cisco.com>
|
|
* Author: Jeff Squyres <jsquyres@cisco.com>
|
|
* Author: Kiran Thirumalai <kithirum@cisco.com>
|
|
* Author: Xuyang Wang <xuywang@cisco.com>
|
|
* Author: Reese Faucette <rfaucett@cisco.com>
|
|
*
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/inetdevice.h>
|
|
#include <linux/init.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/netdevice.h>
|
|
|
|
#include <rdma/ib_user_verbs.h>
|
|
#include <rdma/ib_addr.h>
|
|
|
|
#include "usnic_abi.h"
|
|
#include "usnic_common_util.h"
|
|
#include "usnic_ib.h"
|
|
#include "usnic_ib_qp_grp.h"
|
|
#include "usnic_log.h"
|
|
#include "usnic_fwd.h"
|
|
#include "usnic_debugfs.h"
|
|
#include "usnic_ib_verbs.h"
|
|
#include "usnic_transport.h"
|
|
#include "usnic_uiom.h"
|
|
#include "usnic_ib_sysfs.h"
|
|
|
|
unsigned int usnic_log_lvl = USNIC_LOG_LVL_ERR;
|
|
unsigned int usnic_ib_share_vf = 1;
|
|
|
|
static const char usnic_version[] =
|
|
DRV_NAME ": Cisco VIC (USNIC) Verbs Driver v"
|
|
DRV_VERSION " (" DRV_RELDATE ")\n";
|
|
|
|
static DEFINE_MUTEX(usnic_ib_ibdev_list_lock);
|
|
static LIST_HEAD(usnic_ib_ibdev_list);
|
|
|
|
/* Callback dump funcs */
|
|
static int usnic_ib_dump_vf_hdr(void *obj, char *buf, int buf_sz)
|
|
{
|
|
struct usnic_ib_vf *vf = obj;
|
|
return scnprintf(buf, buf_sz, "PF: %s ", vf->pf->ib_dev.name);
|
|
}
|
|
/* End callback dump funcs */
|
|
|
|
static void usnic_ib_dump_vf(struct usnic_ib_vf *vf, char *buf, int buf_sz)
|
|
{
|
|
usnic_vnic_dump(vf->vnic, buf, buf_sz, vf,
|
|
usnic_ib_dump_vf_hdr,
|
|
usnic_ib_qp_grp_dump_hdr, usnic_ib_qp_grp_dump_rows);
|
|
}
|
|
|
|
void usnic_ib_log_vf(struct usnic_ib_vf *vf)
|
|
{
|
|
char buf[1000];
|
|
usnic_ib_dump_vf(vf, buf, sizeof(buf));
|
|
usnic_dbg("%s\n", buf);
|
|
}
|
|
|
|
/* Start of netdev section */
|
|
static void usnic_ib_qp_grp_modify_active_to_err(struct usnic_ib_dev *us_ibdev)
|
|
{
|
|
struct usnic_ib_ucontext *ctx;
|
|
struct usnic_ib_qp_grp *qp_grp;
|
|
enum ib_qp_state cur_state;
|
|
int status;
|
|
|
|
BUG_ON(!mutex_is_locked(&us_ibdev->usdev_lock));
|
|
|
|
list_for_each_entry(ctx, &us_ibdev->ctx_list, link) {
|
|
list_for_each_entry(qp_grp, &ctx->qp_grp_list, link) {
|
|
cur_state = qp_grp->state;
|
|
if (cur_state == IB_QPS_INIT ||
|
|
cur_state == IB_QPS_RTR ||
|
|
cur_state == IB_QPS_RTS) {
|
|
status = usnic_ib_qp_grp_modify(qp_grp,
|
|
IB_QPS_ERR,
|
|
NULL);
|
|
if (status) {
|
|
usnic_err("Failed to transistion qp grp %u from %s to %s\n",
|
|
qp_grp->grp_id,
|
|
usnic_ib_qp_grp_state_to_string
|
|
(cur_state),
|
|
usnic_ib_qp_grp_state_to_string
|
|
(IB_QPS_ERR));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
static void usnic_ib_handle_usdev_event(struct usnic_ib_dev *us_ibdev,
|
|
unsigned long event)
|
|
{
|
|
struct net_device *netdev;
|
|
struct ib_event ib_event;
|
|
|
|
memset(&ib_event, 0, sizeof(ib_event));
|
|
|
|
mutex_lock(&us_ibdev->usdev_lock);
|
|
netdev = us_ibdev->netdev;
|
|
switch (event) {
|
|
case NETDEV_REBOOT:
|
|
usnic_info("PF Reset on %s\n", us_ibdev->ib_dev.name);
|
|
usnic_ib_qp_grp_modify_active_to_err(us_ibdev);
|
|
ib_event.event = IB_EVENT_PORT_ERR;
|
|
ib_event.device = &us_ibdev->ib_dev;
|
|
ib_event.element.port_num = 1;
|
|
ib_dispatch_event(&ib_event);
|
|
break;
|
|
case NETDEV_UP:
|
|
case NETDEV_DOWN:
|
|
case NETDEV_CHANGE:
|
|
if (!us_ibdev->ufdev->link_up &&
|
|
netif_carrier_ok(netdev)) {
|
|
usnic_fwd_carrier_up(us_ibdev->ufdev);
|
|
usnic_info("Link UP on %s\n", us_ibdev->ib_dev.name);
|
|
ib_event.event = IB_EVENT_PORT_ACTIVE;
|
|
ib_event.device = &us_ibdev->ib_dev;
|
|
ib_event.element.port_num = 1;
|
|
ib_dispatch_event(&ib_event);
|
|
} else if (us_ibdev->ufdev->link_up &&
|
|
!netif_carrier_ok(netdev)) {
|
|
usnic_fwd_carrier_down(us_ibdev->ufdev);
|
|
usnic_info("Link DOWN on %s\n", us_ibdev->ib_dev.name);
|
|
usnic_ib_qp_grp_modify_active_to_err(us_ibdev);
|
|
ib_event.event = IB_EVENT_PORT_ERR;
|
|
ib_event.device = &us_ibdev->ib_dev;
|
|
ib_event.element.port_num = 1;
|
|
ib_dispatch_event(&ib_event);
|
|
} else {
|
|
usnic_dbg("Ignoring %s on %s\n",
|
|
netdev_cmd_to_name(event),
|
|
us_ibdev->ib_dev.name);
|
|
}
|
|
break;
|
|
case NETDEV_CHANGEADDR:
|
|
if (!memcmp(us_ibdev->ufdev->mac, netdev->dev_addr,
|
|
sizeof(us_ibdev->ufdev->mac))) {
|
|
usnic_dbg("Ignoring addr change on %s\n",
|
|
us_ibdev->ib_dev.name);
|
|
} else {
|
|
usnic_info(" %s old mac: %pM new mac: %pM\n",
|
|
us_ibdev->ib_dev.name,
|
|
us_ibdev->ufdev->mac,
|
|
netdev->dev_addr);
|
|
usnic_fwd_set_mac(us_ibdev->ufdev, netdev->dev_addr);
|
|
usnic_ib_qp_grp_modify_active_to_err(us_ibdev);
|
|
ib_event.event = IB_EVENT_GID_CHANGE;
|
|
ib_event.device = &us_ibdev->ib_dev;
|
|
ib_event.element.port_num = 1;
|
|
ib_dispatch_event(&ib_event);
|
|
}
|
|
|
|
break;
|
|
case NETDEV_CHANGEMTU:
|
|
if (us_ibdev->ufdev->mtu != netdev->mtu) {
|
|
usnic_info("MTU Change on %s old: %u new: %u\n",
|
|
us_ibdev->ib_dev.name,
|
|
us_ibdev->ufdev->mtu, netdev->mtu);
|
|
usnic_fwd_set_mtu(us_ibdev->ufdev, netdev->mtu);
|
|
usnic_ib_qp_grp_modify_active_to_err(us_ibdev);
|
|
} else {
|
|
usnic_dbg("Ignoring MTU change on %s\n",
|
|
us_ibdev->ib_dev.name);
|
|
}
|
|
break;
|
|
default:
|
|
usnic_dbg("Ignoring event %s on %s",
|
|
netdev_cmd_to_name(event),
|
|
us_ibdev->ib_dev.name);
|
|
}
|
|
mutex_unlock(&us_ibdev->usdev_lock);
|
|
}
|
|
|
|
static int usnic_ib_netdevice_event(struct notifier_block *notifier,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct usnic_ib_dev *us_ibdev;
|
|
|
|
struct net_device *netdev = netdev_notifier_info_to_dev(ptr);
|
|
|
|
mutex_lock(&usnic_ib_ibdev_list_lock);
|
|
list_for_each_entry(us_ibdev, &usnic_ib_ibdev_list, ib_dev_link) {
|
|
if (us_ibdev->netdev == netdev) {
|
|
usnic_ib_handle_usdev_event(us_ibdev, event);
|
|
break;
|
|
}
|
|
}
|
|
mutex_unlock(&usnic_ib_ibdev_list_lock);
|
|
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
static struct notifier_block usnic_ib_netdevice_notifier = {
|
|
.notifier_call = usnic_ib_netdevice_event
|
|
};
|
|
/* End of netdev section */
|
|
|
|
/* Start of inet section */
|
|
static int usnic_ib_handle_inet_event(struct usnic_ib_dev *us_ibdev,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct in_ifaddr *ifa = ptr;
|
|
struct ib_event ib_event;
|
|
|
|
mutex_lock(&us_ibdev->usdev_lock);
|
|
|
|
switch (event) {
|
|
case NETDEV_DOWN:
|
|
usnic_info("%s via ip notifiers",
|
|
netdev_cmd_to_name(event));
|
|
usnic_fwd_del_ipaddr(us_ibdev->ufdev);
|
|
usnic_ib_qp_grp_modify_active_to_err(us_ibdev);
|
|
ib_event.event = IB_EVENT_GID_CHANGE;
|
|
ib_event.device = &us_ibdev->ib_dev;
|
|
ib_event.element.port_num = 1;
|
|
ib_dispatch_event(&ib_event);
|
|
break;
|
|
case NETDEV_UP:
|
|
usnic_fwd_add_ipaddr(us_ibdev->ufdev, ifa->ifa_address);
|
|
usnic_info("%s via ip notifiers: ip %pI4",
|
|
netdev_cmd_to_name(event),
|
|
&us_ibdev->ufdev->inaddr);
|
|
ib_event.event = IB_EVENT_GID_CHANGE;
|
|
ib_event.device = &us_ibdev->ib_dev;
|
|
ib_event.element.port_num = 1;
|
|
ib_dispatch_event(&ib_event);
|
|
break;
|
|
default:
|
|
usnic_info("Ignoring event %s on %s",
|
|
netdev_cmd_to_name(event),
|
|
us_ibdev->ib_dev.name);
|
|
}
|
|
mutex_unlock(&us_ibdev->usdev_lock);
|
|
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
static int usnic_ib_inetaddr_event(struct notifier_block *notifier,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct usnic_ib_dev *us_ibdev;
|
|
struct in_ifaddr *ifa = ptr;
|
|
struct net_device *netdev = ifa->ifa_dev->dev;
|
|
|
|
mutex_lock(&usnic_ib_ibdev_list_lock);
|
|
list_for_each_entry(us_ibdev, &usnic_ib_ibdev_list, ib_dev_link) {
|
|
if (us_ibdev->netdev == netdev) {
|
|
usnic_ib_handle_inet_event(us_ibdev, event, ptr);
|
|
break;
|
|
}
|
|
}
|
|
mutex_unlock(&usnic_ib_ibdev_list_lock);
|
|
|
|
return NOTIFY_DONE;
|
|
}
|
|
static struct notifier_block usnic_ib_inetaddr_notifier = {
|
|
.notifier_call = usnic_ib_inetaddr_event
|
|
};
|
|
/* End of inet section*/
|
|
|
|
static int usnic_port_immutable(struct ib_device *ibdev, u8 port_num,
|
|
struct ib_port_immutable *immutable)
|
|
{
|
|
struct ib_port_attr attr;
|
|
int err;
|
|
|
|
immutable->core_cap_flags = RDMA_CORE_PORT_USNIC;
|
|
|
|
err = ib_query_port(ibdev, port_num, &attr);
|
|
if (err)
|
|
return err;
|
|
|
|
immutable->pkey_tbl_len = attr.pkey_tbl_len;
|
|
immutable->gid_tbl_len = attr.gid_tbl_len;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void usnic_get_dev_fw_str(struct ib_device *device, char *str)
|
|
{
|
|
struct usnic_ib_dev *us_ibdev =
|
|
container_of(device, struct usnic_ib_dev, ib_dev);
|
|
struct ethtool_drvinfo info;
|
|
|
|
mutex_lock(&us_ibdev->usdev_lock);
|
|
us_ibdev->netdev->ethtool_ops->get_drvinfo(us_ibdev->netdev, &info);
|
|
mutex_unlock(&us_ibdev->usdev_lock);
|
|
|
|
snprintf(str, IB_FW_VERSION_NAME_MAX, "%s", info.fw_version);
|
|
}
|
|
|
|
/* Start of PF discovery section */
|
|
static void *usnic_ib_device_add(struct pci_dev *dev)
|
|
{
|
|
struct usnic_ib_dev *us_ibdev;
|
|
union ib_gid gid;
|
|
struct in_device *ind;
|
|
struct net_device *netdev;
|
|
|
|
usnic_dbg("\n");
|
|
netdev = pci_get_drvdata(dev);
|
|
|
|
us_ibdev = (struct usnic_ib_dev *)ib_alloc_device(sizeof(*us_ibdev));
|
|
if (!us_ibdev) {
|
|
usnic_err("Device %s context alloc failed\n",
|
|
netdev_name(pci_get_drvdata(dev)));
|
|
return ERR_PTR(-EFAULT);
|
|
}
|
|
|
|
us_ibdev->ufdev = usnic_fwd_dev_alloc(dev);
|
|
if (!us_ibdev->ufdev) {
|
|
usnic_err("Failed to alloc ufdev for %s\n", pci_name(dev));
|
|
goto err_dealloc;
|
|
}
|
|
|
|
mutex_init(&us_ibdev->usdev_lock);
|
|
INIT_LIST_HEAD(&us_ibdev->vf_dev_list);
|
|
INIT_LIST_HEAD(&us_ibdev->ctx_list);
|
|
|
|
us_ibdev->pdev = dev;
|
|
us_ibdev->netdev = pci_get_drvdata(dev);
|
|
us_ibdev->ib_dev.owner = THIS_MODULE;
|
|
us_ibdev->ib_dev.node_type = RDMA_NODE_USNIC_UDP;
|
|
us_ibdev->ib_dev.phys_port_cnt = USNIC_IB_PORT_CNT;
|
|
us_ibdev->ib_dev.num_comp_vectors = USNIC_IB_NUM_COMP_VECTORS;
|
|
us_ibdev->ib_dev.dev.parent = &dev->dev;
|
|
us_ibdev->ib_dev.uverbs_abi_ver = USNIC_UVERBS_ABI_VERSION;
|
|
strlcpy(us_ibdev->ib_dev.name, "usnic_%d", IB_DEVICE_NAME_MAX);
|
|
|
|
us_ibdev->ib_dev.uverbs_cmd_mask =
|
|
(1ull << IB_USER_VERBS_CMD_GET_CONTEXT) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_PORT) |
|
|
(1ull << IB_USER_VERBS_CMD_ALLOC_PD) |
|
|
(1ull << IB_USER_VERBS_CMD_DEALLOC_PD) |
|
|
(1ull << IB_USER_VERBS_CMD_REG_MR) |
|
|
(1ull << IB_USER_VERBS_CMD_DEREG_MR) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_DESTROY_CQ) |
|
|
(1ull << IB_USER_VERBS_CMD_CREATE_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_MODIFY_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_QUERY_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_DESTROY_QP) |
|
|
(1ull << IB_USER_VERBS_CMD_ATTACH_MCAST) |
|
|
(1ull << IB_USER_VERBS_CMD_DETACH_MCAST) |
|
|
(1ull << IB_USER_VERBS_CMD_OPEN_QP);
|
|
|
|
us_ibdev->ib_dev.query_device = usnic_ib_query_device;
|
|
us_ibdev->ib_dev.query_port = usnic_ib_query_port;
|
|
us_ibdev->ib_dev.query_pkey = usnic_ib_query_pkey;
|
|
us_ibdev->ib_dev.query_gid = usnic_ib_query_gid;
|
|
us_ibdev->ib_dev.get_netdev = usnic_get_netdev;
|
|
us_ibdev->ib_dev.get_link_layer = usnic_ib_port_link_layer;
|
|
us_ibdev->ib_dev.alloc_pd = usnic_ib_alloc_pd;
|
|
us_ibdev->ib_dev.dealloc_pd = usnic_ib_dealloc_pd;
|
|
us_ibdev->ib_dev.create_qp = usnic_ib_create_qp;
|
|
us_ibdev->ib_dev.modify_qp = usnic_ib_modify_qp;
|
|
us_ibdev->ib_dev.query_qp = usnic_ib_query_qp;
|
|
us_ibdev->ib_dev.destroy_qp = usnic_ib_destroy_qp;
|
|
us_ibdev->ib_dev.create_cq = usnic_ib_create_cq;
|
|
us_ibdev->ib_dev.destroy_cq = usnic_ib_destroy_cq;
|
|
us_ibdev->ib_dev.reg_user_mr = usnic_ib_reg_mr;
|
|
us_ibdev->ib_dev.dereg_mr = usnic_ib_dereg_mr;
|
|
us_ibdev->ib_dev.alloc_ucontext = usnic_ib_alloc_ucontext;
|
|
us_ibdev->ib_dev.dealloc_ucontext = usnic_ib_dealloc_ucontext;
|
|
us_ibdev->ib_dev.mmap = usnic_ib_mmap;
|
|
us_ibdev->ib_dev.create_ah = usnic_ib_create_ah;
|
|
us_ibdev->ib_dev.destroy_ah = usnic_ib_destroy_ah;
|
|
us_ibdev->ib_dev.post_send = usnic_ib_post_send;
|
|
us_ibdev->ib_dev.post_recv = usnic_ib_post_recv;
|
|
us_ibdev->ib_dev.poll_cq = usnic_ib_poll_cq;
|
|
us_ibdev->ib_dev.req_notify_cq = usnic_ib_req_notify_cq;
|
|
us_ibdev->ib_dev.get_dma_mr = usnic_ib_get_dma_mr;
|
|
us_ibdev->ib_dev.get_port_immutable = usnic_port_immutable;
|
|
us_ibdev->ib_dev.get_dev_fw_str = usnic_get_dev_fw_str;
|
|
|
|
|
|
us_ibdev->ib_dev.driver_id = RDMA_DRIVER_USNIC;
|
|
if (ib_register_device(&us_ibdev->ib_dev, NULL))
|
|
goto err_fwd_dealloc;
|
|
|
|
usnic_fwd_set_mtu(us_ibdev->ufdev, us_ibdev->netdev->mtu);
|
|
usnic_fwd_set_mac(us_ibdev->ufdev, us_ibdev->netdev->dev_addr);
|
|
if (netif_carrier_ok(us_ibdev->netdev))
|
|
usnic_fwd_carrier_up(us_ibdev->ufdev);
|
|
|
|
ind = in_dev_get(netdev);
|
|
if (ind->ifa_list)
|
|
usnic_fwd_add_ipaddr(us_ibdev->ufdev,
|
|
ind->ifa_list->ifa_address);
|
|
in_dev_put(ind);
|
|
|
|
usnic_mac_ip_to_gid(us_ibdev->netdev->perm_addr,
|
|
us_ibdev->ufdev->inaddr, &gid.raw[0]);
|
|
memcpy(&us_ibdev->ib_dev.node_guid, &gid.global.interface_id,
|
|
sizeof(gid.global.interface_id));
|
|
kref_init(&us_ibdev->vf_cnt);
|
|
|
|
usnic_info("Added ibdev: %s netdev: %s with mac %pM Link: %u MTU: %u\n",
|
|
us_ibdev->ib_dev.name, netdev_name(us_ibdev->netdev),
|
|
us_ibdev->ufdev->mac, us_ibdev->ufdev->link_up,
|
|
us_ibdev->ufdev->mtu);
|
|
return us_ibdev;
|
|
|
|
err_fwd_dealloc:
|
|
usnic_fwd_dev_free(us_ibdev->ufdev);
|
|
err_dealloc:
|
|
usnic_err("failed -- deallocing device\n");
|
|
ib_dealloc_device(&us_ibdev->ib_dev);
|
|
return NULL;
|
|
}
|
|
|
|
static void usnic_ib_device_remove(struct usnic_ib_dev *us_ibdev)
|
|
{
|
|
usnic_info("Unregistering %s\n", us_ibdev->ib_dev.name);
|
|
usnic_ib_sysfs_unregister_usdev(us_ibdev);
|
|
usnic_fwd_dev_free(us_ibdev->ufdev);
|
|
ib_unregister_device(&us_ibdev->ib_dev);
|
|
ib_dealloc_device(&us_ibdev->ib_dev);
|
|
}
|
|
|
|
static void usnic_ib_undiscover_pf(struct kref *kref)
|
|
{
|
|
struct usnic_ib_dev *us_ibdev, *tmp;
|
|
struct pci_dev *dev;
|
|
bool found = false;
|
|
|
|
dev = container_of(kref, struct usnic_ib_dev, vf_cnt)->pdev;
|
|
mutex_lock(&usnic_ib_ibdev_list_lock);
|
|
list_for_each_entry_safe(us_ibdev, tmp,
|
|
&usnic_ib_ibdev_list, ib_dev_link) {
|
|
if (us_ibdev->pdev == dev) {
|
|
list_del(&us_ibdev->ib_dev_link);
|
|
usnic_ib_device_remove(us_ibdev);
|
|
found = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
WARN(!found, "Failed to remove PF %s\n", pci_name(dev));
|
|
|
|
mutex_unlock(&usnic_ib_ibdev_list_lock);
|
|
}
|
|
|
|
static struct usnic_ib_dev *usnic_ib_discover_pf(struct usnic_vnic *vnic)
|
|
{
|
|
struct usnic_ib_dev *us_ibdev;
|
|
struct pci_dev *parent_pci, *vf_pci;
|
|
int err;
|
|
|
|
vf_pci = usnic_vnic_get_pdev(vnic);
|
|
parent_pci = pci_physfn(vf_pci);
|
|
|
|
BUG_ON(!parent_pci);
|
|
|
|
mutex_lock(&usnic_ib_ibdev_list_lock);
|
|
list_for_each_entry(us_ibdev, &usnic_ib_ibdev_list, ib_dev_link) {
|
|
if (us_ibdev->pdev == parent_pci) {
|
|
kref_get(&us_ibdev->vf_cnt);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
us_ibdev = usnic_ib_device_add(parent_pci);
|
|
if (IS_ERR_OR_NULL(us_ibdev)) {
|
|
us_ibdev = us_ibdev ? us_ibdev : ERR_PTR(-EFAULT);
|
|
goto out;
|
|
}
|
|
|
|
err = usnic_ib_sysfs_register_usdev(us_ibdev);
|
|
if (err) {
|
|
usnic_ib_device_remove(us_ibdev);
|
|
us_ibdev = ERR_PTR(err);
|
|
goto out;
|
|
}
|
|
|
|
list_add(&us_ibdev->ib_dev_link, &usnic_ib_ibdev_list);
|
|
out:
|
|
mutex_unlock(&usnic_ib_ibdev_list_lock);
|
|
return us_ibdev;
|
|
}
|
|
/* End of PF discovery section */
|
|
|
|
/* Start of PCI section */
|
|
|
|
static const struct pci_device_id usnic_ib_pci_ids[] = {
|
|
{PCI_DEVICE(PCI_VENDOR_ID_CISCO, PCI_DEVICE_ID_CISCO_VIC_USPACE_NIC)},
|
|
{0,}
|
|
};
|
|
|
|
static int usnic_ib_pci_probe(struct pci_dev *pdev,
|
|
const struct pci_device_id *id)
|
|
{
|
|
int err;
|
|
struct usnic_ib_dev *pf;
|
|
struct usnic_ib_vf *vf;
|
|
enum usnic_vnic_res_type res_type;
|
|
|
|
vf = kzalloc(sizeof(*vf), GFP_KERNEL);
|
|
if (!vf)
|
|
return -ENOMEM;
|
|
|
|
err = pci_enable_device(pdev);
|
|
if (err) {
|
|
usnic_err("Failed to enable %s with err %d\n",
|
|
pci_name(pdev), err);
|
|
goto out_clean_vf;
|
|
}
|
|
|
|
err = pci_request_regions(pdev, DRV_NAME);
|
|
if (err) {
|
|
usnic_err("Failed to request region for %s with err %d\n",
|
|
pci_name(pdev), err);
|
|
goto out_disable_device;
|
|
}
|
|
|
|
pci_set_master(pdev);
|
|
pci_set_drvdata(pdev, vf);
|
|
|
|
vf->vnic = usnic_vnic_alloc(pdev);
|
|
if (IS_ERR_OR_NULL(vf->vnic)) {
|
|
err = vf->vnic ? PTR_ERR(vf->vnic) : -ENOMEM;
|
|
usnic_err("Failed to alloc vnic for %s with err %d\n",
|
|
pci_name(pdev), err);
|
|
goto out_release_regions;
|
|
}
|
|
|
|
pf = usnic_ib_discover_pf(vf->vnic);
|
|
if (IS_ERR_OR_NULL(pf)) {
|
|
usnic_err("Failed to discover pf of vnic %s with err%ld\n",
|
|
pci_name(pdev), PTR_ERR(pf));
|
|
err = pf ? PTR_ERR(pf) : -EFAULT;
|
|
goto out_clean_vnic;
|
|
}
|
|
|
|
vf->pf = pf;
|
|
spin_lock_init(&vf->lock);
|
|
mutex_lock(&pf->usdev_lock);
|
|
list_add_tail(&vf->link, &pf->vf_dev_list);
|
|
/*
|
|
* Save max settings (will be same for each VF, easier to re-write than
|
|
* to say "if (!set) { set_values(); set=1; }
|
|
*/
|
|
for (res_type = USNIC_VNIC_RES_TYPE_EOL+1;
|
|
res_type < USNIC_VNIC_RES_TYPE_MAX;
|
|
res_type++) {
|
|
pf->vf_res_cnt[res_type] = usnic_vnic_res_cnt(vf->vnic,
|
|
res_type);
|
|
}
|
|
|
|
mutex_unlock(&pf->usdev_lock);
|
|
|
|
usnic_info("Registering usnic VF %s into PF %s\n", pci_name(pdev),
|
|
pf->ib_dev.name);
|
|
usnic_ib_log_vf(vf);
|
|
return 0;
|
|
|
|
out_clean_vnic:
|
|
usnic_vnic_free(vf->vnic);
|
|
out_release_regions:
|
|
pci_set_drvdata(pdev, NULL);
|
|
pci_clear_master(pdev);
|
|
pci_release_regions(pdev);
|
|
out_disable_device:
|
|
pci_disable_device(pdev);
|
|
out_clean_vf:
|
|
kfree(vf);
|
|
return err;
|
|
}
|
|
|
|
static void usnic_ib_pci_remove(struct pci_dev *pdev)
|
|
{
|
|
struct usnic_ib_vf *vf = pci_get_drvdata(pdev);
|
|
struct usnic_ib_dev *pf = vf->pf;
|
|
|
|
mutex_lock(&pf->usdev_lock);
|
|
list_del(&vf->link);
|
|
mutex_unlock(&pf->usdev_lock);
|
|
|
|
kref_put(&pf->vf_cnt, usnic_ib_undiscover_pf);
|
|
usnic_vnic_free(vf->vnic);
|
|
pci_set_drvdata(pdev, NULL);
|
|
pci_clear_master(pdev);
|
|
pci_release_regions(pdev);
|
|
pci_disable_device(pdev);
|
|
kfree(vf);
|
|
|
|
usnic_info("Removed VF %s\n", pci_name(pdev));
|
|
}
|
|
|
|
/* PCI driver entry points */
|
|
static struct pci_driver usnic_ib_pci_driver = {
|
|
.name = DRV_NAME,
|
|
.id_table = usnic_ib_pci_ids,
|
|
.probe = usnic_ib_pci_probe,
|
|
.remove = usnic_ib_pci_remove,
|
|
};
|
|
/* End of PCI section */
|
|
|
|
/* Start of module section */
|
|
static int __init usnic_ib_init(void)
|
|
{
|
|
int err;
|
|
|
|
printk_once(KERN_INFO "%s", usnic_version);
|
|
|
|
err = usnic_uiom_init(DRV_NAME);
|
|
if (err) {
|
|
usnic_err("Unable to initalize umem with err %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
err = pci_register_driver(&usnic_ib_pci_driver);
|
|
if (err) {
|
|
usnic_err("Unable to register with PCI\n");
|
|
goto out_umem_fini;
|
|
}
|
|
|
|
err = register_netdevice_notifier(&usnic_ib_netdevice_notifier);
|
|
if (err) {
|
|
usnic_err("Failed to register netdev notifier\n");
|
|
goto out_pci_unreg;
|
|
}
|
|
|
|
err = register_inetaddr_notifier(&usnic_ib_inetaddr_notifier);
|
|
if (err) {
|
|
usnic_err("Failed to register inet addr notifier\n");
|
|
goto out_unreg_netdev_notifier;
|
|
}
|
|
|
|
err = usnic_transport_init();
|
|
if (err) {
|
|
usnic_err("Failed to initialize transport\n");
|
|
goto out_unreg_inetaddr_notifier;
|
|
}
|
|
|
|
usnic_debugfs_init();
|
|
|
|
return 0;
|
|
|
|
out_unreg_inetaddr_notifier:
|
|
unregister_inetaddr_notifier(&usnic_ib_inetaddr_notifier);
|
|
out_unreg_netdev_notifier:
|
|
unregister_netdevice_notifier(&usnic_ib_netdevice_notifier);
|
|
out_pci_unreg:
|
|
pci_unregister_driver(&usnic_ib_pci_driver);
|
|
out_umem_fini:
|
|
usnic_uiom_fini();
|
|
|
|
return err;
|
|
}
|
|
|
|
static void __exit usnic_ib_destroy(void)
|
|
{
|
|
usnic_dbg("\n");
|
|
usnic_debugfs_exit();
|
|
usnic_transport_fini();
|
|
unregister_inetaddr_notifier(&usnic_ib_inetaddr_notifier);
|
|
unregister_netdevice_notifier(&usnic_ib_netdevice_notifier);
|
|
pci_unregister_driver(&usnic_ib_pci_driver);
|
|
usnic_uiom_fini();
|
|
}
|
|
|
|
MODULE_DESCRIPTION("Cisco VIC (usNIC) Verbs Driver");
|
|
MODULE_AUTHOR("Upinder Malhi <umalhi@cisco.com>");
|
|
MODULE_LICENSE("Dual BSD/GPL");
|
|
module_param(usnic_log_lvl, uint, S_IRUGO | S_IWUSR);
|
|
module_param(usnic_ib_share_vf, uint, S_IRUGO | S_IWUSR);
|
|
MODULE_PARM_DESC(usnic_log_lvl, " Off=0, Err=1, Info=2, Debug=3");
|
|
MODULE_PARM_DESC(usnic_ib_share_vf, "Off=0, On=1 VF sharing amongst QPs");
|
|
MODULE_DEVICE_TABLE(pci, usnic_ib_pci_ids);
|
|
|
|
module_init(usnic_ib_init);
|
|
module_exit(usnic_ib_destroy);
|
|
/* End of module section */
|