mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 05:32:00 +00:00
RDMA/mana_ib: Introduce mana_ib_get_netdev helper function
Use a helper function to access netdevs using a port number. This patch removes code repetitions as well as removes the need to explicitly use gdma_dev, which was error-prone. Signed-off-by: Konstantin Taranov <kotaranov@microsoft.com> Link: https://lore.kernel.org/r/1705965781-3235-3-git-send-email-kotaranov@linux.microsoft.com Signed-off-by: Leon Romanovsky <leon@kernel.org>
This commit is contained in:
parent
71c8cbfcdc
commit
3b73eb3a4a
@ -8,13 +8,10 @@
|
||||
void mana_ib_uncfg_vport(struct mana_ib_dev *dev, struct mana_ib_pd *pd,
|
||||
u32 port)
|
||||
{
|
||||
struct gdma_dev *gd = &dev->gdma_dev->gdma_context->mana;
|
||||
struct mana_port_context *mpc;
|
||||
struct net_device *ndev;
|
||||
struct mana_context *mc;
|
||||
|
||||
mc = gd->driver_data;
|
||||
ndev = mc->ports[port];
|
||||
ndev = mana_ib_get_netdev(&dev->ib_dev, port);
|
||||
mpc = netdev_priv(ndev);
|
||||
|
||||
mutex_lock(&pd->vport_mutex);
|
||||
@ -31,14 +28,11 @@ void mana_ib_uncfg_vport(struct mana_ib_dev *dev, struct mana_ib_pd *pd,
|
||||
int mana_ib_cfg_vport(struct mana_ib_dev *dev, u32 port, struct mana_ib_pd *pd,
|
||||
u32 doorbell_id)
|
||||
{
|
||||
struct gdma_dev *mdev = &dev->gdma_dev->gdma_context->mana;
|
||||
struct mana_port_context *mpc;
|
||||
struct mana_context *mc;
|
||||
struct net_device *ndev;
|
||||
int err;
|
||||
|
||||
mc = mdev->driver_data;
|
||||
ndev = mc->ports[port];
|
||||
ndev = mana_ib_get_netdev(&dev->ib_dev, port);
|
||||
mpc = netdev_priv(ndev);
|
||||
|
||||
mutex_lock(&pd->vport_mutex);
|
||||
|
@ -147,6 +147,17 @@ static inline struct gdma_context *mdev_to_gc(struct mana_ib_dev *mdev)
|
||||
return mdev->gdma_dev->gdma_context;
|
||||
}
|
||||
|
||||
static inline struct net_device *mana_ib_get_netdev(struct ib_device *ibdev, u32 port)
|
||||
{
|
||||
struct mana_ib_dev *mdev = container_of(ibdev, struct mana_ib_dev, ib_dev);
|
||||
struct gdma_context *gc = mdev_to_gc(mdev);
|
||||
struct mana_context *mc = gc->mana.driver_data;
|
||||
|
||||
if (port < 1 || port > mc->num_ports)
|
||||
return NULL;
|
||||
return mc->ports[port - 1];
|
||||
}
|
||||
|
||||
int mana_ib_gd_create_dma_region(struct mana_ib_dev *dev, struct ib_umem *umem,
|
||||
mana_handle_t *gdma_region);
|
||||
|
||||
|
@ -106,11 +106,9 @@ static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd,
|
||||
struct mana_port_context *mpc;
|
||||
struct gdma_queue *gdma_cq;
|
||||
unsigned int ind_tbl_size;
|
||||
struct mana_context *mc;
|
||||
struct net_device *ndev;
|
||||
struct mana_ib_cq *cq;
|
||||
struct mana_ib_wq *wq;
|
||||
struct gdma_dev *gd;
|
||||
struct mana_eq *eq;
|
||||
struct ib_cq *ibcq;
|
||||
struct ib_wq *ibwq;
|
||||
@ -118,9 +116,6 @@ static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd,
|
||||
u32 port;
|
||||
int ret;
|
||||
|
||||
gd = &gc->mana;
|
||||
mc = gd->driver_data;
|
||||
|
||||
if (!udata || udata->inlen < sizeof(ucmd))
|
||||
return -EINVAL;
|
||||
|
||||
@ -163,12 +158,12 @@ static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd,
|
||||
|
||||
/* IB ports start with 1, MANA start with 0 */
|
||||
port = ucmd.port;
|
||||
if (port < 1 || port > mc->num_ports) {
|
||||
ndev = mana_ib_get_netdev(pd->device, port);
|
||||
if (!ndev) {
|
||||
ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n",
|
||||
port);
|
||||
return -EINVAL;
|
||||
}
|
||||
ndev = mc->ports[port - 1];
|
||||
mpc = netdev_priv(ndev);
|
||||
|
||||
ibdev_dbg(&mdev->ib_dev, "rx_hash_function %d port %d\n",
|
||||
@ -206,7 +201,7 @@ static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd,
|
||||
cq_spec.gdma_region = cq->gdma_region;
|
||||
cq_spec.queue_size = cq->cqe * COMP_ENTRY_SIZE;
|
||||
cq_spec.modr_ctx_id = 0;
|
||||
eq = &mc->eqs[cq->comp_vector % gc->max_num_queues];
|
||||
eq = &mpc->ac->eqs[cq->comp_vector % gc->max_num_queues];
|
||||
cq_spec.attached_eq = eq->eq->id;
|
||||
|
||||
ret = mana_create_wq_obj(mpc, mpc->port_handle, GDMA_RQ,
|
||||
@ -303,7 +298,6 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
struct mana_ib_ucontext *mana_ucontext =
|
||||
rdma_udata_to_drv_context(udata, struct mana_ib_ucontext,
|
||||
ibucontext);
|
||||
struct gdma_dev *gd = &mdev->gdma_dev->gdma_context->mana;
|
||||
struct gdma_context *gc = mdev_to_gc(mdev);
|
||||
struct mana_ib_create_qp_resp resp = {};
|
||||
struct mana_ib_create_qp ucmd = {};
|
||||
@ -311,7 +305,6 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
struct mana_obj_spec wq_spec = {};
|
||||
struct mana_obj_spec cq_spec = {};
|
||||
struct mana_port_context *mpc;
|
||||
struct mana_context *mc;
|
||||
struct net_device *ndev;
|
||||
struct ib_umem *umem;
|
||||
struct mana_eq *eq;
|
||||
@ -319,8 +312,6 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
u32 port;
|
||||
int err;
|
||||
|
||||
mc = gd->driver_data;
|
||||
|
||||
if (!mana_ucontext || udata->inlen < sizeof(ucmd))
|
||||
return -EINVAL;
|
||||
|
||||
@ -331,11 +322,6 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
return err;
|
||||
}
|
||||
|
||||
/* IB ports start with 1, MANA Ethernet ports start with 0 */
|
||||
port = ucmd.port;
|
||||
if (port < 1 || port > mc->num_ports)
|
||||
return -EINVAL;
|
||||
|
||||
if (attr->cap.max_send_wr > mdev->adapter_caps.max_qp_wr) {
|
||||
ibdev_dbg(&mdev->ib_dev,
|
||||
"Requested max_send_wr %d exceeding limit\n",
|
||||
@ -350,11 +336,17 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
ndev = mc->ports[port - 1];
|
||||
port = ucmd.port;
|
||||
ndev = mana_ib_get_netdev(ibpd->device, port);
|
||||
if (!ndev) {
|
||||
ibdev_dbg(&mdev->ib_dev, "Invalid port %u in creating qp\n",
|
||||
port);
|
||||
return -EINVAL;
|
||||
}
|
||||
mpc = netdev_priv(ndev);
|
||||
ibdev_dbg(&mdev->ib_dev, "port %u ndev %p mpc %p\n", port, ndev, mpc);
|
||||
|
||||
err = mana_ib_cfg_vport(mdev, port - 1, pd, mana_ucontext->doorbell);
|
||||
err = mana_ib_cfg_vport(mdev, port, pd, mana_ucontext->doorbell);
|
||||
if (err)
|
||||
return -ENODEV;
|
||||
|
||||
@ -394,8 +386,8 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
cq_spec.gdma_region = send_cq->gdma_region;
|
||||
cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE;
|
||||
cq_spec.modr_ctx_id = 0;
|
||||
eq_vec = send_cq->comp_vector % gd->gdma_context->max_num_queues;
|
||||
eq = &mc->eqs[eq_vec];
|
||||
eq_vec = send_cq->comp_vector % gc->max_num_queues;
|
||||
eq = &mpc->ac->eqs[eq_vec];
|
||||
cq_spec.attached_eq = eq->eq->id;
|
||||
|
||||
err = mana_create_wq_obj(mpc, mpc->port_handle, GDMA_SQ, &wq_spec,
|
||||
@ -415,7 +407,7 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
send_cq->id = cq_spec.queue_index;
|
||||
|
||||
/* Create CQ table entry */
|
||||
WARN_ON(gd->gdma_context->cq_table[send_cq->id]);
|
||||
WARN_ON(gc->cq_table[send_cq->id]);
|
||||
gdma_cq = kzalloc(sizeof(*gdma_cq), GFP_KERNEL);
|
||||
if (!gdma_cq) {
|
||||
err = -ENOMEM;
|
||||
@ -426,7 +418,7 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
gdma_cq->type = GDMA_CQ;
|
||||
gdma_cq->cq.callback = mana_ib_cq_handler;
|
||||
gdma_cq->id = send_cq->id;
|
||||
gd->gdma_context->cq_table[send_cq->id] = gdma_cq;
|
||||
gc->cq_table[send_cq->id] = gdma_cq;
|
||||
|
||||
ibdev_dbg(&mdev->ib_dev,
|
||||
"ret %d qp->tx_object 0x%llx sq id %llu cq id %llu\n", err,
|
||||
@ -460,7 +452,7 @@ err_release_umem:
|
||||
ib_umem_release(umem);
|
||||
|
||||
err_free_vport:
|
||||
mana_ib_uncfg_vport(mdev, pd, port - 1);
|
||||
mana_ib_uncfg_vport(mdev, pd, port);
|
||||
|
||||
return err;
|
||||
}
|
||||
@ -498,16 +490,13 @@ static int mana_ib_destroy_qp_rss(struct mana_ib_qp *qp,
|
||||
{
|
||||
struct mana_ib_dev *mdev =
|
||||
container_of(qp->ibqp.device, struct mana_ib_dev, ib_dev);
|
||||
struct gdma_dev *gd = &mdev->gdma_dev->gdma_context->mana;
|
||||
struct mana_port_context *mpc;
|
||||
struct mana_context *mc;
|
||||
struct net_device *ndev;
|
||||
struct mana_ib_wq *wq;
|
||||
struct ib_wq *ibwq;
|
||||
int i;
|
||||
|
||||
mc = gd->driver_data;
|
||||
ndev = mc->ports[qp->port - 1];
|
||||
ndev = mana_ib_get_netdev(qp->ibqp.device, qp->port);
|
||||
mpc = netdev_priv(ndev);
|
||||
|
||||
for (i = 0; i < (1 << ind_tbl->log_ind_tbl_size); i++) {
|
||||
@ -525,15 +514,12 @@ static int mana_ib_destroy_qp_raw(struct mana_ib_qp *qp, struct ib_udata *udata)
|
||||
{
|
||||
struct mana_ib_dev *mdev =
|
||||
container_of(qp->ibqp.device, struct mana_ib_dev, ib_dev);
|
||||
struct gdma_dev *gd = &mdev->gdma_dev->gdma_context->mana;
|
||||
struct ib_pd *ibpd = qp->ibqp.pd;
|
||||
struct mana_port_context *mpc;
|
||||
struct mana_context *mc;
|
||||
struct net_device *ndev;
|
||||
struct mana_ib_pd *pd;
|
||||
|
||||
mc = gd->driver_data;
|
||||
ndev = mc->ports[qp->port - 1];
|
||||
ndev = mana_ib_get_netdev(qp->ibqp.device, qp->port);
|
||||
mpc = netdev_priv(ndev);
|
||||
pd = container_of(ibpd, struct mana_ib_pd, ibpd);
|
||||
|
||||
@ -544,7 +530,7 @@ static int mana_ib_destroy_qp_raw(struct mana_ib_qp *qp, struct ib_udata *udata)
|
||||
ib_umem_release(qp->sq_umem);
|
||||
}
|
||||
|
||||
mana_ib_uncfg_vport(mdev, pd, qp->port - 1);
|
||||
mana_ib_uncfg_vport(mdev, pd, qp->port);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user