mirror of
https://github.com/torvalds/linux.git
synced 2024-11-28 15:11:31 +00:00
xprtrdma: No qp_event disconnect
After thinking about this more, and auditing other kernel ULP imple- mentations, I believe that a DISCONNECT cm_event will occur after a fatal QP event. If that's the case, there's no need for an explicit disconnect in the QP event handler. Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
This commit is contained in:
parent
6d2d0ee27c
commit
3d433ad812
@ -99,25 +99,6 @@ static void rpcrdma_xprt_drain(struct rpcrdma_xprt *r_xprt)
|
|||||||
ib_drain_sq(ia->ri_id->qp);
|
ib_drain_sq(ia->ri_id->qp);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* rpcrdma_disconnect_worker - Force a disconnect
|
|
||||||
* @work: endpoint to be disconnected
|
|
||||||
*
|
|
||||||
* Provider callbacks can possibly run in an IRQ context. This function
|
|
||||||
* is invoked in a worker thread to guarantee that disconnect wake-up
|
|
||||||
* calls are always done in process context.
|
|
||||||
*/
|
|
||||||
static void
|
|
||||||
rpcrdma_disconnect_worker(struct work_struct *work)
|
|
||||||
{
|
|
||||||
struct rpcrdma_ep *ep = container_of(work, struct rpcrdma_ep,
|
|
||||||
rep_disconnect_worker.work);
|
|
||||||
struct rpcrdma_xprt *r_xprt =
|
|
||||||
container_of(ep, struct rpcrdma_xprt, rx_ep);
|
|
||||||
|
|
||||||
xprt_force_disconnect(&r_xprt->rx_xprt);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* rpcrdma_qp_event_handler - Handle one QP event (error notification)
|
* rpcrdma_qp_event_handler - Handle one QP event (error notification)
|
||||||
* @event: details of the event
|
* @event: details of the event
|
||||||
@ -134,15 +115,6 @@ rpcrdma_qp_event_handler(struct ib_event *event, void *context)
|
|||||||
rx_ep);
|
rx_ep);
|
||||||
|
|
||||||
trace_xprtrdma_qp_event(r_xprt, event);
|
trace_xprtrdma_qp_event(r_xprt, event);
|
||||||
pr_err("rpcrdma: %s on device %s connected to %s:%s\n",
|
|
||||||
ib_event_msg(event->event), event->device->name,
|
|
||||||
rpcrdma_addrstr(r_xprt), rpcrdma_portstr(r_xprt));
|
|
||||||
|
|
||||||
if (ep->rep_connected == 1) {
|
|
||||||
ep->rep_connected = -EIO;
|
|
||||||
schedule_delayed_work(&ep->rep_disconnect_worker, 0);
|
|
||||||
wake_up_all(&ep->rep_connect_wait);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -571,8 +543,6 @@ rpcrdma_ep_create(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia,
|
|||||||
cdata->max_requests >> 2);
|
cdata->max_requests >> 2);
|
||||||
ep->rep_send_count = ep->rep_send_batch;
|
ep->rep_send_count = ep->rep_send_batch;
|
||||||
init_waitqueue_head(&ep->rep_connect_wait);
|
init_waitqueue_head(&ep->rep_connect_wait);
|
||||||
INIT_DELAYED_WORK(&ep->rep_disconnect_worker,
|
|
||||||
rpcrdma_disconnect_worker);
|
|
||||||
ep->rep_receive_count = 0;
|
ep->rep_receive_count = 0;
|
||||||
|
|
||||||
sendcq = ib_alloc_cq(ia->ri_device, NULL,
|
sendcq = ib_alloc_cq(ia->ri_device, NULL,
|
||||||
@ -646,8 +616,6 @@ out1:
|
|||||||
void
|
void
|
||||||
rpcrdma_ep_destroy(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia)
|
rpcrdma_ep_destroy(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia)
|
||||||
{
|
{
|
||||||
cancel_delayed_work_sync(&ep->rep_disconnect_worker);
|
|
||||||
|
|
||||||
if (ia->ri_id && ia->ri_id->qp) {
|
if (ia->ri_id && ia->ri_id->qp) {
|
||||||
rpcrdma_ep_disconnect(ep, ia);
|
rpcrdma_ep_disconnect(ep, ia);
|
||||||
rdma_destroy_qp(ia->ri_id);
|
rdma_destroy_qp(ia->ri_id);
|
||||||
|
@ -101,7 +101,6 @@ struct rpcrdma_ep {
|
|||||||
wait_queue_head_t rep_connect_wait;
|
wait_queue_head_t rep_connect_wait;
|
||||||
struct rpcrdma_connect_private rep_cm_private;
|
struct rpcrdma_connect_private rep_cm_private;
|
||||||
struct rdma_conn_param rep_remote_cma;
|
struct rdma_conn_param rep_remote_cma;
|
||||||
struct delayed_work rep_disconnect_worker;
|
|
||||||
int rep_receive_count;
|
int rep_receive_count;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user