mirror of
https://github.com/torvalds/linux.git
synced 2024-11-24 05:02:12 +00:00
9b4b7c1f9f
Replace tasklets by work queues for the three main rxe tasklets: rxe_requester, rxe_completer and rxe_responder. work queues are a more modern way to process work from an IRQ and provide more control over how that work is run for future patches. Link: https://lore.kernel.org/r/20230428171321.5774-1-rpearsonhpe@gmail.com Signed-off-by: Ian Ziemba <ian.ziemba@hpe.com> Signed-off-by: Bob Pearson <rpearsonhpe@gmail.com> Reviewed-by: Daisuke Matsuda <matsuda-daisuke@fujitsu.com> Tested-by: Daisuke Matsuda <matsuda-daisuke@fujitsu.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
314 lines
7.6 KiB
C
314 lines
7.6 KiB
C
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
|
|
/*
|
|
* Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved.
|
|
* Copyright (c) 2015 System Fabric Works, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include "rxe.h"
|
|
|
|
static struct workqueue_struct *rxe_wq;
|
|
|
|
int rxe_alloc_wq(void)
|
|
{
|
|
rxe_wq = alloc_workqueue("rxe_wq", WQ_UNBOUND, WQ_MAX_ACTIVE);
|
|
if (!rxe_wq)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void rxe_destroy_wq(void)
|
|
{
|
|
destroy_workqueue(rxe_wq);
|
|
}
|
|
|
|
/* Check if task is idle i.e. not running, not scheduled in
|
|
* work queue and not draining. If so move to busy to
|
|
* reserve a slot in do_task() by setting to busy and taking
|
|
* a qp reference to cover the gap from now until the task finishes.
|
|
* state will move out of busy if task returns a non zero value
|
|
* in do_task(). If state is already busy it is raised to armed
|
|
* to indicate to do_task that additional pass should be made
|
|
* over the task.
|
|
* Context: caller should hold task->lock.
|
|
* Returns: true if state transitioned from idle to busy else false.
|
|
*/
|
|
static bool __reserve_if_idle(struct rxe_task *task)
|
|
{
|
|
WARN_ON(rxe_read(task->qp) <= 0);
|
|
|
|
if (task->state == TASK_STATE_IDLE) {
|
|
rxe_get(task->qp);
|
|
task->state = TASK_STATE_BUSY;
|
|
task->num_sched++;
|
|
return true;
|
|
}
|
|
|
|
if (task->state == TASK_STATE_BUSY)
|
|
task->state = TASK_STATE_ARMED;
|
|
|
|
return false;
|
|
}
|
|
|
|
/* check if task is idle or drained and not currently
|
|
* scheduled in the work queue. This routine is
|
|
* called by rxe_cleanup_task or rxe_disable_task to
|
|
* see if the queue is empty.
|
|
* Context: caller should hold task->lock.
|
|
* Returns true if done else false.
|
|
*/
|
|
static bool __is_done(struct rxe_task *task)
|
|
{
|
|
if (work_pending(&task->work))
|
|
return false;
|
|
|
|
if (task->state == TASK_STATE_IDLE ||
|
|
task->state == TASK_STATE_DRAINED) {
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/* a locked version of __is_done */
|
|
static bool is_done(struct rxe_task *task)
|
|
{
|
|
unsigned long flags;
|
|
int done;
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
done = __is_done(task);
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
|
|
return done;
|
|
}
|
|
|
|
/* do_task is a wrapper for the three tasks (requester,
|
|
* completer, responder) and calls them in a loop until
|
|
* they return a non-zero value. It is called either
|
|
* directly by rxe_run_task or indirectly if rxe_sched_task
|
|
* schedules the task. They must call __reserve_if_idle to
|
|
* move the task to busy before calling or scheduling.
|
|
* The task can also be moved to drained or invalid
|
|
* by calls to rxe_cleanup_task or rxe_disable_task.
|
|
* In that case tasks which get here are not executed but
|
|
* just flushed. The tasks are designed to look to see if
|
|
* there is work to do and then do part of it before returning
|
|
* here with a return value of zero until all the work
|
|
* has been consumed then it returns a non-zero value.
|
|
* The number of times the task can be run is limited by
|
|
* max iterations so one task cannot hold the cpu forever.
|
|
* If the limit is hit and work remains the task is rescheduled.
|
|
*/
|
|
static void do_task(struct rxe_task *task)
|
|
{
|
|
unsigned int iterations;
|
|
unsigned long flags;
|
|
int resched = 0;
|
|
int cont;
|
|
int ret;
|
|
|
|
WARN_ON(rxe_read(task->qp) <= 0);
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
if (task->state >= TASK_STATE_DRAINED) {
|
|
rxe_put(task->qp);
|
|
task->num_done++;
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
return;
|
|
}
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
|
|
do {
|
|
iterations = RXE_MAX_ITERATIONS;
|
|
cont = 0;
|
|
|
|
do {
|
|
ret = task->func(task->qp);
|
|
} while (ret == 0 && iterations-- > 0);
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
/* we're not done yet but we ran out of iterations.
|
|
* yield the cpu and reschedule the task
|
|
*/
|
|
if (!ret) {
|
|
task->state = TASK_STATE_IDLE;
|
|
resched = 1;
|
|
goto exit;
|
|
}
|
|
|
|
switch (task->state) {
|
|
case TASK_STATE_BUSY:
|
|
task->state = TASK_STATE_IDLE;
|
|
break;
|
|
|
|
/* someone tried to schedule the task while we
|
|
* were running, keep going
|
|
*/
|
|
case TASK_STATE_ARMED:
|
|
task->state = TASK_STATE_BUSY;
|
|
cont = 1;
|
|
break;
|
|
|
|
case TASK_STATE_DRAINING:
|
|
task->state = TASK_STATE_DRAINED;
|
|
break;
|
|
|
|
default:
|
|
WARN_ON(1);
|
|
rxe_dbg_qp(task->qp, "unexpected task state = %d",
|
|
task->state);
|
|
task->state = TASK_STATE_IDLE;
|
|
}
|
|
|
|
exit:
|
|
if (!cont) {
|
|
task->num_done++;
|
|
if (WARN_ON(task->num_done != task->num_sched))
|
|
rxe_dbg_qp(
|
|
task->qp,
|
|
"%ld tasks scheduled, %ld tasks done",
|
|
task->num_sched, task->num_done);
|
|
}
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
} while (cont);
|
|
|
|
task->ret = ret;
|
|
|
|
if (resched)
|
|
rxe_sched_task(task);
|
|
|
|
rxe_put(task->qp);
|
|
}
|
|
|
|
/* wrapper around do_task to fix argument for work queue */
|
|
static void do_work(struct work_struct *work)
|
|
{
|
|
do_task(container_of(work, struct rxe_task, work));
|
|
}
|
|
|
|
int rxe_init_task(struct rxe_task *task, struct rxe_qp *qp,
|
|
int (*func)(struct rxe_qp *))
|
|
{
|
|
WARN_ON(rxe_read(qp) <= 0);
|
|
|
|
task->qp = qp;
|
|
task->func = func;
|
|
task->state = TASK_STATE_IDLE;
|
|
spin_lock_init(&task->lock);
|
|
INIT_WORK(&task->work, do_work);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* rxe_cleanup_task is only called from rxe_do_qp_cleanup in
|
|
* process context. The qp is already completed with no
|
|
* remaining references. Once the queue is drained the
|
|
* task is moved to invalid and returns. The qp cleanup
|
|
* code then calls the task functions directly without
|
|
* using the task struct to drain any late arriving packets
|
|
* or work requests.
|
|
*/
|
|
void rxe_cleanup_task(struct rxe_task *task)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
if (!__is_done(task) && task->state < TASK_STATE_DRAINED) {
|
|
task->state = TASK_STATE_DRAINING;
|
|
} else {
|
|
task->state = TASK_STATE_INVALID;
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
return;
|
|
}
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
|
|
/* now the task cannot be scheduled or run just wait
|
|
* for the previously scheduled tasks to finish.
|
|
*/
|
|
while (!is_done(task))
|
|
cond_resched();
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
task->state = TASK_STATE_INVALID;
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
}
|
|
|
|
/* run the task inline if it is currently idle
|
|
* cannot call do_task holding the lock
|
|
*/
|
|
void rxe_run_task(struct rxe_task *task)
|
|
{
|
|
unsigned long flags;
|
|
bool run;
|
|
|
|
WARN_ON(rxe_read(task->qp) <= 0);
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
run = __reserve_if_idle(task);
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
|
|
if (run)
|
|
do_task(task);
|
|
}
|
|
|
|
/* schedule the task to run later as a work queue entry.
|
|
* the queue_work call can be called holding
|
|
* the lock.
|
|
*/
|
|
void rxe_sched_task(struct rxe_task *task)
|
|
{
|
|
unsigned long flags;
|
|
|
|
WARN_ON(rxe_read(task->qp) <= 0);
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
if (__reserve_if_idle(task))
|
|
queue_work(rxe_wq, &task->work);
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
}
|
|
|
|
/* rxe_disable/enable_task are only called from
|
|
* rxe_modify_qp in process context. Task is moved
|
|
* to the drained state by do_task.
|
|
*/
|
|
void rxe_disable_task(struct rxe_task *task)
|
|
{
|
|
unsigned long flags;
|
|
|
|
WARN_ON(rxe_read(task->qp) <= 0);
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
if (!__is_done(task) && task->state < TASK_STATE_DRAINED) {
|
|
task->state = TASK_STATE_DRAINING;
|
|
} else {
|
|
task->state = TASK_STATE_DRAINED;
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
return;
|
|
}
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
|
|
while (!is_done(task))
|
|
cond_resched();
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
task->state = TASK_STATE_DRAINED;
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
}
|
|
|
|
void rxe_enable_task(struct rxe_task *task)
|
|
{
|
|
unsigned long flags;
|
|
|
|
WARN_ON(rxe_read(task->qp) <= 0);
|
|
|
|
spin_lock_irqsave(&task->lock, flags);
|
|
if (task->state == TASK_STATE_INVALID) {
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
return;
|
|
}
|
|
|
|
task->state = TASK_STATE_IDLE;
|
|
spin_unlock_irqrestore(&task->lock, flags);
|
|
}
|