net/mlx5: Refcount mlx5_irq with integer

Currently, all access to mlx5 IRQs are done undere a lock. Hance, there
isn't a reason to have kref in struct mlx5_irq.
Switch it to integer.

Signed-off-by: Shay Drory <shayd@nvidia.com>
Reviewed-by: Parav Pandit <parav@nvidia.com>
Signed-off-by: Saeed Mahameed <saeedm@nvidia.com>
This commit is contained in:
Shay Drory 2021-06-22 14:20:16 +03:00 committed by Saeed Mahameed
parent 68fefb7089
commit 2d0b41a376

View File

@ -32,7 +32,7 @@ struct mlx5_irq {
cpumask_var_t mask; cpumask_var_t mask;
char name[MLX5_MAX_IRQ_NAME]; char name[MLX5_MAX_IRQ_NAME];
struct mlx5_irq_pool *pool; struct mlx5_irq_pool *pool;
struct kref kref; int refcount;
u32 index; u32 index;
int irqn; int irqn;
}; };
@ -138,9 +138,8 @@ out:
return ret; return ret;
} }
static void irq_release(struct kref *kref) static void irq_release(struct mlx5_irq *irq)
{ {
struct mlx5_irq *irq = container_of(kref, struct mlx5_irq, kref);
struct mlx5_irq_pool *pool = irq->pool; struct mlx5_irq_pool *pool = irq->pool;
xa_erase(&pool->irqs, irq->index); xa_erase(&pool->irqs, irq->index);
@ -159,10 +158,31 @@ static void irq_put(struct mlx5_irq *irq)
struct mlx5_irq_pool *pool = irq->pool; struct mlx5_irq_pool *pool = irq->pool;
mutex_lock(&pool->lock); mutex_lock(&pool->lock);
kref_put(&irq->kref, irq_release); irq->refcount--;
if (!irq->refcount)
irq_release(irq);
mutex_unlock(&pool->lock); mutex_unlock(&pool->lock);
} }
static int irq_get_locked(struct mlx5_irq *irq)
{
lockdep_assert_held(&irq->pool->lock);
if (WARN_ON_ONCE(!irq->refcount))
return 0;
irq->refcount++;
return 1;
}
static int irq_get(struct mlx5_irq *irq)
{
int err;
mutex_lock(&irq->pool->lock);
err = irq_get_locked(irq);
mutex_unlock(&irq->pool->lock);
return err;
}
static irqreturn_t irq_int_handler(int irq, void *nh) static irqreturn_t irq_int_handler(int irq, void *nh)
{ {
atomic_notifier_call_chain(nh, 0, NULL); atomic_notifier_call_chain(nh, 0, NULL);
@ -214,7 +234,7 @@ static struct mlx5_irq *irq_request(struct mlx5_irq_pool *pool, int i)
err = -ENOMEM; err = -ENOMEM;
goto err_cpumask; goto err_cpumask;
} }
kref_init(&irq->kref); irq->refcount = 1;
irq->index = i; irq->index = i;
err = xa_err(xa_store(&pool->irqs, irq->index, irq, GFP_KERNEL)); err = xa_err(xa_store(&pool->irqs, irq->index, irq, GFP_KERNEL));
if (err) { if (err) {
@ -235,18 +255,18 @@ err_req_irq:
int mlx5_irq_attach_nb(struct mlx5_irq *irq, struct notifier_block *nb) int mlx5_irq_attach_nb(struct mlx5_irq *irq, struct notifier_block *nb)
{ {
int err; int ret;
err = kref_get_unless_zero(&irq->kref); ret = irq_get(irq);
if (WARN_ON_ONCE(!err)) if (!ret)
/* Something very bad happens here, we are enabling EQ /* Something very bad happens here, we are enabling EQ
* on non-existing IRQ. * on non-existing IRQ.
*/ */
return -ENOENT; return -ENOENT;
err = atomic_notifier_chain_register(&irq->nh, nb); ret = atomic_notifier_chain_register(&irq->nh, nb);
if (err) if (ret)
irq_put(irq); irq_put(irq);
return err; return ret;
} }
int mlx5_irq_detach_nb(struct mlx5_irq *irq, struct notifier_block *nb) int mlx5_irq_detach_nb(struct mlx5_irq *irq, struct notifier_block *nb)
@ -301,10 +321,9 @@ static struct mlx5_irq *irq_pool_find_least_loaded(struct mlx5_irq_pool *pool,
xa_for_each_range(&pool->irqs, index, iter, start, end) { xa_for_each_range(&pool->irqs, index, iter, start, end) {
if (!cpumask_equal(iter->mask, affinity)) if (!cpumask_equal(iter->mask, affinity))
continue; continue;
if (kref_read(&iter->kref) < pool->min_threshold) if (iter->refcount < pool->min_threshold)
return iter; return iter;
if (!irq || kref_read(&iter->kref) < if (!irq || iter->refcount < irq->refcount)
kref_read(&irq->kref))
irq = iter; irq = iter;
} }
return irq; return irq;
@ -319,7 +338,7 @@ static struct mlx5_irq *irq_pool_request_affinity(struct mlx5_irq_pool *pool,
mutex_lock(&pool->lock); mutex_lock(&pool->lock);
least_loaded_irq = irq_pool_find_least_loaded(pool, affinity); least_loaded_irq = irq_pool_find_least_loaded(pool, affinity);
if (least_loaded_irq && if (least_loaded_irq &&
kref_read(&least_loaded_irq->kref) < pool->min_threshold) least_loaded_irq->refcount < pool->min_threshold)
goto out; goto out;
new_irq = irq_pool_create_irq(pool, affinity); new_irq = irq_pool_create_irq(pool, affinity);
if (IS_ERR(new_irq)) { if (IS_ERR(new_irq)) {
@ -337,11 +356,11 @@ static struct mlx5_irq *irq_pool_request_affinity(struct mlx5_irq_pool *pool,
least_loaded_irq = new_irq; least_loaded_irq = new_irq;
goto unlock; goto unlock;
out: out:
kref_get(&least_loaded_irq->kref); irq_get_locked(least_loaded_irq);
if (kref_read(&least_loaded_irq->kref) > pool->max_threshold) if (least_loaded_irq->refcount > pool->max_threshold)
mlx5_core_dbg(pool->dev, "IRQ %u overloaded, pool_name: %s, %u EQs on this irq\n", mlx5_core_dbg(pool->dev, "IRQ %u overloaded, pool_name: %s, %u EQs on this irq\n",
least_loaded_irq->irqn, pool->name, least_loaded_irq->irqn, pool->name,
kref_read(&least_loaded_irq->kref) / MLX5_EQ_REFS_PER_IRQ); least_loaded_irq->refcount / MLX5_EQ_REFS_PER_IRQ);
unlock: unlock:
mutex_unlock(&pool->lock); mutex_unlock(&pool->lock);
return least_loaded_irq; return least_loaded_irq;
@ -357,7 +376,7 @@ irq_pool_request_vector(struct mlx5_irq_pool *pool, int vecidx,
mutex_lock(&pool->lock); mutex_lock(&pool->lock);
irq = xa_load(&pool->irqs, vecidx); irq = xa_load(&pool->irqs, vecidx);
if (irq) { if (irq) {
kref_get(&irq->kref); irq_get_locked(irq);
goto unlock; goto unlock;
} }
irq = irq_request(pool, vecidx); irq = irq_request(pool, vecidx);
@ -424,7 +443,7 @@ out:
return irq; return irq;
mlx5_core_dbg(dev, "irq %u mapped to cpu %*pbl, %u EQs on this irq\n", mlx5_core_dbg(dev, "irq %u mapped to cpu %*pbl, %u EQs on this irq\n",
irq->irqn, cpumask_pr_args(affinity), irq->irqn, cpumask_pr_args(affinity),
kref_read(&irq->kref) / MLX5_EQ_REFS_PER_IRQ); irq->refcount / MLX5_EQ_REFS_PER_IRQ);
return irq; return irq;
} }
@ -456,8 +475,12 @@ static void irq_pool_free(struct mlx5_irq_pool *pool)
struct mlx5_irq *irq; struct mlx5_irq *irq;
unsigned long index; unsigned long index;
/* There are cases in which we are destrying the irq_table before
* freeing all the IRQs, fast teardown for example. Hence, free the irqs
* which might not have been freed.
*/
xa_for_each(&pool->irqs, index, irq) xa_for_each(&pool->irqs, index, irq)
irq_release(&irq->kref); irq_release(irq);
xa_destroy(&pool->irqs); xa_destroy(&pool->irqs);
kvfree(pool); kvfree(pool);
} }