2018-03-14 21:15:19 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2015-09-24 09:32:13 +00:00
|
|
|
/*
|
|
|
|
* Generic cpu hotunplug interrupt migration code copied from the
|
|
|
|
* arch/arm implementation
|
|
|
|
*
|
|
|
|
* Copyright (C) Russell King
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/ratelimit.h>
|
|
|
|
#include <linux/irq.h>
|
2020-01-20 09:16:25 +00:00
|
|
|
#include <linux/sched/isolation.h>
|
2015-09-24 09:32:13 +00:00
|
|
|
|
|
|
|
#include "internals.h"
|
|
|
|
|
2017-06-19 23:37:39 +00:00
|
|
|
/* For !GENERIC_IRQ_EFFECTIVE_AFF_MASK this looks at general affinity mask */
|
|
|
|
static inline bool irq_needs_fixup(struct irq_data *d)
|
|
|
|
{
|
|
|
|
const struct cpumask *m = irq_data_get_effective_affinity_mask(d);
|
2017-10-09 10:47:24 +00:00
|
|
|
unsigned int cpu = smp_processor_id();
|
2017-06-19 23:37:39 +00:00
|
|
|
|
2017-10-09 10:47:24 +00:00
|
|
|
#ifdef CONFIG_GENERIC_IRQ_EFFECTIVE_AFF_MASK
|
|
|
|
/*
|
|
|
|
* The cpumask_empty() check is a workaround for interrupt chips,
|
|
|
|
* which do not implement effective affinity, but the architecture has
|
|
|
|
* enabled the config switch. Use the general affinity mask instead.
|
|
|
|
*/
|
|
|
|
if (cpumask_empty(m))
|
|
|
|
m = irq_data_get_affinity_mask(d);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Sanity check. If the mask is not empty when excluding the outgoing
|
|
|
|
* CPU then it must contain at least one online CPU. The outgoing CPU
|
|
|
|
* has been removed from the online mask already.
|
|
|
|
*/
|
|
|
|
if (cpumask_any_but(m, cpu) < nr_cpu_ids &&
|
|
|
|
cpumask_any_and(m, cpu_online_mask) >= nr_cpu_ids) {
|
|
|
|
/*
|
|
|
|
* If this happens then there was a missed IRQ fixup at some
|
|
|
|
* point. Warn about it and enforce fixup.
|
|
|
|
*/
|
|
|
|
pr_warn("Eff. affinity %*pbl of IRQ %u contains only offline CPUs after offlining CPU %u\n",
|
|
|
|
cpumask_pr_args(m), d->irq, cpu);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
return cpumask_test_cpu(cpu, m);
|
2017-06-19 23:37:39 +00:00
|
|
|
}
|
|
|
|
|
2015-09-24 09:32:13 +00:00
|
|
|
static bool migrate_one_irq(struct irq_desc *desc)
|
|
|
|
{
|
|
|
|
struct irq_data *d = irq_desc_get_irq_data(desc);
|
2017-06-19 23:37:27 +00:00
|
|
|
struct irq_chip *chip = irq_data_get_irq_chip(d);
|
2017-06-19 23:37:30 +00:00
|
|
|
bool maskchip = !irq_can_move_pcntxt(d) && !irqd_irq_masked(d);
|
2017-06-19 23:37:29 +00:00
|
|
|
const struct cpumask *affinity;
|
2017-06-19 23:37:27 +00:00
|
|
|
bool brokeaff = false;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* IRQ chip might be already torn down, but the irq descriptor is
|
|
|
|
* still in the radix tree. Also if the chip has no affinity setter,
|
|
|
|
* nothing can be done here.
|
|
|
|
*/
|
|
|
|
if (!chip || !chip->irq_set_affinity) {
|
|
|
|
pr_debug("IRQ %u: Unable to migrate away\n", d->irq);
|
|
|
|
return false;
|
|
|
|
}
|
2015-09-24 09:32:13 +00:00
|
|
|
|
genirq/cpuhotplug, x86/vector: Prevent vector leak during CPU offline
The absence of IRQD_MOVE_PCNTXT prevents immediate effectiveness of
interrupt affinity reconfiguration via procfs. Instead, the change is
deferred until the next instance of the interrupt being triggered on the
original CPU.
When the interrupt next triggers on the original CPU, the new affinity is
enforced within __irq_move_irq(). A vector is allocated from the new CPU,
but the old vector on the original CPU remains and is not immediately
reclaimed. Instead, apicd->move_in_progress is flagged, and the reclaiming
process is delayed until the next trigger of the interrupt on the new CPU.
Upon the subsequent triggering of the interrupt on the new CPU,
irq_complete_move() adds a task to the old CPU's vector_cleanup list if it
remains online. Subsequently, the timer on the old CPU iterates over its
vector_cleanup list, reclaiming old vectors.
However, a rare scenario arises if the old CPU is outgoing before the
interrupt triggers again on the new CPU.
In that case irq_force_complete_move() is not invoked on the outgoing CPU
to reclaim the old apicd->prev_vector because the interrupt isn't currently
affine to the outgoing CPU, and irq_needs_fixup() returns false. Even
though __vector_schedule_cleanup() is later called on the new CPU, it
doesn't reclaim apicd->prev_vector; instead, it simply resets both
apicd->move_in_progress and apicd->prev_vector to 0.
As a result, the vector remains unreclaimed in vector_matrix, leading to a
CPU vector leak.
To address this issue, move the invocation of irq_force_complete_move()
before the irq_needs_fixup() call to reclaim apicd->prev_vector, if the
interrupt is currently or used to be affine to the outgoing CPU.
Additionally, reclaim the vector in __vector_schedule_cleanup() as well,
following a warning message, although theoretically it should never see
apicd->move_in_progress with apicd->prev_cpu pointing to an offline CPU.
Fixes: f0383c24b485 ("genirq/cpuhotplug: Add support for cleaning up move in progress")
Signed-off-by: Dongli Zhang <dongli.zhang@oracle.com>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Cc: stable@vger.kernel.org
Link: https://lore.kernel.org/r/20240522220218.162423-1-dongli.zhang@oracle.com
2024-05-22 22:02:18 +00:00
|
|
|
/*
|
|
|
|
* Complete an eventually pending irq move cleanup. If this
|
|
|
|
* interrupt was moved in hard irq context, then the vectors need
|
|
|
|
* to be cleaned up. It can't wait until this interrupt actually
|
|
|
|
* happens and this CPU was involved.
|
|
|
|
*/
|
|
|
|
irq_force_complete_move(desc);
|
|
|
|
|
2015-09-24 09:32:13 +00:00
|
|
|
/*
|
2017-06-19 23:37:28 +00:00
|
|
|
* No move required, if:
|
|
|
|
* - Interrupt is per cpu
|
|
|
|
* - Interrupt is not started
|
|
|
|
* - Affinity mask does not include this CPU.
|
|
|
|
*
|
|
|
|
* Note: Do not check desc->action as this might be a chained
|
|
|
|
* interrupt.
|
2015-09-24 09:32:13 +00:00
|
|
|
*/
|
2017-06-19 23:37:39 +00:00
|
|
|
if (irqd_is_per_cpu(d) || !irqd_is_started(d) || !irq_needs_fixup(d)) {
|
2017-06-19 23:37:29 +00:00
|
|
|
/*
|
|
|
|
* If an irq move is pending, abort it if the dying CPU is
|
|
|
|
* the sole target.
|
|
|
|
*/
|
|
|
|
irq_fixup_move_pending(desc, false);
|
2015-09-24 09:32:13 +00:00
|
|
|
return false;
|
2017-06-19 23:37:29 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If there is a setaffinity pending, then try to reuse the pending
|
|
|
|
* mask, so the last change of the affinity does not get lost. If
|
|
|
|
* there is no move pending or the pending mask does not contain
|
|
|
|
* any online CPU, use the current affinity mask.
|
|
|
|
*/
|
|
|
|
if (irq_fixup_move_pending(desc, true))
|
|
|
|
affinity = irq_desc_get_pending_mask(desc);
|
2017-06-19 23:37:39 +00:00
|
|
|
else
|
|
|
|
affinity = irq_data_get_affinity_mask(d);
|
2015-09-24 09:32:13 +00:00
|
|
|
|
2017-06-19 23:37:30 +00:00
|
|
|
/* Mask the chip for interrupts which cannot move in process context */
|
|
|
|
if (maskchip && chip->irq_mask)
|
|
|
|
chip->irq_mask(d);
|
|
|
|
|
2015-09-24 09:32:13 +00:00
|
|
|
if (cpumask_any_and(affinity, cpu_online_mask) >= nr_cpu_ids) {
|
2017-06-19 23:37:51 +00:00
|
|
|
/*
|
|
|
|
* If the interrupt is managed, then shut it down and leave
|
|
|
|
* the affinity untouched.
|
|
|
|
*/
|
|
|
|
if (irqd_affinity_is_managed(d)) {
|
|
|
|
irqd_set_managed_shutdown(d);
|
2019-06-28 11:11:49 +00:00
|
|
|
irq_shutdown_and_deactivate(desc);
|
2017-06-19 23:37:51 +00:00
|
|
|
return false;
|
|
|
|
}
|
2015-09-24 09:32:13 +00:00
|
|
|
affinity = cpu_online_mask;
|
2017-06-19 23:37:27 +00:00
|
|
|
brokeaff = true;
|
2015-09-24 09:32:13 +00:00
|
|
|
}
|
2017-07-27 10:21:11 +00:00
|
|
|
/*
|
|
|
|
* Do not set the force argument of irq_do_set_affinity() as this
|
|
|
|
* disables the masking of offline CPUs from the supplied affinity
|
|
|
|
* mask and therefore might keep/reassign the irq to the outgoing
|
|
|
|
* CPU.
|
|
|
|
*/
|
|
|
|
err = irq_do_set_affinity(d, affinity, false);
|
2024-04-23 07:34:13 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If there are online CPUs in the affinity mask, but they have no
|
|
|
|
* vectors left to make the migration work, try to break the
|
|
|
|
* affinity by migrating to any online CPU.
|
|
|
|
*/
|
|
|
|
if (err == -ENOSPC && !irqd_affinity_is_managed(d) && affinity != cpu_online_mask) {
|
|
|
|
pr_debug("IRQ%u: set affinity failed for %*pbl, re-try with online CPUs\n",
|
|
|
|
d->irq, cpumask_pr_args(affinity));
|
|
|
|
|
|
|
|
affinity = cpu_online_mask;
|
|
|
|
brokeaff = true;
|
|
|
|
|
|
|
|
err = irq_do_set_affinity(d, affinity, false);
|
|
|
|
}
|
|
|
|
|
2017-06-19 23:37:27 +00:00
|
|
|
if (err) {
|
|
|
|
pr_warn_ratelimited("IRQ%u: set affinity failed(%d).\n",
|
|
|
|
d->irq, err);
|
2017-06-19 23:37:30 +00:00
|
|
|
brokeaff = false;
|
2015-09-24 09:32:13 +00:00
|
|
|
}
|
2017-06-19 23:37:30 +00:00
|
|
|
|
|
|
|
if (maskchip && chip->irq_unmask)
|
|
|
|
chip->irq_unmask(d);
|
|
|
|
|
2017-06-19 23:37:27 +00:00
|
|
|
return brokeaff;
|
2015-09-24 09:32:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* irq_migrate_all_off_this_cpu - Migrate irqs away from offline cpu
|
|
|
|
*
|
|
|
|
* The current CPU has been marked offline. Migrate IRQs off this CPU.
|
|
|
|
* If the affinity settings do not allow other CPUs, force them onto any
|
|
|
|
* available CPU.
|
|
|
|
*
|
|
|
|
* Note: we must iterate over all IRQs, whether they have an attached
|
|
|
|
* action structure or not, as we need to get chained interrupts too.
|
|
|
|
*/
|
|
|
|
void irq_migrate_all_off_this_cpu(void)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc;
|
2017-06-19 23:37:25 +00:00
|
|
|
unsigned int irq;
|
2015-09-24 09:32:13 +00:00
|
|
|
|
|
|
|
for_each_active_irq(irq) {
|
|
|
|
bool affinity_broken;
|
|
|
|
|
|
|
|
desc = irq_to_desc(irq);
|
|
|
|
raw_spin_lock(&desc->lock);
|
|
|
|
affinity_broken = migrate_one_irq(desc);
|
|
|
|
raw_spin_unlock(&desc->lock);
|
|
|
|
|
2017-06-19 23:37:25 +00:00
|
|
|
if (affinity_broken) {
|
2021-08-16 13:48:17 +00:00
|
|
|
pr_debug_ratelimited("IRQ %u: no longer affine to CPU%u\n",
|
2015-09-24 09:32:13 +00:00
|
|
|
irq, smp_processor_id());
|
2017-06-19 23:37:25 +00:00
|
|
|
}
|
2015-09-24 09:32:13 +00:00
|
|
|
}
|
|
|
|
}
|
2017-06-19 23:37:51 +00:00
|
|
|
|
2020-01-20 09:16:25 +00:00
|
|
|
static bool hk_should_isolate(struct irq_data *data, unsigned int cpu)
|
|
|
|
{
|
|
|
|
const struct cpumask *hk_mask;
|
|
|
|
|
2022-02-07 15:59:06 +00:00
|
|
|
if (!housekeeping_enabled(HK_TYPE_MANAGED_IRQ))
|
2020-01-20 09:16:25 +00:00
|
|
|
return false;
|
|
|
|
|
2022-02-07 15:59:06 +00:00
|
|
|
hk_mask = housekeeping_cpumask(HK_TYPE_MANAGED_IRQ);
|
2020-01-20 09:16:25 +00:00
|
|
|
if (cpumask_subset(irq_data_get_effective_affinity_mask(data), hk_mask))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return cpumask_test_cpu(cpu, hk_mask);
|
|
|
|
}
|
|
|
|
|
2017-06-19 23:37:51 +00:00
|
|
|
static void irq_restore_affinity_of_irq(struct irq_desc *desc, unsigned int cpu)
|
|
|
|
{
|
|
|
|
struct irq_data *data = irq_desc_get_irq_data(desc);
|
|
|
|
const struct cpumask *affinity = irq_data_get_affinity_mask(data);
|
|
|
|
|
|
|
|
if (!irqd_affinity_is_managed(data) || !desc->action ||
|
|
|
|
!irq_data_get_irq_chip(data) || !cpumask_test_cpu(cpu, affinity))
|
|
|
|
return;
|
|
|
|
|
2024-04-24 09:03:41 +00:00
|
|
|
/*
|
|
|
|
* Don't restore suspended interrupts here when a system comes back
|
|
|
|
* from S3. They are reenabled via resume_device_irqs().
|
|
|
|
*/
|
|
|
|
if (desc->istate & IRQS_SUSPENDED)
|
2017-06-19 23:37:53 +00:00
|
|
|
return;
|
2024-04-24 09:03:41 +00:00
|
|
|
|
|
|
|
if (irqd_is_managed_and_shutdown(data))
|
|
|
|
irq_startup(desc, IRQ_RESEND, IRQ_START_COND);
|
2017-06-19 23:37:53 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If the interrupt can only be directed to a single target
|
|
|
|
* CPU then it is already assigned to a CPU in the affinity
|
2020-01-20 09:16:25 +00:00
|
|
|
* mask. No point in trying to move it around unless the
|
|
|
|
* isolation mechanism requests to move it to an upcoming
|
|
|
|
* housekeeping CPU.
|
2017-06-19 23:37:53 +00:00
|
|
|
*/
|
2020-01-20 09:16:25 +00:00
|
|
|
if (!irqd_is_single_target(data) || hk_should_isolate(data, cpu))
|
2017-06-19 23:37:51 +00:00
|
|
|
irq_set_affinity_locked(data, affinity, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* irq_affinity_online_cpu - Restore affinity for managed interrupts
|
|
|
|
* @cpu: Upcoming CPU for which interrupts should be restored
|
|
|
|
*/
|
|
|
|
int irq_affinity_online_cpu(unsigned int cpu)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc;
|
|
|
|
unsigned int irq;
|
|
|
|
|
|
|
|
irq_lock_sparse();
|
|
|
|
for_each_active_irq(irq) {
|
|
|
|
desc = irq_to_desc(irq);
|
|
|
|
raw_spin_lock_irq(&desc->lock);
|
|
|
|
irq_restore_affinity_of_irq(desc, cpu);
|
|
|
|
raw_spin_unlock_irq(&desc->lock);
|
|
|
|
}
|
|
|
|
irq_unlock_sparse();
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|