mirror of
https://github.com/torvalds/linux.git
synced 2024-11-23 04:31:50 +00:00
5009121214
The xtensa-mx driver only targets a single CPU at a time, even if the notional affinity is wider. Let's inform the core code about this. Signed-off-by: Marc Zyngier <marc.zyngier@arm.com> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: Andrew Lunn <andrew@lunn.ch> Cc: James Hogan <james.hogan@imgtec.com> Cc: Jason Cooper <jason@lakedaemon.net> Cc: Paul Burton <paul.burton@imgtec.com> Cc: Chris Zankel <chris@zankel.net> Cc: Kevin Cernekee <cernekee@gmail.com> Cc: Wei Xu <xuwei5@hisilicon.com> Cc: Max Filippov <jcmvbkbc@gmail.com> Cc: Florian Fainelli <f.fainelli@gmail.com> Cc: Gregory Clement <gregory.clement@free-electrons.com> Cc: Matt Redfearn <matt.redfearn@imgtec.com> Cc: Sebastian Hesselbarth <sebastian.hesselbarth@gmail.com> Link: http://lkml.kernel.org/r/20170818083925.10108-13-marc.zyngier@arm.com
168 lines
4.3 KiB
C
168 lines
4.3 KiB
C
/*
|
|
* Xtensa MX interrupt distributor
|
|
*
|
|
* Copyright (C) 2002 - 2013 Tensilica, Inc.
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/irqdomain.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/irqchip.h>
|
|
#include <linux/of.h>
|
|
|
|
#include <asm/mxregs.h>
|
|
|
|
#define HW_IRQ_IPI_COUNT 2
|
|
#define HW_IRQ_MX_BASE 2
|
|
#define HW_IRQ_EXTERN_BASE 3
|
|
|
|
static DEFINE_PER_CPU(unsigned int, cached_irq_mask);
|
|
|
|
static int xtensa_mx_irq_map(struct irq_domain *d, unsigned int irq,
|
|
irq_hw_number_t hw)
|
|
{
|
|
if (hw < HW_IRQ_IPI_COUNT) {
|
|
struct irq_chip *irq_chip = d->host_data;
|
|
irq_set_chip_and_handler_name(irq, irq_chip,
|
|
handle_percpu_irq, "ipi");
|
|
irq_set_status_flags(irq, IRQ_LEVEL);
|
|
return 0;
|
|
}
|
|
irqd_set_single_target(irq_desc_get_irq_data(irq_to_desc(irq)));
|
|
return xtensa_irq_map(d, irq, hw);
|
|
}
|
|
|
|
/*
|
|
* Device Tree IRQ specifier translation function which works with one or
|
|
* two cell bindings. First cell value maps directly to the hwirq number.
|
|
* Second cell if present specifies whether hwirq number is external (1) or
|
|
* internal (0).
|
|
*/
|
|
static int xtensa_mx_irq_domain_xlate(struct irq_domain *d,
|
|
struct device_node *ctrlr,
|
|
const u32 *intspec, unsigned int intsize,
|
|
unsigned long *out_hwirq, unsigned int *out_type)
|
|
{
|
|
return xtensa_irq_domain_xlate(intspec, intsize,
|
|
intspec[0], intspec[0] + HW_IRQ_EXTERN_BASE,
|
|
out_hwirq, out_type);
|
|
}
|
|
|
|
static const struct irq_domain_ops xtensa_mx_irq_domain_ops = {
|
|
.xlate = xtensa_mx_irq_domain_xlate,
|
|
.map = xtensa_mx_irq_map,
|
|
};
|
|
|
|
void secondary_init_irq(void)
|
|
{
|
|
__this_cpu_write(cached_irq_mask,
|
|
XCHAL_INTTYPE_MASK_EXTERN_EDGE |
|
|
XCHAL_INTTYPE_MASK_EXTERN_LEVEL);
|
|
set_sr(XCHAL_INTTYPE_MASK_EXTERN_EDGE |
|
|
XCHAL_INTTYPE_MASK_EXTERN_LEVEL, intenable);
|
|
}
|
|
|
|
static void xtensa_mx_irq_mask(struct irq_data *d)
|
|
{
|
|
unsigned int mask = 1u << d->hwirq;
|
|
|
|
if (mask & (XCHAL_INTTYPE_MASK_EXTERN_EDGE |
|
|
XCHAL_INTTYPE_MASK_EXTERN_LEVEL)) {
|
|
set_er(1u << (xtensa_get_ext_irq_no(d->hwirq) -
|
|
HW_IRQ_MX_BASE), MIENG);
|
|
} else {
|
|
mask = __this_cpu_read(cached_irq_mask) & ~mask;
|
|
__this_cpu_write(cached_irq_mask, mask);
|
|
set_sr(mask, intenable);
|
|
}
|
|
}
|
|
|
|
static void xtensa_mx_irq_unmask(struct irq_data *d)
|
|
{
|
|
unsigned int mask = 1u << d->hwirq;
|
|
|
|
if (mask & (XCHAL_INTTYPE_MASK_EXTERN_EDGE |
|
|
XCHAL_INTTYPE_MASK_EXTERN_LEVEL)) {
|
|
set_er(1u << (xtensa_get_ext_irq_no(d->hwirq) -
|
|
HW_IRQ_MX_BASE), MIENGSET);
|
|
} else {
|
|
mask |= __this_cpu_read(cached_irq_mask);
|
|
__this_cpu_write(cached_irq_mask, mask);
|
|
set_sr(mask, intenable);
|
|
}
|
|
}
|
|
|
|
static void xtensa_mx_irq_enable(struct irq_data *d)
|
|
{
|
|
variant_irq_enable(d->hwirq);
|
|
xtensa_mx_irq_unmask(d);
|
|
}
|
|
|
|
static void xtensa_mx_irq_disable(struct irq_data *d)
|
|
{
|
|
xtensa_mx_irq_mask(d);
|
|
variant_irq_disable(d->hwirq);
|
|
}
|
|
|
|
static void xtensa_mx_irq_ack(struct irq_data *d)
|
|
{
|
|
set_sr(1 << d->hwirq, intclear);
|
|
}
|
|
|
|
static int xtensa_mx_irq_retrigger(struct irq_data *d)
|
|
{
|
|
set_sr(1 << d->hwirq, intset);
|
|
return 1;
|
|
}
|
|
|
|
static int xtensa_mx_irq_set_affinity(struct irq_data *d,
|
|
const struct cpumask *dest, bool force)
|
|
{
|
|
int cpu = cpumask_any_and(dest, cpu_online_mask);
|
|
unsigned mask = 1u << cpu;
|
|
|
|
set_er(mask, MIROUT(d->hwirq - HW_IRQ_MX_BASE));
|
|
irq_data_update_effective_affinity(d, cpumask_of(cpu));
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static struct irq_chip xtensa_mx_irq_chip = {
|
|
.name = "xtensa-mx",
|
|
.irq_enable = xtensa_mx_irq_enable,
|
|
.irq_disable = xtensa_mx_irq_disable,
|
|
.irq_mask = xtensa_mx_irq_mask,
|
|
.irq_unmask = xtensa_mx_irq_unmask,
|
|
.irq_ack = xtensa_mx_irq_ack,
|
|
.irq_retrigger = xtensa_mx_irq_retrigger,
|
|
.irq_set_affinity = xtensa_mx_irq_set_affinity,
|
|
};
|
|
|
|
int __init xtensa_mx_init_legacy(struct device_node *interrupt_parent)
|
|
{
|
|
struct irq_domain *root_domain =
|
|
irq_domain_add_legacy(NULL, NR_IRQS - 1, 1, 0,
|
|
&xtensa_mx_irq_domain_ops,
|
|
&xtensa_mx_irq_chip);
|
|
irq_set_default_host(root_domain);
|
|
secondary_init_irq();
|
|
return 0;
|
|
}
|
|
|
|
static int __init xtensa_mx_init(struct device_node *np,
|
|
struct device_node *interrupt_parent)
|
|
{
|
|
struct irq_domain *root_domain =
|
|
irq_domain_add_linear(np, NR_IRQS, &xtensa_mx_irq_domain_ops,
|
|
&xtensa_mx_irq_chip);
|
|
irq_set_default_host(root_domain);
|
|
secondary_init_irq();
|
|
return 0;
|
|
}
|
|
IRQCHIP_DECLARE(xtensa_mx_irq_chip, "cdns,xtensa-mx", xtensa_mx_init);
|