mirror of
https://github.com/torvalds/linux.git
synced 2024-12-21 02:21:36 +00:00
4892c9b4ad
Add support for MSI message groups for Xen Dom0 using the MAP_PIRQ_TYPE_MULTI_MSI pirq map type. In order to keep track of which pirq is the first one in the group all pirqs in the MSI group except for the first one have the newly introduced PIRQ_MSI_GROUP flag set. This prevents calling PHYSDEVOP_unmap_pirq on them, since the unmap must be done with the first pirq in the group. Signed-off-by: Roger Pau Monné <roger.pau@citrix.com> Signed-off-by: David Vrabel <david.vrabel@citrix.com> Cc: Boris Ostrovsky <boris.ostrovsky@oracle.com>
152 lines
3.2 KiB
C
152 lines
3.2 KiB
C
/*
|
|
* Xen Event Channels (internal header)
|
|
*
|
|
* Copyright (C) 2013 Citrix Systems R&D Ltd.
|
|
*
|
|
* This source code is licensed under the GNU General Public License,
|
|
* Version 2 or later. See the file COPYING for more details.
|
|
*/
|
|
#ifndef __EVENTS_INTERNAL_H__
|
|
#define __EVENTS_INTERNAL_H__
|
|
|
|
/* Interrupt types. */
|
|
enum xen_irq_type {
|
|
IRQT_UNBOUND = 0,
|
|
IRQT_PIRQ,
|
|
IRQT_VIRQ,
|
|
IRQT_IPI,
|
|
IRQT_EVTCHN
|
|
};
|
|
|
|
/*
|
|
* Packed IRQ information:
|
|
* type - enum xen_irq_type
|
|
* event channel - irq->event channel mapping
|
|
* cpu - cpu this event channel is bound to
|
|
* index - type-specific information:
|
|
* PIRQ - vector, with MSB being "needs EIO", or physical IRQ of the HVM
|
|
* guest, or GSI (real passthrough IRQ) of the device.
|
|
* VIRQ - virq number
|
|
* IPI - IPI vector
|
|
* EVTCHN -
|
|
*/
|
|
struct irq_info {
|
|
struct list_head list;
|
|
int refcnt;
|
|
enum xen_irq_type type; /* type */
|
|
unsigned irq;
|
|
unsigned int evtchn; /* event channel */
|
|
unsigned short cpu; /* cpu bound */
|
|
|
|
union {
|
|
unsigned short virq;
|
|
enum ipi_vector ipi;
|
|
struct {
|
|
unsigned short pirq;
|
|
unsigned short gsi;
|
|
unsigned char vector;
|
|
unsigned char flags;
|
|
uint16_t domid;
|
|
} pirq;
|
|
} u;
|
|
};
|
|
|
|
#define PIRQ_NEEDS_EOI (1 << 0)
|
|
#define PIRQ_SHAREABLE (1 << 1)
|
|
#define PIRQ_MSI_GROUP (1 << 2)
|
|
|
|
struct evtchn_ops {
|
|
unsigned (*max_channels)(void);
|
|
unsigned (*nr_channels)(void);
|
|
|
|
int (*setup)(struct irq_info *info);
|
|
void (*bind_to_cpu)(struct irq_info *info, unsigned cpu);
|
|
|
|
void (*clear_pending)(unsigned port);
|
|
void (*set_pending)(unsigned port);
|
|
bool (*is_pending)(unsigned port);
|
|
bool (*test_and_set_mask)(unsigned port);
|
|
void (*mask)(unsigned port);
|
|
void (*unmask)(unsigned port);
|
|
|
|
void (*handle_events)(unsigned cpu);
|
|
void (*resume)(void);
|
|
};
|
|
|
|
extern const struct evtchn_ops *evtchn_ops;
|
|
|
|
extern int **evtchn_to_irq;
|
|
int get_evtchn_to_irq(unsigned int evtchn);
|
|
|
|
struct irq_info *info_for_irq(unsigned irq);
|
|
unsigned cpu_from_irq(unsigned irq);
|
|
unsigned cpu_from_evtchn(unsigned int evtchn);
|
|
|
|
static inline unsigned xen_evtchn_max_channels(void)
|
|
{
|
|
return evtchn_ops->max_channels();
|
|
}
|
|
|
|
/*
|
|
* Do any ABI specific setup for a bound event channel before it can
|
|
* be unmasked and used.
|
|
*/
|
|
static inline int xen_evtchn_port_setup(struct irq_info *info)
|
|
{
|
|
if (evtchn_ops->setup)
|
|
return evtchn_ops->setup(info);
|
|
return 0;
|
|
}
|
|
|
|
static inline void xen_evtchn_port_bind_to_cpu(struct irq_info *info,
|
|
unsigned cpu)
|
|
{
|
|
evtchn_ops->bind_to_cpu(info, cpu);
|
|
}
|
|
|
|
static inline void clear_evtchn(unsigned port)
|
|
{
|
|
evtchn_ops->clear_pending(port);
|
|
}
|
|
|
|
static inline void set_evtchn(unsigned port)
|
|
{
|
|
evtchn_ops->set_pending(port);
|
|
}
|
|
|
|
static inline bool test_evtchn(unsigned port)
|
|
{
|
|
return evtchn_ops->is_pending(port);
|
|
}
|
|
|
|
static inline bool test_and_set_mask(unsigned port)
|
|
{
|
|
return evtchn_ops->test_and_set_mask(port);
|
|
}
|
|
|
|
static inline void mask_evtchn(unsigned port)
|
|
{
|
|
return evtchn_ops->mask(port);
|
|
}
|
|
|
|
static inline void unmask_evtchn(unsigned port)
|
|
{
|
|
return evtchn_ops->unmask(port);
|
|
}
|
|
|
|
static inline void xen_evtchn_handle_events(unsigned cpu)
|
|
{
|
|
return evtchn_ops->handle_events(cpu);
|
|
}
|
|
|
|
static inline void xen_evtchn_resume(void)
|
|
{
|
|
if (evtchn_ops->resume)
|
|
evtchn_ops->resume();
|
|
}
|
|
|
|
void xen_evtchn_2l_init(void);
|
|
int xen_evtchn_fifo_init(void);
|
|
|
|
#endif /* #ifndef __EVENTS_INTERNAL_H__ */
|