28323add09
Sponsored by: Dell EMC Isilon
1668 lines
43 KiB
C
1668 lines
43 KiB
C
/******************************************************************************
|
|
* xen_intr.c
|
|
*
|
|
* Xen event and interrupt services for x86 HVM guests.
|
|
*
|
|
* Copyright (c) 2002-2005, K A Fraser
|
|
* Copyright (c) 2005, Intel Corporation <xiaofeng.ling@intel.com>
|
|
* Copyright (c) 2012, Spectra Logic Corporation
|
|
*
|
|
* This file may be distributed separately from the Linux kernel, or
|
|
* incorporated into other software packages, subject to the following license:
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this source file (the "Software"), to deal in the Software without
|
|
* restriction, including without limitation the rights to use, copy, modify,
|
|
* merge, publish, distribute, sublicense, and/or sell copies of the Software,
|
|
* and to permit persons to whom the Software is furnished to do so, subject to
|
|
* the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include "opt_ddb.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/bus.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/limits.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/mutex.h>
|
|
#include <sys/interrupt.h>
|
|
#include <sys/pcpu.h>
|
|
#include <sys/smp.h>
|
|
#include <sys/refcount.h>
|
|
|
|
#include <vm/vm.h>
|
|
#include <vm/pmap.h>
|
|
|
|
#include <machine/intr_machdep.h>
|
|
#include <x86/apicvar.h>
|
|
#include <x86/apicreg.h>
|
|
#include <machine/smp.h>
|
|
#include <machine/stdarg.h>
|
|
|
|
#include <machine/xen/synch_bitops.h>
|
|
#include <machine/xen/xen-os.h>
|
|
|
|
#include <xen/hypervisor.h>
|
|
#include <xen/xen_intr.h>
|
|
#include <xen/evtchn/evtchnvar.h>
|
|
|
|
#include <dev/xen/xenpci/xenpcivar.h>
|
|
#include <dev/pci/pcivar.h>
|
|
|
|
#ifdef DDB
|
|
#include <ddb/ddb.h>
|
|
#endif
|
|
|
|
static MALLOC_DEFINE(M_XENINTR, "xen_intr", "Xen Interrupt Services");
|
|
|
|
/**
|
|
* Per-cpu event channel processing state.
|
|
*/
|
|
struct xen_intr_pcpu_data {
|
|
/**
|
|
* The last event channel bitmap section (level one bit) processed.
|
|
* This is used to ensure we scan all ports before
|
|
* servicing an already servied port again.
|
|
*/
|
|
u_int last_processed_l1i;
|
|
|
|
/**
|
|
* The last event channel processed within the event channel
|
|
* bitmap being scanned.
|
|
*/
|
|
u_int last_processed_l2i;
|
|
|
|
/** Pointer to this CPU's interrupt statistic counter. */
|
|
u_long *evtchn_intrcnt;
|
|
|
|
/**
|
|
* A bitmap of ports that can be serviced from this CPU.
|
|
* A set bit means interrupt handling is enabled.
|
|
*/
|
|
u_long evtchn_enabled[sizeof(u_long) * 8];
|
|
};
|
|
|
|
/*
|
|
* Start the scan at port 0 by initializing the last scanned
|
|
* location as the highest numbered event channel port.
|
|
*/
|
|
static DPCPU_DEFINE(struct xen_intr_pcpu_data, xen_intr_pcpu) = {
|
|
.last_processed_l1i = LONG_BIT - 1,
|
|
.last_processed_l2i = LONG_BIT - 1
|
|
};
|
|
|
|
DPCPU_DECLARE(struct vcpu_info *, vcpu_info);
|
|
|
|
#define XEN_EEXIST 17 /* Xen "already exists" error */
|
|
#define XEN_ALLOCATE_VECTOR 0 /* Allocate a vector for this event channel */
|
|
#define XEN_INVALID_EVTCHN 0 /* Invalid event channel */
|
|
|
|
#define is_valid_evtchn(x) ((x) != XEN_INVALID_EVTCHN)
|
|
|
|
struct xenisrc {
|
|
struct intsrc xi_intsrc;
|
|
enum evtchn_type xi_type;
|
|
int xi_cpu; /* VCPU for delivery. */
|
|
int xi_vector; /* Global isrc vector number. */
|
|
evtchn_port_t xi_port;
|
|
int xi_pirq;
|
|
int xi_virq;
|
|
void *xi_cookie;
|
|
u_int xi_close:1; /* close on unbind? */
|
|
u_int xi_activehi:1;
|
|
u_int xi_edgetrigger:1;
|
|
u_int xi_masked:1;
|
|
volatile u_int xi_refcount;
|
|
};
|
|
|
|
static void xen_intr_suspend(struct pic *);
|
|
static void xen_intr_resume(struct pic *, bool suspend_cancelled);
|
|
static void xen_intr_enable_source(struct intsrc *isrc);
|
|
static void xen_intr_disable_source(struct intsrc *isrc, int eoi);
|
|
static void xen_intr_eoi_source(struct intsrc *isrc);
|
|
static void xen_intr_enable_intr(struct intsrc *isrc);
|
|
static void xen_intr_disable_intr(struct intsrc *isrc);
|
|
static int xen_intr_vector(struct intsrc *isrc);
|
|
static int xen_intr_source_pending(struct intsrc *isrc);
|
|
static int xen_intr_config_intr(struct intsrc *isrc,
|
|
enum intr_trigger trig, enum intr_polarity pol);
|
|
static int xen_intr_assign_cpu(struct intsrc *isrc, u_int apic_id);
|
|
|
|
static void xen_intr_pirq_enable_source(struct intsrc *isrc);
|
|
static void xen_intr_pirq_disable_source(struct intsrc *isrc, int eoi);
|
|
static void xen_intr_pirq_eoi_source(struct intsrc *isrc);
|
|
static void xen_intr_pirq_enable_intr(struct intsrc *isrc);
|
|
static void xen_intr_pirq_disable_intr(struct intsrc *isrc);
|
|
static int xen_intr_pirq_config_intr(struct intsrc *isrc,
|
|
enum intr_trigger trig, enum intr_polarity pol);
|
|
|
|
/**
|
|
* PIC interface for all event channel port types except physical IRQs.
|
|
*/
|
|
struct pic xen_intr_pic = {
|
|
.pic_enable_source = xen_intr_enable_source,
|
|
.pic_disable_source = xen_intr_disable_source,
|
|
.pic_eoi_source = xen_intr_eoi_source,
|
|
.pic_enable_intr = xen_intr_enable_intr,
|
|
.pic_disable_intr = xen_intr_disable_intr,
|
|
.pic_vector = xen_intr_vector,
|
|
.pic_source_pending = xen_intr_source_pending,
|
|
.pic_suspend = xen_intr_suspend,
|
|
.pic_resume = xen_intr_resume,
|
|
.pic_config_intr = xen_intr_config_intr,
|
|
.pic_assign_cpu = xen_intr_assign_cpu
|
|
};
|
|
|
|
/**
|
|
* PIC interface for all event channel representing
|
|
* physical interrupt sources.
|
|
*/
|
|
struct pic xen_intr_pirq_pic = {
|
|
.pic_enable_source = xen_intr_pirq_enable_source,
|
|
.pic_disable_source = xen_intr_pirq_disable_source,
|
|
.pic_eoi_source = xen_intr_pirq_eoi_source,
|
|
.pic_enable_intr = xen_intr_pirq_enable_intr,
|
|
.pic_disable_intr = xen_intr_pirq_disable_intr,
|
|
.pic_vector = xen_intr_vector,
|
|
.pic_source_pending = xen_intr_source_pending,
|
|
.pic_config_intr = xen_intr_pirq_config_intr,
|
|
.pic_assign_cpu = xen_intr_assign_cpu
|
|
};
|
|
|
|
static struct mtx xen_intr_isrc_lock;
|
|
static int xen_intr_auto_vector_count;
|
|
static struct xenisrc *xen_intr_port_to_isrc[NR_EVENT_CHANNELS];
|
|
static u_long *xen_intr_pirq_eoi_map;
|
|
static boolean_t xen_intr_pirq_eoi_map_enabled;
|
|
|
|
/*------------------------- Private Functions --------------------------------*/
|
|
/**
|
|
* Disable signal delivery for an event channel port on the
|
|
* specified CPU.
|
|
*
|
|
* \param port The event channel port to mask.
|
|
*
|
|
* This API is used to manage the port<=>CPU binding of event
|
|
* channel handlers.
|
|
*
|
|
* \note This operation does not preclude reception of an event
|
|
* for this event channel on another CPU. To mask the
|
|
* event channel globally, use evtchn_mask().
|
|
*/
|
|
static inline void
|
|
evtchn_cpu_mask_port(u_int cpu, evtchn_port_t port)
|
|
{
|
|
struct xen_intr_pcpu_data *pcpu;
|
|
|
|
pcpu = DPCPU_ID_PTR(cpu, xen_intr_pcpu);
|
|
xen_clear_bit(port, pcpu->evtchn_enabled);
|
|
}
|
|
|
|
/**
|
|
* Enable signal delivery for an event channel port on the
|
|
* specified CPU.
|
|
*
|
|
* \param port The event channel port to unmask.
|
|
*
|
|
* This API is used to manage the port<=>CPU binding of event
|
|
* channel handlers.
|
|
*
|
|
* \note This operation does not guarantee that event delivery
|
|
* is enabled for this event channel port. The port must
|
|
* also be globally enabled. See evtchn_unmask().
|
|
*/
|
|
static inline void
|
|
evtchn_cpu_unmask_port(u_int cpu, evtchn_port_t port)
|
|
{
|
|
struct xen_intr_pcpu_data *pcpu;
|
|
|
|
pcpu = DPCPU_ID_PTR(cpu, xen_intr_pcpu);
|
|
xen_set_bit(port, pcpu->evtchn_enabled);
|
|
}
|
|
|
|
/**
|
|
* Allocate and register a per-cpu Xen upcall interrupt counter.
|
|
*
|
|
* \param cpu The cpu for which to register this interrupt count.
|
|
*/
|
|
static void
|
|
xen_intr_intrcnt_add(u_int cpu)
|
|
{
|
|
char buf[MAXCOMLEN + 1];
|
|
struct xen_intr_pcpu_data *pcpu;
|
|
|
|
pcpu = DPCPU_ID_PTR(cpu, xen_intr_pcpu);
|
|
if (pcpu->evtchn_intrcnt != NULL)
|
|
return;
|
|
|
|
snprintf(buf, sizeof(buf), "cpu%d:xen", cpu);
|
|
intrcnt_add(buf, &pcpu->evtchn_intrcnt);
|
|
}
|
|
|
|
/**
|
|
* Search for an already allocated but currently unused Xen interrupt
|
|
* source object.
|
|
*
|
|
* \param type Restrict the search to interrupt sources of the given
|
|
* type.
|
|
*
|
|
* \return A pointer to a free Xen interrupt source object or NULL.
|
|
*/
|
|
static struct xenisrc *
|
|
xen_intr_find_unused_isrc(enum evtchn_type type)
|
|
{
|
|
int isrc_idx;
|
|
|
|
KASSERT(mtx_owned(&xen_intr_isrc_lock), ("Evtchn isrc lock not held"));
|
|
|
|
for (isrc_idx = 0; isrc_idx < xen_intr_auto_vector_count; isrc_idx ++) {
|
|
struct xenisrc *isrc;
|
|
u_int vector;
|
|
|
|
vector = FIRST_EVTCHN_INT + isrc_idx;
|
|
isrc = (struct xenisrc *)intr_lookup_source(vector);
|
|
if (isrc != NULL
|
|
&& isrc->xi_type == EVTCHN_TYPE_UNBOUND) {
|
|
KASSERT(isrc->xi_intsrc.is_handlers == 0,
|
|
("Free evtchn still has handlers"));
|
|
isrc->xi_type = type;
|
|
return (isrc);
|
|
}
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
/**
|
|
* Allocate a Xen interrupt source object.
|
|
*
|
|
* \param type The type of interrupt source to create.
|
|
*
|
|
* \return A pointer to a newly allocated Xen interrupt source
|
|
* object or NULL.
|
|
*/
|
|
static struct xenisrc *
|
|
xen_intr_alloc_isrc(enum evtchn_type type, int vector)
|
|
{
|
|
static int warned;
|
|
struct xenisrc *isrc;
|
|
|
|
KASSERT(mtx_owned(&xen_intr_isrc_lock), ("Evtchn alloc lock not held"));
|
|
|
|
if (xen_intr_auto_vector_count > NR_EVENT_CHANNELS) {
|
|
if (!warned) {
|
|
warned = 1;
|
|
printf("xen_intr_alloc: Event channels exhausted.\n");
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
if (type != EVTCHN_TYPE_PIRQ) {
|
|
vector = FIRST_EVTCHN_INT + xen_intr_auto_vector_count;
|
|
xen_intr_auto_vector_count++;
|
|
}
|
|
|
|
KASSERT((intr_lookup_source(vector) == NULL),
|
|
("Trying to use an already allocated vector"));
|
|
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
isrc = malloc(sizeof(*isrc), M_XENINTR, M_WAITOK | M_ZERO);
|
|
isrc->xi_intsrc.is_pic =
|
|
(type == EVTCHN_TYPE_PIRQ) ? &xen_intr_pirq_pic : &xen_intr_pic;
|
|
isrc->xi_vector = vector;
|
|
isrc->xi_type = type;
|
|
intr_register_source(&isrc->xi_intsrc);
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
|
|
return (isrc);
|
|
}
|
|
|
|
/**
|
|
* Attempt to free an active Xen interrupt source object.
|
|
*
|
|
* \param isrc The interrupt source object to release.
|
|
*
|
|
* \returns EBUSY if the source is still in use, otherwise 0.
|
|
*/
|
|
static int
|
|
xen_intr_release_isrc(struct xenisrc *isrc)
|
|
{
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
KASSERT(isrc->xi_intsrc.is_handlers == 0,
|
|
("Release called, but xenisrc still in use"));
|
|
evtchn_mask_port(isrc->xi_port);
|
|
evtchn_clear_port(isrc->xi_port);
|
|
|
|
/* Rebind port to CPU 0. */
|
|
evtchn_cpu_mask_port(isrc->xi_cpu, isrc->xi_port);
|
|
evtchn_cpu_unmask_port(0, isrc->xi_port);
|
|
|
|
if (isrc->xi_close != 0 && is_valid_evtchn(isrc->xi_port)) {
|
|
struct evtchn_close close = { .port = isrc->xi_port };
|
|
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
|
|
panic("EVTCHNOP_close failed");
|
|
}
|
|
|
|
xen_intr_port_to_isrc[isrc->xi_port] = NULL;
|
|
isrc->xi_cpu = 0;
|
|
isrc->xi_type = EVTCHN_TYPE_UNBOUND;
|
|
isrc->xi_port = 0;
|
|
isrc->xi_cookie = NULL;
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
return (0);
|
|
}
|
|
|
|
/**
|
|
* Associate an interrupt handler with an already allocated local Xen
|
|
* event channel port.
|
|
*
|
|
* \param isrcp The returned Xen interrupt object associated with
|
|
* the specified local port.
|
|
* \param local_port The event channel to bind.
|
|
* \param type The event channel type of local_port.
|
|
* \param intr_owner The device making this bind request.
|
|
* \param filter An interrupt filter handler. Specify NULL
|
|
* to always dispatch to the ithread handler.
|
|
* \param handler An interrupt ithread handler. Optional (can
|
|
* specify NULL) if all necessary event actions
|
|
* are performed by filter.
|
|
* \param arg Argument to present to both filter and handler.
|
|
* \param irqflags Interrupt handler flags. See sys/bus.h.
|
|
* \param handlep Pointer to an opaque handle used to manage this
|
|
* registration.
|
|
*
|
|
* \returns 0 on success, otherwise an errno.
|
|
*/
|
|
static int
|
|
xen_intr_bind_isrc(struct xenisrc **isrcp, evtchn_port_t local_port,
|
|
enum evtchn_type type, device_t intr_owner, driver_filter_t filter,
|
|
driver_intr_t handler, void *arg, enum intr_type flags,
|
|
xen_intr_handle_t *port_handlep)
|
|
{
|
|
struct xenisrc *isrc;
|
|
int error;
|
|
|
|
*isrcp = NULL;
|
|
if (port_handlep == NULL) {
|
|
device_printf(intr_owner,
|
|
"xen_intr_bind_isrc: Bad event handle\n");
|
|
return (EINVAL);
|
|
}
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
isrc = xen_intr_find_unused_isrc(type);
|
|
if (isrc == NULL) {
|
|
isrc = xen_intr_alloc_isrc(type, XEN_ALLOCATE_VECTOR);
|
|
if (isrc == NULL) {
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
return (ENOSPC);
|
|
}
|
|
}
|
|
isrc->xi_port = local_port;
|
|
xen_intr_port_to_isrc[local_port] = isrc;
|
|
refcount_init(&isrc->xi_refcount, 1);
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
|
|
/* Assign the opaque handler (the event channel port) */
|
|
*port_handlep = &isrc->xi_vector;
|
|
|
|
#ifdef SMP
|
|
if (type == EVTCHN_TYPE_PORT) {
|
|
/*
|
|
* By default all interrupts are assigned to vCPU#0
|
|
* unless specified otherwise, so shuffle them to balance
|
|
* the interrupt load.
|
|
*/
|
|
xen_intr_assign_cpu(&isrc->xi_intsrc, intr_next_cpu());
|
|
}
|
|
#endif
|
|
|
|
if (filter == NULL && handler == NULL) {
|
|
/*
|
|
* No filter/handler provided, leave the event channel
|
|
* masked and without a valid handler, the caller is
|
|
* in charge of setting that up.
|
|
*/
|
|
*isrcp = isrc;
|
|
return (0);
|
|
}
|
|
|
|
error = xen_intr_add_handler(intr_owner, filter, handler, arg, flags,
|
|
*port_handlep);
|
|
if (error != 0) {
|
|
xen_intr_release_isrc(isrc);
|
|
return (error);
|
|
}
|
|
*isrcp = isrc;
|
|
return (0);
|
|
}
|
|
|
|
/**
|
|
* Lookup a Xen interrupt source object given an interrupt binding handle.
|
|
*
|
|
* \param handle A handle initialized by a previous call to
|
|
* xen_intr_bind_isrc().
|
|
*
|
|
* \returns A pointer to the Xen interrupt source object associated
|
|
* with the given interrupt handle. NULL if no association
|
|
* currently exists.
|
|
*/
|
|
static struct xenisrc *
|
|
xen_intr_isrc(xen_intr_handle_t handle)
|
|
{
|
|
int vector;
|
|
|
|
if (handle == NULL)
|
|
return (NULL);
|
|
|
|
vector = *(int *)handle;
|
|
KASSERT(vector >= FIRST_EVTCHN_INT &&
|
|
vector < (FIRST_EVTCHN_INT + xen_intr_auto_vector_count),
|
|
("Xen interrupt vector is out of range"));
|
|
|
|
return ((struct xenisrc *)intr_lookup_source(vector));
|
|
}
|
|
|
|
/**
|
|
* Determine the event channel ports at the given section of the
|
|
* event port bitmap which have pending events for the given cpu.
|
|
*
|
|
* \param pcpu The Xen interrupt pcpu data for the cpu being querried.
|
|
* \param sh The Xen shared info area.
|
|
* \param idx The index of the section of the event channel bitmap to
|
|
* inspect.
|
|
*
|
|
* \returns A u_long with bits set for every event channel with pending
|
|
* events.
|
|
*/
|
|
static inline u_long
|
|
xen_intr_active_ports(struct xen_intr_pcpu_data *pcpu, shared_info_t *sh,
|
|
u_int idx)
|
|
{
|
|
|
|
CTASSERT(sizeof(sh->evtchn_mask[0]) == sizeof(sh->evtchn_pending[0]));
|
|
CTASSERT(sizeof(sh->evtchn_mask[0]) == sizeof(pcpu->evtchn_enabled[0]));
|
|
CTASSERT(sizeof(sh->evtchn_mask) == sizeof(sh->evtchn_pending));
|
|
CTASSERT(sizeof(sh->evtchn_mask) == sizeof(pcpu->evtchn_enabled));
|
|
return (sh->evtchn_pending[idx]
|
|
& ~sh->evtchn_mask[idx]
|
|
& pcpu->evtchn_enabled[idx]);
|
|
}
|
|
|
|
/**
|
|
* Interrupt handler for processing all Xen event channel events.
|
|
*
|
|
* \param trap_frame The trap frame context for the current interrupt.
|
|
*/
|
|
void
|
|
xen_intr_handle_upcall(struct trapframe *trap_frame)
|
|
{
|
|
u_int l1i, l2i, port, cpu;
|
|
u_long masked_l1, masked_l2;
|
|
struct xenisrc *isrc;
|
|
shared_info_t *s;
|
|
vcpu_info_t *v;
|
|
struct xen_intr_pcpu_data *pc;
|
|
u_long l1, l2;
|
|
|
|
/*
|
|
* Disable preemption in order to always check and fire events
|
|
* on the right vCPU
|
|
*/
|
|
critical_enter();
|
|
|
|
cpu = PCPU_GET(cpuid);
|
|
pc = DPCPU_PTR(xen_intr_pcpu);
|
|
s = HYPERVISOR_shared_info;
|
|
v = DPCPU_GET(vcpu_info);
|
|
|
|
if (xen_hvm_domain() && !xen_vector_callback_enabled) {
|
|
KASSERT((cpu == 0), ("Fired PCI event callback on wrong CPU"));
|
|
}
|
|
|
|
v->evtchn_upcall_pending = 0;
|
|
|
|
#if 0
|
|
#ifndef CONFIG_X86 /* No need for a barrier -- XCHG is a barrier on x86. */
|
|
/* Clear master flag /before/ clearing selector flag. */
|
|
wmb();
|
|
#endif
|
|
#endif
|
|
|
|
l1 = atomic_readandclear_long(&v->evtchn_pending_sel);
|
|
|
|
l1i = pc->last_processed_l1i;
|
|
l2i = pc->last_processed_l2i;
|
|
(*pc->evtchn_intrcnt)++;
|
|
|
|
while (l1 != 0) {
|
|
|
|
l1i = (l1i + 1) % LONG_BIT;
|
|
masked_l1 = l1 & ((~0UL) << l1i);
|
|
|
|
if (masked_l1 == 0) {
|
|
/*
|
|
* if we masked out all events, wrap around
|
|
* to the beginning.
|
|
*/
|
|
l1i = LONG_BIT - 1;
|
|
l2i = LONG_BIT - 1;
|
|
continue;
|
|
}
|
|
l1i = ffsl(masked_l1) - 1;
|
|
|
|
do {
|
|
l2 = xen_intr_active_ports(pc, s, l1i);
|
|
|
|
l2i = (l2i + 1) % LONG_BIT;
|
|
masked_l2 = l2 & ((~0UL) << l2i);
|
|
|
|
if (masked_l2 == 0) {
|
|
/* if we masked out all events, move on */
|
|
l2i = LONG_BIT - 1;
|
|
break;
|
|
}
|
|
l2i = ffsl(masked_l2) - 1;
|
|
|
|
/* process port */
|
|
port = (l1i * LONG_BIT) + l2i;
|
|
synch_clear_bit(port, &s->evtchn_pending[0]);
|
|
|
|
isrc = xen_intr_port_to_isrc[port];
|
|
if (__predict_false(isrc == NULL))
|
|
continue;
|
|
|
|
/* Make sure we are firing on the right vCPU */
|
|
KASSERT((isrc->xi_cpu == PCPU_GET(cpuid)),
|
|
("Received unexpected event on vCPU#%d, event bound to vCPU#%d",
|
|
PCPU_GET(cpuid), isrc->xi_cpu));
|
|
|
|
intr_execute_handlers(&isrc->xi_intsrc, trap_frame);
|
|
|
|
/*
|
|
* If this is the final port processed,
|
|
* we'll pick up here+1 next time.
|
|
*/
|
|
pc->last_processed_l1i = l1i;
|
|
pc->last_processed_l2i = l2i;
|
|
|
|
} while (l2i != LONG_BIT - 1);
|
|
|
|
l2 = xen_intr_active_ports(pc, s, l1i);
|
|
if (l2 == 0) {
|
|
/*
|
|
* We handled all ports, so we can clear the
|
|
* selector bit.
|
|
*/
|
|
l1 &= ~(1UL << l1i);
|
|
}
|
|
}
|
|
critical_exit();
|
|
}
|
|
|
|
static int
|
|
xen_intr_init(void *dummy __unused)
|
|
{
|
|
shared_info_t *s = HYPERVISOR_shared_info;
|
|
struct xen_intr_pcpu_data *pcpu;
|
|
struct physdev_pirq_eoi_gmfn eoi_gmfn;
|
|
int i, rc;
|
|
|
|
if (!xen_domain())
|
|
return (0);
|
|
|
|
mtx_init(&xen_intr_isrc_lock, "xen-irq-lock", NULL, MTX_DEF);
|
|
|
|
/*
|
|
* Register interrupt count manually as we aren't
|
|
* guaranteed to see a call to xen_intr_assign_cpu()
|
|
* before our first interrupt. Also set the per-cpu
|
|
* mask of CPU#0 to enable all, since by default
|
|
* all event channels are bound to CPU#0.
|
|
*/
|
|
CPU_FOREACH(i) {
|
|
pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
|
|
memset(pcpu->evtchn_enabled, i == 0 ? ~0 : 0,
|
|
sizeof(pcpu->evtchn_enabled));
|
|
xen_intr_intrcnt_add(i);
|
|
}
|
|
|
|
for (i = 0; i < nitems(s->evtchn_mask); i++)
|
|
atomic_store_rel_long(&s->evtchn_mask[i], ~0);
|
|
|
|
/* Try to register PIRQ EOI map */
|
|
xen_intr_pirq_eoi_map = malloc(PAGE_SIZE, M_XENINTR, M_WAITOK | M_ZERO);
|
|
eoi_gmfn.gmfn = atop(vtophys(xen_intr_pirq_eoi_map));
|
|
rc = HYPERVISOR_physdev_op(PHYSDEVOP_pirq_eoi_gmfn_v2, &eoi_gmfn);
|
|
if (rc != 0 && bootverbose)
|
|
printf("Xen interrupts: unable to register PIRQ EOI map\n");
|
|
else
|
|
xen_intr_pirq_eoi_map_enabled = true;
|
|
|
|
intr_register_pic(&xen_intr_pic);
|
|
intr_register_pic(&xen_intr_pirq_pic);
|
|
|
|
if (bootverbose)
|
|
printf("Xen interrupt system initialized\n");
|
|
|
|
return (0);
|
|
}
|
|
SYSINIT(xen_intr_init, SI_SUB_INTR, SI_ORDER_SECOND, xen_intr_init, NULL);
|
|
|
|
/*--------------------------- Common PIC Functions ---------------------------*/
|
|
/**
|
|
* Prepare this PIC for system suspension.
|
|
*/
|
|
static void
|
|
xen_intr_suspend(struct pic *unused)
|
|
{
|
|
}
|
|
|
|
static void
|
|
xen_rebind_ipi(struct xenisrc *isrc)
|
|
{
|
|
#ifdef SMP
|
|
int cpu = isrc->xi_cpu;
|
|
int vcpu_id = pcpu_find(cpu)->pc_vcpu_id;
|
|
int error;
|
|
struct evtchn_bind_ipi bind_ipi = { .vcpu = vcpu_id };
|
|
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_ipi,
|
|
&bind_ipi);
|
|
if (error != 0)
|
|
panic("unable to rebind xen IPI: %d", error);
|
|
|
|
isrc->xi_port = bind_ipi.port;
|
|
isrc->xi_cpu = 0;
|
|
xen_intr_port_to_isrc[bind_ipi.port] = isrc;
|
|
|
|
error = xen_intr_assign_cpu(&isrc->xi_intsrc,
|
|
cpu_apic_ids[cpu]);
|
|
if (error)
|
|
panic("unable to bind xen IPI to CPU#%d: %d",
|
|
cpu, error);
|
|
|
|
evtchn_unmask_port(bind_ipi.port);
|
|
#else
|
|
panic("Resume IPI event channel on UP");
|
|
#endif
|
|
}
|
|
|
|
static void
|
|
xen_rebind_virq(struct xenisrc *isrc)
|
|
{
|
|
int cpu = isrc->xi_cpu;
|
|
int vcpu_id = pcpu_find(cpu)->pc_vcpu_id;
|
|
int error;
|
|
struct evtchn_bind_virq bind_virq = { .virq = isrc->xi_virq,
|
|
.vcpu = vcpu_id };
|
|
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_virq,
|
|
&bind_virq);
|
|
if (error != 0)
|
|
panic("unable to rebind xen VIRQ#%d: %d", isrc->xi_virq, error);
|
|
|
|
isrc->xi_port = bind_virq.port;
|
|
isrc->xi_cpu = 0;
|
|
xen_intr_port_to_isrc[bind_virq.port] = isrc;
|
|
|
|
#ifdef SMP
|
|
error = xen_intr_assign_cpu(&isrc->xi_intsrc,
|
|
cpu_apic_ids[cpu]);
|
|
if (error)
|
|
panic("unable to bind xen VIRQ#%d to CPU#%d: %d",
|
|
isrc->xi_virq, cpu, error);
|
|
#endif
|
|
|
|
evtchn_unmask_port(bind_virq.port);
|
|
}
|
|
|
|
/**
|
|
* Return this PIC to service after being suspended.
|
|
*/
|
|
static void
|
|
xen_intr_resume(struct pic *unused, bool suspend_cancelled)
|
|
{
|
|
shared_info_t *s = HYPERVISOR_shared_info;
|
|
struct xenisrc *isrc;
|
|
u_int isrc_idx;
|
|
int i;
|
|
|
|
if (suspend_cancelled)
|
|
return;
|
|
|
|
/* Reset the per-CPU masks */
|
|
CPU_FOREACH(i) {
|
|
struct xen_intr_pcpu_data *pcpu;
|
|
|
|
pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
|
|
memset(pcpu->evtchn_enabled, i == 0 ? ~0 : 0,
|
|
sizeof(pcpu->evtchn_enabled));
|
|
}
|
|
|
|
/* Mask all event channels. */
|
|
for (i = 0; i < nitems(s->evtchn_mask); i++)
|
|
atomic_store_rel_long(&s->evtchn_mask[i], ~0);
|
|
|
|
/* Remove port -> isrc mappings */
|
|
memset(xen_intr_port_to_isrc, 0, sizeof(xen_intr_port_to_isrc));
|
|
|
|
/* Free unused isrcs and rebind VIRQs and IPIs */
|
|
for (isrc_idx = 0; isrc_idx < xen_intr_auto_vector_count; isrc_idx++) {
|
|
u_int vector;
|
|
|
|
vector = FIRST_EVTCHN_INT + isrc_idx;
|
|
isrc = (struct xenisrc *)intr_lookup_source(vector);
|
|
if (isrc != NULL) {
|
|
isrc->xi_port = 0;
|
|
switch (isrc->xi_type) {
|
|
case EVTCHN_TYPE_IPI:
|
|
xen_rebind_ipi(isrc);
|
|
break;
|
|
case EVTCHN_TYPE_VIRQ:
|
|
xen_rebind_virq(isrc);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Disable a Xen interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to disable.
|
|
*/
|
|
static void
|
|
xen_intr_disable_intr(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc = (struct xenisrc *)base_isrc;
|
|
|
|
evtchn_mask_port(isrc->xi_port);
|
|
}
|
|
|
|
/**
|
|
* Determine the global interrupt vector number for
|
|
* a Xen interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to query.
|
|
*
|
|
* \return The vector number corresponding to the given interrupt source.
|
|
*/
|
|
static int
|
|
xen_intr_vector(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc = (struct xenisrc *)base_isrc;
|
|
|
|
return (isrc->xi_vector);
|
|
}
|
|
|
|
/**
|
|
* Determine whether or not interrupt events are pending on the
|
|
* the given interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to query.
|
|
*
|
|
* \returns 0 if no events are pending, otherwise non-zero.
|
|
*/
|
|
static int
|
|
xen_intr_source_pending(struct intsrc *isrc)
|
|
{
|
|
/*
|
|
* EventChannels are edge triggered and never masked.
|
|
* There can be no pending events.
|
|
*/
|
|
return (0);
|
|
}
|
|
|
|
/**
|
|
* Perform configuration of an interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to configure.
|
|
* \param trig Edge or level.
|
|
* \param pol Active high or low.
|
|
*
|
|
* \returns 0 if no events are pending, otherwise non-zero.
|
|
*/
|
|
static int
|
|
xen_intr_config_intr(struct intsrc *isrc, enum intr_trigger trig,
|
|
enum intr_polarity pol)
|
|
{
|
|
/* Configuration is only possible via the evtchn apis. */
|
|
return (ENODEV);
|
|
}
|
|
|
|
/**
|
|
* Configure CPU affinity for interrupt source event delivery.
|
|
*
|
|
* \param isrc The interrupt source to configure.
|
|
* \param apic_id The apic id of the CPU for handling future events.
|
|
*
|
|
* \returns 0 if successful, otherwise an errno.
|
|
*/
|
|
static int
|
|
xen_intr_assign_cpu(struct intsrc *base_isrc, u_int apic_id)
|
|
{
|
|
#ifdef SMP
|
|
struct evtchn_bind_vcpu bind_vcpu;
|
|
struct xenisrc *isrc;
|
|
u_int to_cpu, vcpu_id;
|
|
int error, masked;
|
|
|
|
if (xen_vector_callback_enabled == 0)
|
|
return (EOPNOTSUPP);
|
|
|
|
to_cpu = apic_cpuid(apic_id);
|
|
vcpu_id = pcpu_find(to_cpu)->pc_vcpu_id;
|
|
xen_intr_intrcnt_add(to_cpu);
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
if (!is_valid_evtchn(isrc->xi_port)) {
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
return (EINVAL);
|
|
}
|
|
|
|
/*
|
|
* Mask the event channel while binding it to prevent interrupt
|
|
* delivery with an inconsistent state in isrc->xi_cpu.
|
|
*/
|
|
masked = evtchn_test_and_set_mask(isrc->xi_port);
|
|
if ((isrc->xi_type == EVTCHN_TYPE_VIRQ) ||
|
|
(isrc->xi_type == EVTCHN_TYPE_IPI)) {
|
|
/*
|
|
* Virtual IRQs are associated with a cpu by
|
|
* the Hypervisor at evtchn_bind_virq time, so
|
|
* all we need to do is update the per-CPU masks.
|
|
*/
|
|
evtchn_cpu_mask_port(isrc->xi_cpu, isrc->xi_port);
|
|
isrc->xi_cpu = to_cpu;
|
|
evtchn_cpu_unmask_port(isrc->xi_cpu, isrc->xi_port);
|
|
goto out;
|
|
}
|
|
|
|
bind_vcpu.port = isrc->xi_port;
|
|
bind_vcpu.vcpu = vcpu_id;
|
|
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_vcpu, &bind_vcpu);
|
|
if (isrc->xi_cpu != to_cpu) {
|
|
if (error == 0) {
|
|
/* Commit to new binding by removing the old one. */
|
|
evtchn_cpu_mask_port(isrc->xi_cpu, isrc->xi_port);
|
|
isrc->xi_cpu = to_cpu;
|
|
evtchn_cpu_unmask_port(isrc->xi_cpu, isrc->xi_port);
|
|
}
|
|
}
|
|
|
|
out:
|
|
if (masked == 0)
|
|
evtchn_unmask_port(isrc->xi_port);
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
return (0);
|
|
#else
|
|
return (EOPNOTSUPP);
|
|
#endif
|
|
}
|
|
|
|
/*------------------- Virtual Interrupt Source PIC Functions -----------------*/
|
|
/*
|
|
* Mask a level triggered interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to mask (if necessary).
|
|
* \param eoi If non-zero, perform any necessary end-of-interrupt
|
|
* acknowledgements.
|
|
*/
|
|
static void
|
|
xen_intr_disable_source(struct intsrc *base_isrc, int eoi)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
/*
|
|
* NB: checking if the event channel is already masked is
|
|
* needed because the event channel user-space device
|
|
* masks event channels on its filter as part of its
|
|
* normal operation, and those shouldn't be automatically
|
|
* unmasked by the generic interrupt code. The event channel
|
|
* device will unmask them when needed.
|
|
*/
|
|
isrc->xi_masked = !!evtchn_test_and_set_mask(isrc->xi_port);
|
|
}
|
|
|
|
/*
|
|
* Unmask a level triggered interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to unmask (if necessary).
|
|
*/
|
|
static void
|
|
xen_intr_enable_source(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
if (isrc->xi_masked == 0)
|
|
evtchn_unmask_port(isrc->xi_port);
|
|
}
|
|
|
|
/*
|
|
* Perform any necessary end-of-interrupt acknowledgements.
|
|
*
|
|
* \param isrc The interrupt source to EOI.
|
|
*/
|
|
static void
|
|
xen_intr_eoi_source(struct intsrc *base_isrc)
|
|
{
|
|
}
|
|
|
|
/*
|
|
* Enable and unmask the interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to enable.
|
|
*/
|
|
static void
|
|
xen_intr_enable_intr(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc = (struct xenisrc *)base_isrc;
|
|
|
|
evtchn_unmask_port(isrc->xi_port);
|
|
}
|
|
|
|
/*------------------ Physical Interrupt Source PIC Functions -----------------*/
|
|
/*
|
|
* Mask a level triggered interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to mask (if necessary).
|
|
* \param eoi If non-zero, perform any necessary end-of-interrupt
|
|
* acknowledgements.
|
|
*/
|
|
static void
|
|
xen_intr_pirq_disable_source(struct intsrc *base_isrc, int eoi)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
if (isrc->xi_edgetrigger == 0)
|
|
evtchn_mask_port(isrc->xi_port);
|
|
if (eoi == PIC_EOI)
|
|
xen_intr_pirq_eoi_source(base_isrc);
|
|
}
|
|
|
|
/*
|
|
* Unmask a level triggered interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to unmask (if necessary).
|
|
*/
|
|
static void
|
|
xen_intr_pirq_enable_source(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
if (isrc->xi_edgetrigger == 0)
|
|
evtchn_unmask_port(isrc->xi_port);
|
|
}
|
|
|
|
/*
|
|
* Perform any necessary end-of-interrupt acknowledgements.
|
|
*
|
|
* \param isrc The interrupt source to EOI.
|
|
*/
|
|
static void
|
|
xen_intr_pirq_eoi_source(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc;
|
|
int error;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
if (xen_test_bit(isrc->xi_pirq, xen_intr_pirq_eoi_map)) {
|
|
struct physdev_eoi eoi = { .irq = isrc->xi_pirq };
|
|
|
|
error = HYPERVISOR_physdev_op(PHYSDEVOP_eoi, &eoi);
|
|
if (error != 0)
|
|
panic("Unable to EOI PIRQ#%d: %d\n",
|
|
isrc->xi_pirq, error);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Enable and unmask the interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to enable.
|
|
*/
|
|
static void
|
|
xen_intr_pirq_enable_intr(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc;
|
|
struct evtchn_bind_pirq bind_pirq;
|
|
struct physdev_irq_status_query irq_status;
|
|
int error;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
if (!xen_intr_pirq_eoi_map_enabled) {
|
|
irq_status.irq = isrc->xi_pirq;
|
|
error = HYPERVISOR_physdev_op(PHYSDEVOP_irq_status_query,
|
|
&irq_status);
|
|
if (error)
|
|
panic("unable to get status of IRQ#%d", isrc->xi_pirq);
|
|
|
|
if (irq_status.flags & XENIRQSTAT_needs_eoi) {
|
|
/*
|
|
* Since the dynamic PIRQ EOI map is not available
|
|
* mark the PIRQ as needing EOI unconditionally.
|
|
*/
|
|
xen_set_bit(isrc->xi_pirq, xen_intr_pirq_eoi_map);
|
|
}
|
|
}
|
|
|
|
bind_pirq.pirq = isrc->xi_pirq;
|
|
bind_pirq.flags = isrc->xi_edgetrigger ? 0 : BIND_PIRQ__WILL_SHARE;
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_pirq, &bind_pirq);
|
|
if (error)
|
|
panic("unable to bind IRQ#%d", isrc->xi_pirq);
|
|
|
|
isrc->xi_port = bind_pirq.port;
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
KASSERT((xen_intr_port_to_isrc[bind_pirq.port] == NULL),
|
|
("trying to override an already setup event channel port"));
|
|
xen_intr_port_to_isrc[bind_pirq.port] = isrc;
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
|
|
evtchn_unmask_port(isrc->xi_port);
|
|
}
|
|
|
|
/*
|
|
* Disable an interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to disable.
|
|
*/
|
|
static void
|
|
xen_intr_pirq_disable_intr(struct intsrc *base_isrc)
|
|
{
|
|
struct xenisrc *isrc;
|
|
struct evtchn_close close;
|
|
int error;
|
|
|
|
isrc = (struct xenisrc *)base_isrc;
|
|
|
|
evtchn_mask_port(isrc->xi_port);
|
|
|
|
close.port = isrc->xi_port;
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_close, &close);
|
|
if (error)
|
|
panic("unable to close event channel %d IRQ#%d",
|
|
isrc->xi_port, isrc->xi_pirq);
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
xen_intr_port_to_isrc[isrc->xi_port] = NULL;
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
|
|
isrc->xi_port = 0;
|
|
}
|
|
|
|
/**
|
|
* Perform configuration of an interrupt source.
|
|
*
|
|
* \param isrc The interrupt source to configure.
|
|
* \param trig Edge or level.
|
|
* \param pol Active high or low.
|
|
*
|
|
* \returns 0 if no events are pending, otherwise non-zero.
|
|
*/
|
|
static int
|
|
xen_intr_pirq_config_intr(struct intsrc *base_isrc, enum intr_trigger trig,
|
|
enum intr_polarity pol)
|
|
{
|
|
struct xenisrc *isrc = (struct xenisrc *)base_isrc;
|
|
struct physdev_setup_gsi setup_gsi;
|
|
int error;
|
|
|
|
KASSERT(!(trig == INTR_TRIGGER_CONFORM || pol == INTR_POLARITY_CONFORM),
|
|
("%s: Conforming trigger or polarity\n", __func__));
|
|
|
|
setup_gsi.gsi = isrc->xi_pirq;
|
|
setup_gsi.triggering = trig == INTR_TRIGGER_EDGE ? 0 : 1;
|
|
setup_gsi.polarity = pol == INTR_POLARITY_HIGH ? 0 : 1;
|
|
|
|
error = HYPERVISOR_physdev_op(PHYSDEVOP_setup_gsi, &setup_gsi);
|
|
if (error == -XEN_EEXIST) {
|
|
if ((isrc->xi_edgetrigger && (trig != INTR_TRIGGER_EDGE)) ||
|
|
(isrc->xi_activehi && (pol != INTR_POLARITY_HIGH)))
|
|
panic("unable to reconfigure interrupt IRQ#%d",
|
|
isrc->xi_pirq);
|
|
error = 0;
|
|
}
|
|
if (error)
|
|
panic("unable to configure IRQ#%d\n", isrc->xi_pirq);
|
|
|
|
isrc->xi_activehi = pol == INTR_POLARITY_HIGH ? 1 : 0;
|
|
isrc->xi_edgetrigger = trig == INTR_TRIGGER_EDGE ? 1 : 0;
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*--------------------------- Public Functions -------------------------------*/
|
|
/*------- API comments for these methods can be found in xen/xenintr.h -------*/
|
|
int
|
|
xen_intr_bind_local_port(device_t dev, evtchn_port_t local_port,
|
|
driver_filter_t filter, driver_intr_t handler, void *arg,
|
|
enum intr_type flags, xen_intr_handle_t *port_handlep)
|
|
{
|
|
struct xenisrc *isrc;
|
|
int error;
|
|
|
|
error = xen_intr_bind_isrc(&isrc, local_port, EVTCHN_TYPE_PORT, dev,
|
|
filter, handler, arg, flags, port_handlep);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
/*
|
|
* The Event Channel API didn't open this port, so it is not
|
|
* responsible for closing it automatically on unbind.
|
|
*/
|
|
isrc->xi_close = 0;
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_intr_alloc_and_bind_local_port(device_t dev, u_int remote_domain,
|
|
driver_filter_t filter, driver_intr_t handler, void *arg,
|
|
enum intr_type flags, xen_intr_handle_t *port_handlep)
|
|
{
|
|
struct xenisrc *isrc;
|
|
struct evtchn_alloc_unbound alloc_unbound;
|
|
int error;
|
|
|
|
alloc_unbound.dom = DOMID_SELF;
|
|
alloc_unbound.remote_dom = remote_domain;
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_alloc_unbound,
|
|
&alloc_unbound);
|
|
if (error != 0) {
|
|
/*
|
|
* XXX Trap Hypercall error code Linuxisms in
|
|
* the HYPERCALL layer.
|
|
*/
|
|
return (-error);
|
|
}
|
|
|
|
error = xen_intr_bind_isrc(&isrc, alloc_unbound.port, EVTCHN_TYPE_PORT,
|
|
dev, filter, handler, arg, flags,
|
|
port_handlep);
|
|
if (error != 0) {
|
|
evtchn_close_t close = { .port = alloc_unbound.port };
|
|
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
|
|
panic("EVTCHNOP_close failed");
|
|
return (error);
|
|
}
|
|
|
|
isrc->xi_close = 1;
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_intr_bind_remote_port(device_t dev, u_int remote_domain,
|
|
u_int remote_port, driver_filter_t filter, driver_intr_t handler,
|
|
void *arg, enum intr_type flags, xen_intr_handle_t *port_handlep)
|
|
{
|
|
struct xenisrc *isrc;
|
|
struct evtchn_bind_interdomain bind_interdomain;
|
|
int error;
|
|
|
|
bind_interdomain.remote_dom = remote_domain;
|
|
bind_interdomain.remote_port = remote_port;
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_interdomain,
|
|
&bind_interdomain);
|
|
if (error != 0) {
|
|
/*
|
|
* XXX Trap Hypercall error code Linuxisms in
|
|
* the HYPERCALL layer.
|
|
*/
|
|
return (-error);
|
|
}
|
|
|
|
error = xen_intr_bind_isrc(&isrc, bind_interdomain.local_port,
|
|
EVTCHN_TYPE_PORT, dev, filter, handler,
|
|
arg, flags, port_handlep);
|
|
if (error) {
|
|
evtchn_close_t close = { .port = bind_interdomain.local_port };
|
|
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
|
|
panic("EVTCHNOP_close failed");
|
|
return (error);
|
|
}
|
|
|
|
/*
|
|
* The Event Channel API opened this port, so it is
|
|
* responsible for closing it automatically on unbind.
|
|
*/
|
|
isrc->xi_close = 1;
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_intr_bind_virq(device_t dev, u_int virq, u_int cpu,
|
|
driver_filter_t filter, driver_intr_t handler, void *arg,
|
|
enum intr_type flags, xen_intr_handle_t *port_handlep)
|
|
{
|
|
int vcpu_id = pcpu_find(cpu)->pc_vcpu_id;
|
|
struct xenisrc *isrc;
|
|
struct evtchn_bind_virq bind_virq = { .virq = virq, .vcpu = vcpu_id };
|
|
int error;
|
|
|
|
/* Ensure the target CPU is ready to handle evtchn interrupts. */
|
|
xen_intr_intrcnt_add(cpu);
|
|
|
|
isrc = NULL;
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_virq, &bind_virq);
|
|
if (error != 0) {
|
|
/*
|
|
* XXX Trap Hypercall error code Linuxisms in
|
|
* the HYPERCALL layer.
|
|
*/
|
|
return (-error);
|
|
}
|
|
|
|
error = xen_intr_bind_isrc(&isrc, bind_virq.port, EVTCHN_TYPE_VIRQ, dev,
|
|
filter, handler, arg, flags, port_handlep);
|
|
|
|
#ifdef SMP
|
|
if (error == 0)
|
|
error = intr_event_bind(isrc->xi_intsrc.is_event, cpu);
|
|
#endif
|
|
|
|
if (error != 0) {
|
|
evtchn_close_t close = { .port = bind_virq.port };
|
|
|
|
xen_intr_unbind(*port_handlep);
|
|
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
|
|
panic("EVTCHNOP_close failed");
|
|
return (error);
|
|
}
|
|
|
|
#ifdef SMP
|
|
if (isrc->xi_cpu != cpu) {
|
|
/*
|
|
* Too early in the boot process for the generic interrupt
|
|
* code to perform the binding. Update our event channel
|
|
* masks manually so events can't fire on the wrong cpu
|
|
* during AP startup.
|
|
*/
|
|
xen_intr_assign_cpu(&isrc->xi_intsrc, cpu_apic_ids[cpu]);
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* The Event Channel API opened this port, so it is
|
|
* responsible for closing it automatically on unbind.
|
|
*/
|
|
isrc->xi_close = 1;
|
|
isrc->xi_virq = virq;
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_intr_alloc_and_bind_ipi(device_t dev, u_int cpu,
|
|
driver_filter_t filter, enum intr_type flags,
|
|
xen_intr_handle_t *port_handlep)
|
|
{
|
|
#ifdef SMP
|
|
int vcpu_id = pcpu_find(cpu)->pc_vcpu_id;
|
|
struct xenisrc *isrc;
|
|
struct evtchn_bind_ipi bind_ipi = { .vcpu = vcpu_id };
|
|
int error;
|
|
|
|
/* Ensure the target CPU is ready to handle evtchn interrupts. */
|
|
xen_intr_intrcnt_add(cpu);
|
|
|
|
isrc = NULL;
|
|
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_ipi, &bind_ipi);
|
|
if (error != 0) {
|
|
/*
|
|
* XXX Trap Hypercall error code Linuxisms in
|
|
* the HYPERCALL layer.
|
|
*/
|
|
return (-error);
|
|
}
|
|
|
|
error = xen_intr_bind_isrc(&isrc, bind_ipi.port, EVTCHN_TYPE_IPI,
|
|
dev, filter, NULL, NULL, flags,
|
|
port_handlep);
|
|
if (error == 0)
|
|
error = intr_event_bind(isrc->xi_intsrc.is_event, cpu);
|
|
|
|
if (error != 0) {
|
|
evtchn_close_t close = { .port = bind_ipi.port };
|
|
|
|
xen_intr_unbind(*port_handlep);
|
|
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close))
|
|
panic("EVTCHNOP_close failed");
|
|
return (error);
|
|
}
|
|
|
|
if (isrc->xi_cpu != cpu) {
|
|
/*
|
|
* Too early in the boot process for the generic interrupt
|
|
* code to perform the binding. Update our event channel
|
|
* masks manually so events can't fire on the wrong cpu
|
|
* during AP startup.
|
|
*/
|
|
xen_intr_assign_cpu(&isrc->xi_intsrc, cpu_apic_ids[cpu]);
|
|
}
|
|
|
|
/*
|
|
* The Event Channel API opened this port, so it is
|
|
* responsible for closing it automatically on unbind.
|
|
*/
|
|
isrc->xi_close = 1;
|
|
return (0);
|
|
#else
|
|
return (EOPNOTSUPP);
|
|
#endif
|
|
}
|
|
|
|
int
|
|
xen_register_pirq(int vector, enum intr_trigger trig, enum intr_polarity pol)
|
|
{
|
|
struct physdev_map_pirq map_pirq;
|
|
struct xenisrc *isrc;
|
|
int error;
|
|
|
|
if (vector == 0)
|
|
return (EINVAL);
|
|
|
|
if (bootverbose)
|
|
printf("xen: register IRQ#%d\n", vector);
|
|
|
|
map_pirq.domid = DOMID_SELF;
|
|
map_pirq.type = MAP_PIRQ_TYPE_GSI;
|
|
map_pirq.index = vector;
|
|
map_pirq.pirq = vector;
|
|
|
|
error = HYPERVISOR_physdev_op(PHYSDEVOP_map_pirq, &map_pirq);
|
|
if (error) {
|
|
printf("xen: unable to map IRQ#%d\n", vector);
|
|
return (error);
|
|
}
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
isrc = xen_intr_alloc_isrc(EVTCHN_TYPE_PIRQ, vector);
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
KASSERT((isrc != NULL), ("xen: unable to allocate isrc for interrupt"));
|
|
isrc->xi_pirq = vector;
|
|
isrc->xi_activehi = pol == INTR_POLARITY_HIGH ? 1 : 0;
|
|
isrc->xi_edgetrigger = trig == INTR_TRIGGER_EDGE ? 1 : 0;
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_register_msi(device_t dev, int vector, int count)
|
|
{
|
|
struct physdev_map_pirq msi_irq;
|
|
struct xenisrc *isrc;
|
|
int ret;
|
|
|
|
memset(&msi_irq, 0, sizeof(msi_irq));
|
|
msi_irq.domid = DOMID_SELF;
|
|
msi_irq.type = count == 1 ?
|
|
MAP_PIRQ_TYPE_MSI_SEG : MAP_PIRQ_TYPE_MULTI_MSI;
|
|
msi_irq.index = -1;
|
|
msi_irq.pirq = -1;
|
|
msi_irq.bus = pci_get_bus(dev) | (pci_get_domain(dev) << 16);
|
|
msi_irq.devfn = (pci_get_slot(dev) << 3) | pci_get_function(dev);
|
|
msi_irq.entry_nr = count;
|
|
|
|
ret = HYPERVISOR_physdev_op(PHYSDEVOP_map_pirq, &msi_irq);
|
|
if (ret != 0)
|
|
return (ret);
|
|
if (count != msi_irq.entry_nr) {
|
|
panic("unable to setup all requested MSI vectors "
|
|
"(expected %d got %d)", count, msi_irq.entry_nr);
|
|
}
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
for (int i = 0; i < count; i++) {
|
|
isrc = xen_intr_alloc_isrc(EVTCHN_TYPE_PIRQ, vector + i);
|
|
KASSERT(isrc != NULL,
|
|
("xen: unable to allocate isrc for interrupt"));
|
|
isrc->xi_pirq = msi_irq.pirq + i;
|
|
/* MSI interrupts are always edge triggered */
|
|
isrc->xi_edgetrigger = 1;
|
|
}
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_release_msi(int vector)
|
|
{
|
|
struct physdev_unmap_pirq unmap;
|
|
struct xenisrc *isrc;
|
|
int ret;
|
|
|
|
isrc = (struct xenisrc *)intr_lookup_source(vector);
|
|
if (isrc == NULL)
|
|
return (ENXIO);
|
|
|
|
unmap.pirq = isrc->xi_pirq;
|
|
ret = HYPERVISOR_physdev_op(PHYSDEVOP_unmap_pirq, &unmap);
|
|
if (ret != 0)
|
|
return (ret);
|
|
|
|
xen_intr_release_isrc(isrc);
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
xen_intr_describe(xen_intr_handle_t port_handle, const char *fmt, ...)
|
|
{
|
|
char descr[MAXCOMLEN + 1];
|
|
struct xenisrc *isrc;
|
|
va_list ap;
|
|
|
|
isrc = xen_intr_isrc(port_handle);
|
|
if (isrc == NULL)
|
|
return (EINVAL);
|
|
|
|
va_start(ap, fmt);
|
|
vsnprintf(descr, sizeof(descr), fmt, ap);
|
|
va_end(ap);
|
|
return (intr_describe(isrc->xi_vector, isrc->xi_cookie, descr));
|
|
}
|
|
|
|
void
|
|
xen_intr_unbind(xen_intr_handle_t *port_handlep)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
KASSERT(port_handlep != NULL,
|
|
("NULL xen_intr_handle_t passed to xen_intr_unbind"));
|
|
|
|
isrc = xen_intr_isrc(*port_handlep);
|
|
*port_handlep = NULL;
|
|
if (isrc == NULL)
|
|
return;
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
if (refcount_release(&isrc->xi_refcount) == 0) {
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
return;
|
|
}
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
|
|
if (isrc->xi_cookie != NULL)
|
|
intr_remove_handler(isrc->xi_cookie);
|
|
xen_intr_release_isrc(isrc);
|
|
}
|
|
|
|
void
|
|
xen_intr_signal(xen_intr_handle_t handle)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = xen_intr_isrc(handle);
|
|
if (isrc != NULL) {
|
|
KASSERT(isrc->xi_type == EVTCHN_TYPE_PORT ||
|
|
isrc->xi_type == EVTCHN_TYPE_IPI,
|
|
("evtchn_signal on something other than a local port"));
|
|
struct evtchn_send send = { .port = isrc->xi_port };
|
|
(void)HYPERVISOR_event_channel_op(EVTCHNOP_send, &send);
|
|
}
|
|
}
|
|
|
|
evtchn_port_t
|
|
xen_intr_port(xen_intr_handle_t handle)
|
|
{
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = xen_intr_isrc(handle);
|
|
if (isrc == NULL)
|
|
return (0);
|
|
|
|
return (isrc->xi_port);
|
|
}
|
|
|
|
int
|
|
xen_intr_add_handler(device_t dev, driver_filter_t filter,
|
|
driver_intr_t handler, void *arg, enum intr_type flags,
|
|
xen_intr_handle_t handle)
|
|
{
|
|
struct xenisrc *isrc;
|
|
int error;
|
|
|
|
isrc = xen_intr_isrc(handle);
|
|
if (isrc == NULL || isrc->xi_cookie != NULL)
|
|
return (EINVAL);
|
|
|
|
error = intr_add_handler(device_get_nameunit(dev), isrc->xi_vector,
|
|
filter, handler, arg, flags|INTR_EXCL, &isrc->xi_cookie);
|
|
if (error != 0) {
|
|
device_printf(dev,
|
|
"xen_intr_add_handler: intr_add_handler failed: %d\n",
|
|
error);
|
|
}
|
|
|
|
return (error);
|
|
}
|
|
|
|
int
|
|
xen_intr_get_evtchn_from_port(evtchn_port_t port, xen_intr_handle_t *handlep)
|
|
{
|
|
|
|
if (!is_valid_evtchn(port) || port >= NR_EVENT_CHANNELS)
|
|
return (EINVAL);
|
|
|
|
if (handlep == NULL) {
|
|
return (EINVAL);
|
|
}
|
|
|
|
mtx_lock(&xen_intr_isrc_lock);
|
|
if (xen_intr_port_to_isrc[port] == NULL) {
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
return (EINVAL);
|
|
}
|
|
refcount_acquire(&xen_intr_port_to_isrc[port]->xi_refcount);
|
|
mtx_unlock(&xen_intr_isrc_lock);
|
|
|
|
/* Assign the opaque handler (the event channel port) */
|
|
*handlep = &xen_intr_port_to_isrc[port]->xi_vector;
|
|
|
|
return (0);
|
|
}
|
|
|
|
#ifdef DDB
|
|
static const char *
|
|
xen_intr_print_type(enum evtchn_type type)
|
|
{
|
|
static const char *evtchn_type_to_string[EVTCHN_TYPE_COUNT] = {
|
|
[EVTCHN_TYPE_UNBOUND] = "UNBOUND",
|
|
[EVTCHN_TYPE_PIRQ] = "PIRQ",
|
|
[EVTCHN_TYPE_VIRQ] = "VIRQ",
|
|
[EVTCHN_TYPE_IPI] = "IPI",
|
|
[EVTCHN_TYPE_PORT] = "PORT",
|
|
};
|
|
|
|
if (type >= EVTCHN_TYPE_COUNT)
|
|
return ("UNKNOWN");
|
|
|
|
return (evtchn_type_to_string[type]);
|
|
}
|
|
|
|
static void
|
|
xen_intr_dump_port(struct xenisrc *isrc)
|
|
{
|
|
struct xen_intr_pcpu_data *pcpu;
|
|
shared_info_t *s = HYPERVISOR_shared_info;
|
|
int i;
|
|
|
|
db_printf("Port %d Type: %s\n",
|
|
isrc->xi_port, xen_intr_print_type(isrc->xi_type));
|
|
if (isrc->xi_type == EVTCHN_TYPE_PIRQ) {
|
|
db_printf("\tPirq: %d ActiveHi: %d EdgeTrigger: %d "
|
|
"NeedsEOI: %d\n",
|
|
isrc->xi_pirq, isrc->xi_activehi, isrc->xi_edgetrigger,
|
|
!!xen_test_bit(isrc->xi_pirq, xen_intr_pirq_eoi_map));
|
|
}
|
|
if (isrc->xi_type == EVTCHN_TYPE_VIRQ)
|
|
db_printf("\tVirq: %d\n", isrc->xi_virq);
|
|
|
|
db_printf("\tMasked: %d Pending: %d\n",
|
|
!!xen_test_bit(isrc->xi_port, &s->evtchn_mask[0]),
|
|
!!xen_test_bit(isrc->xi_port, &s->evtchn_pending[0]));
|
|
|
|
db_printf("\tPer-CPU Masks: ");
|
|
CPU_FOREACH(i) {
|
|
pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
|
|
db_printf("cpu#%d: %d ", i,
|
|
!!xen_test_bit(isrc->xi_port, pcpu->evtchn_enabled));
|
|
}
|
|
db_printf("\n");
|
|
}
|
|
|
|
DB_SHOW_COMMAND(xen_evtchn, db_show_xen_evtchn)
|
|
{
|
|
int i;
|
|
|
|
if (!xen_domain()) {
|
|
db_printf("Only available on Xen guests\n");
|
|
return;
|
|
}
|
|
|
|
for (i = 0; i < NR_EVENT_CHANNELS; i++) {
|
|
struct xenisrc *isrc;
|
|
|
|
isrc = xen_intr_port_to_isrc[i];
|
|
if (isrc == NULL)
|
|
continue;
|
|
|
|
xen_intr_dump_port(isrc);
|
|
}
|
|
}
|
|
#endif /* DDB */
|