You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
Merge commit 'konrad/stable/xen-pcifront-0.8.2' into 2.6.36-rc8-initial-domain-v6
This commit is contained in:
+2
-1
@@ -74,6 +74,7 @@ config XEN_PLATFORM_PCI
|
||||
|
||||
config SWIOTLB_XEN
|
||||
def_bool y
|
||||
depends on SWIOTLB
|
||||
depends on PCI
|
||||
select SWIOTLB
|
||||
|
||||
endmenu
|
||||
|
||||
@@ -4,6 +4,7 @@ obj-y += xenbus/
|
||||
nostackp := $(call cc-option, -fno-stack-protector)
|
||||
CFLAGS_features.o := $(nostackp)
|
||||
|
||||
obj-$(CONFIG_BLOCK) += biomerge.o
|
||||
obj-$(CONFIG_HOTPLUG_CPU) += cpu_hotplug.o
|
||||
obj-$(CONFIG_XEN_XENCOMM) += xencomm.o
|
||||
obj-$(CONFIG_XEN_BALLOON) += balloon.o
|
||||
|
||||
@@ -0,0 +1,13 @@
|
||||
#include <linux/bio.h>
|
||||
#include <linux/io.h>
|
||||
#include <xen/page.h>
|
||||
|
||||
bool xen_biovec_phys_mergeable(const struct bio_vec *vec1,
|
||||
const struct bio_vec *vec2)
|
||||
{
|
||||
unsigned long mfn1 = pfn_to_mfn(page_to_pfn(vec1->bv_page));
|
||||
unsigned long mfn2 = pfn_to_mfn(page_to_pfn(vec2->bv_page));
|
||||
|
||||
return __BIOVEC_PHYS_MERGEABLE(vec1, vec2) &&
|
||||
((mfn1 == mfn2) || ((mfn1+1) == mfn2));
|
||||
}
|
||||
+338
-27
@@ -16,7 +16,7 @@
|
||||
* (typically dom0).
|
||||
* 2. VIRQs, typically used for timers. These are per-cpu events.
|
||||
* 3. IPIs.
|
||||
* 4. Hardware interrupts. Not supported at present.
|
||||
* 4. PIRQs - Hardware interrupts.
|
||||
*
|
||||
* Jeremy Fitzhardinge <jeremy@xensource.com>, XenSource Inc, 2007
|
||||
*/
|
||||
@@ -28,11 +28,13 @@
|
||||
#include <linux/string.h>
|
||||
#include <linux/bootmem.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/irqnr.h>
|
||||
|
||||
#include <asm/desc.h>
|
||||
#include <asm/ptrace.h>
|
||||
#include <asm/irq.h>
|
||||
#include <asm/idle.h>
|
||||
#include <asm/io_apic.h>
|
||||
#include <asm/sync_bitops.h>
|
||||
#include <asm/xen/hypercall.h>
|
||||
#include <asm/xen/hypervisor.h>
|
||||
@@ -89,20 +91,26 @@ struct irq_info
|
||||
enum ipi_vector ipi;
|
||||
struct {
|
||||
unsigned short gsi;
|
||||
unsigned short vector;
|
||||
unsigned char vector;
|
||||
unsigned char flags;
|
||||
} pirq;
|
||||
} u;
|
||||
};
|
||||
#define PIRQ_NEEDS_EOI (1 << 0)
|
||||
#define PIRQ_SHAREABLE (1 << 1)
|
||||
|
||||
static struct irq_info irq_info[NR_IRQS];
|
||||
static struct irq_info *irq_info;
|
||||
|
||||
static int evtchn_to_irq[NR_EVENT_CHANNELS] = {
|
||||
[0 ... NR_EVENT_CHANNELS-1] = -1
|
||||
};
|
||||
static int *evtchn_to_irq;
|
||||
struct cpu_evtchn_s {
|
||||
unsigned long bits[NR_EVENT_CHANNELS/BITS_PER_LONG];
|
||||
};
|
||||
static struct cpu_evtchn_s *cpu_evtchn_mask_p;
|
||||
|
||||
static __initdata struct cpu_evtchn_s init_evtchn_mask = {
|
||||
.bits[0 ... (NR_EVENT_CHANNELS/BITS_PER_LONG)-1] = ~0ul,
|
||||
};
|
||||
static struct cpu_evtchn_s *cpu_evtchn_mask_p = &init_evtchn_mask;
|
||||
|
||||
static inline unsigned long *cpu_evtchn_mask(int cpu)
|
||||
{
|
||||
return cpu_evtchn_mask_p[cpu].bits;
|
||||
@@ -113,6 +121,7 @@ static inline unsigned long *cpu_evtchn_mask(int cpu)
|
||||
|
||||
static struct irq_chip xen_dynamic_chip;
|
||||
static struct irq_chip xen_percpu_chip;
|
||||
static struct irq_chip xen_pirq_chip;
|
||||
|
||||
/* Constructor for packed IRQ information. */
|
||||
static struct irq_info mk_unbound_info(void)
|
||||
@@ -225,6 +234,15 @@ static unsigned int cpu_from_evtchn(unsigned int evtchn)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static bool pirq_needs_eoi(unsigned irq)
|
||||
{
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
return info->u.pirq.flags & PIRQ_NEEDS_EOI;
|
||||
}
|
||||
|
||||
static inline unsigned long active_evtchns(unsigned int cpu,
|
||||
struct shared_info *sh,
|
||||
unsigned int idx)
|
||||
@@ -336,36 +354,297 @@ static void unmask_evtchn(int port)
|
||||
put_cpu();
|
||||
}
|
||||
|
||||
static int get_nr_hw_irqs(void)
|
||||
{
|
||||
int ret = 1;
|
||||
|
||||
#ifdef CONFIG_X86_IO_APIC
|
||||
ret = get_nr_irqs_gsi();
|
||||
#endif
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int find_unbound_irq(void)
|
||||
{
|
||||
int irq;
|
||||
struct irq_desc *desc;
|
||||
struct irq_data *data;
|
||||
int irq, res;
|
||||
int start = get_nr_hw_irqs();
|
||||
|
||||
for (irq = 0; irq < nr_irqs; irq++) {
|
||||
desc = irq_to_desc(irq);
|
||||
if (start == nr_irqs)
|
||||
goto no_irqs;
|
||||
|
||||
/* nr_irqs is a magic value. Must not use it.*/
|
||||
for (irq = nr_irqs-1; irq > start; irq--) {
|
||||
data = irq_get_irq_data(irq);
|
||||
/* only 0->15 have init'd desc; handle irq > 16 */
|
||||
if (desc == NULL)
|
||||
if (!data)
|
||||
break;
|
||||
if (desc->chip == &no_irq_chip)
|
||||
if (data->chip == &no_irq_chip)
|
||||
break;
|
||||
if (desc->chip != &xen_dynamic_chip)
|
||||
if (data->chip != &xen_dynamic_chip)
|
||||
continue;
|
||||
if (irq_info[irq].type == IRQT_UNBOUND)
|
||||
break;
|
||||
return irq;
|
||||
}
|
||||
|
||||
if (irq == nr_irqs)
|
||||
panic("No available IRQ to bind to: increase nr_irqs!\n");
|
||||
if (irq == start)
|
||||
goto no_irqs;
|
||||
|
||||
desc = irq_to_desc_alloc_node(irq, 0);
|
||||
if (WARN_ON(desc == NULL))
|
||||
res = irq_alloc_desc_at(irq, 0);
|
||||
|
||||
if (WARN_ON(res != irq))
|
||||
return -1;
|
||||
|
||||
dynamic_irq_init_keep_chip_data(irq);
|
||||
return irq;
|
||||
|
||||
no_irqs:
|
||||
panic("No available IRQ to bind to: increase nr_irqs!\n");
|
||||
}
|
||||
|
||||
static bool identity_mapped_irq(unsigned irq)
|
||||
{
|
||||
/* identity map all the hardware irqs */
|
||||
return irq < get_nr_hw_irqs();
|
||||
}
|
||||
|
||||
static void pirq_unmask_notify(int irq)
|
||||
{
|
||||
struct physdev_eoi eoi = { .irq = irq };
|
||||
|
||||
if (unlikely(pirq_needs_eoi(irq))) {
|
||||
int rc = HYPERVISOR_physdev_op(PHYSDEVOP_eoi, &eoi);
|
||||
WARN_ON(rc);
|
||||
}
|
||||
}
|
||||
|
||||
static void pirq_query_unmask(int irq)
|
||||
{
|
||||
struct physdev_irq_status_query irq_status;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
irq_status.irq = irq;
|
||||
if (HYPERVISOR_physdev_op(PHYSDEVOP_irq_status_query, &irq_status))
|
||||
irq_status.flags = 0;
|
||||
|
||||
info->u.pirq.flags &= ~PIRQ_NEEDS_EOI;
|
||||
if (irq_status.flags & XENIRQSTAT_needs_eoi)
|
||||
info->u.pirq.flags |= PIRQ_NEEDS_EOI;
|
||||
}
|
||||
|
||||
static bool probing_irq(int irq)
|
||||
{
|
||||
struct irq_desc *desc = irq_to_desc(irq);
|
||||
|
||||
return desc && desc->action == NULL;
|
||||
}
|
||||
|
||||
static unsigned int startup_pirq(unsigned int irq)
|
||||
{
|
||||
struct evtchn_bind_pirq bind_pirq;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
int rc;
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
if (VALID_EVTCHN(evtchn))
|
||||
goto out;
|
||||
|
||||
bind_pirq.pirq = irq;
|
||||
/* NB. We are happy to share unless we are probing. */
|
||||
bind_pirq.flags = info->u.pirq.flags & PIRQ_SHAREABLE ?
|
||||
BIND_PIRQ__WILL_SHARE : 0;
|
||||
rc = HYPERVISOR_event_channel_op(EVTCHNOP_bind_pirq, &bind_pirq);
|
||||
if (rc != 0) {
|
||||
if (!probing_irq(irq))
|
||||
printk(KERN_INFO "Failed to obtain physical IRQ %d\n",
|
||||
irq);
|
||||
return 0;
|
||||
}
|
||||
evtchn = bind_pirq.port;
|
||||
|
||||
pirq_query_unmask(irq);
|
||||
|
||||
evtchn_to_irq[evtchn] = irq;
|
||||
bind_evtchn_to_cpu(evtchn, 0);
|
||||
info->evtchn = evtchn;
|
||||
|
||||
out:
|
||||
unmask_evtchn(evtchn);
|
||||
pirq_unmask_notify(irq);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void shutdown_pirq(unsigned int irq)
|
||||
{
|
||||
struct evtchn_close close;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
if (!VALID_EVTCHN(evtchn))
|
||||
return;
|
||||
|
||||
mask_evtchn(evtchn);
|
||||
|
||||
close.port = evtchn;
|
||||
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close) != 0)
|
||||
BUG();
|
||||
|
||||
bind_evtchn_to_cpu(evtchn, 0);
|
||||
evtchn_to_irq[evtchn] = -1;
|
||||
info->evtchn = 0;
|
||||
}
|
||||
|
||||
static void enable_pirq(unsigned int irq)
|
||||
{
|
||||
startup_pirq(irq);
|
||||
}
|
||||
|
||||
static void disable_pirq(unsigned int irq)
|
||||
{
|
||||
}
|
||||
|
||||
static void ack_pirq(unsigned int irq)
|
||||
{
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
|
||||
move_native_irq(irq);
|
||||
|
||||
if (VALID_EVTCHN(evtchn)) {
|
||||
mask_evtchn(evtchn);
|
||||
clear_evtchn(evtchn);
|
||||
}
|
||||
}
|
||||
|
||||
static void end_pirq(unsigned int irq)
|
||||
{
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
struct irq_desc *desc = irq_to_desc(irq);
|
||||
|
||||
if (WARN_ON(!desc))
|
||||
return;
|
||||
|
||||
if ((desc->status & (IRQ_DISABLED|IRQ_PENDING)) ==
|
||||
(IRQ_DISABLED|IRQ_PENDING)) {
|
||||
shutdown_pirq(irq);
|
||||
} else if (VALID_EVTCHN(evtchn)) {
|
||||
unmask_evtchn(evtchn);
|
||||
pirq_unmask_notify(irq);
|
||||
}
|
||||
}
|
||||
|
||||
static int find_irq_by_gsi(unsigned gsi)
|
||||
{
|
||||
int irq;
|
||||
|
||||
for (irq = 0; irq < nr_irqs; irq++) {
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
if (info == NULL || info->type != IRQT_PIRQ)
|
||||
continue;
|
||||
|
||||
if (gsi_from_irq(irq) == gsi)
|
||||
return irq;
|
||||
}
|
||||
|
||||
return -1;
|
||||
}
|
||||
|
||||
/* xen_allocate_irq might allocate irqs from the top down, as a
|
||||
* consequence don't assume that the irq number returned has a low value
|
||||
* or can be used as a pirq number unless you know otherwise.
|
||||
*
|
||||
* One notable exception is when xen_allocate_irq is called passing an
|
||||
* hardware gsi as argument, in that case the irq number returned
|
||||
* matches the gsi number passed as first argument.
|
||||
|
||||
* Note: We don't assign an
|
||||
* event channel until the irq actually started up. Return an
|
||||
* existing irq if we've already got one for the gsi.
|
||||
*/
|
||||
int xen_allocate_pirq(unsigned gsi, int shareable, char *name)
|
||||
{
|
||||
int irq;
|
||||
struct physdev_irq irq_op;
|
||||
|
||||
spin_lock(&irq_mapping_update_lock);
|
||||
|
||||
irq = find_irq_by_gsi(gsi);
|
||||
if (irq != -1) {
|
||||
printk(KERN_INFO "xen_allocate_pirq: returning irq %d for gsi %u\n",
|
||||
irq, gsi);
|
||||
goto out; /* XXX need refcount? */
|
||||
}
|
||||
|
||||
/* If we are a PV guest, we don't have GSIs (no ACPI passed). Therefore
|
||||
* we are using the !xen_initial_domain() to drop in the function.*/
|
||||
if (identity_mapped_irq(gsi) || !xen_initial_domain()) {
|
||||
irq = gsi;
|
||||
irq_alloc_desc_at(irq, 0);
|
||||
} else
|
||||
irq = find_unbound_irq();
|
||||
|
||||
set_irq_chip_and_handler_name(irq, &xen_pirq_chip,
|
||||
handle_level_irq, name);
|
||||
|
||||
irq_op.irq = irq;
|
||||
irq_op.vector = 0;
|
||||
|
||||
/* Only the privileged domain can do this. For non-priv, the pcifront
|
||||
* driver provides a PCI bus that does the call to do exactly
|
||||
* this in the priv domain. */
|
||||
if (xen_initial_domain() &&
|
||||
HYPERVISOR_physdev_op(PHYSDEVOP_alloc_irq_vector, &irq_op)) {
|
||||
irq_free_desc(irq);
|
||||
irq = -ENOSPC;
|
||||
goto out;
|
||||
}
|
||||
|
||||
irq_info[irq] = mk_pirq_info(0, gsi, irq_op.vector);
|
||||
irq_info[irq].u.pirq.flags |= shareable ? PIRQ_SHAREABLE : 0;
|
||||
|
||||
out:
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
|
||||
return irq;
|
||||
}
|
||||
|
||||
int xen_destroy_irq(int irq)
|
||||
{
|
||||
struct irq_desc *desc;
|
||||
int rc = -ENOENT;
|
||||
|
||||
spin_lock(&irq_mapping_update_lock);
|
||||
|
||||
desc = irq_to_desc(irq);
|
||||
if (!desc)
|
||||
goto out;
|
||||
|
||||
irq_info[irq] = mk_unbound_info();
|
||||
|
||||
irq_free_desc(irq);
|
||||
|
||||
out:
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
return rc;
|
||||
}
|
||||
|
||||
int xen_vector_from_irq(unsigned irq)
|
||||
{
|
||||
return vector_from_irq(irq);
|
||||
}
|
||||
|
||||
int xen_gsi_from_irq(unsigned irq)
|
||||
{
|
||||
return gsi_from_irq(irq);
|
||||
}
|
||||
|
||||
int bind_evtchn_to_irq(unsigned int evtchn)
|
||||
{
|
||||
int irq;
|
||||
@@ -495,7 +774,7 @@ static void unbind_from_irq(unsigned int irq)
|
||||
if (irq_info[irq].type != IRQT_UNBOUND) {
|
||||
irq_info[irq] = mk_unbound_info();
|
||||
|
||||
dynamic_irq_cleanup(irq);
|
||||
irq_free_desc(irq);
|
||||
}
|
||||
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
@@ -892,7 +1171,7 @@ void xen_clear_irq_pending(int irq)
|
||||
if (VALID_EVTCHN(evtchn))
|
||||
clear_evtchn(evtchn);
|
||||
}
|
||||
|
||||
EXPORT_SYMBOL(xen_clear_irq_pending);
|
||||
void xen_set_irq_pending(int irq)
|
||||
{
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
@@ -912,9 +1191,9 @@ bool xen_test_irq_pending(int irq)
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* Poll waiting for an irq to become pending. In the usual case, the
|
||||
irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq(int irq)
|
||||
/* Poll waiting for an irq to become pending with timeout. In the usual case,
|
||||
* the irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq_timeout(int irq, u64 timeout)
|
||||
{
|
||||
evtchn_port_t evtchn = evtchn_from_irq(irq);
|
||||
|
||||
@@ -922,13 +1201,20 @@ void xen_poll_irq(int irq)
|
||||
struct sched_poll poll;
|
||||
|
||||
poll.nr_ports = 1;
|
||||
poll.timeout = 0;
|
||||
poll.timeout = timeout;
|
||||
set_xen_guest_handle(poll.ports, &evtchn);
|
||||
|
||||
if (HYPERVISOR_sched_op(SCHEDOP_poll, &poll) != 0)
|
||||
BUG();
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL(xen_poll_irq_timeout);
|
||||
/* Poll waiting for an irq to become pending. In the usual case, the
|
||||
* irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq(int irq)
|
||||
{
|
||||
xen_poll_irq_timeout(irq, 0 /* no timeout */);
|
||||
}
|
||||
|
||||
void xen_irq_resume(void)
|
||||
{
|
||||
@@ -965,6 +1251,26 @@ static struct irq_chip xen_dynamic_chip __read_mostly = {
|
||||
.retrigger = retrigger_dynirq,
|
||||
};
|
||||
|
||||
static struct irq_chip xen_pirq_chip __read_mostly = {
|
||||
.name = "xen-pirq",
|
||||
|
||||
.startup = startup_pirq,
|
||||
.shutdown = shutdown_pirq,
|
||||
|
||||
.enable = enable_pirq,
|
||||
.unmask = enable_pirq,
|
||||
|
||||
.disable = disable_pirq,
|
||||
.mask = disable_pirq,
|
||||
|
||||
.ack = ack_pirq,
|
||||
.end = end_pirq,
|
||||
|
||||
.set_affinity = set_affinity_irq,
|
||||
|
||||
.retrigger = retrigger_dynirq,
|
||||
};
|
||||
|
||||
static struct irq_chip xen_percpu_chip __read_mostly = {
|
||||
.name = "xen-percpu",
|
||||
|
||||
@@ -1019,7 +1325,12 @@ void __init xen_init_IRQ(void)
|
||||
|
||||
cpu_evtchn_mask_p = kcalloc(nr_cpu_ids, sizeof(struct cpu_evtchn_s),
|
||||
GFP_KERNEL);
|
||||
BUG_ON(cpu_evtchn_mask_p == NULL);
|
||||
irq_info = kcalloc(nr_irqs, sizeof(*irq_info), GFP_KERNEL);
|
||||
|
||||
evtchn_to_irq = kcalloc(NR_EVENT_CHANNELS, sizeof(*evtchn_to_irq),
|
||||
GFP_KERNEL);
|
||||
for (i = 0; i < NR_EVENT_CHANNELS; i++)
|
||||
evtchn_to_irq[i] = -1;
|
||||
|
||||
init_evtchn_cpu_bindings();
|
||||
|
||||
|
||||
@@ -50,6 +50,8 @@ const char *xenbus_strstate(enum xenbus_state state)
|
||||
[ XenbusStateConnected ] = "Connected",
|
||||
[ XenbusStateClosing ] = "Closing",
|
||||
[ XenbusStateClosed ] = "Closed",
|
||||
[XenbusStateReconfiguring] = "Reconfiguring",
|
||||
[XenbusStateReconfigured] = "Reconfigured",
|
||||
};
|
||||
return (state < ARRAY_SIZE(name)) ? name[state] : "INVALID";
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user