|
@@ -46,58 +46,31 @@ static __init int setup_intremap(char *str)
|
|
|
}
|
|
|
early_param("intremap", setup_intremap);
|
|
|
|
|
|
+static DEFINE_SPINLOCK(irq_2_ir_lock);
|
|
|
+
|
|
|
static struct irq_2_iommu *irq_2_iommu(unsigned int irq)
|
|
|
{
|
|
|
struct irq_cfg *cfg = get_irq_chip_data(irq);
|
|
|
return cfg ? &cfg->irq_2_iommu : NULL;
|
|
|
}
|
|
|
|
|
|
-static struct irq_2_iommu *irq_2_iommu_alloc(unsigned int irq)
|
|
|
-{
|
|
|
- return irq_2_iommu(irq);
|
|
|
-}
|
|
|
-
|
|
|
-static void irq_2_iommu_free(unsigned int irq)
|
|
|
-{
|
|
|
-}
|
|
|
-
|
|
|
-static DEFINE_SPINLOCK(irq_2_ir_lock);
|
|
|
-
|
|
|
-static struct irq_2_iommu *valid_irq_2_iommu(unsigned int irq)
|
|
|
-{
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
-
|
|
|
- irq_iommu = irq_2_iommu(irq);
|
|
|
-
|
|
|
- if (!irq_iommu)
|
|
|
- return NULL;
|
|
|
-
|
|
|
- if (!irq_iommu->iommu)
|
|
|
- return NULL;
|
|
|
-
|
|
|
- return irq_iommu;
|
|
|
-}
|
|
|
-
|
|
|
int irq_remapped(int irq)
|
|
|
{
|
|
|
- return valid_irq_2_iommu(irq) != NULL;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
+
|
|
|
+ return irq_iommu ? irq_iommu->iommu != NULL : 0;
|
|
|
}
|
|
|
|
|
|
int get_irte(int irq, struct irte *entry)
|
|
|
{
|
|
|
- int index;
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
unsigned long flags;
|
|
|
+ int index;
|
|
|
|
|
|
- if (!entry)
|
|
|
+ if (!entry || !irq_iommu)
|
|
|
return -1;
|
|
|
|
|
|
spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
- irq_iommu = valid_irq_2_iommu(irq);
|
|
|
- if (!irq_iommu) {
|
|
|
- spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
- return -1;
|
|
|
- }
|
|
|
|
|
|
index = irq_iommu->irte_index + irq_iommu->sub_handle;
|
|
|
*entry = *(irq_iommu->iommu->ir_table->base + index);
|
|
@@ -109,20 +82,14 @@ int get_irte(int irq, struct irte *entry)
|
|
|
int alloc_irte(struct intel_iommu *iommu, int irq, u16 count)
|
|
|
{
|
|
|
struct ir_table *table = iommu->ir_table;
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
u16 index, start_index;
|
|
|
unsigned int mask = 0;
|
|
|
unsigned long flags;
|
|
|
int i;
|
|
|
|
|
|
- if (!count)
|
|
|
- return -1;
|
|
|
-
|
|
|
-#ifndef CONFIG_SPARSE_IRQ
|
|
|
- /* protect irq_2_iommu_alloc later */
|
|
|
- if (irq >= nr_irqs)
|
|
|
+ if (!count || !irq_iommu)
|
|
|
return -1;
|
|
|
-#endif
|
|
|
|
|
|
/*
|
|
|
* start the IRTE search from index 0.
|
|
@@ -163,13 +130,6 @@ int alloc_irte(struct intel_iommu *iommu, int irq, u16 count)
|
|
|
for (i = index; i < index + count; i++)
|
|
|
table->base[i].present = 1;
|
|
|
|
|
|
- irq_iommu = irq_2_iommu_alloc(irq);
|
|
|
- if (!irq_iommu) {
|
|
|
- spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
- printk(KERN_ERR "can't allocate irq_2_iommu\n");
|
|
|
- return -1;
|
|
|
- }
|
|
|
-
|
|
|
irq_iommu->iommu = iommu;
|
|
|
irq_iommu->irte_index = index;
|
|
|
irq_iommu->sub_handle = 0;
|
|
@@ -193,17 +153,14 @@ static int qi_flush_iec(struct intel_iommu *iommu, int index, int mask)
|
|
|
|
|
|
int map_irq_to_irte_handle(int irq, u16 *sub_handle)
|
|
|
{
|
|
|
- int index;
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
unsigned long flags;
|
|
|
+ int index;
|
|
|
|
|
|
- spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
- irq_iommu = valid_irq_2_iommu(irq);
|
|
|
- if (!irq_iommu) {
|
|
|
- spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
+ if (!irq_iommu)
|
|
|
return -1;
|
|
|
- }
|
|
|
|
|
|
+ spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
*sub_handle = irq_iommu->sub_handle;
|
|
|
index = irq_iommu->irte_index;
|
|
|
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
@@ -212,18 +169,13 @@ int map_irq_to_irte_handle(int irq, u16 *sub_handle)
|
|
|
|
|
|
int set_irte_irq(int irq, struct intel_iommu *iommu, u16 index, u16 subhandle)
|
|
|
{
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
unsigned long flags;
|
|
|
|
|
|
- spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
-
|
|
|
- irq_iommu = irq_2_iommu_alloc(irq);
|
|
|
-
|
|
|
- if (!irq_iommu) {
|
|
|
- spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
- printk(KERN_ERR "can't allocate irq_2_iommu\n");
|
|
|
+ if (!irq_iommu)
|
|
|
return -1;
|
|
|
- }
|
|
|
+
|
|
|
+ spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
|
|
|
irq_iommu->iommu = iommu;
|
|
|
irq_iommu->irte_index = index;
|
|
@@ -237,19 +189,16 @@ int set_irte_irq(int irq, struct intel_iommu *iommu, u16 index, u16 subhandle)
|
|
|
|
|
|
int modify_irte(int irq, struct irte *irte_modified)
|
|
|
{
|
|
|
- int rc;
|
|
|
- int index;
|
|
|
- struct irte *irte;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
struct intel_iommu *iommu;
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
unsigned long flags;
|
|
|
+ struct irte *irte;
|
|
|
+ int rc, index;
|
|
|
|
|
|
- spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
- irq_iommu = valid_irq_2_iommu(irq);
|
|
|
- if (!irq_iommu) {
|
|
|
- spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
+ if (!irq_iommu)
|
|
|
return -1;
|
|
|
- }
|
|
|
+
|
|
|
+ spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
|
|
|
iommu = irq_iommu->iommu;
|
|
|
|
|
@@ -322,16 +271,14 @@ static int clear_entries(struct irq_2_iommu *irq_iommu)
|
|
|
|
|
|
int free_irte(int irq)
|
|
|
{
|
|
|
- int rc = 0;
|
|
|
- struct irq_2_iommu *irq_iommu;
|
|
|
+ struct irq_2_iommu *irq_iommu = irq_2_iommu(irq);
|
|
|
unsigned long flags;
|
|
|
+ int rc;
|
|
|
|
|
|
- spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
- irq_iommu = valid_irq_2_iommu(irq);
|
|
|
- if (!irq_iommu) {
|
|
|
- spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
+ if (!irq_iommu)
|
|
|
return -1;
|
|
|
- }
|
|
|
+
|
|
|
+ spin_lock_irqsave(&irq_2_ir_lock, flags);
|
|
|
|
|
|
rc = clear_entries(irq_iommu);
|
|
|
|
|
@@ -342,8 +289,6 @@ int free_irte(int irq)
|
|
|
|
|
|
spin_unlock_irqrestore(&irq_2_ir_lock, flags);
|
|
|
|
|
|
- irq_2_iommu_free(irq);
|
|
|
-
|
|
|
return rc;
|
|
|
}
|
|
|
|