|
@@ -120,7 +120,7 @@ int cpu_check_affinity(unsigned int irq, cpumask_t *dest)
|
|
if (CHECK_IRQ_PER_CPU(irq)) {
|
|
if (CHECK_IRQ_PER_CPU(irq)) {
|
|
/* Bad linux design decision. The mask has already
|
|
/* Bad linux design decision. The mask has already
|
|
* been set; we must reset it */
|
|
* been set; we must reset it */
|
|
- irq_desc[irq].affinity = CPU_MASK_ALL;
|
|
|
|
|
|
+ cpumask_setall(irq_desc[irq].affinity);
|
|
return -EINVAL;
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -136,7 +136,7 @@ static void cpu_set_affinity_irq(unsigned int irq, const struct cpumask *dest)
|
|
if (cpu_check_affinity(irq, dest))
|
|
if (cpu_check_affinity(irq, dest))
|
|
return;
|
|
return;
|
|
|
|
|
|
- irq_desc[irq].affinity = *dest;
|
|
|
|
|
|
+ cpumask_copy(irq_desc[irq].affinity, dest);
|
|
}
|
|
}
|
|
#endif
|
|
#endif
|
|
|
|
|
|
@@ -295,7 +295,7 @@ int txn_alloc_irq(unsigned int bits_wide)
|
|
unsigned long txn_affinity_addr(unsigned int irq, int cpu)
|
|
unsigned long txn_affinity_addr(unsigned int irq, int cpu)
|
|
{
|
|
{
|
|
#ifdef CONFIG_SMP
|
|
#ifdef CONFIG_SMP
|
|
- irq_desc[irq].affinity = cpumask_of_cpu(cpu);
|
|
|
|
|
|
+ cpumask_copy(irq_desc[irq].affinity, cpumask_of(cpu));
|
|
#endif
|
|
#endif
|
|
|
|
|
|
return per_cpu(cpu_data, cpu).txn_addr;
|
|
return per_cpu(cpu_data, cpu).txn_addr;
|
|
@@ -352,7 +352,7 @@ void do_cpu_irq_mask(struct pt_regs *regs)
|
|
irq = eirr_to_irq(eirr_val);
|
|
irq = eirr_to_irq(eirr_val);
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
#ifdef CONFIG_SMP
|
|
- dest = irq_desc[irq].affinity;
|
|
|
|
|
|
+ cpumask_copy(&dest, irq_desc[irq].affinity);
|
|
if (CHECK_IRQ_PER_CPU(irq_desc[irq].status) &&
|
|
if (CHECK_IRQ_PER_CPU(irq_desc[irq].status) &&
|
|
!cpu_isset(smp_processor_id(), dest)) {
|
|
!cpu_isset(smp_processor_id(), dest)) {
|
|
int cpu = first_cpu(dest);
|
|
int cpu = first_cpu(dest);
|