ints-priority.c 33 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446
  1. /*
  2. * Set up the interrupt priorities
  3. *
  4. * Copyright 2004-2009 Analog Devices Inc.
  5. * 2003 Bas Vermeulen <bas@buyways.nl>
  6. * 2002 Arcturus Networks Inc. MaTed <mated@sympatico.ca>
  7. * 2000-2001 Lineo, Inc. D. Jefff Dionne <jeff@lineo.ca>
  8. * 1999 D. Jeff Dionne <jeff@uclinux.org>
  9. * 1996 Roman Zippel
  10. *
  11. * Licensed under the GPL-2
  12. */
  13. #include <linux/module.h>
  14. #include <linux/kernel_stat.h>
  15. #include <linux/seq_file.h>
  16. #include <linux/irq.h>
  17. #ifdef CONFIG_IPIPE
  18. #include <linux/ipipe.h>
  19. #endif
  20. #ifdef CONFIG_KGDB
  21. #include <linux/kgdb.h>
  22. #endif
  23. #include <asm/traps.h>
  24. #include <asm/blackfin.h>
  25. #include <asm/gpio.h>
  26. #include <asm/irq_handler.h>
  27. #include <asm/dpmc.h>
  28. #include <asm/bfin5xx_spi.h>
  29. #include <asm/bfin_sport.h>
  30. #include <asm/bfin_can.h>
  31. #define SIC_SYSIRQ(irq) (irq - (IRQ_CORETMR + 1))
  32. #ifdef BF537_FAMILY
  33. # define BF537_GENERIC_ERROR_INT_DEMUX
  34. # define SPI_ERR_MASK (BIT_STAT_TXCOL | BIT_STAT_RBSY | BIT_STAT_MODF | BIT_STAT_TXE) /* SPI_STAT */
  35. # define SPORT_ERR_MASK (ROVF | RUVF | TOVF | TUVF) /* SPORT_STAT */
  36. # define PPI_ERR_MASK (0xFFFF & ~FLD) /* PPI_STATUS */
  37. # define EMAC_ERR_MASK (PHYINT | MMCINT | RXFSINT | TXFSINT | WAKEDET | RXDMAERR | TXDMAERR | STMDONE) /* EMAC_SYSTAT */
  38. # define UART_ERR_MASK (0x6) /* UART_IIR */
  39. # define CAN_ERR_MASK (EWTIF | EWRIF | EPIF | BOIF | WUIF | UIAIF | AAIF | RMLIF | UCEIF | EXTIF | ADIF) /* CAN_GIF */
  40. #else
  41. # undef BF537_GENERIC_ERROR_INT_DEMUX
  42. #endif
  43. /*
  44. * NOTES:
  45. * - we have separated the physical Hardware interrupt from the
  46. * levels that the LINUX kernel sees (see the description in irq.h)
  47. * -
  48. */
  49. #ifndef CONFIG_SMP
  50. /* Initialize this to an actual value to force it into the .data
  51. * section so that we know it is properly initialized at entry into
  52. * the kernel but before bss is initialized to zero (which is where
  53. * it would live otherwise). The 0x1f magic represents the IRQs we
  54. * cannot actually mask out in hardware.
  55. */
  56. unsigned long bfin_irq_flags = 0x1f;
  57. EXPORT_SYMBOL(bfin_irq_flags);
  58. #endif
  59. /* The number of spurious interrupts */
  60. atomic_t num_spurious;
  61. #ifdef CONFIG_PM
  62. unsigned long bfin_sic_iwr[3]; /* Up to 3 SIC_IWRx registers */
  63. unsigned vr_wakeup;
  64. #endif
  65. struct ivgx {
  66. /* irq number for request_irq, available in mach-bf5xx/irq.h */
  67. unsigned int irqno;
  68. /* corresponding bit in the SIC_ISR register */
  69. unsigned int isrflag;
  70. } ivg_table[NR_PERI_INTS];
  71. struct ivg_slice {
  72. /* position of first irq in ivg_table for given ivg */
  73. struct ivgx *ifirst;
  74. struct ivgx *istop;
  75. } ivg7_13[IVG13 - IVG7 + 1];
  76. /*
  77. * Search SIC_IAR and fill tables with the irqvalues
  78. * and their positions in the SIC_ISR register.
  79. */
  80. static void __init search_IAR(void)
  81. {
  82. unsigned ivg, irq_pos = 0;
  83. for (ivg = 0; ivg <= IVG13 - IVG7; ivg++) {
  84. int irqN;
  85. ivg7_13[ivg].istop = ivg7_13[ivg].ifirst = &ivg_table[irq_pos];
  86. for (irqN = 0; irqN < NR_PERI_INTS; irqN += 4) {
  87. int irqn;
  88. u32 iar = bfin_read32((unsigned long *)SIC_IAR0 +
  89. #if defined(CONFIG_BF51x) || defined(CONFIG_BF52x) || \
  90. defined(CONFIG_BF538) || defined(CONFIG_BF539)
  91. ((irqN % 32) >> 3) + ((irqN / 32) * ((SIC_IAR4 - SIC_IAR0) / 4))
  92. #else
  93. (irqN >> 3)
  94. #endif
  95. );
  96. for (irqn = irqN; irqn < irqN + 4; ++irqn) {
  97. int iar_shift = (irqn & 7) * 4;
  98. if (ivg == (0xf & (iar >> iar_shift))) {
  99. ivg_table[irq_pos].irqno = IVG7 + irqn;
  100. ivg_table[irq_pos].isrflag = 1 << (irqn % 32);
  101. ivg7_13[ivg].istop++;
  102. irq_pos++;
  103. }
  104. }
  105. }
  106. }
  107. }
  108. /*
  109. * This is for core internal IRQs
  110. */
  111. static void bfin_ack_noop(unsigned int irq)
  112. {
  113. /* Dummy function. */
  114. }
  115. static void bfin_core_mask_irq(unsigned int irq)
  116. {
  117. bfin_irq_flags &= ~(1 << irq);
  118. if (!irqs_disabled_hw())
  119. local_irq_enable_hw();
  120. }
  121. static void bfin_core_unmask_irq(unsigned int irq)
  122. {
  123. bfin_irq_flags |= 1 << irq;
  124. /*
  125. * If interrupts are enabled, IMASK must contain the same value
  126. * as bfin_irq_flags. Make sure that invariant holds. If interrupts
  127. * are currently disabled we need not do anything; one of the
  128. * callers will take care of setting IMASK to the proper value
  129. * when reenabling interrupts.
  130. * local_irq_enable just does "STI bfin_irq_flags", so it's exactly
  131. * what we need.
  132. */
  133. if (!irqs_disabled_hw())
  134. local_irq_enable_hw();
  135. return;
  136. }
  137. static void bfin_internal_mask_irq(unsigned int irq)
  138. {
  139. unsigned long flags;
  140. #ifdef CONFIG_BF53x
  141. local_irq_save_hw(flags);
  142. bfin_write_SIC_IMASK(bfin_read_SIC_IMASK() &
  143. ~(1 << SIC_SYSIRQ(irq)));
  144. #else
  145. unsigned mask_bank, mask_bit;
  146. local_irq_save_hw(flags);
  147. mask_bank = SIC_SYSIRQ(irq) / 32;
  148. mask_bit = SIC_SYSIRQ(irq) % 32;
  149. bfin_write_SIC_IMASK(mask_bank, bfin_read_SIC_IMASK(mask_bank) &
  150. ~(1 << mask_bit));
  151. #ifdef CONFIG_SMP
  152. bfin_write_SICB_IMASK(mask_bank, bfin_read_SICB_IMASK(mask_bank) &
  153. ~(1 << mask_bit));
  154. #endif
  155. #endif
  156. local_irq_restore_hw(flags);
  157. }
  158. #ifdef CONFIG_SMP
  159. static void bfin_internal_unmask_irq_affinity(unsigned int irq,
  160. const struct cpumask *affinity)
  161. #else
  162. static void bfin_internal_unmask_irq(unsigned int irq)
  163. #endif
  164. {
  165. unsigned long flags;
  166. #ifdef CONFIG_BF53x
  167. local_irq_save_hw(flags);
  168. bfin_write_SIC_IMASK(bfin_read_SIC_IMASK() |
  169. (1 << SIC_SYSIRQ(irq)));
  170. #else
  171. unsigned mask_bank, mask_bit;
  172. local_irq_save_hw(flags);
  173. mask_bank = SIC_SYSIRQ(irq) / 32;
  174. mask_bit = SIC_SYSIRQ(irq) % 32;
  175. #ifdef CONFIG_SMP
  176. if (cpumask_test_cpu(0, affinity))
  177. #endif
  178. bfin_write_SIC_IMASK(mask_bank,
  179. bfin_read_SIC_IMASK(mask_bank) |
  180. (1 << mask_bit));
  181. #ifdef CONFIG_SMP
  182. if (cpumask_test_cpu(1, affinity))
  183. bfin_write_SICB_IMASK(mask_bank,
  184. bfin_read_SICB_IMASK(mask_bank) |
  185. (1 << mask_bit));
  186. #endif
  187. #endif
  188. local_irq_restore_hw(flags);
  189. }
  190. #ifdef CONFIG_SMP
  191. static void bfin_internal_unmask_irq(unsigned int irq)
  192. {
  193. struct irq_desc *desc = irq_to_desc(irq);
  194. bfin_internal_unmask_irq_affinity(irq, desc->affinity);
  195. }
  196. static int bfin_internal_set_affinity(unsigned int irq, const struct cpumask *mask)
  197. {
  198. bfin_internal_mask_irq(irq);
  199. bfin_internal_unmask_irq_affinity(irq, mask);
  200. return 0;
  201. }
  202. #endif
  203. #ifdef CONFIG_PM
  204. int bfin_internal_set_wake(unsigned int irq, unsigned int state)
  205. {
  206. u32 bank, bit, wakeup = 0;
  207. unsigned long flags;
  208. bank = SIC_SYSIRQ(irq) / 32;
  209. bit = SIC_SYSIRQ(irq) % 32;
  210. switch (irq) {
  211. #ifdef IRQ_RTC
  212. case IRQ_RTC:
  213. wakeup |= WAKE;
  214. break;
  215. #endif
  216. #ifdef IRQ_CAN0_RX
  217. case IRQ_CAN0_RX:
  218. wakeup |= CANWE;
  219. break;
  220. #endif
  221. #ifdef IRQ_CAN1_RX
  222. case IRQ_CAN1_RX:
  223. wakeup |= CANWE;
  224. break;
  225. #endif
  226. #ifdef IRQ_USB_INT0
  227. case IRQ_USB_INT0:
  228. wakeup |= USBWE;
  229. break;
  230. #endif
  231. #ifdef CONFIG_BF54x
  232. case IRQ_CNT:
  233. wakeup |= ROTWE;
  234. break;
  235. #endif
  236. default:
  237. break;
  238. }
  239. local_irq_save_hw(flags);
  240. if (state) {
  241. bfin_sic_iwr[bank] |= (1 << bit);
  242. vr_wakeup |= wakeup;
  243. } else {
  244. bfin_sic_iwr[bank] &= ~(1 << bit);
  245. vr_wakeup &= ~wakeup;
  246. }
  247. local_irq_restore_hw(flags);
  248. return 0;
  249. }
  250. #endif
  251. static struct irq_chip bfin_core_irqchip = {
  252. .name = "CORE",
  253. .ack = bfin_ack_noop,
  254. .mask = bfin_core_mask_irq,
  255. .unmask = bfin_core_unmask_irq,
  256. };
  257. static struct irq_chip bfin_internal_irqchip = {
  258. .name = "INTN",
  259. .ack = bfin_ack_noop,
  260. .mask = bfin_internal_mask_irq,
  261. .unmask = bfin_internal_unmask_irq,
  262. .mask_ack = bfin_internal_mask_irq,
  263. .disable = bfin_internal_mask_irq,
  264. .enable = bfin_internal_unmask_irq,
  265. #ifdef CONFIG_SMP
  266. .set_affinity = bfin_internal_set_affinity,
  267. #endif
  268. #ifdef CONFIG_PM
  269. .set_wake = bfin_internal_set_wake,
  270. #endif
  271. };
  272. static void bfin_handle_irq(unsigned irq)
  273. {
  274. #ifdef CONFIG_IPIPE
  275. struct pt_regs regs; /* Contents not used. */
  276. ipipe_trace_irq_entry(irq);
  277. __ipipe_handle_irq(irq, &regs);
  278. ipipe_trace_irq_exit(irq);
  279. #else /* !CONFIG_IPIPE */
  280. struct irq_desc *desc = irq_desc + irq;
  281. desc->handle_irq(irq, desc);
  282. #endif /* !CONFIG_IPIPE */
  283. }
  284. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  285. static int error_int_mask;
  286. static void bfin_generic_error_mask_irq(unsigned int irq)
  287. {
  288. error_int_mask &= ~(1L << (irq - IRQ_PPI_ERROR));
  289. if (!error_int_mask)
  290. bfin_internal_mask_irq(IRQ_GENERIC_ERROR);
  291. }
  292. static void bfin_generic_error_unmask_irq(unsigned int irq)
  293. {
  294. bfin_internal_unmask_irq(IRQ_GENERIC_ERROR);
  295. error_int_mask |= 1L << (irq - IRQ_PPI_ERROR);
  296. }
  297. static struct irq_chip bfin_generic_error_irqchip = {
  298. .name = "ERROR",
  299. .ack = bfin_ack_noop,
  300. .mask_ack = bfin_generic_error_mask_irq,
  301. .mask = bfin_generic_error_mask_irq,
  302. .unmask = bfin_generic_error_unmask_irq,
  303. };
  304. static void bfin_demux_error_irq(unsigned int int_err_irq,
  305. struct irq_desc *inta_desc)
  306. {
  307. int irq = 0;
  308. #if (defined(CONFIG_BF537) || defined(CONFIG_BF536))
  309. if (bfin_read_EMAC_SYSTAT() & EMAC_ERR_MASK)
  310. irq = IRQ_MAC_ERROR;
  311. else
  312. #endif
  313. if (bfin_read_SPORT0_STAT() & SPORT_ERR_MASK)
  314. irq = IRQ_SPORT0_ERROR;
  315. else if (bfin_read_SPORT1_STAT() & SPORT_ERR_MASK)
  316. irq = IRQ_SPORT1_ERROR;
  317. else if (bfin_read_PPI_STATUS() & PPI_ERR_MASK)
  318. irq = IRQ_PPI_ERROR;
  319. else if (bfin_read_CAN_GIF() & CAN_ERR_MASK)
  320. irq = IRQ_CAN_ERROR;
  321. else if (bfin_read_SPI_STAT() & SPI_ERR_MASK)
  322. irq = IRQ_SPI_ERROR;
  323. else if ((bfin_read_UART0_IIR() & UART_ERR_MASK) == UART_ERR_MASK)
  324. irq = IRQ_UART0_ERROR;
  325. else if ((bfin_read_UART1_IIR() & UART_ERR_MASK) == UART_ERR_MASK)
  326. irq = IRQ_UART1_ERROR;
  327. if (irq) {
  328. if (error_int_mask & (1L << (irq - IRQ_PPI_ERROR)))
  329. bfin_handle_irq(irq);
  330. else {
  331. switch (irq) {
  332. case IRQ_PPI_ERROR:
  333. bfin_write_PPI_STATUS(PPI_ERR_MASK);
  334. break;
  335. #if (defined(CONFIG_BF537) || defined(CONFIG_BF536))
  336. case IRQ_MAC_ERROR:
  337. bfin_write_EMAC_SYSTAT(EMAC_ERR_MASK);
  338. break;
  339. #endif
  340. case IRQ_SPORT0_ERROR:
  341. bfin_write_SPORT0_STAT(SPORT_ERR_MASK);
  342. break;
  343. case IRQ_SPORT1_ERROR:
  344. bfin_write_SPORT1_STAT(SPORT_ERR_MASK);
  345. break;
  346. case IRQ_CAN_ERROR:
  347. bfin_write_CAN_GIS(CAN_ERR_MASK);
  348. break;
  349. case IRQ_SPI_ERROR:
  350. bfin_write_SPI_STAT(SPI_ERR_MASK);
  351. break;
  352. default:
  353. break;
  354. }
  355. pr_debug("IRQ %d:"
  356. " MASKED PERIPHERAL ERROR INTERRUPT ASSERTED\n",
  357. irq);
  358. }
  359. } else
  360. printk(KERN_ERR
  361. "%s : %s : LINE %d :\nIRQ ?: PERIPHERAL ERROR"
  362. " INTERRUPT ASSERTED BUT NO SOURCE FOUND\n",
  363. __func__, __FILE__, __LINE__);
  364. }
  365. #endif /* BF537_GENERIC_ERROR_INT_DEMUX */
  366. #if defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE)
  367. static int mac_stat_int_mask;
  368. static void bfin_mac_status_ack_irq(unsigned int irq)
  369. {
  370. switch (irq) {
  371. case IRQ_MAC_MMCINT:
  372. bfin_write_EMAC_MMC_TIRQS(
  373. bfin_read_EMAC_MMC_TIRQE() &
  374. bfin_read_EMAC_MMC_TIRQS());
  375. bfin_write_EMAC_MMC_RIRQS(
  376. bfin_read_EMAC_MMC_RIRQE() &
  377. bfin_read_EMAC_MMC_RIRQS());
  378. break;
  379. case IRQ_MAC_RXFSINT:
  380. bfin_write_EMAC_RX_STKY(
  381. bfin_read_EMAC_RX_IRQE() &
  382. bfin_read_EMAC_RX_STKY());
  383. break;
  384. case IRQ_MAC_TXFSINT:
  385. bfin_write_EMAC_TX_STKY(
  386. bfin_read_EMAC_TX_IRQE() &
  387. bfin_read_EMAC_TX_STKY());
  388. break;
  389. case IRQ_MAC_WAKEDET:
  390. bfin_write_EMAC_WKUP_CTL(
  391. bfin_read_EMAC_WKUP_CTL() | MPKS | RWKS);
  392. break;
  393. default:
  394. /* These bits are W1C */
  395. bfin_write_EMAC_SYSTAT(1L << (irq - IRQ_MAC_PHYINT));
  396. break;
  397. }
  398. }
  399. static void bfin_mac_status_mask_irq(unsigned int irq)
  400. {
  401. mac_stat_int_mask &= ~(1L << (irq - IRQ_MAC_PHYINT));
  402. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  403. switch (irq) {
  404. case IRQ_MAC_PHYINT:
  405. bfin_write_EMAC_SYSCTL(bfin_read_EMAC_SYSCTL() & ~PHYIE);
  406. break;
  407. default:
  408. break;
  409. }
  410. #else
  411. if (!mac_stat_int_mask)
  412. bfin_internal_mask_irq(IRQ_MAC_ERROR);
  413. #endif
  414. bfin_mac_status_ack_irq(irq);
  415. }
  416. static void bfin_mac_status_unmask_irq(unsigned int irq)
  417. {
  418. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  419. switch (irq) {
  420. case IRQ_MAC_PHYINT:
  421. bfin_write_EMAC_SYSCTL(bfin_read_EMAC_SYSCTL() | PHYIE);
  422. break;
  423. default:
  424. break;
  425. }
  426. #else
  427. if (!mac_stat_int_mask)
  428. bfin_internal_unmask_irq(IRQ_MAC_ERROR);
  429. #endif
  430. mac_stat_int_mask |= 1L << (irq - IRQ_MAC_PHYINT);
  431. }
  432. #ifdef CONFIG_PM
  433. int bfin_mac_status_set_wake(unsigned int irq, unsigned int state)
  434. {
  435. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  436. return bfin_internal_set_wake(IRQ_GENERIC_ERROR, state);
  437. #else
  438. return bfin_internal_set_wake(IRQ_MAC_ERROR, state);
  439. #endif
  440. }
  441. #endif
  442. static struct irq_chip bfin_mac_status_irqchip = {
  443. .name = "MACST",
  444. .ack = bfin_ack_noop,
  445. .mask_ack = bfin_mac_status_mask_irq,
  446. .mask = bfin_mac_status_mask_irq,
  447. .unmask = bfin_mac_status_unmask_irq,
  448. #ifdef CONFIG_PM
  449. .set_wake = bfin_mac_status_set_wake,
  450. #endif
  451. };
  452. static void bfin_demux_mac_status_irq(unsigned int int_err_irq,
  453. struct irq_desc *inta_desc)
  454. {
  455. int i, irq = 0;
  456. u32 status = bfin_read_EMAC_SYSTAT();
  457. for (i = 0; i < (IRQ_MAC_STMDONE - IRQ_MAC_PHYINT); i++)
  458. if (status & (1L << i)) {
  459. irq = IRQ_MAC_PHYINT + i;
  460. break;
  461. }
  462. if (irq) {
  463. if (mac_stat_int_mask & (1L << (irq - IRQ_MAC_PHYINT))) {
  464. bfin_handle_irq(irq);
  465. } else {
  466. bfin_mac_status_ack_irq(irq);
  467. pr_debug("IRQ %d:"
  468. " MASKED MAC ERROR INTERRUPT ASSERTED\n",
  469. irq);
  470. }
  471. } else
  472. printk(KERN_ERR
  473. "%s : %s : LINE %d :\nIRQ ?: MAC ERROR"
  474. " INTERRUPT ASSERTED BUT NO SOURCE FOUND\n",
  475. __func__, __FILE__, __LINE__);
  476. }
  477. #endif
  478. static inline void bfin_set_irq_handler(unsigned irq, irq_flow_handler_t handle)
  479. {
  480. #ifdef CONFIG_IPIPE
  481. _set_irq_handler(irq, handle_level_irq);
  482. #else
  483. struct irq_desc *desc = irq_desc + irq;
  484. /* May not call generic set_irq_handler() due to spinlock
  485. recursion. */
  486. desc->handle_irq = handle;
  487. #endif
  488. }
  489. static DECLARE_BITMAP(gpio_enabled, MAX_BLACKFIN_GPIOS);
  490. extern void bfin_gpio_irq_prepare(unsigned gpio);
  491. #if !defined(CONFIG_BF54x)
  492. static void bfin_gpio_ack_irq(unsigned int irq)
  493. {
  494. /* AFAIK ack_irq in case mask_ack is provided
  495. * get's only called for edge sense irqs
  496. */
  497. set_gpio_data(irq_to_gpio(irq), 0);
  498. }
  499. static void bfin_gpio_mask_ack_irq(unsigned int irq)
  500. {
  501. struct irq_desc *desc = irq_desc + irq;
  502. u32 gpionr = irq_to_gpio(irq);
  503. if (desc->handle_irq == handle_edge_irq)
  504. set_gpio_data(gpionr, 0);
  505. set_gpio_maska(gpionr, 0);
  506. }
  507. static void bfin_gpio_mask_irq(unsigned int irq)
  508. {
  509. set_gpio_maska(irq_to_gpio(irq), 0);
  510. }
  511. static void bfin_gpio_unmask_irq(unsigned int irq)
  512. {
  513. set_gpio_maska(irq_to_gpio(irq), 1);
  514. }
  515. static unsigned int bfin_gpio_irq_startup(unsigned int irq)
  516. {
  517. u32 gpionr = irq_to_gpio(irq);
  518. if (__test_and_set_bit(gpionr, gpio_enabled))
  519. bfin_gpio_irq_prepare(gpionr);
  520. bfin_gpio_unmask_irq(irq);
  521. return 0;
  522. }
  523. static void bfin_gpio_irq_shutdown(unsigned int irq)
  524. {
  525. u32 gpionr = irq_to_gpio(irq);
  526. bfin_gpio_mask_irq(irq);
  527. __clear_bit(gpionr, gpio_enabled);
  528. bfin_gpio_irq_free(gpionr);
  529. }
  530. static int bfin_gpio_irq_type(unsigned int irq, unsigned int type)
  531. {
  532. int ret;
  533. char buf[16];
  534. u32 gpionr = irq_to_gpio(irq);
  535. if (type == IRQ_TYPE_PROBE) {
  536. /* only probe unenabled GPIO interrupt lines */
  537. if (test_bit(gpionr, gpio_enabled))
  538. return 0;
  539. type = IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING;
  540. }
  541. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING |
  542. IRQ_TYPE_LEVEL_HIGH | IRQ_TYPE_LEVEL_LOW)) {
  543. snprintf(buf, 16, "gpio-irq%d", irq);
  544. ret = bfin_gpio_irq_request(gpionr, buf);
  545. if (ret)
  546. return ret;
  547. if (__test_and_set_bit(gpionr, gpio_enabled))
  548. bfin_gpio_irq_prepare(gpionr);
  549. } else {
  550. __clear_bit(gpionr, gpio_enabled);
  551. return 0;
  552. }
  553. set_gpio_inen(gpionr, 0);
  554. set_gpio_dir(gpionr, 0);
  555. if ((type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  556. == (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  557. set_gpio_both(gpionr, 1);
  558. else
  559. set_gpio_both(gpionr, 0);
  560. if ((type & (IRQ_TYPE_EDGE_FALLING | IRQ_TYPE_LEVEL_LOW)))
  561. set_gpio_polar(gpionr, 1); /* low or falling edge denoted by one */
  562. else
  563. set_gpio_polar(gpionr, 0); /* high or rising edge denoted by zero */
  564. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  565. set_gpio_edge(gpionr, 1);
  566. set_gpio_inen(gpionr, 1);
  567. set_gpio_data(gpionr, 0);
  568. } else {
  569. set_gpio_edge(gpionr, 0);
  570. set_gpio_inen(gpionr, 1);
  571. }
  572. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  573. bfin_set_irq_handler(irq, handle_edge_irq);
  574. else
  575. bfin_set_irq_handler(irq, handle_level_irq);
  576. return 0;
  577. }
  578. #ifdef CONFIG_PM
  579. int bfin_gpio_set_wake(unsigned int irq, unsigned int state)
  580. {
  581. return gpio_pm_wakeup_ctrl(irq_to_gpio(irq), state);
  582. }
  583. #endif
  584. static void bfin_demux_gpio_irq(unsigned int inta_irq,
  585. struct irq_desc *desc)
  586. {
  587. unsigned int i, gpio, mask, irq, search = 0;
  588. switch (inta_irq) {
  589. #if defined(CONFIG_BF53x)
  590. case IRQ_PROG_INTA:
  591. irq = IRQ_PF0;
  592. search = 1;
  593. break;
  594. # if defined(BF537_FAMILY) && !(defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE))
  595. case IRQ_MAC_RX:
  596. irq = IRQ_PH0;
  597. break;
  598. # endif
  599. #elif defined(CONFIG_BF538) || defined(CONFIG_BF539)
  600. case IRQ_PORTF_INTA:
  601. irq = IRQ_PF0;
  602. break;
  603. #elif defined(CONFIG_BF52x) || defined(CONFIG_BF51x)
  604. case IRQ_PORTF_INTA:
  605. irq = IRQ_PF0;
  606. break;
  607. case IRQ_PORTG_INTA:
  608. irq = IRQ_PG0;
  609. break;
  610. case IRQ_PORTH_INTA:
  611. irq = IRQ_PH0;
  612. break;
  613. #elif defined(CONFIG_BF561)
  614. case IRQ_PROG0_INTA:
  615. irq = IRQ_PF0;
  616. break;
  617. case IRQ_PROG1_INTA:
  618. irq = IRQ_PF16;
  619. break;
  620. case IRQ_PROG2_INTA:
  621. irq = IRQ_PF32;
  622. break;
  623. #endif
  624. default:
  625. BUG();
  626. return;
  627. }
  628. if (search) {
  629. for (i = 0; i < MAX_BLACKFIN_GPIOS; i += GPIO_BANKSIZE) {
  630. irq += i;
  631. mask = get_gpiop_data(i) & get_gpiop_maska(i);
  632. while (mask) {
  633. if (mask & 1)
  634. bfin_handle_irq(irq);
  635. irq++;
  636. mask >>= 1;
  637. }
  638. }
  639. } else {
  640. gpio = irq_to_gpio(irq);
  641. mask = get_gpiop_data(gpio) & get_gpiop_maska(gpio);
  642. do {
  643. if (mask & 1)
  644. bfin_handle_irq(irq);
  645. irq++;
  646. mask >>= 1;
  647. } while (mask);
  648. }
  649. }
  650. #else /* CONFIG_BF54x */
  651. #define NR_PINT_SYS_IRQS 4
  652. #define NR_PINT_BITS 32
  653. #define NR_PINTS 160
  654. #define IRQ_NOT_AVAIL 0xFF
  655. #define PINT_2_BANK(x) ((x) >> 5)
  656. #define PINT_2_BIT(x) ((x) & 0x1F)
  657. #define PINT_BIT(x) (1 << (PINT_2_BIT(x)))
  658. static unsigned char irq2pint_lut[NR_PINTS];
  659. static unsigned char pint2irq_lut[NR_PINT_SYS_IRQS * NR_PINT_BITS];
  660. struct pin_int_t {
  661. unsigned int mask_set;
  662. unsigned int mask_clear;
  663. unsigned int request;
  664. unsigned int assign;
  665. unsigned int edge_set;
  666. unsigned int edge_clear;
  667. unsigned int invert_set;
  668. unsigned int invert_clear;
  669. unsigned int pinstate;
  670. unsigned int latch;
  671. };
  672. static struct pin_int_t *pint[NR_PINT_SYS_IRQS] = {
  673. (struct pin_int_t *)PINT0_MASK_SET,
  674. (struct pin_int_t *)PINT1_MASK_SET,
  675. (struct pin_int_t *)PINT2_MASK_SET,
  676. (struct pin_int_t *)PINT3_MASK_SET,
  677. };
  678. inline unsigned int get_irq_base(u32 bank, u8 bmap)
  679. {
  680. unsigned int irq_base;
  681. if (bank < 2) { /*PA-PB */
  682. irq_base = IRQ_PA0 + bmap * 16;
  683. } else { /*PC-PJ */
  684. irq_base = IRQ_PC0 + bmap * 16;
  685. }
  686. return irq_base;
  687. }
  688. /* Whenever PINTx_ASSIGN is altered init_pint_lut() must be executed! */
  689. void init_pint_lut(void)
  690. {
  691. u16 bank, bit, irq_base, bit_pos;
  692. u32 pint_assign;
  693. u8 bmap;
  694. memset(irq2pint_lut, IRQ_NOT_AVAIL, sizeof(irq2pint_lut));
  695. for (bank = 0; bank < NR_PINT_SYS_IRQS; bank++) {
  696. pint_assign = pint[bank]->assign;
  697. for (bit = 0; bit < NR_PINT_BITS; bit++) {
  698. bmap = (pint_assign >> ((bit / 8) * 8)) & 0xFF;
  699. irq_base = get_irq_base(bank, bmap);
  700. irq_base += (bit % 8) + ((bit / 8) & 1 ? 8 : 0);
  701. bit_pos = bit + bank * NR_PINT_BITS;
  702. pint2irq_lut[bit_pos] = irq_base - SYS_IRQS;
  703. irq2pint_lut[irq_base - SYS_IRQS] = bit_pos;
  704. }
  705. }
  706. }
  707. static void bfin_gpio_ack_irq(unsigned int irq)
  708. {
  709. struct irq_desc *desc = irq_desc + irq;
  710. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  711. u32 pintbit = PINT_BIT(pint_val);
  712. u32 bank = PINT_2_BANK(pint_val);
  713. if ((desc->status & IRQ_TYPE_SENSE_MASK) == IRQ_TYPE_EDGE_BOTH) {
  714. if (pint[bank]->invert_set & pintbit)
  715. pint[bank]->invert_clear = pintbit;
  716. else
  717. pint[bank]->invert_set = pintbit;
  718. }
  719. pint[bank]->request = pintbit;
  720. }
  721. static void bfin_gpio_mask_ack_irq(unsigned int irq)
  722. {
  723. struct irq_desc *desc = irq_desc + irq;
  724. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  725. u32 pintbit = PINT_BIT(pint_val);
  726. u32 bank = PINT_2_BANK(pint_val);
  727. if ((desc->status & IRQ_TYPE_SENSE_MASK) == IRQ_TYPE_EDGE_BOTH) {
  728. if (pint[bank]->invert_set & pintbit)
  729. pint[bank]->invert_clear = pintbit;
  730. else
  731. pint[bank]->invert_set = pintbit;
  732. }
  733. pint[bank]->request = pintbit;
  734. pint[bank]->mask_clear = pintbit;
  735. }
  736. static void bfin_gpio_mask_irq(unsigned int irq)
  737. {
  738. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  739. pint[PINT_2_BANK(pint_val)]->mask_clear = PINT_BIT(pint_val);
  740. }
  741. static void bfin_gpio_unmask_irq(unsigned int irq)
  742. {
  743. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  744. u32 pintbit = PINT_BIT(pint_val);
  745. u32 bank = PINT_2_BANK(pint_val);
  746. pint[bank]->request = pintbit;
  747. pint[bank]->mask_set = pintbit;
  748. }
  749. static unsigned int bfin_gpio_irq_startup(unsigned int irq)
  750. {
  751. u32 gpionr = irq_to_gpio(irq);
  752. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  753. if (pint_val == IRQ_NOT_AVAIL) {
  754. printk(KERN_ERR
  755. "GPIO IRQ %d :Not in PINT Assign table "
  756. "Reconfigure Interrupt to Port Assignemt\n", irq);
  757. return -ENODEV;
  758. }
  759. if (__test_and_set_bit(gpionr, gpio_enabled))
  760. bfin_gpio_irq_prepare(gpionr);
  761. bfin_gpio_unmask_irq(irq);
  762. return 0;
  763. }
  764. static void bfin_gpio_irq_shutdown(unsigned int irq)
  765. {
  766. u32 gpionr = irq_to_gpio(irq);
  767. bfin_gpio_mask_irq(irq);
  768. __clear_bit(gpionr, gpio_enabled);
  769. bfin_gpio_irq_free(gpionr);
  770. }
  771. static int bfin_gpio_irq_type(unsigned int irq, unsigned int type)
  772. {
  773. int ret;
  774. char buf[16];
  775. u32 gpionr = irq_to_gpio(irq);
  776. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  777. u32 pintbit = PINT_BIT(pint_val);
  778. u32 bank = PINT_2_BANK(pint_val);
  779. if (pint_val == IRQ_NOT_AVAIL)
  780. return -ENODEV;
  781. if (type == IRQ_TYPE_PROBE) {
  782. /* only probe unenabled GPIO interrupt lines */
  783. if (test_bit(gpionr, gpio_enabled))
  784. return 0;
  785. type = IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING;
  786. }
  787. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING |
  788. IRQ_TYPE_LEVEL_HIGH | IRQ_TYPE_LEVEL_LOW)) {
  789. snprintf(buf, 16, "gpio-irq%d", irq);
  790. ret = bfin_gpio_irq_request(gpionr, buf);
  791. if (ret)
  792. return ret;
  793. if (__test_and_set_bit(gpionr, gpio_enabled))
  794. bfin_gpio_irq_prepare(gpionr);
  795. } else {
  796. __clear_bit(gpionr, gpio_enabled);
  797. return 0;
  798. }
  799. if ((type & (IRQ_TYPE_EDGE_FALLING | IRQ_TYPE_LEVEL_LOW)))
  800. pint[bank]->invert_set = pintbit; /* low or falling edge denoted by one */
  801. else
  802. pint[bank]->invert_clear = pintbit; /* high or rising edge denoted by zero */
  803. if ((type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  804. == (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  805. if (gpio_get_value(gpionr))
  806. pint[bank]->invert_set = pintbit;
  807. else
  808. pint[bank]->invert_clear = pintbit;
  809. }
  810. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  811. pint[bank]->edge_set = pintbit;
  812. bfin_set_irq_handler(irq, handle_edge_irq);
  813. } else {
  814. pint[bank]->edge_clear = pintbit;
  815. bfin_set_irq_handler(irq, handle_level_irq);
  816. }
  817. return 0;
  818. }
  819. #ifdef CONFIG_PM
  820. u32 pint_saved_masks[NR_PINT_SYS_IRQS];
  821. u32 pint_wakeup_masks[NR_PINT_SYS_IRQS];
  822. int bfin_gpio_set_wake(unsigned int irq, unsigned int state)
  823. {
  824. u32 pint_irq;
  825. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  826. u32 bank = PINT_2_BANK(pint_val);
  827. u32 pintbit = PINT_BIT(pint_val);
  828. switch (bank) {
  829. case 0:
  830. pint_irq = IRQ_PINT0;
  831. break;
  832. case 2:
  833. pint_irq = IRQ_PINT2;
  834. break;
  835. case 3:
  836. pint_irq = IRQ_PINT3;
  837. break;
  838. case 1:
  839. pint_irq = IRQ_PINT1;
  840. break;
  841. default:
  842. return -EINVAL;
  843. }
  844. bfin_internal_set_wake(pint_irq, state);
  845. if (state)
  846. pint_wakeup_masks[bank] |= pintbit;
  847. else
  848. pint_wakeup_masks[bank] &= ~pintbit;
  849. return 0;
  850. }
  851. u32 bfin_pm_setup(void)
  852. {
  853. u32 val, i;
  854. for (i = 0; i < NR_PINT_SYS_IRQS; i++) {
  855. val = pint[i]->mask_clear;
  856. pint_saved_masks[i] = val;
  857. if (val ^ pint_wakeup_masks[i]) {
  858. pint[i]->mask_clear = val;
  859. pint[i]->mask_set = pint_wakeup_masks[i];
  860. }
  861. }
  862. return 0;
  863. }
  864. void bfin_pm_restore(void)
  865. {
  866. u32 i, val;
  867. for (i = 0; i < NR_PINT_SYS_IRQS; i++) {
  868. val = pint_saved_masks[i];
  869. if (val ^ pint_wakeup_masks[i]) {
  870. pint[i]->mask_clear = pint[i]->mask_clear;
  871. pint[i]->mask_set = val;
  872. }
  873. }
  874. }
  875. #endif
  876. static void bfin_demux_gpio_irq(unsigned int inta_irq,
  877. struct irq_desc *desc)
  878. {
  879. u32 bank, pint_val;
  880. u32 request, irq;
  881. switch (inta_irq) {
  882. case IRQ_PINT0:
  883. bank = 0;
  884. break;
  885. case IRQ_PINT2:
  886. bank = 2;
  887. break;
  888. case IRQ_PINT3:
  889. bank = 3;
  890. break;
  891. case IRQ_PINT1:
  892. bank = 1;
  893. break;
  894. default:
  895. return;
  896. }
  897. pint_val = bank * NR_PINT_BITS;
  898. request = pint[bank]->request;
  899. while (request) {
  900. if (request & 1) {
  901. irq = pint2irq_lut[pint_val] + SYS_IRQS;
  902. bfin_handle_irq(irq);
  903. }
  904. pint_val++;
  905. request >>= 1;
  906. }
  907. }
  908. #endif
  909. static struct irq_chip bfin_gpio_irqchip = {
  910. .name = "GPIO",
  911. .ack = bfin_gpio_ack_irq,
  912. .mask = bfin_gpio_mask_irq,
  913. .mask_ack = bfin_gpio_mask_ack_irq,
  914. .unmask = bfin_gpio_unmask_irq,
  915. .disable = bfin_gpio_mask_irq,
  916. .enable = bfin_gpio_unmask_irq,
  917. .set_type = bfin_gpio_irq_type,
  918. .startup = bfin_gpio_irq_startup,
  919. .shutdown = bfin_gpio_irq_shutdown,
  920. #ifdef CONFIG_PM
  921. .set_wake = bfin_gpio_set_wake,
  922. #endif
  923. };
  924. void __cpuinit init_exception_vectors(void)
  925. {
  926. /* cannot program in software:
  927. * evt0 - emulation (jtag)
  928. * evt1 - reset
  929. */
  930. bfin_write_EVT2(evt_nmi);
  931. bfin_write_EVT3(trap);
  932. bfin_write_EVT5(evt_ivhw);
  933. bfin_write_EVT6(evt_timer);
  934. bfin_write_EVT7(evt_evt7);
  935. bfin_write_EVT8(evt_evt8);
  936. bfin_write_EVT9(evt_evt9);
  937. bfin_write_EVT10(evt_evt10);
  938. bfin_write_EVT11(evt_evt11);
  939. bfin_write_EVT12(evt_evt12);
  940. bfin_write_EVT13(evt_evt13);
  941. bfin_write_EVT14(evt_evt14);
  942. bfin_write_EVT15(evt_system_call);
  943. CSYNC();
  944. }
  945. /*
  946. * This function should be called during kernel startup to initialize
  947. * the BFin IRQ handling routines.
  948. */
  949. int __init init_arch_irq(void)
  950. {
  951. int irq;
  952. unsigned long ilat = 0;
  953. /* Disable all the peripheral intrs - page 4-29 HW Ref manual */
  954. #if defined(CONFIG_BF54x) || defined(CONFIG_BF52x) || defined(CONFIG_BF561) \
  955. || defined(BF538_FAMILY) || defined(CONFIG_BF51x)
  956. bfin_write_SIC_IMASK0(SIC_UNMASK_ALL);
  957. bfin_write_SIC_IMASK1(SIC_UNMASK_ALL);
  958. # ifdef CONFIG_BF54x
  959. bfin_write_SIC_IMASK2(SIC_UNMASK_ALL);
  960. # endif
  961. # ifdef CONFIG_SMP
  962. bfin_write_SICB_IMASK0(SIC_UNMASK_ALL);
  963. bfin_write_SICB_IMASK1(SIC_UNMASK_ALL);
  964. # endif
  965. #else
  966. bfin_write_SIC_IMASK(SIC_UNMASK_ALL);
  967. #endif
  968. local_irq_disable();
  969. #if (defined(CONFIG_BF537) || defined(CONFIG_BF536))
  970. /* Clear EMAC Interrupt Status bits so we can demux it later */
  971. bfin_write_EMAC_SYSTAT(-1);
  972. #endif
  973. #ifdef CONFIG_BF54x
  974. # ifdef CONFIG_PINTx_REASSIGN
  975. pint[0]->assign = CONFIG_PINT0_ASSIGN;
  976. pint[1]->assign = CONFIG_PINT1_ASSIGN;
  977. pint[2]->assign = CONFIG_PINT2_ASSIGN;
  978. pint[3]->assign = CONFIG_PINT3_ASSIGN;
  979. # endif
  980. /* Whenever PINTx_ASSIGN is altered init_pint_lut() must be executed! */
  981. init_pint_lut();
  982. #endif
  983. for (irq = 0; irq <= SYS_IRQS; irq++) {
  984. if (irq <= IRQ_CORETMR)
  985. set_irq_chip(irq, &bfin_core_irqchip);
  986. else
  987. set_irq_chip(irq, &bfin_internal_irqchip);
  988. switch (irq) {
  989. #if defined(CONFIG_BF53x)
  990. case IRQ_PROG_INTA:
  991. # if defined(BF537_FAMILY) && !(defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE))
  992. case IRQ_MAC_RX:
  993. # endif
  994. #elif defined(CONFIG_BF54x)
  995. case IRQ_PINT0:
  996. case IRQ_PINT1:
  997. case IRQ_PINT2:
  998. case IRQ_PINT3:
  999. #elif defined(CONFIG_BF52x) || defined(CONFIG_BF51x)
  1000. case IRQ_PORTF_INTA:
  1001. case IRQ_PORTG_INTA:
  1002. case IRQ_PORTH_INTA:
  1003. #elif defined(CONFIG_BF561)
  1004. case IRQ_PROG0_INTA:
  1005. case IRQ_PROG1_INTA:
  1006. case IRQ_PROG2_INTA:
  1007. #elif defined(CONFIG_BF538) || defined(CONFIG_BF539)
  1008. case IRQ_PORTF_INTA:
  1009. #endif
  1010. set_irq_chained_handler(irq,
  1011. bfin_demux_gpio_irq);
  1012. break;
  1013. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  1014. case IRQ_GENERIC_ERROR:
  1015. set_irq_chained_handler(irq, bfin_demux_error_irq);
  1016. break;
  1017. #endif
  1018. #if defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE)
  1019. case IRQ_MAC_ERROR:
  1020. set_irq_chained_handler(irq, bfin_demux_mac_status_irq);
  1021. break;
  1022. #endif
  1023. #ifdef CONFIG_SMP
  1024. case IRQ_SUPPLE_0:
  1025. case IRQ_SUPPLE_1:
  1026. set_irq_handler(irq, handle_percpu_irq);
  1027. break;
  1028. #endif
  1029. #ifdef CONFIG_TICKSOURCE_CORETMR
  1030. case IRQ_CORETMR:
  1031. # ifdef CONFIG_SMP
  1032. set_irq_handler(irq, handle_percpu_irq);
  1033. break;
  1034. # else
  1035. set_irq_handler(irq, handle_simple_irq);
  1036. break;
  1037. # endif
  1038. #endif
  1039. #ifdef CONFIG_TICKSOURCE_GPTMR0
  1040. case IRQ_TIMER0:
  1041. set_irq_handler(irq, handle_simple_irq);
  1042. break;
  1043. #endif
  1044. #ifdef CONFIG_IPIPE
  1045. default:
  1046. set_irq_handler(irq, handle_level_irq);
  1047. break;
  1048. #else /* !CONFIG_IPIPE */
  1049. default:
  1050. set_irq_handler(irq, handle_simple_irq);
  1051. break;
  1052. #endif /* !CONFIG_IPIPE */
  1053. }
  1054. }
  1055. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  1056. for (irq = IRQ_PPI_ERROR; irq <= IRQ_UART1_ERROR; irq++)
  1057. set_irq_chip_and_handler(irq, &bfin_generic_error_irqchip,
  1058. handle_level_irq);
  1059. #if defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE)
  1060. set_irq_chained_handler(IRQ_MAC_ERROR, bfin_demux_mac_status_irq);
  1061. #endif
  1062. #endif
  1063. #if defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE)
  1064. for (irq = IRQ_MAC_PHYINT; irq <= IRQ_MAC_STMDONE; irq++)
  1065. set_irq_chip_and_handler(irq, &bfin_mac_status_irqchip,
  1066. handle_level_irq);
  1067. #endif
  1068. /* if configured as edge, then will be changed to do_edge_IRQ */
  1069. for (irq = GPIO_IRQ_BASE;
  1070. irq < (GPIO_IRQ_BASE + MAX_BLACKFIN_GPIOS); irq++)
  1071. set_irq_chip_and_handler(irq, &bfin_gpio_irqchip,
  1072. handle_level_irq);
  1073. bfin_write_IMASK(0);
  1074. CSYNC();
  1075. ilat = bfin_read_ILAT();
  1076. CSYNC();
  1077. bfin_write_ILAT(ilat);
  1078. CSYNC();
  1079. printk(KERN_INFO "Configuring Blackfin Priority Driven Interrupts\n");
  1080. /* IMASK=xxx is equivalent to STI xx or bfin_irq_flags=xx,
  1081. * local_irq_enable()
  1082. */
  1083. program_IAR();
  1084. /* Therefore it's better to setup IARs before interrupts enabled */
  1085. search_IAR();
  1086. /* Enable interrupts IVG7-15 */
  1087. bfin_irq_flags |= IMASK_IVG15 |
  1088. IMASK_IVG14 | IMASK_IVG13 | IMASK_IVG12 | IMASK_IVG11 |
  1089. IMASK_IVG10 | IMASK_IVG9 | IMASK_IVG8 | IMASK_IVG7 | IMASK_IVGHW;
  1090. /* This implicitly covers ANOMALY_05000171
  1091. * Boot-ROM code modifies SICA_IWRx wakeup registers
  1092. */
  1093. #ifdef SIC_IWR0
  1094. bfin_write_SIC_IWR0(IWR_DISABLE_ALL);
  1095. # ifdef SIC_IWR1
  1096. /* BF52x/BF51x system reset does not properly reset SIC_IWR1 which
  1097. * will screw up the bootrom as it relies on MDMA0/1 waking it
  1098. * up from IDLE instructions. See this report for more info:
  1099. * http://blackfin.uclinux.org/gf/tracker/4323
  1100. */
  1101. if (ANOMALY_05000435)
  1102. bfin_write_SIC_IWR1(IWR_ENABLE(10) | IWR_ENABLE(11));
  1103. else
  1104. bfin_write_SIC_IWR1(IWR_DISABLE_ALL);
  1105. # endif
  1106. # ifdef SIC_IWR2
  1107. bfin_write_SIC_IWR2(IWR_DISABLE_ALL);
  1108. # endif
  1109. #else
  1110. bfin_write_SIC_IWR(IWR_DISABLE_ALL);
  1111. #endif
  1112. return 0;
  1113. }
  1114. #ifdef CONFIG_DO_IRQ_L1
  1115. __attribute__((l1_text))
  1116. #endif
  1117. void do_irq(int vec, struct pt_regs *fp)
  1118. {
  1119. if (vec == EVT_IVTMR_P) {
  1120. vec = IRQ_CORETMR;
  1121. } else {
  1122. struct ivgx *ivg = ivg7_13[vec - IVG7].ifirst;
  1123. struct ivgx *ivg_stop = ivg7_13[vec - IVG7].istop;
  1124. #if defined(SIC_ISR0) || defined(SICA_ISR0)
  1125. unsigned long sic_status[3];
  1126. if (smp_processor_id()) {
  1127. # ifdef SICB_ISR0
  1128. /* This will be optimized out in UP mode. */
  1129. sic_status[0] = bfin_read_SICB_ISR0() & bfin_read_SICB_IMASK0();
  1130. sic_status[1] = bfin_read_SICB_ISR1() & bfin_read_SICB_IMASK1();
  1131. # endif
  1132. } else {
  1133. sic_status[0] = bfin_read_SIC_ISR0() & bfin_read_SIC_IMASK0();
  1134. sic_status[1] = bfin_read_SIC_ISR1() & bfin_read_SIC_IMASK1();
  1135. }
  1136. # ifdef SIC_ISR2
  1137. sic_status[2] = bfin_read_SIC_ISR2() & bfin_read_SIC_IMASK2();
  1138. # endif
  1139. for (;; ivg++) {
  1140. if (ivg >= ivg_stop) {
  1141. atomic_inc(&num_spurious);
  1142. return;
  1143. }
  1144. if (sic_status[(ivg->irqno - IVG7) / 32] & ivg->isrflag)
  1145. break;
  1146. }
  1147. #else
  1148. unsigned long sic_status;
  1149. sic_status = bfin_read_SIC_IMASK() & bfin_read_SIC_ISR();
  1150. for (;; ivg++) {
  1151. if (ivg >= ivg_stop) {
  1152. atomic_inc(&num_spurious);
  1153. return;
  1154. } else if (sic_status & ivg->isrflag)
  1155. break;
  1156. }
  1157. #endif
  1158. vec = ivg->irqno;
  1159. }
  1160. asm_do_IRQ(vec, fp);
  1161. }
  1162. #ifdef CONFIG_IPIPE
  1163. int __ipipe_get_irq_priority(unsigned irq)
  1164. {
  1165. int ient, prio;
  1166. if (irq <= IRQ_CORETMR)
  1167. return irq;
  1168. for (ient = 0; ient < NR_PERI_INTS; ient++) {
  1169. struct ivgx *ivg = ivg_table + ient;
  1170. if (ivg->irqno == irq) {
  1171. for (prio = 0; prio <= IVG13-IVG7; prio++) {
  1172. if (ivg7_13[prio].ifirst <= ivg &&
  1173. ivg7_13[prio].istop > ivg)
  1174. return IVG7 + prio;
  1175. }
  1176. }
  1177. }
  1178. return IVG15;
  1179. }
  1180. /* Hw interrupts are disabled on entry (check SAVE_CONTEXT). */
  1181. #ifdef CONFIG_DO_IRQ_L1
  1182. __attribute__((l1_text))
  1183. #endif
  1184. asmlinkage int __ipipe_grab_irq(int vec, struct pt_regs *regs)
  1185. {
  1186. struct ipipe_percpu_domain_data *p = ipipe_root_cpudom_ptr();
  1187. struct ipipe_domain *this_domain = __ipipe_current_domain;
  1188. struct ivgx *ivg_stop = ivg7_13[vec-IVG7].istop;
  1189. struct ivgx *ivg = ivg7_13[vec-IVG7].ifirst;
  1190. int irq, s;
  1191. if (likely(vec == EVT_IVTMR_P))
  1192. irq = IRQ_CORETMR;
  1193. else {
  1194. #if defined(SIC_ISR0) || defined(SICA_ISR0)
  1195. unsigned long sic_status[3];
  1196. sic_status[0] = bfin_read_SIC_ISR0() & bfin_read_SIC_IMASK0();
  1197. sic_status[1] = bfin_read_SIC_ISR1() & bfin_read_SIC_IMASK1();
  1198. # ifdef SIC_ISR2
  1199. sic_status[2] = bfin_read_SIC_ISR2() & bfin_read_SIC_IMASK2();
  1200. # endif
  1201. for (;; ivg++) {
  1202. if (ivg >= ivg_stop) {
  1203. atomic_inc(&num_spurious);
  1204. return 0;
  1205. }
  1206. if (sic_status[(ivg->irqno - IVG7) / 32] & ivg->isrflag)
  1207. break;
  1208. }
  1209. #else
  1210. unsigned long sic_status;
  1211. sic_status = bfin_read_SIC_IMASK() & bfin_read_SIC_ISR();
  1212. for (;; ivg++) {
  1213. if (ivg >= ivg_stop) {
  1214. atomic_inc(&num_spurious);
  1215. return 0;
  1216. } else if (sic_status & ivg->isrflag)
  1217. break;
  1218. }
  1219. #endif
  1220. irq = ivg->irqno;
  1221. }
  1222. if (irq == IRQ_SYSTMR) {
  1223. #if !defined(CONFIG_GENERIC_CLOCKEVENTS) || defined(CONFIG_TICKSOURCE_GPTMR0)
  1224. bfin_write_TIMER_STATUS(1); /* Latch TIMIL0 */
  1225. #endif
  1226. /* This is basically what we need from the register frame. */
  1227. __raw_get_cpu_var(__ipipe_tick_regs).ipend = regs->ipend;
  1228. __raw_get_cpu_var(__ipipe_tick_regs).pc = regs->pc;
  1229. if (this_domain != ipipe_root_domain)
  1230. __raw_get_cpu_var(__ipipe_tick_regs).ipend &= ~0x10;
  1231. else
  1232. __raw_get_cpu_var(__ipipe_tick_regs).ipend |= 0x10;
  1233. }
  1234. if (this_domain == ipipe_root_domain) {
  1235. s = __test_and_set_bit(IPIPE_SYNCDEFER_FLAG, &p->status);
  1236. barrier();
  1237. }
  1238. ipipe_trace_irq_entry(irq);
  1239. __ipipe_handle_irq(irq, regs);
  1240. ipipe_trace_irq_exit(irq);
  1241. if (this_domain == ipipe_root_domain) {
  1242. set_thread_flag(TIF_IRQ_SYNC);
  1243. if (!s) {
  1244. __clear_bit(IPIPE_SYNCDEFER_FLAG, &p->status);
  1245. return !test_bit(IPIPE_STALL_FLAG, &p->status);
  1246. }
  1247. }
  1248. return 0;
  1249. }
  1250. #endif /* CONFIG_IPIPE */