ints-priority.c 30 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323
  1. /*
  2. * Set up the interrupt priorities
  3. *
  4. * Copyright 2004-2009 Analog Devices Inc.
  5. * 2003 Bas Vermeulen <bas@buyways.nl>
  6. * 2002 Arcturus Networks Inc. MaTed <mated@sympatico.ca>
  7. * 2000-2001 Lineo, Inc. D. Jefff Dionne <jeff@lineo.ca>
  8. * 1999 D. Jeff Dionne <jeff@uclinux.org>
  9. * 1996 Roman Zippel
  10. *
  11. * Licensed under the GPL-2
  12. */
  13. #include <linux/module.h>
  14. #include <linux/kernel_stat.h>
  15. #include <linux/seq_file.h>
  16. #include <linux/irq.h>
  17. #ifdef CONFIG_IPIPE
  18. #include <linux/ipipe.h>
  19. #endif
  20. #ifdef CONFIG_KGDB
  21. #include <linux/kgdb.h>
  22. #endif
  23. #include <asm/traps.h>
  24. #include <asm/blackfin.h>
  25. #include <asm/gpio.h>
  26. #include <asm/irq_handler.h>
  27. #include <asm/dpmc.h>
  28. #include <asm/bfin5xx_spi.h>
  29. #include <asm/bfin_sport.h>
  30. #include <asm/bfin_can.h>
  31. #define SIC_SYSIRQ(irq) (irq - (IRQ_CORETMR + 1))
  32. #ifdef BF537_FAMILY
  33. # define BF537_GENERIC_ERROR_INT_DEMUX
  34. # define SPI_ERR_MASK (BIT_STAT_TXCOL | BIT_STAT_RBSY | BIT_STAT_MODF | BIT_STAT_TXE) /* SPI_STAT */
  35. # define SPORT_ERR_MASK (ROVF | RUVF | TOVF | TUVF) /* SPORT_STAT */
  36. # define PPI_ERR_MASK (0xFFFF & ~FLD) /* PPI_STATUS */
  37. # define EMAC_ERR_MASK (PHYINT | MMCINT | RXFSINT | TXFSINT | WAKEDET | RXDMAERR | TXDMAERR | STMDONE) /* EMAC_SYSTAT */
  38. # define UART_ERR_MASK (0x6) /* UART_IIR */
  39. # define CAN_ERR_MASK (EWTIF | EWRIF | EPIF | BOIF | WUIF | UIAIF | AAIF | RMLIF | UCEIF | EXTIF | ADIF) /* CAN_GIF */
  40. #else
  41. # undef BF537_GENERIC_ERROR_INT_DEMUX
  42. #endif
  43. /*
  44. * NOTES:
  45. * - we have separated the physical Hardware interrupt from the
  46. * levels that the LINUX kernel sees (see the description in irq.h)
  47. * -
  48. */
  49. #ifndef CONFIG_SMP
  50. /* Initialize this to an actual value to force it into the .data
  51. * section so that we know it is properly initialized at entry into
  52. * the kernel but before bss is initialized to zero (which is where
  53. * it would live otherwise). The 0x1f magic represents the IRQs we
  54. * cannot actually mask out in hardware.
  55. */
  56. unsigned long bfin_irq_flags = 0x1f;
  57. EXPORT_SYMBOL(bfin_irq_flags);
  58. #endif
  59. /* The number of spurious interrupts */
  60. atomic_t num_spurious;
  61. #ifdef CONFIG_PM
  62. unsigned long bfin_sic_iwr[3]; /* Up to 3 SIC_IWRx registers */
  63. unsigned vr_wakeup;
  64. #endif
  65. struct ivgx {
  66. /* irq number for request_irq, available in mach-bf5xx/irq.h */
  67. unsigned int irqno;
  68. /* corresponding bit in the SIC_ISR register */
  69. unsigned int isrflag;
  70. } ivg_table[NR_PERI_INTS];
  71. struct ivg_slice {
  72. /* position of first irq in ivg_table for given ivg */
  73. struct ivgx *ifirst;
  74. struct ivgx *istop;
  75. } ivg7_13[IVG13 - IVG7 + 1];
  76. /*
  77. * Search SIC_IAR and fill tables with the irqvalues
  78. * and their positions in the SIC_ISR register.
  79. */
  80. static void __init search_IAR(void)
  81. {
  82. unsigned ivg, irq_pos = 0;
  83. for (ivg = 0; ivg <= IVG13 - IVG7; ivg++) {
  84. int irqn;
  85. ivg7_13[ivg].istop = ivg7_13[ivg].ifirst = &ivg_table[irq_pos];
  86. for (irqn = 0; irqn < NR_PERI_INTS; irqn++) {
  87. int iar_shift = (irqn & 7) * 4;
  88. if (ivg == (0xf &
  89. #if defined(CONFIG_BF52x) || defined(CONFIG_BF538) \
  90. || defined(CONFIG_BF539) || defined(CONFIG_BF51x)
  91. bfin_read32((unsigned long *)SIC_IAR0 +
  92. ((irqn % 32) >> 3) + ((irqn / 32) *
  93. ((SIC_IAR4 - SIC_IAR0) / 4))) >> iar_shift)) {
  94. #else
  95. bfin_read32((unsigned long *)SIC_IAR0 +
  96. (irqn >> 3)) >> iar_shift)) {
  97. #endif
  98. ivg_table[irq_pos].irqno = IVG7 + irqn;
  99. ivg_table[irq_pos].isrflag = 1 << (irqn % 32);
  100. ivg7_13[ivg].istop++;
  101. irq_pos++;
  102. }
  103. }
  104. }
  105. }
  106. /*
  107. * This is for core internal IRQs
  108. */
  109. static void bfin_ack_noop(unsigned int irq)
  110. {
  111. /* Dummy function. */
  112. }
  113. static void bfin_core_mask_irq(unsigned int irq)
  114. {
  115. bfin_irq_flags &= ~(1 << irq);
  116. if (!irqs_disabled_hw())
  117. local_irq_enable_hw();
  118. }
  119. static void bfin_core_unmask_irq(unsigned int irq)
  120. {
  121. bfin_irq_flags |= 1 << irq;
  122. /*
  123. * If interrupts are enabled, IMASK must contain the same value
  124. * as bfin_irq_flags. Make sure that invariant holds. If interrupts
  125. * are currently disabled we need not do anything; one of the
  126. * callers will take care of setting IMASK to the proper value
  127. * when reenabling interrupts.
  128. * local_irq_enable just does "STI bfin_irq_flags", so it's exactly
  129. * what we need.
  130. */
  131. if (!irqs_disabled_hw())
  132. local_irq_enable_hw();
  133. return;
  134. }
  135. static void bfin_internal_mask_irq(unsigned int irq)
  136. {
  137. unsigned long flags;
  138. #ifdef CONFIG_BF53x
  139. local_irq_save_hw(flags);
  140. bfin_write_SIC_IMASK(bfin_read_SIC_IMASK() &
  141. ~(1 << SIC_SYSIRQ(irq)));
  142. #else
  143. unsigned mask_bank, mask_bit;
  144. local_irq_save_hw(flags);
  145. mask_bank = SIC_SYSIRQ(irq) / 32;
  146. mask_bit = SIC_SYSIRQ(irq) % 32;
  147. bfin_write_SIC_IMASK(mask_bank, bfin_read_SIC_IMASK(mask_bank) &
  148. ~(1 << mask_bit));
  149. #ifdef CONFIG_SMP
  150. bfin_write_SICB_IMASK(mask_bank, bfin_read_SICB_IMASK(mask_bank) &
  151. ~(1 << mask_bit));
  152. #endif
  153. #endif
  154. local_irq_restore_hw(flags);
  155. }
  156. #ifdef CONFIG_SMP
  157. static void bfin_internal_unmask_irq_affinity(unsigned int irq,
  158. const struct cpumask *affinity)
  159. #else
  160. static void bfin_internal_unmask_irq(unsigned int irq)
  161. #endif
  162. {
  163. unsigned long flags;
  164. #ifdef CONFIG_BF53x
  165. local_irq_save_hw(flags);
  166. bfin_write_SIC_IMASK(bfin_read_SIC_IMASK() |
  167. (1 << SIC_SYSIRQ(irq)));
  168. #else
  169. unsigned mask_bank, mask_bit;
  170. local_irq_save_hw(flags);
  171. mask_bank = SIC_SYSIRQ(irq) / 32;
  172. mask_bit = SIC_SYSIRQ(irq) % 32;
  173. #ifdef CONFIG_SMP
  174. if (cpumask_test_cpu(0, affinity))
  175. #endif
  176. bfin_write_SIC_IMASK(mask_bank,
  177. bfin_read_SIC_IMASK(mask_bank) |
  178. (1 << mask_bit));
  179. #ifdef CONFIG_SMP
  180. if (cpumask_test_cpu(1, affinity))
  181. bfin_write_SICB_IMASK(mask_bank,
  182. bfin_read_SICB_IMASK(mask_bank) |
  183. (1 << mask_bit));
  184. #endif
  185. #endif
  186. local_irq_restore_hw(flags);
  187. }
  188. #ifdef CONFIG_SMP
  189. static void bfin_internal_unmask_irq(unsigned int irq)
  190. {
  191. struct irq_desc *desc = irq_to_desc(irq);
  192. bfin_internal_unmask_irq_affinity(irq, desc->affinity);
  193. }
  194. static int bfin_internal_set_affinity(unsigned int irq, const struct cpumask *mask)
  195. {
  196. bfin_internal_mask_irq(irq);
  197. bfin_internal_unmask_irq_affinity(irq, mask);
  198. return 0;
  199. }
  200. #endif
  201. #ifdef CONFIG_PM
  202. int bfin_internal_set_wake(unsigned int irq, unsigned int state)
  203. {
  204. u32 bank, bit, wakeup = 0;
  205. unsigned long flags;
  206. bank = SIC_SYSIRQ(irq) / 32;
  207. bit = SIC_SYSIRQ(irq) % 32;
  208. switch (irq) {
  209. #ifdef IRQ_RTC
  210. case IRQ_RTC:
  211. wakeup |= WAKE;
  212. break;
  213. #endif
  214. #ifdef IRQ_CAN0_RX
  215. case IRQ_CAN0_RX:
  216. wakeup |= CANWE;
  217. break;
  218. #endif
  219. #ifdef IRQ_CAN1_RX
  220. case IRQ_CAN1_RX:
  221. wakeup |= CANWE;
  222. break;
  223. #endif
  224. #ifdef IRQ_USB_INT0
  225. case IRQ_USB_INT0:
  226. wakeup |= USBWE;
  227. break;
  228. #endif
  229. #ifdef IRQ_KEY
  230. case IRQ_KEY:
  231. wakeup |= KPADWE;
  232. break;
  233. #endif
  234. #ifdef CONFIG_BF54x
  235. case IRQ_CNT:
  236. wakeup |= ROTWE;
  237. break;
  238. #endif
  239. default:
  240. break;
  241. }
  242. local_irq_save_hw(flags);
  243. if (state) {
  244. bfin_sic_iwr[bank] |= (1 << bit);
  245. vr_wakeup |= wakeup;
  246. } else {
  247. bfin_sic_iwr[bank] &= ~(1 << bit);
  248. vr_wakeup &= ~wakeup;
  249. }
  250. local_irq_restore_hw(flags);
  251. return 0;
  252. }
  253. #endif
  254. static struct irq_chip bfin_core_irqchip = {
  255. .name = "CORE",
  256. .ack = bfin_ack_noop,
  257. .mask = bfin_core_mask_irq,
  258. .unmask = bfin_core_unmask_irq,
  259. };
  260. static struct irq_chip bfin_internal_irqchip = {
  261. .name = "INTN",
  262. .ack = bfin_ack_noop,
  263. .mask = bfin_internal_mask_irq,
  264. .unmask = bfin_internal_unmask_irq,
  265. .mask_ack = bfin_internal_mask_irq,
  266. .disable = bfin_internal_mask_irq,
  267. .enable = bfin_internal_unmask_irq,
  268. #ifdef CONFIG_SMP
  269. .set_affinity = bfin_internal_set_affinity,
  270. #endif
  271. #ifdef CONFIG_PM
  272. .set_wake = bfin_internal_set_wake,
  273. #endif
  274. };
  275. static void bfin_handle_irq(unsigned irq)
  276. {
  277. #ifdef CONFIG_IPIPE
  278. struct pt_regs regs; /* Contents not used. */
  279. ipipe_trace_irq_entry(irq);
  280. __ipipe_handle_irq(irq, &regs);
  281. ipipe_trace_irq_exit(irq);
  282. #else /* !CONFIG_IPIPE */
  283. struct irq_desc *desc = irq_desc + irq;
  284. desc->handle_irq(irq, desc);
  285. #endif /* !CONFIG_IPIPE */
  286. }
  287. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  288. static int error_int_mask;
  289. static void bfin_generic_error_mask_irq(unsigned int irq)
  290. {
  291. error_int_mask &= ~(1L << (irq - IRQ_PPI_ERROR));
  292. if (!error_int_mask)
  293. bfin_internal_mask_irq(IRQ_GENERIC_ERROR);
  294. }
  295. static void bfin_generic_error_unmask_irq(unsigned int irq)
  296. {
  297. bfin_internal_unmask_irq(IRQ_GENERIC_ERROR);
  298. error_int_mask |= 1L << (irq - IRQ_PPI_ERROR);
  299. }
  300. static struct irq_chip bfin_generic_error_irqchip = {
  301. .name = "ERROR",
  302. .ack = bfin_ack_noop,
  303. .mask_ack = bfin_generic_error_mask_irq,
  304. .mask = bfin_generic_error_mask_irq,
  305. .unmask = bfin_generic_error_unmask_irq,
  306. };
  307. static void bfin_demux_error_irq(unsigned int int_err_irq,
  308. struct irq_desc *inta_desc)
  309. {
  310. int irq = 0;
  311. #if (defined(CONFIG_BF537) || defined(CONFIG_BF536))
  312. if (bfin_read_EMAC_SYSTAT() & EMAC_ERR_MASK)
  313. irq = IRQ_MAC_ERROR;
  314. else
  315. #endif
  316. if (bfin_read_SPORT0_STAT() & SPORT_ERR_MASK)
  317. irq = IRQ_SPORT0_ERROR;
  318. else if (bfin_read_SPORT1_STAT() & SPORT_ERR_MASK)
  319. irq = IRQ_SPORT1_ERROR;
  320. else if (bfin_read_PPI_STATUS() & PPI_ERR_MASK)
  321. irq = IRQ_PPI_ERROR;
  322. else if (bfin_read_CAN_GIF() & CAN_ERR_MASK)
  323. irq = IRQ_CAN_ERROR;
  324. else if (bfin_read_SPI_STAT() & SPI_ERR_MASK)
  325. irq = IRQ_SPI_ERROR;
  326. else if ((bfin_read_UART0_IIR() & UART_ERR_MASK) == UART_ERR_MASK)
  327. irq = IRQ_UART0_ERROR;
  328. else if ((bfin_read_UART1_IIR() & UART_ERR_MASK) == UART_ERR_MASK)
  329. irq = IRQ_UART1_ERROR;
  330. if (irq) {
  331. if (error_int_mask & (1L << (irq - IRQ_PPI_ERROR)))
  332. bfin_handle_irq(irq);
  333. else {
  334. switch (irq) {
  335. case IRQ_PPI_ERROR:
  336. bfin_write_PPI_STATUS(PPI_ERR_MASK);
  337. break;
  338. #if (defined(CONFIG_BF537) || defined(CONFIG_BF536))
  339. case IRQ_MAC_ERROR:
  340. bfin_write_EMAC_SYSTAT(EMAC_ERR_MASK);
  341. break;
  342. #endif
  343. case IRQ_SPORT0_ERROR:
  344. bfin_write_SPORT0_STAT(SPORT_ERR_MASK);
  345. break;
  346. case IRQ_SPORT1_ERROR:
  347. bfin_write_SPORT1_STAT(SPORT_ERR_MASK);
  348. break;
  349. case IRQ_CAN_ERROR:
  350. bfin_write_CAN_GIS(CAN_ERR_MASK);
  351. break;
  352. case IRQ_SPI_ERROR:
  353. bfin_write_SPI_STAT(SPI_ERR_MASK);
  354. break;
  355. default:
  356. break;
  357. }
  358. pr_debug("IRQ %d:"
  359. " MASKED PERIPHERAL ERROR INTERRUPT ASSERTED\n",
  360. irq);
  361. }
  362. } else
  363. printk(KERN_ERR
  364. "%s : %s : LINE %d :\nIRQ ?: PERIPHERAL ERROR"
  365. " INTERRUPT ASSERTED BUT NO SOURCE FOUND\n",
  366. __func__, __FILE__, __LINE__);
  367. }
  368. #endif /* BF537_GENERIC_ERROR_INT_DEMUX */
  369. static inline void bfin_set_irq_handler(unsigned irq, irq_flow_handler_t handle)
  370. {
  371. #ifdef CONFIG_IPIPE
  372. _set_irq_handler(irq, handle_level_irq);
  373. #else
  374. struct irq_desc *desc = irq_desc + irq;
  375. /* May not call generic set_irq_handler() due to spinlock
  376. recursion. */
  377. desc->handle_irq = handle;
  378. #endif
  379. }
  380. static DECLARE_BITMAP(gpio_enabled, MAX_BLACKFIN_GPIOS);
  381. extern void bfin_gpio_irq_prepare(unsigned gpio);
  382. #if !defined(CONFIG_BF54x)
  383. static void bfin_gpio_ack_irq(unsigned int irq)
  384. {
  385. /* AFAIK ack_irq in case mask_ack is provided
  386. * get's only called for edge sense irqs
  387. */
  388. set_gpio_data(irq_to_gpio(irq), 0);
  389. }
  390. static void bfin_gpio_mask_ack_irq(unsigned int irq)
  391. {
  392. struct irq_desc *desc = irq_desc + irq;
  393. u32 gpionr = irq_to_gpio(irq);
  394. if (desc->handle_irq == handle_edge_irq)
  395. set_gpio_data(gpionr, 0);
  396. set_gpio_maska(gpionr, 0);
  397. }
  398. static void bfin_gpio_mask_irq(unsigned int irq)
  399. {
  400. set_gpio_maska(irq_to_gpio(irq), 0);
  401. }
  402. static void bfin_gpio_unmask_irq(unsigned int irq)
  403. {
  404. set_gpio_maska(irq_to_gpio(irq), 1);
  405. }
  406. static unsigned int bfin_gpio_irq_startup(unsigned int irq)
  407. {
  408. u32 gpionr = irq_to_gpio(irq);
  409. if (__test_and_set_bit(gpionr, gpio_enabled))
  410. bfin_gpio_irq_prepare(gpionr);
  411. bfin_gpio_unmask_irq(irq);
  412. return 0;
  413. }
  414. static void bfin_gpio_irq_shutdown(unsigned int irq)
  415. {
  416. u32 gpionr = irq_to_gpio(irq);
  417. bfin_gpio_mask_irq(irq);
  418. __clear_bit(gpionr, gpio_enabled);
  419. bfin_gpio_irq_free(gpionr);
  420. }
  421. static int bfin_gpio_irq_type(unsigned int irq, unsigned int type)
  422. {
  423. int ret;
  424. char buf[16];
  425. u32 gpionr = irq_to_gpio(irq);
  426. if (type == IRQ_TYPE_PROBE) {
  427. /* only probe unenabled GPIO interrupt lines */
  428. if (test_bit(gpionr, gpio_enabled))
  429. return 0;
  430. type = IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING;
  431. }
  432. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING |
  433. IRQ_TYPE_LEVEL_HIGH | IRQ_TYPE_LEVEL_LOW)) {
  434. snprintf(buf, 16, "gpio-irq%d", irq);
  435. ret = bfin_gpio_irq_request(gpionr, buf);
  436. if (ret)
  437. return ret;
  438. if (__test_and_set_bit(gpionr, gpio_enabled))
  439. bfin_gpio_irq_prepare(gpionr);
  440. } else {
  441. __clear_bit(gpionr, gpio_enabled);
  442. return 0;
  443. }
  444. set_gpio_inen(gpionr, 0);
  445. set_gpio_dir(gpionr, 0);
  446. if ((type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  447. == (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  448. set_gpio_both(gpionr, 1);
  449. else
  450. set_gpio_both(gpionr, 0);
  451. if ((type & (IRQ_TYPE_EDGE_FALLING | IRQ_TYPE_LEVEL_LOW)))
  452. set_gpio_polar(gpionr, 1); /* low or falling edge denoted by one */
  453. else
  454. set_gpio_polar(gpionr, 0); /* high or rising edge denoted by zero */
  455. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  456. set_gpio_edge(gpionr, 1);
  457. set_gpio_inen(gpionr, 1);
  458. set_gpio_data(gpionr, 0);
  459. } else {
  460. set_gpio_edge(gpionr, 0);
  461. set_gpio_inen(gpionr, 1);
  462. }
  463. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  464. bfin_set_irq_handler(irq, handle_edge_irq);
  465. else
  466. bfin_set_irq_handler(irq, handle_level_irq);
  467. return 0;
  468. }
  469. #ifdef CONFIG_PM
  470. int bfin_gpio_set_wake(unsigned int irq, unsigned int state)
  471. {
  472. unsigned gpio = irq_to_gpio(irq);
  473. if (state)
  474. gpio_pm_wakeup_request(gpio, PM_WAKE_IGNORE);
  475. else
  476. gpio_pm_wakeup_free(gpio);
  477. return 0;
  478. }
  479. #endif
  480. static void bfin_demux_gpio_irq(unsigned int inta_irq,
  481. struct irq_desc *desc)
  482. {
  483. unsigned int i, gpio, mask, irq, search = 0;
  484. switch (inta_irq) {
  485. #if defined(CONFIG_BF53x)
  486. case IRQ_PROG_INTA:
  487. irq = IRQ_PF0;
  488. search = 1;
  489. break;
  490. # if defined(BF537_FAMILY) && !(defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE))
  491. case IRQ_MAC_RX:
  492. irq = IRQ_PH0;
  493. break;
  494. # endif
  495. #elif defined(CONFIG_BF538) || defined(CONFIG_BF539)
  496. case IRQ_PORTF_INTA:
  497. irq = IRQ_PF0;
  498. break;
  499. #elif defined(CONFIG_BF52x) || defined(CONFIG_BF51x)
  500. case IRQ_PORTF_INTA:
  501. irq = IRQ_PF0;
  502. break;
  503. case IRQ_PORTG_INTA:
  504. irq = IRQ_PG0;
  505. break;
  506. case IRQ_PORTH_INTA:
  507. irq = IRQ_PH0;
  508. break;
  509. #elif defined(CONFIG_BF561)
  510. case IRQ_PROG0_INTA:
  511. irq = IRQ_PF0;
  512. break;
  513. case IRQ_PROG1_INTA:
  514. irq = IRQ_PF16;
  515. break;
  516. case IRQ_PROG2_INTA:
  517. irq = IRQ_PF32;
  518. break;
  519. #endif
  520. default:
  521. BUG();
  522. return;
  523. }
  524. if (search) {
  525. for (i = 0; i < MAX_BLACKFIN_GPIOS; i += GPIO_BANKSIZE) {
  526. irq += i;
  527. mask = get_gpiop_data(i) & get_gpiop_maska(i);
  528. while (mask) {
  529. if (mask & 1)
  530. bfin_handle_irq(irq);
  531. irq++;
  532. mask >>= 1;
  533. }
  534. }
  535. } else {
  536. gpio = irq_to_gpio(irq);
  537. mask = get_gpiop_data(gpio) & get_gpiop_maska(gpio);
  538. do {
  539. if (mask & 1)
  540. bfin_handle_irq(irq);
  541. irq++;
  542. mask >>= 1;
  543. } while (mask);
  544. }
  545. }
  546. #else /* CONFIG_BF54x */
  547. #define NR_PINT_SYS_IRQS 4
  548. #define NR_PINT_BITS 32
  549. #define NR_PINTS 160
  550. #define IRQ_NOT_AVAIL 0xFF
  551. #define PINT_2_BANK(x) ((x) >> 5)
  552. #define PINT_2_BIT(x) ((x) & 0x1F)
  553. #define PINT_BIT(x) (1 << (PINT_2_BIT(x)))
  554. static unsigned char irq2pint_lut[NR_PINTS];
  555. static unsigned char pint2irq_lut[NR_PINT_SYS_IRQS * NR_PINT_BITS];
  556. struct pin_int_t {
  557. unsigned int mask_set;
  558. unsigned int mask_clear;
  559. unsigned int request;
  560. unsigned int assign;
  561. unsigned int edge_set;
  562. unsigned int edge_clear;
  563. unsigned int invert_set;
  564. unsigned int invert_clear;
  565. unsigned int pinstate;
  566. unsigned int latch;
  567. };
  568. static struct pin_int_t *pint[NR_PINT_SYS_IRQS] = {
  569. (struct pin_int_t *)PINT0_MASK_SET,
  570. (struct pin_int_t *)PINT1_MASK_SET,
  571. (struct pin_int_t *)PINT2_MASK_SET,
  572. (struct pin_int_t *)PINT3_MASK_SET,
  573. };
  574. inline unsigned int get_irq_base(u32 bank, u8 bmap)
  575. {
  576. unsigned int irq_base;
  577. if (bank < 2) { /*PA-PB */
  578. irq_base = IRQ_PA0 + bmap * 16;
  579. } else { /*PC-PJ */
  580. irq_base = IRQ_PC0 + bmap * 16;
  581. }
  582. return irq_base;
  583. }
  584. /* Whenever PINTx_ASSIGN is altered init_pint_lut() must be executed! */
  585. void init_pint_lut(void)
  586. {
  587. u16 bank, bit, irq_base, bit_pos;
  588. u32 pint_assign;
  589. u8 bmap;
  590. memset(irq2pint_lut, IRQ_NOT_AVAIL, sizeof(irq2pint_lut));
  591. for (bank = 0; bank < NR_PINT_SYS_IRQS; bank++) {
  592. pint_assign = pint[bank]->assign;
  593. for (bit = 0; bit < NR_PINT_BITS; bit++) {
  594. bmap = (pint_assign >> ((bit / 8) * 8)) & 0xFF;
  595. irq_base = get_irq_base(bank, bmap);
  596. irq_base += (bit % 8) + ((bit / 8) & 1 ? 8 : 0);
  597. bit_pos = bit + bank * NR_PINT_BITS;
  598. pint2irq_lut[bit_pos] = irq_base - SYS_IRQS;
  599. irq2pint_lut[irq_base - SYS_IRQS] = bit_pos;
  600. }
  601. }
  602. }
  603. static void bfin_gpio_ack_irq(unsigned int irq)
  604. {
  605. struct irq_desc *desc = irq_desc + irq;
  606. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  607. u32 pintbit = PINT_BIT(pint_val);
  608. u32 bank = PINT_2_BANK(pint_val);
  609. if ((desc->status & IRQ_TYPE_SENSE_MASK) == IRQ_TYPE_EDGE_BOTH) {
  610. if (pint[bank]->invert_set & pintbit)
  611. pint[bank]->invert_clear = pintbit;
  612. else
  613. pint[bank]->invert_set = pintbit;
  614. }
  615. pint[bank]->request = pintbit;
  616. }
  617. static void bfin_gpio_mask_ack_irq(unsigned int irq)
  618. {
  619. struct irq_desc *desc = irq_desc + irq;
  620. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  621. u32 pintbit = PINT_BIT(pint_val);
  622. u32 bank = PINT_2_BANK(pint_val);
  623. if ((desc->status & IRQ_TYPE_SENSE_MASK) == IRQ_TYPE_EDGE_BOTH) {
  624. if (pint[bank]->invert_set & pintbit)
  625. pint[bank]->invert_clear = pintbit;
  626. else
  627. pint[bank]->invert_set = pintbit;
  628. }
  629. pint[bank]->request = pintbit;
  630. pint[bank]->mask_clear = pintbit;
  631. }
  632. static void bfin_gpio_mask_irq(unsigned int irq)
  633. {
  634. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  635. pint[PINT_2_BANK(pint_val)]->mask_clear = PINT_BIT(pint_val);
  636. }
  637. static void bfin_gpio_unmask_irq(unsigned int irq)
  638. {
  639. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  640. u32 pintbit = PINT_BIT(pint_val);
  641. u32 bank = PINT_2_BANK(pint_val);
  642. pint[bank]->request = pintbit;
  643. pint[bank]->mask_set = pintbit;
  644. }
  645. static unsigned int bfin_gpio_irq_startup(unsigned int irq)
  646. {
  647. u32 gpionr = irq_to_gpio(irq);
  648. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  649. if (pint_val == IRQ_NOT_AVAIL) {
  650. printk(KERN_ERR
  651. "GPIO IRQ %d :Not in PINT Assign table "
  652. "Reconfigure Interrupt to Port Assignemt\n", irq);
  653. return -ENODEV;
  654. }
  655. if (__test_and_set_bit(gpionr, gpio_enabled))
  656. bfin_gpio_irq_prepare(gpionr);
  657. bfin_gpio_unmask_irq(irq);
  658. return 0;
  659. }
  660. static void bfin_gpio_irq_shutdown(unsigned int irq)
  661. {
  662. u32 gpionr = irq_to_gpio(irq);
  663. bfin_gpio_mask_irq(irq);
  664. __clear_bit(gpionr, gpio_enabled);
  665. bfin_gpio_irq_free(gpionr);
  666. }
  667. static int bfin_gpio_irq_type(unsigned int irq, unsigned int type)
  668. {
  669. int ret;
  670. char buf[16];
  671. u32 gpionr = irq_to_gpio(irq);
  672. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  673. u32 pintbit = PINT_BIT(pint_val);
  674. u32 bank = PINT_2_BANK(pint_val);
  675. if (pint_val == IRQ_NOT_AVAIL)
  676. return -ENODEV;
  677. if (type == IRQ_TYPE_PROBE) {
  678. /* only probe unenabled GPIO interrupt lines */
  679. if (test_bit(gpionr, gpio_enabled))
  680. return 0;
  681. type = IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING;
  682. }
  683. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING |
  684. IRQ_TYPE_LEVEL_HIGH | IRQ_TYPE_LEVEL_LOW)) {
  685. snprintf(buf, 16, "gpio-irq%d", irq);
  686. ret = bfin_gpio_irq_request(gpionr, buf);
  687. if (ret)
  688. return ret;
  689. if (__test_and_set_bit(gpionr, gpio_enabled))
  690. bfin_gpio_irq_prepare(gpionr);
  691. } else {
  692. __clear_bit(gpionr, gpio_enabled);
  693. return 0;
  694. }
  695. if ((type & (IRQ_TYPE_EDGE_FALLING | IRQ_TYPE_LEVEL_LOW)))
  696. pint[bank]->invert_set = pintbit; /* low or falling edge denoted by one */
  697. else
  698. pint[bank]->invert_clear = pintbit; /* high or rising edge denoted by zero */
  699. if ((type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING))
  700. == (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  701. if (gpio_get_value(gpionr))
  702. pint[bank]->invert_set = pintbit;
  703. else
  704. pint[bank]->invert_clear = pintbit;
  705. }
  706. if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
  707. pint[bank]->edge_set = pintbit;
  708. bfin_set_irq_handler(irq, handle_edge_irq);
  709. } else {
  710. pint[bank]->edge_clear = pintbit;
  711. bfin_set_irq_handler(irq, handle_level_irq);
  712. }
  713. return 0;
  714. }
  715. #ifdef CONFIG_PM
  716. u32 pint_saved_masks[NR_PINT_SYS_IRQS];
  717. u32 pint_wakeup_masks[NR_PINT_SYS_IRQS];
  718. int bfin_gpio_set_wake(unsigned int irq, unsigned int state)
  719. {
  720. u32 pint_irq;
  721. u32 pint_val = irq2pint_lut[irq - SYS_IRQS];
  722. u32 bank = PINT_2_BANK(pint_val);
  723. u32 pintbit = PINT_BIT(pint_val);
  724. switch (bank) {
  725. case 0:
  726. pint_irq = IRQ_PINT0;
  727. break;
  728. case 2:
  729. pint_irq = IRQ_PINT2;
  730. break;
  731. case 3:
  732. pint_irq = IRQ_PINT3;
  733. break;
  734. case 1:
  735. pint_irq = IRQ_PINT1;
  736. break;
  737. default:
  738. return -EINVAL;
  739. }
  740. bfin_internal_set_wake(pint_irq, state);
  741. if (state)
  742. pint_wakeup_masks[bank] |= pintbit;
  743. else
  744. pint_wakeup_masks[bank] &= ~pintbit;
  745. return 0;
  746. }
  747. u32 bfin_pm_setup(void)
  748. {
  749. u32 val, i;
  750. for (i = 0; i < NR_PINT_SYS_IRQS; i++) {
  751. val = pint[i]->mask_clear;
  752. pint_saved_masks[i] = val;
  753. if (val ^ pint_wakeup_masks[i]) {
  754. pint[i]->mask_clear = val;
  755. pint[i]->mask_set = pint_wakeup_masks[i];
  756. }
  757. }
  758. return 0;
  759. }
  760. void bfin_pm_restore(void)
  761. {
  762. u32 i, val;
  763. for (i = 0; i < NR_PINT_SYS_IRQS; i++) {
  764. val = pint_saved_masks[i];
  765. if (val ^ pint_wakeup_masks[i]) {
  766. pint[i]->mask_clear = pint[i]->mask_clear;
  767. pint[i]->mask_set = val;
  768. }
  769. }
  770. }
  771. #endif
  772. static void bfin_demux_gpio_irq(unsigned int inta_irq,
  773. struct irq_desc *desc)
  774. {
  775. u32 bank, pint_val;
  776. u32 request, irq;
  777. switch (inta_irq) {
  778. case IRQ_PINT0:
  779. bank = 0;
  780. break;
  781. case IRQ_PINT2:
  782. bank = 2;
  783. break;
  784. case IRQ_PINT3:
  785. bank = 3;
  786. break;
  787. case IRQ_PINT1:
  788. bank = 1;
  789. break;
  790. default:
  791. return;
  792. }
  793. pint_val = bank * NR_PINT_BITS;
  794. request = pint[bank]->request;
  795. while (request) {
  796. if (request & 1) {
  797. irq = pint2irq_lut[pint_val] + SYS_IRQS;
  798. bfin_handle_irq(irq);
  799. }
  800. pint_val++;
  801. request >>= 1;
  802. }
  803. }
  804. #endif
  805. static struct irq_chip bfin_gpio_irqchip = {
  806. .name = "GPIO",
  807. .ack = bfin_gpio_ack_irq,
  808. .mask = bfin_gpio_mask_irq,
  809. .mask_ack = bfin_gpio_mask_ack_irq,
  810. .unmask = bfin_gpio_unmask_irq,
  811. .disable = bfin_gpio_mask_irq,
  812. .enable = bfin_gpio_unmask_irq,
  813. .set_type = bfin_gpio_irq_type,
  814. .startup = bfin_gpio_irq_startup,
  815. .shutdown = bfin_gpio_irq_shutdown,
  816. #ifdef CONFIG_PM
  817. .set_wake = bfin_gpio_set_wake,
  818. #endif
  819. };
  820. void __cpuinit init_exception_vectors(void)
  821. {
  822. /* cannot program in software:
  823. * evt0 - emulation (jtag)
  824. * evt1 - reset
  825. */
  826. bfin_write_EVT2(evt_nmi);
  827. bfin_write_EVT3(trap);
  828. bfin_write_EVT5(evt_ivhw);
  829. bfin_write_EVT6(evt_timer);
  830. bfin_write_EVT7(evt_evt7);
  831. bfin_write_EVT8(evt_evt8);
  832. bfin_write_EVT9(evt_evt9);
  833. bfin_write_EVT10(evt_evt10);
  834. bfin_write_EVT11(evt_evt11);
  835. bfin_write_EVT12(evt_evt12);
  836. bfin_write_EVT13(evt_evt13);
  837. bfin_write_EVT14(evt_evt14);
  838. bfin_write_EVT15(evt_system_call);
  839. CSYNC();
  840. }
  841. /*
  842. * This function should be called during kernel startup to initialize
  843. * the BFin IRQ handling routines.
  844. */
  845. int __init init_arch_irq(void)
  846. {
  847. int irq;
  848. unsigned long ilat = 0;
  849. /* Disable all the peripheral intrs - page 4-29 HW Ref manual */
  850. #if defined(CONFIG_BF54x) || defined(CONFIG_BF52x) || defined(CONFIG_BF561) \
  851. || defined(BF538_FAMILY) || defined(CONFIG_BF51x)
  852. bfin_write_SIC_IMASK0(SIC_UNMASK_ALL);
  853. bfin_write_SIC_IMASK1(SIC_UNMASK_ALL);
  854. # ifdef CONFIG_BF54x
  855. bfin_write_SIC_IMASK2(SIC_UNMASK_ALL);
  856. # endif
  857. # ifdef CONFIG_SMP
  858. bfin_write_SICB_IMASK0(SIC_UNMASK_ALL);
  859. bfin_write_SICB_IMASK1(SIC_UNMASK_ALL);
  860. # endif
  861. #else
  862. bfin_write_SIC_IMASK(SIC_UNMASK_ALL);
  863. #endif
  864. local_irq_disable();
  865. #if (defined(CONFIG_BF537) || defined(CONFIG_BF536))
  866. /* Clear EMAC Interrupt Status bits so we can demux it later */
  867. bfin_write_EMAC_SYSTAT(-1);
  868. #endif
  869. #ifdef CONFIG_BF54x
  870. # ifdef CONFIG_PINTx_REASSIGN
  871. pint[0]->assign = CONFIG_PINT0_ASSIGN;
  872. pint[1]->assign = CONFIG_PINT1_ASSIGN;
  873. pint[2]->assign = CONFIG_PINT2_ASSIGN;
  874. pint[3]->assign = CONFIG_PINT3_ASSIGN;
  875. # endif
  876. /* Whenever PINTx_ASSIGN is altered init_pint_lut() must be executed! */
  877. init_pint_lut();
  878. #endif
  879. for (irq = 0; irq <= SYS_IRQS; irq++) {
  880. if (irq <= IRQ_CORETMR)
  881. set_irq_chip(irq, &bfin_core_irqchip);
  882. else
  883. set_irq_chip(irq, &bfin_internal_irqchip);
  884. switch (irq) {
  885. #if defined(CONFIG_BF53x)
  886. case IRQ_PROG_INTA:
  887. # if defined(BF537_FAMILY) && !(defined(CONFIG_BFIN_MAC) || defined(CONFIG_BFIN_MAC_MODULE))
  888. case IRQ_MAC_RX:
  889. # endif
  890. #elif defined(CONFIG_BF54x)
  891. case IRQ_PINT0:
  892. case IRQ_PINT1:
  893. case IRQ_PINT2:
  894. case IRQ_PINT3:
  895. #elif defined(CONFIG_BF52x) || defined(CONFIG_BF51x)
  896. case IRQ_PORTF_INTA:
  897. case IRQ_PORTG_INTA:
  898. case IRQ_PORTH_INTA:
  899. #elif defined(CONFIG_BF561)
  900. case IRQ_PROG0_INTA:
  901. case IRQ_PROG1_INTA:
  902. case IRQ_PROG2_INTA:
  903. #elif defined(CONFIG_BF538) || defined(CONFIG_BF539)
  904. case IRQ_PORTF_INTA:
  905. #endif
  906. set_irq_chained_handler(irq,
  907. bfin_demux_gpio_irq);
  908. break;
  909. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  910. case IRQ_GENERIC_ERROR:
  911. set_irq_chained_handler(irq, bfin_demux_error_irq);
  912. break;
  913. #endif
  914. #ifdef CONFIG_SMP
  915. case IRQ_SUPPLE_0:
  916. case IRQ_SUPPLE_1:
  917. set_irq_handler(irq, handle_percpu_irq);
  918. break;
  919. #endif
  920. #ifdef CONFIG_TICKSOURCE_CORETMR
  921. case IRQ_CORETMR:
  922. # ifdef CONFIG_SMP
  923. set_irq_handler(irq, handle_percpu_irq);
  924. break;
  925. # else
  926. set_irq_handler(irq, handle_simple_irq);
  927. break;
  928. # endif
  929. #endif
  930. #ifdef CONFIG_TICKSOURCE_GPTMR0
  931. case IRQ_TIMER0:
  932. set_irq_handler(irq, handle_simple_irq);
  933. break;
  934. #endif
  935. #ifdef CONFIG_IPIPE
  936. default:
  937. set_irq_handler(irq, handle_level_irq);
  938. break;
  939. #else /* !CONFIG_IPIPE */
  940. default:
  941. set_irq_handler(irq, handle_simple_irq);
  942. break;
  943. #endif /* !CONFIG_IPIPE */
  944. }
  945. }
  946. #ifdef BF537_GENERIC_ERROR_INT_DEMUX
  947. for (irq = IRQ_PPI_ERROR; irq <= IRQ_UART1_ERROR; irq++)
  948. set_irq_chip_and_handler(irq, &bfin_generic_error_irqchip,
  949. handle_level_irq);
  950. #endif
  951. /* if configured as edge, then will be changed to do_edge_IRQ */
  952. for (irq = GPIO_IRQ_BASE; irq < NR_MACH_IRQS; irq++)
  953. set_irq_chip_and_handler(irq, &bfin_gpio_irqchip,
  954. handle_level_irq);
  955. bfin_write_IMASK(0);
  956. CSYNC();
  957. ilat = bfin_read_ILAT();
  958. CSYNC();
  959. bfin_write_ILAT(ilat);
  960. CSYNC();
  961. printk(KERN_INFO "Configuring Blackfin Priority Driven Interrupts\n");
  962. /* IMASK=xxx is equivalent to STI xx or bfin_irq_flags=xx,
  963. * local_irq_enable()
  964. */
  965. program_IAR();
  966. /* Therefore it's better to setup IARs before interrupts enabled */
  967. search_IAR();
  968. /* Enable interrupts IVG7-15 */
  969. bfin_irq_flags |= IMASK_IVG15 |
  970. IMASK_IVG14 | IMASK_IVG13 | IMASK_IVG12 | IMASK_IVG11 |
  971. IMASK_IVG10 | IMASK_IVG9 | IMASK_IVG8 | IMASK_IVG7 | IMASK_IVGHW;
  972. /* This implicitly covers ANOMALY_05000171
  973. * Boot-ROM code modifies SICA_IWRx wakeup registers
  974. */
  975. #ifdef SIC_IWR0
  976. bfin_write_SIC_IWR0(IWR_DISABLE_ALL);
  977. # ifdef SIC_IWR1
  978. /* BF52x/BF51x system reset does not properly reset SIC_IWR1 which
  979. * will screw up the bootrom as it relies on MDMA0/1 waking it
  980. * up from IDLE instructions. See this report for more info:
  981. * http://blackfin.uclinux.org/gf/tracker/4323
  982. */
  983. if (ANOMALY_05000435)
  984. bfin_write_SIC_IWR1(IWR_ENABLE(10) | IWR_ENABLE(11));
  985. else
  986. bfin_write_SIC_IWR1(IWR_DISABLE_ALL);
  987. # endif
  988. # ifdef SIC_IWR2
  989. bfin_write_SIC_IWR2(IWR_DISABLE_ALL);
  990. # endif
  991. #else
  992. bfin_write_SIC_IWR(IWR_DISABLE_ALL);
  993. #endif
  994. return 0;
  995. }
  996. #ifdef CONFIG_DO_IRQ_L1
  997. __attribute__((l1_text))
  998. #endif
  999. void do_irq(int vec, struct pt_regs *fp)
  1000. {
  1001. if (vec == EVT_IVTMR_P) {
  1002. vec = IRQ_CORETMR;
  1003. } else {
  1004. struct ivgx *ivg = ivg7_13[vec - IVG7].ifirst;
  1005. struct ivgx *ivg_stop = ivg7_13[vec - IVG7].istop;
  1006. #if defined(SIC_ISR0) || defined(SICA_ISR0)
  1007. unsigned long sic_status[3];
  1008. if (smp_processor_id()) {
  1009. # ifdef SICB_ISR0
  1010. /* This will be optimized out in UP mode. */
  1011. sic_status[0] = bfin_read_SICB_ISR0() & bfin_read_SICB_IMASK0();
  1012. sic_status[1] = bfin_read_SICB_ISR1() & bfin_read_SICB_IMASK1();
  1013. # endif
  1014. } else {
  1015. sic_status[0] = bfin_read_SIC_ISR0() & bfin_read_SIC_IMASK0();
  1016. sic_status[1] = bfin_read_SIC_ISR1() & bfin_read_SIC_IMASK1();
  1017. }
  1018. # ifdef SIC_ISR2
  1019. sic_status[2] = bfin_read_SIC_ISR2() & bfin_read_SIC_IMASK2();
  1020. # endif
  1021. for (;; ivg++) {
  1022. if (ivg >= ivg_stop) {
  1023. atomic_inc(&num_spurious);
  1024. return;
  1025. }
  1026. if (sic_status[(ivg->irqno - IVG7) / 32] & ivg->isrflag)
  1027. break;
  1028. }
  1029. #else
  1030. unsigned long sic_status;
  1031. sic_status = bfin_read_SIC_IMASK() & bfin_read_SIC_ISR();
  1032. for (;; ivg++) {
  1033. if (ivg >= ivg_stop) {
  1034. atomic_inc(&num_spurious);
  1035. return;
  1036. } else if (sic_status & ivg->isrflag)
  1037. break;
  1038. }
  1039. #endif
  1040. vec = ivg->irqno;
  1041. }
  1042. asm_do_IRQ(vec, fp);
  1043. }
  1044. #ifdef CONFIG_IPIPE
  1045. int __ipipe_get_irq_priority(unsigned irq)
  1046. {
  1047. int ient, prio;
  1048. if (irq <= IRQ_CORETMR)
  1049. return irq;
  1050. for (ient = 0; ient < NR_PERI_INTS; ient++) {
  1051. struct ivgx *ivg = ivg_table + ient;
  1052. if (ivg->irqno == irq) {
  1053. for (prio = 0; prio <= IVG13-IVG7; prio++) {
  1054. if (ivg7_13[prio].ifirst <= ivg &&
  1055. ivg7_13[prio].istop > ivg)
  1056. return IVG7 + prio;
  1057. }
  1058. }
  1059. }
  1060. return IVG15;
  1061. }
  1062. /* Hw interrupts are disabled on entry (check SAVE_CONTEXT). */
  1063. #ifdef CONFIG_DO_IRQ_L1
  1064. __attribute__((l1_text))
  1065. #endif
  1066. asmlinkage int __ipipe_grab_irq(int vec, struct pt_regs *regs)
  1067. {
  1068. struct ipipe_percpu_domain_data *p = ipipe_root_cpudom_ptr();
  1069. struct ipipe_domain *this_domain = __ipipe_current_domain;
  1070. struct ivgx *ivg_stop = ivg7_13[vec-IVG7].istop;
  1071. struct ivgx *ivg = ivg7_13[vec-IVG7].ifirst;
  1072. int irq, s;
  1073. if (likely(vec == EVT_IVTMR_P))
  1074. irq = IRQ_CORETMR;
  1075. else {
  1076. #if defined(SIC_ISR0) || defined(SICA_ISR0)
  1077. unsigned long sic_status[3];
  1078. sic_status[0] = bfin_read_SIC_ISR0() & bfin_read_SIC_IMASK0();
  1079. sic_status[1] = bfin_read_SIC_ISR1() & bfin_read_SIC_IMASK1();
  1080. # ifdef SIC_ISR2
  1081. sic_status[2] = bfin_read_SIC_ISR2() & bfin_read_SIC_IMASK2();
  1082. # endif
  1083. for (;; ivg++) {
  1084. if (ivg >= ivg_stop) {
  1085. atomic_inc(&num_spurious);
  1086. return 0;
  1087. }
  1088. if (sic_status[(ivg->irqno - IVG7) / 32] & ivg->isrflag)
  1089. break;
  1090. }
  1091. #else
  1092. unsigned long sic_status;
  1093. sic_status = bfin_read_SIC_IMASK() & bfin_read_SIC_ISR();
  1094. for (;; ivg++) {
  1095. if (ivg >= ivg_stop) {
  1096. atomic_inc(&num_spurious);
  1097. return 0;
  1098. } else if (sic_status & ivg->isrflag)
  1099. break;
  1100. }
  1101. #endif
  1102. irq = ivg->irqno;
  1103. }
  1104. if (irq == IRQ_SYSTMR) {
  1105. #if !defined(CONFIG_GENERIC_CLOCKEVENTS) || defined(CONFIG_TICKSOURCE_GPTMR0)
  1106. bfin_write_TIMER_STATUS(1); /* Latch TIMIL0 */
  1107. #endif
  1108. /* This is basically what we need from the register frame. */
  1109. __raw_get_cpu_var(__ipipe_tick_regs).ipend = regs->ipend;
  1110. __raw_get_cpu_var(__ipipe_tick_regs).pc = regs->pc;
  1111. if (this_domain != ipipe_root_domain)
  1112. __raw_get_cpu_var(__ipipe_tick_regs).ipend &= ~0x10;
  1113. else
  1114. __raw_get_cpu_var(__ipipe_tick_regs).ipend |= 0x10;
  1115. }
  1116. if (this_domain == ipipe_root_domain) {
  1117. s = __test_and_set_bit(IPIPE_SYNCDEFER_FLAG, &p->status);
  1118. barrier();
  1119. }
  1120. ipipe_trace_irq_entry(irq);
  1121. __ipipe_handle_irq(irq, regs);
  1122. ipipe_trace_irq_exit(irq);
  1123. if (this_domain == ipipe_root_domain) {
  1124. set_thread_flag(TIF_IRQ_SYNC);
  1125. if (!s) {
  1126. __clear_bit(IPIPE_SYNCDEFER_FLAG, &p->status);
  1127. return !test_bit(IPIPE_STALL_FLAG, &p->status);
  1128. }
  1129. }
  1130. return 0;
  1131. }
  1132. #endif /* CONFIG_IPIPE */