dma.h 8.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221
  1. /* $Id: dma.h,v 1.21 2001/12/13 04:16:52 davem Exp $
  2. * include/asm-sparc64/dma.h
  3. *
  4. * Copyright 1996 (C) David S. Miller (davem@caip.rutgers.edu)
  5. */
  6. #ifndef _ASM_SPARC64_DMA_H
  7. #define _ASM_SPARC64_DMA_H
  8. #include <linux/config.h>
  9. #include <linux/kernel.h>
  10. #include <linux/types.h>
  11. #include <linux/spinlock.h>
  12. #include <asm/sbus.h>
  13. #include <asm/delay.h>
  14. #include <asm/oplib.h>
  15. extern spinlock_t dma_spin_lock;
  16. #define claim_dma_lock() \
  17. ({ unsigned long flags; \
  18. spin_lock_irqsave(&dma_spin_lock, flags); \
  19. flags; \
  20. })
  21. #define release_dma_lock(__flags) \
  22. spin_unlock_irqrestore(&dma_spin_lock, __flags);
  23. /* These are irrelevant for Sparc DMA, but we leave it in so that
  24. * things can compile.
  25. */
  26. #define MAX_DMA_CHANNELS 8
  27. #define DMA_MODE_READ 1
  28. #define DMA_MODE_WRITE 2
  29. #define MAX_DMA_ADDRESS (~0UL)
  30. /* Useful constants */
  31. #define SIZE_16MB (16*1024*1024)
  32. #define SIZE_64K (64*1024)
  33. /* SBUS DMA controller reg offsets */
  34. #define DMA_CSR 0x00UL /* rw DMA control/status register 0x00 */
  35. #define DMA_ADDR 0x04UL /* rw DMA transfer address register 0x04 */
  36. #define DMA_COUNT 0x08UL /* rw DMA transfer count register 0x08 */
  37. #define DMA_TEST 0x0cUL /* rw DMA test/debug register 0x0c */
  38. /* DVMA chip revisions */
  39. enum dvma_rev {
  40. dvmarev0,
  41. dvmaesc1,
  42. dvmarev1,
  43. dvmarev2,
  44. dvmarev3,
  45. dvmarevplus,
  46. dvmahme
  47. };
  48. #define DMA_HASCOUNT(rev) ((rev)==dvmaesc1)
  49. /* Linux DMA information structure, filled during probe. */
  50. struct sbus_dma {
  51. struct sbus_dma *next;
  52. struct sbus_dev *sdev;
  53. void __iomem *regs;
  54. /* Status, misc info */
  55. int node; /* Prom node for this DMA device */
  56. int running; /* Are we doing DMA now? */
  57. int allocated; /* Are we "owned" by anyone yet? */
  58. /* Transfer information. */
  59. u32 addr; /* Start address of current transfer */
  60. int nbytes; /* Size of current transfer */
  61. int realbytes; /* For splitting up large transfers, etc. */
  62. /* DMA revision */
  63. enum dvma_rev revision;
  64. };
  65. extern struct sbus_dma *dma_chain;
  66. /* Broken hardware... */
  67. #define DMA_ISBROKEN(dma) ((dma)->revision == dvmarev1)
  68. #define DMA_ISESC1(dma) ((dma)->revision == dvmaesc1)
  69. /* Main routines in dma.c */
  70. extern void dvma_init(struct sbus_bus *);
  71. /* Fields in the cond_reg register */
  72. /* First, the version identification bits */
  73. #define DMA_DEVICE_ID 0xf0000000 /* Device identification bits */
  74. #define DMA_VERS0 0x00000000 /* Sunray DMA version */
  75. #define DMA_ESCV1 0x40000000 /* DMA ESC Version 1 */
  76. #define DMA_VERS1 0x80000000 /* DMA rev 1 */
  77. #define DMA_VERS2 0xa0000000 /* DMA rev 2 */
  78. #define DMA_VERHME 0xb0000000 /* DMA hme gate array */
  79. #define DMA_VERSPLUS 0x90000000 /* DMA rev 1 PLUS */
  80. #define DMA_HNDL_INTR 0x00000001 /* An IRQ needs to be handled */
  81. #define DMA_HNDL_ERROR 0x00000002 /* We need to take an error */
  82. #define DMA_FIFO_ISDRAIN 0x0000000c /* The DMA FIFO is draining */
  83. #define DMA_INT_ENAB 0x00000010 /* Turn on interrupts */
  84. #define DMA_FIFO_INV 0x00000020 /* Invalidate the FIFO */
  85. #define DMA_ACC_SZ_ERR 0x00000040 /* The access size was bad */
  86. #define DMA_FIFO_STDRAIN 0x00000040 /* DMA_VERS1 Drain the FIFO */
  87. #define DMA_RST_SCSI 0x00000080 /* Reset the SCSI controller */
  88. #define DMA_RST_ENET DMA_RST_SCSI /* Reset the ENET controller */
  89. #define DMA_ST_WRITE 0x00000100 /* write from device to memory */
  90. #define DMA_ENABLE 0x00000200 /* Fire up DMA, handle requests */
  91. #define DMA_PEND_READ 0x00000400 /* DMA_VERS1/0/PLUS Pending Read */
  92. #define DMA_ESC_BURST 0x00000800 /* 1=16byte 0=32byte */
  93. #define DMA_READ_AHEAD 0x00001800 /* DMA read ahead partial longword */
  94. #define DMA_DSBL_RD_DRN 0x00001000 /* No EC drain on slave reads */
  95. #define DMA_BCNT_ENAB 0x00002000 /* If on, use the byte counter */
  96. #define DMA_TERM_CNTR 0x00004000 /* Terminal counter */
  97. #define DMA_SCSI_SBUS64 0x00008000 /* HME: Enable 64-bit SBUS mode. */
  98. #define DMA_CSR_DISAB 0x00010000 /* No FIFO drains during csr */
  99. #define DMA_SCSI_DISAB 0x00020000 /* No FIFO drains during reg */
  100. #define DMA_DSBL_WR_INV 0x00020000 /* No EC inval. on slave writes */
  101. #define DMA_ADD_ENABLE 0x00040000 /* Special ESC DVMA optimization */
  102. #define DMA_E_BURSTS 0x000c0000 /* ENET: SBUS r/w burst mask */
  103. #define DMA_E_BURST32 0x00040000 /* ENET: SBUS 32 byte r/w burst */
  104. #define DMA_E_BURST16 0x00000000 /* ENET: SBUS 16 byte r/w burst */
  105. #define DMA_BRST_SZ 0x000c0000 /* SCSI: SBUS r/w burst size */
  106. #define DMA_BRST64 0x000c0000 /* SCSI: 64byte bursts (HME on UltraSparc only) */
  107. #define DMA_BRST32 0x00040000 /* SCSI: 32byte bursts */
  108. #define DMA_BRST16 0x00000000 /* SCSI: 16byte bursts */
  109. #define DMA_BRST0 0x00080000 /* SCSI: no bursts (non-HME gate arrays) */
  110. #define DMA_ADDR_DISAB 0x00100000 /* No FIFO drains during addr */
  111. #define DMA_2CLKS 0x00200000 /* Each transfer = 2 clock ticks */
  112. #define DMA_3CLKS 0x00400000 /* Each transfer = 3 clock ticks */
  113. #define DMA_EN_ENETAUI DMA_3CLKS /* Put lance into AUI-cable mode */
  114. #define DMA_CNTR_DISAB 0x00800000 /* No IRQ when DMA_TERM_CNTR set */
  115. #define DMA_AUTO_NADDR 0x01000000 /* Use "auto nxt addr" feature */
  116. #define DMA_SCSI_ON 0x02000000 /* Enable SCSI dma */
  117. #define DMA_PARITY_OFF 0x02000000 /* HME: disable parity checking */
  118. #define DMA_LOADED_ADDR 0x04000000 /* Address has been loaded */
  119. #define DMA_LOADED_NADDR 0x08000000 /* Next address has been loaded */
  120. #define DMA_RESET_FAS366 0x08000000 /* HME: Assert RESET to FAS366 */
  121. /* Values describing the burst-size property from the PROM */
  122. #define DMA_BURST1 0x01
  123. #define DMA_BURST2 0x02
  124. #define DMA_BURST4 0x04
  125. #define DMA_BURST8 0x08
  126. #define DMA_BURST16 0x10
  127. #define DMA_BURST32 0x20
  128. #define DMA_BURST64 0x40
  129. #define DMA_BURSTBITS 0x7f
  130. /* Determine highest possible final transfer address given a base */
  131. #define DMA_MAXEND(addr) (0x01000000UL-(((unsigned long)(addr))&0x00ffffffUL))
  132. /* Yes, I hack a lot of elisp in my spare time... */
  133. #define DMA_ERROR_P(regs) (((sbus_readl((regs) + DMA_CSR) & DMA_HNDL_ERROR))
  134. #define DMA_IRQ_P(regs) (((sbus_readl((regs) + DMA_CSR)) & (DMA_HNDL_INTR | DMA_HNDL_ERROR)))
  135. #define DMA_WRITE_P(regs) (((sbus_readl((regs) + DMA_CSR) & DMA_ST_WRITE))
  136. #define DMA_OFF(__regs) \
  137. do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
  138. tmp &= ~DMA_ENABLE; \
  139. sbus_writel(tmp, (__regs) + DMA_CSR); \
  140. } while(0)
  141. #define DMA_INTSOFF(__regs) \
  142. do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
  143. tmp &= ~DMA_INT_ENAB; \
  144. sbus_writel(tmp, (__regs) + DMA_CSR); \
  145. } while(0)
  146. #define DMA_INTSON(__regs) \
  147. do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
  148. tmp |= DMA_INT_ENAB; \
  149. sbus_writel(tmp, (__regs) + DMA_CSR); \
  150. } while(0)
  151. #define DMA_PUNTFIFO(__regs) \
  152. do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
  153. tmp |= DMA_FIFO_INV; \
  154. sbus_writel(tmp, (__regs) + DMA_CSR); \
  155. } while(0)
  156. #define DMA_SETSTART(__regs, __addr) \
  157. sbus_writel((u32)(__addr), (__regs) + DMA_ADDR);
  158. #define DMA_BEGINDMA_W(__regs) \
  159. do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
  160. tmp |= (DMA_ST_WRITE|DMA_ENABLE|DMA_INT_ENAB); \
  161. sbus_writel(tmp, (__regs) + DMA_CSR); \
  162. } while(0)
  163. #define DMA_BEGINDMA_R(__regs) \
  164. do { u32 tmp = sbus_readl((__regs) + DMA_CSR); \
  165. tmp |= (DMA_ENABLE|DMA_INT_ENAB); \
  166. tmp &= ~DMA_ST_WRITE; \
  167. sbus_writel(tmp, (__regs) + DMA_CSR); \
  168. } while(0)
  169. /* For certain DMA chips, we need to disable ints upon irq entry
  170. * and turn them back on when we are done. So in any ESP interrupt
  171. * handler you *must* call DMA_IRQ_ENTRY upon entry and DMA_IRQ_EXIT
  172. * when leaving the handler. You have been warned...
  173. */
  174. #define DMA_IRQ_ENTRY(dma, dregs) do { \
  175. if(DMA_ISBROKEN(dma)) DMA_INTSOFF(dregs); \
  176. } while (0)
  177. #define DMA_IRQ_EXIT(dma, dregs) do { \
  178. if(DMA_ISBROKEN(dma)) DMA_INTSON(dregs); \
  179. } while(0)
  180. #define for_each_dvma(dma) \
  181. for((dma) = dma_chain; (dma); (dma) = (dma)->next)
  182. extern int get_dma_list(char *);
  183. extern int request_dma(unsigned int, __const__ char *);
  184. extern void free_dma(unsigned int);
  185. /* From PCI */
  186. #ifdef CONFIG_PCI
  187. extern int isa_dma_bridge_buggy;
  188. #else
  189. #define isa_dma_bridge_buggy (0)
  190. #endif
  191. #endif /* !(_ASM_SPARC64_DMA_H) */