mpsc.c 52 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115
  1. /*
  2. * Generic driver for the MPSC (UART mode) on Marvell parts (e.g., GT64240,
  3. * GT64260, MV64340, MV64360, GT96100, ... ).
  4. *
  5. * Author: Mark A. Greer <mgreer@mvista.com>
  6. *
  7. * Based on an old MPSC driver that was in the linuxppc tree. It appears to
  8. * have been created by Chris Zankel (formerly of MontaVista) but there
  9. * is no proper Copyright so I'm not sure. Apparently, parts were also
  10. * taken from PPCBoot (now U-Boot). Also based on drivers/serial/8250.c
  11. * by Russell King.
  12. *
  13. * 2004 (c) MontaVista, Software, Inc. This file is licensed under
  14. * the terms of the GNU General Public License version 2. This program
  15. * is licensed "as is" without any warranty of any kind, whether express
  16. * or implied.
  17. */
  18. /*
  19. * The MPSC interface is much like a typical network controller's interface.
  20. * That is, you set up separate rings of descriptors for transmitting and
  21. * receiving data. There is also a pool of buffers with (one buffer per
  22. * descriptor) that incoming data are dma'd into or outgoing data are dma'd
  23. * out of.
  24. *
  25. * The MPSC requires two other controllers to be able to work. The Baud Rate
  26. * Generator (BRG) provides a clock at programmable frequencies which determines
  27. * the baud rate. The Serial DMA Controller (SDMA) takes incoming data from the
  28. * MPSC and DMA's it into memory or DMA's outgoing data and passes it to the
  29. * MPSC. It is actually the SDMA interrupt that the driver uses to keep the
  30. * transmit and receive "engines" going (i.e., indicate data has been
  31. * transmitted or received).
  32. *
  33. * NOTES:
  34. *
  35. * 1) Some chips have an erratum where several regs cannot be
  36. * read. To work around that, we keep a local copy of those regs in
  37. * 'mpsc_port_info'.
  38. *
  39. * 2) Some chips have an erratum where the ctlr will hang when the SDMA ctlr
  40. * accesses system mem with coherency enabled. For that reason, the driver
  41. * assumes that coherency for that ctlr has been disabled. This means
  42. * that when in a cache coherent system, the driver has to manually manage
  43. * the data cache on the areas that it touches because the dma_* macro are
  44. * basically no-ops.
  45. *
  46. * 3) There is an erratum (on PPC) where you can't use the instruction to do
  47. * a DMA_TO_DEVICE/cache clean so DMA_BIDIRECTIONAL/flushes are used in places
  48. * where a DMA_TO_DEVICE/clean would have [otherwise] sufficed.
  49. *
  50. * 4) AFAICT, hardware flow control isn't supported by the controller --MAG.
  51. */
  52. #if defined(CONFIG_SERIAL_MPSC_CONSOLE) && defined(CONFIG_MAGIC_SYSRQ)
  53. #define SUPPORT_SYSRQ
  54. #endif
  55. #include <linux/module.h>
  56. #include <linux/moduleparam.h>
  57. #include <linux/tty.h>
  58. #include <linux/tty_flip.h>
  59. #include <linux/ioport.h>
  60. #include <linux/init.h>
  61. #include <linux/console.h>
  62. #include <linux/sysrq.h>
  63. #include <linux/serial.h>
  64. #include <linux/serial_core.h>
  65. #include <linux/delay.h>
  66. #include <linux/device.h>
  67. #include <linux/dma-mapping.h>
  68. #include <linux/mv643xx.h>
  69. #include <linux/platform_device.h>
  70. #include <asm/io.h>
  71. #include <asm/irq.h>
  72. #if defined(CONFIG_SERIAL_MPSC_CONSOLE) && defined(CONFIG_MAGIC_SYSRQ)
  73. #define SUPPORT_SYSRQ
  74. #endif
  75. #define MPSC_NUM_CTLRS 2
  76. /*
  77. * Descriptors and buffers must be cache line aligned.
  78. * Buffers lengths must be multiple of cache line size.
  79. * Number of Tx & Rx descriptors must be powers of 2.
  80. */
  81. #define MPSC_RXR_ENTRIES 32
  82. #define MPSC_RXRE_SIZE dma_get_cache_alignment()
  83. #define MPSC_RXR_SIZE (MPSC_RXR_ENTRIES * MPSC_RXRE_SIZE)
  84. #define MPSC_RXBE_SIZE dma_get_cache_alignment()
  85. #define MPSC_RXB_SIZE (MPSC_RXR_ENTRIES * MPSC_RXBE_SIZE)
  86. #define MPSC_TXR_ENTRIES 32
  87. #define MPSC_TXRE_SIZE dma_get_cache_alignment()
  88. #define MPSC_TXR_SIZE (MPSC_TXR_ENTRIES * MPSC_TXRE_SIZE)
  89. #define MPSC_TXBE_SIZE dma_get_cache_alignment()
  90. #define MPSC_TXB_SIZE (MPSC_TXR_ENTRIES * MPSC_TXBE_SIZE)
  91. #define MPSC_DMA_ALLOC_SIZE (MPSC_RXR_SIZE + MPSC_RXB_SIZE + \
  92. MPSC_TXR_SIZE + MPSC_TXB_SIZE + \
  93. dma_get_cache_alignment() /* for alignment */)
  94. /* Rx and Tx Ring entry descriptors -- assume entry size is <= cacheline size */
  95. struct mpsc_rx_desc {
  96. u16 bufsize;
  97. u16 bytecnt;
  98. u32 cmdstat;
  99. u32 link;
  100. u32 buf_ptr;
  101. } __attribute((packed));
  102. struct mpsc_tx_desc {
  103. u16 bytecnt;
  104. u16 shadow;
  105. u32 cmdstat;
  106. u32 link;
  107. u32 buf_ptr;
  108. } __attribute((packed));
  109. /*
  110. * Some regs that have the erratum that you can't read them are are shared
  111. * between the two MPSC controllers. This struct contains those shared regs.
  112. */
  113. struct mpsc_shared_regs {
  114. phys_addr_t mpsc_routing_base_p;
  115. phys_addr_t sdma_intr_base_p;
  116. void __iomem *mpsc_routing_base;
  117. void __iomem *sdma_intr_base;
  118. u32 MPSC_MRR_m;
  119. u32 MPSC_RCRR_m;
  120. u32 MPSC_TCRR_m;
  121. u32 SDMA_INTR_CAUSE_m;
  122. u32 SDMA_INTR_MASK_m;
  123. };
  124. /* The main driver data structure */
  125. struct mpsc_port_info {
  126. struct uart_port port; /* Overlay uart_port structure */
  127. /* Internal driver state for this ctlr */
  128. u8 ready;
  129. u8 rcv_data;
  130. tcflag_t c_iflag; /* save termios->c_iflag */
  131. tcflag_t c_cflag; /* save termios->c_cflag */
  132. /* Info passed in from platform */
  133. u8 mirror_regs; /* Need to mirror regs? */
  134. u8 cache_mgmt; /* Need manual cache mgmt? */
  135. u8 brg_can_tune; /* BRG has baud tuning? */
  136. u32 brg_clk_src;
  137. u16 mpsc_max_idle;
  138. int default_baud;
  139. int default_bits;
  140. int default_parity;
  141. int default_flow;
  142. /* Physical addresses of various blocks of registers (from platform) */
  143. phys_addr_t mpsc_base_p;
  144. phys_addr_t sdma_base_p;
  145. phys_addr_t brg_base_p;
  146. /* Virtual addresses of various blocks of registers (from platform) */
  147. void __iomem *mpsc_base;
  148. void __iomem *sdma_base;
  149. void __iomem *brg_base;
  150. /* Descriptor ring and buffer allocations */
  151. void *dma_region;
  152. dma_addr_t dma_region_p;
  153. dma_addr_t rxr; /* Rx descriptor ring */
  154. dma_addr_t rxr_p; /* Phys addr of rxr */
  155. u8 *rxb; /* Rx Ring I/O buf */
  156. u8 *rxb_p; /* Phys addr of rxb */
  157. u32 rxr_posn; /* First desc w/ Rx data */
  158. dma_addr_t txr; /* Tx descriptor ring */
  159. dma_addr_t txr_p; /* Phys addr of txr */
  160. u8 *txb; /* Tx Ring I/O buf */
  161. u8 *txb_p; /* Phys addr of txb */
  162. int txr_head; /* Where new data goes */
  163. int txr_tail; /* Where sent data comes off */
  164. spinlock_t tx_lock; /* transmit lock */
  165. /* Mirrored values of regs we can't read (if 'mirror_regs' set) */
  166. u32 MPSC_MPCR_m;
  167. u32 MPSC_CHR_1_m;
  168. u32 MPSC_CHR_2_m;
  169. u32 MPSC_CHR_10_m;
  170. u32 BRG_BCR_m;
  171. struct mpsc_shared_regs *shared_regs;
  172. };
  173. /* Hooks to platform-specific code */
  174. int mpsc_platform_register_driver(void);
  175. void mpsc_platform_unregister_driver(void);
  176. /* Hooks back in to mpsc common to be called by platform-specific code */
  177. struct mpsc_port_info *mpsc_device_probe(int index);
  178. struct mpsc_port_info *mpsc_device_remove(int index);
  179. /* Main MPSC Configuration Register Offsets */
  180. #define MPSC_MMCRL 0x0000
  181. #define MPSC_MMCRH 0x0004
  182. #define MPSC_MPCR 0x0008
  183. #define MPSC_CHR_1 0x000c
  184. #define MPSC_CHR_2 0x0010
  185. #define MPSC_CHR_3 0x0014
  186. #define MPSC_CHR_4 0x0018
  187. #define MPSC_CHR_5 0x001c
  188. #define MPSC_CHR_6 0x0020
  189. #define MPSC_CHR_7 0x0024
  190. #define MPSC_CHR_8 0x0028
  191. #define MPSC_CHR_9 0x002c
  192. #define MPSC_CHR_10 0x0030
  193. #define MPSC_CHR_11 0x0034
  194. #define MPSC_MPCR_FRZ (1 << 9)
  195. #define MPSC_MPCR_CL_5 0
  196. #define MPSC_MPCR_CL_6 1
  197. #define MPSC_MPCR_CL_7 2
  198. #define MPSC_MPCR_CL_8 3
  199. #define MPSC_MPCR_SBL_1 0
  200. #define MPSC_MPCR_SBL_2 1
  201. #define MPSC_CHR_2_TEV (1<<1)
  202. #define MPSC_CHR_2_TA (1<<7)
  203. #define MPSC_CHR_2_TTCS (1<<9)
  204. #define MPSC_CHR_2_REV (1<<17)
  205. #define MPSC_CHR_2_RA (1<<23)
  206. #define MPSC_CHR_2_CRD (1<<25)
  207. #define MPSC_CHR_2_EH (1<<31)
  208. #define MPSC_CHR_2_PAR_ODD 0
  209. #define MPSC_CHR_2_PAR_SPACE 1
  210. #define MPSC_CHR_2_PAR_EVEN 2
  211. #define MPSC_CHR_2_PAR_MARK 3
  212. /* MPSC Signal Routing */
  213. #define MPSC_MRR 0x0000
  214. #define MPSC_RCRR 0x0004
  215. #define MPSC_TCRR 0x0008
  216. /* Serial DMA Controller Interface Registers */
  217. #define SDMA_SDC 0x0000
  218. #define SDMA_SDCM 0x0008
  219. #define SDMA_RX_DESC 0x0800
  220. #define SDMA_RX_BUF_PTR 0x0808
  221. #define SDMA_SCRDP 0x0810
  222. #define SDMA_TX_DESC 0x0c00
  223. #define SDMA_SCTDP 0x0c10
  224. #define SDMA_SFTDP 0x0c14
  225. #define SDMA_DESC_CMDSTAT_PE (1<<0)
  226. #define SDMA_DESC_CMDSTAT_CDL (1<<1)
  227. #define SDMA_DESC_CMDSTAT_FR (1<<3)
  228. #define SDMA_DESC_CMDSTAT_OR (1<<6)
  229. #define SDMA_DESC_CMDSTAT_BR (1<<9)
  230. #define SDMA_DESC_CMDSTAT_MI (1<<10)
  231. #define SDMA_DESC_CMDSTAT_A (1<<11)
  232. #define SDMA_DESC_CMDSTAT_AM (1<<12)
  233. #define SDMA_DESC_CMDSTAT_CT (1<<13)
  234. #define SDMA_DESC_CMDSTAT_C (1<<14)
  235. #define SDMA_DESC_CMDSTAT_ES (1<<15)
  236. #define SDMA_DESC_CMDSTAT_L (1<<16)
  237. #define SDMA_DESC_CMDSTAT_F (1<<17)
  238. #define SDMA_DESC_CMDSTAT_P (1<<18)
  239. #define SDMA_DESC_CMDSTAT_EI (1<<23)
  240. #define SDMA_DESC_CMDSTAT_O (1<<31)
  241. #define SDMA_DESC_DFLT (SDMA_DESC_CMDSTAT_O | \
  242. SDMA_DESC_CMDSTAT_EI)
  243. #define SDMA_SDC_RFT (1<<0)
  244. #define SDMA_SDC_SFM (1<<1)
  245. #define SDMA_SDC_BLMR (1<<6)
  246. #define SDMA_SDC_BLMT (1<<7)
  247. #define SDMA_SDC_POVR (1<<8)
  248. #define SDMA_SDC_RIFB (1<<9)
  249. #define SDMA_SDCM_ERD (1<<7)
  250. #define SDMA_SDCM_AR (1<<15)
  251. #define SDMA_SDCM_STD (1<<16)
  252. #define SDMA_SDCM_TXD (1<<23)
  253. #define SDMA_SDCM_AT (1<<31)
  254. #define SDMA_0_CAUSE_RXBUF (1<<0)
  255. #define SDMA_0_CAUSE_RXERR (1<<1)
  256. #define SDMA_0_CAUSE_TXBUF (1<<2)
  257. #define SDMA_0_CAUSE_TXEND (1<<3)
  258. #define SDMA_1_CAUSE_RXBUF (1<<8)
  259. #define SDMA_1_CAUSE_RXERR (1<<9)
  260. #define SDMA_1_CAUSE_TXBUF (1<<10)
  261. #define SDMA_1_CAUSE_TXEND (1<<11)
  262. #define SDMA_CAUSE_RX_MASK (SDMA_0_CAUSE_RXBUF | SDMA_0_CAUSE_RXERR | \
  263. SDMA_1_CAUSE_RXBUF | SDMA_1_CAUSE_RXERR)
  264. #define SDMA_CAUSE_TX_MASK (SDMA_0_CAUSE_TXBUF | SDMA_0_CAUSE_TXEND | \
  265. SDMA_1_CAUSE_TXBUF | SDMA_1_CAUSE_TXEND)
  266. /* SDMA Interrupt registers */
  267. #define SDMA_INTR_CAUSE 0x0000
  268. #define SDMA_INTR_MASK 0x0080
  269. /* Baud Rate Generator Interface Registers */
  270. #define BRG_BCR 0x0000
  271. #define BRG_BTR 0x0004
  272. /*
  273. * Define how this driver is known to the outside (we've been assigned a
  274. * range on the "Low-density serial ports" major).
  275. */
  276. #define MPSC_MAJOR 204
  277. #define MPSC_MINOR_START 44
  278. #define MPSC_DRIVER_NAME "MPSC"
  279. #define MPSC_DEV_NAME "ttyMM"
  280. #define MPSC_VERSION "1.00"
  281. static struct mpsc_port_info mpsc_ports[MPSC_NUM_CTLRS];
  282. static struct mpsc_shared_regs mpsc_shared_regs;
  283. static struct uart_driver mpsc_reg;
  284. static void mpsc_start_rx(struct mpsc_port_info *pi);
  285. static void mpsc_free_ring_mem(struct mpsc_port_info *pi);
  286. static void mpsc_release_port(struct uart_port *port);
  287. /*
  288. ******************************************************************************
  289. *
  290. * Baud Rate Generator Routines (BRG)
  291. *
  292. ******************************************************************************
  293. */
  294. static void
  295. mpsc_brg_init(struct mpsc_port_info *pi, u32 clk_src)
  296. {
  297. u32 v;
  298. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  299. v = (v & ~(0xf << 18)) | ((clk_src & 0xf) << 18);
  300. if (pi->brg_can_tune)
  301. v &= ~(1 << 25);
  302. if (pi->mirror_regs)
  303. pi->BRG_BCR_m = v;
  304. writel(v, pi->brg_base + BRG_BCR);
  305. writel(readl(pi->brg_base + BRG_BTR) & 0xffff0000,
  306. pi->brg_base + BRG_BTR);
  307. return;
  308. }
  309. static void
  310. mpsc_brg_enable(struct mpsc_port_info *pi)
  311. {
  312. u32 v;
  313. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  314. v |= (1 << 16);
  315. if (pi->mirror_regs)
  316. pi->BRG_BCR_m = v;
  317. writel(v, pi->brg_base + BRG_BCR);
  318. return;
  319. }
  320. static void
  321. mpsc_brg_disable(struct mpsc_port_info *pi)
  322. {
  323. u32 v;
  324. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  325. v &= ~(1 << 16);
  326. if (pi->mirror_regs)
  327. pi->BRG_BCR_m = v;
  328. writel(v, pi->brg_base + BRG_BCR);
  329. return;
  330. }
  331. static inline void
  332. mpsc_set_baudrate(struct mpsc_port_info *pi, u32 baud)
  333. {
  334. /*
  335. * To set the baud, we adjust the CDV field in the BRG_BCR reg.
  336. * From manual: Baud = clk / ((CDV+1)*2) ==> CDV = (clk / (baud*2)) - 1.
  337. * However, the input clock is divided by 16 in the MPSC b/c of how
  338. * 'MPSC_MMCRH' was set up so we have to divide the 'clk' used in our
  339. * calculation by 16 to account for that. So the real calculation
  340. * that accounts for the way the mpsc is set up is:
  341. * CDV = (clk / (baud*2*16)) - 1 ==> CDV = (clk / (baud << 5)) - 1.
  342. */
  343. u32 cdv = (pi->port.uartclk / (baud << 5)) - 1;
  344. u32 v;
  345. mpsc_brg_disable(pi);
  346. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  347. v = (v & 0xffff0000) | (cdv & 0xffff);
  348. if (pi->mirror_regs)
  349. pi->BRG_BCR_m = v;
  350. writel(v, pi->brg_base + BRG_BCR);
  351. mpsc_brg_enable(pi);
  352. return;
  353. }
  354. /*
  355. ******************************************************************************
  356. *
  357. * Serial DMA Routines (SDMA)
  358. *
  359. ******************************************************************************
  360. */
  361. static void
  362. mpsc_sdma_burstsize(struct mpsc_port_info *pi, u32 burst_size)
  363. {
  364. u32 v;
  365. pr_debug("mpsc_sdma_burstsize[%d]: burst_size: %d\n",
  366. pi->port.line, burst_size);
  367. burst_size >>= 3; /* Divide by 8 b/c reg values are 8-byte chunks */
  368. if (burst_size < 2)
  369. v = 0x0; /* 1 64-bit word */
  370. else if (burst_size < 4)
  371. v = 0x1; /* 2 64-bit words */
  372. else if (burst_size < 8)
  373. v = 0x2; /* 4 64-bit words */
  374. else
  375. v = 0x3; /* 8 64-bit words */
  376. writel((readl(pi->sdma_base + SDMA_SDC) & (0x3 << 12)) | (v << 12),
  377. pi->sdma_base + SDMA_SDC);
  378. return;
  379. }
  380. static void
  381. mpsc_sdma_init(struct mpsc_port_info *pi, u32 burst_size)
  382. {
  383. pr_debug("mpsc_sdma_init[%d]: burst_size: %d\n", pi->port.line,
  384. burst_size);
  385. writel((readl(pi->sdma_base + SDMA_SDC) & 0x3ff) | 0x03f,
  386. pi->sdma_base + SDMA_SDC);
  387. mpsc_sdma_burstsize(pi, burst_size);
  388. return;
  389. }
  390. static inline u32
  391. mpsc_sdma_intr_mask(struct mpsc_port_info *pi, u32 mask)
  392. {
  393. u32 old, v;
  394. pr_debug("mpsc_sdma_intr_mask[%d]: mask: 0x%x\n", pi->port.line, mask);
  395. old = v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  396. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  397. mask &= 0xf;
  398. if (pi->port.line)
  399. mask <<= 8;
  400. v &= ~mask;
  401. if (pi->mirror_regs)
  402. pi->shared_regs->SDMA_INTR_MASK_m = v;
  403. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  404. if (pi->port.line)
  405. old >>= 8;
  406. return old & 0xf;
  407. }
  408. static inline void
  409. mpsc_sdma_intr_unmask(struct mpsc_port_info *pi, u32 mask)
  410. {
  411. u32 v;
  412. pr_debug("mpsc_sdma_intr_unmask[%d]: mask: 0x%x\n", pi->port.line,mask);
  413. v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  414. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  415. mask &= 0xf;
  416. if (pi->port.line)
  417. mask <<= 8;
  418. v |= mask;
  419. if (pi->mirror_regs)
  420. pi->shared_regs->SDMA_INTR_MASK_m = v;
  421. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  422. return;
  423. }
  424. static inline void
  425. mpsc_sdma_intr_ack(struct mpsc_port_info *pi)
  426. {
  427. pr_debug("mpsc_sdma_intr_ack[%d]: Acknowledging IRQ\n", pi->port.line);
  428. if (pi->mirror_regs)
  429. pi->shared_regs->SDMA_INTR_CAUSE_m = 0;
  430. writeb(0x00, pi->shared_regs->sdma_intr_base + SDMA_INTR_CAUSE +
  431. pi->port.line);
  432. return;
  433. }
  434. static inline void
  435. mpsc_sdma_set_rx_ring(struct mpsc_port_info *pi, struct mpsc_rx_desc *rxre_p)
  436. {
  437. pr_debug("mpsc_sdma_set_rx_ring[%d]: rxre_p: 0x%x\n",
  438. pi->port.line, (u32) rxre_p);
  439. writel((u32)rxre_p, pi->sdma_base + SDMA_SCRDP);
  440. return;
  441. }
  442. static inline void
  443. mpsc_sdma_set_tx_ring(struct mpsc_port_info *pi, struct mpsc_tx_desc *txre_p)
  444. {
  445. writel((u32)txre_p, pi->sdma_base + SDMA_SFTDP);
  446. writel((u32)txre_p, pi->sdma_base + SDMA_SCTDP);
  447. return;
  448. }
  449. static inline void
  450. mpsc_sdma_cmd(struct mpsc_port_info *pi, u32 val)
  451. {
  452. u32 v;
  453. v = readl(pi->sdma_base + SDMA_SDCM);
  454. if (val)
  455. v |= val;
  456. else
  457. v = 0;
  458. wmb();
  459. writel(v, pi->sdma_base + SDMA_SDCM);
  460. wmb();
  461. return;
  462. }
  463. static inline uint
  464. mpsc_sdma_tx_active(struct mpsc_port_info *pi)
  465. {
  466. return readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_TXD;
  467. }
  468. static inline void
  469. mpsc_sdma_start_tx(struct mpsc_port_info *pi)
  470. {
  471. struct mpsc_tx_desc *txre, *txre_p;
  472. /* If tx isn't running & there's a desc ready to go, start it */
  473. if (!mpsc_sdma_tx_active(pi)) {
  474. txre = (struct mpsc_tx_desc *)(pi->txr +
  475. (pi->txr_tail * MPSC_TXRE_SIZE));
  476. dma_cache_sync(pi->port.dev, (void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  477. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  478. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  479. invalidate_dcache_range((ulong)txre,
  480. (ulong)txre + MPSC_TXRE_SIZE);
  481. #endif
  482. if (be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O) {
  483. txre_p = (struct mpsc_tx_desc *)(pi->txr_p +
  484. (pi->txr_tail *
  485. MPSC_TXRE_SIZE));
  486. mpsc_sdma_set_tx_ring(pi, txre_p);
  487. mpsc_sdma_cmd(pi, SDMA_SDCM_STD | SDMA_SDCM_TXD);
  488. }
  489. }
  490. return;
  491. }
  492. static inline void
  493. mpsc_sdma_stop(struct mpsc_port_info *pi)
  494. {
  495. pr_debug("mpsc_sdma_stop[%d]: Stopping SDMA\n", pi->port.line);
  496. /* Abort any SDMA transfers */
  497. mpsc_sdma_cmd(pi, 0);
  498. mpsc_sdma_cmd(pi, SDMA_SDCM_AR | SDMA_SDCM_AT);
  499. /* Clear the SDMA current and first TX and RX pointers */
  500. mpsc_sdma_set_tx_ring(pi, NULL);
  501. mpsc_sdma_set_rx_ring(pi, NULL);
  502. /* Disable interrupts */
  503. mpsc_sdma_intr_mask(pi, 0xf);
  504. mpsc_sdma_intr_ack(pi);
  505. return;
  506. }
  507. /*
  508. ******************************************************************************
  509. *
  510. * Multi-Protocol Serial Controller Routines (MPSC)
  511. *
  512. ******************************************************************************
  513. */
  514. static void
  515. mpsc_hw_init(struct mpsc_port_info *pi)
  516. {
  517. u32 v;
  518. pr_debug("mpsc_hw_init[%d]: Initializing hardware\n", pi->port.line);
  519. /* Set up clock routing */
  520. if (pi->mirror_regs) {
  521. v = pi->shared_regs->MPSC_MRR_m;
  522. v &= ~0x1c7;
  523. pi->shared_regs->MPSC_MRR_m = v;
  524. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  525. v = pi->shared_regs->MPSC_RCRR_m;
  526. v = (v & ~0xf0f) | 0x100;
  527. pi->shared_regs->MPSC_RCRR_m = v;
  528. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  529. v = pi->shared_regs->MPSC_TCRR_m;
  530. v = (v & ~0xf0f) | 0x100;
  531. pi->shared_regs->MPSC_TCRR_m = v;
  532. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  533. }
  534. else {
  535. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  536. v &= ~0x1c7;
  537. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  538. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  539. v = (v & ~0xf0f) | 0x100;
  540. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  541. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  542. v = (v & ~0xf0f) | 0x100;
  543. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  544. }
  545. /* Put MPSC in UART mode & enabel Tx/Rx egines */
  546. writel(0x000004c4, pi->mpsc_base + MPSC_MMCRL);
  547. /* No preamble, 16x divider, low-latency, */
  548. writel(0x04400400, pi->mpsc_base + MPSC_MMCRH);
  549. if (pi->mirror_regs) {
  550. pi->MPSC_CHR_1_m = 0;
  551. pi->MPSC_CHR_2_m = 0;
  552. }
  553. writel(0, pi->mpsc_base + MPSC_CHR_1);
  554. writel(0, pi->mpsc_base + MPSC_CHR_2);
  555. writel(pi->mpsc_max_idle, pi->mpsc_base + MPSC_CHR_3);
  556. writel(0, pi->mpsc_base + MPSC_CHR_4);
  557. writel(0, pi->mpsc_base + MPSC_CHR_5);
  558. writel(0, pi->mpsc_base + MPSC_CHR_6);
  559. writel(0, pi->mpsc_base + MPSC_CHR_7);
  560. writel(0, pi->mpsc_base + MPSC_CHR_8);
  561. writel(0, pi->mpsc_base + MPSC_CHR_9);
  562. writel(0, pi->mpsc_base + MPSC_CHR_10);
  563. return;
  564. }
  565. static inline void
  566. mpsc_enter_hunt(struct mpsc_port_info *pi)
  567. {
  568. pr_debug("mpsc_enter_hunt[%d]: Hunting...\n", pi->port.line);
  569. if (pi->mirror_regs) {
  570. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_EH,
  571. pi->mpsc_base + MPSC_CHR_2);
  572. /* Erratum prevents reading CHR_2 so just delay for a while */
  573. udelay(100);
  574. }
  575. else {
  576. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_EH,
  577. pi->mpsc_base + MPSC_CHR_2);
  578. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_EH)
  579. udelay(10);
  580. }
  581. return;
  582. }
  583. static inline void
  584. mpsc_freeze(struct mpsc_port_info *pi)
  585. {
  586. u32 v;
  587. pr_debug("mpsc_freeze[%d]: Freezing\n", pi->port.line);
  588. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  589. readl(pi->mpsc_base + MPSC_MPCR);
  590. v |= MPSC_MPCR_FRZ;
  591. if (pi->mirror_regs)
  592. pi->MPSC_MPCR_m = v;
  593. writel(v, pi->mpsc_base + MPSC_MPCR);
  594. return;
  595. }
  596. static inline void
  597. mpsc_unfreeze(struct mpsc_port_info *pi)
  598. {
  599. u32 v;
  600. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  601. readl(pi->mpsc_base + MPSC_MPCR);
  602. v &= ~MPSC_MPCR_FRZ;
  603. if (pi->mirror_regs)
  604. pi->MPSC_MPCR_m = v;
  605. writel(v, pi->mpsc_base + MPSC_MPCR);
  606. pr_debug("mpsc_unfreeze[%d]: Unfrozen\n", pi->port.line);
  607. return;
  608. }
  609. static inline void
  610. mpsc_set_char_length(struct mpsc_port_info *pi, u32 len)
  611. {
  612. u32 v;
  613. pr_debug("mpsc_set_char_length[%d]: char len: %d\n", pi->port.line,len);
  614. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  615. readl(pi->mpsc_base + MPSC_MPCR);
  616. v = (v & ~(0x3 << 12)) | ((len & 0x3) << 12);
  617. if (pi->mirror_regs)
  618. pi->MPSC_MPCR_m = v;
  619. writel(v, pi->mpsc_base + MPSC_MPCR);
  620. return;
  621. }
  622. static inline void
  623. mpsc_set_stop_bit_length(struct mpsc_port_info *pi, u32 len)
  624. {
  625. u32 v;
  626. pr_debug("mpsc_set_stop_bit_length[%d]: stop bits: %d\n",
  627. pi->port.line, len);
  628. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  629. readl(pi->mpsc_base + MPSC_MPCR);
  630. v = (v & ~(1 << 14)) | ((len & 0x1) << 14);
  631. if (pi->mirror_regs)
  632. pi->MPSC_MPCR_m = v;
  633. writel(v, pi->mpsc_base + MPSC_MPCR);
  634. return;
  635. }
  636. static inline void
  637. mpsc_set_parity(struct mpsc_port_info *pi, u32 p)
  638. {
  639. u32 v;
  640. pr_debug("mpsc_set_parity[%d]: parity bits: 0x%x\n", pi->port.line, p);
  641. v = (pi->mirror_regs) ? pi->MPSC_CHR_2_m :
  642. readl(pi->mpsc_base + MPSC_CHR_2);
  643. p &= 0x3;
  644. v = (v & ~0xc000c) | (p << 18) | (p << 2);
  645. if (pi->mirror_regs)
  646. pi->MPSC_CHR_2_m = v;
  647. writel(v, pi->mpsc_base + MPSC_CHR_2);
  648. return;
  649. }
  650. /*
  651. ******************************************************************************
  652. *
  653. * Driver Init Routines
  654. *
  655. ******************************************************************************
  656. */
  657. static void
  658. mpsc_init_hw(struct mpsc_port_info *pi)
  659. {
  660. pr_debug("mpsc_init_hw[%d]: Initializing\n", pi->port.line);
  661. mpsc_brg_init(pi, pi->brg_clk_src);
  662. mpsc_brg_enable(pi);
  663. mpsc_sdma_init(pi, dma_get_cache_alignment()); /* burst a cacheline */
  664. mpsc_sdma_stop(pi);
  665. mpsc_hw_init(pi);
  666. return;
  667. }
  668. static int
  669. mpsc_alloc_ring_mem(struct mpsc_port_info *pi)
  670. {
  671. int rc = 0;
  672. pr_debug("mpsc_alloc_ring_mem[%d]: Allocating ring mem\n",
  673. pi->port.line);
  674. if (!pi->dma_region) {
  675. if (!dma_supported(pi->port.dev, 0xffffffff)) {
  676. printk(KERN_ERR "MPSC: Inadequate DMA support\n");
  677. rc = -ENXIO;
  678. }
  679. else if ((pi->dma_region = dma_alloc_noncoherent(pi->port.dev,
  680. MPSC_DMA_ALLOC_SIZE, &pi->dma_region_p, GFP_KERNEL))
  681. == NULL) {
  682. printk(KERN_ERR "MPSC: Can't alloc Desc region\n");
  683. rc = -ENOMEM;
  684. }
  685. }
  686. return rc;
  687. }
  688. static void
  689. mpsc_free_ring_mem(struct mpsc_port_info *pi)
  690. {
  691. pr_debug("mpsc_free_ring_mem[%d]: Freeing ring mem\n", pi->port.line);
  692. if (pi->dma_region) {
  693. dma_free_noncoherent(pi->port.dev, MPSC_DMA_ALLOC_SIZE,
  694. pi->dma_region, pi->dma_region_p);
  695. pi->dma_region = NULL;
  696. pi->dma_region_p = (dma_addr_t) NULL;
  697. }
  698. return;
  699. }
  700. static void
  701. mpsc_init_rings(struct mpsc_port_info *pi)
  702. {
  703. struct mpsc_rx_desc *rxre;
  704. struct mpsc_tx_desc *txre;
  705. dma_addr_t dp, dp_p;
  706. u8 *bp, *bp_p;
  707. int i;
  708. pr_debug("mpsc_init_rings[%d]: Initializing rings\n", pi->port.line);
  709. BUG_ON(pi->dma_region == NULL);
  710. memset(pi->dma_region, 0, MPSC_DMA_ALLOC_SIZE);
  711. /*
  712. * Descriptors & buffers are multiples of cacheline size and must be
  713. * cacheline aligned.
  714. */
  715. dp = ALIGN((u32) pi->dma_region, dma_get_cache_alignment());
  716. dp_p = ALIGN((u32) pi->dma_region_p, dma_get_cache_alignment());
  717. /*
  718. * Partition dma region into rx ring descriptor, rx buffers,
  719. * tx ring descriptors, and tx buffers.
  720. */
  721. pi->rxr = dp;
  722. pi->rxr_p = dp_p;
  723. dp += MPSC_RXR_SIZE;
  724. dp_p += MPSC_RXR_SIZE;
  725. pi->rxb = (u8 *) dp;
  726. pi->rxb_p = (u8 *) dp_p;
  727. dp += MPSC_RXB_SIZE;
  728. dp_p += MPSC_RXB_SIZE;
  729. pi->rxr_posn = 0;
  730. pi->txr = dp;
  731. pi->txr_p = dp_p;
  732. dp += MPSC_TXR_SIZE;
  733. dp_p += MPSC_TXR_SIZE;
  734. pi->txb = (u8 *) dp;
  735. pi->txb_p = (u8 *) dp_p;
  736. pi->txr_head = 0;
  737. pi->txr_tail = 0;
  738. /* Init rx ring descriptors */
  739. dp = pi->rxr;
  740. dp_p = pi->rxr_p;
  741. bp = pi->rxb;
  742. bp_p = pi->rxb_p;
  743. for (i = 0; i < MPSC_RXR_ENTRIES; i++) {
  744. rxre = (struct mpsc_rx_desc *)dp;
  745. rxre->bufsize = cpu_to_be16(MPSC_RXBE_SIZE);
  746. rxre->bytecnt = cpu_to_be16(0);
  747. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  748. SDMA_DESC_CMDSTAT_EI |
  749. SDMA_DESC_CMDSTAT_F |
  750. SDMA_DESC_CMDSTAT_L);
  751. rxre->link = cpu_to_be32(dp_p + MPSC_RXRE_SIZE);
  752. rxre->buf_ptr = cpu_to_be32(bp_p);
  753. dp += MPSC_RXRE_SIZE;
  754. dp_p += MPSC_RXRE_SIZE;
  755. bp += MPSC_RXBE_SIZE;
  756. bp_p += MPSC_RXBE_SIZE;
  757. }
  758. rxre->link = cpu_to_be32(pi->rxr_p); /* Wrap last back to first */
  759. /* Init tx ring descriptors */
  760. dp = pi->txr;
  761. dp_p = pi->txr_p;
  762. bp = pi->txb;
  763. bp_p = pi->txb_p;
  764. for (i = 0; i < MPSC_TXR_ENTRIES; i++) {
  765. txre = (struct mpsc_tx_desc *)dp;
  766. txre->link = cpu_to_be32(dp_p + MPSC_TXRE_SIZE);
  767. txre->buf_ptr = cpu_to_be32(bp_p);
  768. dp += MPSC_TXRE_SIZE;
  769. dp_p += MPSC_TXRE_SIZE;
  770. bp += MPSC_TXBE_SIZE;
  771. bp_p += MPSC_TXBE_SIZE;
  772. }
  773. txre->link = cpu_to_be32(pi->txr_p); /* Wrap last back to first */
  774. dma_cache_sync(pi->port.dev, (void *) pi->dma_region, MPSC_DMA_ALLOC_SIZE,
  775. DMA_BIDIRECTIONAL);
  776. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  777. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  778. flush_dcache_range((ulong)pi->dma_region,
  779. (ulong)pi->dma_region + MPSC_DMA_ALLOC_SIZE);
  780. #endif
  781. return;
  782. }
  783. static void
  784. mpsc_uninit_rings(struct mpsc_port_info *pi)
  785. {
  786. pr_debug("mpsc_uninit_rings[%d]: Uninitializing rings\n",pi->port.line);
  787. BUG_ON(pi->dma_region == NULL);
  788. pi->rxr = 0;
  789. pi->rxr_p = 0;
  790. pi->rxb = NULL;
  791. pi->rxb_p = NULL;
  792. pi->rxr_posn = 0;
  793. pi->txr = 0;
  794. pi->txr_p = 0;
  795. pi->txb = NULL;
  796. pi->txb_p = NULL;
  797. pi->txr_head = 0;
  798. pi->txr_tail = 0;
  799. return;
  800. }
  801. static int
  802. mpsc_make_ready(struct mpsc_port_info *pi)
  803. {
  804. int rc;
  805. pr_debug("mpsc_make_ready[%d]: Making cltr ready\n", pi->port.line);
  806. if (!pi->ready) {
  807. mpsc_init_hw(pi);
  808. if ((rc = mpsc_alloc_ring_mem(pi)))
  809. return rc;
  810. mpsc_init_rings(pi);
  811. pi->ready = 1;
  812. }
  813. return 0;
  814. }
  815. /*
  816. ******************************************************************************
  817. *
  818. * Interrupt Handling Routines
  819. *
  820. ******************************************************************************
  821. */
  822. static inline int
  823. mpsc_rx_intr(struct mpsc_port_info *pi)
  824. {
  825. struct mpsc_rx_desc *rxre;
  826. struct tty_struct *tty = pi->port.info->tty;
  827. u32 cmdstat, bytes_in, i;
  828. int rc = 0;
  829. u8 *bp;
  830. char flag = TTY_NORMAL;
  831. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  832. rxre = (struct mpsc_rx_desc *)(pi->rxr + (pi->rxr_posn*MPSC_RXRE_SIZE));
  833. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  834. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  835. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  836. invalidate_dcache_range((ulong)rxre,
  837. (ulong)rxre + MPSC_RXRE_SIZE);
  838. #endif
  839. /*
  840. * Loop through Rx descriptors handling ones that have been completed.
  841. */
  842. while (!((cmdstat = be32_to_cpu(rxre->cmdstat)) & SDMA_DESC_CMDSTAT_O)){
  843. bytes_in = be16_to_cpu(rxre->bytecnt);
  844. /* Following use of tty struct directly is deprecated */
  845. if (unlikely(tty_buffer_request_room(tty, bytes_in) < bytes_in)) {
  846. if (tty->low_latency)
  847. tty_flip_buffer_push(tty);
  848. /*
  849. * If this failed then we will throw away the bytes
  850. * but must do so to clear interrupts.
  851. */
  852. }
  853. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  854. dma_cache_sync(pi->port.dev, (void *) bp, MPSC_RXBE_SIZE, DMA_FROM_DEVICE);
  855. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  856. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  857. invalidate_dcache_range((ulong)bp,
  858. (ulong)bp + MPSC_RXBE_SIZE);
  859. #endif
  860. /*
  861. * Other than for parity error, the manual provides little
  862. * info on what data will be in a frame flagged by any of
  863. * these errors. For parity error, it is the last byte in
  864. * the buffer that had the error. As for the rest, I guess
  865. * we'll assume there is no data in the buffer.
  866. * If there is...it gets lost.
  867. */
  868. if (unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  869. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) {
  870. pi->port.icount.rx++;
  871. if (cmdstat & SDMA_DESC_CMDSTAT_BR) { /* Break */
  872. pi->port.icount.brk++;
  873. if (uart_handle_break(&pi->port))
  874. goto next_frame;
  875. }
  876. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)/* Framing */
  877. pi->port.icount.frame++;
  878. else if (cmdstat & SDMA_DESC_CMDSTAT_OR) /* Overrun */
  879. pi->port.icount.overrun++;
  880. cmdstat &= pi->port.read_status_mask;
  881. if (cmdstat & SDMA_DESC_CMDSTAT_BR)
  882. flag = TTY_BREAK;
  883. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)
  884. flag = TTY_FRAME;
  885. else if (cmdstat & SDMA_DESC_CMDSTAT_OR)
  886. flag = TTY_OVERRUN;
  887. else if (cmdstat & SDMA_DESC_CMDSTAT_PE)
  888. flag = TTY_PARITY;
  889. }
  890. if (uart_handle_sysrq_char(&pi->port, *bp)) {
  891. bp++;
  892. bytes_in--;
  893. goto next_frame;
  894. }
  895. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  896. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) &&
  897. !(cmdstat & pi->port.ignore_status_mask))
  898. tty_insert_flip_char(tty, *bp, flag);
  899. else {
  900. for (i=0; i<bytes_in; i++)
  901. tty_insert_flip_char(tty, *bp++, TTY_NORMAL);
  902. pi->port.icount.rx += bytes_in;
  903. }
  904. next_frame:
  905. rxre->bytecnt = cpu_to_be16(0);
  906. wmb();
  907. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  908. SDMA_DESC_CMDSTAT_EI |
  909. SDMA_DESC_CMDSTAT_F |
  910. SDMA_DESC_CMDSTAT_L);
  911. wmb();
  912. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE, DMA_BIDIRECTIONAL);
  913. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  914. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  915. flush_dcache_range((ulong)rxre,
  916. (ulong)rxre + MPSC_RXRE_SIZE);
  917. #endif
  918. /* Advance to next descriptor */
  919. pi->rxr_posn = (pi->rxr_posn + 1) & (MPSC_RXR_ENTRIES - 1);
  920. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  921. (pi->rxr_posn * MPSC_RXRE_SIZE));
  922. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  923. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  924. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  925. invalidate_dcache_range((ulong)rxre,
  926. (ulong)rxre + MPSC_RXRE_SIZE);
  927. #endif
  928. rc = 1;
  929. }
  930. /* Restart rx engine, if its stopped */
  931. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  932. mpsc_start_rx(pi);
  933. tty_flip_buffer_push(tty);
  934. return rc;
  935. }
  936. static inline void
  937. mpsc_setup_tx_desc(struct mpsc_port_info *pi, u32 count, u32 intr)
  938. {
  939. struct mpsc_tx_desc *txre;
  940. txre = (struct mpsc_tx_desc *)(pi->txr +
  941. (pi->txr_head * MPSC_TXRE_SIZE));
  942. txre->bytecnt = cpu_to_be16(count);
  943. txre->shadow = txre->bytecnt;
  944. wmb(); /* ensure cmdstat is last field updated */
  945. txre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O | SDMA_DESC_CMDSTAT_F |
  946. SDMA_DESC_CMDSTAT_L | ((intr) ?
  947. SDMA_DESC_CMDSTAT_EI
  948. : 0));
  949. wmb();
  950. dma_cache_sync(pi->port.dev, (void *) txre, MPSC_TXRE_SIZE, DMA_BIDIRECTIONAL);
  951. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  952. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  953. flush_dcache_range((ulong)txre,
  954. (ulong)txre + MPSC_TXRE_SIZE);
  955. #endif
  956. return;
  957. }
  958. static inline void
  959. mpsc_copy_tx_data(struct mpsc_port_info *pi)
  960. {
  961. struct circ_buf *xmit = &pi->port.info->xmit;
  962. u8 *bp;
  963. u32 i;
  964. /* Make sure the desc ring isn't full */
  965. while (CIRC_CNT(pi->txr_head, pi->txr_tail, MPSC_TXR_ENTRIES) <
  966. (MPSC_TXR_ENTRIES - 1)) {
  967. if (pi->port.x_char) {
  968. /*
  969. * Ideally, we should use the TCS field in
  970. * CHR_1 to put the x_char out immediately but
  971. * errata prevents us from being able to read
  972. * CHR_2 to know that its safe to write to
  973. * CHR_1. Instead, just put it in-band with
  974. * all the other Tx data.
  975. */
  976. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  977. *bp = pi->port.x_char;
  978. pi->port.x_char = 0;
  979. i = 1;
  980. }
  981. else if (!uart_circ_empty(xmit) && !uart_tx_stopped(&pi->port)){
  982. i = min((u32) MPSC_TXBE_SIZE,
  983. (u32) uart_circ_chars_pending(xmit));
  984. i = min(i, (u32) CIRC_CNT_TO_END(xmit->head, xmit->tail,
  985. UART_XMIT_SIZE));
  986. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  987. memcpy(bp, &xmit->buf[xmit->tail], i);
  988. xmit->tail = (xmit->tail + i) & (UART_XMIT_SIZE - 1);
  989. if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
  990. uart_write_wakeup(&pi->port);
  991. }
  992. else /* All tx data copied into ring bufs */
  993. return;
  994. dma_cache_sync(pi->port.dev, (void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  995. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  996. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  997. flush_dcache_range((ulong)bp,
  998. (ulong)bp + MPSC_TXBE_SIZE);
  999. #endif
  1000. mpsc_setup_tx_desc(pi, i, 1);
  1001. /* Advance to next descriptor */
  1002. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1003. }
  1004. return;
  1005. }
  1006. static inline int
  1007. mpsc_tx_intr(struct mpsc_port_info *pi)
  1008. {
  1009. struct mpsc_tx_desc *txre;
  1010. int rc = 0;
  1011. unsigned long iflags;
  1012. spin_lock_irqsave(&pi->tx_lock, iflags);
  1013. if (!mpsc_sdma_tx_active(pi)) {
  1014. txre = (struct mpsc_tx_desc *)(pi->txr +
  1015. (pi->txr_tail * MPSC_TXRE_SIZE));
  1016. dma_cache_sync(pi->port.dev, (void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  1017. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1018. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1019. invalidate_dcache_range((ulong)txre,
  1020. (ulong)txre + MPSC_TXRE_SIZE);
  1021. #endif
  1022. while (!(be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O)) {
  1023. rc = 1;
  1024. pi->port.icount.tx += be16_to_cpu(txre->bytecnt);
  1025. pi->txr_tail = (pi->txr_tail+1) & (MPSC_TXR_ENTRIES-1);
  1026. /* If no more data to tx, fall out of loop */
  1027. if (pi->txr_head == pi->txr_tail)
  1028. break;
  1029. txre = (struct mpsc_tx_desc *)(pi->txr +
  1030. (pi->txr_tail * MPSC_TXRE_SIZE));
  1031. dma_cache_sync(pi->port.dev, (void *) txre, MPSC_TXRE_SIZE,
  1032. DMA_FROM_DEVICE);
  1033. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1034. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1035. invalidate_dcache_range((ulong)txre,
  1036. (ulong)txre + MPSC_TXRE_SIZE);
  1037. #endif
  1038. }
  1039. mpsc_copy_tx_data(pi);
  1040. mpsc_sdma_start_tx(pi); /* start next desc if ready */
  1041. }
  1042. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1043. return rc;
  1044. }
  1045. /*
  1046. * This is the driver's interrupt handler. To avoid a race, we first clear
  1047. * the interrupt, then handle any completed Rx/Tx descriptors. When done
  1048. * handling those descriptors, we restart the Rx/Tx engines if they're stopped.
  1049. */
  1050. static irqreturn_t
  1051. mpsc_sdma_intr(int irq, void *dev_id)
  1052. {
  1053. struct mpsc_port_info *pi = dev_id;
  1054. ulong iflags;
  1055. int rc = IRQ_NONE;
  1056. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Received\n",pi->port.line);
  1057. spin_lock_irqsave(&pi->port.lock, iflags);
  1058. mpsc_sdma_intr_ack(pi);
  1059. if (mpsc_rx_intr(pi))
  1060. rc = IRQ_HANDLED;
  1061. if (mpsc_tx_intr(pi))
  1062. rc = IRQ_HANDLED;
  1063. spin_unlock_irqrestore(&pi->port.lock, iflags);
  1064. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Handled\n", pi->port.line);
  1065. return rc;
  1066. }
  1067. /*
  1068. ******************************************************************************
  1069. *
  1070. * serial_core.c Interface routines
  1071. *
  1072. ******************************************************************************
  1073. */
  1074. static uint
  1075. mpsc_tx_empty(struct uart_port *port)
  1076. {
  1077. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1078. ulong iflags;
  1079. uint rc;
  1080. spin_lock_irqsave(&pi->port.lock, iflags);
  1081. rc = mpsc_sdma_tx_active(pi) ? 0 : TIOCSER_TEMT;
  1082. spin_unlock_irqrestore(&pi->port.lock, iflags);
  1083. return rc;
  1084. }
  1085. static void
  1086. mpsc_set_mctrl(struct uart_port *port, uint mctrl)
  1087. {
  1088. /* Have no way to set modem control lines AFAICT */
  1089. return;
  1090. }
  1091. static uint
  1092. mpsc_get_mctrl(struct uart_port *port)
  1093. {
  1094. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1095. u32 mflags, status;
  1096. status = (pi->mirror_regs) ? pi->MPSC_CHR_10_m :
  1097. readl(pi->mpsc_base + MPSC_CHR_10);
  1098. mflags = 0;
  1099. if (status & 0x1)
  1100. mflags |= TIOCM_CTS;
  1101. if (status & 0x2)
  1102. mflags |= TIOCM_CAR;
  1103. return mflags | TIOCM_DSR; /* No way to tell if DSR asserted */
  1104. }
  1105. static void
  1106. mpsc_stop_tx(struct uart_port *port)
  1107. {
  1108. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1109. pr_debug("mpsc_stop_tx[%d]\n", port->line);
  1110. mpsc_freeze(pi);
  1111. return;
  1112. }
  1113. static void
  1114. mpsc_start_tx(struct uart_port *port)
  1115. {
  1116. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1117. unsigned long iflags;
  1118. spin_lock_irqsave(&pi->tx_lock, iflags);
  1119. mpsc_unfreeze(pi);
  1120. mpsc_copy_tx_data(pi);
  1121. mpsc_sdma_start_tx(pi);
  1122. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1123. pr_debug("mpsc_start_tx[%d]\n", port->line);
  1124. return;
  1125. }
  1126. static void
  1127. mpsc_start_rx(struct mpsc_port_info *pi)
  1128. {
  1129. pr_debug("mpsc_start_rx[%d]: Starting...\n", pi->port.line);
  1130. /* Issue a Receive Abort to clear any receive errors */
  1131. writel(MPSC_CHR_2_RA, pi->mpsc_base + MPSC_CHR_2);
  1132. if (pi->rcv_data) {
  1133. mpsc_enter_hunt(pi);
  1134. mpsc_sdma_cmd(pi, SDMA_SDCM_ERD);
  1135. }
  1136. return;
  1137. }
  1138. static void
  1139. mpsc_stop_rx(struct uart_port *port)
  1140. {
  1141. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1142. pr_debug("mpsc_stop_rx[%d]: Stopping...\n", port->line);
  1143. mpsc_sdma_cmd(pi, SDMA_SDCM_AR);
  1144. return;
  1145. }
  1146. static void
  1147. mpsc_enable_ms(struct uart_port *port)
  1148. {
  1149. return; /* Not supported */
  1150. }
  1151. static void
  1152. mpsc_break_ctl(struct uart_port *port, int ctl)
  1153. {
  1154. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1155. ulong flags;
  1156. u32 v;
  1157. v = ctl ? 0x00ff0000 : 0;
  1158. spin_lock_irqsave(&pi->port.lock, flags);
  1159. if (pi->mirror_regs)
  1160. pi->MPSC_CHR_1_m = v;
  1161. writel(v, pi->mpsc_base + MPSC_CHR_1);
  1162. spin_unlock_irqrestore(&pi->port.lock, flags);
  1163. return;
  1164. }
  1165. static int
  1166. mpsc_startup(struct uart_port *port)
  1167. {
  1168. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1169. u32 flag = 0;
  1170. int rc;
  1171. pr_debug("mpsc_startup[%d]: Starting up MPSC, irq: %d\n",
  1172. port->line, pi->port.irq);
  1173. if ((rc = mpsc_make_ready(pi)) == 0) {
  1174. /* Setup IRQ handler */
  1175. mpsc_sdma_intr_ack(pi);
  1176. /* If irq's are shared, need to set flag */
  1177. if (mpsc_ports[0].port.irq == mpsc_ports[1].port.irq)
  1178. flag = IRQF_SHARED;
  1179. if (request_irq(pi->port.irq, mpsc_sdma_intr, flag,
  1180. "mpsc-sdma", pi))
  1181. printk(KERN_ERR "MPSC: Can't get SDMA IRQ %d\n",
  1182. pi->port.irq);
  1183. mpsc_sdma_intr_unmask(pi, 0xf);
  1184. mpsc_sdma_set_rx_ring(pi, (struct mpsc_rx_desc *)(pi->rxr_p +
  1185. (pi->rxr_posn * MPSC_RXRE_SIZE)));
  1186. }
  1187. return rc;
  1188. }
  1189. static void
  1190. mpsc_shutdown(struct uart_port *port)
  1191. {
  1192. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1193. pr_debug("mpsc_shutdown[%d]: Shutting down MPSC\n", port->line);
  1194. mpsc_sdma_stop(pi);
  1195. free_irq(pi->port.irq, pi);
  1196. return;
  1197. }
  1198. static void
  1199. mpsc_set_termios(struct uart_port *port, struct ktermios *termios,
  1200. struct ktermios *old)
  1201. {
  1202. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1203. u32 baud;
  1204. ulong flags;
  1205. u32 chr_bits, stop_bits, par;
  1206. pi->c_iflag = termios->c_iflag;
  1207. pi->c_cflag = termios->c_cflag;
  1208. switch (termios->c_cflag & CSIZE) {
  1209. case CS5:
  1210. chr_bits = MPSC_MPCR_CL_5;
  1211. break;
  1212. case CS6:
  1213. chr_bits = MPSC_MPCR_CL_6;
  1214. break;
  1215. case CS7:
  1216. chr_bits = MPSC_MPCR_CL_7;
  1217. break;
  1218. case CS8:
  1219. default:
  1220. chr_bits = MPSC_MPCR_CL_8;
  1221. break;
  1222. }
  1223. if (termios->c_cflag & CSTOPB)
  1224. stop_bits = MPSC_MPCR_SBL_2;
  1225. else
  1226. stop_bits = MPSC_MPCR_SBL_1;
  1227. par = MPSC_CHR_2_PAR_EVEN;
  1228. if (termios->c_cflag & PARENB)
  1229. if (termios->c_cflag & PARODD)
  1230. par = MPSC_CHR_2_PAR_ODD;
  1231. #ifdef CMSPAR
  1232. if (termios->c_cflag & CMSPAR) {
  1233. if (termios->c_cflag & PARODD)
  1234. par = MPSC_CHR_2_PAR_MARK;
  1235. else
  1236. par = MPSC_CHR_2_PAR_SPACE;
  1237. }
  1238. #endif
  1239. baud = uart_get_baud_rate(port, termios, old, 0, port->uartclk);
  1240. spin_lock_irqsave(&pi->port.lock, flags);
  1241. uart_update_timeout(port, termios->c_cflag, baud);
  1242. mpsc_set_char_length(pi, chr_bits);
  1243. mpsc_set_stop_bit_length(pi, stop_bits);
  1244. mpsc_set_parity(pi, par);
  1245. mpsc_set_baudrate(pi, baud);
  1246. /* Characters/events to read */
  1247. pi->rcv_data = 1;
  1248. pi->port.read_status_mask = SDMA_DESC_CMDSTAT_OR;
  1249. if (termios->c_iflag & INPCK)
  1250. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1251. SDMA_DESC_CMDSTAT_FR;
  1252. if (termios->c_iflag & (BRKINT | PARMRK))
  1253. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1254. /* Characters/events to ignore */
  1255. pi->port.ignore_status_mask = 0;
  1256. if (termios->c_iflag & IGNPAR)
  1257. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1258. SDMA_DESC_CMDSTAT_FR;
  1259. if (termios->c_iflag & IGNBRK) {
  1260. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1261. if (termios->c_iflag & IGNPAR)
  1262. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_OR;
  1263. }
  1264. /* Ignore all chars if CREAD not set */
  1265. if (!(termios->c_cflag & CREAD))
  1266. pi->rcv_data = 0;
  1267. else
  1268. mpsc_start_rx(pi);
  1269. spin_unlock_irqrestore(&pi->port.lock, flags);
  1270. return;
  1271. }
  1272. static const char *
  1273. mpsc_type(struct uart_port *port)
  1274. {
  1275. pr_debug("mpsc_type[%d]: port type: %s\n", port->line,MPSC_DRIVER_NAME);
  1276. return MPSC_DRIVER_NAME;
  1277. }
  1278. static int
  1279. mpsc_request_port(struct uart_port *port)
  1280. {
  1281. /* Should make chip/platform specific call */
  1282. return 0;
  1283. }
  1284. static void
  1285. mpsc_release_port(struct uart_port *port)
  1286. {
  1287. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1288. if (pi->ready) {
  1289. mpsc_uninit_rings(pi);
  1290. mpsc_free_ring_mem(pi);
  1291. pi->ready = 0;
  1292. }
  1293. return;
  1294. }
  1295. static void
  1296. mpsc_config_port(struct uart_port *port, int flags)
  1297. {
  1298. return;
  1299. }
  1300. static int
  1301. mpsc_verify_port(struct uart_port *port, struct serial_struct *ser)
  1302. {
  1303. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1304. int rc = 0;
  1305. pr_debug("mpsc_verify_port[%d]: Verifying port data\n", pi->port.line);
  1306. if (ser->type != PORT_UNKNOWN && ser->type != PORT_MPSC)
  1307. rc = -EINVAL;
  1308. else if (pi->port.irq != ser->irq)
  1309. rc = -EINVAL;
  1310. else if (ser->io_type != SERIAL_IO_MEM)
  1311. rc = -EINVAL;
  1312. else if (pi->port.uartclk / 16 != ser->baud_base) /* Not sure */
  1313. rc = -EINVAL;
  1314. else if ((void *)pi->port.mapbase != ser->iomem_base)
  1315. rc = -EINVAL;
  1316. else if (pi->port.iobase != ser->port)
  1317. rc = -EINVAL;
  1318. else if (ser->hub6 != 0)
  1319. rc = -EINVAL;
  1320. return rc;
  1321. }
  1322. static struct uart_ops mpsc_pops = {
  1323. .tx_empty = mpsc_tx_empty,
  1324. .set_mctrl = mpsc_set_mctrl,
  1325. .get_mctrl = mpsc_get_mctrl,
  1326. .stop_tx = mpsc_stop_tx,
  1327. .start_tx = mpsc_start_tx,
  1328. .stop_rx = mpsc_stop_rx,
  1329. .enable_ms = mpsc_enable_ms,
  1330. .break_ctl = mpsc_break_ctl,
  1331. .startup = mpsc_startup,
  1332. .shutdown = mpsc_shutdown,
  1333. .set_termios = mpsc_set_termios,
  1334. .type = mpsc_type,
  1335. .release_port = mpsc_release_port,
  1336. .request_port = mpsc_request_port,
  1337. .config_port = mpsc_config_port,
  1338. .verify_port = mpsc_verify_port,
  1339. };
  1340. /*
  1341. ******************************************************************************
  1342. *
  1343. * Console Interface Routines
  1344. *
  1345. ******************************************************************************
  1346. */
  1347. #ifdef CONFIG_SERIAL_MPSC_CONSOLE
  1348. static void
  1349. mpsc_console_write(struct console *co, const char *s, uint count)
  1350. {
  1351. struct mpsc_port_info *pi = &mpsc_ports[co->index];
  1352. u8 *bp, *dp, add_cr = 0;
  1353. int i;
  1354. unsigned long iflags;
  1355. spin_lock_irqsave(&pi->tx_lock, iflags);
  1356. while (pi->txr_head != pi->txr_tail) {
  1357. while (mpsc_sdma_tx_active(pi))
  1358. udelay(100);
  1359. mpsc_sdma_intr_ack(pi);
  1360. mpsc_tx_intr(pi);
  1361. }
  1362. while (mpsc_sdma_tx_active(pi))
  1363. udelay(100);
  1364. while (count > 0) {
  1365. bp = dp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  1366. for (i = 0; i < MPSC_TXBE_SIZE; i++) {
  1367. if (count == 0)
  1368. break;
  1369. if (add_cr) {
  1370. *(dp++) = '\r';
  1371. add_cr = 0;
  1372. }
  1373. else {
  1374. *(dp++) = *s;
  1375. if (*(s++) == '\n') { /* add '\r' after '\n' */
  1376. add_cr = 1;
  1377. count++;
  1378. }
  1379. }
  1380. count--;
  1381. }
  1382. dma_cache_sync(pi->port.dev, (void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  1383. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1384. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1385. flush_dcache_range((ulong)bp,
  1386. (ulong)bp + MPSC_TXBE_SIZE);
  1387. #endif
  1388. mpsc_setup_tx_desc(pi, i, 0);
  1389. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1390. mpsc_sdma_start_tx(pi);
  1391. while (mpsc_sdma_tx_active(pi))
  1392. udelay(100);
  1393. pi->txr_tail = (pi->txr_tail + 1) & (MPSC_TXR_ENTRIES - 1);
  1394. }
  1395. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1396. return;
  1397. }
  1398. static int __init
  1399. mpsc_console_setup(struct console *co, char *options)
  1400. {
  1401. struct mpsc_port_info *pi;
  1402. int baud, bits, parity, flow;
  1403. pr_debug("mpsc_console_setup[%d]: options: %s\n", co->index, options);
  1404. if (co->index >= MPSC_NUM_CTLRS)
  1405. co->index = 0;
  1406. pi = &mpsc_ports[co->index];
  1407. baud = pi->default_baud;
  1408. bits = pi->default_bits;
  1409. parity = pi->default_parity;
  1410. flow = pi->default_flow;
  1411. if (!pi->port.ops)
  1412. return -ENODEV;
  1413. spin_lock_init(&pi->port.lock); /* Temporary fix--copied from 8250.c */
  1414. if (options)
  1415. uart_parse_options(options, &baud, &parity, &bits, &flow);
  1416. return uart_set_options(&pi->port, co, baud, parity, bits, flow);
  1417. }
  1418. static struct console mpsc_console = {
  1419. .name = MPSC_DEV_NAME,
  1420. .write = mpsc_console_write,
  1421. .device = uart_console_device,
  1422. .setup = mpsc_console_setup,
  1423. .flags = CON_PRINTBUFFER,
  1424. .index = -1,
  1425. .data = &mpsc_reg,
  1426. };
  1427. static int __init
  1428. mpsc_late_console_init(void)
  1429. {
  1430. pr_debug("mpsc_late_console_init: Enter\n");
  1431. if (!(mpsc_console.flags & CON_ENABLED))
  1432. register_console(&mpsc_console);
  1433. return 0;
  1434. }
  1435. late_initcall(mpsc_late_console_init);
  1436. #define MPSC_CONSOLE &mpsc_console
  1437. #else
  1438. #define MPSC_CONSOLE NULL
  1439. #endif
  1440. /*
  1441. ******************************************************************************
  1442. *
  1443. * Dummy Platform Driver to extract & map shared register regions
  1444. *
  1445. ******************************************************************************
  1446. */
  1447. static void
  1448. mpsc_resource_err(char *s)
  1449. {
  1450. printk(KERN_WARNING "MPSC: Platform device resource error in %s\n", s);
  1451. return;
  1452. }
  1453. static int
  1454. mpsc_shared_map_regs(struct platform_device *pd)
  1455. {
  1456. struct resource *r;
  1457. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1458. MPSC_ROUTING_BASE_ORDER)) && request_mem_region(r->start,
  1459. MPSC_ROUTING_REG_BLOCK_SIZE, "mpsc_routing_regs")) {
  1460. mpsc_shared_regs.mpsc_routing_base = ioremap(r->start,
  1461. MPSC_ROUTING_REG_BLOCK_SIZE);
  1462. mpsc_shared_regs.mpsc_routing_base_p = r->start;
  1463. }
  1464. else {
  1465. mpsc_resource_err("MPSC routing base");
  1466. return -ENOMEM;
  1467. }
  1468. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1469. MPSC_SDMA_INTR_BASE_ORDER)) && request_mem_region(r->start,
  1470. MPSC_SDMA_INTR_REG_BLOCK_SIZE, "sdma_intr_regs")) {
  1471. mpsc_shared_regs.sdma_intr_base = ioremap(r->start,
  1472. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1473. mpsc_shared_regs.sdma_intr_base_p = r->start;
  1474. }
  1475. else {
  1476. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1477. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1478. MPSC_ROUTING_REG_BLOCK_SIZE);
  1479. mpsc_resource_err("SDMA intr base");
  1480. return -ENOMEM;
  1481. }
  1482. return 0;
  1483. }
  1484. static void
  1485. mpsc_shared_unmap_regs(void)
  1486. {
  1487. if (!mpsc_shared_regs.mpsc_routing_base) {
  1488. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1489. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1490. MPSC_ROUTING_REG_BLOCK_SIZE);
  1491. }
  1492. if (!mpsc_shared_regs.sdma_intr_base) {
  1493. iounmap(mpsc_shared_regs.sdma_intr_base);
  1494. release_mem_region(mpsc_shared_regs.sdma_intr_base_p,
  1495. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1496. }
  1497. mpsc_shared_regs.mpsc_routing_base = NULL;
  1498. mpsc_shared_regs.sdma_intr_base = NULL;
  1499. mpsc_shared_regs.mpsc_routing_base_p = 0;
  1500. mpsc_shared_regs.sdma_intr_base_p = 0;
  1501. return;
  1502. }
  1503. static int
  1504. mpsc_shared_drv_probe(struct platform_device *dev)
  1505. {
  1506. struct mpsc_shared_pdata *pdata;
  1507. int rc = -ENODEV;
  1508. if (dev->id == 0) {
  1509. if (!(rc = mpsc_shared_map_regs(dev))) {
  1510. pdata = (struct mpsc_shared_pdata *)dev->dev.platform_data;
  1511. mpsc_shared_regs.MPSC_MRR_m = pdata->mrr_val;
  1512. mpsc_shared_regs.MPSC_RCRR_m= pdata->rcrr_val;
  1513. mpsc_shared_regs.MPSC_TCRR_m= pdata->tcrr_val;
  1514. mpsc_shared_regs.SDMA_INTR_CAUSE_m =
  1515. pdata->intr_cause_val;
  1516. mpsc_shared_regs.SDMA_INTR_MASK_m =
  1517. pdata->intr_mask_val;
  1518. rc = 0;
  1519. }
  1520. }
  1521. return rc;
  1522. }
  1523. static int
  1524. mpsc_shared_drv_remove(struct platform_device *dev)
  1525. {
  1526. int rc = -ENODEV;
  1527. if (dev->id == 0) {
  1528. mpsc_shared_unmap_regs();
  1529. mpsc_shared_regs.MPSC_MRR_m = 0;
  1530. mpsc_shared_regs.MPSC_RCRR_m = 0;
  1531. mpsc_shared_regs.MPSC_TCRR_m = 0;
  1532. mpsc_shared_regs.SDMA_INTR_CAUSE_m = 0;
  1533. mpsc_shared_regs.SDMA_INTR_MASK_m = 0;
  1534. rc = 0;
  1535. }
  1536. return rc;
  1537. }
  1538. static struct platform_driver mpsc_shared_driver = {
  1539. .probe = mpsc_shared_drv_probe,
  1540. .remove = mpsc_shared_drv_remove,
  1541. .driver = {
  1542. .name = MPSC_SHARED_NAME,
  1543. },
  1544. };
  1545. /*
  1546. ******************************************************************************
  1547. *
  1548. * Driver Interface Routines
  1549. *
  1550. ******************************************************************************
  1551. */
  1552. static struct uart_driver mpsc_reg = {
  1553. .owner = THIS_MODULE,
  1554. .driver_name = MPSC_DRIVER_NAME,
  1555. .dev_name = MPSC_DEV_NAME,
  1556. .major = MPSC_MAJOR,
  1557. .minor = MPSC_MINOR_START,
  1558. .nr = MPSC_NUM_CTLRS,
  1559. .cons = MPSC_CONSOLE,
  1560. };
  1561. static int
  1562. mpsc_drv_map_regs(struct mpsc_port_info *pi, struct platform_device *pd)
  1563. {
  1564. struct resource *r;
  1565. if ((r = platform_get_resource(pd, IORESOURCE_MEM, MPSC_BASE_ORDER)) &&
  1566. request_mem_region(r->start, MPSC_REG_BLOCK_SIZE, "mpsc_regs")){
  1567. pi->mpsc_base = ioremap(r->start, MPSC_REG_BLOCK_SIZE);
  1568. pi->mpsc_base_p = r->start;
  1569. }
  1570. else {
  1571. mpsc_resource_err("MPSC base");
  1572. return -ENOMEM;
  1573. }
  1574. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1575. MPSC_SDMA_BASE_ORDER)) && request_mem_region(r->start,
  1576. MPSC_SDMA_REG_BLOCK_SIZE, "sdma_regs")) {
  1577. pi->sdma_base = ioremap(r->start,MPSC_SDMA_REG_BLOCK_SIZE);
  1578. pi->sdma_base_p = r->start;
  1579. }
  1580. else {
  1581. mpsc_resource_err("SDMA base");
  1582. if (pi->mpsc_base) {
  1583. iounmap(pi->mpsc_base);
  1584. pi->mpsc_base = NULL;
  1585. }
  1586. return -ENOMEM;
  1587. }
  1588. if ((r = platform_get_resource(pd,IORESOURCE_MEM,MPSC_BRG_BASE_ORDER))
  1589. && request_mem_region(r->start, MPSC_BRG_REG_BLOCK_SIZE,
  1590. "brg_regs")) {
  1591. pi->brg_base = ioremap(r->start, MPSC_BRG_REG_BLOCK_SIZE);
  1592. pi->brg_base_p = r->start;
  1593. }
  1594. else {
  1595. mpsc_resource_err("BRG base");
  1596. if (pi->mpsc_base) {
  1597. iounmap(pi->mpsc_base);
  1598. pi->mpsc_base = NULL;
  1599. }
  1600. if (pi->sdma_base) {
  1601. iounmap(pi->sdma_base);
  1602. pi->sdma_base = NULL;
  1603. }
  1604. return -ENOMEM;
  1605. }
  1606. return 0;
  1607. }
  1608. static void
  1609. mpsc_drv_unmap_regs(struct mpsc_port_info *pi)
  1610. {
  1611. if (!pi->mpsc_base) {
  1612. iounmap(pi->mpsc_base);
  1613. release_mem_region(pi->mpsc_base_p, MPSC_REG_BLOCK_SIZE);
  1614. }
  1615. if (!pi->sdma_base) {
  1616. iounmap(pi->sdma_base);
  1617. release_mem_region(pi->sdma_base_p, MPSC_SDMA_REG_BLOCK_SIZE);
  1618. }
  1619. if (!pi->brg_base) {
  1620. iounmap(pi->brg_base);
  1621. release_mem_region(pi->brg_base_p, MPSC_BRG_REG_BLOCK_SIZE);
  1622. }
  1623. pi->mpsc_base = NULL;
  1624. pi->sdma_base = NULL;
  1625. pi->brg_base = NULL;
  1626. pi->mpsc_base_p = 0;
  1627. pi->sdma_base_p = 0;
  1628. pi->brg_base_p = 0;
  1629. return;
  1630. }
  1631. static void
  1632. mpsc_drv_get_platform_data(struct mpsc_port_info *pi,
  1633. struct platform_device *pd, int num)
  1634. {
  1635. struct mpsc_pdata *pdata;
  1636. pdata = (struct mpsc_pdata *)pd->dev.platform_data;
  1637. pi->port.uartclk = pdata->brg_clk_freq;
  1638. pi->port.iotype = UPIO_MEM;
  1639. pi->port.line = num;
  1640. pi->port.type = PORT_MPSC;
  1641. pi->port.fifosize = MPSC_TXBE_SIZE;
  1642. pi->port.membase = pi->mpsc_base;
  1643. pi->port.mapbase = (ulong)pi->mpsc_base;
  1644. pi->port.ops = &mpsc_pops;
  1645. pi->mirror_regs = pdata->mirror_regs;
  1646. pi->cache_mgmt = pdata->cache_mgmt;
  1647. pi->brg_can_tune = pdata->brg_can_tune;
  1648. pi->brg_clk_src = pdata->brg_clk_src;
  1649. pi->mpsc_max_idle = pdata->max_idle;
  1650. pi->default_baud = pdata->default_baud;
  1651. pi->default_bits = pdata->default_bits;
  1652. pi->default_parity = pdata->default_parity;
  1653. pi->default_flow = pdata->default_flow;
  1654. /* Initial values of mirrored regs */
  1655. pi->MPSC_CHR_1_m = pdata->chr_1_val;
  1656. pi->MPSC_CHR_2_m = pdata->chr_2_val;
  1657. pi->MPSC_CHR_10_m = pdata->chr_10_val;
  1658. pi->MPSC_MPCR_m = pdata->mpcr_val;
  1659. pi->BRG_BCR_m = pdata->bcr_val;
  1660. pi->shared_regs = &mpsc_shared_regs;
  1661. pi->port.irq = platform_get_irq(pd, 0);
  1662. return;
  1663. }
  1664. static int
  1665. mpsc_drv_probe(struct platform_device *dev)
  1666. {
  1667. struct mpsc_port_info *pi;
  1668. int rc = -ENODEV;
  1669. pr_debug("mpsc_drv_probe: Adding MPSC %d\n", dev->id);
  1670. if (dev->id < MPSC_NUM_CTLRS) {
  1671. pi = &mpsc_ports[dev->id];
  1672. if (!(rc = mpsc_drv_map_regs(pi, dev))) {
  1673. mpsc_drv_get_platform_data(pi, dev, dev->id);
  1674. if (!(rc = mpsc_make_ready(pi))) {
  1675. spin_lock_init(&pi->tx_lock);
  1676. if (!(rc = uart_add_one_port(&mpsc_reg,
  1677. &pi->port)))
  1678. rc = 0;
  1679. else {
  1680. mpsc_release_port(
  1681. (struct uart_port *)pi);
  1682. mpsc_drv_unmap_regs(pi);
  1683. }
  1684. }
  1685. else
  1686. mpsc_drv_unmap_regs(pi);
  1687. }
  1688. }
  1689. return rc;
  1690. }
  1691. static int
  1692. mpsc_drv_remove(struct platform_device *dev)
  1693. {
  1694. pr_debug("mpsc_drv_exit: Removing MPSC %d\n", dev->id);
  1695. if (dev->id < MPSC_NUM_CTLRS) {
  1696. uart_remove_one_port(&mpsc_reg, &mpsc_ports[dev->id].port);
  1697. mpsc_release_port((struct uart_port *)&mpsc_ports[dev->id].port);
  1698. mpsc_drv_unmap_regs(&mpsc_ports[dev->id]);
  1699. return 0;
  1700. }
  1701. else
  1702. return -ENODEV;
  1703. }
  1704. static struct platform_driver mpsc_driver = {
  1705. .probe = mpsc_drv_probe,
  1706. .remove = mpsc_drv_remove,
  1707. .driver = {
  1708. .name = MPSC_CTLR_NAME,
  1709. },
  1710. };
  1711. static int __init
  1712. mpsc_drv_init(void)
  1713. {
  1714. int rc;
  1715. printk(KERN_INFO "Serial: MPSC driver $Revision: 1.00 $\n");
  1716. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1717. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1718. if (!(rc = uart_register_driver(&mpsc_reg))) {
  1719. if (!(rc = platform_driver_register(&mpsc_shared_driver))) {
  1720. if ((rc = platform_driver_register(&mpsc_driver))) {
  1721. platform_driver_unregister(&mpsc_shared_driver);
  1722. uart_unregister_driver(&mpsc_reg);
  1723. }
  1724. }
  1725. else
  1726. uart_unregister_driver(&mpsc_reg);
  1727. }
  1728. return rc;
  1729. }
  1730. static void __exit
  1731. mpsc_drv_exit(void)
  1732. {
  1733. platform_driver_unregister(&mpsc_driver);
  1734. platform_driver_unregister(&mpsc_shared_driver);
  1735. uart_unregister_driver(&mpsc_reg);
  1736. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1737. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1738. return;
  1739. }
  1740. module_init(mpsc_drv_init);
  1741. module_exit(mpsc_drv_exit);
  1742. MODULE_AUTHOR("Mark A. Greer <mgreer@mvista.com>");
  1743. MODULE_DESCRIPTION("Generic Marvell MPSC serial/UART driver $Revision: 1.00 $");
  1744. MODULE_VERSION(MPSC_VERSION);
  1745. MODULE_LICENSE("GPL");
  1746. MODULE_ALIAS_CHARDEV_MAJOR(MPSC_MAJOR);