mpsc.c 56 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159
  1. /*
  2. * Generic driver for the MPSC (UART mode) on Marvell parts (e.g., GT64240,
  3. * GT64260, MV64340, MV64360, GT96100, ... ).
  4. *
  5. * Author: Mark A. Greer <mgreer@mvista.com>
  6. *
  7. * Based on an old MPSC driver that was in the linuxppc tree. It appears to
  8. * have been created by Chris Zankel (formerly of MontaVista) but there
  9. * is no proper Copyright so I'm not sure. Apparently, parts were also
  10. * taken from PPCBoot (now U-Boot). Also based on drivers/serial/8250.c
  11. * by Russell King.
  12. *
  13. * 2004 (c) MontaVista, Software, Inc. This file is licensed under
  14. * the terms of the GNU General Public License version 2. This program
  15. * is licensed "as is" without any warranty of any kind, whether express
  16. * or implied.
  17. */
  18. /*
  19. * The MPSC interface is much like a typical network controller's interface.
  20. * That is, you set up separate rings of descriptors for transmitting and
  21. * receiving data. There is also a pool of buffers with (one buffer per
  22. * descriptor) that incoming data are dma'd into or outgoing data are dma'd
  23. * out of.
  24. *
  25. * The MPSC requires two other controllers to be able to work. The Baud Rate
  26. * Generator (BRG) provides a clock at programmable frequencies which determines
  27. * the baud rate. The Serial DMA Controller (SDMA) takes incoming data from the
  28. * MPSC and DMA's it into memory or DMA's outgoing data and passes it to the
  29. * MPSC. It is actually the SDMA interrupt that the driver uses to keep the
  30. * transmit and receive "engines" going (i.e., indicate data has been
  31. * transmitted or received).
  32. *
  33. * NOTES:
  34. *
  35. * 1) Some chips have an erratum where several regs cannot be
  36. * read. To work around that, we keep a local copy of those regs in
  37. * 'mpsc_port_info'.
  38. *
  39. * 2) Some chips have an erratum where the ctlr will hang when the SDMA ctlr
  40. * accesses system mem with coherency enabled. For that reason, the driver
  41. * assumes that coherency for that ctlr has been disabled. This means
  42. * that when in a cache coherent system, the driver has to manually manage
  43. * the data cache on the areas that it touches because the dma_* macro are
  44. * basically no-ops.
  45. *
  46. * 3) There is an erratum (on PPC) where you can't use the instruction to do
  47. * a DMA_TO_DEVICE/cache clean so DMA_BIDIRECTIONAL/flushes are used in places
  48. * where a DMA_TO_DEVICE/clean would have [otherwise] sufficed.
  49. *
  50. * 4) AFAICT, hardware flow control isn't supported by the controller --MAG.
  51. */
  52. #if defined(CONFIG_SERIAL_MPSC_CONSOLE) && defined(CONFIG_MAGIC_SYSRQ)
  53. #define SUPPORT_SYSRQ
  54. #endif
  55. #include <linux/module.h>
  56. #include <linux/moduleparam.h>
  57. #include <linux/tty.h>
  58. #include <linux/tty_flip.h>
  59. #include <linux/ioport.h>
  60. #include <linux/init.h>
  61. #include <linux/console.h>
  62. #include <linux/sysrq.h>
  63. #include <linux/serial.h>
  64. #include <linux/serial_core.h>
  65. #include <linux/delay.h>
  66. #include <linux/device.h>
  67. #include <linux/dma-mapping.h>
  68. #include <linux/mv643xx.h>
  69. #include <linux/platform_device.h>
  70. #include <linux/gfp.h>
  71. #include <asm/io.h>
  72. #include <asm/irq.h>
  73. #define MPSC_NUM_CTLRS 2
  74. /*
  75. * Descriptors and buffers must be cache line aligned.
  76. * Buffers lengths must be multiple of cache line size.
  77. * Number of Tx & Rx descriptors must be powers of 2.
  78. */
  79. #define MPSC_RXR_ENTRIES 32
  80. #define MPSC_RXRE_SIZE dma_get_cache_alignment()
  81. #define MPSC_RXR_SIZE (MPSC_RXR_ENTRIES * MPSC_RXRE_SIZE)
  82. #define MPSC_RXBE_SIZE dma_get_cache_alignment()
  83. #define MPSC_RXB_SIZE (MPSC_RXR_ENTRIES * MPSC_RXBE_SIZE)
  84. #define MPSC_TXR_ENTRIES 32
  85. #define MPSC_TXRE_SIZE dma_get_cache_alignment()
  86. #define MPSC_TXR_SIZE (MPSC_TXR_ENTRIES * MPSC_TXRE_SIZE)
  87. #define MPSC_TXBE_SIZE dma_get_cache_alignment()
  88. #define MPSC_TXB_SIZE (MPSC_TXR_ENTRIES * MPSC_TXBE_SIZE)
  89. #define MPSC_DMA_ALLOC_SIZE (MPSC_RXR_SIZE + MPSC_RXB_SIZE + MPSC_TXR_SIZE \
  90. + MPSC_TXB_SIZE + dma_get_cache_alignment() /* for alignment */)
  91. /* Rx and Tx Ring entry descriptors -- assume entry size is <= cacheline size */
  92. struct mpsc_rx_desc {
  93. u16 bufsize;
  94. u16 bytecnt;
  95. u32 cmdstat;
  96. u32 link;
  97. u32 buf_ptr;
  98. } __attribute((packed));
  99. struct mpsc_tx_desc {
  100. u16 bytecnt;
  101. u16 shadow;
  102. u32 cmdstat;
  103. u32 link;
  104. u32 buf_ptr;
  105. } __attribute((packed));
  106. /*
  107. * Some regs that have the erratum that you can't read them are are shared
  108. * between the two MPSC controllers. This struct contains those shared regs.
  109. */
  110. struct mpsc_shared_regs {
  111. phys_addr_t mpsc_routing_base_p;
  112. phys_addr_t sdma_intr_base_p;
  113. void __iomem *mpsc_routing_base;
  114. void __iomem *sdma_intr_base;
  115. u32 MPSC_MRR_m;
  116. u32 MPSC_RCRR_m;
  117. u32 MPSC_TCRR_m;
  118. u32 SDMA_INTR_CAUSE_m;
  119. u32 SDMA_INTR_MASK_m;
  120. };
  121. /* The main driver data structure */
  122. struct mpsc_port_info {
  123. struct uart_port port; /* Overlay uart_port structure */
  124. /* Internal driver state for this ctlr */
  125. u8 ready;
  126. u8 rcv_data;
  127. tcflag_t c_iflag; /* save termios->c_iflag */
  128. tcflag_t c_cflag; /* save termios->c_cflag */
  129. /* Info passed in from platform */
  130. u8 mirror_regs; /* Need to mirror regs? */
  131. u8 cache_mgmt; /* Need manual cache mgmt? */
  132. u8 brg_can_tune; /* BRG has baud tuning? */
  133. u32 brg_clk_src;
  134. u16 mpsc_max_idle;
  135. int default_baud;
  136. int default_bits;
  137. int default_parity;
  138. int default_flow;
  139. /* Physical addresses of various blocks of registers (from platform) */
  140. phys_addr_t mpsc_base_p;
  141. phys_addr_t sdma_base_p;
  142. phys_addr_t brg_base_p;
  143. /* Virtual addresses of various blocks of registers (from platform) */
  144. void __iomem *mpsc_base;
  145. void __iomem *sdma_base;
  146. void __iomem *brg_base;
  147. /* Descriptor ring and buffer allocations */
  148. void *dma_region;
  149. dma_addr_t dma_region_p;
  150. dma_addr_t rxr; /* Rx descriptor ring */
  151. dma_addr_t rxr_p; /* Phys addr of rxr */
  152. u8 *rxb; /* Rx Ring I/O buf */
  153. u8 *rxb_p; /* Phys addr of rxb */
  154. u32 rxr_posn; /* First desc w/ Rx data */
  155. dma_addr_t txr; /* Tx descriptor ring */
  156. dma_addr_t txr_p; /* Phys addr of txr */
  157. u8 *txb; /* Tx Ring I/O buf */
  158. u8 *txb_p; /* Phys addr of txb */
  159. int txr_head; /* Where new data goes */
  160. int txr_tail; /* Where sent data comes off */
  161. spinlock_t tx_lock; /* transmit lock */
  162. /* Mirrored values of regs we can't read (if 'mirror_regs' set) */
  163. u32 MPSC_MPCR_m;
  164. u32 MPSC_CHR_1_m;
  165. u32 MPSC_CHR_2_m;
  166. u32 MPSC_CHR_10_m;
  167. u32 BRG_BCR_m;
  168. struct mpsc_shared_regs *shared_regs;
  169. };
  170. /* Hooks to platform-specific code */
  171. int mpsc_platform_register_driver(void);
  172. void mpsc_platform_unregister_driver(void);
  173. /* Hooks back in to mpsc common to be called by platform-specific code */
  174. struct mpsc_port_info *mpsc_device_probe(int index);
  175. struct mpsc_port_info *mpsc_device_remove(int index);
  176. /* Main MPSC Configuration Register Offsets */
  177. #define MPSC_MMCRL 0x0000
  178. #define MPSC_MMCRH 0x0004
  179. #define MPSC_MPCR 0x0008
  180. #define MPSC_CHR_1 0x000c
  181. #define MPSC_CHR_2 0x0010
  182. #define MPSC_CHR_3 0x0014
  183. #define MPSC_CHR_4 0x0018
  184. #define MPSC_CHR_5 0x001c
  185. #define MPSC_CHR_6 0x0020
  186. #define MPSC_CHR_7 0x0024
  187. #define MPSC_CHR_8 0x0028
  188. #define MPSC_CHR_9 0x002c
  189. #define MPSC_CHR_10 0x0030
  190. #define MPSC_CHR_11 0x0034
  191. #define MPSC_MPCR_FRZ (1 << 9)
  192. #define MPSC_MPCR_CL_5 0
  193. #define MPSC_MPCR_CL_6 1
  194. #define MPSC_MPCR_CL_7 2
  195. #define MPSC_MPCR_CL_8 3
  196. #define MPSC_MPCR_SBL_1 0
  197. #define MPSC_MPCR_SBL_2 1
  198. #define MPSC_CHR_2_TEV (1<<1)
  199. #define MPSC_CHR_2_TA (1<<7)
  200. #define MPSC_CHR_2_TTCS (1<<9)
  201. #define MPSC_CHR_2_REV (1<<17)
  202. #define MPSC_CHR_2_RA (1<<23)
  203. #define MPSC_CHR_2_CRD (1<<25)
  204. #define MPSC_CHR_2_EH (1<<31)
  205. #define MPSC_CHR_2_PAR_ODD 0
  206. #define MPSC_CHR_2_PAR_SPACE 1
  207. #define MPSC_CHR_2_PAR_EVEN 2
  208. #define MPSC_CHR_2_PAR_MARK 3
  209. /* MPSC Signal Routing */
  210. #define MPSC_MRR 0x0000
  211. #define MPSC_RCRR 0x0004
  212. #define MPSC_TCRR 0x0008
  213. /* Serial DMA Controller Interface Registers */
  214. #define SDMA_SDC 0x0000
  215. #define SDMA_SDCM 0x0008
  216. #define SDMA_RX_DESC 0x0800
  217. #define SDMA_RX_BUF_PTR 0x0808
  218. #define SDMA_SCRDP 0x0810
  219. #define SDMA_TX_DESC 0x0c00
  220. #define SDMA_SCTDP 0x0c10
  221. #define SDMA_SFTDP 0x0c14
  222. #define SDMA_DESC_CMDSTAT_PE (1<<0)
  223. #define SDMA_DESC_CMDSTAT_CDL (1<<1)
  224. #define SDMA_DESC_CMDSTAT_FR (1<<3)
  225. #define SDMA_DESC_CMDSTAT_OR (1<<6)
  226. #define SDMA_DESC_CMDSTAT_BR (1<<9)
  227. #define SDMA_DESC_CMDSTAT_MI (1<<10)
  228. #define SDMA_DESC_CMDSTAT_A (1<<11)
  229. #define SDMA_DESC_CMDSTAT_AM (1<<12)
  230. #define SDMA_DESC_CMDSTAT_CT (1<<13)
  231. #define SDMA_DESC_CMDSTAT_C (1<<14)
  232. #define SDMA_DESC_CMDSTAT_ES (1<<15)
  233. #define SDMA_DESC_CMDSTAT_L (1<<16)
  234. #define SDMA_DESC_CMDSTAT_F (1<<17)
  235. #define SDMA_DESC_CMDSTAT_P (1<<18)
  236. #define SDMA_DESC_CMDSTAT_EI (1<<23)
  237. #define SDMA_DESC_CMDSTAT_O (1<<31)
  238. #define SDMA_DESC_DFLT (SDMA_DESC_CMDSTAT_O \
  239. | SDMA_DESC_CMDSTAT_EI)
  240. #define SDMA_SDC_RFT (1<<0)
  241. #define SDMA_SDC_SFM (1<<1)
  242. #define SDMA_SDC_BLMR (1<<6)
  243. #define SDMA_SDC_BLMT (1<<7)
  244. #define SDMA_SDC_POVR (1<<8)
  245. #define SDMA_SDC_RIFB (1<<9)
  246. #define SDMA_SDCM_ERD (1<<7)
  247. #define SDMA_SDCM_AR (1<<15)
  248. #define SDMA_SDCM_STD (1<<16)
  249. #define SDMA_SDCM_TXD (1<<23)
  250. #define SDMA_SDCM_AT (1<<31)
  251. #define SDMA_0_CAUSE_RXBUF (1<<0)
  252. #define SDMA_0_CAUSE_RXERR (1<<1)
  253. #define SDMA_0_CAUSE_TXBUF (1<<2)
  254. #define SDMA_0_CAUSE_TXEND (1<<3)
  255. #define SDMA_1_CAUSE_RXBUF (1<<8)
  256. #define SDMA_1_CAUSE_RXERR (1<<9)
  257. #define SDMA_1_CAUSE_TXBUF (1<<10)
  258. #define SDMA_1_CAUSE_TXEND (1<<11)
  259. #define SDMA_CAUSE_RX_MASK (SDMA_0_CAUSE_RXBUF | SDMA_0_CAUSE_RXERR \
  260. | SDMA_1_CAUSE_RXBUF | SDMA_1_CAUSE_RXERR)
  261. #define SDMA_CAUSE_TX_MASK (SDMA_0_CAUSE_TXBUF | SDMA_0_CAUSE_TXEND \
  262. | SDMA_1_CAUSE_TXBUF | SDMA_1_CAUSE_TXEND)
  263. /* SDMA Interrupt registers */
  264. #define SDMA_INTR_CAUSE 0x0000
  265. #define SDMA_INTR_MASK 0x0080
  266. /* Baud Rate Generator Interface Registers */
  267. #define BRG_BCR 0x0000
  268. #define BRG_BTR 0x0004
  269. /*
  270. * Define how this driver is known to the outside (we've been assigned a
  271. * range on the "Low-density serial ports" major).
  272. */
  273. #define MPSC_MAJOR 204
  274. #define MPSC_MINOR_START 44
  275. #define MPSC_DRIVER_NAME "MPSC"
  276. #define MPSC_DEV_NAME "ttyMM"
  277. #define MPSC_VERSION "1.00"
  278. static struct mpsc_port_info mpsc_ports[MPSC_NUM_CTLRS];
  279. static struct mpsc_shared_regs mpsc_shared_regs;
  280. static struct uart_driver mpsc_reg;
  281. static void mpsc_start_rx(struct mpsc_port_info *pi);
  282. static void mpsc_free_ring_mem(struct mpsc_port_info *pi);
  283. static void mpsc_release_port(struct uart_port *port);
  284. /*
  285. ******************************************************************************
  286. *
  287. * Baud Rate Generator Routines (BRG)
  288. *
  289. ******************************************************************************
  290. */
  291. static void mpsc_brg_init(struct mpsc_port_info *pi, u32 clk_src)
  292. {
  293. u32 v;
  294. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  295. v = (v & ~(0xf << 18)) | ((clk_src & 0xf) << 18);
  296. if (pi->brg_can_tune)
  297. v &= ~(1 << 25);
  298. if (pi->mirror_regs)
  299. pi->BRG_BCR_m = v;
  300. writel(v, pi->brg_base + BRG_BCR);
  301. writel(readl(pi->brg_base + BRG_BTR) & 0xffff0000,
  302. pi->brg_base + BRG_BTR);
  303. }
  304. static void mpsc_brg_enable(struct mpsc_port_info *pi)
  305. {
  306. u32 v;
  307. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  308. v |= (1 << 16);
  309. if (pi->mirror_regs)
  310. pi->BRG_BCR_m = v;
  311. writel(v, pi->brg_base + BRG_BCR);
  312. }
  313. static void mpsc_brg_disable(struct mpsc_port_info *pi)
  314. {
  315. u32 v;
  316. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  317. v &= ~(1 << 16);
  318. if (pi->mirror_regs)
  319. pi->BRG_BCR_m = v;
  320. writel(v, pi->brg_base + BRG_BCR);
  321. }
  322. /*
  323. * To set the baud, we adjust the CDV field in the BRG_BCR reg.
  324. * From manual: Baud = clk / ((CDV+1)*2) ==> CDV = (clk / (baud*2)) - 1.
  325. * However, the input clock is divided by 16 in the MPSC b/c of how
  326. * 'MPSC_MMCRH' was set up so we have to divide the 'clk' used in our
  327. * calculation by 16 to account for that. So the real calculation
  328. * that accounts for the way the mpsc is set up is:
  329. * CDV = (clk / (baud*2*16)) - 1 ==> CDV = (clk / (baud << 5)) - 1.
  330. */
  331. static void mpsc_set_baudrate(struct mpsc_port_info *pi, u32 baud)
  332. {
  333. u32 cdv = (pi->port.uartclk / (baud << 5)) - 1;
  334. u32 v;
  335. mpsc_brg_disable(pi);
  336. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  337. v = (v & 0xffff0000) | (cdv & 0xffff);
  338. if (pi->mirror_regs)
  339. pi->BRG_BCR_m = v;
  340. writel(v, pi->brg_base + BRG_BCR);
  341. mpsc_brg_enable(pi);
  342. }
  343. /*
  344. ******************************************************************************
  345. *
  346. * Serial DMA Routines (SDMA)
  347. *
  348. ******************************************************************************
  349. */
  350. static void mpsc_sdma_burstsize(struct mpsc_port_info *pi, u32 burst_size)
  351. {
  352. u32 v;
  353. pr_debug("mpsc_sdma_burstsize[%d]: burst_size: %d\n",
  354. pi->port.line, burst_size);
  355. burst_size >>= 3; /* Divide by 8 b/c reg values are 8-byte chunks */
  356. if (burst_size < 2)
  357. v = 0x0; /* 1 64-bit word */
  358. else if (burst_size < 4)
  359. v = 0x1; /* 2 64-bit words */
  360. else if (burst_size < 8)
  361. v = 0x2; /* 4 64-bit words */
  362. else
  363. v = 0x3; /* 8 64-bit words */
  364. writel((readl(pi->sdma_base + SDMA_SDC) & (0x3 << 12)) | (v << 12),
  365. pi->sdma_base + SDMA_SDC);
  366. }
  367. static void mpsc_sdma_init(struct mpsc_port_info *pi, u32 burst_size)
  368. {
  369. pr_debug("mpsc_sdma_init[%d]: burst_size: %d\n", pi->port.line,
  370. burst_size);
  371. writel((readl(pi->sdma_base + SDMA_SDC) & 0x3ff) | 0x03f,
  372. pi->sdma_base + SDMA_SDC);
  373. mpsc_sdma_burstsize(pi, burst_size);
  374. }
  375. static u32 mpsc_sdma_intr_mask(struct mpsc_port_info *pi, u32 mask)
  376. {
  377. u32 old, v;
  378. pr_debug("mpsc_sdma_intr_mask[%d]: mask: 0x%x\n", pi->port.line, mask);
  379. old = v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  380. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  381. mask &= 0xf;
  382. if (pi->port.line)
  383. mask <<= 8;
  384. v &= ~mask;
  385. if (pi->mirror_regs)
  386. pi->shared_regs->SDMA_INTR_MASK_m = v;
  387. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  388. if (pi->port.line)
  389. old >>= 8;
  390. return old & 0xf;
  391. }
  392. static void mpsc_sdma_intr_unmask(struct mpsc_port_info *pi, u32 mask)
  393. {
  394. u32 v;
  395. pr_debug("mpsc_sdma_intr_unmask[%d]: mask: 0x%x\n", pi->port.line,mask);
  396. v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m
  397. : readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  398. mask &= 0xf;
  399. if (pi->port.line)
  400. mask <<= 8;
  401. v |= mask;
  402. if (pi->mirror_regs)
  403. pi->shared_regs->SDMA_INTR_MASK_m = v;
  404. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  405. }
  406. static void mpsc_sdma_intr_ack(struct mpsc_port_info *pi)
  407. {
  408. pr_debug("mpsc_sdma_intr_ack[%d]: Acknowledging IRQ\n", pi->port.line);
  409. if (pi->mirror_regs)
  410. pi->shared_regs->SDMA_INTR_CAUSE_m = 0;
  411. writeb(0x00, pi->shared_regs->sdma_intr_base + SDMA_INTR_CAUSE
  412. + pi->port.line);
  413. }
  414. static void mpsc_sdma_set_rx_ring(struct mpsc_port_info *pi,
  415. struct mpsc_rx_desc *rxre_p)
  416. {
  417. pr_debug("mpsc_sdma_set_rx_ring[%d]: rxre_p: 0x%x\n",
  418. pi->port.line, (u32)rxre_p);
  419. writel((u32)rxre_p, pi->sdma_base + SDMA_SCRDP);
  420. }
  421. static void mpsc_sdma_set_tx_ring(struct mpsc_port_info *pi,
  422. struct mpsc_tx_desc *txre_p)
  423. {
  424. writel((u32)txre_p, pi->sdma_base + SDMA_SFTDP);
  425. writel((u32)txre_p, pi->sdma_base + SDMA_SCTDP);
  426. }
  427. static void mpsc_sdma_cmd(struct mpsc_port_info *pi, u32 val)
  428. {
  429. u32 v;
  430. v = readl(pi->sdma_base + SDMA_SDCM);
  431. if (val)
  432. v |= val;
  433. else
  434. v = 0;
  435. wmb();
  436. writel(v, pi->sdma_base + SDMA_SDCM);
  437. wmb();
  438. }
  439. static uint mpsc_sdma_tx_active(struct mpsc_port_info *pi)
  440. {
  441. return readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_TXD;
  442. }
  443. static void mpsc_sdma_start_tx(struct mpsc_port_info *pi)
  444. {
  445. struct mpsc_tx_desc *txre, *txre_p;
  446. /* If tx isn't running & there's a desc ready to go, start it */
  447. if (!mpsc_sdma_tx_active(pi)) {
  448. txre = (struct mpsc_tx_desc *)(pi->txr
  449. + (pi->txr_tail * MPSC_TXRE_SIZE));
  450. dma_cache_sync(pi->port.dev, (void *)txre, MPSC_TXRE_SIZE,
  451. DMA_FROM_DEVICE);
  452. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  453. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  454. invalidate_dcache_range((ulong)txre,
  455. (ulong)txre + MPSC_TXRE_SIZE);
  456. #endif
  457. if (be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O) {
  458. txre_p = (struct mpsc_tx_desc *)
  459. (pi->txr_p + (pi->txr_tail * MPSC_TXRE_SIZE));
  460. mpsc_sdma_set_tx_ring(pi, txre_p);
  461. mpsc_sdma_cmd(pi, SDMA_SDCM_STD | SDMA_SDCM_TXD);
  462. }
  463. }
  464. }
  465. static void mpsc_sdma_stop(struct mpsc_port_info *pi)
  466. {
  467. pr_debug("mpsc_sdma_stop[%d]: Stopping SDMA\n", pi->port.line);
  468. /* Abort any SDMA transfers */
  469. mpsc_sdma_cmd(pi, 0);
  470. mpsc_sdma_cmd(pi, SDMA_SDCM_AR | SDMA_SDCM_AT);
  471. /* Clear the SDMA current and first TX and RX pointers */
  472. mpsc_sdma_set_tx_ring(pi, NULL);
  473. mpsc_sdma_set_rx_ring(pi, NULL);
  474. /* Disable interrupts */
  475. mpsc_sdma_intr_mask(pi, 0xf);
  476. mpsc_sdma_intr_ack(pi);
  477. }
  478. /*
  479. ******************************************************************************
  480. *
  481. * Multi-Protocol Serial Controller Routines (MPSC)
  482. *
  483. ******************************************************************************
  484. */
  485. static void mpsc_hw_init(struct mpsc_port_info *pi)
  486. {
  487. u32 v;
  488. pr_debug("mpsc_hw_init[%d]: Initializing hardware\n", pi->port.line);
  489. /* Set up clock routing */
  490. if (pi->mirror_regs) {
  491. v = pi->shared_regs->MPSC_MRR_m;
  492. v &= ~0x1c7;
  493. pi->shared_regs->MPSC_MRR_m = v;
  494. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  495. v = pi->shared_regs->MPSC_RCRR_m;
  496. v = (v & ~0xf0f) | 0x100;
  497. pi->shared_regs->MPSC_RCRR_m = v;
  498. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  499. v = pi->shared_regs->MPSC_TCRR_m;
  500. v = (v & ~0xf0f) | 0x100;
  501. pi->shared_regs->MPSC_TCRR_m = v;
  502. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  503. } else {
  504. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  505. v &= ~0x1c7;
  506. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  507. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  508. v = (v & ~0xf0f) | 0x100;
  509. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  510. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  511. v = (v & ~0xf0f) | 0x100;
  512. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  513. }
  514. /* Put MPSC in UART mode & enabel Tx/Rx egines */
  515. writel(0x000004c4, pi->mpsc_base + MPSC_MMCRL);
  516. /* No preamble, 16x divider, low-latency, */
  517. writel(0x04400400, pi->mpsc_base + MPSC_MMCRH);
  518. mpsc_set_baudrate(pi, pi->default_baud);
  519. if (pi->mirror_regs) {
  520. pi->MPSC_CHR_1_m = 0;
  521. pi->MPSC_CHR_2_m = 0;
  522. }
  523. writel(0, pi->mpsc_base + MPSC_CHR_1);
  524. writel(0, pi->mpsc_base + MPSC_CHR_2);
  525. writel(pi->mpsc_max_idle, pi->mpsc_base + MPSC_CHR_3);
  526. writel(0, pi->mpsc_base + MPSC_CHR_4);
  527. writel(0, pi->mpsc_base + MPSC_CHR_5);
  528. writel(0, pi->mpsc_base + MPSC_CHR_6);
  529. writel(0, pi->mpsc_base + MPSC_CHR_7);
  530. writel(0, pi->mpsc_base + MPSC_CHR_8);
  531. writel(0, pi->mpsc_base + MPSC_CHR_9);
  532. writel(0, pi->mpsc_base + MPSC_CHR_10);
  533. }
  534. static void mpsc_enter_hunt(struct mpsc_port_info *pi)
  535. {
  536. pr_debug("mpsc_enter_hunt[%d]: Hunting...\n", pi->port.line);
  537. if (pi->mirror_regs) {
  538. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_EH,
  539. pi->mpsc_base + MPSC_CHR_2);
  540. /* Erratum prevents reading CHR_2 so just delay for a while */
  541. udelay(100);
  542. } else {
  543. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_EH,
  544. pi->mpsc_base + MPSC_CHR_2);
  545. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_EH)
  546. udelay(10);
  547. }
  548. }
  549. static void mpsc_freeze(struct mpsc_port_info *pi)
  550. {
  551. u32 v;
  552. pr_debug("mpsc_freeze[%d]: Freezing\n", pi->port.line);
  553. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  554. readl(pi->mpsc_base + MPSC_MPCR);
  555. v |= MPSC_MPCR_FRZ;
  556. if (pi->mirror_regs)
  557. pi->MPSC_MPCR_m = v;
  558. writel(v, pi->mpsc_base + MPSC_MPCR);
  559. }
  560. static void mpsc_unfreeze(struct mpsc_port_info *pi)
  561. {
  562. u32 v;
  563. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  564. readl(pi->mpsc_base + MPSC_MPCR);
  565. v &= ~MPSC_MPCR_FRZ;
  566. if (pi->mirror_regs)
  567. pi->MPSC_MPCR_m = v;
  568. writel(v, pi->mpsc_base + MPSC_MPCR);
  569. pr_debug("mpsc_unfreeze[%d]: Unfrozen\n", pi->port.line);
  570. }
  571. static void mpsc_set_char_length(struct mpsc_port_info *pi, u32 len)
  572. {
  573. u32 v;
  574. pr_debug("mpsc_set_char_length[%d]: char len: %d\n", pi->port.line,len);
  575. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  576. readl(pi->mpsc_base + MPSC_MPCR);
  577. v = (v & ~(0x3 << 12)) | ((len & 0x3) << 12);
  578. if (pi->mirror_regs)
  579. pi->MPSC_MPCR_m = v;
  580. writel(v, pi->mpsc_base + MPSC_MPCR);
  581. }
  582. static void mpsc_set_stop_bit_length(struct mpsc_port_info *pi, u32 len)
  583. {
  584. u32 v;
  585. pr_debug("mpsc_set_stop_bit_length[%d]: stop bits: %d\n",
  586. pi->port.line, len);
  587. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  588. readl(pi->mpsc_base + MPSC_MPCR);
  589. v = (v & ~(1 << 14)) | ((len & 0x1) << 14);
  590. if (pi->mirror_regs)
  591. pi->MPSC_MPCR_m = v;
  592. writel(v, pi->mpsc_base + MPSC_MPCR);
  593. }
  594. static void mpsc_set_parity(struct mpsc_port_info *pi, u32 p)
  595. {
  596. u32 v;
  597. pr_debug("mpsc_set_parity[%d]: parity bits: 0x%x\n", pi->port.line, p);
  598. v = (pi->mirror_regs) ? pi->MPSC_CHR_2_m :
  599. readl(pi->mpsc_base + MPSC_CHR_2);
  600. p &= 0x3;
  601. v = (v & ~0xc000c) | (p << 18) | (p << 2);
  602. if (pi->mirror_regs)
  603. pi->MPSC_CHR_2_m = v;
  604. writel(v, pi->mpsc_base + MPSC_CHR_2);
  605. }
  606. /*
  607. ******************************************************************************
  608. *
  609. * Driver Init Routines
  610. *
  611. ******************************************************************************
  612. */
  613. static void mpsc_init_hw(struct mpsc_port_info *pi)
  614. {
  615. pr_debug("mpsc_init_hw[%d]: Initializing\n", pi->port.line);
  616. mpsc_brg_init(pi, pi->brg_clk_src);
  617. mpsc_brg_enable(pi);
  618. mpsc_sdma_init(pi, dma_get_cache_alignment()); /* burst a cacheline */
  619. mpsc_sdma_stop(pi);
  620. mpsc_hw_init(pi);
  621. }
  622. static int mpsc_alloc_ring_mem(struct mpsc_port_info *pi)
  623. {
  624. int rc = 0;
  625. pr_debug("mpsc_alloc_ring_mem[%d]: Allocating ring mem\n",
  626. pi->port.line);
  627. if (!pi->dma_region) {
  628. if (!dma_supported(pi->port.dev, 0xffffffff)) {
  629. printk(KERN_ERR "MPSC: Inadequate DMA support\n");
  630. rc = -ENXIO;
  631. } else if ((pi->dma_region = dma_alloc_noncoherent(pi->port.dev,
  632. MPSC_DMA_ALLOC_SIZE,
  633. &pi->dma_region_p, GFP_KERNEL))
  634. == NULL) {
  635. printk(KERN_ERR "MPSC: Can't alloc Desc region\n");
  636. rc = -ENOMEM;
  637. }
  638. }
  639. return rc;
  640. }
  641. static void mpsc_free_ring_mem(struct mpsc_port_info *pi)
  642. {
  643. pr_debug("mpsc_free_ring_mem[%d]: Freeing ring mem\n", pi->port.line);
  644. if (pi->dma_region) {
  645. dma_free_noncoherent(pi->port.dev, MPSC_DMA_ALLOC_SIZE,
  646. pi->dma_region, pi->dma_region_p);
  647. pi->dma_region = NULL;
  648. pi->dma_region_p = (dma_addr_t)NULL;
  649. }
  650. }
  651. static void mpsc_init_rings(struct mpsc_port_info *pi)
  652. {
  653. struct mpsc_rx_desc *rxre;
  654. struct mpsc_tx_desc *txre;
  655. dma_addr_t dp, dp_p;
  656. u8 *bp, *bp_p;
  657. int i;
  658. pr_debug("mpsc_init_rings[%d]: Initializing rings\n", pi->port.line);
  659. BUG_ON(pi->dma_region == NULL);
  660. memset(pi->dma_region, 0, MPSC_DMA_ALLOC_SIZE);
  661. /*
  662. * Descriptors & buffers are multiples of cacheline size and must be
  663. * cacheline aligned.
  664. */
  665. dp = ALIGN((u32)pi->dma_region, dma_get_cache_alignment());
  666. dp_p = ALIGN((u32)pi->dma_region_p, dma_get_cache_alignment());
  667. /*
  668. * Partition dma region into rx ring descriptor, rx buffers,
  669. * tx ring descriptors, and tx buffers.
  670. */
  671. pi->rxr = dp;
  672. pi->rxr_p = dp_p;
  673. dp += MPSC_RXR_SIZE;
  674. dp_p += MPSC_RXR_SIZE;
  675. pi->rxb = (u8 *)dp;
  676. pi->rxb_p = (u8 *)dp_p;
  677. dp += MPSC_RXB_SIZE;
  678. dp_p += MPSC_RXB_SIZE;
  679. pi->rxr_posn = 0;
  680. pi->txr = dp;
  681. pi->txr_p = dp_p;
  682. dp += MPSC_TXR_SIZE;
  683. dp_p += MPSC_TXR_SIZE;
  684. pi->txb = (u8 *)dp;
  685. pi->txb_p = (u8 *)dp_p;
  686. pi->txr_head = 0;
  687. pi->txr_tail = 0;
  688. /* Init rx ring descriptors */
  689. dp = pi->rxr;
  690. dp_p = pi->rxr_p;
  691. bp = pi->rxb;
  692. bp_p = pi->rxb_p;
  693. for (i = 0; i < MPSC_RXR_ENTRIES; i++) {
  694. rxre = (struct mpsc_rx_desc *)dp;
  695. rxre->bufsize = cpu_to_be16(MPSC_RXBE_SIZE);
  696. rxre->bytecnt = cpu_to_be16(0);
  697. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O
  698. | SDMA_DESC_CMDSTAT_EI | SDMA_DESC_CMDSTAT_F
  699. | SDMA_DESC_CMDSTAT_L);
  700. rxre->link = cpu_to_be32(dp_p + MPSC_RXRE_SIZE);
  701. rxre->buf_ptr = cpu_to_be32(bp_p);
  702. dp += MPSC_RXRE_SIZE;
  703. dp_p += MPSC_RXRE_SIZE;
  704. bp += MPSC_RXBE_SIZE;
  705. bp_p += MPSC_RXBE_SIZE;
  706. }
  707. rxre->link = cpu_to_be32(pi->rxr_p); /* Wrap last back to first */
  708. /* Init tx ring descriptors */
  709. dp = pi->txr;
  710. dp_p = pi->txr_p;
  711. bp = pi->txb;
  712. bp_p = pi->txb_p;
  713. for (i = 0; i < MPSC_TXR_ENTRIES; i++) {
  714. txre = (struct mpsc_tx_desc *)dp;
  715. txre->link = cpu_to_be32(dp_p + MPSC_TXRE_SIZE);
  716. txre->buf_ptr = cpu_to_be32(bp_p);
  717. dp += MPSC_TXRE_SIZE;
  718. dp_p += MPSC_TXRE_SIZE;
  719. bp += MPSC_TXBE_SIZE;
  720. bp_p += MPSC_TXBE_SIZE;
  721. }
  722. txre->link = cpu_to_be32(pi->txr_p); /* Wrap last back to first */
  723. dma_cache_sync(pi->port.dev, (void *)pi->dma_region,
  724. MPSC_DMA_ALLOC_SIZE, DMA_BIDIRECTIONAL);
  725. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  726. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  727. flush_dcache_range((ulong)pi->dma_region,
  728. (ulong)pi->dma_region
  729. + MPSC_DMA_ALLOC_SIZE);
  730. #endif
  731. return;
  732. }
  733. static void mpsc_uninit_rings(struct mpsc_port_info *pi)
  734. {
  735. pr_debug("mpsc_uninit_rings[%d]: Uninitializing rings\n",pi->port.line);
  736. BUG_ON(pi->dma_region == NULL);
  737. pi->rxr = 0;
  738. pi->rxr_p = 0;
  739. pi->rxb = NULL;
  740. pi->rxb_p = NULL;
  741. pi->rxr_posn = 0;
  742. pi->txr = 0;
  743. pi->txr_p = 0;
  744. pi->txb = NULL;
  745. pi->txb_p = NULL;
  746. pi->txr_head = 0;
  747. pi->txr_tail = 0;
  748. }
  749. static int mpsc_make_ready(struct mpsc_port_info *pi)
  750. {
  751. int rc;
  752. pr_debug("mpsc_make_ready[%d]: Making cltr ready\n", pi->port.line);
  753. if (!pi->ready) {
  754. mpsc_init_hw(pi);
  755. if ((rc = mpsc_alloc_ring_mem(pi)))
  756. return rc;
  757. mpsc_init_rings(pi);
  758. pi->ready = 1;
  759. }
  760. return 0;
  761. }
  762. #ifdef CONFIG_CONSOLE_POLL
  763. static int serial_polled;
  764. #endif
  765. /*
  766. ******************************************************************************
  767. *
  768. * Interrupt Handling Routines
  769. *
  770. ******************************************************************************
  771. */
  772. static int mpsc_rx_intr(struct mpsc_port_info *pi)
  773. {
  774. struct mpsc_rx_desc *rxre;
  775. struct tty_struct *tty = pi->port.state->port.tty;
  776. u32 cmdstat, bytes_in, i;
  777. int rc = 0;
  778. u8 *bp;
  779. char flag = TTY_NORMAL;
  780. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  781. rxre = (struct mpsc_rx_desc *)(pi->rxr + (pi->rxr_posn*MPSC_RXRE_SIZE));
  782. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE,
  783. DMA_FROM_DEVICE);
  784. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  785. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  786. invalidate_dcache_range((ulong)rxre,
  787. (ulong)rxre + MPSC_RXRE_SIZE);
  788. #endif
  789. /*
  790. * Loop through Rx descriptors handling ones that have been completed.
  791. */
  792. while (!((cmdstat = be32_to_cpu(rxre->cmdstat))
  793. & SDMA_DESC_CMDSTAT_O)) {
  794. bytes_in = be16_to_cpu(rxre->bytecnt);
  795. #ifdef CONFIG_CONSOLE_POLL
  796. if (unlikely(serial_polled)) {
  797. serial_polled = 0;
  798. return 0;
  799. }
  800. #endif
  801. /* Following use of tty struct directly is deprecated */
  802. if (unlikely(tty_buffer_request_room(tty, bytes_in)
  803. < bytes_in)) {
  804. if (tty->low_latency)
  805. tty_flip_buffer_push(tty);
  806. /*
  807. * If this failed then we will throw away the bytes
  808. * but must do so to clear interrupts.
  809. */
  810. }
  811. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  812. dma_cache_sync(pi->port.dev, (void *)bp, MPSC_RXBE_SIZE,
  813. DMA_FROM_DEVICE);
  814. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  815. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  816. invalidate_dcache_range((ulong)bp,
  817. (ulong)bp + MPSC_RXBE_SIZE);
  818. #endif
  819. /*
  820. * Other than for parity error, the manual provides little
  821. * info on what data will be in a frame flagged by any of
  822. * these errors. For parity error, it is the last byte in
  823. * the buffer that had the error. As for the rest, I guess
  824. * we'll assume there is no data in the buffer.
  825. * If there is...it gets lost.
  826. */
  827. if (unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR
  828. | SDMA_DESC_CMDSTAT_FR
  829. | SDMA_DESC_CMDSTAT_OR))) {
  830. pi->port.icount.rx++;
  831. if (cmdstat & SDMA_DESC_CMDSTAT_BR) { /* Break */
  832. pi->port.icount.brk++;
  833. if (uart_handle_break(&pi->port))
  834. goto next_frame;
  835. } else if (cmdstat & SDMA_DESC_CMDSTAT_FR) {
  836. pi->port.icount.frame++;
  837. } else if (cmdstat & SDMA_DESC_CMDSTAT_OR) {
  838. pi->port.icount.overrun++;
  839. }
  840. cmdstat &= pi->port.read_status_mask;
  841. if (cmdstat & SDMA_DESC_CMDSTAT_BR)
  842. flag = TTY_BREAK;
  843. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)
  844. flag = TTY_FRAME;
  845. else if (cmdstat & SDMA_DESC_CMDSTAT_OR)
  846. flag = TTY_OVERRUN;
  847. else if (cmdstat & SDMA_DESC_CMDSTAT_PE)
  848. flag = TTY_PARITY;
  849. }
  850. if (uart_handle_sysrq_char(&pi->port, *bp)) {
  851. bp++;
  852. bytes_in--;
  853. #ifdef CONFIG_CONSOLE_POLL
  854. if (unlikely(serial_polled)) {
  855. serial_polled = 0;
  856. return 0;
  857. }
  858. #endif
  859. goto next_frame;
  860. }
  861. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR
  862. | SDMA_DESC_CMDSTAT_FR
  863. | SDMA_DESC_CMDSTAT_OR)))
  864. && !(cmdstat & pi->port.ignore_status_mask)) {
  865. tty_insert_flip_char(tty, *bp, flag);
  866. } else {
  867. for (i=0; i<bytes_in; i++)
  868. tty_insert_flip_char(tty, *bp++, TTY_NORMAL);
  869. pi->port.icount.rx += bytes_in;
  870. }
  871. next_frame:
  872. rxre->bytecnt = cpu_to_be16(0);
  873. wmb();
  874. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O
  875. | SDMA_DESC_CMDSTAT_EI | SDMA_DESC_CMDSTAT_F
  876. | SDMA_DESC_CMDSTAT_L);
  877. wmb();
  878. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE,
  879. DMA_BIDIRECTIONAL);
  880. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  881. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  882. flush_dcache_range((ulong)rxre,
  883. (ulong)rxre + MPSC_RXRE_SIZE);
  884. #endif
  885. /* Advance to next descriptor */
  886. pi->rxr_posn = (pi->rxr_posn + 1) & (MPSC_RXR_ENTRIES - 1);
  887. rxre = (struct mpsc_rx_desc *)
  888. (pi->rxr + (pi->rxr_posn * MPSC_RXRE_SIZE));
  889. dma_cache_sync(pi->port.dev, (void *)rxre, MPSC_RXRE_SIZE,
  890. DMA_FROM_DEVICE);
  891. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  892. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  893. invalidate_dcache_range((ulong)rxre,
  894. (ulong)rxre + MPSC_RXRE_SIZE);
  895. #endif
  896. rc = 1;
  897. }
  898. /* Restart rx engine, if its stopped */
  899. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  900. mpsc_start_rx(pi);
  901. tty_flip_buffer_push(tty);
  902. return rc;
  903. }
  904. static void mpsc_setup_tx_desc(struct mpsc_port_info *pi, u32 count, u32 intr)
  905. {
  906. struct mpsc_tx_desc *txre;
  907. txre = (struct mpsc_tx_desc *)(pi->txr
  908. + (pi->txr_head * MPSC_TXRE_SIZE));
  909. txre->bytecnt = cpu_to_be16(count);
  910. txre->shadow = txre->bytecnt;
  911. wmb(); /* ensure cmdstat is last field updated */
  912. txre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O | SDMA_DESC_CMDSTAT_F
  913. | SDMA_DESC_CMDSTAT_L
  914. | ((intr) ? SDMA_DESC_CMDSTAT_EI : 0));
  915. wmb();
  916. dma_cache_sync(pi->port.dev, (void *)txre, MPSC_TXRE_SIZE,
  917. DMA_BIDIRECTIONAL);
  918. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  919. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  920. flush_dcache_range((ulong)txre,
  921. (ulong)txre + MPSC_TXRE_SIZE);
  922. #endif
  923. }
  924. static void mpsc_copy_tx_data(struct mpsc_port_info *pi)
  925. {
  926. struct circ_buf *xmit = &pi->port.state->xmit;
  927. u8 *bp;
  928. u32 i;
  929. /* Make sure the desc ring isn't full */
  930. while (CIRC_CNT(pi->txr_head, pi->txr_tail, MPSC_TXR_ENTRIES)
  931. < (MPSC_TXR_ENTRIES - 1)) {
  932. if (pi->port.x_char) {
  933. /*
  934. * Ideally, we should use the TCS field in
  935. * CHR_1 to put the x_char out immediately but
  936. * errata prevents us from being able to read
  937. * CHR_2 to know that its safe to write to
  938. * CHR_1. Instead, just put it in-band with
  939. * all the other Tx data.
  940. */
  941. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  942. *bp = pi->port.x_char;
  943. pi->port.x_char = 0;
  944. i = 1;
  945. } else if (!uart_circ_empty(xmit)
  946. && !uart_tx_stopped(&pi->port)) {
  947. i = min((u32)MPSC_TXBE_SIZE,
  948. (u32)uart_circ_chars_pending(xmit));
  949. i = min(i, (u32)CIRC_CNT_TO_END(xmit->head, xmit->tail,
  950. UART_XMIT_SIZE));
  951. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  952. memcpy(bp, &xmit->buf[xmit->tail], i);
  953. xmit->tail = (xmit->tail + i) & (UART_XMIT_SIZE - 1);
  954. if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
  955. uart_write_wakeup(&pi->port);
  956. } else { /* All tx data copied into ring bufs */
  957. return;
  958. }
  959. dma_cache_sync(pi->port.dev, (void *)bp, MPSC_TXBE_SIZE,
  960. DMA_BIDIRECTIONAL);
  961. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  962. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  963. flush_dcache_range((ulong)bp,
  964. (ulong)bp + MPSC_TXBE_SIZE);
  965. #endif
  966. mpsc_setup_tx_desc(pi, i, 1);
  967. /* Advance to next descriptor */
  968. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  969. }
  970. }
  971. static int mpsc_tx_intr(struct mpsc_port_info *pi)
  972. {
  973. struct mpsc_tx_desc *txre;
  974. int rc = 0;
  975. unsigned long iflags;
  976. spin_lock_irqsave(&pi->tx_lock, iflags);
  977. if (!mpsc_sdma_tx_active(pi)) {
  978. txre = (struct mpsc_tx_desc *)(pi->txr
  979. + (pi->txr_tail * MPSC_TXRE_SIZE));
  980. dma_cache_sync(pi->port.dev, (void *)txre, MPSC_TXRE_SIZE,
  981. DMA_FROM_DEVICE);
  982. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  983. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  984. invalidate_dcache_range((ulong)txre,
  985. (ulong)txre + MPSC_TXRE_SIZE);
  986. #endif
  987. while (!(be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O)) {
  988. rc = 1;
  989. pi->port.icount.tx += be16_to_cpu(txre->bytecnt);
  990. pi->txr_tail = (pi->txr_tail+1) & (MPSC_TXR_ENTRIES-1);
  991. /* If no more data to tx, fall out of loop */
  992. if (pi->txr_head == pi->txr_tail)
  993. break;
  994. txre = (struct mpsc_tx_desc *)(pi->txr
  995. + (pi->txr_tail * MPSC_TXRE_SIZE));
  996. dma_cache_sync(pi->port.dev, (void *)txre,
  997. MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  998. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  999. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1000. invalidate_dcache_range((ulong)txre,
  1001. (ulong)txre + MPSC_TXRE_SIZE);
  1002. #endif
  1003. }
  1004. mpsc_copy_tx_data(pi);
  1005. mpsc_sdma_start_tx(pi); /* start next desc if ready */
  1006. }
  1007. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1008. return rc;
  1009. }
  1010. /*
  1011. * This is the driver's interrupt handler. To avoid a race, we first clear
  1012. * the interrupt, then handle any completed Rx/Tx descriptors. When done
  1013. * handling those descriptors, we restart the Rx/Tx engines if they're stopped.
  1014. */
  1015. static irqreturn_t mpsc_sdma_intr(int irq, void *dev_id)
  1016. {
  1017. struct mpsc_port_info *pi = dev_id;
  1018. ulong iflags;
  1019. int rc = IRQ_NONE;
  1020. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Received\n",pi->port.line);
  1021. spin_lock_irqsave(&pi->port.lock, iflags);
  1022. mpsc_sdma_intr_ack(pi);
  1023. if (mpsc_rx_intr(pi))
  1024. rc = IRQ_HANDLED;
  1025. if (mpsc_tx_intr(pi))
  1026. rc = IRQ_HANDLED;
  1027. spin_unlock_irqrestore(&pi->port.lock, iflags);
  1028. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Handled\n", pi->port.line);
  1029. return rc;
  1030. }
  1031. /*
  1032. ******************************************************************************
  1033. *
  1034. * serial_core.c Interface routines
  1035. *
  1036. ******************************************************************************
  1037. */
  1038. static uint mpsc_tx_empty(struct uart_port *port)
  1039. {
  1040. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1041. ulong iflags;
  1042. uint rc;
  1043. spin_lock_irqsave(&pi->port.lock, iflags);
  1044. rc = mpsc_sdma_tx_active(pi) ? 0 : TIOCSER_TEMT;
  1045. spin_unlock_irqrestore(&pi->port.lock, iflags);
  1046. return rc;
  1047. }
  1048. static void mpsc_set_mctrl(struct uart_port *port, uint mctrl)
  1049. {
  1050. /* Have no way to set modem control lines AFAICT */
  1051. }
  1052. static uint mpsc_get_mctrl(struct uart_port *port)
  1053. {
  1054. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1055. u32 mflags, status;
  1056. status = (pi->mirror_regs) ? pi->MPSC_CHR_10_m
  1057. : readl(pi->mpsc_base + MPSC_CHR_10);
  1058. mflags = 0;
  1059. if (status & 0x1)
  1060. mflags |= TIOCM_CTS;
  1061. if (status & 0x2)
  1062. mflags |= TIOCM_CAR;
  1063. return mflags | TIOCM_DSR; /* No way to tell if DSR asserted */
  1064. }
  1065. static void mpsc_stop_tx(struct uart_port *port)
  1066. {
  1067. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1068. pr_debug("mpsc_stop_tx[%d]\n", port->line);
  1069. mpsc_freeze(pi);
  1070. }
  1071. static void mpsc_start_tx(struct uart_port *port)
  1072. {
  1073. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1074. unsigned long iflags;
  1075. spin_lock_irqsave(&pi->tx_lock, iflags);
  1076. mpsc_unfreeze(pi);
  1077. mpsc_copy_tx_data(pi);
  1078. mpsc_sdma_start_tx(pi);
  1079. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1080. pr_debug("mpsc_start_tx[%d]\n", port->line);
  1081. }
  1082. static void mpsc_start_rx(struct mpsc_port_info *pi)
  1083. {
  1084. pr_debug("mpsc_start_rx[%d]: Starting...\n", pi->port.line);
  1085. if (pi->rcv_data) {
  1086. mpsc_enter_hunt(pi);
  1087. mpsc_sdma_cmd(pi, SDMA_SDCM_ERD);
  1088. }
  1089. }
  1090. static void mpsc_stop_rx(struct uart_port *port)
  1091. {
  1092. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1093. pr_debug("mpsc_stop_rx[%d]: Stopping...\n", port->line);
  1094. if (pi->mirror_regs) {
  1095. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_RA,
  1096. pi->mpsc_base + MPSC_CHR_2);
  1097. /* Erratum prevents reading CHR_2 so just delay for a while */
  1098. udelay(100);
  1099. } else {
  1100. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_RA,
  1101. pi->mpsc_base + MPSC_CHR_2);
  1102. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_RA)
  1103. udelay(10);
  1104. }
  1105. mpsc_sdma_cmd(pi, SDMA_SDCM_AR);
  1106. }
  1107. static void mpsc_enable_ms(struct uart_port *port)
  1108. {
  1109. }
  1110. static void mpsc_break_ctl(struct uart_port *port, int ctl)
  1111. {
  1112. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1113. ulong flags;
  1114. u32 v;
  1115. v = ctl ? 0x00ff0000 : 0;
  1116. spin_lock_irqsave(&pi->port.lock, flags);
  1117. if (pi->mirror_regs)
  1118. pi->MPSC_CHR_1_m = v;
  1119. writel(v, pi->mpsc_base + MPSC_CHR_1);
  1120. spin_unlock_irqrestore(&pi->port.lock, flags);
  1121. }
  1122. static int mpsc_startup(struct uart_port *port)
  1123. {
  1124. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1125. u32 flag = 0;
  1126. int rc;
  1127. pr_debug("mpsc_startup[%d]: Starting up MPSC, irq: %d\n",
  1128. port->line, pi->port.irq);
  1129. if ((rc = mpsc_make_ready(pi)) == 0) {
  1130. /* Setup IRQ handler */
  1131. mpsc_sdma_intr_ack(pi);
  1132. /* If irq's are shared, need to set flag */
  1133. if (mpsc_ports[0].port.irq == mpsc_ports[1].port.irq)
  1134. flag = IRQF_SHARED;
  1135. if (request_irq(pi->port.irq, mpsc_sdma_intr, flag,
  1136. "mpsc-sdma", pi))
  1137. printk(KERN_ERR "MPSC: Can't get SDMA IRQ %d\n",
  1138. pi->port.irq);
  1139. mpsc_sdma_intr_unmask(pi, 0xf);
  1140. mpsc_sdma_set_rx_ring(pi, (struct mpsc_rx_desc *)(pi->rxr_p
  1141. + (pi->rxr_posn * MPSC_RXRE_SIZE)));
  1142. }
  1143. return rc;
  1144. }
  1145. static void mpsc_shutdown(struct uart_port *port)
  1146. {
  1147. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1148. pr_debug("mpsc_shutdown[%d]: Shutting down MPSC\n", port->line);
  1149. mpsc_sdma_stop(pi);
  1150. free_irq(pi->port.irq, pi);
  1151. }
  1152. static void mpsc_set_termios(struct uart_port *port, struct ktermios *termios,
  1153. struct ktermios *old)
  1154. {
  1155. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1156. u32 baud;
  1157. ulong flags;
  1158. u32 chr_bits, stop_bits, par;
  1159. pi->c_iflag = termios->c_iflag;
  1160. pi->c_cflag = termios->c_cflag;
  1161. switch (termios->c_cflag & CSIZE) {
  1162. case CS5:
  1163. chr_bits = MPSC_MPCR_CL_5;
  1164. break;
  1165. case CS6:
  1166. chr_bits = MPSC_MPCR_CL_6;
  1167. break;
  1168. case CS7:
  1169. chr_bits = MPSC_MPCR_CL_7;
  1170. break;
  1171. case CS8:
  1172. default:
  1173. chr_bits = MPSC_MPCR_CL_8;
  1174. break;
  1175. }
  1176. if (termios->c_cflag & CSTOPB)
  1177. stop_bits = MPSC_MPCR_SBL_2;
  1178. else
  1179. stop_bits = MPSC_MPCR_SBL_1;
  1180. par = MPSC_CHR_2_PAR_EVEN;
  1181. if (termios->c_cflag & PARENB)
  1182. if (termios->c_cflag & PARODD)
  1183. par = MPSC_CHR_2_PAR_ODD;
  1184. #ifdef CMSPAR
  1185. if (termios->c_cflag & CMSPAR) {
  1186. if (termios->c_cflag & PARODD)
  1187. par = MPSC_CHR_2_PAR_MARK;
  1188. else
  1189. par = MPSC_CHR_2_PAR_SPACE;
  1190. }
  1191. #endif
  1192. baud = uart_get_baud_rate(port, termios, old, 0, port->uartclk);
  1193. spin_lock_irqsave(&pi->port.lock, flags);
  1194. uart_update_timeout(port, termios->c_cflag, baud);
  1195. mpsc_set_char_length(pi, chr_bits);
  1196. mpsc_set_stop_bit_length(pi, stop_bits);
  1197. mpsc_set_parity(pi, par);
  1198. mpsc_set_baudrate(pi, baud);
  1199. /* Characters/events to read */
  1200. pi->port.read_status_mask = SDMA_DESC_CMDSTAT_OR;
  1201. if (termios->c_iflag & INPCK)
  1202. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_PE
  1203. | SDMA_DESC_CMDSTAT_FR;
  1204. if (termios->c_iflag & (BRKINT | PARMRK))
  1205. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1206. /* Characters/events to ignore */
  1207. pi->port.ignore_status_mask = 0;
  1208. if (termios->c_iflag & IGNPAR)
  1209. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_PE
  1210. | SDMA_DESC_CMDSTAT_FR;
  1211. if (termios->c_iflag & IGNBRK) {
  1212. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1213. if (termios->c_iflag & IGNPAR)
  1214. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_OR;
  1215. }
  1216. if ((termios->c_cflag & CREAD)) {
  1217. if (!pi->rcv_data) {
  1218. pi->rcv_data = 1;
  1219. mpsc_start_rx(pi);
  1220. }
  1221. } else if (pi->rcv_data) {
  1222. mpsc_stop_rx(port);
  1223. pi->rcv_data = 0;
  1224. }
  1225. spin_unlock_irqrestore(&pi->port.lock, flags);
  1226. }
  1227. static const char *mpsc_type(struct uart_port *port)
  1228. {
  1229. pr_debug("mpsc_type[%d]: port type: %s\n", port->line,MPSC_DRIVER_NAME);
  1230. return MPSC_DRIVER_NAME;
  1231. }
  1232. static int mpsc_request_port(struct uart_port *port)
  1233. {
  1234. /* Should make chip/platform specific call */
  1235. return 0;
  1236. }
  1237. static void mpsc_release_port(struct uart_port *port)
  1238. {
  1239. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1240. if (pi->ready) {
  1241. mpsc_uninit_rings(pi);
  1242. mpsc_free_ring_mem(pi);
  1243. pi->ready = 0;
  1244. }
  1245. }
  1246. static void mpsc_config_port(struct uart_port *port, int flags)
  1247. {
  1248. }
  1249. static int mpsc_verify_port(struct uart_port *port, struct serial_struct *ser)
  1250. {
  1251. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1252. int rc = 0;
  1253. pr_debug("mpsc_verify_port[%d]: Verifying port data\n", pi->port.line);
  1254. if (ser->type != PORT_UNKNOWN && ser->type != PORT_MPSC)
  1255. rc = -EINVAL;
  1256. else if (pi->port.irq != ser->irq)
  1257. rc = -EINVAL;
  1258. else if (ser->io_type != SERIAL_IO_MEM)
  1259. rc = -EINVAL;
  1260. else if (pi->port.uartclk / 16 != ser->baud_base) /* Not sure */
  1261. rc = -EINVAL;
  1262. else if ((void *)pi->port.mapbase != ser->iomem_base)
  1263. rc = -EINVAL;
  1264. else if (pi->port.iobase != ser->port)
  1265. rc = -EINVAL;
  1266. else if (ser->hub6 != 0)
  1267. rc = -EINVAL;
  1268. return rc;
  1269. }
  1270. #ifdef CONFIG_CONSOLE_POLL
  1271. /* Serial polling routines for writing and reading from the uart while
  1272. * in an interrupt or debug context.
  1273. */
  1274. static char poll_buf[2048];
  1275. static int poll_ptr;
  1276. static int poll_cnt;
  1277. static void mpsc_put_poll_char(struct uart_port *port,
  1278. unsigned char c);
  1279. static int mpsc_get_poll_char(struct uart_port *port)
  1280. {
  1281. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1282. struct mpsc_rx_desc *rxre;
  1283. u32 cmdstat, bytes_in, i;
  1284. u8 *bp;
  1285. if (!serial_polled)
  1286. serial_polled = 1;
  1287. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  1288. if (poll_cnt) {
  1289. poll_cnt--;
  1290. return poll_buf[poll_ptr++];
  1291. }
  1292. poll_ptr = 0;
  1293. poll_cnt = 0;
  1294. while (poll_cnt == 0) {
  1295. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  1296. (pi->rxr_posn*MPSC_RXRE_SIZE));
  1297. dma_cache_sync(pi->port.dev, (void *)rxre,
  1298. MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  1299. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1300. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1301. invalidate_dcache_range((ulong)rxre,
  1302. (ulong)rxre + MPSC_RXRE_SIZE);
  1303. #endif
  1304. /*
  1305. * Loop through Rx descriptors handling ones that have
  1306. * been completed.
  1307. */
  1308. while (poll_cnt == 0 &&
  1309. !((cmdstat = be32_to_cpu(rxre->cmdstat)) &
  1310. SDMA_DESC_CMDSTAT_O)){
  1311. bytes_in = be16_to_cpu(rxre->bytecnt);
  1312. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  1313. dma_cache_sync(pi->port.dev, (void *) bp,
  1314. MPSC_RXBE_SIZE, DMA_FROM_DEVICE);
  1315. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1316. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1317. invalidate_dcache_range((ulong)bp,
  1318. (ulong)bp + MPSC_RXBE_SIZE);
  1319. #endif
  1320. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  1321. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) &&
  1322. !(cmdstat & pi->port.ignore_status_mask)) {
  1323. poll_buf[poll_cnt] = *bp;
  1324. poll_cnt++;
  1325. } else {
  1326. for (i = 0; i < bytes_in; i++) {
  1327. poll_buf[poll_cnt] = *bp++;
  1328. poll_cnt++;
  1329. }
  1330. pi->port.icount.rx += bytes_in;
  1331. }
  1332. rxre->bytecnt = cpu_to_be16(0);
  1333. wmb();
  1334. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  1335. SDMA_DESC_CMDSTAT_EI |
  1336. SDMA_DESC_CMDSTAT_F |
  1337. SDMA_DESC_CMDSTAT_L);
  1338. wmb();
  1339. dma_cache_sync(pi->port.dev, (void *)rxre,
  1340. MPSC_RXRE_SIZE, DMA_BIDIRECTIONAL);
  1341. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1342. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1343. flush_dcache_range((ulong)rxre,
  1344. (ulong)rxre + MPSC_RXRE_SIZE);
  1345. #endif
  1346. /* Advance to next descriptor */
  1347. pi->rxr_posn = (pi->rxr_posn + 1) &
  1348. (MPSC_RXR_ENTRIES - 1);
  1349. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  1350. (pi->rxr_posn * MPSC_RXRE_SIZE));
  1351. dma_cache_sync(pi->port.dev, (void *)rxre,
  1352. MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  1353. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1354. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1355. invalidate_dcache_range((ulong)rxre,
  1356. (ulong)rxre + MPSC_RXRE_SIZE);
  1357. #endif
  1358. }
  1359. /* Restart rx engine, if its stopped */
  1360. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  1361. mpsc_start_rx(pi);
  1362. }
  1363. if (poll_cnt) {
  1364. poll_cnt--;
  1365. return poll_buf[poll_ptr++];
  1366. }
  1367. return 0;
  1368. }
  1369. static void mpsc_put_poll_char(struct uart_port *port,
  1370. unsigned char c)
  1371. {
  1372. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1373. u32 data;
  1374. data = readl(pi->mpsc_base + MPSC_MPCR);
  1375. writeb(c, pi->mpsc_base + MPSC_CHR_1);
  1376. mb();
  1377. data = readl(pi->mpsc_base + MPSC_CHR_2);
  1378. data |= MPSC_CHR_2_TTCS;
  1379. writel(data, pi->mpsc_base + MPSC_CHR_2);
  1380. mb();
  1381. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_TTCS);
  1382. }
  1383. #endif
  1384. static struct uart_ops mpsc_pops = {
  1385. .tx_empty = mpsc_tx_empty,
  1386. .set_mctrl = mpsc_set_mctrl,
  1387. .get_mctrl = mpsc_get_mctrl,
  1388. .stop_tx = mpsc_stop_tx,
  1389. .start_tx = mpsc_start_tx,
  1390. .stop_rx = mpsc_stop_rx,
  1391. .enable_ms = mpsc_enable_ms,
  1392. .break_ctl = mpsc_break_ctl,
  1393. .startup = mpsc_startup,
  1394. .shutdown = mpsc_shutdown,
  1395. .set_termios = mpsc_set_termios,
  1396. .type = mpsc_type,
  1397. .release_port = mpsc_release_port,
  1398. .request_port = mpsc_request_port,
  1399. .config_port = mpsc_config_port,
  1400. .verify_port = mpsc_verify_port,
  1401. #ifdef CONFIG_CONSOLE_POLL
  1402. .poll_get_char = mpsc_get_poll_char,
  1403. .poll_put_char = mpsc_put_poll_char,
  1404. #endif
  1405. };
  1406. /*
  1407. ******************************************************************************
  1408. *
  1409. * Console Interface Routines
  1410. *
  1411. ******************************************************************************
  1412. */
  1413. #ifdef CONFIG_SERIAL_MPSC_CONSOLE
  1414. static void mpsc_console_write(struct console *co, const char *s, uint count)
  1415. {
  1416. struct mpsc_port_info *pi = &mpsc_ports[co->index];
  1417. u8 *bp, *dp, add_cr = 0;
  1418. int i;
  1419. unsigned long iflags;
  1420. spin_lock_irqsave(&pi->tx_lock, iflags);
  1421. while (pi->txr_head != pi->txr_tail) {
  1422. while (mpsc_sdma_tx_active(pi))
  1423. udelay(100);
  1424. mpsc_sdma_intr_ack(pi);
  1425. mpsc_tx_intr(pi);
  1426. }
  1427. while (mpsc_sdma_tx_active(pi))
  1428. udelay(100);
  1429. while (count > 0) {
  1430. bp = dp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  1431. for (i = 0; i < MPSC_TXBE_SIZE; i++) {
  1432. if (count == 0)
  1433. break;
  1434. if (add_cr) {
  1435. *(dp++) = '\r';
  1436. add_cr = 0;
  1437. } else {
  1438. *(dp++) = *s;
  1439. if (*(s++) == '\n') { /* add '\r' after '\n' */
  1440. add_cr = 1;
  1441. count++;
  1442. }
  1443. }
  1444. count--;
  1445. }
  1446. dma_cache_sync(pi->port.dev, (void *)bp, MPSC_TXBE_SIZE,
  1447. DMA_BIDIRECTIONAL);
  1448. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1449. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1450. flush_dcache_range((ulong)bp,
  1451. (ulong)bp + MPSC_TXBE_SIZE);
  1452. #endif
  1453. mpsc_setup_tx_desc(pi, i, 0);
  1454. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1455. mpsc_sdma_start_tx(pi);
  1456. while (mpsc_sdma_tx_active(pi))
  1457. udelay(100);
  1458. pi->txr_tail = (pi->txr_tail + 1) & (MPSC_TXR_ENTRIES - 1);
  1459. }
  1460. spin_unlock_irqrestore(&pi->tx_lock, iflags);
  1461. }
  1462. static int __init mpsc_console_setup(struct console *co, char *options)
  1463. {
  1464. struct mpsc_port_info *pi;
  1465. int baud, bits, parity, flow;
  1466. pr_debug("mpsc_console_setup[%d]: options: %s\n", co->index, options);
  1467. if (co->index >= MPSC_NUM_CTLRS)
  1468. co->index = 0;
  1469. pi = &mpsc_ports[co->index];
  1470. baud = pi->default_baud;
  1471. bits = pi->default_bits;
  1472. parity = pi->default_parity;
  1473. flow = pi->default_flow;
  1474. if (!pi->port.ops)
  1475. return -ENODEV;
  1476. spin_lock_init(&pi->port.lock); /* Temporary fix--copied from 8250.c */
  1477. if (options)
  1478. uart_parse_options(options, &baud, &parity, &bits, &flow);
  1479. return uart_set_options(&pi->port, co, baud, parity, bits, flow);
  1480. }
  1481. static struct console mpsc_console = {
  1482. .name = MPSC_DEV_NAME,
  1483. .write = mpsc_console_write,
  1484. .device = uart_console_device,
  1485. .setup = mpsc_console_setup,
  1486. .flags = CON_PRINTBUFFER,
  1487. .index = -1,
  1488. .data = &mpsc_reg,
  1489. };
  1490. static int __init mpsc_late_console_init(void)
  1491. {
  1492. pr_debug("mpsc_late_console_init: Enter\n");
  1493. if (!(mpsc_console.flags & CON_ENABLED))
  1494. register_console(&mpsc_console);
  1495. return 0;
  1496. }
  1497. late_initcall(mpsc_late_console_init);
  1498. #define MPSC_CONSOLE &mpsc_console
  1499. #else
  1500. #define MPSC_CONSOLE NULL
  1501. #endif
  1502. /*
  1503. ******************************************************************************
  1504. *
  1505. * Dummy Platform Driver to extract & map shared register regions
  1506. *
  1507. ******************************************************************************
  1508. */
  1509. static void mpsc_resource_err(char *s)
  1510. {
  1511. printk(KERN_WARNING "MPSC: Platform device resource error in %s\n", s);
  1512. }
  1513. static int mpsc_shared_map_regs(struct platform_device *pd)
  1514. {
  1515. struct resource *r;
  1516. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1517. MPSC_ROUTING_BASE_ORDER))
  1518. && request_mem_region(r->start,
  1519. MPSC_ROUTING_REG_BLOCK_SIZE,
  1520. "mpsc_routing_regs")) {
  1521. mpsc_shared_regs.mpsc_routing_base = ioremap(r->start,
  1522. MPSC_ROUTING_REG_BLOCK_SIZE);
  1523. mpsc_shared_regs.mpsc_routing_base_p = r->start;
  1524. } else {
  1525. mpsc_resource_err("MPSC routing base");
  1526. return -ENOMEM;
  1527. }
  1528. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1529. MPSC_SDMA_INTR_BASE_ORDER))
  1530. && request_mem_region(r->start,
  1531. MPSC_SDMA_INTR_REG_BLOCK_SIZE,
  1532. "sdma_intr_regs")) {
  1533. mpsc_shared_regs.sdma_intr_base = ioremap(r->start,
  1534. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1535. mpsc_shared_regs.sdma_intr_base_p = r->start;
  1536. } else {
  1537. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1538. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1539. MPSC_ROUTING_REG_BLOCK_SIZE);
  1540. mpsc_resource_err("SDMA intr base");
  1541. return -ENOMEM;
  1542. }
  1543. return 0;
  1544. }
  1545. static void mpsc_shared_unmap_regs(void)
  1546. {
  1547. if (!mpsc_shared_regs.mpsc_routing_base) {
  1548. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1549. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1550. MPSC_ROUTING_REG_BLOCK_SIZE);
  1551. }
  1552. if (!mpsc_shared_regs.sdma_intr_base) {
  1553. iounmap(mpsc_shared_regs.sdma_intr_base);
  1554. release_mem_region(mpsc_shared_regs.sdma_intr_base_p,
  1555. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1556. }
  1557. mpsc_shared_regs.mpsc_routing_base = NULL;
  1558. mpsc_shared_regs.sdma_intr_base = NULL;
  1559. mpsc_shared_regs.mpsc_routing_base_p = 0;
  1560. mpsc_shared_regs.sdma_intr_base_p = 0;
  1561. }
  1562. static int mpsc_shared_drv_probe(struct platform_device *dev)
  1563. {
  1564. struct mpsc_shared_pdata *pdata;
  1565. int rc = -ENODEV;
  1566. if (dev->id == 0) {
  1567. if (!(rc = mpsc_shared_map_regs(dev))) {
  1568. pdata = (struct mpsc_shared_pdata *)
  1569. dev->dev.platform_data;
  1570. mpsc_shared_regs.MPSC_MRR_m = pdata->mrr_val;
  1571. mpsc_shared_regs.MPSC_RCRR_m= pdata->rcrr_val;
  1572. mpsc_shared_regs.MPSC_TCRR_m= pdata->tcrr_val;
  1573. mpsc_shared_regs.SDMA_INTR_CAUSE_m =
  1574. pdata->intr_cause_val;
  1575. mpsc_shared_regs.SDMA_INTR_MASK_m =
  1576. pdata->intr_mask_val;
  1577. rc = 0;
  1578. }
  1579. }
  1580. return rc;
  1581. }
  1582. static int mpsc_shared_drv_remove(struct platform_device *dev)
  1583. {
  1584. int rc = -ENODEV;
  1585. if (dev->id == 0) {
  1586. mpsc_shared_unmap_regs();
  1587. mpsc_shared_regs.MPSC_MRR_m = 0;
  1588. mpsc_shared_regs.MPSC_RCRR_m = 0;
  1589. mpsc_shared_regs.MPSC_TCRR_m = 0;
  1590. mpsc_shared_regs.SDMA_INTR_CAUSE_m = 0;
  1591. mpsc_shared_regs.SDMA_INTR_MASK_m = 0;
  1592. rc = 0;
  1593. }
  1594. return rc;
  1595. }
  1596. static struct platform_driver mpsc_shared_driver = {
  1597. .probe = mpsc_shared_drv_probe,
  1598. .remove = mpsc_shared_drv_remove,
  1599. .driver = {
  1600. .name = MPSC_SHARED_NAME,
  1601. },
  1602. };
  1603. /*
  1604. ******************************************************************************
  1605. *
  1606. * Driver Interface Routines
  1607. *
  1608. ******************************************************************************
  1609. */
  1610. static struct uart_driver mpsc_reg = {
  1611. .owner = THIS_MODULE,
  1612. .driver_name = MPSC_DRIVER_NAME,
  1613. .dev_name = MPSC_DEV_NAME,
  1614. .major = MPSC_MAJOR,
  1615. .minor = MPSC_MINOR_START,
  1616. .nr = MPSC_NUM_CTLRS,
  1617. .cons = MPSC_CONSOLE,
  1618. };
  1619. static int mpsc_drv_map_regs(struct mpsc_port_info *pi,
  1620. struct platform_device *pd)
  1621. {
  1622. struct resource *r;
  1623. if ((r = platform_get_resource(pd, IORESOURCE_MEM, MPSC_BASE_ORDER))
  1624. && request_mem_region(r->start, MPSC_REG_BLOCK_SIZE,
  1625. "mpsc_regs")) {
  1626. pi->mpsc_base = ioremap(r->start, MPSC_REG_BLOCK_SIZE);
  1627. pi->mpsc_base_p = r->start;
  1628. } else {
  1629. mpsc_resource_err("MPSC base");
  1630. goto err;
  1631. }
  1632. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1633. MPSC_SDMA_BASE_ORDER))
  1634. && request_mem_region(r->start,
  1635. MPSC_SDMA_REG_BLOCK_SIZE, "sdma_regs")) {
  1636. pi->sdma_base = ioremap(r->start,MPSC_SDMA_REG_BLOCK_SIZE);
  1637. pi->sdma_base_p = r->start;
  1638. } else {
  1639. mpsc_resource_err("SDMA base");
  1640. if (pi->mpsc_base) {
  1641. iounmap(pi->mpsc_base);
  1642. pi->mpsc_base = NULL;
  1643. }
  1644. goto err;
  1645. }
  1646. if ((r = platform_get_resource(pd,IORESOURCE_MEM,MPSC_BRG_BASE_ORDER))
  1647. && request_mem_region(r->start,
  1648. MPSC_BRG_REG_BLOCK_SIZE, "brg_regs")) {
  1649. pi->brg_base = ioremap(r->start, MPSC_BRG_REG_BLOCK_SIZE);
  1650. pi->brg_base_p = r->start;
  1651. } else {
  1652. mpsc_resource_err("BRG base");
  1653. if (pi->mpsc_base) {
  1654. iounmap(pi->mpsc_base);
  1655. pi->mpsc_base = NULL;
  1656. }
  1657. if (pi->sdma_base) {
  1658. iounmap(pi->sdma_base);
  1659. pi->sdma_base = NULL;
  1660. }
  1661. goto err;
  1662. }
  1663. return 0;
  1664. err:
  1665. return -ENOMEM;
  1666. }
  1667. static void mpsc_drv_unmap_regs(struct mpsc_port_info *pi)
  1668. {
  1669. if (!pi->mpsc_base) {
  1670. iounmap(pi->mpsc_base);
  1671. release_mem_region(pi->mpsc_base_p, MPSC_REG_BLOCK_SIZE);
  1672. }
  1673. if (!pi->sdma_base) {
  1674. iounmap(pi->sdma_base);
  1675. release_mem_region(pi->sdma_base_p, MPSC_SDMA_REG_BLOCK_SIZE);
  1676. }
  1677. if (!pi->brg_base) {
  1678. iounmap(pi->brg_base);
  1679. release_mem_region(pi->brg_base_p, MPSC_BRG_REG_BLOCK_SIZE);
  1680. }
  1681. pi->mpsc_base = NULL;
  1682. pi->sdma_base = NULL;
  1683. pi->brg_base = NULL;
  1684. pi->mpsc_base_p = 0;
  1685. pi->sdma_base_p = 0;
  1686. pi->brg_base_p = 0;
  1687. }
  1688. static void mpsc_drv_get_platform_data(struct mpsc_port_info *pi,
  1689. struct platform_device *pd, int num)
  1690. {
  1691. struct mpsc_pdata *pdata;
  1692. pdata = (struct mpsc_pdata *)pd->dev.platform_data;
  1693. pi->port.uartclk = pdata->brg_clk_freq;
  1694. pi->port.iotype = UPIO_MEM;
  1695. pi->port.line = num;
  1696. pi->port.type = PORT_MPSC;
  1697. pi->port.fifosize = MPSC_TXBE_SIZE;
  1698. pi->port.membase = pi->mpsc_base;
  1699. pi->port.mapbase = (ulong)pi->mpsc_base;
  1700. pi->port.ops = &mpsc_pops;
  1701. pi->mirror_regs = pdata->mirror_regs;
  1702. pi->cache_mgmt = pdata->cache_mgmt;
  1703. pi->brg_can_tune = pdata->brg_can_tune;
  1704. pi->brg_clk_src = pdata->brg_clk_src;
  1705. pi->mpsc_max_idle = pdata->max_idle;
  1706. pi->default_baud = pdata->default_baud;
  1707. pi->default_bits = pdata->default_bits;
  1708. pi->default_parity = pdata->default_parity;
  1709. pi->default_flow = pdata->default_flow;
  1710. /* Initial values of mirrored regs */
  1711. pi->MPSC_CHR_1_m = pdata->chr_1_val;
  1712. pi->MPSC_CHR_2_m = pdata->chr_2_val;
  1713. pi->MPSC_CHR_10_m = pdata->chr_10_val;
  1714. pi->MPSC_MPCR_m = pdata->mpcr_val;
  1715. pi->BRG_BCR_m = pdata->bcr_val;
  1716. pi->shared_regs = &mpsc_shared_regs;
  1717. pi->port.irq = platform_get_irq(pd, 0);
  1718. }
  1719. static int mpsc_drv_probe(struct platform_device *dev)
  1720. {
  1721. struct mpsc_port_info *pi;
  1722. int rc = -ENODEV;
  1723. pr_debug("mpsc_drv_probe: Adding MPSC %d\n", dev->id);
  1724. if (dev->id < MPSC_NUM_CTLRS) {
  1725. pi = &mpsc_ports[dev->id];
  1726. if (!(rc = mpsc_drv_map_regs(pi, dev))) {
  1727. mpsc_drv_get_platform_data(pi, dev, dev->id);
  1728. pi->port.dev = &dev->dev;
  1729. if (!(rc = mpsc_make_ready(pi))) {
  1730. spin_lock_init(&pi->tx_lock);
  1731. if (!(rc = uart_add_one_port(&mpsc_reg,
  1732. &pi->port))) {
  1733. rc = 0;
  1734. } else {
  1735. mpsc_release_port((struct uart_port *)
  1736. pi);
  1737. mpsc_drv_unmap_regs(pi);
  1738. }
  1739. } else {
  1740. mpsc_drv_unmap_regs(pi);
  1741. }
  1742. }
  1743. }
  1744. return rc;
  1745. }
  1746. static int mpsc_drv_remove(struct platform_device *dev)
  1747. {
  1748. pr_debug("mpsc_drv_exit: Removing MPSC %d\n", dev->id);
  1749. if (dev->id < MPSC_NUM_CTLRS) {
  1750. uart_remove_one_port(&mpsc_reg, &mpsc_ports[dev->id].port);
  1751. mpsc_release_port((struct uart_port *)
  1752. &mpsc_ports[dev->id].port);
  1753. mpsc_drv_unmap_regs(&mpsc_ports[dev->id]);
  1754. return 0;
  1755. } else {
  1756. return -ENODEV;
  1757. }
  1758. }
  1759. static struct platform_driver mpsc_driver = {
  1760. .probe = mpsc_drv_probe,
  1761. .remove = mpsc_drv_remove,
  1762. .driver = {
  1763. .name = MPSC_CTLR_NAME,
  1764. .owner = THIS_MODULE,
  1765. },
  1766. };
  1767. static int __init mpsc_drv_init(void)
  1768. {
  1769. int rc;
  1770. printk(KERN_INFO "Serial: MPSC driver\n");
  1771. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1772. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1773. if (!(rc = uart_register_driver(&mpsc_reg))) {
  1774. if (!(rc = platform_driver_register(&mpsc_shared_driver))) {
  1775. if ((rc = platform_driver_register(&mpsc_driver))) {
  1776. platform_driver_unregister(&mpsc_shared_driver);
  1777. uart_unregister_driver(&mpsc_reg);
  1778. }
  1779. } else {
  1780. uart_unregister_driver(&mpsc_reg);
  1781. }
  1782. }
  1783. return rc;
  1784. }
  1785. static void __exit mpsc_drv_exit(void)
  1786. {
  1787. platform_driver_unregister(&mpsc_driver);
  1788. platform_driver_unregister(&mpsc_shared_driver);
  1789. uart_unregister_driver(&mpsc_reg);
  1790. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1791. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1792. }
  1793. module_init(mpsc_drv_init);
  1794. module_exit(mpsc_drv_exit);
  1795. MODULE_AUTHOR("Mark A. Greer <mgreer@mvista.com>");
  1796. MODULE_DESCRIPTION("Generic Marvell MPSC serial/UART driver");
  1797. MODULE_VERSION(MPSC_VERSION);
  1798. MODULE_LICENSE("GPL");
  1799. MODULE_ALIAS_CHARDEV_MAJOR(MPSC_MAJOR);
  1800. MODULE_ALIAS("platform:" MPSC_CTLR_NAME);