mpsc.c 44 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830
  1. /*
  2. * drivers/serial/mpsc.c
  3. *
  4. * Generic driver for the MPSC (UART mode) on Marvell parts (e.g., GT64240,
  5. * GT64260, MV64340, MV64360, GT96100, ... ).
  6. *
  7. * Author: Mark A. Greer <mgreer@mvista.com>
  8. *
  9. * Based on an old MPSC driver that was in the linuxppc tree. It appears to
  10. * have been created by Chris Zankel (formerly of MontaVista) but there
  11. * is no proper Copyright so I'm not sure. Apparently, parts were also
  12. * taken from PPCBoot (now U-Boot). Also based on drivers/serial/8250.c
  13. * by Russell King.
  14. *
  15. * 2004 (c) MontaVista, Software, Inc. This file is licensed under
  16. * the terms of the GNU General Public License version 2. This program
  17. * is licensed "as is" without any warranty of any kind, whether express
  18. * or implied.
  19. */
  20. /*
  21. * The MPSC interface is much like a typical network controller's interface.
  22. * That is, you set up separate rings of descriptors for transmitting and
  23. * receiving data. There is also a pool of buffers with (one buffer per
  24. * descriptor) that incoming data are dma'd into or outgoing data are dma'd
  25. * out of.
  26. *
  27. * The MPSC requires two other controllers to be able to work. The Baud Rate
  28. * Generator (BRG) provides a clock at programmable frequencies which determines
  29. * the baud rate. The Serial DMA Controller (SDMA) takes incoming data from the
  30. * MPSC and DMA's it into memory or DMA's outgoing data and passes it to the
  31. * MPSC. It is actually the SDMA interrupt that the driver uses to keep the
  32. * transmit and receive "engines" going (i.e., indicate data has been
  33. * transmitted or received).
  34. *
  35. * NOTES:
  36. *
  37. * 1) Some chips have an erratum where several regs cannot be
  38. * read. To work around that, we keep a local copy of those regs in
  39. * 'mpsc_port_info'.
  40. *
  41. * 2) Some chips have an erratum where the ctlr will hang when the SDMA ctlr
  42. * accesses system mem with coherency enabled. For that reason, the driver
  43. * assumes that coherency for that ctlr has been disabled. This means
  44. * that when in a cache coherent system, the driver has to manually manage
  45. * the data cache on the areas that it touches because the dma_* macro are
  46. * basically no-ops.
  47. *
  48. * 3) There is an erratum (on PPC) where you can't use the instruction to do
  49. * a DMA_TO_DEVICE/cache clean so DMA_BIDIRECTIONAL/flushes are used in places
  50. * where a DMA_TO_DEVICE/clean would have [otherwise] sufficed.
  51. *
  52. * 4) AFAICT, hardware flow control isn't supported by the controller --MAG.
  53. */
  54. #include <linux/platform_device.h>
  55. #include "mpsc.h"
  56. /*
  57. * Define how this driver is known to the outside (we've been assigned a
  58. * range on the "Low-density serial ports" major).
  59. */
  60. #define MPSC_MAJOR 204
  61. #define MPSC_MINOR_START 44
  62. #define MPSC_DRIVER_NAME "MPSC"
  63. #define MPSC_DEVFS_NAME "ttymm/"
  64. #define MPSC_DEV_NAME "ttyMM"
  65. #define MPSC_VERSION "1.00"
  66. static struct mpsc_port_info mpsc_ports[MPSC_NUM_CTLRS];
  67. static struct mpsc_shared_regs mpsc_shared_regs;
  68. static struct uart_driver mpsc_reg;
  69. static void mpsc_start_rx(struct mpsc_port_info *pi);
  70. static void mpsc_free_ring_mem(struct mpsc_port_info *pi);
  71. static void mpsc_release_port(struct uart_port *port);
  72. /*
  73. ******************************************************************************
  74. *
  75. * Baud Rate Generator Routines (BRG)
  76. *
  77. ******************************************************************************
  78. */
  79. static void
  80. mpsc_brg_init(struct mpsc_port_info *pi, u32 clk_src)
  81. {
  82. u32 v;
  83. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  84. v = (v & ~(0xf << 18)) | ((clk_src & 0xf) << 18);
  85. if (pi->brg_can_tune)
  86. v &= ~(1 << 25);
  87. if (pi->mirror_regs)
  88. pi->BRG_BCR_m = v;
  89. writel(v, pi->brg_base + BRG_BCR);
  90. writel(readl(pi->brg_base + BRG_BTR) & 0xffff0000,
  91. pi->brg_base + BRG_BTR);
  92. return;
  93. }
  94. static void
  95. mpsc_brg_enable(struct mpsc_port_info *pi)
  96. {
  97. u32 v;
  98. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  99. v |= (1 << 16);
  100. if (pi->mirror_regs)
  101. pi->BRG_BCR_m = v;
  102. writel(v, pi->brg_base + BRG_BCR);
  103. return;
  104. }
  105. static void
  106. mpsc_brg_disable(struct mpsc_port_info *pi)
  107. {
  108. u32 v;
  109. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  110. v &= ~(1 << 16);
  111. if (pi->mirror_regs)
  112. pi->BRG_BCR_m = v;
  113. writel(v, pi->brg_base + BRG_BCR);
  114. return;
  115. }
  116. static inline void
  117. mpsc_set_baudrate(struct mpsc_port_info *pi, u32 baud)
  118. {
  119. /*
  120. * To set the baud, we adjust the CDV field in the BRG_BCR reg.
  121. * From manual: Baud = clk / ((CDV+1)*2) ==> CDV = (clk / (baud*2)) - 1.
  122. * However, the input clock is divided by 16 in the MPSC b/c of how
  123. * 'MPSC_MMCRH' was set up so we have to divide the 'clk' used in our
  124. * calculation by 16 to account for that. So the real calculation
  125. * that accounts for the way the mpsc is set up is:
  126. * CDV = (clk / (baud*2*16)) - 1 ==> CDV = (clk / (baud << 5)) - 1.
  127. */
  128. u32 cdv = (pi->port.uartclk / (baud << 5)) - 1;
  129. u32 v;
  130. mpsc_brg_disable(pi);
  131. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  132. v = (v & 0xffff0000) | (cdv & 0xffff);
  133. if (pi->mirror_regs)
  134. pi->BRG_BCR_m = v;
  135. writel(v, pi->brg_base + BRG_BCR);
  136. mpsc_brg_enable(pi);
  137. return;
  138. }
  139. /*
  140. ******************************************************************************
  141. *
  142. * Serial DMA Routines (SDMA)
  143. *
  144. ******************************************************************************
  145. */
  146. static void
  147. mpsc_sdma_burstsize(struct mpsc_port_info *pi, u32 burst_size)
  148. {
  149. u32 v;
  150. pr_debug("mpsc_sdma_burstsize[%d]: burst_size: %d\n",
  151. pi->port.line, burst_size);
  152. burst_size >>= 3; /* Divide by 8 b/c reg values are 8-byte chunks */
  153. if (burst_size < 2)
  154. v = 0x0; /* 1 64-bit word */
  155. else if (burst_size < 4)
  156. v = 0x1; /* 2 64-bit words */
  157. else if (burst_size < 8)
  158. v = 0x2; /* 4 64-bit words */
  159. else
  160. v = 0x3; /* 8 64-bit words */
  161. writel((readl(pi->sdma_base + SDMA_SDC) & (0x3 << 12)) | (v << 12),
  162. pi->sdma_base + SDMA_SDC);
  163. return;
  164. }
  165. static void
  166. mpsc_sdma_init(struct mpsc_port_info *pi, u32 burst_size)
  167. {
  168. pr_debug("mpsc_sdma_init[%d]: burst_size: %d\n", pi->port.line,
  169. burst_size);
  170. writel((readl(pi->sdma_base + SDMA_SDC) & 0x3ff) | 0x03f,
  171. pi->sdma_base + SDMA_SDC);
  172. mpsc_sdma_burstsize(pi, burst_size);
  173. return;
  174. }
  175. static inline u32
  176. mpsc_sdma_intr_mask(struct mpsc_port_info *pi, u32 mask)
  177. {
  178. u32 old, v;
  179. pr_debug("mpsc_sdma_intr_mask[%d]: mask: 0x%x\n", pi->port.line, mask);
  180. old = v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  181. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  182. mask &= 0xf;
  183. if (pi->port.line)
  184. mask <<= 8;
  185. v &= ~mask;
  186. if (pi->mirror_regs)
  187. pi->shared_regs->SDMA_INTR_MASK_m = v;
  188. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  189. if (pi->port.line)
  190. old >>= 8;
  191. return old & 0xf;
  192. }
  193. static inline void
  194. mpsc_sdma_intr_unmask(struct mpsc_port_info *pi, u32 mask)
  195. {
  196. u32 v;
  197. pr_debug("mpsc_sdma_intr_unmask[%d]: mask: 0x%x\n", pi->port.line,mask);
  198. v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  199. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  200. mask &= 0xf;
  201. if (pi->port.line)
  202. mask <<= 8;
  203. v |= mask;
  204. if (pi->mirror_regs)
  205. pi->shared_regs->SDMA_INTR_MASK_m = v;
  206. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  207. return;
  208. }
  209. static inline void
  210. mpsc_sdma_intr_ack(struct mpsc_port_info *pi)
  211. {
  212. pr_debug("mpsc_sdma_intr_ack[%d]: Acknowledging IRQ\n", pi->port.line);
  213. if (pi->mirror_regs)
  214. pi->shared_regs->SDMA_INTR_CAUSE_m = 0;
  215. writel(0, pi->shared_regs->sdma_intr_base + SDMA_INTR_CAUSE);
  216. return;
  217. }
  218. static inline void
  219. mpsc_sdma_set_rx_ring(struct mpsc_port_info *pi, struct mpsc_rx_desc *rxre_p)
  220. {
  221. pr_debug("mpsc_sdma_set_rx_ring[%d]: rxre_p: 0x%x\n",
  222. pi->port.line, (u32) rxre_p);
  223. writel((u32)rxre_p, pi->sdma_base + SDMA_SCRDP);
  224. return;
  225. }
  226. static inline void
  227. mpsc_sdma_set_tx_ring(struct mpsc_port_info *pi, struct mpsc_tx_desc *txre_p)
  228. {
  229. writel((u32)txre_p, pi->sdma_base + SDMA_SFTDP);
  230. writel((u32)txre_p, pi->sdma_base + SDMA_SCTDP);
  231. return;
  232. }
  233. static inline void
  234. mpsc_sdma_cmd(struct mpsc_port_info *pi, u32 val)
  235. {
  236. u32 v;
  237. v = readl(pi->sdma_base + SDMA_SDCM);
  238. if (val)
  239. v |= val;
  240. else
  241. v = 0;
  242. wmb();
  243. writel(v, pi->sdma_base + SDMA_SDCM);
  244. wmb();
  245. return;
  246. }
  247. static inline uint
  248. mpsc_sdma_tx_active(struct mpsc_port_info *pi)
  249. {
  250. return readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_TXD;
  251. }
  252. static inline void
  253. mpsc_sdma_start_tx(struct mpsc_port_info *pi)
  254. {
  255. struct mpsc_tx_desc *txre, *txre_p;
  256. /* If tx isn't running & there's a desc ready to go, start it */
  257. if (!mpsc_sdma_tx_active(pi)) {
  258. txre = (struct mpsc_tx_desc *)(pi->txr +
  259. (pi->txr_tail * MPSC_TXRE_SIZE));
  260. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  261. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  262. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  263. invalidate_dcache_range((ulong)txre,
  264. (ulong)txre + MPSC_TXRE_SIZE);
  265. #endif
  266. if (be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O) {
  267. txre_p = (struct mpsc_tx_desc *)(pi->txr_p +
  268. (pi->txr_tail *
  269. MPSC_TXRE_SIZE));
  270. mpsc_sdma_set_tx_ring(pi, txre_p);
  271. mpsc_sdma_cmd(pi, SDMA_SDCM_STD | SDMA_SDCM_TXD);
  272. }
  273. }
  274. return;
  275. }
  276. static inline void
  277. mpsc_sdma_stop(struct mpsc_port_info *pi)
  278. {
  279. pr_debug("mpsc_sdma_stop[%d]: Stopping SDMA\n", pi->port.line);
  280. /* Abort any SDMA transfers */
  281. mpsc_sdma_cmd(pi, 0);
  282. mpsc_sdma_cmd(pi, SDMA_SDCM_AR | SDMA_SDCM_AT);
  283. /* Clear the SDMA current and first TX and RX pointers */
  284. mpsc_sdma_set_tx_ring(pi, NULL);
  285. mpsc_sdma_set_rx_ring(pi, NULL);
  286. /* Disable interrupts */
  287. mpsc_sdma_intr_mask(pi, 0xf);
  288. mpsc_sdma_intr_ack(pi);
  289. return;
  290. }
  291. /*
  292. ******************************************************************************
  293. *
  294. * Multi-Protocol Serial Controller Routines (MPSC)
  295. *
  296. ******************************************************************************
  297. */
  298. static void
  299. mpsc_hw_init(struct mpsc_port_info *pi)
  300. {
  301. u32 v;
  302. pr_debug("mpsc_hw_init[%d]: Initializing hardware\n", pi->port.line);
  303. /* Set up clock routing */
  304. if (pi->mirror_regs) {
  305. v = pi->shared_regs->MPSC_MRR_m;
  306. v &= ~0x1c7;
  307. pi->shared_regs->MPSC_MRR_m = v;
  308. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  309. v = pi->shared_regs->MPSC_RCRR_m;
  310. v = (v & ~0xf0f) | 0x100;
  311. pi->shared_regs->MPSC_RCRR_m = v;
  312. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  313. v = pi->shared_regs->MPSC_TCRR_m;
  314. v = (v & ~0xf0f) | 0x100;
  315. pi->shared_regs->MPSC_TCRR_m = v;
  316. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  317. }
  318. else {
  319. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  320. v &= ~0x1c7;
  321. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  322. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  323. v = (v & ~0xf0f) | 0x100;
  324. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  325. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  326. v = (v & ~0xf0f) | 0x100;
  327. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  328. }
  329. /* Put MPSC in UART mode & enabel Tx/Rx egines */
  330. writel(0x000004c4, pi->mpsc_base + MPSC_MMCRL);
  331. /* No preamble, 16x divider, low-latency, */
  332. writel(0x04400400, pi->mpsc_base + MPSC_MMCRH);
  333. if (pi->mirror_regs) {
  334. pi->MPSC_CHR_1_m = 0;
  335. pi->MPSC_CHR_2_m = 0;
  336. }
  337. writel(0, pi->mpsc_base + MPSC_CHR_1);
  338. writel(0, pi->mpsc_base + MPSC_CHR_2);
  339. writel(pi->mpsc_max_idle, pi->mpsc_base + MPSC_CHR_3);
  340. writel(0, pi->mpsc_base + MPSC_CHR_4);
  341. writel(0, pi->mpsc_base + MPSC_CHR_5);
  342. writel(0, pi->mpsc_base + MPSC_CHR_6);
  343. writel(0, pi->mpsc_base + MPSC_CHR_7);
  344. writel(0, pi->mpsc_base + MPSC_CHR_8);
  345. writel(0, pi->mpsc_base + MPSC_CHR_9);
  346. writel(0, pi->mpsc_base + MPSC_CHR_10);
  347. return;
  348. }
  349. static inline void
  350. mpsc_enter_hunt(struct mpsc_port_info *pi)
  351. {
  352. pr_debug("mpsc_enter_hunt[%d]: Hunting...\n", pi->port.line);
  353. if (pi->mirror_regs) {
  354. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_EH,
  355. pi->mpsc_base + MPSC_CHR_2);
  356. /* Erratum prevents reading CHR_2 so just delay for a while */
  357. udelay(100);
  358. }
  359. else {
  360. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_EH,
  361. pi->mpsc_base + MPSC_CHR_2);
  362. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_EH)
  363. udelay(10);
  364. }
  365. return;
  366. }
  367. static inline void
  368. mpsc_freeze(struct mpsc_port_info *pi)
  369. {
  370. u32 v;
  371. pr_debug("mpsc_freeze[%d]: Freezing\n", pi->port.line);
  372. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  373. readl(pi->mpsc_base + MPSC_MPCR);
  374. v |= MPSC_MPCR_FRZ;
  375. if (pi->mirror_regs)
  376. pi->MPSC_MPCR_m = v;
  377. writel(v, pi->mpsc_base + MPSC_MPCR);
  378. return;
  379. }
  380. static inline void
  381. mpsc_unfreeze(struct mpsc_port_info *pi)
  382. {
  383. u32 v;
  384. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  385. readl(pi->mpsc_base + MPSC_MPCR);
  386. v &= ~MPSC_MPCR_FRZ;
  387. if (pi->mirror_regs)
  388. pi->MPSC_MPCR_m = v;
  389. writel(v, pi->mpsc_base + MPSC_MPCR);
  390. pr_debug("mpsc_unfreeze[%d]: Unfrozen\n", pi->port.line);
  391. return;
  392. }
  393. static inline void
  394. mpsc_set_char_length(struct mpsc_port_info *pi, u32 len)
  395. {
  396. u32 v;
  397. pr_debug("mpsc_set_char_length[%d]: char len: %d\n", pi->port.line,len);
  398. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  399. readl(pi->mpsc_base + MPSC_MPCR);
  400. v = (v & ~(0x3 << 12)) | ((len & 0x3) << 12);
  401. if (pi->mirror_regs)
  402. pi->MPSC_MPCR_m = v;
  403. writel(v, pi->mpsc_base + MPSC_MPCR);
  404. return;
  405. }
  406. static inline void
  407. mpsc_set_stop_bit_length(struct mpsc_port_info *pi, u32 len)
  408. {
  409. u32 v;
  410. pr_debug("mpsc_set_stop_bit_length[%d]: stop bits: %d\n",
  411. pi->port.line, len);
  412. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  413. readl(pi->mpsc_base + MPSC_MPCR);
  414. v = (v & ~(1 << 14)) | ((len & 0x1) << 14);
  415. if (pi->mirror_regs)
  416. pi->MPSC_MPCR_m = v;
  417. writel(v, pi->mpsc_base + MPSC_MPCR);
  418. return;
  419. }
  420. static inline void
  421. mpsc_set_parity(struct mpsc_port_info *pi, u32 p)
  422. {
  423. u32 v;
  424. pr_debug("mpsc_set_parity[%d]: parity bits: 0x%x\n", pi->port.line, p);
  425. v = (pi->mirror_regs) ? pi->MPSC_CHR_2_m :
  426. readl(pi->mpsc_base + MPSC_CHR_2);
  427. p &= 0x3;
  428. v = (v & ~0xc000c) | (p << 18) | (p << 2);
  429. if (pi->mirror_regs)
  430. pi->MPSC_CHR_2_m = v;
  431. writel(v, pi->mpsc_base + MPSC_CHR_2);
  432. return;
  433. }
  434. /*
  435. ******************************************************************************
  436. *
  437. * Driver Init Routines
  438. *
  439. ******************************************************************************
  440. */
  441. static void
  442. mpsc_init_hw(struct mpsc_port_info *pi)
  443. {
  444. pr_debug("mpsc_init_hw[%d]: Initializing\n", pi->port.line);
  445. mpsc_brg_init(pi, pi->brg_clk_src);
  446. mpsc_brg_enable(pi);
  447. mpsc_sdma_init(pi, dma_get_cache_alignment()); /* burst a cacheline */
  448. mpsc_sdma_stop(pi);
  449. mpsc_hw_init(pi);
  450. return;
  451. }
  452. static int
  453. mpsc_alloc_ring_mem(struct mpsc_port_info *pi)
  454. {
  455. int rc = 0;
  456. pr_debug("mpsc_alloc_ring_mem[%d]: Allocating ring mem\n",
  457. pi->port.line);
  458. if (!pi->dma_region) {
  459. if (!dma_supported(pi->port.dev, 0xffffffff)) {
  460. printk(KERN_ERR "MPSC: Inadequate DMA support\n");
  461. rc = -ENXIO;
  462. }
  463. else if ((pi->dma_region = dma_alloc_noncoherent(pi->port.dev,
  464. MPSC_DMA_ALLOC_SIZE, &pi->dma_region_p, GFP_KERNEL))
  465. == NULL) {
  466. printk(KERN_ERR "MPSC: Can't alloc Desc region\n");
  467. rc = -ENOMEM;
  468. }
  469. }
  470. return rc;
  471. }
  472. static void
  473. mpsc_free_ring_mem(struct mpsc_port_info *pi)
  474. {
  475. pr_debug("mpsc_free_ring_mem[%d]: Freeing ring mem\n", pi->port.line);
  476. if (pi->dma_region) {
  477. dma_free_noncoherent(pi->port.dev, MPSC_DMA_ALLOC_SIZE,
  478. pi->dma_region, pi->dma_region_p);
  479. pi->dma_region = NULL;
  480. pi->dma_region_p = (dma_addr_t) NULL;
  481. }
  482. return;
  483. }
  484. static void
  485. mpsc_init_rings(struct mpsc_port_info *pi)
  486. {
  487. struct mpsc_rx_desc *rxre;
  488. struct mpsc_tx_desc *txre;
  489. dma_addr_t dp, dp_p;
  490. u8 *bp, *bp_p;
  491. int i;
  492. pr_debug("mpsc_init_rings[%d]: Initializing rings\n", pi->port.line);
  493. BUG_ON(pi->dma_region == NULL);
  494. memset(pi->dma_region, 0, MPSC_DMA_ALLOC_SIZE);
  495. /*
  496. * Descriptors & buffers are multiples of cacheline size and must be
  497. * cacheline aligned.
  498. */
  499. dp = ALIGN((u32) pi->dma_region, dma_get_cache_alignment());
  500. dp_p = ALIGN((u32) pi->dma_region_p, dma_get_cache_alignment());
  501. /*
  502. * Partition dma region into rx ring descriptor, rx buffers,
  503. * tx ring descriptors, and tx buffers.
  504. */
  505. pi->rxr = dp;
  506. pi->rxr_p = dp_p;
  507. dp += MPSC_RXR_SIZE;
  508. dp_p += MPSC_RXR_SIZE;
  509. pi->rxb = (u8 *) dp;
  510. pi->rxb_p = (u8 *) dp_p;
  511. dp += MPSC_RXB_SIZE;
  512. dp_p += MPSC_RXB_SIZE;
  513. pi->rxr_posn = 0;
  514. pi->txr = dp;
  515. pi->txr_p = dp_p;
  516. dp += MPSC_TXR_SIZE;
  517. dp_p += MPSC_TXR_SIZE;
  518. pi->txb = (u8 *) dp;
  519. pi->txb_p = (u8 *) dp_p;
  520. pi->txr_head = 0;
  521. pi->txr_tail = 0;
  522. /* Init rx ring descriptors */
  523. dp = pi->rxr;
  524. dp_p = pi->rxr_p;
  525. bp = pi->rxb;
  526. bp_p = pi->rxb_p;
  527. for (i = 0; i < MPSC_RXR_ENTRIES; i++) {
  528. rxre = (struct mpsc_rx_desc *)dp;
  529. rxre->bufsize = cpu_to_be16(MPSC_RXBE_SIZE);
  530. rxre->bytecnt = cpu_to_be16(0);
  531. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  532. SDMA_DESC_CMDSTAT_EI |
  533. SDMA_DESC_CMDSTAT_F |
  534. SDMA_DESC_CMDSTAT_L);
  535. rxre->link = cpu_to_be32(dp_p + MPSC_RXRE_SIZE);
  536. rxre->buf_ptr = cpu_to_be32(bp_p);
  537. dp += MPSC_RXRE_SIZE;
  538. dp_p += MPSC_RXRE_SIZE;
  539. bp += MPSC_RXBE_SIZE;
  540. bp_p += MPSC_RXBE_SIZE;
  541. }
  542. rxre->link = cpu_to_be32(pi->rxr_p); /* Wrap last back to first */
  543. /* Init tx ring descriptors */
  544. dp = pi->txr;
  545. dp_p = pi->txr_p;
  546. bp = pi->txb;
  547. bp_p = pi->txb_p;
  548. for (i = 0; i < MPSC_TXR_ENTRIES; i++) {
  549. txre = (struct mpsc_tx_desc *)dp;
  550. txre->link = cpu_to_be32(dp_p + MPSC_TXRE_SIZE);
  551. txre->buf_ptr = cpu_to_be32(bp_p);
  552. dp += MPSC_TXRE_SIZE;
  553. dp_p += MPSC_TXRE_SIZE;
  554. bp += MPSC_TXBE_SIZE;
  555. bp_p += MPSC_TXBE_SIZE;
  556. }
  557. txre->link = cpu_to_be32(pi->txr_p); /* Wrap last back to first */
  558. dma_cache_sync((void *) pi->dma_region, MPSC_DMA_ALLOC_SIZE,
  559. DMA_BIDIRECTIONAL);
  560. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  561. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  562. flush_dcache_range((ulong)pi->dma_region,
  563. (ulong)pi->dma_region + MPSC_DMA_ALLOC_SIZE);
  564. #endif
  565. return;
  566. }
  567. static void
  568. mpsc_uninit_rings(struct mpsc_port_info *pi)
  569. {
  570. pr_debug("mpsc_uninit_rings[%d]: Uninitializing rings\n",pi->port.line);
  571. BUG_ON(pi->dma_region == NULL);
  572. pi->rxr = 0;
  573. pi->rxr_p = 0;
  574. pi->rxb = NULL;
  575. pi->rxb_p = NULL;
  576. pi->rxr_posn = 0;
  577. pi->txr = 0;
  578. pi->txr_p = 0;
  579. pi->txb = NULL;
  580. pi->txb_p = NULL;
  581. pi->txr_head = 0;
  582. pi->txr_tail = 0;
  583. return;
  584. }
  585. static int
  586. mpsc_make_ready(struct mpsc_port_info *pi)
  587. {
  588. int rc;
  589. pr_debug("mpsc_make_ready[%d]: Making cltr ready\n", pi->port.line);
  590. if (!pi->ready) {
  591. mpsc_init_hw(pi);
  592. if ((rc = mpsc_alloc_ring_mem(pi)))
  593. return rc;
  594. mpsc_init_rings(pi);
  595. pi->ready = 1;
  596. }
  597. return 0;
  598. }
  599. /*
  600. ******************************************************************************
  601. *
  602. * Interrupt Handling Routines
  603. *
  604. ******************************************************************************
  605. */
  606. static inline int
  607. mpsc_rx_intr(struct mpsc_port_info *pi, struct pt_regs *regs)
  608. {
  609. struct mpsc_rx_desc *rxre;
  610. struct tty_struct *tty = pi->port.info->tty;
  611. u32 cmdstat, bytes_in, i;
  612. int rc = 0;
  613. u8 *bp;
  614. char flag = TTY_NORMAL;
  615. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  616. rxre = (struct mpsc_rx_desc *)(pi->rxr + (pi->rxr_posn*MPSC_RXRE_SIZE));
  617. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  618. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  619. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  620. invalidate_dcache_range((ulong)rxre,
  621. (ulong)rxre + MPSC_RXRE_SIZE);
  622. #endif
  623. /*
  624. * Loop through Rx descriptors handling ones that have been completed.
  625. */
  626. while (!((cmdstat = be32_to_cpu(rxre->cmdstat)) & SDMA_DESC_CMDSTAT_O)){
  627. bytes_in = be16_to_cpu(rxre->bytecnt);
  628. /* Following use of tty struct directly is deprecated */
  629. if (unlikely(tty_buffer_request_room(tty, bytes_in) < bytes_in)) {
  630. if (tty->low_latency)
  631. tty_flip_buffer_push(tty);
  632. /*
  633. * If this failed then we will throw away the bytes
  634. * but must do so to clear interrupts.
  635. */
  636. }
  637. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  638. dma_cache_sync((void *) bp, MPSC_RXBE_SIZE, DMA_FROM_DEVICE);
  639. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  640. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  641. invalidate_dcache_range((ulong)bp,
  642. (ulong)bp + MPSC_RXBE_SIZE);
  643. #endif
  644. /*
  645. * Other than for parity error, the manual provides little
  646. * info on what data will be in a frame flagged by any of
  647. * these errors. For parity error, it is the last byte in
  648. * the buffer that had the error. As for the rest, I guess
  649. * we'll assume there is no data in the buffer.
  650. * If there is...it gets lost.
  651. */
  652. if (unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  653. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) {
  654. pi->port.icount.rx++;
  655. if (cmdstat & SDMA_DESC_CMDSTAT_BR) { /* Break */
  656. pi->port.icount.brk++;
  657. if (uart_handle_break(&pi->port))
  658. goto next_frame;
  659. }
  660. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)/* Framing */
  661. pi->port.icount.frame++;
  662. else if (cmdstat & SDMA_DESC_CMDSTAT_OR) /* Overrun */
  663. pi->port.icount.overrun++;
  664. cmdstat &= pi->port.read_status_mask;
  665. if (cmdstat & SDMA_DESC_CMDSTAT_BR)
  666. flag = TTY_BREAK;
  667. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)
  668. flag = TTY_FRAME;
  669. else if (cmdstat & SDMA_DESC_CMDSTAT_OR)
  670. flag = TTY_OVERRUN;
  671. else if (cmdstat & SDMA_DESC_CMDSTAT_PE)
  672. flag = TTY_PARITY;
  673. }
  674. if (uart_handle_sysrq_char(&pi->port, *bp, regs)) {
  675. bp++;
  676. bytes_in--;
  677. goto next_frame;
  678. }
  679. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  680. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) &&
  681. !(cmdstat & pi->port.ignore_status_mask))
  682. tty_insert_flip_char(tty, *bp, flag);
  683. else {
  684. for (i=0; i<bytes_in; i++)
  685. tty_insert_flip_char(tty, *bp++, TTY_NORMAL);
  686. pi->port.icount.rx += bytes_in;
  687. }
  688. next_frame:
  689. rxre->bytecnt = cpu_to_be16(0);
  690. wmb();
  691. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  692. SDMA_DESC_CMDSTAT_EI |
  693. SDMA_DESC_CMDSTAT_F |
  694. SDMA_DESC_CMDSTAT_L);
  695. wmb();
  696. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_BIDIRECTIONAL);
  697. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  698. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  699. flush_dcache_range((ulong)rxre,
  700. (ulong)rxre + MPSC_RXRE_SIZE);
  701. #endif
  702. /* Advance to next descriptor */
  703. pi->rxr_posn = (pi->rxr_posn + 1) & (MPSC_RXR_ENTRIES - 1);
  704. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  705. (pi->rxr_posn * MPSC_RXRE_SIZE));
  706. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  707. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  708. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  709. invalidate_dcache_range((ulong)rxre,
  710. (ulong)rxre + MPSC_RXRE_SIZE);
  711. #endif
  712. rc = 1;
  713. }
  714. /* Restart rx engine, if its stopped */
  715. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  716. mpsc_start_rx(pi);
  717. tty_flip_buffer_push(tty);
  718. return rc;
  719. }
  720. static inline void
  721. mpsc_setup_tx_desc(struct mpsc_port_info *pi, u32 count, u32 intr)
  722. {
  723. struct mpsc_tx_desc *txre;
  724. txre = (struct mpsc_tx_desc *)(pi->txr +
  725. (pi->txr_head * MPSC_TXRE_SIZE));
  726. txre->bytecnt = cpu_to_be16(count);
  727. txre->shadow = txre->bytecnt;
  728. wmb(); /* ensure cmdstat is last field updated */
  729. txre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O | SDMA_DESC_CMDSTAT_F |
  730. SDMA_DESC_CMDSTAT_L | ((intr) ?
  731. SDMA_DESC_CMDSTAT_EI
  732. : 0));
  733. wmb();
  734. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_BIDIRECTIONAL);
  735. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  736. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  737. flush_dcache_range((ulong)txre,
  738. (ulong)txre + MPSC_TXRE_SIZE);
  739. #endif
  740. return;
  741. }
  742. static inline void
  743. mpsc_copy_tx_data(struct mpsc_port_info *pi)
  744. {
  745. struct circ_buf *xmit = &pi->port.info->xmit;
  746. u8 *bp;
  747. u32 i;
  748. /* Make sure the desc ring isn't full */
  749. while (CIRC_CNT(pi->txr_head, pi->txr_tail, MPSC_TXR_ENTRIES) <
  750. (MPSC_TXR_ENTRIES - 1)) {
  751. if (pi->port.x_char) {
  752. /*
  753. * Ideally, we should use the TCS field in
  754. * CHR_1 to put the x_char out immediately but
  755. * errata prevents us from being able to read
  756. * CHR_2 to know that its safe to write to
  757. * CHR_1. Instead, just put it in-band with
  758. * all the other Tx data.
  759. */
  760. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  761. *bp = pi->port.x_char;
  762. pi->port.x_char = 0;
  763. i = 1;
  764. }
  765. else if (!uart_circ_empty(xmit) && !uart_tx_stopped(&pi->port)){
  766. i = min((u32) MPSC_TXBE_SIZE,
  767. (u32) uart_circ_chars_pending(xmit));
  768. i = min(i, (u32) CIRC_CNT_TO_END(xmit->head, xmit->tail,
  769. UART_XMIT_SIZE));
  770. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  771. memcpy(bp, &xmit->buf[xmit->tail], i);
  772. xmit->tail = (xmit->tail + i) & (UART_XMIT_SIZE - 1);
  773. if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
  774. uart_write_wakeup(&pi->port);
  775. }
  776. else /* All tx data copied into ring bufs */
  777. return;
  778. dma_cache_sync((void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  779. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  780. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  781. flush_dcache_range((ulong)bp,
  782. (ulong)bp + MPSC_TXBE_SIZE);
  783. #endif
  784. mpsc_setup_tx_desc(pi, i, 1);
  785. /* Advance to next descriptor */
  786. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  787. }
  788. return;
  789. }
  790. static inline int
  791. mpsc_tx_intr(struct mpsc_port_info *pi)
  792. {
  793. struct mpsc_tx_desc *txre;
  794. int rc = 0;
  795. if (!mpsc_sdma_tx_active(pi)) {
  796. txre = (struct mpsc_tx_desc *)(pi->txr +
  797. (pi->txr_tail * MPSC_TXRE_SIZE));
  798. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  799. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  800. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  801. invalidate_dcache_range((ulong)txre,
  802. (ulong)txre + MPSC_TXRE_SIZE);
  803. #endif
  804. while (!(be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O)) {
  805. rc = 1;
  806. pi->port.icount.tx += be16_to_cpu(txre->bytecnt);
  807. pi->txr_tail = (pi->txr_tail+1) & (MPSC_TXR_ENTRIES-1);
  808. /* If no more data to tx, fall out of loop */
  809. if (pi->txr_head == pi->txr_tail)
  810. break;
  811. txre = (struct mpsc_tx_desc *)(pi->txr +
  812. (pi->txr_tail * MPSC_TXRE_SIZE));
  813. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE,
  814. DMA_FROM_DEVICE);
  815. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  816. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  817. invalidate_dcache_range((ulong)txre,
  818. (ulong)txre + MPSC_TXRE_SIZE);
  819. #endif
  820. }
  821. mpsc_copy_tx_data(pi);
  822. mpsc_sdma_start_tx(pi); /* start next desc if ready */
  823. }
  824. return rc;
  825. }
  826. /*
  827. * This is the driver's interrupt handler. To avoid a race, we first clear
  828. * the interrupt, then handle any completed Rx/Tx descriptors. When done
  829. * handling those descriptors, we restart the Rx/Tx engines if they're stopped.
  830. */
  831. static irqreturn_t
  832. mpsc_sdma_intr(int irq, void *dev_id, struct pt_regs *regs)
  833. {
  834. struct mpsc_port_info *pi = dev_id;
  835. ulong iflags;
  836. int rc = IRQ_NONE;
  837. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Received\n",pi->port.line);
  838. spin_lock_irqsave(&pi->port.lock, iflags);
  839. mpsc_sdma_intr_ack(pi);
  840. if (mpsc_rx_intr(pi, regs))
  841. rc = IRQ_HANDLED;
  842. if (mpsc_tx_intr(pi))
  843. rc = IRQ_HANDLED;
  844. spin_unlock_irqrestore(&pi->port.lock, iflags);
  845. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Handled\n", pi->port.line);
  846. return rc;
  847. }
  848. /*
  849. ******************************************************************************
  850. *
  851. * serial_core.c Interface routines
  852. *
  853. ******************************************************************************
  854. */
  855. static uint
  856. mpsc_tx_empty(struct uart_port *port)
  857. {
  858. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  859. ulong iflags;
  860. uint rc;
  861. spin_lock_irqsave(&pi->port.lock, iflags);
  862. rc = mpsc_sdma_tx_active(pi) ? 0 : TIOCSER_TEMT;
  863. spin_unlock_irqrestore(&pi->port.lock, iflags);
  864. return rc;
  865. }
  866. static void
  867. mpsc_set_mctrl(struct uart_port *port, uint mctrl)
  868. {
  869. /* Have no way to set modem control lines AFAICT */
  870. return;
  871. }
  872. static uint
  873. mpsc_get_mctrl(struct uart_port *port)
  874. {
  875. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  876. u32 mflags, status;
  877. status = (pi->mirror_regs) ? pi->MPSC_CHR_10_m :
  878. readl(pi->mpsc_base + MPSC_CHR_10);
  879. mflags = 0;
  880. if (status & 0x1)
  881. mflags |= TIOCM_CTS;
  882. if (status & 0x2)
  883. mflags |= TIOCM_CAR;
  884. return mflags | TIOCM_DSR; /* No way to tell if DSR asserted */
  885. }
  886. static void
  887. mpsc_stop_tx(struct uart_port *port)
  888. {
  889. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  890. pr_debug("mpsc_stop_tx[%d]\n", port->line);
  891. mpsc_freeze(pi);
  892. return;
  893. }
  894. static void
  895. mpsc_start_tx(struct uart_port *port)
  896. {
  897. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  898. mpsc_unfreeze(pi);
  899. mpsc_copy_tx_data(pi);
  900. mpsc_sdma_start_tx(pi);
  901. pr_debug("mpsc_start_tx[%d]\n", port->line);
  902. return;
  903. }
  904. static void
  905. mpsc_start_rx(struct mpsc_port_info *pi)
  906. {
  907. pr_debug("mpsc_start_rx[%d]: Starting...\n", pi->port.line);
  908. /* Issue a Receive Abort to clear any receive errors */
  909. writel(MPSC_CHR_2_RA, pi->mpsc_base + MPSC_CHR_2);
  910. if (pi->rcv_data) {
  911. mpsc_enter_hunt(pi);
  912. mpsc_sdma_cmd(pi, SDMA_SDCM_ERD);
  913. }
  914. return;
  915. }
  916. static void
  917. mpsc_stop_rx(struct uart_port *port)
  918. {
  919. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  920. pr_debug("mpsc_stop_rx[%d]: Stopping...\n", port->line);
  921. mpsc_sdma_cmd(pi, SDMA_SDCM_AR);
  922. return;
  923. }
  924. static void
  925. mpsc_enable_ms(struct uart_port *port)
  926. {
  927. return; /* Not supported */
  928. }
  929. static void
  930. mpsc_break_ctl(struct uart_port *port, int ctl)
  931. {
  932. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  933. ulong flags;
  934. u32 v;
  935. v = ctl ? 0x00ff0000 : 0;
  936. spin_lock_irqsave(&pi->port.lock, flags);
  937. if (pi->mirror_regs)
  938. pi->MPSC_CHR_1_m = v;
  939. writel(v, pi->mpsc_base + MPSC_CHR_1);
  940. spin_unlock_irqrestore(&pi->port.lock, flags);
  941. return;
  942. }
  943. static int
  944. mpsc_startup(struct uart_port *port)
  945. {
  946. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  947. u32 flag = 0;
  948. int rc;
  949. pr_debug("mpsc_startup[%d]: Starting up MPSC, irq: %d\n",
  950. port->line, pi->port.irq);
  951. if ((rc = mpsc_make_ready(pi)) == 0) {
  952. /* Setup IRQ handler */
  953. mpsc_sdma_intr_ack(pi);
  954. /* If irq's are shared, need to set flag */
  955. if (mpsc_ports[0].port.irq == mpsc_ports[1].port.irq)
  956. flag = SA_SHIRQ;
  957. if (request_irq(pi->port.irq, mpsc_sdma_intr, flag,
  958. "mpsc/sdma", pi))
  959. printk(KERN_ERR "MPSC: Can't get SDMA IRQ %d\n",
  960. pi->port.irq);
  961. mpsc_sdma_intr_unmask(pi, 0xf);
  962. mpsc_sdma_set_rx_ring(pi, (struct mpsc_rx_desc *)(pi->rxr_p +
  963. (pi->rxr_posn * MPSC_RXRE_SIZE)));
  964. }
  965. return rc;
  966. }
  967. static void
  968. mpsc_shutdown(struct uart_port *port)
  969. {
  970. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  971. pr_debug("mpsc_shutdown[%d]: Shutting down MPSC\n", port->line);
  972. mpsc_sdma_stop(pi);
  973. free_irq(pi->port.irq, pi);
  974. return;
  975. }
  976. static void
  977. mpsc_set_termios(struct uart_port *port, struct termios *termios,
  978. struct termios *old)
  979. {
  980. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  981. u32 baud;
  982. ulong flags;
  983. u32 chr_bits, stop_bits, par;
  984. pi->c_iflag = termios->c_iflag;
  985. pi->c_cflag = termios->c_cflag;
  986. switch (termios->c_cflag & CSIZE) {
  987. case CS5:
  988. chr_bits = MPSC_MPCR_CL_5;
  989. break;
  990. case CS6:
  991. chr_bits = MPSC_MPCR_CL_6;
  992. break;
  993. case CS7:
  994. chr_bits = MPSC_MPCR_CL_7;
  995. break;
  996. case CS8:
  997. default:
  998. chr_bits = MPSC_MPCR_CL_8;
  999. break;
  1000. }
  1001. if (termios->c_cflag & CSTOPB)
  1002. stop_bits = MPSC_MPCR_SBL_2;
  1003. else
  1004. stop_bits = MPSC_MPCR_SBL_1;
  1005. par = MPSC_CHR_2_PAR_EVEN;
  1006. if (termios->c_cflag & PARENB)
  1007. if (termios->c_cflag & PARODD)
  1008. par = MPSC_CHR_2_PAR_ODD;
  1009. #ifdef CMSPAR
  1010. if (termios->c_cflag & CMSPAR) {
  1011. if (termios->c_cflag & PARODD)
  1012. par = MPSC_CHR_2_PAR_MARK;
  1013. else
  1014. par = MPSC_CHR_2_PAR_SPACE;
  1015. }
  1016. #endif
  1017. baud = uart_get_baud_rate(port, termios, old, 0, port->uartclk);
  1018. spin_lock_irqsave(&pi->port.lock, flags);
  1019. uart_update_timeout(port, termios->c_cflag, baud);
  1020. mpsc_set_char_length(pi, chr_bits);
  1021. mpsc_set_stop_bit_length(pi, stop_bits);
  1022. mpsc_set_parity(pi, par);
  1023. mpsc_set_baudrate(pi, baud);
  1024. /* Characters/events to read */
  1025. pi->rcv_data = 1;
  1026. pi->port.read_status_mask = SDMA_DESC_CMDSTAT_OR;
  1027. if (termios->c_iflag & INPCK)
  1028. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1029. SDMA_DESC_CMDSTAT_FR;
  1030. if (termios->c_iflag & (BRKINT | PARMRK))
  1031. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1032. /* Characters/events to ignore */
  1033. pi->port.ignore_status_mask = 0;
  1034. if (termios->c_iflag & IGNPAR)
  1035. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1036. SDMA_DESC_CMDSTAT_FR;
  1037. if (termios->c_iflag & IGNBRK) {
  1038. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1039. if (termios->c_iflag & IGNPAR)
  1040. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_OR;
  1041. }
  1042. /* Ignore all chars if CREAD not set */
  1043. if (!(termios->c_cflag & CREAD))
  1044. pi->rcv_data = 0;
  1045. else
  1046. mpsc_start_rx(pi);
  1047. spin_unlock_irqrestore(&pi->port.lock, flags);
  1048. return;
  1049. }
  1050. static const char *
  1051. mpsc_type(struct uart_port *port)
  1052. {
  1053. pr_debug("mpsc_type[%d]: port type: %s\n", port->line,MPSC_DRIVER_NAME);
  1054. return MPSC_DRIVER_NAME;
  1055. }
  1056. static int
  1057. mpsc_request_port(struct uart_port *port)
  1058. {
  1059. /* Should make chip/platform specific call */
  1060. return 0;
  1061. }
  1062. static void
  1063. mpsc_release_port(struct uart_port *port)
  1064. {
  1065. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1066. if (pi->ready) {
  1067. mpsc_uninit_rings(pi);
  1068. mpsc_free_ring_mem(pi);
  1069. pi->ready = 0;
  1070. }
  1071. return;
  1072. }
  1073. static void
  1074. mpsc_config_port(struct uart_port *port, int flags)
  1075. {
  1076. return;
  1077. }
  1078. static int
  1079. mpsc_verify_port(struct uart_port *port, struct serial_struct *ser)
  1080. {
  1081. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1082. int rc = 0;
  1083. pr_debug("mpsc_verify_port[%d]: Verifying port data\n", pi->port.line);
  1084. if (ser->type != PORT_UNKNOWN && ser->type != PORT_MPSC)
  1085. rc = -EINVAL;
  1086. else if (pi->port.irq != ser->irq)
  1087. rc = -EINVAL;
  1088. else if (ser->io_type != SERIAL_IO_MEM)
  1089. rc = -EINVAL;
  1090. else if (pi->port.uartclk / 16 != ser->baud_base) /* Not sure */
  1091. rc = -EINVAL;
  1092. else if ((void *)pi->port.mapbase != ser->iomem_base)
  1093. rc = -EINVAL;
  1094. else if (pi->port.iobase != ser->port)
  1095. rc = -EINVAL;
  1096. else if (ser->hub6 != 0)
  1097. rc = -EINVAL;
  1098. return rc;
  1099. }
  1100. static struct uart_ops mpsc_pops = {
  1101. .tx_empty = mpsc_tx_empty,
  1102. .set_mctrl = mpsc_set_mctrl,
  1103. .get_mctrl = mpsc_get_mctrl,
  1104. .stop_tx = mpsc_stop_tx,
  1105. .start_tx = mpsc_start_tx,
  1106. .stop_rx = mpsc_stop_rx,
  1107. .enable_ms = mpsc_enable_ms,
  1108. .break_ctl = mpsc_break_ctl,
  1109. .startup = mpsc_startup,
  1110. .shutdown = mpsc_shutdown,
  1111. .set_termios = mpsc_set_termios,
  1112. .type = mpsc_type,
  1113. .release_port = mpsc_release_port,
  1114. .request_port = mpsc_request_port,
  1115. .config_port = mpsc_config_port,
  1116. .verify_port = mpsc_verify_port,
  1117. };
  1118. /*
  1119. ******************************************************************************
  1120. *
  1121. * Console Interface Routines
  1122. *
  1123. ******************************************************************************
  1124. */
  1125. #ifdef CONFIG_SERIAL_MPSC_CONSOLE
  1126. static void
  1127. mpsc_console_write(struct console *co, const char *s, uint count)
  1128. {
  1129. struct mpsc_port_info *pi = &mpsc_ports[co->index];
  1130. u8 *bp, *dp, add_cr = 0;
  1131. int i;
  1132. while (mpsc_sdma_tx_active(pi))
  1133. udelay(100);
  1134. while (count > 0) {
  1135. bp = dp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  1136. for (i = 0; i < MPSC_TXBE_SIZE; i++) {
  1137. if (count == 0)
  1138. break;
  1139. if (add_cr) {
  1140. *(dp++) = '\r';
  1141. add_cr = 0;
  1142. }
  1143. else {
  1144. *(dp++) = *s;
  1145. if (*(s++) == '\n') { /* add '\r' after '\n' */
  1146. add_cr = 1;
  1147. count++;
  1148. }
  1149. }
  1150. count--;
  1151. }
  1152. dma_cache_sync((void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  1153. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1154. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1155. flush_dcache_range((ulong)bp,
  1156. (ulong)bp + MPSC_TXBE_SIZE);
  1157. #endif
  1158. mpsc_setup_tx_desc(pi, i, 0);
  1159. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1160. mpsc_sdma_start_tx(pi);
  1161. while (mpsc_sdma_tx_active(pi))
  1162. udelay(100);
  1163. pi->txr_tail = (pi->txr_tail + 1) & (MPSC_TXR_ENTRIES - 1);
  1164. }
  1165. return;
  1166. }
  1167. static int __init
  1168. mpsc_console_setup(struct console *co, char *options)
  1169. {
  1170. struct mpsc_port_info *pi;
  1171. int baud, bits, parity, flow;
  1172. pr_debug("mpsc_console_setup[%d]: options: %s\n", co->index, options);
  1173. if (co->index >= MPSC_NUM_CTLRS)
  1174. co->index = 0;
  1175. pi = &mpsc_ports[co->index];
  1176. baud = pi->default_baud;
  1177. bits = pi->default_bits;
  1178. parity = pi->default_parity;
  1179. flow = pi->default_flow;
  1180. if (!pi->port.ops)
  1181. return -ENODEV;
  1182. spin_lock_init(&pi->port.lock); /* Temporary fix--copied from 8250.c */
  1183. if (options)
  1184. uart_parse_options(options, &baud, &parity, &bits, &flow);
  1185. return uart_set_options(&pi->port, co, baud, parity, bits, flow);
  1186. }
  1187. static struct console mpsc_console = {
  1188. .name = MPSC_DEV_NAME,
  1189. .write = mpsc_console_write,
  1190. .device = uart_console_device,
  1191. .setup = mpsc_console_setup,
  1192. .flags = CON_PRINTBUFFER,
  1193. .index = -1,
  1194. .data = &mpsc_reg,
  1195. };
  1196. static int __init
  1197. mpsc_late_console_init(void)
  1198. {
  1199. pr_debug("mpsc_late_console_init: Enter\n");
  1200. if (!(mpsc_console.flags & CON_ENABLED))
  1201. register_console(&mpsc_console);
  1202. return 0;
  1203. }
  1204. late_initcall(mpsc_late_console_init);
  1205. #define MPSC_CONSOLE &mpsc_console
  1206. #else
  1207. #define MPSC_CONSOLE NULL
  1208. #endif
  1209. /*
  1210. ******************************************************************************
  1211. *
  1212. * Dummy Platform Driver to extract & map shared register regions
  1213. *
  1214. ******************************************************************************
  1215. */
  1216. static void
  1217. mpsc_resource_err(char *s)
  1218. {
  1219. printk(KERN_WARNING "MPSC: Platform device resource error in %s\n", s);
  1220. return;
  1221. }
  1222. static int
  1223. mpsc_shared_map_regs(struct platform_device *pd)
  1224. {
  1225. struct resource *r;
  1226. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1227. MPSC_ROUTING_BASE_ORDER)) && request_mem_region(r->start,
  1228. MPSC_ROUTING_REG_BLOCK_SIZE, "mpsc_routing_regs")) {
  1229. mpsc_shared_regs.mpsc_routing_base = ioremap(r->start,
  1230. MPSC_ROUTING_REG_BLOCK_SIZE);
  1231. mpsc_shared_regs.mpsc_routing_base_p = r->start;
  1232. }
  1233. else {
  1234. mpsc_resource_err("MPSC routing base");
  1235. return -ENOMEM;
  1236. }
  1237. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1238. MPSC_SDMA_INTR_BASE_ORDER)) && request_mem_region(r->start,
  1239. MPSC_SDMA_INTR_REG_BLOCK_SIZE, "sdma_intr_regs")) {
  1240. mpsc_shared_regs.sdma_intr_base = ioremap(r->start,
  1241. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1242. mpsc_shared_regs.sdma_intr_base_p = r->start;
  1243. }
  1244. else {
  1245. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1246. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1247. MPSC_ROUTING_REG_BLOCK_SIZE);
  1248. mpsc_resource_err("SDMA intr base");
  1249. return -ENOMEM;
  1250. }
  1251. return 0;
  1252. }
  1253. static void
  1254. mpsc_shared_unmap_regs(void)
  1255. {
  1256. if (!mpsc_shared_regs.mpsc_routing_base) {
  1257. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1258. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1259. MPSC_ROUTING_REG_BLOCK_SIZE);
  1260. }
  1261. if (!mpsc_shared_regs.sdma_intr_base) {
  1262. iounmap(mpsc_shared_regs.sdma_intr_base);
  1263. release_mem_region(mpsc_shared_regs.sdma_intr_base_p,
  1264. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1265. }
  1266. mpsc_shared_regs.mpsc_routing_base = NULL;
  1267. mpsc_shared_regs.sdma_intr_base = NULL;
  1268. mpsc_shared_regs.mpsc_routing_base_p = 0;
  1269. mpsc_shared_regs.sdma_intr_base_p = 0;
  1270. return;
  1271. }
  1272. static int
  1273. mpsc_shared_drv_probe(struct platform_device *dev)
  1274. {
  1275. struct mpsc_shared_pdata *pdata;
  1276. int rc = -ENODEV;
  1277. if (dev->id == 0) {
  1278. if (!(rc = mpsc_shared_map_regs(dev))) {
  1279. pdata = (struct mpsc_shared_pdata *)dev->dev.platform_data;
  1280. mpsc_shared_regs.MPSC_MRR_m = pdata->mrr_val;
  1281. mpsc_shared_regs.MPSC_RCRR_m= pdata->rcrr_val;
  1282. mpsc_shared_regs.MPSC_TCRR_m= pdata->tcrr_val;
  1283. mpsc_shared_regs.SDMA_INTR_CAUSE_m =
  1284. pdata->intr_cause_val;
  1285. mpsc_shared_regs.SDMA_INTR_MASK_m =
  1286. pdata->intr_mask_val;
  1287. rc = 0;
  1288. }
  1289. }
  1290. return rc;
  1291. }
  1292. static int
  1293. mpsc_shared_drv_remove(struct platform_device *dev)
  1294. {
  1295. int rc = -ENODEV;
  1296. if (dev->id == 0) {
  1297. mpsc_shared_unmap_regs();
  1298. mpsc_shared_regs.MPSC_MRR_m = 0;
  1299. mpsc_shared_regs.MPSC_RCRR_m = 0;
  1300. mpsc_shared_regs.MPSC_TCRR_m = 0;
  1301. mpsc_shared_regs.SDMA_INTR_CAUSE_m = 0;
  1302. mpsc_shared_regs.SDMA_INTR_MASK_m = 0;
  1303. rc = 0;
  1304. }
  1305. return rc;
  1306. }
  1307. static struct platform_driver mpsc_shared_driver = {
  1308. .probe = mpsc_shared_drv_probe,
  1309. .remove = mpsc_shared_drv_remove,
  1310. .driver = {
  1311. .name = MPSC_SHARED_NAME,
  1312. },
  1313. };
  1314. /*
  1315. ******************************************************************************
  1316. *
  1317. * Driver Interface Routines
  1318. *
  1319. ******************************************************************************
  1320. */
  1321. static struct uart_driver mpsc_reg = {
  1322. .owner = THIS_MODULE,
  1323. .driver_name = MPSC_DRIVER_NAME,
  1324. .devfs_name = MPSC_DEVFS_NAME,
  1325. .dev_name = MPSC_DEV_NAME,
  1326. .major = MPSC_MAJOR,
  1327. .minor = MPSC_MINOR_START,
  1328. .nr = MPSC_NUM_CTLRS,
  1329. .cons = MPSC_CONSOLE,
  1330. };
  1331. static int
  1332. mpsc_drv_map_regs(struct mpsc_port_info *pi, struct platform_device *pd)
  1333. {
  1334. struct resource *r;
  1335. if ((r = platform_get_resource(pd, IORESOURCE_MEM, MPSC_BASE_ORDER)) &&
  1336. request_mem_region(r->start, MPSC_REG_BLOCK_SIZE, "mpsc_regs")){
  1337. pi->mpsc_base = ioremap(r->start, MPSC_REG_BLOCK_SIZE);
  1338. pi->mpsc_base_p = r->start;
  1339. }
  1340. else {
  1341. mpsc_resource_err("MPSC base");
  1342. return -ENOMEM;
  1343. }
  1344. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1345. MPSC_SDMA_BASE_ORDER)) && request_mem_region(r->start,
  1346. MPSC_SDMA_REG_BLOCK_SIZE, "sdma_regs")) {
  1347. pi->sdma_base = ioremap(r->start,MPSC_SDMA_REG_BLOCK_SIZE);
  1348. pi->sdma_base_p = r->start;
  1349. }
  1350. else {
  1351. mpsc_resource_err("SDMA base");
  1352. return -ENOMEM;
  1353. }
  1354. if ((r = platform_get_resource(pd,IORESOURCE_MEM,MPSC_BRG_BASE_ORDER))
  1355. && request_mem_region(r->start, MPSC_BRG_REG_BLOCK_SIZE,
  1356. "brg_regs")) {
  1357. pi->brg_base = ioremap(r->start, MPSC_BRG_REG_BLOCK_SIZE);
  1358. pi->brg_base_p = r->start;
  1359. }
  1360. else {
  1361. mpsc_resource_err("BRG base");
  1362. return -ENOMEM;
  1363. }
  1364. return 0;
  1365. }
  1366. static void
  1367. mpsc_drv_unmap_regs(struct mpsc_port_info *pi)
  1368. {
  1369. if (!pi->mpsc_base) {
  1370. iounmap(pi->mpsc_base);
  1371. release_mem_region(pi->mpsc_base_p, MPSC_REG_BLOCK_SIZE);
  1372. }
  1373. if (!pi->sdma_base) {
  1374. iounmap(pi->sdma_base);
  1375. release_mem_region(pi->sdma_base_p, MPSC_SDMA_REG_BLOCK_SIZE);
  1376. }
  1377. if (!pi->brg_base) {
  1378. iounmap(pi->brg_base);
  1379. release_mem_region(pi->brg_base_p, MPSC_BRG_REG_BLOCK_SIZE);
  1380. }
  1381. pi->mpsc_base = NULL;
  1382. pi->sdma_base = NULL;
  1383. pi->brg_base = NULL;
  1384. pi->mpsc_base_p = 0;
  1385. pi->sdma_base_p = 0;
  1386. pi->brg_base_p = 0;
  1387. return;
  1388. }
  1389. static void
  1390. mpsc_drv_get_platform_data(struct mpsc_port_info *pi,
  1391. struct platform_device *pd, int num)
  1392. {
  1393. struct mpsc_pdata *pdata;
  1394. pdata = (struct mpsc_pdata *)pd->dev.platform_data;
  1395. pi->port.uartclk = pdata->brg_clk_freq;
  1396. pi->port.iotype = UPIO_MEM;
  1397. pi->port.line = num;
  1398. pi->port.type = PORT_MPSC;
  1399. pi->port.fifosize = MPSC_TXBE_SIZE;
  1400. pi->port.membase = pi->mpsc_base;
  1401. pi->port.mapbase = (ulong)pi->mpsc_base;
  1402. pi->port.ops = &mpsc_pops;
  1403. pi->mirror_regs = pdata->mirror_regs;
  1404. pi->cache_mgmt = pdata->cache_mgmt;
  1405. pi->brg_can_tune = pdata->brg_can_tune;
  1406. pi->brg_clk_src = pdata->brg_clk_src;
  1407. pi->mpsc_max_idle = pdata->max_idle;
  1408. pi->default_baud = pdata->default_baud;
  1409. pi->default_bits = pdata->default_bits;
  1410. pi->default_parity = pdata->default_parity;
  1411. pi->default_flow = pdata->default_flow;
  1412. /* Initial values of mirrored regs */
  1413. pi->MPSC_CHR_1_m = pdata->chr_1_val;
  1414. pi->MPSC_CHR_2_m = pdata->chr_2_val;
  1415. pi->MPSC_CHR_10_m = pdata->chr_10_val;
  1416. pi->MPSC_MPCR_m = pdata->mpcr_val;
  1417. pi->BRG_BCR_m = pdata->bcr_val;
  1418. pi->shared_regs = &mpsc_shared_regs;
  1419. pi->port.irq = platform_get_irq(pd, 0);
  1420. return;
  1421. }
  1422. static int
  1423. mpsc_drv_probe(struct platform_device *dev)
  1424. {
  1425. struct mpsc_port_info *pi;
  1426. int rc = -ENODEV;
  1427. pr_debug("mpsc_drv_probe: Adding MPSC %d\n", dev->id);
  1428. if (dev->id < MPSC_NUM_CTLRS) {
  1429. pi = &mpsc_ports[dev->id];
  1430. if (!(rc = mpsc_drv_map_regs(pi, dev))) {
  1431. mpsc_drv_get_platform_data(pi, dev, dev->id);
  1432. if (!(rc = mpsc_make_ready(pi)))
  1433. if (!(rc = uart_add_one_port(&mpsc_reg,
  1434. &pi->port)))
  1435. rc = 0;
  1436. else {
  1437. mpsc_release_port(
  1438. (struct uart_port *)pi);
  1439. mpsc_drv_unmap_regs(pi);
  1440. }
  1441. else
  1442. mpsc_drv_unmap_regs(pi);
  1443. }
  1444. }
  1445. return rc;
  1446. }
  1447. static int
  1448. mpsc_drv_remove(struct platform_device *dev)
  1449. {
  1450. pr_debug("mpsc_drv_exit: Removing MPSC %d\n", dev->id);
  1451. if (dev->id < MPSC_NUM_CTLRS) {
  1452. uart_remove_one_port(&mpsc_reg, &mpsc_ports[dev->id].port);
  1453. mpsc_release_port((struct uart_port *)&mpsc_ports[dev->id].port);
  1454. mpsc_drv_unmap_regs(&mpsc_ports[dev->id]);
  1455. return 0;
  1456. }
  1457. else
  1458. return -ENODEV;
  1459. }
  1460. static struct platform_driver mpsc_driver = {
  1461. .probe = mpsc_drv_probe,
  1462. .remove = mpsc_drv_remove,
  1463. .driver = {
  1464. .name = MPSC_CTLR_NAME,
  1465. },
  1466. };
  1467. static int __init
  1468. mpsc_drv_init(void)
  1469. {
  1470. int rc;
  1471. printk(KERN_INFO "Serial: MPSC driver $Revision: 1.00 $\n");
  1472. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1473. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1474. if (!(rc = uart_register_driver(&mpsc_reg))) {
  1475. if (!(rc = platform_driver_register(&mpsc_shared_driver))) {
  1476. if ((rc = platform_driver_register(&mpsc_driver))) {
  1477. platform_driver_unregister(&mpsc_shared_driver);
  1478. uart_unregister_driver(&mpsc_reg);
  1479. }
  1480. }
  1481. else
  1482. uart_unregister_driver(&mpsc_reg);
  1483. }
  1484. return rc;
  1485. }
  1486. static void __exit
  1487. mpsc_drv_exit(void)
  1488. {
  1489. platform_driver_unregister(&mpsc_driver);
  1490. platform_driver_unregister(&mpsc_shared_driver);
  1491. uart_unregister_driver(&mpsc_reg);
  1492. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1493. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1494. return;
  1495. }
  1496. module_init(mpsc_drv_init);
  1497. module_exit(mpsc_drv_exit);
  1498. MODULE_AUTHOR("Mark A. Greer <mgreer@mvista.com>");
  1499. MODULE_DESCRIPTION("Generic Marvell MPSC serial/UART driver $Revision: 1.00 $");
  1500. MODULE_VERSION(MPSC_VERSION);
  1501. MODULE_LICENSE("GPL");
  1502. MODULE_ALIAS_CHARDEV_MAJOR(MPSC_MAJOR);