mpsc.c 44 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831
  1. /*
  2. * drivers/serial/mpsc.c
  3. *
  4. * Generic driver for the MPSC (UART mode) on Marvell parts (e.g., GT64240,
  5. * GT64260, MV64340, MV64360, GT96100, ... ).
  6. *
  7. * Author: Mark A. Greer <mgreer@mvista.com>
  8. *
  9. * Based on an old MPSC driver that was in the linuxppc tree. It appears to
  10. * have been created by Chris Zankel (formerly of MontaVista) but there
  11. * is no proper Copyright so I'm not sure. Apparently, parts were also
  12. * taken from PPCBoot (now U-Boot). Also based on drivers/serial/8250.c
  13. * by Russell King.
  14. *
  15. * 2004 (c) MontaVista, Software, Inc. This file is licensed under
  16. * the terms of the GNU General Public License version 2. This program
  17. * is licensed "as is" without any warranty of any kind, whether express
  18. * or implied.
  19. */
  20. /*
  21. * The MPSC interface is much like a typical network controller's interface.
  22. * That is, you set up separate rings of descriptors for transmitting and
  23. * receiving data. There is also a pool of buffers with (one buffer per
  24. * descriptor) that incoming data are dma'd into or outgoing data are dma'd
  25. * out of.
  26. *
  27. * The MPSC requires two other controllers to be able to work. The Baud Rate
  28. * Generator (BRG) provides a clock at programmable frequencies which determines
  29. * the baud rate. The Serial DMA Controller (SDMA) takes incoming data from the
  30. * MPSC and DMA's it into memory or DMA's outgoing data and passes it to the
  31. * MPSC. It is actually the SDMA interrupt that the driver uses to keep the
  32. * transmit and receive "engines" going (i.e., indicate data has been
  33. * transmitted or received).
  34. *
  35. * NOTES:
  36. *
  37. * 1) Some chips have an erratum where several regs cannot be
  38. * read. To work around that, we keep a local copy of those regs in
  39. * 'mpsc_port_info'.
  40. *
  41. * 2) Some chips have an erratum where the ctlr will hang when the SDMA ctlr
  42. * accesses system mem with coherency enabled. For that reason, the driver
  43. * assumes that coherency for that ctlr has been disabled. This means
  44. * that when in a cache coherent system, the driver has to manually manage
  45. * the data cache on the areas that it touches because the dma_* macro are
  46. * basically no-ops.
  47. *
  48. * 3) There is an erratum (on PPC) where you can't use the instruction to do
  49. * a DMA_TO_DEVICE/cache clean so DMA_BIDIRECTIONAL/flushes are used in places
  50. * where a DMA_TO_DEVICE/clean would have [otherwise] sufficed.
  51. *
  52. * 4) AFAICT, hardware flow control isn't supported by the controller --MAG.
  53. */
  54. #include "mpsc.h"
  55. /*
  56. * Define how this driver is known to the outside (we've been assigned a
  57. * range on the "Low-density serial ports" major).
  58. */
  59. #define MPSC_MAJOR 204
  60. #define MPSC_MINOR_START 44
  61. #define MPSC_DRIVER_NAME "MPSC"
  62. #define MPSC_DEVFS_NAME "ttymm/"
  63. #define MPSC_DEV_NAME "ttyMM"
  64. #define MPSC_VERSION "1.00"
  65. static struct mpsc_port_info mpsc_ports[MPSC_NUM_CTLRS];
  66. static struct mpsc_shared_regs mpsc_shared_regs;
  67. static struct uart_driver mpsc_reg;
  68. static void mpsc_start_rx(struct mpsc_port_info *pi);
  69. static void mpsc_free_ring_mem(struct mpsc_port_info *pi);
  70. static void mpsc_release_port(struct uart_port *port);
  71. /*
  72. ******************************************************************************
  73. *
  74. * Baud Rate Generator Routines (BRG)
  75. *
  76. ******************************************************************************
  77. */
  78. static void
  79. mpsc_brg_init(struct mpsc_port_info *pi, u32 clk_src)
  80. {
  81. u32 v;
  82. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  83. v = (v & ~(0xf << 18)) | ((clk_src & 0xf) << 18);
  84. if (pi->brg_can_tune)
  85. v &= ~(1 << 25);
  86. if (pi->mirror_regs)
  87. pi->BRG_BCR_m = v;
  88. writel(v, pi->brg_base + BRG_BCR);
  89. writel(readl(pi->brg_base + BRG_BTR) & 0xffff0000,
  90. pi->brg_base + BRG_BTR);
  91. return;
  92. }
  93. static void
  94. mpsc_brg_enable(struct mpsc_port_info *pi)
  95. {
  96. u32 v;
  97. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  98. v |= (1 << 16);
  99. if (pi->mirror_regs)
  100. pi->BRG_BCR_m = v;
  101. writel(v, pi->brg_base + BRG_BCR);
  102. return;
  103. }
  104. static void
  105. mpsc_brg_disable(struct mpsc_port_info *pi)
  106. {
  107. u32 v;
  108. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  109. v &= ~(1 << 16);
  110. if (pi->mirror_regs)
  111. pi->BRG_BCR_m = v;
  112. writel(v, pi->brg_base + BRG_BCR);
  113. return;
  114. }
  115. static inline void
  116. mpsc_set_baudrate(struct mpsc_port_info *pi, u32 baud)
  117. {
  118. /*
  119. * To set the baud, we adjust the CDV field in the BRG_BCR reg.
  120. * From manual: Baud = clk / ((CDV+1)*2) ==> CDV = (clk / (baud*2)) - 1.
  121. * However, the input clock is divided by 16 in the MPSC b/c of how
  122. * 'MPSC_MMCRH' was set up so we have to divide the 'clk' used in our
  123. * calculation by 16 to account for that. So the real calculation
  124. * that accounts for the way the mpsc is set up is:
  125. * CDV = (clk / (baud*2*16)) - 1 ==> CDV = (clk / (baud << 5)) - 1.
  126. */
  127. u32 cdv = (pi->port.uartclk / (baud << 5)) - 1;
  128. u32 v;
  129. mpsc_brg_disable(pi);
  130. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  131. v = (v & 0xffff0000) | (cdv & 0xffff);
  132. if (pi->mirror_regs)
  133. pi->BRG_BCR_m = v;
  134. writel(v, pi->brg_base + BRG_BCR);
  135. mpsc_brg_enable(pi);
  136. return;
  137. }
  138. /*
  139. ******************************************************************************
  140. *
  141. * Serial DMA Routines (SDMA)
  142. *
  143. ******************************************************************************
  144. */
  145. static void
  146. mpsc_sdma_burstsize(struct mpsc_port_info *pi, u32 burst_size)
  147. {
  148. u32 v;
  149. pr_debug("mpsc_sdma_burstsize[%d]: burst_size: %d\n",
  150. pi->port.line, burst_size);
  151. burst_size >>= 3; /* Divide by 8 b/c reg values are 8-byte chunks */
  152. if (burst_size < 2)
  153. v = 0x0; /* 1 64-bit word */
  154. else if (burst_size < 4)
  155. v = 0x1; /* 2 64-bit words */
  156. else if (burst_size < 8)
  157. v = 0x2; /* 4 64-bit words */
  158. else
  159. v = 0x3; /* 8 64-bit words */
  160. writel((readl(pi->sdma_base + SDMA_SDC) & (0x3 << 12)) | (v << 12),
  161. pi->sdma_base + SDMA_SDC);
  162. return;
  163. }
  164. static void
  165. mpsc_sdma_init(struct mpsc_port_info *pi, u32 burst_size)
  166. {
  167. pr_debug("mpsc_sdma_init[%d]: burst_size: %d\n", pi->port.line,
  168. burst_size);
  169. writel((readl(pi->sdma_base + SDMA_SDC) & 0x3ff) | 0x03f,
  170. pi->sdma_base + SDMA_SDC);
  171. mpsc_sdma_burstsize(pi, burst_size);
  172. return;
  173. }
  174. static inline u32
  175. mpsc_sdma_intr_mask(struct mpsc_port_info *pi, u32 mask)
  176. {
  177. u32 old, v;
  178. pr_debug("mpsc_sdma_intr_mask[%d]: mask: 0x%x\n", pi->port.line, mask);
  179. old = v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  180. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  181. mask &= 0xf;
  182. if (pi->port.line)
  183. mask <<= 8;
  184. v &= ~mask;
  185. if (pi->mirror_regs)
  186. pi->shared_regs->SDMA_INTR_MASK_m = v;
  187. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  188. if (pi->port.line)
  189. old >>= 8;
  190. return old & 0xf;
  191. }
  192. static inline void
  193. mpsc_sdma_intr_unmask(struct mpsc_port_info *pi, u32 mask)
  194. {
  195. u32 v;
  196. pr_debug("mpsc_sdma_intr_unmask[%d]: mask: 0x%x\n", pi->port.line,mask);
  197. v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  198. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  199. mask &= 0xf;
  200. if (pi->port.line)
  201. mask <<= 8;
  202. v |= mask;
  203. if (pi->mirror_regs)
  204. pi->shared_regs->SDMA_INTR_MASK_m = v;
  205. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  206. return;
  207. }
  208. static inline void
  209. mpsc_sdma_intr_ack(struct mpsc_port_info *pi)
  210. {
  211. pr_debug("mpsc_sdma_intr_ack[%d]: Acknowledging IRQ\n", pi->port.line);
  212. if (pi->mirror_regs)
  213. pi->shared_regs->SDMA_INTR_CAUSE_m = 0;
  214. writel(0, pi->shared_regs->sdma_intr_base + SDMA_INTR_CAUSE);
  215. return;
  216. }
  217. static inline void
  218. mpsc_sdma_set_rx_ring(struct mpsc_port_info *pi, struct mpsc_rx_desc *rxre_p)
  219. {
  220. pr_debug("mpsc_sdma_set_rx_ring[%d]: rxre_p: 0x%x\n",
  221. pi->port.line, (u32) rxre_p);
  222. writel((u32)rxre_p, pi->sdma_base + SDMA_SCRDP);
  223. return;
  224. }
  225. static inline void
  226. mpsc_sdma_set_tx_ring(struct mpsc_port_info *pi, struct mpsc_tx_desc *txre_p)
  227. {
  228. writel((u32)txre_p, pi->sdma_base + SDMA_SFTDP);
  229. writel((u32)txre_p, pi->sdma_base + SDMA_SCTDP);
  230. return;
  231. }
  232. static inline void
  233. mpsc_sdma_cmd(struct mpsc_port_info *pi, u32 val)
  234. {
  235. u32 v;
  236. v = readl(pi->sdma_base + SDMA_SDCM);
  237. if (val)
  238. v |= val;
  239. else
  240. v = 0;
  241. wmb();
  242. writel(v, pi->sdma_base + SDMA_SDCM);
  243. wmb();
  244. return;
  245. }
  246. static inline uint
  247. mpsc_sdma_tx_active(struct mpsc_port_info *pi)
  248. {
  249. return readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_TXD;
  250. }
  251. static inline void
  252. mpsc_sdma_start_tx(struct mpsc_port_info *pi)
  253. {
  254. struct mpsc_tx_desc *txre, *txre_p;
  255. /* If tx isn't running & there's a desc ready to go, start it */
  256. if (!mpsc_sdma_tx_active(pi)) {
  257. txre = (struct mpsc_tx_desc *)(pi->txr +
  258. (pi->txr_tail * MPSC_TXRE_SIZE));
  259. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  260. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  261. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  262. invalidate_dcache_range((ulong)txre,
  263. (ulong)txre + MPSC_TXRE_SIZE);
  264. #endif
  265. if (be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O) {
  266. txre_p = (struct mpsc_tx_desc *)(pi->txr_p +
  267. (pi->txr_tail *
  268. MPSC_TXRE_SIZE));
  269. mpsc_sdma_set_tx_ring(pi, txre_p);
  270. mpsc_sdma_cmd(pi, SDMA_SDCM_STD | SDMA_SDCM_TXD);
  271. }
  272. }
  273. return;
  274. }
  275. static inline void
  276. mpsc_sdma_stop(struct mpsc_port_info *pi)
  277. {
  278. pr_debug("mpsc_sdma_stop[%d]: Stopping SDMA\n", pi->port.line);
  279. /* Abort any SDMA transfers */
  280. mpsc_sdma_cmd(pi, 0);
  281. mpsc_sdma_cmd(pi, SDMA_SDCM_AR | SDMA_SDCM_AT);
  282. /* Clear the SDMA current and first TX and RX pointers */
  283. mpsc_sdma_set_tx_ring(pi, NULL);
  284. mpsc_sdma_set_rx_ring(pi, NULL);
  285. /* Disable interrupts */
  286. mpsc_sdma_intr_mask(pi, 0xf);
  287. mpsc_sdma_intr_ack(pi);
  288. return;
  289. }
  290. /*
  291. ******************************************************************************
  292. *
  293. * Multi-Protocol Serial Controller Routines (MPSC)
  294. *
  295. ******************************************************************************
  296. */
  297. static void
  298. mpsc_hw_init(struct mpsc_port_info *pi)
  299. {
  300. u32 v;
  301. pr_debug("mpsc_hw_init[%d]: Initializing hardware\n", pi->port.line);
  302. /* Set up clock routing */
  303. if (pi->mirror_regs) {
  304. v = pi->shared_regs->MPSC_MRR_m;
  305. v &= ~0x1c7;
  306. pi->shared_regs->MPSC_MRR_m = v;
  307. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  308. v = pi->shared_regs->MPSC_RCRR_m;
  309. v = (v & ~0xf0f) | 0x100;
  310. pi->shared_regs->MPSC_RCRR_m = v;
  311. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  312. v = pi->shared_regs->MPSC_TCRR_m;
  313. v = (v & ~0xf0f) | 0x100;
  314. pi->shared_regs->MPSC_TCRR_m = v;
  315. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  316. }
  317. else {
  318. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  319. v &= ~0x1c7;
  320. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  321. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  322. v = (v & ~0xf0f) | 0x100;
  323. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  324. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  325. v = (v & ~0xf0f) | 0x100;
  326. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  327. }
  328. /* Put MPSC in UART mode & enabel Tx/Rx egines */
  329. writel(0x000004c4, pi->mpsc_base + MPSC_MMCRL);
  330. /* No preamble, 16x divider, low-latency, */
  331. writel(0x04400400, pi->mpsc_base + MPSC_MMCRH);
  332. if (pi->mirror_regs) {
  333. pi->MPSC_CHR_1_m = 0;
  334. pi->MPSC_CHR_2_m = 0;
  335. }
  336. writel(0, pi->mpsc_base + MPSC_CHR_1);
  337. writel(0, pi->mpsc_base + MPSC_CHR_2);
  338. writel(pi->mpsc_max_idle, pi->mpsc_base + MPSC_CHR_3);
  339. writel(0, pi->mpsc_base + MPSC_CHR_4);
  340. writel(0, pi->mpsc_base + MPSC_CHR_5);
  341. writel(0, pi->mpsc_base + MPSC_CHR_6);
  342. writel(0, pi->mpsc_base + MPSC_CHR_7);
  343. writel(0, pi->mpsc_base + MPSC_CHR_8);
  344. writel(0, pi->mpsc_base + MPSC_CHR_9);
  345. writel(0, pi->mpsc_base + MPSC_CHR_10);
  346. return;
  347. }
  348. static inline void
  349. mpsc_enter_hunt(struct mpsc_port_info *pi)
  350. {
  351. pr_debug("mpsc_enter_hunt[%d]: Hunting...\n", pi->port.line);
  352. if (pi->mirror_regs) {
  353. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_EH,
  354. pi->mpsc_base + MPSC_CHR_2);
  355. /* Erratum prevents reading CHR_2 so just delay for a while */
  356. udelay(100);
  357. }
  358. else {
  359. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_EH,
  360. pi->mpsc_base + MPSC_CHR_2);
  361. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_EH)
  362. udelay(10);
  363. }
  364. return;
  365. }
  366. static inline void
  367. mpsc_freeze(struct mpsc_port_info *pi)
  368. {
  369. u32 v;
  370. pr_debug("mpsc_freeze[%d]: Freezing\n", pi->port.line);
  371. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  372. readl(pi->mpsc_base + MPSC_MPCR);
  373. v |= MPSC_MPCR_FRZ;
  374. if (pi->mirror_regs)
  375. pi->MPSC_MPCR_m = v;
  376. writel(v, pi->mpsc_base + MPSC_MPCR);
  377. return;
  378. }
  379. static inline void
  380. mpsc_unfreeze(struct mpsc_port_info *pi)
  381. {
  382. u32 v;
  383. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  384. readl(pi->mpsc_base + MPSC_MPCR);
  385. v &= ~MPSC_MPCR_FRZ;
  386. if (pi->mirror_regs)
  387. pi->MPSC_MPCR_m = v;
  388. writel(v, pi->mpsc_base + MPSC_MPCR);
  389. pr_debug("mpsc_unfreeze[%d]: Unfrozen\n", pi->port.line);
  390. return;
  391. }
  392. static inline void
  393. mpsc_set_char_length(struct mpsc_port_info *pi, u32 len)
  394. {
  395. u32 v;
  396. pr_debug("mpsc_set_char_length[%d]: char len: %d\n", pi->port.line,len);
  397. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  398. readl(pi->mpsc_base + MPSC_MPCR);
  399. v = (v & ~(0x3 << 12)) | ((len & 0x3) << 12);
  400. if (pi->mirror_regs)
  401. pi->MPSC_MPCR_m = v;
  402. writel(v, pi->mpsc_base + MPSC_MPCR);
  403. return;
  404. }
  405. static inline void
  406. mpsc_set_stop_bit_length(struct mpsc_port_info *pi, u32 len)
  407. {
  408. u32 v;
  409. pr_debug("mpsc_set_stop_bit_length[%d]: stop bits: %d\n",
  410. pi->port.line, len);
  411. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  412. readl(pi->mpsc_base + MPSC_MPCR);
  413. v = (v & ~(1 << 14)) | ((len & 0x1) << 14);
  414. if (pi->mirror_regs)
  415. pi->MPSC_MPCR_m = v;
  416. writel(v, pi->mpsc_base + MPSC_MPCR);
  417. return;
  418. }
  419. static inline void
  420. mpsc_set_parity(struct mpsc_port_info *pi, u32 p)
  421. {
  422. u32 v;
  423. pr_debug("mpsc_set_parity[%d]: parity bits: 0x%x\n", pi->port.line, p);
  424. v = (pi->mirror_regs) ? pi->MPSC_CHR_2_m :
  425. readl(pi->mpsc_base + MPSC_CHR_2);
  426. p &= 0x3;
  427. v = (v & ~0xc000c) | (p << 18) | (p << 2);
  428. if (pi->mirror_regs)
  429. pi->MPSC_CHR_2_m = v;
  430. writel(v, pi->mpsc_base + MPSC_CHR_2);
  431. return;
  432. }
  433. /*
  434. ******************************************************************************
  435. *
  436. * Driver Init Routines
  437. *
  438. ******************************************************************************
  439. */
  440. static void
  441. mpsc_init_hw(struct mpsc_port_info *pi)
  442. {
  443. pr_debug("mpsc_init_hw[%d]: Initializing\n", pi->port.line);
  444. mpsc_brg_init(pi, pi->brg_clk_src);
  445. mpsc_brg_enable(pi);
  446. mpsc_sdma_init(pi, dma_get_cache_alignment()); /* burst a cacheline */
  447. mpsc_sdma_stop(pi);
  448. mpsc_hw_init(pi);
  449. return;
  450. }
  451. static int
  452. mpsc_alloc_ring_mem(struct mpsc_port_info *pi)
  453. {
  454. int rc = 0;
  455. pr_debug("mpsc_alloc_ring_mem[%d]: Allocating ring mem\n",
  456. pi->port.line);
  457. if (!pi->dma_region) {
  458. if (!dma_supported(pi->port.dev, 0xffffffff)) {
  459. printk(KERN_ERR "MPSC: Inadequate DMA support\n");
  460. rc = -ENXIO;
  461. }
  462. else if ((pi->dma_region = dma_alloc_noncoherent(pi->port.dev,
  463. MPSC_DMA_ALLOC_SIZE, &pi->dma_region_p, GFP_KERNEL))
  464. == NULL) {
  465. printk(KERN_ERR "MPSC: Can't alloc Desc region\n");
  466. rc = -ENOMEM;
  467. }
  468. }
  469. return rc;
  470. }
  471. static void
  472. mpsc_free_ring_mem(struct mpsc_port_info *pi)
  473. {
  474. pr_debug("mpsc_free_ring_mem[%d]: Freeing ring mem\n", pi->port.line);
  475. if (pi->dma_region) {
  476. dma_free_noncoherent(pi->port.dev, MPSC_DMA_ALLOC_SIZE,
  477. pi->dma_region, pi->dma_region_p);
  478. pi->dma_region = NULL;
  479. pi->dma_region_p = (dma_addr_t) NULL;
  480. }
  481. return;
  482. }
  483. static void
  484. mpsc_init_rings(struct mpsc_port_info *pi)
  485. {
  486. struct mpsc_rx_desc *rxre;
  487. struct mpsc_tx_desc *txre;
  488. dma_addr_t dp, dp_p;
  489. u8 *bp, *bp_p;
  490. int i;
  491. pr_debug("mpsc_init_rings[%d]: Initializing rings\n", pi->port.line);
  492. BUG_ON(pi->dma_region == NULL);
  493. memset(pi->dma_region, 0, MPSC_DMA_ALLOC_SIZE);
  494. /*
  495. * Descriptors & buffers are multiples of cacheline size and must be
  496. * cacheline aligned.
  497. */
  498. dp = ALIGN((u32) pi->dma_region, dma_get_cache_alignment());
  499. dp_p = ALIGN((u32) pi->dma_region_p, dma_get_cache_alignment());
  500. /*
  501. * Partition dma region into rx ring descriptor, rx buffers,
  502. * tx ring descriptors, and tx buffers.
  503. */
  504. pi->rxr = dp;
  505. pi->rxr_p = dp_p;
  506. dp += MPSC_RXR_SIZE;
  507. dp_p += MPSC_RXR_SIZE;
  508. pi->rxb = (u8 *) dp;
  509. pi->rxb_p = (u8 *) dp_p;
  510. dp += MPSC_RXB_SIZE;
  511. dp_p += MPSC_RXB_SIZE;
  512. pi->rxr_posn = 0;
  513. pi->txr = dp;
  514. pi->txr_p = dp_p;
  515. dp += MPSC_TXR_SIZE;
  516. dp_p += MPSC_TXR_SIZE;
  517. pi->txb = (u8 *) dp;
  518. pi->txb_p = (u8 *) dp_p;
  519. pi->txr_head = 0;
  520. pi->txr_tail = 0;
  521. /* Init rx ring descriptors */
  522. dp = pi->rxr;
  523. dp_p = pi->rxr_p;
  524. bp = pi->rxb;
  525. bp_p = pi->rxb_p;
  526. for (i = 0; i < MPSC_RXR_ENTRIES; i++) {
  527. rxre = (struct mpsc_rx_desc *)dp;
  528. rxre->bufsize = cpu_to_be16(MPSC_RXBE_SIZE);
  529. rxre->bytecnt = cpu_to_be16(0);
  530. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  531. SDMA_DESC_CMDSTAT_EI |
  532. SDMA_DESC_CMDSTAT_F |
  533. SDMA_DESC_CMDSTAT_L);
  534. rxre->link = cpu_to_be32(dp_p + MPSC_RXRE_SIZE);
  535. rxre->buf_ptr = cpu_to_be32(bp_p);
  536. dp += MPSC_RXRE_SIZE;
  537. dp_p += MPSC_RXRE_SIZE;
  538. bp += MPSC_RXBE_SIZE;
  539. bp_p += MPSC_RXBE_SIZE;
  540. }
  541. rxre->link = cpu_to_be32(pi->rxr_p); /* Wrap last back to first */
  542. /* Init tx ring descriptors */
  543. dp = pi->txr;
  544. dp_p = pi->txr_p;
  545. bp = pi->txb;
  546. bp_p = pi->txb_p;
  547. for (i = 0; i < MPSC_TXR_ENTRIES; i++) {
  548. txre = (struct mpsc_tx_desc *)dp;
  549. txre->link = cpu_to_be32(dp_p + MPSC_TXRE_SIZE);
  550. txre->buf_ptr = cpu_to_be32(bp_p);
  551. dp += MPSC_TXRE_SIZE;
  552. dp_p += MPSC_TXRE_SIZE;
  553. bp += MPSC_TXBE_SIZE;
  554. bp_p += MPSC_TXBE_SIZE;
  555. }
  556. txre->link = cpu_to_be32(pi->txr_p); /* Wrap last back to first */
  557. dma_cache_sync((void *) pi->dma_region, MPSC_DMA_ALLOC_SIZE,
  558. DMA_BIDIRECTIONAL);
  559. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  560. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  561. flush_dcache_range((ulong)pi->dma_region,
  562. (ulong)pi->dma_region + MPSC_DMA_ALLOC_SIZE);
  563. #endif
  564. return;
  565. }
  566. static void
  567. mpsc_uninit_rings(struct mpsc_port_info *pi)
  568. {
  569. pr_debug("mpsc_uninit_rings[%d]: Uninitializing rings\n",pi->port.line);
  570. BUG_ON(pi->dma_region == NULL);
  571. pi->rxr = 0;
  572. pi->rxr_p = 0;
  573. pi->rxb = NULL;
  574. pi->rxb_p = NULL;
  575. pi->rxr_posn = 0;
  576. pi->txr = 0;
  577. pi->txr_p = 0;
  578. pi->txb = NULL;
  579. pi->txb_p = NULL;
  580. pi->txr_head = 0;
  581. pi->txr_tail = 0;
  582. return;
  583. }
  584. static int
  585. mpsc_make_ready(struct mpsc_port_info *pi)
  586. {
  587. int rc;
  588. pr_debug("mpsc_make_ready[%d]: Making cltr ready\n", pi->port.line);
  589. if (!pi->ready) {
  590. mpsc_init_hw(pi);
  591. if ((rc = mpsc_alloc_ring_mem(pi)))
  592. return rc;
  593. mpsc_init_rings(pi);
  594. pi->ready = 1;
  595. }
  596. return 0;
  597. }
  598. /*
  599. ******************************************************************************
  600. *
  601. * Interrupt Handling Routines
  602. *
  603. ******************************************************************************
  604. */
  605. static inline int
  606. mpsc_rx_intr(struct mpsc_port_info *pi, struct pt_regs *regs)
  607. {
  608. struct mpsc_rx_desc *rxre;
  609. struct tty_struct *tty = pi->port.info->tty;
  610. u32 cmdstat, bytes_in, i;
  611. int rc = 0;
  612. u8 *bp;
  613. char flag = TTY_NORMAL;
  614. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  615. rxre = (struct mpsc_rx_desc *)(pi->rxr + (pi->rxr_posn*MPSC_RXRE_SIZE));
  616. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  617. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  618. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  619. invalidate_dcache_range((ulong)rxre,
  620. (ulong)rxre + MPSC_RXRE_SIZE);
  621. #endif
  622. /*
  623. * Loop through Rx descriptors handling ones that have been completed.
  624. */
  625. while (!((cmdstat = be32_to_cpu(rxre->cmdstat)) & SDMA_DESC_CMDSTAT_O)){
  626. bytes_in = be16_to_cpu(rxre->bytecnt);
  627. /* Following use of tty struct directly is deprecated */
  628. if (unlikely((tty->flip.count + bytes_in) >= TTY_FLIPBUF_SIZE)){
  629. if (tty->low_latency)
  630. tty_flip_buffer_push(tty);
  631. /*
  632. * If this failed then we will throw awa the bytes
  633. * but mst do so to clear interrupts.
  634. */
  635. }
  636. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  637. dma_cache_sync((void *) bp, MPSC_RXBE_SIZE, DMA_FROM_DEVICE);
  638. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  639. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  640. invalidate_dcache_range((ulong)bp,
  641. (ulong)bp + MPSC_RXBE_SIZE);
  642. #endif
  643. /*
  644. * Other than for parity error, the manual provides little
  645. * info on what data will be in a frame flagged by any of
  646. * these errors. For parity error, it is the last byte in
  647. * the buffer that had the error. As for the rest, I guess
  648. * we'll assume there is no data in the buffer.
  649. * If there is...it gets lost.
  650. */
  651. if (unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  652. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) {
  653. pi->port.icount.rx++;
  654. if (cmdstat & SDMA_DESC_CMDSTAT_BR) { /* Break */
  655. pi->port.icount.brk++;
  656. if (uart_handle_break(&pi->port))
  657. goto next_frame;
  658. }
  659. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)/* Framing */
  660. pi->port.icount.frame++;
  661. else if (cmdstat & SDMA_DESC_CMDSTAT_OR) /* Overrun */
  662. pi->port.icount.overrun++;
  663. cmdstat &= pi->port.read_status_mask;
  664. if (cmdstat & SDMA_DESC_CMDSTAT_BR)
  665. flag = TTY_BREAK;
  666. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)
  667. flag = TTY_FRAME;
  668. else if (cmdstat & SDMA_DESC_CMDSTAT_OR)
  669. flag = TTY_OVERRUN;
  670. else if (cmdstat & SDMA_DESC_CMDSTAT_PE)
  671. flag = TTY_PARITY;
  672. }
  673. if (uart_handle_sysrq_char(&pi->port, *bp, regs)) {
  674. bp++;
  675. bytes_in--;
  676. goto next_frame;
  677. }
  678. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  679. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) &&
  680. !(cmdstat & pi->port.ignore_status_mask))
  681. tty_insert_flip_char(tty, *bp, flag);
  682. else {
  683. for (i=0; i<bytes_in; i++)
  684. tty_insert_flip_char(tty, *bp++, TTY_NORMAL);
  685. pi->port.icount.rx += bytes_in;
  686. }
  687. next_frame:
  688. rxre->bytecnt = cpu_to_be16(0);
  689. wmb();
  690. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  691. SDMA_DESC_CMDSTAT_EI |
  692. SDMA_DESC_CMDSTAT_F |
  693. SDMA_DESC_CMDSTAT_L);
  694. wmb();
  695. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_BIDIRECTIONAL);
  696. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  697. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  698. flush_dcache_range((ulong)rxre,
  699. (ulong)rxre + MPSC_RXRE_SIZE);
  700. #endif
  701. /* Advance to next descriptor */
  702. pi->rxr_posn = (pi->rxr_posn + 1) & (MPSC_RXR_ENTRIES - 1);
  703. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  704. (pi->rxr_posn * MPSC_RXRE_SIZE));
  705. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  706. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  707. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  708. invalidate_dcache_range((ulong)rxre,
  709. (ulong)rxre + MPSC_RXRE_SIZE);
  710. #endif
  711. rc = 1;
  712. }
  713. /* Restart rx engine, if its stopped */
  714. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  715. mpsc_start_rx(pi);
  716. tty_flip_buffer_push(tty);
  717. return rc;
  718. }
  719. static inline void
  720. mpsc_setup_tx_desc(struct mpsc_port_info *pi, u32 count, u32 intr)
  721. {
  722. struct mpsc_tx_desc *txre;
  723. txre = (struct mpsc_tx_desc *)(pi->txr +
  724. (pi->txr_head * MPSC_TXRE_SIZE));
  725. txre->bytecnt = cpu_to_be16(count);
  726. txre->shadow = txre->bytecnt;
  727. wmb(); /* ensure cmdstat is last field updated */
  728. txre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O | SDMA_DESC_CMDSTAT_F |
  729. SDMA_DESC_CMDSTAT_L | ((intr) ?
  730. SDMA_DESC_CMDSTAT_EI
  731. : 0));
  732. wmb();
  733. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_BIDIRECTIONAL);
  734. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  735. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  736. flush_dcache_range((ulong)txre,
  737. (ulong)txre + MPSC_TXRE_SIZE);
  738. #endif
  739. return;
  740. }
  741. static inline void
  742. mpsc_copy_tx_data(struct mpsc_port_info *pi)
  743. {
  744. struct circ_buf *xmit = &pi->port.info->xmit;
  745. u8 *bp;
  746. u32 i;
  747. /* Make sure the desc ring isn't full */
  748. while (CIRC_CNT(pi->txr_head, pi->txr_tail, MPSC_TXR_ENTRIES) <
  749. (MPSC_TXR_ENTRIES - 1)) {
  750. if (pi->port.x_char) {
  751. /*
  752. * Ideally, we should use the TCS field in
  753. * CHR_1 to put the x_char out immediately but
  754. * errata prevents us from being able to read
  755. * CHR_2 to know that its safe to write to
  756. * CHR_1. Instead, just put it in-band with
  757. * all the other Tx data.
  758. */
  759. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  760. *bp = pi->port.x_char;
  761. pi->port.x_char = 0;
  762. i = 1;
  763. }
  764. else if (!uart_circ_empty(xmit) && !uart_tx_stopped(&pi->port)){
  765. i = min((u32) MPSC_TXBE_SIZE,
  766. (u32) uart_circ_chars_pending(xmit));
  767. i = min(i, (u32) CIRC_CNT_TO_END(xmit->head, xmit->tail,
  768. UART_XMIT_SIZE));
  769. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  770. memcpy(bp, &xmit->buf[xmit->tail], i);
  771. xmit->tail = (xmit->tail + i) & (UART_XMIT_SIZE - 1);
  772. if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
  773. uart_write_wakeup(&pi->port);
  774. }
  775. else /* All tx data copied into ring bufs */
  776. return;
  777. dma_cache_sync((void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  778. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  779. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  780. flush_dcache_range((ulong)bp,
  781. (ulong)bp + MPSC_TXBE_SIZE);
  782. #endif
  783. mpsc_setup_tx_desc(pi, i, 1);
  784. /* Advance to next descriptor */
  785. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  786. }
  787. return;
  788. }
  789. static inline int
  790. mpsc_tx_intr(struct mpsc_port_info *pi)
  791. {
  792. struct mpsc_tx_desc *txre;
  793. int rc = 0;
  794. if (!mpsc_sdma_tx_active(pi)) {
  795. txre = (struct mpsc_tx_desc *)(pi->txr +
  796. (pi->txr_tail * MPSC_TXRE_SIZE));
  797. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  798. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  799. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  800. invalidate_dcache_range((ulong)txre,
  801. (ulong)txre + MPSC_TXRE_SIZE);
  802. #endif
  803. while (!(be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O)) {
  804. rc = 1;
  805. pi->port.icount.tx += be16_to_cpu(txre->bytecnt);
  806. pi->txr_tail = (pi->txr_tail+1) & (MPSC_TXR_ENTRIES-1);
  807. /* If no more data to tx, fall out of loop */
  808. if (pi->txr_head == pi->txr_tail)
  809. break;
  810. txre = (struct mpsc_tx_desc *)(pi->txr +
  811. (pi->txr_tail * MPSC_TXRE_SIZE));
  812. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE,
  813. DMA_FROM_DEVICE);
  814. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  815. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  816. invalidate_dcache_range((ulong)txre,
  817. (ulong)txre + MPSC_TXRE_SIZE);
  818. #endif
  819. }
  820. mpsc_copy_tx_data(pi);
  821. mpsc_sdma_start_tx(pi); /* start next desc if ready */
  822. }
  823. return rc;
  824. }
  825. /*
  826. * This is the driver's interrupt handler. To avoid a race, we first clear
  827. * the interrupt, then handle any completed Rx/Tx descriptors. When done
  828. * handling those descriptors, we restart the Rx/Tx engines if they're stopped.
  829. */
  830. static irqreturn_t
  831. mpsc_sdma_intr(int irq, void *dev_id, struct pt_regs *regs)
  832. {
  833. struct mpsc_port_info *pi = dev_id;
  834. ulong iflags;
  835. int rc = IRQ_NONE;
  836. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Received\n",pi->port.line);
  837. spin_lock_irqsave(&pi->port.lock, iflags);
  838. mpsc_sdma_intr_ack(pi);
  839. if (mpsc_rx_intr(pi, regs))
  840. rc = IRQ_HANDLED;
  841. if (mpsc_tx_intr(pi))
  842. rc = IRQ_HANDLED;
  843. spin_unlock_irqrestore(&pi->port.lock, iflags);
  844. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Handled\n", pi->port.line);
  845. return rc;
  846. }
  847. /*
  848. ******************************************************************************
  849. *
  850. * serial_core.c Interface routines
  851. *
  852. ******************************************************************************
  853. */
  854. static uint
  855. mpsc_tx_empty(struct uart_port *port)
  856. {
  857. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  858. ulong iflags;
  859. uint rc;
  860. spin_lock_irqsave(&pi->port.lock, iflags);
  861. rc = mpsc_sdma_tx_active(pi) ? 0 : TIOCSER_TEMT;
  862. spin_unlock_irqrestore(&pi->port.lock, iflags);
  863. return rc;
  864. }
  865. static void
  866. mpsc_set_mctrl(struct uart_port *port, uint mctrl)
  867. {
  868. /* Have no way to set modem control lines AFAICT */
  869. return;
  870. }
  871. static uint
  872. mpsc_get_mctrl(struct uart_port *port)
  873. {
  874. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  875. u32 mflags, status;
  876. status = (pi->mirror_regs) ? pi->MPSC_CHR_10_m :
  877. readl(pi->mpsc_base + MPSC_CHR_10);
  878. mflags = 0;
  879. if (status & 0x1)
  880. mflags |= TIOCM_CTS;
  881. if (status & 0x2)
  882. mflags |= TIOCM_CAR;
  883. return mflags | TIOCM_DSR; /* No way to tell if DSR asserted */
  884. }
  885. static void
  886. mpsc_stop_tx(struct uart_port *port)
  887. {
  888. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  889. pr_debug("mpsc_stop_tx[%d]\n", port->line);
  890. mpsc_freeze(pi);
  891. return;
  892. }
  893. static void
  894. mpsc_start_tx(struct uart_port *port)
  895. {
  896. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  897. mpsc_unfreeze(pi);
  898. mpsc_copy_tx_data(pi);
  899. mpsc_sdma_start_tx(pi);
  900. pr_debug("mpsc_start_tx[%d]\n", port->line);
  901. return;
  902. }
  903. static void
  904. mpsc_start_rx(struct mpsc_port_info *pi)
  905. {
  906. pr_debug("mpsc_start_rx[%d]: Starting...\n", pi->port.line);
  907. /* Issue a Receive Abort to clear any receive errors */
  908. writel(MPSC_CHR_2_RA, pi->mpsc_base + MPSC_CHR_2);
  909. if (pi->rcv_data) {
  910. mpsc_enter_hunt(pi);
  911. mpsc_sdma_cmd(pi, SDMA_SDCM_ERD);
  912. }
  913. return;
  914. }
  915. static void
  916. mpsc_stop_rx(struct uart_port *port)
  917. {
  918. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  919. pr_debug("mpsc_stop_rx[%d]: Stopping...\n", port->line);
  920. mpsc_sdma_cmd(pi, SDMA_SDCM_AR);
  921. return;
  922. }
  923. static void
  924. mpsc_enable_ms(struct uart_port *port)
  925. {
  926. return; /* Not supported */
  927. }
  928. static void
  929. mpsc_break_ctl(struct uart_port *port, int ctl)
  930. {
  931. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  932. ulong flags;
  933. u32 v;
  934. v = ctl ? 0x00ff0000 : 0;
  935. spin_lock_irqsave(&pi->port.lock, flags);
  936. if (pi->mirror_regs)
  937. pi->MPSC_CHR_1_m = v;
  938. writel(v, pi->mpsc_base + MPSC_CHR_1);
  939. spin_unlock_irqrestore(&pi->port.lock, flags);
  940. return;
  941. }
  942. static int
  943. mpsc_startup(struct uart_port *port)
  944. {
  945. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  946. u32 flag = 0;
  947. int rc;
  948. pr_debug("mpsc_startup[%d]: Starting up MPSC, irq: %d\n",
  949. port->line, pi->port.irq);
  950. if ((rc = mpsc_make_ready(pi)) == 0) {
  951. /* Setup IRQ handler */
  952. mpsc_sdma_intr_ack(pi);
  953. /* If irq's are shared, need to set flag */
  954. if (mpsc_ports[0].port.irq == mpsc_ports[1].port.irq)
  955. flag = SA_SHIRQ;
  956. if (request_irq(pi->port.irq, mpsc_sdma_intr, flag,
  957. "mpsc/sdma", pi))
  958. printk(KERN_ERR "MPSC: Can't get SDMA IRQ %d\n",
  959. pi->port.irq);
  960. mpsc_sdma_intr_unmask(pi, 0xf);
  961. mpsc_sdma_set_rx_ring(pi, (struct mpsc_rx_desc *)(pi->rxr_p +
  962. (pi->rxr_posn * MPSC_RXRE_SIZE)));
  963. }
  964. return rc;
  965. }
  966. static void
  967. mpsc_shutdown(struct uart_port *port)
  968. {
  969. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  970. pr_debug("mpsc_shutdown[%d]: Shutting down MPSC\n", port->line);
  971. mpsc_sdma_stop(pi);
  972. free_irq(pi->port.irq, pi);
  973. return;
  974. }
  975. static void
  976. mpsc_set_termios(struct uart_port *port, struct termios *termios,
  977. struct termios *old)
  978. {
  979. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  980. u32 baud;
  981. ulong flags;
  982. u32 chr_bits, stop_bits, par;
  983. pi->c_iflag = termios->c_iflag;
  984. pi->c_cflag = termios->c_cflag;
  985. switch (termios->c_cflag & CSIZE) {
  986. case CS5:
  987. chr_bits = MPSC_MPCR_CL_5;
  988. break;
  989. case CS6:
  990. chr_bits = MPSC_MPCR_CL_6;
  991. break;
  992. case CS7:
  993. chr_bits = MPSC_MPCR_CL_7;
  994. break;
  995. case CS8:
  996. default:
  997. chr_bits = MPSC_MPCR_CL_8;
  998. break;
  999. }
  1000. if (termios->c_cflag & CSTOPB)
  1001. stop_bits = MPSC_MPCR_SBL_2;
  1002. else
  1003. stop_bits = MPSC_MPCR_SBL_1;
  1004. par = MPSC_CHR_2_PAR_EVEN;
  1005. if (termios->c_cflag & PARENB)
  1006. if (termios->c_cflag & PARODD)
  1007. par = MPSC_CHR_2_PAR_ODD;
  1008. #ifdef CMSPAR
  1009. if (termios->c_cflag & CMSPAR) {
  1010. if (termios->c_cflag & PARODD)
  1011. par = MPSC_CHR_2_PAR_MARK;
  1012. else
  1013. par = MPSC_CHR_2_PAR_SPACE;
  1014. }
  1015. #endif
  1016. baud = uart_get_baud_rate(port, termios, old, 0, port->uartclk);
  1017. spin_lock_irqsave(&pi->port.lock, flags);
  1018. uart_update_timeout(port, termios->c_cflag, baud);
  1019. mpsc_set_char_length(pi, chr_bits);
  1020. mpsc_set_stop_bit_length(pi, stop_bits);
  1021. mpsc_set_parity(pi, par);
  1022. mpsc_set_baudrate(pi, baud);
  1023. /* Characters/events to read */
  1024. pi->rcv_data = 1;
  1025. pi->port.read_status_mask = SDMA_DESC_CMDSTAT_OR;
  1026. if (termios->c_iflag & INPCK)
  1027. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1028. SDMA_DESC_CMDSTAT_FR;
  1029. if (termios->c_iflag & (BRKINT | PARMRK))
  1030. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1031. /* Characters/events to ignore */
  1032. pi->port.ignore_status_mask = 0;
  1033. if (termios->c_iflag & IGNPAR)
  1034. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1035. SDMA_DESC_CMDSTAT_FR;
  1036. if (termios->c_iflag & IGNBRK) {
  1037. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1038. if (termios->c_iflag & IGNPAR)
  1039. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_OR;
  1040. }
  1041. /* Ignore all chars if CREAD not set */
  1042. if (!(termios->c_cflag & CREAD))
  1043. pi->rcv_data = 0;
  1044. else
  1045. mpsc_start_rx(pi);
  1046. spin_unlock_irqrestore(&pi->port.lock, flags);
  1047. return;
  1048. }
  1049. static const char *
  1050. mpsc_type(struct uart_port *port)
  1051. {
  1052. pr_debug("mpsc_type[%d]: port type: %s\n", port->line,MPSC_DRIVER_NAME);
  1053. return MPSC_DRIVER_NAME;
  1054. }
  1055. static int
  1056. mpsc_request_port(struct uart_port *port)
  1057. {
  1058. /* Should make chip/platform specific call */
  1059. return 0;
  1060. }
  1061. static void
  1062. mpsc_release_port(struct uart_port *port)
  1063. {
  1064. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1065. if (pi->ready) {
  1066. mpsc_uninit_rings(pi);
  1067. mpsc_free_ring_mem(pi);
  1068. pi->ready = 0;
  1069. }
  1070. return;
  1071. }
  1072. static void
  1073. mpsc_config_port(struct uart_port *port, int flags)
  1074. {
  1075. return;
  1076. }
  1077. static int
  1078. mpsc_verify_port(struct uart_port *port, struct serial_struct *ser)
  1079. {
  1080. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1081. int rc = 0;
  1082. pr_debug("mpsc_verify_port[%d]: Verifying port data\n", pi->port.line);
  1083. if (ser->type != PORT_UNKNOWN && ser->type != PORT_MPSC)
  1084. rc = -EINVAL;
  1085. else if (pi->port.irq != ser->irq)
  1086. rc = -EINVAL;
  1087. else if (ser->io_type != SERIAL_IO_MEM)
  1088. rc = -EINVAL;
  1089. else if (pi->port.uartclk / 16 != ser->baud_base) /* Not sure */
  1090. rc = -EINVAL;
  1091. else if ((void *)pi->port.mapbase != ser->iomem_base)
  1092. rc = -EINVAL;
  1093. else if (pi->port.iobase != ser->port)
  1094. rc = -EINVAL;
  1095. else if (ser->hub6 != 0)
  1096. rc = -EINVAL;
  1097. return rc;
  1098. }
  1099. static struct uart_ops mpsc_pops = {
  1100. .tx_empty = mpsc_tx_empty,
  1101. .set_mctrl = mpsc_set_mctrl,
  1102. .get_mctrl = mpsc_get_mctrl,
  1103. .stop_tx = mpsc_stop_tx,
  1104. .start_tx = mpsc_start_tx,
  1105. .stop_rx = mpsc_stop_rx,
  1106. .enable_ms = mpsc_enable_ms,
  1107. .break_ctl = mpsc_break_ctl,
  1108. .startup = mpsc_startup,
  1109. .shutdown = mpsc_shutdown,
  1110. .set_termios = mpsc_set_termios,
  1111. .type = mpsc_type,
  1112. .release_port = mpsc_release_port,
  1113. .request_port = mpsc_request_port,
  1114. .config_port = mpsc_config_port,
  1115. .verify_port = mpsc_verify_port,
  1116. };
  1117. /*
  1118. ******************************************************************************
  1119. *
  1120. * Console Interface Routines
  1121. *
  1122. ******************************************************************************
  1123. */
  1124. #ifdef CONFIG_SERIAL_MPSC_CONSOLE
  1125. static void
  1126. mpsc_console_write(struct console *co, const char *s, uint count)
  1127. {
  1128. struct mpsc_port_info *pi = &mpsc_ports[co->index];
  1129. u8 *bp, *dp, add_cr = 0;
  1130. int i;
  1131. while (mpsc_sdma_tx_active(pi))
  1132. udelay(100);
  1133. while (count > 0) {
  1134. bp = dp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  1135. for (i = 0; i < MPSC_TXBE_SIZE; i++) {
  1136. if (count == 0)
  1137. break;
  1138. if (add_cr) {
  1139. *(dp++) = '\r';
  1140. add_cr = 0;
  1141. }
  1142. else {
  1143. *(dp++) = *s;
  1144. if (*(s++) == '\n') { /* add '\r' after '\n' */
  1145. add_cr = 1;
  1146. count++;
  1147. }
  1148. }
  1149. count--;
  1150. }
  1151. dma_cache_sync((void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  1152. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1153. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1154. flush_dcache_range((ulong)bp,
  1155. (ulong)bp + MPSC_TXBE_SIZE);
  1156. #endif
  1157. mpsc_setup_tx_desc(pi, i, 0);
  1158. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1159. mpsc_sdma_start_tx(pi);
  1160. while (mpsc_sdma_tx_active(pi))
  1161. udelay(100);
  1162. pi->txr_tail = (pi->txr_tail + 1) & (MPSC_TXR_ENTRIES - 1);
  1163. }
  1164. return;
  1165. }
  1166. static int __init
  1167. mpsc_console_setup(struct console *co, char *options)
  1168. {
  1169. struct mpsc_port_info *pi;
  1170. int baud, bits, parity, flow;
  1171. pr_debug("mpsc_console_setup[%d]: options: %s\n", co->index, options);
  1172. if (co->index >= MPSC_NUM_CTLRS)
  1173. co->index = 0;
  1174. pi = &mpsc_ports[co->index];
  1175. baud = pi->default_baud;
  1176. bits = pi->default_bits;
  1177. parity = pi->default_parity;
  1178. flow = pi->default_flow;
  1179. if (!pi->port.ops)
  1180. return -ENODEV;
  1181. spin_lock_init(&pi->port.lock); /* Temporary fix--copied from 8250.c */
  1182. if (options)
  1183. uart_parse_options(options, &baud, &parity, &bits, &flow);
  1184. return uart_set_options(&pi->port, co, baud, parity, bits, flow);
  1185. }
  1186. static struct console mpsc_console = {
  1187. .name = MPSC_DEV_NAME,
  1188. .write = mpsc_console_write,
  1189. .device = uart_console_device,
  1190. .setup = mpsc_console_setup,
  1191. .flags = CON_PRINTBUFFER,
  1192. .index = -1,
  1193. .data = &mpsc_reg,
  1194. };
  1195. static int __init
  1196. mpsc_late_console_init(void)
  1197. {
  1198. pr_debug("mpsc_late_console_init: Enter\n");
  1199. if (!(mpsc_console.flags & CON_ENABLED))
  1200. register_console(&mpsc_console);
  1201. return 0;
  1202. }
  1203. late_initcall(mpsc_late_console_init);
  1204. #define MPSC_CONSOLE &mpsc_console
  1205. #else
  1206. #define MPSC_CONSOLE NULL
  1207. #endif
  1208. /*
  1209. ******************************************************************************
  1210. *
  1211. * Dummy Platform Driver to extract & map shared register regions
  1212. *
  1213. ******************************************************************************
  1214. */
  1215. static void
  1216. mpsc_resource_err(char *s)
  1217. {
  1218. printk(KERN_WARNING "MPSC: Platform device resource error in %s\n", s);
  1219. return;
  1220. }
  1221. static int
  1222. mpsc_shared_map_regs(struct platform_device *pd)
  1223. {
  1224. struct resource *r;
  1225. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1226. MPSC_ROUTING_BASE_ORDER)) && request_mem_region(r->start,
  1227. MPSC_ROUTING_REG_BLOCK_SIZE, "mpsc_routing_regs")) {
  1228. mpsc_shared_regs.mpsc_routing_base = ioremap(r->start,
  1229. MPSC_ROUTING_REG_BLOCK_SIZE);
  1230. mpsc_shared_regs.mpsc_routing_base_p = r->start;
  1231. }
  1232. else {
  1233. mpsc_resource_err("MPSC routing base");
  1234. return -ENOMEM;
  1235. }
  1236. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1237. MPSC_SDMA_INTR_BASE_ORDER)) && request_mem_region(r->start,
  1238. MPSC_SDMA_INTR_REG_BLOCK_SIZE, "sdma_intr_regs")) {
  1239. mpsc_shared_regs.sdma_intr_base = ioremap(r->start,
  1240. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1241. mpsc_shared_regs.sdma_intr_base_p = r->start;
  1242. }
  1243. else {
  1244. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1245. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1246. MPSC_ROUTING_REG_BLOCK_SIZE);
  1247. mpsc_resource_err("SDMA intr base");
  1248. return -ENOMEM;
  1249. }
  1250. return 0;
  1251. }
  1252. static void
  1253. mpsc_shared_unmap_regs(void)
  1254. {
  1255. if (!mpsc_shared_regs.mpsc_routing_base) {
  1256. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1257. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1258. MPSC_ROUTING_REG_BLOCK_SIZE);
  1259. }
  1260. if (!mpsc_shared_regs.sdma_intr_base) {
  1261. iounmap(mpsc_shared_regs.sdma_intr_base);
  1262. release_mem_region(mpsc_shared_regs.sdma_intr_base_p,
  1263. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1264. }
  1265. mpsc_shared_regs.mpsc_routing_base = NULL;
  1266. mpsc_shared_regs.sdma_intr_base = NULL;
  1267. mpsc_shared_regs.mpsc_routing_base_p = 0;
  1268. mpsc_shared_regs.sdma_intr_base_p = 0;
  1269. return;
  1270. }
  1271. static int
  1272. mpsc_shared_drv_probe(struct device *dev)
  1273. {
  1274. struct platform_device *pd = to_platform_device(dev);
  1275. struct mpsc_shared_pdata *pdata;
  1276. int rc = -ENODEV;
  1277. if (pd->id == 0) {
  1278. if (!(rc = mpsc_shared_map_regs(pd))) {
  1279. pdata = (struct mpsc_shared_pdata *)dev->platform_data;
  1280. mpsc_shared_regs.MPSC_MRR_m = pdata->mrr_val;
  1281. mpsc_shared_regs.MPSC_RCRR_m= pdata->rcrr_val;
  1282. mpsc_shared_regs.MPSC_TCRR_m= pdata->tcrr_val;
  1283. mpsc_shared_regs.SDMA_INTR_CAUSE_m =
  1284. pdata->intr_cause_val;
  1285. mpsc_shared_regs.SDMA_INTR_MASK_m =
  1286. pdata->intr_mask_val;
  1287. rc = 0;
  1288. }
  1289. }
  1290. return rc;
  1291. }
  1292. static int
  1293. mpsc_shared_drv_remove(struct device *dev)
  1294. {
  1295. struct platform_device *pd = to_platform_device(dev);
  1296. int rc = -ENODEV;
  1297. if (pd->id == 0) {
  1298. mpsc_shared_unmap_regs();
  1299. mpsc_shared_regs.MPSC_MRR_m = 0;
  1300. mpsc_shared_regs.MPSC_RCRR_m = 0;
  1301. mpsc_shared_regs.MPSC_TCRR_m = 0;
  1302. mpsc_shared_regs.SDMA_INTR_CAUSE_m = 0;
  1303. mpsc_shared_regs.SDMA_INTR_MASK_m = 0;
  1304. rc = 0;
  1305. }
  1306. return rc;
  1307. }
  1308. static struct device_driver mpsc_shared_driver = {
  1309. .name = MPSC_SHARED_NAME,
  1310. .bus = &platform_bus_type,
  1311. .probe = mpsc_shared_drv_probe,
  1312. .remove = mpsc_shared_drv_remove,
  1313. };
  1314. /*
  1315. ******************************************************************************
  1316. *
  1317. * Driver Interface Routines
  1318. *
  1319. ******************************************************************************
  1320. */
  1321. static struct uart_driver mpsc_reg = {
  1322. .owner = THIS_MODULE,
  1323. .driver_name = MPSC_DRIVER_NAME,
  1324. .devfs_name = MPSC_DEVFS_NAME,
  1325. .dev_name = MPSC_DEV_NAME,
  1326. .major = MPSC_MAJOR,
  1327. .minor = MPSC_MINOR_START,
  1328. .nr = MPSC_NUM_CTLRS,
  1329. .cons = MPSC_CONSOLE,
  1330. };
  1331. static int
  1332. mpsc_drv_map_regs(struct mpsc_port_info *pi, struct platform_device *pd)
  1333. {
  1334. struct resource *r;
  1335. if ((r = platform_get_resource(pd, IORESOURCE_MEM, MPSC_BASE_ORDER)) &&
  1336. request_mem_region(r->start, MPSC_REG_BLOCK_SIZE, "mpsc_regs")){
  1337. pi->mpsc_base = ioremap(r->start, MPSC_REG_BLOCK_SIZE);
  1338. pi->mpsc_base_p = r->start;
  1339. }
  1340. else {
  1341. mpsc_resource_err("MPSC base");
  1342. return -ENOMEM;
  1343. }
  1344. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1345. MPSC_SDMA_BASE_ORDER)) && request_mem_region(r->start,
  1346. MPSC_SDMA_REG_BLOCK_SIZE, "sdma_regs")) {
  1347. pi->sdma_base = ioremap(r->start,MPSC_SDMA_REG_BLOCK_SIZE);
  1348. pi->sdma_base_p = r->start;
  1349. }
  1350. else {
  1351. mpsc_resource_err("SDMA base");
  1352. return -ENOMEM;
  1353. }
  1354. if ((r = platform_get_resource(pd,IORESOURCE_MEM,MPSC_BRG_BASE_ORDER))
  1355. && request_mem_region(r->start, MPSC_BRG_REG_BLOCK_SIZE,
  1356. "brg_regs")) {
  1357. pi->brg_base = ioremap(r->start, MPSC_BRG_REG_BLOCK_SIZE);
  1358. pi->brg_base_p = r->start;
  1359. }
  1360. else {
  1361. mpsc_resource_err("BRG base");
  1362. return -ENOMEM;
  1363. }
  1364. return 0;
  1365. }
  1366. static void
  1367. mpsc_drv_unmap_regs(struct mpsc_port_info *pi)
  1368. {
  1369. if (!pi->mpsc_base) {
  1370. iounmap(pi->mpsc_base);
  1371. release_mem_region(pi->mpsc_base_p, MPSC_REG_BLOCK_SIZE);
  1372. }
  1373. if (!pi->sdma_base) {
  1374. iounmap(pi->sdma_base);
  1375. release_mem_region(pi->sdma_base_p, MPSC_SDMA_REG_BLOCK_SIZE);
  1376. }
  1377. if (!pi->brg_base) {
  1378. iounmap(pi->brg_base);
  1379. release_mem_region(pi->brg_base_p, MPSC_BRG_REG_BLOCK_SIZE);
  1380. }
  1381. pi->mpsc_base = NULL;
  1382. pi->sdma_base = NULL;
  1383. pi->brg_base = NULL;
  1384. pi->mpsc_base_p = 0;
  1385. pi->sdma_base_p = 0;
  1386. pi->brg_base_p = 0;
  1387. return;
  1388. }
  1389. static void
  1390. mpsc_drv_get_platform_data(struct mpsc_port_info *pi,
  1391. struct platform_device *pd, int num)
  1392. {
  1393. struct mpsc_pdata *pdata;
  1394. pdata = (struct mpsc_pdata *)pd->dev.platform_data;
  1395. pi->port.uartclk = pdata->brg_clk_freq;
  1396. pi->port.iotype = UPIO_MEM;
  1397. pi->port.line = num;
  1398. pi->port.type = PORT_MPSC;
  1399. pi->port.fifosize = MPSC_TXBE_SIZE;
  1400. pi->port.membase = pi->mpsc_base;
  1401. pi->port.mapbase = (ulong)pi->mpsc_base;
  1402. pi->port.ops = &mpsc_pops;
  1403. pi->mirror_regs = pdata->mirror_regs;
  1404. pi->cache_mgmt = pdata->cache_mgmt;
  1405. pi->brg_can_tune = pdata->brg_can_tune;
  1406. pi->brg_clk_src = pdata->brg_clk_src;
  1407. pi->mpsc_max_idle = pdata->max_idle;
  1408. pi->default_baud = pdata->default_baud;
  1409. pi->default_bits = pdata->default_bits;
  1410. pi->default_parity = pdata->default_parity;
  1411. pi->default_flow = pdata->default_flow;
  1412. /* Initial values of mirrored regs */
  1413. pi->MPSC_CHR_1_m = pdata->chr_1_val;
  1414. pi->MPSC_CHR_2_m = pdata->chr_2_val;
  1415. pi->MPSC_CHR_10_m = pdata->chr_10_val;
  1416. pi->MPSC_MPCR_m = pdata->mpcr_val;
  1417. pi->BRG_BCR_m = pdata->bcr_val;
  1418. pi->shared_regs = &mpsc_shared_regs;
  1419. pi->port.irq = platform_get_irq(pd, 0);
  1420. return;
  1421. }
  1422. static int
  1423. mpsc_drv_probe(struct device *dev)
  1424. {
  1425. struct platform_device *pd = to_platform_device(dev);
  1426. struct mpsc_port_info *pi;
  1427. int rc = -ENODEV;
  1428. pr_debug("mpsc_drv_probe: Adding MPSC %d\n", pd->id);
  1429. if (pd->id < MPSC_NUM_CTLRS) {
  1430. pi = &mpsc_ports[pd->id];
  1431. if (!(rc = mpsc_drv_map_regs(pi, pd))) {
  1432. mpsc_drv_get_platform_data(pi, pd, pd->id);
  1433. if (!(rc = mpsc_make_ready(pi)))
  1434. if (!(rc = uart_add_one_port(&mpsc_reg,
  1435. &pi->port)))
  1436. rc = 0;
  1437. else {
  1438. mpsc_release_port(
  1439. (struct uart_port *)pi);
  1440. mpsc_drv_unmap_regs(pi);
  1441. }
  1442. else
  1443. mpsc_drv_unmap_regs(pi);
  1444. }
  1445. }
  1446. return rc;
  1447. }
  1448. static int
  1449. mpsc_drv_remove(struct device *dev)
  1450. {
  1451. struct platform_device *pd = to_platform_device(dev);
  1452. pr_debug("mpsc_drv_exit: Removing MPSC %d\n", pd->id);
  1453. if (pd->id < MPSC_NUM_CTLRS) {
  1454. uart_remove_one_port(&mpsc_reg, &mpsc_ports[pd->id].port);
  1455. mpsc_release_port((struct uart_port *)&mpsc_ports[pd->id].port);
  1456. mpsc_drv_unmap_regs(&mpsc_ports[pd->id]);
  1457. return 0;
  1458. }
  1459. else
  1460. return -ENODEV;
  1461. }
  1462. static struct device_driver mpsc_driver = {
  1463. .name = MPSC_CTLR_NAME,
  1464. .bus = &platform_bus_type,
  1465. .probe = mpsc_drv_probe,
  1466. .remove = mpsc_drv_remove,
  1467. };
  1468. static int __init
  1469. mpsc_drv_init(void)
  1470. {
  1471. int rc;
  1472. printk(KERN_INFO "Serial: MPSC driver $Revision: 1.00 $\n");
  1473. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1474. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1475. if (!(rc = uart_register_driver(&mpsc_reg))) {
  1476. if (!(rc = driver_register(&mpsc_shared_driver))) {
  1477. if ((rc = driver_register(&mpsc_driver))) {
  1478. driver_unregister(&mpsc_shared_driver);
  1479. uart_unregister_driver(&mpsc_reg);
  1480. }
  1481. }
  1482. else
  1483. uart_unregister_driver(&mpsc_reg);
  1484. }
  1485. return rc;
  1486. }
  1487. static void __exit
  1488. mpsc_drv_exit(void)
  1489. {
  1490. driver_unregister(&mpsc_driver);
  1491. driver_unregister(&mpsc_shared_driver);
  1492. uart_unregister_driver(&mpsc_reg);
  1493. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1494. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1495. return;
  1496. }
  1497. module_init(mpsc_drv_init);
  1498. module_exit(mpsc_drv_exit);
  1499. MODULE_AUTHOR("Mark A. Greer <mgreer@mvista.com>");
  1500. MODULE_DESCRIPTION("Generic Marvell MPSC serial/UART driver $Revision: 1.00 $");
  1501. MODULE_VERSION(MPSC_VERSION);
  1502. MODULE_LICENSE("GPL");
  1503. MODULE_ALIAS_CHARDEV_MAJOR(MPSC_MAJOR);