mpsc.c 44 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832
  1. /*
  2. * drivers/serial/mpsc.c
  3. *
  4. * Generic driver for the MPSC (UART mode) on Marvell parts (e.g., GT64240,
  5. * GT64260, MV64340, MV64360, GT96100, ... ).
  6. *
  7. * Author: Mark A. Greer <mgreer@mvista.com>
  8. *
  9. * Based on an old MPSC driver that was in the linuxppc tree. It appears to
  10. * have been created by Chris Zankel (formerly of MontaVista) but there
  11. * is no proper Copyright so I'm not sure. Apparently, parts were also
  12. * taken from PPCBoot (now U-Boot). Also based on drivers/serial/8250.c
  13. * by Russell King.
  14. *
  15. * 2004 (c) MontaVista, Software, Inc. This file is licensed under
  16. * the terms of the GNU General Public License version 2. This program
  17. * is licensed "as is" without any warranty of any kind, whether express
  18. * or implied.
  19. */
  20. /*
  21. * The MPSC interface is much like a typical network controller's interface.
  22. * That is, you set up separate rings of descriptors for transmitting and
  23. * receiving data. There is also a pool of buffers with (one buffer per
  24. * descriptor) that incoming data are dma'd into or outgoing data are dma'd
  25. * out of.
  26. *
  27. * The MPSC requires two other controllers to be able to work. The Baud Rate
  28. * Generator (BRG) provides a clock at programmable frequencies which determines
  29. * the baud rate. The Serial DMA Controller (SDMA) takes incoming data from the
  30. * MPSC and DMA's it into memory or DMA's outgoing data and passes it to the
  31. * MPSC. It is actually the SDMA interrupt that the driver uses to keep the
  32. * transmit and receive "engines" going (i.e., indicate data has been
  33. * transmitted or received).
  34. *
  35. * NOTES:
  36. *
  37. * 1) Some chips have an erratum where several regs cannot be
  38. * read. To work around that, we keep a local copy of those regs in
  39. * 'mpsc_port_info'.
  40. *
  41. * 2) Some chips have an erratum where the ctlr will hang when the SDMA ctlr
  42. * accesses system mem with coherency enabled. For that reason, the driver
  43. * assumes that coherency for that ctlr has been disabled. This means
  44. * that when in a cache coherent system, the driver has to manually manage
  45. * the data cache on the areas that it touches because the dma_* macro are
  46. * basically no-ops.
  47. *
  48. * 3) There is an erratum (on PPC) where you can't use the instruction to do
  49. * a DMA_TO_DEVICE/cache clean so DMA_BIDIRECTIONAL/flushes are used in places
  50. * where a DMA_TO_DEVICE/clean would have [otherwise] sufficed.
  51. *
  52. * 4) AFAICT, hardware flow control isn't supported by the controller --MAG.
  53. */
  54. #include "mpsc.h"
  55. /*
  56. * Define how this driver is known to the outside (we've been assigned a
  57. * range on the "Low-density serial ports" major).
  58. */
  59. #define MPSC_MAJOR 204
  60. #define MPSC_MINOR_START 44
  61. #define MPSC_DRIVER_NAME "MPSC"
  62. #define MPSC_DEVFS_NAME "ttymm/"
  63. #define MPSC_DEV_NAME "ttyMM"
  64. #define MPSC_VERSION "1.00"
  65. static struct mpsc_port_info mpsc_ports[MPSC_NUM_CTLRS];
  66. static struct mpsc_shared_regs mpsc_shared_regs;
  67. /*
  68. ******************************************************************************
  69. *
  70. * Baud Rate Generator Routines (BRG)
  71. *
  72. ******************************************************************************
  73. */
  74. static void
  75. mpsc_brg_init(struct mpsc_port_info *pi, u32 clk_src)
  76. {
  77. u32 v;
  78. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  79. v = (v & ~(0xf << 18)) | ((clk_src & 0xf) << 18);
  80. if (pi->brg_can_tune)
  81. v &= ~(1 << 25);
  82. if (pi->mirror_regs)
  83. pi->BRG_BCR_m = v;
  84. writel(v, pi->brg_base + BRG_BCR);
  85. writel(readl(pi->brg_base + BRG_BTR) & 0xffff0000,
  86. pi->brg_base + BRG_BTR);
  87. return;
  88. }
  89. static void
  90. mpsc_brg_enable(struct mpsc_port_info *pi)
  91. {
  92. u32 v;
  93. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  94. v |= (1 << 16);
  95. if (pi->mirror_regs)
  96. pi->BRG_BCR_m = v;
  97. writel(v, pi->brg_base + BRG_BCR);
  98. return;
  99. }
  100. static void
  101. mpsc_brg_disable(struct mpsc_port_info *pi)
  102. {
  103. u32 v;
  104. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  105. v &= ~(1 << 16);
  106. if (pi->mirror_regs)
  107. pi->BRG_BCR_m = v;
  108. writel(v, pi->brg_base + BRG_BCR);
  109. return;
  110. }
  111. static inline void
  112. mpsc_set_baudrate(struct mpsc_port_info *pi, u32 baud)
  113. {
  114. /*
  115. * To set the baud, we adjust the CDV field in the BRG_BCR reg.
  116. * From manual: Baud = clk / ((CDV+1)*2) ==> CDV = (clk / (baud*2)) - 1.
  117. * However, the input clock is divided by 16 in the MPSC b/c of how
  118. * 'MPSC_MMCRH' was set up so we have to divide the 'clk' used in our
  119. * calculation by 16 to account for that. So the real calculation
  120. * that accounts for the way the mpsc is set up is:
  121. * CDV = (clk / (baud*2*16)) - 1 ==> CDV = (clk / (baud << 5)) - 1.
  122. */
  123. u32 cdv = (pi->port.uartclk / (baud << 5)) - 1;
  124. u32 v;
  125. mpsc_brg_disable(pi);
  126. v = (pi->mirror_regs) ? pi->BRG_BCR_m : readl(pi->brg_base + BRG_BCR);
  127. v = (v & 0xffff0000) | (cdv & 0xffff);
  128. if (pi->mirror_regs)
  129. pi->BRG_BCR_m = v;
  130. writel(v, pi->brg_base + BRG_BCR);
  131. mpsc_brg_enable(pi);
  132. return;
  133. }
  134. /*
  135. ******************************************************************************
  136. *
  137. * Serial DMA Routines (SDMA)
  138. *
  139. ******************************************************************************
  140. */
  141. static void
  142. mpsc_sdma_burstsize(struct mpsc_port_info *pi, u32 burst_size)
  143. {
  144. u32 v;
  145. pr_debug("mpsc_sdma_burstsize[%d]: burst_size: %d\n",
  146. pi->port.line, burst_size);
  147. burst_size >>= 3; /* Divide by 8 b/c reg values are 8-byte chunks */
  148. if (burst_size < 2)
  149. v = 0x0; /* 1 64-bit word */
  150. else if (burst_size < 4)
  151. v = 0x1; /* 2 64-bit words */
  152. else if (burst_size < 8)
  153. v = 0x2; /* 4 64-bit words */
  154. else
  155. v = 0x3; /* 8 64-bit words */
  156. writel((readl(pi->sdma_base + SDMA_SDC) & (0x3 << 12)) | (v << 12),
  157. pi->sdma_base + SDMA_SDC);
  158. return;
  159. }
  160. static void
  161. mpsc_sdma_init(struct mpsc_port_info *pi, u32 burst_size)
  162. {
  163. pr_debug("mpsc_sdma_init[%d]: burst_size: %d\n", pi->port.line,
  164. burst_size);
  165. writel((readl(pi->sdma_base + SDMA_SDC) & 0x3ff) | 0x03f,
  166. pi->sdma_base + SDMA_SDC);
  167. mpsc_sdma_burstsize(pi, burst_size);
  168. return;
  169. }
  170. static inline u32
  171. mpsc_sdma_intr_mask(struct mpsc_port_info *pi, u32 mask)
  172. {
  173. u32 old, v;
  174. pr_debug("mpsc_sdma_intr_mask[%d]: mask: 0x%x\n", pi->port.line, mask);
  175. old = v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  176. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  177. mask &= 0xf;
  178. if (pi->port.line)
  179. mask <<= 8;
  180. v &= ~mask;
  181. if (pi->mirror_regs)
  182. pi->shared_regs->SDMA_INTR_MASK_m = v;
  183. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  184. if (pi->port.line)
  185. old >>= 8;
  186. return old & 0xf;
  187. }
  188. static inline void
  189. mpsc_sdma_intr_unmask(struct mpsc_port_info *pi, u32 mask)
  190. {
  191. u32 v;
  192. pr_debug("mpsc_sdma_intr_unmask[%d]: mask: 0x%x\n", pi->port.line,mask);
  193. v = (pi->mirror_regs) ? pi->shared_regs->SDMA_INTR_MASK_m :
  194. readl(pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  195. mask &= 0xf;
  196. if (pi->port.line)
  197. mask <<= 8;
  198. v |= mask;
  199. if (pi->mirror_regs)
  200. pi->shared_regs->SDMA_INTR_MASK_m = v;
  201. writel(v, pi->shared_regs->sdma_intr_base + SDMA_INTR_MASK);
  202. return;
  203. }
  204. static inline void
  205. mpsc_sdma_intr_ack(struct mpsc_port_info *pi)
  206. {
  207. pr_debug("mpsc_sdma_intr_ack[%d]: Acknowledging IRQ\n", pi->port.line);
  208. if (pi->mirror_regs)
  209. pi->shared_regs->SDMA_INTR_CAUSE_m = 0;
  210. writel(0, pi->shared_regs->sdma_intr_base + SDMA_INTR_CAUSE);
  211. return;
  212. }
  213. static inline void
  214. mpsc_sdma_set_rx_ring(struct mpsc_port_info *pi, struct mpsc_rx_desc *rxre_p)
  215. {
  216. pr_debug("mpsc_sdma_set_rx_ring[%d]: rxre_p: 0x%x\n",
  217. pi->port.line, (u32) rxre_p);
  218. writel((u32)rxre_p, pi->sdma_base + SDMA_SCRDP);
  219. return;
  220. }
  221. static inline void
  222. mpsc_sdma_set_tx_ring(struct mpsc_port_info *pi, struct mpsc_tx_desc *txre_p)
  223. {
  224. writel((u32)txre_p, pi->sdma_base + SDMA_SFTDP);
  225. writel((u32)txre_p, pi->sdma_base + SDMA_SCTDP);
  226. return;
  227. }
  228. static inline void
  229. mpsc_sdma_cmd(struct mpsc_port_info *pi, u32 val)
  230. {
  231. u32 v;
  232. v = readl(pi->sdma_base + SDMA_SDCM);
  233. if (val)
  234. v |= val;
  235. else
  236. v = 0;
  237. wmb();
  238. writel(v, pi->sdma_base + SDMA_SDCM);
  239. wmb();
  240. return;
  241. }
  242. static inline uint
  243. mpsc_sdma_tx_active(struct mpsc_port_info *pi)
  244. {
  245. return readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_TXD;
  246. }
  247. static inline void
  248. mpsc_sdma_start_tx(struct mpsc_port_info *pi)
  249. {
  250. struct mpsc_tx_desc *txre, *txre_p;
  251. /* If tx isn't running & there's a desc ready to go, start it */
  252. if (!mpsc_sdma_tx_active(pi)) {
  253. txre = (struct mpsc_tx_desc *)(pi->txr +
  254. (pi->txr_tail * MPSC_TXRE_SIZE));
  255. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  256. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  257. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  258. invalidate_dcache_range((ulong)txre,
  259. (ulong)txre + MPSC_TXRE_SIZE);
  260. #endif
  261. if (be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O) {
  262. txre_p = (struct mpsc_tx_desc *)(pi->txr_p +
  263. (pi->txr_tail *
  264. MPSC_TXRE_SIZE));
  265. mpsc_sdma_set_tx_ring(pi, txre_p);
  266. mpsc_sdma_cmd(pi, SDMA_SDCM_STD | SDMA_SDCM_TXD);
  267. }
  268. }
  269. return;
  270. }
  271. static inline void
  272. mpsc_sdma_stop(struct mpsc_port_info *pi)
  273. {
  274. pr_debug("mpsc_sdma_stop[%d]: Stopping SDMA\n", pi->port.line);
  275. /* Abort any SDMA transfers */
  276. mpsc_sdma_cmd(pi, 0);
  277. mpsc_sdma_cmd(pi, SDMA_SDCM_AR | SDMA_SDCM_AT);
  278. /* Clear the SDMA current and first TX and RX pointers */
  279. mpsc_sdma_set_tx_ring(pi, NULL);
  280. mpsc_sdma_set_rx_ring(pi, NULL);
  281. /* Disable interrupts */
  282. mpsc_sdma_intr_mask(pi, 0xf);
  283. mpsc_sdma_intr_ack(pi);
  284. return;
  285. }
  286. /*
  287. ******************************************************************************
  288. *
  289. * Multi-Protocol Serial Controller Routines (MPSC)
  290. *
  291. ******************************************************************************
  292. */
  293. static void
  294. mpsc_hw_init(struct mpsc_port_info *pi)
  295. {
  296. u32 v;
  297. pr_debug("mpsc_hw_init[%d]: Initializing hardware\n", pi->port.line);
  298. /* Set up clock routing */
  299. if (pi->mirror_regs) {
  300. v = pi->shared_regs->MPSC_MRR_m;
  301. v &= ~0x1c7;
  302. pi->shared_regs->MPSC_MRR_m = v;
  303. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  304. v = pi->shared_regs->MPSC_RCRR_m;
  305. v = (v & ~0xf0f) | 0x100;
  306. pi->shared_regs->MPSC_RCRR_m = v;
  307. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  308. v = pi->shared_regs->MPSC_TCRR_m;
  309. v = (v & ~0xf0f) | 0x100;
  310. pi->shared_regs->MPSC_TCRR_m = v;
  311. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  312. }
  313. else {
  314. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  315. v &= ~0x1c7;
  316. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_MRR);
  317. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  318. v = (v & ~0xf0f) | 0x100;
  319. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_RCRR);
  320. v = readl(pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  321. v = (v & ~0xf0f) | 0x100;
  322. writel(v, pi->shared_regs->mpsc_routing_base + MPSC_TCRR);
  323. }
  324. /* Put MPSC in UART mode & enabel Tx/Rx egines */
  325. writel(0x000004c4, pi->mpsc_base + MPSC_MMCRL);
  326. /* No preamble, 16x divider, low-latency, */
  327. writel(0x04400400, pi->mpsc_base + MPSC_MMCRH);
  328. if (pi->mirror_regs) {
  329. pi->MPSC_CHR_1_m = 0;
  330. pi->MPSC_CHR_2_m = 0;
  331. }
  332. writel(0, pi->mpsc_base + MPSC_CHR_1);
  333. writel(0, pi->mpsc_base + MPSC_CHR_2);
  334. writel(pi->mpsc_max_idle, pi->mpsc_base + MPSC_CHR_3);
  335. writel(0, pi->mpsc_base + MPSC_CHR_4);
  336. writel(0, pi->mpsc_base + MPSC_CHR_5);
  337. writel(0, pi->mpsc_base + MPSC_CHR_6);
  338. writel(0, pi->mpsc_base + MPSC_CHR_7);
  339. writel(0, pi->mpsc_base + MPSC_CHR_8);
  340. writel(0, pi->mpsc_base + MPSC_CHR_9);
  341. writel(0, pi->mpsc_base + MPSC_CHR_10);
  342. return;
  343. }
  344. static inline void
  345. mpsc_enter_hunt(struct mpsc_port_info *pi)
  346. {
  347. pr_debug("mpsc_enter_hunt[%d]: Hunting...\n", pi->port.line);
  348. if (pi->mirror_regs) {
  349. writel(pi->MPSC_CHR_2_m | MPSC_CHR_2_EH,
  350. pi->mpsc_base + MPSC_CHR_2);
  351. /* Erratum prevents reading CHR_2 so just delay for a while */
  352. udelay(100);
  353. }
  354. else {
  355. writel(readl(pi->mpsc_base + MPSC_CHR_2) | MPSC_CHR_2_EH,
  356. pi->mpsc_base + MPSC_CHR_2);
  357. while (readl(pi->mpsc_base + MPSC_CHR_2) & MPSC_CHR_2_EH)
  358. udelay(10);
  359. }
  360. return;
  361. }
  362. static inline void
  363. mpsc_freeze(struct mpsc_port_info *pi)
  364. {
  365. u32 v;
  366. pr_debug("mpsc_freeze[%d]: Freezing\n", pi->port.line);
  367. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  368. readl(pi->mpsc_base + MPSC_MPCR);
  369. v |= MPSC_MPCR_FRZ;
  370. if (pi->mirror_regs)
  371. pi->MPSC_MPCR_m = v;
  372. writel(v, pi->mpsc_base + MPSC_MPCR);
  373. return;
  374. }
  375. static inline void
  376. mpsc_unfreeze(struct mpsc_port_info *pi)
  377. {
  378. u32 v;
  379. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  380. readl(pi->mpsc_base + MPSC_MPCR);
  381. v &= ~MPSC_MPCR_FRZ;
  382. if (pi->mirror_regs)
  383. pi->MPSC_MPCR_m = v;
  384. writel(v, pi->mpsc_base + MPSC_MPCR);
  385. pr_debug("mpsc_unfreeze[%d]: Unfrozen\n", pi->port.line);
  386. return;
  387. }
  388. static inline void
  389. mpsc_set_char_length(struct mpsc_port_info *pi, u32 len)
  390. {
  391. u32 v;
  392. pr_debug("mpsc_set_char_length[%d]: char len: %d\n", pi->port.line,len);
  393. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  394. readl(pi->mpsc_base + MPSC_MPCR);
  395. v = (v & ~(0x3 << 12)) | ((len & 0x3) << 12);
  396. if (pi->mirror_regs)
  397. pi->MPSC_MPCR_m = v;
  398. writel(v, pi->mpsc_base + MPSC_MPCR);
  399. return;
  400. }
  401. static inline void
  402. mpsc_set_stop_bit_length(struct mpsc_port_info *pi, u32 len)
  403. {
  404. u32 v;
  405. pr_debug("mpsc_set_stop_bit_length[%d]: stop bits: %d\n",
  406. pi->port.line, len);
  407. v = (pi->mirror_regs) ? pi->MPSC_MPCR_m :
  408. readl(pi->mpsc_base + MPSC_MPCR);
  409. v = (v & ~(1 << 14)) | ((len & 0x1) << 14);
  410. if (pi->mirror_regs)
  411. pi->MPSC_MPCR_m = v;
  412. writel(v, pi->mpsc_base + MPSC_MPCR);
  413. return;
  414. }
  415. static inline void
  416. mpsc_set_parity(struct mpsc_port_info *pi, u32 p)
  417. {
  418. u32 v;
  419. pr_debug("mpsc_set_parity[%d]: parity bits: 0x%x\n", pi->port.line, p);
  420. v = (pi->mirror_regs) ? pi->MPSC_CHR_2_m :
  421. readl(pi->mpsc_base + MPSC_CHR_2);
  422. p &= 0x3;
  423. v = (v & ~0xc000c) | (p << 18) | (p << 2);
  424. if (pi->mirror_regs)
  425. pi->MPSC_CHR_2_m = v;
  426. writel(v, pi->mpsc_base + MPSC_CHR_2);
  427. return;
  428. }
  429. /*
  430. ******************************************************************************
  431. *
  432. * Driver Init Routines
  433. *
  434. ******************************************************************************
  435. */
  436. static void
  437. mpsc_init_hw(struct mpsc_port_info *pi)
  438. {
  439. pr_debug("mpsc_init_hw[%d]: Initializing\n", pi->port.line);
  440. mpsc_brg_init(pi, pi->brg_clk_src);
  441. mpsc_brg_enable(pi);
  442. mpsc_sdma_init(pi, dma_get_cache_alignment()); /* burst a cacheline */
  443. mpsc_sdma_stop(pi);
  444. mpsc_hw_init(pi);
  445. return;
  446. }
  447. static int
  448. mpsc_alloc_ring_mem(struct mpsc_port_info *pi)
  449. {
  450. int rc = 0;
  451. static void mpsc_free_ring_mem(struct mpsc_port_info *pi);
  452. pr_debug("mpsc_alloc_ring_mem[%d]: Allocating ring mem\n",
  453. pi->port.line);
  454. if (!pi->dma_region) {
  455. if (!dma_supported(pi->port.dev, 0xffffffff)) {
  456. printk(KERN_ERR "MPSC: Inadequate DMA support\n");
  457. rc = -ENXIO;
  458. }
  459. else if ((pi->dma_region = dma_alloc_noncoherent(pi->port.dev,
  460. MPSC_DMA_ALLOC_SIZE, &pi->dma_region_p, GFP_KERNEL))
  461. == NULL) {
  462. printk(KERN_ERR "MPSC: Can't alloc Desc region\n");
  463. rc = -ENOMEM;
  464. }
  465. }
  466. return rc;
  467. }
  468. static void
  469. mpsc_free_ring_mem(struct mpsc_port_info *pi)
  470. {
  471. pr_debug("mpsc_free_ring_mem[%d]: Freeing ring mem\n", pi->port.line);
  472. if (pi->dma_region) {
  473. dma_free_noncoherent(pi->port.dev, MPSC_DMA_ALLOC_SIZE,
  474. pi->dma_region, pi->dma_region_p);
  475. pi->dma_region = NULL;
  476. pi->dma_region_p = (dma_addr_t) NULL;
  477. }
  478. return;
  479. }
  480. static void
  481. mpsc_init_rings(struct mpsc_port_info *pi)
  482. {
  483. struct mpsc_rx_desc *rxre;
  484. struct mpsc_tx_desc *txre;
  485. dma_addr_t dp, dp_p;
  486. u8 *bp, *bp_p;
  487. int i;
  488. pr_debug("mpsc_init_rings[%d]: Initializing rings\n", pi->port.line);
  489. BUG_ON(pi->dma_region == NULL);
  490. memset(pi->dma_region, 0, MPSC_DMA_ALLOC_SIZE);
  491. /*
  492. * Descriptors & buffers are multiples of cacheline size and must be
  493. * cacheline aligned.
  494. */
  495. dp = ALIGN((u32) pi->dma_region, dma_get_cache_alignment());
  496. dp_p = ALIGN((u32) pi->dma_region_p, dma_get_cache_alignment());
  497. /*
  498. * Partition dma region into rx ring descriptor, rx buffers,
  499. * tx ring descriptors, and tx buffers.
  500. */
  501. pi->rxr = dp;
  502. pi->rxr_p = dp_p;
  503. dp += MPSC_RXR_SIZE;
  504. dp_p += MPSC_RXR_SIZE;
  505. pi->rxb = (u8 *) dp;
  506. pi->rxb_p = (u8 *) dp_p;
  507. dp += MPSC_RXB_SIZE;
  508. dp_p += MPSC_RXB_SIZE;
  509. pi->rxr_posn = 0;
  510. pi->txr = dp;
  511. pi->txr_p = dp_p;
  512. dp += MPSC_TXR_SIZE;
  513. dp_p += MPSC_TXR_SIZE;
  514. pi->txb = (u8 *) dp;
  515. pi->txb_p = (u8 *) dp_p;
  516. pi->txr_head = 0;
  517. pi->txr_tail = 0;
  518. /* Init rx ring descriptors */
  519. dp = pi->rxr;
  520. dp_p = pi->rxr_p;
  521. bp = pi->rxb;
  522. bp_p = pi->rxb_p;
  523. for (i = 0; i < MPSC_RXR_ENTRIES; i++) {
  524. rxre = (struct mpsc_rx_desc *)dp;
  525. rxre->bufsize = cpu_to_be16(MPSC_RXBE_SIZE);
  526. rxre->bytecnt = cpu_to_be16(0);
  527. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  528. SDMA_DESC_CMDSTAT_EI |
  529. SDMA_DESC_CMDSTAT_F |
  530. SDMA_DESC_CMDSTAT_L);
  531. rxre->link = cpu_to_be32(dp_p + MPSC_RXRE_SIZE);
  532. rxre->buf_ptr = cpu_to_be32(bp_p);
  533. dp += MPSC_RXRE_SIZE;
  534. dp_p += MPSC_RXRE_SIZE;
  535. bp += MPSC_RXBE_SIZE;
  536. bp_p += MPSC_RXBE_SIZE;
  537. }
  538. rxre->link = cpu_to_be32(pi->rxr_p); /* Wrap last back to first */
  539. /* Init tx ring descriptors */
  540. dp = pi->txr;
  541. dp_p = pi->txr_p;
  542. bp = pi->txb;
  543. bp_p = pi->txb_p;
  544. for (i = 0; i < MPSC_TXR_ENTRIES; i++) {
  545. txre = (struct mpsc_tx_desc *)dp;
  546. txre->link = cpu_to_be32(dp_p + MPSC_TXRE_SIZE);
  547. txre->buf_ptr = cpu_to_be32(bp_p);
  548. dp += MPSC_TXRE_SIZE;
  549. dp_p += MPSC_TXRE_SIZE;
  550. bp += MPSC_TXBE_SIZE;
  551. bp_p += MPSC_TXBE_SIZE;
  552. }
  553. txre->link = cpu_to_be32(pi->txr_p); /* Wrap last back to first */
  554. dma_cache_sync((void *) pi->dma_region, MPSC_DMA_ALLOC_SIZE,
  555. DMA_BIDIRECTIONAL);
  556. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  557. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  558. flush_dcache_range((ulong)pi->dma_region,
  559. (ulong)pi->dma_region + MPSC_DMA_ALLOC_SIZE);
  560. #endif
  561. return;
  562. }
  563. static void
  564. mpsc_uninit_rings(struct mpsc_port_info *pi)
  565. {
  566. pr_debug("mpsc_uninit_rings[%d]: Uninitializing rings\n",pi->port.line);
  567. BUG_ON(pi->dma_region == NULL);
  568. pi->rxr = 0;
  569. pi->rxr_p = 0;
  570. pi->rxb = NULL;
  571. pi->rxb_p = NULL;
  572. pi->rxr_posn = 0;
  573. pi->txr = 0;
  574. pi->txr_p = 0;
  575. pi->txb = NULL;
  576. pi->txb_p = NULL;
  577. pi->txr_head = 0;
  578. pi->txr_tail = 0;
  579. return;
  580. }
  581. static int
  582. mpsc_make_ready(struct mpsc_port_info *pi)
  583. {
  584. int rc;
  585. pr_debug("mpsc_make_ready[%d]: Making cltr ready\n", pi->port.line);
  586. if (!pi->ready) {
  587. mpsc_init_hw(pi);
  588. if ((rc = mpsc_alloc_ring_mem(pi)))
  589. return rc;
  590. mpsc_init_rings(pi);
  591. pi->ready = 1;
  592. }
  593. return 0;
  594. }
  595. /*
  596. ******************************************************************************
  597. *
  598. * Interrupt Handling Routines
  599. *
  600. ******************************************************************************
  601. */
  602. static inline int
  603. mpsc_rx_intr(struct mpsc_port_info *pi, struct pt_regs *regs)
  604. {
  605. struct mpsc_rx_desc *rxre;
  606. struct tty_struct *tty = pi->port.info->tty;
  607. u32 cmdstat, bytes_in, i;
  608. int rc = 0;
  609. u8 *bp;
  610. char flag = TTY_NORMAL;
  611. static void mpsc_start_rx(struct mpsc_port_info *pi);
  612. pr_debug("mpsc_rx_intr[%d]: Handling Rx intr\n", pi->port.line);
  613. rxre = (struct mpsc_rx_desc *)(pi->rxr + (pi->rxr_posn*MPSC_RXRE_SIZE));
  614. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  615. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  616. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  617. invalidate_dcache_range((ulong)rxre,
  618. (ulong)rxre + MPSC_RXRE_SIZE);
  619. #endif
  620. /*
  621. * Loop through Rx descriptors handling ones that have been completed.
  622. */
  623. while (!((cmdstat = be32_to_cpu(rxre->cmdstat)) & SDMA_DESC_CMDSTAT_O)){
  624. bytes_in = be16_to_cpu(rxre->bytecnt);
  625. /* Following use of tty struct directly is deprecated */
  626. if (unlikely((tty->flip.count + bytes_in) >= TTY_FLIPBUF_SIZE)){
  627. if (tty->low_latency)
  628. tty_flip_buffer_push(tty);
  629. /*
  630. * If this failed then we will throw awa the bytes
  631. * but mst do so to clear interrupts.
  632. */
  633. }
  634. bp = pi->rxb + (pi->rxr_posn * MPSC_RXBE_SIZE);
  635. dma_cache_sync((void *) bp, MPSC_RXBE_SIZE, DMA_FROM_DEVICE);
  636. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  637. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  638. invalidate_dcache_range((ulong)bp,
  639. (ulong)bp + MPSC_RXBE_SIZE);
  640. #endif
  641. /*
  642. * Other than for parity error, the manual provides little
  643. * info on what data will be in a frame flagged by any of
  644. * these errors. For parity error, it is the last byte in
  645. * the buffer that had the error. As for the rest, I guess
  646. * we'll assume there is no data in the buffer.
  647. * If there is...it gets lost.
  648. */
  649. if (unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  650. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) {
  651. pi->port.icount.rx++;
  652. if (cmdstat & SDMA_DESC_CMDSTAT_BR) { /* Break */
  653. pi->port.icount.brk++;
  654. if (uart_handle_break(&pi->port))
  655. goto next_frame;
  656. }
  657. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)/* Framing */
  658. pi->port.icount.frame++;
  659. else if (cmdstat & SDMA_DESC_CMDSTAT_OR) /* Overrun */
  660. pi->port.icount.overrun++;
  661. cmdstat &= pi->port.read_status_mask;
  662. if (cmdstat & SDMA_DESC_CMDSTAT_BR)
  663. flag = TTY_BREAK;
  664. else if (cmdstat & SDMA_DESC_CMDSTAT_FR)
  665. flag = TTY_FRAME;
  666. else if (cmdstat & SDMA_DESC_CMDSTAT_OR)
  667. flag = TTY_OVERRUN;
  668. else if (cmdstat & SDMA_DESC_CMDSTAT_PE)
  669. flag = TTY_PARITY;
  670. }
  671. if (uart_handle_sysrq_char(&pi->port, *bp, regs)) {
  672. bp++;
  673. bytes_in--;
  674. goto next_frame;
  675. }
  676. if ((unlikely(cmdstat & (SDMA_DESC_CMDSTAT_BR |
  677. SDMA_DESC_CMDSTAT_FR | SDMA_DESC_CMDSTAT_OR))) &&
  678. !(cmdstat & pi->port.ignore_status_mask))
  679. tty_insert_flip_char(tty, *bp, flag);
  680. else {
  681. for (i=0; i<bytes_in; i++)
  682. tty_insert_flip_char(tty, *bp++, TTY_NORMAL);
  683. pi->port.icount.rx += bytes_in;
  684. }
  685. next_frame:
  686. rxre->bytecnt = cpu_to_be16(0);
  687. wmb();
  688. rxre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O |
  689. SDMA_DESC_CMDSTAT_EI |
  690. SDMA_DESC_CMDSTAT_F |
  691. SDMA_DESC_CMDSTAT_L);
  692. wmb();
  693. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_BIDIRECTIONAL);
  694. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  695. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  696. flush_dcache_range((ulong)rxre,
  697. (ulong)rxre + MPSC_RXRE_SIZE);
  698. #endif
  699. /* Advance to next descriptor */
  700. pi->rxr_posn = (pi->rxr_posn + 1) & (MPSC_RXR_ENTRIES - 1);
  701. rxre = (struct mpsc_rx_desc *)(pi->rxr +
  702. (pi->rxr_posn * MPSC_RXRE_SIZE));
  703. dma_cache_sync((void *)rxre, MPSC_RXRE_SIZE, DMA_FROM_DEVICE);
  704. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  705. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  706. invalidate_dcache_range((ulong)rxre,
  707. (ulong)rxre + MPSC_RXRE_SIZE);
  708. #endif
  709. rc = 1;
  710. }
  711. /* Restart rx engine, if its stopped */
  712. if ((readl(pi->sdma_base + SDMA_SDCM) & SDMA_SDCM_ERD) == 0)
  713. mpsc_start_rx(pi);
  714. tty_flip_buffer_push(tty);
  715. return rc;
  716. }
  717. static inline void
  718. mpsc_setup_tx_desc(struct mpsc_port_info *pi, u32 count, u32 intr)
  719. {
  720. struct mpsc_tx_desc *txre;
  721. txre = (struct mpsc_tx_desc *)(pi->txr +
  722. (pi->txr_head * MPSC_TXRE_SIZE));
  723. txre->bytecnt = cpu_to_be16(count);
  724. txre->shadow = txre->bytecnt;
  725. wmb(); /* ensure cmdstat is last field updated */
  726. txre->cmdstat = cpu_to_be32(SDMA_DESC_CMDSTAT_O | SDMA_DESC_CMDSTAT_F |
  727. SDMA_DESC_CMDSTAT_L | ((intr) ?
  728. SDMA_DESC_CMDSTAT_EI
  729. : 0));
  730. wmb();
  731. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_BIDIRECTIONAL);
  732. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  733. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  734. flush_dcache_range((ulong)txre,
  735. (ulong)txre + MPSC_TXRE_SIZE);
  736. #endif
  737. return;
  738. }
  739. static inline void
  740. mpsc_copy_tx_data(struct mpsc_port_info *pi)
  741. {
  742. struct circ_buf *xmit = &pi->port.info->xmit;
  743. u8 *bp;
  744. u32 i;
  745. /* Make sure the desc ring isn't full */
  746. while (CIRC_CNT(pi->txr_head, pi->txr_tail, MPSC_TXR_ENTRIES) <
  747. (MPSC_TXR_ENTRIES - 1)) {
  748. if (pi->port.x_char) {
  749. /*
  750. * Ideally, we should use the TCS field in
  751. * CHR_1 to put the x_char out immediately but
  752. * errata prevents us from being able to read
  753. * CHR_2 to know that its safe to write to
  754. * CHR_1. Instead, just put it in-band with
  755. * all the other Tx data.
  756. */
  757. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  758. *bp = pi->port.x_char;
  759. pi->port.x_char = 0;
  760. i = 1;
  761. }
  762. else if (!uart_circ_empty(xmit) && !uart_tx_stopped(&pi->port)){
  763. i = min((u32) MPSC_TXBE_SIZE,
  764. (u32) uart_circ_chars_pending(xmit));
  765. i = min(i, (u32) CIRC_CNT_TO_END(xmit->head, xmit->tail,
  766. UART_XMIT_SIZE));
  767. bp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  768. memcpy(bp, &xmit->buf[xmit->tail], i);
  769. xmit->tail = (xmit->tail + i) & (UART_XMIT_SIZE - 1);
  770. if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS)
  771. uart_write_wakeup(&pi->port);
  772. }
  773. else /* All tx data copied into ring bufs */
  774. return;
  775. dma_cache_sync((void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  776. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  777. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  778. flush_dcache_range((ulong)bp,
  779. (ulong)bp + MPSC_TXBE_SIZE);
  780. #endif
  781. mpsc_setup_tx_desc(pi, i, 1);
  782. /* Advance to next descriptor */
  783. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  784. }
  785. return;
  786. }
  787. static inline int
  788. mpsc_tx_intr(struct mpsc_port_info *pi)
  789. {
  790. struct mpsc_tx_desc *txre;
  791. int rc = 0;
  792. if (!mpsc_sdma_tx_active(pi)) {
  793. txre = (struct mpsc_tx_desc *)(pi->txr +
  794. (pi->txr_tail * MPSC_TXRE_SIZE));
  795. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE, DMA_FROM_DEVICE);
  796. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  797. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  798. invalidate_dcache_range((ulong)txre,
  799. (ulong)txre + MPSC_TXRE_SIZE);
  800. #endif
  801. while (!(be32_to_cpu(txre->cmdstat) & SDMA_DESC_CMDSTAT_O)) {
  802. rc = 1;
  803. pi->port.icount.tx += be16_to_cpu(txre->bytecnt);
  804. pi->txr_tail = (pi->txr_tail+1) & (MPSC_TXR_ENTRIES-1);
  805. /* If no more data to tx, fall out of loop */
  806. if (pi->txr_head == pi->txr_tail)
  807. break;
  808. txre = (struct mpsc_tx_desc *)(pi->txr +
  809. (pi->txr_tail * MPSC_TXRE_SIZE));
  810. dma_cache_sync((void *) txre, MPSC_TXRE_SIZE,
  811. DMA_FROM_DEVICE);
  812. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  813. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  814. invalidate_dcache_range((ulong)txre,
  815. (ulong)txre + MPSC_TXRE_SIZE);
  816. #endif
  817. }
  818. mpsc_copy_tx_data(pi);
  819. mpsc_sdma_start_tx(pi); /* start next desc if ready */
  820. }
  821. return rc;
  822. }
  823. /*
  824. * This is the driver's interrupt handler. To avoid a race, we first clear
  825. * the interrupt, then handle any completed Rx/Tx descriptors. When done
  826. * handling those descriptors, we restart the Rx/Tx engines if they're stopped.
  827. */
  828. static irqreturn_t
  829. mpsc_sdma_intr(int irq, void *dev_id, struct pt_regs *regs)
  830. {
  831. struct mpsc_port_info *pi = dev_id;
  832. ulong iflags;
  833. int rc = IRQ_NONE;
  834. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Received\n",pi->port.line);
  835. spin_lock_irqsave(&pi->port.lock, iflags);
  836. mpsc_sdma_intr_ack(pi);
  837. if (mpsc_rx_intr(pi, regs))
  838. rc = IRQ_HANDLED;
  839. if (mpsc_tx_intr(pi))
  840. rc = IRQ_HANDLED;
  841. spin_unlock_irqrestore(&pi->port.lock, iflags);
  842. pr_debug("mpsc_sdma_intr[%d]: SDMA Interrupt Handled\n", pi->port.line);
  843. return rc;
  844. }
  845. /*
  846. ******************************************************************************
  847. *
  848. * serial_core.c Interface routines
  849. *
  850. ******************************************************************************
  851. */
  852. static uint
  853. mpsc_tx_empty(struct uart_port *port)
  854. {
  855. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  856. ulong iflags;
  857. uint rc;
  858. spin_lock_irqsave(&pi->port.lock, iflags);
  859. rc = mpsc_sdma_tx_active(pi) ? 0 : TIOCSER_TEMT;
  860. spin_unlock_irqrestore(&pi->port.lock, iflags);
  861. return rc;
  862. }
  863. static void
  864. mpsc_set_mctrl(struct uart_port *port, uint mctrl)
  865. {
  866. /* Have no way to set modem control lines AFAICT */
  867. return;
  868. }
  869. static uint
  870. mpsc_get_mctrl(struct uart_port *port)
  871. {
  872. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  873. u32 mflags, status;
  874. ulong iflags;
  875. spin_lock_irqsave(&pi->port.lock, iflags);
  876. status = (pi->mirror_regs) ? pi->MPSC_CHR_10_m :
  877. readl(pi->mpsc_base + MPSC_CHR_10);
  878. spin_unlock_irqrestore(&pi->port.lock, iflags);
  879. mflags = 0;
  880. if (status & 0x1)
  881. mflags |= TIOCM_CTS;
  882. if (status & 0x2)
  883. mflags |= TIOCM_CAR;
  884. return mflags | TIOCM_DSR; /* No way to tell if DSR asserted */
  885. }
  886. static void
  887. mpsc_stop_tx(struct uart_port *port, uint tty_start)
  888. {
  889. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  890. pr_debug("mpsc_stop_tx[%d]: tty_start: %d\n", port->line, tty_start);
  891. mpsc_freeze(pi);
  892. return;
  893. }
  894. static void
  895. mpsc_start_tx(struct uart_port *port, uint tty_start)
  896. {
  897. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  898. mpsc_unfreeze(pi);
  899. mpsc_copy_tx_data(pi);
  900. mpsc_sdma_start_tx(pi);
  901. pr_debug("mpsc_start_tx[%d]: tty_start: %d\n", port->line, tty_start);
  902. return;
  903. }
  904. static void
  905. mpsc_start_rx(struct mpsc_port_info *pi)
  906. {
  907. pr_debug("mpsc_start_rx[%d]: Starting...\n", pi->port.line);
  908. if (pi->rcv_data) {
  909. mpsc_enter_hunt(pi);
  910. mpsc_sdma_cmd(pi, SDMA_SDCM_ERD);
  911. }
  912. return;
  913. }
  914. static void
  915. mpsc_stop_rx(struct uart_port *port)
  916. {
  917. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  918. pr_debug("mpsc_stop_rx[%d]: Stopping...\n", port->line);
  919. mpsc_sdma_cmd(pi, SDMA_SDCM_AR);
  920. return;
  921. }
  922. static void
  923. mpsc_enable_ms(struct uart_port *port)
  924. {
  925. return; /* Not supported */
  926. }
  927. static void
  928. mpsc_break_ctl(struct uart_port *port, int ctl)
  929. {
  930. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  931. ulong flags;
  932. u32 v;
  933. v = ctl ? 0x00ff0000 : 0;
  934. spin_lock_irqsave(&pi->port.lock, flags);
  935. if (pi->mirror_regs)
  936. pi->MPSC_CHR_1_m = v;
  937. writel(v, pi->mpsc_base + MPSC_CHR_1);
  938. spin_unlock_irqrestore(&pi->port.lock, flags);
  939. return;
  940. }
  941. static int
  942. mpsc_startup(struct uart_port *port)
  943. {
  944. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  945. u32 flag = 0;
  946. int rc;
  947. pr_debug("mpsc_startup[%d]: Starting up MPSC, irq: %d\n",
  948. port->line, pi->port.irq);
  949. if ((rc = mpsc_make_ready(pi)) == 0) {
  950. /* Setup IRQ handler */
  951. mpsc_sdma_intr_ack(pi);
  952. /* If irq's are shared, need to set flag */
  953. if (mpsc_ports[0].port.irq == mpsc_ports[1].port.irq)
  954. flag = SA_SHIRQ;
  955. if (request_irq(pi->port.irq, mpsc_sdma_intr, flag,
  956. "mpsc/sdma", pi))
  957. printk(KERN_ERR "MPSC: Can't get SDMA IRQ %d\n",
  958. pi->port.irq);
  959. mpsc_sdma_intr_unmask(pi, 0xf);
  960. mpsc_sdma_set_rx_ring(pi, (struct mpsc_rx_desc *)(pi->rxr_p +
  961. (pi->rxr_posn * MPSC_RXRE_SIZE)));
  962. }
  963. return rc;
  964. }
  965. static void
  966. mpsc_shutdown(struct uart_port *port)
  967. {
  968. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  969. static void mpsc_release_port(struct uart_port *port);
  970. pr_debug("mpsc_shutdown[%d]: Shutting down MPSC\n", port->line);
  971. mpsc_sdma_stop(pi);
  972. free_irq(pi->port.irq, pi);
  973. return;
  974. }
  975. static void
  976. mpsc_set_termios(struct uart_port *port, struct termios *termios,
  977. struct termios *old)
  978. {
  979. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  980. u32 baud;
  981. ulong flags;
  982. u32 chr_bits, stop_bits, par;
  983. pi->c_iflag = termios->c_iflag;
  984. pi->c_cflag = termios->c_cflag;
  985. switch (termios->c_cflag & CSIZE) {
  986. case CS5:
  987. chr_bits = MPSC_MPCR_CL_5;
  988. break;
  989. case CS6:
  990. chr_bits = MPSC_MPCR_CL_6;
  991. break;
  992. case CS7:
  993. chr_bits = MPSC_MPCR_CL_7;
  994. break;
  995. case CS8:
  996. default:
  997. chr_bits = MPSC_MPCR_CL_8;
  998. break;
  999. }
  1000. if (termios->c_cflag & CSTOPB)
  1001. stop_bits = MPSC_MPCR_SBL_2;
  1002. else
  1003. stop_bits = MPSC_MPCR_SBL_1;
  1004. par = MPSC_CHR_2_PAR_EVEN;
  1005. if (termios->c_cflag & PARENB)
  1006. if (termios->c_cflag & PARODD)
  1007. par = MPSC_CHR_2_PAR_ODD;
  1008. #ifdef CMSPAR
  1009. if (termios->c_cflag & CMSPAR) {
  1010. if (termios->c_cflag & PARODD)
  1011. par = MPSC_CHR_2_PAR_MARK;
  1012. else
  1013. par = MPSC_CHR_2_PAR_SPACE;
  1014. }
  1015. #endif
  1016. baud = uart_get_baud_rate(port, termios, old, 0, port->uartclk);
  1017. spin_lock_irqsave(&pi->port.lock, flags);
  1018. uart_update_timeout(port, termios->c_cflag, baud);
  1019. mpsc_set_char_length(pi, chr_bits);
  1020. mpsc_set_stop_bit_length(pi, stop_bits);
  1021. mpsc_set_parity(pi, par);
  1022. mpsc_set_baudrate(pi, baud);
  1023. /* Characters/events to read */
  1024. pi->rcv_data = 1;
  1025. pi->port.read_status_mask = SDMA_DESC_CMDSTAT_OR;
  1026. if (termios->c_iflag & INPCK)
  1027. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1028. SDMA_DESC_CMDSTAT_FR;
  1029. if (termios->c_iflag & (BRKINT | PARMRK))
  1030. pi->port.read_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1031. /* Characters/events to ignore */
  1032. pi->port.ignore_status_mask = 0;
  1033. if (termios->c_iflag & IGNPAR)
  1034. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_PE |
  1035. SDMA_DESC_CMDSTAT_FR;
  1036. if (termios->c_iflag & IGNBRK) {
  1037. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_BR;
  1038. if (termios->c_iflag & IGNPAR)
  1039. pi->port.ignore_status_mask |= SDMA_DESC_CMDSTAT_OR;
  1040. }
  1041. /* Ignore all chars if CREAD not set */
  1042. if (!(termios->c_cflag & CREAD))
  1043. pi->rcv_data = 0;
  1044. else
  1045. mpsc_start_rx(pi);
  1046. spin_unlock_irqrestore(&pi->port.lock, flags);
  1047. return;
  1048. }
  1049. static const char *
  1050. mpsc_type(struct uart_port *port)
  1051. {
  1052. pr_debug("mpsc_type[%d]: port type: %s\n", port->line,MPSC_DRIVER_NAME);
  1053. return MPSC_DRIVER_NAME;
  1054. }
  1055. static int
  1056. mpsc_request_port(struct uart_port *port)
  1057. {
  1058. /* Should make chip/platform specific call */
  1059. return 0;
  1060. }
  1061. static void
  1062. mpsc_release_port(struct uart_port *port)
  1063. {
  1064. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1065. if (pi->ready) {
  1066. mpsc_uninit_rings(pi);
  1067. mpsc_free_ring_mem(pi);
  1068. pi->ready = 0;
  1069. }
  1070. return;
  1071. }
  1072. static void
  1073. mpsc_config_port(struct uart_port *port, int flags)
  1074. {
  1075. return;
  1076. }
  1077. static int
  1078. mpsc_verify_port(struct uart_port *port, struct serial_struct *ser)
  1079. {
  1080. struct mpsc_port_info *pi = (struct mpsc_port_info *)port;
  1081. int rc = 0;
  1082. pr_debug("mpsc_verify_port[%d]: Verifying port data\n", pi->port.line);
  1083. if (ser->type != PORT_UNKNOWN && ser->type != PORT_MPSC)
  1084. rc = -EINVAL;
  1085. else if (pi->port.irq != ser->irq)
  1086. rc = -EINVAL;
  1087. else if (ser->io_type != SERIAL_IO_MEM)
  1088. rc = -EINVAL;
  1089. else if (pi->port.uartclk / 16 != ser->baud_base) /* Not sure */
  1090. rc = -EINVAL;
  1091. else if ((void *)pi->port.mapbase != ser->iomem_base)
  1092. rc = -EINVAL;
  1093. else if (pi->port.iobase != ser->port)
  1094. rc = -EINVAL;
  1095. else if (ser->hub6 != 0)
  1096. rc = -EINVAL;
  1097. return rc;
  1098. }
  1099. static struct uart_ops mpsc_pops = {
  1100. .tx_empty = mpsc_tx_empty,
  1101. .set_mctrl = mpsc_set_mctrl,
  1102. .get_mctrl = mpsc_get_mctrl,
  1103. .stop_tx = mpsc_stop_tx,
  1104. .start_tx = mpsc_start_tx,
  1105. .stop_rx = mpsc_stop_rx,
  1106. .enable_ms = mpsc_enable_ms,
  1107. .break_ctl = mpsc_break_ctl,
  1108. .startup = mpsc_startup,
  1109. .shutdown = mpsc_shutdown,
  1110. .set_termios = mpsc_set_termios,
  1111. .type = mpsc_type,
  1112. .release_port = mpsc_release_port,
  1113. .request_port = mpsc_request_port,
  1114. .config_port = mpsc_config_port,
  1115. .verify_port = mpsc_verify_port,
  1116. };
  1117. /*
  1118. ******************************************************************************
  1119. *
  1120. * Console Interface Routines
  1121. *
  1122. ******************************************************************************
  1123. */
  1124. #ifdef CONFIG_SERIAL_MPSC_CONSOLE
  1125. static void
  1126. mpsc_console_write(struct console *co, const char *s, uint count)
  1127. {
  1128. struct mpsc_port_info *pi = &mpsc_ports[co->index];
  1129. u8 *bp, *dp, add_cr = 0;
  1130. int i;
  1131. while (mpsc_sdma_tx_active(pi))
  1132. udelay(100);
  1133. while (count > 0) {
  1134. bp = dp = pi->txb + (pi->txr_head * MPSC_TXBE_SIZE);
  1135. for (i = 0; i < MPSC_TXBE_SIZE; i++) {
  1136. if (count == 0)
  1137. break;
  1138. if (add_cr) {
  1139. *(dp++) = '\r';
  1140. add_cr = 0;
  1141. }
  1142. else {
  1143. *(dp++) = *s;
  1144. if (*(s++) == '\n') { /* add '\r' after '\n' */
  1145. add_cr = 1;
  1146. count++;
  1147. }
  1148. }
  1149. count--;
  1150. }
  1151. dma_cache_sync((void *) bp, MPSC_TXBE_SIZE, DMA_BIDIRECTIONAL);
  1152. #if defined(CONFIG_PPC32) && !defined(CONFIG_NOT_COHERENT_CACHE)
  1153. if (pi->cache_mgmt) /* GT642[46]0 Res #COMM-2 */
  1154. flush_dcache_range((ulong)bp,
  1155. (ulong)bp + MPSC_TXBE_SIZE);
  1156. #endif
  1157. mpsc_setup_tx_desc(pi, i, 0);
  1158. pi->txr_head = (pi->txr_head + 1) & (MPSC_TXR_ENTRIES - 1);
  1159. mpsc_sdma_start_tx(pi);
  1160. while (mpsc_sdma_tx_active(pi))
  1161. udelay(100);
  1162. pi->txr_tail = (pi->txr_tail + 1) & (MPSC_TXR_ENTRIES - 1);
  1163. }
  1164. return;
  1165. }
  1166. static int __init
  1167. mpsc_console_setup(struct console *co, char *options)
  1168. {
  1169. struct mpsc_port_info *pi;
  1170. int baud, bits, parity, flow;
  1171. pr_debug("mpsc_console_setup[%d]: options: %s\n", co->index, options);
  1172. if (co->index >= MPSC_NUM_CTLRS)
  1173. co->index = 0;
  1174. pi = &mpsc_ports[co->index];
  1175. baud = pi->default_baud;
  1176. bits = pi->default_bits;
  1177. parity = pi->default_parity;
  1178. flow = pi->default_flow;
  1179. if (!pi->port.ops)
  1180. return -ENODEV;
  1181. spin_lock_init(&pi->port.lock); /* Temporary fix--copied from 8250.c */
  1182. if (options)
  1183. uart_parse_options(options, &baud, &parity, &bits, &flow);
  1184. return uart_set_options(&pi->port, co, baud, parity, bits, flow);
  1185. }
  1186. extern struct uart_driver mpsc_reg;
  1187. static struct console mpsc_console = {
  1188. .name = MPSC_DEV_NAME,
  1189. .write = mpsc_console_write,
  1190. .device = uart_console_device,
  1191. .setup = mpsc_console_setup,
  1192. .flags = CON_PRINTBUFFER,
  1193. .index = -1,
  1194. .data = &mpsc_reg,
  1195. };
  1196. static int __init
  1197. mpsc_late_console_init(void)
  1198. {
  1199. pr_debug("mpsc_late_console_init: Enter\n");
  1200. if (!(mpsc_console.flags & CON_ENABLED))
  1201. register_console(&mpsc_console);
  1202. return 0;
  1203. }
  1204. late_initcall(mpsc_late_console_init);
  1205. #define MPSC_CONSOLE &mpsc_console
  1206. #else
  1207. #define MPSC_CONSOLE NULL
  1208. #endif
  1209. /*
  1210. ******************************************************************************
  1211. *
  1212. * Dummy Platform Driver to extract & map shared register regions
  1213. *
  1214. ******************************************************************************
  1215. */
  1216. static void
  1217. mpsc_resource_err(char *s)
  1218. {
  1219. printk(KERN_WARNING "MPSC: Platform device resource error in %s\n", s);
  1220. return;
  1221. }
  1222. static int
  1223. mpsc_shared_map_regs(struct platform_device *pd)
  1224. {
  1225. struct resource *r;
  1226. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1227. MPSC_ROUTING_BASE_ORDER)) && request_mem_region(r->start,
  1228. MPSC_ROUTING_REG_BLOCK_SIZE, "mpsc_routing_regs")) {
  1229. mpsc_shared_regs.mpsc_routing_base = ioremap(r->start,
  1230. MPSC_ROUTING_REG_BLOCK_SIZE);
  1231. mpsc_shared_regs.mpsc_routing_base_p = r->start;
  1232. }
  1233. else {
  1234. mpsc_resource_err("MPSC routing base");
  1235. return -ENOMEM;
  1236. }
  1237. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1238. MPSC_SDMA_INTR_BASE_ORDER)) && request_mem_region(r->start,
  1239. MPSC_SDMA_INTR_REG_BLOCK_SIZE, "sdma_intr_regs")) {
  1240. mpsc_shared_regs.sdma_intr_base = ioremap(r->start,
  1241. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1242. mpsc_shared_regs.sdma_intr_base_p = r->start;
  1243. }
  1244. else {
  1245. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1246. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1247. MPSC_ROUTING_REG_BLOCK_SIZE);
  1248. mpsc_resource_err("SDMA intr base");
  1249. return -ENOMEM;
  1250. }
  1251. return 0;
  1252. }
  1253. static void
  1254. mpsc_shared_unmap_regs(void)
  1255. {
  1256. if (!mpsc_shared_regs.mpsc_routing_base) {
  1257. iounmap(mpsc_shared_regs.mpsc_routing_base);
  1258. release_mem_region(mpsc_shared_regs.mpsc_routing_base_p,
  1259. MPSC_ROUTING_REG_BLOCK_SIZE);
  1260. }
  1261. if (!mpsc_shared_regs.sdma_intr_base) {
  1262. iounmap(mpsc_shared_regs.sdma_intr_base);
  1263. release_mem_region(mpsc_shared_regs.sdma_intr_base_p,
  1264. MPSC_SDMA_INTR_REG_BLOCK_SIZE);
  1265. }
  1266. mpsc_shared_regs.mpsc_routing_base = NULL;
  1267. mpsc_shared_regs.sdma_intr_base = NULL;
  1268. mpsc_shared_regs.mpsc_routing_base_p = 0;
  1269. mpsc_shared_regs.sdma_intr_base_p = 0;
  1270. return;
  1271. }
  1272. static int
  1273. mpsc_shared_drv_probe(struct device *dev)
  1274. {
  1275. struct platform_device *pd = to_platform_device(dev);
  1276. struct mpsc_shared_pdata *pdata;
  1277. int rc = -ENODEV;
  1278. if (pd->id == 0) {
  1279. if (!(rc = mpsc_shared_map_regs(pd))) {
  1280. pdata = (struct mpsc_shared_pdata *)dev->platform_data;
  1281. mpsc_shared_regs.MPSC_MRR_m = pdata->mrr_val;
  1282. mpsc_shared_regs.MPSC_RCRR_m= pdata->rcrr_val;
  1283. mpsc_shared_regs.MPSC_TCRR_m= pdata->tcrr_val;
  1284. mpsc_shared_regs.SDMA_INTR_CAUSE_m =
  1285. pdata->intr_cause_val;
  1286. mpsc_shared_regs.SDMA_INTR_MASK_m =
  1287. pdata->intr_mask_val;
  1288. rc = 0;
  1289. }
  1290. }
  1291. return rc;
  1292. }
  1293. static int
  1294. mpsc_shared_drv_remove(struct device *dev)
  1295. {
  1296. struct platform_device *pd = to_platform_device(dev);
  1297. int rc = -ENODEV;
  1298. if (pd->id == 0) {
  1299. mpsc_shared_unmap_regs();
  1300. mpsc_shared_regs.MPSC_MRR_m = 0;
  1301. mpsc_shared_regs.MPSC_RCRR_m = 0;
  1302. mpsc_shared_regs.MPSC_TCRR_m = 0;
  1303. mpsc_shared_regs.SDMA_INTR_CAUSE_m = 0;
  1304. mpsc_shared_regs.SDMA_INTR_MASK_m = 0;
  1305. rc = 0;
  1306. }
  1307. return rc;
  1308. }
  1309. static struct device_driver mpsc_shared_driver = {
  1310. .name = MPSC_SHARED_NAME,
  1311. .bus = &platform_bus_type,
  1312. .probe = mpsc_shared_drv_probe,
  1313. .remove = mpsc_shared_drv_remove,
  1314. };
  1315. /*
  1316. ******************************************************************************
  1317. *
  1318. * Driver Interface Routines
  1319. *
  1320. ******************************************************************************
  1321. */
  1322. static struct uart_driver mpsc_reg = {
  1323. .owner = THIS_MODULE,
  1324. .driver_name = MPSC_DRIVER_NAME,
  1325. .devfs_name = MPSC_DEVFS_NAME,
  1326. .dev_name = MPSC_DEV_NAME,
  1327. .major = MPSC_MAJOR,
  1328. .minor = MPSC_MINOR_START,
  1329. .nr = MPSC_NUM_CTLRS,
  1330. .cons = MPSC_CONSOLE,
  1331. };
  1332. static int
  1333. mpsc_drv_map_regs(struct mpsc_port_info *pi, struct platform_device *pd)
  1334. {
  1335. struct resource *r;
  1336. if ((r = platform_get_resource(pd, IORESOURCE_MEM, MPSC_BASE_ORDER)) &&
  1337. request_mem_region(r->start, MPSC_REG_BLOCK_SIZE, "mpsc_regs")){
  1338. pi->mpsc_base = ioremap(r->start, MPSC_REG_BLOCK_SIZE);
  1339. pi->mpsc_base_p = r->start;
  1340. }
  1341. else {
  1342. mpsc_resource_err("MPSC base");
  1343. return -ENOMEM;
  1344. }
  1345. if ((r = platform_get_resource(pd, IORESOURCE_MEM,
  1346. MPSC_SDMA_BASE_ORDER)) && request_mem_region(r->start,
  1347. MPSC_SDMA_REG_BLOCK_SIZE, "sdma_regs")) {
  1348. pi->sdma_base = ioremap(r->start,MPSC_SDMA_REG_BLOCK_SIZE);
  1349. pi->sdma_base_p = r->start;
  1350. }
  1351. else {
  1352. mpsc_resource_err("SDMA base");
  1353. return -ENOMEM;
  1354. }
  1355. if ((r = platform_get_resource(pd,IORESOURCE_MEM,MPSC_BRG_BASE_ORDER))
  1356. && request_mem_region(r->start, MPSC_BRG_REG_BLOCK_SIZE,
  1357. "brg_regs")) {
  1358. pi->brg_base = ioremap(r->start, MPSC_BRG_REG_BLOCK_SIZE);
  1359. pi->brg_base_p = r->start;
  1360. }
  1361. else {
  1362. mpsc_resource_err("BRG base");
  1363. return -ENOMEM;
  1364. }
  1365. return 0;
  1366. }
  1367. static void
  1368. mpsc_drv_unmap_regs(struct mpsc_port_info *pi)
  1369. {
  1370. if (!pi->mpsc_base) {
  1371. iounmap(pi->mpsc_base);
  1372. release_mem_region(pi->mpsc_base_p, MPSC_REG_BLOCK_SIZE);
  1373. }
  1374. if (!pi->sdma_base) {
  1375. iounmap(pi->sdma_base);
  1376. release_mem_region(pi->sdma_base_p, MPSC_SDMA_REG_BLOCK_SIZE);
  1377. }
  1378. if (!pi->brg_base) {
  1379. iounmap(pi->brg_base);
  1380. release_mem_region(pi->brg_base_p, MPSC_BRG_REG_BLOCK_SIZE);
  1381. }
  1382. pi->mpsc_base = NULL;
  1383. pi->sdma_base = NULL;
  1384. pi->brg_base = NULL;
  1385. pi->mpsc_base_p = 0;
  1386. pi->sdma_base_p = 0;
  1387. pi->brg_base_p = 0;
  1388. return;
  1389. }
  1390. static void
  1391. mpsc_drv_get_platform_data(struct mpsc_port_info *pi,
  1392. struct platform_device *pd, int num)
  1393. {
  1394. struct mpsc_pdata *pdata;
  1395. pdata = (struct mpsc_pdata *)pd->dev.platform_data;
  1396. pi->port.uartclk = pdata->brg_clk_freq;
  1397. pi->port.iotype = UPIO_MEM;
  1398. pi->port.line = num;
  1399. pi->port.type = PORT_MPSC;
  1400. pi->port.fifosize = MPSC_TXBE_SIZE;
  1401. pi->port.membase = pi->mpsc_base;
  1402. pi->port.mapbase = (ulong)pi->mpsc_base;
  1403. pi->port.ops = &mpsc_pops;
  1404. pi->mirror_regs = pdata->mirror_regs;
  1405. pi->cache_mgmt = pdata->cache_mgmt;
  1406. pi->brg_can_tune = pdata->brg_can_tune;
  1407. pi->brg_clk_src = pdata->brg_clk_src;
  1408. pi->mpsc_max_idle = pdata->max_idle;
  1409. pi->default_baud = pdata->default_baud;
  1410. pi->default_bits = pdata->default_bits;
  1411. pi->default_parity = pdata->default_parity;
  1412. pi->default_flow = pdata->default_flow;
  1413. /* Initial values of mirrored regs */
  1414. pi->MPSC_CHR_1_m = pdata->chr_1_val;
  1415. pi->MPSC_CHR_2_m = pdata->chr_2_val;
  1416. pi->MPSC_CHR_10_m = pdata->chr_10_val;
  1417. pi->MPSC_MPCR_m = pdata->mpcr_val;
  1418. pi->BRG_BCR_m = pdata->bcr_val;
  1419. pi->shared_regs = &mpsc_shared_regs;
  1420. pi->port.irq = platform_get_irq(pd, 0);
  1421. return;
  1422. }
  1423. static int
  1424. mpsc_drv_probe(struct device *dev)
  1425. {
  1426. struct platform_device *pd = to_platform_device(dev);
  1427. struct mpsc_port_info *pi;
  1428. int rc = -ENODEV;
  1429. pr_debug("mpsc_drv_probe: Adding MPSC %d\n", pd->id);
  1430. if (pd->id < MPSC_NUM_CTLRS) {
  1431. pi = &mpsc_ports[pd->id];
  1432. if (!(rc = mpsc_drv_map_regs(pi, pd))) {
  1433. mpsc_drv_get_platform_data(pi, pd, pd->id);
  1434. if (!(rc = mpsc_make_ready(pi)))
  1435. if (!(rc = uart_add_one_port(&mpsc_reg,
  1436. &pi->port)))
  1437. rc = 0;
  1438. else {
  1439. mpsc_release_port(
  1440. (struct uart_port *)pi);
  1441. mpsc_drv_unmap_regs(pi);
  1442. }
  1443. else
  1444. mpsc_drv_unmap_regs(pi);
  1445. }
  1446. }
  1447. return rc;
  1448. }
  1449. static int
  1450. mpsc_drv_remove(struct device *dev)
  1451. {
  1452. struct platform_device *pd = to_platform_device(dev);
  1453. pr_debug("mpsc_drv_exit: Removing MPSC %d\n", pd->id);
  1454. if (pd->id < MPSC_NUM_CTLRS) {
  1455. uart_remove_one_port(&mpsc_reg, &mpsc_ports[pd->id].port);
  1456. mpsc_release_port((struct uart_port *)&mpsc_ports[pd->id].port);
  1457. mpsc_drv_unmap_regs(&mpsc_ports[pd->id]);
  1458. return 0;
  1459. }
  1460. else
  1461. return -ENODEV;
  1462. }
  1463. static struct device_driver mpsc_driver = {
  1464. .name = MPSC_CTLR_NAME,
  1465. .bus = &platform_bus_type,
  1466. .probe = mpsc_drv_probe,
  1467. .remove = mpsc_drv_remove,
  1468. };
  1469. static int __init
  1470. mpsc_drv_init(void)
  1471. {
  1472. int rc;
  1473. printk(KERN_INFO "Serial: MPSC driver $Revision: 1.00 $\n");
  1474. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1475. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1476. if (!(rc = uart_register_driver(&mpsc_reg))) {
  1477. if (!(rc = driver_register(&mpsc_shared_driver))) {
  1478. if ((rc = driver_register(&mpsc_driver))) {
  1479. driver_unregister(&mpsc_shared_driver);
  1480. uart_unregister_driver(&mpsc_reg);
  1481. }
  1482. }
  1483. else
  1484. uart_unregister_driver(&mpsc_reg);
  1485. }
  1486. return rc;
  1487. }
  1488. static void __exit
  1489. mpsc_drv_exit(void)
  1490. {
  1491. driver_unregister(&mpsc_driver);
  1492. driver_unregister(&mpsc_shared_driver);
  1493. uart_unregister_driver(&mpsc_reg);
  1494. memset(mpsc_ports, 0, sizeof(mpsc_ports));
  1495. memset(&mpsc_shared_regs, 0, sizeof(mpsc_shared_regs));
  1496. return;
  1497. }
  1498. module_init(mpsc_drv_init);
  1499. module_exit(mpsc_drv_exit);
  1500. MODULE_AUTHOR("Mark A. Greer <mgreer@mvista.com>");
  1501. MODULE_DESCRIPTION("Generic Marvell MPSC serial/UART driver $Revision: 1.00 $");
  1502. MODULE_VERSION(MPSC_VERSION);
  1503. MODULE_LICENSE("GPL");
  1504. MODULE_ALIAS_CHARDEV_MAJOR(MPSC_MAJOR);