w83977af_ir.c 32 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372
  1. /*********************************************************************
  2. *
  3. * Filename: w83977af_ir.c
  4. * Version: 1.0
  5. * Description: FIR driver for the Winbond W83977AF Super I/O chip
  6. * Status: Experimental.
  7. * Author: Paul VanderSpek
  8. * Created at: Wed Nov 4 11:46:16 1998
  9. * Modified at: Fri Jan 28 12:10:59 2000
  10. * Modified by: Dag Brattli <dagb@cs.uit.no>
  11. *
  12. * Copyright (c) 1998-2000 Dag Brattli <dagb@cs.uit.no>
  13. * Copyright (c) 1998-1999 Rebel.com
  14. *
  15. * This program is free software; you can redistribute it and/or
  16. * modify it under the terms of the GNU General Public License as
  17. * published by the Free Software Foundation; either version 2 of
  18. * the License, or (at your option) any later version.
  19. *
  20. * Neither Paul VanderSpek nor Rebel.com admit liability nor provide
  21. * warranty for any of this software. This material is provided "AS-IS"
  22. * and at no charge.
  23. *
  24. * If you find bugs in this file, its very likely that the same bug
  25. * will also be in pc87108.c since the implementations are quite
  26. * similar.
  27. *
  28. * Notice that all functions that needs to access the chip in _any_
  29. * way, must save BSR register on entry, and restore it on exit.
  30. * It is _very_ important to follow this policy!
  31. *
  32. * __u8 bank;
  33. *
  34. * bank = inb( iobase+BSR);
  35. *
  36. * do_your_stuff_here();
  37. *
  38. * outb( bank, iobase+BSR);
  39. *
  40. ********************************************************************/
  41. #include <linux/module.h>
  42. #include <linux/kernel.h>
  43. #include <linux/types.h>
  44. #include <linux/skbuff.h>
  45. #include <linux/netdevice.h>
  46. #include <linux/ioport.h>
  47. #include <linux/delay.h>
  48. #include <linux/slab.h>
  49. #include <linux/init.h>
  50. #include <linux/rtnetlink.h>
  51. #include <linux/dma-mapping.h>
  52. #include <asm/io.h>
  53. #include <asm/dma.h>
  54. #include <asm/byteorder.h>
  55. #include <net/irda/irda.h>
  56. #include <net/irda/wrapper.h>
  57. #include <net/irda/irda_device.h>
  58. #include "w83977af.h"
  59. #include "w83977af_ir.h"
  60. #ifdef CONFIG_ARCH_NETWINDER /* Adjust to NetWinder differences */
  61. #undef CONFIG_NETWINDER_TX_DMA_PROBLEMS /* Not needed */
  62. #define CONFIG_NETWINDER_RX_DMA_PROBLEMS /* Must have this one! */
  63. #endif
  64. #undef CONFIG_USE_INTERNAL_TIMER /* Just cannot make that timer work */
  65. #define CONFIG_USE_W977_PNP /* Currently needed */
  66. #define PIO_MAX_SPEED 115200
  67. static char *driver_name = "w83977af_ir";
  68. static int qos_mtt_bits = 0x07; /* 1 ms or more */
  69. #define CHIP_IO_EXTENT 8
  70. static unsigned int io[] = { 0x180, ~0, ~0, ~0 };
  71. #ifdef CONFIG_ARCH_NETWINDER /* Adjust to NetWinder differences */
  72. static unsigned int irq[] = { 6, 0, 0, 0 };
  73. #else
  74. static unsigned int irq[] = { 11, 0, 0, 0 };
  75. #endif
  76. static unsigned int dma[] = { 1, 0, 0, 0 };
  77. static unsigned int efbase[] = { W977_EFIO_BASE, W977_EFIO2_BASE };
  78. static unsigned int efio = W977_EFIO_BASE;
  79. static struct w83977af_ir *dev_self[] = { NULL, NULL, NULL, NULL};
  80. /* Some prototypes */
  81. static int w83977af_open(int i, unsigned int iobase, unsigned int irq,
  82. unsigned int dma);
  83. static int w83977af_close(struct w83977af_ir *self);
  84. static int w83977af_probe(int iobase, int irq, int dma);
  85. static int w83977af_dma_receive(struct w83977af_ir *self);
  86. static int w83977af_dma_receive_complete(struct w83977af_ir *self);
  87. static int w83977af_hard_xmit(struct sk_buff *skb, struct net_device *dev);
  88. static int w83977af_pio_write(int iobase, __u8 *buf, int len, int fifo_size);
  89. static void w83977af_dma_write(struct w83977af_ir *self, int iobase);
  90. static void w83977af_change_speed(struct w83977af_ir *self, __u32 speed);
  91. static int w83977af_is_receiving(struct w83977af_ir *self);
  92. static int w83977af_net_open(struct net_device *dev);
  93. static int w83977af_net_close(struct net_device *dev);
  94. static int w83977af_net_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
  95. static struct net_device_stats *w83977af_net_get_stats(struct net_device *dev);
  96. /*
  97. * Function w83977af_init ()
  98. *
  99. * Initialize chip. Just try to find out how many chips we are dealing with
  100. * and where they are
  101. */
  102. static int __init w83977af_init(void)
  103. {
  104. int i;
  105. IRDA_DEBUG(0, "%s()\n", __FUNCTION__ );
  106. for (i=0; (io[i] < 2000) && (i < ARRAY_SIZE(dev_self)); i++) {
  107. if (w83977af_open(i, io[i], irq[i], dma[i]) == 0)
  108. return 0;
  109. }
  110. return -ENODEV;
  111. }
  112. /*
  113. * Function w83977af_cleanup ()
  114. *
  115. * Close all configured chips
  116. *
  117. */
  118. static void __exit w83977af_cleanup(void)
  119. {
  120. int i;
  121. IRDA_DEBUG(4, "%s()\n", __FUNCTION__ );
  122. for (i=0; i < ARRAY_SIZE(dev_self); i++) {
  123. if (dev_self[i])
  124. w83977af_close(dev_self[i]);
  125. }
  126. }
  127. /*
  128. * Function w83977af_open (iobase, irq)
  129. *
  130. * Open driver instance
  131. *
  132. */
  133. int w83977af_open(int i, unsigned int iobase, unsigned int irq,
  134. unsigned int dma)
  135. {
  136. struct net_device *dev;
  137. struct w83977af_ir *self;
  138. int err;
  139. IRDA_DEBUG(0, "%s()\n", __FUNCTION__ );
  140. /* Lock the port that we need */
  141. if (!request_region(iobase, CHIP_IO_EXTENT, driver_name)) {
  142. IRDA_DEBUG(0, "%s(), can't get iobase of 0x%03x\n",
  143. __FUNCTION__ , iobase);
  144. return -ENODEV;
  145. }
  146. if (w83977af_probe(iobase, irq, dma) == -1) {
  147. err = -1;
  148. goto err_out;
  149. }
  150. /*
  151. * Allocate new instance of the driver
  152. */
  153. dev = alloc_irdadev(sizeof(struct w83977af_ir));
  154. if (dev == NULL) {
  155. printk( KERN_ERR "IrDA: Can't allocate memory for "
  156. "IrDA control block!\n");
  157. err = -ENOMEM;
  158. goto err_out;
  159. }
  160. self = dev->priv;
  161. spin_lock_init(&self->lock);
  162. /* Initialize IO */
  163. self->io.fir_base = iobase;
  164. self->io.irq = irq;
  165. self->io.fir_ext = CHIP_IO_EXTENT;
  166. self->io.dma = dma;
  167. self->io.fifo_size = 32;
  168. /* Initialize QoS for this device */
  169. irda_init_max_qos_capabilies(&self->qos);
  170. /* The only value we must override it the baudrate */
  171. /* FIXME: The HP HDLS-1100 does not support 1152000! */
  172. self->qos.baud_rate.bits = IR_9600|IR_19200|IR_38400|IR_57600|
  173. IR_115200|IR_576000|IR_1152000|(IR_4000000 << 8);
  174. /* The HP HDLS-1100 needs 1 ms according to the specs */
  175. self->qos.min_turn_time.bits = qos_mtt_bits;
  176. irda_qos_bits_to_value(&self->qos);
  177. /* Max DMA buffer size needed = (data_size + 6) * (window_size) + 6; */
  178. self->rx_buff.truesize = 14384;
  179. self->tx_buff.truesize = 4000;
  180. /* Allocate memory if needed */
  181. self->rx_buff.head =
  182. dma_alloc_coherent(NULL, self->rx_buff.truesize,
  183. &self->rx_buff_dma, GFP_KERNEL);
  184. if (self->rx_buff.head == NULL) {
  185. err = -ENOMEM;
  186. goto err_out1;
  187. }
  188. memset(self->rx_buff.head, 0, self->rx_buff.truesize);
  189. self->tx_buff.head =
  190. dma_alloc_coherent(NULL, self->tx_buff.truesize,
  191. &self->tx_buff_dma, GFP_KERNEL);
  192. if (self->tx_buff.head == NULL) {
  193. err = -ENOMEM;
  194. goto err_out2;
  195. }
  196. memset(self->tx_buff.head, 0, self->tx_buff.truesize);
  197. self->rx_buff.in_frame = FALSE;
  198. self->rx_buff.state = OUTSIDE_FRAME;
  199. self->tx_buff.data = self->tx_buff.head;
  200. self->rx_buff.data = self->rx_buff.head;
  201. self->netdev = dev;
  202. /* Keep track of module usage */
  203. SET_MODULE_OWNER(dev);
  204. /* Override the network functions we need to use */
  205. dev->hard_start_xmit = w83977af_hard_xmit;
  206. dev->open = w83977af_net_open;
  207. dev->stop = w83977af_net_close;
  208. dev->do_ioctl = w83977af_net_ioctl;
  209. dev->get_stats = w83977af_net_get_stats;
  210. err = register_netdev(dev);
  211. if (err) {
  212. IRDA_ERROR("%s(), register_netdevice() failed!\n", __FUNCTION__);
  213. goto err_out3;
  214. }
  215. IRDA_MESSAGE("IrDA: Registered device %s\n", dev->name);
  216. /* Need to store self somewhere */
  217. dev_self[i] = self;
  218. return 0;
  219. err_out3:
  220. dma_free_coherent(NULL, self->tx_buff.truesize,
  221. self->tx_buff.head, self->tx_buff_dma);
  222. err_out2:
  223. dma_free_coherent(NULL, self->rx_buff.truesize,
  224. self->rx_buff.head, self->rx_buff_dma);
  225. err_out1:
  226. free_netdev(dev);
  227. err_out:
  228. release_region(iobase, CHIP_IO_EXTENT);
  229. return err;
  230. }
  231. /*
  232. * Function w83977af_close (self)
  233. *
  234. * Close driver instance
  235. *
  236. */
  237. static int w83977af_close(struct w83977af_ir *self)
  238. {
  239. int iobase;
  240. IRDA_DEBUG(0, "%s()\n", __FUNCTION__ );
  241. iobase = self->io.fir_base;
  242. #ifdef CONFIG_USE_W977_PNP
  243. /* enter PnP configuration mode */
  244. w977_efm_enter(efio);
  245. w977_select_device(W977_DEVICE_IR, efio);
  246. /* Deactivate device */
  247. w977_write_reg(0x30, 0x00, efio);
  248. w977_efm_exit(efio);
  249. #endif /* CONFIG_USE_W977_PNP */
  250. /* Remove netdevice */
  251. unregister_netdev(self->netdev);
  252. /* Release the PORT that this driver is using */
  253. IRDA_DEBUG(0 , "%s(), Releasing Region %03x\n",
  254. __FUNCTION__ , self->io.fir_base);
  255. release_region(self->io.fir_base, self->io.fir_ext);
  256. if (self->tx_buff.head)
  257. dma_free_coherent(NULL, self->tx_buff.truesize,
  258. self->tx_buff.head, self->tx_buff_dma);
  259. if (self->rx_buff.head)
  260. dma_free_coherent(NULL, self->rx_buff.truesize,
  261. self->rx_buff.head, self->rx_buff_dma);
  262. free_netdev(self->netdev);
  263. return 0;
  264. }
  265. int w83977af_probe( int iobase, int irq, int dma)
  266. {
  267. int version;
  268. int i;
  269. for (i=0; i < 2; i++) {
  270. IRDA_DEBUG( 0, "%s()\n", __FUNCTION__ );
  271. #ifdef CONFIG_USE_W977_PNP
  272. /* Enter PnP configuration mode */
  273. w977_efm_enter(efbase[i]);
  274. w977_select_device(W977_DEVICE_IR, efbase[i]);
  275. /* Configure PnP port, IRQ, and DMA channel */
  276. w977_write_reg(0x60, (iobase >> 8) & 0xff, efbase[i]);
  277. w977_write_reg(0x61, (iobase) & 0xff, efbase[i]);
  278. w977_write_reg(0x70, irq, efbase[i]);
  279. #ifdef CONFIG_ARCH_NETWINDER
  280. /* Netwinder uses 1 higher than Linux */
  281. w977_write_reg(0x74, dma+1, efbase[i]);
  282. #else
  283. w977_write_reg(0x74, dma, efbase[i]);
  284. #endif /*CONFIG_ARCH_NETWINDER */
  285. w977_write_reg(0x75, 0x04, efbase[i]); /* Disable Tx DMA */
  286. /* Set append hardware CRC, enable IR bank selection */
  287. w977_write_reg(0xf0, APEDCRC|ENBNKSEL, efbase[i]);
  288. /* Activate device */
  289. w977_write_reg(0x30, 0x01, efbase[i]);
  290. w977_efm_exit(efbase[i]);
  291. #endif /* CONFIG_USE_W977_PNP */
  292. /* Disable Advanced mode */
  293. switch_bank(iobase, SET2);
  294. outb(iobase+2, 0x00);
  295. /* Turn on UART (global) interrupts */
  296. switch_bank(iobase, SET0);
  297. outb(HCR_EN_IRQ, iobase+HCR);
  298. /* Switch to advanced mode */
  299. switch_bank(iobase, SET2);
  300. outb(inb(iobase+ADCR1) | ADCR1_ADV_SL, iobase+ADCR1);
  301. /* Set default IR-mode */
  302. switch_bank(iobase, SET0);
  303. outb(HCR_SIR, iobase+HCR);
  304. /* Read the Advanced IR ID */
  305. switch_bank(iobase, SET3);
  306. version = inb(iobase+AUID);
  307. /* Should be 0x1? */
  308. if (0x10 == (version & 0xf0)) {
  309. efio = efbase[i];
  310. /* Set FIFO size to 32 */
  311. switch_bank(iobase, SET2);
  312. outb(ADCR2_RXFS32|ADCR2_TXFS32, iobase+ADCR2);
  313. /* Set FIFO threshold to TX17, RX16 */
  314. switch_bank(iobase, SET0);
  315. outb(UFR_RXTL|UFR_TXTL|UFR_TXF_RST|UFR_RXF_RST|
  316. UFR_EN_FIFO,iobase+UFR);
  317. /* Receiver frame length */
  318. switch_bank(iobase, SET4);
  319. outb(2048 & 0xff, iobase+6);
  320. outb((2048 >> 8) & 0x1f, iobase+7);
  321. /*
  322. * Init HP HSDL-1100 transceiver.
  323. *
  324. * Set IRX_MSL since we have 2 * receive paths IRRX,
  325. * and IRRXH. Clear IRSL0D since we want IRSL0 * to
  326. * be a input pin used for IRRXH
  327. *
  328. * IRRX pin 37 connected to receiver
  329. * IRTX pin 38 connected to transmitter
  330. * FIRRX pin 39 connected to receiver (IRSL0)
  331. * CIRRX pin 40 connected to pin 37
  332. */
  333. switch_bank(iobase, SET7);
  334. outb(0x40, iobase+7);
  335. IRDA_MESSAGE("W83977AF (IR) driver loaded. "
  336. "Version: 0x%02x\n", version);
  337. return 0;
  338. } else {
  339. /* Try next extented function register address */
  340. IRDA_DEBUG( 0, "%s(), Wrong chip version", __FUNCTION__ );
  341. }
  342. }
  343. return -1;
  344. }
  345. void w83977af_change_speed(struct w83977af_ir *self, __u32 speed)
  346. {
  347. int ir_mode = HCR_SIR;
  348. int iobase;
  349. __u8 set;
  350. iobase = self->io.fir_base;
  351. /* Update accounting for new speed */
  352. self->io.speed = speed;
  353. /* Save current bank */
  354. set = inb(iobase+SSR);
  355. /* Disable interrupts */
  356. switch_bank(iobase, SET0);
  357. outb(0, iobase+ICR);
  358. /* Select Set 2 */
  359. switch_bank(iobase, SET2);
  360. outb(0x00, iobase+ABHL);
  361. switch (speed) {
  362. case 9600: outb(0x0c, iobase+ABLL); break;
  363. case 19200: outb(0x06, iobase+ABLL); break;
  364. case 38400: outb(0x03, iobase+ABLL); break;
  365. case 57600: outb(0x02, iobase+ABLL); break;
  366. case 115200: outb(0x01, iobase+ABLL); break;
  367. case 576000:
  368. ir_mode = HCR_MIR_576;
  369. IRDA_DEBUG(0, "%s(), handling baud of 576000\n", __FUNCTION__ );
  370. break;
  371. case 1152000:
  372. ir_mode = HCR_MIR_1152;
  373. IRDA_DEBUG(0, "%s(), handling baud of 1152000\n", __FUNCTION__ );
  374. break;
  375. case 4000000:
  376. ir_mode = HCR_FIR;
  377. IRDA_DEBUG(0, "%s(), handling baud of 4000000\n", __FUNCTION__ );
  378. break;
  379. default:
  380. ir_mode = HCR_FIR;
  381. IRDA_DEBUG(0, "%s(), unknown baud rate of %d\n", __FUNCTION__ , speed);
  382. break;
  383. }
  384. /* Set speed mode */
  385. switch_bank(iobase, SET0);
  386. outb(ir_mode, iobase+HCR);
  387. /* set FIFO size to 32 */
  388. switch_bank(iobase, SET2);
  389. outb(ADCR2_RXFS32|ADCR2_TXFS32, iobase+ADCR2);
  390. /* set FIFO threshold to TX17, RX16 */
  391. switch_bank(iobase, SET0);
  392. outb(0x00, iobase+UFR); /* Reset */
  393. outb(UFR_EN_FIFO, iobase+UFR); /* First we must enable FIFO */
  394. outb(0xa7, iobase+UFR);
  395. netif_wake_queue(self->netdev);
  396. /* Enable some interrupts so we can receive frames */
  397. switch_bank(iobase, SET0);
  398. if (speed > PIO_MAX_SPEED) {
  399. outb(ICR_EFSFI, iobase+ICR);
  400. w83977af_dma_receive(self);
  401. } else
  402. outb(ICR_ERBRI, iobase+ICR);
  403. /* Restore SSR */
  404. outb(set, iobase+SSR);
  405. }
  406. /*
  407. * Function w83977af_hard_xmit (skb, dev)
  408. *
  409. * Sets up a DMA transfer to send the current frame.
  410. *
  411. */
  412. int w83977af_hard_xmit(struct sk_buff *skb, struct net_device *dev)
  413. {
  414. struct w83977af_ir *self;
  415. __s32 speed;
  416. int iobase;
  417. __u8 set;
  418. int mtt;
  419. self = (struct w83977af_ir *) dev->priv;
  420. iobase = self->io.fir_base;
  421. IRDA_DEBUG(4, "%s(%ld), skb->len=%d\n", __FUNCTION__ , jiffies,
  422. (int) skb->len);
  423. /* Lock transmit buffer */
  424. netif_stop_queue(dev);
  425. /* Check if we need to change the speed */
  426. speed = irda_get_next_speed(skb);
  427. if ((speed != self->io.speed) && (speed != -1)) {
  428. /* Check for empty frame */
  429. if (!skb->len) {
  430. w83977af_change_speed(self, speed);
  431. dev->trans_start = jiffies;
  432. dev_kfree_skb(skb);
  433. return 0;
  434. } else
  435. self->new_speed = speed;
  436. }
  437. /* Save current set */
  438. set = inb(iobase+SSR);
  439. /* Decide if we should use PIO or DMA transfer */
  440. if (self->io.speed > PIO_MAX_SPEED) {
  441. self->tx_buff.data = self->tx_buff.head;
  442. memcpy(self->tx_buff.data, skb->data, skb->len);
  443. self->tx_buff.len = skb->len;
  444. mtt = irda_get_mtt(skb);
  445. #ifdef CONFIG_USE_INTERNAL_TIMER
  446. if (mtt > 50) {
  447. /* Adjust for timer resolution */
  448. mtt /= 1000+1;
  449. /* Setup timer */
  450. switch_bank(iobase, SET4);
  451. outb(mtt & 0xff, iobase+TMRL);
  452. outb((mtt >> 8) & 0x0f, iobase+TMRH);
  453. /* Start timer */
  454. outb(IR_MSL_EN_TMR, iobase+IR_MSL);
  455. self->io.direction = IO_XMIT;
  456. /* Enable timer interrupt */
  457. switch_bank(iobase, SET0);
  458. outb(ICR_ETMRI, iobase+ICR);
  459. } else {
  460. #endif
  461. IRDA_DEBUG(4, "%s(%ld), mtt=%d\n", __FUNCTION__ , jiffies, mtt);
  462. if (mtt)
  463. udelay(mtt);
  464. /* Enable DMA interrupt */
  465. switch_bank(iobase, SET0);
  466. outb(ICR_EDMAI, iobase+ICR);
  467. w83977af_dma_write(self, iobase);
  468. #ifdef CONFIG_USE_INTERNAL_TIMER
  469. }
  470. #endif
  471. } else {
  472. self->tx_buff.data = self->tx_buff.head;
  473. self->tx_buff.len = async_wrap_skb(skb, self->tx_buff.data,
  474. self->tx_buff.truesize);
  475. /* Add interrupt on tx low level (will fire immediately) */
  476. switch_bank(iobase, SET0);
  477. outb(ICR_ETXTHI, iobase+ICR);
  478. }
  479. dev->trans_start = jiffies;
  480. dev_kfree_skb(skb);
  481. /* Restore set register */
  482. outb(set, iobase+SSR);
  483. return 0;
  484. }
  485. /*
  486. * Function w83977af_dma_write (self, iobase)
  487. *
  488. * Send frame using DMA
  489. *
  490. */
  491. static void w83977af_dma_write(struct w83977af_ir *self, int iobase)
  492. {
  493. __u8 set;
  494. #ifdef CONFIG_NETWINDER_TX_DMA_PROBLEMS
  495. unsigned long flags;
  496. __u8 hcr;
  497. #endif
  498. IRDA_DEBUG(4, "%s(), len=%d\n", __FUNCTION__ , self->tx_buff.len);
  499. /* Save current set */
  500. set = inb(iobase+SSR);
  501. /* Disable DMA */
  502. switch_bank(iobase, SET0);
  503. outb(inb(iobase+HCR) & ~HCR_EN_DMA, iobase+HCR);
  504. /* Choose transmit DMA channel */
  505. switch_bank(iobase, SET2);
  506. outb(ADCR1_D_CHSW|/*ADCR1_DMA_F|*/ADCR1_ADV_SL, iobase+ADCR1);
  507. #ifdef CONFIG_NETWINDER_TX_DMA_PROBLEMS
  508. spin_lock_irqsave(&self->lock, flags);
  509. disable_dma(self->io.dma);
  510. clear_dma_ff(self->io.dma);
  511. set_dma_mode(self->io.dma, DMA_MODE_READ);
  512. set_dma_addr(self->io.dma, self->tx_buff_dma);
  513. set_dma_count(self->io.dma, self->tx_buff.len);
  514. #else
  515. irda_setup_dma(self->io.dma, self->tx_buff_dma, self->tx_buff.len,
  516. DMA_MODE_WRITE);
  517. #endif
  518. self->io.direction = IO_XMIT;
  519. /* Enable DMA */
  520. switch_bank(iobase, SET0);
  521. #ifdef CONFIG_NETWINDER_TX_DMA_PROBLEMS
  522. hcr = inb(iobase+HCR);
  523. outb(hcr | HCR_EN_DMA, iobase+HCR);
  524. enable_dma(self->io.dma);
  525. spin_unlock_irqrestore(&self->lock, flags);
  526. #else
  527. outb(inb(iobase+HCR) | HCR_EN_DMA | HCR_TX_WT, iobase+HCR);
  528. #endif
  529. /* Restore set register */
  530. outb(set, iobase+SSR);
  531. }
  532. /*
  533. * Function w83977af_pio_write (iobase, buf, len, fifo_size)
  534. *
  535. *
  536. *
  537. */
  538. static int w83977af_pio_write(int iobase, __u8 *buf, int len, int fifo_size)
  539. {
  540. int actual = 0;
  541. __u8 set;
  542. IRDA_DEBUG(4, "%s()\n", __FUNCTION__ );
  543. /* Save current bank */
  544. set = inb(iobase+SSR);
  545. switch_bank(iobase, SET0);
  546. if (!(inb_p(iobase+USR) & USR_TSRE)) {
  547. IRDA_DEBUG(4,
  548. "%s(), warning, FIFO not empty yet!\n", __FUNCTION__ );
  549. fifo_size -= 17;
  550. IRDA_DEBUG(4, "%s(), %d bytes left in tx fifo\n",
  551. __FUNCTION__ , fifo_size);
  552. }
  553. /* Fill FIFO with current frame */
  554. while ((fifo_size-- > 0) && (actual < len)) {
  555. /* Transmit next byte */
  556. outb(buf[actual++], iobase+TBR);
  557. }
  558. IRDA_DEBUG(4, "%s(), fifo_size %d ; %d sent of %d\n",
  559. __FUNCTION__ , fifo_size, actual, len);
  560. /* Restore bank */
  561. outb(set, iobase+SSR);
  562. return actual;
  563. }
  564. /*
  565. * Function w83977af_dma_xmit_complete (self)
  566. *
  567. * The transfer of a frame in finished. So do the necessary things
  568. *
  569. *
  570. */
  571. static void w83977af_dma_xmit_complete(struct w83977af_ir *self)
  572. {
  573. int iobase;
  574. __u8 set;
  575. IRDA_DEBUG(4, "%s(%ld)\n", __FUNCTION__ , jiffies);
  576. IRDA_ASSERT(self != NULL, return;);
  577. iobase = self->io.fir_base;
  578. /* Save current set */
  579. set = inb(iobase+SSR);
  580. /* Disable DMA */
  581. switch_bank(iobase, SET0);
  582. outb(inb(iobase+HCR) & ~HCR_EN_DMA, iobase+HCR);
  583. /* Check for underrrun! */
  584. if (inb(iobase+AUDR) & AUDR_UNDR) {
  585. IRDA_DEBUG(0, "%s(), Transmit underrun!\n", __FUNCTION__ );
  586. self->stats.tx_errors++;
  587. self->stats.tx_fifo_errors++;
  588. /* Clear bit, by writing 1 to it */
  589. outb(AUDR_UNDR, iobase+AUDR);
  590. } else
  591. self->stats.tx_packets++;
  592. if (self->new_speed) {
  593. w83977af_change_speed(self, self->new_speed);
  594. self->new_speed = 0;
  595. }
  596. /* Unlock tx_buff and request another frame */
  597. /* Tell the network layer, that we want more frames */
  598. netif_wake_queue(self->netdev);
  599. /* Restore set */
  600. outb(set, iobase+SSR);
  601. }
  602. /*
  603. * Function w83977af_dma_receive (self)
  604. *
  605. * Get ready for receiving a frame. The device will initiate a DMA
  606. * if it starts to receive a frame.
  607. *
  608. */
  609. int w83977af_dma_receive(struct w83977af_ir *self)
  610. {
  611. int iobase;
  612. __u8 set;
  613. #ifdef CONFIG_NETWINDER_RX_DMA_PROBLEMS
  614. unsigned long flags;
  615. __u8 hcr;
  616. #endif
  617. IRDA_ASSERT(self != NULL, return -1;);
  618. IRDA_DEBUG(4, "%s\n", __FUNCTION__ );
  619. iobase= self->io.fir_base;
  620. /* Save current set */
  621. set = inb(iobase+SSR);
  622. /* Disable DMA */
  623. switch_bank(iobase, SET0);
  624. outb(inb(iobase+HCR) & ~HCR_EN_DMA, iobase+HCR);
  625. /* Choose DMA Rx, DMA Fairness, and Advanced mode */
  626. switch_bank(iobase, SET2);
  627. outb((inb(iobase+ADCR1) & ~ADCR1_D_CHSW)/*|ADCR1_DMA_F*/|ADCR1_ADV_SL,
  628. iobase+ADCR1);
  629. self->io.direction = IO_RECV;
  630. self->rx_buff.data = self->rx_buff.head;
  631. #ifdef CONFIG_NETWINDER_RX_DMA_PROBLEMS
  632. spin_lock_irqsave(&self->lock, flags);
  633. disable_dma(self->io.dma);
  634. clear_dma_ff(self->io.dma);
  635. set_dma_mode(self->io.dma, DMA_MODE_READ);
  636. set_dma_addr(self->io.dma, self->rx_buff_dma);
  637. set_dma_count(self->io.dma, self->rx_buff.truesize);
  638. #else
  639. irda_setup_dma(self->io.dma, self->rx_buff_dma, self->rx_buff.truesize,
  640. DMA_MODE_READ);
  641. #endif
  642. /*
  643. * Reset Rx FIFO. This will also flush the ST_FIFO, it's very
  644. * important that we don't reset the Tx FIFO since it might not
  645. * be finished transmitting yet
  646. */
  647. switch_bank(iobase, SET0);
  648. outb(UFR_RXTL|UFR_TXTL|UFR_RXF_RST|UFR_EN_FIFO, iobase+UFR);
  649. self->st_fifo.len = self->st_fifo.tail = self->st_fifo.head = 0;
  650. /* Enable DMA */
  651. switch_bank(iobase, SET0);
  652. #ifdef CONFIG_NETWINDER_RX_DMA_PROBLEMS
  653. hcr = inb(iobase+HCR);
  654. outb(hcr | HCR_EN_DMA, iobase+HCR);
  655. enable_dma(self->io.dma);
  656. spin_unlock_irqrestore(&self->lock, flags);
  657. #else
  658. outb(inb(iobase+HCR) | HCR_EN_DMA, iobase+HCR);
  659. #endif
  660. /* Restore set */
  661. outb(set, iobase+SSR);
  662. return 0;
  663. }
  664. /*
  665. * Function w83977af_receive_complete (self)
  666. *
  667. * Finished with receiving a frame
  668. *
  669. */
  670. int w83977af_dma_receive_complete(struct w83977af_ir *self)
  671. {
  672. struct sk_buff *skb;
  673. struct st_fifo *st_fifo;
  674. int len;
  675. int iobase;
  676. __u8 set;
  677. __u8 status;
  678. IRDA_DEBUG(4, "%s\n", __FUNCTION__ );
  679. st_fifo = &self->st_fifo;
  680. iobase = self->io.fir_base;
  681. /* Save current set */
  682. set = inb(iobase+SSR);
  683. iobase = self->io.fir_base;
  684. /* Read status FIFO */
  685. switch_bank(iobase, SET5);
  686. while ((status = inb(iobase+FS_FO)) & FS_FO_FSFDR) {
  687. st_fifo->entries[st_fifo->tail].status = status;
  688. st_fifo->entries[st_fifo->tail].len = inb(iobase+RFLFL);
  689. st_fifo->entries[st_fifo->tail].len |= inb(iobase+RFLFH) << 8;
  690. st_fifo->tail++;
  691. st_fifo->len++;
  692. }
  693. while (st_fifo->len) {
  694. /* Get first entry */
  695. status = st_fifo->entries[st_fifo->head].status;
  696. len = st_fifo->entries[st_fifo->head].len;
  697. st_fifo->head++;
  698. st_fifo->len--;
  699. /* Check for errors */
  700. if (status & FS_FO_ERR_MSK) {
  701. if (status & FS_FO_LST_FR) {
  702. /* Add number of lost frames to stats */
  703. self->stats.rx_errors += len;
  704. } else {
  705. /* Skip frame */
  706. self->stats.rx_errors++;
  707. self->rx_buff.data += len;
  708. if (status & FS_FO_MX_LEX)
  709. self->stats.rx_length_errors++;
  710. if (status & FS_FO_PHY_ERR)
  711. self->stats.rx_frame_errors++;
  712. if (status & FS_FO_CRC_ERR)
  713. self->stats.rx_crc_errors++;
  714. }
  715. /* The errors below can be reported in both cases */
  716. if (status & FS_FO_RX_OV)
  717. self->stats.rx_fifo_errors++;
  718. if (status & FS_FO_FSF_OV)
  719. self->stats.rx_fifo_errors++;
  720. } else {
  721. /* Check if we have transferred all data to memory */
  722. switch_bank(iobase, SET0);
  723. if (inb(iobase+USR) & USR_RDR) {
  724. #ifdef CONFIG_USE_INTERNAL_TIMER
  725. /* Put this entry back in fifo */
  726. st_fifo->head--;
  727. st_fifo->len++;
  728. st_fifo->entries[st_fifo->head].status = status;
  729. st_fifo->entries[st_fifo->head].len = len;
  730. /* Restore set register */
  731. outb(set, iobase+SSR);
  732. return FALSE; /* I'll be back! */
  733. #else
  734. udelay(80); /* Should be enough!? */
  735. #endif
  736. }
  737. skb = dev_alloc_skb(len+1);
  738. if (skb == NULL) {
  739. printk(KERN_INFO
  740. "%s(), memory squeeze, dropping frame.\n", __FUNCTION__);
  741. /* Restore set register */
  742. outb(set, iobase+SSR);
  743. return FALSE;
  744. }
  745. /* Align to 20 bytes */
  746. skb_reserve(skb, 1);
  747. /* Copy frame without CRC */
  748. if (self->io.speed < 4000000) {
  749. skb_put(skb, len-2);
  750. memcpy(skb->data, self->rx_buff.data, len-2);
  751. } else {
  752. skb_put(skb, len-4);
  753. memcpy(skb->data, self->rx_buff.data, len-4);
  754. }
  755. /* Move to next frame */
  756. self->rx_buff.data += len;
  757. self->stats.rx_packets++;
  758. skb->dev = self->netdev;
  759. skb->mac.raw = skb->data;
  760. skb->protocol = htons(ETH_P_IRDA);
  761. netif_rx(skb);
  762. self->netdev->last_rx = jiffies;
  763. }
  764. }
  765. /* Restore set register */
  766. outb(set, iobase+SSR);
  767. return TRUE;
  768. }
  769. /*
  770. * Function pc87108_pio_receive (self)
  771. *
  772. * Receive all data in receiver FIFO
  773. *
  774. */
  775. static void w83977af_pio_receive(struct w83977af_ir *self)
  776. {
  777. __u8 byte = 0x00;
  778. int iobase;
  779. IRDA_DEBUG(4, "%s()\n", __FUNCTION__ );
  780. IRDA_ASSERT(self != NULL, return;);
  781. iobase = self->io.fir_base;
  782. /* Receive all characters in Rx FIFO */
  783. do {
  784. byte = inb(iobase+RBR);
  785. async_unwrap_char(self->netdev, &self->stats, &self->rx_buff,
  786. byte);
  787. } while (inb(iobase+USR) & USR_RDR); /* Data available */
  788. }
  789. /*
  790. * Function w83977af_sir_interrupt (self, eir)
  791. *
  792. * Handle SIR interrupt
  793. *
  794. */
  795. static __u8 w83977af_sir_interrupt(struct w83977af_ir *self, int isr)
  796. {
  797. int actual;
  798. __u8 new_icr = 0;
  799. __u8 set;
  800. int iobase;
  801. IRDA_DEBUG(4, "%s(), isr=%#x\n", __FUNCTION__ , isr);
  802. iobase = self->io.fir_base;
  803. /* Transmit FIFO low on data */
  804. if (isr & ISR_TXTH_I) {
  805. /* Write data left in transmit buffer */
  806. actual = w83977af_pio_write(self->io.fir_base,
  807. self->tx_buff.data,
  808. self->tx_buff.len,
  809. self->io.fifo_size);
  810. self->tx_buff.data += actual;
  811. self->tx_buff.len -= actual;
  812. self->io.direction = IO_XMIT;
  813. /* Check if finished */
  814. if (self->tx_buff.len > 0) {
  815. new_icr |= ICR_ETXTHI;
  816. } else {
  817. set = inb(iobase+SSR);
  818. switch_bank(iobase, SET0);
  819. outb(AUDR_SFEND, iobase+AUDR);
  820. outb(set, iobase+SSR);
  821. self->stats.tx_packets++;
  822. /* Feed me more packets */
  823. netif_wake_queue(self->netdev);
  824. new_icr |= ICR_ETBREI;
  825. }
  826. }
  827. /* Check if transmission has completed */
  828. if (isr & ISR_TXEMP_I) {
  829. /* Check if we need to change the speed? */
  830. if (self->new_speed) {
  831. IRDA_DEBUG(2,
  832. "%s(), Changing speed!\n", __FUNCTION__ );
  833. w83977af_change_speed(self, self->new_speed);
  834. self->new_speed = 0;
  835. }
  836. /* Turn around and get ready to receive some data */
  837. self->io.direction = IO_RECV;
  838. new_icr |= ICR_ERBRI;
  839. }
  840. /* Rx FIFO threshold or timeout */
  841. if (isr & ISR_RXTH_I) {
  842. w83977af_pio_receive(self);
  843. /* Keep receiving */
  844. new_icr |= ICR_ERBRI;
  845. }
  846. return new_icr;
  847. }
  848. /*
  849. * Function pc87108_fir_interrupt (self, eir)
  850. *
  851. * Handle MIR/FIR interrupt
  852. *
  853. */
  854. static __u8 w83977af_fir_interrupt(struct w83977af_ir *self, int isr)
  855. {
  856. __u8 new_icr = 0;
  857. __u8 set;
  858. int iobase;
  859. iobase = self->io.fir_base;
  860. set = inb(iobase+SSR);
  861. /* End of frame detected in FIFO */
  862. if (isr & (ISR_FEND_I|ISR_FSF_I)) {
  863. if (w83977af_dma_receive_complete(self)) {
  864. /* Wait for next status FIFO interrupt */
  865. new_icr |= ICR_EFSFI;
  866. } else {
  867. /* DMA not finished yet */
  868. /* Set timer value, resolution 1 ms */
  869. switch_bank(iobase, SET4);
  870. outb(0x01, iobase+TMRL); /* 1 ms */
  871. outb(0x00, iobase+TMRH);
  872. /* Start timer */
  873. outb(IR_MSL_EN_TMR, iobase+IR_MSL);
  874. new_icr |= ICR_ETMRI;
  875. }
  876. }
  877. /* Timer finished */
  878. if (isr & ISR_TMR_I) {
  879. /* Disable timer */
  880. switch_bank(iobase, SET4);
  881. outb(0, iobase+IR_MSL);
  882. /* Clear timer event */
  883. /* switch_bank(iobase, SET0); */
  884. /* outb(ASCR_CTE, iobase+ASCR); */
  885. /* Check if this is a TX timer interrupt */
  886. if (self->io.direction == IO_XMIT) {
  887. w83977af_dma_write(self, iobase);
  888. new_icr |= ICR_EDMAI;
  889. } else {
  890. /* Check if DMA has now finished */
  891. w83977af_dma_receive_complete(self);
  892. new_icr |= ICR_EFSFI;
  893. }
  894. }
  895. /* Finished with DMA */
  896. if (isr & ISR_DMA_I) {
  897. w83977af_dma_xmit_complete(self);
  898. /* Check if there are more frames to be transmitted */
  899. /* if (irda_device_txqueue_empty(self)) { */
  900. /* Prepare for receive
  901. *
  902. * ** Netwinder Tx DMA likes that we do this anyway **
  903. */
  904. w83977af_dma_receive(self);
  905. new_icr = ICR_EFSFI;
  906. /* } */
  907. }
  908. /* Restore set */
  909. outb(set, iobase+SSR);
  910. return new_icr;
  911. }
  912. /*
  913. * Function w83977af_interrupt (irq, dev_id, regs)
  914. *
  915. * An interrupt from the chip has arrived. Time to do some work
  916. *
  917. */
  918. static irqreturn_t w83977af_interrupt(int irq, void *dev_id)
  919. {
  920. struct net_device *dev = dev_id;
  921. struct w83977af_ir *self;
  922. __u8 set, icr, isr;
  923. int iobase;
  924. self = dev->priv;
  925. iobase = self->io.fir_base;
  926. /* Save current bank */
  927. set = inb(iobase+SSR);
  928. switch_bank(iobase, SET0);
  929. icr = inb(iobase+ICR);
  930. isr = inb(iobase+ISR) & icr; /* Mask out the interesting ones */
  931. outb(0, iobase+ICR); /* Disable interrupts */
  932. if (isr) {
  933. /* Dispatch interrupt handler for the current speed */
  934. if (self->io.speed > PIO_MAX_SPEED )
  935. icr = w83977af_fir_interrupt(self, isr);
  936. else
  937. icr = w83977af_sir_interrupt(self, isr);
  938. }
  939. outb(icr, iobase+ICR); /* Restore (new) interrupts */
  940. outb(set, iobase+SSR); /* Restore bank register */
  941. return IRQ_RETVAL(isr);
  942. }
  943. /*
  944. * Function w83977af_is_receiving (self)
  945. *
  946. * Return TRUE is we are currently receiving a frame
  947. *
  948. */
  949. static int w83977af_is_receiving(struct w83977af_ir *self)
  950. {
  951. int status = FALSE;
  952. int iobase;
  953. __u8 set;
  954. IRDA_ASSERT(self != NULL, return FALSE;);
  955. if (self->io.speed > 115200) {
  956. iobase = self->io.fir_base;
  957. /* Check if rx FIFO is not empty */
  958. set = inb(iobase+SSR);
  959. switch_bank(iobase, SET2);
  960. if ((inb(iobase+RXFDTH) & 0x3f) != 0) {
  961. /* We are receiving something */
  962. status = TRUE;
  963. }
  964. outb(set, iobase+SSR);
  965. } else
  966. status = (self->rx_buff.state != OUTSIDE_FRAME);
  967. return status;
  968. }
  969. /*
  970. * Function w83977af_net_open (dev)
  971. *
  972. * Start the device
  973. *
  974. */
  975. static int w83977af_net_open(struct net_device *dev)
  976. {
  977. struct w83977af_ir *self;
  978. int iobase;
  979. char hwname[32];
  980. __u8 set;
  981. IRDA_DEBUG(0, "%s()\n", __FUNCTION__ );
  982. IRDA_ASSERT(dev != NULL, return -1;);
  983. self = (struct w83977af_ir *) dev->priv;
  984. IRDA_ASSERT(self != NULL, return 0;);
  985. iobase = self->io.fir_base;
  986. if (request_irq(self->io.irq, w83977af_interrupt, 0, dev->name,
  987. (void *) dev)) {
  988. return -EAGAIN;
  989. }
  990. /*
  991. * Always allocate the DMA channel after the IRQ,
  992. * and clean up on failure.
  993. */
  994. if (request_dma(self->io.dma, dev->name)) {
  995. free_irq(self->io.irq, self);
  996. return -EAGAIN;
  997. }
  998. /* Save current set */
  999. set = inb(iobase+SSR);
  1000. /* Enable some interrupts so we can receive frames again */
  1001. switch_bank(iobase, SET0);
  1002. if (self->io.speed > 115200) {
  1003. outb(ICR_EFSFI, iobase+ICR);
  1004. w83977af_dma_receive(self);
  1005. } else
  1006. outb(ICR_ERBRI, iobase+ICR);
  1007. /* Restore bank register */
  1008. outb(set, iobase+SSR);
  1009. /* Ready to play! */
  1010. netif_start_queue(dev);
  1011. /* Give self a hardware name */
  1012. sprintf(hwname, "w83977af @ 0x%03x", self->io.fir_base);
  1013. /*
  1014. * Open new IrLAP layer instance, now that everything should be
  1015. * initialized properly
  1016. */
  1017. self->irlap = irlap_open(dev, &self->qos, hwname);
  1018. return 0;
  1019. }
  1020. /*
  1021. * Function w83977af_net_close (dev)
  1022. *
  1023. * Stop the device
  1024. *
  1025. */
  1026. static int w83977af_net_close(struct net_device *dev)
  1027. {
  1028. struct w83977af_ir *self;
  1029. int iobase;
  1030. __u8 set;
  1031. IRDA_DEBUG(0, "%s()\n", __FUNCTION__ );
  1032. IRDA_ASSERT(dev != NULL, return -1;);
  1033. self = (struct w83977af_ir *) dev->priv;
  1034. IRDA_ASSERT(self != NULL, return 0;);
  1035. iobase = self->io.fir_base;
  1036. /* Stop device */
  1037. netif_stop_queue(dev);
  1038. /* Stop and remove instance of IrLAP */
  1039. if (self->irlap)
  1040. irlap_close(self->irlap);
  1041. self->irlap = NULL;
  1042. disable_dma(self->io.dma);
  1043. /* Save current set */
  1044. set = inb(iobase+SSR);
  1045. /* Disable interrupts */
  1046. switch_bank(iobase, SET0);
  1047. outb(0, iobase+ICR);
  1048. free_irq(self->io.irq, dev);
  1049. free_dma(self->io.dma);
  1050. /* Restore bank register */
  1051. outb(set, iobase+SSR);
  1052. return 0;
  1053. }
  1054. /*
  1055. * Function w83977af_net_ioctl (dev, rq, cmd)
  1056. *
  1057. * Process IOCTL commands for this device
  1058. *
  1059. */
  1060. static int w83977af_net_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
  1061. {
  1062. struct if_irda_req *irq = (struct if_irda_req *) rq;
  1063. struct w83977af_ir *self;
  1064. unsigned long flags;
  1065. int ret = 0;
  1066. IRDA_ASSERT(dev != NULL, return -1;);
  1067. self = dev->priv;
  1068. IRDA_ASSERT(self != NULL, return -1;);
  1069. IRDA_DEBUG(2, "%s(), %s, (cmd=0x%X)\n", __FUNCTION__ , dev->name, cmd);
  1070. spin_lock_irqsave(&self->lock, flags);
  1071. switch (cmd) {
  1072. case SIOCSBANDWIDTH: /* Set bandwidth */
  1073. if (!capable(CAP_NET_ADMIN)) {
  1074. ret = -EPERM;
  1075. goto out;
  1076. }
  1077. w83977af_change_speed(self, irq->ifr_baudrate);
  1078. break;
  1079. case SIOCSMEDIABUSY: /* Set media busy */
  1080. if (!capable(CAP_NET_ADMIN)) {
  1081. ret = -EPERM;
  1082. goto out;
  1083. }
  1084. irda_device_set_media_busy(self->netdev, TRUE);
  1085. break;
  1086. case SIOCGRECEIVING: /* Check if we are receiving right now */
  1087. irq->ifr_receiving = w83977af_is_receiving(self);
  1088. break;
  1089. default:
  1090. ret = -EOPNOTSUPP;
  1091. }
  1092. out:
  1093. spin_unlock_irqrestore(&self->lock, flags);
  1094. return ret;
  1095. }
  1096. static struct net_device_stats *w83977af_net_get_stats(struct net_device *dev)
  1097. {
  1098. struct w83977af_ir *self = (struct w83977af_ir *) dev->priv;
  1099. return &self->stats;
  1100. }
  1101. MODULE_AUTHOR("Dag Brattli <dagb@cs.uit.no>");
  1102. MODULE_DESCRIPTION("Winbond W83977AF IrDA Device Driver");
  1103. MODULE_LICENSE("GPL");
  1104. module_param(qos_mtt_bits, int, 0);
  1105. MODULE_PARM_DESC(qos_mtt_bits, "Mimimum Turn Time");
  1106. module_param_array(io, int, NULL, 0);
  1107. MODULE_PARM_DESC(io, "Base I/O addresses");
  1108. module_param_array(irq, int, NULL, 0);
  1109. MODULE_PARM_DESC(irq, "IRQ lines");
  1110. /*
  1111. * Function init_module (void)
  1112. *
  1113. *
  1114. *
  1115. */
  1116. module_init(w83977af_init);
  1117. /*
  1118. * Function cleanup_module (void)
  1119. *
  1120. *
  1121. *
  1122. */
  1123. module_exit(w83977af_cleanup);