dmabuf.c 35 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269
  1. /*
  2. * sound/oss/dmabuf.c
  3. *
  4. * The DMA buffer manager for digitized voice applications
  5. */
  6. /*
  7. * Copyright (C) by Hannu Savolainen 1993-1997
  8. *
  9. * OSS/Free for Linux is distributed under the GNU GENERAL PUBLIC LICENSE (GPL)
  10. * Version 2 (June 1991). See the "COPYING" file distributed with this software
  11. * for more info.
  12. *
  13. * Thomas Sailer : moved several static variables into struct audio_operations
  14. * (which is grossly misnamed btw.) because they have the same
  15. * lifetime as the rest in there and dynamic allocation saves
  16. * 12k or so
  17. * Thomas Sailer : remove {in,out}_sleep_flag. It was used for the sleeper to
  18. * determine if it was woken up by the expiring timeout or by
  19. * an explicit wake_up. The return value from schedule_timeout
  20. * can be used instead; if 0, the wakeup was due to the timeout.
  21. *
  22. * Rob Riggs Added persistent DMA buffers (1998/10/17)
  23. */
  24. #define BE_CONSERVATIVE
  25. #define SAMPLE_ROUNDUP 0
  26. #include <linux/mm.h>
  27. #include <linux/gfp.h>
  28. #include "sound_config.h"
  29. #define DMAP_FREE_ON_CLOSE 0
  30. #define DMAP_KEEP_ON_CLOSE 1
  31. extern int sound_dmap_flag;
  32. static void dma_reset_output(int dev);
  33. static void dma_reset_input(int dev);
  34. static int local_start_dma(struct audio_operations *adev, unsigned long physaddr, int count, int dma_mode);
  35. static int debugmem; /* switched off by default */
  36. static int dma_buffsize = DSP_BUFFSIZE;
  37. static long dmabuf_timeout(struct dma_buffparms *dmap)
  38. {
  39. long tmout;
  40. tmout = (dmap->fragment_size * HZ) / dmap->data_rate;
  41. tmout += HZ / 5; /* Some safety distance */
  42. if (tmout < (HZ / 2))
  43. tmout = HZ / 2;
  44. if (tmout > 20 * HZ)
  45. tmout = 20 * HZ;
  46. return tmout;
  47. }
  48. static int sound_alloc_dmap(struct dma_buffparms *dmap)
  49. {
  50. char *start_addr, *end_addr;
  51. int dma_pagesize;
  52. int sz, size;
  53. struct page *page;
  54. dmap->mapping_flags &= ~DMA_MAP_MAPPED;
  55. if (dmap->raw_buf != NULL)
  56. return 0; /* Already done */
  57. if (dma_buffsize < 4096)
  58. dma_buffsize = 4096;
  59. dma_pagesize = (dmap->dma < 4) ? (64 * 1024) : (128 * 1024);
  60. /*
  61. * Now check for the Cyrix problem.
  62. */
  63. if(isa_dma_bridge_buggy==2)
  64. dma_pagesize=32768;
  65. dmap->raw_buf = NULL;
  66. dmap->buffsize = dma_buffsize;
  67. if (dmap->buffsize > dma_pagesize)
  68. dmap->buffsize = dma_pagesize;
  69. start_addr = NULL;
  70. /*
  71. * Now loop until we get a free buffer. Try to get smaller buffer if
  72. * it fails. Don't accept smaller than 8k buffer for performance
  73. * reasons.
  74. */
  75. while (start_addr == NULL && dmap->buffsize > PAGE_SIZE) {
  76. for (sz = 0, size = PAGE_SIZE; size < dmap->buffsize; sz++, size <<= 1);
  77. dmap->buffsize = PAGE_SIZE * (1 << sz);
  78. start_addr = (char *) __get_free_pages(GFP_ATOMIC|GFP_DMA|__GFP_NOWARN, sz);
  79. if (start_addr == NULL)
  80. dmap->buffsize /= 2;
  81. }
  82. if (start_addr == NULL) {
  83. printk(KERN_WARNING "Sound error: Couldn't allocate DMA buffer\n");
  84. return -ENOMEM;
  85. } else {
  86. /* make some checks */
  87. end_addr = start_addr + dmap->buffsize - 1;
  88. if (debugmem)
  89. printk(KERN_DEBUG "sound: start 0x%lx, end 0x%lx\n", (long) start_addr, (long) end_addr);
  90. /* now check if it fits into the same dma-pagesize */
  91. if (((long) start_addr & ~(dma_pagesize - 1)) != ((long) end_addr & ~(dma_pagesize - 1))
  92. || end_addr >= (char *) (MAX_DMA_ADDRESS)) {
  93. printk(KERN_ERR "sound: Got invalid address 0x%lx for %db DMA-buffer\n", (long) start_addr, dmap->buffsize);
  94. return -EFAULT;
  95. }
  96. }
  97. dmap->raw_buf = start_addr;
  98. dmap->raw_buf_phys = dma_map_single(NULL, start_addr, dmap->buffsize, DMA_BIDIRECTIONAL);
  99. for (page = virt_to_page(start_addr); page <= virt_to_page(end_addr); page++)
  100. SetPageReserved(page);
  101. return 0;
  102. }
  103. static void sound_free_dmap(struct dma_buffparms *dmap)
  104. {
  105. int sz, size;
  106. struct page *page;
  107. unsigned long start_addr, end_addr;
  108. if (dmap->raw_buf == NULL)
  109. return;
  110. if (dmap->mapping_flags & DMA_MAP_MAPPED)
  111. return; /* Don't free mmapped buffer. Will use it next time */
  112. for (sz = 0, size = PAGE_SIZE; size < dmap->buffsize; sz++, size <<= 1);
  113. start_addr = (unsigned long) dmap->raw_buf;
  114. end_addr = start_addr + dmap->buffsize;
  115. for (page = virt_to_page(start_addr); page <= virt_to_page(end_addr); page++)
  116. ClearPageReserved(page);
  117. dma_unmap_single(NULL, dmap->raw_buf_phys, dmap->buffsize, DMA_BIDIRECTIONAL);
  118. free_pages((unsigned long) dmap->raw_buf, sz);
  119. dmap->raw_buf = NULL;
  120. }
  121. /* Intel version !!!!!!!!! */
  122. static int sound_start_dma(struct dma_buffparms *dmap, unsigned long physaddr, int count, int dma_mode)
  123. {
  124. unsigned long flags;
  125. int chan = dmap->dma;
  126. /* printk( "Start DMA%d %d, %d\n", chan, (int)(physaddr-dmap->raw_buf_phys), count); */
  127. flags = claim_dma_lock();
  128. disable_dma(chan);
  129. clear_dma_ff(chan);
  130. set_dma_mode(chan, dma_mode);
  131. set_dma_addr(chan, physaddr);
  132. set_dma_count(chan, count);
  133. enable_dma(chan);
  134. release_dma_lock(flags);
  135. return 0;
  136. }
  137. static void dma_init_buffers(struct dma_buffparms *dmap)
  138. {
  139. dmap->qlen = dmap->qhead = dmap->qtail = dmap->user_counter = 0;
  140. dmap->byte_counter = 0;
  141. dmap->max_byte_counter = 8000 * 60 * 60;
  142. dmap->bytes_in_use = dmap->buffsize;
  143. dmap->dma_mode = DMODE_NONE;
  144. dmap->mapping_flags = 0;
  145. dmap->neutral_byte = 0x80;
  146. dmap->data_rate = 8000;
  147. dmap->cfrag = -1;
  148. dmap->closing = 0;
  149. dmap->nbufs = 1;
  150. dmap->flags = DMA_BUSY; /* Other flags off */
  151. }
  152. static int open_dmap(struct audio_operations *adev, int mode, struct dma_buffparms *dmap)
  153. {
  154. int err;
  155. if (dmap->flags & DMA_BUSY)
  156. return -EBUSY;
  157. if ((err = sound_alloc_dmap(dmap)) < 0)
  158. return err;
  159. if (dmap->raw_buf == NULL) {
  160. printk(KERN_WARNING "Sound: DMA buffers not available\n");
  161. return -ENOSPC; /* Memory allocation failed during boot */
  162. }
  163. if (dmap->dma >= 0 && sound_open_dma(dmap->dma, adev->name)) {
  164. printk(KERN_WARNING "Unable to grab(2) DMA%d for the audio driver\n", dmap->dma);
  165. return -EBUSY;
  166. }
  167. dma_init_buffers(dmap);
  168. spin_lock_init(&dmap->lock);
  169. dmap->open_mode = mode;
  170. dmap->subdivision = dmap->underrun_count = 0;
  171. dmap->fragment_size = 0;
  172. dmap->max_fragments = 65536; /* Just a large value */
  173. dmap->byte_counter = 0;
  174. dmap->max_byte_counter = 8000 * 60 * 60;
  175. dmap->applic_profile = APF_NORMAL;
  176. dmap->needs_reorg = 1;
  177. dmap->audio_callback = NULL;
  178. dmap->callback_parm = 0;
  179. return 0;
  180. }
  181. static void close_dmap(struct audio_operations *adev, struct dma_buffparms *dmap)
  182. {
  183. unsigned long flags;
  184. if (dmap->dma >= 0) {
  185. sound_close_dma(dmap->dma);
  186. flags=claim_dma_lock();
  187. disable_dma(dmap->dma);
  188. release_dma_lock(flags);
  189. }
  190. if (dmap->flags & DMA_BUSY)
  191. dmap->dma_mode = DMODE_NONE;
  192. dmap->flags &= ~DMA_BUSY;
  193. if (sound_dmap_flag == DMAP_FREE_ON_CLOSE)
  194. sound_free_dmap(dmap);
  195. }
  196. static unsigned int default_set_bits(int dev, unsigned int bits)
  197. {
  198. mm_segment_t fs = get_fs();
  199. set_fs(get_ds());
  200. audio_devs[dev]->d->ioctl(dev, SNDCTL_DSP_SETFMT, (void __user *)&bits);
  201. set_fs(fs);
  202. return bits;
  203. }
  204. static int default_set_speed(int dev, int speed)
  205. {
  206. mm_segment_t fs = get_fs();
  207. set_fs(get_ds());
  208. audio_devs[dev]->d->ioctl(dev, SNDCTL_DSP_SPEED, (void __user *)&speed);
  209. set_fs(fs);
  210. return speed;
  211. }
  212. static short default_set_channels(int dev, short channels)
  213. {
  214. int c = channels;
  215. mm_segment_t fs = get_fs();
  216. set_fs(get_ds());
  217. audio_devs[dev]->d->ioctl(dev, SNDCTL_DSP_CHANNELS, (void __user *)&c);
  218. set_fs(fs);
  219. return c;
  220. }
  221. static void check_driver(struct audio_driver *d)
  222. {
  223. if (d->set_speed == NULL)
  224. d->set_speed = default_set_speed;
  225. if (d->set_bits == NULL)
  226. d->set_bits = default_set_bits;
  227. if (d->set_channels == NULL)
  228. d->set_channels = default_set_channels;
  229. }
  230. int DMAbuf_open(int dev, int mode)
  231. {
  232. struct audio_operations *adev = audio_devs[dev];
  233. int retval;
  234. struct dma_buffparms *dmap_in = NULL;
  235. struct dma_buffparms *dmap_out = NULL;
  236. if (!adev)
  237. return -ENXIO;
  238. if (!(adev->flags & DMA_DUPLEX))
  239. adev->dmap_in = adev->dmap_out;
  240. check_driver(adev->d);
  241. if ((retval = adev->d->open(dev, mode)) < 0)
  242. return retval;
  243. dmap_out = adev->dmap_out;
  244. dmap_in = adev->dmap_in;
  245. if (dmap_in == dmap_out)
  246. adev->flags &= ~DMA_DUPLEX;
  247. if (mode & OPEN_WRITE) {
  248. if ((retval = open_dmap(adev, mode, dmap_out)) < 0) {
  249. adev->d->close(dev);
  250. return retval;
  251. }
  252. }
  253. adev->enable_bits = mode;
  254. if (mode == OPEN_READ || (mode != OPEN_WRITE && (adev->flags & DMA_DUPLEX))) {
  255. if ((retval = open_dmap(adev, mode, dmap_in)) < 0) {
  256. adev->d->close(dev);
  257. if (mode & OPEN_WRITE)
  258. close_dmap(adev, dmap_out);
  259. return retval;
  260. }
  261. }
  262. adev->open_mode = mode;
  263. adev->go = 1;
  264. adev->d->set_bits(dev, 8);
  265. adev->d->set_channels(dev, 1);
  266. adev->d->set_speed(dev, DSP_DEFAULT_SPEED);
  267. if (adev->dmap_out->dma_mode == DMODE_OUTPUT)
  268. memset(adev->dmap_out->raw_buf, adev->dmap_out->neutral_byte,
  269. adev->dmap_out->bytes_in_use);
  270. return 0;
  271. }
  272. /* MUST not hold the spinlock */
  273. void DMAbuf_reset(int dev)
  274. {
  275. if (audio_devs[dev]->open_mode & OPEN_WRITE)
  276. dma_reset_output(dev);
  277. if (audio_devs[dev]->open_mode & OPEN_READ)
  278. dma_reset_input(dev);
  279. }
  280. static void dma_reset_output(int dev)
  281. {
  282. struct audio_operations *adev = audio_devs[dev];
  283. unsigned long flags,f ;
  284. struct dma_buffparms *dmap = adev->dmap_out;
  285. if (!(dmap->flags & DMA_STARTED)) /* DMA is not active */
  286. return;
  287. /*
  288. * First wait until the current fragment has been played completely
  289. */
  290. spin_lock_irqsave(&dmap->lock,flags);
  291. adev->dmap_out->flags |= DMA_SYNCING;
  292. adev->dmap_out->underrun_count = 0;
  293. if (!signal_pending(current) && adev->dmap_out->qlen &&
  294. adev->dmap_out->underrun_count == 0){
  295. spin_unlock_irqrestore(&dmap->lock,flags);
  296. interruptible_sleep_on_timeout(&adev->out_sleeper,
  297. dmabuf_timeout(dmap));
  298. spin_lock_irqsave(&dmap->lock,flags);
  299. }
  300. adev->dmap_out->flags &= ~(DMA_SYNCING | DMA_ACTIVE);
  301. /*
  302. * Finally shut the device off
  303. */
  304. if (!(adev->flags & DMA_DUPLEX) || !adev->d->halt_output)
  305. adev->d->halt_io(dev);
  306. else
  307. adev->d->halt_output(dev);
  308. adev->dmap_out->flags &= ~DMA_STARTED;
  309. f=claim_dma_lock();
  310. clear_dma_ff(dmap->dma);
  311. disable_dma(dmap->dma);
  312. release_dma_lock(f);
  313. dmap->byte_counter = 0;
  314. reorganize_buffers(dev, adev->dmap_out, 0);
  315. dmap->qlen = dmap->qhead = dmap->qtail = dmap->user_counter = 0;
  316. spin_unlock_irqrestore(&dmap->lock,flags);
  317. }
  318. static void dma_reset_input(int dev)
  319. {
  320. struct audio_operations *adev = audio_devs[dev];
  321. unsigned long flags;
  322. struct dma_buffparms *dmap = adev->dmap_in;
  323. spin_lock_irqsave(&dmap->lock,flags);
  324. if (!(adev->flags & DMA_DUPLEX) || !adev->d->halt_input)
  325. adev->d->halt_io(dev);
  326. else
  327. adev->d->halt_input(dev);
  328. adev->dmap_in->flags &= ~DMA_STARTED;
  329. dmap->qlen = dmap->qhead = dmap->qtail = dmap->user_counter = 0;
  330. dmap->byte_counter = 0;
  331. reorganize_buffers(dev, adev->dmap_in, 1);
  332. spin_unlock_irqrestore(&dmap->lock,flags);
  333. }
  334. /* MUST be called with holding the dmap->lock */
  335. void DMAbuf_launch_output(int dev, struct dma_buffparms *dmap)
  336. {
  337. struct audio_operations *adev = audio_devs[dev];
  338. if (!((adev->enable_bits * adev->go) & PCM_ENABLE_OUTPUT))
  339. return; /* Don't start DMA yet */
  340. dmap->dma_mode = DMODE_OUTPUT;
  341. if (!(dmap->flags & DMA_ACTIVE) || !(adev->flags & DMA_AUTOMODE) || (dmap->flags & DMA_NODMA)) {
  342. if (!(dmap->flags & DMA_STARTED)) {
  343. reorganize_buffers(dev, dmap, 0);
  344. if (adev->d->prepare_for_output(dev, dmap->fragment_size, dmap->nbufs))
  345. return;
  346. if (!(dmap->flags & DMA_NODMA))
  347. local_start_dma(adev, dmap->raw_buf_phys, dmap->bytes_in_use,DMA_MODE_WRITE);
  348. dmap->flags |= DMA_STARTED;
  349. }
  350. if (dmap->counts[dmap->qhead] == 0)
  351. dmap->counts[dmap->qhead] = dmap->fragment_size;
  352. dmap->dma_mode = DMODE_OUTPUT;
  353. adev->d->output_block(dev, dmap->raw_buf_phys + dmap->qhead * dmap->fragment_size,
  354. dmap->counts[dmap->qhead], 1);
  355. if (adev->d->trigger)
  356. adev->d->trigger(dev,adev->enable_bits * adev->go);
  357. }
  358. dmap->flags |= DMA_ACTIVE;
  359. }
  360. int DMAbuf_sync(int dev)
  361. {
  362. struct audio_operations *adev = audio_devs[dev];
  363. unsigned long flags;
  364. int n = 0;
  365. struct dma_buffparms *dmap;
  366. if (!adev->go && !(adev->enable_bits & PCM_ENABLE_OUTPUT))
  367. return 0;
  368. if (adev->dmap_out->dma_mode == DMODE_OUTPUT) {
  369. dmap = adev->dmap_out;
  370. spin_lock_irqsave(&dmap->lock,flags);
  371. if (dmap->qlen > 0 && !(dmap->flags & DMA_ACTIVE))
  372. DMAbuf_launch_output(dev, dmap);
  373. adev->dmap_out->flags |= DMA_SYNCING;
  374. adev->dmap_out->underrun_count = 0;
  375. while (!signal_pending(current) && n++ < adev->dmap_out->nbufs &&
  376. adev->dmap_out->qlen && adev->dmap_out->underrun_count == 0) {
  377. long t = dmabuf_timeout(dmap);
  378. spin_unlock_irqrestore(&dmap->lock,flags);
  379. /* FIXME: not safe may miss events */
  380. t = interruptible_sleep_on_timeout(&adev->out_sleeper, t);
  381. spin_lock_irqsave(&dmap->lock,flags);
  382. if (!t) {
  383. adev->dmap_out->flags &= ~DMA_SYNCING;
  384. spin_unlock_irqrestore(&dmap->lock,flags);
  385. return adev->dmap_out->qlen;
  386. }
  387. }
  388. adev->dmap_out->flags &= ~(DMA_SYNCING | DMA_ACTIVE);
  389. /*
  390. * Some devices such as GUS have huge amount of on board RAM for the
  391. * audio data. We have to wait until the device has finished playing.
  392. */
  393. /* still holding the lock */
  394. if (adev->d->local_qlen) { /* Device has hidden buffers */
  395. while (!signal_pending(current) &&
  396. adev->d->local_qlen(dev)){
  397. spin_unlock_irqrestore(&dmap->lock,flags);
  398. interruptible_sleep_on_timeout(&adev->out_sleeper,
  399. dmabuf_timeout(dmap));
  400. spin_lock_irqsave(&dmap->lock,flags);
  401. }
  402. }
  403. spin_unlock_irqrestore(&dmap->lock,flags);
  404. }
  405. adev->dmap_out->dma_mode = DMODE_NONE;
  406. return adev->dmap_out->qlen;
  407. }
  408. int DMAbuf_release(int dev, int mode)
  409. {
  410. struct audio_operations *adev = audio_devs[dev];
  411. struct dma_buffparms *dmap;
  412. unsigned long flags;
  413. dmap = adev->dmap_out;
  414. if (adev->open_mode & OPEN_WRITE)
  415. adev->dmap_out->closing = 1;
  416. if (adev->open_mode & OPEN_READ){
  417. adev->dmap_in->closing = 1;
  418. dmap = adev->dmap_in;
  419. }
  420. if (adev->open_mode & OPEN_WRITE)
  421. if (!(adev->dmap_out->mapping_flags & DMA_MAP_MAPPED))
  422. if (!signal_pending(current) && (adev->dmap_out->dma_mode == DMODE_OUTPUT))
  423. DMAbuf_sync(dev);
  424. if (adev->dmap_out->dma_mode == DMODE_OUTPUT)
  425. memset(adev->dmap_out->raw_buf, adev->dmap_out->neutral_byte, adev->dmap_out->bytes_in_use);
  426. DMAbuf_reset(dev);
  427. spin_lock_irqsave(&dmap->lock,flags);
  428. adev->d->close(dev);
  429. if (adev->open_mode & OPEN_WRITE)
  430. close_dmap(adev, adev->dmap_out);
  431. if (adev->open_mode == OPEN_READ ||
  432. (adev->open_mode != OPEN_WRITE &&
  433. (adev->flags & DMA_DUPLEX)))
  434. close_dmap(adev, adev->dmap_in);
  435. adev->open_mode = 0;
  436. spin_unlock_irqrestore(&dmap->lock,flags);
  437. return 0;
  438. }
  439. /* called with dmap->lock dold */
  440. int DMAbuf_activate_recording(int dev, struct dma_buffparms *dmap)
  441. {
  442. struct audio_operations *adev = audio_devs[dev];
  443. int err;
  444. if (!(adev->open_mode & OPEN_READ))
  445. return 0;
  446. if (!(adev->enable_bits & PCM_ENABLE_INPUT))
  447. return 0;
  448. if (dmap->dma_mode == DMODE_OUTPUT) { /* Direction change */
  449. /* release lock - it's not recursive */
  450. spin_unlock_irq(&dmap->lock);
  451. DMAbuf_sync(dev);
  452. DMAbuf_reset(dev);
  453. spin_lock_irq(&dmap->lock);
  454. dmap->dma_mode = DMODE_NONE;
  455. }
  456. if (!dmap->dma_mode) {
  457. reorganize_buffers(dev, dmap, 1);
  458. if ((err = adev->d->prepare_for_input(dev,
  459. dmap->fragment_size, dmap->nbufs)) < 0)
  460. return err;
  461. dmap->dma_mode = DMODE_INPUT;
  462. }
  463. if (!(dmap->flags & DMA_ACTIVE)) {
  464. if (dmap->needs_reorg)
  465. reorganize_buffers(dev, dmap, 0);
  466. local_start_dma(adev, dmap->raw_buf_phys, dmap->bytes_in_use, DMA_MODE_READ);
  467. adev->d->start_input(dev, dmap->raw_buf_phys + dmap->qtail * dmap->fragment_size,
  468. dmap->fragment_size, 0);
  469. dmap->flags |= DMA_ACTIVE;
  470. if (adev->d->trigger)
  471. adev->d->trigger(dev, adev->enable_bits * adev->go);
  472. }
  473. return 0;
  474. }
  475. /* acquires lock */
  476. int DMAbuf_getrdbuffer(int dev, char **buf, int *len, int dontblock)
  477. {
  478. struct audio_operations *adev = audio_devs[dev];
  479. unsigned long flags;
  480. int err = 0, n = 0;
  481. struct dma_buffparms *dmap = adev->dmap_in;
  482. int go;
  483. if (!(adev->open_mode & OPEN_READ))
  484. return -EIO;
  485. spin_lock_irqsave(&dmap->lock,flags);
  486. if (dmap->needs_reorg)
  487. reorganize_buffers(dev, dmap, 0);
  488. if (adev->dmap_in->mapping_flags & DMA_MAP_MAPPED) {
  489. /* printk(KERN_WARNING "Sound: Can't read from mmapped device (1)\n");*/
  490. spin_unlock_irqrestore(&dmap->lock,flags);
  491. return -EINVAL;
  492. } else while (dmap->qlen <= 0 && n++ < 10) {
  493. long timeout = MAX_SCHEDULE_TIMEOUT;
  494. if (!(adev->enable_bits & PCM_ENABLE_INPUT) || !adev->go) {
  495. spin_unlock_irqrestore(&dmap->lock,flags);
  496. return -EAGAIN;
  497. }
  498. if ((err = DMAbuf_activate_recording(dev, dmap)) < 0) {
  499. spin_unlock_irqrestore(&dmap->lock,flags);
  500. return err;
  501. }
  502. /* Wait for the next block */
  503. if (dontblock) {
  504. spin_unlock_irqrestore(&dmap->lock,flags);
  505. return -EAGAIN;
  506. }
  507. if ((go = adev->go))
  508. timeout = dmabuf_timeout(dmap);
  509. spin_unlock_irqrestore(&dmap->lock,flags);
  510. timeout = interruptible_sleep_on_timeout(&adev->in_sleeper,
  511. timeout);
  512. if (!timeout) {
  513. /* FIXME: include device name */
  514. err = -EIO;
  515. printk(KERN_WARNING "Sound: DMA (input) timed out - IRQ/DRQ config error?\n");
  516. dma_reset_input(dev);
  517. } else
  518. err = -EINTR;
  519. spin_lock_irqsave(&dmap->lock,flags);
  520. }
  521. spin_unlock_irqrestore(&dmap->lock,flags);
  522. if (dmap->qlen <= 0)
  523. return err ? err : -EINTR;
  524. *buf = &dmap->raw_buf[dmap->qhead * dmap->fragment_size + dmap->counts[dmap->qhead]];
  525. *len = dmap->fragment_size - dmap->counts[dmap->qhead];
  526. return dmap->qhead;
  527. }
  528. int DMAbuf_rmchars(int dev, int buff_no, int c)
  529. {
  530. struct audio_operations *adev = audio_devs[dev];
  531. struct dma_buffparms *dmap = adev->dmap_in;
  532. int p = dmap->counts[dmap->qhead] + c;
  533. if (dmap->mapping_flags & DMA_MAP_MAPPED)
  534. {
  535. /* printk("Sound: Can't read from mmapped device (2)\n");*/
  536. return -EINVAL;
  537. }
  538. else if (dmap->qlen <= 0)
  539. return -EIO;
  540. else if (p >= dmap->fragment_size) { /* This buffer is completely empty */
  541. dmap->counts[dmap->qhead] = 0;
  542. dmap->qlen--;
  543. dmap->qhead = (dmap->qhead + 1) % dmap->nbufs;
  544. }
  545. else dmap->counts[dmap->qhead] = p;
  546. return 0;
  547. }
  548. /* MUST be called with dmap->lock hold */
  549. int DMAbuf_get_buffer_pointer(int dev, struct dma_buffparms *dmap, int direction)
  550. {
  551. /*
  552. * Try to approximate the active byte position of the DMA pointer within the
  553. * buffer area as well as possible.
  554. */
  555. int pos;
  556. unsigned long f;
  557. if (!(dmap->flags & DMA_ACTIVE))
  558. pos = 0;
  559. else {
  560. int chan = dmap->dma;
  561. f=claim_dma_lock();
  562. clear_dma_ff(chan);
  563. if(!isa_dma_bridge_buggy)
  564. disable_dma(dmap->dma);
  565. pos = get_dma_residue(chan);
  566. pos = dmap->bytes_in_use - pos;
  567. if (!(dmap->mapping_flags & DMA_MAP_MAPPED)) {
  568. if (direction == DMODE_OUTPUT) {
  569. if (dmap->qhead == 0)
  570. if (pos > dmap->fragment_size)
  571. pos = 0;
  572. } else {
  573. if (dmap->qtail == 0)
  574. if (pos > dmap->fragment_size)
  575. pos = 0;
  576. }
  577. }
  578. if (pos < 0)
  579. pos = 0;
  580. if (pos >= dmap->bytes_in_use)
  581. pos = 0;
  582. if(!isa_dma_bridge_buggy)
  583. enable_dma(dmap->dma);
  584. release_dma_lock(f);
  585. }
  586. /* printk( "%04x ", pos); */
  587. return pos;
  588. }
  589. /*
  590. * DMAbuf_start_devices() is called by the /dev/music driver to start
  591. * one or more audio devices at desired moment.
  592. */
  593. void DMAbuf_start_devices(unsigned int devmask)
  594. {
  595. struct audio_operations *adev;
  596. int dev;
  597. for (dev = 0; dev < num_audiodevs; dev++) {
  598. if (!(devmask & (1 << dev)))
  599. continue;
  600. if (!(adev = audio_devs[dev]))
  601. continue;
  602. if (adev->open_mode == 0)
  603. continue;
  604. if (adev->go)
  605. continue;
  606. /* OK to start the device */
  607. adev->go = 1;
  608. if (adev->d->trigger)
  609. adev->d->trigger(dev,adev->enable_bits * adev->go);
  610. }
  611. }
  612. /* via poll called without a lock ?*/
  613. int DMAbuf_space_in_queue(int dev)
  614. {
  615. struct audio_operations *adev = audio_devs[dev];
  616. int len, max, tmp;
  617. struct dma_buffparms *dmap = adev->dmap_out;
  618. int lim = dmap->nbufs;
  619. if (lim < 2)
  620. lim = 2;
  621. if (dmap->qlen >= lim) /* No space at all */
  622. return 0;
  623. /*
  624. * Verify that there are no more pending buffers than the limit
  625. * defined by the process.
  626. */
  627. max = dmap->max_fragments;
  628. if (max > lim)
  629. max = lim;
  630. len = dmap->qlen;
  631. if (adev->d->local_qlen) {
  632. tmp = adev->d->local_qlen(dev);
  633. if (tmp && len)
  634. tmp--; /* This buffer has been counted twice */
  635. len += tmp;
  636. }
  637. if (dmap->byte_counter % dmap->fragment_size) /* There is a partial fragment */
  638. len = len + 1;
  639. if (len >= max)
  640. return 0;
  641. return max - len;
  642. }
  643. /* MUST not hold the spinlock - this function may sleep */
  644. static int output_sleep(int dev, int dontblock)
  645. {
  646. struct audio_operations *adev = audio_devs[dev];
  647. int err = 0;
  648. struct dma_buffparms *dmap = adev->dmap_out;
  649. long timeout;
  650. long timeout_value;
  651. if (dontblock)
  652. return -EAGAIN;
  653. if (!(adev->enable_bits & PCM_ENABLE_OUTPUT))
  654. return -EAGAIN;
  655. /*
  656. * Wait for free space
  657. */
  658. if (signal_pending(current))
  659. return -EINTR;
  660. timeout = (adev->go && !(dmap->flags & DMA_NOTIMEOUT));
  661. if (timeout)
  662. timeout_value = dmabuf_timeout(dmap);
  663. else
  664. timeout_value = MAX_SCHEDULE_TIMEOUT;
  665. timeout_value = interruptible_sleep_on_timeout(&adev->out_sleeper,
  666. timeout_value);
  667. if (timeout != MAX_SCHEDULE_TIMEOUT && !timeout_value) {
  668. printk(KERN_WARNING "Sound: DMA (output) timed out - IRQ/DRQ config error?\n");
  669. dma_reset_output(dev);
  670. } else {
  671. if (signal_pending(current))
  672. err = -EINTR;
  673. }
  674. return err;
  675. }
  676. /* called with the lock held */
  677. static int find_output_space(int dev, char **buf, int *size)
  678. {
  679. struct audio_operations *adev = audio_devs[dev];
  680. struct dma_buffparms *dmap = adev->dmap_out;
  681. unsigned long active_offs;
  682. long len, offs;
  683. int maxfrags;
  684. int occupied_bytes = (dmap->user_counter % dmap->fragment_size);
  685. *buf = dmap->raw_buf;
  686. if (!(maxfrags = DMAbuf_space_in_queue(dev)) && !occupied_bytes)
  687. return 0;
  688. #ifdef BE_CONSERVATIVE
  689. active_offs = dmap->byte_counter + dmap->qhead * dmap->fragment_size;
  690. #else
  691. active_offs = max(DMAbuf_get_buffer_pointer(dev, dmap, DMODE_OUTPUT), 0);
  692. /* Check for pointer wrapping situation */
  693. if (active_offs >= dmap->bytes_in_use)
  694. active_offs = 0;
  695. active_offs += dmap->byte_counter;
  696. #endif
  697. offs = (dmap->user_counter % dmap->bytes_in_use) & ~SAMPLE_ROUNDUP;
  698. if (offs < 0 || offs >= dmap->bytes_in_use) {
  699. printk(KERN_ERR "Sound: Got unexpected offs %ld. Giving up.\n", offs);
  700. printk("Counter = %ld, bytes=%d\n", dmap->user_counter, dmap->bytes_in_use);
  701. return 0;
  702. }
  703. *buf = dmap->raw_buf + offs;
  704. len = active_offs + dmap->bytes_in_use - dmap->user_counter; /* Number of unused bytes in buffer */
  705. if ((offs + len) > dmap->bytes_in_use)
  706. len = dmap->bytes_in_use - offs;
  707. if (len < 0) {
  708. return 0;
  709. }
  710. if (len > ((maxfrags * dmap->fragment_size) - occupied_bytes))
  711. len = (maxfrags * dmap->fragment_size) - occupied_bytes;
  712. *size = len & ~SAMPLE_ROUNDUP;
  713. return (*size > 0);
  714. }
  715. /* acquires lock */
  716. int DMAbuf_getwrbuffer(int dev, char **buf, int *size, int dontblock)
  717. {
  718. struct audio_operations *adev = audio_devs[dev];
  719. unsigned long flags;
  720. int err = -EIO;
  721. struct dma_buffparms *dmap = adev->dmap_out;
  722. if (dmap->mapping_flags & DMA_MAP_MAPPED) {
  723. /* printk(KERN_DEBUG "Sound: Can't write to mmapped device (3)\n");*/
  724. return -EINVAL;
  725. }
  726. spin_lock_irqsave(&dmap->lock,flags);
  727. if (dmap->needs_reorg)
  728. reorganize_buffers(dev, dmap, 0);
  729. if (dmap->dma_mode == DMODE_INPUT) { /* Direction change */
  730. spin_unlock_irqrestore(&dmap->lock,flags);
  731. DMAbuf_reset(dev);
  732. spin_lock_irqsave(&dmap->lock,flags);
  733. }
  734. dmap->dma_mode = DMODE_OUTPUT;
  735. while (find_output_space(dev, buf, size) <= 0) {
  736. spin_unlock_irqrestore(&dmap->lock,flags);
  737. if ((err = output_sleep(dev, dontblock)) < 0) {
  738. return err;
  739. }
  740. spin_lock_irqsave(&dmap->lock,flags);
  741. }
  742. spin_unlock_irqrestore(&dmap->lock,flags);
  743. return 0;
  744. }
  745. /* has to acquire dmap->lock */
  746. int DMAbuf_move_wrpointer(int dev, int l)
  747. {
  748. struct audio_operations *adev = audio_devs[dev];
  749. struct dma_buffparms *dmap = adev->dmap_out;
  750. unsigned long ptr;
  751. unsigned long end_ptr, p;
  752. int post;
  753. unsigned long flags;
  754. spin_lock_irqsave(&dmap->lock,flags);
  755. post= (dmap->flags & DMA_POST);
  756. ptr = (dmap->user_counter / dmap->fragment_size) * dmap->fragment_size;
  757. dmap->flags &= ~DMA_POST;
  758. dmap->cfrag = -1;
  759. dmap->user_counter += l;
  760. dmap->flags |= DMA_DIRTY;
  761. if (dmap->byte_counter >= dmap->max_byte_counter) {
  762. /* Wrap the byte counters */
  763. long decr = dmap->byte_counter;
  764. dmap->byte_counter = (dmap->byte_counter % dmap->bytes_in_use);
  765. decr -= dmap->byte_counter;
  766. dmap->user_counter -= decr;
  767. }
  768. end_ptr = (dmap->user_counter / dmap->fragment_size) * dmap->fragment_size;
  769. p = (dmap->user_counter - 1) % dmap->bytes_in_use;
  770. dmap->neutral_byte = dmap->raw_buf[p];
  771. /* Update the fragment based bookkeeping too */
  772. while (ptr < end_ptr) {
  773. dmap->counts[dmap->qtail] = dmap->fragment_size;
  774. dmap->qtail = (dmap->qtail + 1) % dmap->nbufs;
  775. dmap->qlen++;
  776. ptr += dmap->fragment_size;
  777. }
  778. dmap->counts[dmap->qtail] = dmap->user_counter - ptr;
  779. /*
  780. * Let the low level driver perform some postprocessing to
  781. * the written data.
  782. */
  783. if (adev->d->postprocess_write)
  784. adev->d->postprocess_write(dev);
  785. if (!(dmap->flags & DMA_ACTIVE))
  786. if (dmap->qlen > 1 || (dmap->qlen > 0 && (post || dmap->qlen >= dmap->nbufs - 1)))
  787. DMAbuf_launch_output(dev, dmap);
  788. spin_unlock_irqrestore(&dmap->lock,flags);
  789. return 0;
  790. }
  791. int DMAbuf_start_dma(int dev, unsigned long physaddr, int count, int dma_mode)
  792. {
  793. struct audio_operations *adev = audio_devs[dev];
  794. struct dma_buffparms *dmap = (dma_mode == DMA_MODE_WRITE) ? adev->dmap_out : adev->dmap_in;
  795. if (dmap->raw_buf == NULL) {
  796. printk(KERN_ERR "sound: DMA buffer(1) == NULL\n");
  797. printk("Device %d, chn=%s\n", dev, (dmap == adev->dmap_out) ? "out" : "in");
  798. return 0;
  799. }
  800. if (dmap->dma < 0)
  801. return 0;
  802. sound_start_dma(dmap, physaddr, count, dma_mode);
  803. return count;
  804. }
  805. EXPORT_SYMBOL(DMAbuf_start_dma);
  806. static int local_start_dma(struct audio_operations *adev, unsigned long physaddr, int count, int dma_mode)
  807. {
  808. struct dma_buffparms *dmap = (dma_mode == DMA_MODE_WRITE) ? adev->dmap_out : adev->dmap_in;
  809. if (dmap->raw_buf == NULL) {
  810. printk(KERN_ERR "sound: DMA buffer(2) == NULL\n");
  811. printk(KERN_ERR "Device %s, chn=%s\n", adev->name, (dmap == adev->dmap_out) ? "out" : "in");
  812. return 0;
  813. }
  814. if (dmap->flags & DMA_NODMA)
  815. return 1;
  816. if (dmap->dma < 0)
  817. return 0;
  818. sound_start_dma(dmap, dmap->raw_buf_phys, dmap->bytes_in_use, dma_mode | DMA_AUTOINIT);
  819. dmap->flags |= DMA_STARTED;
  820. return count;
  821. }
  822. static void finish_output_interrupt(int dev, struct dma_buffparms *dmap)
  823. {
  824. struct audio_operations *adev = audio_devs[dev];
  825. if (dmap->audio_callback != NULL)
  826. dmap->audio_callback(dev, dmap->callback_parm);
  827. wake_up(&adev->out_sleeper);
  828. wake_up(&adev->poll_sleeper);
  829. }
  830. /* called with dmap->lock held in irq context*/
  831. static void do_outputintr(int dev, int dummy)
  832. {
  833. struct audio_operations *adev = audio_devs[dev];
  834. struct dma_buffparms *dmap = adev->dmap_out;
  835. int this_fragment;
  836. if (dmap->raw_buf == NULL) {
  837. printk(KERN_ERR "Sound: Error. Audio interrupt (%d) after freeing buffers.\n", dev);
  838. return;
  839. }
  840. if (dmap->mapping_flags & DMA_MAP_MAPPED) { /* Virtual memory mapped access */
  841. /* mmapped access */
  842. dmap->qhead = (dmap->qhead + 1) % dmap->nbufs;
  843. if (dmap->qhead == 0) { /* Wrapped */
  844. dmap->byte_counter += dmap->bytes_in_use;
  845. if (dmap->byte_counter >= dmap->max_byte_counter) { /* Overflow */
  846. long decr = dmap->byte_counter;
  847. dmap->byte_counter = (dmap->byte_counter % dmap->bytes_in_use);
  848. decr -= dmap->byte_counter;
  849. dmap->user_counter -= decr;
  850. }
  851. }
  852. dmap->qlen++; /* Yes increment it (don't decrement) */
  853. if (!(adev->flags & DMA_AUTOMODE))
  854. dmap->flags &= ~DMA_ACTIVE;
  855. dmap->counts[dmap->qhead] = dmap->fragment_size;
  856. DMAbuf_launch_output(dev, dmap);
  857. finish_output_interrupt(dev, dmap);
  858. return;
  859. }
  860. dmap->qlen--;
  861. this_fragment = dmap->qhead;
  862. dmap->qhead = (dmap->qhead + 1) % dmap->nbufs;
  863. if (dmap->qhead == 0) { /* Wrapped */
  864. dmap->byte_counter += dmap->bytes_in_use;
  865. if (dmap->byte_counter >= dmap->max_byte_counter) { /* Overflow */
  866. long decr = dmap->byte_counter;
  867. dmap->byte_counter = (dmap->byte_counter % dmap->bytes_in_use);
  868. decr -= dmap->byte_counter;
  869. dmap->user_counter -= decr;
  870. }
  871. }
  872. if (!(adev->flags & DMA_AUTOMODE))
  873. dmap->flags &= ~DMA_ACTIVE;
  874. /*
  875. * This is dmap->qlen <= 0 except when closing when
  876. * dmap->qlen < 0
  877. */
  878. while (dmap->qlen <= -dmap->closing) {
  879. dmap->underrun_count++;
  880. dmap->qlen++;
  881. if ((dmap->flags & DMA_DIRTY) && dmap->applic_profile != APF_CPUINTENS) {
  882. dmap->flags &= ~DMA_DIRTY;
  883. memset(adev->dmap_out->raw_buf, adev->dmap_out->neutral_byte,
  884. adev->dmap_out->buffsize);
  885. }
  886. dmap->user_counter += dmap->fragment_size;
  887. dmap->qtail = (dmap->qtail + 1) % dmap->nbufs;
  888. }
  889. if (dmap->qlen > 0)
  890. DMAbuf_launch_output(dev, dmap);
  891. finish_output_interrupt(dev, dmap);
  892. }
  893. /* called in irq context */
  894. void DMAbuf_outputintr(int dev, int notify_only)
  895. {
  896. struct audio_operations *adev = audio_devs[dev];
  897. unsigned long flags;
  898. struct dma_buffparms *dmap = adev->dmap_out;
  899. spin_lock_irqsave(&dmap->lock,flags);
  900. if (!(dmap->flags & DMA_NODMA)) {
  901. int chan = dmap->dma, pos, n;
  902. unsigned long f;
  903. f=claim_dma_lock();
  904. if(!isa_dma_bridge_buggy)
  905. disable_dma(dmap->dma);
  906. clear_dma_ff(chan);
  907. pos = dmap->bytes_in_use - get_dma_residue(chan);
  908. if(!isa_dma_bridge_buggy)
  909. enable_dma(dmap->dma);
  910. release_dma_lock(f);
  911. pos = pos / dmap->fragment_size; /* Actual qhead */
  912. if (pos < 0 || pos >= dmap->nbufs)
  913. pos = 0;
  914. n = 0;
  915. while (dmap->qhead != pos && n++ < dmap->nbufs)
  916. do_outputintr(dev, notify_only);
  917. }
  918. else
  919. do_outputintr(dev, notify_only);
  920. spin_unlock_irqrestore(&dmap->lock,flags);
  921. }
  922. EXPORT_SYMBOL(DMAbuf_outputintr);
  923. /* called with dmap->lock held in irq context */
  924. static void do_inputintr(int dev)
  925. {
  926. struct audio_operations *adev = audio_devs[dev];
  927. struct dma_buffparms *dmap = adev->dmap_in;
  928. if (dmap->raw_buf == NULL) {
  929. printk(KERN_ERR "Sound: Fatal error. Audio interrupt after freeing buffers.\n");
  930. return;
  931. }
  932. if (dmap->mapping_flags & DMA_MAP_MAPPED) {
  933. dmap->qtail = (dmap->qtail + 1) % dmap->nbufs;
  934. if (dmap->qtail == 0) { /* Wrapped */
  935. dmap->byte_counter += dmap->bytes_in_use;
  936. if (dmap->byte_counter >= dmap->max_byte_counter) { /* Overflow */
  937. long decr = dmap->byte_counter;
  938. dmap->byte_counter = (dmap->byte_counter % dmap->bytes_in_use) + dmap->bytes_in_use;
  939. decr -= dmap->byte_counter;
  940. dmap->user_counter -= decr;
  941. }
  942. }
  943. dmap->qlen++;
  944. if (!(adev->flags & DMA_AUTOMODE)) {
  945. if (dmap->needs_reorg)
  946. reorganize_buffers(dev, dmap, 0);
  947. local_start_dma(adev, dmap->raw_buf_phys, dmap->bytes_in_use,DMA_MODE_READ);
  948. adev->d->start_input(dev, dmap->raw_buf_phys + dmap->qtail * dmap->fragment_size,
  949. dmap->fragment_size, 1);
  950. if (adev->d->trigger)
  951. adev->d->trigger(dev, adev->enable_bits * adev->go);
  952. }
  953. dmap->flags |= DMA_ACTIVE;
  954. } else if (dmap->qlen >= (dmap->nbufs - 1)) {
  955. printk(KERN_WARNING "Sound: Recording overrun\n");
  956. dmap->underrun_count++;
  957. /* Just throw away the oldest fragment but keep the engine running */
  958. dmap->qhead = (dmap->qhead + 1) % dmap->nbufs;
  959. dmap->qtail = (dmap->qtail + 1) % dmap->nbufs;
  960. } else if (dmap->qlen >= 0 && dmap->qlen < dmap->nbufs) {
  961. dmap->qlen++;
  962. dmap->qtail = (dmap->qtail + 1) % dmap->nbufs;
  963. if (dmap->qtail == 0) { /* Wrapped */
  964. dmap->byte_counter += dmap->bytes_in_use;
  965. if (dmap->byte_counter >= dmap->max_byte_counter) { /* Overflow */
  966. long decr = dmap->byte_counter;
  967. dmap->byte_counter = (dmap->byte_counter % dmap->bytes_in_use) + dmap->bytes_in_use;
  968. decr -= dmap->byte_counter;
  969. dmap->user_counter -= decr;
  970. }
  971. }
  972. }
  973. if (!(adev->flags & DMA_AUTOMODE) || (dmap->flags & DMA_NODMA)) {
  974. local_start_dma(adev, dmap->raw_buf_phys, dmap->bytes_in_use, DMA_MODE_READ);
  975. adev->d->start_input(dev, dmap->raw_buf_phys + dmap->qtail * dmap->fragment_size, dmap->fragment_size, 1);
  976. if (adev->d->trigger)
  977. adev->d->trigger(dev,adev->enable_bits * adev->go);
  978. }
  979. dmap->flags |= DMA_ACTIVE;
  980. if (dmap->qlen > 0)
  981. {
  982. wake_up(&adev->in_sleeper);
  983. wake_up(&adev->poll_sleeper);
  984. }
  985. }
  986. /* called in irq context */
  987. void DMAbuf_inputintr(int dev)
  988. {
  989. struct audio_operations *adev = audio_devs[dev];
  990. struct dma_buffparms *dmap = adev->dmap_in;
  991. unsigned long flags;
  992. spin_lock_irqsave(&dmap->lock,flags);
  993. if (!(dmap->flags & DMA_NODMA)) {
  994. int chan = dmap->dma, pos, n;
  995. unsigned long f;
  996. f=claim_dma_lock();
  997. if(!isa_dma_bridge_buggy)
  998. disable_dma(dmap->dma);
  999. clear_dma_ff(chan);
  1000. pos = dmap->bytes_in_use - get_dma_residue(chan);
  1001. if(!isa_dma_bridge_buggy)
  1002. enable_dma(dmap->dma);
  1003. release_dma_lock(f);
  1004. pos = pos / dmap->fragment_size; /* Actual qhead */
  1005. if (pos < 0 || pos >= dmap->nbufs)
  1006. pos = 0;
  1007. n = 0;
  1008. while (dmap->qtail != pos && ++n < dmap->nbufs)
  1009. do_inputintr(dev);
  1010. } else
  1011. do_inputintr(dev);
  1012. spin_unlock_irqrestore(&dmap->lock,flags);
  1013. }
  1014. EXPORT_SYMBOL(DMAbuf_inputintr);
  1015. void DMAbuf_init(int dev, int dma1, int dma2)
  1016. {
  1017. struct audio_operations *adev = audio_devs[dev];
  1018. /*
  1019. * NOTE! This routine could be called several times.
  1020. */
  1021. if (adev && adev->dmap_out == NULL) {
  1022. if (adev->d == NULL)
  1023. panic("OSS: audio_devs[%d]->d == NULL\n", dev);
  1024. if (adev->parent_dev) { /* Use DMA map of the parent dev */
  1025. int parent = adev->parent_dev - 1;
  1026. adev->dmap_out = audio_devs[parent]->dmap_out;
  1027. adev->dmap_in = audio_devs[parent]->dmap_in;
  1028. } else {
  1029. adev->dmap_out = adev->dmap_in = &adev->dmaps[0];
  1030. adev->dmap_out->dma = dma1;
  1031. if (adev->flags & DMA_DUPLEX) {
  1032. adev->dmap_in = &adev->dmaps[1];
  1033. adev->dmap_in->dma = dma2;
  1034. }
  1035. }
  1036. /* Persistent DMA buffers allocated here */
  1037. if (sound_dmap_flag == DMAP_KEEP_ON_CLOSE) {
  1038. if (adev->dmap_in->raw_buf == NULL)
  1039. sound_alloc_dmap(adev->dmap_in);
  1040. if (adev->dmap_out->raw_buf == NULL)
  1041. sound_alloc_dmap(adev->dmap_out);
  1042. }
  1043. }
  1044. }
  1045. /* No kernel lock - DMAbuf_activate_recording protected by global cli/sti */
  1046. static unsigned int poll_input(struct file * file, int dev, poll_table *wait)
  1047. {
  1048. struct audio_operations *adev = audio_devs[dev];
  1049. struct dma_buffparms *dmap = adev->dmap_in;
  1050. if (!(adev->open_mode & OPEN_READ))
  1051. return 0;
  1052. if (dmap->mapping_flags & DMA_MAP_MAPPED) {
  1053. if (dmap->qlen)
  1054. return POLLIN | POLLRDNORM;
  1055. return 0;
  1056. }
  1057. if (dmap->dma_mode != DMODE_INPUT) {
  1058. if (dmap->dma_mode == DMODE_NONE &&
  1059. adev->enable_bits & PCM_ENABLE_INPUT &&
  1060. !dmap->qlen && adev->go) {
  1061. unsigned long flags;
  1062. spin_lock_irqsave(&dmap->lock,flags);
  1063. DMAbuf_activate_recording(dev, dmap);
  1064. spin_unlock_irqrestore(&dmap->lock,flags);
  1065. }
  1066. return 0;
  1067. }
  1068. if (!dmap->qlen)
  1069. return 0;
  1070. return POLLIN | POLLRDNORM;
  1071. }
  1072. static unsigned int poll_output(struct file * file, int dev, poll_table *wait)
  1073. {
  1074. struct audio_operations *adev = audio_devs[dev];
  1075. struct dma_buffparms *dmap = adev->dmap_out;
  1076. if (!(adev->open_mode & OPEN_WRITE))
  1077. return 0;
  1078. if (dmap->mapping_flags & DMA_MAP_MAPPED) {
  1079. if (dmap->qlen)
  1080. return POLLOUT | POLLWRNORM;
  1081. return 0;
  1082. }
  1083. if (dmap->dma_mode == DMODE_INPUT)
  1084. return 0;
  1085. if (dmap->dma_mode == DMODE_NONE)
  1086. return POLLOUT | POLLWRNORM;
  1087. if (!DMAbuf_space_in_queue(dev))
  1088. return 0;
  1089. return POLLOUT | POLLWRNORM;
  1090. }
  1091. unsigned int DMAbuf_poll(struct file * file, int dev, poll_table *wait)
  1092. {
  1093. struct audio_operations *adev = audio_devs[dev];
  1094. poll_wait(file, &adev->poll_sleeper, wait);
  1095. return poll_input(file, dev, wait) | poll_output(file, dev, wait);
  1096. }
  1097. void DMAbuf_deinit(int dev)
  1098. {
  1099. struct audio_operations *adev = audio_devs[dev];
  1100. /* This routine is called when driver is being unloaded */
  1101. if (!adev)
  1102. return;
  1103. /* Persistent DMA buffers deallocated here */
  1104. if (sound_dmap_flag == DMAP_KEEP_ON_CLOSE) {
  1105. sound_free_dmap(adev->dmap_out);
  1106. if (adev->flags & DMA_DUPLEX)
  1107. sound_free_dmap(adev->dmap_in);
  1108. }
  1109. }