|
@@ -128,8 +128,8 @@ async_tx_channel_switch(struct dma_async_tx_descriptor *depend_tx,
|
|
|
}
|
|
|
device->device_issue_pending(chan);
|
|
|
} else {
|
|
|
- if (dma_wait_for_async_tx(depend_tx) == DMA_ERROR)
|
|
|
- panic("%s: DMA_ERROR waiting for depend_tx\n",
|
|
|
+ if (dma_wait_for_async_tx(depend_tx) != DMA_SUCCESS)
|
|
|
+ panic("%s: DMA error waiting for depend_tx\n",
|
|
|
__func__);
|
|
|
tx->tx_submit(tx);
|
|
|
}
|
|
@@ -280,8 +280,9 @@ void async_tx_quiesce(struct dma_async_tx_descriptor **tx)
|
|
|
* we are referring to the correct operation
|
|
|
*/
|
|
|
BUG_ON(async_tx_test_ack(*tx));
|
|
|
- if (dma_wait_for_async_tx(*tx) == DMA_ERROR)
|
|
|
- panic("DMA_ERROR waiting for transaction\n");
|
|
|
+ if (dma_wait_for_async_tx(*tx) != DMA_SUCCESS)
|
|
|
+ panic("%s: DMA error waiting for transaction\n",
|
|
|
+ __func__);
|
|
|
async_tx_ack(*tx);
|
|
|
*tx = NULL;
|
|
|
}
|