|
@@ -1406,7 +1406,7 @@ static void tcp_service_net_dma(struct sock *sk, bool wait)
|
|
|
return;
|
|
|
|
|
|
last_issued = tp->ucopy.dma_cookie;
|
|
|
- dma_async_memcpy_issue_pending(tp->ucopy.dma_chan);
|
|
|
+ dma_async_issue_pending(tp->ucopy.dma_chan);
|
|
|
|
|
|
do {
|
|
|
if (dma_async_memcpy_complete(tp->ucopy.dma_chan,
|
|
@@ -1744,7 +1744,7 @@ int tcp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
|
|
|
tcp_service_net_dma(sk, true);
|
|
|
tcp_cleanup_rbuf(sk, copied);
|
|
|
} else
|
|
|
- dma_async_memcpy_issue_pending(tp->ucopy.dma_chan);
|
|
|
+ dma_async_issue_pending(tp->ucopy.dma_chan);
|
|
|
}
|
|
|
#endif
|
|
|
if (copied >= target) {
|
|
@@ -1837,7 +1837,7 @@ do_prequeue:
|
|
|
break;
|
|
|
}
|
|
|
|
|
|
- dma_async_memcpy_issue_pending(tp->ucopy.dma_chan);
|
|
|
+ dma_async_issue_pending(tp->ucopy.dma_chan);
|
|
|
|
|
|
if ((offset + used) == skb->len)
|
|
|
copied_early = true;
|