|
@@ -283,10 +283,14 @@ static int ep93xx_rx(struct net_device *dev, int processed, int budget)
|
|
|
|
|
|
skb = dev_alloc_skb(length + 2);
|
|
|
if (likely(skb != NULL)) {
|
|
|
+ struct ep93xx_rdesc *rxd = &ep->descs->rdesc[entry];
|
|
|
skb_reserve(skb, 2);
|
|
|
- dma_sync_single_for_cpu(dev->dev.parent, ep->descs->rdesc[entry].buf_addr,
|
|
|
+ dma_sync_single_for_cpu(dev->dev.parent, rxd->buf_addr,
|
|
|
length, DMA_FROM_DEVICE);
|
|
|
skb_copy_to_linear_data(skb, ep->rx_buf[entry], length);
|
|
|
+ dma_sync_single_for_device(dev->dev.parent,
|
|
|
+ rxd->buf_addr, length,
|
|
|
+ DMA_FROM_DEVICE);
|
|
|
skb_put(skb, length);
|
|
|
skb->protocol = eth_type_trans(skb, dev);
|
|
|
|
|
@@ -348,6 +352,7 @@ poll_some_more:
|
|
|
static int ep93xx_xmit(struct sk_buff *skb, struct net_device *dev)
|
|
|
{
|
|
|
struct ep93xx_priv *ep = netdev_priv(dev);
|
|
|
+ struct ep93xx_tdesc *txd;
|
|
|
int entry;
|
|
|
|
|
|
if (unlikely(skb->len > MAX_PKT_SIZE)) {
|
|
@@ -359,11 +364,14 @@ static int ep93xx_xmit(struct sk_buff *skb, struct net_device *dev)
|
|
|
entry = ep->tx_pointer;
|
|
|
ep->tx_pointer = (ep->tx_pointer + 1) & (TX_QUEUE_ENTRIES - 1);
|
|
|
|
|
|
- ep->descs->tdesc[entry].tdesc1 =
|
|
|
- TDESC1_EOF | (entry << 16) | (skb->len & 0xfff);
|
|
|
+ txd = &ep->descs->tdesc[entry];
|
|
|
+
|
|
|
+ txd->tdesc1 = TDESC1_EOF | (entry << 16) | (skb->len & 0xfff);
|
|
|
+ dma_sync_single_for_cpu(dev->dev.parent, txd->buf_addr, skb->len,
|
|
|
+ DMA_TO_DEVICE);
|
|
|
skb_copy_and_csum_dev(skb, ep->tx_buf[entry]);
|
|
|
- dma_sync_single_for_cpu(dev->dev.parent, ep->descs->tdesc[entry].buf_addr,
|
|
|
- skb->len, DMA_TO_DEVICE);
|
|
|
+ dma_sync_single_for_device(dev->dev.parent, txd->buf_addr, skb->len,
|
|
|
+ DMA_TO_DEVICE);
|
|
|
dev_kfree_skb(skb);
|
|
|
|
|
|
spin_lock_irq(&ep->tx_pending_lock);
|