|
@@ -198,8 +198,8 @@ static inline void ath5k_txbuf_free_skb(struct ath5k_softc *sc,
|
|
|
BUG_ON(!bf);
|
|
|
if (!bf->skb)
|
|
|
return;
|
|
|
- pci_unmap_single(sc->pdev, bf->skbaddr, bf->skb->len,
|
|
|
- PCI_DMA_TODEVICE);
|
|
|
+ dma_unmap_single(sc->dev, bf->skbaddr, bf->skb->len,
|
|
|
+ DMA_TO_DEVICE);
|
|
|
dev_kfree_skb_any(bf->skb);
|
|
|
bf->skb = NULL;
|
|
|
bf->skbaddr = 0;
|
|
@@ -215,8 +215,8 @@ static inline void ath5k_rxbuf_free_skb(struct ath5k_softc *sc,
|
|
|
BUG_ON(!bf);
|
|
|
if (!bf->skb)
|
|
|
return;
|
|
|
- pci_unmap_single(sc->pdev, bf->skbaddr, common->rx_bufsize,
|
|
|
- PCI_DMA_FROMDEVICE);
|
|
|
+ dma_unmap_single(sc->dev, bf->skbaddr, common->rx_bufsize,
|
|
|
+ DMA_FROM_DEVICE);
|
|
|
dev_kfree_skb_any(bf->skb);
|
|
|
bf->skb = NULL;
|
|
|
bf->skbaddr = 0;
|
|
@@ -647,10 +647,11 @@ struct sk_buff *ath5k_rx_skb_alloc(struct ath5k_softc *sc, dma_addr_t *skb_addr)
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
|
- *skb_addr = pci_map_single(sc->pdev,
|
|
|
+ *skb_addr = dma_map_single(sc->dev,
|
|
|
skb->data, common->rx_bufsize,
|
|
|
- PCI_DMA_FROMDEVICE);
|
|
|
- if (unlikely(pci_dma_mapping_error(sc->pdev, *skb_addr))) {
|
|
|
+ DMA_FROM_DEVICE);
|
|
|
+
|
|
|
+ if (unlikely(dma_mapping_error(sc->dev, *skb_addr))) {
|
|
|
ATH5K_ERR(sc, "%s: DMA mapping failed\n", __func__);
|
|
|
dev_kfree_skb(skb);
|
|
|
return NULL;
|
|
@@ -746,8 +747,8 @@ ath5k_txbuf_setup(struct ath5k_softc *sc, struct ath5k_buf *bf,
|
|
|
flags = AR5K_TXDESC_INTREQ | AR5K_TXDESC_CLRDMASK;
|
|
|
|
|
|
/* XXX endianness */
|
|
|
- bf->skbaddr = pci_map_single(sc->pdev, skb->data, skb->len,
|
|
|
- PCI_DMA_TODEVICE);
|
|
|
+ bf->skbaddr = dma_map_single(sc->dev, skb->data, skb->len,
|
|
|
+ DMA_TO_DEVICE);
|
|
|
|
|
|
rate = ieee80211_get_tx_rate(sc->hw, info);
|
|
|
if (!rate) {
|
|
@@ -827,7 +828,7 @@ ath5k_txbuf_setup(struct ath5k_softc *sc, struct ath5k_buf *bf,
|
|
|
|
|
|
return 0;
|
|
|
err_unmap:
|
|
|
- pci_unmap_single(sc->pdev, bf->skbaddr, skb->len, PCI_DMA_TODEVICE);
|
|
|
+ dma_unmap_single(sc->dev, bf->skbaddr, skb->len, DMA_TO_DEVICE);
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
@@ -836,7 +837,7 @@ err_unmap:
|
|
|
\*******************/
|
|
|
|
|
|
static int
|
|
|
-ath5k_desc_alloc(struct ath5k_softc *sc, struct pci_dev *pdev)
|
|
|
+ath5k_desc_alloc(struct ath5k_softc *sc)
|
|
|
{
|
|
|
struct ath5k_desc *ds;
|
|
|
struct ath5k_buf *bf;
|
|
@@ -847,7 +848,9 @@ ath5k_desc_alloc(struct ath5k_softc *sc, struct pci_dev *pdev)
|
|
|
/* allocate descriptors */
|
|
|
sc->desc_len = sizeof(struct ath5k_desc) *
|
|
|
(ATH_TXBUF + ATH_RXBUF + ATH_BCBUF + 1);
|
|
|
- sc->desc = pci_alloc_consistent(pdev, sc->desc_len, &sc->desc_daddr);
|
|
|
+
|
|
|
+ sc->desc = dma_alloc_coherent(sc->dev, sc->desc_len,
|
|
|
+ &sc->desc_daddr, GFP_KERNEL);
|
|
|
if (sc->desc == NULL) {
|
|
|
ATH5K_ERR(sc, "can't allocate descriptors\n");
|
|
|
ret = -ENOMEM;
|
|
@@ -893,14 +896,14 @@ ath5k_desc_alloc(struct ath5k_softc *sc, struct pci_dev *pdev)
|
|
|
|
|
|
return 0;
|
|
|
err_free:
|
|
|
- pci_free_consistent(pdev, sc->desc_len, sc->desc, sc->desc_daddr);
|
|
|
+ dma_free_coherent(sc->dev, sc->desc_len, sc->desc, sc->desc_daddr);
|
|
|
err:
|
|
|
sc->desc = NULL;
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
static void
|
|
|
-ath5k_desc_free(struct ath5k_softc *sc, struct pci_dev *pdev)
|
|
|
+ath5k_desc_free(struct ath5k_softc *sc)
|
|
|
{
|
|
|
struct ath5k_buf *bf;
|
|
|
|
|
@@ -912,7 +915,7 @@ ath5k_desc_free(struct ath5k_softc *sc, struct pci_dev *pdev)
|
|
|
ath5k_txbuf_free_skb(sc, bf);
|
|
|
|
|
|
/* Free memory associated with all descriptors */
|
|
|
- pci_free_consistent(pdev, sc->desc_len, sc->desc, sc->desc_daddr);
|
|
|
+ dma_free_coherent(sc->dev, sc->desc_len, sc->desc, sc->desc_daddr);
|
|
|
sc->desc = NULL;
|
|
|
sc->desc_daddr = 0;
|
|
|
|
|
@@ -1523,9 +1526,9 @@ ath5k_tasklet_rx(unsigned long data)
|
|
|
if (!next_skb)
|
|
|
goto next;
|
|
|
|
|
|
- pci_unmap_single(sc->pdev, bf->skbaddr,
|
|
|
+ dma_unmap_single(sc->dev, bf->skbaddr,
|
|
|
common->rx_bufsize,
|
|
|
- PCI_DMA_FROMDEVICE);
|
|
|
+ DMA_FROM_DEVICE);
|
|
|
|
|
|
skb_put(skb, rs.rs_datalen);
|
|
|
|
|
@@ -1688,8 +1691,9 @@ ath5k_tx_processq(struct ath5k_softc *sc, struct ath5k_txq *txq)
|
|
|
|
|
|
skb = bf->skb;
|
|
|
bf->skb = NULL;
|
|
|
- pci_unmap_single(sc->pdev, bf->skbaddr, skb->len,
|
|
|
- PCI_DMA_TODEVICE);
|
|
|
+
|
|
|
+ dma_unmap_single(sc->dev, bf->skbaddr, skb->len,
|
|
|
+ DMA_TO_DEVICE);
|
|
|
ath5k_tx_frame_completed(sc, skb, &ts);
|
|
|
}
|
|
|
|
|
@@ -1743,12 +1747,13 @@ ath5k_beacon_setup(struct ath5k_softc *sc, struct ath5k_buf *bf)
|
|
|
u32 flags;
|
|
|
const int padsize = 0;
|
|
|
|
|
|
- bf->skbaddr = pci_map_single(sc->pdev, skb->data, skb->len,
|
|
|
- PCI_DMA_TODEVICE);
|
|
|
+ bf->skbaddr = dma_map_single(sc->dev, skb->data, skb->len,
|
|
|
+ DMA_TO_DEVICE);
|
|
|
ATH5K_DBG(sc, ATH5K_DEBUG_BEACON, "skb %p [data %p len %u] "
|
|
|
"skbaddr %llx\n", skb, skb->data, skb->len,
|
|
|
(unsigned long long)bf->skbaddr);
|
|
|
- if (pci_dma_mapping_error(sc->pdev, bf->skbaddr)) {
|
|
|
+
|
|
|
+ if (dma_mapping_error(sc->dev, bf->skbaddr)) {
|
|
|
ATH5K_ERR(sc, "beacon DMA mapping failed\n");
|
|
|
return -EIO;
|
|
|
}
|
|
@@ -1800,7 +1805,7 @@ ath5k_beacon_setup(struct ath5k_softc *sc, struct ath5k_buf *bf)
|
|
|
|
|
|
return 0;
|
|
|
err_unmap:
|
|
|
- pci_unmap_single(sc->pdev, bf->skbaddr, skb->len, PCI_DMA_TODEVICE);
|
|
|
+ dma_unmap_single(sc->dev, bf->skbaddr, skb->len, DMA_TO_DEVICE);
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
@@ -2361,7 +2366,7 @@ ath5k_stop_locked(struct ath5k_softc *sc)
|
|
|
if (!test_bit(ATH_STAT_INVALID, sc->status)) {
|
|
|
ath5k_led_off(sc);
|
|
|
ath5k_hw_set_imr(ah, 0);
|
|
|
- synchronize_irq(sc->pdev->irq);
|
|
|
+ synchronize_irq(sc->irq);
|
|
|
ath5k_rx_stop(sc);
|
|
|
ath5k_hw_dma_stop(ah);
|
|
|
ath5k_drain_tx_buffs(sc);
|
|
@@ -2509,7 +2514,7 @@ ath5k_reset(struct ath5k_softc *sc, struct ieee80211_channel *chan,
|
|
|
ATH5K_DBG(sc, ATH5K_DEBUG_RESET, "resetting\n");
|
|
|
|
|
|
ath5k_hw_set_imr(ah, 0);
|
|
|
- synchronize_irq(sc->pdev->irq);
|
|
|
+ synchronize_irq(sc->irq);
|
|
|
stop_tasklets(sc);
|
|
|
|
|
|
if (chan) {
|
|
@@ -2616,7 +2621,7 @@ ath5k_attach(struct pci_dev *pdev, struct ieee80211_hw *hw)
|
|
|
/*
|
|
|
* Allocate tx+rx descriptors and populate the lists.
|
|
|
*/
|
|
|
- ret = ath5k_desc_alloc(sc, pdev);
|
|
|
+ ret = ath5k_desc_alloc(sc);
|
|
|
if (ret) {
|
|
|
ATH5K_ERR(sc, "can't allocate descriptors\n");
|
|
|
goto err;
|
|
@@ -2680,8 +2685,7 @@ ath5k_attach(struct pci_dev *pdev, struct ieee80211_hw *hw)
|
|
|
|
|
|
ret = ath5k_eeprom_read_mac(ah, mac);
|
|
|
if (ret) {
|
|
|
- ATH5K_ERR(sc, "unable to read address from EEPROM: 0x%04x\n",
|
|
|
- sc->pdev->device);
|
|
|
+ ATH5K_ERR(sc, "unable to read address from EEPROM\n");
|
|
|
goto err_queues;
|
|
|
}
|
|
|
|
|
@@ -2716,7 +2720,7 @@ err_queues:
|
|
|
err_bhal:
|
|
|
ath5k_hw_release_tx_queue(ah, sc->bhalq);
|
|
|
err_desc:
|
|
|
- ath5k_desc_free(sc, pdev);
|
|
|
+ ath5k_desc_free(sc);
|
|
|
err:
|
|
|
return ret;
|
|
|
}
|
|
@@ -2740,7 +2744,7 @@ ath5k_detach(struct pci_dev *pdev, struct ieee80211_hw *hw)
|
|
|
* Other than that, it's straightforward...
|
|
|
*/
|
|
|
ieee80211_unregister_hw(hw);
|
|
|
- ath5k_desc_free(sc, pdev);
|
|
|
+ ath5k_desc_free(sc);
|
|
|
ath5k_txq_release(sc);
|
|
|
ath5k_hw_release_tx_queue(sc->ah, sc->bhalq);
|
|
|
ath5k_unregister_leds(sc);
|
|
@@ -3565,6 +3569,8 @@ ath5k_pci_probe(struct pci_dev *pdev,
|
|
|
sc = hw->priv;
|
|
|
sc->hw = hw;
|
|
|
sc->pdev = pdev;
|
|
|
+ sc->dev = &pdev->dev;
|
|
|
+ sc->irq = pdev->irq;
|
|
|
|
|
|
/*
|
|
|
* Mark the device as detached to avoid processing
|