|
@@ -436,6 +436,8 @@ static int bgmac_dma_alloc(struct bgmac *bgmac)
|
|
}
|
|
}
|
|
|
|
|
|
for (i = 0; i < BGMAC_MAX_RX_RINGS; i++) {
|
|
for (i = 0; i < BGMAC_MAX_RX_RINGS; i++) {
|
|
|
|
+ int j;
|
|
|
|
+
|
|
ring = &bgmac->rx_ring[i];
|
|
ring = &bgmac->rx_ring[i];
|
|
ring->num_slots = BGMAC_RX_RING_SLOTS;
|
|
ring->num_slots = BGMAC_RX_RING_SLOTS;
|
|
ring->mmio_base = ring_base[i];
|
|
ring->mmio_base = ring_base[i];
|
|
@@ -458,8 +460,8 @@ static int bgmac_dma_alloc(struct bgmac *bgmac)
|
|
bgmac_warn(bgmac, "DMA address using 0xC0000000 bit(s), it may need translation trick\n");
|
|
bgmac_warn(bgmac, "DMA address using 0xC0000000 bit(s), it may need translation trick\n");
|
|
|
|
|
|
/* Alloc RX slots */
|
|
/* Alloc RX slots */
|
|
- for (i = 0; i < ring->num_slots; i++) {
|
|
|
|
- err = bgmac_dma_rx_skb_for_slot(bgmac, &ring->slots[i]);
|
|
|
|
|
|
+ for (j = 0; j < ring->num_slots; j++) {
|
|
|
|
+ err = bgmac_dma_rx_skb_for_slot(bgmac, &ring->slots[j]);
|
|
if (err) {
|
|
if (err) {
|
|
bgmac_err(bgmac, "Can't allocate skb for slot in RX ring\n");
|
|
bgmac_err(bgmac, "Can't allocate skb for slot in RX ring\n");
|
|
goto err_dma_free;
|
|
goto err_dma_free;
|
|
@@ -496,6 +498,8 @@ static void bgmac_dma_init(struct bgmac *bgmac)
|
|
}
|
|
}
|
|
|
|
|
|
for (i = 0; i < BGMAC_MAX_RX_RINGS; i++) {
|
|
for (i = 0; i < BGMAC_MAX_RX_RINGS; i++) {
|
|
|
|
+ int j;
|
|
|
|
+
|
|
ring = &bgmac->rx_ring[i];
|
|
ring = &bgmac->rx_ring[i];
|
|
|
|
|
|
/* We don't implement unaligned addressing, so enable first */
|
|
/* We don't implement unaligned addressing, so enable first */
|
|
@@ -505,11 +509,11 @@ static void bgmac_dma_init(struct bgmac *bgmac)
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_RINGHI,
|
|
bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_RINGHI,
|
|
upper_32_bits(ring->dma_base));
|
|
upper_32_bits(ring->dma_base));
|
|
|
|
|
|
- for (i = 0, dma_desc = ring->cpu_base; i < ring->num_slots;
|
|
|
|
- i++, dma_desc++) {
|
|
|
|
|
|
+ for (j = 0, dma_desc = ring->cpu_base; j < ring->num_slots;
|
|
|
|
+ j++, dma_desc++) {
|
|
ctl0 = ctl1 = 0;
|
|
ctl0 = ctl1 = 0;
|
|
|
|
|
|
- if (i == ring->num_slots - 1)
|
|
|
|
|
|
+ if (j == ring->num_slots - 1)
|
|
ctl0 |= BGMAC_DESC_CTL0_EOT;
|
|
ctl0 |= BGMAC_DESC_CTL0_EOT;
|
|
ctl1 |= BGMAC_RX_BUF_SIZE & BGMAC_DESC_CTL1_LEN;
|
|
ctl1 |= BGMAC_RX_BUF_SIZE & BGMAC_DESC_CTL1_LEN;
|
|
/* Is there any BGMAC device that requires extension? */
|
|
/* Is there any BGMAC device that requires extension? */
|
|
@@ -517,8 +521,8 @@ static void bgmac_dma_init(struct bgmac *bgmac)
|
|
* B43_DMA64_DCTL1_ADDREXT_MASK;
|
|
* B43_DMA64_DCTL1_ADDREXT_MASK;
|
|
*/
|
|
*/
|
|
|
|
|
|
- dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[i].dma_addr));
|
|
|
|
- dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[i].dma_addr));
|
|
|
|
|
|
+ dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[j].dma_addr));
|
|
|
|
+ dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[j].dma_addr));
|
|
dma_desc->ctl0 = cpu_to_le32(ctl0);
|
|
dma_desc->ctl0 = cpu_to_le32(ctl0);
|
|
dma_desc->ctl1 = cpu_to_le32(ctl1);
|
|
dma_desc->ctl1 = cpu_to_le32(ctl1);
|
|
}
|
|
}
|