|
@@ -195,15 +195,6 @@ static inline void _tg3_flag_clear(enum TG3_FLAGS flag, unsigned long *bits)
|
|
#define TG3_RX_OFFSET(tp) (NET_SKB_PAD)
|
|
#define TG3_RX_OFFSET(tp) (NET_SKB_PAD)
|
|
#endif
|
|
#endif
|
|
|
|
|
|
-/* This driver uses the new build_skb() API providing a frag as skb->head
|
|
|
|
- * This strategy permits better GRO aggregation, better TCP coalescing, and
|
|
|
|
- * better splice() implementation (avoids a copy from head to a page), at
|
|
|
|
- * minimal memory cost.
|
|
|
|
- * In this 2048 bytes block, we have enough room to store the MTU=1500 frame
|
|
|
|
- * and the struct skb_shared_info.
|
|
|
|
- */
|
|
|
|
-#define TG3_FRAGSIZE 2048
|
|
|
|
-
|
|
|
|
/* minimum number of free TX descriptors required to wake up TX process */
|
|
/* minimum number of free TX descriptors required to wake up TX process */
|
|
#define TG3_TX_WAKEUP_THRESH(tnapi) ((tnapi)->tx_pending / 4)
|
|
#define TG3_TX_WAKEUP_THRESH(tnapi) ((tnapi)->tx_pending / 4)
|
|
#define TG3_TX_BD_DMA_MAX_2K 2048
|
|
#define TG3_TX_BD_DMA_MAX_2K 2048
|
|
@@ -5631,25 +5622,6 @@ static void tg3_tx(struct tg3_napi *tnapi)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
-static void *tg3_frag_alloc(struct tg3_rx_prodring_set *tpr)
|
|
|
|
-{
|
|
|
|
- void *data;
|
|
|
|
-
|
|
|
|
- if (tpr->rx_page_size < TG3_FRAGSIZE) {
|
|
|
|
- struct page *page = alloc_page(GFP_ATOMIC);
|
|
|
|
-
|
|
|
|
- if (!page)
|
|
|
|
- return NULL;
|
|
|
|
- atomic_add((PAGE_SIZE / TG3_FRAGSIZE) - 1, &page->_count);
|
|
|
|
- tpr->rx_page_addr = page_address(page);
|
|
|
|
- tpr->rx_page_size = PAGE_SIZE;
|
|
|
|
- }
|
|
|
|
- data = tpr->rx_page_addr;
|
|
|
|
- tpr->rx_page_addr += TG3_FRAGSIZE;
|
|
|
|
- tpr->rx_page_size -= TG3_FRAGSIZE;
|
|
|
|
- return data;
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
static void tg3_frag_free(bool is_frag, void *data)
|
|
static void tg3_frag_free(bool is_frag, void *data)
|
|
{
|
|
{
|
|
if (is_frag)
|
|
if (is_frag)
|
|
@@ -5668,7 +5640,7 @@ static void tg3_rx_data_free(struct tg3 *tp, struct ring_info *ri, u32 map_sz)
|
|
|
|
|
|
pci_unmap_single(tp->pdev, dma_unmap_addr(ri, mapping),
|
|
pci_unmap_single(tp->pdev, dma_unmap_addr(ri, mapping),
|
|
map_sz, PCI_DMA_FROMDEVICE);
|
|
map_sz, PCI_DMA_FROMDEVICE);
|
|
- tg3_frag_free(skb_size <= TG3_FRAGSIZE, ri->data);
|
|
|
|
|
|
+ tg3_frag_free(skb_size <= PAGE_SIZE, ri->data);
|
|
ri->data = NULL;
|
|
ri->data = NULL;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -5721,9 +5693,9 @@ static int tg3_alloc_rx_data(struct tg3 *tp, struct tg3_rx_prodring_set *tpr,
|
|
*/
|
|
*/
|
|
skb_size = SKB_DATA_ALIGN(data_size + TG3_RX_OFFSET(tp)) +
|
|
skb_size = SKB_DATA_ALIGN(data_size + TG3_RX_OFFSET(tp)) +
|
|
SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
|
|
SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
|
|
- if (skb_size <= TG3_FRAGSIZE) {
|
|
|
|
- data = tg3_frag_alloc(tpr);
|
|
|
|
- *frag_size = TG3_FRAGSIZE;
|
|
|
|
|
|
+ if (skb_size <= PAGE_SIZE) {
|
|
|
|
+ data = netdev_alloc_frag(skb_size);
|
|
|
|
+ *frag_size = skb_size;
|
|
} else {
|
|
} else {
|
|
data = kmalloc(skb_size, GFP_ATOMIC);
|
|
data = kmalloc(skb_size, GFP_ATOMIC);
|
|
*frag_size = 0;
|
|
*frag_size = 0;
|
|
@@ -5736,7 +5708,7 @@ static int tg3_alloc_rx_data(struct tg3 *tp, struct tg3_rx_prodring_set *tpr,
|
|
data_size,
|
|
data_size,
|
|
PCI_DMA_FROMDEVICE);
|
|
PCI_DMA_FROMDEVICE);
|
|
if (unlikely(pci_dma_mapping_error(tp->pdev, mapping))) {
|
|
if (unlikely(pci_dma_mapping_error(tp->pdev, mapping))) {
|
|
- tg3_frag_free(skb_size <= TG3_FRAGSIZE, data);
|
|
|
|
|
|
+ tg3_frag_free(skb_size <= PAGE_SIZE, data);
|
|
return -EIO;
|
|
return -EIO;
|
|
}
|
|
}
|
|
|
|
|