|
@@ -2014,6 +2014,7 @@ static void e1000_clean_tx_ring(struct e1000_adapter *adapter,
|
|
|
e1000_unmap_and_free_tx_resource(adapter, buffer_info);
|
|
|
}
|
|
|
|
|
|
+ netdev_reset_queue(adapter->netdev);
|
|
|
size = sizeof(struct e1000_buffer) * tx_ring->count;
|
|
|
memset(tx_ring->buffer_info, 0, size);
|
|
|
|
|
@@ -3262,6 +3263,7 @@ static netdev_tx_t e1000_xmit_frame(struct sk_buff *skb,
|
|
|
nr_frags, mss);
|
|
|
|
|
|
if (count) {
|
|
|
+ netdev_sent_queue(netdev, skb->len);
|
|
|
skb_tx_timestamp(skb);
|
|
|
|
|
|
e1000_tx_queue(adapter, tx_ring, tx_flags, count);
|
|
@@ -3849,6 +3851,7 @@ static bool e1000_clean_tx_irq(struct e1000_adapter *adapter,
|
|
|
unsigned int i, eop;
|
|
|
unsigned int count = 0;
|
|
|
unsigned int total_tx_bytes=0, total_tx_packets=0;
|
|
|
+ unsigned int bytes_compl = 0, pkts_compl = 0;
|
|
|
|
|
|
i = tx_ring->next_to_clean;
|
|
|
eop = tx_ring->buffer_info[i].next_to_watch;
|
|
@@ -3866,6 +3869,11 @@ static bool e1000_clean_tx_irq(struct e1000_adapter *adapter,
|
|
|
if (cleaned) {
|
|
|
total_tx_packets += buffer_info->segs;
|
|
|
total_tx_bytes += buffer_info->bytecount;
|
|
|
+ if (buffer_info->skb) {
|
|
|
+ bytes_compl += buffer_info->skb->len;
|
|
|
+ pkts_compl++;
|
|
|
+ }
|
|
|
+
|
|
|
}
|
|
|
e1000_unmap_and_free_tx_resource(adapter, buffer_info);
|
|
|
tx_desc->upper.data = 0;
|
|
@@ -3879,6 +3887,8 @@ static bool e1000_clean_tx_irq(struct e1000_adapter *adapter,
|
|
|
|
|
|
tx_ring->next_to_clean = i;
|
|
|
|
|
|
+ netdev_completed_queue(netdev, pkts_compl, bytes_compl);
|
|
|
+
|
|
|
#define TX_WAKE_THRESHOLD 32
|
|
|
if (unlikely(count && netif_carrier_ok(netdev) &&
|
|
|
E1000_DESC_UNUSED(tx_ring) >= TX_WAKE_THRESHOLD)) {
|