|
@@ -190,7 +190,7 @@ int bnx2x_tx_int(struct bnx2x *bp, struct bnx2x_fp_txdata *txdata)
|
|
|
|
|
|
if ((netif_tx_queue_stopped(txq)) &&
|
|
|
(bp->state == BNX2X_STATE_OPEN) &&
|
|
|
- (bnx2x_tx_avail(bp, txdata) >= MAX_SKB_FRAGS + 3))
|
|
|
+ (bnx2x_tx_avail(bp, txdata) >= MAX_SKB_FRAGS + 4))
|
|
|
netif_tx_wake_queue(txq);
|
|
|
|
|
|
__netif_tx_unlock(txq);
|
|
@@ -2516,8 +2516,6 @@ int bnx2x_poll(struct napi_struct *napi, int budget)
|
|
|
/* we split the first BD into headers and data BDs
|
|
|
* to ease the pain of our fellow microcode engineers
|
|
|
* we use one mapping for both BDs
|
|
|
- * So far this has only been observed to happen
|
|
|
- * in Other Operating Systems(TM)
|
|
|
*/
|
|
|
static noinline u16 bnx2x_tx_split(struct bnx2x *bp,
|
|
|
struct bnx2x_fp_txdata *txdata,
|
|
@@ -3171,7 +3169,7 @@ netdev_tx_t bnx2x_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|
|
|
|
|
txdata->tx_bd_prod += nbd;
|
|
|
|
|
|
- if (unlikely(bnx2x_tx_avail(bp, txdata) < MAX_SKB_FRAGS + 3)) {
|
|
|
+ if (unlikely(bnx2x_tx_avail(bp, txdata) < MAX_SKB_FRAGS + 4)) {
|
|
|
netif_tx_stop_queue(txq);
|
|
|
|
|
|
/* paired memory barrier is in bnx2x_tx_int(), we have to keep
|
|
@@ -3180,7 +3178,7 @@ netdev_tx_t bnx2x_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|
|
smp_mb();
|
|
|
|
|
|
fp->eth_q_stats.driver_xoff++;
|
|
|
- if (bnx2x_tx_avail(bp, txdata) >= MAX_SKB_FRAGS + 3)
|
|
|
+ if (bnx2x_tx_avail(bp, txdata) >= MAX_SKB_FRAGS + 4)
|
|
|
netif_tx_wake_queue(txq);
|
|
|
}
|
|
|
txdata->tx_pkt++;
|