|
@@ -138,30 +138,6 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
|
|
|
}
|
|
|
#endif
|
|
|
|
|
|
-/**
|
|
|
- * ixgbe_cache_ring_fdir - Descriptor ring to register mapping for Flow Director
|
|
|
- * @adapter: board private structure to initialize
|
|
|
- *
|
|
|
- * Cache the descriptor ring offsets for Flow Director to the assigned rings.
|
|
|
- *
|
|
|
- **/
|
|
|
-static inline bool ixgbe_cache_ring_fdir(struct ixgbe_adapter *adapter)
|
|
|
-{
|
|
|
- int i;
|
|
|
- bool ret = false;
|
|
|
-
|
|
|
- if ((adapter->flags & IXGBE_FLAG_RSS_ENABLED) &&
|
|
|
- (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE)) {
|
|
|
- for (i = 0; i < adapter->num_rx_queues; i++)
|
|
|
- adapter->rx_ring[i]->reg_idx = i;
|
|
|
- for (i = 0; i < adapter->num_tx_queues; i++)
|
|
|
- adapter->tx_ring[i]->reg_idx = i;
|
|
|
- ret = true;
|
|
|
- }
|
|
|
-
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
#ifdef IXGBE_FCOE
|
|
|
/**
|
|
|
* ixgbe_cache_ring_fcoe - Descriptor ring to register mapping for the FCoE
|
|
@@ -180,10 +156,7 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
|
|
|
return false;
|
|
|
|
|
|
if (adapter->flags & IXGBE_FLAG_RSS_ENABLED) {
|
|
|
- if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE)
|
|
|
- ixgbe_cache_ring_fdir(adapter);
|
|
|
- else
|
|
|
- ixgbe_cache_ring_rss(adapter);
|
|
|
+ ixgbe_cache_ring_rss(adapter);
|
|
|
|
|
|
fcoe_rx_i = f->offset;
|
|
|
fcoe_tx_i = f->offset;
|
|
@@ -244,9 +217,6 @@ static void ixgbe_cache_ring_register(struct ixgbe_adapter *adapter)
|
|
|
return;
|
|
|
#endif /* IXGBE_FCOE */
|
|
|
|
|
|
- if (ixgbe_cache_ring_fdir(adapter))
|
|
|
- return;
|
|
|
-
|
|
|
if (ixgbe_cache_ring_rss(adapter))
|
|
|
return;
|
|
|
}
|
|
@@ -272,53 +242,39 @@ static inline bool ixgbe_set_sriov_queues(struct ixgbe_adapter *adapter)
|
|
|
* to allocate one Rx queue per CPU, and if available, one Tx queue per CPU.
|
|
|
*
|
|
|
**/
|
|
|
-static inline bool ixgbe_set_rss_queues(struct ixgbe_adapter *adapter)
|
|
|
+static bool ixgbe_set_rss_queues(struct ixgbe_adapter *adapter)
|
|
|
{
|
|
|
- bool ret = false;
|
|
|
- struct ixgbe_ring_feature *f = &adapter->ring_feature[RING_F_RSS];
|
|
|
+ struct ixgbe_ring_feature *f;
|
|
|
+ u16 rss_i;
|
|
|
|
|
|
- if (adapter->flags & IXGBE_FLAG_RSS_ENABLED) {
|
|
|
- f->mask = 0xF;
|
|
|
- adapter->num_rx_queues = f->indices;
|
|
|
- adapter->num_tx_queues = f->indices;
|
|
|
- ret = true;
|
|
|
+ if (!(adapter->flags & IXGBE_FLAG_RSS_ENABLED)) {
|
|
|
+ adapter->flags &= ~IXGBE_FLAG_FDIR_HASH_CAPABLE;
|
|
|
+ return false;
|
|
|
}
|
|
|
|
|
|
- return ret;
|
|
|
-}
|
|
|
+ /* set mask for 16 queue limit of RSS */
|
|
|
+ f = &adapter->ring_feature[RING_F_RSS];
|
|
|
+ rss_i = f->limit;
|
|
|
|
|
|
-/**
|
|
|
- * ixgbe_set_fdir_queues - Allocate queues for Flow Director
|
|
|
- * @adapter: board private structure to initialize
|
|
|
- *
|
|
|
- * Flow Director is an advanced Rx filter, attempting to get Rx flows back
|
|
|
- * to the original CPU that initiated the Tx session. This runs in addition
|
|
|
- * to RSS, so if a packet doesn't match an FDIR filter, we can still spread the
|
|
|
- * Rx load across CPUs using RSS.
|
|
|
- *
|
|
|
- **/
|
|
|
-static inline bool ixgbe_set_fdir_queues(struct ixgbe_adapter *adapter)
|
|
|
-{
|
|
|
- bool ret = false;
|
|
|
- struct ixgbe_ring_feature *f_fdir = &adapter->ring_feature[RING_F_FDIR];
|
|
|
-
|
|
|
- f_fdir->indices = min_t(int, num_online_cpus(), f_fdir->limit);
|
|
|
- f_fdir->mask = 0;
|
|
|
+ f->indices = rss_i;
|
|
|
+ f->mask = 0xF;
|
|
|
|
|
|
/*
|
|
|
- * Use RSS in addition to Flow Director to ensure the best
|
|
|
+ * Use Flow Director in addition to RSS to ensure the best
|
|
|
* distribution of flows across cores, even when an FDIR flow
|
|
|
* isn't matched.
|
|
|
*/
|
|
|
- if ((adapter->flags & IXGBE_FLAG_RSS_ENABLED) &&
|
|
|
- (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE)) {
|
|
|
- adapter->num_tx_queues = f_fdir->indices;
|
|
|
- adapter->num_rx_queues = f_fdir->indices;
|
|
|
- ret = true;
|
|
|
- } else {
|
|
|
- adapter->flags &= ~IXGBE_FLAG_FDIR_HASH_CAPABLE;
|
|
|
+ if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) {
|
|
|
+ f = &adapter->ring_feature[RING_F_FDIR];
|
|
|
+
|
|
|
+ f->indices = min_t(u16, num_online_cpus(), f->limit);
|
|
|
+ rss_i = max_t(u16, rss_i, f->indices);
|
|
|
}
|
|
|
- return ret;
|
|
|
+
|
|
|
+ adapter->num_rx_queues = rss_i;
|
|
|
+ adapter->num_tx_queues = rss_i;
|
|
|
+
|
|
|
+ return true;
|
|
|
}
|
|
|
|
|
|
#ifdef IXGBE_FCOE
|
|
@@ -343,10 +299,7 @@ static inline bool ixgbe_set_fcoe_queues(struct ixgbe_adapter *adapter)
|
|
|
|
|
|
if (adapter->flags & IXGBE_FLAG_RSS_ENABLED) {
|
|
|
e_info(probe, "FCoE enabled with RSS\n");
|
|
|
- if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE)
|
|
|
- ixgbe_set_fdir_queues(adapter);
|
|
|
- else
|
|
|
- ixgbe_set_rss_queues(adapter);
|
|
|
+ ixgbe_set_rss_queues(adapter);
|
|
|
}
|
|
|
|
|
|
/* adding FCoE rx rings to the end */
|
|
@@ -438,9 +391,6 @@ static int ixgbe_set_num_queues(struct ixgbe_adapter *adapter)
|
|
|
goto done;
|
|
|
|
|
|
#endif /* IXGBE_FCOE */
|
|
|
- if (ixgbe_set_fdir_queues(adapter))
|
|
|
- goto done;
|
|
|
-
|
|
|
if (ixgbe_set_rss_queues(adapter))
|
|
|
goto done;
|
|
|
|