|
@@ -668,9 +668,9 @@ static inline u16 seq_sub(u16 sq1, u16 sq2)
|
|
|
|
|
|
static void ieee80211_release_reorder_frame(struct ieee80211_sub_if_data *sdata,
|
|
|
struct tid_ampdu_rx *tid_agg_rx,
|
|
|
- int index)
|
|
|
+ int index,
|
|
|
+ struct sk_buff_head *frames)
|
|
|
{
|
|
|
- struct ieee80211_local *local = sdata->local;
|
|
|
struct sk_buff *skb = tid_agg_rx->reorder_buf[index];
|
|
|
struct ieee80211_rx_status *status;
|
|
|
|
|
@@ -684,7 +684,7 @@ static void ieee80211_release_reorder_frame(struct ieee80211_sub_if_data *sdata,
|
|
|
tid_agg_rx->reorder_buf[index] = NULL;
|
|
|
status = IEEE80211_SKB_RXCB(skb);
|
|
|
status->rx_flags |= IEEE80211_RX_DEFERRED_RELEASE;
|
|
|
- skb_queue_tail(&local->rx_skb_queue, skb);
|
|
|
+ __skb_queue_tail(frames, skb);
|
|
|
|
|
|
no_frame:
|
|
|
tid_agg_rx->head_seq_num = seq_inc(tid_agg_rx->head_seq_num);
|
|
@@ -692,7 +692,8 @@ no_frame:
|
|
|
|
|
|
static void ieee80211_release_reorder_frames(struct ieee80211_sub_if_data *sdata,
|
|
|
struct tid_ampdu_rx *tid_agg_rx,
|
|
|
- u16 head_seq_num)
|
|
|
+ u16 head_seq_num,
|
|
|
+ struct sk_buff_head *frames)
|
|
|
{
|
|
|
int index;
|
|
|
|
|
@@ -701,7 +702,8 @@ static void ieee80211_release_reorder_frames(struct ieee80211_sub_if_data *sdata
|
|
|
while (seq_less(tid_agg_rx->head_seq_num, head_seq_num)) {
|
|
|
index = seq_sub(tid_agg_rx->head_seq_num, tid_agg_rx->ssn) %
|
|
|
tid_agg_rx->buf_size;
|
|
|
- ieee80211_release_reorder_frame(sdata, tid_agg_rx, index);
|
|
|
+ ieee80211_release_reorder_frame(sdata, tid_agg_rx, index,
|
|
|
+ frames);
|
|
|
}
|
|
|
}
|
|
|
|
|
@@ -717,7 +719,8 @@ static void ieee80211_release_reorder_frames(struct ieee80211_sub_if_data *sdata
|
|
|
#define HT_RX_REORDER_BUF_TIMEOUT (HZ / 10)
|
|
|
|
|
|
static void ieee80211_sta_reorder_release(struct ieee80211_sub_if_data *sdata,
|
|
|
- struct tid_ampdu_rx *tid_agg_rx)
|
|
|
+ struct tid_ampdu_rx *tid_agg_rx,
|
|
|
+ struct sk_buff_head *frames)
|
|
|
{
|
|
|
int index, j;
|
|
|
|
|
@@ -746,7 +749,8 @@ static void ieee80211_sta_reorder_release(struct ieee80211_sub_if_data *sdata,
|
|
|
|
|
|
ht_dbg_ratelimited(sdata,
|
|
|
"release an RX reorder frame due to timeout on earlier frames\n");
|
|
|
- ieee80211_release_reorder_frame(sdata, tid_agg_rx, j);
|
|
|
+ ieee80211_release_reorder_frame(sdata, tid_agg_rx, j,
|
|
|
+ frames);
|
|
|
|
|
|
/*
|
|
|
* Increment the head seq# also for the skipped slots.
|
|
@@ -756,7 +760,8 @@ static void ieee80211_sta_reorder_release(struct ieee80211_sub_if_data *sdata,
|
|
|
skipped = 0;
|
|
|
}
|
|
|
} else while (tid_agg_rx->reorder_buf[index]) {
|
|
|
- ieee80211_release_reorder_frame(sdata, tid_agg_rx, index);
|
|
|
+ ieee80211_release_reorder_frame(sdata, tid_agg_rx, index,
|
|
|
+ frames);
|
|
|
index = seq_sub(tid_agg_rx->head_seq_num, tid_agg_rx->ssn) %
|
|
|
tid_agg_rx->buf_size;
|
|
|
}
|
|
@@ -788,7 +793,8 @@ static void ieee80211_sta_reorder_release(struct ieee80211_sub_if_data *sdata,
|
|
|
*/
|
|
|
static bool ieee80211_sta_manage_reorder_buf(struct ieee80211_sub_if_data *sdata,
|
|
|
struct tid_ampdu_rx *tid_agg_rx,
|
|
|
- struct sk_buff *skb)
|
|
|
+ struct sk_buff *skb,
|
|
|
+ struct sk_buff_head *frames)
|
|
|
{
|
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data;
|
|
|
u16 sc = le16_to_cpu(hdr->seq_ctrl);
|
|
@@ -816,7 +822,7 @@ static bool ieee80211_sta_manage_reorder_buf(struct ieee80211_sub_if_data *sdata
|
|
|
head_seq_num = seq_inc(seq_sub(mpdu_seq_num, buf_size));
|
|
|
/* release stored frames up to new head to stack */
|
|
|
ieee80211_release_reorder_frames(sdata, tid_agg_rx,
|
|
|
- head_seq_num);
|
|
|
+ head_seq_num, frames);
|
|
|
}
|
|
|
|
|
|
/* Now the new frame is always in the range of the reordering buffer */
|
|
@@ -846,7 +852,7 @@ static bool ieee80211_sta_manage_reorder_buf(struct ieee80211_sub_if_data *sdata
|
|
|
tid_agg_rx->reorder_buf[index] = skb;
|
|
|
tid_agg_rx->reorder_time[index] = jiffies;
|
|
|
tid_agg_rx->stored_mpdu_num++;
|
|
|
- ieee80211_sta_reorder_release(sdata, tid_agg_rx);
|
|
|
+ ieee80211_sta_reorder_release(sdata, tid_agg_rx, frames);
|
|
|
|
|
|
out:
|
|
|
spin_unlock(&tid_agg_rx->reorder_lock);
|
|
@@ -857,7 +863,8 @@ static bool ieee80211_sta_manage_reorder_buf(struct ieee80211_sub_if_data *sdata
|
|
|
* Reorder MPDUs from A-MPDUs, keeping them on a buffer. Returns
|
|
|
* true if the MPDU was buffered, false if it should be processed.
|
|
|
*/
|
|
|
-static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx)
|
|
|
+static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx,
|
|
|
+ struct sk_buff_head *frames)
|
|
|
{
|
|
|
struct sk_buff *skb = rx->skb;
|
|
|
struct ieee80211_local *local = rx->local;
|
|
@@ -922,11 +929,12 @@ static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx)
|
|
|
* sure that we cannot get to it any more before doing
|
|
|
* anything with it.
|
|
|
*/
|
|
|
- if (ieee80211_sta_manage_reorder_buf(rx->sdata, tid_agg_rx, skb))
|
|
|
+ if (ieee80211_sta_manage_reorder_buf(rx->sdata, tid_agg_rx, skb,
|
|
|
+ frames))
|
|
|
return;
|
|
|
|
|
|
dont_reorder:
|
|
|
- skb_queue_tail(&local->rx_skb_queue, skb);
|
|
|
+ __skb_queue_tail(frames, skb);
|
|
|
}
|
|
|
|
|
|
static ieee80211_rx_result debug_noinline
|
|
@@ -2184,7 +2192,7 @@ ieee80211_rx_h_data(struct ieee80211_rx_data *rx)
|
|
|
}
|
|
|
|
|
|
static ieee80211_rx_result debug_noinline
|
|
|
-ieee80211_rx_h_ctrl(struct ieee80211_rx_data *rx)
|
|
|
+ieee80211_rx_h_ctrl(struct ieee80211_rx_data *rx, struct sk_buff_head *frames)
|
|
|
{
|
|
|
struct sk_buff *skb = rx->skb;
|
|
|
struct ieee80211_bar *bar = (struct ieee80211_bar *)skb->data;
|
|
@@ -2223,7 +2231,7 @@ ieee80211_rx_h_ctrl(struct ieee80211_rx_data *rx)
|
|
|
spin_lock(&tid_agg_rx->reorder_lock);
|
|
|
/* release stored frames up to start of BAR */
|
|
|
ieee80211_release_reorder_frames(rx->sdata, tid_agg_rx,
|
|
|
- start_seq_num);
|
|
|
+ start_seq_num, frames);
|
|
|
spin_unlock(&tid_agg_rx->reorder_lock);
|
|
|
|
|
|
kfree_skb(skb);
|
|
@@ -2808,7 +2816,8 @@ static void ieee80211_rx_handlers_result(struct ieee80211_rx_data *rx,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
+static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx,
|
|
|
+ struct sk_buff_head *frames)
|
|
|
{
|
|
|
ieee80211_rx_result res = RX_DROP_MONITOR;
|
|
|
struct sk_buff *skb;
|
|
@@ -2820,15 +2829,9 @@ static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
goto rxh_next; \
|
|
|
} while (0);
|
|
|
|
|
|
- spin_lock(&rx->local->rx_skb_queue.lock);
|
|
|
- if (rx->local->running_rx_handler)
|
|
|
- goto unlock;
|
|
|
-
|
|
|
- rx->local->running_rx_handler = true;
|
|
|
-
|
|
|
- while ((skb = __skb_dequeue(&rx->local->rx_skb_queue))) {
|
|
|
- spin_unlock(&rx->local->rx_skb_queue.lock);
|
|
|
+ spin_lock_bh(&rx->local->rx_path_lock);
|
|
|
|
|
|
+ while ((skb = __skb_dequeue(frames))) {
|
|
|
/*
|
|
|
* all the other fields are valid across frames
|
|
|
* that belong to an aMPDU since they are on the
|
|
@@ -2849,7 +2852,12 @@ static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
#endif
|
|
|
CALL_RXH(ieee80211_rx_h_amsdu)
|
|
|
CALL_RXH(ieee80211_rx_h_data)
|
|
|
- CALL_RXH(ieee80211_rx_h_ctrl);
|
|
|
+
|
|
|
+ /* special treatment -- needs the queue */
|
|
|
+ res = ieee80211_rx_h_ctrl(rx, frames);
|
|
|
+ if (res != RX_CONTINUE)
|
|
|
+ goto rxh_next;
|
|
|
+
|
|
|
CALL_RXH(ieee80211_rx_h_mgmt_check)
|
|
|
CALL_RXH(ieee80211_rx_h_action)
|
|
|
CALL_RXH(ieee80211_rx_h_userspace_mgmt)
|
|
@@ -2858,20 +2866,20 @@ static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
|
|
|
rxh_next:
|
|
|
ieee80211_rx_handlers_result(rx, res);
|
|
|
- spin_lock(&rx->local->rx_skb_queue.lock);
|
|
|
+
|
|
|
#undef CALL_RXH
|
|
|
}
|
|
|
|
|
|
- rx->local->running_rx_handler = false;
|
|
|
-
|
|
|
- unlock:
|
|
|
- spin_unlock(&rx->local->rx_skb_queue.lock);
|
|
|
+ spin_unlock_bh(&rx->local->rx_path_lock);
|
|
|
}
|
|
|
|
|
|
static void ieee80211_invoke_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
{
|
|
|
+ struct sk_buff_head reorder_release;
|
|
|
ieee80211_rx_result res = RX_DROP_MONITOR;
|
|
|
|
|
|
+ __skb_queue_head_init(&reorder_release);
|
|
|
+
|
|
|
#define CALL_RXH(rxh) \
|
|
|
do { \
|
|
|
res = rxh(rx); \
|
|
@@ -2881,9 +2889,9 @@ static void ieee80211_invoke_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
|
|
|
CALL_RXH(ieee80211_rx_h_check)
|
|
|
|
|
|
- ieee80211_rx_reorder_ampdu(rx);
|
|
|
+ ieee80211_rx_reorder_ampdu(rx, &reorder_release);
|
|
|
|
|
|
- ieee80211_rx_handlers(rx);
|
|
|
+ ieee80211_rx_handlers(rx, &reorder_release);
|
|
|
return;
|
|
|
|
|
|
rxh_next:
|
|
@@ -2898,6 +2906,7 @@ static void ieee80211_invoke_rx_handlers(struct ieee80211_rx_data *rx)
|
|
|
*/
|
|
|
void ieee80211_release_reorder_timeout(struct sta_info *sta, int tid)
|
|
|
{
|
|
|
+ struct sk_buff_head frames;
|
|
|
struct ieee80211_rx_data rx = {
|
|
|
.sta = sta,
|
|
|
.sdata = sta->sdata,
|
|
@@ -2913,11 +2922,13 @@ void ieee80211_release_reorder_timeout(struct sta_info *sta, int tid)
|
|
|
if (!tid_agg_rx)
|
|
|
return;
|
|
|
|
|
|
+ __skb_queue_head_init(&frames);
|
|
|
+
|
|
|
spin_lock(&tid_agg_rx->reorder_lock);
|
|
|
- ieee80211_sta_reorder_release(sta->sdata, tid_agg_rx);
|
|
|
+ ieee80211_sta_reorder_release(sta->sdata, tid_agg_rx, &frames);
|
|
|
spin_unlock(&tid_agg_rx->reorder_lock);
|
|
|
|
|
|
- ieee80211_rx_handlers(&rx);
|
|
|
+ ieee80211_rx_handlers(&rx, &frames);
|
|
|
}
|
|
|
|
|
|
/* main receive path */
|