|
@@ -135,6 +135,9 @@ static struct ath_frame_info *get_frame_info(struct sk_buff *skb)
|
|
|
|
|
|
static void ath_send_bar(struct ath_atx_tid *tid, u16 seqno)
|
|
|
{
|
|
|
+ if (!tid->an->sta)
|
|
|
+ return;
|
|
|
+
|
|
|
ieee80211_send_bar(tid->an->vif, tid->an->sta->addr, tid->tidno,
|
|
|
seqno << IEEE80211_SEQ_SEQ_SHIFT);
|
|
|
}
|
|
@@ -168,6 +171,71 @@ static void ath_txq_skb_done(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
+static struct ath_atx_tid *
|
|
|
+ath_get_skb_tid(struct ath_softc *sc, struct ath_node *an, struct sk_buff *skb)
|
|
|
+{
|
|
|
+ struct ieee80211_hdr *hdr;
|
|
|
+ u8 tidno = 0;
|
|
|
+
|
|
|
+ hdr = (struct ieee80211_hdr *) skb->data;
|
|
|
+ if (ieee80211_is_data_qos(hdr->frame_control))
|
|
|
+ tidno = ieee80211_get_qos_ctl(hdr)[0];
|
|
|
+
|
|
|
+ tidno &= IEEE80211_QOS_CTL_TID_MASK;
|
|
|
+ return ATH_AN_2_TID(an, tidno);
|
|
|
+}
|
|
|
+
|
|
|
+static bool ath_tid_has_buffered(struct ath_atx_tid *tid)
|
|
|
+{
|
|
|
+ return !skb_queue_empty(&tid->buf_q) || !skb_queue_empty(&tid->retry_q);
|
|
|
+}
|
|
|
+
|
|
|
+static struct sk_buff *ath_tid_dequeue(struct ath_atx_tid *tid)
|
|
|
+{
|
|
|
+ struct sk_buff *skb;
|
|
|
+
|
|
|
+ skb = __skb_dequeue(&tid->retry_q);
|
|
|
+ if (!skb)
|
|
|
+ skb = __skb_dequeue(&tid->buf_q);
|
|
|
+
|
|
|
+ return skb;
|
|
|
+}
|
|
|
+
|
|
|
+/*
|
|
|
+ * ath_tx_tid_change_state:
|
|
|
+ * - clears a-mpdu flag of previous session
|
|
|
+ * - force sequence number allocation to fix next BlockAck Window
|
|
|
+ */
|
|
|
+static void
|
|
|
+ath_tx_tid_change_state(struct ath_softc *sc, struct ath_atx_tid *tid)
|
|
|
+{
|
|
|
+ struct ath_txq *txq = tid->ac->txq;
|
|
|
+ struct ieee80211_tx_info *tx_info;
|
|
|
+ struct sk_buff *skb, *tskb;
|
|
|
+ struct ath_buf *bf;
|
|
|
+ struct ath_frame_info *fi;
|
|
|
+
|
|
|
+ skb_queue_walk_safe(&tid->buf_q, skb, tskb) {
|
|
|
+ fi = get_frame_info(skb);
|
|
|
+ bf = fi->bf;
|
|
|
+
|
|
|
+ tx_info = IEEE80211_SKB_CB(skb);
|
|
|
+ tx_info->flags &= ~IEEE80211_TX_CTL_AMPDU;
|
|
|
+
|
|
|
+ if (bf)
|
|
|
+ continue;
|
|
|
+
|
|
|
+ bf = ath_tx_setup_buffer(sc, txq, tid, skb);
|
|
|
+ if (!bf) {
|
|
|
+ __skb_unlink(skb, &tid->buf_q);
|
|
|
+ ath_txq_skb_done(sc, txq, skb);
|
|
|
+ ieee80211_free_txskb(sc->hw, skb);
|
|
|
+ continue;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+}
|
|
|
+
|
|
|
static void ath_tx_flush_tid(struct ath_softc *sc, struct ath_atx_tid *tid)
|
|
|
{
|
|
|
struct ath_txq *txq = tid->ac->txq;
|
|
@@ -182,28 +250,22 @@ static void ath_tx_flush_tid(struct ath_softc *sc, struct ath_atx_tid *tid)
|
|
|
|
|
|
memset(&ts, 0, sizeof(ts));
|
|
|
|
|
|
- while ((skb = __skb_dequeue(&tid->buf_q))) {
|
|
|
+ while ((skb = __skb_dequeue(&tid->retry_q))) {
|
|
|
fi = get_frame_info(skb);
|
|
|
bf = fi->bf;
|
|
|
-
|
|
|
if (!bf) {
|
|
|
- bf = ath_tx_setup_buffer(sc, txq, tid, skb);
|
|
|
- if (!bf) {
|
|
|
- ath_txq_skb_done(sc, txq, skb);
|
|
|
- ieee80211_free_txskb(sc->hw, skb);
|
|
|
- continue;
|
|
|
- }
|
|
|
+ ath_txq_skb_done(sc, txq, skb);
|
|
|
+ ieee80211_free_txskb(sc->hw, skb);
|
|
|
+ continue;
|
|
|
}
|
|
|
|
|
|
- if (fi->retries) {
|
|
|
- list_add_tail(&bf->list, &bf_head);
|
|
|
+ if (fi->baw_tracked) {
|
|
|
ath_tx_update_baw(sc, tid, bf->bf_state.seqno);
|
|
|
- ath_tx_complete_buf(sc, bf, txq, &bf_head, &ts, 0);
|
|
|
sendbar = true;
|
|
|
- } else {
|
|
|
- ath_set_rates(tid->an->vif, tid->an->sta, bf);
|
|
|
- ath_tx_send_normal(sc, txq, NULL, skb);
|
|
|
}
|
|
|
+
|
|
|
+ list_add_tail(&bf->list, &bf_head);
|
|
|
+ ath_tx_complete_buf(sc, bf, txq, &bf_head, &ts, 0);
|
|
|
}
|
|
|
|
|
|
if (sendbar) {
|
|
@@ -232,13 +294,16 @@ static void ath_tx_update_baw(struct ath_softc *sc, struct ath_atx_tid *tid,
|
|
|
}
|
|
|
|
|
|
static void ath_tx_addto_baw(struct ath_softc *sc, struct ath_atx_tid *tid,
|
|
|
- u16 seqno)
|
|
|
+ struct ath_buf *bf)
|
|
|
{
|
|
|
+ struct ath_frame_info *fi = get_frame_info(bf->bf_mpdu);
|
|
|
+ u16 seqno = bf->bf_state.seqno;
|
|
|
int index, cindex;
|
|
|
|
|
|
index = ATH_BA_INDEX(tid->seq_start, seqno);
|
|
|
cindex = (tid->baw_head + index) & (ATH_TID_MAX_BUFS - 1);
|
|
|
__set_bit(cindex, tid->tx_buf);
|
|
|
+ fi->baw_tracked = 1;
|
|
|
|
|
|
if (index >= ((tid->baw_tail - tid->baw_head) &
|
|
|
(ATH_TID_MAX_BUFS - 1))) {
|
|
@@ -266,7 +331,7 @@ static void ath_tid_drain(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
memset(&ts, 0, sizeof(ts));
|
|
|
INIT_LIST_HEAD(&bf_head);
|
|
|
|
|
|
- while ((skb = __skb_dequeue(&tid->buf_q))) {
|
|
|
+ while ((skb = ath_tid_dequeue(tid))) {
|
|
|
fi = get_frame_info(skb);
|
|
|
bf = fi->bf;
|
|
|
|
|
@@ -403,7 +468,6 @@ static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
struct ieee80211_tx_rate rates[4];
|
|
|
struct ath_frame_info *fi;
|
|
|
int nframes;
|
|
|
- u8 tidno;
|
|
|
bool flush = !!(ts->ts_status & ATH9K_TX_FLUSH);
|
|
|
int i, retries;
|
|
|
int bar_index = -1;
|
|
@@ -440,8 +504,7 @@ static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
}
|
|
|
|
|
|
an = (struct ath_node *)sta->drv_priv;
|
|
|
- tidno = ieee80211_get_qos_ctl(hdr)[0] & IEEE80211_QOS_CTL_TID_MASK;
|
|
|
- tid = ATH_AN_2_TID(an, tidno);
|
|
|
+ tid = ath_get_skb_tid(sc, an, skb);
|
|
|
seq_first = tid->seq_start;
|
|
|
isba = ts->ts_flags & ATH9K_TX_BA;
|
|
|
|
|
@@ -453,7 +516,7 @@ static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
* Only BlockAcks have a TID and therefore normal Acks cannot be
|
|
|
* checked
|
|
|
*/
|
|
|
- if (isba && tidno != ts->tid)
|
|
|
+ if (isba && tid->tidno != ts->tid)
|
|
|
txok = false;
|
|
|
|
|
|
isaggr = bf_isaggr(bf);
|
|
@@ -489,7 +552,8 @@ static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
tx_info = IEEE80211_SKB_CB(skb);
|
|
|
fi = get_frame_info(skb);
|
|
|
|
|
|
- if (!BAW_WITHIN(tid->seq_start, tid->baw_size, seqno)) {
|
|
|
+ if (!BAW_WITHIN(tid->seq_start, tid->baw_size, seqno) ||
|
|
|
+ !tid->active) {
|
|
|
/*
|
|
|
* Outside of the current BlockAck window,
|
|
|
* maybe part of a previous session
|
|
@@ -583,7 +647,7 @@ static void ath_tx_complete_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
if (an->sleeping)
|
|
|
ieee80211_sta_set_buffered(sta, tid->tidno, true);
|
|
|
|
|
|
- skb_queue_splice(&bf_pending, &tid->buf_q);
|
|
|
+ skb_queue_splice_tail(&bf_pending, &tid->retry_q);
|
|
|
if (!an->sleeping) {
|
|
|
ath_tx_queue_tid(txq, tid);
|
|
|
|
|
@@ -641,7 +705,7 @@ static void ath_tx_process_buffer(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
} else
|
|
|
ath_tx_complete_aggr(sc, txq, bf, bf_head, ts, txok);
|
|
|
|
|
|
- if ((sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_HT) && !flush)
|
|
|
+ if (!flush)
|
|
|
ath_txq_schedule(sc, txq);
|
|
|
}
|
|
|
|
|
@@ -815,15 +879,20 @@ static int ath_compute_num_delims(struct ath_softc *sc, struct ath_atx_tid *tid,
|
|
|
|
|
|
static struct ath_buf *
|
|
|
ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
- struct ath_atx_tid *tid)
|
|
|
+ struct ath_atx_tid *tid, struct sk_buff_head **q)
|
|
|
{
|
|
|
+ struct ieee80211_tx_info *tx_info;
|
|
|
struct ath_frame_info *fi;
|
|
|
struct sk_buff *skb;
|
|
|
struct ath_buf *bf;
|
|
|
u16 seqno;
|
|
|
|
|
|
while (1) {
|
|
|
- skb = skb_peek(&tid->buf_q);
|
|
|
+ *q = &tid->retry_q;
|
|
|
+ if (skb_queue_empty(*q))
|
|
|
+ *q = &tid->buf_q;
|
|
|
+
|
|
|
+ skb = skb_peek(*q);
|
|
|
if (!skb)
|
|
|
break;
|
|
|
|
|
@@ -833,12 +902,22 @@ ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
bf = ath_tx_setup_buffer(sc, txq, tid, skb);
|
|
|
|
|
|
if (!bf) {
|
|
|
- __skb_unlink(skb, &tid->buf_q);
|
|
|
+ __skb_unlink(skb, *q);
|
|
|
ath_txq_skb_done(sc, txq, skb);
|
|
|
ieee80211_free_txskb(sc->hw, skb);
|
|
|
continue;
|
|
|
}
|
|
|
|
|
|
+ bf->bf_next = NULL;
|
|
|
+ bf->bf_lastbf = bf;
|
|
|
+
|
|
|
+ tx_info = IEEE80211_SKB_CB(skb);
|
|
|
+ tx_info->flags &= ~IEEE80211_TX_CTL_CLEAR_PS_FILT;
|
|
|
+ if (!(tx_info->flags & IEEE80211_TX_CTL_AMPDU)) {
|
|
|
+ bf->bf_state.bf_type = 0;
|
|
|
+ return bf;
|
|
|
+ }
|
|
|
+
|
|
|
bf->bf_state.bf_type = BUF_AMPDU | BUF_AGGR;
|
|
|
seqno = bf->bf_state.seqno;
|
|
|
|
|
@@ -852,73 +931,52 @@ ath_tx_get_tid_subframe(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
|
|
|
INIT_LIST_HEAD(&bf_head);
|
|
|
list_add(&bf->list, &bf_head);
|
|
|
- __skb_unlink(skb, &tid->buf_q);
|
|
|
+ __skb_unlink(skb, *q);
|
|
|
ath_tx_update_baw(sc, tid, seqno);
|
|
|
ath_tx_complete_buf(sc, bf, txq, &bf_head, &ts, 0);
|
|
|
continue;
|
|
|
}
|
|
|
|
|
|
- bf->bf_next = NULL;
|
|
|
- bf->bf_lastbf = bf;
|
|
|
return bf;
|
|
|
}
|
|
|
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
|
-static enum ATH_AGGR_STATUS ath_tx_form_aggr(struct ath_softc *sc,
|
|
|
- struct ath_txq *txq,
|
|
|
- struct ath_atx_tid *tid,
|
|
|
- struct list_head *bf_q,
|
|
|
- int *aggr_len)
|
|
|
+static bool
|
|
|
+ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
+ struct ath_atx_tid *tid, struct list_head *bf_q,
|
|
|
+ struct ath_buf *bf_first, struct sk_buff_head *tid_q,
|
|
|
+ int *aggr_len)
|
|
|
{
|
|
|
#define PADBYTES(_len) ((4 - ((_len) % 4)) % 4)
|
|
|
- struct ath_buf *bf, *bf_first = NULL, *bf_prev = NULL;
|
|
|
- int rl = 0, nframes = 0, ndelim, prev_al = 0;
|
|
|
+ struct ath_buf *bf = bf_first, *bf_prev = NULL;
|
|
|
+ int nframes = 0, ndelim;
|
|
|
u16 aggr_limit = 0, al = 0, bpad = 0,
|
|
|
- al_delta, h_baw = tid->baw_size / 2;
|
|
|
- enum ATH_AGGR_STATUS status = ATH_AGGR_DONE;
|
|
|
+ al_delta, h_baw = tid->baw_size / 2;
|
|
|
struct ieee80211_tx_info *tx_info;
|
|
|
struct ath_frame_info *fi;
|
|
|
struct sk_buff *skb;
|
|
|
+ bool closed = false;
|
|
|
|
|
|
- do {
|
|
|
- bf = ath_tx_get_tid_subframe(sc, txq, tid);
|
|
|
- if (!bf) {
|
|
|
- status = ATH_AGGR_BAW_CLOSED;
|
|
|
- break;
|
|
|
- }
|
|
|
+ bf = bf_first;
|
|
|
+ aggr_limit = ath_lookup_rate(sc, bf, tid);
|
|
|
|
|
|
+ do {
|
|
|
skb = bf->bf_mpdu;
|
|
|
fi = get_frame_info(skb);
|
|
|
|
|
|
- if (!bf_first)
|
|
|
- bf_first = bf;
|
|
|
-
|
|
|
- if (!rl) {
|
|
|
- ath_set_rates(tid->an->vif, tid->an->sta, bf);
|
|
|
- aggr_limit = ath_lookup_rate(sc, bf, tid);
|
|
|
- rl = 1;
|
|
|
- }
|
|
|
-
|
|
|
/* do not exceed aggregation limit */
|
|
|
al_delta = ATH_AGGR_DELIM_SZ + fi->framelen;
|
|
|
+ if (nframes) {
|
|
|
+ if (aggr_limit < al + bpad + al_delta ||
|
|
|
+ ath_lookup_legacy(bf) || nframes >= h_baw)
|
|
|
+ break;
|
|
|
|
|
|
- if (nframes &&
|
|
|
- ((aggr_limit < (al + bpad + al_delta + prev_al)) ||
|
|
|
- ath_lookup_legacy(bf))) {
|
|
|
- status = ATH_AGGR_LIMITED;
|
|
|
- break;
|
|
|
- }
|
|
|
-
|
|
|
- tx_info = IEEE80211_SKB_CB(bf->bf_mpdu);
|
|
|
- if (nframes && (tx_info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE))
|
|
|
- break;
|
|
|
-
|
|
|
- /* do not exceed subframe limit */
|
|
|
- if (nframes >= min((int)h_baw, ATH_AMPDU_SUBFRAME_DEFAULT)) {
|
|
|
- status = ATH_AGGR_LIMITED;
|
|
|
- break;
|
|
|
+ tx_info = IEEE80211_SKB_CB(bf->bf_mpdu);
|
|
|
+ if ((tx_info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE) ||
|
|
|
+ !(tx_info->flags & IEEE80211_TX_CTL_AMPDU))
|
|
|
+ break;
|
|
|
}
|
|
|
|
|
|
/* add padding for previous frame to aggregation length */
|
|
@@ -936,22 +994,37 @@ static enum ATH_AGGR_STATUS ath_tx_form_aggr(struct ath_softc *sc,
|
|
|
bf->bf_next = NULL;
|
|
|
|
|
|
/* link buffers of this frame to the aggregate */
|
|
|
- if (!fi->retries)
|
|
|
- ath_tx_addto_baw(sc, tid, bf->bf_state.seqno);
|
|
|
+ if (!fi->baw_tracked)
|
|
|
+ ath_tx_addto_baw(sc, tid, bf);
|
|
|
bf->bf_state.ndelim = ndelim;
|
|
|
|
|
|
- __skb_unlink(skb, &tid->buf_q);
|
|
|
+ __skb_unlink(skb, tid_q);
|
|
|
list_add_tail(&bf->list, bf_q);
|
|
|
if (bf_prev)
|
|
|
bf_prev->bf_next = bf;
|
|
|
|
|
|
bf_prev = bf;
|
|
|
|
|
|
- } while (!skb_queue_empty(&tid->buf_q));
|
|
|
+ bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q);
|
|
|
+ if (!bf) {
|
|
|
+ closed = true;
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ } while (ath_tid_has_buffered(tid));
|
|
|
+
|
|
|
+ bf = bf_first;
|
|
|
+ bf->bf_lastbf = bf_prev;
|
|
|
+
|
|
|
+ if (bf == bf_prev) {
|
|
|
+ al = get_frame_info(bf->bf_mpdu)->framelen;
|
|
|
+ bf->bf_state.bf_type = BUF_AMPDU;
|
|
|
+ } else {
|
|
|
+ TX_STAT_INC(txq->axq_qnum, a_aggr);
|
|
|
+ }
|
|
|
|
|
|
*aggr_len = al;
|
|
|
|
|
|
- return status;
|
|
|
+ return closed;
|
|
|
#undef PADBYTES
|
|
|
}
|
|
|
|
|
@@ -1023,7 +1096,7 @@ void ath_update_max_aggr_framelen(struct ath_softc *sc, int queue, int txop)
|
|
|
}
|
|
|
|
|
|
static void ath_buf_set_rate(struct ath_softc *sc, struct ath_buf *bf,
|
|
|
- struct ath_tx_info *info, int len)
|
|
|
+ struct ath_tx_info *info, int len, bool rts)
|
|
|
{
|
|
|
struct ath_hw *ah = sc->sc_ah;
|
|
|
struct sk_buff *skb;
|
|
@@ -1032,6 +1105,7 @@ static void ath_buf_set_rate(struct ath_softc *sc, struct ath_buf *bf,
|
|
|
const struct ieee80211_rate *rate;
|
|
|
struct ieee80211_hdr *hdr;
|
|
|
struct ath_frame_info *fi = get_frame_info(bf->bf_mpdu);
|
|
|
+ u32 rts_thresh = sc->hw->wiphy->rts_threshold;
|
|
|
int i;
|
|
|
u8 rix = 0;
|
|
|
|
|
@@ -1054,7 +1128,17 @@ static void ath_buf_set_rate(struct ath_softc *sc, struct ath_buf *bf,
|
|
|
rix = rates[i].idx;
|
|
|
info->rates[i].Tries = rates[i].count;
|
|
|
|
|
|
- if (rates[i].flags & IEEE80211_TX_RC_USE_RTS_CTS) {
|
|
|
+ /*
|
|
|
+ * Handle RTS threshold for unaggregated HT frames.
|
|
|
+ */
|
|
|
+ if (bf_isampdu(bf) && !bf_isaggr(bf) &&
|
|
|
+ (rates[i].flags & IEEE80211_TX_RC_MCS) &&
|
|
|
+ unlikely(rts_thresh != (u32) -1)) {
|
|
|
+ if (!rts_thresh || (len > rts_thresh))
|
|
|
+ rts = true;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (rts || rates[i].flags & IEEE80211_TX_RC_USE_RTS_CTS) {
|
|
|
info->rates[i].RateFlags |= ATH9K_RATESERIES_RTS_CTS;
|
|
|
info->flags |= ATH9K_TXDESC_RTSENA;
|
|
|
} else if (rates[i].flags & IEEE80211_TX_RC_USE_CTS_PROTECT) {
|
|
@@ -1147,6 +1231,8 @@ static void ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf,
|
|
|
struct ath_hw *ah = sc->sc_ah;
|
|
|
struct ath_buf *bf_first = NULL;
|
|
|
struct ath_tx_info info;
|
|
|
+ u32 rts_thresh = sc->hw->wiphy->rts_threshold;
|
|
|
+ bool rts = false;
|
|
|
|
|
|
memset(&info, 0, sizeof(info));
|
|
|
info.is_first = true;
|
|
@@ -1183,7 +1269,22 @@ static void ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf,
|
|
|
info.flags |= (u32) bf->bf_state.bfs_paprd <<
|
|
|
ATH9K_TXDESC_PAPRD_S;
|
|
|
|
|
|
- ath_buf_set_rate(sc, bf, &info, len);
|
|
|
+ /*
|
|
|
+ * mac80211 doesn't handle RTS threshold for HT because
|
|
|
+ * the decision has to be taken based on AMPDU length
|
|
|
+ * and aggregation is done entirely inside ath9k.
|
|
|
+ * Set the RTS/CTS flag for the first subframe based
|
|
|
+ * on the threshold.
|
|
|
+ */
|
|
|
+ if (aggr && (bf == bf_first) &&
|
|
|
+ unlikely(rts_thresh != (u32) -1)) {
|
|
|
+ /*
|
|
|
+ * "len" is the size of the entire AMPDU.
|
|
|
+ */
|
|
|
+ if (!rts_thresh || (len > rts_thresh))
|
|
|
+ rts = true;
|
|
|
+ }
|
|
|
+ ath_buf_set_rate(sc, bf, &info, len, rts);
|
|
|
}
|
|
|
|
|
|
info.buf_addr[0] = bf->bf_buf_addr;
|
|
@@ -1212,53 +1313,86 @@ static void ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static void ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
- struct ath_atx_tid *tid)
|
|
|
+static void
|
|
|
+ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
+ struct ath_atx_tid *tid, struct list_head *bf_q,
|
|
|
+ struct ath_buf *bf_first, struct sk_buff_head *tid_q)
|
|
|
{
|
|
|
- struct ath_buf *bf;
|
|
|
- enum ATH_AGGR_STATUS status;
|
|
|
- struct ieee80211_tx_info *tx_info;
|
|
|
- struct list_head bf_q;
|
|
|
- int aggr_len;
|
|
|
+ struct ath_buf *bf = bf_first, *bf_prev = NULL;
|
|
|
+ struct sk_buff *skb;
|
|
|
+ int nframes = 0;
|
|
|
|
|
|
do {
|
|
|
- if (skb_queue_empty(&tid->buf_q))
|
|
|
- return;
|
|
|
+ struct ieee80211_tx_info *tx_info;
|
|
|
+ skb = bf->bf_mpdu;
|
|
|
|
|
|
- INIT_LIST_HEAD(&bf_q);
|
|
|
+ nframes++;
|
|
|
+ __skb_unlink(skb, tid_q);
|
|
|
+ list_add_tail(&bf->list, bf_q);
|
|
|
+ if (bf_prev)
|
|
|
+ bf_prev->bf_next = bf;
|
|
|
+ bf_prev = bf;
|
|
|
|
|
|
- status = ath_tx_form_aggr(sc, txq, tid, &bf_q, &aggr_len);
|
|
|
+ if (nframes >= 2)
|
|
|
+ break;
|
|
|
|
|
|
- /*
|
|
|
- * no frames picked up to be aggregated;
|
|
|
- * block-ack window is not open.
|
|
|
- */
|
|
|
- if (list_empty(&bf_q))
|
|
|
+ bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q);
|
|
|
+ if (!bf)
|
|
|
break;
|
|
|
|
|
|
- bf = list_first_entry(&bf_q, struct ath_buf, list);
|
|
|
- bf->bf_lastbf = list_entry(bf_q.prev, struct ath_buf, list);
|
|
|
tx_info = IEEE80211_SKB_CB(bf->bf_mpdu);
|
|
|
+ if (tx_info->flags & IEEE80211_TX_CTL_AMPDU)
|
|
|
+ break;
|
|
|
|
|
|
- if (tid->ac->clear_ps_filter) {
|
|
|
- tid->ac->clear_ps_filter = false;
|
|
|
- tx_info->flags |= IEEE80211_TX_CTL_CLEAR_PS_FILT;
|
|
|
- } else {
|
|
|
- tx_info->flags &= ~IEEE80211_TX_CTL_CLEAR_PS_FILT;
|
|
|
- }
|
|
|
+ ath_set_rates(tid->an->vif, tid->an->sta, bf);
|
|
|
+ } while (1);
|
|
|
+}
|
|
|
|
|
|
- /* if only one frame, send as non-aggregate */
|
|
|
- if (bf == bf->bf_lastbf) {
|
|
|
- aggr_len = get_frame_info(bf->bf_mpdu)->framelen;
|
|
|
- bf->bf_state.bf_type = BUF_AMPDU;
|
|
|
- } else {
|
|
|
- TX_STAT_INC(txq->axq_qnum, a_aggr);
|
|
|
- }
|
|
|
+static bool ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
+ struct ath_atx_tid *tid, bool *stop)
|
|
|
+{
|
|
|
+ struct ath_buf *bf;
|
|
|
+ struct ieee80211_tx_info *tx_info;
|
|
|
+ struct sk_buff_head *tid_q;
|
|
|
+ struct list_head bf_q;
|
|
|
+ int aggr_len = 0;
|
|
|
+ bool aggr, last = true;
|
|
|
+
|
|
|
+ if (!ath_tid_has_buffered(tid))
|
|
|
+ return false;
|
|
|
+
|
|
|
+ INIT_LIST_HEAD(&bf_q);
|
|
|
+
|
|
|
+ bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q);
|
|
|
+ if (!bf)
|
|
|
+ return false;
|
|
|
+
|
|
|
+ tx_info = IEEE80211_SKB_CB(bf->bf_mpdu);
|
|
|
+ aggr = !!(tx_info->flags & IEEE80211_TX_CTL_AMPDU);
|
|
|
+ if ((aggr && txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH) ||
|
|
|
+ (!aggr && txq->axq_depth >= ATH_NON_AGGR_MIN_QDEPTH)) {
|
|
|
+ *stop = true;
|
|
|
+ return false;
|
|
|
+ }
|
|
|
|
|
|
- ath_tx_fill_desc(sc, bf, txq, aggr_len);
|
|
|
- ath_tx_txqaddbuf(sc, txq, &bf_q, false);
|
|
|
- } while (txq->axq_ampdu_depth < ATH_AGGR_MIN_QDEPTH &&
|
|
|
- status != ATH_AGGR_BAW_CLOSED);
|
|
|
+ ath_set_rates(tid->an->vif, tid->an->sta, bf);
|
|
|
+ if (aggr)
|
|
|
+ last = ath_tx_form_aggr(sc, txq, tid, &bf_q, bf,
|
|
|
+ tid_q, &aggr_len);
|
|
|
+ else
|
|
|
+ ath_tx_form_burst(sc, txq, tid, &bf_q, bf, tid_q);
|
|
|
+
|
|
|
+ if (list_empty(&bf_q))
|
|
|
+ return false;
|
|
|
+
|
|
|
+ if (tid->ac->clear_ps_filter || tid->an->no_ps_filter) {
|
|
|
+ tid->ac->clear_ps_filter = false;
|
|
|
+ tx_info->flags |= IEEE80211_TX_CTL_CLEAR_PS_FILT;
|
|
|
+ }
|
|
|
+
|
|
|
+ ath_tx_fill_desc(sc, bf, txq, aggr_len);
|
|
|
+ ath_tx_txqaddbuf(sc, txq, &bf_q, false);
|
|
|
+ return true;
|
|
|
}
|
|
|
|
|
|
int ath_tx_aggr_start(struct ath_softc *sc, struct ieee80211_sta *sta,
|
|
@@ -1282,6 +1416,9 @@ int ath_tx_aggr_start(struct ath_softc *sc, struct ieee80211_sta *sta,
|
|
|
an->mpdudensity = density;
|
|
|
}
|
|
|
|
|
|
+ /* force sequence number allocation for pending frames */
|
|
|
+ ath_tx_tid_change_state(sc, txtid);
|
|
|
+
|
|
|
txtid->active = true;
|
|
|
txtid->paused = true;
|
|
|
*ssn = txtid->seq_start = txtid->seq_next;
|
|
@@ -1301,8 +1438,9 @@ void ath_tx_aggr_stop(struct ath_softc *sc, struct ieee80211_sta *sta, u16 tid)
|
|
|
|
|
|
ath_txq_lock(sc, txq);
|
|
|
txtid->active = false;
|
|
|
- txtid->paused = true;
|
|
|
+ txtid->paused = false;
|
|
|
ath_tx_flush_tid(sc, txtid);
|
|
|
+ ath_tx_tid_change_state(sc, txtid);
|
|
|
ath_txq_unlock_complete(sc, txq);
|
|
|
}
|
|
|
|
|
@@ -1326,7 +1464,7 @@ void ath_tx_aggr_sleep(struct ieee80211_sta *sta, struct ath_softc *sc,
|
|
|
|
|
|
ath_txq_lock(sc, txq);
|
|
|
|
|
|
- buffered = !skb_queue_empty(&tid->buf_q);
|
|
|
+ buffered = ath_tid_has_buffered(tid);
|
|
|
|
|
|
tid->sched = false;
|
|
|
list_del(&tid->list);
|
|
@@ -1358,7 +1496,7 @@ void ath_tx_aggr_wakeup(struct ath_softc *sc, struct ath_node *an)
|
|
|
ath_txq_lock(sc, txq);
|
|
|
ac->clear_ps_filter = true;
|
|
|
|
|
|
- if (!skb_queue_empty(&tid->buf_q) && !tid->paused) {
|
|
|
+ if (!tid->paused && ath_tid_has_buffered(tid)) {
|
|
|
ath_tx_queue_tid(txq, tid);
|
|
|
ath_txq_schedule(sc, txq);
|
|
|
}
|
|
@@ -1383,7 +1521,7 @@ void ath_tx_aggr_resume(struct ath_softc *sc, struct ieee80211_sta *sta,
|
|
|
tid->baw_size = IEEE80211_MIN_AMPDU_BUF << sta->ht_cap.ampdu_factor;
|
|
|
tid->paused = false;
|
|
|
|
|
|
- if (!skb_queue_empty(&tid->buf_q)) {
|
|
|
+ if (ath_tid_has_buffered(tid)) {
|
|
|
ath_tx_queue_tid(txq, tid);
|
|
|
ath_txq_schedule(sc, txq);
|
|
|
}
|
|
@@ -1403,6 +1541,7 @@ void ath9k_release_buffered_frames(struct ieee80211_hw *hw,
|
|
|
struct ieee80211_tx_info *info;
|
|
|
struct list_head bf_q;
|
|
|
struct ath_buf *bf_tail = NULL, *bf;
|
|
|
+ struct sk_buff_head *tid_q;
|
|
|
int sent = 0;
|
|
|
int i;
|
|
|
|
|
@@ -1418,15 +1557,15 @@ void ath9k_release_buffered_frames(struct ieee80211_hw *hw,
|
|
|
continue;
|
|
|
|
|
|
ath_txq_lock(sc, tid->ac->txq);
|
|
|
- while (!skb_queue_empty(&tid->buf_q) && nframes > 0) {
|
|
|
- bf = ath_tx_get_tid_subframe(sc, sc->tx.uapsdq, tid);
|
|
|
+ while (nframes > 0) {
|
|
|
+ bf = ath_tx_get_tid_subframe(sc, sc->tx.uapsdq, tid, &tid_q);
|
|
|
if (!bf)
|
|
|
break;
|
|
|
|
|
|
- __skb_unlink(bf->bf_mpdu, &tid->buf_q);
|
|
|
+ __skb_unlink(bf->bf_mpdu, tid_q);
|
|
|
list_add_tail(&bf->list, &bf_q);
|
|
|
ath_set_rates(tid->an->vif, tid->an->sta, bf);
|
|
|
- ath_tx_addto_baw(sc, tid, bf->bf_state.seqno);
|
|
|
+ ath_tx_addto_baw(sc, tid, bf);
|
|
|
bf->bf_state.bf_type &= ~BUF_AGGR;
|
|
|
if (bf_tail)
|
|
|
bf_tail->bf_next = bf;
|
|
@@ -1436,7 +1575,7 @@ void ath9k_release_buffered_frames(struct ieee80211_hw *hw,
|
|
|
sent++;
|
|
|
TX_STAT_INC(txq->axq_qnum, a_queued_hw);
|
|
|
|
|
|
- if (skb_queue_empty(&tid->buf_q))
|
|
|
+ if (an->sta && !ath_tid_has_buffered(tid))
|
|
|
ieee80211_sta_set_buffered(an->sta, i, false);
|
|
|
}
|
|
|
ath_txq_unlock_complete(sc, tid->ac->txq);
|
|
@@ -1689,25 +1828,27 @@ void ath_tx_cleanupq(struct ath_softc *sc, struct ath_txq *txq)
|
|
|
*/
|
|
|
void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq)
|
|
|
{
|
|
|
- struct ath_atx_ac *ac, *ac_tmp, *last_ac;
|
|
|
+ struct ath_atx_ac *ac, *last_ac;
|
|
|
struct ath_atx_tid *tid, *last_tid;
|
|
|
+ bool sent = false;
|
|
|
|
|
|
if (test_bit(SC_OP_HW_RESET, &sc->sc_flags) ||
|
|
|
- list_empty(&txq->axq_acq) ||
|
|
|
- txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH)
|
|
|
+ list_empty(&txq->axq_acq))
|
|
|
return;
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
|
|
- ac = list_first_entry(&txq->axq_acq, struct ath_atx_ac, list);
|
|
|
last_ac = list_entry(txq->axq_acq.prev, struct ath_atx_ac, list);
|
|
|
+ while (!list_empty(&txq->axq_acq)) {
|
|
|
+ bool stop = false;
|
|
|
|
|
|
- list_for_each_entry_safe(ac, ac_tmp, &txq->axq_acq, list) {
|
|
|
+ ac = list_first_entry(&txq->axq_acq, struct ath_atx_ac, list);
|
|
|
last_tid = list_entry(ac->tid_q.prev, struct ath_atx_tid, list);
|
|
|
list_del(&ac->list);
|
|
|
ac->sched = false;
|
|
|
|
|
|
while (!list_empty(&ac->tid_q)) {
|
|
|
+
|
|
|
tid = list_first_entry(&ac->tid_q, struct ath_atx_tid,
|
|
|
list);
|
|
|
list_del(&tid->list);
|
|
@@ -1716,17 +1857,17 @@ void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq)
|
|
|
if (tid->paused)
|
|
|
continue;
|
|
|
|
|
|
- ath_tx_sched_aggr(sc, txq, tid);
|
|
|
+ if (ath_tx_sched_aggr(sc, txq, tid, &stop))
|
|
|
+ sent = true;
|
|
|
|
|
|
/*
|
|
|
* add tid to round-robin queue if more frames
|
|
|
* are pending for the tid
|
|
|
*/
|
|
|
- if (!skb_queue_empty(&tid->buf_q))
|
|
|
+ if (ath_tid_has_buffered(tid))
|
|
|
ath_tx_queue_tid(txq, tid);
|
|
|
|
|
|
- if (tid == last_tid ||
|
|
|
- txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH)
|
|
|
+ if (stop || tid == last_tid)
|
|
|
break;
|
|
|
}
|
|
|
|
|
@@ -1735,9 +1876,17 @@ void ath_txq_schedule(struct ath_softc *sc, struct ath_txq *txq)
|
|
|
list_add_tail(&ac->list, &txq->axq_acq);
|
|
|
}
|
|
|
|
|
|
- if (ac == last_ac ||
|
|
|
- txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH)
|
|
|
+ if (stop)
|
|
|
break;
|
|
|
+
|
|
|
+ if (ac == last_ac) {
|
|
|
+ if (!sent)
|
|
|
+ break;
|
|
|
+
|
|
|
+ sent = false;
|
|
|
+ last_ac = list_entry(txq->axq_acq.prev,
|
|
|
+ struct ath_atx_ac, list);
|
|
|
+ }
|
|
|
}
|
|
|
|
|
|
rcu_read_unlock();
|
|
@@ -1816,58 +1965,6 @@ static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static void ath_tx_send_ampdu(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
- struct ath_atx_tid *tid, struct sk_buff *skb,
|
|
|
- struct ath_tx_control *txctl)
|
|
|
-{
|
|
|
- struct ath_frame_info *fi = get_frame_info(skb);
|
|
|
- struct list_head bf_head;
|
|
|
- struct ath_buf *bf;
|
|
|
-
|
|
|
- /*
|
|
|
- * Do not queue to h/w when any of the following conditions is true:
|
|
|
- * - there are pending frames in software queue
|
|
|
- * - the TID is currently paused for ADDBA/BAR request
|
|
|
- * - seqno is not within block-ack window
|
|
|
- * - h/w queue depth exceeds low water mark
|
|
|
- */
|
|
|
- if ((!skb_queue_empty(&tid->buf_q) || tid->paused ||
|
|
|
- !BAW_WITHIN(tid->seq_start, tid->baw_size, tid->seq_next) ||
|
|
|
- txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH) &&
|
|
|
- txq != sc->tx.uapsdq) {
|
|
|
- /*
|
|
|
- * Add this frame to software queue for scheduling later
|
|
|
- * for aggregation.
|
|
|
- */
|
|
|
- TX_STAT_INC(txq->axq_qnum, a_queued_sw);
|
|
|
- __skb_queue_tail(&tid->buf_q, skb);
|
|
|
- if (!txctl->an || !txctl->an->sleeping)
|
|
|
- ath_tx_queue_tid(txq, tid);
|
|
|
- return;
|
|
|
- }
|
|
|
-
|
|
|
- bf = ath_tx_setup_buffer(sc, txq, tid, skb);
|
|
|
- if (!bf) {
|
|
|
- ath_txq_skb_done(sc, txq, skb);
|
|
|
- ieee80211_free_txskb(sc->hw, skb);
|
|
|
- return;
|
|
|
- }
|
|
|
-
|
|
|
- ath_set_rates(tid->an->vif, tid->an->sta, bf);
|
|
|
- bf->bf_state.bf_type = BUF_AMPDU;
|
|
|
- INIT_LIST_HEAD(&bf_head);
|
|
|
- list_add(&bf->list, &bf_head);
|
|
|
-
|
|
|
- /* Add sub-frame to BAW */
|
|
|
- ath_tx_addto_baw(sc, tid, bf->bf_state.seqno);
|
|
|
-
|
|
|
- /* Queue to h/w without aggregation */
|
|
|
- TX_STAT_INC(txq->axq_qnum, a_queued_hw);
|
|
|
- bf->bf_lastbf = bf;
|
|
|
- ath_tx_fill_desc(sc, bf, txq, fi->framelen);
|
|
|
- ath_tx_txqaddbuf(sc, txq, &bf_head, false);
|
|
|
-}
|
|
|
-
|
|
|
static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
|
|
|
struct ath_atx_tid *tid, struct sk_buff *skb)
|
|
|
{
|
|
@@ -2010,6 +2107,7 @@ static int ath_tx_prepare(struct ieee80211_hw *hw, struct sk_buff *skb,
|
|
|
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
|
|
|
struct ieee80211_sta *sta = txctl->sta;
|
|
|
struct ieee80211_vif *vif = info->control.vif;
|
|
|
+ struct ath_vif *avp;
|
|
|
struct ath_softc *sc = hw->priv;
|
|
|
int frmlen = skb->len + FCS_LEN;
|
|
|
int padpos, padsize;
|
|
@@ -2017,6 +2115,10 @@ static int ath_tx_prepare(struct ieee80211_hw *hw, struct sk_buff *skb,
|
|
|
/* NOTE: sta can be NULL according to net/mac80211.h */
|
|
|
if (sta)
|
|
|
txctl->an = (struct ath_node *)sta->drv_priv;
|
|
|
+ else if (vif && ieee80211_is_data(hdr->frame_control)) {
|
|
|
+ avp = (void *)vif->drv_priv;
|
|
|
+ txctl->an = &avp->mcast_node;
|
|
|
+ }
|
|
|
|
|
|
if (info->control.hw_key)
|
|
|
frmlen += info->control.hw_key->icv_len;
|
|
@@ -2066,7 +2168,6 @@ int ath_tx_start(struct ieee80211_hw *hw, struct sk_buff *skb,
|
|
|
struct ath_txq *txq = txctl->txq;
|
|
|
struct ath_atx_tid *tid = NULL;
|
|
|
struct ath_buf *bf;
|
|
|
- u8 tidno;
|
|
|
int q;
|
|
|
int ret;
|
|
|
|
|
@@ -2094,22 +2195,25 @@ int ath_tx_start(struct ieee80211_hw *hw, struct sk_buff *skb,
|
|
|
ath_txq_unlock(sc, txq);
|
|
|
txq = sc->tx.uapsdq;
|
|
|
ath_txq_lock(sc, txq);
|
|
|
- }
|
|
|
-
|
|
|
- if (txctl->an && ieee80211_is_data_qos(hdr->frame_control)) {
|
|
|
- tidno = ieee80211_get_qos_ctl(hdr)[0] &
|
|
|
- IEEE80211_QOS_CTL_TID_MASK;
|
|
|
- tid = ATH_AN_2_TID(txctl->an, tidno);
|
|
|
+ } else if (txctl->an &&
|
|
|
+ ieee80211_is_data_present(hdr->frame_control)) {
|
|
|
+ tid = ath_get_skb_tid(sc, txctl->an, skb);
|
|
|
|
|
|
WARN_ON(tid->ac->txq != txctl->txq);
|
|
|
- }
|
|
|
|
|
|
- if ((info->flags & IEEE80211_TX_CTL_AMPDU) && tid) {
|
|
|
+ if (info->flags & IEEE80211_TX_CTL_CLEAR_PS_FILT)
|
|
|
+ tid->ac->clear_ps_filter = true;
|
|
|
+
|
|
|
/*
|
|
|
- * Try aggregation if it's a unicast data frame
|
|
|
- * and the destination is HT capable.
|
|
|
+ * Add this frame to software queue for scheduling later
|
|
|
+ * for aggregation.
|
|
|
*/
|
|
|
- ath_tx_send_ampdu(sc, txq, tid, skb, txctl);
|
|
|
+ TX_STAT_INC(txq->axq_qnum, a_queued_sw);
|
|
|
+ __skb_queue_tail(&tid->buf_q, skb);
|
|
|
+ if (!txctl->an->sleeping)
|
|
|
+ ath_tx_queue_tid(txq, tid);
|
|
|
+
|
|
|
+ ath_txq_schedule(sc, txq);
|
|
|
goto out;
|
|
|
}
|
|
|
|
|
@@ -2168,7 +2272,7 @@ void ath_tx_cabq(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
|
|
|
|
|
|
bf->bf_lastbf = bf;
|
|
|
ath_set_rates(vif, NULL, bf);
|
|
|
- ath_buf_set_rate(sc, bf, &info, fi->framelen);
|
|
|
+ ath_buf_set_rate(sc, bf, &info, fi->framelen, false);
|
|
|
duration += info.rates[0].PktDuration;
|
|
|
if (bf_tail)
|
|
|
bf_tail->bf_next = bf;
|
|
@@ -2372,8 +2476,7 @@ static void ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq)
|
|
|
|
|
|
if (list_empty(&txq->axq_q)) {
|
|
|
txq->axq_link = NULL;
|
|
|
- if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_HT)
|
|
|
- ath_txq_schedule(sc, txq);
|
|
|
+ ath_txq_schedule(sc, txq);
|
|
|
break;
|
|
|
}
|
|
|
bf = list_first_entry(&txq->axq_q, struct ath_buf, list);
|
|
@@ -2595,6 +2698,7 @@ void ath_tx_node_init(struct ath_softc *sc, struct ath_node *an)
|
|
|
tid->paused = false;
|
|
|
tid->active = false;
|
|
|
__skb_queue_head_init(&tid->buf_q);
|
|
|
+ __skb_queue_head_init(&tid->retry_q);
|
|
|
acno = TID_TO_WME_AC(tidno);
|
|
|
tid->ac = &an->ac[acno];
|
|
|
}
|
|
@@ -2602,6 +2706,7 @@ void ath_tx_node_init(struct ath_softc *sc, struct ath_node *an)
|
|
|
for (acno = 0, ac = &an->ac[acno];
|
|
|
acno < IEEE80211_NUM_ACS; acno++, ac++) {
|
|
|
ac->sched = false;
|
|
|
+ ac->clear_ps_filter = true;
|
|
|
ac->txq = sc->tx.txq_map[acno];
|
|
|
INIT_LIST_HEAD(&ac->tid_q);
|
|
|
}
|