|
@@ -358,34 +358,6 @@ restart:
|
|
|
queue_work(priv->workqueue, &priv->restart);
|
|
|
}
|
|
|
|
|
|
-static void iwl5000_set_wr_ptrs(struct iwl_priv *priv,
|
|
|
- int txq_id, u32 index)
|
|
|
-{
|
|
|
- iwl_write_direct32(priv, HBUS_TARG_WRPTR,
|
|
|
- (index & 0xff) | (txq_id << 8));
|
|
|
- iwl_write_prph(priv, IWL50_SCD_QUEUE_RDPTR(txq_id), index);
|
|
|
-}
|
|
|
-
|
|
|
-static void iwl5000_tx_queue_set_status(struct iwl_priv *priv,
|
|
|
- struct iwl_tx_queue *txq,
|
|
|
- int tx_fifo_id, int scd_retry)
|
|
|
-{
|
|
|
- int txq_id = txq->q.id;
|
|
|
- int active = test_bit(txq_id, &priv->txq_ctx_active_msk) ? 1 : 0;
|
|
|
-
|
|
|
- iwl_write_prph(priv, IWL50_SCD_QUEUE_STATUS_BITS(txq_id),
|
|
|
- (active << IWL50_SCD_QUEUE_STTS_REG_POS_ACTIVE) |
|
|
|
- (tx_fifo_id << IWL50_SCD_QUEUE_STTS_REG_POS_TXF) |
|
|
|
- (1 << IWL50_SCD_QUEUE_STTS_REG_POS_WSL) |
|
|
|
- IWL50_SCD_QUEUE_STTS_REG_MSK);
|
|
|
-
|
|
|
- txq->sched_retry = scd_retry;
|
|
|
-
|
|
|
- IWL_DEBUG_INFO(priv, "%s %s Queue %d on FIFO %d\n",
|
|
|
- active ? "Activate" : "Deactivate",
|
|
|
- scd_retry ? "BA" : "AC/CMD", txq_id, tx_fifo_id);
|
|
|
-}
|
|
|
-
|
|
|
int iwl5000_alive_notify(struct iwl_priv *priv)
|
|
|
{
|
|
|
u32 a;
|
|
@@ -448,7 +420,7 @@ int iwl5000_alive_notify(struct iwl_priv *priv)
|
|
|
/* Activate all Tx DMA/FIFO channels */
|
|
|
priv->cfg->ops->lib->txq_set_sched(priv, IWL_MASK(0, 7));
|
|
|
|
|
|
- iwl5000_set_wr_ptrs(priv, IWL_CMD_QUEUE_NUM, 0);
|
|
|
+ iwlagn_set_wr_ptrs(priv, IWL_CMD_QUEUE_NUM, 0);
|
|
|
|
|
|
/* make sure all queue are not stopped */
|
|
|
memset(&priv->queue_stopped[0], 0, sizeof(priv->queue_stopped));
|
|
@@ -468,7 +440,7 @@ int iwl5000_alive_notify(struct iwl_priv *priv)
|
|
|
if (ac == IWL_TX_FIFO_UNUSED)
|
|
|
continue;
|
|
|
|
|
|
- iwl5000_tx_queue_set_status(priv, &priv->txq[i], ac, 0);
|
|
|
+ iwlagn_tx_queue_set_status(priv, &priv->txq[i], ac, 0);
|
|
|
}
|
|
|
|
|
|
spin_unlock_irqrestore(&priv->lock, flags);
|
|
@@ -539,207 +511,6 @@ int iwl5000_hw_set_hw_params(struct iwl_priv *priv)
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-/**
|
|
|
- * iwl5000_txq_update_byte_cnt_tbl - Set up entry in Tx byte-count array
|
|
|
- */
|
|
|
-void iwl5000_txq_update_byte_cnt_tbl(struct iwl_priv *priv,
|
|
|
- struct iwl_tx_queue *txq,
|
|
|
- u16 byte_cnt)
|
|
|
-{
|
|
|
- struct iwl5000_scd_bc_tbl *scd_bc_tbl = priv->scd_bc_tbls.addr;
|
|
|
- int write_ptr = txq->q.write_ptr;
|
|
|
- int txq_id = txq->q.id;
|
|
|
- u8 sec_ctl = 0;
|
|
|
- u8 sta_id = 0;
|
|
|
- u16 len = byte_cnt + IWL_TX_CRC_SIZE + IWL_TX_DELIMITER_SIZE;
|
|
|
- __le16 bc_ent;
|
|
|
-
|
|
|
- WARN_ON(len > 0xFFF || write_ptr >= TFD_QUEUE_SIZE_MAX);
|
|
|
-
|
|
|
- if (txq_id != IWL_CMD_QUEUE_NUM) {
|
|
|
- sta_id = txq->cmd[txq->q.write_ptr]->cmd.tx.sta_id;
|
|
|
- sec_ctl = txq->cmd[txq->q.write_ptr]->cmd.tx.sec_ctl;
|
|
|
-
|
|
|
- switch (sec_ctl & TX_CMD_SEC_MSK) {
|
|
|
- case TX_CMD_SEC_CCM:
|
|
|
- len += CCMP_MIC_LEN;
|
|
|
- break;
|
|
|
- case TX_CMD_SEC_TKIP:
|
|
|
- len += TKIP_ICV_LEN;
|
|
|
- break;
|
|
|
- case TX_CMD_SEC_WEP:
|
|
|
- len += WEP_IV_LEN + WEP_ICV_LEN;
|
|
|
- break;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- bc_ent = cpu_to_le16((len & 0xFFF) | (sta_id << 12));
|
|
|
-
|
|
|
- scd_bc_tbl[txq_id].tfd_offset[write_ptr] = bc_ent;
|
|
|
-
|
|
|
- if (write_ptr < TFD_QUEUE_SIZE_BC_DUP)
|
|
|
- scd_bc_tbl[txq_id].
|
|
|
- tfd_offset[TFD_QUEUE_SIZE_MAX + write_ptr] = bc_ent;
|
|
|
-}
|
|
|
-
|
|
|
-void iwl5000_txq_inval_byte_cnt_tbl(struct iwl_priv *priv,
|
|
|
- struct iwl_tx_queue *txq)
|
|
|
-{
|
|
|
- struct iwl5000_scd_bc_tbl *scd_bc_tbl = priv->scd_bc_tbls.addr;
|
|
|
- int txq_id = txq->q.id;
|
|
|
- int read_ptr = txq->q.read_ptr;
|
|
|
- u8 sta_id = 0;
|
|
|
- __le16 bc_ent;
|
|
|
-
|
|
|
- WARN_ON(read_ptr >= TFD_QUEUE_SIZE_MAX);
|
|
|
-
|
|
|
- if (txq_id != IWL_CMD_QUEUE_NUM)
|
|
|
- sta_id = txq->cmd[read_ptr]->cmd.tx.sta_id;
|
|
|
-
|
|
|
- bc_ent = cpu_to_le16(1 | (sta_id << 12));
|
|
|
- scd_bc_tbl[txq_id].tfd_offset[read_ptr] = bc_ent;
|
|
|
-
|
|
|
- if (read_ptr < TFD_QUEUE_SIZE_BC_DUP)
|
|
|
- scd_bc_tbl[txq_id].
|
|
|
- tfd_offset[TFD_QUEUE_SIZE_MAX + read_ptr] = bc_ent;
|
|
|
-}
|
|
|
-
|
|
|
-static int iwl5000_tx_queue_set_q2ratid(struct iwl_priv *priv, u16 ra_tid,
|
|
|
- u16 txq_id)
|
|
|
-{
|
|
|
- u32 tbl_dw_addr;
|
|
|
- u32 tbl_dw;
|
|
|
- u16 scd_q2ratid;
|
|
|
-
|
|
|
- scd_q2ratid = ra_tid & IWL_SCD_QUEUE_RA_TID_MAP_RATID_MSK;
|
|
|
-
|
|
|
- tbl_dw_addr = priv->scd_base_addr +
|
|
|
- IWL50_SCD_TRANSLATE_TBL_OFFSET_QUEUE(txq_id);
|
|
|
-
|
|
|
- tbl_dw = iwl_read_targ_mem(priv, tbl_dw_addr);
|
|
|
-
|
|
|
- if (txq_id & 0x1)
|
|
|
- tbl_dw = (scd_q2ratid << 16) | (tbl_dw & 0x0000FFFF);
|
|
|
- else
|
|
|
- tbl_dw = scd_q2ratid | (tbl_dw & 0xFFFF0000);
|
|
|
-
|
|
|
- iwl_write_targ_mem(priv, tbl_dw_addr, tbl_dw);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-static void iwl5000_tx_queue_stop_scheduler(struct iwl_priv *priv, u16 txq_id)
|
|
|
-{
|
|
|
- /* Simply stop the queue, but don't change any configuration;
|
|
|
- * the SCD_ACT_EN bit is the write-enable mask for the ACTIVE bit. */
|
|
|
- iwl_write_prph(priv,
|
|
|
- IWL50_SCD_QUEUE_STATUS_BITS(txq_id),
|
|
|
- (0 << IWL50_SCD_QUEUE_STTS_REG_POS_ACTIVE)|
|
|
|
- (1 << IWL50_SCD_QUEUE_STTS_REG_POS_SCD_ACT_EN));
|
|
|
-}
|
|
|
-
|
|
|
-int iwl5000_txq_agg_enable(struct iwl_priv *priv, int txq_id,
|
|
|
- int tx_fifo, int sta_id, int tid, u16 ssn_idx)
|
|
|
-{
|
|
|
- unsigned long flags;
|
|
|
- u16 ra_tid;
|
|
|
-
|
|
|
- if ((IWL50_FIRST_AMPDU_QUEUE > txq_id) ||
|
|
|
- (IWL50_FIRST_AMPDU_QUEUE + priv->cfg->num_of_ampdu_queues
|
|
|
- <= txq_id)) {
|
|
|
- IWL_WARN(priv,
|
|
|
- "queue number out of range: %d, must be %d to %d\n",
|
|
|
- txq_id, IWL50_FIRST_AMPDU_QUEUE,
|
|
|
- IWL50_FIRST_AMPDU_QUEUE +
|
|
|
- priv->cfg->num_of_ampdu_queues - 1);
|
|
|
- return -EINVAL;
|
|
|
- }
|
|
|
-
|
|
|
- ra_tid = BUILD_RAxTID(sta_id, tid);
|
|
|
-
|
|
|
- /* Modify device's station table to Tx this TID */
|
|
|
- iwl_sta_tx_modify_enable_tid(priv, sta_id, tid);
|
|
|
-
|
|
|
- spin_lock_irqsave(&priv->lock, flags);
|
|
|
-
|
|
|
- /* Stop this Tx queue before configuring it */
|
|
|
- iwl5000_tx_queue_stop_scheduler(priv, txq_id);
|
|
|
-
|
|
|
- /* Map receiver-address / traffic-ID to this queue */
|
|
|
- iwl5000_tx_queue_set_q2ratid(priv, ra_tid, txq_id);
|
|
|
-
|
|
|
- /* Set this queue as a chain-building queue */
|
|
|
- iwl_set_bits_prph(priv, IWL50_SCD_QUEUECHAIN_SEL, (1<<txq_id));
|
|
|
-
|
|
|
- /* enable aggregations for the queue */
|
|
|
- iwl_set_bits_prph(priv, IWL50_SCD_AGGR_SEL, (1<<txq_id));
|
|
|
-
|
|
|
- /* Place first TFD at index corresponding to start sequence number.
|
|
|
- * Assumes that ssn_idx is valid (!= 0xFFF) */
|
|
|
- priv->txq[txq_id].q.read_ptr = (ssn_idx & 0xff);
|
|
|
- priv->txq[txq_id].q.write_ptr = (ssn_idx & 0xff);
|
|
|
- iwl5000_set_wr_ptrs(priv, txq_id, ssn_idx);
|
|
|
-
|
|
|
- /* Set up Tx window size and frame limit for this queue */
|
|
|
- iwl_write_targ_mem(priv, priv->scd_base_addr +
|
|
|
- IWL50_SCD_CONTEXT_QUEUE_OFFSET(txq_id) +
|
|
|
- sizeof(u32),
|
|
|
- ((SCD_WIN_SIZE <<
|
|
|
- IWL50_SCD_QUEUE_CTX_REG2_WIN_SIZE_POS) &
|
|
|
- IWL50_SCD_QUEUE_CTX_REG2_WIN_SIZE_MSK) |
|
|
|
- ((SCD_FRAME_LIMIT <<
|
|
|
- IWL50_SCD_QUEUE_CTX_REG2_FRAME_LIMIT_POS) &
|
|
|
- IWL50_SCD_QUEUE_CTX_REG2_FRAME_LIMIT_MSK));
|
|
|
-
|
|
|
- iwl_set_bits_prph(priv, IWL50_SCD_INTERRUPT_MASK, (1 << txq_id));
|
|
|
-
|
|
|
- /* Set up Status area in SRAM, map to Tx DMA/FIFO, activate the queue */
|
|
|
- iwl5000_tx_queue_set_status(priv, &priv->txq[txq_id], tx_fifo, 1);
|
|
|
-
|
|
|
- spin_unlock_irqrestore(&priv->lock, flags);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-int iwl5000_txq_agg_disable(struct iwl_priv *priv, u16 txq_id,
|
|
|
- u16 ssn_idx, u8 tx_fifo)
|
|
|
-{
|
|
|
- if ((IWL50_FIRST_AMPDU_QUEUE > txq_id) ||
|
|
|
- (IWL50_FIRST_AMPDU_QUEUE + priv->cfg->num_of_ampdu_queues
|
|
|
- <= txq_id)) {
|
|
|
- IWL_ERR(priv,
|
|
|
- "queue number out of range: %d, must be %d to %d\n",
|
|
|
- txq_id, IWL50_FIRST_AMPDU_QUEUE,
|
|
|
- IWL50_FIRST_AMPDU_QUEUE +
|
|
|
- priv->cfg->num_of_ampdu_queues - 1);
|
|
|
- return -EINVAL;
|
|
|
- }
|
|
|
-
|
|
|
- iwl5000_tx_queue_stop_scheduler(priv, txq_id);
|
|
|
-
|
|
|
- iwl_clear_bits_prph(priv, IWL50_SCD_AGGR_SEL, (1 << txq_id));
|
|
|
-
|
|
|
- priv->txq[txq_id].q.read_ptr = (ssn_idx & 0xff);
|
|
|
- priv->txq[txq_id].q.write_ptr = (ssn_idx & 0xff);
|
|
|
- /* supposes that ssn_idx is valid (!= 0xFFF) */
|
|
|
- iwl5000_set_wr_ptrs(priv, txq_id, ssn_idx);
|
|
|
-
|
|
|
- iwl_clear_bits_prph(priv, IWL50_SCD_INTERRUPT_MASK, (1 << txq_id));
|
|
|
- iwl_txq_ctx_deactivate(priv, txq_id);
|
|
|
- iwl5000_tx_queue_set_status(priv, &priv->txq[txq_id], tx_fifo, 0);
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-/*
|
|
|
- * Activate/Deactivate Tx DMA/FIFO channels according tx fifos mask
|
|
|
- * must be called under priv->lock and mac access
|
|
|
- */
|
|
|
-void iwl5000_txq_set_sched(struct iwl_priv *priv, u32 mask)
|
|
|
-{
|
|
|
- iwl_write_prph(priv, IWL50_SCD_TXFACT, mask);
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
static inline u32 iwl5000_get_scd_ssn(struct iwl5000_tx_resp *tx_resp)
|
|
|
{
|
|
|
return le32_to_cpup((__le32 *)&tx_resp->status +
|
|
@@ -1063,11 +834,11 @@ static int iwl5000_hw_channel_switch(struct iwl_priv *priv, u16 channel)
|
|
|
|
|
|
struct iwl_lib_ops iwl5000_lib = {
|
|
|
.set_hw_params = iwl5000_hw_set_hw_params,
|
|
|
- .txq_update_byte_cnt_tbl = iwl5000_txq_update_byte_cnt_tbl,
|
|
|
- .txq_inval_byte_cnt_tbl = iwl5000_txq_inval_byte_cnt_tbl,
|
|
|
- .txq_set_sched = iwl5000_txq_set_sched,
|
|
|
- .txq_agg_enable = iwl5000_txq_agg_enable,
|
|
|
- .txq_agg_disable = iwl5000_txq_agg_disable,
|
|
|
+ .txq_update_byte_cnt_tbl = iwlagn_txq_update_byte_cnt_tbl,
|
|
|
+ .txq_inval_byte_cnt_tbl = iwlagn_txq_inval_byte_cnt_tbl,
|
|
|
+ .txq_set_sched = iwlagn_txq_set_sched,
|
|
|
+ .txq_agg_enable = iwlagn_txq_agg_enable,
|
|
|
+ .txq_agg_disable = iwlagn_txq_agg_disable,
|
|
|
.txq_attach_buf_to_tfd = iwl_hw_txq_attach_buf_to_tfd,
|
|
|
.txq_free_tfd = iwl_hw_txq_free_tfd,
|
|
|
.txq_init = iwl_hw_tx_queue_init,
|
|
@@ -1121,11 +892,11 @@ struct iwl_lib_ops iwl5000_lib = {
|
|
|
|
|
|
static struct iwl_lib_ops iwl5150_lib = {
|
|
|
.set_hw_params = iwl5000_hw_set_hw_params,
|
|
|
- .txq_update_byte_cnt_tbl = iwl5000_txq_update_byte_cnt_tbl,
|
|
|
- .txq_inval_byte_cnt_tbl = iwl5000_txq_inval_byte_cnt_tbl,
|
|
|
- .txq_set_sched = iwl5000_txq_set_sched,
|
|
|
- .txq_agg_enable = iwl5000_txq_agg_enable,
|
|
|
- .txq_agg_disable = iwl5000_txq_agg_disable,
|
|
|
+ .txq_update_byte_cnt_tbl = iwlagn_txq_update_byte_cnt_tbl,
|
|
|
+ .txq_inval_byte_cnt_tbl = iwlagn_txq_inval_byte_cnt_tbl,
|
|
|
+ .txq_set_sched = iwlagn_txq_set_sched,
|
|
|
+ .txq_agg_enable = iwlagn_txq_agg_enable,
|
|
|
+ .txq_agg_disable = iwlagn_txq_agg_disable,
|
|
|
.txq_attach_buf_to_tfd = iwl_hw_txq_attach_buf_to_tfd,
|
|
|
.txq_free_tfd = iwl_hw_txq_free_tfd,
|
|
|
.txq_init = iwl_hw_tx_queue_init,
|