mwifiex: provision for holding and restoring packets during TDLS setup

While TDLS link is being setup, few packets from this station to
peer station may be buffered at AP. It may happen that once TDLS
link is setup, packets sent from station to peer on direct link
get delivered before traffic from AP arrives at peer station.
This results into packet reordering issue at peer station.

To avoid this, we hold data packets destined to TDLS peer during
TDLS setup. These packets are moved to temperory TDLS TX queue.
Upon successful TDLS setup, they are moved to RA list created for
this peer. Upon failure, packets are moved back to AP's RA list
for that particular TID.

Signed-off-by: Avinash Patil <patila@marvell.com>
Signed-off-by: Bing Zhao <bzhao@marvell.com>
Signed-off-by: John W. Linville <linville@tuxdriver.com>
This commit is contained in:
Avinash Patil 2014-02-07 16:30:35 -08:00 committed by John W. Linville
parent 1f4dfd8a1e
commit 56bd24a18e
4 changed files with 100 additions and 0 deletions

View File

@ -452,6 +452,7 @@ int mwifiex_init_lock_list(struct mwifiex_adapter *adapter)
INIT_LIST_HEAD(&priv->tx_ba_stream_tbl_ptr); INIT_LIST_HEAD(&priv->tx_ba_stream_tbl_ptr);
INIT_LIST_HEAD(&priv->rx_reorder_tbl_ptr); INIT_LIST_HEAD(&priv->rx_reorder_tbl_ptr);
INIT_LIST_HEAD(&priv->sta_list); INIT_LIST_HEAD(&priv->sta_list);
skb_queue_head_init(&priv->tdls_txq);
spin_lock_init(&priv->tx_ba_stream_tbl_lock); spin_lock_init(&priv->tx_ba_stream_tbl_lock);
spin_lock_init(&priv->rx_reorder_tbl_lock); spin_lock_init(&priv->rx_reorder_tbl_lock);

View File

@ -530,6 +530,7 @@ struct mwifiex_private {
u8 del_list_idx; u8 del_list_idx;
bool hs2_enabled; bool hs2_enabled;
struct station_parameters *sta_params; struct station_parameters *sta_params;
struct sk_buff_head tdls_txq;
}; };
enum mwifiex_ba_status { enum mwifiex_ba_status {

View File

@ -24,6 +24,96 @@
#define TDLS_RESP_FIX_LEN 8 #define TDLS_RESP_FIX_LEN 8
#define TDLS_CONFIRM_FIX_LEN 6 #define TDLS_CONFIRM_FIX_LEN 6
static void
mwifiex_restore_tdls_packets(struct mwifiex_private *priv, u8 *mac, u8 status)
{
struct mwifiex_ra_list_tbl *ra_list;
struct list_head *tid_list;
struct sk_buff *skb, *tmp;
struct mwifiex_txinfo *tx_info;
unsigned long flags;
u32 tid;
u8 tid_down;
dev_dbg(priv->adapter->dev, "%s: %pM\n", __func__, mac);
spin_lock_irqsave(&priv->wmm.ra_list_spinlock, flags);
skb_queue_walk_safe(&priv->tdls_txq, skb, tmp) {
if (!ether_addr_equal(mac, skb->data))
continue;
__skb_unlink(skb, &priv->tdls_txq);
tx_info = MWIFIEX_SKB_TXCB(skb);
tid = skb->priority;
tid_down = mwifiex_wmm_downgrade_tid(priv, tid);
if (status == TDLS_SETUP_COMPLETE) {
ra_list = mwifiex_wmm_get_queue_raptr(priv, tid, mac);
tx_info->flags |= MWIFIEX_BUF_FLAG_TDLS_PKT;
} else {
tid_list = &priv->wmm.tid_tbl_ptr[tid_down].ra_list;
if (!list_empty(tid_list))
ra_list = list_first_entry(tid_list,
struct mwifiex_ra_list_tbl, list);
else
ra_list = NULL;
tx_info->flags &= ~MWIFIEX_BUF_FLAG_TDLS_PKT;
}
if (!ra_list) {
mwifiex_write_data_complete(priv->adapter, skb, 0, -1);
continue;
}
skb_queue_tail(&ra_list->skb_head, skb);
ra_list->ba_pkt_count++;
ra_list->total_pkt_count++;
if (atomic_read(&priv->wmm.highest_queued_prio) <
tos_to_tid_inv[tid_down])
atomic_set(&priv->wmm.highest_queued_prio,
tos_to_tid_inv[tid_down]);
atomic_inc(&priv->wmm.tx_pkts_queued);
}
spin_unlock_irqrestore(&priv->wmm.ra_list_spinlock, flags);
return;
}
static void mwifiex_hold_tdls_packets(struct mwifiex_private *priv, u8 *mac)
{
struct mwifiex_ra_list_tbl *ra_list;
struct list_head *ra_list_head;
struct sk_buff *skb, *tmp;
unsigned long flags;
int i;
dev_dbg(priv->adapter->dev, "%s: %pM\n", __func__, mac);
spin_lock_irqsave(&priv->wmm.ra_list_spinlock, flags);
for (i = 0; i < MAX_NUM_TID; i++) {
if (!list_empty(&priv->wmm.tid_tbl_ptr[i].ra_list)) {
ra_list_head = &priv->wmm.tid_tbl_ptr[i].ra_list;
list_for_each_entry(ra_list, ra_list_head, list) {
skb_queue_walk_safe(&ra_list->skb_head, skb,
tmp) {
if (!ether_addr_equal(mac, skb->data))
continue;
__skb_unlink(skb, &ra_list->skb_head);
atomic_dec(&priv->wmm.tx_pkts_queued);
ra_list->total_pkt_count--;
skb_queue_tail(&priv->tdls_txq, skb);
}
}
}
}
spin_unlock_irqrestore(&priv->wmm.ra_list_spinlock, flags);
return;
}
/* This function appends rate TLV to scan config command. */ /* This function appends rate TLV to scan config command. */
static int static int
mwifiex_tdls_append_rates_ie(struct mwifiex_private *priv, mwifiex_tdls_append_rates_ie(struct mwifiex_private *priv,
@ -584,6 +674,7 @@ mwifiex_tdls_process_create_link(struct mwifiex_private *priv, u8 *peer)
return -ENOMEM; return -ENOMEM;
sta_ptr->tdls_status = TDLS_SETUP_INPROGRESS; sta_ptr->tdls_status = TDLS_SETUP_INPROGRESS;
mwifiex_hold_tdls_packets(priv, peer);
memcpy(&tdls_oper.peer_mac, peer, ETH_ALEN); memcpy(&tdls_oper.peer_mac, peer, ETH_ALEN);
tdls_oper.tdls_action = MWIFIEX_TDLS_CREATE_LINK; tdls_oper.tdls_action = MWIFIEX_TDLS_CREATE_LINK;
return mwifiex_send_cmd_sync(priv, HostCmd_CMD_TDLS_OPER, return mwifiex_send_cmd_sync(priv, HostCmd_CMD_TDLS_OPER,
@ -612,6 +703,7 @@ mwifiex_tdls_process_disable_link(struct mwifiex_private *priv, u8 *peer)
mwifiex_del_sta_entry(priv, peer); mwifiex_del_sta_entry(priv, peer);
} }
mwifiex_restore_tdls_packets(priv, peer, TDLS_LINK_TEARDOWN);
memcpy(&tdls_oper.peer_mac, peer, ETH_ALEN); memcpy(&tdls_oper.peer_mac, peer, ETH_ALEN);
tdls_oper.tdls_action = MWIFIEX_TDLS_DISABLE_LINK; tdls_oper.tdls_action = MWIFIEX_TDLS_DISABLE_LINK;
return mwifiex_send_cmd_sync(priv, HostCmd_CMD_TDLS_OPER, return mwifiex_send_cmd_sync(priv, HostCmd_CMD_TDLS_OPER,
@ -655,6 +747,7 @@ mwifiex_tdls_process_enable_link(struct mwifiex_private *priv, u8 *peer)
} }
memset(sta_ptr->rx_seq, 0xff, sizeof(sta_ptr->rx_seq)); memset(sta_ptr->rx_seq, 0xff, sizeof(sta_ptr->rx_seq));
mwifiex_restore_tdls_packets(priv, peer, TDLS_SETUP_COMPLETE);
} else { } else {
dev_dbg(priv->adapter->dev, dev_dbg(priv->adapter->dev,
"tdls: enable link %pM failed\n", peer); "tdls: enable link %pM failed\n", peer);
@ -667,6 +760,7 @@ mwifiex_tdls_process_enable_link(struct mwifiex_private *priv, u8 *peer)
flags); flags);
mwifiex_del_sta_entry(priv, peer); mwifiex_del_sta_entry(priv, peer);
} }
mwifiex_restore_tdls_packets(priv, peer, TDLS_LINK_TEARDOWN);
return -1; return -1;
} }

View File

@ -533,6 +533,7 @@ void
mwifiex_clean_txrx(struct mwifiex_private *priv) mwifiex_clean_txrx(struct mwifiex_private *priv)
{ {
unsigned long flags; unsigned long flags;
struct sk_buff *skb, *tmp;
mwifiex_11n_cleanup_reorder_tbl(priv); mwifiex_11n_cleanup_reorder_tbl(priv);
spin_lock_irqsave(&priv->wmm.ra_list_spinlock, flags); spin_lock_irqsave(&priv->wmm.ra_list_spinlock, flags);
@ -549,6 +550,9 @@ mwifiex_clean_txrx(struct mwifiex_private *priv)
if (priv->adapter->if_ops.clean_pcie_ring) if (priv->adapter->if_ops.clean_pcie_ring)
priv->adapter->if_ops.clean_pcie_ring(priv->adapter); priv->adapter->if_ops.clean_pcie_ring(priv->adapter);
spin_unlock_irqrestore(&priv->wmm.ra_list_spinlock, flags); spin_unlock_irqrestore(&priv->wmm.ra_list_spinlock, flags);
skb_queue_walk_safe(&priv->tdls_txq, skb, tmp)
mwifiex_write_data_complete(priv->adapter, skb, 0, -1);
} }
/* /*