1 /* 2 * Copyright (c) 2005-2011 Atheros Communications Inc. 3 * Copyright (c) 2011-2013 Qualcomm Atheros, Inc. 4 * 5 * Permission to use, copy, modify, and/or distribute this software for any 6 * purpose with or without fee is hereby granted, provided that the above 7 * copyright notice and this permission notice appear in all copies. 8 * 9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES 10 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF 11 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR 12 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES 13 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 14 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 15 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 16 */ 17 18 #include "core.h" 19 #include "txrx.h" 20 #include "htt.h" 21 #include "mac.h" 22 #include "debug.h" 23 24 static void ath10k_report_offchan_tx(struct ath10k *ar, struct sk_buff *skb) 25 { 26 if (!ATH10K_SKB_CB(skb)->htt.is_offchan) 27 return; 28 29 /* If the original wait_for_completion() timed out before 30 * {data,mgmt}_tx_completed() was called then we could complete 31 * offchan_tx_completed for a different skb. Prevent this by using 32 * offchan_tx_skb. */ 33 spin_lock_bh(&ar->data_lock); 34 if (ar->offchan_tx_skb != skb) { 35 ath10k_warn(ar, "completed old offchannel frame\n"); 36 goto out; 37 } 38 39 complete(&ar->offchan_tx_completed); 40 ar->offchan_tx_skb = NULL; /* just for sanity */ 41 42 ath10k_dbg(ar, ATH10K_DBG_HTT, "completed offchannel skb %p\n", skb); 43 out: 44 spin_unlock_bh(&ar->data_lock); 45 } 46 47 void ath10k_txrx_tx_unref(struct ath10k_htt *htt, 48 const struct htt_tx_done *tx_done) 49 { 50 struct ath10k *ar = htt->ar; 51 struct device *dev = ar->dev; 52 struct ieee80211_tx_info *info; 53 struct ath10k_skb_cb *skb_cb; 54 struct sk_buff *msdu; 55 struct ieee80211_hdr *hdr; 56 __le16 fc; 57 bool limit_mgmt_desc = false; 58 59 ath10k_dbg(ar, ATH10K_DBG_HTT, 60 "htt tx completion msdu_id %u discard %d no_ack %d success %d\n", 61 tx_done->msdu_id, !!tx_done->discard, 62 !!tx_done->no_ack, !!tx_done->success); 63 64 if (tx_done->msdu_id >= htt->max_num_pending_tx) { 65 ath10k_warn(ar, "warning: msdu_id %d too big, ignoring\n", 66 tx_done->msdu_id); 67 return; 68 } 69 70 spin_lock_bh(&htt->tx_lock); 71 msdu = idr_find(&htt->pending_tx, tx_done->msdu_id); 72 if (!msdu) { 73 ath10k_warn(ar, "received tx completion for invalid msdu_id: %d\n", 74 tx_done->msdu_id); 75 spin_unlock_bh(&htt->tx_lock); 76 return; 77 } 78 79 hdr = (struct ieee80211_hdr *)msdu->data; 80 fc = hdr->frame_control; 81 82 if (unlikely(ieee80211_is_mgmt(fc)) && 83 ar->hw_params.max_probe_resp_desc_thres) 84 limit_mgmt_desc = true; 85 86 ath10k_htt_tx_free_msdu_id(htt, tx_done->msdu_id); 87 __ath10k_htt_tx_dec_pending(htt, limit_mgmt_desc); 88 if (htt->num_pending_tx == 0) 89 wake_up(&htt->empty_tx_wq); 90 spin_unlock_bh(&htt->tx_lock); 91 92 skb_cb = ATH10K_SKB_CB(msdu); 93 dma_unmap_single(dev, skb_cb->paddr, msdu->len, DMA_TO_DEVICE); 94 95 ath10k_report_offchan_tx(htt->ar, msdu); 96 97 info = IEEE80211_SKB_CB(msdu); 98 memset(&info->status, 0, sizeof(info->status)); 99 trace_ath10k_txrx_tx_unref(ar, tx_done->msdu_id); 100 101 if (tx_done->discard) { 102 ieee80211_free_txskb(htt->ar->hw, msdu); 103 return; 104 } 105 106 if (!(info->flags & IEEE80211_TX_CTL_NO_ACK)) 107 info->flags |= IEEE80211_TX_STAT_ACK; 108 109 if (tx_done->no_ack) 110 info->flags &= ~IEEE80211_TX_STAT_ACK; 111 112 if (tx_done->success && (info->flags & IEEE80211_TX_CTL_NO_ACK)) 113 info->flags |= IEEE80211_TX_STAT_NOACK_TRANSMITTED; 114 115 ieee80211_tx_status(htt->ar->hw, msdu); 116 /* we do not own the msdu anymore */ 117 } 118 119 struct ath10k_peer *ath10k_peer_find(struct ath10k *ar, int vdev_id, 120 const u8 *addr) 121 { 122 struct ath10k_peer *peer; 123 124 lockdep_assert_held(&ar->data_lock); 125 126 list_for_each_entry(peer, &ar->peers, list) { 127 if (peer->vdev_id != vdev_id) 128 continue; 129 if (memcmp(peer->addr, addr, ETH_ALEN)) 130 continue; 131 132 return peer; 133 } 134 135 return NULL; 136 } 137 138 struct ath10k_peer *ath10k_peer_find_by_id(struct ath10k *ar, int peer_id) 139 { 140 struct ath10k_peer *peer; 141 142 lockdep_assert_held(&ar->data_lock); 143 144 list_for_each_entry(peer, &ar->peers, list) 145 if (test_bit(peer_id, peer->peer_ids)) 146 return peer; 147 148 return NULL; 149 } 150 151 static int ath10k_wait_for_peer_common(struct ath10k *ar, int vdev_id, 152 const u8 *addr, bool expect_mapped) 153 { 154 long time_left; 155 156 time_left = wait_event_timeout(ar->peer_mapping_wq, ({ 157 bool mapped; 158 159 spin_lock_bh(&ar->data_lock); 160 mapped = !!ath10k_peer_find(ar, vdev_id, addr); 161 spin_unlock_bh(&ar->data_lock); 162 163 (mapped == expect_mapped || 164 test_bit(ATH10K_FLAG_CRASH_FLUSH, &ar->dev_flags)); 165 }), 3*HZ); 166 167 if (time_left == 0) 168 return -ETIMEDOUT; 169 170 return 0; 171 } 172 173 int ath10k_wait_for_peer_created(struct ath10k *ar, int vdev_id, const u8 *addr) 174 { 175 return ath10k_wait_for_peer_common(ar, vdev_id, addr, true); 176 } 177 178 int ath10k_wait_for_peer_deleted(struct ath10k *ar, int vdev_id, const u8 *addr) 179 { 180 return ath10k_wait_for_peer_common(ar, vdev_id, addr, false); 181 } 182 183 void ath10k_peer_map_event(struct ath10k_htt *htt, 184 struct htt_peer_map_event *ev) 185 { 186 struct ath10k *ar = htt->ar; 187 struct ath10k_peer *peer; 188 189 spin_lock_bh(&ar->data_lock); 190 peer = ath10k_peer_find(ar, ev->vdev_id, ev->addr); 191 if (!peer) { 192 peer = kzalloc(sizeof(*peer), GFP_ATOMIC); 193 if (!peer) 194 goto exit; 195 196 peer->vdev_id = ev->vdev_id; 197 ether_addr_copy(peer->addr, ev->addr); 198 list_add(&peer->list, &ar->peers); 199 wake_up(&ar->peer_mapping_wq); 200 } 201 202 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt peer map vdev %d peer %pM id %d\n", 203 ev->vdev_id, ev->addr, ev->peer_id); 204 205 set_bit(ev->peer_id, peer->peer_ids); 206 exit: 207 spin_unlock_bh(&ar->data_lock); 208 } 209 210 void ath10k_peer_unmap_event(struct ath10k_htt *htt, 211 struct htt_peer_unmap_event *ev) 212 { 213 struct ath10k *ar = htt->ar; 214 struct ath10k_peer *peer; 215 216 spin_lock_bh(&ar->data_lock); 217 peer = ath10k_peer_find_by_id(ar, ev->peer_id); 218 if (!peer) { 219 ath10k_warn(ar, "peer-unmap-event: unknown peer id %d\n", 220 ev->peer_id); 221 goto exit; 222 } 223 224 ath10k_dbg(ar, ATH10K_DBG_HTT, "htt peer unmap vdev %d peer %pM id %d\n", 225 peer->vdev_id, peer->addr, ev->peer_id); 226 227 clear_bit(ev->peer_id, peer->peer_ids); 228 229 if (bitmap_empty(peer->peer_ids, ATH10K_MAX_NUM_PEER_IDS)) { 230 list_del(&peer->list); 231 kfree(peer); 232 wake_up(&ar->peer_mapping_wq); 233 } 234 235 exit: 236 spin_unlock_bh(&ar->data_lock); 237 } 238