1 /* 2 * Copyright 2002-2005, Instant802 Networks, Inc. 3 * Copyright 2005-2006, Devicescape Software, Inc. 4 * Copyright 2006-2007 Jiri Benc <jbenc@suse.cz> 5 * Copyright 2008-2010 Johannes Berg <johannes@sipsolutions.net> 6 * Copyright 2013-2014 Intel Mobile Communications GmbH 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License version 2 as 10 * published by the Free Software Foundation. 11 */ 12 13 #include <linux/export.h> 14 #include <linux/etherdevice.h> 15 #include <linux/time.h> 16 #include <net/mac80211.h> 17 #include <asm/unaligned.h> 18 #include "ieee80211_i.h" 19 #include "rate.h" 20 #include "mesh.h" 21 #include "led.h" 22 #include "wme.h" 23 24 25 void ieee80211_tx_status_irqsafe(struct ieee80211_hw *hw, 26 struct sk_buff *skb) 27 { 28 struct ieee80211_local *local = hw_to_local(hw); 29 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 30 int tmp; 31 32 skb->pkt_type = IEEE80211_TX_STATUS_MSG; 33 skb_queue_tail(info->flags & IEEE80211_TX_CTL_REQ_TX_STATUS ? 34 &local->skb_queue : &local->skb_queue_unreliable, skb); 35 tmp = skb_queue_len(&local->skb_queue) + 36 skb_queue_len(&local->skb_queue_unreliable); 37 while (tmp > IEEE80211_IRQSAFE_QUEUE_LIMIT && 38 (skb = skb_dequeue(&local->skb_queue_unreliable))) { 39 ieee80211_free_txskb(hw, skb); 40 tmp--; 41 I802_DEBUG_INC(local->tx_status_drop); 42 } 43 tasklet_schedule(&local->tasklet); 44 } 45 EXPORT_SYMBOL(ieee80211_tx_status_irqsafe); 46 47 static void ieee80211_handle_filtered_frame(struct ieee80211_local *local, 48 struct sta_info *sta, 49 struct sk_buff *skb) 50 { 51 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 52 struct ieee80211_hdr *hdr = (void *)skb->data; 53 int ac; 54 55 /* 56 * This skb 'survived' a round-trip through the driver, and 57 * hopefully the driver didn't mangle it too badly. However, 58 * we can definitely not rely on the control information 59 * being correct. Clear it so we don't get junk there, and 60 * indicate that it needs new processing, but must not be 61 * modified/encrypted again. 62 */ 63 memset(&info->control, 0, sizeof(info->control)); 64 65 info->control.jiffies = jiffies; 66 info->control.vif = &sta->sdata->vif; 67 info->flags |= IEEE80211_TX_INTFL_NEED_TXPROCESSING | 68 IEEE80211_TX_INTFL_RETRANSMISSION; 69 info->flags &= ~IEEE80211_TX_TEMPORARY_FLAGS; 70 71 sta->tx_filtered_count++; 72 73 /* 74 * Clear more-data bit on filtered frames, it might be set 75 * but later frames might time out so it might have to be 76 * clear again ... It's all rather unlikely (this frame 77 * should time out first, right?) but let's not confuse 78 * peers unnecessarily. 79 */ 80 if (hdr->frame_control & cpu_to_le16(IEEE80211_FCTL_MOREDATA)) 81 hdr->frame_control &= ~cpu_to_le16(IEEE80211_FCTL_MOREDATA); 82 83 if (ieee80211_is_data_qos(hdr->frame_control)) { 84 u8 *p = ieee80211_get_qos_ctl(hdr); 85 int tid = *p & IEEE80211_QOS_CTL_TID_MASK; 86 87 /* 88 * Clear EOSP if set, this could happen e.g. 89 * if an absence period (us being a P2P GO) 90 * shortens the SP. 91 */ 92 if (*p & IEEE80211_QOS_CTL_EOSP) 93 *p &= ~IEEE80211_QOS_CTL_EOSP; 94 ac = ieee802_1d_to_ac[tid & 7]; 95 } else { 96 ac = IEEE80211_AC_BE; 97 } 98 99 /* 100 * Clear the TX filter mask for this STA when sending the next 101 * packet. If the STA went to power save mode, this will happen 102 * when it wakes up for the next time. 103 */ 104 set_sta_flag(sta, WLAN_STA_CLEAR_PS_FILT); 105 106 /* 107 * This code races in the following way: 108 * 109 * (1) STA sends frame indicating it will go to sleep and does so 110 * (2) hardware/firmware adds STA to filter list, passes frame up 111 * (3) hardware/firmware processes TX fifo and suppresses a frame 112 * (4) we get TX status before having processed the frame and 113 * knowing that the STA has gone to sleep. 114 * 115 * This is actually quite unlikely even when both those events are 116 * processed from interrupts coming in quickly after one another or 117 * even at the same time because we queue both TX status events and 118 * RX frames to be processed by a tasklet and process them in the 119 * same order that they were received or TX status last. Hence, there 120 * is no race as long as the frame RX is processed before the next TX 121 * status, which drivers can ensure, see below. 122 * 123 * Note that this can only happen if the hardware or firmware can 124 * actually add STAs to the filter list, if this is done by the 125 * driver in response to set_tim() (which will only reduce the race 126 * this whole filtering tries to solve, not completely solve it) 127 * this situation cannot happen. 128 * 129 * To completely solve this race drivers need to make sure that they 130 * (a) don't mix the irq-safe/not irq-safe TX status/RX processing 131 * functions and 132 * (b) always process RX events before TX status events if ordering 133 * can be unknown, for example with different interrupt status 134 * bits. 135 * (c) if PS mode transitions are manual (i.e. the flag 136 * %IEEE80211_HW_AP_LINK_PS is set), always process PS state 137 * changes before calling TX status events if ordering can be 138 * unknown. 139 */ 140 if (test_sta_flag(sta, WLAN_STA_PS_STA) && 141 skb_queue_len(&sta->tx_filtered[ac]) < STA_MAX_TX_BUFFER) { 142 skb_queue_tail(&sta->tx_filtered[ac], skb); 143 sta_info_recalc_tim(sta); 144 145 if (!timer_pending(&local->sta_cleanup)) 146 mod_timer(&local->sta_cleanup, 147 round_jiffies(jiffies + 148 STA_INFO_CLEANUP_INTERVAL)); 149 return; 150 } 151 152 if (!test_sta_flag(sta, WLAN_STA_PS_STA) && 153 !(info->flags & IEEE80211_TX_INTFL_RETRIED)) { 154 /* Software retry the packet once */ 155 info->flags |= IEEE80211_TX_INTFL_RETRIED; 156 ieee80211_add_pending_skb(local, skb); 157 return; 158 } 159 160 ps_dbg_ratelimited(sta->sdata, 161 "dropped TX filtered frame, queue_len=%d PS=%d @%lu\n", 162 skb_queue_len(&sta->tx_filtered[ac]), 163 !!test_sta_flag(sta, WLAN_STA_PS_STA), jiffies); 164 ieee80211_free_txskb(&local->hw, skb); 165 } 166 167 static void ieee80211_check_pending_bar(struct sta_info *sta, u8 *addr, u8 tid) 168 { 169 struct tid_ampdu_tx *tid_tx; 170 171 tid_tx = rcu_dereference(sta->ampdu_mlme.tid_tx[tid]); 172 if (!tid_tx || !tid_tx->bar_pending) 173 return; 174 175 tid_tx->bar_pending = false; 176 ieee80211_send_bar(&sta->sdata->vif, addr, tid, tid_tx->failed_bar_ssn); 177 } 178 179 static void ieee80211_frame_acked(struct sta_info *sta, struct sk_buff *skb) 180 { 181 struct ieee80211_mgmt *mgmt = (void *) skb->data; 182 struct ieee80211_local *local = sta->local; 183 struct ieee80211_sub_if_data *sdata = sta->sdata; 184 185 if (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS) 186 sta->last_rx = jiffies; 187 188 if (ieee80211_is_data_qos(mgmt->frame_control)) { 189 struct ieee80211_hdr *hdr = (void *) skb->data; 190 u8 *qc = ieee80211_get_qos_ctl(hdr); 191 u16 tid = qc[0] & 0xf; 192 193 ieee80211_check_pending_bar(sta, hdr->addr1, tid); 194 } 195 196 if (ieee80211_is_action(mgmt->frame_control) && 197 mgmt->u.action.category == WLAN_CATEGORY_HT && 198 mgmt->u.action.u.ht_smps.action == WLAN_HT_ACTION_SMPS && 199 ieee80211_sdata_running(sdata)) { 200 enum ieee80211_smps_mode smps_mode; 201 202 switch (mgmt->u.action.u.ht_smps.smps_control) { 203 case WLAN_HT_SMPS_CONTROL_DYNAMIC: 204 smps_mode = IEEE80211_SMPS_DYNAMIC; 205 break; 206 case WLAN_HT_SMPS_CONTROL_STATIC: 207 smps_mode = IEEE80211_SMPS_STATIC; 208 break; 209 case WLAN_HT_SMPS_CONTROL_DISABLED: 210 default: /* shouldn't happen since we don't send that */ 211 smps_mode = IEEE80211_SMPS_OFF; 212 break; 213 } 214 215 if (sdata->vif.type == NL80211_IFTYPE_STATION) { 216 /* 217 * This update looks racy, but isn't -- if we come 218 * here we've definitely got a station that we're 219 * talking to, and on a managed interface that can 220 * only be the AP. And the only other place updating 221 * this variable in managed mode is before association. 222 */ 223 sdata->smps_mode = smps_mode; 224 ieee80211_queue_work(&local->hw, &sdata->recalc_smps); 225 } else if (sdata->vif.type == NL80211_IFTYPE_AP || 226 sdata->vif.type == NL80211_IFTYPE_AP_VLAN) { 227 sta->known_smps_mode = smps_mode; 228 } 229 } 230 } 231 232 static void ieee80211_set_bar_pending(struct sta_info *sta, u8 tid, u16 ssn) 233 { 234 struct tid_ampdu_tx *tid_tx; 235 236 tid_tx = rcu_dereference(sta->ampdu_mlme.tid_tx[tid]); 237 if (!tid_tx) 238 return; 239 240 tid_tx->failed_bar_ssn = ssn; 241 tid_tx->bar_pending = true; 242 } 243 244 static int ieee80211_tx_radiotap_len(struct ieee80211_tx_info *info) 245 { 246 int len = sizeof(struct ieee80211_radiotap_header); 247 248 /* IEEE80211_RADIOTAP_RATE rate */ 249 if (info->status.rates[0].idx >= 0 && 250 !(info->status.rates[0].flags & (IEEE80211_TX_RC_MCS | 251 IEEE80211_TX_RC_VHT_MCS))) 252 len += 2; 253 254 /* IEEE80211_RADIOTAP_TX_FLAGS */ 255 len += 2; 256 257 /* IEEE80211_RADIOTAP_DATA_RETRIES */ 258 len += 1; 259 260 /* IEEE80211_RADIOTAP_MCS 261 * IEEE80211_RADIOTAP_VHT */ 262 if (info->status.rates[0].idx >= 0) { 263 if (info->status.rates[0].flags & IEEE80211_TX_RC_MCS) 264 len += 3; 265 else if (info->status.rates[0].flags & IEEE80211_TX_RC_VHT_MCS) 266 len = ALIGN(len, 2) + 12; 267 } 268 269 return len; 270 } 271 272 static void 273 ieee80211_add_tx_radiotap_header(struct ieee80211_local *local, 274 struct ieee80211_supported_band *sband, 275 struct sk_buff *skb, int retry_count, 276 int rtap_len, int shift) 277 { 278 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 279 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data; 280 struct ieee80211_radiotap_header *rthdr; 281 unsigned char *pos; 282 u16 txflags; 283 284 rthdr = (struct ieee80211_radiotap_header *) skb_push(skb, rtap_len); 285 286 memset(rthdr, 0, rtap_len); 287 rthdr->it_len = cpu_to_le16(rtap_len); 288 rthdr->it_present = 289 cpu_to_le32((1 << IEEE80211_RADIOTAP_TX_FLAGS) | 290 (1 << IEEE80211_RADIOTAP_DATA_RETRIES)); 291 pos = (unsigned char *)(rthdr + 1); 292 293 /* 294 * XXX: Once radiotap gets the bitmap reset thing the vendor 295 * extensions proposal contains, we can actually report 296 * the whole set of tries we did. 297 */ 298 299 /* IEEE80211_RADIOTAP_RATE */ 300 if (info->status.rates[0].idx >= 0 && 301 !(info->status.rates[0].flags & (IEEE80211_TX_RC_MCS | 302 IEEE80211_TX_RC_VHT_MCS))) { 303 u16 rate; 304 305 rthdr->it_present |= cpu_to_le32(1 << IEEE80211_RADIOTAP_RATE); 306 rate = sband->bitrates[info->status.rates[0].idx].bitrate; 307 *pos = DIV_ROUND_UP(rate, 5 * (1 << shift)); 308 /* padding for tx flags */ 309 pos += 2; 310 } 311 312 /* IEEE80211_RADIOTAP_TX_FLAGS */ 313 txflags = 0; 314 if (!(info->flags & IEEE80211_TX_STAT_ACK) && 315 !is_multicast_ether_addr(hdr->addr1)) 316 txflags |= IEEE80211_RADIOTAP_F_TX_FAIL; 317 318 if (info->status.rates[0].flags & IEEE80211_TX_RC_USE_CTS_PROTECT) 319 txflags |= IEEE80211_RADIOTAP_F_TX_CTS; 320 if (info->status.rates[0].flags & IEEE80211_TX_RC_USE_RTS_CTS) 321 txflags |= IEEE80211_RADIOTAP_F_TX_RTS; 322 323 put_unaligned_le16(txflags, pos); 324 pos += 2; 325 326 /* IEEE80211_RADIOTAP_DATA_RETRIES */ 327 /* for now report the total retry_count */ 328 *pos = retry_count; 329 pos++; 330 331 if (info->status.rates[0].idx < 0) 332 return; 333 334 /* IEEE80211_RADIOTAP_MCS 335 * IEEE80211_RADIOTAP_VHT */ 336 if (info->status.rates[0].flags & IEEE80211_TX_RC_MCS) { 337 rthdr->it_present |= cpu_to_le32(1 << IEEE80211_RADIOTAP_MCS); 338 pos[0] = IEEE80211_RADIOTAP_MCS_HAVE_MCS | 339 IEEE80211_RADIOTAP_MCS_HAVE_GI | 340 IEEE80211_RADIOTAP_MCS_HAVE_BW; 341 if (info->status.rates[0].flags & IEEE80211_TX_RC_SHORT_GI) 342 pos[1] |= IEEE80211_RADIOTAP_MCS_SGI; 343 if (info->status.rates[0].flags & IEEE80211_TX_RC_40_MHZ_WIDTH) 344 pos[1] |= IEEE80211_RADIOTAP_MCS_BW_40; 345 if (info->status.rates[0].flags & IEEE80211_TX_RC_GREEN_FIELD) 346 pos[1] |= IEEE80211_RADIOTAP_MCS_FMT_GF; 347 pos[2] = info->status.rates[0].idx; 348 pos += 3; 349 } else if (info->status.rates[0].flags & IEEE80211_TX_RC_VHT_MCS) { 350 u16 known = local->hw.radiotap_vht_details & 351 (IEEE80211_RADIOTAP_VHT_KNOWN_GI | 352 IEEE80211_RADIOTAP_VHT_KNOWN_BANDWIDTH); 353 354 rthdr->it_present |= cpu_to_le32(1 << IEEE80211_RADIOTAP_VHT); 355 356 /* required alignment from rthdr */ 357 pos = (u8 *)rthdr + ALIGN(pos - (u8 *)rthdr, 2); 358 359 /* u16 known - IEEE80211_RADIOTAP_VHT_KNOWN_* */ 360 put_unaligned_le16(known, pos); 361 pos += 2; 362 363 /* u8 flags - IEEE80211_RADIOTAP_VHT_FLAG_* */ 364 if (info->status.rates[0].flags & IEEE80211_TX_RC_SHORT_GI) 365 *pos |= IEEE80211_RADIOTAP_VHT_FLAG_SGI; 366 pos++; 367 368 /* u8 bandwidth */ 369 if (info->status.rates[0].flags & IEEE80211_TX_RC_40_MHZ_WIDTH) 370 *pos = 1; 371 else if (info->status.rates[0].flags & IEEE80211_TX_RC_80_MHZ_WIDTH) 372 *pos = 4; 373 else if (info->status.rates[0].flags & IEEE80211_TX_RC_160_MHZ_WIDTH) 374 *pos = 11; 375 else /* IEEE80211_TX_RC_{20_MHZ_WIDTH,FIXME:DUP_DATA} */ 376 *pos = 0; 377 pos++; 378 379 /* u8 mcs_nss[4] */ 380 *pos = (ieee80211_rate_get_vht_mcs(&info->status.rates[0]) << 4) | 381 ieee80211_rate_get_vht_nss(&info->status.rates[0]); 382 pos += 4; 383 384 /* u8 coding */ 385 pos++; 386 /* u8 group_id */ 387 pos++; 388 /* u16 partial_aid */ 389 pos += 2; 390 } 391 } 392 393 static void ieee80211_report_used_skb(struct ieee80211_local *local, 394 struct sk_buff *skb, bool dropped) 395 { 396 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 397 struct ieee80211_hdr *hdr = (void *)skb->data; 398 bool acked = info->flags & IEEE80211_TX_STAT_ACK; 399 400 if (dropped) 401 acked = false; 402 403 if (info->flags & (IEEE80211_TX_INTFL_NL80211_FRAME_TX | 404 IEEE80211_TX_INTFL_MLME_CONN_TX)) { 405 struct ieee80211_sub_if_data *sdata = NULL; 406 struct ieee80211_sub_if_data *iter_sdata; 407 u64 cookie = (unsigned long)skb; 408 409 rcu_read_lock(); 410 411 if (skb->dev) { 412 list_for_each_entry_rcu(iter_sdata, &local->interfaces, 413 list) { 414 if (!iter_sdata->dev) 415 continue; 416 417 if (skb->dev == iter_sdata->dev) { 418 sdata = iter_sdata; 419 break; 420 } 421 } 422 } else { 423 sdata = rcu_dereference(local->p2p_sdata); 424 } 425 426 if (!sdata) { 427 skb->dev = NULL; 428 } else if (info->flags & IEEE80211_TX_INTFL_MLME_CONN_TX) { 429 ieee80211_mgd_conn_tx_status(sdata, hdr->frame_control, 430 acked); 431 } else if (ieee80211_is_nullfunc(hdr->frame_control) || 432 ieee80211_is_qos_nullfunc(hdr->frame_control)) { 433 cfg80211_probe_status(sdata->dev, hdr->addr1, 434 cookie, acked, GFP_ATOMIC); 435 } else { 436 cfg80211_mgmt_tx_status(&sdata->wdev, cookie, skb->data, 437 skb->len, acked, GFP_ATOMIC); 438 } 439 440 rcu_read_unlock(); 441 } 442 443 if (unlikely(info->ack_frame_id)) { 444 struct sk_buff *ack_skb; 445 unsigned long flags; 446 447 spin_lock_irqsave(&local->ack_status_lock, flags); 448 ack_skb = idr_find(&local->ack_status_frames, 449 info->ack_frame_id); 450 if (ack_skb) 451 idr_remove(&local->ack_status_frames, 452 info->ack_frame_id); 453 spin_unlock_irqrestore(&local->ack_status_lock, flags); 454 455 if (ack_skb) { 456 if (!dropped) { 457 /* consumes ack_skb */ 458 skb_complete_wifi_ack(ack_skb, acked); 459 } else { 460 dev_kfree_skb_any(ack_skb); 461 } 462 } 463 } 464 } 465 466 /* 467 * Measure Tx frame completion and removal time for Tx latency statistics 468 * calculation. A single Tx frame latency should be measured from when it 469 * is entering the Kernel until we receive Tx complete confirmation indication 470 * and remove the skb. 471 */ 472 static void ieee80211_tx_latency_end_msrmnt(struct ieee80211_local *local, 473 struct sk_buff *skb, 474 struct sta_info *sta, 475 struct ieee80211_hdr *hdr) 476 { 477 u32 msrmnt; 478 u16 tid; 479 u8 *qc; 480 int i, bin_range_count; 481 u32 *bin_ranges; 482 __le16 fc; 483 struct ieee80211_tx_latency_stat *tx_lat; 484 struct ieee80211_tx_latency_bin_ranges *tx_latency; 485 ktime_t skb_arv = skb->tstamp; 486 487 tx_latency = rcu_dereference(local->tx_latency); 488 489 /* assert Tx latency stats are enabled & frame arrived when enabled */ 490 if (!tx_latency || !ktime_to_ns(skb_arv)) 491 return; 492 493 fc = hdr->frame_control; 494 495 if (!ieee80211_is_data(fc)) /* make sure it is a data frame */ 496 return; 497 498 /* get frame tid */ 499 if (ieee80211_is_data_qos(hdr->frame_control)) { 500 qc = ieee80211_get_qos_ctl(hdr); 501 tid = qc[0] & IEEE80211_QOS_CTL_TID_MASK; 502 } else { 503 tid = 0; 504 } 505 506 tx_lat = &sta->tx_lat[tid]; 507 508 /* Calculate the latency */ 509 msrmnt = ktime_to_ms(ktime_sub(ktime_get(), skb_arv)); 510 511 if (tx_lat->max < msrmnt) /* update stats */ 512 tx_lat->max = msrmnt; 513 tx_lat->counter++; 514 tx_lat->sum += msrmnt; 515 516 if (!tx_lat->bins) /* bins not activated */ 517 return; 518 519 /* count how many Tx frames transmitted with the appropriate latency */ 520 bin_range_count = tx_latency->n_ranges; 521 bin_ranges = tx_latency->ranges; 522 523 for (i = 0; i < bin_range_count; i++) { 524 if (msrmnt <= bin_ranges[i]) { 525 tx_lat->bins[i]++; 526 break; 527 } 528 } 529 if (i == bin_range_count) /* msrmnt is bigger than the biggest range */ 530 tx_lat->bins[i]++; 531 } 532 533 /* 534 * Use a static threshold for now, best value to be determined 535 * by testing ... 536 * Should it depend on: 537 * - on # of retransmissions 538 * - current throughput (higher value for higher tpt)? 539 */ 540 #define STA_LOST_PKT_THRESHOLD 50 541 #define STA_LOST_TDLS_PKT_THRESHOLD 10 542 #define STA_LOST_TDLS_PKT_TIME (10*HZ) /* 10secs since last ACK */ 543 544 static void ieee80211_lost_packet(struct sta_info *sta, struct sk_buff *skb) 545 { 546 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 547 548 /* This packet was aggregated but doesn't carry status info */ 549 if ((info->flags & IEEE80211_TX_CTL_AMPDU) && 550 !(info->flags & IEEE80211_TX_STAT_AMPDU)) 551 return; 552 553 sta->lost_packets++; 554 if (!sta->sta.tdls && sta->lost_packets < STA_LOST_PKT_THRESHOLD) 555 return; 556 557 /* 558 * If we're in TDLS mode, make sure that all STA_LOST_TDLS_PKT_THRESHOLD 559 * of the last packets were lost, and that no ACK was received in the 560 * last STA_LOST_TDLS_PKT_TIME ms, before triggering the CQM packet-loss 561 * mechanism. 562 */ 563 if (sta->sta.tdls && 564 (sta->lost_packets < STA_LOST_TDLS_PKT_THRESHOLD || 565 time_before(jiffies, 566 sta->last_tdls_pkt_time + STA_LOST_TDLS_PKT_TIME))) 567 return; 568 569 cfg80211_cqm_pktloss_notify(sta->sdata->dev, sta->sta.addr, 570 sta->lost_packets, GFP_ATOMIC); 571 sta->lost_packets = 0; 572 } 573 574 void ieee80211_tx_status(struct ieee80211_hw *hw, struct sk_buff *skb) 575 { 576 struct sk_buff *skb2; 577 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data; 578 struct ieee80211_local *local = hw_to_local(hw); 579 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 580 __le16 fc; 581 struct ieee80211_supported_band *sband; 582 struct ieee80211_sub_if_data *sdata; 583 struct net_device *prev_dev = NULL; 584 struct sta_info *sta, *tmp; 585 int retry_count = -1, i; 586 int rates_idx = -1; 587 bool send_to_cooked; 588 bool acked; 589 struct ieee80211_bar *bar; 590 int rtap_len; 591 int shift = 0; 592 593 for (i = 0; i < IEEE80211_TX_MAX_RATES; i++) { 594 if ((info->flags & IEEE80211_TX_CTL_AMPDU) && 595 !(info->flags & IEEE80211_TX_STAT_AMPDU)) { 596 /* just the first aggr frame carry status info */ 597 info->status.rates[i].idx = -1; 598 info->status.rates[i].count = 0; 599 break; 600 } else if (info->status.rates[i].idx < 0) { 601 break; 602 } else if (i >= hw->max_report_rates) { 603 /* the HW cannot have attempted that rate */ 604 info->status.rates[i].idx = -1; 605 info->status.rates[i].count = 0; 606 break; 607 } 608 609 retry_count += info->status.rates[i].count; 610 } 611 rates_idx = i - 1; 612 613 if (retry_count < 0) 614 retry_count = 0; 615 616 rcu_read_lock(); 617 618 sband = local->hw.wiphy->bands[info->band]; 619 fc = hdr->frame_control; 620 621 for_each_sta_info(local, hdr->addr1, sta, tmp) { 622 /* skip wrong virtual interface */ 623 if (!ether_addr_equal(hdr->addr2, sta->sdata->vif.addr)) 624 continue; 625 626 shift = ieee80211_vif_get_shift(&sta->sdata->vif); 627 628 if (info->flags & IEEE80211_TX_STATUS_EOSP) 629 clear_sta_flag(sta, WLAN_STA_SP); 630 631 acked = !!(info->flags & IEEE80211_TX_STAT_ACK); 632 if (!acked && test_sta_flag(sta, WLAN_STA_PS_STA)) { 633 /* 634 * The STA is in power save mode, so assume 635 * that this TX packet failed because of that. 636 */ 637 ieee80211_handle_filtered_frame(local, sta, skb); 638 rcu_read_unlock(); 639 return; 640 } 641 642 /* mesh Peer Service Period support */ 643 if (ieee80211_vif_is_mesh(&sta->sdata->vif) && 644 ieee80211_is_data_qos(fc)) 645 ieee80211_mpsp_trigger_process( 646 ieee80211_get_qos_ctl(hdr), 647 sta, true, acked); 648 649 if ((local->hw.flags & IEEE80211_HW_HAS_RATE_CONTROL) && 650 (ieee80211_is_data(hdr->frame_control)) && 651 (rates_idx != -1)) 652 sta->last_tx_rate = info->status.rates[rates_idx]; 653 654 if ((info->flags & IEEE80211_TX_STAT_AMPDU_NO_BACK) && 655 (ieee80211_is_data_qos(fc))) { 656 u16 tid, ssn; 657 u8 *qc; 658 659 qc = ieee80211_get_qos_ctl(hdr); 660 tid = qc[0] & 0xf; 661 ssn = ((le16_to_cpu(hdr->seq_ctrl) + 0x10) 662 & IEEE80211_SCTL_SEQ); 663 ieee80211_send_bar(&sta->sdata->vif, hdr->addr1, 664 tid, ssn); 665 } 666 667 if (!acked && ieee80211_is_back_req(fc)) { 668 u16 tid, control; 669 670 /* 671 * BAR failed, store the last SSN and retry sending 672 * the BAR when the next unicast transmission on the 673 * same TID succeeds. 674 */ 675 bar = (struct ieee80211_bar *) skb->data; 676 control = le16_to_cpu(bar->control); 677 if (!(control & IEEE80211_BAR_CTRL_MULTI_TID)) { 678 u16 ssn = le16_to_cpu(bar->start_seq_num); 679 680 tid = (control & 681 IEEE80211_BAR_CTRL_TID_INFO_MASK) >> 682 IEEE80211_BAR_CTRL_TID_INFO_SHIFT; 683 684 ieee80211_set_bar_pending(sta, tid, ssn); 685 } 686 } 687 688 if (info->flags & IEEE80211_TX_STAT_TX_FILTERED) { 689 ieee80211_handle_filtered_frame(local, sta, skb); 690 rcu_read_unlock(); 691 return; 692 } else { 693 if (!acked) 694 sta->tx_retry_failed++; 695 sta->tx_retry_count += retry_count; 696 } 697 698 rate_control_tx_status(local, sband, sta, skb); 699 if (ieee80211_vif_is_mesh(&sta->sdata->vif)) 700 ieee80211s_update_metric(local, sta, skb); 701 702 if (!(info->flags & IEEE80211_TX_CTL_INJECTED) && acked) 703 ieee80211_frame_acked(sta, skb); 704 705 if ((sta->sdata->vif.type == NL80211_IFTYPE_STATION) && 706 (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS)) 707 ieee80211_sta_tx_notify(sta->sdata, (void *) skb->data, acked); 708 709 if (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS) { 710 if (info->flags & IEEE80211_TX_STAT_ACK) { 711 if (sta->lost_packets) 712 sta->lost_packets = 0; 713 714 /* Track when last TDLS packet was ACKed */ 715 if (test_sta_flag(sta, WLAN_STA_TDLS_PEER_AUTH)) 716 sta->last_tdls_pkt_time = jiffies; 717 } else { 718 ieee80211_lost_packet(sta, skb); 719 } 720 } 721 722 if (acked) 723 sta->last_ack_signal = info->status.ack_signal; 724 725 /* 726 * Measure frame removal for tx latency 727 * statistics calculation 728 */ 729 ieee80211_tx_latency_end_msrmnt(local, skb, sta, hdr); 730 } 731 732 rcu_read_unlock(); 733 734 ieee80211_led_tx(local); 735 736 /* SNMP counters 737 * Fragments are passed to low-level drivers as separate skbs, so these 738 * are actually fragments, not frames. Update frame counters only for 739 * the first fragment of the frame. */ 740 if (info->flags & IEEE80211_TX_STAT_ACK) { 741 if (ieee80211_is_first_frag(hdr->seq_ctrl)) { 742 local->dot11TransmittedFrameCount++; 743 if (is_multicast_ether_addr(hdr->addr1)) 744 local->dot11MulticastTransmittedFrameCount++; 745 if (retry_count > 0) 746 local->dot11RetryCount++; 747 if (retry_count > 1) 748 local->dot11MultipleRetryCount++; 749 } 750 751 /* This counter shall be incremented for an acknowledged MPDU 752 * with an individual address in the address 1 field or an MPDU 753 * with a multicast address in the address 1 field of type Data 754 * or Management. */ 755 if (!is_multicast_ether_addr(hdr->addr1) || 756 ieee80211_is_data(fc) || 757 ieee80211_is_mgmt(fc)) 758 local->dot11TransmittedFragmentCount++; 759 } else { 760 if (ieee80211_is_first_frag(hdr->seq_ctrl)) 761 local->dot11FailedCount++; 762 } 763 764 if (ieee80211_is_nullfunc(fc) && ieee80211_has_pm(fc) && 765 (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS) && 766 !(info->flags & IEEE80211_TX_CTL_INJECTED) && 767 local->ps_sdata && !(local->scanning)) { 768 if (info->flags & IEEE80211_TX_STAT_ACK) { 769 local->ps_sdata->u.mgd.flags |= 770 IEEE80211_STA_NULLFUNC_ACKED; 771 } else 772 mod_timer(&local->dynamic_ps_timer, jiffies + 773 msecs_to_jiffies(10)); 774 } 775 776 ieee80211_report_used_skb(local, skb, false); 777 778 /* this was a transmitted frame, but now we want to reuse it */ 779 skb_orphan(skb); 780 781 /* Need to make a copy before skb->cb gets cleared */ 782 send_to_cooked = !!(info->flags & IEEE80211_TX_CTL_INJECTED) || 783 !(ieee80211_is_data(fc)); 784 785 /* 786 * This is a bit racy but we can avoid a lot of work 787 * with this test... 788 */ 789 if (!local->monitors && (!send_to_cooked || !local->cooked_mntrs)) { 790 dev_kfree_skb(skb); 791 return; 792 } 793 794 /* send frame to monitor interfaces now */ 795 rtap_len = ieee80211_tx_radiotap_len(info); 796 if (WARN_ON_ONCE(skb_headroom(skb) < rtap_len)) { 797 pr_err("ieee80211_tx_status: headroom too small\n"); 798 dev_kfree_skb(skb); 799 return; 800 } 801 ieee80211_add_tx_radiotap_header(local, sband, skb, retry_count, 802 rtap_len, shift); 803 804 /* XXX: is this sufficient for BPF? */ 805 skb_set_mac_header(skb, 0); 806 skb->ip_summed = CHECKSUM_UNNECESSARY; 807 skb->pkt_type = PACKET_OTHERHOST; 808 skb->protocol = htons(ETH_P_802_2); 809 memset(skb->cb, 0, sizeof(skb->cb)); 810 811 rcu_read_lock(); 812 list_for_each_entry_rcu(sdata, &local->interfaces, list) { 813 if (sdata->vif.type == NL80211_IFTYPE_MONITOR) { 814 if (!ieee80211_sdata_running(sdata)) 815 continue; 816 817 if ((sdata->u.mntr_flags & MONITOR_FLAG_COOK_FRAMES) && 818 !send_to_cooked) 819 continue; 820 821 if (prev_dev) { 822 skb2 = skb_clone(skb, GFP_ATOMIC); 823 if (skb2) { 824 skb2->dev = prev_dev; 825 netif_rx(skb2); 826 } 827 } 828 829 prev_dev = sdata->dev; 830 } 831 } 832 if (prev_dev) { 833 skb->dev = prev_dev; 834 netif_rx(skb); 835 skb = NULL; 836 } 837 rcu_read_unlock(); 838 dev_kfree_skb(skb); 839 } 840 EXPORT_SYMBOL(ieee80211_tx_status); 841 842 void ieee80211_report_low_ack(struct ieee80211_sta *pubsta, u32 num_packets) 843 { 844 struct sta_info *sta = container_of(pubsta, struct sta_info, sta); 845 cfg80211_cqm_pktloss_notify(sta->sdata->dev, sta->sta.addr, 846 num_packets, GFP_ATOMIC); 847 } 848 EXPORT_SYMBOL(ieee80211_report_low_ack); 849 850 void ieee80211_free_txskb(struct ieee80211_hw *hw, struct sk_buff *skb) 851 { 852 struct ieee80211_local *local = hw_to_local(hw); 853 854 ieee80211_report_used_skb(local, skb, true); 855 dev_kfree_skb_any(skb); 856 } 857 EXPORT_SYMBOL(ieee80211_free_txskb); 858 859 void ieee80211_purge_tx_queue(struct ieee80211_hw *hw, 860 struct sk_buff_head *skbs) 861 { 862 struct sk_buff *skb; 863 864 while ((skb = __skb_dequeue(skbs))) 865 ieee80211_free_txskb(hw, skb); 866 } 867