1 /* 2 * Copyright 2002-2005, Instant802 Networks, Inc. 3 * Copyright 2005-2006, Devicescape Software, Inc. 4 * Copyright 2006-2007 Jiri Benc <jbenc@suse.cz> 5 * Copyright 2008-2010 Johannes Berg <johannes@sipsolutions.net> 6 * 7 * This program is free software; you can redistribute it and/or modify 8 * it under the terms of the GNU General Public License version 2 as 9 * published by the Free Software Foundation. 10 */ 11 12 #include <linux/export.h> 13 #include <linux/etherdevice.h> 14 #include <linux/time.h> 15 #include <net/mac80211.h> 16 #include <asm/unaligned.h> 17 #include "ieee80211_i.h" 18 #include "rate.h" 19 #include "mesh.h" 20 #include "led.h" 21 #include "wme.h" 22 23 24 void ieee80211_tx_status_irqsafe(struct ieee80211_hw *hw, 25 struct sk_buff *skb) 26 { 27 struct ieee80211_local *local = hw_to_local(hw); 28 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 29 int tmp; 30 31 skb->pkt_type = IEEE80211_TX_STATUS_MSG; 32 skb_queue_tail(info->flags & IEEE80211_TX_CTL_REQ_TX_STATUS ? 33 &local->skb_queue : &local->skb_queue_unreliable, skb); 34 tmp = skb_queue_len(&local->skb_queue) + 35 skb_queue_len(&local->skb_queue_unreliable); 36 while (tmp > IEEE80211_IRQSAFE_QUEUE_LIMIT && 37 (skb = skb_dequeue(&local->skb_queue_unreliable))) { 38 ieee80211_free_txskb(hw, skb); 39 tmp--; 40 I802_DEBUG_INC(local->tx_status_drop); 41 } 42 tasklet_schedule(&local->tasklet); 43 } 44 EXPORT_SYMBOL(ieee80211_tx_status_irqsafe); 45 46 static void ieee80211_handle_filtered_frame(struct ieee80211_local *local, 47 struct sta_info *sta, 48 struct sk_buff *skb) 49 { 50 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 51 struct ieee80211_hdr *hdr = (void *)skb->data; 52 int ac; 53 54 /* 55 * This skb 'survived' a round-trip through the driver, and 56 * hopefully the driver didn't mangle it too badly. However, 57 * we can definitely not rely on the control information 58 * being correct. Clear it so we don't get junk there, and 59 * indicate that it needs new processing, but must not be 60 * modified/encrypted again. 61 */ 62 memset(&info->control, 0, sizeof(info->control)); 63 64 info->control.jiffies = jiffies; 65 info->control.vif = &sta->sdata->vif; 66 info->flags |= IEEE80211_TX_INTFL_NEED_TXPROCESSING | 67 IEEE80211_TX_INTFL_RETRANSMISSION; 68 info->flags &= ~IEEE80211_TX_TEMPORARY_FLAGS; 69 70 sta->tx_filtered_count++; 71 72 /* 73 * Clear more-data bit on filtered frames, it might be set 74 * but later frames might time out so it might have to be 75 * clear again ... It's all rather unlikely (this frame 76 * should time out first, right?) but let's not confuse 77 * peers unnecessarily. 78 */ 79 if (hdr->frame_control & cpu_to_le16(IEEE80211_FCTL_MOREDATA)) 80 hdr->frame_control &= ~cpu_to_le16(IEEE80211_FCTL_MOREDATA); 81 82 if (ieee80211_is_data_qos(hdr->frame_control)) { 83 u8 *p = ieee80211_get_qos_ctl(hdr); 84 int tid = *p & IEEE80211_QOS_CTL_TID_MASK; 85 86 /* 87 * Clear EOSP if set, this could happen e.g. 88 * if an absence period (us being a P2P GO) 89 * shortens the SP. 90 */ 91 if (*p & IEEE80211_QOS_CTL_EOSP) 92 *p &= ~IEEE80211_QOS_CTL_EOSP; 93 ac = ieee802_1d_to_ac[tid & 7]; 94 } else { 95 ac = IEEE80211_AC_BE; 96 } 97 98 /* 99 * Clear the TX filter mask for this STA when sending the next 100 * packet. If the STA went to power save mode, this will happen 101 * when it wakes up for the next time. 102 */ 103 set_sta_flag(sta, WLAN_STA_CLEAR_PS_FILT); 104 105 /* 106 * This code races in the following way: 107 * 108 * (1) STA sends frame indicating it will go to sleep and does so 109 * (2) hardware/firmware adds STA to filter list, passes frame up 110 * (3) hardware/firmware processes TX fifo and suppresses a frame 111 * (4) we get TX status before having processed the frame and 112 * knowing that the STA has gone to sleep. 113 * 114 * This is actually quite unlikely even when both those events are 115 * processed from interrupts coming in quickly after one another or 116 * even at the same time because we queue both TX status events and 117 * RX frames to be processed by a tasklet and process them in the 118 * same order that they were received or TX status last. Hence, there 119 * is no race as long as the frame RX is processed before the next TX 120 * status, which drivers can ensure, see below. 121 * 122 * Note that this can only happen if the hardware or firmware can 123 * actually add STAs to the filter list, if this is done by the 124 * driver in response to set_tim() (which will only reduce the race 125 * this whole filtering tries to solve, not completely solve it) 126 * this situation cannot happen. 127 * 128 * To completely solve this race drivers need to make sure that they 129 * (a) don't mix the irq-safe/not irq-safe TX status/RX processing 130 * functions and 131 * (b) always process RX events before TX status events if ordering 132 * can be unknown, for example with different interrupt status 133 * bits. 134 * (c) if PS mode transitions are manual (i.e. the flag 135 * %IEEE80211_HW_AP_LINK_PS is set), always process PS state 136 * changes before calling TX status events if ordering can be 137 * unknown. 138 */ 139 if (test_sta_flag(sta, WLAN_STA_PS_STA) && 140 skb_queue_len(&sta->tx_filtered[ac]) < STA_MAX_TX_BUFFER) { 141 skb_queue_tail(&sta->tx_filtered[ac], skb); 142 sta_info_recalc_tim(sta); 143 144 if (!timer_pending(&local->sta_cleanup)) 145 mod_timer(&local->sta_cleanup, 146 round_jiffies(jiffies + 147 STA_INFO_CLEANUP_INTERVAL)); 148 return; 149 } 150 151 if (!test_sta_flag(sta, WLAN_STA_PS_STA) && 152 !(info->flags & IEEE80211_TX_INTFL_RETRIED)) { 153 /* Software retry the packet once */ 154 info->flags |= IEEE80211_TX_INTFL_RETRIED; 155 ieee80211_add_pending_skb(local, skb); 156 return; 157 } 158 159 ps_dbg_ratelimited(sta->sdata, 160 "dropped TX filtered frame, queue_len=%d PS=%d @%lu\n", 161 skb_queue_len(&sta->tx_filtered[ac]), 162 !!test_sta_flag(sta, WLAN_STA_PS_STA), jiffies); 163 ieee80211_free_txskb(&local->hw, skb); 164 } 165 166 static void ieee80211_check_pending_bar(struct sta_info *sta, u8 *addr, u8 tid) 167 { 168 struct tid_ampdu_tx *tid_tx; 169 170 tid_tx = rcu_dereference(sta->ampdu_mlme.tid_tx[tid]); 171 if (!tid_tx || !tid_tx->bar_pending) 172 return; 173 174 tid_tx->bar_pending = false; 175 ieee80211_send_bar(&sta->sdata->vif, addr, tid, tid_tx->failed_bar_ssn); 176 } 177 178 static void ieee80211_frame_acked(struct sta_info *sta, struct sk_buff *skb) 179 { 180 struct ieee80211_mgmt *mgmt = (void *) skb->data; 181 struct ieee80211_local *local = sta->local; 182 struct ieee80211_sub_if_data *sdata = sta->sdata; 183 184 if (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS) 185 sta->last_rx = jiffies; 186 187 if (ieee80211_is_data_qos(mgmt->frame_control)) { 188 struct ieee80211_hdr *hdr = (void *) skb->data; 189 u8 *qc = ieee80211_get_qos_ctl(hdr); 190 u16 tid = qc[0] & 0xf; 191 192 ieee80211_check_pending_bar(sta, hdr->addr1, tid); 193 } 194 195 if (ieee80211_is_action(mgmt->frame_control) && 196 mgmt->u.action.category == WLAN_CATEGORY_HT && 197 mgmt->u.action.u.ht_smps.action == WLAN_HT_ACTION_SMPS && 198 ieee80211_sdata_running(sdata)) { 199 enum ieee80211_smps_mode smps_mode; 200 201 switch (mgmt->u.action.u.ht_smps.smps_control) { 202 case WLAN_HT_SMPS_CONTROL_DYNAMIC: 203 smps_mode = IEEE80211_SMPS_DYNAMIC; 204 break; 205 case WLAN_HT_SMPS_CONTROL_STATIC: 206 smps_mode = IEEE80211_SMPS_STATIC; 207 break; 208 case WLAN_HT_SMPS_CONTROL_DISABLED: 209 default: /* shouldn't happen since we don't send that */ 210 smps_mode = IEEE80211_SMPS_OFF; 211 break; 212 } 213 214 if (sdata->vif.type == NL80211_IFTYPE_STATION) { 215 /* 216 * This update looks racy, but isn't -- if we come 217 * here we've definitely got a station that we're 218 * talking to, and on a managed interface that can 219 * only be the AP. And the only other place updating 220 * this variable in managed mode is before association. 221 */ 222 sdata->smps_mode = smps_mode; 223 ieee80211_queue_work(&local->hw, &sdata->recalc_smps); 224 } else if (sdata->vif.type == NL80211_IFTYPE_AP || 225 sdata->vif.type == NL80211_IFTYPE_AP_VLAN) { 226 sta->known_smps_mode = smps_mode; 227 } 228 } 229 } 230 231 static void ieee80211_set_bar_pending(struct sta_info *sta, u8 tid, u16 ssn) 232 { 233 struct tid_ampdu_tx *tid_tx; 234 235 tid_tx = rcu_dereference(sta->ampdu_mlme.tid_tx[tid]); 236 if (!tid_tx) 237 return; 238 239 tid_tx->failed_bar_ssn = ssn; 240 tid_tx->bar_pending = true; 241 } 242 243 static int ieee80211_tx_radiotap_len(struct ieee80211_tx_info *info) 244 { 245 int len = sizeof(struct ieee80211_radiotap_header); 246 247 /* IEEE80211_RADIOTAP_RATE rate */ 248 if (info->status.rates[0].idx >= 0 && 249 !(info->status.rates[0].flags & (IEEE80211_TX_RC_MCS | 250 IEEE80211_TX_RC_VHT_MCS))) 251 len += 2; 252 253 /* IEEE80211_RADIOTAP_TX_FLAGS */ 254 len += 2; 255 256 /* IEEE80211_RADIOTAP_DATA_RETRIES */ 257 len += 1; 258 259 /* IEEE80211_RADIOTAP_MCS 260 * IEEE80211_RADIOTAP_VHT */ 261 if (info->status.rates[0].idx >= 0) { 262 if (info->status.rates[0].flags & IEEE80211_TX_RC_MCS) 263 len += 3; 264 else if (info->status.rates[0].flags & IEEE80211_TX_RC_VHT_MCS) 265 len = ALIGN(len, 2) + 12; 266 } 267 268 return len; 269 } 270 271 static void 272 ieee80211_add_tx_radiotap_header(struct ieee80211_local *local, 273 struct ieee80211_supported_band *sband, 274 struct sk_buff *skb, int retry_count, 275 int rtap_len, int shift) 276 { 277 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 278 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data; 279 struct ieee80211_radiotap_header *rthdr; 280 unsigned char *pos; 281 u16 txflags; 282 283 rthdr = (struct ieee80211_radiotap_header *) skb_push(skb, rtap_len); 284 285 memset(rthdr, 0, rtap_len); 286 rthdr->it_len = cpu_to_le16(rtap_len); 287 rthdr->it_present = 288 cpu_to_le32((1 << IEEE80211_RADIOTAP_TX_FLAGS) | 289 (1 << IEEE80211_RADIOTAP_DATA_RETRIES)); 290 pos = (unsigned char *)(rthdr + 1); 291 292 /* 293 * XXX: Once radiotap gets the bitmap reset thing the vendor 294 * extensions proposal contains, we can actually report 295 * the whole set of tries we did. 296 */ 297 298 /* IEEE80211_RADIOTAP_RATE */ 299 if (info->status.rates[0].idx >= 0 && 300 !(info->status.rates[0].flags & (IEEE80211_TX_RC_MCS | 301 IEEE80211_TX_RC_VHT_MCS))) { 302 u16 rate; 303 304 rthdr->it_present |= cpu_to_le32(1 << IEEE80211_RADIOTAP_RATE); 305 rate = sband->bitrates[info->status.rates[0].idx].bitrate; 306 *pos = DIV_ROUND_UP(rate, 5 * (1 << shift)); 307 /* padding for tx flags */ 308 pos += 2; 309 } 310 311 /* IEEE80211_RADIOTAP_TX_FLAGS */ 312 txflags = 0; 313 if (!(info->flags & IEEE80211_TX_STAT_ACK) && 314 !is_multicast_ether_addr(hdr->addr1)) 315 txflags |= IEEE80211_RADIOTAP_F_TX_FAIL; 316 317 if (info->status.rates[0].flags & IEEE80211_TX_RC_USE_CTS_PROTECT) 318 txflags |= IEEE80211_RADIOTAP_F_TX_CTS; 319 if (info->status.rates[0].flags & IEEE80211_TX_RC_USE_RTS_CTS) 320 txflags |= IEEE80211_RADIOTAP_F_TX_RTS; 321 322 put_unaligned_le16(txflags, pos); 323 pos += 2; 324 325 /* IEEE80211_RADIOTAP_DATA_RETRIES */ 326 /* for now report the total retry_count */ 327 *pos = retry_count; 328 pos++; 329 330 if (info->status.rates[0].idx < 0) 331 return; 332 333 /* IEEE80211_RADIOTAP_MCS 334 * IEEE80211_RADIOTAP_VHT */ 335 if (info->status.rates[0].flags & IEEE80211_TX_RC_MCS) { 336 rthdr->it_present |= cpu_to_le32(1 << IEEE80211_RADIOTAP_MCS); 337 pos[0] = IEEE80211_RADIOTAP_MCS_HAVE_MCS | 338 IEEE80211_RADIOTAP_MCS_HAVE_GI | 339 IEEE80211_RADIOTAP_MCS_HAVE_BW; 340 if (info->status.rates[0].flags & IEEE80211_TX_RC_SHORT_GI) 341 pos[1] |= IEEE80211_RADIOTAP_MCS_SGI; 342 if (info->status.rates[0].flags & IEEE80211_TX_RC_40_MHZ_WIDTH) 343 pos[1] |= IEEE80211_RADIOTAP_MCS_BW_40; 344 if (info->status.rates[0].flags & IEEE80211_TX_RC_GREEN_FIELD) 345 pos[1] |= IEEE80211_RADIOTAP_MCS_FMT_GF; 346 pos[2] = info->status.rates[0].idx; 347 pos += 3; 348 } else if (info->status.rates[0].flags & IEEE80211_TX_RC_VHT_MCS) { 349 u16 known = local->hw.radiotap_vht_details & 350 (IEEE80211_RADIOTAP_VHT_KNOWN_GI | 351 IEEE80211_RADIOTAP_VHT_KNOWN_BANDWIDTH); 352 353 rthdr->it_present |= cpu_to_le32(1 << IEEE80211_RADIOTAP_VHT); 354 355 /* required alignment from rthdr */ 356 pos = (u8 *)rthdr + ALIGN(pos - (u8 *)rthdr, 2); 357 358 /* u16 known - IEEE80211_RADIOTAP_VHT_KNOWN_* */ 359 put_unaligned_le16(known, pos); 360 pos += 2; 361 362 /* u8 flags - IEEE80211_RADIOTAP_VHT_FLAG_* */ 363 if (info->status.rates[0].flags & IEEE80211_TX_RC_SHORT_GI) 364 *pos |= IEEE80211_RADIOTAP_VHT_FLAG_SGI; 365 pos++; 366 367 /* u8 bandwidth */ 368 if (info->status.rates[0].flags & IEEE80211_TX_RC_40_MHZ_WIDTH) 369 *pos = 1; 370 else if (info->status.rates[0].flags & IEEE80211_TX_RC_80_MHZ_WIDTH) 371 *pos = 4; 372 else if (info->status.rates[0].flags & IEEE80211_TX_RC_160_MHZ_WIDTH) 373 *pos = 11; 374 else /* IEEE80211_TX_RC_{20_MHZ_WIDTH,FIXME:DUP_DATA} */ 375 *pos = 0; 376 pos++; 377 378 /* u8 mcs_nss[4] */ 379 *pos = (ieee80211_rate_get_vht_mcs(&info->status.rates[0]) << 4) | 380 ieee80211_rate_get_vht_nss(&info->status.rates[0]); 381 pos += 4; 382 383 /* u8 coding */ 384 pos++; 385 /* u8 group_id */ 386 pos++; 387 /* u16 partial_aid */ 388 pos += 2; 389 } 390 } 391 392 static void ieee80211_report_used_skb(struct ieee80211_local *local, 393 struct sk_buff *skb, bool dropped) 394 { 395 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 396 struct ieee80211_hdr *hdr = (void *)skb->data; 397 bool acked = info->flags & IEEE80211_TX_STAT_ACK; 398 399 if (dropped) 400 acked = false; 401 402 if (info->flags & (IEEE80211_TX_INTFL_NL80211_FRAME_TX | 403 IEEE80211_TX_INTFL_MLME_CONN_TX)) { 404 struct ieee80211_sub_if_data *sdata = NULL; 405 struct ieee80211_sub_if_data *iter_sdata; 406 u64 cookie = (unsigned long)skb; 407 408 rcu_read_lock(); 409 410 if (skb->dev) { 411 list_for_each_entry_rcu(iter_sdata, &local->interfaces, 412 list) { 413 if (!iter_sdata->dev) 414 continue; 415 416 if (skb->dev == iter_sdata->dev) { 417 sdata = iter_sdata; 418 break; 419 } 420 } 421 } else { 422 sdata = rcu_dereference(local->p2p_sdata); 423 } 424 425 if (!sdata) { 426 skb->dev = NULL; 427 } else if (info->flags & IEEE80211_TX_INTFL_MLME_CONN_TX) { 428 ieee80211_mgd_conn_tx_status(sdata, hdr->frame_control, 429 acked); 430 } else if (ieee80211_is_nullfunc(hdr->frame_control) || 431 ieee80211_is_qos_nullfunc(hdr->frame_control)) { 432 cfg80211_probe_status(sdata->dev, hdr->addr1, 433 cookie, acked, GFP_ATOMIC); 434 } else { 435 cfg80211_mgmt_tx_status(&sdata->wdev, cookie, skb->data, 436 skb->len, acked, GFP_ATOMIC); 437 } 438 439 rcu_read_unlock(); 440 } 441 442 if (unlikely(info->ack_frame_id)) { 443 struct sk_buff *ack_skb; 444 unsigned long flags; 445 446 spin_lock_irqsave(&local->ack_status_lock, flags); 447 ack_skb = idr_find(&local->ack_status_frames, 448 info->ack_frame_id); 449 if (ack_skb) 450 idr_remove(&local->ack_status_frames, 451 info->ack_frame_id); 452 spin_unlock_irqrestore(&local->ack_status_lock, flags); 453 454 if (ack_skb) { 455 if (!dropped) { 456 /* consumes ack_skb */ 457 skb_complete_wifi_ack(ack_skb, acked); 458 } else { 459 dev_kfree_skb_any(ack_skb); 460 } 461 } 462 } 463 } 464 465 /* 466 * Measure Tx frame completion and removal time for Tx latency statistics 467 * calculation. A single Tx frame latency should be measured from when it 468 * is entering the Kernel until we receive Tx complete confirmation indication 469 * and remove the skb. 470 */ 471 static void ieee80211_tx_latency_end_msrmnt(struct ieee80211_local *local, 472 struct sk_buff *skb, 473 struct sta_info *sta, 474 struct ieee80211_hdr *hdr) 475 { 476 ktime_t skb_dprt; 477 struct timespec dprt_time; 478 u32 msrmnt; 479 u16 tid; 480 u8 *qc; 481 int i, bin_range_count; 482 u32 *bin_ranges; 483 __le16 fc; 484 struct ieee80211_tx_latency_stat *tx_lat; 485 struct ieee80211_tx_latency_bin_ranges *tx_latency; 486 ktime_t skb_arv = skb->tstamp; 487 488 tx_latency = rcu_dereference(local->tx_latency); 489 490 /* assert Tx latency stats are enabled & frame arrived when enabled */ 491 if (!tx_latency || !ktime_to_ns(skb_arv)) 492 return; 493 494 fc = hdr->frame_control; 495 496 if (!ieee80211_is_data(fc)) /* make sure it is a data frame */ 497 return; 498 499 /* get frame tid */ 500 if (ieee80211_is_data_qos(hdr->frame_control)) { 501 qc = ieee80211_get_qos_ctl(hdr); 502 tid = qc[0] & IEEE80211_QOS_CTL_TID_MASK; 503 } else { 504 tid = 0; 505 } 506 507 tx_lat = &sta->tx_lat[tid]; 508 509 ktime_get_ts(&dprt_time); /* time stamp completion time */ 510 skb_dprt = ktime_set(dprt_time.tv_sec, dprt_time.tv_nsec); 511 msrmnt = ktime_to_ms(ktime_sub(skb_dprt, skb_arv)); 512 513 if (tx_lat->max < msrmnt) /* update stats */ 514 tx_lat->max = msrmnt; 515 tx_lat->counter++; 516 tx_lat->sum += msrmnt; 517 518 if (!tx_lat->bins) /* bins not activated */ 519 return; 520 521 /* count how many Tx frames transmitted with the appropriate latency */ 522 bin_range_count = tx_latency->n_ranges; 523 bin_ranges = tx_latency->ranges; 524 525 for (i = 0; i < bin_range_count; i++) { 526 if (msrmnt <= bin_ranges[i]) { 527 tx_lat->bins[i]++; 528 break; 529 } 530 } 531 if (i == bin_range_count) /* msrmnt is bigger than the biggest range */ 532 tx_lat->bins[i]++; 533 } 534 535 /* 536 * Use a static threshold for now, best value to be determined 537 * by testing ... 538 * Should it depend on: 539 * - on # of retransmissions 540 * - current throughput (higher value for higher tpt)? 541 */ 542 #define STA_LOST_PKT_THRESHOLD 50 543 544 void ieee80211_tx_status(struct ieee80211_hw *hw, struct sk_buff *skb) 545 { 546 struct sk_buff *skb2; 547 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data; 548 struct ieee80211_local *local = hw_to_local(hw); 549 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 550 __le16 fc; 551 struct ieee80211_supported_band *sband; 552 struct ieee80211_sub_if_data *sdata; 553 struct net_device *prev_dev = NULL; 554 struct sta_info *sta, *tmp; 555 int retry_count = -1, i; 556 int rates_idx = -1; 557 bool send_to_cooked; 558 bool acked; 559 struct ieee80211_bar *bar; 560 int rtap_len; 561 int shift = 0; 562 563 for (i = 0; i < IEEE80211_TX_MAX_RATES; i++) { 564 if ((info->flags & IEEE80211_TX_CTL_AMPDU) && 565 !(info->flags & IEEE80211_TX_STAT_AMPDU)) { 566 /* just the first aggr frame carry status info */ 567 info->status.rates[i].idx = -1; 568 info->status.rates[i].count = 0; 569 break; 570 } else if (info->status.rates[i].idx < 0) { 571 break; 572 } else if (i >= hw->max_report_rates) { 573 /* the HW cannot have attempted that rate */ 574 info->status.rates[i].idx = -1; 575 info->status.rates[i].count = 0; 576 break; 577 } 578 579 retry_count += info->status.rates[i].count; 580 } 581 rates_idx = i - 1; 582 583 if (retry_count < 0) 584 retry_count = 0; 585 586 rcu_read_lock(); 587 588 sband = local->hw.wiphy->bands[info->band]; 589 fc = hdr->frame_control; 590 591 for_each_sta_info(local, hdr->addr1, sta, tmp) { 592 /* skip wrong virtual interface */ 593 if (!ether_addr_equal(hdr->addr2, sta->sdata->vif.addr)) 594 continue; 595 596 shift = ieee80211_vif_get_shift(&sta->sdata->vif); 597 598 if (info->flags & IEEE80211_TX_STATUS_EOSP) 599 clear_sta_flag(sta, WLAN_STA_SP); 600 601 acked = !!(info->flags & IEEE80211_TX_STAT_ACK); 602 if (!acked && test_sta_flag(sta, WLAN_STA_PS_STA)) { 603 /* 604 * The STA is in power save mode, so assume 605 * that this TX packet failed because of that. 606 */ 607 ieee80211_handle_filtered_frame(local, sta, skb); 608 rcu_read_unlock(); 609 return; 610 } 611 612 /* mesh Peer Service Period support */ 613 if (ieee80211_vif_is_mesh(&sta->sdata->vif) && 614 ieee80211_is_data_qos(fc)) 615 ieee80211_mpsp_trigger_process( 616 ieee80211_get_qos_ctl(hdr), 617 sta, true, acked); 618 619 if ((local->hw.flags & IEEE80211_HW_HAS_RATE_CONTROL) && 620 (ieee80211_is_data(hdr->frame_control)) && 621 (rates_idx != -1)) 622 sta->last_tx_rate = info->status.rates[rates_idx]; 623 624 if ((info->flags & IEEE80211_TX_STAT_AMPDU_NO_BACK) && 625 (ieee80211_is_data_qos(fc))) { 626 u16 tid, ssn; 627 u8 *qc; 628 629 qc = ieee80211_get_qos_ctl(hdr); 630 tid = qc[0] & 0xf; 631 ssn = ((le16_to_cpu(hdr->seq_ctrl) + 0x10) 632 & IEEE80211_SCTL_SEQ); 633 ieee80211_send_bar(&sta->sdata->vif, hdr->addr1, 634 tid, ssn); 635 } 636 637 if (!acked && ieee80211_is_back_req(fc)) { 638 u16 tid, control; 639 640 /* 641 * BAR failed, store the last SSN and retry sending 642 * the BAR when the next unicast transmission on the 643 * same TID succeeds. 644 */ 645 bar = (struct ieee80211_bar *) skb->data; 646 control = le16_to_cpu(bar->control); 647 if (!(control & IEEE80211_BAR_CTRL_MULTI_TID)) { 648 u16 ssn = le16_to_cpu(bar->start_seq_num); 649 650 tid = (control & 651 IEEE80211_BAR_CTRL_TID_INFO_MASK) >> 652 IEEE80211_BAR_CTRL_TID_INFO_SHIFT; 653 654 ieee80211_set_bar_pending(sta, tid, ssn); 655 } 656 } 657 658 if (info->flags & IEEE80211_TX_STAT_TX_FILTERED) { 659 ieee80211_handle_filtered_frame(local, sta, skb); 660 rcu_read_unlock(); 661 return; 662 } else { 663 if (!acked) 664 sta->tx_retry_failed++; 665 sta->tx_retry_count += retry_count; 666 } 667 668 rate_control_tx_status(local, sband, sta, skb); 669 if (ieee80211_vif_is_mesh(&sta->sdata->vif)) 670 ieee80211s_update_metric(local, sta, skb); 671 672 if (!(info->flags & IEEE80211_TX_CTL_INJECTED) && acked) 673 ieee80211_frame_acked(sta, skb); 674 675 if ((sta->sdata->vif.type == NL80211_IFTYPE_STATION) && 676 (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS)) 677 ieee80211_sta_tx_notify(sta->sdata, (void *) skb->data, acked); 678 679 if (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS) { 680 if (info->flags & IEEE80211_TX_STAT_ACK) { 681 if (sta->lost_packets) 682 sta->lost_packets = 0; 683 } else if (++sta->lost_packets >= STA_LOST_PKT_THRESHOLD) { 684 cfg80211_cqm_pktloss_notify(sta->sdata->dev, 685 sta->sta.addr, 686 sta->lost_packets, 687 GFP_ATOMIC); 688 sta->lost_packets = 0; 689 } 690 } 691 692 if (acked) 693 sta->last_ack_signal = info->status.ack_signal; 694 695 /* 696 * Measure frame removal for tx latency 697 * statistics calculation 698 */ 699 ieee80211_tx_latency_end_msrmnt(local, skb, sta, hdr); 700 } 701 702 rcu_read_unlock(); 703 704 ieee80211_led_tx(local); 705 706 /* SNMP counters 707 * Fragments are passed to low-level drivers as separate skbs, so these 708 * are actually fragments, not frames. Update frame counters only for 709 * the first fragment of the frame. */ 710 if (info->flags & IEEE80211_TX_STAT_ACK) { 711 if (ieee80211_is_first_frag(hdr->seq_ctrl)) { 712 local->dot11TransmittedFrameCount++; 713 if (is_multicast_ether_addr(hdr->addr1)) 714 local->dot11MulticastTransmittedFrameCount++; 715 if (retry_count > 0) 716 local->dot11RetryCount++; 717 if (retry_count > 1) 718 local->dot11MultipleRetryCount++; 719 } 720 721 /* This counter shall be incremented for an acknowledged MPDU 722 * with an individual address in the address 1 field or an MPDU 723 * with a multicast address in the address 1 field of type Data 724 * or Management. */ 725 if (!is_multicast_ether_addr(hdr->addr1) || 726 ieee80211_is_data(fc) || 727 ieee80211_is_mgmt(fc)) 728 local->dot11TransmittedFragmentCount++; 729 } else { 730 if (ieee80211_is_first_frag(hdr->seq_ctrl)) 731 local->dot11FailedCount++; 732 } 733 734 if (ieee80211_is_nullfunc(fc) && ieee80211_has_pm(fc) && 735 (local->hw.flags & IEEE80211_HW_REPORTS_TX_ACK_STATUS) && 736 !(info->flags & IEEE80211_TX_CTL_INJECTED) && 737 local->ps_sdata && !(local->scanning)) { 738 if (info->flags & IEEE80211_TX_STAT_ACK) { 739 local->ps_sdata->u.mgd.flags |= 740 IEEE80211_STA_NULLFUNC_ACKED; 741 } else 742 mod_timer(&local->dynamic_ps_timer, jiffies + 743 msecs_to_jiffies(10)); 744 } 745 746 ieee80211_report_used_skb(local, skb, false); 747 748 /* this was a transmitted frame, but now we want to reuse it */ 749 skb_orphan(skb); 750 751 /* Need to make a copy before skb->cb gets cleared */ 752 send_to_cooked = !!(info->flags & IEEE80211_TX_CTL_INJECTED) || 753 !(ieee80211_is_data(fc)); 754 755 /* 756 * This is a bit racy but we can avoid a lot of work 757 * with this test... 758 */ 759 if (!local->monitors && (!send_to_cooked || !local->cooked_mntrs)) { 760 dev_kfree_skb(skb); 761 return; 762 } 763 764 /* send frame to monitor interfaces now */ 765 rtap_len = ieee80211_tx_radiotap_len(info); 766 if (WARN_ON_ONCE(skb_headroom(skb) < rtap_len)) { 767 pr_err("ieee80211_tx_status: headroom too small\n"); 768 dev_kfree_skb(skb); 769 return; 770 } 771 ieee80211_add_tx_radiotap_header(local, sband, skb, retry_count, 772 rtap_len, shift); 773 774 /* XXX: is this sufficient for BPF? */ 775 skb_set_mac_header(skb, 0); 776 skb->ip_summed = CHECKSUM_UNNECESSARY; 777 skb->pkt_type = PACKET_OTHERHOST; 778 skb->protocol = htons(ETH_P_802_2); 779 memset(skb->cb, 0, sizeof(skb->cb)); 780 781 rcu_read_lock(); 782 list_for_each_entry_rcu(sdata, &local->interfaces, list) { 783 if (sdata->vif.type == NL80211_IFTYPE_MONITOR) { 784 if (!ieee80211_sdata_running(sdata)) 785 continue; 786 787 if ((sdata->u.mntr_flags & MONITOR_FLAG_COOK_FRAMES) && 788 !send_to_cooked) 789 continue; 790 791 if (prev_dev) { 792 skb2 = skb_clone(skb, GFP_ATOMIC); 793 if (skb2) { 794 skb2->dev = prev_dev; 795 netif_rx(skb2); 796 } 797 } 798 799 prev_dev = sdata->dev; 800 } 801 } 802 if (prev_dev) { 803 skb->dev = prev_dev; 804 netif_rx(skb); 805 skb = NULL; 806 } 807 rcu_read_unlock(); 808 dev_kfree_skb(skb); 809 } 810 EXPORT_SYMBOL(ieee80211_tx_status); 811 812 void ieee80211_report_low_ack(struct ieee80211_sta *pubsta, u32 num_packets) 813 { 814 struct sta_info *sta = container_of(pubsta, struct sta_info, sta); 815 cfg80211_cqm_pktloss_notify(sta->sdata->dev, sta->sta.addr, 816 num_packets, GFP_ATOMIC); 817 } 818 EXPORT_SYMBOL(ieee80211_report_low_ack); 819 820 void ieee80211_free_txskb(struct ieee80211_hw *hw, struct sk_buff *skb) 821 { 822 struct ieee80211_local *local = hw_to_local(hw); 823 824 ieee80211_report_used_skb(local, skb, true); 825 dev_kfree_skb_any(skb); 826 } 827 EXPORT_SYMBOL(ieee80211_free_txskb); 828 829 void ieee80211_purge_tx_queue(struct ieee80211_hw *hw, 830 struct sk_buff_head *skbs) 831 { 832 struct sk_buff *skb; 833 834 while ((skb = __skb_dequeue(skbs))) 835 ieee80211_free_txskb(hw, skb); 836 } 837