1 /****************************************************************************** 2 * 3 * This file is provided under a dual BSD/GPLv2 license. When using or 4 * redistributing this file, you may do so under either license. 5 * 6 * GPL LICENSE SUMMARY 7 * 8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. 9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH 11 * Copyright(c) 2018 Intel Corporation 12 * 13 * This program is free software; you can redistribute it and/or modify 14 * it under the terms of version 2 of the GNU General Public License as 15 * published by the Free Software Foundation. 16 * 17 * This program is distributed in the hope that it will be useful, but 18 * WITHOUT ANY WARRANTY; without even the implied warranty of 19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 20 * General Public License for more details. 21 * 22 * The full GNU General Public License is included in this distribution 23 * in the file called COPYING. 24 * 25 * Contact Information: 26 * Intel Linux Wireless <linuxwifi@intel.com> 27 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 28 * 29 * BSD LICENSE 30 * 31 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. 32 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 33 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH 34 * Copyright(c) 2018 Intel Corporation 35 * All rights reserved. 36 * 37 * Redistribution and use in source and binary forms, with or without 38 * modification, are permitted provided that the following conditions 39 * are met: 40 * 41 * * Redistributions of source code must retain the above copyright 42 * notice, this list of conditions and the following disclaimer. 43 * * Redistributions in binary form must reproduce the above copyright 44 * notice, this list of conditions and the following disclaimer in 45 * the documentation and/or other materials provided with the 46 * distribution. 47 * * Neither the name Intel Corporation nor the names of its 48 * contributors may be used to endorse or promote products derived 49 * from this software without specific prior written permission. 50 * 51 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 52 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 53 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 54 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 55 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 56 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 57 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 58 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 59 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 60 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 61 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 62 * 63 *****************************************************************************/ 64 #include <linux/kernel.h> 65 #include <linux/slab.h> 66 #include <linux/skbuff.h> 67 #include <linux/netdevice.h> 68 #include <linux/etherdevice.h> 69 #include <linux/ip.h> 70 #include <linux/if_arp.h> 71 #include <linux/time.h> 72 #include <net/mac80211.h> 73 #include <net/ieee80211_radiotap.h> 74 #include <net/tcp.h> 75 76 #include "iwl-op-mode.h" 77 #include "iwl-io.h" 78 #include "mvm.h" 79 #include "sta.h" 80 #include "time-event.h" 81 #include "iwl-eeprom-parse.h" 82 #include "iwl-phy-db.h" 83 #include "testmode.h" 84 #include "fw/error-dump.h" 85 #include "iwl-prph.h" 86 #include "iwl-nvm-parse.h" 87 88 static const struct ieee80211_iface_limit iwl_mvm_limits[] = { 89 { 90 .max = 1, 91 .types = BIT(NL80211_IFTYPE_STATION), 92 }, 93 { 94 .max = 1, 95 .types = BIT(NL80211_IFTYPE_AP) | 96 BIT(NL80211_IFTYPE_P2P_CLIENT) | 97 BIT(NL80211_IFTYPE_P2P_GO), 98 }, 99 { 100 .max = 1, 101 .types = BIT(NL80211_IFTYPE_P2P_DEVICE), 102 }, 103 }; 104 105 static const struct ieee80211_iface_combination iwl_mvm_iface_combinations[] = { 106 { 107 .num_different_channels = 2, 108 .max_interfaces = 3, 109 .limits = iwl_mvm_limits, 110 .n_limits = ARRAY_SIZE(iwl_mvm_limits), 111 }, 112 }; 113 114 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING 115 /* 116 * Use the reserved field to indicate magic values. 117 * these values will only be used internally by the driver, 118 * and won't make it to the fw (reserved will be 0). 119 * BC_FILTER_MAGIC_IP - configure the val of this attribute to 120 * be the vif's ip address. in case there is not a single 121 * ip address (0, or more than 1), this attribute will 122 * be skipped. 123 * BC_FILTER_MAGIC_MAC - set the val of this attribute to 124 * the LSB bytes of the vif's mac address 125 */ 126 enum { 127 BC_FILTER_MAGIC_NONE = 0, 128 BC_FILTER_MAGIC_IP, 129 BC_FILTER_MAGIC_MAC, 130 }; 131 132 static const struct iwl_fw_bcast_filter iwl_mvm_default_bcast_filters[] = { 133 { 134 /* arp */ 135 .discard = 0, 136 .frame_type = BCAST_FILTER_FRAME_TYPE_ALL, 137 .attrs = { 138 { 139 /* frame type - arp, hw type - ethernet */ 140 .offset_type = 141 BCAST_FILTER_OFFSET_PAYLOAD_START, 142 .offset = sizeof(rfc1042_header), 143 .val = cpu_to_be32(0x08060001), 144 .mask = cpu_to_be32(0xffffffff), 145 }, 146 { 147 /* arp dest ip */ 148 .offset_type = 149 BCAST_FILTER_OFFSET_PAYLOAD_START, 150 .offset = sizeof(rfc1042_header) + 2 + 151 sizeof(struct arphdr) + 152 ETH_ALEN + sizeof(__be32) + 153 ETH_ALEN, 154 .mask = cpu_to_be32(0xffffffff), 155 /* mark it as special field */ 156 .reserved1 = cpu_to_le16(BC_FILTER_MAGIC_IP), 157 }, 158 }, 159 }, 160 { 161 /* dhcp offer bcast */ 162 .discard = 0, 163 .frame_type = BCAST_FILTER_FRAME_TYPE_IPV4, 164 .attrs = { 165 { 166 /* udp dest port - 68 (bootp client)*/ 167 .offset_type = BCAST_FILTER_OFFSET_IP_END, 168 .offset = offsetof(struct udphdr, dest), 169 .val = cpu_to_be32(0x00440000), 170 .mask = cpu_to_be32(0xffff0000), 171 }, 172 { 173 /* dhcp - lsb bytes of client hw address */ 174 .offset_type = BCAST_FILTER_OFFSET_IP_END, 175 .offset = 38, 176 .mask = cpu_to_be32(0xffffffff), 177 /* mark it as special field */ 178 .reserved1 = cpu_to_le16(BC_FILTER_MAGIC_MAC), 179 }, 180 }, 181 }, 182 /* last filter must be empty */ 183 {}, 184 }; 185 #endif 186 187 static const struct cfg80211_pmsr_capabilities iwl_mvm_pmsr_capa = { 188 .max_peers = IWL_MVM_TOF_MAX_APS, 189 .report_ap_tsf = 1, 190 .randomize_mac_addr = 1, 191 192 .ftm = { 193 .supported = 1, 194 .asap = 1, 195 .non_asap = 1, 196 .request_lci = 1, 197 .request_civicloc = 1, 198 .max_bursts_exponent = -1, /* all supported */ 199 .max_ftms_per_burst = 0, /* no limits */ 200 .bandwidths = BIT(NL80211_CHAN_WIDTH_20_NOHT) | 201 BIT(NL80211_CHAN_WIDTH_20) | 202 BIT(NL80211_CHAN_WIDTH_40) | 203 BIT(NL80211_CHAN_WIDTH_80), 204 .preambles = BIT(NL80211_PREAMBLE_LEGACY) | 205 BIT(NL80211_PREAMBLE_HT) | 206 BIT(NL80211_PREAMBLE_VHT), 207 }, 208 }; 209 210 void iwl_mvm_ref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type) 211 { 212 if (!iwl_mvm_is_d0i3_supported(mvm)) 213 return; 214 215 IWL_DEBUG_RPM(mvm, "Take mvm reference - type %d\n", ref_type); 216 spin_lock_bh(&mvm->refs_lock); 217 mvm->refs[ref_type]++; 218 spin_unlock_bh(&mvm->refs_lock); 219 iwl_trans_ref(mvm->trans); 220 } 221 222 void iwl_mvm_unref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type) 223 { 224 if (!iwl_mvm_is_d0i3_supported(mvm)) 225 return; 226 227 IWL_DEBUG_RPM(mvm, "Leave mvm reference - type %d\n", ref_type); 228 spin_lock_bh(&mvm->refs_lock); 229 if (WARN_ON(!mvm->refs[ref_type])) { 230 spin_unlock_bh(&mvm->refs_lock); 231 return; 232 } 233 mvm->refs[ref_type]--; 234 spin_unlock_bh(&mvm->refs_lock); 235 iwl_trans_unref(mvm->trans); 236 } 237 238 static void iwl_mvm_unref_all_except(struct iwl_mvm *mvm, 239 enum iwl_mvm_ref_type except_ref) 240 { 241 int i, j; 242 243 if (!iwl_mvm_is_d0i3_supported(mvm)) 244 return; 245 246 spin_lock_bh(&mvm->refs_lock); 247 for (i = 0; i < IWL_MVM_REF_COUNT; i++) { 248 if (except_ref == i || !mvm->refs[i]) 249 continue; 250 251 IWL_DEBUG_RPM(mvm, "Cleanup: remove mvm ref type %d (%d)\n", 252 i, mvm->refs[i]); 253 for (j = 0; j < mvm->refs[i]; j++) 254 iwl_trans_unref(mvm->trans); 255 mvm->refs[i] = 0; 256 } 257 spin_unlock_bh(&mvm->refs_lock); 258 } 259 260 bool iwl_mvm_ref_taken(struct iwl_mvm *mvm) 261 { 262 int i; 263 bool taken = false; 264 265 if (!iwl_mvm_is_d0i3_supported(mvm)) 266 return true; 267 268 spin_lock_bh(&mvm->refs_lock); 269 for (i = 0; i < IWL_MVM_REF_COUNT; i++) { 270 if (mvm->refs[i]) { 271 taken = true; 272 break; 273 } 274 } 275 spin_unlock_bh(&mvm->refs_lock); 276 277 return taken; 278 } 279 280 int iwl_mvm_ref_sync(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type) 281 { 282 iwl_mvm_ref(mvm, ref_type); 283 284 if (!wait_event_timeout(mvm->d0i3_exit_waitq, 285 !test_bit(IWL_MVM_STATUS_IN_D0I3, &mvm->status), 286 HZ)) { 287 WARN_ON_ONCE(1); 288 iwl_mvm_unref(mvm, ref_type); 289 return -EIO; 290 } 291 292 return 0; 293 } 294 295 static void iwl_mvm_reset_phy_ctxts(struct iwl_mvm *mvm) 296 { 297 int i; 298 299 memset(mvm->phy_ctxts, 0, sizeof(mvm->phy_ctxts)); 300 for (i = 0; i < NUM_PHY_CTX; i++) { 301 mvm->phy_ctxts[i].id = i; 302 mvm->phy_ctxts[i].ref = 0; 303 } 304 } 305 306 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy, 307 const char *alpha2, 308 enum iwl_mcc_source src_id, 309 bool *changed) 310 { 311 struct ieee80211_regdomain *regd = NULL; 312 struct ieee80211_hw *hw = wiphy_to_ieee80211_hw(wiphy); 313 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 314 struct iwl_mcc_update_resp *resp; 315 316 IWL_DEBUG_LAR(mvm, "Getting regdomain data for %s from FW\n", alpha2); 317 318 lockdep_assert_held(&mvm->mutex); 319 320 resp = iwl_mvm_update_mcc(mvm, alpha2, src_id); 321 if (IS_ERR_OR_NULL(resp)) { 322 IWL_DEBUG_LAR(mvm, "Could not get update from FW %d\n", 323 PTR_ERR_OR_ZERO(resp)); 324 goto out; 325 } 326 327 if (changed) { 328 u32 status = le32_to_cpu(resp->status); 329 330 *changed = (status == MCC_RESP_NEW_CHAN_PROFILE || 331 status == MCC_RESP_ILLEGAL); 332 } 333 334 regd = iwl_parse_nvm_mcc_info(mvm->trans->dev, mvm->cfg, 335 __le32_to_cpu(resp->n_channels), 336 resp->channels, 337 __le16_to_cpu(resp->mcc), 338 __le16_to_cpu(resp->geo_info)); 339 /* Store the return source id */ 340 src_id = resp->source_id; 341 kfree(resp); 342 if (IS_ERR_OR_NULL(regd)) { 343 IWL_DEBUG_LAR(mvm, "Could not get parse update from FW %d\n", 344 PTR_ERR_OR_ZERO(regd)); 345 goto out; 346 } 347 348 IWL_DEBUG_LAR(mvm, "setting alpha2 from FW to %s (0x%x, 0x%x) src=%d\n", 349 regd->alpha2, regd->alpha2[0], regd->alpha2[1], src_id); 350 mvm->lar_regdom_set = true; 351 mvm->mcc_src = src_id; 352 353 out: 354 return regd; 355 } 356 357 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm) 358 { 359 bool changed; 360 struct ieee80211_regdomain *regd; 361 362 if (!iwl_mvm_is_lar_supported(mvm)) 363 return; 364 365 regd = iwl_mvm_get_current_regdomain(mvm, &changed); 366 if (!IS_ERR_OR_NULL(regd)) { 367 /* only update the regulatory core if changed */ 368 if (changed) 369 regulatory_set_wiphy_regd(mvm->hw->wiphy, regd); 370 371 kfree(regd); 372 } 373 } 374 375 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm, 376 bool *changed) 377 { 378 return iwl_mvm_get_regdomain(mvm->hw->wiphy, "ZZ", 379 iwl_mvm_is_wifi_mcc_supported(mvm) ? 380 MCC_SOURCE_GET_CURRENT : 381 MCC_SOURCE_OLD_FW, changed); 382 } 383 384 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm) 385 { 386 enum iwl_mcc_source used_src; 387 struct ieee80211_regdomain *regd; 388 int ret; 389 bool changed; 390 const struct ieee80211_regdomain *r = 391 rtnl_dereference(mvm->hw->wiphy->regd); 392 393 if (!r) 394 return -ENOENT; 395 396 /* save the last source in case we overwrite it below */ 397 used_src = mvm->mcc_src; 398 if (iwl_mvm_is_wifi_mcc_supported(mvm)) { 399 /* Notify the firmware we support wifi location updates */ 400 regd = iwl_mvm_get_current_regdomain(mvm, NULL); 401 if (!IS_ERR_OR_NULL(regd)) 402 kfree(regd); 403 } 404 405 /* Now set our last stored MCC and source */ 406 regd = iwl_mvm_get_regdomain(mvm->hw->wiphy, r->alpha2, used_src, 407 &changed); 408 if (IS_ERR_OR_NULL(regd)) 409 return -EIO; 410 411 /* update cfg80211 if the regdomain was changed */ 412 if (changed) 413 ret = regulatory_set_wiphy_regd_sync_rtnl(mvm->hw->wiphy, regd); 414 else 415 ret = 0; 416 417 kfree(regd); 418 return ret; 419 } 420 421 const static u8 he_if_types_ext_capa_sta[] = { 422 [0] = WLAN_EXT_CAPA1_EXT_CHANNEL_SWITCHING, 423 [7] = WLAN_EXT_CAPA8_OPMODE_NOTIF, 424 [9] = WLAN_EXT_CAPA10_TWT_REQUESTER_SUPPORT, 425 }; 426 427 const static struct wiphy_iftype_ext_capab he_iftypes_ext_capa[] = { 428 { 429 .iftype = NL80211_IFTYPE_STATION, 430 .extended_capabilities = he_if_types_ext_capa_sta, 431 .extended_capabilities_mask = he_if_types_ext_capa_sta, 432 .extended_capabilities_len = sizeof(he_if_types_ext_capa_sta), 433 }, 434 }; 435 436 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm) 437 { 438 struct ieee80211_hw *hw = mvm->hw; 439 int num_mac, ret, i; 440 static const u32 mvm_ciphers[] = { 441 WLAN_CIPHER_SUITE_WEP40, 442 WLAN_CIPHER_SUITE_WEP104, 443 WLAN_CIPHER_SUITE_TKIP, 444 WLAN_CIPHER_SUITE_CCMP, 445 }; 446 #ifdef CONFIG_PM_SLEEP 447 bool unified = fw_has_capa(&mvm->fw->ucode_capa, 448 IWL_UCODE_TLV_CAPA_CNSLDTD_D3_D0_IMG); 449 #endif 450 451 /* Tell mac80211 our characteristics */ 452 ieee80211_hw_set(hw, SIGNAL_DBM); 453 ieee80211_hw_set(hw, SPECTRUM_MGMT); 454 ieee80211_hw_set(hw, REPORTS_TX_ACK_STATUS); 455 ieee80211_hw_set(hw, WANT_MONITOR_VIF); 456 ieee80211_hw_set(hw, SUPPORTS_PS); 457 ieee80211_hw_set(hw, SUPPORTS_DYNAMIC_PS); 458 ieee80211_hw_set(hw, AMPDU_AGGREGATION); 459 ieee80211_hw_set(hw, TIMING_BEACON_ONLY); 460 ieee80211_hw_set(hw, CONNECTION_MONITOR); 461 ieee80211_hw_set(hw, CHANCTX_STA_CSA); 462 ieee80211_hw_set(hw, SUPPORT_FAST_XMIT); 463 ieee80211_hw_set(hw, SUPPORTS_CLONED_SKBS); 464 ieee80211_hw_set(hw, SUPPORTS_AMSDU_IN_AMPDU); 465 ieee80211_hw_set(hw, NEEDS_UNIQUE_STA_ADDR); 466 ieee80211_hw_set(hw, DEAUTH_NEED_MGD_TX_PREP); 467 ieee80211_hw_set(hw, SUPPORTS_VHT_EXT_NSS_BW); 468 ieee80211_hw_set(hw, BUFF_MMPDU_TXQ); 469 ieee80211_hw_set(hw, STA_MMPDU_TXQ); 470 ieee80211_hw_set(hw, TX_AMSDU); 471 ieee80211_hw_set(hw, TX_FRAG_LIST); 472 473 if (iwl_mvm_has_tlc_offload(mvm)) { 474 ieee80211_hw_set(hw, TX_AMPDU_SETUP_IN_HW); 475 ieee80211_hw_set(hw, HAS_RATE_CONTROL); 476 } 477 478 if (iwl_mvm_has_new_rx_api(mvm)) 479 ieee80211_hw_set(hw, SUPPORTS_REORDERING_BUFFER); 480 481 if (fw_has_capa(&mvm->fw->ucode_capa, 482 IWL_UCODE_TLV_CAPA_STA_PM_NOTIF)) { 483 ieee80211_hw_set(hw, AP_LINK_PS); 484 } else if (WARN_ON(iwl_mvm_has_new_tx_api(mvm))) { 485 /* 486 * we absolutely need this for the new TX API since that comes 487 * with many more queues than the current code can deal with 488 * for station powersave 489 */ 490 return -EINVAL; 491 } 492 493 if (mvm->trans->num_rx_queues > 1) 494 ieee80211_hw_set(hw, USES_RSS); 495 496 if (mvm->trans->max_skb_frags) 497 hw->netdev_features = NETIF_F_HIGHDMA | NETIF_F_SG; 498 499 hw->queues = IEEE80211_MAX_QUEUES; 500 hw->offchannel_tx_hw_queue = IWL_MVM_OFFCHANNEL_QUEUE; 501 hw->radiotap_mcs_details |= IEEE80211_RADIOTAP_MCS_HAVE_FEC | 502 IEEE80211_RADIOTAP_MCS_HAVE_STBC; 503 hw->radiotap_vht_details |= IEEE80211_RADIOTAP_VHT_KNOWN_STBC | 504 IEEE80211_RADIOTAP_VHT_KNOWN_BEAMFORMED; 505 506 hw->radiotap_timestamp.units_pos = 507 IEEE80211_RADIOTAP_TIMESTAMP_UNIT_US | 508 IEEE80211_RADIOTAP_TIMESTAMP_SPOS_PLCP_SIG_ACQ; 509 /* this is the case for CCK frames, it's better (only 8) for OFDM */ 510 hw->radiotap_timestamp.accuracy = 22; 511 512 if (!iwl_mvm_has_tlc_offload(mvm)) 513 hw->rate_control_algorithm = RS_NAME; 514 515 hw->uapsd_queues = IWL_MVM_UAPSD_QUEUES; 516 hw->uapsd_max_sp_len = IWL_UAPSD_MAX_SP; 517 hw->max_tx_fragments = mvm->trans->max_skb_frags; 518 519 BUILD_BUG_ON(ARRAY_SIZE(mvm->ciphers) < ARRAY_SIZE(mvm_ciphers) + 6); 520 memcpy(mvm->ciphers, mvm_ciphers, sizeof(mvm_ciphers)); 521 hw->wiphy->n_cipher_suites = ARRAY_SIZE(mvm_ciphers); 522 hw->wiphy->cipher_suites = mvm->ciphers; 523 524 if (iwl_mvm_has_new_rx_api(mvm)) { 525 mvm->ciphers[hw->wiphy->n_cipher_suites] = 526 WLAN_CIPHER_SUITE_GCMP; 527 hw->wiphy->n_cipher_suites++; 528 mvm->ciphers[hw->wiphy->n_cipher_suites] = 529 WLAN_CIPHER_SUITE_GCMP_256; 530 hw->wiphy->n_cipher_suites++; 531 } 532 533 /* Enable 11w if software crypto is not enabled (as the 534 * firmware will interpret some mgmt packets, so enabling it 535 * with software crypto isn't safe). 536 */ 537 if (!iwlwifi_mod_params.swcrypto) { 538 ieee80211_hw_set(hw, MFP_CAPABLE); 539 mvm->ciphers[hw->wiphy->n_cipher_suites] = 540 WLAN_CIPHER_SUITE_AES_CMAC; 541 hw->wiphy->n_cipher_suites++; 542 if (iwl_mvm_has_new_rx_api(mvm)) { 543 mvm->ciphers[hw->wiphy->n_cipher_suites] = 544 WLAN_CIPHER_SUITE_BIP_GMAC_128; 545 hw->wiphy->n_cipher_suites++; 546 mvm->ciphers[hw->wiphy->n_cipher_suites] = 547 WLAN_CIPHER_SUITE_BIP_GMAC_256; 548 hw->wiphy->n_cipher_suites++; 549 } 550 } 551 552 /* currently FW API supports only one optional cipher scheme */ 553 if (mvm->fw->cs[0].cipher) { 554 const struct iwl_fw_cipher_scheme *fwcs = &mvm->fw->cs[0]; 555 struct ieee80211_cipher_scheme *cs = &mvm->cs[0]; 556 557 mvm->hw->n_cipher_schemes = 1; 558 559 cs->cipher = le32_to_cpu(fwcs->cipher); 560 cs->iftype = BIT(NL80211_IFTYPE_STATION); 561 cs->hdr_len = fwcs->hdr_len; 562 cs->pn_len = fwcs->pn_len; 563 cs->pn_off = fwcs->pn_off; 564 cs->key_idx_off = fwcs->key_idx_off; 565 cs->key_idx_mask = fwcs->key_idx_mask; 566 cs->key_idx_shift = fwcs->key_idx_shift; 567 cs->mic_len = fwcs->mic_len; 568 569 mvm->hw->cipher_schemes = mvm->cs; 570 mvm->ciphers[hw->wiphy->n_cipher_suites] = cs->cipher; 571 hw->wiphy->n_cipher_suites++; 572 } 573 574 if (fw_has_capa(&mvm->fw->ucode_capa, 575 IWL_UCODE_TLV_CAPA_FTM_CALIBRATED)) { 576 wiphy_ext_feature_set(hw->wiphy, 577 NL80211_EXT_FEATURE_ENABLE_FTM_RESPONDER); 578 hw->wiphy->pmsr_capa = &iwl_mvm_pmsr_capa; 579 } 580 581 ieee80211_hw_set(hw, SINGLE_SCAN_ON_ALL_BANDS); 582 hw->wiphy->features |= 583 NL80211_FEATURE_SCHED_SCAN_RANDOM_MAC_ADDR | 584 NL80211_FEATURE_SCAN_RANDOM_MAC_ADDR | 585 NL80211_FEATURE_ND_RANDOM_MAC_ADDR; 586 587 hw->sta_data_size = sizeof(struct iwl_mvm_sta); 588 hw->vif_data_size = sizeof(struct iwl_mvm_vif); 589 hw->chanctx_data_size = sizeof(u16); 590 hw->txq_data_size = sizeof(struct iwl_mvm_txq); 591 592 hw->wiphy->interface_modes = BIT(NL80211_IFTYPE_STATION) | 593 BIT(NL80211_IFTYPE_P2P_CLIENT) | 594 BIT(NL80211_IFTYPE_AP) | 595 BIT(NL80211_IFTYPE_P2P_GO) | 596 BIT(NL80211_IFTYPE_P2P_DEVICE) | 597 BIT(NL80211_IFTYPE_ADHOC); 598 599 hw->wiphy->flags |= WIPHY_FLAG_IBSS_RSN; 600 hw->wiphy->regulatory_flags |= REGULATORY_ENABLE_RELAX_NO_IR; 601 if (iwl_mvm_is_lar_supported(mvm)) 602 hw->wiphy->regulatory_flags |= REGULATORY_WIPHY_SELF_MANAGED; 603 else 604 hw->wiphy->regulatory_flags |= REGULATORY_CUSTOM_REG | 605 REGULATORY_DISABLE_BEACON_HINTS; 606 607 hw->wiphy->flags |= WIPHY_FLAG_AP_UAPSD; 608 hw->wiphy->flags |= WIPHY_FLAG_HAS_CHANNEL_SWITCH; 609 610 hw->wiphy->iface_combinations = iwl_mvm_iface_combinations; 611 hw->wiphy->n_iface_combinations = 612 ARRAY_SIZE(iwl_mvm_iface_combinations); 613 614 hw->wiphy->max_remain_on_channel_duration = 10000; 615 hw->max_listen_interval = IWL_CONN_MAX_LISTEN_INTERVAL; 616 617 /* Extract MAC address */ 618 memcpy(mvm->addresses[0].addr, mvm->nvm_data->hw_addr, ETH_ALEN); 619 hw->wiphy->addresses = mvm->addresses; 620 hw->wiphy->n_addresses = 1; 621 622 /* Extract additional MAC addresses if available */ 623 num_mac = (mvm->nvm_data->n_hw_addrs > 1) ? 624 min(IWL_MVM_MAX_ADDRESSES, mvm->nvm_data->n_hw_addrs) : 1; 625 626 for (i = 1; i < num_mac; i++) { 627 memcpy(mvm->addresses[i].addr, mvm->addresses[i-1].addr, 628 ETH_ALEN); 629 mvm->addresses[i].addr[5]++; 630 hw->wiphy->n_addresses++; 631 } 632 633 iwl_mvm_reset_phy_ctxts(mvm); 634 635 hw->wiphy->max_scan_ie_len = iwl_mvm_max_scan_ie_len(mvm); 636 637 hw->wiphy->max_scan_ssids = PROBE_OPTION_MAX; 638 639 BUILD_BUG_ON(IWL_MVM_SCAN_STOPPING_MASK & IWL_MVM_SCAN_MASK); 640 BUILD_BUG_ON(IWL_MVM_MAX_UMAC_SCANS > HWEIGHT32(IWL_MVM_SCAN_MASK) || 641 IWL_MVM_MAX_LMAC_SCANS > HWEIGHT32(IWL_MVM_SCAN_MASK)); 642 643 if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN)) 644 mvm->max_scans = IWL_MVM_MAX_UMAC_SCANS; 645 else 646 mvm->max_scans = IWL_MVM_MAX_LMAC_SCANS; 647 648 if (mvm->nvm_data->bands[NL80211_BAND_2GHZ].n_channels) 649 hw->wiphy->bands[NL80211_BAND_2GHZ] = 650 &mvm->nvm_data->bands[NL80211_BAND_2GHZ]; 651 if (mvm->nvm_data->bands[NL80211_BAND_5GHZ].n_channels) { 652 hw->wiphy->bands[NL80211_BAND_5GHZ] = 653 &mvm->nvm_data->bands[NL80211_BAND_5GHZ]; 654 655 if (fw_has_capa(&mvm->fw->ucode_capa, 656 IWL_UCODE_TLV_CAPA_BEAMFORMER) && 657 fw_has_api(&mvm->fw->ucode_capa, 658 IWL_UCODE_TLV_API_LQ_SS_PARAMS)) 659 hw->wiphy->bands[NL80211_BAND_5GHZ]->vht_cap.cap |= 660 IEEE80211_VHT_CAP_SU_BEAMFORMER_CAPABLE; 661 } 662 663 hw->wiphy->hw_version = mvm->trans->hw_id; 664 665 if (iwlmvm_mod_params.power_scheme != IWL_POWER_SCHEME_CAM) 666 hw->wiphy->flags |= WIPHY_FLAG_PS_ON_BY_DEFAULT; 667 else 668 hw->wiphy->flags &= ~WIPHY_FLAG_PS_ON_BY_DEFAULT; 669 670 hw->wiphy->max_sched_scan_reqs = 1; 671 hw->wiphy->max_sched_scan_ssids = PROBE_OPTION_MAX; 672 hw->wiphy->max_match_sets = IWL_SCAN_MAX_PROFILES; 673 /* we create the 802.11 header and zero length SSID IE. */ 674 hw->wiphy->max_sched_scan_ie_len = 675 SCAN_OFFLOAD_PROBE_REQ_SIZE - 24 - 2; 676 hw->wiphy->max_sched_scan_plans = IWL_MAX_SCHED_SCAN_PLANS; 677 hw->wiphy->max_sched_scan_plan_interval = U16_MAX; 678 679 /* 680 * the firmware uses u8 for num of iterations, but 0xff is saved for 681 * infinite loop, so the maximum number of iterations is actually 254. 682 */ 683 hw->wiphy->max_sched_scan_plan_iterations = 254; 684 685 hw->wiphy->features |= NL80211_FEATURE_P2P_GO_CTWIN | 686 NL80211_FEATURE_LOW_PRIORITY_SCAN | 687 NL80211_FEATURE_P2P_GO_OPPPS | 688 NL80211_FEATURE_AP_MODE_CHAN_WIDTH_CHANGE | 689 NL80211_FEATURE_DYNAMIC_SMPS | 690 NL80211_FEATURE_STATIC_SMPS | 691 NL80211_FEATURE_SUPPORTS_WMM_ADMISSION; 692 693 if (fw_has_capa(&mvm->fw->ucode_capa, 694 IWL_UCODE_TLV_CAPA_TXPOWER_INSERTION_SUPPORT)) 695 hw->wiphy->features |= NL80211_FEATURE_TX_POWER_INSERTION; 696 if (fw_has_capa(&mvm->fw->ucode_capa, 697 IWL_UCODE_TLV_CAPA_QUIET_PERIOD_SUPPORT)) 698 hw->wiphy->features |= NL80211_FEATURE_QUIET; 699 700 if (fw_has_capa(&mvm->fw->ucode_capa, 701 IWL_UCODE_TLV_CAPA_DS_PARAM_SET_IE_SUPPORT)) 702 hw->wiphy->features |= 703 NL80211_FEATURE_DS_PARAM_SET_IE_IN_PROBES; 704 705 if (fw_has_capa(&mvm->fw->ucode_capa, 706 IWL_UCODE_TLV_CAPA_WFA_TPC_REP_IE_SUPPORT)) 707 hw->wiphy->features |= NL80211_FEATURE_WFA_TPC_IE_IN_PROBES; 708 709 if (fw_has_api(&mvm->fw->ucode_capa, 710 IWL_UCODE_TLV_API_SCAN_TSF_REPORT)) { 711 wiphy_ext_feature_set(hw->wiphy, 712 NL80211_EXT_FEATURE_SCAN_START_TIME); 713 wiphy_ext_feature_set(hw->wiphy, 714 NL80211_EXT_FEATURE_BSS_PARENT_TSF); 715 wiphy_ext_feature_set(hw->wiphy, 716 NL80211_EXT_FEATURE_SET_SCAN_DWELL); 717 } 718 719 if (iwl_mvm_is_oce_supported(mvm)) { 720 wiphy_ext_feature_set(hw->wiphy, 721 NL80211_EXT_FEATURE_ACCEPT_BCAST_PROBE_RESP); 722 wiphy_ext_feature_set(hw->wiphy, 723 NL80211_EXT_FEATURE_FILS_MAX_CHANNEL_TIME); 724 wiphy_ext_feature_set(hw->wiphy, 725 NL80211_EXT_FEATURE_OCE_PROBE_REQ_DEFERRAL_SUPPRESSION); 726 wiphy_ext_feature_set(hw->wiphy, 727 NL80211_EXT_FEATURE_OCE_PROBE_REQ_HIGH_TX_RATE); 728 } 729 730 if (mvm->nvm_data->sku_cap_11ax_enable && 731 !iwlwifi_mod_params.disable_11ax) { 732 hw->wiphy->iftype_ext_capab = he_iftypes_ext_capa; 733 hw->wiphy->num_iftype_ext_capab = 734 ARRAY_SIZE(he_iftypes_ext_capa); 735 } 736 737 mvm->rts_threshold = IEEE80211_MAX_RTS_THRESHOLD; 738 739 #ifdef CONFIG_PM_SLEEP 740 if (iwl_mvm_is_d0i3_supported(mvm) && 741 device_can_wakeup(mvm->trans->dev)) { 742 mvm->wowlan.flags = WIPHY_WOWLAN_ANY; 743 hw->wiphy->wowlan = &mvm->wowlan; 744 } 745 746 if ((unified || mvm->fw->img[IWL_UCODE_WOWLAN].num_sec) && 747 mvm->trans->ops->d3_suspend && 748 mvm->trans->ops->d3_resume && 749 device_can_wakeup(mvm->trans->dev)) { 750 mvm->wowlan.flags |= WIPHY_WOWLAN_MAGIC_PKT | 751 WIPHY_WOWLAN_DISCONNECT | 752 WIPHY_WOWLAN_EAP_IDENTITY_REQ | 753 WIPHY_WOWLAN_RFKILL_RELEASE | 754 WIPHY_WOWLAN_NET_DETECT; 755 if (!iwlwifi_mod_params.swcrypto) 756 mvm->wowlan.flags |= WIPHY_WOWLAN_SUPPORTS_GTK_REKEY | 757 WIPHY_WOWLAN_GTK_REKEY_FAILURE | 758 WIPHY_WOWLAN_4WAY_HANDSHAKE; 759 760 mvm->wowlan.n_patterns = IWL_WOWLAN_MAX_PATTERNS; 761 mvm->wowlan.pattern_min_len = IWL_WOWLAN_MIN_PATTERN_LEN; 762 mvm->wowlan.pattern_max_len = IWL_WOWLAN_MAX_PATTERN_LEN; 763 mvm->wowlan.max_nd_match_sets = IWL_SCAN_MAX_PROFILES; 764 hw->wiphy->wowlan = &mvm->wowlan; 765 } 766 #endif 767 768 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING 769 /* assign default bcast filtering configuration */ 770 mvm->bcast_filters = iwl_mvm_default_bcast_filters; 771 #endif 772 773 ret = iwl_mvm_leds_init(mvm); 774 if (ret) 775 return ret; 776 777 if (fw_has_capa(&mvm->fw->ucode_capa, 778 IWL_UCODE_TLV_CAPA_TDLS_SUPPORT)) { 779 IWL_DEBUG_TDLS(mvm, "TDLS supported\n"); 780 hw->wiphy->flags |= WIPHY_FLAG_SUPPORTS_TDLS; 781 ieee80211_hw_set(hw, TDLS_WIDER_BW); 782 } 783 784 if (fw_has_capa(&mvm->fw->ucode_capa, 785 IWL_UCODE_TLV_CAPA_TDLS_CHANNEL_SWITCH)) { 786 IWL_DEBUG_TDLS(mvm, "TDLS channel switch supported\n"); 787 hw->wiphy->features |= NL80211_FEATURE_TDLS_CHANNEL_SWITCH; 788 } 789 790 hw->netdev_features |= mvm->cfg->features; 791 if (!iwl_mvm_is_csum_supported(mvm)) { 792 hw->netdev_features &= ~(IWL_TX_CSUM_NETIF_FLAGS | 793 NETIF_F_RXCSUM); 794 /* We may support SW TX CSUM */ 795 if (IWL_MVM_SW_TX_CSUM_OFFLOAD) 796 hw->netdev_features |= IWL_TX_CSUM_NETIF_FLAGS; 797 } 798 799 if (mvm->cfg->vht_mu_mimo_supported) 800 wiphy_ext_feature_set(hw->wiphy, 801 NL80211_EXT_FEATURE_MU_MIMO_AIR_SNIFFER); 802 803 ret = ieee80211_register_hw(mvm->hw); 804 if (ret) { 805 iwl_mvm_leds_exit(mvm); 806 } 807 808 return ret; 809 } 810 811 static bool iwl_mvm_defer_tx(struct iwl_mvm *mvm, 812 struct ieee80211_sta *sta, 813 struct sk_buff *skb) 814 { 815 struct iwl_mvm_sta *mvmsta; 816 bool defer = false; 817 818 /* 819 * double check the IN_D0I3 flag both before and after 820 * taking the spinlock, in order to prevent taking 821 * the spinlock when not needed. 822 */ 823 if (likely(!test_bit(IWL_MVM_STATUS_IN_D0I3, &mvm->status))) 824 return false; 825 826 spin_lock(&mvm->d0i3_tx_lock); 827 /* 828 * testing the flag again ensures the skb dequeue 829 * loop (on d0i3 exit) hasn't run yet. 830 */ 831 if (!test_bit(IWL_MVM_STATUS_IN_D0I3, &mvm->status)) 832 goto out; 833 834 mvmsta = iwl_mvm_sta_from_mac80211(sta); 835 if (mvmsta->sta_id == IWL_MVM_INVALID_STA || 836 mvmsta->sta_id != mvm->d0i3_ap_sta_id) 837 goto out; 838 839 __skb_queue_tail(&mvm->d0i3_tx, skb); 840 841 /* trigger wakeup */ 842 iwl_mvm_ref(mvm, IWL_MVM_REF_TX); 843 iwl_mvm_unref(mvm, IWL_MVM_REF_TX); 844 845 defer = true; 846 out: 847 spin_unlock(&mvm->d0i3_tx_lock); 848 return defer; 849 } 850 851 static void iwl_mvm_mac_tx(struct ieee80211_hw *hw, 852 struct ieee80211_tx_control *control, 853 struct sk_buff *skb) 854 { 855 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 856 struct ieee80211_sta *sta = control->sta; 857 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 858 struct ieee80211_hdr *hdr = (void *)skb->data; 859 bool offchannel = IEEE80211_SKB_CB(skb)->flags & 860 IEEE80211_TX_CTL_TX_OFFCHAN; 861 862 if (iwl_mvm_is_radio_killed(mvm)) { 863 IWL_DEBUG_DROP(mvm, "Dropping - RF/CT KILL\n"); 864 goto drop; 865 } 866 867 if (offchannel && 868 !test_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status) && 869 !test_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status)) 870 goto drop; 871 872 /* treat non-bufferable MMPDUs on AP interfaces as broadcast */ 873 if ((info->control.vif->type == NL80211_IFTYPE_AP || 874 info->control.vif->type == NL80211_IFTYPE_ADHOC) && 875 ieee80211_is_mgmt(hdr->frame_control) && 876 !ieee80211_is_bufferable_mmpdu(hdr->frame_control)) 877 sta = NULL; 878 879 /* If there is no sta, and it's not offchannel - send through AP */ 880 if (!sta && info->control.vif->type == NL80211_IFTYPE_STATION && 881 !offchannel) { 882 struct iwl_mvm_vif *mvmvif = 883 iwl_mvm_vif_from_mac80211(info->control.vif); 884 u8 ap_sta_id = READ_ONCE(mvmvif->ap_sta_id); 885 886 if (ap_sta_id < IWL_MVM_STATION_COUNT) { 887 /* mac80211 holds rcu read lock */ 888 sta = rcu_dereference(mvm->fw_id_to_mac_id[ap_sta_id]); 889 if (IS_ERR_OR_NULL(sta)) 890 goto drop; 891 } 892 } 893 894 if (sta) { 895 if (iwl_mvm_defer_tx(mvm, sta, skb)) 896 return; 897 if (iwl_mvm_tx_skb(mvm, skb, sta)) 898 goto drop; 899 return; 900 } 901 902 if (iwl_mvm_tx_skb_non_sta(mvm, skb)) 903 goto drop; 904 return; 905 drop: 906 ieee80211_free_txskb(hw, skb); 907 } 908 909 void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq) 910 { 911 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 912 struct iwl_mvm_txq *mvmtxq = iwl_mvm_txq_from_mac80211(txq); 913 struct sk_buff *skb = NULL; 914 915 /* 916 * No need for threads to be pending here, they can leave the first 917 * taker all the work. 918 * 919 * mvmtxq->tx_request logic: 920 * 921 * If 0, no one is currently TXing, set to 1 to indicate current thread 922 * will now start TX and other threads should quit. 923 * 924 * If 1, another thread is currently TXing, set to 2 to indicate to 925 * that thread that there was another request. Since that request may 926 * have raced with the check whether the queue is empty, the TXing 927 * thread should check the queue's status one more time before leaving. 928 * This check is done in order to not leave any TX hanging in the queue 929 * until the next TX invocation (which may not even happen). 930 * 931 * If 2, another thread is currently TXing, and it will already double 932 * check the queue, so do nothing. 933 */ 934 if (atomic_fetch_add_unless(&mvmtxq->tx_request, 1, 2)) 935 return; 936 937 rcu_read_lock(); 938 do { 939 while (likely(!mvmtxq->stopped && 940 (mvm->trans->system_pm_mode == 941 IWL_PLAT_PM_MODE_DISABLED))) { 942 skb = ieee80211_tx_dequeue(hw, txq); 943 944 if (!skb) { 945 if (txq->sta) 946 IWL_DEBUG_TX(mvm, 947 "TXQ of sta %pM tid %d is now empty\n", 948 txq->sta->addr, 949 txq->tid); 950 break; 951 } 952 953 if (!txq->sta) 954 iwl_mvm_tx_skb_non_sta(mvm, skb); 955 else 956 iwl_mvm_tx_skb(mvm, skb, txq->sta); 957 } 958 } while (atomic_dec_return(&mvmtxq->tx_request)); 959 rcu_read_unlock(); 960 } 961 962 static void iwl_mvm_mac_wake_tx_queue(struct ieee80211_hw *hw, 963 struct ieee80211_txq *txq) 964 { 965 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 966 struct iwl_mvm_txq *mvmtxq = iwl_mvm_txq_from_mac80211(txq); 967 968 /* 969 * Please note that racing is handled very carefully here: 970 * mvmtxq->txq_id is updated during allocation, and mvmtxq->list is 971 * deleted afterwards. 972 * This means that if: 973 * mvmtxq->txq_id != INVALID_QUEUE && list_empty(&mvmtxq->list): 974 * queue is allocated and we can TX. 975 * mvmtxq->txq_id != INVALID_QUEUE && !list_empty(&mvmtxq->list): 976 * a race, should defer the frame. 977 * mvmtxq->txq_id == INVALID_QUEUE && list_empty(&mvmtxq->list): 978 * need to allocate the queue and defer the frame. 979 * mvmtxq->txq_id == INVALID_QUEUE && !list_empty(&mvmtxq->list): 980 * queue is already scheduled for allocation, no need to allocate, 981 * should defer the frame. 982 */ 983 984 /* If the queue is allocated TX and return. */ 985 if (!txq->sta || mvmtxq->txq_id != IWL_MVM_INVALID_QUEUE) { 986 /* 987 * Check that list is empty to avoid a race where txq_id is 988 * already updated, but the queue allocation work wasn't 989 * finished 990 */ 991 if (unlikely(txq->sta && !list_empty(&mvmtxq->list))) 992 return; 993 994 iwl_mvm_mac_itxq_xmit(hw, txq); 995 return; 996 } 997 998 /* The list is being deleted only after the queue is fully allocated. */ 999 if (!list_empty(&mvmtxq->list)) 1000 return; 1001 1002 list_add_tail(&mvmtxq->list, &mvm->add_stream_txqs); 1003 schedule_work(&mvm->add_stream_wk); 1004 } 1005 1006 #define CHECK_BA_TRIGGER(_mvm, _trig, _tid_bm, _tid, _fmt...) \ 1007 do { \ 1008 if (!(le16_to_cpu(_tid_bm) & BIT(_tid))) \ 1009 break; \ 1010 iwl_fw_dbg_collect_trig(&(_mvm)->fwrt, _trig, _fmt); \ 1011 } while (0) 1012 1013 static void 1014 iwl_mvm_ampdu_check_trigger(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1015 struct ieee80211_sta *sta, u16 tid, u16 rx_ba_ssn, 1016 enum ieee80211_ampdu_mlme_action action) 1017 { 1018 struct iwl_fw_dbg_trigger_tlv *trig; 1019 struct iwl_fw_dbg_trigger_ba *ba_trig; 1020 1021 trig = iwl_fw_dbg_trigger_on(&mvm->fwrt, ieee80211_vif_to_wdev(vif), 1022 FW_DBG_TRIGGER_BA); 1023 if (!trig) 1024 return; 1025 1026 ba_trig = (void *)trig->data; 1027 1028 switch (action) { 1029 case IEEE80211_AMPDU_TX_OPERATIONAL: { 1030 struct iwl_mvm_sta *mvmsta = iwl_mvm_sta_from_mac80211(sta); 1031 struct iwl_mvm_tid_data *tid_data = &mvmsta->tid_data[tid]; 1032 1033 CHECK_BA_TRIGGER(mvm, trig, ba_trig->tx_ba_start, tid, 1034 "TX AGG START: MAC %pM tid %d ssn %d\n", 1035 sta->addr, tid, tid_data->ssn); 1036 break; 1037 } 1038 case IEEE80211_AMPDU_TX_STOP_CONT: 1039 CHECK_BA_TRIGGER(mvm, trig, ba_trig->tx_ba_stop, tid, 1040 "TX AGG STOP: MAC %pM tid %d\n", 1041 sta->addr, tid); 1042 break; 1043 case IEEE80211_AMPDU_RX_START: 1044 CHECK_BA_TRIGGER(mvm, trig, ba_trig->rx_ba_start, tid, 1045 "RX AGG START: MAC %pM tid %d ssn %d\n", 1046 sta->addr, tid, rx_ba_ssn); 1047 break; 1048 case IEEE80211_AMPDU_RX_STOP: 1049 CHECK_BA_TRIGGER(mvm, trig, ba_trig->rx_ba_stop, tid, 1050 "RX AGG STOP: MAC %pM tid %d\n", 1051 sta->addr, tid); 1052 break; 1053 default: 1054 break; 1055 } 1056 } 1057 1058 static int iwl_mvm_mac_ampdu_action(struct ieee80211_hw *hw, 1059 struct ieee80211_vif *vif, 1060 struct ieee80211_ampdu_params *params) 1061 { 1062 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1063 int ret; 1064 bool tx_agg_ref = false; 1065 struct ieee80211_sta *sta = params->sta; 1066 enum ieee80211_ampdu_mlme_action action = params->action; 1067 u16 tid = params->tid; 1068 u16 *ssn = ¶ms->ssn; 1069 u16 buf_size = params->buf_size; 1070 bool amsdu = params->amsdu; 1071 u16 timeout = params->timeout; 1072 1073 IWL_DEBUG_HT(mvm, "A-MPDU action on addr %pM tid %d: action %d\n", 1074 sta->addr, tid, action); 1075 1076 if (!(mvm->nvm_data->sku_cap_11n_enable)) 1077 return -EACCES; 1078 1079 /* return from D0i3 before starting a new Tx aggregation */ 1080 switch (action) { 1081 case IEEE80211_AMPDU_TX_START: 1082 case IEEE80211_AMPDU_TX_STOP_CONT: 1083 case IEEE80211_AMPDU_TX_STOP_FLUSH: 1084 case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT: 1085 case IEEE80211_AMPDU_TX_OPERATIONAL: 1086 /* 1087 * for tx start, wait synchronously until D0i3 exit to 1088 * get the correct sequence number for the tid. 1089 * additionally, some other ampdu actions use direct 1090 * target access, which is not handled automatically 1091 * by the trans layer (unlike commands), so wait for 1092 * d0i3 exit in these cases as well. 1093 */ 1094 ret = iwl_mvm_ref_sync(mvm, IWL_MVM_REF_TX_AGG); 1095 if (ret) 1096 return ret; 1097 1098 tx_agg_ref = true; 1099 break; 1100 default: 1101 break; 1102 } 1103 1104 mutex_lock(&mvm->mutex); 1105 1106 switch (action) { 1107 case IEEE80211_AMPDU_RX_START: 1108 if (iwl_mvm_vif_from_mac80211(vif)->ap_sta_id == 1109 iwl_mvm_sta_from_mac80211(sta)->sta_id) { 1110 struct iwl_mvm_vif *mvmvif; 1111 u16 macid = iwl_mvm_vif_from_mac80211(vif)->id; 1112 struct iwl_mvm_tcm_mac *mdata = &mvm->tcm.data[macid]; 1113 1114 mdata->opened_rx_ba_sessions = true; 1115 mvmvif = iwl_mvm_vif_from_mac80211(vif); 1116 cancel_delayed_work(&mvmvif->uapsd_nonagg_detected_wk); 1117 } 1118 if (!iwl_enable_rx_ampdu()) { 1119 ret = -EINVAL; 1120 break; 1121 } 1122 ret = iwl_mvm_sta_rx_agg(mvm, sta, tid, *ssn, true, buf_size, 1123 timeout); 1124 break; 1125 case IEEE80211_AMPDU_RX_STOP: 1126 ret = iwl_mvm_sta_rx_agg(mvm, sta, tid, 0, false, buf_size, 1127 timeout); 1128 break; 1129 case IEEE80211_AMPDU_TX_START: 1130 if (!iwl_enable_tx_ampdu()) { 1131 ret = -EINVAL; 1132 break; 1133 } 1134 ret = iwl_mvm_sta_tx_agg_start(mvm, vif, sta, tid, ssn); 1135 break; 1136 case IEEE80211_AMPDU_TX_STOP_CONT: 1137 ret = iwl_mvm_sta_tx_agg_stop(mvm, vif, sta, tid); 1138 break; 1139 case IEEE80211_AMPDU_TX_STOP_FLUSH: 1140 case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT: 1141 ret = iwl_mvm_sta_tx_agg_flush(mvm, vif, sta, tid); 1142 break; 1143 case IEEE80211_AMPDU_TX_OPERATIONAL: 1144 ret = iwl_mvm_sta_tx_agg_oper(mvm, vif, sta, tid, 1145 buf_size, amsdu); 1146 break; 1147 default: 1148 WARN_ON_ONCE(1); 1149 ret = -EINVAL; 1150 break; 1151 } 1152 1153 if (!ret) { 1154 u16 rx_ba_ssn = 0; 1155 1156 if (action == IEEE80211_AMPDU_RX_START) 1157 rx_ba_ssn = *ssn; 1158 1159 iwl_mvm_ampdu_check_trigger(mvm, vif, sta, tid, 1160 rx_ba_ssn, action); 1161 } 1162 mutex_unlock(&mvm->mutex); 1163 1164 /* 1165 * If the tid is marked as started, we won't use it for offloaded 1166 * traffic on the next D0i3 entry. It's safe to unref. 1167 */ 1168 if (tx_agg_ref) 1169 iwl_mvm_unref(mvm, IWL_MVM_REF_TX_AGG); 1170 1171 return ret; 1172 } 1173 1174 static void iwl_mvm_cleanup_iterator(void *data, u8 *mac, 1175 struct ieee80211_vif *vif) 1176 { 1177 struct iwl_mvm *mvm = data; 1178 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 1179 1180 mvmvif->uploaded = false; 1181 mvmvif->ap_sta_id = IWL_MVM_INVALID_STA; 1182 1183 spin_lock_bh(&mvm->time_event_lock); 1184 iwl_mvm_te_clear_data(mvm, &mvmvif->time_event_data); 1185 spin_unlock_bh(&mvm->time_event_lock); 1186 1187 mvmvif->phy_ctxt = NULL; 1188 memset(&mvmvif->bf_data, 0, sizeof(mvmvif->bf_data)); 1189 memset(&mvmvif->probe_resp_data, 0, sizeof(mvmvif->probe_resp_data)); 1190 } 1191 1192 static void iwl_mvm_restart_cleanup(struct iwl_mvm *mvm) 1193 { 1194 /* clear the D3 reconfig, we only need it to avoid dumping a 1195 * firmware coredump on reconfiguration, we shouldn't do that 1196 * on D3->D0 transition 1197 */ 1198 if (!test_and_clear_bit(IWL_MVM_STATUS_D3_RECONFIG, &mvm->status)) { 1199 mvm->fwrt.dump.desc = &iwl_dump_desc_assert; 1200 iwl_fw_error_dump(&mvm->fwrt); 1201 } 1202 1203 /* cleanup all stale references (scan, roc), but keep the 1204 * ucode_down ref until reconfig is complete 1205 */ 1206 iwl_mvm_unref_all_except(mvm, IWL_MVM_REF_UCODE_DOWN); 1207 1208 iwl_mvm_stop_device(mvm); 1209 1210 mvm->cur_aid = 0; 1211 1212 mvm->scan_status = 0; 1213 mvm->ps_disabled = false; 1214 mvm->calibrating = false; 1215 1216 /* just in case one was running */ 1217 iwl_mvm_cleanup_roc_te(mvm); 1218 ieee80211_remain_on_channel_expired(mvm->hw); 1219 1220 iwl_mvm_ftm_restart(mvm); 1221 1222 /* 1223 * cleanup all interfaces, even inactive ones, as some might have 1224 * gone down during the HW restart 1225 */ 1226 ieee80211_iterate_interfaces(mvm->hw, 0, iwl_mvm_cleanup_iterator, mvm); 1227 1228 mvm->p2p_device_vif = NULL; 1229 mvm->d0i3_ap_sta_id = IWL_MVM_INVALID_STA; 1230 1231 iwl_mvm_reset_phy_ctxts(mvm); 1232 memset(mvm->fw_key_table, 0, sizeof(mvm->fw_key_table)); 1233 memset(&mvm->last_bt_notif, 0, sizeof(mvm->last_bt_notif)); 1234 memset(&mvm->last_bt_ci_cmd, 0, sizeof(mvm->last_bt_ci_cmd)); 1235 1236 ieee80211_wake_queues(mvm->hw); 1237 1238 /* clear any stale d0i3 state */ 1239 clear_bit(IWL_MVM_STATUS_IN_D0I3, &mvm->status); 1240 1241 mvm->vif_count = 0; 1242 mvm->rx_ba_sessions = 0; 1243 mvm->fwrt.dump.conf = FW_DBG_INVALID; 1244 mvm->monitor_on = false; 1245 1246 /* keep statistics ticking */ 1247 iwl_mvm_accu_radio_stats(mvm); 1248 } 1249 1250 int __iwl_mvm_mac_start(struct iwl_mvm *mvm) 1251 { 1252 int ret; 1253 1254 lockdep_assert_held(&mvm->mutex); 1255 1256 if (test_bit(IWL_MVM_STATUS_HW_RESTART_REQUESTED, &mvm->status)) { 1257 /* 1258 * Now convert the HW_RESTART_REQUESTED flag to IN_HW_RESTART 1259 * so later code will - from now on - see that we're doing it. 1260 */ 1261 set_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status); 1262 clear_bit(IWL_MVM_STATUS_HW_RESTART_REQUESTED, &mvm->status); 1263 /* Clean up some internal and mac80211 state on restart */ 1264 iwl_mvm_restart_cleanup(mvm); 1265 } else { 1266 /* Hold the reference to prevent runtime suspend while 1267 * the start procedure runs. It's a bit confusing 1268 * that the UCODE_DOWN reference is taken, but it just 1269 * means "UCODE is not UP yet". ( TODO: rename this 1270 * reference). 1271 */ 1272 iwl_mvm_ref(mvm, IWL_MVM_REF_UCODE_DOWN); 1273 } 1274 ret = iwl_mvm_up(mvm); 1275 1276 iwl_fw_dbg_apply_point(&mvm->fwrt, IWL_FW_INI_APPLY_POST_INIT); 1277 1278 if (ret && test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status)) { 1279 /* Something went wrong - we need to finish some cleanup 1280 * that normally iwl_mvm_mac_restart_complete() below 1281 * would do. 1282 */ 1283 clear_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status); 1284 #ifdef CONFIG_PM 1285 iwl_mvm_d0i3_enable_tx(mvm, NULL); 1286 #endif 1287 } 1288 1289 return ret; 1290 } 1291 1292 static int iwl_mvm_mac_start(struct ieee80211_hw *hw) 1293 { 1294 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1295 int ret; 1296 1297 /* Some hw restart cleanups must not hold the mutex */ 1298 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status)) { 1299 /* 1300 * Make sure we are out of d0i3. This is needed 1301 * to make sure the reference accounting is correct 1302 * (and there is no stale d0i3_exit_work). 1303 */ 1304 wait_event_timeout(mvm->d0i3_exit_waitq, 1305 !test_bit(IWL_MVM_STATUS_IN_D0I3, 1306 &mvm->status), 1307 HZ); 1308 } 1309 1310 mutex_lock(&mvm->mutex); 1311 ret = __iwl_mvm_mac_start(mvm); 1312 mutex_unlock(&mvm->mutex); 1313 1314 return ret; 1315 } 1316 1317 static void iwl_mvm_restart_complete(struct iwl_mvm *mvm) 1318 { 1319 int ret; 1320 1321 mutex_lock(&mvm->mutex); 1322 1323 clear_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status); 1324 #ifdef CONFIG_PM 1325 iwl_mvm_d0i3_enable_tx(mvm, NULL); 1326 #endif 1327 ret = iwl_mvm_update_quotas(mvm, true, NULL); 1328 if (ret) 1329 IWL_ERR(mvm, "Failed to update quotas after restart (%d)\n", 1330 ret); 1331 1332 /* allow transport/FW low power modes */ 1333 iwl_mvm_unref(mvm, IWL_MVM_REF_UCODE_DOWN); 1334 1335 iwl_mvm_send_recovery_cmd(mvm, ERROR_RECOVERY_END_OF_RECOVERY); 1336 1337 /* 1338 * If we have TDLS peers, remove them. We don't know the last seqno/PN 1339 * of packets the FW sent out, so we must reconnect. 1340 */ 1341 iwl_mvm_teardown_tdls_peers(mvm); 1342 1343 mutex_unlock(&mvm->mutex); 1344 } 1345 1346 static void iwl_mvm_resume_complete(struct iwl_mvm *mvm) 1347 { 1348 if (iwl_mvm_is_d0i3_supported(mvm) && 1349 iwl_mvm_enter_d0i3_on_suspend(mvm)) 1350 WARN_ONCE(!wait_event_timeout(mvm->d0i3_exit_waitq, 1351 !test_bit(IWL_MVM_STATUS_IN_D0I3, 1352 &mvm->status), 1353 HZ), 1354 "D0i3 exit on resume timed out\n"); 1355 } 1356 1357 static void 1358 iwl_mvm_mac_reconfig_complete(struct ieee80211_hw *hw, 1359 enum ieee80211_reconfig_type reconfig_type) 1360 { 1361 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1362 1363 switch (reconfig_type) { 1364 case IEEE80211_RECONFIG_TYPE_RESTART: 1365 iwl_mvm_restart_complete(mvm); 1366 break; 1367 case IEEE80211_RECONFIG_TYPE_SUSPEND: 1368 iwl_mvm_resume_complete(mvm); 1369 break; 1370 } 1371 } 1372 1373 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm) 1374 { 1375 lockdep_assert_held(&mvm->mutex); 1376 1377 /* firmware counters are obviously reset now, but we shouldn't 1378 * partially track so also clear the fw_reset_accu counters. 1379 */ 1380 memset(&mvm->accu_radio_stats, 0, sizeof(mvm->accu_radio_stats)); 1381 1382 /* async_handlers_wk is now blocked */ 1383 1384 /* 1385 * The work item could be running or queued if the 1386 * ROC time event stops just as we get here. 1387 */ 1388 flush_work(&mvm->roc_done_wk); 1389 1390 iwl_mvm_stop_device(mvm); 1391 1392 iwl_mvm_async_handlers_purge(mvm); 1393 /* async_handlers_list is empty and will stay empty: HW is stopped */ 1394 1395 /* the fw is stopped, the aux sta is dead: clean up driver state */ 1396 iwl_mvm_del_aux_sta(mvm); 1397 1398 /* 1399 * Clear IN_HW_RESTART and HW_RESTART_REQUESTED flag when stopping the 1400 * hw (as restart_complete() won't be called in this case) and mac80211 1401 * won't execute the restart. 1402 * But make sure to cleanup interfaces that have gone down before/during 1403 * HW restart was requested. 1404 */ 1405 if (test_and_clear_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status) || 1406 test_and_clear_bit(IWL_MVM_STATUS_HW_RESTART_REQUESTED, 1407 &mvm->status)) 1408 ieee80211_iterate_interfaces(mvm->hw, 0, 1409 iwl_mvm_cleanup_iterator, mvm); 1410 1411 /* We shouldn't have any UIDs still set. Loop over all the UIDs to 1412 * make sure there's nothing left there and warn if any is found. 1413 */ 1414 if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN)) { 1415 int i; 1416 1417 for (i = 0; i < mvm->max_scans; i++) { 1418 if (WARN_ONCE(mvm->scan_uid_status[i], 1419 "UMAC scan UID %d status was not cleaned\n", 1420 i)) 1421 mvm->scan_uid_status[i] = 0; 1422 } 1423 } 1424 } 1425 1426 static void iwl_mvm_mac_stop(struct ieee80211_hw *hw) 1427 { 1428 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1429 1430 flush_work(&mvm->d0i3_exit_work); 1431 flush_work(&mvm->async_handlers_wk); 1432 flush_work(&mvm->add_stream_wk); 1433 1434 /* 1435 * Lock and clear the firmware running bit here already, so that 1436 * new commands coming in elsewhere, e.g. from debugfs, will not 1437 * be able to proceed. This is important here because one of those 1438 * debugfs files causes the firmware dump to be triggered, and if we 1439 * don't stop debugfs accesses before canceling that it could be 1440 * retriggered after we flush it but before we've cleared the bit. 1441 */ 1442 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status); 1443 1444 iwl_fw_cancel_dump(&mvm->fwrt); 1445 cancel_delayed_work_sync(&mvm->cs_tx_unblock_dwork); 1446 cancel_delayed_work_sync(&mvm->scan_timeout_dwork); 1447 iwl_fw_free_dump_desc(&mvm->fwrt); 1448 1449 mutex_lock(&mvm->mutex); 1450 __iwl_mvm_mac_stop(mvm); 1451 mutex_unlock(&mvm->mutex); 1452 1453 /* 1454 * The worker might have been waiting for the mutex, let it run and 1455 * discover that its list is now empty. 1456 */ 1457 cancel_work_sync(&mvm->async_handlers_wk); 1458 } 1459 1460 static struct iwl_mvm_phy_ctxt *iwl_mvm_get_free_phy_ctxt(struct iwl_mvm *mvm) 1461 { 1462 u16 i; 1463 1464 lockdep_assert_held(&mvm->mutex); 1465 1466 for (i = 0; i < NUM_PHY_CTX; i++) 1467 if (!mvm->phy_ctxts[i].ref) 1468 return &mvm->phy_ctxts[i]; 1469 1470 IWL_ERR(mvm, "No available PHY context\n"); 1471 return NULL; 1472 } 1473 1474 static int iwl_mvm_set_tx_power(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1475 s16 tx_power) 1476 { 1477 int len; 1478 union { 1479 struct iwl_dev_tx_power_cmd v5; 1480 struct iwl_dev_tx_power_cmd_v4 v4; 1481 } cmd = { 1482 .v5.v3.set_mode = cpu_to_le32(IWL_TX_POWER_MODE_SET_MAC), 1483 .v5.v3.mac_context_id = 1484 cpu_to_le32(iwl_mvm_vif_from_mac80211(vif)->id), 1485 .v5.v3.pwr_restriction = cpu_to_le16(8 * tx_power), 1486 }; 1487 1488 if (tx_power == IWL_DEFAULT_MAX_TX_POWER) 1489 cmd.v5.v3.pwr_restriction = cpu_to_le16(IWL_DEV_MAX_TX_POWER); 1490 1491 if (fw_has_api(&mvm->fw->ucode_capa, 1492 IWL_UCODE_TLV_API_REDUCE_TX_POWER)) 1493 len = sizeof(cmd.v5); 1494 else if (fw_has_capa(&mvm->fw->ucode_capa, 1495 IWL_UCODE_TLV_CAPA_TX_POWER_ACK)) 1496 len = sizeof(cmd.v4); 1497 else 1498 len = sizeof(cmd.v4.v3); 1499 1500 return iwl_mvm_send_cmd_pdu(mvm, REDUCE_TX_POWER_CMD, 0, len, &cmd); 1501 } 1502 1503 static int iwl_mvm_mac_add_interface(struct ieee80211_hw *hw, 1504 struct ieee80211_vif *vif) 1505 { 1506 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1507 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 1508 int ret; 1509 1510 mvmvif->mvm = mvm; 1511 RCU_INIT_POINTER(mvmvif->probe_resp_data, NULL); 1512 1513 /* 1514 * make sure D0i3 exit is completed, otherwise a target access 1515 * during tx queue configuration could be done when still in 1516 * D0i3 state. 1517 */ 1518 ret = iwl_mvm_ref_sync(mvm, IWL_MVM_REF_ADD_IF); 1519 if (ret) 1520 return ret; 1521 1522 /* 1523 * Not much to do here. The stack will not allow interface 1524 * types or combinations that we didn't advertise, so we 1525 * don't really have to check the types. 1526 */ 1527 1528 mutex_lock(&mvm->mutex); 1529 1530 /* make sure that beacon statistics don't go backwards with FW reset */ 1531 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status)) 1532 mvmvif->beacon_stats.accu_num_beacons += 1533 mvmvif->beacon_stats.num_beacons; 1534 1535 /* Allocate resources for the MAC context, and add it to the fw */ 1536 ret = iwl_mvm_mac_ctxt_init(mvm, vif); 1537 if (ret) 1538 goto out_unlock; 1539 1540 rcu_assign_pointer(mvm->vif_id_to_mac[mvmvif->id], vif); 1541 1542 /* Counting number of interfaces is needed for legacy PM */ 1543 if (vif->type != NL80211_IFTYPE_P2P_DEVICE) 1544 mvm->vif_count++; 1545 1546 /* 1547 * The AP binding flow can be done only after the beacon 1548 * template is configured (which happens only in the mac80211 1549 * start_ap() flow), and adding the broadcast station can happen 1550 * only after the binding. 1551 * In addition, since modifying the MAC before adding a bcast 1552 * station is not allowed by the FW, delay the adding of MAC context to 1553 * the point where we can also add the bcast station. 1554 * In short: there's not much we can do at this point, other than 1555 * allocating resources :) 1556 */ 1557 if (vif->type == NL80211_IFTYPE_AP || 1558 vif->type == NL80211_IFTYPE_ADHOC) { 1559 ret = iwl_mvm_alloc_bcast_sta(mvm, vif); 1560 if (ret) { 1561 IWL_ERR(mvm, "Failed to allocate bcast sta\n"); 1562 goto out_release; 1563 } 1564 1565 /* 1566 * Only queue for this station is the mcast queue, 1567 * which shouldn't be in TFD mask anyway 1568 */ 1569 ret = iwl_mvm_allocate_int_sta(mvm, &mvmvif->mcast_sta, 1570 0, vif->type, 1571 IWL_STA_MULTICAST); 1572 if (ret) 1573 goto out_release; 1574 1575 iwl_mvm_vif_dbgfs_register(mvm, vif); 1576 goto out_unlock; 1577 } 1578 1579 mvmvif->features |= hw->netdev_features; 1580 1581 ret = iwl_mvm_mac_ctxt_add(mvm, vif); 1582 if (ret) 1583 goto out_release; 1584 1585 ret = iwl_mvm_power_update_mac(mvm); 1586 if (ret) 1587 goto out_remove_mac; 1588 1589 /* beacon filtering */ 1590 ret = iwl_mvm_disable_beacon_filter(mvm, vif, 0); 1591 if (ret) 1592 goto out_remove_mac; 1593 1594 if (!mvm->bf_allowed_vif && 1595 vif->type == NL80211_IFTYPE_STATION && !vif->p2p) { 1596 mvm->bf_allowed_vif = mvmvif; 1597 vif->driver_flags |= IEEE80211_VIF_BEACON_FILTER | 1598 IEEE80211_VIF_SUPPORTS_CQM_RSSI; 1599 } 1600 1601 /* 1602 * P2P_DEVICE interface does not have a channel context assigned to it, 1603 * so a dedicated PHY context is allocated to it and the corresponding 1604 * MAC context is bound to it at this stage. 1605 */ 1606 if (vif->type == NL80211_IFTYPE_P2P_DEVICE) { 1607 1608 mvmvif->phy_ctxt = iwl_mvm_get_free_phy_ctxt(mvm); 1609 if (!mvmvif->phy_ctxt) { 1610 ret = -ENOSPC; 1611 goto out_free_bf; 1612 } 1613 1614 iwl_mvm_phy_ctxt_ref(mvm, mvmvif->phy_ctxt); 1615 ret = iwl_mvm_binding_add_vif(mvm, vif); 1616 if (ret) 1617 goto out_unref_phy; 1618 1619 ret = iwl_mvm_add_p2p_bcast_sta(mvm, vif); 1620 if (ret) 1621 goto out_unbind; 1622 1623 /* Save a pointer to p2p device vif, so it can later be used to 1624 * update the p2p device MAC when a GO is started/stopped */ 1625 mvm->p2p_device_vif = vif; 1626 } 1627 1628 iwl_mvm_tcm_add_vif(mvm, vif); 1629 1630 if (vif->type == NL80211_IFTYPE_MONITOR) 1631 mvm->monitor_on = true; 1632 1633 iwl_mvm_vif_dbgfs_register(mvm, vif); 1634 goto out_unlock; 1635 1636 out_unbind: 1637 iwl_mvm_binding_remove_vif(mvm, vif); 1638 out_unref_phy: 1639 iwl_mvm_phy_ctxt_unref(mvm, mvmvif->phy_ctxt); 1640 out_free_bf: 1641 if (mvm->bf_allowed_vif == mvmvif) { 1642 mvm->bf_allowed_vif = NULL; 1643 vif->driver_flags &= ~(IEEE80211_VIF_BEACON_FILTER | 1644 IEEE80211_VIF_SUPPORTS_CQM_RSSI); 1645 } 1646 out_remove_mac: 1647 mvmvif->phy_ctxt = NULL; 1648 iwl_mvm_mac_ctxt_remove(mvm, vif); 1649 out_release: 1650 if (vif->type != NL80211_IFTYPE_P2P_DEVICE) 1651 mvm->vif_count--; 1652 out_unlock: 1653 mutex_unlock(&mvm->mutex); 1654 1655 iwl_mvm_unref(mvm, IWL_MVM_REF_ADD_IF); 1656 1657 return ret; 1658 } 1659 1660 static void iwl_mvm_prepare_mac_removal(struct iwl_mvm *mvm, 1661 struct ieee80211_vif *vif) 1662 { 1663 if (vif->type == NL80211_IFTYPE_P2P_DEVICE) { 1664 /* 1665 * Flush the ROC worker which will flush the OFFCHANNEL queue. 1666 * We assume here that all the packets sent to the OFFCHANNEL 1667 * queue are sent in ROC session. 1668 */ 1669 flush_work(&mvm->roc_done_wk); 1670 } 1671 } 1672 1673 static void iwl_mvm_mac_remove_interface(struct ieee80211_hw *hw, 1674 struct ieee80211_vif *vif) 1675 { 1676 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1677 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 1678 struct iwl_probe_resp_data *probe_data; 1679 1680 iwl_mvm_prepare_mac_removal(mvm, vif); 1681 1682 if (!(vif->type == NL80211_IFTYPE_AP || 1683 vif->type == NL80211_IFTYPE_ADHOC)) 1684 iwl_mvm_tcm_rm_vif(mvm, vif); 1685 1686 mutex_lock(&mvm->mutex); 1687 1688 probe_data = rcu_dereference_protected(mvmvif->probe_resp_data, 1689 lockdep_is_held(&mvm->mutex)); 1690 RCU_INIT_POINTER(mvmvif->probe_resp_data, NULL); 1691 if (probe_data) 1692 kfree_rcu(probe_data, rcu_head); 1693 1694 if (mvm->bf_allowed_vif == mvmvif) { 1695 mvm->bf_allowed_vif = NULL; 1696 vif->driver_flags &= ~(IEEE80211_VIF_BEACON_FILTER | 1697 IEEE80211_VIF_SUPPORTS_CQM_RSSI); 1698 } 1699 1700 if (vif->bss_conf.ftm_responder) 1701 memset(&mvm->ftm_resp_stats, 0, sizeof(mvm->ftm_resp_stats)); 1702 1703 iwl_mvm_vif_dbgfs_clean(mvm, vif); 1704 1705 /* 1706 * For AP/GO interface, the tear down of the resources allocated to the 1707 * interface is be handled as part of the stop_ap flow. 1708 */ 1709 if (vif->type == NL80211_IFTYPE_AP || 1710 vif->type == NL80211_IFTYPE_ADHOC) { 1711 #ifdef CONFIG_NL80211_TESTMODE 1712 if (vif == mvm->noa_vif) { 1713 mvm->noa_vif = NULL; 1714 mvm->noa_duration = 0; 1715 } 1716 #endif 1717 iwl_mvm_dealloc_int_sta(mvm, &mvmvif->mcast_sta); 1718 iwl_mvm_dealloc_bcast_sta(mvm, vif); 1719 goto out_release; 1720 } 1721 1722 if (vif->type == NL80211_IFTYPE_P2P_DEVICE) { 1723 mvm->p2p_device_vif = NULL; 1724 iwl_mvm_rm_p2p_bcast_sta(mvm, vif); 1725 iwl_mvm_binding_remove_vif(mvm, vif); 1726 iwl_mvm_phy_ctxt_unref(mvm, mvmvif->phy_ctxt); 1727 mvmvif->phy_ctxt = NULL; 1728 } 1729 1730 if (mvm->vif_count && vif->type != NL80211_IFTYPE_P2P_DEVICE) 1731 mvm->vif_count--; 1732 1733 iwl_mvm_power_update_mac(mvm); 1734 iwl_mvm_mac_ctxt_remove(mvm, vif); 1735 1736 RCU_INIT_POINTER(mvm->vif_id_to_mac[mvmvif->id], NULL); 1737 1738 if (vif->type == NL80211_IFTYPE_MONITOR) 1739 mvm->monitor_on = false; 1740 1741 out_release: 1742 mutex_unlock(&mvm->mutex); 1743 } 1744 1745 static int iwl_mvm_mac_config(struct ieee80211_hw *hw, u32 changed) 1746 { 1747 return 0; 1748 } 1749 1750 struct iwl_mvm_mc_iter_data { 1751 struct iwl_mvm *mvm; 1752 int port_id; 1753 }; 1754 1755 static void iwl_mvm_mc_iface_iterator(void *_data, u8 *mac, 1756 struct ieee80211_vif *vif) 1757 { 1758 struct iwl_mvm_mc_iter_data *data = _data; 1759 struct iwl_mvm *mvm = data->mvm; 1760 struct iwl_mcast_filter_cmd *cmd = mvm->mcast_filter_cmd; 1761 struct iwl_host_cmd hcmd = { 1762 .id = MCAST_FILTER_CMD, 1763 .flags = CMD_ASYNC, 1764 .dataflags[0] = IWL_HCMD_DFL_NOCOPY, 1765 }; 1766 int ret, len; 1767 1768 /* if we don't have free ports, mcast frames will be dropped */ 1769 if (WARN_ON_ONCE(data->port_id >= MAX_PORT_ID_NUM)) 1770 return; 1771 1772 if (vif->type != NL80211_IFTYPE_STATION || 1773 !vif->bss_conf.assoc) 1774 return; 1775 1776 cmd->port_id = data->port_id++; 1777 memcpy(cmd->bssid, vif->bss_conf.bssid, ETH_ALEN); 1778 len = roundup(sizeof(*cmd) + cmd->count * ETH_ALEN, 4); 1779 1780 hcmd.len[0] = len; 1781 hcmd.data[0] = cmd; 1782 1783 ret = iwl_mvm_send_cmd(mvm, &hcmd); 1784 if (ret) 1785 IWL_ERR(mvm, "mcast filter cmd error. ret=%d\n", ret); 1786 } 1787 1788 static void iwl_mvm_recalc_multicast(struct iwl_mvm *mvm) 1789 { 1790 struct iwl_mvm_mc_iter_data iter_data = { 1791 .mvm = mvm, 1792 }; 1793 1794 lockdep_assert_held(&mvm->mutex); 1795 1796 if (WARN_ON_ONCE(!mvm->mcast_filter_cmd)) 1797 return; 1798 1799 ieee80211_iterate_active_interfaces_atomic( 1800 mvm->hw, IEEE80211_IFACE_ITER_NORMAL, 1801 iwl_mvm_mc_iface_iterator, &iter_data); 1802 } 1803 1804 static u64 iwl_mvm_prepare_multicast(struct ieee80211_hw *hw, 1805 struct netdev_hw_addr_list *mc_list) 1806 { 1807 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1808 struct iwl_mcast_filter_cmd *cmd; 1809 struct netdev_hw_addr *addr; 1810 int addr_count; 1811 bool pass_all; 1812 int len; 1813 1814 addr_count = netdev_hw_addr_list_count(mc_list); 1815 pass_all = addr_count > MAX_MCAST_FILTERING_ADDRESSES || 1816 IWL_MVM_FW_MCAST_FILTER_PASS_ALL; 1817 if (pass_all) 1818 addr_count = 0; 1819 1820 len = roundup(sizeof(*cmd) + addr_count * ETH_ALEN, 4); 1821 cmd = kzalloc(len, GFP_ATOMIC); 1822 if (!cmd) 1823 return 0; 1824 1825 if (pass_all) { 1826 cmd->pass_all = 1; 1827 return (u64)(unsigned long)cmd; 1828 } 1829 1830 netdev_hw_addr_list_for_each(addr, mc_list) { 1831 IWL_DEBUG_MAC80211(mvm, "mcast addr (%d): %pM\n", 1832 cmd->count, addr->addr); 1833 memcpy(&cmd->addr_list[cmd->count * ETH_ALEN], 1834 addr->addr, ETH_ALEN); 1835 cmd->count++; 1836 } 1837 1838 return (u64)(unsigned long)cmd; 1839 } 1840 1841 static void iwl_mvm_configure_filter(struct ieee80211_hw *hw, 1842 unsigned int changed_flags, 1843 unsigned int *total_flags, 1844 u64 multicast) 1845 { 1846 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1847 struct iwl_mcast_filter_cmd *cmd = (void *)(unsigned long)multicast; 1848 1849 mutex_lock(&mvm->mutex); 1850 1851 /* replace previous configuration */ 1852 kfree(mvm->mcast_filter_cmd); 1853 mvm->mcast_filter_cmd = cmd; 1854 1855 if (!cmd) 1856 goto out; 1857 1858 if (changed_flags & FIF_ALLMULTI) 1859 cmd->pass_all = !!(*total_flags & FIF_ALLMULTI); 1860 1861 if (cmd->pass_all) 1862 cmd->count = 0; 1863 1864 iwl_mvm_recalc_multicast(mvm); 1865 out: 1866 mutex_unlock(&mvm->mutex); 1867 *total_flags = 0; 1868 } 1869 1870 static void iwl_mvm_config_iface_filter(struct ieee80211_hw *hw, 1871 struct ieee80211_vif *vif, 1872 unsigned int filter_flags, 1873 unsigned int changed_flags) 1874 { 1875 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1876 1877 /* We support only filter for probe requests */ 1878 if (!(changed_flags & FIF_PROBE_REQ)) 1879 return; 1880 1881 /* Supported only for p2p client interfaces */ 1882 if (vif->type != NL80211_IFTYPE_STATION || !vif->bss_conf.assoc || 1883 !vif->p2p) 1884 return; 1885 1886 mutex_lock(&mvm->mutex); 1887 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 1888 mutex_unlock(&mvm->mutex); 1889 } 1890 1891 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING 1892 struct iwl_bcast_iter_data { 1893 struct iwl_mvm *mvm; 1894 struct iwl_bcast_filter_cmd *cmd; 1895 u8 current_filter; 1896 }; 1897 1898 static void 1899 iwl_mvm_set_bcast_filter(struct ieee80211_vif *vif, 1900 const struct iwl_fw_bcast_filter *in_filter, 1901 struct iwl_fw_bcast_filter *out_filter) 1902 { 1903 struct iwl_fw_bcast_filter_attr *attr; 1904 int i; 1905 1906 memcpy(out_filter, in_filter, sizeof(*out_filter)); 1907 1908 for (i = 0; i < ARRAY_SIZE(out_filter->attrs); i++) { 1909 attr = &out_filter->attrs[i]; 1910 1911 if (!attr->mask) 1912 break; 1913 1914 switch (attr->reserved1) { 1915 case cpu_to_le16(BC_FILTER_MAGIC_IP): 1916 if (vif->bss_conf.arp_addr_cnt != 1) { 1917 attr->mask = 0; 1918 continue; 1919 } 1920 1921 attr->val = vif->bss_conf.arp_addr_list[0]; 1922 break; 1923 case cpu_to_le16(BC_FILTER_MAGIC_MAC): 1924 attr->val = *(__be32 *)&vif->addr[2]; 1925 break; 1926 default: 1927 break; 1928 } 1929 attr->reserved1 = 0; 1930 out_filter->num_attrs++; 1931 } 1932 } 1933 1934 static void iwl_mvm_bcast_filter_iterator(void *_data, u8 *mac, 1935 struct ieee80211_vif *vif) 1936 { 1937 struct iwl_bcast_iter_data *data = _data; 1938 struct iwl_mvm *mvm = data->mvm; 1939 struct iwl_bcast_filter_cmd *cmd = data->cmd; 1940 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 1941 struct iwl_fw_bcast_mac *bcast_mac; 1942 int i; 1943 1944 if (WARN_ON(mvmvif->id >= ARRAY_SIZE(cmd->macs))) 1945 return; 1946 1947 bcast_mac = &cmd->macs[mvmvif->id]; 1948 1949 /* 1950 * enable filtering only for associated stations, but not for P2P 1951 * Clients 1952 */ 1953 if (vif->type != NL80211_IFTYPE_STATION || vif->p2p || 1954 !vif->bss_conf.assoc) 1955 return; 1956 1957 bcast_mac->default_discard = 1; 1958 1959 /* copy all configured filters */ 1960 for (i = 0; mvm->bcast_filters[i].attrs[0].mask; i++) { 1961 /* 1962 * Make sure we don't exceed our filters limit. 1963 * if there is still a valid filter to be configured, 1964 * be on the safe side and just allow bcast for this mac. 1965 */ 1966 if (WARN_ON_ONCE(data->current_filter >= 1967 ARRAY_SIZE(cmd->filters))) { 1968 bcast_mac->default_discard = 0; 1969 bcast_mac->attached_filters = 0; 1970 break; 1971 } 1972 1973 iwl_mvm_set_bcast_filter(vif, 1974 &mvm->bcast_filters[i], 1975 &cmd->filters[data->current_filter]); 1976 1977 /* skip current filter if it contains no attributes */ 1978 if (!cmd->filters[data->current_filter].num_attrs) 1979 continue; 1980 1981 /* attach the filter to current mac */ 1982 bcast_mac->attached_filters |= 1983 cpu_to_le16(BIT(data->current_filter)); 1984 1985 data->current_filter++; 1986 } 1987 } 1988 1989 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm, 1990 struct iwl_bcast_filter_cmd *cmd) 1991 { 1992 struct iwl_bcast_iter_data iter_data = { 1993 .mvm = mvm, 1994 .cmd = cmd, 1995 }; 1996 1997 if (IWL_MVM_FW_BCAST_FILTER_PASS_ALL) 1998 return false; 1999 2000 memset(cmd, 0, sizeof(*cmd)); 2001 cmd->max_bcast_filters = ARRAY_SIZE(cmd->filters); 2002 cmd->max_macs = ARRAY_SIZE(cmd->macs); 2003 2004 #ifdef CONFIG_IWLWIFI_DEBUGFS 2005 /* use debugfs filters/macs if override is configured */ 2006 if (mvm->dbgfs_bcast_filtering.override) { 2007 memcpy(cmd->filters, &mvm->dbgfs_bcast_filtering.cmd.filters, 2008 sizeof(cmd->filters)); 2009 memcpy(cmd->macs, &mvm->dbgfs_bcast_filtering.cmd.macs, 2010 sizeof(cmd->macs)); 2011 return true; 2012 } 2013 #endif 2014 2015 /* if no filters are configured, do nothing */ 2016 if (!mvm->bcast_filters) 2017 return false; 2018 2019 /* configure and attach these filters for each associated sta vif */ 2020 ieee80211_iterate_active_interfaces( 2021 mvm->hw, IEEE80211_IFACE_ITER_NORMAL, 2022 iwl_mvm_bcast_filter_iterator, &iter_data); 2023 2024 return true; 2025 } 2026 2027 static int iwl_mvm_configure_bcast_filter(struct iwl_mvm *mvm) 2028 { 2029 struct iwl_bcast_filter_cmd cmd; 2030 2031 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_BCAST_FILTERING)) 2032 return 0; 2033 2034 if (!iwl_mvm_bcast_filter_build_cmd(mvm, &cmd)) 2035 return 0; 2036 2037 return iwl_mvm_send_cmd_pdu(mvm, BCAST_FILTER_CMD, 0, 2038 sizeof(cmd), &cmd); 2039 } 2040 #else 2041 static inline int iwl_mvm_configure_bcast_filter(struct iwl_mvm *mvm) 2042 { 2043 return 0; 2044 } 2045 #endif 2046 2047 static int iwl_mvm_update_mu_groups(struct iwl_mvm *mvm, 2048 struct ieee80211_vif *vif) 2049 { 2050 struct iwl_mu_group_mgmt_cmd cmd = {}; 2051 2052 memcpy(cmd.membership_status, vif->bss_conf.mu_group.membership, 2053 WLAN_MEMBERSHIP_LEN); 2054 memcpy(cmd.user_position, vif->bss_conf.mu_group.position, 2055 WLAN_USER_POSITION_LEN); 2056 2057 return iwl_mvm_send_cmd_pdu(mvm, 2058 WIDE_ID(DATA_PATH_GROUP, 2059 UPDATE_MU_GROUPS_CMD), 2060 0, sizeof(cmd), &cmd); 2061 } 2062 2063 static void iwl_mvm_mu_mimo_iface_iterator(void *_data, u8 *mac, 2064 struct ieee80211_vif *vif) 2065 { 2066 if (vif->mu_mimo_owner) { 2067 struct iwl_mu_group_mgmt_notif *notif = _data; 2068 2069 /* 2070 * MU-MIMO Group Id action frame is little endian. We treat 2071 * the data received from firmware as if it came from the 2072 * action frame, so no conversion is needed. 2073 */ 2074 ieee80211_update_mu_groups(vif, 2075 (u8 *)¬if->membership_status, 2076 (u8 *)¬if->user_position); 2077 } 2078 } 2079 2080 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm, 2081 struct iwl_rx_cmd_buffer *rxb) 2082 { 2083 struct iwl_rx_packet *pkt = rxb_addr(rxb); 2084 struct iwl_mu_group_mgmt_notif *notif = (void *)pkt->data; 2085 2086 ieee80211_iterate_active_interfaces_atomic( 2087 mvm->hw, IEEE80211_IFACE_ITER_NORMAL, 2088 iwl_mvm_mu_mimo_iface_iterator, notif); 2089 } 2090 2091 static u8 iwl_mvm_he_get_ppe_val(u8 *ppe, u8 ppe_pos_bit) 2092 { 2093 u8 byte_num = ppe_pos_bit / 8; 2094 u8 bit_num = ppe_pos_bit % 8; 2095 u8 residue_bits; 2096 u8 res; 2097 2098 if (bit_num <= 5) 2099 return (ppe[byte_num] >> bit_num) & 2100 (BIT(IEEE80211_PPE_THRES_INFO_PPET_SIZE) - 1); 2101 2102 /* 2103 * If bit_num > 5, we have to combine bits with next byte. 2104 * Calculate how many bits we need to take from current byte (called 2105 * here "residue_bits"), and add them to bits from next byte. 2106 */ 2107 2108 residue_bits = 8 - bit_num; 2109 2110 res = (ppe[byte_num + 1] & 2111 (BIT(IEEE80211_PPE_THRES_INFO_PPET_SIZE - residue_bits) - 1)) << 2112 residue_bits; 2113 res += (ppe[byte_num] >> bit_num) & (BIT(residue_bits) - 1); 2114 2115 return res; 2116 } 2117 2118 static void iwl_mvm_cfg_he_sta(struct iwl_mvm *mvm, 2119 struct ieee80211_vif *vif, u8 sta_id) 2120 { 2121 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 2122 struct iwl_he_sta_context_cmd sta_ctxt_cmd = { 2123 .sta_id = sta_id, 2124 .tid_limit = IWL_MAX_TID_COUNT, 2125 .bss_color = vif->bss_conf.bss_color, 2126 .htc_trig_based_pkt_ext = vif->bss_conf.htc_trig_based_pkt_ext, 2127 .frame_time_rts_th = 2128 cpu_to_le16(vif->bss_conf.frame_time_rts_th), 2129 }; 2130 struct ieee80211_sta *sta; 2131 u32 flags; 2132 int i; 2133 2134 rcu_read_lock(); 2135 2136 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_ctxt_cmd.sta_id]); 2137 if (IS_ERR(sta)) { 2138 rcu_read_unlock(); 2139 WARN(1, "Can't find STA to configure HE\n"); 2140 return; 2141 } 2142 2143 if (!sta->he_cap.has_he) { 2144 rcu_read_unlock(); 2145 return; 2146 } 2147 2148 flags = 0; 2149 2150 /* HTC flags */ 2151 if (sta->he_cap.he_cap_elem.mac_cap_info[0] & 2152 IEEE80211_HE_MAC_CAP0_HTC_HE) 2153 sta_ctxt_cmd.htc_flags |= cpu_to_le32(IWL_HE_HTC_SUPPORT); 2154 if ((sta->he_cap.he_cap_elem.mac_cap_info[1] & 2155 IEEE80211_HE_MAC_CAP1_LINK_ADAPTATION) || 2156 (sta->he_cap.he_cap_elem.mac_cap_info[2] & 2157 IEEE80211_HE_MAC_CAP2_LINK_ADAPTATION)) { 2158 u8 link_adap = 2159 ((sta->he_cap.he_cap_elem.mac_cap_info[2] & 2160 IEEE80211_HE_MAC_CAP2_LINK_ADAPTATION) << 1) + 2161 (sta->he_cap.he_cap_elem.mac_cap_info[1] & 2162 IEEE80211_HE_MAC_CAP1_LINK_ADAPTATION); 2163 2164 if (link_adap == 2) 2165 sta_ctxt_cmd.htc_flags |= 2166 cpu_to_le32(IWL_HE_HTC_LINK_ADAP_UNSOLICITED); 2167 else if (link_adap == 3) 2168 sta_ctxt_cmd.htc_flags |= 2169 cpu_to_le32(IWL_HE_HTC_LINK_ADAP_BOTH); 2170 } 2171 if (sta->he_cap.he_cap_elem.mac_cap_info[2] & IEEE80211_HE_MAC_CAP2_BSR) 2172 sta_ctxt_cmd.htc_flags |= cpu_to_le32(IWL_HE_HTC_BSR_SUPP); 2173 if (sta->he_cap.he_cap_elem.mac_cap_info[3] & 2174 IEEE80211_HE_MAC_CAP3_OMI_CONTROL) 2175 sta_ctxt_cmd.htc_flags |= cpu_to_le32(IWL_HE_HTC_OMI_SUPP); 2176 if (sta->he_cap.he_cap_elem.mac_cap_info[4] & IEEE80211_HE_MAC_CAP4_BQR) 2177 sta_ctxt_cmd.htc_flags |= cpu_to_le32(IWL_HE_HTC_BQR_SUPP); 2178 2179 /* 2180 * Initialize the PPE thresholds to "None" (7), as described in Table 2181 * 9-262ac of 80211.ax/D3.0. 2182 */ 2183 memset(&sta_ctxt_cmd.pkt_ext, 7, sizeof(sta_ctxt_cmd.pkt_ext)); 2184 2185 /* If PPE Thresholds exist, parse them into a FW-familiar format. */ 2186 if (sta->he_cap.he_cap_elem.phy_cap_info[6] & 2187 IEEE80211_HE_PHY_CAP6_PPE_THRESHOLD_PRESENT) { 2188 u8 nss = (sta->he_cap.ppe_thres[0] & 2189 IEEE80211_PPE_THRES_NSS_MASK) + 1; 2190 u8 ru_index_bitmap = 2191 (sta->he_cap.ppe_thres[0] & 2192 IEEE80211_PPE_THRES_RU_INDEX_BITMASK_MASK) >> 2193 IEEE80211_PPE_THRES_RU_INDEX_BITMASK_POS; 2194 u8 *ppe = &sta->he_cap.ppe_thres[0]; 2195 u8 ppe_pos_bit = 7; /* Starting after PPE header */ 2196 2197 /* 2198 * FW currently supports only nss == MAX_HE_SUPP_NSS 2199 * 2200 * If nss > MAX: we can ignore values we don't support 2201 * If nss < MAX: we can set zeros in other streams 2202 */ 2203 if (nss > MAX_HE_SUPP_NSS) { 2204 IWL_INFO(mvm, "Got NSS = %d - trimming to %d\n", nss, 2205 MAX_HE_SUPP_NSS); 2206 nss = MAX_HE_SUPP_NSS; 2207 } 2208 2209 for (i = 0; i < nss; i++) { 2210 u8 ru_index_tmp = ru_index_bitmap << 1; 2211 u8 bw; 2212 2213 for (bw = 0; bw < MAX_HE_CHANNEL_BW_INDX; bw++) { 2214 ru_index_tmp >>= 1; 2215 if (!(ru_index_tmp & 1)) 2216 continue; 2217 2218 sta_ctxt_cmd.pkt_ext.pkt_ext_qam_th[i][bw][1] = 2219 iwl_mvm_he_get_ppe_val(ppe, 2220 ppe_pos_bit); 2221 ppe_pos_bit += 2222 IEEE80211_PPE_THRES_INFO_PPET_SIZE; 2223 sta_ctxt_cmd.pkt_ext.pkt_ext_qam_th[i][bw][0] = 2224 iwl_mvm_he_get_ppe_val(ppe, 2225 ppe_pos_bit); 2226 ppe_pos_bit += 2227 IEEE80211_PPE_THRES_INFO_PPET_SIZE; 2228 } 2229 } 2230 2231 flags |= STA_CTXT_HE_PACKET_EXT; 2232 } else if ((sta->he_cap.he_cap_elem.phy_cap_info[9] & 2233 IEEE80211_HE_PHY_CAP9_NOMIMAL_PKT_PADDING_MASK) != 2234 IEEE80211_HE_PHY_CAP9_NOMIMAL_PKT_PADDING_RESERVED) { 2235 int low_th = -1; 2236 int high_th = -1; 2237 2238 /* Take the PPE thresholds from the nominal padding info */ 2239 switch (sta->he_cap.he_cap_elem.phy_cap_info[9] & 2240 IEEE80211_HE_PHY_CAP9_NOMIMAL_PKT_PADDING_MASK) { 2241 case IEEE80211_HE_PHY_CAP9_NOMIMAL_PKT_PADDING_0US: 2242 low_th = IWL_HE_PKT_EXT_NONE; 2243 high_th = IWL_HE_PKT_EXT_NONE; 2244 break; 2245 case IEEE80211_HE_PHY_CAP9_NOMIMAL_PKT_PADDING_8US: 2246 low_th = IWL_HE_PKT_EXT_BPSK; 2247 high_th = IWL_HE_PKT_EXT_NONE; 2248 break; 2249 case IEEE80211_HE_PHY_CAP9_NOMIMAL_PKT_PADDING_16US: 2250 low_th = IWL_HE_PKT_EXT_NONE; 2251 high_th = IWL_HE_PKT_EXT_BPSK; 2252 break; 2253 } 2254 2255 /* Set the PPE thresholds accordingly */ 2256 if (low_th >= 0 && high_th >= 0) { 2257 u8 ***pkt_ext_qam = 2258 (void *)sta_ctxt_cmd.pkt_ext.pkt_ext_qam_th; 2259 2260 for (i = 0; i < MAX_HE_SUPP_NSS; i++) { 2261 u8 bw; 2262 2263 for (bw = 0; bw < MAX_HE_CHANNEL_BW_INDX; 2264 bw++) { 2265 pkt_ext_qam[i][bw][0] = low_th; 2266 pkt_ext_qam[i][bw][1] = high_th; 2267 } 2268 } 2269 2270 flags |= STA_CTXT_HE_PACKET_EXT; 2271 } 2272 } 2273 rcu_read_unlock(); 2274 2275 /* Mark MU EDCA as enabled, unless none detected on some AC */ 2276 flags |= STA_CTXT_HE_MU_EDCA_CW; 2277 for (i = 0; i < AC_NUM; i++) { 2278 struct ieee80211_he_mu_edca_param_ac_rec *mu_edca = 2279 &mvmvif->queue_params[i].mu_edca_param_rec; 2280 2281 if (!mvmvif->queue_params[i].mu_edca) { 2282 flags &= ~STA_CTXT_HE_MU_EDCA_CW; 2283 break; 2284 } 2285 2286 sta_ctxt_cmd.trig_based_txf[i].cwmin = 2287 cpu_to_le16(mu_edca->ecw_min_max & 0xf); 2288 sta_ctxt_cmd.trig_based_txf[i].cwmax = 2289 cpu_to_le16((mu_edca->ecw_min_max & 0xf0) >> 4); 2290 sta_ctxt_cmd.trig_based_txf[i].aifsn = 2291 cpu_to_le16(mu_edca->aifsn); 2292 sta_ctxt_cmd.trig_based_txf[i].mu_time = 2293 cpu_to_le16(mu_edca->mu_edca_timer); 2294 } 2295 2296 if (vif->bss_conf.multi_sta_back_32bit) 2297 flags |= STA_CTXT_HE_32BIT_BA_BITMAP; 2298 2299 if (vif->bss_conf.ack_enabled) 2300 flags |= STA_CTXT_HE_ACK_ENABLED; 2301 2302 if (vif->bss_conf.uora_exists) { 2303 flags |= STA_CTXT_HE_TRIG_RND_ALLOC; 2304 2305 sta_ctxt_cmd.rand_alloc_ecwmin = 2306 vif->bss_conf.uora_ocw_range & 0x7; 2307 sta_ctxt_cmd.rand_alloc_ecwmax = 2308 (vif->bss_conf.uora_ocw_range >> 3) & 0x7; 2309 } 2310 2311 /* TODO: support Multi BSSID IE */ 2312 2313 sta_ctxt_cmd.flags = cpu_to_le32(flags); 2314 2315 if (iwl_mvm_send_cmd_pdu(mvm, iwl_cmd_id(STA_HE_CTXT_CMD, 2316 DATA_PATH_GROUP, 0), 2317 0, sizeof(sta_ctxt_cmd), &sta_ctxt_cmd)) 2318 IWL_ERR(mvm, "Failed to config FW to work HE!\n"); 2319 } 2320 2321 static void iwl_mvm_bss_info_changed_station(struct iwl_mvm *mvm, 2322 struct ieee80211_vif *vif, 2323 struct ieee80211_bss_conf *bss_conf, 2324 u32 changes) 2325 { 2326 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 2327 int ret; 2328 2329 /* 2330 * Re-calculate the tsf id, as the master-slave relations depend on the 2331 * beacon interval, which was not known when the station interface was 2332 * added. 2333 */ 2334 if (changes & BSS_CHANGED_ASSOC && bss_conf->assoc) { 2335 if (vif->bss_conf.he_support && 2336 !iwlwifi_mod_params.disable_11ax) 2337 iwl_mvm_cfg_he_sta(mvm, vif, mvmvif->ap_sta_id); 2338 2339 iwl_mvm_mac_ctxt_recalc_tsf_id(mvm, vif); 2340 } 2341 2342 /* Update MU EDCA params */ 2343 if (changes & BSS_CHANGED_QOS && mvmvif->associated && 2344 bss_conf->assoc && vif->bss_conf.he_support && 2345 !iwlwifi_mod_params.disable_11ax) 2346 iwl_mvm_cfg_he_sta(mvm, vif, mvmvif->ap_sta_id); 2347 2348 /* 2349 * If we're not associated yet, take the (new) BSSID before associating 2350 * so the firmware knows. If we're already associated, then use the old 2351 * BSSID here, and we'll send a cleared one later in the CHANGED_ASSOC 2352 * branch for disassociation below. 2353 */ 2354 if (changes & BSS_CHANGED_BSSID && !mvmvif->associated) 2355 memcpy(mvmvif->bssid, bss_conf->bssid, ETH_ALEN); 2356 2357 ret = iwl_mvm_mac_ctxt_changed(mvm, vif, false, mvmvif->bssid); 2358 if (ret) 2359 IWL_ERR(mvm, "failed to update MAC %pM\n", vif->addr); 2360 2361 /* after sending it once, adopt mac80211 data */ 2362 memcpy(mvmvif->bssid, bss_conf->bssid, ETH_ALEN); 2363 mvmvif->associated = bss_conf->assoc; 2364 2365 if (changes & BSS_CHANGED_ASSOC) { 2366 if (bss_conf->assoc) { 2367 /* clear statistics to get clean beacon counter */ 2368 iwl_mvm_request_statistics(mvm, true); 2369 memset(&mvmvif->beacon_stats, 0, 2370 sizeof(mvmvif->beacon_stats)); 2371 2372 /* add quota for this interface */ 2373 ret = iwl_mvm_update_quotas(mvm, true, NULL); 2374 if (ret) { 2375 IWL_ERR(mvm, "failed to update quotas\n"); 2376 return; 2377 } 2378 2379 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART, 2380 &mvm->status)) { 2381 /* 2382 * If we're restarting then the firmware will 2383 * obviously have lost synchronisation with 2384 * the AP. It will attempt to synchronise by 2385 * itself, but we can make it more reliable by 2386 * scheduling a session protection time event. 2387 * 2388 * The firmware needs to receive a beacon to 2389 * catch up with synchronisation, use 110% of 2390 * the beacon interval. 2391 * 2392 * Set a large maximum delay to allow for more 2393 * than a single interface. 2394 */ 2395 u32 dur = (11 * vif->bss_conf.beacon_int) / 10; 2396 iwl_mvm_protect_session(mvm, vif, dur, dur, 2397 5 * dur, false); 2398 } 2399 2400 iwl_mvm_sf_update(mvm, vif, false); 2401 iwl_mvm_power_vif_assoc(mvm, vif); 2402 if (vif->p2p) { 2403 iwl_mvm_ref(mvm, IWL_MVM_REF_P2P_CLIENT); 2404 iwl_mvm_update_smps(mvm, vif, 2405 IWL_MVM_SMPS_REQ_PROT, 2406 IEEE80211_SMPS_DYNAMIC); 2407 } 2408 } else if (mvmvif->ap_sta_id != IWL_MVM_INVALID_STA) { 2409 /* 2410 * If update fails - SF might be running in associated 2411 * mode while disassociated - which is forbidden. 2412 */ 2413 ret = iwl_mvm_sf_update(mvm, vif, false); 2414 WARN_ONCE(ret && 2415 !test_bit(IWL_MVM_STATUS_HW_RESTART_REQUESTED, 2416 &mvm->status), 2417 "Failed to update SF upon disassociation\n"); 2418 2419 /* 2420 * If we get an assert during the connection (after the 2421 * station has been added, but before the vif is set 2422 * to associated), mac80211 will re-add the station and 2423 * then configure the vif. Since the vif is not 2424 * associated, we would remove the station here and 2425 * this would fail the recovery. 2426 */ 2427 if (!test_bit(IWL_MVM_STATUS_IN_HW_RESTART, 2428 &mvm->status)) { 2429 /* 2430 * Remove AP station now that 2431 * the MAC is unassoc 2432 */ 2433 ret = iwl_mvm_rm_sta_id(mvm, vif, 2434 mvmvif->ap_sta_id); 2435 if (ret) 2436 IWL_ERR(mvm, 2437 "failed to remove AP station\n"); 2438 2439 if (mvm->d0i3_ap_sta_id == mvmvif->ap_sta_id) 2440 mvm->d0i3_ap_sta_id = 2441 IWL_MVM_INVALID_STA; 2442 mvmvif->ap_sta_id = IWL_MVM_INVALID_STA; 2443 } 2444 2445 /* remove quota for this interface */ 2446 ret = iwl_mvm_update_quotas(mvm, false, NULL); 2447 if (ret) 2448 IWL_ERR(mvm, "failed to update quotas\n"); 2449 2450 if (vif->p2p) 2451 iwl_mvm_unref(mvm, IWL_MVM_REF_P2P_CLIENT); 2452 2453 /* this will take the cleared BSSID from bss_conf */ 2454 ret = iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 2455 if (ret) 2456 IWL_ERR(mvm, 2457 "failed to update MAC %pM (clear after unassoc)\n", 2458 vif->addr); 2459 } 2460 2461 /* 2462 * The firmware tracks the MU-MIMO group on its own. 2463 * However, on HW restart we should restore this data. 2464 */ 2465 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status) && 2466 (changes & BSS_CHANGED_MU_GROUPS) && vif->mu_mimo_owner) { 2467 ret = iwl_mvm_update_mu_groups(mvm, vif); 2468 if (ret) 2469 IWL_ERR(mvm, 2470 "failed to update VHT MU_MIMO groups\n"); 2471 } 2472 2473 iwl_mvm_recalc_multicast(mvm); 2474 iwl_mvm_configure_bcast_filter(mvm); 2475 2476 /* reset rssi values */ 2477 mvmvif->bf_data.ave_beacon_signal = 0; 2478 2479 iwl_mvm_bt_coex_vif_change(mvm); 2480 iwl_mvm_update_smps(mvm, vif, IWL_MVM_SMPS_REQ_TT, 2481 IEEE80211_SMPS_AUTOMATIC); 2482 if (fw_has_capa(&mvm->fw->ucode_capa, 2483 IWL_UCODE_TLV_CAPA_UMAC_SCAN)) 2484 iwl_mvm_config_scan(mvm); 2485 } 2486 2487 if (changes & BSS_CHANGED_BEACON_INFO) { 2488 /* 2489 * We received a beacon from the associated AP so 2490 * remove the session protection. 2491 */ 2492 iwl_mvm_stop_session_protection(mvm, vif); 2493 2494 iwl_mvm_sf_update(mvm, vif, false); 2495 WARN_ON(iwl_mvm_enable_beacon_filter(mvm, vif, 0)); 2496 } 2497 2498 if (changes & (BSS_CHANGED_PS | BSS_CHANGED_P2P_PS | BSS_CHANGED_QOS | 2499 /* 2500 * Send power command on every beacon change, 2501 * because we may have not enabled beacon abort yet. 2502 */ 2503 BSS_CHANGED_BEACON_INFO)) { 2504 ret = iwl_mvm_power_update_mac(mvm); 2505 if (ret) 2506 IWL_ERR(mvm, "failed to update power mode\n"); 2507 } 2508 2509 if (changes & BSS_CHANGED_TXPOWER) { 2510 IWL_DEBUG_CALIB(mvm, "Changing TX Power to %d\n", 2511 bss_conf->txpower); 2512 iwl_mvm_set_tx_power(mvm, vif, bss_conf->txpower); 2513 } 2514 2515 if (changes & BSS_CHANGED_CQM) { 2516 IWL_DEBUG_MAC80211(mvm, "cqm info_changed\n"); 2517 /* reset cqm events tracking */ 2518 mvmvif->bf_data.last_cqm_event = 0; 2519 if (mvmvif->bf_data.bf_enabled) { 2520 ret = iwl_mvm_enable_beacon_filter(mvm, vif, 0); 2521 if (ret) 2522 IWL_ERR(mvm, 2523 "failed to update CQM thresholds\n"); 2524 } 2525 } 2526 2527 if (changes & BSS_CHANGED_ARP_FILTER) { 2528 IWL_DEBUG_MAC80211(mvm, "arp filter changed\n"); 2529 iwl_mvm_configure_bcast_filter(mvm); 2530 } 2531 } 2532 2533 static int iwl_mvm_start_ap_ibss(struct ieee80211_hw *hw, 2534 struct ieee80211_vif *vif) 2535 { 2536 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2537 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 2538 int ret; 2539 2540 /* 2541 * iwl_mvm_mac_ctxt_add() might read directly from the device 2542 * (the system time), so make sure it is available. 2543 */ 2544 ret = iwl_mvm_ref_sync(mvm, IWL_MVM_REF_START_AP); 2545 if (ret) 2546 return ret; 2547 2548 mutex_lock(&mvm->mutex); 2549 2550 /* Send the beacon template */ 2551 ret = iwl_mvm_mac_ctxt_beacon_changed(mvm, vif); 2552 if (ret) 2553 goto out_unlock; 2554 2555 /* 2556 * Re-calculate the tsf id, as the master-slave relations depend on the 2557 * beacon interval, which was not known when the AP interface was added. 2558 */ 2559 if (vif->type == NL80211_IFTYPE_AP) 2560 iwl_mvm_mac_ctxt_recalc_tsf_id(mvm, vif); 2561 2562 mvmvif->ap_assoc_sta_count = 0; 2563 2564 /* Add the mac context */ 2565 ret = iwl_mvm_mac_ctxt_add(mvm, vif); 2566 if (ret) 2567 goto out_unlock; 2568 2569 /* Perform the binding */ 2570 ret = iwl_mvm_binding_add_vif(mvm, vif); 2571 if (ret) 2572 goto out_remove; 2573 2574 /* 2575 * This is not very nice, but the simplest: 2576 * For older FWs adding the mcast sta before the bcast station may 2577 * cause assert 0x2b00. 2578 * This is fixed in later FW so make the order of removal depend on 2579 * the TLV 2580 */ 2581 if (fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_STA_TYPE)) { 2582 ret = iwl_mvm_add_mcast_sta(mvm, vif); 2583 if (ret) 2584 goto out_unbind; 2585 /* 2586 * Send the bcast station. At this stage the TBTT and DTIM time 2587 * events are added and applied to the scheduler 2588 */ 2589 ret = iwl_mvm_send_add_bcast_sta(mvm, vif); 2590 if (ret) { 2591 iwl_mvm_rm_mcast_sta(mvm, vif); 2592 goto out_unbind; 2593 } 2594 } else { 2595 /* 2596 * Send the bcast station. At this stage the TBTT and DTIM time 2597 * events are added and applied to the scheduler 2598 */ 2599 ret = iwl_mvm_send_add_bcast_sta(mvm, vif); 2600 if (ret) 2601 goto out_unbind; 2602 ret = iwl_mvm_add_mcast_sta(mvm, vif); 2603 if (ret) { 2604 iwl_mvm_send_rm_bcast_sta(mvm, vif); 2605 goto out_unbind; 2606 } 2607 } 2608 2609 /* must be set before quota calculations */ 2610 mvmvif->ap_ibss_active = true; 2611 2612 if (vif->type == NL80211_IFTYPE_AP && !vif->p2p) { 2613 iwl_mvm_vif_set_low_latency(mvmvif, true, 2614 LOW_LATENCY_VIF_TYPE); 2615 iwl_mvm_send_low_latency_cmd(mvm, true, mvmvif->id); 2616 } 2617 2618 /* power updated needs to be done before quotas */ 2619 iwl_mvm_power_update_mac(mvm); 2620 2621 ret = iwl_mvm_update_quotas(mvm, false, NULL); 2622 if (ret) 2623 goto out_quota_failed; 2624 2625 /* Need to update the P2P Device MAC (only GO, IBSS is single vif) */ 2626 if (vif->p2p && mvm->p2p_device_vif) 2627 iwl_mvm_mac_ctxt_changed(mvm, mvm->p2p_device_vif, false, NULL); 2628 2629 iwl_mvm_ref(mvm, IWL_MVM_REF_AP_IBSS); 2630 2631 iwl_mvm_bt_coex_vif_change(mvm); 2632 2633 /* we don't support TDLS during DCM */ 2634 if (iwl_mvm_phy_ctx_count(mvm) > 1) 2635 iwl_mvm_teardown_tdls_peers(mvm); 2636 2637 iwl_mvm_ftm_restart_responder(mvm, vif); 2638 2639 goto out_unlock; 2640 2641 out_quota_failed: 2642 iwl_mvm_power_update_mac(mvm); 2643 mvmvif->ap_ibss_active = false; 2644 iwl_mvm_send_rm_bcast_sta(mvm, vif); 2645 iwl_mvm_rm_mcast_sta(mvm, vif); 2646 out_unbind: 2647 iwl_mvm_binding_remove_vif(mvm, vif); 2648 out_remove: 2649 iwl_mvm_mac_ctxt_remove(mvm, vif); 2650 out_unlock: 2651 mutex_unlock(&mvm->mutex); 2652 iwl_mvm_unref(mvm, IWL_MVM_REF_START_AP); 2653 return ret; 2654 } 2655 2656 static void iwl_mvm_stop_ap_ibss(struct ieee80211_hw *hw, 2657 struct ieee80211_vif *vif) 2658 { 2659 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2660 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 2661 2662 iwl_mvm_prepare_mac_removal(mvm, vif); 2663 2664 mutex_lock(&mvm->mutex); 2665 2666 /* Handle AP stop while in CSA */ 2667 if (rcu_access_pointer(mvm->csa_vif) == vif) { 2668 iwl_mvm_remove_time_event(mvm, mvmvif, 2669 &mvmvif->time_event_data); 2670 RCU_INIT_POINTER(mvm->csa_vif, NULL); 2671 mvmvif->csa_countdown = false; 2672 } 2673 2674 if (rcu_access_pointer(mvm->csa_tx_blocked_vif) == vif) { 2675 RCU_INIT_POINTER(mvm->csa_tx_blocked_vif, NULL); 2676 mvm->csa_tx_block_bcn_timeout = 0; 2677 } 2678 2679 mvmvif->ap_ibss_active = false; 2680 mvm->ap_last_beacon_gp2 = 0; 2681 2682 if (vif->type == NL80211_IFTYPE_AP && !vif->p2p) { 2683 iwl_mvm_vif_set_low_latency(mvmvif, false, 2684 LOW_LATENCY_VIF_TYPE); 2685 iwl_mvm_send_low_latency_cmd(mvm, false, mvmvif->id); 2686 } 2687 2688 iwl_mvm_bt_coex_vif_change(mvm); 2689 2690 iwl_mvm_unref(mvm, IWL_MVM_REF_AP_IBSS); 2691 2692 /* Need to update the P2P Device MAC (only GO, IBSS is single vif) */ 2693 if (vif->p2p && mvm->p2p_device_vif) 2694 iwl_mvm_mac_ctxt_changed(mvm, mvm->p2p_device_vif, false, NULL); 2695 2696 iwl_mvm_update_quotas(mvm, false, NULL); 2697 2698 /* 2699 * This is not very nice, but the simplest: 2700 * For older FWs removing the mcast sta before the bcast station may 2701 * cause assert 0x2b00. 2702 * This is fixed in later FW (which will stop beaconing when removing 2703 * bcast station). 2704 * So make the order of removal depend on the TLV 2705 */ 2706 if (!fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_STA_TYPE)) 2707 iwl_mvm_rm_mcast_sta(mvm, vif); 2708 iwl_mvm_send_rm_bcast_sta(mvm, vif); 2709 if (fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_STA_TYPE)) 2710 iwl_mvm_rm_mcast_sta(mvm, vif); 2711 iwl_mvm_binding_remove_vif(mvm, vif); 2712 2713 iwl_mvm_power_update_mac(mvm); 2714 2715 iwl_mvm_mac_ctxt_remove(mvm, vif); 2716 2717 mutex_unlock(&mvm->mutex); 2718 } 2719 2720 static void 2721 iwl_mvm_bss_info_changed_ap_ibss(struct iwl_mvm *mvm, 2722 struct ieee80211_vif *vif, 2723 struct ieee80211_bss_conf *bss_conf, 2724 u32 changes) 2725 { 2726 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 2727 2728 /* Changes will be applied when the AP/IBSS is started */ 2729 if (!mvmvif->ap_ibss_active) 2730 return; 2731 2732 if (changes & (BSS_CHANGED_ERP_CTS_PROT | BSS_CHANGED_HT | 2733 BSS_CHANGED_BANDWIDTH | BSS_CHANGED_QOS) && 2734 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL)) 2735 IWL_ERR(mvm, "failed to update MAC %pM\n", vif->addr); 2736 2737 /* Need to send a new beacon template to the FW */ 2738 if (changes & BSS_CHANGED_BEACON && 2739 iwl_mvm_mac_ctxt_beacon_changed(mvm, vif)) 2740 IWL_WARN(mvm, "Failed updating beacon data\n"); 2741 2742 if (changes & BSS_CHANGED_TXPOWER) { 2743 IWL_DEBUG_CALIB(mvm, "Changing TX Power to %d\n", 2744 bss_conf->txpower); 2745 iwl_mvm_set_tx_power(mvm, vif, bss_conf->txpower); 2746 } 2747 2748 if (changes & BSS_CHANGED_FTM_RESPONDER) { 2749 int ret = iwl_mvm_ftm_start_responder(mvm, vif); 2750 2751 if (ret) 2752 IWL_WARN(mvm, "Failed to enable FTM responder (%d)\n", 2753 ret); 2754 } 2755 2756 } 2757 2758 static void iwl_mvm_bss_info_changed(struct ieee80211_hw *hw, 2759 struct ieee80211_vif *vif, 2760 struct ieee80211_bss_conf *bss_conf, 2761 u32 changes) 2762 { 2763 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2764 2765 /* 2766 * iwl_mvm_bss_info_changed_station() might call 2767 * iwl_mvm_protect_session(), which reads directly from 2768 * the device (the system time), so make sure it is available. 2769 */ 2770 if (iwl_mvm_ref_sync(mvm, IWL_MVM_REF_BSS_CHANGED)) 2771 return; 2772 2773 mutex_lock(&mvm->mutex); 2774 2775 if (changes & BSS_CHANGED_IDLE && !bss_conf->idle) 2776 iwl_mvm_scan_stop(mvm, IWL_MVM_SCAN_SCHED, true); 2777 2778 switch (vif->type) { 2779 case NL80211_IFTYPE_STATION: 2780 iwl_mvm_bss_info_changed_station(mvm, vif, bss_conf, changes); 2781 break; 2782 case NL80211_IFTYPE_AP: 2783 case NL80211_IFTYPE_ADHOC: 2784 iwl_mvm_bss_info_changed_ap_ibss(mvm, vif, bss_conf, changes); 2785 break; 2786 case NL80211_IFTYPE_MONITOR: 2787 if (changes & BSS_CHANGED_MU_GROUPS) 2788 iwl_mvm_update_mu_groups(mvm, vif); 2789 break; 2790 default: 2791 /* shouldn't happen */ 2792 WARN_ON_ONCE(1); 2793 } 2794 2795 mutex_unlock(&mvm->mutex); 2796 iwl_mvm_unref(mvm, IWL_MVM_REF_BSS_CHANGED); 2797 } 2798 2799 static int iwl_mvm_mac_hw_scan(struct ieee80211_hw *hw, 2800 struct ieee80211_vif *vif, 2801 struct ieee80211_scan_request *hw_req) 2802 { 2803 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2804 int ret; 2805 2806 if (hw_req->req.n_channels == 0 || 2807 hw_req->req.n_channels > mvm->fw->ucode_capa.n_scan_channels) 2808 return -EINVAL; 2809 2810 mutex_lock(&mvm->mutex); 2811 ret = iwl_mvm_reg_scan_start(mvm, vif, &hw_req->req, &hw_req->ies); 2812 mutex_unlock(&mvm->mutex); 2813 2814 return ret; 2815 } 2816 2817 static void iwl_mvm_mac_cancel_hw_scan(struct ieee80211_hw *hw, 2818 struct ieee80211_vif *vif) 2819 { 2820 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2821 2822 mutex_lock(&mvm->mutex); 2823 2824 /* Due to a race condition, it's possible that mac80211 asks 2825 * us to stop a hw_scan when it's already stopped. This can 2826 * happen, for instance, if we stopped the scan ourselves, 2827 * called ieee80211_scan_completed() and the userspace called 2828 * cancel scan scan before ieee80211_scan_work() could run. 2829 * To handle that, simply return if the scan is not running. 2830 */ 2831 if (mvm->scan_status & IWL_MVM_SCAN_REGULAR) 2832 iwl_mvm_scan_stop(mvm, IWL_MVM_SCAN_REGULAR, true); 2833 2834 mutex_unlock(&mvm->mutex); 2835 } 2836 2837 static void 2838 iwl_mvm_mac_allow_buffered_frames(struct ieee80211_hw *hw, 2839 struct ieee80211_sta *sta, u16 tids, 2840 int num_frames, 2841 enum ieee80211_frame_release_type reason, 2842 bool more_data) 2843 { 2844 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2845 2846 /* Called when we need to transmit (a) frame(s) from mac80211 */ 2847 2848 iwl_mvm_sta_modify_sleep_tx_count(mvm, sta, reason, num_frames, 2849 tids, more_data, false); 2850 } 2851 2852 static void 2853 iwl_mvm_mac_release_buffered_frames(struct ieee80211_hw *hw, 2854 struct ieee80211_sta *sta, u16 tids, 2855 int num_frames, 2856 enum ieee80211_frame_release_type reason, 2857 bool more_data) 2858 { 2859 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2860 2861 /* Called when we need to transmit (a) frame(s) from agg or dqa queue */ 2862 2863 iwl_mvm_sta_modify_sleep_tx_count(mvm, sta, reason, num_frames, 2864 tids, more_data, true); 2865 } 2866 2867 static void __iwl_mvm_mac_sta_notify(struct ieee80211_hw *hw, 2868 enum sta_notify_cmd cmd, 2869 struct ieee80211_sta *sta) 2870 { 2871 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2872 struct iwl_mvm_sta *mvmsta = iwl_mvm_sta_from_mac80211(sta); 2873 unsigned long txqs = 0, tids = 0; 2874 int tid; 2875 2876 /* 2877 * If we have TVQM then we get too high queue numbers - luckily 2878 * we really shouldn't get here with that because such hardware 2879 * should have firmware supporting buffer station offload. 2880 */ 2881 if (WARN_ON(iwl_mvm_has_new_tx_api(mvm))) 2882 return; 2883 2884 spin_lock_bh(&mvmsta->lock); 2885 for (tid = 0; tid < ARRAY_SIZE(mvmsta->tid_data); tid++) { 2886 struct iwl_mvm_tid_data *tid_data = &mvmsta->tid_data[tid]; 2887 2888 if (tid_data->txq_id == IWL_MVM_INVALID_QUEUE) 2889 continue; 2890 2891 __set_bit(tid_data->txq_id, &txqs); 2892 2893 if (iwl_mvm_tid_queued(mvm, tid_data) == 0) 2894 continue; 2895 2896 __set_bit(tid, &tids); 2897 } 2898 2899 switch (cmd) { 2900 case STA_NOTIFY_SLEEP: 2901 for_each_set_bit(tid, &tids, IWL_MAX_TID_COUNT) 2902 ieee80211_sta_set_buffered(sta, tid, true); 2903 2904 if (txqs) 2905 iwl_trans_freeze_txq_timer(mvm->trans, txqs, true); 2906 /* 2907 * The fw updates the STA to be asleep. Tx packets on the Tx 2908 * queues to this station will not be transmitted. The fw will 2909 * send a Tx response with TX_STATUS_FAIL_DEST_PS. 2910 */ 2911 break; 2912 case STA_NOTIFY_AWAKE: 2913 if (WARN_ON(mvmsta->sta_id == IWL_MVM_INVALID_STA)) 2914 break; 2915 2916 if (txqs) 2917 iwl_trans_freeze_txq_timer(mvm->trans, txqs, false); 2918 iwl_mvm_sta_modify_ps_wake(mvm, sta); 2919 break; 2920 default: 2921 break; 2922 } 2923 spin_unlock_bh(&mvmsta->lock); 2924 } 2925 2926 static void iwl_mvm_mac_sta_notify(struct ieee80211_hw *hw, 2927 struct ieee80211_vif *vif, 2928 enum sta_notify_cmd cmd, 2929 struct ieee80211_sta *sta) 2930 { 2931 __iwl_mvm_mac_sta_notify(hw, cmd, sta); 2932 } 2933 2934 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb) 2935 { 2936 struct iwl_rx_packet *pkt = rxb_addr(rxb); 2937 struct iwl_mvm_pm_state_notification *notif = (void *)pkt->data; 2938 struct ieee80211_sta *sta; 2939 struct iwl_mvm_sta *mvmsta; 2940 bool sleeping = (notif->type != IWL_MVM_PM_EVENT_AWAKE); 2941 2942 if (WARN_ON(notif->sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id))) 2943 return; 2944 2945 rcu_read_lock(); 2946 sta = rcu_dereference(mvm->fw_id_to_mac_id[notif->sta_id]); 2947 if (WARN_ON(IS_ERR_OR_NULL(sta))) { 2948 rcu_read_unlock(); 2949 return; 2950 } 2951 2952 mvmsta = iwl_mvm_sta_from_mac80211(sta); 2953 2954 if (!mvmsta->vif || 2955 mvmsta->vif->type != NL80211_IFTYPE_AP) { 2956 rcu_read_unlock(); 2957 return; 2958 } 2959 2960 if (mvmsta->sleeping != sleeping) { 2961 mvmsta->sleeping = sleeping; 2962 __iwl_mvm_mac_sta_notify(mvm->hw, 2963 sleeping ? STA_NOTIFY_SLEEP : STA_NOTIFY_AWAKE, 2964 sta); 2965 ieee80211_sta_ps_transition(sta, sleeping); 2966 } 2967 2968 if (sleeping) { 2969 switch (notif->type) { 2970 case IWL_MVM_PM_EVENT_AWAKE: 2971 case IWL_MVM_PM_EVENT_ASLEEP: 2972 break; 2973 case IWL_MVM_PM_EVENT_UAPSD: 2974 ieee80211_sta_uapsd_trigger(sta, IEEE80211_NUM_TIDS); 2975 break; 2976 case IWL_MVM_PM_EVENT_PS_POLL: 2977 ieee80211_sta_pspoll(sta); 2978 break; 2979 default: 2980 break; 2981 } 2982 } 2983 2984 rcu_read_unlock(); 2985 } 2986 2987 static void iwl_mvm_sta_pre_rcu_remove(struct ieee80211_hw *hw, 2988 struct ieee80211_vif *vif, 2989 struct ieee80211_sta *sta) 2990 { 2991 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 2992 struct iwl_mvm_sta *mvm_sta = iwl_mvm_sta_from_mac80211(sta); 2993 2994 /* 2995 * This is called before mac80211 does RCU synchronisation, 2996 * so here we already invalidate our internal RCU-protected 2997 * station pointer. The rest of the code will thus no longer 2998 * be able to find the station this way, and we don't rely 2999 * on further RCU synchronisation after the sta_state() 3000 * callback deleted the station. 3001 */ 3002 mutex_lock(&mvm->mutex); 3003 if (sta == rcu_access_pointer(mvm->fw_id_to_mac_id[mvm_sta->sta_id])) 3004 rcu_assign_pointer(mvm->fw_id_to_mac_id[mvm_sta->sta_id], 3005 ERR_PTR(-ENOENT)); 3006 3007 mutex_unlock(&mvm->mutex); 3008 } 3009 3010 static void iwl_mvm_check_uapsd(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 3011 const u8 *bssid) 3012 { 3013 int i; 3014 3015 if (!test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status)) { 3016 struct iwl_mvm_tcm_mac *mdata; 3017 3018 mdata = &mvm->tcm.data[iwl_mvm_vif_from_mac80211(vif)->id]; 3019 ewma_rate_init(&mdata->uapsd_nonagg_detect.rate); 3020 mdata->opened_rx_ba_sessions = false; 3021 } 3022 3023 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_UAPSD_SUPPORT)) 3024 return; 3025 3026 if (vif->p2p && !iwl_mvm_is_p2p_scm_uapsd_supported(mvm)) { 3027 vif->driver_flags &= ~IEEE80211_VIF_SUPPORTS_UAPSD; 3028 return; 3029 } 3030 3031 if (!vif->p2p && 3032 (iwlwifi_mod_params.uapsd_disable & IWL_DISABLE_UAPSD_BSS)) { 3033 vif->driver_flags &= ~IEEE80211_VIF_SUPPORTS_UAPSD; 3034 return; 3035 } 3036 3037 for (i = 0; i < IWL_MVM_UAPSD_NOAGG_LIST_LEN; i++) { 3038 if (ether_addr_equal(mvm->uapsd_noagg_bssids[i].addr, bssid)) { 3039 vif->driver_flags &= ~IEEE80211_VIF_SUPPORTS_UAPSD; 3040 return; 3041 } 3042 } 3043 3044 vif->driver_flags |= IEEE80211_VIF_SUPPORTS_UAPSD; 3045 } 3046 3047 static void 3048 iwl_mvm_tdls_check_trigger(struct iwl_mvm *mvm, 3049 struct ieee80211_vif *vif, u8 *peer_addr, 3050 enum nl80211_tdls_operation action) 3051 { 3052 struct iwl_fw_dbg_trigger_tlv *trig; 3053 struct iwl_fw_dbg_trigger_tdls *tdls_trig; 3054 3055 trig = iwl_fw_dbg_trigger_on(&mvm->fwrt, ieee80211_vif_to_wdev(vif), 3056 FW_DBG_TRIGGER_TDLS); 3057 if (!trig) 3058 return; 3059 3060 tdls_trig = (void *)trig->data; 3061 3062 if (!(tdls_trig->action_bitmap & BIT(action))) 3063 return; 3064 3065 if (tdls_trig->peer_mode && 3066 memcmp(tdls_trig->peer, peer_addr, ETH_ALEN) != 0) 3067 return; 3068 3069 iwl_fw_dbg_collect_trig(&mvm->fwrt, trig, 3070 "TDLS event occurred, peer %pM, action %d", 3071 peer_addr, action); 3072 } 3073 3074 static int iwl_mvm_mac_sta_state(struct ieee80211_hw *hw, 3075 struct ieee80211_vif *vif, 3076 struct ieee80211_sta *sta, 3077 enum ieee80211_sta_state old_state, 3078 enum ieee80211_sta_state new_state) 3079 { 3080 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3081 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 3082 struct iwl_mvm_sta *mvm_sta = iwl_mvm_sta_from_mac80211(sta); 3083 int ret; 3084 3085 IWL_DEBUG_MAC80211(mvm, "station %pM state change %d->%d\n", 3086 sta->addr, old_state, new_state); 3087 3088 /* this would be a mac80211 bug ... but don't crash */ 3089 if (WARN_ON_ONCE(!mvmvif->phy_ctxt)) 3090 return -EINVAL; 3091 3092 /* 3093 * If we are in a STA removal flow and in DQA mode: 3094 * 3095 * This is after the sync_rcu part, so the queues have already been 3096 * flushed. No more TXs on their way in mac80211's path, and no more in 3097 * the queues. 3098 * Also, we won't be getting any new TX frames for this station. 3099 * What we might have are deferred TX frames that need to be taken care 3100 * of. 3101 * 3102 * Drop any still-queued deferred-frame before removing the STA, and 3103 * make sure the worker is no longer handling frames for this STA. 3104 */ 3105 if (old_state == IEEE80211_STA_NONE && 3106 new_state == IEEE80211_STA_NOTEXIST) { 3107 flush_work(&mvm->add_stream_wk); 3108 3109 /* 3110 * No need to make sure deferred TX indication is off since the 3111 * worker will already remove it if it was on 3112 */ 3113 } 3114 3115 mutex_lock(&mvm->mutex); 3116 /* track whether or not the station is associated */ 3117 mvm_sta->sta_state = new_state; 3118 3119 if (old_state == IEEE80211_STA_NOTEXIST && 3120 new_state == IEEE80211_STA_NONE) { 3121 /* 3122 * Firmware bug - it'll crash if the beacon interval is less 3123 * than 16. We can't avoid connecting at all, so refuse the 3124 * station state change, this will cause mac80211 to abandon 3125 * attempts to connect to this AP, and eventually wpa_s will 3126 * blacklist the AP... 3127 */ 3128 if (vif->type == NL80211_IFTYPE_STATION && 3129 vif->bss_conf.beacon_int < 16) { 3130 IWL_ERR(mvm, 3131 "AP %pM beacon interval is %d, refusing due to firmware bug!\n", 3132 sta->addr, vif->bss_conf.beacon_int); 3133 ret = -EINVAL; 3134 goto out_unlock; 3135 } 3136 3137 if (sta->tdls && 3138 (vif->p2p || 3139 iwl_mvm_tdls_sta_count(mvm, NULL) == 3140 IWL_MVM_TDLS_STA_COUNT || 3141 iwl_mvm_phy_ctx_count(mvm) > 1)) { 3142 IWL_DEBUG_MAC80211(mvm, "refusing TDLS sta\n"); 3143 ret = -EBUSY; 3144 goto out_unlock; 3145 } 3146 3147 ret = iwl_mvm_add_sta(mvm, vif, sta); 3148 if (sta->tdls && ret == 0) { 3149 iwl_mvm_recalc_tdls_state(mvm, vif, true); 3150 iwl_mvm_tdls_check_trigger(mvm, vif, sta->addr, 3151 NL80211_TDLS_SETUP); 3152 } 3153 3154 sta->max_rc_amsdu_len = 1; 3155 } else if (old_state == IEEE80211_STA_NONE && 3156 new_state == IEEE80211_STA_AUTH) { 3157 /* 3158 * EBS may be disabled due to previous failures reported by FW. 3159 * Reset EBS status here assuming environment has been changed. 3160 */ 3161 mvm->last_ebs_successful = true; 3162 iwl_mvm_check_uapsd(mvm, vif, sta->addr); 3163 ret = 0; 3164 } else if (old_state == IEEE80211_STA_AUTH && 3165 new_state == IEEE80211_STA_ASSOC) { 3166 if (vif->type == NL80211_IFTYPE_AP) { 3167 vif->bss_conf.he_support = sta->he_cap.has_he; 3168 mvmvif->ap_assoc_sta_count++; 3169 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 3170 if (vif->bss_conf.he_support && 3171 !iwlwifi_mod_params.disable_11ax) 3172 iwl_mvm_cfg_he_sta(mvm, vif, mvm_sta->sta_id); 3173 } else if (vif->type == NL80211_IFTYPE_STATION) { 3174 vif->bss_conf.he_support = sta->he_cap.has_he; 3175 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 3176 } 3177 3178 iwl_mvm_rs_rate_init(mvm, sta, mvmvif->phy_ctxt->channel->band, 3179 false); 3180 ret = iwl_mvm_update_sta(mvm, vif, sta); 3181 } else if (old_state == IEEE80211_STA_ASSOC && 3182 new_state == IEEE80211_STA_AUTHORIZED) { 3183 ret = 0; 3184 3185 /* we don't support TDLS during DCM */ 3186 if (iwl_mvm_phy_ctx_count(mvm) > 1) 3187 iwl_mvm_teardown_tdls_peers(mvm); 3188 3189 if (sta->tdls) 3190 iwl_mvm_tdls_check_trigger(mvm, vif, sta->addr, 3191 NL80211_TDLS_ENABLE_LINK); 3192 3193 /* enable beacon filtering */ 3194 WARN_ON(iwl_mvm_enable_beacon_filter(mvm, vif, 0)); 3195 3196 iwl_mvm_rs_rate_init(mvm, sta, mvmvif->phy_ctxt->channel->band, 3197 true); 3198 } else if (old_state == IEEE80211_STA_AUTHORIZED && 3199 new_state == IEEE80211_STA_ASSOC) { 3200 /* disable beacon filtering */ 3201 ret = iwl_mvm_disable_beacon_filter(mvm, vif, 0); 3202 WARN_ON(ret && 3203 !test_bit(IWL_MVM_STATUS_HW_RESTART_REQUESTED, 3204 &mvm->status)); 3205 ret = 0; 3206 } else if (old_state == IEEE80211_STA_ASSOC && 3207 new_state == IEEE80211_STA_AUTH) { 3208 if (vif->type == NL80211_IFTYPE_AP) { 3209 mvmvif->ap_assoc_sta_count--; 3210 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 3211 } 3212 ret = 0; 3213 } else if (old_state == IEEE80211_STA_AUTH && 3214 new_state == IEEE80211_STA_NONE) { 3215 ret = 0; 3216 } else if (old_state == IEEE80211_STA_NONE && 3217 new_state == IEEE80211_STA_NOTEXIST) { 3218 ret = iwl_mvm_rm_sta(mvm, vif, sta); 3219 if (sta->tdls) { 3220 iwl_mvm_recalc_tdls_state(mvm, vif, false); 3221 iwl_mvm_tdls_check_trigger(mvm, vif, sta->addr, 3222 NL80211_TDLS_DISABLE_LINK); 3223 } 3224 3225 if (unlikely(ret && 3226 test_bit(IWL_MVM_STATUS_HW_RESTART_REQUESTED, 3227 &mvm->status))) 3228 ret = 0; 3229 } else { 3230 ret = -EIO; 3231 } 3232 out_unlock: 3233 mutex_unlock(&mvm->mutex); 3234 3235 if (sta->tdls && ret == 0) { 3236 if (old_state == IEEE80211_STA_NOTEXIST && 3237 new_state == IEEE80211_STA_NONE) 3238 ieee80211_reserve_tid(sta, IWL_MVM_TDLS_FW_TID); 3239 else if (old_state == IEEE80211_STA_NONE && 3240 new_state == IEEE80211_STA_NOTEXIST) 3241 ieee80211_unreserve_tid(sta, IWL_MVM_TDLS_FW_TID); 3242 } 3243 3244 return ret; 3245 } 3246 3247 static int iwl_mvm_mac_set_rts_threshold(struct ieee80211_hw *hw, u32 value) 3248 { 3249 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3250 3251 mvm->rts_threshold = value; 3252 3253 return 0; 3254 } 3255 3256 static void iwl_mvm_sta_rc_update(struct ieee80211_hw *hw, 3257 struct ieee80211_vif *vif, 3258 struct ieee80211_sta *sta, u32 changed) 3259 { 3260 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3261 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 3262 3263 if (changed & (IEEE80211_RC_BW_CHANGED | 3264 IEEE80211_RC_SUPP_RATES_CHANGED | 3265 IEEE80211_RC_NSS_CHANGED)) 3266 iwl_mvm_rs_rate_init(mvm, sta, mvmvif->phy_ctxt->channel->band, 3267 true); 3268 3269 if (vif->type == NL80211_IFTYPE_STATION && 3270 changed & IEEE80211_RC_NSS_CHANGED) 3271 iwl_mvm_sf_update(mvm, vif, false); 3272 } 3273 3274 static int iwl_mvm_mac_conf_tx(struct ieee80211_hw *hw, 3275 struct ieee80211_vif *vif, u16 ac, 3276 const struct ieee80211_tx_queue_params *params) 3277 { 3278 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3279 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 3280 3281 mvmvif->queue_params[ac] = *params; 3282 3283 /* 3284 * No need to update right away, we'll get BSS_CHANGED_QOS 3285 * The exception is P2P_DEVICE interface which needs immediate update. 3286 */ 3287 if (vif->type == NL80211_IFTYPE_P2P_DEVICE) { 3288 int ret; 3289 3290 mutex_lock(&mvm->mutex); 3291 ret = iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 3292 mutex_unlock(&mvm->mutex); 3293 return ret; 3294 } 3295 return 0; 3296 } 3297 3298 static void iwl_mvm_mac_mgd_prepare_tx(struct ieee80211_hw *hw, 3299 struct ieee80211_vif *vif, 3300 u16 req_duration) 3301 { 3302 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3303 u32 duration = IWL_MVM_TE_SESSION_PROTECTION_MAX_TIME_MS; 3304 u32 min_duration = IWL_MVM_TE_SESSION_PROTECTION_MIN_TIME_MS; 3305 3306 /* 3307 * iwl_mvm_protect_session() reads directly from the device 3308 * (the system time), so make sure it is available. 3309 */ 3310 if (iwl_mvm_ref_sync(mvm, IWL_MVM_REF_PREPARE_TX)) 3311 return; 3312 3313 if (req_duration > duration) 3314 duration = req_duration; 3315 3316 mutex_lock(&mvm->mutex); 3317 /* Try really hard to protect the session and hear a beacon */ 3318 iwl_mvm_protect_session(mvm, vif, duration, min_duration, 500, false); 3319 mutex_unlock(&mvm->mutex); 3320 3321 iwl_mvm_unref(mvm, IWL_MVM_REF_PREPARE_TX); 3322 } 3323 3324 static int iwl_mvm_mac_sched_scan_start(struct ieee80211_hw *hw, 3325 struct ieee80211_vif *vif, 3326 struct cfg80211_sched_scan_request *req, 3327 struct ieee80211_scan_ies *ies) 3328 { 3329 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3330 3331 int ret; 3332 3333 mutex_lock(&mvm->mutex); 3334 3335 if (!vif->bss_conf.idle) { 3336 ret = -EBUSY; 3337 goto out; 3338 } 3339 3340 ret = iwl_mvm_sched_scan_start(mvm, vif, req, ies, IWL_MVM_SCAN_SCHED); 3341 3342 out: 3343 mutex_unlock(&mvm->mutex); 3344 return ret; 3345 } 3346 3347 static int iwl_mvm_mac_sched_scan_stop(struct ieee80211_hw *hw, 3348 struct ieee80211_vif *vif) 3349 { 3350 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3351 int ret; 3352 3353 mutex_lock(&mvm->mutex); 3354 3355 /* Due to a race condition, it's possible that mac80211 asks 3356 * us to stop a sched_scan when it's already stopped. This 3357 * can happen, for instance, if we stopped the scan ourselves, 3358 * called ieee80211_sched_scan_stopped() and the userspace called 3359 * stop sched scan scan before ieee80211_sched_scan_stopped_work() 3360 * could run. To handle this, simply return if the scan is 3361 * not running. 3362 */ 3363 if (!(mvm->scan_status & IWL_MVM_SCAN_SCHED)) { 3364 mutex_unlock(&mvm->mutex); 3365 return 0; 3366 } 3367 3368 ret = iwl_mvm_scan_stop(mvm, IWL_MVM_SCAN_SCHED, false); 3369 mutex_unlock(&mvm->mutex); 3370 iwl_mvm_wait_for_async_handlers(mvm); 3371 3372 return ret; 3373 } 3374 3375 static int iwl_mvm_mac_set_key(struct ieee80211_hw *hw, 3376 enum set_key_cmd cmd, 3377 struct ieee80211_vif *vif, 3378 struct ieee80211_sta *sta, 3379 struct ieee80211_key_conf *key) 3380 { 3381 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3382 struct iwl_mvm_sta *mvmsta; 3383 struct iwl_mvm_key_pn *ptk_pn; 3384 int keyidx = key->keyidx; 3385 int ret; 3386 u8 key_offset; 3387 3388 if (iwlwifi_mod_params.swcrypto) { 3389 IWL_DEBUG_MAC80211(mvm, "leave - hwcrypto disabled\n"); 3390 return -EOPNOTSUPP; 3391 } 3392 3393 switch (key->cipher) { 3394 case WLAN_CIPHER_SUITE_TKIP: 3395 if (!mvm->trans->cfg->gen2) { 3396 key->flags |= IEEE80211_KEY_FLAG_GENERATE_MMIC; 3397 key->flags |= IEEE80211_KEY_FLAG_PUT_IV_SPACE; 3398 } else if (vif->type == NL80211_IFTYPE_STATION) { 3399 key->flags |= IEEE80211_KEY_FLAG_PUT_MIC_SPACE; 3400 } else { 3401 IWL_DEBUG_MAC80211(mvm, "Use SW encryption for TKIP\n"); 3402 return -EOPNOTSUPP; 3403 } 3404 break; 3405 case WLAN_CIPHER_SUITE_CCMP: 3406 case WLAN_CIPHER_SUITE_GCMP: 3407 case WLAN_CIPHER_SUITE_GCMP_256: 3408 if (!iwl_mvm_has_new_tx_api(mvm)) 3409 key->flags |= IEEE80211_KEY_FLAG_PUT_IV_SPACE; 3410 break; 3411 case WLAN_CIPHER_SUITE_AES_CMAC: 3412 case WLAN_CIPHER_SUITE_BIP_GMAC_128: 3413 case WLAN_CIPHER_SUITE_BIP_GMAC_256: 3414 WARN_ON_ONCE(!ieee80211_hw_check(hw, MFP_CAPABLE)); 3415 break; 3416 case WLAN_CIPHER_SUITE_WEP40: 3417 case WLAN_CIPHER_SUITE_WEP104: 3418 if (vif->type == NL80211_IFTYPE_STATION) 3419 break; 3420 if (iwl_mvm_has_new_tx_api(mvm)) 3421 return -EOPNOTSUPP; 3422 /* support HW crypto on TX */ 3423 return 0; 3424 default: 3425 /* currently FW supports only one optional cipher scheme */ 3426 if (hw->n_cipher_schemes && 3427 hw->cipher_schemes->cipher == key->cipher) 3428 key->flags |= IEEE80211_KEY_FLAG_PUT_IV_SPACE; 3429 else 3430 return -EOPNOTSUPP; 3431 } 3432 3433 mutex_lock(&mvm->mutex); 3434 3435 switch (cmd) { 3436 case SET_KEY: 3437 if ((vif->type == NL80211_IFTYPE_ADHOC || 3438 vif->type == NL80211_IFTYPE_AP) && !sta) { 3439 /* 3440 * GTK on AP interface is a TX-only key, return 0; 3441 * on IBSS they're per-station and because we're lazy 3442 * we don't support them for RX, so do the same. 3443 * CMAC/GMAC in AP/IBSS modes must be done in software. 3444 */ 3445 if (key->cipher == WLAN_CIPHER_SUITE_AES_CMAC || 3446 key->cipher == WLAN_CIPHER_SUITE_BIP_GMAC_128 || 3447 key->cipher == WLAN_CIPHER_SUITE_BIP_GMAC_256) 3448 ret = -EOPNOTSUPP; 3449 else 3450 ret = 0; 3451 3452 if (key->cipher != WLAN_CIPHER_SUITE_GCMP && 3453 key->cipher != WLAN_CIPHER_SUITE_GCMP_256 && 3454 !iwl_mvm_has_new_tx_api(mvm)) { 3455 key->hw_key_idx = STA_KEY_IDX_INVALID; 3456 break; 3457 } 3458 } 3459 3460 /* During FW restart, in order to restore the state as it was, 3461 * don't try to reprogram keys we previously failed for. 3462 */ 3463 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status) && 3464 key->hw_key_idx == STA_KEY_IDX_INVALID) { 3465 IWL_DEBUG_MAC80211(mvm, 3466 "skip invalid idx key programming during restart\n"); 3467 ret = 0; 3468 break; 3469 } 3470 3471 if (!test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status) && 3472 sta && iwl_mvm_has_new_rx_api(mvm) && 3473 key->flags & IEEE80211_KEY_FLAG_PAIRWISE && 3474 (key->cipher == WLAN_CIPHER_SUITE_CCMP || 3475 key->cipher == WLAN_CIPHER_SUITE_GCMP || 3476 key->cipher == WLAN_CIPHER_SUITE_GCMP_256)) { 3477 struct ieee80211_key_seq seq; 3478 int tid, q; 3479 3480 mvmsta = iwl_mvm_sta_from_mac80211(sta); 3481 WARN_ON(rcu_access_pointer(mvmsta->ptk_pn[keyidx])); 3482 ptk_pn = kzalloc(struct_size(ptk_pn, q, 3483 mvm->trans->num_rx_queues), 3484 GFP_KERNEL); 3485 if (!ptk_pn) { 3486 ret = -ENOMEM; 3487 break; 3488 } 3489 3490 for (tid = 0; tid < IWL_MAX_TID_COUNT; tid++) { 3491 ieee80211_get_key_rx_seq(key, tid, &seq); 3492 for (q = 0; q < mvm->trans->num_rx_queues; q++) 3493 memcpy(ptk_pn->q[q].pn[tid], 3494 seq.ccmp.pn, 3495 IEEE80211_CCMP_PN_LEN); 3496 } 3497 3498 rcu_assign_pointer(mvmsta->ptk_pn[keyidx], ptk_pn); 3499 } 3500 3501 /* in HW restart reuse the index, otherwise request a new one */ 3502 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status)) 3503 key_offset = key->hw_key_idx; 3504 else 3505 key_offset = STA_KEY_IDX_INVALID; 3506 3507 IWL_DEBUG_MAC80211(mvm, "set hwcrypto key\n"); 3508 ret = iwl_mvm_set_sta_key(mvm, vif, sta, key, key_offset); 3509 if (ret) { 3510 IWL_WARN(mvm, "set key failed\n"); 3511 key->hw_key_idx = STA_KEY_IDX_INVALID; 3512 /* 3513 * can't add key for RX, but we don't need it 3514 * in the device for TX so still return 0, 3515 * unless we have new TX API where we cannot 3516 * put key material into the TX_CMD 3517 */ 3518 if (iwl_mvm_has_new_tx_api(mvm)) 3519 ret = -EOPNOTSUPP; 3520 else 3521 ret = 0; 3522 } 3523 3524 break; 3525 case DISABLE_KEY: 3526 if (key->hw_key_idx == STA_KEY_IDX_INVALID) { 3527 ret = 0; 3528 break; 3529 } 3530 3531 if (sta && iwl_mvm_has_new_rx_api(mvm) && 3532 key->flags & IEEE80211_KEY_FLAG_PAIRWISE && 3533 (key->cipher == WLAN_CIPHER_SUITE_CCMP || 3534 key->cipher == WLAN_CIPHER_SUITE_GCMP || 3535 key->cipher == WLAN_CIPHER_SUITE_GCMP_256)) { 3536 mvmsta = iwl_mvm_sta_from_mac80211(sta); 3537 ptk_pn = rcu_dereference_protected( 3538 mvmsta->ptk_pn[keyidx], 3539 lockdep_is_held(&mvm->mutex)); 3540 RCU_INIT_POINTER(mvmsta->ptk_pn[keyidx], NULL); 3541 if (ptk_pn) 3542 kfree_rcu(ptk_pn, rcu_head); 3543 } 3544 3545 IWL_DEBUG_MAC80211(mvm, "disable hwcrypto key\n"); 3546 ret = iwl_mvm_remove_sta_key(mvm, vif, sta, key); 3547 break; 3548 default: 3549 ret = -EINVAL; 3550 } 3551 3552 mutex_unlock(&mvm->mutex); 3553 return ret; 3554 } 3555 3556 static void iwl_mvm_mac_update_tkip_key(struct ieee80211_hw *hw, 3557 struct ieee80211_vif *vif, 3558 struct ieee80211_key_conf *keyconf, 3559 struct ieee80211_sta *sta, 3560 u32 iv32, u16 *phase1key) 3561 { 3562 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3563 3564 if (keyconf->hw_key_idx == STA_KEY_IDX_INVALID) 3565 return; 3566 3567 iwl_mvm_update_tkip_key(mvm, vif, keyconf, sta, iv32, phase1key); 3568 } 3569 3570 3571 static bool iwl_mvm_rx_aux_roc(struct iwl_notif_wait_data *notif_wait, 3572 struct iwl_rx_packet *pkt, void *data) 3573 { 3574 struct iwl_mvm *mvm = 3575 container_of(notif_wait, struct iwl_mvm, notif_wait); 3576 struct iwl_hs20_roc_res *resp; 3577 int resp_len = iwl_rx_packet_payload_len(pkt); 3578 struct iwl_mvm_time_event_data *te_data = data; 3579 3580 if (WARN_ON(pkt->hdr.cmd != HOT_SPOT_CMD)) 3581 return true; 3582 3583 if (WARN_ON_ONCE(resp_len != sizeof(*resp))) { 3584 IWL_ERR(mvm, "Invalid HOT_SPOT_CMD response\n"); 3585 return true; 3586 } 3587 3588 resp = (void *)pkt->data; 3589 3590 IWL_DEBUG_TE(mvm, 3591 "Aux ROC: Received response from ucode: status=%d uid=%d\n", 3592 resp->status, resp->event_unique_id); 3593 3594 te_data->uid = le32_to_cpu(resp->event_unique_id); 3595 IWL_DEBUG_TE(mvm, "TIME_EVENT_CMD response - UID = 0x%x\n", 3596 te_data->uid); 3597 3598 spin_lock_bh(&mvm->time_event_lock); 3599 list_add_tail(&te_data->list, &mvm->aux_roc_te_list); 3600 spin_unlock_bh(&mvm->time_event_lock); 3601 3602 return true; 3603 } 3604 3605 #define AUX_ROC_MIN_DURATION MSEC_TO_TU(100) 3606 #define AUX_ROC_MIN_DELAY MSEC_TO_TU(200) 3607 #define AUX_ROC_MAX_DELAY MSEC_TO_TU(600) 3608 #define AUX_ROC_SAFETY_BUFFER MSEC_TO_TU(20) 3609 #define AUX_ROC_MIN_SAFETY_BUFFER MSEC_TO_TU(10) 3610 static int iwl_mvm_send_aux_roc_cmd(struct iwl_mvm *mvm, 3611 struct ieee80211_channel *channel, 3612 struct ieee80211_vif *vif, 3613 int duration) 3614 { 3615 int res, time_reg = DEVICE_SYSTEM_TIME_REG; 3616 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 3617 struct iwl_mvm_time_event_data *te_data = &mvmvif->hs_time_event_data; 3618 static const u16 time_event_response[] = { HOT_SPOT_CMD }; 3619 struct iwl_notification_wait wait_time_event; 3620 u32 dtim_interval = vif->bss_conf.dtim_period * 3621 vif->bss_conf.beacon_int; 3622 u32 req_dur, delay; 3623 struct iwl_hs20_roc_req aux_roc_req = { 3624 .action = cpu_to_le32(FW_CTXT_ACTION_ADD), 3625 .id_and_color = 3626 cpu_to_le32(FW_CMD_ID_AND_COLOR(MAC_INDEX_AUX, 0)), 3627 .sta_id_and_color = cpu_to_le32(mvm->aux_sta.sta_id), 3628 }; 3629 struct iwl_hs20_roc_req_tail *tail = iwl_mvm_chan_info_cmd_tail(mvm, 3630 &aux_roc_req.channel_info); 3631 u16 len = sizeof(aux_roc_req) - iwl_mvm_chan_info_padding(mvm); 3632 3633 /* Set the channel info data */ 3634 iwl_mvm_set_chan_info(mvm, &aux_roc_req.channel_info, channel->hw_value, 3635 (channel->band == NL80211_BAND_2GHZ) ? 3636 PHY_BAND_24 : PHY_BAND_5, 3637 PHY_VHT_CHANNEL_MODE20, 3638 0); 3639 3640 /* Set the time and duration */ 3641 tail->apply_time = cpu_to_le32(iwl_read_prph(mvm->trans, time_reg)); 3642 3643 delay = AUX_ROC_MIN_DELAY; 3644 req_dur = MSEC_TO_TU(duration); 3645 3646 /* 3647 * If we are associated we want the delay time to be at least one 3648 * dtim interval so that the FW can wait until after the DTIM and 3649 * then start the time event, this will potentially allow us to 3650 * remain off-channel for the max duration. 3651 * Since we want to use almost a whole dtim interval we would also 3652 * like the delay to be for 2-3 dtim intervals, in case there are 3653 * other time events with higher priority. 3654 */ 3655 if (vif->bss_conf.assoc) { 3656 delay = min_t(u32, dtim_interval * 3, AUX_ROC_MAX_DELAY); 3657 /* We cannot remain off-channel longer than the DTIM interval */ 3658 if (dtim_interval <= req_dur) { 3659 req_dur = dtim_interval - AUX_ROC_SAFETY_BUFFER; 3660 if (req_dur <= AUX_ROC_MIN_DURATION) 3661 req_dur = dtim_interval - 3662 AUX_ROC_MIN_SAFETY_BUFFER; 3663 } 3664 } 3665 3666 tail->duration = cpu_to_le32(req_dur); 3667 tail->apply_time_max_delay = cpu_to_le32(delay); 3668 3669 IWL_DEBUG_TE(mvm, 3670 "ROC: Requesting to remain on channel %u for %ums (requested = %ums, max_delay = %ums, dtim_interval = %ums)\n", 3671 channel->hw_value, req_dur, duration, delay, 3672 dtim_interval); 3673 /* Set the node address */ 3674 memcpy(tail->node_addr, vif->addr, ETH_ALEN); 3675 3676 lockdep_assert_held(&mvm->mutex); 3677 3678 spin_lock_bh(&mvm->time_event_lock); 3679 3680 if (WARN_ON(te_data->id == HOT_SPOT_CMD)) { 3681 spin_unlock_bh(&mvm->time_event_lock); 3682 return -EIO; 3683 } 3684 3685 te_data->vif = vif; 3686 te_data->duration = duration; 3687 te_data->id = HOT_SPOT_CMD; 3688 3689 spin_unlock_bh(&mvm->time_event_lock); 3690 3691 /* 3692 * Use a notification wait, which really just processes the 3693 * command response and doesn't wait for anything, in order 3694 * to be able to process the response and get the UID inside 3695 * the RX path. Using CMD_WANT_SKB doesn't work because it 3696 * stores the buffer and then wakes up this thread, by which 3697 * time another notification (that the time event started) 3698 * might already be processed unsuccessfully. 3699 */ 3700 iwl_init_notification_wait(&mvm->notif_wait, &wait_time_event, 3701 time_event_response, 3702 ARRAY_SIZE(time_event_response), 3703 iwl_mvm_rx_aux_roc, te_data); 3704 3705 res = iwl_mvm_send_cmd_pdu(mvm, HOT_SPOT_CMD, 0, len, 3706 &aux_roc_req); 3707 3708 if (res) { 3709 IWL_ERR(mvm, "Couldn't send HOT_SPOT_CMD: %d\n", res); 3710 iwl_remove_notification(&mvm->notif_wait, &wait_time_event); 3711 goto out_clear_te; 3712 } 3713 3714 /* No need to wait for anything, so just pass 1 (0 isn't valid) */ 3715 res = iwl_wait_notification(&mvm->notif_wait, &wait_time_event, 1); 3716 /* should never fail */ 3717 WARN_ON_ONCE(res); 3718 3719 if (res) { 3720 out_clear_te: 3721 spin_lock_bh(&mvm->time_event_lock); 3722 iwl_mvm_te_clear_data(mvm, te_data); 3723 spin_unlock_bh(&mvm->time_event_lock); 3724 } 3725 3726 return res; 3727 } 3728 3729 static int iwl_mvm_roc(struct ieee80211_hw *hw, 3730 struct ieee80211_vif *vif, 3731 struct ieee80211_channel *channel, 3732 int duration, 3733 enum ieee80211_roc_type type) 3734 { 3735 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3736 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 3737 struct cfg80211_chan_def chandef; 3738 struct iwl_mvm_phy_ctxt *phy_ctxt; 3739 int ret, i; 3740 3741 IWL_DEBUG_MAC80211(mvm, "enter (%d, %d, %d)\n", channel->hw_value, 3742 duration, type); 3743 3744 /* 3745 * Flush the done work, just in case it's still pending, so that 3746 * the work it does can complete and we can accept new frames. 3747 */ 3748 flush_work(&mvm->roc_done_wk); 3749 3750 mutex_lock(&mvm->mutex); 3751 3752 switch (vif->type) { 3753 case NL80211_IFTYPE_STATION: 3754 if (fw_has_capa(&mvm->fw->ucode_capa, 3755 IWL_UCODE_TLV_CAPA_HOTSPOT_SUPPORT)) { 3756 /* Use aux roc framework (HS20) */ 3757 ret = iwl_mvm_send_aux_roc_cmd(mvm, channel, 3758 vif, duration); 3759 goto out_unlock; 3760 } 3761 IWL_ERR(mvm, "hotspot not supported\n"); 3762 ret = -EINVAL; 3763 goto out_unlock; 3764 case NL80211_IFTYPE_P2P_DEVICE: 3765 /* handle below */ 3766 break; 3767 default: 3768 IWL_ERR(mvm, "vif isn't P2P_DEVICE: %d\n", vif->type); 3769 ret = -EINVAL; 3770 goto out_unlock; 3771 } 3772 3773 for (i = 0; i < NUM_PHY_CTX; i++) { 3774 phy_ctxt = &mvm->phy_ctxts[i]; 3775 if (phy_ctxt->ref == 0 || mvmvif->phy_ctxt == phy_ctxt) 3776 continue; 3777 3778 if (phy_ctxt->ref && channel == phy_ctxt->channel) { 3779 /* 3780 * Unbind the P2P_DEVICE from the current PHY context, 3781 * and if the PHY context is not used remove it. 3782 */ 3783 ret = iwl_mvm_binding_remove_vif(mvm, vif); 3784 if (WARN(ret, "Failed unbinding P2P_DEVICE\n")) 3785 goto out_unlock; 3786 3787 iwl_mvm_phy_ctxt_unref(mvm, mvmvif->phy_ctxt); 3788 3789 /* Bind the P2P_DEVICE to the current PHY Context */ 3790 mvmvif->phy_ctxt = phy_ctxt; 3791 3792 ret = iwl_mvm_binding_add_vif(mvm, vif); 3793 if (WARN(ret, "Failed binding P2P_DEVICE\n")) 3794 goto out_unlock; 3795 3796 iwl_mvm_phy_ctxt_ref(mvm, mvmvif->phy_ctxt); 3797 goto schedule_time_event; 3798 } 3799 } 3800 3801 /* Need to update the PHY context only if the ROC channel changed */ 3802 if (channel == mvmvif->phy_ctxt->channel) 3803 goto schedule_time_event; 3804 3805 cfg80211_chandef_create(&chandef, channel, NL80211_CHAN_NO_HT); 3806 3807 /* 3808 * Change the PHY context configuration as it is currently referenced 3809 * only by the P2P Device MAC 3810 */ 3811 if (mvmvif->phy_ctxt->ref == 1) { 3812 ret = iwl_mvm_phy_ctxt_changed(mvm, mvmvif->phy_ctxt, 3813 &chandef, 1, 1); 3814 if (ret) 3815 goto out_unlock; 3816 } else { 3817 /* 3818 * The PHY context is shared with other MACs. Need to remove the 3819 * P2P Device from the binding, allocate an new PHY context and 3820 * create a new binding 3821 */ 3822 phy_ctxt = iwl_mvm_get_free_phy_ctxt(mvm); 3823 if (!phy_ctxt) { 3824 ret = -ENOSPC; 3825 goto out_unlock; 3826 } 3827 3828 ret = iwl_mvm_phy_ctxt_changed(mvm, phy_ctxt, &chandef, 3829 1, 1); 3830 if (ret) { 3831 IWL_ERR(mvm, "Failed to change PHY context\n"); 3832 goto out_unlock; 3833 } 3834 3835 /* Unbind the P2P_DEVICE from the current PHY context */ 3836 ret = iwl_mvm_binding_remove_vif(mvm, vif); 3837 if (WARN(ret, "Failed unbinding P2P_DEVICE\n")) 3838 goto out_unlock; 3839 3840 iwl_mvm_phy_ctxt_unref(mvm, mvmvif->phy_ctxt); 3841 3842 /* Bind the P2P_DEVICE to the new allocated PHY context */ 3843 mvmvif->phy_ctxt = phy_ctxt; 3844 3845 ret = iwl_mvm_binding_add_vif(mvm, vif); 3846 if (WARN(ret, "Failed binding P2P_DEVICE\n")) 3847 goto out_unlock; 3848 3849 iwl_mvm_phy_ctxt_ref(mvm, mvmvif->phy_ctxt); 3850 } 3851 3852 schedule_time_event: 3853 /* Schedule the time events */ 3854 ret = iwl_mvm_start_p2p_roc(mvm, vif, duration, type); 3855 3856 out_unlock: 3857 mutex_unlock(&mvm->mutex); 3858 IWL_DEBUG_MAC80211(mvm, "leave\n"); 3859 return ret; 3860 } 3861 3862 static int iwl_mvm_cancel_roc(struct ieee80211_hw *hw) 3863 { 3864 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3865 3866 IWL_DEBUG_MAC80211(mvm, "enter\n"); 3867 3868 mutex_lock(&mvm->mutex); 3869 iwl_mvm_stop_roc(mvm); 3870 mutex_unlock(&mvm->mutex); 3871 3872 IWL_DEBUG_MAC80211(mvm, "leave\n"); 3873 return 0; 3874 } 3875 3876 struct iwl_mvm_ftm_responder_iter_data { 3877 bool responder; 3878 struct ieee80211_chanctx_conf *ctx; 3879 }; 3880 3881 static void iwl_mvm_ftm_responder_chanctx_iter(void *_data, u8 *mac, 3882 struct ieee80211_vif *vif) 3883 { 3884 struct iwl_mvm_ftm_responder_iter_data *data = _data; 3885 3886 if (rcu_access_pointer(vif->chanctx_conf) == data->ctx && 3887 vif->type == NL80211_IFTYPE_AP && vif->bss_conf.ftmr_params) 3888 data->responder = true; 3889 } 3890 3891 static bool iwl_mvm_is_ftm_responder_chanctx(struct iwl_mvm *mvm, 3892 struct ieee80211_chanctx_conf *ctx) 3893 { 3894 struct iwl_mvm_ftm_responder_iter_data data = { 3895 .responder = false, 3896 .ctx = ctx, 3897 }; 3898 3899 ieee80211_iterate_active_interfaces_atomic(mvm->hw, 3900 IEEE80211_IFACE_ITER_NORMAL, 3901 iwl_mvm_ftm_responder_chanctx_iter, 3902 &data); 3903 return data.responder; 3904 } 3905 3906 static int __iwl_mvm_add_chanctx(struct iwl_mvm *mvm, 3907 struct ieee80211_chanctx_conf *ctx) 3908 { 3909 u16 *phy_ctxt_id = (u16 *)ctx->drv_priv; 3910 struct iwl_mvm_phy_ctxt *phy_ctxt; 3911 bool responder = iwl_mvm_is_ftm_responder_chanctx(mvm, ctx); 3912 struct cfg80211_chan_def *def = responder ? &ctx->def : &ctx->min_def; 3913 int ret; 3914 3915 lockdep_assert_held(&mvm->mutex); 3916 3917 IWL_DEBUG_MAC80211(mvm, "Add channel context\n"); 3918 3919 phy_ctxt = iwl_mvm_get_free_phy_ctxt(mvm); 3920 if (!phy_ctxt) { 3921 ret = -ENOSPC; 3922 goto out; 3923 } 3924 3925 ret = iwl_mvm_phy_ctxt_changed(mvm, phy_ctxt, def, 3926 ctx->rx_chains_static, 3927 ctx->rx_chains_dynamic); 3928 if (ret) { 3929 IWL_ERR(mvm, "Failed to add PHY context\n"); 3930 goto out; 3931 } 3932 3933 iwl_mvm_phy_ctxt_ref(mvm, phy_ctxt); 3934 *phy_ctxt_id = phy_ctxt->id; 3935 out: 3936 return ret; 3937 } 3938 3939 static int iwl_mvm_add_chanctx(struct ieee80211_hw *hw, 3940 struct ieee80211_chanctx_conf *ctx) 3941 { 3942 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3943 int ret; 3944 3945 mutex_lock(&mvm->mutex); 3946 ret = __iwl_mvm_add_chanctx(mvm, ctx); 3947 mutex_unlock(&mvm->mutex); 3948 3949 return ret; 3950 } 3951 3952 static void __iwl_mvm_remove_chanctx(struct iwl_mvm *mvm, 3953 struct ieee80211_chanctx_conf *ctx) 3954 { 3955 u16 *phy_ctxt_id = (u16 *)ctx->drv_priv; 3956 struct iwl_mvm_phy_ctxt *phy_ctxt = &mvm->phy_ctxts[*phy_ctxt_id]; 3957 3958 lockdep_assert_held(&mvm->mutex); 3959 3960 iwl_mvm_phy_ctxt_unref(mvm, phy_ctxt); 3961 } 3962 3963 static void iwl_mvm_remove_chanctx(struct ieee80211_hw *hw, 3964 struct ieee80211_chanctx_conf *ctx) 3965 { 3966 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3967 3968 mutex_lock(&mvm->mutex); 3969 __iwl_mvm_remove_chanctx(mvm, ctx); 3970 mutex_unlock(&mvm->mutex); 3971 } 3972 3973 static void iwl_mvm_change_chanctx(struct ieee80211_hw *hw, 3974 struct ieee80211_chanctx_conf *ctx, 3975 u32 changed) 3976 { 3977 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 3978 u16 *phy_ctxt_id = (u16 *)ctx->drv_priv; 3979 struct iwl_mvm_phy_ctxt *phy_ctxt = &mvm->phy_ctxts[*phy_ctxt_id]; 3980 bool responder = iwl_mvm_is_ftm_responder_chanctx(mvm, ctx); 3981 struct cfg80211_chan_def *def = responder ? &ctx->def : &ctx->min_def; 3982 3983 if (WARN_ONCE((phy_ctxt->ref > 1) && 3984 (changed & ~(IEEE80211_CHANCTX_CHANGE_WIDTH | 3985 IEEE80211_CHANCTX_CHANGE_RX_CHAINS | 3986 IEEE80211_CHANCTX_CHANGE_RADAR | 3987 IEEE80211_CHANCTX_CHANGE_MIN_WIDTH)), 3988 "Cannot change PHY. Ref=%d, changed=0x%X\n", 3989 phy_ctxt->ref, changed)) 3990 return; 3991 3992 mutex_lock(&mvm->mutex); 3993 3994 /* we are only changing the min_width, may be a noop */ 3995 if (changed == IEEE80211_CHANCTX_CHANGE_MIN_WIDTH) { 3996 if (phy_ctxt->width == def->width) 3997 goto out_unlock; 3998 3999 /* we are just toggling between 20_NOHT and 20 */ 4000 if (phy_ctxt->width <= NL80211_CHAN_WIDTH_20 && 4001 def->width <= NL80211_CHAN_WIDTH_20) 4002 goto out_unlock; 4003 } 4004 4005 iwl_mvm_bt_coex_vif_change(mvm); 4006 iwl_mvm_phy_ctxt_changed(mvm, phy_ctxt, def, 4007 ctx->rx_chains_static, 4008 ctx->rx_chains_dynamic); 4009 4010 out_unlock: 4011 mutex_unlock(&mvm->mutex); 4012 } 4013 4014 static int __iwl_mvm_assign_vif_chanctx(struct iwl_mvm *mvm, 4015 struct ieee80211_vif *vif, 4016 struct ieee80211_chanctx_conf *ctx, 4017 bool switching_chanctx) 4018 { 4019 u16 *phy_ctxt_id = (u16 *)ctx->drv_priv; 4020 struct iwl_mvm_phy_ctxt *phy_ctxt = &mvm->phy_ctxts[*phy_ctxt_id]; 4021 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4022 int ret; 4023 4024 lockdep_assert_held(&mvm->mutex); 4025 4026 mvmvif->phy_ctxt = phy_ctxt; 4027 4028 switch (vif->type) { 4029 case NL80211_IFTYPE_AP: 4030 /* only needed if we're switching chanctx (i.e. during CSA) */ 4031 if (switching_chanctx) { 4032 mvmvif->ap_ibss_active = true; 4033 break; 4034 } 4035 /* fall through */ 4036 case NL80211_IFTYPE_ADHOC: 4037 /* 4038 * The AP binding flow is handled as part of the start_ap flow 4039 * (in bss_info_changed), similarly for IBSS. 4040 */ 4041 ret = 0; 4042 goto out; 4043 case NL80211_IFTYPE_STATION: 4044 mvmvif->csa_bcn_pending = false; 4045 break; 4046 case NL80211_IFTYPE_MONITOR: 4047 /* always disable PS when a monitor interface is active */ 4048 mvmvif->ps_disabled = true; 4049 break; 4050 default: 4051 ret = -EINVAL; 4052 goto out; 4053 } 4054 4055 ret = iwl_mvm_binding_add_vif(mvm, vif); 4056 if (ret) 4057 goto out; 4058 4059 /* 4060 * Power state must be updated before quotas, 4061 * otherwise fw will complain. 4062 */ 4063 iwl_mvm_power_update_mac(mvm); 4064 4065 /* Setting the quota at this stage is only required for monitor 4066 * interfaces. For the other types, the bss_info changed flow 4067 * will handle quota settings. 4068 */ 4069 if (vif->type == NL80211_IFTYPE_MONITOR) { 4070 mvmvif->monitor_active = true; 4071 ret = iwl_mvm_update_quotas(mvm, false, NULL); 4072 if (ret) 4073 goto out_remove_binding; 4074 4075 ret = iwl_mvm_add_snif_sta(mvm, vif); 4076 if (ret) 4077 goto out_remove_binding; 4078 4079 } 4080 4081 /* Handle binding during CSA */ 4082 if (vif->type == NL80211_IFTYPE_AP) { 4083 iwl_mvm_update_quotas(mvm, false, NULL); 4084 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 4085 } 4086 4087 if (switching_chanctx && vif->type == NL80211_IFTYPE_STATION) { 4088 mvmvif->csa_bcn_pending = true; 4089 4090 if (!fw_has_capa(&mvm->fw->ucode_capa, 4091 IWL_UCODE_TLV_CAPA_CHANNEL_SWITCH_CMD)) { 4092 u32 duration = 3 * vif->bss_conf.beacon_int; 4093 4094 4095 /* iwl_mvm_protect_session() reads directly from the 4096 * device (the system time), so make sure it is 4097 * available. 4098 */ 4099 ret = iwl_mvm_ref_sync(mvm, IWL_MVM_REF_PROTECT_CSA); 4100 if (ret) 4101 goto out_remove_binding; 4102 4103 /* Protect the session to make sure we hear the first 4104 * beacon on the new channel. 4105 */ 4106 iwl_mvm_protect_session(mvm, vif, duration, duration, 4107 vif->bss_conf.beacon_int / 2, 4108 true); 4109 4110 iwl_mvm_unref(mvm, IWL_MVM_REF_PROTECT_CSA); 4111 } 4112 4113 iwl_mvm_update_quotas(mvm, false, NULL); 4114 } 4115 4116 goto out; 4117 4118 out_remove_binding: 4119 iwl_mvm_binding_remove_vif(mvm, vif); 4120 iwl_mvm_power_update_mac(mvm); 4121 out: 4122 if (ret) 4123 mvmvif->phy_ctxt = NULL; 4124 return ret; 4125 } 4126 static int iwl_mvm_assign_vif_chanctx(struct ieee80211_hw *hw, 4127 struct ieee80211_vif *vif, 4128 struct ieee80211_chanctx_conf *ctx) 4129 { 4130 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4131 int ret; 4132 4133 mutex_lock(&mvm->mutex); 4134 ret = __iwl_mvm_assign_vif_chanctx(mvm, vif, ctx, false); 4135 mutex_unlock(&mvm->mutex); 4136 4137 return ret; 4138 } 4139 4140 static void __iwl_mvm_unassign_vif_chanctx(struct iwl_mvm *mvm, 4141 struct ieee80211_vif *vif, 4142 struct ieee80211_chanctx_conf *ctx, 4143 bool switching_chanctx) 4144 { 4145 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4146 struct ieee80211_vif *disabled_vif = NULL; 4147 4148 lockdep_assert_held(&mvm->mutex); 4149 4150 iwl_mvm_remove_time_event(mvm, mvmvif, &mvmvif->time_event_data); 4151 4152 switch (vif->type) { 4153 case NL80211_IFTYPE_ADHOC: 4154 goto out; 4155 case NL80211_IFTYPE_MONITOR: 4156 mvmvif->monitor_active = false; 4157 mvmvif->ps_disabled = false; 4158 iwl_mvm_rm_snif_sta(mvm, vif); 4159 break; 4160 case NL80211_IFTYPE_AP: 4161 /* This part is triggered only during CSA */ 4162 if (!switching_chanctx || !mvmvif->ap_ibss_active) 4163 goto out; 4164 4165 mvmvif->csa_countdown = false; 4166 4167 /* Set CS bit on all the stations */ 4168 iwl_mvm_modify_all_sta_disable_tx(mvm, mvmvif, true); 4169 4170 /* Save blocked iface, the timeout is set on the next beacon */ 4171 rcu_assign_pointer(mvm->csa_tx_blocked_vif, vif); 4172 4173 mvmvif->ap_ibss_active = false; 4174 break; 4175 case NL80211_IFTYPE_STATION: 4176 if (!switching_chanctx) 4177 break; 4178 4179 disabled_vif = vif; 4180 4181 if (!fw_has_capa(&mvm->fw->ucode_capa, 4182 IWL_UCODE_TLV_CAPA_CHANNEL_SWITCH_CMD)) 4183 iwl_mvm_mac_ctxt_changed(mvm, vif, true, NULL); 4184 break; 4185 default: 4186 break; 4187 } 4188 4189 iwl_mvm_update_quotas(mvm, false, disabled_vif); 4190 iwl_mvm_binding_remove_vif(mvm, vif); 4191 4192 out: 4193 mvmvif->phy_ctxt = NULL; 4194 iwl_mvm_power_update_mac(mvm); 4195 } 4196 4197 static void iwl_mvm_unassign_vif_chanctx(struct ieee80211_hw *hw, 4198 struct ieee80211_vif *vif, 4199 struct ieee80211_chanctx_conf *ctx) 4200 { 4201 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4202 4203 mutex_lock(&mvm->mutex); 4204 __iwl_mvm_unassign_vif_chanctx(mvm, vif, ctx, false); 4205 mutex_unlock(&mvm->mutex); 4206 } 4207 4208 static int 4209 iwl_mvm_switch_vif_chanctx_swap(struct iwl_mvm *mvm, 4210 struct ieee80211_vif_chanctx_switch *vifs) 4211 { 4212 int ret; 4213 4214 mutex_lock(&mvm->mutex); 4215 __iwl_mvm_unassign_vif_chanctx(mvm, vifs[0].vif, vifs[0].old_ctx, true); 4216 __iwl_mvm_remove_chanctx(mvm, vifs[0].old_ctx); 4217 4218 ret = __iwl_mvm_add_chanctx(mvm, vifs[0].new_ctx); 4219 if (ret) { 4220 IWL_ERR(mvm, "failed to add new_ctx during channel switch\n"); 4221 goto out_reassign; 4222 } 4223 4224 ret = __iwl_mvm_assign_vif_chanctx(mvm, vifs[0].vif, vifs[0].new_ctx, 4225 true); 4226 if (ret) { 4227 IWL_ERR(mvm, 4228 "failed to assign new_ctx during channel switch\n"); 4229 goto out_remove; 4230 } 4231 4232 /* we don't support TDLS during DCM - can be caused by channel switch */ 4233 if (iwl_mvm_phy_ctx_count(mvm) > 1) 4234 iwl_mvm_teardown_tdls_peers(mvm); 4235 4236 goto out; 4237 4238 out_remove: 4239 __iwl_mvm_remove_chanctx(mvm, vifs[0].new_ctx); 4240 4241 out_reassign: 4242 if (__iwl_mvm_add_chanctx(mvm, vifs[0].old_ctx)) { 4243 IWL_ERR(mvm, "failed to add old_ctx back after failure.\n"); 4244 goto out_restart; 4245 } 4246 4247 if (__iwl_mvm_assign_vif_chanctx(mvm, vifs[0].vif, vifs[0].old_ctx, 4248 true)) { 4249 IWL_ERR(mvm, "failed to reassign old_ctx after failure.\n"); 4250 goto out_restart; 4251 } 4252 4253 goto out; 4254 4255 out_restart: 4256 /* things keep failing, better restart the hw */ 4257 iwl_mvm_nic_restart(mvm, false); 4258 4259 out: 4260 mutex_unlock(&mvm->mutex); 4261 4262 return ret; 4263 } 4264 4265 static int 4266 iwl_mvm_switch_vif_chanctx_reassign(struct iwl_mvm *mvm, 4267 struct ieee80211_vif_chanctx_switch *vifs) 4268 { 4269 int ret; 4270 4271 mutex_lock(&mvm->mutex); 4272 __iwl_mvm_unassign_vif_chanctx(mvm, vifs[0].vif, vifs[0].old_ctx, true); 4273 4274 ret = __iwl_mvm_assign_vif_chanctx(mvm, vifs[0].vif, vifs[0].new_ctx, 4275 true); 4276 if (ret) { 4277 IWL_ERR(mvm, 4278 "failed to assign new_ctx during channel switch\n"); 4279 goto out_reassign; 4280 } 4281 4282 goto out; 4283 4284 out_reassign: 4285 if (__iwl_mvm_assign_vif_chanctx(mvm, vifs[0].vif, vifs[0].old_ctx, 4286 true)) { 4287 IWL_ERR(mvm, "failed to reassign old_ctx after failure.\n"); 4288 goto out_restart; 4289 } 4290 4291 goto out; 4292 4293 out_restart: 4294 /* things keep failing, better restart the hw */ 4295 iwl_mvm_nic_restart(mvm, false); 4296 4297 out: 4298 mutex_unlock(&mvm->mutex); 4299 4300 return ret; 4301 } 4302 4303 static int iwl_mvm_switch_vif_chanctx(struct ieee80211_hw *hw, 4304 struct ieee80211_vif_chanctx_switch *vifs, 4305 int n_vifs, 4306 enum ieee80211_chanctx_switch_mode mode) 4307 { 4308 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4309 int ret; 4310 4311 /* we only support a single-vif right now */ 4312 if (n_vifs > 1) 4313 return -EOPNOTSUPP; 4314 4315 switch (mode) { 4316 case CHANCTX_SWMODE_SWAP_CONTEXTS: 4317 ret = iwl_mvm_switch_vif_chanctx_swap(mvm, vifs); 4318 break; 4319 case CHANCTX_SWMODE_REASSIGN_VIF: 4320 ret = iwl_mvm_switch_vif_chanctx_reassign(mvm, vifs); 4321 break; 4322 default: 4323 ret = -EOPNOTSUPP; 4324 break; 4325 } 4326 4327 return ret; 4328 } 4329 4330 static int iwl_mvm_tx_last_beacon(struct ieee80211_hw *hw) 4331 { 4332 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4333 4334 return mvm->ibss_manager; 4335 } 4336 4337 static int iwl_mvm_set_tim(struct ieee80211_hw *hw, 4338 struct ieee80211_sta *sta, 4339 bool set) 4340 { 4341 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4342 struct iwl_mvm_sta *mvm_sta = iwl_mvm_sta_from_mac80211(sta); 4343 4344 if (!mvm_sta || !mvm_sta->vif) { 4345 IWL_ERR(mvm, "Station is not associated to a vif\n"); 4346 return -EINVAL; 4347 } 4348 4349 return iwl_mvm_mac_ctxt_beacon_changed(mvm, mvm_sta->vif); 4350 } 4351 4352 #ifdef CONFIG_NL80211_TESTMODE 4353 static const struct nla_policy iwl_mvm_tm_policy[IWL_MVM_TM_ATTR_MAX + 1] = { 4354 [IWL_MVM_TM_ATTR_CMD] = { .type = NLA_U32 }, 4355 [IWL_MVM_TM_ATTR_NOA_DURATION] = { .type = NLA_U32 }, 4356 [IWL_MVM_TM_ATTR_BEACON_FILTER_STATE] = { .type = NLA_U32 }, 4357 }; 4358 4359 static int __iwl_mvm_mac_testmode_cmd(struct iwl_mvm *mvm, 4360 struct ieee80211_vif *vif, 4361 void *data, int len) 4362 { 4363 struct nlattr *tb[IWL_MVM_TM_ATTR_MAX + 1]; 4364 int err; 4365 u32 noa_duration; 4366 4367 err = nla_parse(tb, IWL_MVM_TM_ATTR_MAX, data, len, iwl_mvm_tm_policy, 4368 NULL); 4369 if (err) 4370 return err; 4371 4372 if (!tb[IWL_MVM_TM_ATTR_CMD]) 4373 return -EINVAL; 4374 4375 switch (nla_get_u32(tb[IWL_MVM_TM_ATTR_CMD])) { 4376 case IWL_MVM_TM_CMD_SET_NOA: 4377 if (!vif || vif->type != NL80211_IFTYPE_AP || !vif->p2p || 4378 !vif->bss_conf.enable_beacon || 4379 !tb[IWL_MVM_TM_ATTR_NOA_DURATION]) 4380 return -EINVAL; 4381 4382 noa_duration = nla_get_u32(tb[IWL_MVM_TM_ATTR_NOA_DURATION]); 4383 if (noa_duration >= vif->bss_conf.beacon_int) 4384 return -EINVAL; 4385 4386 mvm->noa_duration = noa_duration; 4387 mvm->noa_vif = vif; 4388 4389 return iwl_mvm_update_quotas(mvm, true, NULL); 4390 case IWL_MVM_TM_CMD_SET_BEACON_FILTER: 4391 /* must be associated client vif - ignore authorized */ 4392 if (!vif || vif->type != NL80211_IFTYPE_STATION || 4393 !vif->bss_conf.assoc || !vif->bss_conf.dtim_period || 4394 !tb[IWL_MVM_TM_ATTR_BEACON_FILTER_STATE]) 4395 return -EINVAL; 4396 4397 if (nla_get_u32(tb[IWL_MVM_TM_ATTR_BEACON_FILTER_STATE])) 4398 return iwl_mvm_enable_beacon_filter(mvm, vif, 0); 4399 return iwl_mvm_disable_beacon_filter(mvm, vif, 0); 4400 } 4401 4402 return -EOPNOTSUPP; 4403 } 4404 4405 static int iwl_mvm_mac_testmode_cmd(struct ieee80211_hw *hw, 4406 struct ieee80211_vif *vif, 4407 void *data, int len) 4408 { 4409 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4410 int err; 4411 4412 mutex_lock(&mvm->mutex); 4413 err = __iwl_mvm_mac_testmode_cmd(mvm, vif, data, len); 4414 mutex_unlock(&mvm->mutex); 4415 4416 return err; 4417 } 4418 #endif 4419 4420 static void iwl_mvm_channel_switch(struct ieee80211_hw *hw, 4421 struct ieee80211_vif *vif, 4422 struct ieee80211_channel_switch *chsw) 4423 { 4424 /* By implementing this operation, we prevent mac80211 from 4425 * starting its own channel switch timer, so that we can call 4426 * ieee80211_chswitch_done() ourselves at the right time 4427 * (which is when the absence time event starts). 4428 */ 4429 4430 IWL_DEBUG_MAC80211(IWL_MAC80211_GET_MVM(hw), 4431 "dummy channel switch op\n"); 4432 } 4433 4434 static int iwl_mvm_schedule_client_csa(struct iwl_mvm *mvm, 4435 struct ieee80211_vif *vif, 4436 struct ieee80211_channel_switch *chsw) 4437 { 4438 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4439 struct iwl_chan_switch_te_cmd cmd = { 4440 .mac_id = cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, 4441 mvmvif->color)), 4442 .action = cpu_to_le32(FW_CTXT_ACTION_ADD), 4443 .tsf = cpu_to_le32(chsw->timestamp), 4444 .cs_count = chsw->count, 4445 }; 4446 4447 lockdep_assert_held(&mvm->mutex); 4448 4449 return iwl_mvm_send_cmd_pdu(mvm, 4450 WIDE_ID(MAC_CONF_GROUP, 4451 CHANNEL_SWITCH_TIME_EVENT_CMD), 4452 0, sizeof(cmd), &cmd); 4453 } 4454 4455 static int iwl_mvm_pre_channel_switch(struct ieee80211_hw *hw, 4456 struct ieee80211_vif *vif, 4457 struct ieee80211_channel_switch *chsw) 4458 { 4459 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4460 struct ieee80211_vif *csa_vif; 4461 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4462 u32 apply_time; 4463 int ret; 4464 4465 mutex_lock(&mvm->mutex); 4466 4467 mvmvif->csa_failed = false; 4468 4469 IWL_DEBUG_MAC80211(mvm, "pre CSA to freq %d\n", 4470 chsw->chandef.center_freq1); 4471 4472 iwl_fw_dbg_trigger_simple_stop(&mvm->fwrt, 4473 ieee80211_vif_to_wdev(vif), 4474 FW_DBG_TRIGGER_CHANNEL_SWITCH); 4475 4476 switch (vif->type) { 4477 case NL80211_IFTYPE_AP: 4478 csa_vif = 4479 rcu_dereference_protected(mvm->csa_vif, 4480 lockdep_is_held(&mvm->mutex)); 4481 if (WARN_ONCE(csa_vif && csa_vif->csa_active, 4482 "Another CSA is already in progress")) { 4483 ret = -EBUSY; 4484 goto out_unlock; 4485 } 4486 4487 /* we still didn't unblock tx. prevent new CS meanwhile */ 4488 if (rcu_dereference_protected(mvm->csa_tx_blocked_vif, 4489 lockdep_is_held(&mvm->mutex))) { 4490 ret = -EBUSY; 4491 goto out_unlock; 4492 } 4493 4494 rcu_assign_pointer(mvm->csa_vif, vif); 4495 4496 if (WARN_ONCE(mvmvif->csa_countdown, 4497 "Previous CSA countdown didn't complete")) { 4498 ret = -EBUSY; 4499 goto out_unlock; 4500 } 4501 4502 mvmvif->csa_target_freq = chsw->chandef.chan->center_freq; 4503 4504 break; 4505 case NL80211_IFTYPE_STATION: 4506 /* Schedule the time event to a bit before beacon 1, 4507 * to make sure we're in the new channel when the 4508 * GO/AP arrives. In case count <= 1 immediately schedule the 4509 * TE (this might result with some packet loss or connection 4510 * loss). 4511 */ 4512 if (chsw->count <= 1) 4513 apply_time = 0; 4514 else 4515 apply_time = chsw->device_timestamp + 4516 ((vif->bss_conf.beacon_int * (chsw->count - 1) - 4517 IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT) * 1024); 4518 4519 if (chsw->block_tx) 4520 iwl_mvm_csa_client_absent(mvm, vif); 4521 4522 if (mvmvif->bf_data.bf_enabled) { 4523 ret = iwl_mvm_disable_beacon_filter(mvm, vif, 0); 4524 if (ret) 4525 goto out_unlock; 4526 } 4527 4528 if (fw_has_capa(&mvm->fw->ucode_capa, 4529 IWL_UCODE_TLV_CAPA_CHANNEL_SWITCH_CMD)) 4530 iwl_mvm_schedule_client_csa(mvm, vif, chsw); 4531 else 4532 iwl_mvm_schedule_csa_period(mvm, vif, 4533 vif->bss_conf.beacon_int, 4534 apply_time); 4535 break; 4536 default: 4537 break; 4538 } 4539 4540 mvmvif->ps_disabled = true; 4541 4542 ret = iwl_mvm_power_update_ps(mvm); 4543 if (ret) 4544 goto out_unlock; 4545 4546 /* we won't be on this channel any longer */ 4547 iwl_mvm_teardown_tdls_peers(mvm); 4548 4549 out_unlock: 4550 mutex_unlock(&mvm->mutex); 4551 4552 return ret; 4553 } 4554 4555 static int iwl_mvm_post_channel_switch(struct ieee80211_hw *hw, 4556 struct ieee80211_vif *vif) 4557 { 4558 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4559 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4560 int ret; 4561 4562 mutex_lock(&mvm->mutex); 4563 4564 if (mvmvif->csa_failed) { 4565 mvmvif->csa_failed = false; 4566 ret = -EIO; 4567 goto out_unlock; 4568 } 4569 4570 if (vif->type == NL80211_IFTYPE_STATION) { 4571 struct iwl_mvm_sta *mvmsta; 4572 4573 mvmvif->csa_bcn_pending = false; 4574 mvmsta = iwl_mvm_sta_from_staid_protected(mvm, 4575 mvmvif->ap_sta_id); 4576 4577 if (WARN_ON(!mvmsta)) { 4578 ret = -EIO; 4579 goto out_unlock; 4580 } 4581 4582 iwl_mvm_sta_modify_disable_tx(mvm, mvmsta, false); 4583 4584 iwl_mvm_mac_ctxt_changed(mvm, vif, false, NULL); 4585 4586 ret = iwl_mvm_enable_beacon_filter(mvm, vif, 0); 4587 if (ret) 4588 goto out_unlock; 4589 4590 iwl_mvm_stop_session_protection(mvm, vif); 4591 } 4592 4593 mvmvif->ps_disabled = false; 4594 4595 ret = iwl_mvm_power_update_ps(mvm); 4596 4597 out_unlock: 4598 mutex_unlock(&mvm->mutex); 4599 4600 return ret; 4601 } 4602 4603 static void iwl_mvm_flush_no_vif(struct iwl_mvm *mvm, u32 queues, bool drop) 4604 { 4605 int i; 4606 4607 if (!iwl_mvm_has_new_tx_api(mvm)) { 4608 if (drop) { 4609 mutex_lock(&mvm->mutex); 4610 iwl_mvm_flush_tx_path(mvm, 4611 iwl_mvm_flushable_queues(mvm) & queues, 0); 4612 mutex_unlock(&mvm->mutex); 4613 } else { 4614 iwl_trans_wait_tx_queues_empty(mvm->trans, queues); 4615 } 4616 return; 4617 } 4618 4619 mutex_lock(&mvm->mutex); 4620 for (i = 0; i < ARRAY_SIZE(mvm->fw_id_to_mac_id); i++) { 4621 struct ieee80211_sta *sta; 4622 4623 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[i], 4624 lockdep_is_held(&mvm->mutex)); 4625 if (IS_ERR_OR_NULL(sta)) 4626 continue; 4627 4628 if (drop) 4629 iwl_mvm_flush_sta_tids(mvm, i, 0xFF, 0); 4630 else 4631 iwl_mvm_wait_sta_queues_empty(mvm, 4632 iwl_mvm_sta_from_mac80211(sta)); 4633 } 4634 mutex_unlock(&mvm->mutex); 4635 } 4636 4637 static void iwl_mvm_mac_flush(struct ieee80211_hw *hw, 4638 struct ieee80211_vif *vif, u32 queues, bool drop) 4639 { 4640 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4641 struct iwl_mvm_vif *mvmvif; 4642 struct iwl_mvm_sta *mvmsta; 4643 struct ieee80211_sta *sta; 4644 int i; 4645 u32 msk = 0; 4646 4647 if (!vif) { 4648 iwl_mvm_flush_no_vif(mvm, queues, drop); 4649 return; 4650 } 4651 4652 if (vif->type != NL80211_IFTYPE_STATION) 4653 return; 4654 4655 /* Make sure we're done with the deferred traffic before flushing */ 4656 flush_work(&mvm->add_stream_wk); 4657 4658 mutex_lock(&mvm->mutex); 4659 mvmvif = iwl_mvm_vif_from_mac80211(vif); 4660 4661 /* flush the AP-station and all TDLS peers */ 4662 for (i = 0; i < ARRAY_SIZE(mvm->fw_id_to_mac_id); i++) { 4663 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[i], 4664 lockdep_is_held(&mvm->mutex)); 4665 if (IS_ERR_OR_NULL(sta)) 4666 continue; 4667 4668 mvmsta = iwl_mvm_sta_from_mac80211(sta); 4669 if (mvmsta->vif != vif) 4670 continue; 4671 4672 /* make sure only TDLS peers or the AP are flushed */ 4673 WARN_ON(i != mvmvif->ap_sta_id && !sta->tdls); 4674 4675 if (drop) { 4676 if (iwl_mvm_flush_sta(mvm, mvmsta, false, 0)) 4677 IWL_ERR(mvm, "flush request fail\n"); 4678 } else { 4679 msk |= mvmsta->tfd_queue_msk; 4680 if (iwl_mvm_has_new_tx_api(mvm)) 4681 iwl_mvm_wait_sta_queues_empty(mvm, mvmsta); 4682 } 4683 } 4684 4685 mutex_unlock(&mvm->mutex); 4686 4687 /* this can take a while, and we may need/want other operations 4688 * to succeed while doing this, so do it without the mutex held 4689 */ 4690 if (!drop && !iwl_mvm_has_new_tx_api(mvm)) 4691 iwl_trans_wait_tx_queues_empty(mvm->trans, msk); 4692 } 4693 4694 static int iwl_mvm_mac_get_survey(struct ieee80211_hw *hw, int idx, 4695 struct survey_info *survey) 4696 { 4697 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4698 int ret; 4699 4700 memset(survey, 0, sizeof(*survey)); 4701 4702 /* only support global statistics right now */ 4703 if (idx != 0) 4704 return -ENOENT; 4705 4706 if (!fw_has_capa(&mvm->fw->ucode_capa, 4707 IWL_UCODE_TLV_CAPA_RADIO_BEACON_STATS)) 4708 return -ENOENT; 4709 4710 mutex_lock(&mvm->mutex); 4711 4712 if (iwl_mvm_firmware_running(mvm)) { 4713 ret = iwl_mvm_request_statistics(mvm, false); 4714 if (ret) 4715 goto out; 4716 } 4717 4718 survey->filled = SURVEY_INFO_TIME | 4719 SURVEY_INFO_TIME_RX | 4720 SURVEY_INFO_TIME_TX | 4721 SURVEY_INFO_TIME_SCAN; 4722 survey->time = mvm->accu_radio_stats.on_time_rf + 4723 mvm->radio_stats.on_time_rf; 4724 do_div(survey->time, USEC_PER_MSEC); 4725 4726 survey->time_rx = mvm->accu_radio_stats.rx_time + 4727 mvm->radio_stats.rx_time; 4728 do_div(survey->time_rx, USEC_PER_MSEC); 4729 4730 survey->time_tx = mvm->accu_radio_stats.tx_time + 4731 mvm->radio_stats.tx_time; 4732 do_div(survey->time_tx, USEC_PER_MSEC); 4733 4734 survey->time_scan = mvm->accu_radio_stats.on_time_scan + 4735 mvm->radio_stats.on_time_scan; 4736 do_div(survey->time_scan, USEC_PER_MSEC); 4737 4738 ret = 0; 4739 out: 4740 mutex_unlock(&mvm->mutex); 4741 return ret; 4742 } 4743 4744 static void iwl_mvm_mac_sta_statistics(struct ieee80211_hw *hw, 4745 struct ieee80211_vif *vif, 4746 struct ieee80211_sta *sta, 4747 struct station_info *sinfo) 4748 { 4749 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4750 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4751 struct iwl_mvm_sta *mvmsta = iwl_mvm_sta_from_mac80211(sta); 4752 4753 if (mvmsta->avg_energy) { 4754 sinfo->signal_avg = mvmsta->avg_energy; 4755 sinfo->filled |= BIT_ULL(NL80211_STA_INFO_SIGNAL_AVG); 4756 } 4757 4758 /* if beacon filtering isn't on mac80211 does it anyway */ 4759 if (!(vif->driver_flags & IEEE80211_VIF_BEACON_FILTER)) 4760 return; 4761 4762 if (!vif->bss_conf.assoc) 4763 return; 4764 4765 mutex_lock(&mvm->mutex); 4766 4767 if (mvmvif->ap_sta_id != mvmsta->sta_id) 4768 goto unlock; 4769 4770 if (iwl_mvm_request_statistics(mvm, false)) 4771 goto unlock; 4772 4773 sinfo->rx_beacon = mvmvif->beacon_stats.num_beacons + 4774 mvmvif->beacon_stats.accu_num_beacons; 4775 sinfo->filled |= BIT_ULL(NL80211_STA_INFO_BEACON_RX); 4776 if (mvmvif->beacon_stats.avg_signal) { 4777 /* firmware only reports a value after RXing a few beacons */ 4778 sinfo->rx_beacon_signal_avg = mvmvif->beacon_stats.avg_signal; 4779 sinfo->filled |= BIT_ULL(NL80211_STA_INFO_BEACON_SIGNAL_AVG); 4780 } 4781 unlock: 4782 mutex_unlock(&mvm->mutex); 4783 } 4784 4785 static void iwl_mvm_event_mlme_callback(struct iwl_mvm *mvm, 4786 struct ieee80211_vif *vif, 4787 const struct ieee80211_event *event) 4788 { 4789 #define CHECK_MLME_TRIGGER(_cnt, _fmt...) \ 4790 do { \ 4791 if ((trig_mlme->_cnt) && --(trig_mlme->_cnt)) \ 4792 break; \ 4793 iwl_fw_dbg_collect_trig(&(mvm)->fwrt, trig, _fmt); \ 4794 } while (0) 4795 4796 struct iwl_fw_dbg_trigger_tlv *trig; 4797 struct iwl_fw_dbg_trigger_mlme *trig_mlme; 4798 4799 trig = iwl_fw_dbg_trigger_on(&mvm->fwrt, ieee80211_vif_to_wdev(vif), 4800 FW_DBG_TRIGGER_MLME); 4801 if (!trig) 4802 return; 4803 4804 trig_mlme = (void *)trig->data; 4805 4806 if (event->u.mlme.data == ASSOC_EVENT) { 4807 if (event->u.mlme.status == MLME_DENIED) 4808 CHECK_MLME_TRIGGER(stop_assoc_denied, 4809 "DENIED ASSOC: reason %d", 4810 event->u.mlme.reason); 4811 else if (event->u.mlme.status == MLME_TIMEOUT) 4812 CHECK_MLME_TRIGGER(stop_assoc_timeout, 4813 "ASSOC TIMEOUT"); 4814 } else if (event->u.mlme.data == AUTH_EVENT) { 4815 if (event->u.mlme.status == MLME_DENIED) 4816 CHECK_MLME_TRIGGER(stop_auth_denied, 4817 "DENIED AUTH: reason %d", 4818 event->u.mlme.reason); 4819 else if (event->u.mlme.status == MLME_TIMEOUT) 4820 CHECK_MLME_TRIGGER(stop_auth_timeout, 4821 "AUTH TIMEOUT"); 4822 } else if (event->u.mlme.data == DEAUTH_RX_EVENT) { 4823 CHECK_MLME_TRIGGER(stop_rx_deauth, 4824 "DEAUTH RX %d", event->u.mlme.reason); 4825 } else if (event->u.mlme.data == DEAUTH_TX_EVENT) { 4826 CHECK_MLME_TRIGGER(stop_tx_deauth, 4827 "DEAUTH TX %d", event->u.mlme.reason); 4828 } 4829 #undef CHECK_MLME_TRIGGER 4830 } 4831 4832 static void iwl_mvm_event_bar_rx_callback(struct iwl_mvm *mvm, 4833 struct ieee80211_vif *vif, 4834 const struct ieee80211_event *event) 4835 { 4836 struct iwl_fw_dbg_trigger_tlv *trig; 4837 struct iwl_fw_dbg_trigger_ba *ba_trig; 4838 4839 trig = iwl_fw_dbg_trigger_on(&mvm->fwrt, ieee80211_vif_to_wdev(vif), 4840 FW_DBG_TRIGGER_BA); 4841 if (!trig) 4842 return; 4843 4844 ba_trig = (void *)trig->data; 4845 4846 if (!(le16_to_cpu(ba_trig->rx_bar) & BIT(event->u.ba.tid))) 4847 return; 4848 4849 iwl_fw_dbg_collect_trig(&mvm->fwrt, trig, 4850 "BAR received from %pM, tid %d, ssn %d", 4851 event->u.ba.sta->addr, event->u.ba.tid, 4852 event->u.ba.ssn); 4853 } 4854 4855 static void iwl_mvm_mac_event_callback(struct ieee80211_hw *hw, 4856 struct ieee80211_vif *vif, 4857 const struct ieee80211_event *event) 4858 { 4859 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4860 4861 switch (event->type) { 4862 case MLME_EVENT: 4863 iwl_mvm_event_mlme_callback(mvm, vif, event); 4864 break; 4865 case BAR_RX_EVENT: 4866 iwl_mvm_event_bar_rx_callback(mvm, vif, event); 4867 break; 4868 case BA_FRAME_TIMEOUT: 4869 iwl_mvm_event_frame_timeout_callback(mvm, vif, event->u.ba.sta, 4870 event->u.ba.tid); 4871 break; 4872 default: 4873 break; 4874 } 4875 } 4876 4877 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm, 4878 struct iwl_mvm_internal_rxq_notif *notif, 4879 u32 size) 4880 { 4881 u32 qmask = BIT(mvm->trans->num_rx_queues) - 1; 4882 int ret; 4883 4884 lockdep_assert_held(&mvm->mutex); 4885 4886 if (!iwl_mvm_has_new_rx_api(mvm)) 4887 return; 4888 4889 notif->cookie = mvm->queue_sync_cookie; 4890 4891 if (notif->sync) 4892 atomic_set(&mvm->queue_sync_counter, 4893 mvm->trans->num_rx_queues); 4894 4895 ret = iwl_mvm_notify_rx_queue(mvm, qmask, (u8 *)notif, size); 4896 if (ret) { 4897 IWL_ERR(mvm, "Failed to trigger RX queues sync (%d)\n", ret); 4898 goto out; 4899 } 4900 4901 if (notif->sync) { 4902 ret = wait_event_timeout(mvm->rx_sync_waitq, 4903 atomic_read(&mvm->queue_sync_counter) == 0 || 4904 iwl_mvm_is_radio_killed(mvm), 4905 HZ); 4906 WARN_ON_ONCE(!ret && !iwl_mvm_is_radio_killed(mvm)); 4907 } 4908 4909 out: 4910 atomic_set(&mvm->queue_sync_counter, 0); 4911 mvm->queue_sync_cookie++; 4912 } 4913 4914 static void iwl_mvm_sync_rx_queues(struct ieee80211_hw *hw) 4915 { 4916 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4917 struct iwl_mvm_internal_rxq_notif data = { 4918 .type = IWL_MVM_RXQ_EMPTY, 4919 .sync = 1, 4920 }; 4921 4922 mutex_lock(&mvm->mutex); 4923 iwl_mvm_sync_rx_queues_internal(mvm, &data, sizeof(data)); 4924 mutex_unlock(&mvm->mutex); 4925 } 4926 4927 static int 4928 iwl_mvm_mac_get_ftm_responder_stats(struct ieee80211_hw *hw, 4929 struct ieee80211_vif *vif, 4930 struct cfg80211_ftm_responder_stats *stats) 4931 { 4932 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4933 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 4934 4935 if (vif->p2p || vif->type != NL80211_IFTYPE_AP || 4936 !mvmvif->ap_ibss_active || !vif->bss_conf.ftm_responder) 4937 return -EINVAL; 4938 4939 mutex_lock(&mvm->mutex); 4940 *stats = mvm->ftm_resp_stats; 4941 mutex_unlock(&mvm->mutex); 4942 4943 stats->filled = BIT(NL80211_FTM_STATS_SUCCESS_NUM) | 4944 BIT(NL80211_FTM_STATS_PARTIAL_NUM) | 4945 BIT(NL80211_FTM_STATS_FAILED_NUM) | 4946 BIT(NL80211_FTM_STATS_ASAP_NUM) | 4947 BIT(NL80211_FTM_STATS_NON_ASAP_NUM) | 4948 BIT(NL80211_FTM_STATS_TOTAL_DURATION_MSEC) | 4949 BIT(NL80211_FTM_STATS_UNKNOWN_TRIGGERS_NUM) | 4950 BIT(NL80211_FTM_STATS_RESCHEDULE_REQUESTS_NUM) | 4951 BIT(NL80211_FTM_STATS_OUT_OF_WINDOW_TRIGGERS_NUM); 4952 4953 return 0; 4954 } 4955 4956 static int iwl_mvm_start_pmsr(struct ieee80211_hw *hw, 4957 struct ieee80211_vif *vif, 4958 struct cfg80211_pmsr_request *request) 4959 { 4960 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4961 int ret; 4962 4963 mutex_lock(&mvm->mutex); 4964 ret = iwl_mvm_ftm_start(mvm, vif, request); 4965 mutex_unlock(&mvm->mutex); 4966 4967 return ret; 4968 } 4969 4970 static void iwl_mvm_abort_pmsr(struct ieee80211_hw *hw, 4971 struct ieee80211_vif *vif, 4972 struct cfg80211_pmsr_request *request) 4973 { 4974 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4975 4976 mutex_lock(&mvm->mutex); 4977 iwl_mvm_ftm_abort(mvm, request); 4978 mutex_unlock(&mvm->mutex); 4979 } 4980 4981 static bool iwl_mvm_can_hw_csum(struct sk_buff *skb) 4982 { 4983 u8 protocol = ip_hdr(skb)->protocol; 4984 4985 if (!IS_ENABLED(CONFIG_INET)) 4986 return false; 4987 4988 return protocol == IPPROTO_TCP || protocol == IPPROTO_UDP; 4989 } 4990 4991 static bool iwl_mvm_mac_can_aggregate(struct ieee80211_hw *hw, 4992 struct sk_buff *head, 4993 struct sk_buff *skb) 4994 { 4995 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 4996 4997 /* For now don't aggregate IPv6 in AMSDU */ 4998 if (skb->protocol != htons(ETH_P_IP)) 4999 return false; 5000 5001 if (!iwl_mvm_is_csum_supported(mvm)) 5002 return true; 5003 5004 return iwl_mvm_can_hw_csum(skb) == iwl_mvm_can_hw_csum(head); 5005 } 5006 5007 const struct ieee80211_ops iwl_mvm_hw_ops = { 5008 .tx = iwl_mvm_mac_tx, 5009 .wake_tx_queue = iwl_mvm_mac_wake_tx_queue, 5010 .ampdu_action = iwl_mvm_mac_ampdu_action, 5011 .start = iwl_mvm_mac_start, 5012 .reconfig_complete = iwl_mvm_mac_reconfig_complete, 5013 .stop = iwl_mvm_mac_stop, 5014 .add_interface = iwl_mvm_mac_add_interface, 5015 .remove_interface = iwl_mvm_mac_remove_interface, 5016 .config = iwl_mvm_mac_config, 5017 .prepare_multicast = iwl_mvm_prepare_multicast, 5018 .configure_filter = iwl_mvm_configure_filter, 5019 .config_iface_filter = iwl_mvm_config_iface_filter, 5020 .bss_info_changed = iwl_mvm_bss_info_changed, 5021 .hw_scan = iwl_mvm_mac_hw_scan, 5022 .cancel_hw_scan = iwl_mvm_mac_cancel_hw_scan, 5023 .sta_pre_rcu_remove = iwl_mvm_sta_pre_rcu_remove, 5024 .sta_state = iwl_mvm_mac_sta_state, 5025 .sta_notify = iwl_mvm_mac_sta_notify, 5026 .allow_buffered_frames = iwl_mvm_mac_allow_buffered_frames, 5027 .release_buffered_frames = iwl_mvm_mac_release_buffered_frames, 5028 .set_rts_threshold = iwl_mvm_mac_set_rts_threshold, 5029 .sta_rc_update = iwl_mvm_sta_rc_update, 5030 .conf_tx = iwl_mvm_mac_conf_tx, 5031 .mgd_prepare_tx = iwl_mvm_mac_mgd_prepare_tx, 5032 .mgd_protect_tdls_discover = iwl_mvm_mac_mgd_protect_tdls_discover, 5033 .flush = iwl_mvm_mac_flush, 5034 .sched_scan_start = iwl_mvm_mac_sched_scan_start, 5035 .sched_scan_stop = iwl_mvm_mac_sched_scan_stop, 5036 .set_key = iwl_mvm_mac_set_key, 5037 .update_tkip_key = iwl_mvm_mac_update_tkip_key, 5038 .remain_on_channel = iwl_mvm_roc, 5039 .cancel_remain_on_channel = iwl_mvm_cancel_roc, 5040 .add_chanctx = iwl_mvm_add_chanctx, 5041 .remove_chanctx = iwl_mvm_remove_chanctx, 5042 .change_chanctx = iwl_mvm_change_chanctx, 5043 .assign_vif_chanctx = iwl_mvm_assign_vif_chanctx, 5044 .unassign_vif_chanctx = iwl_mvm_unassign_vif_chanctx, 5045 .switch_vif_chanctx = iwl_mvm_switch_vif_chanctx, 5046 5047 .start_ap = iwl_mvm_start_ap_ibss, 5048 .stop_ap = iwl_mvm_stop_ap_ibss, 5049 .join_ibss = iwl_mvm_start_ap_ibss, 5050 .leave_ibss = iwl_mvm_stop_ap_ibss, 5051 5052 .tx_last_beacon = iwl_mvm_tx_last_beacon, 5053 5054 .set_tim = iwl_mvm_set_tim, 5055 5056 .channel_switch = iwl_mvm_channel_switch, 5057 .pre_channel_switch = iwl_mvm_pre_channel_switch, 5058 .post_channel_switch = iwl_mvm_post_channel_switch, 5059 5060 .tdls_channel_switch = iwl_mvm_tdls_channel_switch, 5061 .tdls_cancel_channel_switch = iwl_mvm_tdls_cancel_channel_switch, 5062 .tdls_recv_channel_switch = iwl_mvm_tdls_recv_channel_switch, 5063 5064 .event_callback = iwl_mvm_mac_event_callback, 5065 5066 .sync_rx_queues = iwl_mvm_sync_rx_queues, 5067 5068 CFG80211_TESTMODE_CMD(iwl_mvm_mac_testmode_cmd) 5069 5070 #ifdef CONFIG_PM_SLEEP 5071 /* look at d3.c */ 5072 .suspend = iwl_mvm_suspend, 5073 .resume = iwl_mvm_resume, 5074 .set_wakeup = iwl_mvm_set_wakeup, 5075 .set_rekey_data = iwl_mvm_set_rekey_data, 5076 #if IS_ENABLED(CONFIG_IPV6) 5077 .ipv6_addr_change = iwl_mvm_ipv6_addr_change, 5078 #endif 5079 .set_default_unicast_key = iwl_mvm_set_default_unicast_key, 5080 #endif 5081 .get_survey = iwl_mvm_mac_get_survey, 5082 .sta_statistics = iwl_mvm_mac_sta_statistics, 5083 .get_ftm_responder_stats = iwl_mvm_mac_get_ftm_responder_stats, 5084 .start_pmsr = iwl_mvm_start_pmsr, 5085 .abort_pmsr = iwl_mvm_abort_pmsr, 5086 5087 .can_aggregate_in_amsdu = iwl_mvm_mac_can_aggregate, 5088 #ifdef CONFIG_IWLWIFI_DEBUGFS 5089 .sta_add_debugfs = iwl_mvm_sta_add_debugfs, 5090 #endif 5091 }; 5092