1 /****************************************************************************** 2 * 3 * This file is provided under a dual BSD/GPLv2 license. When using or 4 * redistributing this file, you may do so under either license. 5 * 6 * GPL LICENSE SUMMARY 7 * 8 * Copyright(c) 2012 - 2014, 2018 - 2020 Intel Corporation. All rights reserved. 9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH 11 * 12 * This program is free software; you can redistribute it and/or modify 13 * it under the terms of version 2 of the GNU General Public License as 14 * published by the Free Software Foundation. 15 * 16 * This program is distributed in the hope that it will be useful, but 17 * WITHOUT ANY WARRANTY; without even the implied warranty of 18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 19 * General Public License for more details. 20 * 21 * The full GNU General Public License is included in this distribution 22 * in the file called COPYING. 23 * 24 * Contact Information: 25 * Intel Linux Wireless <linuxwifi@intel.com> 26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 27 * 28 * BSD LICENSE 29 * 30 * Copyright(c) 2012 - 2014, 2018 - 2020 Intel Corporation. All rights reserved. 31 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 32 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH 33 * All rights reserved. 34 * 35 * Redistribution and use in source and binary forms, with or without 36 * modification, are permitted provided that the following conditions 37 * are met: 38 * 39 * * Redistributions of source code must retain the above copyright 40 * notice, this list of conditions and the following disclaimer. 41 * * Redistributions in binary form must reproduce the above copyright 42 * notice, this list of conditions and the following disclaimer in 43 * the documentation and/or other materials provided with the 44 * distribution. 45 * * Neither the name Intel Corporation nor the names of its 46 * contributors may be used to endorse or promote products derived 47 * from this software without specific prior written permission. 48 * 49 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 50 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 51 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 52 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 53 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 54 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 55 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 56 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 57 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 58 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 59 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 60 * 61 *****************************************************************************/ 62 63 #ifndef __IWL_MVM_H__ 64 #define __IWL_MVM_H__ 65 66 #include <linux/list.h> 67 #include <linux/spinlock.h> 68 #include <linux/leds.h> 69 #include <linux/in6.h> 70 71 #ifdef CONFIG_THERMAL 72 #include <linux/thermal.h> 73 #endif 74 75 #include "iwl-op-mode.h" 76 #include "iwl-trans.h" 77 #include "fw/notif-wait.h" 78 #include "iwl-eeprom-parse.h" 79 #include "fw/file.h" 80 #include "iwl-config.h" 81 #include "sta.h" 82 #include "fw-api.h" 83 #include "constants.h" 84 #include "fw/runtime.h" 85 #include "fw/dbg.h" 86 #include "fw/acpi.h" 87 #include "iwl-nvm-parse.h" 88 89 #include <linux/average.h> 90 91 #define IWL_MVM_MAX_ADDRESSES 5 92 /* RSSI offset for WkP */ 93 #define IWL_RSSI_OFFSET 50 94 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8 95 #define IWL_MVM_MISSED_BEACONS_THRESHOLD_LONG 16 96 97 /* A TimeUnit is 1024 microsecond */ 98 #define MSEC_TO_TU(_msec) (_msec*1000/1024) 99 100 /* For GO, this value represents the number of TUs before CSA "beacon 101 * 0" TBTT when the CSA time-event needs to be scheduled to start. It 102 * must be big enough to ensure that we switch in time. 103 */ 104 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40 105 106 /* For client, this value represents the number of TUs before CSA 107 * "beacon 1" TBTT, instead. This is because we don't know when the 108 * GO/AP will be in the new channel, so we switch early enough. 109 */ 110 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10 111 112 /* 113 * This value (in TUs) is used to fine tune the CSA NoA end time which should 114 * be just before "beacon 0" TBTT. 115 */ 116 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4 117 118 /* 119 * Number of beacons to transmit on a new channel until we unblock tx to 120 * the stations, even if we didn't identify them on a new channel 121 */ 122 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3 123 124 /* offchannel queue towards mac80211 */ 125 #define IWL_MVM_OFFCHANNEL_QUEUE 0 126 127 extern const struct ieee80211_ops iwl_mvm_hw_ops; 128 129 /** 130 * struct iwl_mvm_mod_params - module parameters for iwlmvm 131 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted. 132 * We will register to mac80211 to have testmode working. The NIC must not 133 * be up'ed after the INIT fw asserted. This is useful to be able to use 134 * proprietary tools over testmode to debug the INIT fw. 135 * @power_scheme: one of enum iwl_power_scheme 136 */ 137 struct iwl_mvm_mod_params { 138 bool init_dbg; 139 int power_scheme; 140 }; 141 extern struct iwl_mvm_mod_params iwlmvm_mod_params; 142 143 struct iwl_mvm_phy_ctxt { 144 u16 id; 145 u16 color; 146 u32 ref; 147 148 enum nl80211_chan_width width; 149 150 /* 151 * TODO: This should probably be removed. Currently here only for rate 152 * scaling algorithm 153 */ 154 struct ieee80211_channel *channel; 155 }; 156 157 struct iwl_mvm_time_event_data { 158 struct ieee80211_vif *vif; 159 struct list_head list; 160 unsigned long end_jiffies; 161 u32 duration; 162 bool running; 163 u32 uid; 164 165 /* 166 * The access to the 'id' field must be done when the 167 * mvm->time_event_lock is held, as it value is used to indicate 168 * if the te is in the time event list or not (when id == TE_MAX) 169 */ 170 u32 id; 171 }; 172 173 /* Power management */ 174 175 /** 176 * enum iwl_power_scheme 177 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode 178 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default) 179 * @IWL_POWER_LEVEL_LP - Low Power 180 */ 181 enum iwl_power_scheme { 182 IWL_POWER_SCHEME_CAM = 1, 183 IWL_POWER_SCHEME_BPS, 184 IWL_POWER_SCHEME_LP 185 }; 186 187 #define IWL_CONN_MAX_LISTEN_INTERVAL 10 188 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL 189 190 #ifdef CONFIG_IWLWIFI_DEBUGFS 191 enum iwl_dbgfs_pm_mask { 192 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0), 193 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1), 194 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2), 195 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3), 196 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4), 197 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6), 198 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7), 199 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8), 200 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9), 201 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10), 202 }; 203 204 struct iwl_dbgfs_pm { 205 u16 keep_alive_seconds; 206 u32 rx_data_timeout; 207 u32 tx_data_timeout; 208 bool skip_over_dtim; 209 u8 skip_dtim_periods; 210 bool lprx_ena; 211 u32 lprx_rssi_threshold; 212 bool snooze_ena; 213 bool uapsd_misbehaving; 214 bool use_ps_poll; 215 int mask; 216 }; 217 218 /* beacon filtering */ 219 220 enum iwl_dbgfs_bf_mask { 221 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0), 222 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1), 223 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2), 224 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3), 225 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4), 226 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5), 227 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6), 228 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7), 229 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8), 230 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9), 231 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10), 232 }; 233 234 struct iwl_dbgfs_bf { 235 u32 bf_energy_delta; 236 u32 bf_roaming_energy_delta; 237 u32 bf_roaming_state; 238 u32 bf_temp_threshold; 239 u32 bf_temp_fast_filter; 240 u32 bf_temp_slow_filter; 241 u32 bf_enable_beacon_filter; 242 u32 bf_debug_flag; 243 u32 bf_escape_timer; 244 u32 ba_escape_timer; 245 u32 ba_enable_beacon_abort; 246 int mask; 247 }; 248 #endif 249 250 enum iwl_mvm_smps_type_request { 251 IWL_MVM_SMPS_REQ_BT_COEX, 252 IWL_MVM_SMPS_REQ_TT, 253 IWL_MVM_SMPS_REQ_PROT, 254 NUM_IWL_MVM_SMPS_REQ, 255 }; 256 257 enum iwl_bt_force_ant_mode { 258 BT_FORCE_ANT_DIS = 0, 259 BT_FORCE_ANT_AUTO, 260 BT_FORCE_ANT_BT, 261 BT_FORCE_ANT_WIFI, 262 263 BT_FORCE_ANT_MAX, 264 }; 265 266 /** 267 * struct iwl_mvm_low_latency_force - low latency force mode set by debugfs 268 * @LOW_LATENCY_FORCE_UNSET: unset force mode 269 * @LOW_LATENCY_FORCE_ON: for low latency on 270 * @LOW_LATENCY_FORCE_OFF: for low latency off 271 * @NUM_LOW_LATENCY_FORCE: max num of modes 272 */ 273 enum iwl_mvm_low_latency_force { 274 LOW_LATENCY_FORCE_UNSET, 275 LOW_LATENCY_FORCE_ON, 276 LOW_LATENCY_FORCE_OFF, 277 NUM_LOW_LATENCY_FORCE 278 }; 279 280 /** 281 * struct iwl_mvm_low_latency_cause - low latency set causes 282 * @LOW_LATENCY_TRAFFIC: indicates low latency traffic was detected 283 * @LOW_LATENCY_DEBUGFS: low latency mode set from debugfs 284 * @LOW_LATENCY_VCMD: low latency mode set from vendor command 285 * @LOW_LATENCY_VIF_TYPE: low latency mode set because of vif type (ap) 286 * @LOW_LATENCY_DEBUGFS_FORCE_ENABLE: indicate that force mode is enabled 287 * the actual set/unset is done with LOW_LATENCY_DEBUGFS_FORCE 288 * @LOW_LATENCY_DEBUGFS_FORCE: low latency force mode from debugfs 289 * set this with LOW_LATENCY_DEBUGFS_FORCE_ENABLE flag 290 * in low_latency. 291 */ 292 enum iwl_mvm_low_latency_cause { 293 LOW_LATENCY_TRAFFIC = BIT(0), 294 LOW_LATENCY_DEBUGFS = BIT(1), 295 LOW_LATENCY_VCMD = BIT(2), 296 LOW_LATENCY_VIF_TYPE = BIT(3), 297 LOW_LATENCY_DEBUGFS_FORCE_ENABLE = BIT(4), 298 LOW_LATENCY_DEBUGFS_FORCE = BIT(5), 299 }; 300 301 /** 302 * struct iwl_mvm_vif_bf_data - beacon filtering related data 303 * @bf_enabled: indicates if beacon filtering is enabled 304 * @ba_enabled: indicated if beacon abort is enabled 305 * @ave_beacon_signal: average beacon signal 306 * @last_cqm_event: rssi of the last cqm event 307 * @bt_coex_min_thold: minimum threshold for BT coex 308 * @bt_coex_max_thold: maximum threshold for BT coex 309 * @last_bt_coex_event: rssi of the last BT coex event 310 */ 311 struct iwl_mvm_vif_bf_data { 312 bool bf_enabled; 313 bool ba_enabled; 314 int ave_beacon_signal; 315 int last_cqm_event; 316 int bt_coex_min_thold; 317 int bt_coex_max_thold; 318 int last_bt_coex_event; 319 }; 320 321 /** 322 * struct iwl_probe_resp_data - data for NoA/CSA updates 323 * @rcu_head: used for freeing the data on update 324 * @notif: notification data 325 * @noa_len: length of NoA attribute, calculated from the notification 326 */ 327 struct iwl_probe_resp_data { 328 struct rcu_head rcu_head; 329 struct iwl_probe_resp_data_notif notif; 330 int noa_len; 331 }; 332 333 /** 334 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context 335 * @id: between 0 and 3 336 * @color: to solve races upon MAC addition and removal 337 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA 338 * @bssid: BSSID for this (client) interface 339 * @associated: indicates that we're currently associated, used only for 340 * managing the firmware state in iwl_mvm_bss_info_changed_station() 341 * @ap_assoc_sta_count: count of stations associated to us - valid only 342 * if VIF type is AP 343 * @uploaded: indicates the MAC context has been added to the device 344 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface 345 * should get quota etc. 346 * @pm_enabled - Indicate if MAC power management is allowed 347 * @monitor_active: indicates that monitor context is configured, and that the 348 * interface should get quota etc. 349 * @low_latency: bit flags for low latency 350 * see enum &iwl_mvm_low_latency_cause for causes. 351 * @low_latency_actual: boolean, indicates low latency is set, 352 * as a result from low_latency bit flags and takes force into account. 353 * @ps_disabled: indicates that this interface requires PS to be disabled 354 * @queue_params: QoS params for this MAC 355 * @bcast_sta: station used for broadcast packets. Used by the following 356 * vifs: P2P_DEVICE, GO and AP. 357 * @beacon_skb: the skb used to hold the AP/GO beacon template 358 * @smps_requests: the SMPS requests of different parts of the driver, 359 * combined on update to yield the overall request to mac80211. 360 * @beacon_stats: beacon statistics, containing the # of received beacons, 361 * # of received beacons accumulated over FW restart, and the current 362 * average signal of beacons retrieved from the firmware 363 * @csa_failed: CSA failed to schedule time event, report an error later 364 * @features: hw features active for this vif 365 * @probe_resp_data: data from FW notification to store NOA and CSA related 366 * data to be inserted into probe response. 367 */ 368 struct iwl_mvm_vif { 369 struct iwl_mvm *mvm; 370 u16 id; 371 u16 color; 372 u8 ap_sta_id; 373 374 u8 bssid[ETH_ALEN]; 375 bool associated; 376 u8 ap_assoc_sta_count; 377 378 u16 cab_queue; 379 380 bool uploaded; 381 bool ap_ibss_active; 382 bool pm_enabled; 383 bool monitor_active; 384 u8 low_latency: 6; 385 u8 low_latency_actual: 1; 386 bool ps_disabled; 387 struct iwl_mvm_vif_bf_data bf_data; 388 389 struct { 390 u32 num_beacons, accu_num_beacons; 391 u8 avg_signal; 392 } beacon_stats; 393 394 u32 ap_beacon_time; 395 396 enum iwl_tsf_id tsf_id; 397 398 /* 399 * QoS data from mac80211, need to store this here 400 * as mac80211 has a separate callback but we need 401 * to have the data for the MAC context 402 */ 403 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS]; 404 struct iwl_mvm_time_event_data time_event_data; 405 struct iwl_mvm_time_event_data hs_time_event_data; 406 407 struct iwl_mvm_int_sta bcast_sta; 408 struct iwl_mvm_int_sta mcast_sta; 409 410 /* 411 * Assigned while mac80211 has the interface in a channel context, 412 * or, for P2P Device, while it exists. 413 */ 414 struct iwl_mvm_phy_ctxt *phy_ctxt; 415 416 #ifdef CONFIG_PM 417 /* WoWLAN GTK rekey data */ 418 struct { 419 u8 kck[NL80211_KCK_EXT_LEN]; 420 u8 kek[NL80211_KEK_EXT_LEN]; 421 size_t kek_len; 422 size_t kck_len; 423 u32 akm; 424 __le64 replay_ctr; 425 bool valid; 426 } rekey_data; 427 428 int tx_key_idx; 429 430 bool seqno_valid; 431 u16 seqno; 432 #endif 433 434 #if IS_ENABLED(CONFIG_IPV6) 435 /* IPv6 addresses for WoWLAN */ 436 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX]; 437 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)]; 438 int num_target_ipv6_addrs; 439 #endif 440 441 #ifdef CONFIG_IWLWIFI_DEBUGFS 442 struct dentry *dbgfs_dir; 443 struct dentry *dbgfs_slink; 444 struct iwl_dbgfs_pm dbgfs_pm; 445 struct iwl_dbgfs_bf dbgfs_bf; 446 struct iwl_mac_power_cmd mac_pwr_cmd; 447 int dbgfs_quota_min; 448 #endif 449 450 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ]; 451 452 /* FW identified misbehaving AP */ 453 u8 uapsd_misbehaving_bssid[ETH_ALEN]; 454 455 struct delayed_work uapsd_nonagg_detected_wk; 456 457 /* Indicates that CSA countdown may be started */ 458 bool csa_countdown; 459 bool csa_failed; 460 u16 csa_target_freq; 461 u16 csa_count; 462 u16 csa_misbehave; 463 struct delayed_work csa_work; 464 465 /* Indicates that we are waiting for a beacon on a new channel */ 466 bool csa_bcn_pending; 467 468 /* TCP Checksum Offload */ 469 netdev_features_t features; 470 471 struct iwl_probe_resp_data __rcu *probe_resp_data; 472 473 /* we can only have 2 GTK + 2 IGTK active at a time */ 474 struct ieee80211_key_conf *ap_early_keys[4]; 475 476 /* 26-tone RU OFDMA transmissions should be blocked */ 477 bool he_ru_2mhz_block; 478 }; 479 480 static inline struct iwl_mvm_vif * 481 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif) 482 { 483 if (!vif) 484 return NULL; 485 return (void *)vif->drv_priv; 486 } 487 488 extern const u8 tid_to_mac80211_ac[]; 489 490 #define IWL_MVM_SCAN_STOPPING_SHIFT 8 491 492 enum iwl_scan_status { 493 IWL_MVM_SCAN_REGULAR = BIT(0), 494 IWL_MVM_SCAN_SCHED = BIT(1), 495 IWL_MVM_SCAN_NETDETECT = BIT(2), 496 497 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8), 498 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9), 499 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10), 500 501 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR | 502 IWL_MVM_SCAN_STOPPING_REGULAR, 503 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED | 504 IWL_MVM_SCAN_STOPPING_SCHED, 505 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT | 506 IWL_MVM_SCAN_STOPPING_NETDETECT, 507 508 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT, 509 IWL_MVM_SCAN_MASK = 0xff, 510 }; 511 512 enum iwl_mvm_scan_type { 513 IWL_SCAN_TYPE_NOT_SET, 514 IWL_SCAN_TYPE_UNASSOC, 515 IWL_SCAN_TYPE_WILD, 516 IWL_SCAN_TYPE_MILD, 517 IWL_SCAN_TYPE_FRAGMENTED, 518 IWL_SCAN_TYPE_FAST_BALANCE, 519 }; 520 521 enum iwl_mvm_sched_scan_pass_all_states { 522 SCHED_SCAN_PASS_ALL_DISABLED, 523 SCHED_SCAN_PASS_ALL_ENABLED, 524 SCHED_SCAN_PASS_ALL_FOUND, 525 }; 526 527 /** 528 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure 529 * @ct_kill_exit: worker to exit thermal kill 530 * @dynamic_smps: Is thermal throttling enabled dynamic_smps? 531 * @tx_backoff: The current thremal throttling tx backoff in uSec. 532 * @min_backoff: The minimal tx backoff due to power restrictions 533 * @params: Parameters to configure the thermal throttling algorithm. 534 * @throttle: Is thermal throttling is active? 535 */ 536 struct iwl_mvm_tt_mgmt { 537 struct delayed_work ct_kill_exit; 538 bool dynamic_smps; 539 u32 tx_backoff; 540 u32 min_backoff; 541 struct iwl_tt_params params; 542 bool throttle; 543 }; 544 545 #ifdef CONFIG_THERMAL 546 /** 547 *struct iwl_mvm_thermal_device - thermal zone related data 548 * @temp_trips: temperature thresholds for report 549 * @fw_trips_index: keep indexes to original array - temp_trips 550 * @tzone: thermal zone device data 551 */ 552 struct iwl_mvm_thermal_device { 553 s16 temp_trips[IWL_MAX_DTS_TRIPS]; 554 u8 fw_trips_index[IWL_MAX_DTS_TRIPS]; 555 struct thermal_zone_device *tzone; 556 }; 557 558 /* 559 * struct iwl_mvm_cooling_device 560 * @cur_state: current state 561 * @cdev: struct thermal cooling device 562 */ 563 struct iwl_mvm_cooling_device { 564 u32 cur_state; 565 struct thermal_cooling_device *cdev; 566 }; 567 #endif 568 569 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8 570 571 struct iwl_mvm_frame_stats { 572 u32 legacy_frames; 573 u32 ht_frames; 574 u32 vht_frames; 575 u32 bw_20_frames; 576 u32 bw_40_frames; 577 u32 bw_80_frames; 578 u32 bw_160_frames; 579 u32 sgi_frames; 580 u32 ngi_frames; 581 u32 siso_frames; 582 u32 mimo2_frames; 583 u32 agg_frames; 584 u32 ampdu_count; 585 u32 success_frames; 586 u32 fail_frames; 587 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES]; 588 int last_frame_idx; 589 }; 590 591 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff 592 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100 593 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200 594 595 enum iwl_mvm_tdls_cs_state { 596 IWL_MVM_TDLS_SW_IDLE = 0, 597 IWL_MVM_TDLS_SW_REQ_SENT, 598 IWL_MVM_TDLS_SW_RESP_RCVD, 599 IWL_MVM_TDLS_SW_REQ_RCVD, 600 IWL_MVM_TDLS_SW_ACTIVE, 601 }; 602 603 enum iwl_mvm_traffic_load { 604 IWL_MVM_TRAFFIC_LOW, 605 IWL_MVM_TRAFFIC_MEDIUM, 606 IWL_MVM_TRAFFIC_HIGH, 607 }; 608 609 DECLARE_EWMA(rate, 16, 16) 610 611 struct iwl_mvm_tcm_mac { 612 struct { 613 u32 pkts[IEEE80211_NUM_ACS]; 614 u32 airtime; 615 } tx; 616 struct { 617 u32 pkts[IEEE80211_NUM_ACS]; 618 u32 airtime; 619 u32 last_ampdu_ref; 620 } rx; 621 struct { 622 /* track AP's transfer in client mode */ 623 u64 rx_bytes; 624 struct ewma_rate rate; 625 bool detected; 626 } uapsd_nonagg_detect; 627 bool opened_rx_ba_sessions; 628 }; 629 630 struct iwl_mvm_tcm { 631 struct delayed_work work; 632 spinlock_t lock; /* used when time elapsed */ 633 unsigned long ts; /* timestamp when period ends */ 634 unsigned long ll_ts; 635 unsigned long uapsd_nonagg_ts; 636 bool paused; 637 struct iwl_mvm_tcm_mac data[NUM_MAC_INDEX_DRIVER]; 638 struct { 639 u32 elapsed; /* milliseconds for this TCM period */ 640 u32 airtime[NUM_MAC_INDEX_DRIVER]; 641 enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER]; 642 enum iwl_mvm_traffic_load band_load[NUM_NL80211_BANDS]; 643 enum iwl_mvm_traffic_load global_load; 644 bool low_latency[NUM_MAC_INDEX_DRIVER]; 645 bool change[NUM_MAC_INDEX_DRIVER]; 646 bool global_change; 647 } result; 648 }; 649 650 /** 651 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer 652 * @head_sn: reorder window head sn 653 * @num_stored: number of mpdus stored in the buffer 654 * @buf_size: the reorder buffer size as set by the last addba request 655 * @queue: queue of this reorder buffer 656 * @last_amsdu: track last ASMDU SN for duplication detection 657 * @last_sub_index: track ASMDU sub frame index for duplication detection 658 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU 659 * it is the time of last received sub-frame 660 * @removed: prevent timer re-arming 661 * @valid: reordering is valid for this queue 662 * @lock: protect reorder buffer internal state 663 * @mvm: mvm pointer, needed for frame timer context 664 * @consec_oldsn_drops: consecutive drops due to old SN 665 * @consec_oldsn_ampdu_gp2: A-MPDU GP2 timestamp to track 666 * when to apply old SN consecutive drop workaround 667 * @consec_oldsn_prev_drop: track whether or not an MPDU 668 * that was single/part of the previous A-MPDU was 669 * dropped due to old SN 670 */ 671 struct iwl_mvm_reorder_buffer { 672 u16 head_sn; 673 u16 num_stored; 674 u16 buf_size; 675 int queue; 676 u16 last_amsdu; 677 u8 last_sub_index; 678 struct timer_list reorder_timer; 679 bool removed; 680 bool valid; 681 spinlock_t lock; 682 struct iwl_mvm *mvm; 683 unsigned int consec_oldsn_drops; 684 u32 consec_oldsn_ampdu_gp2; 685 unsigned int consec_oldsn_prev_drop:1; 686 } ____cacheline_aligned_in_smp; 687 688 /** 689 * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno 690 * @frames: list of skbs stored 691 * @reorder_time: time the packet was stored in the reorder buffer 692 */ 693 struct _iwl_mvm_reorder_buf_entry { 694 struct sk_buff_head frames; 695 unsigned long reorder_time; 696 }; 697 698 /* make this indirection to get the aligned thing */ 699 struct iwl_mvm_reorder_buf_entry { 700 struct _iwl_mvm_reorder_buf_entry e; 701 } 702 #ifndef __CHECKER__ 703 /* sparse doesn't like this construct: "bad integer constant expression" */ 704 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry))) 705 #endif 706 ; 707 708 /** 709 * struct iwl_mvm_baid_data - BA session data 710 * @sta_id: station id 711 * @tid: tid of the session 712 * @baid baid of the session 713 * @timeout: the timeout set in the addba request 714 * @entries_per_queue: # of buffers per queue, this actually gets 715 * aligned up to avoid cache line sharing between queues 716 * @last_rx: last rx jiffies, updated only if timeout passed from last update 717 * @session_timer: timer to check if BA session expired, runs at 2 * timeout 718 * @mvm: mvm pointer, needed for timer context 719 * @reorder_buf: reorder buffer, allocated per queue 720 * @reorder_buf_data: data 721 */ 722 struct iwl_mvm_baid_data { 723 struct rcu_head rcu_head; 724 u8 sta_id; 725 u8 tid; 726 u8 baid; 727 u16 timeout; 728 u16 entries_per_queue; 729 unsigned long last_rx; 730 struct timer_list session_timer; 731 struct iwl_mvm_baid_data __rcu **rcu_ptr; 732 struct iwl_mvm *mvm; 733 struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES]; 734 struct iwl_mvm_reorder_buf_entry entries[]; 735 }; 736 737 static inline struct iwl_mvm_baid_data * 738 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf) 739 { 740 return (void *)((u8 *)buf - 741 offsetof(struct iwl_mvm_baid_data, reorder_buf) - 742 sizeof(*buf) * buf->queue); 743 } 744 745 /* 746 * enum iwl_mvm_queue_status - queue status 747 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved 748 * Basically, this means that this queue can be used for any purpose 749 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use 750 * This is the state of a queue that has been dedicated for some RATID 751 * (agg'd or not), but that hasn't yet gone through the actual enablement 752 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet. 753 * Note that in this state there is no requirement to already know what TID 754 * should be used with this queue, it is just marked as a queue that will 755 * be used, and shouldn't be allocated to anyone else. 756 * @IWL_MVM_QUEUE_READY: queue is ready to be used 757 * This is the state of a queue that has been fully configured (including 758 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be 759 * used to send traffic. 760 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared 761 * This is a state in which a single queue serves more than one TID, all of 762 * which are not aggregated. Note that the queue is only associated to one 763 * RA. 764 */ 765 enum iwl_mvm_queue_status { 766 IWL_MVM_QUEUE_FREE, 767 IWL_MVM_QUEUE_RESERVED, 768 IWL_MVM_QUEUE_READY, 769 IWL_MVM_QUEUE_SHARED, 770 }; 771 772 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ) 773 #define IWL_MVM_INVALID_QUEUE 0xFFFF 774 775 #define IWL_MVM_NUM_CIPHERS 10 776 777 778 struct iwl_mvm_txq { 779 struct list_head list; 780 u16 txq_id; 781 atomic_t tx_request; 782 bool stopped; 783 }; 784 785 static inline struct iwl_mvm_txq * 786 iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq) 787 { 788 return (void *)txq->drv_priv; 789 } 790 791 static inline struct iwl_mvm_txq * 792 iwl_mvm_txq_from_tid(struct ieee80211_sta *sta, u8 tid) 793 { 794 if (tid == IWL_MAX_TID_COUNT) 795 tid = IEEE80211_NUM_TIDS; 796 797 return (void *)sta->txq[tid]->drv_priv; 798 } 799 800 /** 801 * struct iwl_mvm_tvqm_txq_info - maps TVQM hw queue to tid 802 * 803 * @sta_id: sta id 804 * @txq_tid: txq tid 805 */ 806 struct iwl_mvm_tvqm_txq_info { 807 u8 sta_id; 808 u8 txq_tid; 809 }; 810 811 struct iwl_mvm_dqa_txq_info { 812 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */ 813 bool reserved; /* Is this the TXQ reserved for a STA */ 814 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */ 815 u8 txq_tid; /* The TID "owner" of this queue*/ 816 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */ 817 /* Timestamp for inactivation per TID of this queue */ 818 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1]; 819 enum iwl_mvm_queue_status status; 820 }; 821 822 struct iwl_mvm { 823 /* for logger access */ 824 struct device *dev; 825 826 struct iwl_trans *trans; 827 const struct iwl_fw *fw; 828 const struct iwl_cfg *cfg; 829 struct iwl_phy_db *phy_db; 830 struct ieee80211_hw *hw; 831 832 /* for protecting access to iwl_mvm */ 833 struct mutex mutex; 834 struct list_head async_handlers_list; 835 spinlock_t async_handlers_lock; 836 struct work_struct async_handlers_wk; 837 838 struct work_struct roc_done_wk; 839 840 unsigned long init_status; 841 842 unsigned long status; 843 844 u32 queue_sync_cookie; 845 atomic_t queue_sync_counter; 846 /* 847 * for beacon filtering - 848 * currently only one interface can be supported 849 */ 850 struct iwl_mvm_vif *bf_allowed_vif; 851 852 bool hw_registered; 853 bool rfkill_safe_init_done; 854 855 u32 ampdu_ref; 856 bool ampdu_toggle; 857 858 struct iwl_notif_wait_data notif_wait; 859 860 union { 861 struct mvm_statistics_rx_v3 rx_stats_v3; 862 struct mvm_statistics_rx rx_stats; 863 }; 864 865 struct { 866 u64 rx_time; 867 u64 tx_time; 868 u64 on_time_rf; 869 u64 on_time_scan; 870 } radio_stats, accu_radio_stats; 871 872 struct list_head add_stream_txqs; 873 union { 874 struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES]; 875 struct iwl_mvm_tvqm_txq_info tvqm_info[IWL_MAX_TVQM_QUEUES]; 876 }; 877 struct work_struct add_stream_wk; /* To add streams to queues */ 878 879 const char *nvm_file_name; 880 struct iwl_nvm_data *nvm_data; 881 /* NVM sections */ 882 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS]; 883 884 struct iwl_fw_runtime fwrt; 885 886 /* EEPROM MAC addresses */ 887 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES]; 888 889 /* data related to data path */ 890 struct iwl_rx_phy_info last_phy_info; 891 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT]; 892 u8 rx_ba_sessions; 893 894 /* configured by mac80211 */ 895 u32 rts_threshold; 896 897 /* Scan status, cmd (pre-allocated) and auxiliary station */ 898 unsigned int scan_status; 899 void *scan_cmd; 900 struct iwl_mcast_filter_cmd *mcast_filter_cmd; 901 /* For CDB this is low band scan type, for non-CDB - type. */ 902 enum iwl_mvm_scan_type scan_type; 903 enum iwl_mvm_scan_type hb_scan_type; 904 905 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all; 906 struct delayed_work scan_timeout_dwork; 907 908 /* max number of simultaneous scans the FW supports */ 909 unsigned int max_scans; 910 911 /* UMAC scan tracking */ 912 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS]; 913 914 /* start time of last scan in TSF of the mac that requested the scan */ 915 u64 scan_start; 916 917 /* the vif that requested the current scan */ 918 struct iwl_mvm_vif *scan_vif; 919 920 /* rx chain antennas set through debugfs for the scan command */ 921 u8 scan_rx_ant; 922 923 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING 924 /* broadcast filters to configure for each associated station */ 925 const struct iwl_fw_bcast_filter *bcast_filters; 926 #ifdef CONFIG_IWLWIFI_DEBUGFS 927 struct { 928 bool override; 929 struct iwl_bcast_filter_cmd cmd; 930 } dbgfs_bcast_filtering; 931 #endif 932 #endif 933 934 /* Internal station */ 935 struct iwl_mvm_int_sta aux_sta; 936 struct iwl_mvm_int_sta snif_sta; 937 938 bool last_ebs_successful; 939 940 u8 scan_last_antenna_idx; /* to toggle TX between antennas */ 941 u8 mgmt_last_antenna_idx; 942 943 /* last smart fifo state that was successfully sent to firmware */ 944 enum iwl_sf_state sf_state; 945 946 #ifdef CONFIG_IWLWIFI_DEBUGFS 947 struct dentry *debugfs_dir; 948 u32 dbgfs_sram_offset, dbgfs_sram_len; 949 u32 dbgfs_prph_reg_addr; 950 bool disable_power_off; 951 bool disable_power_off_d3; 952 bool beacon_inject_active; 953 954 bool scan_iter_notif_enabled; 955 956 struct debugfs_blob_wrapper nvm_hw_blob; 957 struct debugfs_blob_wrapper nvm_sw_blob; 958 struct debugfs_blob_wrapper nvm_calib_blob; 959 struct debugfs_blob_wrapper nvm_prod_blob; 960 struct debugfs_blob_wrapper nvm_phy_sku_blob; 961 struct debugfs_blob_wrapper nvm_reg_blob; 962 963 struct iwl_mvm_frame_stats drv_rx_stats; 964 spinlock_t drv_stats_lock; 965 u16 dbgfs_rx_phyinfo; 966 #endif 967 968 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX]; 969 970 struct list_head time_event_list; 971 spinlock_t time_event_lock; 972 973 /* 974 * A bitmap indicating the index of the key in use. The firmware 975 * can hold 16 keys at most. Reflect this fact. 976 */ 977 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)]; 978 u8 fw_key_deleted[STA_KEY_MAX_NUM]; 979 980 u8 vif_count; 981 struct ieee80211_vif __rcu *vif_id_to_mac[NUM_MAC_INDEX_DRIVER]; 982 983 /* -1 for always, 0 for never, >0 for that many times */ 984 s8 fw_restart; 985 u8 *error_recovery_buf; 986 987 #ifdef CONFIG_IWLWIFI_LEDS 988 struct led_classdev led; 989 #endif 990 991 struct ieee80211_vif *p2p_device_vif; 992 993 #ifdef CONFIG_PM 994 struct wiphy_wowlan_support wowlan; 995 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen; 996 997 /* sched scan settings for net detect */ 998 struct ieee80211_scan_ies nd_ies; 999 struct cfg80211_match_set *nd_match_sets; 1000 int n_nd_match_sets; 1001 struct ieee80211_channel **nd_channels; 1002 int n_nd_channels; 1003 bool net_detect; 1004 u8 offload_tid; 1005 #ifdef CONFIG_IWLWIFI_DEBUGFS 1006 bool d3_wake_sysassert; 1007 bool d3_test_active; 1008 u32 d3_test_pme_ptr; 1009 struct ieee80211_vif *keep_vif; 1010 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */ 1011 #endif 1012 #endif 1013 1014 wait_queue_head_t rx_sync_waitq; 1015 1016 /* BT-Coex */ 1017 struct iwl_bt_coex_profile_notif last_bt_notif; 1018 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd; 1019 1020 u8 bt_tx_prio; 1021 enum iwl_bt_force_ant_mode bt_force_ant_mode; 1022 1023 /* Aux ROC */ 1024 struct list_head aux_roc_te_list; 1025 1026 /* Thermal Throttling and CTkill */ 1027 struct iwl_mvm_tt_mgmt thermal_throttle; 1028 #ifdef CONFIG_THERMAL 1029 struct iwl_mvm_thermal_device tz_device; 1030 struct iwl_mvm_cooling_device cooling_dev; 1031 #endif 1032 1033 s32 temperature; /* Celsius */ 1034 /* 1035 * Debug option to set the NIC temperature. This option makes the 1036 * driver think this is the actual NIC temperature, and ignore the 1037 * real temperature that is received from the fw 1038 */ 1039 bool temperature_test; /* Debug test temperature is enabled */ 1040 1041 unsigned long bt_coex_last_tcm_ts; 1042 struct iwl_mvm_tcm tcm; 1043 1044 u8 uapsd_noagg_bssid_write_idx; 1045 struct mac_address uapsd_noagg_bssids[IWL_MVM_UAPSD_NOAGG_BSSIDS_NUM] 1046 __aligned(2); 1047 1048 struct iwl_time_quota_cmd last_quota_cmd; 1049 1050 #ifdef CONFIG_NL80211_TESTMODE 1051 u32 noa_duration; 1052 struct ieee80211_vif *noa_vif; 1053 #endif 1054 1055 /* Tx queues */ 1056 u16 aux_queue; 1057 u16 snif_queue; 1058 u16 probe_queue; 1059 u16 p2p_dev_queue; 1060 1061 /* Indicate if device power save is allowed */ 1062 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */ 1063 /* Indicate if 32Khz external clock is valid */ 1064 u32 ext_clock_valid; 1065 1066 struct ieee80211_vif __rcu *csa_vif; 1067 struct ieee80211_vif __rcu *csa_tx_blocked_vif; 1068 u8 csa_tx_block_bcn_timeout; 1069 1070 /* system time of last beacon (for AP/GO interface) */ 1071 u32 ap_last_beacon_gp2; 1072 1073 /* indicates that we transmitted the last beacon */ 1074 bool ibss_manager; 1075 1076 bool lar_regdom_set; 1077 enum iwl_mcc_source mcc_src; 1078 1079 /* TDLS channel switch data */ 1080 struct { 1081 struct delayed_work dwork; 1082 enum iwl_mvm_tdls_cs_state state; 1083 1084 /* 1085 * Current cs sta - might be different from periodic cs peer 1086 * station. Value is meaningless when the cs-state is idle. 1087 */ 1088 u8 cur_sta_id; 1089 1090 /* TDLS periodic channel-switch peer */ 1091 struct { 1092 u8 sta_id; 1093 u8 op_class; 1094 bool initiator; /* are we the link initiator */ 1095 struct cfg80211_chan_def chandef; 1096 struct sk_buff *skb; /* ch sw template */ 1097 u32 ch_sw_tm_ie; 1098 1099 /* timestamp of last ch-sw request sent (GP2 time) */ 1100 u32 sent_timestamp; 1101 } peer; 1102 } tdls_cs; 1103 1104 1105 u32 ciphers[IWL_MVM_NUM_CIPHERS]; 1106 struct ieee80211_cipher_scheme cs[IWL_UCODE_MAX_CS]; 1107 1108 struct cfg80211_ftm_responder_stats ftm_resp_stats; 1109 struct { 1110 struct cfg80211_pmsr_request *req; 1111 struct wireless_dev *req_wdev; 1112 struct list_head loc_list; 1113 int responses[IWL_MVM_TOF_MAX_APS]; 1114 struct { 1115 struct list_head resp; 1116 } smooth; 1117 struct list_head pasn_list; 1118 } ftm_initiator; 1119 1120 struct list_head resp_pasn_list; 1121 1122 struct { 1123 u8 d0i3_resp; 1124 u8 range_resp; 1125 } cmd_ver; 1126 1127 struct ieee80211_vif *nan_vif; 1128 #define IWL_MAX_BAID 32 1129 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID]; 1130 1131 /* 1132 * Drop beacons from other APs in AP mode when there are no connected 1133 * clients. 1134 */ 1135 bool drop_bcn_ap_mode; 1136 1137 struct delayed_work cs_tx_unblock_dwork; 1138 1139 /* does a monitor vif exist (only one can exist hence bool) */ 1140 bool monitor_on; 1141 1142 /* sniffer data to include in radiotap */ 1143 __le16 cur_aid; 1144 u8 cur_bssid[ETH_ALEN]; 1145 }; 1146 1147 /* Extract MVM priv from op_mode and _hw */ 1148 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \ 1149 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific) 1150 1151 #define IWL_MAC80211_GET_MVM(_hw) \ 1152 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv)) 1153 1154 /** 1155 * enum iwl_mvm_status - MVM status bits 1156 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted 1157 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active 1158 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running 1159 * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested 1160 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active 1161 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running 1162 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running 1163 * @IWL_MVM_STATUS_NEED_FLUSH_P2P: need to flush P2P bcast STA 1164 * @IWL_MVM_STATUS_IN_D3: in D3 (or at least about to go into it) 1165 */ 1166 enum iwl_mvm_status { 1167 IWL_MVM_STATUS_HW_RFKILL, 1168 IWL_MVM_STATUS_HW_CTKILL, 1169 IWL_MVM_STATUS_ROC_RUNNING, 1170 IWL_MVM_STATUS_HW_RESTART_REQUESTED, 1171 IWL_MVM_STATUS_IN_HW_RESTART, 1172 IWL_MVM_STATUS_ROC_AUX_RUNNING, 1173 IWL_MVM_STATUS_FIRMWARE_RUNNING, 1174 IWL_MVM_STATUS_NEED_FLUSH_P2P, 1175 IWL_MVM_STATUS_IN_D3, 1176 }; 1177 1178 /* Keep track of completed init configuration */ 1179 enum iwl_mvm_init_status { 1180 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0), 1181 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1), 1182 }; 1183 1184 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm) 1185 { 1186 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) || 1187 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status); 1188 } 1189 1190 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm) 1191 { 1192 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status); 1193 } 1194 1195 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm) 1196 { 1197 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status); 1198 } 1199 1200 /* Must be called with rcu_read_lock() held and it can only be 1201 * released when mvmsta is not needed anymore. 1202 */ 1203 static inline struct iwl_mvm_sta * 1204 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id) 1205 { 1206 struct ieee80211_sta *sta; 1207 1208 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id)) 1209 return NULL; 1210 1211 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]); 1212 1213 /* This can happen if the station has been removed right now */ 1214 if (IS_ERR_OR_NULL(sta)) 1215 return NULL; 1216 1217 return iwl_mvm_sta_from_mac80211(sta); 1218 } 1219 1220 static inline struct iwl_mvm_sta * 1221 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id) 1222 { 1223 struct ieee80211_sta *sta; 1224 1225 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id)) 1226 return NULL; 1227 1228 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id], 1229 lockdep_is_held(&mvm->mutex)); 1230 1231 /* This can happen if the station has been removed right now */ 1232 if (IS_ERR_OR_NULL(sta)) 1233 return NULL; 1234 1235 return iwl_mvm_sta_from_mac80211(sta); 1236 } 1237 1238 static inline struct ieee80211_vif * 1239 iwl_mvm_rcu_dereference_vif_id(struct iwl_mvm *mvm, u8 vif_id, bool rcu) 1240 { 1241 if (WARN_ON(vif_id >= ARRAY_SIZE(mvm->vif_id_to_mac))) 1242 return NULL; 1243 1244 if (rcu) 1245 return rcu_dereference(mvm->vif_id_to_mac[vif_id]); 1246 1247 return rcu_dereference_protected(mvm->vif_id_to_mac[vif_id], 1248 lockdep_is_held(&mvm->mutex)); 1249 } 1250 1251 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm) 1252 { 1253 return fw_has_api(&mvm->fw->ucode_capa, 1254 IWL_UCODE_TLV_API_ADAPTIVE_DWELL); 1255 } 1256 1257 static inline bool iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm *mvm) 1258 { 1259 return fw_has_api(&mvm->fw->ucode_capa, 1260 IWL_UCODE_TLV_API_ADAPTIVE_DWELL_V2); 1261 } 1262 1263 static inline bool iwl_mvm_is_adwell_hb_ap_num_supported(struct iwl_mvm *mvm) 1264 { 1265 return fw_has_api(&mvm->fw->ucode_capa, 1266 IWL_UCODE_TLV_API_ADWELL_HB_DEF_N_AP); 1267 } 1268 1269 static inline bool iwl_mvm_is_oce_supported(struct iwl_mvm *mvm) 1270 { 1271 /* OCE should never be enabled for LMAC scan FWs */ 1272 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_OCE); 1273 } 1274 1275 static inline bool iwl_mvm_is_frag_ebs_supported(struct iwl_mvm *mvm) 1276 { 1277 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAG_EBS); 1278 } 1279 1280 static inline bool iwl_mvm_is_short_beacon_notif_supported(struct iwl_mvm *mvm) 1281 { 1282 return fw_has_api(&mvm->fw->ucode_capa, 1283 IWL_UCODE_TLV_API_SHORT_BEACON_NOTIF); 1284 } 1285 1286 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue) 1287 { 1288 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) && 1289 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE); 1290 } 1291 1292 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue) 1293 { 1294 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) && 1295 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE); 1296 } 1297 1298 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm) 1299 { 1300 bool nvm_lar = mvm->nvm_data->lar_enabled; 1301 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa, 1302 IWL_UCODE_TLV_CAPA_LAR_SUPPORT); 1303 1304 /* 1305 * Enable LAR only if it is supported by the FW (TLV) && 1306 * enabled in the NVM 1307 */ 1308 if (mvm->cfg->nvm_type == IWL_NVM_EXT) 1309 return nvm_lar && tlv_lar; 1310 else 1311 return tlv_lar; 1312 } 1313 1314 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm) 1315 { 1316 return fw_has_api(&mvm->fw->ucode_capa, 1317 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) || 1318 fw_has_capa(&mvm->fw->ucode_capa, 1319 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC); 1320 } 1321 1322 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm) 1323 { 1324 return fw_has_capa(&mvm->fw->ucode_capa, 1325 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) && 1326 IWL_MVM_BT_COEX_RRC; 1327 } 1328 1329 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm) 1330 { 1331 return fw_has_capa(&mvm->fw->ucode_capa, 1332 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) && 1333 !IWL_MVM_HW_CSUM_DISABLE; 1334 } 1335 1336 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm) 1337 { 1338 return fw_has_capa(&mvm->fw->ucode_capa, 1339 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) && 1340 IWL_MVM_BT_COEX_MPLUT; 1341 } 1342 1343 static inline 1344 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm) 1345 { 1346 return fw_has_capa(&mvm->fw->ucode_capa, 1347 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) && 1348 !(iwlwifi_mod_params.uapsd_disable & 1349 IWL_DISABLE_UAPSD_P2P_CLIENT); 1350 } 1351 1352 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm) 1353 { 1354 return fw_has_capa(&mvm->fw->ucode_capa, 1355 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT); 1356 } 1357 1358 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm) 1359 { 1360 /* TODO - replace with TLV once defined */ 1361 return mvm->trans->trans_cfg->use_tfh; 1362 } 1363 1364 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm) 1365 { 1366 /* TODO - better define this */ 1367 return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000; 1368 } 1369 1370 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm) 1371 { 1372 /* 1373 * TODO: 1374 * The issue of how to determine CDB APIs and usage is still not fully 1375 * defined. 1376 * There is a compilation for CDB and non-CDB FW, but there may 1377 * be also runtime check. 1378 * For now there is a TLV for checking compilation mode, but a 1379 * runtime check will also have to be here - once defined. 1380 */ 1381 return fw_has_capa(&mvm->fw->ucode_capa, 1382 IWL_UCODE_TLV_CAPA_CDB_SUPPORT); 1383 } 1384 1385 static inline bool iwl_mvm_cdb_scan_api(struct iwl_mvm *mvm) 1386 { 1387 /* 1388 * TODO: should this be the same as iwl_mvm_is_cdb_supported()? 1389 * but then there's a little bit of code in scan that won't make 1390 * any sense... 1391 */ 1392 return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000; 1393 } 1394 1395 static inline bool iwl_mvm_is_scan_ext_chan_supported(struct iwl_mvm *mvm) 1396 { 1397 return fw_has_api(&mvm->fw->ucode_capa, 1398 IWL_UCODE_TLV_API_SCAN_EXT_CHAN_VER); 1399 } 1400 1401 1402 static inline bool iwl_mvm_is_reduced_config_scan_supported(struct iwl_mvm *mvm) 1403 { 1404 return fw_has_api(&mvm->fw->ucode_capa, 1405 IWL_UCODE_TLV_API_REDUCED_SCAN_CONFIG); 1406 } 1407 1408 static inline bool iwl_mvm_is_band_in_rx_supported(struct iwl_mvm *mvm) 1409 { 1410 return fw_has_api(&mvm->fw->ucode_capa, 1411 IWL_UCODE_TLV_API_BAND_IN_RX_DATA); 1412 } 1413 1414 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm) 1415 { 1416 return fw_has_api(&mvm->fw->ucode_capa, 1417 IWL_UCODE_TLV_API_NEW_RX_STATS); 1418 } 1419 1420 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm) 1421 { 1422 return fw_has_api(&mvm->fw->ucode_capa, 1423 IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY); 1424 } 1425 1426 static inline bool iwl_mvm_has_tlc_offload(const struct iwl_mvm *mvm) 1427 { 1428 return fw_has_capa(&mvm->fw->ucode_capa, 1429 IWL_UCODE_TLV_CAPA_TLC_OFFLOAD); 1430 } 1431 1432 static inline struct agg_tx_status * 1433 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp) 1434 { 1435 if (iwl_mvm_has_new_tx_api(mvm)) 1436 return &((struct iwl_mvm_tx_resp *)tx_resp)->status; 1437 else 1438 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status; 1439 } 1440 1441 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm) 1442 { 1443 /* these two TLV are redundant since the responsibility to CT-kill by 1444 * FW happens only after we send at least one command of 1445 * temperature THs report. 1446 */ 1447 return fw_has_capa(&mvm->fw->ucode_capa, 1448 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) && 1449 fw_has_capa(&mvm->fw->ucode_capa, 1450 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT); 1451 } 1452 1453 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm) 1454 { 1455 return fw_has_capa(&mvm->fw->ucode_capa, 1456 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT); 1457 } 1458 1459 extern const u8 iwl_mvm_ac_to_tx_fifo[]; 1460 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[]; 1461 1462 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm, 1463 enum ieee80211_ac_numbers ac) 1464 { 1465 return iwl_mvm_has_new_tx_api(mvm) ? 1466 iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac]; 1467 } 1468 1469 struct iwl_rate_info { 1470 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */ 1471 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */ 1472 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */ 1473 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */ 1474 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */ 1475 }; 1476 1477 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm); 1478 int __iwl_mvm_mac_start(struct iwl_mvm *mvm); 1479 1480 /****************** 1481 * MVM Methods 1482 ******************/ 1483 /* uCode */ 1484 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm); 1485 1486 /* Utils */ 1487 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags, 1488 enum nl80211_band band); 1489 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags, 1490 enum nl80211_band band, 1491 struct ieee80211_tx_rate *r); 1492 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx); 1493 u8 iwl_mvm_mac80211_ac_to_ucode_ac(enum ieee80211_ac_numbers ac); 1494 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm); 1495 u8 first_antenna(u8 mask); 1496 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx); 1497 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime); 1498 u32 iwl_mvm_get_systime(struct iwl_mvm *mvm); 1499 1500 /* Tx / Host Commands */ 1501 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm, 1502 struct iwl_host_cmd *cmd); 1503 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id, 1504 u32 flags, u16 len, const void *data); 1505 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm, 1506 struct iwl_host_cmd *cmd, 1507 u32 *status); 1508 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id, 1509 u16 len, const void *data, 1510 u32 *status); 1511 int iwl_mvm_tx_skb_sta(struct iwl_mvm *mvm, struct sk_buff *skb, 1512 struct ieee80211_sta *sta); 1513 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb); 1514 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb, 1515 struct iwl_tx_cmd *tx_cmd, 1516 struct ieee80211_tx_info *info, u8 sta_id); 1517 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd, 1518 struct ieee80211_tx_info *info, 1519 struct ieee80211_sta *sta, __le16 fc); 1520 void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq); 1521 unsigned int iwl_mvm_max_amsdu_size(struct iwl_mvm *mvm, 1522 struct ieee80211_sta *sta, 1523 unsigned int tid); 1524 1525 #ifdef CONFIG_IWLWIFI_DEBUG 1526 const char *iwl_mvm_get_tx_fail_reason(u32 status); 1527 #else 1528 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; } 1529 #endif 1530 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags); 1531 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal, u32 flags); 1532 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id, 1533 u16 tids, u32 flags); 1534 1535 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm); 1536 1537 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info, 1538 struct iwl_tx_cmd *tx_cmd) 1539 { 1540 struct ieee80211_key_conf *keyconf = info->control.hw_key; 1541 1542 tx_cmd->sec_ctl = TX_CMD_SEC_CCM; 1543 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen); 1544 } 1545 1546 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm) 1547 { 1548 flush_work(&mvm->async_handlers_wk); 1549 } 1550 1551 /* Statistics */ 1552 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm, 1553 struct iwl_rx_packet *pkt); 1554 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm, 1555 struct iwl_rx_cmd_buffer *rxb); 1556 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear); 1557 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm); 1558 1559 /* NVM */ 1560 int iwl_nvm_init(struct iwl_mvm *mvm); 1561 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm); 1562 1563 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm) 1564 { 1565 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ? 1566 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant : 1567 mvm->fw->valid_tx_ant; 1568 } 1569 1570 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm) 1571 { 1572 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ? 1573 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant : 1574 mvm->fw->valid_rx_ant; 1575 } 1576 1577 static inline void iwl_mvm_toggle_tx_ant(struct iwl_mvm *mvm, u8 *ant) 1578 { 1579 *ant = iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm), *ant); 1580 } 1581 1582 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm) 1583 { 1584 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN | 1585 FW_PHY_CFG_RX_CHAIN); 1586 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm); 1587 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm); 1588 1589 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS | 1590 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS; 1591 1592 return mvm->fw->phy_config & phy_config; 1593 } 1594 1595 int iwl_mvm_up(struct iwl_mvm *mvm); 1596 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm); 1597 1598 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm); 1599 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm, 1600 struct iwl_bcast_filter_cmd *cmd); 1601 1602 /* 1603 * FW notifications / CMD responses handlers 1604 * Convention: iwl_mvm_rx_<NAME OF THE CMD> 1605 */ 1606 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1607 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi, 1608 struct iwl_rx_cmd_buffer *rxb); 1609 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi, 1610 struct iwl_rx_cmd_buffer *rxb, int queue); 1611 void iwl_mvm_rx_monitor_no_data(struct iwl_mvm *mvm, struct napi_struct *napi, 1612 struct iwl_rx_cmd_buffer *rxb, int queue); 1613 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi, 1614 struct iwl_rx_cmd_buffer *rxb, int queue); 1615 void iwl_mvm_rx_bar_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi, 1616 struct iwl_rx_cmd_buffer *rxb, int queue); 1617 int iwl_mvm_notify_rx_queue(struct iwl_mvm *mvm, u32 rxq_mask, 1618 const u8 *data, u32 count, bool async); 1619 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct napi_struct *napi, 1620 struct iwl_rx_cmd_buffer *rxb, int queue); 1621 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1622 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm, 1623 struct iwl_rx_cmd_buffer *rxb); 1624 void iwl_mvm_send_recovery_cmd(struct iwl_mvm *mvm, u32 flags); 1625 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1626 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm, 1627 struct iwl_rx_cmd_buffer *rxb); 1628 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1629 void iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm, 1630 struct iwl_rx_cmd_buffer *rxb); 1631 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm, 1632 struct iwl_rx_cmd_buffer *rxb); 1633 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm, 1634 struct iwl_rx_cmd_buffer *rxb); 1635 1636 /* MVM PHY */ 1637 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt, 1638 struct cfg80211_chan_def *chandef, 1639 u8 chains_static, u8 chains_dynamic); 1640 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt, 1641 struct cfg80211_chan_def *chandef, 1642 u8 chains_static, u8 chains_dynamic); 1643 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm, 1644 struct iwl_mvm_phy_ctxt *ctxt); 1645 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm, 1646 struct iwl_mvm_phy_ctxt *ctxt); 1647 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm); 1648 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef); 1649 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef); 1650 1651 /* MAC (virtual interface) programming */ 1652 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1653 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1654 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1655 bool force_assoc_off, const u8 *bssid_override); 1656 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1657 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm, 1658 struct ieee80211_vif *vif); 1659 int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm, 1660 struct ieee80211_vif *vif, 1661 struct sk_buff *beacon); 1662 int iwl_mvm_mac_ctxt_send_beacon_cmd(struct iwl_mvm *mvm, 1663 struct sk_buff *beacon, 1664 void *data, int len); 1665 u8 iwl_mvm_mac_ctxt_get_lowest_rate(struct ieee80211_tx_info *info, 1666 struct ieee80211_vif *vif); 1667 void iwl_mvm_mac_ctxt_set_tim(struct iwl_mvm *mvm, 1668 __le32 *tim_index, __le32 *tim_size, 1669 u8 *beacon, u32 frame_size); 1670 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm, 1671 struct iwl_rx_cmd_buffer *rxb); 1672 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm, 1673 struct iwl_rx_cmd_buffer *rxb); 1674 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm, 1675 struct iwl_rx_cmd_buffer *rxb); 1676 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm, 1677 struct iwl_rx_cmd_buffer *rxb); 1678 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1679 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm, 1680 struct iwl_rx_cmd_buffer *rxb); 1681 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm, 1682 struct ieee80211_vif *vif); 1683 void iwl_mvm_probe_resp_data_notif(struct iwl_mvm *mvm, 1684 struct iwl_rx_cmd_buffer *rxb); 1685 void iwl_mvm_rx_missed_vap_notif(struct iwl_mvm *mvm, 1686 struct iwl_rx_cmd_buffer *rxb); 1687 void iwl_mvm_channel_switch_noa_notif(struct iwl_mvm *mvm, 1688 struct iwl_rx_cmd_buffer *rxb); 1689 /* Bindings */ 1690 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1691 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1692 1693 /* Quota management */ 1694 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm) 1695 { 1696 return iwl_mvm_has_quota_low_latency(mvm) ? 1697 sizeof(struct iwl_time_quota_cmd) : 1698 sizeof(struct iwl_time_quota_cmd_v1); 1699 } 1700 1701 static inline struct iwl_time_quota_data 1702 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm, 1703 struct iwl_time_quota_cmd *cmd, 1704 int i) 1705 { 1706 struct iwl_time_quota_data_v1 *quotas; 1707 1708 if (iwl_mvm_has_quota_low_latency(mvm)) 1709 return &cmd->quotas[i]; 1710 1711 quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas; 1712 return (struct iwl_time_quota_data *)"as[i]; 1713 } 1714 1715 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload, 1716 struct ieee80211_vif *disabled_vif); 1717 1718 /* Scanning */ 1719 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1720 struct cfg80211_scan_request *req, 1721 struct ieee80211_scan_ies *ies); 1722 int iwl_mvm_scan_size(struct iwl_mvm *mvm); 1723 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify); 1724 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm); 1725 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm); 1726 void iwl_mvm_scan_timeout_wk(struct work_struct *work); 1727 1728 /* Scheduled scan */ 1729 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm, 1730 struct iwl_rx_cmd_buffer *rxb); 1731 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm, 1732 struct iwl_rx_cmd_buffer *rxb); 1733 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm, 1734 struct ieee80211_vif *vif, 1735 struct cfg80211_sched_scan_request *req, 1736 struct ieee80211_scan_ies *ies, 1737 int type); 1738 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm, 1739 struct iwl_rx_cmd_buffer *rxb); 1740 1741 /* UMAC scan */ 1742 int iwl_mvm_config_scan(struct iwl_mvm *mvm); 1743 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm, 1744 struct iwl_rx_cmd_buffer *rxb); 1745 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm, 1746 struct iwl_rx_cmd_buffer *rxb); 1747 1748 /* MVM debugfs */ 1749 #ifdef CONFIG_IWLWIFI_DEBUGFS 1750 void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir); 1751 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1752 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1753 #else 1754 static inline void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, 1755 struct dentry *dbgfs_dir) 1756 { 1757 } 1758 static inline void 1759 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 1760 { 1761 } 1762 static inline void 1763 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 1764 { 1765 } 1766 #endif /* CONFIG_IWLWIFI_DEBUGFS */ 1767 1768 /* rate scaling */ 1769 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq); 1770 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg); 1771 int rs_pretty_print_rate(char *buf, int bufsz, const u32 rate); 1772 void rs_update_last_rssi(struct iwl_mvm *mvm, 1773 struct iwl_mvm_sta *mvmsta, 1774 struct ieee80211_rx_status *rx_status); 1775 1776 /* power management */ 1777 int iwl_mvm_power_update_device(struct iwl_mvm *mvm); 1778 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm); 1779 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm); 1780 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1781 char *buf, int bufsz); 1782 1783 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1784 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm, 1785 struct iwl_rx_cmd_buffer *rxb); 1786 1787 #ifdef CONFIG_IWLWIFI_LEDS 1788 int iwl_mvm_leds_init(struct iwl_mvm *mvm); 1789 void iwl_mvm_leds_exit(struct iwl_mvm *mvm); 1790 void iwl_mvm_leds_sync(struct iwl_mvm *mvm); 1791 #else 1792 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm) 1793 { 1794 return 0; 1795 } 1796 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm) 1797 { 1798 } 1799 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm) 1800 { 1801 } 1802 #endif 1803 1804 /* D3 (WoWLAN, NetDetect) */ 1805 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan); 1806 int iwl_mvm_resume(struct ieee80211_hw *hw); 1807 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled); 1808 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw, 1809 struct ieee80211_vif *vif, 1810 struct cfg80211_gtk_rekey_data *data); 1811 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw, 1812 struct ieee80211_vif *vif, 1813 struct inet6_dev *idev); 1814 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw, 1815 struct ieee80211_vif *vif, int idx); 1816 extern const struct file_operations iwl_dbgfs_d3_test_ops; 1817 struct iwl_wowlan_status *iwl_mvm_send_wowlan_get_status(struct iwl_mvm *mvm); 1818 #ifdef CONFIG_PM 1819 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, 1820 struct ieee80211_vif *vif); 1821 #else 1822 static inline void 1823 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 1824 { 1825 } 1826 #endif 1827 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta, 1828 struct iwl_wowlan_config_cmd *cmd); 1829 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm, 1830 struct ieee80211_vif *vif, 1831 bool disable_offloading, 1832 bool offload_ns, 1833 u32 cmd_flags); 1834 1835 /* BT Coex */ 1836 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm); 1837 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm, 1838 struct iwl_rx_cmd_buffer *rxb); 1839 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1840 enum ieee80211_rssi_event_data); 1841 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm); 1842 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm, 1843 struct ieee80211_sta *sta); 1844 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm, 1845 struct ieee80211_sta *sta); 1846 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant); 1847 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm); 1848 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm, 1849 enum nl80211_band band); 1850 u8 iwl_mvm_bt_coex_get_single_ant_msk(struct iwl_mvm *mvm, u8 enabled_ants); 1851 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr, 1852 struct ieee80211_tx_info *info, u8 ac); 1853 1854 /* beacon filtering */ 1855 #ifdef CONFIG_IWLWIFI_DEBUGFS 1856 void 1857 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif, 1858 struct iwl_beacon_filter_cmd *cmd); 1859 #else 1860 static inline void 1861 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif, 1862 struct iwl_beacon_filter_cmd *cmd) 1863 {} 1864 #endif 1865 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm, 1866 struct ieee80211_vif *vif, 1867 u32 flags); 1868 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm, 1869 struct ieee80211_vif *vif, 1870 u32 flags); 1871 /* SMPS */ 1872 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1873 enum iwl_mvm_smps_type_request req_type, 1874 enum ieee80211_smps_mode smps_request); 1875 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm); 1876 1877 /* Low latency */ 1878 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1879 bool low_latency, 1880 enum iwl_mvm_low_latency_cause cause); 1881 /* get SystemLowLatencyMode - only needed for beacon threshold? */ 1882 bool iwl_mvm_low_latency(struct iwl_mvm *mvm); 1883 bool iwl_mvm_low_latency_band(struct iwl_mvm *mvm, enum nl80211_band band); 1884 void iwl_mvm_send_low_latency_cmd(struct iwl_mvm *mvm, bool low_latency, 1885 u16 mac_id); 1886 1887 /* get VMACLowLatencyMode */ 1888 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif) 1889 { 1890 /* 1891 * should this consider associated/active/... state? 1892 * 1893 * Normally low-latency should only be active on interfaces 1894 * that are active, but at least with debugfs it can also be 1895 * enabled on interfaces that aren't active. However, when 1896 * interface aren't active then they aren't added into the 1897 * binding, so this has no real impact. For now, just return 1898 * the current desired low-latency state. 1899 */ 1900 return mvmvif->low_latency_actual; 1901 } 1902 1903 static inline 1904 void iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif *mvmvif, bool set, 1905 enum iwl_mvm_low_latency_cause cause) 1906 { 1907 u8 new_state; 1908 1909 if (set) 1910 mvmvif->low_latency |= cause; 1911 else 1912 mvmvif->low_latency &= ~cause; 1913 1914 /* 1915 * if LOW_LATENCY_DEBUGFS_FORCE_ENABLE is enabled no changes are 1916 * allowed to actual mode. 1917 */ 1918 if (mvmvif->low_latency & LOW_LATENCY_DEBUGFS_FORCE_ENABLE && 1919 cause != LOW_LATENCY_DEBUGFS_FORCE_ENABLE) 1920 return; 1921 1922 if (cause == LOW_LATENCY_DEBUGFS_FORCE_ENABLE && set) 1923 /* 1924 * We enter force state 1925 */ 1926 new_state = !!(mvmvif->low_latency & 1927 LOW_LATENCY_DEBUGFS_FORCE); 1928 else 1929 /* 1930 * Check if any other one set low latency 1931 */ 1932 new_state = !!(mvmvif->low_latency & 1933 ~(LOW_LATENCY_DEBUGFS_FORCE_ENABLE | 1934 LOW_LATENCY_DEBUGFS_FORCE)); 1935 1936 mvmvif->low_latency_actual = new_state; 1937 } 1938 1939 /* Return a bitmask with all the hw supported queues, except for the 1940 * command queue, which can't be flushed. 1941 */ 1942 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm) 1943 { 1944 return ((BIT(mvm->trans->trans_cfg->base_params->num_of_queues) - 1) & 1945 ~BIT(IWL_MVM_DQA_CMD_QUEUE)); 1946 } 1947 1948 static inline void iwl_mvm_stop_device(struct iwl_mvm *mvm) 1949 { 1950 lockdep_assert_held(&mvm->mutex); 1951 iwl_fw_cancel_timestamp(&mvm->fwrt); 1952 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status); 1953 iwl_fw_dbg_stop_sync(&mvm->fwrt); 1954 iwl_trans_stop_device(mvm->trans); 1955 iwl_free_fw_paging(&mvm->fwrt); 1956 iwl_fw_dump_conf_clear(&mvm->fwrt); 1957 } 1958 1959 /* Re-configure the SCD for a queue that has already been configured */ 1960 int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id, 1961 int tid, int frame_limit, u16 ssn); 1962 1963 /* Thermal management and CT-kill */ 1964 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff); 1965 void iwl_mvm_tt_temp_changed(struct iwl_mvm *mvm, u32 temp); 1966 void iwl_mvm_temp_notif(struct iwl_mvm *mvm, 1967 struct iwl_rx_cmd_buffer *rxb); 1968 void iwl_mvm_tt_handler(struct iwl_mvm *mvm); 1969 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff); 1970 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm); 1971 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state); 1972 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp); 1973 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1974 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm); 1975 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm); 1976 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget); 1977 1978 /* Location Aware Regulatory */ 1979 struct iwl_mcc_update_resp * 1980 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2, 1981 enum iwl_mcc_source src_id); 1982 int iwl_mvm_init_mcc(struct iwl_mvm *mvm); 1983 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm, 1984 struct iwl_rx_cmd_buffer *rxb); 1985 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy, 1986 const char *alpha2, 1987 enum iwl_mcc_source src_id, 1988 bool *changed); 1989 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm, 1990 bool *changed); 1991 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm); 1992 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm); 1993 1994 /* smart fifo */ 1995 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1996 bool added_vif); 1997 1998 /* FTM responder */ 1999 int iwl_mvm_ftm_start_responder(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2000 void iwl_mvm_ftm_restart_responder(struct iwl_mvm *mvm, 2001 struct ieee80211_vif *vif); 2002 void iwl_mvm_ftm_responder_stats(struct iwl_mvm *mvm, 2003 struct iwl_rx_cmd_buffer *rxb); 2004 int iwl_mvm_ftm_resp_remove_pasn_sta(struct iwl_mvm *mvm, 2005 struct ieee80211_vif *vif, u8 *addr); 2006 int iwl_mvm_ftm_respoder_add_pasn_sta(struct iwl_mvm *mvm, 2007 struct ieee80211_vif *vif, 2008 u8 *addr, u32 cipher, u8 *tk, u32 tk_len, 2009 u8 *hltk, u32 hltk_len); 2010 void iwl_mvm_ftm_responder_clear(struct iwl_mvm *mvm, 2011 struct ieee80211_vif *vif); 2012 2013 /* FTM initiator */ 2014 void iwl_mvm_ftm_restart(struct iwl_mvm *mvm); 2015 void iwl_mvm_ftm_range_resp(struct iwl_mvm *mvm, 2016 struct iwl_rx_cmd_buffer *rxb); 2017 void iwl_mvm_ftm_lc_notif(struct iwl_mvm *mvm, 2018 struct iwl_rx_cmd_buffer *rxb); 2019 int iwl_mvm_ftm_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2020 struct cfg80211_pmsr_request *request); 2021 void iwl_mvm_ftm_abort(struct iwl_mvm *mvm, struct cfg80211_pmsr_request *req); 2022 void iwl_mvm_ftm_initiator_smooth_config(struct iwl_mvm *mvm); 2023 void iwl_mvm_ftm_initiator_smooth_stop(struct iwl_mvm *mvm); 2024 int iwl_mvm_ftm_add_pasn_sta(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2025 u8 *addr, u32 cipher, u8 *tk, u32 tk_len, 2026 u8 *hltk, u32 hltk_len); 2027 void iwl_mvm_ftm_remove_pasn_sta(struct iwl_mvm *mvm, u8 *addr); 2028 2029 /* TDLS */ 2030 2031 /* 2032 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present. 2033 * This TID is marked as used vs the AP and all connected TDLS peers. 2034 */ 2035 #define IWL_MVM_TDLS_FW_TID 4 2036 2037 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2038 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm); 2039 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2040 bool sta_added); 2041 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw, 2042 struct ieee80211_vif *vif); 2043 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw, 2044 struct ieee80211_vif *vif, 2045 struct ieee80211_sta *sta, u8 oper_class, 2046 struct cfg80211_chan_def *chandef, 2047 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie); 2048 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw, 2049 struct ieee80211_vif *vif, 2050 struct ieee80211_tdls_ch_sw_params *params); 2051 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw, 2052 struct ieee80211_vif *vif, 2053 struct ieee80211_sta *sta); 2054 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 2055 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work); 2056 2057 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm, 2058 struct iwl_mvm_internal_rxq_notif *notif, 2059 u32 size); 2060 void iwl_mvm_reorder_timer_expired(struct timer_list *t); 2061 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm); 2062 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm); 2063 2064 #define MVM_TCM_PERIOD_MSEC 500 2065 #define MVM_TCM_PERIOD (HZ * MVM_TCM_PERIOD_MSEC / 1000) 2066 #define MVM_LL_PERIOD (10 * HZ) 2067 void iwl_mvm_tcm_work(struct work_struct *work); 2068 void iwl_mvm_recalc_tcm(struct iwl_mvm *mvm); 2069 void iwl_mvm_pause_tcm(struct iwl_mvm *mvm, bool with_cancel); 2070 void iwl_mvm_resume_tcm(struct iwl_mvm *mvm); 2071 void iwl_mvm_tcm_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2072 void iwl_mvm_tcm_rm_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2073 u8 iwl_mvm_tcm_load_percentage(u32 airtime, u32 elapsed); 2074 2075 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error); 2076 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm, 2077 struct ieee80211_vif *vif, 2078 bool tdls, bool cmd_q); 2079 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2080 const char *errmsg); 2081 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm, 2082 struct ieee80211_vif *vif, 2083 const struct ieee80211_sta *sta, 2084 u16 tid); 2085 2086 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b); 2087 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm); 2088 int iwl_mvm_ppag_send_cmd(struct iwl_mvm *mvm); 2089 #ifdef CONFIG_IWLWIFI_DEBUGFS 2090 void iwl_mvm_sta_add_debugfs(struct ieee80211_hw *hw, 2091 struct ieee80211_vif *vif, 2092 struct ieee80211_sta *sta, 2093 struct dentry *dir); 2094 #endif 2095 2096 static inline u8 iwl_mvm_phy_band_from_nl80211(enum nl80211_band band) 2097 { 2098 switch (band) { 2099 case NL80211_BAND_2GHZ: 2100 return PHY_BAND_24; 2101 case NL80211_BAND_5GHZ: 2102 return PHY_BAND_5; 2103 default: 2104 WARN_ONCE(1, "Unsupported band (%u)\n", band); 2105 return PHY_BAND_5; 2106 } 2107 } 2108 2109 /* Channel info utils */ 2110 static inline bool iwl_mvm_has_ultra_hb_channel(struct iwl_mvm *mvm) 2111 { 2112 return fw_has_capa(&mvm->fw->ucode_capa, 2113 IWL_UCODE_TLV_CAPA_ULTRA_HB_CHANNELS); 2114 } 2115 2116 static inline void *iwl_mvm_chan_info_cmd_tail(struct iwl_mvm *mvm, 2117 struct iwl_fw_channel_info *ci) 2118 { 2119 return (u8 *)ci + (iwl_mvm_has_ultra_hb_channel(mvm) ? 2120 sizeof(struct iwl_fw_channel_info) : 2121 sizeof(struct iwl_fw_channel_info_v1)); 2122 } 2123 2124 static inline size_t iwl_mvm_chan_info_padding(struct iwl_mvm *mvm) 2125 { 2126 return iwl_mvm_has_ultra_hb_channel(mvm) ? 0 : 2127 sizeof(struct iwl_fw_channel_info) - 2128 sizeof(struct iwl_fw_channel_info_v1); 2129 } 2130 2131 static inline void iwl_mvm_set_chan_info(struct iwl_mvm *mvm, 2132 struct iwl_fw_channel_info *ci, 2133 u32 chan, u8 band, u8 width, 2134 u8 ctrl_pos) 2135 { 2136 if (iwl_mvm_has_ultra_hb_channel(mvm)) { 2137 ci->channel = cpu_to_le32(chan); 2138 ci->band = band; 2139 ci->width = width; 2140 ci->ctrl_pos = ctrl_pos; 2141 } else { 2142 struct iwl_fw_channel_info_v1 *ci_v1 = 2143 (struct iwl_fw_channel_info_v1 *)ci; 2144 2145 ci_v1->channel = chan; 2146 ci_v1->band = band; 2147 ci_v1->width = width; 2148 ci_v1->ctrl_pos = ctrl_pos; 2149 } 2150 } 2151 2152 static inline void 2153 iwl_mvm_set_chan_info_chandef(struct iwl_mvm *mvm, 2154 struct iwl_fw_channel_info *ci, 2155 struct cfg80211_chan_def *chandef) 2156 { 2157 enum nl80211_band band = chandef->chan->band; 2158 2159 iwl_mvm_set_chan_info(mvm, ci, chandef->chan->hw_value, 2160 iwl_mvm_phy_band_from_nl80211(band), 2161 iwl_mvm_get_channel_width(chandef), 2162 iwl_mvm_get_ctrl_pos(chandef)); 2163 } 2164 2165 static inline int iwl_umac_scan_get_max_profiles(const struct iwl_fw *fw) 2166 { 2167 u8 ver = iwl_fw_lookup_cmd_ver(fw, IWL_ALWAYS_LONG_GROUP, 2168 SCAN_OFFLOAD_UPDATE_PROFILES_CMD, 2169 IWL_FW_CMD_VER_UNKNOWN); 2170 return (ver == IWL_FW_CMD_VER_UNKNOWN || ver < 3) ? 2171 IWL_SCAN_MAX_PROFILES : IWL_SCAN_MAX_PROFILES_V2; 2172 } 2173 2174 static inline 2175 enum iwl_location_cipher iwl_mvm_cipher_to_location_cipher(u32 cipher) 2176 { 2177 switch (cipher) { 2178 case WLAN_CIPHER_SUITE_CCMP: 2179 return IWL_LOCATION_CIPHER_CCMP_128; 2180 case WLAN_CIPHER_SUITE_GCMP: 2181 return IWL_LOCATION_CIPHER_GCMP_128; 2182 case WLAN_CIPHER_SUITE_GCMP_256: 2183 return IWL_LOCATION_CIPHER_GCMP_256; 2184 default: 2185 return IWL_LOCATION_CIPHER_INVALID; 2186 } 2187 } 2188 #endif /* __IWL_MVM_H__ */ 2189