1 /****************************************************************************** 2 * 3 * This file is provided under a dual BSD/GPLv2 license. When using or 4 * redistributing this file, you may do so under either license. 5 * 6 * GPL LICENSE SUMMARY 7 * 8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. 9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH 11 * Copyright(c) 2018 - 2019 Intel Corporation 12 * 13 * This program is free software; you can redistribute it and/or modify 14 * it under the terms of version 2 of the GNU General Public License as 15 * published by the Free Software Foundation. 16 * 17 * This program is distributed in the hope that it will be useful, but 18 * WITHOUT ANY WARRANTY; without even the implied warranty of 19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 20 * General Public License for more details. 21 * 22 * The full GNU General Public License is included in this distribution 23 * in the file called COPYING. 24 * 25 * Contact Information: 26 * Intel Linux Wireless <linuxwifi@intel.com> 27 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497 28 * 29 * BSD LICENSE 30 * 31 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved. 32 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH 33 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH 34 * Copyright(c) 2018 - 2019 Intel Corporation 35 * All rights reserved. 36 * 37 * Redistribution and use in source and binary forms, with or without 38 * modification, are permitted provided that the following conditions 39 * are met: 40 * 41 * * Redistributions of source code must retain the above copyright 42 * notice, this list of conditions and the following disclaimer. 43 * * Redistributions in binary form must reproduce the above copyright 44 * notice, this list of conditions and the following disclaimer in 45 * the documentation and/or other materials provided with the 46 * distribution. 47 * * Neither the name Intel Corporation nor the names of its 48 * contributors may be used to endorse or promote products derived 49 * from this software without specific prior written permission. 50 * 51 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 52 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 53 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 54 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 55 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 56 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 57 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 58 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 59 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 60 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 61 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 62 * 63 *****************************************************************************/ 64 65 #ifndef __IWL_MVM_H__ 66 #define __IWL_MVM_H__ 67 68 #include <linux/list.h> 69 #include <linux/spinlock.h> 70 #include <linux/leds.h> 71 #include <linux/in6.h> 72 73 #ifdef CONFIG_THERMAL 74 #include <linux/thermal.h> 75 #endif 76 77 #include "iwl-op-mode.h" 78 #include "iwl-trans.h" 79 #include "fw/notif-wait.h" 80 #include "iwl-eeprom-parse.h" 81 #include "fw/file.h" 82 #include "iwl-config.h" 83 #include "sta.h" 84 #include "fw-api.h" 85 #include "constants.h" 86 #include "fw/runtime.h" 87 #include "fw/dbg.h" 88 #include "fw/acpi.h" 89 #include "iwl-nvm-parse.h" 90 91 #include <linux/average.h> 92 93 #define IWL_MVM_MAX_ADDRESSES 5 94 /* RSSI offset for WkP */ 95 #define IWL_RSSI_OFFSET 50 96 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8 97 #define IWL_MVM_MISSED_BEACONS_THRESHOLD_LONG 16 98 99 /* A TimeUnit is 1024 microsecond */ 100 #define MSEC_TO_TU(_msec) (_msec*1000/1024) 101 102 /* For GO, this value represents the number of TUs before CSA "beacon 103 * 0" TBTT when the CSA time-event needs to be scheduled to start. It 104 * must be big enough to ensure that we switch in time. 105 */ 106 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO 40 107 108 /* For client, this value represents the number of TUs before CSA 109 * "beacon 1" TBTT, instead. This is because we don't know when the 110 * GO/AP will be in the new channel, so we switch early enough. 111 */ 112 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT 10 113 114 /* 115 * This value (in TUs) is used to fine tune the CSA NoA end time which should 116 * be just before "beacon 0" TBTT. 117 */ 118 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4 119 120 /* 121 * Number of beacons to transmit on a new channel until we unblock tx to 122 * the stations, even if we didn't identify them on a new channel 123 */ 124 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3 125 126 /* offchannel queue towards mac80211 */ 127 #define IWL_MVM_OFFCHANNEL_QUEUE 0 128 129 extern const struct ieee80211_ops iwl_mvm_hw_ops; 130 131 /** 132 * struct iwl_mvm_mod_params - module parameters for iwlmvm 133 * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted. 134 * We will register to mac80211 to have testmode working. The NIC must not 135 * be up'ed after the INIT fw asserted. This is useful to be able to use 136 * proprietary tools over testmode to debug the INIT fw. 137 * @tfd_q_hang_detect: enabled the detection of hung transmit queues 138 * @power_scheme: one of enum iwl_power_scheme 139 */ 140 struct iwl_mvm_mod_params { 141 bool init_dbg; 142 bool tfd_q_hang_detect; 143 int power_scheme; 144 }; 145 extern struct iwl_mvm_mod_params iwlmvm_mod_params; 146 147 struct iwl_mvm_phy_ctxt { 148 u16 id; 149 u16 color; 150 u32 ref; 151 152 enum nl80211_chan_width width; 153 154 /* 155 * TODO: This should probably be removed. Currently here only for rate 156 * scaling algorithm 157 */ 158 struct ieee80211_channel *channel; 159 }; 160 161 struct iwl_mvm_time_event_data { 162 struct ieee80211_vif *vif; 163 struct list_head list; 164 unsigned long end_jiffies; 165 u32 duration; 166 bool running; 167 u32 uid; 168 169 /* 170 * The access to the 'id' field must be done when the 171 * mvm->time_event_lock is held, as it value is used to indicate 172 * if the te is in the time event list or not (when id == TE_MAX) 173 */ 174 u32 id; 175 }; 176 177 /* Power management */ 178 179 /** 180 * enum iwl_power_scheme 181 * @IWL_POWER_LEVEL_CAM - Continuously Active Mode 182 * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default) 183 * @IWL_POWER_LEVEL_LP - Low Power 184 */ 185 enum iwl_power_scheme { 186 IWL_POWER_SCHEME_CAM = 1, 187 IWL_POWER_SCHEME_BPS, 188 IWL_POWER_SCHEME_LP 189 }; 190 191 #define IWL_CONN_MAX_LISTEN_INTERVAL 10 192 #define IWL_UAPSD_MAX_SP IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL 193 194 #ifdef CONFIG_IWLWIFI_DEBUGFS 195 enum iwl_dbgfs_pm_mask { 196 MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0), 197 MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1), 198 MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2), 199 MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3), 200 MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4), 201 MVM_DEBUGFS_PM_LPRX_ENA = BIT(6), 202 MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7), 203 MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8), 204 MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9), 205 MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10), 206 }; 207 208 struct iwl_dbgfs_pm { 209 u16 keep_alive_seconds; 210 u32 rx_data_timeout; 211 u32 tx_data_timeout; 212 bool skip_over_dtim; 213 u8 skip_dtim_periods; 214 bool lprx_ena; 215 u32 lprx_rssi_threshold; 216 bool snooze_ena; 217 bool uapsd_misbehaving; 218 bool use_ps_poll; 219 int mask; 220 }; 221 222 /* beacon filtering */ 223 224 enum iwl_dbgfs_bf_mask { 225 MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0), 226 MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1), 227 MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2), 228 MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3), 229 MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4), 230 MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5), 231 MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6), 232 MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7), 233 MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8), 234 MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9), 235 MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10), 236 }; 237 238 struct iwl_dbgfs_bf { 239 u32 bf_energy_delta; 240 u32 bf_roaming_energy_delta; 241 u32 bf_roaming_state; 242 u32 bf_temp_threshold; 243 u32 bf_temp_fast_filter; 244 u32 bf_temp_slow_filter; 245 u32 bf_enable_beacon_filter; 246 u32 bf_debug_flag; 247 u32 bf_escape_timer; 248 u32 ba_escape_timer; 249 u32 ba_enable_beacon_abort; 250 int mask; 251 }; 252 #endif 253 254 enum iwl_mvm_smps_type_request { 255 IWL_MVM_SMPS_REQ_BT_COEX, 256 IWL_MVM_SMPS_REQ_TT, 257 IWL_MVM_SMPS_REQ_PROT, 258 NUM_IWL_MVM_SMPS_REQ, 259 }; 260 261 enum iwl_mvm_ref_type { 262 IWL_MVM_REF_UCODE_DOWN, 263 IWL_MVM_REF_SCAN, 264 IWL_MVM_REF_ROC, 265 IWL_MVM_REF_ROC_AUX, 266 IWL_MVM_REF_P2P_CLIENT, 267 IWL_MVM_REF_AP_IBSS, 268 IWL_MVM_REF_USER, 269 IWL_MVM_REF_TX, 270 IWL_MVM_REF_TX_AGG, 271 IWL_MVM_REF_ADD_IF, 272 IWL_MVM_REF_START_AP, 273 IWL_MVM_REF_BSS_CHANGED, 274 IWL_MVM_REF_PREPARE_TX, 275 IWL_MVM_REF_PROTECT_TDLS, 276 IWL_MVM_REF_CHECK_CTKILL, 277 IWL_MVM_REF_PRPH_READ, 278 IWL_MVM_REF_PRPH_WRITE, 279 IWL_MVM_REF_NMI, 280 IWL_MVM_REF_TM_CMD, 281 IWL_MVM_REF_EXIT_WORK, 282 IWL_MVM_REF_PROTECT_CSA, 283 IWL_MVM_REF_FW_DBG_COLLECT, 284 IWL_MVM_REF_INIT_UCODE, 285 IWL_MVM_REF_SENDING_CMD, 286 IWL_MVM_REF_RX, 287 288 /* update debugfs.c when changing this */ 289 290 IWL_MVM_REF_COUNT, 291 }; 292 293 enum iwl_bt_force_ant_mode { 294 BT_FORCE_ANT_DIS = 0, 295 BT_FORCE_ANT_AUTO, 296 BT_FORCE_ANT_BT, 297 BT_FORCE_ANT_WIFI, 298 299 BT_FORCE_ANT_MAX, 300 }; 301 302 /** 303 * struct iwl_mvm_low_latency_force - low latency force mode set by debugfs 304 * @LOW_LATENCY_FORCE_UNSET: unset force mode 305 * @LOW_LATENCY_FORCE_ON: for low latency on 306 * @LOW_LATENCY_FORCE_OFF: for low latency off 307 * @NUM_LOW_LATENCY_FORCE: max num of modes 308 */ 309 enum iwl_mvm_low_latency_force { 310 LOW_LATENCY_FORCE_UNSET, 311 LOW_LATENCY_FORCE_ON, 312 LOW_LATENCY_FORCE_OFF, 313 NUM_LOW_LATENCY_FORCE 314 }; 315 316 /** 317 * struct iwl_mvm_low_latency_cause - low latency set causes 318 * @LOW_LATENCY_TRAFFIC: indicates low latency traffic was detected 319 * @LOW_LATENCY_DEBUGFS: low latency mode set from debugfs 320 * @LOW_LATENCY_VCMD: low latency mode set from vendor command 321 * @LOW_LATENCY_VIF_TYPE: low latency mode set because of vif type (ap) 322 * @LOW_LATENCY_DEBUGFS_FORCE_ENABLE: indicate that force mode is enabled 323 * the actual set/unset is done with LOW_LATENCY_DEBUGFS_FORCE 324 * @LOW_LATENCY_DEBUGFS_FORCE: low latency force mode from debugfs 325 * set this with LOW_LATENCY_DEBUGFS_FORCE_ENABLE flag 326 * in low_latency. 327 */ 328 enum iwl_mvm_low_latency_cause { 329 LOW_LATENCY_TRAFFIC = BIT(0), 330 LOW_LATENCY_DEBUGFS = BIT(1), 331 LOW_LATENCY_VCMD = BIT(2), 332 LOW_LATENCY_VIF_TYPE = BIT(3), 333 LOW_LATENCY_DEBUGFS_FORCE_ENABLE = BIT(4), 334 LOW_LATENCY_DEBUGFS_FORCE = BIT(5), 335 }; 336 337 /** 338 * struct iwl_mvm_vif_bf_data - beacon filtering related data 339 * @bf_enabled: indicates if beacon filtering is enabled 340 * @ba_enabled: indicated if beacon abort is enabled 341 * @ave_beacon_signal: average beacon signal 342 * @last_cqm_event: rssi of the last cqm event 343 * @bt_coex_min_thold: minimum threshold for BT coex 344 * @bt_coex_max_thold: maximum threshold for BT coex 345 * @last_bt_coex_event: rssi of the last BT coex event 346 */ 347 struct iwl_mvm_vif_bf_data { 348 bool bf_enabled; 349 bool ba_enabled; 350 int ave_beacon_signal; 351 int last_cqm_event; 352 int bt_coex_min_thold; 353 int bt_coex_max_thold; 354 int last_bt_coex_event; 355 }; 356 357 /** 358 * struct iwl_probe_resp_data - data for NoA/CSA updates 359 * @rcu_head: used for freeing the data on update 360 * @notif: notification data 361 * @noa_len: length of NoA attribute, calculated from the notification 362 */ 363 struct iwl_probe_resp_data { 364 struct rcu_head rcu_head; 365 struct iwl_probe_resp_data_notif notif; 366 int noa_len; 367 }; 368 369 /** 370 * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context 371 * @id: between 0 and 3 372 * @color: to solve races upon MAC addition and removal 373 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA 374 * @bssid: BSSID for this (client) interface 375 * @associated: indicates that we're currently associated, used only for 376 * managing the firmware state in iwl_mvm_bss_info_changed_station() 377 * @ap_assoc_sta_count: count of stations associated to us - valid only 378 * if VIF type is AP 379 * @uploaded: indicates the MAC context has been added to the device 380 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface 381 * should get quota etc. 382 * @pm_enabled - Indicate if MAC power management is allowed 383 * @monitor_active: indicates that monitor context is configured, and that the 384 * interface should get quota etc. 385 * @low_latency: bit flags for low latency 386 * see enum &iwl_mvm_low_latency_cause for causes. 387 * @low_latency_actual: boolean, indicates low latency is set, 388 * as a result from low_latency bit flags and takes force into account. 389 * @ps_disabled: indicates that this interface requires PS to be disabled 390 * @queue_params: QoS params for this MAC 391 * @bcast_sta: station used for broadcast packets. Used by the following 392 * vifs: P2P_DEVICE, GO and AP. 393 * @beacon_skb: the skb used to hold the AP/GO beacon template 394 * @smps_requests: the SMPS requests of different parts of the driver, 395 * combined on update to yield the overall request to mac80211. 396 * @beacon_stats: beacon statistics, containing the # of received beacons, 397 * # of received beacons accumulated over FW restart, and the current 398 * average signal of beacons retrieved from the firmware 399 * @csa_failed: CSA failed to schedule time event, report an error later 400 * @features: hw features active for this vif 401 * @probe_resp_data: data from FW notification to store NOA and CSA related 402 * data to be inserted into probe response. 403 */ 404 struct iwl_mvm_vif { 405 struct iwl_mvm *mvm; 406 u16 id; 407 u16 color; 408 u8 ap_sta_id; 409 410 u8 bssid[ETH_ALEN]; 411 bool associated; 412 u8 ap_assoc_sta_count; 413 414 u16 cab_queue; 415 416 bool uploaded; 417 bool ap_ibss_active; 418 bool pm_enabled; 419 bool monitor_active; 420 u8 low_latency: 6; 421 u8 low_latency_actual: 1; 422 bool ps_disabled; 423 struct iwl_mvm_vif_bf_data bf_data; 424 425 struct { 426 u32 num_beacons, accu_num_beacons; 427 u8 avg_signal; 428 } beacon_stats; 429 430 u32 ap_beacon_time; 431 432 enum iwl_tsf_id tsf_id; 433 434 /* 435 * QoS data from mac80211, need to store this here 436 * as mac80211 has a separate callback but we need 437 * to have the data for the MAC context 438 */ 439 struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS]; 440 struct iwl_mvm_time_event_data time_event_data; 441 struct iwl_mvm_time_event_data hs_time_event_data; 442 443 struct iwl_mvm_int_sta bcast_sta; 444 struct iwl_mvm_int_sta mcast_sta; 445 446 /* 447 * Assigned while mac80211 has the interface in a channel context, 448 * or, for P2P Device, while it exists. 449 */ 450 struct iwl_mvm_phy_ctxt *phy_ctxt; 451 452 #ifdef CONFIG_PM 453 /* WoWLAN GTK rekey data */ 454 struct { 455 u8 kck[NL80211_KCK_LEN], kek[NL80211_KEK_LEN]; 456 __le64 replay_ctr; 457 bool valid; 458 } rekey_data; 459 460 int tx_key_idx; 461 462 bool seqno_valid; 463 u16 seqno; 464 #endif 465 466 #if IS_ENABLED(CONFIG_IPV6) 467 /* IPv6 addresses for WoWLAN */ 468 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX]; 469 unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)]; 470 int num_target_ipv6_addrs; 471 #endif 472 473 #ifdef CONFIG_IWLWIFI_DEBUGFS 474 struct dentry *dbgfs_dir; 475 struct dentry *dbgfs_slink; 476 struct iwl_dbgfs_pm dbgfs_pm; 477 struct iwl_dbgfs_bf dbgfs_bf; 478 struct iwl_mac_power_cmd mac_pwr_cmd; 479 int dbgfs_quota_min; 480 #endif 481 482 enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ]; 483 484 /* FW identified misbehaving AP */ 485 u8 uapsd_misbehaving_bssid[ETH_ALEN]; 486 487 struct delayed_work uapsd_nonagg_detected_wk; 488 489 /* Indicates that CSA countdown may be started */ 490 bool csa_countdown; 491 bool csa_failed; 492 u16 csa_target_freq; 493 u16 csa_count; 494 u16 csa_misbehave; 495 struct delayed_work csa_work; 496 497 /* Indicates that we are waiting for a beacon on a new channel */ 498 bool csa_bcn_pending; 499 500 /* TCP Checksum Offload */ 501 netdev_features_t features; 502 503 struct iwl_probe_resp_data __rcu *probe_resp_data; 504 }; 505 506 static inline struct iwl_mvm_vif * 507 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif) 508 { 509 if (!vif) 510 return NULL; 511 return (void *)vif->drv_priv; 512 } 513 514 extern const u8 tid_to_mac80211_ac[]; 515 516 #define IWL_MVM_SCAN_STOPPING_SHIFT 8 517 518 enum iwl_scan_status { 519 IWL_MVM_SCAN_REGULAR = BIT(0), 520 IWL_MVM_SCAN_SCHED = BIT(1), 521 IWL_MVM_SCAN_NETDETECT = BIT(2), 522 523 IWL_MVM_SCAN_STOPPING_REGULAR = BIT(8), 524 IWL_MVM_SCAN_STOPPING_SCHED = BIT(9), 525 IWL_MVM_SCAN_STOPPING_NETDETECT = BIT(10), 526 527 IWL_MVM_SCAN_REGULAR_MASK = IWL_MVM_SCAN_REGULAR | 528 IWL_MVM_SCAN_STOPPING_REGULAR, 529 IWL_MVM_SCAN_SCHED_MASK = IWL_MVM_SCAN_SCHED | 530 IWL_MVM_SCAN_STOPPING_SCHED, 531 IWL_MVM_SCAN_NETDETECT_MASK = IWL_MVM_SCAN_NETDETECT | 532 IWL_MVM_SCAN_STOPPING_NETDETECT, 533 534 IWL_MVM_SCAN_STOPPING_MASK = 0xff << IWL_MVM_SCAN_STOPPING_SHIFT, 535 IWL_MVM_SCAN_MASK = 0xff, 536 }; 537 538 enum iwl_mvm_scan_type { 539 IWL_SCAN_TYPE_NOT_SET, 540 IWL_SCAN_TYPE_UNASSOC, 541 IWL_SCAN_TYPE_WILD, 542 IWL_SCAN_TYPE_MILD, 543 IWL_SCAN_TYPE_FRAGMENTED, 544 IWL_SCAN_TYPE_FAST_BALANCE, 545 }; 546 547 enum iwl_mvm_sched_scan_pass_all_states { 548 SCHED_SCAN_PASS_ALL_DISABLED, 549 SCHED_SCAN_PASS_ALL_ENABLED, 550 SCHED_SCAN_PASS_ALL_FOUND, 551 }; 552 553 /** 554 * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure 555 * @ct_kill_exit: worker to exit thermal kill 556 * @dynamic_smps: Is thermal throttling enabled dynamic_smps? 557 * @tx_backoff: The current thremal throttling tx backoff in uSec. 558 * @min_backoff: The minimal tx backoff due to power restrictions 559 * @params: Parameters to configure the thermal throttling algorithm. 560 * @throttle: Is thermal throttling is active? 561 */ 562 struct iwl_mvm_tt_mgmt { 563 struct delayed_work ct_kill_exit; 564 bool dynamic_smps; 565 u32 tx_backoff; 566 u32 min_backoff; 567 struct iwl_tt_params params; 568 bool throttle; 569 }; 570 571 #ifdef CONFIG_THERMAL 572 /** 573 *struct iwl_mvm_thermal_device - thermal zone related data 574 * @temp_trips: temperature thresholds for report 575 * @fw_trips_index: keep indexes to original array - temp_trips 576 * @tzone: thermal zone device data 577 */ 578 struct iwl_mvm_thermal_device { 579 s16 temp_trips[IWL_MAX_DTS_TRIPS]; 580 u8 fw_trips_index[IWL_MAX_DTS_TRIPS]; 581 struct thermal_zone_device *tzone; 582 }; 583 584 /* 585 * struct iwl_mvm_cooling_device 586 * @cur_state: current state 587 * @cdev: struct thermal cooling device 588 */ 589 struct iwl_mvm_cooling_device { 590 u32 cur_state; 591 struct thermal_cooling_device *cdev; 592 }; 593 #endif 594 595 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8 596 597 struct iwl_mvm_frame_stats { 598 u32 legacy_frames; 599 u32 ht_frames; 600 u32 vht_frames; 601 u32 bw_20_frames; 602 u32 bw_40_frames; 603 u32 bw_80_frames; 604 u32 bw_160_frames; 605 u32 sgi_frames; 606 u32 ngi_frames; 607 u32 siso_frames; 608 u32 mimo2_frames; 609 u32 agg_frames; 610 u32 ampdu_count; 611 u32 success_frames; 612 u32 fail_frames; 613 u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES]; 614 int last_frame_idx; 615 }; 616 617 enum { 618 D0I3_DEFER_WAKEUP, 619 D0I3_PENDING_WAKEUP, 620 }; 621 622 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff 623 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100 624 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200 625 626 enum iwl_mvm_tdls_cs_state { 627 IWL_MVM_TDLS_SW_IDLE = 0, 628 IWL_MVM_TDLS_SW_REQ_SENT, 629 IWL_MVM_TDLS_SW_RESP_RCVD, 630 IWL_MVM_TDLS_SW_REQ_RCVD, 631 IWL_MVM_TDLS_SW_ACTIVE, 632 }; 633 634 enum iwl_mvm_traffic_load { 635 IWL_MVM_TRAFFIC_LOW, 636 IWL_MVM_TRAFFIC_MEDIUM, 637 IWL_MVM_TRAFFIC_HIGH, 638 }; 639 640 DECLARE_EWMA(rate, 16, 16) 641 642 struct iwl_mvm_tcm_mac { 643 struct { 644 u32 pkts[IEEE80211_NUM_ACS]; 645 u32 airtime; 646 } tx; 647 struct { 648 u32 pkts[IEEE80211_NUM_ACS]; 649 u32 airtime; 650 u32 last_ampdu_ref; 651 } rx; 652 struct { 653 /* track AP's transfer in client mode */ 654 u64 rx_bytes; 655 struct ewma_rate rate; 656 bool detected; 657 } uapsd_nonagg_detect; 658 bool opened_rx_ba_sessions; 659 }; 660 661 struct iwl_mvm_tcm { 662 struct delayed_work work; 663 spinlock_t lock; /* used when time elapsed */ 664 unsigned long ts; /* timestamp when period ends */ 665 unsigned long ll_ts; 666 unsigned long uapsd_nonagg_ts; 667 bool paused; 668 struct iwl_mvm_tcm_mac data[NUM_MAC_INDEX_DRIVER]; 669 struct { 670 u32 elapsed; /* milliseconds for this TCM period */ 671 u32 airtime[NUM_MAC_INDEX_DRIVER]; 672 enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER]; 673 enum iwl_mvm_traffic_load band_load[NUM_NL80211_BANDS]; 674 enum iwl_mvm_traffic_load global_load; 675 bool low_latency[NUM_MAC_INDEX_DRIVER]; 676 bool change[NUM_MAC_INDEX_DRIVER]; 677 bool global_change; 678 } result; 679 }; 680 681 /** 682 * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer 683 * @head_sn: reorder window head sn 684 * @num_stored: number of mpdus stored in the buffer 685 * @buf_size: the reorder buffer size as set by the last addba request 686 * @queue: queue of this reorder buffer 687 * @last_amsdu: track last ASMDU SN for duplication detection 688 * @last_sub_index: track ASMDU sub frame index for duplication detection 689 * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU 690 * it is the time of last received sub-frame 691 * @removed: prevent timer re-arming 692 * @valid: reordering is valid for this queue 693 * @lock: protect reorder buffer internal state 694 * @mvm: mvm pointer, needed for frame timer context 695 */ 696 struct iwl_mvm_reorder_buffer { 697 u16 head_sn; 698 u16 num_stored; 699 u16 buf_size; 700 int queue; 701 u16 last_amsdu; 702 u8 last_sub_index; 703 struct timer_list reorder_timer; 704 bool removed; 705 bool valid; 706 spinlock_t lock; 707 struct iwl_mvm *mvm; 708 } ____cacheline_aligned_in_smp; 709 710 /** 711 * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno 712 * @frames: list of skbs stored 713 * @reorder_time: time the packet was stored in the reorder buffer 714 */ 715 struct _iwl_mvm_reorder_buf_entry { 716 struct sk_buff_head frames; 717 unsigned long reorder_time; 718 }; 719 720 /* make this indirection to get the aligned thing */ 721 struct iwl_mvm_reorder_buf_entry { 722 struct _iwl_mvm_reorder_buf_entry e; 723 } 724 #ifndef __CHECKER__ 725 /* sparse doesn't like this construct: "bad integer constant expression" */ 726 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry))) 727 #endif 728 ; 729 730 /** 731 * struct iwl_mvm_baid_data - BA session data 732 * @sta_id: station id 733 * @tid: tid of the session 734 * @baid baid of the session 735 * @timeout: the timeout set in the addba request 736 * @entries_per_queue: # of buffers per queue, this actually gets 737 * aligned up to avoid cache line sharing between queues 738 * @last_rx: last rx jiffies, updated only if timeout passed from last update 739 * @session_timer: timer to check if BA session expired, runs at 2 * timeout 740 * @mvm: mvm pointer, needed for timer context 741 * @reorder_buf: reorder buffer, allocated per queue 742 * @reorder_buf_data: data 743 */ 744 struct iwl_mvm_baid_data { 745 struct rcu_head rcu_head; 746 u8 sta_id; 747 u8 tid; 748 u8 baid; 749 u16 timeout; 750 u16 entries_per_queue; 751 unsigned long last_rx; 752 struct timer_list session_timer; 753 struct iwl_mvm_baid_data __rcu **rcu_ptr; 754 struct iwl_mvm *mvm; 755 struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES]; 756 struct iwl_mvm_reorder_buf_entry entries[]; 757 }; 758 759 static inline struct iwl_mvm_baid_data * 760 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf) 761 { 762 return (void *)((u8 *)buf - 763 offsetof(struct iwl_mvm_baid_data, reorder_buf) - 764 sizeof(*buf) * buf->queue); 765 } 766 767 /* 768 * enum iwl_mvm_queue_status - queue status 769 * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved 770 * Basically, this means that this queue can be used for any purpose 771 * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use 772 * This is the state of a queue that has been dedicated for some RATID 773 * (agg'd or not), but that hasn't yet gone through the actual enablement 774 * of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet. 775 * Note that in this state there is no requirement to already know what TID 776 * should be used with this queue, it is just marked as a queue that will 777 * be used, and shouldn't be allocated to anyone else. 778 * @IWL_MVM_QUEUE_READY: queue is ready to be used 779 * This is the state of a queue that has been fully configured (including 780 * SCD pointers, etc), has a specific RA/TID assigned to it, and can be 781 * used to send traffic. 782 * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared 783 * This is a state in which a single queue serves more than one TID, all of 784 * which are not aggregated. Note that the queue is only associated to one 785 * RA. 786 */ 787 enum iwl_mvm_queue_status { 788 IWL_MVM_QUEUE_FREE, 789 IWL_MVM_QUEUE_RESERVED, 790 IWL_MVM_QUEUE_READY, 791 IWL_MVM_QUEUE_SHARED, 792 }; 793 794 #define IWL_MVM_DQA_QUEUE_TIMEOUT (5 * HZ) 795 #define IWL_MVM_INVALID_QUEUE 0xFFFF 796 797 #define IWL_MVM_NUM_CIPHERS 10 798 799 struct iwl_mvm_sar_profile { 800 bool enabled; 801 u8 table[ACPI_SAR_TABLE_SIZE]; 802 }; 803 804 struct iwl_mvm_geo_profile { 805 u8 values[ACPI_GEO_TABLE_SIZE]; 806 }; 807 808 struct iwl_mvm_txq { 809 struct list_head list; 810 u16 txq_id; 811 atomic_t tx_request; 812 bool stopped; 813 }; 814 815 static inline struct iwl_mvm_txq * 816 iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq) 817 { 818 return (void *)txq->drv_priv; 819 } 820 821 static inline struct iwl_mvm_txq * 822 iwl_mvm_txq_from_tid(struct ieee80211_sta *sta, u8 tid) 823 { 824 if (tid == IWL_MAX_TID_COUNT) 825 tid = IEEE80211_NUM_TIDS; 826 827 return (void *)sta->txq[tid]->drv_priv; 828 } 829 830 /** 831 * struct iwl_mvm_tvqm_txq_info - maps TVQM hw queue to tid 832 * 833 * @sta_id: sta id 834 * @txq_tid: txq tid 835 */ 836 struct iwl_mvm_tvqm_txq_info { 837 u8 sta_id; 838 u8 txq_tid; 839 }; 840 841 struct iwl_mvm_dqa_txq_info { 842 u8 ra_sta_id; /* The RA this queue is mapped to, if exists */ 843 bool reserved; /* Is this the TXQ reserved for a STA */ 844 u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */ 845 u8 txq_tid; /* The TID "owner" of this queue*/ 846 u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */ 847 /* Timestamp for inactivation per TID of this queue */ 848 unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1]; 849 enum iwl_mvm_queue_status status; 850 }; 851 852 struct iwl_mvm { 853 /* for logger access */ 854 struct device *dev; 855 856 struct iwl_trans *trans; 857 const struct iwl_fw *fw; 858 const struct iwl_cfg *cfg; 859 struct iwl_phy_db *phy_db; 860 struct ieee80211_hw *hw; 861 862 /* for protecting access to iwl_mvm */ 863 struct mutex mutex; 864 struct list_head async_handlers_list; 865 spinlock_t async_handlers_lock; 866 struct work_struct async_handlers_wk; 867 868 struct work_struct roc_done_wk; 869 870 unsigned long init_status; 871 872 unsigned long status; 873 874 u32 queue_sync_cookie; 875 atomic_t queue_sync_counter; 876 /* 877 * for beacon filtering - 878 * currently only one interface can be supported 879 */ 880 struct iwl_mvm_vif *bf_allowed_vif; 881 882 bool hw_registered; 883 bool calibrating; 884 bool support_umac_log; 885 886 u32 ampdu_ref; 887 bool ampdu_toggle; 888 889 struct iwl_notif_wait_data notif_wait; 890 891 union { 892 struct mvm_statistics_rx_v3 rx_stats_v3; 893 struct mvm_statistics_rx rx_stats; 894 }; 895 896 struct { 897 u64 rx_time; 898 u64 tx_time; 899 u64 on_time_rf; 900 u64 on_time_scan; 901 } radio_stats, accu_radio_stats; 902 903 struct list_head add_stream_txqs; 904 union { 905 struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES]; 906 struct iwl_mvm_tvqm_txq_info tvqm_info[IWL_MAX_TVQM_QUEUES]; 907 }; 908 struct work_struct add_stream_wk; /* To add streams to queues */ 909 910 const char *nvm_file_name; 911 struct iwl_nvm_data *nvm_data; 912 /* NVM sections */ 913 struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS]; 914 915 struct iwl_fw_runtime fwrt; 916 917 /* EEPROM MAC addresses */ 918 struct mac_address addresses[IWL_MVM_MAX_ADDRESSES]; 919 920 /* data related to data path */ 921 struct iwl_rx_phy_info last_phy_info; 922 struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT]; 923 u8 rx_ba_sessions; 924 925 /* configured by mac80211 */ 926 u32 rts_threshold; 927 928 /* Scan status, cmd (pre-allocated) and auxiliary station */ 929 unsigned int scan_status; 930 void *scan_cmd; 931 struct iwl_mcast_filter_cmd *mcast_filter_cmd; 932 /* For CDB this is low band scan type, for non-CDB - type. */ 933 enum iwl_mvm_scan_type scan_type; 934 enum iwl_mvm_scan_type hb_scan_type; 935 936 enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all; 937 struct delayed_work scan_timeout_dwork; 938 939 /* max number of simultaneous scans the FW supports */ 940 unsigned int max_scans; 941 942 /* UMAC scan tracking */ 943 u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS]; 944 945 /* start time of last scan in TSF of the mac that requested the scan */ 946 u64 scan_start; 947 948 /* the vif that requested the current scan */ 949 struct iwl_mvm_vif *scan_vif; 950 951 /* rx chain antennas set through debugfs for the scan command */ 952 u8 scan_rx_ant; 953 954 #ifdef CONFIG_IWLWIFI_BCAST_FILTERING 955 /* broadcast filters to configure for each associated station */ 956 const struct iwl_fw_bcast_filter *bcast_filters; 957 #ifdef CONFIG_IWLWIFI_DEBUGFS 958 struct { 959 bool override; 960 struct iwl_bcast_filter_cmd cmd; 961 } dbgfs_bcast_filtering; 962 #endif 963 #endif 964 965 /* Internal station */ 966 struct iwl_mvm_int_sta aux_sta; 967 struct iwl_mvm_int_sta snif_sta; 968 969 bool last_ebs_successful; 970 971 u8 scan_last_antenna_idx; /* to toggle TX between antennas */ 972 u8 mgmt_last_antenna_idx; 973 974 /* last smart fifo state that was successfully sent to firmware */ 975 enum iwl_sf_state sf_state; 976 977 #ifdef CONFIG_IWLWIFI_DEBUGFS 978 struct dentry *debugfs_dir; 979 u32 dbgfs_sram_offset, dbgfs_sram_len; 980 u32 dbgfs_prph_reg_addr; 981 bool disable_power_off; 982 bool disable_power_off_d3; 983 bool beacon_inject_active; 984 985 bool scan_iter_notif_enabled; 986 987 struct debugfs_blob_wrapper nvm_hw_blob; 988 struct debugfs_blob_wrapper nvm_sw_blob; 989 struct debugfs_blob_wrapper nvm_calib_blob; 990 struct debugfs_blob_wrapper nvm_prod_blob; 991 struct debugfs_blob_wrapper nvm_phy_sku_blob; 992 struct debugfs_blob_wrapper nvm_reg_blob; 993 994 struct iwl_mvm_frame_stats drv_rx_stats; 995 spinlock_t drv_stats_lock; 996 u16 dbgfs_rx_phyinfo; 997 #endif 998 999 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX]; 1000 1001 struct list_head time_event_list; 1002 spinlock_t time_event_lock; 1003 1004 /* 1005 * A bitmap indicating the index of the key in use. The firmware 1006 * can hold 16 keys at most. Reflect this fact. 1007 */ 1008 unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)]; 1009 u8 fw_key_deleted[STA_KEY_MAX_NUM]; 1010 1011 /* references taken by the driver and spinlock protecting them */ 1012 spinlock_t refs_lock; 1013 u8 refs[IWL_MVM_REF_COUNT]; 1014 1015 u8 vif_count; 1016 struct ieee80211_vif __rcu *vif_id_to_mac[NUM_MAC_INDEX_DRIVER]; 1017 1018 /* -1 for always, 0 for never, >0 for that many times */ 1019 s8 fw_restart; 1020 u8 *error_recovery_buf; 1021 1022 #ifdef CONFIG_IWLWIFI_LEDS 1023 struct led_classdev led; 1024 #endif 1025 1026 struct ieee80211_vif *p2p_device_vif; 1027 1028 #ifdef CONFIG_PM 1029 struct wiphy_wowlan_support wowlan; 1030 int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen; 1031 1032 /* sched scan settings for net detect */ 1033 struct ieee80211_scan_ies nd_ies; 1034 struct cfg80211_match_set *nd_match_sets; 1035 int n_nd_match_sets; 1036 struct ieee80211_channel **nd_channels; 1037 int n_nd_channels; 1038 bool net_detect; 1039 #ifdef CONFIG_IWLWIFI_DEBUGFS 1040 bool d3_wake_sysassert; 1041 bool d3_test_active; 1042 bool store_d3_resume_sram; 1043 void *d3_resume_sram; 1044 u32 d3_test_pme_ptr; 1045 struct ieee80211_vif *keep_vif; 1046 u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */ 1047 #endif 1048 #endif 1049 1050 /* d0i3 */ 1051 u8 d0i3_ap_sta_id; 1052 bool d0i3_offloading; 1053 struct work_struct d0i3_exit_work; 1054 struct sk_buff_head d0i3_tx; 1055 /* protect d0i3_suspend_flags */ 1056 struct mutex d0i3_suspend_mutex; 1057 unsigned long d0i3_suspend_flags; 1058 /* sync d0i3_tx queue and IWL_MVM_STATUS_IN_D0I3 status flag */ 1059 spinlock_t d0i3_tx_lock; 1060 wait_queue_head_t d0i3_exit_waitq; 1061 wait_queue_head_t rx_sync_waitq; 1062 1063 /* BT-Coex */ 1064 struct iwl_bt_coex_profile_notif last_bt_notif; 1065 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd; 1066 1067 u8 bt_tx_prio; 1068 enum iwl_bt_force_ant_mode bt_force_ant_mode; 1069 1070 /* Aux ROC */ 1071 struct list_head aux_roc_te_list; 1072 1073 /* Thermal Throttling and CTkill */ 1074 struct iwl_mvm_tt_mgmt thermal_throttle; 1075 #ifdef CONFIG_THERMAL 1076 struct iwl_mvm_thermal_device tz_device; 1077 struct iwl_mvm_cooling_device cooling_dev; 1078 #endif 1079 1080 s32 temperature; /* Celsius */ 1081 /* 1082 * Debug option to set the NIC temperature. This option makes the 1083 * driver think this is the actual NIC temperature, and ignore the 1084 * real temperature that is received from the fw 1085 */ 1086 bool temperature_test; /* Debug test temperature is enabled */ 1087 1088 unsigned long bt_coex_last_tcm_ts; 1089 struct iwl_mvm_tcm tcm; 1090 1091 u8 uapsd_noagg_bssid_write_idx; 1092 struct mac_address uapsd_noagg_bssids[IWL_MVM_UAPSD_NOAGG_BSSIDS_NUM] 1093 __aligned(2); 1094 1095 struct iwl_time_quota_cmd last_quota_cmd; 1096 1097 #ifdef CONFIG_NL80211_TESTMODE 1098 u32 noa_duration; 1099 struct ieee80211_vif *noa_vif; 1100 #endif 1101 1102 /* Tx queues */ 1103 u16 aux_queue; 1104 u16 snif_queue; 1105 u16 probe_queue; 1106 u16 p2p_dev_queue; 1107 1108 /* Indicate if device power save is allowed */ 1109 u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */ 1110 /* Indicate if 32Khz external clock is valid */ 1111 u32 ext_clock_valid; 1112 unsigned int max_amsdu_len; /* used for debugfs only */ 1113 1114 struct ieee80211_vif __rcu *csa_vif; 1115 struct ieee80211_vif __rcu *csa_tx_blocked_vif; 1116 u8 csa_tx_block_bcn_timeout; 1117 1118 /* system time of last beacon (for AP/GO interface) */ 1119 u32 ap_last_beacon_gp2; 1120 1121 /* indicates that we transmitted the last beacon */ 1122 bool ibss_manager; 1123 1124 bool lar_regdom_set; 1125 enum iwl_mcc_source mcc_src; 1126 1127 /* TDLS channel switch data */ 1128 struct { 1129 struct delayed_work dwork; 1130 enum iwl_mvm_tdls_cs_state state; 1131 1132 /* 1133 * Current cs sta - might be different from periodic cs peer 1134 * station. Value is meaningless when the cs-state is idle. 1135 */ 1136 u8 cur_sta_id; 1137 1138 /* TDLS periodic channel-switch peer */ 1139 struct { 1140 u8 sta_id; 1141 u8 op_class; 1142 bool initiator; /* are we the link initiator */ 1143 struct cfg80211_chan_def chandef; 1144 struct sk_buff *skb; /* ch sw template */ 1145 u32 ch_sw_tm_ie; 1146 1147 /* timestamp of last ch-sw request sent (GP2 time) */ 1148 u32 sent_timestamp; 1149 } peer; 1150 } tdls_cs; 1151 1152 1153 u32 ciphers[IWL_MVM_NUM_CIPHERS]; 1154 struct ieee80211_cipher_scheme cs[IWL_UCODE_MAX_CS]; 1155 1156 struct cfg80211_ftm_responder_stats ftm_resp_stats; 1157 struct { 1158 struct cfg80211_pmsr_request *req; 1159 struct wireless_dev *req_wdev; 1160 struct list_head loc_list; 1161 int responses[IWL_MVM_TOF_MAX_APS]; 1162 } ftm_initiator; 1163 1164 struct ieee80211_vif *nan_vif; 1165 #define IWL_MAX_BAID 32 1166 struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID]; 1167 1168 /* 1169 * Drop beacons from other APs in AP mode when there are no connected 1170 * clients. 1171 */ 1172 bool drop_bcn_ap_mode; 1173 1174 struct delayed_work cs_tx_unblock_dwork; 1175 1176 /* does a monitor vif exist (only one can exist hence bool) */ 1177 bool monitor_on; 1178 1179 /* sniffer data to include in radiotap */ 1180 __le16 cur_aid; 1181 u8 cur_bssid[ETH_ALEN]; 1182 1183 #ifdef CONFIG_ACPI 1184 struct iwl_mvm_sar_profile sar_profiles[ACPI_SAR_PROFILE_NUM]; 1185 struct iwl_mvm_geo_profile geo_profiles[ACPI_NUM_GEO_PROFILES]; 1186 #endif 1187 }; 1188 1189 /* Extract MVM priv from op_mode and _hw */ 1190 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode) \ 1191 ((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific) 1192 1193 #define IWL_MAC80211_GET_MVM(_hw) \ 1194 IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv)) 1195 1196 /** 1197 * enum iwl_mvm_status - MVM status bits 1198 * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted 1199 * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active 1200 * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running 1201 * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested 1202 * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active 1203 * @IWL_MVM_STATUS_IN_D0I3: NIC is in D0i3 1204 * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running 1205 * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running 1206 * @IWL_MVM_STATUS_NEED_FLUSH_P2P: need to flush P2P bcast STA 1207 */ 1208 enum iwl_mvm_status { 1209 IWL_MVM_STATUS_HW_RFKILL, 1210 IWL_MVM_STATUS_HW_CTKILL, 1211 IWL_MVM_STATUS_ROC_RUNNING, 1212 IWL_MVM_STATUS_HW_RESTART_REQUESTED, 1213 IWL_MVM_STATUS_IN_HW_RESTART, 1214 IWL_MVM_STATUS_IN_D0I3, 1215 IWL_MVM_STATUS_ROC_AUX_RUNNING, 1216 IWL_MVM_STATUS_FIRMWARE_RUNNING, 1217 IWL_MVM_STATUS_NEED_FLUSH_P2P, 1218 }; 1219 1220 /* Keep track of completed init configuration */ 1221 enum iwl_mvm_init_status { 1222 IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0), 1223 IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1), 1224 }; 1225 1226 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm) 1227 { 1228 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) || 1229 test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status); 1230 } 1231 1232 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm) 1233 { 1234 return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status); 1235 } 1236 1237 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm) 1238 { 1239 return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status); 1240 } 1241 1242 /* Must be called with rcu_read_lock() held and it can only be 1243 * released when mvmsta is not needed anymore. 1244 */ 1245 static inline struct iwl_mvm_sta * 1246 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id) 1247 { 1248 struct ieee80211_sta *sta; 1249 1250 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id)) 1251 return NULL; 1252 1253 sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]); 1254 1255 /* This can happen if the station has been removed right now */ 1256 if (IS_ERR_OR_NULL(sta)) 1257 return NULL; 1258 1259 return iwl_mvm_sta_from_mac80211(sta); 1260 } 1261 1262 static inline struct iwl_mvm_sta * 1263 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id) 1264 { 1265 struct ieee80211_sta *sta; 1266 1267 if (sta_id >= ARRAY_SIZE(mvm->fw_id_to_mac_id)) 1268 return NULL; 1269 1270 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id], 1271 lockdep_is_held(&mvm->mutex)); 1272 1273 /* This can happen if the station has been removed right now */ 1274 if (IS_ERR_OR_NULL(sta)) 1275 return NULL; 1276 1277 return iwl_mvm_sta_from_mac80211(sta); 1278 } 1279 1280 static inline struct ieee80211_vif * 1281 iwl_mvm_rcu_dereference_vif_id(struct iwl_mvm *mvm, u8 vif_id, bool rcu) 1282 { 1283 if (WARN_ON(vif_id >= ARRAY_SIZE(mvm->vif_id_to_mac))) 1284 return NULL; 1285 1286 if (rcu) 1287 return rcu_dereference(mvm->vif_id_to_mac[vif_id]); 1288 1289 return rcu_dereference_protected(mvm->vif_id_to_mac[vif_id], 1290 lockdep_is_held(&mvm->mutex)); 1291 } 1292 1293 static inline bool iwl_mvm_is_d0i3_supported(struct iwl_mvm *mvm) 1294 { 1295 return !iwlwifi_mod_params.d0i3_disable && 1296 fw_has_capa(&mvm->fw->ucode_capa, 1297 IWL_UCODE_TLV_CAPA_D0I3_SUPPORT); 1298 } 1299 1300 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm) 1301 { 1302 return fw_has_api(&mvm->fw->ucode_capa, 1303 IWL_UCODE_TLV_API_ADAPTIVE_DWELL); 1304 } 1305 1306 static inline bool iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm *mvm) 1307 { 1308 return fw_has_api(&mvm->fw->ucode_capa, 1309 IWL_UCODE_TLV_API_ADAPTIVE_DWELL_V2); 1310 } 1311 1312 static inline bool iwl_mvm_is_oce_supported(struct iwl_mvm *mvm) 1313 { 1314 /* OCE should never be enabled for LMAC scan FWs */ 1315 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_OCE); 1316 } 1317 1318 static inline bool iwl_mvm_is_frag_ebs_supported(struct iwl_mvm *mvm) 1319 { 1320 return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAG_EBS); 1321 } 1322 1323 static inline bool iwl_mvm_is_short_beacon_notif_supported(struct iwl_mvm *mvm) 1324 { 1325 return fw_has_api(&mvm->fw->ucode_capa, 1326 IWL_UCODE_TLV_API_SHORT_BEACON_NOTIF); 1327 } 1328 1329 static inline bool iwl_mvm_enter_d0i3_on_suspend(struct iwl_mvm *mvm) 1330 { 1331 /* For now we only use this mode to differentiate between 1332 * slave transports, which handle D0i3 entry in suspend by 1333 * themselves in conjunction with runtime PM D0i3. So, this 1334 * function is used to check whether we need to do anything 1335 * when entering suspend or if the transport layer has already 1336 * done it. 1337 */ 1338 return (mvm->trans->system_pm_mode == IWL_PLAT_PM_MODE_D0I3) && 1339 (mvm->trans->runtime_pm_mode != IWL_PLAT_PM_MODE_D0I3); 1340 } 1341 1342 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue) 1343 { 1344 return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) && 1345 (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE); 1346 } 1347 1348 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue) 1349 { 1350 return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) && 1351 (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE); 1352 } 1353 1354 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm) 1355 { 1356 bool nvm_lar = mvm->nvm_data->lar_enabled; 1357 bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa, 1358 IWL_UCODE_TLV_CAPA_LAR_SUPPORT); 1359 1360 if (iwlwifi_mod_params.lar_disable) 1361 return false; 1362 1363 /* 1364 * Enable LAR only if it is supported by the FW (TLV) && 1365 * enabled in the NVM 1366 */ 1367 if (mvm->cfg->nvm_type == IWL_NVM_EXT) 1368 return nvm_lar && tlv_lar; 1369 else 1370 return tlv_lar; 1371 } 1372 1373 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm) 1374 { 1375 return fw_has_api(&mvm->fw->ucode_capa, 1376 IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) || 1377 fw_has_capa(&mvm->fw->ucode_capa, 1378 IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC); 1379 } 1380 1381 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm) 1382 { 1383 return fw_has_capa(&mvm->fw->ucode_capa, 1384 IWL_UCODE_TLV_CAPA_BT_COEX_RRC) && 1385 IWL_MVM_BT_COEX_RRC; 1386 } 1387 1388 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm) 1389 { 1390 return fw_has_capa(&mvm->fw->ucode_capa, 1391 IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) && 1392 !IWL_MVM_HW_CSUM_DISABLE; 1393 } 1394 1395 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm) 1396 { 1397 return fw_has_capa(&mvm->fw->ucode_capa, 1398 IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) && 1399 IWL_MVM_BT_COEX_MPLUT; 1400 } 1401 1402 static inline 1403 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm) 1404 { 1405 return fw_has_capa(&mvm->fw->ucode_capa, 1406 IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) && 1407 !(iwlwifi_mod_params.uapsd_disable & 1408 IWL_DISABLE_UAPSD_P2P_CLIENT); 1409 } 1410 1411 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm) 1412 { 1413 return fw_has_capa(&mvm->fw->ucode_capa, 1414 IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT); 1415 } 1416 1417 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm) 1418 { 1419 /* TODO - replace with TLV once defined */ 1420 return mvm->trans->cfg->use_tfh; 1421 } 1422 1423 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm) 1424 { 1425 /* TODO - better define this */ 1426 return mvm->trans->cfg->device_family >= IWL_DEVICE_FAMILY_22000; 1427 } 1428 1429 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm) 1430 { 1431 /* 1432 * TODO: 1433 * The issue of how to determine CDB APIs and usage is still not fully 1434 * defined. 1435 * There is a compilation for CDB and non-CDB FW, but there may 1436 * be also runtime check. 1437 * For now there is a TLV for checking compilation mode, but a 1438 * runtime check will also have to be here - once defined. 1439 */ 1440 return fw_has_capa(&mvm->fw->ucode_capa, 1441 IWL_UCODE_TLV_CAPA_CDB_SUPPORT); 1442 } 1443 1444 static inline bool iwl_mvm_cdb_scan_api(struct iwl_mvm *mvm) 1445 { 1446 /* 1447 * TODO: should this be the same as iwl_mvm_is_cdb_supported()? 1448 * but then there's a little bit of code in scan that won't make 1449 * any sense... 1450 */ 1451 return mvm->trans->cfg->device_family >= IWL_DEVICE_FAMILY_22000; 1452 } 1453 1454 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm) 1455 { 1456 return fw_has_api(&mvm->fw->ucode_capa, 1457 IWL_UCODE_TLV_API_NEW_RX_STATS); 1458 } 1459 1460 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm) 1461 { 1462 return fw_has_api(&mvm->fw->ucode_capa, 1463 IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY); 1464 } 1465 1466 static inline bool iwl_mvm_has_tlc_offload(const struct iwl_mvm *mvm) 1467 { 1468 return fw_has_capa(&mvm->fw->ucode_capa, 1469 IWL_UCODE_TLV_CAPA_TLC_OFFLOAD); 1470 } 1471 1472 static inline struct agg_tx_status * 1473 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp) 1474 { 1475 if (iwl_mvm_has_new_tx_api(mvm)) 1476 return &((struct iwl_mvm_tx_resp *)tx_resp)->status; 1477 else 1478 return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status; 1479 } 1480 1481 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm) 1482 { 1483 #ifdef CONFIG_THERMAL 1484 /* these two TLV are redundant since the responsibility to CT-kill by 1485 * FW happens only after we send at least one command of 1486 * temperature THs report. 1487 */ 1488 return fw_has_capa(&mvm->fw->ucode_capa, 1489 IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) && 1490 fw_has_capa(&mvm->fw->ucode_capa, 1491 IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT); 1492 #else /* CONFIG_THERMAL */ 1493 return false; 1494 #endif /* CONFIG_THERMAL */ 1495 } 1496 1497 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm) 1498 { 1499 return fw_has_capa(&mvm->fw->ucode_capa, 1500 IWL_UCODE_TLV_CAPA_CTDP_SUPPORT); 1501 } 1502 1503 extern const u8 iwl_mvm_ac_to_tx_fifo[]; 1504 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[]; 1505 1506 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm, 1507 enum ieee80211_ac_numbers ac) 1508 { 1509 return iwl_mvm_has_new_tx_api(mvm) ? 1510 iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac]; 1511 } 1512 1513 struct iwl_rate_info { 1514 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */ 1515 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */ 1516 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */ 1517 u8 plcp_mimo3; /* uCode API: IWL_RATE_MIMO3_6M_PLCP, etc. */ 1518 u8 ieee; /* MAC header: IWL_RATE_6M_IEEE, etc. */ 1519 }; 1520 1521 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm); 1522 int __iwl_mvm_mac_start(struct iwl_mvm *mvm); 1523 1524 /****************** 1525 * MVM Methods 1526 ******************/ 1527 /* uCode */ 1528 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm); 1529 1530 /* Utils */ 1531 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags, 1532 enum nl80211_band band); 1533 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags, 1534 enum nl80211_band band, 1535 struct ieee80211_tx_rate *r); 1536 u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx); 1537 void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm); 1538 u8 first_antenna(u8 mask); 1539 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx); 1540 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, u32 *gp2, u64 *boottime); 1541 u32 iwl_mvm_get_systime(struct iwl_mvm *mvm); 1542 1543 /* Tx / Host Commands */ 1544 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm, 1545 struct iwl_host_cmd *cmd); 1546 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id, 1547 u32 flags, u16 len, const void *data); 1548 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm, 1549 struct iwl_host_cmd *cmd, 1550 u32 *status); 1551 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id, 1552 u16 len, const void *data, 1553 u32 *status); 1554 int iwl_mvm_tx_skb(struct iwl_mvm *mvm, struct sk_buff *skb, 1555 struct ieee80211_sta *sta); 1556 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb); 1557 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb, 1558 struct iwl_tx_cmd *tx_cmd, 1559 struct ieee80211_tx_info *info, u8 sta_id); 1560 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd, 1561 struct ieee80211_tx_info *info, 1562 struct ieee80211_sta *sta, __le16 fc); 1563 void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq); 1564 unsigned int iwl_mvm_max_amsdu_size(struct iwl_mvm *mvm, 1565 struct ieee80211_sta *sta, 1566 unsigned int tid); 1567 1568 #ifdef CONFIG_IWLWIFI_DEBUG 1569 const char *iwl_mvm_get_tx_fail_reason(u32 status); 1570 #else 1571 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; } 1572 #endif 1573 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk, u32 flags); 1574 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal, u32 flags); 1575 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id, 1576 u16 tids, u32 flags); 1577 1578 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm); 1579 1580 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info, 1581 struct iwl_tx_cmd *tx_cmd) 1582 { 1583 struct ieee80211_key_conf *keyconf = info->control.hw_key; 1584 1585 tx_cmd->sec_ctl = TX_CMD_SEC_CCM; 1586 memcpy(tx_cmd->key, keyconf->key, keyconf->keylen); 1587 } 1588 1589 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm) 1590 { 1591 flush_work(&mvm->async_handlers_wk); 1592 } 1593 1594 /* Statistics */ 1595 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm, 1596 struct iwl_rx_packet *pkt); 1597 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm, 1598 struct iwl_rx_cmd_buffer *rxb); 1599 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear); 1600 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm); 1601 1602 /* NVM */ 1603 int iwl_nvm_init(struct iwl_mvm *mvm); 1604 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm); 1605 1606 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm) 1607 { 1608 return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ? 1609 mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant : 1610 mvm->fw->valid_tx_ant; 1611 } 1612 1613 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm) 1614 { 1615 return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ? 1616 mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant : 1617 mvm->fw->valid_rx_ant; 1618 } 1619 1620 static inline void iwl_mvm_toggle_tx_ant(struct iwl_mvm *mvm, u8 *ant) 1621 { 1622 *ant = iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm), *ant); 1623 } 1624 1625 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm) 1626 { 1627 u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN | 1628 FW_PHY_CFG_RX_CHAIN); 1629 u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm); 1630 u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm); 1631 1632 phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS | 1633 valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS; 1634 1635 return mvm->fw->phy_config & phy_config; 1636 } 1637 1638 int iwl_mvm_up(struct iwl_mvm *mvm); 1639 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm); 1640 1641 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm); 1642 bool iwl_mvm_bcast_filter_build_cmd(struct iwl_mvm *mvm, 1643 struct iwl_bcast_filter_cmd *cmd); 1644 1645 /* 1646 * FW notifications / CMD responses handlers 1647 * Convention: iwl_mvm_rx_<NAME OF THE CMD> 1648 */ 1649 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1650 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi, 1651 struct iwl_rx_cmd_buffer *rxb); 1652 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi, 1653 struct iwl_rx_cmd_buffer *rxb, int queue); 1654 void iwl_mvm_rx_monitor_no_data(struct iwl_mvm *mvm, struct napi_struct *napi, 1655 struct iwl_rx_cmd_buffer *rxb, int queue); 1656 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi, 1657 struct iwl_rx_cmd_buffer *rxb, int queue); 1658 int iwl_mvm_notify_rx_queue(struct iwl_mvm *mvm, u32 rxq_mask, 1659 const u8 *data, u32 count); 1660 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb, 1661 int queue); 1662 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1663 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm, 1664 struct iwl_rx_cmd_buffer *rxb); 1665 void iwl_mvm_send_recovery_cmd(struct iwl_mvm *mvm, u32 flags); 1666 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1667 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm, 1668 struct iwl_rx_cmd_buffer *rxb); 1669 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1670 void iwl_mvm_rx_card_state_notif(struct iwl_mvm *mvm, 1671 struct iwl_rx_cmd_buffer *rxb); 1672 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm, 1673 struct iwl_rx_cmd_buffer *rxb); 1674 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm, 1675 struct iwl_rx_cmd_buffer *rxb); 1676 1677 /* MVM PHY */ 1678 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt, 1679 struct cfg80211_chan_def *chandef, 1680 u8 chains_static, u8 chains_dynamic); 1681 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt, 1682 struct cfg80211_chan_def *chandef, 1683 u8 chains_static, u8 chains_dynamic); 1684 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm, 1685 struct iwl_mvm_phy_ctxt *ctxt); 1686 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm, 1687 struct iwl_mvm_phy_ctxt *ctxt); 1688 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm); 1689 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef); 1690 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef); 1691 1692 /* MAC (virtual interface) programming */ 1693 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1694 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1695 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1696 bool force_assoc_off, const u8 *bssid_override); 1697 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1698 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm, 1699 struct ieee80211_vif *vif); 1700 int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm, 1701 struct ieee80211_vif *vif, 1702 struct sk_buff *beacon); 1703 int iwl_mvm_mac_ctxt_send_beacon_cmd(struct iwl_mvm *mvm, 1704 struct sk_buff *beacon, 1705 void *data, int len); 1706 u8 iwl_mvm_mac_ctxt_get_lowest_rate(struct ieee80211_tx_info *info, 1707 struct ieee80211_vif *vif); 1708 void iwl_mvm_mac_ctxt_set_tim(struct iwl_mvm *mvm, 1709 __le32 *tim_index, __le32 *tim_size, 1710 u8 *beacon, u32 frame_size); 1711 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm, 1712 struct iwl_rx_cmd_buffer *rxb); 1713 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm, 1714 struct iwl_rx_cmd_buffer *rxb); 1715 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm, 1716 struct iwl_rx_cmd_buffer *rxb); 1717 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm, 1718 struct iwl_rx_cmd_buffer *rxb); 1719 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 1720 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm, 1721 struct iwl_rx_cmd_buffer *rxb); 1722 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm, 1723 struct ieee80211_vif *vif); 1724 void iwl_mvm_probe_resp_data_notif(struct iwl_mvm *mvm, 1725 struct iwl_rx_cmd_buffer *rxb); 1726 void iwl_mvm_channel_switch_noa_notif(struct iwl_mvm *mvm, 1727 struct iwl_rx_cmd_buffer *rxb); 1728 /* Bindings */ 1729 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1730 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1731 1732 /* Quota management */ 1733 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm) 1734 { 1735 return iwl_mvm_has_quota_low_latency(mvm) ? 1736 sizeof(struct iwl_time_quota_cmd) : 1737 sizeof(struct iwl_time_quota_cmd_v1); 1738 } 1739 1740 static inline struct iwl_time_quota_data 1741 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm, 1742 struct iwl_time_quota_cmd *cmd, 1743 int i) 1744 { 1745 struct iwl_time_quota_data_v1 *quotas; 1746 1747 if (iwl_mvm_has_quota_low_latency(mvm)) 1748 return &cmd->quotas[i]; 1749 1750 quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas; 1751 return (struct iwl_time_quota_data *)"as[i]; 1752 } 1753 1754 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload, 1755 struct ieee80211_vif *disabled_vif); 1756 1757 /* Scanning */ 1758 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1759 struct cfg80211_scan_request *req, 1760 struct ieee80211_scan_ies *ies); 1761 int iwl_mvm_scan_size(struct iwl_mvm *mvm); 1762 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify); 1763 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm); 1764 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm); 1765 void iwl_mvm_scan_timeout_wk(struct work_struct *work); 1766 1767 /* Scheduled scan */ 1768 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm, 1769 struct iwl_rx_cmd_buffer *rxb); 1770 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm, 1771 struct iwl_rx_cmd_buffer *rxb); 1772 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm, 1773 struct ieee80211_vif *vif, 1774 struct cfg80211_sched_scan_request *req, 1775 struct ieee80211_scan_ies *ies, 1776 int type); 1777 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm, 1778 struct iwl_rx_cmd_buffer *rxb); 1779 1780 /* UMAC scan */ 1781 int iwl_mvm_config_scan(struct iwl_mvm *mvm); 1782 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm, 1783 struct iwl_rx_cmd_buffer *rxb); 1784 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm, 1785 struct iwl_rx_cmd_buffer *rxb); 1786 1787 /* MVM debugfs */ 1788 #ifdef CONFIG_IWLWIFI_DEBUGFS 1789 void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir); 1790 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1791 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1792 #else 1793 static inline void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, 1794 struct dentry *dbgfs_dir) 1795 { 1796 } 1797 static inline void 1798 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 1799 { 1800 } 1801 static inline void 1802 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 1803 { 1804 } 1805 #endif /* CONFIG_IWLWIFI_DEBUGFS */ 1806 1807 /* rate scaling */ 1808 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq, bool sync); 1809 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg); 1810 int rs_pretty_print_rate(char *buf, int bufsz, const u32 rate); 1811 void rs_update_last_rssi(struct iwl_mvm *mvm, 1812 struct iwl_mvm_sta *mvmsta, 1813 struct ieee80211_rx_status *rx_status); 1814 1815 /* power management */ 1816 int iwl_mvm_power_update_device(struct iwl_mvm *mvm); 1817 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm); 1818 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm); 1819 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1820 char *buf, int bufsz); 1821 1822 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 1823 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm, 1824 struct iwl_rx_cmd_buffer *rxb); 1825 1826 #ifdef CONFIG_IWLWIFI_LEDS 1827 int iwl_mvm_leds_init(struct iwl_mvm *mvm); 1828 void iwl_mvm_leds_exit(struct iwl_mvm *mvm); 1829 void iwl_mvm_leds_sync(struct iwl_mvm *mvm); 1830 #else 1831 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm) 1832 { 1833 return 0; 1834 } 1835 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm) 1836 { 1837 } 1838 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm) 1839 { 1840 } 1841 #endif 1842 1843 /* D3 (WoWLAN, NetDetect) */ 1844 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan); 1845 int iwl_mvm_resume(struct ieee80211_hw *hw); 1846 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled); 1847 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw, 1848 struct ieee80211_vif *vif, 1849 struct cfg80211_gtk_rekey_data *data); 1850 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw, 1851 struct ieee80211_vif *vif, 1852 struct inet6_dev *idev); 1853 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw, 1854 struct ieee80211_vif *vif, int idx); 1855 extern const struct file_operations iwl_dbgfs_d3_test_ops; 1856 struct iwl_wowlan_status *iwl_mvm_send_wowlan_get_status(struct iwl_mvm *mvm); 1857 #ifdef CONFIG_PM 1858 int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm, 1859 struct ieee80211_vif *vif, 1860 bool host_awake, 1861 u32 cmd_flags); 1862 void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm, 1863 struct ieee80211_vif *vif, 1864 struct iwl_wowlan_status *status); 1865 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, 1866 struct ieee80211_vif *vif); 1867 #else 1868 static inline int iwl_mvm_wowlan_config_key_params(struct iwl_mvm *mvm, 1869 struct ieee80211_vif *vif, 1870 bool host_awake, 1871 u32 cmd_flags) 1872 { 1873 return 0; 1874 } 1875 1876 static inline void iwl_mvm_d0i3_update_keys(struct iwl_mvm *mvm, 1877 struct ieee80211_vif *vif, 1878 struct iwl_wowlan_status *status) 1879 { 1880 } 1881 1882 static inline void 1883 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 1884 { 1885 } 1886 #endif 1887 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta, 1888 struct iwl_wowlan_config_cmd *cmd); 1889 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm, 1890 struct ieee80211_vif *vif, 1891 bool disable_offloading, 1892 bool offload_ns, 1893 u32 cmd_flags); 1894 1895 /* D0i3 */ 1896 void iwl_mvm_ref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type); 1897 void iwl_mvm_unref(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type); 1898 int iwl_mvm_ref_sync(struct iwl_mvm *mvm, enum iwl_mvm_ref_type ref_type); 1899 bool iwl_mvm_ref_taken(struct iwl_mvm *mvm); 1900 1901 #ifdef CONFIG_PM 1902 void iwl_mvm_d0i3_enable_tx(struct iwl_mvm *mvm, __le16 *qos_seq); 1903 int iwl_mvm_enter_d0i3(struct iwl_op_mode *op_mode); 1904 int iwl_mvm_exit_d0i3(struct iwl_op_mode *op_mode); 1905 int _iwl_mvm_exit_d0i3(struct iwl_mvm *mvm); 1906 #endif 1907 1908 /* BT Coex */ 1909 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm); 1910 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm, 1911 struct iwl_rx_cmd_buffer *rxb); 1912 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1913 enum ieee80211_rssi_event_data); 1914 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm); 1915 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm, 1916 struct ieee80211_sta *sta); 1917 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm, 1918 struct ieee80211_sta *sta); 1919 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant); 1920 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm); 1921 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm, 1922 enum nl80211_band band); 1923 u8 iwl_mvm_bt_coex_get_single_ant_msk(struct iwl_mvm *mvm, u8 enabled_ants); 1924 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr, 1925 struct ieee80211_tx_info *info, u8 ac); 1926 1927 /* beacon filtering */ 1928 #ifdef CONFIG_IWLWIFI_DEBUGFS 1929 void 1930 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif, 1931 struct iwl_beacon_filter_cmd *cmd); 1932 #else 1933 static inline void 1934 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif, 1935 struct iwl_beacon_filter_cmd *cmd) 1936 {} 1937 #endif 1938 int iwl_mvm_update_d0i3_power_mode(struct iwl_mvm *mvm, 1939 struct ieee80211_vif *vif, 1940 bool enable, u32 flags); 1941 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm, 1942 struct ieee80211_vif *vif, 1943 u32 flags); 1944 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm, 1945 struct ieee80211_vif *vif, 1946 u32 flags); 1947 /* SMPS */ 1948 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1949 enum iwl_mvm_smps_type_request req_type, 1950 enum ieee80211_smps_mode smps_request); 1951 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm); 1952 1953 /* Low latency */ 1954 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 1955 bool low_latency, 1956 enum iwl_mvm_low_latency_cause cause); 1957 /* get SystemLowLatencyMode - only needed for beacon threshold? */ 1958 bool iwl_mvm_low_latency(struct iwl_mvm *mvm); 1959 bool iwl_mvm_low_latency_band(struct iwl_mvm *mvm, enum nl80211_band band); 1960 void iwl_mvm_send_low_latency_cmd(struct iwl_mvm *mvm, bool low_latency, 1961 u16 mac_id); 1962 1963 /* get VMACLowLatencyMode */ 1964 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif) 1965 { 1966 /* 1967 * should this consider associated/active/... state? 1968 * 1969 * Normally low-latency should only be active on interfaces 1970 * that are active, but at least with debugfs it can also be 1971 * enabled on interfaces that aren't active. However, when 1972 * interface aren't active then they aren't added into the 1973 * binding, so this has no real impact. For now, just return 1974 * the current desired low-latency state. 1975 */ 1976 return mvmvif->low_latency_actual; 1977 } 1978 1979 static inline 1980 void iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif *mvmvif, bool set, 1981 enum iwl_mvm_low_latency_cause cause) 1982 { 1983 u8 new_state; 1984 1985 if (set) 1986 mvmvif->low_latency |= cause; 1987 else 1988 mvmvif->low_latency &= ~cause; 1989 1990 /* 1991 * if LOW_LATENCY_DEBUGFS_FORCE_ENABLE is enabled no changes are 1992 * allowed to actual mode. 1993 */ 1994 if (mvmvif->low_latency & LOW_LATENCY_DEBUGFS_FORCE_ENABLE && 1995 cause != LOW_LATENCY_DEBUGFS_FORCE_ENABLE) 1996 return; 1997 1998 if (cause == LOW_LATENCY_DEBUGFS_FORCE_ENABLE && set) 1999 /* 2000 * We enter force state 2001 */ 2002 new_state = !!(mvmvif->low_latency & 2003 LOW_LATENCY_DEBUGFS_FORCE); 2004 else 2005 /* 2006 * Check if any other one set low latency 2007 */ 2008 new_state = !!(mvmvif->low_latency & 2009 ~(LOW_LATENCY_DEBUGFS_FORCE_ENABLE | 2010 LOW_LATENCY_DEBUGFS_FORCE)); 2011 2012 mvmvif->low_latency_actual = new_state; 2013 } 2014 2015 /* Return a bitmask with all the hw supported queues, except for the 2016 * command queue, which can't be flushed. 2017 */ 2018 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm) 2019 { 2020 return ((BIT(mvm->cfg->base_params->num_of_queues) - 1) & 2021 ~BIT(IWL_MVM_DQA_CMD_QUEUE)); 2022 } 2023 2024 static inline void iwl_mvm_stop_device(struct iwl_mvm *mvm) 2025 { 2026 lockdep_assert_held(&mvm->mutex); 2027 iwl_fw_cancel_timestamp(&mvm->fwrt); 2028 clear_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status); 2029 iwl_fwrt_stop_device(&mvm->fwrt); 2030 iwl_free_fw_paging(&mvm->fwrt); 2031 iwl_fw_dump_conf_clear(&mvm->fwrt); 2032 } 2033 2034 /* Re-configure the SCD for a queue that has already been configured */ 2035 int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo, int sta_id, 2036 int tid, int frame_limit, u16 ssn); 2037 2038 /* Thermal management and CT-kill */ 2039 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff); 2040 void iwl_mvm_tt_temp_changed(struct iwl_mvm *mvm, u32 temp); 2041 void iwl_mvm_temp_notif(struct iwl_mvm *mvm, 2042 struct iwl_rx_cmd_buffer *rxb); 2043 void iwl_mvm_tt_handler(struct iwl_mvm *mvm); 2044 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff); 2045 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm); 2046 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state); 2047 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp); 2048 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 2049 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm); 2050 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm); 2051 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget); 2052 2053 /* Location Aware Regulatory */ 2054 struct iwl_mcc_update_resp * 2055 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2, 2056 enum iwl_mcc_source src_id); 2057 int iwl_mvm_init_mcc(struct iwl_mvm *mvm); 2058 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm, 2059 struct iwl_rx_cmd_buffer *rxb); 2060 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy, 2061 const char *alpha2, 2062 enum iwl_mcc_source src_id, 2063 bool *changed); 2064 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm, 2065 bool *changed); 2066 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm); 2067 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm); 2068 2069 /* smart fifo */ 2070 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2071 bool added_vif); 2072 2073 /* FTM responder */ 2074 int iwl_mvm_ftm_start_responder(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2075 void iwl_mvm_ftm_restart_responder(struct iwl_mvm *mvm, 2076 struct ieee80211_vif *vif); 2077 void iwl_mvm_ftm_responder_stats(struct iwl_mvm *mvm, 2078 struct iwl_rx_cmd_buffer *rxb); 2079 2080 /* FTM initiator */ 2081 void iwl_mvm_ftm_restart(struct iwl_mvm *mvm); 2082 void iwl_mvm_ftm_range_resp(struct iwl_mvm *mvm, 2083 struct iwl_rx_cmd_buffer *rxb); 2084 void iwl_mvm_ftm_lc_notif(struct iwl_mvm *mvm, 2085 struct iwl_rx_cmd_buffer *rxb); 2086 int iwl_mvm_ftm_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2087 struct cfg80211_pmsr_request *request); 2088 void iwl_mvm_ftm_abort(struct iwl_mvm *mvm, struct cfg80211_pmsr_request *req); 2089 2090 /* TDLS */ 2091 2092 /* 2093 * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present. 2094 * This TID is marked as used vs the AP and all connected TDLS peers. 2095 */ 2096 #define IWL_MVM_TDLS_FW_TID 4 2097 2098 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2099 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm); 2100 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2101 bool sta_added); 2102 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw, 2103 struct ieee80211_vif *vif); 2104 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw, 2105 struct ieee80211_vif *vif, 2106 struct ieee80211_sta *sta, u8 oper_class, 2107 struct cfg80211_chan_def *chandef, 2108 struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie); 2109 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw, 2110 struct ieee80211_vif *vif, 2111 struct ieee80211_tdls_ch_sw_params *params); 2112 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw, 2113 struct ieee80211_vif *vif, 2114 struct ieee80211_sta *sta); 2115 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb); 2116 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work); 2117 2118 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm, 2119 struct iwl_mvm_internal_rxq_notif *notif, 2120 u32 size); 2121 void iwl_mvm_reorder_timer_expired(struct timer_list *t); 2122 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm); 2123 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm); 2124 2125 #define MVM_TCM_PERIOD_MSEC 500 2126 #define MVM_TCM_PERIOD (HZ * MVM_TCM_PERIOD_MSEC / 1000) 2127 #define MVM_LL_PERIOD (10 * HZ) 2128 void iwl_mvm_tcm_work(struct work_struct *work); 2129 void iwl_mvm_recalc_tcm(struct iwl_mvm *mvm); 2130 void iwl_mvm_pause_tcm(struct iwl_mvm *mvm, bool with_cancel); 2131 void iwl_mvm_resume_tcm(struct iwl_mvm *mvm); 2132 void iwl_mvm_tcm_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2133 void iwl_mvm_tcm_rm_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 2134 u8 iwl_mvm_tcm_load_percentage(u32 airtime, u32 elapsed); 2135 2136 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error); 2137 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm, 2138 struct ieee80211_vif *vif, 2139 bool tdls, bool cmd_q); 2140 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 2141 const char *errmsg); 2142 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm, 2143 struct ieee80211_vif *vif, 2144 const struct ieee80211_sta *sta, 2145 u16 tid); 2146 2147 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b); 2148 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm); 2149 #ifdef CONFIG_IWLWIFI_DEBUGFS 2150 void iwl_mvm_sta_add_debugfs(struct ieee80211_hw *hw, 2151 struct ieee80211_vif *vif, 2152 struct ieee80211_sta *sta, 2153 struct dentry *dir); 2154 #endif 2155 2156 /* Channel info utils */ 2157 static inline bool iwl_mvm_has_ultra_hb_channel(struct iwl_mvm *mvm) 2158 { 2159 return fw_has_capa(&mvm->fw->ucode_capa, 2160 IWL_UCODE_TLV_CAPA_ULTRA_HB_CHANNELS); 2161 } 2162 2163 static inline void *iwl_mvm_chan_info_cmd_tail(struct iwl_mvm *mvm, 2164 struct iwl_fw_channel_info *ci) 2165 { 2166 return (u8 *)ci + (iwl_mvm_has_ultra_hb_channel(mvm) ? 2167 sizeof(struct iwl_fw_channel_info) : 2168 sizeof(struct iwl_fw_channel_info_v1)); 2169 } 2170 2171 static inline size_t iwl_mvm_chan_info_padding(struct iwl_mvm *mvm) 2172 { 2173 return iwl_mvm_has_ultra_hb_channel(mvm) ? 0 : 2174 sizeof(struct iwl_fw_channel_info) - 2175 sizeof(struct iwl_fw_channel_info_v1); 2176 } 2177 2178 static inline void iwl_mvm_set_chan_info(struct iwl_mvm *mvm, 2179 struct iwl_fw_channel_info *ci, 2180 u32 chan, u8 band, u8 width, 2181 u8 ctrl_pos) 2182 { 2183 if (iwl_mvm_has_ultra_hb_channel(mvm)) { 2184 ci->channel = cpu_to_le32(chan); 2185 ci->band = band; 2186 ci->width = width; 2187 ci->ctrl_pos = ctrl_pos; 2188 } else { 2189 struct iwl_fw_channel_info_v1 *ci_v1 = 2190 (struct iwl_fw_channel_info_v1 *)ci; 2191 2192 ci_v1->channel = chan; 2193 ci_v1->band = band; 2194 ci_v1->width = width; 2195 ci_v1->ctrl_pos = ctrl_pos; 2196 } 2197 } 2198 2199 static inline void 2200 iwl_mvm_set_chan_info_chandef(struct iwl_mvm *mvm, 2201 struct iwl_fw_channel_info *ci, 2202 struct cfg80211_chan_def *chandef) 2203 { 2204 iwl_mvm_set_chan_info(mvm, ci, chandef->chan->hw_value, 2205 (chandef->chan->band == NL80211_BAND_2GHZ ? 2206 PHY_BAND_24 : PHY_BAND_5), 2207 iwl_mvm_get_channel_width(chandef), 2208 iwl_mvm_get_ctrl_pos(chandef)); 2209 } 2210 2211 #endif /* __IWL_MVM_H__ */ 2212