1 /* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */
2 /*
3  * Copyright (C) 2012-2014, 2018-2022 Intel Corporation
4  * Copyright (C) 2013-2015 Intel Mobile Communications GmbH
5  * Copyright (C) 2016-2017 Intel Deutschland GmbH
6  */
7 #ifndef __IWL_MVM_H__
8 #define __IWL_MVM_H__
9 
10 #include <linux/list.h>
11 #include <linux/spinlock.h>
12 #include <linux/leds.h>
13 #include <linux/in6.h>
14 
15 #ifdef CONFIG_THERMAL
16 #include <linux/thermal.h>
17 #endif
18 
19 #include <linux/ktime.h>
20 
21 #include "iwl-op-mode.h"
22 #include "iwl-trans.h"
23 #include "fw/notif-wait.h"
24 #include "iwl-eeprom-parse.h"
25 #include "fw/file.h"
26 #include "iwl-config.h"
27 #include "sta.h"
28 #include "fw-api.h"
29 #include "constants.h"
30 #include "fw/runtime.h"
31 #include "fw/dbg.h"
32 #include "fw/acpi.h"
33 #include "mei/iwl-mei.h"
34 #include "iwl-nvm-parse.h"
35 
36 #include <linux/average.h>
37 
38 #define IWL_MVM_MAX_ADDRESSES		5
39 /* RSSI offset for WkP */
40 #define IWL_RSSI_OFFSET 50
41 #define IWL_MVM_MISSED_BEACONS_THRESHOLD 8
42 #define IWL_MVM_MISSED_BEACONS_THRESHOLD_LONG 16
43 
44 /* A TimeUnit is 1024 microsecond */
45 #define MSEC_TO_TU(_msec)	(_msec*1000/1024)
46 
47 /* For GO, this value represents the number of TUs before CSA "beacon
48  * 0" TBTT when the CSA time-event needs to be scheduled to start.  It
49  * must be big enough to ensure that we switch in time.
50  */
51 #define IWL_MVM_CHANNEL_SWITCH_TIME_GO		40
52 
53 /* For client, this value represents the number of TUs before CSA
54  * "beacon 1" TBTT, instead.  This is because we don't know when the
55  * GO/AP will be in the new channel, so we switch early enough.
56  */
57 #define IWL_MVM_CHANNEL_SWITCH_TIME_CLIENT	10
58 
59 /*
60  * This value (in TUs) is used to fine tune the CSA NoA end time which should
61  * be just before "beacon 0" TBTT.
62  */
63 #define IWL_MVM_CHANNEL_SWITCH_MARGIN 4
64 
65 /*
66  * Number of beacons to transmit on a new channel until we unblock tx to
67  * the stations, even if we didn't identify them on a new channel
68  */
69 #define IWL_MVM_CS_UNBLOCK_TX_TIMEOUT 3
70 
71 /* offchannel queue towards mac80211 */
72 #define IWL_MVM_OFFCHANNEL_QUEUE 0
73 
74 extern const struct ieee80211_ops iwl_mvm_hw_ops;
75 
76 /**
77  * struct iwl_mvm_mod_params - module parameters for iwlmvm
78  * @init_dbg: if true, then the NIC won't be stopped if the INIT fw asserted.
79  *	We will register to mac80211 to have testmode working. The NIC must not
80  *	be up'ed after the INIT fw asserted. This is useful to be able to use
81  *	proprietary tools over testmode to debug the INIT fw.
82  * @power_scheme: one of enum iwl_power_scheme
83  */
84 struct iwl_mvm_mod_params {
85 	bool init_dbg;
86 	int power_scheme;
87 };
88 extern struct iwl_mvm_mod_params iwlmvm_mod_params;
89 
90 struct iwl_mvm_phy_ctxt {
91 	u16 id;
92 	u16 color;
93 	u32 ref;
94 
95 	enum nl80211_chan_width width;
96 
97 	struct ieee80211_channel *channel;
98 
99 	/* track for RLC config command */
100 	u32 center_freq1;
101 };
102 
103 struct iwl_mvm_time_event_data {
104 	struct ieee80211_vif *vif;
105 	struct list_head list;
106 	unsigned long end_jiffies;
107 	u32 duration;
108 	bool running;
109 	u32 uid;
110 
111 	/*
112 	 * The access to the 'id' field must be done when the
113 	 * mvm->time_event_lock is held, as it value is used to indicate
114 	 * if the te is in the time event list or not (when id == TE_MAX)
115 	 */
116 	u32 id;
117 };
118 
119  /* Power management */
120 
121 /**
122  * enum iwl_power_scheme
123  * @IWL_POWER_LEVEL_CAM - Continuously Active Mode
124  * @IWL_POWER_LEVEL_BPS - Balanced Power Save (default)
125  * @IWL_POWER_LEVEL_LP  - Low Power
126  */
127 enum iwl_power_scheme {
128 	IWL_POWER_SCHEME_CAM = 1,
129 	IWL_POWER_SCHEME_BPS,
130 	IWL_POWER_SCHEME_LP
131 };
132 
133 #define IWL_CONN_MAX_LISTEN_INTERVAL	10
134 #define IWL_UAPSD_MAX_SP		IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL
135 
136 #ifdef CONFIG_IWLWIFI_DEBUGFS
137 enum iwl_dbgfs_pm_mask {
138 	MVM_DEBUGFS_PM_KEEP_ALIVE = BIT(0),
139 	MVM_DEBUGFS_PM_SKIP_OVER_DTIM = BIT(1),
140 	MVM_DEBUGFS_PM_SKIP_DTIM_PERIODS = BIT(2),
141 	MVM_DEBUGFS_PM_RX_DATA_TIMEOUT = BIT(3),
142 	MVM_DEBUGFS_PM_TX_DATA_TIMEOUT = BIT(4),
143 	MVM_DEBUGFS_PM_LPRX_ENA = BIT(6),
144 	MVM_DEBUGFS_PM_LPRX_RSSI_THRESHOLD = BIT(7),
145 	MVM_DEBUGFS_PM_SNOOZE_ENABLE = BIT(8),
146 	MVM_DEBUGFS_PM_UAPSD_MISBEHAVING = BIT(9),
147 	MVM_DEBUGFS_PM_USE_PS_POLL = BIT(10),
148 };
149 
150 struct iwl_dbgfs_pm {
151 	u16 keep_alive_seconds;
152 	u32 rx_data_timeout;
153 	u32 tx_data_timeout;
154 	bool skip_over_dtim;
155 	u8 skip_dtim_periods;
156 	bool lprx_ena;
157 	u32 lprx_rssi_threshold;
158 	bool snooze_ena;
159 	bool uapsd_misbehaving;
160 	bool use_ps_poll;
161 	int mask;
162 };
163 
164 /* beacon filtering */
165 
166 enum iwl_dbgfs_bf_mask {
167 	MVM_DEBUGFS_BF_ENERGY_DELTA = BIT(0),
168 	MVM_DEBUGFS_BF_ROAMING_ENERGY_DELTA = BIT(1),
169 	MVM_DEBUGFS_BF_ROAMING_STATE = BIT(2),
170 	MVM_DEBUGFS_BF_TEMP_THRESHOLD = BIT(3),
171 	MVM_DEBUGFS_BF_TEMP_FAST_FILTER = BIT(4),
172 	MVM_DEBUGFS_BF_TEMP_SLOW_FILTER = BIT(5),
173 	MVM_DEBUGFS_BF_ENABLE_BEACON_FILTER = BIT(6),
174 	MVM_DEBUGFS_BF_DEBUG_FLAG = BIT(7),
175 	MVM_DEBUGFS_BF_ESCAPE_TIMER = BIT(8),
176 	MVM_DEBUGFS_BA_ESCAPE_TIMER = BIT(9),
177 	MVM_DEBUGFS_BA_ENABLE_BEACON_ABORT = BIT(10),
178 };
179 
180 struct iwl_dbgfs_bf {
181 	u32 bf_energy_delta;
182 	u32 bf_roaming_energy_delta;
183 	u32 bf_roaming_state;
184 	u32 bf_temp_threshold;
185 	u32 bf_temp_fast_filter;
186 	u32 bf_temp_slow_filter;
187 	u32 bf_enable_beacon_filter;
188 	u32 bf_debug_flag;
189 	u32 bf_escape_timer;
190 	u32 ba_escape_timer;
191 	u32 ba_enable_beacon_abort;
192 	int mask;
193 };
194 #endif
195 
196 enum iwl_mvm_smps_type_request {
197 	IWL_MVM_SMPS_REQ_BT_COEX,
198 	IWL_MVM_SMPS_REQ_TT,
199 	IWL_MVM_SMPS_REQ_PROT,
200 	IWL_MVM_SMPS_REQ_FW,
201 	NUM_IWL_MVM_SMPS_REQ,
202 };
203 
204 enum iwl_bt_force_ant_mode {
205 	BT_FORCE_ANT_DIS = 0,
206 	BT_FORCE_ANT_AUTO,
207 	BT_FORCE_ANT_BT,
208 	BT_FORCE_ANT_WIFI,
209 
210 	BT_FORCE_ANT_MAX,
211 };
212 
213 /**
214  * struct iwl_mvm_low_latency_force - low latency force mode set by debugfs
215  * @LOW_LATENCY_FORCE_UNSET: unset force mode
216  * @LOW_LATENCY_FORCE_ON: for low latency on
217  * @LOW_LATENCY_FORCE_OFF: for low latency off
218  * @NUM_LOW_LATENCY_FORCE: max num of modes
219  */
220 enum iwl_mvm_low_latency_force {
221 	LOW_LATENCY_FORCE_UNSET,
222 	LOW_LATENCY_FORCE_ON,
223 	LOW_LATENCY_FORCE_OFF,
224 	NUM_LOW_LATENCY_FORCE
225 };
226 
227 /**
228 * struct iwl_mvm_low_latency_cause - low latency set causes
229 * @LOW_LATENCY_TRAFFIC: indicates low latency traffic was detected
230 * @LOW_LATENCY_DEBUGFS: low latency mode set from debugfs
231 * @LOW_LATENCY_VCMD: low latency mode set from vendor command
232 * @LOW_LATENCY_VIF_TYPE: low latency mode set because of vif type (ap)
233 * @LOW_LATENCY_DEBUGFS_FORCE_ENABLE: indicate that force mode is enabled
234 *	the actual set/unset is done with LOW_LATENCY_DEBUGFS_FORCE
235 * @LOW_LATENCY_DEBUGFS_FORCE: low latency force mode from debugfs
236 *	set this with LOW_LATENCY_DEBUGFS_FORCE_ENABLE flag
237 *	in low_latency.
238 */
239 enum iwl_mvm_low_latency_cause {
240 	LOW_LATENCY_TRAFFIC = BIT(0),
241 	LOW_LATENCY_DEBUGFS = BIT(1),
242 	LOW_LATENCY_VCMD = BIT(2),
243 	LOW_LATENCY_VIF_TYPE = BIT(3),
244 	LOW_LATENCY_DEBUGFS_FORCE_ENABLE = BIT(4),
245 	LOW_LATENCY_DEBUGFS_FORCE = BIT(5),
246 };
247 
248 /**
249 * struct iwl_mvm_vif_bf_data - beacon filtering related data
250 * @bf_enabled: indicates if beacon filtering is enabled
251 * @ba_enabled: indicated if beacon abort is enabled
252 * @ave_beacon_signal: average beacon signal
253 * @last_cqm_event: rssi of the last cqm event
254 * @bt_coex_min_thold: minimum threshold for BT coex
255 * @bt_coex_max_thold: maximum threshold for BT coex
256 * @last_bt_coex_event: rssi of the last BT coex event
257 */
258 struct iwl_mvm_vif_bf_data {
259 	bool bf_enabled;
260 	bool ba_enabled;
261 	int ave_beacon_signal;
262 	int last_cqm_event;
263 	int bt_coex_min_thold;
264 	int bt_coex_max_thold;
265 	int last_bt_coex_event;
266 };
267 
268 /**
269  * struct iwl_probe_resp_data - data for NoA/CSA updates
270  * @rcu_head: used for freeing the data on update
271  * @notif: notification data
272  * @noa_len: length of NoA attribute, calculated from the notification
273  */
274 struct iwl_probe_resp_data {
275 	struct rcu_head rcu_head;
276 	struct iwl_probe_resp_data_notif notif;
277 	int noa_len;
278 };
279 
280 /**
281  * struct iwl_mvm_vif - data per Virtual Interface, it is a MAC context
282  * @id: between 0 and 3
283  * @color: to solve races upon MAC addition and removal
284  * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
285  * @bssid: BSSID for this (client) interface
286  * @associated: indicates that we're currently associated, used only for
287  *	managing the firmware state in iwl_mvm_bss_info_changed_station()
288  * @ap_assoc_sta_count: count of stations associated to us - valid only
289  *	if VIF type is AP
290  * @uploaded: indicates the MAC context has been added to the device
291  * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
292  *	should get quota etc.
293  * @pm_enabled - Indicate if MAC power management is allowed
294  * @monitor_active: indicates that monitor context is configured, and that the
295  *	interface should get quota etc.
296  * @low_latency: bit flags for low latency
297  *	see enum &iwl_mvm_low_latency_cause for causes.
298  * @low_latency_actual: boolean, indicates low latency is set,
299  *	as a result from low_latency bit flags and takes force into account.
300  * @authorized: indicates the AP station was set to authorized
301  * @ps_disabled: indicates that this interface requires PS to be disabled
302  * @queue_params: QoS params for this MAC
303  * @bcast_sta: station used for broadcast packets. Used by the following
304  *  vifs: P2P_DEVICE, GO and AP.
305  * @beacon_skb: the skb used to hold the AP/GO beacon template
306  * @smps_requests: the SMPS requests of different parts of the driver,
307  *	combined on update to yield the overall request to mac80211.
308  * @beacon_stats: beacon statistics, containing the # of received beacons,
309  *	# of received beacons accumulated over FW restart, and the current
310  *	average signal of beacons retrieved from the firmware
311  * @csa_failed: CSA failed to schedule time event, report an error later
312  * @features: hw features active for this vif
313  * @probe_resp_data: data from FW notification to store NOA and CSA related
314  *	data to be inserted into probe response.
315  */
316 struct iwl_mvm_vif {
317 	struct iwl_mvm *mvm;
318 	u16 id;
319 	u16 color;
320 	u8 ap_sta_id;
321 
322 	u8 bssid[ETH_ALEN];
323 	bool associated;
324 	u8 ap_assoc_sta_count;
325 
326 	u16 cab_queue;
327 
328 	bool uploaded;
329 	bool ap_ibss_active;
330 	bool pm_enabled;
331 	bool monitor_active;
332 	u8 low_latency: 6;
333 	u8 low_latency_actual: 1;
334 	u8 authorized:1;
335 	bool ps_disabled;
336 	struct iwl_mvm_vif_bf_data bf_data;
337 
338 	struct {
339 		u32 num_beacons, accu_num_beacons;
340 		u8 avg_signal;
341 	} beacon_stats;
342 
343 	u32 ap_beacon_time;
344 
345 	enum iwl_tsf_id tsf_id;
346 
347 	/*
348 	 * QoS data from mac80211, need to store this here
349 	 * as mac80211 has a separate callback but we need
350 	 * to have the data for the MAC context
351 	 */
352 	struct ieee80211_tx_queue_params queue_params[IEEE80211_NUM_ACS];
353 	struct iwl_mvm_time_event_data time_event_data;
354 	struct iwl_mvm_time_event_data hs_time_event_data;
355 
356 	struct iwl_mvm_int_sta bcast_sta;
357 	struct iwl_mvm_int_sta mcast_sta;
358 
359 	/*
360 	 * Assigned while mac80211 has the interface in a channel context,
361 	 * or, for P2P Device, while it exists.
362 	 */
363 	struct iwl_mvm_phy_ctxt *phy_ctxt;
364 
365 #ifdef CONFIG_PM
366 	/* WoWLAN GTK rekey data */
367 	struct {
368 		u8 kck[NL80211_KCK_EXT_LEN];
369 		u8 kek[NL80211_KEK_EXT_LEN];
370 		size_t kek_len;
371 		size_t kck_len;
372 		u32 akm;
373 		__le64 replay_ctr;
374 		bool valid;
375 	} rekey_data;
376 
377 	int tx_key_idx;
378 
379 	bool seqno_valid;
380 	u16 seqno;
381 #endif
382 
383 #if IS_ENABLED(CONFIG_IPV6)
384 	/* IPv6 addresses for WoWLAN */
385 	struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
386 	unsigned long tentative_addrs[BITS_TO_LONGS(IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX)];
387 	int num_target_ipv6_addrs;
388 #endif
389 
390 #ifdef CONFIG_IWLWIFI_DEBUGFS
391 	struct dentry *dbgfs_dir;
392 	struct dentry *dbgfs_slink;
393 	struct iwl_dbgfs_pm dbgfs_pm;
394 	struct iwl_dbgfs_bf dbgfs_bf;
395 	struct iwl_mac_power_cmd mac_pwr_cmd;
396 	int dbgfs_quota_min;
397 #endif
398 
399 	enum ieee80211_smps_mode smps_requests[NUM_IWL_MVM_SMPS_REQ];
400 
401 	/* FW identified misbehaving AP */
402 	u8 uapsd_misbehaving_bssid[ETH_ALEN];
403 
404 	struct delayed_work uapsd_nonagg_detected_wk;
405 
406 	/* Indicates that CSA countdown may be started */
407 	bool csa_countdown;
408 	bool csa_failed;
409 	u16 csa_target_freq;
410 	u16 csa_count;
411 	u16 csa_misbehave;
412 	struct delayed_work csa_work;
413 
414 	/* Indicates that we are waiting for a beacon on a new channel */
415 	bool csa_bcn_pending;
416 
417 	/* TCP Checksum Offload */
418 	netdev_features_t features;
419 
420 	struct iwl_probe_resp_data __rcu *probe_resp_data;
421 
422 	/* we can only have 2 GTK + 2 IGTK active at a time */
423 	struct ieee80211_key_conf *ap_early_keys[4];
424 
425 	/* 26-tone RU OFDMA transmissions should be blocked */
426 	bool he_ru_2mhz_block;
427 
428 	struct {
429 		struct ieee80211_key_conf __rcu *keys[2];
430 	} bcn_prot;
431 };
432 
433 static inline struct iwl_mvm_vif *
434 iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
435 {
436 	return (void *)vif->drv_priv;
437 }
438 
439 extern const u8 tid_to_mac80211_ac[];
440 
441 #define IWL_MVM_SCAN_STOPPING_SHIFT	8
442 
443 enum iwl_scan_status {
444 	IWL_MVM_SCAN_REGULAR		= BIT(0),
445 	IWL_MVM_SCAN_SCHED		= BIT(1),
446 	IWL_MVM_SCAN_NETDETECT		= BIT(2),
447 
448 	IWL_MVM_SCAN_STOPPING_REGULAR	= BIT(8),
449 	IWL_MVM_SCAN_STOPPING_SCHED	= BIT(9),
450 	IWL_MVM_SCAN_STOPPING_NETDETECT	= BIT(10),
451 
452 	IWL_MVM_SCAN_REGULAR_MASK	= IWL_MVM_SCAN_REGULAR |
453 					  IWL_MVM_SCAN_STOPPING_REGULAR,
454 	IWL_MVM_SCAN_SCHED_MASK		= IWL_MVM_SCAN_SCHED |
455 					  IWL_MVM_SCAN_STOPPING_SCHED,
456 	IWL_MVM_SCAN_NETDETECT_MASK	= IWL_MVM_SCAN_NETDETECT |
457 					  IWL_MVM_SCAN_STOPPING_NETDETECT,
458 
459 	IWL_MVM_SCAN_STOPPING_MASK	= 0xff << IWL_MVM_SCAN_STOPPING_SHIFT,
460 	IWL_MVM_SCAN_MASK		= 0xff,
461 };
462 
463 enum iwl_mvm_scan_type {
464 	IWL_SCAN_TYPE_NOT_SET,
465 	IWL_SCAN_TYPE_UNASSOC,
466 	IWL_SCAN_TYPE_WILD,
467 	IWL_SCAN_TYPE_MILD,
468 	IWL_SCAN_TYPE_FRAGMENTED,
469 	IWL_SCAN_TYPE_FAST_BALANCE,
470 };
471 
472 enum iwl_mvm_sched_scan_pass_all_states {
473 	SCHED_SCAN_PASS_ALL_DISABLED,
474 	SCHED_SCAN_PASS_ALL_ENABLED,
475 	SCHED_SCAN_PASS_ALL_FOUND,
476 };
477 
478 /**
479  * struct iwl_mvm_tt_mgnt - Thermal Throttling Management structure
480  * @ct_kill_exit: worker to exit thermal kill
481  * @dynamic_smps: Is thermal throttling enabled dynamic_smps?
482  * @tx_backoff: The current thremal throttling tx backoff in uSec.
483  * @min_backoff: The minimal tx backoff due to power restrictions
484  * @params: Parameters to configure the thermal throttling algorithm.
485  * @throttle: Is thermal throttling is active?
486  */
487 struct iwl_mvm_tt_mgmt {
488 	struct delayed_work ct_kill_exit;
489 	bool dynamic_smps;
490 	u32 tx_backoff;
491 	u32 min_backoff;
492 	struct iwl_tt_params params;
493 	bool throttle;
494 };
495 
496 #ifdef CONFIG_THERMAL
497 /**
498  *struct iwl_mvm_thermal_device - thermal zone related data
499  * @temp_trips: temperature thresholds for report
500  * @fw_trips_index: keep indexes to original array - temp_trips
501  * @tzone: thermal zone device data
502 */
503 struct iwl_mvm_thermal_device {
504 	struct thermal_trip trips[IWL_MAX_DTS_TRIPS];
505 	u8 fw_trips_index[IWL_MAX_DTS_TRIPS];
506 	struct thermal_zone_device *tzone;
507 };
508 
509 /*
510  * struct iwl_mvm_cooling_device
511  * @cur_state: current state
512  * @cdev: struct thermal cooling device
513  */
514 struct iwl_mvm_cooling_device {
515 	u32 cur_state;
516 	struct thermal_cooling_device *cdev;
517 };
518 #endif
519 
520 #define IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES 8
521 
522 struct iwl_mvm_frame_stats {
523 	u32 legacy_frames;
524 	u32 ht_frames;
525 	u32 vht_frames;
526 	u32 bw_20_frames;
527 	u32 bw_40_frames;
528 	u32 bw_80_frames;
529 	u32 bw_160_frames;
530 	u32 sgi_frames;
531 	u32 ngi_frames;
532 	u32 siso_frames;
533 	u32 mimo2_frames;
534 	u32 agg_frames;
535 	u32 ampdu_count;
536 	u32 success_frames;
537 	u32 fail_frames;
538 	u32 last_rates[IWL_MVM_NUM_LAST_FRAMES_UCODE_RATES];
539 	int last_frame_idx;
540 };
541 
542 #define IWL_MVM_DEBUG_SET_TEMPERATURE_DISABLE 0xff
543 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MIN -100
544 #define IWL_MVM_DEBUG_SET_TEMPERATURE_MAX 200
545 
546 enum iwl_mvm_tdls_cs_state {
547 	IWL_MVM_TDLS_SW_IDLE = 0,
548 	IWL_MVM_TDLS_SW_REQ_SENT,
549 	IWL_MVM_TDLS_SW_RESP_RCVD,
550 	IWL_MVM_TDLS_SW_REQ_RCVD,
551 	IWL_MVM_TDLS_SW_ACTIVE,
552 };
553 
554 enum iwl_mvm_traffic_load {
555 	IWL_MVM_TRAFFIC_LOW,
556 	IWL_MVM_TRAFFIC_MEDIUM,
557 	IWL_MVM_TRAFFIC_HIGH,
558 };
559 
560 DECLARE_EWMA(rate, 16, 16)
561 
562 struct iwl_mvm_tcm_mac {
563 	struct {
564 		u32 pkts[IEEE80211_NUM_ACS];
565 		u32 airtime;
566 	} tx;
567 	struct {
568 		u32 pkts[IEEE80211_NUM_ACS];
569 		u32 airtime;
570 		u32 last_ampdu_ref;
571 	} rx;
572 	struct {
573 		/* track AP's transfer in client mode */
574 		u64 rx_bytes;
575 		struct ewma_rate rate;
576 		bool detected;
577 	} uapsd_nonagg_detect;
578 	bool opened_rx_ba_sessions;
579 };
580 
581 struct iwl_mvm_tcm {
582 	struct delayed_work work;
583 	spinlock_t lock; /* used when time elapsed */
584 	unsigned long ts; /* timestamp when period ends */
585 	unsigned long ll_ts;
586 	unsigned long uapsd_nonagg_ts;
587 	bool paused;
588 	struct iwl_mvm_tcm_mac data[NUM_MAC_INDEX_DRIVER];
589 	struct {
590 		u32 elapsed; /* milliseconds for this TCM period */
591 		u32 airtime[NUM_MAC_INDEX_DRIVER];
592 		enum iwl_mvm_traffic_load load[NUM_MAC_INDEX_DRIVER];
593 		enum iwl_mvm_traffic_load band_load[NUM_NL80211_BANDS];
594 		enum iwl_mvm_traffic_load global_load;
595 		bool low_latency[NUM_MAC_INDEX_DRIVER];
596 		bool change[NUM_MAC_INDEX_DRIVER];
597 	} result;
598 };
599 
600 /**
601  * struct iwl_mvm_reorder_buffer - per ra/tid/queue reorder buffer
602  * @head_sn: reorder window head sn
603  * @num_stored: number of mpdus stored in the buffer
604  * @buf_size: the reorder buffer size as set by the last addba request
605  * @queue: queue of this reorder buffer
606  * @last_amsdu: track last ASMDU SN for duplication detection
607  * @last_sub_index: track ASMDU sub frame index for duplication detection
608  * @reorder_timer: timer for frames are in the reorder buffer. For AMSDU
609  *	it is the time of last received sub-frame
610  * @removed: prevent timer re-arming
611  * @valid: reordering is valid for this queue
612  * @lock: protect reorder buffer internal state
613  * @mvm: mvm pointer, needed for frame timer context
614  * @consec_oldsn_drops: consecutive drops due to old SN
615  * @consec_oldsn_ampdu_gp2: A-MPDU GP2 timestamp to track
616  *	when to apply old SN consecutive drop workaround
617  * @consec_oldsn_prev_drop: track whether or not an MPDU
618  *	that was single/part of the previous A-MPDU was
619  *	dropped due to old SN
620  */
621 struct iwl_mvm_reorder_buffer {
622 	u16 head_sn;
623 	u16 num_stored;
624 	u16 buf_size;
625 	int queue;
626 	u16 last_amsdu;
627 	u8 last_sub_index;
628 	struct timer_list reorder_timer;
629 	bool removed;
630 	bool valid;
631 	spinlock_t lock;
632 	struct iwl_mvm *mvm;
633 	unsigned int consec_oldsn_drops;
634 	u32 consec_oldsn_ampdu_gp2;
635 	unsigned int consec_oldsn_prev_drop:1;
636 } ____cacheline_aligned_in_smp;
637 
638 /**
639  * struct _iwl_mvm_reorder_buf_entry - reorder buffer entry per-queue/per-seqno
640  * @frames: list of skbs stored
641  * @reorder_time: time the packet was stored in the reorder buffer
642  */
643 struct _iwl_mvm_reorder_buf_entry {
644 	struct sk_buff_head frames;
645 	unsigned long reorder_time;
646 };
647 
648 /* make this indirection to get the aligned thing */
649 struct iwl_mvm_reorder_buf_entry {
650 	struct _iwl_mvm_reorder_buf_entry e;
651 }
652 #ifndef __CHECKER__
653 /* sparse doesn't like this construct: "bad integer constant expression" */
654 __aligned(roundup_pow_of_two(sizeof(struct _iwl_mvm_reorder_buf_entry)))
655 #endif
656 ;
657 
658 /**
659  * struct iwl_mvm_baid_data - BA session data
660  * @sta_id: station id
661  * @tid: tid of the session
662  * @baid baid of the session
663  * @timeout: the timeout set in the addba request
664  * @entries_per_queue: # of buffers per queue, this actually gets
665  *	aligned up to avoid cache line sharing between queues
666  * @last_rx: last rx jiffies, updated only if timeout passed from last update
667  * @session_timer: timer to check if BA session expired, runs at 2 * timeout
668  * @mvm: mvm pointer, needed for timer context
669  * @reorder_buf: reorder buffer, allocated per queue
670  * @reorder_buf_data: data
671  */
672 struct iwl_mvm_baid_data {
673 	struct rcu_head rcu_head;
674 	u8 sta_id;
675 	u8 tid;
676 	u8 baid;
677 	u16 timeout;
678 	u16 entries_per_queue;
679 	unsigned long last_rx;
680 	struct timer_list session_timer;
681 	struct iwl_mvm_baid_data __rcu **rcu_ptr;
682 	struct iwl_mvm *mvm;
683 	struct iwl_mvm_reorder_buffer reorder_buf[IWL_MAX_RX_HW_QUEUES];
684 	struct iwl_mvm_reorder_buf_entry entries[];
685 };
686 
687 static inline struct iwl_mvm_baid_data *
688 iwl_mvm_baid_data_from_reorder_buf(struct iwl_mvm_reorder_buffer *buf)
689 {
690 	return (void *)((u8 *)buf -
691 			offsetof(struct iwl_mvm_baid_data, reorder_buf) -
692 			sizeof(*buf) * buf->queue);
693 }
694 
695 /*
696  * enum iwl_mvm_queue_status - queue status
697  * @IWL_MVM_QUEUE_FREE: the queue is not allocated nor reserved
698  *	Basically, this means that this queue can be used for any purpose
699  * @IWL_MVM_QUEUE_RESERVED: queue is reserved but not yet in use
700  *	This is the state of a queue that has been dedicated for some RATID
701  *	(agg'd or not), but that hasn't yet gone through the actual enablement
702  *	of iwl_mvm_enable_txq(), and therefore no traffic can go through it yet.
703  *	Note that in this state there is no requirement to already know what TID
704  *	should be used with this queue, it is just marked as a queue that will
705  *	be used, and shouldn't be allocated to anyone else.
706  * @IWL_MVM_QUEUE_READY: queue is ready to be used
707  *	This is the state of a queue that has been fully configured (including
708  *	SCD pointers, etc), has a specific RA/TID assigned to it, and can be
709  *	used to send traffic.
710  * @IWL_MVM_QUEUE_SHARED: queue is shared, or in a process of becoming shared
711  *	This is a state in which a single queue serves more than one TID, all of
712  *	which are not aggregated. Note that the queue is only associated to one
713  *	RA.
714  */
715 enum iwl_mvm_queue_status {
716 	IWL_MVM_QUEUE_FREE,
717 	IWL_MVM_QUEUE_RESERVED,
718 	IWL_MVM_QUEUE_READY,
719 	IWL_MVM_QUEUE_SHARED,
720 };
721 
722 #define IWL_MVM_DQA_QUEUE_TIMEOUT	(5 * HZ)
723 #define IWL_MVM_INVALID_QUEUE		0xFFFF
724 
725 #define IWL_MVM_NUM_CIPHERS             10
726 
727 
728 struct iwl_mvm_txq {
729 	struct list_head list;
730 	u16 txq_id;
731 	atomic_t tx_request;
732 #define IWL_MVM_TXQ_STATE_STOP_FULL	0
733 #define IWL_MVM_TXQ_STATE_STOP_REDIRECT	1
734 #define IWL_MVM_TXQ_STATE_READY		2
735 	unsigned long state;
736 };
737 
738 static inline struct iwl_mvm_txq *
739 iwl_mvm_txq_from_mac80211(struct ieee80211_txq *txq)
740 {
741 	return (void *)txq->drv_priv;
742 }
743 
744 static inline struct iwl_mvm_txq *
745 iwl_mvm_txq_from_tid(struct ieee80211_sta *sta, u8 tid)
746 {
747 	if (tid == IWL_MAX_TID_COUNT)
748 		tid = IEEE80211_NUM_TIDS;
749 
750 	return (void *)sta->txq[tid]->drv_priv;
751 }
752 
753 /**
754  * struct iwl_mvm_tvqm_txq_info - maps TVQM hw queue to tid
755  *
756  * @sta_id: sta id
757  * @txq_tid: txq tid
758  */
759 struct iwl_mvm_tvqm_txq_info {
760 	u8 sta_id;
761 	u8 txq_tid;
762 };
763 
764 struct iwl_mvm_dqa_txq_info {
765 	u8 ra_sta_id; /* The RA this queue is mapped to, if exists */
766 	bool reserved; /* Is this the TXQ reserved for a STA */
767 	u8 mac80211_ac; /* The mac80211 AC this queue is mapped to */
768 	u8 txq_tid; /* The TID "owner" of this queue*/
769 	u16 tid_bitmap; /* Bitmap of the TIDs mapped to this queue */
770 	/* Timestamp for inactivation per TID of this queue */
771 	unsigned long last_frame_time[IWL_MAX_TID_COUNT + 1];
772 	enum iwl_mvm_queue_status status;
773 };
774 
775 struct iwl_mvm {
776 	/* for logger access */
777 	struct device *dev;
778 
779 	struct iwl_trans *trans;
780 	const struct iwl_fw *fw;
781 	const struct iwl_cfg *cfg;
782 	struct iwl_phy_db *phy_db;
783 	struct ieee80211_hw *hw;
784 
785 	/* for protecting access to iwl_mvm */
786 	struct mutex mutex;
787 	struct list_head async_handlers_list;
788 	spinlock_t async_handlers_lock;
789 	struct work_struct async_handlers_wk;
790 
791 	struct work_struct roc_done_wk;
792 
793 	unsigned long init_status;
794 
795 	unsigned long status;
796 
797 	u32 queue_sync_cookie;
798 	unsigned long queue_sync_state;
799 	/*
800 	 * for beacon filtering -
801 	 * currently only one interface can be supported
802 	 */
803 	struct iwl_mvm_vif *bf_allowed_vif;
804 
805 	bool hw_registered;
806 	bool rfkill_safe_init_done;
807 
808 	u8 cca_40mhz_workaround;
809 
810 	u32 ampdu_ref;
811 	bool ampdu_toggle;
812 
813 	struct iwl_notif_wait_data notif_wait;
814 
815 	union {
816 		struct mvm_statistics_rx_v3 rx_stats_v3;
817 		struct mvm_statistics_rx rx_stats;
818 	};
819 
820 	struct {
821 		u64 rx_time;
822 		u64 tx_time;
823 		u64 on_time_rf;
824 		u64 on_time_scan;
825 	} radio_stats, accu_radio_stats;
826 
827 	struct list_head add_stream_txqs;
828 	union {
829 		struct iwl_mvm_dqa_txq_info queue_info[IWL_MAX_HW_QUEUES];
830 		struct iwl_mvm_tvqm_txq_info tvqm_info[IWL_MAX_TVQM_QUEUES];
831 	};
832 	struct work_struct add_stream_wk; /* To add streams to queues */
833 	spinlock_t add_stream_lock;
834 
835 	const char *nvm_file_name;
836 	struct iwl_nvm_data *nvm_data;
837 	struct iwl_mei_nvm *mei_nvm_data;
838 	struct iwl_mvm_csme_conn_info __rcu *csme_conn_info;
839 	bool mei_rfkill_blocked;
840 	bool mei_registered;
841 	struct work_struct sap_connected_wk;
842 
843 	/*
844 	 * NVM built based on the SAP data but that we can't free even after
845 	 * we get ownership because it contains the cfg80211's channel.
846 	 */
847 	struct iwl_nvm_data *temp_nvm_data;
848 
849 	/* NVM sections */
850 	struct iwl_nvm_section nvm_sections[NVM_MAX_NUM_SECTIONS];
851 
852 	struct iwl_fw_runtime fwrt;
853 
854 	/* EEPROM MAC addresses */
855 	struct mac_address addresses[IWL_MVM_MAX_ADDRESSES];
856 
857 	/* data related to data path */
858 	struct iwl_rx_phy_info last_phy_info;
859 	struct ieee80211_sta __rcu *fw_id_to_mac_id[IWL_MVM_STATION_COUNT_MAX];
860 	u8 rx_ba_sessions;
861 
862 	/* configured by mac80211 */
863 	u32 rts_threshold;
864 
865 	/* Scan status, cmd (pre-allocated) and auxiliary station */
866 	unsigned int scan_status;
867 	size_t scan_cmd_size;
868 	void *scan_cmd;
869 	struct iwl_mcast_filter_cmd *mcast_filter_cmd;
870 	/* For CDB this is low band scan type, for non-CDB - type. */
871 	enum iwl_mvm_scan_type scan_type;
872 	enum iwl_mvm_scan_type hb_scan_type;
873 
874 	enum iwl_mvm_sched_scan_pass_all_states sched_scan_pass_all;
875 	struct delayed_work scan_timeout_dwork;
876 
877 	/* max number of simultaneous scans the FW supports */
878 	unsigned int max_scans;
879 
880 	/* UMAC scan tracking */
881 	u32 scan_uid_status[IWL_MVM_MAX_UMAC_SCANS];
882 
883 	/* start time of last scan in TSF of the mac that requested the scan */
884 	u64 scan_start;
885 
886 	/* the vif that requested the current scan */
887 	struct iwl_mvm_vif *scan_vif;
888 
889 	/* rx chain antennas set through debugfs for the scan command */
890 	u8 scan_rx_ant;
891 
892 	/* Internal station */
893 	struct iwl_mvm_int_sta aux_sta;
894 	struct iwl_mvm_int_sta snif_sta;
895 
896 	bool last_ebs_successful;
897 
898 	u8 scan_last_antenna_idx; /* to toggle TX between antennas */
899 	u8 mgmt_last_antenna_idx;
900 
901 	/* last smart fifo state that was successfully sent to firmware */
902 	enum iwl_sf_state sf_state;
903 
904 	/*
905 	 * Leave this pointer outside the ifdef below so that it can be
906 	 * assigned without ifdef in the source code.
907 	 */
908 	struct dentry *debugfs_dir;
909 #ifdef CONFIG_IWLWIFI_DEBUGFS
910 	u32 dbgfs_sram_offset, dbgfs_sram_len;
911 	u32 dbgfs_prph_reg_addr;
912 	bool disable_power_off;
913 	bool disable_power_off_d3;
914 	bool beacon_inject_active;
915 
916 	bool scan_iter_notif_enabled;
917 
918 	struct debugfs_blob_wrapper nvm_hw_blob;
919 	struct debugfs_blob_wrapper nvm_sw_blob;
920 	struct debugfs_blob_wrapper nvm_calib_blob;
921 	struct debugfs_blob_wrapper nvm_prod_blob;
922 	struct debugfs_blob_wrapper nvm_phy_sku_blob;
923 	struct debugfs_blob_wrapper nvm_reg_blob;
924 
925 	struct iwl_mvm_frame_stats drv_rx_stats;
926 	spinlock_t drv_stats_lock;
927 	u16 dbgfs_rx_phyinfo;
928 #endif
929 
930 	struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
931 
932 	struct list_head time_event_list;
933 	spinlock_t time_event_lock;
934 
935 	/*
936 	 * A bitmap indicating the index of the key in use. The firmware
937 	 * can hold 16 keys at most. Reflect this fact.
938 	 */
939 	unsigned long fw_key_table[BITS_TO_LONGS(STA_KEY_MAX_NUM)];
940 	u8 fw_key_deleted[STA_KEY_MAX_NUM];
941 
942 	struct ieee80211_vif __rcu *vif_id_to_mac[NUM_MAC_INDEX_DRIVER];
943 
944 	/* -1 for always, 0 for never, >0 for that many times */
945 	s8 fw_restart;
946 	u8 *error_recovery_buf;
947 
948 #ifdef CONFIG_IWLWIFI_LEDS
949 	struct led_classdev led;
950 #endif
951 
952 	struct ieee80211_vif *p2p_device_vif;
953 
954 #ifdef CONFIG_PM
955 	struct wiphy_wowlan_support wowlan;
956 	int gtk_ivlen, gtk_icvlen, ptk_ivlen, ptk_icvlen;
957 
958 	/* sched scan settings for net detect */
959 	struct ieee80211_scan_ies nd_ies;
960 	struct cfg80211_match_set *nd_match_sets;
961 	int n_nd_match_sets;
962 	struct ieee80211_channel **nd_channels;
963 	int n_nd_channels;
964 	bool net_detect;
965 	u8 offload_tid;
966 #ifdef CONFIG_IWLWIFI_DEBUGFS
967 	bool d3_wake_sysassert;
968 	bool d3_test_active;
969 	u32 d3_test_pme_ptr;
970 	struct ieee80211_vif *keep_vif;
971 	u32 last_netdetect_scans; /* no. of scans in the last net-detect wake */
972 #endif
973 #endif
974 
975 	wait_queue_head_t rx_sync_waitq;
976 
977 	/* BT-Coex */
978 	struct iwl_bt_coex_profile_notif last_bt_notif;
979 	struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
980 
981 	u8 bt_tx_prio;
982 	enum iwl_bt_force_ant_mode bt_force_ant_mode;
983 
984 	/* Aux ROC */
985 	struct list_head aux_roc_te_list;
986 
987 	/* Thermal Throttling and CTkill */
988 	struct iwl_mvm_tt_mgmt thermal_throttle;
989 #ifdef CONFIG_THERMAL
990 	struct iwl_mvm_thermal_device tz_device;
991 	struct iwl_mvm_cooling_device cooling_dev;
992 #endif
993 
994 	s32 temperature;	/* Celsius */
995 	/*
996 	 * Debug option to set the NIC temperature. This option makes the
997 	 * driver think this is the actual NIC temperature, and ignore the
998 	 * real temperature that is received from the fw
999 	 */
1000 	bool temperature_test;  /* Debug test temperature is enabled */
1001 
1002 	bool fw_static_smps_request;
1003 
1004 	unsigned long bt_coex_last_tcm_ts;
1005 	struct iwl_mvm_tcm tcm;
1006 
1007 	u8 uapsd_noagg_bssid_write_idx;
1008 	struct mac_address uapsd_noagg_bssids[IWL_MVM_UAPSD_NOAGG_BSSIDS_NUM]
1009 		__aligned(2);
1010 
1011 	struct iwl_time_quota_cmd last_quota_cmd;
1012 
1013 #ifdef CONFIG_NL80211_TESTMODE
1014 	u32 noa_duration;
1015 	struct ieee80211_vif *noa_vif;
1016 #endif
1017 
1018 	/* Tx queues */
1019 	u16 aux_queue;
1020 	u16 snif_queue;
1021 	u16 probe_queue;
1022 	u16 p2p_dev_queue;
1023 
1024 	/* Indicate if device power save is allowed */
1025 	u8 ps_disabled; /* u8 instead of bool to ease debugfs_create_* usage */
1026 	/* Indicate if 32Khz external clock is valid */
1027 	u32 ext_clock_valid;
1028 
1029 	/* This vif used by CSME to send / receive traffic */
1030 	struct ieee80211_vif *csme_vif;
1031 	struct ieee80211_vif __rcu *csa_vif;
1032 	struct ieee80211_vif __rcu *csa_tx_blocked_vif;
1033 	u8 csa_tx_block_bcn_timeout;
1034 
1035 	/* system time of last beacon (for AP/GO interface) */
1036 	u32 ap_last_beacon_gp2;
1037 
1038 	/* indicates that we transmitted the last beacon */
1039 	bool ibss_manager;
1040 
1041 	bool lar_regdom_set;
1042 	enum iwl_mcc_source mcc_src;
1043 
1044 	/* TDLS channel switch data */
1045 	struct {
1046 		struct delayed_work dwork;
1047 		enum iwl_mvm_tdls_cs_state state;
1048 
1049 		/*
1050 		 * Current cs sta - might be different from periodic cs peer
1051 		 * station. Value is meaningless when the cs-state is idle.
1052 		 */
1053 		u8 cur_sta_id;
1054 
1055 		/* TDLS periodic channel-switch peer */
1056 		struct {
1057 			u8 sta_id;
1058 			u8 op_class;
1059 			bool initiator; /* are we the link initiator */
1060 			struct cfg80211_chan_def chandef;
1061 			struct sk_buff *skb; /* ch sw template */
1062 			u32 ch_sw_tm_ie;
1063 
1064 			/* timestamp of last ch-sw request sent (GP2 time) */
1065 			u32 sent_timestamp;
1066 		} peer;
1067 	} tdls_cs;
1068 
1069 
1070 	u32 ciphers[IWL_MVM_NUM_CIPHERS];
1071 
1072 	struct cfg80211_ftm_responder_stats ftm_resp_stats;
1073 	struct {
1074 		struct cfg80211_pmsr_request *req;
1075 		struct wireless_dev *req_wdev;
1076 		struct list_head loc_list;
1077 		int responses[IWL_MVM_TOF_MAX_APS];
1078 		struct {
1079 			struct list_head resp;
1080 		} smooth;
1081 		struct list_head pasn_list;
1082 	} ftm_initiator;
1083 
1084 	struct list_head resp_pasn_list;
1085 
1086 	struct {
1087 		u8 range_resp;
1088 	} cmd_ver;
1089 
1090 	struct ieee80211_vif *nan_vif;
1091 	struct iwl_mvm_baid_data __rcu *baid_map[IWL_MAX_BAID];
1092 
1093 	/*
1094 	 * Drop beacons from other APs in AP mode when there are no connected
1095 	 * clients.
1096 	 */
1097 	bool drop_bcn_ap_mode;
1098 
1099 	struct delayed_work cs_tx_unblock_dwork;
1100 
1101 	/* does a monitor vif exist (only one can exist hence bool) */
1102 	bool monitor_on;
1103 
1104 	/* sniffer data to include in radiotap */
1105 	__le16 cur_aid;
1106 	u8 cur_bssid[ETH_ALEN];
1107 
1108 	unsigned long last_6ghz_passive_scan_jiffies;
1109 	unsigned long last_reset_or_resume_time_jiffies;
1110 
1111 	bool sta_remove_requires_queue_remove;
1112 
1113 	bool pldr_sync;
1114 };
1115 
1116 /* Extract MVM priv from op_mode and _hw */
1117 #define IWL_OP_MODE_GET_MVM(_iwl_op_mode)		\
1118 	((struct iwl_mvm *)(_iwl_op_mode)->op_mode_specific)
1119 
1120 #define IWL_MAC80211_GET_MVM(_hw)			\
1121 	IWL_OP_MODE_GET_MVM((struct iwl_op_mode *)((_hw)->priv))
1122 
1123 /**
1124  * enum iwl_mvm_status - MVM status bits
1125  * @IWL_MVM_STATUS_HW_RFKILL: HW RF-kill is asserted
1126  * @IWL_MVM_STATUS_HW_CTKILL: CT-kill is active
1127  * @IWL_MVM_STATUS_ROC_RUNNING: remain-on-channel is running
1128  * @IWL_MVM_STATUS_HW_RESTART_REQUESTED: HW restart was requested
1129  * @IWL_MVM_STATUS_IN_HW_RESTART: HW restart is active
1130  * @IWL_MVM_STATUS_ROC_AUX_RUNNING: AUX remain-on-channel is running
1131  * @IWL_MVM_STATUS_FIRMWARE_RUNNING: firmware is running
1132  * @IWL_MVM_STATUS_NEED_FLUSH_P2P: need to flush P2P bcast STA
1133  * @IWL_MVM_STATUS_IN_D3: in D3 (or at least about to go into it)
1134  * @IWL_MVM_STATUS_SUPPRESS_ERROR_LOG_ONCE: suppress one error log
1135  *	if this is set, when intentionally triggered
1136  * @IWL_MVM_STATUS_STARTING: starting mac,
1137  *	used to disable restart flow while in STARTING state
1138  */
1139 enum iwl_mvm_status {
1140 	IWL_MVM_STATUS_HW_RFKILL,
1141 	IWL_MVM_STATUS_HW_CTKILL,
1142 	IWL_MVM_STATUS_ROC_RUNNING,
1143 	IWL_MVM_STATUS_HW_RESTART_REQUESTED,
1144 	IWL_MVM_STATUS_IN_HW_RESTART,
1145 	IWL_MVM_STATUS_ROC_AUX_RUNNING,
1146 	IWL_MVM_STATUS_FIRMWARE_RUNNING,
1147 	IWL_MVM_STATUS_NEED_FLUSH_P2P,
1148 	IWL_MVM_STATUS_IN_D3,
1149 	IWL_MVM_STATUS_SUPPRESS_ERROR_LOG_ONCE,
1150 	IWL_MVM_STATUS_STARTING,
1151 };
1152 
1153 struct iwl_mvm_csme_conn_info {
1154 	struct rcu_head rcu_head;
1155 	struct iwl_mei_conn_info conn_info;
1156 };
1157 
1158 /* Keep track of completed init configuration */
1159 enum iwl_mvm_init_status {
1160 	IWL_MVM_INIT_STATUS_THERMAL_INIT_COMPLETE = BIT(0),
1161 	IWL_MVM_INIT_STATUS_LEDS_INIT_COMPLETE = BIT(1),
1162 };
1163 
1164 static inline bool iwl_mvm_is_radio_killed(struct iwl_mvm *mvm)
1165 {
1166 	return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status) ||
1167 	       test_bit(IWL_MVM_STATUS_HW_CTKILL, &mvm->status);
1168 }
1169 
1170 static inline bool iwl_mvm_is_radio_hw_killed(struct iwl_mvm *mvm)
1171 {
1172 	return test_bit(IWL_MVM_STATUS_HW_RFKILL, &mvm->status);
1173 }
1174 
1175 static inline bool iwl_mvm_firmware_running(struct iwl_mvm *mvm)
1176 {
1177 	return test_bit(IWL_MVM_STATUS_FIRMWARE_RUNNING, &mvm->status);
1178 }
1179 
1180 /* Must be called with rcu_read_lock() held and it can only be
1181  * released when mvmsta is not needed anymore.
1182  */
1183 static inline struct iwl_mvm_sta *
1184 iwl_mvm_sta_from_staid_rcu(struct iwl_mvm *mvm, u8 sta_id)
1185 {
1186 	struct ieee80211_sta *sta;
1187 
1188 	if (sta_id >= mvm->fw->ucode_capa.num_stations)
1189 		return NULL;
1190 
1191 	sta = rcu_dereference(mvm->fw_id_to_mac_id[sta_id]);
1192 
1193 	/* This can happen if the station has been removed right now */
1194 	if (IS_ERR_OR_NULL(sta))
1195 		return NULL;
1196 
1197 	return iwl_mvm_sta_from_mac80211(sta);
1198 }
1199 
1200 static inline struct iwl_mvm_sta *
1201 iwl_mvm_sta_from_staid_protected(struct iwl_mvm *mvm, u8 sta_id)
1202 {
1203 	struct ieee80211_sta *sta;
1204 
1205 	if (sta_id >= mvm->fw->ucode_capa.num_stations)
1206 		return NULL;
1207 
1208 	sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
1209 					lockdep_is_held(&mvm->mutex));
1210 
1211 	/* This can happen if the station has been removed right now */
1212 	if (IS_ERR_OR_NULL(sta))
1213 		return NULL;
1214 
1215 	return iwl_mvm_sta_from_mac80211(sta);
1216 }
1217 
1218 static inline struct ieee80211_vif *
1219 iwl_mvm_rcu_dereference_vif_id(struct iwl_mvm *mvm, u8 vif_id, bool rcu)
1220 {
1221 	if (WARN_ON(vif_id >= ARRAY_SIZE(mvm->vif_id_to_mac)))
1222 		return NULL;
1223 
1224 	if (rcu)
1225 		return rcu_dereference(mvm->vif_id_to_mac[vif_id]);
1226 
1227 	return rcu_dereference_protected(mvm->vif_id_to_mac[vif_id],
1228 					 lockdep_is_held(&mvm->mutex));
1229 }
1230 
1231 static inline bool iwl_mvm_is_adaptive_dwell_supported(struct iwl_mvm *mvm)
1232 {
1233 	return fw_has_api(&mvm->fw->ucode_capa,
1234 			  IWL_UCODE_TLV_API_ADAPTIVE_DWELL);
1235 }
1236 
1237 static inline bool iwl_mvm_is_adaptive_dwell_v2_supported(struct iwl_mvm *mvm)
1238 {
1239 	return fw_has_api(&mvm->fw->ucode_capa,
1240 			  IWL_UCODE_TLV_API_ADAPTIVE_DWELL_V2);
1241 }
1242 
1243 static inline bool iwl_mvm_is_adwell_hb_ap_num_supported(struct iwl_mvm *mvm)
1244 {
1245 	return fw_has_api(&mvm->fw->ucode_capa,
1246 			  IWL_UCODE_TLV_API_ADWELL_HB_DEF_N_AP);
1247 }
1248 
1249 static inline bool iwl_mvm_is_oce_supported(struct iwl_mvm *mvm)
1250 {
1251 	/* OCE should never be enabled for LMAC scan FWs */
1252 	return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_OCE);
1253 }
1254 
1255 static inline bool iwl_mvm_is_frag_ebs_supported(struct iwl_mvm *mvm)
1256 {
1257 	return fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAG_EBS);
1258 }
1259 
1260 static inline bool iwl_mvm_is_short_beacon_notif_supported(struct iwl_mvm *mvm)
1261 {
1262 	return fw_has_api(&mvm->fw->ucode_capa,
1263 			  IWL_UCODE_TLV_API_SHORT_BEACON_NOTIF);
1264 }
1265 
1266 static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
1267 {
1268 	return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
1269 	       (queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
1270 }
1271 
1272 static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
1273 {
1274 	return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
1275 	       (queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
1276 }
1277 
1278 static inline bool iwl_mvm_is_lar_supported(struct iwl_mvm *mvm)
1279 {
1280 	bool nvm_lar = mvm->nvm_data->lar_enabled;
1281 	bool tlv_lar = fw_has_capa(&mvm->fw->ucode_capa,
1282 				   IWL_UCODE_TLV_CAPA_LAR_SUPPORT);
1283 
1284 	/*
1285 	 * Enable LAR only if it is supported by the FW (TLV) &&
1286 	 * enabled in the NVM
1287 	 */
1288 	if (mvm->cfg->nvm_type == IWL_NVM_EXT)
1289 		return nvm_lar && tlv_lar;
1290 	else
1291 		return tlv_lar;
1292 }
1293 
1294 static inline bool iwl_mvm_is_wifi_mcc_supported(struct iwl_mvm *mvm)
1295 {
1296 	return fw_has_api(&mvm->fw->ucode_capa,
1297 			  IWL_UCODE_TLV_API_WIFI_MCC_UPDATE) ||
1298 	       fw_has_capa(&mvm->fw->ucode_capa,
1299 			   IWL_UCODE_TLV_CAPA_LAR_MULTI_MCC);
1300 }
1301 
1302 static inline bool iwl_mvm_bt_is_rrc_supported(struct iwl_mvm *mvm)
1303 {
1304 	return fw_has_capa(&mvm->fw->ucode_capa,
1305 			   IWL_UCODE_TLV_CAPA_BT_COEX_RRC) &&
1306 		IWL_MVM_BT_COEX_RRC;
1307 }
1308 
1309 static inline bool iwl_mvm_is_csum_supported(struct iwl_mvm *mvm)
1310 {
1311 	return fw_has_capa(&mvm->fw->ucode_capa,
1312 			   IWL_UCODE_TLV_CAPA_CSUM_SUPPORT) &&
1313                !IWL_MVM_HW_CSUM_DISABLE;
1314 }
1315 
1316 static inline bool iwl_mvm_is_mplut_supported(struct iwl_mvm *mvm)
1317 {
1318 	return fw_has_capa(&mvm->fw->ucode_capa,
1319 			   IWL_UCODE_TLV_CAPA_BT_MPLUT_SUPPORT) &&
1320 		IWL_MVM_BT_COEX_MPLUT;
1321 }
1322 
1323 static inline
1324 bool iwl_mvm_is_p2p_scm_uapsd_supported(struct iwl_mvm *mvm)
1325 {
1326 	return fw_has_capa(&mvm->fw->ucode_capa,
1327 			   IWL_UCODE_TLV_CAPA_P2P_SCM_UAPSD) &&
1328 		!(iwlwifi_mod_params.uapsd_disable &
1329 		  IWL_DISABLE_UAPSD_P2P_CLIENT);
1330 }
1331 
1332 static inline bool iwl_mvm_has_new_rx_api(struct iwl_mvm *mvm)
1333 {
1334 	return fw_has_capa(&mvm->fw->ucode_capa,
1335 			   IWL_UCODE_TLV_CAPA_MULTI_QUEUE_RX_SUPPORT);
1336 }
1337 
1338 static inline bool iwl_mvm_has_new_tx_api(struct iwl_mvm *mvm)
1339 {
1340 	/* TODO - replace with TLV once defined */
1341 	return mvm->trans->trans_cfg->use_tfh;
1342 }
1343 
1344 static inline bool iwl_mvm_has_unified_ucode(struct iwl_mvm *mvm)
1345 {
1346 	/* TODO - better define this */
1347 	return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1348 }
1349 
1350 static inline bool iwl_mvm_is_cdb_supported(struct iwl_mvm *mvm)
1351 {
1352 	/*
1353 	 * TODO:
1354 	 * The issue of how to determine CDB APIs and usage is still not fully
1355 	 * defined.
1356 	 * There is a compilation for CDB and non-CDB FW, but there may
1357 	 * be also runtime check.
1358 	 * For now there is a TLV for checking compilation mode, but a
1359 	 * runtime check will also have to be here - once defined.
1360 	 */
1361 	return fw_has_capa(&mvm->fw->ucode_capa,
1362 			   IWL_UCODE_TLV_CAPA_CDB_SUPPORT);
1363 }
1364 
1365 static inline bool iwl_mvm_cdb_scan_api(struct iwl_mvm *mvm)
1366 {
1367 	/*
1368 	 * TODO: should this be the same as iwl_mvm_is_cdb_supported()?
1369 	 * but then there's a little bit of code in scan that won't make
1370 	 * any sense...
1371 	 */
1372 	return mvm->trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_22000;
1373 }
1374 
1375 static inline bool iwl_mvm_is_scan_ext_chan_supported(struct iwl_mvm *mvm)
1376 {
1377 	return fw_has_api(&mvm->fw->ucode_capa,
1378 			  IWL_UCODE_TLV_API_SCAN_EXT_CHAN_VER);
1379 }
1380 
1381 
1382 static inline bool iwl_mvm_is_reduced_config_scan_supported(struct iwl_mvm *mvm)
1383 {
1384 	return fw_has_api(&mvm->fw->ucode_capa,
1385 			  IWL_UCODE_TLV_API_REDUCED_SCAN_CONFIG);
1386 }
1387 
1388 static inline bool iwl_mvm_is_band_in_rx_supported(struct iwl_mvm *mvm)
1389 {
1390 	return fw_has_api(&mvm->fw->ucode_capa,
1391 			   IWL_UCODE_TLV_API_BAND_IN_RX_DATA);
1392 }
1393 
1394 static inline bool iwl_mvm_has_new_rx_stats_api(struct iwl_mvm *mvm)
1395 {
1396 	return fw_has_api(&mvm->fw->ucode_capa,
1397 			  IWL_UCODE_TLV_API_NEW_RX_STATS);
1398 }
1399 
1400 static inline bool iwl_mvm_has_quota_low_latency(struct iwl_mvm *mvm)
1401 {
1402 	return fw_has_api(&mvm->fw->ucode_capa,
1403 			  IWL_UCODE_TLV_API_QUOTA_LOW_LATENCY);
1404 }
1405 
1406 static inline bool iwl_mvm_has_tlc_offload(const struct iwl_mvm *mvm)
1407 {
1408 	return fw_has_capa(&mvm->fw->ucode_capa,
1409 			   IWL_UCODE_TLV_CAPA_TLC_OFFLOAD);
1410 }
1411 
1412 static inline struct agg_tx_status *
1413 iwl_mvm_get_agg_status(struct iwl_mvm *mvm, void *tx_resp)
1414 {
1415 	if (iwl_mvm_has_new_tx_api(mvm))
1416 		return &((struct iwl_mvm_tx_resp *)tx_resp)->status;
1417 	else
1418 		return ((struct iwl_mvm_tx_resp_v3 *)tx_resp)->status;
1419 }
1420 
1421 static inline bool iwl_mvm_is_tt_in_fw(struct iwl_mvm *mvm)
1422 {
1423 	/* these two TLV are redundant since the responsibility to CT-kill by
1424 	 * FW happens only after we send at least one command of
1425 	 * temperature THs report.
1426 	 */
1427 	return fw_has_capa(&mvm->fw->ucode_capa,
1428 			   IWL_UCODE_TLV_CAPA_CT_KILL_BY_FW) &&
1429 	       fw_has_capa(&mvm->fw->ucode_capa,
1430 			   IWL_UCODE_TLV_CAPA_TEMP_THS_REPORT_SUPPORT);
1431 }
1432 
1433 static inline bool iwl_mvm_is_ctdp_supported(struct iwl_mvm *mvm)
1434 {
1435 	return fw_has_capa(&mvm->fw->ucode_capa,
1436 			   IWL_UCODE_TLV_CAPA_CTDP_SUPPORT);
1437 }
1438 
1439 extern const u8 iwl_mvm_ac_to_tx_fifo[];
1440 extern const u8 iwl_mvm_ac_to_gen2_tx_fifo[];
1441 
1442 static inline u8 iwl_mvm_mac_ac_to_tx_fifo(struct iwl_mvm *mvm,
1443 					   enum ieee80211_ac_numbers ac)
1444 {
1445 	return iwl_mvm_has_new_tx_api(mvm) ?
1446 		iwl_mvm_ac_to_gen2_tx_fifo[ac] : iwl_mvm_ac_to_tx_fifo[ac];
1447 }
1448 
1449 struct iwl_rate_info {
1450 	u8 plcp;	/* uCode API:  IWL_RATE_6M_PLCP, etc. */
1451 	u8 plcp_siso;	/* uCode API:  IWL_RATE_SISO_6M_PLCP, etc. */
1452 	u8 plcp_mimo2;	/* uCode API:  IWL_RATE_MIMO2_6M_PLCP, etc. */
1453 	u8 plcp_mimo3;  /* uCode API:  IWL_RATE_MIMO3_6M_PLCP, etc. */
1454 	u8 ieee;	/* MAC header:  IWL_RATE_6M_IEEE, etc. */
1455 };
1456 
1457 void __iwl_mvm_mac_stop(struct iwl_mvm *mvm);
1458 int __iwl_mvm_mac_start(struct iwl_mvm *mvm);
1459 
1460 /******************
1461  * MVM Methods
1462  ******************/
1463 /* uCode */
1464 int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm);
1465 
1466 /* Utils */
1467 int iwl_mvm_legacy_hw_idx_to_mac80211_idx(u32 rate_n_flags,
1468 					  enum nl80211_band band);
1469 int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
1470 					enum nl80211_band band);
1471 void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
1472 			       enum nl80211_band band,
1473 			       struct ieee80211_tx_rate *r);
1474 void iwl_mvm_hwrate_to_tx_rate_v1(u32 rate_n_flags,
1475 				  enum nl80211_band band,
1476 				  struct ieee80211_tx_rate *r);
1477 u8 iwl_mvm_mac80211_idx_to_hwrate(const struct iwl_fw *fw, int rate_idx);
1478 u8 iwl_mvm_mac80211_ac_to_ucode_ac(enum ieee80211_ac_numbers ac);
1479 
1480 static inline void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm)
1481 {
1482 	iwl_fwrt_dump_error_logs(&mvm->fwrt);
1483 }
1484 
1485 u8 first_antenna(u8 mask);
1486 u8 iwl_mvm_next_antenna(struct iwl_mvm *mvm, u8 valid, u8 last_idx);
1487 void iwl_mvm_get_sync_time(struct iwl_mvm *mvm, int clock_type, u32 *gp2,
1488 			   u64 *boottime, ktime_t *realtime);
1489 u32 iwl_mvm_get_systime(struct iwl_mvm *mvm);
1490 
1491 /* Tx / Host Commands */
1492 int __must_check iwl_mvm_send_cmd(struct iwl_mvm *mvm,
1493 				  struct iwl_host_cmd *cmd);
1494 int __must_check iwl_mvm_send_cmd_pdu(struct iwl_mvm *mvm, u32 id,
1495 				      u32 flags, u16 len, const void *data);
1496 int __must_check iwl_mvm_send_cmd_status(struct iwl_mvm *mvm,
1497 					 struct iwl_host_cmd *cmd,
1498 					 u32 *status);
1499 int __must_check iwl_mvm_send_cmd_pdu_status(struct iwl_mvm *mvm, u32 id,
1500 					     u16 len, const void *data,
1501 					     u32 *status);
1502 int iwl_mvm_tx_skb_sta(struct iwl_mvm *mvm, struct sk_buff *skb,
1503 		       struct ieee80211_sta *sta);
1504 int iwl_mvm_tx_skb_non_sta(struct iwl_mvm *mvm, struct sk_buff *skb);
1505 void iwl_mvm_set_tx_cmd(struct iwl_mvm *mvm, struct sk_buff *skb,
1506 			struct iwl_tx_cmd *tx_cmd,
1507 			struct ieee80211_tx_info *info, u8 sta_id);
1508 void iwl_mvm_set_tx_cmd_rate(struct iwl_mvm *mvm, struct iwl_tx_cmd *tx_cmd,
1509 			    struct ieee80211_tx_info *info,
1510 			    struct ieee80211_sta *sta, __le16 fc);
1511 void iwl_mvm_mac_itxq_xmit(struct ieee80211_hw *hw, struct ieee80211_txq *txq);
1512 unsigned int iwl_mvm_max_amsdu_size(struct iwl_mvm *mvm,
1513 				    struct ieee80211_sta *sta,
1514 				    unsigned int tid);
1515 u32 iwl_mvm_tx_csum_bz(struct iwl_mvm *mvm, struct sk_buff *skb, bool amsdu);
1516 
1517 #ifdef CONFIG_IWLWIFI_DEBUG
1518 const char *iwl_mvm_get_tx_fail_reason(u32 status);
1519 #else
1520 static inline const char *iwl_mvm_get_tx_fail_reason(u32 status) { return ""; }
1521 #endif
1522 int iwl_mvm_flush_tx_path(struct iwl_mvm *mvm, u32 tfd_msk);
1523 int iwl_mvm_flush_sta(struct iwl_mvm *mvm, void *sta, bool internal);
1524 int iwl_mvm_flush_sta_tids(struct iwl_mvm *mvm, u32 sta_id, u16 tids);
1525 
1526 void iwl_mvm_async_handlers_purge(struct iwl_mvm *mvm);
1527 
1528 static inline void iwl_mvm_set_tx_cmd_ccmp(struct ieee80211_tx_info *info,
1529 					   struct iwl_tx_cmd *tx_cmd)
1530 {
1531 	struct ieee80211_key_conf *keyconf = info->control.hw_key;
1532 
1533 	tx_cmd->sec_ctl = TX_CMD_SEC_CCM;
1534 	memcpy(tx_cmd->key, keyconf->key, keyconf->keylen);
1535 }
1536 
1537 static inline void iwl_mvm_wait_for_async_handlers(struct iwl_mvm *mvm)
1538 {
1539 	flush_work(&mvm->async_handlers_wk);
1540 }
1541 
1542 /* Statistics */
1543 void iwl_mvm_handle_rx_statistics(struct iwl_mvm *mvm,
1544 				  struct iwl_rx_packet *pkt);
1545 void iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
1546 			   struct iwl_rx_cmd_buffer *rxb);
1547 int iwl_mvm_request_statistics(struct iwl_mvm *mvm, bool clear);
1548 void iwl_mvm_accu_radio_stats(struct iwl_mvm *mvm);
1549 
1550 /* NVM */
1551 int iwl_nvm_init(struct iwl_mvm *mvm);
1552 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
1553 
1554 static inline u8 iwl_mvm_get_valid_tx_ant(struct iwl_mvm *mvm)
1555 {
1556 	return mvm->nvm_data && mvm->nvm_data->valid_tx_ant ?
1557 	       mvm->fw->valid_tx_ant & mvm->nvm_data->valid_tx_ant :
1558 	       mvm->fw->valid_tx_ant;
1559 }
1560 
1561 static inline u8 iwl_mvm_get_valid_rx_ant(struct iwl_mvm *mvm)
1562 {
1563 	return mvm->nvm_data && mvm->nvm_data->valid_rx_ant ?
1564 	       mvm->fw->valid_rx_ant & mvm->nvm_data->valid_rx_ant :
1565 	       mvm->fw->valid_rx_ant;
1566 }
1567 
1568 static inline void iwl_mvm_toggle_tx_ant(struct iwl_mvm *mvm, u8 *ant)
1569 {
1570 	*ant = iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm), *ant);
1571 }
1572 
1573 static inline u32 iwl_mvm_get_phy_config(struct iwl_mvm *mvm)
1574 {
1575 	u32 phy_config = ~(FW_PHY_CFG_TX_CHAIN |
1576 			   FW_PHY_CFG_RX_CHAIN);
1577 	u32 valid_rx_ant = iwl_mvm_get_valid_rx_ant(mvm);
1578 	u32 valid_tx_ant = iwl_mvm_get_valid_tx_ant(mvm);
1579 
1580 	phy_config |= valid_tx_ant << FW_PHY_CFG_TX_CHAIN_POS |
1581 		      valid_rx_ant << FW_PHY_CFG_RX_CHAIN_POS;
1582 
1583 	return mvm->fw->phy_config & phy_config;
1584 }
1585 
1586 int iwl_mvm_up(struct iwl_mvm *mvm);
1587 int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
1588 
1589 int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm);
1590 
1591 /*
1592  * FW notifications / CMD responses handlers
1593  * Convention: iwl_mvm_rx_<NAME OF THE CMD>
1594  */
1595 void iwl_mvm_rx_mq(struct iwl_op_mode *op_mode,
1596 		   struct napi_struct *napi,
1597 		   struct iwl_rx_cmd_buffer *rxb);
1598 void iwl_mvm_rx_rx_phy_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1599 void iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct napi_struct *napi,
1600 			struct iwl_rx_cmd_buffer *rxb);
1601 void iwl_mvm_rx_mpdu_mq(struct iwl_mvm *mvm, struct napi_struct *napi,
1602 			struct iwl_rx_cmd_buffer *rxb, int queue);
1603 void iwl_mvm_rx_monitor_no_data(struct iwl_mvm *mvm, struct napi_struct *napi,
1604 				struct iwl_rx_cmd_buffer *rxb, int queue);
1605 void iwl_mvm_rx_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1606 			      struct iwl_rx_cmd_buffer *rxb, int queue);
1607 void iwl_mvm_rx_bar_frame_release(struct iwl_mvm *mvm, struct napi_struct *napi,
1608 				  struct iwl_rx_cmd_buffer *rxb, int queue);
1609 void iwl_mvm_rx_queue_notif(struct iwl_mvm *mvm, struct napi_struct *napi,
1610 			    struct iwl_rx_cmd_buffer *rxb, int queue);
1611 void iwl_mvm_rx_tx_cmd(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1612 void iwl_mvm_mfu_assert_dump_notif(struct iwl_mvm *mvm,
1613 				   struct iwl_rx_cmd_buffer *rxb);
1614 void iwl_mvm_send_recovery_cmd(struct iwl_mvm *mvm, u32 flags);
1615 void iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1616 void iwl_mvm_rx_ant_coupling_notif(struct iwl_mvm *mvm,
1617 				   struct iwl_rx_cmd_buffer *rxb);
1618 void iwl_mvm_rx_fw_error(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1619 void iwl_mvm_rx_mfuart_notif(struct iwl_mvm *mvm,
1620 			     struct iwl_rx_cmd_buffer *rxb);
1621 void iwl_mvm_rx_shared_mem_cfg_notif(struct iwl_mvm *mvm,
1622 				     struct iwl_rx_cmd_buffer *rxb);
1623 
1624 /* MVM PHY */
1625 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1626 			 struct cfg80211_chan_def *chandef,
1627 			 u8 chains_static, u8 chains_dynamic);
1628 int iwl_mvm_phy_ctxt_changed(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt,
1629 			     struct cfg80211_chan_def *chandef,
1630 			     u8 chains_static, u8 chains_dynamic);
1631 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm,
1632 			  struct iwl_mvm_phy_ctxt *ctxt);
1633 void iwl_mvm_phy_ctxt_unref(struct iwl_mvm *mvm,
1634 			    struct iwl_mvm_phy_ctxt *ctxt);
1635 int iwl_mvm_phy_ctx_count(struct iwl_mvm *mvm);
1636 u8 iwl_mvm_get_channel_width(struct cfg80211_chan_def *chandef);
1637 u8 iwl_mvm_get_ctrl_pos(struct cfg80211_chan_def *chandef);
1638 
1639 /* MAC (virtual interface) programming */
1640 int iwl_mvm_mac_ctxt_init(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1641 int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1642 int iwl_mvm_mac_ctxt_changed(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1643 			     bool force_assoc_off, const u8 *bssid_override);
1644 int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1645 int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
1646 				    struct ieee80211_vif *vif);
1647 int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm,
1648 				 struct ieee80211_vif *vif,
1649 				 struct sk_buff *beacon);
1650 int iwl_mvm_mac_ctxt_send_beacon_cmd(struct iwl_mvm *mvm,
1651 				     struct sk_buff *beacon,
1652 				     void *data, int len);
1653 u8 iwl_mvm_mac_ctxt_get_beacon_rate(struct iwl_mvm *mvm,
1654 				    struct ieee80211_tx_info *info,
1655 				    struct ieee80211_vif *vif);
1656 u16 iwl_mvm_mac_ctxt_get_beacon_flags(const struct iwl_fw *fw,
1657 				      u8 rate_idx);
1658 void iwl_mvm_mac_ctxt_set_tim(struct iwl_mvm *mvm,
1659 			      __le32 *tim_index, __le32 *tim_size,
1660 			      u8 *beacon, u32 frame_size);
1661 void iwl_mvm_rx_beacon_notif(struct iwl_mvm *mvm,
1662 			     struct iwl_rx_cmd_buffer *rxb);
1663 void iwl_mvm_rx_missed_beacons_notif(struct iwl_mvm *mvm,
1664 				     struct iwl_rx_cmd_buffer *rxb);
1665 void iwl_mvm_rx_stored_beacon_notif(struct iwl_mvm *mvm,
1666 				    struct iwl_rx_cmd_buffer *rxb);
1667 void iwl_mvm_mu_mimo_grp_notif(struct iwl_mvm *mvm,
1668 			       struct iwl_rx_cmd_buffer *rxb);
1669 void iwl_mvm_sta_pm_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1670 void iwl_mvm_window_status_notif(struct iwl_mvm *mvm,
1671 				 struct iwl_rx_cmd_buffer *rxb);
1672 void iwl_mvm_mac_ctxt_recalc_tsf_id(struct iwl_mvm *mvm,
1673 				    struct ieee80211_vif *vif);
1674 void iwl_mvm_probe_resp_data_notif(struct iwl_mvm *mvm,
1675 				   struct iwl_rx_cmd_buffer *rxb);
1676 void iwl_mvm_rx_missed_vap_notif(struct iwl_mvm *mvm,
1677 				 struct iwl_rx_cmd_buffer *rxb);
1678 void iwl_mvm_channel_switch_start_notif(struct iwl_mvm *mvm,
1679 					struct iwl_rx_cmd_buffer *rxb);
1680 void iwl_mvm_channel_switch_error_notif(struct iwl_mvm *mvm,
1681 					struct iwl_rx_cmd_buffer *rxb);
1682 /* Bindings */
1683 int iwl_mvm_binding_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1684 int iwl_mvm_binding_remove_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1685 
1686 /* Quota management */
1687 static inline size_t iwl_mvm_quota_cmd_size(struct iwl_mvm *mvm)
1688 {
1689 	return iwl_mvm_has_quota_low_latency(mvm) ?
1690 		sizeof(struct iwl_time_quota_cmd) :
1691 		sizeof(struct iwl_time_quota_cmd_v1);
1692 }
1693 
1694 static inline struct iwl_time_quota_data
1695 *iwl_mvm_quota_cmd_get_quota(struct iwl_mvm *mvm,
1696 			     struct iwl_time_quota_cmd *cmd,
1697 			     int i)
1698 {
1699 	struct iwl_time_quota_data_v1 *quotas;
1700 
1701 	if (iwl_mvm_has_quota_low_latency(mvm))
1702 		return &cmd->quotas[i];
1703 
1704 	quotas = (struct iwl_time_quota_data_v1 *)cmd->quotas;
1705 	return (struct iwl_time_quota_data *)&quotas[i];
1706 }
1707 
1708 int iwl_mvm_update_quotas(struct iwl_mvm *mvm, bool force_upload,
1709 			  struct ieee80211_vif *disabled_vif);
1710 
1711 /* Scanning */
1712 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1713 			   struct cfg80211_scan_request *req,
1714 			   struct ieee80211_scan_ies *ies);
1715 size_t iwl_mvm_scan_size(struct iwl_mvm *mvm);
1716 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify);
1717 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm);
1718 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm);
1719 void iwl_mvm_scan_timeout_wk(struct work_struct *work);
1720 
1721 /* Scheduled scan */
1722 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
1723 					 struct iwl_rx_cmd_buffer *rxb);
1724 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1725 					      struct iwl_rx_cmd_buffer *rxb);
1726 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1727 			     struct ieee80211_vif *vif,
1728 			     struct cfg80211_sched_scan_request *req,
1729 			     struct ieee80211_scan_ies *ies,
1730 			     int type);
1731 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
1732 				 struct iwl_rx_cmd_buffer *rxb);
1733 
1734 /* UMAC scan */
1735 int iwl_mvm_config_scan(struct iwl_mvm *mvm);
1736 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1737 					 struct iwl_rx_cmd_buffer *rxb);
1738 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1739 					      struct iwl_rx_cmd_buffer *rxb);
1740 
1741 /* MVM debugfs */
1742 #ifdef CONFIG_IWLWIFI_DEBUGFS
1743 void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm);
1744 void iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1745 void iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1746 #else
1747 static inline void iwl_mvm_dbgfs_register(struct iwl_mvm *mvm)
1748 {
1749 }
1750 static inline void
1751 iwl_mvm_vif_dbgfs_register(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1752 {
1753 }
1754 static inline void
1755 iwl_mvm_vif_dbgfs_clean(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1756 {
1757 }
1758 #endif /* CONFIG_IWLWIFI_DEBUGFS */
1759 
1760 /* rate scaling */
1761 int iwl_mvm_send_lq_cmd(struct iwl_mvm *mvm, struct iwl_lq_cmd *lq);
1762 void iwl_mvm_update_frame_stats(struct iwl_mvm *mvm, u32 rate, bool agg);
1763 int rs_pretty_print_rate_v1(char *buf, int bufsz, const u32 rate);
1764 void rs_update_last_rssi(struct iwl_mvm *mvm,
1765 			 struct iwl_mvm_sta *mvmsta,
1766 			 struct ieee80211_rx_status *rx_status);
1767 
1768 /* power management */
1769 int iwl_mvm_power_update_device(struct iwl_mvm *mvm);
1770 int iwl_mvm_power_update_mac(struct iwl_mvm *mvm);
1771 int iwl_mvm_power_update_ps(struct iwl_mvm *mvm);
1772 int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1773 				 char *buf, int bufsz);
1774 
1775 void iwl_mvm_power_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1776 void iwl_mvm_power_uapsd_misbehaving_ap_notif(struct iwl_mvm *mvm,
1777 					      struct iwl_rx_cmd_buffer *rxb);
1778 
1779 #ifdef CONFIG_IWLWIFI_LEDS
1780 int iwl_mvm_leds_init(struct iwl_mvm *mvm);
1781 void iwl_mvm_leds_exit(struct iwl_mvm *mvm);
1782 void iwl_mvm_leds_sync(struct iwl_mvm *mvm);
1783 #else
1784 static inline int iwl_mvm_leds_init(struct iwl_mvm *mvm)
1785 {
1786 	return 0;
1787 }
1788 static inline void iwl_mvm_leds_exit(struct iwl_mvm *mvm)
1789 {
1790 }
1791 static inline void iwl_mvm_leds_sync(struct iwl_mvm *mvm)
1792 {
1793 }
1794 #endif
1795 
1796 /* D3 (WoWLAN, NetDetect) */
1797 int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan);
1798 int iwl_mvm_resume(struct ieee80211_hw *hw);
1799 void iwl_mvm_set_wakeup(struct ieee80211_hw *hw, bool enabled);
1800 void iwl_mvm_set_rekey_data(struct ieee80211_hw *hw,
1801 			    struct ieee80211_vif *vif,
1802 			    struct cfg80211_gtk_rekey_data *data);
1803 void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
1804 			      struct ieee80211_vif *vif,
1805 			      struct inet6_dev *idev);
1806 void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
1807 				     struct ieee80211_vif *vif, int idx);
1808 extern const struct file_operations iwl_dbgfs_d3_test_ops;
1809 #ifdef CONFIG_PM
1810 void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
1811 				 struct ieee80211_vif *vif);
1812 #else
1813 static inline void
1814 iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1815 {
1816 }
1817 #endif
1818 void iwl_mvm_set_wowlan_qos_seq(struct iwl_mvm_sta *mvm_ap_sta,
1819 				struct iwl_wowlan_config_cmd *cmd);
1820 int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
1821 			       struct ieee80211_vif *vif,
1822 			       bool disable_offloading,
1823 			       bool offload_ns,
1824 			       u32 cmd_flags);
1825 
1826 /* BT Coex */
1827 int iwl_mvm_send_bt_init_conf(struct iwl_mvm *mvm);
1828 void iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
1829 			      struct iwl_rx_cmd_buffer *rxb);
1830 void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1831 			   enum ieee80211_rssi_event_data);
1832 void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
1833 u16 iwl_mvm_coex_agg_time_limit(struct iwl_mvm *mvm,
1834 				struct ieee80211_sta *sta);
1835 bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
1836 				     struct ieee80211_sta *sta);
1837 bool iwl_mvm_bt_coex_is_ant_avail(struct iwl_mvm *mvm, u8 ant);
1838 bool iwl_mvm_bt_coex_is_shared_ant_avail(struct iwl_mvm *mvm);
1839 bool iwl_mvm_bt_coex_is_tpc_allowed(struct iwl_mvm *mvm,
1840 				    enum nl80211_band band);
1841 u8 iwl_mvm_bt_coex_get_single_ant_msk(struct iwl_mvm *mvm, u8 enabled_ants);
1842 u8 iwl_mvm_bt_coex_tx_prio(struct iwl_mvm *mvm, struct ieee80211_hdr *hdr,
1843 			   struct ieee80211_tx_info *info, u8 ac);
1844 
1845 /* beacon filtering */
1846 #ifdef CONFIG_IWLWIFI_DEBUGFS
1847 void
1848 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1849 					 struct iwl_beacon_filter_cmd *cmd);
1850 #else
1851 static inline void
1852 iwl_mvm_beacon_filter_debugfs_parameters(struct ieee80211_vif *vif,
1853 					 struct iwl_beacon_filter_cmd *cmd)
1854 {}
1855 #endif
1856 int iwl_mvm_enable_beacon_filter(struct iwl_mvm *mvm,
1857 				 struct ieee80211_vif *vif,
1858 				 u32 flags);
1859 int iwl_mvm_disable_beacon_filter(struct iwl_mvm *mvm,
1860 				  struct ieee80211_vif *vif,
1861 				  u32 flags);
1862 /* SMPS */
1863 void iwl_mvm_update_smps(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1864 				enum iwl_mvm_smps_type_request req_type,
1865 				enum ieee80211_smps_mode smps_request);
1866 bool iwl_mvm_rx_diversity_allowed(struct iwl_mvm *mvm,
1867 				  struct iwl_mvm_phy_ctxt *ctxt);
1868 void iwl_mvm_apply_fw_smps_request(struct ieee80211_vif *vif);
1869 
1870 /* Low latency */
1871 int iwl_mvm_update_low_latency(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1872 			      bool low_latency,
1873 			      enum iwl_mvm_low_latency_cause cause);
1874 /* get SystemLowLatencyMode - only needed for beacon threshold? */
1875 bool iwl_mvm_low_latency(struct iwl_mvm *mvm);
1876 bool iwl_mvm_low_latency_band(struct iwl_mvm *mvm, enum nl80211_band band);
1877 void iwl_mvm_send_low_latency_cmd(struct iwl_mvm *mvm, bool low_latency,
1878 				  u16 mac_id);
1879 
1880 /* get VMACLowLatencyMode */
1881 static inline bool iwl_mvm_vif_low_latency(struct iwl_mvm_vif *mvmvif)
1882 {
1883 	/*
1884 	 * should this consider associated/active/... state?
1885 	 *
1886 	 * Normally low-latency should only be active on interfaces
1887 	 * that are active, but at least with debugfs it can also be
1888 	 * enabled on interfaces that aren't active. However, when
1889 	 * interface aren't active then they aren't added into the
1890 	 * binding, so this has no real impact. For now, just return
1891 	 * the current desired low-latency state.
1892 	 */
1893 	return mvmvif->low_latency_actual;
1894 }
1895 
1896 static inline
1897 void iwl_mvm_vif_set_low_latency(struct iwl_mvm_vif *mvmvif, bool set,
1898 				 enum iwl_mvm_low_latency_cause cause)
1899 {
1900 	u8 new_state;
1901 
1902 	if (set)
1903 		mvmvif->low_latency |= cause;
1904 	else
1905 		mvmvif->low_latency &= ~cause;
1906 
1907 	/*
1908 	 * if LOW_LATENCY_DEBUGFS_FORCE_ENABLE is enabled no changes are
1909 	 * allowed to actual mode.
1910 	 */
1911 	if (mvmvif->low_latency & LOW_LATENCY_DEBUGFS_FORCE_ENABLE &&
1912 	    cause != LOW_LATENCY_DEBUGFS_FORCE_ENABLE)
1913 		return;
1914 
1915 	if (cause == LOW_LATENCY_DEBUGFS_FORCE_ENABLE && set)
1916 		/*
1917 		 * We enter force state
1918 		 */
1919 		new_state = !!(mvmvif->low_latency &
1920 			       LOW_LATENCY_DEBUGFS_FORCE);
1921 	else
1922 		/*
1923 		 * Check if any other one set low latency
1924 		 */
1925 		new_state = !!(mvmvif->low_latency &
1926 				  ~(LOW_LATENCY_DEBUGFS_FORCE_ENABLE |
1927 				    LOW_LATENCY_DEBUGFS_FORCE));
1928 
1929 	mvmvif->low_latency_actual = new_state;
1930 }
1931 
1932 /* Return a bitmask with all the hw supported queues, except for the
1933  * command queue, which can't be flushed.
1934  */
1935 static inline u32 iwl_mvm_flushable_queues(struct iwl_mvm *mvm)
1936 {
1937 	return ((BIT(mvm->trans->trans_cfg->base_params->num_of_queues) - 1) &
1938 		~BIT(IWL_MVM_DQA_CMD_QUEUE));
1939 }
1940 
1941 void iwl_mvm_stop_device(struct iwl_mvm *mvm);
1942 
1943 /* Thermal management and CT-kill */
1944 void iwl_mvm_tt_tx_backoff(struct iwl_mvm *mvm, u32 backoff);
1945 void iwl_mvm_temp_notif(struct iwl_mvm *mvm,
1946 			struct iwl_rx_cmd_buffer *rxb);
1947 void iwl_mvm_tt_handler(struct iwl_mvm *mvm);
1948 void iwl_mvm_thermal_initialize(struct iwl_mvm *mvm, u32 min_backoff);
1949 void iwl_mvm_thermal_exit(struct iwl_mvm *mvm);
1950 void iwl_mvm_set_hw_ctkill_state(struct iwl_mvm *mvm, bool state);
1951 int iwl_mvm_get_temp(struct iwl_mvm *mvm, s32 *temp);
1952 void iwl_mvm_ct_kill_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
1953 void iwl_mvm_enter_ctkill(struct iwl_mvm *mvm);
1954 int iwl_mvm_send_temp_report_ths_cmd(struct iwl_mvm *mvm);
1955 int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
1956 
1957 #if IS_ENABLED(CONFIG_IWLMEI)
1958 
1959 /* vendor commands */
1960 void iwl_mvm_vendor_cmds_register(struct iwl_mvm *mvm);
1961 
1962 #else
1963 
1964 static inline void iwl_mvm_vendor_cmds_register(struct iwl_mvm *mvm) {}
1965 
1966 #endif
1967 
1968 /* Location Aware Regulatory */
1969 struct iwl_mcc_update_resp *
1970 iwl_mvm_update_mcc(struct iwl_mvm *mvm, const char *alpha2,
1971 		   enum iwl_mcc_source src_id);
1972 int iwl_mvm_init_mcc(struct iwl_mvm *mvm);
1973 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm *mvm,
1974 				struct iwl_rx_cmd_buffer *rxb);
1975 struct ieee80211_regdomain *iwl_mvm_get_regdomain(struct wiphy *wiphy,
1976 						  const char *alpha2,
1977 						  enum iwl_mcc_source src_id,
1978 						  bool *changed);
1979 struct ieee80211_regdomain *iwl_mvm_get_current_regdomain(struct iwl_mvm *mvm,
1980 							  bool *changed);
1981 int iwl_mvm_init_fw_regd(struct iwl_mvm *mvm);
1982 void iwl_mvm_update_changed_regdom(struct iwl_mvm *mvm);
1983 
1984 /* smart fifo */
1985 int iwl_mvm_sf_update(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1986 		      bool added_vif);
1987 
1988 /* FTM responder */
1989 int iwl_mvm_ftm_start_responder(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
1990 void iwl_mvm_ftm_restart_responder(struct iwl_mvm *mvm,
1991 				   struct ieee80211_vif *vif);
1992 void iwl_mvm_ftm_responder_stats(struct iwl_mvm *mvm,
1993 				 struct iwl_rx_cmd_buffer *rxb);
1994 int iwl_mvm_ftm_resp_remove_pasn_sta(struct iwl_mvm *mvm,
1995 				     struct ieee80211_vif *vif, u8 *addr);
1996 int iwl_mvm_ftm_respoder_add_pasn_sta(struct iwl_mvm *mvm,
1997 				      struct ieee80211_vif *vif,
1998 				      u8 *addr, u32 cipher, u8 *tk, u32 tk_len,
1999 				      u8 *hltk, u32 hltk_len);
2000 void iwl_mvm_ftm_responder_clear(struct iwl_mvm *mvm,
2001 				 struct ieee80211_vif *vif);
2002 
2003 /* FTM initiator */
2004 void iwl_mvm_ftm_restart(struct iwl_mvm *mvm);
2005 void iwl_mvm_ftm_range_resp(struct iwl_mvm *mvm,
2006 			    struct iwl_rx_cmd_buffer *rxb);
2007 void iwl_mvm_ftm_lc_notif(struct iwl_mvm *mvm,
2008 			  struct iwl_rx_cmd_buffer *rxb);
2009 int iwl_mvm_ftm_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2010 		      struct cfg80211_pmsr_request *request);
2011 void iwl_mvm_ftm_abort(struct iwl_mvm *mvm, struct cfg80211_pmsr_request *req);
2012 void iwl_mvm_ftm_initiator_smooth_config(struct iwl_mvm *mvm);
2013 void iwl_mvm_ftm_initiator_smooth_stop(struct iwl_mvm *mvm);
2014 int iwl_mvm_ftm_add_pasn_sta(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2015 			     u8 *addr, u32 cipher, u8 *tk, u32 tk_len,
2016 			     u8 *hltk, u32 hltk_len);
2017 void iwl_mvm_ftm_remove_pasn_sta(struct iwl_mvm *mvm, u8 *addr);
2018 
2019 /* TDLS */
2020 
2021 /*
2022  * We use TID 4 (VI) as a FW-used-only TID when TDLS connections are present.
2023  * This TID is marked as used vs the AP and all connected TDLS peers.
2024  */
2025 #define IWL_MVM_TDLS_FW_TID 4
2026 
2027 int iwl_mvm_tdls_sta_count(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2028 void iwl_mvm_teardown_tdls_peers(struct iwl_mvm *mvm);
2029 void iwl_mvm_recalc_tdls_state(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2030 			       bool sta_added);
2031 void iwl_mvm_mac_mgd_protect_tdls_discover(struct ieee80211_hw *hw,
2032 					   struct ieee80211_vif *vif);
2033 int iwl_mvm_tdls_channel_switch(struct ieee80211_hw *hw,
2034 				struct ieee80211_vif *vif,
2035 				struct ieee80211_sta *sta, u8 oper_class,
2036 				struct cfg80211_chan_def *chandef,
2037 				struct sk_buff *tmpl_skb, u32 ch_sw_tm_ie);
2038 void iwl_mvm_tdls_recv_channel_switch(struct ieee80211_hw *hw,
2039 				      struct ieee80211_vif *vif,
2040 				      struct ieee80211_tdls_ch_sw_params *params);
2041 void iwl_mvm_tdls_cancel_channel_switch(struct ieee80211_hw *hw,
2042 					struct ieee80211_vif *vif,
2043 					struct ieee80211_sta *sta);
2044 void iwl_mvm_rx_tdls_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb);
2045 void iwl_mvm_tdls_ch_switch_work(struct work_struct *work);
2046 
2047 void iwl_mvm_sync_rx_queues_internal(struct iwl_mvm *mvm,
2048 				     enum iwl_mvm_rxq_notif_type type,
2049 				     bool sync,
2050 				     const void *data, u32 size);
2051 void iwl_mvm_reorder_timer_expired(struct timer_list *t);
2052 struct ieee80211_vif *iwl_mvm_get_bss_vif(struct iwl_mvm *mvm);
2053 struct ieee80211_vif *iwl_mvm_get_vif_by_macid(struct iwl_mvm *mvm, u32 macid);
2054 bool iwl_mvm_is_vif_assoc(struct iwl_mvm *mvm);
2055 
2056 #define MVM_TCM_PERIOD_MSEC 500
2057 #define MVM_TCM_PERIOD (HZ * MVM_TCM_PERIOD_MSEC / 1000)
2058 #define MVM_LL_PERIOD (10 * HZ)
2059 void iwl_mvm_tcm_work(struct work_struct *work);
2060 void iwl_mvm_recalc_tcm(struct iwl_mvm *mvm);
2061 void iwl_mvm_pause_tcm(struct iwl_mvm *mvm, bool with_cancel);
2062 void iwl_mvm_resume_tcm(struct iwl_mvm *mvm);
2063 void iwl_mvm_tcm_add_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2064 void iwl_mvm_tcm_rm_vif(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
2065 u8 iwl_mvm_tcm_load_percentage(u32 airtime, u32 elapsed);
2066 
2067 void iwl_mvm_nic_restart(struct iwl_mvm *mvm, bool fw_error);
2068 unsigned int iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm,
2069 				    struct ieee80211_vif *vif,
2070 				    bool tdls, bool cmd_q);
2071 void iwl_mvm_connection_loss(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
2072 			     const char *errmsg);
2073 void iwl_mvm_event_frame_timeout_callback(struct iwl_mvm *mvm,
2074 					  struct ieee80211_vif *vif,
2075 					  const struct ieee80211_sta *sta,
2076 					  u16 tid);
2077 
2078 int iwl_mvm_sar_select_profile(struct iwl_mvm *mvm, int prof_a, int prof_b);
2079 int iwl_mvm_get_sar_geo_profile(struct iwl_mvm *mvm);
2080 int iwl_mvm_ppag_send_cmd(struct iwl_mvm *mvm);
2081 void iwl_mvm_get_acpi_tables(struct iwl_mvm *mvm);
2082 #ifdef CONFIG_IWLWIFI_DEBUGFS
2083 void iwl_mvm_sta_add_debugfs(struct ieee80211_hw *hw,
2084 			     struct ieee80211_vif *vif,
2085 			     struct ieee80211_sta *sta,
2086 			     struct dentry *dir);
2087 #endif
2088 
2089 /* new MLD related APIs */
2090 int iwl_mvm_sec_key_add(struct iwl_mvm *mvm,
2091 			struct ieee80211_vif *vif,
2092 			struct ieee80211_sta *sta,
2093 			struct ieee80211_key_conf *keyconf);
2094 int iwl_mvm_sec_key_del(struct iwl_mvm *mvm,
2095 			struct ieee80211_vif *vif,
2096 			struct ieee80211_sta *sta,
2097 			struct ieee80211_key_conf *keyconf);
2098 void iwl_mvm_sec_key_remove_ap(struct iwl_mvm *mvm,
2099 			       struct ieee80211_vif *vif);
2100 
2101 int iwl_rfi_send_config_cmd(struct iwl_mvm *mvm,
2102 			    struct iwl_rfi_lut_entry *rfi_table);
2103 struct iwl_rfi_freq_table_resp_cmd *iwl_rfi_get_freq_table(struct iwl_mvm *mvm);
2104 void iwl_rfi_deactivate_notif_handler(struct iwl_mvm *mvm,
2105 				      struct iwl_rx_cmd_buffer *rxb);
2106 
2107 static inline u8 iwl_mvm_phy_band_from_nl80211(enum nl80211_band band)
2108 {
2109 	switch (band) {
2110 	case NL80211_BAND_2GHZ:
2111 		return PHY_BAND_24;
2112 	case NL80211_BAND_5GHZ:
2113 		return PHY_BAND_5;
2114 	case NL80211_BAND_6GHZ:
2115 		return PHY_BAND_6;
2116 	default:
2117 		WARN_ONCE(1, "Unsupported band (%u)\n", band);
2118 		return PHY_BAND_5;
2119 	}
2120 }
2121 
2122 /* Channel info utils */
2123 static inline bool iwl_mvm_has_ultra_hb_channel(struct iwl_mvm *mvm)
2124 {
2125 	return fw_has_capa(&mvm->fw->ucode_capa,
2126 			   IWL_UCODE_TLV_CAPA_ULTRA_HB_CHANNELS);
2127 }
2128 
2129 static inline void *iwl_mvm_chan_info_cmd_tail(struct iwl_mvm *mvm,
2130 					       struct iwl_fw_channel_info *ci)
2131 {
2132 	return (u8 *)ci + (iwl_mvm_has_ultra_hb_channel(mvm) ?
2133 			   sizeof(struct iwl_fw_channel_info) :
2134 			   sizeof(struct iwl_fw_channel_info_v1));
2135 }
2136 
2137 static inline size_t iwl_mvm_chan_info_padding(struct iwl_mvm *mvm)
2138 {
2139 	return iwl_mvm_has_ultra_hb_channel(mvm) ? 0 :
2140 		sizeof(struct iwl_fw_channel_info) -
2141 		sizeof(struct iwl_fw_channel_info_v1);
2142 }
2143 
2144 static inline void iwl_mvm_set_chan_info(struct iwl_mvm *mvm,
2145 					 struct iwl_fw_channel_info *ci,
2146 					 u32 chan, u8 band, u8 width,
2147 					 u8 ctrl_pos)
2148 {
2149 	if (iwl_mvm_has_ultra_hb_channel(mvm)) {
2150 		ci->channel = cpu_to_le32(chan);
2151 		ci->band = band;
2152 		ci->width = width;
2153 		ci->ctrl_pos = ctrl_pos;
2154 	} else {
2155 		struct iwl_fw_channel_info_v1 *ci_v1 =
2156 					(struct iwl_fw_channel_info_v1 *)ci;
2157 
2158 		ci_v1->channel = chan;
2159 		ci_v1->band = band;
2160 		ci_v1->width = width;
2161 		ci_v1->ctrl_pos = ctrl_pos;
2162 	}
2163 }
2164 
2165 static inline void
2166 iwl_mvm_set_chan_info_chandef(struct iwl_mvm *mvm,
2167 			      struct iwl_fw_channel_info *ci,
2168 			      struct cfg80211_chan_def *chandef)
2169 {
2170 	enum nl80211_band band = chandef->chan->band;
2171 
2172 	iwl_mvm_set_chan_info(mvm, ci, chandef->chan->hw_value,
2173 			      iwl_mvm_phy_band_from_nl80211(band),
2174 			      iwl_mvm_get_channel_width(chandef),
2175 			      iwl_mvm_get_ctrl_pos(chandef));
2176 }
2177 
2178 static inline int iwl_umac_scan_get_max_profiles(const struct iwl_fw *fw)
2179 {
2180 	u8 ver = iwl_fw_lookup_cmd_ver(fw, SCAN_OFFLOAD_UPDATE_PROFILES_CMD,
2181 				       IWL_FW_CMD_VER_UNKNOWN);
2182 	return (ver == IWL_FW_CMD_VER_UNKNOWN || ver < 3) ?
2183 		IWL_SCAN_MAX_PROFILES : IWL_SCAN_MAX_PROFILES_V2;
2184 }
2185 
2186 static inline
2187 enum iwl_location_cipher iwl_mvm_cipher_to_location_cipher(u32 cipher)
2188 {
2189 	switch (cipher) {
2190 	case WLAN_CIPHER_SUITE_CCMP:
2191 		return IWL_LOCATION_CIPHER_CCMP_128;
2192 	case WLAN_CIPHER_SUITE_GCMP:
2193 		return IWL_LOCATION_CIPHER_GCMP_128;
2194 	case WLAN_CIPHER_SUITE_GCMP_256:
2195 		return IWL_LOCATION_CIPHER_GCMP_256;
2196 	default:
2197 		return IWL_LOCATION_CIPHER_INVALID;
2198 	}
2199 }
2200 
2201 struct iwl_mvm_csme_conn_info *iwl_mvm_get_csme_conn_info(struct iwl_mvm *mvm);
2202 static inline int iwl_mvm_mei_get_ownership(struct iwl_mvm *mvm)
2203 {
2204 	if (mvm->mei_registered)
2205 		return iwl_mei_get_ownership();
2206 	return 0;
2207 }
2208 
2209 static inline void iwl_mvm_mei_tx_copy_to_csme(struct iwl_mvm *mvm,
2210 					       struct sk_buff *skb,
2211 					       unsigned int ivlen)
2212 {
2213 	if (mvm->mei_registered)
2214 		iwl_mei_tx_copy_to_csme(skb, ivlen);
2215 }
2216 
2217 static inline void iwl_mvm_mei_host_disassociated(struct iwl_mvm *mvm)
2218 {
2219 	if (mvm->mei_registered)
2220 		iwl_mei_host_disassociated();
2221 }
2222 
2223 static inline void iwl_mvm_mei_device_state(struct iwl_mvm *mvm, bool up)
2224 {
2225 	if (mvm->mei_registered)
2226 		iwl_mei_device_state(up);
2227 }
2228 
2229 static inline void iwl_mvm_mei_set_sw_rfkill_state(struct iwl_mvm *mvm)
2230 {
2231 	bool sw_rfkill =
2232 		mvm->hw_registered ? rfkill_soft_blocked(mvm->hw->wiphy->rfkill) : false;
2233 
2234 	if (mvm->mei_registered)
2235 		iwl_mei_set_rfkill_state(iwl_mvm_is_radio_killed(mvm),
2236 					 sw_rfkill);
2237 }
2238 
2239 void iwl_mvm_send_roaming_forbidden_event(struct iwl_mvm *mvm,
2240 					  struct ieee80211_vif *vif,
2241 					  bool forbidden);
2242 
2243 #endif /* __IWL_MVM_H__ */
2244