1 /******************************************************************************
2  *
3  * This file is provided under a dual BSD/GPLv2 license.  When using or
4  * redistributing this file, you may do so under either license.
5  *
6  * GPL LICENSE SUMMARY
7  *
8  * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9  * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10  * Copyright(c) 2016 Intel Deutschland GmbH
11  *
12  * This program is free software; you can redistribute it and/or modify
13  * it under the terms of version 2 of the GNU General Public License as
14  * published by the Free Software Foundation.
15  *
16  * This program is distributed in the hope that it will be useful, but
17  * WITHOUT ANY WARRANTY; without even the implied warranty of
18  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
19  * General Public License for more details.
20  *
21  * You should have received a copy of the GNU General Public License
22  * along with this program; if not, write to the Free Software
23  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
24  * USA
25  *
26  * The full GNU General Public License is included in this distribution
27  * in the file called COPYING.
28  *
29  * Contact Information:
30  *  Intel Linux Wireless <linuxwifi@intel.com>
31  * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
32  *
33  * BSD LICENSE
34  *
35  * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
36  * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
37  * Copyright(c) 2016 Intel Deutschland GmbH
38  * All rights reserved.
39  *
40  * Redistribution and use in source and binary forms, with or without
41  * modification, are permitted provided that the following conditions
42  * are met:
43  *
44  *  * Redistributions of source code must retain the above copyright
45  *    notice, this list of conditions and the following disclaimer.
46  *  * Redistributions in binary form must reproduce the above copyright
47  *    notice, this list of conditions and the following disclaimer in
48  *    the documentation and/or other materials provided with the
49  *    distribution.
50  *  * Neither the name Intel Corporation nor the names of its
51  *    contributors may be used to endorse or promote products derived
52  *    from this software without specific prior written permission.
53  *
54  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
55  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
56  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
57  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
58  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
59  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
60  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
61  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
62  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
63  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
64  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
65  *
66  *****************************************************************************/
67 
68 #include <linux/etherdevice.h>
69 #include <net/mac80211.h>
70 
71 #include "mvm.h"
72 #include "fw-api-scan.h"
73 #include "iwl-io.h"
74 
75 #define IWL_DENSE_EBS_SCAN_RATIO 5
76 #define IWL_SPARSE_EBS_SCAN_RATIO 1
77 
78 enum iwl_mvm_traffic_load {
79 	IWL_MVM_TRAFFIC_LOW,
80 	IWL_MVM_TRAFFIC_MEDIUM,
81 	IWL_MVM_TRAFFIC_HIGH,
82 };
83 
84 struct iwl_mvm_scan_timing_params {
85 	u32 dwell_active;
86 	u32 dwell_passive;
87 	u32 dwell_fragmented;
88 	u32 dwell_extended;
89 	u32 suspend_time;
90 	u32 max_out_time;
91 };
92 
93 static struct iwl_mvm_scan_timing_params scan_timing[] = {
94 	[IWL_SCAN_TYPE_UNASSOC] = {
95 		.dwell_active = 10,
96 		.dwell_passive = 110,
97 		.dwell_fragmented = 44,
98 		.dwell_extended = 90,
99 		.suspend_time = 0,
100 		.max_out_time = 0,
101 	},
102 	[IWL_SCAN_TYPE_WILD] = {
103 		.dwell_active = 10,
104 		.dwell_passive = 110,
105 		.dwell_fragmented = 44,
106 		.dwell_extended = 90,
107 		.suspend_time = 30,
108 		.max_out_time = 120,
109 	},
110 	[IWL_SCAN_TYPE_MILD] = {
111 		.dwell_active = 10,
112 		.dwell_passive = 110,
113 		.dwell_fragmented = 44,
114 		.dwell_extended = 90,
115 		.suspend_time = 120,
116 		.max_out_time = 120,
117 	},
118 	[IWL_SCAN_TYPE_FRAGMENTED] = {
119 		.dwell_active = 10,
120 		.dwell_passive = 110,
121 		.dwell_fragmented = 44,
122 		.suspend_time = 95,
123 		.max_out_time = 44,
124 	},
125 };
126 
127 struct iwl_mvm_scan_params {
128 	enum iwl_mvm_scan_type type;
129 	u32 n_channels;
130 	u16 delay;
131 	int n_ssids;
132 	struct cfg80211_ssid *ssids;
133 	struct ieee80211_channel **channels;
134 	u32 flags;
135 	u8 *mac_addr;
136 	u8 *mac_addr_mask;
137 	bool no_cck;
138 	bool pass_all;
139 	int n_match_sets;
140 	struct iwl_scan_probe_req preq;
141 	struct cfg80211_match_set *match_sets;
142 	int n_scan_plans;
143 	struct cfg80211_sched_scan_plan *scan_plans;
144 };
145 
146 static u8 iwl_mvm_scan_rx_ant(struct iwl_mvm *mvm)
147 {
148 	if (mvm->scan_rx_ant != ANT_NONE)
149 		return mvm->scan_rx_ant;
150 	return iwl_mvm_get_valid_rx_ant(mvm);
151 }
152 
153 static inline __le16 iwl_mvm_scan_rx_chain(struct iwl_mvm *mvm)
154 {
155 	u16 rx_chain;
156 	u8 rx_ant;
157 
158 	rx_ant = iwl_mvm_scan_rx_ant(mvm);
159 	rx_chain = rx_ant << PHY_RX_CHAIN_VALID_POS;
160 	rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_MIMO_SEL_POS;
161 	rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_SEL_POS;
162 	rx_chain |= 0x1 << PHY_RX_CHAIN_DRIVER_FORCE_POS;
163 	return cpu_to_le16(rx_chain);
164 }
165 
166 static __le32 iwl_mvm_scan_rxon_flags(enum nl80211_band band)
167 {
168 	if (band == NL80211_BAND_2GHZ)
169 		return cpu_to_le32(PHY_BAND_24);
170 	else
171 		return cpu_to_le32(PHY_BAND_5);
172 }
173 
174 static inline __le32
175 iwl_mvm_scan_rate_n_flags(struct iwl_mvm *mvm, enum nl80211_band band,
176 			  bool no_cck)
177 {
178 	u32 tx_ant;
179 
180 	mvm->scan_last_antenna_idx =
181 		iwl_mvm_next_antenna(mvm, iwl_mvm_get_valid_tx_ant(mvm),
182 				     mvm->scan_last_antenna_idx);
183 	tx_ant = BIT(mvm->scan_last_antenna_idx) << RATE_MCS_ANT_POS;
184 
185 	if (band == NL80211_BAND_2GHZ && !no_cck)
186 		return cpu_to_le32(IWL_RATE_1M_PLCP | RATE_MCS_CCK_MSK |
187 				   tx_ant);
188 	else
189 		return cpu_to_le32(IWL_RATE_6M_PLCP | tx_ant);
190 }
191 
192 static void iwl_mvm_scan_condition_iterator(void *data, u8 *mac,
193 					    struct ieee80211_vif *vif)
194 {
195 	struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
196 	int *global_cnt = data;
197 
198 	if (vif->type != NL80211_IFTYPE_P2P_DEVICE && mvmvif->phy_ctxt &&
199 	    mvmvif->phy_ctxt->id < MAX_PHYS)
200 		*global_cnt += 1;
201 }
202 
203 static enum iwl_mvm_traffic_load iwl_mvm_get_traffic_load(struct iwl_mvm *mvm)
204 {
205 	return IWL_MVM_TRAFFIC_LOW;
206 }
207 
208 static enum
209 iwl_mvm_scan_type iwl_mvm_get_scan_type(struct iwl_mvm *mvm, bool p2p_device)
210 {
211 	int global_cnt = 0;
212 	enum iwl_mvm_traffic_load load;
213 	bool low_latency;
214 
215 	ieee80211_iterate_active_interfaces_atomic(mvm->hw,
216 					    IEEE80211_IFACE_ITER_NORMAL,
217 					    iwl_mvm_scan_condition_iterator,
218 					    &global_cnt);
219 	if (!global_cnt)
220 		return IWL_SCAN_TYPE_UNASSOC;
221 
222 	load = iwl_mvm_get_traffic_load(mvm);
223 	low_latency = iwl_mvm_low_latency(mvm);
224 
225 	if ((load == IWL_MVM_TRAFFIC_HIGH || low_latency) && !p2p_device &&
226 	    fw_has_api(&mvm->fw->ucode_capa, IWL_UCODE_TLV_API_FRAGMENTED_SCAN))
227 		return IWL_SCAN_TYPE_FRAGMENTED;
228 
229 	if (load >= IWL_MVM_TRAFFIC_MEDIUM || low_latency)
230 		return IWL_SCAN_TYPE_MILD;
231 
232 	return IWL_SCAN_TYPE_WILD;
233 }
234 
235 static inline bool iwl_mvm_rrm_scan_needed(struct iwl_mvm *mvm)
236 {
237 	/* require rrm scan whenever the fw supports it */
238 	return fw_has_capa(&mvm->fw->ucode_capa,
239 			   IWL_UCODE_TLV_CAPA_DS_PARAM_SET_IE_SUPPORT);
240 }
241 
242 static int iwl_mvm_max_scan_ie_fw_cmd_room(struct iwl_mvm *mvm)
243 {
244 	int max_probe_len;
245 
246 	max_probe_len = SCAN_OFFLOAD_PROBE_REQ_SIZE;
247 
248 	/* we create the 802.11 header and SSID element */
249 	max_probe_len -= 24 + 2;
250 
251 	/* DS parameter set element is added on 2.4GHZ band if required */
252 	if (iwl_mvm_rrm_scan_needed(mvm))
253 		max_probe_len -= 3;
254 
255 	return max_probe_len;
256 }
257 
258 int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm)
259 {
260 	int max_ie_len = iwl_mvm_max_scan_ie_fw_cmd_room(mvm);
261 
262 	/* TODO: [BUG] This function should return the maximum allowed size of
263 	 * scan IEs, however the LMAC scan api contains both 2GHZ and 5GHZ IEs
264 	 * in the same command. So the correct implementation of this function
265 	 * is just iwl_mvm_max_scan_ie_fw_cmd_room() / 2. Currently the scan
266 	 * command has only 512 bytes and it would leave us with about 240
267 	 * bytes for scan IEs, which is clearly not enough. So meanwhile
268 	 * we will report an incorrect value. This may result in a failure to
269 	 * issue a scan in unified_scan_lmac and unified_sched_scan_lmac
270 	 * functions with -ENOBUFS, if a large enough probe will be provided.
271 	 */
272 	return max_ie_len;
273 }
274 
275 static u8 *iwl_mvm_dump_channel_list(struct iwl_scan_results_notif *res,
276 				     int num_res, u8 *buf, size_t buf_size)
277 {
278 	int i;
279 	u8 *pos = buf, *end = buf + buf_size;
280 
281 	for (i = 0; pos < end && i < num_res; i++)
282 		pos += snprintf(pos, end - pos, " %u", res[i].channel);
283 
284 	/* terminate the string in case the buffer was too short */
285 	*(buf + buf_size - 1) = '\0';
286 
287 	return buf;
288 }
289 
290 void iwl_mvm_rx_lmac_scan_iter_complete_notif(struct iwl_mvm *mvm,
291 					      struct iwl_rx_cmd_buffer *rxb)
292 {
293 	struct iwl_rx_packet *pkt = rxb_addr(rxb);
294 	struct iwl_lmac_scan_complete_notif *notif = (void *)pkt->data;
295 	u8 buf[256];
296 
297 	IWL_DEBUG_SCAN(mvm,
298 		       "Scan offload iteration complete: status=0x%x scanned channels=%d channels list: %s\n",
299 		       notif->status, notif->scanned_channels,
300 		       iwl_mvm_dump_channel_list(notif->results,
301 						 notif->scanned_channels, buf,
302 						 sizeof(buf)));
303 
304 	if (mvm->sched_scan_pass_all == SCHED_SCAN_PASS_ALL_FOUND) {
305 		IWL_DEBUG_SCAN(mvm, "Pass all scheduled scan results found\n");
306 		ieee80211_sched_scan_results(mvm->hw);
307 		mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_ENABLED;
308 	}
309 }
310 
311 void iwl_mvm_rx_scan_match_found(struct iwl_mvm *mvm,
312 				 struct iwl_rx_cmd_buffer *rxb)
313 {
314 	IWL_DEBUG_SCAN(mvm, "Scheduled scan results\n");
315 	ieee80211_sched_scan_results(mvm->hw);
316 }
317 
318 static const char *iwl_mvm_ebs_status_str(enum iwl_scan_ebs_status status)
319 {
320 	switch (status) {
321 	case IWL_SCAN_EBS_SUCCESS:
322 		return "successful";
323 	case IWL_SCAN_EBS_INACTIVE:
324 		return "inactive";
325 	case IWL_SCAN_EBS_FAILED:
326 	case IWL_SCAN_EBS_CHAN_NOT_FOUND:
327 	default:
328 		return "failed";
329 	}
330 }
331 
332 void iwl_mvm_rx_lmac_scan_complete_notif(struct iwl_mvm *mvm,
333 					 struct iwl_rx_cmd_buffer *rxb)
334 {
335 	struct iwl_rx_packet *pkt = rxb_addr(rxb);
336 	struct iwl_periodic_scan_complete *scan_notif = (void *)pkt->data;
337 	bool aborted = (scan_notif->status == IWL_SCAN_OFFLOAD_ABORTED);
338 
339 	/* If this happens, the firmware has mistakenly sent an LMAC
340 	 * notification during UMAC scans -- warn and ignore it.
341 	 */
342 	if (WARN_ON_ONCE(fw_has_capa(&mvm->fw->ucode_capa,
343 				     IWL_UCODE_TLV_CAPA_UMAC_SCAN)))
344 		return;
345 
346 	/* scan status must be locked for proper checking */
347 	lockdep_assert_held(&mvm->mutex);
348 
349 	/* We first check if we were stopping a scan, in which case we
350 	 * just clear the stopping flag.  Then we check if it was a
351 	 * firmware initiated stop, in which case we need to inform
352 	 * mac80211.
353 	 * Note that we can have a stopping and a running scan
354 	 * simultaneously, but we can't have two different types of
355 	 * scans stopping or running at the same time (since LMAC
356 	 * doesn't support it).
357 	 */
358 
359 	if (mvm->scan_status & IWL_MVM_SCAN_STOPPING_SCHED) {
360 		WARN_ON_ONCE(mvm->scan_status & IWL_MVM_SCAN_STOPPING_REGULAR);
361 
362 		IWL_DEBUG_SCAN(mvm, "Scheduled scan %s, EBS status %s\n",
363 			       aborted ? "aborted" : "completed",
364 			       iwl_mvm_ebs_status_str(scan_notif->ebs_status));
365 		IWL_DEBUG_SCAN(mvm,
366 			       "Last line %d, Last iteration %d, Time after last iteration %d\n",
367 			       scan_notif->last_schedule_line,
368 			       scan_notif->last_schedule_iteration,
369 			       __le32_to_cpu(scan_notif->time_after_last_iter));
370 
371 		mvm->scan_status &= ~IWL_MVM_SCAN_STOPPING_SCHED;
372 	} else if (mvm->scan_status & IWL_MVM_SCAN_STOPPING_REGULAR) {
373 		IWL_DEBUG_SCAN(mvm, "Regular scan %s, EBS status %s\n",
374 			       aborted ? "aborted" : "completed",
375 			       iwl_mvm_ebs_status_str(scan_notif->ebs_status));
376 
377 		mvm->scan_status &= ~IWL_MVM_SCAN_STOPPING_REGULAR;
378 	} else if (mvm->scan_status & IWL_MVM_SCAN_SCHED) {
379 		WARN_ON_ONCE(mvm->scan_status & IWL_MVM_SCAN_REGULAR);
380 
381 		IWL_DEBUG_SCAN(mvm, "Scheduled scan %s, EBS status %s\n",
382 			       aborted ? "aborted" : "completed",
383 			       iwl_mvm_ebs_status_str(scan_notif->ebs_status));
384 		IWL_DEBUG_SCAN(mvm,
385 			       "Last line %d, Last iteration %d, Time after last iteration %d (FW)\n",
386 			       scan_notif->last_schedule_line,
387 			       scan_notif->last_schedule_iteration,
388 			       __le32_to_cpu(scan_notif->time_after_last_iter));
389 
390 		mvm->scan_status &= ~IWL_MVM_SCAN_SCHED;
391 		ieee80211_sched_scan_stopped(mvm->hw);
392 		mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_DISABLED;
393 	} else if (mvm->scan_status & IWL_MVM_SCAN_REGULAR) {
394 		struct cfg80211_scan_info info = {
395 			.aborted = aborted,
396 		};
397 
398 		IWL_DEBUG_SCAN(mvm, "Regular scan %s, EBS status %s (FW)\n",
399 			       aborted ? "aborted" : "completed",
400 			       iwl_mvm_ebs_status_str(scan_notif->ebs_status));
401 
402 		mvm->scan_status &= ~IWL_MVM_SCAN_REGULAR;
403 		ieee80211_scan_completed(mvm->hw, &info);
404 		iwl_mvm_unref(mvm, IWL_MVM_REF_SCAN);
405 		cancel_delayed_work(&mvm->scan_timeout_dwork);
406 	} else {
407 		IWL_ERR(mvm,
408 			"got scan complete notification but no scan is running\n");
409 	}
410 
411 	mvm->last_ebs_successful =
412 			scan_notif->ebs_status == IWL_SCAN_EBS_SUCCESS ||
413 			scan_notif->ebs_status == IWL_SCAN_EBS_INACTIVE;
414 }
415 
416 static int iwl_ssid_exist(u8 *ssid, u8 ssid_len, struct iwl_ssid_ie *ssid_list)
417 {
418 	int i;
419 
420 	for (i = 0; i < PROBE_OPTION_MAX; i++) {
421 		if (!ssid_list[i].len)
422 			break;
423 		if (ssid_list[i].len == ssid_len &&
424 		    !memcmp(ssid_list->ssid, ssid, ssid_len))
425 			return i;
426 	}
427 	return -1;
428 }
429 
430 /* We insert the SSIDs in an inverted order, because the FW will
431  * invert it back.
432  */
433 static void iwl_scan_build_ssids(struct iwl_mvm_scan_params *params,
434 				 struct iwl_ssid_ie *ssids,
435 				 u32 *ssid_bitmap)
436 {
437 	int i, j;
438 	int index;
439 
440 	/*
441 	 * copy SSIDs from match list.
442 	 * iwl_config_sched_scan_profiles() uses the order of these ssids to
443 	 * config match list.
444 	 */
445 	for (i = 0, j = params->n_match_sets - 1;
446 	     j >= 0 && i < PROBE_OPTION_MAX;
447 	     i++, j--) {
448 		/* skip empty SSID matchsets */
449 		if (!params->match_sets[j].ssid.ssid_len)
450 			continue;
451 		ssids[i].id = WLAN_EID_SSID;
452 		ssids[i].len = params->match_sets[j].ssid.ssid_len;
453 		memcpy(ssids[i].ssid, params->match_sets[j].ssid.ssid,
454 		       ssids[i].len);
455 	}
456 
457 	/* add SSIDs from scan SSID list */
458 	*ssid_bitmap = 0;
459 	for (j = params->n_ssids - 1;
460 	     j >= 0 && i < PROBE_OPTION_MAX;
461 	     i++, j--) {
462 		index = iwl_ssid_exist(params->ssids[j].ssid,
463 				       params->ssids[j].ssid_len,
464 				       ssids);
465 		if (index < 0) {
466 			ssids[i].id = WLAN_EID_SSID;
467 			ssids[i].len = params->ssids[j].ssid_len;
468 			memcpy(ssids[i].ssid, params->ssids[j].ssid,
469 			       ssids[i].len);
470 			*ssid_bitmap |= BIT(i);
471 		} else {
472 			*ssid_bitmap |= BIT(index);
473 		}
474 	}
475 }
476 
477 static int
478 iwl_mvm_config_sched_scan_profiles(struct iwl_mvm *mvm,
479 				   struct cfg80211_sched_scan_request *req)
480 {
481 	struct iwl_scan_offload_profile *profile;
482 	struct iwl_scan_offload_profile_cfg *profile_cfg;
483 	struct iwl_scan_offload_blacklist *blacklist;
484 	struct iwl_host_cmd cmd = {
485 		.id = SCAN_OFFLOAD_UPDATE_PROFILES_CMD,
486 		.len[1] = sizeof(*profile_cfg),
487 		.dataflags[0] = IWL_HCMD_DFL_NOCOPY,
488 		.dataflags[1] = IWL_HCMD_DFL_NOCOPY,
489 	};
490 	int blacklist_len;
491 	int i;
492 	int ret;
493 
494 	if (WARN_ON(req->n_match_sets > IWL_SCAN_MAX_PROFILES))
495 		return -EIO;
496 
497 	if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_SHORT_BL)
498 		blacklist_len = IWL_SCAN_SHORT_BLACKLIST_LEN;
499 	else
500 		blacklist_len = IWL_SCAN_MAX_BLACKLIST_LEN;
501 
502 	blacklist = kzalloc(sizeof(*blacklist) * blacklist_len, GFP_KERNEL);
503 	if (!blacklist)
504 		return -ENOMEM;
505 
506 	profile_cfg = kzalloc(sizeof(*profile_cfg), GFP_KERNEL);
507 	if (!profile_cfg) {
508 		ret = -ENOMEM;
509 		goto free_blacklist;
510 	}
511 
512 	cmd.data[0] = blacklist;
513 	cmd.len[0] = sizeof(*blacklist) * blacklist_len;
514 	cmd.data[1] = profile_cfg;
515 
516 	/* No blacklist configuration */
517 
518 	profile_cfg->num_profiles = req->n_match_sets;
519 	profile_cfg->active_clients = SCAN_CLIENT_SCHED_SCAN;
520 	profile_cfg->pass_match = SCAN_CLIENT_SCHED_SCAN;
521 	profile_cfg->match_notify = SCAN_CLIENT_SCHED_SCAN;
522 	if (!req->n_match_sets || !req->match_sets[0].ssid.ssid_len)
523 		profile_cfg->any_beacon_notify = SCAN_CLIENT_SCHED_SCAN;
524 
525 	for (i = 0; i < req->n_match_sets; i++) {
526 		profile = &profile_cfg->profiles[i];
527 		profile->ssid_index = i;
528 		/* Support any cipher and auth algorithm */
529 		profile->unicast_cipher = 0xff;
530 		profile->auth_alg = 0xff;
531 		profile->network_type = IWL_NETWORK_TYPE_ANY;
532 		profile->band_selection = IWL_SCAN_OFFLOAD_SELECT_ANY;
533 		profile->client_bitmap = SCAN_CLIENT_SCHED_SCAN;
534 	}
535 
536 	IWL_DEBUG_SCAN(mvm, "Sending scheduled scan profile config\n");
537 
538 	ret = iwl_mvm_send_cmd(mvm, &cmd);
539 	kfree(profile_cfg);
540 free_blacklist:
541 	kfree(blacklist);
542 
543 	return ret;
544 }
545 
546 static bool iwl_mvm_scan_pass_all(struct iwl_mvm *mvm,
547 				  struct cfg80211_sched_scan_request *req)
548 {
549 	if (req->n_match_sets && req->match_sets[0].ssid.ssid_len) {
550 		IWL_DEBUG_SCAN(mvm,
551 			       "Sending scheduled scan with filtering, n_match_sets %d\n",
552 			       req->n_match_sets);
553 		mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_DISABLED;
554 		return false;
555 	}
556 
557 	IWL_DEBUG_SCAN(mvm, "Sending Scheduled scan without filtering\n");
558 
559 	mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_ENABLED;
560 	return true;
561 }
562 
563 static int iwl_mvm_lmac_scan_abort(struct iwl_mvm *mvm)
564 {
565 	int ret;
566 	struct iwl_host_cmd cmd = {
567 		.id = SCAN_OFFLOAD_ABORT_CMD,
568 	};
569 	u32 status;
570 
571 	ret = iwl_mvm_send_cmd_status(mvm, &cmd, &status);
572 	if (ret)
573 		return ret;
574 
575 	if (status != CAN_ABORT_STATUS) {
576 		/*
577 		 * The scan abort will return 1 for success or
578 		 * 2 for "failure".  A failure condition can be
579 		 * due to simply not being in an active scan which
580 		 * can occur if we send the scan abort before the
581 		 * microcode has notified us that a scan is completed.
582 		 */
583 		IWL_DEBUG_SCAN(mvm, "SCAN OFFLOAD ABORT ret %d.\n", status);
584 		ret = -ENOENT;
585 	}
586 
587 	return ret;
588 }
589 
590 static void iwl_mvm_scan_fill_tx_cmd(struct iwl_mvm *mvm,
591 				     struct iwl_scan_req_tx_cmd *tx_cmd,
592 				     bool no_cck)
593 {
594 	tx_cmd[0].tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL |
595 					 TX_CMD_FLG_BT_DIS);
596 	tx_cmd[0].rate_n_flags = iwl_mvm_scan_rate_n_flags(mvm,
597 							   NL80211_BAND_2GHZ,
598 							   no_cck);
599 	tx_cmd[0].sta_id = mvm->aux_sta.sta_id;
600 
601 	tx_cmd[1].tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL |
602 					 TX_CMD_FLG_BT_DIS);
603 	tx_cmd[1].rate_n_flags = iwl_mvm_scan_rate_n_flags(mvm,
604 							   NL80211_BAND_5GHZ,
605 							   no_cck);
606 	tx_cmd[1].sta_id = mvm->aux_sta.sta_id;
607 }
608 
609 static void
610 iwl_mvm_lmac_scan_cfg_channels(struct iwl_mvm *mvm,
611 			       struct ieee80211_channel **channels,
612 			       int n_channels, u32 ssid_bitmap,
613 			       struct iwl_scan_req_lmac *cmd)
614 {
615 	struct iwl_scan_channel_cfg_lmac *channel_cfg = (void *)&cmd->data;
616 	int i;
617 
618 	for (i = 0; i < n_channels; i++) {
619 		channel_cfg[i].channel_num =
620 			cpu_to_le16(channels[i]->hw_value);
621 		channel_cfg[i].iter_count = cpu_to_le16(1);
622 		channel_cfg[i].iter_interval = 0;
623 		channel_cfg[i].flags =
624 			cpu_to_le32(IWL_UNIFIED_SCAN_CHANNEL_PARTIAL |
625 				    ssid_bitmap);
626 	}
627 }
628 
629 static u8 *iwl_mvm_copy_and_insert_ds_elem(struct iwl_mvm *mvm, const u8 *ies,
630 					   size_t len, u8 *const pos)
631 {
632 	static const u8 before_ds_params[] = {
633 			WLAN_EID_SSID,
634 			WLAN_EID_SUPP_RATES,
635 			WLAN_EID_REQUEST,
636 			WLAN_EID_EXT_SUPP_RATES,
637 	};
638 	size_t offs;
639 	u8 *newpos = pos;
640 
641 	if (!iwl_mvm_rrm_scan_needed(mvm)) {
642 		memcpy(newpos, ies, len);
643 		return newpos + len;
644 	}
645 
646 	offs = ieee80211_ie_split(ies, len,
647 				  before_ds_params,
648 				  ARRAY_SIZE(before_ds_params),
649 				  0);
650 
651 	memcpy(newpos, ies, offs);
652 	newpos += offs;
653 
654 	/* Add a placeholder for DS Parameter Set element */
655 	*newpos++ = WLAN_EID_DS_PARAMS;
656 	*newpos++ = 1;
657 	*newpos++ = 0;
658 
659 	memcpy(newpos, ies + offs, len - offs);
660 	newpos += len - offs;
661 
662 	return newpos;
663 }
664 
665 static void
666 iwl_mvm_build_scan_probe(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
667 			 struct ieee80211_scan_ies *ies,
668 			 struct iwl_mvm_scan_params *params)
669 {
670 	struct ieee80211_mgmt *frame = (void *)params->preq.buf;
671 	u8 *pos, *newpos;
672 	const u8 *mac_addr = params->flags & NL80211_SCAN_FLAG_RANDOM_ADDR ?
673 		params->mac_addr : NULL;
674 
675 	/*
676 	 * Unfortunately, right now the offload scan doesn't support randomising
677 	 * within the firmware, so until the firmware API is ready we implement
678 	 * it in the driver. This means that the scan iterations won't really be
679 	 * random, only when it's restarted, but at least that helps a bit.
680 	 */
681 	if (mac_addr)
682 		get_random_mask_addr(frame->sa, mac_addr,
683 				     params->mac_addr_mask);
684 	else
685 		memcpy(frame->sa, vif->addr, ETH_ALEN);
686 
687 	frame->frame_control = cpu_to_le16(IEEE80211_STYPE_PROBE_REQ);
688 	eth_broadcast_addr(frame->da);
689 	eth_broadcast_addr(frame->bssid);
690 	frame->seq_ctrl = 0;
691 
692 	pos = frame->u.probe_req.variable;
693 	*pos++ = WLAN_EID_SSID;
694 	*pos++ = 0;
695 
696 	params->preq.mac_header.offset = 0;
697 	params->preq.mac_header.len = cpu_to_le16(24 + 2);
698 
699 	/* Insert ds parameter set element on 2.4 GHz band */
700 	newpos = iwl_mvm_copy_and_insert_ds_elem(mvm,
701 						 ies->ies[NL80211_BAND_2GHZ],
702 						 ies->len[NL80211_BAND_2GHZ],
703 						 pos);
704 	params->preq.band_data[0].offset = cpu_to_le16(pos - params->preq.buf);
705 	params->preq.band_data[0].len = cpu_to_le16(newpos - pos);
706 	pos = newpos;
707 
708 	memcpy(pos, ies->ies[NL80211_BAND_5GHZ],
709 	       ies->len[NL80211_BAND_5GHZ]);
710 	params->preq.band_data[1].offset = cpu_to_le16(pos - params->preq.buf);
711 	params->preq.band_data[1].len =
712 		cpu_to_le16(ies->len[NL80211_BAND_5GHZ]);
713 	pos += ies->len[NL80211_BAND_5GHZ];
714 
715 	memcpy(pos, ies->common_ies, ies->common_ie_len);
716 	params->preq.common_data.offset = cpu_to_le16(pos - params->preq.buf);
717 	params->preq.common_data.len = cpu_to_le16(ies->common_ie_len);
718 }
719 
720 static __le32 iwl_mvm_scan_priority(struct iwl_mvm *mvm,
721 				    enum iwl_scan_priority_ext prio)
722 {
723 	if (fw_has_api(&mvm->fw->ucode_capa,
724 		       IWL_UCODE_TLV_API_EXT_SCAN_PRIORITY))
725 		return cpu_to_le32(prio);
726 
727 	if (prio <= IWL_SCAN_PRIORITY_EXT_2)
728 		return cpu_to_le32(IWL_SCAN_PRIORITY_LOW);
729 
730 	if (prio <= IWL_SCAN_PRIORITY_EXT_4)
731 		return cpu_to_le32(IWL_SCAN_PRIORITY_MEDIUM);
732 
733 	return cpu_to_le32(IWL_SCAN_PRIORITY_HIGH);
734 }
735 
736 static void iwl_mvm_scan_lmac_dwell(struct iwl_mvm *mvm,
737 				    struct iwl_scan_req_lmac *cmd,
738 				    struct iwl_mvm_scan_params *params)
739 {
740 	cmd->active_dwell = scan_timing[params->type].dwell_active;
741 	cmd->passive_dwell = scan_timing[params->type].dwell_passive;
742 	cmd->fragmented_dwell = scan_timing[params->type].dwell_fragmented;
743 	cmd->extended_dwell = scan_timing[params->type].dwell_extended;
744 	cmd->max_out_time = cpu_to_le32(scan_timing[params->type].max_out_time);
745 	cmd->suspend_time = cpu_to_le32(scan_timing[params->type].suspend_time);
746 	cmd->scan_prio = iwl_mvm_scan_priority(mvm, IWL_SCAN_PRIORITY_EXT_6);
747 }
748 
749 static inline bool iwl_mvm_scan_fits(struct iwl_mvm *mvm, int n_ssids,
750 				     struct ieee80211_scan_ies *ies,
751 				     int n_channels)
752 {
753 	return ((n_ssids <= PROBE_OPTION_MAX) &&
754 		(n_channels <= mvm->fw->ucode_capa.n_scan_channels) &
755 		(ies->common_ie_len +
756 		 ies->len[NL80211_BAND_2GHZ] +
757 		 ies->len[NL80211_BAND_5GHZ] <=
758 		 iwl_mvm_max_scan_ie_fw_cmd_room(mvm)));
759 }
760 
761 static inline bool iwl_mvm_scan_use_ebs(struct iwl_mvm *mvm,
762 					struct ieee80211_vif *vif)
763 {
764 	const struct iwl_ucode_capabilities *capa = &mvm->fw->ucode_capa;
765 
766 	/* We can only use EBS if:
767 	 *	1. the feature is supported;
768 	 *	2. the last EBS was successful;
769 	 *	3. if only single scan, the single scan EBS API is supported;
770 	 *	4. it's not a p2p find operation.
771 	 */
772 	return ((capa->flags & IWL_UCODE_TLV_FLAGS_EBS_SUPPORT) &&
773 		mvm->last_ebs_successful &&
774 		vif->type != NL80211_IFTYPE_P2P_DEVICE);
775 }
776 
777 static inline bool iwl_mvm_is_regular_scan(struct iwl_mvm_scan_params *params)
778 {
779 	return params->n_scan_plans == 1 &&
780 		params->scan_plans[0].iterations == 1;
781 }
782 
783 static int iwl_mvm_scan_lmac_flags(struct iwl_mvm *mvm,
784 				   struct iwl_mvm_scan_params *params,
785 				   struct ieee80211_vif *vif)
786 {
787 	int flags = 0;
788 
789 	if (params->n_ssids == 0)
790 		flags |= IWL_MVM_LMAC_SCAN_FLAG_PASSIVE;
791 
792 	if (params->n_ssids == 1 && params->ssids[0].ssid_len != 0)
793 		flags |= IWL_MVM_LMAC_SCAN_FLAG_PRE_CONNECTION;
794 
795 	if (params->type == IWL_SCAN_TYPE_FRAGMENTED)
796 		flags |= IWL_MVM_LMAC_SCAN_FLAG_FRAGMENTED;
797 
798 	if (iwl_mvm_rrm_scan_needed(mvm))
799 		flags |= IWL_MVM_LMAC_SCAN_FLAGS_RRM_ENABLED;
800 
801 	if (params->pass_all)
802 		flags |= IWL_MVM_LMAC_SCAN_FLAG_PASS_ALL;
803 	else
804 		flags |= IWL_MVM_LMAC_SCAN_FLAG_MATCH;
805 
806 #ifdef CONFIG_IWLWIFI_DEBUGFS
807 	if (mvm->scan_iter_notif_enabled)
808 		flags |= IWL_MVM_LMAC_SCAN_FLAG_ITER_COMPLETE;
809 #endif
810 
811 	if (mvm->sched_scan_pass_all == SCHED_SCAN_PASS_ALL_ENABLED)
812 		flags |= IWL_MVM_LMAC_SCAN_FLAG_ITER_COMPLETE;
813 
814 	if (iwl_mvm_is_regular_scan(params) &&
815 	    vif->type != NL80211_IFTYPE_P2P_DEVICE &&
816 	    params->type != IWL_SCAN_TYPE_FRAGMENTED)
817 		flags |= IWL_MVM_LMAC_SCAN_FLAG_EXTENDED_DWELL;
818 
819 	return flags;
820 }
821 
822 static int iwl_mvm_scan_lmac(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
823 			     struct iwl_mvm_scan_params *params)
824 {
825 	struct iwl_scan_req_lmac *cmd = mvm->scan_cmd;
826 	struct iwl_scan_probe_req *preq =
827 		(void *)(cmd->data + sizeof(struct iwl_scan_channel_cfg_lmac) *
828 			 mvm->fw->ucode_capa.n_scan_channels);
829 	u32 ssid_bitmap = 0;
830 	int i;
831 
832 	lockdep_assert_held(&mvm->mutex);
833 
834 	memset(cmd, 0, ksize(cmd));
835 
836 	if (WARN_ON(params->n_scan_plans > IWL_MAX_SCHED_SCAN_PLANS))
837 		return -EINVAL;
838 
839 	iwl_mvm_scan_lmac_dwell(mvm, cmd, params);
840 
841 	cmd->rx_chain_select = iwl_mvm_scan_rx_chain(mvm);
842 	cmd->iter_num = cpu_to_le32(1);
843 	cmd->n_channels = (u8)params->n_channels;
844 
845 	cmd->delay = cpu_to_le32(params->delay);
846 
847 	cmd->scan_flags = cpu_to_le32(iwl_mvm_scan_lmac_flags(mvm, params,
848 							      vif));
849 
850 	cmd->flags = iwl_mvm_scan_rxon_flags(params->channels[0]->band);
851 	cmd->filter_flags = cpu_to_le32(MAC_FILTER_ACCEPT_GRP |
852 					MAC_FILTER_IN_BEACON);
853 	iwl_mvm_scan_fill_tx_cmd(mvm, cmd->tx_cmd, params->no_cck);
854 	iwl_scan_build_ssids(params, cmd->direct_scan, &ssid_bitmap);
855 
856 	/* this API uses bits 1-20 instead of 0-19 */
857 	ssid_bitmap <<= 1;
858 
859 	for (i = 0; i < params->n_scan_plans; i++) {
860 		struct cfg80211_sched_scan_plan *scan_plan =
861 			&params->scan_plans[i];
862 
863 		cmd->schedule[i].delay =
864 			cpu_to_le16(scan_plan->interval);
865 		cmd->schedule[i].iterations = scan_plan->iterations;
866 		cmd->schedule[i].full_scan_mul = 1;
867 	}
868 
869 	/*
870 	 * If the number of iterations of the last scan plan is set to
871 	 * zero, it should run infinitely. However, this is not always the case.
872 	 * For example, when regular scan is requested the driver sets one scan
873 	 * plan with one iteration.
874 	 */
875 	if (!cmd->schedule[i - 1].iterations)
876 		cmd->schedule[i - 1].iterations = 0xff;
877 
878 	if (iwl_mvm_scan_use_ebs(mvm, vif)) {
879 		cmd->channel_opt[0].flags =
880 			cpu_to_le16(IWL_SCAN_CHANNEL_FLAG_EBS |
881 				    IWL_SCAN_CHANNEL_FLAG_EBS_ACCURATE |
882 				    IWL_SCAN_CHANNEL_FLAG_CACHE_ADD);
883 		cmd->channel_opt[0].non_ebs_ratio =
884 			cpu_to_le16(IWL_DENSE_EBS_SCAN_RATIO);
885 		cmd->channel_opt[1].flags =
886 			cpu_to_le16(IWL_SCAN_CHANNEL_FLAG_EBS |
887 				    IWL_SCAN_CHANNEL_FLAG_EBS_ACCURATE |
888 				    IWL_SCAN_CHANNEL_FLAG_CACHE_ADD);
889 		cmd->channel_opt[1].non_ebs_ratio =
890 			cpu_to_le16(IWL_SPARSE_EBS_SCAN_RATIO);
891 	}
892 
893 	iwl_mvm_lmac_scan_cfg_channels(mvm, params->channels,
894 				       params->n_channels, ssid_bitmap, cmd);
895 
896 	*preq = params->preq;
897 
898 	return 0;
899 }
900 
901 static int rate_to_scan_rate_flag(unsigned int rate)
902 {
903 	static const int rate_to_scan_rate[IWL_RATE_COUNT] = {
904 		[IWL_RATE_1M_INDEX]	= SCAN_CONFIG_RATE_1M,
905 		[IWL_RATE_2M_INDEX]	= SCAN_CONFIG_RATE_2M,
906 		[IWL_RATE_5M_INDEX]	= SCAN_CONFIG_RATE_5M,
907 		[IWL_RATE_11M_INDEX]	= SCAN_CONFIG_RATE_11M,
908 		[IWL_RATE_6M_INDEX]	= SCAN_CONFIG_RATE_6M,
909 		[IWL_RATE_9M_INDEX]	= SCAN_CONFIG_RATE_9M,
910 		[IWL_RATE_12M_INDEX]	= SCAN_CONFIG_RATE_12M,
911 		[IWL_RATE_18M_INDEX]	= SCAN_CONFIG_RATE_18M,
912 		[IWL_RATE_24M_INDEX]	= SCAN_CONFIG_RATE_24M,
913 		[IWL_RATE_36M_INDEX]	= SCAN_CONFIG_RATE_36M,
914 		[IWL_RATE_48M_INDEX]	= SCAN_CONFIG_RATE_48M,
915 		[IWL_RATE_54M_INDEX]	= SCAN_CONFIG_RATE_54M,
916 	};
917 
918 	return rate_to_scan_rate[rate];
919 }
920 
921 static __le32 iwl_mvm_scan_config_rates(struct iwl_mvm *mvm)
922 {
923 	struct ieee80211_supported_band *band;
924 	unsigned int rates = 0;
925 	int i;
926 
927 	band = &mvm->nvm_data->bands[NL80211_BAND_2GHZ];
928 	for (i = 0; i < band->n_bitrates; i++)
929 		rates |= rate_to_scan_rate_flag(band->bitrates[i].hw_value);
930 	band = &mvm->nvm_data->bands[NL80211_BAND_5GHZ];
931 	for (i = 0; i < band->n_bitrates; i++)
932 		rates |= rate_to_scan_rate_flag(band->bitrates[i].hw_value);
933 
934 	/* Set both basic rates and supported rates */
935 	rates |= SCAN_CONFIG_SUPPORTED_RATE(rates);
936 
937 	return cpu_to_le32(rates);
938 }
939 
940 int iwl_mvm_config_scan(struct iwl_mvm *mvm)
941 {
942 	struct iwl_scan_config *scan_config;
943 	struct ieee80211_supported_band *band;
944 	int num_channels =
945 		mvm->nvm_data->bands[NL80211_BAND_2GHZ].n_channels +
946 		mvm->nvm_data->bands[NL80211_BAND_5GHZ].n_channels;
947 	int ret, i, j = 0, cmd_size;
948 	struct iwl_host_cmd cmd = {
949 		.id = iwl_cmd_id(SCAN_CFG_CMD, IWL_ALWAYS_LONG_GROUP, 0),
950 	};
951 	enum iwl_mvm_scan_type type = iwl_mvm_get_scan_type(mvm, false);
952 
953 	if (WARN_ON(num_channels > mvm->fw->ucode_capa.n_scan_channels))
954 		return -ENOBUFS;
955 
956 	if (type == mvm->scan_type) {
957 		IWL_DEBUG_SCAN(mvm,
958 			       "Ignoring UMAC scan config of the same type\n");
959 		return 0;
960 	}
961 
962 	cmd_size = sizeof(*scan_config) + mvm->fw->ucode_capa.n_scan_channels;
963 
964 	scan_config = kzalloc(cmd_size, GFP_KERNEL);
965 	if (!scan_config)
966 		return -ENOMEM;
967 
968 	scan_config->flags = cpu_to_le32(SCAN_CONFIG_FLAG_ACTIVATE |
969 					 SCAN_CONFIG_FLAG_ALLOW_CHUB_REQS |
970 					 SCAN_CONFIG_FLAG_SET_TX_CHAINS |
971 					 SCAN_CONFIG_FLAG_SET_RX_CHAINS |
972 					 SCAN_CONFIG_FLAG_SET_AUX_STA_ID |
973 					 SCAN_CONFIG_FLAG_SET_ALL_TIMES |
974 					 SCAN_CONFIG_FLAG_SET_LEGACY_RATES |
975 					 SCAN_CONFIG_FLAG_SET_MAC_ADDR |
976 					 SCAN_CONFIG_FLAG_SET_CHANNEL_FLAGS|
977 					 SCAN_CONFIG_N_CHANNELS(num_channels) |
978 					 (type == IWL_SCAN_TYPE_FRAGMENTED ?
979 					  SCAN_CONFIG_FLAG_SET_FRAGMENTED :
980 					  SCAN_CONFIG_FLAG_CLEAR_FRAGMENTED));
981 	scan_config->tx_chains = cpu_to_le32(iwl_mvm_get_valid_tx_ant(mvm));
982 	scan_config->rx_chains = cpu_to_le32(iwl_mvm_scan_rx_ant(mvm));
983 	scan_config->legacy_rates = iwl_mvm_scan_config_rates(mvm);
984 	scan_config->out_of_channel_time =
985 		cpu_to_le32(scan_timing[type].max_out_time);
986 	scan_config->suspend_time = cpu_to_le32(scan_timing[type].suspend_time);
987 	scan_config->dwell_active = scan_timing[type].dwell_active;
988 	scan_config->dwell_passive = scan_timing[type].dwell_passive;
989 	scan_config->dwell_fragmented = scan_timing[type].dwell_fragmented;
990 	scan_config->dwell_extended = scan_timing[type].dwell_extended;
991 
992 	memcpy(&scan_config->mac_addr, &mvm->addresses[0].addr, ETH_ALEN);
993 
994 	scan_config->bcast_sta_id = mvm->aux_sta.sta_id;
995 	scan_config->channel_flags = IWL_CHANNEL_FLAG_EBS |
996 				     IWL_CHANNEL_FLAG_ACCURATE_EBS |
997 				     IWL_CHANNEL_FLAG_EBS_ADD |
998 				     IWL_CHANNEL_FLAG_PRE_SCAN_PASSIVE2ACTIVE;
999 
1000 	band = &mvm->nvm_data->bands[NL80211_BAND_2GHZ];
1001 	for (i = 0; i < band->n_channels; i++, j++)
1002 		scan_config->channel_array[j] = band->channels[i].hw_value;
1003 	band = &mvm->nvm_data->bands[NL80211_BAND_5GHZ];
1004 	for (i = 0; i < band->n_channels; i++, j++)
1005 		scan_config->channel_array[j] = band->channels[i].hw_value;
1006 
1007 	cmd.data[0] = scan_config;
1008 	cmd.len[0] = cmd_size;
1009 	cmd.dataflags[0] = IWL_HCMD_DFL_NOCOPY;
1010 
1011 	IWL_DEBUG_SCAN(mvm, "Sending UMAC scan config\n");
1012 
1013 	ret = iwl_mvm_send_cmd(mvm, &cmd);
1014 	if (!ret)
1015 		mvm->scan_type = type;
1016 
1017 	kfree(scan_config);
1018 	return ret;
1019 }
1020 
1021 static int iwl_mvm_scan_uid_by_status(struct iwl_mvm *mvm, int status)
1022 {
1023 	int i;
1024 
1025 	for (i = 0; i < mvm->max_scans; i++)
1026 		if (mvm->scan_uid_status[i] == status)
1027 			return i;
1028 
1029 	return -ENOENT;
1030 }
1031 
1032 static void iwl_mvm_scan_umac_dwell(struct iwl_mvm *mvm,
1033 				    struct iwl_scan_req_umac *cmd,
1034 				    struct iwl_mvm_scan_params *params)
1035 {
1036 	cmd->extended_dwell = scan_timing[params->type].dwell_extended;
1037 	cmd->active_dwell = scan_timing[params->type].dwell_active;
1038 	cmd->passive_dwell = scan_timing[params->type].dwell_passive;
1039 	cmd->fragmented_dwell = scan_timing[params->type].dwell_fragmented;
1040 	cmd->max_out_time = cpu_to_le32(scan_timing[params->type].max_out_time);
1041 	cmd->suspend_time = cpu_to_le32(scan_timing[params->type].suspend_time);
1042 	cmd->scan_priority =
1043 		iwl_mvm_scan_priority(mvm, IWL_SCAN_PRIORITY_EXT_6);
1044 
1045 	if (iwl_mvm_is_regular_scan(params))
1046 		cmd->ooc_priority =
1047 			iwl_mvm_scan_priority(mvm, IWL_SCAN_PRIORITY_EXT_6);
1048 	else
1049 		cmd->ooc_priority =
1050 			iwl_mvm_scan_priority(mvm, IWL_SCAN_PRIORITY_EXT_2);
1051 }
1052 
1053 static void
1054 iwl_mvm_umac_scan_cfg_channels(struct iwl_mvm *mvm,
1055 			       struct ieee80211_channel **channels,
1056 			       int n_channels, u32 ssid_bitmap,
1057 			       struct iwl_scan_req_umac *cmd)
1058 {
1059 	struct iwl_scan_channel_cfg_umac *channel_cfg = (void *)&cmd->data;
1060 	int i;
1061 
1062 	for (i = 0; i < n_channels; i++) {
1063 		channel_cfg[i].flags = cpu_to_le32(ssid_bitmap);
1064 		channel_cfg[i].channel_num = channels[i]->hw_value;
1065 		channel_cfg[i].iter_count = 1;
1066 		channel_cfg[i].iter_interval = 0;
1067 	}
1068 }
1069 
1070 static u32 iwl_mvm_scan_umac_flags(struct iwl_mvm *mvm,
1071 				   struct iwl_mvm_scan_params *params,
1072 				   struct ieee80211_vif *vif)
1073 {
1074 	int flags = 0;
1075 
1076 	if (params->n_ssids == 0)
1077 		flags = IWL_UMAC_SCAN_GEN_FLAGS_PASSIVE;
1078 
1079 	if (params->n_ssids == 1 && params->ssids[0].ssid_len != 0)
1080 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_PRE_CONNECT;
1081 
1082 	if (params->type == IWL_SCAN_TYPE_FRAGMENTED)
1083 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_FRAGMENTED;
1084 
1085 	if (iwl_mvm_rrm_scan_needed(mvm))
1086 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_RRM_ENABLED;
1087 
1088 	if (params->pass_all)
1089 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_PASS_ALL;
1090 	else
1091 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_MATCH;
1092 
1093 	if (!iwl_mvm_is_regular_scan(params))
1094 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_PERIODIC;
1095 
1096 #ifdef CONFIG_IWLWIFI_DEBUGFS
1097 	if (mvm->scan_iter_notif_enabled)
1098 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_ITER_COMPLETE;
1099 #endif
1100 
1101 	if (mvm->sched_scan_pass_all == SCHED_SCAN_PASS_ALL_ENABLED)
1102 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_ITER_COMPLETE;
1103 
1104 	if (iwl_mvm_is_regular_scan(params) &&
1105 	    vif->type != NL80211_IFTYPE_P2P_DEVICE &&
1106 	    params->type != IWL_SCAN_TYPE_FRAGMENTED)
1107 		flags |= IWL_UMAC_SCAN_GEN_FLAGS_EXTENDED_DWELL;
1108 
1109 	return flags;
1110 }
1111 
1112 static int iwl_mvm_scan_umac(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1113 			     struct iwl_mvm_scan_params *params,
1114 			     int type)
1115 {
1116 	struct iwl_scan_req_umac *cmd = mvm->scan_cmd;
1117 	struct iwl_scan_req_umac_tail *sec_part = (void *)&cmd->data +
1118 		sizeof(struct iwl_scan_channel_cfg_umac) *
1119 			mvm->fw->ucode_capa.n_scan_channels;
1120 	int uid, i;
1121 	u32 ssid_bitmap = 0;
1122 
1123 	lockdep_assert_held(&mvm->mutex);
1124 
1125 	if (WARN_ON(params->n_scan_plans > IWL_MAX_SCHED_SCAN_PLANS))
1126 		return -EINVAL;
1127 
1128 	uid = iwl_mvm_scan_uid_by_status(mvm, 0);
1129 	if (uid < 0)
1130 		return uid;
1131 
1132 	memset(cmd, 0, ksize(cmd));
1133 
1134 	iwl_mvm_scan_umac_dwell(mvm, cmd, params);
1135 
1136 	mvm->scan_uid_status[uid] = type;
1137 
1138 	cmd->uid = cpu_to_le32(uid);
1139 	cmd->general_flags = cpu_to_le32(iwl_mvm_scan_umac_flags(mvm, params,
1140 								 vif));
1141 
1142 	if (type == IWL_MVM_SCAN_SCHED || type == IWL_MVM_SCAN_NETDETECT)
1143 		cmd->flags = cpu_to_le32(IWL_UMAC_SCAN_FLAG_PREEMPTIVE);
1144 
1145 	if (iwl_mvm_scan_use_ebs(mvm, vif))
1146 		cmd->channel_flags = IWL_SCAN_CHANNEL_FLAG_EBS |
1147 				     IWL_SCAN_CHANNEL_FLAG_EBS_ACCURATE |
1148 				     IWL_SCAN_CHANNEL_FLAG_CACHE_ADD;
1149 
1150 	cmd->n_channels = params->n_channels;
1151 
1152 	iwl_scan_build_ssids(params, sec_part->direct_scan, &ssid_bitmap);
1153 
1154 	iwl_mvm_umac_scan_cfg_channels(mvm, params->channels,
1155 				       params->n_channels, ssid_bitmap, cmd);
1156 
1157 	for (i = 0; i < params->n_scan_plans; i++) {
1158 		struct cfg80211_sched_scan_plan *scan_plan =
1159 			&params->scan_plans[i];
1160 
1161 		sec_part->schedule[i].iter_count = scan_plan->iterations;
1162 		sec_part->schedule[i].interval =
1163 			cpu_to_le16(scan_plan->interval);
1164 	}
1165 
1166 	/*
1167 	 * If the number of iterations of the last scan plan is set to
1168 	 * zero, it should run infinitely. However, this is not always the case.
1169 	 * For example, when regular scan is requested the driver sets one scan
1170 	 * plan with one iteration.
1171 	 */
1172 	if (!sec_part->schedule[i - 1].iter_count)
1173 		sec_part->schedule[i - 1].iter_count = 0xff;
1174 
1175 	sec_part->delay = cpu_to_le16(params->delay);
1176 	sec_part->preq = params->preq;
1177 
1178 	return 0;
1179 }
1180 
1181 static int iwl_mvm_num_scans(struct iwl_mvm *mvm)
1182 {
1183 	return hweight32(mvm->scan_status & IWL_MVM_SCAN_MASK);
1184 }
1185 
1186 static int iwl_mvm_check_running_scans(struct iwl_mvm *mvm, int type)
1187 {
1188 	/* This looks a bit arbitrary, but the idea is that if we run
1189 	 * out of possible simultaneous scans and the userspace is
1190 	 * trying to run a scan type that is already running, we
1191 	 * return -EBUSY.  But if the userspace wants to start a
1192 	 * different type of scan, we stop the opposite type to make
1193 	 * space for the new request.  The reason is backwards
1194 	 * compatibility with old wpa_supplicant that wouldn't stop a
1195 	 * scheduled scan before starting a normal scan.
1196 	 */
1197 
1198 	if (iwl_mvm_num_scans(mvm) < mvm->max_scans)
1199 		return 0;
1200 
1201 	/* Use a switch, even though this is a bitmask, so that more
1202 	 * than one bits set will fall in default and we will warn.
1203 	 */
1204 	switch (type) {
1205 	case IWL_MVM_SCAN_REGULAR:
1206 		if (mvm->scan_status & IWL_MVM_SCAN_REGULAR_MASK)
1207 			return -EBUSY;
1208 		return iwl_mvm_scan_stop(mvm, IWL_MVM_SCAN_SCHED, true);
1209 	case IWL_MVM_SCAN_SCHED:
1210 		if (mvm->scan_status & IWL_MVM_SCAN_SCHED_MASK)
1211 			return -EBUSY;
1212 		return iwl_mvm_scan_stop(mvm, IWL_MVM_SCAN_REGULAR, true);
1213 	case IWL_MVM_SCAN_NETDETECT:
1214 		/* No need to stop anything for net-detect since the
1215 		 * firmware is restarted anyway.  This way, any sched
1216 		 * scans that were running will be restarted when we
1217 		 * resume.
1218 		*/
1219 		return 0;
1220 	default:
1221 		WARN_ON(1);
1222 		break;
1223 	}
1224 
1225 	return -EIO;
1226 }
1227 
1228 #define SCAN_TIMEOUT 20000
1229 
1230 void iwl_mvm_scan_timeout_wk(struct work_struct *work)
1231 {
1232 	struct delayed_work *delayed_work = to_delayed_work(work);
1233 	struct iwl_mvm *mvm = container_of(delayed_work, struct iwl_mvm,
1234 					   scan_timeout_dwork);
1235 
1236 	IWL_ERR(mvm, "regular scan timed out\n");
1237 
1238 	iwl_force_nmi(mvm->trans);
1239 }
1240 
1241 int iwl_mvm_reg_scan_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1242 			   struct cfg80211_scan_request *req,
1243 			   struct ieee80211_scan_ies *ies)
1244 {
1245 	struct iwl_host_cmd hcmd = {
1246 		.len = { iwl_mvm_scan_size(mvm), },
1247 		.data = { mvm->scan_cmd, },
1248 		.dataflags = { IWL_HCMD_DFL_NOCOPY, },
1249 	};
1250 	struct iwl_mvm_scan_params params = {};
1251 	int ret;
1252 	struct cfg80211_sched_scan_plan scan_plan = { .iterations = 1 };
1253 
1254 	lockdep_assert_held(&mvm->mutex);
1255 
1256 	if (iwl_mvm_is_lar_supported(mvm) && !mvm->lar_regdom_set) {
1257 		IWL_ERR(mvm, "scan while LAR regdomain is not set\n");
1258 		return -EBUSY;
1259 	}
1260 
1261 	ret = iwl_mvm_check_running_scans(mvm, IWL_MVM_SCAN_REGULAR);
1262 	if (ret)
1263 		return ret;
1264 
1265 	/* we should have failed registration if scan_cmd was NULL */
1266 	if (WARN_ON(!mvm->scan_cmd))
1267 		return -ENOMEM;
1268 
1269 	if (!iwl_mvm_scan_fits(mvm, req->n_ssids, ies, req->n_channels))
1270 		return -ENOBUFS;
1271 
1272 	params.n_ssids = req->n_ssids;
1273 	params.flags = req->flags;
1274 	params.n_channels = req->n_channels;
1275 	params.delay = 0;
1276 	params.ssids = req->ssids;
1277 	params.channels = req->channels;
1278 	params.mac_addr = req->mac_addr;
1279 	params.mac_addr_mask = req->mac_addr_mask;
1280 	params.no_cck = req->no_cck;
1281 	params.pass_all = true;
1282 	params.n_match_sets = 0;
1283 	params.match_sets = NULL;
1284 
1285 	params.scan_plans = &scan_plan;
1286 	params.n_scan_plans = 1;
1287 
1288 	params.type =
1289 		iwl_mvm_get_scan_type(mvm,
1290 				      vif->type == NL80211_IFTYPE_P2P_DEVICE);
1291 
1292 	iwl_mvm_build_scan_probe(mvm, vif, ies, &params);
1293 
1294 	if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN)) {
1295 		hcmd.id = iwl_cmd_id(SCAN_REQ_UMAC, IWL_ALWAYS_LONG_GROUP, 0);
1296 		ret = iwl_mvm_scan_umac(mvm, vif, &params,
1297 					IWL_MVM_SCAN_REGULAR);
1298 	} else {
1299 		hcmd.id = SCAN_OFFLOAD_REQUEST_CMD;
1300 		ret = iwl_mvm_scan_lmac(mvm, vif, &params);
1301 	}
1302 
1303 	if (ret)
1304 		return ret;
1305 
1306 	ret = iwl_mvm_send_cmd(mvm, &hcmd);
1307 	if (ret) {
1308 		/* If the scan failed, it usually means that the FW was unable
1309 		 * to allocate the time events. Warn on it, but maybe we
1310 		 * should try to send the command again with different params.
1311 		 */
1312 		IWL_ERR(mvm, "Scan failed! ret %d\n", ret);
1313 		return ret;
1314 	}
1315 
1316 	IWL_DEBUG_SCAN(mvm, "Scan request was sent successfully\n");
1317 	mvm->scan_status |= IWL_MVM_SCAN_REGULAR;
1318 	iwl_mvm_ref(mvm, IWL_MVM_REF_SCAN);
1319 
1320 	queue_delayed_work(system_wq, &mvm->scan_timeout_dwork,
1321 			   msecs_to_jiffies(SCAN_TIMEOUT));
1322 
1323 	return 0;
1324 }
1325 
1326 int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
1327 			     struct ieee80211_vif *vif,
1328 			     struct cfg80211_sched_scan_request *req,
1329 			     struct ieee80211_scan_ies *ies,
1330 			     int type)
1331 {
1332 	struct iwl_host_cmd hcmd = {
1333 		.len = { iwl_mvm_scan_size(mvm), },
1334 		.data = { mvm->scan_cmd, },
1335 		.dataflags = { IWL_HCMD_DFL_NOCOPY, },
1336 	};
1337 	struct iwl_mvm_scan_params params = {};
1338 	int ret;
1339 
1340 	lockdep_assert_held(&mvm->mutex);
1341 
1342 	if (iwl_mvm_is_lar_supported(mvm) && !mvm->lar_regdom_set) {
1343 		IWL_ERR(mvm, "sched-scan while LAR regdomain is not set\n");
1344 		return -EBUSY;
1345 	}
1346 
1347 	ret = iwl_mvm_check_running_scans(mvm, type);
1348 	if (ret)
1349 		return ret;
1350 
1351 	/* we should have failed registration if scan_cmd was NULL */
1352 	if (WARN_ON(!mvm->scan_cmd))
1353 		return -ENOMEM;
1354 
1355 	if (!iwl_mvm_scan_fits(mvm, req->n_ssids, ies, req->n_channels))
1356 		return -ENOBUFS;
1357 
1358 	params.n_ssids = req->n_ssids;
1359 	params.flags = req->flags;
1360 	params.n_channels = req->n_channels;
1361 	params.ssids = req->ssids;
1362 	params.channels = req->channels;
1363 	params.mac_addr = req->mac_addr;
1364 	params.mac_addr_mask = req->mac_addr_mask;
1365 	params.no_cck = false;
1366 	params.pass_all =  iwl_mvm_scan_pass_all(mvm, req);
1367 	params.n_match_sets = req->n_match_sets;
1368 	params.match_sets = req->match_sets;
1369 	if (!req->n_scan_plans)
1370 		return -EINVAL;
1371 
1372 	params.n_scan_plans = req->n_scan_plans;
1373 	params.scan_plans = req->scan_plans;
1374 
1375 	params.type =
1376 		iwl_mvm_get_scan_type(mvm,
1377 				      vif->type == NL80211_IFTYPE_P2P_DEVICE);
1378 
1379 	/* In theory, LMAC scans can handle a 32-bit delay, but since
1380 	 * waiting for over 18 hours to start the scan is a bit silly
1381 	 * and to keep it aligned with UMAC scans (which only support
1382 	 * 16-bit delays), trim it down to 16-bits.
1383 	 */
1384 	if (req->delay > U16_MAX) {
1385 		IWL_DEBUG_SCAN(mvm,
1386 			       "delay value is > 16-bits, set to max possible\n");
1387 		params.delay = U16_MAX;
1388 	} else {
1389 		params.delay = req->delay;
1390 	}
1391 
1392 	ret = iwl_mvm_config_sched_scan_profiles(mvm, req);
1393 	if (ret)
1394 		return ret;
1395 
1396 	iwl_mvm_build_scan_probe(mvm, vif, ies, &params);
1397 
1398 	if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN)) {
1399 		hcmd.id = iwl_cmd_id(SCAN_REQ_UMAC, IWL_ALWAYS_LONG_GROUP, 0);
1400 		ret = iwl_mvm_scan_umac(mvm, vif, &params, type);
1401 	} else {
1402 		hcmd.id = SCAN_OFFLOAD_REQUEST_CMD;
1403 		ret = iwl_mvm_scan_lmac(mvm, vif, &params);
1404 	}
1405 
1406 	if (ret)
1407 		return ret;
1408 
1409 	ret = iwl_mvm_send_cmd(mvm, &hcmd);
1410 	if (!ret) {
1411 		IWL_DEBUG_SCAN(mvm,
1412 			       "Sched scan request was sent successfully\n");
1413 		mvm->scan_status |= type;
1414 	} else {
1415 		/* If the scan failed, it usually means that the FW was unable
1416 		 * to allocate the time events. Warn on it, but maybe we
1417 		 * should try to send the command again with different params.
1418 		 */
1419 		IWL_ERR(mvm, "Sched scan failed! ret %d\n", ret);
1420 	}
1421 
1422 	return ret;
1423 }
1424 
1425 void iwl_mvm_rx_umac_scan_complete_notif(struct iwl_mvm *mvm,
1426 					 struct iwl_rx_cmd_buffer *rxb)
1427 {
1428 	struct iwl_rx_packet *pkt = rxb_addr(rxb);
1429 	struct iwl_umac_scan_complete *notif = (void *)pkt->data;
1430 	u32 uid = __le32_to_cpu(notif->uid);
1431 	bool aborted = (notif->status == IWL_SCAN_OFFLOAD_ABORTED);
1432 
1433 	if (WARN_ON(!(mvm->scan_uid_status[uid] & mvm->scan_status)))
1434 		return;
1435 
1436 	/* if the scan is already stopping, we don't need to notify mac80211 */
1437 	if (mvm->scan_uid_status[uid] == IWL_MVM_SCAN_REGULAR) {
1438 		struct cfg80211_scan_info info = {
1439 			.aborted = aborted,
1440 		};
1441 
1442 		ieee80211_scan_completed(mvm->hw, &info);
1443 		iwl_mvm_unref(mvm, IWL_MVM_REF_SCAN);
1444 		cancel_delayed_work(&mvm->scan_timeout_dwork);
1445 	} else if (mvm->scan_uid_status[uid] == IWL_MVM_SCAN_SCHED) {
1446 		ieee80211_sched_scan_stopped(mvm->hw);
1447 		mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_DISABLED;
1448 	}
1449 
1450 	mvm->scan_status &= ~mvm->scan_uid_status[uid];
1451 	IWL_DEBUG_SCAN(mvm,
1452 		       "Scan completed, uid %u type %u, status %s, EBS status %s\n",
1453 		       uid, mvm->scan_uid_status[uid],
1454 		       notif->status == IWL_SCAN_OFFLOAD_COMPLETED ?
1455 				"completed" : "aborted",
1456 		       iwl_mvm_ebs_status_str(notif->ebs_status));
1457 	IWL_DEBUG_SCAN(mvm,
1458 		       "Last line %d, Last iteration %d, Time from last iteration %d\n",
1459 		       notif->last_schedule, notif->last_iter,
1460 		       __le32_to_cpu(notif->time_from_last_iter));
1461 
1462 	if (notif->ebs_status != IWL_SCAN_EBS_SUCCESS &&
1463 	    notif->ebs_status != IWL_SCAN_EBS_INACTIVE)
1464 		mvm->last_ebs_successful = false;
1465 
1466 	mvm->scan_uid_status[uid] = 0;
1467 }
1468 
1469 void iwl_mvm_rx_umac_scan_iter_complete_notif(struct iwl_mvm *mvm,
1470 					      struct iwl_rx_cmd_buffer *rxb)
1471 {
1472 	struct iwl_rx_packet *pkt = rxb_addr(rxb);
1473 	struct iwl_umac_scan_iter_complete_notif *notif = (void *)pkt->data;
1474 	u8 buf[256];
1475 
1476 	IWL_DEBUG_SCAN(mvm,
1477 		       "UMAC Scan iteration complete: status=0x%x scanned_channels=%d channels list: %s\n",
1478 		       notif->status, notif->scanned_channels,
1479 		       iwl_mvm_dump_channel_list(notif->results,
1480 						 notif->scanned_channels, buf,
1481 						 sizeof(buf)));
1482 
1483 	if (mvm->sched_scan_pass_all == SCHED_SCAN_PASS_ALL_FOUND) {
1484 		IWL_DEBUG_SCAN(mvm, "Pass all scheduled scan results found\n");
1485 		ieee80211_sched_scan_results(mvm->hw);
1486 		mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_ENABLED;
1487 	}
1488 }
1489 
1490 static int iwl_mvm_umac_scan_abort(struct iwl_mvm *mvm, int type)
1491 {
1492 	struct iwl_umac_scan_abort cmd = {};
1493 	int uid, ret;
1494 
1495 	lockdep_assert_held(&mvm->mutex);
1496 
1497 	/* We should always get a valid index here, because we already
1498 	 * checked that this type of scan was running in the generic
1499 	 * code.
1500 	 */
1501 	uid = iwl_mvm_scan_uid_by_status(mvm, type);
1502 	if (WARN_ON_ONCE(uid < 0))
1503 		return uid;
1504 
1505 	cmd.uid = cpu_to_le32(uid);
1506 
1507 	IWL_DEBUG_SCAN(mvm, "Sending scan abort, uid %u\n", uid);
1508 
1509 	ret = iwl_mvm_send_cmd_pdu(mvm,
1510 				   iwl_cmd_id(SCAN_ABORT_UMAC,
1511 					      IWL_ALWAYS_LONG_GROUP, 0),
1512 				   0, sizeof(cmd), &cmd);
1513 	if (!ret)
1514 		mvm->scan_uid_status[uid] = type << IWL_MVM_SCAN_STOPPING_SHIFT;
1515 
1516 	return ret;
1517 }
1518 
1519 static int iwl_mvm_scan_stop_wait(struct iwl_mvm *mvm, int type)
1520 {
1521 	struct iwl_notification_wait wait_scan_done;
1522 	static const u16 scan_done_notif[] = { SCAN_COMPLETE_UMAC,
1523 					      SCAN_OFFLOAD_COMPLETE, };
1524 	int ret;
1525 
1526 	lockdep_assert_held(&mvm->mutex);
1527 
1528 	iwl_init_notification_wait(&mvm->notif_wait, &wait_scan_done,
1529 				   scan_done_notif,
1530 				   ARRAY_SIZE(scan_done_notif),
1531 				   NULL, NULL);
1532 
1533 	IWL_DEBUG_SCAN(mvm, "Preparing to stop scan, type %x\n", type);
1534 
1535 	if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN))
1536 		ret = iwl_mvm_umac_scan_abort(mvm, type);
1537 	else
1538 		ret = iwl_mvm_lmac_scan_abort(mvm);
1539 
1540 	if (ret) {
1541 		IWL_DEBUG_SCAN(mvm, "couldn't stop scan type %d\n", type);
1542 		iwl_remove_notification(&mvm->notif_wait, &wait_scan_done);
1543 		return ret;
1544 	}
1545 
1546 	ret = iwl_wait_notification(&mvm->notif_wait, &wait_scan_done, 1 * HZ);
1547 
1548 	return ret;
1549 }
1550 
1551 int iwl_mvm_scan_size(struct iwl_mvm *mvm)
1552 {
1553 	if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN))
1554 		return sizeof(struct iwl_scan_req_umac) +
1555 			sizeof(struct iwl_scan_channel_cfg_umac) *
1556 				mvm->fw->ucode_capa.n_scan_channels +
1557 			sizeof(struct iwl_scan_req_umac_tail);
1558 
1559 	return sizeof(struct iwl_scan_req_lmac) +
1560 		sizeof(struct iwl_scan_channel_cfg_lmac) *
1561 		mvm->fw->ucode_capa.n_scan_channels +
1562 		sizeof(struct iwl_scan_probe_req);
1563 }
1564 
1565 /*
1566  * This function is used in nic restart flow, to inform mac80211 about scans
1567  * that was aborted by restart flow or by an assert.
1568  */
1569 void iwl_mvm_report_scan_aborted(struct iwl_mvm *mvm)
1570 {
1571 	if (fw_has_capa(&mvm->fw->ucode_capa, IWL_UCODE_TLV_CAPA_UMAC_SCAN)) {
1572 		int uid, i;
1573 
1574 		uid = iwl_mvm_scan_uid_by_status(mvm, IWL_MVM_SCAN_REGULAR);
1575 		if (uid >= 0) {
1576 			struct cfg80211_scan_info info = {
1577 				.aborted = true,
1578 			};
1579 
1580 			ieee80211_scan_completed(mvm->hw, &info);
1581 			mvm->scan_uid_status[uid] = 0;
1582 		}
1583 		uid = iwl_mvm_scan_uid_by_status(mvm, IWL_MVM_SCAN_SCHED);
1584 		if (uid >= 0 && !mvm->restart_fw) {
1585 			ieee80211_sched_scan_stopped(mvm->hw);
1586 			mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_DISABLED;
1587 			mvm->scan_uid_status[uid] = 0;
1588 		}
1589 
1590 		/* We shouldn't have any UIDs still set.  Loop over all the
1591 		 * UIDs to make sure there's nothing left there and warn if
1592 		 * any is found.
1593 		 */
1594 		for (i = 0; i < mvm->max_scans; i++) {
1595 			if (WARN_ONCE(mvm->scan_uid_status[i],
1596 				      "UMAC scan UID %d status was not cleaned\n",
1597 				      i))
1598 				mvm->scan_uid_status[i] = 0;
1599 		}
1600 	} else {
1601 		if (mvm->scan_status & IWL_MVM_SCAN_REGULAR) {
1602 			struct cfg80211_scan_info info = {
1603 				.aborted = true,
1604 			};
1605 
1606 			ieee80211_scan_completed(mvm->hw, &info);
1607 		}
1608 
1609 		/* Sched scan will be restarted by mac80211 in
1610 		 * restart_hw, so do not report if FW is about to be
1611 		 * restarted.
1612 		 */
1613 		if ((mvm->scan_status & IWL_MVM_SCAN_SCHED) &&
1614 		    !mvm->restart_fw) {
1615 			ieee80211_sched_scan_stopped(mvm->hw);
1616 			mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_DISABLED;
1617 		}
1618 	}
1619 }
1620 
1621 int iwl_mvm_scan_stop(struct iwl_mvm *mvm, int type, bool notify)
1622 {
1623 	int ret;
1624 
1625 	if (!(mvm->scan_status & type))
1626 		return 0;
1627 
1628 	if (iwl_mvm_is_radio_killed(mvm)) {
1629 		ret = 0;
1630 		goto out;
1631 	}
1632 
1633 	ret = iwl_mvm_scan_stop_wait(mvm, type);
1634 	if (!ret)
1635 		mvm->scan_status |= type << IWL_MVM_SCAN_STOPPING_SHIFT;
1636 out:
1637 	/* Clear the scan status so the next scan requests will
1638 	 * succeed and mark the scan as stopping, so that the Rx
1639 	 * handler doesn't do anything, as the scan was stopped from
1640 	 * above.
1641 	 */
1642 	mvm->scan_status &= ~type;
1643 
1644 	if (type == IWL_MVM_SCAN_REGULAR) {
1645 		/* Since the rx handler won't do anything now, we have
1646 		 * to release the scan reference here.
1647 		 */
1648 		iwl_mvm_unref(mvm, IWL_MVM_REF_SCAN);
1649 		cancel_delayed_work(&mvm->scan_timeout_dwork);
1650 		if (notify) {
1651 			struct cfg80211_scan_info info = {
1652 				.aborted = true,
1653 			};
1654 
1655 			ieee80211_scan_completed(mvm->hw, &info);
1656 		}
1657 	} else if (notify) {
1658 		ieee80211_sched_scan_stopped(mvm->hw);
1659 		mvm->sched_scan_pass_all = SCHED_SCAN_PASS_ALL_DISABLED;
1660 	}
1661 
1662 	return ret;
1663 }
1664