xref: /openbmc/linux/drivers/net/wireless/ti/wlcore/main.c (revision 462cd772)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * This file is part of wlcore
4  *
5  * Copyright (C) 2008-2010 Nokia Corporation
6  * Copyright (C) 2011-2013 Texas Instruments Inc.
7  */
8 
9 #include <linux/module.h>
10 #include <linux/firmware.h>
11 #include <linux/etherdevice.h>
12 #include <linux/vmalloc.h>
13 #include <linux/interrupt.h>
14 #include <linux/irq.h>
15 #include <linux/pm_runtime.h>
16 #include <linux/pm_wakeirq.h>
17 
18 #include "wlcore.h"
19 #include "debug.h"
20 #include "wl12xx_80211.h"
21 #include "io.h"
22 #include "tx.h"
23 #include "ps.h"
24 #include "init.h"
25 #include "debugfs.h"
26 #include "testmode.h"
27 #include "vendor_cmd.h"
28 #include "scan.h"
29 #include "hw_ops.h"
30 #include "sysfs.h"
31 
32 #define WL1271_BOOT_RETRIES 3
33 #define WL1271_WAKEUP_TIMEOUT 500
34 
35 static char *fwlog_param;
36 static int fwlog_mem_blocks = -1;
37 static int bug_on_recovery = -1;
38 static int no_recovery     = -1;
39 
40 static void __wl1271_op_remove_interface(struct wl1271 *wl,
41 					 struct ieee80211_vif *vif,
42 					 bool reset_tx_queues);
43 static void wlcore_op_stop_locked(struct wl1271 *wl);
44 static void wl1271_free_ap_keys(struct wl1271 *wl, struct wl12xx_vif *wlvif);
45 
46 static int wl12xx_set_authorized(struct wl1271 *wl, struct wl12xx_vif *wlvif)
47 {
48 	int ret;
49 
50 	if (WARN_ON(wlvif->bss_type != BSS_TYPE_STA_BSS))
51 		return -EINVAL;
52 
53 	if (!test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
54 		return 0;
55 
56 	if (test_and_set_bit(WLVIF_FLAG_STA_STATE_SENT, &wlvif->flags))
57 		return 0;
58 
59 	ret = wl12xx_cmd_set_peer_state(wl, wlvif, wlvif->sta.hlid);
60 	if (ret < 0)
61 		return ret;
62 
63 	wl1271_info("Association completed.");
64 	return 0;
65 }
66 
67 static void wl1271_reg_notify(struct wiphy *wiphy,
68 			      struct regulatory_request *request)
69 {
70 	struct ieee80211_hw *hw = wiphy_to_ieee80211_hw(wiphy);
71 	struct wl1271 *wl = hw->priv;
72 
73 	/* copy the current dfs region */
74 	if (request)
75 		wl->dfs_region = request->dfs_region;
76 
77 	wlcore_regdomain_config(wl);
78 }
79 
80 static int wl1271_set_rx_streaming(struct wl1271 *wl, struct wl12xx_vif *wlvif,
81 				   bool enable)
82 {
83 	int ret = 0;
84 
85 	/* we should hold wl->mutex */
86 	ret = wl1271_acx_ps_rx_streaming(wl, wlvif, enable);
87 	if (ret < 0)
88 		goto out;
89 
90 	if (enable)
91 		set_bit(WLVIF_FLAG_RX_STREAMING_STARTED, &wlvif->flags);
92 	else
93 		clear_bit(WLVIF_FLAG_RX_STREAMING_STARTED, &wlvif->flags);
94 out:
95 	return ret;
96 }
97 
98 /*
99  * this function is being called when the rx_streaming interval
100  * has beed changed or rx_streaming should be disabled
101  */
102 int wl1271_recalc_rx_streaming(struct wl1271 *wl, struct wl12xx_vif *wlvif)
103 {
104 	int ret = 0;
105 	int period = wl->conf.rx_streaming.interval;
106 
107 	/* don't reconfigure if rx_streaming is disabled */
108 	if (!test_bit(WLVIF_FLAG_RX_STREAMING_STARTED, &wlvif->flags))
109 		goto out;
110 
111 	/* reconfigure/disable according to new streaming_period */
112 	if (period &&
113 	    test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags) &&
114 	    (wl->conf.rx_streaming.always ||
115 	     test_bit(WL1271_FLAG_SOFT_GEMINI, &wl->flags)))
116 		ret = wl1271_set_rx_streaming(wl, wlvif, true);
117 	else {
118 		ret = wl1271_set_rx_streaming(wl, wlvif, false);
119 		/* don't cancel_work_sync since we might deadlock */
120 		del_timer_sync(&wlvif->rx_streaming_timer);
121 	}
122 out:
123 	return ret;
124 }
125 
126 static void wl1271_rx_streaming_enable_work(struct work_struct *work)
127 {
128 	int ret;
129 	struct wl12xx_vif *wlvif = container_of(work, struct wl12xx_vif,
130 						rx_streaming_enable_work);
131 	struct wl1271 *wl = wlvif->wl;
132 
133 	mutex_lock(&wl->mutex);
134 
135 	if (test_bit(WLVIF_FLAG_RX_STREAMING_STARTED, &wlvif->flags) ||
136 	    !test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags) ||
137 	    (!wl->conf.rx_streaming.always &&
138 	     !test_bit(WL1271_FLAG_SOFT_GEMINI, &wl->flags)))
139 		goto out;
140 
141 	if (!wl->conf.rx_streaming.interval)
142 		goto out;
143 
144 	ret = pm_runtime_resume_and_get(wl->dev);
145 	if (ret < 0)
146 		goto out;
147 
148 	ret = wl1271_set_rx_streaming(wl, wlvif, true);
149 	if (ret < 0)
150 		goto out_sleep;
151 
152 	/* stop it after some time of inactivity */
153 	mod_timer(&wlvif->rx_streaming_timer,
154 		  jiffies + msecs_to_jiffies(wl->conf.rx_streaming.duration));
155 
156 out_sleep:
157 	pm_runtime_mark_last_busy(wl->dev);
158 	pm_runtime_put_autosuspend(wl->dev);
159 out:
160 	mutex_unlock(&wl->mutex);
161 }
162 
163 static void wl1271_rx_streaming_disable_work(struct work_struct *work)
164 {
165 	int ret;
166 	struct wl12xx_vif *wlvif = container_of(work, struct wl12xx_vif,
167 						rx_streaming_disable_work);
168 	struct wl1271 *wl = wlvif->wl;
169 
170 	mutex_lock(&wl->mutex);
171 
172 	if (!test_bit(WLVIF_FLAG_RX_STREAMING_STARTED, &wlvif->flags))
173 		goto out;
174 
175 	ret = pm_runtime_resume_and_get(wl->dev);
176 	if (ret < 0)
177 		goto out;
178 
179 	ret = wl1271_set_rx_streaming(wl, wlvif, false);
180 	if (ret)
181 		goto out_sleep;
182 
183 out_sleep:
184 	pm_runtime_mark_last_busy(wl->dev);
185 	pm_runtime_put_autosuspend(wl->dev);
186 out:
187 	mutex_unlock(&wl->mutex);
188 }
189 
190 static void wl1271_rx_streaming_timer(struct timer_list *t)
191 {
192 	struct wl12xx_vif *wlvif = from_timer(wlvif, t, rx_streaming_timer);
193 	struct wl1271 *wl = wlvif->wl;
194 	ieee80211_queue_work(wl->hw, &wlvif->rx_streaming_disable_work);
195 }
196 
197 /* wl->mutex must be taken */
198 void wl12xx_rearm_tx_watchdog_locked(struct wl1271 *wl)
199 {
200 	/* if the watchdog is not armed, don't do anything */
201 	if (wl->tx_allocated_blocks == 0)
202 		return;
203 
204 	cancel_delayed_work(&wl->tx_watchdog_work);
205 	ieee80211_queue_delayed_work(wl->hw, &wl->tx_watchdog_work,
206 		msecs_to_jiffies(wl->conf.tx.tx_watchdog_timeout));
207 }
208 
209 static void wlcore_rc_update_work(struct work_struct *work)
210 {
211 	int ret;
212 	struct wl12xx_vif *wlvif = container_of(work, struct wl12xx_vif,
213 						rc_update_work);
214 	struct wl1271 *wl = wlvif->wl;
215 	struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
216 
217 	mutex_lock(&wl->mutex);
218 
219 	if (unlikely(wl->state != WLCORE_STATE_ON))
220 		goto out;
221 
222 	ret = pm_runtime_resume_and_get(wl->dev);
223 	if (ret < 0)
224 		goto out;
225 
226 	if (ieee80211_vif_is_mesh(vif)) {
227 		ret = wl1271_acx_set_ht_capabilities(wl, &wlvif->rc_ht_cap,
228 						     true, wlvif->sta.hlid);
229 		if (ret < 0)
230 			goto out_sleep;
231 	} else {
232 		wlcore_hw_sta_rc_update(wl, wlvif);
233 	}
234 
235 out_sleep:
236 	pm_runtime_mark_last_busy(wl->dev);
237 	pm_runtime_put_autosuspend(wl->dev);
238 out:
239 	mutex_unlock(&wl->mutex);
240 }
241 
242 static void wl12xx_tx_watchdog_work(struct work_struct *work)
243 {
244 	struct delayed_work *dwork;
245 	struct wl1271 *wl;
246 
247 	dwork = to_delayed_work(work);
248 	wl = container_of(dwork, struct wl1271, tx_watchdog_work);
249 
250 	mutex_lock(&wl->mutex);
251 
252 	if (unlikely(wl->state != WLCORE_STATE_ON))
253 		goto out;
254 
255 	/* Tx went out in the meantime - everything is ok */
256 	if (unlikely(wl->tx_allocated_blocks == 0))
257 		goto out;
258 
259 	/*
260 	 * if a ROC is in progress, we might not have any Tx for a long
261 	 * time (e.g. pending Tx on the non-ROC channels)
262 	 */
263 	if (find_first_bit(wl->roc_map, WL12XX_MAX_ROLES) < WL12XX_MAX_ROLES) {
264 		wl1271_debug(DEBUG_TX, "No Tx (in FW) for %d ms due to ROC",
265 			     wl->conf.tx.tx_watchdog_timeout);
266 		wl12xx_rearm_tx_watchdog_locked(wl);
267 		goto out;
268 	}
269 
270 	/*
271 	 * if a scan is in progress, we might not have any Tx for a long
272 	 * time
273 	 */
274 	if (wl->scan.state != WL1271_SCAN_STATE_IDLE) {
275 		wl1271_debug(DEBUG_TX, "No Tx (in FW) for %d ms due to scan",
276 			     wl->conf.tx.tx_watchdog_timeout);
277 		wl12xx_rearm_tx_watchdog_locked(wl);
278 		goto out;
279 	}
280 
281 	/*
282 	* AP might cache a frame for a long time for a sleeping station,
283 	* so rearm the timer if there's an AP interface with stations. If
284 	* Tx is genuinely stuck we will most hopefully discover it when all
285 	* stations are removed due to inactivity.
286 	*/
287 	if (wl->active_sta_count) {
288 		wl1271_debug(DEBUG_TX, "No Tx (in FW) for %d ms. AP has "
289 			     " %d stations",
290 			      wl->conf.tx.tx_watchdog_timeout,
291 			      wl->active_sta_count);
292 		wl12xx_rearm_tx_watchdog_locked(wl);
293 		goto out;
294 	}
295 
296 	wl1271_error("Tx stuck (in FW) for %d ms. Starting recovery",
297 		     wl->conf.tx.tx_watchdog_timeout);
298 	wl12xx_queue_recovery_work(wl);
299 
300 out:
301 	mutex_unlock(&wl->mutex);
302 }
303 
304 static void wlcore_adjust_conf(struct wl1271 *wl)
305 {
306 
307 	if (fwlog_param) {
308 		if (!strcmp(fwlog_param, "continuous")) {
309 			wl->conf.fwlog.mode = WL12XX_FWLOG_CONTINUOUS;
310 			wl->conf.fwlog.output = WL12XX_FWLOG_OUTPUT_HOST;
311 		} else if (!strcmp(fwlog_param, "dbgpins")) {
312 			wl->conf.fwlog.mode = WL12XX_FWLOG_CONTINUOUS;
313 			wl->conf.fwlog.output = WL12XX_FWLOG_OUTPUT_DBG_PINS;
314 		} else if (!strcmp(fwlog_param, "disable")) {
315 			wl->conf.fwlog.mem_blocks = 0;
316 			wl->conf.fwlog.output = WL12XX_FWLOG_OUTPUT_NONE;
317 		} else {
318 			wl1271_error("Unknown fwlog parameter %s", fwlog_param);
319 		}
320 	}
321 
322 	if (bug_on_recovery != -1)
323 		wl->conf.recovery.bug_on_recovery = (u8) bug_on_recovery;
324 
325 	if (no_recovery != -1)
326 		wl->conf.recovery.no_recovery = (u8) no_recovery;
327 }
328 
329 static void wl12xx_irq_ps_regulate_link(struct wl1271 *wl,
330 					struct wl12xx_vif *wlvif,
331 					u8 hlid, u8 tx_pkts)
332 {
333 	bool fw_ps;
334 
335 	fw_ps = test_bit(hlid, &wl->ap_fw_ps_map);
336 
337 	/*
338 	 * Wake up from high level PS if the STA is asleep with too little
339 	 * packets in FW or if the STA is awake.
340 	 */
341 	if (!fw_ps || tx_pkts < WL1271_PS_STA_MAX_PACKETS)
342 		wl12xx_ps_link_end(wl, wlvif, hlid);
343 
344 	/*
345 	 * Start high-level PS if the STA is asleep with enough blocks in FW.
346 	 * Make an exception if this is the only connected link. In this
347 	 * case FW-memory congestion is less of a problem.
348 	 * Note that a single connected STA means 2*ap_count + 1 active links,
349 	 * since we must account for the global and broadcast AP links
350 	 * for each AP. The "fw_ps" check assures us the other link is a STA
351 	 * connected to the AP. Otherwise the FW would not set the PSM bit.
352 	 */
353 	else if (wl->active_link_count > (wl->ap_count*2 + 1) && fw_ps &&
354 		 tx_pkts >= WL1271_PS_STA_MAX_PACKETS)
355 		wl12xx_ps_link_start(wl, wlvif, hlid, true);
356 }
357 
358 static void wl12xx_irq_update_links_status(struct wl1271 *wl,
359 					   struct wl12xx_vif *wlvif,
360 					   struct wl_fw_status *status)
361 {
362 	unsigned long cur_fw_ps_map;
363 	u8 hlid;
364 
365 	cur_fw_ps_map = status->link_ps_bitmap;
366 	if (wl->ap_fw_ps_map != cur_fw_ps_map) {
367 		wl1271_debug(DEBUG_PSM,
368 			     "link ps prev 0x%lx cur 0x%lx changed 0x%lx",
369 			     wl->ap_fw_ps_map, cur_fw_ps_map,
370 			     wl->ap_fw_ps_map ^ cur_fw_ps_map);
371 
372 		wl->ap_fw_ps_map = cur_fw_ps_map;
373 	}
374 
375 	for_each_set_bit(hlid, wlvif->ap.sta_hlid_map, wl->num_links)
376 		wl12xx_irq_ps_regulate_link(wl, wlvif, hlid,
377 					    wl->links[hlid].allocated_pkts);
378 }
379 
380 static int wlcore_fw_status(struct wl1271 *wl, struct wl_fw_status *status)
381 {
382 	struct wl12xx_vif *wlvif;
383 	u32 old_tx_blk_count = wl->tx_blocks_available;
384 	int avail, freed_blocks;
385 	int i;
386 	int ret;
387 	struct wl1271_link *lnk;
388 
389 	ret = wlcore_raw_read_data(wl, REG_RAW_FW_STATUS_ADDR,
390 				   wl->raw_fw_status,
391 				   wl->fw_status_len, false);
392 	if (ret < 0)
393 		return ret;
394 
395 	wlcore_hw_convert_fw_status(wl, wl->raw_fw_status, wl->fw_status);
396 
397 	wl1271_debug(DEBUG_IRQ, "intr: 0x%x (fw_rx_counter = %d, "
398 		     "drv_rx_counter = %d, tx_results_counter = %d)",
399 		     status->intr,
400 		     status->fw_rx_counter,
401 		     status->drv_rx_counter,
402 		     status->tx_results_counter);
403 
404 	for (i = 0; i < NUM_TX_QUEUES; i++) {
405 		/* prevent wrap-around in freed-packets counter */
406 		wl->tx_allocated_pkts[i] -=
407 				(status->counters.tx_released_pkts[i] -
408 				wl->tx_pkts_freed[i]) & 0xff;
409 
410 		wl->tx_pkts_freed[i] = status->counters.tx_released_pkts[i];
411 	}
412 
413 
414 	for_each_set_bit(i, wl->links_map, wl->num_links) {
415 		u8 diff;
416 		lnk = &wl->links[i];
417 
418 		/* prevent wrap-around in freed-packets counter */
419 		diff = (status->counters.tx_lnk_free_pkts[i] -
420 		       lnk->prev_freed_pkts) & 0xff;
421 
422 		if (diff == 0)
423 			continue;
424 
425 		lnk->allocated_pkts -= diff;
426 		lnk->prev_freed_pkts = status->counters.tx_lnk_free_pkts[i];
427 
428 		/* accumulate the prev_freed_pkts counter */
429 		lnk->total_freed_pkts += diff;
430 	}
431 
432 	/* prevent wrap-around in total blocks counter */
433 	if (likely(wl->tx_blocks_freed <= status->total_released_blks))
434 		freed_blocks = status->total_released_blks -
435 			       wl->tx_blocks_freed;
436 	else
437 		freed_blocks = 0x100000000LL - wl->tx_blocks_freed +
438 			       status->total_released_blks;
439 
440 	wl->tx_blocks_freed = status->total_released_blks;
441 
442 	wl->tx_allocated_blocks -= freed_blocks;
443 
444 	/*
445 	 * If the FW freed some blocks:
446 	 * If we still have allocated blocks - re-arm the timer, Tx is
447 	 * not stuck. Otherwise, cancel the timer (no Tx currently).
448 	 */
449 	if (freed_blocks) {
450 		if (wl->tx_allocated_blocks)
451 			wl12xx_rearm_tx_watchdog_locked(wl);
452 		else
453 			cancel_delayed_work(&wl->tx_watchdog_work);
454 	}
455 
456 	avail = status->tx_total - wl->tx_allocated_blocks;
457 
458 	/*
459 	 * The FW might change the total number of TX memblocks before
460 	 * we get a notification about blocks being released. Thus, the
461 	 * available blocks calculation might yield a temporary result
462 	 * which is lower than the actual available blocks. Keeping in
463 	 * mind that only blocks that were allocated can be moved from
464 	 * TX to RX, tx_blocks_available should never decrease here.
465 	 */
466 	wl->tx_blocks_available = max((int)wl->tx_blocks_available,
467 				      avail);
468 
469 	/* if more blocks are available now, tx work can be scheduled */
470 	if (wl->tx_blocks_available > old_tx_blk_count)
471 		clear_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags);
472 
473 	/* for AP update num of allocated TX blocks per link and ps status */
474 	wl12xx_for_each_wlvif_ap(wl, wlvif) {
475 		wl12xx_irq_update_links_status(wl, wlvif, status);
476 	}
477 
478 	/* update the host-chipset time offset */
479 	wl->time_offset = (ktime_get_boottime_ns() >> 10) -
480 		(s64)(status->fw_localtime);
481 
482 	wl->fw_fast_lnk_map = status->link_fast_bitmap;
483 
484 	return 0;
485 }
486 
487 static void wl1271_flush_deferred_work(struct wl1271 *wl)
488 {
489 	struct sk_buff *skb;
490 
491 	/* Pass all received frames to the network stack */
492 	while ((skb = skb_dequeue(&wl->deferred_rx_queue)))
493 		ieee80211_rx_ni(wl->hw, skb);
494 
495 	/* Return sent skbs to the network stack */
496 	while ((skb = skb_dequeue(&wl->deferred_tx_queue)))
497 		ieee80211_tx_status_ni(wl->hw, skb);
498 }
499 
500 static void wl1271_netstack_work(struct work_struct *work)
501 {
502 	struct wl1271 *wl =
503 		container_of(work, struct wl1271, netstack_work);
504 
505 	do {
506 		wl1271_flush_deferred_work(wl);
507 	} while (skb_queue_len(&wl->deferred_rx_queue));
508 }
509 
510 #define WL1271_IRQ_MAX_LOOPS 256
511 
512 static int wlcore_irq_locked(struct wl1271 *wl)
513 {
514 	int ret = 0;
515 	u32 intr;
516 	int loopcount = WL1271_IRQ_MAX_LOOPS;
517 	bool run_tx_queue = true;
518 	bool done = false;
519 	unsigned int defer_count;
520 	unsigned long flags;
521 
522 	/*
523 	 * In case edge triggered interrupt must be used, we cannot iterate
524 	 * more than once without introducing race conditions with the hardirq.
525 	 */
526 	if (wl->irq_flags & (IRQF_TRIGGER_RISING | IRQF_TRIGGER_FALLING))
527 		loopcount = 1;
528 
529 	wl1271_debug(DEBUG_IRQ, "IRQ work");
530 
531 	if (unlikely(wl->state != WLCORE_STATE_ON))
532 		goto out;
533 
534 	ret = pm_runtime_resume_and_get(wl->dev);
535 	if (ret < 0)
536 		goto out;
537 
538 	while (!done && loopcount--) {
539 		smp_mb__after_atomic();
540 
541 		ret = wlcore_fw_status(wl, wl->fw_status);
542 		if (ret < 0)
543 			goto err_ret;
544 
545 		wlcore_hw_tx_immediate_compl(wl);
546 
547 		intr = wl->fw_status->intr;
548 		intr &= WLCORE_ALL_INTR_MASK;
549 		if (!intr) {
550 			done = true;
551 			continue;
552 		}
553 
554 		if (unlikely(intr & WL1271_ACX_INTR_WATCHDOG)) {
555 			wl1271_error("HW watchdog interrupt received! starting recovery.");
556 			wl->watchdog_recovery = true;
557 			ret = -EIO;
558 
559 			/* restarting the chip. ignore any other interrupt. */
560 			goto err_ret;
561 		}
562 
563 		if (unlikely(intr & WL1271_ACX_SW_INTR_WATCHDOG)) {
564 			wl1271_error("SW watchdog interrupt received! "
565 				     "starting recovery.");
566 			wl->watchdog_recovery = true;
567 			ret = -EIO;
568 
569 			/* restarting the chip. ignore any other interrupt. */
570 			goto err_ret;
571 		}
572 
573 		if (likely(intr & WL1271_ACX_INTR_DATA)) {
574 			wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_DATA");
575 
576 			ret = wlcore_rx(wl, wl->fw_status);
577 			if (ret < 0)
578 				goto err_ret;
579 
580 			/* Check if any tx blocks were freed */
581 			if (!test_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags)) {
582 				if (spin_trylock_irqsave(&wl->wl_lock, flags)) {
583 					if (!wl1271_tx_total_queue_count(wl))
584 						run_tx_queue = false;
585 					spin_unlock_irqrestore(&wl->wl_lock, flags);
586 				}
587 
588 				/*
589 				 * In order to avoid starvation of the TX path,
590 				 * call the work function directly.
591 				 */
592 				if (run_tx_queue) {
593 					ret = wlcore_tx_work_locked(wl);
594 					if (ret < 0)
595 						goto err_ret;
596 				}
597 			}
598 
599 			/* check for tx results */
600 			ret = wlcore_hw_tx_delayed_compl(wl);
601 			if (ret < 0)
602 				goto err_ret;
603 
604 			/* Make sure the deferred queues don't get too long */
605 			defer_count = skb_queue_len(&wl->deferred_tx_queue) +
606 				      skb_queue_len(&wl->deferred_rx_queue);
607 			if (defer_count > WL1271_DEFERRED_QUEUE_LIMIT)
608 				wl1271_flush_deferred_work(wl);
609 		}
610 
611 		if (intr & WL1271_ACX_INTR_EVENT_A) {
612 			wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_A");
613 			ret = wl1271_event_handle(wl, 0);
614 			if (ret < 0)
615 				goto err_ret;
616 		}
617 
618 		if (intr & WL1271_ACX_INTR_EVENT_B) {
619 			wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_B");
620 			ret = wl1271_event_handle(wl, 1);
621 			if (ret < 0)
622 				goto err_ret;
623 		}
624 
625 		if (intr & WL1271_ACX_INTR_INIT_COMPLETE)
626 			wl1271_debug(DEBUG_IRQ,
627 				     "WL1271_ACX_INTR_INIT_COMPLETE");
628 
629 		if (intr & WL1271_ACX_INTR_HW_AVAILABLE)
630 			wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_HW_AVAILABLE");
631 	}
632 
633 err_ret:
634 	pm_runtime_mark_last_busy(wl->dev);
635 	pm_runtime_put_autosuspend(wl->dev);
636 
637 out:
638 	return ret;
639 }
640 
641 static irqreturn_t wlcore_irq(int irq, void *cookie)
642 {
643 	int ret;
644 	unsigned long flags;
645 	struct wl1271 *wl = cookie;
646 	bool queue_tx_work = true;
647 
648 	set_bit(WL1271_FLAG_IRQ_RUNNING, &wl->flags);
649 
650 	/* complete the ELP completion */
651 	if (test_bit(WL1271_FLAG_IN_ELP, &wl->flags)) {
652 		spin_lock_irqsave(&wl->wl_lock, flags);
653 		if (wl->elp_compl)
654 			complete(wl->elp_compl);
655 		spin_unlock_irqrestore(&wl->wl_lock, flags);
656 	}
657 
658 	if (test_bit(WL1271_FLAG_SUSPENDED, &wl->flags)) {
659 		/* don't enqueue a work right now. mark it as pending */
660 		set_bit(WL1271_FLAG_PENDING_WORK, &wl->flags);
661 		wl1271_debug(DEBUG_IRQ, "should not enqueue work");
662 		spin_lock_irqsave(&wl->wl_lock, flags);
663 		disable_irq_nosync(wl->irq);
664 		pm_wakeup_event(wl->dev, 0);
665 		spin_unlock_irqrestore(&wl->wl_lock, flags);
666 		goto out_handled;
667 	}
668 
669 	/* TX might be handled here, avoid redundant work */
670 	set_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
671 	cancel_work_sync(&wl->tx_work);
672 
673 	mutex_lock(&wl->mutex);
674 
675 	ret = wlcore_irq_locked(wl);
676 	if (ret)
677 		wl12xx_queue_recovery_work(wl);
678 
679 	/* In case TX was not handled in wlcore_irq_locked(), queue TX work */
680 	clear_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
681 	if (!test_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags)) {
682 		if (spin_trylock_irqsave(&wl->wl_lock, flags)) {
683 			if (!wl1271_tx_total_queue_count(wl))
684 				queue_tx_work = false;
685 			spin_unlock_irqrestore(&wl->wl_lock, flags);
686 		}
687 		if (queue_tx_work)
688 			ieee80211_queue_work(wl->hw, &wl->tx_work);
689 	}
690 
691 	mutex_unlock(&wl->mutex);
692 
693 out_handled:
694 	clear_bit(WL1271_FLAG_IRQ_RUNNING, &wl->flags);
695 
696 	return IRQ_HANDLED;
697 }
698 
699 struct vif_counter_data {
700 	u8 counter;
701 
702 	struct ieee80211_vif *cur_vif;
703 	bool cur_vif_running;
704 };
705 
706 static void wl12xx_vif_count_iter(void *data, u8 *mac,
707 				  struct ieee80211_vif *vif)
708 {
709 	struct vif_counter_data *counter = data;
710 
711 	counter->counter++;
712 	if (counter->cur_vif == vif)
713 		counter->cur_vif_running = true;
714 }
715 
716 /* caller must not hold wl->mutex, as it might deadlock */
717 static void wl12xx_get_vif_count(struct ieee80211_hw *hw,
718 			       struct ieee80211_vif *cur_vif,
719 			       struct vif_counter_data *data)
720 {
721 	memset(data, 0, sizeof(*data));
722 	data->cur_vif = cur_vif;
723 
724 	ieee80211_iterate_active_interfaces(hw, IEEE80211_IFACE_ITER_RESUME_ALL,
725 					    wl12xx_vif_count_iter, data);
726 }
727 
728 static int wl12xx_fetch_firmware(struct wl1271 *wl, bool plt)
729 {
730 	const struct firmware *fw;
731 	const char *fw_name;
732 	enum wl12xx_fw_type fw_type;
733 	int ret;
734 
735 	if (plt) {
736 		fw_type = WL12XX_FW_TYPE_PLT;
737 		fw_name = wl->plt_fw_name;
738 	} else {
739 		/*
740 		 * we can't call wl12xx_get_vif_count() here because
741 		 * wl->mutex is taken, so use the cached last_vif_count value
742 		 */
743 		if (wl->last_vif_count > 1 && wl->mr_fw_name) {
744 			fw_type = WL12XX_FW_TYPE_MULTI;
745 			fw_name = wl->mr_fw_name;
746 		} else {
747 			fw_type = WL12XX_FW_TYPE_NORMAL;
748 			fw_name = wl->sr_fw_name;
749 		}
750 	}
751 
752 	if (wl->fw_type == fw_type)
753 		return 0;
754 
755 	wl1271_debug(DEBUG_BOOT, "booting firmware %s", fw_name);
756 
757 	ret = request_firmware(&fw, fw_name, wl->dev);
758 
759 	if (ret < 0) {
760 		wl1271_error("could not get firmware %s: %d", fw_name, ret);
761 		return ret;
762 	}
763 
764 	if (fw->size % 4) {
765 		wl1271_error("firmware size is not multiple of 32 bits: %zu",
766 			     fw->size);
767 		ret = -EILSEQ;
768 		goto out;
769 	}
770 
771 	vfree(wl->fw);
772 	wl->fw_type = WL12XX_FW_TYPE_NONE;
773 	wl->fw_len = fw->size;
774 	wl->fw = vmalloc(wl->fw_len);
775 
776 	if (!wl->fw) {
777 		wl1271_error("could not allocate memory for the firmware");
778 		ret = -ENOMEM;
779 		goto out;
780 	}
781 
782 	memcpy(wl->fw, fw->data, wl->fw_len);
783 	ret = 0;
784 	wl->fw_type = fw_type;
785 out:
786 	release_firmware(fw);
787 
788 	return ret;
789 }
790 
791 void wl12xx_queue_recovery_work(struct wl1271 *wl)
792 {
793 	/* Avoid a recursive recovery */
794 	if (wl->state == WLCORE_STATE_ON) {
795 		WARN_ON(!test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY,
796 				  &wl->flags));
797 
798 		wl->state = WLCORE_STATE_RESTARTING;
799 		set_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags);
800 		ieee80211_queue_work(wl->hw, &wl->recovery_work);
801 	}
802 }
803 
804 size_t wl12xx_copy_fwlog(struct wl1271 *wl, u8 *memblock, size_t maxlen)
805 {
806 	size_t len;
807 
808 	/* Make sure we have enough room */
809 	len = min_t(size_t, maxlen, PAGE_SIZE - wl->fwlog_size);
810 
811 	/* Fill the FW log file, consumed by the sysfs fwlog entry */
812 	memcpy(wl->fwlog + wl->fwlog_size, memblock, len);
813 	wl->fwlog_size += len;
814 
815 	return len;
816 }
817 
818 static void wl12xx_read_fwlog_panic(struct wl1271 *wl)
819 {
820 	u32 end_of_log = 0;
821 	int error;
822 
823 	if (wl->quirks & WLCORE_QUIRK_FWLOG_NOT_IMPLEMENTED)
824 		return;
825 
826 	wl1271_info("Reading FW panic log");
827 
828 	/*
829 	 * Make sure the chip is awake and the logger isn't active.
830 	 * Do not send a stop fwlog command if the fw is hanged or if
831 	 * dbgpins are used (due to some fw bug).
832 	 */
833 	error = pm_runtime_resume_and_get(wl->dev);
834 	if (error < 0)
835 		return;
836 	if (!wl->watchdog_recovery &&
837 	    wl->conf.fwlog.output != WL12XX_FWLOG_OUTPUT_DBG_PINS)
838 		wl12xx_cmd_stop_fwlog(wl);
839 
840 	/* Traverse the memory blocks linked list */
841 	do {
842 		end_of_log = wlcore_event_fw_logger(wl);
843 		if (end_of_log == 0) {
844 			msleep(100);
845 			end_of_log = wlcore_event_fw_logger(wl);
846 		}
847 	} while (end_of_log != 0);
848 }
849 
850 static void wlcore_save_freed_pkts(struct wl1271 *wl, struct wl12xx_vif *wlvif,
851 				   u8 hlid, struct ieee80211_sta *sta)
852 {
853 	struct wl1271_station *wl_sta;
854 	u32 sqn_recovery_padding = WL1271_TX_SQN_POST_RECOVERY_PADDING;
855 
856 	wl_sta = (void *)sta->drv_priv;
857 	wl_sta->total_freed_pkts = wl->links[hlid].total_freed_pkts;
858 
859 	/*
860 	 * increment the initial seq number on recovery to account for
861 	 * transmitted packets that we haven't yet got in the FW status
862 	 */
863 	if (wlvif->encryption_type == KEY_GEM)
864 		sqn_recovery_padding = WL1271_TX_SQN_POST_RECOVERY_PADDING_GEM;
865 
866 	if (test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags))
867 		wl_sta->total_freed_pkts += sqn_recovery_padding;
868 }
869 
870 static void wlcore_save_freed_pkts_addr(struct wl1271 *wl,
871 					struct wl12xx_vif *wlvif,
872 					u8 hlid, const u8 *addr)
873 {
874 	struct ieee80211_sta *sta;
875 	struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
876 
877 	if (WARN_ON(hlid == WL12XX_INVALID_LINK_ID ||
878 		    is_zero_ether_addr(addr)))
879 		return;
880 
881 	rcu_read_lock();
882 	sta = ieee80211_find_sta(vif, addr);
883 	if (sta)
884 		wlcore_save_freed_pkts(wl, wlvif, hlid, sta);
885 	rcu_read_unlock();
886 }
887 
888 static void wlcore_print_recovery(struct wl1271 *wl)
889 {
890 	u32 pc = 0;
891 	u32 hint_sts = 0;
892 	int ret;
893 
894 	wl1271_info("Hardware recovery in progress. FW ver: %s",
895 		    wl->chip.fw_ver_str);
896 
897 	/* change partitions momentarily so we can read the FW pc */
898 	ret = wlcore_set_partition(wl, &wl->ptable[PART_BOOT]);
899 	if (ret < 0)
900 		return;
901 
902 	ret = wlcore_read_reg(wl, REG_PC_ON_RECOVERY, &pc);
903 	if (ret < 0)
904 		return;
905 
906 	ret = wlcore_read_reg(wl, REG_INTERRUPT_NO_CLEAR, &hint_sts);
907 	if (ret < 0)
908 		return;
909 
910 	wl1271_info("pc: 0x%x, hint_sts: 0x%08x count: %d",
911 				pc, hint_sts, ++wl->recovery_count);
912 
913 	wlcore_set_partition(wl, &wl->ptable[PART_WORK]);
914 }
915 
916 
917 static void wl1271_recovery_work(struct work_struct *work)
918 {
919 	struct wl1271 *wl =
920 		container_of(work, struct wl1271, recovery_work);
921 	struct wl12xx_vif *wlvif;
922 	struct ieee80211_vif *vif;
923 	int error;
924 
925 	mutex_lock(&wl->mutex);
926 
927 	if (wl->state == WLCORE_STATE_OFF || wl->plt)
928 		goto out_unlock;
929 
930 	error = pm_runtime_resume_and_get(wl->dev);
931 	if (error < 0)
932 		wl1271_warning("Enable for recovery failed");
933 	wlcore_disable_interrupts_nosync(wl);
934 
935 	if (!test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags)) {
936 		if (wl->conf.fwlog.output == WL12XX_FWLOG_OUTPUT_HOST)
937 			wl12xx_read_fwlog_panic(wl);
938 		wlcore_print_recovery(wl);
939 	}
940 
941 	BUG_ON(wl->conf.recovery.bug_on_recovery &&
942 	       !test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags));
943 
944 	clear_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags);
945 
946 	if (wl->conf.recovery.no_recovery) {
947 		wl1271_info("No recovery (chosen on module load). Fw will remain stuck.");
948 		goto out_unlock;
949 	}
950 
951 	/* Prevent spurious TX during FW restart */
952 	wlcore_stop_queues(wl, WLCORE_QUEUE_STOP_REASON_FW_RESTART);
953 
954 	/* reboot the chipset */
955 	while (!list_empty(&wl->wlvif_list)) {
956 		wlvif = list_first_entry(&wl->wlvif_list,
957 				       struct wl12xx_vif, list);
958 		vif = wl12xx_wlvif_to_vif(wlvif);
959 
960 		if (wlvif->bss_type == BSS_TYPE_STA_BSS &&
961 		    test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags)) {
962 			wlcore_save_freed_pkts_addr(wl, wlvif, wlvif->sta.hlid,
963 						    vif->bss_conf.bssid);
964 		}
965 
966 		__wl1271_op_remove_interface(wl, vif, false);
967 	}
968 
969 	wlcore_op_stop_locked(wl);
970 	pm_runtime_mark_last_busy(wl->dev);
971 	pm_runtime_put_autosuspend(wl->dev);
972 
973 	ieee80211_restart_hw(wl->hw);
974 
975 	/*
976 	 * Its safe to enable TX now - the queues are stopped after a request
977 	 * to restart the HW.
978 	 */
979 	wlcore_wake_queues(wl, WLCORE_QUEUE_STOP_REASON_FW_RESTART);
980 
981 out_unlock:
982 	wl->watchdog_recovery = false;
983 	clear_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags);
984 	mutex_unlock(&wl->mutex);
985 }
986 
987 static int wlcore_fw_wakeup(struct wl1271 *wl)
988 {
989 	return wlcore_raw_write32(wl, HW_ACCESS_ELP_CTRL_REG, ELPCTRL_WAKE_UP);
990 }
991 
992 static int wl1271_setup(struct wl1271 *wl)
993 {
994 	wl->raw_fw_status = kzalloc(wl->fw_status_len, GFP_KERNEL);
995 	if (!wl->raw_fw_status)
996 		goto err;
997 
998 	wl->fw_status = kzalloc(sizeof(*wl->fw_status), GFP_KERNEL);
999 	if (!wl->fw_status)
1000 		goto err;
1001 
1002 	wl->tx_res_if = kzalloc(sizeof(*wl->tx_res_if), GFP_KERNEL);
1003 	if (!wl->tx_res_if)
1004 		goto err;
1005 
1006 	return 0;
1007 err:
1008 	kfree(wl->fw_status);
1009 	kfree(wl->raw_fw_status);
1010 	return -ENOMEM;
1011 }
1012 
1013 static int wl12xx_set_power_on(struct wl1271 *wl)
1014 {
1015 	int ret;
1016 
1017 	msleep(WL1271_PRE_POWER_ON_SLEEP);
1018 	ret = wl1271_power_on(wl);
1019 	if (ret < 0)
1020 		goto out;
1021 	msleep(WL1271_POWER_ON_SLEEP);
1022 	wl1271_io_reset(wl);
1023 	wl1271_io_init(wl);
1024 
1025 	ret = wlcore_set_partition(wl, &wl->ptable[PART_BOOT]);
1026 	if (ret < 0)
1027 		goto fail;
1028 
1029 	/* ELP module wake up */
1030 	ret = wlcore_fw_wakeup(wl);
1031 	if (ret < 0)
1032 		goto fail;
1033 
1034 out:
1035 	return ret;
1036 
1037 fail:
1038 	wl1271_power_off(wl);
1039 	return ret;
1040 }
1041 
1042 static int wl12xx_chip_wakeup(struct wl1271 *wl, bool plt)
1043 {
1044 	int ret = 0;
1045 
1046 	ret = wl12xx_set_power_on(wl);
1047 	if (ret < 0)
1048 		goto out;
1049 
1050 	/*
1051 	 * For wl127x based devices we could use the default block
1052 	 * size (512 bytes), but due to a bug in the sdio driver, we
1053 	 * need to set it explicitly after the chip is powered on.  To
1054 	 * simplify the code and since the performance impact is
1055 	 * negligible, we use the same block size for all different
1056 	 * chip types.
1057 	 *
1058 	 * Check if the bus supports blocksize alignment and, if it
1059 	 * doesn't, make sure we don't have the quirk.
1060 	 */
1061 	if (!wl1271_set_block_size(wl))
1062 		wl->quirks &= ~WLCORE_QUIRK_TX_BLOCKSIZE_ALIGN;
1063 
1064 	/* TODO: make sure the lower driver has set things up correctly */
1065 
1066 	ret = wl1271_setup(wl);
1067 	if (ret < 0)
1068 		goto out;
1069 
1070 	ret = wl12xx_fetch_firmware(wl, plt);
1071 	if (ret < 0) {
1072 		kfree(wl->fw_status);
1073 		kfree(wl->raw_fw_status);
1074 		kfree(wl->tx_res_if);
1075 	}
1076 
1077 out:
1078 	return ret;
1079 }
1080 
1081 int wl1271_plt_start(struct wl1271 *wl, const enum plt_mode plt_mode)
1082 {
1083 	int retries = WL1271_BOOT_RETRIES;
1084 	struct wiphy *wiphy = wl->hw->wiphy;
1085 
1086 	static const char* const PLT_MODE[] = {
1087 		"PLT_OFF",
1088 		"PLT_ON",
1089 		"PLT_FEM_DETECT",
1090 		"PLT_CHIP_AWAKE"
1091 	};
1092 
1093 	int ret;
1094 
1095 	mutex_lock(&wl->mutex);
1096 
1097 	wl1271_notice("power up");
1098 
1099 	if (wl->state != WLCORE_STATE_OFF) {
1100 		wl1271_error("cannot go into PLT state because not "
1101 			     "in off state: %d", wl->state);
1102 		ret = -EBUSY;
1103 		goto out;
1104 	}
1105 
1106 	/* Indicate to lower levels that we are now in PLT mode */
1107 	wl->plt = true;
1108 	wl->plt_mode = plt_mode;
1109 
1110 	while (retries) {
1111 		retries--;
1112 		ret = wl12xx_chip_wakeup(wl, true);
1113 		if (ret < 0)
1114 			goto power_off;
1115 
1116 		if (plt_mode != PLT_CHIP_AWAKE) {
1117 			ret = wl->ops->plt_init(wl);
1118 			if (ret < 0)
1119 				goto power_off;
1120 		}
1121 
1122 		wl->state = WLCORE_STATE_ON;
1123 		wl1271_notice("firmware booted in PLT mode %s (%s)",
1124 			      PLT_MODE[plt_mode],
1125 			      wl->chip.fw_ver_str);
1126 
1127 		/* update hw/fw version info in wiphy struct */
1128 		wiphy->hw_version = wl->chip.id;
1129 		strncpy(wiphy->fw_version, wl->chip.fw_ver_str,
1130 			sizeof(wiphy->fw_version));
1131 
1132 		goto out;
1133 
1134 power_off:
1135 		wl1271_power_off(wl);
1136 	}
1137 
1138 	wl->plt = false;
1139 	wl->plt_mode = PLT_OFF;
1140 
1141 	wl1271_error("firmware boot in PLT mode failed despite %d retries",
1142 		     WL1271_BOOT_RETRIES);
1143 out:
1144 	mutex_unlock(&wl->mutex);
1145 
1146 	return ret;
1147 }
1148 
1149 int wl1271_plt_stop(struct wl1271 *wl)
1150 {
1151 	int ret = 0;
1152 
1153 	wl1271_notice("power down");
1154 
1155 	/*
1156 	 * Interrupts must be disabled before setting the state to OFF.
1157 	 * Otherwise, the interrupt handler might be called and exit without
1158 	 * reading the interrupt status.
1159 	 */
1160 	wlcore_disable_interrupts(wl);
1161 	mutex_lock(&wl->mutex);
1162 	if (!wl->plt) {
1163 		mutex_unlock(&wl->mutex);
1164 
1165 		/*
1166 		 * This will not necessarily enable interrupts as interrupts
1167 		 * may have been disabled when op_stop was called. It will,
1168 		 * however, balance the above call to disable_interrupts().
1169 		 */
1170 		wlcore_enable_interrupts(wl);
1171 
1172 		wl1271_error("cannot power down because not in PLT "
1173 			     "state: %d", wl->state);
1174 		ret = -EBUSY;
1175 		goto out;
1176 	}
1177 
1178 	mutex_unlock(&wl->mutex);
1179 
1180 	wl1271_flush_deferred_work(wl);
1181 	cancel_work_sync(&wl->netstack_work);
1182 	cancel_work_sync(&wl->recovery_work);
1183 	cancel_delayed_work_sync(&wl->tx_watchdog_work);
1184 
1185 	mutex_lock(&wl->mutex);
1186 	wl1271_power_off(wl);
1187 	wl->flags = 0;
1188 	wl->sleep_auth = WL1271_PSM_ILLEGAL;
1189 	wl->state = WLCORE_STATE_OFF;
1190 	wl->plt = false;
1191 	wl->plt_mode = PLT_OFF;
1192 	wl->rx_counter = 0;
1193 	mutex_unlock(&wl->mutex);
1194 
1195 out:
1196 	return ret;
1197 }
1198 
1199 static void wl1271_op_tx(struct ieee80211_hw *hw,
1200 			 struct ieee80211_tx_control *control,
1201 			 struct sk_buff *skb)
1202 {
1203 	struct wl1271 *wl = hw->priv;
1204 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
1205 	struct ieee80211_vif *vif = info->control.vif;
1206 	struct wl12xx_vif *wlvif = NULL;
1207 	unsigned long flags;
1208 	int q, mapping;
1209 	u8 hlid;
1210 
1211 	if (!vif) {
1212 		wl1271_debug(DEBUG_TX, "DROP skb with no vif");
1213 		ieee80211_free_txskb(hw, skb);
1214 		return;
1215 	}
1216 
1217 	wlvif = wl12xx_vif_to_data(vif);
1218 	mapping = skb_get_queue_mapping(skb);
1219 	q = wl1271_tx_get_queue(mapping);
1220 
1221 	hlid = wl12xx_tx_get_hlid(wl, wlvif, skb, control->sta);
1222 
1223 	spin_lock_irqsave(&wl->wl_lock, flags);
1224 
1225 	/*
1226 	 * drop the packet if the link is invalid or the queue is stopped
1227 	 * for any reason but watermark. Watermark is a "soft"-stop so we
1228 	 * allow these packets through.
1229 	 */
1230 	if (hlid == WL12XX_INVALID_LINK_ID ||
1231 	    (!test_bit(hlid, wlvif->links_map)) ||
1232 	     (wlcore_is_queue_stopped_locked(wl, wlvif, q) &&
1233 	      !wlcore_is_queue_stopped_by_reason_locked(wl, wlvif, q,
1234 			WLCORE_QUEUE_STOP_REASON_WATERMARK))) {
1235 		wl1271_debug(DEBUG_TX, "DROP skb hlid %d q %d", hlid, q);
1236 		ieee80211_free_txskb(hw, skb);
1237 		goto out;
1238 	}
1239 
1240 	wl1271_debug(DEBUG_TX, "queue skb hlid %d q %d len %d",
1241 		     hlid, q, skb->len);
1242 	skb_queue_tail(&wl->links[hlid].tx_queue[q], skb);
1243 
1244 	wl->tx_queue_count[q]++;
1245 	wlvif->tx_queue_count[q]++;
1246 
1247 	/*
1248 	 * The workqueue is slow to process the tx_queue and we need stop
1249 	 * the queue here, otherwise the queue will get too long.
1250 	 */
1251 	if (wlvif->tx_queue_count[q] >= WL1271_TX_QUEUE_HIGH_WATERMARK &&
1252 	    !wlcore_is_queue_stopped_by_reason_locked(wl, wlvif, q,
1253 					WLCORE_QUEUE_STOP_REASON_WATERMARK)) {
1254 		wl1271_debug(DEBUG_TX, "op_tx: stopping queues for q %d", q);
1255 		wlcore_stop_queue_locked(wl, wlvif, q,
1256 					 WLCORE_QUEUE_STOP_REASON_WATERMARK);
1257 	}
1258 
1259 	/*
1260 	 * The chip specific setup must run before the first TX packet -
1261 	 * before that, the tx_work will not be initialized!
1262 	 */
1263 
1264 	if (!test_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags) &&
1265 	    !test_bit(WL1271_FLAG_TX_PENDING, &wl->flags))
1266 		ieee80211_queue_work(wl->hw, &wl->tx_work);
1267 
1268 out:
1269 	spin_unlock_irqrestore(&wl->wl_lock, flags);
1270 }
1271 
1272 int wl1271_tx_dummy_packet(struct wl1271 *wl)
1273 {
1274 	unsigned long flags;
1275 	int q;
1276 
1277 	/* no need to queue a new dummy packet if one is already pending */
1278 	if (test_bit(WL1271_FLAG_DUMMY_PACKET_PENDING, &wl->flags))
1279 		return 0;
1280 
1281 	q = wl1271_tx_get_queue(skb_get_queue_mapping(wl->dummy_packet));
1282 
1283 	spin_lock_irqsave(&wl->wl_lock, flags);
1284 	set_bit(WL1271_FLAG_DUMMY_PACKET_PENDING, &wl->flags);
1285 	wl->tx_queue_count[q]++;
1286 	spin_unlock_irqrestore(&wl->wl_lock, flags);
1287 
1288 	/* The FW is low on RX memory blocks, so send the dummy packet asap */
1289 	if (!test_bit(WL1271_FLAG_FW_TX_BUSY, &wl->flags))
1290 		return wlcore_tx_work_locked(wl);
1291 
1292 	/*
1293 	 * If the FW TX is busy, TX work will be scheduled by the threaded
1294 	 * interrupt handler function
1295 	 */
1296 	return 0;
1297 }
1298 
1299 /*
1300  * The size of the dummy packet should be at least 1400 bytes. However, in
1301  * order to minimize the number of bus transactions, aligning it to 512 bytes
1302  * boundaries could be beneficial, performance wise
1303  */
1304 #define TOTAL_TX_DUMMY_PACKET_SIZE (ALIGN(1400, 512))
1305 
1306 static struct sk_buff *wl12xx_alloc_dummy_packet(struct wl1271 *wl)
1307 {
1308 	struct sk_buff *skb;
1309 	struct ieee80211_hdr_3addr *hdr;
1310 	unsigned int dummy_packet_size;
1311 
1312 	dummy_packet_size = TOTAL_TX_DUMMY_PACKET_SIZE -
1313 			    sizeof(struct wl1271_tx_hw_descr) - sizeof(*hdr);
1314 
1315 	skb = dev_alloc_skb(TOTAL_TX_DUMMY_PACKET_SIZE);
1316 	if (!skb) {
1317 		wl1271_warning("Failed to allocate a dummy packet skb");
1318 		return NULL;
1319 	}
1320 
1321 	skb_reserve(skb, sizeof(struct wl1271_tx_hw_descr));
1322 
1323 	hdr = skb_put_zero(skb, sizeof(*hdr));
1324 	hdr->frame_control = cpu_to_le16(IEEE80211_FTYPE_DATA |
1325 					 IEEE80211_STYPE_NULLFUNC |
1326 					 IEEE80211_FCTL_TODS);
1327 
1328 	skb_put_zero(skb, dummy_packet_size);
1329 
1330 	/* Dummy packets require the TID to be management */
1331 	skb->priority = WL1271_TID_MGMT;
1332 
1333 	/* Initialize all fields that might be used */
1334 	skb_set_queue_mapping(skb, 0);
1335 	memset(IEEE80211_SKB_CB(skb), 0, sizeof(struct ieee80211_tx_info));
1336 
1337 	return skb;
1338 }
1339 
1340 
1341 static int
1342 wl1271_validate_wowlan_pattern(struct cfg80211_pkt_pattern *p)
1343 {
1344 	int num_fields = 0, in_field = 0, fields_size = 0;
1345 	int i, pattern_len = 0;
1346 
1347 	if (!p->mask) {
1348 		wl1271_warning("No mask in WoWLAN pattern");
1349 		return -EINVAL;
1350 	}
1351 
1352 	/*
1353 	 * The pattern is broken up into segments of bytes at different offsets
1354 	 * that need to be checked by the FW filter. Each segment is called
1355 	 * a field in the FW API. We verify that the total number of fields
1356 	 * required for this pattern won't exceed FW limits (8)
1357 	 * as well as the total fields buffer won't exceed the FW limit.
1358 	 * Note that if there's a pattern which crosses Ethernet/IP header
1359 	 * boundary a new field is required.
1360 	 */
1361 	for (i = 0; i < p->pattern_len; i++) {
1362 		if (test_bit(i, (unsigned long *)p->mask)) {
1363 			if (!in_field) {
1364 				in_field = 1;
1365 				pattern_len = 1;
1366 			} else {
1367 				if (i == WL1271_RX_FILTER_ETH_HEADER_SIZE) {
1368 					num_fields++;
1369 					fields_size += pattern_len +
1370 						RX_FILTER_FIELD_OVERHEAD;
1371 					pattern_len = 1;
1372 				} else
1373 					pattern_len++;
1374 			}
1375 		} else {
1376 			if (in_field) {
1377 				in_field = 0;
1378 				fields_size += pattern_len +
1379 					RX_FILTER_FIELD_OVERHEAD;
1380 				num_fields++;
1381 			}
1382 		}
1383 	}
1384 
1385 	if (in_field) {
1386 		fields_size += pattern_len + RX_FILTER_FIELD_OVERHEAD;
1387 		num_fields++;
1388 	}
1389 
1390 	if (num_fields > WL1271_RX_FILTER_MAX_FIELDS) {
1391 		wl1271_warning("RX Filter too complex. Too many segments");
1392 		return -EINVAL;
1393 	}
1394 
1395 	if (fields_size > WL1271_RX_FILTER_MAX_FIELDS_SIZE) {
1396 		wl1271_warning("RX filter pattern is too big");
1397 		return -E2BIG;
1398 	}
1399 
1400 	return 0;
1401 }
1402 
1403 struct wl12xx_rx_filter *wl1271_rx_filter_alloc(void)
1404 {
1405 	return kzalloc(sizeof(struct wl12xx_rx_filter), GFP_KERNEL);
1406 }
1407 
1408 void wl1271_rx_filter_free(struct wl12xx_rx_filter *filter)
1409 {
1410 	int i;
1411 
1412 	if (filter == NULL)
1413 		return;
1414 
1415 	for (i = 0; i < filter->num_fields; i++)
1416 		kfree(filter->fields[i].pattern);
1417 
1418 	kfree(filter);
1419 }
1420 
1421 int wl1271_rx_filter_alloc_field(struct wl12xx_rx_filter *filter,
1422 				 u16 offset, u8 flags,
1423 				 const u8 *pattern, u8 len)
1424 {
1425 	struct wl12xx_rx_filter_field *field;
1426 
1427 	if (filter->num_fields == WL1271_RX_FILTER_MAX_FIELDS) {
1428 		wl1271_warning("Max fields per RX filter. can't alloc another");
1429 		return -EINVAL;
1430 	}
1431 
1432 	field = &filter->fields[filter->num_fields];
1433 
1434 	field->pattern = kmemdup(pattern, len, GFP_KERNEL);
1435 	if (!field->pattern) {
1436 		wl1271_warning("Failed to allocate RX filter pattern");
1437 		return -ENOMEM;
1438 	}
1439 
1440 	filter->num_fields++;
1441 
1442 	field->offset = cpu_to_le16(offset);
1443 	field->flags = flags;
1444 	field->len = len;
1445 
1446 	return 0;
1447 }
1448 
1449 int wl1271_rx_filter_get_fields_size(struct wl12xx_rx_filter *filter)
1450 {
1451 	int i, fields_size = 0;
1452 
1453 	for (i = 0; i < filter->num_fields; i++)
1454 		fields_size += filter->fields[i].len +
1455 			sizeof(struct wl12xx_rx_filter_field) -
1456 			sizeof(u8 *);
1457 
1458 	return fields_size;
1459 }
1460 
1461 void wl1271_rx_filter_flatten_fields(struct wl12xx_rx_filter *filter,
1462 				    u8 *buf)
1463 {
1464 	int i;
1465 	struct wl12xx_rx_filter_field *field;
1466 
1467 	for (i = 0; i < filter->num_fields; i++) {
1468 		field = (struct wl12xx_rx_filter_field *)buf;
1469 
1470 		field->offset = filter->fields[i].offset;
1471 		field->flags = filter->fields[i].flags;
1472 		field->len = filter->fields[i].len;
1473 
1474 		memcpy(&field->pattern, filter->fields[i].pattern, field->len);
1475 		buf += sizeof(struct wl12xx_rx_filter_field) -
1476 			sizeof(u8 *) + field->len;
1477 	}
1478 }
1479 
1480 /*
1481  * Allocates an RX filter returned through f
1482  * which needs to be freed using rx_filter_free()
1483  */
1484 static int
1485 wl1271_convert_wowlan_pattern_to_rx_filter(struct cfg80211_pkt_pattern *p,
1486 					   struct wl12xx_rx_filter **f)
1487 {
1488 	int i, j, ret = 0;
1489 	struct wl12xx_rx_filter *filter;
1490 	u16 offset;
1491 	u8 flags, len;
1492 
1493 	filter = wl1271_rx_filter_alloc();
1494 	if (!filter) {
1495 		wl1271_warning("Failed to alloc rx filter");
1496 		ret = -ENOMEM;
1497 		goto err;
1498 	}
1499 
1500 	i = 0;
1501 	while (i < p->pattern_len) {
1502 		if (!test_bit(i, (unsigned long *)p->mask)) {
1503 			i++;
1504 			continue;
1505 		}
1506 
1507 		for (j = i; j < p->pattern_len; j++) {
1508 			if (!test_bit(j, (unsigned long *)p->mask))
1509 				break;
1510 
1511 			if (i < WL1271_RX_FILTER_ETH_HEADER_SIZE &&
1512 			    j >= WL1271_RX_FILTER_ETH_HEADER_SIZE)
1513 				break;
1514 		}
1515 
1516 		if (i < WL1271_RX_FILTER_ETH_HEADER_SIZE) {
1517 			offset = i;
1518 			flags = WL1271_RX_FILTER_FLAG_ETHERNET_HEADER;
1519 		} else {
1520 			offset = i - WL1271_RX_FILTER_ETH_HEADER_SIZE;
1521 			flags = WL1271_RX_FILTER_FLAG_IP_HEADER;
1522 		}
1523 
1524 		len = j - i;
1525 
1526 		ret = wl1271_rx_filter_alloc_field(filter,
1527 						   offset,
1528 						   flags,
1529 						   &p->pattern[i], len);
1530 		if (ret)
1531 			goto err;
1532 
1533 		i = j;
1534 	}
1535 
1536 	filter->action = FILTER_SIGNAL;
1537 
1538 	*f = filter;
1539 	return 0;
1540 
1541 err:
1542 	wl1271_rx_filter_free(filter);
1543 	*f = NULL;
1544 
1545 	return ret;
1546 }
1547 
1548 static int wl1271_configure_wowlan(struct wl1271 *wl,
1549 				   struct cfg80211_wowlan *wow)
1550 {
1551 	int i, ret;
1552 
1553 	if (!wow || wow->any || !wow->n_patterns) {
1554 		ret = wl1271_acx_default_rx_filter_enable(wl, 0,
1555 							  FILTER_SIGNAL);
1556 		if (ret)
1557 			goto out;
1558 
1559 		ret = wl1271_rx_filter_clear_all(wl);
1560 		if (ret)
1561 			goto out;
1562 
1563 		return 0;
1564 	}
1565 
1566 	if (WARN_ON(wow->n_patterns > WL1271_MAX_RX_FILTERS))
1567 		return -EINVAL;
1568 
1569 	/* Validate all incoming patterns before clearing current FW state */
1570 	for (i = 0; i < wow->n_patterns; i++) {
1571 		ret = wl1271_validate_wowlan_pattern(&wow->patterns[i]);
1572 		if (ret) {
1573 			wl1271_warning("Bad wowlan pattern %d", i);
1574 			return ret;
1575 		}
1576 	}
1577 
1578 	ret = wl1271_acx_default_rx_filter_enable(wl, 0, FILTER_SIGNAL);
1579 	if (ret)
1580 		goto out;
1581 
1582 	ret = wl1271_rx_filter_clear_all(wl);
1583 	if (ret)
1584 		goto out;
1585 
1586 	/* Translate WoWLAN patterns into filters */
1587 	for (i = 0; i < wow->n_patterns; i++) {
1588 		struct cfg80211_pkt_pattern *p;
1589 		struct wl12xx_rx_filter *filter = NULL;
1590 
1591 		p = &wow->patterns[i];
1592 
1593 		ret = wl1271_convert_wowlan_pattern_to_rx_filter(p, &filter);
1594 		if (ret) {
1595 			wl1271_warning("Failed to create an RX filter from "
1596 				       "wowlan pattern %d", i);
1597 			goto out;
1598 		}
1599 
1600 		ret = wl1271_rx_filter_enable(wl, i, 1, filter);
1601 
1602 		wl1271_rx_filter_free(filter);
1603 		if (ret)
1604 			goto out;
1605 	}
1606 
1607 	ret = wl1271_acx_default_rx_filter_enable(wl, 1, FILTER_DROP);
1608 
1609 out:
1610 	return ret;
1611 }
1612 
1613 static int wl1271_configure_suspend_sta(struct wl1271 *wl,
1614 					struct wl12xx_vif *wlvif,
1615 					struct cfg80211_wowlan *wow)
1616 {
1617 	int ret = 0;
1618 
1619 	if (!test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
1620 		goto out;
1621 
1622 	ret = wl1271_configure_wowlan(wl, wow);
1623 	if (ret < 0)
1624 		goto out;
1625 
1626 	if ((wl->conf.conn.suspend_wake_up_event ==
1627 	     wl->conf.conn.wake_up_event) &&
1628 	    (wl->conf.conn.suspend_listen_interval ==
1629 	     wl->conf.conn.listen_interval))
1630 		goto out;
1631 
1632 	ret = wl1271_acx_wake_up_conditions(wl, wlvif,
1633 				    wl->conf.conn.suspend_wake_up_event,
1634 				    wl->conf.conn.suspend_listen_interval);
1635 
1636 	if (ret < 0)
1637 		wl1271_error("suspend: set wake up conditions failed: %d", ret);
1638 out:
1639 	return ret;
1640 
1641 }
1642 
1643 static int wl1271_configure_suspend_ap(struct wl1271 *wl,
1644 					struct wl12xx_vif *wlvif,
1645 					struct cfg80211_wowlan *wow)
1646 {
1647 	int ret = 0;
1648 
1649 	if (!test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags))
1650 		goto out;
1651 
1652 	ret = wl1271_acx_beacon_filter_opt(wl, wlvif, true);
1653 	if (ret < 0)
1654 		goto out;
1655 
1656 	ret = wl1271_configure_wowlan(wl, wow);
1657 	if (ret < 0)
1658 		goto out;
1659 
1660 out:
1661 	return ret;
1662 
1663 }
1664 
1665 static int wl1271_configure_suspend(struct wl1271 *wl,
1666 				    struct wl12xx_vif *wlvif,
1667 				    struct cfg80211_wowlan *wow)
1668 {
1669 	if (wlvif->bss_type == BSS_TYPE_STA_BSS)
1670 		return wl1271_configure_suspend_sta(wl, wlvif, wow);
1671 	if (wlvif->bss_type == BSS_TYPE_AP_BSS)
1672 		return wl1271_configure_suspend_ap(wl, wlvif, wow);
1673 	return 0;
1674 }
1675 
1676 static void wl1271_configure_resume(struct wl1271 *wl, struct wl12xx_vif *wlvif)
1677 {
1678 	int ret = 0;
1679 	bool is_ap = wlvif->bss_type == BSS_TYPE_AP_BSS;
1680 	bool is_sta = wlvif->bss_type == BSS_TYPE_STA_BSS;
1681 
1682 	if ((!is_ap) && (!is_sta))
1683 		return;
1684 
1685 	if ((is_sta && !test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags)) ||
1686 	    (is_ap && !test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags)))
1687 		return;
1688 
1689 	wl1271_configure_wowlan(wl, NULL);
1690 
1691 	if (is_sta) {
1692 		if ((wl->conf.conn.suspend_wake_up_event ==
1693 		     wl->conf.conn.wake_up_event) &&
1694 		    (wl->conf.conn.suspend_listen_interval ==
1695 		     wl->conf.conn.listen_interval))
1696 			return;
1697 
1698 		ret = wl1271_acx_wake_up_conditions(wl, wlvif,
1699 				    wl->conf.conn.wake_up_event,
1700 				    wl->conf.conn.listen_interval);
1701 
1702 		if (ret < 0)
1703 			wl1271_error("resume: wake up conditions failed: %d",
1704 				     ret);
1705 
1706 	} else if (is_ap) {
1707 		ret = wl1271_acx_beacon_filter_opt(wl, wlvif, false);
1708 	}
1709 }
1710 
1711 static int __maybe_unused wl1271_op_suspend(struct ieee80211_hw *hw,
1712 					    struct cfg80211_wowlan *wow)
1713 {
1714 	struct wl1271 *wl = hw->priv;
1715 	struct wl12xx_vif *wlvif;
1716 	unsigned long flags;
1717 	int ret;
1718 
1719 	wl1271_debug(DEBUG_MAC80211, "mac80211 suspend wow=%d", !!wow);
1720 	WARN_ON(!wow);
1721 
1722 	/* we want to perform the recovery before suspending */
1723 	if (test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags)) {
1724 		wl1271_warning("postponing suspend to perform recovery");
1725 		return -EBUSY;
1726 	}
1727 
1728 	wl1271_tx_flush(wl);
1729 
1730 	mutex_lock(&wl->mutex);
1731 
1732 	ret = pm_runtime_resume_and_get(wl->dev);
1733 	if (ret < 0) {
1734 		mutex_unlock(&wl->mutex);
1735 		return ret;
1736 	}
1737 
1738 	wl->wow_enabled = true;
1739 	wl12xx_for_each_wlvif(wl, wlvif) {
1740 		if (wlcore_is_p2p_mgmt(wlvif))
1741 			continue;
1742 
1743 		ret = wl1271_configure_suspend(wl, wlvif, wow);
1744 		if (ret < 0) {
1745 			goto out_sleep;
1746 		}
1747 	}
1748 
1749 	/* disable fast link flow control notifications from FW */
1750 	ret = wlcore_hw_interrupt_notify(wl, false);
1751 	if (ret < 0)
1752 		goto out_sleep;
1753 
1754 	/* if filtering is enabled, configure the FW to drop all RX BA frames */
1755 	ret = wlcore_hw_rx_ba_filter(wl,
1756 				     !!wl->conf.conn.suspend_rx_ba_activity);
1757 	if (ret < 0)
1758 		goto out_sleep;
1759 
1760 out_sleep:
1761 	pm_runtime_put_noidle(wl->dev);
1762 	mutex_unlock(&wl->mutex);
1763 
1764 	if (ret < 0) {
1765 		wl1271_warning("couldn't prepare device to suspend");
1766 		return ret;
1767 	}
1768 
1769 	/* flush any remaining work */
1770 	wl1271_debug(DEBUG_MAC80211, "flushing remaining works");
1771 
1772 	flush_work(&wl->tx_work);
1773 
1774 	/*
1775 	 * Cancel the watchdog even if above tx_flush failed. We will detect
1776 	 * it on resume anyway.
1777 	 */
1778 	cancel_delayed_work(&wl->tx_watchdog_work);
1779 
1780 	/*
1781 	 * set suspended flag to avoid triggering a new threaded_irq
1782 	 * work.
1783 	 */
1784 	spin_lock_irqsave(&wl->wl_lock, flags);
1785 	set_bit(WL1271_FLAG_SUSPENDED, &wl->flags);
1786 	spin_unlock_irqrestore(&wl->wl_lock, flags);
1787 
1788 	return pm_runtime_force_suspend(wl->dev);
1789 }
1790 
1791 static int __maybe_unused wl1271_op_resume(struct ieee80211_hw *hw)
1792 {
1793 	struct wl1271 *wl = hw->priv;
1794 	struct wl12xx_vif *wlvif;
1795 	unsigned long flags;
1796 	bool run_irq_work = false, pending_recovery;
1797 	int ret;
1798 
1799 	wl1271_debug(DEBUG_MAC80211, "mac80211 resume wow=%d",
1800 		     wl->wow_enabled);
1801 	WARN_ON(!wl->wow_enabled);
1802 
1803 	ret = pm_runtime_force_resume(wl->dev);
1804 	if (ret < 0) {
1805 		wl1271_error("ELP wakeup failure!");
1806 		goto out_sleep;
1807 	}
1808 
1809 	/*
1810 	 * re-enable irq_work enqueuing, and call irq_work directly if
1811 	 * there is a pending work.
1812 	 */
1813 	spin_lock_irqsave(&wl->wl_lock, flags);
1814 	clear_bit(WL1271_FLAG_SUSPENDED, &wl->flags);
1815 	if (test_and_clear_bit(WL1271_FLAG_PENDING_WORK, &wl->flags))
1816 		run_irq_work = true;
1817 	spin_unlock_irqrestore(&wl->wl_lock, flags);
1818 
1819 	mutex_lock(&wl->mutex);
1820 
1821 	/* test the recovery flag before calling any SDIO functions */
1822 	pending_recovery = test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS,
1823 				    &wl->flags);
1824 
1825 	if (run_irq_work) {
1826 		wl1271_debug(DEBUG_MAC80211,
1827 			     "run postponed irq_work directly");
1828 
1829 		/* don't talk to the HW if recovery is pending */
1830 		if (!pending_recovery) {
1831 			ret = wlcore_irq_locked(wl);
1832 			if (ret)
1833 				wl12xx_queue_recovery_work(wl);
1834 		}
1835 
1836 		wlcore_enable_interrupts(wl);
1837 	}
1838 
1839 	if (pending_recovery) {
1840 		wl1271_warning("queuing forgotten recovery on resume");
1841 		ieee80211_queue_work(wl->hw, &wl->recovery_work);
1842 		goto out_sleep;
1843 	}
1844 
1845 	ret = pm_runtime_resume_and_get(wl->dev);
1846 	if (ret < 0)
1847 		goto out;
1848 
1849 	wl12xx_for_each_wlvif(wl, wlvif) {
1850 		if (wlcore_is_p2p_mgmt(wlvif))
1851 			continue;
1852 
1853 		wl1271_configure_resume(wl, wlvif);
1854 	}
1855 
1856 	ret = wlcore_hw_interrupt_notify(wl, true);
1857 	if (ret < 0)
1858 		goto out_sleep;
1859 
1860 	/* if filtering is enabled, configure the FW to drop all RX BA frames */
1861 	ret = wlcore_hw_rx_ba_filter(wl, false);
1862 	if (ret < 0)
1863 		goto out_sleep;
1864 
1865 out_sleep:
1866 	pm_runtime_mark_last_busy(wl->dev);
1867 	pm_runtime_put_autosuspend(wl->dev);
1868 
1869 out:
1870 	wl->wow_enabled = false;
1871 
1872 	/*
1873 	 * Set a flag to re-init the watchdog on the first Tx after resume.
1874 	 * That way we avoid possible conditions where Tx-complete interrupts
1875 	 * fail to arrive and we perform a spurious recovery.
1876 	 */
1877 	set_bit(WL1271_FLAG_REINIT_TX_WDOG, &wl->flags);
1878 	mutex_unlock(&wl->mutex);
1879 
1880 	return 0;
1881 }
1882 
1883 static int wl1271_op_start(struct ieee80211_hw *hw)
1884 {
1885 	wl1271_debug(DEBUG_MAC80211, "mac80211 start");
1886 
1887 	/*
1888 	 * We have to delay the booting of the hardware because
1889 	 * we need to know the local MAC address before downloading and
1890 	 * initializing the firmware. The MAC address cannot be changed
1891 	 * after boot, and without the proper MAC address, the firmware
1892 	 * will not function properly.
1893 	 *
1894 	 * The MAC address is first known when the corresponding interface
1895 	 * is added. That is where we will initialize the hardware.
1896 	 */
1897 
1898 	return 0;
1899 }
1900 
1901 static void wlcore_op_stop_locked(struct wl1271 *wl)
1902 {
1903 	int i;
1904 
1905 	if (wl->state == WLCORE_STATE_OFF) {
1906 		if (test_and_clear_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS,
1907 					&wl->flags))
1908 			wlcore_enable_interrupts(wl);
1909 
1910 		return;
1911 	}
1912 
1913 	/*
1914 	 * this must be before the cancel_work calls below, so that the work
1915 	 * functions don't perform further work.
1916 	 */
1917 	wl->state = WLCORE_STATE_OFF;
1918 
1919 	/*
1920 	 * Use the nosync variant to disable interrupts, so the mutex could be
1921 	 * held while doing so without deadlocking.
1922 	 */
1923 	wlcore_disable_interrupts_nosync(wl);
1924 
1925 	mutex_unlock(&wl->mutex);
1926 
1927 	wlcore_synchronize_interrupts(wl);
1928 	if (!test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags))
1929 		cancel_work_sync(&wl->recovery_work);
1930 	wl1271_flush_deferred_work(wl);
1931 	cancel_delayed_work_sync(&wl->scan_complete_work);
1932 	cancel_work_sync(&wl->netstack_work);
1933 	cancel_work_sync(&wl->tx_work);
1934 	cancel_delayed_work_sync(&wl->tx_watchdog_work);
1935 
1936 	/* let's notify MAC80211 about the remaining pending TX frames */
1937 	mutex_lock(&wl->mutex);
1938 	wl12xx_tx_reset(wl);
1939 
1940 	wl1271_power_off(wl);
1941 	/*
1942 	 * In case a recovery was scheduled, interrupts were disabled to avoid
1943 	 * an interrupt storm. Now that the power is down, it is safe to
1944 	 * re-enable interrupts to balance the disable depth
1945 	 */
1946 	if (test_and_clear_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags))
1947 		wlcore_enable_interrupts(wl);
1948 
1949 	wl->band = NL80211_BAND_2GHZ;
1950 
1951 	wl->rx_counter = 0;
1952 	wl->power_level = WL1271_DEFAULT_POWER_LEVEL;
1953 	wl->channel_type = NL80211_CHAN_NO_HT;
1954 	wl->tx_blocks_available = 0;
1955 	wl->tx_allocated_blocks = 0;
1956 	wl->tx_results_count = 0;
1957 	wl->tx_packets_count = 0;
1958 	wl->time_offset = 0;
1959 	wl->ap_fw_ps_map = 0;
1960 	wl->ap_ps_map = 0;
1961 	wl->sleep_auth = WL1271_PSM_ILLEGAL;
1962 	memset(wl->roles_map, 0, sizeof(wl->roles_map));
1963 	memset(wl->links_map, 0, sizeof(wl->links_map));
1964 	memset(wl->roc_map, 0, sizeof(wl->roc_map));
1965 	memset(wl->session_ids, 0, sizeof(wl->session_ids));
1966 	memset(wl->rx_filter_enabled, 0, sizeof(wl->rx_filter_enabled));
1967 	wl->active_sta_count = 0;
1968 	wl->active_link_count = 0;
1969 
1970 	/* The system link is always allocated */
1971 	wl->links[WL12XX_SYSTEM_HLID].allocated_pkts = 0;
1972 	wl->links[WL12XX_SYSTEM_HLID].prev_freed_pkts = 0;
1973 	__set_bit(WL12XX_SYSTEM_HLID, wl->links_map);
1974 
1975 	/*
1976 	 * this is performed after the cancel_work calls and the associated
1977 	 * mutex_lock, so that wl1271_op_add_interface does not accidentally
1978 	 * get executed before all these vars have been reset.
1979 	 */
1980 	wl->flags = 0;
1981 
1982 	wl->tx_blocks_freed = 0;
1983 
1984 	for (i = 0; i < NUM_TX_QUEUES; i++) {
1985 		wl->tx_pkts_freed[i] = 0;
1986 		wl->tx_allocated_pkts[i] = 0;
1987 	}
1988 
1989 	wl1271_debugfs_reset(wl);
1990 
1991 	kfree(wl->raw_fw_status);
1992 	wl->raw_fw_status = NULL;
1993 	kfree(wl->fw_status);
1994 	wl->fw_status = NULL;
1995 	kfree(wl->tx_res_if);
1996 	wl->tx_res_if = NULL;
1997 	kfree(wl->target_mem_map);
1998 	wl->target_mem_map = NULL;
1999 
2000 	/*
2001 	 * FW channels must be re-calibrated after recovery,
2002 	 * save current Reg-Domain channel configuration and clear it.
2003 	 */
2004 	memcpy(wl->reg_ch_conf_pending, wl->reg_ch_conf_last,
2005 	       sizeof(wl->reg_ch_conf_pending));
2006 	memset(wl->reg_ch_conf_last, 0, sizeof(wl->reg_ch_conf_last));
2007 }
2008 
2009 static void wlcore_op_stop(struct ieee80211_hw *hw)
2010 {
2011 	struct wl1271 *wl = hw->priv;
2012 
2013 	wl1271_debug(DEBUG_MAC80211, "mac80211 stop");
2014 
2015 	mutex_lock(&wl->mutex);
2016 
2017 	wlcore_op_stop_locked(wl);
2018 
2019 	mutex_unlock(&wl->mutex);
2020 }
2021 
2022 static void wlcore_channel_switch_work(struct work_struct *work)
2023 {
2024 	struct delayed_work *dwork;
2025 	struct wl1271 *wl;
2026 	struct ieee80211_vif *vif;
2027 	struct wl12xx_vif *wlvif;
2028 	int ret;
2029 
2030 	dwork = to_delayed_work(work);
2031 	wlvif = container_of(dwork, struct wl12xx_vif, channel_switch_work);
2032 	wl = wlvif->wl;
2033 
2034 	wl1271_info("channel switch failed (role_id: %d).", wlvif->role_id);
2035 
2036 	mutex_lock(&wl->mutex);
2037 
2038 	if (unlikely(wl->state != WLCORE_STATE_ON))
2039 		goto out;
2040 
2041 	/* check the channel switch is still ongoing */
2042 	if (!test_and_clear_bit(WLVIF_FLAG_CS_PROGRESS, &wlvif->flags))
2043 		goto out;
2044 
2045 	vif = wl12xx_wlvif_to_vif(wlvif);
2046 	ieee80211_chswitch_done(vif, false);
2047 
2048 	ret = pm_runtime_resume_and_get(wl->dev);
2049 	if (ret < 0)
2050 		goto out;
2051 
2052 	wl12xx_cmd_stop_channel_switch(wl, wlvif);
2053 
2054 	pm_runtime_mark_last_busy(wl->dev);
2055 	pm_runtime_put_autosuspend(wl->dev);
2056 out:
2057 	mutex_unlock(&wl->mutex);
2058 }
2059 
2060 static void wlcore_connection_loss_work(struct work_struct *work)
2061 {
2062 	struct delayed_work *dwork;
2063 	struct wl1271 *wl;
2064 	struct ieee80211_vif *vif;
2065 	struct wl12xx_vif *wlvif;
2066 
2067 	dwork = to_delayed_work(work);
2068 	wlvif = container_of(dwork, struct wl12xx_vif, connection_loss_work);
2069 	wl = wlvif->wl;
2070 
2071 	wl1271_info("Connection loss work (role_id: %d).", wlvif->role_id);
2072 
2073 	mutex_lock(&wl->mutex);
2074 
2075 	if (unlikely(wl->state != WLCORE_STATE_ON))
2076 		goto out;
2077 
2078 	/* Call mac80211 connection loss */
2079 	if (!test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
2080 		goto out;
2081 
2082 	vif = wl12xx_wlvif_to_vif(wlvif);
2083 	ieee80211_connection_loss(vif);
2084 out:
2085 	mutex_unlock(&wl->mutex);
2086 }
2087 
2088 static void wlcore_pending_auth_complete_work(struct work_struct *work)
2089 {
2090 	struct delayed_work *dwork;
2091 	struct wl1271 *wl;
2092 	struct wl12xx_vif *wlvif;
2093 	unsigned long time_spare;
2094 	int ret;
2095 
2096 	dwork = to_delayed_work(work);
2097 	wlvif = container_of(dwork, struct wl12xx_vif,
2098 			     pending_auth_complete_work);
2099 	wl = wlvif->wl;
2100 
2101 	mutex_lock(&wl->mutex);
2102 
2103 	if (unlikely(wl->state != WLCORE_STATE_ON))
2104 		goto out;
2105 
2106 	/*
2107 	 * Make sure a second really passed since the last auth reply. Maybe
2108 	 * a second auth reply arrived while we were stuck on the mutex.
2109 	 * Check for a little less than the timeout to protect from scheduler
2110 	 * irregularities.
2111 	 */
2112 	time_spare = jiffies +
2113 			msecs_to_jiffies(WLCORE_PEND_AUTH_ROC_TIMEOUT - 50);
2114 	if (!time_after(time_spare, wlvif->pending_auth_reply_time))
2115 		goto out;
2116 
2117 	ret = pm_runtime_resume_and_get(wl->dev);
2118 	if (ret < 0)
2119 		goto out;
2120 
2121 	/* cancel the ROC if active */
2122 	wlcore_update_inconn_sta(wl, wlvif, NULL, false);
2123 
2124 	pm_runtime_mark_last_busy(wl->dev);
2125 	pm_runtime_put_autosuspend(wl->dev);
2126 out:
2127 	mutex_unlock(&wl->mutex);
2128 }
2129 
2130 static int wl12xx_allocate_rate_policy(struct wl1271 *wl, u8 *idx)
2131 {
2132 	u8 policy = find_first_zero_bit(wl->rate_policies_map,
2133 					WL12XX_MAX_RATE_POLICIES);
2134 	if (policy >= WL12XX_MAX_RATE_POLICIES)
2135 		return -EBUSY;
2136 
2137 	__set_bit(policy, wl->rate_policies_map);
2138 	*idx = policy;
2139 	return 0;
2140 }
2141 
2142 static void wl12xx_free_rate_policy(struct wl1271 *wl, u8 *idx)
2143 {
2144 	if (WARN_ON(*idx >= WL12XX_MAX_RATE_POLICIES))
2145 		return;
2146 
2147 	__clear_bit(*idx, wl->rate_policies_map);
2148 	*idx = WL12XX_MAX_RATE_POLICIES;
2149 }
2150 
2151 static int wlcore_allocate_klv_template(struct wl1271 *wl, u8 *idx)
2152 {
2153 	u8 policy = find_first_zero_bit(wl->klv_templates_map,
2154 					WLCORE_MAX_KLV_TEMPLATES);
2155 	if (policy >= WLCORE_MAX_KLV_TEMPLATES)
2156 		return -EBUSY;
2157 
2158 	__set_bit(policy, wl->klv_templates_map);
2159 	*idx = policy;
2160 	return 0;
2161 }
2162 
2163 static void wlcore_free_klv_template(struct wl1271 *wl, u8 *idx)
2164 {
2165 	if (WARN_ON(*idx >= WLCORE_MAX_KLV_TEMPLATES))
2166 		return;
2167 
2168 	__clear_bit(*idx, wl->klv_templates_map);
2169 	*idx = WLCORE_MAX_KLV_TEMPLATES;
2170 }
2171 
2172 static u8 wl12xx_get_role_type(struct wl1271 *wl, struct wl12xx_vif *wlvif)
2173 {
2174 	struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
2175 
2176 	switch (wlvif->bss_type) {
2177 	case BSS_TYPE_AP_BSS:
2178 		if (wlvif->p2p)
2179 			return WL1271_ROLE_P2P_GO;
2180 		else if (ieee80211_vif_is_mesh(vif))
2181 			return WL1271_ROLE_MESH_POINT;
2182 		else
2183 			return WL1271_ROLE_AP;
2184 
2185 	case BSS_TYPE_STA_BSS:
2186 		if (wlvif->p2p)
2187 			return WL1271_ROLE_P2P_CL;
2188 		else
2189 			return WL1271_ROLE_STA;
2190 
2191 	case BSS_TYPE_IBSS:
2192 		return WL1271_ROLE_IBSS;
2193 
2194 	default:
2195 		wl1271_error("invalid bss_type: %d", wlvif->bss_type);
2196 	}
2197 	return WL12XX_INVALID_ROLE_TYPE;
2198 }
2199 
2200 static int wl12xx_init_vif_data(struct wl1271 *wl, struct ieee80211_vif *vif)
2201 {
2202 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
2203 	int i;
2204 
2205 	/* clear everything but the persistent data */
2206 	memset(wlvif, 0, offsetof(struct wl12xx_vif, persistent));
2207 
2208 	switch (ieee80211_vif_type_p2p(vif)) {
2209 	case NL80211_IFTYPE_P2P_CLIENT:
2210 		wlvif->p2p = 1;
2211 		fallthrough;
2212 	case NL80211_IFTYPE_STATION:
2213 	case NL80211_IFTYPE_P2P_DEVICE:
2214 		wlvif->bss_type = BSS_TYPE_STA_BSS;
2215 		break;
2216 	case NL80211_IFTYPE_ADHOC:
2217 		wlvif->bss_type = BSS_TYPE_IBSS;
2218 		break;
2219 	case NL80211_IFTYPE_P2P_GO:
2220 		wlvif->p2p = 1;
2221 		fallthrough;
2222 	case NL80211_IFTYPE_AP:
2223 	case NL80211_IFTYPE_MESH_POINT:
2224 		wlvif->bss_type = BSS_TYPE_AP_BSS;
2225 		break;
2226 	default:
2227 		wlvif->bss_type = MAX_BSS_TYPE;
2228 		return -EOPNOTSUPP;
2229 	}
2230 
2231 	wlvif->role_id = WL12XX_INVALID_ROLE_ID;
2232 	wlvif->dev_role_id = WL12XX_INVALID_ROLE_ID;
2233 	wlvif->dev_hlid = WL12XX_INVALID_LINK_ID;
2234 
2235 	if (wlvif->bss_type == BSS_TYPE_STA_BSS ||
2236 	    wlvif->bss_type == BSS_TYPE_IBSS) {
2237 		/* init sta/ibss data */
2238 		wlvif->sta.hlid = WL12XX_INVALID_LINK_ID;
2239 		wl12xx_allocate_rate_policy(wl, &wlvif->sta.basic_rate_idx);
2240 		wl12xx_allocate_rate_policy(wl, &wlvif->sta.ap_rate_idx);
2241 		wl12xx_allocate_rate_policy(wl, &wlvif->sta.p2p_rate_idx);
2242 		wlcore_allocate_klv_template(wl, &wlvif->sta.klv_template_id);
2243 		wlvif->basic_rate_set = CONF_TX_RATE_MASK_BASIC;
2244 		wlvif->basic_rate = CONF_TX_RATE_MASK_BASIC;
2245 		wlvif->rate_set = CONF_TX_RATE_MASK_BASIC;
2246 	} else {
2247 		/* init ap data */
2248 		wlvif->ap.bcast_hlid = WL12XX_INVALID_LINK_ID;
2249 		wlvif->ap.global_hlid = WL12XX_INVALID_LINK_ID;
2250 		wl12xx_allocate_rate_policy(wl, &wlvif->ap.mgmt_rate_idx);
2251 		wl12xx_allocate_rate_policy(wl, &wlvif->ap.bcast_rate_idx);
2252 		for (i = 0; i < CONF_TX_MAX_AC_COUNT; i++)
2253 			wl12xx_allocate_rate_policy(wl,
2254 						&wlvif->ap.ucast_rate_idx[i]);
2255 		wlvif->basic_rate_set = CONF_TX_ENABLED_RATES;
2256 		/*
2257 		 * TODO: check if basic_rate shouldn't be
2258 		 * wl1271_tx_min_rate_get(wl, wlvif->basic_rate_set);
2259 		 * instead (the same thing for STA above).
2260 		*/
2261 		wlvif->basic_rate = CONF_TX_ENABLED_RATES;
2262 		/* TODO: this seems to be used only for STA, check it */
2263 		wlvif->rate_set = CONF_TX_ENABLED_RATES;
2264 	}
2265 
2266 	wlvif->bitrate_masks[NL80211_BAND_2GHZ] = wl->conf.tx.basic_rate;
2267 	wlvif->bitrate_masks[NL80211_BAND_5GHZ] = wl->conf.tx.basic_rate_5;
2268 	wlvif->beacon_int = WL1271_DEFAULT_BEACON_INT;
2269 
2270 	/*
2271 	 * mac80211 configures some values globally, while we treat them
2272 	 * per-interface. thus, on init, we have to copy them from wl
2273 	 */
2274 	wlvif->band = wl->band;
2275 	wlvif->channel = wl->channel;
2276 	wlvif->power_level = wl->power_level;
2277 	wlvif->channel_type = wl->channel_type;
2278 
2279 	INIT_WORK(&wlvif->rx_streaming_enable_work,
2280 		  wl1271_rx_streaming_enable_work);
2281 	INIT_WORK(&wlvif->rx_streaming_disable_work,
2282 		  wl1271_rx_streaming_disable_work);
2283 	INIT_WORK(&wlvif->rc_update_work, wlcore_rc_update_work);
2284 	INIT_DELAYED_WORK(&wlvif->channel_switch_work,
2285 			  wlcore_channel_switch_work);
2286 	INIT_DELAYED_WORK(&wlvif->connection_loss_work,
2287 			  wlcore_connection_loss_work);
2288 	INIT_DELAYED_WORK(&wlvif->pending_auth_complete_work,
2289 			  wlcore_pending_auth_complete_work);
2290 	INIT_LIST_HEAD(&wlvif->list);
2291 
2292 	timer_setup(&wlvif->rx_streaming_timer, wl1271_rx_streaming_timer, 0);
2293 	return 0;
2294 }
2295 
2296 static int wl12xx_init_fw(struct wl1271 *wl)
2297 {
2298 	int retries = WL1271_BOOT_RETRIES;
2299 	bool booted = false;
2300 	struct wiphy *wiphy = wl->hw->wiphy;
2301 	int ret;
2302 
2303 	while (retries) {
2304 		retries--;
2305 		ret = wl12xx_chip_wakeup(wl, false);
2306 		if (ret < 0)
2307 			goto power_off;
2308 
2309 		ret = wl->ops->boot(wl);
2310 		if (ret < 0)
2311 			goto power_off;
2312 
2313 		ret = wl1271_hw_init(wl);
2314 		if (ret < 0)
2315 			goto irq_disable;
2316 
2317 		booted = true;
2318 		break;
2319 
2320 irq_disable:
2321 		mutex_unlock(&wl->mutex);
2322 		/* Unlocking the mutex in the middle of handling is
2323 		   inherently unsafe. In this case we deem it safe to do,
2324 		   because we need to let any possibly pending IRQ out of
2325 		   the system (and while we are WLCORE_STATE_OFF the IRQ
2326 		   work function will not do anything.) Also, any other
2327 		   possible concurrent operations will fail due to the
2328 		   current state, hence the wl1271 struct should be safe. */
2329 		wlcore_disable_interrupts(wl);
2330 		wl1271_flush_deferred_work(wl);
2331 		cancel_work_sync(&wl->netstack_work);
2332 		mutex_lock(&wl->mutex);
2333 power_off:
2334 		wl1271_power_off(wl);
2335 	}
2336 
2337 	if (!booted) {
2338 		wl1271_error("firmware boot failed despite %d retries",
2339 			     WL1271_BOOT_RETRIES);
2340 		goto out;
2341 	}
2342 
2343 	wl1271_info("firmware booted (%s)", wl->chip.fw_ver_str);
2344 
2345 	/* update hw/fw version info in wiphy struct */
2346 	wiphy->hw_version = wl->chip.id;
2347 	strncpy(wiphy->fw_version, wl->chip.fw_ver_str,
2348 		sizeof(wiphy->fw_version));
2349 
2350 	/*
2351 	 * Now we know if 11a is supported (info from the NVS), so disable
2352 	 * 11a channels if not supported
2353 	 */
2354 	if (!wl->enable_11a)
2355 		wiphy->bands[NL80211_BAND_5GHZ]->n_channels = 0;
2356 
2357 	wl1271_debug(DEBUG_MAC80211, "11a is %ssupported",
2358 		     wl->enable_11a ? "" : "not ");
2359 
2360 	wl->state = WLCORE_STATE_ON;
2361 out:
2362 	return ret;
2363 }
2364 
2365 static bool wl12xx_dev_role_started(struct wl12xx_vif *wlvif)
2366 {
2367 	return wlvif->dev_hlid != WL12XX_INVALID_LINK_ID;
2368 }
2369 
2370 /*
2371  * Check whether a fw switch (i.e. moving from one loaded
2372  * fw to another) is needed. This function is also responsible
2373  * for updating wl->last_vif_count, so it must be called before
2374  * loading a non-plt fw (so the correct fw (single-role/multi-role)
2375  * will be used).
2376  */
2377 static bool wl12xx_need_fw_change(struct wl1271 *wl,
2378 				  struct vif_counter_data vif_counter_data,
2379 				  bool add)
2380 {
2381 	enum wl12xx_fw_type current_fw = wl->fw_type;
2382 	u8 vif_count = vif_counter_data.counter;
2383 
2384 	if (test_bit(WL1271_FLAG_VIF_CHANGE_IN_PROGRESS, &wl->flags))
2385 		return false;
2386 
2387 	/* increase the vif count if this is a new vif */
2388 	if (add && !vif_counter_data.cur_vif_running)
2389 		vif_count++;
2390 
2391 	wl->last_vif_count = vif_count;
2392 
2393 	/* no need for fw change if the device is OFF */
2394 	if (wl->state == WLCORE_STATE_OFF)
2395 		return false;
2396 
2397 	/* no need for fw change if a single fw is used */
2398 	if (!wl->mr_fw_name)
2399 		return false;
2400 
2401 	if (vif_count > 1 && current_fw == WL12XX_FW_TYPE_NORMAL)
2402 		return true;
2403 	if (vif_count <= 1 && current_fw == WL12XX_FW_TYPE_MULTI)
2404 		return true;
2405 
2406 	return false;
2407 }
2408 
2409 /*
2410  * Enter "forced psm". Make sure the sta is in psm against the ap,
2411  * to make the fw switch a bit more disconnection-persistent.
2412  */
2413 static void wl12xx_force_active_psm(struct wl1271 *wl)
2414 {
2415 	struct wl12xx_vif *wlvif;
2416 
2417 	wl12xx_for_each_wlvif_sta(wl, wlvif) {
2418 		wl1271_ps_set_mode(wl, wlvif, STATION_POWER_SAVE_MODE);
2419 	}
2420 }
2421 
2422 struct wlcore_hw_queue_iter_data {
2423 	unsigned long hw_queue_map[BITS_TO_LONGS(WLCORE_NUM_MAC_ADDRESSES)];
2424 	/* current vif */
2425 	struct ieee80211_vif *vif;
2426 	/* is the current vif among those iterated */
2427 	bool cur_running;
2428 };
2429 
2430 static void wlcore_hw_queue_iter(void *data, u8 *mac,
2431 				 struct ieee80211_vif *vif)
2432 {
2433 	struct wlcore_hw_queue_iter_data *iter_data = data;
2434 
2435 	if (vif->type == NL80211_IFTYPE_P2P_DEVICE ||
2436 	    WARN_ON_ONCE(vif->hw_queue[0] == IEEE80211_INVAL_HW_QUEUE))
2437 		return;
2438 
2439 	if (iter_data->cur_running || vif == iter_data->vif) {
2440 		iter_data->cur_running = true;
2441 		return;
2442 	}
2443 
2444 	__set_bit(vif->hw_queue[0] / NUM_TX_QUEUES, iter_data->hw_queue_map);
2445 }
2446 
2447 static int wlcore_allocate_hw_queue_base(struct wl1271 *wl,
2448 					 struct wl12xx_vif *wlvif)
2449 {
2450 	struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
2451 	struct wlcore_hw_queue_iter_data iter_data = {};
2452 	int i, q_base;
2453 
2454 	if (vif->type == NL80211_IFTYPE_P2P_DEVICE) {
2455 		vif->cab_queue = IEEE80211_INVAL_HW_QUEUE;
2456 		return 0;
2457 	}
2458 
2459 	iter_data.vif = vif;
2460 
2461 	/* mark all bits taken by active interfaces */
2462 	ieee80211_iterate_active_interfaces_atomic(wl->hw,
2463 					IEEE80211_IFACE_ITER_RESUME_ALL,
2464 					wlcore_hw_queue_iter, &iter_data);
2465 
2466 	/* the current vif is already running in mac80211 (resume/recovery) */
2467 	if (iter_data.cur_running) {
2468 		wlvif->hw_queue_base = vif->hw_queue[0];
2469 		wl1271_debug(DEBUG_MAC80211,
2470 			     "using pre-allocated hw queue base %d",
2471 			     wlvif->hw_queue_base);
2472 
2473 		/* interface type might have changed type */
2474 		goto adjust_cab_queue;
2475 	}
2476 
2477 	q_base = find_first_zero_bit(iter_data.hw_queue_map,
2478 				     WLCORE_NUM_MAC_ADDRESSES);
2479 	if (q_base >= WLCORE_NUM_MAC_ADDRESSES)
2480 		return -EBUSY;
2481 
2482 	wlvif->hw_queue_base = q_base * NUM_TX_QUEUES;
2483 	wl1271_debug(DEBUG_MAC80211, "allocating hw queue base: %d",
2484 		     wlvif->hw_queue_base);
2485 
2486 	for (i = 0; i < NUM_TX_QUEUES; i++) {
2487 		wl->queue_stop_reasons[wlvif->hw_queue_base + i] = 0;
2488 		/* register hw queues in mac80211 */
2489 		vif->hw_queue[i] = wlvif->hw_queue_base + i;
2490 	}
2491 
2492 adjust_cab_queue:
2493 	/* the last places are reserved for cab queues per interface */
2494 	if (wlvif->bss_type == BSS_TYPE_AP_BSS)
2495 		vif->cab_queue = NUM_TX_QUEUES * WLCORE_NUM_MAC_ADDRESSES +
2496 				 wlvif->hw_queue_base / NUM_TX_QUEUES;
2497 	else
2498 		vif->cab_queue = IEEE80211_INVAL_HW_QUEUE;
2499 
2500 	return 0;
2501 }
2502 
2503 static int wl1271_op_add_interface(struct ieee80211_hw *hw,
2504 				   struct ieee80211_vif *vif)
2505 {
2506 	struct wl1271 *wl = hw->priv;
2507 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
2508 	struct vif_counter_data vif_count;
2509 	int ret = 0;
2510 	u8 role_type;
2511 
2512 	if (wl->plt) {
2513 		wl1271_error("Adding Interface not allowed while in PLT mode");
2514 		return -EBUSY;
2515 	}
2516 
2517 	vif->driver_flags |= IEEE80211_VIF_BEACON_FILTER |
2518 			     IEEE80211_VIF_SUPPORTS_UAPSD |
2519 			     IEEE80211_VIF_SUPPORTS_CQM_RSSI;
2520 
2521 	wl1271_debug(DEBUG_MAC80211, "mac80211 add interface type %d mac %pM",
2522 		     ieee80211_vif_type_p2p(vif), vif->addr);
2523 
2524 	wl12xx_get_vif_count(hw, vif, &vif_count);
2525 
2526 	mutex_lock(&wl->mutex);
2527 
2528 	/*
2529 	 * in some very corner case HW recovery scenarios its possible to
2530 	 * get here before __wl1271_op_remove_interface is complete, so
2531 	 * opt out if that is the case.
2532 	 */
2533 	if (test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags) ||
2534 	    test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags)) {
2535 		ret = -EBUSY;
2536 		goto out;
2537 	}
2538 
2539 
2540 	ret = wl12xx_init_vif_data(wl, vif);
2541 	if (ret < 0)
2542 		goto out;
2543 
2544 	wlvif->wl = wl;
2545 	role_type = wl12xx_get_role_type(wl, wlvif);
2546 	if (role_type == WL12XX_INVALID_ROLE_TYPE) {
2547 		ret = -EINVAL;
2548 		goto out;
2549 	}
2550 
2551 	ret = wlcore_allocate_hw_queue_base(wl, wlvif);
2552 	if (ret < 0)
2553 		goto out;
2554 
2555 	/*
2556 	 * TODO: after the nvs issue will be solved, move this block
2557 	 * to start(), and make sure here the driver is ON.
2558 	 */
2559 	if (wl->state == WLCORE_STATE_OFF) {
2560 		/*
2561 		 * we still need this in order to configure the fw
2562 		 * while uploading the nvs
2563 		 */
2564 		memcpy(wl->addresses[0].addr, vif->addr, ETH_ALEN);
2565 
2566 		ret = wl12xx_init_fw(wl);
2567 		if (ret < 0)
2568 			goto out;
2569 	}
2570 
2571 	/*
2572 	 * Call runtime PM only after possible wl12xx_init_fw() above
2573 	 * is done. Otherwise we do not have interrupts enabled.
2574 	 */
2575 	ret = pm_runtime_resume_and_get(wl->dev);
2576 	if (ret < 0)
2577 		goto out_unlock;
2578 
2579 	if (wl12xx_need_fw_change(wl, vif_count, true)) {
2580 		wl12xx_force_active_psm(wl);
2581 		set_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags);
2582 		mutex_unlock(&wl->mutex);
2583 		wl1271_recovery_work(&wl->recovery_work);
2584 		return 0;
2585 	}
2586 
2587 	if (!wlcore_is_p2p_mgmt(wlvif)) {
2588 		ret = wl12xx_cmd_role_enable(wl, vif->addr,
2589 					     role_type, &wlvif->role_id);
2590 		if (ret < 0)
2591 			goto out;
2592 
2593 		ret = wl1271_init_vif_specific(wl, vif);
2594 		if (ret < 0)
2595 			goto out;
2596 
2597 	} else {
2598 		ret = wl12xx_cmd_role_enable(wl, vif->addr, WL1271_ROLE_DEVICE,
2599 					     &wlvif->dev_role_id);
2600 		if (ret < 0)
2601 			goto out;
2602 
2603 		/* needed mainly for configuring rate policies */
2604 		ret = wl1271_sta_hw_init(wl, wlvif);
2605 		if (ret < 0)
2606 			goto out;
2607 	}
2608 
2609 	list_add(&wlvif->list, &wl->wlvif_list);
2610 	set_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags);
2611 
2612 	if (wlvif->bss_type == BSS_TYPE_AP_BSS)
2613 		wl->ap_count++;
2614 	else
2615 		wl->sta_count++;
2616 out:
2617 	pm_runtime_mark_last_busy(wl->dev);
2618 	pm_runtime_put_autosuspend(wl->dev);
2619 out_unlock:
2620 	mutex_unlock(&wl->mutex);
2621 
2622 	return ret;
2623 }
2624 
2625 static void __wl1271_op_remove_interface(struct wl1271 *wl,
2626 					 struct ieee80211_vif *vif,
2627 					 bool reset_tx_queues)
2628 {
2629 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
2630 	int i, ret;
2631 	bool is_ap = (wlvif->bss_type == BSS_TYPE_AP_BSS);
2632 
2633 	wl1271_debug(DEBUG_MAC80211, "mac80211 remove interface");
2634 
2635 	if (!test_and_clear_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags))
2636 		return;
2637 
2638 	/* because of hardware recovery, we may get here twice */
2639 	if (wl->state == WLCORE_STATE_OFF)
2640 		return;
2641 
2642 	wl1271_info("down");
2643 
2644 	if (wl->scan.state != WL1271_SCAN_STATE_IDLE &&
2645 	    wl->scan_wlvif == wlvif) {
2646 		struct cfg80211_scan_info info = {
2647 			.aborted = true,
2648 		};
2649 
2650 		/*
2651 		 * Rearm the tx watchdog just before idling scan. This
2652 		 * prevents just-finished scans from triggering the watchdog
2653 		 */
2654 		wl12xx_rearm_tx_watchdog_locked(wl);
2655 
2656 		wl->scan.state = WL1271_SCAN_STATE_IDLE;
2657 		memset(wl->scan.scanned_ch, 0, sizeof(wl->scan.scanned_ch));
2658 		wl->scan_wlvif = NULL;
2659 		wl->scan.req = NULL;
2660 		ieee80211_scan_completed(wl->hw, &info);
2661 	}
2662 
2663 	if (wl->sched_vif == wlvif)
2664 		wl->sched_vif = NULL;
2665 
2666 	if (wl->roc_vif == vif) {
2667 		wl->roc_vif = NULL;
2668 		ieee80211_remain_on_channel_expired(wl->hw);
2669 	}
2670 
2671 	if (!test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags)) {
2672 		/* disable active roles */
2673 		ret = pm_runtime_resume_and_get(wl->dev);
2674 		if (ret < 0)
2675 			goto deinit;
2676 
2677 		if (wlvif->bss_type == BSS_TYPE_STA_BSS ||
2678 		    wlvif->bss_type == BSS_TYPE_IBSS) {
2679 			if (wl12xx_dev_role_started(wlvif))
2680 				wl12xx_stop_dev(wl, wlvif);
2681 		}
2682 
2683 		if (!wlcore_is_p2p_mgmt(wlvif)) {
2684 			ret = wl12xx_cmd_role_disable(wl, &wlvif->role_id);
2685 			if (ret < 0) {
2686 				pm_runtime_put_noidle(wl->dev);
2687 				goto deinit;
2688 			}
2689 		} else {
2690 			ret = wl12xx_cmd_role_disable(wl, &wlvif->dev_role_id);
2691 			if (ret < 0) {
2692 				pm_runtime_put_noidle(wl->dev);
2693 				goto deinit;
2694 			}
2695 		}
2696 
2697 		pm_runtime_mark_last_busy(wl->dev);
2698 		pm_runtime_put_autosuspend(wl->dev);
2699 	}
2700 deinit:
2701 	wl12xx_tx_reset_wlvif(wl, wlvif);
2702 
2703 	/* clear all hlids (except system_hlid) */
2704 	wlvif->dev_hlid = WL12XX_INVALID_LINK_ID;
2705 
2706 	if (wlvif->bss_type == BSS_TYPE_STA_BSS ||
2707 	    wlvif->bss_type == BSS_TYPE_IBSS) {
2708 		wlvif->sta.hlid = WL12XX_INVALID_LINK_ID;
2709 		wl12xx_free_rate_policy(wl, &wlvif->sta.basic_rate_idx);
2710 		wl12xx_free_rate_policy(wl, &wlvif->sta.ap_rate_idx);
2711 		wl12xx_free_rate_policy(wl, &wlvif->sta.p2p_rate_idx);
2712 		wlcore_free_klv_template(wl, &wlvif->sta.klv_template_id);
2713 	} else {
2714 		wlvif->ap.bcast_hlid = WL12XX_INVALID_LINK_ID;
2715 		wlvif->ap.global_hlid = WL12XX_INVALID_LINK_ID;
2716 		wl12xx_free_rate_policy(wl, &wlvif->ap.mgmt_rate_idx);
2717 		wl12xx_free_rate_policy(wl, &wlvif->ap.bcast_rate_idx);
2718 		for (i = 0; i < CONF_TX_MAX_AC_COUNT; i++)
2719 			wl12xx_free_rate_policy(wl,
2720 						&wlvif->ap.ucast_rate_idx[i]);
2721 		wl1271_free_ap_keys(wl, wlvif);
2722 	}
2723 
2724 	dev_kfree_skb(wlvif->probereq);
2725 	wlvif->probereq = NULL;
2726 	if (wl->last_wlvif == wlvif)
2727 		wl->last_wlvif = NULL;
2728 	list_del(&wlvif->list);
2729 	memset(wlvif->ap.sta_hlid_map, 0, sizeof(wlvif->ap.sta_hlid_map));
2730 	wlvif->role_id = WL12XX_INVALID_ROLE_ID;
2731 	wlvif->dev_role_id = WL12XX_INVALID_ROLE_ID;
2732 
2733 	if (is_ap)
2734 		wl->ap_count--;
2735 	else
2736 		wl->sta_count--;
2737 
2738 	/*
2739 	 * Last AP, have more stations. Configure sleep auth according to STA.
2740 	 * Don't do thin on unintended recovery.
2741 	 */
2742 	if (test_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags) &&
2743 	    !test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags))
2744 		goto unlock;
2745 
2746 	if (wl->ap_count == 0 && is_ap) {
2747 		/* mask ap events */
2748 		wl->event_mask &= ~wl->ap_event_mask;
2749 		wl1271_event_unmask(wl);
2750 	}
2751 
2752 	if (wl->ap_count == 0 && is_ap && wl->sta_count) {
2753 		u8 sta_auth = wl->conf.conn.sta_sleep_auth;
2754 		/* Configure for power according to debugfs */
2755 		if (sta_auth != WL1271_PSM_ILLEGAL)
2756 			wl1271_acx_sleep_auth(wl, sta_auth);
2757 		/* Configure for ELP power saving */
2758 		else
2759 			wl1271_acx_sleep_auth(wl, WL1271_PSM_ELP);
2760 	}
2761 
2762 unlock:
2763 	mutex_unlock(&wl->mutex);
2764 
2765 	del_timer_sync(&wlvif->rx_streaming_timer);
2766 	cancel_work_sync(&wlvif->rx_streaming_enable_work);
2767 	cancel_work_sync(&wlvif->rx_streaming_disable_work);
2768 	cancel_work_sync(&wlvif->rc_update_work);
2769 	cancel_delayed_work_sync(&wlvif->connection_loss_work);
2770 	cancel_delayed_work_sync(&wlvif->channel_switch_work);
2771 	cancel_delayed_work_sync(&wlvif->pending_auth_complete_work);
2772 
2773 	mutex_lock(&wl->mutex);
2774 }
2775 
2776 static void wl1271_op_remove_interface(struct ieee80211_hw *hw,
2777 				       struct ieee80211_vif *vif)
2778 {
2779 	struct wl1271 *wl = hw->priv;
2780 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
2781 	struct wl12xx_vif *iter;
2782 	struct vif_counter_data vif_count;
2783 
2784 	wl12xx_get_vif_count(hw, vif, &vif_count);
2785 	mutex_lock(&wl->mutex);
2786 
2787 	if (wl->state == WLCORE_STATE_OFF ||
2788 	    !test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags))
2789 		goto out;
2790 
2791 	/*
2792 	 * wl->vif can be null here if someone shuts down the interface
2793 	 * just when hardware recovery has been started.
2794 	 */
2795 	wl12xx_for_each_wlvif(wl, iter) {
2796 		if (iter != wlvif)
2797 			continue;
2798 
2799 		__wl1271_op_remove_interface(wl, vif, true);
2800 		break;
2801 	}
2802 	WARN_ON(iter != wlvif);
2803 	if (wl12xx_need_fw_change(wl, vif_count, false)) {
2804 		wl12xx_force_active_psm(wl);
2805 		set_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags);
2806 		wl12xx_queue_recovery_work(wl);
2807 	}
2808 out:
2809 	mutex_unlock(&wl->mutex);
2810 }
2811 
2812 static int wl12xx_op_change_interface(struct ieee80211_hw *hw,
2813 				      struct ieee80211_vif *vif,
2814 				      enum nl80211_iftype new_type, bool p2p)
2815 {
2816 	struct wl1271 *wl = hw->priv;
2817 	int ret;
2818 
2819 	set_bit(WL1271_FLAG_VIF_CHANGE_IN_PROGRESS, &wl->flags);
2820 	wl1271_op_remove_interface(hw, vif);
2821 
2822 	vif->type = new_type;
2823 	vif->p2p = p2p;
2824 	ret = wl1271_op_add_interface(hw, vif);
2825 
2826 	clear_bit(WL1271_FLAG_VIF_CHANGE_IN_PROGRESS, &wl->flags);
2827 	return ret;
2828 }
2829 
2830 static int wlcore_join(struct wl1271 *wl, struct wl12xx_vif *wlvif)
2831 {
2832 	int ret;
2833 	bool is_ibss = (wlvif->bss_type == BSS_TYPE_IBSS);
2834 
2835 	/*
2836 	 * One of the side effects of the JOIN command is that is clears
2837 	 * WPA/WPA2 keys from the chipset. Performing a JOIN while associated
2838 	 * to a WPA/WPA2 access point will therefore kill the data-path.
2839 	 * Currently the only valid scenario for JOIN during association
2840 	 * is on roaming, in which case we will also be given new keys.
2841 	 * Keep the below message for now, unless it starts bothering
2842 	 * users who really like to roam a lot :)
2843 	 */
2844 	if (test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
2845 		wl1271_info("JOIN while associated.");
2846 
2847 	/* clear encryption type */
2848 	wlvif->encryption_type = KEY_NONE;
2849 
2850 	if (is_ibss)
2851 		ret = wl12xx_cmd_role_start_ibss(wl, wlvif);
2852 	else
2853 		ret = wl12xx_cmd_role_start_sta(wl, wlvif);
2854 
2855 	return ret;
2856 }
2857 
2858 static int wl1271_ssid_set(struct wl12xx_vif *wlvif, struct sk_buff *skb,
2859 			    int offset)
2860 {
2861 	u8 ssid_len;
2862 	const u8 *ptr = cfg80211_find_ie(WLAN_EID_SSID, skb->data + offset,
2863 					 skb->len - offset);
2864 
2865 	if (!ptr) {
2866 		wl1271_error("No SSID in IEs!");
2867 		return -ENOENT;
2868 	}
2869 
2870 	ssid_len = ptr[1];
2871 	if (ssid_len > IEEE80211_MAX_SSID_LEN) {
2872 		wl1271_error("SSID is too long!");
2873 		return -EINVAL;
2874 	}
2875 
2876 	wlvif->ssid_len = ssid_len;
2877 	memcpy(wlvif->ssid, ptr+2, ssid_len);
2878 	return 0;
2879 }
2880 
2881 static int wlcore_set_ssid(struct wl1271 *wl, struct wl12xx_vif *wlvif)
2882 {
2883 	struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
2884 	struct sk_buff *skb;
2885 	int ieoffset;
2886 
2887 	/* we currently only support setting the ssid from the ap probe req */
2888 	if (wlvif->bss_type != BSS_TYPE_STA_BSS)
2889 		return -EINVAL;
2890 
2891 	skb = ieee80211_ap_probereq_get(wl->hw, vif);
2892 	if (!skb)
2893 		return -EINVAL;
2894 
2895 	ieoffset = offsetof(struct ieee80211_mgmt,
2896 			    u.probe_req.variable);
2897 	wl1271_ssid_set(wlvif, skb, ieoffset);
2898 	dev_kfree_skb(skb);
2899 
2900 	return 0;
2901 }
2902 
2903 static int wlcore_set_assoc(struct wl1271 *wl, struct wl12xx_vif *wlvif,
2904 			    struct ieee80211_bss_conf *bss_conf,
2905 			    u32 sta_rate_set)
2906 {
2907 	struct ieee80211_vif *vif = container_of(bss_conf, struct ieee80211_vif,
2908 						 bss_conf);
2909 	int ieoffset;
2910 	int ret;
2911 
2912 	wlvif->aid = vif->cfg.aid;
2913 	wlvif->channel_type = cfg80211_get_chandef_type(&bss_conf->chandef);
2914 	wlvif->beacon_int = bss_conf->beacon_int;
2915 	wlvif->wmm_enabled = bss_conf->qos;
2916 
2917 	set_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags);
2918 
2919 	/*
2920 	 * with wl1271, we don't need to update the
2921 	 * beacon_int and dtim_period, because the firmware
2922 	 * updates it by itself when the first beacon is
2923 	 * received after a join.
2924 	 */
2925 	ret = wl1271_cmd_build_ps_poll(wl, wlvif, wlvif->aid);
2926 	if (ret < 0)
2927 		return ret;
2928 
2929 	/*
2930 	 * Get a template for hardware connection maintenance
2931 	 */
2932 	dev_kfree_skb(wlvif->probereq);
2933 	wlvif->probereq = wl1271_cmd_build_ap_probe_req(wl,
2934 							wlvif,
2935 							NULL);
2936 	ieoffset = offsetof(struct ieee80211_mgmt,
2937 			    u.probe_req.variable);
2938 	wl1271_ssid_set(wlvif, wlvif->probereq, ieoffset);
2939 
2940 	/* enable the connection monitoring feature */
2941 	ret = wl1271_acx_conn_monit_params(wl, wlvif, true);
2942 	if (ret < 0)
2943 		return ret;
2944 
2945 	/*
2946 	 * The join command disable the keep-alive mode, shut down its process,
2947 	 * and also clear the template config, so we need to reset it all after
2948 	 * the join. The acx_aid starts the keep-alive process, and the order
2949 	 * of the commands below is relevant.
2950 	 */
2951 	ret = wl1271_acx_keep_alive_mode(wl, wlvif, true);
2952 	if (ret < 0)
2953 		return ret;
2954 
2955 	ret = wl1271_acx_aid(wl, wlvif, wlvif->aid);
2956 	if (ret < 0)
2957 		return ret;
2958 
2959 	ret = wl12xx_cmd_build_klv_null_data(wl, wlvif);
2960 	if (ret < 0)
2961 		return ret;
2962 
2963 	ret = wl1271_acx_keep_alive_config(wl, wlvif,
2964 					   wlvif->sta.klv_template_id,
2965 					   ACX_KEEP_ALIVE_TPL_VALID);
2966 	if (ret < 0)
2967 		return ret;
2968 
2969 	/*
2970 	 * The default fw psm configuration is AUTO, while mac80211 default
2971 	 * setting is off (ACTIVE), so sync the fw with the correct value.
2972 	 */
2973 	ret = wl1271_ps_set_mode(wl, wlvif, STATION_ACTIVE_MODE);
2974 	if (ret < 0)
2975 		return ret;
2976 
2977 	if (sta_rate_set) {
2978 		wlvif->rate_set =
2979 			wl1271_tx_enabled_rates_get(wl,
2980 						    sta_rate_set,
2981 						    wlvif->band);
2982 		ret = wl1271_acx_sta_rate_policies(wl, wlvif);
2983 		if (ret < 0)
2984 			return ret;
2985 	}
2986 
2987 	return ret;
2988 }
2989 
2990 static int wlcore_unset_assoc(struct wl1271 *wl, struct wl12xx_vif *wlvif)
2991 {
2992 	int ret;
2993 	bool sta = wlvif->bss_type == BSS_TYPE_STA_BSS;
2994 
2995 	/* make sure we are connected (sta) joined */
2996 	if (sta &&
2997 	    !test_and_clear_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
2998 		return false;
2999 
3000 	/* make sure we are joined (ibss) */
3001 	if (!sta &&
3002 	    test_and_clear_bit(WLVIF_FLAG_IBSS_JOINED, &wlvif->flags))
3003 		return false;
3004 
3005 	if (sta) {
3006 		/* use defaults when not associated */
3007 		wlvif->aid = 0;
3008 
3009 		/* free probe-request template */
3010 		dev_kfree_skb(wlvif->probereq);
3011 		wlvif->probereq = NULL;
3012 
3013 		/* disable connection monitor features */
3014 		ret = wl1271_acx_conn_monit_params(wl, wlvif, false);
3015 		if (ret < 0)
3016 			return ret;
3017 
3018 		/* Disable the keep-alive feature */
3019 		ret = wl1271_acx_keep_alive_mode(wl, wlvif, false);
3020 		if (ret < 0)
3021 			return ret;
3022 
3023 		/* disable beacon filtering */
3024 		ret = wl1271_acx_beacon_filter_opt(wl, wlvif, false);
3025 		if (ret < 0)
3026 			return ret;
3027 	}
3028 
3029 	if (test_and_clear_bit(WLVIF_FLAG_CS_PROGRESS, &wlvif->flags)) {
3030 		struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
3031 
3032 		wl12xx_cmd_stop_channel_switch(wl, wlvif);
3033 		ieee80211_chswitch_done(vif, false);
3034 		cancel_delayed_work(&wlvif->channel_switch_work);
3035 	}
3036 
3037 	/* invalidate keep-alive template */
3038 	wl1271_acx_keep_alive_config(wl, wlvif,
3039 				     wlvif->sta.klv_template_id,
3040 				     ACX_KEEP_ALIVE_TPL_INVALID);
3041 
3042 	return 0;
3043 }
3044 
3045 static void wl1271_set_band_rate(struct wl1271 *wl, struct wl12xx_vif *wlvif)
3046 {
3047 	wlvif->basic_rate_set = wlvif->bitrate_masks[wlvif->band];
3048 	wlvif->rate_set = wlvif->basic_rate_set;
3049 }
3050 
3051 static void wl1271_sta_handle_idle(struct wl1271 *wl, struct wl12xx_vif *wlvif,
3052 				   bool idle)
3053 {
3054 	bool cur_idle = !test_bit(WLVIF_FLAG_ACTIVE, &wlvif->flags);
3055 
3056 	if (idle == cur_idle)
3057 		return;
3058 
3059 	if (idle) {
3060 		clear_bit(WLVIF_FLAG_ACTIVE, &wlvif->flags);
3061 	} else {
3062 		/* The current firmware only supports sched_scan in idle */
3063 		if (wl->sched_vif == wlvif)
3064 			wl->ops->sched_scan_stop(wl, wlvif);
3065 
3066 		set_bit(WLVIF_FLAG_ACTIVE, &wlvif->flags);
3067 	}
3068 }
3069 
3070 static int wl12xx_config_vif(struct wl1271 *wl, struct wl12xx_vif *wlvif,
3071 			     struct ieee80211_conf *conf, u32 changed)
3072 {
3073 	int ret;
3074 
3075 	if (wlcore_is_p2p_mgmt(wlvif))
3076 		return 0;
3077 
3078 	if (conf->power_level != wlvif->power_level) {
3079 		ret = wl1271_acx_tx_power(wl, wlvif, conf->power_level);
3080 		if (ret < 0)
3081 			return ret;
3082 
3083 		wlvif->power_level = conf->power_level;
3084 	}
3085 
3086 	return 0;
3087 }
3088 
3089 static int wl1271_op_config(struct ieee80211_hw *hw, u32 changed)
3090 {
3091 	struct wl1271 *wl = hw->priv;
3092 	struct wl12xx_vif *wlvif;
3093 	struct ieee80211_conf *conf = &hw->conf;
3094 	int ret = 0;
3095 
3096 	wl1271_debug(DEBUG_MAC80211, "mac80211 config psm %s power %d %s"
3097 		     " changed 0x%x",
3098 		     conf->flags & IEEE80211_CONF_PS ? "on" : "off",
3099 		     conf->power_level,
3100 		     conf->flags & IEEE80211_CONF_IDLE ? "idle" : "in use",
3101 			 changed);
3102 
3103 	mutex_lock(&wl->mutex);
3104 
3105 	if (changed & IEEE80211_CONF_CHANGE_POWER)
3106 		wl->power_level = conf->power_level;
3107 
3108 	if (unlikely(wl->state != WLCORE_STATE_ON))
3109 		goto out;
3110 
3111 	ret = pm_runtime_resume_and_get(wl->dev);
3112 	if (ret < 0)
3113 		goto out;
3114 
3115 	/* configure each interface */
3116 	wl12xx_for_each_wlvif(wl, wlvif) {
3117 		ret = wl12xx_config_vif(wl, wlvif, conf, changed);
3118 		if (ret < 0)
3119 			goto out_sleep;
3120 	}
3121 
3122 out_sleep:
3123 	pm_runtime_mark_last_busy(wl->dev);
3124 	pm_runtime_put_autosuspend(wl->dev);
3125 
3126 out:
3127 	mutex_unlock(&wl->mutex);
3128 
3129 	return ret;
3130 }
3131 
3132 struct wl1271_filter_params {
3133 	bool enabled;
3134 	int mc_list_length;
3135 	u8 mc_list[ACX_MC_ADDRESS_GROUP_MAX][ETH_ALEN];
3136 };
3137 
3138 static u64 wl1271_op_prepare_multicast(struct ieee80211_hw *hw,
3139 				       struct netdev_hw_addr_list *mc_list)
3140 {
3141 	struct wl1271_filter_params *fp;
3142 	struct netdev_hw_addr *ha;
3143 
3144 	fp = kzalloc(sizeof(*fp), GFP_ATOMIC);
3145 	if (!fp) {
3146 		wl1271_error("Out of memory setting filters.");
3147 		return 0;
3148 	}
3149 
3150 	/* update multicast filtering parameters */
3151 	fp->mc_list_length = 0;
3152 	if (netdev_hw_addr_list_count(mc_list) > ACX_MC_ADDRESS_GROUP_MAX) {
3153 		fp->enabled = false;
3154 	} else {
3155 		fp->enabled = true;
3156 		netdev_hw_addr_list_for_each(ha, mc_list) {
3157 			memcpy(fp->mc_list[fp->mc_list_length],
3158 					ha->addr, ETH_ALEN);
3159 			fp->mc_list_length++;
3160 		}
3161 	}
3162 
3163 	return (u64)(unsigned long)fp;
3164 }
3165 
3166 #define WL1271_SUPPORTED_FILTERS (FIF_ALLMULTI | \
3167 				  FIF_FCSFAIL | \
3168 				  FIF_BCN_PRBRESP_PROMISC | \
3169 				  FIF_CONTROL | \
3170 				  FIF_OTHER_BSS)
3171 
3172 static void wl1271_op_configure_filter(struct ieee80211_hw *hw,
3173 				       unsigned int changed,
3174 				       unsigned int *total, u64 multicast)
3175 {
3176 	struct wl1271_filter_params *fp = (void *)(unsigned long)multicast;
3177 	struct wl1271 *wl = hw->priv;
3178 	struct wl12xx_vif *wlvif;
3179 
3180 	int ret;
3181 
3182 	wl1271_debug(DEBUG_MAC80211, "mac80211 configure filter changed %x"
3183 		     " total %x", changed, *total);
3184 
3185 	mutex_lock(&wl->mutex);
3186 
3187 	*total &= WL1271_SUPPORTED_FILTERS;
3188 	changed &= WL1271_SUPPORTED_FILTERS;
3189 
3190 	if (unlikely(wl->state != WLCORE_STATE_ON))
3191 		goto out;
3192 
3193 	ret = pm_runtime_resume_and_get(wl->dev);
3194 	if (ret < 0)
3195 		goto out;
3196 
3197 	wl12xx_for_each_wlvif(wl, wlvif) {
3198 		if (wlcore_is_p2p_mgmt(wlvif))
3199 			continue;
3200 
3201 		if (wlvif->bss_type != BSS_TYPE_AP_BSS) {
3202 			if (*total & FIF_ALLMULTI)
3203 				ret = wl1271_acx_group_address_tbl(wl, wlvif,
3204 								   false,
3205 								   NULL, 0);
3206 			else if (fp)
3207 				ret = wl1271_acx_group_address_tbl(wl, wlvif,
3208 							fp->enabled,
3209 							fp->mc_list,
3210 							fp->mc_list_length);
3211 			if (ret < 0)
3212 				goto out_sleep;
3213 		}
3214 
3215 		/*
3216 		 * If interface in AP mode and created with allmulticast then disable
3217 		 * the firmware filters so that all multicast packets are passed
3218 		 * This is mandatory for MDNS based discovery protocols
3219 		 */
3220 		if (wlvif->bss_type == BSS_TYPE_AP_BSS) {
3221 			if (*total & FIF_ALLMULTI) {
3222 				ret = wl1271_acx_group_address_tbl(wl, wlvif,
3223 							false,
3224 							NULL, 0);
3225 				if (ret < 0)
3226 					goto out_sleep;
3227 			}
3228 		}
3229 	}
3230 
3231 	/*
3232 	 * the fw doesn't provide an api to configure the filters. instead,
3233 	 * the filters configuration is based on the active roles / ROC
3234 	 * state.
3235 	 */
3236 
3237 out_sleep:
3238 	pm_runtime_mark_last_busy(wl->dev);
3239 	pm_runtime_put_autosuspend(wl->dev);
3240 
3241 out:
3242 	mutex_unlock(&wl->mutex);
3243 	kfree(fp);
3244 }
3245 
3246 static int wl1271_record_ap_key(struct wl1271 *wl, struct wl12xx_vif *wlvif,
3247 				u8 id, u8 key_type, u8 key_size,
3248 				const u8 *key, u8 hlid, u32 tx_seq_32,
3249 				u16 tx_seq_16, bool is_pairwise)
3250 {
3251 	struct wl1271_ap_key *ap_key;
3252 	int i;
3253 
3254 	wl1271_debug(DEBUG_CRYPT, "record ap key id %d", (int)id);
3255 
3256 	if (key_size > MAX_KEY_SIZE)
3257 		return -EINVAL;
3258 
3259 	/*
3260 	 * Find next free entry in ap_keys. Also check we are not replacing
3261 	 * an existing key.
3262 	 */
3263 	for (i = 0; i < MAX_NUM_KEYS; i++) {
3264 		if (wlvif->ap.recorded_keys[i] == NULL)
3265 			break;
3266 
3267 		if (wlvif->ap.recorded_keys[i]->id == id) {
3268 			wl1271_warning("trying to record key replacement");
3269 			return -EINVAL;
3270 		}
3271 	}
3272 
3273 	if (i == MAX_NUM_KEYS)
3274 		return -EBUSY;
3275 
3276 	ap_key = kzalloc(sizeof(*ap_key), GFP_KERNEL);
3277 	if (!ap_key)
3278 		return -ENOMEM;
3279 
3280 	ap_key->id = id;
3281 	ap_key->key_type = key_type;
3282 	ap_key->key_size = key_size;
3283 	memcpy(ap_key->key, key, key_size);
3284 	ap_key->hlid = hlid;
3285 	ap_key->tx_seq_32 = tx_seq_32;
3286 	ap_key->tx_seq_16 = tx_seq_16;
3287 	ap_key->is_pairwise = is_pairwise;
3288 
3289 	wlvif->ap.recorded_keys[i] = ap_key;
3290 	return 0;
3291 }
3292 
3293 static void wl1271_free_ap_keys(struct wl1271 *wl, struct wl12xx_vif *wlvif)
3294 {
3295 	int i;
3296 
3297 	for (i = 0; i < MAX_NUM_KEYS; i++) {
3298 		kfree(wlvif->ap.recorded_keys[i]);
3299 		wlvif->ap.recorded_keys[i] = NULL;
3300 	}
3301 }
3302 
3303 static int wl1271_ap_init_hwenc(struct wl1271 *wl, struct wl12xx_vif *wlvif)
3304 {
3305 	int i, ret = 0;
3306 	struct wl1271_ap_key *key;
3307 	bool wep_key_added = false;
3308 
3309 	for (i = 0; i < MAX_NUM_KEYS; i++) {
3310 		u8 hlid;
3311 		if (wlvif->ap.recorded_keys[i] == NULL)
3312 			break;
3313 
3314 		key = wlvif->ap.recorded_keys[i];
3315 		hlid = key->hlid;
3316 		if (hlid == WL12XX_INVALID_LINK_ID)
3317 			hlid = wlvif->ap.bcast_hlid;
3318 
3319 		ret = wl1271_cmd_set_ap_key(wl, wlvif, KEY_ADD_OR_REPLACE,
3320 					    key->id, key->key_type,
3321 					    key->key_size, key->key,
3322 					    hlid, key->tx_seq_32,
3323 					    key->tx_seq_16, key->is_pairwise);
3324 		if (ret < 0)
3325 			goto out;
3326 
3327 		if (key->key_type == KEY_WEP)
3328 			wep_key_added = true;
3329 	}
3330 
3331 	if (wep_key_added) {
3332 		ret = wl12xx_cmd_set_default_wep_key(wl, wlvif->default_key,
3333 						     wlvif->ap.bcast_hlid);
3334 		if (ret < 0)
3335 			goto out;
3336 	}
3337 
3338 out:
3339 	wl1271_free_ap_keys(wl, wlvif);
3340 	return ret;
3341 }
3342 
3343 static int wl1271_set_key(struct wl1271 *wl, struct wl12xx_vif *wlvif,
3344 		       u16 action, u8 id, u8 key_type,
3345 		       u8 key_size, const u8 *key, u32 tx_seq_32,
3346 		       u16 tx_seq_16, struct ieee80211_sta *sta,
3347 		       bool is_pairwise)
3348 {
3349 	int ret;
3350 	bool is_ap = (wlvif->bss_type == BSS_TYPE_AP_BSS);
3351 
3352 	if (is_ap) {
3353 		struct wl1271_station *wl_sta;
3354 		u8 hlid;
3355 
3356 		if (sta) {
3357 			wl_sta = (struct wl1271_station *)sta->drv_priv;
3358 			hlid = wl_sta->hlid;
3359 		} else {
3360 			hlid = wlvif->ap.bcast_hlid;
3361 		}
3362 
3363 		if (!test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags)) {
3364 			/*
3365 			 * We do not support removing keys after AP shutdown.
3366 			 * Pretend we do to make mac80211 happy.
3367 			 */
3368 			if (action != KEY_ADD_OR_REPLACE)
3369 				return 0;
3370 
3371 			ret = wl1271_record_ap_key(wl, wlvif, id,
3372 					     key_type, key_size,
3373 					     key, hlid, tx_seq_32,
3374 					     tx_seq_16, is_pairwise);
3375 		} else {
3376 			ret = wl1271_cmd_set_ap_key(wl, wlvif, action,
3377 					     id, key_type, key_size,
3378 					     key, hlid, tx_seq_32,
3379 					     tx_seq_16, is_pairwise);
3380 		}
3381 
3382 		if (ret < 0)
3383 			return ret;
3384 	} else {
3385 		const u8 *addr;
3386 		static const u8 bcast_addr[ETH_ALEN] = {
3387 			0xff, 0xff, 0xff, 0xff, 0xff, 0xff
3388 		};
3389 
3390 		addr = sta ? sta->addr : bcast_addr;
3391 
3392 		if (is_zero_ether_addr(addr)) {
3393 			/* We dont support TX only encryption */
3394 			return -EOPNOTSUPP;
3395 		}
3396 
3397 		/* The wl1271 does not allow to remove unicast keys - they
3398 		   will be cleared automatically on next CMD_JOIN. Ignore the
3399 		   request silently, as we dont want the mac80211 to emit
3400 		   an error message. */
3401 		if (action == KEY_REMOVE && !is_broadcast_ether_addr(addr))
3402 			return 0;
3403 
3404 		/* don't remove key if hlid was already deleted */
3405 		if (action == KEY_REMOVE &&
3406 		    wlvif->sta.hlid == WL12XX_INVALID_LINK_ID)
3407 			return 0;
3408 
3409 		ret = wl1271_cmd_set_sta_key(wl, wlvif, action,
3410 					     id, key_type, key_size,
3411 					     key, addr, tx_seq_32,
3412 					     tx_seq_16);
3413 		if (ret < 0)
3414 			return ret;
3415 
3416 	}
3417 
3418 	return 0;
3419 }
3420 
3421 static int wlcore_op_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
3422 			     struct ieee80211_vif *vif,
3423 			     struct ieee80211_sta *sta,
3424 			     struct ieee80211_key_conf *key_conf)
3425 {
3426 	struct wl1271 *wl = hw->priv;
3427 	int ret;
3428 	bool might_change_spare =
3429 		key_conf->cipher == WL1271_CIPHER_SUITE_GEM ||
3430 		key_conf->cipher == WLAN_CIPHER_SUITE_TKIP;
3431 
3432 	if (might_change_spare) {
3433 		/*
3434 		 * stop the queues and flush to ensure the next packets are
3435 		 * in sync with FW spare block accounting
3436 		 */
3437 		wlcore_stop_queues(wl, WLCORE_QUEUE_STOP_REASON_SPARE_BLK);
3438 		wl1271_tx_flush(wl);
3439 	}
3440 
3441 	mutex_lock(&wl->mutex);
3442 
3443 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
3444 		ret = -EAGAIN;
3445 		goto out_wake_queues;
3446 	}
3447 
3448 	ret = pm_runtime_resume_and_get(wl->dev);
3449 	if (ret < 0)
3450 		goto out_wake_queues;
3451 
3452 	ret = wlcore_hw_set_key(wl, cmd, vif, sta, key_conf);
3453 
3454 	pm_runtime_mark_last_busy(wl->dev);
3455 	pm_runtime_put_autosuspend(wl->dev);
3456 
3457 out_wake_queues:
3458 	if (might_change_spare)
3459 		wlcore_wake_queues(wl, WLCORE_QUEUE_STOP_REASON_SPARE_BLK);
3460 
3461 	mutex_unlock(&wl->mutex);
3462 
3463 	return ret;
3464 }
3465 
3466 int wlcore_set_key(struct wl1271 *wl, enum set_key_cmd cmd,
3467 		   struct ieee80211_vif *vif,
3468 		   struct ieee80211_sta *sta,
3469 		   struct ieee80211_key_conf *key_conf)
3470 {
3471 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3472 	int ret;
3473 	u32 tx_seq_32 = 0;
3474 	u16 tx_seq_16 = 0;
3475 	u8 key_type;
3476 	u8 hlid;
3477 	bool is_pairwise;
3478 
3479 	wl1271_debug(DEBUG_MAC80211, "mac80211 set key");
3480 
3481 	wl1271_debug(DEBUG_CRYPT, "CMD: 0x%x sta: %p", cmd, sta);
3482 	wl1271_debug(DEBUG_CRYPT, "Key: algo:0x%x, id:%d, len:%d flags 0x%x",
3483 		     key_conf->cipher, key_conf->keyidx,
3484 		     key_conf->keylen, key_conf->flags);
3485 	wl1271_dump(DEBUG_CRYPT, "KEY: ", key_conf->key, key_conf->keylen);
3486 
3487 	if (wlvif->bss_type == BSS_TYPE_AP_BSS)
3488 		if (sta) {
3489 			struct wl1271_station *wl_sta = (void *)sta->drv_priv;
3490 			hlid = wl_sta->hlid;
3491 		} else {
3492 			hlid = wlvif->ap.bcast_hlid;
3493 		}
3494 	else
3495 		hlid = wlvif->sta.hlid;
3496 
3497 	if (hlid != WL12XX_INVALID_LINK_ID) {
3498 		u64 tx_seq = wl->links[hlid].total_freed_pkts;
3499 		tx_seq_32 = WL1271_TX_SECURITY_HI32(tx_seq);
3500 		tx_seq_16 = WL1271_TX_SECURITY_LO16(tx_seq);
3501 	}
3502 
3503 	switch (key_conf->cipher) {
3504 	case WLAN_CIPHER_SUITE_WEP40:
3505 	case WLAN_CIPHER_SUITE_WEP104:
3506 		key_type = KEY_WEP;
3507 
3508 		key_conf->hw_key_idx = key_conf->keyidx;
3509 		break;
3510 	case WLAN_CIPHER_SUITE_TKIP:
3511 		key_type = KEY_TKIP;
3512 		key_conf->hw_key_idx = key_conf->keyidx;
3513 		break;
3514 	case WLAN_CIPHER_SUITE_CCMP:
3515 		key_type = KEY_AES;
3516 		key_conf->flags |= IEEE80211_KEY_FLAG_PUT_IV_SPACE;
3517 		break;
3518 	case WL1271_CIPHER_SUITE_GEM:
3519 		key_type = KEY_GEM;
3520 		break;
3521 	default:
3522 		wl1271_error("Unknown key algo 0x%x", key_conf->cipher);
3523 
3524 		return -EOPNOTSUPP;
3525 	}
3526 
3527 	is_pairwise = key_conf->flags & IEEE80211_KEY_FLAG_PAIRWISE;
3528 
3529 	switch (cmd) {
3530 	case SET_KEY:
3531 		ret = wl1271_set_key(wl, wlvif, KEY_ADD_OR_REPLACE,
3532 				 key_conf->keyidx, key_type,
3533 				 key_conf->keylen, key_conf->key,
3534 				 tx_seq_32, tx_seq_16, sta, is_pairwise);
3535 		if (ret < 0) {
3536 			wl1271_error("Could not add or replace key");
3537 			return ret;
3538 		}
3539 
3540 		/*
3541 		 * reconfiguring arp response if the unicast (or common)
3542 		 * encryption key type was changed
3543 		 */
3544 		if (wlvif->bss_type == BSS_TYPE_STA_BSS &&
3545 		    (sta || key_type == KEY_WEP) &&
3546 		    wlvif->encryption_type != key_type) {
3547 			wlvif->encryption_type = key_type;
3548 			ret = wl1271_cmd_build_arp_rsp(wl, wlvif);
3549 			if (ret < 0) {
3550 				wl1271_warning("build arp rsp failed: %d", ret);
3551 				return ret;
3552 			}
3553 		}
3554 		break;
3555 
3556 	case DISABLE_KEY:
3557 		ret = wl1271_set_key(wl, wlvif, KEY_REMOVE,
3558 				     key_conf->keyidx, key_type,
3559 				     key_conf->keylen, key_conf->key,
3560 				     0, 0, sta, is_pairwise);
3561 		if (ret < 0) {
3562 			wl1271_error("Could not remove key");
3563 			return ret;
3564 		}
3565 		break;
3566 
3567 	default:
3568 		wl1271_error("Unsupported key cmd 0x%x", cmd);
3569 		return -EOPNOTSUPP;
3570 	}
3571 
3572 	return ret;
3573 }
3574 EXPORT_SYMBOL_GPL(wlcore_set_key);
3575 
3576 static void wl1271_op_set_default_key_idx(struct ieee80211_hw *hw,
3577 					  struct ieee80211_vif *vif,
3578 					  int key_idx)
3579 {
3580 	struct wl1271 *wl = hw->priv;
3581 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3582 	int ret;
3583 
3584 	wl1271_debug(DEBUG_MAC80211, "mac80211 set default key idx %d",
3585 		     key_idx);
3586 
3587 	/* we don't handle unsetting of default key */
3588 	if (key_idx == -1)
3589 		return;
3590 
3591 	mutex_lock(&wl->mutex);
3592 
3593 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
3594 		ret = -EAGAIN;
3595 		goto out_unlock;
3596 	}
3597 
3598 	ret = pm_runtime_resume_and_get(wl->dev);
3599 	if (ret < 0)
3600 		goto out_unlock;
3601 
3602 	wlvif->default_key = key_idx;
3603 
3604 	/* the default WEP key needs to be configured at least once */
3605 	if (wlvif->encryption_type == KEY_WEP) {
3606 		ret = wl12xx_cmd_set_default_wep_key(wl,
3607 				key_idx,
3608 				wlvif->sta.hlid);
3609 		if (ret < 0)
3610 			goto out_sleep;
3611 	}
3612 
3613 out_sleep:
3614 	pm_runtime_mark_last_busy(wl->dev);
3615 	pm_runtime_put_autosuspend(wl->dev);
3616 
3617 out_unlock:
3618 	mutex_unlock(&wl->mutex);
3619 }
3620 
3621 void wlcore_regdomain_config(struct wl1271 *wl)
3622 {
3623 	int ret;
3624 
3625 	if (!(wl->quirks & WLCORE_QUIRK_REGDOMAIN_CONF))
3626 		return;
3627 
3628 	mutex_lock(&wl->mutex);
3629 
3630 	if (unlikely(wl->state != WLCORE_STATE_ON))
3631 		goto out;
3632 
3633 	ret = pm_runtime_resume_and_get(wl->dev);
3634 	if (ret < 0)
3635 		goto out;
3636 
3637 	ret = wlcore_cmd_regdomain_config_locked(wl);
3638 	if (ret < 0) {
3639 		wl12xx_queue_recovery_work(wl);
3640 		goto out;
3641 	}
3642 
3643 	pm_runtime_mark_last_busy(wl->dev);
3644 	pm_runtime_put_autosuspend(wl->dev);
3645 out:
3646 	mutex_unlock(&wl->mutex);
3647 }
3648 
3649 static int wl1271_op_hw_scan(struct ieee80211_hw *hw,
3650 			     struct ieee80211_vif *vif,
3651 			     struct ieee80211_scan_request *hw_req)
3652 {
3653 	struct cfg80211_scan_request *req = &hw_req->req;
3654 	struct wl1271 *wl = hw->priv;
3655 	int ret;
3656 	u8 *ssid = NULL;
3657 	size_t len = 0;
3658 
3659 	wl1271_debug(DEBUG_MAC80211, "mac80211 hw scan");
3660 
3661 	if (req->n_ssids) {
3662 		ssid = req->ssids[0].ssid;
3663 		len = req->ssids[0].ssid_len;
3664 	}
3665 
3666 	mutex_lock(&wl->mutex);
3667 
3668 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
3669 		/*
3670 		 * We cannot return -EBUSY here because cfg80211 will expect
3671 		 * a call to ieee80211_scan_completed if we do - in this case
3672 		 * there won't be any call.
3673 		 */
3674 		ret = -EAGAIN;
3675 		goto out;
3676 	}
3677 
3678 	ret = pm_runtime_resume_and_get(wl->dev);
3679 	if (ret < 0)
3680 		goto out;
3681 
3682 	/* fail if there is any role in ROC */
3683 	if (find_first_bit(wl->roc_map, WL12XX_MAX_ROLES) < WL12XX_MAX_ROLES) {
3684 		/* don't allow scanning right now */
3685 		ret = -EBUSY;
3686 		goto out_sleep;
3687 	}
3688 
3689 	ret = wlcore_scan(hw->priv, vif, ssid, len, req);
3690 out_sleep:
3691 	pm_runtime_mark_last_busy(wl->dev);
3692 	pm_runtime_put_autosuspend(wl->dev);
3693 out:
3694 	mutex_unlock(&wl->mutex);
3695 
3696 	return ret;
3697 }
3698 
3699 static void wl1271_op_cancel_hw_scan(struct ieee80211_hw *hw,
3700 				     struct ieee80211_vif *vif)
3701 {
3702 	struct wl1271 *wl = hw->priv;
3703 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3704 	struct cfg80211_scan_info info = {
3705 		.aborted = true,
3706 	};
3707 	int ret;
3708 
3709 	wl1271_debug(DEBUG_MAC80211, "mac80211 cancel hw scan");
3710 
3711 	mutex_lock(&wl->mutex);
3712 
3713 	if (unlikely(wl->state != WLCORE_STATE_ON))
3714 		goto out;
3715 
3716 	if (wl->scan.state == WL1271_SCAN_STATE_IDLE)
3717 		goto out;
3718 
3719 	ret = pm_runtime_resume_and_get(wl->dev);
3720 	if (ret < 0)
3721 		goto out;
3722 
3723 	if (wl->scan.state != WL1271_SCAN_STATE_DONE) {
3724 		ret = wl->ops->scan_stop(wl, wlvif);
3725 		if (ret < 0)
3726 			goto out_sleep;
3727 	}
3728 
3729 	/*
3730 	 * Rearm the tx watchdog just before idling scan. This
3731 	 * prevents just-finished scans from triggering the watchdog
3732 	 */
3733 	wl12xx_rearm_tx_watchdog_locked(wl);
3734 
3735 	wl->scan.state = WL1271_SCAN_STATE_IDLE;
3736 	memset(wl->scan.scanned_ch, 0, sizeof(wl->scan.scanned_ch));
3737 	wl->scan_wlvif = NULL;
3738 	wl->scan.req = NULL;
3739 	ieee80211_scan_completed(wl->hw, &info);
3740 
3741 out_sleep:
3742 	pm_runtime_mark_last_busy(wl->dev);
3743 	pm_runtime_put_autosuspend(wl->dev);
3744 out:
3745 	mutex_unlock(&wl->mutex);
3746 
3747 	cancel_delayed_work_sync(&wl->scan_complete_work);
3748 }
3749 
3750 static int wl1271_op_sched_scan_start(struct ieee80211_hw *hw,
3751 				      struct ieee80211_vif *vif,
3752 				      struct cfg80211_sched_scan_request *req,
3753 				      struct ieee80211_scan_ies *ies)
3754 {
3755 	struct wl1271 *wl = hw->priv;
3756 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3757 	int ret;
3758 
3759 	wl1271_debug(DEBUG_MAC80211, "wl1271_op_sched_scan_start");
3760 
3761 	mutex_lock(&wl->mutex);
3762 
3763 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
3764 		ret = -EAGAIN;
3765 		goto out;
3766 	}
3767 
3768 	ret = pm_runtime_resume_and_get(wl->dev);
3769 	if (ret < 0)
3770 		goto out;
3771 
3772 	ret = wl->ops->sched_scan_start(wl, wlvif, req, ies);
3773 	if (ret < 0)
3774 		goto out_sleep;
3775 
3776 	wl->sched_vif = wlvif;
3777 
3778 out_sleep:
3779 	pm_runtime_mark_last_busy(wl->dev);
3780 	pm_runtime_put_autosuspend(wl->dev);
3781 out:
3782 	mutex_unlock(&wl->mutex);
3783 	return ret;
3784 }
3785 
3786 static int wl1271_op_sched_scan_stop(struct ieee80211_hw *hw,
3787 				     struct ieee80211_vif *vif)
3788 {
3789 	struct wl1271 *wl = hw->priv;
3790 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3791 	int ret;
3792 
3793 	wl1271_debug(DEBUG_MAC80211, "wl1271_op_sched_scan_stop");
3794 
3795 	mutex_lock(&wl->mutex);
3796 
3797 	if (unlikely(wl->state != WLCORE_STATE_ON))
3798 		goto out;
3799 
3800 	ret = pm_runtime_resume_and_get(wl->dev);
3801 	if (ret < 0)
3802 		goto out;
3803 
3804 	wl->ops->sched_scan_stop(wl, wlvif);
3805 
3806 	pm_runtime_mark_last_busy(wl->dev);
3807 	pm_runtime_put_autosuspend(wl->dev);
3808 out:
3809 	mutex_unlock(&wl->mutex);
3810 
3811 	return 0;
3812 }
3813 
3814 static int wl1271_op_set_frag_threshold(struct ieee80211_hw *hw, u32 value)
3815 {
3816 	struct wl1271 *wl = hw->priv;
3817 	int ret = 0;
3818 
3819 	mutex_lock(&wl->mutex);
3820 
3821 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
3822 		ret = -EAGAIN;
3823 		goto out;
3824 	}
3825 
3826 	ret = pm_runtime_resume_and_get(wl->dev);
3827 	if (ret < 0)
3828 		goto out;
3829 
3830 	ret = wl1271_acx_frag_threshold(wl, value);
3831 	if (ret < 0)
3832 		wl1271_warning("wl1271_op_set_frag_threshold failed: %d", ret);
3833 
3834 	pm_runtime_mark_last_busy(wl->dev);
3835 	pm_runtime_put_autosuspend(wl->dev);
3836 
3837 out:
3838 	mutex_unlock(&wl->mutex);
3839 
3840 	return ret;
3841 }
3842 
3843 static int wl1271_op_set_rts_threshold(struct ieee80211_hw *hw, u32 value)
3844 {
3845 	struct wl1271 *wl = hw->priv;
3846 	struct wl12xx_vif *wlvif;
3847 	int ret = 0;
3848 
3849 	mutex_lock(&wl->mutex);
3850 
3851 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
3852 		ret = -EAGAIN;
3853 		goto out;
3854 	}
3855 
3856 	ret = pm_runtime_resume_and_get(wl->dev);
3857 	if (ret < 0)
3858 		goto out;
3859 
3860 	wl12xx_for_each_wlvif(wl, wlvif) {
3861 		ret = wl1271_acx_rts_threshold(wl, wlvif, value);
3862 		if (ret < 0)
3863 			wl1271_warning("set rts threshold failed: %d", ret);
3864 	}
3865 	pm_runtime_mark_last_busy(wl->dev);
3866 	pm_runtime_put_autosuspend(wl->dev);
3867 
3868 out:
3869 	mutex_unlock(&wl->mutex);
3870 
3871 	return ret;
3872 }
3873 
3874 static void wl12xx_remove_ie(struct sk_buff *skb, u8 eid, int ieoffset)
3875 {
3876 	int len;
3877 	const u8 *next, *end = skb->data + skb->len;
3878 	u8 *ie = (u8 *)cfg80211_find_ie(eid, skb->data + ieoffset,
3879 					skb->len - ieoffset);
3880 	if (!ie)
3881 		return;
3882 	len = ie[1] + 2;
3883 	next = ie + len;
3884 	memmove(ie, next, end - next);
3885 	skb_trim(skb, skb->len - len);
3886 }
3887 
3888 static void wl12xx_remove_vendor_ie(struct sk_buff *skb,
3889 					    unsigned int oui, u8 oui_type,
3890 					    int ieoffset)
3891 {
3892 	int len;
3893 	const u8 *next, *end = skb->data + skb->len;
3894 	u8 *ie = (u8 *)cfg80211_find_vendor_ie(oui, oui_type,
3895 					       skb->data + ieoffset,
3896 					       skb->len - ieoffset);
3897 	if (!ie)
3898 		return;
3899 	len = ie[1] + 2;
3900 	next = ie + len;
3901 	memmove(ie, next, end - next);
3902 	skb_trim(skb, skb->len - len);
3903 }
3904 
3905 static int wl1271_ap_set_probe_resp_tmpl(struct wl1271 *wl, u32 rates,
3906 					 struct ieee80211_vif *vif)
3907 {
3908 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3909 	struct sk_buff *skb;
3910 	int ret;
3911 
3912 	skb = ieee80211_proberesp_get(wl->hw, vif);
3913 	if (!skb)
3914 		return -EOPNOTSUPP;
3915 
3916 	ret = wl1271_cmd_template_set(wl, wlvif->role_id,
3917 				      CMD_TEMPL_AP_PROBE_RESPONSE,
3918 				      skb->data,
3919 				      skb->len, 0,
3920 				      rates);
3921 	dev_kfree_skb(skb);
3922 
3923 	if (ret < 0)
3924 		goto out;
3925 
3926 	wl1271_debug(DEBUG_AP, "probe response updated");
3927 	set_bit(WLVIF_FLAG_AP_PROBE_RESP_SET, &wlvif->flags);
3928 
3929 out:
3930 	return ret;
3931 }
3932 
3933 static int wl1271_ap_set_probe_resp_tmpl_legacy(struct wl1271 *wl,
3934 					     struct ieee80211_vif *vif,
3935 					     u8 *probe_rsp_data,
3936 					     size_t probe_rsp_len,
3937 					     u32 rates)
3938 {
3939 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3940 	u8 probe_rsp_templ[WL1271_CMD_TEMPL_MAX_SIZE];
3941 	int ssid_ie_offset, ie_offset, templ_len;
3942 	const u8 *ptr;
3943 
3944 	/* no need to change probe response if the SSID is set correctly */
3945 	if (wlvif->ssid_len > 0)
3946 		return wl1271_cmd_template_set(wl, wlvif->role_id,
3947 					       CMD_TEMPL_AP_PROBE_RESPONSE,
3948 					       probe_rsp_data,
3949 					       probe_rsp_len, 0,
3950 					       rates);
3951 
3952 	if (probe_rsp_len + vif->cfg.ssid_len > WL1271_CMD_TEMPL_MAX_SIZE) {
3953 		wl1271_error("probe_rsp template too big");
3954 		return -EINVAL;
3955 	}
3956 
3957 	/* start searching from IE offset */
3958 	ie_offset = offsetof(struct ieee80211_mgmt, u.probe_resp.variable);
3959 
3960 	ptr = cfg80211_find_ie(WLAN_EID_SSID, probe_rsp_data + ie_offset,
3961 			       probe_rsp_len - ie_offset);
3962 	if (!ptr) {
3963 		wl1271_error("No SSID in beacon!");
3964 		return -EINVAL;
3965 	}
3966 
3967 	ssid_ie_offset = ptr - probe_rsp_data;
3968 	ptr += (ptr[1] + 2);
3969 
3970 	memcpy(probe_rsp_templ, probe_rsp_data, ssid_ie_offset);
3971 
3972 	/* insert SSID from bss_conf */
3973 	probe_rsp_templ[ssid_ie_offset] = WLAN_EID_SSID;
3974 	probe_rsp_templ[ssid_ie_offset + 1] = vif->cfg.ssid_len;
3975 	memcpy(probe_rsp_templ + ssid_ie_offset + 2,
3976 	       vif->cfg.ssid, vif->cfg.ssid_len);
3977 	templ_len = ssid_ie_offset + 2 + vif->cfg.ssid_len;
3978 
3979 	memcpy(probe_rsp_templ + ssid_ie_offset + 2 + vif->cfg.ssid_len,
3980 	       ptr, probe_rsp_len - (ptr - probe_rsp_data));
3981 	templ_len += probe_rsp_len - (ptr - probe_rsp_data);
3982 
3983 	return wl1271_cmd_template_set(wl, wlvif->role_id,
3984 				       CMD_TEMPL_AP_PROBE_RESPONSE,
3985 				       probe_rsp_templ,
3986 				       templ_len, 0,
3987 				       rates);
3988 }
3989 
3990 static int wl1271_bss_erp_info_changed(struct wl1271 *wl,
3991 				       struct ieee80211_vif *vif,
3992 				       struct ieee80211_bss_conf *bss_conf,
3993 				       u32 changed)
3994 {
3995 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
3996 	int ret = 0;
3997 
3998 	if (changed & BSS_CHANGED_ERP_SLOT) {
3999 		if (bss_conf->use_short_slot)
4000 			ret = wl1271_acx_slot(wl, wlvif, SLOT_TIME_SHORT);
4001 		else
4002 			ret = wl1271_acx_slot(wl, wlvif, SLOT_TIME_LONG);
4003 		if (ret < 0) {
4004 			wl1271_warning("Set slot time failed %d", ret);
4005 			goto out;
4006 		}
4007 	}
4008 
4009 	if (changed & BSS_CHANGED_ERP_PREAMBLE) {
4010 		if (bss_conf->use_short_preamble)
4011 			wl1271_acx_set_preamble(wl, wlvif, ACX_PREAMBLE_SHORT);
4012 		else
4013 			wl1271_acx_set_preamble(wl, wlvif, ACX_PREAMBLE_LONG);
4014 	}
4015 
4016 	if (changed & BSS_CHANGED_ERP_CTS_PROT) {
4017 		if (bss_conf->use_cts_prot)
4018 			ret = wl1271_acx_cts_protect(wl, wlvif,
4019 						     CTSPROTECT_ENABLE);
4020 		else
4021 			ret = wl1271_acx_cts_protect(wl, wlvif,
4022 						     CTSPROTECT_DISABLE);
4023 		if (ret < 0) {
4024 			wl1271_warning("Set ctsprotect failed %d", ret);
4025 			goto out;
4026 		}
4027 	}
4028 
4029 out:
4030 	return ret;
4031 }
4032 
4033 static int wlcore_set_beacon_template(struct wl1271 *wl,
4034 				      struct ieee80211_vif *vif,
4035 				      bool is_ap)
4036 {
4037 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4038 	struct ieee80211_hdr *hdr;
4039 	u32 min_rate;
4040 	int ret;
4041 	int ieoffset = offsetof(struct ieee80211_mgmt, u.beacon.variable);
4042 	struct sk_buff *beacon = ieee80211_beacon_get(wl->hw, vif, 0);
4043 	u16 tmpl_id;
4044 
4045 	if (!beacon) {
4046 		ret = -EINVAL;
4047 		goto out;
4048 	}
4049 
4050 	wl1271_debug(DEBUG_MASTER, "beacon updated");
4051 
4052 	ret = wl1271_ssid_set(wlvif, beacon, ieoffset);
4053 	if (ret < 0) {
4054 		dev_kfree_skb(beacon);
4055 		goto out;
4056 	}
4057 	min_rate = wl1271_tx_min_rate_get(wl, wlvif->basic_rate_set);
4058 	tmpl_id = is_ap ? CMD_TEMPL_AP_BEACON :
4059 		CMD_TEMPL_BEACON;
4060 	ret = wl1271_cmd_template_set(wl, wlvif->role_id, tmpl_id,
4061 				      beacon->data,
4062 				      beacon->len, 0,
4063 				      min_rate);
4064 	if (ret < 0) {
4065 		dev_kfree_skb(beacon);
4066 		goto out;
4067 	}
4068 
4069 	wlvif->wmm_enabled =
4070 		cfg80211_find_vendor_ie(WLAN_OUI_MICROSOFT,
4071 					WLAN_OUI_TYPE_MICROSOFT_WMM,
4072 					beacon->data + ieoffset,
4073 					beacon->len - ieoffset);
4074 
4075 	/*
4076 	 * In case we already have a probe-resp beacon set explicitly
4077 	 * by usermode, don't use the beacon data.
4078 	 */
4079 	if (test_bit(WLVIF_FLAG_AP_PROBE_RESP_SET, &wlvif->flags))
4080 		goto end_bcn;
4081 
4082 	/* remove TIM ie from probe response */
4083 	wl12xx_remove_ie(beacon, WLAN_EID_TIM, ieoffset);
4084 
4085 	/*
4086 	 * remove p2p ie from probe response.
4087 	 * the fw reponds to probe requests that don't include
4088 	 * the p2p ie. probe requests with p2p ie will be passed,
4089 	 * and will be responded by the supplicant (the spec
4090 	 * forbids including the p2p ie when responding to probe
4091 	 * requests that didn't include it).
4092 	 */
4093 	wl12xx_remove_vendor_ie(beacon, WLAN_OUI_WFA,
4094 				WLAN_OUI_TYPE_WFA_P2P, ieoffset);
4095 
4096 	hdr = (struct ieee80211_hdr *) beacon->data;
4097 	hdr->frame_control = cpu_to_le16(IEEE80211_FTYPE_MGMT |
4098 					 IEEE80211_STYPE_PROBE_RESP);
4099 	if (is_ap)
4100 		ret = wl1271_ap_set_probe_resp_tmpl_legacy(wl, vif,
4101 							   beacon->data,
4102 							   beacon->len,
4103 							   min_rate);
4104 	else
4105 		ret = wl1271_cmd_template_set(wl, wlvif->role_id,
4106 					      CMD_TEMPL_PROBE_RESPONSE,
4107 					      beacon->data,
4108 					      beacon->len, 0,
4109 					      min_rate);
4110 end_bcn:
4111 	dev_kfree_skb(beacon);
4112 	if (ret < 0)
4113 		goto out;
4114 
4115 out:
4116 	return ret;
4117 }
4118 
4119 static int wl1271_bss_beacon_info_changed(struct wl1271 *wl,
4120 					  struct ieee80211_vif *vif,
4121 					  struct ieee80211_bss_conf *bss_conf,
4122 					  u32 changed)
4123 {
4124 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4125 	bool is_ap = (wlvif->bss_type == BSS_TYPE_AP_BSS);
4126 	int ret = 0;
4127 
4128 	if (changed & BSS_CHANGED_BEACON_INT) {
4129 		wl1271_debug(DEBUG_MASTER, "beacon interval updated: %d",
4130 			bss_conf->beacon_int);
4131 
4132 		wlvif->beacon_int = bss_conf->beacon_int;
4133 	}
4134 
4135 	if ((changed & BSS_CHANGED_AP_PROBE_RESP) && is_ap) {
4136 		u32 rate = wl1271_tx_min_rate_get(wl, wlvif->basic_rate_set);
4137 
4138 		wl1271_ap_set_probe_resp_tmpl(wl, rate, vif);
4139 	}
4140 
4141 	if (changed & BSS_CHANGED_BEACON) {
4142 		ret = wlcore_set_beacon_template(wl, vif, is_ap);
4143 		if (ret < 0)
4144 			goto out;
4145 
4146 		if (test_and_clear_bit(WLVIF_FLAG_BEACON_DISABLED,
4147 				       &wlvif->flags)) {
4148 			ret = wlcore_hw_dfs_master_restart(wl, wlvif);
4149 			if (ret < 0)
4150 				goto out;
4151 		}
4152 	}
4153 out:
4154 	if (ret != 0)
4155 		wl1271_error("beacon info change failed: %d", ret);
4156 	return ret;
4157 }
4158 
4159 /* AP mode changes */
4160 static void wl1271_bss_info_changed_ap(struct wl1271 *wl,
4161 				       struct ieee80211_vif *vif,
4162 				       struct ieee80211_bss_conf *bss_conf,
4163 				       u32 changed)
4164 {
4165 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4166 	int ret = 0;
4167 
4168 	if (changed & BSS_CHANGED_BASIC_RATES) {
4169 		u32 rates = bss_conf->basic_rates;
4170 
4171 		wlvif->basic_rate_set = wl1271_tx_enabled_rates_get(wl, rates,
4172 								 wlvif->band);
4173 		wlvif->basic_rate = wl1271_tx_min_rate_get(wl,
4174 							wlvif->basic_rate_set);
4175 
4176 		ret = wl1271_init_ap_rates(wl, wlvif);
4177 		if (ret < 0) {
4178 			wl1271_error("AP rate policy change failed %d", ret);
4179 			goto out;
4180 		}
4181 
4182 		ret = wl1271_ap_init_templates(wl, vif);
4183 		if (ret < 0)
4184 			goto out;
4185 
4186 		/* No need to set probe resp template for mesh */
4187 		if (!ieee80211_vif_is_mesh(vif)) {
4188 			ret = wl1271_ap_set_probe_resp_tmpl(wl,
4189 							    wlvif->basic_rate,
4190 							    vif);
4191 			if (ret < 0)
4192 				goto out;
4193 		}
4194 
4195 		ret = wlcore_set_beacon_template(wl, vif, true);
4196 		if (ret < 0)
4197 			goto out;
4198 	}
4199 
4200 	ret = wl1271_bss_beacon_info_changed(wl, vif, bss_conf, changed);
4201 	if (ret < 0)
4202 		goto out;
4203 
4204 	if (changed & BSS_CHANGED_BEACON_ENABLED) {
4205 		if (bss_conf->enable_beacon) {
4206 			if (!test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags)) {
4207 				ret = wl12xx_cmd_role_start_ap(wl, wlvif);
4208 				if (ret < 0)
4209 					goto out;
4210 
4211 				ret = wl1271_ap_init_hwenc(wl, wlvif);
4212 				if (ret < 0)
4213 					goto out;
4214 
4215 				set_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags);
4216 				wl1271_debug(DEBUG_AP, "started AP");
4217 			}
4218 		} else {
4219 			if (test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags)) {
4220 				/*
4221 				 * AP might be in ROC in case we have just
4222 				 * sent auth reply. handle it.
4223 				 */
4224 				if (test_bit(wlvif->role_id, wl->roc_map))
4225 					wl12xx_croc(wl, wlvif->role_id);
4226 
4227 				ret = wl12xx_cmd_role_stop_ap(wl, wlvif);
4228 				if (ret < 0)
4229 					goto out;
4230 
4231 				clear_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags);
4232 				clear_bit(WLVIF_FLAG_AP_PROBE_RESP_SET,
4233 					  &wlvif->flags);
4234 				wl1271_debug(DEBUG_AP, "stopped AP");
4235 			}
4236 		}
4237 	}
4238 
4239 	ret = wl1271_bss_erp_info_changed(wl, vif, bss_conf, changed);
4240 	if (ret < 0)
4241 		goto out;
4242 
4243 	/* Handle HT information change */
4244 	if ((changed & BSS_CHANGED_HT) &&
4245 	    (bss_conf->chandef.width != NL80211_CHAN_WIDTH_20_NOHT)) {
4246 		ret = wl1271_acx_set_ht_information(wl, wlvif,
4247 					bss_conf->ht_operation_mode);
4248 		if (ret < 0) {
4249 			wl1271_warning("Set ht information failed %d", ret);
4250 			goto out;
4251 		}
4252 	}
4253 
4254 out:
4255 	return;
4256 }
4257 
4258 static int wlcore_set_bssid(struct wl1271 *wl, struct wl12xx_vif *wlvif,
4259 			    struct ieee80211_vif *vif, u32 sta_rate_set)
4260 {
4261 	struct ieee80211_bss_conf *bss_conf = &vif->bss_conf;
4262 	u32 rates;
4263 	int ret;
4264 
4265 	wl1271_debug(DEBUG_MAC80211,
4266 	     "changed_bssid: %pM, aid: %d, bcn_int: %d, brates: 0x%x sta_rate_set: 0x%x",
4267 	     bss_conf->bssid, vif->cfg.aid,
4268 	     bss_conf->beacon_int,
4269 	     bss_conf->basic_rates, sta_rate_set);
4270 
4271 	wlvif->beacon_int = bss_conf->beacon_int;
4272 	rates = bss_conf->basic_rates;
4273 	wlvif->basic_rate_set =
4274 		wl1271_tx_enabled_rates_get(wl, rates,
4275 					    wlvif->band);
4276 	wlvif->basic_rate =
4277 		wl1271_tx_min_rate_get(wl,
4278 				       wlvif->basic_rate_set);
4279 
4280 	if (sta_rate_set)
4281 		wlvif->rate_set =
4282 			wl1271_tx_enabled_rates_get(wl,
4283 						sta_rate_set,
4284 						wlvif->band);
4285 
4286 	/* we only support sched_scan while not connected */
4287 	if (wl->sched_vif == wlvif)
4288 		wl->ops->sched_scan_stop(wl, wlvif);
4289 
4290 	ret = wl1271_acx_sta_rate_policies(wl, wlvif);
4291 	if (ret < 0)
4292 		return ret;
4293 
4294 	ret = wl12xx_cmd_build_null_data(wl, wlvif);
4295 	if (ret < 0)
4296 		return ret;
4297 
4298 	ret = wl1271_build_qos_null_data(wl, wl12xx_wlvif_to_vif(wlvif));
4299 	if (ret < 0)
4300 		return ret;
4301 
4302 	wlcore_set_ssid(wl, wlvif);
4303 
4304 	set_bit(WLVIF_FLAG_IN_USE, &wlvif->flags);
4305 
4306 	return 0;
4307 }
4308 
4309 static int wlcore_clear_bssid(struct wl1271 *wl, struct wl12xx_vif *wlvif)
4310 {
4311 	int ret;
4312 
4313 	/* revert back to minimum rates for the current band */
4314 	wl1271_set_band_rate(wl, wlvif);
4315 	wlvif->basic_rate = wl1271_tx_min_rate_get(wl, wlvif->basic_rate_set);
4316 
4317 	ret = wl1271_acx_sta_rate_policies(wl, wlvif);
4318 	if (ret < 0)
4319 		return ret;
4320 
4321 	if (wlvif->bss_type == BSS_TYPE_STA_BSS &&
4322 	    test_bit(WLVIF_FLAG_IN_USE, &wlvif->flags)) {
4323 		ret = wl12xx_cmd_role_stop_sta(wl, wlvif);
4324 		if (ret < 0)
4325 			return ret;
4326 	}
4327 
4328 	clear_bit(WLVIF_FLAG_IN_USE, &wlvif->flags);
4329 	return 0;
4330 }
4331 /* STA/IBSS mode changes */
4332 static void wl1271_bss_info_changed_sta(struct wl1271 *wl,
4333 					struct ieee80211_vif *vif,
4334 					struct ieee80211_bss_conf *bss_conf,
4335 					u32 changed)
4336 {
4337 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4338 	bool do_join = false;
4339 	bool is_ibss = (wlvif->bss_type == BSS_TYPE_IBSS);
4340 	bool ibss_joined = false;
4341 	u32 sta_rate_set = 0;
4342 	int ret;
4343 	struct ieee80211_sta *sta;
4344 	bool sta_exists = false;
4345 	struct ieee80211_sta_ht_cap sta_ht_cap;
4346 
4347 	if (is_ibss) {
4348 		ret = wl1271_bss_beacon_info_changed(wl, vif, bss_conf,
4349 						     changed);
4350 		if (ret < 0)
4351 			goto out;
4352 	}
4353 
4354 	if (changed & BSS_CHANGED_IBSS) {
4355 		if (vif->cfg.ibss_joined) {
4356 			set_bit(WLVIF_FLAG_IBSS_JOINED, &wlvif->flags);
4357 			ibss_joined = true;
4358 		} else {
4359 			wlcore_unset_assoc(wl, wlvif);
4360 			wl12xx_cmd_role_stop_sta(wl, wlvif);
4361 		}
4362 	}
4363 
4364 	if ((changed & BSS_CHANGED_BEACON_INT) && ibss_joined)
4365 		do_join = true;
4366 
4367 	/* Need to update the SSID (for filtering etc) */
4368 	if ((changed & BSS_CHANGED_BEACON) && ibss_joined)
4369 		do_join = true;
4370 
4371 	if ((changed & BSS_CHANGED_BEACON_ENABLED) && ibss_joined) {
4372 		wl1271_debug(DEBUG_ADHOC, "ad-hoc beaconing: %s",
4373 			     bss_conf->enable_beacon ? "enabled" : "disabled");
4374 
4375 		do_join = true;
4376 	}
4377 
4378 	if (changed & BSS_CHANGED_IDLE && !is_ibss)
4379 		wl1271_sta_handle_idle(wl, wlvif, vif->cfg.idle);
4380 
4381 	if (changed & BSS_CHANGED_CQM) {
4382 		bool enable = false;
4383 		if (bss_conf->cqm_rssi_thold)
4384 			enable = true;
4385 		ret = wl1271_acx_rssi_snr_trigger(wl, wlvif, enable,
4386 						  bss_conf->cqm_rssi_thold,
4387 						  bss_conf->cqm_rssi_hyst);
4388 		if (ret < 0)
4389 			goto out;
4390 		wlvif->rssi_thold = bss_conf->cqm_rssi_thold;
4391 	}
4392 
4393 	if (changed & (BSS_CHANGED_BSSID | BSS_CHANGED_HT |
4394 		       BSS_CHANGED_ASSOC)) {
4395 		rcu_read_lock();
4396 		sta = ieee80211_find_sta(vif, bss_conf->bssid);
4397 		if (sta) {
4398 			u8 *rx_mask = sta->deflink.ht_cap.mcs.rx_mask;
4399 
4400 			/* save the supp_rates of the ap */
4401 			sta_rate_set = sta->deflink.supp_rates[wlvif->band];
4402 			if (sta->deflink.ht_cap.ht_supported)
4403 				sta_rate_set |=
4404 					(rx_mask[0] << HW_HT_RATES_OFFSET) |
4405 					(rx_mask[1] << HW_MIMO_RATES_OFFSET);
4406 			sta_ht_cap = sta->deflink.ht_cap;
4407 			sta_exists = true;
4408 		}
4409 
4410 		rcu_read_unlock();
4411 	}
4412 
4413 	if (changed & BSS_CHANGED_BSSID) {
4414 		if (!is_zero_ether_addr(bss_conf->bssid)) {
4415 			ret = wlcore_set_bssid(wl, wlvif, vif,
4416 					       sta_rate_set);
4417 			if (ret < 0)
4418 				goto out;
4419 
4420 			/* Need to update the BSSID (for filtering etc) */
4421 			do_join = true;
4422 		} else {
4423 			ret = wlcore_clear_bssid(wl, wlvif);
4424 			if (ret < 0)
4425 				goto out;
4426 		}
4427 	}
4428 
4429 	if (changed & BSS_CHANGED_IBSS) {
4430 		wl1271_debug(DEBUG_ADHOC, "ibss_joined: %d",
4431 			     vif->cfg.ibss_joined);
4432 
4433 		if (vif->cfg.ibss_joined) {
4434 			u32 rates = bss_conf->basic_rates;
4435 			wlvif->basic_rate_set =
4436 				wl1271_tx_enabled_rates_get(wl, rates,
4437 							    wlvif->band);
4438 			wlvif->basic_rate =
4439 				wl1271_tx_min_rate_get(wl,
4440 						       wlvif->basic_rate_set);
4441 
4442 			/* by default, use 11b + OFDM rates */
4443 			wlvif->rate_set = CONF_TX_IBSS_DEFAULT_RATES;
4444 			ret = wl1271_acx_sta_rate_policies(wl, wlvif);
4445 			if (ret < 0)
4446 				goto out;
4447 		}
4448 	}
4449 
4450 	if ((changed & BSS_CHANGED_BEACON_INFO) && bss_conf->dtim_period) {
4451 		/* enable beacon filtering */
4452 		ret = wl1271_acx_beacon_filter_opt(wl, wlvif, true);
4453 		if (ret < 0)
4454 			goto out;
4455 	}
4456 
4457 	ret = wl1271_bss_erp_info_changed(wl, vif, bss_conf, changed);
4458 	if (ret < 0)
4459 		goto out;
4460 
4461 	if (do_join) {
4462 		ret = wlcore_join(wl, wlvif);
4463 		if (ret < 0) {
4464 			wl1271_warning("cmd join failed %d", ret);
4465 			goto out;
4466 		}
4467 	}
4468 
4469 	if (changed & BSS_CHANGED_ASSOC) {
4470 		if (vif->cfg.assoc) {
4471 			ret = wlcore_set_assoc(wl, wlvif, bss_conf,
4472 					       sta_rate_set);
4473 			if (ret < 0)
4474 				goto out;
4475 
4476 			if (test_bit(WLVIF_FLAG_STA_AUTHORIZED, &wlvif->flags))
4477 				wl12xx_set_authorized(wl, wlvif);
4478 		} else {
4479 			wlcore_unset_assoc(wl, wlvif);
4480 		}
4481 	}
4482 
4483 	if (changed & BSS_CHANGED_PS) {
4484 		if (vif->cfg.ps &&
4485 		    test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags) &&
4486 		    !test_bit(WLVIF_FLAG_IN_PS, &wlvif->flags)) {
4487 			int ps_mode;
4488 			char *ps_mode_str;
4489 
4490 			if (wl->conf.conn.forced_ps) {
4491 				ps_mode = STATION_POWER_SAVE_MODE;
4492 				ps_mode_str = "forced";
4493 			} else {
4494 				ps_mode = STATION_AUTO_PS_MODE;
4495 				ps_mode_str = "auto";
4496 			}
4497 
4498 			wl1271_debug(DEBUG_PSM, "%s ps enabled", ps_mode_str);
4499 
4500 			ret = wl1271_ps_set_mode(wl, wlvif, ps_mode);
4501 			if (ret < 0)
4502 				wl1271_warning("enter %s ps failed %d",
4503 					       ps_mode_str, ret);
4504 		} else if (!vif->cfg.ps &&
4505 			   test_bit(WLVIF_FLAG_IN_PS, &wlvif->flags)) {
4506 			wl1271_debug(DEBUG_PSM, "auto ps disabled");
4507 
4508 			ret = wl1271_ps_set_mode(wl, wlvif,
4509 						 STATION_ACTIVE_MODE);
4510 			if (ret < 0)
4511 				wl1271_warning("exit auto ps failed %d", ret);
4512 		}
4513 	}
4514 
4515 	/* Handle new association with HT. Do this after join. */
4516 	if (sta_exists) {
4517 		bool enabled =
4518 			bss_conf->chandef.width != NL80211_CHAN_WIDTH_20_NOHT;
4519 
4520 		ret = wlcore_hw_set_peer_cap(wl,
4521 					     &sta_ht_cap,
4522 					     enabled,
4523 					     wlvif->rate_set,
4524 					     wlvif->sta.hlid);
4525 		if (ret < 0) {
4526 			wl1271_warning("Set ht cap failed %d", ret);
4527 			goto out;
4528 
4529 		}
4530 
4531 		if (enabled) {
4532 			ret = wl1271_acx_set_ht_information(wl, wlvif,
4533 						bss_conf->ht_operation_mode);
4534 			if (ret < 0) {
4535 				wl1271_warning("Set ht information failed %d",
4536 					       ret);
4537 				goto out;
4538 			}
4539 		}
4540 	}
4541 
4542 	/* Handle arp filtering. Done after join. */
4543 	if ((changed & BSS_CHANGED_ARP_FILTER) ||
4544 	    (!is_ibss && (changed & BSS_CHANGED_QOS))) {
4545 		__be32 addr = vif->cfg.arp_addr_list[0];
4546 		wlvif->sta.qos = bss_conf->qos;
4547 		WARN_ON(wlvif->bss_type != BSS_TYPE_STA_BSS);
4548 
4549 		if (vif->cfg.arp_addr_cnt == 1 && vif->cfg.assoc) {
4550 			wlvif->ip_addr = addr;
4551 			/*
4552 			 * The template should have been configured only upon
4553 			 * association. however, it seems that the correct ip
4554 			 * isn't being set (when sending), so we have to
4555 			 * reconfigure the template upon every ip change.
4556 			 */
4557 			ret = wl1271_cmd_build_arp_rsp(wl, wlvif);
4558 			if (ret < 0) {
4559 				wl1271_warning("build arp rsp failed: %d", ret);
4560 				goto out;
4561 			}
4562 
4563 			ret = wl1271_acx_arp_ip_filter(wl, wlvif,
4564 				(ACX_ARP_FILTER_ARP_FILTERING |
4565 				 ACX_ARP_FILTER_AUTO_ARP),
4566 				addr);
4567 		} else {
4568 			wlvif->ip_addr = 0;
4569 			ret = wl1271_acx_arp_ip_filter(wl, wlvif, 0, addr);
4570 		}
4571 
4572 		if (ret < 0)
4573 			goto out;
4574 	}
4575 
4576 out:
4577 	return;
4578 }
4579 
4580 static void wl1271_op_bss_info_changed(struct ieee80211_hw *hw,
4581 				       struct ieee80211_vif *vif,
4582 				       struct ieee80211_bss_conf *bss_conf,
4583 				       u64 changed)
4584 {
4585 	struct wl1271 *wl = hw->priv;
4586 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4587 	bool is_ap = (wlvif->bss_type == BSS_TYPE_AP_BSS);
4588 	int ret;
4589 
4590 	wl1271_debug(DEBUG_MAC80211, "mac80211 bss info role %d changed 0x%x",
4591 		     wlvif->role_id, (int)changed);
4592 
4593 	/*
4594 	 * make sure to cancel pending disconnections if our association
4595 	 * state changed
4596 	 */
4597 	if (!is_ap && (changed & BSS_CHANGED_ASSOC))
4598 		cancel_delayed_work_sync(&wlvif->connection_loss_work);
4599 
4600 	if (is_ap && (changed & BSS_CHANGED_BEACON_ENABLED) &&
4601 	    !bss_conf->enable_beacon)
4602 		wl1271_tx_flush(wl);
4603 
4604 	mutex_lock(&wl->mutex);
4605 
4606 	if (unlikely(wl->state != WLCORE_STATE_ON))
4607 		goto out;
4608 
4609 	if (unlikely(!test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags)))
4610 		goto out;
4611 
4612 	ret = pm_runtime_resume_and_get(wl->dev);
4613 	if (ret < 0)
4614 		goto out;
4615 
4616 	if ((changed & BSS_CHANGED_TXPOWER) &&
4617 	    bss_conf->txpower != wlvif->power_level) {
4618 
4619 		ret = wl1271_acx_tx_power(wl, wlvif, bss_conf->txpower);
4620 		if (ret < 0)
4621 			goto out;
4622 
4623 		wlvif->power_level = bss_conf->txpower;
4624 	}
4625 
4626 	if (is_ap)
4627 		wl1271_bss_info_changed_ap(wl, vif, bss_conf, changed);
4628 	else
4629 		wl1271_bss_info_changed_sta(wl, vif, bss_conf, changed);
4630 
4631 	pm_runtime_mark_last_busy(wl->dev);
4632 	pm_runtime_put_autosuspend(wl->dev);
4633 
4634 out:
4635 	mutex_unlock(&wl->mutex);
4636 }
4637 
4638 static int wlcore_op_add_chanctx(struct ieee80211_hw *hw,
4639 				 struct ieee80211_chanctx_conf *ctx)
4640 {
4641 	wl1271_debug(DEBUG_MAC80211, "mac80211 add chanctx %d (type %d)",
4642 		     ieee80211_frequency_to_channel(ctx->def.chan->center_freq),
4643 		     cfg80211_get_chandef_type(&ctx->def));
4644 	return 0;
4645 }
4646 
4647 static void wlcore_op_remove_chanctx(struct ieee80211_hw *hw,
4648 				     struct ieee80211_chanctx_conf *ctx)
4649 {
4650 	wl1271_debug(DEBUG_MAC80211, "mac80211 remove chanctx %d (type %d)",
4651 		     ieee80211_frequency_to_channel(ctx->def.chan->center_freq),
4652 		     cfg80211_get_chandef_type(&ctx->def));
4653 }
4654 
4655 static void wlcore_op_change_chanctx(struct ieee80211_hw *hw,
4656 				     struct ieee80211_chanctx_conf *ctx,
4657 				     u32 changed)
4658 {
4659 	struct wl1271 *wl = hw->priv;
4660 	struct wl12xx_vif *wlvif;
4661 	int ret;
4662 	int channel = ieee80211_frequency_to_channel(
4663 		ctx->def.chan->center_freq);
4664 
4665 	wl1271_debug(DEBUG_MAC80211,
4666 		     "mac80211 change chanctx %d (type %d) changed 0x%x",
4667 		     channel, cfg80211_get_chandef_type(&ctx->def), changed);
4668 
4669 	mutex_lock(&wl->mutex);
4670 
4671 	ret = pm_runtime_resume_and_get(wl->dev);
4672 	if (ret < 0)
4673 		goto out;
4674 
4675 	wl12xx_for_each_wlvif(wl, wlvif) {
4676 		struct ieee80211_vif *vif = wl12xx_wlvif_to_vif(wlvif);
4677 
4678 		rcu_read_lock();
4679 		if (rcu_access_pointer(vif->bss_conf.chanctx_conf) != ctx) {
4680 			rcu_read_unlock();
4681 			continue;
4682 		}
4683 		rcu_read_unlock();
4684 
4685 		/* start radar if needed */
4686 		if (changed & IEEE80211_CHANCTX_CHANGE_RADAR &&
4687 		    wlvif->bss_type == BSS_TYPE_AP_BSS &&
4688 		    ctx->radar_enabled && !wlvif->radar_enabled &&
4689 		    ctx->def.chan->dfs_state == NL80211_DFS_USABLE) {
4690 			wl1271_debug(DEBUG_MAC80211, "Start radar detection");
4691 			wlcore_hw_set_cac(wl, wlvif, true);
4692 			wlvif->radar_enabled = true;
4693 		}
4694 	}
4695 
4696 	pm_runtime_mark_last_busy(wl->dev);
4697 	pm_runtime_put_autosuspend(wl->dev);
4698 out:
4699 	mutex_unlock(&wl->mutex);
4700 }
4701 
4702 static int wlcore_op_assign_vif_chanctx(struct ieee80211_hw *hw,
4703 					struct ieee80211_vif *vif,
4704 					struct ieee80211_bss_conf *link_conf,
4705 					struct ieee80211_chanctx_conf *ctx)
4706 {
4707 	struct wl1271 *wl = hw->priv;
4708 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4709 	int channel = ieee80211_frequency_to_channel(
4710 		ctx->def.chan->center_freq);
4711 	int ret = -EINVAL;
4712 
4713 	wl1271_debug(DEBUG_MAC80211,
4714 		     "mac80211 assign chanctx (role %d) %d (type %d) (radar %d dfs_state %d)",
4715 		     wlvif->role_id, channel,
4716 		     cfg80211_get_chandef_type(&ctx->def),
4717 		     ctx->radar_enabled, ctx->def.chan->dfs_state);
4718 
4719 	mutex_lock(&wl->mutex);
4720 
4721 	if (unlikely(wl->state != WLCORE_STATE_ON))
4722 		goto out;
4723 
4724 	if (unlikely(!test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags)))
4725 		goto out;
4726 
4727 	ret = pm_runtime_resume_and_get(wl->dev);
4728 	if (ret < 0)
4729 		goto out;
4730 
4731 	wlvif->band = ctx->def.chan->band;
4732 	wlvif->channel = channel;
4733 	wlvif->channel_type = cfg80211_get_chandef_type(&ctx->def);
4734 
4735 	/* update default rates according to the band */
4736 	wl1271_set_band_rate(wl, wlvif);
4737 
4738 	if (ctx->radar_enabled &&
4739 	    ctx->def.chan->dfs_state == NL80211_DFS_USABLE) {
4740 		wl1271_debug(DEBUG_MAC80211, "Start radar detection");
4741 		wlcore_hw_set_cac(wl, wlvif, true);
4742 		wlvif->radar_enabled = true;
4743 	}
4744 
4745 	pm_runtime_mark_last_busy(wl->dev);
4746 	pm_runtime_put_autosuspend(wl->dev);
4747 out:
4748 	mutex_unlock(&wl->mutex);
4749 
4750 	return 0;
4751 }
4752 
4753 static void wlcore_op_unassign_vif_chanctx(struct ieee80211_hw *hw,
4754 					   struct ieee80211_vif *vif,
4755 					   struct ieee80211_bss_conf *link_conf,
4756 					   struct ieee80211_chanctx_conf *ctx)
4757 {
4758 	struct wl1271 *wl = hw->priv;
4759 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4760 	int ret;
4761 
4762 	wl1271_debug(DEBUG_MAC80211,
4763 		     "mac80211 unassign chanctx (role %d) %d (type %d)",
4764 		     wlvif->role_id,
4765 		     ieee80211_frequency_to_channel(ctx->def.chan->center_freq),
4766 		     cfg80211_get_chandef_type(&ctx->def));
4767 
4768 	wl1271_tx_flush(wl);
4769 
4770 	mutex_lock(&wl->mutex);
4771 
4772 	if (unlikely(wl->state != WLCORE_STATE_ON))
4773 		goto out;
4774 
4775 	if (unlikely(!test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags)))
4776 		goto out;
4777 
4778 	ret = pm_runtime_resume_and_get(wl->dev);
4779 	if (ret < 0)
4780 		goto out;
4781 
4782 	if (wlvif->radar_enabled) {
4783 		wl1271_debug(DEBUG_MAC80211, "Stop radar detection");
4784 		wlcore_hw_set_cac(wl, wlvif, false);
4785 		wlvif->radar_enabled = false;
4786 	}
4787 
4788 	pm_runtime_mark_last_busy(wl->dev);
4789 	pm_runtime_put_autosuspend(wl->dev);
4790 out:
4791 	mutex_unlock(&wl->mutex);
4792 }
4793 
4794 static int __wlcore_switch_vif_chan(struct wl1271 *wl,
4795 				    struct wl12xx_vif *wlvif,
4796 				    struct ieee80211_chanctx_conf *new_ctx)
4797 {
4798 	int channel = ieee80211_frequency_to_channel(
4799 		new_ctx->def.chan->center_freq);
4800 
4801 	wl1271_debug(DEBUG_MAC80211,
4802 		     "switch vif (role %d) %d -> %d chan_type: %d",
4803 		     wlvif->role_id, wlvif->channel, channel,
4804 		     cfg80211_get_chandef_type(&new_ctx->def));
4805 
4806 	if (WARN_ON_ONCE(wlvif->bss_type != BSS_TYPE_AP_BSS))
4807 		return 0;
4808 
4809 	WARN_ON(!test_bit(WLVIF_FLAG_BEACON_DISABLED, &wlvif->flags));
4810 
4811 	if (wlvif->radar_enabled) {
4812 		wl1271_debug(DEBUG_MAC80211, "Stop radar detection");
4813 		wlcore_hw_set_cac(wl, wlvif, false);
4814 		wlvif->radar_enabled = false;
4815 	}
4816 
4817 	wlvif->band = new_ctx->def.chan->band;
4818 	wlvif->channel = channel;
4819 	wlvif->channel_type = cfg80211_get_chandef_type(&new_ctx->def);
4820 
4821 	/* start radar if needed */
4822 	if (new_ctx->radar_enabled) {
4823 		wl1271_debug(DEBUG_MAC80211, "Start radar detection");
4824 		wlcore_hw_set_cac(wl, wlvif, true);
4825 		wlvif->radar_enabled = true;
4826 	}
4827 
4828 	return 0;
4829 }
4830 
4831 static int
4832 wlcore_op_switch_vif_chanctx(struct ieee80211_hw *hw,
4833 			     struct ieee80211_vif_chanctx_switch *vifs,
4834 			     int n_vifs,
4835 			     enum ieee80211_chanctx_switch_mode mode)
4836 {
4837 	struct wl1271 *wl = hw->priv;
4838 	int i, ret;
4839 
4840 	wl1271_debug(DEBUG_MAC80211,
4841 		     "mac80211 switch chanctx n_vifs %d mode %d",
4842 		     n_vifs, mode);
4843 
4844 	mutex_lock(&wl->mutex);
4845 
4846 	ret = pm_runtime_resume_and_get(wl->dev);
4847 	if (ret < 0)
4848 		goto out;
4849 
4850 	for (i = 0; i < n_vifs; i++) {
4851 		struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vifs[i].vif);
4852 
4853 		ret = __wlcore_switch_vif_chan(wl, wlvif, vifs[i].new_ctx);
4854 		if (ret)
4855 			goto out_sleep;
4856 	}
4857 out_sleep:
4858 	pm_runtime_mark_last_busy(wl->dev);
4859 	pm_runtime_put_autosuspend(wl->dev);
4860 out:
4861 	mutex_unlock(&wl->mutex);
4862 
4863 	return 0;
4864 }
4865 
4866 static int wl1271_op_conf_tx(struct ieee80211_hw *hw,
4867 			     struct ieee80211_vif *vif,
4868 			     unsigned int link_id, u16 queue,
4869 			     const struct ieee80211_tx_queue_params *params)
4870 {
4871 	struct wl1271 *wl = hw->priv;
4872 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4873 	u8 ps_scheme;
4874 	int ret = 0;
4875 
4876 	if (wlcore_is_p2p_mgmt(wlvif))
4877 		return 0;
4878 
4879 	mutex_lock(&wl->mutex);
4880 
4881 	wl1271_debug(DEBUG_MAC80211, "mac80211 conf tx %d", queue);
4882 
4883 	if (params->uapsd)
4884 		ps_scheme = CONF_PS_SCHEME_UPSD_TRIGGER;
4885 	else
4886 		ps_scheme = CONF_PS_SCHEME_LEGACY;
4887 
4888 	if (!test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags))
4889 		goto out;
4890 
4891 	ret = pm_runtime_resume_and_get(wl->dev);
4892 	if (ret < 0)
4893 		goto out;
4894 
4895 	/*
4896 	 * the txop is confed in units of 32us by the mac80211,
4897 	 * we need us
4898 	 */
4899 	ret = wl1271_acx_ac_cfg(wl, wlvif, wl1271_tx_get_queue(queue),
4900 				params->cw_min, params->cw_max,
4901 				params->aifs, params->txop << 5);
4902 	if (ret < 0)
4903 		goto out_sleep;
4904 
4905 	ret = wl1271_acx_tid_cfg(wl, wlvif, wl1271_tx_get_queue(queue),
4906 				 CONF_CHANNEL_TYPE_EDCF,
4907 				 wl1271_tx_get_queue(queue),
4908 				 ps_scheme, CONF_ACK_POLICY_LEGACY,
4909 				 0, 0);
4910 
4911 out_sleep:
4912 	pm_runtime_mark_last_busy(wl->dev);
4913 	pm_runtime_put_autosuspend(wl->dev);
4914 
4915 out:
4916 	mutex_unlock(&wl->mutex);
4917 
4918 	return ret;
4919 }
4920 
4921 static u64 wl1271_op_get_tsf(struct ieee80211_hw *hw,
4922 			     struct ieee80211_vif *vif)
4923 {
4924 
4925 	struct wl1271 *wl = hw->priv;
4926 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
4927 	u64 mactime = ULLONG_MAX;
4928 	int ret;
4929 
4930 	wl1271_debug(DEBUG_MAC80211, "mac80211 get tsf");
4931 
4932 	mutex_lock(&wl->mutex);
4933 
4934 	if (unlikely(wl->state != WLCORE_STATE_ON))
4935 		goto out;
4936 
4937 	ret = pm_runtime_resume_and_get(wl->dev);
4938 	if (ret < 0)
4939 		goto out;
4940 
4941 	ret = wl12xx_acx_tsf_info(wl, wlvif, &mactime);
4942 	if (ret < 0)
4943 		goto out_sleep;
4944 
4945 out_sleep:
4946 	pm_runtime_mark_last_busy(wl->dev);
4947 	pm_runtime_put_autosuspend(wl->dev);
4948 
4949 out:
4950 	mutex_unlock(&wl->mutex);
4951 	return mactime;
4952 }
4953 
4954 static int wl1271_op_get_survey(struct ieee80211_hw *hw, int idx,
4955 				struct survey_info *survey)
4956 {
4957 	struct ieee80211_conf *conf = &hw->conf;
4958 
4959 	if (idx != 0)
4960 		return -ENOENT;
4961 
4962 	survey->channel = conf->chandef.chan;
4963 	survey->filled = 0;
4964 	return 0;
4965 }
4966 
4967 static int wl1271_allocate_sta(struct wl1271 *wl,
4968 			     struct wl12xx_vif *wlvif,
4969 			     struct ieee80211_sta *sta)
4970 {
4971 	struct wl1271_station *wl_sta;
4972 	int ret;
4973 
4974 
4975 	if (wl->active_sta_count >= wl->max_ap_stations) {
4976 		wl1271_warning("could not allocate HLID - too much stations");
4977 		return -EBUSY;
4978 	}
4979 
4980 	wl_sta = (struct wl1271_station *)sta->drv_priv;
4981 	ret = wl12xx_allocate_link(wl, wlvif, &wl_sta->hlid);
4982 	if (ret < 0) {
4983 		wl1271_warning("could not allocate HLID - too many links");
4984 		return -EBUSY;
4985 	}
4986 
4987 	/* use the previous security seq, if this is a recovery/resume */
4988 	wl->links[wl_sta->hlid].total_freed_pkts = wl_sta->total_freed_pkts;
4989 
4990 	set_bit(wl_sta->hlid, wlvif->ap.sta_hlid_map);
4991 	memcpy(wl->links[wl_sta->hlid].addr, sta->addr, ETH_ALEN);
4992 	wl->active_sta_count++;
4993 	return 0;
4994 }
4995 
4996 void wl1271_free_sta(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 hlid)
4997 {
4998 	if (!test_bit(hlid, wlvif->ap.sta_hlid_map))
4999 		return;
5000 
5001 	clear_bit(hlid, wlvif->ap.sta_hlid_map);
5002 	__clear_bit(hlid, &wl->ap_ps_map);
5003 	__clear_bit(hlid, &wl->ap_fw_ps_map);
5004 
5005 	/*
5006 	 * save the last used PN in the private part of iee80211_sta,
5007 	 * in case of recovery/suspend
5008 	 */
5009 	wlcore_save_freed_pkts_addr(wl, wlvif, hlid, wl->links[hlid].addr);
5010 
5011 	wl12xx_free_link(wl, wlvif, &hlid);
5012 	wl->active_sta_count--;
5013 
5014 	/*
5015 	 * rearm the tx watchdog when the last STA is freed - give the FW a
5016 	 * chance to return STA-buffered packets before complaining.
5017 	 */
5018 	if (wl->active_sta_count == 0)
5019 		wl12xx_rearm_tx_watchdog_locked(wl);
5020 }
5021 
5022 static int wl12xx_sta_add(struct wl1271 *wl,
5023 			  struct wl12xx_vif *wlvif,
5024 			  struct ieee80211_sta *sta)
5025 {
5026 	struct wl1271_station *wl_sta;
5027 	int ret = 0;
5028 	u8 hlid;
5029 
5030 	wl1271_debug(DEBUG_MAC80211, "mac80211 add sta %d", (int)sta->aid);
5031 
5032 	ret = wl1271_allocate_sta(wl, wlvif, sta);
5033 	if (ret < 0)
5034 		return ret;
5035 
5036 	wl_sta = (struct wl1271_station *)sta->drv_priv;
5037 	hlid = wl_sta->hlid;
5038 
5039 	ret = wl12xx_cmd_add_peer(wl, wlvif, sta, hlid);
5040 	if (ret < 0)
5041 		wl1271_free_sta(wl, wlvif, hlid);
5042 
5043 	return ret;
5044 }
5045 
5046 static int wl12xx_sta_remove(struct wl1271 *wl,
5047 			     struct wl12xx_vif *wlvif,
5048 			     struct ieee80211_sta *sta)
5049 {
5050 	struct wl1271_station *wl_sta;
5051 	int ret = 0, id;
5052 
5053 	wl1271_debug(DEBUG_MAC80211, "mac80211 remove sta %d", (int)sta->aid);
5054 
5055 	wl_sta = (struct wl1271_station *)sta->drv_priv;
5056 	id = wl_sta->hlid;
5057 	if (WARN_ON(!test_bit(id, wlvif->ap.sta_hlid_map)))
5058 		return -EINVAL;
5059 
5060 	ret = wl12xx_cmd_remove_peer(wl, wlvif, wl_sta->hlid);
5061 	if (ret < 0)
5062 		return ret;
5063 
5064 	wl1271_free_sta(wl, wlvif, wl_sta->hlid);
5065 	return ret;
5066 }
5067 
5068 static void wlcore_roc_if_possible(struct wl1271 *wl,
5069 				   struct wl12xx_vif *wlvif)
5070 {
5071 	if (find_first_bit(wl->roc_map,
5072 			   WL12XX_MAX_ROLES) < WL12XX_MAX_ROLES)
5073 		return;
5074 
5075 	if (WARN_ON(wlvif->role_id == WL12XX_INVALID_ROLE_ID))
5076 		return;
5077 
5078 	wl12xx_roc(wl, wlvif, wlvif->role_id, wlvif->band, wlvif->channel);
5079 }
5080 
5081 /*
5082  * when wl_sta is NULL, we treat this call as if coming from a
5083  * pending auth reply.
5084  * wl->mutex must be taken and the FW must be awake when the call
5085  * takes place.
5086  */
5087 void wlcore_update_inconn_sta(struct wl1271 *wl, struct wl12xx_vif *wlvif,
5088 			      struct wl1271_station *wl_sta, bool in_conn)
5089 {
5090 	if (in_conn) {
5091 		if (WARN_ON(wl_sta && wl_sta->in_connection))
5092 			return;
5093 
5094 		if (!wlvif->ap_pending_auth_reply &&
5095 		    !wlvif->inconn_count)
5096 			wlcore_roc_if_possible(wl, wlvif);
5097 
5098 		if (wl_sta) {
5099 			wl_sta->in_connection = true;
5100 			wlvif->inconn_count++;
5101 		} else {
5102 			wlvif->ap_pending_auth_reply = true;
5103 		}
5104 	} else {
5105 		if (wl_sta && !wl_sta->in_connection)
5106 			return;
5107 
5108 		if (WARN_ON(!wl_sta && !wlvif->ap_pending_auth_reply))
5109 			return;
5110 
5111 		if (WARN_ON(wl_sta && !wlvif->inconn_count))
5112 			return;
5113 
5114 		if (wl_sta) {
5115 			wl_sta->in_connection = false;
5116 			wlvif->inconn_count--;
5117 		} else {
5118 			wlvif->ap_pending_auth_reply = false;
5119 		}
5120 
5121 		if (!wlvif->inconn_count && !wlvif->ap_pending_auth_reply &&
5122 		    test_bit(wlvif->role_id, wl->roc_map))
5123 			wl12xx_croc(wl, wlvif->role_id);
5124 	}
5125 }
5126 
5127 static int wl12xx_update_sta_state(struct wl1271 *wl,
5128 				   struct wl12xx_vif *wlvif,
5129 				   struct ieee80211_sta *sta,
5130 				   enum ieee80211_sta_state old_state,
5131 				   enum ieee80211_sta_state new_state)
5132 {
5133 	struct wl1271_station *wl_sta;
5134 	bool is_ap = wlvif->bss_type == BSS_TYPE_AP_BSS;
5135 	bool is_sta = wlvif->bss_type == BSS_TYPE_STA_BSS;
5136 	int ret;
5137 
5138 	wl_sta = (struct wl1271_station *)sta->drv_priv;
5139 
5140 	/* Add station (AP mode) */
5141 	if (is_ap &&
5142 	    old_state == IEEE80211_STA_NOTEXIST &&
5143 	    new_state == IEEE80211_STA_NONE) {
5144 		ret = wl12xx_sta_add(wl, wlvif, sta);
5145 		if (ret)
5146 			return ret;
5147 
5148 		wlcore_update_inconn_sta(wl, wlvif, wl_sta, true);
5149 	}
5150 
5151 	/* Remove station (AP mode) */
5152 	if (is_ap &&
5153 	    old_state == IEEE80211_STA_NONE &&
5154 	    new_state == IEEE80211_STA_NOTEXIST) {
5155 		/* must not fail */
5156 		wl12xx_sta_remove(wl, wlvif, sta);
5157 
5158 		wlcore_update_inconn_sta(wl, wlvif, wl_sta, false);
5159 	}
5160 
5161 	/* Authorize station (AP mode) */
5162 	if (is_ap &&
5163 	    new_state == IEEE80211_STA_AUTHORIZED) {
5164 		ret = wl12xx_cmd_set_peer_state(wl, wlvif, wl_sta->hlid);
5165 		if (ret < 0)
5166 			return ret;
5167 
5168 		/* reconfigure rates */
5169 		ret = wl12xx_cmd_add_peer(wl, wlvif, sta, wl_sta->hlid);
5170 		if (ret < 0)
5171 			return ret;
5172 
5173 		ret = wl1271_acx_set_ht_capabilities(wl, &sta->deflink.ht_cap,
5174 						     true,
5175 						     wl_sta->hlid);
5176 		if (ret)
5177 			return ret;
5178 
5179 		wlcore_update_inconn_sta(wl, wlvif, wl_sta, false);
5180 	}
5181 
5182 	/* Authorize station */
5183 	if (is_sta &&
5184 	    new_state == IEEE80211_STA_AUTHORIZED) {
5185 		set_bit(WLVIF_FLAG_STA_AUTHORIZED, &wlvif->flags);
5186 		ret = wl12xx_set_authorized(wl, wlvif);
5187 		if (ret)
5188 			return ret;
5189 	}
5190 
5191 	if (is_sta &&
5192 	    old_state == IEEE80211_STA_AUTHORIZED &&
5193 	    new_state == IEEE80211_STA_ASSOC) {
5194 		clear_bit(WLVIF_FLAG_STA_AUTHORIZED, &wlvif->flags);
5195 		clear_bit(WLVIF_FLAG_STA_STATE_SENT, &wlvif->flags);
5196 	}
5197 
5198 	/* save seq number on disassoc (suspend) */
5199 	if (is_sta &&
5200 	    old_state == IEEE80211_STA_ASSOC &&
5201 	    new_state == IEEE80211_STA_AUTH) {
5202 		wlcore_save_freed_pkts(wl, wlvif, wlvif->sta.hlid, sta);
5203 		wlvif->total_freed_pkts = 0;
5204 	}
5205 
5206 	/* restore seq number on assoc (resume) */
5207 	if (is_sta &&
5208 	    old_state == IEEE80211_STA_AUTH &&
5209 	    new_state == IEEE80211_STA_ASSOC) {
5210 		wlvif->total_freed_pkts = wl_sta->total_freed_pkts;
5211 	}
5212 
5213 	/* clear ROCs on failure or authorization */
5214 	if (is_sta &&
5215 	    (new_state == IEEE80211_STA_AUTHORIZED ||
5216 	     new_state == IEEE80211_STA_NOTEXIST)) {
5217 		if (test_bit(wlvif->role_id, wl->roc_map))
5218 			wl12xx_croc(wl, wlvif->role_id);
5219 	}
5220 
5221 	if (is_sta &&
5222 	    old_state == IEEE80211_STA_NOTEXIST &&
5223 	    new_state == IEEE80211_STA_NONE) {
5224 		if (find_first_bit(wl->roc_map,
5225 				   WL12XX_MAX_ROLES) >= WL12XX_MAX_ROLES) {
5226 			WARN_ON(wlvif->role_id == WL12XX_INVALID_ROLE_ID);
5227 			wl12xx_roc(wl, wlvif, wlvif->role_id,
5228 				   wlvif->band, wlvif->channel);
5229 		}
5230 	}
5231 	return 0;
5232 }
5233 
5234 static int wl12xx_op_sta_state(struct ieee80211_hw *hw,
5235 			       struct ieee80211_vif *vif,
5236 			       struct ieee80211_sta *sta,
5237 			       enum ieee80211_sta_state old_state,
5238 			       enum ieee80211_sta_state new_state)
5239 {
5240 	struct wl1271 *wl = hw->priv;
5241 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5242 	int ret;
5243 
5244 	wl1271_debug(DEBUG_MAC80211, "mac80211 sta %d state=%d->%d",
5245 		     sta->aid, old_state, new_state);
5246 
5247 	mutex_lock(&wl->mutex);
5248 
5249 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
5250 		ret = -EBUSY;
5251 		goto out;
5252 	}
5253 
5254 	ret = pm_runtime_resume_and_get(wl->dev);
5255 	if (ret < 0)
5256 		goto out;
5257 
5258 	ret = wl12xx_update_sta_state(wl, wlvif, sta, old_state, new_state);
5259 
5260 	pm_runtime_mark_last_busy(wl->dev);
5261 	pm_runtime_put_autosuspend(wl->dev);
5262 out:
5263 	mutex_unlock(&wl->mutex);
5264 	if (new_state < old_state)
5265 		return 0;
5266 	return ret;
5267 }
5268 
5269 static int wl1271_op_ampdu_action(struct ieee80211_hw *hw,
5270 				  struct ieee80211_vif *vif,
5271 				  struct ieee80211_ampdu_params *params)
5272 {
5273 	struct wl1271 *wl = hw->priv;
5274 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5275 	int ret;
5276 	u8 hlid, *ba_bitmap;
5277 	struct ieee80211_sta *sta = params->sta;
5278 	enum ieee80211_ampdu_mlme_action action = params->action;
5279 	u16 tid = params->tid;
5280 	u16 *ssn = &params->ssn;
5281 
5282 	wl1271_debug(DEBUG_MAC80211, "mac80211 ampdu action %d tid %d", action,
5283 		     tid);
5284 
5285 	/* sanity check - the fields in FW are only 8bits wide */
5286 	if (WARN_ON(tid > 0xFF))
5287 		return -ENOTSUPP;
5288 
5289 	mutex_lock(&wl->mutex);
5290 
5291 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
5292 		ret = -EAGAIN;
5293 		goto out;
5294 	}
5295 
5296 	if (wlvif->bss_type == BSS_TYPE_STA_BSS) {
5297 		hlid = wlvif->sta.hlid;
5298 	} else if (wlvif->bss_type == BSS_TYPE_AP_BSS) {
5299 		struct wl1271_station *wl_sta;
5300 
5301 		wl_sta = (struct wl1271_station *)sta->drv_priv;
5302 		hlid = wl_sta->hlid;
5303 	} else {
5304 		ret = -EINVAL;
5305 		goto out;
5306 	}
5307 
5308 	ba_bitmap = &wl->links[hlid].ba_bitmap;
5309 
5310 	ret = pm_runtime_resume_and_get(wl->dev);
5311 	if (ret < 0)
5312 		goto out;
5313 
5314 	wl1271_debug(DEBUG_MAC80211, "mac80211 ampdu: Rx tid %d action %d",
5315 		     tid, action);
5316 
5317 	switch (action) {
5318 	case IEEE80211_AMPDU_RX_START:
5319 		if (!wlvif->ba_support || !wlvif->ba_allowed) {
5320 			ret = -ENOTSUPP;
5321 			break;
5322 		}
5323 
5324 		if (wl->ba_rx_session_count >= wl->ba_rx_session_count_max) {
5325 			ret = -EBUSY;
5326 			wl1271_debug(DEBUG_RX, "exceeded max RX BA sessions");
5327 			break;
5328 		}
5329 
5330 		if (*ba_bitmap & BIT(tid)) {
5331 			ret = -EINVAL;
5332 			wl1271_error("cannot enable RX BA session on active "
5333 				     "tid: %d", tid);
5334 			break;
5335 		}
5336 
5337 		ret = wl12xx_acx_set_ba_receiver_session(wl, tid, *ssn, true,
5338 				hlid,
5339 				params->buf_size);
5340 
5341 		if (!ret) {
5342 			*ba_bitmap |= BIT(tid);
5343 			wl->ba_rx_session_count++;
5344 		}
5345 		break;
5346 
5347 	case IEEE80211_AMPDU_RX_STOP:
5348 		if (!(*ba_bitmap & BIT(tid))) {
5349 			/*
5350 			 * this happens on reconfig - so only output a debug
5351 			 * message for now, and don't fail the function.
5352 			 */
5353 			wl1271_debug(DEBUG_MAC80211,
5354 				     "no active RX BA session on tid: %d",
5355 				     tid);
5356 			ret = 0;
5357 			break;
5358 		}
5359 
5360 		ret = wl12xx_acx_set_ba_receiver_session(wl, tid, 0, false,
5361 							 hlid, 0);
5362 		if (!ret) {
5363 			*ba_bitmap &= ~BIT(tid);
5364 			wl->ba_rx_session_count--;
5365 		}
5366 		break;
5367 
5368 	/*
5369 	 * The BA initiator session management in FW independently.
5370 	 * Falling break here on purpose for all TX APDU commands.
5371 	 */
5372 	case IEEE80211_AMPDU_TX_START:
5373 	case IEEE80211_AMPDU_TX_STOP_CONT:
5374 	case IEEE80211_AMPDU_TX_STOP_FLUSH:
5375 	case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT:
5376 	case IEEE80211_AMPDU_TX_OPERATIONAL:
5377 		ret = -EINVAL;
5378 		break;
5379 
5380 	default:
5381 		wl1271_error("Incorrect ampdu action id=%x\n", action);
5382 		ret = -EINVAL;
5383 	}
5384 
5385 	pm_runtime_mark_last_busy(wl->dev);
5386 	pm_runtime_put_autosuspend(wl->dev);
5387 
5388 out:
5389 	mutex_unlock(&wl->mutex);
5390 
5391 	return ret;
5392 }
5393 
5394 static int wl12xx_set_bitrate_mask(struct ieee80211_hw *hw,
5395 				   struct ieee80211_vif *vif,
5396 				   const struct cfg80211_bitrate_mask *mask)
5397 {
5398 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5399 	struct wl1271 *wl = hw->priv;
5400 	int i, ret = 0;
5401 
5402 	wl1271_debug(DEBUG_MAC80211, "mac80211 set_bitrate_mask 0x%x 0x%x",
5403 		mask->control[NL80211_BAND_2GHZ].legacy,
5404 		mask->control[NL80211_BAND_5GHZ].legacy);
5405 
5406 	mutex_lock(&wl->mutex);
5407 
5408 	for (i = 0; i < WLCORE_NUM_BANDS; i++)
5409 		wlvif->bitrate_masks[i] =
5410 			wl1271_tx_enabled_rates_get(wl,
5411 						    mask->control[i].legacy,
5412 						    i);
5413 
5414 	if (unlikely(wl->state != WLCORE_STATE_ON))
5415 		goto out;
5416 
5417 	if (wlvif->bss_type == BSS_TYPE_STA_BSS &&
5418 	    !test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags)) {
5419 
5420 		ret = pm_runtime_resume_and_get(wl->dev);
5421 		if (ret < 0)
5422 			goto out;
5423 
5424 		wl1271_set_band_rate(wl, wlvif);
5425 		wlvif->basic_rate =
5426 			wl1271_tx_min_rate_get(wl, wlvif->basic_rate_set);
5427 		ret = wl1271_acx_sta_rate_policies(wl, wlvif);
5428 
5429 		pm_runtime_mark_last_busy(wl->dev);
5430 		pm_runtime_put_autosuspend(wl->dev);
5431 	}
5432 out:
5433 	mutex_unlock(&wl->mutex);
5434 
5435 	return ret;
5436 }
5437 
5438 static void wl12xx_op_channel_switch(struct ieee80211_hw *hw,
5439 				     struct ieee80211_vif *vif,
5440 				     struct ieee80211_channel_switch *ch_switch)
5441 {
5442 	struct wl1271 *wl = hw->priv;
5443 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5444 	int ret;
5445 
5446 	wl1271_debug(DEBUG_MAC80211, "mac80211 channel switch");
5447 
5448 	wl1271_tx_flush(wl);
5449 
5450 	mutex_lock(&wl->mutex);
5451 
5452 	if (unlikely(wl->state == WLCORE_STATE_OFF)) {
5453 		if (test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
5454 			ieee80211_chswitch_done(vif, false);
5455 		goto out;
5456 	} else if (unlikely(wl->state != WLCORE_STATE_ON)) {
5457 		goto out;
5458 	}
5459 
5460 	ret = pm_runtime_resume_and_get(wl->dev);
5461 	if (ret < 0)
5462 		goto out;
5463 
5464 	/* TODO: change mac80211 to pass vif as param */
5465 
5466 	if (test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags)) {
5467 		unsigned long delay_usec;
5468 
5469 		ret = wl->ops->channel_switch(wl, wlvif, ch_switch);
5470 		if (ret)
5471 			goto out_sleep;
5472 
5473 		set_bit(WLVIF_FLAG_CS_PROGRESS, &wlvif->flags);
5474 
5475 		/* indicate failure 5 seconds after channel switch time */
5476 		delay_usec = ieee80211_tu_to_usec(wlvif->beacon_int) *
5477 			ch_switch->count;
5478 		ieee80211_queue_delayed_work(hw, &wlvif->channel_switch_work,
5479 					     usecs_to_jiffies(delay_usec) +
5480 					     msecs_to_jiffies(5000));
5481 	}
5482 
5483 out_sleep:
5484 	pm_runtime_mark_last_busy(wl->dev);
5485 	pm_runtime_put_autosuspend(wl->dev);
5486 
5487 out:
5488 	mutex_unlock(&wl->mutex);
5489 }
5490 
5491 static const void *wlcore_get_beacon_ie(struct wl1271 *wl,
5492 					struct wl12xx_vif *wlvif,
5493 					u8 eid)
5494 {
5495 	int ieoffset = offsetof(struct ieee80211_mgmt, u.beacon.variable);
5496 	struct sk_buff *beacon =
5497 		ieee80211_beacon_get(wl->hw, wl12xx_wlvif_to_vif(wlvif), 0);
5498 
5499 	if (!beacon)
5500 		return NULL;
5501 
5502 	return cfg80211_find_ie(eid,
5503 				beacon->data + ieoffset,
5504 				beacon->len - ieoffset);
5505 }
5506 
5507 static int wlcore_get_csa_count(struct wl1271 *wl, struct wl12xx_vif *wlvif,
5508 				u8 *csa_count)
5509 {
5510 	const u8 *ie;
5511 	const struct ieee80211_channel_sw_ie *ie_csa;
5512 
5513 	ie = wlcore_get_beacon_ie(wl, wlvif, WLAN_EID_CHANNEL_SWITCH);
5514 	if (!ie)
5515 		return -EINVAL;
5516 
5517 	ie_csa = (struct ieee80211_channel_sw_ie *)&ie[2];
5518 	*csa_count = ie_csa->count;
5519 
5520 	return 0;
5521 }
5522 
5523 static void wlcore_op_channel_switch_beacon(struct ieee80211_hw *hw,
5524 					    struct ieee80211_vif *vif,
5525 					    struct cfg80211_chan_def *chandef)
5526 {
5527 	struct wl1271 *wl = hw->priv;
5528 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5529 	struct ieee80211_channel_switch ch_switch = {
5530 		.block_tx = true,
5531 		.chandef = *chandef,
5532 	};
5533 	int ret;
5534 
5535 	wl1271_debug(DEBUG_MAC80211,
5536 		     "mac80211 channel switch beacon (role %d)",
5537 		     wlvif->role_id);
5538 
5539 	ret = wlcore_get_csa_count(wl, wlvif, &ch_switch.count);
5540 	if (ret < 0) {
5541 		wl1271_error("error getting beacon (for CSA counter)");
5542 		return;
5543 	}
5544 
5545 	mutex_lock(&wl->mutex);
5546 
5547 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
5548 		ret = -EBUSY;
5549 		goto out;
5550 	}
5551 
5552 	ret = pm_runtime_resume_and_get(wl->dev);
5553 	if (ret < 0)
5554 		goto out;
5555 
5556 	ret = wl->ops->channel_switch(wl, wlvif, &ch_switch);
5557 	if (ret)
5558 		goto out_sleep;
5559 
5560 	set_bit(WLVIF_FLAG_CS_PROGRESS, &wlvif->flags);
5561 
5562 out_sleep:
5563 	pm_runtime_mark_last_busy(wl->dev);
5564 	pm_runtime_put_autosuspend(wl->dev);
5565 out:
5566 	mutex_unlock(&wl->mutex);
5567 }
5568 
5569 static void wlcore_op_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
5570 			    u32 queues, bool drop)
5571 {
5572 	struct wl1271 *wl = hw->priv;
5573 
5574 	wl1271_tx_flush(wl);
5575 }
5576 
5577 static int wlcore_op_remain_on_channel(struct ieee80211_hw *hw,
5578 				       struct ieee80211_vif *vif,
5579 				       struct ieee80211_channel *chan,
5580 				       int duration,
5581 				       enum ieee80211_roc_type type)
5582 {
5583 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5584 	struct wl1271 *wl = hw->priv;
5585 	int channel, active_roc, ret = 0;
5586 
5587 	channel = ieee80211_frequency_to_channel(chan->center_freq);
5588 
5589 	wl1271_debug(DEBUG_MAC80211, "mac80211 roc %d (%d)",
5590 		     channel, wlvif->role_id);
5591 
5592 	mutex_lock(&wl->mutex);
5593 
5594 	if (unlikely(wl->state != WLCORE_STATE_ON))
5595 		goto out;
5596 
5597 	/* return EBUSY if we can't ROC right now */
5598 	active_roc = find_first_bit(wl->roc_map, WL12XX_MAX_ROLES);
5599 	if (wl->roc_vif || active_roc < WL12XX_MAX_ROLES) {
5600 		wl1271_warning("active roc on role %d", active_roc);
5601 		ret = -EBUSY;
5602 		goto out;
5603 	}
5604 
5605 	ret = pm_runtime_resume_and_get(wl->dev);
5606 	if (ret < 0)
5607 		goto out;
5608 
5609 	ret = wl12xx_start_dev(wl, wlvif, chan->band, channel);
5610 	if (ret < 0)
5611 		goto out_sleep;
5612 
5613 	wl->roc_vif = vif;
5614 	ieee80211_queue_delayed_work(hw, &wl->roc_complete_work,
5615 				     msecs_to_jiffies(duration));
5616 out_sleep:
5617 	pm_runtime_mark_last_busy(wl->dev);
5618 	pm_runtime_put_autosuspend(wl->dev);
5619 out:
5620 	mutex_unlock(&wl->mutex);
5621 	return ret;
5622 }
5623 
5624 static int __wlcore_roc_completed(struct wl1271 *wl)
5625 {
5626 	struct wl12xx_vif *wlvif;
5627 	int ret;
5628 
5629 	/* already completed */
5630 	if (unlikely(!wl->roc_vif))
5631 		return 0;
5632 
5633 	wlvif = wl12xx_vif_to_data(wl->roc_vif);
5634 
5635 	if (!test_bit(WLVIF_FLAG_INITIALIZED, &wlvif->flags))
5636 		return -EBUSY;
5637 
5638 	ret = wl12xx_stop_dev(wl, wlvif);
5639 	if (ret < 0)
5640 		return ret;
5641 
5642 	wl->roc_vif = NULL;
5643 
5644 	return 0;
5645 }
5646 
5647 static int wlcore_roc_completed(struct wl1271 *wl)
5648 {
5649 	int ret;
5650 
5651 	wl1271_debug(DEBUG_MAC80211, "roc complete");
5652 
5653 	mutex_lock(&wl->mutex);
5654 
5655 	if (unlikely(wl->state != WLCORE_STATE_ON)) {
5656 		ret = -EBUSY;
5657 		goto out;
5658 	}
5659 
5660 	ret = pm_runtime_resume_and_get(wl->dev);
5661 	if (ret < 0)
5662 		goto out;
5663 
5664 	ret = __wlcore_roc_completed(wl);
5665 
5666 	pm_runtime_mark_last_busy(wl->dev);
5667 	pm_runtime_put_autosuspend(wl->dev);
5668 out:
5669 	mutex_unlock(&wl->mutex);
5670 
5671 	return ret;
5672 }
5673 
5674 static void wlcore_roc_complete_work(struct work_struct *work)
5675 {
5676 	struct delayed_work *dwork;
5677 	struct wl1271 *wl;
5678 	int ret;
5679 
5680 	dwork = to_delayed_work(work);
5681 	wl = container_of(dwork, struct wl1271, roc_complete_work);
5682 
5683 	ret = wlcore_roc_completed(wl);
5684 	if (!ret)
5685 		ieee80211_remain_on_channel_expired(wl->hw);
5686 }
5687 
5688 static int wlcore_op_cancel_remain_on_channel(struct ieee80211_hw *hw,
5689 					      struct ieee80211_vif *vif)
5690 {
5691 	struct wl1271 *wl = hw->priv;
5692 
5693 	wl1271_debug(DEBUG_MAC80211, "mac80211 croc");
5694 
5695 	/* TODO: per-vif */
5696 	wl1271_tx_flush(wl);
5697 
5698 	/*
5699 	 * we can't just flush_work here, because it might deadlock
5700 	 * (as we might get called from the same workqueue)
5701 	 */
5702 	cancel_delayed_work_sync(&wl->roc_complete_work);
5703 	wlcore_roc_completed(wl);
5704 
5705 	return 0;
5706 }
5707 
5708 static void wlcore_op_sta_rc_update(struct ieee80211_hw *hw,
5709 				    struct ieee80211_vif *vif,
5710 				    struct ieee80211_sta *sta,
5711 				    u32 changed)
5712 {
5713 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5714 
5715 	wl1271_debug(DEBUG_MAC80211, "mac80211 sta_rc_update");
5716 
5717 	if (!(changed & IEEE80211_RC_BW_CHANGED))
5718 		return;
5719 
5720 	/* this callback is atomic, so schedule a new work */
5721 	wlvif->rc_update_bw = sta->deflink.bandwidth;
5722 	memcpy(&wlvif->rc_ht_cap, &sta->deflink.ht_cap,
5723 	       sizeof(sta->deflink.ht_cap));
5724 	ieee80211_queue_work(hw, &wlvif->rc_update_work);
5725 }
5726 
5727 static void wlcore_op_sta_statistics(struct ieee80211_hw *hw,
5728 				     struct ieee80211_vif *vif,
5729 				     struct ieee80211_sta *sta,
5730 				     struct station_info *sinfo)
5731 {
5732 	struct wl1271 *wl = hw->priv;
5733 	struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
5734 	s8 rssi_dbm;
5735 	int ret;
5736 
5737 	wl1271_debug(DEBUG_MAC80211, "mac80211 get_rssi");
5738 
5739 	mutex_lock(&wl->mutex);
5740 
5741 	if (unlikely(wl->state != WLCORE_STATE_ON))
5742 		goto out;
5743 
5744 	ret = pm_runtime_resume_and_get(wl->dev);
5745 	if (ret < 0)
5746 		goto out_sleep;
5747 
5748 	ret = wlcore_acx_average_rssi(wl, wlvif, &rssi_dbm);
5749 	if (ret < 0)
5750 		goto out_sleep;
5751 
5752 	sinfo->filled |= BIT_ULL(NL80211_STA_INFO_SIGNAL);
5753 	sinfo->signal = rssi_dbm;
5754 
5755 out_sleep:
5756 	pm_runtime_mark_last_busy(wl->dev);
5757 	pm_runtime_put_autosuspend(wl->dev);
5758 
5759 out:
5760 	mutex_unlock(&wl->mutex);
5761 }
5762 
5763 static u32 wlcore_op_get_expected_throughput(struct ieee80211_hw *hw,
5764 					     struct ieee80211_sta *sta)
5765 {
5766 	struct wl1271_station *wl_sta = (struct wl1271_station *)sta->drv_priv;
5767 	struct wl1271 *wl = hw->priv;
5768 	u8 hlid = wl_sta->hlid;
5769 
5770 	/* return in units of Kbps */
5771 	return (wl->links[hlid].fw_rate_mbps * 1000);
5772 }
5773 
5774 static bool wl1271_tx_frames_pending(struct ieee80211_hw *hw)
5775 {
5776 	struct wl1271 *wl = hw->priv;
5777 	bool ret = false;
5778 
5779 	mutex_lock(&wl->mutex);
5780 
5781 	if (unlikely(wl->state != WLCORE_STATE_ON))
5782 		goto out;
5783 
5784 	/* packets are considered pending if in the TX queue or the FW */
5785 	ret = (wl1271_tx_total_queue_count(wl) > 0) || (wl->tx_frames_cnt > 0);
5786 out:
5787 	mutex_unlock(&wl->mutex);
5788 
5789 	return ret;
5790 }
5791 
5792 /* can't be const, mac80211 writes to this */
5793 static struct ieee80211_rate wl1271_rates[] = {
5794 	{ .bitrate = 10,
5795 	  .hw_value = CONF_HW_BIT_RATE_1MBPS,
5796 	  .hw_value_short = CONF_HW_BIT_RATE_1MBPS, },
5797 	{ .bitrate = 20,
5798 	  .hw_value = CONF_HW_BIT_RATE_2MBPS,
5799 	  .hw_value_short = CONF_HW_BIT_RATE_2MBPS,
5800 	  .flags = IEEE80211_RATE_SHORT_PREAMBLE },
5801 	{ .bitrate = 55,
5802 	  .hw_value = CONF_HW_BIT_RATE_5_5MBPS,
5803 	  .hw_value_short = CONF_HW_BIT_RATE_5_5MBPS,
5804 	  .flags = IEEE80211_RATE_SHORT_PREAMBLE },
5805 	{ .bitrate = 110,
5806 	  .hw_value = CONF_HW_BIT_RATE_11MBPS,
5807 	  .hw_value_short = CONF_HW_BIT_RATE_11MBPS,
5808 	  .flags = IEEE80211_RATE_SHORT_PREAMBLE },
5809 	{ .bitrate = 60,
5810 	  .hw_value = CONF_HW_BIT_RATE_6MBPS,
5811 	  .hw_value_short = CONF_HW_BIT_RATE_6MBPS, },
5812 	{ .bitrate = 90,
5813 	  .hw_value = CONF_HW_BIT_RATE_9MBPS,
5814 	  .hw_value_short = CONF_HW_BIT_RATE_9MBPS, },
5815 	{ .bitrate = 120,
5816 	  .hw_value = CONF_HW_BIT_RATE_12MBPS,
5817 	  .hw_value_short = CONF_HW_BIT_RATE_12MBPS, },
5818 	{ .bitrate = 180,
5819 	  .hw_value = CONF_HW_BIT_RATE_18MBPS,
5820 	  .hw_value_short = CONF_HW_BIT_RATE_18MBPS, },
5821 	{ .bitrate = 240,
5822 	  .hw_value = CONF_HW_BIT_RATE_24MBPS,
5823 	  .hw_value_short = CONF_HW_BIT_RATE_24MBPS, },
5824 	{ .bitrate = 360,
5825 	 .hw_value = CONF_HW_BIT_RATE_36MBPS,
5826 	 .hw_value_short = CONF_HW_BIT_RATE_36MBPS, },
5827 	{ .bitrate = 480,
5828 	  .hw_value = CONF_HW_BIT_RATE_48MBPS,
5829 	  .hw_value_short = CONF_HW_BIT_RATE_48MBPS, },
5830 	{ .bitrate = 540,
5831 	  .hw_value = CONF_HW_BIT_RATE_54MBPS,
5832 	  .hw_value_short = CONF_HW_BIT_RATE_54MBPS, },
5833 };
5834 
5835 /* can't be const, mac80211 writes to this */
5836 static struct ieee80211_channel wl1271_channels[] = {
5837 	{ .hw_value = 1, .center_freq = 2412, .max_power = WLCORE_MAX_TXPWR },
5838 	{ .hw_value = 2, .center_freq = 2417, .max_power = WLCORE_MAX_TXPWR },
5839 	{ .hw_value = 3, .center_freq = 2422, .max_power = WLCORE_MAX_TXPWR },
5840 	{ .hw_value = 4, .center_freq = 2427, .max_power = WLCORE_MAX_TXPWR },
5841 	{ .hw_value = 5, .center_freq = 2432, .max_power = WLCORE_MAX_TXPWR },
5842 	{ .hw_value = 6, .center_freq = 2437, .max_power = WLCORE_MAX_TXPWR },
5843 	{ .hw_value = 7, .center_freq = 2442, .max_power = WLCORE_MAX_TXPWR },
5844 	{ .hw_value = 8, .center_freq = 2447, .max_power = WLCORE_MAX_TXPWR },
5845 	{ .hw_value = 9, .center_freq = 2452, .max_power = WLCORE_MAX_TXPWR },
5846 	{ .hw_value = 10, .center_freq = 2457, .max_power = WLCORE_MAX_TXPWR },
5847 	{ .hw_value = 11, .center_freq = 2462, .max_power = WLCORE_MAX_TXPWR },
5848 	{ .hw_value = 12, .center_freq = 2467, .max_power = WLCORE_MAX_TXPWR },
5849 	{ .hw_value = 13, .center_freq = 2472, .max_power = WLCORE_MAX_TXPWR },
5850 	{ .hw_value = 14, .center_freq = 2484, .max_power = WLCORE_MAX_TXPWR },
5851 };
5852 
5853 /* can't be const, mac80211 writes to this */
5854 static struct ieee80211_supported_band wl1271_band_2ghz = {
5855 	.channels = wl1271_channels,
5856 	.n_channels = ARRAY_SIZE(wl1271_channels),
5857 	.bitrates = wl1271_rates,
5858 	.n_bitrates = ARRAY_SIZE(wl1271_rates),
5859 };
5860 
5861 /* 5 GHz data rates for WL1273 */
5862 static struct ieee80211_rate wl1271_rates_5ghz[] = {
5863 	{ .bitrate = 60,
5864 	  .hw_value = CONF_HW_BIT_RATE_6MBPS,
5865 	  .hw_value_short = CONF_HW_BIT_RATE_6MBPS, },
5866 	{ .bitrate = 90,
5867 	  .hw_value = CONF_HW_BIT_RATE_9MBPS,
5868 	  .hw_value_short = CONF_HW_BIT_RATE_9MBPS, },
5869 	{ .bitrate = 120,
5870 	  .hw_value = CONF_HW_BIT_RATE_12MBPS,
5871 	  .hw_value_short = CONF_HW_BIT_RATE_12MBPS, },
5872 	{ .bitrate = 180,
5873 	  .hw_value = CONF_HW_BIT_RATE_18MBPS,
5874 	  .hw_value_short = CONF_HW_BIT_RATE_18MBPS, },
5875 	{ .bitrate = 240,
5876 	  .hw_value = CONF_HW_BIT_RATE_24MBPS,
5877 	  .hw_value_short = CONF_HW_BIT_RATE_24MBPS, },
5878 	{ .bitrate = 360,
5879 	 .hw_value = CONF_HW_BIT_RATE_36MBPS,
5880 	 .hw_value_short = CONF_HW_BIT_RATE_36MBPS, },
5881 	{ .bitrate = 480,
5882 	  .hw_value = CONF_HW_BIT_RATE_48MBPS,
5883 	  .hw_value_short = CONF_HW_BIT_RATE_48MBPS, },
5884 	{ .bitrate = 540,
5885 	  .hw_value = CONF_HW_BIT_RATE_54MBPS,
5886 	  .hw_value_short = CONF_HW_BIT_RATE_54MBPS, },
5887 };
5888 
5889 /* 5 GHz band channels for WL1273 */
5890 static struct ieee80211_channel wl1271_channels_5ghz[] = {
5891 	{ .hw_value = 8, .center_freq = 5040, .max_power = WLCORE_MAX_TXPWR },
5892 	{ .hw_value = 12, .center_freq = 5060, .max_power = WLCORE_MAX_TXPWR },
5893 	{ .hw_value = 16, .center_freq = 5080, .max_power = WLCORE_MAX_TXPWR },
5894 	{ .hw_value = 34, .center_freq = 5170, .max_power = WLCORE_MAX_TXPWR },
5895 	{ .hw_value = 36, .center_freq = 5180, .max_power = WLCORE_MAX_TXPWR },
5896 	{ .hw_value = 38, .center_freq = 5190, .max_power = WLCORE_MAX_TXPWR },
5897 	{ .hw_value = 40, .center_freq = 5200, .max_power = WLCORE_MAX_TXPWR },
5898 	{ .hw_value = 42, .center_freq = 5210, .max_power = WLCORE_MAX_TXPWR },
5899 	{ .hw_value = 44, .center_freq = 5220, .max_power = WLCORE_MAX_TXPWR },
5900 	{ .hw_value = 46, .center_freq = 5230, .max_power = WLCORE_MAX_TXPWR },
5901 	{ .hw_value = 48, .center_freq = 5240, .max_power = WLCORE_MAX_TXPWR },
5902 	{ .hw_value = 52, .center_freq = 5260, .max_power = WLCORE_MAX_TXPWR },
5903 	{ .hw_value = 56, .center_freq = 5280, .max_power = WLCORE_MAX_TXPWR },
5904 	{ .hw_value = 60, .center_freq = 5300, .max_power = WLCORE_MAX_TXPWR },
5905 	{ .hw_value = 64, .center_freq = 5320, .max_power = WLCORE_MAX_TXPWR },
5906 	{ .hw_value = 100, .center_freq = 5500, .max_power = WLCORE_MAX_TXPWR },
5907 	{ .hw_value = 104, .center_freq = 5520, .max_power = WLCORE_MAX_TXPWR },
5908 	{ .hw_value = 108, .center_freq = 5540, .max_power = WLCORE_MAX_TXPWR },
5909 	{ .hw_value = 112, .center_freq = 5560, .max_power = WLCORE_MAX_TXPWR },
5910 	{ .hw_value = 116, .center_freq = 5580, .max_power = WLCORE_MAX_TXPWR },
5911 	{ .hw_value = 120, .center_freq = 5600, .max_power = WLCORE_MAX_TXPWR },
5912 	{ .hw_value = 124, .center_freq = 5620, .max_power = WLCORE_MAX_TXPWR },
5913 	{ .hw_value = 128, .center_freq = 5640, .max_power = WLCORE_MAX_TXPWR },
5914 	{ .hw_value = 132, .center_freq = 5660, .max_power = WLCORE_MAX_TXPWR },
5915 	{ .hw_value = 136, .center_freq = 5680, .max_power = WLCORE_MAX_TXPWR },
5916 	{ .hw_value = 140, .center_freq = 5700, .max_power = WLCORE_MAX_TXPWR },
5917 	{ .hw_value = 149, .center_freq = 5745, .max_power = WLCORE_MAX_TXPWR },
5918 	{ .hw_value = 153, .center_freq = 5765, .max_power = WLCORE_MAX_TXPWR },
5919 	{ .hw_value = 157, .center_freq = 5785, .max_power = WLCORE_MAX_TXPWR },
5920 	{ .hw_value = 161, .center_freq = 5805, .max_power = WLCORE_MAX_TXPWR },
5921 	{ .hw_value = 165, .center_freq = 5825, .max_power = WLCORE_MAX_TXPWR },
5922 };
5923 
5924 static struct ieee80211_supported_band wl1271_band_5ghz = {
5925 	.channels = wl1271_channels_5ghz,
5926 	.n_channels = ARRAY_SIZE(wl1271_channels_5ghz),
5927 	.bitrates = wl1271_rates_5ghz,
5928 	.n_bitrates = ARRAY_SIZE(wl1271_rates_5ghz),
5929 };
5930 
5931 static const struct ieee80211_ops wl1271_ops = {
5932 	.start = wl1271_op_start,
5933 	.stop = wlcore_op_stop,
5934 	.add_interface = wl1271_op_add_interface,
5935 	.remove_interface = wl1271_op_remove_interface,
5936 	.change_interface = wl12xx_op_change_interface,
5937 #ifdef CONFIG_PM
5938 	.suspend = wl1271_op_suspend,
5939 	.resume = wl1271_op_resume,
5940 #endif
5941 	.config = wl1271_op_config,
5942 	.prepare_multicast = wl1271_op_prepare_multicast,
5943 	.configure_filter = wl1271_op_configure_filter,
5944 	.tx = wl1271_op_tx,
5945 	.set_key = wlcore_op_set_key,
5946 	.hw_scan = wl1271_op_hw_scan,
5947 	.cancel_hw_scan = wl1271_op_cancel_hw_scan,
5948 	.sched_scan_start = wl1271_op_sched_scan_start,
5949 	.sched_scan_stop = wl1271_op_sched_scan_stop,
5950 	.bss_info_changed = wl1271_op_bss_info_changed,
5951 	.set_frag_threshold = wl1271_op_set_frag_threshold,
5952 	.set_rts_threshold = wl1271_op_set_rts_threshold,
5953 	.conf_tx = wl1271_op_conf_tx,
5954 	.get_tsf = wl1271_op_get_tsf,
5955 	.get_survey = wl1271_op_get_survey,
5956 	.sta_state = wl12xx_op_sta_state,
5957 	.ampdu_action = wl1271_op_ampdu_action,
5958 	.tx_frames_pending = wl1271_tx_frames_pending,
5959 	.set_bitrate_mask = wl12xx_set_bitrate_mask,
5960 	.set_default_unicast_key = wl1271_op_set_default_key_idx,
5961 	.channel_switch = wl12xx_op_channel_switch,
5962 	.channel_switch_beacon = wlcore_op_channel_switch_beacon,
5963 	.flush = wlcore_op_flush,
5964 	.remain_on_channel = wlcore_op_remain_on_channel,
5965 	.cancel_remain_on_channel = wlcore_op_cancel_remain_on_channel,
5966 	.add_chanctx = wlcore_op_add_chanctx,
5967 	.remove_chanctx = wlcore_op_remove_chanctx,
5968 	.change_chanctx = wlcore_op_change_chanctx,
5969 	.assign_vif_chanctx = wlcore_op_assign_vif_chanctx,
5970 	.unassign_vif_chanctx = wlcore_op_unassign_vif_chanctx,
5971 	.switch_vif_chanctx = wlcore_op_switch_vif_chanctx,
5972 	.sta_rc_update = wlcore_op_sta_rc_update,
5973 	.sta_statistics = wlcore_op_sta_statistics,
5974 	.get_expected_throughput = wlcore_op_get_expected_throughput,
5975 	CFG80211_TESTMODE_CMD(wl1271_tm_cmd)
5976 };
5977 
5978 
5979 u8 wlcore_rate_to_idx(struct wl1271 *wl, u8 rate, enum nl80211_band band)
5980 {
5981 	u8 idx;
5982 
5983 	BUG_ON(band >= 2);
5984 
5985 	if (unlikely(rate >= wl->hw_tx_rate_tbl_size)) {
5986 		wl1271_error("Illegal RX rate from HW: %d", rate);
5987 		return 0;
5988 	}
5989 
5990 	idx = wl->band_rate_to_idx[band][rate];
5991 	if (unlikely(idx == CONF_HW_RXTX_RATE_UNSUPPORTED)) {
5992 		wl1271_error("Unsupported RX rate from HW: %d", rate);
5993 		return 0;
5994 	}
5995 
5996 	return idx;
5997 }
5998 
5999 static void wl12xx_derive_mac_addresses(struct wl1271 *wl, u32 oui, u32 nic)
6000 {
6001 	int i;
6002 
6003 	wl1271_debug(DEBUG_PROBE, "base address: oui %06x nic %06x",
6004 		     oui, nic);
6005 
6006 	if (nic + WLCORE_NUM_MAC_ADDRESSES - wl->num_mac_addr > 0xffffff)
6007 		wl1271_warning("NIC part of the MAC address wraps around!");
6008 
6009 	for (i = 0; i < wl->num_mac_addr; i++) {
6010 		wl->addresses[i].addr[0] = (u8)(oui >> 16);
6011 		wl->addresses[i].addr[1] = (u8)(oui >> 8);
6012 		wl->addresses[i].addr[2] = (u8) oui;
6013 		wl->addresses[i].addr[3] = (u8)(nic >> 16);
6014 		wl->addresses[i].addr[4] = (u8)(nic >> 8);
6015 		wl->addresses[i].addr[5] = (u8) nic;
6016 		nic++;
6017 	}
6018 
6019 	/* we may be one address short at the most */
6020 	WARN_ON(wl->num_mac_addr + 1 < WLCORE_NUM_MAC_ADDRESSES);
6021 
6022 	/*
6023 	 * turn on the LAA bit in the first address and use it as
6024 	 * the last address.
6025 	 */
6026 	if (wl->num_mac_addr < WLCORE_NUM_MAC_ADDRESSES) {
6027 		int idx = WLCORE_NUM_MAC_ADDRESSES - 1;
6028 		memcpy(&wl->addresses[idx], &wl->addresses[0],
6029 		       sizeof(wl->addresses[0]));
6030 		/* LAA bit */
6031 		wl->addresses[idx].addr[0] |= BIT(1);
6032 	}
6033 
6034 	wl->hw->wiphy->n_addresses = WLCORE_NUM_MAC_ADDRESSES;
6035 	wl->hw->wiphy->addresses = wl->addresses;
6036 }
6037 
6038 static int wl12xx_get_hw_info(struct wl1271 *wl)
6039 {
6040 	int ret;
6041 
6042 	ret = wlcore_read_reg(wl, REG_CHIP_ID_B, &wl->chip.id);
6043 	if (ret < 0)
6044 		goto out;
6045 
6046 	wl->fuse_oui_addr = 0;
6047 	wl->fuse_nic_addr = 0;
6048 
6049 	ret = wl->ops->get_pg_ver(wl, &wl->hw_pg_ver);
6050 	if (ret < 0)
6051 		goto out;
6052 
6053 	if (wl->ops->get_mac)
6054 		ret = wl->ops->get_mac(wl);
6055 
6056 out:
6057 	return ret;
6058 }
6059 
6060 static int wl1271_register_hw(struct wl1271 *wl)
6061 {
6062 	int ret;
6063 	u32 oui_addr = 0, nic_addr = 0;
6064 	struct platform_device *pdev = wl->pdev;
6065 	struct wlcore_platdev_data *pdev_data = dev_get_platdata(&pdev->dev);
6066 
6067 	if (wl->mac80211_registered)
6068 		return 0;
6069 
6070 	if (wl->nvs_len >= 12) {
6071 		/* NOTE: The wl->nvs->nvs element must be first, in
6072 		 * order to simplify the casting, we assume it is at
6073 		 * the beginning of the wl->nvs structure.
6074 		 */
6075 		u8 *nvs_ptr = (u8 *)wl->nvs;
6076 
6077 		oui_addr =
6078 			(nvs_ptr[11] << 16) + (nvs_ptr[10] << 8) + nvs_ptr[6];
6079 		nic_addr =
6080 			(nvs_ptr[5] << 16) + (nvs_ptr[4] << 8) + nvs_ptr[3];
6081 	}
6082 
6083 	/* if the MAC address is zeroed in the NVS derive from fuse */
6084 	if (oui_addr == 0 && nic_addr == 0) {
6085 		oui_addr = wl->fuse_oui_addr;
6086 		/* fuse has the BD_ADDR, the WLAN addresses are the next two */
6087 		nic_addr = wl->fuse_nic_addr + 1;
6088 	}
6089 
6090 	if (oui_addr == 0xdeadbe && nic_addr == 0xef0000) {
6091 		wl1271_warning("Detected unconfigured mac address in nvs, derive from fuse instead.");
6092 		if (!strcmp(pdev_data->family->name, "wl18xx")) {
6093 			wl1271_warning("This default nvs file can be removed from the file system");
6094 		} else {
6095 			wl1271_warning("Your device performance is not optimized.");
6096 			wl1271_warning("Please use the calibrator tool to configure your device.");
6097 		}
6098 
6099 		if (wl->fuse_oui_addr == 0 && wl->fuse_nic_addr == 0) {
6100 			wl1271_warning("Fuse mac address is zero. using random mac");
6101 			/* Use TI oui and a random nic */
6102 			oui_addr = WLCORE_TI_OUI_ADDRESS;
6103 			nic_addr = get_random_int();
6104 		} else {
6105 			oui_addr = wl->fuse_oui_addr;
6106 			/* fuse has the BD_ADDR, the WLAN addresses are the next two */
6107 			nic_addr = wl->fuse_nic_addr + 1;
6108 		}
6109 	}
6110 
6111 	wl12xx_derive_mac_addresses(wl, oui_addr, nic_addr);
6112 
6113 	ret = ieee80211_register_hw(wl->hw);
6114 	if (ret < 0) {
6115 		wl1271_error("unable to register mac80211 hw: %d", ret);
6116 		goto out;
6117 	}
6118 
6119 	wl->mac80211_registered = true;
6120 
6121 	wl1271_debugfs_init(wl);
6122 
6123 	wl1271_notice("loaded");
6124 
6125 out:
6126 	return ret;
6127 }
6128 
6129 static void wl1271_unregister_hw(struct wl1271 *wl)
6130 {
6131 	if (wl->plt)
6132 		wl1271_plt_stop(wl);
6133 
6134 	ieee80211_unregister_hw(wl->hw);
6135 	wl->mac80211_registered = false;
6136 
6137 }
6138 
6139 static int wl1271_init_ieee80211(struct wl1271 *wl)
6140 {
6141 	int i;
6142 	static const u32 cipher_suites[] = {
6143 		WLAN_CIPHER_SUITE_WEP40,
6144 		WLAN_CIPHER_SUITE_WEP104,
6145 		WLAN_CIPHER_SUITE_TKIP,
6146 		WLAN_CIPHER_SUITE_CCMP,
6147 		WL1271_CIPHER_SUITE_GEM,
6148 	};
6149 
6150 	/* The tx descriptor buffer */
6151 	wl->hw->extra_tx_headroom = sizeof(struct wl1271_tx_hw_descr);
6152 
6153 	if (wl->quirks & WLCORE_QUIRK_TKIP_HEADER_SPACE)
6154 		wl->hw->extra_tx_headroom += WL1271_EXTRA_SPACE_TKIP;
6155 
6156 	/* unit us */
6157 	/* FIXME: find a proper value */
6158 	wl->hw->max_listen_interval = wl->conf.conn.max_listen_interval;
6159 
6160 	ieee80211_hw_set(wl->hw, SUPPORT_FAST_XMIT);
6161 	ieee80211_hw_set(wl->hw, CHANCTX_STA_CSA);
6162 	ieee80211_hw_set(wl->hw, SUPPORTS_PER_STA_GTK);
6163 	ieee80211_hw_set(wl->hw, QUEUE_CONTROL);
6164 	ieee80211_hw_set(wl->hw, TX_AMPDU_SETUP_IN_HW);
6165 	ieee80211_hw_set(wl->hw, AMPDU_AGGREGATION);
6166 	ieee80211_hw_set(wl->hw, AP_LINK_PS);
6167 	ieee80211_hw_set(wl->hw, SPECTRUM_MGMT);
6168 	ieee80211_hw_set(wl->hw, REPORTS_TX_ACK_STATUS);
6169 	ieee80211_hw_set(wl->hw, CONNECTION_MONITOR);
6170 	ieee80211_hw_set(wl->hw, HAS_RATE_CONTROL);
6171 	ieee80211_hw_set(wl->hw, SUPPORTS_DYNAMIC_PS);
6172 	ieee80211_hw_set(wl->hw, SIGNAL_DBM);
6173 	ieee80211_hw_set(wl->hw, SUPPORTS_PS);
6174 	ieee80211_hw_set(wl->hw, SUPPORTS_TX_FRAG);
6175 
6176 	wl->hw->wiphy->cipher_suites = cipher_suites;
6177 	wl->hw->wiphy->n_cipher_suites = ARRAY_SIZE(cipher_suites);
6178 
6179 	wl->hw->wiphy->interface_modes = BIT(NL80211_IFTYPE_STATION) |
6180 					 BIT(NL80211_IFTYPE_AP) |
6181 					 BIT(NL80211_IFTYPE_P2P_DEVICE) |
6182 					 BIT(NL80211_IFTYPE_P2P_CLIENT) |
6183 #ifdef CONFIG_MAC80211_MESH
6184 					 BIT(NL80211_IFTYPE_MESH_POINT) |
6185 #endif
6186 					 BIT(NL80211_IFTYPE_P2P_GO);
6187 
6188 	wl->hw->wiphy->max_scan_ssids = 1;
6189 	wl->hw->wiphy->max_sched_scan_ssids = 16;
6190 	wl->hw->wiphy->max_match_sets = 16;
6191 	/*
6192 	 * Maximum length of elements in scanning probe request templates
6193 	 * should be the maximum length possible for a template, without
6194 	 * the IEEE80211 header of the template
6195 	 */
6196 	wl->hw->wiphy->max_scan_ie_len = WL1271_CMD_TEMPL_MAX_SIZE -
6197 			sizeof(struct ieee80211_header);
6198 
6199 	wl->hw->wiphy->max_sched_scan_reqs = 1;
6200 	wl->hw->wiphy->max_sched_scan_ie_len = WL1271_CMD_TEMPL_MAX_SIZE -
6201 		sizeof(struct ieee80211_header);
6202 
6203 	wl->hw->wiphy->max_remain_on_channel_duration = 30000;
6204 
6205 	wl->hw->wiphy->flags |= WIPHY_FLAG_AP_UAPSD |
6206 				WIPHY_FLAG_HAS_REMAIN_ON_CHANNEL |
6207 				WIPHY_FLAG_HAS_CHANNEL_SWITCH |
6208 				WIPHY_FLAG_IBSS_RSN;
6209 
6210 	wl->hw->wiphy->features |= NL80211_FEATURE_AP_SCAN;
6211 
6212 	/* make sure all our channels fit in the scanned_ch bitmask */
6213 	BUILD_BUG_ON(ARRAY_SIZE(wl1271_channels) +
6214 		     ARRAY_SIZE(wl1271_channels_5ghz) >
6215 		     WL1271_MAX_CHANNELS);
6216 	/*
6217 	* clear channel flags from the previous usage
6218 	* and restore max_power & max_antenna_gain values.
6219 	*/
6220 	for (i = 0; i < ARRAY_SIZE(wl1271_channels); i++) {
6221 		wl1271_band_2ghz.channels[i].flags = 0;
6222 		wl1271_band_2ghz.channels[i].max_power = WLCORE_MAX_TXPWR;
6223 		wl1271_band_2ghz.channels[i].max_antenna_gain = 0;
6224 	}
6225 
6226 	for (i = 0; i < ARRAY_SIZE(wl1271_channels_5ghz); i++) {
6227 		wl1271_band_5ghz.channels[i].flags = 0;
6228 		wl1271_band_5ghz.channels[i].max_power = WLCORE_MAX_TXPWR;
6229 		wl1271_band_5ghz.channels[i].max_antenna_gain = 0;
6230 	}
6231 
6232 	/*
6233 	 * We keep local copies of the band structs because we need to
6234 	 * modify them on a per-device basis.
6235 	 */
6236 	memcpy(&wl->bands[NL80211_BAND_2GHZ], &wl1271_band_2ghz,
6237 	       sizeof(wl1271_band_2ghz));
6238 	memcpy(&wl->bands[NL80211_BAND_2GHZ].ht_cap,
6239 	       &wl->ht_cap[NL80211_BAND_2GHZ],
6240 	       sizeof(*wl->ht_cap));
6241 	memcpy(&wl->bands[NL80211_BAND_5GHZ], &wl1271_band_5ghz,
6242 	       sizeof(wl1271_band_5ghz));
6243 	memcpy(&wl->bands[NL80211_BAND_5GHZ].ht_cap,
6244 	       &wl->ht_cap[NL80211_BAND_5GHZ],
6245 	       sizeof(*wl->ht_cap));
6246 
6247 	wl->hw->wiphy->bands[NL80211_BAND_2GHZ] =
6248 		&wl->bands[NL80211_BAND_2GHZ];
6249 	wl->hw->wiphy->bands[NL80211_BAND_5GHZ] =
6250 		&wl->bands[NL80211_BAND_5GHZ];
6251 
6252 	/*
6253 	 * allow 4 queues per mac address we support +
6254 	 * 1 cab queue per mac + one global offchannel Tx queue
6255 	 */
6256 	wl->hw->queues = (NUM_TX_QUEUES + 1) * WLCORE_NUM_MAC_ADDRESSES + 1;
6257 
6258 	/* the last queue is the offchannel queue */
6259 	wl->hw->offchannel_tx_hw_queue = wl->hw->queues - 1;
6260 	wl->hw->max_rates = 1;
6261 
6262 	wl->hw->wiphy->reg_notifier = wl1271_reg_notify;
6263 
6264 	/* the FW answers probe-requests in AP-mode */
6265 	wl->hw->wiphy->flags |= WIPHY_FLAG_AP_PROBE_RESP_OFFLOAD;
6266 	wl->hw->wiphy->probe_resp_offload =
6267 		NL80211_PROBE_RESP_OFFLOAD_SUPPORT_WPS |
6268 		NL80211_PROBE_RESP_OFFLOAD_SUPPORT_WPS2 |
6269 		NL80211_PROBE_RESP_OFFLOAD_SUPPORT_P2P;
6270 
6271 	/* allowed interface combinations */
6272 	wl->hw->wiphy->iface_combinations = wl->iface_combinations;
6273 	wl->hw->wiphy->n_iface_combinations = wl->n_iface_combinations;
6274 
6275 	/* register vendor commands */
6276 	wlcore_set_vendor_commands(wl->hw->wiphy);
6277 
6278 	SET_IEEE80211_DEV(wl->hw, wl->dev);
6279 
6280 	wl->hw->sta_data_size = sizeof(struct wl1271_station);
6281 	wl->hw->vif_data_size = sizeof(struct wl12xx_vif);
6282 
6283 	wl->hw->max_rx_aggregation_subframes = wl->conf.ht.rx_ba_win_size;
6284 
6285 	return 0;
6286 }
6287 
6288 struct ieee80211_hw *wlcore_alloc_hw(size_t priv_size, u32 aggr_buf_size,
6289 				     u32 mbox_size)
6290 {
6291 	struct ieee80211_hw *hw;
6292 	struct wl1271 *wl;
6293 	int i, j, ret;
6294 	unsigned int order;
6295 
6296 	hw = ieee80211_alloc_hw(sizeof(*wl), &wl1271_ops);
6297 	if (!hw) {
6298 		wl1271_error("could not alloc ieee80211_hw");
6299 		ret = -ENOMEM;
6300 		goto err_hw_alloc;
6301 	}
6302 
6303 	wl = hw->priv;
6304 	memset(wl, 0, sizeof(*wl));
6305 
6306 	wl->priv = kzalloc(priv_size, GFP_KERNEL);
6307 	if (!wl->priv) {
6308 		wl1271_error("could not alloc wl priv");
6309 		ret = -ENOMEM;
6310 		goto err_priv_alloc;
6311 	}
6312 
6313 	INIT_LIST_HEAD(&wl->wlvif_list);
6314 
6315 	wl->hw = hw;
6316 
6317 	/*
6318 	 * wl->num_links is not configured yet, so just use WLCORE_MAX_LINKS.
6319 	 * we don't allocate any additional resource here, so that's fine.
6320 	 */
6321 	for (i = 0; i < NUM_TX_QUEUES; i++)
6322 		for (j = 0; j < WLCORE_MAX_LINKS; j++)
6323 			skb_queue_head_init(&wl->links[j].tx_queue[i]);
6324 
6325 	skb_queue_head_init(&wl->deferred_rx_queue);
6326 	skb_queue_head_init(&wl->deferred_tx_queue);
6327 
6328 	INIT_WORK(&wl->netstack_work, wl1271_netstack_work);
6329 	INIT_WORK(&wl->tx_work, wl1271_tx_work);
6330 	INIT_WORK(&wl->recovery_work, wl1271_recovery_work);
6331 	INIT_DELAYED_WORK(&wl->scan_complete_work, wl1271_scan_complete_work);
6332 	INIT_DELAYED_WORK(&wl->roc_complete_work, wlcore_roc_complete_work);
6333 	INIT_DELAYED_WORK(&wl->tx_watchdog_work, wl12xx_tx_watchdog_work);
6334 
6335 	wl->freezable_wq = create_freezable_workqueue("wl12xx_wq");
6336 	if (!wl->freezable_wq) {
6337 		ret = -ENOMEM;
6338 		goto err_hw;
6339 	}
6340 
6341 	wl->channel = 0;
6342 	wl->rx_counter = 0;
6343 	wl->power_level = WL1271_DEFAULT_POWER_LEVEL;
6344 	wl->band = NL80211_BAND_2GHZ;
6345 	wl->channel_type = NL80211_CHAN_NO_HT;
6346 	wl->flags = 0;
6347 	wl->sg_enabled = true;
6348 	wl->sleep_auth = WL1271_PSM_ILLEGAL;
6349 	wl->recovery_count = 0;
6350 	wl->hw_pg_ver = -1;
6351 	wl->ap_ps_map = 0;
6352 	wl->ap_fw_ps_map = 0;
6353 	wl->quirks = 0;
6354 	wl->system_hlid = WL12XX_SYSTEM_HLID;
6355 	wl->active_sta_count = 0;
6356 	wl->active_link_count = 0;
6357 	wl->fwlog_size = 0;
6358 
6359 	/* The system link is always allocated */
6360 	__set_bit(WL12XX_SYSTEM_HLID, wl->links_map);
6361 
6362 	memset(wl->tx_frames_map, 0, sizeof(wl->tx_frames_map));
6363 	for (i = 0; i < wl->num_tx_desc; i++)
6364 		wl->tx_frames[i] = NULL;
6365 
6366 	spin_lock_init(&wl->wl_lock);
6367 
6368 	wl->state = WLCORE_STATE_OFF;
6369 	wl->fw_type = WL12XX_FW_TYPE_NONE;
6370 	mutex_init(&wl->mutex);
6371 	mutex_init(&wl->flush_mutex);
6372 	init_completion(&wl->nvs_loading_complete);
6373 
6374 	order = get_order(aggr_buf_size);
6375 	wl->aggr_buf = (u8 *)__get_free_pages(GFP_KERNEL, order);
6376 	if (!wl->aggr_buf) {
6377 		ret = -ENOMEM;
6378 		goto err_wq;
6379 	}
6380 	wl->aggr_buf_size = aggr_buf_size;
6381 
6382 	wl->dummy_packet = wl12xx_alloc_dummy_packet(wl);
6383 	if (!wl->dummy_packet) {
6384 		ret = -ENOMEM;
6385 		goto err_aggr;
6386 	}
6387 
6388 	/* Allocate one page for the FW log */
6389 	wl->fwlog = (u8 *)get_zeroed_page(GFP_KERNEL);
6390 	if (!wl->fwlog) {
6391 		ret = -ENOMEM;
6392 		goto err_dummy_packet;
6393 	}
6394 
6395 	wl->mbox_size = mbox_size;
6396 	wl->mbox = kmalloc(wl->mbox_size, GFP_KERNEL | GFP_DMA);
6397 	if (!wl->mbox) {
6398 		ret = -ENOMEM;
6399 		goto err_fwlog;
6400 	}
6401 
6402 	wl->buffer_32 = kmalloc(sizeof(*wl->buffer_32), GFP_KERNEL);
6403 	if (!wl->buffer_32) {
6404 		ret = -ENOMEM;
6405 		goto err_mbox;
6406 	}
6407 
6408 	return hw;
6409 
6410 err_mbox:
6411 	kfree(wl->mbox);
6412 
6413 err_fwlog:
6414 	free_page((unsigned long)wl->fwlog);
6415 
6416 err_dummy_packet:
6417 	dev_kfree_skb(wl->dummy_packet);
6418 
6419 err_aggr:
6420 	free_pages((unsigned long)wl->aggr_buf, order);
6421 
6422 err_wq:
6423 	destroy_workqueue(wl->freezable_wq);
6424 
6425 err_hw:
6426 	wl1271_debugfs_exit(wl);
6427 	kfree(wl->priv);
6428 
6429 err_priv_alloc:
6430 	ieee80211_free_hw(hw);
6431 
6432 err_hw_alloc:
6433 
6434 	return ERR_PTR(ret);
6435 }
6436 EXPORT_SYMBOL_GPL(wlcore_alloc_hw);
6437 
6438 int wlcore_free_hw(struct wl1271 *wl)
6439 {
6440 	/* Unblock any fwlog readers */
6441 	mutex_lock(&wl->mutex);
6442 	wl->fwlog_size = -1;
6443 	mutex_unlock(&wl->mutex);
6444 
6445 	wlcore_sysfs_free(wl);
6446 
6447 	kfree(wl->buffer_32);
6448 	kfree(wl->mbox);
6449 	free_page((unsigned long)wl->fwlog);
6450 	dev_kfree_skb(wl->dummy_packet);
6451 	free_pages((unsigned long)wl->aggr_buf, get_order(wl->aggr_buf_size));
6452 
6453 	wl1271_debugfs_exit(wl);
6454 
6455 	vfree(wl->fw);
6456 	wl->fw = NULL;
6457 	wl->fw_type = WL12XX_FW_TYPE_NONE;
6458 	kfree(wl->nvs);
6459 	wl->nvs = NULL;
6460 
6461 	kfree(wl->raw_fw_status);
6462 	kfree(wl->fw_status);
6463 	kfree(wl->tx_res_if);
6464 	destroy_workqueue(wl->freezable_wq);
6465 
6466 	kfree(wl->priv);
6467 	ieee80211_free_hw(wl->hw);
6468 
6469 	return 0;
6470 }
6471 EXPORT_SYMBOL_GPL(wlcore_free_hw);
6472 
6473 #ifdef CONFIG_PM
6474 static const struct wiphy_wowlan_support wlcore_wowlan_support = {
6475 	.flags = WIPHY_WOWLAN_ANY,
6476 	.n_patterns = WL1271_MAX_RX_FILTERS,
6477 	.pattern_min_len = 1,
6478 	.pattern_max_len = WL1271_RX_FILTER_MAX_PATTERN_SIZE,
6479 };
6480 #endif
6481 
6482 static irqreturn_t wlcore_hardirq(int irq, void *cookie)
6483 {
6484 	return IRQ_WAKE_THREAD;
6485 }
6486 
6487 static void wlcore_nvs_cb(const struct firmware *fw, void *context)
6488 {
6489 	struct wl1271 *wl = context;
6490 	struct platform_device *pdev = wl->pdev;
6491 	struct wlcore_platdev_data *pdev_data = dev_get_platdata(&pdev->dev);
6492 	struct resource *res;
6493 
6494 	int ret;
6495 	irq_handler_t hardirq_fn = NULL;
6496 
6497 	if (fw) {
6498 		wl->nvs = kmemdup(fw->data, fw->size, GFP_KERNEL);
6499 		if (!wl->nvs) {
6500 			wl1271_error("Could not allocate nvs data");
6501 			goto out;
6502 		}
6503 		wl->nvs_len = fw->size;
6504 	} else if (pdev_data->family->nvs_name) {
6505 		wl1271_debug(DEBUG_BOOT, "Could not get nvs file %s",
6506 			     pdev_data->family->nvs_name);
6507 		wl->nvs = NULL;
6508 		wl->nvs_len = 0;
6509 	} else {
6510 		wl->nvs = NULL;
6511 		wl->nvs_len = 0;
6512 	}
6513 
6514 	ret = wl->ops->setup(wl);
6515 	if (ret < 0)
6516 		goto out_free_nvs;
6517 
6518 	BUG_ON(wl->num_tx_desc > WLCORE_MAX_TX_DESCRIPTORS);
6519 
6520 	/* adjust some runtime configuration parameters */
6521 	wlcore_adjust_conf(wl);
6522 
6523 	res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
6524 	if (!res) {
6525 		wl1271_error("Could not get IRQ resource");
6526 		goto out_free_nvs;
6527 	}
6528 
6529 	wl->irq = res->start;
6530 	wl->irq_flags = res->flags & IRQF_TRIGGER_MASK;
6531 	wl->if_ops = pdev_data->if_ops;
6532 
6533 	if (wl->irq_flags & (IRQF_TRIGGER_RISING | IRQF_TRIGGER_FALLING))
6534 		hardirq_fn = wlcore_hardirq;
6535 	else
6536 		wl->irq_flags |= IRQF_ONESHOT;
6537 
6538 	ret = wl12xx_set_power_on(wl);
6539 	if (ret < 0)
6540 		goto out_free_nvs;
6541 
6542 	ret = wl12xx_get_hw_info(wl);
6543 	if (ret < 0) {
6544 		wl1271_error("couldn't get hw info");
6545 		wl1271_power_off(wl);
6546 		goto out_free_nvs;
6547 	}
6548 
6549 	ret = request_threaded_irq(wl->irq, hardirq_fn, wlcore_irq,
6550 				   wl->irq_flags, pdev->name, wl);
6551 	if (ret < 0) {
6552 		wl1271_error("interrupt configuration failed");
6553 		wl1271_power_off(wl);
6554 		goto out_free_nvs;
6555 	}
6556 
6557 #ifdef CONFIG_PM
6558 	device_init_wakeup(wl->dev, true);
6559 
6560 	ret = enable_irq_wake(wl->irq);
6561 	if (!ret) {
6562 		wl->irq_wake_enabled = true;
6563 		if (pdev_data->pwr_in_suspend)
6564 			wl->hw->wiphy->wowlan = &wlcore_wowlan_support;
6565 	}
6566 
6567 	res = platform_get_resource(pdev, IORESOURCE_IRQ, 1);
6568 	if (res) {
6569 		wl->wakeirq = res->start;
6570 		wl->wakeirq_flags = res->flags & IRQF_TRIGGER_MASK;
6571 		ret = dev_pm_set_dedicated_wake_irq(wl->dev, wl->wakeirq);
6572 		if (ret)
6573 			wl->wakeirq = -ENODEV;
6574 	} else {
6575 		wl->wakeirq = -ENODEV;
6576 	}
6577 #endif
6578 	disable_irq(wl->irq);
6579 	wl1271_power_off(wl);
6580 
6581 	ret = wl->ops->identify_chip(wl);
6582 	if (ret < 0)
6583 		goto out_irq;
6584 
6585 	ret = wl1271_init_ieee80211(wl);
6586 	if (ret)
6587 		goto out_irq;
6588 
6589 	ret = wl1271_register_hw(wl);
6590 	if (ret)
6591 		goto out_irq;
6592 
6593 	ret = wlcore_sysfs_init(wl);
6594 	if (ret)
6595 		goto out_unreg;
6596 
6597 	wl->initialized = true;
6598 	goto out;
6599 
6600 out_unreg:
6601 	wl1271_unregister_hw(wl);
6602 
6603 out_irq:
6604 	if (wl->wakeirq >= 0)
6605 		dev_pm_clear_wake_irq(wl->dev);
6606 	device_init_wakeup(wl->dev, false);
6607 	free_irq(wl->irq, wl);
6608 
6609 out_free_nvs:
6610 	kfree(wl->nvs);
6611 
6612 out:
6613 	release_firmware(fw);
6614 	complete_all(&wl->nvs_loading_complete);
6615 }
6616 
6617 static int __maybe_unused wlcore_runtime_suspend(struct device *dev)
6618 {
6619 	struct wl1271 *wl = dev_get_drvdata(dev);
6620 	struct wl12xx_vif *wlvif;
6621 	int error;
6622 
6623 	/* We do not enter elp sleep in PLT mode */
6624 	if (wl->plt)
6625 		return 0;
6626 
6627 	/* Nothing to do if no ELP mode requested */
6628 	if (wl->sleep_auth != WL1271_PSM_ELP)
6629 		return 0;
6630 
6631 	wl12xx_for_each_wlvif(wl, wlvif) {
6632 		if (!test_bit(WLVIF_FLAG_IN_PS, &wlvif->flags) &&
6633 		    test_bit(WLVIF_FLAG_IN_USE, &wlvif->flags))
6634 			return -EBUSY;
6635 	}
6636 
6637 	wl1271_debug(DEBUG_PSM, "chip to elp");
6638 	error = wlcore_raw_write32(wl, HW_ACCESS_ELP_CTRL_REG, ELPCTRL_SLEEP);
6639 	if (error < 0) {
6640 		wl12xx_queue_recovery_work(wl);
6641 
6642 		return error;
6643 	}
6644 
6645 	set_bit(WL1271_FLAG_IN_ELP, &wl->flags);
6646 
6647 	return 0;
6648 }
6649 
6650 static int __maybe_unused wlcore_runtime_resume(struct device *dev)
6651 {
6652 	struct wl1271 *wl = dev_get_drvdata(dev);
6653 	DECLARE_COMPLETION_ONSTACK(compl);
6654 	unsigned long flags;
6655 	int ret;
6656 	unsigned long start_time = jiffies;
6657 	bool recovery = false;
6658 
6659 	/* Nothing to do if no ELP mode requested */
6660 	if (!test_bit(WL1271_FLAG_IN_ELP, &wl->flags))
6661 		return 0;
6662 
6663 	wl1271_debug(DEBUG_PSM, "waking up chip from elp");
6664 
6665 	spin_lock_irqsave(&wl->wl_lock, flags);
6666 	wl->elp_compl = &compl;
6667 	spin_unlock_irqrestore(&wl->wl_lock, flags);
6668 
6669 	ret = wlcore_raw_write32(wl, HW_ACCESS_ELP_CTRL_REG, ELPCTRL_WAKE_UP);
6670 	if (ret < 0) {
6671 		recovery = true;
6672 	} else if (!test_bit(WL1271_FLAG_IRQ_RUNNING, &wl->flags)) {
6673 		ret = wait_for_completion_timeout(&compl,
6674 			msecs_to_jiffies(WL1271_WAKEUP_TIMEOUT));
6675 		if (ret == 0) {
6676 			wl1271_warning("ELP wakeup timeout!");
6677 			recovery = true;
6678 		}
6679 	}
6680 
6681 	spin_lock_irqsave(&wl->wl_lock, flags);
6682 	wl->elp_compl = NULL;
6683 	spin_unlock_irqrestore(&wl->wl_lock, flags);
6684 	clear_bit(WL1271_FLAG_IN_ELP, &wl->flags);
6685 
6686 	if (recovery) {
6687 		set_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags);
6688 		wl12xx_queue_recovery_work(wl);
6689 	} else {
6690 		wl1271_debug(DEBUG_PSM, "wakeup time: %u ms",
6691 			     jiffies_to_msecs(jiffies - start_time));
6692 	}
6693 
6694 	return 0;
6695 }
6696 
6697 static const struct dev_pm_ops wlcore_pm_ops = {
6698 	SET_RUNTIME_PM_OPS(wlcore_runtime_suspend,
6699 			   wlcore_runtime_resume,
6700 			   NULL)
6701 };
6702 
6703 int wlcore_probe(struct wl1271 *wl, struct platform_device *pdev)
6704 {
6705 	struct wlcore_platdev_data *pdev_data = dev_get_platdata(&pdev->dev);
6706 	const char *nvs_name;
6707 	int ret = 0;
6708 
6709 	if (!wl->ops || !wl->ptable || !pdev_data)
6710 		return -EINVAL;
6711 
6712 	wl->dev = &pdev->dev;
6713 	wl->pdev = pdev;
6714 	platform_set_drvdata(pdev, wl);
6715 
6716 	if (pdev_data->family && pdev_data->family->nvs_name) {
6717 		nvs_name = pdev_data->family->nvs_name;
6718 		ret = request_firmware_nowait(THIS_MODULE, FW_ACTION_UEVENT,
6719 					      nvs_name, &pdev->dev, GFP_KERNEL,
6720 					      wl, wlcore_nvs_cb);
6721 		if (ret < 0) {
6722 			wl1271_error("request_firmware_nowait failed for %s: %d",
6723 				     nvs_name, ret);
6724 			complete_all(&wl->nvs_loading_complete);
6725 		}
6726 	} else {
6727 		wlcore_nvs_cb(NULL, wl);
6728 	}
6729 
6730 	wl->dev->driver->pm = &wlcore_pm_ops;
6731 	pm_runtime_set_autosuspend_delay(wl->dev, 50);
6732 	pm_runtime_use_autosuspend(wl->dev);
6733 	pm_runtime_enable(wl->dev);
6734 
6735 	return ret;
6736 }
6737 EXPORT_SYMBOL_GPL(wlcore_probe);
6738 
6739 int wlcore_remove(struct platform_device *pdev)
6740 {
6741 	struct wlcore_platdev_data *pdev_data = dev_get_platdata(&pdev->dev);
6742 	struct wl1271 *wl = platform_get_drvdata(pdev);
6743 	int error;
6744 
6745 	error = pm_runtime_get_sync(wl->dev);
6746 	if (error < 0)
6747 		dev_warn(wl->dev, "PM runtime failed: %i\n", error);
6748 
6749 	wl->dev->driver->pm = NULL;
6750 
6751 	if (pdev_data->family && pdev_data->family->nvs_name)
6752 		wait_for_completion(&wl->nvs_loading_complete);
6753 	if (!wl->initialized)
6754 		return 0;
6755 
6756 	if (wl->wakeirq >= 0) {
6757 		dev_pm_clear_wake_irq(wl->dev);
6758 		wl->wakeirq = -ENODEV;
6759 	}
6760 
6761 	device_init_wakeup(wl->dev, false);
6762 
6763 	if (wl->irq_wake_enabled)
6764 		disable_irq_wake(wl->irq);
6765 
6766 	wl1271_unregister_hw(wl);
6767 
6768 	pm_runtime_put_sync(wl->dev);
6769 	pm_runtime_dont_use_autosuspend(wl->dev);
6770 	pm_runtime_disable(wl->dev);
6771 
6772 	free_irq(wl->irq, wl);
6773 	wlcore_free_hw(wl);
6774 
6775 	return 0;
6776 }
6777 EXPORT_SYMBOL_GPL(wlcore_remove);
6778 
6779 u32 wl12xx_debug_level = DEBUG_NONE;
6780 EXPORT_SYMBOL_GPL(wl12xx_debug_level);
6781 module_param_named(debug_level, wl12xx_debug_level, uint, 0600);
6782 MODULE_PARM_DESC(debug_level, "wl12xx debugging level");
6783 
6784 module_param_named(fwlog, fwlog_param, charp, 0);
6785 MODULE_PARM_DESC(fwlog,
6786 		 "FW logger options: continuous, dbgpins or disable");
6787 
6788 module_param(fwlog_mem_blocks, int, 0600);
6789 MODULE_PARM_DESC(fwlog_mem_blocks, "fwlog mem_blocks");
6790 
6791 module_param(bug_on_recovery, int, 0600);
6792 MODULE_PARM_DESC(bug_on_recovery, "BUG() on fw recovery");
6793 
6794 module_param(no_recovery, int, 0600);
6795 MODULE_PARM_DESC(no_recovery, "Prevent HW recovery. FW will remain stuck.");
6796 
6797 MODULE_LICENSE("GPL");
6798 MODULE_AUTHOR("Luciano Coelho <coelho@ti.com>");
6799 MODULE_AUTHOR("Juuso Oikarinen <juuso.oikarinen@nokia.com>");
6800