1 /* 2 * Copyright (c) 2012 Qualcomm Atheros, Inc. 3 * 4 * Permission to use, copy, modify, and/or distribute this software for any 5 * purpose with or without fee is hereby granted, provided that the above 6 * copyright notice and this permission notice appear in all copies. 7 * 8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES 9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF 10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR 11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES 12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 15 */ 16 17 #include <linux/etherdevice.h> 18 #include <linux/if_arp.h> 19 20 #include "wil6210.h" 21 #include "txrx.h" 22 #include "wmi.h" 23 #include "trace.h" 24 25 /** 26 * WMI event receiving - theory of operations 27 * 28 * When firmware about to report WMI event, it fills memory area 29 * in the mailbox and raises misc. IRQ. Thread interrupt handler invoked for 30 * the misc IRQ, function @wmi_recv_cmd called by thread IRQ handler. 31 * 32 * @wmi_recv_cmd reads event, allocates memory chunk and attaches it to the 33 * event list @wil->pending_wmi_ev. Then, work queue @wil->wmi_wq wakes up 34 * and handles events within the @wmi_event_worker. Every event get detached 35 * from list, processed and deleted. 36 * 37 * Purpose for this mechanism is to release IRQ thread; otherwise, 38 * if WMI event handling involves another WMI command flow, this 2-nd flow 39 * won't be completed because of blocked IRQ thread. 40 */ 41 42 /** 43 * Addressing - theory of operations 44 * 45 * There are several buses present on the WIL6210 card. 46 * Same memory areas are visible at different address on 47 * the different busses. There are 3 main bus masters: 48 * - MAC CPU (ucode) 49 * - User CPU (firmware) 50 * - AHB (host) 51 * 52 * On the PCI bus, there is one BAR (BAR0) of 2Mb size, exposing 53 * AHB addresses starting from 0x880000 54 * 55 * Internally, firmware uses addresses that allows faster access but 56 * are invisible from the host. To read from these addresses, alternative 57 * AHB address must be used. 58 * 59 * Memory mapping 60 * Linker address PCI/Host address 61 * 0x880000 .. 0xa80000 2Mb BAR0 62 * 0x800000 .. 0x807000 0x900000 .. 0x907000 28k DCCM 63 * 0x840000 .. 0x857000 0x908000 .. 0x91f000 92k PERIPH 64 */ 65 66 /** 67 * @fw_mapping provides memory remapping table 68 * 69 * array size should be in sync with the declaration in the wil6210.h 70 */ 71 const struct fw_map fw_mapping[] = { 72 {0x000000, 0x040000, 0x8c0000, "fw_code"}, /* FW code RAM 256k */ 73 {0x800000, 0x808000, 0x900000, "fw_data"}, /* FW data RAM 32k */ 74 {0x840000, 0x860000, 0x908000, "fw_peri"}, /* periph. data RAM 128k */ 75 {0x880000, 0x88a000, 0x880000, "rgf"}, /* various RGF 40k */ 76 {0x88a000, 0x88b000, 0x88a000, "AGC_tbl"}, /* AGC table 4k */ 77 {0x88b000, 0x88c000, 0x88b000, "rgf_ext"}, /* Pcie_ext_rgf 4k */ 78 {0x8c0000, 0x949000, 0x8c0000, "upper"}, /* upper area 548k */ 79 /* 80 * 920000..930000 ucode code RAM 81 * 930000..932000 ucode data RAM 82 * 932000..949000 back-door debug data 83 */ 84 }; 85 86 /** 87 * return AHB address for given firmware/ucode internal (linker) address 88 * @x - internal address 89 * If address have no valid AHB mapping, return 0 90 */ 91 static u32 wmi_addr_remap(u32 x) 92 { 93 uint i; 94 95 for (i = 0; i < ARRAY_SIZE(fw_mapping); i++) { 96 if ((x >= fw_mapping[i].from) && (x < fw_mapping[i].to)) 97 return x + fw_mapping[i].host - fw_mapping[i].from; 98 } 99 100 return 0; 101 } 102 103 /** 104 * Check address validity for WMI buffer; remap if needed 105 * @ptr - internal (linker) fw/ucode address 106 * 107 * Valid buffer should be DWORD aligned 108 * 109 * return address for accessing buffer from the host; 110 * if buffer is not valid, return NULL. 111 */ 112 void __iomem *wmi_buffer(struct wil6210_priv *wil, __le32 ptr_) 113 { 114 u32 off; 115 u32 ptr = le32_to_cpu(ptr_); 116 117 if (ptr % 4) 118 return NULL; 119 120 ptr = wmi_addr_remap(ptr); 121 if (ptr < WIL6210_FW_HOST_OFF) 122 return NULL; 123 124 off = HOSTADDR(ptr); 125 if (off > WIL6210_MEM_SIZE - 4) 126 return NULL; 127 128 return wil->csr + off; 129 } 130 131 /** 132 * Check address validity 133 */ 134 void __iomem *wmi_addr(struct wil6210_priv *wil, u32 ptr) 135 { 136 u32 off; 137 138 if (ptr % 4) 139 return NULL; 140 141 if (ptr < WIL6210_FW_HOST_OFF) 142 return NULL; 143 144 off = HOSTADDR(ptr); 145 if (off > WIL6210_MEM_SIZE - 4) 146 return NULL; 147 148 return wil->csr + off; 149 } 150 151 int wmi_read_hdr(struct wil6210_priv *wil, __le32 ptr, 152 struct wil6210_mbox_hdr *hdr) 153 { 154 void __iomem *src = wmi_buffer(wil, ptr); 155 if (!src) 156 return -EINVAL; 157 158 wil_memcpy_fromio_32(hdr, src, sizeof(*hdr)); 159 160 return 0; 161 } 162 163 static int __wmi_send(struct wil6210_priv *wil, u16 cmdid, void *buf, u16 len) 164 { 165 struct { 166 struct wil6210_mbox_hdr hdr; 167 struct wil6210_mbox_hdr_wmi wmi; 168 } __packed cmd = { 169 .hdr = { 170 .type = WIL_MBOX_HDR_TYPE_WMI, 171 .flags = 0, 172 .len = cpu_to_le16(sizeof(cmd.wmi) + len), 173 }, 174 .wmi = { 175 .mid = 0, 176 .id = cpu_to_le16(cmdid), 177 }, 178 }; 179 struct wil6210_mbox_ring *r = &wil->mbox_ctl.tx; 180 struct wil6210_mbox_ring_desc d_head; 181 u32 next_head; 182 void __iomem *dst; 183 void __iomem *head = wmi_addr(wil, r->head); 184 uint retry; 185 186 if (sizeof(cmd) + len > r->entry_size) { 187 wil_err(wil, "WMI size too large: %d bytes, max is %d\n", 188 (int)(sizeof(cmd) + len), r->entry_size); 189 return -ERANGE; 190 } 191 192 might_sleep(); 193 194 if (!test_bit(wil_status_fwready, &wil->status)) { 195 wil_err(wil, "WMI: cannot send command while FW not ready\n"); 196 return -EAGAIN; 197 } 198 199 if (!head) { 200 wil_err(wil, "WMI head is garbage: 0x%08x\n", r->head); 201 return -EINVAL; 202 } 203 /* read Tx head till it is not busy */ 204 for (retry = 5; retry > 0; retry--) { 205 wil_memcpy_fromio_32(&d_head, head, sizeof(d_head)); 206 if (d_head.sync == 0) 207 break; 208 msleep(20); 209 } 210 if (d_head.sync != 0) { 211 wil_err(wil, "WMI head busy\n"); 212 return -EBUSY; 213 } 214 /* next head */ 215 next_head = r->base + ((r->head - r->base + sizeof(d_head)) % r->size); 216 wil_dbg_wmi(wil, "Head 0x%08x -> 0x%08x\n", r->head, next_head); 217 /* wait till FW finish with previous command */ 218 for (retry = 5; retry > 0; retry--) { 219 r->tail = ioread32(wil->csr + HOST_MBOX + 220 offsetof(struct wil6210_mbox_ctl, tx.tail)); 221 if (next_head != r->tail) 222 break; 223 msleep(20); 224 } 225 if (next_head == r->tail) { 226 wil_err(wil, "WMI ring full\n"); 227 return -EBUSY; 228 } 229 dst = wmi_buffer(wil, d_head.addr); 230 if (!dst) { 231 wil_err(wil, "invalid WMI buffer: 0x%08x\n", 232 le32_to_cpu(d_head.addr)); 233 return -EINVAL; 234 } 235 cmd.hdr.seq = cpu_to_le16(++wil->wmi_seq); 236 /* set command */ 237 wil_dbg_wmi(wil, "WMI command 0x%04x [%d]\n", cmdid, len); 238 wil_hex_dump_wmi("Cmd ", DUMP_PREFIX_OFFSET, 16, 1, &cmd, 239 sizeof(cmd), true); 240 wil_hex_dump_wmi("cmd ", DUMP_PREFIX_OFFSET, 16, 1, buf, 241 len, true); 242 wil_memcpy_toio_32(dst, &cmd, sizeof(cmd)); 243 wil_memcpy_toio_32(dst + sizeof(cmd), buf, len); 244 /* mark entry as full */ 245 iowrite32(1, wil->csr + HOSTADDR(r->head) + 246 offsetof(struct wil6210_mbox_ring_desc, sync)); 247 /* advance next ptr */ 248 iowrite32(r->head = next_head, wil->csr + HOST_MBOX + 249 offsetof(struct wil6210_mbox_ctl, tx.head)); 250 251 trace_wil6210_wmi_cmd(&cmd.wmi, buf, len); 252 253 /* interrupt to FW */ 254 iowrite32(SW_INT_MBOX, wil->csr + HOST_SW_INT); 255 256 return 0; 257 } 258 259 int wmi_send(struct wil6210_priv *wil, u16 cmdid, void *buf, u16 len) 260 { 261 int rc; 262 263 mutex_lock(&wil->wmi_mutex); 264 rc = __wmi_send(wil, cmdid, buf, len); 265 mutex_unlock(&wil->wmi_mutex); 266 267 return rc; 268 } 269 270 /*=== Event handlers ===*/ 271 static void wmi_evt_ready(struct wil6210_priv *wil, int id, void *d, int len) 272 { 273 struct net_device *ndev = wil_to_ndev(wil); 274 struct wireless_dev *wdev = wil->wdev; 275 struct wmi_ready_event *evt = d; 276 wil->fw_version = le32_to_cpu(evt->sw_version); 277 wil->n_mids = evt->numof_additional_mids; 278 279 wil_info(wil, "FW ver. %d; MAC %pM; %d MID's\n", wil->fw_version, 280 evt->mac, wil->n_mids); 281 282 if (!is_valid_ether_addr(ndev->dev_addr)) { 283 memcpy(ndev->dev_addr, evt->mac, ETH_ALEN); 284 memcpy(ndev->perm_addr, evt->mac, ETH_ALEN); 285 } 286 snprintf(wdev->wiphy->fw_version, sizeof(wdev->wiphy->fw_version), 287 "%d", wil->fw_version); 288 } 289 290 static void wmi_evt_fw_ready(struct wil6210_priv *wil, int id, void *d, 291 int len) 292 { 293 wil_dbg_wmi(wil, "WMI: got FW ready event\n"); 294 295 set_bit(wil_status_fwready, &wil->status); 296 /* reuse wmi_ready for the firmware ready indication */ 297 complete(&wil->wmi_ready); 298 } 299 300 static void wmi_evt_rx_mgmt(struct wil6210_priv *wil, int id, void *d, int len) 301 { 302 struct wmi_rx_mgmt_packet_event *data = d; 303 struct wiphy *wiphy = wil_to_wiphy(wil); 304 struct ieee80211_mgmt *rx_mgmt_frame = 305 (struct ieee80211_mgmt *)data->payload; 306 int ch_no = data->info.channel+1; 307 u32 freq = ieee80211_channel_to_frequency(ch_no, 308 IEEE80211_BAND_60GHZ); 309 struct ieee80211_channel *channel = ieee80211_get_channel(wiphy, freq); 310 s32 signal = data->info.sqi; 311 __le16 fc = rx_mgmt_frame->frame_control; 312 u32 d_len = le32_to_cpu(data->info.len); 313 u16 d_status = le16_to_cpu(data->info.status); 314 315 wil_dbg_wmi(wil, "MGMT: channel %d MCS %d SNR %d SQI %d%%\n", 316 data->info.channel, data->info.mcs, data->info.snr, 317 data->info.sqi); 318 wil_dbg_wmi(wil, "status 0x%04x len %d fc 0x%04x\n", d_status, d_len, 319 le16_to_cpu(fc)); 320 wil_dbg_wmi(wil, "qid %d mid %d cid %d\n", 321 data->info.qid, data->info.mid, data->info.cid); 322 323 if (!channel) { 324 wil_err(wil, "Frame on unsupported channel\n"); 325 return; 326 } 327 328 if (ieee80211_is_beacon(fc) || ieee80211_is_probe_resp(fc)) { 329 struct cfg80211_bss *bss; 330 u64 tsf = le64_to_cpu(rx_mgmt_frame->u.beacon.timestamp); 331 u16 cap = le16_to_cpu(rx_mgmt_frame->u.beacon.capab_info); 332 u16 bi = le16_to_cpu(rx_mgmt_frame->u.beacon.beacon_int); 333 const u8 *ie_buf = rx_mgmt_frame->u.beacon.variable; 334 size_t ie_len = d_len - offsetof(struct ieee80211_mgmt, 335 u.beacon.variable); 336 wil_dbg_wmi(wil, "Capability info : 0x%04x\n", cap); 337 wil_dbg_wmi(wil, "TSF : 0x%016llx\n", tsf); 338 wil_dbg_wmi(wil, "Beacon interval : %d\n", bi); 339 wil_hex_dump_wmi("IE ", DUMP_PREFIX_OFFSET, 16, 1, ie_buf, 340 ie_len, true); 341 342 bss = cfg80211_inform_bss_frame(wiphy, channel, rx_mgmt_frame, 343 d_len, signal, GFP_KERNEL); 344 if (bss) { 345 wil_dbg_wmi(wil, "Added BSS %pM\n", 346 rx_mgmt_frame->bssid); 347 cfg80211_put_bss(wiphy, bss); 348 } else { 349 wil_err(wil, "cfg80211_inform_bss() failed\n"); 350 } 351 } else { 352 cfg80211_rx_mgmt(wil->wdev, freq, signal, 353 (void *)rx_mgmt_frame, d_len, 0, GFP_KERNEL); 354 } 355 } 356 357 static void wmi_evt_scan_complete(struct wil6210_priv *wil, int id, 358 void *d, int len) 359 { 360 if (wil->scan_request) { 361 struct wmi_scan_complete_event *data = d; 362 bool aborted = (data->status != WMI_SCAN_SUCCESS); 363 364 wil_dbg_wmi(wil, "SCAN_COMPLETE(0x%08x)\n", data->status); 365 wil_dbg_misc(wil, "Complete scan_request 0x%p aborted %d\n", 366 wil->scan_request, aborted); 367 368 del_timer_sync(&wil->scan_timer); 369 cfg80211_scan_done(wil->scan_request, aborted); 370 wil->scan_request = NULL; 371 } else { 372 wil_err(wil, "SCAN_COMPLETE while not scanning\n"); 373 } 374 } 375 376 static void wmi_evt_connect(struct wil6210_priv *wil, int id, void *d, int len) 377 { 378 struct net_device *ndev = wil_to_ndev(wil); 379 struct wireless_dev *wdev = wil->wdev; 380 struct wmi_connect_event *evt = d; 381 int ch; /* channel number */ 382 struct station_info sinfo; 383 u8 *assoc_req_ie, *assoc_resp_ie; 384 size_t assoc_req_ielen, assoc_resp_ielen; 385 /* capinfo(u16) + listen_interval(u16) + IEs */ 386 const size_t assoc_req_ie_offset = sizeof(u16) * 2; 387 /* capinfo(u16) + status_code(u16) + associd(u16) + IEs */ 388 const size_t assoc_resp_ie_offset = sizeof(u16) * 3; 389 390 if (len < sizeof(*evt)) { 391 wil_err(wil, "Connect event too short : %d bytes\n", len); 392 return; 393 } 394 if (len != sizeof(*evt) + evt->beacon_ie_len + evt->assoc_req_len + 395 evt->assoc_resp_len) { 396 wil_err(wil, 397 "Connect event corrupted : %d != %d + %d + %d + %d\n", 398 len, (int)sizeof(*evt), evt->beacon_ie_len, 399 evt->assoc_req_len, evt->assoc_resp_len); 400 return; 401 } 402 if (evt->cid >= WIL6210_MAX_CID) { 403 wil_err(wil, "Connect CID invalid : %d\n", evt->cid); 404 return; 405 } 406 407 ch = evt->channel + 1; 408 wil_dbg_wmi(wil, "Connect %pM channel [%d] cid %d\n", 409 evt->bssid, ch, evt->cid); 410 wil_hex_dump_wmi("connect AI : ", DUMP_PREFIX_OFFSET, 16, 1, 411 evt->assoc_info, len - sizeof(*evt), true); 412 413 /* figure out IE's */ 414 assoc_req_ie = &evt->assoc_info[evt->beacon_ie_len + 415 assoc_req_ie_offset]; 416 assoc_req_ielen = evt->assoc_req_len - assoc_req_ie_offset; 417 if (evt->assoc_req_len <= assoc_req_ie_offset) { 418 assoc_req_ie = NULL; 419 assoc_req_ielen = 0; 420 } 421 422 assoc_resp_ie = &evt->assoc_info[evt->beacon_ie_len + 423 evt->assoc_req_len + 424 assoc_resp_ie_offset]; 425 assoc_resp_ielen = evt->assoc_resp_len - assoc_resp_ie_offset; 426 if (evt->assoc_resp_len <= assoc_resp_ie_offset) { 427 assoc_resp_ie = NULL; 428 assoc_resp_ielen = 0; 429 } 430 431 if ((wdev->iftype == NL80211_IFTYPE_STATION) || 432 (wdev->iftype == NL80211_IFTYPE_P2P_CLIENT)) { 433 if (!test_bit(wil_status_fwconnecting, &wil->status)) { 434 wil_err(wil, "Not in connecting state\n"); 435 return; 436 } 437 del_timer_sync(&wil->connect_timer); 438 cfg80211_connect_result(ndev, evt->bssid, 439 assoc_req_ie, assoc_req_ielen, 440 assoc_resp_ie, assoc_resp_ielen, 441 WLAN_STATUS_SUCCESS, GFP_KERNEL); 442 443 } else if ((wdev->iftype == NL80211_IFTYPE_AP) || 444 (wdev->iftype == NL80211_IFTYPE_P2P_GO)) { 445 memset(&sinfo, 0, sizeof(sinfo)); 446 447 sinfo.generation = wil->sinfo_gen++; 448 449 if (assoc_req_ie) { 450 sinfo.assoc_req_ies = assoc_req_ie; 451 sinfo.assoc_req_ies_len = assoc_req_ielen; 452 sinfo.filled |= STATION_INFO_ASSOC_REQ_IES; 453 } 454 455 cfg80211_new_sta(ndev, evt->bssid, &sinfo, GFP_KERNEL); 456 } 457 clear_bit(wil_status_fwconnecting, &wil->status); 458 set_bit(wil_status_fwconnected, &wil->status); 459 460 /* FIXME FW can transmit only ucast frames to peer */ 461 /* FIXME real ring_id instead of hard coded 0 */ 462 memcpy(wil->sta[evt->cid].addr, evt->bssid, ETH_ALEN); 463 wil->sta[evt->cid].status = wil_sta_conn_pending; 464 465 wil->pending_connect_cid = evt->cid; 466 queue_work(wil->wmi_wq_conn, &wil->connect_worker); 467 } 468 469 static void wmi_evt_disconnect(struct wil6210_priv *wil, int id, 470 void *d, int len) 471 { 472 struct wmi_disconnect_event *evt = d; 473 474 wil_dbg_wmi(wil, "Disconnect %pM reason %d proto %d wmi\n", 475 evt->bssid, 476 evt->protocol_reason_status, evt->disconnect_reason); 477 478 wil->sinfo_gen++; 479 480 mutex_lock(&wil->mutex); 481 wil6210_disconnect(wil, evt->bssid); 482 mutex_unlock(&wil->mutex); 483 } 484 485 static void wmi_evt_notify(struct wil6210_priv *wil, int id, void *d, int len) 486 { 487 struct wmi_notify_req_done_event *evt = d; 488 489 if (len < sizeof(*evt)) { 490 wil_err(wil, "Short NOTIFY event\n"); 491 return; 492 } 493 494 wil->stats.tsf = le64_to_cpu(evt->tsf); 495 wil->stats.snr = le32_to_cpu(evt->snr_val); 496 wil->stats.bf_mcs = le16_to_cpu(evt->bf_mcs); 497 wil->stats.my_rx_sector = le16_to_cpu(evt->my_rx_sector); 498 wil->stats.my_tx_sector = le16_to_cpu(evt->my_tx_sector); 499 wil->stats.peer_rx_sector = le16_to_cpu(evt->other_rx_sector); 500 wil->stats.peer_tx_sector = le16_to_cpu(evt->other_tx_sector); 501 wil_dbg_wmi(wil, "Link status, MCS %d TSF 0x%016llx\n" 502 "BF status 0x%08x SNR 0x%08x SQI %d%%\n" 503 "Tx Tpt %d goodput %d Rx goodput %d\n" 504 "Sectors(rx:tx) my %d:%d peer %d:%d\n", 505 wil->stats.bf_mcs, wil->stats.tsf, evt->status, 506 wil->stats.snr, evt->sqi, le32_to_cpu(evt->tx_tpt), 507 le32_to_cpu(evt->tx_goodput), le32_to_cpu(evt->rx_goodput), 508 wil->stats.my_rx_sector, wil->stats.my_tx_sector, 509 wil->stats.peer_rx_sector, wil->stats.peer_tx_sector); 510 } 511 512 /* 513 * Firmware reports EAPOL frame using WME event. 514 * Reconstruct Ethernet frame and deliver it via normal Rx 515 */ 516 static void wmi_evt_eapol_rx(struct wil6210_priv *wil, int id, 517 void *d, int len) 518 { 519 struct net_device *ndev = wil_to_ndev(wil); 520 struct wmi_eapol_rx_event *evt = d; 521 u16 eapol_len = le16_to_cpu(evt->eapol_len); 522 int sz = eapol_len + ETH_HLEN; 523 struct sk_buff *skb; 524 struct ethhdr *eth; 525 int cid; 526 struct wil_net_stats *stats = NULL; 527 528 wil_dbg_wmi(wil, "EAPOL len %d from %pM\n", eapol_len, 529 evt->src_mac); 530 531 cid = wil_find_cid(wil, evt->src_mac); 532 if (cid >= 0) 533 stats = &wil->sta[cid].stats; 534 535 if (eapol_len > 196) { /* TODO: revisit size limit */ 536 wil_err(wil, "EAPOL too large\n"); 537 return; 538 } 539 540 skb = alloc_skb(sz, GFP_KERNEL); 541 if (!skb) { 542 wil_err(wil, "Failed to allocate skb\n"); 543 return; 544 } 545 546 eth = (struct ethhdr *)skb_put(skb, ETH_HLEN); 547 memcpy(eth->h_dest, ndev->dev_addr, ETH_ALEN); 548 memcpy(eth->h_source, evt->src_mac, ETH_ALEN); 549 eth->h_proto = cpu_to_be16(ETH_P_PAE); 550 memcpy(skb_put(skb, eapol_len), evt->eapol, eapol_len); 551 skb->protocol = eth_type_trans(skb, ndev); 552 if (likely(netif_rx_ni(skb) == NET_RX_SUCCESS)) { 553 ndev->stats.rx_packets++; 554 ndev->stats.rx_bytes += sz; 555 if (stats) { 556 stats->rx_packets++; 557 stats->rx_bytes += sz; 558 } 559 } else { 560 ndev->stats.rx_dropped++; 561 if (stats) 562 stats->rx_dropped++; 563 } 564 } 565 566 static void wmi_evt_linkup(struct wil6210_priv *wil, int id, void *d, int len) 567 { 568 struct net_device *ndev = wil_to_ndev(wil); 569 struct wmi_data_port_open_event *evt = d; 570 u8 cid = evt->cid; 571 572 wil_dbg_wmi(wil, "Link UP for CID %d\n", cid); 573 574 if (cid >= ARRAY_SIZE(wil->sta)) { 575 wil_err(wil, "Link UP for invalid CID %d\n", cid); 576 return; 577 } 578 579 wil->sta[cid].data_port_open = true; 580 netif_carrier_on(ndev); 581 } 582 583 static void wmi_evt_linkdown(struct wil6210_priv *wil, int id, void *d, int len) 584 { 585 struct net_device *ndev = wil_to_ndev(wil); 586 struct wmi_wbe_link_down_event *evt = d; 587 u8 cid = evt->cid; 588 589 wil_dbg_wmi(wil, "Link DOWN for CID %d, reason %d\n", 590 cid, le32_to_cpu(evt->reason)); 591 592 if (cid >= ARRAY_SIZE(wil->sta)) { 593 wil_err(wil, "Link DOWN for invalid CID %d\n", cid); 594 return; 595 } 596 597 wil->sta[cid].data_port_open = false; 598 netif_carrier_off(ndev); 599 } 600 601 static void wmi_evt_ba_status(struct wil6210_priv *wil, int id, void *d, 602 int len) 603 { 604 struct wmi_vring_ba_status_event *evt = d; 605 struct wil_sta_info *sta; 606 uint i, cid; 607 608 /* TODO: use Rx BA status, not Tx one */ 609 610 wil_dbg_wmi(wil, "BACK[%d] %s {%d} timeout %d\n", 611 evt->ringid, 612 evt->status == WMI_BA_AGREED ? "OK" : "N/A", 613 evt->agg_wsize, __le16_to_cpu(evt->ba_timeout)); 614 615 if (evt->ringid >= WIL6210_MAX_TX_RINGS) { 616 wil_err(wil, "invalid ring id %d\n", evt->ringid); 617 return; 618 } 619 620 cid = wil->vring2cid_tid[evt->ringid][0]; 621 if (cid >= WIL6210_MAX_CID) { 622 wil_err(wil, "invalid CID %d for vring %d\n", cid, evt->ringid); 623 return; 624 } 625 626 sta = &wil->sta[cid]; 627 if (sta->status == wil_sta_unused) { 628 wil_err(wil, "CID %d unused\n", cid); 629 return; 630 } 631 632 wil_dbg_wmi(wil, "BACK for CID %d %pM\n", cid, sta->addr); 633 for (i = 0; i < WIL_STA_TID_NUM; i++) { 634 struct wil_tid_ampdu_rx *r = sta->tid_rx[i]; 635 sta->tid_rx[i] = NULL; 636 wil_tid_ampdu_rx_free(wil, r); 637 if ((evt->status == WMI_BA_AGREED) && evt->agg_wsize) 638 sta->tid_rx[i] = wil_tid_ampdu_rx_alloc(wil, 639 evt->agg_wsize, 0); 640 } 641 } 642 643 static const struct { 644 int eventid; 645 void (*handler)(struct wil6210_priv *wil, int eventid, 646 void *data, int data_len); 647 } wmi_evt_handlers[] = { 648 {WMI_READY_EVENTID, wmi_evt_ready}, 649 {WMI_FW_READY_EVENTID, wmi_evt_fw_ready}, 650 {WMI_RX_MGMT_PACKET_EVENTID, wmi_evt_rx_mgmt}, 651 {WMI_SCAN_COMPLETE_EVENTID, wmi_evt_scan_complete}, 652 {WMI_CONNECT_EVENTID, wmi_evt_connect}, 653 {WMI_DISCONNECT_EVENTID, wmi_evt_disconnect}, 654 {WMI_NOTIFY_REQ_DONE_EVENTID, wmi_evt_notify}, 655 {WMI_EAPOL_RX_EVENTID, wmi_evt_eapol_rx}, 656 {WMI_DATA_PORT_OPEN_EVENTID, wmi_evt_linkup}, 657 {WMI_WBE_LINKDOWN_EVENTID, wmi_evt_linkdown}, 658 {WMI_BA_STATUS_EVENTID, wmi_evt_ba_status}, 659 }; 660 661 /* 662 * Run in IRQ context 663 * Extract WMI command from mailbox. Queue it to the @wil->pending_wmi_ev 664 * that will be eventually handled by the @wmi_event_worker in the thread 665 * context of thread "wil6210_wmi" 666 */ 667 void wmi_recv_cmd(struct wil6210_priv *wil) 668 { 669 struct wil6210_mbox_ring_desc d_tail; 670 struct wil6210_mbox_hdr hdr; 671 struct wil6210_mbox_ring *r = &wil->mbox_ctl.rx; 672 struct pending_wmi_event *evt; 673 u8 *cmd; 674 void __iomem *src; 675 ulong flags; 676 unsigned n; 677 678 if (!test_bit(wil_status_reset_done, &wil->status)) { 679 wil_err(wil, "Reset not completed\n"); 680 return; 681 } 682 683 for (n = 0;; n++) { 684 u16 len; 685 bool q; 686 687 r->head = ioread32(wil->csr + HOST_MBOX + 688 offsetof(struct wil6210_mbox_ctl, rx.head)); 689 if (r->tail == r->head) 690 break; 691 692 wil_dbg_wmi(wil, "Mbox head %08x tail %08x\n", 693 r->head, r->tail); 694 /* read cmd descriptor from tail */ 695 wil_memcpy_fromio_32(&d_tail, wil->csr + HOSTADDR(r->tail), 696 sizeof(struct wil6210_mbox_ring_desc)); 697 if (d_tail.sync == 0) { 698 wil_err(wil, "Mbox evt not owned by FW?\n"); 699 break; 700 } 701 702 /* read cmd header from descriptor */ 703 if (0 != wmi_read_hdr(wil, d_tail.addr, &hdr)) { 704 wil_err(wil, "Mbox evt at 0x%08x?\n", 705 le32_to_cpu(d_tail.addr)); 706 break; 707 } 708 len = le16_to_cpu(hdr.len); 709 wil_dbg_wmi(wil, "Mbox evt %04x %04x %04x %02x\n", 710 le16_to_cpu(hdr.seq), len, le16_to_cpu(hdr.type), 711 hdr.flags); 712 713 /* read cmd buffer from descriptor */ 714 src = wmi_buffer(wil, d_tail.addr) + 715 sizeof(struct wil6210_mbox_hdr); 716 evt = kmalloc(ALIGN(offsetof(struct pending_wmi_event, 717 event.wmi) + len, 4), 718 GFP_KERNEL); 719 if (!evt) 720 break; 721 722 evt->event.hdr = hdr; 723 cmd = (void *)&evt->event.wmi; 724 wil_memcpy_fromio_32(cmd, src, len); 725 /* mark entry as empty */ 726 iowrite32(0, wil->csr + HOSTADDR(r->tail) + 727 offsetof(struct wil6210_mbox_ring_desc, sync)); 728 /* indicate */ 729 if ((hdr.type == WIL_MBOX_HDR_TYPE_WMI) && 730 (len >= sizeof(struct wil6210_mbox_hdr_wmi))) { 731 struct wil6210_mbox_hdr_wmi *wmi = &evt->event.wmi; 732 u16 id = le16_to_cpu(wmi->id); 733 u32 tstamp = le32_to_cpu(wmi->timestamp); 734 wil_dbg_wmi(wil, "WMI event 0x%04x MID %d @%d msec\n", 735 id, wmi->mid, tstamp); 736 trace_wil6210_wmi_event(wmi, &wmi[1], 737 len - sizeof(*wmi)); 738 } 739 wil_hex_dump_wmi("evt ", DUMP_PREFIX_OFFSET, 16, 1, 740 &evt->event.hdr, sizeof(hdr) + len, true); 741 742 /* advance tail */ 743 r->tail = r->base + ((r->tail - r->base + 744 sizeof(struct wil6210_mbox_ring_desc)) % r->size); 745 iowrite32(r->tail, wil->csr + HOST_MBOX + 746 offsetof(struct wil6210_mbox_ctl, rx.tail)); 747 748 /* add to the pending list */ 749 spin_lock_irqsave(&wil->wmi_ev_lock, flags); 750 list_add_tail(&evt->list, &wil->pending_wmi_ev); 751 spin_unlock_irqrestore(&wil->wmi_ev_lock, flags); 752 q = queue_work(wil->wmi_wq, &wil->wmi_event_worker); 753 wil_dbg_wmi(wil, "queue_work -> %d\n", q); 754 } 755 /* normally, 1 event per IRQ should be processed */ 756 wil_dbg_wmi(wil, "%s -> %d events queued\n", __func__, n); 757 } 758 759 int wmi_call(struct wil6210_priv *wil, u16 cmdid, void *buf, u16 len, 760 u16 reply_id, void *reply, u8 reply_size, int to_msec) 761 { 762 int rc; 763 int remain; 764 765 mutex_lock(&wil->wmi_mutex); 766 767 rc = __wmi_send(wil, cmdid, buf, len); 768 if (rc) 769 goto out; 770 771 wil->reply_id = reply_id; 772 wil->reply_buf = reply; 773 wil->reply_size = reply_size; 774 remain = wait_for_completion_timeout(&wil->wmi_ready, 775 msecs_to_jiffies(to_msec)); 776 if (0 == remain) { 777 wil_err(wil, "wmi_call(0x%04x->0x%04x) timeout %d msec\n", 778 cmdid, reply_id, to_msec); 779 rc = -ETIME; 780 } else { 781 wil_dbg_wmi(wil, 782 "wmi_call(0x%04x->0x%04x) completed in %d msec\n", 783 cmdid, reply_id, 784 to_msec - jiffies_to_msecs(remain)); 785 } 786 wil->reply_id = 0; 787 wil->reply_buf = NULL; 788 wil->reply_size = 0; 789 out: 790 mutex_unlock(&wil->wmi_mutex); 791 792 return rc; 793 } 794 795 int wmi_echo(struct wil6210_priv *wil) 796 { 797 struct wmi_echo_cmd cmd = { 798 .value = cpu_to_le32(0x12345678), 799 }; 800 801 return wmi_call(wil, WMI_ECHO_CMDID, &cmd, sizeof(cmd), 802 WMI_ECHO_RSP_EVENTID, NULL, 0, 20); 803 } 804 805 int wmi_set_mac_address(struct wil6210_priv *wil, void *addr) 806 { 807 struct wmi_set_mac_address_cmd cmd; 808 809 memcpy(cmd.mac, addr, ETH_ALEN); 810 811 wil_dbg_wmi(wil, "Set MAC %pM\n", addr); 812 813 return wmi_send(wil, WMI_SET_MAC_ADDRESS_CMDID, &cmd, sizeof(cmd)); 814 } 815 816 int wmi_pcp_start(struct wil6210_priv *wil, int bi, u8 wmi_nettype, u8 chan) 817 { 818 int rc; 819 820 struct wmi_pcp_start_cmd cmd = { 821 .bcon_interval = cpu_to_le16(bi), 822 .network_type = wmi_nettype, 823 .disable_sec_offload = 1, 824 .channel = chan - 1, 825 .pcp_max_assoc_sta = WIL6210_MAX_CID, 826 }; 827 struct { 828 struct wil6210_mbox_hdr_wmi wmi; 829 struct wmi_pcp_started_event evt; 830 } __packed reply; 831 832 if (!wil->secure_pcp) 833 cmd.disable_sec = 1; 834 835 /* 836 * Processing time may be huge, in case of secure AP it takes about 837 * 3500ms for FW to start AP 838 */ 839 rc = wmi_call(wil, WMI_PCP_START_CMDID, &cmd, sizeof(cmd), 840 WMI_PCP_STARTED_EVENTID, &reply, sizeof(reply), 5000); 841 if (rc) 842 return rc; 843 844 if (reply.evt.status != WMI_FW_STATUS_SUCCESS) 845 rc = -EINVAL; 846 847 return rc; 848 } 849 850 int wmi_pcp_stop(struct wil6210_priv *wil) 851 { 852 return wmi_call(wil, WMI_PCP_STOP_CMDID, NULL, 0, 853 WMI_PCP_STOPPED_EVENTID, NULL, 0, 20); 854 } 855 856 int wmi_set_ssid(struct wil6210_priv *wil, u8 ssid_len, const void *ssid) 857 { 858 struct wmi_set_ssid_cmd cmd = { 859 .ssid_len = cpu_to_le32(ssid_len), 860 }; 861 862 if (ssid_len > sizeof(cmd.ssid)) 863 return -EINVAL; 864 865 memcpy(cmd.ssid, ssid, ssid_len); 866 867 return wmi_send(wil, WMI_SET_SSID_CMDID, &cmd, sizeof(cmd)); 868 } 869 870 int wmi_get_ssid(struct wil6210_priv *wil, u8 *ssid_len, void *ssid) 871 { 872 int rc; 873 struct { 874 struct wil6210_mbox_hdr_wmi wmi; 875 struct wmi_set_ssid_cmd cmd; 876 } __packed reply; 877 int len; /* reply.cmd.ssid_len in CPU order */ 878 879 rc = wmi_call(wil, WMI_GET_SSID_CMDID, NULL, 0, WMI_GET_SSID_EVENTID, 880 &reply, sizeof(reply), 20); 881 if (rc) 882 return rc; 883 884 len = le32_to_cpu(reply.cmd.ssid_len); 885 if (len > sizeof(reply.cmd.ssid)) 886 return -EINVAL; 887 888 *ssid_len = len; 889 memcpy(ssid, reply.cmd.ssid, len); 890 891 return 0; 892 } 893 894 int wmi_set_channel(struct wil6210_priv *wil, int channel) 895 { 896 struct wmi_set_pcp_channel_cmd cmd = { 897 .channel = channel - 1, 898 }; 899 900 return wmi_send(wil, WMI_SET_PCP_CHANNEL_CMDID, &cmd, sizeof(cmd)); 901 } 902 903 int wmi_get_channel(struct wil6210_priv *wil, int *channel) 904 { 905 int rc; 906 struct { 907 struct wil6210_mbox_hdr_wmi wmi; 908 struct wmi_set_pcp_channel_cmd cmd; 909 } __packed reply; 910 911 rc = wmi_call(wil, WMI_GET_PCP_CHANNEL_CMDID, NULL, 0, 912 WMI_GET_PCP_CHANNEL_EVENTID, &reply, sizeof(reply), 20); 913 if (rc) 914 return rc; 915 916 if (reply.cmd.channel > 3) 917 return -EINVAL; 918 919 *channel = reply.cmd.channel + 1; 920 921 return 0; 922 } 923 924 int wmi_p2p_cfg(struct wil6210_priv *wil, int channel) 925 { 926 struct wmi_p2p_cfg_cmd cmd = { 927 .discovery_mode = WMI_DISCOVERY_MODE_NON_OFFLOAD, 928 .channel = channel - 1, 929 }; 930 931 return wmi_send(wil, WMI_P2P_CFG_CMDID, &cmd, sizeof(cmd)); 932 } 933 934 int wmi_del_cipher_key(struct wil6210_priv *wil, u8 key_index, 935 const void *mac_addr) 936 { 937 struct wmi_delete_cipher_key_cmd cmd = { 938 .key_index = key_index, 939 }; 940 941 if (mac_addr) 942 memcpy(cmd.mac, mac_addr, WMI_MAC_LEN); 943 944 return wmi_send(wil, WMI_DELETE_CIPHER_KEY_CMDID, &cmd, sizeof(cmd)); 945 } 946 947 int wmi_add_cipher_key(struct wil6210_priv *wil, u8 key_index, 948 const void *mac_addr, int key_len, const void *key) 949 { 950 struct wmi_add_cipher_key_cmd cmd = { 951 .key_index = key_index, 952 .key_usage = WMI_KEY_USE_PAIRWISE, 953 .key_len = key_len, 954 }; 955 956 if (!key || (key_len > sizeof(cmd.key))) 957 return -EINVAL; 958 959 memcpy(cmd.key, key, key_len); 960 if (mac_addr) 961 memcpy(cmd.mac, mac_addr, WMI_MAC_LEN); 962 963 return wmi_send(wil, WMI_ADD_CIPHER_KEY_CMDID, &cmd, sizeof(cmd)); 964 } 965 966 int wmi_set_ie(struct wil6210_priv *wil, u8 type, u16 ie_len, const void *ie) 967 { 968 int rc; 969 u16 len = sizeof(struct wmi_set_appie_cmd) + ie_len; 970 struct wmi_set_appie_cmd *cmd = kzalloc(len, GFP_KERNEL); 971 if (!cmd) 972 return -ENOMEM; 973 974 cmd->mgmt_frm_type = type; 975 /* BUG: FW API define ieLen as u8. Will fix FW */ 976 cmd->ie_len = cpu_to_le16(ie_len); 977 memcpy(cmd->ie_info, ie, ie_len); 978 rc = wmi_send(wil, WMI_SET_APPIE_CMDID, cmd, len); 979 kfree(cmd); 980 981 return rc; 982 } 983 984 /** 985 * wmi_rxon - turn radio on/off 986 * @on: turn on if true, off otherwise 987 * 988 * Only switch radio. Channel should be set separately. 989 * No timeout for rxon - radio turned on forever unless some other call 990 * turns it off 991 */ 992 int wmi_rxon(struct wil6210_priv *wil, bool on) 993 { 994 int rc; 995 struct { 996 struct wil6210_mbox_hdr_wmi wmi; 997 struct wmi_listen_started_event evt; 998 } __packed reply; 999 1000 wil_info(wil, "%s(%s)\n", __func__, on ? "on" : "off"); 1001 1002 if (on) { 1003 rc = wmi_call(wil, WMI_START_LISTEN_CMDID, NULL, 0, 1004 WMI_LISTEN_STARTED_EVENTID, 1005 &reply, sizeof(reply), 100); 1006 if ((rc == 0) && (reply.evt.status != WMI_FW_STATUS_SUCCESS)) 1007 rc = -EINVAL; 1008 } else { 1009 rc = wmi_call(wil, WMI_DISCOVERY_STOP_CMDID, NULL, 0, 1010 WMI_DISCOVERY_STOPPED_EVENTID, NULL, 0, 20); 1011 } 1012 1013 return rc; 1014 } 1015 1016 int wmi_rx_chain_add(struct wil6210_priv *wil, struct vring *vring) 1017 { 1018 struct wireless_dev *wdev = wil->wdev; 1019 struct net_device *ndev = wil_to_ndev(wil); 1020 struct wmi_cfg_rx_chain_cmd cmd = { 1021 .action = WMI_RX_CHAIN_ADD, 1022 .rx_sw_ring = { 1023 .max_mpdu_size = cpu_to_le16(RX_BUF_LEN), 1024 .ring_mem_base = cpu_to_le64(vring->pa), 1025 .ring_size = cpu_to_le16(vring->size), 1026 }, 1027 .mid = 0, /* TODO - what is it? */ 1028 .decap_trans_type = WMI_DECAP_TYPE_802_3, 1029 .reorder_type = WMI_RX_SW_REORDER, 1030 }; 1031 struct { 1032 struct wil6210_mbox_hdr_wmi wmi; 1033 struct wmi_cfg_rx_chain_done_event evt; 1034 } __packed evt; 1035 int rc; 1036 1037 if (wdev->iftype == NL80211_IFTYPE_MONITOR) { 1038 struct ieee80211_channel *ch = wdev->preset_chandef.chan; 1039 1040 cmd.sniffer_cfg.mode = cpu_to_le32(WMI_SNIFFER_ON); 1041 if (ch) 1042 cmd.sniffer_cfg.channel = ch->hw_value - 1; 1043 cmd.sniffer_cfg.phy_info_mode = 1044 cpu_to_le32(ndev->type == ARPHRD_IEEE80211_RADIOTAP); 1045 cmd.sniffer_cfg.phy_support = 1046 cpu_to_le32((wil->monitor_flags & MONITOR_FLAG_CONTROL) 1047 ? WMI_SNIFFER_CP : WMI_SNIFFER_DP); 1048 } else { 1049 /* Initialize offload (in non-sniffer mode). 1050 * Linux IP stack always calculates IP checksum 1051 * HW always calculate TCP/UDP checksum 1052 */ 1053 cmd.l3_l4_ctrl |= (1 << L3_L4_CTRL_TCPIP_CHECKSUM_EN_POS); 1054 } 1055 /* typical time for secure PCP is 840ms */ 1056 rc = wmi_call(wil, WMI_CFG_RX_CHAIN_CMDID, &cmd, sizeof(cmd), 1057 WMI_CFG_RX_CHAIN_DONE_EVENTID, &evt, sizeof(evt), 2000); 1058 if (rc) 1059 return rc; 1060 1061 vring->hwtail = le32_to_cpu(evt.evt.rx_ring_tail_ptr); 1062 1063 wil_dbg_misc(wil, "Rx init: status %d tail 0x%08x\n", 1064 le32_to_cpu(evt.evt.status), vring->hwtail); 1065 1066 if (le32_to_cpu(evt.evt.status) != WMI_CFG_RX_CHAIN_SUCCESS) 1067 rc = -EINVAL; 1068 1069 return rc; 1070 } 1071 1072 int wmi_get_temperature(struct wil6210_priv *wil, u32 *t_m, u32 *t_r) 1073 { 1074 int rc; 1075 struct wmi_temp_sense_cmd cmd = { 1076 .measure_marlon_m_en = cpu_to_le32(!!t_m), 1077 .measure_marlon_r_en = cpu_to_le32(!!t_r), 1078 }; 1079 struct { 1080 struct wil6210_mbox_hdr_wmi wmi; 1081 struct wmi_temp_sense_done_event evt; 1082 } __packed reply; 1083 1084 rc = wmi_call(wil, WMI_TEMP_SENSE_CMDID, &cmd, sizeof(cmd), 1085 WMI_TEMP_SENSE_DONE_EVENTID, &reply, sizeof(reply), 100); 1086 if (rc) 1087 return rc; 1088 1089 if (t_m) 1090 *t_m = le32_to_cpu(reply.evt.marlon_m_t1000); 1091 if (t_r) 1092 *t_r = le32_to_cpu(reply.evt.marlon_r_t1000); 1093 1094 return 0; 1095 } 1096 1097 int wmi_disconnect_sta(struct wil6210_priv *wil, const u8 *mac, u16 reason) 1098 { 1099 struct wmi_disconnect_sta_cmd cmd = { 1100 .disconnect_reason = cpu_to_le16(reason), 1101 }; 1102 memcpy(cmd.dst_mac, mac, ETH_ALEN); 1103 1104 wil_dbg_wmi(wil, "%s(%pM, reason %d)\n", __func__, mac, reason); 1105 1106 return wmi_send(wil, WMI_DISCONNECT_STA_CMDID, &cmd, sizeof(cmd)); 1107 } 1108 1109 void wmi_event_flush(struct wil6210_priv *wil) 1110 { 1111 struct pending_wmi_event *evt, *t; 1112 1113 wil_dbg_wmi(wil, "%s()\n", __func__); 1114 1115 list_for_each_entry_safe(evt, t, &wil->pending_wmi_ev, list) { 1116 list_del(&evt->list); 1117 kfree(evt); 1118 } 1119 } 1120 1121 static bool wmi_evt_call_handler(struct wil6210_priv *wil, int id, 1122 void *d, int len) 1123 { 1124 uint i; 1125 1126 for (i = 0; i < ARRAY_SIZE(wmi_evt_handlers); i++) { 1127 if (wmi_evt_handlers[i].eventid == id) { 1128 wmi_evt_handlers[i].handler(wil, id, d, len); 1129 return true; 1130 } 1131 } 1132 1133 return false; 1134 } 1135 1136 static void wmi_event_handle(struct wil6210_priv *wil, 1137 struct wil6210_mbox_hdr *hdr) 1138 { 1139 u16 len = le16_to_cpu(hdr->len); 1140 1141 if ((hdr->type == WIL_MBOX_HDR_TYPE_WMI) && 1142 (len >= sizeof(struct wil6210_mbox_hdr_wmi))) { 1143 struct wil6210_mbox_hdr_wmi *wmi = (void *)(&hdr[1]); 1144 void *evt_data = (void *)(&wmi[1]); 1145 u16 id = le16_to_cpu(wmi->id); 1146 /* check if someone waits for this event */ 1147 if (wil->reply_id && wil->reply_id == id) { 1148 if (wil->reply_buf) { 1149 memcpy(wil->reply_buf, wmi, 1150 min(len, wil->reply_size)); 1151 } else { 1152 wmi_evt_call_handler(wil, id, evt_data, 1153 len - sizeof(*wmi)); 1154 } 1155 wil_dbg_wmi(wil, "Complete WMI 0x%04x\n", id); 1156 complete(&wil->wmi_ready); 1157 return; 1158 } 1159 /* unsolicited event */ 1160 /* search for handler */ 1161 if (!wmi_evt_call_handler(wil, id, evt_data, 1162 len - sizeof(*wmi))) { 1163 wil_err(wil, "Unhandled event 0x%04x\n", id); 1164 } 1165 } else { 1166 wil_err(wil, "Unknown event type\n"); 1167 print_hex_dump(KERN_ERR, "evt?? ", DUMP_PREFIX_OFFSET, 16, 1, 1168 hdr, sizeof(*hdr) + len, true); 1169 } 1170 } 1171 1172 /* 1173 * Retrieve next WMI event from the pending list 1174 */ 1175 static struct list_head *next_wmi_ev(struct wil6210_priv *wil) 1176 { 1177 ulong flags; 1178 struct list_head *ret = NULL; 1179 1180 spin_lock_irqsave(&wil->wmi_ev_lock, flags); 1181 1182 if (!list_empty(&wil->pending_wmi_ev)) { 1183 ret = wil->pending_wmi_ev.next; 1184 list_del(ret); 1185 } 1186 1187 spin_unlock_irqrestore(&wil->wmi_ev_lock, flags); 1188 1189 return ret; 1190 } 1191 1192 /* 1193 * Handler for the WMI events 1194 */ 1195 void wmi_event_worker(struct work_struct *work) 1196 { 1197 struct wil6210_priv *wil = container_of(work, struct wil6210_priv, 1198 wmi_event_worker); 1199 struct pending_wmi_event *evt; 1200 struct list_head *lh; 1201 1202 while ((lh = next_wmi_ev(wil)) != NULL) { 1203 evt = list_entry(lh, struct pending_wmi_event, list); 1204 wmi_event_handle(wil, &evt->event.hdr); 1205 kfree(evt); 1206 } 1207 } 1208