xref: /openbmc/linux/drivers/net/wireless/realtek/rtw88/sdio.c (revision 44ad3baf1cca483e418b6aadf2d3994f69e0f16a)
1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
2 /* Copyright (C) 2021 Martin Blumenstingl <martin.blumenstingl@googlemail.com>
3  * Copyright (C) 2021 Jernej Skrabec <jernej.skrabec@gmail.com>
4  *
5  * Based on rtw88/pci.c:
6  *   Copyright(c) 2018-2019  Realtek Corporation
7  */
8 
9 #include <linux/module.h>
10 #include <linux/mmc/host.h>
11 #include <linux/mmc/sdio_func.h>
12 #include "main.h"
13 #include "debug.h"
14 #include "fw.h"
15 #include "ps.h"
16 #include "reg.h"
17 #include "rx.h"
18 #include "sdio.h"
19 #include "tx.h"
20 
21 #define RTW_SDIO_INDIRECT_RW_RETRIES			50
22 
rtw_sdio_is_bus_addr(u32 addr)23 static bool rtw_sdio_is_bus_addr(u32 addr)
24 {
25 	return !!(addr & RTW_SDIO_BUS_MSK);
26 }
27 
rtw_sdio_bus_claim_needed(struct rtw_sdio * rtwsdio)28 static bool rtw_sdio_bus_claim_needed(struct rtw_sdio *rtwsdio)
29 {
30 	return !rtwsdio->irq_thread ||
31 	       rtwsdio->irq_thread != current;
32 }
33 
rtw_sdio_to_bus_offset(struct rtw_dev * rtwdev,u32 addr)34 static u32 rtw_sdio_to_bus_offset(struct rtw_dev *rtwdev, u32 addr)
35 {
36 	switch (addr & RTW_SDIO_BUS_MSK) {
37 	case WLAN_IOREG_OFFSET:
38 		addr &= WLAN_IOREG_REG_MSK;
39 		addr |= FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
40 				   REG_SDIO_CMD_ADDR_MAC_REG);
41 		break;
42 	case SDIO_LOCAL_OFFSET:
43 		addr &= SDIO_LOCAL_REG_MSK;
44 		addr |= FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
45 				   REG_SDIO_CMD_ADDR_SDIO_REG);
46 		break;
47 	default:
48 		rtw_warn(rtwdev, "Cannot convert addr 0x%08x to bus offset",
49 			 addr);
50 	}
51 
52 	return addr;
53 }
54 
rtw_sdio_use_memcpy_io(struct rtw_dev * rtwdev,u32 addr,u8 alignment)55 static bool rtw_sdio_use_memcpy_io(struct rtw_dev *rtwdev, u32 addr,
56 				   u8 alignment)
57 {
58 	return IS_ALIGNED(addr, alignment) &&
59 	       test_bit(RTW_FLAG_POWERON, rtwdev->flags);
60 }
61 
rtw_sdio_writel(struct rtw_dev * rtwdev,u32 val,u32 addr,int * err_ret)62 static void rtw_sdio_writel(struct rtw_dev *rtwdev, u32 val, u32 addr,
63 			    int *err_ret)
64 {
65 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
66 	u8 buf[4];
67 	int i;
68 
69 	if (rtw_sdio_use_memcpy_io(rtwdev, addr, 4)) {
70 		sdio_writel(rtwsdio->sdio_func, val, addr, err_ret);
71 		return;
72 	}
73 
74 	*(__le32 *)buf = cpu_to_le32(val);
75 
76 	for (i = 0; i < 4; i++) {
77 		sdio_writeb(rtwsdio->sdio_func, buf[i], addr + i, err_ret);
78 		if (*err_ret)
79 			return;
80 	}
81 }
82 
rtw_sdio_writew(struct rtw_dev * rtwdev,u16 val,u32 addr,int * err_ret)83 static void rtw_sdio_writew(struct rtw_dev *rtwdev, u16 val, u32 addr,
84 			    int *err_ret)
85 {
86 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
87 	u8 buf[2];
88 	int i;
89 
90 	*(__le16 *)buf = cpu_to_le16(val);
91 
92 	for (i = 0; i < 2; i++) {
93 		sdio_writeb(rtwsdio->sdio_func, buf[i], addr + i, err_ret);
94 		if (*err_ret)
95 			return;
96 	}
97 }
98 
rtw_sdio_readl(struct rtw_dev * rtwdev,u32 addr,int * err_ret)99 static u32 rtw_sdio_readl(struct rtw_dev *rtwdev, u32 addr, int *err_ret)
100 {
101 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
102 	u8 buf[4];
103 	int i;
104 
105 	if (rtw_sdio_use_memcpy_io(rtwdev, addr, 4))
106 		return sdio_readl(rtwsdio->sdio_func, addr, err_ret);
107 
108 	for (i = 0; i < 4; i++) {
109 		buf[i] = sdio_readb(rtwsdio->sdio_func, addr + i, err_ret);
110 		if (*err_ret)
111 			return 0;
112 	}
113 
114 	return le32_to_cpu(*(__le32 *)buf);
115 }
116 
rtw_sdio_readw(struct rtw_dev * rtwdev,u32 addr,int * err_ret)117 static u16 rtw_sdio_readw(struct rtw_dev *rtwdev, u32 addr, int *err_ret)
118 {
119 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
120 	u8 buf[2];
121 	int i;
122 
123 	for (i = 0; i < 2; i++) {
124 		buf[i] = sdio_readb(rtwsdio->sdio_func, addr + i, err_ret);
125 		if (*err_ret)
126 			return 0;
127 	}
128 
129 	return le16_to_cpu(*(__le16 *)buf);
130 }
131 
rtw_sdio_to_io_address(struct rtw_dev * rtwdev,u32 addr,bool direct)132 static u32 rtw_sdio_to_io_address(struct rtw_dev *rtwdev, u32 addr,
133 				  bool direct)
134 {
135 	if (!direct)
136 		return addr;
137 
138 	if (!rtw_sdio_is_bus_addr(addr))
139 		addr |= WLAN_IOREG_OFFSET;
140 
141 	return rtw_sdio_to_bus_offset(rtwdev, addr);
142 }
143 
rtw_sdio_use_direct_io(struct rtw_dev * rtwdev,u32 addr)144 static bool rtw_sdio_use_direct_io(struct rtw_dev *rtwdev, u32 addr)
145 {
146 	return !rtw_sdio_is_sdio30_supported(rtwdev) ||
147 		rtw_sdio_is_bus_addr(addr);
148 }
149 
rtw_sdio_indirect_reg_cfg(struct rtw_dev * rtwdev,u32 addr,u32 cfg)150 static int rtw_sdio_indirect_reg_cfg(struct rtw_dev *rtwdev, u32 addr, u32 cfg)
151 {
152 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
153 	unsigned int retry;
154 	u32 reg_cfg;
155 	int ret;
156 	u8 tmp;
157 
158 	reg_cfg = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_CFG);
159 
160 	rtw_sdio_writel(rtwdev, addr | cfg | BIT_SDIO_INDIRECT_REG_CFG_UNK20,
161 			reg_cfg, &ret);
162 	if (ret)
163 		return ret;
164 
165 	for (retry = 0; retry < RTW_SDIO_INDIRECT_RW_RETRIES; retry++) {
166 		tmp = sdio_readb(rtwsdio->sdio_func, reg_cfg + 2, &ret);
167 		if (!ret && (tmp & BIT(4)))
168 			return 0;
169 	}
170 
171 	return -ETIMEDOUT;
172 }
173 
rtw_sdio_indirect_read8(struct rtw_dev * rtwdev,u32 addr,int * err_ret)174 static u8 rtw_sdio_indirect_read8(struct rtw_dev *rtwdev, u32 addr,
175 				  int *err_ret)
176 {
177 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
178 	u32 reg_data;
179 
180 	*err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
181 					     BIT_SDIO_INDIRECT_REG_CFG_READ);
182 	if (*err_ret)
183 		return 0;
184 
185 	reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
186 	return sdio_readb(rtwsdio->sdio_func, reg_data, err_ret);
187 }
188 
rtw_sdio_indirect_read_bytes(struct rtw_dev * rtwdev,u32 addr,u8 * buf,int count)189 static int rtw_sdio_indirect_read_bytes(struct rtw_dev *rtwdev, u32 addr,
190 					u8 *buf, int count)
191 {
192 	int i, ret = 0;
193 
194 	for (i = 0; i < count; i++) {
195 		buf[i] = rtw_sdio_indirect_read8(rtwdev, addr + i, &ret);
196 		if (ret)
197 			break;
198 	}
199 
200 	return ret;
201 }
202 
rtw_sdio_indirect_read16(struct rtw_dev * rtwdev,u32 addr,int * err_ret)203 static u16 rtw_sdio_indirect_read16(struct rtw_dev *rtwdev, u32 addr,
204 				    int *err_ret)
205 {
206 	u32 reg_data;
207 	u8 buf[2];
208 
209 	if (!IS_ALIGNED(addr, 2)) {
210 		*err_ret = rtw_sdio_indirect_read_bytes(rtwdev, addr, buf, 2);
211 		if (*err_ret)
212 			return 0;
213 
214 		return le16_to_cpu(*(__le16 *)buf);
215 	}
216 
217 	*err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
218 					     BIT_SDIO_INDIRECT_REG_CFG_READ);
219 	if (*err_ret)
220 		return 0;
221 
222 	reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
223 	return rtw_sdio_readw(rtwdev, reg_data, err_ret);
224 }
225 
rtw_sdio_indirect_read32(struct rtw_dev * rtwdev,u32 addr,int * err_ret)226 static u32 rtw_sdio_indirect_read32(struct rtw_dev *rtwdev, u32 addr,
227 				    int *err_ret)
228 {
229 	u32 reg_data;
230 	u8 buf[4];
231 
232 	if (!IS_ALIGNED(addr, 4)) {
233 		*err_ret = rtw_sdio_indirect_read_bytes(rtwdev, addr, buf, 4);
234 		if (*err_ret)
235 			return 0;
236 
237 		return le32_to_cpu(*(__le32 *)buf);
238 	}
239 
240 	*err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
241 					     BIT_SDIO_INDIRECT_REG_CFG_READ);
242 	if (*err_ret)
243 		return 0;
244 
245 	reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
246 	return rtw_sdio_readl(rtwdev, reg_data, err_ret);
247 }
248 
rtw_sdio_read8(struct rtw_dev * rtwdev,u32 addr)249 static u8 rtw_sdio_read8(struct rtw_dev *rtwdev, u32 addr)
250 {
251 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
252 	bool direct, bus_claim;
253 	int ret;
254 	u8 val;
255 
256 	direct = rtw_sdio_use_direct_io(rtwdev, addr);
257 	addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
258 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
259 
260 	if (bus_claim)
261 		sdio_claim_host(rtwsdio->sdio_func);
262 
263 	if (direct)
264 		val = sdio_readb(rtwsdio->sdio_func, addr, &ret);
265 	else
266 		val = rtw_sdio_indirect_read8(rtwdev, addr, &ret);
267 
268 	if (bus_claim)
269 		sdio_release_host(rtwsdio->sdio_func);
270 
271 	if (ret)
272 		rtw_warn(rtwdev, "sdio read8 failed (0x%x): %d", addr, ret);
273 
274 	return val;
275 }
276 
rtw_sdio_read16(struct rtw_dev * rtwdev,u32 addr)277 static u16 rtw_sdio_read16(struct rtw_dev *rtwdev, u32 addr)
278 {
279 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
280 	bool direct, bus_claim;
281 	int ret;
282 	u16 val;
283 
284 	direct = rtw_sdio_use_direct_io(rtwdev, addr);
285 	addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
286 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
287 
288 	if (bus_claim)
289 		sdio_claim_host(rtwsdio->sdio_func);
290 
291 	if (direct)
292 		val = rtw_sdio_readw(rtwdev, addr, &ret);
293 	else
294 		val = rtw_sdio_indirect_read16(rtwdev, addr, &ret);
295 
296 	if (bus_claim)
297 		sdio_release_host(rtwsdio->sdio_func);
298 
299 	if (ret)
300 		rtw_warn(rtwdev, "sdio read16 failed (0x%x): %d", addr, ret);
301 
302 	return val;
303 }
304 
rtw_sdio_read32(struct rtw_dev * rtwdev,u32 addr)305 static u32 rtw_sdio_read32(struct rtw_dev *rtwdev, u32 addr)
306 {
307 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
308 	bool direct, bus_claim;
309 	u32 val;
310 	int ret;
311 
312 	direct = rtw_sdio_use_direct_io(rtwdev, addr);
313 	addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
314 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
315 
316 	if (bus_claim)
317 		sdio_claim_host(rtwsdio->sdio_func);
318 
319 	if (direct)
320 		val = rtw_sdio_readl(rtwdev, addr, &ret);
321 	else
322 		val = rtw_sdio_indirect_read32(rtwdev, addr, &ret);
323 
324 	if (bus_claim)
325 		sdio_release_host(rtwsdio->sdio_func);
326 
327 	if (ret)
328 		rtw_warn(rtwdev, "sdio read32 failed (0x%x): %d", addr, ret);
329 
330 	return val;
331 }
332 
rtw_sdio_indirect_write8(struct rtw_dev * rtwdev,u8 val,u32 addr,int * err_ret)333 static void rtw_sdio_indirect_write8(struct rtw_dev *rtwdev, u8 val, u32 addr,
334 				     int *err_ret)
335 {
336 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
337 	u32 reg_data;
338 
339 	reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
340 	sdio_writeb(rtwsdio->sdio_func, val, reg_data, err_ret);
341 	if (*err_ret)
342 		return;
343 
344 	*err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
345 					     BIT_SDIO_INDIRECT_REG_CFG_WRITE);
346 }
347 
rtw_sdio_indirect_write16(struct rtw_dev * rtwdev,u16 val,u32 addr,int * err_ret)348 static void rtw_sdio_indirect_write16(struct rtw_dev *rtwdev, u16 val, u32 addr,
349 				      int *err_ret)
350 {
351 	u32 reg_data;
352 
353 	if (!IS_ALIGNED(addr, 2)) {
354 		addr = rtw_sdio_to_io_address(rtwdev, addr, true);
355 		rtw_sdio_writew(rtwdev, val, addr, err_ret);
356 		return;
357 	}
358 
359 	reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
360 	rtw_sdio_writew(rtwdev, val, reg_data, err_ret);
361 	if (*err_ret)
362 		return;
363 
364 	*err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
365 					     BIT_SDIO_INDIRECT_REG_CFG_WRITE |
366 					     BIT_SDIO_INDIRECT_REG_CFG_WORD);
367 }
368 
rtw_sdio_indirect_write32(struct rtw_dev * rtwdev,u32 val,u32 addr,int * err_ret)369 static void rtw_sdio_indirect_write32(struct rtw_dev *rtwdev, u32 val,
370 				      u32 addr, int *err_ret)
371 {
372 	u32 reg_data;
373 
374 	if (!IS_ALIGNED(addr, 4)) {
375 		addr = rtw_sdio_to_io_address(rtwdev, addr, true);
376 		rtw_sdio_writel(rtwdev, val, addr, err_ret);
377 		return;
378 	}
379 
380 	reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
381 	rtw_sdio_writel(rtwdev, val, reg_data, err_ret);
382 
383 	*err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
384 					     BIT_SDIO_INDIRECT_REG_CFG_WRITE |
385 					     BIT_SDIO_INDIRECT_REG_CFG_DWORD);
386 }
387 
rtw_sdio_write8(struct rtw_dev * rtwdev,u32 addr,u8 val)388 static void rtw_sdio_write8(struct rtw_dev *rtwdev, u32 addr, u8 val)
389 {
390 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
391 	bool direct, bus_claim;
392 	int ret;
393 
394 	direct = rtw_sdio_use_direct_io(rtwdev, addr);
395 	addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
396 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
397 
398 	if (bus_claim)
399 		sdio_claim_host(rtwsdio->sdio_func);
400 
401 	if (direct)
402 		sdio_writeb(rtwsdio->sdio_func, val, addr, &ret);
403 	else
404 		rtw_sdio_indirect_write8(rtwdev, val, addr, &ret);
405 
406 	if (bus_claim)
407 		sdio_release_host(rtwsdio->sdio_func);
408 
409 	if (ret)
410 		rtw_warn(rtwdev, "sdio write8 failed (0x%x): %d", addr, ret);
411 }
412 
rtw_sdio_write16(struct rtw_dev * rtwdev,u32 addr,u16 val)413 static void rtw_sdio_write16(struct rtw_dev *rtwdev, u32 addr, u16 val)
414 {
415 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
416 	bool direct, bus_claim;
417 	int ret;
418 
419 	direct = rtw_sdio_use_direct_io(rtwdev, addr);
420 	addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
421 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
422 
423 	if (bus_claim)
424 		sdio_claim_host(rtwsdio->sdio_func);
425 
426 	if (direct)
427 		rtw_sdio_writew(rtwdev, val, addr, &ret);
428 	else
429 		rtw_sdio_indirect_write16(rtwdev, val, addr, &ret);
430 
431 	if (bus_claim)
432 		sdio_release_host(rtwsdio->sdio_func);
433 
434 	if (ret)
435 		rtw_warn(rtwdev, "sdio write16 failed (0x%x): %d", addr, ret);
436 }
437 
rtw_sdio_write32(struct rtw_dev * rtwdev,u32 addr,u32 val)438 static void rtw_sdio_write32(struct rtw_dev *rtwdev, u32 addr, u32 val)
439 {
440 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
441 	bool direct, bus_claim;
442 	int ret;
443 
444 	direct = rtw_sdio_use_direct_io(rtwdev, addr);
445 	addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
446 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
447 
448 	if (bus_claim)
449 		sdio_claim_host(rtwsdio->sdio_func);
450 
451 	if (direct)
452 		rtw_sdio_writel(rtwdev, val, addr, &ret);
453 	else
454 		rtw_sdio_indirect_write32(rtwdev, val, addr, &ret);
455 
456 	if (bus_claim)
457 		sdio_release_host(rtwsdio->sdio_func);
458 
459 	if (ret)
460 		rtw_warn(rtwdev, "sdio write32 failed (0x%x): %d", addr, ret);
461 }
462 
rtw_sdio_get_tx_addr(struct rtw_dev * rtwdev,size_t size,enum rtw_tx_queue_type queue)463 static u32 rtw_sdio_get_tx_addr(struct rtw_dev *rtwdev, size_t size,
464 				enum rtw_tx_queue_type queue)
465 {
466 	u32 txaddr;
467 
468 	switch (queue) {
469 	case RTW_TX_QUEUE_BCN:
470 	case RTW_TX_QUEUE_H2C:
471 	case RTW_TX_QUEUE_HI0:
472 		txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
473 				    REG_SDIO_CMD_ADDR_TXFF_HIGH);
474 		break;
475 	case RTW_TX_QUEUE_VI:
476 	case RTW_TX_QUEUE_VO:
477 		txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
478 				    REG_SDIO_CMD_ADDR_TXFF_NORMAL);
479 		break;
480 	case RTW_TX_QUEUE_BE:
481 	case RTW_TX_QUEUE_BK:
482 		txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
483 				    REG_SDIO_CMD_ADDR_TXFF_LOW);
484 		break;
485 	case RTW_TX_QUEUE_MGMT:
486 		txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
487 				    REG_SDIO_CMD_ADDR_TXFF_EXTRA);
488 		break;
489 	default:
490 		rtw_warn(rtwdev, "Unsupported queue for TX addr: 0x%02x\n",
491 			 queue);
492 		return 0;
493 	}
494 
495 	txaddr += DIV_ROUND_UP(size, 4);
496 
497 	return txaddr;
498 };
499 
rtw_sdio_read_port(struct rtw_dev * rtwdev,u8 * buf,size_t count)500 static int rtw_sdio_read_port(struct rtw_dev *rtwdev, u8 *buf, size_t count)
501 {
502 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
503 	struct mmc_host *host = rtwsdio->sdio_func->card->host;
504 	bool bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
505 	u32 rxaddr = rtwsdio->rx_addr++;
506 	int ret = 0, err;
507 	size_t bytes;
508 
509 	if (bus_claim)
510 		sdio_claim_host(rtwsdio->sdio_func);
511 
512 	while (count > 0) {
513 		bytes = min_t(size_t, host->max_req_size, count);
514 
515 		err = sdio_memcpy_fromio(rtwsdio->sdio_func, buf,
516 					 RTW_SDIO_ADDR_RX_RX0FF_GEN(rxaddr),
517 					 bytes);
518 		if (err) {
519 			rtw_warn(rtwdev,
520 				 "Failed to read %zu byte(s) from SDIO port 0x%08x: %d",
521 				 bytes, rxaddr, err);
522 
523 			 /* Signal to the caller that reading did not work and
524 			  * that the data in the buffer is short/corrupted.
525 			  */
526 			ret = err;
527 
528 			/* Don't stop here - instead drain the remaining data
529 			 * from the card's buffer, else the card will return
530 			 * corrupt data for the next rtw_sdio_read_port() call.
531 			 */
532 		}
533 
534 		count -= bytes;
535 		buf += bytes;
536 	}
537 
538 	if (bus_claim)
539 		sdio_release_host(rtwsdio->sdio_func);
540 
541 	return ret;
542 }
543 
rtw_sdio_check_free_txpg(struct rtw_dev * rtwdev,u8 queue,size_t count)544 static int rtw_sdio_check_free_txpg(struct rtw_dev *rtwdev, u8 queue,
545 				    size_t count)
546 {
547 	unsigned int pages_free, pages_needed;
548 
549 	if (rtw_chip_wcpu_11n(rtwdev)) {
550 		u32 free_txpg;
551 
552 		free_txpg = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG);
553 
554 		switch (queue) {
555 		case RTW_TX_QUEUE_BCN:
556 		case RTW_TX_QUEUE_H2C:
557 		case RTW_TX_QUEUE_HI0:
558 		case RTW_TX_QUEUE_MGMT:
559 			/* high */
560 			pages_free = free_txpg & 0xff;
561 			break;
562 		case RTW_TX_QUEUE_VI:
563 		case RTW_TX_QUEUE_VO:
564 			/* normal */
565 			pages_free = (free_txpg >> 8) & 0xff;
566 			break;
567 		case RTW_TX_QUEUE_BE:
568 		case RTW_TX_QUEUE_BK:
569 			/* low */
570 			pages_free = (free_txpg >> 16) & 0xff;
571 			break;
572 		default:
573 			rtw_warn(rtwdev, "Unknown mapping for queue %u\n", queue);
574 			return -EINVAL;
575 		}
576 
577 		/* add the pages from the public queue */
578 		pages_free += (free_txpg >> 24) & 0xff;
579 	} else {
580 		u32 free_txpg[3];
581 
582 		free_txpg[0] = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG);
583 		free_txpg[1] = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG + 4);
584 		free_txpg[2] = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG + 8);
585 
586 		switch (queue) {
587 		case RTW_TX_QUEUE_BCN:
588 		case RTW_TX_QUEUE_H2C:
589 		case RTW_TX_QUEUE_HI0:
590 			/* high */
591 			pages_free = free_txpg[0] & 0xfff;
592 			break;
593 		case RTW_TX_QUEUE_VI:
594 		case RTW_TX_QUEUE_VO:
595 			/* normal */
596 			pages_free = (free_txpg[0] >> 16) & 0xfff;
597 			break;
598 		case RTW_TX_QUEUE_BE:
599 		case RTW_TX_QUEUE_BK:
600 			/* low */
601 			pages_free = free_txpg[1] & 0xfff;
602 			break;
603 		case RTW_TX_QUEUE_MGMT:
604 			/* extra */
605 			pages_free = free_txpg[2] & 0xfff;
606 			break;
607 		default:
608 			rtw_warn(rtwdev, "Unknown mapping for queue %u\n", queue);
609 			return -EINVAL;
610 		}
611 
612 		/* add the pages from the public queue */
613 		pages_free += (free_txpg[1] >> 16) & 0xfff;
614 	}
615 
616 	pages_needed = DIV_ROUND_UP(count, rtwdev->chip->page_size);
617 
618 	if (pages_needed > pages_free) {
619 		rtw_dbg(rtwdev, RTW_DBG_SDIO,
620 			"Not enough free pages (%u needed, %u free) in queue %u for %zu bytes\n",
621 			pages_needed, pages_free, queue, count);
622 		return -EBUSY;
623 	}
624 
625 	return 0;
626 }
627 
rtw_sdio_write_port(struct rtw_dev * rtwdev,struct sk_buff * skb,enum rtw_tx_queue_type queue)628 static int rtw_sdio_write_port(struct rtw_dev *rtwdev, struct sk_buff *skb,
629 			       enum rtw_tx_queue_type queue)
630 {
631 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
632 	bool bus_claim;
633 	size_t txsize;
634 	u32 txaddr;
635 	int ret;
636 
637 	txaddr = rtw_sdio_get_tx_addr(rtwdev, skb->len, queue);
638 	if (!txaddr)
639 		return -EINVAL;
640 
641 	txsize = sdio_align_size(rtwsdio->sdio_func, skb->len);
642 
643 	ret = rtw_sdio_check_free_txpg(rtwdev, queue, txsize);
644 	if (ret)
645 		return ret;
646 
647 	if (!IS_ALIGNED((unsigned long)skb->data, RTW_SDIO_DATA_PTR_ALIGN))
648 		rtw_warn(rtwdev, "Got unaligned SKB in %s() for queue %u\n",
649 			 __func__, queue);
650 
651 	bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
652 
653 	if (bus_claim)
654 		sdio_claim_host(rtwsdio->sdio_func);
655 
656 	ret = sdio_memcpy_toio(rtwsdio->sdio_func, txaddr, skb->data, txsize);
657 
658 	if (bus_claim)
659 		sdio_release_host(rtwsdio->sdio_func);
660 
661 	if (ret)
662 		rtw_warn(rtwdev,
663 			 "Failed to write %zu byte(s) to SDIO port 0x%08x",
664 			 txsize, txaddr);
665 
666 	return ret;
667 }
668 
rtw_sdio_init(struct rtw_dev * rtwdev)669 static void rtw_sdio_init(struct rtw_dev *rtwdev)
670 {
671 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
672 
673 	rtwsdio->irq_mask = REG_SDIO_HIMR_RX_REQUEST | REG_SDIO_HIMR_CPWM1;
674 }
675 
rtw_sdio_enable_rx_aggregation(struct rtw_dev * rtwdev)676 static void rtw_sdio_enable_rx_aggregation(struct rtw_dev *rtwdev)
677 {
678 	u8 size, timeout;
679 
680 	if (rtw_chip_wcpu_11n(rtwdev)) {
681 		size = 0x6;
682 		timeout = 0x6;
683 	} else {
684 		size = 0xff;
685 		timeout = 0x1;
686 	}
687 
688 	/* Make the firmware honor the size limit configured below */
689 	rtw_write32_set(rtwdev, REG_RXDMA_AGG_PG_TH, BIT_EN_PRE_CALC);
690 
691 	rtw_write8_set(rtwdev, REG_TXDMA_PQ_MAP, BIT_RXDMA_AGG_EN);
692 
693 	rtw_write16(rtwdev, REG_RXDMA_AGG_PG_TH,
694 		    FIELD_PREP(BIT_RXDMA_AGG_PG_TH, size) |
695 		    FIELD_PREP(BIT_DMA_AGG_TO_V1, timeout));
696 
697 	rtw_write8_set(rtwdev, REG_RXDMA_MODE, BIT_DMA_MODE);
698 }
699 
rtw_sdio_enable_interrupt(struct rtw_dev * rtwdev)700 static void rtw_sdio_enable_interrupt(struct rtw_dev *rtwdev)
701 {
702 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
703 
704 	rtw_write32(rtwdev, REG_SDIO_HIMR, rtwsdio->irq_mask);
705 }
706 
rtw_sdio_disable_interrupt(struct rtw_dev * rtwdev)707 static void rtw_sdio_disable_interrupt(struct rtw_dev *rtwdev)
708 {
709 	rtw_write32(rtwdev, REG_SDIO_HIMR, 0x0);
710 }
711 
rtw_sdio_get_tx_qsel(struct rtw_dev * rtwdev,struct sk_buff * skb,u8 queue)712 static u8 rtw_sdio_get_tx_qsel(struct rtw_dev *rtwdev, struct sk_buff *skb,
713 			       u8 queue)
714 {
715 	switch (queue) {
716 	case RTW_TX_QUEUE_BCN:
717 		return TX_DESC_QSEL_BEACON;
718 	case RTW_TX_QUEUE_H2C:
719 		return TX_DESC_QSEL_H2C;
720 	case RTW_TX_QUEUE_MGMT:
721 		return TX_DESC_QSEL_MGMT;
722 	case RTW_TX_QUEUE_HI0:
723 		return TX_DESC_QSEL_HIGH;
724 	default:
725 		return skb->priority;
726 	}
727 }
728 
rtw_sdio_setup(struct rtw_dev * rtwdev)729 static int rtw_sdio_setup(struct rtw_dev *rtwdev)
730 {
731 	/* nothing to do */
732 	return 0;
733 }
734 
rtw_sdio_start(struct rtw_dev * rtwdev)735 static int rtw_sdio_start(struct rtw_dev *rtwdev)
736 {
737 	rtw_sdio_enable_rx_aggregation(rtwdev);
738 	rtw_sdio_enable_interrupt(rtwdev);
739 
740 	return 0;
741 }
742 
rtw_sdio_stop(struct rtw_dev * rtwdev)743 static void rtw_sdio_stop(struct rtw_dev *rtwdev)
744 {
745 	rtw_sdio_disable_interrupt(rtwdev);
746 }
747 
rtw_sdio_deep_ps_enter(struct rtw_dev * rtwdev)748 static void rtw_sdio_deep_ps_enter(struct rtw_dev *rtwdev)
749 {
750 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
751 	bool tx_empty = true;
752 	u8 queue;
753 
754 	if (!rtw_fw_feature_check(&rtwdev->fw, FW_FEATURE_TX_WAKE)) {
755 		/* Deep PS state is not allowed to TX-DMA */
756 		for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) {
757 			/* BCN queue is rsvd page, does not have DMA interrupt
758 			 * H2C queue is managed by firmware
759 			 */
760 			if (queue == RTW_TX_QUEUE_BCN ||
761 			    queue == RTW_TX_QUEUE_H2C)
762 				continue;
763 
764 			/* check if there is any skb DMAing */
765 			if (skb_queue_len(&rtwsdio->tx_queue[queue])) {
766 				tx_empty = false;
767 				break;
768 			}
769 		}
770 	}
771 
772 	if (!tx_empty) {
773 		rtw_dbg(rtwdev, RTW_DBG_PS,
774 			"TX path not empty, cannot enter deep power save state\n");
775 		return;
776 	}
777 
778 	set_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags);
779 	rtw_power_mode_change(rtwdev, true);
780 }
781 
rtw_sdio_deep_ps_leave(struct rtw_dev * rtwdev)782 static void rtw_sdio_deep_ps_leave(struct rtw_dev *rtwdev)
783 {
784 	if (test_and_clear_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags))
785 		rtw_power_mode_change(rtwdev, false);
786 }
787 
rtw_sdio_deep_ps(struct rtw_dev * rtwdev,bool enter)788 static void rtw_sdio_deep_ps(struct rtw_dev *rtwdev, bool enter)
789 {
790 	if (enter && !test_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags))
791 		rtw_sdio_deep_ps_enter(rtwdev);
792 
793 	if (!enter && test_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags))
794 		rtw_sdio_deep_ps_leave(rtwdev);
795 }
796 
rtw_sdio_tx_kick_off(struct rtw_dev * rtwdev)797 static void rtw_sdio_tx_kick_off(struct rtw_dev *rtwdev)
798 {
799 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
800 
801 	queue_work(rtwsdio->txwq, &rtwsdio->tx_handler_data->work);
802 }
803 
rtw_sdio_link_ps(struct rtw_dev * rtwdev,bool enter)804 static void rtw_sdio_link_ps(struct rtw_dev *rtwdev, bool enter)
805 {
806 	/* nothing to do */
807 }
808 
rtw_sdio_interface_cfg(struct rtw_dev * rtwdev)809 static void rtw_sdio_interface_cfg(struct rtw_dev *rtwdev)
810 {
811 	u32 val;
812 
813 	rtw_read32(rtwdev, REG_SDIO_FREE_TXPG);
814 
815 	val = rtw_read32(rtwdev, REG_SDIO_TX_CTRL);
816 	val &= 0xfff8;
817 	rtw_write32(rtwdev, REG_SDIO_TX_CTRL, val);
818 }
819 
rtw_sdio_get_tx_data(struct sk_buff * skb)820 static struct rtw_sdio_tx_data *rtw_sdio_get_tx_data(struct sk_buff *skb)
821 {
822 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
823 
824 	BUILD_BUG_ON(sizeof(struct rtw_sdio_tx_data) >
825 		     sizeof(info->status.status_driver_data));
826 
827 	return (struct rtw_sdio_tx_data *)info->status.status_driver_data;
828 }
829 
rtw_sdio_tx_skb_prepare(struct rtw_dev * rtwdev,struct rtw_tx_pkt_info * pkt_info,struct sk_buff * skb,enum rtw_tx_queue_type queue)830 static void rtw_sdio_tx_skb_prepare(struct rtw_dev *rtwdev,
831 				    struct rtw_tx_pkt_info *pkt_info,
832 				    struct sk_buff *skb,
833 				    enum rtw_tx_queue_type queue)
834 {
835 	const struct rtw_chip_info *chip = rtwdev->chip;
836 	unsigned long data_addr, aligned_addr;
837 	size_t offset;
838 	u8 *pkt_desc;
839 
840 	pkt_desc = skb_push(skb, chip->tx_pkt_desc_sz);
841 
842 	data_addr = (unsigned long)pkt_desc;
843 	aligned_addr = ALIGN(data_addr, RTW_SDIO_DATA_PTR_ALIGN);
844 
845 	if (data_addr != aligned_addr) {
846 		/* Ensure that the start of the pkt_desc is always aligned at
847 		 * RTW_SDIO_DATA_PTR_ALIGN.
848 		 */
849 		offset = RTW_SDIO_DATA_PTR_ALIGN - (aligned_addr - data_addr);
850 
851 		pkt_desc = skb_push(skb, offset);
852 
853 		/* By inserting padding to align the start of the pkt_desc we
854 		 * need to inform the firmware that the actual data starts at
855 		 * a different offset than normal.
856 		 */
857 		pkt_info->offset += offset;
858 	}
859 
860 	memset(pkt_desc, 0, chip->tx_pkt_desc_sz);
861 
862 	pkt_info->qsel = rtw_sdio_get_tx_qsel(rtwdev, skb, queue);
863 
864 	rtw_tx_fill_tx_desc(pkt_info, skb);
865 	rtw_tx_fill_txdesc_checksum(rtwdev, pkt_info, pkt_desc);
866 }
867 
rtw_sdio_write_data(struct rtw_dev * rtwdev,struct rtw_tx_pkt_info * pkt_info,struct sk_buff * skb,enum rtw_tx_queue_type queue)868 static int rtw_sdio_write_data(struct rtw_dev *rtwdev,
869 			       struct rtw_tx_pkt_info *pkt_info,
870 			       struct sk_buff *skb,
871 			       enum rtw_tx_queue_type queue)
872 {
873 	int ret;
874 
875 	rtw_sdio_tx_skb_prepare(rtwdev, pkt_info, skb, queue);
876 
877 	ret = rtw_sdio_write_port(rtwdev, skb, queue);
878 	dev_kfree_skb_any(skb);
879 
880 	return ret;
881 }
882 
rtw_sdio_write_data_rsvd_page(struct rtw_dev * rtwdev,u8 * buf,u32 size)883 static int rtw_sdio_write_data_rsvd_page(struct rtw_dev *rtwdev, u8 *buf,
884 					 u32 size)
885 {
886 	struct rtw_tx_pkt_info pkt_info = {};
887 	struct sk_buff *skb;
888 
889 	skb = rtw_tx_write_data_rsvd_page_get(rtwdev, &pkt_info, buf, size);
890 	if (!skb)
891 		return -ENOMEM;
892 
893 	return rtw_sdio_write_data(rtwdev, &pkt_info, skb, RTW_TX_QUEUE_BCN);
894 }
895 
rtw_sdio_write_data_h2c(struct rtw_dev * rtwdev,u8 * buf,u32 size)896 static int rtw_sdio_write_data_h2c(struct rtw_dev *rtwdev, u8 *buf, u32 size)
897 {
898 	struct rtw_tx_pkt_info pkt_info = {};
899 	struct sk_buff *skb;
900 
901 	skb = rtw_tx_write_data_h2c_get(rtwdev, &pkt_info, buf, size);
902 	if (!skb)
903 		return -ENOMEM;
904 
905 	return rtw_sdio_write_data(rtwdev, &pkt_info, skb, RTW_TX_QUEUE_H2C);
906 }
907 
rtw_sdio_tx_write(struct rtw_dev * rtwdev,struct rtw_tx_pkt_info * pkt_info,struct sk_buff * skb)908 static int rtw_sdio_tx_write(struct rtw_dev *rtwdev,
909 			     struct rtw_tx_pkt_info *pkt_info,
910 			     struct sk_buff *skb)
911 {
912 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
913 	enum rtw_tx_queue_type queue = rtw_tx_queue_mapping(skb);
914 	struct rtw_sdio_tx_data *tx_data;
915 
916 	rtw_sdio_tx_skb_prepare(rtwdev, pkt_info, skb, queue);
917 
918 	tx_data = rtw_sdio_get_tx_data(skb);
919 	tx_data->sn = pkt_info->sn;
920 
921 	skb_queue_tail(&rtwsdio->tx_queue[queue], skb);
922 
923 	return 0;
924 }
925 
rtw_sdio_tx_err_isr(struct rtw_dev * rtwdev)926 static void rtw_sdio_tx_err_isr(struct rtw_dev *rtwdev)
927 {
928 	u32 val = rtw_read32(rtwdev, REG_TXDMA_STATUS);
929 
930 	rtw_write32(rtwdev, REG_TXDMA_STATUS, val);
931 }
932 
rtw_sdio_rx_skb(struct rtw_dev * rtwdev,struct sk_buff * skb,u32 pkt_offset,struct rtw_rx_pkt_stat * pkt_stat,struct ieee80211_rx_status * rx_status)933 static void rtw_sdio_rx_skb(struct rtw_dev *rtwdev, struct sk_buff *skb,
934 			    u32 pkt_offset, struct rtw_rx_pkt_stat *pkt_stat,
935 			    struct ieee80211_rx_status *rx_status)
936 {
937 	*IEEE80211_SKB_RXCB(skb) = *rx_status;
938 
939 	if (pkt_stat->is_c2h) {
940 		skb_put(skb, pkt_stat->pkt_len + pkt_offset);
941 		rtw_fw_c2h_cmd_rx_irqsafe(rtwdev, pkt_offset, skb);
942 		return;
943 	}
944 
945 	skb_put(skb, pkt_stat->pkt_len);
946 	skb_reserve(skb, pkt_offset);
947 
948 	rtw_rx_stats(rtwdev, pkt_stat->vif, skb);
949 
950 	ieee80211_rx_irqsafe(rtwdev->hw, skb);
951 }
952 
rtw_sdio_rxfifo_recv(struct rtw_dev * rtwdev,u32 rx_len)953 static void rtw_sdio_rxfifo_recv(struct rtw_dev *rtwdev, u32 rx_len)
954 {
955 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
956 	const struct rtw_chip_info *chip = rtwdev->chip;
957 	u32 pkt_desc_sz = chip->rx_pkt_desc_sz;
958 	struct ieee80211_rx_status rx_status;
959 	struct rtw_rx_pkt_stat pkt_stat;
960 	struct sk_buff *skb, *split_skb;
961 	u32 pkt_offset, curr_pkt_len;
962 	size_t bufsz;
963 	u8 *rx_desc;
964 	int ret;
965 
966 	bufsz = sdio_align_size(rtwsdio->sdio_func, rx_len);
967 
968 	skb = dev_alloc_skb(bufsz);
969 	if (!skb)
970 		return;
971 
972 	ret = rtw_sdio_read_port(rtwdev, skb->data, bufsz);
973 	if (ret) {
974 		dev_kfree_skb_any(skb);
975 		return;
976 	}
977 
978 	while (true) {
979 		rx_desc = skb->data;
980 		chip->ops->query_rx_desc(rtwdev, rx_desc, &pkt_stat,
981 					 &rx_status);
982 		pkt_offset = pkt_desc_sz + pkt_stat.drv_info_sz +
983 			     pkt_stat.shift;
984 
985 		curr_pkt_len = ALIGN(pkt_offset + pkt_stat.pkt_len,
986 				     RTW_SDIO_DATA_PTR_ALIGN);
987 
988 		if ((curr_pkt_len + pkt_desc_sz) >= rx_len) {
989 			/* Use the original skb (with it's adjusted offset)
990 			 * when processing the last (or even the only) entry to
991 			 * have it's memory freed automatically.
992 			 */
993 			rtw_sdio_rx_skb(rtwdev, skb, pkt_offset, &pkt_stat,
994 					&rx_status);
995 			break;
996 		}
997 
998 		split_skb = dev_alloc_skb(curr_pkt_len);
999 		if (!split_skb) {
1000 			rtw_sdio_rx_skb(rtwdev, skb, pkt_offset, &pkt_stat,
1001 					&rx_status);
1002 			break;
1003 		}
1004 
1005 		skb_copy_header(split_skb, skb);
1006 		memcpy(split_skb->data, skb->data, curr_pkt_len);
1007 
1008 		rtw_sdio_rx_skb(rtwdev, split_skb, pkt_offset, &pkt_stat,
1009 				&rx_status);
1010 
1011 		/* Move to the start of the next RX descriptor */
1012 		skb_reserve(skb, curr_pkt_len);
1013 		rx_len -= curr_pkt_len;
1014 	}
1015 }
1016 
rtw_sdio_rx_isr(struct rtw_dev * rtwdev)1017 static void rtw_sdio_rx_isr(struct rtw_dev *rtwdev)
1018 {
1019 	u32 rx_len, hisr, total_rx_bytes = 0;
1020 
1021 	do {
1022 		if (rtw_chip_wcpu_11n(rtwdev))
1023 			rx_len = rtw_read16(rtwdev, REG_SDIO_RX0_REQ_LEN);
1024 		else
1025 			rx_len = rtw_read32(rtwdev, REG_SDIO_RX0_REQ_LEN);
1026 
1027 		if (!rx_len)
1028 			break;
1029 
1030 		rtw_sdio_rxfifo_recv(rtwdev, rx_len);
1031 
1032 		total_rx_bytes += rx_len;
1033 
1034 		if (rtw_chip_wcpu_11n(rtwdev)) {
1035 			/* Stop if no more RX requests are pending, even if
1036 			 * rx_len could be greater than zero in the next
1037 			 * iteration. This is needed because the RX buffer may
1038 			 * already contain data while either HW or FW are not
1039 			 * done filling that buffer yet. Still reading the
1040 			 * buffer can result in packets where
1041 			 * rtw_rx_pkt_stat.pkt_len is zero or points beyond the
1042 			 * end of the buffer.
1043 			 */
1044 			hisr = rtw_read32(rtwdev, REG_SDIO_HISR);
1045 		} else {
1046 			/* RTW_WCPU_11AC chips have improved hardware or
1047 			 * firmware and can use rx_len unconditionally.
1048 			 */
1049 			hisr = REG_SDIO_HISR_RX_REQUEST;
1050 		}
1051 	} while (total_rx_bytes < SZ_64K && hisr & REG_SDIO_HISR_RX_REQUEST);
1052 }
1053 
rtw_sdio_handle_interrupt(struct sdio_func * sdio_func)1054 static void rtw_sdio_handle_interrupt(struct sdio_func *sdio_func)
1055 {
1056 	struct ieee80211_hw *hw = sdio_get_drvdata(sdio_func);
1057 	struct rtw_sdio *rtwsdio;
1058 	struct rtw_dev *rtwdev;
1059 	u32 hisr;
1060 
1061 	rtwdev = hw->priv;
1062 	rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1063 
1064 	rtwsdio->irq_thread = current;
1065 
1066 	hisr = rtw_read32(rtwdev, REG_SDIO_HISR);
1067 
1068 	if (hisr & REG_SDIO_HISR_TXERR)
1069 		rtw_sdio_tx_err_isr(rtwdev);
1070 	if (hisr & REG_SDIO_HISR_RX_REQUEST) {
1071 		hisr &= ~REG_SDIO_HISR_RX_REQUEST;
1072 		rtw_sdio_rx_isr(rtwdev);
1073 	}
1074 
1075 	rtw_write32(rtwdev, REG_SDIO_HISR, hisr);
1076 
1077 	rtwsdio->irq_thread = NULL;
1078 }
1079 
rtw_sdio_suspend(struct device * dev)1080 static int __maybe_unused rtw_sdio_suspend(struct device *dev)
1081 {
1082 	struct sdio_func *func = dev_to_sdio_func(dev);
1083 	struct ieee80211_hw *hw = dev_get_drvdata(dev);
1084 	struct rtw_dev *rtwdev = hw->priv;
1085 	int ret;
1086 
1087 	ret = sdio_set_host_pm_flags(func, MMC_PM_KEEP_POWER);
1088 	if (ret)
1089 		rtw_err(rtwdev, "Failed to host PM flag MMC_PM_KEEP_POWER");
1090 
1091 	return ret;
1092 }
1093 
rtw_sdio_resume(struct device * dev)1094 static int __maybe_unused rtw_sdio_resume(struct device *dev)
1095 {
1096 	return 0;
1097 }
1098 
1099 SIMPLE_DEV_PM_OPS(rtw_sdio_pm_ops, rtw_sdio_suspend, rtw_sdio_resume);
1100 EXPORT_SYMBOL(rtw_sdio_pm_ops);
1101 
rtw_sdio_claim(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1102 static int rtw_sdio_claim(struct rtw_dev *rtwdev, struct sdio_func *sdio_func)
1103 {
1104 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1105 	int ret;
1106 
1107 	sdio_claim_host(sdio_func);
1108 
1109 	ret = sdio_enable_func(sdio_func);
1110 	if (ret) {
1111 		rtw_err(rtwdev, "Failed to enable SDIO func");
1112 		goto err_release_host;
1113 	}
1114 
1115 	ret = sdio_set_block_size(sdio_func, RTW_SDIO_BLOCK_SIZE);
1116 	if (ret) {
1117 		rtw_err(rtwdev, "Failed to set SDIO block size to 512");
1118 		goto err_disable_func;
1119 	}
1120 
1121 	rtwsdio->sdio_func = sdio_func;
1122 
1123 	rtwsdio->sdio3_bus_mode = mmc_card_uhs(sdio_func->card);
1124 
1125 	sdio_set_drvdata(sdio_func, rtwdev->hw);
1126 	SET_IEEE80211_DEV(rtwdev->hw, &sdio_func->dev);
1127 
1128 	sdio_release_host(sdio_func);
1129 
1130 	return 0;
1131 
1132 err_disable_func:
1133 	sdio_disable_func(sdio_func);
1134 err_release_host:
1135 	sdio_release_host(sdio_func);
1136 	return ret;
1137 }
1138 
rtw_sdio_declaim(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1139 static void rtw_sdio_declaim(struct rtw_dev *rtwdev,
1140 			     struct sdio_func *sdio_func)
1141 {
1142 	sdio_claim_host(sdio_func);
1143 	sdio_disable_func(sdio_func);
1144 	sdio_release_host(sdio_func);
1145 }
1146 
1147 static struct rtw_hci_ops rtw_sdio_ops = {
1148 	.tx_write = rtw_sdio_tx_write,
1149 	.tx_kick_off = rtw_sdio_tx_kick_off,
1150 	.setup = rtw_sdio_setup,
1151 	.start = rtw_sdio_start,
1152 	.stop = rtw_sdio_stop,
1153 	.deep_ps = rtw_sdio_deep_ps,
1154 	.link_ps = rtw_sdio_link_ps,
1155 	.interface_cfg = rtw_sdio_interface_cfg,
1156 
1157 	.read8 = rtw_sdio_read8,
1158 	.read16 = rtw_sdio_read16,
1159 	.read32 = rtw_sdio_read32,
1160 	.write8 = rtw_sdio_write8,
1161 	.write16 = rtw_sdio_write16,
1162 	.write32 = rtw_sdio_write32,
1163 	.write_data_rsvd_page = rtw_sdio_write_data_rsvd_page,
1164 	.write_data_h2c = rtw_sdio_write_data_h2c,
1165 };
1166 
rtw_sdio_request_irq(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1167 static int rtw_sdio_request_irq(struct rtw_dev *rtwdev,
1168 				struct sdio_func *sdio_func)
1169 {
1170 	int ret;
1171 
1172 	sdio_claim_host(sdio_func);
1173 	ret = sdio_claim_irq(sdio_func, &rtw_sdio_handle_interrupt);
1174 	sdio_release_host(sdio_func);
1175 
1176 	if (ret) {
1177 		rtw_err(rtwdev, "failed to claim SDIO IRQ");
1178 		return ret;
1179 	}
1180 
1181 	return 0;
1182 }
1183 
rtw_sdio_indicate_tx_status(struct rtw_dev * rtwdev,struct sk_buff * skb)1184 static void rtw_sdio_indicate_tx_status(struct rtw_dev *rtwdev,
1185 					struct sk_buff *skb)
1186 {
1187 	struct rtw_sdio_tx_data *tx_data = rtw_sdio_get_tx_data(skb);
1188 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
1189 	struct ieee80211_hw *hw = rtwdev->hw;
1190 
1191 	skb_pull(skb, rtwdev->chip->tx_pkt_desc_sz);
1192 
1193 	/* enqueue to wait for tx report */
1194 	if (info->flags & IEEE80211_TX_CTL_REQ_TX_STATUS) {
1195 		rtw_tx_report_enqueue(rtwdev, skb, tx_data->sn);
1196 		return;
1197 	}
1198 
1199 	/* always ACK for others, then they won't be marked as drop */
1200 	ieee80211_tx_info_clear_status(info);
1201 	if (info->flags & IEEE80211_TX_CTL_NO_ACK)
1202 		info->flags |= IEEE80211_TX_STAT_NOACK_TRANSMITTED;
1203 	else
1204 		info->flags |= IEEE80211_TX_STAT_ACK;
1205 
1206 	ieee80211_tx_status_irqsafe(hw, skb);
1207 }
1208 
rtw_sdio_process_tx_queue(struct rtw_dev * rtwdev,enum rtw_tx_queue_type queue)1209 static void rtw_sdio_process_tx_queue(struct rtw_dev *rtwdev,
1210 				      enum rtw_tx_queue_type queue)
1211 {
1212 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1213 	struct sk_buff *skb;
1214 	int ret;
1215 
1216 	skb = skb_dequeue(&rtwsdio->tx_queue[queue]);
1217 	if (!skb)
1218 		return;
1219 
1220 	ret = rtw_sdio_write_port(rtwdev, skb, queue);
1221 	if (ret) {
1222 		skb_queue_head(&rtwsdio->tx_queue[queue], skb);
1223 		return;
1224 	}
1225 
1226 	rtw_sdio_indicate_tx_status(rtwdev, skb);
1227 }
1228 
rtw_sdio_tx_handler(struct work_struct * work)1229 static void rtw_sdio_tx_handler(struct work_struct *work)
1230 {
1231 	struct rtw_sdio_work_data *work_data =
1232 		container_of(work, struct rtw_sdio_work_data, work);
1233 	struct rtw_sdio *rtwsdio;
1234 	struct rtw_dev *rtwdev;
1235 	int limit, queue;
1236 
1237 	rtwdev = work_data->rtwdev;
1238 	rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1239 
1240 	if (!rtw_fw_feature_check(&rtwdev->fw, FW_FEATURE_TX_WAKE))
1241 		rtw_sdio_deep_ps_leave(rtwdev);
1242 
1243 	for (queue = RTK_MAX_TX_QUEUE_NUM - 1; queue >= 0; queue--) {
1244 		for (limit = 0; limit < 1000; limit++) {
1245 			rtw_sdio_process_tx_queue(rtwdev, queue);
1246 
1247 			if (skb_queue_empty(&rtwsdio->tx_queue[queue]))
1248 				break;
1249 		}
1250 	}
1251 }
1252 
rtw_sdio_free_irq(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1253 static void rtw_sdio_free_irq(struct rtw_dev *rtwdev,
1254 			      struct sdio_func *sdio_func)
1255 {
1256 	sdio_claim_host(sdio_func);
1257 	sdio_release_irq(sdio_func);
1258 	sdio_release_host(sdio_func);
1259 }
1260 
rtw_sdio_init_tx(struct rtw_dev * rtwdev)1261 static int rtw_sdio_init_tx(struct rtw_dev *rtwdev)
1262 {
1263 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1264 	int i;
1265 
1266 	rtwsdio->txwq = create_singlethread_workqueue("rtw88_sdio: tx wq");
1267 	if (!rtwsdio->txwq) {
1268 		rtw_err(rtwdev, "failed to create TX work queue\n");
1269 		return -ENOMEM;
1270 	}
1271 
1272 	for (i = 0; i < RTK_MAX_TX_QUEUE_NUM; i++)
1273 		skb_queue_head_init(&rtwsdio->tx_queue[i]);
1274 	rtwsdio->tx_handler_data = kmalloc(sizeof(*rtwsdio->tx_handler_data),
1275 					   GFP_KERNEL);
1276 	if (!rtwsdio->tx_handler_data)
1277 		goto err_destroy_wq;
1278 
1279 	rtwsdio->tx_handler_data->rtwdev = rtwdev;
1280 	INIT_WORK(&rtwsdio->tx_handler_data->work, rtw_sdio_tx_handler);
1281 
1282 	return 0;
1283 
1284 err_destroy_wq:
1285 	destroy_workqueue(rtwsdio->txwq);
1286 	return -ENOMEM;
1287 }
1288 
rtw_sdio_deinit_tx(struct rtw_dev * rtwdev)1289 static void rtw_sdio_deinit_tx(struct rtw_dev *rtwdev)
1290 {
1291 	struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1292 	int i;
1293 
1294 	flush_workqueue(rtwsdio->txwq);
1295 	destroy_workqueue(rtwsdio->txwq);
1296 	kfree(rtwsdio->tx_handler_data);
1297 
1298 	for (i = 0; i < RTK_MAX_TX_QUEUE_NUM; i++)
1299 		ieee80211_purge_tx_queue(rtwdev->hw, &rtwsdio->tx_queue[i]);
1300 }
1301 
rtw_sdio_probe(struct sdio_func * sdio_func,const struct sdio_device_id * id)1302 int rtw_sdio_probe(struct sdio_func *sdio_func,
1303 		   const struct sdio_device_id *id)
1304 {
1305 	struct ieee80211_hw *hw;
1306 	struct rtw_dev *rtwdev;
1307 	int drv_data_size;
1308 	int ret;
1309 
1310 	drv_data_size = sizeof(struct rtw_dev) + sizeof(struct rtw_sdio);
1311 	hw = ieee80211_alloc_hw(drv_data_size, &rtw_ops);
1312 	if (!hw) {
1313 		dev_err(&sdio_func->dev, "failed to allocate hw");
1314 		return -ENOMEM;
1315 	}
1316 
1317 	rtwdev = hw->priv;
1318 	rtwdev->hw = hw;
1319 	rtwdev->dev = &sdio_func->dev;
1320 	rtwdev->chip = (struct rtw_chip_info *)id->driver_data;
1321 	rtwdev->hci.ops = &rtw_sdio_ops;
1322 	rtwdev->hci.type = RTW_HCI_TYPE_SDIO;
1323 
1324 	ret = rtw_core_init(rtwdev);
1325 	if (ret)
1326 		goto err_release_hw;
1327 
1328 	rtw_dbg(rtwdev, RTW_DBG_SDIO,
1329 		"rtw88 SDIO probe: vendor=0x%04x device=%04x class=%02x",
1330 		id->vendor, id->device, id->class);
1331 
1332 	ret = rtw_sdio_claim(rtwdev, sdio_func);
1333 	if (ret) {
1334 		rtw_err(rtwdev, "failed to claim SDIO device");
1335 		goto err_deinit_core;
1336 	}
1337 
1338 	rtw_sdio_init(rtwdev);
1339 
1340 	ret = rtw_sdio_init_tx(rtwdev);
1341 	if (ret) {
1342 		rtw_err(rtwdev, "failed to init SDIO TX queue\n");
1343 		goto err_sdio_declaim;
1344 	}
1345 
1346 	ret = rtw_chip_info_setup(rtwdev);
1347 	if (ret) {
1348 		rtw_err(rtwdev, "failed to setup chip information");
1349 		goto err_destroy_txwq;
1350 	}
1351 
1352 	ret = rtw_sdio_request_irq(rtwdev, sdio_func);
1353 	if (ret)
1354 		goto err_destroy_txwq;
1355 
1356 	ret = rtw_register_hw(rtwdev, hw);
1357 	if (ret) {
1358 		rtw_err(rtwdev, "failed to register hw");
1359 		goto err_free_irq;
1360 	}
1361 
1362 	return 0;
1363 
1364 err_free_irq:
1365 	rtw_sdio_free_irq(rtwdev, sdio_func);
1366 err_destroy_txwq:
1367 	rtw_sdio_deinit_tx(rtwdev);
1368 err_sdio_declaim:
1369 	rtw_sdio_declaim(rtwdev, sdio_func);
1370 err_deinit_core:
1371 	rtw_core_deinit(rtwdev);
1372 err_release_hw:
1373 	ieee80211_free_hw(hw);
1374 
1375 	return ret;
1376 }
1377 EXPORT_SYMBOL(rtw_sdio_probe);
1378 
rtw_sdio_remove(struct sdio_func * sdio_func)1379 void rtw_sdio_remove(struct sdio_func *sdio_func)
1380 {
1381 	struct ieee80211_hw *hw = sdio_get_drvdata(sdio_func);
1382 	struct rtw_dev *rtwdev;
1383 
1384 	if (!hw)
1385 		return;
1386 
1387 	rtwdev = hw->priv;
1388 
1389 	rtw_unregister_hw(rtwdev, hw);
1390 	rtw_sdio_disable_interrupt(rtwdev);
1391 	rtw_sdio_free_irq(rtwdev, sdio_func);
1392 	rtw_sdio_declaim(rtwdev, sdio_func);
1393 	rtw_sdio_deinit_tx(rtwdev);
1394 	rtw_core_deinit(rtwdev);
1395 	ieee80211_free_hw(hw);
1396 }
1397 EXPORT_SYMBOL(rtw_sdio_remove);
1398 
rtw_sdio_shutdown(struct device * dev)1399 void rtw_sdio_shutdown(struct device *dev)
1400 {
1401 	struct sdio_func *sdio_func = dev_to_sdio_func(dev);
1402 	const struct rtw_chip_info *chip;
1403 	struct ieee80211_hw *hw;
1404 	struct rtw_dev *rtwdev;
1405 
1406 	hw = sdio_get_drvdata(sdio_func);
1407 	if (!hw)
1408 		return;
1409 
1410 	rtwdev = hw->priv;
1411 	chip = rtwdev->chip;
1412 
1413 	if (chip->ops->shutdown)
1414 		chip->ops->shutdown(rtwdev);
1415 }
1416 EXPORT_SYMBOL(rtw_sdio_shutdown);
1417 
1418 MODULE_AUTHOR("Martin Blumenstingl");
1419 MODULE_AUTHOR("Jernej Skrabec");
1420 MODULE_DESCRIPTION("Realtek 802.11ac wireless SDIO driver");
1421 MODULE_LICENSE("Dual BSD/GPL");
1422