xref: /openbmc/linux/drivers/net/ethernet/nxp/lpc_eth.c (revision ac8b6f14)
1 /*
2  * drivers/net/ethernet/nxp/lpc_eth.c
3  *
4  * Author: Kevin Wells <kevin.wells@nxp.com>
5  *
6  * Copyright (C) 2010 NXP Semiconductors
7  * Copyright (C) 2012 Roland Stigge <stigge@antcom.de>
8  *
9  * This program is free software; you can redistribute it and/or modify
10  * it under the terms of the GNU General Public License as published by
11  * the Free Software Foundation; either version 2 of the License, or
12  * (at your option) any later version.
13  *
14  * This program is distributed in the hope that it will be useful,
15  * but WITHOUT ANY WARRANTY; without even the implied warranty of
16  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
17  * GNU General Public License for more details.
18  */
19 
20 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
21 
22 #include <linux/clk.h>
23 #include <linux/crc32.h>
24 #include <linux/etherdevice.h>
25 #include <linux/module.h>
26 #include <linux/of_net.h>
27 #include <linux/phy.h>
28 #include <linux/platform_device.h>
29 #include <linux/spinlock.h>
30 
31 #include <mach/board.h>
32 #include <mach/hardware.h>
33 #include <mach/platform.h>
34 
35 #define MODNAME "lpc-eth"
36 #define DRV_VERSION "1.00"
37 
38 #define ENET_MAXF_SIZE 1536
39 #define ENET_RX_DESC 48
40 #define ENET_TX_DESC 16
41 
42 #define NAPI_WEIGHT 16
43 
44 /*
45  * Ethernet MAC controller Register offsets
46  */
47 #define LPC_ENET_MAC1(x)			(x + 0x000)
48 #define LPC_ENET_MAC2(x)			(x + 0x004)
49 #define LPC_ENET_IPGT(x)			(x + 0x008)
50 #define LPC_ENET_IPGR(x)			(x + 0x00C)
51 #define LPC_ENET_CLRT(x)			(x + 0x010)
52 #define LPC_ENET_MAXF(x)			(x + 0x014)
53 #define LPC_ENET_SUPP(x)			(x + 0x018)
54 #define LPC_ENET_TEST(x)			(x + 0x01C)
55 #define LPC_ENET_MCFG(x)			(x + 0x020)
56 #define LPC_ENET_MCMD(x)			(x + 0x024)
57 #define LPC_ENET_MADR(x)			(x + 0x028)
58 #define LPC_ENET_MWTD(x)			(x + 0x02C)
59 #define LPC_ENET_MRDD(x)			(x + 0x030)
60 #define LPC_ENET_MIND(x)			(x + 0x034)
61 #define LPC_ENET_SA0(x)				(x + 0x040)
62 #define LPC_ENET_SA1(x)				(x + 0x044)
63 #define LPC_ENET_SA2(x)				(x + 0x048)
64 #define LPC_ENET_COMMAND(x)			(x + 0x100)
65 #define LPC_ENET_STATUS(x)			(x + 0x104)
66 #define LPC_ENET_RXDESCRIPTOR(x)		(x + 0x108)
67 #define LPC_ENET_RXSTATUS(x)			(x + 0x10C)
68 #define LPC_ENET_RXDESCRIPTORNUMBER(x)		(x + 0x110)
69 #define LPC_ENET_RXPRODUCEINDEX(x)		(x + 0x114)
70 #define LPC_ENET_RXCONSUMEINDEX(x)		(x + 0x118)
71 #define LPC_ENET_TXDESCRIPTOR(x)		(x + 0x11C)
72 #define LPC_ENET_TXSTATUS(x)			(x + 0x120)
73 #define LPC_ENET_TXDESCRIPTORNUMBER(x)		(x + 0x124)
74 #define LPC_ENET_TXPRODUCEINDEX(x)		(x + 0x128)
75 #define LPC_ENET_TXCONSUMEINDEX(x)		(x + 0x12C)
76 #define LPC_ENET_TSV0(x)			(x + 0x158)
77 #define LPC_ENET_TSV1(x)			(x + 0x15C)
78 #define LPC_ENET_RSV(x)				(x + 0x160)
79 #define LPC_ENET_FLOWCONTROLCOUNTER(x)		(x + 0x170)
80 #define LPC_ENET_FLOWCONTROLSTATUS(x)		(x + 0x174)
81 #define LPC_ENET_RXFILTER_CTRL(x)		(x + 0x200)
82 #define LPC_ENET_RXFILTERWOLSTATUS(x)		(x + 0x204)
83 #define LPC_ENET_RXFILTERWOLCLEAR(x)		(x + 0x208)
84 #define LPC_ENET_HASHFILTERL(x)			(x + 0x210)
85 #define LPC_ENET_HASHFILTERH(x)			(x + 0x214)
86 #define LPC_ENET_INTSTATUS(x)			(x + 0xFE0)
87 #define LPC_ENET_INTENABLE(x)			(x + 0xFE4)
88 #define LPC_ENET_INTCLEAR(x)			(x + 0xFE8)
89 #define LPC_ENET_INTSET(x)			(x + 0xFEC)
90 #define LPC_ENET_POWERDOWN(x)			(x + 0xFF4)
91 
92 /*
93  * mac1 register definitions
94  */
95 #define LPC_MAC1_RECV_ENABLE			(1 << 0)
96 #define LPC_MAC1_PASS_ALL_RX_FRAMES		(1 << 1)
97 #define LPC_MAC1_RX_FLOW_CONTROL		(1 << 2)
98 #define LPC_MAC1_TX_FLOW_CONTROL		(1 << 3)
99 #define LPC_MAC1_LOOPBACK			(1 << 4)
100 #define LPC_MAC1_RESET_TX			(1 << 8)
101 #define LPC_MAC1_RESET_MCS_TX			(1 << 9)
102 #define LPC_MAC1_RESET_RX			(1 << 10)
103 #define LPC_MAC1_RESET_MCS_RX			(1 << 11)
104 #define LPC_MAC1_SIMULATION_RESET		(1 << 14)
105 #define LPC_MAC1_SOFT_RESET			(1 << 15)
106 
107 /*
108  * mac2 register definitions
109  */
110 #define LPC_MAC2_FULL_DUPLEX			(1 << 0)
111 #define LPC_MAC2_FRAME_LENGTH_CHECKING		(1 << 1)
112 #define LPC_MAC2_HUGH_LENGTH_CHECKING		(1 << 2)
113 #define LPC_MAC2_DELAYED_CRC			(1 << 3)
114 #define LPC_MAC2_CRC_ENABLE			(1 << 4)
115 #define LPC_MAC2_PAD_CRC_ENABLE			(1 << 5)
116 #define LPC_MAC2_VLAN_PAD_ENABLE		(1 << 6)
117 #define LPC_MAC2_AUTO_DETECT_PAD_ENABLE		(1 << 7)
118 #define LPC_MAC2_PURE_PREAMBLE_ENFORCEMENT	(1 << 8)
119 #define LPC_MAC2_LONG_PREAMBLE_ENFORCEMENT	(1 << 9)
120 #define LPC_MAC2_NO_BACKOFF			(1 << 12)
121 #define LPC_MAC2_BACK_PRESSURE			(1 << 13)
122 #define LPC_MAC2_EXCESS_DEFER			(1 << 14)
123 
124 /*
125  * ipgt register definitions
126  */
127 #define LPC_IPGT_LOAD(n)			((n) & 0x7F)
128 
129 /*
130  * ipgr register definitions
131  */
132 #define LPC_IPGR_LOAD_PART2(n)			((n) & 0x7F)
133 #define LPC_IPGR_LOAD_PART1(n)			(((n) & 0x7F) << 8)
134 
135 /*
136  * clrt register definitions
137  */
138 #define LPC_CLRT_LOAD_RETRY_MAX(n)		((n) & 0xF)
139 #define LPC_CLRT_LOAD_COLLISION_WINDOW(n)	(((n) & 0x3F) << 8)
140 
141 /*
142  * maxf register definitions
143  */
144 #define LPC_MAXF_LOAD_MAX_FRAME_LEN(n)		((n) & 0xFFFF)
145 
146 /*
147  * supp register definitions
148  */
149 #define LPC_SUPP_SPEED				(1 << 8)
150 #define LPC_SUPP_RESET_RMII			(1 << 11)
151 
152 /*
153  * test register definitions
154  */
155 #define LPC_TEST_SHORTCUT_PAUSE_QUANTA		(1 << 0)
156 #define LPC_TEST_PAUSE				(1 << 1)
157 #define LPC_TEST_BACKPRESSURE			(1 << 2)
158 
159 /*
160  * mcfg register definitions
161  */
162 #define LPC_MCFG_SCAN_INCREMENT			(1 << 0)
163 #define LPC_MCFG_SUPPRESS_PREAMBLE		(1 << 1)
164 #define LPC_MCFG_CLOCK_SELECT(n)		(((n) & 0x7) << 2)
165 #define LPC_MCFG_CLOCK_HOST_DIV_4		0
166 #define LPC_MCFG_CLOCK_HOST_DIV_6		2
167 #define LPC_MCFG_CLOCK_HOST_DIV_8		3
168 #define LPC_MCFG_CLOCK_HOST_DIV_10		4
169 #define LPC_MCFG_CLOCK_HOST_DIV_14		5
170 #define LPC_MCFG_CLOCK_HOST_DIV_20		6
171 #define LPC_MCFG_CLOCK_HOST_DIV_28		7
172 #define LPC_MCFG_RESET_MII_MGMT			(1 << 15)
173 
174 /*
175  * mcmd register definitions
176  */
177 #define LPC_MCMD_READ				(1 << 0)
178 #define LPC_MCMD_SCAN				(1 << 1)
179 
180 /*
181  * madr register definitions
182  */
183 #define LPC_MADR_REGISTER_ADDRESS(n)		((n) & 0x1F)
184 #define LPC_MADR_PHY_0ADDRESS(n)		(((n) & 0x1F) << 8)
185 
186 /*
187  * mwtd register definitions
188  */
189 #define LPC_MWDT_WRITE(n)			((n) & 0xFFFF)
190 
191 /*
192  * mrdd register definitions
193  */
194 #define LPC_MRDD_READ_MASK			0xFFFF
195 
196 /*
197  * mind register definitions
198  */
199 #define LPC_MIND_BUSY				(1 << 0)
200 #define LPC_MIND_SCANNING			(1 << 1)
201 #define LPC_MIND_NOT_VALID			(1 << 2)
202 #define LPC_MIND_MII_LINK_FAIL			(1 << 3)
203 
204 /*
205  * command register definitions
206  */
207 #define LPC_COMMAND_RXENABLE			(1 << 0)
208 #define LPC_COMMAND_TXENABLE			(1 << 1)
209 #define LPC_COMMAND_REG_RESET			(1 << 3)
210 #define LPC_COMMAND_TXRESET			(1 << 4)
211 #define LPC_COMMAND_RXRESET			(1 << 5)
212 #define LPC_COMMAND_PASSRUNTFRAME		(1 << 6)
213 #define LPC_COMMAND_PASSRXFILTER		(1 << 7)
214 #define LPC_COMMAND_TXFLOWCONTROL		(1 << 8)
215 #define LPC_COMMAND_RMII			(1 << 9)
216 #define LPC_COMMAND_FULLDUPLEX			(1 << 10)
217 
218 /*
219  * status register definitions
220  */
221 #define LPC_STATUS_RXACTIVE			(1 << 0)
222 #define LPC_STATUS_TXACTIVE			(1 << 1)
223 
224 /*
225  * tsv0 register definitions
226  */
227 #define LPC_TSV0_CRC_ERROR			(1 << 0)
228 #define LPC_TSV0_LENGTH_CHECK_ERROR		(1 << 1)
229 #define LPC_TSV0_LENGTH_OUT_OF_RANGE		(1 << 2)
230 #define LPC_TSV0_DONE				(1 << 3)
231 #define LPC_TSV0_MULTICAST			(1 << 4)
232 #define LPC_TSV0_BROADCAST			(1 << 5)
233 #define LPC_TSV0_PACKET_DEFER			(1 << 6)
234 #define LPC_TSV0_ESCESSIVE_DEFER		(1 << 7)
235 #define LPC_TSV0_ESCESSIVE_COLLISION		(1 << 8)
236 #define LPC_TSV0_LATE_COLLISION			(1 << 9)
237 #define LPC_TSV0_GIANT				(1 << 10)
238 #define LPC_TSV0_UNDERRUN			(1 << 11)
239 #define LPC_TSV0_TOTAL_BYTES(n)			(((n) >> 12) & 0xFFFF)
240 #define LPC_TSV0_CONTROL_FRAME			(1 << 28)
241 #define LPC_TSV0_PAUSE				(1 << 29)
242 #define LPC_TSV0_BACKPRESSURE			(1 << 30)
243 #define LPC_TSV0_VLAN				(1 << 31)
244 
245 /*
246  * tsv1 register definitions
247  */
248 #define LPC_TSV1_TRANSMIT_BYTE_COUNT(n)		((n) & 0xFFFF)
249 #define LPC_TSV1_COLLISION_COUNT(n)		(((n) >> 16) & 0xF)
250 
251 /*
252  * rsv register definitions
253  */
254 #define LPC_RSV_RECEIVED_BYTE_COUNT(n)		((n) & 0xFFFF)
255 #define LPC_RSV_RXDV_EVENT_IGNORED		(1 << 16)
256 #define LPC_RSV_RXDV_EVENT_PREVIOUSLY_SEEN	(1 << 17)
257 #define LPC_RSV_CARRIER_EVNT_PREVIOUS_SEEN	(1 << 18)
258 #define LPC_RSV_RECEIVE_CODE_VIOLATION		(1 << 19)
259 #define LPC_RSV_CRC_ERROR			(1 << 20)
260 #define LPC_RSV_LENGTH_CHECK_ERROR		(1 << 21)
261 #define LPC_RSV_LENGTH_OUT_OF_RANGE		(1 << 22)
262 #define LPC_RSV_RECEIVE_OK			(1 << 23)
263 #define LPC_RSV_MULTICAST			(1 << 24)
264 #define LPC_RSV_BROADCAST			(1 << 25)
265 #define LPC_RSV_DRIBBLE_NIBBLE			(1 << 26)
266 #define LPC_RSV_CONTROL_FRAME			(1 << 27)
267 #define LPC_RSV_PAUSE				(1 << 28)
268 #define LPC_RSV_UNSUPPORTED_OPCODE		(1 << 29)
269 #define LPC_RSV_VLAN				(1 << 30)
270 
271 /*
272  * flowcontrolcounter register definitions
273  */
274 #define LPC_FCCR_MIRRORCOUNTER(n)		((n) & 0xFFFF)
275 #define LPC_FCCR_PAUSETIMER(n)			(((n) >> 16) & 0xFFFF)
276 
277 /*
278  * flowcontrolstatus register definitions
279  */
280 #define LPC_FCCR_MIRRORCOUNTERCURRENT(n)	((n) & 0xFFFF)
281 
282 /*
283  * rxfliterctrl, rxfilterwolstatus, and rxfilterwolclear shared
284  * register definitions
285  */
286 #define LPC_RXFLTRW_ACCEPTUNICAST		(1 << 0)
287 #define LPC_RXFLTRW_ACCEPTUBROADCAST		(1 << 1)
288 #define LPC_RXFLTRW_ACCEPTUMULTICAST		(1 << 2)
289 #define LPC_RXFLTRW_ACCEPTUNICASTHASH		(1 << 3)
290 #define LPC_RXFLTRW_ACCEPTUMULTICASTHASH	(1 << 4)
291 #define LPC_RXFLTRW_ACCEPTPERFECT		(1 << 5)
292 
293 /*
294  * rxfliterctrl register definitions
295  */
296 #define LPC_RXFLTRWSTS_MAGICPACKETENWOL		(1 << 12)
297 #define LPC_RXFLTRWSTS_RXFILTERENWOL		(1 << 13)
298 
299 /*
300  * rxfilterwolstatus/rxfilterwolclear register definitions
301  */
302 #define LPC_RXFLTRWSTS_RXFILTERWOL		(1 << 7)
303 #define LPC_RXFLTRWSTS_MAGICPACKETWOL		(1 << 8)
304 
305 /*
306  * intstatus, intenable, intclear, and Intset shared register
307  * definitions
308  */
309 #define LPC_MACINT_RXOVERRUNINTEN		(1 << 0)
310 #define LPC_MACINT_RXERRORONINT			(1 << 1)
311 #define LPC_MACINT_RXFINISHEDINTEN		(1 << 2)
312 #define LPC_MACINT_RXDONEINTEN			(1 << 3)
313 #define LPC_MACINT_TXUNDERRUNINTEN		(1 << 4)
314 #define LPC_MACINT_TXERRORINTEN			(1 << 5)
315 #define LPC_MACINT_TXFINISHEDINTEN		(1 << 6)
316 #define LPC_MACINT_TXDONEINTEN			(1 << 7)
317 #define LPC_MACINT_SOFTINTEN			(1 << 12)
318 #define LPC_MACINT_WAKEUPINTEN			(1 << 13)
319 
320 /*
321  * powerdown register definitions
322  */
323 #define LPC_POWERDOWN_MACAHB			(1 << 31)
324 
325 static phy_interface_t lpc_phy_interface_mode(struct device *dev)
326 {
327 	if (dev && dev->of_node) {
328 		const char *mode = of_get_property(dev->of_node,
329 						   "phy-mode", NULL);
330 		if (mode && !strcmp(mode, "mii"))
331 			return PHY_INTERFACE_MODE_MII;
332 	}
333 	return PHY_INTERFACE_MODE_RMII;
334 }
335 
336 static bool use_iram_for_net(struct device *dev)
337 {
338 	if (dev && dev->of_node)
339 		return of_property_read_bool(dev->of_node, "use-iram");
340 	return false;
341 }
342 
343 /* Receive Status information word */
344 #define RXSTATUS_SIZE			0x000007FF
345 #define RXSTATUS_CONTROL		(1 << 18)
346 #define RXSTATUS_VLAN			(1 << 19)
347 #define RXSTATUS_FILTER			(1 << 20)
348 #define RXSTATUS_MULTICAST		(1 << 21)
349 #define RXSTATUS_BROADCAST		(1 << 22)
350 #define RXSTATUS_CRC			(1 << 23)
351 #define RXSTATUS_SYMBOL			(1 << 24)
352 #define RXSTATUS_LENGTH			(1 << 25)
353 #define RXSTATUS_RANGE			(1 << 26)
354 #define RXSTATUS_ALIGN			(1 << 27)
355 #define RXSTATUS_OVERRUN		(1 << 28)
356 #define RXSTATUS_NODESC			(1 << 29)
357 #define RXSTATUS_LAST			(1 << 30)
358 #define RXSTATUS_ERROR			(1 << 31)
359 
360 #define RXSTATUS_STATUS_ERROR \
361 	(RXSTATUS_NODESC | RXSTATUS_OVERRUN | RXSTATUS_ALIGN | \
362 	 RXSTATUS_RANGE | RXSTATUS_LENGTH | RXSTATUS_SYMBOL | RXSTATUS_CRC)
363 
364 /* Receive Descriptor control word */
365 #define RXDESC_CONTROL_SIZE		0x000007FF
366 #define RXDESC_CONTROL_INT		(1 << 31)
367 
368 /* Transmit Status information word */
369 #define TXSTATUS_COLLISIONS_GET(x)	(((x) >> 21) & 0xF)
370 #define TXSTATUS_DEFER			(1 << 25)
371 #define TXSTATUS_EXCESSDEFER		(1 << 26)
372 #define TXSTATUS_EXCESSCOLL		(1 << 27)
373 #define TXSTATUS_LATECOLL		(1 << 28)
374 #define TXSTATUS_UNDERRUN		(1 << 29)
375 #define TXSTATUS_NODESC			(1 << 30)
376 #define TXSTATUS_ERROR			(1 << 31)
377 
378 /* Transmit Descriptor control word */
379 #define TXDESC_CONTROL_SIZE		0x000007FF
380 #define TXDESC_CONTROL_OVERRIDE		(1 << 26)
381 #define TXDESC_CONTROL_HUGE		(1 << 27)
382 #define TXDESC_CONTROL_PAD		(1 << 28)
383 #define TXDESC_CONTROL_CRC		(1 << 29)
384 #define TXDESC_CONTROL_LAST		(1 << 30)
385 #define TXDESC_CONTROL_INT		(1 << 31)
386 
387 /*
388  * Structure of a TX/RX descriptors and RX status
389  */
390 struct txrx_desc_t {
391 	__le32 packet;
392 	__le32 control;
393 };
394 struct rx_status_t {
395 	__le32 statusinfo;
396 	__le32 statushashcrc;
397 };
398 
399 /*
400  * Device driver data structure
401  */
402 struct netdata_local {
403 	struct platform_device	*pdev;
404 	struct net_device	*ndev;
405 	spinlock_t		lock;
406 	void __iomem		*net_base;
407 	u32			msg_enable;
408 	unsigned int		skblen[ENET_TX_DESC];
409 	unsigned int		last_tx_idx;
410 	unsigned int		num_used_tx_buffs;
411 	struct mii_bus		*mii_bus;
412 	struct clk		*clk;
413 	dma_addr_t		dma_buff_base_p;
414 	void			*dma_buff_base_v;
415 	size_t			dma_buff_size;
416 	struct txrx_desc_t	*tx_desc_v;
417 	u32			*tx_stat_v;
418 	void			*tx_buff_v;
419 	struct txrx_desc_t	*rx_desc_v;
420 	struct rx_status_t	*rx_stat_v;
421 	void			*rx_buff_v;
422 	int			link;
423 	int			speed;
424 	int			duplex;
425 	struct napi_struct	napi;
426 };
427 
428 /*
429  * MAC support functions
430  */
431 static void __lpc_set_mac(struct netdata_local *pldat, u8 *mac)
432 {
433 	u32 tmp;
434 
435 	/* Set station address */
436 	tmp = mac[0] | ((u32)mac[1] << 8);
437 	writel(tmp, LPC_ENET_SA2(pldat->net_base));
438 	tmp = mac[2] | ((u32)mac[3] << 8);
439 	writel(tmp, LPC_ENET_SA1(pldat->net_base));
440 	tmp = mac[4] | ((u32)mac[5] << 8);
441 	writel(tmp, LPC_ENET_SA0(pldat->net_base));
442 
443 	netdev_dbg(pldat->ndev, "Ethernet MAC address %pM\n", mac);
444 }
445 
446 static void __lpc_get_mac(struct netdata_local *pldat, u8 *mac)
447 {
448 	u32 tmp;
449 
450 	/* Get station address */
451 	tmp = readl(LPC_ENET_SA2(pldat->net_base));
452 	mac[0] = tmp & 0xFF;
453 	mac[1] = tmp >> 8;
454 	tmp = readl(LPC_ENET_SA1(pldat->net_base));
455 	mac[2] = tmp & 0xFF;
456 	mac[3] = tmp >> 8;
457 	tmp = readl(LPC_ENET_SA0(pldat->net_base));
458 	mac[4] = tmp & 0xFF;
459 	mac[5] = tmp >> 8;
460 }
461 
462 static void __lpc_params_setup(struct netdata_local *pldat)
463 {
464 	u32 tmp;
465 
466 	if (pldat->duplex == DUPLEX_FULL) {
467 		tmp = readl(LPC_ENET_MAC2(pldat->net_base));
468 		tmp |= LPC_MAC2_FULL_DUPLEX;
469 		writel(tmp, LPC_ENET_MAC2(pldat->net_base));
470 		tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
471 		tmp |= LPC_COMMAND_FULLDUPLEX;
472 		writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
473 		writel(LPC_IPGT_LOAD(0x15), LPC_ENET_IPGT(pldat->net_base));
474 	} else {
475 		tmp = readl(LPC_ENET_MAC2(pldat->net_base));
476 		tmp &= ~LPC_MAC2_FULL_DUPLEX;
477 		writel(tmp, LPC_ENET_MAC2(pldat->net_base));
478 		tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
479 		tmp &= ~LPC_COMMAND_FULLDUPLEX;
480 		writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
481 		writel(LPC_IPGT_LOAD(0x12), LPC_ENET_IPGT(pldat->net_base));
482 	}
483 
484 	if (pldat->speed == SPEED_100)
485 		writel(LPC_SUPP_SPEED, LPC_ENET_SUPP(pldat->net_base));
486 	else
487 		writel(0, LPC_ENET_SUPP(pldat->net_base));
488 }
489 
490 static void __lpc_eth_reset(struct netdata_local *pldat)
491 {
492 	/* Reset all MAC logic */
493 	writel((LPC_MAC1_RESET_TX | LPC_MAC1_RESET_MCS_TX | LPC_MAC1_RESET_RX |
494 		LPC_MAC1_RESET_MCS_RX | LPC_MAC1_SIMULATION_RESET |
495 		LPC_MAC1_SOFT_RESET), LPC_ENET_MAC1(pldat->net_base));
496 	writel((LPC_COMMAND_REG_RESET | LPC_COMMAND_TXRESET |
497 		LPC_COMMAND_RXRESET), LPC_ENET_COMMAND(pldat->net_base));
498 }
499 
500 static int __lpc_mii_mngt_reset(struct netdata_local *pldat)
501 {
502 	/* Reset MII management hardware */
503 	writel(LPC_MCFG_RESET_MII_MGMT, LPC_ENET_MCFG(pldat->net_base));
504 
505 	/* Setup MII clock to slowest rate with a /28 divider */
506 	writel(LPC_MCFG_CLOCK_SELECT(LPC_MCFG_CLOCK_HOST_DIV_28),
507 	       LPC_ENET_MCFG(pldat->net_base));
508 
509 	return 0;
510 }
511 
512 static inline phys_addr_t __va_to_pa(void *addr, struct netdata_local *pldat)
513 {
514 	phys_addr_t phaddr;
515 
516 	phaddr = addr - pldat->dma_buff_base_v;
517 	phaddr += pldat->dma_buff_base_p;
518 
519 	return phaddr;
520 }
521 
522 static void lpc_eth_enable_int(void __iomem *regbase)
523 {
524 	writel((LPC_MACINT_RXDONEINTEN | LPC_MACINT_TXDONEINTEN),
525 	       LPC_ENET_INTENABLE(regbase));
526 }
527 
528 static void lpc_eth_disable_int(void __iomem *regbase)
529 {
530 	writel(0, LPC_ENET_INTENABLE(regbase));
531 }
532 
533 /* Setup TX/RX descriptors */
534 static void __lpc_txrx_desc_setup(struct netdata_local *pldat)
535 {
536 	u32 *ptxstat;
537 	void *tbuff;
538 	int i;
539 	struct txrx_desc_t *ptxrxdesc;
540 	struct rx_status_t *prxstat;
541 
542 	tbuff = PTR_ALIGN(pldat->dma_buff_base_v, 16);
543 
544 	/* Setup TX descriptors, status, and buffers */
545 	pldat->tx_desc_v = tbuff;
546 	tbuff += sizeof(struct txrx_desc_t) * ENET_TX_DESC;
547 
548 	pldat->tx_stat_v = tbuff;
549 	tbuff += sizeof(u32) * ENET_TX_DESC;
550 
551 	tbuff = PTR_ALIGN(tbuff, 16);
552 	pldat->tx_buff_v = tbuff;
553 	tbuff += ENET_MAXF_SIZE * ENET_TX_DESC;
554 
555 	/* Setup RX descriptors, status, and buffers */
556 	pldat->rx_desc_v = tbuff;
557 	tbuff += sizeof(struct txrx_desc_t) * ENET_RX_DESC;
558 
559 	tbuff = PTR_ALIGN(tbuff, 16);
560 	pldat->rx_stat_v = tbuff;
561 	tbuff += sizeof(struct rx_status_t) * ENET_RX_DESC;
562 
563 	tbuff = PTR_ALIGN(tbuff, 16);
564 	pldat->rx_buff_v = tbuff;
565 	tbuff += ENET_MAXF_SIZE * ENET_RX_DESC;
566 
567 	/* Map the TX descriptors to the TX buffers in hardware */
568 	for (i = 0; i < ENET_TX_DESC; i++) {
569 		ptxstat = &pldat->tx_stat_v[i];
570 		ptxrxdesc = &pldat->tx_desc_v[i];
571 
572 		ptxrxdesc->packet = __va_to_pa(
573 				pldat->tx_buff_v + i * ENET_MAXF_SIZE, pldat);
574 		ptxrxdesc->control = 0;
575 		*ptxstat = 0;
576 	}
577 
578 	/* Map the RX descriptors to the RX buffers in hardware */
579 	for (i = 0; i < ENET_RX_DESC; i++) {
580 		prxstat = &pldat->rx_stat_v[i];
581 		ptxrxdesc = &pldat->rx_desc_v[i];
582 
583 		ptxrxdesc->packet = __va_to_pa(
584 				pldat->rx_buff_v + i * ENET_MAXF_SIZE, pldat);
585 		ptxrxdesc->control = RXDESC_CONTROL_INT | (ENET_MAXF_SIZE - 1);
586 		prxstat->statusinfo = 0;
587 		prxstat->statushashcrc = 0;
588 	}
589 
590 	/* Setup base addresses in hardware to point to buffers and
591 	 * descriptors
592 	 */
593 	writel((ENET_TX_DESC - 1),
594 	       LPC_ENET_TXDESCRIPTORNUMBER(pldat->net_base));
595 	writel(__va_to_pa(pldat->tx_desc_v, pldat),
596 	       LPC_ENET_TXDESCRIPTOR(pldat->net_base));
597 	writel(__va_to_pa(pldat->tx_stat_v, pldat),
598 	       LPC_ENET_TXSTATUS(pldat->net_base));
599 	writel((ENET_RX_DESC - 1),
600 	       LPC_ENET_RXDESCRIPTORNUMBER(pldat->net_base));
601 	writel(__va_to_pa(pldat->rx_desc_v, pldat),
602 	       LPC_ENET_RXDESCRIPTOR(pldat->net_base));
603 	writel(__va_to_pa(pldat->rx_stat_v, pldat),
604 	       LPC_ENET_RXSTATUS(pldat->net_base));
605 }
606 
607 static void __lpc_eth_init(struct netdata_local *pldat)
608 {
609 	u32 tmp;
610 
611 	/* Disable controller and reset */
612 	tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
613 	tmp &= ~LPC_COMMAND_RXENABLE | LPC_COMMAND_TXENABLE;
614 	writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
615 	tmp = readl(LPC_ENET_MAC1(pldat->net_base));
616 	tmp &= ~LPC_MAC1_RECV_ENABLE;
617 	writel(tmp, LPC_ENET_MAC1(pldat->net_base));
618 
619 	/* Initial MAC setup */
620 	writel(LPC_MAC1_PASS_ALL_RX_FRAMES, LPC_ENET_MAC1(pldat->net_base));
621 	writel((LPC_MAC2_PAD_CRC_ENABLE | LPC_MAC2_CRC_ENABLE),
622 	       LPC_ENET_MAC2(pldat->net_base));
623 	writel(ENET_MAXF_SIZE, LPC_ENET_MAXF(pldat->net_base));
624 
625 	/* Collision window, gap */
626 	writel((LPC_CLRT_LOAD_RETRY_MAX(0xF) |
627 		LPC_CLRT_LOAD_COLLISION_WINDOW(0x37)),
628 	       LPC_ENET_CLRT(pldat->net_base));
629 	writel(LPC_IPGR_LOAD_PART2(0x12), LPC_ENET_IPGR(pldat->net_base));
630 
631 	if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII)
632 		writel(LPC_COMMAND_PASSRUNTFRAME,
633 		       LPC_ENET_COMMAND(pldat->net_base));
634 	else {
635 		writel((LPC_COMMAND_PASSRUNTFRAME | LPC_COMMAND_RMII),
636 		       LPC_ENET_COMMAND(pldat->net_base));
637 		writel(LPC_SUPP_RESET_RMII, LPC_ENET_SUPP(pldat->net_base));
638 	}
639 
640 	__lpc_params_setup(pldat);
641 
642 	/* Setup TX and RX descriptors */
643 	__lpc_txrx_desc_setup(pldat);
644 
645 	/* Setup packet filtering */
646 	writel((LPC_RXFLTRW_ACCEPTUBROADCAST | LPC_RXFLTRW_ACCEPTPERFECT),
647 	       LPC_ENET_RXFILTER_CTRL(pldat->net_base));
648 
649 	/* Get the next TX buffer output index */
650 	pldat->num_used_tx_buffs = 0;
651 	pldat->last_tx_idx =
652 		readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base));
653 
654 	/* Clear and enable interrupts */
655 	writel(0xFFFF, LPC_ENET_INTCLEAR(pldat->net_base));
656 	smp_wmb();
657 	lpc_eth_enable_int(pldat->net_base);
658 
659 	/* Enable controller */
660 	tmp = readl(LPC_ENET_COMMAND(pldat->net_base));
661 	tmp |= LPC_COMMAND_RXENABLE | LPC_COMMAND_TXENABLE;
662 	writel(tmp, LPC_ENET_COMMAND(pldat->net_base));
663 	tmp = readl(LPC_ENET_MAC1(pldat->net_base));
664 	tmp |= LPC_MAC1_RECV_ENABLE;
665 	writel(tmp, LPC_ENET_MAC1(pldat->net_base));
666 }
667 
668 static void __lpc_eth_shutdown(struct netdata_local *pldat)
669 {
670 	/* Reset ethernet and power down PHY */
671 	__lpc_eth_reset(pldat);
672 	writel(0, LPC_ENET_MAC1(pldat->net_base));
673 	writel(0, LPC_ENET_MAC2(pldat->net_base));
674 }
675 
676 /*
677  * MAC<--->PHY support functions
678  */
679 static int lpc_mdio_read(struct mii_bus *bus, int phy_id, int phyreg)
680 {
681 	struct netdata_local *pldat = bus->priv;
682 	unsigned long timeout = jiffies + msecs_to_jiffies(100);
683 	int lps;
684 
685 	writel(((phy_id << 8) | phyreg), LPC_ENET_MADR(pldat->net_base));
686 	writel(LPC_MCMD_READ, LPC_ENET_MCMD(pldat->net_base));
687 
688 	/* Wait for unbusy status */
689 	while (readl(LPC_ENET_MIND(pldat->net_base)) & LPC_MIND_BUSY) {
690 		if (time_after(jiffies, timeout))
691 			return -EIO;
692 		cpu_relax();
693 	}
694 
695 	lps = readl(LPC_ENET_MRDD(pldat->net_base));
696 	writel(0, LPC_ENET_MCMD(pldat->net_base));
697 
698 	return lps;
699 }
700 
701 static int lpc_mdio_write(struct mii_bus *bus, int phy_id, int phyreg,
702 			u16 phydata)
703 {
704 	struct netdata_local *pldat = bus->priv;
705 	unsigned long timeout = jiffies + msecs_to_jiffies(100);
706 
707 	writel(((phy_id << 8) | phyreg), LPC_ENET_MADR(pldat->net_base));
708 	writel(phydata, LPC_ENET_MWTD(pldat->net_base));
709 
710 	/* Wait for completion */
711 	while (readl(LPC_ENET_MIND(pldat->net_base)) & LPC_MIND_BUSY) {
712 		if (time_after(jiffies, timeout))
713 			return -EIO;
714 		cpu_relax();
715 	}
716 
717 	return 0;
718 }
719 
720 static int lpc_mdio_reset(struct mii_bus *bus)
721 {
722 	return __lpc_mii_mngt_reset((struct netdata_local *)bus->priv);
723 }
724 
725 static void lpc_handle_link_change(struct net_device *ndev)
726 {
727 	struct netdata_local *pldat = netdev_priv(ndev);
728 	struct phy_device *phydev = ndev->phydev;
729 	unsigned long flags;
730 
731 	bool status_change = false;
732 
733 	spin_lock_irqsave(&pldat->lock, flags);
734 
735 	if (phydev->link) {
736 		if ((pldat->speed != phydev->speed) ||
737 		    (pldat->duplex != phydev->duplex)) {
738 			pldat->speed = phydev->speed;
739 			pldat->duplex = phydev->duplex;
740 			status_change = true;
741 		}
742 	}
743 
744 	if (phydev->link != pldat->link) {
745 		if (!phydev->link) {
746 			pldat->speed = 0;
747 			pldat->duplex = -1;
748 		}
749 		pldat->link = phydev->link;
750 
751 		status_change = true;
752 	}
753 
754 	spin_unlock_irqrestore(&pldat->lock, flags);
755 
756 	if (status_change)
757 		__lpc_params_setup(pldat);
758 }
759 
760 static int lpc_mii_probe(struct net_device *ndev)
761 {
762 	struct netdata_local *pldat = netdev_priv(ndev);
763 	struct phy_device *phydev = phy_find_first(pldat->mii_bus);
764 
765 	if (!phydev) {
766 		netdev_err(ndev, "no PHY found\n");
767 		return -ENODEV;
768 	}
769 
770 	/* Attach to the PHY */
771 	if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII)
772 		netdev_info(ndev, "using MII interface\n");
773 	else
774 		netdev_info(ndev, "using RMII interface\n");
775 	phydev = phy_connect(ndev, phydev_name(phydev),
776 			     &lpc_handle_link_change,
777 			     lpc_phy_interface_mode(&pldat->pdev->dev));
778 
779 	if (IS_ERR(phydev)) {
780 		netdev_err(ndev, "Could not attach to PHY\n");
781 		return PTR_ERR(phydev);
782 	}
783 
784 	phy_set_max_speed(phydev, SPEED_100);
785 
786 	phydev->advertising = phydev->supported;
787 
788 	pldat->link = 0;
789 	pldat->speed = 0;
790 	pldat->duplex = -1;
791 
792 	phy_attached_info(phydev);
793 
794 	return 0;
795 }
796 
797 static int lpc_mii_init(struct netdata_local *pldat)
798 {
799 	int err = -ENXIO;
800 
801 	pldat->mii_bus = mdiobus_alloc();
802 	if (!pldat->mii_bus) {
803 		err = -ENOMEM;
804 		goto err_out;
805 	}
806 
807 	/* Setup MII mode */
808 	if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII)
809 		writel(LPC_COMMAND_PASSRUNTFRAME,
810 		       LPC_ENET_COMMAND(pldat->net_base));
811 	else {
812 		writel((LPC_COMMAND_PASSRUNTFRAME | LPC_COMMAND_RMII),
813 		       LPC_ENET_COMMAND(pldat->net_base));
814 		writel(LPC_SUPP_RESET_RMII, LPC_ENET_SUPP(pldat->net_base));
815 	}
816 
817 	pldat->mii_bus->name = "lpc_mii_bus";
818 	pldat->mii_bus->read = &lpc_mdio_read;
819 	pldat->mii_bus->write = &lpc_mdio_write;
820 	pldat->mii_bus->reset = &lpc_mdio_reset;
821 	snprintf(pldat->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
822 		 pldat->pdev->name, pldat->pdev->id);
823 	pldat->mii_bus->priv = pldat;
824 	pldat->mii_bus->parent = &pldat->pdev->dev;
825 
826 	platform_set_drvdata(pldat->pdev, pldat->mii_bus);
827 
828 	if (mdiobus_register(pldat->mii_bus))
829 		goto err_out_unregister_bus;
830 
831 	if (lpc_mii_probe(pldat->ndev) != 0)
832 		goto err_out_unregister_bus;
833 
834 	return 0;
835 
836 err_out_unregister_bus:
837 	mdiobus_unregister(pldat->mii_bus);
838 	mdiobus_free(pldat->mii_bus);
839 err_out:
840 	return err;
841 }
842 
843 static void __lpc_handle_xmit(struct net_device *ndev)
844 {
845 	struct netdata_local *pldat = netdev_priv(ndev);
846 	u32 txcidx, *ptxstat, txstat;
847 
848 	txcidx = readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base));
849 	while (pldat->last_tx_idx != txcidx) {
850 		unsigned int skblen = pldat->skblen[pldat->last_tx_idx];
851 
852 		/* A buffer is available, get buffer status */
853 		ptxstat = &pldat->tx_stat_v[pldat->last_tx_idx];
854 		txstat = *ptxstat;
855 
856 		/* Next buffer and decrement used buffer counter */
857 		pldat->num_used_tx_buffs--;
858 		pldat->last_tx_idx++;
859 		if (pldat->last_tx_idx >= ENET_TX_DESC)
860 			pldat->last_tx_idx = 0;
861 
862 		/* Update collision counter */
863 		ndev->stats.collisions += TXSTATUS_COLLISIONS_GET(txstat);
864 
865 		/* Any errors occurred? */
866 		if (txstat & TXSTATUS_ERROR) {
867 			if (txstat & TXSTATUS_UNDERRUN) {
868 				/* FIFO underrun */
869 				ndev->stats.tx_fifo_errors++;
870 			}
871 			if (txstat & TXSTATUS_LATECOLL) {
872 				/* Late collision */
873 				ndev->stats.tx_aborted_errors++;
874 			}
875 			if (txstat & TXSTATUS_EXCESSCOLL) {
876 				/* Excessive collision */
877 				ndev->stats.tx_aborted_errors++;
878 			}
879 			if (txstat & TXSTATUS_EXCESSDEFER) {
880 				/* Defer limit */
881 				ndev->stats.tx_aborted_errors++;
882 			}
883 			ndev->stats.tx_errors++;
884 		} else {
885 			/* Update stats */
886 			ndev->stats.tx_packets++;
887 			ndev->stats.tx_bytes += skblen;
888 		}
889 
890 		txcidx = readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base));
891 	}
892 
893 	if (pldat->num_used_tx_buffs <= ENET_TX_DESC/2) {
894 		if (netif_queue_stopped(ndev))
895 			netif_wake_queue(ndev);
896 	}
897 }
898 
899 static int __lpc_handle_recv(struct net_device *ndev, int budget)
900 {
901 	struct netdata_local *pldat = netdev_priv(ndev);
902 	struct sk_buff *skb;
903 	u32 rxconsidx, len, ethst;
904 	struct rx_status_t *prxstat;
905 	int rx_done = 0;
906 
907 	/* Get the current RX buffer indexes */
908 	rxconsidx = readl(LPC_ENET_RXCONSUMEINDEX(pldat->net_base));
909 	while (rx_done < budget && rxconsidx !=
910 			readl(LPC_ENET_RXPRODUCEINDEX(pldat->net_base))) {
911 		/* Get pointer to receive status */
912 		prxstat = &pldat->rx_stat_v[rxconsidx];
913 		len = (prxstat->statusinfo & RXSTATUS_SIZE) + 1;
914 
915 		/* Status error? */
916 		ethst = prxstat->statusinfo;
917 		if ((ethst & (RXSTATUS_ERROR | RXSTATUS_STATUS_ERROR)) ==
918 		    (RXSTATUS_ERROR | RXSTATUS_RANGE))
919 			ethst &= ~RXSTATUS_ERROR;
920 
921 		if (ethst & RXSTATUS_ERROR) {
922 			int si = prxstat->statusinfo;
923 			/* Check statuses */
924 			if (si & RXSTATUS_OVERRUN) {
925 				/* Overrun error */
926 				ndev->stats.rx_fifo_errors++;
927 			} else if (si & RXSTATUS_CRC) {
928 				/* CRC error */
929 				ndev->stats.rx_crc_errors++;
930 			} else if (si & RXSTATUS_LENGTH) {
931 				/* Length error */
932 				ndev->stats.rx_length_errors++;
933 			} else if (si & RXSTATUS_ERROR) {
934 				/* Other error */
935 				ndev->stats.rx_length_errors++;
936 			}
937 			ndev->stats.rx_errors++;
938 		} else {
939 			/* Packet is good */
940 			skb = dev_alloc_skb(len);
941 			if (!skb) {
942 				ndev->stats.rx_dropped++;
943 			} else {
944 				/* Copy packet from buffer */
945 				skb_put_data(skb,
946 					     pldat->rx_buff_v + rxconsidx * ENET_MAXF_SIZE,
947 					     len);
948 
949 				/* Pass to upper layer */
950 				skb->protocol = eth_type_trans(skb, ndev);
951 				netif_receive_skb(skb);
952 				ndev->stats.rx_packets++;
953 				ndev->stats.rx_bytes += len;
954 			}
955 		}
956 
957 		/* Increment consume index */
958 		rxconsidx = rxconsidx + 1;
959 		if (rxconsidx >= ENET_RX_DESC)
960 			rxconsidx = 0;
961 		writel(rxconsidx,
962 		       LPC_ENET_RXCONSUMEINDEX(pldat->net_base));
963 		rx_done++;
964 	}
965 
966 	return rx_done;
967 }
968 
969 static int lpc_eth_poll(struct napi_struct *napi, int budget)
970 {
971 	struct netdata_local *pldat = container_of(napi,
972 			struct netdata_local, napi);
973 	struct net_device *ndev = pldat->ndev;
974 	int rx_done = 0;
975 	struct netdev_queue *txq = netdev_get_tx_queue(ndev, 0);
976 
977 	__netif_tx_lock(txq, smp_processor_id());
978 	__lpc_handle_xmit(ndev);
979 	__netif_tx_unlock(txq);
980 	rx_done = __lpc_handle_recv(ndev, budget);
981 
982 	if (rx_done < budget) {
983 		napi_complete_done(napi, rx_done);
984 		lpc_eth_enable_int(pldat->net_base);
985 	}
986 
987 	return rx_done;
988 }
989 
990 static irqreturn_t __lpc_eth_interrupt(int irq, void *dev_id)
991 {
992 	struct net_device *ndev = dev_id;
993 	struct netdata_local *pldat = netdev_priv(ndev);
994 	u32 tmp;
995 
996 	spin_lock(&pldat->lock);
997 
998 	tmp = readl(LPC_ENET_INTSTATUS(pldat->net_base));
999 	/* Clear interrupts */
1000 	writel(tmp, LPC_ENET_INTCLEAR(pldat->net_base));
1001 
1002 	lpc_eth_disable_int(pldat->net_base);
1003 	if (likely(napi_schedule_prep(&pldat->napi)))
1004 		__napi_schedule(&pldat->napi);
1005 
1006 	spin_unlock(&pldat->lock);
1007 
1008 	return IRQ_HANDLED;
1009 }
1010 
1011 static int lpc_eth_close(struct net_device *ndev)
1012 {
1013 	unsigned long flags;
1014 	struct netdata_local *pldat = netdev_priv(ndev);
1015 
1016 	if (netif_msg_ifdown(pldat))
1017 		dev_dbg(&pldat->pdev->dev, "shutting down %s\n", ndev->name);
1018 
1019 	napi_disable(&pldat->napi);
1020 	netif_stop_queue(ndev);
1021 
1022 	if (ndev->phydev)
1023 		phy_stop(ndev->phydev);
1024 
1025 	spin_lock_irqsave(&pldat->lock, flags);
1026 	__lpc_eth_reset(pldat);
1027 	netif_carrier_off(ndev);
1028 	writel(0, LPC_ENET_MAC1(pldat->net_base));
1029 	writel(0, LPC_ENET_MAC2(pldat->net_base));
1030 	spin_unlock_irqrestore(&pldat->lock, flags);
1031 
1032 	clk_disable_unprepare(pldat->clk);
1033 
1034 	return 0;
1035 }
1036 
1037 static int lpc_eth_hard_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1038 {
1039 	struct netdata_local *pldat = netdev_priv(ndev);
1040 	u32 len, txidx;
1041 	u32 *ptxstat;
1042 	struct txrx_desc_t *ptxrxdesc;
1043 
1044 	len = skb->len;
1045 
1046 	spin_lock_irq(&pldat->lock);
1047 
1048 	if (pldat->num_used_tx_buffs >= (ENET_TX_DESC - 1)) {
1049 		/* This function should never be called when there are no
1050 		   buffers */
1051 		netif_stop_queue(ndev);
1052 		spin_unlock_irq(&pldat->lock);
1053 		WARN(1, "BUG! TX request when no free TX buffers!\n");
1054 		return NETDEV_TX_BUSY;
1055 	}
1056 
1057 	/* Get the next TX descriptor index */
1058 	txidx = readl(LPC_ENET_TXPRODUCEINDEX(pldat->net_base));
1059 
1060 	/* Setup control for the transfer */
1061 	ptxstat = &pldat->tx_stat_v[txidx];
1062 	*ptxstat = 0;
1063 	ptxrxdesc = &pldat->tx_desc_v[txidx];
1064 	ptxrxdesc->control =
1065 		(len - 1) | TXDESC_CONTROL_LAST | TXDESC_CONTROL_INT;
1066 
1067 	/* Copy data to the DMA buffer */
1068 	memcpy(pldat->tx_buff_v + txidx * ENET_MAXF_SIZE, skb->data, len);
1069 
1070 	/* Save the buffer and increment the buffer counter */
1071 	pldat->skblen[txidx] = len;
1072 	pldat->num_used_tx_buffs++;
1073 
1074 	/* Start transmit */
1075 	txidx++;
1076 	if (txidx >= ENET_TX_DESC)
1077 		txidx = 0;
1078 	writel(txidx, LPC_ENET_TXPRODUCEINDEX(pldat->net_base));
1079 
1080 	/* Stop queue if no more TX buffers */
1081 	if (pldat->num_used_tx_buffs >= (ENET_TX_DESC - 1))
1082 		netif_stop_queue(ndev);
1083 
1084 	spin_unlock_irq(&pldat->lock);
1085 
1086 	dev_kfree_skb(skb);
1087 	return NETDEV_TX_OK;
1088 }
1089 
1090 static int lpc_set_mac_address(struct net_device *ndev, void *p)
1091 {
1092 	struct sockaddr *addr = p;
1093 	struct netdata_local *pldat = netdev_priv(ndev);
1094 	unsigned long flags;
1095 
1096 	if (!is_valid_ether_addr(addr->sa_data))
1097 		return -EADDRNOTAVAIL;
1098 	memcpy(ndev->dev_addr, addr->sa_data, ETH_ALEN);
1099 
1100 	spin_lock_irqsave(&pldat->lock, flags);
1101 
1102 	/* Set station address */
1103 	__lpc_set_mac(pldat, ndev->dev_addr);
1104 
1105 	spin_unlock_irqrestore(&pldat->lock, flags);
1106 
1107 	return 0;
1108 }
1109 
1110 static void lpc_eth_set_multicast_list(struct net_device *ndev)
1111 {
1112 	struct netdata_local *pldat = netdev_priv(ndev);
1113 	struct netdev_hw_addr_list *mcptr = &ndev->mc;
1114 	struct netdev_hw_addr *ha;
1115 	u32 tmp32, hash_val, hashlo, hashhi;
1116 	unsigned long flags;
1117 
1118 	spin_lock_irqsave(&pldat->lock, flags);
1119 
1120 	/* Set station address */
1121 	__lpc_set_mac(pldat, ndev->dev_addr);
1122 
1123 	tmp32 =  LPC_RXFLTRW_ACCEPTUBROADCAST | LPC_RXFLTRW_ACCEPTPERFECT;
1124 
1125 	if (ndev->flags & IFF_PROMISC)
1126 		tmp32 |= LPC_RXFLTRW_ACCEPTUNICAST |
1127 			LPC_RXFLTRW_ACCEPTUMULTICAST;
1128 	if (ndev->flags & IFF_ALLMULTI)
1129 		tmp32 |= LPC_RXFLTRW_ACCEPTUMULTICAST;
1130 
1131 	if (netdev_hw_addr_list_count(mcptr))
1132 		tmp32 |= LPC_RXFLTRW_ACCEPTUMULTICASTHASH;
1133 
1134 	writel(tmp32, LPC_ENET_RXFILTER_CTRL(pldat->net_base));
1135 
1136 
1137 	/* Set initial hash table */
1138 	hashlo = 0x0;
1139 	hashhi = 0x0;
1140 
1141 	/* 64 bits : multicast address in hash table */
1142 	netdev_hw_addr_list_for_each(ha, mcptr) {
1143 		hash_val = (ether_crc(6, ha->addr) >> 23) & 0x3F;
1144 
1145 		if (hash_val >= 32)
1146 			hashhi |= 1 << (hash_val - 32);
1147 		else
1148 			hashlo |= 1 << hash_val;
1149 	}
1150 
1151 	writel(hashlo, LPC_ENET_HASHFILTERL(pldat->net_base));
1152 	writel(hashhi, LPC_ENET_HASHFILTERH(pldat->net_base));
1153 
1154 	spin_unlock_irqrestore(&pldat->lock, flags);
1155 }
1156 
1157 static int lpc_eth_ioctl(struct net_device *ndev, struct ifreq *req, int cmd)
1158 {
1159 	struct phy_device *phydev = ndev->phydev;
1160 
1161 	if (!netif_running(ndev))
1162 		return -EINVAL;
1163 
1164 	if (!phydev)
1165 		return -ENODEV;
1166 
1167 	return phy_mii_ioctl(phydev, req, cmd);
1168 }
1169 
1170 static int lpc_eth_open(struct net_device *ndev)
1171 {
1172 	struct netdata_local *pldat = netdev_priv(ndev);
1173 	int ret;
1174 
1175 	if (netif_msg_ifup(pldat))
1176 		dev_dbg(&pldat->pdev->dev, "enabling %s\n", ndev->name);
1177 
1178 	ret = clk_prepare_enable(pldat->clk);
1179 	if (ret)
1180 		return ret;
1181 
1182 	/* Suspended PHY makes LPC ethernet core block, so resume now */
1183 	phy_resume(ndev->phydev);
1184 
1185 	/* Reset and initialize */
1186 	__lpc_eth_reset(pldat);
1187 	__lpc_eth_init(pldat);
1188 
1189 	/* schedule a link state check */
1190 	phy_start(ndev->phydev);
1191 	netif_start_queue(ndev);
1192 	napi_enable(&pldat->napi);
1193 
1194 	return 0;
1195 }
1196 
1197 /*
1198  * Ethtool ops
1199  */
1200 static void lpc_eth_ethtool_getdrvinfo(struct net_device *ndev,
1201 	struct ethtool_drvinfo *info)
1202 {
1203 	strlcpy(info->driver, MODNAME, sizeof(info->driver));
1204 	strlcpy(info->version, DRV_VERSION, sizeof(info->version));
1205 	strlcpy(info->bus_info, dev_name(ndev->dev.parent),
1206 		sizeof(info->bus_info));
1207 }
1208 
1209 static u32 lpc_eth_ethtool_getmsglevel(struct net_device *ndev)
1210 {
1211 	struct netdata_local *pldat = netdev_priv(ndev);
1212 
1213 	return pldat->msg_enable;
1214 }
1215 
1216 static void lpc_eth_ethtool_setmsglevel(struct net_device *ndev, u32 level)
1217 {
1218 	struct netdata_local *pldat = netdev_priv(ndev);
1219 
1220 	pldat->msg_enable = level;
1221 }
1222 
1223 static const struct ethtool_ops lpc_eth_ethtool_ops = {
1224 	.get_drvinfo	= lpc_eth_ethtool_getdrvinfo,
1225 	.get_msglevel	= lpc_eth_ethtool_getmsglevel,
1226 	.set_msglevel	= lpc_eth_ethtool_setmsglevel,
1227 	.get_link	= ethtool_op_get_link,
1228 	.get_link_ksettings = phy_ethtool_get_link_ksettings,
1229 	.set_link_ksettings = phy_ethtool_set_link_ksettings,
1230 };
1231 
1232 static const struct net_device_ops lpc_netdev_ops = {
1233 	.ndo_open		= lpc_eth_open,
1234 	.ndo_stop		= lpc_eth_close,
1235 	.ndo_start_xmit		= lpc_eth_hard_start_xmit,
1236 	.ndo_set_rx_mode	= lpc_eth_set_multicast_list,
1237 	.ndo_do_ioctl		= lpc_eth_ioctl,
1238 	.ndo_set_mac_address	= lpc_set_mac_address,
1239 	.ndo_validate_addr	= eth_validate_addr,
1240 };
1241 
1242 static int lpc_eth_drv_probe(struct platform_device *pdev)
1243 {
1244 	struct device *dev = &pdev->dev;
1245 	struct device_node *np = dev->of_node;
1246 	struct netdata_local *pldat;
1247 	struct net_device *ndev;
1248 	dma_addr_t dma_handle;
1249 	struct resource *res;
1250 	int irq, ret;
1251 	u32 tmp;
1252 
1253 	/* Setup network interface for RMII or MII mode */
1254 	tmp = __raw_readl(LPC32XX_CLKPWR_MACCLK_CTRL);
1255 	tmp &= ~LPC32XX_CLKPWR_MACCTRL_PINS_MSK;
1256 	if (lpc_phy_interface_mode(dev) == PHY_INTERFACE_MODE_MII)
1257 		tmp |= LPC32XX_CLKPWR_MACCTRL_USE_MII_PINS;
1258 	else
1259 		tmp |= LPC32XX_CLKPWR_MACCTRL_USE_RMII_PINS;
1260 	__raw_writel(tmp, LPC32XX_CLKPWR_MACCLK_CTRL);
1261 
1262 	/* Get platform resources */
1263 	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1264 	irq = platform_get_irq(pdev, 0);
1265 	if (!res || irq < 0) {
1266 		dev_err(dev, "error getting resources.\n");
1267 		ret = -ENXIO;
1268 		goto err_exit;
1269 	}
1270 
1271 	/* Allocate net driver data structure */
1272 	ndev = alloc_etherdev(sizeof(struct netdata_local));
1273 	if (!ndev) {
1274 		dev_err(dev, "could not allocate device.\n");
1275 		ret = -ENOMEM;
1276 		goto err_exit;
1277 	}
1278 
1279 	SET_NETDEV_DEV(ndev, dev);
1280 
1281 	pldat = netdev_priv(ndev);
1282 	pldat->pdev = pdev;
1283 	pldat->ndev = ndev;
1284 
1285 	spin_lock_init(&pldat->lock);
1286 
1287 	/* Save resources */
1288 	ndev->irq = irq;
1289 
1290 	/* Get clock for the device */
1291 	pldat->clk = clk_get(dev, NULL);
1292 	if (IS_ERR(pldat->clk)) {
1293 		dev_err(dev, "error getting clock.\n");
1294 		ret = PTR_ERR(pldat->clk);
1295 		goto err_out_free_dev;
1296 	}
1297 
1298 	/* Enable network clock */
1299 	ret = clk_prepare_enable(pldat->clk);
1300 	if (ret)
1301 		goto err_out_clk_put;
1302 
1303 	/* Map IO space */
1304 	pldat->net_base = ioremap(res->start, resource_size(res));
1305 	if (!pldat->net_base) {
1306 		dev_err(dev, "failed to map registers\n");
1307 		ret = -ENOMEM;
1308 		goto err_out_disable_clocks;
1309 	}
1310 	ret = request_irq(ndev->irq, __lpc_eth_interrupt, 0,
1311 			  ndev->name, ndev);
1312 	if (ret) {
1313 		dev_err(dev, "error requesting interrupt.\n");
1314 		goto err_out_iounmap;
1315 	}
1316 
1317 	/* Setup driver functions */
1318 	ndev->netdev_ops = &lpc_netdev_ops;
1319 	ndev->ethtool_ops = &lpc_eth_ethtool_ops;
1320 	ndev->watchdog_timeo = msecs_to_jiffies(2500);
1321 
1322 	/* Get size of DMA buffers/descriptors region */
1323 	pldat->dma_buff_size = (ENET_TX_DESC + ENET_RX_DESC) * (ENET_MAXF_SIZE +
1324 		sizeof(struct txrx_desc_t) + sizeof(struct rx_status_t));
1325 	pldat->dma_buff_base_v = 0;
1326 
1327 	if (use_iram_for_net(dev)) {
1328 		dma_handle = LPC32XX_IRAM_BASE;
1329 		if (pldat->dma_buff_size <= lpc32xx_return_iram_size())
1330 			pldat->dma_buff_base_v =
1331 				io_p2v(LPC32XX_IRAM_BASE);
1332 		else
1333 			netdev_err(ndev,
1334 				"IRAM not big enough for net buffers, using SDRAM instead.\n");
1335 	}
1336 
1337 	if (pldat->dma_buff_base_v == 0) {
1338 		ret = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(32));
1339 		if (ret)
1340 			goto err_out_free_irq;
1341 
1342 		pldat->dma_buff_size = PAGE_ALIGN(pldat->dma_buff_size);
1343 
1344 		/* Allocate a chunk of memory for the DMA ethernet buffers
1345 		   and descriptors */
1346 		pldat->dma_buff_base_v =
1347 			dma_alloc_coherent(dev,
1348 					   pldat->dma_buff_size, &dma_handle,
1349 					   GFP_KERNEL);
1350 		if (pldat->dma_buff_base_v == NULL) {
1351 			ret = -ENOMEM;
1352 			goto err_out_free_irq;
1353 		}
1354 	}
1355 	pldat->dma_buff_base_p = dma_handle;
1356 
1357 	netdev_dbg(ndev, "IO address space     :%pR\n", res);
1358 	netdev_dbg(ndev, "IO address size      :%d\n", resource_size(res));
1359 	netdev_dbg(ndev, "IO address (mapped)  :0x%p\n",
1360 			pldat->net_base);
1361 	netdev_dbg(ndev, "IRQ number           :%d\n", ndev->irq);
1362 	netdev_dbg(ndev, "DMA buffer size      :%d\n", pldat->dma_buff_size);
1363 	netdev_dbg(ndev, "DMA buffer P address :0x%08x\n",
1364 			pldat->dma_buff_base_p);
1365 	netdev_dbg(ndev, "DMA buffer V address :0x%p\n",
1366 			pldat->dma_buff_base_v);
1367 
1368 	/* Get MAC address from current HW setting (POR state is all zeros) */
1369 	__lpc_get_mac(pldat, ndev->dev_addr);
1370 
1371 	if (!is_valid_ether_addr(ndev->dev_addr)) {
1372 		const char *macaddr = of_get_mac_address(np);
1373 		if (macaddr)
1374 			memcpy(ndev->dev_addr, macaddr, ETH_ALEN);
1375 	}
1376 	if (!is_valid_ether_addr(ndev->dev_addr))
1377 		eth_hw_addr_random(ndev);
1378 
1379 	/* Reset the ethernet controller */
1380 	__lpc_eth_reset(pldat);
1381 
1382 	/* then shut everything down to save power */
1383 	__lpc_eth_shutdown(pldat);
1384 
1385 	/* Set default parameters */
1386 	pldat->msg_enable = NETIF_MSG_LINK;
1387 
1388 	/* Force an MII interface reset and clock setup */
1389 	__lpc_mii_mngt_reset(pldat);
1390 
1391 	/* Force default PHY interface setup in chip, this will probably be
1392 	   changed by the PHY driver */
1393 	pldat->link = 0;
1394 	pldat->speed = 100;
1395 	pldat->duplex = DUPLEX_FULL;
1396 	__lpc_params_setup(pldat);
1397 
1398 	netif_napi_add(ndev, &pldat->napi, lpc_eth_poll, NAPI_WEIGHT);
1399 
1400 	ret = register_netdev(ndev);
1401 	if (ret) {
1402 		dev_err(dev, "Cannot register net device, aborting.\n");
1403 		goto err_out_dma_unmap;
1404 	}
1405 	platform_set_drvdata(pdev, ndev);
1406 
1407 	ret = lpc_mii_init(pldat);
1408 	if (ret)
1409 		goto err_out_unregister_netdev;
1410 
1411 	netdev_info(ndev, "LPC mac at 0x%08x irq %d\n",
1412 	       res->start, ndev->irq);
1413 
1414 	device_init_wakeup(dev, 1);
1415 	device_set_wakeup_enable(dev, 0);
1416 
1417 	return 0;
1418 
1419 err_out_unregister_netdev:
1420 	unregister_netdev(ndev);
1421 err_out_dma_unmap:
1422 	if (!use_iram_for_net(dev) ||
1423 	    pldat->dma_buff_size > lpc32xx_return_iram_size())
1424 		dma_free_coherent(dev, pldat->dma_buff_size,
1425 				  pldat->dma_buff_base_v,
1426 				  pldat->dma_buff_base_p);
1427 err_out_free_irq:
1428 	free_irq(ndev->irq, ndev);
1429 err_out_iounmap:
1430 	iounmap(pldat->net_base);
1431 err_out_disable_clocks:
1432 	clk_disable_unprepare(pldat->clk);
1433 err_out_clk_put:
1434 	clk_put(pldat->clk);
1435 err_out_free_dev:
1436 	free_netdev(ndev);
1437 err_exit:
1438 	pr_err("%s: not found (%d).\n", MODNAME, ret);
1439 	return ret;
1440 }
1441 
1442 static int lpc_eth_drv_remove(struct platform_device *pdev)
1443 {
1444 	struct net_device *ndev = platform_get_drvdata(pdev);
1445 	struct netdata_local *pldat = netdev_priv(ndev);
1446 
1447 	unregister_netdev(ndev);
1448 
1449 	if (!use_iram_for_net(&pldat->pdev->dev) ||
1450 	    pldat->dma_buff_size > lpc32xx_return_iram_size())
1451 		dma_free_coherent(&pldat->pdev->dev, pldat->dma_buff_size,
1452 				  pldat->dma_buff_base_v,
1453 				  pldat->dma_buff_base_p);
1454 	free_irq(ndev->irq, ndev);
1455 	iounmap(pldat->net_base);
1456 	mdiobus_unregister(pldat->mii_bus);
1457 	mdiobus_free(pldat->mii_bus);
1458 	clk_disable_unprepare(pldat->clk);
1459 	clk_put(pldat->clk);
1460 	free_netdev(ndev);
1461 
1462 	return 0;
1463 }
1464 
1465 #ifdef CONFIG_PM
1466 static int lpc_eth_drv_suspend(struct platform_device *pdev,
1467 	pm_message_t state)
1468 {
1469 	struct net_device *ndev = platform_get_drvdata(pdev);
1470 	struct netdata_local *pldat = netdev_priv(ndev);
1471 
1472 	if (device_may_wakeup(&pdev->dev))
1473 		enable_irq_wake(ndev->irq);
1474 
1475 	if (ndev) {
1476 		if (netif_running(ndev)) {
1477 			netif_device_detach(ndev);
1478 			__lpc_eth_shutdown(pldat);
1479 			clk_disable_unprepare(pldat->clk);
1480 
1481 			/*
1482 			 * Reset again now clock is disable to be sure
1483 			 * EMC_MDC is down
1484 			 */
1485 			__lpc_eth_reset(pldat);
1486 		}
1487 	}
1488 
1489 	return 0;
1490 }
1491 
1492 static int lpc_eth_drv_resume(struct platform_device *pdev)
1493 {
1494 	struct net_device *ndev = platform_get_drvdata(pdev);
1495 	struct netdata_local *pldat;
1496 
1497 	if (device_may_wakeup(&pdev->dev))
1498 		disable_irq_wake(ndev->irq);
1499 
1500 	if (ndev) {
1501 		if (netif_running(ndev)) {
1502 			pldat = netdev_priv(ndev);
1503 
1504 			/* Enable interface clock */
1505 			clk_enable(pldat->clk);
1506 
1507 			/* Reset and initialize */
1508 			__lpc_eth_reset(pldat);
1509 			__lpc_eth_init(pldat);
1510 
1511 			netif_device_attach(ndev);
1512 		}
1513 	}
1514 
1515 	return 0;
1516 }
1517 #endif
1518 
1519 static const struct of_device_id lpc_eth_match[] = {
1520 	{ .compatible = "nxp,lpc-eth" },
1521 	{ }
1522 };
1523 MODULE_DEVICE_TABLE(of, lpc_eth_match);
1524 
1525 static struct platform_driver lpc_eth_driver = {
1526 	.probe		= lpc_eth_drv_probe,
1527 	.remove		= lpc_eth_drv_remove,
1528 #ifdef CONFIG_PM
1529 	.suspend	= lpc_eth_drv_suspend,
1530 	.resume		= lpc_eth_drv_resume,
1531 #endif
1532 	.driver		= {
1533 		.name	= MODNAME,
1534 		.of_match_table = lpc_eth_match,
1535 	},
1536 };
1537 
1538 module_platform_driver(lpc_eth_driver);
1539 
1540 MODULE_AUTHOR("Kevin Wells <kevin.wells@nxp.com>");
1541 MODULE_AUTHOR("Roland Stigge <stigge@antcom.de>");
1542 MODULE_DESCRIPTION("LPC Ethernet Driver");
1543 MODULE_LICENSE("GPL");
1544