xref: /openbmc/linux/drivers/net/fjes/fjes_main.c (revision 4ed91d48259d9ddd378424d008f2e6559f7e78f8)
1 /*
2  *  FUJITSU Extended Socket Network Device driver
3  *  Copyright (c) 2015 FUJITSU LIMITED
4  *
5  * This program is free software; you can redistribute it and/or modify it
6  * under the terms and conditions of the GNU General Public License,
7  * version 2, as published by the Free Software Foundation.
8  *
9  * This program is distributed in the hope it will be useful, but WITHOUT
10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
12  * more details.
13  *
14  * You should have received a copy of the GNU General Public License along with
15  * this program; if not, see <http://www.gnu.org/licenses/>.
16  *
17  * The full GNU General Public License is included in this distribution in
18  * the file called "COPYING".
19  *
20  */
21 
22 #include <linux/module.h>
23 #include <linux/types.h>
24 #include <linux/nls.h>
25 #include <linux/platform_device.h>
26 #include <linux/netdevice.h>
27 #include <linux/interrupt.h>
28 
29 #include "fjes.h"
30 #include "fjes_trace.h"
31 
32 #define MAJ 1
33 #define MIN 2
34 #define DRV_VERSION __stringify(MAJ) "." __stringify(MIN)
35 #define DRV_NAME	"fjes"
36 char fjes_driver_name[] = DRV_NAME;
37 char fjes_driver_version[] = DRV_VERSION;
38 static const char fjes_driver_string[] =
39 		"FUJITSU Extended Socket Network Device Driver";
40 static const char fjes_copyright[] =
41 		"Copyright (c) 2015 FUJITSU LIMITED";
42 
43 MODULE_AUTHOR("Taku Izumi <izumi.taku@jp.fujitsu.com>");
44 MODULE_DESCRIPTION("FUJITSU Extended Socket Network Device Driver");
45 MODULE_LICENSE("GPL");
46 MODULE_VERSION(DRV_VERSION);
47 
48 static int fjes_request_irq(struct fjes_adapter *);
49 static void fjes_free_irq(struct fjes_adapter *);
50 
51 static int fjes_open(struct net_device *);
52 static int fjes_close(struct net_device *);
53 static int fjes_setup_resources(struct fjes_adapter *);
54 static void fjes_free_resources(struct fjes_adapter *);
55 static netdev_tx_t fjes_xmit_frame(struct sk_buff *, struct net_device *);
56 static void fjes_raise_intr_rxdata_task(struct work_struct *);
57 static void fjes_tx_stall_task(struct work_struct *);
58 static void fjes_force_close_task(struct work_struct *);
59 static irqreturn_t fjes_intr(int, void*);
60 static void fjes_get_stats64(struct net_device *, struct rtnl_link_stats64 *);
61 static int fjes_change_mtu(struct net_device *, int);
62 static int fjes_vlan_rx_add_vid(struct net_device *, __be16 proto, u16);
63 static int fjes_vlan_rx_kill_vid(struct net_device *, __be16 proto, u16);
64 static void fjes_tx_retry(struct net_device *);
65 
66 static int fjes_acpi_add(struct acpi_device *);
67 static int fjes_acpi_remove(struct acpi_device *);
68 static acpi_status fjes_get_acpi_resource(struct acpi_resource *, void*);
69 
70 static int fjes_probe(struct platform_device *);
71 static int fjes_remove(struct platform_device *);
72 
73 static int fjes_sw_init(struct fjes_adapter *);
74 static void fjes_netdev_setup(struct net_device *);
75 static void fjes_irq_watch_task(struct work_struct *);
76 static void fjes_watch_unshare_task(struct work_struct *);
77 static void fjes_rx_irq(struct fjes_adapter *, int);
78 static int fjes_poll(struct napi_struct *, int);
79 
80 static const struct acpi_device_id fjes_acpi_ids[] = {
81 	{"PNP0C02", 0},
82 	{"", 0},
83 };
84 MODULE_DEVICE_TABLE(acpi, fjes_acpi_ids);
85 
86 static struct acpi_driver fjes_acpi_driver = {
87 	.name = DRV_NAME,
88 	.class = DRV_NAME,
89 	.owner = THIS_MODULE,
90 	.ids = fjes_acpi_ids,
91 	.ops = {
92 		.add = fjes_acpi_add,
93 		.remove = fjes_acpi_remove,
94 	},
95 };
96 
97 static struct platform_driver fjes_driver = {
98 	.driver = {
99 		.name = DRV_NAME,
100 	},
101 	.probe = fjes_probe,
102 	.remove = fjes_remove,
103 };
104 
105 static struct resource fjes_resource[] = {
106 	{
107 		.flags = IORESOURCE_MEM,
108 		.start = 0,
109 		.end = 0,
110 	},
111 	{
112 		.flags = IORESOURCE_IRQ,
113 		.start = 0,
114 		.end = 0,
115 	},
116 };
117 
118 static int fjes_acpi_add(struct acpi_device *device)
119 {
120 	struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL};
121 	char str_buf[sizeof(FJES_ACPI_SYMBOL) + 1];
122 	struct platform_device *plat_dev;
123 	union acpi_object *str;
124 	acpi_status status;
125 	int result;
126 
127 	status = acpi_evaluate_object(device->handle, "_STR", NULL, &buffer);
128 	if (ACPI_FAILURE(status))
129 		return -ENODEV;
130 
131 	str = buffer.pointer;
132 	result = utf16s_to_utf8s((wchar_t *)str->string.pointer,
133 				 str->string.length, UTF16_LITTLE_ENDIAN,
134 				 str_buf, sizeof(str_buf) - 1);
135 	str_buf[result] = 0;
136 
137 	if (strncmp(FJES_ACPI_SYMBOL, str_buf, strlen(FJES_ACPI_SYMBOL)) != 0) {
138 		kfree(buffer.pointer);
139 		return -ENODEV;
140 	}
141 	kfree(buffer.pointer);
142 
143 	status = acpi_walk_resources(device->handle, METHOD_NAME__CRS,
144 				     fjes_get_acpi_resource, fjes_resource);
145 	if (ACPI_FAILURE(status))
146 		return -ENODEV;
147 
148 	/* create platform_device */
149 	plat_dev = platform_device_register_simple(DRV_NAME, 0, fjes_resource,
150 						   ARRAY_SIZE(fjes_resource));
151 	device->driver_data = plat_dev;
152 
153 	return 0;
154 }
155 
156 static int fjes_acpi_remove(struct acpi_device *device)
157 {
158 	struct platform_device *plat_dev;
159 
160 	plat_dev = (struct platform_device *)acpi_driver_data(device);
161 	platform_device_unregister(plat_dev);
162 
163 	return 0;
164 }
165 
166 static acpi_status
167 fjes_get_acpi_resource(struct acpi_resource *acpi_res, void *data)
168 {
169 	struct acpi_resource_address32 *addr;
170 	struct acpi_resource_irq *irq;
171 	struct resource *res = data;
172 
173 	switch (acpi_res->type) {
174 	case ACPI_RESOURCE_TYPE_ADDRESS32:
175 		addr = &acpi_res->data.address32;
176 		res[0].start = addr->address.minimum;
177 		res[0].end = addr->address.minimum +
178 			addr->address.address_length - 1;
179 		break;
180 
181 	case ACPI_RESOURCE_TYPE_IRQ:
182 		irq = &acpi_res->data.irq;
183 		if (irq->interrupt_count != 1)
184 			return AE_ERROR;
185 		res[1].start = irq->interrupts[0];
186 		res[1].end = irq->interrupts[0];
187 		break;
188 
189 	default:
190 		break;
191 	}
192 
193 	return AE_OK;
194 }
195 
196 static int fjes_request_irq(struct fjes_adapter *adapter)
197 {
198 	struct net_device *netdev = adapter->netdev;
199 	int result = -1;
200 
201 	adapter->interrupt_watch_enable = true;
202 	if (!delayed_work_pending(&adapter->interrupt_watch_task)) {
203 		queue_delayed_work(adapter->control_wq,
204 				   &adapter->interrupt_watch_task,
205 				   FJES_IRQ_WATCH_DELAY);
206 	}
207 
208 	if (!adapter->irq_registered) {
209 		result = request_irq(adapter->hw.hw_res.irq, fjes_intr,
210 				     IRQF_SHARED, netdev->name, adapter);
211 		if (result)
212 			adapter->irq_registered = false;
213 		else
214 			adapter->irq_registered = true;
215 	}
216 
217 	return result;
218 }
219 
220 static void fjes_free_irq(struct fjes_adapter *adapter)
221 {
222 	struct fjes_hw *hw = &adapter->hw;
223 
224 	adapter->interrupt_watch_enable = false;
225 	cancel_delayed_work_sync(&adapter->interrupt_watch_task);
226 
227 	fjes_hw_set_irqmask(hw, REG_ICTL_MASK_ALL, true);
228 
229 	if (adapter->irq_registered) {
230 		free_irq(adapter->hw.hw_res.irq, adapter);
231 		adapter->irq_registered = false;
232 	}
233 }
234 
235 static const struct net_device_ops fjes_netdev_ops = {
236 	.ndo_open		= fjes_open,
237 	.ndo_stop		= fjes_close,
238 	.ndo_start_xmit		= fjes_xmit_frame,
239 	.ndo_get_stats64	= fjes_get_stats64,
240 	.ndo_change_mtu		= fjes_change_mtu,
241 	.ndo_tx_timeout		= fjes_tx_retry,
242 	.ndo_vlan_rx_add_vid	= fjes_vlan_rx_add_vid,
243 	.ndo_vlan_rx_kill_vid = fjes_vlan_rx_kill_vid,
244 };
245 
246 /* fjes_open - Called when a network interface is made active */
247 static int fjes_open(struct net_device *netdev)
248 {
249 	struct fjes_adapter *adapter = netdev_priv(netdev);
250 	struct fjes_hw *hw = &adapter->hw;
251 	int result;
252 
253 	if (adapter->open_guard)
254 		return -ENXIO;
255 
256 	result = fjes_setup_resources(adapter);
257 	if (result)
258 		goto err_setup_res;
259 
260 	hw->txrx_stop_req_bit = 0;
261 	hw->epstop_req_bit = 0;
262 
263 	napi_enable(&adapter->napi);
264 
265 	fjes_hw_capture_interrupt_status(hw);
266 
267 	result = fjes_request_irq(adapter);
268 	if (result)
269 		goto err_req_irq;
270 
271 	fjes_hw_set_irqmask(hw, REG_ICTL_MASK_ALL, false);
272 
273 	netif_tx_start_all_queues(netdev);
274 	netif_carrier_on(netdev);
275 
276 	return 0;
277 
278 err_req_irq:
279 	fjes_free_irq(adapter);
280 	napi_disable(&adapter->napi);
281 
282 err_setup_res:
283 	fjes_free_resources(adapter);
284 	return result;
285 }
286 
287 /* fjes_close - Disables a network interface */
288 static int fjes_close(struct net_device *netdev)
289 {
290 	struct fjes_adapter *adapter = netdev_priv(netdev);
291 	struct fjes_hw *hw = &adapter->hw;
292 	unsigned long flags;
293 	int epidx;
294 
295 	netif_tx_stop_all_queues(netdev);
296 	netif_carrier_off(netdev);
297 
298 	fjes_hw_raise_epstop(hw);
299 
300 	napi_disable(&adapter->napi);
301 
302 	spin_lock_irqsave(&hw->rx_status_lock, flags);
303 	for (epidx = 0; epidx < hw->max_epid; epidx++) {
304 		if (epidx == hw->my_epid)
305 			continue;
306 
307 		if (fjes_hw_get_partner_ep_status(hw, epidx) ==
308 		    EP_PARTNER_SHARED)
309 			adapter->hw.ep_shm_info[epidx]
310 				   .tx.info->v1i.rx_status &=
311 				~FJES_RX_POLL_WORK;
312 	}
313 	spin_unlock_irqrestore(&hw->rx_status_lock, flags);
314 
315 	fjes_free_irq(adapter);
316 
317 	cancel_delayed_work_sync(&adapter->interrupt_watch_task);
318 	cancel_work_sync(&adapter->unshare_watch_task);
319 	adapter->unshare_watch_bitmask = 0;
320 	cancel_work_sync(&adapter->raise_intr_rxdata_task);
321 	cancel_work_sync(&adapter->tx_stall_task);
322 
323 	cancel_work_sync(&hw->update_zone_task);
324 	cancel_work_sync(&hw->epstop_task);
325 
326 	fjes_hw_wait_epstop(hw);
327 
328 	fjes_free_resources(adapter);
329 
330 	return 0;
331 }
332 
333 static int fjes_setup_resources(struct fjes_adapter *adapter)
334 {
335 	struct net_device *netdev = adapter->netdev;
336 	struct ep_share_mem_info *buf_pair;
337 	struct fjes_hw *hw = &adapter->hw;
338 	unsigned long flags;
339 	int result;
340 	int epidx;
341 
342 	mutex_lock(&hw->hw_info.lock);
343 	result = fjes_hw_request_info(hw);
344 	switch (result) {
345 	case 0:
346 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
347 			hw->ep_shm_info[epidx].es_status =
348 			    hw->hw_info.res_buf->info.info[epidx].es_status;
349 			hw->ep_shm_info[epidx].zone =
350 			    hw->hw_info.res_buf->info.info[epidx].zone;
351 		}
352 		break;
353 	default:
354 	case -ENOMSG:
355 	case -EBUSY:
356 		adapter->force_reset = true;
357 
358 		mutex_unlock(&hw->hw_info.lock);
359 		return result;
360 	}
361 	mutex_unlock(&hw->hw_info.lock);
362 
363 	for (epidx = 0; epidx < (hw->max_epid); epidx++) {
364 		if ((epidx != hw->my_epid) &&
365 		    (hw->ep_shm_info[epidx].es_status ==
366 		     FJES_ZONING_STATUS_ENABLE)) {
367 			fjes_hw_raise_interrupt(hw, epidx,
368 						REG_ICTL_MASK_INFO_UPDATE);
369 			hw->ep_shm_info[epidx].ep_stats
370 				.send_intr_zoneupdate += 1;
371 		}
372 	}
373 
374 	msleep(FJES_OPEN_ZONE_UPDATE_WAIT * hw->max_epid);
375 
376 	for (epidx = 0; epidx < (hw->max_epid); epidx++) {
377 		if (epidx == hw->my_epid)
378 			continue;
379 
380 		buf_pair = &hw->ep_shm_info[epidx];
381 
382 		spin_lock_irqsave(&hw->rx_status_lock, flags);
383 		fjes_hw_setup_epbuf(&buf_pair->tx, netdev->dev_addr,
384 				    netdev->mtu);
385 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
386 
387 		if (fjes_hw_epid_is_same_zone(hw, epidx)) {
388 			mutex_lock(&hw->hw_info.lock);
389 			result =
390 			fjes_hw_register_buff_addr(hw, epidx, buf_pair);
391 			mutex_unlock(&hw->hw_info.lock);
392 
393 			switch (result) {
394 			case 0:
395 				break;
396 			case -ENOMSG:
397 			case -EBUSY:
398 			default:
399 				adapter->force_reset = true;
400 				return result;
401 			}
402 
403 			hw->ep_shm_info[epidx].ep_stats
404 				.com_regist_buf_exec += 1;
405 		}
406 	}
407 
408 	return 0;
409 }
410 
411 static void fjes_free_resources(struct fjes_adapter *adapter)
412 {
413 	struct net_device *netdev = adapter->netdev;
414 	struct fjes_device_command_param param;
415 	struct ep_share_mem_info *buf_pair;
416 	struct fjes_hw *hw = &adapter->hw;
417 	bool reset_flag = false;
418 	unsigned long flags;
419 	int result;
420 	int epidx;
421 
422 	for (epidx = 0; epidx < hw->max_epid; epidx++) {
423 		if (epidx == hw->my_epid)
424 			continue;
425 
426 		mutex_lock(&hw->hw_info.lock);
427 		result = fjes_hw_unregister_buff_addr(hw, epidx);
428 		mutex_unlock(&hw->hw_info.lock);
429 
430 		hw->ep_shm_info[epidx].ep_stats.com_unregist_buf_exec += 1;
431 
432 		if (result)
433 			reset_flag = true;
434 
435 		buf_pair = &hw->ep_shm_info[epidx];
436 
437 		spin_lock_irqsave(&hw->rx_status_lock, flags);
438 		fjes_hw_setup_epbuf(&buf_pair->tx,
439 				    netdev->dev_addr, netdev->mtu);
440 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
441 
442 		clear_bit(epidx, &hw->txrx_stop_req_bit);
443 	}
444 
445 	if (reset_flag || adapter->force_reset) {
446 		result = fjes_hw_reset(hw);
447 
448 		adapter->force_reset = false;
449 
450 		if (result)
451 			adapter->open_guard = true;
452 
453 		hw->hw_info.buffer_share_bit = 0;
454 
455 		memset((void *)&param, 0, sizeof(param));
456 
457 		param.req_len = hw->hw_info.req_buf_size;
458 		param.req_start = __pa(hw->hw_info.req_buf);
459 		param.res_len = hw->hw_info.res_buf_size;
460 		param.res_start = __pa(hw->hw_info.res_buf);
461 		param.share_start = __pa(hw->hw_info.share->ep_status);
462 
463 		fjes_hw_init_command_registers(hw, &param);
464 	}
465 }
466 
467 static void fjes_tx_stall_task(struct work_struct *work)
468 {
469 	struct fjes_adapter *adapter = container_of(work,
470 			struct fjes_adapter, tx_stall_task);
471 	struct net_device *netdev = adapter->netdev;
472 	struct fjes_hw *hw = &adapter->hw;
473 	int all_queue_available, sendable;
474 	enum ep_partner_status pstatus;
475 	int max_epid, my_epid, epid;
476 	union ep_buffer_info *info;
477 	int i;
478 
479 	if (((long)jiffies -
480 		dev_trans_start(netdev)) > FJES_TX_TX_STALL_TIMEOUT) {
481 		netif_wake_queue(netdev);
482 		return;
483 	}
484 
485 	my_epid = hw->my_epid;
486 	max_epid = hw->max_epid;
487 
488 	for (i = 0; i < 5; i++) {
489 		all_queue_available = 1;
490 
491 		for (epid = 0; epid < max_epid; epid++) {
492 			if (my_epid == epid)
493 				continue;
494 
495 			pstatus = fjes_hw_get_partner_ep_status(hw, epid);
496 			sendable = (pstatus == EP_PARTNER_SHARED);
497 			if (!sendable)
498 				continue;
499 
500 			info = adapter->hw.ep_shm_info[epid].tx.info;
501 
502 			if (!(info->v1i.rx_status & FJES_RX_MTU_CHANGING_DONE))
503 				return;
504 
505 			if (EP_RING_FULL(info->v1i.head, info->v1i.tail,
506 					 info->v1i.count_max)) {
507 				all_queue_available = 0;
508 				break;
509 			}
510 		}
511 
512 		if (all_queue_available) {
513 			netif_wake_queue(netdev);
514 			return;
515 		}
516 	}
517 
518 	usleep_range(50, 100);
519 
520 	queue_work(adapter->txrx_wq, &adapter->tx_stall_task);
521 }
522 
523 static void fjes_force_close_task(struct work_struct *work)
524 {
525 	struct fjes_adapter *adapter = container_of(work,
526 			struct fjes_adapter, force_close_task);
527 	struct net_device *netdev = adapter->netdev;
528 
529 	rtnl_lock();
530 	dev_close(netdev);
531 	rtnl_unlock();
532 }
533 
534 static void fjes_raise_intr_rxdata_task(struct work_struct *work)
535 {
536 	struct fjes_adapter *adapter = container_of(work,
537 			struct fjes_adapter, raise_intr_rxdata_task);
538 	struct fjes_hw *hw = &adapter->hw;
539 	enum ep_partner_status pstatus;
540 	int max_epid, my_epid, epid;
541 
542 	my_epid = hw->my_epid;
543 	max_epid = hw->max_epid;
544 
545 	for (epid = 0; epid < max_epid; epid++)
546 		hw->ep_shm_info[epid].tx_status_work = 0;
547 
548 	for (epid = 0; epid < max_epid; epid++) {
549 		if (epid == my_epid)
550 			continue;
551 
552 		pstatus = fjes_hw_get_partner_ep_status(hw, epid);
553 		if (pstatus == EP_PARTNER_SHARED) {
554 			hw->ep_shm_info[epid].tx_status_work =
555 				hw->ep_shm_info[epid].tx.info->v1i.tx_status;
556 
557 			if (hw->ep_shm_info[epid].tx_status_work ==
558 				FJES_TX_DELAY_SEND_PENDING) {
559 				hw->ep_shm_info[epid].tx.info->v1i.tx_status =
560 					FJES_TX_DELAY_SEND_NONE;
561 			}
562 		}
563 	}
564 
565 	for (epid = 0; epid < max_epid; epid++) {
566 		if (epid == my_epid)
567 			continue;
568 
569 		pstatus = fjes_hw_get_partner_ep_status(hw, epid);
570 		if ((hw->ep_shm_info[epid].tx_status_work ==
571 		     FJES_TX_DELAY_SEND_PENDING) &&
572 		    (pstatus == EP_PARTNER_SHARED) &&
573 		    !(hw->ep_shm_info[epid].rx.info->v1i.rx_status &
574 		      FJES_RX_POLL_WORK)) {
575 			fjes_hw_raise_interrupt(hw, epid,
576 						REG_ICTL_MASK_RX_DATA);
577 			hw->ep_shm_info[epid].ep_stats.send_intr_rx += 1;
578 		}
579 	}
580 
581 	usleep_range(500, 1000);
582 }
583 
584 static int fjes_tx_send(struct fjes_adapter *adapter, int dest,
585 			void *data, size_t len)
586 {
587 	int retval;
588 
589 	retval = fjes_hw_epbuf_tx_pkt_send(&adapter->hw.ep_shm_info[dest].tx,
590 					   data, len);
591 	if (retval)
592 		return retval;
593 
594 	adapter->hw.ep_shm_info[dest].tx.info->v1i.tx_status =
595 		FJES_TX_DELAY_SEND_PENDING;
596 	if (!work_pending(&adapter->raise_intr_rxdata_task))
597 		queue_work(adapter->txrx_wq,
598 			   &adapter->raise_intr_rxdata_task);
599 
600 	retval = 0;
601 	return retval;
602 }
603 
604 static netdev_tx_t
605 fjes_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
606 {
607 	struct fjes_adapter *adapter = netdev_priv(netdev);
608 	struct fjes_hw *hw = &adapter->hw;
609 
610 	int max_epid, my_epid, dest_epid;
611 	enum ep_partner_status pstatus;
612 	struct netdev_queue *cur_queue;
613 	char shortpkt[VLAN_ETH_HLEN];
614 	bool is_multi, vlan;
615 	struct ethhdr *eth;
616 	u16 queue_no = 0;
617 	u16 vlan_id = 0;
618 	netdev_tx_t ret;
619 	char *data;
620 	int len;
621 
622 	ret = NETDEV_TX_OK;
623 	is_multi = false;
624 	cur_queue = netdev_get_tx_queue(netdev, queue_no);
625 
626 	eth = (struct ethhdr *)skb->data;
627 	my_epid = hw->my_epid;
628 
629 	vlan = (vlan_get_tag(skb, &vlan_id) == 0) ? true : false;
630 
631 	data = skb->data;
632 	len = skb->len;
633 
634 	if (is_multicast_ether_addr(eth->h_dest)) {
635 		dest_epid = 0;
636 		max_epid = hw->max_epid;
637 		is_multi = true;
638 	} else if (is_local_ether_addr(eth->h_dest)) {
639 		dest_epid = eth->h_dest[ETH_ALEN - 1];
640 		max_epid = dest_epid + 1;
641 
642 		if ((eth->h_dest[0] == 0x02) &&
643 		    (0x00 == (eth->h_dest[1] | eth->h_dest[2] |
644 			      eth->h_dest[3] | eth->h_dest[4])) &&
645 		    (dest_epid < hw->max_epid)) {
646 			;
647 		} else {
648 			dest_epid = 0;
649 			max_epid = 0;
650 			ret = NETDEV_TX_OK;
651 
652 			adapter->stats64.tx_packets += 1;
653 			hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
654 			adapter->stats64.tx_bytes += len;
655 			hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
656 		}
657 	} else {
658 		dest_epid = 0;
659 		max_epid = 0;
660 		ret = NETDEV_TX_OK;
661 
662 		adapter->stats64.tx_packets += 1;
663 		hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
664 		adapter->stats64.tx_bytes += len;
665 		hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
666 	}
667 
668 	for (; dest_epid < max_epid; dest_epid++) {
669 		if (my_epid == dest_epid)
670 			continue;
671 
672 		pstatus = fjes_hw_get_partner_ep_status(hw, dest_epid);
673 		if (pstatus != EP_PARTNER_SHARED) {
674 			if (!is_multi)
675 				hw->ep_shm_info[dest_epid].ep_stats
676 					.tx_dropped_not_shared += 1;
677 			ret = NETDEV_TX_OK;
678 		} else if (!fjes_hw_check_epbuf_version(
679 				&adapter->hw.ep_shm_info[dest_epid].rx, 0)) {
680 			/* version is NOT 0 */
681 			adapter->stats64.tx_carrier_errors += 1;
682 			hw->ep_shm_info[dest_epid].net_stats
683 						.tx_carrier_errors += 1;
684 			hw->ep_shm_info[dest_epid].ep_stats
685 					.tx_dropped_ver_mismatch += 1;
686 
687 			ret = NETDEV_TX_OK;
688 		} else if (!fjes_hw_check_mtu(
689 				&adapter->hw.ep_shm_info[dest_epid].rx,
690 				netdev->mtu)) {
691 			adapter->stats64.tx_dropped += 1;
692 			hw->ep_shm_info[dest_epid].net_stats.tx_dropped += 1;
693 			adapter->stats64.tx_errors += 1;
694 			hw->ep_shm_info[dest_epid].net_stats.tx_errors += 1;
695 			hw->ep_shm_info[dest_epid].ep_stats
696 					.tx_dropped_buf_size_mismatch += 1;
697 
698 			ret = NETDEV_TX_OK;
699 		} else if (vlan &&
700 			   !fjes_hw_check_vlan_id(
701 				&adapter->hw.ep_shm_info[dest_epid].rx,
702 				vlan_id)) {
703 			hw->ep_shm_info[dest_epid].ep_stats
704 				.tx_dropped_vlanid_mismatch += 1;
705 			ret = NETDEV_TX_OK;
706 		} else {
707 			if (len < VLAN_ETH_HLEN) {
708 				memset(shortpkt, 0, VLAN_ETH_HLEN);
709 				memcpy(shortpkt, skb->data, skb->len);
710 				len = VLAN_ETH_HLEN;
711 				data = shortpkt;
712 			}
713 
714 			if (adapter->tx_retry_count == 0) {
715 				adapter->tx_start_jiffies = jiffies;
716 				adapter->tx_retry_count = 1;
717 			} else {
718 				adapter->tx_retry_count++;
719 			}
720 
721 			if (fjes_tx_send(adapter, dest_epid, data, len)) {
722 				if (is_multi) {
723 					ret = NETDEV_TX_OK;
724 				} else if (
725 					   ((long)jiffies -
726 					    (long)adapter->tx_start_jiffies) >=
727 					    FJES_TX_RETRY_TIMEOUT) {
728 					adapter->stats64.tx_fifo_errors += 1;
729 					hw->ep_shm_info[dest_epid].net_stats
730 								.tx_fifo_errors += 1;
731 					adapter->stats64.tx_errors += 1;
732 					hw->ep_shm_info[dest_epid].net_stats
733 								.tx_errors += 1;
734 
735 					ret = NETDEV_TX_OK;
736 				} else {
737 					netif_trans_update(netdev);
738 					hw->ep_shm_info[dest_epid].ep_stats
739 						.tx_buffer_full += 1;
740 					netif_tx_stop_queue(cur_queue);
741 
742 					if (!work_pending(&adapter->tx_stall_task))
743 						queue_work(adapter->txrx_wq,
744 							   &adapter->tx_stall_task);
745 
746 					ret = NETDEV_TX_BUSY;
747 				}
748 			} else {
749 				if (!is_multi) {
750 					adapter->stats64.tx_packets += 1;
751 					hw->ep_shm_info[dest_epid].net_stats
752 								.tx_packets += 1;
753 					adapter->stats64.tx_bytes += len;
754 					hw->ep_shm_info[dest_epid].net_stats
755 								.tx_bytes += len;
756 				}
757 
758 				adapter->tx_retry_count = 0;
759 				ret = NETDEV_TX_OK;
760 			}
761 		}
762 	}
763 
764 	if (ret == NETDEV_TX_OK) {
765 		dev_kfree_skb(skb);
766 		if (is_multi) {
767 			adapter->stats64.tx_packets += 1;
768 			hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
769 			adapter->stats64.tx_bytes += 1;
770 			hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
771 		}
772 	}
773 
774 	return ret;
775 }
776 
777 static void fjes_tx_retry(struct net_device *netdev)
778 {
779 	struct netdev_queue *queue = netdev_get_tx_queue(netdev, 0);
780 
781 	netif_tx_wake_queue(queue);
782 }
783 
784 static void
785 fjes_get_stats64(struct net_device *netdev, struct rtnl_link_stats64 *stats)
786 {
787 	struct fjes_adapter *adapter = netdev_priv(netdev);
788 
789 	memcpy(stats, &adapter->stats64, sizeof(struct rtnl_link_stats64));
790 }
791 
792 static int fjes_change_mtu(struct net_device *netdev, int new_mtu)
793 {
794 	struct fjes_adapter *adapter = netdev_priv(netdev);
795 	bool running = netif_running(netdev);
796 	struct fjes_hw *hw = &adapter->hw;
797 	unsigned long flags;
798 	int ret = -EINVAL;
799 	int idx, epidx;
800 
801 	for (idx = 0; fjes_support_mtu[idx] != 0; idx++) {
802 		if (new_mtu <= fjes_support_mtu[idx]) {
803 			new_mtu = fjes_support_mtu[idx];
804 			if (new_mtu == netdev->mtu)
805 				return 0;
806 
807 			ret = 0;
808 			break;
809 		}
810 	}
811 
812 	if (ret)
813 		return ret;
814 
815 	if (running) {
816 		spin_lock_irqsave(&hw->rx_status_lock, flags);
817 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
818 			if (epidx == hw->my_epid)
819 				continue;
820 			hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
821 				~FJES_RX_MTU_CHANGING_DONE;
822 		}
823 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
824 
825 		netif_tx_stop_all_queues(netdev);
826 		netif_carrier_off(netdev);
827 		cancel_work_sync(&adapter->tx_stall_task);
828 		napi_disable(&adapter->napi);
829 
830 		msleep(1000);
831 
832 		netif_tx_stop_all_queues(netdev);
833 	}
834 
835 	netdev->mtu = new_mtu;
836 
837 	if (running) {
838 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
839 			if (epidx == hw->my_epid)
840 				continue;
841 
842 			spin_lock_irqsave(&hw->rx_status_lock, flags);
843 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
844 					    netdev->dev_addr,
845 					    netdev->mtu);
846 
847 			hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
848 				FJES_RX_MTU_CHANGING_DONE;
849 			spin_unlock_irqrestore(&hw->rx_status_lock, flags);
850 		}
851 
852 		netif_tx_wake_all_queues(netdev);
853 		netif_carrier_on(netdev);
854 		napi_enable(&adapter->napi);
855 		napi_schedule(&adapter->napi);
856 	}
857 
858 	return ret;
859 }
860 
861 static int fjes_vlan_rx_add_vid(struct net_device *netdev,
862 				__be16 proto, u16 vid)
863 {
864 	struct fjes_adapter *adapter = netdev_priv(netdev);
865 	bool ret = true;
866 	int epid;
867 
868 	for (epid = 0; epid < adapter->hw.max_epid; epid++) {
869 		if (epid == adapter->hw.my_epid)
870 			continue;
871 
872 		if (!fjes_hw_check_vlan_id(
873 			&adapter->hw.ep_shm_info[epid].tx, vid))
874 			ret = fjes_hw_set_vlan_id(
875 				&adapter->hw.ep_shm_info[epid].tx, vid);
876 	}
877 
878 	return ret ? 0 : -ENOSPC;
879 }
880 
881 static int fjes_vlan_rx_kill_vid(struct net_device *netdev,
882 				 __be16 proto, u16 vid)
883 {
884 	struct fjes_adapter *adapter = netdev_priv(netdev);
885 	int epid;
886 
887 	for (epid = 0; epid < adapter->hw.max_epid; epid++) {
888 		if (epid == adapter->hw.my_epid)
889 			continue;
890 
891 		fjes_hw_del_vlan_id(&adapter->hw.ep_shm_info[epid].tx, vid);
892 	}
893 
894 	return 0;
895 }
896 
897 static void fjes_txrx_stop_req_irq(struct fjes_adapter *adapter,
898 				   int src_epid)
899 {
900 	struct fjes_hw *hw = &adapter->hw;
901 	enum ep_partner_status status;
902 	unsigned long flags;
903 
904 	status = fjes_hw_get_partner_ep_status(hw, src_epid);
905 	trace_fjes_txrx_stop_req_irq_pre(hw, src_epid, status);
906 	switch (status) {
907 	case EP_PARTNER_UNSHARE:
908 	case EP_PARTNER_COMPLETE:
909 	default:
910 		break;
911 	case EP_PARTNER_WAITING:
912 		if (src_epid < hw->my_epid) {
913 			spin_lock_irqsave(&hw->rx_status_lock, flags);
914 			hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
915 				FJES_RX_STOP_REQ_DONE;
916 			spin_unlock_irqrestore(&hw->rx_status_lock, flags);
917 
918 			clear_bit(src_epid, &hw->txrx_stop_req_bit);
919 			set_bit(src_epid, &adapter->unshare_watch_bitmask);
920 
921 			if (!work_pending(&adapter->unshare_watch_task))
922 				queue_work(adapter->control_wq,
923 					   &adapter->unshare_watch_task);
924 		}
925 		break;
926 	case EP_PARTNER_SHARED:
927 		if (hw->ep_shm_info[src_epid].rx.info->v1i.rx_status &
928 		    FJES_RX_STOP_REQ_REQUEST) {
929 			set_bit(src_epid, &hw->epstop_req_bit);
930 			if (!work_pending(&hw->epstop_task))
931 				queue_work(adapter->control_wq,
932 					   &hw->epstop_task);
933 		}
934 		break;
935 	}
936 	trace_fjes_txrx_stop_req_irq_post(hw, src_epid);
937 }
938 
939 static void fjes_stop_req_irq(struct fjes_adapter *adapter, int src_epid)
940 {
941 	struct fjes_hw *hw = &adapter->hw;
942 	enum ep_partner_status status;
943 	unsigned long flags;
944 
945 	set_bit(src_epid, &hw->hw_info.buffer_unshare_reserve_bit);
946 
947 	status = fjes_hw_get_partner_ep_status(hw, src_epid);
948 	trace_fjes_stop_req_irq_pre(hw, src_epid, status);
949 	switch (status) {
950 	case EP_PARTNER_WAITING:
951 		spin_lock_irqsave(&hw->rx_status_lock, flags);
952 		hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
953 				FJES_RX_STOP_REQ_DONE;
954 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
955 		clear_bit(src_epid, &hw->txrx_stop_req_bit);
956 		/* fall through */
957 	case EP_PARTNER_UNSHARE:
958 	case EP_PARTNER_COMPLETE:
959 	default:
960 		set_bit(src_epid, &adapter->unshare_watch_bitmask);
961 		if (!work_pending(&adapter->unshare_watch_task))
962 			queue_work(adapter->control_wq,
963 				   &adapter->unshare_watch_task);
964 		break;
965 	case EP_PARTNER_SHARED:
966 		set_bit(src_epid, &hw->epstop_req_bit);
967 
968 		if (!work_pending(&hw->epstop_task))
969 			queue_work(adapter->control_wq, &hw->epstop_task);
970 		break;
971 	}
972 	trace_fjes_stop_req_irq_post(hw, src_epid);
973 }
974 
975 static void fjes_update_zone_irq(struct fjes_adapter *adapter,
976 				 int src_epid)
977 {
978 	struct fjes_hw *hw = &adapter->hw;
979 
980 	if (!work_pending(&hw->update_zone_task))
981 		queue_work(adapter->control_wq, &hw->update_zone_task);
982 }
983 
984 static irqreturn_t fjes_intr(int irq, void *data)
985 {
986 	struct fjes_adapter *adapter = data;
987 	struct fjes_hw *hw = &adapter->hw;
988 	irqreturn_t ret;
989 	u32 icr;
990 
991 	icr = fjes_hw_capture_interrupt_status(hw);
992 
993 	if (icr & REG_IS_MASK_IS_ASSERT) {
994 		if (icr & REG_ICTL_MASK_RX_DATA) {
995 			fjes_rx_irq(adapter, icr & REG_IS_MASK_EPID);
996 			hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
997 				.recv_intr_rx += 1;
998 		}
999 
1000 		if (icr & REG_ICTL_MASK_DEV_STOP_REQ) {
1001 			fjes_stop_req_irq(adapter, icr & REG_IS_MASK_EPID);
1002 			hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
1003 				.recv_intr_stop += 1;
1004 		}
1005 
1006 		if (icr & REG_ICTL_MASK_TXRX_STOP_REQ) {
1007 			fjes_txrx_stop_req_irq(adapter, icr & REG_IS_MASK_EPID);
1008 			hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
1009 				.recv_intr_unshare += 1;
1010 		}
1011 
1012 		if (icr & REG_ICTL_MASK_TXRX_STOP_DONE)
1013 			fjes_hw_set_irqmask(hw,
1014 					    REG_ICTL_MASK_TXRX_STOP_DONE, true);
1015 
1016 		if (icr & REG_ICTL_MASK_INFO_UPDATE) {
1017 			fjes_update_zone_irq(adapter, icr & REG_IS_MASK_EPID);
1018 			hw->ep_shm_info[icr & REG_IS_MASK_EPID].ep_stats
1019 				.recv_intr_zoneupdate += 1;
1020 		}
1021 
1022 		ret = IRQ_HANDLED;
1023 	} else {
1024 		ret = IRQ_NONE;
1025 	}
1026 
1027 	return ret;
1028 }
1029 
1030 static int fjes_rxframe_search_exist(struct fjes_adapter *adapter,
1031 				     int start_epid)
1032 {
1033 	struct fjes_hw *hw = &adapter->hw;
1034 	enum ep_partner_status pstatus;
1035 	int max_epid, cur_epid;
1036 	int i;
1037 
1038 	max_epid = hw->max_epid;
1039 	start_epid = (start_epid + 1 + max_epid) % max_epid;
1040 
1041 	for (i = 0; i < max_epid; i++) {
1042 		cur_epid = (start_epid + i) % max_epid;
1043 		if (cur_epid == hw->my_epid)
1044 			continue;
1045 
1046 		pstatus = fjes_hw_get_partner_ep_status(hw, cur_epid);
1047 		if (pstatus == EP_PARTNER_SHARED) {
1048 			if (!fjes_hw_epbuf_rx_is_empty(
1049 				&hw->ep_shm_info[cur_epid].rx))
1050 				return cur_epid;
1051 		}
1052 	}
1053 	return -1;
1054 }
1055 
1056 static void *fjes_rxframe_get(struct fjes_adapter *adapter, size_t *psize,
1057 			      int *cur_epid)
1058 {
1059 	void *frame;
1060 
1061 	*cur_epid = fjes_rxframe_search_exist(adapter, *cur_epid);
1062 	if (*cur_epid < 0)
1063 		return NULL;
1064 
1065 	frame =
1066 	fjes_hw_epbuf_rx_curpkt_get_addr(
1067 		&adapter->hw.ep_shm_info[*cur_epid].rx, psize);
1068 
1069 	return frame;
1070 }
1071 
1072 static void fjes_rxframe_release(struct fjes_adapter *adapter, int cur_epid)
1073 {
1074 	fjes_hw_epbuf_rx_curpkt_drop(&adapter->hw.ep_shm_info[cur_epid].rx);
1075 }
1076 
1077 static void fjes_rx_irq(struct fjes_adapter *adapter, int src_epid)
1078 {
1079 	struct fjes_hw *hw = &adapter->hw;
1080 
1081 	fjes_hw_set_irqmask(hw, REG_ICTL_MASK_RX_DATA, true);
1082 
1083 	adapter->unset_rx_last = true;
1084 	napi_schedule(&adapter->napi);
1085 }
1086 
1087 static int fjes_poll(struct napi_struct *napi, int budget)
1088 {
1089 	struct fjes_adapter *adapter =
1090 			container_of(napi, struct fjes_adapter, napi);
1091 	struct net_device *netdev = napi->dev;
1092 	struct fjes_hw *hw = &adapter->hw;
1093 	struct sk_buff *skb;
1094 	int work_done = 0;
1095 	int cur_epid = 0;
1096 	int epidx;
1097 	size_t frame_len;
1098 	void *frame;
1099 
1100 	spin_lock(&hw->rx_status_lock);
1101 	for (epidx = 0; epidx < hw->max_epid; epidx++) {
1102 		if (epidx == hw->my_epid)
1103 			continue;
1104 
1105 		if (fjes_hw_get_partner_ep_status(hw, epidx) ==
1106 		    EP_PARTNER_SHARED)
1107 			adapter->hw.ep_shm_info[epidx]
1108 				   .tx.info->v1i.rx_status |= FJES_RX_POLL_WORK;
1109 	}
1110 	spin_unlock(&hw->rx_status_lock);
1111 
1112 	while (work_done < budget) {
1113 		prefetch(&adapter->hw);
1114 		frame = fjes_rxframe_get(adapter, &frame_len, &cur_epid);
1115 
1116 		if (frame) {
1117 			skb = napi_alloc_skb(napi, frame_len);
1118 			if (!skb) {
1119 				adapter->stats64.rx_dropped += 1;
1120 				hw->ep_shm_info[cur_epid].net_stats
1121 							 .rx_dropped += 1;
1122 				adapter->stats64.rx_errors += 1;
1123 				hw->ep_shm_info[cur_epid].net_stats
1124 							 .rx_errors += 1;
1125 			} else {
1126 				memcpy(skb_put(skb, frame_len),
1127 				       frame, frame_len);
1128 				skb->protocol = eth_type_trans(skb, netdev);
1129 				skb->ip_summed = CHECKSUM_UNNECESSARY;
1130 
1131 				netif_receive_skb(skb);
1132 
1133 				work_done++;
1134 
1135 				adapter->stats64.rx_packets += 1;
1136 				hw->ep_shm_info[cur_epid].net_stats
1137 							 .rx_packets += 1;
1138 				adapter->stats64.rx_bytes += frame_len;
1139 				hw->ep_shm_info[cur_epid].net_stats
1140 							 .rx_bytes += frame_len;
1141 
1142 				if (is_multicast_ether_addr(
1143 					((struct ethhdr *)frame)->h_dest)) {
1144 					adapter->stats64.multicast += 1;
1145 					hw->ep_shm_info[cur_epid].net_stats
1146 								 .multicast += 1;
1147 				}
1148 			}
1149 
1150 			fjes_rxframe_release(adapter, cur_epid);
1151 			adapter->unset_rx_last = true;
1152 		} else {
1153 			break;
1154 		}
1155 	}
1156 
1157 	if (work_done < budget) {
1158 		napi_complete_done(napi, work_done);
1159 
1160 		if (adapter->unset_rx_last) {
1161 			adapter->rx_last_jiffies = jiffies;
1162 			adapter->unset_rx_last = false;
1163 		}
1164 
1165 		if (((long)jiffies - (long)adapter->rx_last_jiffies) < 3) {
1166 			napi_reschedule(napi);
1167 		} else {
1168 			spin_lock(&hw->rx_status_lock);
1169 			for (epidx = 0; epidx < hw->max_epid; epidx++) {
1170 				if (epidx == hw->my_epid)
1171 					continue;
1172 				if (fjes_hw_get_partner_ep_status(hw, epidx) ==
1173 				    EP_PARTNER_SHARED)
1174 					adapter->hw.ep_shm_info[epidx].tx
1175 						   .info->v1i.rx_status &=
1176 						~FJES_RX_POLL_WORK;
1177 			}
1178 			spin_unlock(&hw->rx_status_lock);
1179 
1180 			fjes_hw_set_irqmask(hw, REG_ICTL_MASK_RX_DATA, false);
1181 		}
1182 	}
1183 
1184 	return work_done;
1185 }
1186 
1187 /* fjes_probe - Device Initialization Routine */
1188 static int fjes_probe(struct platform_device *plat_dev)
1189 {
1190 	struct fjes_adapter *adapter;
1191 	struct net_device *netdev;
1192 	struct resource *res;
1193 	struct fjes_hw *hw;
1194 	int err;
1195 
1196 	err = -ENOMEM;
1197 	netdev = alloc_netdev_mq(sizeof(struct fjes_adapter), "es%d",
1198 				 NET_NAME_UNKNOWN, fjes_netdev_setup,
1199 				 FJES_MAX_QUEUES);
1200 
1201 	if (!netdev)
1202 		goto err_out;
1203 
1204 	SET_NETDEV_DEV(netdev, &plat_dev->dev);
1205 
1206 	dev_set_drvdata(&plat_dev->dev, netdev);
1207 	adapter = netdev_priv(netdev);
1208 	adapter->netdev = netdev;
1209 	adapter->plat_dev = plat_dev;
1210 	hw = &adapter->hw;
1211 	hw->back = adapter;
1212 
1213 	/* setup the private structure */
1214 	err = fjes_sw_init(adapter);
1215 	if (err)
1216 		goto err_free_netdev;
1217 
1218 	INIT_WORK(&adapter->force_close_task, fjes_force_close_task);
1219 	adapter->force_reset = false;
1220 	adapter->open_guard = false;
1221 
1222 	adapter->txrx_wq = alloc_workqueue(DRV_NAME "/txrx", WQ_MEM_RECLAIM, 0);
1223 	adapter->control_wq = alloc_workqueue(DRV_NAME "/control",
1224 					      WQ_MEM_RECLAIM, 0);
1225 
1226 	INIT_WORK(&adapter->tx_stall_task, fjes_tx_stall_task);
1227 	INIT_WORK(&adapter->raise_intr_rxdata_task,
1228 		  fjes_raise_intr_rxdata_task);
1229 	INIT_WORK(&adapter->unshare_watch_task, fjes_watch_unshare_task);
1230 	adapter->unshare_watch_bitmask = 0;
1231 
1232 	INIT_DELAYED_WORK(&adapter->interrupt_watch_task, fjes_irq_watch_task);
1233 	adapter->interrupt_watch_enable = false;
1234 
1235 	res = platform_get_resource(plat_dev, IORESOURCE_MEM, 0);
1236 	hw->hw_res.start = res->start;
1237 	hw->hw_res.size = resource_size(res);
1238 	hw->hw_res.irq = platform_get_irq(plat_dev, 0);
1239 	err = fjes_hw_init(&adapter->hw);
1240 	if (err)
1241 		goto err_free_netdev;
1242 
1243 	/* setup MAC address (02:00:00:00:00:[epid])*/
1244 	netdev->dev_addr[0] = 2;
1245 	netdev->dev_addr[1] = 0;
1246 	netdev->dev_addr[2] = 0;
1247 	netdev->dev_addr[3] = 0;
1248 	netdev->dev_addr[4] = 0;
1249 	netdev->dev_addr[5] = hw->my_epid; /* EPID */
1250 
1251 	err = register_netdev(netdev);
1252 	if (err)
1253 		goto err_hw_exit;
1254 
1255 	netif_carrier_off(netdev);
1256 
1257 	fjes_dbg_adapter_init(adapter);
1258 
1259 	return 0;
1260 
1261 err_hw_exit:
1262 	fjes_hw_exit(&adapter->hw);
1263 err_free_netdev:
1264 	free_netdev(netdev);
1265 err_out:
1266 	return err;
1267 }
1268 
1269 /* fjes_remove - Device Removal Routine */
1270 static int fjes_remove(struct platform_device *plat_dev)
1271 {
1272 	struct net_device *netdev = dev_get_drvdata(&plat_dev->dev);
1273 	struct fjes_adapter *adapter = netdev_priv(netdev);
1274 	struct fjes_hw *hw = &adapter->hw;
1275 
1276 	fjes_dbg_adapter_exit(adapter);
1277 
1278 	cancel_delayed_work_sync(&adapter->interrupt_watch_task);
1279 	cancel_work_sync(&adapter->unshare_watch_task);
1280 	cancel_work_sync(&adapter->raise_intr_rxdata_task);
1281 	cancel_work_sync(&adapter->tx_stall_task);
1282 	if (adapter->control_wq)
1283 		destroy_workqueue(adapter->control_wq);
1284 	if (adapter->txrx_wq)
1285 		destroy_workqueue(adapter->txrx_wq);
1286 
1287 	unregister_netdev(netdev);
1288 
1289 	fjes_hw_exit(hw);
1290 
1291 	netif_napi_del(&adapter->napi);
1292 
1293 	free_netdev(netdev);
1294 
1295 	return 0;
1296 }
1297 
1298 static int fjes_sw_init(struct fjes_adapter *adapter)
1299 {
1300 	struct net_device *netdev = adapter->netdev;
1301 
1302 	netif_napi_add(netdev, &adapter->napi, fjes_poll, 64);
1303 
1304 	return 0;
1305 }
1306 
1307 /* fjes_netdev_setup - netdevice initialization routine */
1308 static void fjes_netdev_setup(struct net_device *netdev)
1309 {
1310 	ether_setup(netdev);
1311 
1312 	netdev->watchdog_timeo = FJES_TX_RETRY_INTERVAL;
1313 	netdev->netdev_ops = &fjes_netdev_ops;
1314 	fjes_set_ethtool_ops(netdev);
1315 	netdev->mtu = fjes_support_mtu[3];
1316 	netdev->min_mtu = fjes_support_mtu[0];
1317 	netdev->max_mtu = fjes_support_mtu[3];
1318 	netdev->flags |= IFF_BROADCAST;
1319 	netdev->features |= NETIF_F_HW_CSUM | NETIF_F_HW_VLAN_CTAG_FILTER;
1320 }
1321 
1322 static void fjes_irq_watch_task(struct work_struct *work)
1323 {
1324 	struct fjes_adapter *adapter = container_of(to_delayed_work(work),
1325 			struct fjes_adapter, interrupt_watch_task);
1326 
1327 	local_irq_disable();
1328 	fjes_intr(adapter->hw.hw_res.irq, adapter);
1329 	local_irq_enable();
1330 
1331 	if (fjes_rxframe_search_exist(adapter, 0) >= 0)
1332 		napi_schedule(&adapter->napi);
1333 
1334 	if (adapter->interrupt_watch_enable) {
1335 		if (!delayed_work_pending(&adapter->interrupt_watch_task))
1336 			queue_delayed_work(adapter->control_wq,
1337 					   &adapter->interrupt_watch_task,
1338 					   FJES_IRQ_WATCH_DELAY);
1339 	}
1340 }
1341 
1342 static void fjes_watch_unshare_task(struct work_struct *work)
1343 {
1344 	struct fjes_adapter *adapter =
1345 	container_of(work, struct fjes_adapter, unshare_watch_task);
1346 
1347 	struct net_device *netdev = adapter->netdev;
1348 	struct fjes_hw *hw = &adapter->hw;
1349 
1350 	int unshare_watch, unshare_reserve;
1351 	int max_epid, my_epid, epidx;
1352 	int stop_req, stop_req_done;
1353 	ulong unshare_watch_bitmask;
1354 	unsigned long flags;
1355 	int wait_time = 0;
1356 	int is_shared;
1357 	int ret;
1358 
1359 	my_epid = hw->my_epid;
1360 	max_epid = hw->max_epid;
1361 
1362 	unshare_watch_bitmask = adapter->unshare_watch_bitmask;
1363 	adapter->unshare_watch_bitmask = 0;
1364 
1365 	while ((unshare_watch_bitmask || hw->txrx_stop_req_bit) &&
1366 	       (wait_time < 3000)) {
1367 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
1368 			if (epidx == hw->my_epid)
1369 				continue;
1370 
1371 			is_shared = fjes_hw_epid_is_shared(hw->hw_info.share,
1372 							   epidx);
1373 
1374 			stop_req = test_bit(epidx, &hw->txrx_stop_req_bit);
1375 
1376 			stop_req_done = hw->ep_shm_info[epidx].rx.info->v1i.rx_status &
1377 					FJES_RX_STOP_REQ_DONE;
1378 
1379 			unshare_watch = test_bit(epidx, &unshare_watch_bitmask);
1380 
1381 			unshare_reserve = test_bit(epidx,
1382 						   &hw->hw_info.buffer_unshare_reserve_bit);
1383 
1384 			if ((!stop_req ||
1385 			     (is_shared && (!is_shared || !stop_req_done))) &&
1386 			    (is_shared || !unshare_watch || !unshare_reserve))
1387 				continue;
1388 
1389 			mutex_lock(&hw->hw_info.lock);
1390 			ret = fjes_hw_unregister_buff_addr(hw, epidx);
1391 			switch (ret) {
1392 			case 0:
1393 				break;
1394 			case -ENOMSG:
1395 			case -EBUSY:
1396 			default:
1397 				if (!work_pending(
1398 					&adapter->force_close_task)) {
1399 					adapter->force_reset = true;
1400 					schedule_work(
1401 						&adapter->force_close_task);
1402 				}
1403 				break;
1404 			}
1405 			mutex_unlock(&hw->hw_info.lock);
1406 			hw->ep_shm_info[epidx].ep_stats
1407 					.com_unregist_buf_exec += 1;
1408 
1409 			spin_lock_irqsave(&hw->rx_status_lock, flags);
1410 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
1411 					    netdev->dev_addr, netdev->mtu);
1412 			spin_unlock_irqrestore(&hw->rx_status_lock, flags);
1413 
1414 			clear_bit(epidx, &hw->txrx_stop_req_bit);
1415 			clear_bit(epidx, &unshare_watch_bitmask);
1416 			clear_bit(epidx,
1417 				  &hw->hw_info.buffer_unshare_reserve_bit);
1418 		}
1419 
1420 		msleep(100);
1421 		wait_time += 100;
1422 	}
1423 
1424 	if (hw->hw_info.buffer_unshare_reserve_bit) {
1425 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
1426 			if (epidx == hw->my_epid)
1427 				continue;
1428 
1429 			if (test_bit(epidx,
1430 				     &hw->hw_info.buffer_unshare_reserve_bit)) {
1431 				mutex_lock(&hw->hw_info.lock);
1432 
1433 				ret = fjes_hw_unregister_buff_addr(hw, epidx);
1434 				switch (ret) {
1435 				case 0:
1436 					break;
1437 				case -ENOMSG:
1438 				case -EBUSY:
1439 				default:
1440 					if (!work_pending(
1441 						&adapter->force_close_task)) {
1442 						adapter->force_reset = true;
1443 						schedule_work(
1444 							&adapter->force_close_task);
1445 					}
1446 					break;
1447 				}
1448 				mutex_unlock(&hw->hw_info.lock);
1449 
1450 				hw->ep_shm_info[epidx].ep_stats
1451 					.com_unregist_buf_exec += 1;
1452 
1453 				spin_lock_irqsave(&hw->rx_status_lock, flags);
1454 				fjes_hw_setup_epbuf(
1455 					&hw->ep_shm_info[epidx].tx,
1456 					netdev->dev_addr, netdev->mtu);
1457 				spin_unlock_irqrestore(&hw->rx_status_lock,
1458 						       flags);
1459 
1460 				clear_bit(epidx, &hw->txrx_stop_req_bit);
1461 				clear_bit(epidx, &unshare_watch_bitmask);
1462 				clear_bit(epidx, &hw->hw_info.buffer_unshare_reserve_bit);
1463 			}
1464 
1465 			if (test_bit(epidx, &unshare_watch_bitmask)) {
1466 				spin_lock_irqsave(&hw->rx_status_lock, flags);
1467 				hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
1468 						~FJES_RX_STOP_REQ_DONE;
1469 				spin_unlock_irqrestore(&hw->rx_status_lock,
1470 						       flags);
1471 			}
1472 		}
1473 	}
1474 }
1475 
1476 /* fjes_init_module - Driver Registration Routine */
1477 static int __init fjes_init_module(void)
1478 {
1479 	int result;
1480 
1481 	pr_info("%s - version %s - %s\n",
1482 		fjes_driver_string, fjes_driver_version, fjes_copyright);
1483 
1484 	fjes_dbg_init();
1485 
1486 	result = platform_driver_register(&fjes_driver);
1487 	if (result < 0) {
1488 		fjes_dbg_exit();
1489 		return result;
1490 	}
1491 
1492 	result = acpi_bus_register_driver(&fjes_acpi_driver);
1493 	if (result < 0)
1494 		goto fail_acpi_driver;
1495 
1496 	return 0;
1497 
1498 fail_acpi_driver:
1499 	platform_driver_unregister(&fjes_driver);
1500 	fjes_dbg_exit();
1501 	return result;
1502 }
1503 
1504 module_init(fjes_init_module);
1505 
1506 /* fjes_exit_module - Driver Exit Cleanup Routine */
1507 static void __exit fjes_exit_module(void)
1508 {
1509 	acpi_bus_unregister_driver(&fjes_acpi_driver);
1510 	platform_driver_unregister(&fjes_driver);
1511 	fjes_dbg_exit();
1512 }
1513 
1514 module_exit(fjes_exit_module);
1515