xref: /openbmc/linux/drivers/net/fjes/fjes_main.c (revision 03b262f2)
1 /*
2  *  FUJITSU Extended Socket Network Device driver
3  *  Copyright (c) 2015 FUJITSU LIMITED
4  *
5  * This program is free software; you can redistribute it and/or modify it
6  * under the terms and conditions of the GNU General Public License,
7  * version 2, as published by the Free Software Foundation.
8  *
9  * This program is distributed in the hope it will be useful, but WITHOUT
10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
12  * more details.
13  *
14  * You should have received a copy of the GNU General Public License along with
15  * this program; if not, see <http://www.gnu.org/licenses/>.
16  *
17  * The full GNU General Public License is included in this distribution in
18  * the file called "COPYING".
19  *
20  */
21 
22 #include <linux/module.h>
23 #include <linux/types.h>
24 #include <linux/nls.h>
25 #include <linux/platform_device.h>
26 #include <linux/netdevice.h>
27 #include <linux/interrupt.h>
28 
29 #include "fjes.h"
30 
31 #define MAJ 1
32 #define MIN 1
33 #define DRV_VERSION __stringify(MAJ) "." __stringify(MIN)
34 #define DRV_NAME	"fjes"
35 char fjes_driver_name[] = DRV_NAME;
36 char fjes_driver_version[] = DRV_VERSION;
37 static const char fjes_driver_string[] =
38 		"FUJITSU Extended Socket Network Device Driver";
39 static const char fjes_copyright[] =
40 		"Copyright (c) 2015 FUJITSU LIMITED";
41 
42 MODULE_AUTHOR("Taku Izumi <izumi.taku@jp.fujitsu.com>");
43 MODULE_DESCRIPTION("FUJITSU Extended Socket Network Device Driver");
44 MODULE_LICENSE("GPL");
45 MODULE_VERSION(DRV_VERSION);
46 
47 static int fjes_request_irq(struct fjes_adapter *);
48 static void fjes_free_irq(struct fjes_adapter *);
49 
50 static int fjes_open(struct net_device *);
51 static int fjes_close(struct net_device *);
52 static int fjes_setup_resources(struct fjes_adapter *);
53 static void fjes_free_resources(struct fjes_adapter *);
54 static netdev_tx_t fjes_xmit_frame(struct sk_buff *, struct net_device *);
55 static void fjes_raise_intr_rxdata_task(struct work_struct *);
56 static void fjes_tx_stall_task(struct work_struct *);
57 static void fjes_force_close_task(struct work_struct *);
58 static irqreturn_t fjes_intr(int, void*);
59 static struct rtnl_link_stats64 *
60 fjes_get_stats64(struct net_device *, struct rtnl_link_stats64 *);
61 static int fjes_change_mtu(struct net_device *, int);
62 static int fjes_vlan_rx_add_vid(struct net_device *, __be16 proto, u16);
63 static int fjes_vlan_rx_kill_vid(struct net_device *, __be16 proto, u16);
64 static void fjes_tx_retry(struct net_device *);
65 
66 static int fjes_acpi_add(struct acpi_device *);
67 static int fjes_acpi_remove(struct acpi_device *);
68 static acpi_status fjes_get_acpi_resource(struct acpi_resource *, void*);
69 
70 static int fjes_probe(struct platform_device *);
71 static int fjes_remove(struct platform_device *);
72 
73 static int fjes_sw_init(struct fjes_adapter *);
74 static void fjes_netdev_setup(struct net_device *);
75 static void fjes_irq_watch_task(struct work_struct *);
76 static void fjes_watch_unshare_task(struct work_struct *);
77 static void fjes_rx_irq(struct fjes_adapter *, int);
78 static int fjes_poll(struct napi_struct *, int);
79 
80 static const struct acpi_device_id fjes_acpi_ids[] = {
81 	{"PNP0C02", 0},
82 	{"", 0},
83 };
84 MODULE_DEVICE_TABLE(acpi, fjes_acpi_ids);
85 
86 static struct acpi_driver fjes_acpi_driver = {
87 	.name = DRV_NAME,
88 	.class = DRV_NAME,
89 	.owner = THIS_MODULE,
90 	.ids = fjes_acpi_ids,
91 	.ops = {
92 		.add = fjes_acpi_add,
93 		.remove = fjes_acpi_remove,
94 	},
95 };
96 
97 static struct platform_driver fjes_driver = {
98 	.driver = {
99 		.name = DRV_NAME,
100 		.owner = THIS_MODULE,
101 	},
102 	.probe = fjes_probe,
103 	.remove = fjes_remove,
104 };
105 
106 static struct resource fjes_resource[] = {
107 	{
108 		.flags = IORESOURCE_MEM,
109 		.start = 0,
110 		.end = 0,
111 	},
112 	{
113 		.flags = IORESOURCE_IRQ,
114 		.start = 0,
115 		.end = 0,
116 	},
117 };
118 
119 static int fjes_acpi_add(struct acpi_device *device)
120 {
121 	struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL};
122 	char str_buf[sizeof(FJES_ACPI_SYMBOL) + 1];
123 	struct platform_device *plat_dev;
124 	union acpi_object *str;
125 	acpi_status status;
126 	int result;
127 
128 	status = acpi_evaluate_object(device->handle, "_STR", NULL, &buffer);
129 	if (ACPI_FAILURE(status))
130 		return -ENODEV;
131 
132 	str = buffer.pointer;
133 	result = utf16s_to_utf8s((wchar_t *)str->string.pointer,
134 				 str->string.length, UTF16_LITTLE_ENDIAN,
135 				 str_buf, sizeof(str_buf) - 1);
136 	str_buf[result] = 0;
137 
138 	if (strncmp(FJES_ACPI_SYMBOL, str_buf, strlen(FJES_ACPI_SYMBOL)) != 0) {
139 		kfree(buffer.pointer);
140 		return -ENODEV;
141 	}
142 	kfree(buffer.pointer);
143 
144 	status = acpi_walk_resources(device->handle, METHOD_NAME__CRS,
145 				     fjes_get_acpi_resource, fjes_resource);
146 	if (ACPI_FAILURE(status))
147 		return -ENODEV;
148 
149 	/* create platform_device */
150 	plat_dev = platform_device_register_simple(DRV_NAME, 0, fjes_resource,
151 						   ARRAY_SIZE(fjes_resource));
152 	device->driver_data = plat_dev;
153 
154 	return 0;
155 }
156 
157 static int fjes_acpi_remove(struct acpi_device *device)
158 {
159 	struct platform_device *plat_dev;
160 
161 	plat_dev = (struct platform_device *)acpi_driver_data(device);
162 	platform_device_unregister(plat_dev);
163 
164 	return 0;
165 }
166 
167 static acpi_status
168 fjes_get_acpi_resource(struct acpi_resource *acpi_res, void *data)
169 {
170 	struct acpi_resource_address32 *addr;
171 	struct acpi_resource_irq *irq;
172 	struct resource *res = data;
173 
174 	switch (acpi_res->type) {
175 	case ACPI_RESOURCE_TYPE_ADDRESS32:
176 		addr = &acpi_res->data.address32;
177 		res[0].start = addr->address.minimum;
178 		res[0].end = addr->address.minimum +
179 			addr->address.address_length - 1;
180 		break;
181 
182 	case ACPI_RESOURCE_TYPE_IRQ:
183 		irq = &acpi_res->data.irq;
184 		if (irq->interrupt_count != 1)
185 			return AE_ERROR;
186 		res[1].start = irq->interrupts[0];
187 		res[1].end = irq->interrupts[0];
188 		break;
189 
190 	default:
191 		break;
192 	}
193 
194 	return AE_OK;
195 }
196 
197 static int fjes_request_irq(struct fjes_adapter *adapter)
198 {
199 	struct net_device *netdev = adapter->netdev;
200 	int result = -1;
201 
202 	adapter->interrupt_watch_enable = true;
203 	if (!delayed_work_pending(&adapter->interrupt_watch_task)) {
204 		queue_delayed_work(adapter->control_wq,
205 				   &adapter->interrupt_watch_task,
206 				   FJES_IRQ_WATCH_DELAY);
207 	}
208 
209 	if (!adapter->irq_registered) {
210 		result = request_irq(adapter->hw.hw_res.irq, fjes_intr,
211 				     IRQF_SHARED, netdev->name, adapter);
212 		if (result)
213 			adapter->irq_registered = false;
214 		else
215 			adapter->irq_registered = true;
216 	}
217 
218 	return result;
219 }
220 
221 static void fjes_free_irq(struct fjes_adapter *adapter)
222 {
223 	struct fjes_hw *hw = &adapter->hw;
224 
225 	adapter->interrupt_watch_enable = false;
226 	cancel_delayed_work_sync(&adapter->interrupt_watch_task);
227 
228 	fjes_hw_set_irqmask(hw, REG_ICTL_MASK_ALL, true);
229 
230 	if (adapter->irq_registered) {
231 		free_irq(adapter->hw.hw_res.irq, adapter);
232 		adapter->irq_registered = false;
233 	}
234 }
235 
236 static const struct net_device_ops fjes_netdev_ops = {
237 	.ndo_open		= fjes_open,
238 	.ndo_stop		= fjes_close,
239 	.ndo_start_xmit		= fjes_xmit_frame,
240 	.ndo_get_stats64	= fjes_get_stats64,
241 	.ndo_change_mtu		= fjes_change_mtu,
242 	.ndo_tx_timeout		= fjes_tx_retry,
243 	.ndo_vlan_rx_add_vid	= fjes_vlan_rx_add_vid,
244 	.ndo_vlan_rx_kill_vid = fjes_vlan_rx_kill_vid,
245 };
246 
247 /* fjes_open - Called when a network interface is made active */
248 static int fjes_open(struct net_device *netdev)
249 {
250 	struct fjes_adapter *adapter = netdev_priv(netdev);
251 	struct fjes_hw *hw = &adapter->hw;
252 	int result;
253 
254 	if (adapter->open_guard)
255 		return -ENXIO;
256 
257 	result = fjes_setup_resources(adapter);
258 	if (result)
259 		goto err_setup_res;
260 
261 	hw->txrx_stop_req_bit = 0;
262 	hw->epstop_req_bit = 0;
263 
264 	napi_enable(&adapter->napi);
265 
266 	fjes_hw_capture_interrupt_status(hw);
267 
268 	result = fjes_request_irq(adapter);
269 	if (result)
270 		goto err_req_irq;
271 
272 	fjes_hw_set_irqmask(hw, REG_ICTL_MASK_ALL, false);
273 
274 	netif_tx_start_all_queues(netdev);
275 	netif_carrier_on(netdev);
276 
277 	return 0;
278 
279 err_req_irq:
280 	fjes_free_irq(adapter);
281 	napi_disable(&adapter->napi);
282 
283 err_setup_res:
284 	fjes_free_resources(adapter);
285 	return result;
286 }
287 
288 /* fjes_close - Disables a network interface */
289 static int fjes_close(struct net_device *netdev)
290 {
291 	struct fjes_adapter *adapter = netdev_priv(netdev);
292 	struct fjes_hw *hw = &adapter->hw;
293 	unsigned long flags;
294 	int epidx;
295 
296 	netif_tx_stop_all_queues(netdev);
297 	netif_carrier_off(netdev);
298 
299 	fjes_hw_raise_epstop(hw);
300 
301 	napi_disable(&adapter->napi);
302 
303 	spin_lock_irqsave(&hw->rx_status_lock, flags);
304 	for (epidx = 0; epidx < hw->max_epid; epidx++) {
305 		if (epidx == hw->my_epid)
306 			continue;
307 
308 		if (fjes_hw_get_partner_ep_status(hw, epidx) ==
309 		    EP_PARTNER_SHARED)
310 			adapter->hw.ep_shm_info[epidx]
311 				   .tx.info->v1i.rx_status &=
312 				~FJES_RX_POLL_WORK;
313 	}
314 	spin_unlock_irqrestore(&hw->rx_status_lock, flags);
315 
316 	fjes_free_irq(adapter);
317 
318 	cancel_delayed_work_sync(&adapter->interrupt_watch_task);
319 	cancel_work_sync(&adapter->unshare_watch_task);
320 	adapter->unshare_watch_bitmask = 0;
321 	cancel_work_sync(&adapter->raise_intr_rxdata_task);
322 	cancel_work_sync(&adapter->tx_stall_task);
323 
324 	cancel_work_sync(&hw->update_zone_task);
325 	cancel_work_sync(&hw->epstop_task);
326 
327 	fjes_hw_wait_epstop(hw);
328 
329 	fjes_free_resources(adapter);
330 
331 	return 0;
332 }
333 
334 static int fjes_setup_resources(struct fjes_adapter *adapter)
335 {
336 	struct net_device *netdev = adapter->netdev;
337 	struct ep_share_mem_info *buf_pair;
338 	struct fjes_hw *hw = &adapter->hw;
339 	unsigned long flags;
340 	int result;
341 	int epidx;
342 
343 	mutex_lock(&hw->hw_info.lock);
344 	result = fjes_hw_request_info(hw);
345 	switch (result) {
346 	case 0:
347 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
348 			hw->ep_shm_info[epidx].es_status =
349 			    hw->hw_info.res_buf->info.info[epidx].es_status;
350 			hw->ep_shm_info[epidx].zone =
351 			    hw->hw_info.res_buf->info.info[epidx].zone;
352 		}
353 		break;
354 	default:
355 	case -ENOMSG:
356 	case -EBUSY:
357 		adapter->force_reset = true;
358 
359 		mutex_unlock(&hw->hw_info.lock);
360 		return result;
361 	}
362 	mutex_unlock(&hw->hw_info.lock);
363 
364 	for (epidx = 0; epidx < (hw->max_epid); epidx++) {
365 		if ((epidx != hw->my_epid) &&
366 		    (hw->ep_shm_info[epidx].es_status ==
367 		     FJES_ZONING_STATUS_ENABLE)) {
368 			fjes_hw_raise_interrupt(hw, epidx,
369 						REG_ICTL_MASK_INFO_UPDATE);
370 		}
371 	}
372 
373 	msleep(FJES_OPEN_ZONE_UPDATE_WAIT * hw->max_epid);
374 
375 	for (epidx = 0; epidx < (hw->max_epid); epidx++) {
376 		if (epidx == hw->my_epid)
377 			continue;
378 
379 		buf_pair = &hw->ep_shm_info[epidx];
380 
381 		spin_lock_irqsave(&hw->rx_status_lock, flags);
382 		fjes_hw_setup_epbuf(&buf_pair->tx, netdev->dev_addr,
383 				    netdev->mtu);
384 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
385 
386 		if (fjes_hw_epid_is_same_zone(hw, epidx)) {
387 			mutex_lock(&hw->hw_info.lock);
388 			result =
389 			fjes_hw_register_buff_addr(hw, epidx, buf_pair);
390 			mutex_unlock(&hw->hw_info.lock);
391 
392 			switch (result) {
393 			case 0:
394 				break;
395 			case -ENOMSG:
396 			case -EBUSY:
397 			default:
398 				adapter->force_reset = true;
399 				return result;
400 			}
401 		}
402 	}
403 
404 	return 0;
405 }
406 
407 static void fjes_free_resources(struct fjes_adapter *adapter)
408 {
409 	struct net_device *netdev = adapter->netdev;
410 	struct fjes_device_command_param param;
411 	struct ep_share_mem_info *buf_pair;
412 	struct fjes_hw *hw = &adapter->hw;
413 	bool reset_flag = false;
414 	unsigned long flags;
415 	int result;
416 	int epidx;
417 
418 	for (epidx = 0; epidx < hw->max_epid; epidx++) {
419 		if (epidx == hw->my_epid)
420 			continue;
421 
422 		mutex_lock(&hw->hw_info.lock);
423 		result = fjes_hw_unregister_buff_addr(hw, epidx);
424 		mutex_unlock(&hw->hw_info.lock);
425 
426 		if (result)
427 			reset_flag = true;
428 
429 		buf_pair = &hw->ep_shm_info[epidx];
430 
431 		spin_lock_irqsave(&hw->rx_status_lock, flags);
432 		fjes_hw_setup_epbuf(&buf_pair->tx,
433 				    netdev->dev_addr, netdev->mtu);
434 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
435 
436 		clear_bit(epidx, &hw->txrx_stop_req_bit);
437 	}
438 
439 	if (reset_flag || adapter->force_reset) {
440 		result = fjes_hw_reset(hw);
441 
442 		adapter->force_reset = false;
443 
444 		if (result)
445 			adapter->open_guard = true;
446 
447 		hw->hw_info.buffer_share_bit = 0;
448 
449 		memset((void *)&param, 0, sizeof(param));
450 
451 		param.req_len = hw->hw_info.req_buf_size;
452 		param.req_start = __pa(hw->hw_info.req_buf);
453 		param.res_len = hw->hw_info.res_buf_size;
454 		param.res_start = __pa(hw->hw_info.res_buf);
455 		param.share_start = __pa(hw->hw_info.share->ep_status);
456 
457 		fjes_hw_init_command_registers(hw, &param);
458 	}
459 }
460 
461 static void fjes_tx_stall_task(struct work_struct *work)
462 {
463 	struct fjes_adapter *adapter = container_of(work,
464 			struct fjes_adapter, tx_stall_task);
465 	struct net_device *netdev = adapter->netdev;
466 	struct fjes_hw *hw = &adapter->hw;
467 	int all_queue_available, sendable;
468 	enum ep_partner_status pstatus;
469 	int max_epid, my_epid, epid;
470 	union ep_buffer_info *info;
471 	int i;
472 
473 	if (((long)jiffies -
474 		dev_trans_start(netdev)) > FJES_TX_TX_STALL_TIMEOUT) {
475 		netif_wake_queue(netdev);
476 		return;
477 	}
478 
479 	my_epid = hw->my_epid;
480 	max_epid = hw->max_epid;
481 
482 	for (i = 0; i < 5; i++) {
483 		all_queue_available = 1;
484 
485 		for (epid = 0; epid < max_epid; epid++) {
486 			if (my_epid == epid)
487 				continue;
488 
489 			pstatus = fjes_hw_get_partner_ep_status(hw, epid);
490 			sendable = (pstatus == EP_PARTNER_SHARED);
491 			if (!sendable)
492 				continue;
493 
494 			info = adapter->hw.ep_shm_info[epid].tx.info;
495 
496 			if (!(info->v1i.rx_status & FJES_RX_MTU_CHANGING_DONE))
497 				return;
498 
499 			if (EP_RING_FULL(info->v1i.head, info->v1i.tail,
500 					 info->v1i.count_max)) {
501 				all_queue_available = 0;
502 				break;
503 			}
504 		}
505 
506 		if (all_queue_available) {
507 			netif_wake_queue(netdev);
508 			return;
509 		}
510 	}
511 
512 	usleep_range(50, 100);
513 
514 	queue_work(adapter->txrx_wq, &adapter->tx_stall_task);
515 }
516 
517 static void fjes_force_close_task(struct work_struct *work)
518 {
519 	struct fjes_adapter *adapter = container_of(work,
520 			struct fjes_adapter, force_close_task);
521 	struct net_device *netdev = adapter->netdev;
522 
523 	rtnl_lock();
524 	dev_close(netdev);
525 	rtnl_unlock();
526 }
527 
528 static void fjes_raise_intr_rxdata_task(struct work_struct *work)
529 {
530 	struct fjes_adapter *adapter = container_of(work,
531 			struct fjes_adapter, raise_intr_rxdata_task);
532 	struct fjes_hw *hw = &adapter->hw;
533 	enum ep_partner_status pstatus;
534 	int max_epid, my_epid, epid;
535 
536 	my_epid = hw->my_epid;
537 	max_epid = hw->max_epid;
538 
539 	for (epid = 0; epid < max_epid; epid++)
540 		hw->ep_shm_info[epid].tx_status_work = 0;
541 
542 	for (epid = 0; epid < max_epid; epid++) {
543 		if (epid == my_epid)
544 			continue;
545 
546 		pstatus = fjes_hw_get_partner_ep_status(hw, epid);
547 		if (pstatus == EP_PARTNER_SHARED) {
548 			hw->ep_shm_info[epid].tx_status_work =
549 				hw->ep_shm_info[epid].tx.info->v1i.tx_status;
550 
551 			if (hw->ep_shm_info[epid].tx_status_work ==
552 				FJES_TX_DELAY_SEND_PENDING) {
553 				hw->ep_shm_info[epid].tx.info->v1i.tx_status =
554 					FJES_TX_DELAY_SEND_NONE;
555 			}
556 		}
557 	}
558 
559 	for (epid = 0; epid < max_epid; epid++) {
560 		if (epid == my_epid)
561 			continue;
562 
563 		pstatus = fjes_hw_get_partner_ep_status(hw, epid);
564 		if ((hw->ep_shm_info[epid].tx_status_work ==
565 		     FJES_TX_DELAY_SEND_PENDING) &&
566 		    (pstatus == EP_PARTNER_SHARED) &&
567 		    !(hw->ep_shm_info[epid].rx.info->v1i.rx_status &
568 		      FJES_RX_POLL_WORK)) {
569 			fjes_hw_raise_interrupt(hw, epid,
570 						REG_ICTL_MASK_RX_DATA);
571 		}
572 	}
573 
574 	usleep_range(500, 1000);
575 }
576 
577 static int fjes_tx_send(struct fjes_adapter *adapter, int dest,
578 			void *data, size_t len)
579 {
580 	int retval;
581 
582 	retval = fjes_hw_epbuf_tx_pkt_send(&adapter->hw.ep_shm_info[dest].tx,
583 					   data, len);
584 	if (retval)
585 		return retval;
586 
587 	adapter->hw.ep_shm_info[dest].tx.info->v1i.tx_status =
588 		FJES_TX_DELAY_SEND_PENDING;
589 	if (!work_pending(&adapter->raise_intr_rxdata_task))
590 		queue_work(adapter->txrx_wq,
591 			   &adapter->raise_intr_rxdata_task);
592 
593 	retval = 0;
594 	return retval;
595 }
596 
597 static netdev_tx_t
598 fjes_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
599 {
600 	struct fjes_adapter *adapter = netdev_priv(netdev);
601 	struct fjes_hw *hw = &adapter->hw;
602 
603 	int max_epid, my_epid, dest_epid;
604 	enum ep_partner_status pstatus;
605 	struct netdev_queue *cur_queue;
606 	char shortpkt[VLAN_ETH_HLEN];
607 	bool is_multi, vlan;
608 	struct ethhdr *eth;
609 	u16 queue_no = 0;
610 	u16 vlan_id = 0;
611 	netdev_tx_t ret;
612 	char *data;
613 	int len;
614 
615 	ret = NETDEV_TX_OK;
616 	is_multi = false;
617 	cur_queue = netdev_get_tx_queue(netdev, queue_no);
618 
619 	eth = (struct ethhdr *)skb->data;
620 	my_epid = hw->my_epid;
621 
622 	vlan = (vlan_get_tag(skb, &vlan_id) == 0) ? true : false;
623 
624 	data = skb->data;
625 	len = skb->len;
626 
627 	if (is_multicast_ether_addr(eth->h_dest)) {
628 		dest_epid = 0;
629 		max_epid = hw->max_epid;
630 		is_multi = true;
631 	} else if (is_local_ether_addr(eth->h_dest)) {
632 		dest_epid = eth->h_dest[ETH_ALEN - 1];
633 		max_epid = dest_epid + 1;
634 
635 		if ((eth->h_dest[0] == 0x02) &&
636 		    (0x00 == (eth->h_dest[1] | eth->h_dest[2] |
637 			      eth->h_dest[3] | eth->h_dest[4])) &&
638 		    (dest_epid < hw->max_epid)) {
639 			;
640 		} else {
641 			dest_epid = 0;
642 			max_epid = 0;
643 			ret = NETDEV_TX_OK;
644 
645 			adapter->stats64.tx_packets += 1;
646 			hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
647 			adapter->stats64.tx_bytes += len;
648 			hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
649 		}
650 	} else {
651 		dest_epid = 0;
652 		max_epid = 0;
653 		ret = NETDEV_TX_OK;
654 
655 		adapter->stats64.tx_packets += 1;
656 		hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
657 		adapter->stats64.tx_bytes += len;
658 		hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
659 	}
660 
661 	for (; dest_epid < max_epid; dest_epid++) {
662 		if (my_epid == dest_epid)
663 			continue;
664 
665 		pstatus = fjes_hw_get_partner_ep_status(hw, dest_epid);
666 		if (pstatus != EP_PARTNER_SHARED) {
667 			ret = NETDEV_TX_OK;
668 		} else if (!fjes_hw_check_epbuf_version(
669 				&adapter->hw.ep_shm_info[dest_epid].rx, 0)) {
670 			/* version is NOT 0 */
671 			adapter->stats64.tx_carrier_errors += 1;
672 			hw->ep_shm_info[dest_epid].net_stats
673 						.tx_carrier_errors += 1;
674 
675 			ret = NETDEV_TX_OK;
676 		} else if (!fjes_hw_check_mtu(
677 				&adapter->hw.ep_shm_info[dest_epid].rx,
678 				netdev->mtu)) {
679 			adapter->stats64.tx_dropped += 1;
680 			hw->ep_shm_info[dest_epid].net_stats.tx_dropped += 1;
681 			adapter->stats64.tx_errors += 1;
682 			hw->ep_shm_info[dest_epid].net_stats.tx_errors += 1;
683 
684 			ret = NETDEV_TX_OK;
685 		} else if (vlan &&
686 			   !fjes_hw_check_vlan_id(
687 				&adapter->hw.ep_shm_info[dest_epid].rx,
688 				vlan_id)) {
689 			ret = NETDEV_TX_OK;
690 		} else {
691 			if (len < VLAN_ETH_HLEN) {
692 				memset(shortpkt, 0, VLAN_ETH_HLEN);
693 				memcpy(shortpkt, skb->data, skb->len);
694 				len = VLAN_ETH_HLEN;
695 				data = shortpkt;
696 			}
697 
698 			if (adapter->tx_retry_count == 0) {
699 				adapter->tx_start_jiffies = jiffies;
700 				adapter->tx_retry_count = 1;
701 			} else {
702 				adapter->tx_retry_count++;
703 			}
704 
705 			if (fjes_tx_send(adapter, dest_epid, data, len)) {
706 				if (is_multi) {
707 					ret = NETDEV_TX_OK;
708 				} else if (
709 					   ((long)jiffies -
710 					    (long)adapter->tx_start_jiffies) >=
711 					    FJES_TX_RETRY_TIMEOUT) {
712 					adapter->stats64.tx_fifo_errors += 1;
713 					hw->ep_shm_info[dest_epid].net_stats
714 								.tx_fifo_errors += 1;
715 					adapter->stats64.tx_errors += 1;
716 					hw->ep_shm_info[dest_epid].net_stats
717 								.tx_errors += 1;
718 
719 					ret = NETDEV_TX_OK;
720 				} else {
721 					netif_trans_update(netdev);
722 					netif_tx_stop_queue(cur_queue);
723 
724 					if (!work_pending(&adapter->tx_stall_task))
725 						queue_work(adapter->txrx_wq,
726 							   &adapter->tx_stall_task);
727 
728 					ret = NETDEV_TX_BUSY;
729 				}
730 			} else {
731 				if (!is_multi) {
732 					adapter->stats64.tx_packets += 1;
733 					hw->ep_shm_info[dest_epid].net_stats
734 								.tx_packets += 1;
735 					adapter->stats64.tx_bytes += len;
736 					hw->ep_shm_info[dest_epid].net_stats
737 								.tx_bytes += len;
738 				}
739 
740 				adapter->tx_retry_count = 0;
741 				ret = NETDEV_TX_OK;
742 			}
743 		}
744 	}
745 
746 	if (ret == NETDEV_TX_OK) {
747 		dev_kfree_skb(skb);
748 		if (is_multi) {
749 			adapter->stats64.tx_packets += 1;
750 			hw->ep_shm_info[my_epid].net_stats.tx_packets += 1;
751 			adapter->stats64.tx_bytes += 1;
752 			hw->ep_shm_info[my_epid].net_stats.tx_bytes += len;
753 		}
754 	}
755 
756 	return ret;
757 }
758 
759 static void fjes_tx_retry(struct net_device *netdev)
760 {
761 	struct netdev_queue *queue = netdev_get_tx_queue(netdev, 0);
762 
763 	netif_tx_wake_queue(queue);
764 }
765 
766 static struct rtnl_link_stats64 *
767 fjes_get_stats64(struct net_device *netdev, struct rtnl_link_stats64 *stats)
768 {
769 	struct fjes_adapter *adapter = netdev_priv(netdev);
770 
771 	memcpy(stats, &adapter->stats64, sizeof(struct rtnl_link_stats64));
772 
773 	return stats;
774 }
775 
776 static int fjes_change_mtu(struct net_device *netdev, int new_mtu)
777 {
778 	struct fjes_adapter *adapter = netdev_priv(netdev);
779 	bool running = netif_running(netdev);
780 	struct fjes_hw *hw = &adapter->hw;
781 	unsigned long flags;
782 	int ret = -EINVAL;
783 	int idx, epidx;
784 
785 	for (idx = 0; fjes_support_mtu[idx] != 0; idx++) {
786 		if (new_mtu <= fjes_support_mtu[idx]) {
787 			new_mtu = fjes_support_mtu[idx];
788 			if (new_mtu == netdev->mtu)
789 				return 0;
790 
791 			ret = 0;
792 			break;
793 		}
794 	}
795 
796 	if (ret)
797 		return ret;
798 
799 	if (running) {
800 		spin_lock_irqsave(&hw->rx_status_lock, flags);
801 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
802 			if (epidx == hw->my_epid)
803 				continue;
804 			hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
805 				~FJES_RX_MTU_CHANGING_DONE;
806 		}
807 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
808 
809 		netif_tx_stop_all_queues(netdev);
810 		netif_carrier_off(netdev);
811 		cancel_work_sync(&adapter->tx_stall_task);
812 		napi_disable(&adapter->napi);
813 
814 		msleep(1000);
815 
816 		netif_tx_stop_all_queues(netdev);
817 	}
818 
819 	netdev->mtu = new_mtu;
820 
821 	if (running) {
822 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
823 			if (epidx == hw->my_epid)
824 				continue;
825 
826 			spin_lock_irqsave(&hw->rx_status_lock, flags);
827 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
828 					    netdev->dev_addr,
829 					    netdev->mtu);
830 
831 			hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
832 				FJES_RX_MTU_CHANGING_DONE;
833 			spin_unlock_irqrestore(&hw->rx_status_lock, flags);
834 		}
835 
836 		netif_tx_wake_all_queues(netdev);
837 		netif_carrier_on(netdev);
838 		napi_enable(&adapter->napi);
839 		napi_schedule(&adapter->napi);
840 	}
841 
842 	return ret;
843 }
844 
845 static int fjes_vlan_rx_add_vid(struct net_device *netdev,
846 				__be16 proto, u16 vid)
847 {
848 	struct fjes_adapter *adapter = netdev_priv(netdev);
849 	bool ret = true;
850 	int epid;
851 
852 	for (epid = 0; epid < adapter->hw.max_epid; epid++) {
853 		if (epid == adapter->hw.my_epid)
854 			continue;
855 
856 		if (!fjes_hw_check_vlan_id(
857 			&adapter->hw.ep_shm_info[epid].tx, vid))
858 			ret = fjes_hw_set_vlan_id(
859 				&adapter->hw.ep_shm_info[epid].tx, vid);
860 	}
861 
862 	return ret ? 0 : -ENOSPC;
863 }
864 
865 static int fjes_vlan_rx_kill_vid(struct net_device *netdev,
866 				 __be16 proto, u16 vid)
867 {
868 	struct fjes_adapter *adapter = netdev_priv(netdev);
869 	int epid;
870 
871 	for (epid = 0; epid < adapter->hw.max_epid; epid++) {
872 		if (epid == adapter->hw.my_epid)
873 			continue;
874 
875 		fjes_hw_del_vlan_id(&adapter->hw.ep_shm_info[epid].tx, vid);
876 	}
877 
878 	return 0;
879 }
880 
881 static void fjes_txrx_stop_req_irq(struct fjes_adapter *adapter,
882 				   int src_epid)
883 {
884 	struct fjes_hw *hw = &adapter->hw;
885 	enum ep_partner_status status;
886 	unsigned long flags;
887 
888 	status = fjes_hw_get_partner_ep_status(hw, src_epid);
889 	switch (status) {
890 	case EP_PARTNER_UNSHARE:
891 	case EP_PARTNER_COMPLETE:
892 	default:
893 		break;
894 	case EP_PARTNER_WAITING:
895 		if (src_epid < hw->my_epid) {
896 			spin_lock_irqsave(&hw->rx_status_lock, flags);
897 			hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
898 				FJES_RX_STOP_REQ_DONE;
899 			spin_unlock_irqrestore(&hw->rx_status_lock, flags);
900 
901 			clear_bit(src_epid, &hw->txrx_stop_req_bit);
902 			set_bit(src_epid, &adapter->unshare_watch_bitmask);
903 
904 			if (!work_pending(&adapter->unshare_watch_task))
905 				queue_work(adapter->control_wq,
906 					   &adapter->unshare_watch_task);
907 		}
908 		break;
909 	case EP_PARTNER_SHARED:
910 		if (hw->ep_shm_info[src_epid].rx.info->v1i.rx_status &
911 		    FJES_RX_STOP_REQ_REQUEST) {
912 			set_bit(src_epid, &hw->epstop_req_bit);
913 			if (!work_pending(&hw->epstop_task))
914 				queue_work(adapter->control_wq,
915 					   &hw->epstop_task);
916 		}
917 		break;
918 	}
919 }
920 
921 static void fjes_stop_req_irq(struct fjes_adapter *adapter, int src_epid)
922 {
923 	struct fjes_hw *hw = &adapter->hw;
924 	enum ep_partner_status status;
925 	unsigned long flags;
926 
927 	set_bit(src_epid, &hw->hw_info.buffer_unshare_reserve_bit);
928 
929 	status = fjes_hw_get_partner_ep_status(hw, src_epid);
930 	switch (status) {
931 	case EP_PARTNER_WAITING:
932 		spin_lock_irqsave(&hw->rx_status_lock, flags);
933 		hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
934 				FJES_RX_STOP_REQ_DONE;
935 		spin_unlock_irqrestore(&hw->rx_status_lock, flags);
936 		clear_bit(src_epid, &hw->txrx_stop_req_bit);
937 		/* fall through */
938 	case EP_PARTNER_UNSHARE:
939 	case EP_PARTNER_COMPLETE:
940 	default:
941 		set_bit(src_epid, &adapter->unshare_watch_bitmask);
942 		if (!work_pending(&adapter->unshare_watch_task))
943 			queue_work(adapter->control_wq,
944 				   &adapter->unshare_watch_task);
945 		break;
946 	case EP_PARTNER_SHARED:
947 		set_bit(src_epid, &hw->epstop_req_bit);
948 
949 		if (!work_pending(&hw->epstop_task))
950 			queue_work(adapter->control_wq, &hw->epstop_task);
951 		break;
952 	}
953 }
954 
955 static void fjes_update_zone_irq(struct fjes_adapter *adapter,
956 				 int src_epid)
957 {
958 	struct fjes_hw *hw = &adapter->hw;
959 
960 	if (!work_pending(&hw->update_zone_task))
961 		queue_work(adapter->control_wq, &hw->update_zone_task);
962 }
963 
964 static irqreturn_t fjes_intr(int irq, void *data)
965 {
966 	struct fjes_adapter *adapter = data;
967 	struct fjes_hw *hw = &adapter->hw;
968 	irqreturn_t ret;
969 	u32 icr;
970 
971 	icr = fjes_hw_capture_interrupt_status(hw);
972 
973 	if (icr & REG_IS_MASK_IS_ASSERT) {
974 		if (icr & REG_ICTL_MASK_RX_DATA)
975 			fjes_rx_irq(adapter, icr & REG_IS_MASK_EPID);
976 
977 		if (icr & REG_ICTL_MASK_DEV_STOP_REQ)
978 			fjes_stop_req_irq(adapter, icr & REG_IS_MASK_EPID);
979 
980 		if (icr & REG_ICTL_MASK_TXRX_STOP_REQ)
981 			fjes_txrx_stop_req_irq(adapter, icr & REG_IS_MASK_EPID);
982 
983 		if (icr & REG_ICTL_MASK_TXRX_STOP_DONE)
984 			fjes_hw_set_irqmask(hw,
985 					    REG_ICTL_MASK_TXRX_STOP_DONE, true);
986 
987 		if (icr & REG_ICTL_MASK_INFO_UPDATE)
988 			fjes_update_zone_irq(adapter, icr & REG_IS_MASK_EPID);
989 
990 		ret = IRQ_HANDLED;
991 	} else {
992 		ret = IRQ_NONE;
993 	}
994 
995 	return ret;
996 }
997 
998 static int fjes_rxframe_search_exist(struct fjes_adapter *adapter,
999 				     int start_epid)
1000 {
1001 	struct fjes_hw *hw = &adapter->hw;
1002 	enum ep_partner_status pstatus;
1003 	int max_epid, cur_epid;
1004 	int i;
1005 
1006 	max_epid = hw->max_epid;
1007 	start_epid = (start_epid + 1 + max_epid) % max_epid;
1008 
1009 	for (i = 0; i < max_epid; i++) {
1010 		cur_epid = (start_epid + i) % max_epid;
1011 		if (cur_epid == hw->my_epid)
1012 			continue;
1013 
1014 		pstatus = fjes_hw_get_partner_ep_status(hw, cur_epid);
1015 		if (pstatus == EP_PARTNER_SHARED) {
1016 			if (!fjes_hw_epbuf_rx_is_empty(
1017 				&hw->ep_shm_info[cur_epid].rx))
1018 				return cur_epid;
1019 		}
1020 	}
1021 	return -1;
1022 }
1023 
1024 static void *fjes_rxframe_get(struct fjes_adapter *adapter, size_t *psize,
1025 			      int *cur_epid)
1026 {
1027 	void *frame;
1028 
1029 	*cur_epid = fjes_rxframe_search_exist(adapter, *cur_epid);
1030 	if (*cur_epid < 0)
1031 		return NULL;
1032 
1033 	frame =
1034 	fjes_hw_epbuf_rx_curpkt_get_addr(
1035 		&adapter->hw.ep_shm_info[*cur_epid].rx, psize);
1036 
1037 	return frame;
1038 }
1039 
1040 static void fjes_rxframe_release(struct fjes_adapter *adapter, int cur_epid)
1041 {
1042 	fjes_hw_epbuf_rx_curpkt_drop(&adapter->hw.ep_shm_info[cur_epid].rx);
1043 }
1044 
1045 static void fjes_rx_irq(struct fjes_adapter *adapter, int src_epid)
1046 {
1047 	struct fjes_hw *hw = &adapter->hw;
1048 
1049 	fjes_hw_set_irqmask(hw, REG_ICTL_MASK_RX_DATA, true);
1050 
1051 	adapter->unset_rx_last = true;
1052 	napi_schedule(&adapter->napi);
1053 }
1054 
1055 static int fjes_poll(struct napi_struct *napi, int budget)
1056 {
1057 	struct fjes_adapter *adapter =
1058 			container_of(napi, struct fjes_adapter, napi);
1059 	struct net_device *netdev = napi->dev;
1060 	struct fjes_hw *hw = &adapter->hw;
1061 	struct sk_buff *skb;
1062 	int work_done = 0;
1063 	int cur_epid = 0;
1064 	int epidx;
1065 	size_t frame_len;
1066 	void *frame;
1067 
1068 	spin_lock(&hw->rx_status_lock);
1069 	for (epidx = 0; epidx < hw->max_epid; epidx++) {
1070 		if (epidx == hw->my_epid)
1071 			continue;
1072 
1073 		if (fjes_hw_get_partner_ep_status(hw, epidx) ==
1074 		    EP_PARTNER_SHARED)
1075 			adapter->hw.ep_shm_info[epidx]
1076 				   .tx.info->v1i.rx_status |= FJES_RX_POLL_WORK;
1077 	}
1078 	spin_unlock(&hw->rx_status_lock);
1079 
1080 	while (work_done < budget) {
1081 		prefetch(&adapter->hw);
1082 		frame = fjes_rxframe_get(adapter, &frame_len, &cur_epid);
1083 
1084 		if (frame) {
1085 			skb = napi_alloc_skb(napi, frame_len);
1086 			if (!skb) {
1087 				adapter->stats64.rx_dropped += 1;
1088 				hw->ep_shm_info[cur_epid].net_stats
1089 							 .rx_dropped += 1;
1090 				adapter->stats64.rx_errors += 1;
1091 				hw->ep_shm_info[cur_epid].net_stats
1092 							 .rx_errors += 1;
1093 			} else {
1094 				memcpy(skb_put(skb, frame_len),
1095 				       frame, frame_len);
1096 				skb->protocol = eth_type_trans(skb, netdev);
1097 				skb->ip_summed = CHECKSUM_UNNECESSARY;
1098 
1099 				netif_receive_skb(skb);
1100 
1101 				work_done++;
1102 
1103 				adapter->stats64.rx_packets += 1;
1104 				hw->ep_shm_info[cur_epid].net_stats
1105 							 .rx_packets += 1;
1106 				adapter->stats64.rx_bytes += frame_len;
1107 				hw->ep_shm_info[cur_epid].net_stats
1108 							 .rx_bytes += frame_len;
1109 
1110 				if (is_multicast_ether_addr(
1111 					((struct ethhdr *)frame)->h_dest)) {
1112 					adapter->stats64.multicast += 1;
1113 					hw->ep_shm_info[cur_epid].net_stats
1114 								 .multicast += 1;
1115 				}
1116 			}
1117 
1118 			fjes_rxframe_release(adapter, cur_epid);
1119 			adapter->unset_rx_last = true;
1120 		} else {
1121 			break;
1122 		}
1123 	}
1124 
1125 	if (work_done < budget) {
1126 		napi_complete(napi);
1127 
1128 		if (adapter->unset_rx_last) {
1129 			adapter->rx_last_jiffies = jiffies;
1130 			adapter->unset_rx_last = false;
1131 		}
1132 
1133 		if (((long)jiffies - (long)adapter->rx_last_jiffies) < 3) {
1134 			napi_reschedule(napi);
1135 		} else {
1136 			spin_lock(&hw->rx_status_lock);
1137 			for (epidx = 0; epidx < hw->max_epid; epidx++) {
1138 				if (epidx == hw->my_epid)
1139 					continue;
1140 				if (fjes_hw_get_partner_ep_status(hw, epidx) ==
1141 				    EP_PARTNER_SHARED)
1142 					adapter->hw.ep_shm_info[epidx].tx
1143 						   .info->v1i.rx_status &=
1144 						~FJES_RX_POLL_WORK;
1145 			}
1146 			spin_unlock(&hw->rx_status_lock);
1147 
1148 			fjes_hw_set_irqmask(hw, REG_ICTL_MASK_RX_DATA, false);
1149 		}
1150 	}
1151 
1152 	return work_done;
1153 }
1154 
1155 /* fjes_probe - Device Initialization Routine */
1156 static int fjes_probe(struct platform_device *plat_dev)
1157 {
1158 	struct fjes_adapter *adapter;
1159 	struct net_device *netdev;
1160 	struct resource *res;
1161 	struct fjes_hw *hw;
1162 	int err;
1163 
1164 	err = -ENOMEM;
1165 	netdev = alloc_netdev_mq(sizeof(struct fjes_adapter), "es%d",
1166 				 NET_NAME_UNKNOWN, fjes_netdev_setup,
1167 				 FJES_MAX_QUEUES);
1168 
1169 	if (!netdev)
1170 		goto err_out;
1171 
1172 	SET_NETDEV_DEV(netdev, &plat_dev->dev);
1173 
1174 	dev_set_drvdata(&plat_dev->dev, netdev);
1175 	adapter = netdev_priv(netdev);
1176 	adapter->netdev = netdev;
1177 	adapter->plat_dev = plat_dev;
1178 	hw = &adapter->hw;
1179 	hw->back = adapter;
1180 
1181 	/* setup the private structure */
1182 	err = fjes_sw_init(adapter);
1183 	if (err)
1184 		goto err_free_netdev;
1185 
1186 	INIT_WORK(&adapter->force_close_task, fjes_force_close_task);
1187 	adapter->force_reset = false;
1188 	adapter->open_guard = false;
1189 
1190 	adapter->txrx_wq = alloc_workqueue(DRV_NAME "/txrx", WQ_MEM_RECLAIM, 0);
1191 	adapter->control_wq = alloc_workqueue(DRV_NAME "/control",
1192 					      WQ_MEM_RECLAIM, 0);
1193 
1194 	INIT_WORK(&adapter->tx_stall_task, fjes_tx_stall_task);
1195 	INIT_WORK(&adapter->raise_intr_rxdata_task,
1196 		  fjes_raise_intr_rxdata_task);
1197 	INIT_WORK(&adapter->unshare_watch_task, fjes_watch_unshare_task);
1198 	adapter->unshare_watch_bitmask = 0;
1199 
1200 	INIT_DELAYED_WORK(&adapter->interrupt_watch_task, fjes_irq_watch_task);
1201 	adapter->interrupt_watch_enable = false;
1202 
1203 	res = platform_get_resource(plat_dev, IORESOURCE_MEM, 0);
1204 	hw->hw_res.start = res->start;
1205 	hw->hw_res.size = resource_size(res);
1206 	hw->hw_res.irq = platform_get_irq(plat_dev, 0);
1207 	err = fjes_hw_init(&adapter->hw);
1208 	if (err)
1209 		goto err_free_netdev;
1210 
1211 	/* setup MAC address (02:00:00:00:00:[epid])*/
1212 	netdev->dev_addr[0] = 2;
1213 	netdev->dev_addr[1] = 0;
1214 	netdev->dev_addr[2] = 0;
1215 	netdev->dev_addr[3] = 0;
1216 	netdev->dev_addr[4] = 0;
1217 	netdev->dev_addr[5] = hw->my_epid; /* EPID */
1218 
1219 	err = register_netdev(netdev);
1220 	if (err)
1221 		goto err_hw_exit;
1222 
1223 	netif_carrier_off(netdev);
1224 
1225 	return 0;
1226 
1227 err_hw_exit:
1228 	fjes_hw_exit(&adapter->hw);
1229 err_free_netdev:
1230 	free_netdev(netdev);
1231 err_out:
1232 	return err;
1233 }
1234 
1235 /* fjes_remove - Device Removal Routine */
1236 static int fjes_remove(struct platform_device *plat_dev)
1237 {
1238 	struct net_device *netdev = dev_get_drvdata(&plat_dev->dev);
1239 	struct fjes_adapter *adapter = netdev_priv(netdev);
1240 	struct fjes_hw *hw = &adapter->hw;
1241 
1242 	cancel_delayed_work_sync(&adapter->interrupt_watch_task);
1243 	cancel_work_sync(&adapter->unshare_watch_task);
1244 	cancel_work_sync(&adapter->raise_intr_rxdata_task);
1245 	cancel_work_sync(&adapter->tx_stall_task);
1246 	if (adapter->control_wq)
1247 		destroy_workqueue(adapter->control_wq);
1248 	if (adapter->txrx_wq)
1249 		destroy_workqueue(adapter->txrx_wq);
1250 
1251 	unregister_netdev(netdev);
1252 
1253 	fjes_hw_exit(hw);
1254 
1255 	netif_napi_del(&adapter->napi);
1256 
1257 	free_netdev(netdev);
1258 
1259 	return 0;
1260 }
1261 
1262 static int fjes_sw_init(struct fjes_adapter *adapter)
1263 {
1264 	struct net_device *netdev = adapter->netdev;
1265 
1266 	netif_napi_add(netdev, &adapter->napi, fjes_poll, 64);
1267 
1268 	return 0;
1269 }
1270 
1271 /* fjes_netdev_setup - netdevice initialization routine */
1272 static void fjes_netdev_setup(struct net_device *netdev)
1273 {
1274 	ether_setup(netdev);
1275 
1276 	netdev->watchdog_timeo = FJES_TX_RETRY_INTERVAL;
1277 	netdev->netdev_ops = &fjes_netdev_ops;
1278 	fjes_set_ethtool_ops(netdev);
1279 	netdev->mtu = fjes_support_mtu[3];
1280 	netdev->flags |= IFF_BROADCAST;
1281 	netdev->features |= NETIF_F_HW_CSUM | NETIF_F_HW_VLAN_CTAG_FILTER;
1282 }
1283 
1284 static void fjes_irq_watch_task(struct work_struct *work)
1285 {
1286 	struct fjes_adapter *adapter = container_of(to_delayed_work(work),
1287 			struct fjes_adapter, interrupt_watch_task);
1288 
1289 	local_irq_disable();
1290 	fjes_intr(adapter->hw.hw_res.irq, adapter);
1291 	local_irq_enable();
1292 
1293 	if (fjes_rxframe_search_exist(adapter, 0) >= 0)
1294 		napi_schedule(&adapter->napi);
1295 
1296 	if (adapter->interrupt_watch_enable) {
1297 		if (!delayed_work_pending(&adapter->interrupt_watch_task))
1298 			queue_delayed_work(adapter->control_wq,
1299 					   &adapter->interrupt_watch_task,
1300 					   FJES_IRQ_WATCH_DELAY);
1301 	}
1302 }
1303 
1304 static void fjes_watch_unshare_task(struct work_struct *work)
1305 {
1306 	struct fjes_adapter *adapter =
1307 	container_of(work, struct fjes_adapter, unshare_watch_task);
1308 
1309 	struct net_device *netdev = adapter->netdev;
1310 	struct fjes_hw *hw = &adapter->hw;
1311 
1312 	int unshare_watch, unshare_reserve;
1313 	int max_epid, my_epid, epidx;
1314 	int stop_req, stop_req_done;
1315 	ulong unshare_watch_bitmask;
1316 	unsigned long flags;
1317 	int wait_time = 0;
1318 	int is_shared;
1319 	int ret;
1320 
1321 	my_epid = hw->my_epid;
1322 	max_epid = hw->max_epid;
1323 
1324 	unshare_watch_bitmask = adapter->unshare_watch_bitmask;
1325 	adapter->unshare_watch_bitmask = 0;
1326 
1327 	while ((unshare_watch_bitmask || hw->txrx_stop_req_bit) &&
1328 	       (wait_time < 3000)) {
1329 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
1330 			if (epidx == hw->my_epid)
1331 				continue;
1332 
1333 			is_shared = fjes_hw_epid_is_shared(hw->hw_info.share,
1334 							   epidx);
1335 
1336 			stop_req = test_bit(epidx, &hw->txrx_stop_req_bit);
1337 
1338 			stop_req_done = hw->ep_shm_info[epidx].rx.info->v1i.rx_status &
1339 					FJES_RX_STOP_REQ_DONE;
1340 
1341 			unshare_watch = test_bit(epidx, &unshare_watch_bitmask);
1342 
1343 			unshare_reserve = test_bit(epidx,
1344 						   &hw->hw_info.buffer_unshare_reserve_bit);
1345 
1346 			if ((!stop_req ||
1347 			     (is_shared && (!is_shared || !stop_req_done))) &&
1348 			    (is_shared || !unshare_watch || !unshare_reserve))
1349 				continue;
1350 
1351 			mutex_lock(&hw->hw_info.lock);
1352 			ret = fjes_hw_unregister_buff_addr(hw, epidx);
1353 			switch (ret) {
1354 			case 0:
1355 				break;
1356 			case -ENOMSG:
1357 			case -EBUSY:
1358 			default:
1359 				if (!work_pending(
1360 					&adapter->force_close_task)) {
1361 					adapter->force_reset = true;
1362 					schedule_work(
1363 						&adapter->force_close_task);
1364 				}
1365 				break;
1366 			}
1367 			mutex_unlock(&hw->hw_info.lock);
1368 
1369 			spin_lock_irqsave(&hw->rx_status_lock, flags);
1370 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
1371 					    netdev->dev_addr, netdev->mtu);
1372 			spin_unlock_irqrestore(&hw->rx_status_lock, flags);
1373 
1374 			clear_bit(epidx, &hw->txrx_stop_req_bit);
1375 			clear_bit(epidx, &unshare_watch_bitmask);
1376 			clear_bit(epidx,
1377 				  &hw->hw_info.buffer_unshare_reserve_bit);
1378 		}
1379 
1380 		msleep(100);
1381 		wait_time += 100;
1382 	}
1383 
1384 	if (hw->hw_info.buffer_unshare_reserve_bit) {
1385 		for (epidx = 0; epidx < hw->max_epid; epidx++) {
1386 			if (epidx == hw->my_epid)
1387 				continue;
1388 
1389 			if (test_bit(epidx,
1390 				     &hw->hw_info.buffer_unshare_reserve_bit)) {
1391 				mutex_lock(&hw->hw_info.lock);
1392 
1393 				ret = fjes_hw_unregister_buff_addr(hw, epidx);
1394 				switch (ret) {
1395 				case 0:
1396 					break;
1397 				case -ENOMSG:
1398 				case -EBUSY:
1399 				default:
1400 					if (!work_pending(
1401 						&adapter->force_close_task)) {
1402 						adapter->force_reset = true;
1403 						schedule_work(
1404 							&adapter->force_close_task);
1405 					}
1406 					break;
1407 				}
1408 				mutex_unlock(&hw->hw_info.lock);
1409 
1410 				spin_lock_irqsave(&hw->rx_status_lock, flags);
1411 				fjes_hw_setup_epbuf(
1412 					&hw->ep_shm_info[epidx].tx,
1413 					netdev->dev_addr, netdev->mtu);
1414 				spin_unlock_irqrestore(&hw->rx_status_lock,
1415 						       flags);
1416 
1417 				clear_bit(epidx, &hw->txrx_stop_req_bit);
1418 				clear_bit(epidx, &unshare_watch_bitmask);
1419 				clear_bit(epidx, &hw->hw_info.buffer_unshare_reserve_bit);
1420 			}
1421 
1422 			if (test_bit(epidx, &unshare_watch_bitmask)) {
1423 				spin_lock_irqsave(&hw->rx_status_lock, flags);
1424 				hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
1425 						~FJES_RX_STOP_REQ_DONE;
1426 				spin_unlock_irqrestore(&hw->rx_status_lock,
1427 						       flags);
1428 			}
1429 		}
1430 	}
1431 }
1432 
1433 /* fjes_init_module - Driver Registration Routine */
1434 static int __init fjes_init_module(void)
1435 {
1436 	int result;
1437 
1438 	pr_info("%s - version %s - %s\n",
1439 		fjes_driver_string, fjes_driver_version, fjes_copyright);
1440 
1441 	result = platform_driver_register(&fjes_driver);
1442 	if (result < 0)
1443 		return result;
1444 
1445 	result = acpi_bus_register_driver(&fjes_acpi_driver);
1446 	if (result < 0)
1447 		goto fail_acpi_driver;
1448 
1449 	return 0;
1450 
1451 fail_acpi_driver:
1452 	platform_driver_unregister(&fjes_driver);
1453 	return result;
1454 }
1455 
1456 module_init(fjes_init_module);
1457 
1458 /* fjes_exit_module - Driver Exit Cleanup Routine */
1459 static void __exit fjes_exit_module(void)
1460 {
1461 	acpi_bus_unregister_driver(&fjes_acpi_driver);
1462 	platform_driver_unregister(&fjes_driver);
1463 }
1464 
1465 module_exit(fjes_exit_module);
1466