xref: /openbmc/linux/drivers/net/wan/wanxl.c (revision bfe655d1)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * wanXL serial card driver for Linux
4  * host part
5  *
6  * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
7  *
8  * Status:
9  *   - Only DTE (external clock) support with NRZ and NRZI encodings
10  *   - wanXL100 will require minor driver modifications, no access to hw
11  */
12 
13 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14 
15 #include <linux/module.h>
16 #include <linux/kernel.h>
17 #include <linux/slab.h>
18 #include <linux/sched.h>
19 #include <linux/types.h>
20 #include <linux/fcntl.h>
21 #include <linux/string.h>
22 #include <linux/errno.h>
23 #include <linux/init.h>
24 #include <linux/interrupt.h>
25 #include <linux/ioport.h>
26 #include <linux/netdevice.h>
27 #include <linux/hdlc.h>
28 #include <linux/pci.h>
29 #include <linux/dma-mapping.h>
30 #include <linux/delay.h>
31 #include <asm/io.h>
32 
33 #include "wanxl.h"
34 
35 static const char* version = "wanXL serial card driver version: 0.48";
36 
37 #define PLX_CTL_RESET   0x40000000 /* adapter reset */
38 
39 #undef DEBUG_PKT
40 #undef DEBUG_PCI
41 
42 /* MAILBOX #1 - PUTS COMMANDS */
43 #define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
44 #ifdef __LITTLE_ENDIAN
45 #define MBX1_CMD_BSWAP  0x8C000001 /* little-endian Byte Swap Mode */
46 #else
47 #define MBX1_CMD_BSWAP  0x8C000000 /* big-endian Byte Swap Mode */
48 #endif
49 
50 /* MAILBOX #2 - DRAM SIZE */
51 #define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
52 
53 
54 struct port {
55 	struct net_device *dev;
56 	struct card *card;
57 	spinlock_t lock;	/* for wanxl_xmit */
58         int node;		/* physical port #0 - 3 */
59 	unsigned int clock_type;
60 	int tx_in, tx_out;
61 	struct sk_buff *tx_skbs[TX_BUFFERS];
62 };
63 
64 
65 struct card_status {
66 	desc_t rx_descs[RX_QUEUE_LENGTH];
67 	port_status_t port_status[4];
68 };
69 
70 
71 struct card {
72 	int n_ports;		/* 1, 2 or 4 ports */
73 	u8 irq;
74 
75 	u8 __iomem *plx;	/* PLX PCI9060 virtual base address */
76 	struct pci_dev *pdev;	/* for pci_name(pdev) */
77 	int rx_in;
78 	struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
79 	struct card_status *status;	/* shared between host and card */
80 	dma_addr_t status_address;
81 	struct port ports[0];	/* 1 - 4 port structures follow */
82 };
83 
84 
85 
86 static inline struct port *dev_to_port(struct net_device *dev)
87 {
88 	return (struct port *)dev_to_hdlc(dev)->priv;
89 }
90 
91 
92 static inline port_status_t *get_status(struct port *port)
93 {
94 	return &port->card->status->port_status[port->node];
95 }
96 
97 
98 #ifdef DEBUG_PCI
99 static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
100 					      size_t size, int direction)
101 {
102 	dma_addr_t addr = pci_map_single(pdev, ptr, size, direction);
103 	if (addr + size > 0x100000000LL)
104 		pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
105 			pci_name(pdev), (unsigned long long)addr);
106 	return addr;
107 }
108 
109 #undef pci_map_single
110 #define pci_map_single pci_map_single_debug
111 #endif
112 
113 
114 /* Cable and/or personality module change interrupt service */
115 static inline void wanxl_cable_intr(struct port *port)
116 {
117 	u32 value = get_status(port)->cable;
118 	int valid = 1;
119 	const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
120 
121 	switch(value & 0x7) {
122 	case STATUS_CABLE_V35: cable = "V.35"; break;
123 	case STATUS_CABLE_X21: cable = "X.21"; break;
124 	case STATUS_CABLE_V24: cable = "V.24"; break;
125 	case STATUS_CABLE_EIA530: cable = "EIA530"; break;
126 	case STATUS_CABLE_NONE: cable = "no"; break;
127 	default: cable = "invalid";
128 	}
129 
130 	switch((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
131 	case STATUS_CABLE_V35: pm = "V.35"; break;
132 	case STATUS_CABLE_X21: pm = "X.21"; break;
133 	case STATUS_CABLE_V24: pm = "V.24"; break;
134 	case STATUS_CABLE_EIA530: pm = "EIA530"; break;
135 	case STATUS_CABLE_NONE: pm = "no personality"; valid = 0; break;
136 	default: pm = "invalid personality"; valid = 0;
137 	}
138 
139 	if (valid) {
140 		if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
141 			dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
142 				", DSR off";
143 			dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
144 				", carrier off";
145 		}
146 		dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
147 	}
148 	netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
149 		    pm, dte, cable, dsr, dcd);
150 
151 	if (value & STATUS_CABLE_DCD)
152 		netif_carrier_on(port->dev);
153 	else
154 		netif_carrier_off(port->dev);
155 }
156 
157 
158 
159 /* Transmit complete interrupt service */
160 static inline void wanxl_tx_intr(struct port *port)
161 {
162 	struct net_device *dev = port->dev;
163 	while (1) {
164                 desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
165 		struct sk_buff *skb = port->tx_skbs[port->tx_in];
166 
167 		switch (desc->stat) {
168 		case PACKET_FULL:
169 		case PACKET_EMPTY:
170 			netif_wake_queue(dev);
171 			return;
172 
173 		case PACKET_UNDERRUN:
174 			dev->stats.tx_errors++;
175 			dev->stats.tx_fifo_errors++;
176 			break;
177 
178 		default:
179 			dev->stats.tx_packets++;
180 			dev->stats.tx_bytes += skb->len;
181 		}
182                 desc->stat = PACKET_EMPTY; /* Free descriptor */
183 		pci_unmap_single(port->card->pdev, desc->address, skb->len,
184 				 PCI_DMA_TODEVICE);
185 		dev_consume_skb_irq(skb);
186                 port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
187         }
188 }
189 
190 
191 
192 /* Receive complete interrupt service */
193 static inline void wanxl_rx_intr(struct card *card)
194 {
195 	desc_t *desc;
196 	while (desc = &card->status->rx_descs[card->rx_in],
197 	       desc->stat != PACKET_EMPTY) {
198 		if ((desc->stat & PACKET_PORT_MASK) > card->n_ports)
199 			pr_crit("%s: received packet for nonexistent port\n",
200 				pci_name(card->pdev));
201 		else {
202 			struct sk_buff *skb = card->rx_skbs[card->rx_in];
203 			struct port *port = &card->ports[desc->stat &
204 						    PACKET_PORT_MASK];
205 			struct net_device *dev = port->dev;
206 
207 			if (!skb)
208 				dev->stats.rx_dropped++;
209 			else {
210 				pci_unmap_single(card->pdev, desc->address,
211 						 BUFFER_LENGTH,
212 						 PCI_DMA_FROMDEVICE);
213 				skb_put(skb, desc->length);
214 
215 #ifdef DEBUG_PKT
216 				printk(KERN_DEBUG "%s RX(%i):", dev->name,
217 				       skb->len);
218 				debug_frame(skb);
219 #endif
220 				dev->stats.rx_packets++;
221 				dev->stats.rx_bytes += skb->len;
222 				skb->protocol = hdlc_type_trans(skb, dev);
223 				netif_rx(skb);
224 				skb = NULL;
225 			}
226 
227 			if (!skb) {
228 				skb = dev_alloc_skb(BUFFER_LENGTH);
229 				desc->address = skb ?
230 					pci_map_single(card->pdev, skb->data,
231 						       BUFFER_LENGTH,
232 						       PCI_DMA_FROMDEVICE) : 0;
233 				card->rx_skbs[card->rx_in] = skb;
234 			}
235 		}
236 		desc->stat = PACKET_EMPTY; /* Free descriptor */
237 		card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
238 	}
239 }
240 
241 
242 
243 static irqreturn_t wanxl_intr(int irq, void* dev_id)
244 {
245 	struct card *card = dev_id;
246         int i;
247         u32 stat;
248         int handled = 0;
249 
250 
251         while((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
252                 handled = 1;
253 		writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
254 
255                 for (i = 0; i < card->n_ports; i++) {
256 			if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
257 				wanxl_tx_intr(&card->ports[i]);
258 			if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
259 				wanxl_cable_intr(&card->ports[i]);
260 		}
261 		if (stat & (1 << DOORBELL_FROM_CARD_RX))
262 			wanxl_rx_intr(card);
263         }
264 
265         return IRQ_RETVAL(handled);
266 }
267 
268 
269 
270 static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
271 {
272 	struct port *port = dev_to_port(dev);
273 	desc_t *desc;
274 
275         spin_lock(&port->lock);
276 
277 	desc = &get_status(port)->tx_descs[port->tx_out];
278         if (desc->stat != PACKET_EMPTY) {
279                 /* should never happen - previous xmit should stop queue */
280 #ifdef DEBUG_PKT
281                 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
282 #endif
283 		netif_stop_queue(dev);
284 		spin_unlock(&port->lock);
285 		return NETDEV_TX_BUSY;       /* request packet to be queued */
286 	}
287 
288 #ifdef DEBUG_PKT
289 	printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
290 	debug_frame(skb);
291 #endif
292 
293 	port->tx_skbs[port->tx_out] = skb;
294 	desc->address = pci_map_single(port->card->pdev, skb->data, skb->len,
295 				       PCI_DMA_TODEVICE);
296 	desc->length = skb->len;
297 	desc->stat = PACKET_FULL;
298 	writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
299 	       port->card->plx + PLX_DOORBELL_TO_CARD);
300 
301 	port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
302 
303 	if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
304 		netif_stop_queue(dev);
305 #ifdef DEBUG_PKT
306 		printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
307 #endif
308 	}
309 
310 	spin_unlock(&port->lock);
311 	return NETDEV_TX_OK;
312 }
313 
314 
315 
316 static int wanxl_attach(struct net_device *dev, unsigned short encoding,
317 			unsigned short parity)
318 {
319 	struct port *port = dev_to_port(dev);
320 
321 	if (encoding != ENCODING_NRZ &&
322 	    encoding != ENCODING_NRZI)
323 		return -EINVAL;
324 
325 	if (parity != PARITY_NONE &&
326 	    parity != PARITY_CRC32_PR1_CCITT &&
327 	    parity != PARITY_CRC16_PR1_CCITT &&
328 	    parity != PARITY_CRC32_PR0_CCITT &&
329 	    parity != PARITY_CRC16_PR0_CCITT)
330 		return -EINVAL;
331 
332 	get_status(port)->encoding = encoding;
333 	get_status(port)->parity = parity;
334 	return 0;
335 }
336 
337 
338 
339 static int wanxl_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
340 {
341 	const size_t size = sizeof(sync_serial_settings);
342 	sync_serial_settings line;
343 	struct port *port = dev_to_port(dev);
344 
345 	if (cmd != SIOCWANDEV)
346 		return hdlc_ioctl(dev, ifr, cmd);
347 
348 	switch (ifr->ifr_settings.type) {
349 	case IF_GET_IFACE:
350 		ifr->ifr_settings.type = IF_IFACE_SYNC_SERIAL;
351 		if (ifr->ifr_settings.size < size) {
352 			ifr->ifr_settings.size = size; /* data size wanted */
353 			return -ENOBUFS;
354 		}
355 		memset(&line, 0, sizeof(line));
356 		line.clock_type = get_status(port)->clocking;
357 		line.clock_rate = 0;
358 		line.loopback = 0;
359 
360 		if (copy_to_user(ifr->ifr_settings.ifs_ifsu.sync, &line, size))
361 			return -EFAULT;
362 		return 0;
363 
364 	case IF_IFACE_SYNC_SERIAL:
365 		if (!capable(CAP_NET_ADMIN))
366 			return -EPERM;
367 		if (dev->flags & IFF_UP)
368 			return -EBUSY;
369 
370 		if (copy_from_user(&line, ifr->ifr_settings.ifs_ifsu.sync,
371 				   size))
372 			return -EFAULT;
373 
374 		if (line.clock_type != CLOCK_EXT &&
375 		    line.clock_type != CLOCK_TXFROMRX)
376 			return -EINVAL; /* No such clock setting */
377 
378 		if (line.loopback != 0)
379 			return -EINVAL;
380 
381 		get_status(port)->clocking = line.clock_type;
382 		return 0;
383 
384 	default:
385 		return hdlc_ioctl(dev, ifr, cmd);
386         }
387 }
388 
389 
390 
391 static int wanxl_open(struct net_device *dev)
392 {
393 	struct port *port = dev_to_port(dev);
394 	u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
395 	unsigned long timeout;
396 	int i;
397 
398 	if (get_status(port)->open) {
399 		netdev_err(dev, "port already open\n");
400 		return -EIO;
401 	}
402 	if ((i = hdlc_open(dev)) != 0)
403 		return i;
404 
405 	port->tx_in = port->tx_out = 0;
406 	for (i = 0; i < TX_BUFFERS; i++)
407 		get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
408 	/* signal the card */
409 	writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
410 
411 	timeout = jiffies + HZ;
412 	do {
413 		if (get_status(port)->open) {
414 			netif_start_queue(dev);
415 			return 0;
416 		}
417 	} while (time_after(timeout, jiffies));
418 
419 	netdev_err(dev, "unable to open port\n");
420 	/* ask the card to close the port, should it be still alive */
421 	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
422 	return -EFAULT;
423 }
424 
425 
426 
427 static int wanxl_close(struct net_device *dev)
428 {
429 	struct port *port = dev_to_port(dev);
430 	unsigned long timeout;
431 	int i;
432 
433 	hdlc_close(dev);
434 	/* signal the card */
435 	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
436 	       port->card->plx + PLX_DOORBELL_TO_CARD);
437 
438 	timeout = jiffies + HZ;
439 	do {
440 		if (!get_status(port)->open)
441 			break;
442 	} while (time_after(timeout, jiffies));
443 
444 	if (get_status(port)->open)
445 		netdev_err(dev, "unable to close port\n");
446 
447 	netif_stop_queue(dev);
448 
449 	for (i = 0; i < TX_BUFFERS; i++) {
450 		desc_t *desc = &get_status(port)->tx_descs[i];
451 
452 		if (desc->stat != PACKET_EMPTY) {
453 			desc->stat = PACKET_EMPTY;
454 			pci_unmap_single(port->card->pdev, desc->address,
455 					 port->tx_skbs[i]->len,
456 					 PCI_DMA_TODEVICE);
457 			dev_kfree_skb(port->tx_skbs[i]);
458 		}
459 	}
460 	return 0;
461 }
462 
463 
464 
465 static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
466 {
467 	struct port *port = dev_to_port(dev);
468 
469 	dev->stats.rx_over_errors = get_status(port)->rx_overruns;
470 	dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
471 	dev->stats.rx_errors = dev->stats.rx_over_errors +
472 		dev->stats.rx_frame_errors;
473 	return &dev->stats;
474 }
475 
476 
477 
478 static int wanxl_puts_command(struct card *card, u32 cmd)
479 {
480 	unsigned long timeout = jiffies + 5 * HZ;
481 
482 	writel(cmd, card->plx + PLX_MAILBOX_1);
483 	do {
484 		if (readl(card->plx + PLX_MAILBOX_1) == 0)
485 			return 0;
486 
487 		schedule();
488 	}while (time_after(timeout, jiffies));
489 
490 	return -1;
491 }
492 
493 
494 
495 static void wanxl_reset(struct card *card)
496 {
497 	u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
498 
499 	writel(0x80, card->plx + PLX_MAILBOX_0);
500 	writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
501 	readl(card->plx + PLX_CONTROL); /* wait for posted write */
502 	udelay(1);
503 	writel(old_value, card->plx + PLX_CONTROL);
504 	readl(card->plx + PLX_CONTROL); /* wait for posted write */
505 }
506 
507 
508 
509 static void wanxl_pci_remove_one(struct pci_dev *pdev)
510 {
511 	struct card *card = pci_get_drvdata(pdev);
512 	int i;
513 
514 	for (i = 0; i < card->n_ports; i++) {
515 		unregister_hdlc_device(card->ports[i].dev);
516 		free_netdev(card->ports[i].dev);
517 	}
518 
519 	/* unregister and free all host resources */
520 	if (card->irq)
521 		free_irq(card->irq, card);
522 
523 	wanxl_reset(card);
524 
525 	for (i = 0; i < RX_QUEUE_LENGTH; i++)
526 		if (card->rx_skbs[i]) {
527 			pci_unmap_single(card->pdev,
528 					 card->status->rx_descs[i].address,
529 					 BUFFER_LENGTH, PCI_DMA_FROMDEVICE);
530 			dev_kfree_skb(card->rx_skbs[i]);
531 		}
532 
533 	if (card->plx)
534 		iounmap(card->plx);
535 
536 	if (card->status)
537 		pci_free_consistent(pdev, sizeof(struct card_status),
538 				    card->status, card->status_address);
539 
540 	pci_release_regions(pdev);
541 	pci_disable_device(pdev);
542 	kfree(card);
543 }
544 
545 
546 #include "wanxlfw.inc"
547 
548 static const struct net_device_ops wanxl_ops = {
549 	.ndo_open       = wanxl_open,
550 	.ndo_stop       = wanxl_close,
551 	.ndo_start_xmit = hdlc_start_xmit,
552 	.ndo_do_ioctl   = wanxl_ioctl,
553 	.ndo_get_stats  = wanxl_get_stats,
554 };
555 
556 static int wanxl_pci_init_one(struct pci_dev *pdev,
557 			      const struct pci_device_id *ent)
558 {
559 	struct card *card;
560 	u32 ramsize, stat;
561 	unsigned long timeout;
562 	u32 plx_phy;		/* PLX PCI base address */
563 	u32 mem_phy;		/* memory PCI base addr */
564 	u8 __iomem *mem;	/* memory virtual base addr */
565 	int i, ports;
566 
567 #ifndef MODULE
568 	pr_info_once("%s\n", version);
569 #endif
570 
571 	i = pci_enable_device(pdev);
572 	if (i)
573 		return i;
574 
575 	/* QUICC can only access first 256 MB of host RAM directly,
576 	   but PLX9060 DMA does 32-bits for actual packet data transfers */
577 
578 	/* FIXME when PCI/DMA subsystems are fixed.
579 	   We set both dma_mask and consistent_dma_mask to 28 bits
580 	   and pray pci_alloc_consistent() will use this info. It should
581 	   work on most platforms */
582 	if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(28)) ||
583 	    pci_set_dma_mask(pdev, DMA_BIT_MASK(28))) {
584 		pr_err("No usable DMA configuration\n");
585 		pci_disable_device(pdev);
586 		return -EIO;
587 	}
588 
589 	i = pci_request_regions(pdev, "wanXL");
590 	if (i) {
591 		pci_disable_device(pdev);
592 		return i;
593 	}
594 
595 	switch (pdev->device) {
596 	case PCI_DEVICE_ID_SBE_WANXL100: ports = 1; break;
597 	case PCI_DEVICE_ID_SBE_WANXL200: ports = 2; break;
598 	default: ports = 4;
599 	}
600 
601 	card = kzalloc(struct_size(card, ports, ports), GFP_KERNEL);
602 	if (card == NULL) {
603 		pci_release_regions(pdev);
604 		pci_disable_device(pdev);
605 		return -ENOBUFS;
606 	}
607 
608 	pci_set_drvdata(pdev, card);
609 	card->pdev = pdev;
610 
611 	card->status = pci_alloc_consistent(pdev,
612 					    sizeof(struct card_status),
613 					    &card->status_address);
614 	if (card->status == NULL) {
615 		wanxl_pci_remove_one(pdev);
616 		return -ENOBUFS;
617 	}
618 
619 #ifdef DEBUG_PCI
620 	printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
621 	       " at 0x%LX\n", pci_name(pdev),
622 	       (unsigned long long)card->status_address);
623 #endif
624 
625 	/* FIXME when PCI/DMA subsystems are fixed.
626 	   We set both dma_mask and consistent_dma_mask back to 32 bits
627 	   to indicate the card can do 32-bit DMA addressing */
628 	if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) ||
629 	    pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
630 		pr_err("No usable DMA configuration\n");
631 		wanxl_pci_remove_one(pdev);
632 		return -EIO;
633 	}
634 
635 	/* set up PLX mapping */
636 	plx_phy = pci_resource_start(pdev, 0);
637 
638 	card->plx = ioremap_nocache(plx_phy, 0x70);
639 	if (!card->plx) {
640 		pr_err("ioremap() failed\n");
641  		wanxl_pci_remove_one(pdev);
642 		return -EFAULT;
643 	}
644 
645 #if RESET_WHILE_LOADING
646 	wanxl_reset(card);
647 #endif
648 
649 	timeout = jiffies + 20 * HZ;
650 	while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
651 		if (time_before(timeout, jiffies)) {
652 			pr_warn("%s: timeout waiting for PUTS to complete\n",
653 				pci_name(pdev));
654 			wanxl_pci_remove_one(pdev);
655 			return -ENODEV;
656 		}
657 
658 		switch(stat & 0xC0) {
659 		case 0x00:	/* hmm - PUTS completed with non-zero code? */
660 		case 0x80:	/* PUTS still testing the hardware */
661 			break;
662 
663 		default:
664 			pr_warn("%s: PUTS test 0x%X failed\n",
665 				pci_name(pdev), stat & 0x30);
666 			wanxl_pci_remove_one(pdev);
667 			return -ENODEV;
668 		}
669 
670 		schedule();
671 	}
672 
673 	/* get on-board memory size (PUTS detects no more than 4 MB) */
674 	ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
675 
676 	/* set up on-board RAM mapping */
677 	mem_phy = pci_resource_start(pdev, 2);
678 
679 
680 	/* sanity check the board's reported memory size */
681 	if (ramsize < BUFFERS_ADDR +
682 	    (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
683 		pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
684 			pci_name(pdev), ramsize,
685 			BUFFERS_ADDR +
686 			(TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
687 		wanxl_pci_remove_one(pdev);
688 		return -ENODEV;
689 	}
690 
691 	if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
692 		pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
693 		wanxl_pci_remove_one(pdev);
694 		return -ENODEV;
695 	}
696 
697 	for (i = 0; i < RX_QUEUE_LENGTH; i++) {
698 		struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
699 		card->rx_skbs[i] = skb;
700 		if (skb)
701 			card->status->rx_descs[i].address =
702 				pci_map_single(card->pdev, skb->data,
703 					       BUFFER_LENGTH,
704 					       PCI_DMA_FROMDEVICE);
705 	}
706 
707 	mem = ioremap_nocache(mem_phy, PDM_OFFSET + sizeof(firmware));
708 	if (!mem) {
709 		pr_err("ioremap() failed\n");
710  		wanxl_pci_remove_one(pdev);
711 		return -EFAULT;
712 	}
713 
714 	for (i = 0; i < sizeof(firmware); i += 4)
715 		writel(ntohl(*(__be32*)(firmware + i)), mem + PDM_OFFSET + i);
716 
717 	for (i = 0; i < ports; i++)
718 		writel(card->status_address +
719 		       (void *)&card->status->port_status[i] -
720 		       (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
721 	writel(card->status_address, mem + PDM_OFFSET + 20);
722 	writel(PDM_OFFSET, mem);
723 	iounmap(mem);
724 
725 	writel(0, card->plx + PLX_MAILBOX_5);
726 
727 	if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
728 		pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
729 		wanxl_pci_remove_one(pdev);
730 		return -ENODEV;
731 	}
732 
733 	timeout = jiffies + 5 * HZ;
734 	do {
735 		if ((stat = readl(card->plx + PLX_MAILBOX_5)) != 0)
736 			break;
737 		schedule();
738 	}while (time_after(timeout, jiffies));
739 
740 	if (!stat) {
741 		pr_warn("%s: timeout while initializing card firmware\n",
742 			pci_name(pdev));
743 		wanxl_pci_remove_one(pdev);
744 		return -ENODEV;
745 	}
746 
747 #if DETECT_RAM
748 	ramsize = stat;
749 #endif
750 
751 	pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
752 		pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
753 
754 	/* Allocate IRQ */
755 	if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
756 		pr_warn("%s: could not allocate IRQ%i\n",
757 			pci_name(pdev), pdev->irq);
758 		wanxl_pci_remove_one(pdev);
759 		return -EBUSY;
760 	}
761 	card->irq = pdev->irq;
762 
763 	for (i = 0; i < ports; i++) {
764 		hdlc_device *hdlc;
765 		struct port *port = &card->ports[i];
766 		struct net_device *dev = alloc_hdlcdev(port);
767 		if (!dev) {
768 			pr_err("%s: unable to allocate memory\n",
769 			       pci_name(pdev));
770 			wanxl_pci_remove_one(pdev);
771 			return -ENOMEM;
772 		}
773 
774 		port->dev = dev;
775 		hdlc = dev_to_hdlc(dev);
776 		spin_lock_init(&port->lock);
777 		dev->tx_queue_len = 50;
778 		dev->netdev_ops = &wanxl_ops;
779 		hdlc->attach = wanxl_attach;
780 		hdlc->xmit = wanxl_xmit;
781 		port->card = card;
782 		port->node = i;
783 		get_status(port)->clocking = CLOCK_EXT;
784 		if (register_hdlc_device(dev)) {
785 			pr_err("%s: unable to register hdlc device\n",
786 			       pci_name(pdev));
787 			free_netdev(dev);
788 			wanxl_pci_remove_one(pdev);
789 			return -ENOBUFS;
790 		}
791 		card->n_ports++;
792 	}
793 
794 	pr_info("%s: port", pci_name(pdev));
795 	for (i = 0; i < ports; i++)
796 		pr_cont("%s #%i: %s",
797 			i ? "," : "", i, card->ports[i].dev->name);
798 	pr_cont("\n");
799 
800 	for (i = 0; i < ports; i++)
801 		wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
802 
803 	return 0;
804 }
805 
806 static const struct pci_device_id wanxl_pci_tbl[] = {
807 	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
808 	  PCI_ANY_ID, 0, 0, 0 },
809 	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
810 	  PCI_ANY_ID, 0, 0, 0 },
811 	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
812 	  PCI_ANY_ID, 0, 0, 0 },
813 	{ 0, }
814 };
815 
816 
817 static struct pci_driver wanxl_pci_driver = {
818 	.name		= "wanXL",
819 	.id_table	= wanxl_pci_tbl,
820 	.probe		= wanxl_pci_init_one,
821 	.remove		= wanxl_pci_remove_one,
822 };
823 
824 
825 static int __init wanxl_init_module(void)
826 {
827 #ifdef MODULE
828 	pr_info("%s\n", version);
829 #endif
830 	return pci_register_driver(&wanxl_pci_driver);
831 }
832 
833 static void __exit wanxl_cleanup_module(void)
834 {
835 	pci_unregister_driver(&wanxl_pci_driver);
836 }
837 
838 
839 MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
840 MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
841 MODULE_LICENSE("GPL v2");
842 MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
843 
844 module_init(wanxl_init_module);
845 module_exit(wanxl_cleanup_module);
846