1 /*****************************************************************************
2  *                                                                           *
3  * File: cxgb2.c                                                             *
4  * $Revision: 1.25 $                                                         *
5  * $Date: 2005/06/22 00:43:25 $                                              *
6  * Description:                                                              *
7  *  Chelsio 10Gb Ethernet Driver.                                            *
8  *                                                                           *
9  * This program is free software; you can redistribute it and/or modify      *
10  * it under the terms of the GNU General Public License, version 2, as       *
11  * published by the Free Software Foundation.                                *
12  *                                                                           *
13  * You should have received a copy of the GNU General Public License along   *
14  * with this program; if not, see <http://www.gnu.org/licenses/>.            *
15  *                                                                           *
16  * THIS SOFTWARE IS PROVIDED ``AS IS'' AND WITHOUT ANY EXPRESS OR IMPLIED    *
17  * WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE IMPLIED WARRANTIES OF      *
18  * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE.                     *
19  *                                                                           *
20  * http://www.chelsio.com                                                    *
21  *                                                                           *
22  * Copyright (c) 2003 - 2005 Chelsio Communications, Inc.                    *
23  * All rights reserved.                                                      *
24  *                                                                           *
25  * Maintainers: maintainers@chelsio.com                                      *
26  *                                                                           *
27  * Authors: Dimitrios Michailidis   <dm@chelsio.com>                         *
28  *          Tina Yang               <tainay@chelsio.com>                     *
29  *          Felix Marti             <felix@chelsio.com>                      *
30  *          Scott Bardone           <sbardone@chelsio.com>                   *
31  *          Kurt Ottaway            <kottaway@chelsio.com>                   *
32  *          Frank DiMambro          <frank@chelsio.com>                      *
33  *                                                                           *
34  * History:                                                                  *
35  *                                                                           *
36  ****************************************************************************/
37 
38 #include "common.h"
39 #include <linux/module.h>
40 #include <linux/pci.h>
41 #include <linux/netdevice.h>
42 #include <linux/etherdevice.h>
43 #include <linux/if_vlan.h>
44 #include <linux/mii.h>
45 #include <linux/sockios.h>
46 #include <linux/dma-mapping.h>
47 #include <linux/uaccess.h>
48 
49 #include "cpl5_cmd.h"
50 #include "regs.h"
51 #include "gmac.h"
52 #include "cphy.h"
53 #include "sge.h"
54 #include "tp.h"
55 #include "espi.h"
56 #include "elmer0.h"
57 
58 #include <linux/workqueue.h>
59 
60 static inline void schedule_mac_stats_update(struct adapter *ap, int secs)
61 {
62 	schedule_delayed_work(&ap->stats_update_task, secs * HZ);
63 }
64 
65 static inline void cancel_mac_stats_update(struct adapter *ap)
66 {
67 	cancel_delayed_work(&ap->stats_update_task);
68 }
69 
70 #define MAX_CMDQ_ENTRIES	16384
71 #define MAX_CMDQ1_ENTRIES	1024
72 #define MAX_RX_BUFFERS		16384
73 #define MAX_RX_JUMBO_BUFFERS	16384
74 #define MAX_TX_BUFFERS_HIGH	16384U
75 #define MAX_TX_BUFFERS_LOW	1536U
76 #define MAX_TX_BUFFERS		1460U
77 #define MIN_FL_ENTRIES		32
78 
79 #define DFLT_MSG_ENABLE (NETIF_MSG_DRV | NETIF_MSG_PROBE | NETIF_MSG_LINK | \
80 			 NETIF_MSG_TIMER | NETIF_MSG_IFDOWN | NETIF_MSG_IFUP |\
81 			 NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR)
82 
83 /*
84  * The EEPROM is actually bigger but only the first few bytes are used so we
85  * only report those.
86  */
87 #define EEPROM_SIZE 32
88 
89 MODULE_DESCRIPTION(DRV_DESCRIPTION);
90 MODULE_AUTHOR("Chelsio Communications");
91 MODULE_LICENSE("GPL");
92 
93 static int dflt_msg_enable = DFLT_MSG_ENABLE;
94 
95 module_param(dflt_msg_enable, int, 0);
96 MODULE_PARM_DESC(dflt_msg_enable, "Chelsio T1 default message enable bitmap");
97 
98 #define HCLOCK 0x0
99 #define LCLOCK 0x1
100 
101 /* T1 cards powersave mode */
102 static int t1_clock(struct adapter *adapter, int mode);
103 static int t1powersave = 1;	/* HW default is powersave mode. */
104 
105 module_param(t1powersave, int, 0);
106 MODULE_PARM_DESC(t1powersave, "Enable/Disable T1 powersaving mode");
107 
108 static int disable_msi = 0;
109 module_param(disable_msi, int, 0);
110 MODULE_PARM_DESC(disable_msi, "Disable Message Signaled Interrupt (MSI)");
111 
112 /*
113  * Setup MAC to receive the types of packets we want.
114  */
115 static void t1_set_rxmode(struct net_device *dev)
116 {
117 	struct adapter *adapter = dev->ml_priv;
118 	struct cmac *mac = adapter->port[dev->if_port].mac;
119 	struct t1_rx_mode rm;
120 
121 	rm.dev = dev;
122 	mac->ops->set_rx_mode(mac, &rm);
123 }
124 
125 static void link_report(struct port_info *p)
126 {
127 	if (!netif_carrier_ok(p->dev))
128 		netdev_info(p->dev, "link down\n");
129 	else {
130 		const char *s = "10Mbps";
131 
132 		switch (p->link_config.speed) {
133 			case SPEED_10000: s = "10Gbps"; break;
134 			case SPEED_1000:  s = "1000Mbps"; break;
135 			case SPEED_100:   s = "100Mbps"; break;
136 		}
137 
138 		netdev_info(p->dev, "link up, %s, %s-duplex\n",
139 			    s, p->link_config.duplex == DUPLEX_FULL
140 			    ? "full" : "half");
141 	}
142 }
143 
144 void t1_link_negotiated(struct adapter *adapter, int port_id, int link_stat,
145 			int speed, int duplex, int pause)
146 {
147 	struct port_info *p = &adapter->port[port_id];
148 
149 	if (link_stat != netif_carrier_ok(p->dev)) {
150 		if (link_stat)
151 			netif_carrier_on(p->dev);
152 		else
153 			netif_carrier_off(p->dev);
154 		link_report(p);
155 
156 		/* multi-ports: inform toe */
157 		if ((speed > 0) && (adapter->params.nports > 1)) {
158 			unsigned int sched_speed = 10;
159 			switch (speed) {
160 			case SPEED_1000:
161 				sched_speed = 1000;
162 				break;
163 			case SPEED_100:
164 				sched_speed = 100;
165 				break;
166 			case SPEED_10:
167 				sched_speed = 10;
168 				break;
169 			}
170 			t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed);
171 		}
172 	}
173 }
174 
175 static void link_start(struct port_info *p)
176 {
177 	struct cmac *mac = p->mac;
178 
179 	mac->ops->reset(mac);
180 	if (mac->ops->macaddress_set)
181 		mac->ops->macaddress_set(mac, p->dev->dev_addr);
182 	t1_set_rxmode(p->dev);
183 	t1_link_start(p->phy, mac, &p->link_config);
184 	mac->ops->enable(mac, MAC_DIRECTION_RX | MAC_DIRECTION_TX);
185 }
186 
187 static void enable_hw_csum(struct adapter *adapter)
188 {
189 	if (adapter->port[0].dev->hw_features & NETIF_F_TSO)
190 		t1_tp_set_ip_checksum_offload(adapter->tp, 1);	/* for TSO only */
191 	t1_tp_set_tcp_checksum_offload(adapter->tp, 1);
192 }
193 
194 /*
195  * Things to do upon first use of a card.
196  * This must run with the rtnl lock held.
197  */
198 static int cxgb_up(struct adapter *adapter)
199 {
200 	int err = 0;
201 
202 	if (!(adapter->flags & FULL_INIT_DONE)) {
203 		err = t1_init_hw_modules(adapter);
204 		if (err)
205 			goto out_err;
206 
207 		enable_hw_csum(adapter);
208 		adapter->flags |= FULL_INIT_DONE;
209 	}
210 
211 	t1_interrupts_clear(adapter);
212 
213 	adapter->params.has_msi = !disable_msi && !pci_enable_msi(adapter->pdev);
214 	err = request_threaded_irq(adapter->pdev->irq, t1_interrupt,
215 				   t1_interrupt_thread,
216 				   adapter->params.has_msi ? 0 : IRQF_SHARED,
217 				   adapter->name, adapter);
218 	if (err) {
219 		if (adapter->params.has_msi)
220 			pci_disable_msi(adapter->pdev);
221 
222 		goto out_err;
223 	}
224 
225 	t1_sge_start(adapter->sge);
226 	t1_interrupts_enable(adapter);
227 out_err:
228 	return err;
229 }
230 
231 /*
232  * Release resources when all the ports have been stopped.
233  */
234 static void cxgb_down(struct adapter *adapter)
235 {
236 	t1_sge_stop(adapter->sge);
237 	t1_interrupts_disable(adapter);
238 	free_irq(adapter->pdev->irq, adapter);
239 	if (adapter->params.has_msi)
240 		pci_disable_msi(adapter->pdev);
241 }
242 
243 static int cxgb_open(struct net_device *dev)
244 {
245 	int err;
246 	struct adapter *adapter = dev->ml_priv;
247 	int other_ports = adapter->open_device_map & PORT_MASK;
248 
249 	napi_enable(&adapter->napi);
250 	if (!adapter->open_device_map && (err = cxgb_up(adapter)) < 0) {
251 		napi_disable(&adapter->napi);
252 		return err;
253 	}
254 
255 	__set_bit(dev->if_port, &adapter->open_device_map);
256 	link_start(&adapter->port[dev->if_port]);
257 	netif_start_queue(dev);
258 	if (!other_ports && adapter->params.stats_update_period)
259 		schedule_mac_stats_update(adapter,
260 					  adapter->params.stats_update_period);
261 
262 	t1_vlan_mode(adapter, dev->features);
263 	return 0;
264 }
265 
266 static int cxgb_close(struct net_device *dev)
267 {
268 	struct adapter *adapter = dev->ml_priv;
269 	struct port_info *p = &adapter->port[dev->if_port];
270 	struct cmac *mac = p->mac;
271 
272 	netif_stop_queue(dev);
273 	napi_disable(&adapter->napi);
274 	mac->ops->disable(mac, MAC_DIRECTION_TX | MAC_DIRECTION_RX);
275 	netif_carrier_off(dev);
276 
277 	clear_bit(dev->if_port, &adapter->open_device_map);
278 	if (adapter->params.stats_update_period &&
279 	    !(adapter->open_device_map & PORT_MASK)) {
280 		/* Stop statistics accumulation. */
281 		smp_mb__after_atomic();
282 		spin_lock(&adapter->work_lock);   /* sync with update task */
283 		spin_unlock(&adapter->work_lock);
284 		cancel_mac_stats_update(adapter);
285 	}
286 
287 	if (!adapter->open_device_map)
288 		cxgb_down(adapter);
289 	return 0;
290 }
291 
292 static struct net_device_stats *t1_get_stats(struct net_device *dev)
293 {
294 	struct adapter *adapter = dev->ml_priv;
295 	struct port_info *p = &adapter->port[dev->if_port];
296 	struct net_device_stats *ns = &dev->stats;
297 	const struct cmac_statistics *pstats;
298 
299 	/* Do a full update of the MAC stats */
300 	pstats = p->mac->ops->statistics_update(p->mac,
301 						MAC_STATS_UPDATE_FULL);
302 
303 	ns->tx_packets = pstats->TxUnicastFramesOK +
304 		pstats->TxMulticastFramesOK + pstats->TxBroadcastFramesOK;
305 
306 	ns->rx_packets = pstats->RxUnicastFramesOK +
307 		pstats->RxMulticastFramesOK + pstats->RxBroadcastFramesOK;
308 
309 	ns->tx_bytes = pstats->TxOctetsOK;
310 	ns->rx_bytes = pstats->RxOctetsOK;
311 
312 	ns->tx_errors = pstats->TxLateCollisions + pstats->TxLengthErrors +
313 		pstats->TxUnderrun + pstats->TxFramesAbortedDueToXSCollisions;
314 	ns->rx_errors = pstats->RxDataErrors + pstats->RxJabberErrors +
315 		pstats->RxFCSErrors + pstats->RxAlignErrors +
316 		pstats->RxSequenceErrors + pstats->RxFrameTooLongErrors +
317 		pstats->RxSymbolErrors + pstats->RxRuntErrors;
318 
319 	ns->multicast  = pstats->RxMulticastFramesOK;
320 	ns->collisions = pstats->TxTotalCollisions;
321 
322 	/* detailed rx_errors */
323 	ns->rx_length_errors = pstats->RxFrameTooLongErrors +
324 		pstats->RxJabberErrors;
325 	ns->rx_over_errors   = 0;
326 	ns->rx_crc_errors    = pstats->RxFCSErrors;
327 	ns->rx_frame_errors  = pstats->RxAlignErrors;
328 	ns->rx_fifo_errors   = 0;
329 	ns->rx_missed_errors = 0;
330 
331 	/* detailed tx_errors */
332 	ns->tx_aborted_errors   = pstats->TxFramesAbortedDueToXSCollisions;
333 	ns->tx_carrier_errors   = 0;
334 	ns->tx_fifo_errors      = pstats->TxUnderrun;
335 	ns->tx_heartbeat_errors = 0;
336 	ns->tx_window_errors    = pstats->TxLateCollisions;
337 	return ns;
338 }
339 
340 static u32 get_msglevel(struct net_device *dev)
341 {
342 	struct adapter *adapter = dev->ml_priv;
343 
344 	return adapter->msg_enable;
345 }
346 
347 static void set_msglevel(struct net_device *dev, u32 val)
348 {
349 	struct adapter *adapter = dev->ml_priv;
350 
351 	adapter->msg_enable = val;
352 }
353 
354 static const char stats_strings[][ETH_GSTRING_LEN] = {
355 	"TxOctetsOK",
356 	"TxOctetsBad",
357 	"TxUnicastFramesOK",
358 	"TxMulticastFramesOK",
359 	"TxBroadcastFramesOK",
360 	"TxPauseFrames",
361 	"TxFramesWithDeferredXmissions",
362 	"TxLateCollisions",
363 	"TxTotalCollisions",
364 	"TxFramesAbortedDueToXSCollisions",
365 	"TxUnderrun",
366 	"TxLengthErrors",
367 	"TxInternalMACXmitError",
368 	"TxFramesWithExcessiveDeferral",
369 	"TxFCSErrors",
370 	"TxJumboFramesOk",
371 	"TxJumboOctetsOk",
372 
373 	"RxOctetsOK",
374 	"RxOctetsBad",
375 	"RxUnicastFramesOK",
376 	"RxMulticastFramesOK",
377 	"RxBroadcastFramesOK",
378 	"RxPauseFrames",
379 	"RxFCSErrors",
380 	"RxAlignErrors",
381 	"RxSymbolErrors",
382 	"RxDataErrors",
383 	"RxSequenceErrors",
384 	"RxRuntErrors",
385 	"RxJabberErrors",
386 	"RxInternalMACRcvError",
387 	"RxInRangeLengthErrors",
388 	"RxOutOfRangeLengthField",
389 	"RxFrameTooLongErrors",
390 	"RxJumboFramesOk",
391 	"RxJumboOctetsOk",
392 
393 	/* Port stats */
394 	"RxCsumGood",
395 	"TxCsumOffload",
396 	"TxTso",
397 	"RxVlan",
398 	"TxVlan",
399 	"TxNeedHeadroom",
400 
401 	/* Interrupt stats */
402 	"rx drops",
403 	"pure_rsps",
404 	"unhandled irqs",
405 	"respQ_empty",
406 	"respQ_overflow",
407 	"freelistQ_empty",
408 	"pkt_too_big",
409 	"pkt_mismatch",
410 	"cmdQ_full0",
411 	"cmdQ_full1",
412 
413 	"espi_DIP2ParityErr",
414 	"espi_DIP4Err",
415 	"espi_RxDrops",
416 	"espi_TxDrops",
417 	"espi_RxOvfl",
418 	"espi_ParityErr"
419 };
420 
421 #define T2_REGMAP_SIZE (3 * 1024)
422 
423 static int get_regs_len(struct net_device *dev)
424 {
425 	return T2_REGMAP_SIZE;
426 }
427 
428 static void get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
429 {
430 	struct adapter *adapter = dev->ml_priv;
431 
432 	strlcpy(info->driver, DRV_NAME, sizeof(info->driver));
433 	strlcpy(info->bus_info, pci_name(adapter->pdev),
434 		sizeof(info->bus_info));
435 }
436 
437 static int get_sset_count(struct net_device *dev, int sset)
438 {
439 	switch (sset) {
440 	case ETH_SS_STATS:
441 		return ARRAY_SIZE(stats_strings);
442 	default:
443 		return -EOPNOTSUPP;
444 	}
445 }
446 
447 static void get_strings(struct net_device *dev, u32 stringset, u8 *data)
448 {
449 	if (stringset == ETH_SS_STATS)
450 		memcpy(data, stats_strings, sizeof(stats_strings));
451 }
452 
453 static void get_stats(struct net_device *dev, struct ethtool_stats *stats,
454 		      u64 *data)
455 {
456 	struct adapter *adapter = dev->ml_priv;
457 	struct cmac *mac = adapter->port[dev->if_port].mac;
458 	const struct cmac_statistics *s;
459 	const struct sge_intr_counts *t;
460 	struct sge_port_stats ss;
461 
462 	s = mac->ops->statistics_update(mac, MAC_STATS_UPDATE_FULL);
463 	t = t1_sge_get_intr_counts(adapter->sge);
464 	t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss);
465 
466 	*data++ = s->TxOctetsOK;
467 	*data++ = s->TxOctetsBad;
468 	*data++ = s->TxUnicastFramesOK;
469 	*data++ = s->TxMulticastFramesOK;
470 	*data++ = s->TxBroadcastFramesOK;
471 	*data++ = s->TxPauseFrames;
472 	*data++ = s->TxFramesWithDeferredXmissions;
473 	*data++ = s->TxLateCollisions;
474 	*data++ = s->TxTotalCollisions;
475 	*data++ = s->TxFramesAbortedDueToXSCollisions;
476 	*data++ = s->TxUnderrun;
477 	*data++ = s->TxLengthErrors;
478 	*data++ = s->TxInternalMACXmitError;
479 	*data++ = s->TxFramesWithExcessiveDeferral;
480 	*data++ = s->TxFCSErrors;
481 	*data++ = s->TxJumboFramesOK;
482 	*data++ = s->TxJumboOctetsOK;
483 
484 	*data++ = s->RxOctetsOK;
485 	*data++ = s->RxOctetsBad;
486 	*data++ = s->RxUnicastFramesOK;
487 	*data++ = s->RxMulticastFramesOK;
488 	*data++ = s->RxBroadcastFramesOK;
489 	*data++ = s->RxPauseFrames;
490 	*data++ = s->RxFCSErrors;
491 	*data++ = s->RxAlignErrors;
492 	*data++ = s->RxSymbolErrors;
493 	*data++ = s->RxDataErrors;
494 	*data++ = s->RxSequenceErrors;
495 	*data++ = s->RxRuntErrors;
496 	*data++ = s->RxJabberErrors;
497 	*data++ = s->RxInternalMACRcvError;
498 	*data++ = s->RxInRangeLengthErrors;
499 	*data++ = s->RxOutOfRangeLengthField;
500 	*data++ = s->RxFrameTooLongErrors;
501 	*data++ = s->RxJumboFramesOK;
502 	*data++ = s->RxJumboOctetsOK;
503 
504 	*data++ = ss.rx_cso_good;
505 	*data++ = ss.tx_cso;
506 	*data++ = ss.tx_tso;
507 	*data++ = ss.vlan_xtract;
508 	*data++ = ss.vlan_insert;
509 	*data++ = ss.tx_need_hdrroom;
510 
511 	*data++ = t->rx_drops;
512 	*data++ = t->pure_rsps;
513 	*data++ = t->unhandled_irqs;
514 	*data++ = t->respQ_empty;
515 	*data++ = t->respQ_overflow;
516 	*data++ = t->freelistQ_empty;
517 	*data++ = t->pkt_too_big;
518 	*data++ = t->pkt_mismatch;
519 	*data++ = t->cmdQ_full[0];
520 	*data++ = t->cmdQ_full[1];
521 
522 	if (adapter->espi) {
523 		const struct espi_intr_counts *e;
524 
525 		e = t1_espi_get_intr_counts(adapter->espi);
526 		*data++ = e->DIP2_parity_err;
527 		*data++ = e->DIP4_err;
528 		*data++ = e->rx_drops;
529 		*data++ = e->tx_drops;
530 		*data++ = e->rx_ovflw;
531 		*data++ = e->parity_err;
532 	}
533 }
534 
535 static inline void reg_block_dump(struct adapter *ap, void *buf,
536 				  unsigned int start, unsigned int end)
537 {
538 	u32 *p = buf + start;
539 
540 	for ( ; start <= end; start += sizeof(u32))
541 		*p++ = readl(ap->regs + start);
542 }
543 
544 static void get_regs(struct net_device *dev, struct ethtool_regs *regs,
545 		     void *buf)
546 {
547 	struct adapter *ap = dev->ml_priv;
548 
549 	/*
550 	 * Version scheme: bits 0..9: chip version, bits 10..15: chip revision
551 	 */
552 	regs->version = 2;
553 
554 	memset(buf, 0, T2_REGMAP_SIZE);
555 	reg_block_dump(ap, buf, 0, A_SG_RESPACCUTIMER);
556 	reg_block_dump(ap, buf, A_MC3_CFG, A_MC4_INT_CAUSE);
557 	reg_block_dump(ap, buf, A_TPI_ADDR, A_TPI_PAR);
558 	reg_block_dump(ap, buf, A_TP_IN_CONFIG, A_TP_TX_DROP_COUNT);
559 	reg_block_dump(ap, buf, A_RAT_ROUTE_CONTROL, A_RAT_INTR_CAUSE);
560 	reg_block_dump(ap, buf, A_CSPI_RX_AE_WM, A_CSPI_INTR_ENABLE);
561 	reg_block_dump(ap, buf, A_ESPI_SCH_TOKEN0, A_ESPI_GOSTAT);
562 	reg_block_dump(ap, buf, A_ULP_ULIMIT, A_ULP_PIO_CTRL);
563 	reg_block_dump(ap, buf, A_PL_ENABLE, A_PL_CAUSE);
564 	reg_block_dump(ap, buf, A_MC5_CONFIG, A_MC5_MASK_WRITE_CMD);
565 }
566 
567 static int get_link_ksettings(struct net_device *dev,
568 			      struct ethtool_link_ksettings *cmd)
569 {
570 	struct adapter *adapter = dev->ml_priv;
571 	struct port_info *p = &adapter->port[dev->if_port];
572 	u32 supported, advertising;
573 
574 	supported = p->link_config.supported;
575 	advertising = p->link_config.advertising;
576 
577 	if (netif_carrier_ok(dev)) {
578 		cmd->base.speed = p->link_config.speed;
579 		cmd->base.duplex = p->link_config.duplex;
580 	} else {
581 		cmd->base.speed = SPEED_UNKNOWN;
582 		cmd->base.duplex = DUPLEX_UNKNOWN;
583 	}
584 
585 	cmd->base.port = (supported & SUPPORTED_TP) ? PORT_TP : PORT_FIBRE;
586 	cmd->base.phy_address = p->phy->mdio.prtad;
587 	cmd->base.autoneg = p->link_config.autoneg;
588 
589 	ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported,
590 						supported);
591 	ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising,
592 						advertising);
593 
594 	return 0;
595 }
596 
597 static int speed_duplex_to_caps(int speed, int duplex)
598 {
599 	int cap = 0;
600 
601 	switch (speed) {
602 	case SPEED_10:
603 		if (duplex == DUPLEX_FULL)
604 			cap = SUPPORTED_10baseT_Full;
605 		else
606 			cap = SUPPORTED_10baseT_Half;
607 		break;
608 	case SPEED_100:
609 		if (duplex == DUPLEX_FULL)
610 			cap = SUPPORTED_100baseT_Full;
611 		else
612 			cap = SUPPORTED_100baseT_Half;
613 		break;
614 	case SPEED_1000:
615 		if (duplex == DUPLEX_FULL)
616 			cap = SUPPORTED_1000baseT_Full;
617 		else
618 			cap = SUPPORTED_1000baseT_Half;
619 		break;
620 	case SPEED_10000:
621 		if (duplex == DUPLEX_FULL)
622 			cap = SUPPORTED_10000baseT_Full;
623 	}
624 	return cap;
625 }
626 
627 #define ADVERTISED_MASK (ADVERTISED_10baseT_Half | ADVERTISED_10baseT_Full | \
628 		      ADVERTISED_100baseT_Half | ADVERTISED_100baseT_Full | \
629 		      ADVERTISED_1000baseT_Half | ADVERTISED_1000baseT_Full | \
630 		      ADVERTISED_10000baseT_Full)
631 
632 static int set_link_ksettings(struct net_device *dev,
633 			      const struct ethtool_link_ksettings *cmd)
634 {
635 	struct adapter *adapter = dev->ml_priv;
636 	struct port_info *p = &adapter->port[dev->if_port];
637 	struct link_config *lc = &p->link_config;
638 	u32 advertising;
639 
640 	ethtool_convert_link_mode_to_legacy_u32(&advertising,
641 						cmd->link_modes.advertising);
642 
643 	if (!(lc->supported & SUPPORTED_Autoneg))
644 		return -EOPNOTSUPP;             /* can't change speed/duplex */
645 
646 	if (cmd->base.autoneg == AUTONEG_DISABLE) {
647 		u32 speed = cmd->base.speed;
648 		int cap = speed_duplex_to_caps(speed, cmd->base.duplex);
649 
650 		if (!(lc->supported & cap) || (speed == SPEED_1000))
651 			return -EINVAL;
652 		lc->requested_speed = speed;
653 		lc->requested_duplex = cmd->base.duplex;
654 		lc->advertising = 0;
655 	} else {
656 		advertising &= ADVERTISED_MASK;
657 		if (advertising & (advertising - 1))
658 			advertising = lc->supported;
659 		advertising &= lc->supported;
660 		if (!advertising)
661 			return -EINVAL;
662 		lc->requested_speed = SPEED_INVALID;
663 		lc->requested_duplex = DUPLEX_INVALID;
664 		lc->advertising = advertising | ADVERTISED_Autoneg;
665 	}
666 	lc->autoneg = cmd->base.autoneg;
667 	if (netif_running(dev))
668 		t1_link_start(p->phy, p->mac, lc);
669 	return 0;
670 }
671 
672 static void get_pauseparam(struct net_device *dev,
673 			   struct ethtool_pauseparam *epause)
674 {
675 	struct adapter *adapter = dev->ml_priv;
676 	struct port_info *p = &adapter->port[dev->if_port];
677 
678 	epause->autoneg = (p->link_config.requested_fc & PAUSE_AUTONEG) != 0;
679 	epause->rx_pause = (p->link_config.fc & PAUSE_RX) != 0;
680 	epause->tx_pause = (p->link_config.fc & PAUSE_TX) != 0;
681 }
682 
683 static int set_pauseparam(struct net_device *dev,
684 			  struct ethtool_pauseparam *epause)
685 {
686 	struct adapter *adapter = dev->ml_priv;
687 	struct port_info *p = &adapter->port[dev->if_port];
688 	struct link_config *lc = &p->link_config;
689 
690 	if (epause->autoneg == AUTONEG_DISABLE)
691 		lc->requested_fc = 0;
692 	else if (lc->supported & SUPPORTED_Autoneg)
693 		lc->requested_fc = PAUSE_AUTONEG;
694 	else
695 		return -EINVAL;
696 
697 	if (epause->rx_pause)
698 		lc->requested_fc |= PAUSE_RX;
699 	if (epause->tx_pause)
700 		lc->requested_fc |= PAUSE_TX;
701 	if (lc->autoneg == AUTONEG_ENABLE) {
702 		if (netif_running(dev))
703 			t1_link_start(p->phy, p->mac, lc);
704 	} else {
705 		lc->fc = lc->requested_fc & (PAUSE_RX | PAUSE_TX);
706 		if (netif_running(dev))
707 			p->mac->ops->set_speed_duplex_fc(p->mac, -1, -1,
708 							 lc->fc);
709 	}
710 	return 0;
711 }
712 
713 static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e)
714 {
715 	struct adapter *adapter = dev->ml_priv;
716 	int jumbo_fl = t1_is_T1B(adapter) ? 1 : 0;
717 
718 	e->rx_max_pending = MAX_RX_BUFFERS;
719 	e->rx_jumbo_max_pending = MAX_RX_JUMBO_BUFFERS;
720 	e->tx_max_pending = MAX_CMDQ_ENTRIES;
721 
722 	e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl];
723 	e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl];
724 	e->tx_pending = adapter->params.sge.cmdQ_size[0];
725 }
726 
727 static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e)
728 {
729 	struct adapter *adapter = dev->ml_priv;
730 	int jumbo_fl = t1_is_T1B(adapter) ? 1 : 0;
731 
732 	if (e->rx_pending > MAX_RX_BUFFERS || e->rx_mini_pending ||
733 	    e->rx_jumbo_pending > MAX_RX_JUMBO_BUFFERS ||
734 	    e->tx_pending > MAX_CMDQ_ENTRIES ||
735 	    e->rx_pending < MIN_FL_ENTRIES ||
736 	    e->rx_jumbo_pending < MIN_FL_ENTRIES ||
737 	    e->tx_pending < (adapter->params.nports + 1) * (MAX_SKB_FRAGS + 1))
738 		return -EINVAL;
739 
740 	if (adapter->flags & FULL_INIT_DONE)
741 		return -EBUSY;
742 
743 	adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending;
744 	adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending;
745 	adapter->params.sge.cmdQ_size[0] = e->tx_pending;
746 	adapter->params.sge.cmdQ_size[1] = e->tx_pending > MAX_CMDQ1_ENTRIES ?
747 		MAX_CMDQ1_ENTRIES : e->tx_pending;
748 	return 0;
749 }
750 
751 static int set_coalesce(struct net_device *dev, struct ethtool_coalesce *c)
752 {
753 	struct adapter *adapter = dev->ml_priv;
754 
755 	adapter->params.sge.rx_coalesce_usecs = c->rx_coalesce_usecs;
756 	adapter->params.sge.coalesce_enable = c->use_adaptive_rx_coalesce;
757 	adapter->params.sge.sample_interval_usecs = c->rate_sample_interval;
758 	t1_sge_set_coalesce_params(adapter->sge, &adapter->params.sge);
759 	return 0;
760 }
761 
762 static int get_coalesce(struct net_device *dev, struct ethtool_coalesce *c)
763 {
764 	struct adapter *adapter = dev->ml_priv;
765 
766 	c->rx_coalesce_usecs = adapter->params.sge.rx_coalesce_usecs;
767 	c->rate_sample_interval = adapter->params.sge.sample_interval_usecs;
768 	c->use_adaptive_rx_coalesce = adapter->params.sge.coalesce_enable;
769 	return 0;
770 }
771 
772 static int get_eeprom_len(struct net_device *dev)
773 {
774 	struct adapter *adapter = dev->ml_priv;
775 
776 	return t1_is_asic(adapter) ? EEPROM_SIZE : 0;
777 }
778 
779 #define EEPROM_MAGIC(ap) \
780 	(PCI_VENDOR_ID_CHELSIO | ((ap)->params.chip_version << 16))
781 
782 static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e,
783 		      u8 *data)
784 {
785 	int i;
786 	u8 buf[EEPROM_SIZE] __attribute__((aligned(4)));
787 	struct adapter *adapter = dev->ml_priv;
788 
789 	e->magic = EEPROM_MAGIC(adapter);
790 	for (i = e->offset & ~3; i < e->offset + e->len; i += sizeof(u32))
791 		t1_seeprom_read(adapter, i, (__le32 *)&buf[i]);
792 	memcpy(data, buf + e->offset, e->len);
793 	return 0;
794 }
795 
796 static const struct ethtool_ops t1_ethtool_ops = {
797 	.supported_coalesce_params = ETHTOOL_COALESCE_RX_USECS |
798 				     ETHTOOL_COALESCE_USE_ADAPTIVE_RX |
799 				     ETHTOOL_COALESCE_RATE_SAMPLE_INTERVAL,
800 	.get_drvinfo       = get_drvinfo,
801 	.get_msglevel      = get_msglevel,
802 	.set_msglevel      = set_msglevel,
803 	.get_ringparam     = get_sge_param,
804 	.set_ringparam     = set_sge_param,
805 	.get_coalesce      = get_coalesce,
806 	.set_coalesce      = set_coalesce,
807 	.get_eeprom_len    = get_eeprom_len,
808 	.get_eeprom        = get_eeprom,
809 	.get_pauseparam    = get_pauseparam,
810 	.set_pauseparam    = set_pauseparam,
811 	.get_link          = ethtool_op_get_link,
812 	.get_strings       = get_strings,
813 	.get_sset_count	   = get_sset_count,
814 	.get_ethtool_stats = get_stats,
815 	.get_regs_len      = get_regs_len,
816 	.get_regs          = get_regs,
817 	.get_link_ksettings = get_link_ksettings,
818 	.set_link_ksettings = set_link_ksettings,
819 };
820 
821 static int t1_ioctl(struct net_device *dev, struct ifreq *req, int cmd)
822 {
823 	struct adapter *adapter = dev->ml_priv;
824 	struct mdio_if_info *mdio = &adapter->port[dev->if_port].phy->mdio;
825 
826 	return mdio_mii_ioctl(mdio, if_mii(req), cmd);
827 }
828 
829 static int t1_change_mtu(struct net_device *dev, int new_mtu)
830 {
831 	int ret;
832 	struct adapter *adapter = dev->ml_priv;
833 	struct cmac *mac = adapter->port[dev->if_port].mac;
834 
835 	if (!mac->ops->set_mtu)
836 		return -EOPNOTSUPP;
837 	if ((ret = mac->ops->set_mtu(mac, new_mtu)))
838 		return ret;
839 	dev->mtu = new_mtu;
840 	return 0;
841 }
842 
843 static int t1_set_mac_addr(struct net_device *dev, void *p)
844 {
845 	struct adapter *adapter = dev->ml_priv;
846 	struct cmac *mac = adapter->port[dev->if_port].mac;
847 	struct sockaddr *addr = p;
848 
849 	if (!mac->ops->macaddress_set)
850 		return -EOPNOTSUPP;
851 
852 	memcpy(dev->dev_addr, addr->sa_data, dev->addr_len);
853 	mac->ops->macaddress_set(mac, dev->dev_addr);
854 	return 0;
855 }
856 
857 static netdev_features_t t1_fix_features(struct net_device *dev,
858 	netdev_features_t features)
859 {
860 	/*
861 	 * Since there is no support for separate rx/tx vlan accel
862 	 * enable/disable make sure tx flag is always in same state as rx.
863 	 */
864 	if (features & NETIF_F_HW_VLAN_CTAG_RX)
865 		features |= NETIF_F_HW_VLAN_CTAG_TX;
866 	else
867 		features &= ~NETIF_F_HW_VLAN_CTAG_TX;
868 
869 	return features;
870 }
871 
872 static int t1_set_features(struct net_device *dev, netdev_features_t features)
873 {
874 	netdev_features_t changed = dev->features ^ features;
875 	struct adapter *adapter = dev->ml_priv;
876 
877 	if (changed & NETIF_F_HW_VLAN_CTAG_RX)
878 		t1_vlan_mode(adapter, features);
879 
880 	return 0;
881 }
882 #ifdef CONFIG_NET_POLL_CONTROLLER
883 static void t1_netpoll(struct net_device *dev)
884 {
885 	unsigned long flags;
886 	struct adapter *adapter = dev->ml_priv;
887 
888 	local_irq_save(flags);
889 	t1_interrupt(adapter->pdev->irq, adapter);
890 	local_irq_restore(flags);
891 }
892 #endif
893 
894 /*
895  * Periodic accumulation of MAC statistics.  This is used only if the MAC
896  * does not have any other way to prevent stats counter overflow.
897  */
898 static void mac_stats_task(struct work_struct *work)
899 {
900 	int i;
901 	struct adapter *adapter =
902 		container_of(work, struct adapter, stats_update_task.work);
903 
904 	for_each_port(adapter, i) {
905 		struct port_info *p = &adapter->port[i];
906 
907 		if (netif_running(p->dev))
908 			p->mac->ops->statistics_update(p->mac,
909 						       MAC_STATS_UPDATE_FAST);
910 	}
911 
912 	/* Schedule the next statistics update if any port is active. */
913 	spin_lock(&adapter->work_lock);
914 	if (adapter->open_device_map & PORT_MASK)
915 		schedule_mac_stats_update(adapter,
916 					  adapter->params.stats_update_period);
917 	spin_unlock(&adapter->work_lock);
918 }
919 
920 static const struct net_device_ops cxgb_netdev_ops = {
921 	.ndo_open		= cxgb_open,
922 	.ndo_stop		= cxgb_close,
923 	.ndo_start_xmit		= t1_start_xmit,
924 	.ndo_get_stats		= t1_get_stats,
925 	.ndo_validate_addr	= eth_validate_addr,
926 	.ndo_set_rx_mode	= t1_set_rxmode,
927 	.ndo_eth_ioctl		= t1_ioctl,
928 	.ndo_change_mtu		= t1_change_mtu,
929 	.ndo_set_mac_address	= t1_set_mac_addr,
930 	.ndo_fix_features	= t1_fix_features,
931 	.ndo_set_features	= t1_set_features,
932 #ifdef CONFIG_NET_POLL_CONTROLLER
933 	.ndo_poll_controller	= t1_netpoll,
934 #endif
935 };
936 
937 static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
938 {
939 	int i, err, pci_using_dac = 0;
940 	unsigned long mmio_start, mmio_len;
941 	const struct board_info *bi;
942 	struct adapter *adapter = NULL;
943 	struct port_info *pi;
944 
945 	err = pci_enable_device(pdev);
946 	if (err)
947 		return err;
948 
949 	if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) {
950 		pr_err("%s: cannot find PCI device memory base address\n",
951 		       pci_name(pdev));
952 		err = -ENODEV;
953 		goto out_disable_pdev;
954 	}
955 
956 	if (!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64))) {
957 		pci_using_dac = 1;
958 
959 		if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64))) {
960 			pr_err("%s: unable to obtain 64-bit DMA for coherent allocations\n",
961 			       pci_name(pdev));
962 			err = -ENODEV;
963 			goto out_disable_pdev;
964 		}
965 
966 	} else if ((err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) != 0) {
967 		pr_err("%s: no usable DMA configuration\n", pci_name(pdev));
968 		goto out_disable_pdev;
969 	}
970 
971 	err = pci_request_regions(pdev, DRV_NAME);
972 	if (err) {
973 		pr_err("%s: cannot obtain PCI resources\n", pci_name(pdev));
974 		goto out_disable_pdev;
975 	}
976 
977 	pci_set_master(pdev);
978 
979 	mmio_start = pci_resource_start(pdev, 0);
980 	mmio_len = pci_resource_len(pdev, 0);
981 	bi = t1_get_board_info(ent->driver_data);
982 
983 	for (i = 0; i < bi->port_number; ++i) {
984 		struct net_device *netdev;
985 
986 		netdev = alloc_etherdev(adapter ? 0 : sizeof(*adapter));
987 		if (!netdev) {
988 			err = -ENOMEM;
989 			goto out_free_dev;
990 		}
991 
992 		SET_NETDEV_DEV(netdev, &pdev->dev);
993 
994 		if (!adapter) {
995 			adapter = netdev_priv(netdev);
996 			adapter->pdev = pdev;
997 			adapter->port[0].dev = netdev;  /* so we don't leak it */
998 
999 			adapter->regs = ioremap(mmio_start, mmio_len);
1000 			if (!adapter->regs) {
1001 				pr_err("%s: cannot map device registers\n",
1002 				       pci_name(pdev));
1003 				err = -ENOMEM;
1004 				goto out_free_dev;
1005 			}
1006 
1007 			if (t1_get_board_rev(adapter, bi, &adapter->params)) {
1008 				err = -ENODEV;	  /* Can't handle this chip rev */
1009 				goto out_free_dev;
1010 			}
1011 
1012 			adapter->name = pci_name(pdev);
1013 			adapter->msg_enable = dflt_msg_enable;
1014 			adapter->mmio_len = mmio_len;
1015 
1016 			spin_lock_init(&adapter->tpi_lock);
1017 			spin_lock_init(&adapter->work_lock);
1018 			spin_lock_init(&adapter->async_lock);
1019 			spin_lock_init(&adapter->mac_lock);
1020 
1021 			INIT_DELAYED_WORK(&adapter->stats_update_task,
1022 					  mac_stats_task);
1023 
1024 			pci_set_drvdata(pdev, netdev);
1025 		}
1026 
1027 		pi = &adapter->port[i];
1028 		pi->dev = netdev;
1029 		netif_carrier_off(netdev);
1030 		netdev->irq = pdev->irq;
1031 		netdev->if_port = i;
1032 		netdev->mem_start = mmio_start;
1033 		netdev->mem_end = mmio_start + mmio_len - 1;
1034 		netdev->ml_priv = adapter;
1035 		netdev->hw_features |= NETIF_F_SG | NETIF_F_IP_CSUM |
1036 			NETIF_F_RXCSUM;
1037 		netdev->features |= NETIF_F_SG | NETIF_F_IP_CSUM |
1038 			NETIF_F_RXCSUM | NETIF_F_LLTX;
1039 
1040 		if (pci_using_dac)
1041 			netdev->features |= NETIF_F_HIGHDMA;
1042 		if (vlan_tso_capable(adapter)) {
1043 			netdev->features |=
1044 				NETIF_F_HW_VLAN_CTAG_TX |
1045 				NETIF_F_HW_VLAN_CTAG_RX;
1046 			netdev->hw_features |= NETIF_F_HW_VLAN_CTAG_RX;
1047 
1048 			/* T204: disable TSO */
1049 			if (!(is_T2(adapter)) || bi->port_number != 4) {
1050 				netdev->hw_features |= NETIF_F_TSO;
1051 				netdev->features |= NETIF_F_TSO;
1052 			}
1053 		}
1054 
1055 		netdev->netdev_ops = &cxgb_netdev_ops;
1056 		netdev->hard_header_len += (netdev->hw_features & NETIF_F_TSO) ?
1057 			sizeof(struct cpl_tx_pkt_lso) : sizeof(struct cpl_tx_pkt);
1058 
1059 		netif_napi_add(netdev, &adapter->napi, t1_poll, 64);
1060 
1061 		netdev->ethtool_ops = &t1_ethtool_ops;
1062 
1063 		switch (bi->board) {
1064 		case CHBT_BOARD_CHT110:
1065 		case CHBT_BOARD_N110:
1066 		case CHBT_BOARD_N210:
1067 		case CHBT_BOARD_CHT210:
1068 			netdev->max_mtu = PM3393_MAX_FRAME_SIZE -
1069 					  (ETH_HLEN + ETH_FCS_LEN);
1070 			break;
1071 		case CHBT_BOARD_CHN204:
1072 			netdev->max_mtu = VSC7326_MAX_MTU;
1073 			break;
1074 		default:
1075 			netdev->max_mtu = ETH_DATA_LEN;
1076 			break;
1077 		}
1078 	}
1079 
1080 	if (t1_init_sw_modules(adapter, bi) < 0) {
1081 		err = -ENODEV;
1082 		goto out_free_dev;
1083 	}
1084 
1085 	/*
1086 	 * The card is now ready to go.  If any errors occur during device
1087 	 * registration we do not fail the whole card but rather proceed only
1088 	 * with the ports we manage to register successfully.  However we must
1089 	 * register at least one net device.
1090 	 */
1091 	for (i = 0; i < bi->port_number; ++i) {
1092 		err = register_netdev(adapter->port[i].dev);
1093 		if (err)
1094 			pr_warn("%s: cannot register net device %s, skipping\n",
1095 				pci_name(pdev), adapter->port[i].dev->name);
1096 		else {
1097 			/*
1098 			 * Change the name we use for messages to the name of
1099 			 * the first successfully registered interface.
1100 			 */
1101 			if (!adapter->registered_device_map)
1102 				adapter->name = adapter->port[i].dev->name;
1103 
1104 			__set_bit(i, &adapter->registered_device_map);
1105 		}
1106 	}
1107 	if (!adapter->registered_device_map) {
1108 		pr_err("%s: could not register any net devices\n",
1109 		       pci_name(pdev));
1110 		goto out_release_adapter_res;
1111 	}
1112 
1113 	pr_info("%s: %s (rev %d), %s %dMHz/%d-bit\n",
1114 		adapter->name, bi->desc, adapter->params.chip_revision,
1115 		adapter->params.pci.is_pcix ? "PCIX" : "PCI",
1116 		adapter->params.pci.speed, adapter->params.pci.width);
1117 
1118 	/*
1119 	 * Set the T1B ASIC and memory clocks.
1120 	 */
1121 	if (t1powersave)
1122 		adapter->t1powersave = LCLOCK;	/* HW default is powersave mode. */
1123 	else
1124 		adapter->t1powersave = HCLOCK;
1125 	if (t1_is_T1B(adapter))
1126 		t1_clock(adapter, t1powersave);
1127 
1128 	return 0;
1129 
1130 out_release_adapter_res:
1131 	t1_free_sw_modules(adapter);
1132 out_free_dev:
1133 	if (adapter) {
1134 		if (adapter->regs)
1135 			iounmap(adapter->regs);
1136 		for (i = bi->port_number - 1; i >= 0; --i)
1137 			if (adapter->port[i].dev)
1138 				free_netdev(adapter->port[i].dev);
1139 	}
1140 	pci_release_regions(pdev);
1141 out_disable_pdev:
1142 	pci_disable_device(pdev);
1143 	return err;
1144 }
1145 
1146 static void bit_bang(struct adapter *adapter, int bitdata, int nbits)
1147 {
1148 	int data;
1149 	int i;
1150 	u32 val;
1151 
1152 	enum {
1153 		S_CLOCK = 1 << 3,
1154 		S_DATA = 1 << 4
1155 	};
1156 
1157 	for (i = (nbits - 1); i > -1; i--) {
1158 
1159 		udelay(50);
1160 
1161 		data = ((bitdata >> i) & 0x1);
1162 		__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1163 
1164 		if (data)
1165 			val |= S_DATA;
1166 		else
1167 			val &= ~S_DATA;
1168 
1169 		udelay(50);
1170 
1171 		/* Set SCLOCK low */
1172 		val &= ~S_CLOCK;
1173 		__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1174 
1175 		udelay(50);
1176 
1177 		/* Write SCLOCK high */
1178 		val |= S_CLOCK;
1179 		__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1180 
1181 	}
1182 }
1183 
1184 static int t1_clock(struct adapter *adapter, int mode)
1185 {
1186 	u32 val;
1187 	int M_CORE_VAL;
1188 	int M_MEM_VAL;
1189 
1190 	enum {
1191 		M_CORE_BITS	= 9,
1192 		T_CORE_VAL	= 0,
1193 		T_CORE_BITS	= 2,
1194 		N_CORE_VAL	= 0,
1195 		N_CORE_BITS	= 2,
1196 		M_MEM_BITS	= 9,
1197 		T_MEM_VAL	= 0,
1198 		T_MEM_BITS	= 2,
1199 		N_MEM_VAL	= 0,
1200 		N_MEM_BITS	= 2,
1201 		NP_LOAD		= 1 << 17,
1202 		S_LOAD_MEM	= 1 << 5,
1203 		S_LOAD_CORE	= 1 << 6,
1204 		S_CLOCK		= 1 << 3
1205 	};
1206 
1207 	if (!t1_is_T1B(adapter))
1208 		return -ENODEV;	/* Can't re-clock this chip. */
1209 
1210 	if (mode & 2)
1211 		return 0;	/* show current mode. */
1212 
1213 	if ((adapter->t1powersave & 1) == (mode & 1))
1214 		return -EALREADY;	/* ASIC already running in mode. */
1215 
1216 	if ((mode & 1) == HCLOCK) {
1217 		M_CORE_VAL = 0x14;
1218 		M_MEM_VAL = 0x18;
1219 		adapter->t1powersave = HCLOCK;	/* overclock */
1220 	} else {
1221 		M_CORE_VAL = 0xe;
1222 		M_MEM_VAL = 0x10;
1223 		adapter->t1powersave = LCLOCK;	/* underclock */
1224 	}
1225 
1226 	/* Don't interrupt this serial stream! */
1227 	spin_lock(&adapter->tpi_lock);
1228 
1229 	/* Initialize for ASIC core */
1230 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1231 	val |= NP_LOAD;
1232 	udelay(50);
1233 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1234 	udelay(50);
1235 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1236 	val &= ~S_LOAD_CORE;
1237 	val &= ~S_CLOCK;
1238 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1239 	udelay(50);
1240 
1241 	/* Serial program the ASIC clock synthesizer */
1242 	bit_bang(adapter, T_CORE_VAL, T_CORE_BITS);
1243 	bit_bang(adapter, N_CORE_VAL, N_CORE_BITS);
1244 	bit_bang(adapter, M_CORE_VAL, M_CORE_BITS);
1245 	udelay(50);
1246 
1247 	/* Finish ASIC core */
1248 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1249 	val |= S_LOAD_CORE;
1250 	udelay(50);
1251 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1252 	udelay(50);
1253 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1254 	val &= ~S_LOAD_CORE;
1255 	udelay(50);
1256 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1257 	udelay(50);
1258 
1259 	/* Initialize for memory */
1260 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1261 	val |= NP_LOAD;
1262 	udelay(50);
1263 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1264 	udelay(50);
1265 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1266 	val &= ~S_LOAD_MEM;
1267 	val &= ~S_CLOCK;
1268 	udelay(50);
1269 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1270 	udelay(50);
1271 
1272 	/* Serial program the memory clock synthesizer */
1273 	bit_bang(adapter, T_MEM_VAL, T_MEM_BITS);
1274 	bit_bang(adapter, N_MEM_VAL, N_MEM_BITS);
1275 	bit_bang(adapter, M_MEM_VAL, M_MEM_BITS);
1276 	udelay(50);
1277 
1278 	/* Finish memory */
1279 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1280 	val |= S_LOAD_MEM;
1281 	udelay(50);
1282 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1283 	udelay(50);
1284 	__t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1285 	val &= ~S_LOAD_MEM;
1286 	udelay(50);
1287 	__t1_tpi_write(adapter, A_ELMER0_GPO, val);
1288 
1289 	spin_unlock(&adapter->tpi_lock);
1290 
1291 	return 0;
1292 }
1293 
1294 static inline void t1_sw_reset(struct pci_dev *pdev)
1295 {
1296 	pci_write_config_dword(pdev, A_PCICFG_PM_CSR, 3);
1297 	pci_write_config_dword(pdev, A_PCICFG_PM_CSR, 0);
1298 }
1299 
1300 static void remove_one(struct pci_dev *pdev)
1301 {
1302 	struct net_device *dev = pci_get_drvdata(pdev);
1303 	struct adapter *adapter = dev->ml_priv;
1304 	int i;
1305 
1306 	for_each_port(adapter, i) {
1307 		if (test_bit(i, &adapter->registered_device_map))
1308 			unregister_netdev(adapter->port[i].dev);
1309 	}
1310 
1311 	t1_free_sw_modules(adapter);
1312 	iounmap(adapter->regs);
1313 
1314 	while (--i >= 0) {
1315 		if (adapter->port[i].dev)
1316 			free_netdev(adapter->port[i].dev);
1317 	}
1318 
1319 	pci_release_regions(pdev);
1320 	pci_disable_device(pdev);
1321 	t1_sw_reset(pdev);
1322 }
1323 
1324 static struct pci_driver cxgb_pci_driver = {
1325 	.name     = DRV_NAME,
1326 	.id_table = t1_pci_tbl,
1327 	.probe    = init_one,
1328 	.remove   = remove_one,
1329 };
1330 
1331 module_pci_driver(cxgb_pci_driver);
1332