[netdrvr] Stop using legacy hooks ->self_test_count, ->get_stats_count
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / drivers / net / chelsio / cxgb2.c
1 /*****************************************************************************
2 * *
3 * File: cxgb2.c *
4 * $Revision: 1.25 $ *
5 * $Date: 2005/06/22 00:43:25 $ *
6 * Description: *
7 * Chelsio 10Gb Ethernet Driver. *
8 * *
9 * This program is free software; you can redistribute it and/or modify *
10 * it under the terms of the GNU General Public License, version 2, as *
11 * published by the Free Software Foundation. *
12 * *
13 * You should have received a copy of the GNU General Public License along *
14 * with this program; if not, write to the Free Software Foundation, Inc., *
15 * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. *
16 * *
17 * THIS SOFTWARE IS PROVIDED ``AS IS'' AND WITHOUT ANY EXPRESS OR IMPLIED *
18 * WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE IMPLIED WARRANTIES OF *
19 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. *
20 * *
21 * http://www.chelsio.com *
22 * *
23 * Copyright (c) 2003 - 2005 Chelsio Communications, Inc. *
24 * All rights reserved. *
25 * *
26 * Maintainers: maintainers@chelsio.com *
27 * *
28 * Authors: Dimitrios Michailidis <dm@chelsio.com> *
29 * Tina Yang <tainay@chelsio.com> *
30 * Felix Marti <felix@chelsio.com> *
31 * Scott Bardone <sbardone@chelsio.com> *
32 * Kurt Ottaway <kottaway@chelsio.com> *
33 * Frank DiMambro <frank@chelsio.com> *
34 * *
35 * History: *
36 * *
37 ****************************************************************************/
38
39 #include "common.h"
40 #include <linux/module.h>
41 #include <linux/init.h>
42 #include <linux/pci.h>
43 #include <linux/netdevice.h>
44 #include <linux/etherdevice.h>
45 #include <linux/if_vlan.h>
46 #include <linux/mii.h>
47 #include <linux/sockios.h>
48 #include <linux/dma-mapping.h>
49 #include <asm/uaccess.h>
50
51 #include "cpl5_cmd.h"
52 #include "regs.h"
53 #include "gmac.h"
54 #include "cphy.h"
55 #include "sge.h"
56 #include "tp.h"
57 #include "espi.h"
58 #include "elmer0.h"
59
60 #include <linux/workqueue.h>
61
62 static inline void schedule_mac_stats_update(struct adapter *ap, int secs)
63 {
64 schedule_delayed_work(&ap->stats_update_task, secs * HZ);
65 }
66
67 static inline void cancel_mac_stats_update(struct adapter *ap)
68 {
69 cancel_delayed_work(&ap->stats_update_task);
70 }
71
72 #define MAX_CMDQ_ENTRIES 16384
73 #define MAX_CMDQ1_ENTRIES 1024
74 #define MAX_RX_BUFFERS 16384
75 #define MAX_RX_JUMBO_BUFFERS 16384
76 #define MAX_TX_BUFFERS_HIGH 16384U
77 #define MAX_TX_BUFFERS_LOW 1536U
78 #define MAX_TX_BUFFERS 1460U
79 #define MIN_FL_ENTRIES 32
80
81 #define DFLT_MSG_ENABLE (NETIF_MSG_DRV | NETIF_MSG_PROBE | NETIF_MSG_LINK | \
82 NETIF_MSG_TIMER | NETIF_MSG_IFDOWN | NETIF_MSG_IFUP |\
83 NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR)
84
85 /*
86 * The EEPROM is actually bigger but only the first few bytes are used so we
87 * only report those.
88 */
89 #define EEPROM_SIZE 32
90
91 MODULE_DESCRIPTION(DRV_DESCRIPTION);
92 MODULE_AUTHOR("Chelsio Communications");
93 MODULE_LICENSE("GPL");
94
95 static int dflt_msg_enable = DFLT_MSG_ENABLE;
96
97 module_param(dflt_msg_enable, int, 0);
98 MODULE_PARM_DESC(dflt_msg_enable, "Chelsio T1 default message enable bitmap");
99
100 #define HCLOCK 0x0
101 #define LCLOCK 0x1
102
103 /* T1 cards powersave mode */
104 static int t1_clock(struct adapter *adapter, int mode);
105 static int t1powersave = 1; /* HW default is powersave mode. */
106
107 module_param(t1powersave, int, 0);
108 MODULE_PARM_DESC(t1powersave, "Enable/Disable T1 powersaving mode");
109
110 static int disable_msi = 0;
111 module_param(disable_msi, int, 0);
112 MODULE_PARM_DESC(disable_msi, "Disable Message Signaled Interrupt (MSI)");
113
114 static const char pci_speed[][4] = {
115 "33", "66", "100", "133"
116 };
117
118 /*
119 * Setup MAC to receive the types of packets we want.
120 */
121 static void t1_set_rxmode(struct net_device *dev)
122 {
123 struct adapter *adapter = dev->priv;
124 struct cmac *mac = adapter->port[dev->if_port].mac;
125 struct t1_rx_mode rm;
126
127 rm.dev = dev;
128 rm.idx = 0;
129 rm.list = dev->mc_list;
130 mac->ops->set_rx_mode(mac, &rm);
131 }
132
133 static void link_report(struct port_info *p)
134 {
135 if (!netif_carrier_ok(p->dev))
136 printk(KERN_INFO "%s: link down\n", p->dev->name);
137 else {
138 const char *s = "10Mbps";
139
140 switch (p->link_config.speed) {
141 case SPEED_10000: s = "10Gbps"; break;
142 case SPEED_1000: s = "1000Mbps"; break;
143 case SPEED_100: s = "100Mbps"; break;
144 }
145
146 printk(KERN_INFO "%s: link up, %s, %s-duplex\n",
147 p->dev->name, s,
148 p->link_config.duplex == DUPLEX_FULL ? "full" : "half");
149 }
150 }
151
152 void t1_link_negotiated(struct adapter *adapter, int port_id, int link_stat,
153 int speed, int duplex, int pause)
154 {
155 struct port_info *p = &adapter->port[port_id];
156
157 if (link_stat != netif_carrier_ok(p->dev)) {
158 if (link_stat)
159 netif_carrier_on(p->dev);
160 else
161 netif_carrier_off(p->dev);
162 link_report(p);
163
164 /* multi-ports: inform toe */
165 if ((speed > 0) && (adapter->params.nports > 1)) {
166 unsigned int sched_speed = 10;
167 switch (speed) {
168 case SPEED_1000:
169 sched_speed = 1000;
170 break;
171 case SPEED_100:
172 sched_speed = 100;
173 break;
174 case SPEED_10:
175 sched_speed = 10;
176 break;
177 }
178 t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed);
179 }
180 }
181 }
182
183 static void link_start(struct port_info *p)
184 {
185 struct cmac *mac = p->mac;
186
187 mac->ops->reset(mac);
188 if (mac->ops->macaddress_set)
189 mac->ops->macaddress_set(mac, p->dev->dev_addr);
190 t1_set_rxmode(p->dev);
191 t1_link_start(p->phy, mac, &p->link_config);
192 mac->ops->enable(mac, MAC_DIRECTION_RX | MAC_DIRECTION_TX);
193 }
194
195 static void enable_hw_csum(struct adapter *adapter)
196 {
197 if (adapter->flags & TSO_CAPABLE)
198 t1_tp_set_ip_checksum_offload(adapter->tp, 1); /* for TSO only */
199 if (adapter->flags & UDP_CSUM_CAPABLE)
200 t1_tp_set_udp_checksum_offload(adapter->tp, 1);
201 t1_tp_set_tcp_checksum_offload(adapter->tp, 1);
202 }
203
204 /*
205 * Things to do upon first use of a card.
206 * This must run with the rtnl lock held.
207 */
208 static int cxgb_up(struct adapter *adapter)
209 {
210 int err = 0;
211
212 if (!(adapter->flags & FULL_INIT_DONE)) {
213 err = t1_init_hw_modules(adapter);
214 if (err)
215 goto out_err;
216
217 enable_hw_csum(adapter);
218 adapter->flags |= FULL_INIT_DONE;
219 }
220
221 t1_interrupts_clear(adapter);
222
223 adapter->params.has_msi = !disable_msi && !pci_enable_msi(adapter->pdev);
224 err = request_irq(adapter->pdev->irq, t1_interrupt,
225 adapter->params.has_msi ? 0 : IRQF_SHARED,
226 adapter->name, adapter);
227 if (err) {
228 if (adapter->params.has_msi)
229 pci_disable_msi(adapter->pdev);
230
231 goto out_err;
232 }
233
234 t1_sge_start(adapter->sge);
235 t1_interrupts_enable(adapter);
236 out_err:
237 return err;
238 }
239
240 /*
241 * Release resources when all the ports have been stopped.
242 */
243 static void cxgb_down(struct adapter *adapter)
244 {
245 t1_sge_stop(adapter->sge);
246 t1_interrupts_disable(adapter);
247 free_irq(adapter->pdev->irq, adapter);
248 if (adapter->params.has_msi)
249 pci_disable_msi(adapter->pdev);
250 }
251
252 static int cxgb_open(struct net_device *dev)
253 {
254 int err;
255 struct adapter *adapter = dev->priv;
256 int other_ports = adapter->open_device_map & PORT_MASK;
257
258 napi_enable(&adapter->napi);
259 if (!adapter->open_device_map && (err = cxgb_up(adapter)) < 0) {
260 napi_disable(&adapter->napi);
261 return err;
262 }
263
264 __set_bit(dev->if_port, &adapter->open_device_map);
265 link_start(&adapter->port[dev->if_port]);
266 netif_start_queue(dev);
267 if (!other_ports && adapter->params.stats_update_period)
268 schedule_mac_stats_update(adapter,
269 adapter->params.stats_update_period);
270 return 0;
271 }
272
273 static int cxgb_close(struct net_device *dev)
274 {
275 struct adapter *adapter = dev->priv;
276 struct port_info *p = &adapter->port[dev->if_port];
277 struct cmac *mac = p->mac;
278
279 netif_stop_queue(dev);
280 napi_disable(&adapter->napi);
281 mac->ops->disable(mac, MAC_DIRECTION_TX | MAC_DIRECTION_RX);
282 netif_carrier_off(dev);
283
284 clear_bit(dev->if_port, &adapter->open_device_map);
285 if (adapter->params.stats_update_period &&
286 !(adapter->open_device_map & PORT_MASK)) {
287 /* Stop statistics accumulation. */
288 smp_mb__after_clear_bit();
289 spin_lock(&adapter->work_lock); /* sync with update task */
290 spin_unlock(&adapter->work_lock);
291 cancel_mac_stats_update(adapter);
292 }
293
294 if (!adapter->open_device_map)
295 cxgb_down(adapter);
296 return 0;
297 }
298
299 static struct net_device_stats *t1_get_stats(struct net_device *dev)
300 {
301 struct adapter *adapter = dev->priv;
302 struct port_info *p = &adapter->port[dev->if_port];
303 struct net_device_stats *ns = &p->netstats;
304 const struct cmac_statistics *pstats;
305
306 /* Do a full update of the MAC stats */
307 pstats = p->mac->ops->statistics_update(p->mac,
308 MAC_STATS_UPDATE_FULL);
309
310 ns->tx_packets = pstats->TxUnicastFramesOK +
311 pstats->TxMulticastFramesOK + pstats->TxBroadcastFramesOK;
312
313 ns->rx_packets = pstats->RxUnicastFramesOK +
314 pstats->RxMulticastFramesOK + pstats->RxBroadcastFramesOK;
315
316 ns->tx_bytes = pstats->TxOctetsOK;
317 ns->rx_bytes = pstats->RxOctetsOK;
318
319 ns->tx_errors = pstats->TxLateCollisions + pstats->TxLengthErrors +
320 pstats->TxUnderrun + pstats->TxFramesAbortedDueToXSCollisions;
321 ns->rx_errors = pstats->RxDataErrors + pstats->RxJabberErrors +
322 pstats->RxFCSErrors + pstats->RxAlignErrors +
323 pstats->RxSequenceErrors + pstats->RxFrameTooLongErrors +
324 pstats->RxSymbolErrors + pstats->RxRuntErrors;
325
326 ns->multicast = pstats->RxMulticastFramesOK;
327 ns->collisions = pstats->TxTotalCollisions;
328
329 /* detailed rx_errors */
330 ns->rx_length_errors = pstats->RxFrameTooLongErrors +
331 pstats->RxJabberErrors;
332 ns->rx_over_errors = 0;
333 ns->rx_crc_errors = pstats->RxFCSErrors;
334 ns->rx_frame_errors = pstats->RxAlignErrors;
335 ns->rx_fifo_errors = 0;
336 ns->rx_missed_errors = 0;
337
338 /* detailed tx_errors */
339 ns->tx_aborted_errors = pstats->TxFramesAbortedDueToXSCollisions;
340 ns->tx_carrier_errors = 0;
341 ns->tx_fifo_errors = pstats->TxUnderrun;
342 ns->tx_heartbeat_errors = 0;
343 ns->tx_window_errors = pstats->TxLateCollisions;
344 return ns;
345 }
346
347 static u32 get_msglevel(struct net_device *dev)
348 {
349 struct adapter *adapter = dev->priv;
350
351 return adapter->msg_enable;
352 }
353
354 static void set_msglevel(struct net_device *dev, u32 val)
355 {
356 struct adapter *adapter = dev->priv;
357
358 adapter->msg_enable = val;
359 }
360
361 static char stats_strings[][ETH_GSTRING_LEN] = {
362 "TxOctetsOK",
363 "TxOctetsBad",
364 "TxUnicastFramesOK",
365 "TxMulticastFramesOK",
366 "TxBroadcastFramesOK",
367 "TxPauseFrames",
368 "TxFramesWithDeferredXmissions",
369 "TxLateCollisions",
370 "TxTotalCollisions",
371 "TxFramesAbortedDueToXSCollisions",
372 "TxUnderrun",
373 "TxLengthErrors",
374 "TxInternalMACXmitError",
375 "TxFramesWithExcessiveDeferral",
376 "TxFCSErrors",
377
378 "RxOctetsOK",
379 "RxOctetsBad",
380 "RxUnicastFramesOK",
381 "RxMulticastFramesOK",
382 "RxBroadcastFramesOK",
383 "RxPauseFrames",
384 "RxFCSErrors",
385 "RxAlignErrors",
386 "RxSymbolErrors",
387 "RxDataErrors",
388 "RxSequenceErrors",
389 "RxRuntErrors",
390 "RxJabberErrors",
391 "RxInternalMACRcvError",
392 "RxInRangeLengthErrors",
393 "RxOutOfRangeLengthField",
394 "RxFrameTooLongErrors",
395
396 /* Port stats */
397 "RxPackets",
398 "RxCsumGood",
399 "TxPackets",
400 "TxCsumOffload",
401 "TxTso",
402 "RxVlan",
403 "TxVlan",
404
405 /* Interrupt stats */
406 "rx drops",
407 "pure_rsps",
408 "unhandled irqs",
409 "respQ_empty",
410 "respQ_overflow",
411 "freelistQ_empty",
412 "pkt_too_big",
413 "pkt_mismatch",
414 "cmdQ_full0",
415 "cmdQ_full1",
416
417 "espi_DIP2ParityErr",
418 "espi_DIP4Err",
419 "espi_RxDrops",
420 "espi_TxDrops",
421 "espi_RxOvfl",
422 "espi_ParityErr"
423 };
424
425 #define T2_REGMAP_SIZE (3 * 1024)
426
427 static int get_regs_len(struct net_device *dev)
428 {
429 return T2_REGMAP_SIZE;
430 }
431
432 static void get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
433 {
434 struct adapter *adapter = dev->priv;
435
436 strcpy(info->driver, DRV_NAME);
437 strcpy(info->version, DRV_VERSION);
438 strcpy(info->fw_version, "N/A");
439 strcpy(info->bus_info, pci_name(adapter->pdev));
440 }
441
442 static int get_sset_count(struct net_device *dev, int sset)
443 {
444 switch (sset) {
445 case ETH_SS_STATS:
446 return ARRAY_SIZE(stats_strings);
447 default:
448 return -EOPNOTSUPP;
449 }
450 }
451
452 static void get_strings(struct net_device *dev, u32 stringset, u8 *data)
453 {
454 if (stringset == ETH_SS_STATS)
455 memcpy(data, stats_strings, sizeof(stats_strings));
456 }
457
458 static void get_stats(struct net_device *dev, struct ethtool_stats *stats,
459 u64 *data)
460 {
461 struct adapter *adapter = dev->priv;
462 struct cmac *mac = adapter->port[dev->if_port].mac;
463 const struct cmac_statistics *s;
464 const struct sge_intr_counts *t;
465 struct sge_port_stats ss;
466 unsigned int len;
467
468 s = mac->ops->statistics_update(mac, MAC_STATS_UPDATE_FULL);
469
470 len = sizeof(u64)*(&s->TxFCSErrors + 1 - &s->TxOctetsOK);
471 memcpy(data, &s->TxOctetsOK, len);
472 data += len;
473
474 len = sizeof(u64)*(&s->RxFrameTooLongErrors + 1 - &s->RxOctetsOK);
475 memcpy(data, &s->RxOctetsOK, len);
476 data += len;
477
478 t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss);
479 memcpy(data, &ss, sizeof(ss));
480 data += sizeof(ss);
481
482 t = t1_sge_get_intr_counts(adapter->sge);
483 *data++ = t->rx_drops;
484 *data++ = t->pure_rsps;
485 *data++ = t->unhandled_irqs;
486 *data++ = t->respQ_empty;
487 *data++ = t->respQ_overflow;
488 *data++ = t->freelistQ_empty;
489 *data++ = t->pkt_too_big;
490 *data++ = t->pkt_mismatch;
491 *data++ = t->cmdQ_full[0];
492 *data++ = t->cmdQ_full[1];
493
494 if (adapter->espi) {
495 const struct espi_intr_counts *e;
496
497 e = t1_espi_get_intr_counts(adapter->espi);
498 *data++ = e->DIP2_parity_err;
499 *data++ = e->DIP4_err;
500 *data++ = e->rx_drops;
501 *data++ = e->tx_drops;
502 *data++ = e->rx_ovflw;
503 *data++ = e->parity_err;
504 }
505 }
506
507 static inline void reg_block_dump(struct adapter *ap, void *buf,
508 unsigned int start, unsigned int end)
509 {
510 u32 *p = buf + start;
511
512 for ( ; start <= end; start += sizeof(u32))
513 *p++ = readl(ap->regs + start);
514 }
515
516 static void get_regs(struct net_device *dev, struct ethtool_regs *regs,
517 void *buf)
518 {
519 struct adapter *ap = dev->priv;
520
521 /*
522 * Version scheme: bits 0..9: chip version, bits 10..15: chip revision
523 */
524 regs->version = 2;
525
526 memset(buf, 0, T2_REGMAP_SIZE);
527 reg_block_dump(ap, buf, 0, A_SG_RESPACCUTIMER);
528 reg_block_dump(ap, buf, A_MC3_CFG, A_MC4_INT_CAUSE);
529 reg_block_dump(ap, buf, A_TPI_ADDR, A_TPI_PAR);
530 reg_block_dump(ap, buf, A_TP_IN_CONFIG, A_TP_TX_DROP_COUNT);
531 reg_block_dump(ap, buf, A_RAT_ROUTE_CONTROL, A_RAT_INTR_CAUSE);
532 reg_block_dump(ap, buf, A_CSPI_RX_AE_WM, A_CSPI_INTR_ENABLE);
533 reg_block_dump(ap, buf, A_ESPI_SCH_TOKEN0, A_ESPI_GOSTAT);
534 reg_block_dump(ap, buf, A_ULP_ULIMIT, A_ULP_PIO_CTRL);
535 reg_block_dump(ap, buf, A_PL_ENABLE, A_PL_CAUSE);
536 reg_block_dump(ap, buf, A_MC5_CONFIG, A_MC5_MASK_WRITE_CMD);
537 }
538
539 static int get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
540 {
541 struct adapter *adapter = dev->priv;
542 struct port_info *p = &adapter->port[dev->if_port];
543
544 cmd->supported = p->link_config.supported;
545 cmd->advertising = p->link_config.advertising;
546
547 if (netif_carrier_ok(dev)) {
548 cmd->speed = p->link_config.speed;
549 cmd->duplex = p->link_config.duplex;
550 } else {
551 cmd->speed = -1;
552 cmd->duplex = -1;
553 }
554
555 cmd->port = (cmd->supported & SUPPORTED_TP) ? PORT_TP : PORT_FIBRE;
556 cmd->phy_address = p->phy->addr;
557 cmd->transceiver = XCVR_EXTERNAL;
558 cmd->autoneg = p->link_config.autoneg;
559 cmd->maxtxpkt = 0;
560 cmd->maxrxpkt = 0;
561 return 0;
562 }
563
564 static int speed_duplex_to_caps(int speed, int duplex)
565 {
566 int cap = 0;
567
568 switch (speed) {
569 case SPEED_10:
570 if (duplex == DUPLEX_FULL)
571 cap = SUPPORTED_10baseT_Full;
572 else
573 cap = SUPPORTED_10baseT_Half;
574 break;
575 case SPEED_100:
576 if (duplex == DUPLEX_FULL)
577 cap = SUPPORTED_100baseT_Full;
578 else
579 cap = SUPPORTED_100baseT_Half;
580 break;
581 case SPEED_1000:
582 if (duplex == DUPLEX_FULL)
583 cap = SUPPORTED_1000baseT_Full;
584 else
585 cap = SUPPORTED_1000baseT_Half;
586 break;
587 case SPEED_10000:
588 if (duplex == DUPLEX_FULL)
589 cap = SUPPORTED_10000baseT_Full;
590 }
591 return cap;
592 }
593
594 #define ADVERTISED_MASK (ADVERTISED_10baseT_Half | ADVERTISED_10baseT_Full | \
595 ADVERTISED_100baseT_Half | ADVERTISED_100baseT_Full | \
596 ADVERTISED_1000baseT_Half | ADVERTISED_1000baseT_Full | \
597 ADVERTISED_10000baseT_Full)
598
599 static int set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
600 {
601 struct adapter *adapter = dev->priv;
602 struct port_info *p = &adapter->port[dev->if_port];
603 struct link_config *lc = &p->link_config;
604
605 if (!(lc->supported & SUPPORTED_Autoneg))
606 return -EOPNOTSUPP; /* can't change speed/duplex */
607
608 if (cmd->autoneg == AUTONEG_DISABLE) {
609 int cap = speed_duplex_to_caps(cmd->speed, cmd->duplex);
610
611 if (!(lc->supported & cap) || cmd->speed == SPEED_1000)
612 return -EINVAL;
613 lc->requested_speed = cmd->speed;
614 lc->requested_duplex = cmd->duplex;
615 lc->advertising = 0;
616 } else {
617 cmd->advertising &= ADVERTISED_MASK;
618 if (cmd->advertising & (cmd->advertising - 1))
619 cmd->advertising = lc->supported;
620 cmd->advertising &= lc->supported;
621 if (!cmd->advertising)
622 return -EINVAL;
623 lc->requested_speed = SPEED_INVALID;
624 lc->requested_duplex = DUPLEX_INVALID;
625 lc->advertising = cmd->advertising | ADVERTISED_Autoneg;
626 }
627 lc->autoneg = cmd->autoneg;
628 if (netif_running(dev))
629 t1_link_start(p->phy, p->mac, lc);
630 return 0;
631 }
632
633 static void get_pauseparam(struct net_device *dev,
634 struct ethtool_pauseparam *epause)
635 {
636 struct adapter *adapter = dev->priv;
637 struct port_info *p = &adapter->port[dev->if_port];
638
639 epause->autoneg = (p->link_config.requested_fc & PAUSE_AUTONEG) != 0;
640 epause->rx_pause = (p->link_config.fc & PAUSE_RX) != 0;
641 epause->tx_pause = (p->link_config.fc & PAUSE_TX) != 0;
642 }
643
644 static int set_pauseparam(struct net_device *dev,
645 struct ethtool_pauseparam *epause)
646 {
647 struct adapter *adapter = dev->priv;
648 struct port_info *p = &adapter->port[dev->if_port];
649 struct link_config *lc = &p->link_config;
650
651 if (epause->autoneg == AUTONEG_DISABLE)
652 lc->requested_fc = 0;
653 else if (lc->supported & SUPPORTED_Autoneg)
654 lc->requested_fc = PAUSE_AUTONEG;
655 else
656 return -EINVAL;
657
658 if (epause->rx_pause)
659 lc->requested_fc |= PAUSE_RX;
660 if (epause->tx_pause)
661 lc->requested_fc |= PAUSE_TX;
662 if (lc->autoneg == AUTONEG_ENABLE) {
663 if (netif_running(dev))
664 t1_link_start(p->phy, p->mac, lc);
665 } else {
666 lc->fc = lc->requested_fc & (PAUSE_RX | PAUSE_TX);
667 if (netif_running(dev))
668 p->mac->ops->set_speed_duplex_fc(p->mac, -1, -1,
669 lc->fc);
670 }
671 return 0;
672 }
673
674 static u32 get_rx_csum(struct net_device *dev)
675 {
676 struct adapter *adapter = dev->priv;
677
678 return (adapter->flags & RX_CSUM_ENABLED) != 0;
679 }
680
681 static int set_rx_csum(struct net_device *dev, u32 data)
682 {
683 struct adapter *adapter = dev->priv;
684
685 if (data)
686 adapter->flags |= RX_CSUM_ENABLED;
687 else
688 adapter->flags &= ~RX_CSUM_ENABLED;
689 return 0;
690 }
691
692 static int set_tso(struct net_device *dev, u32 value)
693 {
694 struct adapter *adapter = dev->priv;
695
696 if (!(adapter->flags & TSO_CAPABLE))
697 return value ? -EOPNOTSUPP : 0;
698 return ethtool_op_set_tso(dev, value);
699 }
700
701 static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e)
702 {
703 struct adapter *adapter = dev->priv;
704 int jumbo_fl = t1_is_T1B(adapter) ? 1 : 0;
705
706 e->rx_max_pending = MAX_RX_BUFFERS;
707 e->rx_mini_max_pending = 0;
708 e->rx_jumbo_max_pending = MAX_RX_JUMBO_BUFFERS;
709 e->tx_max_pending = MAX_CMDQ_ENTRIES;
710
711 e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl];
712 e->rx_mini_pending = 0;
713 e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl];
714 e->tx_pending = adapter->params.sge.cmdQ_size[0];
715 }
716
717 static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e)
718 {
719 struct adapter *adapter = dev->priv;
720 int jumbo_fl = t1_is_T1B(adapter) ? 1 : 0;
721
722 if (e->rx_pending > MAX_RX_BUFFERS || e->rx_mini_pending ||
723 e->rx_jumbo_pending > MAX_RX_JUMBO_BUFFERS ||
724 e->tx_pending > MAX_CMDQ_ENTRIES ||
725 e->rx_pending < MIN_FL_ENTRIES ||
726 e->rx_jumbo_pending < MIN_FL_ENTRIES ||
727 e->tx_pending < (adapter->params.nports + 1) * (MAX_SKB_FRAGS + 1))
728 return -EINVAL;
729
730 if (adapter->flags & FULL_INIT_DONE)
731 return -EBUSY;
732
733 adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending;
734 adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending;
735 adapter->params.sge.cmdQ_size[0] = e->tx_pending;
736 adapter->params.sge.cmdQ_size[1] = e->tx_pending > MAX_CMDQ1_ENTRIES ?
737 MAX_CMDQ1_ENTRIES : e->tx_pending;
738 return 0;
739 }
740
741 static int set_coalesce(struct net_device *dev, struct ethtool_coalesce *c)
742 {
743 struct adapter *adapter = dev->priv;
744
745 adapter->params.sge.rx_coalesce_usecs = c->rx_coalesce_usecs;
746 adapter->params.sge.coalesce_enable = c->use_adaptive_rx_coalesce;
747 adapter->params.sge.sample_interval_usecs = c->rate_sample_interval;
748 t1_sge_set_coalesce_params(adapter->sge, &adapter->params.sge);
749 return 0;
750 }
751
752 static int get_coalesce(struct net_device *dev, struct ethtool_coalesce *c)
753 {
754 struct adapter *adapter = dev->priv;
755
756 c->rx_coalesce_usecs = adapter->params.sge.rx_coalesce_usecs;
757 c->rate_sample_interval = adapter->params.sge.sample_interval_usecs;
758 c->use_adaptive_rx_coalesce = adapter->params.sge.coalesce_enable;
759 return 0;
760 }
761
762 static int get_eeprom_len(struct net_device *dev)
763 {
764 struct adapter *adapter = dev->priv;
765
766 return t1_is_asic(adapter) ? EEPROM_SIZE : 0;
767 }
768
769 #define EEPROM_MAGIC(ap) \
770 (PCI_VENDOR_ID_CHELSIO | ((ap)->params.chip_version << 16))
771
772 static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e,
773 u8 *data)
774 {
775 int i;
776 u8 buf[EEPROM_SIZE] __attribute__((aligned(4)));
777 struct adapter *adapter = dev->priv;
778
779 e->magic = EEPROM_MAGIC(adapter);
780 for (i = e->offset & ~3; i < e->offset + e->len; i += sizeof(u32))
781 t1_seeprom_read(adapter, i, (u32 *)&buf[i]);
782 memcpy(data, buf + e->offset, e->len);
783 return 0;
784 }
785
786 static const struct ethtool_ops t1_ethtool_ops = {
787 .get_settings = get_settings,
788 .set_settings = set_settings,
789 .get_drvinfo = get_drvinfo,
790 .get_msglevel = get_msglevel,
791 .set_msglevel = set_msglevel,
792 .get_ringparam = get_sge_param,
793 .set_ringparam = set_sge_param,
794 .get_coalesce = get_coalesce,
795 .set_coalesce = set_coalesce,
796 .get_eeprom_len = get_eeprom_len,
797 .get_eeprom = get_eeprom,
798 .get_pauseparam = get_pauseparam,
799 .set_pauseparam = set_pauseparam,
800 .get_rx_csum = get_rx_csum,
801 .set_rx_csum = set_rx_csum,
802 .set_tx_csum = ethtool_op_set_tx_csum,
803 .set_sg = ethtool_op_set_sg,
804 .get_link = ethtool_op_get_link,
805 .get_strings = get_strings,
806 .get_sset_count = get_sset_count,
807 .get_ethtool_stats = get_stats,
808 .get_regs_len = get_regs_len,
809 .get_regs = get_regs,
810 .set_tso = set_tso,
811 };
812
813 static int t1_ioctl(struct net_device *dev, struct ifreq *req, int cmd)
814 {
815 struct adapter *adapter = dev->priv;
816 struct mii_ioctl_data *data = if_mii(req);
817
818 switch (cmd) {
819 case SIOCGMIIPHY:
820 data->phy_id = adapter->port[dev->if_port].phy->addr;
821 /* FALLTHRU */
822 case SIOCGMIIREG: {
823 struct cphy *phy = adapter->port[dev->if_port].phy;
824 u32 val;
825
826 if (!phy->mdio_read)
827 return -EOPNOTSUPP;
828 phy->mdio_read(adapter, data->phy_id, 0, data->reg_num & 0x1f,
829 &val);
830 data->val_out = val;
831 break;
832 }
833 case SIOCSMIIREG: {
834 struct cphy *phy = adapter->port[dev->if_port].phy;
835
836 if (!capable(CAP_NET_ADMIN))
837 return -EPERM;
838 if (!phy->mdio_write)
839 return -EOPNOTSUPP;
840 phy->mdio_write(adapter, data->phy_id, 0, data->reg_num & 0x1f,
841 data->val_in);
842 break;
843 }
844
845 default:
846 return -EOPNOTSUPP;
847 }
848 return 0;
849 }
850
851 static int t1_change_mtu(struct net_device *dev, int new_mtu)
852 {
853 int ret;
854 struct adapter *adapter = dev->priv;
855 struct cmac *mac = adapter->port[dev->if_port].mac;
856
857 if (!mac->ops->set_mtu)
858 return -EOPNOTSUPP;
859 if (new_mtu < 68)
860 return -EINVAL;
861 if ((ret = mac->ops->set_mtu(mac, new_mtu)))
862 return ret;
863 dev->mtu = new_mtu;
864 return 0;
865 }
866
867 static int t1_set_mac_addr(struct net_device *dev, void *p)
868 {
869 struct adapter *adapter = dev->priv;
870 struct cmac *mac = adapter->port[dev->if_port].mac;
871 struct sockaddr *addr = p;
872
873 if (!mac->ops->macaddress_set)
874 return -EOPNOTSUPP;
875
876 memcpy(dev->dev_addr, addr->sa_data, dev->addr_len);
877 mac->ops->macaddress_set(mac, dev->dev_addr);
878 return 0;
879 }
880
881 #if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE)
882 static void vlan_rx_register(struct net_device *dev,
883 struct vlan_group *grp)
884 {
885 struct adapter *adapter = dev->priv;
886
887 spin_lock_irq(&adapter->async_lock);
888 adapter->vlan_grp = grp;
889 t1_set_vlan_accel(adapter, grp != NULL);
890 spin_unlock_irq(&adapter->async_lock);
891 }
892 #endif
893
894 #ifdef CONFIG_NET_POLL_CONTROLLER
895 static void t1_netpoll(struct net_device *dev)
896 {
897 unsigned long flags;
898 struct adapter *adapter = dev->priv;
899
900 local_irq_save(flags);
901 t1_interrupt(adapter->pdev->irq, adapter);
902 local_irq_restore(flags);
903 }
904 #endif
905
906 /*
907 * Periodic accumulation of MAC statistics. This is used only if the MAC
908 * does not have any other way to prevent stats counter overflow.
909 */
910 static void mac_stats_task(struct work_struct *work)
911 {
912 int i;
913 struct adapter *adapter =
914 container_of(work, struct adapter, stats_update_task.work);
915
916 for_each_port(adapter, i) {
917 struct port_info *p = &adapter->port[i];
918
919 if (netif_running(p->dev))
920 p->mac->ops->statistics_update(p->mac,
921 MAC_STATS_UPDATE_FAST);
922 }
923
924 /* Schedule the next statistics update if any port is active. */
925 spin_lock(&adapter->work_lock);
926 if (adapter->open_device_map & PORT_MASK)
927 schedule_mac_stats_update(adapter,
928 adapter->params.stats_update_period);
929 spin_unlock(&adapter->work_lock);
930 }
931
932 /*
933 * Processes elmer0 external interrupts in process context.
934 */
935 static void ext_intr_task(struct work_struct *work)
936 {
937 struct adapter *adapter =
938 container_of(work, struct adapter, ext_intr_handler_task);
939
940 t1_elmer0_ext_intr_handler(adapter);
941
942 /* Now reenable external interrupts */
943 spin_lock_irq(&adapter->async_lock);
944 adapter->slow_intr_mask |= F_PL_INTR_EXT;
945 writel(F_PL_INTR_EXT, adapter->regs + A_PL_CAUSE);
946 writel(adapter->slow_intr_mask | F_PL_INTR_SGE_DATA,
947 adapter->regs + A_PL_ENABLE);
948 spin_unlock_irq(&adapter->async_lock);
949 }
950
951 /*
952 * Interrupt-context handler for elmer0 external interrupts.
953 */
954 void t1_elmer0_ext_intr(struct adapter *adapter)
955 {
956 /*
957 * Schedule a task to handle external interrupts as we require
958 * a process context. We disable EXT interrupts in the interim
959 * and let the task reenable them when it's done.
960 */
961 adapter->slow_intr_mask &= ~F_PL_INTR_EXT;
962 writel(adapter->slow_intr_mask | F_PL_INTR_SGE_DATA,
963 adapter->regs + A_PL_ENABLE);
964 schedule_work(&adapter->ext_intr_handler_task);
965 }
966
967 void t1_fatal_err(struct adapter *adapter)
968 {
969 if (adapter->flags & FULL_INIT_DONE) {
970 t1_sge_stop(adapter->sge);
971 t1_interrupts_disable(adapter);
972 }
973 CH_ALERT("%s: encountered fatal error, operation suspended\n",
974 adapter->name);
975 }
976
977 static int __devinit init_one(struct pci_dev *pdev,
978 const struct pci_device_id *ent)
979 {
980 static int version_printed;
981
982 int i, err, pci_using_dac = 0;
983 unsigned long mmio_start, mmio_len;
984 const struct board_info *bi;
985 struct adapter *adapter = NULL;
986 struct port_info *pi;
987
988 if (!version_printed) {
989 printk(KERN_INFO "%s - version %s\n", DRV_DESCRIPTION,
990 DRV_VERSION);
991 ++version_printed;
992 }
993
994 err = pci_enable_device(pdev);
995 if (err)
996 return err;
997
998 if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) {
999 CH_ERR("%s: cannot find PCI device memory base address\n",
1000 pci_name(pdev));
1001 err = -ENODEV;
1002 goto out_disable_pdev;
1003 }
1004
1005 if (!pci_set_dma_mask(pdev, DMA_64BIT_MASK)) {
1006 pci_using_dac = 1;
1007
1008 if (pci_set_consistent_dma_mask(pdev, DMA_64BIT_MASK)) {
1009 CH_ERR("%s: unable to obtain 64-bit DMA for"
1010 "consistent allocations\n", pci_name(pdev));
1011 err = -ENODEV;
1012 goto out_disable_pdev;
1013 }
1014
1015 } else if ((err = pci_set_dma_mask(pdev, DMA_32BIT_MASK)) != 0) {
1016 CH_ERR("%s: no usable DMA configuration\n", pci_name(pdev));
1017 goto out_disable_pdev;
1018 }
1019
1020 err = pci_request_regions(pdev, DRV_NAME);
1021 if (err) {
1022 CH_ERR("%s: cannot obtain PCI resources\n", pci_name(pdev));
1023 goto out_disable_pdev;
1024 }
1025
1026 pci_set_master(pdev);
1027
1028 mmio_start = pci_resource_start(pdev, 0);
1029 mmio_len = pci_resource_len(pdev, 0);
1030 bi = t1_get_board_info(ent->driver_data);
1031
1032 for (i = 0; i < bi->port_number; ++i) {
1033 struct net_device *netdev;
1034
1035 netdev = alloc_etherdev(adapter ? 0 : sizeof(*adapter));
1036 if (!netdev) {
1037 err = -ENOMEM;
1038 goto out_free_dev;
1039 }
1040
1041 SET_NETDEV_DEV(netdev, &pdev->dev);
1042
1043 if (!adapter) {
1044 adapter = netdev->priv;
1045 adapter->pdev = pdev;
1046 adapter->port[0].dev = netdev; /* so we don't leak it */
1047
1048 adapter->regs = ioremap(mmio_start, mmio_len);
1049 if (!adapter->regs) {
1050 CH_ERR("%s: cannot map device registers\n",
1051 pci_name(pdev));
1052 err = -ENOMEM;
1053 goto out_free_dev;
1054 }
1055
1056 if (t1_get_board_rev(adapter, bi, &adapter->params)) {
1057 err = -ENODEV; /* Can't handle this chip rev */
1058 goto out_free_dev;
1059 }
1060
1061 adapter->name = pci_name(pdev);
1062 adapter->msg_enable = dflt_msg_enable;
1063 adapter->mmio_len = mmio_len;
1064
1065 spin_lock_init(&adapter->tpi_lock);
1066 spin_lock_init(&adapter->work_lock);
1067 spin_lock_init(&adapter->async_lock);
1068 spin_lock_init(&adapter->mac_lock);
1069
1070 INIT_WORK(&adapter->ext_intr_handler_task,
1071 ext_intr_task);
1072 INIT_DELAYED_WORK(&adapter->stats_update_task,
1073 mac_stats_task);
1074
1075 pci_set_drvdata(pdev, netdev);
1076 }
1077
1078 pi = &adapter->port[i];
1079 pi->dev = netdev;
1080 netif_carrier_off(netdev);
1081 netdev->irq = pdev->irq;
1082 netdev->if_port = i;
1083 netdev->mem_start = mmio_start;
1084 netdev->mem_end = mmio_start + mmio_len - 1;
1085 netdev->priv = adapter;
1086 netdev->features |= NETIF_F_SG | NETIF_F_IP_CSUM;
1087 netdev->features |= NETIF_F_LLTX;
1088
1089 adapter->flags |= RX_CSUM_ENABLED | TCP_CSUM_CAPABLE;
1090 if (pci_using_dac)
1091 netdev->features |= NETIF_F_HIGHDMA;
1092 if (vlan_tso_capable(adapter)) {
1093 #if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE)
1094 adapter->flags |= VLAN_ACCEL_CAPABLE;
1095 netdev->features |=
1096 NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX;
1097 netdev->vlan_rx_register = vlan_rx_register;
1098 #endif
1099
1100 /* T204: disable TSO */
1101 if (!(is_T2(adapter)) || bi->port_number != 4) {
1102 adapter->flags |= TSO_CAPABLE;
1103 netdev->features |= NETIF_F_TSO;
1104 }
1105 }
1106
1107 netdev->open = cxgb_open;
1108 netdev->stop = cxgb_close;
1109 netdev->hard_start_xmit = t1_start_xmit;
1110 netdev->hard_header_len += (adapter->flags & TSO_CAPABLE) ?
1111 sizeof(struct cpl_tx_pkt_lso) : sizeof(struct cpl_tx_pkt);
1112 netdev->get_stats = t1_get_stats;
1113 netdev->set_multicast_list = t1_set_rxmode;
1114 netdev->do_ioctl = t1_ioctl;
1115 netdev->change_mtu = t1_change_mtu;
1116 netdev->set_mac_address = t1_set_mac_addr;
1117 #ifdef CONFIG_NET_POLL_CONTROLLER
1118 netdev->poll_controller = t1_netpoll;
1119 #endif
1120 #ifdef CONFIG_CHELSIO_T1_NAPI
1121 netif_napi_add(netdev, &adapter->napi, t1_poll, 64);
1122 #endif
1123
1124 SET_ETHTOOL_OPS(netdev, &t1_ethtool_ops);
1125 }
1126
1127 if (t1_init_sw_modules(adapter, bi) < 0) {
1128 err = -ENODEV;
1129 goto out_free_dev;
1130 }
1131
1132 /*
1133 * The card is now ready to go. If any errors occur during device
1134 * registration we do not fail the whole card but rather proceed only
1135 * with the ports we manage to register successfully. However we must
1136 * register at least one net device.
1137 */
1138 for (i = 0; i < bi->port_number; ++i) {
1139 err = register_netdev(adapter->port[i].dev);
1140 if (err)
1141 CH_WARN("%s: cannot register net device %s, skipping\n",
1142 pci_name(pdev), adapter->port[i].dev->name);
1143 else {
1144 /*
1145 * Change the name we use for messages to the name of
1146 * the first successfully registered interface.
1147 */
1148 if (!adapter->registered_device_map)
1149 adapter->name = adapter->port[i].dev->name;
1150
1151 __set_bit(i, &adapter->registered_device_map);
1152 }
1153 }
1154 if (!adapter->registered_device_map) {
1155 CH_ERR("%s: could not register any net devices\n",
1156 pci_name(pdev));
1157 goto out_release_adapter_res;
1158 }
1159
1160 printk(KERN_INFO "%s: %s (rev %d), %s %dMHz/%d-bit\n", adapter->name,
1161 bi->desc, adapter->params.chip_revision,
1162 adapter->params.pci.is_pcix ? "PCIX" : "PCI",
1163 adapter->params.pci.speed, adapter->params.pci.width);
1164
1165 /*
1166 * Set the T1B ASIC and memory clocks.
1167 */
1168 if (t1powersave)
1169 adapter->t1powersave = LCLOCK; /* HW default is powersave mode. */
1170 else
1171 adapter->t1powersave = HCLOCK;
1172 if (t1_is_T1B(adapter))
1173 t1_clock(adapter, t1powersave);
1174
1175 return 0;
1176
1177 out_release_adapter_res:
1178 t1_free_sw_modules(adapter);
1179 out_free_dev:
1180 if (adapter) {
1181 if (adapter->regs)
1182 iounmap(adapter->regs);
1183 for (i = bi->port_number - 1; i >= 0; --i)
1184 if (adapter->port[i].dev)
1185 free_netdev(adapter->port[i].dev);
1186 }
1187 pci_release_regions(pdev);
1188 out_disable_pdev:
1189 pci_disable_device(pdev);
1190 pci_set_drvdata(pdev, NULL);
1191 return err;
1192 }
1193
1194 static void bit_bang(struct adapter *adapter, int bitdata, int nbits)
1195 {
1196 int data;
1197 int i;
1198 u32 val;
1199
1200 enum {
1201 S_CLOCK = 1 << 3,
1202 S_DATA = 1 << 4
1203 };
1204
1205 for (i = (nbits - 1); i > -1; i--) {
1206
1207 udelay(50);
1208
1209 data = ((bitdata >> i) & 0x1);
1210 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1211
1212 if (data)
1213 val |= S_DATA;
1214 else
1215 val &= ~S_DATA;
1216
1217 udelay(50);
1218
1219 /* Set SCLOCK low */
1220 val &= ~S_CLOCK;
1221 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1222
1223 udelay(50);
1224
1225 /* Write SCLOCK high */
1226 val |= S_CLOCK;
1227 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1228
1229 }
1230 }
1231
1232 static int t1_clock(struct adapter *adapter, int mode)
1233 {
1234 u32 val;
1235 int M_CORE_VAL;
1236 int M_MEM_VAL;
1237
1238 enum {
1239 M_CORE_BITS = 9,
1240 T_CORE_VAL = 0,
1241 T_CORE_BITS = 2,
1242 N_CORE_VAL = 0,
1243 N_CORE_BITS = 2,
1244 M_MEM_BITS = 9,
1245 T_MEM_VAL = 0,
1246 T_MEM_BITS = 2,
1247 N_MEM_VAL = 0,
1248 N_MEM_BITS = 2,
1249 NP_LOAD = 1 << 17,
1250 S_LOAD_MEM = 1 << 5,
1251 S_LOAD_CORE = 1 << 6,
1252 S_CLOCK = 1 << 3
1253 };
1254
1255 if (!t1_is_T1B(adapter))
1256 return -ENODEV; /* Can't re-clock this chip. */
1257
1258 if (mode & 2)
1259 return 0; /* show current mode. */
1260
1261 if ((adapter->t1powersave & 1) == (mode & 1))
1262 return -EALREADY; /* ASIC already running in mode. */
1263
1264 if ((mode & 1) == HCLOCK) {
1265 M_CORE_VAL = 0x14;
1266 M_MEM_VAL = 0x18;
1267 adapter->t1powersave = HCLOCK; /* overclock */
1268 } else {
1269 M_CORE_VAL = 0xe;
1270 M_MEM_VAL = 0x10;
1271 adapter->t1powersave = LCLOCK; /* underclock */
1272 }
1273
1274 /* Don't interrupt this serial stream! */
1275 spin_lock(&adapter->tpi_lock);
1276
1277 /* Initialize for ASIC core */
1278 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1279 val |= NP_LOAD;
1280 udelay(50);
1281 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1282 udelay(50);
1283 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1284 val &= ~S_LOAD_CORE;
1285 val &= ~S_CLOCK;
1286 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1287 udelay(50);
1288
1289 /* Serial program the ASIC clock synthesizer */
1290 bit_bang(adapter, T_CORE_VAL, T_CORE_BITS);
1291 bit_bang(adapter, N_CORE_VAL, N_CORE_BITS);
1292 bit_bang(adapter, M_CORE_VAL, M_CORE_BITS);
1293 udelay(50);
1294
1295 /* Finish ASIC core */
1296 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1297 val |= S_LOAD_CORE;
1298 udelay(50);
1299 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1300 udelay(50);
1301 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1302 val &= ~S_LOAD_CORE;
1303 udelay(50);
1304 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1305 udelay(50);
1306
1307 /* Initialize for memory */
1308 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1309 val |= NP_LOAD;
1310 udelay(50);
1311 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1312 udelay(50);
1313 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1314 val &= ~S_LOAD_MEM;
1315 val &= ~S_CLOCK;
1316 udelay(50);
1317 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1318 udelay(50);
1319
1320 /* Serial program the memory clock synthesizer */
1321 bit_bang(adapter, T_MEM_VAL, T_MEM_BITS);
1322 bit_bang(adapter, N_MEM_VAL, N_MEM_BITS);
1323 bit_bang(adapter, M_MEM_VAL, M_MEM_BITS);
1324 udelay(50);
1325
1326 /* Finish memory */
1327 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1328 val |= S_LOAD_MEM;
1329 udelay(50);
1330 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1331 udelay(50);
1332 __t1_tpi_read(adapter, A_ELMER0_GPO, &val);
1333 val &= ~S_LOAD_MEM;
1334 udelay(50);
1335 __t1_tpi_write(adapter, A_ELMER0_GPO, val);
1336
1337 spin_unlock(&adapter->tpi_lock);
1338
1339 return 0;
1340 }
1341
1342 static inline void t1_sw_reset(struct pci_dev *pdev)
1343 {
1344 pci_write_config_dword(pdev, A_PCICFG_PM_CSR, 3);
1345 pci_write_config_dword(pdev, A_PCICFG_PM_CSR, 0);
1346 }
1347
1348 static void __devexit remove_one(struct pci_dev *pdev)
1349 {
1350 struct net_device *dev = pci_get_drvdata(pdev);
1351 struct adapter *adapter = dev->priv;
1352 int i;
1353
1354 for_each_port(adapter, i) {
1355 if (test_bit(i, &adapter->registered_device_map))
1356 unregister_netdev(adapter->port[i].dev);
1357 }
1358
1359 t1_free_sw_modules(adapter);
1360 iounmap(adapter->regs);
1361
1362 while (--i >= 0) {
1363 if (adapter->port[i].dev)
1364 free_netdev(adapter->port[i].dev);
1365 }
1366
1367 pci_release_regions(pdev);
1368 pci_disable_device(pdev);
1369 pci_set_drvdata(pdev, NULL);
1370 t1_sw_reset(pdev);
1371 }
1372
1373 static struct pci_driver driver = {
1374 .name = DRV_NAME,
1375 .id_table = t1_pci_tbl,
1376 .probe = init_one,
1377 .remove = __devexit_p(remove_one),
1378 };
1379
1380 static int __init t1_init_module(void)
1381 {
1382 return pci_register_driver(&driver);
1383 }
1384
1385 static void __exit t1_cleanup_module(void)
1386 {
1387 pci_unregister_driver(&driver);
1388 }
1389
1390 module_init(t1_init_module);
1391 module_exit(t1_cleanup_module);