]> bbs.cooldavid.org Git - net-next-2.6.git/blame - drivers/net/tg3.c
[AF_PACKET]: Convert PACKET_MMAP over to vm_insert_page().
[net-next-2.6.git] / drivers / net / tg3.c
CommitLineData
1da177e4
LT
1/*
2 * tg3.c: Broadcom Tigon3 ethernet driver.
3 *
4 * Copyright (C) 2001, 2002, 2003, 2004 David S. Miller (davem@redhat.com)
5 * Copyright (C) 2001, 2002, 2003 Jeff Garzik (jgarzik@pobox.com)
6 * Copyright (C) 2004 Sun Microsystems Inc.
7 * Copyright (C) 2005 Broadcom Corporation.
8 *
9 * Firmware is:
49cabf49
MC
10 * Derived from proprietary unpublished source code,
11 * Copyright (C) 2000-2003 Broadcom Corporation.
12 *
13 * Permission is hereby granted for the distribution of this firmware
14 * data in hexadecimal or equivalent format, provided this copyright
15 * notice is accompanying it.
1da177e4
LT
16 */
17
18#include <linux/config.h>
19
20#include <linux/module.h>
21#include <linux/moduleparam.h>
22#include <linux/kernel.h>
23#include <linux/types.h>
24#include <linux/compiler.h>
25#include <linux/slab.h>
26#include <linux/delay.h>
27#include <linux/init.h>
28#include <linux/ioport.h>
29#include <linux/pci.h>
30#include <linux/netdevice.h>
31#include <linux/etherdevice.h>
32#include <linux/skbuff.h>
33#include <linux/ethtool.h>
34#include <linux/mii.h>
35#include <linux/if_vlan.h>
36#include <linux/ip.h>
37#include <linux/tcp.h>
38#include <linux/workqueue.h>
61487480 39#include <linux/prefetch.h>
f9a5f7d3 40#include <linux/dma-mapping.h>
1da177e4
LT
41
42#include <net/checksum.h>
43
44#include <asm/system.h>
45#include <asm/io.h>
46#include <asm/byteorder.h>
47#include <asm/uaccess.h>
48
49#ifdef CONFIG_SPARC64
50#include <asm/idprom.h>
51#include <asm/oplib.h>
52#include <asm/pbm.h>
53#endif
54
55#if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE)
56#define TG3_VLAN_TAG_USED 1
57#else
58#define TG3_VLAN_TAG_USED 0
59#endif
60
61#ifdef NETIF_F_TSO
62#define TG3_TSO_SUPPORT 1
63#else
64#define TG3_TSO_SUPPORT 0
65#endif
66
67#include "tg3.h"
68
69#define DRV_MODULE_NAME "tg3"
70#define PFX DRV_MODULE_NAME ": "
37ff238d
MC
71#define DRV_MODULE_VERSION "3.43"
72#define DRV_MODULE_RELDATE "Oct 24, 2005"
1da177e4
LT
73
74#define TG3_DEF_MAC_MODE 0
75#define TG3_DEF_RX_MODE 0
76#define TG3_DEF_TX_MODE 0
77#define TG3_DEF_MSG_ENABLE \
78 (NETIF_MSG_DRV | \
79 NETIF_MSG_PROBE | \
80 NETIF_MSG_LINK | \
81 NETIF_MSG_TIMER | \
82 NETIF_MSG_IFDOWN | \
83 NETIF_MSG_IFUP | \
84 NETIF_MSG_RX_ERR | \
85 NETIF_MSG_TX_ERR)
86
87/* length of time before we decide the hardware is borked,
88 * and dev->tx_timeout() should be called to fix the problem
89 */
90#define TG3_TX_TIMEOUT (5 * HZ)
91
92/* hardware minimum and maximum for a single frame's data payload */
93#define TG3_MIN_MTU 60
94#define TG3_MAX_MTU(tp) \
0f893dc6 95 ((tp->tg3_flags2 & TG3_FLG2_JUMBO_CAPABLE) ? 9000 : 1500)
1da177e4
LT
96
97/* These numbers seem to be hard coded in the NIC firmware somehow.
98 * You can't change the ring sizes, but you can change where you place
99 * them in the NIC onboard memory.
100 */
101#define TG3_RX_RING_SIZE 512
102#define TG3_DEF_RX_RING_PENDING 200
103#define TG3_RX_JUMBO_RING_SIZE 256
104#define TG3_DEF_RX_JUMBO_RING_PENDING 100
105
106/* Do not place this n-ring entries value into the tp struct itself,
107 * we really want to expose these constants to GCC so that modulo et
108 * al. operations are done with shifts and masks instead of with
109 * hw multiply/modulo instructions. Another solution would be to
110 * replace things like '% foo' with '& (foo - 1)'.
111 */
112#define TG3_RX_RCB_RING_SIZE(tp) \
113 ((tp->tg3_flags2 & TG3_FLG2_5705_PLUS) ? 512 : 1024)
114
115#define TG3_TX_RING_SIZE 512
116#define TG3_DEF_TX_RING_PENDING (TG3_TX_RING_SIZE - 1)
117
118#define TG3_RX_RING_BYTES (sizeof(struct tg3_rx_buffer_desc) * \
119 TG3_RX_RING_SIZE)
120#define TG3_RX_JUMBO_RING_BYTES (sizeof(struct tg3_rx_buffer_desc) * \
121 TG3_RX_JUMBO_RING_SIZE)
122#define TG3_RX_RCB_RING_BYTES(tp) (sizeof(struct tg3_rx_buffer_desc) * \
123 TG3_RX_RCB_RING_SIZE(tp))
124#define TG3_TX_RING_BYTES (sizeof(struct tg3_tx_buffer_desc) * \
125 TG3_TX_RING_SIZE)
1da177e4 126#define TX_BUFFS_AVAIL(TP) \
51b91468
MC
127 ((TP)->tx_pending - \
128 (((TP)->tx_prod - (TP)->tx_cons) & (TG3_TX_RING_SIZE - 1)))
1da177e4
LT
129#define NEXT_TX(N) (((N) + 1) & (TG3_TX_RING_SIZE - 1))
130
131#define RX_PKT_BUF_SZ (1536 + tp->rx_offset + 64)
132#define RX_JUMBO_PKT_BUF_SZ (9046 + tp->rx_offset + 64)
133
134/* minimum number of free TX descriptors required to wake up TX process */
135#define TG3_TX_WAKEUP_THRESH (TG3_TX_RING_SIZE / 4)
136
137/* number of ETHTOOL_GSTATS u64's */
138#define TG3_NUM_STATS (sizeof(struct tg3_ethtool_stats)/sizeof(u64))
139
4cafd3f5
MC
140#define TG3_NUM_TEST 6
141
1da177e4
LT
142static char version[] __devinitdata =
143 DRV_MODULE_NAME ".c:v" DRV_MODULE_VERSION " (" DRV_MODULE_RELDATE ")\n";
144
145MODULE_AUTHOR("David S. Miller (davem@redhat.com) and Jeff Garzik (jgarzik@pobox.com)");
146MODULE_DESCRIPTION("Broadcom Tigon3 ethernet driver");
147MODULE_LICENSE("GPL");
148MODULE_VERSION(DRV_MODULE_VERSION);
149
150static int tg3_debug = -1; /* -1 == use TG3_DEF_MSG_ENABLE as value */
151module_param(tg3_debug, int, 0);
152MODULE_PARM_DESC(tg3_debug, "Tigon3 bitmapped debugging message enable value");
153
154static struct pci_device_id tg3_pci_tbl[] = {
155 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5700,
156 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
157 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5701,
158 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
159 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5702,
160 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
161 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5703,
162 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
163 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5704,
164 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
165 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5702FE,
166 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
167 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5705,
168 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
169 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5705_2,
170 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
171 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5705M,
172 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
173 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5705M_2,
174 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
175 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5702X,
176 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
177 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5703X,
178 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
179 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5704S,
180 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
181 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5702A3,
182 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
183 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5703A3,
184 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
185 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5782,
186 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
187 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5788,
188 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
189 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5789,
190 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
191 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5901,
192 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
193 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5901_2,
194 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
195 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5704S_2,
196 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
197 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5705F,
198 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
199 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5720,
200 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
201 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5721,
202 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
203 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5750,
204 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
205 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5751,
206 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
207 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5750M,
208 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
209 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5751M,
210 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
211 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5751F,
212 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
6e9017a7 213 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5752,
af2bcd97 214 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
d8659255
XVP
215 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5752M,
216 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1da177e4
LT
217 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5753,
218 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
219 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5753M,
220 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
221 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5753F,
222 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
a4e2b347
MC
223 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5714,
224 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
225 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5715,
226 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
4cf78e4f
MC
227 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5780,
228 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
229 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5780S,
230 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
1da177e4
LT
231 { PCI_VENDOR_ID_BROADCOM, PCI_DEVICE_ID_TIGON3_5781,
232 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
233 { PCI_VENDOR_ID_SYSKONNECT, PCI_DEVICE_ID_SYSKONNECT_9DXX,
234 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
235 { PCI_VENDOR_ID_SYSKONNECT, PCI_DEVICE_ID_SYSKONNECT_9MXX,
236 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
237 { PCI_VENDOR_ID_ALTIMA, PCI_DEVICE_ID_ALTIMA_AC1000,
238 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
239 { PCI_VENDOR_ID_ALTIMA, PCI_DEVICE_ID_ALTIMA_AC1001,
240 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
241 { PCI_VENDOR_ID_ALTIMA, PCI_DEVICE_ID_ALTIMA_AC1003,
242 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
243 { PCI_VENDOR_ID_ALTIMA, PCI_DEVICE_ID_ALTIMA_AC9100,
244 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
245 { PCI_VENDOR_ID_APPLE, PCI_DEVICE_ID_APPLE_TIGON3,
246 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL },
247 { 0, }
248};
249
250MODULE_DEVICE_TABLE(pci, tg3_pci_tbl);
251
252static struct {
253 const char string[ETH_GSTRING_LEN];
254} ethtool_stats_keys[TG3_NUM_STATS] = {
255 { "rx_octets" },
256 { "rx_fragments" },
257 { "rx_ucast_packets" },
258 { "rx_mcast_packets" },
259 { "rx_bcast_packets" },
260 { "rx_fcs_errors" },
261 { "rx_align_errors" },
262 { "rx_xon_pause_rcvd" },
263 { "rx_xoff_pause_rcvd" },
264 { "rx_mac_ctrl_rcvd" },
265 { "rx_xoff_entered" },
266 { "rx_frame_too_long_errors" },
267 { "rx_jabbers" },
268 { "rx_undersize_packets" },
269 { "rx_in_length_errors" },
270 { "rx_out_length_errors" },
271 { "rx_64_or_less_octet_packets" },
272 { "rx_65_to_127_octet_packets" },
273 { "rx_128_to_255_octet_packets" },
274 { "rx_256_to_511_octet_packets" },
275 { "rx_512_to_1023_octet_packets" },
276 { "rx_1024_to_1522_octet_packets" },
277 { "rx_1523_to_2047_octet_packets" },
278 { "rx_2048_to_4095_octet_packets" },
279 { "rx_4096_to_8191_octet_packets" },
280 { "rx_8192_to_9022_octet_packets" },
281
282 { "tx_octets" },
283 { "tx_collisions" },
284
285 { "tx_xon_sent" },
286 { "tx_xoff_sent" },
287 { "tx_flow_control" },
288 { "tx_mac_errors" },
289 { "tx_single_collisions" },
290 { "tx_mult_collisions" },
291 { "tx_deferred" },
292 { "tx_excessive_collisions" },
293 { "tx_late_collisions" },
294 { "tx_collide_2times" },
295 { "tx_collide_3times" },
296 { "tx_collide_4times" },
297 { "tx_collide_5times" },
298 { "tx_collide_6times" },
299 { "tx_collide_7times" },
300 { "tx_collide_8times" },
301 { "tx_collide_9times" },
302 { "tx_collide_10times" },
303 { "tx_collide_11times" },
304 { "tx_collide_12times" },
305 { "tx_collide_13times" },
306 { "tx_collide_14times" },
307 { "tx_collide_15times" },
308 { "tx_ucast_packets" },
309 { "tx_mcast_packets" },
310 { "tx_bcast_packets" },
311 { "tx_carrier_sense_errors" },
312 { "tx_discards" },
313 { "tx_errors" },
314
315 { "dma_writeq_full" },
316 { "dma_write_prioq_full" },
317 { "rxbds_empty" },
318 { "rx_discards" },
319 { "rx_errors" },
320 { "rx_threshold_hit" },
321
322 { "dma_readq_full" },
323 { "dma_read_prioq_full" },
324 { "tx_comp_queue_full" },
325
326 { "ring_set_send_prod_index" },
327 { "ring_status_update" },
328 { "nic_irqs" },
329 { "nic_avoided_irqs" },
330 { "nic_tx_threshold_hit" }
331};
332
4cafd3f5
MC
333static struct {
334 const char string[ETH_GSTRING_LEN];
335} ethtool_test_keys[TG3_NUM_TEST] = {
336 { "nvram test (online) " },
337 { "link test (online) " },
338 { "register test (offline)" },
339 { "memory test (offline)" },
340 { "loopback test (offline)" },
341 { "interrupt test (offline)" },
342};
343
1da177e4
LT
344static void tg3_write_indirect_reg32(struct tg3 *tp, u32 off, u32 val)
345{
6892914f
MC
346 unsigned long flags;
347
348 spin_lock_irqsave(&tp->indirect_lock, flags);
1ee582d8
MC
349 pci_write_config_dword(tp->pdev, TG3PCI_REG_BASE_ADDR, off);
350 pci_write_config_dword(tp->pdev, TG3PCI_REG_DATA, val);
6892914f 351 spin_unlock_irqrestore(&tp->indirect_lock, flags);
1ee582d8
MC
352}
353
354static void tg3_write_flush_reg32(struct tg3 *tp, u32 off, u32 val)
355{
356 writel(val, tp->regs + off);
357 readl(tp->regs + off);
1da177e4
LT
358}
359
6892914f 360static u32 tg3_read_indirect_reg32(struct tg3 *tp, u32 off)
1da177e4 361{
6892914f
MC
362 unsigned long flags;
363 u32 val;
364
365 spin_lock_irqsave(&tp->indirect_lock, flags);
366 pci_write_config_dword(tp->pdev, TG3PCI_REG_BASE_ADDR, off);
367 pci_read_config_dword(tp->pdev, TG3PCI_REG_DATA, &val);
368 spin_unlock_irqrestore(&tp->indirect_lock, flags);
369 return val;
370}
371
372static void tg3_write_indirect_mbox(struct tg3 *tp, u32 off, u32 val)
373{
374 unsigned long flags;
375
376 if (off == (MAILBOX_RCVRET_CON_IDX_0 + TG3_64BIT_REG_LOW)) {
377 pci_write_config_dword(tp->pdev, TG3PCI_RCV_RET_RING_CON_IDX +
378 TG3_64BIT_REG_LOW, val);
379 return;
380 }
381 if (off == (MAILBOX_RCV_STD_PROD_IDX + TG3_64BIT_REG_LOW)) {
382 pci_write_config_dword(tp->pdev, TG3PCI_STD_RING_PROD_IDX +
383 TG3_64BIT_REG_LOW, val);
384 return;
1da177e4 385 }
6892914f
MC
386
387 spin_lock_irqsave(&tp->indirect_lock, flags);
388 pci_write_config_dword(tp->pdev, TG3PCI_REG_BASE_ADDR, off + 0x5600);
389 pci_write_config_dword(tp->pdev, TG3PCI_REG_DATA, val);
390 spin_unlock_irqrestore(&tp->indirect_lock, flags);
391
392 /* In indirect mode when disabling interrupts, we also need
393 * to clear the interrupt bit in the GRC local ctrl register.
394 */
395 if ((off == (MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW)) &&
396 (val == 0x1)) {
397 pci_write_config_dword(tp->pdev, TG3PCI_MISC_LOCAL_CTRL,
398 tp->grc_local_ctrl|GRC_LCLCTRL_CLEARINT);
399 }
400}
401
402static u32 tg3_read_indirect_mbox(struct tg3 *tp, u32 off)
403{
404 unsigned long flags;
405 u32 val;
406
407 spin_lock_irqsave(&tp->indirect_lock, flags);
408 pci_write_config_dword(tp->pdev, TG3PCI_REG_BASE_ADDR, off + 0x5600);
409 pci_read_config_dword(tp->pdev, TG3PCI_REG_DATA, &val);
410 spin_unlock_irqrestore(&tp->indirect_lock, flags);
411 return val;
412}
413
414static void _tw32_flush(struct tg3 *tp, u32 off, u32 val)
415{
416 tp->write32(tp, off, val);
417 if (!(tp->tg3_flags & TG3_FLAG_PCIX_TARGET_HWBUG) &&
418 !(tp->tg3_flags & TG3_FLAG_5701_REG_WRITE_BUG) &&
419 !(tp->tg3_flags2 & TG3_FLG2_ICH_WORKAROUND))
420 tp->read32(tp, off); /* flush */
1da177e4
LT
421}
422
09ee929c
MC
423static inline void tw32_mailbox_flush(struct tg3 *tp, u32 off, u32 val)
424{
425 tp->write32_mbox(tp, off, val);
6892914f
MC
426 if (!(tp->tg3_flags & TG3_FLAG_MBOX_WRITE_REORDER) &&
427 !(tp->tg3_flags2 & TG3_FLG2_ICH_WORKAROUND))
428 tp->read32_mbox(tp, off);
09ee929c
MC
429}
430
20094930 431static void tg3_write32_tx_mbox(struct tg3 *tp, u32 off, u32 val)
1da177e4
LT
432{
433 void __iomem *mbox = tp->regs + off;
434 writel(val, mbox);
435 if (tp->tg3_flags & TG3_FLAG_TXD_MBOX_HWBUG)
436 writel(val, mbox);
437 if (tp->tg3_flags & TG3_FLAG_MBOX_WRITE_REORDER)
438 readl(mbox);
439}
440
20094930
MC
441static void tg3_write32(struct tg3 *tp, u32 off, u32 val)
442{
443 writel(val, tp->regs + off);
444}
1da177e4 445
20094930
MC
446static u32 tg3_read32(struct tg3 *tp, u32 off)
447{
448 return (readl(tp->regs + off));
449}
450
451#define tw32_mailbox(reg, val) tp->write32_mbox(tp, reg, val)
09ee929c 452#define tw32_mailbox_f(reg, val) tw32_mailbox_flush(tp, (reg), (val))
20094930
MC
453#define tw32_rx_mbox(reg, val) tp->write32_rx_mbox(tp, reg, val)
454#define tw32_tx_mbox(reg, val) tp->write32_tx_mbox(tp, reg, val)
09ee929c 455#define tr32_mailbox(reg) tp->read32_mbox(tp, reg)
20094930
MC
456
457#define tw32(reg,val) tp->write32(tp, reg, val)
1da177e4 458#define tw32_f(reg,val) _tw32_flush(tp,(reg),(val))
20094930 459#define tr32(reg) tp->read32(tp, reg)
1da177e4
LT
460
461static void tg3_write_mem(struct tg3 *tp, u32 off, u32 val)
462{
6892914f
MC
463 unsigned long flags;
464
465 spin_lock_irqsave(&tp->indirect_lock, flags);
1da177e4
LT
466 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_BASE_ADDR, off);
467 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_DATA, val);
468
469 /* Always leave this as zero. */
470 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_BASE_ADDR, 0);
6892914f 471 spin_unlock_irqrestore(&tp->indirect_lock, flags);
1da177e4
LT
472}
473
28fbef78
MC
474static void tg3_write_mem_fast(struct tg3 *tp, u32 off, u32 val)
475{
476 /* If no workaround is needed, write to mem space directly */
477 if (tp->write32 != tg3_write_indirect_reg32)
478 tw32(NIC_SRAM_WIN_BASE + off, val);
479 else
480 tg3_write_mem(tp, off, val);
481}
482
1da177e4
LT
483static void tg3_read_mem(struct tg3 *tp, u32 off, u32 *val)
484{
6892914f
MC
485 unsigned long flags;
486
487 spin_lock_irqsave(&tp->indirect_lock, flags);
1da177e4
LT
488 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_BASE_ADDR, off);
489 pci_read_config_dword(tp->pdev, TG3PCI_MEM_WIN_DATA, val);
490
491 /* Always leave this as zero. */
492 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_BASE_ADDR, 0);
6892914f 493 spin_unlock_irqrestore(&tp->indirect_lock, flags);
1da177e4
LT
494}
495
496static void tg3_disable_ints(struct tg3 *tp)
497{
498 tw32(TG3PCI_MISC_HOST_CTRL,
499 (tp->misc_host_ctrl | MISC_HOST_CTRL_MASK_PCI_INT));
09ee929c 500 tw32_mailbox_f(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0x00000001);
1da177e4
LT
501}
502
503static inline void tg3_cond_int(struct tg3 *tp)
504{
38f3843e
MC
505 if (!(tp->tg3_flags & TG3_FLAG_TAGGED_STATUS) &&
506 (tp->hw_status->status & SD_STATUS_UPDATED))
1da177e4
LT
507 tw32(GRC_LOCAL_CTRL, tp->grc_local_ctrl | GRC_LCLCTRL_SETINT);
508}
509
510static void tg3_enable_ints(struct tg3 *tp)
511{
bbe832c0
MC
512 tp->irq_sync = 0;
513 wmb();
514
1da177e4
LT
515 tw32(TG3PCI_MISC_HOST_CTRL,
516 (tp->misc_host_ctrl & ~MISC_HOST_CTRL_MASK_PCI_INT));
09ee929c
MC
517 tw32_mailbox_f(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW,
518 (tp->last_tag << 24));
1da177e4
LT
519 tg3_cond_int(tp);
520}
521
04237ddd
MC
522static inline unsigned int tg3_has_work(struct tg3 *tp)
523{
524 struct tg3_hw_status *sblk = tp->hw_status;
525 unsigned int work_exists = 0;
526
527 /* check for phy events */
528 if (!(tp->tg3_flags &
529 (TG3_FLAG_USE_LINKCHG_REG |
530 TG3_FLAG_POLL_SERDES))) {
531 if (sblk->status & SD_STATUS_LINK_CHG)
532 work_exists = 1;
533 }
534 /* check for RX/TX work to do */
535 if (sblk->idx[0].tx_consumer != tp->tx_cons ||
536 sblk->idx[0].rx_producer != tp->rx_rcb_ptr)
537 work_exists = 1;
538
539 return work_exists;
540}
541
1da177e4 542/* tg3_restart_ints
04237ddd
MC
543 * similar to tg3_enable_ints, but it accurately determines whether there
544 * is new work pending and can return without flushing the PIO write
545 * which reenables interrupts
1da177e4
LT
546 */
547static void tg3_restart_ints(struct tg3 *tp)
548{
fac9b83e
DM
549 tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW,
550 tp->last_tag << 24);
1da177e4
LT
551 mmiowb();
552
fac9b83e
DM
553 /* When doing tagged status, this work check is unnecessary.
554 * The last_tag we write above tells the chip which piece of
555 * work we've completed.
556 */
557 if (!(tp->tg3_flags & TG3_FLAG_TAGGED_STATUS) &&
558 tg3_has_work(tp))
04237ddd
MC
559 tw32(HOSTCC_MODE, tp->coalesce_mode |
560 (HOSTCC_MODE_ENABLE | HOSTCC_MODE_NOW));
1da177e4
LT
561}
562
563static inline void tg3_netif_stop(struct tg3 *tp)
564{
bbe832c0 565 tp->dev->trans_start = jiffies; /* prevent tx timeout */
1da177e4
LT
566 netif_poll_disable(tp->dev);
567 netif_tx_disable(tp->dev);
568}
569
570static inline void tg3_netif_start(struct tg3 *tp)
571{
572 netif_wake_queue(tp->dev);
573 /* NOTE: unconditional netif_wake_queue is only appropriate
574 * so long as all callers are assured to have free tx slots
575 * (such as after tg3_init_hw)
576 */
577 netif_poll_enable(tp->dev);
f47c11ee
DM
578 tp->hw_status->status |= SD_STATUS_UPDATED;
579 tg3_enable_ints(tp);
1da177e4
LT
580}
581
582static void tg3_switch_clocks(struct tg3 *tp)
583{
584 u32 clock_ctrl = tr32(TG3PCI_CLOCK_CTRL);
585 u32 orig_clock_ctrl;
586
a4e2b347 587 if (tp->tg3_flags2 & TG3_FLG2_5780_CLASS)
4cf78e4f
MC
588 return;
589
1da177e4
LT
590 orig_clock_ctrl = clock_ctrl;
591 clock_ctrl &= (CLOCK_CTRL_FORCE_CLKRUN |
592 CLOCK_CTRL_CLKRUN_OENABLE |
593 0x1f);
594 tp->pci_clock_ctrl = clock_ctrl;
595
596 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS) {
597 if (orig_clock_ctrl & CLOCK_CTRL_625_CORE) {
598 tw32_f(TG3PCI_CLOCK_CTRL,
599 clock_ctrl | CLOCK_CTRL_625_CORE);
600 udelay(40);
601 }
602 } else if ((orig_clock_ctrl & CLOCK_CTRL_44MHZ_CORE) != 0) {
603 tw32_f(TG3PCI_CLOCK_CTRL,
604 clock_ctrl |
605 (CLOCK_CTRL_44MHZ_CORE | CLOCK_CTRL_ALTCLK));
606 udelay(40);
607 tw32_f(TG3PCI_CLOCK_CTRL,
608 clock_ctrl | (CLOCK_CTRL_ALTCLK));
609 udelay(40);
610 }
611 tw32_f(TG3PCI_CLOCK_CTRL, clock_ctrl);
612 udelay(40);
613}
614
615#define PHY_BUSY_LOOPS 5000
616
617static int tg3_readphy(struct tg3 *tp, int reg, u32 *val)
618{
619 u32 frame_val;
620 unsigned int loops;
621 int ret;
622
623 if ((tp->mi_mode & MAC_MI_MODE_AUTO_POLL) != 0) {
624 tw32_f(MAC_MI_MODE,
625 (tp->mi_mode & ~MAC_MI_MODE_AUTO_POLL));
626 udelay(80);
627 }
628
629 *val = 0x0;
630
631 frame_val = ((PHY_ADDR << MI_COM_PHY_ADDR_SHIFT) &
632 MI_COM_PHY_ADDR_MASK);
633 frame_val |= ((reg << MI_COM_REG_ADDR_SHIFT) &
634 MI_COM_REG_ADDR_MASK);
635 frame_val |= (MI_COM_CMD_READ | MI_COM_START);
636
637 tw32_f(MAC_MI_COM, frame_val);
638
639 loops = PHY_BUSY_LOOPS;
640 while (loops != 0) {
641 udelay(10);
642 frame_val = tr32(MAC_MI_COM);
643
644 if ((frame_val & MI_COM_BUSY) == 0) {
645 udelay(5);
646 frame_val = tr32(MAC_MI_COM);
647 break;
648 }
649 loops -= 1;
650 }
651
652 ret = -EBUSY;
653 if (loops != 0) {
654 *val = frame_val & MI_COM_DATA_MASK;
655 ret = 0;
656 }
657
658 if ((tp->mi_mode & MAC_MI_MODE_AUTO_POLL) != 0) {
659 tw32_f(MAC_MI_MODE, tp->mi_mode);
660 udelay(80);
661 }
662
663 return ret;
664}
665
666static int tg3_writephy(struct tg3 *tp, int reg, u32 val)
667{
668 u32 frame_val;
669 unsigned int loops;
670 int ret;
671
672 if ((tp->mi_mode & MAC_MI_MODE_AUTO_POLL) != 0) {
673 tw32_f(MAC_MI_MODE,
674 (tp->mi_mode & ~MAC_MI_MODE_AUTO_POLL));
675 udelay(80);
676 }
677
678 frame_val = ((PHY_ADDR << MI_COM_PHY_ADDR_SHIFT) &
679 MI_COM_PHY_ADDR_MASK);
680 frame_val |= ((reg << MI_COM_REG_ADDR_SHIFT) &
681 MI_COM_REG_ADDR_MASK);
682 frame_val |= (val & MI_COM_DATA_MASK);
683 frame_val |= (MI_COM_CMD_WRITE | MI_COM_START);
684
685 tw32_f(MAC_MI_COM, frame_val);
686
687 loops = PHY_BUSY_LOOPS;
688 while (loops != 0) {
689 udelay(10);
690 frame_val = tr32(MAC_MI_COM);
691 if ((frame_val & MI_COM_BUSY) == 0) {
692 udelay(5);
693 frame_val = tr32(MAC_MI_COM);
694 break;
695 }
696 loops -= 1;
697 }
698
699 ret = -EBUSY;
700 if (loops != 0)
701 ret = 0;
702
703 if ((tp->mi_mode & MAC_MI_MODE_AUTO_POLL) != 0) {
704 tw32_f(MAC_MI_MODE, tp->mi_mode);
705 udelay(80);
706 }
707
708 return ret;
709}
710
711static void tg3_phy_set_wirespeed(struct tg3 *tp)
712{
713 u32 val;
714
715 if (tp->tg3_flags2 & TG3_FLG2_NO_ETH_WIRE_SPEED)
716 return;
717
718 if (!tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x7007) &&
719 !tg3_readphy(tp, MII_TG3_AUX_CTRL, &val))
720 tg3_writephy(tp, MII_TG3_AUX_CTRL,
721 (val | (1 << 15) | (1 << 4)));
722}
723
724static int tg3_bmcr_reset(struct tg3 *tp)
725{
726 u32 phy_control;
727 int limit, err;
728
729 /* OK, reset it, and poll the BMCR_RESET bit until it
730 * clears or we time out.
731 */
732 phy_control = BMCR_RESET;
733 err = tg3_writephy(tp, MII_BMCR, phy_control);
734 if (err != 0)
735 return -EBUSY;
736
737 limit = 5000;
738 while (limit--) {
739 err = tg3_readphy(tp, MII_BMCR, &phy_control);
740 if (err != 0)
741 return -EBUSY;
742
743 if ((phy_control & BMCR_RESET) == 0) {
744 udelay(40);
745 break;
746 }
747 udelay(10);
748 }
749 if (limit <= 0)
750 return -EBUSY;
751
752 return 0;
753}
754
755static int tg3_wait_macro_done(struct tg3 *tp)
756{
757 int limit = 100;
758
759 while (limit--) {
760 u32 tmp32;
761
762 if (!tg3_readphy(tp, 0x16, &tmp32)) {
763 if ((tmp32 & 0x1000) == 0)
764 break;
765 }
766 }
767 if (limit <= 0)
768 return -EBUSY;
769
770 return 0;
771}
772
773static int tg3_phy_write_and_check_testpat(struct tg3 *tp, int *resetp)
774{
775 static const u32 test_pat[4][6] = {
776 { 0x00005555, 0x00000005, 0x00002aaa, 0x0000000a, 0x00003456, 0x00000003 },
777 { 0x00002aaa, 0x0000000a, 0x00003333, 0x00000003, 0x0000789a, 0x00000005 },
778 { 0x00005a5a, 0x00000005, 0x00002a6a, 0x0000000a, 0x00001bcd, 0x00000003 },
779 { 0x00002a5a, 0x0000000a, 0x000033c3, 0x00000003, 0x00002ef1, 0x00000005 }
780 };
781 int chan;
782
783 for (chan = 0; chan < 4; chan++) {
784 int i;
785
786 tg3_writephy(tp, MII_TG3_DSP_ADDRESS,
787 (chan * 0x2000) | 0x0200);
788 tg3_writephy(tp, 0x16, 0x0002);
789
790 for (i = 0; i < 6; i++)
791 tg3_writephy(tp, MII_TG3_DSP_RW_PORT,
792 test_pat[chan][i]);
793
794 tg3_writephy(tp, 0x16, 0x0202);
795 if (tg3_wait_macro_done(tp)) {
796 *resetp = 1;
797 return -EBUSY;
798 }
799
800 tg3_writephy(tp, MII_TG3_DSP_ADDRESS,
801 (chan * 0x2000) | 0x0200);
802 tg3_writephy(tp, 0x16, 0x0082);
803 if (tg3_wait_macro_done(tp)) {
804 *resetp = 1;
805 return -EBUSY;
806 }
807
808 tg3_writephy(tp, 0x16, 0x0802);
809 if (tg3_wait_macro_done(tp)) {
810 *resetp = 1;
811 return -EBUSY;
812 }
813
814 for (i = 0; i < 6; i += 2) {
815 u32 low, high;
816
817 if (tg3_readphy(tp, MII_TG3_DSP_RW_PORT, &low) ||
818 tg3_readphy(tp, MII_TG3_DSP_RW_PORT, &high) ||
819 tg3_wait_macro_done(tp)) {
820 *resetp = 1;
821 return -EBUSY;
822 }
823 low &= 0x7fff;
824 high &= 0x000f;
825 if (low != test_pat[chan][i] ||
826 high != test_pat[chan][i+1]) {
827 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x000b);
828 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x4001);
829 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x4005);
830
831 return -EBUSY;
832 }
833 }
834 }
835
836 return 0;
837}
838
839static int tg3_phy_reset_chanpat(struct tg3 *tp)
840{
841 int chan;
842
843 for (chan = 0; chan < 4; chan++) {
844 int i;
845
846 tg3_writephy(tp, MII_TG3_DSP_ADDRESS,
847 (chan * 0x2000) | 0x0200);
848 tg3_writephy(tp, 0x16, 0x0002);
849 for (i = 0; i < 6; i++)
850 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x000);
851 tg3_writephy(tp, 0x16, 0x0202);
852 if (tg3_wait_macro_done(tp))
853 return -EBUSY;
854 }
855
856 return 0;
857}
858
859static int tg3_phy_reset_5703_4_5(struct tg3 *tp)
860{
861 u32 reg32, phy9_orig;
862 int retries, do_phy_reset, err;
863
864 retries = 10;
865 do_phy_reset = 1;
866 do {
867 if (do_phy_reset) {
868 err = tg3_bmcr_reset(tp);
869 if (err)
870 return err;
871 do_phy_reset = 0;
872 }
873
874 /* Disable transmitter and interrupt. */
875 if (tg3_readphy(tp, MII_TG3_EXT_CTRL, &reg32))
876 continue;
877
878 reg32 |= 0x3000;
879 tg3_writephy(tp, MII_TG3_EXT_CTRL, reg32);
880
881 /* Set full-duplex, 1000 mbps. */
882 tg3_writephy(tp, MII_BMCR,
883 BMCR_FULLDPLX | TG3_BMCR_SPEED1000);
884
885 /* Set to master mode. */
886 if (tg3_readphy(tp, MII_TG3_CTRL, &phy9_orig))
887 continue;
888
889 tg3_writephy(tp, MII_TG3_CTRL,
890 (MII_TG3_CTRL_AS_MASTER |
891 MII_TG3_CTRL_ENABLE_AS_MASTER));
892
893 /* Enable SM_DSP_CLOCK and 6dB. */
894 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0c00);
895
896 /* Block the PHY control access. */
897 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x8005);
898 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x0800);
899
900 err = tg3_phy_write_and_check_testpat(tp, &do_phy_reset);
901 if (!err)
902 break;
903 } while (--retries);
904
905 err = tg3_phy_reset_chanpat(tp);
906 if (err)
907 return err;
908
909 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x8005);
910 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x0000);
911
912 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x8200);
913 tg3_writephy(tp, 0x16, 0x0000);
914
915 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 ||
916 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704) {
917 /* Set Extended packet length bit for jumbo frames */
918 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x4400);
919 }
920 else {
921 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0400);
922 }
923
924 tg3_writephy(tp, MII_TG3_CTRL, phy9_orig);
925
926 if (!tg3_readphy(tp, MII_TG3_EXT_CTRL, &reg32)) {
927 reg32 &= ~0x3000;
928 tg3_writephy(tp, MII_TG3_EXT_CTRL, reg32);
929 } else if (!err)
930 err = -EBUSY;
931
932 return err;
933}
934
935/* This will reset the tigon3 PHY if there is no valid
936 * link unless the FORCE argument is non-zero.
937 */
938static int tg3_phy_reset(struct tg3 *tp)
939{
940 u32 phy_status;
941 int err;
942
943 err = tg3_readphy(tp, MII_BMSR, &phy_status);
944 err |= tg3_readphy(tp, MII_BMSR, &phy_status);
945 if (err != 0)
946 return -EBUSY;
947
948 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 ||
949 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704 ||
950 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705) {
951 err = tg3_phy_reset_5703_4_5(tp);
952 if (err)
953 return err;
954 goto out;
955 }
956
957 err = tg3_bmcr_reset(tp);
958 if (err)
959 return err;
960
961out:
962 if (tp->tg3_flags2 & TG3_FLG2_PHY_ADC_BUG) {
963 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0c00);
964 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x201f);
965 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x2aaa);
966 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x000a);
967 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x0323);
968 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0400);
969 }
970 if (tp->tg3_flags2 & TG3_FLG2_PHY_5704_A0_BUG) {
971 tg3_writephy(tp, 0x1c, 0x8d68);
972 tg3_writephy(tp, 0x1c, 0x8d68);
973 }
974 if (tp->tg3_flags2 & TG3_FLG2_PHY_BER_BUG) {
975 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0c00);
976 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x000a);
977 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x310b);
978 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x201f);
979 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x9506);
980 tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x401f);
981 tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x14e2);
982 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0400);
983 }
984 /* Set Extended packet length bit (bit 14) on all chips that */
985 /* support jumbo frames */
986 if ((tp->phy_id & PHY_ID_MASK) == PHY_ID_BCM5401) {
987 /* Cannot do read-modify-write on 5401 */
988 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x4c20);
0f893dc6 989 } else if (tp->tg3_flags2 & TG3_FLG2_JUMBO_CAPABLE) {
1da177e4
LT
990 u32 phy_reg;
991
992 /* Set bit 14 with read-modify-write to preserve other bits */
993 if (!tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x0007) &&
994 !tg3_readphy(tp, MII_TG3_AUX_CTRL, &phy_reg))
995 tg3_writephy(tp, MII_TG3_AUX_CTRL, phy_reg | 0x4000);
996 }
997
998 /* Set phy register 0x10 bit 0 to high fifo elasticity to support
999 * jumbo frames transmission.
1000 */
0f893dc6 1001 if (tp->tg3_flags2 & TG3_FLG2_JUMBO_CAPABLE) {
1da177e4
LT
1002 u32 phy_reg;
1003
1004 if (!tg3_readphy(tp, MII_TG3_EXT_CTRL, &phy_reg))
1005 tg3_writephy(tp, MII_TG3_EXT_CTRL,
1006 phy_reg | MII_TG3_EXT_CTRL_FIFO_ELASTIC);
1007 }
1008
1009 tg3_phy_set_wirespeed(tp);
1010 return 0;
1011}
1012
1013static void tg3_frob_aux_power(struct tg3 *tp)
1014{
1015 struct tg3 *tp_peer = tp;
1016
1017 if ((tp->tg3_flags & TG3_FLAG_EEPROM_WRITE_PROT) != 0)
1018 return;
1019
1020 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704) {
1021 tp_peer = pci_get_drvdata(tp->pdev_peer);
1022 if (!tp_peer)
1023 BUG();
1024 }
1025
1026
1027 if ((tp->tg3_flags & TG3_FLAG_WOL_ENABLE) != 0 ||
1028 (tp_peer->tg3_flags & TG3_FLAG_WOL_ENABLE) != 0) {
1029 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
1030 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701) {
1031 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1032 (GRC_LCLCTRL_GPIO_OE0 |
1033 GRC_LCLCTRL_GPIO_OE1 |
1034 GRC_LCLCTRL_GPIO_OE2 |
1035 GRC_LCLCTRL_GPIO_OUTPUT0 |
1036 GRC_LCLCTRL_GPIO_OUTPUT1));
1037 udelay(100);
1038 } else {
1039 u32 no_gpio2;
1040 u32 grc_local_ctrl;
1041
1042 if (tp_peer != tp &&
1043 (tp_peer->tg3_flags & TG3_FLAG_INIT_COMPLETE) != 0)
1044 return;
1045
1046 /* On 5753 and variants, GPIO2 cannot be used. */
1047 no_gpio2 = tp->nic_sram_data_cfg &
1048 NIC_SRAM_DATA_CFG_NO_GPIO2;
1049
1050 grc_local_ctrl = GRC_LCLCTRL_GPIO_OE0 |
1051 GRC_LCLCTRL_GPIO_OE1 |
1052 GRC_LCLCTRL_GPIO_OE2 |
1053 GRC_LCLCTRL_GPIO_OUTPUT1 |
1054 GRC_LCLCTRL_GPIO_OUTPUT2;
1055 if (no_gpio2) {
1056 grc_local_ctrl &= ~(GRC_LCLCTRL_GPIO_OE2 |
1057 GRC_LCLCTRL_GPIO_OUTPUT2);
1058 }
1059 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1060 grc_local_ctrl);
1061 udelay(100);
1062
1063 grc_local_ctrl |= GRC_LCLCTRL_GPIO_OUTPUT0;
1064
1065 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1066 grc_local_ctrl);
1067 udelay(100);
1068
1069 if (!no_gpio2) {
1070 grc_local_ctrl &= ~GRC_LCLCTRL_GPIO_OUTPUT2;
1071 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1072 grc_local_ctrl);
1073 udelay(100);
1074 }
1075 }
1076 } else {
1077 if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700 &&
1078 GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5701) {
1079 if (tp_peer != tp &&
1080 (tp_peer->tg3_flags & TG3_FLAG_INIT_COMPLETE) != 0)
1081 return;
1082
1083 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1084 (GRC_LCLCTRL_GPIO_OE1 |
1085 GRC_LCLCTRL_GPIO_OUTPUT1));
1086 udelay(100);
1087
1088 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1089 (GRC_LCLCTRL_GPIO_OE1));
1090 udelay(100);
1091
1092 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl |
1093 (GRC_LCLCTRL_GPIO_OE1 |
1094 GRC_LCLCTRL_GPIO_OUTPUT1));
1095 udelay(100);
1096 }
1097 }
1098}
1099
1100static int tg3_setup_phy(struct tg3 *, int);
1101
1102#define RESET_KIND_SHUTDOWN 0
1103#define RESET_KIND_INIT 1
1104#define RESET_KIND_SUSPEND 2
1105
1106static void tg3_write_sig_post_reset(struct tg3 *, int);
1107static int tg3_halt_cpu(struct tg3 *, u32);
1108
1109static int tg3_set_power_state(struct tg3 *tp, int state)
1110{
1111 u32 misc_host_ctrl;
1112 u16 power_control, power_caps;
1113 int pm = tp->pm_cap;
1114
1115 /* Make sure register accesses (indirect or otherwise)
1116 * will function correctly.
1117 */
1118 pci_write_config_dword(tp->pdev,
1119 TG3PCI_MISC_HOST_CTRL,
1120 tp->misc_host_ctrl);
1121
1122 pci_read_config_word(tp->pdev,
1123 pm + PCI_PM_CTRL,
1124 &power_control);
1125 power_control |= PCI_PM_CTRL_PME_STATUS;
1126 power_control &= ~(PCI_PM_CTRL_STATE_MASK);
1127 switch (state) {
1128 case 0:
1129 power_control |= 0;
1130 pci_write_config_word(tp->pdev,
1131 pm + PCI_PM_CTRL,
1132 power_control);
8c6bda1a
MC
1133 udelay(100); /* Delay after power state change */
1134
1135 /* Switch out of Vaux if it is not a LOM */
1136 if (!(tp->tg3_flags & TG3_FLAG_EEPROM_WRITE_PROT)) {
1137 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl);
1138 udelay(100);
1139 }
1da177e4
LT
1140
1141 return 0;
1142
1143 case 1:
1144 power_control |= 1;
1145 break;
1146
1147 case 2:
1148 power_control |= 2;
1149 break;
1150
1151 case 3:
1152 power_control |= 3;
1153 break;
1154
1155 default:
1156 printk(KERN_WARNING PFX "%s: Invalid power state (%d) "
1157 "requested.\n",
1158 tp->dev->name, state);
1159 return -EINVAL;
1160 };
1161
1162 power_control |= PCI_PM_CTRL_PME_ENABLE;
1163
1164 misc_host_ctrl = tr32(TG3PCI_MISC_HOST_CTRL);
1165 tw32(TG3PCI_MISC_HOST_CTRL,
1166 misc_host_ctrl | MISC_HOST_CTRL_MASK_PCI_INT);
1167
1168 if (tp->link_config.phy_is_low_power == 0) {
1169 tp->link_config.phy_is_low_power = 1;
1170 tp->link_config.orig_speed = tp->link_config.speed;
1171 tp->link_config.orig_duplex = tp->link_config.duplex;
1172 tp->link_config.orig_autoneg = tp->link_config.autoneg;
1173 }
1174
747e8f8b 1175 if (!(tp->tg3_flags2 & TG3_FLG2_ANY_SERDES)) {
1da177e4
LT
1176 tp->link_config.speed = SPEED_10;
1177 tp->link_config.duplex = DUPLEX_HALF;
1178 tp->link_config.autoneg = AUTONEG_ENABLE;
1179 tg3_setup_phy(tp, 0);
1180 }
1181
1182 pci_read_config_word(tp->pdev, pm + PCI_PM_PMC, &power_caps);
1183
1184 if (tp->tg3_flags & TG3_FLAG_WOL_ENABLE) {
1185 u32 mac_mode;
1186
1187 if (!(tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)) {
1188 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x5a);
1189 udelay(40);
1190
1191 mac_mode = MAC_MODE_PORT_MODE_MII;
1192
1193 if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700 ||
1194 !(tp->tg3_flags & TG3_FLAG_WOL_SPEED_100MB))
1195 mac_mode |= MAC_MODE_LINK_POLARITY;
1196 } else {
1197 mac_mode = MAC_MODE_PORT_MODE_TBI;
1198 }
1199
cbf46853 1200 if (!(tp->tg3_flags2 & TG3_FLG2_5750_PLUS))
1da177e4
LT
1201 tw32(MAC_LED_CTRL, tp->led_ctrl);
1202
1203 if (((power_caps & PCI_PM_CAP_PME_D3cold) &&
1204 (tp->tg3_flags & TG3_FLAG_WOL_ENABLE)))
1205 mac_mode |= MAC_MODE_MAGIC_PKT_ENABLE;
1206
1207 tw32_f(MAC_MODE, mac_mode);
1208 udelay(100);
1209
1210 tw32_f(MAC_RX_MODE, RX_MODE_ENABLE);
1211 udelay(10);
1212 }
1213
1214 if (!(tp->tg3_flags & TG3_FLAG_WOL_SPEED_100MB) &&
1215 (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
1216 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701)) {
1217 u32 base_val;
1218
1219 base_val = tp->pci_clock_ctrl;
1220 base_val |= (CLOCK_CTRL_RXCLK_DISABLE |
1221 CLOCK_CTRL_TXCLK_DISABLE);
1222
1223 tw32_f(TG3PCI_CLOCK_CTRL, base_val |
1224 CLOCK_CTRL_ALTCLK |
1225 CLOCK_CTRL_PWRDOWN_PLL133);
1226 udelay(40);
a4e2b347 1227 } else if (tp->tg3_flags2 & TG3_FLG2_5780_CLASS) {
4cf78e4f 1228 /* do nothing */
85e94ced 1229 } else if (!((tp->tg3_flags2 & TG3_FLG2_5750_PLUS) &&
1da177e4
LT
1230 (tp->tg3_flags & TG3_FLAG_ENABLE_ASF))) {
1231 u32 newbits1, newbits2;
1232
1233 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
1234 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701) {
1235 newbits1 = (CLOCK_CTRL_RXCLK_DISABLE |
1236 CLOCK_CTRL_TXCLK_DISABLE |
1237 CLOCK_CTRL_ALTCLK);
1238 newbits2 = newbits1 | CLOCK_CTRL_44MHZ_CORE;
1239 } else if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS) {
1240 newbits1 = CLOCK_CTRL_625_CORE;
1241 newbits2 = newbits1 | CLOCK_CTRL_ALTCLK;
1242 } else {
1243 newbits1 = CLOCK_CTRL_ALTCLK;
1244 newbits2 = newbits1 | CLOCK_CTRL_44MHZ_CORE;
1245 }
1246
1247 tw32_f(TG3PCI_CLOCK_CTRL, tp->pci_clock_ctrl | newbits1);
1248 udelay(40);
1249
1250 tw32_f(TG3PCI_CLOCK_CTRL, tp->pci_clock_ctrl | newbits2);
1251 udelay(40);
1252
1253 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
1254 u32 newbits3;
1255
1256 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
1257 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701) {
1258 newbits3 = (CLOCK_CTRL_RXCLK_DISABLE |
1259 CLOCK_CTRL_TXCLK_DISABLE |
1260 CLOCK_CTRL_44MHZ_CORE);
1261 } else {
1262 newbits3 = CLOCK_CTRL_44MHZ_CORE;
1263 }
1264
1265 tw32_f(TG3PCI_CLOCK_CTRL,
1266 tp->pci_clock_ctrl | newbits3);
1267 udelay(40);
1268 }
1269 }
1270
1271 tg3_frob_aux_power(tp);
1272
1273 /* Workaround for unstable PLL clock */
1274 if ((GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5750_AX) ||
1275 (GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5750_BX)) {
1276 u32 val = tr32(0x7d00);
1277
1278 val &= ~((1 << 16) | (1 << 4) | (1 << 2) | (1 << 1) | 1);
1279 tw32(0x7d00, val);
1280 if (!(tp->tg3_flags & TG3_FLAG_ENABLE_ASF))
1281 tg3_halt_cpu(tp, RX_CPU_BASE);
1282 }
1283
1284 /* Finally, set the new power state. */
1285 pci_write_config_word(tp->pdev, pm + PCI_PM_CTRL, power_control);
8c6bda1a 1286 udelay(100); /* Delay after power state change */
1da177e4
LT
1287
1288 tg3_write_sig_post_reset(tp, RESET_KIND_SHUTDOWN);
1289
1290 return 0;
1291}
1292
1293static void tg3_link_report(struct tg3 *tp)
1294{
1295 if (!netif_carrier_ok(tp->dev)) {
1296 printk(KERN_INFO PFX "%s: Link is down.\n", tp->dev->name);
1297 } else {
1298 printk(KERN_INFO PFX "%s: Link is up at %d Mbps, %s duplex.\n",
1299 tp->dev->name,
1300 (tp->link_config.active_speed == SPEED_1000 ?
1301 1000 :
1302 (tp->link_config.active_speed == SPEED_100 ?
1303 100 : 10)),
1304 (tp->link_config.active_duplex == DUPLEX_FULL ?
1305 "full" : "half"));
1306
1307 printk(KERN_INFO PFX "%s: Flow control is %s for TX and "
1308 "%s for RX.\n",
1309 tp->dev->name,
1310 (tp->tg3_flags & TG3_FLAG_TX_PAUSE) ? "on" : "off",
1311 (tp->tg3_flags & TG3_FLAG_RX_PAUSE) ? "on" : "off");
1312 }
1313}
1314
1315static void tg3_setup_flow_control(struct tg3 *tp, u32 local_adv, u32 remote_adv)
1316{
1317 u32 new_tg3_flags = 0;
1318 u32 old_rx_mode = tp->rx_mode;
1319 u32 old_tx_mode = tp->tx_mode;
1320
1321 if (tp->tg3_flags & TG3_FLAG_PAUSE_AUTONEG) {
747e8f8b
MC
1322
1323 /* Convert 1000BaseX flow control bits to 1000BaseT
1324 * bits before resolving flow control.
1325 */
1326 if (tp->tg3_flags2 & TG3_FLG2_MII_SERDES) {
1327 local_adv &= ~(ADVERTISE_PAUSE_CAP |
1328 ADVERTISE_PAUSE_ASYM);
1329 remote_adv &= ~(LPA_PAUSE_CAP | LPA_PAUSE_ASYM);
1330
1331 if (local_adv & ADVERTISE_1000XPAUSE)
1332 local_adv |= ADVERTISE_PAUSE_CAP;
1333 if (local_adv & ADVERTISE_1000XPSE_ASYM)
1334 local_adv |= ADVERTISE_PAUSE_ASYM;
1335 if (remote_adv & LPA_1000XPAUSE)
1336 remote_adv |= LPA_PAUSE_CAP;
1337 if (remote_adv & LPA_1000XPAUSE_ASYM)
1338 remote_adv |= LPA_PAUSE_ASYM;
1339 }
1340
1da177e4
LT
1341 if (local_adv & ADVERTISE_PAUSE_CAP) {
1342 if (local_adv & ADVERTISE_PAUSE_ASYM) {
1343 if (remote_adv & LPA_PAUSE_CAP)
1344 new_tg3_flags |=
1345 (TG3_FLAG_RX_PAUSE |
1346 TG3_FLAG_TX_PAUSE);
1347 else if (remote_adv & LPA_PAUSE_ASYM)
1348 new_tg3_flags |=
1349 (TG3_FLAG_RX_PAUSE);
1350 } else {
1351 if (remote_adv & LPA_PAUSE_CAP)
1352 new_tg3_flags |=
1353 (TG3_FLAG_RX_PAUSE |
1354 TG3_FLAG_TX_PAUSE);
1355 }
1356 } else if (local_adv & ADVERTISE_PAUSE_ASYM) {
1357 if ((remote_adv & LPA_PAUSE_CAP) &&
1358 (remote_adv & LPA_PAUSE_ASYM))
1359 new_tg3_flags |= TG3_FLAG_TX_PAUSE;
1360 }
1361
1362 tp->tg3_flags &= ~(TG3_FLAG_RX_PAUSE | TG3_FLAG_TX_PAUSE);
1363 tp->tg3_flags |= new_tg3_flags;
1364 } else {
1365 new_tg3_flags = tp->tg3_flags;
1366 }
1367
1368 if (new_tg3_flags & TG3_FLAG_RX_PAUSE)
1369 tp->rx_mode |= RX_MODE_FLOW_CTRL_ENABLE;
1370 else
1371 tp->rx_mode &= ~RX_MODE_FLOW_CTRL_ENABLE;
1372
1373 if (old_rx_mode != tp->rx_mode) {
1374 tw32_f(MAC_RX_MODE, tp->rx_mode);
1375 }
1376
1377 if (new_tg3_flags & TG3_FLAG_TX_PAUSE)
1378 tp->tx_mode |= TX_MODE_FLOW_CTRL_ENABLE;
1379 else
1380 tp->tx_mode &= ~TX_MODE_FLOW_CTRL_ENABLE;
1381
1382 if (old_tx_mode != tp->tx_mode) {
1383 tw32_f(MAC_TX_MODE, tp->tx_mode);
1384 }
1385}
1386
1387static void tg3_aux_stat_to_speed_duplex(struct tg3 *tp, u32 val, u16 *speed, u8 *duplex)
1388{
1389 switch (val & MII_TG3_AUX_STAT_SPDMASK) {
1390 case MII_TG3_AUX_STAT_10HALF:
1391 *speed = SPEED_10;
1392 *duplex = DUPLEX_HALF;
1393 break;
1394
1395 case MII_TG3_AUX_STAT_10FULL:
1396 *speed = SPEED_10;
1397 *duplex = DUPLEX_FULL;
1398 break;
1399
1400 case MII_TG3_AUX_STAT_100HALF:
1401 *speed = SPEED_100;
1402 *duplex = DUPLEX_HALF;
1403 break;
1404
1405 case MII_TG3_AUX_STAT_100FULL:
1406 *speed = SPEED_100;
1407 *duplex = DUPLEX_FULL;
1408 break;
1409
1410 case MII_TG3_AUX_STAT_1000HALF:
1411 *speed = SPEED_1000;
1412 *duplex = DUPLEX_HALF;
1413 break;
1414
1415 case MII_TG3_AUX_STAT_1000FULL:
1416 *speed = SPEED_1000;
1417 *duplex = DUPLEX_FULL;
1418 break;
1419
1420 default:
1421 *speed = SPEED_INVALID;
1422 *duplex = DUPLEX_INVALID;
1423 break;
1424 };
1425}
1426
1427static void tg3_phy_copper_begin(struct tg3 *tp)
1428{
1429 u32 new_adv;
1430 int i;
1431
1432 if (tp->link_config.phy_is_low_power) {
1433 /* Entering low power mode. Disable gigabit and
1434 * 100baseT advertisements.
1435 */
1436 tg3_writephy(tp, MII_TG3_CTRL, 0);
1437
1438 new_adv = (ADVERTISE_10HALF | ADVERTISE_10FULL |
1439 ADVERTISE_CSMA | ADVERTISE_PAUSE_CAP);
1440 if (tp->tg3_flags & TG3_FLAG_WOL_SPEED_100MB)
1441 new_adv |= (ADVERTISE_100HALF | ADVERTISE_100FULL);
1442
1443 tg3_writephy(tp, MII_ADVERTISE, new_adv);
1444 } else if (tp->link_config.speed == SPEED_INVALID) {
1445 tp->link_config.advertising =
1446 (ADVERTISED_10baseT_Half | ADVERTISED_10baseT_Full |
1447 ADVERTISED_100baseT_Half | ADVERTISED_100baseT_Full |
1448 ADVERTISED_1000baseT_Half | ADVERTISED_1000baseT_Full |
1449 ADVERTISED_Autoneg | ADVERTISED_MII);
1450
1451 if (tp->tg3_flags & TG3_FLAG_10_100_ONLY)
1452 tp->link_config.advertising &=
1453 ~(ADVERTISED_1000baseT_Half |
1454 ADVERTISED_1000baseT_Full);
1455
1456 new_adv = (ADVERTISE_CSMA | ADVERTISE_PAUSE_CAP);
1457 if (tp->link_config.advertising & ADVERTISED_10baseT_Half)
1458 new_adv |= ADVERTISE_10HALF;
1459 if (tp->link_config.advertising & ADVERTISED_10baseT_Full)
1460 new_adv |= ADVERTISE_10FULL;
1461 if (tp->link_config.advertising & ADVERTISED_100baseT_Half)
1462 new_adv |= ADVERTISE_100HALF;
1463 if (tp->link_config.advertising & ADVERTISED_100baseT_Full)
1464 new_adv |= ADVERTISE_100FULL;
1465 tg3_writephy(tp, MII_ADVERTISE, new_adv);
1466
1467 if (tp->link_config.advertising &
1468 (ADVERTISED_1000baseT_Half | ADVERTISED_1000baseT_Full)) {
1469 new_adv = 0;
1470 if (tp->link_config.advertising & ADVERTISED_1000baseT_Half)
1471 new_adv |= MII_TG3_CTRL_ADV_1000_HALF;
1472 if (tp->link_config.advertising & ADVERTISED_1000baseT_Full)
1473 new_adv |= MII_TG3_CTRL_ADV_1000_FULL;
1474 if (!(tp->tg3_flags & TG3_FLAG_10_100_ONLY) &&
1475 (tp->pci_chip_rev_id == CHIPREV_ID_5701_A0 ||
1476 tp->pci_chip_rev_id == CHIPREV_ID_5701_B0))
1477 new_adv |= (MII_TG3_CTRL_AS_MASTER |
1478 MII_TG3_CTRL_ENABLE_AS_MASTER);
1479 tg3_writephy(tp, MII_TG3_CTRL, new_adv);
1480 } else {
1481 tg3_writephy(tp, MII_TG3_CTRL, 0);
1482 }
1483 } else {
1484 /* Asking for a specific link mode. */
1485 if (tp->link_config.speed == SPEED_1000) {
1486 new_adv = ADVERTISE_CSMA | ADVERTISE_PAUSE_CAP;
1487 tg3_writephy(tp, MII_ADVERTISE, new_adv);
1488
1489 if (tp->link_config.duplex == DUPLEX_FULL)
1490 new_adv = MII_TG3_CTRL_ADV_1000_FULL;
1491 else
1492 new_adv = MII_TG3_CTRL_ADV_1000_HALF;
1493 if (tp->pci_chip_rev_id == CHIPREV_ID_5701_A0 ||
1494 tp->pci_chip_rev_id == CHIPREV_ID_5701_B0)
1495 new_adv |= (MII_TG3_CTRL_AS_MASTER |
1496 MII_TG3_CTRL_ENABLE_AS_MASTER);
1497 tg3_writephy(tp, MII_TG3_CTRL, new_adv);
1498 } else {
1499 tg3_writephy(tp, MII_TG3_CTRL, 0);
1500
1501 new_adv = ADVERTISE_CSMA | ADVERTISE_PAUSE_CAP;
1502 if (tp->link_config.speed == SPEED_100) {
1503 if (tp->link_config.duplex == DUPLEX_FULL)
1504 new_adv |= ADVERTISE_100FULL;
1505 else
1506 new_adv |= ADVERTISE_100HALF;
1507 } else {
1508 if (tp->link_config.duplex == DUPLEX_FULL)
1509 new_adv |= ADVERTISE_10FULL;
1510 else
1511 new_adv |= ADVERTISE_10HALF;
1512 }
1513 tg3_writephy(tp, MII_ADVERTISE, new_adv);
1514 }
1515 }
1516
1517 if (tp->link_config.autoneg == AUTONEG_DISABLE &&
1518 tp->link_config.speed != SPEED_INVALID) {
1519 u32 bmcr, orig_bmcr;
1520
1521 tp->link_config.active_speed = tp->link_config.speed;
1522 tp->link_config.active_duplex = tp->link_config.duplex;
1523
1524 bmcr = 0;
1525 switch (tp->link_config.speed) {
1526 default:
1527 case SPEED_10:
1528 break;
1529
1530 case SPEED_100:
1531 bmcr |= BMCR_SPEED100;
1532 break;
1533
1534 case SPEED_1000:
1535 bmcr |= TG3_BMCR_SPEED1000;
1536 break;
1537 };
1538
1539 if (tp->link_config.duplex == DUPLEX_FULL)
1540 bmcr |= BMCR_FULLDPLX;
1541
1542 if (!tg3_readphy(tp, MII_BMCR, &orig_bmcr) &&
1543 (bmcr != orig_bmcr)) {
1544 tg3_writephy(tp, MII_BMCR, BMCR_LOOPBACK);
1545 for (i = 0; i < 1500; i++) {
1546 u32 tmp;
1547
1548 udelay(10);
1549 if (tg3_readphy(tp, MII_BMSR, &tmp) ||
1550 tg3_readphy(tp, MII_BMSR, &tmp))
1551 continue;
1552 if (!(tmp & BMSR_LSTATUS)) {
1553 udelay(40);
1554 break;
1555 }
1556 }
1557 tg3_writephy(tp, MII_BMCR, bmcr);
1558 udelay(40);
1559 }
1560 } else {
1561 tg3_writephy(tp, MII_BMCR,
1562 BMCR_ANENABLE | BMCR_ANRESTART);
1563 }
1564}
1565
1566static int tg3_init_5401phy_dsp(struct tg3 *tp)
1567{
1568 int err;
1569
1570 /* Turn off tap power management. */
1571 /* Set Extended packet length bit */
1572 err = tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x4c20);
1573
1574 err |= tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x0012);
1575 err |= tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x1804);
1576
1577 err |= tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x0013);
1578 err |= tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x1204);
1579
1580 err |= tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x8006);
1581 err |= tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x0132);
1582
1583 err |= tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x8006);
1584 err |= tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x0232);
1585
1586 err |= tg3_writephy(tp, MII_TG3_DSP_ADDRESS, 0x201f);
1587 err |= tg3_writephy(tp, MII_TG3_DSP_RW_PORT, 0x0a20);
1588
1589 udelay(40);
1590
1591 return err;
1592}
1593
1594static int tg3_copper_is_advertising_all(struct tg3 *tp)
1595{
1596 u32 adv_reg, all_mask;
1597
1598 if (tg3_readphy(tp, MII_ADVERTISE, &adv_reg))
1599 return 0;
1600
1601 all_mask = (ADVERTISE_10HALF | ADVERTISE_10FULL |
1602 ADVERTISE_100HALF | ADVERTISE_100FULL);
1603 if ((adv_reg & all_mask) != all_mask)
1604 return 0;
1605 if (!(tp->tg3_flags & TG3_FLAG_10_100_ONLY)) {
1606 u32 tg3_ctrl;
1607
1608 if (tg3_readphy(tp, MII_TG3_CTRL, &tg3_ctrl))
1609 return 0;
1610
1611 all_mask = (MII_TG3_CTRL_ADV_1000_HALF |
1612 MII_TG3_CTRL_ADV_1000_FULL);
1613 if ((tg3_ctrl & all_mask) != all_mask)
1614 return 0;
1615 }
1616 return 1;
1617}
1618
1619static int tg3_setup_copper_phy(struct tg3 *tp, int force_reset)
1620{
1621 int current_link_up;
1622 u32 bmsr, dummy;
1623 u16 current_speed;
1624 u8 current_duplex;
1625 int i, err;
1626
1627 tw32(MAC_EVENT, 0);
1628
1629 tw32_f(MAC_STATUS,
1630 (MAC_STATUS_SYNC_CHANGED |
1631 MAC_STATUS_CFG_CHANGED |
1632 MAC_STATUS_MI_COMPLETION |
1633 MAC_STATUS_LNKSTATE_CHANGED));
1634 udelay(40);
1635
1636 tp->mi_mode = MAC_MI_MODE_BASE;
1637 tw32_f(MAC_MI_MODE, tp->mi_mode);
1638 udelay(80);
1639
1640 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x02);
1641
1642 /* Some third-party PHYs need to be reset on link going
1643 * down.
1644 */
1645 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 ||
1646 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704 ||
1647 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705) &&
1648 netif_carrier_ok(tp->dev)) {
1649 tg3_readphy(tp, MII_BMSR, &bmsr);
1650 if (!tg3_readphy(tp, MII_BMSR, &bmsr) &&
1651 !(bmsr & BMSR_LSTATUS))
1652 force_reset = 1;
1653 }
1654 if (force_reset)
1655 tg3_phy_reset(tp);
1656
1657 if ((tp->phy_id & PHY_ID_MASK) == PHY_ID_BCM5401) {
1658 tg3_readphy(tp, MII_BMSR, &bmsr);
1659 if (tg3_readphy(tp, MII_BMSR, &bmsr) ||
1660 !(tp->tg3_flags & TG3_FLAG_INIT_COMPLETE))
1661 bmsr = 0;
1662
1663 if (!(bmsr & BMSR_LSTATUS)) {
1664 err = tg3_init_5401phy_dsp(tp);
1665 if (err)
1666 return err;
1667
1668 tg3_readphy(tp, MII_BMSR, &bmsr);
1669 for (i = 0; i < 1000; i++) {
1670 udelay(10);
1671 if (!tg3_readphy(tp, MII_BMSR, &bmsr) &&
1672 (bmsr & BMSR_LSTATUS)) {
1673 udelay(40);
1674 break;
1675 }
1676 }
1677
1678 if ((tp->phy_id & PHY_ID_REV_MASK) == PHY_REV_BCM5401_B0 &&
1679 !(bmsr & BMSR_LSTATUS) &&
1680 tp->link_config.active_speed == SPEED_1000) {
1681 err = tg3_phy_reset(tp);
1682 if (!err)
1683 err = tg3_init_5401phy_dsp(tp);
1684 if (err)
1685 return err;
1686 }
1687 }
1688 } else if (tp->pci_chip_rev_id == CHIPREV_ID_5701_A0 ||
1689 tp->pci_chip_rev_id == CHIPREV_ID_5701_B0) {
1690 /* 5701 {A0,B0} CRC bug workaround */
1691 tg3_writephy(tp, 0x15, 0x0a75);
1692 tg3_writephy(tp, 0x1c, 0x8c68);
1693 tg3_writephy(tp, 0x1c, 0x8d68);
1694 tg3_writephy(tp, 0x1c, 0x8c68);
1695 }
1696
1697 /* Clear pending interrupts... */
1698 tg3_readphy(tp, MII_TG3_ISTAT, &dummy);
1699 tg3_readphy(tp, MII_TG3_ISTAT, &dummy);
1700
1701 if (tp->tg3_flags & TG3_FLAG_USE_MI_INTERRUPT)
1702 tg3_writephy(tp, MII_TG3_IMASK, ~MII_TG3_INT_LINKCHG);
1703 else
1704 tg3_writephy(tp, MII_TG3_IMASK, ~0);
1705
1706 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
1707 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701) {
1708 if (tp->led_ctrl == LED_CTRL_MODE_PHY_1)
1709 tg3_writephy(tp, MII_TG3_EXT_CTRL,
1710 MII_TG3_EXT_CTRL_LNK3_LED_MODE);
1711 else
1712 tg3_writephy(tp, MII_TG3_EXT_CTRL, 0);
1713 }
1714
1715 current_link_up = 0;
1716 current_speed = SPEED_INVALID;
1717 current_duplex = DUPLEX_INVALID;
1718
1719 if (tp->tg3_flags2 & TG3_FLG2_CAPACITIVE_COUPLING) {
1720 u32 val;
1721
1722 tg3_writephy(tp, MII_TG3_AUX_CTRL, 0x4007);
1723 tg3_readphy(tp, MII_TG3_AUX_CTRL, &val);
1724 if (!(val & (1 << 10))) {
1725 val |= (1 << 10);
1726 tg3_writephy(tp, MII_TG3_AUX_CTRL, val);
1727 goto relink;
1728 }
1729 }
1730
1731 bmsr = 0;
1732 for (i = 0; i < 100; i++) {
1733 tg3_readphy(tp, MII_BMSR, &bmsr);
1734 if (!tg3_readphy(tp, MII_BMSR, &bmsr) &&
1735 (bmsr & BMSR_LSTATUS))
1736 break;
1737 udelay(40);
1738 }
1739
1740 if (bmsr & BMSR_LSTATUS) {
1741 u32 aux_stat, bmcr;
1742
1743 tg3_readphy(tp, MII_TG3_AUX_STAT, &aux_stat);
1744 for (i = 0; i < 2000; i++) {
1745 udelay(10);
1746 if (!tg3_readphy(tp, MII_TG3_AUX_STAT, &aux_stat) &&
1747 aux_stat)
1748 break;
1749 }
1750
1751 tg3_aux_stat_to_speed_duplex(tp, aux_stat,
1752 &current_speed,
1753 &current_duplex);
1754
1755 bmcr = 0;
1756 for (i = 0; i < 200; i++) {
1757 tg3_readphy(tp, MII_BMCR, &bmcr);
1758 if (tg3_readphy(tp, MII_BMCR, &bmcr))
1759 continue;
1760 if (bmcr && bmcr != 0x7fff)
1761 break;
1762 udelay(10);
1763 }
1764
1765 if (tp->link_config.autoneg == AUTONEG_ENABLE) {
1766 if (bmcr & BMCR_ANENABLE) {
1767 current_link_up = 1;
1768
1769 /* Force autoneg restart if we are exiting
1770 * low power mode.
1771 */
1772 if (!tg3_copper_is_advertising_all(tp))
1773 current_link_up = 0;
1774 } else {
1775 current_link_up = 0;
1776 }
1777 } else {
1778 if (!(bmcr & BMCR_ANENABLE) &&
1779 tp->link_config.speed == current_speed &&
1780 tp->link_config.duplex == current_duplex) {
1781 current_link_up = 1;
1782 } else {
1783 current_link_up = 0;
1784 }
1785 }
1786
1787 tp->link_config.active_speed = current_speed;
1788 tp->link_config.active_duplex = current_duplex;
1789 }
1790
1791 if (current_link_up == 1 &&
1792 (tp->link_config.active_duplex == DUPLEX_FULL) &&
1793 (tp->link_config.autoneg == AUTONEG_ENABLE)) {
1794 u32 local_adv, remote_adv;
1795
1796 if (tg3_readphy(tp, MII_ADVERTISE, &local_adv))
1797 local_adv = 0;
1798 local_adv &= (ADVERTISE_PAUSE_CAP | ADVERTISE_PAUSE_ASYM);
1799
1800 if (tg3_readphy(tp, MII_LPA, &remote_adv))
1801 remote_adv = 0;
1802
1803 remote_adv &= (LPA_PAUSE_CAP | LPA_PAUSE_ASYM);
1804
1805 /* If we are not advertising full pause capability,
1806 * something is wrong. Bring the link down and reconfigure.
1807 */
1808 if (local_adv != ADVERTISE_PAUSE_CAP) {
1809 current_link_up = 0;
1810 } else {
1811 tg3_setup_flow_control(tp, local_adv, remote_adv);
1812 }
1813 }
1814relink:
1815 if (current_link_up == 0) {
1816 u32 tmp;
1817
1818 tg3_phy_copper_begin(tp);
1819
1820 tg3_readphy(tp, MII_BMSR, &tmp);
1821 if (!tg3_readphy(tp, MII_BMSR, &tmp) &&
1822 (tmp & BMSR_LSTATUS))
1823 current_link_up = 1;
1824 }
1825
1826 tp->mac_mode &= ~MAC_MODE_PORT_MODE_MASK;
1827 if (current_link_up == 1) {
1828 if (tp->link_config.active_speed == SPEED_100 ||
1829 tp->link_config.active_speed == SPEED_10)
1830 tp->mac_mode |= MAC_MODE_PORT_MODE_MII;
1831 else
1832 tp->mac_mode |= MAC_MODE_PORT_MODE_GMII;
1833 } else
1834 tp->mac_mode |= MAC_MODE_PORT_MODE_GMII;
1835
1836 tp->mac_mode &= ~MAC_MODE_HALF_DUPLEX;
1837 if (tp->link_config.active_duplex == DUPLEX_HALF)
1838 tp->mac_mode |= MAC_MODE_HALF_DUPLEX;
1839
1840 tp->mac_mode &= ~MAC_MODE_LINK_POLARITY;
1841 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700) {
1842 if ((tp->led_ctrl == LED_CTRL_MODE_PHY_2) ||
1843 (current_link_up == 1 &&
1844 tp->link_config.active_speed == SPEED_10))
1845 tp->mac_mode |= MAC_MODE_LINK_POLARITY;
1846 } else {
1847 if (current_link_up == 1)
1848 tp->mac_mode |= MAC_MODE_LINK_POLARITY;
1849 }
1850
1851 /* ??? Without this setting Netgear GA302T PHY does not
1852 * ??? send/receive packets...
1853 */
1854 if ((tp->phy_id & PHY_ID_MASK) == PHY_ID_BCM5411 &&
1855 tp->pci_chip_rev_id == CHIPREV_ID_5700_ALTIMA) {
1856 tp->mi_mode |= MAC_MI_MODE_AUTO_POLL;
1857 tw32_f(MAC_MI_MODE, tp->mi_mode);
1858 udelay(80);
1859 }
1860
1861 tw32_f(MAC_MODE, tp->mac_mode);
1862 udelay(40);
1863
1864 if (tp->tg3_flags & TG3_FLAG_USE_LINKCHG_REG) {
1865 /* Polled via timer. */
1866 tw32_f(MAC_EVENT, 0);
1867 } else {
1868 tw32_f(MAC_EVENT, MAC_EVENT_LNKSTATE_CHANGED);
1869 }
1870 udelay(40);
1871
1872 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 &&
1873 current_link_up == 1 &&
1874 tp->link_config.active_speed == SPEED_1000 &&
1875 ((tp->tg3_flags & TG3_FLAG_PCIX_MODE) ||
1876 (tp->tg3_flags & TG3_FLAG_PCI_HIGH_SPEED))) {
1877 udelay(120);
1878 tw32_f(MAC_STATUS,
1879 (MAC_STATUS_SYNC_CHANGED |
1880 MAC_STATUS_CFG_CHANGED));
1881 udelay(40);
1882 tg3_write_mem(tp,
1883 NIC_SRAM_FIRMWARE_MBOX,
1884 NIC_SRAM_FIRMWARE_MBOX_MAGIC2);
1885 }
1886
1887 if (current_link_up != netif_carrier_ok(tp->dev)) {
1888 if (current_link_up)
1889 netif_carrier_on(tp->dev);
1890 else
1891 netif_carrier_off(tp->dev);
1892 tg3_link_report(tp);
1893 }
1894
1895 return 0;
1896}
1897
1898struct tg3_fiber_aneginfo {
1899 int state;
1900#define ANEG_STATE_UNKNOWN 0
1901#define ANEG_STATE_AN_ENABLE 1
1902#define ANEG_STATE_RESTART_INIT 2
1903#define ANEG_STATE_RESTART 3
1904#define ANEG_STATE_DISABLE_LINK_OK 4
1905#define ANEG_STATE_ABILITY_DETECT_INIT 5
1906#define ANEG_STATE_ABILITY_DETECT 6
1907#define ANEG_STATE_ACK_DETECT_INIT 7
1908#define ANEG_STATE_ACK_DETECT 8
1909#define ANEG_STATE_COMPLETE_ACK_INIT 9
1910#define ANEG_STATE_COMPLETE_ACK 10
1911#define ANEG_STATE_IDLE_DETECT_INIT 11
1912#define ANEG_STATE_IDLE_DETECT 12
1913#define ANEG_STATE_LINK_OK 13
1914#define ANEG_STATE_NEXT_PAGE_WAIT_INIT 14
1915#define ANEG_STATE_NEXT_PAGE_WAIT 15
1916
1917 u32 flags;
1918#define MR_AN_ENABLE 0x00000001
1919#define MR_RESTART_AN 0x00000002
1920#define MR_AN_COMPLETE 0x00000004
1921#define MR_PAGE_RX 0x00000008
1922#define MR_NP_LOADED 0x00000010
1923#define MR_TOGGLE_TX 0x00000020
1924#define MR_LP_ADV_FULL_DUPLEX 0x00000040
1925#define MR_LP_ADV_HALF_DUPLEX 0x00000080
1926#define MR_LP_ADV_SYM_PAUSE 0x00000100
1927#define MR_LP_ADV_ASYM_PAUSE 0x00000200
1928#define MR_LP_ADV_REMOTE_FAULT1 0x00000400
1929#define MR_LP_ADV_REMOTE_FAULT2 0x00000800
1930#define MR_LP_ADV_NEXT_PAGE 0x00001000
1931#define MR_TOGGLE_RX 0x00002000
1932#define MR_NP_RX 0x00004000
1933
1934#define MR_LINK_OK 0x80000000
1935
1936 unsigned long link_time, cur_time;
1937
1938 u32 ability_match_cfg;
1939 int ability_match_count;
1940
1941 char ability_match, idle_match, ack_match;
1942
1943 u32 txconfig, rxconfig;
1944#define ANEG_CFG_NP 0x00000080
1945#define ANEG_CFG_ACK 0x00000040
1946#define ANEG_CFG_RF2 0x00000020
1947#define ANEG_CFG_RF1 0x00000010
1948#define ANEG_CFG_PS2 0x00000001
1949#define ANEG_CFG_PS1 0x00008000
1950#define ANEG_CFG_HD 0x00004000
1951#define ANEG_CFG_FD 0x00002000
1952#define ANEG_CFG_INVAL 0x00001f06
1953
1954};
1955#define ANEG_OK 0
1956#define ANEG_DONE 1
1957#define ANEG_TIMER_ENAB 2
1958#define ANEG_FAILED -1
1959
1960#define ANEG_STATE_SETTLE_TIME 10000
1961
1962static int tg3_fiber_aneg_smachine(struct tg3 *tp,
1963 struct tg3_fiber_aneginfo *ap)
1964{
1965 unsigned long delta;
1966 u32 rx_cfg_reg;
1967 int ret;
1968
1969 if (ap->state == ANEG_STATE_UNKNOWN) {
1970 ap->rxconfig = 0;
1971 ap->link_time = 0;
1972 ap->cur_time = 0;
1973 ap->ability_match_cfg = 0;
1974 ap->ability_match_count = 0;
1975 ap->ability_match = 0;
1976 ap->idle_match = 0;
1977 ap->ack_match = 0;
1978 }
1979 ap->cur_time++;
1980
1981 if (tr32(MAC_STATUS) & MAC_STATUS_RCVD_CFG) {
1982 rx_cfg_reg = tr32(MAC_RX_AUTO_NEG);
1983
1984 if (rx_cfg_reg != ap->ability_match_cfg) {
1985 ap->ability_match_cfg = rx_cfg_reg;
1986 ap->ability_match = 0;
1987 ap->ability_match_count = 0;
1988 } else {
1989 if (++ap->ability_match_count > 1) {
1990 ap->ability_match = 1;
1991 ap->ability_match_cfg = rx_cfg_reg;
1992 }
1993 }
1994 if (rx_cfg_reg & ANEG_CFG_ACK)
1995 ap->ack_match = 1;
1996 else
1997 ap->ack_match = 0;
1998
1999 ap->idle_match = 0;
2000 } else {
2001 ap->idle_match = 1;
2002 ap->ability_match_cfg = 0;
2003 ap->ability_match_count = 0;
2004 ap->ability_match = 0;
2005 ap->ack_match = 0;
2006
2007 rx_cfg_reg = 0;
2008 }
2009
2010 ap->rxconfig = rx_cfg_reg;
2011 ret = ANEG_OK;
2012
2013 switch(ap->state) {
2014 case ANEG_STATE_UNKNOWN:
2015 if (ap->flags & (MR_AN_ENABLE | MR_RESTART_AN))
2016 ap->state = ANEG_STATE_AN_ENABLE;
2017
2018 /* fallthru */
2019 case ANEG_STATE_AN_ENABLE:
2020 ap->flags &= ~(MR_AN_COMPLETE | MR_PAGE_RX);
2021 if (ap->flags & MR_AN_ENABLE) {
2022 ap->link_time = 0;
2023 ap->cur_time = 0;
2024 ap->ability_match_cfg = 0;
2025 ap->ability_match_count = 0;
2026 ap->ability_match = 0;
2027 ap->idle_match = 0;
2028 ap->ack_match = 0;
2029
2030 ap->state = ANEG_STATE_RESTART_INIT;
2031 } else {
2032 ap->state = ANEG_STATE_DISABLE_LINK_OK;
2033 }
2034 break;
2035
2036 case ANEG_STATE_RESTART_INIT:
2037 ap->link_time = ap->cur_time;
2038 ap->flags &= ~(MR_NP_LOADED);
2039 ap->txconfig = 0;
2040 tw32(MAC_TX_AUTO_NEG, 0);
2041 tp->mac_mode |= MAC_MODE_SEND_CONFIGS;
2042 tw32_f(MAC_MODE, tp->mac_mode);
2043 udelay(40);
2044
2045 ret = ANEG_TIMER_ENAB;
2046 ap->state = ANEG_STATE_RESTART;
2047
2048 /* fallthru */
2049 case ANEG_STATE_RESTART:
2050 delta = ap->cur_time - ap->link_time;
2051 if (delta > ANEG_STATE_SETTLE_TIME) {
2052 ap->state = ANEG_STATE_ABILITY_DETECT_INIT;
2053 } else {
2054 ret = ANEG_TIMER_ENAB;
2055 }
2056 break;
2057
2058 case ANEG_STATE_DISABLE_LINK_OK:
2059 ret = ANEG_DONE;
2060 break;
2061
2062 case ANEG_STATE_ABILITY_DETECT_INIT:
2063 ap->flags &= ~(MR_TOGGLE_TX);
2064 ap->txconfig = (ANEG_CFG_FD | ANEG_CFG_PS1);
2065 tw32(MAC_TX_AUTO_NEG, ap->txconfig);
2066 tp->mac_mode |= MAC_MODE_SEND_CONFIGS;
2067 tw32_f(MAC_MODE, tp->mac_mode);
2068 udelay(40);
2069
2070 ap->state = ANEG_STATE_ABILITY_DETECT;
2071 break;
2072
2073 case ANEG_STATE_ABILITY_DETECT:
2074 if (ap->ability_match != 0 && ap->rxconfig != 0) {
2075 ap->state = ANEG_STATE_ACK_DETECT_INIT;
2076 }
2077 break;
2078
2079 case ANEG_STATE_ACK_DETECT_INIT:
2080 ap->txconfig |= ANEG_CFG_ACK;
2081 tw32(MAC_TX_AUTO_NEG, ap->txconfig);
2082 tp->mac_mode |= MAC_MODE_SEND_CONFIGS;
2083 tw32_f(MAC_MODE, tp->mac_mode);
2084 udelay(40);
2085
2086 ap->state = ANEG_STATE_ACK_DETECT;
2087
2088 /* fallthru */
2089 case ANEG_STATE_ACK_DETECT:
2090 if (ap->ack_match != 0) {
2091 if ((ap->rxconfig & ~ANEG_CFG_ACK) ==
2092 (ap->ability_match_cfg & ~ANEG_CFG_ACK)) {
2093 ap->state = ANEG_STATE_COMPLETE_ACK_INIT;
2094 } else {
2095 ap->state = ANEG_STATE_AN_ENABLE;
2096 }
2097 } else if (ap->ability_match != 0 &&
2098 ap->rxconfig == 0) {
2099 ap->state = ANEG_STATE_AN_ENABLE;
2100 }
2101 break;
2102
2103 case ANEG_STATE_COMPLETE_ACK_INIT:
2104 if (ap->rxconfig & ANEG_CFG_INVAL) {
2105 ret = ANEG_FAILED;
2106 break;
2107 }
2108 ap->flags &= ~(MR_LP_ADV_FULL_DUPLEX |
2109 MR_LP_ADV_HALF_DUPLEX |
2110 MR_LP_ADV_SYM_PAUSE |
2111 MR_LP_ADV_ASYM_PAUSE |
2112 MR_LP_ADV_REMOTE_FAULT1 |
2113 MR_LP_ADV_REMOTE_FAULT2 |
2114 MR_LP_ADV_NEXT_PAGE |
2115 MR_TOGGLE_RX |
2116 MR_NP_RX);
2117 if (ap->rxconfig & ANEG_CFG_FD)
2118 ap->flags |= MR_LP_ADV_FULL_DUPLEX;
2119 if (ap->rxconfig & ANEG_CFG_HD)
2120 ap->flags |= MR_LP_ADV_HALF_DUPLEX;
2121 if (ap->rxconfig & ANEG_CFG_PS1)
2122 ap->flags |= MR_LP_ADV_SYM_PAUSE;
2123 if (ap->rxconfig & ANEG_CFG_PS2)
2124 ap->flags |= MR_LP_ADV_ASYM_PAUSE;
2125 if (ap->rxconfig & ANEG_CFG_RF1)
2126 ap->flags |= MR_LP_ADV_REMOTE_FAULT1;
2127 if (ap->rxconfig & ANEG_CFG_RF2)
2128 ap->flags |= MR_LP_ADV_REMOTE_FAULT2;
2129 if (ap->rxconfig & ANEG_CFG_NP)
2130 ap->flags |= MR_LP_ADV_NEXT_PAGE;
2131
2132 ap->link_time = ap->cur_time;
2133
2134 ap->flags ^= (MR_TOGGLE_TX);
2135 if (ap->rxconfig & 0x0008)
2136 ap->flags |= MR_TOGGLE_RX;
2137 if (ap->rxconfig & ANEG_CFG_NP)
2138 ap->flags |= MR_NP_RX;
2139 ap->flags |= MR_PAGE_RX;
2140
2141 ap->state = ANEG_STATE_COMPLETE_ACK;
2142 ret = ANEG_TIMER_ENAB;
2143 break;
2144
2145 case ANEG_STATE_COMPLETE_ACK:
2146 if (ap->ability_match != 0 &&
2147 ap->rxconfig == 0) {
2148 ap->state = ANEG_STATE_AN_ENABLE;
2149 break;
2150 }
2151 delta = ap->cur_time - ap->link_time;
2152 if (delta > ANEG_STATE_SETTLE_TIME) {
2153 if (!(ap->flags & (MR_LP_ADV_NEXT_PAGE))) {
2154 ap->state = ANEG_STATE_IDLE_DETECT_INIT;
2155 } else {
2156 if ((ap->txconfig & ANEG_CFG_NP) == 0 &&
2157 !(ap->flags & MR_NP_RX)) {
2158 ap->state = ANEG_STATE_IDLE_DETECT_INIT;
2159 } else {
2160 ret = ANEG_FAILED;
2161 }
2162 }
2163 }
2164 break;
2165
2166 case ANEG_STATE_IDLE_DETECT_INIT:
2167 ap->link_time = ap->cur_time;
2168 tp->mac_mode &= ~MAC_MODE_SEND_CONFIGS;
2169 tw32_f(MAC_MODE, tp->mac_mode);
2170 udelay(40);
2171
2172 ap->state = ANEG_STATE_IDLE_DETECT;
2173 ret = ANEG_TIMER_ENAB;
2174 break;
2175
2176 case ANEG_STATE_IDLE_DETECT:
2177 if (ap->ability_match != 0 &&
2178 ap->rxconfig == 0) {
2179 ap->state = ANEG_STATE_AN_ENABLE;
2180 break;
2181 }
2182 delta = ap->cur_time - ap->link_time;
2183 if (delta > ANEG_STATE_SETTLE_TIME) {
2184 /* XXX another gem from the Broadcom driver :( */
2185 ap->state = ANEG_STATE_LINK_OK;
2186 }
2187 break;
2188
2189 case ANEG_STATE_LINK_OK:
2190 ap->flags |= (MR_AN_COMPLETE | MR_LINK_OK);
2191 ret = ANEG_DONE;
2192 break;
2193
2194 case ANEG_STATE_NEXT_PAGE_WAIT_INIT:
2195 /* ??? unimplemented */
2196 break;
2197
2198 case ANEG_STATE_NEXT_PAGE_WAIT:
2199 /* ??? unimplemented */
2200 break;
2201
2202 default:
2203 ret = ANEG_FAILED;
2204 break;
2205 };
2206
2207 return ret;
2208}
2209
2210static int fiber_autoneg(struct tg3 *tp, u32 *flags)
2211{
2212 int res = 0;
2213 struct tg3_fiber_aneginfo aninfo;
2214 int status = ANEG_FAILED;
2215 unsigned int tick;
2216 u32 tmp;
2217
2218 tw32_f(MAC_TX_AUTO_NEG, 0);
2219
2220 tmp = tp->mac_mode & ~MAC_MODE_PORT_MODE_MASK;
2221 tw32_f(MAC_MODE, tmp | MAC_MODE_PORT_MODE_GMII);
2222 udelay(40);
2223
2224 tw32_f(MAC_MODE, tp->mac_mode | MAC_MODE_SEND_CONFIGS);
2225 udelay(40);
2226
2227 memset(&aninfo, 0, sizeof(aninfo));
2228 aninfo.flags |= MR_AN_ENABLE;
2229 aninfo.state = ANEG_STATE_UNKNOWN;
2230 aninfo.cur_time = 0;
2231 tick = 0;
2232 while (++tick < 195000) {
2233 status = tg3_fiber_aneg_smachine(tp, &aninfo);
2234 if (status == ANEG_DONE || status == ANEG_FAILED)
2235 break;
2236
2237 udelay(1);
2238 }
2239
2240 tp->mac_mode &= ~MAC_MODE_SEND_CONFIGS;
2241 tw32_f(MAC_MODE, tp->mac_mode);
2242 udelay(40);
2243
2244 *flags = aninfo.flags;
2245
2246 if (status == ANEG_DONE &&
2247 (aninfo.flags & (MR_AN_COMPLETE | MR_LINK_OK |
2248 MR_LP_ADV_FULL_DUPLEX)))
2249 res = 1;
2250
2251 return res;
2252}
2253
2254static void tg3_init_bcm8002(struct tg3 *tp)
2255{
2256 u32 mac_status = tr32(MAC_STATUS);
2257 int i;
2258
2259 /* Reset when initting first time or we have a link. */
2260 if ((tp->tg3_flags & TG3_FLAG_INIT_COMPLETE) &&
2261 !(mac_status & MAC_STATUS_PCS_SYNCED))
2262 return;
2263
2264 /* Set PLL lock range. */
2265 tg3_writephy(tp, 0x16, 0x8007);
2266
2267 /* SW reset */
2268 tg3_writephy(tp, MII_BMCR, BMCR_RESET);
2269
2270 /* Wait for reset to complete. */
2271 /* XXX schedule_timeout() ... */
2272 for (i = 0; i < 500; i++)
2273 udelay(10);
2274
2275 /* Config mode; select PMA/Ch 1 regs. */
2276 tg3_writephy(tp, 0x10, 0x8411);
2277
2278 /* Enable auto-lock and comdet, select txclk for tx. */
2279 tg3_writephy(tp, 0x11, 0x0a10);
2280
2281 tg3_writephy(tp, 0x18, 0x00a0);
2282 tg3_writephy(tp, 0x16, 0x41ff);
2283
2284 /* Assert and deassert POR. */
2285 tg3_writephy(tp, 0x13, 0x0400);
2286 udelay(40);
2287 tg3_writephy(tp, 0x13, 0x0000);
2288
2289 tg3_writephy(tp, 0x11, 0x0a50);
2290 udelay(40);
2291 tg3_writephy(tp, 0x11, 0x0a10);
2292
2293 /* Wait for signal to stabilize */
2294 /* XXX schedule_timeout() ... */
2295 for (i = 0; i < 15000; i++)
2296 udelay(10);
2297
2298 /* Deselect the channel register so we can read the PHYID
2299 * later.
2300 */
2301 tg3_writephy(tp, 0x10, 0x8011);
2302}
2303
2304static int tg3_setup_fiber_hw_autoneg(struct tg3 *tp, u32 mac_status)
2305{
2306 u32 sg_dig_ctrl, sg_dig_status;
2307 u32 serdes_cfg, expected_sg_dig_ctrl;
2308 int workaround, port_a;
2309 int current_link_up;
2310
2311 serdes_cfg = 0;
2312 expected_sg_dig_ctrl = 0;
2313 workaround = 0;
2314 port_a = 1;
2315 current_link_up = 0;
2316
2317 if (tp->pci_chip_rev_id != CHIPREV_ID_5704_A0 &&
2318 tp->pci_chip_rev_id != CHIPREV_ID_5704_A1) {
2319 workaround = 1;
2320 if (tr32(TG3PCI_DUAL_MAC_CTRL) & DUAL_MAC_CTRL_ID)
2321 port_a = 0;
2322
2323 /* preserve bits 0-11,13,14 for signal pre-emphasis */
2324 /* preserve bits 20-23 for voltage regulator */
2325 serdes_cfg = tr32(MAC_SERDES_CFG) & 0x00f06fff;
2326 }
2327
2328 sg_dig_ctrl = tr32(SG_DIG_CTRL);
2329
2330 if (tp->link_config.autoneg != AUTONEG_ENABLE) {
2331 if (sg_dig_ctrl & (1 << 31)) {
2332 if (workaround) {
2333 u32 val = serdes_cfg;
2334
2335 if (port_a)
2336 val |= 0xc010000;
2337 else
2338 val |= 0x4010000;
2339 tw32_f(MAC_SERDES_CFG, val);
2340 }
2341 tw32_f(SG_DIG_CTRL, 0x01388400);
2342 }
2343 if (mac_status & MAC_STATUS_PCS_SYNCED) {
2344 tg3_setup_flow_control(tp, 0, 0);
2345 current_link_up = 1;
2346 }
2347 goto out;
2348 }
2349
2350 /* Want auto-negotiation. */
2351 expected_sg_dig_ctrl = 0x81388400;
2352
2353 /* Pause capability */
2354 expected_sg_dig_ctrl |= (1 << 11);
2355
2356 /* Asymettric pause */
2357 expected_sg_dig_ctrl |= (1 << 12);
2358
2359 if (sg_dig_ctrl != expected_sg_dig_ctrl) {
2360 if (workaround)
2361 tw32_f(MAC_SERDES_CFG, serdes_cfg | 0xc011000);
2362 tw32_f(SG_DIG_CTRL, expected_sg_dig_ctrl | (1 << 30));
2363 udelay(5);
2364 tw32_f(SG_DIG_CTRL, expected_sg_dig_ctrl);
2365
2366 tp->tg3_flags2 |= TG3_FLG2_PHY_JUST_INITTED;
2367 } else if (mac_status & (MAC_STATUS_PCS_SYNCED |
2368 MAC_STATUS_SIGNAL_DET)) {
2369 int i;
2370
2371 /* Giver time to negotiate (~200ms) */
2372 for (i = 0; i < 40000; i++) {
2373 sg_dig_status = tr32(SG_DIG_STATUS);
2374 if (sg_dig_status & (0x3))
2375 break;
2376 udelay(5);
2377 }
2378 mac_status = tr32(MAC_STATUS);
2379
2380 if ((sg_dig_status & (1 << 1)) &&
2381 (mac_status & MAC_STATUS_PCS_SYNCED)) {
2382 u32 local_adv, remote_adv;
2383
2384 local_adv = ADVERTISE_PAUSE_CAP;
2385 remote_adv = 0;
2386 if (sg_dig_status & (1 << 19))
2387 remote_adv |= LPA_PAUSE_CAP;
2388 if (sg_dig_status & (1 << 20))
2389 remote_adv |= LPA_PAUSE_ASYM;
2390
2391 tg3_setup_flow_control(tp, local_adv, remote_adv);
2392 current_link_up = 1;
2393 tp->tg3_flags2 &= ~TG3_FLG2_PHY_JUST_INITTED;
2394 } else if (!(sg_dig_status & (1 << 1))) {
2395 if (tp->tg3_flags2 & TG3_FLG2_PHY_JUST_INITTED)
2396 tp->tg3_flags2 &= ~TG3_FLG2_PHY_JUST_INITTED;
2397 else {
2398 if (workaround) {
2399 u32 val = serdes_cfg;
2400
2401 if (port_a)
2402 val |= 0xc010000;
2403 else
2404 val |= 0x4010000;
2405
2406 tw32_f(MAC_SERDES_CFG, val);
2407 }
2408
2409 tw32_f(SG_DIG_CTRL, 0x01388400);
2410 udelay(40);
2411
2412 /* Link parallel detection - link is up */
2413 /* only if we have PCS_SYNC and not */
2414 /* receiving config code words */
2415 mac_status = tr32(MAC_STATUS);
2416 if ((mac_status & MAC_STATUS_PCS_SYNCED) &&
2417 !(mac_status & MAC_STATUS_RCVD_CFG)) {
2418 tg3_setup_flow_control(tp, 0, 0);
2419 current_link_up = 1;
2420 }
2421 }
2422 }
2423 }
2424
2425out:
2426 return current_link_up;
2427}
2428
2429static int tg3_setup_fiber_by_hand(struct tg3 *tp, u32 mac_status)
2430{
2431 int current_link_up = 0;
2432
2433 if (!(mac_status & MAC_STATUS_PCS_SYNCED)) {
2434 tp->tg3_flags &= ~TG3_FLAG_GOT_SERDES_FLOWCTL;
2435 goto out;
2436 }
2437
2438 if (tp->link_config.autoneg == AUTONEG_ENABLE) {
2439 u32 flags;
2440 int i;
2441
2442 if (fiber_autoneg(tp, &flags)) {
2443 u32 local_adv, remote_adv;
2444
2445 local_adv = ADVERTISE_PAUSE_CAP;
2446 remote_adv = 0;
2447 if (flags & MR_LP_ADV_SYM_PAUSE)
2448 remote_adv |= LPA_PAUSE_CAP;
2449 if (flags & MR_LP_ADV_ASYM_PAUSE)
2450 remote_adv |= LPA_PAUSE_ASYM;
2451
2452 tg3_setup_flow_control(tp, local_adv, remote_adv);
2453
2454 tp->tg3_flags |= TG3_FLAG_GOT_SERDES_FLOWCTL;
2455 current_link_up = 1;
2456 }
2457 for (i = 0; i < 30; i++) {
2458 udelay(20);
2459 tw32_f(MAC_STATUS,
2460 (MAC_STATUS_SYNC_CHANGED |
2461 MAC_STATUS_CFG_CHANGED));
2462 udelay(40);
2463 if ((tr32(MAC_STATUS) &
2464 (MAC_STATUS_SYNC_CHANGED |
2465 MAC_STATUS_CFG_CHANGED)) == 0)
2466 break;
2467 }
2468
2469 mac_status = tr32(MAC_STATUS);
2470 if (current_link_up == 0 &&
2471 (mac_status & MAC_STATUS_PCS_SYNCED) &&
2472 !(mac_status & MAC_STATUS_RCVD_CFG))
2473 current_link_up = 1;
2474 } else {
2475 /* Forcing 1000FD link up. */
2476 current_link_up = 1;
2477 tp->tg3_flags |= TG3_FLAG_GOT_SERDES_FLOWCTL;
2478
2479 tw32_f(MAC_MODE, (tp->mac_mode | MAC_MODE_SEND_CONFIGS));
2480 udelay(40);
2481 }
2482
2483out:
2484 return current_link_up;
2485}
2486
2487static int tg3_setup_fiber_phy(struct tg3 *tp, int force_reset)
2488{
2489 u32 orig_pause_cfg;
2490 u16 orig_active_speed;
2491 u8 orig_active_duplex;
2492 u32 mac_status;
2493 int current_link_up;
2494 int i;
2495
2496 orig_pause_cfg =
2497 (tp->tg3_flags & (TG3_FLAG_RX_PAUSE |
2498 TG3_FLAG_TX_PAUSE));
2499 orig_active_speed = tp->link_config.active_speed;
2500 orig_active_duplex = tp->link_config.active_duplex;
2501
2502 if (!(tp->tg3_flags2 & TG3_FLG2_HW_AUTONEG) &&
2503 netif_carrier_ok(tp->dev) &&
2504 (tp->tg3_flags & TG3_FLAG_INIT_COMPLETE)) {
2505 mac_status = tr32(MAC_STATUS);
2506 mac_status &= (MAC_STATUS_PCS_SYNCED |
2507 MAC_STATUS_SIGNAL_DET |
2508 MAC_STATUS_CFG_CHANGED |
2509 MAC_STATUS_RCVD_CFG);
2510 if (mac_status == (MAC_STATUS_PCS_SYNCED |
2511 MAC_STATUS_SIGNAL_DET)) {
2512 tw32_f(MAC_STATUS, (MAC_STATUS_SYNC_CHANGED |
2513 MAC_STATUS_CFG_CHANGED));
2514 return 0;
2515 }
2516 }
2517
2518 tw32_f(MAC_TX_AUTO_NEG, 0);
2519
2520 tp->mac_mode &= ~(MAC_MODE_PORT_MODE_MASK | MAC_MODE_HALF_DUPLEX);
2521 tp->mac_mode |= MAC_MODE_PORT_MODE_TBI;
2522 tw32_f(MAC_MODE, tp->mac_mode);
2523 udelay(40);
2524
2525 if (tp->phy_id == PHY_ID_BCM8002)
2526 tg3_init_bcm8002(tp);
2527
2528 /* Enable link change event even when serdes polling. */
2529 tw32_f(MAC_EVENT, MAC_EVENT_LNKSTATE_CHANGED);
2530 udelay(40);
2531
2532 current_link_up = 0;
2533 mac_status = tr32(MAC_STATUS);
2534
2535 if (tp->tg3_flags2 & TG3_FLG2_HW_AUTONEG)
2536 current_link_up = tg3_setup_fiber_hw_autoneg(tp, mac_status);
2537 else
2538 current_link_up = tg3_setup_fiber_by_hand(tp, mac_status);
2539
2540 tp->mac_mode &= ~MAC_MODE_LINK_POLARITY;
2541 tw32_f(MAC_MODE, tp->mac_mode);
2542 udelay(40);
2543
2544 tp->hw_status->status =
2545 (SD_STATUS_UPDATED |
2546 (tp->hw_status->status & ~SD_STATUS_LINK_CHG));
2547
2548 for (i = 0; i < 100; i++) {
2549 tw32_f(MAC_STATUS, (MAC_STATUS_SYNC_CHANGED |
2550 MAC_STATUS_CFG_CHANGED));
2551 udelay(5);
2552 if ((tr32(MAC_STATUS) & (MAC_STATUS_SYNC_CHANGED |
2553 MAC_STATUS_CFG_CHANGED)) == 0)
2554 break;
2555 }
2556
2557 mac_status = tr32(MAC_STATUS);
2558 if ((mac_status & MAC_STATUS_PCS_SYNCED) == 0) {
2559 current_link_up = 0;
2560 if (tp->link_config.autoneg == AUTONEG_ENABLE) {
2561 tw32_f(MAC_MODE, (tp->mac_mode |
2562 MAC_MODE_SEND_CONFIGS));
2563 udelay(1);
2564 tw32_f(MAC_MODE, tp->mac_mode);
2565 }
2566 }
2567
2568 if (current_link_up == 1) {
2569 tp->link_config.active_speed = SPEED_1000;
2570 tp->link_config.active_duplex = DUPLEX_FULL;
2571 tw32(MAC_LED_CTRL, (tp->led_ctrl |
2572 LED_CTRL_LNKLED_OVERRIDE |
2573 LED_CTRL_1000MBPS_ON));
2574 } else {
2575 tp->link_config.active_speed = SPEED_INVALID;
2576 tp->link_config.active_duplex = DUPLEX_INVALID;
2577 tw32(MAC_LED_CTRL, (tp->led_ctrl |
2578 LED_CTRL_LNKLED_OVERRIDE |
2579 LED_CTRL_TRAFFIC_OVERRIDE));
2580 }
2581
2582 if (current_link_up != netif_carrier_ok(tp->dev)) {
2583 if (current_link_up)
2584 netif_carrier_on(tp->dev);
2585 else
2586 netif_carrier_off(tp->dev);
2587 tg3_link_report(tp);
2588 } else {
2589 u32 now_pause_cfg =
2590 tp->tg3_flags & (TG3_FLAG_RX_PAUSE |
2591 TG3_FLAG_TX_PAUSE);
2592 if (orig_pause_cfg != now_pause_cfg ||
2593 orig_active_speed != tp->link_config.active_speed ||
2594 orig_active_duplex != tp->link_config.active_duplex)
2595 tg3_link_report(tp);
2596 }
2597
2598 return 0;
2599}
2600
747e8f8b
MC
2601static int tg3_setup_fiber_mii_phy(struct tg3 *tp, int force_reset)
2602{
2603 int current_link_up, err = 0;
2604 u32 bmsr, bmcr;
2605 u16 current_speed;
2606 u8 current_duplex;
2607
2608 tp->mac_mode |= MAC_MODE_PORT_MODE_GMII;
2609 tw32_f(MAC_MODE, tp->mac_mode);
2610 udelay(40);
2611
2612 tw32(MAC_EVENT, 0);
2613
2614 tw32_f(MAC_STATUS,
2615 (MAC_STATUS_SYNC_CHANGED |
2616 MAC_STATUS_CFG_CHANGED |
2617 MAC_STATUS_MI_COMPLETION |
2618 MAC_STATUS_LNKSTATE_CHANGED));
2619 udelay(40);
2620
2621 if (force_reset)
2622 tg3_phy_reset(tp);
2623
2624 current_link_up = 0;
2625 current_speed = SPEED_INVALID;
2626 current_duplex = DUPLEX_INVALID;
2627
2628 err |= tg3_readphy(tp, MII_BMSR, &bmsr);
2629 err |= tg3_readphy(tp, MII_BMSR, &bmsr);
2630
2631 err |= tg3_readphy(tp, MII_BMCR, &bmcr);
2632
2633 if ((tp->link_config.autoneg == AUTONEG_ENABLE) && !force_reset &&
2634 (tp->tg3_flags2 & TG3_FLG2_PARALLEL_DETECT)) {
2635 /* do nothing, just check for link up at the end */
2636 } else if (tp->link_config.autoneg == AUTONEG_ENABLE) {
2637 u32 adv, new_adv;
2638
2639 err |= tg3_readphy(tp, MII_ADVERTISE, &adv);
2640 new_adv = adv & ~(ADVERTISE_1000XFULL | ADVERTISE_1000XHALF |
2641 ADVERTISE_1000XPAUSE |
2642 ADVERTISE_1000XPSE_ASYM |
2643 ADVERTISE_SLCT);
2644
2645 /* Always advertise symmetric PAUSE just like copper */
2646 new_adv |= ADVERTISE_1000XPAUSE;
2647
2648 if (tp->link_config.advertising & ADVERTISED_1000baseT_Half)
2649 new_adv |= ADVERTISE_1000XHALF;
2650 if (tp->link_config.advertising & ADVERTISED_1000baseT_Full)
2651 new_adv |= ADVERTISE_1000XFULL;
2652
2653 if ((new_adv != adv) || !(bmcr & BMCR_ANENABLE)) {
2654 tg3_writephy(tp, MII_ADVERTISE, new_adv);
2655 bmcr |= BMCR_ANENABLE | BMCR_ANRESTART;
2656 tg3_writephy(tp, MII_BMCR, bmcr);
2657
2658 tw32_f(MAC_EVENT, MAC_EVENT_LNKSTATE_CHANGED);
2659 tp->tg3_flags2 |= TG3_FLG2_PHY_JUST_INITTED;
2660 tp->tg3_flags2 &= ~TG3_FLG2_PARALLEL_DETECT;
2661
2662 return err;
2663 }
2664 } else {
2665 u32 new_bmcr;
2666
2667 bmcr &= ~BMCR_SPEED1000;
2668 new_bmcr = bmcr & ~(BMCR_ANENABLE | BMCR_FULLDPLX);
2669
2670 if (tp->link_config.duplex == DUPLEX_FULL)
2671 new_bmcr |= BMCR_FULLDPLX;
2672
2673 if (new_bmcr != bmcr) {
2674 /* BMCR_SPEED1000 is a reserved bit that needs
2675 * to be set on write.
2676 */
2677 new_bmcr |= BMCR_SPEED1000;
2678
2679 /* Force a linkdown */
2680 if (netif_carrier_ok(tp->dev)) {
2681 u32 adv;
2682
2683 err |= tg3_readphy(tp, MII_ADVERTISE, &adv);
2684 adv &= ~(ADVERTISE_1000XFULL |
2685 ADVERTISE_1000XHALF |
2686 ADVERTISE_SLCT);
2687 tg3_writephy(tp, MII_ADVERTISE, adv);
2688 tg3_writephy(tp, MII_BMCR, bmcr |
2689 BMCR_ANRESTART |
2690 BMCR_ANENABLE);
2691 udelay(10);
2692 netif_carrier_off(tp->dev);
2693 }
2694 tg3_writephy(tp, MII_BMCR, new_bmcr);
2695 bmcr = new_bmcr;
2696 err |= tg3_readphy(tp, MII_BMSR, &bmsr);
2697 err |= tg3_readphy(tp, MII_BMSR, &bmsr);
2698 tp->tg3_flags2 &= ~TG3_FLG2_PARALLEL_DETECT;
2699 }
2700 }
2701
2702 if (bmsr & BMSR_LSTATUS) {
2703 current_speed = SPEED_1000;
2704 current_link_up = 1;
2705 if (bmcr & BMCR_FULLDPLX)
2706 current_duplex = DUPLEX_FULL;
2707 else
2708 current_duplex = DUPLEX_HALF;
2709
2710 if (bmcr & BMCR_ANENABLE) {
2711 u32 local_adv, remote_adv, common;
2712
2713 err |= tg3_readphy(tp, MII_ADVERTISE, &local_adv);
2714 err |= tg3_readphy(tp, MII_LPA, &remote_adv);
2715 common = local_adv & remote_adv;
2716 if (common & (ADVERTISE_1000XHALF |
2717 ADVERTISE_1000XFULL)) {
2718 if (common & ADVERTISE_1000XFULL)
2719 current_duplex = DUPLEX_FULL;
2720 else
2721 current_duplex = DUPLEX_HALF;
2722
2723 tg3_setup_flow_control(tp, local_adv,
2724 remote_adv);
2725 }
2726 else
2727 current_link_up = 0;
2728 }
2729 }
2730
2731 tp->mac_mode &= ~MAC_MODE_HALF_DUPLEX;
2732 if (tp->link_config.active_duplex == DUPLEX_HALF)
2733 tp->mac_mode |= MAC_MODE_HALF_DUPLEX;
2734
2735 tw32_f(MAC_MODE, tp->mac_mode);
2736 udelay(40);
2737
2738 tw32_f(MAC_EVENT, MAC_EVENT_LNKSTATE_CHANGED);
2739
2740 tp->link_config.active_speed = current_speed;
2741 tp->link_config.active_duplex = current_duplex;
2742
2743 if (current_link_up != netif_carrier_ok(tp->dev)) {
2744 if (current_link_up)
2745 netif_carrier_on(tp->dev);
2746 else {
2747 netif_carrier_off(tp->dev);
2748 tp->tg3_flags2 &= ~TG3_FLG2_PARALLEL_DETECT;
2749 }
2750 tg3_link_report(tp);
2751 }
2752 return err;
2753}
2754
2755static void tg3_serdes_parallel_detect(struct tg3 *tp)
2756{
2757 if (tp->tg3_flags2 & TG3_FLG2_PHY_JUST_INITTED) {
2758 /* Give autoneg time to complete. */
2759 tp->tg3_flags2 &= ~TG3_FLG2_PHY_JUST_INITTED;
2760 return;
2761 }
2762 if (!netif_carrier_ok(tp->dev) &&
2763 (tp->link_config.autoneg == AUTONEG_ENABLE)) {
2764 u32 bmcr;
2765
2766 tg3_readphy(tp, MII_BMCR, &bmcr);
2767 if (bmcr & BMCR_ANENABLE) {
2768 u32 phy1, phy2;
2769
2770 /* Select shadow register 0x1f */
2771 tg3_writephy(tp, 0x1c, 0x7c00);
2772 tg3_readphy(tp, 0x1c, &phy1);
2773
2774 /* Select expansion interrupt status register */
2775 tg3_writephy(tp, 0x17, 0x0f01);
2776 tg3_readphy(tp, 0x15, &phy2);
2777 tg3_readphy(tp, 0x15, &phy2);
2778
2779 if ((phy1 & 0x10) && !(phy2 & 0x20)) {
2780 /* We have signal detect and not receiving
2781 * config code words, link is up by parallel
2782 * detection.
2783 */
2784
2785 bmcr &= ~BMCR_ANENABLE;
2786 bmcr |= BMCR_SPEED1000 | BMCR_FULLDPLX;
2787 tg3_writephy(tp, MII_BMCR, bmcr);
2788 tp->tg3_flags2 |= TG3_FLG2_PARALLEL_DETECT;
2789 }
2790 }
2791 }
2792 else if (netif_carrier_ok(tp->dev) &&
2793 (tp->link_config.autoneg == AUTONEG_ENABLE) &&
2794 (tp->tg3_flags2 & TG3_FLG2_PARALLEL_DETECT)) {
2795 u32 phy2;
2796
2797 /* Select expansion interrupt status register */
2798 tg3_writephy(tp, 0x17, 0x0f01);
2799 tg3_readphy(tp, 0x15, &phy2);
2800 if (phy2 & 0x20) {
2801 u32 bmcr;
2802
2803 /* Config code words received, turn on autoneg. */
2804 tg3_readphy(tp, MII_BMCR, &bmcr);
2805 tg3_writephy(tp, MII_BMCR, bmcr | BMCR_ANENABLE);
2806
2807 tp->tg3_flags2 &= ~TG3_FLG2_PARALLEL_DETECT;
2808
2809 }
2810 }
2811}
2812
1da177e4
LT
2813static int tg3_setup_phy(struct tg3 *tp, int force_reset)
2814{
2815 int err;
2816
2817 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES) {
2818 err = tg3_setup_fiber_phy(tp, force_reset);
747e8f8b
MC
2819 } else if (tp->tg3_flags2 & TG3_FLG2_MII_SERDES) {
2820 err = tg3_setup_fiber_mii_phy(tp, force_reset);
1da177e4
LT
2821 } else {
2822 err = tg3_setup_copper_phy(tp, force_reset);
2823 }
2824
2825 if (tp->link_config.active_speed == SPEED_1000 &&
2826 tp->link_config.active_duplex == DUPLEX_HALF)
2827 tw32(MAC_TX_LENGTHS,
2828 ((2 << TX_LENGTHS_IPG_CRS_SHIFT) |
2829 (6 << TX_LENGTHS_IPG_SHIFT) |
2830 (0xff << TX_LENGTHS_SLOT_TIME_SHIFT)));
2831 else
2832 tw32(MAC_TX_LENGTHS,
2833 ((2 << TX_LENGTHS_IPG_CRS_SHIFT) |
2834 (6 << TX_LENGTHS_IPG_SHIFT) |
2835 (32 << TX_LENGTHS_SLOT_TIME_SHIFT)));
2836
2837 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
2838 if (netif_carrier_ok(tp->dev)) {
2839 tw32(HOSTCC_STAT_COAL_TICKS,
15f9850d 2840 tp->coal.stats_block_coalesce_usecs);
1da177e4
LT
2841 } else {
2842 tw32(HOSTCC_STAT_COAL_TICKS, 0);
2843 }
2844 }
2845
2846 return err;
2847}
2848
2849/* Tigon3 never reports partial packet sends. So we do not
2850 * need special logic to handle SKBs that have not had all
2851 * of their frags sent yet, like SunGEM does.
2852 */
2853static void tg3_tx(struct tg3 *tp)
2854{
2855 u32 hw_idx = tp->hw_status->idx[0].tx_consumer;
2856 u32 sw_idx = tp->tx_cons;
2857
2858 while (sw_idx != hw_idx) {
2859 struct tx_ring_info *ri = &tp->tx_buffers[sw_idx];
2860 struct sk_buff *skb = ri->skb;
2861 int i;
2862
2863 if (unlikely(skb == NULL))
2864 BUG();
2865
2866 pci_unmap_single(tp->pdev,
2867 pci_unmap_addr(ri, mapping),
2868 skb_headlen(skb),
2869 PCI_DMA_TODEVICE);
2870
2871 ri->skb = NULL;
2872
2873 sw_idx = NEXT_TX(sw_idx);
2874
2875 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
2876 if (unlikely(sw_idx == hw_idx))
2877 BUG();
2878
2879 ri = &tp->tx_buffers[sw_idx];
2880 if (unlikely(ri->skb != NULL))
2881 BUG();
2882
2883 pci_unmap_page(tp->pdev,
2884 pci_unmap_addr(ri, mapping),
2885 skb_shinfo(skb)->frags[i].size,
2886 PCI_DMA_TODEVICE);
2887
2888 sw_idx = NEXT_TX(sw_idx);
2889 }
2890
f47c11ee 2891 dev_kfree_skb(skb);
1da177e4
LT
2892 }
2893
2894 tp->tx_cons = sw_idx;
2895
51b91468
MC
2896 if (unlikely(netif_queue_stopped(tp->dev))) {
2897 spin_lock(&tp->tx_lock);
2898 if (netif_queue_stopped(tp->dev) &&
2899 (TX_BUFFS_AVAIL(tp) > TG3_TX_WAKEUP_THRESH))
2900 netif_wake_queue(tp->dev);
2901 spin_unlock(&tp->tx_lock);
2902 }
1da177e4
LT
2903}
2904
2905/* Returns size of skb allocated or < 0 on error.
2906 *
2907 * We only need to fill in the address because the other members
2908 * of the RX descriptor are invariant, see tg3_init_rings.
2909 *
2910 * Note the purposeful assymetry of cpu vs. chip accesses. For
2911 * posting buffers we only dirty the first cache line of the RX
2912 * descriptor (containing the address). Whereas for the RX status
2913 * buffers the cpu only reads the last cacheline of the RX descriptor
2914 * (to fetch the error flags, vlan tag, checksum, and opaque cookie).
2915 */
2916static int tg3_alloc_rx_skb(struct tg3 *tp, u32 opaque_key,
2917 int src_idx, u32 dest_idx_unmasked)
2918{
2919 struct tg3_rx_buffer_desc *desc;
2920 struct ring_info *map, *src_map;
2921 struct sk_buff *skb;
2922 dma_addr_t mapping;
2923 int skb_size, dest_idx;
2924
2925 src_map = NULL;
2926 switch (opaque_key) {
2927 case RXD_OPAQUE_RING_STD:
2928 dest_idx = dest_idx_unmasked % TG3_RX_RING_SIZE;
2929 desc = &tp->rx_std[dest_idx];
2930 map = &tp->rx_std_buffers[dest_idx];
2931 if (src_idx >= 0)
2932 src_map = &tp->rx_std_buffers[src_idx];
7e72aad4 2933 skb_size = tp->rx_pkt_buf_sz;
1da177e4
LT
2934 break;
2935
2936 case RXD_OPAQUE_RING_JUMBO:
2937 dest_idx = dest_idx_unmasked % TG3_RX_JUMBO_RING_SIZE;
2938 desc = &tp->rx_jumbo[dest_idx];
2939 map = &tp->rx_jumbo_buffers[dest_idx];
2940 if (src_idx >= 0)
2941 src_map = &tp->rx_jumbo_buffers[src_idx];
2942 skb_size = RX_JUMBO_PKT_BUF_SZ;
2943 break;
2944
2945 default:
2946 return -EINVAL;
2947 };
2948
2949 /* Do not overwrite any of the map or rp information
2950 * until we are sure we can commit to a new buffer.
2951 *
2952 * Callers depend upon this behavior and assume that
2953 * we leave everything unchanged if we fail.
2954 */
2955 skb = dev_alloc_skb(skb_size);
2956 if (skb == NULL)
2957 return -ENOMEM;
2958
2959 skb->dev = tp->dev;
2960 skb_reserve(skb, tp->rx_offset);
2961
2962 mapping = pci_map_single(tp->pdev, skb->data,
2963 skb_size - tp->rx_offset,
2964 PCI_DMA_FROMDEVICE);
2965
2966 map->skb = skb;
2967 pci_unmap_addr_set(map, mapping, mapping);
2968
2969 if (src_map != NULL)
2970 src_map->skb = NULL;
2971
2972 desc->addr_hi = ((u64)mapping >> 32);
2973 desc->addr_lo = ((u64)mapping & 0xffffffff);
2974
2975 return skb_size;
2976}
2977
2978/* We only need to move over in the address because the other
2979 * members of the RX descriptor are invariant. See notes above
2980 * tg3_alloc_rx_skb for full details.
2981 */
2982static void tg3_recycle_rx(struct tg3 *tp, u32 opaque_key,
2983 int src_idx, u32 dest_idx_unmasked)
2984{
2985 struct tg3_rx_buffer_desc *src_desc, *dest_desc;
2986 struct ring_info *src_map, *dest_map;
2987 int dest_idx;
2988
2989 switch (opaque_key) {
2990 case RXD_OPAQUE_RING_STD:
2991 dest_idx = dest_idx_unmasked % TG3_RX_RING_SIZE;
2992 dest_desc = &tp->rx_std[dest_idx];
2993 dest_map = &tp->rx_std_buffers[dest_idx];
2994 src_desc = &tp->rx_std[src_idx];
2995 src_map = &tp->rx_std_buffers[src_idx];
2996 break;
2997
2998 case RXD_OPAQUE_RING_JUMBO:
2999 dest_idx = dest_idx_unmasked % TG3_RX_JUMBO_RING_SIZE;
3000 dest_desc = &tp->rx_jumbo[dest_idx];
3001 dest_map = &tp->rx_jumbo_buffers[dest_idx];
3002 src_desc = &tp->rx_jumbo[src_idx];
3003 src_map = &tp->rx_jumbo_buffers[src_idx];
3004 break;
3005
3006 default:
3007 return;
3008 };
3009
3010 dest_map->skb = src_map->skb;
3011 pci_unmap_addr_set(dest_map, mapping,
3012 pci_unmap_addr(src_map, mapping));
3013 dest_desc->addr_hi = src_desc->addr_hi;
3014 dest_desc->addr_lo = src_desc->addr_lo;
3015
3016 src_map->skb = NULL;
3017}
3018
3019#if TG3_VLAN_TAG_USED
3020static int tg3_vlan_rx(struct tg3 *tp, struct sk_buff *skb, u16 vlan_tag)
3021{
3022 return vlan_hwaccel_receive_skb(skb, tp->vlgrp, vlan_tag);
3023}
3024#endif
3025
3026/* The RX ring scheme is composed of multiple rings which post fresh
3027 * buffers to the chip, and one special ring the chip uses to report
3028 * status back to the host.
3029 *
3030 * The special ring reports the status of received packets to the
3031 * host. The chip does not write into the original descriptor the
3032 * RX buffer was obtained from. The chip simply takes the original
3033 * descriptor as provided by the host, updates the status and length
3034 * field, then writes this into the next status ring entry.
3035 *
3036 * Each ring the host uses to post buffers to the chip is described
3037 * by a TG3_BDINFO entry in the chips SRAM area. When a packet arrives,
3038 * it is first placed into the on-chip ram. When the packet's length
3039 * is known, it walks down the TG3_BDINFO entries to select the ring.
3040 * Each TG3_BDINFO specifies a MAXLEN field and the first TG3_BDINFO
3041 * which is within the range of the new packet's length is chosen.
3042 *
3043 * The "separate ring for rx status" scheme may sound queer, but it makes
3044 * sense from a cache coherency perspective. If only the host writes
3045 * to the buffer post rings, and only the chip writes to the rx status
3046 * rings, then cache lines never move beyond shared-modified state.
3047 * If both the host and chip were to write into the same ring, cache line
3048 * eviction could occur since both entities want it in an exclusive state.
3049 */
3050static int tg3_rx(struct tg3 *tp, int budget)
3051{
3052 u32 work_mask;
483ba50b
MC
3053 u32 sw_idx = tp->rx_rcb_ptr;
3054 u16 hw_idx;
1da177e4
LT
3055 int received;
3056
3057 hw_idx = tp->hw_status->idx[0].rx_producer;
3058 /*
3059 * We need to order the read of hw_idx and the read of
3060 * the opaque cookie.
3061 */
3062 rmb();
1da177e4
LT
3063 work_mask = 0;
3064 received = 0;
3065 while (sw_idx != hw_idx && budget > 0) {
3066 struct tg3_rx_buffer_desc *desc = &tp->rx_rcb[sw_idx];
3067 unsigned int len;
3068 struct sk_buff *skb;
3069 dma_addr_t dma_addr;
3070 u32 opaque_key, desc_idx, *post_ptr;
3071
3072 desc_idx = desc->opaque & RXD_OPAQUE_INDEX_MASK;
3073 opaque_key = desc->opaque & RXD_OPAQUE_RING_MASK;
3074 if (opaque_key == RXD_OPAQUE_RING_STD) {
3075 dma_addr = pci_unmap_addr(&tp->rx_std_buffers[desc_idx],
3076 mapping);
3077 skb = tp->rx_std_buffers[desc_idx].skb;
3078 post_ptr = &tp->rx_std_ptr;
3079 } else if (opaque_key == RXD_OPAQUE_RING_JUMBO) {
3080 dma_addr = pci_unmap_addr(&tp->rx_jumbo_buffers[desc_idx],
3081 mapping);
3082 skb = tp->rx_jumbo_buffers[desc_idx].skb;
3083 post_ptr = &tp->rx_jumbo_ptr;
3084 }
3085 else {
3086 goto next_pkt_nopost;
3087 }
3088
3089 work_mask |= opaque_key;
3090
3091 if ((desc->err_vlan & RXD_ERR_MASK) != 0 &&
3092 (desc->err_vlan != RXD_ERR_ODD_NIBBLE_RCVD_MII)) {
3093 drop_it:
3094 tg3_recycle_rx(tp, opaque_key,
3095 desc_idx, *post_ptr);
3096 drop_it_no_recycle:
3097 /* Other statistics kept track of by card. */
3098 tp->net_stats.rx_dropped++;
3099 goto next_pkt;
3100 }
3101
3102 len = ((desc->idx_len & RXD_LEN_MASK) >> RXD_LEN_SHIFT) - 4; /* omit crc */
3103
3104 if (len > RX_COPY_THRESHOLD
3105 && tp->rx_offset == 2
3106 /* rx_offset != 2 iff this is a 5701 card running
3107 * in PCI-X mode [see tg3_get_invariants()] */
3108 ) {
3109 int skb_size;
3110
3111 skb_size = tg3_alloc_rx_skb(tp, opaque_key,
3112 desc_idx, *post_ptr);
3113 if (skb_size < 0)
3114 goto drop_it;
3115
3116 pci_unmap_single(tp->pdev, dma_addr,
3117 skb_size - tp->rx_offset,
3118 PCI_DMA_FROMDEVICE);
3119
3120 skb_put(skb, len);
3121 } else {
3122 struct sk_buff *copy_skb;
3123
3124 tg3_recycle_rx(tp, opaque_key,
3125 desc_idx, *post_ptr);
3126
3127 copy_skb = dev_alloc_skb(len + 2);
3128 if (copy_skb == NULL)
3129 goto drop_it_no_recycle;
3130
3131 copy_skb->dev = tp->dev;
3132 skb_reserve(copy_skb, 2);
3133 skb_put(copy_skb, len);
3134 pci_dma_sync_single_for_cpu(tp->pdev, dma_addr, len, PCI_DMA_FROMDEVICE);
3135 memcpy(copy_skb->data, skb->data, len);
3136 pci_dma_sync_single_for_device(tp->pdev, dma_addr, len, PCI_DMA_FROMDEVICE);
3137
3138 /* We'll reuse the original ring buffer. */
3139 skb = copy_skb;
3140 }
3141
3142 if ((tp->tg3_flags & TG3_FLAG_RX_CHECKSUMS) &&
3143 (desc->type_flags & RXD_FLAG_TCPUDP_CSUM) &&
3144 (((desc->ip_tcp_csum & RXD_TCPCSUM_MASK)
3145 >> RXD_TCPCSUM_SHIFT) == 0xffff))
3146 skb->ip_summed = CHECKSUM_UNNECESSARY;
3147 else
3148 skb->ip_summed = CHECKSUM_NONE;
3149
3150 skb->protocol = eth_type_trans(skb, tp->dev);
3151#if TG3_VLAN_TAG_USED
3152 if (tp->vlgrp != NULL &&
3153 desc->type_flags & RXD_FLAG_VLAN) {
3154 tg3_vlan_rx(tp, skb,
3155 desc->err_vlan & RXD_VLAN_MASK);
3156 } else
3157#endif
3158 netif_receive_skb(skb);
3159
3160 tp->dev->last_rx = jiffies;
3161 received++;
3162 budget--;
3163
3164next_pkt:
3165 (*post_ptr)++;
3166next_pkt_nopost:
483ba50b
MC
3167 sw_idx++;
3168 sw_idx %= TG3_RX_RCB_RING_SIZE(tp);
52f6d697
MC
3169
3170 /* Refresh hw_idx to see if there is new work */
3171 if (sw_idx == hw_idx) {
3172 hw_idx = tp->hw_status->idx[0].rx_producer;
3173 rmb();
3174 }
1da177e4
LT
3175 }
3176
3177 /* ACK the status ring. */
483ba50b
MC
3178 tp->rx_rcb_ptr = sw_idx;
3179 tw32_rx_mbox(MAILBOX_RCVRET_CON_IDX_0 + TG3_64BIT_REG_LOW, sw_idx);
1da177e4
LT
3180
3181 /* Refill RX ring(s). */
3182 if (work_mask & RXD_OPAQUE_RING_STD) {
3183 sw_idx = tp->rx_std_ptr % TG3_RX_RING_SIZE;
3184 tw32_rx_mbox(MAILBOX_RCV_STD_PROD_IDX + TG3_64BIT_REG_LOW,
3185 sw_idx);
3186 }
3187 if (work_mask & RXD_OPAQUE_RING_JUMBO) {
3188 sw_idx = tp->rx_jumbo_ptr % TG3_RX_JUMBO_RING_SIZE;
3189 tw32_rx_mbox(MAILBOX_RCV_JUMBO_PROD_IDX + TG3_64BIT_REG_LOW,
3190 sw_idx);
3191 }
3192 mmiowb();
3193
3194 return received;
3195}
3196
3197static int tg3_poll(struct net_device *netdev, int *budget)
3198{
3199 struct tg3 *tp = netdev_priv(netdev);
3200 struct tg3_hw_status *sblk = tp->hw_status;
1da177e4
LT
3201 int done;
3202
1da177e4
LT
3203 /* handle link change and other phy events */
3204 if (!(tp->tg3_flags &
3205 (TG3_FLAG_USE_LINKCHG_REG |
3206 TG3_FLAG_POLL_SERDES))) {
3207 if (sblk->status & SD_STATUS_LINK_CHG) {
3208 sblk->status = SD_STATUS_UPDATED |
3209 (sblk->status & ~SD_STATUS_LINK_CHG);
f47c11ee 3210 spin_lock(&tp->lock);
1da177e4 3211 tg3_setup_phy(tp, 0);
f47c11ee 3212 spin_unlock(&tp->lock);
1da177e4
LT
3213 }
3214 }
3215
3216 /* run TX completion thread */
3217 if (sblk->idx[0].tx_consumer != tp->tx_cons) {
1da177e4 3218 tg3_tx(tp);
1da177e4
LT
3219 }
3220
1da177e4
LT
3221 /* run RX thread, within the bounds set by NAPI.
3222 * All RX "locking" is done by ensuring outside
3223 * code synchronizes with dev->poll()
3224 */
1da177e4
LT
3225 if (sblk->idx[0].rx_producer != tp->rx_rcb_ptr) {
3226 int orig_budget = *budget;
3227 int work_done;
3228
3229 if (orig_budget > netdev->quota)
3230 orig_budget = netdev->quota;
3231
3232 work_done = tg3_rx(tp, orig_budget);
3233
3234 *budget -= work_done;
3235 netdev->quota -= work_done;
1da177e4
LT
3236 }
3237
38f3843e 3238 if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS) {
f7383c22 3239 tp->last_tag = sblk->status_tag;
38f3843e
MC
3240 rmb();
3241 } else
3242 sblk->status &= ~SD_STATUS_UPDATED;
f7383c22 3243
1da177e4 3244 /* if no more work, tell net stack and NIC we're done */
f7383c22 3245 done = !tg3_has_work(tp);
1da177e4 3246 if (done) {
f47c11ee 3247 netif_rx_complete(netdev);
1da177e4 3248 tg3_restart_ints(tp);
1da177e4
LT
3249 }
3250
3251 return (done ? 0 : 1);
3252}
3253
f47c11ee
DM
3254static void tg3_irq_quiesce(struct tg3 *tp)
3255{
3256 BUG_ON(tp->irq_sync);
3257
3258 tp->irq_sync = 1;
3259 smp_mb();
3260
3261 synchronize_irq(tp->pdev->irq);
3262}
3263
3264static inline int tg3_irq_sync(struct tg3 *tp)
3265{
3266 return tp->irq_sync;
3267}
3268
3269/* Fully shutdown all tg3 driver activity elsewhere in the system.
3270 * If irq_sync is non-zero, then the IRQ handler must be synchronized
3271 * with as well. Most of the time, this is not necessary except when
3272 * shutting down the device.
3273 */
3274static inline void tg3_full_lock(struct tg3 *tp, int irq_sync)
3275{
3276 if (irq_sync)
3277 tg3_irq_quiesce(tp);
3278 spin_lock_bh(&tp->lock);
3279 spin_lock(&tp->tx_lock);
3280}
3281
3282static inline void tg3_full_unlock(struct tg3 *tp)
3283{
3284 spin_unlock(&tp->tx_lock);
3285 spin_unlock_bh(&tp->lock);
3286}
3287
88b06bc2
MC
3288/* MSI ISR - No need to check for interrupt sharing and no need to
3289 * flush status block and interrupt mailbox. PCI ordering rules
3290 * guarantee that MSI will arrive after the status block.
3291 */
3292static irqreturn_t tg3_msi(int irq, void *dev_id, struct pt_regs *regs)
3293{
3294 struct net_device *dev = dev_id;
3295 struct tg3 *tp = netdev_priv(dev);
88b06bc2 3296
61487480
MC
3297 prefetch(tp->hw_status);
3298 prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]);
88b06bc2 3299 /*
fac9b83e 3300 * Writing any value to intr-mbox-0 clears PCI INTA# and
88b06bc2 3301 * chip-internal interrupt pending events.
fac9b83e 3302 * Writing non-zero to intr-mbox-0 additional tells the
88b06bc2
MC
3303 * NIC to stop sending us irqs, engaging "in-intr-handler"
3304 * event coalescing.
3305 */
3306 tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0x00000001);
61487480 3307 if (likely(!tg3_irq_sync(tp)))
88b06bc2 3308 netif_rx_schedule(dev); /* schedule NAPI poll */
61487480 3309
88b06bc2
MC
3310 return IRQ_RETVAL(1);
3311}
3312
1da177e4
LT
3313static irqreturn_t tg3_interrupt(int irq, void *dev_id, struct pt_regs *regs)
3314{
3315 struct net_device *dev = dev_id;
3316 struct tg3 *tp = netdev_priv(dev);
3317 struct tg3_hw_status *sblk = tp->hw_status;
1da177e4
LT
3318 unsigned int handled = 1;
3319
1da177e4
LT
3320 /* In INTx mode, it is possible for the interrupt to arrive at
3321 * the CPU before the status block posted prior to the interrupt.
3322 * Reading the PCI State register will confirm whether the
3323 * interrupt is ours and will flush the status block.
3324 */
3325 if ((sblk->status & SD_STATUS_UPDATED) ||
3326 !(tr32(TG3PCI_PCISTATE) & PCISTATE_INT_NOT_ACTIVE)) {
3327 /*
fac9b83e 3328 * Writing any value to intr-mbox-0 clears PCI INTA# and
1da177e4 3329 * chip-internal interrupt pending events.
fac9b83e 3330 * Writing non-zero to intr-mbox-0 additional tells the
1da177e4
LT
3331 * NIC to stop sending us irqs, engaging "in-intr-handler"
3332 * event coalescing.
3333 */
3334 tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW,
3335 0x00000001);
f47c11ee
DM
3336 if (tg3_irq_sync(tp))
3337 goto out;
fac9b83e 3338 sblk->status &= ~SD_STATUS_UPDATED;
61487480
MC
3339 if (likely(tg3_has_work(tp))) {
3340 prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]);
fac9b83e 3341 netif_rx_schedule(dev); /* schedule NAPI poll */
61487480 3342 } else {
fac9b83e
DM
3343 /* No work, shared interrupt perhaps? re-enable
3344 * interrupts, and flush that PCI write
3345 */
09ee929c 3346 tw32_mailbox_f(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW,
fac9b83e 3347 0x00000000);
fac9b83e
DM
3348 }
3349 } else { /* shared interrupt */
3350 handled = 0;
3351 }
f47c11ee 3352out:
fac9b83e
DM
3353 return IRQ_RETVAL(handled);
3354}
3355
3356static irqreturn_t tg3_interrupt_tagged(int irq, void *dev_id, struct pt_regs *regs)
3357{
3358 struct net_device *dev = dev_id;
3359 struct tg3 *tp = netdev_priv(dev);
3360 struct tg3_hw_status *sblk = tp->hw_status;
fac9b83e
DM
3361 unsigned int handled = 1;
3362
fac9b83e
DM
3363 /* In INTx mode, it is possible for the interrupt to arrive at
3364 * the CPU before the status block posted prior to the interrupt.
3365 * Reading the PCI State register will confirm whether the
3366 * interrupt is ours and will flush the status block.
3367 */
38f3843e 3368 if ((sblk->status_tag != tp->last_tag) ||
fac9b83e 3369 !(tr32(TG3PCI_PCISTATE) & PCISTATE_INT_NOT_ACTIVE)) {
1da177e4 3370 /*
fac9b83e
DM
3371 * writing any value to intr-mbox-0 clears PCI INTA# and
3372 * chip-internal interrupt pending events.
3373 * writing non-zero to intr-mbox-0 additional tells the
3374 * NIC to stop sending us irqs, engaging "in-intr-handler"
3375 * event coalescing.
1da177e4 3376 */
fac9b83e
DM
3377 tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW,
3378 0x00000001);
f47c11ee
DM
3379 if (tg3_irq_sync(tp))
3380 goto out;
38f3843e 3381 if (netif_rx_schedule_prep(dev)) {
61487480 3382 prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]);
38f3843e
MC
3383 /* Update last_tag to mark that this status has been
3384 * seen. Because interrupt may be shared, we may be
3385 * racing with tg3_poll(), so only update last_tag
3386 * if tg3_poll() is not scheduled.
1da177e4 3387 */
38f3843e
MC
3388 tp->last_tag = sblk->status_tag;
3389 __netif_rx_schedule(dev);
1da177e4
LT
3390 }
3391 } else { /* shared interrupt */
3392 handled = 0;
3393 }
f47c11ee 3394out:
1da177e4
LT
3395 return IRQ_RETVAL(handled);
3396}
3397
7938109f
MC
3398/* ISR for interrupt test */
3399static irqreturn_t tg3_test_isr(int irq, void *dev_id,
3400 struct pt_regs *regs)
3401{
3402 struct net_device *dev = dev_id;
3403 struct tg3 *tp = netdev_priv(dev);
3404 struct tg3_hw_status *sblk = tp->hw_status;
3405
f9804ddb
MC
3406 if ((sblk->status & SD_STATUS_UPDATED) ||
3407 !(tr32(TG3PCI_PCISTATE) & PCISTATE_INT_NOT_ACTIVE)) {
7938109f
MC
3408 tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW,
3409 0x00000001);
3410 return IRQ_RETVAL(1);
3411 }
3412 return IRQ_RETVAL(0);
3413}
3414
1da177e4 3415static int tg3_init_hw(struct tg3 *);
944d980e 3416static int tg3_halt(struct tg3 *, int, int);
1da177e4
LT
3417
3418#ifdef CONFIG_NET_POLL_CONTROLLER
3419static void tg3_poll_controller(struct net_device *dev)
3420{
88b06bc2
MC
3421 struct tg3 *tp = netdev_priv(dev);
3422
3423 tg3_interrupt(tp->pdev->irq, dev, NULL);
1da177e4
LT
3424}
3425#endif
3426
3427static void tg3_reset_task(void *_data)
3428{
3429 struct tg3 *tp = _data;
3430 unsigned int restart_timer;
3431
3432 tg3_netif_stop(tp);
3433
f47c11ee 3434 tg3_full_lock(tp, 1);
1da177e4
LT
3435
3436 restart_timer = tp->tg3_flags2 & TG3_FLG2_RESTART_TIMER;
3437 tp->tg3_flags2 &= ~TG3_FLG2_RESTART_TIMER;
3438
944d980e 3439 tg3_halt(tp, RESET_KIND_SHUTDOWN, 0);
1da177e4
LT
3440 tg3_init_hw(tp);
3441
3442 tg3_netif_start(tp);
3443
f47c11ee 3444 tg3_full_unlock(tp);
1da177e4
LT
3445
3446 if (restart_timer)
3447 mod_timer(&tp->timer, jiffies + 1);
3448}
3449
3450static void tg3_tx_timeout(struct net_device *dev)
3451{
3452 struct tg3 *tp = netdev_priv(dev);
3453
3454 printk(KERN_ERR PFX "%s: transmit timed out, resetting\n",
3455 dev->name);
3456
3457 schedule_work(&tp->reset_task);
3458}
3459
c58ec932
MC
3460/* Test for DMA buffers crossing any 4GB boundaries: 4G, 8G, etc */
3461static inline int tg3_4g_overflow_test(dma_addr_t mapping, int len)
3462{
3463 u32 base = (u32) mapping & 0xffffffff;
3464
3465 return ((base > 0xffffdcc0) &&
3466 (base + len + 8 < base));
3467}
3468
1da177e4
LT
3469static void tg3_set_txd(struct tg3 *, int, dma_addr_t, int, u32, u32);
3470
3471static int tigon3_4gb_hwbug_workaround(struct tg3 *tp, struct sk_buff *skb,
c58ec932
MC
3472 u32 last_plus_one, u32 *start,
3473 u32 base_flags, u32 mss)
1da177e4
LT
3474{
3475 struct sk_buff *new_skb = skb_copy(skb, GFP_ATOMIC);
c58ec932 3476 dma_addr_t new_addr = 0;
1da177e4 3477 u32 entry = *start;
c58ec932 3478 int i, ret = 0;
1da177e4
LT
3479
3480 if (!new_skb) {
c58ec932
MC
3481 ret = -1;
3482 } else {
3483 /* New SKB is guaranteed to be linear. */
3484 entry = *start;
3485 new_addr = pci_map_single(tp->pdev, new_skb->data, new_skb->len,
3486 PCI_DMA_TODEVICE);
3487 /* Make sure new skb does not cross any 4G boundaries.
3488 * Drop the packet if it does.
3489 */
3490 if (tg3_4g_overflow_test(new_addr, new_skb->len)) {
3491 ret = -1;
3492 dev_kfree_skb(new_skb);
3493 new_skb = NULL;
3494 } else {
3495 tg3_set_txd(tp, entry, new_addr, new_skb->len,
3496 base_flags, 1 | (mss << 1));
3497 *start = NEXT_TX(entry);
3498 }
1da177e4
LT
3499 }
3500
1da177e4
LT
3501 /* Now clean up the sw ring entries. */
3502 i = 0;
3503 while (entry != last_plus_one) {
3504 int len;
3505
3506 if (i == 0)
3507 len = skb_headlen(skb);
3508 else
3509 len = skb_shinfo(skb)->frags[i-1].size;
3510 pci_unmap_single(tp->pdev,
3511 pci_unmap_addr(&tp->tx_buffers[entry], mapping),
3512 len, PCI_DMA_TODEVICE);
3513 if (i == 0) {
3514 tp->tx_buffers[entry].skb = new_skb;
3515 pci_unmap_addr_set(&tp->tx_buffers[entry], mapping, new_addr);
3516 } else {
3517 tp->tx_buffers[entry].skb = NULL;
3518 }
3519 entry = NEXT_TX(entry);
3520 i++;
3521 }
3522
3523 dev_kfree_skb(skb);
3524
c58ec932 3525 return ret;
1da177e4
LT
3526}
3527
3528static void tg3_set_txd(struct tg3 *tp, int entry,
3529 dma_addr_t mapping, int len, u32 flags,
3530 u32 mss_and_is_end)
3531{
3532 struct tg3_tx_buffer_desc *txd = &tp->tx_ring[entry];
3533 int is_end = (mss_and_is_end & 0x1);
3534 u32 mss = (mss_and_is_end >> 1);
3535 u32 vlan_tag = 0;
3536
3537 if (is_end)
3538 flags |= TXD_FLAG_END;
3539 if (flags & TXD_FLAG_VLAN) {
3540 vlan_tag = flags >> 16;
3541 flags &= 0xffff;
3542 }
3543 vlan_tag |= (mss << TXD_MSS_SHIFT);
3544
3545 txd->addr_hi = ((u64) mapping >> 32);
3546 txd->addr_lo = ((u64) mapping & 0xffffffff);
3547 txd->len_flags = (len << TXD_LEN_SHIFT) | flags;
3548 txd->vlan_tag = vlan_tag << TXD_VLAN_TAG_SHIFT;
3549}
3550
1da177e4
LT
3551static int tg3_start_xmit(struct sk_buff *skb, struct net_device *dev)
3552{
3553 struct tg3 *tp = netdev_priv(dev);
3554 dma_addr_t mapping;
1da177e4
LT
3555 u32 len, entry, base_flags, mss;
3556 int would_hit_hwbug;
1da177e4
LT
3557
3558 len = skb_headlen(skb);
3559
3560 /* No BH disabling for tx_lock here. We are running in BH disabled
3561 * context and TX reclaim runs via tp->poll inside of a software
f47c11ee
DM
3562 * interrupt. Furthermore, IRQ processing runs lockless so we have
3563 * no IRQ context deadlocks to worry about either. Rejoice!
1da177e4 3564 */
f47c11ee 3565 if (!spin_trylock(&tp->tx_lock))
1da177e4 3566 return NETDEV_TX_LOCKED;
1da177e4
LT
3567
3568 /* This is a hard error, log it. */
3569 if (unlikely(TX_BUFFS_AVAIL(tp) <= (skb_shinfo(skb)->nr_frags + 1))) {
3570 netif_stop_queue(dev);
f47c11ee 3571 spin_unlock(&tp->tx_lock);
1da177e4
LT
3572 printk(KERN_ERR PFX "%s: BUG! Tx Ring full when queue awake!\n",
3573 dev->name);
3574 return NETDEV_TX_BUSY;
3575 }
3576
3577 entry = tp->tx_prod;
3578 base_flags = 0;
3579 if (skb->ip_summed == CHECKSUM_HW)
3580 base_flags |= TXD_FLAG_TCPUDP_CSUM;
3581#if TG3_TSO_SUPPORT != 0
3582 mss = 0;
3583 if (skb->len > (tp->dev->mtu + ETH_HLEN) &&
3584 (mss = skb_shinfo(skb)->tso_size) != 0) {
3585 int tcp_opt_len, ip_tcp_len;
3586
3587 if (skb_header_cloned(skb) &&
3588 pskb_expand_head(skb, 0, 0, GFP_ATOMIC)) {
3589 dev_kfree_skb(skb);
3590 goto out_unlock;
3591 }
3592
3593 tcp_opt_len = ((skb->h.th->doff - 5) * 4);
3594 ip_tcp_len = (skb->nh.iph->ihl * 4) + sizeof(struct tcphdr);
3595
3596 base_flags |= (TXD_FLAG_CPU_PRE_DMA |
3597 TXD_FLAG_CPU_POST_DMA);
3598
3599 skb->nh.iph->check = 0;
3600 skb->nh.iph->tot_len = ntohs(mss + ip_tcp_len + tcp_opt_len);
3601 if (tp->tg3_flags2 & TG3_FLG2_HW_TSO) {
3602 skb->h.th->check = 0;
3603 base_flags &= ~TXD_FLAG_TCPUDP_CSUM;
3604 }
3605 else {
3606 skb->h.th->check =
3607 ~csum_tcpudp_magic(skb->nh.iph->saddr,
3608 skb->nh.iph->daddr,
3609 0, IPPROTO_TCP, 0);
3610 }
3611
3612 if ((tp->tg3_flags2 & TG3_FLG2_HW_TSO) ||
3613 (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705)) {
3614 if (tcp_opt_len || skb->nh.iph->ihl > 5) {
3615 int tsflags;
3616
3617 tsflags = ((skb->nh.iph->ihl - 5) +
3618 (tcp_opt_len >> 2));
3619 mss |= (tsflags << 11);
3620 }
3621 } else {
3622 if (tcp_opt_len || skb->nh.iph->ihl > 5) {
3623 int tsflags;
3624
3625 tsflags = ((skb->nh.iph->ihl - 5) +
3626 (tcp_opt_len >> 2));
3627 base_flags |= tsflags << 12;
3628 }
3629 }
3630 }
3631#else
3632 mss = 0;
3633#endif
3634#if TG3_VLAN_TAG_USED
3635 if (tp->vlgrp != NULL && vlan_tx_tag_present(skb))
3636 base_flags |= (TXD_FLAG_VLAN |
3637 (vlan_tx_tag_get(skb) << 16));
3638#endif
3639
3640 /* Queue skb data, a.k.a. the main skb fragment. */
3641 mapping = pci_map_single(tp->pdev, skb->data, len, PCI_DMA_TODEVICE);
3642
3643 tp->tx_buffers[entry].skb = skb;
3644 pci_unmap_addr_set(&tp->tx_buffers[entry], mapping, mapping);
3645
3646 would_hit_hwbug = 0;
3647
3648 if (tg3_4g_overflow_test(mapping, len))
c58ec932 3649 would_hit_hwbug = 1;
1da177e4
LT
3650
3651 tg3_set_txd(tp, entry, mapping, len, base_flags,
3652 (skb_shinfo(skb)->nr_frags == 0) | (mss << 1));
3653
3654 entry = NEXT_TX(entry);
3655
3656 /* Now loop through additional data fragments, and queue them. */
3657 if (skb_shinfo(skb)->nr_frags > 0) {
3658 unsigned int i, last;
3659
3660 last = skb_shinfo(skb)->nr_frags - 1;
3661 for (i = 0; i <= last; i++) {
3662 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
3663
3664 len = frag->size;
3665 mapping = pci_map_page(tp->pdev,
3666 frag->page,
3667 frag->page_offset,
3668 len, PCI_DMA_TODEVICE);
3669
3670 tp->tx_buffers[entry].skb = NULL;
3671 pci_unmap_addr_set(&tp->tx_buffers[entry], mapping, mapping);
3672
c58ec932
MC
3673 if (tg3_4g_overflow_test(mapping, len))
3674 would_hit_hwbug = 1;
1da177e4
LT
3675
3676 if (tp->tg3_flags2 & TG3_FLG2_HW_TSO)
3677 tg3_set_txd(tp, entry, mapping, len,
3678 base_flags, (i == last)|(mss << 1));
3679 else
3680 tg3_set_txd(tp, entry, mapping, len,
3681 base_flags, (i == last));
3682
3683 entry = NEXT_TX(entry);
3684 }
3685 }
3686
3687 if (would_hit_hwbug) {
3688 u32 last_plus_one = entry;
3689 u32 start;
1da177e4 3690
c58ec932
MC
3691 start = entry - 1 - skb_shinfo(skb)->nr_frags;
3692 start &= (TG3_TX_RING_SIZE - 1);
1da177e4
LT
3693
3694 /* If the workaround fails due to memory/mapping
3695 * failure, silently drop this packet.
3696 */
c58ec932
MC
3697 if (tigon3_4gb_hwbug_workaround(tp, skb, last_plus_one,
3698 &start, base_flags, mss))
1da177e4
LT
3699 goto out_unlock;
3700
3701 entry = start;
3702 }
3703
3704 /* Packets are ready, update Tx producer idx local and on card. */
3705 tw32_tx_mbox((MAILBOX_SNDHOST_PROD_IDX_0 + TG3_64BIT_REG_LOW), entry);
3706
3707 tp->tx_prod = entry;
51b91468 3708 if (TX_BUFFS_AVAIL(tp) <= (MAX_SKB_FRAGS + 1)) {
1da177e4 3709 netif_stop_queue(dev);
51b91468
MC
3710 if (TX_BUFFS_AVAIL(tp) > TG3_TX_WAKEUP_THRESH)
3711 netif_wake_queue(tp->dev);
3712 }
1da177e4
LT
3713
3714out_unlock:
3715 mmiowb();
f47c11ee 3716 spin_unlock(&tp->tx_lock);
1da177e4
LT
3717
3718 dev->trans_start = jiffies;
3719
3720 return NETDEV_TX_OK;
3721}
3722
3723static inline void tg3_set_mtu(struct net_device *dev, struct tg3 *tp,
3724 int new_mtu)
3725{
3726 dev->mtu = new_mtu;
3727
ef7f5ec0 3728 if (new_mtu > ETH_DATA_LEN) {
a4e2b347 3729 if (tp->tg3_flags2 & TG3_FLG2_5780_CLASS) {
ef7f5ec0
MC
3730 tp->tg3_flags2 &= ~TG3_FLG2_TSO_CAPABLE;
3731 ethtool_op_set_tso(dev, 0);
3732 }
3733 else
3734 tp->tg3_flags |= TG3_FLAG_JUMBO_RING_ENABLE;
3735 } else {
a4e2b347 3736 if (tp->tg3_flags2 & TG3_FLG2_5780_CLASS)
ef7f5ec0 3737 tp->tg3_flags2 |= TG3_FLG2_TSO_CAPABLE;
0f893dc6 3738 tp->tg3_flags &= ~TG3_FLAG_JUMBO_RING_ENABLE;
ef7f5ec0 3739 }
1da177e4
LT
3740}
3741
3742static int tg3_change_mtu(struct net_device *dev, int new_mtu)
3743{
3744 struct tg3 *tp = netdev_priv(dev);
3745
3746 if (new_mtu < TG3_MIN_MTU || new_mtu > TG3_MAX_MTU(tp))
3747 return -EINVAL;
3748
3749 if (!netif_running(dev)) {
3750 /* We'll just catch it later when the
3751 * device is up'd.
3752 */
3753 tg3_set_mtu(dev, tp, new_mtu);
3754 return 0;
3755 }
3756
3757 tg3_netif_stop(tp);
f47c11ee
DM
3758
3759 tg3_full_lock(tp, 1);
1da177e4 3760
944d980e 3761 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
1da177e4
LT
3762
3763 tg3_set_mtu(dev, tp, new_mtu);
3764
3765 tg3_init_hw(tp);
3766
3767 tg3_netif_start(tp);
3768
f47c11ee 3769 tg3_full_unlock(tp);
1da177e4
LT
3770
3771 return 0;
3772}
3773
3774/* Free up pending packets in all rx/tx rings.
3775 *
3776 * The chip has been shut down and the driver detached from
3777 * the networking, so no interrupts or new tx packets will
3778 * end up in the driver. tp->{tx,}lock is not held and we are not
3779 * in an interrupt context and thus may sleep.
3780 */
3781static void tg3_free_rings(struct tg3 *tp)
3782{
3783 struct ring_info *rxp;
3784 int i;
3785
3786 for (i = 0; i < TG3_RX_RING_SIZE; i++) {
3787 rxp = &tp->rx_std_buffers[i];
3788
3789 if (rxp->skb == NULL)
3790 continue;
3791 pci_unmap_single(tp->pdev,
3792 pci_unmap_addr(rxp, mapping),
7e72aad4 3793 tp->rx_pkt_buf_sz - tp->rx_offset,
1da177e4
LT
3794 PCI_DMA_FROMDEVICE);
3795 dev_kfree_skb_any(rxp->skb);
3796 rxp->skb = NULL;
3797 }
3798
3799 for (i = 0; i < TG3_RX_JUMBO_RING_SIZE; i++) {
3800 rxp = &tp->rx_jumbo_buffers[i];
3801
3802 if (rxp->skb == NULL)
3803 continue;
3804 pci_unmap_single(tp->pdev,
3805 pci_unmap_addr(rxp, mapping),
3806 RX_JUMBO_PKT_BUF_SZ - tp->rx_offset,
3807 PCI_DMA_FROMDEVICE);
3808 dev_kfree_skb_any(rxp->skb);
3809 rxp->skb = NULL;
3810 }
3811
3812 for (i = 0; i < TG3_TX_RING_SIZE; ) {
3813 struct tx_ring_info *txp;
3814 struct sk_buff *skb;
3815 int j;
3816
3817 txp = &tp->tx_buffers[i];
3818 skb = txp->skb;
3819
3820 if (skb == NULL) {
3821 i++;
3822 continue;
3823 }
3824
3825 pci_unmap_single(tp->pdev,
3826 pci_unmap_addr(txp, mapping),
3827 skb_headlen(skb),
3828 PCI_DMA_TODEVICE);
3829 txp->skb = NULL;
3830
3831 i++;
3832
3833 for (j = 0; j < skb_shinfo(skb)->nr_frags; j++) {
3834 txp = &tp->tx_buffers[i & (TG3_TX_RING_SIZE - 1)];
3835 pci_unmap_page(tp->pdev,
3836 pci_unmap_addr(txp, mapping),
3837 skb_shinfo(skb)->frags[j].size,
3838 PCI_DMA_TODEVICE);
3839 i++;
3840 }
3841
3842 dev_kfree_skb_any(skb);
3843 }
3844}
3845
3846/* Initialize tx/rx rings for packet processing.
3847 *
3848 * The chip has been shut down and the driver detached from
3849 * the networking, so no interrupts or new tx packets will
3850 * end up in the driver. tp->{tx,}lock are held and thus
3851 * we may not sleep.
3852 */
3853static void tg3_init_rings(struct tg3 *tp)
3854{
3855 u32 i;
3856
3857 /* Free up all the SKBs. */
3858 tg3_free_rings(tp);
3859
3860 /* Zero out all descriptors. */
3861 memset(tp->rx_std, 0, TG3_RX_RING_BYTES);
3862 memset(tp->rx_jumbo, 0, TG3_RX_JUMBO_RING_BYTES);
3863 memset(tp->rx_rcb, 0, TG3_RX_RCB_RING_BYTES(tp));
3864 memset(tp->tx_ring, 0, TG3_TX_RING_BYTES);
3865
7e72aad4 3866 tp->rx_pkt_buf_sz = RX_PKT_BUF_SZ;
a4e2b347 3867 if ((tp->tg3_flags2 & TG3_FLG2_5780_CLASS) &&
7e72aad4
MC
3868 (tp->dev->mtu > ETH_DATA_LEN))
3869 tp->rx_pkt_buf_sz = RX_JUMBO_PKT_BUF_SZ;
3870
1da177e4
LT
3871 /* Initialize invariants of the rings, we only set this
3872 * stuff once. This works because the card does not
3873 * write into the rx buffer posting rings.
3874 */
3875 for (i = 0; i < TG3_RX_RING_SIZE; i++) {
3876 struct tg3_rx_buffer_desc *rxd;
3877
3878 rxd = &tp->rx_std[i];
7e72aad4 3879 rxd->idx_len = (tp->rx_pkt_buf_sz - tp->rx_offset - 64)
1da177e4
LT
3880 << RXD_LEN_SHIFT;
3881 rxd->type_flags = (RXD_FLAG_END << RXD_FLAGS_SHIFT);
3882 rxd->opaque = (RXD_OPAQUE_RING_STD |
3883 (i << RXD_OPAQUE_INDEX_SHIFT));
3884 }
3885
0f893dc6 3886 if (tp->tg3_flags & TG3_FLAG_JUMBO_RING_ENABLE) {
1da177e4
LT
3887 for (i = 0; i < TG3_RX_JUMBO_RING_SIZE; i++) {
3888 struct tg3_rx_buffer_desc *rxd;
3889
3890 rxd = &tp->rx_jumbo[i];
3891 rxd->idx_len = (RX_JUMBO_PKT_BUF_SZ - tp->rx_offset - 64)
3892 << RXD_LEN_SHIFT;
3893 rxd->type_flags = (RXD_FLAG_END << RXD_FLAGS_SHIFT) |
3894 RXD_FLAG_JUMBO;
3895 rxd->opaque = (RXD_OPAQUE_RING_JUMBO |
3896 (i << RXD_OPAQUE_INDEX_SHIFT));
3897 }
3898 }
3899
3900 /* Now allocate fresh SKBs for each rx ring. */
3901 for (i = 0; i < tp->rx_pending; i++) {
3902 if (tg3_alloc_rx_skb(tp, RXD_OPAQUE_RING_STD,
3903 -1, i) < 0)
3904 break;
3905 }
3906
0f893dc6 3907 if (tp->tg3_flags & TG3_FLAG_JUMBO_RING_ENABLE) {
1da177e4
LT
3908 for (i = 0; i < tp->rx_jumbo_pending; i++) {
3909 if (tg3_alloc_rx_skb(tp, RXD_OPAQUE_RING_JUMBO,
3910 -1, i) < 0)
3911 break;
3912 }
3913 }
3914}
3915
3916/*
3917 * Must not be invoked with interrupt sources disabled and
3918 * the hardware shutdown down.
3919 */
3920static void tg3_free_consistent(struct tg3 *tp)
3921{
b4558ea9
JJ
3922 kfree(tp->rx_std_buffers);
3923 tp->rx_std_buffers = NULL;
1da177e4
LT
3924 if (tp->rx_std) {
3925 pci_free_consistent(tp->pdev, TG3_RX_RING_BYTES,
3926 tp->rx_std, tp->rx_std_mapping);
3927 tp->rx_std = NULL;
3928 }
3929 if (tp->rx_jumbo) {
3930 pci_free_consistent(tp->pdev, TG3_RX_JUMBO_RING_BYTES,
3931 tp->rx_jumbo, tp->rx_jumbo_mapping);
3932 tp->rx_jumbo = NULL;
3933 }
3934 if (tp->rx_rcb) {
3935 pci_free_consistent(tp->pdev, TG3_RX_RCB_RING_BYTES(tp),
3936 tp->rx_rcb, tp->rx_rcb_mapping);
3937 tp->rx_rcb = NULL;
3938 }
3939 if (tp->tx_ring) {
3940 pci_free_consistent(tp->pdev, TG3_TX_RING_BYTES,
3941 tp->tx_ring, tp->tx_desc_mapping);
3942 tp->tx_ring = NULL;
3943 }
3944 if (tp->hw_status) {
3945 pci_free_consistent(tp->pdev, TG3_HW_STATUS_SIZE,
3946 tp->hw_status, tp->status_mapping);
3947 tp->hw_status = NULL;
3948 }
3949 if (tp->hw_stats) {
3950 pci_free_consistent(tp->pdev, sizeof(struct tg3_hw_stats),
3951 tp->hw_stats, tp->stats_mapping);
3952 tp->hw_stats = NULL;
3953 }
3954}
3955
3956/*
3957 * Must not be invoked with interrupt sources disabled and
3958 * the hardware shutdown down. Can sleep.
3959 */
3960static int tg3_alloc_consistent(struct tg3 *tp)
3961{
3962 tp->rx_std_buffers = kmalloc((sizeof(struct ring_info) *
3963 (TG3_RX_RING_SIZE +
3964 TG3_RX_JUMBO_RING_SIZE)) +
3965 (sizeof(struct tx_ring_info) *
3966 TG3_TX_RING_SIZE),
3967 GFP_KERNEL);
3968 if (!tp->rx_std_buffers)
3969 return -ENOMEM;
3970
3971 memset(tp->rx_std_buffers, 0,
3972 (sizeof(struct ring_info) *
3973 (TG3_RX_RING_SIZE +
3974 TG3_RX_JUMBO_RING_SIZE)) +
3975 (sizeof(struct tx_ring_info) *
3976 TG3_TX_RING_SIZE));
3977
3978 tp->rx_jumbo_buffers = &tp->rx_std_buffers[TG3_RX_RING_SIZE];
3979 tp->tx_buffers = (struct tx_ring_info *)
3980 &tp->rx_jumbo_buffers[TG3_RX_JUMBO_RING_SIZE];
3981
3982 tp->rx_std = pci_alloc_consistent(tp->pdev, TG3_RX_RING_BYTES,
3983 &tp->rx_std_mapping);
3984 if (!tp->rx_std)
3985 goto err_out;
3986
3987 tp->rx_jumbo = pci_alloc_consistent(tp->pdev, TG3_RX_JUMBO_RING_BYTES,
3988 &tp->rx_jumbo_mapping);
3989
3990 if (!tp->rx_jumbo)
3991 goto err_out;
3992
3993 tp->rx_rcb = pci_alloc_consistent(tp->pdev, TG3_RX_RCB_RING_BYTES(tp),
3994 &tp->rx_rcb_mapping);
3995 if (!tp->rx_rcb)
3996 goto err_out;
3997
3998 tp->tx_ring = pci_alloc_consistent(tp->pdev, TG3_TX_RING_BYTES,
3999 &tp->tx_desc_mapping);
4000 if (!tp->tx_ring)
4001 goto err_out;
4002
4003 tp->hw_status = pci_alloc_consistent(tp->pdev,
4004 TG3_HW_STATUS_SIZE,
4005 &tp->status_mapping);
4006 if (!tp->hw_status)
4007 goto err_out;
4008
4009 tp->hw_stats = pci_alloc_consistent(tp->pdev,
4010 sizeof(struct tg3_hw_stats),
4011 &tp->stats_mapping);
4012 if (!tp->hw_stats)
4013 goto err_out;
4014
4015 memset(tp->hw_status, 0, TG3_HW_STATUS_SIZE);
4016 memset(tp->hw_stats, 0, sizeof(struct tg3_hw_stats));
4017
4018 return 0;
4019
4020err_out:
4021 tg3_free_consistent(tp);
4022 return -ENOMEM;
4023}
4024
4025#define MAX_WAIT_CNT 1000
4026
4027/* To stop a block, clear the enable bit and poll till it
4028 * clears. tp->lock is held.
4029 */
b3b7d6be 4030static int tg3_stop_block(struct tg3 *tp, unsigned long ofs, u32 enable_bit, int silent)
1da177e4
LT
4031{
4032 unsigned int i;
4033 u32 val;
4034
4035 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS) {
4036 switch (ofs) {
4037 case RCVLSC_MODE:
4038 case DMAC_MODE:
4039 case MBFREE_MODE:
4040 case BUFMGR_MODE:
4041 case MEMARB_MODE:
4042 /* We can't enable/disable these bits of the
4043 * 5705/5750, just say success.
4044 */
4045 return 0;
4046
4047 default:
4048 break;
4049 };
4050 }
4051
4052 val = tr32(ofs);
4053 val &= ~enable_bit;
4054 tw32_f(ofs, val);
4055
4056 for (i = 0; i < MAX_WAIT_CNT; i++) {
4057 udelay(100);
4058 val = tr32(ofs);
4059 if ((val & enable_bit) == 0)
4060 break;
4061 }
4062
b3b7d6be 4063 if (i == MAX_WAIT_CNT && !silent) {
1da177e4
LT
4064 printk(KERN_ERR PFX "tg3_stop_block timed out, "
4065 "ofs=%lx enable_bit=%x\n",
4066 ofs, enable_bit);
4067 return -ENODEV;
4068 }
4069
4070 return 0;
4071}
4072
4073/* tp->lock is held. */
b3b7d6be 4074static int tg3_abort_hw(struct tg3 *tp, int silent)
1da177e4
LT
4075{
4076 int i, err;
4077
4078 tg3_disable_ints(tp);
4079
4080 tp->rx_mode &= ~RX_MODE_ENABLE;
4081 tw32_f(MAC_RX_MODE, tp->rx_mode);
4082 udelay(10);
4083
b3b7d6be
DM
4084 err = tg3_stop_block(tp, RCVBDI_MODE, RCVBDI_MODE_ENABLE, silent);
4085 err |= tg3_stop_block(tp, RCVLPC_MODE, RCVLPC_MODE_ENABLE, silent);
4086 err |= tg3_stop_block(tp, RCVLSC_MODE, RCVLSC_MODE_ENABLE, silent);
4087 err |= tg3_stop_block(tp, RCVDBDI_MODE, RCVDBDI_MODE_ENABLE, silent);
4088 err |= tg3_stop_block(tp, RCVDCC_MODE, RCVDCC_MODE_ENABLE, silent);
4089 err |= tg3_stop_block(tp, RCVCC_MODE, RCVCC_MODE_ENABLE, silent);
4090
4091 err |= tg3_stop_block(tp, SNDBDS_MODE, SNDBDS_MODE_ENABLE, silent);
4092 err |= tg3_stop_block(tp, SNDBDI_MODE, SNDBDI_MODE_ENABLE, silent);
4093 err |= tg3_stop_block(tp, SNDDATAI_MODE, SNDDATAI_MODE_ENABLE, silent);
4094 err |= tg3_stop_block(tp, RDMAC_MODE, RDMAC_MODE_ENABLE, silent);
4095 err |= tg3_stop_block(tp, SNDDATAC_MODE, SNDDATAC_MODE_ENABLE, silent);
4096 err |= tg3_stop_block(tp, DMAC_MODE, DMAC_MODE_ENABLE, silent);
4097 err |= tg3_stop_block(tp, SNDBDC_MODE, SNDBDC_MODE_ENABLE, silent);
1da177e4
LT
4098
4099 tp->mac_mode &= ~MAC_MODE_TDE_ENABLE;
4100 tw32_f(MAC_MODE, tp->mac_mode);
4101 udelay(40);
4102
4103 tp->tx_mode &= ~TX_MODE_ENABLE;
4104 tw32_f(MAC_TX_MODE, tp->tx_mode);
4105
4106 for (i = 0; i < MAX_WAIT_CNT; i++) {
4107 udelay(100);
4108 if (!(tr32(MAC_TX_MODE) & TX_MODE_ENABLE))
4109 break;
4110 }
4111 if (i >= MAX_WAIT_CNT) {
4112 printk(KERN_ERR PFX "tg3_abort_hw timed out for %s, "
4113 "TX_MODE_ENABLE will not clear MAC_TX_MODE=%08x\n",
4114 tp->dev->name, tr32(MAC_TX_MODE));
e6de8ad1 4115 err |= -ENODEV;
1da177e4
LT
4116 }
4117
e6de8ad1 4118 err |= tg3_stop_block(tp, HOSTCC_MODE, HOSTCC_MODE_ENABLE, silent);
b3b7d6be
DM
4119 err |= tg3_stop_block(tp, WDMAC_MODE, WDMAC_MODE_ENABLE, silent);
4120 err |= tg3_stop_block(tp, MBFREE_MODE, MBFREE_MODE_ENABLE, silent);
1da177e4
LT
4121
4122 tw32(FTQ_RESET, 0xffffffff);
4123 tw32(FTQ_RESET, 0x00000000);
4124
b3b7d6be
DM
4125 err |= tg3_stop_block(tp, BUFMGR_MODE, BUFMGR_MODE_ENABLE, silent);
4126 err |= tg3_stop_block(tp, MEMARB_MODE, MEMARB_MODE_ENABLE, silent);
1da177e4
LT
4127
4128 if (tp->hw_status)
4129 memset(tp->hw_status, 0, TG3_HW_STATUS_SIZE);
4130 if (tp->hw_stats)
4131 memset(tp->hw_stats, 0, sizeof(struct tg3_hw_stats));
4132
1da177e4
LT
4133 return err;
4134}
4135
4136/* tp->lock is held. */
4137static int tg3_nvram_lock(struct tg3 *tp)
4138{
4139 if (tp->tg3_flags & TG3_FLAG_NVRAM) {
4140 int i;
4141
4142 tw32(NVRAM_SWARB, SWARB_REQ_SET1);
4143 for (i = 0; i < 8000; i++) {
4144 if (tr32(NVRAM_SWARB) & SWARB_GNT1)
4145 break;
4146 udelay(20);
4147 }
4148 if (i == 8000)
4149 return -ENODEV;
4150 }
4151 return 0;
4152}
4153
4154/* tp->lock is held. */
4155static void tg3_nvram_unlock(struct tg3 *tp)
4156{
4157 if (tp->tg3_flags & TG3_FLAG_NVRAM)
4158 tw32_f(NVRAM_SWARB, SWARB_REQ_CLR1);
4159}
4160
e6af301b
MC
4161/* tp->lock is held. */
4162static void tg3_enable_nvram_access(struct tg3 *tp)
4163{
4164 if ((tp->tg3_flags2 & TG3_FLG2_5750_PLUS) &&
4165 !(tp->tg3_flags2 & TG3_FLG2_PROTECTED_NVRAM)) {
4166 u32 nvaccess = tr32(NVRAM_ACCESS);
4167
4168 tw32(NVRAM_ACCESS, nvaccess | ACCESS_ENABLE);
4169 }
4170}
4171
4172/* tp->lock is held. */
4173static void tg3_disable_nvram_access(struct tg3 *tp)
4174{
4175 if ((tp->tg3_flags2 & TG3_FLG2_5750_PLUS) &&
4176 !(tp->tg3_flags2 & TG3_FLG2_PROTECTED_NVRAM)) {
4177 u32 nvaccess = tr32(NVRAM_ACCESS);
4178
4179 tw32(NVRAM_ACCESS, nvaccess & ~ACCESS_ENABLE);
4180 }
4181}
4182
1da177e4
LT
4183/* tp->lock is held. */
4184static void tg3_write_sig_pre_reset(struct tg3 *tp, int kind)
4185{
4186 if (!(tp->tg3_flags2 & TG3_FLG2_SUN_570X))
4187 tg3_write_mem(tp, NIC_SRAM_FIRMWARE_MBOX,
4188 NIC_SRAM_FIRMWARE_MBOX_MAGIC1);
4189
4190 if (tp->tg3_flags2 & TG3_FLG2_ASF_NEW_HANDSHAKE) {
4191 switch (kind) {
4192 case RESET_KIND_INIT:
4193 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4194 DRV_STATE_START);
4195 break;
4196
4197 case RESET_KIND_SHUTDOWN:
4198 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4199 DRV_STATE_UNLOAD);
4200 break;
4201
4202 case RESET_KIND_SUSPEND:
4203 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4204 DRV_STATE_SUSPEND);
4205 break;
4206
4207 default:
4208 break;
4209 };
4210 }
4211}
4212
4213/* tp->lock is held. */
4214static void tg3_write_sig_post_reset(struct tg3 *tp, int kind)
4215{
4216 if (tp->tg3_flags2 & TG3_FLG2_ASF_NEW_HANDSHAKE) {
4217 switch (kind) {
4218 case RESET_KIND_INIT:
4219 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4220 DRV_STATE_START_DONE);
4221 break;
4222
4223 case RESET_KIND_SHUTDOWN:
4224 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4225 DRV_STATE_UNLOAD_DONE);
4226 break;
4227
4228 default:
4229 break;
4230 };
4231 }
4232}
4233
4234/* tp->lock is held. */
4235static void tg3_write_sig_legacy(struct tg3 *tp, int kind)
4236{
4237 if (tp->tg3_flags & TG3_FLAG_ENABLE_ASF) {
4238 switch (kind) {
4239 case RESET_KIND_INIT:
4240 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4241 DRV_STATE_START);
4242 break;
4243
4244 case RESET_KIND_SHUTDOWN:
4245 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4246 DRV_STATE_UNLOAD);
4247 break;
4248
4249 case RESET_KIND_SUSPEND:
4250 tg3_write_mem(tp, NIC_SRAM_FW_DRV_STATE_MBOX,
4251 DRV_STATE_SUSPEND);
4252 break;
4253
4254 default:
4255 break;
4256 };
4257 }
4258}
4259
4260static void tg3_stop_fw(struct tg3 *);
4261
4262/* tp->lock is held. */
4263static int tg3_chip_reset(struct tg3 *tp)
4264{
4265 u32 val;
1ee582d8 4266 void (*write_op)(struct tg3 *, u32, u32);
1da177e4
LT
4267 int i;
4268
4269 if (!(tp->tg3_flags2 & TG3_FLG2_SUN_570X))
4270 tg3_nvram_lock(tp);
4271
4272 /*
4273 * We must avoid the readl() that normally takes place.
4274 * It locks machines, causes machine checks, and other
4275 * fun things. So, temporarily disable the 5701
4276 * hardware workaround, while we do the reset.
4277 */
1ee582d8
MC
4278 write_op = tp->write32;
4279 if (write_op == tg3_write_flush_reg32)
4280 tp->write32 = tg3_write32;
1da177e4
LT
4281
4282 /* do the reset */
4283 val = GRC_MISC_CFG_CORECLK_RESET;
4284
4285 if (tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) {
4286 if (tr32(0x7e2c) == 0x60) {
4287 tw32(0x7e2c, 0x20);
4288 }
4289 if (tp->pci_chip_rev_id != CHIPREV_ID_5750_A0) {
4290 tw32(GRC_MISC_CFG, (1 << 29));
4291 val |= (1 << 29);
4292 }
4293 }
4294
4295 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)
4296 val |= GRC_MISC_CFG_KEEP_GPHY_POWER;
4297 tw32(GRC_MISC_CFG, val);
4298
1ee582d8
MC
4299 /* restore 5701 hardware bug workaround write method */
4300 tp->write32 = write_op;
1da177e4
LT
4301
4302 /* Unfortunately, we have to delay before the PCI read back.
4303 * Some 575X chips even will not respond to a PCI cfg access
4304 * when the reset command is given to the chip.
4305 *
4306 * How do these hardware designers expect things to work
4307 * properly if the PCI write is posted for a long period
4308 * of time? It is always necessary to have some method by
4309 * which a register read back can occur to push the write
4310 * out which does the reset.
4311 *
4312 * For most tg3 variants the trick below was working.
4313 * Ho hum...
4314 */
4315 udelay(120);
4316
4317 /* Flush PCI posted writes. The normal MMIO registers
4318 * are inaccessible at this time so this is the only
4319 * way to make this reliably (actually, this is no longer
4320 * the case, see above). I tried to use indirect
4321 * register read/write but this upset some 5701 variants.
4322 */
4323 pci_read_config_dword(tp->pdev, PCI_COMMAND, &val);
4324
4325 udelay(120);
4326
4327 if (tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) {
4328 if (tp->pci_chip_rev_id == CHIPREV_ID_5750_A0) {
4329 int i;
4330 u32 cfg_val;
4331
4332 /* Wait for link training to complete. */
4333 for (i = 0; i < 5000; i++)
4334 udelay(100);
4335
4336 pci_read_config_dword(tp->pdev, 0xc4, &cfg_val);
4337 pci_write_config_dword(tp->pdev, 0xc4,
4338 cfg_val | (1 << 15));
4339 }
4340 /* Set PCIE max payload size and clear error status. */
4341 pci_write_config_dword(tp->pdev, 0xd8, 0xf5000);
4342 }
4343
4344 /* Re-enable indirect register accesses. */
4345 pci_write_config_dword(tp->pdev, TG3PCI_MISC_HOST_CTRL,
4346 tp->misc_host_ctrl);
4347
4348 /* Set MAX PCI retry to zero. */
4349 val = (PCISTATE_ROM_ENABLE | PCISTATE_ROM_RETRY_ENABLE);
4350 if (tp->pci_chip_rev_id == CHIPREV_ID_5704_A0 &&
4351 (tp->tg3_flags & TG3_FLAG_PCIX_MODE))
4352 val |= PCISTATE_RETRY_SAME_DMA;
4353 pci_write_config_dword(tp->pdev, TG3PCI_PCISTATE, val);
4354
4355 pci_restore_state(tp->pdev);
4356
4357 /* Make sure PCI-X relaxed ordering bit is clear. */
4358 pci_read_config_dword(tp->pdev, TG3PCI_X_CAPS, &val);
4359 val &= ~PCIX_CAPS_RELAXED_ORDERING;
4360 pci_write_config_dword(tp->pdev, TG3PCI_X_CAPS, val);
4361
a4e2b347 4362 if (tp->tg3_flags2 & TG3_FLG2_5780_CLASS) {
4cf78e4f
MC
4363 u32 val;
4364
4365 /* Chip reset on 5780 will reset MSI enable bit,
4366 * so need to restore it.
4367 */
4368 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) {
4369 u16 ctrl;
4370
4371 pci_read_config_word(tp->pdev,
4372 tp->msi_cap + PCI_MSI_FLAGS,
4373 &ctrl);
4374 pci_write_config_word(tp->pdev,
4375 tp->msi_cap + PCI_MSI_FLAGS,
4376 ctrl | PCI_MSI_FLAGS_ENABLE);
4377 val = tr32(MSGINT_MODE);
4378 tw32(MSGINT_MODE, val | MSGINT_MODE_ENABLE);
4379 }
4380
4381 val = tr32(MEMARB_MODE);
4382 tw32(MEMARB_MODE, val | MEMARB_MODE_ENABLE);
4383
4384 } else
4385 tw32(MEMARB_MODE, MEMARB_MODE_ENABLE);
1da177e4
LT
4386
4387 if (tp->pci_chip_rev_id == CHIPREV_ID_5750_A3) {
4388 tg3_stop_fw(tp);
4389 tw32(0x5000, 0x400);
4390 }
4391
4392 tw32(GRC_MODE, tp->grc_mode);
4393
4394 if (tp->pci_chip_rev_id == CHIPREV_ID_5705_A0) {
4395 u32 val = tr32(0xc4);
4396
4397 tw32(0xc4, val | (1 << 15));
4398 }
4399
4400 if ((tp->nic_sram_data_cfg & NIC_SRAM_DATA_CFG_MINI_PCI) != 0 &&
4401 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705) {
4402 tp->pci_clock_ctrl |= CLOCK_CTRL_CLKRUN_OENABLE;
4403 if (tp->pci_chip_rev_id == CHIPREV_ID_5705_A0)
4404 tp->pci_clock_ctrl |= CLOCK_CTRL_FORCE_CLKRUN;
4405 tw32(TG3PCI_CLOCK_CTRL, tp->pci_clock_ctrl);
4406 }
4407
4408 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES) {
4409 tp->mac_mode = MAC_MODE_PORT_MODE_TBI;
4410 tw32_f(MAC_MODE, tp->mac_mode);
747e8f8b
MC
4411 } else if (tp->tg3_flags2 & TG3_FLG2_MII_SERDES) {
4412 tp->mac_mode = MAC_MODE_PORT_MODE_GMII;
4413 tw32_f(MAC_MODE, tp->mac_mode);
1da177e4
LT
4414 } else
4415 tw32_f(MAC_MODE, 0);
4416 udelay(40);
4417
4418 if (!(tp->tg3_flags2 & TG3_FLG2_SUN_570X)) {
4419 /* Wait for firmware initialization to complete. */
4420 for (i = 0; i < 100000; i++) {
4421 tg3_read_mem(tp, NIC_SRAM_FIRMWARE_MBOX, &val);
4422 if (val == ~NIC_SRAM_FIRMWARE_MBOX_MAGIC1)
4423 break;
4424 udelay(10);
4425 }
4426 if (i >= 100000) {
4427 printk(KERN_ERR PFX "tg3_reset_hw timed out for %s, "
4428 "firmware will not restart magic=%08x\n",
4429 tp->dev->name, val);
4430 return -ENODEV;
4431 }
4432 }
4433
4434 if ((tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) &&
4435 tp->pci_chip_rev_id != CHIPREV_ID_5750_A0) {
4436 u32 val = tr32(0x7c00);
4437
4438 tw32(0x7c00, val | (1 << 25));
4439 }
4440
4441 /* Reprobe ASF enable state. */
4442 tp->tg3_flags &= ~TG3_FLAG_ENABLE_ASF;
4443 tp->tg3_flags2 &= ~TG3_FLG2_ASF_NEW_HANDSHAKE;
4444 tg3_read_mem(tp, NIC_SRAM_DATA_SIG, &val);
4445 if (val == NIC_SRAM_DATA_SIG_MAGIC) {
4446 u32 nic_cfg;
4447
4448 tg3_read_mem(tp, NIC_SRAM_DATA_CFG, &nic_cfg);
4449 if (nic_cfg & NIC_SRAM_DATA_CFG_ASF_ENABLE) {
4450 tp->tg3_flags |= TG3_FLAG_ENABLE_ASF;
cbf46853 4451 if (tp->tg3_flags2 & TG3_FLG2_5750_PLUS)
1da177e4
LT
4452 tp->tg3_flags2 |= TG3_FLG2_ASF_NEW_HANDSHAKE;
4453 }
4454 }
4455
4456 return 0;
4457}
4458
4459/* tp->lock is held. */
4460static void tg3_stop_fw(struct tg3 *tp)
4461{
4462 if (tp->tg3_flags & TG3_FLAG_ENABLE_ASF) {
4463 u32 val;
4464 int i;
4465
4466 tg3_write_mem(tp, NIC_SRAM_FW_CMD_MBOX, FWCMD_NICDRV_PAUSE_FW);
4467 val = tr32(GRC_RX_CPU_EVENT);
4468 val |= (1 << 14);
4469 tw32(GRC_RX_CPU_EVENT, val);
4470
4471 /* Wait for RX cpu to ACK the event. */
4472 for (i = 0; i < 100; i++) {
4473 if (!(tr32(GRC_RX_CPU_EVENT) & (1 << 14)))
4474 break;
4475 udelay(1);
4476 }
4477 }
4478}
4479
4480/* tp->lock is held. */
944d980e 4481static int tg3_halt(struct tg3 *tp, int kind, int silent)
1da177e4
LT
4482{
4483 int err;
4484
4485 tg3_stop_fw(tp);
4486
944d980e 4487 tg3_write_sig_pre_reset(tp, kind);
1da177e4 4488
b3b7d6be 4489 tg3_abort_hw(tp, silent);
1da177e4
LT
4490 err = tg3_chip_reset(tp);
4491
944d980e
MC
4492 tg3_write_sig_legacy(tp, kind);
4493 tg3_write_sig_post_reset(tp, kind);
1da177e4
LT
4494
4495 if (err)
4496 return err;
4497
4498 return 0;
4499}
4500
4501#define TG3_FW_RELEASE_MAJOR 0x0
4502#define TG3_FW_RELASE_MINOR 0x0
4503#define TG3_FW_RELEASE_FIX 0x0
4504#define TG3_FW_START_ADDR 0x08000000
4505#define TG3_FW_TEXT_ADDR 0x08000000
4506#define TG3_FW_TEXT_LEN 0x9c0
4507#define TG3_FW_RODATA_ADDR 0x080009c0
4508#define TG3_FW_RODATA_LEN 0x60
4509#define TG3_FW_DATA_ADDR 0x08000a40
4510#define TG3_FW_DATA_LEN 0x20
4511#define TG3_FW_SBSS_ADDR 0x08000a60
4512#define TG3_FW_SBSS_LEN 0xc
4513#define TG3_FW_BSS_ADDR 0x08000a70
4514#define TG3_FW_BSS_LEN 0x10
4515
4516static u32 tg3FwText[(TG3_FW_TEXT_LEN / sizeof(u32)) + 1] = {
4517 0x00000000, 0x10000003, 0x00000000, 0x0000000d, 0x0000000d, 0x3c1d0800,
4518 0x37bd3ffc, 0x03a0f021, 0x3c100800, 0x26100000, 0x0e000018, 0x00000000,
4519 0x0000000d, 0x3c1d0800, 0x37bd3ffc, 0x03a0f021, 0x3c100800, 0x26100034,
4520 0x0e00021c, 0x00000000, 0x0000000d, 0x00000000, 0x00000000, 0x00000000,
4521 0x27bdffe0, 0x3c1cc000, 0xafbf0018, 0xaf80680c, 0x0e00004c, 0x241b2105,
4522 0x97850000, 0x97870002, 0x9782002c, 0x9783002e, 0x3c040800, 0x248409c0,
4523 0xafa00014, 0x00021400, 0x00621825, 0x00052c00, 0xafa30010, 0x8f860010,
4524 0x00e52825, 0x0e000060, 0x24070102, 0x3c02ac00, 0x34420100, 0x3c03ac01,
4525 0x34630100, 0xaf820490, 0x3c02ffff, 0xaf820494, 0xaf830498, 0xaf82049c,
4526 0x24020001, 0xaf825ce0, 0x0e00003f, 0xaf825d00, 0x0e000140, 0x00000000,
4527 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x2402ffff, 0xaf825404, 0x8f835400,
4528 0x34630400, 0xaf835400, 0xaf825404, 0x3c020800, 0x24420034, 0xaf82541c,
4529 0x03e00008, 0xaf805400, 0x00000000, 0x00000000, 0x3c020800, 0x34423000,
4530 0x3c030800, 0x34633000, 0x3c040800, 0x348437ff, 0x3c010800, 0xac220a64,
4531 0x24020040, 0x3c010800, 0xac220a68, 0x3c010800, 0xac200a60, 0xac600000,
4532 0x24630004, 0x0083102b, 0x5040fffd, 0xac600000, 0x03e00008, 0x00000000,
4533 0x00804821, 0x8faa0010, 0x3c020800, 0x8c420a60, 0x3c040800, 0x8c840a68,
4534 0x8fab0014, 0x24430001, 0x0044102b, 0x3c010800, 0xac230a60, 0x14400003,
4535 0x00004021, 0x3c010800, 0xac200a60, 0x3c020800, 0x8c420a60, 0x3c030800,
4536 0x8c630a64, 0x91240000, 0x00021140, 0x00431021, 0x00481021, 0x25080001,
4537 0xa0440000, 0x29020008, 0x1440fff4, 0x25290001, 0x3c020800, 0x8c420a60,
4538 0x3c030800, 0x8c630a64, 0x8f84680c, 0x00021140, 0x00431021, 0xac440008,
4539 0xac45000c, 0xac460010, 0xac470014, 0xac4a0018, 0x03e00008, 0xac4b001c,
4540 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
4541 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
4542 0, 0, 0, 0, 0, 0,
4543 0x02000008, 0x00000000, 0x0a0001e3, 0x3c0a0001, 0x0a0001e3, 0x3c0a0002,
4544 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
4545 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
4546 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
4547 0x0a0001e3, 0x3c0a0007, 0x0a0001e3, 0x3c0a0008, 0x0a0001e3, 0x3c0a0009,
4548 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x3c0a000b,
4549 0x0a0001e3, 0x3c0a000c, 0x0a0001e3, 0x3c0a000d, 0x0a0001e3, 0x00000000,
4550 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x3c0a000e, 0x0a0001e3, 0x00000000,
4551 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
4552 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
4553 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x3c0a0013, 0x0a0001e3, 0x3c0a0014,
4554 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
4555 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
4556 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
4557 0x27bdffe0, 0x00001821, 0x00001021, 0xafbf0018, 0xafb10014, 0xafb00010,
4558 0x3c010800, 0x00220821, 0xac200a70, 0x3c010800, 0x00220821, 0xac200a74,
4559 0x3c010800, 0x00220821, 0xac200a78, 0x24630001, 0x1860fff5, 0x2442000c,
4560 0x24110001, 0x8f906810, 0x32020004, 0x14400005, 0x24040001, 0x3c020800,
4561 0x8c420a78, 0x18400003, 0x00002021, 0x0e000182, 0x00000000, 0x32020001,
4562 0x10400003, 0x00000000, 0x0e000169, 0x00000000, 0x0a000153, 0xaf915028,
4563 0x8fbf0018, 0x8fb10014, 0x8fb00010, 0x03e00008, 0x27bd0020, 0x3c050800,
4564 0x8ca50a70, 0x3c060800, 0x8cc60a80, 0x3c070800, 0x8ce70a78, 0x27bdffe0,
4565 0x3c040800, 0x248409d0, 0xafbf0018, 0xafa00010, 0x0e000060, 0xafa00014,
4566 0x0e00017b, 0x00002021, 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x24020001,
4567 0x8f836810, 0x00821004, 0x00021027, 0x00621824, 0x03e00008, 0xaf836810,
4568 0x27bdffd8, 0xafbf0024, 0x1080002e, 0xafb00020, 0x8f825cec, 0xafa20018,
4569 0x8f825cec, 0x3c100800, 0x26100a78, 0xafa2001c, 0x34028000, 0xaf825cec,
4570 0x8e020000, 0x18400016, 0x00000000, 0x3c020800, 0x94420a74, 0x8fa3001c,
4571 0x000221c0, 0xac830004, 0x8fa2001c, 0x3c010800, 0x0e000201, 0xac220a74,
4572 0x10400005, 0x00000000, 0x8e020000, 0x24420001, 0x0a0001df, 0xae020000,
4573 0x3c020800, 0x8c420a70, 0x00021c02, 0x000321c0, 0x0a0001c5, 0xafa2001c,
4574 0x0e000201, 0x00000000, 0x1040001f, 0x00000000, 0x8e020000, 0x8fa3001c,
4575 0x24420001, 0x3c010800, 0xac230a70, 0x3c010800, 0xac230a74, 0x0a0001df,
4576 0xae020000, 0x3c100800, 0x26100a78, 0x8e020000, 0x18400028, 0x00000000,
4577 0x0e000201, 0x00000000, 0x14400024, 0x00000000, 0x8e020000, 0x3c030800,
4578 0x8c630a70, 0x2442ffff, 0xafa3001c, 0x18400006, 0xae020000, 0x00031402,
4579 0x000221c0, 0x8c820004, 0x3c010800, 0xac220a70, 0x97a2001e, 0x2442ff00,
4580 0x2c420300, 0x1440000b, 0x24024000, 0x3c040800, 0x248409dc, 0xafa00010,
4581 0xafa00014, 0x8fa6001c, 0x24050008, 0x0e000060, 0x00003821, 0x0a0001df,
4582 0x00000000, 0xaf825cf8, 0x3c020800, 0x8c420a40, 0x8fa3001c, 0x24420001,
4583 0xaf835cf8, 0x3c010800, 0xac220a40, 0x8fbf0024, 0x8fb00020, 0x03e00008,
4584 0x27bd0028, 0x27bdffe0, 0x3c040800, 0x248409e8, 0x00002821, 0x00003021,
4585 0x00003821, 0xafbf0018, 0xafa00010, 0x0e000060, 0xafa00014, 0x8fbf0018,
4586 0x03e00008, 0x27bd0020, 0x8f82680c, 0x8f85680c, 0x00021827, 0x0003182b,
4587 0x00031823, 0x00431024, 0x00441021, 0x00a2282b, 0x10a00006, 0x00000000,
4588 0x00401821, 0x8f82680c, 0x0043102b, 0x1440fffd, 0x00000000, 0x03e00008,
4589 0x00000000, 0x3c040800, 0x8c840000, 0x3c030800, 0x8c630a40, 0x0064102b,
4590 0x54400002, 0x00831023, 0x00641023, 0x2c420008, 0x03e00008, 0x38420001,
4591 0x27bdffe0, 0x00802821, 0x3c040800, 0x24840a00, 0x00003021, 0x00003821,
4592 0xafbf0018, 0xafa00010, 0x0e000060, 0xafa00014, 0x0a000216, 0x00000000,
4593 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x00000000, 0x27bdffe0, 0x3c1cc000,
4594 0xafbf0018, 0x0e00004c, 0xaf80680c, 0x3c040800, 0x24840a10, 0x03802821,
4595 0x00003021, 0x00003821, 0xafa00010, 0x0e000060, 0xafa00014, 0x2402ffff,
4596 0xaf825404, 0x3c0200aa, 0x0e000234, 0xaf825434, 0x8fbf0018, 0x03e00008,
4597 0x27bd0020, 0x00000000, 0x00000000, 0x00000000, 0x27bdffe8, 0xafb00010,
4598 0x24100001, 0xafbf0014, 0x3c01c003, 0xac200000, 0x8f826810, 0x30422000,
4599 0x10400003, 0x00000000, 0x0e000246, 0x00000000, 0x0a00023a, 0xaf905428,
4600 0x8fbf0014, 0x8fb00010, 0x03e00008, 0x27bd0018, 0x27bdfff8, 0x8f845d0c,
4601 0x3c0200ff, 0x3c030800, 0x8c630a50, 0x3442fff8, 0x00821024, 0x1043001e,
4602 0x3c0500ff, 0x34a5fff8, 0x3c06c003, 0x3c074000, 0x00851824, 0x8c620010,
4603 0x3c010800, 0xac230a50, 0x30420008, 0x10400005, 0x00871025, 0x8cc20000,
4604 0x24420001, 0xacc20000, 0x00871025, 0xaf825d0c, 0x8fa20000, 0x24420001,
4605 0xafa20000, 0x8fa20000, 0x8fa20000, 0x24420001, 0xafa20000, 0x8fa20000,
4606 0x8f845d0c, 0x3c030800, 0x8c630a50, 0x00851024, 0x1443ffe8, 0x00851824,
4607 0x27bd0008, 0x03e00008, 0x00000000, 0x00000000, 0x00000000
4608};
4609
4610static u32 tg3FwRodata[(TG3_FW_RODATA_LEN / sizeof(u32)) + 1] = {
4611 0x35373031, 0x726c7341, 0x00000000, 0x00000000, 0x53774576, 0x656e7430,
4612 0x00000000, 0x726c7045, 0x76656e74, 0x31000000, 0x556e6b6e, 0x45766e74,
4613 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x66617461, 0x6c457272,
4614 0x00000000, 0x00000000, 0x4d61696e, 0x43707542, 0x00000000, 0x00000000,
4615 0x00000000
4616};
4617
4618#if 0 /* All zeros, don't eat up space with it. */
4619u32 tg3FwData[(TG3_FW_DATA_LEN / sizeof(u32)) + 1] = {
4620 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x00000000,
4621 0x00000000, 0x00000000, 0x00000000, 0x00000000
4622};
4623#endif
4624
4625#define RX_CPU_SCRATCH_BASE 0x30000
4626#define RX_CPU_SCRATCH_SIZE 0x04000
4627#define TX_CPU_SCRATCH_BASE 0x34000
4628#define TX_CPU_SCRATCH_SIZE 0x04000
4629
4630/* tp->lock is held. */
4631static int tg3_halt_cpu(struct tg3 *tp, u32 offset)
4632{
4633 int i;
4634
4635 if (offset == TX_CPU_BASE &&
4636 (tp->tg3_flags2 & TG3_FLG2_5705_PLUS))
4637 BUG();
4638
4639 if (offset == RX_CPU_BASE) {
4640 for (i = 0; i < 10000; i++) {
4641 tw32(offset + CPU_STATE, 0xffffffff);
4642 tw32(offset + CPU_MODE, CPU_MODE_HALT);
4643 if (tr32(offset + CPU_MODE) & CPU_MODE_HALT)
4644 break;
4645 }
4646
4647 tw32(offset + CPU_STATE, 0xffffffff);
4648 tw32_f(offset + CPU_MODE, CPU_MODE_HALT);
4649 udelay(10);
4650 } else {
4651 for (i = 0; i < 10000; i++) {
4652 tw32(offset + CPU_STATE, 0xffffffff);
4653 tw32(offset + CPU_MODE, CPU_MODE_HALT);
4654 if (tr32(offset + CPU_MODE) & CPU_MODE_HALT)
4655 break;
4656 }
4657 }
4658
4659 if (i >= 10000) {
4660 printk(KERN_ERR PFX "tg3_reset_cpu timed out for %s, "
4661 "and %s CPU\n",
4662 tp->dev->name,
4663 (offset == RX_CPU_BASE ? "RX" : "TX"));
4664 return -ENODEV;
4665 }
4666 return 0;
4667}
4668
4669struct fw_info {
4670 unsigned int text_base;
4671 unsigned int text_len;
4672 u32 *text_data;
4673 unsigned int rodata_base;
4674 unsigned int rodata_len;
4675 u32 *rodata_data;
4676 unsigned int data_base;
4677 unsigned int data_len;
4678 u32 *data_data;
4679};
4680
4681/* tp->lock is held. */
4682static int tg3_load_firmware_cpu(struct tg3 *tp, u32 cpu_base, u32 cpu_scratch_base,
4683 int cpu_scratch_size, struct fw_info *info)
4684{
4685 int err, i;
1da177e4
LT
4686 void (*write_op)(struct tg3 *, u32, u32);
4687
4688 if (cpu_base == TX_CPU_BASE &&
4689 (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
4690 printk(KERN_ERR PFX "tg3_load_firmware_cpu: Trying to load "
4691 "TX cpu firmware on %s which is 5705.\n",
4692 tp->dev->name);
4693 return -EINVAL;
4694 }
4695
4696 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)
4697 write_op = tg3_write_mem;
4698 else
4699 write_op = tg3_write_indirect_reg32;
4700
1b628151
MC
4701 /* It is possible that bootcode is still loading at this point.
4702 * Get the nvram lock first before halting the cpu.
4703 */
4704 tg3_nvram_lock(tp);
1da177e4 4705 err = tg3_halt_cpu(tp, cpu_base);
1b628151 4706 tg3_nvram_unlock(tp);
1da177e4
LT
4707 if (err)
4708 goto out;
4709
4710 for (i = 0; i < cpu_scratch_size; i += sizeof(u32))
4711 write_op(tp, cpu_scratch_base + i, 0);
4712 tw32(cpu_base + CPU_STATE, 0xffffffff);
4713 tw32(cpu_base + CPU_MODE, tr32(cpu_base+CPU_MODE)|CPU_MODE_HALT);
4714 for (i = 0; i < (info->text_len / sizeof(u32)); i++)
4715 write_op(tp, (cpu_scratch_base +
4716 (info->text_base & 0xffff) +
4717 (i * sizeof(u32))),
4718 (info->text_data ?
4719 info->text_data[i] : 0));
4720 for (i = 0; i < (info->rodata_len / sizeof(u32)); i++)
4721 write_op(tp, (cpu_scratch_base +
4722 (info->rodata_base & 0xffff) +
4723 (i * sizeof(u32))),
4724 (info->rodata_data ?
4725 info->rodata_data[i] : 0));
4726 for (i = 0; i < (info->data_len / sizeof(u32)); i++)
4727 write_op(tp, (cpu_scratch_base +
4728 (info->data_base & 0xffff) +
4729 (i * sizeof(u32))),
4730 (info->data_data ?
4731 info->data_data[i] : 0));
4732
4733 err = 0;
4734
4735out:
1da177e4
LT
4736 return err;
4737}
4738
4739/* tp->lock is held. */
4740static int tg3_load_5701_a0_firmware_fix(struct tg3 *tp)
4741{
4742 struct fw_info info;
4743 int err, i;
4744
4745 info.text_base = TG3_FW_TEXT_ADDR;
4746 info.text_len = TG3_FW_TEXT_LEN;
4747 info.text_data = &tg3FwText[0];
4748 info.rodata_base = TG3_FW_RODATA_ADDR;
4749 info.rodata_len = TG3_FW_RODATA_LEN;
4750 info.rodata_data = &tg3FwRodata[0];
4751 info.data_base = TG3_FW_DATA_ADDR;
4752 info.data_len = TG3_FW_DATA_LEN;
4753 info.data_data = NULL;
4754
4755 err = tg3_load_firmware_cpu(tp, RX_CPU_BASE,
4756 RX_CPU_SCRATCH_BASE, RX_CPU_SCRATCH_SIZE,
4757 &info);
4758 if (err)
4759 return err;
4760
4761 err = tg3_load_firmware_cpu(tp, TX_CPU_BASE,
4762 TX_CPU_SCRATCH_BASE, TX_CPU_SCRATCH_SIZE,
4763 &info);
4764 if (err)
4765 return err;
4766
4767 /* Now startup only the RX cpu. */
4768 tw32(RX_CPU_BASE + CPU_STATE, 0xffffffff);
4769 tw32_f(RX_CPU_BASE + CPU_PC, TG3_FW_TEXT_ADDR);
4770
4771 for (i = 0; i < 5; i++) {
4772 if (tr32(RX_CPU_BASE + CPU_PC) == TG3_FW_TEXT_ADDR)
4773 break;
4774 tw32(RX_CPU_BASE + CPU_STATE, 0xffffffff);
4775 tw32(RX_CPU_BASE + CPU_MODE, CPU_MODE_HALT);
4776 tw32_f(RX_CPU_BASE + CPU_PC, TG3_FW_TEXT_ADDR);
4777 udelay(1000);
4778 }
4779 if (i >= 5) {
4780 printk(KERN_ERR PFX "tg3_load_firmware fails for %s "
4781 "to set RX CPU PC, is %08x should be %08x\n",
4782 tp->dev->name, tr32(RX_CPU_BASE + CPU_PC),
4783 TG3_FW_TEXT_ADDR);
4784 return -ENODEV;
4785 }
4786 tw32(RX_CPU_BASE + CPU_STATE, 0xffffffff);
4787 tw32_f(RX_CPU_BASE + CPU_MODE, 0x00000000);
4788
4789 return 0;
4790}
4791
4792#if TG3_TSO_SUPPORT != 0
4793
4794#define TG3_TSO_FW_RELEASE_MAJOR 0x1
4795#define TG3_TSO_FW_RELASE_MINOR 0x6
4796#define TG3_TSO_FW_RELEASE_FIX 0x0
4797#define TG3_TSO_FW_START_ADDR 0x08000000
4798#define TG3_TSO_FW_TEXT_ADDR 0x08000000
4799#define TG3_TSO_FW_TEXT_LEN 0x1aa0
4800#define TG3_TSO_FW_RODATA_ADDR 0x08001aa0
4801#define TG3_TSO_FW_RODATA_LEN 0x60
4802#define TG3_TSO_FW_DATA_ADDR 0x08001b20
4803#define TG3_TSO_FW_DATA_LEN 0x30
4804#define TG3_TSO_FW_SBSS_ADDR 0x08001b50
4805#define TG3_TSO_FW_SBSS_LEN 0x2c
4806#define TG3_TSO_FW_BSS_ADDR 0x08001b80
4807#define TG3_TSO_FW_BSS_LEN 0x894
4808
4809static u32 tg3TsoFwText[(TG3_TSO_FW_TEXT_LEN / 4) + 1] = {
4810 0x0e000003, 0x00000000, 0x08001b24, 0x00000000, 0x10000003, 0x00000000,
4811 0x0000000d, 0x0000000d, 0x3c1d0800, 0x37bd4000, 0x03a0f021, 0x3c100800,
4812 0x26100000, 0x0e000010, 0x00000000, 0x0000000d, 0x27bdffe0, 0x3c04fefe,
4813 0xafbf0018, 0x0e0005d8, 0x34840002, 0x0e000668, 0x00000000, 0x3c030800,
4814 0x90631b68, 0x24020002, 0x3c040800, 0x24841aac, 0x14620003, 0x24050001,
4815 0x3c040800, 0x24841aa0, 0x24060006, 0x00003821, 0xafa00010, 0x0e00067c,
4816 0xafa00014, 0x8f625c50, 0x34420001, 0xaf625c50, 0x8f625c90, 0x34420001,
4817 0xaf625c90, 0x2402ffff, 0x0e000034, 0xaf625404, 0x8fbf0018, 0x03e00008,
4818 0x27bd0020, 0x00000000, 0x00000000, 0x00000000, 0x27bdffe0, 0xafbf001c,
4819 0xafb20018, 0xafb10014, 0x0e00005b, 0xafb00010, 0x24120002, 0x24110001,
4820 0x8f706820, 0x32020100, 0x10400003, 0x00000000, 0x0e0000bb, 0x00000000,
4821 0x8f706820, 0x32022000, 0x10400004, 0x32020001, 0x0e0001f0, 0x24040001,
4822 0x32020001, 0x10400003, 0x00000000, 0x0e0000a3, 0x00000000, 0x3c020800,
4823 0x90421b98, 0x14520003, 0x00000000, 0x0e0004c0, 0x00000000, 0x0a00003c,
4824 0xaf715028, 0x8fbf001c, 0x8fb20018, 0x8fb10014, 0x8fb00010, 0x03e00008,
4825 0x27bd0020, 0x27bdffe0, 0x3c040800, 0x24841ac0, 0x00002821, 0x00003021,
4826 0x00003821, 0xafbf0018, 0xafa00010, 0x0e00067c, 0xafa00014, 0x3c040800,
4827 0x248423d8, 0xa4800000, 0x3c010800, 0xa0201b98, 0x3c010800, 0xac201b9c,
4828 0x3c010800, 0xac201ba0, 0x3c010800, 0xac201ba4, 0x3c010800, 0xac201bac,
4829 0x3c010800, 0xac201bb8, 0x3c010800, 0xac201bbc, 0x8f624434, 0x3c010800,
4830 0xac221b88, 0x8f624438, 0x3c010800, 0xac221b8c, 0x8f624410, 0xac80f7a8,
4831 0x3c010800, 0xac201b84, 0x3c010800, 0xac2023e0, 0x3c010800, 0xac2023c8,
4832 0x3c010800, 0xac2023cc, 0x3c010800, 0xac202400, 0x3c010800, 0xac221b90,
4833 0x8f620068, 0x24030007, 0x00021702, 0x10430005, 0x00000000, 0x8f620068,
4834 0x00021702, 0x14400004, 0x24020001, 0x3c010800, 0x0a000097, 0xac20240c,
4835 0xac820034, 0x3c040800, 0x24841acc, 0x3c050800, 0x8ca5240c, 0x00003021,
4836 0x00003821, 0xafa00010, 0x0e00067c, 0xafa00014, 0x8fbf0018, 0x03e00008,
4837 0x27bd0020, 0x27bdffe0, 0x3c040800, 0x24841ad8, 0x00002821, 0x00003021,
4838 0x00003821, 0xafbf0018, 0xafa00010, 0x0e00067c, 0xafa00014, 0x0e00005b,
4839 0x00000000, 0x0e0000b4, 0x00002021, 0x8fbf0018, 0x03e00008, 0x27bd0020,
4840 0x24020001, 0x8f636820, 0x00821004, 0x00021027, 0x00621824, 0x03e00008,
4841 0xaf636820, 0x27bdffd0, 0xafbf002c, 0xafb60028, 0xafb50024, 0xafb40020,
4842 0xafb3001c, 0xafb20018, 0xafb10014, 0xafb00010, 0x8f675c5c, 0x3c030800,
4843 0x24631bbc, 0x8c620000, 0x14470005, 0x3c0200ff, 0x3c020800, 0x90421b98,
4844 0x14400119, 0x3c0200ff, 0x3442fff8, 0x00e28824, 0xac670000, 0x00111902,
4845 0x306300ff, 0x30e20003, 0x000211c0, 0x00622825, 0x00a04021, 0x00071602,
4846 0x3c030800, 0x90631b98, 0x3044000f, 0x14600036, 0x00804821, 0x24020001,
4847 0x3c010800, 0xa0221b98, 0x00051100, 0x00821025, 0x3c010800, 0xac201b9c,
4848 0x3c010800, 0xac201ba0, 0x3c010800, 0xac201ba4, 0x3c010800, 0xac201bac,
4849 0x3c010800, 0xac201bb8, 0x3c010800, 0xac201bb0, 0x3c010800, 0xac201bb4,
4850 0x3c010800, 0xa42223d8, 0x9622000c, 0x30437fff, 0x3c010800, 0xa4222410,
4851 0x30428000, 0x3c010800, 0xa4231bc6, 0x10400005, 0x24020001, 0x3c010800,
4852 0xac2223f4, 0x0a000102, 0x2406003e, 0x24060036, 0x3c010800, 0xac2023f4,
4853 0x9622000a, 0x3c030800, 0x94631bc6, 0x3c010800, 0xac2023f0, 0x3c010800,
4854 0xac2023f8, 0x00021302, 0x00021080, 0x00c21021, 0x00621821, 0x3c010800,
4855 0xa42223d0, 0x3c010800, 0x0a000115, 0xa4231b96, 0x9622000c, 0x3c010800,
4856 0xa42223ec, 0x3c040800, 0x24841b9c, 0x8c820000, 0x00021100, 0x3c010800,
4857 0x00220821, 0xac311bc8, 0x8c820000, 0x00021100, 0x3c010800, 0x00220821,
4858 0xac271bcc, 0x8c820000, 0x25030001, 0x306601ff, 0x00021100, 0x3c010800,
4859 0x00220821, 0xac261bd0, 0x8c820000, 0x00021100, 0x3c010800, 0x00220821,
4860 0xac291bd4, 0x96230008, 0x3c020800, 0x8c421bac, 0x00432821, 0x3c010800,
4861 0xac251bac, 0x9622000a, 0x30420004, 0x14400018, 0x00061100, 0x8f630c14,
4862 0x3063000f, 0x2c620002, 0x1440000b, 0x3c02c000, 0x8f630c14, 0x3c020800,
4863 0x8c421b40, 0x3063000f, 0x24420001, 0x3c010800, 0xac221b40, 0x2c620002,
4864 0x1040fff7, 0x3c02c000, 0x00e21825, 0xaf635c5c, 0x8f625c50, 0x30420002,
4865 0x10400014, 0x00000000, 0x0a000147, 0x00000000, 0x3c030800, 0x8c631b80,
4866 0x3c040800, 0x94841b94, 0x01221025, 0x3c010800, 0xa42223da, 0x24020001,
4867 0x3c010800, 0xac221bb8, 0x24630001, 0x0085202a, 0x3c010800, 0x10800003,
4868 0xac231b80, 0x3c010800, 0xa4251b94, 0x3c060800, 0x24c61b9c, 0x8cc20000,
4869 0x24420001, 0xacc20000, 0x28420080, 0x14400005, 0x00000000, 0x0e000656,
4870 0x24040002, 0x0a0001e6, 0x00000000, 0x3c020800, 0x8c421bb8, 0x10400078,
4871 0x24020001, 0x3c050800, 0x90a51b98, 0x14a20072, 0x00000000, 0x3c150800,
4872 0x96b51b96, 0x3c040800, 0x8c841bac, 0x32a3ffff, 0x0083102a, 0x1440006c,
4873 0x00000000, 0x14830003, 0x00000000, 0x3c010800, 0xac2523f0, 0x1060005c,
4874 0x00009021, 0x24d60004, 0x0060a021, 0x24d30014, 0x8ec20000, 0x00028100,
4875 0x3c110800, 0x02308821, 0x0e000625, 0x8e311bc8, 0x00402821, 0x10a00054,
4876 0x00000000, 0x9628000a, 0x31020040, 0x10400005, 0x2407180c, 0x8e22000c,
4877 0x2407188c, 0x00021400, 0xaca20018, 0x3c030800, 0x00701821, 0x8c631bd0,
4878 0x3c020800, 0x00501021, 0x8c421bd4, 0x00031d00, 0x00021400, 0x00621825,
4879 0xaca30014, 0x8ec30004, 0x96220008, 0x00432023, 0x3242ffff, 0x3083ffff,
4880 0x00431021, 0x0282102a, 0x14400002, 0x02b23023, 0x00803021, 0x8e620000,
4881 0x30c4ffff, 0x00441021, 0xae620000, 0x8e220000, 0xaca20000, 0x8e220004,
4882 0x8e63fff4, 0x00431021, 0xaca20004, 0xa4a6000e, 0x8e62fff4, 0x00441021,
4883 0xae62fff4, 0x96230008, 0x0043102a, 0x14400005, 0x02469021, 0x8e62fff0,
4884 0xae60fff4, 0x24420001, 0xae62fff0, 0xaca00008, 0x3242ffff, 0x14540008,
4885 0x24020305, 0x31020080, 0x54400001, 0x34e70010, 0x24020905, 0xa4a2000c,
4886 0x0a0001cb, 0x34e70020, 0xa4a2000c, 0x3c020800, 0x8c4223f0, 0x10400003,
4887 0x3c024b65, 0x0a0001d3, 0x34427654, 0x3c02b49a, 0x344289ab, 0xaca2001c,
4888 0x30e2ffff, 0xaca20010, 0x0e0005a2, 0x00a02021, 0x3242ffff, 0x0054102b,
4889 0x1440ffa9, 0x00000000, 0x24020002, 0x3c010800, 0x0a0001e6, 0xa0221b98,
4890 0x8ec2083c, 0x24420001, 0x0a0001e6, 0xaec2083c, 0x0e0004c0, 0x00000000,
4891 0x8fbf002c, 0x8fb60028, 0x8fb50024, 0x8fb40020, 0x8fb3001c, 0x8fb20018,
4892 0x8fb10014, 0x8fb00010, 0x03e00008, 0x27bd0030, 0x27bdffd0, 0xafbf0028,
4893 0xafb30024, 0xafb20020, 0xafb1001c, 0xafb00018, 0x8f725c9c, 0x3c0200ff,
4894 0x3442fff8, 0x3c070800, 0x24e71bb4, 0x02428824, 0x9623000e, 0x8ce20000,
4895 0x00431021, 0xace20000, 0x8e220010, 0x30420020, 0x14400011, 0x00809821,
4896 0x0e00063b, 0x02202021, 0x3c02c000, 0x02421825, 0xaf635c9c, 0x8f625c90,
4897 0x30420002, 0x1040011e, 0x00000000, 0xaf635c9c, 0x8f625c90, 0x30420002,
4898 0x10400119, 0x00000000, 0x0a00020d, 0x00000000, 0x8e240008, 0x8e230014,
4899 0x00041402, 0x000231c0, 0x00031502, 0x304201ff, 0x2442ffff, 0x3042007f,
4900 0x00031942, 0x30637800, 0x00021100, 0x24424000, 0x00624821, 0x9522000a,
4901 0x3084ffff, 0x30420008, 0x104000b0, 0x000429c0, 0x3c020800, 0x8c422400,
4902 0x14400024, 0x24c50008, 0x94c20014, 0x3c010800, 0xa42223d0, 0x8cc40010,
4903 0x00041402, 0x3c010800, 0xa42223d2, 0x3c010800, 0xa42423d4, 0x94c2000e,
4904 0x3083ffff, 0x00431023, 0x3c010800, 0xac222408, 0x94c2001a, 0x3c010800,
4905 0xac262400, 0x3c010800, 0xac322404, 0x3c010800, 0xac2223fc, 0x3c02c000,
4906 0x02421825, 0xaf635c9c, 0x8f625c90, 0x30420002, 0x104000e5, 0x00000000,
4907 0xaf635c9c, 0x8f625c90, 0x30420002, 0x104000e0, 0x00000000, 0x0a000246,
4908 0x00000000, 0x94c2000e, 0x3c030800, 0x946323d4, 0x00434023, 0x3103ffff,
4909 0x2c620008, 0x1040001c, 0x00000000, 0x94c20014, 0x24420028, 0x00a22821,
4910 0x00031042, 0x1840000b, 0x00002021, 0x24e60848, 0x00403821, 0x94a30000,
4911 0x8cc20000, 0x24840001, 0x00431021, 0xacc20000, 0x0087102a, 0x1440fff9,
4912 0x24a50002, 0x31020001, 0x1040001f, 0x3c024000, 0x3c040800, 0x248423fc,
4913 0xa0a00001, 0x94a30000, 0x8c820000, 0x00431021, 0x0a000285, 0xac820000,
4914 0x8f626800, 0x3c030010, 0x00431024, 0x10400009, 0x00000000, 0x94c2001a,
4915 0x3c030800, 0x8c6323fc, 0x00431021, 0x3c010800, 0xac2223fc, 0x0a000286,
4916 0x3c024000, 0x94c2001a, 0x94c4001c, 0x3c030800, 0x8c6323fc, 0x00441023,
4917 0x00621821, 0x3c010800, 0xac2323fc, 0x3c024000, 0x02421825, 0xaf635c9c,
4918 0x8f625c90, 0x30420002, 0x1440fffc, 0x00000000, 0x9522000a, 0x30420010,
4919 0x1040009b, 0x00000000, 0x3c030800, 0x946323d4, 0x3c070800, 0x24e72400,
4920 0x8ce40000, 0x8f626800, 0x24630030, 0x00832821, 0x3c030010, 0x00431024,
4921 0x1440000a, 0x00000000, 0x94a20004, 0x3c040800, 0x8c842408, 0x3c030800,
4922 0x8c6323fc, 0x00441023, 0x00621821, 0x3c010800, 0xac2323fc, 0x3c040800,
4923 0x8c8423fc, 0x00041c02, 0x3082ffff, 0x00622021, 0x00041402, 0x00822021,
4924 0x00041027, 0xa4a20006, 0x3c030800, 0x8c632404, 0x3c0200ff, 0x3442fff8,
4925 0x00628824, 0x96220008, 0x24050001, 0x24034000, 0x000231c0, 0x00801021,
4926 0xa4c2001a, 0xa4c0001c, 0xace00000, 0x3c010800, 0xac251b60, 0xaf635cb8,
4927 0x8f625cb0, 0x30420002, 0x10400003, 0x00000000, 0x3c010800, 0xac201b60,
4928 0x8e220008, 0xaf625cb8, 0x8f625cb0, 0x30420002, 0x10400003, 0x00000000,
4929 0x3c010800, 0xac201b60, 0x3c020800, 0x8c421b60, 0x1040ffec, 0x00000000,
4930 0x3c040800, 0x0e00063b, 0x8c842404, 0x0a00032a, 0x00000000, 0x3c030800,
4931 0x90631b98, 0x24020002, 0x14620003, 0x3c034b65, 0x0a0002e1, 0x00008021,
4932 0x8e22001c, 0x34637654, 0x10430002, 0x24100002, 0x24100001, 0x00c02021,
4933 0x0e000350, 0x02003021, 0x24020003, 0x3c010800, 0xa0221b98, 0x24020002,
4934 0x1202000a, 0x24020001, 0x3c030800, 0x8c6323f0, 0x10620006, 0x00000000,
4935 0x3c020800, 0x944223d8, 0x00021400, 0x0a00031f, 0xae220014, 0x3c040800,
4936 0x248423da, 0x94820000, 0x00021400, 0xae220014, 0x3c020800, 0x8c421bbc,
4937 0x3c03c000, 0x3c010800, 0xa0201b98, 0x00431025, 0xaf625c5c, 0x8f625c50,
4938 0x30420002, 0x10400009, 0x00000000, 0x2484f7e2, 0x8c820000, 0x00431025,
4939 0xaf625c5c, 0x8f625c50, 0x30420002, 0x1440fffa, 0x00000000, 0x3c020800,
4940 0x24421b84, 0x8c430000, 0x24630001, 0xac430000, 0x8f630c14, 0x3063000f,
4941 0x2c620002, 0x1440000c, 0x3c024000, 0x8f630c14, 0x3c020800, 0x8c421b40,
4942 0x3063000f, 0x24420001, 0x3c010800, 0xac221b40, 0x2c620002, 0x1040fff7,
4943 0x00000000, 0x3c024000, 0x02421825, 0xaf635c9c, 0x8f625c90, 0x30420002,
4944 0x1440fffc, 0x00000000, 0x12600003, 0x00000000, 0x0e0004c0, 0x00000000,
4945 0x8fbf0028, 0x8fb30024, 0x8fb20020, 0x8fb1001c, 0x8fb00018, 0x03e00008,
4946 0x27bd0030, 0x8f634450, 0x3c040800, 0x24841b88, 0x8c820000, 0x00031c02,
4947 0x0043102b, 0x14400007, 0x3c038000, 0x8c840004, 0x8f624450, 0x00021c02,
4948 0x0083102b, 0x1040fffc, 0x3c038000, 0xaf634444, 0x8f624444, 0x00431024,
4949 0x1440fffd, 0x00000000, 0x8f624448, 0x03e00008, 0x3042ffff, 0x3c024000,
4950 0x00822025, 0xaf645c38, 0x8f625c30, 0x30420002, 0x1440fffc, 0x00000000,
4951 0x03e00008, 0x00000000, 0x27bdffe0, 0x00805821, 0x14c00011, 0x256e0008,
4952 0x3c020800, 0x8c4223f4, 0x10400007, 0x24020016, 0x3c010800, 0xa42223d2,
4953 0x2402002a, 0x3c010800, 0x0a000364, 0xa42223d4, 0x8d670010, 0x00071402,
4954 0x3c010800, 0xa42223d2, 0x3c010800, 0xa42723d4, 0x3c040800, 0x948423d4,
4955 0x3c030800, 0x946323d2, 0x95cf0006, 0x3c020800, 0x944223d0, 0x00832023,
4956 0x01e2c023, 0x3065ffff, 0x24a20028, 0x01c24821, 0x3082ffff, 0x14c0001a,
4957 0x01226021, 0x9582000c, 0x3042003f, 0x3c010800, 0xa42223d6, 0x95820004,
4958 0x95830006, 0x3c010800, 0xac2023e4, 0x3c010800, 0xac2023e8, 0x00021400,
4959 0x00431025, 0x3c010800, 0xac221bc0, 0x95220004, 0x3c010800, 0xa4221bc4,
4960 0x95230002, 0x01e51023, 0x0043102a, 0x10400010, 0x24020001, 0x3c010800,
4961 0x0a000398, 0xac2223f8, 0x3c030800, 0x8c6323e8, 0x3c020800, 0x94421bc4,
4962 0x00431021, 0xa5220004, 0x3c020800, 0x94421bc0, 0xa5820004, 0x3c020800,
4963 0x8c421bc0, 0xa5820006, 0x3c020800, 0x8c4223f0, 0x3c0d0800, 0x8dad23e4,
4964 0x3c0a0800, 0x144000e5, 0x8d4a23e8, 0x3c020800, 0x94421bc4, 0x004a1821,
4965 0x3063ffff, 0x0062182b, 0x24020002, 0x10c2000d, 0x01435023, 0x3c020800,
4966 0x944223d6, 0x30420009, 0x10400008, 0x00000000, 0x9582000c, 0x3042fff6,
4967 0xa582000c, 0x3c020800, 0x944223d6, 0x30420009, 0x01a26823, 0x3c020800,
4968 0x8c4223f8, 0x1040004a, 0x01203821, 0x3c020800, 0x944223d2, 0x00004021,
4969 0xa520000a, 0x01e21023, 0xa5220002, 0x3082ffff, 0x00021042, 0x18400008,
4970 0x00003021, 0x00401821, 0x94e20000, 0x25080001, 0x00c23021, 0x0103102a,
4971 0x1440fffb, 0x24e70002, 0x00061c02, 0x30c2ffff, 0x00623021, 0x00061402,
4972 0x00c23021, 0x00c02821, 0x00061027, 0xa522000a, 0x00003021, 0x2527000c,
4973 0x00004021, 0x94e20000, 0x25080001, 0x00c23021, 0x2d020004, 0x1440fffb,
4974 0x24e70002, 0x95220002, 0x00004021, 0x91230009, 0x00442023, 0x01803821,
4975 0x3082ffff, 0xa4e00010, 0x00621821, 0x00021042, 0x18400010, 0x00c33021,
4976 0x00404821, 0x94e20000, 0x24e70002, 0x00c23021, 0x30e2007f, 0x14400006,
4977 0x25080001, 0x8d630000, 0x3c02007f, 0x3442ff80, 0x00625824, 0x25670008,
4978 0x0109102a, 0x1440fff3, 0x00000000, 0x30820001, 0x10400005, 0x00061c02,
4979 0xa0e00001, 0x94e20000, 0x00c23021, 0x00061c02, 0x30c2ffff, 0x00623021,
4980 0x00061402, 0x00c23021, 0x0a00047d, 0x30c6ffff, 0x24020002, 0x14c20081,
4981 0x00000000, 0x3c020800, 0x8c42240c, 0x14400007, 0x00000000, 0x3c020800,
4982 0x944223d2, 0x95230002, 0x01e21023, 0x10620077, 0x00000000, 0x3c020800,
4983 0x944223d2, 0x01e21023, 0xa5220002, 0x3c020800, 0x8c42240c, 0x1040001a,
4984 0x31e3ffff, 0x8dc70010, 0x3c020800, 0x94421b96, 0x00e04021, 0x00072c02,
4985 0x00aa2021, 0x00431023, 0x00823823, 0x00072402, 0x30e2ffff, 0x00823821,
4986 0x00071027, 0xa522000a, 0x3102ffff, 0x3c040800, 0x948423d4, 0x00453023,
4987 0x00e02821, 0x00641823, 0x006d1821, 0x00c33021, 0x00061c02, 0x30c2ffff,
4988 0x0a00047d, 0x00623021, 0x01203821, 0x00004021, 0x3082ffff, 0x00021042,
4989 0x18400008, 0x00003021, 0x00401821, 0x94e20000, 0x25080001, 0x00c23021,
4990 0x0103102a, 0x1440fffb, 0x24e70002, 0x00061c02, 0x30c2ffff, 0x00623021,
4991 0x00061402, 0x00c23021, 0x00c02821, 0x00061027, 0xa522000a, 0x00003021,
4992 0x2527000c, 0x00004021, 0x94e20000, 0x25080001, 0x00c23021, 0x2d020004,
4993 0x1440fffb, 0x24e70002, 0x95220002, 0x00004021, 0x91230009, 0x00442023,
4994 0x01803821, 0x3082ffff, 0xa4e00010, 0x3c040800, 0x948423d4, 0x00621821,
4995 0x00c33021, 0x00061c02, 0x30c2ffff, 0x00623021, 0x00061c02, 0x3c020800,
4996 0x944223d0, 0x00c34821, 0x00441023, 0x00021fc2, 0x00431021, 0x00021043,
4997 0x18400010, 0x00003021, 0x00402021, 0x94e20000, 0x24e70002, 0x00c23021,
4998 0x30e2007f, 0x14400006, 0x25080001, 0x8d630000, 0x3c02007f, 0x3442ff80,
4999 0x00625824, 0x25670008, 0x0104102a, 0x1440fff3, 0x00000000, 0x3c020800,
5000 0x944223ec, 0x00c23021, 0x3122ffff, 0x00c23021, 0x00061c02, 0x30c2ffff,
5001 0x00623021, 0x00061402, 0x00c23021, 0x00c04021, 0x00061027, 0xa5820010,
5002 0xadc00014, 0x0a00049d, 0xadc00000, 0x8dc70010, 0x00e04021, 0x11400007,
5003 0x00072c02, 0x00aa3021, 0x00061402, 0x30c3ffff, 0x00433021, 0x00061402,
5004 0x00c22821, 0x00051027, 0xa522000a, 0x3c030800, 0x946323d4, 0x3102ffff,
5005 0x01e21021, 0x00433023, 0x00cd3021, 0x00061c02, 0x30c2ffff, 0x00623021,
5006 0x00061402, 0x00c23021, 0x00c04021, 0x00061027, 0xa5820010, 0x3102ffff,
5007 0x00051c00, 0x00431025, 0xadc20010, 0x3c020800, 0x8c4223f4, 0x10400005,
5008 0x2de205eb, 0x14400002, 0x25e2fff2, 0x34028870, 0xa5c20034, 0x3c030800,
5009 0x246323e8, 0x8c620000, 0x24420001, 0xac620000, 0x3c040800, 0x8c8423e4,
5010 0x3c020800, 0x8c421bc0, 0x3303ffff, 0x00832021, 0x00431821, 0x0062102b,
5011 0x3c010800, 0xac2423e4, 0x10400003, 0x2482ffff, 0x3c010800, 0xac2223e4,
5012 0x3c010800, 0xac231bc0, 0x03e00008, 0x27bd0020, 0x27bdffb8, 0x3c050800,
5013 0x24a51b96, 0xafbf0044, 0xafbe0040, 0xafb7003c, 0xafb60038, 0xafb50034,
5014 0xafb40030, 0xafb3002c, 0xafb20028, 0xafb10024, 0xafb00020, 0x94a90000,
5015 0x3c020800, 0x944223d0, 0x3c030800, 0x8c631bb0, 0x3c040800, 0x8c841bac,
5016 0x01221023, 0x0064182a, 0xa7a9001e, 0x106000be, 0xa7a20016, 0x24be0022,
5017 0x97b6001e, 0x24b3001a, 0x24b70016, 0x8fc20000, 0x14400008, 0x00000000,
5018 0x8fc2fff8, 0x97a30016, 0x8fc4fff4, 0x00431021, 0x0082202a, 0x148000b0,
5019 0x00000000, 0x97d50818, 0x32a2ffff, 0x104000a3, 0x00009021, 0x0040a021,
5020 0x00008821, 0x0e000625, 0x00000000, 0x00403021, 0x14c00007, 0x00000000,
5021 0x3c020800, 0x8c4223dc, 0x24420001, 0x3c010800, 0x0a000596, 0xac2223dc,
5022 0x3c100800, 0x02118021, 0x8e101bc8, 0x9608000a, 0x31020040, 0x10400005,
5023 0x2407180c, 0x8e02000c, 0x2407188c, 0x00021400, 0xacc20018, 0x31020080,
5024 0x54400001, 0x34e70010, 0x3c020800, 0x00511021, 0x8c421bd0, 0x3c030800,
5025 0x00711821, 0x8c631bd4, 0x00021500, 0x00031c00, 0x00431025, 0xacc20014,
5026 0x96040008, 0x3242ffff, 0x00821021, 0x0282102a, 0x14400002, 0x02b22823,
5027 0x00802821, 0x8e020000, 0x02459021, 0xacc20000, 0x8e020004, 0x00c02021,
5028 0x26310010, 0xac820004, 0x30e2ffff, 0xac800008, 0xa485000e, 0xac820010,
5029 0x24020305, 0x0e0005a2, 0xa482000c, 0x3242ffff, 0x0054102b, 0x1440ffc5,
5030 0x3242ffff, 0x0a00058e, 0x00000000, 0x8e620000, 0x8e63fffc, 0x0043102a,
5031 0x10400067, 0x00000000, 0x8e62fff0, 0x00028900, 0x3c100800, 0x02118021,
5032 0x0e000625, 0x8e101bc8, 0x00403021, 0x14c00005, 0x00000000, 0x8e62082c,
5033 0x24420001, 0x0a000596, 0xae62082c, 0x9608000a, 0x31020040, 0x10400005,
5034 0x2407180c, 0x8e02000c, 0x2407188c, 0x00021400, 0xacc20018, 0x3c020800,
5035 0x00511021, 0x8c421bd0, 0x3c030800, 0x00711821, 0x8c631bd4, 0x00021500,
5036 0x00031c00, 0x00431025, 0xacc20014, 0x8e63fff4, 0x96020008, 0x00432023,
5037 0x3242ffff, 0x3083ffff, 0x00431021, 0x02c2102a, 0x10400003, 0x00802821,
5038 0x97a9001e, 0x01322823, 0x8e620000, 0x30a4ffff, 0x00441021, 0xae620000,
5039 0xa4c5000e, 0x8e020000, 0xacc20000, 0x8e020004, 0x8e63fff4, 0x00431021,
5040 0xacc20004, 0x8e63fff4, 0x96020008, 0x00641821, 0x0062102a, 0x14400006,
5041 0x02459021, 0x8e62fff0, 0xae60fff4, 0x24420001, 0x0a000571, 0xae62fff0,
5042 0xae63fff4, 0xacc00008, 0x3242ffff, 0x10560003, 0x31020004, 0x10400006,
5043 0x24020305, 0x31020080, 0x54400001, 0x34e70010, 0x34e70020, 0x24020905,
5044 0xa4c2000c, 0x8ee30000, 0x8ee20004, 0x14620007, 0x3c02b49a, 0x8ee20860,
5045 0x54400001, 0x34e70400, 0x3c024b65, 0x0a000588, 0x34427654, 0x344289ab,
5046 0xacc2001c, 0x30e2ffff, 0xacc20010, 0x0e0005a2, 0x00c02021, 0x3242ffff,
5047 0x0056102b, 0x1440ff9b, 0x00000000, 0x8e620000, 0x8e63fffc, 0x0043102a,
5048 0x1440ff48, 0x00000000, 0x8fbf0044, 0x8fbe0040, 0x8fb7003c, 0x8fb60038,
5049 0x8fb50034, 0x8fb40030, 0x8fb3002c, 0x8fb20028, 0x8fb10024, 0x8fb00020,
5050 0x03e00008, 0x27bd0048, 0x27bdffe8, 0xafbf0014, 0xafb00010, 0x8f624450,
5051 0x8f634410, 0x0a0005b1, 0x00808021, 0x8f626820, 0x30422000, 0x10400003,
5052 0x00000000, 0x0e0001f0, 0x00002021, 0x8f624450, 0x8f634410, 0x3042ffff,
5053 0x0043102b, 0x1440fff5, 0x00000000, 0x8f630c14, 0x3063000f, 0x2c620002,
5054 0x1440000b, 0x00000000, 0x8f630c14, 0x3c020800, 0x8c421b40, 0x3063000f,
5055 0x24420001, 0x3c010800, 0xac221b40, 0x2c620002, 0x1040fff7, 0x00000000,
5056 0xaf705c18, 0x8f625c10, 0x30420002, 0x10400009, 0x00000000, 0x8f626820,
5057 0x30422000, 0x1040fff8, 0x00000000, 0x0e0001f0, 0x00002021, 0x0a0005c4,
5058 0x00000000, 0x8fbf0014, 0x8fb00010, 0x03e00008, 0x27bd0018, 0x00000000,
5059 0x00000000, 0x00000000, 0x27bdffe8, 0x3c1bc000, 0xafbf0014, 0xafb00010,
5060 0xaf60680c, 0x8f626804, 0x34420082, 0xaf626804, 0x8f634000, 0x24020b50,
5061 0x3c010800, 0xac221b54, 0x24020b78, 0x3c010800, 0xac221b64, 0x34630002,
5062 0xaf634000, 0x0e000605, 0x00808021, 0x3c010800, 0xa0221b68, 0x304200ff,
5063 0x24030002, 0x14430005, 0x00000000, 0x3c020800, 0x8c421b54, 0x0a0005f8,
5064 0xac5000c0, 0x3c020800, 0x8c421b54, 0xac5000bc, 0x8f624434, 0x8f634438,
5065 0x8f644410, 0x3c010800, 0xac221b5c, 0x3c010800, 0xac231b6c, 0x3c010800,
5066 0xac241b58, 0x8fbf0014, 0x8fb00010, 0x03e00008, 0x27bd0018, 0x3c040800,
5067 0x8c870000, 0x3c03aa55, 0x3463aa55, 0x3c06c003, 0xac830000, 0x8cc20000,
5068 0x14430007, 0x24050002, 0x3c0355aa, 0x346355aa, 0xac830000, 0x8cc20000,
5069 0x50430001, 0x24050001, 0x3c020800, 0xac470000, 0x03e00008, 0x00a01021,
5070 0x27bdfff8, 0x18800009, 0x00002821, 0x8f63680c, 0x8f62680c, 0x1043fffe,
5071 0x00000000, 0x24a50001, 0x00a4102a, 0x1440fff9, 0x00000000, 0x03e00008,
5072 0x27bd0008, 0x8f634450, 0x3c020800, 0x8c421b5c, 0x00031c02, 0x0043102b,
5073 0x14400008, 0x3c038000, 0x3c040800, 0x8c841b6c, 0x8f624450, 0x00021c02,
5074 0x0083102b, 0x1040fffc, 0x3c038000, 0xaf634444, 0x8f624444, 0x00431024,
5075 0x1440fffd, 0x00000000, 0x8f624448, 0x03e00008, 0x3042ffff, 0x3082ffff,
5076 0x2442e000, 0x2c422001, 0x14400003, 0x3c024000, 0x0a000648, 0x2402ffff,
5077 0x00822025, 0xaf645c38, 0x8f625c30, 0x30420002, 0x1440fffc, 0x00001021,
5078 0x03e00008, 0x00000000, 0x8f624450, 0x3c030800, 0x8c631b58, 0x0a000651,
5079 0x3042ffff, 0x8f624450, 0x3042ffff, 0x0043102b, 0x1440fffc, 0x00000000,
5080 0x03e00008, 0x00000000, 0x27bdffe0, 0x00802821, 0x3c040800, 0x24841af0,
5081 0x00003021, 0x00003821, 0xafbf0018, 0xafa00010, 0x0e00067c, 0xafa00014,
5082 0x0a000660, 0x00000000, 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x00000000,
5083 0x00000000, 0x00000000, 0x3c020800, 0x34423000, 0x3c030800, 0x34633000,
5084 0x3c040800, 0x348437ff, 0x3c010800, 0xac221b74, 0x24020040, 0x3c010800,
5085 0xac221b78, 0x3c010800, 0xac201b70, 0xac600000, 0x24630004, 0x0083102b,
5086 0x5040fffd, 0xac600000, 0x03e00008, 0x00000000, 0x00804821, 0x8faa0010,
5087 0x3c020800, 0x8c421b70, 0x3c040800, 0x8c841b78, 0x8fab0014, 0x24430001,
5088 0x0044102b, 0x3c010800, 0xac231b70, 0x14400003, 0x00004021, 0x3c010800,
5089 0xac201b70, 0x3c020800, 0x8c421b70, 0x3c030800, 0x8c631b74, 0x91240000,
5090 0x00021140, 0x00431021, 0x00481021, 0x25080001, 0xa0440000, 0x29020008,
5091 0x1440fff4, 0x25290001, 0x3c020800, 0x8c421b70, 0x3c030800, 0x8c631b74,
5092 0x8f64680c, 0x00021140, 0x00431021, 0xac440008, 0xac45000c, 0xac460010,
5093 0xac470014, 0xac4a0018, 0x03e00008, 0xac4b001c, 0x00000000, 0x00000000,
5094};
5095
5096static u32 tg3TsoFwRodata[] = {
5097 0x4d61696e, 0x43707542, 0x00000000, 0x4d61696e, 0x43707541, 0x00000000,
5098 0x00000000, 0x00000000, 0x73746b6f, 0x66666c64, 0x496e0000, 0x73746b6f,
5099 0x66662a2a, 0x00000000, 0x53774576, 0x656e7430, 0x00000000, 0x00000000,
5100 0x00000000, 0x00000000, 0x66617461, 0x6c457272, 0x00000000, 0x00000000,
5101 0x00000000,
5102};
5103
5104static u32 tg3TsoFwData[] = {
5105 0x00000000, 0x73746b6f, 0x66666c64, 0x5f76312e, 0x362e3000, 0x00000000,
5106 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x00000000,
5107 0x00000000,
5108};
5109
5110/* 5705 needs a special version of the TSO firmware. */
5111#define TG3_TSO5_FW_RELEASE_MAJOR 0x1
5112#define TG3_TSO5_FW_RELASE_MINOR 0x2
5113#define TG3_TSO5_FW_RELEASE_FIX 0x0
5114#define TG3_TSO5_FW_START_ADDR 0x00010000
5115#define TG3_TSO5_FW_TEXT_ADDR 0x00010000
5116#define TG3_TSO5_FW_TEXT_LEN 0xe90
5117#define TG3_TSO5_FW_RODATA_ADDR 0x00010e90
5118#define TG3_TSO5_FW_RODATA_LEN 0x50
5119#define TG3_TSO5_FW_DATA_ADDR 0x00010f00
5120#define TG3_TSO5_FW_DATA_LEN 0x20
5121#define TG3_TSO5_FW_SBSS_ADDR 0x00010f20
5122#define TG3_TSO5_FW_SBSS_LEN 0x28
5123#define TG3_TSO5_FW_BSS_ADDR 0x00010f50
5124#define TG3_TSO5_FW_BSS_LEN 0x88
5125
5126static u32 tg3Tso5FwText[(TG3_TSO5_FW_TEXT_LEN / 4) + 1] = {
5127 0x0c004003, 0x00000000, 0x00010f04, 0x00000000, 0x10000003, 0x00000000,
5128 0x0000000d, 0x0000000d, 0x3c1d0001, 0x37bde000, 0x03a0f021, 0x3c100001,
5129 0x26100000, 0x0c004010, 0x00000000, 0x0000000d, 0x27bdffe0, 0x3c04fefe,
5130 0xafbf0018, 0x0c0042e8, 0x34840002, 0x0c004364, 0x00000000, 0x3c030001,
5131 0x90630f34, 0x24020002, 0x3c040001, 0x24840e9c, 0x14620003, 0x24050001,
5132 0x3c040001, 0x24840e90, 0x24060002, 0x00003821, 0xafa00010, 0x0c004378,
5133 0xafa00014, 0x0c00402c, 0x00000000, 0x8fbf0018, 0x03e00008, 0x27bd0020,
5134 0x00000000, 0x00000000, 0x27bdffe0, 0xafbf001c, 0xafb20018, 0xafb10014,
5135 0x0c0042d4, 0xafb00010, 0x3c128000, 0x24110001, 0x8f706810, 0x32020400,
5136 0x10400007, 0x00000000, 0x8f641008, 0x00921024, 0x14400003, 0x00000000,
5137 0x0c004064, 0x00000000, 0x3c020001, 0x90420f56, 0x10510003, 0x32020200,
5138 0x1040fff1, 0x00000000, 0x0c0041b4, 0x00000000, 0x08004034, 0x00000000,
5139 0x8fbf001c, 0x8fb20018, 0x8fb10014, 0x8fb00010, 0x03e00008, 0x27bd0020,
5140 0x27bdffe0, 0x3c040001, 0x24840eb0, 0x00002821, 0x00003021, 0x00003821,
5141 0xafbf0018, 0xafa00010, 0x0c004378, 0xafa00014, 0x0000d021, 0x24020130,
5142 0xaf625000, 0x3c010001, 0xa4200f50, 0x3c010001, 0xa0200f57, 0x8fbf0018,
5143 0x03e00008, 0x27bd0020, 0x00000000, 0x00000000, 0x3c030001, 0x24630f60,
5144 0x90620000, 0x27bdfff0, 0x14400003, 0x0080c021, 0x08004073, 0x00004821,
5145 0x3c022000, 0x03021024, 0x10400003, 0x24090002, 0x08004073, 0xa0600000,
5146 0x24090001, 0x00181040, 0x30431f80, 0x346f8008, 0x1520004b, 0x25eb0028,
5147 0x3c040001, 0x00832021, 0x8c848010, 0x3c050001, 0x24a50f7a, 0x00041402,
5148 0xa0a20000, 0x3c010001, 0xa0240f7b, 0x3c020001, 0x00431021, 0x94428014,
5149 0x3c010001, 0xa0220f7c, 0x3c0c0001, 0x01836021, 0x8d8c8018, 0x304200ff,
5150 0x24420008, 0x000220c3, 0x24020001, 0x3c010001, 0xa0220f60, 0x0124102b,
5151 0x1040000c, 0x00003821, 0x24a6000e, 0x01602821, 0x8ca20000, 0x8ca30004,
5152 0x24a50008, 0x24e70001, 0xacc20000, 0xacc30004, 0x00e4102b, 0x1440fff8,
5153 0x24c60008, 0x00003821, 0x3c080001, 0x25080f7b, 0x91060000, 0x3c020001,
5154 0x90420f7c, 0x2503000d, 0x00c32821, 0x00461023, 0x00021fc2, 0x00431021,
5155 0x00021043, 0x1840000c, 0x00002021, 0x91020001, 0x00461023, 0x00021fc2,
5156 0x00431021, 0x00021843, 0x94a20000, 0x24e70001, 0x00822021, 0x00e3102a,
5157 0x1440fffb, 0x24a50002, 0x00041c02, 0x3082ffff, 0x00622021, 0x00041402,
5158 0x00822021, 0x3c02ffff, 0x01821024, 0x3083ffff, 0x00431025, 0x3c010001,
5159 0x080040fa, 0xac220f80, 0x3c050001, 0x24a50f7c, 0x90a20000, 0x3c0c0001,
5160 0x01836021, 0x8d8c8018, 0x000220c2, 0x1080000e, 0x00003821, 0x01603021,
5161 0x24a5000c, 0x8ca20000, 0x8ca30004, 0x24a50008, 0x24e70001, 0xacc20000,
5162 0xacc30004, 0x00e4102b, 0x1440fff8, 0x24c60008, 0x3c050001, 0x24a50f7c,
5163 0x90a20000, 0x30430007, 0x24020004, 0x10620011, 0x28620005, 0x10400005,
5164 0x24020002, 0x10620008, 0x000710c0, 0x080040fa, 0x00000000, 0x24020006,
5165 0x1062000e, 0x000710c0, 0x080040fa, 0x00000000, 0x00a21821, 0x9463000c,
5166 0x004b1021, 0x080040fa, 0xa4430000, 0x000710c0, 0x00a21821, 0x8c63000c,
5167 0x004b1021, 0x080040fa, 0xac430000, 0x00a21821, 0x8c63000c, 0x004b2021,
5168 0x00a21021, 0xac830000, 0x94420010, 0xa4820004, 0x95e70006, 0x3c020001,
5169 0x90420f7c, 0x3c030001, 0x90630f7a, 0x00e2c823, 0x3c020001, 0x90420f7b,
5170 0x24630028, 0x01e34021, 0x24420028, 0x15200012, 0x01e23021, 0x94c2000c,
5171 0x3c010001, 0xa4220f78, 0x94c20004, 0x94c30006, 0x3c010001, 0xa4200f76,
5172 0x3c010001, 0xa4200f72, 0x00021400, 0x00431025, 0x3c010001, 0xac220f6c,
5173 0x95020004, 0x3c010001, 0x08004124, 0xa4220f70, 0x3c020001, 0x94420f70,
5174 0x3c030001, 0x94630f72, 0x00431021, 0xa5020004, 0x3c020001, 0x94420f6c,
5175 0xa4c20004, 0x3c020001, 0x8c420f6c, 0xa4c20006, 0x3c040001, 0x94840f72,
5176 0x3c020001, 0x94420f70, 0x3c0a0001, 0x954a0f76, 0x00441821, 0x3063ffff,
5177 0x0062182a, 0x24020002, 0x1122000b, 0x00832023, 0x3c030001, 0x94630f78,
5178 0x30620009, 0x10400006, 0x3062fff6, 0xa4c2000c, 0x3c020001, 0x94420f78,
5179 0x30420009, 0x01425023, 0x24020001, 0x1122001b, 0x29220002, 0x50400005,
5180 0x24020002, 0x11200007, 0x31a2ffff, 0x08004197, 0x00000000, 0x1122001d,
5181 0x24020016, 0x08004197, 0x31a2ffff, 0x3c0e0001, 0x95ce0f80, 0x10800005,
5182 0x01806821, 0x01c42021, 0x00041c02, 0x3082ffff, 0x00627021, 0x000e1027,
5183 0xa502000a, 0x3c030001, 0x90630f7b, 0x31a2ffff, 0x00e21021, 0x0800418d,
5184 0x00432023, 0x3c020001, 0x94420f80, 0x00442021, 0x00041c02, 0x3082ffff,
5185 0x00622021, 0x00807021, 0x00041027, 0x08004185, 0xa502000a, 0x3c050001,
5186 0x24a50f7a, 0x90a30000, 0x14620002, 0x24e2fff2, 0xa5e20034, 0x90a20000,
5187 0x00e21023, 0xa5020002, 0x3c030001, 0x94630f80, 0x3c020001, 0x94420f5a,
5188 0x30e5ffff, 0x00641821, 0x00451023, 0x00622023, 0x00041c02, 0x3082ffff,
5189 0x00622021, 0x00041027, 0xa502000a, 0x3c030001, 0x90630f7c, 0x24620001,
5190 0x14a20005, 0x00807021, 0x01631021, 0x90420000, 0x08004185, 0x00026200,
5191 0x24620002, 0x14a20003, 0x306200fe, 0x004b1021, 0x944c0000, 0x3c020001,
5192 0x94420f82, 0x3183ffff, 0x3c040001, 0x90840f7b, 0x00431021, 0x00e21021,
5193 0x00442023, 0x008a2021, 0x00041c02, 0x3082ffff, 0x00622021, 0x00041402,
5194 0x00822021, 0x00806821, 0x00041027, 0xa4c20010, 0x31a2ffff, 0x000e1c00,
5195 0x00431025, 0x3c040001, 0x24840f72, 0xade20010, 0x94820000, 0x3c050001,
5196 0x94a50f76, 0x3c030001, 0x8c630f6c, 0x24420001, 0x00b92821, 0xa4820000,
5197 0x3322ffff, 0x00622021, 0x0083182b, 0x3c010001, 0xa4250f76, 0x10600003,
5198 0x24a2ffff, 0x3c010001, 0xa4220f76, 0x3c024000, 0x03021025, 0x3c010001,
5199 0xac240f6c, 0xaf621008, 0x03e00008, 0x27bd0010, 0x3c030001, 0x90630f56,
5200 0x27bdffe8, 0x24020001, 0xafbf0014, 0x10620026, 0xafb00010, 0x8f620cf4,
5201 0x2442ffff, 0x3042007f, 0x00021100, 0x8c434000, 0x3c010001, 0xac230f64,
5202 0x8c434008, 0x24444000, 0x8c5c4004, 0x30620040, 0x14400002, 0x24020088,
5203 0x24020008, 0x3c010001, 0xa4220f68, 0x30620004, 0x10400005, 0x24020001,
5204 0x3c010001, 0xa0220f57, 0x080041d5, 0x00031402, 0x3c010001, 0xa0200f57,
5205 0x00031402, 0x3c010001, 0xa4220f54, 0x9483000c, 0x24020001, 0x3c010001,
5206 0xa4200f50, 0x3c010001, 0xa0220f56, 0x3c010001, 0xa4230f62, 0x24020001,
5207 0x1342001e, 0x00000000, 0x13400005, 0x24020003, 0x13420067, 0x00000000,
5208 0x080042cf, 0x00000000, 0x3c020001, 0x94420f62, 0x241a0001, 0x3c010001,
5209 0xa4200f5e, 0x3c010001, 0xa4200f52, 0x304407ff, 0x00021bc2, 0x00031823,
5210 0x3063003e, 0x34630036, 0x00021242, 0x3042003c, 0x00621821, 0x3c010001,
5211 0xa4240f58, 0x00832021, 0x24630030, 0x3c010001, 0xa4240f5a, 0x3c010001,
5212 0xa4230f5c, 0x3c060001, 0x24c60f52, 0x94c50000, 0x94c30002, 0x3c040001,
5213 0x94840f5a, 0x00651021, 0x0044102a, 0x10400013, 0x3c108000, 0x00a31021,
5214 0xa4c20000, 0x3c02a000, 0xaf620cf4, 0x3c010001, 0xa0200f56, 0x8f641008,
5215 0x00901024, 0x14400003, 0x00000000, 0x0c004064, 0x00000000, 0x8f620cf4,
5216 0x00501024, 0x104000b7, 0x00000000, 0x0800420f, 0x00000000, 0x3c030001,
5217 0x94630f50, 0x00851023, 0xa4c40000, 0x00621821, 0x3042ffff, 0x3c010001,
5218 0xa4230f50, 0xaf620ce8, 0x3c020001, 0x94420f68, 0x34420024, 0xaf620cec,
5219 0x94c30002, 0x3c020001, 0x94420f50, 0x14620012, 0x3c028000, 0x3c108000,
5220 0x3c02a000, 0xaf620cf4, 0x3c010001, 0xa0200f56, 0x8f641008, 0x00901024,
5221 0x14400003, 0x00000000, 0x0c004064, 0x00000000, 0x8f620cf4, 0x00501024,
5222 0x1440fff7, 0x00000000, 0x080042cf, 0x241a0003, 0xaf620cf4, 0x3c108000,
5223 0x8f641008, 0x00901024, 0x14400003, 0x00000000, 0x0c004064, 0x00000000,
5224 0x8f620cf4, 0x00501024, 0x1440fff7, 0x00000000, 0x080042cf, 0x241a0003,
5225 0x3c070001, 0x24e70f50, 0x94e20000, 0x03821021, 0xaf620ce0, 0x3c020001,
5226 0x8c420f64, 0xaf620ce4, 0x3c050001, 0x94a50f54, 0x94e30000, 0x3c040001,
5227 0x94840f58, 0x3c020001, 0x94420f5e, 0x00a32823, 0x00822023, 0x30a6ffff,
5228 0x3083ffff, 0x00c3102b, 0x14400043, 0x00000000, 0x3c020001, 0x94420f5c,
5229 0x00021400, 0x00621025, 0xaf620ce8, 0x94e20000, 0x3c030001, 0x94630f54,
5230 0x00441021, 0xa4e20000, 0x3042ffff, 0x14430021, 0x3c020008, 0x3c020001,
5231 0x90420f57, 0x10400006, 0x3c03000c, 0x3c020001, 0x94420f68, 0x34630624,
5232 0x0800427c, 0x0000d021, 0x3c020001, 0x94420f68, 0x3c030008, 0x34630624,
5233 0x00431025, 0xaf620cec, 0x3c108000, 0x3c02a000, 0xaf620cf4, 0x3c010001,
5234 0xa0200f56, 0x8f641008, 0x00901024, 0x14400003, 0x00000000, 0x0c004064,
5235 0x00000000, 0x8f620cf4, 0x00501024, 0x10400015, 0x00000000, 0x08004283,
5236 0x00000000, 0x3c030001, 0x94630f68, 0x34420624, 0x3c108000, 0x00621825,
5237 0x3c028000, 0xaf630cec, 0xaf620cf4, 0x8f641008, 0x00901024, 0x14400003,
5238 0x00000000, 0x0c004064, 0x00000000, 0x8f620cf4, 0x00501024, 0x1440fff7,
5239 0x00000000, 0x3c010001, 0x080042cf, 0xa4200f5e, 0x3c020001, 0x94420f5c,
5240 0x00021400, 0x00c21025, 0xaf620ce8, 0x3c020001, 0x90420f57, 0x10400009,
5241 0x3c03000c, 0x3c020001, 0x94420f68, 0x34630624, 0x0000d021, 0x00431025,
5242 0xaf620cec, 0x080042c1, 0x3c108000, 0x3c020001, 0x94420f68, 0x3c030008,
5243 0x34630604, 0x00431025, 0xaf620cec, 0x3c020001, 0x94420f5e, 0x00451021,
5244 0x3c010001, 0xa4220f5e, 0x3c108000, 0x3c02a000, 0xaf620cf4, 0x3c010001,
5245 0xa0200f56, 0x8f641008, 0x00901024, 0x14400003, 0x00000000, 0x0c004064,
5246 0x00000000, 0x8f620cf4, 0x00501024, 0x1440fff7, 0x00000000, 0x8fbf0014,
5247 0x8fb00010, 0x03e00008, 0x27bd0018, 0x00000000, 0x27bdffe0, 0x3c040001,
5248 0x24840ec0, 0x00002821, 0x00003021, 0x00003821, 0xafbf0018, 0xafa00010,
5249 0x0c004378, 0xafa00014, 0x0000d021, 0x24020130, 0xaf625000, 0x3c010001,
5250 0xa4200f50, 0x3c010001, 0xa0200f57, 0x8fbf0018, 0x03e00008, 0x27bd0020,
5251 0x27bdffe8, 0x3c1bc000, 0xafbf0014, 0xafb00010, 0xaf60680c, 0x8f626804,
5252 0x34420082, 0xaf626804, 0x8f634000, 0x24020b50, 0x3c010001, 0xac220f20,
5253 0x24020b78, 0x3c010001, 0xac220f30, 0x34630002, 0xaf634000, 0x0c004315,
5254 0x00808021, 0x3c010001, 0xa0220f34, 0x304200ff, 0x24030002, 0x14430005,
5255 0x00000000, 0x3c020001, 0x8c420f20, 0x08004308, 0xac5000c0, 0x3c020001,
5256 0x8c420f20, 0xac5000bc, 0x8f624434, 0x8f634438, 0x8f644410, 0x3c010001,
5257 0xac220f28, 0x3c010001, 0xac230f38, 0x3c010001, 0xac240f24, 0x8fbf0014,
5258 0x8fb00010, 0x03e00008, 0x27bd0018, 0x03e00008, 0x24020001, 0x27bdfff8,
5259 0x18800009, 0x00002821, 0x8f63680c, 0x8f62680c, 0x1043fffe, 0x00000000,
5260 0x24a50001, 0x00a4102a, 0x1440fff9, 0x00000000, 0x03e00008, 0x27bd0008,
5261 0x8f634450, 0x3c020001, 0x8c420f28, 0x00031c02, 0x0043102b, 0x14400008,
5262 0x3c038000, 0x3c040001, 0x8c840f38, 0x8f624450, 0x00021c02, 0x0083102b,
5263 0x1040fffc, 0x3c038000, 0xaf634444, 0x8f624444, 0x00431024, 0x1440fffd,
5264 0x00000000, 0x8f624448, 0x03e00008, 0x3042ffff, 0x3082ffff, 0x2442e000,
5265 0x2c422001, 0x14400003, 0x3c024000, 0x08004347, 0x2402ffff, 0x00822025,
5266 0xaf645c38, 0x8f625c30, 0x30420002, 0x1440fffc, 0x00001021, 0x03e00008,
5267 0x00000000, 0x8f624450, 0x3c030001, 0x8c630f24, 0x08004350, 0x3042ffff,
5268 0x8f624450, 0x3042ffff, 0x0043102b, 0x1440fffc, 0x00000000, 0x03e00008,
5269 0x00000000, 0x27bdffe0, 0x00802821, 0x3c040001, 0x24840ed0, 0x00003021,
5270 0x00003821, 0xafbf0018, 0xafa00010, 0x0c004378, 0xafa00014, 0x0800435f,
5271 0x00000000, 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x3c020001, 0x3442d600,
5272 0x3c030001, 0x3463d600, 0x3c040001, 0x3484ddff, 0x3c010001, 0xac220f40,
5273 0x24020040, 0x3c010001, 0xac220f44, 0x3c010001, 0xac200f3c, 0xac600000,
5274 0x24630004, 0x0083102b, 0x5040fffd, 0xac600000, 0x03e00008, 0x00000000,
5275 0x00804821, 0x8faa0010, 0x3c020001, 0x8c420f3c, 0x3c040001, 0x8c840f44,
5276 0x8fab0014, 0x24430001, 0x0044102b, 0x3c010001, 0xac230f3c, 0x14400003,
5277 0x00004021, 0x3c010001, 0xac200f3c, 0x3c020001, 0x8c420f3c, 0x3c030001,
5278 0x8c630f40, 0x91240000, 0x00021140, 0x00431021, 0x00481021, 0x25080001,
5279 0xa0440000, 0x29020008, 0x1440fff4, 0x25290001, 0x3c020001, 0x8c420f3c,
5280 0x3c030001, 0x8c630f40, 0x8f64680c, 0x00021140, 0x00431021, 0xac440008,
5281 0xac45000c, 0xac460010, 0xac470014, 0xac4a0018, 0x03e00008, 0xac4b001c,
5282 0x00000000, 0x00000000, 0x00000000,
5283};
5284
5285static u32 tg3Tso5FwRodata[(TG3_TSO5_FW_RODATA_LEN / 4) + 1] = {
5286 0x4d61696e, 0x43707542, 0x00000000, 0x4d61696e, 0x43707541, 0x00000000,
5287 0x00000000, 0x00000000, 0x73746b6f, 0x66666c64, 0x00000000, 0x00000000,
5288 0x73746b6f, 0x66666c64, 0x00000000, 0x00000000, 0x66617461, 0x6c457272,
5289 0x00000000, 0x00000000, 0x00000000,
5290};
5291
5292static u32 tg3Tso5FwData[(TG3_TSO5_FW_DATA_LEN / 4) + 1] = {
5293 0x00000000, 0x73746b6f, 0x66666c64, 0x5f76312e, 0x322e3000, 0x00000000,
5294 0x00000000, 0x00000000, 0x00000000,
5295};
5296
5297/* tp->lock is held. */
5298static int tg3_load_tso_firmware(struct tg3 *tp)
5299{
5300 struct fw_info info;
5301 unsigned long cpu_base, cpu_scratch_base, cpu_scratch_size;
5302 int err, i;
5303
5304 if (tp->tg3_flags2 & TG3_FLG2_HW_TSO)
5305 return 0;
5306
5307 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705) {
5308 info.text_base = TG3_TSO5_FW_TEXT_ADDR;
5309 info.text_len = TG3_TSO5_FW_TEXT_LEN;
5310 info.text_data = &tg3Tso5FwText[0];
5311 info.rodata_base = TG3_TSO5_FW_RODATA_ADDR;
5312 info.rodata_len = TG3_TSO5_FW_RODATA_LEN;
5313 info.rodata_data = &tg3Tso5FwRodata[0];
5314 info.data_base = TG3_TSO5_FW_DATA_ADDR;
5315 info.data_len = TG3_TSO5_FW_DATA_LEN;
5316 info.data_data = &tg3Tso5FwData[0];
5317 cpu_base = RX_CPU_BASE;
5318 cpu_scratch_base = NIC_SRAM_MBUF_POOL_BASE5705;
5319 cpu_scratch_size = (info.text_len +
5320 info.rodata_len +
5321 info.data_len +
5322 TG3_TSO5_FW_SBSS_LEN +
5323 TG3_TSO5_FW_BSS_LEN);
5324 } else {
5325 info.text_base = TG3_TSO_FW_TEXT_ADDR;
5326 info.text_len = TG3_TSO_FW_TEXT_LEN;
5327 info.text_data = &tg3TsoFwText[0];
5328 info.rodata_base = TG3_TSO_FW_RODATA_ADDR;
5329 info.rodata_len = TG3_TSO_FW_RODATA_LEN;
5330 info.rodata_data = &tg3TsoFwRodata[0];
5331 info.data_base = TG3_TSO_FW_DATA_ADDR;
5332 info.data_len = TG3_TSO_FW_DATA_LEN;
5333 info.data_data = &tg3TsoFwData[0];
5334 cpu_base = TX_CPU_BASE;
5335 cpu_scratch_base = TX_CPU_SCRATCH_BASE;
5336 cpu_scratch_size = TX_CPU_SCRATCH_SIZE;
5337 }
5338
5339 err = tg3_load_firmware_cpu(tp, cpu_base,
5340 cpu_scratch_base, cpu_scratch_size,
5341 &info);
5342 if (err)
5343 return err;
5344
5345 /* Now startup the cpu. */
5346 tw32(cpu_base + CPU_STATE, 0xffffffff);
5347 tw32_f(cpu_base + CPU_PC, info.text_base);
5348
5349 for (i = 0; i < 5; i++) {
5350 if (tr32(cpu_base + CPU_PC) == info.text_base)
5351 break;
5352 tw32(cpu_base + CPU_STATE, 0xffffffff);
5353 tw32(cpu_base + CPU_MODE, CPU_MODE_HALT);
5354 tw32_f(cpu_base + CPU_PC, info.text_base);
5355 udelay(1000);
5356 }
5357 if (i >= 5) {
5358 printk(KERN_ERR PFX "tg3_load_tso_firmware fails for %s "
5359 "to set CPU PC, is %08x should be %08x\n",
5360 tp->dev->name, tr32(cpu_base + CPU_PC),
5361 info.text_base);
5362 return -ENODEV;
5363 }
5364 tw32(cpu_base + CPU_STATE, 0xffffffff);
5365 tw32_f(cpu_base + CPU_MODE, 0x00000000);
5366 return 0;
5367}
5368
5369#endif /* TG3_TSO_SUPPORT != 0 */
5370
5371/* tp->lock is held. */
5372static void __tg3_set_mac_addr(struct tg3 *tp)
5373{
5374 u32 addr_high, addr_low;
5375 int i;
5376
5377 addr_high = ((tp->dev->dev_addr[0] << 8) |
5378 tp->dev->dev_addr[1]);
5379 addr_low = ((tp->dev->dev_addr[2] << 24) |
5380 (tp->dev->dev_addr[3] << 16) |
5381 (tp->dev->dev_addr[4] << 8) |
5382 (tp->dev->dev_addr[5] << 0));
5383 for (i = 0; i < 4; i++) {
5384 tw32(MAC_ADDR_0_HIGH + (i * 8), addr_high);
5385 tw32(MAC_ADDR_0_LOW + (i * 8), addr_low);
5386 }
5387
5388 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 ||
5389 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704) {
5390 for (i = 0; i < 12; i++) {
5391 tw32(MAC_EXTADDR_0_HIGH + (i * 8), addr_high);
5392 tw32(MAC_EXTADDR_0_LOW + (i * 8), addr_low);
5393 }
5394 }
5395
5396 addr_high = (tp->dev->dev_addr[0] +
5397 tp->dev->dev_addr[1] +
5398 tp->dev->dev_addr[2] +
5399 tp->dev->dev_addr[3] +
5400 tp->dev->dev_addr[4] +
5401 tp->dev->dev_addr[5]) &
5402 TX_BACKOFF_SEED_MASK;
5403 tw32(MAC_TX_BACKOFF_SEED, addr_high);
5404}
5405
5406static int tg3_set_mac_addr(struct net_device *dev, void *p)
5407{
5408 struct tg3 *tp = netdev_priv(dev);
5409 struct sockaddr *addr = p;
5410
f9804ddb
MC
5411 if (!is_valid_ether_addr(addr->sa_data))
5412 return -EINVAL;
5413
1da177e4
LT
5414 memcpy(dev->dev_addr, addr->sa_data, dev->addr_len);
5415
f47c11ee 5416 spin_lock_bh(&tp->lock);
1da177e4 5417 __tg3_set_mac_addr(tp);
f47c11ee 5418 spin_unlock_bh(&tp->lock);
1da177e4
LT
5419
5420 return 0;
5421}
5422
5423/* tp->lock is held. */
5424static void tg3_set_bdinfo(struct tg3 *tp, u32 bdinfo_addr,
5425 dma_addr_t mapping, u32 maxlen_flags,
5426 u32 nic_addr)
5427{
5428 tg3_write_mem(tp,
5429 (bdinfo_addr + TG3_BDINFO_HOST_ADDR + TG3_64BIT_REG_HIGH),
5430 ((u64) mapping >> 32));
5431 tg3_write_mem(tp,
5432 (bdinfo_addr + TG3_BDINFO_HOST_ADDR + TG3_64BIT_REG_LOW),
5433 ((u64) mapping & 0xffffffff));
5434 tg3_write_mem(tp,
5435 (bdinfo_addr + TG3_BDINFO_MAXLEN_FLAGS),
5436 maxlen_flags);
5437
5438 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS))
5439 tg3_write_mem(tp,
5440 (bdinfo_addr + TG3_BDINFO_NIC_ADDR),
5441 nic_addr);
5442}
5443
5444static void __tg3_set_rx_mode(struct net_device *);
d244c892 5445static void __tg3_set_coalesce(struct tg3 *tp, struct ethtool_coalesce *ec)
15f9850d
DM
5446{
5447 tw32(HOSTCC_RXCOL_TICKS, ec->rx_coalesce_usecs);
5448 tw32(HOSTCC_TXCOL_TICKS, ec->tx_coalesce_usecs);
5449 tw32(HOSTCC_RXMAX_FRAMES, ec->rx_max_coalesced_frames);
5450 tw32(HOSTCC_TXMAX_FRAMES, ec->tx_max_coalesced_frames);
5451 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
5452 tw32(HOSTCC_RXCOAL_TICK_INT, ec->rx_coalesce_usecs_irq);
5453 tw32(HOSTCC_TXCOAL_TICK_INT, ec->tx_coalesce_usecs_irq);
5454 }
5455 tw32(HOSTCC_RXCOAL_MAXF_INT, ec->rx_max_coalesced_frames_irq);
5456 tw32(HOSTCC_TXCOAL_MAXF_INT, ec->tx_max_coalesced_frames_irq);
5457 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
5458 u32 val = ec->stats_block_coalesce_usecs;
5459
5460 if (!netif_carrier_ok(tp->dev))
5461 val = 0;
5462
5463 tw32(HOSTCC_STAT_COAL_TICKS, val);
5464 }
5465}
1da177e4
LT
5466
5467/* tp->lock is held. */
5468static int tg3_reset_hw(struct tg3 *tp)
5469{
5470 u32 val, rdmac_mode;
5471 int i, err, limit;
5472
5473 tg3_disable_ints(tp);
5474
5475 tg3_stop_fw(tp);
5476
5477 tg3_write_sig_pre_reset(tp, RESET_KIND_INIT);
5478
5479 if (tp->tg3_flags & TG3_FLAG_INIT_COMPLETE) {
e6de8ad1 5480 tg3_abort_hw(tp, 1);
1da177e4
LT
5481 }
5482
5483 err = tg3_chip_reset(tp);
5484 if (err)
5485 return err;
5486
5487 tg3_write_sig_legacy(tp, RESET_KIND_INIT);
5488
5489 /* This works around an issue with Athlon chipsets on
5490 * B3 tigon3 silicon. This bit has no effect on any
5491 * other revision. But do not set this on PCI Express
5492 * chips.
5493 */
5494 if (!(tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS))
5495 tp->pci_clock_ctrl |= CLOCK_CTRL_DELAY_PCI_GRANT;
5496 tw32_f(TG3PCI_CLOCK_CTRL, tp->pci_clock_ctrl);
5497
5498 if (tp->pci_chip_rev_id == CHIPREV_ID_5704_A0 &&
5499 (tp->tg3_flags & TG3_FLAG_PCIX_MODE)) {
5500 val = tr32(TG3PCI_PCISTATE);
5501 val |= PCISTATE_RETRY_SAME_DMA;
5502 tw32(TG3PCI_PCISTATE, val);
5503 }
5504
5505 if (GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5704_BX) {
5506 /* Enable some hw fixes. */
5507 val = tr32(TG3PCI_MSI_DATA);
5508 val |= (1 << 26) | (1 << 28) | (1 << 29);
5509 tw32(TG3PCI_MSI_DATA, val);
5510 }
5511
5512 /* Descriptor ring init may make accesses to the
5513 * NIC SRAM area to setup the TX descriptors, so we
5514 * can only do this after the hardware has been
5515 * successfully reset.
5516 */
5517 tg3_init_rings(tp);
5518
5519 /* This value is determined during the probe time DMA
5520 * engine test, tg3_test_dma.
5521 */
5522 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl);
5523
5524 tp->grc_mode &= ~(GRC_MODE_HOST_SENDBDS |
5525 GRC_MODE_4X_NIC_SEND_RINGS |
5526 GRC_MODE_NO_TX_PHDR_CSUM |
5527 GRC_MODE_NO_RX_PHDR_CSUM);
5528 tp->grc_mode |= GRC_MODE_HOST_SENDBDS;
5529 if (tp->tg3_flags & TG3_FLAG_NO_TX_PSEUDO_CSUM)
5530 tp->grc_mode |= GRC_MODE_NO_TX_PHDR_CSUM;
5531 if (tp->tg3_flags & TG3_FLAG_NO_RX_PSEUDO_CSUM)
5532 tp->grc_mode |= GRC_MODE_NO_RX_PHDR_CSUM;
5533
5534 tw32(GRC_MODE,
5535 tp->grc_mode |
5536 (GRC_MODE_IRQ_ON_MAC_ATTN | GRC_MODE_HOST_STACKUP));
5537
5538 /* Setup the timer prescalar register. Clock is always 66Mhz. */
5539 val = tr32(GRC_MISC_CFG);
5540 val &= ~0xff;
5541 val |= (65 << GRC_MISC_CFG_PRESCALAR_SHIFT);
5542 tw32(GRC_MISC_CFG, val);
5543
5544 /* Initialize MBUF/DESC pool. */
cbf46853 5545 if (tp->tg3_flags2 & TG3_FLG2_5750_PLUS) {
1da177e4
LT
5546 /* Do nothing. */
5547 } else if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5705) {
5548 tw32(BUFMGR_MB_POOL_ADDR, NIC_SRAM_MBUF_POOL_BASE);
5549 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704)
5550 tw32(BUFMGR_MB_POOL_SIZE, NIC_SRAM_MBUF_POOL_SIZE64);
5551 else
5552 tw32(BUFMGR_MB_POOL_SIZE, NIC_SRAM_MBUF_POOL_SIZE96);
5553 tw32(BUFMGR_DMA_DESC_POOL_ADDR, NIC_SRAM_DMA_DESC_POOL_BASE);
5554 tw32(BUFMGR_DMA_DESC_POOL_SIZE, NIC_SRAM_DMA_DESC_POOL_SIZE);
5555 }
5556#if TG3_TSO_SUPPORT != 0
5557 else if (tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE) {
5558 int fw_len;
5559
5560 fw_len = (TG3_TSO5_FW_TEXT_LEN +
5561 TG3_TSO5_FW_RODATA_LEN +
5562 TG3_TSO5_FW_DATA_LEN +
5563 TG3_TSO5_FW_SBSS_LEN +
5564 TG3_TSO5_FW_BSS_LEN);
5565 fw_len = (fw_len + (0x80 - 1)) & ~(0x80 - 1);
5566 tw32(BUFMGR_MB_POOL_ADDR,
5567 NIC_SRAM_MBUF_POOL_BASE5705 + fw_len);
5568 tw32(BUFMGR_MB_POOL_SIZE,
5569 NIC_SRAM_MBUF_POOL_SIZE5705 - fw_len - 0xa00);
5570 }
5571#endif
5572
0f893dc6 5573 if (tp->dev->mtu <= ETH_DATA_LEN) {
1da177e4
LT
5574 tw32(BUFMGR_MB_RDMA_LOW_WATER,
5575 tp->bufmgr_config.mbuf_read_dma_low_water);
5576 tw32(BUFMGR_MB_MACRX_LOW_WATER,
5577 tp->bufmgr_config.mbuf_mac_rx_low_water);
5578 tw32(BUFMGR_MB_HIGH_WATER,
5579 tp->bufmgr_config.mbuf_high_water);
5580 } else {
5581 tw32(BUFMGR_MB_RDMA_LOW_WATER,
5582 tp->bufmgr_config.mbuf_read_dma_low_water_jumbo);
5583 tw32(BUFMGR_MB_MACRX_LOW_WATER,
5584 tp->bufmgr_config.mbuf_mac_rx_low_water_jumbo);
5585 tw32(BUFMGR_MB_HIGH_WATER,
5586 tp->bufmgr_config.mbuf_high_water_jumbo);
5587 }
5588 tw32(BUFMGR_DMA_LOW_WATER,
5589 tp->bufmgr_config.dma_low_water);
5590 tw32(BUFMGR_DMA_HIGH_WATER,
5591 tp->bufmgr_config.dma_high_water);
5592
5593 tw32(BUFMGR_MODE, BUFMGR_MODE_ENABLE | BUFMGR_MODE_ATTN_ENABLE);
5594 for (i = 0; i < 2000; i++) {
5595 if (tr32(BUFMGR_MODE) & BUFMGR_MODE_ENABLE)
5596 break;
5597 udelay(10);
5598 }
5599 if (i >= 2000) {
5600 printk(KERN_ERR PFX "tg3_reset_hw cannot enable BUFMGR for %s.\n",
5601 tp->dev->name);
5602 return -ENODEV;
5603 }
5604
5605 /* Setup replenish threshold. */
5606 tw32(RCVBDI_STD_THRESH, tp->rx_pending / 8);
5607
5608 /* Initialize TG3_BDINFO's at:
5609 * RCVDBDI_STD_BD: standard eth size rx ring
5610 * RCVDBDI_JUMBO_BD: jumbo frame rx ring
5611 * RCVDBDI_MINI_BD: small frame rx ring (??? does not work)
5612 *
5613 * like so:
5614 * TG3_BDINFO_HOST_ADDR: high/low parts of DMA address of ring
5615 * TG3_BDINFO_MAXLEN_FLAGS: (rx max buffer size << 16) |
5616 * ring attribute flags
5617 * TG3_BDINFO_NIC_ADDR: location of descriptors in nic SRAM
5618 *
5619 * Standard receive ring @ NIC_SRAM_RX_BUFFER_DESC, 512 entries.
5620 * Jumbo receive ring @ NIC_SRAM_RX_JUMBO_BUFFER_DESC, 256 entries.
5621 *
5622 * The size of each ring is fixed in the firmware, but the location is
5623 * configurable.
5624 */
5625 tw32(RCVDBDI_STD_BD + TG3_BDINFO_HOST_ADDR + TG3_64BIT_REG_HIGH,
5626 ((u64) tp->rx_std_mapping >> 32));
5627 tw32(RCVDBDI_STD_BD + TG3_BDINFO_HOST_ADDR + TG3_64BIT_REG_LOW,
5628 ((u64) tp->rx_std_mapping & 0xffffffff));
5629 tw32(RCVDBDI_STD_BD + TG3_BDINFO_NIC_ADDR,
5630 NIC_SRAM_RX_BUFFER_DESC);
5631
5632 /* Don't even try to program the JUMBO/MINI buffer descriptor
5633 * configs on 5705.
5634 */
5635 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS) {
5636 tw32(RCVDBDI_STD_BD + TG3_BDINFO_MAXLEN_FLAGS,
5637 RX_STD_MAX_SIZE_5705 << BDINFO_FLAGS_MAXLEN_SHIFT);
5638 } else {
5639 tw32(RCVDBDI_STD_BD + TG3_BDINFO_MAXLEN_FLAGS,
5640 RX_STD_MAX_SIZE << BDINFO_FLAGS_MAXLEN_SHIFT);
5641
5642 tw32(RCVDBDI_MINI_BD + TG3_BDINFO_MAXLEN_FLAGS,
5643 BDINFO_FLAGS_DISABLED);
5644
5645 /* Setup replenish threshold. */
5646 tw32(RCVBDI_JUMBO_THRESH, tp->rx_jumbo_pending / 8);
5647
0f893dc6 5648 if (tp->tg3_flags & TG3_FLAG_JUMBO_RING_ENABLE) {
1da177e4
LT
5649 tw32(RCVDBDI_JUMBO_BD + TG3_BDINFO_HOST_ADDR + TG3_64BIT_REG_HIGH,
5650 ((u64) tp->rx_jumbo_mapping >> 32));
5651 tw32(RCVDBDI_JUMBO_BD + TG3_BDINFO_HOST_ADDR + TG3_64BIT_REG_LOW,
5652 ((u64) tp->rx_jumbo_mapping & 0xffffffff));
5653 tw32(RCVDBDI_JUMBO_BD + TG3_BDINFO_MAXLEN_FLAGS,
5654 RX_JUMBO_MAX_SIZE << BDINFO_FLAGS_MAXLEN_SHIFT);
5655 tw32(RCVDBDI_JUMBO_BD + TG3_BDINFO_NIC_ADDR,
5656 NIC_SRAM_RX_JUMBO_BUFFER_DESC);
5657 } else {
5658 tw32(RCVDBDI_JUMBO_BD + TG3_BDINFO_MAXLEN_FLAGS,
5659 BDINFO_FLAGS_DISABLED);
5660 }
5661
5662 }
5663
5664 /* There is only one send ring on 5705/5750, no need to explicitly
5665 * disable the others.
5666 */
5667 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
5668 /* Clear out send RCB ring in SRAM. */
5669 for (i = NIC_SRAM_SEND_RCB; i < NIC_SRAM_RCV_RET_RCB; i += TG3_BDINFO_SIZE)
5670 tg3_write_mem(tp, i + TG3_BDINFO_MAXLEN_FLAGS,
5671 BDINFO_FLAGS_DISABLED);
5672 }
5673
5674 tp->tx_prod = 0;
5675 tp->tx_cons = 0;
5676 tw32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0 + TG3_64BIT_REG_LOW, 0);
5677 tw32_tx_mbox(MAILBOX_SNDNIC_PROD_IDX_0 + TG3_64BIT_REG_LOW, 0);
5678
5679 tg3_set_bdinfo(tp, NIC_SRAM_SEND_RCB,
5680 tp->tx_desc_mapping,
5681 (TG3_TX_RING_SIZE <<
5682 BDINFO_FLAGS_MAXLEN_SHIFT),
5683 NIC_SRAM_TX_BUFFER_DESC);
5684
5685 /* There is only one receive return ring on 5705/5750, no need
5686 * to explicitly disable the others.
5687 */
5688 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
5689 for (i = NIC_SRAM_RCV_RET_RCB; i < NIC_SRAM_STATS_BLK;
5690 i += TG3_BDINFO_SIZE) {
5691 tg3_write_mem(tp, i + TG3_BDINFO_MAXLEN_FLAGS,
5692 BDINFO_FLAGS_DISABLED);
5693 }
5694 }
5695
5696 tp->rx_rcb_ptr = 0;
5697 tw32_rx_mbox(MAILBOX_RCVRET_CON_IDX_0 + TG3_64BIT_REG_LOW, 0);
5698
5699 tg3_set_bdinfo(tp, NIC_SRAM_RCV_RET_RCB,
5700 tp->rx_rcb_mapping,
5701 (TG3_RX_RCB_RING_SIZE(tp) <<
5702 BDINFO_FLAGS_MAXLEN_SHIFT),
5703 0);
5704
5705 tp->rx_std_ptr = tp->rx_pending;
5706 tw32_rx_mbox(MAILBOX_RCV_STD_PROD_IDX + TG3_64BIT_REG_LOW,
5707 tp->rx_std_ptr);
5708
0f893dc6 5709 tp->rx_jumbo_ptr = (tp->tg3_flags & TG3_FLAG_JUMBO_RING_ENABLE) ?
1da177e4
LT
5710 tp->rx_jumbo_pending : 0;
5711 tw32_rx_mbox(MAILBOX_RCV_JUMBO_PROD_IDX + TG3_64BIT_REG_LOW,
5712 tp->rx_jumbo_ptr);
5713
5714 /* Initialize MAC address and backoff seed. */
5715 __tg3_set_mac_addr(tp);
5716
5717 /* MTU + ethernet header + FCS + optional VLAN tag */
5718 tw32(MAC_RX_MTU_SIZE, tp->dev->mtu + ETH_HLEN + 8);
5719
5720 /* The slot time is changed by tg3_setup_phy if we
5721 * run at gigabit with half duplex.
5722 */
5723 tw32(MAC_TX_LENGTHS,
5724 (2 << TX_LENGTHS_IPG_CRS_SHIFT) |
5725 (6 << TX_LENGTHS_IPG_SHIFT) |
5726 (32 << TX_LENGTHS_SLOT_TIME_SHIFT));
5727
5728 /* Receive rules. */
5729 tw32(MAC_RCV_RULE_CFG, RCV_RULE_CFG_DEFAULT_CLASS);
5730 tw32(RCVLPC_CONFIG, 0x0181);
5731
5732 /* Calculate RDMAC_MODE setting early, we need it to determine
5733 * the RCVLPC_STATE_ENABLE mask.
5734 */
5735 rdmac_mode = (RDMAC_MODE_ENABLE | RDMAC_MODE_TGTABORT_ENAB |
5736 RDMAC_MODE_MSTABORT_ENAB | RDMAC_MODE_PARITYERR_ENAB |
5737 RDMAC_MODE_ADDROFLOW_ENAB | RDMAC_MODE_FIFOOFLOW_ENAB |
5738 RDMAC_MODE_FIFOURUN_ENAB | RDMAC_MODE_FIFOOREAD_ENAB |
5739 RDMAC_MODE_LNGREAD_ENAB);
5740 if (tp->tg3_flags & TG3_FLAG_SPLIT_MODE)
5741 rdmac_mode |= RDMAC_MODE_SPLIT_ENABLE;
85e94ced
MC
5742
5743 /* If statement applies to 5705 and 5750 PCI devices only */
5744 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705 &&
5745 tp->pci_chip_rev_id != CHIPREV_ID_5705_A0) ||
5746 (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5750)) {
1da177e4
LT
5747 if (tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE &&
5748 (tp->pci_chip_rev_id == CHIPREV_ID_5705_A1 ||
5749 tp->pci_chip_rev_id == CHIPREV_ID_5705_A2)) {
5750 rdmac_mode |= RDMAC_MODE_FIFO_SIZE_128;
5751 } else if (!(tr32(TG3PCI_PCISTATE) & PCISTATE_BUS_SPEED_HIGH) &&
5752 !(tp->tg3_flags2 & TG3_FLG2_IS_5788)) {
5753 rdmac_mode |= RDMAC_MODE_FIFO_LONG_BURST;
5754 }
5755 }
5756
85e94ced
MC
5757 if (tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS)
5758 rdmac_mode |= RDMAC_MODE_FIFO_LONG_BURST;
5759
1da177e4
LT
5760#if TG3_TSO_SUPPORT != 0
5761 if (tp->tg3_flags2 & TG3_FLG2_HW_TSO)
5762 rdmac_mode |= (1 << 27);
5763#endif
5764
5765 /* Receive/send statistics. */
5766 if ((rdmac_mode & RDMAC_MODE_FIFO_SIZE_128) &&
5767 (tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE)) {
5768 val = tr32(RCVLPC_STATS_ENABLE);
5769 val &= ~RCVLPC_STATSENAB_LNGBRST_RFIX;
5770 tw32(RCVLPC_STATS_ENABLE, val);
5771 } else {
5772 tw32(RCVLPC_STATS_ENABLE, 0xffffff);
5773 }
5774 tw32(RCVLPC_STATSCTRL, RCVLPC_STATSCTRL_ENABLE);
5775 tw32(SNDDATAI_STATSENAB, 0xffffff);
5776 tw32(SNDDATAI_STATSCTRL,
5777 (SNDDATAI_SCTRL_ENABLE |
5778 SNDDATAI_SCTRL_FASTUPD));
5779
5780 /* Setup host coalescing engine. */
5781 tw32(HOSTCC_MODE, 0);
5782 for (i = 0; i < 2000; i++) {
5783 if (!(tr32(HOSTCC_MODE) & HOSTCC_MODE_ENABLE))
5784 break;
5785 udelay(10);
5786 }
5787
d244c892 5788 __tg3_set_coalesce(tp, &tp->coal);
1da177e4
LT
5789
5790 /* set status block DMA address */
5791 tw32(HOSTCC_STATUS_BLK_HOST_ADDR + TG3_64BIT_REG_HIGH,
5792 ((u64) tp->status_mapping >> 32));
5793 tw32(HOSTCC_STATUS_BLK_HOST_ADDR + TG3_64BIT_REG_LOW,
5794 ((u64) tp->status_mapping & 0xffffffff));
5795
5796 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
5797 /* Status/statistics block address. See tg3_timer,
5798 * the tg3_periodic_fetch_stats call there, and
5799 * tg3_get_stats to see how this works for 5705/5750 chips.
5800 */
1da177e4
LT
5801 tw32(HOSTCC_STATS_BLK_HOST_ADDR + TG3_64BIT_REG_HIGH,
5802 ((u64) tp->stats_mapping >> 32));
5803 tw32(HOSTCC_STATS_BLK_HOST_ADDR + TG3_64BIT_REG_LOW,
5804 ((u64) tp->stats_mapping & 0xffffffff));
5805 tw32(HOSTCC_STATS_BLK_NIC_ADDR, NIC_SRAM_STATS_BLK);
5806 tw32(HOSTCC_STATUS_BLK_NIC_ADDR, NIC_SRAM_STATUS_BLK);
5807 }
5808
5809 tw32(HOSTCC_MODE, HOSTCC_MODE_ENABLE | tp->coalesce_mode);
5810
5811 tw32(RCVCC_MODE, RCVCC_MODE_ENABLE | RCVCC_MODE_ATTN_ENABLE);
5812 tw32(RCVLPC_MODE, RCVLPC_MODE_ENABLE);
5813 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS))
5814 tw32(RCVLSC_MODE, RCVLSC_MODE_ENABLE | RCVLSC_MODE_ATTN_ENABLE);
5815
5816 /* Clear statistics/status block in chip, and status block in ram. */
5817 for (i = NIC_SRAM_STATS_BLK;
5818 i < NIC_SRAM_STATUS_BLK + TG3_HW_STATUS_SIZE;
5819 i += sizeof(u32)) {
5820 tg3_write_mem(tp, i, 0);
5821 udelay(40);
5822 }
5823 memset(tp->hw_status, 0, TG3_HW_STATUS_SIZE);
5824
c94e3941
MC
5825 if (tp->tg3_flags2 & TG3_FLG2_MII_SERDES) {
5826 tp->tg3_flags2 &= ~TG3_FLG2_PARALLEL_DETECT;
5827 /* reset to prevent losing 1st rx packet intermittently */
5828 tw32_f(MAC_RX_MODE, RX_MODE_RESET);
5829 udelay(10);
5830 }
5831
1da177e4
LT
5832 tp->mac_mode = MAC_MODE_TXSTAT_ENABLE | MAC_MODE_RXSTAT_ENABLE |
5833 MAC_MODE_TDE_ENABLE | MAC_MODE_RDE_ENABLE | MAC_MODE_FHDE_ENABLE;
5834 tw32_f(MAC_MODE, tp->mac_mode | MAC_MODE_RXSTAT_CLEAR | MAC_MODE_TXSTAT_CLEAR);
5835 udelay(40);
5836
314fba34
MC
5837 /* tp->grc_local_ctrl is partially set up during tg3_get_invariants().
5838 * If TG3_FLAG_EEPROM_WRITE_PROT is set, we should read the
5839 * register to preserve the GPIO settings for LOMs. The GPIOs,
5840 * whether used as inputs or outputs, are set by boot code after
5841 * reset.
5842 */
5843 if (tp->tg3_flags & TG3_FLAG_EEPROM_WRITE_PROT) {
5844 u32 gpio_mask;
5845
5846 gpio_mask = GRC_LCLCTRL_GPIO_OE0 | GRC_LCLCTRL_GPIO_OE2 |
5847 GRC_LCLCTRL_GPIO_OUTPUT0 | GRC_LCLCTRL_GPIO_OUTPUT2;
3e7d83bc
MC
5848
5849 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5752)
5850 gpio_mask |= GRC_LCLCTRL_GPIO_OE3 |
5851 GRC_LCLCTRL_GPIO_OUTPUT3;
5852
314fba34
MC
5853 tp->grc_local_ctrl |= tr32(GRC_LOCAL_CTRL) & gpio_mask;
5854
5855 /* GPIO1 must be driven high for eeprom write protect */
1da177e4
LT
5856 tp->grc_local_ctrl |= (GRC_LCLCTRL_GPIO_OE1 |
5857 GRC_LCLCTRL_GPIO_OUTPUT1);
314fba34 5858 }
1da177e4
LT
5859 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl);
5860 udelay(100);
5861
09ee929c 5862 tw32_mailbox_f(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0);
fac9b83e 5863 tp->last_tag = 0;
1da177e4
LT
5864
5865 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
5866 tw32_f(DMAC_MODE, DMAC_MODE_ENABLE);
5867 udelay(40);
5868 }
5869
5870 val = (WDMAC_MODE_ENABLE | WDMAC_MODE_TGTABORT_ENAB |
5871 WDMAC_MODE_MSTABORT_ENAB | WDMAC_MODE_PARITYERR_ENAB |
5872 WDMAC_MODE_ADDROFLOW_ENAB | WDMAC_MODE_FIFOOFLOW_ENAB |
5873 WDMAC_MODE_FIFOURUN_ENAB | WDMAC_MODE_FIFOOREAD_ENAB |
5874 WDMAC_MODE_LNGREAD_ENAB);
5875
85e94ced
MC
5876 /* If statement applies to 5705 and 5750 PCI devices only */
5877 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705 &&
5878 tp->pci_chip_rev_id != CHIPREV_ID_5705_A0) ||
5879 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5750) {
1da177e4
LT
5880 if ((tp->tg3_flags & TG3_FLG2_TSO_CAPABLE) &&
5881 (tp->pci_chip_rev_id == CHIPREV_ID_5705_A1 ||
5882 tp->pci_chip_rev_id == CHIPREV_ID_5705_A2)) {
5883 /* nothing */
5884 } else if (!(tr32(TG3PCI_PCISTATE) & PCISTATE_BUS_SPEED_HIGH) &&
5885 !(tp->tg3_flags2 & TG3_FLG2_IS_5788) &&
5886 !(tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS)) {
5887 val |= WDMAC_MODE_RX_ACCEL;
5888 }
5889 }
5890
5891 tw32_f(WDMAC_MODE, val);
5892 udelay(40);
5893
5894 if ((tp->tg3_flags & TG3_FLAG_PCIX_MODE) != 0) {
5895 val = tr32(TG3PCI_X_CAPS);
5896 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703) {
5897 val &= ~PCIX_CAPS_BURST_MASK;
5898 val |= (PCIX_CAPS_MAX_BURST_CPIOB << PCIX_CAPS_BURST_SHIFT);
5899 } else if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704) {
5900 val &= ~(PCIX_CAPS_SPLIT_MASK | PCIX_CAPS_BURST_MASK);
5901 val |= (PCIX_CAPS_MAX_BURST_CPIOB << PCIX_CAPS_BURST_SHIFT);
5902 if (tp->tg3_flags & TG3_FLAG_SPLIT_MODE)
5903 val |= (tp->split_mode_max_reqs <<
5904 PCIX_CAPS_SPLIT_SHIFT);
5905 }
5906 tw32(TG3PCI_X_CAPS, val);
5907 }
5908
5909 tw32_f(RDMAC_MODE, rdmac_mode);
5910 udelay(40);
5911
5912 tw32(RCVDCC_MODE, RCVDCC_MODE_ENABLE | RCVDCC_MODE_ATTN_ENABLE);
5913 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS))
5914 tw32(MBFREE_MODE, MBFREE_MODE_ENABLE);
5915 tw32(SNDDATAC_MODE, SNDDATAC_MODE_ENABLE);
5916 tw32(SNDBDC_MODE, SNDBDC_MODE_ENABLE | SNDBDC_MODE_ATTN_ENABLE);
5917 tw32(RCVBDI_MODE, RCVBDI_MODE_ENABLE | RCVBDI_MODE_RCB_ATTN_ENAB);
5918 tw32(RCVDBDI_MODE, RCVDBDI_MODE_ENABLE | RCVDBDI_MODE_INV_RING_SZ);
5919 tw32(SNDDATAI_MODE, SNDDATAI_MODE_ENABLE);
5920#if TG3_TSO_SUPPORT != 0
5921 if (tp->tg3_flags2 & TG3_FLG2_HW_TSO)
5922 tw32(SNDDATAI_MODE, SNDDATAI_MODE_ENABLE | 0x8);
5923#endif
5924 tw32(SNDBDI_MODE, SNDBDI_MODE_ENABLE | SNDBDI_MODE_ATTN_ENABLE);
5925 tw32(SNDBDS_MODE, SNDBDS_MODE_ENABLE | SNDBDS_MODE_ATTN_ENABLE);
5926
5927 if (tp->pci_chip_rev_id == CHIPREV_ID_5701_A0) {
5928 err = tg3_load_5701_a0_firmware_fix(tp);
5929 if (err)
5930 return err;
5931 }
5932
5933#if TG3_TSO_SUPPORT != 0
5934 if (tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE) {
5935 err = tg3_load_tso_firmware(tp);
5936 if (err)
5937 return err;
5938 }
5939#endif
5940
5941 tp->tx_mode = TX_MODE_ENABLE;
5942 tw32_f(MAC_TX_MODE, tp->tx_mode);
5943 udelay(100);
5944
5945 tp->rx_mode = RX_MODE_ENABLE;
5946 tw32_f(MAC_RX_MODE, tp->rx_mode);
5947 udelay(10);
5948
5949 if (tp->link_config.phy_is_low_power) {
5950 tp->link_config.phy_is_low_power = 0;
5951 tp->link_config.speed = tp->link_config.orig_speed;
5952 tp->link_config.duplex = tp->link_config.orig_duplex;
5953 tp->link_config.autoneg = tp->link_config.orig_autoneg;
5954 }
5955
5956 tp->mi_mode = MAC_MI_MODE_BASE;
5957 tw32_f(MAC_MI_MODE, tp->mi_mode);
5958 udelay(80);
5959
5960 tw32(MAC_LED_CTRL, tp->led_ctrl);
5961
5962 tw32(MAC_MI_STAT, MAC_MI_STAT_LNKSTAT_ATTN_ENAB);
c94e3941 5963 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES) {
1da177e4
LT
5964 tw32_f(MAC_RX_MODE, RX_MODE_RESET);
5965 udelay(10);
5966 }
5967 tw32_f(MAC_RX_MODE, tp->rx_mode);
5968 udelay(10);
5969
5970 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES) {
5971 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704) &&
5972 !(tp->tg3_flags2 & TG3_FLG2_SERDES_PREEMPHASIS)) {
5973 /* Set drive transmission level to 1.2V */
5974 /* only if the signal pre-emphasis bit is not set */
5975 val = tr32(MAC_SERDES_CFG);
5976 val &= 0xfffff000;
5977 val |= 0x880;
5978 tw32(MAC_SERDES_CFG, val);
5979 }
5980 if (tp->pci_chip_rev_id == CHIPREV_ID_5703_A1)
5981 tw32(MAC_SERDES_CFG, 0x616000);
5982 }
5983
5984 /* Prevent chip from dropping frames when flow control
5985 * is enabled.
5986 */
5987 tw32_f(MAC_LOW_WMARK_MAX_RX_FRAME, 2);
5988
5989 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704 &&
5990 (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)) {
5991 /* Use hardware link auto-negotiation */
5992 tp->tg3_flags2 |= TG3_FLG2_HW_AUTONEG;
5993 }
5994
5995 err = tg3_setup_phy(tp, 1);
5996 if (err)
5997 return err;
5998
5999 if (!(tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)) {
6000 u32 tmp;
6001
6002 /* Clear CRC stats. */
6003 if (!tg3_readphy(tp, 0x1e, &tmp)) {
6004 tg3_writephy(tp, 0x1e, tmp | 0x8000);
6005 tg3_readphy(tp, 0x14, &tmp);
6006 }
6007 }
6008
6009 __tg3_set_rx_mode(tp->dev);
6010
6011 /* Initialize receive rules. */
6012 tw32(MAC_RCV_RULE_0, 0xc2000000 & RCV_RULE_DISABLE_MASK);
6013 tw32(MAC_RCV_VALUE_0, 0xffffffff & RCV_RULE_DISABLE_MASK);
6014 tw32(MAC_RCV_RULE_1, 0x86000004 & RCV_RULE_DISABLE_MASK);
6015 tw32(MAC_RCV_VALUE_1, 0xffffffff & RCV_RULE_DISABLE_MASK);
6016
4cf78e4f 6017 if ((tp->tg3_flags2 & TG3_FLG2_5705_PLUS) &&
a4e2b347 6018 !(tp->tg3_flags2 & TG3_FLG2_5780_CLASS))
1da177e4
LT
6019 limit = 8;
6020 else
6021 limit = 16;
6022 if (tp->tg3_flags & TG3_FLAG_ENABLE_ASF)
6023 limit -= 4;
6024 switch (limit) {
6025 case 16:
6026 tw32(MAC_RCV_RULE_15, 0); tw32(MAC_RCV_VALUE_15, 0);
6027 case 15:
6028 tw32(MAC_RCV_RULE_14, 0); tw32(MAC_RCV_VALUE_14, 0);
6029 case 14:
6030 tw32(MAC_RCV_RULE_13, 0); tw32(MAC_RCV_VALUE_13, 0);
6031 case 13:
6032 tw32(MAC_RCV_RULE_12, 0); tw32(MAC_RCV_VALUE_12, 0);
6033 case 12:
6034 tw32(MAC_RCV_RULE_11, 0); tw32(MAC_RCV_VALUE_11, 0);
6035 case 11:
6036 tw32(MAC_RCV_RULE_10, 0); tw32(MAC_RCV_VALUE_10, 0);
6037 case 10:
6038 tw32(MAC_RCV_RULE_9, 0); tw32(MAC_RCV_VALUE_9, 0);
6039 case 9:
6040 tw32(MAC_RCV_RULE_8, 0); tw32(MAC_RCV_VALUE_8, 0);
6041 case 8:
6042 tw32(MAC_RCV_RULE_7, 0); tw32(MAC_RCV_VALUE_7, 0);
6043 case 7:
6044 tw32(MAC_RCV_RULE_6, 0); tw32(MAC_RCV_VALUE_6, 0);
6045 case 6:
6046 tw32(MAC_RCV_RULE_5, 0); tw32(MAC_RCV_VALUE_5, 0);
6047 case 5:
6048 tw32(MAC_RCV_RULE_4, 0); tw32(MAC_RCV_VALUE_4, 0);
6049 case 4:
6050 /* tw32(MAC_RCV_RULE_3, 0); tw32(MAC_RCV_VALUE_3, 0); */
6051 case 3:
6052 /* tw32(MAC_RCV_RULE_2, 0); tw32(MAC_RCV_VALUE_2, 0); */
6053 case 2:
6054 case 1:
6055
6056 default:
6057 break;
6058 };
6059
6060 tg3_write_sig_post_reset(tp, RESET_KIND_INIT);
6061
1da177e4
LT
6062 return 0;
6063}
6064
6065/* Called at device open time to get the chip ready for
6066 * packet processing. Invoked with tp->lock held.
6067 */
6068static int tg3_init_hw(struct tg3 *tp)
6069{
6070 int err;
6071
6072 /* Force the chip into D0. */
6073 err = tg3_set_power_state(tp, 0);
6074 if (err)
6075 goto out;
6076
6077 tg3_switch_clocks(tp);
6078
6079 tw32(TG3PCI_MEM_WIN_BASE_ADDR, 0);
6080
6081 err = tg3_reset_hw(tp);
6082
6083out:
6084 return err;
6085}
6086
6087#define TG3_STAT_ADD32(PSTAT, REG) \
6088do { u32 __val = tr32(REG); \
6089 (PSTAT)->low += __val; \
6090 if ((PSTAT)->low < __val) \
6091 (PSTAT)->high += 1; \
6092} while (0)
6093
6094static void tg3_periodic_fetch_stats(struct tg3 *tp)
6095{
6096 struct tg3_hw_stats *sp = tp->hw_stats;
6097
6098 if (!netif_carrier_ok(tp->dev))
6099 return;
6100
6101 TG3_STAT_ADD32(&sp->tx_octets, MAC_TX_STATS_OCTETS);
6102 TG3_STAT_ADD32(&sp->tx_collisions, MAC_TX_STATS_COLLISIONS);
6103 TG3_STAT_ADD32(&sp->tx_xon_sent, MAC_TX_STATS_XON_SENT);
6104 TG3_STAT_ADD32(&sp->tx_xoff_sent, MAC_TX_STATS_XOFF_SENT);
6105 TG3_STAT_ADD32(&sp->tx_mac_errors, MAC_TX_STATS_MAC_ERRORS);
6106 TG3_STAT_ADD32(&sp->tx_single_collisions, MAC_TX_STATS_SINGLE_COLLISIONS);
6107 TG3_STAT_ADD32(&sp->tx_mult_collisions, MAC_TX_STATS_MULT_COLLISIONS);
6108 TG3_STAT_ADD32(&sp->tx_deferred, MAC_TX_STATS_DEFERRED);
6109 TG3_STAT_ADD32(&sp->tx_excessive_collisions, MAC_TX_STATS_EXCESSIVE_COL);
6110 TG3_STAT_ADD32(&sp->tx_late_collisions, MAC_TX_STATS_LATE_COL);
6111 TG3_STAT_ADD32(&sp->tx_ucast_packets, MAC_TX_STATS_UCAST);
6112 TG3_STAT_ADD32(&sp->tx_mcast_packets, MAC_TX_STATS_MCAST);
6113 TG3_STAT_ADD32(&sp->tx_bcast_packets, MAC_TX_STATS_BCAST);
6114
6115 TG3_STAT_ADD32(&sp->rx_octets, MAC_RX_STATS_OCTETS);
6116 TG3_STAT_ADD32(&sp->rx_fragments, MAC_RX_STATS_FRAGMENTS);
6117 TG3_STAT_ADD32(&sp->rx_ucast_packets, MAC_RX_STATS_UCAST);
6118 TG3_STAT_ADD32(&sp->rx_mcast_packets, MAC_RX_STATS_MCAST);
6119 TG3_STAT_ADD32(&sp->rx_bcast_packets, MAC_RX_STATS_BCAST);
6120 TG3_STAT_ADD32(&sp->rx_fcs_errors, MAC_RX_STATS_FCS_ERRORS);
6121 TG3_STAT_ADD32(&sp->rx_align_errors, MAC_RX_STATS_ALIGN_ERRORS);
6122 TG3_STAT_ADD32(&sp->rx_xon_pause_rcvd, MAC_RX_STATS_XON_PAUSE_RECVD);
6123 TG3_STAT_ADD32(&sp->rx_xoff_pause_rcvd, MAC_RX_STATS_XOFF_PAUSE_RECVD);
6124 TG3_STAT_ADD32(&sp->rx_mac_ctrl_rcvd, MAC_RX_STATS_MAC_CTRL_RECVD);
6125 TG3_STAT_ADD32(&sp->rx_xoff_entered, MAC_RX_STATS_XOFF_ENTERED);
6126 TG3_STAT_ADD32(&sp->rx_frame_too_long_errors, MAC_RX_STATS_FRAME_TOO_LONG);
6127 TG3_STAT_ADD32(&sp->rx_jabbers, MAC_RX_STATS_JABBERS);
6128 TG3_STAT_ADD32(&sp->rx_undersize_packets, MAC_RX_STATS_UNDERSIZE);
6129}
6130
6131static void tg3_timer(unsigned long __opaque)
6132{
6133 struct tg3 *tp = (struct tg3 *) __opaque;
1da177e4 6134
f47c11ee 6135 spin_lock(&tp->lock);
1da177e4 6136
fac9b83e
DM
6137 if (!(tp->tg3_flags & TG3_FLAG_TAGGED_STATUS)) {
6138 /* All of this garbage is because when using non-tagged
6139 * IRQ status the mailbox/status_block protocol the chip
6140 * uses with the cpu is race prone.
6141 */
6142 if (tp->hw_status->status & SD_STATUS_UPDATED) {
6143 tw32(GRC_LOCAL_CTRL,
6144 tp->grc_local_ctrl | GRC_LCLCTRL_SETINT);
6145 } else {
6146 tw32(HOSTCC_MODE, tp->coalesce_mode |
6147 (HOSTCC_MODE_ENABLE | HOSTCC_MODE_NOW));
6148 }
1da177e4 6149
fac9b83e
DM
6150 if (!(tr32(WDMAC_MODE) & WDMAC_MODE_ENABLE)) {
6151 tp->tg3_flags2 |= TG3_FLG2_RESTART_TIMER;
f47c11ee 6152 spin_unlock(&tp->lock);
fac9b83e
DM
6153 schedule_work(&tp->reset_task);
6154 return;
6155 }
1da177e4
LT
6156 }
6157
1da177e4
LT
6158 /* This part only runs once per second. */
6159 if (!--tp->timer_counter) {
fac9b83e
DM
6160 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)
6161 tg3_periodic_fetch_stats(tp);
6162
1da177e4
LT
6163 if (tp->tg3_flags & TG3_FLAG_USE_LINKCHG_REG) {
6164 u32 mac_stat;
6165 int phy_event;
6166
6167 mac_stat = tr32(MAC_STATUS);
6168
6169 phy_event = 0;
6170 if (tp->tg3_flags & TG3_FLAG_USE_MI_INTERRUPT) {
6171 if (mac_stat & MAC_STATUS_MI_INTERRUPT)
6172 phy_event = 1;
6173 } else if (mac_stat & MAC_STATUS_LNKSTATE_CHANGED)
6174 phy_event = 1;
6175
6176 if (phy_event)
6177 tg3_setup_phy(tp, 0);
6178 } else if (tp->tg3_flags & TG3_FLAG_POLL_SERDES) {
6179 u32 mac_stat = tr32(MAC_STATUS);
6180 int need_setup = 0;
6181
6182 if (netif_carrier_ok(tp->dev) &&
6183 (mac_stat & MAC_STATUS_LNKSTATE_CHANGED)) {
6184 need_setup = 1;
6185 }
6186 if (! netif_carrier_ok(tp->dev) &&
6187 (mac_stat & (MAC_STATUS_PCS_SYNCED |
6188 MAC_STATUS_SIGNAL_DET))) {
6189 need_setup = 1;
6190 }
6191 if (need_setup) {
6192 tw32_f(MAC_MODE,
6193 (tp->mac_mode &
6194 ~MAC_MODE_PORT_MODE_MASK));
6195 udelay(40);
6196 tw32_f(MAC_MODE, tp->mac_mode);
6197 udelay(40);
6198 tg3_setup_phy(tp, 0);
6199 }
747e8f8b
MC
6200 } else if (tp->tg3_flags2 & TG3_FLG2_MII_SERDES)
6201 tg3_serdes_parallel_detect(tp);
1da177e4
LT
6202
6203 tp->timer_counter = tp->timer_multiplier;
6204 }
6205
28fbef78 6206 /* Heartbeat is only sent once every 2 seconds. */
1da177e4
LT
6207 if (!--tp->asf_counter) {
6208 if (tp->tg3_flags & TG3_FLAG_ENABLE_ASF) {
6209 u32 val;
6210
28fbef78
MC
6211 tg3_write_mem_fast(tp, NIC_SRAM_FW_CMD_MBOX,
6212 FWCMD_NICDRV_ALIVE2);
6213 tg3_write_mem_fast(tp, NIC_SRAM_FW_CMD_LEN_MBOX, 4);
6214 /* 5 seconds timeout */
6215 tg3_write_mem_fast(tp, NIC_SRAM_FW_CMD_DATA_MBOX, 5);
1da177e4
LT
6216 val = tr32(GRC_RX_CPU_EVENT);
6217 val |= (1 << 14);
6218 tw32(GRC_RX_CPU_EVENT, val);
6219 }
6220 tp->asf_counter = tp->asf_multiplier;
6221 }
6222
f47c11ee 6223 spin_unlock(&tp->lock);
1da177e4
LT
6224
6225 tp->timer.expires = jiffies + tp->timer_offset;
6226 add_timer(&tp->timer);
6227}
6228
7938109f
MC
6229static int tg3_test_interrupt(struct tg3 *tp)
6230{
6231 struct net_device *dev = tp->dev;
6232 int err, i;
6233 u32 int_mbox = 0;
6234
d4bc3927
MC
6235 if (!netif_running(dev))
6236 return -ENODEV;
6237
7938109f
MC
6238 tg3_disable_ints(tp);
6239
6240 free_irq(tp->pdev->irq, dev);
6241
6242 err = request_irq(tp->pdev->irq, tg3_test_isr,
f4d0ee98 6243 SA_SHIRQ | SA_SAMPLE_RANDOM, dev->name, dev);
7938109f
MC
6244 if (err)
6245 return err;
6246
38f3843e 6247 tp->hw_status->status &= ~SD_STATUS_UPDATED;
7938109f
MC
6248 tg3_enable_ints(tp);
6249
6250 tw32_f(HOSTCC_MODE, tp->coalesce_mode | HOSTCC_MODE_ENABLE |
6251 HOSTCC_MODE_NOW);
6252
6253 for (i = 0; i < 5; i++) {
09ee929c
MC
6254 int_mbox = tr32_mailbox(MAILBOX_INTERRUPT_0 +
6255 TG3_64BIT_REG_LOW);
7938109f
MC
6256 if (int_mbox != 0)
6257 break;
6258 msleep(10);
6259 }
6260
6261 tg3_disable_ints(tp);
6262
6263 free_irq(tp->pdev->irq, dev);
6264
6265 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI)
6266 err = request_irq(tp->pdev->irq, tg3_msi,
f4d0ee98 6267 SA_SAMPLE_RANDOM, dev->name, dev);
fac9b83e
DM
6268 else {
6269 irqreturn_t (*fn)(int, void *, struct pt_regs *)=tg3_interrupt;
6270 if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS)
6271 fn = tg3_interrupt_tagged;
6272 err = request_irq(tp->pdev->irq, fn,
f4d0ee98 6273 SA_SHIRQ | SA_SAMPLE_RANDOM, dev->name, dev);
fac9b83e 6274 }
7938109f
MC
6275
6276 if (err)
6277 return err;
6278
6279 if (int_mbox != 0)
6280 return 0;
6281
6282 return -EIO;
6283}
6284
6285/* Returns 0 if MSI test succeeds or MSI test fails and INTx mode is
6286 * successfully restored
6287 */
6288static int tg3_test_msi(struct tg3 *tp)
6289{
6290 struct net_device *dev = tp->dev;
6291 int err;
6292 u16 pci_cmd;
6293
6294 if (!(tp->tg3_flags2 & TG3_FLG2_USING_MSI))
6295 return 0;
6296
6297 /* Turn off SERR reporting in case MSI terminates with Master
6298 * Abort.
6299 */
6300 pci_read_config_word(tp->pdev, PCI_COMMAND, &pci_cmd);
6301 pci_write_config_word(tp->pdev, PCI_COMMAND,
6302 pci_cmd & ~PCI_COMMAND_SERR);
6303
6304 err = tg3_test_interrupt(tp);
6305
6306 pci_write_config_word(tp->pdev, PCI_COMMAND, pci_cmd);
6307
6308 if (!err)
6309 return 0;
6310
6311 /* other failures */
6312 if (err != -EIO)
6313 return err;
6314
6315 /* MSI test failed, go back to INTx mode */
6316 printk(KERN_WARNING PFX "%s: No interrupt was generated using MSI, "
6317 "switching to INTx mode. Please report this failure to "
6318 "the PCI maintainer and include system chipset information.\n",
6319 tp->dev->name);
6320
6321 free_irq(tp->pdev->irq, dev);
6322 pci_disable_msi(tp->pdev);
6323
6324 tp->tg3_flags2 &= ~TG3_FLG2_USING_MSI;
6325
fac9b83e
DM
6326 {
6327 irqreturn_t (*fn)(int, void *, struct pt_regs *)=tg3_interrupt;
6328 if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS)
6329 fn = tg3_interrupt_tagged;
7938109f 6330
fac9b83e
DM
6331 err = request_irq(tp->pdev->irq, fn,
6332 SA_SHIRQ | SA_SAMPLE_RANDOM, dev->name, dev);
6333 }
7938109f
MC
6334 if (err)
6335 return err;
6336
6337 /* Need to reset the chip because the MSI cycle may have terminated
6338 * with Master Abort.
6339 */
f47c11ee 6340 tg3_full_lock(tp, 1);
7938109f 6341
944d980e 6342 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
7938109f
MC
6343 err = tg3_init_hw(tp);
6344
f47c11ee 6345 tg3_full_unlock(tp);
7938109f
MC
6346
6347 if (err)
6348 free_irq(tp->pdev->irq, dev);
6349
6350 return err;
6351}
6352
1da177e4
LT
6353static int tg3_open(struct net_device *dev)
6354{
6355 struct tg3 *tp = netdev_priv(dev);
6356 int err;
6357
f47c11ee 6358 tg3_full_lock(tp, 0);
1da177e4
LT
6359
6360 tg3_disable_ints(tp);
6361 tp->tg3_flags &= ~TG3_FLAG_INIT_COMPLETE;
6362
f47c11ee 6363 tg3_full_unlock(tp);
1da177e4
LT
6364
6365 /* The placement of this call is tied
6366 * to the setup and use of Host TX descriptors.
6367 */
6368 err = tg3_alloc_consistent(tp);
6369 if (err)
6370 return err;
6371
88b06bc2
MC
6372 if ((tp->tg3_flags2 & TG3_FLG2_5750_PLUS) &&
6373 (GET_CHIP_REV(tp->pci_chip_rev_id) != CHIPREV_5750_AX) &&
6374 (GET_CHIP_REV(tp->pci_chip_rev_id) != CHIPREV_5750_BX)) {
fac9b83e
DM
6375 /* All MSI supporting chips should support tagged
6376 * status. Assert that this is the case.
6377 */
6378 if (!(tp->tg3_flags & TG3_FLAG_TAGGED_STATUS)) {
6379 printk(KERN_WARNING PFX "%s: MSI without TAGGED? "
6380 "Not using MSI.\n", tp->dev->name);
6381 } else if (pci_enable_msi(tp->pdev) == 0) {
88b06bc2
MC
6382 u32 msi_mode;
6383
6384 msi_mode = tr32(MSGINT_MODE);
6385 tw32(MSGINT_MODE, msi_mode | MSGINT_MODE_ENABLE);
6386 tp->tg3_flags2 |= TG3_FLG2_USING_MSI;
6387 }
6388 }
6389 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI)
6390 err = request_irq(tp->pdev->irq, tg3_msi,
f4d0ee98 6391 SA_SAMPLE_RANDOM, dev->name, dev);
fac9b83e
DM
6392 else {
6393 irqreturn_t (*fn)(int, void *, struct pt_regs *)=tg3_interrupt;
6394 if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS)
6395 fn = tg3_interrupt_tagged;
6396
6397 err = request_irq(tp->pdev->irq, fn,
f4d0ee98 6398 SA_SHIRQ | SA_SAMPLE_RANDOM, dev->name, dev);
fac9b83e 6399 }
1da177e4
LT
6400
6401 if (err) {
88b06bc2
MC
6402 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) {
6403 pci_disable_msi(tp->pdev);
6404 tp->tg3_flags2 &= ~TG3_FLG2_USING_MSI;
6405 }
1da177e4
LT
6406 tg3_free_consistent(tp);
6407 return err;
6408 }
6409
f47c11ee 6410 tg3_full_lock(tp, 0);
1da177e4
LT
6411
6412 err = tg3_init_hw(tp);
6413 if (err) {
944d980e 6414 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
1da177e4
LT
6415 tg3_free_rings(tp);
6416 } else {
fac9b83e
DM
6417 if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS)
6418 tp->timer_offset = HZ;
6419 else
6420 tp->timer_offset = HZ / 10;
6421
6422 BUG_ON(tp->timer_offset > HZ);
6423 tp->timer_counter = tp->timer_multiplier =
6424 (HZ / tp->timer_offset);
6425 tp->asf_counter = tp->asf_multiplier =
28fbef78 6426 ((HZ / tp->timer_offset) * 2);
1da177e4
LT
6427
6428 init_timer(&tp->timer);
6429 tp->timer.expires = jiffies + tp->timer_offset;
6430 tp->timer.data = (unsigned long) tp;
6431 tp->timer.function = tg3_timer;
1da177e4
LT
6432 }
6433
f47c11ee 6434 tg3_full_unlock(tp);
1da177e4
LT
6435
6436 if (err) {
88b06bc2
MC
6437 free_irq(tp->pdev->irq, dev);
6438 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) {
6439 pci_disable_msi(tp->pdev);
6440 tp->tg3_flags2 &= ~TG3_FLG2_USING_MSI;
6441 }
1da177e4
LT
6442 tg3_free_consistent(tp);
6443 return err;
6444 }
6445
7938109f
MC
6446 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) {
6447 err = tg3_test_msi(tp);
fac9b83e 6448
7938109f 6449 if (err) {
f47c11ee 6450 tg3_full_lock(tp, 0);
7938109f
MC
6451
6452 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) {
6453 pci_disable_msi(tp->pdev);
6454 tp->tg3_flags2 &= ~TG3_FLG2_USING_MSI;
6455 }
944d980e 6456 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
7938109f
MC
6457 tg3_free_rings(tp);
6458 tg3_free_consistent(tp);
6459
f47c11ee 6460 tg3_full_unlock(tp);
7938109f
MC
6461
6462 return err;
6463 }
6464 }
6465
f47c11ee 6466 tg3_full_lock(tp, 0);
1da177e4 6467
7938109f
MC
6468 add_timer(&tp->timer);
6469 tp->tg3_flags |= TG3_FLAG_INIT_COMPLETE;
1da177e4
LT
6470 tg3_enable_ints(tp);
6471
f47c11ee 6472 tg3_full_unlock(tp);
1da177e4
LT
6473
6474 netif_start_queue(dev);
6475
6476 return 0;
6477}
6478
6479#if 0
6480/*static*/ void tg3_dump_state(struct tg3 *tp)
6481{
6482 u32 val32, val32_2, val32_3, val32_4, val32_5;
6483 u16 val16;
6484 int i;
6485
6486 pci_read_config_word(tp->pdev, PCI_STATUS, &val16);
6487 pci_read_config_dword(tp->pdev, TG3PCI_PCISTATE, &val32);
6488 printk("DEBUG: PCI status [%04x] TG3PCI state[%08x]\n",
6489 val16, val32);
6490
6491 /* MAC block */
6492 printk("DEBUG: MAC_MODE[%08x] MAC_STATUS[%08x]\n",
6493 tr32(MAC_MODE), tr32(MAC_STATUS));
6494 printk(" MAC_EVENT[%08x] MAC_LED_CTRL[%08x]\n",
6495 tr32(MAC_EVENT), tr32(MAC_LED_CTRL));
6496 printk("DEBUG: MAC_TX_MODE[%08x] MAC_TX_STATUS[%08x]\n",
6497 tr32(MAC_TX_MODE), tr32(MAC_TX_STATUS));
6498 printk(" MAC_RX_MODE[%08x] MAC_RX_STATUS[%08x]\n",
6499 tr32(MAC_RX_MODE), tr32(MAC_RX_STATUS));
6500
6501 /* Send data initiator control block */
6502 printk("DEBUG: SNDDATAI_MODE[%08x] SNDDATAI_STATUS[%08x]\n",
6503 tr32(SNDDATAI_MODE), tr32(SNDDATAI_STATUS));
6504 printk(" SNDDATAI_STATSCTRL[%08x]\n",
6505 tr32(SNDDATAI_STATSCTRL));
6506
6507 /* Send data completion control block */
6508 printk("DEBUG: SNDDATAC_MODE[%08x]\n", tr32(SNDDATAC_MODE));
6509
6510 /* Send BD ring selector block */
6511 printk("DEBUG: SNDBDS_MODE[%08x] SNDBDS_STATUS[%08x]\n",
6512 tr32(SNDBDS_MODE), tr32(SNDBDS_STATUS));
6513
6514 /* Send BD initiator control block */
6515 printk("DEBUG: SNDBDI_MODE[%08x] SNDBDI_STATUS[%08x]\n",
6516 tr32(SNDBDI_MODE), tr32(SNDBDI_STATUS));
6517
6518 /* Send BD completion control block */
6519 printk("DEBUG: SNDBDC_MODE[%08x]\n", tr32(SNDBDC_MODE));
6520
6521 /* Receive list placement control block */
6522 printk("DEBUG: RCVLPC_MODE[%08x] RCVLPC_STATUS[%08x]\n",
6523 tr32(RCVLPC_MODE), tr32(RCVLPC_STATUS));
6524 printk(" RCVLPC_STATSCTRL[%08x]\n",
6525 tr32(RCVLPC_STATSCTRL));
6526
6527 /* Receive data and receive BD initiator control block */
6528 printk("DEBUG: RCVDBDI_MODE[%08x] RCVDBDI_STATUS[%08x]\n",
6529 tr32(RCVDBDI_MODE), tr32(RCVDBDI_STATUS));
6530
6531 /* Receive data completion control block */
6532 printk("DEBUG: RCVDCC_MODE[%08x]\n",
6533 tr32(RCVDCC_MODE));
6534
6535 /* Receive BD initiator control block */
6536 printk("DEBUG: RCVBDI_MODE[%08x] RCVBDI_STATUS[%08x]\n",
6537 tr32(RCVBDI_MODE), tr32(RCVBDI_STATUS));
6538
6539 /* Receive BD completion control block */
6540 printk("DEBUG: RCVCC_MODE[%08x] RCVCC_STATUS[%08x]\n",
6541 tr32(RCVCC_MODE), tr32(RCVCC_STATUS));
6542
6543 /* Receive list selector control block */
6544 printk("DEBUG: RCVLSC_MODE[%08x] RCVLSC_STATUS[%08x]\n",
6545 tr32(RCVLSC_MODE), tr32(RCVLSC_STATUS));
6546
6547 /* Mbuf cluster free block */
6548 printk("DEBUG: MBFREE_MODE[%08x] MBFREE_STATUS[%08x]\n",
6549 tr32(MBFREE_MODE), tr32(MBFREE_STATUS));
6550
6551 /* Host coalescing control block */
6552 printk("DEBUG: HOSTCC_MODE[%08x] HOSTCC_STATUS[%08x]\n",
6553 tr32(HOSTCC_MODE), tr32(HOSTCC_STATUS));
6554 printk("DEBUG: HOSTCC_STATS_BLK_HOST_ADDR[%08x%08x]\n",
6555 tr32(HOSTCC_STATS_BLK_HOST_ADDR + TG3_64BIT_REG_HIGH),
6556 tr32(HOSTCC_STATS_BLK_HOST_ADDR + TG3_64BIT_REG_LOW));
6557 printk("DEBUG: HOSTCC_STATUS_BLK_HOST_ADDR[%08x%08x]\n",
6558 tr32(HOSTCC_STATUS_BLK_HOST_ADDR + TG3_64BIT_REG_HIGH),
6559 tr32(HOSTCC_STATUS_BLK_HOST_ADDR + TG3_64BIT_REG_LOW));
6560 printk("DEBUG: HOSTCC_STATS_BLK_NIC_ADDR[%08x]\n",
6561 tr32(HOSTCC_STATS_BLK_NIC_ADDR));
6562 printk("DEBUG: HOSTCC_STATUS_BLK_NIC_ADDR[%08x]\n",
6563 tr32(HOSTCC_STATUS_BLK_NIC_ADDR));
6564
6565 /* Memory arbiter control block */
6566 printk("DEBUG: MEMARB_MODE[%08x] MEMARB_STATUS[%08x]\n",
6567 tr32(MEMARB_MODE), tr32(MEMARB_STATUS));
6568
6569 /* Buffer manager control block */
6570 printk("DEBUG: BUFMGR_MODE[%08x] BUFMGR_STATUS[%08x]\n",
6571 tr32(BUFMGR_MODE), tr32(BUFMGR_STATUS));
6572 printk("DEBUG: BUFMGR_MB_POOL_ADDR[%08x] BUFMGR_MB_POOL_SIZE[%08x]\n",
6573 tr32(BUFMGR_MB_POOL_ADDR), tr32(BUFMGR_MB_POOL_SIZE));
6574 printk("DEBUG: BUFMGR_DMA_DESC_POOL_ADDR[%08x] "
6575 "BUFMGR_DMA_DESC_POOL_SIZE[%08x]\n",
6576 tr32(BUFMGR_DMA_DESC_POOL_ADDR),
6577 tr32(BUFMGR_DMA_DESC_POOL_SIZE));
6578
6579 /* Read DMA control block */
6580 printk("DEBUG: RDMAC_MODE[%08x] RDMAC_STATUS[%08x]\n",
6581 tr32(RDMAC_MODE), tr32(RDMAC_STATUS));
6582
6583 /* Write DMA control block */
6584 printk("DEBUG: WDMAC_MODE[%08x] WDMAC_STATUS[%08x]\n",
6585 tr32(WDMAC_MODE), tr32(WDMAC_STATUS));
6586
6587 /* DMA completion block */
6588 printk("DEBUG: DMAC_MODE[%08x]\n",
6589 tr32(DMAC_MODE));
6590
6591 /* GRC block */
6592 printk("DEBUG: GRC_MODE[%08x] GRC_MISC_CFG[%08x]\n",
6593 tr32(GRC_MODE), tr32(GRC_MISC_CFG));
6594 printk("DEBUG: GRC_LOCAL_CTRL[%08x]\n",
6595 tr32(GRC_LOCAL_CTRL));
6596
6597 /* TG3_BDINFOs */
6598 printk("DEBUG: RCVDBDI_JUMBO_BD[%08x%08x:%08x:%08x]\n",
6599 tr32(RCVDBDI_JUMBO_BD + 0x0),
6600 tr32(RCVDBDI_JUMBO_BD + 0x4),
6601 tr32(RCVDBDI_JUMBO_BD + 0x8),
6602 tr32(RCVDBDI_JUMBO_BD + 0xc));
6603 printk("DEBUG: RCVDBDI_STD_BD[%08x%08x:%08x:%08x]\n",
6604 tr32(RCVDBDI_STD_BD + 0x0),
6605 tr32(RCVDBDI_STD_BD + 0x4),
6606 tr32(RCVDBDI_STD_BD + 0x8),
6607 tr32(RCVDBDI_STD_BD + 0xc));
6608 printk("DEBUG: RCVDBDI_MINI_BD[%08x%08x:%08x:%08x]\n",
6609 tr32(RCVDBDI_MINI_BD + 0x0),
6610 tr32(RCVDBDI_MINI_BD + 0x4),
6611 tr32(RCVDBDI_MINI_BD + 0x8),
6612 tr32(RCVDBDI_MINI_BD + 0xc));
6613
6614 tg3_read_mem(tp, NIC_SRAM_SEND_RCB + 0x0, &val32);
6615 tg3_read_mem(tp, NIC_SRAM_SEND_RCB + 0x4, &val32_2);
6616 tg3_read_mem(tp, NIC_SRAM_SEND_RCB + 0x8, &val32_3);
6617 tg3_read_mem(tp, NIC_SRAM_SEND_RCB + 0xc, &val32_4);
6618 printk("DEBUG: SRAM_SEND_RCB_0[%08x%08x:%08x:%08x]\n",
6619 val32, val32_2, val32_3, val32_4);
6620
6621 tg3_read_mem(tp, NIC_SRAM_RCV_RET_RCB + 0x0, &val32);
6622 tg3_read_mem(tp, NIC_SRAM_RCV_RET_RCB + 0x4, &val32_2);
6623 tg3_read_mem(tp, NIC_SRAM_RCV_RET_RCB + 0x8, &val32_3);
6624 tg3_read_mem(tp, NIC_SRAM_RCV_RET_RCB + 0xc, &val32_4);
6625 printk("DEBUG: SRAM_RCV_RET_RCB_0[%08x%08x:%08x:%08x]\n",
6626 val32, val32_2, val32_3, val32_4);
6627
6628 tg3_read_mem(tp, NIC_SRAM_STATUS_BLK + 0x0, &val32);
6629 tg3_read_mem(tp, NIC_SRAM_STATUS_BLK + 0x4, &val32_2);
6630 tg3_read_mem(tp, NIC_SRAM_STATUS_BLK + 0x8, &val32_3);
6631 tg3_read_mem(tp, NIC_SRAM_STATUS_BLK + 0xc, &val32_4);
6632 tg3_read_mem(tp, NIC_SRAM_STATUS_BLK + 0x10, &val32_5);
6633 printk("DEBUG: SRAM_STATUS_BLK[%08x:%08x:%08x:%08x:%08x]\n",
6634 val32, val32_2, val32_3, val32_4, val32_5);
6635
6636 /* SW status block */
6637 printk("DEBUG: Host status block [%08x:%08x:(%04x:%04x:%04x):(%04x:%04x)]\n",
6638 tp->hw_status->status,
6639 tp->hw_status->status_tag,
6640 tp->hw_status->rx_jumbo_consumer,
6641 tp->hw_status->rx_consumer,
6642 tp->hw_status->rx_mini_consumer,
6643 tp->hw_status->idx[0].rx_producer,
6644 tp->hw_status->idx[0].tx_consumer);
6645
6646 /* SW statistics block */
6647 printk("DEBUG: Host statistics block [%08x:%08x:%08x:%08x]\n",
6648 ((u32 *)tp->hw_stats)[0],
6649 ((u32 *)tp->hw_stats)[1],
6650 ((u32 *)tp->hw_stats)[2],
6651 ((u32 *)tp->hw_stats)[3]);
6652
6653 /* Mailboxes */
6654 printk("DEBUG: SNDHOST_PROD[%08x%08x] SNDNIC_PROD[%08x%08x]\n",
09ee929c
MC
6655 tr32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0 + 0x0),
6656 tr32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0 + 0x4),
6657 tr32_mailbox(MAILBOX_SNDNIC_PROD_IDX_0 + 0x0),
6658 tr32_mailbox(MAILBOX_SNDNIC_PROD_IDX_0 + 0x4));
1da177e4
LT
6659
6660 /* NIC side send descriptors. */
6661 for (i = 0; i < 6; i++) {
6662 unsigned long txd;
6663
6664 txd = tp->regs + NIC_SRAM_WIN_BASE + NIC_SRAM_TX_BUFFER_DESC
6665 + (i * sizeof(struct tg3_tx_buffer_desc));
6666 printk("DEBUG: NIC TXD(%d)[%08x:%08x:%08x:%08x]\n",
6667 i,
6668 readl(txd + 0x0), readl(txd + 0x4),
6669 readl(txd + 0x8), readl(txd + 0xc));
6670 }
6671
6672 /* NIC side RX descriptors. */
6673 for (i = 0; i < 6; i++) {
6674 unsigned long rxd;
6675
6676 rxd = tp->regs + NIC_SRAM_WIN_BASE + NIC_SRAM_RX_BUFFER_DESC
6677 + (i * sizeof(struct tg3_rx_buffer_desc));
6678 printk("DEBUG: NIC RXD_STD(%d)[0][%08x:%08x:%08x:%08x]\n",
6679 i,
6680 readl(rxd + 0x0), readl(rxd + 0x4),
6681 readl(rxd + 0x8), readl(rxd + 0xc));
6682 rxd += (4 * sizeof(u32));
6683 printk("DEBUG: NIC RXD_STD(%d)[1][%08x:%08x:%08x:%08x]\n",
6684 i,
6685 readl(rxd + 0x0), readl(rxd + 0x4),
6686 readl(rxd + 0x8), readl(rxd + 0xc));
6687 }
6688
6689 for (i = 0; i < 6; i++) {
6690 unsigned long rxd;
6691
6692 rxd = tp->regs + NIC_SRAM_WIN_BASE + NIC_SRAM_RX_JUMBO_BUFFER_DESC
6693 + (i * sizeof(struct tg3_rx_buffer_desc));
6694 printk("DEBUG: NIC RXD_JUMBO(%d)[0][%08x:%08x:%08x:%08x]\n",
6695 i,
6696 readl(rxd + 0x0), readl(rxd + 0x4),
6697 readl(rxd + 0x8), readl(rxd + 0xc));
6698 rxd += (4 * sizeof(u32));
6699 printk("DEBUG: NIC RXD_JUMBO(%d)[1][%08x:%08x:%08x:%08x]\n",
6700 i,
6701 readl(rxd + 0x0), readl(rxd + 0x4),
6702 readl(rxd + 0x8), readl(rxd + 0xc));
6703 }
6704}
6705#endif
6706
6707static struct net_device_stats *tg3_get_stats(struct net_device *);
6708static struct tg3_ethtool_stats *tg3_get_estats(struct tg3 *);
6709
6710static int tg3_close(struct net_device *dev)
6711{
6712 struct tg3 *tp = netdev_priv(dev);
6713
6714 netif_stop_queue(dev);
6715
6716 del_timer_sync(&tp->timer);
6717
f47c11ee 6718 tg3_full_lock(tp, 1);
1da177e4
LT
6719#if 0
6720 tg3_dump_state(tp);
6721#endif
6722
6723 tg3_disable_ints(tp);
6724
944d980e 6725 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
1da177e4
LT
6726 tg3_free_rings(tp);
6727 tp->tg3_flags &=
6728 ~(TG3_FLAG_INIT_COMPLETE |
6729 TG3_FLAG_GOT_SERDES_FLOWCTL);
6730 netif_carrier_off(tp->dev);
6731
f47c11ee 6732 tg3_full_unlock(tp);
1da177e4 6733
88b06bc2
MC
6734 free_irq(tp->pdev->irq, dev);
6735 if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) {
6736 pci_disable_msi(tp->pdev);
6737 tp->tg3_flags2 &= ~TG3_FLG2_USING_MSI;
6738 }
1da177e4
LT
6739
6740 memcpy(&tp->net_stats_prev, tg3_get_stats(tp->dev),
6741 sizeof(tp->net_stats_prev));
6742 memcpy(&tp->estats_prev, tg3_get_estats(tp),
6743 sizeof(tp->estats_prev));
6744
6745 tg3_free_consistent(tp);
6746
6747 return 0;
6748}
6749
6750static inline unsigned long get_stat64(tg3_stat64_t *val)
6751{
6752 unsigned long ret;
6753
6754#if (BITS_PER_LONG == 32)
6755 ret = val->low;
6756#else
6757 ret = ((u64)val->high << 32) | ((u64)val->low);
6758#endif
6759 return ret;
6760}
6761
6762static unsigned long calc_crc_errors(struct tg3 *tp)
6763{
6764 struct tg3_hw_stats *hw_stats = tp->hw_stats;
6765
6766 if (!(tp->tg3_flags2 & TG3_FLG2_PHY_SERDES) &&
6767 (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
6768 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701)) {
1da177e4
LT
6769 u32 val;
6770
f47c11ee 6771 spin_lock_bh(&tp->lock);
1da177e4
LT
6772 if (!tg3_readphy(tp, 0x1e, &val)) {
6773 tg3_writephy(tp, 0x1e, val | 0x8000);
6774 tg3_readphy(tp, 0x14, &val);
6775 } else
6776 val = 0;
f47c11ee 6777 spin_unlock_bh(&tp->lock);
1da177e4
LT
6778
6779 tp->phy_crc_errors += val;
6780
6781 return tp->phy_crc_errors;
6782 }
6783
6784 return get_stat64(&hw_stats->rx_fcs_errors);
6785}
6786
6787#define ESTAT_ADD(member) \
6788 estats->member = old_estats->member + \
6789 get_stat64(&hw_stats->member)
6790
6791static struct tg3_ethtool_stats *tg3_get_estats(struct tg3 *tp)
6792{
6793 struct tg3_ethtool_stats *estats = &tp->estats;
6794 struct tg3_ethtool_stats *old_estats = &tp->estats_prev;
6795 struct tg3_hw_stats *hw_stats = tp->hw_stats;
6796
6797 if (!hw_stats)
6798 return old_estats;
6799
6800 ESTAT_ADD(rx_octets);
6801 ESTAT_ADD(rx_fragments);
6802 ESTAT_ADD(rx_ucast_packets);
6803 ESTAT_ADD(rx_mcast_packets);
6804 ESTAT_ADD(rx_bcast_packets);
6805 ESTAT_ADD(rx_fcs_errors);
6806 ESTAT_ADD(rx_align_errors);
6807 ESTAT_ADD(rx_xon_pause_rcvd);
6808 ESTAT_ADD(rx_xoff_pause_rcvd);
6809 ESTAT_ADD(rx_mac_ctrl_rcvd);
6810 ESTAT_ADD(rx_xoff_entered);
6811 ESTAT_ADD(rx_frame_too_long_errors);
6812 ESTAT_ADD(rx_jabbers);
6813 ESTAT_ADD(rx_undersize_packets);
6814 ESTAT_ADD(rx_in_length_errors);
6815 ESTAT_ADD(rx_out_length_errors);
6816 ESTAT_ADD(rx_64_or_less_octet_packets);
6817 ESTAT_ADD(rx_65_to_127_octet_packets);
6818 ESTAT_ADD(rx_128_to_255_octet_packets);
6819 ESTAT_ADD(rx_256_to_511_octet_packets);
6820 ESTAT_ADD(rx_512_to_1023_octet_packets);
6821 ESTAT_ADD(rx_1024_to_1522_octet_packets);
6822 ESTAT_ADD(rx_1523_to_2047_octet_packets);
6823 ESTAT_ADD(rx_2048_to_4095_octet_packets);
6824 ESTAT_ADD(rx_4096_to_8191_octet_packets);
6825 ESTAT_ADD(rx_8192_to_9022_octet_packets);
6826
6827 ESTAT_ADD(tx_octets);
6828 ESTAT_ADD(tx_collisions);
6829 ESTAT_ADD(tx_xon_sent);
6830 ESTAT_ADD(tx_xoff_sent);
6831 ESTAT_ADD(tx_flow_control);
6832 ESTAT_ADD(tx_mac_errors);
6833 ESTAT_ADD(tx_single_collisions);
6834 ESTAT_ADD(tx_mult_collisions);
6835 ESTAT_ADD(tx_deferred);
6836 ESTAT_ADD(tx_excessive_collisions);
6837 ESTAT_ADD(tx_late_collisions);
6838 ESTAT_ADD(tx_collide_2times);
6839 ESTAT_ADD(tx_collide_3times);
6840 ESTAT_ADD(tx_collide_4times);
6841 ESTAT_ADD(tx_collide_5times);
6842 ESTAT_ADD(tx_collide_6times);
6843 ESTAT_ADD(tx_collide_7times);
6844 ESTAT_ADD(tx_collide_8times);
6845 ESTAT_ADD(tx_collide_9times);
6846 ESTAT_ADD(tx_collide_10times);
6847 ESTAT_ADD(tx_collide_11times);
6848 ESTAT_ADD(tx_collide_12times);
6849 ESTAT_ADD(tx_collide_13times);
6850 ESTAT_ADD(tx_collide_14times);
6851 ESTAT_ADD(tx_collide_15times);
6852 ESTAT_ADD(tx_ucast_packets);
6853 ESTAT_ADD(tx_mcast_packets);
6854 ESTAT_ADD(tx_bcast_packets);
6855 ESTAT_ADD(tx_carrier_sense_errors);
6856 ESTAT_ADD(tx_discards);
6857 ESTAT_ADD(tx_errors);
6858
6859 ESTAT_ADD(dma_writeq_full);
6860 ESTAT_ADD(dma_write_prioq_full);
6861 ESTAT_ADD(rxbds_empty);
6862 ESTAT_ADD(rx_discards);
6863 ESTAT_ADD(rx_errors);
6864 ESTAT_ADD(rx_threshold_hit);
6865
6866 ESTAT_ADD(dma_readq_full);
6867 ESTAT_ADD(dma_read_prioq_full);
6868 ESTAT_ADD(tx_comp_queue_full);
6869
6870 ESTAT_ADD(ring_set_send_prod_index);
6871 ESTAT_ADD(ring_status_update);
6872 ESTAT_ADD(nic_irqs);
6873 ESTAT_ADD(nic_avoided_irqs);
6874 ESTAT_ADD(nic_tx_threshold_hit);
6875
6876 return estats;
6877}
6878
6879static struct net_device_stats *tg3_get_stats(struct net_device *dev)
6880{
6881 struct tg3 *tp = netdev_priv(dev);
6882 struct net_device_stats *stats = &tp->net_stats;
6883 struct net_device_stats *old_stats = &tp->net_stats_prev;
6884 struct tg3_hw_stats *hw_stats = tp->hw_stats;
6885
6886 if (!hw_stats)
6887 return old_stats;
6888
6889 stats->rx_packets = old_stats->rx_packets +
6890 get_stat64(&hw_stats->rx_ucast_packets) +
6891 get_stat64(&hw_stats->rx_mcast_packets) +
6892 get_stat64(&hw_stats->rx_bcast_packets);
6893
6894 stats->tx_packets = old_stats->tx_packets +
6895 get_stat64(&hw_stats->tx_ucast_packets) +
6896 get_stat64(&hw_stats->tx_mcast_packets) +
6897 get_stat64(&hw_stats->tx_bcast_packets);
6898
6899 stats->rx_bytes = old_stats->rx_bytes +
6900 get_stat64(&hw_stats->rx_octets);
6901 stats->tx_bytes = old_stats->tx_bytes +
6902 get_stat64(&hw_stats->tx_octets);
6903
6904 stats->rx_errors = old_stats->rx_errors +
4f63b877 6905 get_stat64(&hw_stats->rx_errors);
1da177e4
LT
6906 stats->tx_errors = old_stats->tx_errors +
6907 get_stat64(&hw_stats->tx_errors) +
6908 get_stat64(&hw_stats->tx_mac_errors) +
6909 get_stat64(&hw_stats->tx_carrier_sense_errors) +
6910 get_stat64(&hw_stats->tx_discards);
6911
6912 stats->multicast = old_stats->multicast +
6913 get_stat64(&hw_stats->rx_mcast_packets);
6914 stats->collisions = old_stats->collisions +
6915 get_stat64(&hw_stats->tx_collisions);
6916
6917 stats->rx_length_errors = old_stats->rx_length_errors +
6918 get_stat64(&hw_stats->rx_frame_too_long_errors) +
6919 get_stat64(&hw_stats->rx_undersize_packets);
6920
6921 stats->rx_over_errors = old_stats->rx_over_errors +
6922 get_stat64(&hw_stats->rxbds_empty);
6923 stats->rx_frame_errors = old_stats->rx_frame_errors +
6924 get_stat64(&hw_stats->rx_align_errors);
6925 stats->tx_aborted_errors = old_stats->tx_aborted_errors +
6926 get_stat64(&hw_stats->tx_discards);
6927 stats->tx_carrier_errors = old_stats->tx_carrier_errors +
6928 get_stat64(&hw_stats->tx_carrier_sense_errors);
6929
6930 stats->rx_crc_errors = old_stats->rx_crc_errors +
6931 calc_crc_errors(tp);
6932
4f63b877
JL
6933 stats->rx_missed_errors = old_stats->rx_missed_errors +
6934 get_stat64(&hw_stats->rx_discards);
6935
1da177e4
LT
6936 return stats;
6937}
6938
6939static inline u32 calc_crc(unsigned char *buf, int len)
6940{
6941 u32 reg;
6942 u32 tmp;
6943 int j, k;
6944
6945 reg = 0xffffffff;
6946
6947 for (j = 0; j < len; j++) {
6948 reg ^= buf[j];
6949
6950 for (k = 0; k < 8; k++) {
6951 tmp = reg & 0x01;
6952
6953 reg >>= 1;
6954
6955 if (tmp) {
6956 reg ^= 0xedb88320;
6957 }
6958 }
6959 }
6960
6961 return ~reg;
6962}
6963
6964static void tg3_set_multi(struct tg3 *tp, unsigned int accept_all)
6965{
6966 /* accept or reject all multicast frames */
6967 tw32(MAC_HASH_REG_0, accept_all ? 0xffffffff : 0);
6968 tw32(MAC_HASH_REG_1, accept_all ? 0xffffffff : 0);
6969 tw32(MAC_HASH_REG_2, accept_all ? 0xffffffff : 0);
6970 tw32(MAC_HASH_REG_3, accept_all ? 0xffffffff : 0);
6971}
6972
6973static void __tg3_set_rx_mode(struct net_device *dev)
6974{
6975 struct tg3 *tp = netdev_priv(dev);
6976 u32 rx_mode;
6977
6978 rx_mode = tp->rx_mode & ~(RX_MODE_PROMISC |
6979 RX_MODE_KEEP_VLAN_TAG);
6980
6981 /* When ASF is in use, we always keep the RX_MODE_KEEP_VLAN_TAG
6982 * flag clear.
6983 */
6984#if TG3_VLAN_TAG_USED
6985 if (!tp->vlgrp &&
6986 !(tp->tg3_flags & TG3_FLAG_ENABLE_ASF))
6987 rx_mode |= RX_MODE_KEEP_VLAN_TAG;
6988#else
6989 /* By definition, VLAN is disabled always in this
6990 * case.
6991 */
6992 if (!(tp->tg3_flags & TG3_FLAG_ENABLE_ASF))
6993 rx_mode |= RX_MODE_KEEP_VLAN_TAG;
6994#endif
6995
6996 if (dev->flags & IFF_PROMISC) {
6997 /* Promiscuous mode. */
6998 rx_mode |= RX_MODE_PROMISC;
6999 } else if (dev->flags & IFF_ALLMULTI) {
7000 /* Accept all multicast. */
7001 tg3_set_multi (tp, 1);
7002 } else if (dev->mc_count < 1) {
7003 /* Reject all multicast. */
7004 tg3_set_multi (tp, 0);
7005 } else {
7006 /* Accept one or more multicast(s). */
7007 struct dev_mc_list *mclist;
7008 unsigned int i;
7009 u32 mc_filter[4] = { 0, };
7010 u32 regidx;
7011 u32 bit;
7012 u32 crc;
7013
7014 for (i = 0, mclist = dev->mc_list; mclist && i < dev->mc_count;
7015 i++, mclist = mclist->next) {
7016
7017 crc = calc_crc (mclist->dmi_addr, ETH_ALEN);
7018 bit = ~crc & 0x7f;
7019 regidx = (bit & 0x60) >> 5;
7020 bit &= 0x1f;
7021 mc_filter[regidx] |= (1 << bit);
7022 }
7023
7024 tw32(MAC_HASH_REG_0, mc_filter[0]);
7025 tw32(MAC_HASH_REG_1, mc_filter[1]);
7026 tw32(MAC_HASH_REG_2, mc_filter[2]);
7027 tw32(MAC_HASH_REG_3, mc_filter[3]);
7028 }
7029
7030 if (rx_mode != tp->rx_mode) {
7031 tp->rx_mode = rx_mode;
7032 tw32_f(MAC_RX_MODE, rx_mode);
7033 udelay(10);
7034 }
7035}
7036
7037static void tg3_set_rx_mode(struct net_device *dev)
7038{
7039 struct tg3 *tp = netdev_priv(dev);
7040
f47c11ee 7041 tg3_full_lock(tp, 0);
1da177e4 7042 __tg3_set_rx_mode(dev);
f47c11ee 7043 tg3_full_unlock(tp);
1da177e4
LT
7044}
7045
7046#define TG3_REGDUMP_LEN (32 * 1024)
7047
7048static int tg3_get_regs_len(struct net_device *dev)
7049{
7050 return TG3_REGDUMP_LEN;
7051}
7052
7053static void tg3_get_regs(struct net_device *dev,
7054 struct ethtool_regs *regs, void *_p)
7055{
7056 u32 *p = _p;
7057 struct tg3 *tp = netdev_priv(dev);
7058 u8 *orig_p = _p;
7059 int i;
7060
7061 regs->version = 0;
7062
7063 memset(p, 0, TG3_REGDUMP_LEN);
7064
f47c11ee 7065 tg3_full_lock(tp, 0);
1da177e4
LT
7066
7067#define __GET_REG32(reg) (*(p)++ = tr32(reg))
7068#define GET_REG32_LOOP(base,len) \
7069do { p = (u32 *)(orig_p + (base)); \
7070 for (i = 0; i < len; i += 4) \
7071 __GET_REG32((base) + i); \
7072} while (0)
7073#define GET_REG32_1(reg) \
7074do { p = (u32 *)(orig_p + (reg)); \
7075 __GET_REG32((reg)); \
7076} while (0)
7077
7078 GET_REG32_LOOP(TG3PCI_VENDOR, 0xb0);
7079 GET_REG32_LOOP(MAILBOX_INTERRUPT_0, 0x200);
7080 GET_REG32_LOOP(MAC_MODE, 0x4f0);
7081 GET_REG32_LOOP(SNDDATAI_MODE, 0xe0);
7082 GET_REG32_1(SNDDATAC_MODE);
7083 GET_REG32_LOOP(SNDBDS_MODE, 0x80);
7084 GET_REG32_LOOP(SNDBDI_MODE, 0x48);
7085 GET_REG32_1(SNDBDC_MODE);
7086 GET_REG32_LOOP(RCVLPC_MODE, 0x20);
7087 GET_REG32_LOOP(RCVLPC_SELLST_BASE, 0x15c);
7088 GET_REG32_LOOP(RCVDBDI_MODE, 0x0c);
7089 GET_REG32_LOOP(RCVDBDI_JUMBO_BD, 0x3c);
7090 GET_REG32_LOOP(RCVDBDI_BD_PROD_IDX_0, 0x44);
7091 GET_REG32_1(RCVDCC_MODE);
7092 GET_REG32_LOOP(RCVBDI_MODE, 0x20);
7093 GET_REG32_LOOP(RCVCC_MODE, 0x14);
7094 GET_REG32_LOOP(RCVLSC_MODE, 0x08);
7095 GET_REG32_1(MBFREE_MODE);
7096 GET_REG32_LOOP(HOSTCC_MODE, 0x100);
7097 GET_REG32_LOOP(MEMARB_MODE, 0x10);
7098 GET_REG32_LOOP(BUFMGR_MODE, 0x58);
7099 GET_REG32_LOOP(RDMAC_MODE, 0x08);
7100 GET_REG32_LOOP(WDMAC_MODE, 0x08);
7101 GET_REG32_LOOP(RX_CPU_BASE, 0x280);
7102 GET_REG32_LOOP(TX_CPU_BASE, 0x280);
7103 GET_REG32_LOOP(GRCMBOX_INTERRUPT_0, 0x110);
7104 GET_REG32_LOOP(FTQ_RESET, 0x120);
7105 GET_REG32_LOOP(MSGINT_MODE, 0x0c);
7106 GET_REG32_1(DMAC_MODE);
7107 GET_REG32_LOOP(GRC_MODE, 0x4c);
7108 if (tp->tg3_flags & TG3_FLAG_NVRAM)
7109 GET_REG32_LOOP(NVRAM_CMD, 0x24);
7110
7111#undef __GET_REG32
7112#undef GET_REG32_LOOP
7113#undef GET_REG32_1
7114
f47c11ee 7115 tg3_full_unlock(tp);
1da177e4
LT
7116}
7117
7118static int tg3_get_eeprom_len(struct net_device *dev)
7119{
7120 struct tg3 *tp = netdev_priv(dev);
7121
7122 return tp->nvram_size;
7123}
7124
7125static int tg3_nvram_read(struct tg3 *tp, u32 offset, u32 *val);
7126
7127static int tg3_get_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, u8 *data)
7128{
7129 struct tg3 *tp = netdev_priv(dev);
7130 int ret;
7131 u8 *pd;
7132 u32 i, offset, len, val, b_offset, b_count;
7133
7134 offset = eeprom->offset;
7135 len = eeprom->len;
7136 eeprom->len = 0;
7137
7138 eeprom->magic = TG3_EEPROM_MAGIC;
7139
7140 if (offset & 3) {
7141 /* adjustments to start on required 4 byte boundary */
7142 b_offset = offset & 3;
7143 b_count = 4 - b_offset;
7144 if (b_count > len) {
7145 /* i.e. offset=1 len=2 */
7146 b_count = len;
7147 }
7148 ret = tg3_nvram_read(tp, offset-b_offset, &val);
7149 if (ret)
7150 return ret;
7151 val = cpu_to_le32(val);
7152 memcpy(data, ((char*)&val) + b_offset, b_count);
7153 len -= b_count;
7154 offset += b_count;
7155 eeprom->len += b_count;
7156 }
7157
7158 /* read bytes upto the last 4 byte boundary */
7159 pd = &data[eeprom->len];
7160 for (i = 0; i < (len - (len & 3)); i += 4) {
7161 ret = tg3_nvram_read(tp, offset + i, &val);
7162 if (ret) {
7163 eeprom->len += i;
7164 return ret;
7165 }
7166 val = cpu_to_le32(val);
7167 memcpy(pd + i, &val, 4);
7168 }
7169 eeprom->len += i;
7170
7171 if (len & 3) {
7172 /* read last bytes not ending on 4 byte boundary */
7173 pd = &data[eeprom->len];
7174 b_count = len & 3;
7175 b_offset = offset + len - b_count;
7176 ret = tg3_nvram_read(tp, b_offset, &val);
7177 if (ret)
7178 return ret;
7179 val = cpu_to_le32(val);
7180 memcpy(pd, ((char*)&val), b_count);
7181 eeprom->len += b_count;
7182 }
7183 return 0;
7184}
7185
7186static int tg3_nvram_write_block(struct tg3 *tp, u32 offset, u32 len, u8 *buf);
7187
7188static int tg3_set_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, u8 *data)
7189{
7190 struct tg3 *tp = netdev_priv(dev);
7191 int ret;
7192 u32 offset, len, b_offset, odd_len, start, end;
7193 u8 *buf;
7194
7195 if (eeprom->magic != TG3_EEPROM_MAGIC)
7196 return -EINVAL;
7197
7198 offset = eeprom->offset;
7199 len = eeprom->len;
7200
7201 if ((b_offset = (offset & 3))) {
7202 /* adjustments to start on required 4 byte boundary */
7203 ret = tg3_nvram_read(tp, offset-b_offset, &start);
7204 if (ret)
7205 return ret;
7206 start = cpu_to_le32(start);
7207 len += b_offset;
7208 offset &= ~3;
1c8594b4
MC
7209 if (len < 4)
7210 len = 4;
1da177e4
LT
7211 }
7212
7213 odd_len = 0;
1c8594b4 7214 if (len & 3) {
1da177e4
LT
7215 /* adjustments to end on required 4 byte boundary */
7216 odd_len = 1;
7217 len = (len + 3) & ~3;
7218 ret = tg3_nvram_read(tp, offset+len-4, &end);
7219 if (ret)
7220 return ret;
7221 end = cpu_to_le32(end);
7222 }
7223
7224 buf = data;
7225 if (b_offset || odd_len) {
7226 buf = kmalloc(len, GFP_KERNEL);
7227 if (buf == 0)
7228 return -ENOMEM;
7229 if (b_offset)
7230 memcpy(buf, &start, 4);
7231 if (odd_len)
7232 memcpy(buf+len-4, &end, 4);
7233 memcpy(buf + b_offset, data, eeprom->len);
7234 }
7235
7236 ret = tg3_nvram_write_block(tp, offset, len, buf);
7237
7238 if (buf != data)
7239 kfree(buf);
7240
7241 return ret;
7242}
7243
7244static int tg3_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
7245{
7246 struct tg3 *tp = netdev_priv(dev);
7247
7248 cmd->supported = (SUPPORTED_Autoneg);
7249
7250 if (!(tp->tg3_flags & TG3_FLAG_10_100_ONLY))
7251 cmd->supported |= (SUPPORTED_1000baseT_Half |
7252 SUPPORTED_1000baseT_Full);
7253
a4e2b347 7254 if (!(tp->tg3_flags2 & TG3_FLG2_ANY_SERDES))
1da177e4
LT
7255 cmd->supported |= (SUPPORTED_100baseT_Half |
7256 SUPPORTED_100baseT_Full |
7257 SUPPORTED_10baseT_Half |
7258 SUPPORTED_10baseT_Full |
7259 SUPPORTED_MII);
7260 else
7261 cmd->supported |= SUPPORTED_FIBRE;
7262
7263 cmd->advertising = tp->link_config.advertising;
7264 if (netif_running(dev)) {
7265 cmd->speed = tp->link_config.active_speed;
7266 cmd->duplex = tp->link_config.active_duplex;
7267 }
7268 cmd->port = 0;
7269 cmd->phy_address = PHY_ADDR;
7270 cmd->transceiver = 0;
7271 cmd->autoneg = tp->link_config.autoneg;
7272 cmd->maxtxpkt = 0;
7273 cmd->maxrxpkt = 0;
7274 return 0;
7275}
7276
7277static int tg3_set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
7278{
7279 struct tg3 *tp = netdev_priv(dev);
7280
37ff238d 7281 if (tp->tg3_flags2 & TG3_FLG2_ANY_SERDES) {
1da177e4
LT
7282 /* These are the only valid advertisement bits allowed. */
7283 if (cmd->autoneg == AUTONEG_ENABLE &&
7284 (cmd->advertising & ~(ADVERTISED_1000baseT_Half |
7285 ADVERTISED_1000baseT_Full |
7286 ADVERTISED_Autoneg |
7287 ADVERTISED_FIBRE)))
7288 return -EINVAL;
37ff238d
MC
7289 /* Fiber can only do SPEED_1000. */
7290 else if ((cmd->autoneg != AUTONEG_ENABLE) &&
7291 (cmd->speed != SPEED_1000))
7292 return -EINVAL;
7293 /* Copper cannot force SPEED_1000. */
7294 } else if ((cmd->autoneg != AUTONEG_ENABLE) &&
7295 (cmd->speed == SPEED_1000))
7296 return -EINVAL;
7297 else if ((cmd->speed == SPEED_1000) &&
7298 (tp->tg3_flags2 & TG3_FLAG_10_100_ONLY))
7299 return -EINVAL;
1da177e4 7300
f47c11ee 7301 tg3_full_lock(tp, 0);
1da177e4
LT
7302
7303 tp->link_config.autoneg = cmd->autoneg;
7304 if (cmd->autoneg == AUTONEG_ENABLE) {
7305 tp->link_config.advertising = cmd->advertising;
7306 tp->link_config.speed = SPEED_INVALID;
7307 tp->link_config.duplex = DUPLEX_INVALID;
7308 } else {
7309 tp->link_config.advertising = 0;
7310 tp->link_config.speed = cmd->speed;
7311 tp->link_config.duplex = cmd->duplex;
7312 }
7313
7314 if (netif_running(dev))
7315 tg3_setup_phy(tp, 1);
7316
f47c11ee 7317 tg3_full_unlock(tp);
1da177e4
LT
7318
7319 return 0;
7320}
7321
7322static void tg3_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
7323{
7324 struct tg3 *tp = netdev_priv(dev);
7325
7326 strcpy(info->driver, DRV_MODULE_NAME);
7327 strcpy(info->version, DRV_MODULE_VERSION);
7328 strcpy(info->bus_info, pci_name(tp->pdev));
7329}
7330
7331static void tg3_get_wol(struct net_device *dev, struct ethtool_wolinfo *wol)
7332{
7333 struct tg3 *tp = netdev_priv(dev);
7334
7335 wol->supported = WAKE_MAGIC;
7336 wol->wolopts = 0;
7337 if (tp->tg3_flags & TG3_FLAG_WOL_ENABLE)
7338 wol->wolopts = WAKE_MAGIC;
7339 memset(&wol->sopass, 0, sizeof(wol->sopass));
7340}
7341
7342static int tg3_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol)
7343{
7344 struct tg3 *tp = netdev_priv(dev);
7345
7346 if (wol->wolopts & ~WAKE_MAGIC)
7347 return -EINVAL;
7348 if ((wol->wolopts & WAKE_MAGIC) &&
7349 tp->tg3_flags2 & TG3_FLG2_PHY_SERDES &&
7350 !(tp->tg3_flags & TG3_FLAG_SERDES_WOL_CAP))
7351 return -EINVAL;
7352
f47c11ee 7353 spin_lock_bh(&tp->lock);
1da177e4
LT
7354 if (wol->wolopts & WAKE_MAGIC)
7355 tp->tg3_flags |= TG3_FLAG_WOL_ENABLE;
7356 else
7357 tp->tg3_flags &= ~TG3_FLAG_WOL_ENABLE;
f47c11ee 7358 spin_unlock_bh(&tp->lock);
1da177e4
LT
7359
7360 return 0;
7361}
7362
7363static u32 tg3_get_msglevel(struct net_device *dev)
7364{
7365 struct tg3 *tp = netdev_priv(dev);
7366 return tp->msg_enable;
7367}
7368
7369static void tg3_set_msglevel(struct net_device *dev, u32 value)
7370{
7371 struct tg3 *tp = netdev_priv(dev);
7372 tp->msg_enable = value;
7373}
7374
7375#if TG3_TSO_SUPPORT != 0
7376static int tg3_set_tso(struct net_device *dev, u32 value)
7377{
7378 struct tg3 *tp = netdev_priv(dev);
7379
7380 if (!(tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE)) {
7381 if (value)
7382 return -EINVAL;
7383 return 0;
7384 }
7385 return ethtool_op_set_tso(dev, value);
7386}
7387#endif
7388
7389static int tg3_nway_reset(struct net_device *dev)
7390{
7391 struct tg3 *tp = netdev_priv(dev);
7392 u32 bmcr;
7393 int r;
7394
7395 if (!netif_running(dev))
7396 return -EAGAIN;
7397
c94e3941
MC
7398 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)
7399 return -EINVAL;
7400
f47c11ee 7401 spin_lock_bh(&tp->lock);
1da177e4
LT
7402 r = -EINVAL;
7403 tg3_readphy(tp, MII_BMCR, &bmcr);
7404 if (!tg3_readphy(tp, MII_BMCR, &bmcr) &&
c94e3941
MC
7405 ((bmcr & BMCR_ANENABLE) ||
7406 (tp->tg3_flags2 & TG3_FLG2_PARALLEL_DETECT))) {
7407 tg3_writephy(tp, MII_BMCR, bmcr | BMCR_ANRESTART |
7408 BMCR_ANENABLE);
1da177e4
LT
7409 r = 0;
7410 }
f47c11ee 7411 spin_unlock_bh(&tp->lock);
1da177e4
LT
7412
7413 return r;
7414}
7415
7416static void tg3_get_ringparam(struct net_device *dev, struct ethtool_ringparam *ering)
7417{
7418 struct tg3 *tp = netdev_priv(dev);
7419
7420 ering->rx_max_pending = TG3_RX_RING_SIZE - 1;
7421 ering->rx_mini_max_pending = 0;
7422 ering->rx_jumbo_max_pending = TG3_RX_JUMBO_RING_SIZE - 1;
7423
7424 ering->rx_pending = tp->rx_pending;
7425 ering->rx_mini_pending = 0;
7426 ering->rx_jumbo_pending = tp->rx_jumbo_pending;
7427 ering->tx_pending = tp->tx_pending;
7428}
7429
7430static int tg3_set_ringparam(struct net_device *dev, struct ethtool_ringparam *ering)
7431{
7432 struct tg3 *tp = netdev_priv(dev);
bbe832c0 7433 int irq_sync = 0;
1da177e4
LT
7434
7435 if ((ering->rx_pending > TG3_RX_RING_SIZE - 1) ||
7436 (ering->rx_jumbo_pending > TG3_RX_JUMBO_RING_SIZE - 1) ||
7437 (ering->tx_pending > TG3_TX_RING_SIZE - 1))
7438 return -EINVAL;
7439
bbe832c0 7440 if (netif_running(dev)) {
1da177e4 7441 tg3_netif_stop(tp);
bbe832c0
MC
7442 irq_sync = 1;
7443 }
1da177e4 7444
bbe832c0 7445 tg3_full_lock(tp, irq_sync);
1da177e4
LT
7446
7447 tp->rx_pending = ering->rx_pending;
7448
7449 if ((tp->tg3_flags2 & TG3_FLG2_MAX_RXPEND_64) &&
7450 tp->rx_pending > 63)
7451 tp->rx_pending = 63;
7452 tp->rx_jumbo_pending = ering->rx_jumbo_pending;
7453 tp->tx_pending = ering->tx_pending;
7454
7455 if (netif_running(dev)) {
944d980e 7456 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
1da177e4
LT
7457 tg3_init_hw(tp);
7458 tg3_netif_start(tp);
7459 }
7460
f47c11ee 7461 tg3_full_unlock(tp);
1da177e4
LT
7462
7463 return 0;
7464}
7465
7466static void tg3_get_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause)
7467{
7468 struct tg3 *tp = netdev_priv(dev);
7469
7470 epause->autoneg = (tp->tg3_flags & TG3_FLAG_PAUSE_AUTONEG) != 0;
7471 epause->rx_pause = (tp->tg3_flags & TG3_FLAG_RX_PAUSE) != 0;
7472 epause->tx_pause = (tp->tg3_flags & TG3_FLAG_TX_PAUSE) != 0;
7473}
7474
7475static int tg3_set_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause)
7476{
7477 struct tg3 *tp = netdev_priv(dev);
bbe832c0 7478 int irq_sync = 0;
1da177e4 7479
bbe832c0 7480 if (netif_running(dev)) {
1da177e4 7481 tg3_netif_stop(tp);
bbe832c0
MC
7482 irq_sync = 1;
7483 }
1da177e4 7484
bbe832c0 7485 tg3_full_lock(tp, irq_sync);
f47c11ee 7486
1da177e4
LT
7487 if (epause->autoneg)
7488 tp->tg3_flags |= TG3_FLAG_PAUSE_AUTONEG;
7489 else
7490 tp->tg3_flags &= ~TG3_FLAG_PAUSE_AUTONEG;
7491 if (epause->rx_pause)
7492 tp->tg3_flags |= TG3_FLAG_RX_PAUSE;
7493 else
7494 tp->tg3_flags &= ~TG3_FLAG_RX_PAUSE;
7495 if (epause->tx_pause)
7496 tp->tg3_flags |= TG3_FLAG_TX_PAUSE;
7497 else
7498 tp->tg3_flags &= ~TG3_FLAG_TX_PAUSE;
7499
7500 if (netif_running(dev)) {
944d980e 7501 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
1da177e4
LT
7502 tg3_init_hw(tp);
7503 tg3_netif_start(tp);
7504 }
f47c11ee
DM
7505
7506 tg3_full_unlock(tp);
1da177e4
LT
7507
7508 return 0;
7509}
7510
7511static u32 tg3_get_rx_csum(struct net_device *dev)
7512{
7513 struct tg3 *tp = netdev_priv(dev);
7514 return (tp->tg3_flags & TG3_FLAG_RX_CHECKSUMS) != 0;
7515}
7516
7517static int tg3_set_rx_csum(struct net_device *dev, u32 data)
7518{
7519 struct tg3 *tp = netdev_priv(dev);
7520
7521 if (tp->tg3_flags & TG3_FLAG_BROKEN_CHECKSUMS) {
7522 if (data != 0)
7523 return -EINVAL;
7524 return 0;
7525 }
7526
f47c11ee 7527 spin_lock_bh(&tp->lock);
1da177e4
LT
7528 if (data)
7529 tp->tg3_flags |= TG3_FLAG_RX_CHECKSUMS;
7530 else
7531 tp->tg3_flags &= ~TG3_FLAG_RX_CHECKSUMS;
f47c11ee 7532 spin_unlock_bh(&tp->lock);
1da177e4
LT
7533
7534 return 0;
7535}
7536
7537static int tg3_set_tx_csum(struct net_device *dev, u32 data)
7538{
7539 struct tg3 *tp = netdev_priv(dev);
7540
7541 if (tp->tg3_flags & TG3_FLAG_BROKEN_CHECKSUMS) {
7542 if (data != 0)
7543 return -EINVAL;
7544 return 0;
7545 }
7546
7547 if (data)
7548 dev->features |= NETIF_F_IP_CSUM;
7549 else
7550 dev->features &= ~NETIF_F_IP_CSUM;
7551
7552 return 0;
7553}
7554
7555static int tg3_get_stats_count (struct net_device *dev)
7556{
7557 return TG3_NUM_STATS;
7558}
7559
4cafd3f5
MC
7560static int tg3_get_test_count (struct net_device *dev)
7561{
7562 return TG3_NUM_TEST;
7563}
7564
1da177e4
LT
7565static void tg3_get_strings (struct net_device *dev, u32 stringset, u8 *buf)
7566{
7567 switch (stringset) {
7568 case ETH_SS_STATS:
7569 memcpy(buf, &ethtool_stats_keys, sizeof(ethtool_stats_keys));
7570 break;
4cafd3f5
MC
7571 case ETH_SS_TEST:
7572 memcpy(buf, &ethtool_test_keys, sizeof(ethtool_test_keys));
7573 break;
1da177e4
LT
7574 default:
7575 WARN_ON(1); /* we need a WARN() */
7576 break;
7577 }
7578}
7579
4009a93d
MC
7580static int tg3_phys_id(struct net_device *dev, u32 data)
7581{
7582 struct tg3 *tp = netdev_priv(dev);
7583 int i;
7584
7585 if (!netif_running(tp->dev))
7586 return -EAGAIN;
7587
7588 if (data == 0)
7589 data = 2;
7590
7591 for (i = 0; i < (data * 2); i++) {
7592 if ((i % 2) == 0)
7593 tw32(MAC_LED_CTRL, LED_CTRL_LNKLED_OVERRIDE |
7594 LED_CTRL_1000MBPS_ON |
7595 LED_CTRL_100MBPS_ON |
7596 LED_CTRL_10MBPS_ON |
7597 LED_CTRL_TRAFFIC_OVERRIDE |
7598 LED_CTRL_TRAFFIC_BLINK |
7599 LED_CTRL_TRAFFIC_LED);
7600
7601 else
7602 tw32(MAC_LED_CTRL, LED_CTRL_LNKLED_OVERRIDE |
7603 LED_CTRL_TRAFFIC_OVERRIDE);
7604
7605 if (msleep_interruptible(500))
7606 break;
7607 }
7608 tw32(MAC_LED_CTRL, tp->led_ctrl);
7609 return 0;
7610}
7611
1da177e4
LT
7612static void tg3_get_ethtool_stats (struct net_device *dev,
7613 struct ethtool_stats *estats, u64 *tmp_stats)
7614{
7615 struct tg3 *tp = netdev_priv(dev);
7616 memcpy(tmp_stats, tg3_get_estats(tp), sizeof(tp->estats));
7617}
7618
566f86ad
MC
7619#define NVRAM_TEST_SIZE 0x100
7620
7621static int tg3_test_nvram(struct tg3 *tp)
7622{
7623 u32 *buf, csum;
7624 int i, j, err = 0;
7625
7626 buf = kmalloc(NVRAM_TEST_SIZE, GFP_KERNEL);
7627 if (buf == NULL)
7628 return -ENOMEM;
7629
7630 for (i = 0, j = 0; i < NVRAM_TEST_SIZE; i += 4, j++) {
7631 u32 val;
7632
7633 if ((err = tg3_nvram_read(tp, i, &val)) != 0)
7634 break;
7635 buf[j] = cpu_to_le32(val);
7636 }
7637 if (i < NVRAM_TEST_SIZE)
7638 goto out;
7639
7640 err = -EIO;
7641 if (cpu_to_be32(buf[0]) != TG3_EEPROM_MAGIC)
7642 goto out;
7643
7644 /* Bootstrap checksum at offset 0x10 */
7645 csum = calc_crc((unsigned char *) buf, 0x10);
7646 if(csum != cpu_to_le32(buf[0x10/4]))
7647 goto out;
7648
7649 /* Manufacturing block starts at offset 0x74, checksum at 0xfc */
7650 csum = calc_crc((unsigned char *) &buf[0x74/4], 0x88);
7651 if (csum != cpu_to_le32(buf[0xfc/4]))
7652 goto out;
7653
7654 err = 0;
7655
7656out:
7657 kfree(buf);
7658 return err;
7659}
7660
ca43007a
MC
7661#define TG3_SERDES_TIMEOUT_SEC 2
7662#define TG3_COPPER_TIMEOUT_SEC 6
7663
7664static int tg3_test_link(struct tg3 *tp)
7665{
7666 int i, max;
7667
7668 if (!netif_running(tp->dev))
7669 return -ENODEV;
7670
4c987487 7671 if (tp->tg3_flags2 & TG3_FLG2_ANY_SERDES)
ca43007a
MC
7672 max = TG3_SERDES_TIMEOUT_SEC;
7673 else
7674 max = TG3_COPPER_TIMEOUT_SEC;
7675
7676 for (i = 0; i < max; i++) {
7677 if (netif_carrier_ok(tp->dev))
7678 return 0;
7679
7680 if (msleep_interruptible(1000))
7681 break;
7682 }
7683
7684 return -EIO;
7685}
7686
a71116d1
MC
7687/* Only test the commonly used registers */
7688static int tg3_test_registers(struct tg3 *tp)
7689{
7690 int i, is_5705;
7691 u32 offset, read_mask, write_mask, val, save_val, read_val;
7692 static struct {
7693 u16 offset;
7694 u16 flags;
7695#define TG3_FL_5705 0x1
7696#define TG3_FL_NOT_5705 0x2
7697#define TG3_FL_NOT_5788 0x4
7698 u32 read_mask;
7699 u32 write_mask;
7700 } reg_tbl[] = {
7701 /* MAC Control Registers */
7702 { MAC_MODE, TG3_FL_NOT_5705,
7703 0x00000000, 0x00ef6f8c },
7704 { MAC_MODE, TG3_FL_5705,
7705 0x00000000, 0x01ef6b8c },
7706 { MAC_STATUS, TG3_FL_NOT_5705,
7707 0x03800107, 0x00000000 },
7708 { MAC_STATUS, TG3_FL_5705,
7709 0x03800100, 0x00000000 },
7710 { MAC_ADDR_0_HIGH, 0x0000,
7711 0x00000000, 0x0000ffff },
7712 { MAC_ADDR_0_LOW, 0x0000,
7713 0x00000000, 0xffffffff },
7714 { MAC_RX_MTU_SIZE, 0x0000,
7715 0x00000000, 0x0000ffff },
7716 { MAC_TX_MODE, 0x0000,
7717 0x00000000, 0x00000070 },
7718 { MAC_TX_LENGTHS, 0x0000,
7719 0x00000000, 0x00003fff },
7720 { MAC_RX_MODE, TG3_FL_NOT_5705,
7721 0x00000000, 0x000007fc },
7722 { MAC_RX_MODE, TG3_FL_5705,
7723 0x00000000, 0x000007dc },
7724 { MAC_HASH_REG_0, 0x0000,
7725 0x00000000, 0xffffffff },
7726 { MAC_HASH_REG_1, 0x0000,
7727 0x00000000, 0xffffffff },
7728 { MAC_HASH_REG_2, 0x0000,
7729 0x00000000, 0xffffffff },
7730 { MAC_HASH_REG_3, 0x0000,
7731 0x00000000, 0xffffffff },
7732
7733 /* Receive Data and Receive BD Initiator Control Registers. */
7734 { RCVDBDI_JUMBO_BD+0, TG3_FL_NOT_5705,
7735 0x00000000, 0xffffffff },
7736 { RCVDBDI_JUMBO_BD+4, TG3_FL_NOT_5705,
7737 0x00000000, 0xffffffff },
7738 { RCVDBDI_JUMBO_BD+8, TG3_FL_NOT_5705,
7739 0x00000000, 0x00000003 },
7740 { RCVDBDI_JUMBO_BD+0xc, TG3_FL_NOT_5705,
7741 0x00000000, 0xffffffff },
7742 { RCVDBDI_STD_BD+0, 0x0000,
7743 0x00000000, 0xffffffff },
7744 { RCVDBDI_STD_BD+4, 0x0000,
7745 0x00000000, 0xffffffff },
7746 { RCVDBDI_STD_BD+8, 0x0000,
7747 0x00000000, 0xffff0002 },
7748 { RCVDBDI_STD_BD+0xc, 0x0000,
7749 0x00000000, 0xffffffff },
7750
7751 /* Receive BD Initiator Control Registers. */
7752 { RCVBDI_STD_THRESH, TG3_FL_NOT_5705,
7753 0x00000000, 0xffffffff },
7754 { RCVBDI_STD_THRESH, TG3_FL_5705,
7755 0x00000000, 0x000003ff },
7756 { RCVBDI_JUMBO_THRESH, TG3_FL_NOT_5705,
7757 0x00000000, 0xffffffff },
7758
7759 /* Host Coalescing Control Registers. */
7760 { HOSTCC_MODE, TG3_FL_NOT_5705,
7761 0x00000000, 0x00000004 },
7762 { HOSTCC_MODE, TG3_FL_5705,
7763 0x00000000, 0x000000f6 },
7764 { HOSTCC_RXCOL_TICKS, TG3_FL_NOT_5705,
7765 0x00000000, 0xffffffff },
7766 { HOSTCC_RXCOL_TICKS, TG3_FL_5705,
7767 0x00000000, 0x000003ff },
7768 { HOSTCC_TXCOL_TICKS, TG3_FL_NOT_5705,
7769 0x00000000, 0xffffffff },
7770 { HOSTCC_TXCOL_TICKS, TG3_FL_5705,
7771 0x00000000, 0x000003ff },
7772 { HOSTCC_RXMAX_FRAMES, TG3_FL_NOT_5705,
7773 0x00000000, 0xffffffff },
7774 { HOSTCC_RXMAX_FRAMES, TG3_FL_5705 | TG3_FL_NOT_5788,
7775 0x00000000, 0x000000ff },
7776 { HOSTCC_TXMAX_FRAMES, TG3_FL_NOT_5705,
7777 0x00000000, 0xffffffff },
7778 { HOSTCC_TXMAX_FRAMES, TG3_FL_5705 | TG3_FL_NOT_5788,
7779 0x00000000, 0x000000ff },
7780 { HOSTCC_RXCOAL_TICK_INT, TG3_FL_NOT_5705,
7781 0x00000000, 0xffffffff },
7782 { HOSTCC_TXCOAL_TICK_INT, TG3_FL_NOT_5705,
7783 0x00000000, 0xffffffff },
7784 { HOSTCC_RXCOAL_MAXF_INT, TG3_FL_NOT_5705,
7785 0x00000000, 0xffffffff },
7786 { HOSTCC_RXCOAL_MAXF_INT, TG3_FL_5705 | TG3_FL_NOT_5788,
7787 0x00000000, 0x000000ff },
7788 { HOSTCC_TXCOAL_MAXF_INT, TG3_FL_NOT_5705,
7789 0x00000000, 0xffffffff },
7790 { HOSTCC_TXCOAL_MAXF_INT, TG3_FL_5705 | TG3_FL_NOT_5788,
7791 0x00000000, 0x000000ff },
7792 { HOSTCC_STAT_COAL_TICKS, TG3_FL_NOT_5705,
7793 0x00000000, 0xffffffff },
7794 { HOSTCC_STATS_BLK_HOST_ADDR, TG3_FL_NOT_5705,
7795 0x00000000, 0xffffffff },
7796 { HOSTCC_STATS_BLK_HOST_ADDR+4, TG3_FL_NOT_5705,
7797 0x00000000, 0xffffffff },
7798 { HOSTCC_STATUS_BLK_HOST_ADDR, 0x0000,
7799 0x00000000, 0xffffffff },
7800 { HOSTCC_STATUS_BLK_HOST_ADDR+4, 0x0000,
7801 0x00000000, 0xffffffff },
7802 { HOSTCC_STATS_BLK_NIC_ADDR, 0x0000,
7803 0xffffffff, 0x00000000 },
7804 { HOSTCC_STATUS_BLK_NIC_ADDR, 0x0000,
7805 0xffffffff, 0x00000000 },
7806
7807 /* Buffer Manager Control Registers. */
7808 { BUFMGR_MB_POOL_ADDR, 0x0000,
7809 0x00000000, 0x007fff80 },
7810 { BUFMGR_MB_POOL_SIZE, 0x0000,
7811 0x00000000, 0x007fffff },
7812 { BUFMGR_MB_RDMA_LOW_WATER, 0x0000,
7813 0x00000000, 0x0000003f },
7814 { BUFMGR_MB_MACRX_LOW_WATER, 0x0000,
7815 0x00000000, 0x000001ff },
7816 { BUFMGR_MB_HIGH_WATER, 0x0000,
7817 0x00000000, 0x000001ff },
7818 { BUFMGR_DMA_DESC_POOL_ADDR, TG3_FL_NOT_5705,
7819 0xffffffff, 0x00000000 },
7820 { BUFMGR_DMA_DESC_POOL_SIZE, TG3_FL_NOT_5705,
7821 0xffffffff, 0x00000000 },
7822
7823 /* Mailbox Registers */
7824 { GRCMBOX_RCVSTD_PROD_IDX+4, 0x0000,
7825 0x00000000, 0x000001ff },
7826 { GRCMBOX_RCVJUMBO_PROD_IDX+4, TG3_FL_NOT_5705,
7827 0x00000000, 0x000001ff },
7828 { GRCMBOX_RCVRET_CON_IDX_0+4, 0x0000,
7829 0x00000000, 0x000007ff },
7830 { GRCMBOX_SNDHOST_PROD_IDX_0+4, 0x0000,
7831 0x00000000, 0x000001ff },
7832
7833 { 0xffff, 0x0000, 0x00000000, 0x00000000 },
7834 };
7835
7836 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)
7837 is_5705 = 1;
7838 else
7839 is_5705 = 0;
7840
7841 for (i = 0; reg_tbl[i].offset != 0xffff; i++) {
7842 if (is_5705 && (reg_tbl[i].flags & TG3_FL_NOT_5705))
7843 continue;
7844
7845 if (!is_5705 && (reg_tbl[i].flags & TG3_FL_5705))
7846 continue;
7847
7848 if ((tp->tg3_flags2 & TG3_FLG2_IS_5788) &&
7849 (reg_tbl[i].flags & TG3_FL_NOT_5788))
7850 continue;
7851
7852 offset = (u32) reg_tbl[i].offset;
7853 read_mask = reg_tbl[i].read_mask;
7854 write_mask = reg_tbl[i].write_mask;
7855
7856 /* Save the original register content */
7857 save_val = tr32(offset);
7858
7859 /* Determine the read-only value. */
7860 read_val = save_val & read_mask;
7861
7862 /* Write zero to the register, then make sure the read-only bits
7863 * are not changed and the read/write bits are all zeros.
7864 */
7865 tw32(offset, 0);
7866
7867 val = tr32(offset);
7868
7869 /* Test the read-only and read/write bits. */
7870 if (((val & read_mask) != read_val) || (val & write_mask))
7871 goto out;
7872
7873 /* Write ones to all the bits defined by RdMask and WrMask, then
7874 * make sure the read-only bits are not changed and the
7875 * read/write bits are all ones.
7876 */
7877 tw32(offset, read_mask | write_mask);
7878
7879 val = tr32(offset);
7880
7881 /* Test the read-only bits. */
7882 if ((val & read_mask) != read_val)
7883 goto out;
7884
7885 /* Test the read/write bits. */
7886 if ((val & write_mask) != write_mask)
7887 goto out;
7888
7889 tw32(offset, save_val);
7890 }
7891
7892 return 0;
7893
7894out:
7895 printk(KERN_ERR PFX "Register test failed at offset %x\n", offset);
7896 tw32(offset, save_val);
7897 return -EIO;
7898}
7899
7942e1db
MC
7900static int tg3_do_mem_test(struct tg3 *tp, u32 offset, u32 len)
7901{
7902 static u32 test_pattern[] = { 0x00000000, 0xffffffff, 0xaa55a55a };
7903 int i;
7904 u32 j;
7905
7906 for (i = 0; i < sizeof(test_pattern)/sizeof(u32); i++) {
7907 for (j = 0; j < len; j += 4) {
7908 u32 val;
7909
7910 tg3_write_mem(tp, offset + j, test_pattern[i]);
7911 tg3_read_mem(tp, offset + j, &val);
7912 if (val != test_pattern[i])
7913 return -EIO;
7914 }
7915 }
7916 return 0;
7917}
7918
7919static int tg3_test_memory(struct tg3 *tp)
7920{
7921 static struct mem_entry {
7922 u32 offset;
7923 u32 len;
7924 } mem_tbl_570x[] = {
7925 { 0x00000000, 0x01000},
7926 { 0x00002000, 0x1c000},
7927 { 0xffffffff, 0x00000}
7928 }, mem_tbl_5705[] = {
7929 { 0x00000100, 0x0000c},
7930 { 0x00000200, 0x00008},
7931 { 0x00000b50, 0x00400},
7932 { 0x00004000, 0x00800},
7933 { 0x00006000, 0x01000},
7934 { 0x00008000, 0x02000},
7935 { 0x00010000, 0x0e000},
7936 { 0xffffffff, 0x00000}
7937 };
7938 struct mem_entry *mem_tbl;
7939 int err = 0;
7940 int i;
7941
7942 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)
7943 mem_tbl = mem_tbl_5705;
7944 else
7945 mem_tbl = mem_tbl_570x;
7946
7947 for (i = 0; mem_tbl[i].offset != 0xffffffff; i++) {
7948 if ((err = tg3_do_mem_test(tp, mem_tbl[i].offset,
7949 mem_tbl[i].len)) != 0)
7950 break;
7951 }
7952
7953 return err;
7954}
7955
9f40dead
MC
7956#define TG3_MAC_LOOPBACK 0
7957#define TG3_PHY_LOOPBACK 1
7958
7959static int tg3_run_loopback(struct tg3 *tp, int loopback_mode)
c76949a6 7960{
9f40dead 7961 u32 mac_mode, rx_start_idx, rx_idx, tx_idx, opaque_key;
c76949a6
MC
7962 u32 desc_idx;
7963 struct sk_buff *skb, *rx_skb;
7964 u8 *tx_data;
7965 dma_addr_t map;
7966 int num_pkts, tx_len, rx_len, i, err;
7967 struct tg3_rx_buffer_desc *desc;
7968
9f40dead 7969 if (loopback_mode == TG3_MAC_LOOPBACK) {
c94e3941
MC
7970 /* HW errata - mac loopback fails in some cases on 5780.
7971 * Normal traffic and PHY loopback are not affected by
7972 * errata.
7973 */
7974 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5780)
7975 return 0;
7976
9f40dead
MC
7977 mac_mode = (tp->mac_mode & ~MAC_MODE_PORT_MODE_MASK) |
7978 MAC_MODE_PORT_INT_LPBACK | MAC_MODE_LINK_POLARITY |
7979 MAC_MODE_PORT_MODE_GMII;
7980 tw32(MAC_MODE, mac_mode);
7981 } else if (loopback_mode == TG3_PHY_LOOPBACK) {
c94e3941
MC
7982 tg3_writephy(tp, MII_BMCR, BMCR_LOOPBACK | BMCR_FULLDPLX |
7983 BMCR_SPEED1000);
7984 udelay(40);
7985 /* reset to prevent losing 1st rx packet intermittently */
7986 if (tp->tg3_flags2 & TG3_FLG2_MII_SERDES) {
7987 tw32_f(MAC_RX_MODE, RX_MODE_RESET);
7988 udelay(10);
7989 tw32_f(MAC_RX_MODE, tp->rx_mode);
7990 }
9f40dead
MC
7991 mac_mode = (tp->mac_mode & ~MAC_MODE_PORT_MODE_MASK) |
7992 MAC_MODE_LINK_POLARITY | MAC_MODE_PORT_MODE_GMII;
7993 if ((tp->phy_id & PHY_ID_MASK) == PHY_ID_BCM5401)
7994 mac_mode &= ~MAC_MODE_LINK_POLARITY;
7995 tw32(MAC_MODE, mac_mode);
9f40dead
MC
7996 }
7997 else
7998 return -EINVAL;
c76949a6
MC
7999
8000 err = -EIO;
8001
c76949a6
MC
8002 tx_len = 1514;
8003 skb = dev_alloc_skb(tx_len);
8004 tx_data = skb_put(skb, tx_len);
8005 memcpy(tx_data, tp->dev->dev_addr, 6);
8006 memset(tx_data + 6, 0x0, 8);
8007
8008 tw32(MAC_RX_MTU_SIZE, tx_len + 4);
8009
8010 for (i = 14; i < tx_len; i++)
8011 tx_data[i] = (u8) (i & 0xff);
8012
8013 map = pci_map_single(tp->pdev, skb->data, tx_len, PCI_DMA_TODEVICE);
8014
8015 tw32_f(HOSTCC_MODE, tp->coalesce_mode | HOSTCC_MODE_ENABLE |
8016 HOSTCC_MODE_NOW);
8017
8018 udelay(10);
8019
8020 rx_start_idx = tp->hw_status->idx[0].rx_producer;
8021
c76949a6
MC
8022 num_pkts = 0;
8023
9f40dead 8024 tg3_set_txd(tp, tp->tx_prod, map, tx_len, 0, 1);
c76949a6 8025
9f40dead 8026 tp->tx_prod++;
c76949a6
MC
8027 num_pkts++;
8028
9f40dead
MC
8029 tw32_tx_mbox(MAILBOX_SNDHOST_PROD_IDX_0 + TG3_64BIT_REG_LOW,
8030 tp->tx_prod);
09ee929c 8031 tr32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0 + TG3_64BIT_REG_LOW);
c76949a6
MC
8032
8033 udelay(10);
8034
8035 for (i = 0; i < 10; i++) {
8036 tw32_f(HOSTCC_MODE, tp->coalesce_mode | HOSTCC_MODE_ENABLE |
8037 HOSTCC_MODE_NOW);
8038
8039 udelay(10);
8040
8041 tx_idx = tp->hw_status->idx[0].tx_consumer;
8042 rx_idx = tp->hw_status->idx[0].rx_producer;
9f40dead 8043 if ((tx_idx == tp->tx_prod) &&
c76949a6
MC
8044 (rx_idx == (rx_start_idx + num_pkts)))
8045 break;
8046 }
8047
8048 pci_unmap_single(tp->pdev, map, tx_len, PCI_DMA_TODEVICE);
8049 dev_kfree_skb(skb);
8050
9f40dead 8051 if (tx_idx != tp->tx_prod)
c76949a6
MC
8052 goto out;
8053
8054 if (rx_idx != rx_start_idx + num_pkts)
8055 goto out;
8056
8057 desc = &tp->rx_rcb[rx_start_idx];
8058 desc_idx = desc->opaque & RXD_OPAQUE_INDEX_MASK;
8059 opaque_key = desc->opaque & RXD_OPAQUE_RING_MASK;
8060 if (opaque_key != RXD_OPAQUE_RING_STD)
8061 goto out;
8062
8063 if ((desc->err_vlan & RXD_ERR_MASK) != 0 &&
8064 (desc->err_vlan != RXD_ERR_ODD_NIBBLE_RCVD_MII))
8065 goto out;
8066
8067 rx_len = ((desc->idx_len & RXD_LEN_MASK) >> RXD_LEN_SHIFT) - 4;
8068 if (rx_len != tx_len)
8069 goto out;
8070
8071 rx_skb = tp->rx_std_buffers[desc_idx].skb;
8072
8073 map = pci_unmap_addr(&tp->rx_std_buffers[desc_idx], mapping);
8074 pci_dma_sync_single_for_cpu(tp->pdev, map, rx_len, PCI_DMA_FROMDEVICE);
8075
8076 for (i = 14; i < tx_len; i++) {
8077 if (*(rx_skb->data + i) != (u8) (i & 0xff))
8078 goto out;
8079 }
8080 err = 0;
8081
8082 /* tg3_free_rings will unmap and free the rx_skb */
8083out:
8084 return err;
8085}
8086
9f40dead
MC
8087#define TG3_MAC_LOOPBACK_FAILED 1
8088#define TG3_PHY_LOOPBACK_FAILED 2
8089#define TG3_LOOPBACK_FAILED (TG3_MAC_LOOPBACK_FAILED | \
8090 TG3_PHY_LOOPBACK_FAILED)
8091
8092static int tg3_test_loopback(struct tg3 *tp)
8093{
8094 int err = 0;
8095
8096 if (!netif_running(tp->dev))
8097 return TG3_LOOPBACK_FAILED;
8098
8099 tg3_reset_hw(tp);
8100
8101 if (tg3_run_loopback(tp, TG3_MAC_LOOPBACK))
8102 err |= TG3_MAC_LOOPBACK_FAILED;
8103 if (!(tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)) {
8104 if (tg3_run_loopback(tp, TG3_PHY_LOOPBACK))
8105 err |= TG3_PHY_LOOPBACK_FAILED;
8106 }
8107
8108 return err;
8109}
8110
4cafd3f5
MC
8111static void tg3_self_test(struct net_device *dev, struct ethtool_test *etest,
8112 u64 *data)
8113{
566f86ad
MC
8114 struct tg3 *tp = netdev_priv(dev);
8115
8116 memset(data, 0, sizeof(u64) * TG3_NUM_TEST);
8117
8118 if (tg3_test_nvram(tp) != 0) {
8119 etest->flags |= ETH_TEST_FL_FAILED;
8120 data[0] = 1;
8121 }
ca43007a
MC
8122 if (tg3_test_link(tp) != 0) {
8123 etest->flags |= ETH_TEST_FL_FAILED;
8124 data[1] = 1;
8125 }
a71116d1 8126 if (etest->flags & ETH_TEST_FL_OFFLINE) {
bbe832c0
MC
8127 int irq_sync = 0;
8128
8129 if (netif_running(dev)) {
a71116d1 8130 tg3_netif_stop(tp);
bbe832c0
MC
8131 irq_sync = 1;
8132 }
a71116d1 8133
bbe832c0 8134 tg3_full_lock(tp, irq_sync);
a71116d1
MC
8135
8136 tg3_halt(tp, RESET_KIND_SUSPEND, 1);
8137 tg3_nvram_lock(tp);
8138 tg3_halt_cpu(tp, RX_CPU_BASE);
8139 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS))
8140 tg3_halt_cpu(tp, TX_CPU_BASE);
8141 tg3_nvram_unlock(tp);
8142
8143 if (tg3_test_registers(tp) != 0) {
8144 etest->flags |= ETH_TEST_FL_FAILED;
8145 data[2] = 1;
8146 }
7942e1db
MC
8147 if (tg3_test_memory(tp) != 0) {
8148 etest->flags |= ETH_TEST_FL_FAILED;
8149 data[3] = 1;
8150 }
9f40dead 8151 if ((data[4] = tg3_test_loopback(tp)) != 0)
c76949a6 8152 etest->flags |= ETH_TEST_FL_FAILED;
a71116d1 8153
f47c11ee
DM
8154 tg3_full_unlock(tp);
8155
d4bc3927
MC
8156 if (tg3_test_interrupt(tp) != 0) {
8157 etest->flags |= ETH_TEST_FL_FAILED;
8158 data[5] = 1;
8159 }
f47c11ee
DM
8160
8161 tg3_full_lock(tp, 0);
d4bc3927 8162
a71116d1
MC
8163 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
8164 if (netif_running(dev)) {
8165 tp->tg3_flags |= TG3_FLAG_INIT_COMPLETE;
8166 tg3_init_hw(tp);
8167 tg3_netif_start(tp);
8168 }
f47c11ee
DM
8169
8170 tg3_full_unlock(tp);
a71116d1 8171 }
4cafd3f5
MC
8172}
8173
1da177e4
LT
8174static int tg3_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
8175{
8176 struct mii_ioctl_data *data = if_mii(ifr);
8177 struct tg3 *tp = netdev_priv(dev);
8178 int err;
8179
8180 switch(cmd) {
8181 case SIOCGMIIPHY:
8182 data->phy_id = PHY_ADDR;
8183
8184 /* fallthru */
8185 case SIOCGMIIREG: {
8186 u32 mii_regval;
8187
8188 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)
8189 break; /* We have no PHY */
8190
f47c11ee 8191 spin_lock_bh(&tp->lock);
1da177e4 8192 err = tg3_readphy(tp, data->reg_num & 0x1f, &mii_regval);
f47c11ee 8193 spin_unlock_bh(&tp->lock);
1da177e4
LT
8194
8195 data->val_out = mii_regval;
8196
8197 return err;
8198 }
8199
8200 case SIOCSMIIREG:
8201 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)
8202 break; /* We have no PHY */
8203
8204 if (!capable(CAP_NET_ADMIN))
8205 return -EPERM;
8206
f47c11ee 8207 spin_lock_bh(&tp->lock);
1da177e4 8208 err = tg3_writephy(tp, data->reg_num & 0x1f, data->val_in);
f47c11ee 8209 spin_unlock_bh(&tp->lock);
1da177e4
LT
8210
8211 return err;
8212
8213 default:
8214 /* do nothing */
8215 break;
8216 }
8217 return -EOPNOTSUPP;
8218}
8219
8220#if TG3_VLAN_TAG_USED
8221static void tg3_vlan_rx_register(struct net_device *dev, struct vlan_group *grp)
8222{
8223 struct tg3 *tp = netdev_priv(dev);
8224
f47c11ee 8225 tg3_full_lock(tp, 0);
1da177e4
LT
8226
8227 tp->vlgrp = grp;
8228
8229 /* Update RX_MODE_KEEP_VLAN_TAG bit in RX_MODE register. */
8230 __tg3_set_rx_mode(dev);
8231
f47c11ee 8232 tg3_full_unlock(tp);
1da177e4
LT
8233}
8234
8235static void tg3_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid)
8236{
8237 struct tg3 *tp = netdev_priv(dev);
8238
f47c11ee 8239 tg3_full_lock(tp, 0);
1da177e4
LT
8240 if (tp->vlgrp)
8241 tp->vlgrp->vlan_devices[vid] = NULL;
f47c11ee 8242 tg3_full_unlock(tp);
1da177e4
LT
8243}
8244#endif
8245
15f9850d
DM
8246static int tg3_get_coalesce(struct net_device *dev, struct ethtool_coalesce *ec)
8247{
8248 struct tg3 *tp = netdev_priv(dev);
8249
8250 memcpy(ec, &tp->coal, sizeof(*ec));
8251 return 0;
8252}
8253
d244c892
MC
8254static int tg3_set_coalesce(struct net_device *dev, struct ethtool_coalesce *ec)
8255{
8256 struct tg3 *tp = netdev_priv(dev);
8257 u32 max_rxcoal_tick_int = 0, max_txcoal_tick_int = 0;
8258 u32 max_stat_coal_ticks = 0, min_stat_coal_ticks = 0;
8259
8260 if (!(tp->tg3_flags2 & TG3_FLG2_5705_PLUS)) {
8261 max_rxcoal_tick_int = MAX_RXCOAL_TICK_INT;
8262 max_txcoal_tick_int = MAX_TXCOAL_TICK_INT;
8263 max_stat_coal_ticks = MAX_STAT_COAL_TICKS;
8264 min_stat_coal_ticks = MIN_STAT_COAL_TICKS;
8265 }
8266
8267 if ((ec->rx_coalesce_usecs > MAX_RXCOL_TICKS) ||
8268 (ec->tx_coalesce_usecs > MAX_TXCOL_TICKS) ||
8269 (ec->rx_max_coalesced_frames > MAX_RXMAX_FRAMES) ||
8270 (ec->tx_max_coalesced_frames > MAX_TXMAX_FRAMES) ||
8271 (ec->rx_coalesce_usecs_irq > max_rxcoal_tick_int) ||
8272 (ec->tx_coalesce_usecs_irq > max_txcoal_tick_int) ||
8273 (ec->rx_max_coalesced_frames_irq > MAX_RXCOAL_MAXF_INT) ||
8274 (ec->tx_max_coalesced_frames_irq > MAX_TXCOAL_MAXF_INT) ||
8275 (ec->stats_block_coalesce_usecs > max_stat_coal_ticks) ||
8276 (ec->stats_block_coalesce_usecs < min_stat_coal_ticks))
8277 return -EINVAL;
8278
8279 /* No rx interrupts will be generated if both are zero */
8280 if ((ec->rx_coalesce_usecs == 0) &&
8281 (ec->rx_max_coalesced_frames == 0))
8282 return -EINVAL;
8283
8284 /* No tx interrupts will be generated if both are zero */
8285 if ((ec->tx_coalesce_usecs == 0) &&
8286 (ec->tx_max_coalesced_frames == 0))
8287 return -EINVAL;
8288
8289 /* Only copy relevant parameters, ignore all others. */
8290 tp->coal.rx_coalesce_usecs = ec->rx_coalesce_usecs;
8291 tp->coal.tx_coalesce_usecs = ec->tx_coalesce_usecs;
8292 tp->coal.rx_max_coalesced_frames = ec->rx_max_coalesced_frames;
8293 tp->coal.tx_max_coalesced_frames = ec->tx_max_coalesced_frames;
8294 tp->coal.rx_coalesce_usecs_irq = ec->rx_coalesce_usecs_irq;
8295 tp->coal.tx_coalesce_usecs_irq = ec->tx_coalesce_usecs_irq;
8296 tp->coal.rx_max_coalesced_frames_irq = ec->rx_max_coalesced_frames_irq;
8297 tp->coal.tx_max_coalesced_frames_irq = ec->tx_max_coalesced_frames_irq;
8298 tp->coal.stats_block_coalesce_usecs = ec->stats_block_coalesce_usecs;
8299
8300 if (netif_running(dev)) {
8301 tg3_full_lock(tp, 0);
8302 __tg3_set_coalesce(tp, &tp->coal);
8303 tg3_full_unlock(tp);
8304 }
8305 return 0;
8306}
8307
1da177e4
LT
8308static struct ethtool_ops tg3_ethtool_ops = {
8309 .get_settings = tg3_get_settings,
8310 .set_settings = tg3_set_settings,
8311 .get_drvinfo = tg3_get_drvinfo,
8312 .get_regs_len = tg3_get_regs_len,
8313 .get_regs = tg3_get_regs,
8314 .get_wol = tg3_get_wol,
8315 .set_wol = tg3_set_wol,
8316 .get_msglevel = tg3_get_msglevel,
8317 .set_msglevel = tg3_set_msglevel,
8318 .nway_reset = tg3_nway_reset,
8319 .get_link = ethtool_op_get_link,
8320 .get_eeprom_len = tg3_get_eeprom_len,
8321 .get_eeprom = tg3_get_eeprom,
8322 .set_eeprom = tg3_set_eeprom,
8323 .get_ringparam = tg3_get_ringparam,
8324 .set_ringparam = tg3_set_ringparam,
8325 .get_pauseparam = tg3_get_pauseparam,
8326 .set_pauseparam = tg3_set_pauseparam,
8327 .get_rx_csum = tg3_get_rx_csum,
8328 .set_rx_csum = tg3_set_rx_csum,
8329 .get_tx_csum = ethtool_op_get_tx_csum,
8330 .set_tx_csum = tg3_set_tx_csum,
8331 .get_sg = ethtool_op_get_sg,
8332 .set_sg = ethtool_op_set_sg,
8333#if TG3_TSO_SUPPORT != 0
8334 .get_tso = ethtool_op_get_tso,
8335 .set_tso = tg3_set_tso,
8336#endif
4cafd3f5
MC
8337 .self_test_count = tg3_get_test_count,
8338 .self_test = tg3_self_test,
1da177e4 8339 .get_strings = tg3_get_strings,
4009a93d 8340 .phys_id = tg3_phys_id,
1da177e4
LT
8341 .get_stats_count = tg3_get_stats_count,
8342 .get_ethtool_stats = tg3_get_ethtool_stats,
15f9850d 8343 .get_coalesce = tg3_get_coalesce,
d244c892 8344 .set_coalesce = tg3_set_coalesce,
2ff43697 8345 .get_perm_addr = ethtool_op_get_perm_addr,
1da177e4
LT
8346};
8347
8348static void __devinit tg3_get_eeprom_size(struct tg3 *tp)
8349{
8350 u32 cursize, val;
8351
8352 tp->nvram_size = EEPROM_CHIP_SIZE;
8353
8354 if (tg3_nvram_read(tp, 0, &val) != 0)
8355 return;
8356
8357 if (swab32(val) != TG3_EEPROM_MAGIC)
8358 return;
8359
8360 /*
8361 * Size the chip by reading offsets at increasing powers of two.
8362 * When we encounter our validation signature, we know the addressing
8363 * has wrapped around, and thus have our chip size.
8364 */
8365 cursize = 0x800;
8366
8367 while (cursize < tp->nvram_size) {
8368 if (tg3_nvram_read(tp, cursize, &val) != 0)
8369 return;
8370
8371 if (swab32(val) == TG3_EEPROM_MAGIC)
8372 break;
8373
8374 cursize <<= 1;
8375 }
8376
8377 tp->nvram_size = cursize;
8378}
8379
8380static void __devinit tg3_get_nvram_size(struct tg3 *tp)
8381{
8382 u32 val;
8383
8384 if (tg3_nvram_read(tp, 0xf0, &val) == 0) {
8385 if (val != 0) {
8386 tp->nvram_size = (val >> 16) * 1024;
8387 return;
8388 }
8389 }
8390 tp->nvram_size = 0x20000;
8391}
8392
8393static void __devinit tg3_get_nvram_info(struct tg3 *tp)
8394{
8395 u32 nvcfg1;
8396
8397 nvcfg1 = tr32(NVRAM_CFG1);
8398 if (nvcfg1 & NVRAM_CFG1_FLASHIF_ENAB) {
8399 tp->tg3_flags2 |= TG3_FLG2_FLASH;
8400 }
8401 else {
8402 nvcfg1 &= ~NVRAM_CFG1_COMPAT_BYPASS;
8403 tw32(NVRAM_CFG1, nvcfg1);
8404 }
8405
4c987487 8406 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5750) ||
a4e2b347 8407 (tp->tg3_flags2 & TG3_FLG2_5780_CLASS)) {
1da177e4
LT
8408 switch (nvcfg1 & NVRAM_CFG1_VENDOR_MASK) {
8409 case FLASH_VENDOR_ATMEL_FLASH_BUFFERED:
8410 tp->nvram_jedecnum = JEDEC_ATMEL;
8411 tp->nvram_pagesize = ATMEL_AT45DB0X1B_PAGE_SIZE;
8412 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8413 break;
8414 case FLASH_VENDOR_ATMEL_FLASH_UNBUFFERED:
8415 tp->nvram_jedecnum = JEDEC_ATMEL;
8416 tp->nvram_pagesize = ATMEL_AT25F512_PAGE_SIZE;
8417 break;
8418 case FLASH_VENDOR_ATMEL_EEPROM:
8419 tp->nvram_jedecnum = JEDEC_ATMEL;
8420 tp->nvram_pagesize = ATMEL_AT24C512_CHIP_SIZE;
8421 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8422 break;
8423 case FLASH_VENDOR_ST:
8424 tp->nvram_jedecnum = JEDEC_ST;
8425 tp->nvram_pagesize = ST_M45PEX0_PAGE_SIZE;
8426 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8427 break;
8428 case FLASH_VENDOR_SAIFUN:
8429 tp->nvram_jedecnum = JEDEC_SAIFUN;
8430 tp->nvram_pagesize = SAIFUN_SA25F0XX_PAGE_SIZE;
8431 break;
8432 case FLASH_VENDOR_SST_SMALL:
8433 case FLASH_VENDOR_SST_LARGE:
8434 tp->nvram_jedecnum = JEDEC_SST;
8435 tp->nvram_pagesize = SST_25VF0X0_PAGE_SIZE;
8436 break;
8437 }
8438 }
8439 else {
8440 tp->nvram_jedecnum = JEDEC_ATMEL;
8441 tp->nvram_pagesize = ATMEL_AT45DB0X1B_PAGE_SIZE;
8442 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8443 }
8444}
8445
361b4ac2
MC
8446static void __devinit tg3_get_5752_nvram_info(struct tg3 *tp)
8447{
8448 u32 nvcfg1;
8449
8450 nvcfg1 = tr32(NVRAM_CFG1);
8451
e6af301b
MC
8452 /* NVRAM protection for TPM */
8453 if (nvcfg1 & (1 << 27))
8454 tp->tg3_flags2 |= TG3_FLG2_PROTECTED_NVRAM;
8455
361b4ac2
MC
8456 switch (nvcfg1 & NVRAM_CFG1_5752VENDOR_MASK) {
8457 case FLASH_5752VENDOR_ATMEL_EEPROM_64KHZ:
8458 case FLASH_5752VENDOR_ATMEL_EEPROM_376KHZ:
8459 tp->nvram_jedecnum = JEDEC_ATMEL;
8460 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8461 break;
8462 case FLASH_5752VENDOR_ATMEL_FLASH_BUFFERED:
8463 tp->nvram_jedecnum = JEDEC_ATMEL;
8464 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8465 tp->tg3_flags2 |= TG3_FLG2_FLASH;
8466 break;
8467 case FLASH_5752VENDOR_ST_M45PE10:
8468 case FLASH_5752VENDOR_ST_M45PE20:
8469 case FLASH_5752VENDOR_ST_M45PE40:
8470 tp->nvram_jedecnum = JEDEC_ST;
8471 tp->tg3_flags |= TG3_FLAG_NVRAM_BUFFERED;
8472 tp->tg3_flags2 |= TG3_FLG2_FLASH;
8473 break;
8474 }
8475
8476 if (tp->tg3_flags2 & TG3_FLG2_FLASH) {
8477 switch (nvcfg1 & NVRAM_CFG1_5752PAGE_SIZE_MASK) {
8478 case FLASH_5752PAGE_SIZE_256:
8479 tp->nvram_pagesize = 256;
8480 break;
8481 case FLASH_5752PAGE_SIZE_512:
8482 tp->nvram_pagesize = 512;
8483 break;
8484 case FLASH_5752PAGE_SIZE_1K:
8485 tp->nvram_pagesize = 1024;
8486 break;
8487 case FLASH_5752PAGE_SIZE_2K:
8488 tp->nvram_pagesize = 2048;
8489 break;
8490 case FLASH_5752PAGE_SIZE_4K:
8491 tp->nvram_pagesize = 4096;
8492 break;
8493 case FLASH_5752PAGE_SIZE_264:
8494 tp->nvram_pagesize = 264;
8495 break;
8496 }
8497 }
8498 else {
8499 /* For eeprom, set pagesize to maximum eeprom size */
8500 tp->nvram_pagesize = ATMEL_AT24C512_CHIP_SIZE;
8501
8502 nvcfg1 &= ~NVRAM_CFG1_COMPAT_BYPASS;
8503 tw32(NVRAM_CFG1, nvcfg1);
8504 }
8505}
8506
1da177e4
LT
8507/* Chips other than 5700/5701 use the NVRAM for fetching info. */
8508static void __devinit tg3_nvram_init(struct tg3 *tp)
8509{
8510 int j;
8511
8512 if (tp->tg3_flags2 & TG3_FLG2_SUN_570X)
8513 return;
8514
8515 tw32_f(GRC_EEPROM_ADDR,
8516 (EEPROM_ADDR_FSM_RESET |
8517 (EEPROM_DEFAULT_CLOCK_PERIOD <<
8518 EEPROM_ADDR_CLKPERD_SHIFT)));
8519
8520 /* XXX schedule_timeout() ... */
8521 for (j = 0; j < 100; j++)
8522 udelay(10);
8523
8524 /* Enable seeprom accesses. */
8525 tw32_f(GRC_LOCAL_CTRL,
8526 tr32(GRC_LOCAL_CTRL) | GRC_LCLCTRL_AUTO_SEEPROM);
8527 udelay(100);
8528
8529 if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700 &&
8530 GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5701) {
8531 tp->tg3_flags |= TG3_FLAG_NVRAM;
8532
e6af301b 8533 tg3_enable_nvram_access(tp);
1da177e4 8534
361b4ac2
MC
8535 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5752)
8536 tg3_get_5752_nvram_info(tp);
8537 else
8538 tg3_get_nvram_info(tp);
8539
1da177e4
LT
8540 tg3_get_nvram_size(tp);
8541
e6af301b 8542 tg3_disable_nvram_access(tp);
1da177e4
LT
8543
8544 } else {
8545 tp->tg3_flags &= ~(TG3_FLAG_NVRAM | TG3_FLAG_NVRAM_BUFFERED);
8546
8547 tg3_get_eeprom_size(tp);
8548 }
8549}
8550
8551static int tg3_nvram_read_using_eeprom(struct tg3 *tp,
8552 u32 offset, u32 *val)
8553{
8554 u32 tmp;
8555 int i;
8556
8557 if (offset > EEPROM_ADDR_ADDR_MASK ||
8558 (offset % 4) != 0)
8559 return -EINVAL;
8560
8561 tmp = tr32(GRC_EEPROM_ADDR) & ~(EEPROM_ADDR_ADDR_MASK |
8562 EEPROM_ADDR_DEVID_MASK |
8563 EEPROM_ADDR_READ);
8564 tw32(GRC_EEPROM_ADDR,
8565 tmp |
8566 (0 << EEPROM_ADDR_DEVID_SHIFT) |
8567 ((offset << EEPROM_ADDR_ADDR_SHIFT) &
8568 EEPROM_ADDR_ADDR_MASK) |
8569 EEPROM_ADDR_READ | EEPROM_ADDR_START);
8570
8571 for (i = 0; i < 10000; i++) {
8572 tmp = tr32(GRC_EEPROM_ADDR);
8573
8574 if (tmp & EEPROM_ADDR_COMPLETE)
8575 break;
8576 udelay(100);
8577 }
8578 if (!(tmp & EEPROM_ADDR_COMPLETE))
8579 return -EBUSY;
8580
8581 *val = tr32(GRC_EEPROM_DATA);
8582 return 0;
8583}
8584
8585#define NVRAM_CMD_TIMEOUT 10000
8586
8587static int tg3_nvram_exec_cmd(struct tg3 *tp, u32 nvram_cmd)
8588{
8589 int i;
8590
8591 tw32(NVRAM_CMD, nvram_cmd);
8592 for (i = 0; i < NVRAM_CMD_TIMEOUT; i++) {
8593 udelay(10);
8594 if (tr32(NVRAM_CMD) & NVRAM_CMD_DONE) {
8595 udelay(10);
8596 break;
8597 }
8598 }
8599 if (i == NVRAM_CMD_TIMEOUT) {
8600 return -EBUSY;
8601 }
8602 return 0;
8603}
8604
8605static int tg3_nvram_read(struct tg3 *tp, u32 offset, u32 *val)
8606{
8607 int ret;
8608
8609 if (tp->tg3_flags2 & TG3_FLG2_SUN_570X) {
8610 printk(KERN_ERR PFX "Attempt to do nvram_read on Sun 570X\n");
8611 return -EINVAL;
8612 }
8613
8614 if (!(tp->tg3_flags & TG3_FLAG_NVRAM))
8615 return tg3_nvram_read_using_eeprom(tp, offset, val);
8616
8617 if ((tp->tg3_flags & TG3_FLAG_NVRAM_BUFFERED) &&
8618 (tp->tg3_flags2 & TG3_FLG2_FLASH) &&
8619 (tp->nvram_jedecnum == JEDEC_ATMEL)) {
8620
8621 offset = ((offset / tp->nvram_pagesize) <<
8622 ATMEL_AT45DB0X1B_PAGE_POS) +
8623 (offset % tp->nvram_pagesize);
8624 }
8625
8626 if (offset > NVRAM_ADDR_MSK)
8627 return -EINVAL;
8628
8629 tg3_nvram_lock(tp);
8630
e6af301b 8631 tg3_enable_nvram_access(tp);
1da177e4
LT
8632
8633 tw32(NVRAM_ADDR, offset);
8634 ret = tg3_nvram_exec_cmd(tp, NVRAM_CMD_RD | NVRAM_CMD_GO |
8635 NVRAM_CMD_FIRST | NVRAM_CMD_LAST | NVRAM_CMD_DONE);
8636
8637 if (ret == 0)
8638 *val = swab32(tr32(NVRAM_RDDATA));
8639
8640 tg3_nvram_unlock(tp);
8641
e6af301b 8642 tg3_disable_nvram_access(tp);
1da177e4
LT
8643
8644 return ret;
8645}
8646
8647static int tg3_nvram_write_block_using_eeprom(struct tg3 *tp,
8648 u32 offset, u32 len, u8 *buf)
8649{
8650 int i, j, rc = 0;
8651 u32 val;
8652
8653 for (i = 0; i < len; i += 4) {
8654 u32 addr, data;
8655
8656 addr = offset + i;
8657
8658 memcpy(&data, buf + i, 4);
8659
8660 tw32(GRC_EEPROM_DATA, cpu_to_le32(data));
8661
8662 val = tr32(GRC_EEPROM_ADDR);
8663 tw32(GRC_EEPROM_ADDR, val | EEPROM_ADDR_COMPLETE);
8664
8665 val &= ~(EEPROM_ADDR_ADDR_MASK | EEPROM_ADDR_DEVID_MASK |
8666 EEPROM_ADDR_READ);
8667 tw32(GRC_EEPROM_ADDR, val |
8668 (0 << EEPROM_ADDR_DEVID_SHIFT) |
8669 (addr & EEPROM_ADDR_ADDR_MASK) |
8670 EEPROM_ADDR_START |
8671 EEPROM_ADDR_WRITE);
8672
8673 for (j = 0; j < 10000; j++) {
8674 val = tr32(GRC_EEPROM_ADDR);
8675
8676 if (val & EEPROM_ADDR_COMPLETE)
8677 break;
8678 udelay(100);
8679 }
8680 if (!(val & EEPROM_ADDR_COMPLETE)) {
8681 rc = -EBUSY;
8682 break;
8683 }
8684 }
8685
8686 return rc;
8687}
8688
8689/* offset and length are dword aligned */
8690static int tg3_nvram_write_block_unbuffered(struct tg3 *tp, u32 offset, u32 len,
8691 u8 *buf)
8692{
8693 int ret = 0;
8694 u32 pagesize = tp->nvram_pagesize;
8695 u32 pagemask = pagesize - 1;
8696 u32 nvram_cmd;
8697 u8 *tmp;
8698
8699 tmp = kmalloc(pagesize, GFP_KERNEL);
8700 if (tmp == NULL)
8701 return -ENOMEM;
8702
8703 while (len) {
8704 int j;
e6af301b 8705 u32 phy_addr, page_off, size;
1da177e4
LT
8706
8707 phy_addr = offset & ~pagemask;
8708
8709 for (j = 0; j < pagesize; j += 4) {
8710 if ((ret = tg3_nvram_read(tp, phy_addr + j,
8711 (u32 *) (tmp + j))))
8712 break;
8713 }
8714 if (ret)
8715 break;
8716
8717 page_off = offset & pagemask;
8718 size = pagesize;
8719 if (len < size)
8720 size = len;
8721
8722 len -= size;
8723
8724 memcpy(tmp + page_off, buf, size);
8725
8726 offset = offset + (pagesize - page_off);
8727
e6af301b 8728 tg3_enable_nvram_access(tp);
1da177e4
LT
8729
8730 /*
8731 * Before we can erase the flash page, we need
8732 * to issue a special "write enable" command.
8733 */
8734 nvram_cmd = NVRAM_CMD_WREN | NVRAM_CMD_GO | NVRAM_CMD_DONE;
8735
8736 if (tg3_nvram_exec_cmd(tp, nvram_cmd))
8737 break;
8738
8739 /* Erase the target page */
8740 tw32(NVRAM_ADDR, phy_addr);
8741
8742 nvram_cmd = NVRAM_CMD_GO | NVRAM_CMD_DONE | NVRAM_CMD_WR |
8743 NVRAM_CMD_FIRST | NVRAM_CMD_LAST | NVRAM_CMD_ERASE;
8744
8745 if (tg3_nvram_exec_cmd(tp, nvram_cmd))
8746 break;
8747
8748 /* Issue another write enable to start the write. */
8749 nvram_cmd = NVRAM_CMD_WREN | NVRAM_CMD_GO | NVRAM_CMD_DONE;
8750
8751 if (tg3_nvram_exec_cmd(tp, nvram_cmd))
8752 break;
8753
8754 for (j = 0; j < pagesize; j += 4) {
8755 u32 data;
8756
8757 data = *((u32 *) (tmp + j));
8758 tw32(NVRAM_WRDATA, cpu_to_be32(data));
8759
8760 tw32(NVRAM_ADDR, phy_addr + j);
8761
8762 nvram_cmd = NVRAM_CMD_GO | NVRAM_CMD_DONE |
8763 NVRAM_CMD_WR;
8764
8765 if (j == 0)
8766 nvram_cmd |= NVRAM_CMD_FIRST;
8767 else if (j == (pagesize - 4))
8768 nvram_cmd |= NVRAM_CMD_LAST;
8769
8770 if ((ret = tg3_nvram_exec_cmd(tp, nvram_cmd)))
8771 break;
8772 }
8773 if (ret)
8774 break;
8775 }
8776
8777 nvram_cmd = NVRAM_CMD_WRDI | NVRAM_CMD_GO | NVRAM_CMD_DONE;
8778 tg3_nvram_exec_cmd(tp, nvram_cmd);
8779
8780 kfree(tmp);
8781
8782 return ret;
8783}
8784
8785/* offset and length are dword aligned */
8786static int tg3_nvram_write_block_buffered(struct tg3 *tp, u32 offset, u32 len,
8787 u8 *buf)
8788{
8789 int i, ret = 0;
8790
8791 for (i = 0; i < len; i += 4, offset += 4) {
8792 u32 data, page_off, phy_addr, nvram_cmd;
8793
8794 memcpy(&data, buf + i, 4);
8795 tw32(NVRAM_WRDATA, cpu_to_be32(data));
8796
8797 page_off = offset % tp->nvram_pagesize;
8798
8799 if ((tp->tg3_flags2 & TG3_FLG2_FLASH) &&
8800 (tp->nvram_jedecnum == JEDEC_ATMEL)) {
8801
8802 phy_addr = ((offset / tp->nvram_pagesize) <<
8803 ATMEL_AT45DB0X1B_PAGE_POS) + page_off;
8804 }
8805 else {
8806 phy_addr = offset;
8807 }
8808
8809 tw32(NVRAM_ADDR, phy_addr);
8810
8811 nvram_cmd = NVRAM_CMD_GO | NVRAM_CMD_DONE | NVRAM_CMD_WR;
8812
8813 if ((page_off == 0) || (i == 0))
8814 nvram_cmd |= NVRAM_CMD_FIRST;
8815 else if (page_off == (tp->nvram_pagesize - 4))
8816 nvram_cmd |= NVRAM_CMD_LAST;
8817
8818 if (i == (len - 4))
8819 nvram_cmd |= NVRAM_CMD_LAST;
8820
4c987487
MC
8821 if ((GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5752) &&
8822 (tp->nvram_jedecnum == JEDEC_ST) &&
8823 (nvram_cmd & NVRAM_CMD_FIRST)) {
1da177e4
LT
8824
8825 if ((ret = tg3_nvram_exec_cmd(tp,
8826 NVRAM_CMD_WREN | NVRAM_CMD_GO |
8827 NVRAM_CMD_DONE)))
8828
8829 break;
8830 }
8831 if (!(tp->tg3_flags2 & TG3_FLG2_FLASH)) {
8832 /* We always do complete word writes to eeprom. */
8833 nvram_cmd |= (NVRAM_CMD_FIRST | NVRAM_CMD_LAST);
8834 }
8835
8836 if ((ret = tg3_nvram_exec_cmd(tp, nvram_cmd)))
8837 break;
8838 }
8839 return ret;
8840}
8841
8842/* offset and length are dword aligned */
8843static int tg3_nvram_write_block(struct tg3 *tp, u32 offset, u32 len, u8 *buf)
8844{
8845 int ret;
8846
8847 if (tp->tg3_flags2 & TG3_FLG2_SUN_570X) {
8848 printk(KERN_ERR PFX "Attempt to do nvram_write on Sun 570X\n");
8849 return -EINVAL;
8850 }
8851
8852 if (tp->tg3_flags & TG3_FLAG_EEPROM_WRITE_PROT) {
314fba34
MC
8853 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl &
8854 ~GRC_LCLCTRL_GPIO_OUTPUT1);
1da177e4
LT
8855 udelay(40);
8856 }
8857
8858 if (!(tp->tg3_flags & TG3_FLAG_NVRAM)) {
8859 ret = tg3_nvram_write_block_using_eeprom(tp, offset, len, buf);
8860 }
8861 else {
8862 u32 grc_mode;
8863
8864 tg3_nvram_lock(tp);
8865
e6af301b
MC
8866 tg3_enable_nvram_access(tp);
8867 if ((tp->tg3_flags2 & TG3_FLG2_5750_PLUS) &&
8868 !(tp->tg3_flags2 & TG3_FLG2_PROTECTED_NVRAM))
1da177e4 8869 tw32(NVRAM_WRITE1, 0x406);
1da177e4
LT
8870
8871 grc_mode = tr32(GRC_MODE);
8872 tw32(GRC_MODE, grc_mode | GRC_MODE_NVRAM_WR_ENABLE);
8873
8874 if ((tp->tg3_flags & TG3_FLAG_NVRAM_BUFFERED) ||
8875 !(tp->tg3_flags2 & TG3_FLG2_FLASH)) {
8876
8877 ret = tg3_nvram_write_block_buffered(tp, offset, len,
8878 buf);
8879 }
8880 else {
8881 ret = tg3_nvram_write_block_unbuffered(tp, offset, len,
8882 buf);
8883 }
8884
8885 grc_mode = tr32(GRC_MODE);
8886 tw32(GRC_MODE, grc_mode & ~GRC_MODE_NVRAM_WR_ENABLE);
8887
e6af301b 8888 tg3_disable_nvram_access(tp);
1da177e4
LT
8889 tg3_nvram_unlock(tp);
8890 }
8891
8892 if (tp->tg3_flags & TG3_FLAG_EEPROM_WRITE_PROT) {
314fba34 8893 tw32_f(GRC_LOCAL_CTRL, tp->grc_local_ctrl);
1da177e4
LT
8894 udelay(40);
8895 }
8896
8897 return ret;
8898}
8899
8900struct subsys_tbl_ent {
8901 u16 subsys_vendor, subsys_devid;
8902 u32 phy_id;
8903};
8904
8905static struct subsys_tbl_ent subsys_id_to_phy_id[] = {
8906 /* Broadcom boards. */
8907 { PCI_VENDOR_ID_BROADCOM, 0x1644, PHY_ID_BCM5401 }, /* BCM95700A6 */
8908 { PCI_VENDOR_ID_BROADCOM, 0x0001, PHY_ID_BCM5701 }, /* BCM95701A5 */
8909 { PCI_VENDOR_ID_BROADCOM, 0x0002, PHY_ID_BCM8002 }, /* BCM95700T6 */
8910 { PCI_VENDOR_ID_BROADCOM, 0x0003, 0 }, /* BCM95700A9 */
8911 { PCI_VENDOR_ID_BROADCOM, 0x0005, PHY_ID_BCM5701 }, /* BCM95701T1 */
8912 { PCI_VENDOR_ID_BROADCOM, 0x0006, PHY_ID_BCM5701 }, /* BCM95701T8 */
8913 { PCI_VENDOR_ID_BROADCOM, 0x0007, 0 }, /* BCM95701A7 */
8914 { PCI_VENDOR_ID_BROADCOM, 0x0008, PHY_ID_BCM5701 }, /* BCM95701A10 */
8915 { PCI_VENDOR_ID_BROADCOM, 0x8008, PHY_ID_BCM5701 }, /* BCM95701A12 */
8916 { PCI_VENDOR_ID_BROADCOM, 0x0009, PHY_ID_BCM5703 }, /* BCM95703Ax1 */
8917 { PCI_VENDOR_ID_BROADCOM, 0x8009, PHY_ID_BCM5703 }, /* BCM95703Ax2 */
8918
8919 /* 3com boards. */
8920 { PCI_VENDOR_ID_3COM, 0x1000, PHY_ID_BCM5401 }, /* 3C996T */
8921 { PCI_VENDOR_ID_3COM, 0x1006, PHY_ID_BCM5701 }, /* 3C996BT */
8922 { PCI_VENDOR_ID_3COM, 0x1004, 0 }, /* 3C996SX */
8923 { PCI_VENDOR_ID_3COM, 0x1007, PHY_ID_BCM5701 }, /* 3C1000T */
8924 { PCI_VENDOR_ID_3COM, 0x1008, PHY_ID_BCM5701 }, /* 3C940BR01 */
8925
8926 /* DELL boards. */
8927 { PCI_VENDOR_ID_DELL, 0x00d1, PHY_ID_BCM5401 }, /* VIPER */
8928 { PCI_VENDOR_ID_DELL, 0x0106, PHY_ID_BCM5401 }, /* JAGUAR */
8929 { PCI_VENDOR_ID_DELL, 0x0109, PHY_ID_BCM5411 }, /* MERLOT */
8930 { PCI_VENDOR_ID_DELL, 0x010a, PHY_ID_BCM5411 }, /* SLIM_MERLOT */
8931
8932 /* Compaq boards. */
8933 { PCI_VENDOR_ID_COMPAQ, 0x007c, PHY_ID_BCM5701 }, /* BANSHEE */
8934 { PCI_VENDOR_ID_COMPAQ, 0x009a, PHY_ID_BCM5701 }, /* BANSHEE_2 */
8935 { PCI_VENDOR_ID_COMPAQ, 0x007d, 0 }, /* CHANGELING */
8936 { PCI_VENDOR_ID_COMPAQ, 0x0085, PHY_ID_BCM5701 }, /* NC7780 */
8937 { PCI_VENDOR_ID_COMPAQ, 0x0099, PHY_ID_BCM5701 }, /* NC7780_2 */
8938
8939 /* IBM boards. */
8940 { PCI_VENDOR_ID_IBM, 0x0281, 0 } /* IBM??? */
8941};
8942
8943static inline struct subsys_tbl_ent *lookup_by_subsys(struct tg3 *tp)
8944{
8945 int i;
8946
8947 for (i = 0; i < ARRAY_SIZE(subsys_id_to_phy_id); i++) {
8948 if ((subsys_id_to_phy_id[i].subsys_vendor ==
8949 tp->pdev->subsystem_vendor) &&
8950 (subsys_id_to_phy_id[i].subsys_devid ==
8951 tp->pdev->subsystem_device))
8952 return &subsys_id_to_phy_id[i];
8953 }
8954 return NULL;
8955}
8956
7d0c41ef
MC
8957/* Since this function may be called in D3-hot power state during
8958 * tg3_init_one(), only config cycles are allowed.
8959 */
8960static void __devinit tg3_get_eeprom_hw_cfg(struct tg3 *tp)
1da177e4 8961{
1da177e4 8962 u32 val;
7d0c41ef
MC
8963
8964 /* Make sure register accesses (indirect or otherwise)
8965 * will function correctly.
8966 */
8967 pci_write_config_dword(tp->pdev, TG3PCI_MISC_HOST_CTRL,
8968 tp->misc_host_ctrl);
1da177e4
LT
8969
8970 tp->phy_id = PHY_ID_INVALID;
7d0c41ef
MC
8971 tp->led_ctrl = LED_CTRL_MODE_PHY_1;
8972
1da177e4
LT
8973 tg3_read_mem(tp, NIC_SRAM_DATA_SIG, &val);
8974 if (val == NIC_SRAM_DATA_SIG_MAGIC) {
8975 u32 nic_cfg, led_cfg;
7d0c41ef
MC
8976 u32 nic_phy_id, ver, cfg2 = 0, eeprom_phy_id;
8977 int eeprom_phy_serdes = 0;
1da177e4
LT
8978
8979 tg3_read_mem(tp, NIC_SRAM_DATA_CFG, &nic_cfg);
8980 tp->nic_sram_data_cfg = nic_cfg;
8981
8982 tg3_read_mem(tp, NIC_SRAM_DATA_VER, &ver);
8983 ver >>= NIC_SRAM_DATA_VER_SHIFT;
8984 if ((GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700) &&
8985 (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5701) &&
8986 (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5703) &&
8987 (ver > 0) && (ver < 0x100))
8988 tg3_read_mem(tp, NIC_SRAM_DATA_CFG_2, &cfg2);
8989
1da177e4
LT
8990 if ((nic_cfg & NIC_SRAM_DATA_CFG_PHY_TYPE_MASK) ==
8991 NIC_SRAM_DATA_CFG_PHY_TYPE_FIBER)
8992 eeprom_phy_serdes = 1;
8993
8994 tg3_read_mem(tp, NIC_SRAM_DATA_PHY_ID, &nic_phy_id);
8995 if (nic_phy_id != 0) {
8996 u32 id1 = nic_phy_id & NIC_SRAM_DATA_PHY_ID1_MASK;
8997 u32 id2 = nic_phy_id & NIC_SRAM_DATA_PHY_ID2_MASK;
8998
8999 eeprom_phy_id = (id1 >> 16) << 10;
9000 eeprom_phy_id |= (id2 & 0xfc00) << 16;
9001 eeprom_phy_id |= (id2 & 0x03ff) << 0;
9002 } else
9003 eeprom_phy_id = 0;
9004
7d0c41ef 9005 tp->phy_id = eeprom_phy_id;
747e8f8b 9006 if (eeprom_phy_serdes) {
a4e2b347 9007 if (tp->tg3_flags2 & TG3_FLG2_5780_CLASS)
747e8f8b
MC
9008 tp->tg3_flags2 |= TG3_FLG2_MII_SERDES;
9009 else
9010 tp->tg3_flags2 |= TG3_FLG2_PHY_SERDES;
9011 }
7d0c41ef 9012
cbf46853 9013 if (tp->tg3_flags2 & TG3_FLG2_5750_PLUS)
1da177e4
LT
9014 led_cfg = cfg2 & (NIC_SRAM_DATA_CFG_LED_MODE_MASK |
9015 SHASTA_EXT_LED_MODE_MASK);
cbf46853 9016 else
1da177e4
LT
9017 led_cfg = nic_cfg & NIC_SRAM_DATA_CFG_LED_MODE_MASK;
9018
9019 switch (led_cfg) {
9020 default:
9021 case NIC_SRAM_DATA_CFG_LED_MODE_PHY_1:
9022 tp->led_ctrl = LED_CTRL_MODE_PHY_1;
9023 break;
9024
9025 case NIC_SRAM_DATA_CFG_LED_MODE_PHY_2:
9026 tp->led_ctrl = LED_CTRL_MODE_PHY_2;
9027 break;
9028
9029 case NIC_SRAM_DATA_CFG_LED_MODE_MAC:
9030 tp->led_ctrl = LED_CTRL_MODE_MAC;
9ba27794
MC
9031
9032 /* Default to PHY_1_MODE if 0 (MAC_MODE) is
9033 * read on some older 5700/5701 bootcode.
9034 */
9035 if (GET_ASIC_REV(tp->pci_chip_rev_id) ==
9036 ASIC_REV_5700 ||
9037 GET_ASIC_REV(tp->pci_chip_rev_id) ==
9038 ASIC_REV_5701)
9039 tp->led_ctrl = LED_CTRL_MODE_PHY_1;
9040
1da177e4
LT
9041 break;
9042
9043 case SHASTA_EXT_LED_SHARED:
9044 tp->led_ctrl = LED_CTRL_MODE_SHARED;
9045 if (tp->pci_chip_rev_id != CHIPREV_ID_5750_A0 &&
9046 tp->pci_chip_rev_id != CHIPREV_ID_5750_A1)
9047 tp->led_ctrl |= (LED_CTRL_MODE_PHY_1 |
9048 LED_CTRL_MODE_PHY_2);
9049 break;
9050
9051 case SHASTA_EXT_LED_MAC:
9052 tp->led_ctrl = LED_CTRL_MODE_SHASTA_MAC;
9053 break;
9054
9055 case SHASTA_EXT_LED_COMBO:
9056 tp->led_ctrl = LED_CTRL_MODE_COMBO;
9057 if (tp->pci_chip_rev_id != CHIPREV_ID_5750_A0)
9058 tp->led_ctrl |= (LED_CTRL_MODE_PHY_1 |
9059 LED_CTRL_MODE_PHY_2);
9060 break;
9061
9062 };
9063
9064 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
9065 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701) &&
9066 tp->pdev->subsystem_vendor == PCI_VENDOR_ID_DELL)
9067 tp->led_ctrl = LED_CTRL_MODE_PHY_2;
9068
9069 if ((GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700) &&
9070 (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5701) &&
9071 (nic_cfg & NIC_SRAM_DATA_CFG_EEPROM_WP))
9072 tp->tg3_flags |= TG3_FLAG_EEPROM_WRITE_PROT;
9073
9074 if (nic_cfg & NIC_SRAM_DATA_CFG_ASF_ENABLE) {
9075 tp->tg3_flags |= TG3_FLAG_ENABLE_ASF;
cbf46853 9076 if (tp->tg3_flags2 & TG3_FLG2_5750_PLUS)
1da177e4
LT
9077 tp->tg3_flags2 |= TG3_FLG2_ASF_NEW_HANDSHAKE;
9078 }
9079 if (nic_cfg & NIC_SRAM_DATA_CFG_FIBER_WOL)
9080 tp->tg3_flags |= TG3_FLAG_SERDES_WOL_CAP;
9081
9082 if (cfg2 & (1 << 17))
9083 tp->tg3_flags2 |= TG3_FLG2_CAPACITIVE_COUPLING;
9084
9085 /* serdes signal pre-emphasis in register 0x590 set by */
9086 /* bootcode if bit 18 is set */
9087 if (cfg2 & (1 << 18))
9088 tp->tg3_flags2 |= TG3_FLG2_SERDES_PREEMPHASIS;
9089 }
7d0c41ef
MC
9090}
9091
9092static int __devinit tg3_phy_probe(struct tg3 *tp)
9093{
9094 u32 hw_phy_id_1, hw_phy_id_2;
9095 u32 hw_phy_id, hw_phy_id_masked;
9096 int err;
1da177e4
LT
9097
9098 /* Reading the PHY ID register can conflict with ASF
9099 * firwmare access to the PHY hardware.
9100 */
9101 err = 0;
9102 if (tp->tg3_flags & TG3_FLAG_ENABLE_ASF) {
9103 hw_phy_id = hw_phy_id_masked = PHY_ID_INVALID;
9104 } else {
9105 /* Now read the physical PHY_ID from the chip and verify
9106 * that it is sane. If it doesn't look good, we fall back
9107 * to either the hard-coded table based PHY_ID and failing
9108 * that the value found in the eeprom area.
9109 */
9110 err |= tg3_readphy(tp, MII_PHYSID1, &hw_phy_id_1);
9111 err |= tg3_readphy(tp, MII_PHYSID2, &hw_phy_id_2);
9112
9113 hw_phy_id = (hw_phy_id_1 & 0xffff) << 10;
9114 hw_phy_id |= (hw_phy_id_2 & 0xfc00) << 16;
9115 hw_phy_id |= (hw_phy_id_2 & 0x03ff) << 0;
9116
9117 hw_phy_id_masked = hw_phy_id & PHY_ID_MASK;
9118 }
9119
9120 if (!err && KNOWN_PHY_ID(hw_phy_id_masked)) {
9121 tp->phy_id = hw_phy_id;
9122 if (hw_phy_id_masked == PHY_ID_BCM8002)
9123 tp->tg3_flags2 |= TG3_FLG2_PHY_SERDES;
da6b2d01
MC
9124 else
9125 tp->tg3_flags2 &= ~TG3_FLG2_PHY_SERDES;
1da177e4 9126 } else {
7d0c41ef
MC
9127 if (tp->phy_id != PHY_ID_INVALID) {
9128 /* Do nothing, phy ID already set up in
9129 * tg3_get_eeprom_hw_cfg().
9130 */
1da177e4
LT
9131 } else {
9132 struct subsys_tbl_ent *p;
9133
9134 /* No eeprom signature? Try the hardcoded
9135 * subsys device table.
9136 */
9137 p = lookup_by_subsys(tp);
9138 if (!p)
9139 return -ENODEV;
9140
9141 tp->phy_id = p->phy_id;
9142 if (!tp->phy_id ||
9143 tp->phy_id == PHY_ID_BCM8002)
9144 tp->tg3_flags2 |= TG3_FLG2_PHY_SERDES;
9145 }
9146 }
9147
747e8f8b 9148 if (!(tp->tg3_flags2 & TG3_FLG2_ANY_SERDES) &&
1da177e4
LT
9149 !(tp->tg3_flags & TG3_FLAG_ENABLE_ASF)) {
9150 u32 bmsr, adv_reg, tg3_ctrl;
9151
9152 tg3_readphy(tp, MII_BMSR, &bmsr);
9153 if (!tg3_readphy(tp, MII_BMSR, &bmsr) &&
9154 (bmsr & BMSR_LSTATUS))
9155 goto skip_phy_reset;
9156
9157 err = tg3_phy_reset(tp);
9158 if (err)
9159 return err;
9160
9161 adv_reg = (ADVERTISE_10HALF | ADVERTISE_10FULL |
9162 ADVERTISE_100HALF | ADVERTISE_100FULL |
9163 ADVERTISE_CSMA | ADVERTISE_PAUSE_CAP);
9164 tg3_ctrl = 0;
9165 if (!(tp->tg3_flags & TG3_FLAG_10_100_ONLY)) {
9166 tg3_ctrl = (MII_TG3_CTRL_ADV_1000_HALF |
9167 MII_TG3_CTRL_ADV_1000_FULL);
9168 if (tp->pci_chip_rev_id == CHIPREV_ID_5701_A0 ||
9169 tp->pci_chip_rev_id == CHIPREV_ID_5701_B0)
9170 tg3_ctrl |= (MII_TG3_CTRL_AS_MASTER |
9171 MII_TG3_CTRL_ENABLE_AS_MASTER);
9172 }
9173
9174 if (!tg3_copper_is_advertising_all(tp)) {
9175 tg3_writephy(tp, MII_ADVERTISE, adv_reg);
9176
9177 if (!(tp->tg3_flags & TG3_FLAG_10_100_ONLY))
9178 tg3_writephy(tp, MII_TG3_CTRL, tg3_ctrl);
9179
9180 tg3_writephy(tp, MII_BMCR,
9181 BMCR_ANENABLE | BMCR_ANRESTART);
9182 }
9183 tg3_phy_set_wirespeed(tp);
9184
9185 tg3_writephy(tp, MII_ADVERTISE, adv_reg);
9186 if (!(tp->tg3_flags & TG3_FLAG_10_100_ONLY))
9187 tg3_writephy(tp, MII_TG3_CTRL, tg3_ctrl);
9188 }
9189
9190skip_phy_reset:
9191 if ((tp->phy_id & PHY_ID_MASK) == PHY_ID_BCM5401) {
9192 err = tg3_init_5401phy_dsp(tp);
9193 if (err)
9194 return err;
9195 }
9196
9197 if (!err && ((tp->phy_id & PHY_ID_MASK) == PHY_ID_BCM5401)) {
9198 err = tg3_init_5401phy_dsp(tp);
9199 }
9200
747e8f8b 9201 if (tp->tg3_flags2 & TG3_FLG2_ANY_SERDES)
1da177e4
LT
9202 tp->link_config.advertising =
9203 (ADVERTISED_1000baseT_Half |
9204 ADVERTISED_1000baseT_Full |
9205 ADVERTISED_Autoneg |
9206 ADVERTISED_FIBRE);
9207 if (tp->tg3_flags & TG3_FLAG_10_100_ONLY)
9208 tp->link_config.advertising &=
9209 ~(ADVERTISED_1000baseT_Half |
9210 ADVERTISED_1000baseT_Full);
9211
9212 return err;
9213}
9214
9215static void __devinit tg3_read_partno(struct tg3 *tp)
9216{
9217 unsigned char vpd_data[256];
9218 int i;
9219
9220 if (tp->tg3_flags2 & TG3_FLG2_SUN_570X) {
9221 /* Sun decided not to put the necessary bits in the
9222 * NVRAM of their onboard tg3 parts :(
9223 */
9224 strcpy(tp->board_part_number, "Sun 570X");
9225 return;
9226 }
9227
9228 for (i = 0; i < 256; i += 4) {
9229 u32 tmp;
9230
9231 if (tg3_nvram_read(tp, 0x100 + i, &tmp))
9232 goto out_not_found;
9233
9234 vpd_data[i + 0] = ((tmp >> 0) & 0xff);
9235 vpd_data[i + 1] = ((tmp >> 8) & 0xff);
9236 vpd_data[i + 2] = ((tmp >> 16) & 0xff);
9237 vpd_data[i + 3] = ((tmp >> 24) & 0xff);
9238 }
9239
9240 /* Now parse and find the part number. */
9241 for (i = 0; i < 256; ) {
9242 unsigned char val = vpd_data[i];
9243 int block_end;
9244
9245 if (val == 0x82 || val == 0x91) {
9246 i = (i + 3 +
9247 (vpd_data[i + 1] +
9248 (vpd_data[i + 2] << 8)));
9249 continue;
9250 }
9251
9252 if (val != 0x90)
9253 goto out_not_found;
9254
9255 block_end = (i + 3 +
9256 (vpd_data[i + 1] +
9257 (vpd_data[i + 2] << 8)));
9258 i += 3;
9259 while (i < block_end) {
9260 if (vpd_data[i + 0] == 'P' &&
9261 vpd_data[i + 1] == 'N') {
9262 int partno_len = vpd_data[i + 2];
9263
9264 if (partno_len > 24)
9265 goto out_not_found;
9266
9267 memcpy(tp->board_part_number,
9268 &vpd_data[i + 3],
9269 partno_len);
9270
9271 /* Success. */
9272 return;
9273 }
9274 }
9275
9276 /* Part number not found. */
9277 goto out_not_found;
9278 }
9279
9280out_not_found:
9281 strcpy(tp->board_part_number, "none");
9282}
9283
9284#ifdef CONFIG_SPARC64
9285static int __devinit tg3_is_sun_570X(struct tg3 *tp)
9286{
9287 struct pci_dev *pdev = tp->pdev;
9288 struct pcidev_cookie *pcp = pdev->sysdata;
9289
9290 if (pcp != NULL) {
9291 int node = pcp->prom_node;
9292 u32 venid;
9293 int err;
9294
9295 err = prom_getproperty(node, "subsystem-vendor-id",
9296 (char *) &venid, sizeof(venid));
9297 if (err == 0 || err == -1)
9298 return 0;
9299 if (venid == PCI_VENDOR_ID_SUN)
9300 return 1;
9301 }
9302 return 0;
9303}
9304#endif
9305
9306static int __devinit tg3_get_invariants(struct tg3 *tp)
9307{
9308 static struct pci_device_id write_reorder_chipsets[] = {
1da177e4
LT
9309 { PCI_DEVICE(PCI_VENDOR_ID_AMD,
9310 PCI_DEVICE_ID_AMD_FE_GATE_700C) },
399de50b
MC
9311 { PCI_DEVICE(PCI_VENDOR_ID_VIA,
9312 PCI_DEVICE_ID_VIA_8385_0) },
1da177e4
LT
9313 { },
9314 };
9315 u32 misc_ctrl_reg;
9316 u32 cacheline_sz_reg;
9317 u32 pci_state_reg, grc_misc_cfg;
9318 u32 val;
9319 u16 pci_cmd;
9320 int err;
9321
9322#ifdef CONFIG_SPARC64
9323 if (tg3_is_sun_570X(tp))
9324 tp->tg3_flags2 |= TG3_FLG2_SUN_570X;
9325#endif
9326
1da177e4
LT
9327 /* Force memory write invalidate off. If we leave it on,
9328 * then on 5700_BX chips we have to enable a workaround.
9329 * The workaround is to set the TG3PCI_DMA_RW_CTRL boundary
9330 * to match the cacheline size. The Broadcom driver have this
9331 * workaround but turns MWI off all the times so never uses
9332 * it. This seems to suggest that the workaround is insufficient.
9333 */
9334 pci_read_config_word(tp->pdev, PCI_COMMAND, &pci_cmd);
9335 pci_cmd &= ~PCI_COMMAND_INVALIDATE;
9336 pci_write_config_word(tp->pdev, PCI_COMMAND, pci_cmd);
9337
9338 /* It is absolutely critical that TG3PCI_MISC_HOST_CTRL
9339 * has the register indirect write enable bit set before
9340 * we try to access any of the MMIO registers. It is also
9341 * critical that the PCI-X hw workaround situation is decided
9342 * before that as well.
9343 */
9344 pci_read_config_dword(tp->pdev, TG3PCI_MISC_HOST_CTRL,
9345 &misc_ctrl_reg);
9346
9347 tp->pci_chip_rev_id = (misc_ctrl_reg >>
9348 MISC_HOST_CTRL_CHIPREV_SHIFT);
9349
ff645bec
MC
9350 /* Wrong chip ID in 5752 A0. This code can be removed later
9351 * as A0 is not in production.
9352 */
9353 if (tp->pci_chip_rev_id == CHIPREV_ID_5752_A0_HW)
9354 tp->pci_chip_rev_id = CHIPREV_ID_5752_A0;
9355
6892914f
MC
9356 /* If we have 5702/03 A1 or A2 on certain ICH chipsets,
9357 * we need to disable memory and use config. cycles
9358 * only to access all registers. The 5702/03 chips
9359 * can mistakenly decode the special cycles from the
9360 * ICH chipsets as memory write cycles, causing corruption
9361 * of register and memory space. Only certain ICH bridges
9362 * will drive special cycles with non-zero data during the
9363 * address phase which can fall within the 5703's address
9364 * range. This is not an ICH bug as the PCI spec allows
9365 * non-zero address during special cycles. However, only
9366 * these ICH bridges are known to drive non-zero addresses
9367 * during special cycles.
9368 *
9369 * Since special cycles do not cross PCI bridges, we only
9370 * enable this workaround if the 5703 is on the secondary
9371 * bus of these ICH bridges.
9372 */
9373 if ((tp->pci_chip_rev_id == CHIPREV_ID_5703_A1) ||
9374 (tp->pci_chip_rev_id == CHIPREV_ID_5703_A2)) {
9375 static struct tg3_dev_id {
9376 u32 vendor;
9377 u32 device;
9378 u32 rev;
9379 } ich_chipsets[] = {
9380 { PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801AA_8,
9381 PCI_ANY_ID },
9382 { PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801AB_8,
9383 PCI_ANY_ID },
9384 { PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801BA_11,
9385 0xa },
9386 { PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801BA_6,
9387 PCI_ANY_ID },
9388 { },
9389 };
9390 struct tg3_dev_id *pci_id = &ich_chipsets[0];
9391 struct pci_dev *bridge = NULL;
9392
9393 while (pci_id->vendor != 0) {
9394 bridge = pci_get_device(pci_id->vendor, pci_id->device,
9395 bridge);
9396 if (!bridge) {
9397 pci_id++;
9398 continue;
9399 }
9400 if (pci_id->rev != PCI_ANY_ID) {
9401 u8 rev;
9402
9403 pci_read_config_byte(bridge, PCI_REVISION_ID,
9404 &rev);
9405 if (rev > pci_id->rev)
9406 continue;
9407 }
9408 if (bridge->subordinate &&
9409 (bridge->subordinate->number ==
9410 tp->pdev->bus->number)) {
9411
9412 tp->tg3_flags2 |= TG3_FLG2_ICH_WORKAROUND;
9413 pci_dev_put(bridge);
9414 break;
9415 }
9416 }
9417 }
9418
4cf78e4f 9419 /* Find msi capability. */
a4e2b347
MC
9420 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5780 ||
9421 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5714) {
9422 tp->tg3_flags2 |= TG3_FLG2_5780_CLASS;
4cf78e4f 9423 tp->msi_cap = pci_find_capability(tp->pdev, PCI_CAP_ID_MSI);
a4e2b347 9424 }
4cf78e4f 9425
1da177e4
LT
9426 /* Initialize misc host control in PCI block. */
9427 tp->misc_host_ctrl |= (misc_ctrl_reg &
9428 MISC_HOST_CTRL_CHIPREV);
9429 pci_write_config_dword(tp->pdev, TG3PCI_MISC_HOST_CTRL,
9430 tp->misc_host_ctrl);
9431
9432 pci_read_config_dword(tp->pdev, TG3PCI_CACHELINESZ,
9433 &cacheline_sz_reg);
9434
9435 tp->pci_cacheline_sz = (cacheline_sz_reg >> 0) & 0xff;
9436 tp->pci_lat_timer = (cacheline_sz_reg >> 8) & 0xff;
9437 tp->pci_hdr_type = (cacheline_sz_reg >> 16) & 0xff;
9438 tp->pci_bist = (cacheline_sz_reg >> 24) & 0xff;
9439
6708e5cc 9440 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5750 ||
4cf78e4f 9441 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5752 ||
a4e2b347 9442 (tp->tg3_flags2 & TG3_FLG2_5780_CLASS))
6708e5cc
JL
9443 tp->tg3_flags2 |= TG3_FLG2_5750_PLUS;
9444
1b440c56
JL
9445 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705) ||
9446 (tp->tg3_flags2 & TG3_FLG2_5750_PLUS))
9447 tp->tg3_flags2 |= TG3_FLG2_5705_PLUS;
9448
bb7064dc 9449 if (tp->tg3_flags2 & TG3_FLG2_5750_PLUS)
1da177e4
LT
9450 tp->tg3_flags2 |= TG3_FLG2_HW_TSO;
9451
0f893dc6
MC
9452 if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5705 &&
9453 GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5750 &&
9454 GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5752)
9455 tp->tg3_flags2 |= TG3_FLG2_JUMBO_CAPABLE;
9456
1da177e4
LT
9457 if (pci_find_capability(tp->pdev, PCI_CAP_ID_EXP) != 0)
9458 tp->tg3_flags2 |= TG3_FLG2_PCI_EXPRESS;
9459
399de50b
MC
9460 /* If we have an AMD 762 or VIA K8T800 chipset, write
9461 * reordering to the mailbox registers done by the host
9462 * controller can cause major troubles. We read back from
9463 * every mailbox register write to force the writes to be
9464 * posted to the chip in order.
9465 */
9466 if (pci_dev_present(write_reorder_chipsets) &&
9467 !(tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS))
9468 tp->tg3_flags |= TG3_FLAG_MBOX_WRITE_REORDER;
9469
1da177e4
LT
9470 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 &&
9471 tp->pci_lat_timer < 64) {
9472 tp->pci_lat_timer = 64;
9473
9474 cacheline_sz_reg = ((tp->pci_cacheline_sz & 0xff) << 0);
9475 cacheline_sz_reg |= ((tp->pci_lat_timer & 0xff) << 8);
9476 cacheline_sz_reg |= ((tp->pci_hdr_type & 0xff) << 16);
9477 cacheline_sz_reg |= ((tp->pci_bist & 0xff) << 24);
9478
9479 pci_write_config_dword(tp->pdev, TG3PCI_CACHELINESZ,
9480 cacheline_sz_reg);
9481 }
9482
9483 pci_read_config_dword(tp->pdev, TG3PCI_PCISTATE,
9484 &pci_state_reg);
9485
9486 if ((pci_state_reg & PCISTATE_CONV_PCI_MODE) == 0) {
9487 tp->tg3_flags |= TG3_FLAG_PCIX_MODE;
9488
9489 /* If this is a 5700 BX chipset, and we are in PCI-X
9490 * mode, enable register write workaround.
9491 *
9492 * The workaround is to use indirect register accesses
9493 * for all chip writes not to mailbox registers.
9494 */
9495 if (GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5700_BX) {
9496 u32 pm_reg;
9497 u16 pci_cmd;
9498
9499 tp->tg3_flags |= TG3_FLAG_PCIX_TARGET_HWBUG;
9500
9501 /* The chip can have it's power management PCI config
9502 * space registers clobbered due to this bug.
9503 * So explicitly force the chip into D0 here.
9504 */
9505 pci_read_config_dword(tp->pdev, TG3PCI_PM_CTRL_STAT,
9506 &pm_reg);
9507 pm_reg &= ~PCI_PM_CTRL_STATE_MASK;
9508 pm_reg |= PCI_PM_CTRL_PME_ENABLE | 0 /* D0 */;
9509 pci_write_config_dword(tp->pdev, TG3PCI_PM_CTRL_STAT,
9510 pm_reg);
9511
9512 /* Also, force SERR#/PERR# in PCI command. */
9513 pci_read_config_word(tp->pdev, PCI_COMMAND, &pci_cmd);
9514 pci_cmd |= PCI_COMMAND_PARITY | PCI_COMMAND_SERR;
9515 pci_write_config_word(tp->pdev, PCI_COMMAND, pci_cmd);
9516 }
9517 }
9518
087fe256
MC
9519 /* 5700 BX chips need to have their TX producer index mailboxes
9520 * written twice to workaround a bug.
9521 */
9522 if (GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5700_BX)
9523 tp->tg3_flags |= TG3_FLAG_TXD_MBOX_HWBUG;
9524
1da177e4
LT
9525 /* Back to back register writes can cause problems on this chip,
9526 * the workaround is to read back all reg writes except those to
9527 * mailbox regs. See tg3_write_indirect_reg32().
9528 *
9529 * PCI Express 5750_A0 rev chips need this workaround too.
9530 */
9531 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701 ||
9532 ((tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) &&
9533 tp->pci_chip_rev_id == CHIPREV_ID_5750_A0))
9534 tp->tg3_flags |= TG3_FLAG_5701_REG_WRITE_BUG;
9535
9536 if ((pci_state_reg & PCISTATE_BUS_SPEED_HIGH) != 0)
9537 tp->tg3_flags |= TG3_FLAG_PCI_HIGH_SPEED;
9538 if ((pci_state_reg & PCISTATE_BUS_32BIT) != 0)
9539 tp->tg3_flags |= TG3_FLAG_PCI_32BIT;
9540
9541 /* Chip-specific fixup from Broadcom driver */
9542 if ((tp->pci_chip_rev_id == CHIPREV_ID_5704_A0) &&
9543 (!(pci_state_reg & PCISTATE_RETRY_SAME_DMA))) {
9544 pci_state_reg |= PCISTATE_RETRY_SAME_DMA;
9545 pci_write_config_dword(tp->pdev, TG3PCI_PCISTATE, pci_state_reg);
9546 }
9547
1ee582d8 9548 /* Default fast path register access methods */
20094930 9549 tp->read32 = tg3_read32;
1ee582d8 9550 tp->write32 = tg3_write32;
09ee929c 9551 tp->read32_mbox = tg3_read32;
20094930 9552 tp->write32_mbox = tg3_write32;
1ee582d8
MC
9553 tp->write32_tx_mbox = tg3_write32;
9554 tp->write32_rx_mbox = tg3_write32;
9555
9556 /* Various workaround register access methods */
9557 if (tp->tg3_flags & TG3_FLAG_PCIX_TARGET_HWBUG)
9558 tp->write32 = tg3_write_indirect_reg32;
9559 else if (tp->tg3_flags & TG3_FLAG_5701_REG_WRITE_BUG)
9560 tp->write32 = tg3_write_flush_reg32;
9561
9562 if ((tp->tg3_flags & TG3_FLAG_TXD_MBOX_HWBUG) ||
9563 (tp->tg3_flags & TG3_FLAG_MBOX_WRITE_REORDER)) {
9564 tp->write32_tx_mbox = tg3_write32_tx_mbox;
9565 if (tp->tg3_flags & TG3_FLAG_MBOX_WRITE_REORDER)
9566 tp->write32_rx_mbox = tg3_write_flush_reg32;
9567 }
20094930 9568
6892914f
MC
9569 if (tp->tg3_flags2 & TG3_FLG2_ICH_WORKAROUND) {
9570 tp->read32 = tg3_read_indirect_reg32;
9571 tp->write32 = tg3_write_indirect_reg32;
9572 tp->read32_mbox = tg3_read_indirect_mbox;
9573 tp->write32_mbox = tg3_write_indirect_mbox;
9574 tp->write32_tx_mbox = tg3_write_indirect_mbox;
9575 tp->write32_rx_mbox = tg3_write_indirect_mbox;
9576
9577 iounmap(tp->regs);
22abe310 9578 tp->regs = NULL;
6892914f
MC
9579
9580 pci_read_config_word(tp->pdev, PCI_COMMAND, &pci_cmd);
9581 pci_cmd &= ~PCI_COMMAND_MEMORY;
9582 pci_write_config_word(tp->pdev, PCI_COMMAND, pci_cmd);
9583 }
9584
7d0c41ef
MC
9585 /* Get eeprom hw config before calling tg3_set_power_state().
9586 * In particular, the TG3_FLAG_EEPROM_WRITE_PROT flag must be
9587 * determined before calling tg3_set_power_state() so that
9588 * we know whether or not to switch out of Vaux power.
9589 * When the flag is set, it means that GPIO1 is used for eeprom
9590 * write protect and also implies that it is a LOM where GPIOs
9591 * are not used to switch power.
9592 */
9593 tg3_get_eeprom_hw_cfg(tp);
9594
314fba34
MC
9595 /* Set up tp->grc_local_ctrl before calling tg3_set_power_state().
9596 * GPIO1 driven high will bring 5700's external PHY out of reset.
9597 * It is also used as eeprom write protect on LOMs.
9598 */
9599 tp->grc_local_ctrl = GRC_LCLCTRL_INT_ON_ATTN | GRC_LCLCTRL_AUTO_SEEPROM;
9600 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700) ||
9601 (tp->tg3_flags & TG3_FLAG_EEPROM_WRITE_PROT))
9602 tp->grc_local_ctrl |= (GRC_LCLCTRL_GPIO_OE1 |
9603 GRC_LCLCTRL_GPIO_OUTPUT1);
3e7d83bc
MC
9604 /* Unused GPIO3 must be driven as output on 5752 because there
9605 * are no pull-up resistors on unused GPIO pins.
9606 */
9607 else if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5752)
9608 tp->grc_local_ctrl |= GRC_LCLCTRL_GPIO_OE3;
314fba34 9609
1da177e4
LT
9610 /* Force the chip into D0. */
9611 err = tg3_set_power_state(tp, 0);
9612 if (err) {
9613 printk(KERN_ERR PFX "(%s) transition to D0 failed\n",
9614 pci_name(tp->pdev));
9615 return err;
9616 }
9617
9618 /* 5700 B0 chips do not support checksumming correctly due
9619 * to hardware bugs.
9620 */
9621 if (tp->pci_chip_rev_id == CHIPREV_ID_5700_B0)
9622 tp->tg3_flags |= TG3_FLAG_BROKEN_CHECKSUMS;
9623
9624 /* Pseudo-header checksum is done by hardware logic and not
9625 * the offload processers, so make the chip do the pseudo-
9626 * header checksums on receive. For transmit it is more
9627 * convenient to do the pseudo-header checksum in software
9628 * as Linux does that on transmit for us in all cases.
9629 */
9630 tp->tg3_flags |= TG3_FLAG_NO_TX_PSEUDO_CSUM;
9631 tp->tg3_flags &= ~TG3_FLAG_NO_RX_PSEUDO_CSUM;
9632
9633 /* Derive initial jumbo mode from MTU assigned in
9634 * ether_setup() via the alloc_etherdev() call
9635 */
0f893dc6 9636 if (tp->dev->mtu > ETH_DATA_LEN &&
a4e2b347 9637 !(tp->tg3_flags2 & TG3_FLG2_5780_CLASS))
0f893dc6 9638 tp->tg3_flags |= TG3_FLAG_JUMBO_RING_ENABLE;
1da177e4
LT
9639
9640 /* Determine WakeOnLan speed to use. */
9641 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
9642 tp->pci_chip_rev_id == CHIPREV_ID_5701_A0 ||
9643 tp->pci_chip_rev_id == CHIPREV_ID_5701_B0 ||
9644 tp->pci_chip_rev_id == CHIPREV_ID_5701_B2) {
9645 tp->tg3_flags &= ~(TG3_FLAG_WOL_SPEED_100MB);
9646 } else {
9647 tp->tg3_flags |= TG3_FLAG_WOL_SPEED_100MB;
9648 }
9649
9650 /* A few boards don't want Ethernet@WireSpeed phy feature */
9651 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700) ||
9652 ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705) &&
9653 (tp->pci_chip_rev_id != CHIPREV_ID_5705_A0) &&
747e8f8b
MC
9654 (tp->pci_chip_rev_id != CHIPREV_ID_5705_A1)) ||
9655 (tp->tg3_flags2 & TG3_FLG2_ANY_SERDES))
1da177e4
LT
9656 tp->tg3_flags2 |= TG3_FLG2_NO_ETH_WIRE_SPEED;
9657
9658 if (GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5703_AX ||
9659 GET_CHIP_REV(tp->pci_chip_rev_id) == CHIPREV_5704_AX)
9660 tp->tg3_flags2 |= TG3_FLG2_PHY_ADC_BUG;
9661 if (tp->pci_chip_rev_id == CHIPREV_ID_5704_A0)
9662 tp->tg3_flags2 |= TG3_FLG2_PHY_5704_A0_BUG;
9663
bb7064dc 9664 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS)
1da177e4
LT
9665 tp->tg3_flags2 |= TG3_FLG2_PHY_BER_BUG;
9666
1da177e4 9667 tp->coalesce_mode = 0;
1da177e4
LT
9668 if (GET_CHIP_REV(tp->pci_chip_rev_id) != CHIPREV_5700_AX &&
9669 GET_CHIP_REV(tp->pci_chip_rev_id) != CHIPREV_5700_BX)
9670 tp->coalesce_mode |= HOSTCC_MODE_32BYTE;
9671
9672 /* Initialize MAC MI mode, polling disabled. */
9673 tw32_f(MAC_MI_MODE, tp->mi_mode);
9674 udelay(80);
9675
9676 /* Initialize data/descriptor byte/word swapping. */
9677 val = tr32(GRC_MODE);
9678 val &= GRC_MODE_HOST_STACKUP;
9679 tw32(GRC_MODE, val | tp->grc_mode);
9680
9681 tg3_switch_clocks(tp);
9682
9683 /* Clear this out for sanity. */
9684 tw32(TG3PCI_MEM_WIN_BASE_ADDR, 0);
9685
9686 pci_read_config_dword(tp->pdev, TG3PCI_PCISTATE,
9687 &pci_state_reg);
9688 if ((pci_state_reg & PCISTATE_CONV_PCI_MODE) == 0 &&
9689 (tp->tg3_flags & TG3_FLAG_PCIX_TARGET_HWBUG) == 0) {
9690 u32 chiprevid = GET_CHIP_REV_ID(tp->misc_host_ctrl);
9691
9692 if (chiprevid == CHIPREV_ID_5701_A0 ||
9693 chiprevid == CHIPREV_ID_5701_B0 ||
9694 chiprevid == CHIPREV_ID_5701_B2 ||
9695 chiprevid == CHIPREV_ID_5701_B5) {
9696 void __iomem *sram_base;
9697
9698 /* Write some dummy words into the SRAM status block
9699 * area, see if it reads back correctly. If the return
9700 * value is bad, force enable the PCIX workaround.
9701 */
9702 sram_base = tp->regs + NIC_SRAM_WIN_BASE + NIC_SRAM_STATS_BLK;
9703
9704 writel(0x00000000, sram_base);
9705 writel(0x00000000, sram_base + 4);
9706 writel(0xffffffff, sram_base + 4);
9707 if (readl(sram_base) != 0x00000000)
9708 tp->tg3_flags |= TG3_FLAG_PCIX_TARGET_HWBUG;
9709 }
9710 }
9711
9712 udelay(50);
9713 tg3_nvram_init(tp);
9714
9715 grc_misc_cfg = tr32(GRC_MISC_CFG);
9716 grc_misc_cfg &= GRC_MISC_CFG_BOARD_ID_MASK;
9717
9718 /* Broadcom's driver says that CIOBE multisplit has a bug */
9719#if 0
9720 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704 &&
9721 grc_misc_cfg == GRC_MISC_CFG_BOARD_ID_5704CIOBE) {
9722 tp->tg3_flags |= TG3_FLAG_SPLIT_MODE;
9723 tp->split_mode_max_reqs = SPLIT_MODE_5704_MAX_REQ;
9724 }
9725#endif
9726 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705 &&
9727 (grc_misc_cfg == GRC_MISC_CFG_BOARD_ID_5788 ||
9728 grc_misc_cfg == GRC_MISC_CFG_BOARD_ID_5788M))
9729 tp->tg3_flags2 |= TG3_FLG2_IS_5788;
9730
fac9b83e
DM
9731 if (!(tp->tg3_flags2 & TG3_FLG2_IS_5788) &&
9732 (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700))
9733 tp->tg3_flags |= TG3_FLAG_TAGGED_STATUS;
9734 if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS) {
9735 tp->coalesce_mode |= (HOSTCC_MODE_CLRTICK_RXBD |
9736 HOSTCC_MODE_CLRTICK_TXBD);
9737
9738 tp->misc_host_ctrl |= MISC_HOST_CTRL_TAGGED_STATUS;
9739 pci_write_config_dword(tp->pdev, TG3PCI_MISC_HOST_CTRL,
9740 tp->misc_host_ctrl);
9741 }
9742
1da177e4
LT
9743 /* these are limited to 10/100 only */
9744 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 &&
9745 (grc_misc_cfg == 0x8000 || grc_misc_cfg == 0x4000)) ||
9746 (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705 &&
9747 tp->pdev->vendor == PCI_VENDOR_ID_BROADCOM &&
9748 (tp->pdev->device == PCI_DEVICE_ID_TIGON3_5901 ||
9749 tp->pdev->device == PCI_DEVICE_ID_TIGON3_5901_2 ||
9750 tp->pdev->device == PCI_DEVICE_ID_TIGON3_5705F)) ||
9751 (tp->pdev->vendor == PCI_VENDOR_ID_BROADCOM &&
9752 (tp->pdev->device == PCI_DEVICE_ID_TIGON3_5751F ||
9753 tp->pdev->device == PCI_DEVICE_ID_TIGON3_5753F)))
9754 tp->tg3_flags |= TG3_FLAG_10_100_ONLY;
9755
9756 err = tg3_phy_probe(tp);
9757 if (err) {
9758 printk(KERN_ERR PFX "(%s) phy probe failed, err %d\n",
9759 pci_name(tp->pdev), err);
9760 /* ... but do not return immediately ... */
9761 }
9762
9763 tg3_read_partno(tp);
9764
9765 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES) {
9766 tp->tg3_flags &= ~TG3_FLAG_USE_MI_INTERRUPT;
9767 } else {
9768 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700)
9769 tp->tg3_flags |= TG3_FLAG_USE_MI_INTERRUPT;
9770 else
9771 tp->tg3_flags &= ~TG3_FLAG_USE_MI_INTERRUPT;
9772 }
9773
9774 /* 5700 {AX,BX} chips have a broken status block link
9775 * change bit implementation, so we must use the
9776 * status register in those cases.
9777 */
9778 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700)
9779 tp->tg3_flags |= TG3_FLAG_USE_LINKCHG_REG;
9780 else
9781 tp->tg3_flags &= ~TG3_FLAG_USE_LINKCHG_REG;
9782
9783 /* The led_ctrl is set during tg3_phy_probe, here we might
9784 * have to force the link status polling mechanism based
9785 * upon subsystem IDs.
9786 */
9787 if (tp->pdev->subsystem_vendor == PCI_VENDOR_ID_DELL &&
9788 !(tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)) {
9789 tp->tg3_flags |= (TG3_FLAG_USE_MI_INTERRUPT |
9790 TG3_FLAG_USE_LINKCHG_REG);
9791 }
9792
9793 /* For all SERDES we poll the MAC status register. */
9794 if (tp->tg3_flags2 & TG3_FLG2_PHY_SERDES)
9795 tp->tg3_flags |= TG3_FLAG_POLL_SERDES;
9796 else
9797 tp->tg3_flags &= ~TG3_FLAG_POLL_SERDES;
9798
1da177e4
LT
9799 /* It seems all chips can get confused if TX buffers
9800 * straddle the 4GB address boundary in some cases.
9801 */
9802 tp->dev->hard_start_xmit = tg3_start_xmit;
9803
9804 tp->rx_offset = 2;
9805 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701 &&
9806 (tp->tg3_flags & TG3_FLAG_PCIX_MODE) != 0)
9807 tp->rx_offset = 0;
9808
9809 /* By default, disable wake-on-lan. User can change this
9810 * using ETHTOOL_SWOL.
9811 */
9812 tp->tg3_flags &= ~TG3_FLAG_WOL_ENABLE;
9813
9814 return err;
9815}
9816
9817#ifdef CONFIG_SPARC64
9818static int __devinit tg3_get_macaddr_sparc(struct tg3 *tp)
9819{
9820 struct net_device *dev = tp->dev;
9821 struct pci_dev *pdev = tp->pdev;
9822 struct pcidev_cookie *pcp = pdev->sysdata;
9823
9824 if (pcp != NULL) {
9825 int node = pcp->prom_node;
9826
9827 if (prom_getproplen(node, "local-mac-address") == 6) {
9828 prom_getproperty(node, "local-mac-address",
9829 dev->dev_addr, 6);
2ff43697 9830 memcpy(dev->perm_addr, dev->dev_addr, 6);
1da177e4
LT
9831 return 0;
9832 }
9833 }
9834 return -ENODEV;
9835}
9836
9837static int __devinit tg3_get_default_macaddr_sparc(struct tg3 *tp)
9838{
9839 struct net_device *dev = tp->dev;
9840
9841 memcpy(dev->dev_addr, idprom->id_ethaddr, 6);
2ff43697 9842 memcpy(dev->perm_addr, idprom->id_ethaddr, 6);
1da177e4
LT
9843 return 0;
9844}
9845#endif
9846
9847static int __devinit tg3_get_device_address(struct tg3 *tp)
9848{
9849 struct net_device *dev = tp->dev;
9850 u32 hi, lo, mac_offset;
9851
9852#ifdef CONFIG_SPARC64
9853 if (!tg3_get_macaddr_sparc(tp))
9854 return 0;
9855#endif
9856
9857 mac_offset = 0x7c;
4cf78e4f
MC
9858 if ((GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704 &&
9859 !(tp->tg3_flags & TG3_FLG2_SUN_570X)) ||
a4e2b347 9860 (tp->tg3_flags2 & TG3_FLG2_5780_CLASS)) {
1da177e4
LT
9861 if (tr32(TG3PCI_DUAL_MAC_CTRL) & DUAL_MAC_CTRL_ID)
9862 mac_offset = 0xcc;
9863 if (tg3_nvram_lock(tp))
9864 tw32_f(NVRAM_CMD, NVRAM_CMD_RESET);
9865 else
9866 tg3_nvram_unlock(tp);
9867 }
9868
9869 /* First try to get it from MAC address mailbox. */
9870 tg3_read_mem(tp, NIC_SRAM_MAC_ADDR_HIGH_MBOX, &hi);
9871 if ((hi >> 16) == 0x484b) {
9872 dev->dev_addr[0] = (hi >> 8) & 0xff;
9873 dev->dev_addr[1] = (hi >> 0) & 0xff;
9874
9875 tg3_read_mem(tp, NIC_SRAM_MAC_ADDR_LOW_MBOX, &lo);
9876 dev->dev_addr[2] = (lo >> 24) & 0xff;
9877 dev->dev_addr[3] = (lo >> 16) & 0xff;
9878 dev->dev_addr[4] = (lo >> 8) & 0xff;
9879 dev->dev_addr[5] = (lo >> 0) & 0xff;
9880 }
9881 /* Next, try NVRAM. */
9882 else if (!(tp->tg3_flags & TG3_FLG2_SUN_570X) &&
9883 !tg3_nvram_read(tp, mac_offset + 0, &hi) &&
9884 !tg3_nvram_read(tp, mac_offset + 4, &lo)) {
9885 dev->dev_addr[0] = ((hi >> 16) & 0xff);
9886 dev->dev_addr[1] = ((hi >> 24) & 0xff);
9887 dev->dev_addr[2] = ((lo >> 0) & 0xff);
9888 dev->dev_addr[3] = ((lo >> 8) & 0xff);
9889 dev->dev_addr[4] = ((lo >> 16) & 0xff);
9890 dev->dev_addr[5] = ((lo >> 24) & 0xff);
9891 }
9892 /* Finally just fetch it out of the MAC control regs. */
9893 else {
9894 hi = tr32(MAC_ADDR_0_HIGH);
9895 lo = tr32(MAC_ADDR_0_LOW);
9896
9897 dev->dev_addr[5] = lo & 0xff;
9898 dev->dev_addr[4] = (lo >> 8) & 0xff;
9899 dev->dev_addr[3] = (lo >> 16) & 0xff;
9900 dev->dev_addr[2] = (lo >> 24) & 0xff;
9901 dev->dev_addr[1] = hi & 0xff;
9902 dev->dev_addr[0] = (hi >> 8) & 0xff;
9903 }
9904
9905 if (!is_valid_ether_addr(&dev->dev_addr[0])) {
9906#ifdef CONFIG_SPARC64
9907 if (!tg3_get_default_macaddr_sparc(tp))
9908 return 0;
9909#endif
9910 return -EINVAL;
9911 }
2ff43697 9912 memcpy(dev->perm_addr, dev->dev_addr, dev->addr_len);
1da177e4
LT
9913 return 0;
9914}
9915
59e6b434
DM
9916#define BOUNDARY_SINGLE_CACHELINE 1
9917#define BOUNDARY_MULTI_CACHELINE 2
9918
9919static u32 __devinit tg3_calc_dma_bndry(struct tg3 *tp, u32 val)
9920{
9921 int cacheline_size;
9922 u8 byte;
9923 int goal;
9924
9925 pci_read_config_byte(tp->pdev, PCI_CACHE_LINE_SIZE, &byte);
9926 if (byte == 0)
9927 cacheline_size = 1024;
9928 else
9929 cacheline_size = (int) byte * 4;
9930
9931 /* On 5703 and later chips, the boundary bits have no
9932 * effect.
9933 */
9934 if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700 &&
9935 GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5701 &&
9936 !(tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS))
9937 goto out;
9938
9939#if defined(CONFIG_PPC64) || defined(CONFIG_IA64) || defined(CONFIG_PARISC)
9940 goal = BOUNDARY_MULTI_CACHELINE;
9941#else
9942#if defined(CONFIG_SPARC64) || defined(CONFIG_ALPHA)
9943 goal = BOUNDARY_SINGLE_CACHELINE;
9944#else
9945 goal = 0;
9946#endif
9947#endif
9948
9949 if (!goal)
9950 goto out;
9951
9952 /* PCI controllers on most RISC systems tend to disconnect
9953 * when a device tries to burst across a cache-line boundary.
9954 * Therefore, letting tg3 do so just wastes PCI bandwidth.
9955 *
9956 * Unfortunately, for PCI-E there are only limited
9957 * write-side controls for this, and thus for reads
9958 * we will still get the disconnects. We'll also waste
9959 * these PCI cycles for both read and write for chips
9960 * other than 5700 and 5701 which do not implement the
9961 * boundary bits.
9962 */
9963 if ((tp->tg3_flags & TG3_FLAG_PCIX_MODE) &&
9964 !(tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS)) {
9965 switch (cacheline_size) {
9966 case 16:
9967 case 32:
9968 case 64:
9969 case 128:
9970 if (goal == BOUNDARY_SINGLE_CACHELINE) {
9971 val |= (DMA_RWCTRL_READ_BNDRY_128_PCIX |
9972 DMA_RWCTRL_WRITE_BNDRY_128_PCIX);
9973 } else {
9974 val |= (DMA_RWCTRL_READ_BNDRY_384_PCIX |
9975 DMA_RWCTRL_WRITE_BNDRY_384_PCIX);
9976 }
9977 break;
9978
9979 case 256:
9980 val |= (DMA_RWCTRL_READ_BNDRY_256_PCIX |
9981 DMA_RWCTRL_WRITE_BNDRY_256_PCIX);
9982 break;
9983
9984 default:
9985 val |= (DMA_RWCTRL_READ_BNDRY_384_PCIX |
9986 DMA_RWCTRL_WRITE_BNDRY_384_PCIX);
9987 break;
9988 };
9989 } else if (tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) {
9990 switch (cacheline_size) {
9991 case 16:
9992 case 32:
9993 case 64:
9994 if (goal == BOUNDARY_SINGLE_CACHELINE) {
9995 val &= ~DMA_RWCTRL_WRITE_BNDRY_DISAB_PCIE;
9996 val |= DMA_RWCTRL_WRITE_BNDRY_64_PCIE;
9997 break;
9998 }
9999 /* fallthrough */
10000 case 128:
10001 default:
10002 val &= ~DMA_RWCTRL_WRITE_BNDRY_DISAB_PCIE;
10003 val |= DMA_RWCTRL_WRITE_BNDRY_128_PCIE;
10004 break;
10005 };
10006 } else {
10007 switch (cacheline_size) {
10008 case 16:
10009 if (goal == BOUNDARY_SINGLE_CACHELINE) {
10010 val |= (DMA_RWCTRL_READ_BNDRY_16 |
10011 DMA_RWCTRL_WRITE_BNDRY_16);
10012 break;
10013 }
10014 /* fallthrough */
10015 case 32:
10016 if (goal == BOUNDARY_SINGLE_CACHELINE) {
10017 val |= (DMA_RWCTRL_READ_BNDRY_32 |
10018 DMA_RWCTRL_WRITE_BNDRY_32);
10019 break;
10020 }
10021 /* fallthrough */
10022 case 64:
10023 if (goal == BOUNDARY_SINGLE_CACHELINE) {
10024 val |= (DMA_RWCTRL_READ_BNDRY_64 |
10025 DMA_RWCTRL_WRITE_BNDRY_64);
10026 break;
10027 }
10028 /* fallthrough */
10029 case 128:
10030 if (goal == BOUNDARY_SINGLE_CACHELINE) {
10031 val |= (DMA_RWCTRL_READ_BNDRY_128 |
10032 DMA_RWCTRL_WRITE_BNDRY_128);
10033 break;
10034 }
10035 /* fallthrough */
10036 case 256:
10037 val |= (DMA_RWCTRL_READ_BNDRY_256 |
10038 DMA_RWCTRL_WRITE_BNDRY_256);
10039 break;
10040 case 512:
10041 val |= (DMA_RWCTRL_READ_BNDRY_512 |
10042 DMA_RWCTRL_WRITE_BNDRY_512);
10043 break;
10044 case 1024:
10045 default:
10046 val |= (DMA_RWCTRL_READ_BNDRY_1024 |
10047 DMA_RWCTRL_WRITE_BNDRY_1024);
10048 break;
10049 };
10050 }
10051
10052out:
10053 return val;
10054}
10055
1da177e4
LT
10056static int __devinit tg3_do_test_dma(struct tg3 *tp, u32 *buf, dma_addr_t buf_dma, int size, int to_device)
10057{
10058 struct tg3_internal_buffer_desc test_desc;
10059 u32 sram_dma_descs;
10060 int i, ret;
10061
10062 sram_dma_descs = NIC_SRAM_DMA_DESC_POOL_BASE;
10063
10064 tw32(FTQ_RCVBD_COMP_FIFO_ENQDEQ, 0);
10065 tw32(FTQ_RCVDATA_COMP_FIFO_ENQDEQ, 0);
10066 tw32(RDMAC_STATUS, 0);
10067 tw32(WDMAC_STATUS, 0);
10068
10069 tw32(BUFMGR_MODE, 0);
10070 tw32(FTQ_RESET, 0);
10071
10072 test_desc.addr_hi = ((u64) buf_dma) >> 32;
10073 test_desc.addr_lo = buf_dma & 0xffffffff;
10074 test_desc.nic_mbuf = 0x00002100;
10075 test_desc.len = size;
10076
10077 /*
10078 * HP ZX1 was seeing test failures for 5701 cards running at 33Mhz
10079 * the *second* time the tg3 driver was getting loaded after an
10080 * initial scan.
10081 *
10082 * Broadcom tells me:
10083 * ...the DMA engine is connected to the GRC block and a DMA
10084 * reset may affect the GRC block in some unpredictable way...
10085 * The behavior of resets to individual blocks has not been tested.
10086 *
10087 * Broadcom noted the GRC reset will also reset all sub-components.
10088 */
10089 if (to_device) {
10090 test_desc.cqid_sqid = (13 << 8) | 2;
10091
10092 tw32_f(RDMAC_MODE, RDMAC_MODE_ENABLE);
10093 udelay(40);
10094 } else {
10095 test_desc.cqid_sqid = (16 << 8) | 7;
10096
10097 tw32_f(WDMAC_MODE, WDMAC_MODE_ENABLE);
10098 udelay(40);
10099 }
10100 test_desc.flags = 0x00000005;
10101
10102 for (i = 0; i < (sizeof(test_desc) / sizeof(u32)); i++) {
10103 u32 val;
10104
10105 val = *(((u32 *)&test_desc) + i);
10106 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_BASE_ADDR,
10107 sram_dma_descs + (i * sizeof(u32)));
10108 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_DATA, val);
10109 }
10110 pci_write_config_dword(tp->pdev, TG3PCI_MEM_WIN_BASE_ADDR, 0);
10111
10112 if (to_device) {
10113 tw32(FTQ_DMA_HIGH_READ_FIFO_ENQDEQ, sram_dma_descs);
10114 } else {
10115 tw32(FTQ_DMA_HIGH_WRITE_FIFO_ENQDEQ, sram_dma_descs);
10116 }
10117
10118 ret = -ENODEV;
10119 for (i = 0; i < 40; i++) {
10120 u32 val;
10121
10122 if (to_device)
10123 val = tr32(FTQ_RCVBD_COMP_FIFO_ENQDEQ);
10124 else
10125 val = tr32(FTQ_RCVDATA_COMP_FIFO_ENQDEQ);
10126 if ((val & 0xffff) == sram_dma_descs) {
10127 ret = 0;
10128 break;
10129 }
10130
10131 udelay(100);
10132 }
10133
10134 return ret;
10135}
10136
ded7340d 10137#define TEST_BUFFER_SIZE 0x2000
1da177e4
LT
10138
10139static int __devinit tg3_test_dma(struct tg3 *tp)
10140{
10141 dma_addr_t buf_dma;
59e6b434 10142 u32 *buf, saved_dma_rwctrl;
1da177e4
LT
10143 int ret;
10144
10145 buf = pci_alloc_consistent(tp->pdev, TEST_BUFFER_SIZE, &buf_dma);
10146 if (!buf) {
10147 ret = -ENOMEM;
10148 goto out_nofree;
10149 }
10150
10151 tp->dma_rwctrl = ((0x7 << DMA_RWCTRL_PCI_WRITE_CMD_SHIFT) |
10152 (0x6 << DMA_RWCTRL_PCI_READ_CMD_SHIFT));
10153
59e6b434 10154 tp->dma_rwctrl = tg3_calc_dma_bndry(tp, tp->dma_rwctrl);
1da177e4
LT
10155
10156 if (tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) {
10157 /* DMA read watermark not used on PCIE */
10158 tp->dma_rwctrl |= 0x00180000;
10159 } else if (!(tp->tg3_flags & TG3_FLAG_PCIX_MODE)) {
85e94ced
MC
10160 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5705 ||
10161 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5750)
1da177e4
LT
10162 tp->dma_rwctrl |= 0x003f0000;
10163 else
10164 tp->dma_rwctrl |= 0x003f000f;
10165 } else {
10166 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 ||
10167 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704) {
10168 u32 ccval = (tr32(TG3PCI_CLOCK_CTRL) & 0x1f);
10169
10170 if (ccval == 0x6 || ccval == 0x7)
10171 tp->dma_rwctrl |= DMA_RWCTRL_ONE_DMA;
10172
59e6b434 10173 /* Set bit 23 to enable PCIX hw bug fix */
1da177e4 10174 tp->dma_rwctrl |= 0x009f0000;
4cf78e4f
MC
10175 } else if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5780) {
10176 /* 5780 always in PCIX mode */
10177 tp->dma_rwctrl |= 0x00144000;
a4e2b347
MC
10178 } else if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5714) {
10179 /* 5714 always in PCIX mode */
10180 tp->dma_rwctrl |= 0x00148000;
1da177e4
LT
10181 } else {
10182 tp->dma_rwctrl |= 0x001b000f;
10183 }
10184 }
10185
10186 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5703 ||
10187 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704)
10188 tp->dma_rwctrl &= 0xfffffff0;
10189
10190 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
10191 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701) {
10192 /* Remove this if it causes problems for some boards. */
10193 tp->dma_rwctrl |= DMA_RWCTRL_USE_MEM_READ_MULT;
10194
10195 /* On 5700/5701 chips, we need to set this bit.
10196 * Otherwise the chip will issue cacheline transactions
10197 * to streamable DMA memory with not all the byte
10198 * enables turned on. This is an error on several
10199 * RISC PCI controllers, in particular sparc64.
10200 *
10201 * On 5703/5704 chips, this bit has been reassigned
10202 * a different meaning. In particular, it is used
10203 * on those chips to enable a PCI-X workaround.
10204 */
10205 tp->dma_rwctrl |= DMA_RWCTRL_ASSERT_ALL_BE;
10206 }
10207
10208 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl);
10209
10210#if 0
10211 /* Unneeded, already done by tg3_get_invariants. */
10212 tg3_switch_clocks(tp);
10213#endif
10214
10215 ret = 0;
10216 if (GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5700 &&
10217 GET_ASIC_REV(tp->pci_chip_rev_id) != ASIC_REV_5701)
10218 goto out;
10219
59e6b434
DM
10220 /* It is best to perform DMA test with maximum write burst size
10221 * to expose the 5700/5701 write DMA bug.
10222 */
10223 saved_dma_rwctrl = tp->dma_rwctrl;
10224 tp->dma_rwctrl &= ~DMA_RWCTRL_WRITE_BNDRY_MASK;
10225 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl);
10226
1da177e4
LT
10227 while (1) {
10228 u32 *p = buf, i;
10229
10230 for (i = 0; i < TEST_BUFFER_SIZE / sizeof(u32); i++)
10231 p[i] = i;
10232
10233 /* Send the buffer to the chip. */
10234 ret = tg3_do_test_dma(tp, buf, buf_dma, TEST_BUFFER_SIZE, 1);
10235 if (ret) {
10236 printk(KERN_ERR "tg3_test_dma() Write the buffer failed %d\n", ret);
10237 break;
10238 }
10239
10240#if 0
10241 /* validate data reached card RAM correctly. */
10242 for (i = 0; i < TEST_BUFFER_SIZE / sizeof(u32); i++) {
10243 u32 val;
10244 tg3_read_mem(tp, 0x2100 + (i*4), &val);
10245 if (le32_to_cpu(val) != p[i]) {
10246 printk(KERN_ERR " tg3_test_dma() Card buffer corrupted on write! (%d != %d)\n", val, i);
10247 /* ret = -ENODEV here? */
10248 }
10249 p[i] = 0;
10250 }
10251#endif
10252 /* Now read it back. */
10253 ret = tg3_do_test_dma(tp, buf, buf_dma, TEST_BUFFER_SIZE, 0);
10254 if (ret) {
10255 printk(KERN_ERR "tg3_test_dma() Read the buffer failed %d\n", ret);
10256
10257 break;
10258 }
10259
10260 /* Verify it. */
10261 for (i = 0; i < TEST_BUFFER_SIZE / sizeof(u32); i++) {
10262 if (p[i] == i)
10263 continue;
10264
59e6b434
DM
10265 if ((tp->dma_rwctrl & DMA_RWCTRL_WRITE_BNDRY_MASK) !=
10266 DMA_RWCTRL_WRITE_BNDRY_16) {
10267 tp->dma_rwctrl &= ~DMA_RWCTRL_WRITE_BNDRY_MASK;
1da177e4
LT
10268 tp->dma_rwctrl |= DMA_RWCTRL_WRITE_BNDRY_16;
10269 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl);
10270 break;
10271 } else {
10272 printk(KERN_ERR "tg3_test_dma() buffer corrupted on read back! (%d != %d)\n", p[i], i);
10273 ret = -ENODEV;
10274 goto out;
10275 }
10276 }
10277
10278 if (i == (TEST_BUFFER_SIZE / sizeof(u32))) {
10279 /* Success. */
10280 ret = 0;
10281 break;
10282 }
10283 }
59e6b434
DM
10284 if ((tp->dma_rwctrl & DMA_RWCTRL_WRITE_BNDRY_MASK) !=
10285 DMA_RWCTRL_WRITE_BNDRY_16) {
6d1cfbab
MC
10286 static struct pci_device_id dma_wait_state_chipsets[] = {
10287 { PCI_DEVICE(PCI_VENDOR_ID_APPLE,
10288 PCI_DEVICE_ID_APPLE_UNI_N_PCI15) },
10289 { },
10290 };
10291
59e6b434 10292 /* DMA test passed without adjusting DMA boundary,
6d1cfbab
MC
10293 * now look for chipsets that are known to expose the
10294 * DMA bug without failing the test.
59e6b434 10295 */
6d1cfbab
MC
10296 if (pci_dev_present(dma_wait_state_chipsets)) {
10297 tp->dma_rwctrl &= ~DMA_RWCTRL_WRITE_BNDRY_MASK;
10298 tp->dma_rwctrl |= DMA_RWCTRL_WRITE_BNDRY_16;
10299 }
10300 else
10301 /* Safe to use the calculated DMA boundary. */
10302 tp->dma_rwctrl = saved_dma_rwctrl;
10303
59e6b434
DM
10304 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl);
10305 }
1da177e4
LT
10306
10307out:
10308 pci_free_consistent(tp->pdev, TEST_BUFFER_SIZE, buf, buf_dma);
10309out_nofree:
10310 return ret;
10311}
10312
10313static void __devinit tg3_init_link_config(struct tg3 *tp)
10314{
10315 tp->link_config.advertising =
10316 (ADVERTISED_10baseT_Half | ADVERTISED_10baseT_Full |
10317 ADVERTISED_100baseT_Half | ADVERTISED_100baseT_Full |
10318 ADVERTISED_1000baseT_Half | ADVERTISED_1000baseT_Full |
10319 ADVERTISED_Autoneg | ADVERTISED_MII);
10320 tp->link_config.speed = SPEED_INVALID;
10321 tp->link_config.duplex = DUPLEX_INVALID;
10322 tp->link_config.autoneg = AUTONEG_ENABLE;
10323 netif_carrier_off(tp->dev);
10324 tp->link_config.active_speed = SPEED_INVALID;
10325 tp->link_config.active_duplex = DUPLEX_INVALID;
10326 tp->link_config.phy_is_low_power = 0;
10327 tp->link_config.orig_speed = SPEED_INVALID;
10328 tp->link_config.orig_duplex = DUPLEX_INVALID;
10329 tp->link_config.orig_autoneg = AUTONEG_INVALID;
10330}
10331
10332static void __devinit tg3_init_bufmgr_config(struct tg3 *tp)
10333{
fdfec172
MC
10334 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS) {
10335 tp->bufmgr_config.mbuf_read_dma_low_water =
10336 DEFAULT_MB_RDMA_LOW_WATER_5705;
10337 tp->bufmgr_config.mbuf_mac_rx_low_water =
10338 DEFAULT_MB_MACRX_LOW_WATER_5705;
10339 tp->bufmgr_config.mbuf_high_water =
10340 DEFAULT_MB_HIGH_WATER_5705;
10341
10342 tp->bufmgr_config.mbuf_read_dma_low_water_jumbo =
10343 DEFAULT_MB_RDMA_LOW_WATER_JUMBO_5780;
10344 tp->bufmgr_config.mbuf_mac_rx_low_water_jumbo =
10345 DEFAULT_MB_MACRX_LOW_WATER_JUMBO_5780;
10346 tp->bufmgr_config.mbuf_high_water_jumbo =
10347 DEFAULT_MB_HIGH_WATER_JUMBO_5780;
10348 } else {
10349 tp->bufmgr_config.mbuf_read_dma_low_water =
10350 DEFAULT_MB_RDMA_LOW_WATER;
10351 tp->bufmgr_config.mbuf_mac_rx_low_water =
10352 DEFAULT_MB_MACRX_LOW_WATER;
10353 tp->bufmgr_config.mbuf_high_water =
10354 DEFAULT_MB_HIGH_WATER;
10355
10356 tp->bufmgr_config.mbuf_read_dma_low_water_jumbo =
10357 DEFAULT_MB_RDMA_LOW_WATER_JUMBO;
10358 tp->bufmgr_config.mbuf_mac_rx_low_water_jumbo =
10359 DEFAULT_MB_MACRX_LOW_WATER_JUMBO;
10360 tp->bufmgr_config.mbuf_high_water_jumbo =
10361 DEFAULT_MB_HIGH_WATER_JUMBO;
10362 }
1da177e4
LT
10363
10364 tp->bufmgr_config.dma_low_water = DEFAULT_DMA_LOW_WATER;
10365 tp->bufmgr_config.dma_high_water = DEFAULT_DMA_HIGH_WATER;
10366}
10367
10368static char * __devinit tg3_phy_string(struct tg3 *tp)
10369{
10370 switch (tp->phy_id & PHY_ID_MASK) {
10371 case PHY_ID_BCM5400: return "5400";
10372 case PHY_ID_BCM5401: return "5401";
10373 case PHY_ID_BCM5411: return "5411";
10374 case PHY_ID_BCM5701: return "5701";
10375 case PHY_ID_BCM5703: return "5703";
10376 case PHY_ID_BCM5704: return "5704";
10377 case PHY_ID_BCM5705: return "5705";
10378 case PHY_ID_BCM5750: return "5750";
85e94ced 10379 case PHY_ID_BCM5752: return "5752";
a4e2b347 10380 case PHY_ID_BCM5714: return "5714";
4cf78e4f 10381 case PHY_ID_BCM5780: return "5780";
1da177e4
LT
10382 case PHY_ID_BCM8002: return "8002/serdes";
10383 case 0: return "serdes";
10384 default: return "unknown";
10385 };
10386}
10387
f9804ddb
MC
10388static char * __devinit tg3_bus_string(struct tg3 *tp, char *str)
10389{
10390 if (tp->tg3_flags2 & TG3_FLG2_PCI_EXPRESS) {
10391 strcpy(str, "PCI Express");
10392 return str;
10393 } else if (tp->tg3_flags & TG3_FLAG_PCIX_MODE) {
10394 u32 clock_ctrl = tr32(TG3PCI_CLOCK_CTRL) & 0x1f;
10395
10396 strcpy(str, "PCIX:");
10397
10398 if ((clock_ctrl == 7) ||
10399 ((tr32(GRC_MISC_CFG) & GRC_MISC_CFG_BOARD_ID_MASK) ==
10400 GRC_MISC_CFG_BOARD_ID_5704CIOBE))
10401 strcat(str, "133MHz");
10402 else if (clock_ctrl == 0)
10403 strcat(str, "33MHz");
10404 else if (clock_ctrl == 2)
10405 strcat(str, "50MHz");
10406 else if (clock_ctrl == 4)
10407 strcat(str, "66MHz");
10408 else if (clock_ctrl == 6)
10409 strcat(str, "100MHz");
10410 else if (clock_ctrl == 7)
10411 strcat(str, "133MHz");
10412 } else {
10413 strcpy(str, "PCI:");
10414 if (tp->tg3_flags & TG3_FLAG_PCI_HIGH_SPEED)
10415 strcat(str, "66MHz");
10416 else
10417 strcat(str, "33MHz");
10418 }
10419 if (tp->tg3_flags & TG3_FLAG_PCI_32BIT)
10420 strcat(str, ":32-bit");
10421 else
10422 strcat(str, ":64-bit");
10423 return str;
10424}
10425
1da177e4
LT
10426static struct pci_dev * __devinit tg3_find_5704_peer(struct tg3 *tp)
10427{
10428 struct pci_dev *peer;
10429 unsigned int func, devnr = tp->pdev->devfn & ~7;
10430
10431 for (func = 0; func < 8; func++) {
10432 peer = pci_get_slot(tp->pdev->bus, devnr | func);
10433 if (peer && peer != tp->pdev)
10434 break;
10435 pci_dev_put(peer);
10436 }
10437 if (!peer || peer == tp->pdev)
10438 BUG();
10439
10440 /*
10441 * We don't need to keep the refcount elevated; there's no way
10442 * to remove one half of this device without removing the other
10443 */
10444 pci_dev_put(peer);
10445
10446 return peer;
10447}
10448
15f9850d
DM
10449static void __devinit tg3_init_coal(struct tg3 *tp)
10450{
10451 struct ethtool_coalesce *ec = &tp->coal;
10452
10453 memset(ec, 0, sizeof(*ec));
10454 ec->cmd = ETHTOOL_GCOALESCE;
10455 ec->rx_coalesce_usecs = LOW_RXCOL_TICKS;
10456 ec->tx_coalesce_usecs = LOW_TXCOL_TICKS;
10457 ec->rx_max_coalesced_frames = LOW_RXMAX_FRAMES;
10458 ec->tx_max_coalesced_frames = LOW_TXMAX_FRAMES;
10459 ec->rx_coalesce_usecs_irq = DEFAULT_RXCOAL_TICK_INT;
10460 ec->tx_coalesce_usecs_irq = DEFAULT_TXCOAL_TICK_INT;
10461 ec->rx_max_coalesced_frames_irq = DEFAULT_RXCOAL_MAXF_INT;
10462 ec->tx_max_coalesced_frames_irq = DEFAULT_TXCOAL_MAXF_INT;
10463 ec->stats_block_coalesce_usecs = DEFAULT_STAT_COAL_TICKS;
10464
10465 if (tp->coalesce_mode & (HOSTCC_MODE_CLRTICK_RXBD |
10466 HOSTCC_MODE_CLRTICK_TXBD)) {
10467 ec->rx_coalesce_usecs = LOW_RXCOL_TICKS_CLRTCKS;
10468 ec->rx_coalesce_usecs_irq = DEFAULT_RXCOAL_TICK_INT_CLRTCKS;
10469 ec->tx_coalesce_usecs = LOW_TXCOL_TICKS_CLRTCKS;
10470 ec->tx_coalesce_usecs_irq = DEFAULT_TXCOAL_TICK_INT_CLRTCKS;
10471 }
d244c892
MC
10472
10473 if (tp->tg3_flags2 & TG3_FLG2_5705_PLUS) {
10474 ec->rx_coalesce_usecs_irq = 0;
10475 ec->tx_coalesce_usecs_irq = 0;
10476 ec->stats_block_coalesce_usecs = 0;
10477 }
15f9850d
DM
10478}
10479
1da177e4
LT
10480static int __devinit tg3_init_one(struct pci_dev *pdev,
10481 const struct pci_device_id *ent)
10482{
10483 static int tg3_version_printed = 0;
10484 unsigned long tg3reg_base, tg3reg_len;
10485 struct net_device *dev;
10486 struct tg3 *tp;
10487 int i, err, pci_using_dac, pm_cap;
f9804ddb 10488 char str[40];
1da177e4
LT
10489
10490 if (tg3_version_printed++ == 0)
10491 printk(KERN_INFO "%s", version);
10492
10493 err = pci_enable_device(pdev);
10494 if (err) {
10495 printk(KERN_ERR PFX "Cannot enable PCI device, "
10496 "aborting.\n");
10497 return err;
10498 }
10499
10500 if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) {
10501 printk(KERN_ERR PFX "Cannot find proper PCI device "
10502 "base address, aborting.\n");
10503 err = -ENODEV;
10504 goto err_out_disable_pdev;
10505 }
10506
10507 err = pci_request_regions(pdev, DRV_MODULE_NAME);
10508 if (err) {
10509 printk(KERN_ERR PFX "Cannot obtain PCI resources, "
10510 "aborting.\n");
10511 goto err_out_disable_pdev;
10512 }
10513
10514 pci_set_master(pdev);
10515
10516 /* Find power-management capability. */
10517 pm_cap = pci_find_capability(pdev, PCI_CAP_ID_PM);
10518 if (pm_cap == 0) {
10519 printk(KERN_ERR PFX "Cannot find PowerManagement capability, "
10520 "aborting.\n");
10521 err = -EIO;
10522 goto err_out_free_res;
10523 }
10524
10525 /* Configure DMA attributes. */
f9a5f7d3 10526 err = pci_set_dma_mask(pdev, DMA_64BIT_MASK);
1da177e4
LT
10527 if (!err) {
10528 pci_using_dac = 1;
f9a5f7d3 10529 err = pci_set_consistent_dma_mask(pdev, DMA_64BIT_MASK);
1da177e4
LT
10530 if (err < 0) {
10531 printk(KERN_ERR PFX "Unable to obtain 64 bit DMA "
10532 "for consistent allocations\n");
10533 goto err_out_free_res;
10534 }
10535 } else {
f9a5f7d3 10536 err = pci_set_dma_mask(pdev, DMA_32BIT_MASK);
1da177e4
LT
10537 if (err) {
10538 printk(KERN_ERR PFX "No usable DMA configuration, "
10539 "aborting.\n");
10540 goto err_out_free_res;
10541 }
10542 pci_using_dac = 0;
10543 }
10544
10545 tg3reg_base = pci_resource_start(pdev, 0);
10546 tg3reg_len = pci_resource_len(pdev, 0);
10547
10548 dev = alloc_etherdev(sizeof(*tp));
10549 if (!dev) {
10550 printk(KERN_ERR PFX "Etherdev alloc failed, aborting.\n");
10551 err = -ENOMEM;
10552 goto err_out_free_res;
10553 }
10554
10555 SET_MODULE_OWNER(dev);
10556 SET_NETDEV_DEV(dev, &pdev->dev);
10557
10558 if (pci_using_dac)
10559 dev->features |= NETIF_F_HIGHDMA;
10560 dev->features |= NETIF_F_LLTX;
10561#if TG3_VLAN_TAG_USED
10562 dev->features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX;
10563 dev->vlan_rx_register = tg3_vlan_rx_register;
10564 dev->vlan_rx_kill_vid = tg3_vlan_rx_kill_vid;
10565#endif
10566
10567 tp = netdev_priv(dev);
10568 tp->pdev = pdev;
10569 tp->dev = dev;
10570 tp->pm_cap = pm_cap;
10571 tp->mac_mode = TG3_DEF_MAC_MODE;
10572 tp->rx_mode = TG3_DEF_RX_MODE;
10573 tp->tx_mode = TG3_DEF_TX_MODE;
10574 tp->mi_mode = MAC_MI_MODE_BASE;
10575 if (tg3_debug > 0)
10576 tp->msg_enable = tg3_debug;
10577 else
10578 tp->msg_enable = TG3_DEF_MSG_ENABLE;
10579
10580 /* The word/byte swap controls here control register access byte
10581 * swapping. DMA data byte swapping is controlled in the GRC_MODE
10582 * setting below.
10583 */
10584 tp->misc_host_ctrl =
10585 MISC_HOST_CTRL_MASK_PCI_INT |
10586 MISC_HOST_CTRL_WORD_SWAP |
10587 MISC_HOST_CTRL_INDIR_ACCESS |
10588 MISC_HOST_CTRL_PCISTATE_RW;
10589
10590 /* The NONFRM (non-frame) byte/word swap controls take effect
10591 * on descriptor entries, anything which isn't packet data.
10592 *
10593 * The StrongARM chips on the board (one for tx, one for rx)
10594 * are running in big-endian mode.
10595 */
10596 tp->grc_mode = (GRC_MODE_WSWAP_DATA | GRC_MODE_BSWAP_DATA |
10597 GRC_MODE_WSWAP_NONFRM_DATA);
10598#ifdef __BIG_ENDIAN
10599 tp->grc_mode |= GRC_MODE_BSWAP_NONFRM_DATA;
10600#endif
10601 spin_lock_init(&tp->lock);
10602 spin_lock_init(&tp->tx_lock);
10603 spin_lock_init(&tp->indirect_lock);
10604 INIT_WORK(&tp->reset_task, tg3_reset_task, tp);
10605
10606 tp->regs = ioremap_nocache(tg3reg_base, tg3reg_len);
10607 if (tp->regs == 0UL) {
10608 printk(KERN_ERR PFX "Cannot map device registers, "
10609 "aborting.\n");
10610 err = -ENOMEM;
10611 goto err_out_free_dev;
10612 }
10613
10614 tg3_init_link_config(tp);
10615
1da177e4
LT
10616 tp->rx_pending = TG3_DEF_RX_RING_PENDING;
10617 tp->rx_jumbo_pending = TG3_DEF_RX_JUMBO_RING_PENDING;
10618 tp->tx_pending = TG3_DEF_TX_RING_PENDING;
10619
10620 dev->open = tg3_open;
10621 dev->stop = tg3_close;
10622 dev->get_stats = tg3_get_stats;
10623 dev->set_multicast_list = tg3_set_rx_mode;
10624 dev->set_mac_address = tg3_set_mac_addr;
10625 dev->do_ioctl = tg3_ioctl;
10626 dev->tx_timeout = tg3_tx_timeout;
10627 dev->poll = tg3_poll;
10628 dev->ethtool_ops = &tg3_ethtool_ops;
10629 dev->weight = 64;
10630 dev->watchdog_timeo = TG3_TX_TIMEOUT;
10631 dev->change_mtu = tg3_change_mtu;
10632 dev->irq = pdev->irq;
10633#ifdef CONFIG_NET_POLL_CONTROLLER
10634 dev->poll_controller = tg3_poll_controller;
10635#endif
10636
10637 err = tg3_get_invariants(tp);
10638 if (err) {
10639 printk(KERN_ERR PFX "Problem fetching invariants of chip, "
10640 "aborting.\n");
10641 goto err_out_iounmap;
10642 }
10643
fdfec172 10644 tg3_init_bufmgr_config(tp);
1da177e4
LT
10645
10646#if TG3_TSO_SUPPORT != 0
10647 if (tp->tg3_flags2 & TG3_FLG2_HW_TSO) {
10648 tp->tg3_flags2 |= TG3_FLG2_TSO_CAPABLE;
10649 }
10650 else if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5700 ||
10651 GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701 ||
10652 tp->pci_chip_rev_id == CHIPREV_ID_5705_A0 ||
10653 (tp->tg3_flags & TG3_FLAG_ENABLE_ASF) != 0) {
10654 tp->tg3_flags2 &= ~TG3_FLG2_TSO_CAPABLE;
10655 } else {
10656 tp->tg3_flags2 |= TG3_FLG2_TSO_CAPABLE;
10657 }
10658
10659 /* TSO is off by default, user can enable using ethtool. */
10660#if 0
10661 if (tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE)
10662 dev->features |= NETIF_F_TSO;
10663#endif
10664
10665#endif
10666
10667 if (tp->pci_chip_rev_id == CHIPREV_ID_5705_A1 &&
10668 !(tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE) &&
10669 !(tr32(TG3PCI_PCISTATE) & PCISTATE_BUS_SPEED_HIGH)) {
10670 tp->tg3_flags2 |= TG3_FLG2_MAX_RXPEND_64;
10671 tp->rx_pending = 63;
10672 }
10673
10674 if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5704)
10675 tp->pdev_peer = tg3_find_5704_peer(tp);
10676
10677 err = tg3_get_device_address(tp);
10678 if (err) {
10679 printk(KERN_ERR PFX "Could not obtain valid ethernet address, "
10680 "aborting.\n");
10681 goto err_out_iounmap;
10682 }
10683
10684 /*
10685 * Reset chip in case UNDI or EFI driver did not shutdown
10686 * DMA self test will enable WDMAC and we'll see (spurious)
10687 * pending DMA on the PCI bus at that point.
10688 */
10689 if ((tr32(HOSTCC_MODE) & HOSTCC_MODE_ENABLE) ||
10690 (tr32(WDMAC_MODE) & WDMAC_MODE_ENABLE)) {
10691 pci_save_state(tp->pdev);
10692 tw32(MEMARB_MODE, MEMARB_MODE_ENABLE);
944d980e 10693 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
1da177e4
LT
10694 }
10695
10696 err = tg3_test_dma(tp);
10697 if (err) {
10698 printk(KERN_ERR PFX "DMA engine test failed, aborting.\n");
10699 goto err_out_iounmap;
10700 }
10701
10702 /* Tigon3 can do ipv4 only... and some chips have buggy
10703 * checksumming.
10704 */
10705 if ((tp->tg3_flags & TG3_FLAG_BROKEN_CHECKSUMS) == 0) {
10706 dev->features |= NETIF_F_SG | NETIF_F_IP_CSUM;
10707 tp->tg3_flags |= TG3_FLAG_RX_CHECKSUMS;
10708 } else
10709 tp->tg3_flags &= ~TG3_FLAG_RX_CHECKSUMS;
10710
10711 if (tp->tg3_flags2 & TG3_FLG2_IS_5788)
10712 dev->features &= ~NETIF_F_HIGHDMA;
10713
10714 /* flow control autonegotiation is default behavior */
10715 tp->tg3_flags |= TG3_FLAG_PAUSE_AUTONEG;
10716
15f9850d
DM
10717 tg3_init_coal(tp);
10718
7d3f4c97
DM
10719 /* Now that we have fully setup the chip, save away a snapshot
10720 * of the PCI config space. We need to restore this after
10721 * GRC_MISC_CFG core clock resets and some resume events.
10722 */
10723 pci_save_state(tp->pdev);
10724
1da177e4
LT
10725 err = register_netdev(dev);
10726 if (err) {
10727 printk(KERN_ERR PFX "Cannot register net device, "
10728 "aborting.\n");
10729 goto err_out_iounmap;
10730 }
10731
10732 pci_set_drvdata(pdev, dev);
10733
f9804ddb 10734 printk(KERN_INFO "%s: Tigon3 [partno(%s) rev %04x PHY(%s)] (%s) %sBaseT Ethernet ",
1da177e4
LT
10735 dev->name,
10736 tp->board_part_number,
10737 tp->pci_chip_rev_id,
10738 tg3_phy_string(tp),
f9804ddb 10739 tg3_bus_string(tp, str),
1da177e4
LT
10740 (tp->tg3_flags & TG3_FLAG_10_100_ONLY) ? "10/100" : "10/100/1000");
10741
10742 for (i = 0; i < 6; i++)
10743 printk("%2.2x%c", dev->dev_addr[i],
10744 i == 5 ? '\n' : ':');
10745
10746 printk(KERN_INFO "%s: RXcsums[%d] LinkChgREG[%d] "
10747 "MIirq[%d] ASF[%d] Split[%d] WireSpeed[%d] "
10748 "TSOcap[%d] \n",
10749 dev->name,
10750 (tp->tg3_flags & TG3_FLAG_RX_CHECKSUMS) != 0,
10751 (tp->tg3_flags & TG3_FLAG_USE_LINKCHG_REG) != 0,
10752 (tp->tg3_flags & TG3_FLAG_USE_MI_INTERRUPT) != 0,
10753 (tp->tg3_flags & TG3_FLAG_ENABLE_ASF) != 0,
10754 (tp->tg3_flags & TG3_FLAG_SPLIT_MODE) != 0,
10755 (tp->tg3_flags2 & TG3_FLG2_NO_ETH_WIRE_SPEED) == 0,
10756 (tp->tg3_flags2 & TG3_FLG2_TSO_CAPABLE) != 0);
59e6b434
DM
10757 printk(KERN_INFO "%s: dma_rwctrl[%08x]\n",
10758 dev->name, tp->dma_rwctrl);
1da177e4
LT
10759
10760 return 0;
10761
10762err_out_iounmap:
6892914f
MC
10763 if (tp->regs) {
10764 iounmap(tp->regs);
22abe310 10765 tp->regs = NULL;
6892914f 10766 }
1da177e4
LT
10767
10768err_out_free_dev:
10769 free_netdev(dev);
10770
10771err_out_free_res:
10772 pci_release_regions(pdev);
10773
10774err_out_disable_pdev:
10775 pci_disable_device(pdev);
10776 pci_set_drvdata(pdev, NULL);
10777 return err;
10778}
10779
10780static void __devexit tg3_remove_one(struct pci_dev *pdev)
10781{
10782 struct net_device *dev = pci_get_drvdata(pdev);
10783
10784 if (dev) {
10785 struct tg3 *tp = netdev_priv(dev);
10786
10787 unregister_netdev(dev);
6892914f
MC
10788 if (tp->regs) {
10789 iounmap(tp->regs);
22abe310 10790 tp->regs = NULL;
6892914f 10791 }
1da177e4
LT
10792 free_netdev(dev);
10793 pci_release_regions(pdev);
10794 pci_disable_device(pdev);
10795 pci_set_drvdata(pdev, NULL);
10796 }
10797}
10798
10799static int tg3_suspend(struct pci_dev *pdev, pm_message_t state)
10800{
10801 struct net_device *dev = pci_get_drvdata(pdev);
10802 struct tg3 *tp = netdev_priv(dev);
10803 int err;
10804
10805 if (!netif_running(dev))
10806 return 0;
10807
10808 tg3_netif_stop(tp);
10809
10810 del_timer_sync(&tp->timer);
10811
f47c11ee 10812 tg3_full_lock(tp, 1);
1da177e4 10813 tg3_disable_ints(tp);
f47c11ee 10814 tg3_full_unlock(tp);
1da177e4
LT
10815
10816 netif_device_detach(dev);
10817
f47c11ee 10818 tg3_full_lock(tp, 0);
944d980e 10819 tg3_halt(tp, RESET_KIND_SHUTDOWN, 1);
f47c11ee 10820 tg3_full_unlock(tp);
1da177e4
LT
10821
10822 err = tg3_set_power_state(tp, pci_choose_state(pdev, state));
10823 if (err) {
f47c11ee 10824 tg3_full_lock(tp, 0);
1da177e4
LT
10825
10826 tg3_init_hw(tp);
10827
10828 tp->timer.expires = jiffies + tp->timer_offset;
10829 add_timer(&tp->timer);
10830
10831 netif_device_attach(dev);
10832 tg3_netif_start(tp);
10833
f47c11ee 10834 tg3_full_unlock(tp);
1da177e4
LT
10835 }
10836
10837 return err;
10838}
10839
10840static int tg3_resume(struct pci_dev *pdev)
10841{
10842 struct net_device *dev = pci_get_drvdata(pdev);
10843 struct tg3 *tp = netdev_priv(dev);
10844 int err;
10845
10846 if (!netif_running(dev))
10847 return 0;
10848
10849 pci_restore_state(tp->pdev);
10850
10851 err = tg3_set_power_state(tp, 0);
10852 if (err)
10853 return err;
10854
10855 netif_device_attach(dev);
10856
f47c11ee 10857 tg3_full_lock(tp, 0);
1da177e4
LT
10858
10859 tg3_init_hw(tp);
10860
10861 tp->timer.expires = jiffies + tp->timer_offset;
10862 add_timer(&tp->timer);
10863
1da177e4
LT
10864 tg3_netif_start(tp);
10865
f47c11ee 10866 tg3_full_unlock(tp);
1da177e4
LT
10867
10868 return 0;
10869}
10870
10871static struct pci_driver tg3_driver = {
10872 .name = DRV_MODULE_NAME,
10873 .id_table = tg3_pci_tbl,
10874 .probe = tg3_init_one,
10875 .remove = __devexit_p(tg3_remove_one),
10876 .suspend = tg3_suspend,
10877 .resume = tg3_resume
10878};
10879
10880static int __init tg3_init(void)
10881{
10882 return pci_module_init(&tg3_driver);
10883}
10884
10885static void __exit tg3_cleanup(void)
10886{
10887 pci_unregister_driver(&tg3_driver);
10888}
10889
10890module_init(tg3_init);
10891module_exit(tg3_cleanup);