xref: /OK3568_Linux_fs/kernel/drivers/net/ethernet/rdc/r6040.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * RDC R6040 Fast Ethernet MAC support
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Copyright (C) 2004 Sten Wang <sten.wang@rdc.com.tw>
6*4882a593Smuzhiyun  * Copyright (C) 2007
7*4882a593Smuzhiyun  *	Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>
8*4882a593Smuzhiyun  * Copyright (C) 2007-2012 Florian Fainelli <f.fainelli@gmail.com>
9*4882a593Smuzhiyun */
10*4882a593Smuzhiyun 
11*4882a593Smuzhiyun #include <linux/kernel.h>
12*4882a593Smuzhiyun #include <linux/module.h>
13*4882a593Smuzhiyun #include <linux/moduleparam.h>
14*4882a593Smuzhiyun #include <linux/string.h>
15*4882a593Smuzhiyun #include <linux/timer.h>
16*4882a593Smuzhiyun #include <linux/errno.h>
17*4882a593Smuzhiyun #include <linux/ioport.h>
18*4882a593Smuzhiyun #include <linux/interrupt.h>
19*4882a593Smuzhiyun #include <linux/pci.h>
20*4882a593Smuzhiyun #include <linux/netdevice.h>
21*4882a593Smuzhiyun #include <linux/etherdevice.h>
22*4882a593Smuzhiyun #include <linux/skbuff.h>
23*4882a593Smuzhiyun #include <linux/delay.h>
24*4882a593Smuzhiyun #include <linux/mii.h>
25*4882a593Smuzhiyun #include <linux/ethtool.h>
26*4882a593Smuzhiyun #include <linux/crc32.h>
27*4882a593Smuzhiyun #include <linux/spinlock.h>
28*4882a593Smuzhiyun #include <linux/bitops.h>
29*4882a593Smuzhiyun #include <linux/io.h>
30*4882a593Smuzhiyun #include <linux/irq.h>
31*4882a593Smuzhiyun #include <linux/uaccess.h>
32*4882a593Smuzhiyun #include <linux/phy.h>
33*4882a593Smuzhiyun 
34*4882a593Smuzhiyun #include <asm/processor.h>
35*4882a593Smuzhiyun 
36*4882a593Smuzhiyun #define DRV_NAME	"r6040"
37*4882a593Smuzhiyun #define DRV_VERSION	"0.29"
38*4882a593Smuzhiyun #define DRV_RELDATE	"04Jul2016"
39*4882a593Smuzhiyun 
40*4882a593Smuzhiyun /* Time in jiffies before concluding the transmitter is hung. */
41*4882a593Smuzhiyun #define TX_TIMEOUT	(6000 * HZ / 1000)
42*4882a593Smuzhiyun 
43*4882a593Smuzhiyun /* RDC MAC I/O Size */
44*4882a593Smuzhiyun #define R6040_IO_SIZE	256
45*4882a593Smuzhiyun 
46*4882a593Smuzhiyun /* MAX RDC MAC */
47*4882a593Smuzhiyun #define MAX_MAC		2
48*4882a593Smuzhiyun 
49*4882a593Smuzhiyun /* MAC registers */
50*4882a593Smuzhiyun #define MCR0		0x00	/* Control register 0 */
51*4882a593Smuzhiyun #define  MCR0_RCVEN	0x0002	/* Receive enable */
52*4882a593Smuzhiyun #define  MCR0_PROMISC	0x0020	/* Promiscuous mode */
53*4882a593Smuzhiyun #define  MCR0_HASH_EN	0x0100	/* Enable multicast hash table function */
54*4882a593Smuzhiyun #define  MCR0_XMTEN	0x1000	/* Transmission enable */
55*4882a593Smuzhiyun #define  MCR0_FD	0x8000	/* Full/Half duplex */
56*4882a593Smuzhiyun #define MCR1		0x04	/* Control register 1 */
57*4882a593Smuzhiyun #define  MAC_RST	0x0001	/* Reset the MAC */
58*4882a593Smuzhiyun #define MBCR		0x08	/* Bus control */
59*4882a593Smuzhiyun #define MT_ICR		0x0C	/* TX interrupt control */
60*4882a593Smuzhiyun #define MR_ICR		0x10	/* RX interrupt control */
61*4882a593Smuzhiyun #define MTPR		0x14	/* TX poll command register */
62*4882a593Smuzhiyun #define  TM2TX		0x0001	/* Trigger MAC to transmit */
63*4882a593Smuzhiyun #define MR_BSR		0x18	/* RX buffer size */
64*4882a593Smuzhiyun #define MR_DCR		0x1A	/* RX descriptor control */
65*4882a593Smuzhiyun #define MLSR		0x1C	/* Last status */
66*4882a593Smuzhiyun #define  TX_FIFO_UNDR	0x0200	/* TX FIFO under-run */
67*4882a593Smuzhiyun #define	 TX_EXCEEDC	0x2000	/* Transmit exceed collision */
68*4882a593Smuzhiyun #define  TX_LATEC	0x4000	/* Transmit late collision */
69*4882a593Smuzhiyun #define MMDIO		0x20	/* MDIO control register */
70*4882a593Smuzhiyun #define  MDIO_WRITE	0x4000	/* MDIO write */
71*4882a593Smuzhiyun #define  MDIO_READ	0x2000	/* MDIO read */
72*4882a593Smuzhiyun #define MMRD		0x24	/* MDIO read data register */
73*4882a593Smuzhiyun #define MMWD		0x28	/* MDIO write data register */
74*4882a593Smuzhiyun #define MTD_SA0		0x2C	/* TX descriptor start address 0 */
75*4882a593Smuzhiyun #define MTD_SA1		0x30	/* TX descriptor start address 1 */
76*4882a593Smuzhiyun #define MRD_SA0		0x34	/* RX descriptor start address 0 */
77*4882a593Smuzhiyun #define MRD_SA1		0x38	/* RX descriptor start address 1 */
78*4882a593Smuzhiyun #define MISR		0x3C	/* Status register */
79*4882a593Smuzhiyun #define MIER		0x40	/* INT enable register */
80*4882a593Smuzhiyun #define  MSK_INT	0x0000	/* Mask off interrupts */
81*4882a593Smuzhiyun #define  RX_FINISH	0x0001  /* RX finished */
82*4882a593Smuzhiyun #define  RX_NO_DESC	0x0002  /* No RX descriptor available */
83*4882a593Smuzhiyun #define  RX_FIFO_FULL	0x0004  /* RX FIFO full */
84*4882a593Smuzhiyun #define  RX_EARLY	0x0008  /* RX early */
85*4882a593Smuzhiyun #define  TX_FINISH	0x0010  /* TX finished */
86*4882a593Smuzhiyun #define  TX_EARLY	0x0080  /* TX early */
87*4882a593Smuzhiyun #define  EVENT_OVRFL	0x0100  /* Event counter overflow */
88*4882a593Smuzhiyun #define  LINK_CHANGED	0x0200  /* PHY link changed */
89*4882a593Smuzhiyun #define ME_CISR		0x44	/* Event counter INT status */
90*4882a593Smuzhiyun #define ME_CIER		0x48	/* Event counter INT enable  */
91*4882a593Smuzhiyun #define MR_CNT		0x50	/* Successfully received packet counter */
92*4882a593Smuzhiyun #define ME_CNT0		0x52	/* Event counter 0 */
93*4882a593Smuzhiyun #define ME_CNT1		0x54	/* Event counter 1 */
94*4882a593Smuzhiyun #define ME_CNT2		0x56	/* Event counter 2 */
95*4882a593Smuzhiyun #define ME_CNT3		0x58	/* Event counter 3 */
96*4882a593Smuzhiyun #define MT_CNT		0x5A	/* Successfully transmit packet counter */
97*4882a593Smuzhiyun #define ME_CNT4		0x5C	/* Event counter 4 */
98*4882a593Smuzhiyun #define MP_CNT		0x5E	/* Pause frame counter register */
99*4882a593Smuzhiyun #define MAR0		0x60	/* Hash table 0 */
100*4882a593Smuzhiyun #define MAR1		0x62	/* Hash table 1 */
101*4882a593Smuzhiyun #define MAR2		0x64	/* Hash table 2 */
102*4882a593Smuzhiyun #define MAR3		0x66	/* Hash table 3 */
103*4882a593Smuzhiyun #define MID_0L		0x68	/* Multicast address MID0 Low */
104*4882a593Smuzhiyun #define MID_0M		0x6A	/* Multicast address MID0 Medium */
105*4882a593Smuzhiyun #define MID_0H		0x6C	/* Multicast address MID0 High */
106*4882a593Smuzhiyun #define MID_1L		0x70	/* MID1 Low */
107*4882a593Smuzhiyun #define MID_1M		0x72	/* MID1 Medium */
108*4882a593Smuzhiyun #define MID_1H		0x74	/* MID1 High */
109*4882a593Smuzhiyun #define MID_2L		0x78	/* MID2 Low */
110*4882a593Smuzhiyun #define MID_2M		0x7A	/* MID2 Medium */
111*4882a593Smuzhiyun #define MID_2H		0x7C	/* MID2 High */
112*4882a593Smuzhiyun #define MID_3L		0x80	/* MID3 Low */
113*4882a593Smuzhiyun #define MID_3M		0x82	/* MID3 Medium */
114*4882a593Smuzhiyun #define MID_3H		0x84	/* MID3 High */
115*4882a593Smuzhiyun #define PHY_CC		0x88	/* PHY status change configuration register */
116*4882a593Smuzhiyun #define  SCEN		0x8000	/* PHY status change enable */
117*4882a593Smuzhiyun #define  PHYAD_SHIFT	8	/* PHY address shift */
118*4882a593Smuzhiyun #define  TMRDIV_SHIFT	0	/* Timer divider shift */
119*4882a593Smuzhiyun #define PHY_ST		0x8A	/* PHY status register */
120*4882a593Smuzhiyun #define MAC_SM		0xAC	/* MAC status machine */
121*4882a593Smuzhiyun #define  MAC_SM_RST	0x0002	/* MAC status machine reset */
122*4882a593Smuzhiyun #define MD_CSC		0xb6	/* MDC speed control register */
123*4882a593Smuzhiyun #define  MD_CSC_DEFAULT	0x0030
124*4882a593Smuzhiyun #define MAC_ID		0xBE	/* Identifier register */
125*4882a593Smuzhiyun 
126*4882a593Smuzhiyun #define TX_DCNT		0x80	/* TX descriptor count */
127*4882a593Smuzhiyun #define RX_DCNT		0x80	/* RX descriptor count */
128*4882a593Smuzhiyun #define MAX_BUF_SIZE	0x600
129*4882a593Smuzhiyun #define RX_DESC_SIZE	(RX_DCNT * sizeof(struct r6040_descriptor))
130*4882a593Smuzhiyun #define TX_DESC_SIZE	(TX_DCNT * sizeof(struct r6040_descriptor))
131*4882a593Smuzhiyun #define MBCR_DEFAULT	0x012A	/* MAC Bus Control Register */
132*4882a593Smuzhiyun #define MCAST_MAX	3	/* Max number multicast addresses to filter */
133*4882a593Smuzhiyun 
134*4882a593Smuzhiyun #define MAC_DEF_TIMEOUT	2048	/* Default MAC read/write operation timeout */
135*4882a593Smuzhiyun 
136*4882a593Smuzhiyun /* Descriptor status */
137*4882a593Smuzhiyun #define DSC_OWNER_MAC	0x8000	/* MAC is the owner of this descriptor */
138*4882a593Smuzhiyun #define DSC_RX_OK	0x4000	/* RX was successful */
139*4882a593Smuzhiyun #define DSC_RX_ERR	0x0800	/* RX PHY error */
140*4882a593Smuzhiyun #define DSC_RX_ERR_DRI	0x0400	/* RX dribble packet */
141*4882a593Smuzhiyun #define DSC_RX_ERR_BUF	0x0200	/* RX length exceeds buffer size */
142*4882a593Smuzhiyun #define DSC_RX_ERR_LONG	0x0100	/* RX length > maximum packet length */
143*4882a593Smuzhiyun #define DSC_RX_ERR_RUNT	0x0080	/* RX packet length < 64 byte */
144*4882a593Smuzhiyun #define DSC_RX_ERR_CRC	0x0040	/* RX CRC error */
145*4882a593Smuzhiyun #define DSC_RX_BCAST	0x0020	/* RX broadcast (no error) */
146*4882a593Smuzhiyun #define DSC_RX_MCAST	0x0010	/* RX multicast (no error) */
147*4882a593Smuzhiyun #define DSC_RX_MCH_HIT	0x0008	/* RX multicast hit in hash table (no error) */
148*4882a593Smuzhiyun #define DSC_RX_MIDH_HIT	0x0004	/* RX MID table hit (no error) */
149*4882a593Smuzhiyun #define DSC_RX_IDX_MID_MASK 3	/* RX mask for the index of matched MIDx */
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun MODULE_AUTHOR("Sten Wang <sten.wang@rdc.com.tw>,"
152*4882a593Smuzhiyun 	"Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>,"
153*4882a593Smuzhiyun 	"Florian Fainelli <f.fainelli@gmail.com>");
154*4882a593Smuzhiyun MODULE_LICENSE("GPL");
155*4882a593Smuzhiyun MODULE_DESCRIPTION("RDC R6040 NAPI PCI FastEthernet driver");
156*4882a593Smuzhiyun MODULE_VERSION(DRV_VERSION " " DRV_RELDATE);
157*4882a593Smuzhiyun 
158*4882a593Smuzhiyun /* RX and TX interrupts that we handle */
159*4882a593Smuzhiyun #define RX_INTS			(RX_FIFO_FULL | RX_NO_DESC | RX_FINISH)
160*4882a593Smuzhiyun #define TX_INTS			(TX_FINISH)
161*4882a593Smuzhiyun #define INT_MASK		(RX_INTS | TX_INTS)
162*4882a593Smuzhiyun 
163*4882a593Smuzhiyun struct r6040_descriptor {
164*4882a593Smuzhiyun 	u16	status, len;		/* 0-3 */
165*4882a593Smuzhiyun 	__le32	buf;			/* 4-7 */
166*4882a593Smuzhiyun 	__le32	ndesc;			/* 8-B */
167*4882a593Smuzhiyun 	u32	rev1;			/* C-F */
168*4882a593Smuzhiyun 	char	*vbufp;			/* 10-13 */
169*4882a593Smuzhiyun 	struct r6040_descriptor *vndescp;	/* 14-17 */
170*4882a593Smuzhiyun 	struct sk_buff *skb_ptr;	/* 18-1B */
171*4882a593Smuzhiyun 	u32	rev2;			/* 1C-1F */
172*4882a593Smuzhiyun } __aligned(32);
173*4882a593Smuzhiyun 
174*4882a593Smuzhiyun struct r6040_private {
175*4882a593Smuzhiyun 	spinlock_t lock;		/* driver lock */
176*4882a593Smuzhiyun 	struct pci_dev *pdev;
177*4882a593Smuzhiyun 	struct r6040_descriptor *rx_insert_ptr;
178*4882a593Smuzhiyun 	struct r6040_descriptor *rx_remove_ptr;
179*4882a593Smuzhiyun 	struct r6040_descriptor *tx_insert_ptr;
180*4882a593Smuzhiyun 	struct r6040_descriptor *tx_remove_ptr;
181*4882a593Smuzhiyun 	struct r6040_descriptor *rx_ring;
182*4882a593Smuzhiyun 	struct r6040_descriptor *tx_ring;
183*4882a593Smuzhiyun 	dma_addr_t rx_ring_dma;
184*4882a593Smuzhiyun 	dma_addr_t tx_ring_dma;
185*4882a593Smuzhiyun 	u16	tx_free_desc;
186*4882a593Smuzhiyun 	u16	mcr0;
187*4882a593Smuzhiyun 	struct net_device *dev;
188*4882a593Smuzhiyun 	struct mii_bus *mii_bus;
189*4882a593Smuzhiyun 	struct napi_struct napi;
190*4882a593Smuzhiyun 	void __iomem *base;
191*4882a593Smuzhiyun 	int old_link;
192*4882a593Smuzhiyun 	int old_duplex;
193*4882a593Smuzhiyun };
194*4882a593Smuzhiyun 
195*4882a593Smuzhiyun static char version[] = DRV_NAME
196*4882a593Smuzhiyun 	": RDC R6040 NAPI net driver,"
197*4882a593Smuzhiyun 	"version "DRV_VERSION " (" DRV_RELDATE ")";
198*4882a593Smuzhiyun 
199*4882a593Smuzhiyun /* Read a word data from PHY Chip */
r6040_phy_read(void __iomem * ioaddr,int phy_addr,int reg)200*4882a593Smuzhiyun static int r6040_phy_read(void __iomem *ioaddr, int phy_addr, int reg)
201*4882a593Smuzhiyun {
202*4882a593Smuzhiyun 	int limit = MAC_DEF_TIMEOUT;
203*4882a593Smuzhiyun 	u16 cmd;
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun 	iowrite16(MDIO_READ + reg + (phy_addr << 8), ioaddr + MMDIO);
206*4882a593Smuzhiyun 	/* Wait for the read bit to be cleared */
207*4882a593Smuzhiyun 	while (limit--) {
208*4882a593Smuzhiyun 		cmd = ioread16(ioaddr + MMDIO);
209*4882a593Smuzhiyun 		if (!(cmd & MDIO_READ))
210*4882a593Smuzhiyun 			break;
211*4882a593Smuzhiyun 		udelay(1);
212*4882a593Smuzhiyun 	}
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 	if (limit < 0)
215*4882a593Smuzhiyun 		return -ETIMEDOUT;
216*4882a593Smuzhiyun 
217*4882a593Smuzhiyun 	return ioread16(ioaddr + MMRD);
218*4882a593Smuzhiyun }
219*4882a593Smuzhiyun 
220*4882a593Smuzhiyun /* Write a word data from PHY Chip */
r6040_phy_write(void __iomem * ioaddr,int phy_addr,int reg,u16 val)221*4882a593Smuzhiyun static int r6040_phy_write(void __iomem *ioaddr,
222*4882a593Smuzhiyun 					int phy_addr, int reg, u16 val)
223*4882a593Smuzhiyun {
224*4882a593Smuzhiyun 	int limit = MAC_DEF_TIMEOUT;
225*4882a593Smuzhiyun 	u16 cmd;
226*4882a593Smuzhiyun 
227*4882a593Smuzhiyun 	iowrite16(val, ioaddr + MMWD);
228*4882a593Smuzhiyun 	/* Write the command to the MDIO bus */
229*4882a593Smuzhiyun 	iowrite16(MDIO_WRITE + reg + (phy_addr << 8), ioaddr + MMDIO);
230*4882a593Smuzhiyun 	/* Wait for the write bit to be cleared */
231*4882a593Smuzhiyun 	while (limit--) {
232*4882a593Smuzhiyun 		cmd = ioread16(ioaddr + MMDIO);
233*4882a593Smuzhiyun 		if (!(cmd & MDIO_WRITE))
234*4882a593Smuzhiyun 			break;
235*4882a593Smuzhiyun 		udelay(1);
236*4882a593Smuzhiyun 	}
237*4882a593Smuzhiyun 
238*4882a593Smuzhiyun 	return (limit < 0) ? -ETIMEDOUT : 0;
239*4882a593Smuzhiyun }
240*4882a593Smuzhiyun 
r6040_mdiobus_read(struct mii_bus * bus,int phy_addr,int reg)241*4882a593Smuzhiyun static int r6040_mdiobus_read(struct mii_bus *bus, int phy_addr, int reg)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun 	struct net_device *dev = bus->priv;
244*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
245*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
246*4882a593Smuzhiyun 
247*4882a593Smuzhiyun 	return r6040_phy_read(ioaddr, phy_addr, reg);
248*4882a593Smuzhiyun }
249*4882a593Smuzhiyun 
r6040_mdiobus_write(struct mii_bus * bus,int phy_addr,int reg,u16 value)250*4882a593Smuzhiyun static int r6040_mdiobus_write(struct mii_bus *bus, int phy_addr,
251*4882a593Smuzhiyun 						int reg, u16 value)
252*4882a593Smuzhiyun {
253*4882a593Smuzhiyun 	struct net_device *dev = bus->priv;
254*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
255*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
256*4882a593Smuzhiyun 
257*4882a593Smuzhiyun 	return r6040_phy_write(ioaddr, phy_addr, reg, value);
258*4882a593Smuzhiyun }
259*4882a593Smuzhiyun 
r6040_free_txbufs(struct net_device * dev)260*4882a593Smuzhiyun static void r6040_free_txbufs(struct net_device *dev)
261*4882a593Smuzhiyun {
262*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
263*4882a593Smuzhiyun 	int i;
264*4882a593Smuzhiyun 
265*4882a593Smuzhiyun 	for (i = 0; i < TX_DCNT; i++) {
266*4882a593Smuzhiyun 		if (lp->tx_insert_ptr->skb_ptr) {
267*4882a593Smuzhiyun 			dma_unmap_single(&lp->pdev->dev,
268*4882a593Smuzhiyun 					 le32_to_cpu(lp->tx_insert_ptr->buf),
269*4882a593Smuzhiyun 					 MAX_BUF_SIZE, DMA_TO_DEVICE);
270*4882a593Smuzhiyun 			dev_kfree_skb(lp->tx_insert_ptr->skb_ptr);
271*4882a593Smuzhiyun 			lp->tx_insert_ptr->skb_ptr = NULL;
272*4882a593Smuzhiyun 		}
273*4882a593Smuzhiyun 		lp->tx_insert_ptr = lp->tx_insert_ptr->vndescp;
274*4882a593Smuzhiyun 	}
275*4882a593Smuzhiyun }
276*4882a593Smuzhiyun 
r6040_free_rxbufs(struct net_device * dev)277*4882a593Smuzhiyun static void r6040_free_rxbufs(struct net_device *dev)
278*4882a593Smuzhiyun {
279*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
280*4882a593Smuzhiyun 	int i;
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 	for (i = 0; i < RX_DCNT; i++) {
283*4882a593Smuzhiyun 		if (lp->rx_insert_ptr->skb_ptr) {
284*4882a593Smuzhiyun 			dma_unmap_single(&lp->pdev->dev,
285*4882a593Smuzhiyun 					 le32_to_cpu(lp->rx_insert_ptr->buf),
286*4882a593Smuzhiyun 					 MAX_BUF_SIZE, DMA_FROM_DEVICE);
287*4882a593Smuzhiyun 			dev_kfree_skb(lp->rx_insert_ptr->skb_ptr);
288*4882a593Smuzhiyun 			lp->rx_insert_ptr->skb_ptr = NULL;
289*4882a593Smuzhiyun 		}
290*4882a593Smuzhiyun 		lp->rx_insert_ptr = lp->rx_insert_ptr->vndescp;
291*4882a593Smuzhiyun 	}
292*4882a593Smuzhiyun }
293*4882a593Smuzhiyun 
r6040_init_ring_desc(struct r6040_descriptor * desc_ring,dma_addr_t desc_dma,int size)294*4882a593Smuzhiyun static void r6040_init_ring_desc(struct r6040_descriptor *desc_ring,
295*4882a593Smuzhiyun 				 dma_addr_t desc_dma, int size)
296*4882a593Smuzhiyun {
297*4882a593Smuzhiyun 	struct r6040_descriptor *desc = desc_ring;
298*4882a593Smuzhiyun 	dma_addr_t mapping = desc_dma;
299*4882a593Smuzhiyun 
300*4882a593Smuzhiyun 	while (size-- > 0) {
301*4882a593Smuzhiyun 		mapping += sizeof(*desc);
302*4882a593Smuzhiyun 		desc->ndesc = cpu_to_le32(mapping);
303*4882a593Smuzhiyun 		desc->vndescp = desc + 1;
304*4882a593Smuzhiyun 		desc++;
305*4882a593Smuzhiyun 	}
306*4882a593Smuzhiyun 	desc--;
307*4882a593Smuzhiyun 	desc->ndesc = cpu_to_le32(desc_dma);
308*4882a593Smuzhiyun 	desc->vndescp = desc_ring;
309*4882a593Smuzhiyun }
310*4882a593Smuzhiyun 
r6040_init_txbufs(struct net_device * dev)311*4882a593Smuzhiyun static void r6040_init_txbufs(struct net_device *dev)
312*4882a593Smuzhiyun {
313*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun 	lp->tx_free_desc = TX_DCNT;
316*4882a593Smuzhiyun 
317*4882a593Smuzhiyun 	lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring;
318*4882a593Smuzhiyun 	r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT);
319*4882a593Smuzhiyun }
320*4882a593Smuzhiyun 
r6040_alloc_rxbufs(struct net_device * dev)321*4882a593Smuzhiyun static int r6040_alloc_rxbufs(struct net_device *dev)
322*4882a593Smuzhiyun {
323*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
324*4882a593Smuzhiyun 	struct r6040_descriptor *desc;
325*4882a593Smuzhiyun 	struct sk_buff *skb;
326*4882a593Smuzhiyun 	int rc;
327*4882a593Smuzhiyun 
328*4882a593Smuzhiyun 	lp->rx_remove_ptr = lp->rx_insert_ptr = lp->rx_ring;
329*4882a593Smuzhiyun 	r6040_init_ring_desc(lp->rx_ring, lp->rx_ring_dma, RX_DCNT);
330*4882a593Smuzhiyun 
331*4882a593Smuzhiyun 	/* Allocate skbs for the rx descriptors */
332*4882a593Smuzhiyun 	desc = lp->rx_ring;
333*4882a593Smuzhiyun 	do {
334*4882a593Smuzhiyun 		skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
335*4882a593Smuzhiyun 		if (!skb) {
336*4882a593Smuzhiyun 			rc = -ENOMEM;
337*4882a593Smuzhiyun 			goto err_exit;
338*4882a593Smuzhiyun 		}
339*4882a593Smuzhiyun 		desc->skb_ptr = skb;
340*4882a593Smuzhiyun 		desc->buf = cpu_to_le32(dma_map_single(&lp->pdev->dev,
341*4882a593Smuzhiyun 						       desc->skb_ptr->data,
342*4882a593Smuzhiyun 						       MAX_BUF_SIZE,
343*4882a593Smuzhiyun 						       DMA_FROM_DEVICE));
344*4882a593Smuzhiyun 		desc->status = DSC_OWNER_MAC;
345*4882a593Smuzhiyun 		desc = desc->vndescp;
346*4882a593Smuzhiyun 	} while (desc != lp->rx_ring);
347*4882a593Smuzhiyun 
348*4882a593Smuzhiyun 	return 0;
349*4882a593Smuzhiyun 
350*4882a593Smuzhiyun err_exit:
351*4882a593Smuzhiyun 	/* Deallocate all previously allocated skbs */
352*4882a593Smuzhiyun 	r6040_free_rxbufs(dev);
353*4882a593Smuzhiyun 	return rc;
354*4882a593Smuzhiyun }
355*4882a593Smuzhiyun 
r6040_reset_mac(struct r6040_private * lp)356*4882a593Smuzhiyun static void r6040_reset_mac(struct r6040_private *lp)
357*4882a593Smuzhiyun {
358*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
359*4882a593Smuzhiyun 	int limit = MAC_DEF_TIMEOUT;
360*4882a593Smuzhiyun 	u16 cmd, md_csc;
361*4882a593Smuzhiyun 
362*4882a593Smuzhiyun 	md_csc = ioread16(ioaddr + MD_CSC);
363*4882a593Smuzhiyun 	iowrite16(MAC_RST, ioaddr + MCR1);
364*4882a593Smuzhiyun 	while (limit--) {
365*4882a593Smuzhiyun 		cmd = ioread16(ioaddr + MCR1);
366*4882a593Smuzhiyun 		if (cmd & MAC_RST)
367*4882a593Smuzhiyun 			break;
368*4882a593Smuzhiyun 	}
369*4882a593Smuzhiyun 
370*4882a593Smuzhiyun 	/* Reset internal state machine */
371*4882a593Smuzhiyun 	iowrite16(MAC_SM_RST, ioaddr + MAC_SM);
372*4882a593Smuzhiyun 	iowrite16(0, ioaddr + MAC_SM);
373*4882a593Smuzhiyun 	mdelay(5);
374*4882a593Smuzhiyun 
375*4882a593Smuzhiyun 	/* Restore MDIO clock frequency */
376*4882a593Smuzhiyun 	if (md_csc != MD_CSC_DEFAULT)
377*4882a593Smuzhiyun 		iowrite16(md_csc, ioaddr + MD_CSC);
378*4882a593Smuzhiyun }
379*4882a593Smuzhiyun 
r6040_init_mac_regs(struct net_device * dev)380*4882a593Smuzhiyun static void r6040_init_mac_regs(struct net_device *dev)
381*4882a593Smuzhiyun {
382*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
383*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
384*4882a593Smuzhiyun 
385*4882a593Smuzhiyun 	/* Mask Off Interrupt */
386*4882a593Smuzhiyun 	iowrite16(MSK_INT, ioaddr + MIER);
387*4882a593Smuzhiyun 
388*4882a593Smuzhiyun 	/* Reset RDC MAC */
389*4882a593Smuzhiyun 	r6040_reset_mac(lp);
390*4882a593Smuzhiyun 
391*4882a593Smuzhiyun 	/* MAC Bus Control Register */
392*4882a593Smuzhiyun 	iowrite16(MBCR_DEFAULT, ioaddr + MBCR);
393*4882a593Smuzhiyun 
394*4882a593Smuzhiyun 	/* Buffer Size Register */
395*4882a593Smuzhiyun 	iowrite16(MAX_BUF_SIZE, ioaddr + MR_BSR);
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	/* Write TX ring start address */
398*4882a593Smuzhiyun 	iowrite16(lp->tx_ring_dma, ioaddr + MTD_SA0);
399*4882a593Smuzhiyun 	iowrite16(lp->tx_ring_dma >> 16, ioaddr + MTD_SA1);
400*4882a593Smuzhiyun 
401*4882a593Smuzhiyun 	/* Write RX ring start address */
402*4882a593Smuzhiyun 	iowrite16(lp->rx_ring_dma, ioaddr + MRD_SA0);
403*4882a593Smuzhiyun 	iowrite16(lp->rx_ring_dma >> 16, ioaddr + MRD_SA1);
404*4882a593Smuzhiyun 
405*4882a593Smuzhiyun 	/* Set interrupt waiting time and packet numbers */
406*4882a593Smuzhiyun 	iowrite16(0, ioaddr + MT_ICR);
407*4882a593Smuzhiyun 	iowrite16(0, ioaddr + MR_ICR);
408*4882a593Smuzhiyun 
409*4882a593Smuzhiyun 	/* Enable interrupts */
410*4882a593Smuzhiyun 	iowrite16(INT_MASK, ioaddr + MIER);
411*4882a593Smuzhiyun 
412*4882a593Smuzhiyun 	/* Enable TX and RX */
413*4882a593Smuzhiyun 	iowrite16(lp->mcr0 | MCR0_RCVEN, ioaddr);
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun 	/* Let TX poll the descriptors
416*4882a593Smuzhiyun 	 * we may got called by r6040_tx_timeout which has left
417*4882a593Smuzhiyun 	 * some unsent tx buffers */
418*4882a593Smuzhiyun 	iowrite16(TM2TX, ioaddr + MTPR);
419*4882a593Smuzhiyun }
420*4882a593Smuzhiyun 
r6040_tx_timeout(struct net_device * dev,unsigned int txqueue)421*4882a593Smuzhiyun static void r6040_tx_timeout(struct net_device *dev, unsigned int txqueue)
422*4882a593Smuzhiyun {
423*4882a593Smuzhiyun 	struct r6040_private *priv = netdev_priv(dev);
424*4882a593Smuzhiyun 	void __iomem *ioaddr = priv->base;
425*4882a593Smuzhiyun 
426*4882a593Smuzhiyun 	netdev_warn(dev, "transmit timed out, int enable %4.4x "
427*4882a593Smuzhiyun 		"status %4.4x\n",
428*4882a593Smuzhiyun 		ioread16(ioaddr + MIER),
429*4882a593Smuzhiyun 		ioread16(ioaddr + MISR));
430*4882a593Smuzhiyun 
431*4882a593Smuzhiyun 	dev->stats.tx_errors++;
432*4882a593Smuzhiyun 
433*4882a593Smuzhiyun 	/* Reset MAC and re-init all registers */
434*4882a593Smuzhiyun 	r6040_init_mac_regs(dev);
435*4882a593Smuzhiyun }
436*4882a593Smuzhiyun 
r6040_get_stats(struct net_device * dev)437*4882a593Smuzhiyun static struct net_device_stats *r6040_get_stats(struct net_device *dev)
438*4882a593Smuzhiyun {
439*4882a593Smuzhiyun 	struct r6040_private *priv = netdev_priv(dev);
440*4882a593Smuzhiyun 	void __iomem *ioaddr = priv->base;
441*4882a593Smuzhiyun 	unsigned long flags;
442*4882a593Smuzhiyun 
443*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->lock, flags);
444*4882a593Smuzhiyun 	dev->stats.rx_crc_errors += ioread8(ioaddr + ME_CNT1);
445*4882a593Smuzhiyun 	dev->stats.multicast += ioread8(ioaddr + ME_CNT0);
446*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->lock, flags);
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	return &dev->stats;
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun 
451*4882a593Smuzhiyun /* Stop RDC MAC and Free the allocated resource */
r6040_down(struct net_device * dev)452*4882a593Smuzhiyun static void r6040_down(struct net_device *dev)
453*4882a593Smuzhiyun {
454*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
455*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
456*4882a593Smuzhiyun 	u16 *adrp;
457*4882a593Smuzhiyun 
458*4882a593Smuzhiyun 	/* Stop MAC */
459*4882a593Smuzhiyun 	iowrite16(MSK_INT, ioaddr + MIER);	/* Mask Off Interrupt */
460*4882a593Smuzhiyun 
461*4882a593Smuzhiyun 	/* Reset RDC MAC */
462*4882a593Smuzhiyun 	r6040_reset_mac(lp);
463*4882a593Smuzhiyun 
464*4882a593Smuzhiyun 	/* Restore MAC Address to MIDx */
465*4882a593Smuzhiyun 	adrp = (u16 *) dev->dev_addr;
466*4882a593Smuzhiyun 	iowrite16(adrp[0], ioaddr + MID_0L);
467*4882a593Smuzhiyun 	iowrite16(adrp[1], ioaddr + MID_0M);
468*4882a593Smuzhiyun 	iowrite16(adrp[2], ioaddr + MID_0H);
469*4882a593Smuzhiyun }
470*4882a593Smuzhiyun 
r6040_close(struct net_device * dev)471*4882a593Smuzhiyun static int r6040_close(struct net_device *dev)
472*4882a593Smuzhiyun {
473*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
474*4882a593Smuzhiyun 	struct pci_dev *pdev = lp->pdev;
475*4882a593Smuzhiyun 
476*4882a593Smuzhiyun 	phy_stop(dev->phydev);
477*4882a593Smuzhiyun 	napi_disable(&lp->napi);
478*4882a593Smuzhiyun 	netif_stop_queue(dev);
479*4882a593Smuzhiyun 
480*4882a593Smuzhiyun 	spin_lock_irq(&lp->lock);
481*4882a593Smuzhiyun 	r6040_down(dev);
482*4882a593Smuzhiyun 
483*4882a593Smuzhiyun 	/* Free RX buffer */
484*4882a593Smuzhiyun 	r6040_free_rxbufs(dev);
485*4882a593Smuzhiyun 
486*4882a593Smuzhiyun 	/* Free TX buffer */
487*4882a593Smuzhiyun 	r6040_free_txbufs(dev);
488*4882a593Smuzhiyun 
489*4882a593Smuzhiyun 	spin_unlock_irq(&lp->lock);
490*4882a593Smuzhiyun 
491*4882a593Smuzhiyun 	free_irq(dev->irq, dev);
492*4882a593Smuzhiyun 
493*4882a593Smuzhiyun 	/* Free Descriptor memory */
494*4882a593Smuzhiyun 	if (lp->rx_ring) {
495*4882a593Smuzhiyun 		dma_free_coherent(&pdev->dev, RX_DESC_SIZE, lp->rx_ring,
496*4882a593Smuzhiyun 				  lp->rx_ring_dma);
497*4882a593Smuzhiyun 		lp->rx_ring = NULL;
498*4882a593Smuzhiyun 	}
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun 	if (lp->tx_ring) {
501*4882a593Smuzhiyun 		dma_free_coherent(&pdev->dev, TX_DESC_SIZE, lp->tx_ring,
502*4882a593Smuzhiyun 				  lp->tx_ring_dma);
503*4882a593Smuzhiyun 		lp->tx_ring = NULL;
504*4882a593Smuzhiyun 	}
505*4882a593Smuzhiyun 
506*4882a593Smuzhiyun 	return 0;
507*4882a593Smuzhiyun }
508*4882a593Smuzhiyun 
r6040_rx(struct net_device * dev,int limit)509*4882a593Smuzhiyun static int r6040_rx(struct net_device *dev, int limit)
510*4882a593Smuzhiyun {
511*4882a593Smuzhiyun 	struct r6040_private *priv = netdev_priv(dev);
512*4882a593Smuzhiyun 	struct r6040_descriptor *descptr = priv->rx_remove_ptr;
513*4882a593Smuzhiyun 	struct sk_buff *skb_ptr, *new_skb;
514*4882a593Smuzhiyun 	int count = 0;
515*4882a593Smuzhiyun 	u16 err;
516*4882a593Smuzhiyun 
517*4882a593Smuzhiyun 	/* Limit not reached and the descriptor belongs to the CPU */
518*4882a593Smuzhiyun 	while (count < limit && !(descptr->status & DSC_OWNER_MAC)) {
519*4882a593Smuzhiyun 		/* Read the descriptor status */
520*4882a593Smuzhiyun 		err = descptr->status;
521*4882a593Smuzhiyun 		/* Global error status set */
522*4882a593Smuzhiyun 		if (err & DSC_RX_ERR) {
523*4882a593Smuzhiyun 			/* RX dribble */
524*4882a593Smuzhiyun 			if (err & DSC_RX_ERR_DRI)
525*4882a593Smuzhiyun 				dev->stats.rx_frame_errors++;
526*4882a593Smuzhiyun 			/* Buffer length exceeded */
527*4882a593Smuzhiyun 			if (err & DSC_RX_ERR_BUF)
528*4882a593Smuzhiyun 				dev->stats.rx_length_errors++;
529*4882a593Smuzhiyun 			/* Packet too long */
530*4882a593Smuzhiyun 			if (err & DSC_RX_ERR_LONG)
531*4882a593Smuzhiyun 				dev->stats.rx_length_errors++;
532*4882a593Smuzhiyun 			/* Packet < 64 bytes */
533*4882a593Smuzhiyun 			if (err & DSC_RX_ERR_RUNT)
534*4882a593Smuzhiyun 				dev->stats.rx_length_errors++;
535*4882a593Smuzhiyun 			/* CRC error */
536*4882a593Smuzhiyun 			if (err & DSC_RX_ERR_CRC) {
537*4882a593Smuzhiyun 				spin_lock(&priv->lock);
538*4882a593Smuzhiyun 				dev->stats.rx_crc_errors++;
539*4882a593Smuzhiyun 				spin_unlock(&priv->lock);
540*4882a593Smuzhiyun 			}
541*4882a593Smuzhiyun 			goto next_descr;
542*4882a593Smuzhiyun 		}
543*4882a593Smuzhiyun 
544*4882a593Smuzhiyun 		/* Packet successfully received */
545*4882a593Smuzhiyun 		new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
546*4882a593Smuzhiyun 		if (!new_skb) {
547*4882a593Smuzhiyun 			dev->stats.rx_dropped++;
548*4882a593Smuzhiyun 			goto next_descr;
549*4882a593Smuzhiyun 		}
550*4882a593Smuzhiyun 		skb_ptr = descptr->skb_ptr;
551*4882a593Smuzhiyun 		skb_ptr->dev = priv->dev;
552*4882a593Smuzhiyun 
553*4882a593Smuzhiyun 		/* Do not count the CRC */
554*4882a593Smuzhiyun 		skb_put(skb_ptr, descptr->len - 4);
555*4882a593Smuzhiyun 		dma_unmap_single(&priv->pdev->dev, le32_to_cpu(descptr->buf),
556*4882a593Smuzhiyun 				 MAX_BUF_SIZE, DMA_FROM_DEVICE);
557*4882a593Smuzhiyun 		skb_ptr->protocol = eth_type_trans(skb_ptr, priv->dev);
558*4882a593Smuzhiyun 
559*4882a593Smuzhiyun 		/* Send to upper layer */
560*4882a593Smuzhiyun 		netif_receive_skb(skb_ptr);
561*4882a593Smuzhiyun 		dev->stats.rx_packets++;
562*4882a593Smuzhiyun 		dev->stats.rx_bytes += descptr->len - 4;
563*4882a593Smuzhiyun 
564*4882a593Smuzhiyun 		/* put new skb into descriptor */
565*4882a593Smuzhiyun 		descptr->skb_ptr = new_skb;
566*4882a593Smuzhiyun 		descptr->buf = cpu_to_le32(dma_map_single(&priv->pdev->dev,
567*4882a593Smuzhiyun 							  descptr->skb_ptr->data,
568*4882a593Smuzhiyun 							  MAX_BUF_SIZE,
569*4882a593Smuzhiyun 							  DMA_FROM_DEVICE));
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun next_descr:
572*4882a593Smuzhiyun 		/* put the descriptor back to the MAC */
573*4882a593Smuzhiyun 		descptr->status = DSC_OWNER_MAC;
574*4882a593Smuzhiyun 		descptr = descptr->vndescp;
575*4882a593Smuzhiyun 		count++;
576*4882a593Smuzhiyun 	}
577*4882a593Smuzhiyun 	priv->rx_remove_ptr = descptr;
578*4882a593Smuzhiyun 
579*4882a593Smuzhiyun 	return count;
580*4882a593Smuzhiyun }
581*4882a593Smuzhiyun 
r6040_tx(struct net_device * dev)582*4882a593Smuzhiyun static void r6040_tx(struct net_device *dev)
583*4882a593Smuzhiyun {
584*4882a593Smuzhiyun 	struct r6040_private *priv = netdev_priv(dev);
585*4882a593Smuzhiyun 	struct r6040_descriptor *descptr;
586*4882a593Smuzhiyun 	void __iomem *ioaddr = priv->base;
587*4882a593Smuzhiyun 	struct sk_buff *skb_ptr;
588*4882a593Smuzhiyun 	u16 err;
589*4882a593Smuzhiyun 
590*4882a593Smuzhiyun 	spin_lock(&priv->lock);
591*4882a593Smuzhiyun 	descptr = priv->tx_remove_ptr;
592*4882a593Smuzhiyun 	while (priv->tx_free_desc < TX_DCNT) {
593*4882a593Smuzhiyun 		/* Check for errors */
594*4882a593Smuzhiyun 		err = ioread16(ioaddr + MLSR);
595*4882a593Smuzhiyun 
596*4882a593Smuzhiyun 		if (err & TX_FIFO_UNDR)
597*4882a593Smuzhiyun 			dev->stats.tx_fifo_errors++;
598*4882a593Smuzhiyun 		if (err & (TX_EXCEEDC | TX_LATEC))
599*4882a593Smuzhiyun 			dev->stats.tx_carrier_errors++;
600*4882a593Smuzhiyun 
601*4882a593Smuzhiyun 		if (descptr->status & DSC_OWNER_MAC)
602*4882a593Smuzhiyun 			break; /* Not complete */
603*4882a593Smuzhiyun 		skb_ptr = descptr->skb_ptr;
604*4882a593Smuzhiyun 
605*4882a593Smuzhiyun 		/* Statistic Counter */
606*4882a593Smuzhiyun 		dev->stats.tx_packets++;
607*4882a593Smuzhiyun 		dev->stats.tx_bytes += skb_ptr->len;
608*4882a593Smuzhiyun 
609*4882a593Smuzhiyun 		dma_unmap_single(&priv->pdev->dev, le32_to_cpu(descptr->buf),
610*4882a593Smuzhiyun 				 skb_ptr->len, DMA_TO_DEVICE);
611*4882a593Smuzhiyun 		/* Free buffer */
612*4882a593Smuzhiyun 		dev_kfree_skb(skb_ptr);
613*4882a593Smuzhiyun 		descptr->skb_ptr = NULL;
614*4882a593Smuzhiyun 		/* To next descriptor */
615*4882a593Smuzhiyun 		descptr = descptr->vndescp;
616*4882a593Smuzhiyun 		priv->tx_free_desc++;
617*4882a593Smuzhiyun 	}
618*4882a593Smuzhiyun 	priv->tx_remove_ptr = descptr;
619*4882a593Smuzhiyun 
620*4882a593Smuzhiyun 	if (priv->tx_free_desc)
621*4882a593Smuzhiyun 		netif_wake_queue(dev);
622*4882a593Smuzhiyun 	spin_unlock(&priv->lock);
623*4882a593Smuzhiyun }
624*4882a593Smuzhiyun 
r6040_poll(struct napi_struct * napi,int budget)625*4882a593Smuzhiyun static int r6040_poll(struct napi_struct *napi, int budget)
626*4882a593Smuzhiyun {
627*4882a593Smuzhiyun 	struct r6040_private *priv =
628*4882a593Smuzhiyun 		container_of(napi, struct r6040_private, napi);
629*4882a593Smuzhiyun 	struct net_device *dev = priv->dev;
630*4882a593Smuzhiyun 	void __iomem *ioaddr = priv->base;
631*4882a593Smuzhiyun 	int work_done;
632*4882a593Smuzhiyun 
633*4882a593Smuzhiyun 	r6040_tx(dev);
634*4882a593Smuzhiyun 
635*4882a593Smuzhiyun 	work_done = r6040_rx(dev, budget);
636*4882a593Smuzhiyun 
637*4882a593Smuzhiyun 	if (work_done < budget) {
638*4882a593Smuzhiyun 		napi_complete_done(napi, work_done);
639*4882a593Smuzhiyun 		/* Enable RX/TX interrupt */
640*4882a593Smuzhiyun 		iowrite16(ioread16(ioaddr + MIER) | RX_INTS | TX_INTS,
641*4882a593Smuzhiyun 			  ioaddr + MIER);
642*4882a593Smuzhiyun 	}
643*4882a593Smuzhiyun 	return work_done;
644*4882a593Smuzhiyun }
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun /* The RDC interrupt handler. */
r6040_interrupt(int irq,void * dev_id)647*4882a593Smuzhiyun static irqreturn_t r6040_interrupt(int irq, void *dev_id)
648*4882a593Smuzhiyun {
649*4882a593Smuzhiyun 	struct net_device *dev = dev_id;
650*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
651*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
652*4882a593Smuzhiyun 	u16 misr, status;
653*4882a593Smuzhiyun 
654*4882a593Smuzhiyun 	/* Save MIER */
655*4882a593Smuzhiyun 	misr = ioread16(ioaddr + MIER);
656*4882a593Smuzhiyun 	/* Mask off RDC MAC interrupt */
657*4882a593Smuzhiyun 	iowrite16(MSK_INT, ioaddr + MIER);
658*4882a593Smuzhiyun 	/* Read MISR status and clear */
659*4882a593Smuzhiyun 	status = ioread16(ioaddr + MISR);
660*4882a593Smuzhiyun 
661*4882a593Smuzhiyun 	if (status == 0x0000 || status == 0xffff) {
662*4882a593Smuzhiyun 		/* Restore RDC MAC interrupt */
663*4882a593Smuzhiyun 		iowrite16(misr, ioaddr + MIER);
664*4882a593Smuzhiyun 		return IRQ_NONE;
665*4882a593Smuzhiyun 	}
666*4882a593Smuzhiyun 
667*4882a593Smuzhiyun 	/* RX interrupt request */
668*4882a593Smuzhiyun 	if (status & (RX_INTS | TX_INTS)) {
669*4882a593Smuzhiyun 		if (status & RX_NO_DESC) {
670*4882a593Smuzhiyun 			/* RX descriptor unavailable */
671*4882a593Smuzhiyun 			dev->stats.rx_dropped++;
672*4882a593Smuzhiyun 			dev->stats.rx_missed_errors++;
673*4882a593Smuzhiyun 		}
674*4882a593Smuzhiyun 		if (status & RX_FIFO_FULL)
675*4882a593Smuzhiyun 			dev->stats.rx_fifo_errors++;
676*4882a593Smuzhiyun 
677*4882a593Smuzhiyun 		if (likely(napi_schedule_prep(&lp->napi))) {
678*4882a593Smuzhiyun 			/* Mask off RX interrupt */
679*4882a593Smuzhiyun 			misr &= ~(RX_INTS | TX_INTS);
680*4882a593Smuzhiyun 			__napi_schedule_irqoff(&lp->napi);
681*4882a593Smuzhiyun 		}
682*4882a593Smuzhiyun 	}
683*4882a593Smuzhiyun 
684*4882a593Smuzhiyun 	/* Restore RDC MAC interrupt */
685*4882a593Smuzhiyun 	iowrite16(misr, ioaddr + MIER);
686*4882a593Smuzhiyun 
687*4882a593Smuzhiyun 	return IRQ_HANDLED;
688*4882a593Smuzhiyun }
689*4882a593Smuzhiyun 
690*4882a593Smuzhiyun #ifdef CONFIG_NET_POLL_CONTROLLER
r6040_poll_controller(struct net_device * dev)691*4882a593Smuzhiyun static void r6040_poll_controller(struct net_device *dev)
692*4882a593Smuzhiyun {
693*4882a593Smuzhiyun 	disable_irq(dev->irq);
694*4882a593Smuzhiyun 	r6040_interrupt(dev->irq, dev);
695*4882a593Smuzhiyun 	enable_irq(dev->irq);
696*4882a593Smuzhiyun }
697*4882a593Smuzhiyun #endif
698*4882a593Smuzhiyun 
699*4882a593Smuzhiyun /* Init RDC MAC */
r6040_up(struct net_device * dev)700*4882a593Smuzhiyun static int r6040_up(struct net_device *dev)
701*4882a593Smuzhiyun {
702*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
703*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
704*4882a593Smuzhiyun 	int ret;
705*4882a593Smuzhiyun 
706*4882a593Smuzhiyun 	/* Initialise and alloc RX/TX buffers */
707*4882a593Smuzhiyun 	r6040_init_txbufs(dev);
708*4882a593Smuzhiyun 	ret = r6040_alloc_rxbufs(dev);
709*4882a593Smuzhiyun 	if (ret)
710*4882a593Smuzhiyun 		return ret;
711*4882a593Smuzhiyun 
712*4882a593Smuzhiyun 	/* improve performance (by RDC guys) */
713*4882a593Smuzhiyun 	r6040_phy_write(ioaddr, 30, 17,
714*4882a593Smuzhiyun 			(r6040_phy_read(ioaddr, 30, 17) | 0x4000));
715*4882a593Smuzhiyun 	r6040_phy_write(ioaddr, 30, 17,
716*4882a593Smuzhiyun 			~((~r6040_phy_read(ioaddr, 30, 17)) | 0x2000));
717*4882a593Smuzhiyun 	r6040_phy_write(ioaddr, 0, 19, 0x0000);
718*4882a593Smuzhiyun 	r6040_phy_write(ioaddr, 0, 30, 0x01F0);
719*4882a593Smuzhiyun 
720*4882a593Smuzhiyun 	/* Initialize all MAC registers */
721*4882a593Smuzhiyun 	r6040_init_mac_regs(dev);
722*4882a593Smuzhiyun 
723*4882a593Smuzhiyun 	phy_start(dev->phydev);
724*4882a593Smuzhiyun 
725*4882a593Smuzhiyun 	return 0;
726*4882a593Smuzhiyun }
727*4882a593Smuzhiyun 
728*4882a593Smuzhiyun 
729*4882a593Smuzhiyun /* Read/set MAC address routines */
r6040_mac_address(struct net_device * dev)730*4882a593Smuzhiyun static void r6040_mac_address(struct net_device *dev)
731*4882a593Smuzhiyun {
732*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
733*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
734*4882a593Smuzhiyun 	u16 *adrp;
735*4882a593Smuzhiyun 
736*4882a593Smuzhiyun 	/* Reset MAC */
737*4882a593Smuzhiyun 	r6040_reset_mac(lp);
738*4882a593Smuzhiyun 
739*4882a593Smuzhiyun 	/* Restore MAC Address */
740*4882a593Smuzhiyun 	adrp = (u16 *) dev->dev_addr;
741*4882a593Smuzhiyun 	iowrite16(adrp[0], ioaddr + MID_0L);
742*4882a593Smuzhiyun 	iowrite16(adrp[1], ioaddr + MID_0M);
743*4882a593Smuzhiyun 	iowrite16(adrp[2], ioaddr + MID_0H);
744*4882a593Smuzhiyun }
745*4882a593Smuzhiyun 
r6040_open(struct net_device * dev)746*4882a593Smuzhiyun static int r6040_open(struct net_device *dev)
747*4882a593Smuzhiyun {
748*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
749*4882a593Smuzhiyun 	int ret;
750*4882a593Smuzhiyun 
751*4882a593Smuzhiyun 	/* Request IRQ and Register interrupt handler */
752*4882a593Smuzhiyun 	ret = request_irq(dev->irq, r6040_interrupt,
753*4882a593Smuzhiyun 		IRQF_SHARED, dev->name, dev);
754*4882a593Smuzhiyun 	if (ret)
755*4882a593Smuzhiyun 		goto out;
756*4882a593Smuzhiyun 
757*4882a593Smuzhiyun 	/* Set MAC address */
758*4882a593Smuzhiyun 	r6040_mac_address(dev);
759*4882a593Smuzhiyun 
760*4882a593Smuzhiyun 	/* Allocate Descriptor memory */
761*4882a593Smuzhiyun 	lp->rx_ring =
762*4882a593Smuzhiyun 		dma_alloc_coherent(&lp->pdev->dev, RX_DESC_SIZE,
763*4882a593Smuzhiyun 				   &lp->rx_ring_dma, GFP_KERNEL);
764*4882a593Smuzhiyun 	if (!lp->rx_ring) {
765*4882a593Smuzhiyun 		ret = -ENOMEM;
766*4882a593Smuzhiyun 		goto err_free_irq;
767*4882a593Smuzhiyun 	}
768*4882a593Smuzhiyun 
769*4882a593Smuzhiyun 	lp->tx_ring =
770*4882a593Smuzhiyun 		dma_alloc_coherent(&lp->pdev->dev, TX_DESC_SIZE,
771*4882a593Smuzhiyun 				   &lp->tx_ring_dma, GFP_KERNEL);
772*4882a593Smuzhiyun 	if (!lp->tx_ring) {
773*4882a593Smuzhiyun 		ret = -ENOMEM;
774*4882a593Smuzhiyun 		goto err_free_rx_ring;
775*4882a593Smuzhiyun 	}
776*4882a593Smuzhiyun 
777*4882a593Smuzhiyun 	ret = r6040_up(dev);
778*4882a593Smuzhiyun 	if (ret)
779*4882a593Smuzhiyun 		goto err_free_tx_ring;
780*4882a593Smuzhiyun 
781*4882a593Smuzhiyun 	napi_enable(&lp->napi);
782*4882a593Smuzhiyun 	netif_start_queue(dev);
783*4882a593Smuzhiyun 
784*4882a593Smuzhiyun 	return 0;
785*4882a593Smuzhiyun 
786*4882a593Smuzhiyun err_free_tx_ring:
787*4882a593Smuzhiyun 	dma_free_coherent(&lp->pdev->dev, TX_DESC_SIZE, lp->tx_ring,
788*4882a593Smuzhiyun 			  lp->tx_ring_dma);
789*4882a593Smuzhiyun err_free_rx_ring:
790*4882a593Smuzhiyun 	dma_free_coherent(&lp->pdev->dev, RX_DESC_SIZE, lp->rx_ring,
791*4882a593Smuzhiyun 			  lp->rx_ring_dma);
792*4882a593Smuzhiyun err_free_irq:
793*4882a593Smuzhiyun 	free_irq(dev->irq, dev);
794*4882a593Smuzhiyun out:
795*4882a593Smuzhiyun 	return ret;
796*4882a593Smuzhiyun }
797*4882a593Smuzhiyun 
r6040_start_xmit(struct sk_buff * skb,struct net_device * dev)798*4882a593Smuzhiyun static netdev_tx_t r6040_start_xmit(struct sk_buff *skb,
799*4882a593Smuzhiyun 				    struct net_device *dev)
800*4882a593Smuzhiyun {
801*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
802*4882a593Smuzhiyun 	struct r6040_descriptor *descptr;
803*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
804*4882a593Smuzhiyun 	unsigned long flags;
805*4882a593Smuzhiyun 
806*4882a593Smuzhiyun 	if (skb_put_padto(skb, ETH_ZLEN) < 0)
807*4882a593Smuzhiyun 		return NETDEV_TX_OK;
808*4882a593Smuzhiyun 
809*4882a593Smuzhiyun 	/* Critical Section */
810*4882a593Smuzhiyun 	spin_lock_irqsave(&lp->lock, flags);
811*4882a593Smuzhiyun 
812*4882a593Smuzhiyun 	/* TX resource check */
813*4882a593Smuzhiyun 	if (!lp->tx_free_desc) {
814*4882a593Smuzhiyun 		spin_unlock_irqrestore(&lp->lock, flags);
815*4882a593Smuzhiyun 		netif_stop_queue(dev);
816*4882a593Smuzhiyun 		netdev_err(dev, ": no tx descriptor\n");
817*4882a593Smuzhiyun 		return NETDEV_TX_BUSY;
818*4882a593Smuzhiyun 	}
819*4882a593Smuzhiyun 
820*4882a593Smuzhiyun 	/* Set TX descriptor & Transmit it */
821*4882a593Smuzhiyun 	lp->tx_free_desc--;
822*4882a593Smuzhiyun 	descptr = lp->tx_insert_ptr;
823*4882a593Smuzhiyun 	descptr->len = skb->len;
824*4882a593Smuzhiyun 	descptr->skb_ptr = skb;
825*4882a593Smuzhiyun 	descptr->buf = cpu_to_le32(dma_map_single(&lp->pdev->dev, skb->data,
826*4882a593Smuzhiyun 						  skb->len, DMA_TO_DEVICE));
827*4882a593Smuzhiyun 	descptr->status = DSC_OWNER_MAC;
828*4882a593Smuzhiyun 
829*4882a593Smuzhiyun 	skb_tx_timestamp(skb);
830*4882a593Smuzhiyun 
831*4882a593Smuzhiyun 	/* Trigger the MAC to check the TX descriptor */
832*4882a593Smuzhiyun 	if (!netdev_xmit_more() || netif_queue_stopped(dev))
833*4882a593Smuzhiyun 		iowrite16(TM2TX, ioaddr + MTPR);
834*4882a593Smuzhiyun 	lp->tx_insert_ptr = descptr->vndescp;
835*4882a593Smuzhiyun 
836*4882a593Smuzhiyun 	/* If no tx resource, stop */
837*4882a593Smuzhiyun 	if (!lp->tx_free_desc)
838*4882a593Smuzhiyun 		netif_stop_queue(dev);
839*4882a593Smuzhiyun 
840*4882a593Smuzhiyun 	spin_unlock_irqrestore(&lp->lock, flags);
841*4882a593Smuzhiyun 
842*4882a593Smuzhiyun 	return NETDEV_TX_OK;
843*4882a593Smuzhiyun }
844*4882a593Smuzhiyun 
r6040_multicast_list(struct net_device * dev)845*4882a593Smuzhiyun static void r6040_multicast_list(struct net_device *dev)
846*4882a593Smuzhiyun {
847*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
848*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
849*4882a593Smuzhiyun 	unsigned long flags;
850*4882a593Smuzhiyun 	struct netdev_hw_addr *ha;
851*4882a593Smuzhiyun 	int i;
852*4882a593Smuzhiyun 	u16 *adrp;
853*4882a593Smuzhiyun 	u16 hash_table[4] = { 0 };
854*4882a593Smuzhiyun 
855*4882a593Smuzhiyun 	spin_lock_irqsave(&lp->lock, flags);
856*4882a593Smuzhiyun 
857*4882a593Smuzhiyun 	/* Keep our MAC Address */
858*4882a593Smuzhiyun 	adrp = (u16 *)dev->dev_addr;
859*4882a593Smuzhiyun 	iowrite16(adrp[0], ioaddr + MID_0L);
860*4882a593Smuzhiyun 	iowrite16(adrp[1], ioaddr + MID_0M);
861*4882a593Smuzhiyun 	iowrite16(adrp[2], ioaddr + MID_0H);
862*4882a593Smuzhiyun 
863*4882a593Smuzhiyun 	/* Clear AMCP & PROM bits */
864*4882a593Smuzhiyun 	lp->mcr0 = ioread16(ioaddr + MCR0) & ~(MCR0_PROMISC | MCR0_HASH_EN);
865*4882a593Smuzhiyun 
866*4882a593Smuzhiyun 	/* Promiscuous mode */
867*4882a593Smuzhiyun 	if (dev->flags & IFF_PROMISC)
868*4882a593Smuzhiyun 		lp->mcr0 |= MCR0_PROMISC;
869*4882a593Smuzhiyun 
870*4882a593Smuzhiyun 	/* Enable multicast hash table function to
871*4882a593Smuzhiyun 	 * receive all multicast packets. */
872*4882a593Smuzhiyun 	else if (dev->flags & IFF_ALLMULTI) {
873*4882a593Smuzhiyun 		lp->mcr0 |= MCR0_HASH_EN;
874*4882a593Smuzhiyun 
875*4882a593Smuzhiyun 		for (i = 0; i < MCAST_MAX ; i++) {
876*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1L + 8 * i);
877*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1M + 8 * i);
878*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1H + 8 * i);
879*4882a593Smuzhiyun 		}
880*4882a593Smuzhiyun 
881*4882a593Smuzhiyun 		for (i = 0; i < 4; i++)
882*4882a593Smuzhiyun 			hash_table[i] = 0xffff;
883*4882a593Smuzhiyun 	}
884*4882a593Smuzhiyun 	/* Use internal multicast address registers if the number of
885*4882a593Smuzhiyun 	 * multicast addresses is not greater than MCAST_MAX. */
886*4882a593Smuzhiyun 	else if (netdev_mc_count(dev) <= MCAST_MAX) {
887*4882a593Smuzhiyun 		i = 0;
888*4882a593Smuzhiyun 		netdev_for_each_mc_addr(ha, dev) {
889*4882a593Smuzhiyun 			u16 *adrp = (u16 *) ha->addr;
890*4882a593Smuzhiyun 			iowrite16(adrp[0], ioaddr + MID_1L + 8 * i);
891*4882a593Smuzhiyun 			iowrite16(adrp[1], ioaddr + MID_1M + 8 * i);
892*4882a593Smuzhiyun 			iowrite16(adrp[2], ioaddr + MID_1H + 8 * i);
893*4882a593Smuzhiyun 			i++;
894*4882a593Smuzhiyun 		}
895*4882a593Smuzhiyun 		while (i < MCAST_MAX) {
896*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1L + 8 * i);
897*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1M + 8 * i);
898*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1H + 8 * i);
899*4882a593Smuzhiyun 			i++;
900*4882a593Smuzhiyun 		}
901*4882a593Smuzhiyun 	}
902*4882a593Smuzhiyun 	/* Otherwise, Enable multicast hash table function. */
903*4882a593Smuzhiyun 	else {
904*4882a593Smuzhiyun 		u32 crc;
905*4882a593Smuzhiyun 
906*4882a593Smuzhiyun 		lp->mcr0 |= MCR0_HASH_EN;
907*4882a593Smuzhiyun 
908*4882a593Smuzhiyun 		for (i = 0; i < MCAST_MAX ; i++) {
909*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1L + 8 * i);
910*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1M + 8 * i);
911*4882a593Smuzhiyun 			iowrite16(0, ioaddr + MID_1H + 8 * i);
912*4882a593Smuzhiyun 		}
913*4882a593Smuzhiyun 
914*4882a593Smuzhiyun 		/* Build multicast hash table */
915*4882a593Smuzhiyun 		netdev_for_each_mc_addr(ha, dev) {
916*4882a593Smuzhiyun 			u8 *addrs = ha->addr;
917*4882a593Smuzhiyun 
918*4882a593Smuzhiyun 			crc = ether_crc(ETH_ALEN, addrs);
919*4882a593Smuzhiyun 			crc >>= 26;
920*4882a593Smuzhiyun 			hash_table[crc >> 4] |= 1 << (crc & 0xf);
921*4882a593Smuzhiyun 		}
922*4882a593Smuzhiyun 	}
923*4882a593Smuzhiyun 
924*4882a593Smuzhiyun 	iowrite16(lp->mcr0, ioaddr + MCR0);
925*4882a593Smuzhiyun 
926*4882a593Smuzhiyun 	/* Fill the MAC hash tables with their values */
927*4882a593Smuzhiyun 	if (lp->mcr0 & MCR0_HASH_EN) {
928*4882a593Smuzhiyun 		iowrite16(hash_table[0], ioaddr + MAR0);
929*4882a593Smuzhiyun 		iowrite16(hash_table[1], ioaddr + MAR1);
930*4882a593Smuzhiyun 		iowrite16(hash_table[2], ioaddr + MAR2);
931*4882a593Smuzhiyun 		iowrite16(hash_table[3], ioaddr + MAR3);
932*4882a593Smuzhiyun 	}
933*4882a593Smuzhiyun 
934*4882a593Smuzhiyun 	spin_unlock_irqrestore(&lp->lock, flags);
935*4882a593Smuzhiyun }
936*4882a593Smuzhiyun 
netdev_get_drvinfo(struct net_device * dev,struct ethtool_drvinfo * info)937*4882a593Smuzhiyun static void netdev_get_drvinfo(struct net_device *dev,
938*4882a593Smuzhiyun 			struct ethtool_drvinfo *info)
939*4882a593Smuzhiyun {
940*4882a593Smuzhiyun 	struct r6040_private *rp = netdev_priv(dev);
941*4882a593Smuzhiyun 
942*4882a593Smuzhiyun 	strlcpy(info->driver, DRV_NAME, sizeof(info->driver));
943*4882a593Smuzhiyun 	strlcpy(info->version, DRV_VERSION, sizeof(info->version));
944*4882a593Smuzhiyun 	strlcpy(info->bus_info, pci_name(rp->pdev), sizeof(info->bus_info));
945*4882a593Smuzhiyun }
946*4882a593Smuzhiyun 
947*4882a593Smuzhiyun static const struct ethtool_ops netdev_ethtool_ops = {
948*4882a593Smuzhiyun 	.get_drvinfo		= netdev_get_drvinfo,
949*4882a593Smuzhiyun 	.get_link		= ethtool_op_get_link,
950*4882a593Smuzhiyun 	.get_ts_info		= ethtool_op_get_ts_info,
951*4882a593Smuzhiyun 	.get_link_ksettings     = phy_ethtool_get_link_ksettings,
952*4882a593Smuzhiyun 	.set_link_ksettings     = phy_ethtool_set_link_ksettings,
953*4882a593Smuzhiyun };
954*4882a593Smuzhiyun 
955*4882a593Smuzhiyun static const struct net_device_ops r6040_netdev_ops = {
956*4882a593Smuzhiyun 	.ndo_open		= r6040_open,
957*4882a593Smuzhiyun 	.ndo_stop		= r6040_close,
958*4882a593Smuzhiyun 	.ndo_start_xmit		= r6040_start_xmit,
959*4882a593Smuzhiyun 	.ndo_get_stats		= r6040_get_stats,
960*4882a593Smuzhiyun 	.ndo_set_rx_mode	= r6040_multicast_list,
961*4882a593Smuzhiyun 	.ndo_validate_addr	= eth_validate_addr,
962*4882a593Smuzhiyun 	.ndo_set_mac_address	= eth_mac_addr,
963*4882a593Smuzhiyun 	.ndo_do_ioctl		= phy_do_ioctl,
964*4882a593Smuzhiyun 	.ndo_tx_timeout		= r6040_tx_timeout,
965*4882a593Smuzhiyun #ifdef CONFIG_NET_POLL_CONTROLLER
966*4882a593Smuzhiyun 	.ndo_poll_controller	= r6040_poll_controller,
967*4882a593Smuzhiyun #endif
968*4882a593Smuzhiyun };
969*4882a593Smuzhiyun 
r6040_adjust_link(struct net_device * dev)970*4882a593Smuzhiyun static void r6040_adjust_link(struct net_device *dev)
971*4882a593Smuzhiyun {
972*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
973*4882a593Smuzhiyun 	struct phy_device *phydev = dev->phydev;
974*4882a593Smuzhiyun 	int status_changed = 0;
975*4882a593Smuzhiyun 	void __iomem *ioaddr = lp->base;
976*4882a593Smuzhiyun 
977*4882a593Smuzhiyun 	BUG_ON(!phydev);
978*4882a593Smuzhiyun 
979*4882a593Smuzhiyun 	if (lp->old_link != phydev->link) {
980*4882a593Smuzhiyun 		status_changed = 1;
981*4882a593Smuzhiyun 		lp->old_link = phydev->link;
982*4882a593Smuzhiyun 	}
983*4882a593Smuzhiyun 
984*4882a593Smuzhiyun 	/* reflect duplex change */
985*4882a593Smuzhiyun 	if (phydev->link && (lp->old_duplex != phydev->duplex)) {
986*4882a593Smuzhiyun 		lp->mcr0 |= (phydev->duplex == DUPLEX_FULL ? MCR0_FD : 0);
987*4882a593Smuzhiyun 		iowrite16(lp->mcr0, ioaddr);
988*4882a593Smuzhiyun 
989*4882a593Smuzhiyun 		status_changed = 1;
990*4882a593Smuzhiyun 		lp->old_duplex = phydev->duplex;
991*4882a593Smuzhiyun 	}
992*4882a593Smuzhiyun 
993*4882a593Smuzhiyun 	if (status_changed)
994*4882a593Smuzhiyun 		phy_print_status(phydev);
995*4882a593Smuzhiyun }
996*4882a593Smuzhiyun 
r6040_mii_probe(struct net_device * dev)997*4882a593Smuzhiyun static int r6040_mii_probe(struct net_device *dev)
998*4882a593Smuzhiyun {
999*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
1000*4882a593Smuzhiyun 	struct phy_device *phydev = NULL;
1001*4882a593Smuzhiyun 
1002*4882a593Smuzhiyun 	phydev = phy_find_first(lp->mii_bus);
1003*4882a593Smuzhiyun 	if (!phydev) {
1004*4882a593Smuzhiyun 		dev_err(&lp->pdev->dev, "no PHY found\n");
1005*4882a593Smuzhiyun 		return -ENODEV;
1006*4882a593Smuzhiyun 	}
1007*4882a593Smuzhiyun 
1008*4882a593Smuzhiyun 	phydev = phy_connect(dev, phydev_name(phydev), &r6040_adjust_link,
1009*4882a593Smuzhiyun 			     PHY_INTERFACE_MODE_MII);
1010*4882a593Smuzhiyun 
1011*4882a593Smuzhiyun 	if (IS_ERR(phydev)) {
1012*4882a593Smuzhiyun 		dev_err(&lp->pdev->dev, "could not attach to PHY\n");
1013*4882a593Smuzhiyun 		return PTR_ERR(phydev);
1014*4882a593Smuzhiyun 	}
1015*4882a593Smuzhiyun 
1016*4882a593Smuzhiyun 	phy_set_max_speed(phydev, SPEED_100);
1017*4882a593Smuzhiyun 
1018*4882a593Smuzhiyun 	lp->old_link = 0;
1019*4882a593Smuzhiyun 	lp->old_duplex = -1;
1020*4882a593Smuzhiyun 
1021*4882a593Smuzhiyun 	phy_attached_info(phydev);
1022*4882a593Smuzhiyun 
1023*4882a593Smuzhiyun 	return 0;
1024*4882a593Smuzhiyun }
1025*4882a593Smuzhiyun 
r6040_init_one(struct pci_dev * pdev,const struct pci_device_id * ent)1026*4882a593Smuzhiyun static int r6040_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
1027*4882a593Smuzhiyun {
1028*4882a593Smuzhiyun 	struct net_device *dev;
1029*4882a593Smuzhiyun 	struct r6040_private *lp;
1030*4882a593Smuzhiyun 	void __iomem *ioaddr;
1031*4882a593Smuzhiyun 	int err, io_size = R6040_IO_SIZE;
1032*4882a593Smuzhiyun 	static int card_idx = -1;
1033*4882a593Smuzhiyun 	int bar = 0;
1034*4882a593Smuzhiyun 	u16 *adrp;
1035*4882a593Smuzhiyun 
1036*4882a593Smuzhiyun 	pr_info("%s\n", version);
1037*4882a593Smuzhiyun 
1038*4882a593Smuzhiyun 	err = pci_enable_device(pdev);
1039*4882a593Smuzhiyun 	if (err)
1040*4882a593Smuzhiyun 		goto err_out;
1041*4882a593Smuzhiyun 
1042*4882a593Smuzhiyun 	/* this should always be supported */
1043*4882a593Smuzhiyun 	err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
1044*4882a593Smuzhiyun 	if (err) {
1045*4882a593Smuzhiyun 		dev_err(&pdev->dev, "32-bit PCI DMA addresses not supported by the card\n");
1046*4882a593Smuzhiyun 		goto err_out_disable_dev;
1047*4882a593Smuzhiyun 	}
1048*4882a593Smuzhiyun 	err = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
1049*4882a593Smuzhiyun 	if (err) {
1050*4882a593Smuzhiyun 		dev_err(&pdev->dev, "32-bit PCI DMA addresses not supported by the card\n");
1051*4882a593Smuzhiyun 		goto err_out_disable_dev;
1052*4882a593Smuzhiyun 	}
1053*4882a593Smuzhiyun 
1054*4882a593Smuzhiyun 	/* IO Size check */
1055*4882a593Smuzhiyun 	if (pci_resource_len(pdev, bar) < io_size) {
1056*4882a593Smuzhiyun 		dev_err(&pdev->dev, "Insufficient PCI resources, aborting\n");
1057*4882a593Smuzhiyun 		err = -EIO;
1058*4882a593Smuzhiyun 		goto err_out_disable_dev;
1059*4882a593Smuzhiyun 	}
1060*4882a593Smuzhiyun 
1061*4882a593Smuzhiyun 	pci_set_master(pdev);
1062*4882a593Smuzhiyun 
1063*4882a593Smuzhiyun 	dev = alloc_etherdev(sizeof(struct r6040_private));
1064*4882a593Smuzhiyun 	if (!dev) {
1065*4882a593Smuzhiyun 		err = -ENOMEM;
1066*4882a593Smuzhiyun 		goto err_out_disable_dev;
1067*4882a593Smuzhiyun 	}
1068*4882a593Smuzhiyun 	SET_NETDEV_DEV(dev, &pdev->dev);
1069*4882a593Smuzhiyun 	lp = netdev_priv(dev);
1070*4882a593Smuzhiyun 
1071*4882a593Smuzhiyun 	err = pci_request_regions(pdev, DRV_NAME);
1072*4882a593Smuzhiyun 
1073*4882a593Smuzhiyun 	if (err) {
1074*4882a593Smuzhiyun 		dev_err(&pdev->dev, "Failed to request PCI regions\n");
1075*4882a593Smuzhiyun 		goto err_out_free_dev;
1076*4882a593Smuzhiyun 	}
1077*4882a593Smuzhiyun 
1078*4882a593Smuzhiyun 	ioaddr = pci_iomap(pdev, bar, io_size);
1079*4882a593Smuzhiyun 	if (!ioaddr) {
1080*4882a593Smuzhiyun 		dev_err(&pdev->dev, "ioremap failed for device\n");
1081*4882a593Smuzhiyun 		err = -EIO;
1082*4882a593Smuzhiyun 		goto err_out_free_res;
1083*4882a593Smuzhiyun 	}
1084*4882a593Smuzhiyun 
1085*4882a593Smuzhiyun 	/* If PHY status change register is still set to zero it means the
1086*4882a593Smuzhiyun 	 * bootloader didn't initialize it, so we set it to:
1087*4882a593Smuzhiyun 	 * - enable phy status change
1088*4882a593Smuzhiyun 	 * - enable all phy addresses
1089*4882a593Smuzhiyun 	 * - set to lowest timer divider */
1090*4882a593Smuzhiyun 	if (ioread16(ioaddr + PHY_CC) == 0)
1091*4882a593Smuzhiyun 		iowrite16(SCEN | PHY_MAX_ADDR << PHYAD_SHIFT |
1092*4882a593Smuzhiyun 				7 << TMRDIV_SHIFT, ioaddr + PHY_CC);
1093*4882a593Smuzhiyun 
1094*4882a593Smuzhiyun 	/* Init system & device */
1095*4882a593Smuzhiyun 	lp->base = ioaddr;
1096*4882a593Smuzhiyun 	dev->irq = pdev->irq;
1097*4882a593Smuzhiyun 
1098*4882a593Smuzhiyun 	spin_lock_init(&lp->lock);
1099*4882a593Smuzhiyun 	pci_set_drvdata(pdev, dev);
1100*4882a593Smuzhiyun 
1101*4882a593Smuzhiyun 	/* Set MAC address */
1102*4882a593Smuzhiyun 	card_idx++;
1103*4882a593Smuzhiyun 
1104*4882a593Smuzhiyun 	adrp = (u16 *)dev->dev_addr;
1105*4882a593Smuzhiyun 	adrp[0] = ioread16(ioaddr + MID_0L);
1106*4882a593Smuzhiyun 	adrp[1] = ioread16(ioaddr + MID_0M);
1107*4882a593Smuzhiyun 	adrp[2] = ioread16(ioaddr + MID_0H);
1108*4882a593Smuzhiyun 
1109*4882a593Smuzhiyun 	/* Some bootloader/BIOSes do not initialize
1110*4882a593Smuzhiyun 	 * MAC address, warn about that */
1111*4882a593Smuzhiyun 	if (!(adrp[0] || adrp[1] || adrp[2])) {
1112*4882a593Smuzhiyun 		netdev_warn(dev, "MAC address not initialized, "
1113*4882a593Smuzhiyun 					"generating random\n");
1114*4882a593Smuzhiyun 		eth_hw_addr_random(dev);
1115*4882a593Smuzhiyun 	}
1116*4882a593Smuzhiyun 
1117*4882a593Smuzhiyun 	/* Link new device into r6040_root_dev */
1118*4882a593Smuzhiyun 	lp->pdev = pdev;
1119*4882a593Smuzhiyun 	lp->dev = dev;
1120*4882a593Smuzhiyun 
1121*4882a593Smuzhiyun 	/* Init RDC private data */
1122*4882a593Smuzhiyun 	lp->mcr0 = MCR0_XMTEN | MCR0_RCVEN;
1123*4882a593Smuzhiyun 
1124*4882a593Smuzhiyun 	/* The RDC-specific entries in the device structure. */
1125*4882a593Smuzhiyun 	dev->netdev_ops = &r6040_netdev_ops;
1126*4882a593Smuzhiyun 	dev->ethtool_ops = &netdev_ethtool_ops;
1127*4882a593Smuzhiyun 	dev->watchdog_timeo = TX_TIMEOUT;
1128*4882a593Smuzhiyun 
1129*4882a593Smuzhiyun 	netif_napi_add(dev, &lp->napi, r6040_poll, 64);
1130*4882a593Smuzhiyun 
1131*4882a593Smuzhiyun 	lp->mii_bus = mdiobus_alloc();
1132*4882a593Smuzhiyun 	if (!lp->mii_bus) {
1133*4882a593Smuzhiyun 		dev_err(&pdev->dev, "mdiobus_alloc() failed\n");
1134*4882a593Smuzhiyun 		err = -ENOMEM;
1135*4882a593Smuzhiyun 		goto err_out_unmap;
1136*4882a593Smuzhiyun 	}
1137*4882a593Smuzhiyun 
1138*4882a593Smuzhiyun 	lp->mii_bus->priv = dev;
1139*4882a593Smuzhiyun 	lp->mii_bus->read = r6040_mdiobus_read;
1140*4882a593Smuzhiyun 	lp->mii_bus->write = r6040_mdiobus_write;
1141*4882a593Smuzhiyun 	lp->mii_bus->name = "r6040_eth_mii";
1142*4882a593Smuzhiyun 	snprintf(lp->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
1143*4882a593Smuzhiyun 		dev_name(&pdev->dev), card_idx);
1144*4882a593Smuzhiyun 
1145*4882a593Smuzhiyun 	err = mdiobus_register(lp->mii_bus);
1146*4882a593Smuzhiyun 	if (err) {
1147*4882a593Smuzhiyun 		dev_err(&pdev->dev, "failed to register MII bus\n");
1148*4882a593Smuzhiyun 		goto err_out_mdio;
1149*4882a593Smuzhiyun 	}
1150*4882a593Smuzhiyun 
1151*4882a593Smuzhiyun 	err = r6040_mii_probe(dev);
1152*4882a593Smuzhiyun 	if (err) {
1153*4882a593Smuzhiyun 		dev_err(&pdev->dev, "failed to probe MII bus\n");
1154*4882a593Smuzhiyun 		goto err_out_mdio_unregister;
1155*4882a593Smuzhiyun 	}
1156*4882a593Smuzhiyun 
1157*4882a593Smuzhiyun 	/* Register net device. After this dev->name assign */
1158*4882a593Smuzhiyun 	err = register_netdev(dev);
1159*4882a593Smuzhiyun 	if (err) {
1160*4882a593Smuzhiyun 		dev_err(&pdev->dev, "Failed to register net device\n");
1161*4882a593Smuzhiyun 		goto err_out_mdio_unregister;
1162*4882a593Smuzhiyun 	}
1163*4882a593Smuzhiyun 	return 0;
1164*4882a593Smuzhiyun 
1165*4882a593Smuzhiyun err_out_mdio_unregister:
1166*4882a593Smuzhiyun 	mdiobus_unregister(lp->mii_bus);
1167*4882a593Smuzhiyun err_out_mdio:
1168*4882a593Smuzhiyun 	mdiobus_free(lp->mii_bus);
1169*4882a593Smuzhiyun err_out_unmap:
1170*4882a593Smuzhiyun 	netif_napi_del(&lp->napi);
1171*4882a593Smuzhiyun 	pci_iounmap(pdev, ioaddr);
1172*4882a593Smuzhiyun err_out_free_res:
1173*4882a593Smuzhiyun 	pci_release_regions(pdev);
1174*4882a593Smuzhiyun err_out_free_dev:
1175*4882a593Smuzhiyun 	free_netdev(dev);
1176*4882a593Smuzhiyun err_out_disable_dev:
1177*4882a593Smuzhiyun 	pci_disable_device(pdev);
1178*4882a593Smuzhiyun err_out:
1179*4882a593Smuzhiyun 	return err;
1180*4882a593Smuzhiyun }
1181*4882a593Smuzhiyun 
r6040_remove_one(struct pci_dev * pdev)1182*4882a593Smuzhiyun static void r6040_remove_one(struct pci_dev *pdev)
1183*4882a593Smuzhiyun {
1184*4882a593Smuzhiyun 	struct net_device *dev = pci_get_drvdata(pdev);
1185*4882a593Smuzhiyun 	struct r6040_private *lp = netdev_priv(dev);
1186*4882a593Smuzhiyun 
1187*4882a593Smuzhiyun 	unregister_netdev(dev);
1188*4882a593Smuzhiyun 	mdiobus_unregister(lp->mii_bus);
1189*4882a593Smuzhiyun 	mdiobus_free(lp->mii_bus);
1190*4882a593Smuzhiyun 	netif_napi_del(&lp->napi);
1191*4882a593Smuzhiyun 	pci_iounmap(pdev, lp->base);
1192*4882a593Smuzhiyun 	pci_release_regions(pdev);
1193*4882a593Smuzhiyun 	free_netdev(dev);
1194*4882a593Smuzhiyun 	pci_disable_device(pdev);
1195*4882a593Smuzhiyun }
1196*4882a593Smuzhiyun 
1197*4882a593Smuzhiyun 
1198*4882a593Smuzhiyun static const struct pci_device_id r6040_pci_tbl[] = {
1199*4882a593Smuzhiyun 	{ PCI_DEVICE(PCI_VENDOR_ID_RDC, 0x6040) },
1200*4882a593Smuzhiyun 	{ 0 }
1201*4882a593Smuzhiyun };
1202*4882a593Smuzhiyun MODULE_DEVICE_TABLE(pci, r6040_pci_tbl);
1203*4882a593Smuzhiyun 
1204*4882a593Smuzhiyun static struct pci_driver r6040_driver = {
1205*4882a593Smuzhiyun 	.name		= DRV_NAME,
1206*4882a593Smuzhiyun 	.id_table	= r6040_pci_tbl,
1207*4882a593Smuzhiyun 	.probe		= r6040_init_one,
1208*4882a593Smuzhiyun 	.remove		= r6040_remove_one,
1209*4882a593Smuzhiyun };
1210*4882a593Smuzhiyun 
1211*4882a593Smuzhiyun module_pci_driver(r6040_driver);
1212