1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * FCC driver for Motorola MPC82xx (PQ2).
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * Copyright (c) 2003 Intracom S.A.
5*4882a593Smuzhiyun * by Pantelis Antoniou <panto@intracom.gr>
6*4882a593Smuzhiyun *
7*4882a593Smuzhiyun * 2005 (c) MontaVista Software, Inc.
8*4882a593Smuzhiyun * Vitaly Bordug <vbordug@ru.mvista.com>
9*4882a593Smuzhiyun *
10*4882a593Smuzhiyun * This file is licensed under the terms of the GNU General Public License
11*4882a593Smuzhiyun * version 2. This program is licensed "as is" without any warranty of any
12*4882a593Smuzhiyun * kind, whether express or implied.
13*4882a593Smuzhiyun */
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun #include <linux/module.h>
16*4882a593Smuzhiyun #include <linux/kernel.h>
17*4882a593Smuzhiyun #include <linux/types.h>
18*4882a593Smuzhiyun #include <linux/string.h>
19*4882a593Smuzhiyun #include <linux/ptrace.h>
20*4882a593Smuzhiyun #include <linux/errno.h>
21*4882a593Smuzhiyun #include <linux/ioport.h>
22*4882a593Smuzhiyun #include <linux/interrupt.h>
23*4882a593Smuzhiyun #include <linux/delay.h>
24*4882a593Smuzhiyun #include <linux/netdevice.h>
25*4882a593Smuzhiyun #include <linux/etherdevice.h>
26*4882a593Smuzhiyun #include <linux/skbuff.h>
27*4882a593Smuzhiyun #include <linux/spinlock.h>
28*4882a593Smuzhiyun #include <linux/mii.h>
29*4882a593Smuzhiyun #include <linux/ethtool.h>
30*4882a593Smuzhiyun #include <linux/bitops.h>
31*4882a593Smuzhiyun #include <linux/fs.h>
32*4882a593Smuzhiyun #include <linux/platform_device.h>
33*4882a593Smuzhiyun #include <linux/phy.h>
34*4882a593Smuzhiyun #include <linux/of_address.h>
35*4882a593Smuzhiyun #include <linux/of_device.h>
36*4882a593Smuzhiyun #include <linux/of_irq.h>
37*4882a593Smuzhiyun #include <linux/gfp.h>
38*4882a593Smuzhiyun #include <linux/pgtable.h>
39*4882a593Smuzhiyun
40*4882a593Smuzhiyun #include <asm/immap_cpm2.h>
41*4882a593Smuzhiyun #include <asm/mpc8260.h>
42*4882a593Smuzhiyun #include <asm/cpm2.h>
43*4882a593Smuzhiyun
44*4882a593Smuzhiyun #include <asm/irq.h>
45*4882a593Smuzhiyun #include <linux/uaccess.h>
46*4882a593Smuzhiyun
47*4882a593Smuzhiyun #include "fs_enet.h"
48*4882a593Smuzhiyun
49*4882a593Smuzhiyun /*************************************************/
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun /* FCC access macros */
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun /* write, read, set bits, clear bits */
54*4882a593Smuzhiyun #define W32(_p, _m, _v) out_be32(&(_p)->_m, (_v))
55*4882a593Smuzhiyun #define R32(_p, _m) in_be32(&(_p)->_m)
56*4882a593Smuzhiyun #define S32(_p, _m, _v) W32(_p, _m, R32(_p, _m) | (_v))
57*4882a593Smuzhiyun #define C32(_p, _m, _v) W32(_p, _m, R32(_p, _m) & ~(_v))
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun #define W16(_p, _m, _v) out_be16(&(_p)->_m, (_v))
60*4882a593Smuzhiyun #define R16(_p, _m) in_be16(&(_p)->_m)
61*4882a593Smuzhiyun #define S16(_p, _m, _v) W16(_p, _m, R16(_p, _m) | (_v))
62*4882a593Smuzhiyun #define C16(_p, _m, _v) W16(_p, _m, R16(_p, _m) & ~(_v))
63*4882a593Smuzhiyun
64*4882a593Smuzhiyun #define W8(_p, _m, _v) out_8(&(_p)->_m, (_v))
65*4882a593Smuzhiyun #define R8(_p, _m) in_8(&(_p)->_m)
66*4882a593Smuzhiyun #define S8(_p, _m, _v) W8(_p, _m, R8(_p, _m) | (_v))
67*4882a593Smuzhiyun #define C8(_p, _m, _v) W8(_p, _m, R8(_p, _m) & ~(_v))
68*4882a593Smuzhiyun
69*4882a593Smuzhiyun /*************************************************/
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun #define FCC_MAX_MULTICAST_ADDRS 64
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun #define mk_mii_read(REG) (0x60020000 | ((REG & 0x1f) << 18))
74*4882a593Smuzhiyun #define mk_mii_write(REG, VAL) (0x50020000 | ((REG & 0x1f) << 18) | (VAL & 0xffff))
75*4882a593Smuzhiyun #define mk_mii_end 0
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun #define MAX_CR_CMD_LOOPS 10000
78*4882a593Smuzhiyun
fcc_cr_cmd(struct fs_enet_private * fep,u32 op)79*4882a593Smuzhiyun static inline int fcc_cr_cmd(struct fs_enet_private *fep, u32 op)
80*4882a593Smuzhiyun {
81*4882a593Smuzhiyun const struct fs_platform_info *fpi = fep->fpi;
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun return cpm_command(fpi->cp_command, op);
84*4882a593Smuzhiyun }
85*4882a593Smuzhiyun
do_pd_setup(struct fs_enet_private * fep)86*4882a593Smuzhiyun static int do_pd_setup(struct fs_enet_private *fep)
87*4882a593Smuzhiyun {
88*4882a593Smuzhiyun struct platform_device *ofdev = to_platform_device(fep->dev);
89*4882a593Smuzhiyun struct fs_platform_info *fpi = fep->fpi;
90*4882a593Smuzhiyun int ret = -EINVAL;
91*4882a593Smuzhiyun
92*4882a593Smuzhiyun fep->interrupt = irq_of_parse_and_map(ofdev->dev.of_node, 0);
93*4882a593Smuzhiyun if (!fep->interrupt)
94*4882a593Smuzhiyun goto out;
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun fep->fcc.fccp = of_iomap(ofdev->dev.of_node, 0);
97*4882a593Smuzhiyun if (!fep->fcc.fccp)
98*4882a593Smuzhiyun goto out;
99*4882a593Smuzhiyun
100*4882a593Smuzhiyun fep->fcc.ep = of_iomap(ofdev->dev.of_node, 1);
101*4882a593Smuzhiyun if (!fep->fcc.ep)
102*4882a593Smuzhiyun goto out_fccp;
103*4882a593Smuzhiyun
104*4882a593Smuzhiyun fep->fcc.fcccp = of_iomap(ofdev->dev.of_node, 2);
105*4882a593Smuzhiyun if (!fep->fcc.fcccp)
106*4882a593Smuzhiyun goto out_ep;
107*4882a593Smuzhiyun
108*4882a593Smuzhiyun fep->fcc.mem = (void __iomem *)cpm2_immr;
109*4882a593Smuzhiyun fpi->dpram_offset = cpm_dpalloc(128, 32);
110*4882a593Smuzhiyun if (IS_ERR_VALUE(fpi->dpram_offset)) {
111*4882a593Smuzhiyun ret = fpi->dpram_offset;
112*4882a593Smuzhiyun goto out_fcccp;
113*4882a593Smuzhiyun }
114*4882a593Smuzhiyun
115*4882a593Smuzhiyun return 0;
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun out_fcccp:
118*4882a593Smuzhiyun iounmap(fep->fcc.fcccp);
119*4882a593Smuzhiyun out_ep:
120*4882a593Smuzhiyun iounmap(fep->fcc.ep);
121*4882a593Smuzhiyun out_fccp:
122*4882a593Smuzhiyun iounmap(fep->fcc.fccp);
123*4882a593Smuzhiyun out:
124*4882a593Smuzhiyun return ret;
125*4882a593Smuzhiyun }
126*4882a593Smuzhiyun
127*4882a593Smuzhiyun #define FCC_NAPI_EVENT_MSK (FCC_ENET_RXF | FCC_ENET_RXB | FCC_ENET_TXB)
128*4882a593Smuzhiyun #define FCC_EVENT (FCC_ENET_RXF | FCC_ENET_TXB)
129*4882a593Smuzhiyun #define FCC_ERR_EVENT_MSK (FCC_ENET_TXE)
130*4882a593Smuzhiyun
setup_data(struct net_device * dev)131*4882a593Smuzhiyun static int setup_data(struct net_device *dev)
132*4882a593Smuzhiyun {
133*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
134*4882a593Smuzhiyun
135*4882a593Smuzhiyun if (do_pd_setup(fep) != 0)
136*4882a593Smuzhiyun return -EINVAL;
137*4882a593Smuzhiyun
138*4882a593Smuzhiyun fep->ev_napi = FCC_NAPI_EVENT_MSK;
139*4882a593Smuzhiyun fep->ev = FCC_EVENT;
140*4882a593Smuzhiyun fep->ev_err = FCC_ERR_EVENT_MSK;
141*4882a593Smuzhiyun
142*4882a593Smuzhiyun return 0;
143*4882a593Smuzhiyun }
144*4882a593Smuzhiyun
allocate_bd(struct net_device * dev)145*4882a593Smuzhiyun static int allocate_bd(struct net_device *dev)
146*4882a593Smuzhiyun {
147*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
148*4882a593Smuzhiyun const struct fs_platform_info *fpi = fep->fpi;
149*4882a593Smuzhiyun
150*4882a593Smuzhiyun fep->ring_base = (void __iomem __force *)dma_alloc_coherent(fep->dev,
151*4882a593Smuzhiyun (fpi->tx_ring + fpi->rx_ring) *
152*4882a593Smuzhiyun sizeof(cbd_t), &fep->ring_mem_addr,
153*4882a593Smuzhiyun GFP_KERNEL);
154*4882a593Smuzhiyun if (fep->ring_base == NULL)
155*4882a593Smuzhiyun return -ENOMEM;
156*4882a593Smuzhiyun
157*4882a593Smuzhiyun return 0;
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun
free_bd(struct net_device * dev)160*4882a593Smuzhiyun static void free_bd(struct net_device *dev)
161*4882a593Smuzhiyun {
162*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
163*4882a593Smuzhiyun const struct fs_platform_info *fpi = fep->fpi;
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun if (fep->ring_base)
166*4882a593Smuzhiyun dma_free_coherent(fep->dev,
167*4882a593Smuzhiyun (fpi->tx_ring + fpi->rx_ring) * sizeof(cbd_t),
168*4882a593Smuzhiyun (void __force *)fep->ring_base, fep->ring_mem_addr);
169*4882a593Smuzhiyun }
170*4882a593Smuzhiyun
cleanup_data(struct net_device * dev)171*4882a593Smuzhiyun static void cleanup_data(struct net_device *dev)
172*4882a593Smuzhiyun {
173*4882a593Smuzhiyun /* nothing */
174*4882a593Smuzhiyun }
175*4882a593Smuzhiyun
set_promiscuous_mode(struct net_device * dev)176*4882a593Smuzhiyun static void set_promiscuous_mode(struct net_device *dev)
177*4882a593Smuzhiyun {
178*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
179*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
180*4882a593Smuzhiyun
181*4882a593Smuzhiyun S32(fccp, fcc_fpsmr, FCC_PSMR_PRO);
182*4882a593Smuzhiyun }
183*4882a593Smuzhiyun
set_multicast_start(struct net_device * dev)184*4882a593Smuzhiyun static void set_multicast_start(struct net_device *dev)
185*4882a593Smuzhiyun {
186*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
187*4882a593Smuzhiyun fcc_enet_t __iomem *ep = fep->fcc.ep;
188*4882a593Smuzhiyun
189*4882a593Smuzhiyun W32(ep, fen_gaddrh, 0);
190*4882a593Smuzhiyun W32(ep, fen_gaddrl, 0);
191*4882a593Smuzhiyun }
192*4882a593Smuzhiyun
set_multicast_one(struct net_device * dev,const u8 * mac)193*4882a593Smuzhiyun static void set_multicast_one(struct net_device *dev, const u8 *mac)
194*4882a593Smuzhiyun {
195*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
196*4882a593Smuzhiyun fcc_enet_t __iomem *ep = fep->fcc.ep;
197*4882a593Smuzhiyun u16 taddrh, taddrm, taddrl;
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun taddrh = ((u16)mac[5] << 8) | mac[4];
200*4882a593Smuzhiyun taddrm = ((u16)mac[3] << 8) | mac[2];
201*4882a593Smuzhiyun taddrl = ((u16)mac[1] << 8) | mac[0];
202*4882a593Smuzhiyun
203*4882a593Smuzhiyun W16(ep, fen_taddrh, taddrh);
204*4882a593Smuzhiyun W16(ep, fen_taddrm, taddrm);
205*4882a593Smuzhiyun W16(ep, fen_taddrl, taddrl);
206*4882a593Smuzhiyun fcc_cr_cmd(fep, CPM_CR_SET_GADDR);
207*4882a593Smuzhiyun }
208*4882a593Smuzhiyun
set_multicast_finish(struct net_device * dev)209*4882a593Smuzhiyun static void set_multicast_finish(struct net_device *dev)
210*4882a593Smuzhiyun {
211*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
212*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
213*4882a593Smuzhiyun fcc_enet_t __iomem *ep = fep->fcc.ep;
214*4882a593Smuzhiyun
215*4882a593Smuzhiyun /* clear promiscuous always */
216*4882a593Smuzhiyun C32(fccp, fcc_fpsmr, FCC_PSMR_PRO);
217*4882a593Smuzhiyun
218*4882a593Smuzhiyun /* if all multi or too many multicasts; just enable all */
219*4882a593Smuzhiyun if ((dev->flags & IFF_ALLMULTI) != 0 ||
220*4882a593Smuzhiyun netdev_mc_count(dev) > FCC_MAX_MULTICAST_ADDRS) {
221*4882a593Smuzhiyun
222*4882a593Smuzhiyun W32(ep, fen_gaddrh, 0xffffffff);
223*4882a593Smuzhiyun W32(ep, fen_gaddrl, 0xffffffff);
224*4882a593Smuzhiyun }
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun /* read back */
227*4882a593Smuzhiyun fep->fcc.gaddrh = R32(ep, fen_gaddrh);
228*4882a593Smuzhiyun fep->fcc.gaddrl = R32(ep, fen_gaddrl);
229*4882a593Smuzhiyun }
230*4882a593Smuzhiyun
set_multicast_list(struct net_device * dev)231*4882a593Smuzhiyun static void set_multicast_list(struct net_device *dev)
232*4882a593Smuzhiyun {
233*4882a593Smuzhiyun struct netdev_hw_addr *ha;
234*4882a593Smuzhiyun
235*4882a593Smuzhiyun if ((dev->flags & IFF_PROMISC) == 0) {
236*4882a593Smuzhiyun set_multicast_start(dev);
237*4882a593Smuzhiyun netdev_for_each_mc_addr(ha, dev)
238*4882a593Smuzhiyun set_multicast_one(dev, ha->addr);
239*4882a593Smuzhiyun set_multicast_finish(dev);
240*4882a593Smuzhiyun } else
241*4882a593Smuzhiyun set_promiscuous_mode(dev);
242*4882a593Smuzhiyun }
243*4882a593Smuzhiyun
restart(struct net_device * dev)244*4882a593Smuzhiyun static void restart(struct net_device *dev)
245*4882a593Smuzhiyun {
246*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
247*4882a593Smuzhiyun const struct fs_platform_info *fpi = fep->fpi;
248*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
249*4882a593Smuzhiyun fcc_c_t __iomem *fcccp = fep->fcc.fcccp;
250*4882a593Smuzhiyun fcc_enet_t __iomem *ep = fep->fcc.ep;
251*4882a593Smuzhiyun dma_addr_t rx_bd_base_phys, tx_bd_base_phys;
252*4882a593Smuzhiyun u16 paddrh, paddrm, paddrl;
253*4882a593Smuzhiyun const unsigned char *mac;
254*4882a593Smuzhiyun int i;
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun C32(fccp, fcc_gfmr, FCC_GFMR_ENR | FCC_GFMR_ENT);
257*4882a593Smuzhiyun
258*4882a593Smuzhiyun /* clear everything (slow & steady does it) */
259*4882a593Smuzhiyun for (i = 0; i < sizeof(*ep); i++)
260*4882a593Smuzhiyun out_8((u8 __iomem *)ep + i, 0);
261*4882a593Smuzhiyun
262*4882a593Smuzhiyun /* get physical address */
263*4882a593Smuzhiyun rx_bd_base_phys = fep->ring_mem_addr;
264*4882a593Smuzhiyun tx_bd_base_phys = rx_bd_base_phys + sizeof(cbd_t) * fpi->rx_ring;
265*4882a593Smuzhiyun
266*4882a593Smuzhiyun /* point to bds */
267*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_rbase, rx_bd_base_phys);
268*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_tbase, tx_bd_base_phys);
269*4882a593Smuzhiyun
270*4882a593Smuzhiyun /* Set maximum bytes per receive buffer.
271*4882a593Smuzhiyun * It must be a multiple of 32.
272*4882a593Smuzhiyun */
273*4882a593Smuzhiyun W16(ep, fen_genfcc.fcc_mrblr, PKT_MAXBLR_SIZE);
274*4882a593Smuzhiyun
275*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_rstate, (CPMFCR_GBL | CPMFCR_EB) << 24);
276*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_tstate, (CPMFCR_GBL | CPMFCR_EB) << 24);
277*4882a593Smuzhiyun
278*4882a593Smuzhiyun /* Allocate space in the reserved FCC area of DPRAM for the
279*4882a593Smuzhiyun * internal buffers. No one uses this space (yet), so we
280*4882a593Smuzhiyun * can do this. Later, we will add resource management for
281*4882a593Smuzhiyun * this area.
282*4882a593Smuzhiyun */
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun W16(ep, fen_genfcc.fcc_riptr, fpi->dpram_offset);
285*4882a593Smuzhiyun W16(ep, fen_genfcc.fcc_tiptr, fpi->dpram_offset + 32);
286*4882a593Smuzhiyun
287*4882a593Smuzhiyun W16(ep, fen_padptr, fpi->dpram_offset + 64);
288*4882a593Smuzhiyun
289*4882a593Smuzhiyun /* fill with special symbol... */
290*4882a593Smuzhiyun memset_io(fep->fcc.mem + fpi->dpram_offset + 64, 0x88, 32);
291*4882a593Smuzhiyun
292*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_rbptr, 0);
293*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_tbptr, 0);
294*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_rcrc, 0);
295*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_tcrc, 0);
296*4882a593Smuzhiyun W16(ep, fen_genfcc.fcc_res1, 0);
297*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_res2, 0);
298*4882a593Smuzhiyun
299*4882a593Smuzhiyun /* no CAM */
300*4882a593Smuzhiyun W32(ep, fen_camptr, 0);
301*4882a593Smuzhiyun
302*4882a593Smuzhiyun /* Set CRC preset and mask */
303*4882a593Smuzhiyun W32(ep, fen_cmask, 0xdebb20e3);
304*4882a593Smuzhiyun W32(ep, fen_cpres, 0xffffffff);
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun W32(ep, fen_crcec, 0); /* CRC Error counter */
307*4882a593Smuzhiyun W32(ep, fen_alec, 0); /* alignment error counter */
308*4882a593Smuzhiyun W32(ep, fen_disfc, 0); /* discard frame counter */
309*4882a593Smuzhiyun W16(ep, fen_retlim, 15); /* Retry limit threshold */
310*4882a593Smuzhiyun W16(ep, fen_pper, 0); /* Normal persistence */
311*4882a593Smuzhiyun
312*4882a593Smuzhiyun /* set group address */
313*4882a593Smuzhiyun W32(ep, fen_gaddrh, fep->fcc.gaddrh);
314*4882a593Smuzhiyun W32(ep, fen_gaddrl, fep->fcc.gaddrh);
315*4882a593Smuzhiyun
316*4882a593Smuzhiyun /* Clear hash filter tables */
317*4882a593Smuzhiyun W32(ep, fen_iaddrh, 0);
318*4882a593Smuzhiyun W32(ep, fen_iaddrl, 0);
319*4882a593Smuzhiyun
320*4882a593Smuzhiyun /* Clear the Out-of-sequence TxBD */
321*4882a593Smuzhiyun W16(ep, fen_tfcstat, 0);
322*4882a593Smuzhiyun W16(ep, fen_tfclen, 0);
323*4882a593Smuzhiyun W32(ep, fen_tfcptr, 0);
324*4882a593Smuzhiyun
325*4882a593Smuzhiyun W16(ep, fen_mflr, PKT_MAXBUF_SIZE); /* maximum frame length register */
326*4882a593Smuzhiyun W16(ep, fen_minflr, PKT_MINBUF_SIZE); /* minimum frame length register */
327*4882a593Smuzhiyun
328*4882a593Smuzhiyun /* set address */
329*4882a593Smuzhiyun mac = dev->dev_addr;
330*4882a593Smuzhiyun paddrh = ((u16)mac[5] << 8) | mac[4];
331*4882a593Smuzhiyun paddrm = ((u16)mac[3] << 8) | mac[2];
332*4882a593Smuzhiyun paddrl = ((u16)mac[1] << 8) | mac[0];
333*4882a593Smuzhiyun
334*4882a593Smuzhiyun W16(ep, fen_paddrh, paddrh);
335*4882a593Smuzhiyun W16(ep, fen_paddrm, paddrm);
336*4882a593Smuzhiyun W16(ep, fen_paddrl, paddrl);
337*4882a593Smuzhiyun
338*4882a593Smuzhiyun W16(ep, fen_taddrh, 0);
339*4882a593Smuzhiyun W16(ep, fen_taddrm, 0);
340*4882a593Smuzhiyun W16(ep, fen_taddrl, 0);
341*4882a593Smuzhiyun
342*4882a593Smuzhiyun W16(ep, fen_maxd1, 1520); /* maximum DMA1 length */
343*4882a593Smuzhiyun W16(ep, fen_maxd2, 1520); /* maximum DMA2 length */
344*4882a593Smuzhiyun
345*4882a593Smuzhiyun /* Clear stat counters, in case we ever enable RMON */
346*4882a593Smuzhiyun W32(ep, fen_octc, 0);
347*4882a593Smuzhiyun W32(ep, fen_colc, 0);
348*4882a593Smuzhiyun W32(ep, fen_broc, 0);
349*4882a593Smuzhiyun W32(ep, fen_mulc, 0);
350*4882a593Smuzhiyun W32(ep, fen_uspc, 0);
351*4882a593Smuzhiyun W32(ep, fen_frgc, 0);
352*4882a593Smuzhiyun W32(ep, fen_ospc, 0);
353*4882a593Smuzhiyun W32(ep, fen_jbrc, 0);
354*4882a593Smuzhiyun W32(ep, fen_p64c, 0);
355*4882a593Smuzhiyun W32(ep, fen_p65c, 0);
356*4882a593Smuzhiyun W32(ep, fen_p128c, 0);
357*4882a593Smuzhiyun W32(ep, fen_p256c, 0);
358*4882a593Smuzhiyun W32(ep, fen_p512c, 0);
359*4882a593Smuzhiyun W32(ep, fen_p1024c, 0);
360*4882a593Smuzhiyun
361*4882a593Smuzhiyun W16(ep, fen_rfthr, 0); /* Suggested by manual */
362*4882a593Smuzhiyun W16(ep, fen_rfcnt, 0);
363*4882a593Smuzhiyun W16(ep, fen_cftype, 0);
364*4882a593Smuzhiyun
365*4882a593Smuzhiyun fs_init_bds(dev);
366*4882a593Smuzhiyun
367*4882a593Smuzhiyun /* adjust to speed (for RMII mode) */
368*4882a593Smuzhiyun if (fpi->use_rmii) {
369*4882a593Smuzhiyun if (dev->phydev->speed == 100)
370*4882a593Smuzhiyun C8(fcccp, fcc_gfemr, 0x20);
371*4882a593Smuzhiyun else
372*4882a593Smuzhiyun S8(fcccp, fcc_gfemr, 0x20);
373*4882a593Smuzhiyun }
374*4882a593Smuzhiyun
375*4882a593Smuzhiyun fcc_cr_cmd(fep, CPM_CR_INIT_TRX);
376*4882a593Smuzhiyun
377*4882a593Smuzhiyun /* clear events */
378*4882a593Smuzhiyun W16(fccp, fcc_fcce, 0xffff);
379*4882a593Smuzhiyun
380*4882a593Smuzhiyun /* Enable interrupts we wish to service */
381*4882a593Smuzhiyun W16(fccp, fcc_fccm, FCC_ENET_TXE | FCC_ENET_RXF | FCC_ENET_TXB);
382*4882a593Smuzhiyun
383*4882a593Smuzhiyun /* Set GFMR to enable Ethernet operating mode */
384*4882a593Smuzhiyun W32(fccp, fcc_gfmr, FCC_GFMR_TCI | FCC_GFMR_MODE_ENET);
385*4882a593Smuzhiyun
386*4882a593Smuzhiyun /* set sync/delimiters */
387*4882a593Smuzhiyun W16(fccp, fcc_fdsr, 0xd555);
388*4882a593Smuzhiyun
389*4882a593Smuzhiyun W32(fccp, fcc_fpsmr, FCC_PSMR_ENCRC);
390*4882a593Smuzhiyun
391*4882a593Smuzhiyun if (fpi->use_rmii)
392*4882a593Smuzhiyun S32(fccp, fcc_fpsmr, FCC_PSMR_RMII);
393*4882a593Smuzhiyun
394*4882a593Smuzhiyun /* adjust to duplex mode */
395*4882a593Smuzhiyun if (dev->phydev->duplex)
396*4882a593Smuzhiyun S32(fccp, fcc_fpsmr, FCC_PSMR_FDE | FCC_PSMR_LPB);
397*4882a593Smuzhiyun else
398*4882a593Smuzhiyun C32(fccp, fcc_fpsmr, FCC_PSMR_FDE | FCC_PSMR_LPB);
399*4882a593Smuzhiyun
400*4882a593Smuzhiyun /* Restore multicast and promiscuous settings */
401*4882a593Smuzhiyun set_multicast_list(dev);
402*4882a593Smuzhiyun
403*4882a593Smuzhiyun S32(fccp, fcc_gfmr, FCC_GFMR_ENR | FCC_GFMR_ENT);
404*4882a593Smuzhiyun }
405*4882a593Smuzhiyun
stop(struct net_device * dev)406*4882a593Smuzhiyun static void stop(struct net_device *dev)
407*4882a593Smuzhiyun {
408*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
409*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
410*4882a593Smuzhiyun
411*4882a593Smuzhiyun /* stop ethernet */
412*4882a593Smuzhiyun C32(fccp, fcc_gfmr, FCC_GFMR_ENR | FCC_GFMR_ENT);
413*4882a593Smuzhiyun
414*4882a593Smuzhiyun /* clear events */
415*4882a593Smuzhiyun W16(fccp, fcc_fcce, 0xffff);
416*4882a593Smuzhiyun
417*4882a593Smuzhiyun /* clear interrupt mask */
418*4882a593Smuzhiyun W16(fccp, fcc_fccm, 0);
419*4882a593Smuzhiyun
420*4882a593Smuzhiyun fs_cleanup_bds(dev);
421*4882a593Smuzhiyun }
422*4882a593Smuzhiyun
napi_clear_event_fs(struct net_device * dev)423*4882a593Smuzhiyun static void napi_clear_event_fs(struct net_device *dev)
424*4882a593Smuzhiyun {
425*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
426*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
427*4882a593Smuzhiyun
428*4882a593Smuzhiyun W16(fccp, fcc_fcce, FCC_NAPI_EVENT_MSK);
429*4882a593Smuzhiyun }
430*4882a593Smuzhiyun
napi_enable_fs(struct net_device * dev)431*4882a593Smuzhiyun static void napi_enable_fs(struct net_device *dev)
432*4882a593Smuzhiyun {
433*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
434*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
435*4882a593Smuzhiyun
436*4882a593Smuzhiyun S16(fccp, fcc_fccm, FCC_NAPI_EVENT_MSK);
437*4882a593Smuzhiyun }
438*4882a593Smuzhiyun
napi_disable_fs(struct net_device * dev)439*4882a593Smuzhiyun static void napi_disable_fs(struct net_device *dev)
440*4882a593Smuzhiyun {
441*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
442*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
443*4882a593Smuzhiyun
444*4882a593Smuzhiyun C16(fccp, fcc_fccm, FCC_NAPI_EVENT_MSK);
445*4882a593Smuzhiyun }
446*4882a593Smuzhiyun
rx_bd_done(struct net_device * dev)447*4882a593Smuzhiyun static void rx_bd_done(struct net_device *dev)
448*4882a593Smuzhiyun {
449*4882a593Smuzhiyun /* nothing */
450*4882a593Smuzhiyun }
451*4882a593Smuzhiyun
tx_kickstart(struct net_device * dev)452*4882a593Smuzhiyun static void tx_kickstart(struct net_device *dev)
453*4882a593Smuzhiyun {
454*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
455*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
456*4882a593Smuzhiyun
457*4882a593Smuzhiyun S16(fccp, fcc_ftodr, 0x8000);
458*4882a593Smuzhiyun }
459*4882a593Smuzhiyun
get_int_events(struct net_device * dev)460*4882a593Smuzhiyun static u32 get_int_events(struct net_device *dev)
461*4882a593Smuzhiyun {
462*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
463*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
464*4882a593Smuzhiyun
465*4882a593Smuzhiyun return (u32)R16(fccp, fcc_fcce);
466*4882a593Smuzhiyun }
467*4882a593Smuzhiyun
clear_int_events(struct net_device * dev,u32 int_events)468*4882a593Smuzhiyun static void clear_int_events(struct net_device *dev, u32 int_events)
469*4882a593Smuzhiyun {
470*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
471*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
472*4882a593Smuzhiyun
473*4882a593Smuzhiyun W16(fccp, fcc_fcce, int_events & 0xffff);
474*4882a593Smuzhiyun }
475*4882a593Smuzhiyun
ev_error(struct net_device * dev,u32 int_events)476*4882a593Smuzhiyun static void ev_error(struct net_device *dev, u32 int_events)
477*4882a593Smuzhiyun {
478*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
479*4882a593Smuzhiyun
480*4882a593Smuzhiyun dev_warn(fep->dev, "FS_ENET ERROR(s) 0x%x\n", int_events);
481*4882a593Smuzhiyun }
482*4882a593Smuzhiyun
get_regs(struct net_device * dev,void * p,int * sizep)483*4882a593Smuzhiyun static int get_regs(struct net_device *dev, void *p, int *sizep)
484*4882a593Smuzhiyun {
485*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
486*4882a593Smuzhiyun
487*4882a593Smuzhiyun if (*sizep < sizeof(fcc_t) + sizeof(fcc_enet_t) + 1)
488*4882a593Smuzhiyun return -EINVAL;
489*4882a593Smuzhiyun
490*4882a593Smuzhiyun memcpy_fromio(p, fep->fcc.fccp, sizeof(fcc_t));
491*4882a593Smuzhiyun p = (char *)p + sizeof(fcc_t);
492*4882a593Smuzhiyun
493*4882a593Smuzhiyun memcpy_fromio(p, fep->fcc.ep, sizeof(fcc_enet_t));
494*4882a593Smuzhiyun p = (char *)p + sizeof(fcc_enet_t);
495*4882a593Smuzhiyun
496*4882a593Smuzhiyun memcpy_fromio(p, fep->fcc.fcccp, 1);
497*4882a593Smuzhiyun return 0;
498*4882a593Smuzhiyun }
499*4882a593Smuzhiyun
get_regs_len(struct net_device * dev)500*4882a593Smuzhiyun static int get_regs_len(struct net_device *dev)
501*4882a593Smuzhiyun {
502*4882a593Smuzhiyun return sizeof(fcc_t) + sizeof(fcc_enet_t) + 1;
503*4882a593Smuzhiyun }
504*4882a593Smuzhiyun
505*4882a593Smuzhiyun /* Some transmit errors cause the transmitter to shut
506*4882a593Smuzhiyun * down. We now issue a restart transmit.
507*4882a593Smuzhiyun * Also, to workaround 8260 device erratum CPM37, we must
508*4882a593Smuzhiyun * disable and then re-enable the transmitterfollowing a
509*4882a593Smuzhiyun * Late Collision, Underrun, or Retry Limit error.
510*4882a593Smuzhiyun * In addition, tbptr may point beyond BDs beyond still marked
511*4882a593Smuzhiyun * as ready due to internal pipelining, so we need to look back
512*4882a593Smuzhiyun * through the BDs and adjust tbptr to point to the last BD
513*4882a593Smuzhiyun * marked as ready. This may result in some buffers being
514*4882a593Smuzhiyun * retransmitted.
515*4882a593Smuzhiyun */
tx_restart(struct net_device * dev)516*4882a593Smuzhiyun static void tx_restart(struct net_device *dev)
517*4882a593Smuzhiyun {
518*4882a593Smuzhiyun struct fs_enet_private *fep = netdev_priv(dev);
519*4882a593Smuzhiyun fcc_t __iomem *fccp = fep->fcc.fccp;
520*4882a593Smuzhiyun const struct fs_platform_info *fpi = fep->fpi;
521*4882a593Smuzhiyun fcc_enet_t __iomem *ep = fep->fcc.ep;
522*4882a593Smuzhiyun cbd_t __iomem *curr_tbptr;
523*4882a593Smuzhiyun cbd_t __iomem *recheck_bd;
524*4882a593Smuzhiyun cbd_t __iomem *prev_bd;
525*4882a593Smuzhiyun cbd_t __iomem *last_tx_bd;
526*4882a593Smuzhiyun
527*4882a593Smuzhiyun last_tx_bd = fep->tx_bd_base + (fpi->tx_ring - 1);
528*4882a593Smuzhiyun
529*4882a593Smuzhiyun /* get the current bd held in TBPTR and scan back from this point */
530*4882a593Smuzhiyun recheck_bd = curr_tbptr = (cbd_t __iomem *)
531*4882a593Smuzhiyun ((R32(ep, fen_genfcc.fcc_tbptr) - fep->ring_mem_addr) +
532*4882a593Smuzhiyun fep->ring_base);
533*4882a593Smuzhiyun
534*4882a593Smuzhiyun prev_bd = (recheck_bd == fep->tx_bd_base) ? last_tx_bd : recheck_bd - 1;
535*4882a593Smuzhiyun
536*4882a593Smuzhiyun /* Move through the bds in reverse, look for the earliest buffer
537*4882a593Smuzhiyun * that is not ready. Adjust TBPTR to the following buffer */
538*4882a593Smuzhiyun while ((CBDR_SC(prev_bd) & BD_ENET_TX_READY) != 0) {
539*4882a593Smuzhiyun /* Go back one buffer */
540*4882a593Smuzhiyun recheck_bd = prev_bd;
541*4882a593Smuzhiyun
542*4882a593Smuzhiyun /* update the previous buffer */
543*4882a593Smuzhiyun prev_bd = (prev_bd == fep->tx_bd_base) ? last_tx_bd : prev_bd - 1;
544*4882a593Smuzhiyun
545*4882a593Smuzhiyun /* We should never see all bds marked as ready, check anyway */
546*4882a593Smuzhiyun if (recheck_bd == curr_tbptr)
547*4882a593Smuzhiyun break;
548*4882a593Smuzhiyun }
549*4882a593Smuzhiyun /* Now update the TBPTR and dirty flag to the current buffer */
550*4882a593Smuzhiyun W32(ep, fen_genfcc.fcc_tbptr,
551*4882a593Smuzhiyun (uint) (((void *)recheck_bd - fep->ring_base) +
552*4882a593Smuzhiyun fep->ring_mem_addr));
553*4882a593Smuzhiyun fep->dirty_tx = recheck_bd;
554*4882a593Smuzhiyun
555*4882a593Smuzhiyun C32(fccp, fcc_gfmr, FCC_GFMR_ENT);
556*4882a593Smuzhiyun udelay(10);
557*4882a593Smuzhiyun S32(fccp, fcc_gfmr, FCC_GFMR_ENT);
558*4882a593Smuzhiyun
559*4882a593Smuzhiyun fcc_cr_cmd(fep, CPM_CR_RESTART_TX);
560*4882a593Smuzhiyun }
561*4882a593Smuzhiyun
562*4882a593Smuzhiyun /*************************************************************************/
563*4882a593Smuzhiyun
564*4882a593Smuzhiyun const struct fs_ops fs_fcc_ops = {
565*4882a593Smuzhiyun .setup_data = setup_data,
566*4882a593Smuzhiyun .cleanup_data = cleanup_data,
567*4882a593Smuzhiyun .set_multicast_list = set_multicast_list,
568*4882a593Smuzhiyun .restart = restart,
569*4882a593Smuzhiyun .stop = stop,
570*4882a593Smuzhiyun .napi_clear_event = napi_clear_event_fs,
571*4882a593Smuzhiyun .napi_enable = napi_enable_fs,
572*4882a593Smuzhiyun .napi_disable = napi_disable_fs,
573*4882a593Smuzhiyun .rx_bd_done = rx_bd_done,
574*4882a593Smuzhiyun .tx_kickstart = tx_kickstart,
575*4882a593Smuzhiyun .get_int_events = get_int_events,
576*4882a593Smuzhiyun .clear_int_events = clear_int_events,
577*4882a593Smuzhiyun .ev_error = ev_error,
578*4882a593Smuzhiyun .get_regs = get_regs,
579*4882a593Smuzhiyun .get_regs_len = get_regs_len,
580*4882a593Smuzhiyun .tx_restart = tx_restart,
581*4882a593Smuzhiyun .allocate_bd = allocate_bd,
582*4882a593Smuzhiyun .free_bd = free_bd,
583*4882a593Smuzhiyun };
584