xref: /OK3568_Linux_fs/kernel/drivers/net/wan/wanxl.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * wanXL serial card driver for Linux
4*4882a593Smuzhiyun  * host part
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
7*4882a593Smuzhiyun  *
8*4882a593Smuzhiyun  * Status:
9*4882a593Smuzhiyun  *   - Only DTE (external clock) support with NRZ and NRZI encodings
10*4882a593Smuzhiyun  *   - wanXL100 will require minor driver modifications, no access to hw
11*4882a593Smuzhiyun  */
12*4882a593Smuzhiyun 
13*4882a593Smuzhiyun #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14*4882a593Smuzhiyun 
15*4882a593Smuzhiyun #include <linux/module.h>
16*4882a593Smuzhiyun #include <linux/kernel.h>
17*4882a593Smuzhiyun #include <linux/slab.h>
18*4882a593Smuzhiyun #include <linux/sched.h>
19*4882a593Smuzhiyun #include <linux/types.h>
20*4882a593Smuzhiyun #include <linux/fcntl.h>
21*4882a593Smuzhiyun #include <linux/string.h>
22*4882a593Smuzhiyun #include <linux/errno.h>
23*4882a593Smuzhiyun #include <linux/init.h>
24*4882a593Smuzhiyun #include <linux/interrupt.h>
25*4882a593Smuzhiyun #include <linux/ioport.h>
26*4882a593Smuzhiyun #include <linux/netdevice.h>
27*4882a593Smuzhiyun #include <linux/hdlc.h>
28*4882a593Smuzhiyun #include <linux/pci.h>
29*4882a593Smuzhiyun #include <linux/dma-mapping.h>
30*4882a593Smuzhiyun #include <linux/delay.h>
31*4882a593Smuzhiyun #include <asm/io.h>
32*4882a593Smuzhiyun 
33*4882a593Smuzhiyun #include "wanxl.h"
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun static const char* version = "wanXL serial card driver version: 0.48";
36*4882a593Smuzhiyun 
37*4882a593Smuzhiyun #define PLX_CTL_RESET   0x40000000 /* adapter reset */
38*4882a593Smuzhiyun 
39*4882a593Smuzhiyun #undef DEBUG_PKT
40*4882a593Smuzhiyun #undef DEBUG_PCI
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun /* MAILBOX #1 - PUTS COMMANDS */
43*4882a593Smuzhiyun #define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
44*4882a593Smuzhiyun #ifdef __LITTLE_ENDIAN
45*4882a593Smuzhiyun #define MBX1_CMD_BSWAP  0x8C000001 /* little-endian Byte Swap Mode */
46*4882a593Smuzhiyun #else
47*4882a593Smuzhiyun #define MBX1_CMD_BSWAP  0x8C000000 /* big-endian Byte Swap Mode */
48*4882a593Smuzhiyun #endif
49*4882a593Smuzhiyun 
50*4882a593Smuzhiyun /* MAILBOX #2 - DRAM SIZE */
51*4882a593Smuzhiyun #define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun 
54*4882a593Smuzhiyun struct port {
55*4882a593Smuzhiyun 	struct net_device *dev;
56*4882a593Smuzhiyun 	struct card *card;
57*4882a593Smuzhiyun 	spinlock_t lock;	/* for wanxl_xmit */
58*4882a593Smuzhiyun         int node;		/* physical port #0 - 3 */
59*4882a593Smuzhiyun 	unsigned int clock_type;
60*4882a593Smuzhiyun 	int tx_in, tx_out;
61*4882a593Smuzhiyun 	struct sk_buff *tx_skbs[TX_BUFFERS];
62*4882a593Smuzhiyun };
63*4882a593Smuzhiyun 
64*4882a593Smuzhiyun 
65*4882a593Smuzhiyun struct card_status {
66*4882a593Smuzhiyun 	desc_t rx_descs[RX_QUEUE_LENGTH];
67*4882a593Smuzhiyun 	port_status_t port_status[4];
68*4882a593Smuzhiyun };
69*4882a593Smuzhiyun 
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun struct card {
72*4882a593Smuzhiyun 	int n_ports;		/* 1, 2 or 4 ports */
73*4882a593Smuzhiyun 	u8 irq;
74*4882a593Smuzhiyun 
75*4882a593Smuzhiyun 	u8 __iomem *plx;	/* PLX PCI9060 virtual base address */
76*4882a593Smuzhiyun 	struct pci_dev *pdev;	/* for pci_name(pdev) */
77*4882a593Smuzhiyun 	int rx_in;
78*4882a593Smuzhiyun 	struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
79*4882a593Smuzhiyun 	struct card_status *status;	/* shared between host and card */
80*4882a593Smuzhiyun 	dma_addr_t status_address;
81*4882a593Smuzhiyun 	struct port ports[];	/* 1 - 4 port structures follow */
82*4882a593Smuzhiyun };
83*4882a593Smuzhiyun 
84*4882a593Smuzhiyun 
85*4882a593Smuzhiyun 
dev_to_port(struct net_device * dev)86*4882a593Smuzhiyun static inline struct port *dev_to_port(struct net_device *dev)
87*4882a593Smuzhiyun {
88*4882a593Smuzhiyun 	return (struct port *)dev_to_hdlc(dev)->priv;
89*4882a593Smuzhiyun }
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 
get_status(struct port * port)92*4882a593Smuzhiyun static inline port_status_t *get_status(struct port *port)
93*4882a593Smuzhiyun {
94*4882a593Smuzhiyun 	return &port->card->status->port_status[port->node];
95*4882a593Smuzhiyun }
96*4882a593Smuzhiyun 
97*4882a593Smuzhiyun 
98*4882a593Smuzhiyun #ifdef DEBUG_PCI
pci_map_single_debug(struct pci_dev * pdev,void * ptr,size_t size,int direction)99*4882a593Smuzhiyun static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
100*4882a593Smuzhiyun 					      size_t size, int direction)
101*4882a593Smuzhiyun {
102*4882a593Smuzhiyun 	dma_addr_t addr = dma_map_single(&pdev->dev, ptr, size, direction);
103*4882a593Smuzhiyun 	if (addr + size > 0x100000000LL)
104*4882a593Smuzhiyun 		pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
105*4882a593Smuzhiyun 			pci_name(pdev), (unsigned long long)addr);
106*4882a593Smuzhiyun 	return addr;
107*4882a593Smuzhiyun }
108*4882a593Smuzhiyun 
109*4882a593Smuzhiyun #undef pci_map_single
110*4882a593Smuzhiyun #define pci_map_single pci_map_single_debug
111*4882a593Smuzhiyun #endif
112*4882a593Smuzhiyun 
113*4882a593Smuzhiyun 
114*4882a593Smuzhiyun /* Cable and/or personality module change interrupt service */
wanxl_cable_intr(struct port * port)115*4882a593Smuzhiyun static inline void wanxl_cable_intr(struct port *port)
116*4882a593Smuzhiyun {
117*4882a593Smuzhiyun 	u32 value = get_status(port)->cable;
118*4882a593Smuzhiyun 	int valid = 1;
119*4882a593Smuzhiyun 	const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
120*4882a593Smuzhiyun 
121*4882a593Smuzhiyun 	switch(value & 0x7) {
122*4882a593Smuzhiyun 	case STATUS_CABLE_V35: cable = "V.35"; break;
123*4882a593Smuzhiyun 	case STATUS_CABLE_X21: cable = "X.21"; break;
124*4882a593Smuzhiyun 	case STATUS_CABLE_V24: cable = "V.24"; break;
125*4882a593Smuzhiyun 	case STATUS_CABLE_EIA530: cable = "EIA530"; break;
126*4882a593Smuzhiyun 	case STATUS_CABLE_NONE: cable = "no"; break;
127*4882a593Smuzhiyun 	default: cable = "invalid";
128*4882a593Smuzhiyun 	}
129*4882a593Smuzhiyun 
130*4882a593Smuzhiyun 	switch((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
131*4882a593Smuzhiyun 	case STATUS_CABLE_V35: pm = "V.35"; break;
132*4882a593Smuzhiyun 	case STATUS_CABLE_X21: pm = "X.21"; break;
133*4882a593Smuzhiyun 	case STATUS_CABLE_V24: pm = "V.24"; break;
134*4882a593Smuzhiyun 	case STATUS_CABLE_EIA530: pm = "EIA530"; break;
135*4882a593Smuzhiyun 	case STATUS_CABLE_NONE: pm = "no personality"; valid = 0; break;
136*4882a593Smuzhiyun 	default: pm = "invalid personality"; valid = 0;
137*4882a593Smuzhiyun 	}
138*4882a593Smuzhiyun 
139*4882a593Smuzhiyun 	if (valid) {
140*4882a593Smuzhiyun 		if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
141*4882a593Smuzhiyun 			dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
142*4882a593Smuzhiyun 				", DSR off";
143*4882a593Smuzhiyun 			dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
144*4882a593Smuzhiyun 				", carrier off";
145*4882a593Smuzhiyun 		}
146*4882a593Smuzhiyun 		dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
147*4882a593Smuzhiyun 	}
148*4882a593Smuzhiyun 	netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
149*4882a593Smuzhiyun 		    pm, dte, cable, dsr, dcd);
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun 	if (value & STATUS_CABLE_DCD)
152*4882a593Smuzhiyun 		netif_carrier_on(port->dev);
153*4882a593Smuzhiyun 	else
154*4882a593Smuzhiyun 		netif_carrier_off(port->dev);
155*4882a593Smuzhiyun }
156*4882a593Smuzhiyun 
157*4882a593Smuzhiyun 
158*4882a593Smuzhiyun 
159*4882a593Smuzhiyun /* Transmit complete interrupt service */
wanxl_tx_intr(struct port * port)160*4882a593Smuzhiyun static inline void wanxl_tx_intr(struct port *port)
161*4882a593Smuzhiyun {
162*4882a593Smuzhiyun 	struct net_device *dev = port->dev;
163*4882a593Smuzhiyun 	while (1) {
164*4882a593Smuzhiyun                 desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
165*4882a593Smuzhiyun 		struct sk_buff *skb = port->tx_skbs[port->tx_in];
166*4882a593Smuzhiyun 
167*4882a593Smuzhiyun 		switch (desc->stat) {
168*4882a593Smuzhiyun 		case PACKET_FULL:
169*4882a593Smuzhiyun 		case PACKET_EMPTY:
170*4882a593Smuzhiyun 			netif_wake_queue(dev);
171*4882a593Smuzhiyun 			return;
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun 		case PACKET_UNDERRUN:
174*4882a593Smuzhiyun 			dev->stats.tx_errors++;
175*4882a593Smuzhiyun 			dev->stats.tx_fifo_errors++;
176*4882a593Smuzhiyun 			break;
177*4882a593Smuzhiyun 
178*4882a593Smuzhiyun 		default:
179*4882a593Smuzhiyun 			dev->stats.tx_packets++;
180*4882a593Smuzhiyun 			dev->stats.tx_bytes += skb->len;
181*4882a593Smuzhiyun 		}
182*4882a593Smuzhiyun                 desc->stat = PACKET_EMPTY; /* Free descriptor */
183*4882a593Smuzhiyun 		dma_unmap_single(&port->card->pdev->dev, desc->address,
184*4882a593Smuzhiyun 				 skb->len, DMA_TO_DEVICE);
185*4882a593Smuzhiyun 		dev_consume_skb_irq(skb);
186*4882a593Smuzhiyun                 port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
187*4882a593Smuzhiyun         }
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun 
190*4882a593Smuzhiyun 
191*4882a593Smuzhiyun 
192*4882a593Smuzhiyun /* Receive complete interrupt service */
wanxl_rx_intr(struct card * card)193*4882a593Smuzhiyun static inline void wanxl_rx_intr(struct card *card)
194*4882a593Smuzhiyun {
195*4882a593Smuzhiyun 	desc_t *desc;
196*4882a593Smuzhiyun 	while (desc = &card->status->rx_descs[card->rx_in],
197*4882a593Smuzhiyun 	       desc->stat != PACKET_EMPTY) {
198*4882a593Smuzhiyun 		if ((desc->stat & PACKET_PORT_MASK) > card->n_ports)
199*4882a593Smuzhiyun 			pr_crit("%s: received packet for nonexistent port\n",
200*4882a593Smuzhiyun 				pci_name(card->pdev));
201*4882a593Smuzhiyun 		else {
202*4882a593Smuzhiyun 			struct sk_buff *skb = card->rx_skbs[card->rx_in];
203*4882a593Smuzhiyun 			struct port *port = &card->ports[desc->stat &
204*4882a593Smuzhiyun 						    PACKET_PORT_MASK];
205*4882a593Smuzhiyun 			struct net_device *dev = port->dev;
206*4882a593Smuzhiyun 
207*4882a593Smuzhiyun 			if (!skb)
208*4882a593Smuzhiyun 				dev->stats.rx_dropped++;
209*4882a593Smuzhiyun 			else {
210*4882a593Smuzhiyun 				dma_unmap_single(&card->pdev->dev,
211*4882a593Smuzhiyun 						 desc->address, BUFFER_LENGTH,
212*4882a593Smuzhiyun 						 DMA_FROM_DEVICE);
213*4882a593Smuzhiyun 				skb_put(skb, desc->length);
214*4882a593Smuzhiyun 
215*4882a593Smuzhiyun #ifdef DEBUG_PKT
216*4882a593Smuzhiyun 				printk(KERN_DEBUG "%s RX(%i):", dev->name,
217*4882a593Smuzhiyun 				       skb->len);
218*4882a593Smuzhiyun 				debug_frame(skb);
219*4882a593Smuzhiyun #endif
220*4882a593Smuzhiyun 				dev->stats.rx_packets++;
221*4882a593Smuzhiyun 				dev->stats.rx_bytes += skb->len;
222*4882a593Smuzhiyun 				skb->protocol = hdlc_type_trans(skb, dev);
223*4882a593Smuzhiyun 				netif_rx(skb);
224*4882a593Smuzhiyun 				skb = NULL;
225*4882a593Smuzhiyun 			}
226*4882a593Smuzhiyun 
227*4882a593Smuzhiyun 			if (!skb) {
228*4882a593Smuzhiyun 				skb = dev_alloc_skb(BUFFER_LENGTH);
229*4882a593Smuzhiyun 				desc->address = skb ?
230*4882a593Smuzhiyun 					dma_map_single(&card->pdev->dev,
231*4882a593Smuzhiyun 						       skb->data,
232*4882a593Smuzhiyun 						       BUFFER_LENGTH,
233*4882a593Smuzhiyun 						       DMA_FROM_DEVICE) : 0;
234*4882a593Smuzhiyun 				card->rx_skbs[card->rx_in] = skb;
235*4882a593Smuzhiyun 			}
236*4882a593Smuzhiyun 		}
237*4882a593Smuzhiyun 		desc->stat = PACKET_EMPTY; /* Free descriptor */
238*4882a593Smuzhiyun 		card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
239*4882a593Smuzhiyun 	}
240*4882a593Smuzhiyun }
241*4882a593Smuzhiyun 
242*4882a593Smuzhiyun 
243*4882a593Smuzhiyun 
wanxl_intr(int irq,void * dev_id)244*4882a593Smuzhiyun static irqreturn_t wanxl_intr(int irq, void* dev_id)
245*4882a593Smuzhiyun {
246*4882a593Smuzhiyun 	struct card *card = dev_id;
247*4882a593Smuzhiyun         int i;
248*4882a593Smuzhiyun         u32 stat;
249*4882a593Smuzhiyun         int handled = 0;
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun 
252*4882a593Smuzhiyun         while((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
253*4882a593Smuzhiyun                 handled = 1;
254*4882a593Smuzhiyun 		writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun                 for (i = 0; i < card->n_ports; i++) {
257*4882a593Smuzhiyun 			if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
258*4882a593Smuzhiyun 				wanxl_tx_intr(&card->ports[i]);
259*4882a593Smuzhiyun 			if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
260*4882a593Smuzhiyun 				wanxl_cable_intr(&card->ports[i]);
261*4882a593Smuzhiyun 		}
262*4882a593Smuzhiyun 		if (stat & (1 << DOORBELL_FROM_CARD_RX))
263*4882a593Smuzhiyun 			wanxl_rx_intr(card);
264*4882a593Smuzhiyun         }
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun         return IRQ_RETVAL(handled);
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun 
269*4882a593Smuzhiyun 
270*4882a593Smuzhiyun 
wanxl_xmit(struct sk_buff * skb,struct net_device * dev)271*4882a593Smuzhiyun static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
272*4882a593Smuzhiyun {
273*4882a593Smuzhiyun 	struct port *port = dev_to_port(dev);
274*4882a593Smuzhiyun 	desc_t *desc;
275*4882a593Smuzhiyun 
276*4882a593Smuzhiyun         spin_lock(&port->lock);
277*4882a593Smuzhiyun 
278*4882a593Smuzhiyun 	desc = &get_status(port)->tx_descs[port->tx_out];
279*4882a593Smuzhiyun         if (desc->stat != PACKET_EMPTY) {
280*4882a593Smuzhiyun                 /* should never happen - previous xmit should stop queue */
281*4882a593Smuzhiyun #ifdef DEBUG_PKT
282*4882a593Smuzhiyun                 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
283*4882a593Smuzhiyun #endif
284*4882a593Smuzhiyun 		netif_stop_queue(dev);
285*4882a593Smuzhiyun 		spin_unlock(&port->lock);
286*4882a593Smuzhiyun 		return NETDEV_TX_BUSY;       /* request packet to be queued */
287*4882a593Smuzhiyun 	}
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun #ifdef DEBUG_PKT
290*4882a593Smuzhiyun 	printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
291*4882a593Smuzhiyun 	debug_frame(skb);
292*4882a593Smuzhiyun #endif
293*4882a593Smuzhiyun 
294*4882a593Smuzhiyun 	port->tx_skbs[port->tx_out] = skb;
295*4882a593Smuzhiyun 	desc->address = dma_map_single(&port->card->pdev->dev, skb->data,
296*4882a593Smuzhiyun 				       skb->len, DMA_TO_DEVICE);
297*4882a593Smuzhiyun 	desc->length = skb->len;
298*4882a593Smuzhiyun 	desc->stat = PACKET_FULL;
299*4882a593Smuzhiyun 	writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
300*4882a593Smuzhiyun 	       port->card->plx + PLX_DOORBELL_TO_CARD);
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun 	port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun 	if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
305*4882a593Smuzhiyun 		netif_stop_queue(dev);
306*4882a593Smuzhiyun #ifdef DEBUG_PKT
307*4882a593Smuzhiyun 		printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
308*4882a593Smuzhiyun #endif
309*4882a593Smuzhiyun 	}
310*4882a593Smuzhiyun 
311*4882a593Smuzhiyun 	spin_unlock(&port->lock);
312*4882a593Smuzhiyun 	return NETDEV_TX_OK;
313*4882a593Smuzhiyun }
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun 
316*4882a593Smuzhiyun 
wanxl_attach(struct net_device * dev,unsigned short encoding,unsigned short parity)317*4882a593Smuzhiyun static int wanxl_attach(struct net_device *dev, unsigned short encoding,
318*4882a593Smuzhiyun 			unsigned short parity)
319*4882a593Smuzhiyun {
320*4882a593Smuzhiyun 	struct port *port = dev_to_port(dev);
321*4882a593Smuzhiyun 
322*4882a593Smuzhiyun 	if (encoding != ENCODING_NRZ &&
323*4882a593Smuzhiyun 	    encoding != ENCODING_NRZI)
324*4882a593Smuzhiyun 		return -EINVAL;
325*4882a593Smuzhiyun 
326*4882a593Smuzhiyun 	if (parity != PARITY_NONE &&
327*4882a593Smuzhiyun 	    parity != PARITY_CRC32_PR1_CCITT &&
328*4882a593Smuzhiyun 	    parity != PARITY_CRC16_PR1_CCITT &&
329*4882a593Smuzhiyun 	    parity != PARITY_CRC32_PR0_CCITT &&
330*4882a593Smuzhiyun 	    parity != PARITY_CRC16_PR0_CCITT)
331*4882a593Smuzhiyun 		return -EINVAL;
332*4882a593Smuzhiyun 
333*4882a593Smuzhiyun 	get_status(port)->encoding = encoding;
334*4882a593Smuzhiyun 	get_status(port)->parity = parity;
335*4882a593Smuzhiyun 	return 0;
336*4882a593Smuzhiyun }
337*4882a593Smuzhiyun 
338*4882a593Smuzhiyun 
339*4882a593Smuzhiyun 
wanxl_ioctl(struct net_device * dev,struct ifreq * ifr,int cmd)340*4882a593Smuzhiyun static int wanxl_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
341*4882a593Smuzhiyun {
342*4882a593Smuzhiyun 	const size_t size = sizeof(sync_serial_settings);
343*4882a593Smuzhiyun 	sync_serial_settings line;
344*4882a593Smuzhiyun 	struct port *port = dev_to_port(dev);
345*4882a593Smuzhiyun 
346*4882a593Smuzhiyun 	if (cmd != SIOCWANDEV)
347*4882a593Smuzhiyun 		return hdlc_ioctl(dev, ifr, cmd);
348*4882a593Smuzhiyun 
349*4882a593Smuzhiyun 	switch (ifr->ifr_settings.type) {
350*4882a593Smuzhiyun 	case IF_GET_IFACE:
351*4882a593Smuzhiyun 		ifr->ifr_settings.type = IF_IFACE_SYNC_SERIAL;
352*4882a593Smuzhiyun 		if (ifr->ifr_settings.size < size) {
353*4882a593Smuzhiyun 			ifr->ifr_settings.size = size; /* data size wanted */
354*4882a593Smuzhiyun 			return -ENOBUFS;
355*4882a593Smuzhiyun 		}
356*4882a593Smuzhiyun 		memset(&line, 0, sizeof(line));
357*4882a593Smuzhiyun 		line.clock_type = get_status(port)->clocking;
358*4882a593Smuzhiyun 		line.clock_rate = 0;
359*4882a593Smuzhiyun 		line.loopback = 0;
360*4882a593Smuzhiyun 
361*4882a593Smuzhiyun 		if (copy_to_user(ifr->ifr_settings.ifs_ifsu.sync, &line, size))
362*4882a593Smuzhiyun 			return -EFAULT;
363*4882a593Smuzhiyun 		return 0;
364*4882a593Smuzhiyun 
365*4882a593Smuzhiyun 	case IF_IFACE_SYNC_SERIAL:
366*4882a593Smuzhiyun 		if (!capable(CAP_NET_ADMIN))
367*4882a593Smuzhiyun 			return -EPERM;
368*4882a593Smuzhiyun 		if (dev->flags & IFF_UP)
369*4882a593Smuzhiyun 			return -EBUSY;
370*4882a593Smuzhiyun 
371*4882a593Smuzhiyun 		if (copy_from_user(&line, ifr->ifr_settings.ifs_ifsu.sync,
372*4882a593Smuzhiyun 				   size))
373*4882a593Smuzhiyun 			return -EFAULT;
374*4882a593Smuzhiyun 
375*4882a593Smuzhiyun 		if (line.clock_type != CLOCK_EXT &&
376*4882a593Smuzhiyun 		    line.clock_type != CLOCK_TXFROMRX)
377*4882a593Smuzhiyun 			return -EINVAL; /* No such clock setting */
378*4882a593Smuzhiyun 
379*4882a593Smuzhiyun 		if (line.loopback != 0)
380*4882a593Smuzhiyun 			return -EINVAL;
381*4882a593Smuzhiyun 
382*4882a593Smuzhiyun 		get_status(port)->clocking = line.clock_type;
383*4882a593Smuzhiyun 		return 0;
384*4882a593Smuzhiyun 
385*4882a593Smuzhiyun 	default:
386*4882a593Smuzhiyun 		return hdlc_ioctl(dev, ifr, cmd);
387*4882a593Smuzhiyun         }
388*4882a593Smuzhiyun }
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 
391*4882a593Smuzhiyun 
wanxl_open(struct net_device * dev)392*4882a593Smuzhiyun static int wanxl_open(struct net_device *dev)
393*4882a593Smuzhiyun {
394*4882a593Smuzhiyun 	struct port *port = dev_to_port(dev);
395*4882a593Smuzhiyun 	u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
396*4882a593Smuzhiyun 	unsigned long timeout;
397*4882a593Smuzhiyun 	int i;
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun 	if (get_status(port)->open) {
400*4882a593Smuzhiyun 		netdev_err(dev, "port already open\n");
401*4882a593Smuzhiyun 		return -EIO;
402*4882a593Smuzhiyun 	}
403*4882a593Smuzhiyun 	if ((i = hdlc_open(dev)) != 0)
404*4882a593Smuzhiyun 		return i;
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun 	port->tx_in = port->tx_out = 0;
407*4882a593Smuzhiyun 	for (i = 0; i < TX_BUFFERS; i++)
408*4882a593Smuzhiyun 		get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
409*4882a593Smuzhiyun 	/* signal the card */
410*4882a593Smuzhiyun 	writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
411*4882a593Smuzhiyun 
412*4882a593Smuzhiyun 	timeout = jiffies + HZ;
413*4882a593Smuzhiyun 	do {
414*4882a593Smuzhiyun 		if (get_status(port)->open) {
415*4882a593Smuzhiyun 			netif_start_queue(dev);
416*4882a593Smuzhiyun 			return 0;
417*4882a593Smuzhiyun 		}
418*4882a593Smuzhiyun 	} while (time_after(timeout, jiffies));
419*4882a593Smuzhiyun 
420*4882a593Smuzhiyun 	netdev_err(dev, "unable to open port\n");
421*4882a593Smuzhiyun 	/* ask the card to close the port, should it be still alive */
422*4882a593Smuzhiyun 	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
423*4882a593Smuzhiyun 	return -EFAULT;
424*4882a593Smuzhiyun }
425*4882a593Smuzhiyun 
426*4882a593Smuzhiyun 
427*4882a593Smuzhiyun 
wanxl_close(struct net_device * dev)428*4882a593Smuzhiyun static int wanxl_close(struct net_device *dev)
429*4882a593Smuzhiyun {
430*4882a593Smuzhiyun 	struct port *port = dev_to_port(dev);
431*4882a593Smuzhiyun 	unsigned long timeout;
432*4882a593Smuzhiyun 	int i;
433*4882a593Smuzhiyun 
434*4882a593Smuzhiyun 	hdlc_close(dev);
435*4882a593Smuzhiyun 	/* signal the card */
436*4882a593Smuzhiyun 	writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
437*4882a593Smuzhiyun 	       port->card->plx + PLX_DOORBELL_TO_CARD);
438*4882a593Smuzhiyun 
439*4882a593Smuzhiyun 	timeout = jiffies + HZ;
440*4882a593Smuzhiyun 	do {
441*4882a593Smuzhiyun 		if (!get_status(port)->open)
442*4882a593Smuzhiyun 			break;
443*4882a593Smuzhiyun 	} while (time_after(timeout, jiffies));
444*4882a593Smuzhiyun 
445*4882a593Smuzhiyun 	if (get_status(port)->open)
446*4882a593Smuzhiyun 		netdev_err(dev, "unable to close port\n");
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	netif_stop_queue(dev);
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	for (i = 0; i < TX_BUFFERS; i++) {
451*4882a593Smuzhiyun 		desc_t *desc = &get_status(port)->tx_descs[i];
452*4882a593Smuzhiyun 
453*4882a593Smuzhiyun 		if (desc->stat != PACKET_EMPTY) {
454*4882a593Smuzhiyun 			desc->stat = PACKET_EMPTY;
455*4882a593Smuzhiyun 			dma_unmap_single(&port->card->pdev->dev,
456*4882a593Smuzhiyun 					 desc->address, port->tx_skbs[i]->len,
457*4882a593Smuzhiyun 					 DMA_TO_DEVICE);
458*4882a593Smuzhiyun 			dev_kfree_skb(port->tx_skbs[i]);
459*4882a593Smuzhiyun 		}
460*4882a593Smuzhiyun 	}
461*4882a593Smuzhiyun 	return 0;
462*4882a593Smuzhiyun }
463*4882a593Smuzhiyun 
464*4882a593Smuzhiyun 
465*4882a593Smuzhiyun 
wanxl_get_stats(struct net_device * dev)466*4882a593Smuzhiyun static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
467*4882a593Smuzhiyun {
468*4882a593Smuzhiyun 	struct port *port = dev_to_port(dev);
469*4882a593Smuzhiyun 
470*4882a593Smuzhiyun 	dev->stats.rx_over_errors = get_status(port)->rx_overruns;
471*4882a593Smuzhiyun 	dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
472*4882a593Smuzhiyun 	dev->stats.rx_errors = dev->stats.rx_over_errors +
473*4882a593Smuzhiyun 		dev->stats.rx_frame_errors;
474*4882a593Smuzhiyun 	return &dev->stats;
475*4882a593Smuzhiyun }
476*4882a593Smuzhiyun 
477*4882a593Smuzhiyun 
478*4882a593Smuzhiyun 
wanxl_puts_command(struct card * card,u32 cmd)479*4882a593Smuzhiyun static int wanxl_puts_command(struct card *card, u32 cmd)
480*4882a593Smuzhiyun {
481*4882a593Smuzhiyun 	unsigned long timeout = jiffies + 5 * HZ;
482*4882a593Smuzhiyun 
483*4882a593Smuzhiyun 	writel(cmd, card->plx + PLX_MAILBOX_1);
484*4882a593Smuzhiyun 	do {
485*4882a593Smuzhiyun 		if (readl(card->plx + PLX_MAILBOX_1) == 0)
486*4882a593Smuzhiyun 			return 0;
487*4882a593Smuzhiyun 
488*4882a593Smuzhiyun 		schedule();
489*4882a593Smuzhiyun 	}while (time_after(timeout, jiffies));
490*4882a593Smuzhiyun 
491*4882a593Smuzhiyun 	return -1;
492*4882a593Smuzhiyun }
493*4882a593Smuzhiyun 
494*4882a593Smuzhiyun 
495*4882a593Smuzhiyun 
wanxl_reset(struct card * card)496*4882a593Smuzhiyun static void wanxl_reset(struct card *card)
497*4882a593Smuzhiyun {
498*4882a593Smuzhiyun 	u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun 	writel(0x80, card->plx + PLX_MAILBOX_0);
501*4882a593Smuzhiyun 	writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
502*4882a593Smuzhiyun 	readl(card->plx + PLX_CONTROL); /* wait for posted write */
503*4882a593Smuzhiyun 	udelay(1);
504*4882a593Smuzhiyun 	writel(old_value, card->plx + PLX_CONTROL);
505*4882a593Smuzhiyun 	readl(card->plx + PLX_CONTROL); /* wait for posted write */
506*4882a593Smuzhiyun }
507*4882a593Smuzhiyun 
508*4882a593Smuzhiyun 
509*4882a593Smuzhiyun 
wanxl_pci_remove_one(struct pci_dev * pdev)510*4882a593Smuzhiyun static void wanxl_pci_remove_one(struct pci_dev *pdev)
511*4882a593Smuzhiyun {
512*4882a593Smuzhiyun 	struct card *card = pci_get_drvdata(pdev);
513*4882a593Smuzhiyun 	int i;
514*4882a593Smuzhiyun 
515*4882a593Smuzhiyun 	for (i = 0; i < card->n_ports; i++) {
516*4882a593Smuzhiyun 		unregister_hdlc_device(card->ports[i].dev);
517*4882a593Smuzhiyun 		free_netdev(card->ports[i].dev);
518*4882a593Smuzhiyun 	}
519*4882a593Smuzhiyun 
520*4882a593Smuzhiyun 	/* unregister and free all host resources */
521*4882a593Smuzhiyun 	if (card->irq)
522*4882a593Smuzhiyun 		free_irq(card->irq, card);
523*4882a593Smuzhiyun 
524*4882a593Smuzhiyun 	wanxl_reset(card);
525*4882a593Smuzhiyun 
526*4882a593Smuzhiyun 	for (i = 0; i < RX_QUEUE_LENGTH; i++)
527*4882a593Smuzhiyun 		if (card->rx_skbs[i]) {
528*4882a593Smuzhiyun 			dma_unmap_single(&card->pdev->dev,
529*4882a593Smuzhiyun 					 card->status->rx_descs[i].address,
530*4882a593Smuzhiyun 					 BUFFER_LENGTH, DMA_FROM_DEVICE);
531*4882a593Smuzhiyun 			dev_kfree_skb(card->rx_skbs[i]);
532*4882a593Smuzhiyun 		}
533*4882a593Smuzhiyun 
534*4882a593Smuzhiyun 	if (card->plx)
535*4882a593Smuzhiyun 		iounmap(card->plx);
536*4882a593Smuzhiyun 
537*4882a593Smuzhiyun 	if (card->status)
538*4882a593Smuzhiyun 		dma_free_coherent(&pdev->dev, sizeof(struct card_status),
539*4882a593Smuzhiyun 				  card->status, card->status_address);
540*4882a593Smuzhiyun 
541*4882a593Smuzhiyun 	pci_release_regions(pdev);
542*4882a593Smuzhiyun 	pci_disable_device(pdev);
543*4882a593Smuzhiyun 	kfree(card);
544*4882a593Smuzhiyun }
545*4882a593Smuzhiyun 
546*4882a593Smuzhiyun 
547*4882a593Smuzhiyun #include "wanxlfw.inc"
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun static const struct net_device_ops wanxl_ops = {
550*4882a593Smuzhiyun 	.ndo_open       = wanxl_open,
551*4882a593Smuzhiyun 	.ndo_stop       = wanxl_close,
552*4882a593Smuzhiyun 	.ndo_start_xmit = hdlc_start_xmit,
553*4882a593Smuzhiyun 	.ndo_do_ioctl   = wanxl_ioctl,
554*4882a593Smuzhiyun 	.ndo_get_stats  = wanxl_get_stats,
555*4882a593Smuzhiyun };
556*4882a593Smuzhiyun 
wanxl_pci_init_one(struct pci_dev * pdev,const struct pci_device_id * ent)557*4882a593Smuzhiyun static int wanxl_pci_init_one(struct pci_dev *pdev,
558*4882a593Smuzhiyun 			      const struct pci_device_id *ent)
559*4882a593Smuzhiyun {
560*4882a593Smuzhiyun 	struct card *card;
561*4882a593Smuzhiyun 	u32 ramsize, stat;
562*4882a593Smuzhiyun 	unsigned long timeout;
563*4882a593Smuzhiyun 	u32 plx_phy;		/* PLX PCI base address */
564*4882a593Smuzhiyun 	u32 mem_phy;		/* memory PCI base addr */
565*4882a593Smuzhiyun 	u8 __iomem *mem;	/* memory virtual base addr */
566*4882a593Smuzhiyun 	int i, ports;
567*4882a593Smuzhiyun 
568*4882a593Smuzhiyun #ifndef MODULE
569*4882a593Smuzhiyun 	pr_info_once("%s\n", version);
570*4882a593Smuzhiyun #endif
571*4882a593Smuzhiyun 
572*4882a593Smuzhiyun 	i = pci_enable_device(pdev);
573*4882a593Smuzhiyun 	if (i)
574*4882a593Smuzhiyun 		return i;
575*4882a593Smuzhiyun 
576*4882a593Smuzhiyun 	/* QUICC can only access first 256 MB of host RAM directly,
577*4882a593Smuzhiyun 	   but PLX9060 DMA does 32-bits for actual packet data transfers */
578*4882a593Smuzhiyun 
579*4882a593Smuzhiyun 	/* FIXME when PCI/DMA subsystems are fixed.
580*4882a593Smuzhiyun 	   We set both dma_mask and consistent_dma_mask to 28 bits
581*4882a593Smuzhiyun 	   and pray pci_alloc_consistent() will use this info. It should
582*4882a593Smuzhiyun 	   work on most platforms */
583*4882a593Smuzhiyun 	if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(28)) ||
584*4882a593Smuzhiyun 	    dma_set_mask(&pdev->dev, DMA_BIT_MASK(28))) {
585*4882a593Smuzhiyun 		pr_err("No usable DMA configuration\n");
586*4882a593Smuzhiyun 		pci_disable_device(pdev);
587*4882a593Smuzhiyun 		return -EIO;
588*4882a593Smuzhiyun 	}
589*4882a593Smuzhiyun 
590*4882a593Smuzhiyun 	i = pci_request_regions(pdev, "wanXL");
591*4882a593Smuzhiyun 	if (i) {
592*4882a593Smuzhiyun 		pci_disable_device(pdev);
593*4882a593Smuzhiyun 		return i;
594*4882a593Smuzhiyun 	}
595*4882a593Smuzhiyun 
596*4882a593Smuzhiyun 	switch (pdev->device) {
597*4882a593Smuzhiyun 	case PCI_DEVICE_ID_SBE_WANXL100: ports = 1; break;
598*4882a593Smuzhiyun 	case PCI_DEVICE_ID_SBE_WANXL200: ports = 2; break;
599*4882a593Smuzhiyun 	default: ports = 4;
600*4882a593Smuzhiyun 	}
601*4882a593Smuzhiyun 
602*4882a593Smuzhiyun 	card = kzalloc(struct_size(card, ports, ports), GFP_KERNEL);
603*4882a593Smuzhiyun 	if (card == NULL) {
604*4882a593Smuzhiyun 		pci_release_regions(pdev);
605*4882a593Smuzhiyun 		pci_disable_device(pdev);
606*4882a593Smuzhiyun 		return -ENOBUFS;
607*4882a593Smuzhiyun 	}
608*4882a593Smuzhiyun 
609*4882a593Smuzhiyun 	pci_set_drvdata(pdev, card);
610*4882a593Smuzhiyun 	card->pdev = pdev;
611*4882a593Smuzhiyun 
612*4882a593Smuzhiyun 	card->status = dma_alloc_coherent(&pdev->dev,
613*4882a593Smuzhiyun 					  sizeof(struct card_status),
614*4882a593Smuzhiyun 					  &card->status_address, GFP_KERNEL);
615*4882a593Smuzhiyun 	if (card->status == NULL) {
616*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
617*4882a593Smuzhiyun 		return -ENOBUFS;
618*4882a593Smuzhiyun 	}
619*4882a593Smuzhiyun 
620*4882a593Smuzhiyun #ifdef DEBUG_PCI
621*4882a593Smuzhiyun 	printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
622*4882a593Smuzhiyun 	       " at 0x%LX\n", pci_name(pdev),
623*4882a593Smuzhiyun 	       (unsigned long long)card->status_address);
624*4882a593Smuzhiyun #endif
625*4882a593Smuzhiyun 
626*4882a593Smuzhiyun 	/* FIXME when PCI/DMA subsystems are fixed.
627*4882a593Smuzhiyun 	   We set both dma_mask and consistent_dma_mask back to 32 bits
628*4882a593Smuzhiyun 	   to indicate the card can do 32-bit DMA addressing */
629*4882a593Smuzhiyun 	if (dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)) ||
630*4882a593Smuzhiyun 	    dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
631*4882a593Smuzhiyun 		pr_err("No usable DMA configuration\n");
632*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
633*4882a593Smuzhiyun 		return -EIO;
634*4882a593Smuzhiyun 	}
635*4882a593Smuzhiyun 
636*4882a593Smuzhiyun 	/* set up PLX mapping */
637*4882a593Smuzhiyun 	plx_phy = pci_resource_start(pdev, 0);
638*4882a593Smuzhiyun 
639*4882a593Smuzhiyun 	card->plx = ioremap(plx_phy, 0x70);
640*4882a593Smuzhiyun 	if (!card->plx) {
641*4882a593Smuzhiyun 		pr_err("ioremap() failed\n");
642*4882a593Smuzhiyun  		wanxl_pci_remove_one(pdev);
643*4882a593Smuzhiyun 		return -EFAULT;
644*4882a593Smuzhiyun 	}
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun #if RESET_WHILE_LOADING
647*4882a593Smuzhiyun 	wanxl_reset(card);
648*4882a593Smuzhiyun #endif
649*4882a593Smuzhiyun 
650*4882a593Smuzhiyun 	timeout = jiffies + 20 * HZ;
651*4882a593Smuzhiyun 	while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
652*4882a593Smuzhiyun 		if (time_before(timeout, jiffies)) {
653*4882a593Smuzhiyun 			pr_warn("%s: timeout waiting for PUTS to complete\n",
654*4882a593Smuzhiyun 				pci_name(pdev));
655*4882a593Smuzhiyun 			wanxl_pci_remove_one(pdev);
656*4882a593Smuzhiyun 			return -ENODEV;
657*4882a593Smuzhiyun 		}
658*4882a593Smuzhiyun 
659*4882a593Smuzhiyun 		switch(stat & 0xC0) {
660*4882a593Smuzhiyun 		case 0x00:	/* hmm - PUTS completed with non-zero code? */
661*4882a593Smuzhiyun 		case 0x80:	/* PUTS still testing the hardware */
662*4882a593Smuzhiyun 			break;
663*4882a593Smuzhiyun 
664*4882a593Smuzhiyun 		default:
665*4882a593Smuzhiyun 			pr_warn("%s: PUTS test 0x%X failed\n",
666*4882a593Smuzhiyun 				pci_name(pdev), stat & 0x30);
667*4882a593Smuzhiyun 			wanxl_pci_remove_one(pdev);
668*4882a593Smuzhiyun 			return -ENODEV;
669*4882a593Smuzhiyun 		}
670*4882a593Smuzhiyun 
671*4882a593Smuzhiyun 		schedule();
672*4882a593Smuzhiyun 	}
673*4882a593Smuzhiyun 
674*4882a593Smuzhiyun 	/* get on-board memory size (PUTS detects no more than 4 MB) */
675*4882a593Smuzhiyun 	ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
676*4882a593Smuzhiyun 
677*4882a593Smuzhiyun 	/* set up on-board RAM mapping */
678*4882a593Smuzhiyun 	mem_phy = pci_resource_start(pdev, 2);
679*4882a593Smuzhiyun 
680*4882a593Smuzhiyun 
681*4882a593Smuzhiyun 	/* sanity check the board's reported memory size */
682*4882a593Smuzhiyun 	if (ramsize < BUFFERS_ADDR +
683*4882a593Smuzhiyun 	    (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
684*4882a593Smuzhiyun 		pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
685*4882a593Smuzhiyun 			pci_name(pdev), ramsize,
686*4882a593Smuzhiyun 			BUFFERS_ADDR +
687*4882a593Smuzhiyun 			(TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
688*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
689*4882a593Smuzhiyun 		return -ENODEV;
690*4882a593Smuzhiyun 	}
691*4882a593Smuzhiyun 
692*4882a593Smuzhiyun 	if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
693*4882a593Smuzhiyun 		pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
694*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
695*4882a593Smuzhiyun 		return -ENODEV;
696*4882a593Smuzhiyun 	}
697*4882a593Smuzhiyun 
698*4882a593Smuzhiyun 	for (i = 0; i < RX_QUEUE_LENGTH; i++) {
699*4882a593Smuzhiyun 		struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
700*4882a593Smuzhiyun 		card->rx_skbs[i] = skb;
701*4882a593Smuzhiyun 		if (skb)
702*4882a593Smuzhiyun 			card->status->rx_descs[i].address =
703*4882a593Smuzhiyun 				dma_map_single(&card->pdev->dev, skb->data,
704*4882a593Smuzhiyun 					       BUFFER_LENGTH, DMA_FROM_DEVICE);
705*4882a593Smuzhiyun 	}
706*4882a593Smuzhiyun 
707*4882a593Smuzhiyun 	mem = ioremap(mem_phy, PDM_OFFSET + sizeof(firmware));
708*4882a593Smuzhiyun 	if (!mem) {
709*4882a593Smuzhiyun 		pr_err("ioremap() failed\n");
710*4882a593Smuzhiyun  		wanxl_pci_remove_one(pdev);
711*4882a593Smuzhiyun 		return -EFAULT;
712*4882a593Smuzhiyun 	}
713*4882a593Smuzhiyun 
714*4882a593Smuzhiyun 	for (i = 0; i < sizeof(firmware); i += 4)
715*4882a593Smuzhiyun 		writel(ntohl(*(__be32*)(firmware + i)), mem + PDM_OFFSET + i);
716*4882a593Smuzhiyun 
717*4882a593Smuzhiyun 	for (i = 0; i < ports; i++)
718*4882a593Smuzhiyun 		writel(card->status_address +
719*4882a593Smuzhiyun 		       (void *)&card->status->port_status[i] -
720*4882a593Smuzhiyun 		       (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
721*4882a593Smuzhiyun 	writel(card->status_address, mem + PDM_OFFSET + 20);
722*4882a593Smuzhiyun 	writel(PDM_OFFSET, mem);
723*4882a593Smuzhiyun 	iounmap(mem);
724*4882a593Smuzhiyun 
725*4882a593Smuzhiyun 	writel(0, card->plx + PLX_MAILBOX_5);
726*4882a593Smuzhiyun 
727*4882a593Smuzhiyun 	if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
728*4882a593Smuzhiyun 		pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
729*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
730*4882a593Smuzhiyun 		return -ENODEV;
731*4882a593Smuzhiyun 	}
732*4882a593Smuzhiyun 
733*4882a593Smuzhiyun 	timeout = jiffies + 5 * HZ;
734*4882a593Smuzhiyun 	do {
735*4882a593Smuzhiyun 		if ((stat = readl(card->plx + PLX_MAILBOX_5)) != 0)
736*4882a593Smuzhiyun 			break;
737*4882a593Smuzhiyun 		schedule();
738*4882a593Smuzhiyun 	}while (time_after(timeout, jiffies));
739*4882a593Smuzhiyun 
740*4882a593Smuzhiyun 	if (!stat) {
741*4882a593Smuzhiyun 		pr_warn("%s: timeout while initializing card firmware\n",
742*4882a593Smuzhiyun 			pci_name(pdev));
743*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
744*4882a593Smuzhiyun 		return -ENODEV;
745*4882a593Smuzhiyun 	}
746*4882a593Smuzhiyun 
747*4882a593Smuzhiyun #if DETECT_RAM
748*4882a593Smuzhiyun 	ramsize = stat;
749*4882a593Smuzhiyun #endif
750*4882a593Smuzhiyun 
751*4882a593Smuzhiyun 	pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
752*4882a593Smuzhiyun 		pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
753*4882a593Smuzhiyun 
754*4882a593Smuzhiyun 	/* Allocate IRQ */
755*4882a593Smuzhiyun 	if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
756*4882a593Smuzhiyun 		pr_warn("%s: could not allocate IRQ%i\n",
757*4882a593Smuzhiyun 			pci_name(pdev), pdev->irq);
758*4882a593Smuzhiyun 		wanxl_pci_remove_one(pdev);
759*4882a593Smuzhiyun 		return -EBUSY;
760*4882a593Smuzhiyun 	}
761*4882a593Smuzhiyun 	card->irq = pdev->irq;
762*4882a593Smuzhiyun 
763*4882a593Smuzhiyun 	for (i = 0; i < ports; i++) {
764*4882a593Smuzhiyun 		hdlc_device *hdlc;
765*4882a593Smuzhiyun 		struct port *port = &card->ports[i];
766*4882a593Smuzhiyun 		struct net_device *dev = alloc_hdlcdev(port);
767*4882a593Smuzhiyun 		if (!dev) {
768*4882a593Smuzhiyun 			pr_err("%s: unable to allocate memory\n",
769*4882a593Smuzhiyun 			       pci_name(pdev));
770*4882a593Smuzhiyun 			wanxl_pci_remove_one(pdev);
771*4882a593Smuzhiyun 			return -ENOMEM;
772*4882a593Smuzhiyun 		}
773*4882a593Smuzhiyun 
774*4882a593Smuzhiyun 		port->dev = dev;
775*4882a593Smuzhiyun 		hdlc = dev_to_hdlc(dev);
776*4882a593Smuzhiyun 		spin_lock_init(&port->lock);
777*4882a593Smuzhiyun 		dev->tx_queue_len = 50;
778*4882a593Smuzhiyun 		dev->netdev_ops = &wanxl_ops;
779*4882a593Smuzhiyun 		hdlc->attach = wanxl_attach;
780*4882a593Smuzhiyun 		hdlc->xmit = wanxl_xmit;
781*4882a593Smuzhiyun 		port->card = card;
782*4882a593Smuzhiyun 		port->node = i;
783*4882a593Smuzhiyun 		get_status(port)->clocking = CLOCK_EXT;
784*4882a593Smuzhiyun 		if (register_hdlc_device(dev)) {
785*4882a593Smuzhiyun 			pr_err("%s: unable to register hdlc device\n",
786*4882a593Smuzhiyun 			       pci_name(pdev));
787*4882a593Smuzhiyun 			free_netdev(dev);
788*4882a593Smuzhiyun 			wanxl_pci_remove_one(pdev);
789*4882a593Smuzhiyun 			return -ENOBUFS;
790*4882a593Smuzhiyun 		}
791*4882a593Smuzhiyun 		card->n_ports++;
792*4882a593Smuzhiyun 	}
793*4882a593Smuzhiyun 
794*4882a593Smuzhiyun 	pr_info("%s: port", pci_name(pdev));
795*4882a593Smuzhiyun 	for (i = 0; i < ports; i++)
796*4882a593Smuzhiyun 		pr_cont("%s #%i: %s",
797*4882a593Smuzhiyun 			i ? "," : "", i, card->ports[i].dev->name);
798*4882a593Smuzhiyun 	pr_cont("\n");
799*4882a593Smuzhiyun 
800*4882a593Smuzhiyun 	for (i = 0; i < ports; i++)
801*4882a593Smuzhiyun 		wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
802*4882a593Smuzhiyun 
803*4882a593Smuzhiyun 	return 0;
804*4882a593Smuzhiyun }
805*4882a593Smuzhiyun 
806*4882a593Smuzhiyun static const struct pci_device_id wanxl_pci_tbl[] = {
807*4882a593Smuzhiyun 	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
808*4882a593Smuzhiyun 	  PCI_ANY_ID, 0, 0, 0 },
809*4882a593Smuzhiyun 	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
810*4882a593Smuzhiyun 	  PCI_ANY_ID, 0, 0, 0 },
811*4882a593Smuzhiyun 	{ PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
812*4882a593Smuzhiyun 	  PCI_ANY_ID, 0, 0, 0 },
813*4882a593Smuzhiyun 	{ 0, }
814*4882a593Smuzhiyun };
815*4882a593Smuzhiyun 
816*4882a593Smuzhiyun 
817*4882a593Smuzhiyun static struct pci_driver wanxl_pci_driver = {
818*4882a593Smuzhiyun 	.name		= "wanXL",
819*4882a593Smuzhiyun 	.id_table	= wanxl_pci_tbl,
820*4882a593Smuzhiyun 	.probe		= wanxl_pci_init_one,
821*4882a593Smuzhiyun 	.remove		= wanxl_pci_remove_one,
822*4882a593Smuzhiyun };
823*4882a593Smuzhiyun 
824*4882a593Smuzhiyun 
wanxl_init_module(void)825*4882a593Smuzhiyun static int __init wanxl_init_module(void)
826*4882a593Smuzhiyun {
827*4882a593Smuzhiyun #ifdef MODULE
828*4882a593Smuzhiyun 	pr_info("%s\n", version);
829*4882a593Smuzhiyun #endif
830*4882a593Smuzhiyun 	return pci_register_driver(&wanxl_pci_driver);
831*4882a593Smuzhiyun }
832*4882a593Smuzhiyun 
wanxl_cleanup_module(void)833*4882a593Smuzhiyun static void __exit wanxl_cleanup_module(void)
834*4882a593Smuzhiyun {
835*4882a593Smuzhiyun 	pci_unregister_driver(&wanxl_pci_driver);
836*4882a593Smuzhiyun }
837*4882a593Smuzhiyun 
838*4882a593Smuzhiyun 
839*4882a593Smuzhiyun MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
840*4882a593Smuzhiyun MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
841*4882a593Smuzhiyun MODULE_LICENSE("GPL v2");
842*4882a593Smuzhiyun MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
843*4882a593Smuzhiyun 
844*4882a593Smuzhiyun module_init(wanxl_init_module);
845*4882a593Smuzhiyun module_exit(wanxl_cleanup_module);
846