xref: /OK3568_Linux_fs/kernel/drivers/ide/cmd64x.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * cmd64x.c: Enable interrupts at initialization time on Ultra/PCI machines.
4*4882a593Smuzhiyun  *           Due to massive hardware bugs, UltraDMA is only supported
5*4882a593Smuzhiyun  *           on the 646U2 and not on the 646U.
6*4882a593Smuzhiyun  *
7*4882a593Smuzhiyun  * Copyright (C) 1998		Eddie C. Dost  (ecd@skynet.be)
8*4882a593Smuzhiyun  * Copyright (C) 1998		David S. Miller (davem@redhat.com)
9*4882a593Smuzhiyun  *
10*4882a593Smuzhiyun  * Copyright (C) 1999-2002	Andre Hedrick <andre@linux-ide.org>
11*4882a593Smuzhiyun  * Copyright (C) 2007-2010	Bartlomiej Zolnierkiewicz
12*4882a593Smuzhiyun  * Copyright (C) 2007,2009	MontaVista Software, Inc. <source@mvista.com>
13*4882a593Smuzhiyun  */
14*4882a593Smuzhiyun 
15*4882a593Smuzhiyun #include <linux/module.h>
16*4882a593Smuzhiyun #include <linux/types.h>
17*4882a593Smuzhiyun #include <linux/pci.h>
18*4882a593Smuzhiyun #include <linux/ide.h>
19*4882a593Smuzhiyun #include <linux/init.h>
20*4882a593Smuzhiyun 
21*4882a593Smuzhiyun #include <asm/io.h>
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun #define DRV_NAME "cmd64x"
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun /*
26*4882a593Smuzhiyun  * CMD64x specific registers definition.
27*4882a593Smuzhiyun  */
28*4882a593Smuzhiyun #define CFR		0x50
29*4882a593Smuzhiyun #define   CFR_INTR_CH0		0x04
30*4882a593Smuzhiyun 
31*4882a593Smuzhiyun #define	CMDTIM		0x52
32*4882a593Smuzhiyun #define	ARTTIM0		0x53
33*4882a593Smuzhiyun #define	DRWTIM0		0x54
34*4882a593Smuzhiyun #define ARTTIM1 	0x55
35*4882a593Smuzhiyun #define DRWTIM1		0x56
36*4882a593Smuzhiyun #define ARTTIM23	0x57
37*4882a593Smuzhiyun #define   ARTTIM23_DIS_RA2	0x04
38*4882a593Smuzhiyun #define   ARTTIM23_DIS_RA3	0x08
39*4882a593Smuzhiyun #define   ARTTIM23_INTR_CH1	0x10
40*4882a593Smuzhiyun #define DRWTIM2		0x58
41*4882a593Smuzhiyun #define BRST		0x59
42*4882a593Smuzhiyun #define DRWTIM3		0x5b
43*4882a593Smuzhiyun 
44*4882a593Smuzhiyun #define BMIDECR0	0x70
45*4882a593Smuzhiyun #define MRDMODE		0x71
46*4882a593Smuzhiyun #define   MRDMODE_INTR_CH0	0x04
47*4882a593Smuzhiyun #define   MRDMODE_INTR_CH1	0x08
48*4882a593Smuzhiyun #define UDIDETCR0	0x73
49*4882a593Smuzhiyun #define DTPR0		0x74
50*4882a593Smuzhiyun #define BMIDECR1	0x78
51*4882a593Smuzhiyun #define BMIDECSR	0x79
52*4882a593Smuzhiyun #define UDIDETCR1	0x7B
53*4882a593Smuzhiyun #define DTPR1		0x7C
54*4882a593Smuzhiyun 
cmd64x_program_timings(ide_drive_t * drive,u8 mode)55*4882a593Smuzhiyun static void cmd64x_program_timings(ide_drive_t *drive, u8 mode)
56*4882a593Smuzhiyun {
57*4882a593Smuzhiyun 	ide_hwif_t *hwif = drive->hwif;
58*4882a593Smuzhiyun 	struct pci_dev *dev = to_pci_dev(drive->hwif->dev);
59*4882a593Smuzhiyun 	int bus_speed = ide_pci_clk ? ide_pci_clk : 33;
60*4882a593Smuzhiyun 	const unsigned long T = 1000000 / bus_speed;
61*4882a593Smuzhiyun 	static const u8 recovery_values[] =
62*4882a593Smuzhiyun 		{15, 15, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 0};
63*4882a593Smuzhiyun 	static const u8 setup_values[] = {0x40, 0x40, 0x40, 0x80, 0, 0xc0};
64*4882a593Smuzhiyun 	static const u8 arttim_regs[4] = {ARTTIM0, ARTTIM1, ARTTIM23, ARTTIM23};
65*4882a593Smuzhiyun 	static const u8 drwtim_regs[4] = {DRWTIM0, DRWTIM1, DRWTIM2, DRWTIM3};
66*4882a593Smuzhiyun 	struct ide_timing t;
67*4882a593Smuzhiyun 	u8 arttim = 0;
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun 	if (drive->dn >= ARRAY_SIZE(drwtim_regs))
70*4882a593Smuzhiyun 		return;
71*4882a593Smuzhiyun 
72*4882a593Smuzhiyun 	ide_timing_compute(drive, mode, &t, T, 0);
73*4882a593Smuzhiyun 
74*4882a593Smuzhiyun 	/*
75*4882a593Smuzhiyun 	 * In case we've got too long recovery phase, try to lengthen
76*4882a593Smuzhiyun 	 * the active phase
77*4882a593Smuzhiyun 	 */
78*4882a593Smuzhiyun 	if (t.recover > 16) {
79*4882a593Smuzhiyun 		t.active += t.recover - 16;
80*4882a593Smuzhiyun 		t.recover = 16;
81*4882a593Smuzhiyun 	}
82*4882a593Smuzhiyun 	if (t.active > 16)		/* shouldn't actually happen... */
83*4882a593Smuzhiyun 		t.active = 16;
84*4882a593Smuzhiyun 
85*4882a593Smuzhiyun 	/*
86*4882a593Smuzhiyun 	 * Convert values to internal chipset representation
87*4882a593Smuzhiyun 	 */
88*4882a593Smuzhiyun 	t.recover = recovery_values[t.recover];
89*4882a593Smuzhiyun 	t.active &= 0x0f;
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 	/* Program the active/recovery counts into the DRWTIM register */
92*4882a593Smuzhiyun 	pci_write_config_byte(dev, drwtim_regs[drive->dn],
93*4882a593Smuzhiyun 			      (t.active << 4) | t.recover);
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun 	/*
96*4882a593Smuzhiyun 	 * The primary channel has individual address setup timing registers
97*4882a593Smuzhiyun 	 * for each drive and the hardware selects the slowest timing itself.
98*4882a593Smuzhiyun 	 * The secondary channel has one common register and we have to select
99*4882a593Smuzhiyun 	 * the slowest address setup timing ourselves.
100*4882a593Smuzhiyun 	 */
101*4882a593Smuzhiyun 	if (hwif->channel) {
102*4882a593Smuzhiyun 		ide_drive_t *pair = ide_get_pair_dev(drive);
103*4882a593Smuzhiyun 
104*4882a593Smuzhiyun 		if (pair) {
105*4882a593Smuzhiyun 			struct ide_timing tp;
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun 			ide_timing_compute(pair, pair->pio_mode, &tp, T, 0);
108*4882a593Smuzhiyun 			ide_timing_merge(&t, &tp, &t, IDE_TIMING_SETUP);
109*4882a593Smuzhiyun 			if (pair->dma_mode) {
110*4882a593Smuzhiyun 				ide_timing_compute(pair, pair->dma_mode,
111*4882a593Smuzhiyun 						&tp, T, 0);
112*4882a593Smuzhiyun 				ide_timing_merge(&tp, &t, &t, IDE_TIMING_SETUP);
113*4882a593Smuzhiyun 			}
114*4882a593Smuzhiyun 		}
115*4882a593Smuzhiyun 	}
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun 	if (t.setup > 5)		/* shouldn't actually happen... */
118*4882a593Smuzhiyun 		t.setup = 5;
119*4882a593Smuzhiyun 
120*4882a593Smuzhiyun 	/*
121*4882a593Smuzhiyun 	 * Program the address setup clocks into the ARTTIM registers.
122*4882a593Smuzhiyun 	 * Avoid clearing the secondary channel's interrupt bit.
123*4882a593Smuzhiyun 	 */
124*4882a593Smuzhiyun 	(void) pci_read_config_byte (dev, arttim_regs[drive->dn], &arttim);
125*4882a593Smuzhiyun 	if (hwif->channel)
126*4882a593Smuzhiyun 		arttim &= ~ARTTIM23_INTR_CH1;
127*4882a593Smuzhiyun 	arttim &= ~0xc0;
128*4882a593Smuzhiyun 	arttim |= setup_values[t.setup];
129*4882a593Smuzhiyun 	(void) pci_write_config_byte(dev, arttim_regs[drive->dn], arttim);
130*4882a593Smuzhiyun }
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun /*
133*4882a593Smuzhiyun  * Attempts to set drive's PIO mode.
134*4882a593Smuzhiyun  * Special cases are 8: prefetch off, 9: prefetch on (both never worked)
135*4882a593Smuzhiyun  */
136*4882a593Smuzhiyun 
cmd64x_set_pio_mode(ide_hwif_t * hwif,ide_drive_t * drive)137*4882a593Smuzhiyun static void cmd64x_set_pio_mode(ide_hwif_t *hwif, ide_drive_t *drive)
138*4882a593Smuzhiyun {
139*4882a593Smuzhiyun 	const u8 pio = drive->pio_mode - XFER_PIO_0;
140*4882a593Smuzhiyun 
141*4882a593Smuzhiyun 	/*
142*4882a593Smuzhiyun 	 * Filter out the prefetch control values
143*4882a593Smuzhiyun 	 * to prevent PIO5 from being programmed
144*4882a593Smuzhiyun 	 */
145*4882a593Smuzhiyun 	if (pio == 8 || pio == 9)
146*4882a593Smuzhiyun 		return;
147*4882a593Smuzhiyun 
148*4882a593Smuzhiyun 	cmd64x_program_timings(drive, XFER_PIO_0 + pio);
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun 
cmd64x_set_dma_mode(ide_hwif_t * hwif,ide_drive_t * drive)151*4882a593Smuzhiyun static void cmd64x_set_dma_mode(ide_hwif_t *hwif, ide_drive_t *drive)
152*4882a593Smuzhiyun {
153*4882a593Smuzhiyun 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
154*4882a593Smuzhiyun 	u8 unit			= drive->dn & 0x01;
155*4882a593Smuzhiyun 	u8 regU = 0, pciU	= hwif->channel ? UDIDETCR1 : UDIDETCR0;
156*4882a593Smuzhiyun 	const u8 speed		= drive->dma_mode;
157*4882a593Smuzhiyun 
158*4882a593Smuzhiyun 	pci_read_config_byte(dev, pciU, &regU);
159*4882a593Smuzhiyun 	regU &= ~(unit ? 0xCA : 0x35);
160*4882a593Smuzhiyun 
161*4882a593Smuzhiyun 	switch(speed) {
162*4882a593Smuzhiyun 	case XFER_UDMA_5:
163*4882a593Smuzhiyun 		regU |= unit ? 0x0A : 0x05;
164*4882a593Smuzhiyun 		break;
165*4882a593Smuzhiyun 	case XFER_UDMA_4:
166*4882a593Smuzhiyun 		regU |= unit ? 0x4A : 0x15;
167*4882a593Smuzhiyun 		break;
168*4882a593Smuzhiyun 	case XFER_UDMA_3:
169*4882a593Smuzhiyun 		regU |= unit ? 0x8A : 0x25;
170*4882a593Smuzhiyun 		break;
171*4882a593Smuzhiyun 	case XFER_UDMA_2:
172*4882a593Smuzhiyun 		regU |= unit ? 0x42 : 0x11;
173*4882a593Smuzhiyun 		break;
174*4882a593Smuzhiyun 	case XFER_UDMA_1:
175*4882a593Smuzhiyun 		regU |= unit ? 0x82 : 0x21;
176*4882a593Smuzhiyun 		break;
177*4882a593Smuzhiyun 	case XFER_UDMA_0:
178*4882a593Smuzhiyun 		regU |= unit ? 0xC2 : 0x31;
179*4882a593Smuzhiyun 		break;
180*4882a593Smuzhiyun 	case XFER_MW_DMA_2:
181*4882a593Smuzhiyun 	case XFER_MW_DMA_1:
182*4882a593Smuzhiyun 	case XFER_MW_DMA_0:
183*4882a593Smuzhiyun 		cmd64x_program_timings(drive, speed);
184*4882a593Smuzhiyun 		break;
185*4882a593Smuzhiyun 	}
186*4882a593Smuzhiyun 
187*4882a593Smuzhiyun 	pci_write_config_byte(dev, pciU, regU);
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun 
cmd648_clear_irq(ide_drive_t * drive)190*4882a593Smuzhiyun static void cmd648_clear_irq(ide_drive_t *drive)
191*4882a593Smuzhiyun {
192*4882a593Smuzhiyun 	ide_hwif_t *hwif	= drive->hwif;
193*4882a593Smuzhiyun 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
194*4882a593Smuzhiyun 	unsigned long base	= pci_resource_start(dev, 4);
195*4882a593Smuzhiyun 	u8  irq_mask		= hwif->channel ? MRDMODE_INTR_CH1 :
196*4882a593Smuzhiyun 						  MRDMODE_INTR_CH0;
197*4882a593Smuzhiyun 	u8  mrdmode		= inb(base + 1);
198*4882a593Smuzhiyun 
199*4882a593Smuzhiyun 	/* clear the interrupt bit */
200*4882a593Smuzhiyun 	outb((mrdmode & ~(MRDMODE_INTR_CH0 | MRDMODE_INTR_CH1)) | irq_mask,
201*4882a593Smuzhiyun 	     base + 1);
202*4882a593Smuzhiyun }
203*4882a593Smuzhiyun 
cmd64x_clear_irq(ide_drive_t * drive)204*4882a593Smuzhiyun static void cmd64x_clear_irq(ide_drive_t *drive)
205*4882a593Smuzhiyun {
206*4882a593Smuzhiyun 	ide_hwif_t *hwif	= drive->hwif;
207*4882a593Smuzhiyun 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
208*4882a593Smuzhiyun 	int irq_reg		= hwif->channel ? ARTTIM23 : CFR;
209*4882a593Smuzhiyun 	u8  irq_mask		= hwif->channel ? ARTTIM23_INTR_CH1 :
210*4882a593Smuzhiyun 						  CFR_INTR_CH0;
211*4882a593Smuzhiyun 	u8  irq_stat		= 0;
212*4882a593Smuzhiyun 
213*4882a593Smuzhiyun 	(void) pci_read_config_byte(dev, irq_reg, &irq_stat);
214*4882a593Smuzhiyun 	/* clear the interrupt bit */
215*4882a593Smuzhiyun 	(void) pci_write_config_byte(dev, irq_reg, irq_stat | irq_mask);
216*4882a593Smuzhiyun }
217*4882a593Smuzhiyun 
cmd648_test_irq(ide_hwif_t * hwif)218*4882a593Smuzhiyun static int cmd648_test_irq(ide_hwif_t *hwif)
219*4882a593Smuzhiyun {
220*4882a593Smuzhiyun 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
221*4882a593Smuzhiyun 	unsigned long base	= pci_resource_start(dev, 4);
222*4882a593Smuzhiyun 	u8 irq_mask		= hwif->channel ? MRDMODE_INTR_CH1 :
223*4882a593Smuzhiyun 						  MRDMODE_INTR_CH0;
224*4882a593Smuzhiyun 	u8 mrdmode		= inb(base + 1);
225*4882a593Smuzhiyun 
226*4882a593Smuzhiyun 	pr_debug("%s: mrdmode: 0x%02x irq_mask: 0x%02x\n",
227*4882a593Smuzhiyun 		 hwif->name, mrdmode, irq_mask);
228*4882a593Smuzhiyun 
229*4882a593Smuzhiyun 	return (mrdmode & irq_mask) ? 1 : 0;
230*4882a593Smuzhiyun }
231*4882a593Smuzhiyun 
cmd64x_test_irq(ide_hwif_t * hwif)232*4882a593Smuzhiyun static int cmd64x_test_irq(ide_hwif_t *hwif)
233*4882a593Smuzhiyun {
234*4882a593Smuzhiyun 	struct pci_dev *dev	= to_pci_dev(hwif->dev);
235*4882a593Smuzhiyun 	int irq_reg		= hwif->channel ? ARTTIM23 : CFR;
236*4882a593Smuzhiyun 	u8  irq_mask		= hwif->channel ? ARTTIM23_INTR_CH1 :
237*4882a593Smuzhiyun 						  CFR_INTR_CH0;
238*4882a593Smuzhiyun 	u8  irq_stat		= 0;
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun 	(void) pci_read_config_byte(dev, irq_reg, &irq_stat);
241*4882a593Smuzhiyun 
242*4882a593Smuzhiyun 	pr_debug("%s: irq_stat: 0x%02x irq_mask: 0x%02x\n",
243*4882a593Smuzhiyun 		 hwif->name, irq_stat, irq_mask);
244*4882a593Smuzhiyun 
245*4882a593Smuzhiyun 	return (irq_stat & irq_mask) ? 1 : 0;
246*4882a593Smuzhiyun }
247*4882a593Smuzhiyun 
248*4882a593Smuzhiyun /*
249*4882a593Smuzhiyun  * ASUS P55T2P4D with CMD646 chipset revision 0x01 requires the old
250*4882a593Smuzhiyun  * event order for DMA transfers.
251*4882a593Smuzhiyun  */
252*4882a593Smuzhiyun 
cmd646_1_dma_end(ide_drive_t * drive)253*4882a593Smuzhiyun static int cmd646_1_dma_end(ide_drive_t *drive)
254*4882a593Smuzhiyun {
255*4882a593Smuzhiyun 	ide_hwif_t *hwif = drive->hwif;
256*4882a593Smuzhiyun 	u8 dma_stat = 0, dma_cmd = 0;
257*4882a593Smuzhiyun 
258*4882a593Smuzhiyun 	/* get DMA status */
259*4882a593Smuzhiyun 	dma_stat = inb(hwif->dma_base + ATA_DMA_STATUS);
260*4882a593Smuzhiyun 	/* read DMA command state */
261*4882a593Smuzhiyun 	dma_cmd = inb(hwif->dma_base + ATA_DMA_CMD);
262*4882a593Smuzhiyun 	/* stop DMA */
263*4882a593Smuzhiyun 	outb(dma_cmd & ~1, hwif->dma_base + ATA_DMA_CMD);
264*4882a593Smuzhiyun 	/* clear the INTR & ERROR bits */
265*4882a593Smuzhiyun 	outb(dma_stat | 6, hwif->dma_base + ATA_DMA_STATUS);
266*4882a593Smuzhiyun 	/* verify good DMA status */
267*4882a593Smuzhiyun 	return (dma_stat & 7) != 4;
268*4882a593Smuzhiyun }
269*4882a593Smuzhiyun 
init_chipset_cmd64x(struct pci_dev * dev)270*4882a593Smuzhiyun static int init_chipset_cmd64x(struct pci_dev *dev)
271*4882a593Smuzhiyun {
272*4882a593Smuzhiyun 	u8 mrdmode = 0;
273*4882a593Smuzhiyun 
274*4882a593Smuzhiyun 	/* Set a good latency timer and cache line size value. */
275*4882a593Smuzhiyun 	(void) pci_write_config_byte(dev, PCI_LATENCY_TIMER, 64);
276*4882a593Smuzhiyun 	/* FIXME: pci_set_master() to ensure a good latency timer value */
277*4882a593Smuzhiyun 
278*4882a593Smuzhiyun 	/*
279*4882a593Smuzhiyun 	 * Enable interrupts, select MEMORY READ LINE for reads.
280*4882a593Smuzhiyun 	 *
281*4882a593Smuzhiyun 	 * NOTE: although not mentioned in the PCI0646U specs,
282*4882a593Smuzhiyun 	 * bits 0-1 are write only and won't be read back as
283*4882a593Smuzhiyun 	 * set or not -- PCI0646U2 specs clarify this point.
284*4882a593Smuzhiyun 	 */
285*4882a593Smuzhiyun 	(void) pci_read_config_byte (dev, MRDMODE, &mrdmode);
286*4882a593Smuzhiyun 	mrdmode &= ~0x30;
287*4882a593Smuzhiyun 	(void) pci_write_config_byte(dev, MRDMODE, (mrdmode | 0x02));
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun 	return 0;
290*4882a593Smuzhiyun }
291*4882a593Smuzhiyun 
cmd64x_cable_detect(ide_hwif_t * hwif)292*4882a593Smuzhiyun static u8 cmd64x_cable_detect(ide_hwif_t *hwif)
293*4882a593Smuzhiyun {
294*4882a593Smuzhiyun 	struct pci_dev  *dev	= to_pci_dev(hwif->dev);
295*4882a593Smuzhiyun 	u8 bmidecsr = 0, mask	= hwif->channel ? 0x02 : 0x01;
296*4882a593Smuzhiyun 
297*4882a593Smuzhiyun 	switch (dev->device) {
298*4882a593Smuzhiyun 	case PCI_DEVICE_ID_CMD_648:
299*4882a593Smuzhiyun 	case PCI_DEVICE_ID_CMD_649:
300*4882a593Smuzhiyun  		pci_read_config_byte(dev, BMIDECSR, &bmidecsr);
301*4882a593Smuzhiyun 		return (bmidecsr & mask) ? ATA_CBL_PATA80 : ATA_CBL_PATA40;
302*4882a593Smuzhiyun 	default:
303*4882a593Smuzhiyun 		return ATA_CBL_PATA40;
304*4882a593Smuzhiyun 	}
305*4882a593Smuzhiyun }
306*4882a593Smuzhiyun 
307*4882a593Smuzhiyun static const struct ide_port_ops cmd64x_port_ops = {
308*4882a593Smuzhiyun 	.set_pio_mode		= cmd64x_set_pio_mode,
309*4882a593Smuzhiyun 	.set_dma_mode		= cmd64x_set_dma_mode,
310*4882a593Smuzhiyun 	.clear_irq		= cmd64x_clear_irq,
311*4882a593Smuzhiyun 	.test_irq		= cmd64x_test_irq,
312*4882a593Smuzhiyun 	.cable_detect		= cmd64x_cable_detect,
313*4882a593Smuzhiyun };
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun static const struct ide_port_ops cmd648_port_ops = {
316*4882a593Smuzhiyun 	.set_pio_mode		= cmd64x_set_pio_mode,
317*4882a593Smuzhiyun 	.set_dma_mode		= cmd64x_set_dma_mode,
318*4882a593Smuzhiyun 	.clear_irq		= cmd648_clear_irq,
319*4882a593Smuzhiyun 	.test_irq		= cmd648_test_irq,
320*4882a593Smuzhiyun 	.cable_detect		= cmd64x_cable_detect,
321*4882a593Smuzhiyun };
322*4882a593Smuzhiyun 
323*4882a593Smuzhiyun static const struct ide_dma_ops cmd646_rev1_dma_ops = {
324*4882a593Smuzhiyun 	.dma_host_set		= ide_dma_host_set,
325*4882a593Smuzhiyun 	.dma_setup		= ide_dma_setup,
326*4882a593Smuzhiyun 	.dma_start		= ide_dma_start,
327*4882a593Smuzhiyun 	.dma_end		= cmd646_1_dma_end,
328*4882a593Smuzhiyun 	.dma_test_irq		= ide_dma_test_irq,
329*4882a593Smuzhiyun 	.dma_lost_irq		= ide_dma_lost_irq,
330*4882a593Smuzhiyun 	.dma_timer_expiry	= ide_dma_sff_timer_expiry,
331*4882a593Smuzhiyun 	.dma_sff_read_status	= ide_dma_sff_read_status,
332*4882a593Smuzhiyun };
333*4882a593Smuzhiyun 
334*4882a593Smuzhiyun static const struct ide_port_info cmd64x_chipsets[] = {
335*4882a593Smuzhiyun 	{	/* 0: CMD643 */
336*4882a593Smuzhiyun 		.name		= DRV_NAME,
337*4882a593Smuzhiyun 		.init_chipset	= init_chipset_cmd64x,
338*4882a593Smuzhiyun 		.enablebits	= {{0x00,0x00,0x00}, {0x51,0x08,0x08}},
339*4882a593Smuzhiyun 		.port_ops	= &cmd64x_port_ops,
340*4882a593Smuzhiyun 		.host_flags	= IDE_HFLAG_CLEAR_SIMPLEX |
341*4882a593Smuzhiyun 				  IDE_HFLAG_ABUSE_PREFETCH |
342*4882a593Smuzhiyun 				  IDE_HFLAG_SERIALIZE,
343*4882a593Smuzhiyun 		.pio_mask	= ATA_PIO5,
344*4882a593Smuzhiyun 		.mwdma_mask	= ATA_MWDMA2,
345*4882a593Smuzhiyun 		.udma_mask	= 0x00, /* no udma */
346*4882a593Smuzhiyun 	},
347*4882a593Smuzhiyun 	{	/* 1: CMD646 */
348*4882a593Smuzhiyun 		.name		= DRV_NAME,
349*4882a593Smuzhiyun 		.init_chipset	= init_chipset_cmd64x,
350*4882a593Smuzhiyun 		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
351*4882a593Smuzhiyun 		.port_ops	= &cmd648_port_ops,
352*4882a593Smuzhiyun 		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH |
353*4882a593Smuzhiyun 				  IDE_HFLAG_SERIALIZE,
354*4882a593Smuzhiyun 		.pio_mask	= ATA_PIO5,
355*4882a593Smuzhiyun 		.mwdma_mask	= ATA_MWDMA2,
356*4882a593Smuzhiyun 		.udma_mask	= ATA_UDMA2,
357*4882a593Smuzhiyun 	},
358*4882a593Smuzhiyun 	{	/* 2: CMD648 */
359*4882a593Smuzhiyun 		.name		= DRV_NAME,
360*4882a593Smuzhiyun 		.init_chipset	= init_chipset_cmd64x,
361*4882a593Smuzhiyun 		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
362*4882a593Smuzhiyun 		.port_ops	= &cmd648_port_ops,
363*4882a593Smuzhiyun 		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH,
364*4882a593Smuzhiyun 		.pio_mask	= ATA_PIO5,
365*4882a593Smuzhiyun 		.mwdma_mask	= ATA_MWDMA2,
366*4882a593Smuzhiyun 		.udma_mask	= ATA_UDMA4,
367*4882a593Smuzhiyun 	},
368*4882a593Smuzhiyun 	{	/* 3: CMD649 */
369*4882a593Smuzhiyun 		.name		= DRV_NAME,
370*4882a593Smuzhiyun 		.init_chipset	= init_chipset_cmd64x,
371*4882a593Smuzhiyun 		.enablebits	= {{0x51,0x04,0x04}, {0x51,0x08,0x08}},
372*4882a593Smuzhiyun 		.port_ops	= &cmd648_port_ops,
373*4882a593Smuzhiyun 		.host_flags	= IDE_HFLAG_ABUSE_PREFETCH,
374*4882a593Smuzhiyun 		.pio_mask	= ATA_PIO5,
375*4882a593Smuzhiyun 		.mwdma_mask	= ATA_MWDMA2,
376*4882a593Smuzhiyun 		.udma_mask	= ATA_UDMA5,
377*4882a593Smuzhiyun 	}
378*4882a593Smuzhiyun };
379*4882a593Smuzhiyun 
cmd64x_init_one(struct pci_dev * dev,const struct pci_device_id * id)380*4882a593Smuzhiyun static int cmd64x_init_one(struct pci_dev *dev, const struct pci_device_id *id)
381*4882a593Smuzhiyun {
382*4882a593Smuzhiyun 	struct ide_port_info d;
383*4882a593Smuzhiyun 	u8 idx = id->driver_data;
384*4882a593Smuzhiyun 
385*4882a593Smuzhiyun 	d = cmd64x_chipsets[idx];
386*4882a593Smuzhiyun 
387*4882a593Smuzhiyun 	if (idx == 1) {
388*4882a593Smuzhiyun 		/*
389*4882a593Smuzhiyun 		 * UltraDMA only supported on PCI646U and PCI646U2, which
390*4882a593Smuzhiyun 		 * correspond to revisions 0x03, 0x05 and 0x07 respectively.
391*4882a593Smuzhiyun 		 * Actually, although the CMD tech support people won't
392*4882a593Smuzhiyun 		 * tell me the details, the 0x03 revision cannot support
393*4882a593Smuzhiyun 		 * UDMA correctly without hardware modifications, and even
394*4882a593Smuzhiyun 		 * then it only works with Quantum disks due to some
395*4882a593Smuzhiyun 		 * hold time assumptions in the 646U part which are fixed
396*4882a593Smuzhiyun 		 * in the 646U2.
397*4882a593Smuzhiyun 		 *
398*4882a593Smuzhiyun 		 * So we only do UltraDMA on revision 0x05 and 0x07 chipsets.
399*4882a593Smuzhiyun 		 */
400*4882a593Smuzhiyun 		if (dev->revision < 5) {
401*4882a593Smuzhiyun 			d.udma_mask = 0x00;
402*4882a593Smuzhiyun 			/*
403*4882a593Smuzhiyun 			 * The original PCI0646 didn't have the primary
404*4882a593Smuzhiyun 			 * channel enable bit, it appeared starting with
405*4882a593Smuzhiyun 			 * PCI0646U (i.e. revision ID 3).
406*4882a593Smuzhiyun 			 */
407*4882a593Smuzhiyun 			if (dev->revision < 3) {
408*4882a593Smuzhiyun 				d.enablebits[0].reg = 0;
409*4882a593Smuzhiyun 				d.port_ops = &cmd64x_port_ops;
410*4882a593Smuzhiyun 				if (dev->revision == 1)
411*4882a593Smuzhiyun 					d.dma_ops = &cmd646_rev1_dma_ops;
412*4882a593Smuzhiyun 			}
413*4882a593Smuzhiyun 		}
414*4882a593Smuzhiyun 	}
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun 	return ide_pci_init_one(dev, &d, NULL);
417*4882a593Smuzhiyun }
418*4882a593Smuzhiyun 
419*4882a593Smuzhiyun static const struct pci_device_id cmd64x_pci_tbl[] = {
420*4882a593Smuzhiyun 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_643), 0 },
421*4882a593Smuzhiyun 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_646), 1 },
422*4882a593Smuzhiyun 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_648), 2 },
423*4882a593Smuzhiyun 	{ PCI_VDEVICE(CMD, PCI_DEVICE_ID_CMD_649), 3 },
424*4882a593Smuzhiyun 	{ 0, },
425*4882a593Smuzhiyun };
426*4882a593Smuzhiyun MODULE_DEVICE_TABLE(pci, cmd64x_pci_tbl);
427*4882a593Smuzhiyun 
428*4882a593Smuzhiyun static struct pci_driver cmd64x_pci_driver = {
429*4882a593Smuzhiyun 	.name		= "CMD64x_IDE",
430*4882a593Smuzhiyun 	.id_table	= cmd64x_pci_tbl,
431*4882a593Smuzhiyun 	.probe		= cmd64x_init_one,
432*4882a593Smuzhiyun 	.remove		= ide_pci_remove,
433*4882a593Smuzhiyun 	.suspend	= ide_pci_suspend,
434*4882a593Smuzhiyun 	.resume		= ide_pci_resume,
435*4882a593Smuzhiyun };
436*4882a593Smuzhiyun 
cmd64x_ide_init(void)437*4882a593Smuzhiyun static int __init cmd64x_ide_init(void)
438*4882a593Smuzhiyun {
439*4882a593Smuzhiyun 	return ide_pci_register_driver(&cmd64x_pci_driver);
440*4882a593Smuzhiyun }
441*4882a593Smuzhiyun 
cmd64x_ide_exit(void)442*4882a593Smuzhiyun static void __exit cmd64x_ide_exit(void)
443*4882a593Smuzhiyun {
444*4882a593Smuzhiyun 	pci_unregister_driver(&cmd64x_pci_driver);
445*4882a593Smuzhiyun }
446*4882a593Smuzhiyun 
447*4882a593Smuzhiyun module_init(cmd64x_ide_init);
448*4882a593Smuzhiyun module_exit(cmd64x_ide_exit);
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun MODULE_AUTHOR("Eddie Dost, David Miller, Andre Hedrick, Bartlomiej Zolnierkiewicz");
451*4882a593Smuzhiyun MODULE_DESCRIPTION("PCI driver module for CMD64x IDE");
452*4882a593Smuzhiyun MODULE_LICENSE("GPL");
453