xref: /OK3568_Linux_fs/kernel/arch/m68k/include/asm/dvma.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0 */
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * include/asm-m68k/dma.h
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Copyright 1995 (C) David S. Miller (davem@caip.rutgers.edu)
6*4882a593Smuzhiyun  *
7*4882a593Smuzhiyun  * Hacked to fit Sun3x needs by Thomas Bogendoerfer
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun 
10*4882a593Smuzhiyun #ifndef __M68K_DVMA_H
11*4882a593Smuzhiyun #define __M68K_DVMA_H
12*4882a593Smuzhiyun 
13*4882a593Smuzhiyun 
14*4882a593Smuzhiyun #define DVMA_PAGE_SHIFT	13
15*4882a593Smuzhiyun #define DVMA_PAGE_SIZE	(1UL << DVMA_PAGE_SHIFT)
16*4882a593Smuzhiyun #define DVMA_PAGE_MASK	(~(DVMA_PAGE_SIZE-1))
17*4882a593Smuzhiyun #define DVMA_PAGE_ALIGN(addr)	ALIGN(addr, DVMA_PAGE_SIZE)
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun extern void dvma_init(void);
20*4882a593Smuzhiyun extern int dvma_map_iommu(unsigned long kaddr, unsigned long baddr,
21*4882a593Smuzhiyun 			  int len);
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun #define dvma_malloc(x) dvma_malloc_align(x, 0)
24*4882a593Smuzhiyun #define dvma_map(x, y) dvma_map_align(x, y, 0)
25*4882a593Smuzhiyun #define dvma_map_vme(x, y) (dvma_map(x, y) & 0xfffff)
26*4882a593Smuzhiyun #define dvma_map_align_vme(x, y, z) (dvma_map_align (x, y, z) & 0xfffff)
27*4882a593Smuzhiyun extern unsigned long dvma_map_align(unsigned long kaddr, int len,
28*4882a593Smuzhiyun 			    int align);
29*4882a593Smuzhiyun extern void *dvma_malloc_align(unsigned long len, unsigned long align);
30*4882a593Smuzhiyun 
31*4882a593Smuzhiyun extern void dvma_unmap(void *baddr);
32*4882a593Smuzhiyun extern void dvma_free(void *vaddr);
33*4882a593Smuzhiyun 
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun #ifdef CONFIG_SUN3
36*4882a593Smuzhiyun /* sun3 dvma page support */
37*4882a593Smuzhiyun 
38*4882a593Smuzhiyun /* memory and pmegs potentially reserved for dvma */
39*4882a593Smuzhiyun #define DVMA_PMEG_START 10
40*4882a593Smuzhiyun #define DVMA_PMEG_END 16
41*4882a593Smuzhiyun #define DVMA_START 0xf00000
42*4882a593Smuzhiyun #define DVMA_END 0xfe0000
43*4882a593Smuzhiyun #define DVMA_SIZE (DVMA_END-DVMA_START)
44*4882a593Smuzhiyun #define IOMMU_TOTAL_ENTRIES 128
45*4882a593Smuzhiyun #define IOMMU_ENTRIES 120
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun /* empirical kludge -- dvma regions only seem to work right on 0x10000
48*4882a593Smuzhiyun    byte boundaries */
49*4882a593Smuzhiyun #define DVMA_REGION_SIZE 0x10000
50*4882a593Smuzhiyun #define DVMA_ALIGN(addr) (((addr)+DVMA_REGION_SIZE-1) & \
51*4882a593Smuzhiyun                          ~(DVMA_REGION_SIZE-1))
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun /* virt <-> phys conversions */
54*4882a593Smuzhiyun #define dvma_vtop(x) ((unsigned long)(x) & 0xffffff)
55*4882a593Smuzhiyun #define dvma_ptov(x) ((unsigned long)(x) | 0xf000000)
56*4882a593Smuzhiyun #define dvma_vtovme(x) ((unsigned long)(x) & 0x00fffff)
57*4882a593Smuzhiyun #define dvma_vmetov(x) ((unsigned long)(x) | 0xff00000)
58*4882a593Smuzhiyun #define dvma_vtob(x) dvma_vtop(x)
59*4882a593Smuzhiyun #define dvma_btov(x) dvma_ptov(x)
60*4882a593Smuzhiyun 
dvma_map_cpu(unsigned long kaddr,unsigned long vaddr,int len)61*4882a593Smuzhiyun static inline int dvma_map_cpu(unsigned long kaddr, unsigned long vaddr,
62*4882a593Smuzhiyun 			       int len)
63*4882a593Smuzhiyun {
64*4882a593Smuzhiyun 	return 0;
65*4882a593Smuzhiyun }
66*4882a593Smuzhiyun 
67*4882a593Smuzhiyun #else /* Sun3x */
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun /* sun3x dvma page support */
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun #define DVMA_START 0x0
72*4882a593Smuzhiyun #define DVMA_END 0xf00000
73*4882a593Smuzhiyun #define DVMA_SIZE (DVMA_END-DVMA_START)
74*4882a593Smuzhiyun #define IOMMU_TOTAL_ENTRIES	   2048
75*4882a593Smuzhiyun /* the prom takes the top meg */
76*4882a593Smuzhiyun #define IOMMU_ENTRIES              (IOMMU_TOTAL_ENTRIES - 0x80)
77*4882a593Smuzhiyun 
78*4882a593Smuzhiyun #define dvma_vtob(x) ((unsigned long)(x) & 0x00ffffff)
79*4882a593Smuzhiyun #define dvma_btov(x) ((unsigned long)(x) | 0xff000000)
80*4882a593Smuzhiyun 
81*4882a593Smuzhiyun extern int dvma_map_cpu(unsigned long kaddr, unsigned long vaddr, int len);
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun 
84*4882a593Smuzhiyun 
85*4882a593Smuzhiyun /* everything below this line is specific to dma used for the onboard
86*4882a593Smuzhiyun    ESP scsi on sun3x */
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun /* Structure to describe the current status of DMA registers on the Sparc */
89*4882a593Smuzhiyun struct sparc_dma_registers {
90*4882a593Smuzhiyun   __volatile__ unsigned long cond_reg;	/* DMA condition register */
91*4882a593Smuzhiyun   __volatile__ unsigned long st_addr;	/* Start address of this transfer */
92*4882a593Smuzhiyun   __volatile__ unsigned long  cnt;	/* How many bytes to transfer */
93*4882a593Smuzhiyun   __volatile__ unsigned long dma_test;	/* DMA test register */
94*4882a593Smuzhiyun };
95*4882a593Smuzhiyun 
96*4882a593Smuzhiyun /* DVMA chip revisions */
97*4882a593Smuzhiyun enum dvma_rev {
98*4882a593Smuzhiyun 	dvmarev0,
99*4882a593Smuzhiyun 	dvmaesc1,
100*4882a593Smuzhiyun 	dvmarev1,
101*4882a593Smuzhiyun 	dvmarev2,
102*4882a593Smuzhiyun 	dvmarev3,
103*4882a593Smuzhiyun 	dvmarevplus,
104*4882a593Smuzhiyun 	dvmahme
105*4882a593Smuzhiyun };
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun #define DMA_HASCOUNT(rev)  ((rev)==dvmaesc1)
108*4882a593Smuzhiyun 
109*4882a593Smuzhiyun /* Linux DMA information structure, filled during probe. */
110*4882a593Smuzhiyun struct Linux_SBus_DMA {
111*4882a593Smuzhiyun 	struct Linux_SBus_DMA *next;
112*4882a593Smuzhiyun 	struct linux_sbus_device *SBus_dev;
113*4882a593Smuzhiyun 	struct sparc_dma_registers *regs;
114*4882a593Smuzhiyun 
115*4882a593Smuzhiyun 	/* Status, misc info */
116*4882a593Smuzhiyun 	int node;                /* Prom node for this DMA device */
117*4882a593Smuzhiyun 	int running;             /* Are we doing DMA now? */
118*4882a593Smuzhiyun 	int allocated;           /* Are we "owned" by anyone yet? */
119*4882a593Smuzhiyun 
120*4882a593Smuzhiyun 	/* Transfer information. */
121*4882a593Smuzhiyun 	unsigned long addr;      /* Start address of current transfer */
122*4882a593Smuzhiyun 	int nbytes;              /* Size of current transfer */
123*4882a593Smuzhiyun 	int realbytes;           /* For splitting up large transfers, etc. */
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 	/* DMA revision */
126*4882a593Smuzhiyun 	enum dvma_rev revision;
127*4882a593Smuzhiyun };
128*4882a593Smuzhiyun 
129*4882a593Smuzhiyun extern struct Linux_SBus_DMA *dma_chain;
130*4882a593Smuzhiyun 
131*4882a593Smuzhiyun /* Broken hardware... */
132*4882a593Smuzhiyun #define DMA_ISBROKEN(dma)    ((dma)->revision == dvmarev1)
133*4882a593Smuzhiyun #define DMA_ISESC1(dma)      ((dma)->revision == dvmaesc1)
134*4882a593Smuzhiyun 
135*4882a593Smuzhiyun /* Fields in the cond_reg register */
136*4882a593Smuzhiyun /* First, the version identification bits */
137*4882a593Smuzhiyun #define DMA_DEVICE_ID    0xf0000000        /* Device identification bits */
138*4882a593Smuzhiyun #define DMA_VERS0        0x00000000        /* Sunray DMA version */
139*4882a593Smuzhiyun #define DMA_ESCV1        0x40000000        /* DMA ESC Version 1 */
140*4882a593Smuzhiyun #define DMA_VERS1        0x80000000        /* DMA rev 1 */
141*4882a593Smuzhiyun #define DMA_VERS2        0xa0000000        /* DMA rev 2 */
142*4882a593Smuzhiyun #define DMA_VERHME       0xb0000000        /* DMA hme gate array */
143*4882a593Smuzhiyun #define DMA_VERSPLUS     0x90000000        /* DMA rev 1 PLUS */
144*4882a593Smuzhiyun 
145*4882a593Smuzhiyun #define DMA_HNDL_INTR    0x00000001        /* An IRQ needs to be handled */
146*4882a593Smuzhiyun #define DMA_HNDL_ERROR   0x00000002        /* We need to take an error */
147*4882a593Smuzhiyun #define DMA_FIFO_ISDRAIN 0x0000000c        /* The DMA FIFO is draining */
148*4882a593Smuzhiyun #define DMA_INT_ENAB     0x00000010        /* Turn on interrupts */
149*4882a593Smuzhiyun #define DMA_FIFO_INV     0x00000020        /* Invalidate the FIFO */
150*4882a593Smuzhiyun #define DMA_ACC_SZ_ERR   0x00000040        /* The access size was bad */
151*4882a593Smuzhiyun #define DMA_FIFO_STDRAIN 0x00000040        /* DMA_VERS1 Drain the FIFO */
152*4882a593Smuzhiyun #define DMA_RST_SCSI     0x00000080        /* Reset the SCSI controller */
153*4882a593Smuzhiyun #define DMA_RST_ENET     DMA_RST_SCSI      /* Reset the ENET controller */
154*4882a593Smuzhiyun #define DMA_ST_WRITE     0x00000100        /* write from device to memory */
155*4882a593Smuzhiyun #define DMA_ENABLE       0x00000200        /* Fire up DMA, handle requests */
156*4882a593Smuzhiyun #define DMA_PEND_READ    0x00000400        /* DMA_VERS1/0/PLUS Pending Read */
157*4882a593Smuzhiyun #define DMA_ESC_BURST    0x00000800        /* 1=16byte 0=32byte */
158*4882a593Smuzhiyun #define DMA_READ_AHEAD   0x00001800        /* DMA read ahead partial longword */
159*4882a593Smuzhiyun #define DMA_DSBL_RD_DRN  0x00001000        /* No EC drain on slave reads */
160*4882a593Smuzhiyun #define DMA_BCNT_ENAB    0x00002000        /* If on, use the byte counter */
161*4882a593Smuzhiyun #define DMA_TERM_CNTR    0x00004000        /* Terminal counter */
162*4882a593Smuzhiyun #define DMA_CSR_DISAB    0x00010000        /* No FIFO drains during csr */
163*4882a593Smuzhiyun #define DMA_SCSI_DISAB   0x00020000        /* No FIFO drains during reg */
164*4882a593Smuzhiyun #define DMA_DSBL_WR_INV  0x00020000        /* No EC inval. on slave writes */
165*4882a593Smuzhiyun #define DMA_ADD_ENABLE   0x00040000        /* Special ESC DVMA optimization */
166*4882a593Smuzhiyun #define DMA_E_BURST8	 0x00040000	   /* ENET: SBUS r/w burst size */
167*4882a593Smuzhiyun #define DMA_BRST_SZ      0x000c0000        /* SCSI: SBUS r/w burst size */
168*4882a593Smuzhiyun #define DMA_BRST64       0x00080000        /* SCSI: 64byte bursts (HME on UltraSparc only) */
169*4882a593Smuzhiyun #define DMA_BRST32       0x00040000        /* SCSI: 32byte bursts */
170*4882a593Smuzhiyun #define DMA_BRST16       0x00000000        /* SCSI: 16byte bursts */
171*4882a593Smuzhiyun #define DMA_BRST0        0x00080000        /* SCSI: no bursts (non-HME gate arrays) */
172*4882a593Smuzhiyun #define DMA_ADDR_DISAB   0x00100000        /* No FIFO drains during addr */
173*4882a593Smuzhiyun #define DMA_2CLKS        0x00200000        /* Each transfer = 2 clock ticks */
174*4882a593Smuzhiyun #define DMA_3CLKS        0x00400000        /* Each transfer = 3 clock ticks */
175*4882a593Smuzhiyun #define DMA_EN_ENETAUI   DMA_3CLKS         /* Put lance into AUI-cable mode */
176*4882a593Smuzhiyun #define DMA_CNTR_DISAB   0x00800000        /* No IRQ when DMA_TERM_CNTR set */
177*4882a593Smuzhiyun #define DMA_AUTO_NADDR   0x01000000        /* Use "auto nxt addr" feature */
178*4882a593Smuzhiyun #define DMA_SCSI_ON      0x02000000        /* Enable SCSI dma */
179*4882a593Smuzhiyun #define DMA_PARITY_OFF   0x02000000        /* HME: disable parity checking */
180*4882a593Smuzhiyun #define DMA_LOADED_ADDR  0x04000000        /* Address has been loaded */
181*4882a593Smuzhiyun #define DMA_LOADED_NADDR 0x08000000        /* Next address has been loaded */
182*4882a593Smuzhiyun 
183*4882a593Smuzhiyun /* Values describing the burst-size property from the PROM */
184*4882a593Smuzhiyun #define DMA_BURST1       0x01
185*4882a593Smuzhiyun #define DMA_BURST2       0x02
186*4882a593Smuzhiyun #define DMA_BURST4       0x04
187*4882a593Smuzhiyun #define DMA_BURST8       0x08
188*4882a593Smuzhiyun #define DMA_BURST16      0x10
189*4882a593Smuzhiyun #define DMA_BURST32      0x20
190*4882a593Smuzhiyun #define DMA_BURST64      0x40
191*4882a593Smuzhiyun #define DMA_BURSTBITS    0x7f
192*4882a593Smuzhiyun 
193*4882a593Smuzhiyun /* Determine highest possible final transfer address given a base */
194*4882a593Smuzhiyun #define DMA_MAXEND(addr) (0x01000000UL-(((unsigned long)(addr))&0x00ffffffUL))
195*4882a593Smuzhiyun 
196*4882a593Smuzhiyun /* Yes, I hack a lot of elisp in my spare time... */
197*4882a593Smuzhiyun #define DMA_ERROR_P(regs)  ((((regs)->cond_reg) & DMA_HNDL_ERROR))
198*4882a593Smuzhiyun #define DMA_IRQ_P(regs)    ((((regs)->cond_reg) & (DMA_HNDL_INTR | DMA_HNDL_ERROR)))
199*4882a593Smuzhiyun #define DMA_WRITE_P(regs)  ((((regs)->cond_reg) & DMA_ST_WRITE))
200*4882a593Smuzhiyun #define DMA_OFF(regs)      ((((regs)->cond_reg) &= (~DMA_ENABLE)))
201*4882a593Smuzhiyun #define DMA_INTSOFF(regs)  ((((regs)->cond_reg) &= (~DMA_INT_ENAB)))
202*4882a593Smuzhiyun #define DMA_INTSON(regs)   ((((regs)->cond_reg) |= (DMA_INT_ENAB)))
203*4882a593Smuzhiyun #define DMA_PUNTFIFO(regs) ((((regs)->cond_reg) |= DMA_FIFO_INV))
204*4882a593Smuzhiyun #define DMA_SETSTART(regs, addr)  ((((regs)->st_addr) = (char *) addr))
205*4882a593Smuzhiyun #define DMA_BEGINDMA_W(regs) \
206*4882a593Smuzhiyun         ((((regs)->cond_reg |= (DMA_ST_WRITE|DMA_ENABLE|DMA_INT_ENAB))))
207*4882a593Smuzhiyun #define DMA_BEGINDMA_R(regs) \
208*4882a593Smuzhiyun         ((((regs)->cond_reg |= ((DMA_ENABLE|DMA_INT_ENAB)&(~DMA_ST_WRITE)))))
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun /* For certain DMA chips, we need to disable ints upon irq entry
211*4882a593Smuzhiyun  * and turn them back on when we are done.  So in any ESP interrupt
212*4882a593Smuzhiyun  * handler you *must* call DMA_IRQ_ENTRY upon entry and DMA_IRQ_EXIT
213*4882a593Smuzhiyun  * when leaving the handler.  You have been warned...
214*4882a593Smuzhiyun  */
215*4882a593Smuzhiyun #define DMA_IRQ_ENTRY(dma, dregs) do { \
216*4882a593Smuzhiyun         if(DMA_ISBROKEN(dma)) DMA_INTSOFF(dregs); \
217*4882a593Smuzhiyun    } while (0)
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun #define DMA_IRQ_EXIT(dma, dregs) do { \
220*4882a593Smuzhiyun 	if(DMA_ISBROKEN(dma)) DMA_INTSON(dregs); \
221*4882a593Smuzhiyun    } while(0)
222*4882a593Smuzhiyun 
223*4882a593Smuzhiyun /* Reset the friggin' thing... */
224*4882a593Smuzhiyun #define DMA_RESET(dma) do { \
225*4882a593Smuzhiyun 	struct sparc_dma_registers *regs = dma->regs;                      \
226*4882a593Smuzhiyun 	/* Let the current FIFO drain itself */                            \
227*4882a593Smuzhiyun 	sparc_dma_pause(regs, (DMA_FIFO_ISDRAIN));                         \
228*4882a593Smuzhiyun 	/* Reset the logic */                                              \
229*4882a593Smuzhiyun 	regs->cond_reg |= (DMA_RST_SCSI);     /* assert */                 \
230*4882a593Smuzhiyun 	__delay(400);                         /* let the bits set ;) */    \
231*4882a593Smuzhiyun 	regs->cond_reg &= ~(DMA_RST_SCSI);    /* de-assert */              \
232*4882a593Smuzhiyun 	sparc_dma_enable_interrupts(regs);    /* Re-enable interrupts */   \
233*4882a593Smuzhiyun 	/* Enable FAST transfers if available */                           \
234*4882a593Smuzhiyun 	if(dma->revision>dvmarev1) regs->cond_reg |= DMA_3CLKS;            \
235*4882a593Smuzhiyun 	dma->running = 0;                                                  \
236*4882a593Smuzhiyun } while(0)
237*4882a593Smuzhiyun 
238*4882a593Smuzhiyun 
239*4882a593Smuzhiyun #endif /* !CONFIG_SUN3 */
240*4882a593Smuzhiyun 
241*4882a593Smuzhiyun #endif /* !(__M68K_DVMA_H) */
242