xref: /OK3568_Linux_fs/kernel/drivers/dma/dw/regs.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0 */
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Driver for the Synopsys DesignWare AHB DMA Controller
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Copyright (C) 2005-2007 Atmel Corporation
6*4882a593Smuzhiyun  * Copyright (C) 2010-2011 ST Microelectronics
7*4882a593Smuzhiyun  * Copyright (C) 2016 Intel Corporation
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun 
10*4882a593Smuzhiyun #include <linux/bitops.h>
11*4882a593Smuzhiyun #include <linux/interrupt.h>
12*4882a593Smuzhiyun #include <linux/dmaengine.h>
13*4882a593Smuzhiyun 
14*4882a593Smuzhiyun #include <linux/io-64-nonatomic-hi-lo.h>
15*4882a593Smuzhiyun 
16*4882a593Smuzhiyun #include "internal.h"
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun #define DW_DMA_MAX_NR_REQUESTS	16
19*4882a593Smuzhiyun 
20*4882a593Smuzhiyun /* flow controller */
21*4882a593Smuzhiyun enum dw_dma_fc {
22*4882a593Smuzhiyun 	DW_DMA_FC_D_M2M,
23*4882a593Smuzhiyun 	DW_DMA_FC_D_M2P,
24*4882a593Smuzhiyun 	DW_DMA_FC_D_P2M,
25*4882a593Smuzhiyun 	DW_DMA_FC_D_P2P,
26*4882a593Smuzhiyun 	DW_DMA_FC_P_P2M,
27*4882a593Smuzhiyun 	DW_DMA_FC_SP_P2P,
28*4882a593Smuzhiyun 	DW_DMA_FC_P_M2P,
29*4882a593Smuzhiyun 	DW_DMA_FC_DP_P2P,
30*4882a593Smuzhiyun };
31*4882a593Smuzhiyun 
32*4882a593Smuzhiyun /*
33*4882a593Smuzhiyun  * Redefine this macro to handle differences between 32- and 64-bit
34*4882a593Smuzhiyun  * addressing, big vs. little endian, etc.
35*4882a593Smuzhiyun  */
36*4882a593Smuzhiyun #define DW_REG(name)		u32 name; u32 __pad_##name
37*4882a593Smuzhiyun 
38*4882a593Smuzhiyun /* Hardware register definitions. */
39*4882a593Smuzhiyun struct dw_dma_chan_regs {
40*4882a593Smuzhiyun 	DW_REG(SAR);		/* Source Address Register */
41*4882a593Smuzhiyun 	DW_REG(DAR);		/* Destination Address Register */
42*4882a593Smuzhiyun 	DW_REG(LLP);		/* Linked List Pointer */
43*4882a593Smuzhiyun 	u32	CTL_LO;		/* Control Register Low */
44*4882a593Smuzhiyun 	u32	CTL_HI;		/* Control Register High */
45*4882a593Smuzhiyun 	DW_REG(SSTAT);
46*4882a593Smuzhiyun 	DW_REG(DSTAT);
47*4882a593Smuzhiyun 	DW_REG(SSTATAR);
48*4882a593Smuzhiyun 	DW_REG(DSTATAR);
49*4882a593Smuzhiyun 	u32	CFG_LO;		/* Configuration Register Low */
50*4882a593Smuzhiyun 	u32	CFG_HI;		/* Configuration Register High */
51*4882a593Smuzhiyun 	DW_REG(SGR);
52*4882a593Smuzhiyun 	DW_REG(DSR);
53*4882a593Smuzhiyun };
54*4882a593Smuzhiyun 
55*4882a593Smuzhiyun struct dw_dma_irq_regs {
56*4882a593Smuzhiyun 	DW_REG(XFER);
57*4882a593Smuzhiyun 	DW_REG(BLOCK);
58*4882a593Smuzhiyun 	DW_REG(SRC_TRAN);
59*4882a593Smuzhiyun 	DW_REG(DST_TRAN);
60*4882a593Smuzhiyun 	DW_REG(ERROR);
61*4882a593Smuzhiyun };
62*4882a593Smuzhiyun 
63*4882a593Smuzhiyun struct dw_dma_regs {
64*4882a593Smuzhiyun 	/* per-channel registers */
65*4882a593Smuzhiyun 	struct dw_dma_chan_regs	CHAN[DW_DMA_MAX_NR_CHANNELS];
66*4882a593Smuzhiyun 
67*4882a593Smuzhiyun 	/* irq handling */
68*4882a593Smuzhiyun 	struct dw_dma_irq_regs	RAW;		/* r */
69*4882a593Smuzhiyun 	struct dw_dma_irq_regs	STATUS;		/* r (raw & mask) */
70*4882a593Smuzhiyun 	struct dw_dma_irq_regs	MASK;		/* rw (set = irq enabled) */
71*4882a593Smuzhiyun 	struct dw_dma_irq_regs	CLEAR;		/* w (ack, affects "raw") */
72*4882a593Smuzhiyun 
73*4882a593Smuzhiyun 	DW_REG(STATUS_INT);			/* r */
74*4882a593Smuzhiyun 
75*4882a593Smuzhiyun 	/* software handshaking */
76*4882a593Smuzhiyun 	DW_REG(REQ_SRC);
77*4882a593Smuzhiyun 	DW_REG(REQ_DST);
78*4882a593Smuzhiyun 	DW_REG(SGL_REQ_SRC);
79*4882a593Smuzhiyun 	DW_REG(SGL_REQ_DST);
80*4882a593Smuzhiyun 	DW_REG(LAST_SRC);
81*4882a593Smuzhiyun 	DW_REG(LAST_DST);
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun 	/* miscellaneous */
84*4882a593Smuzhiyun 	DW_REG(CFG);
85*4882a593Smuzhiyun 	DW_REG(CH_EN);
86*4882a593Smuzhiyun 	DW_REG(ID);
87*4882a593Smuzhiyun 	DW_REG(TEST);
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun 	/* iDMA 32-bit support */
90*4882a593Smuzhiyun 	DW_REG(CLASS_PRIORITY0);
91*4882a593Smuzhiyun 	DW_REG(CLASS_PRIORITY1);
92*4882a593Smuzhiyun 
93*4882a593Smuzhiyun 	/* optional encoded params, 0x3c8..0x3f7 */
94*4882a593Smuzhiyun 	u32	__reserved;
95*4882a593Smuzhiyun 
96*4882a593Smuzhiyun 	/* per-channel configuration registers */
97*4882a593Smuzhiyun 	u32	DWC_PARAMS[DW_DMA_MAX_NR_CHANNELS];
98*4882a593Smuzhiyun 	u32	MULTI_BLK_TYPE;
99*4882a593Smuzhiyun 	u32	MAX_BLK_SIZE;
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun 	/* top-level parameters */
102*4882a593Smuzhiyun 	u32	DW_PARAMS;
103*4882a593Smuzhiyun 
104*4882a593Smuzhiyun 	/* component ID */
105*4882a593Smuzhiyun 	u32	COMP_TYPE;
106*4882a593Smuzhiyun 	u32	COMP_VERSION;
107*4882a593Smuzhiyun 
108*4882a593Smuzhiyun 	/* iDMA 32-bit support */
109*4882a593Smuzhiyun 	DW_REG(FIFO_PARTITION0);
110*4882a593Smuzhiyun 	DW_REG(FIFO_PARTITION1);
111*4882a593Smuzhiyun 
112*4882a593Smuzhiyun 	DW_REG(SAI_ERR);
113*4882a593Smuzhiyun 	DW_REG(GLOBAL_CFG);
114*4882a593Smuzhiyun };
115*4882a593Smuzhiyun 
116*4882a593Smuzhiyun /* Bitfields in DW_PARAMS */
117*4882a593Smuzhiyun #define DW_PARAMS_NR_CHAN	8		/* number of channels */
118*4882a593Smuzhiyun #define DW_PARAMS_NR_MASTER	11		/* number of AHB masters */
119*4882a593Smuzhiyun #define DW_PARAMS_DATA_WIDTH(n)	(15 + 2 * (n))
120*4882a593Smuzhiyun #define DW_PARAMS_DATA_WIDTH1	15		/* master 1 data width */
121*4882a593Smuzhiyun #define DW_PARAMS_DATA_WIDTH2	17		/* master 2 data width */
122*4882a593Smuzhiyun #define DW_PARAMS_DATA_WIDTH3	19		/* master 3 data width */
123*4882a593Smuzhiyun #define DW_PARAMS_DATA_WIDTH4	21		/* master 4 data width */
124*4882a593Smuzhiyun #define DW_PARAMS_EN		28		/* encoded parameters */
125*4882a593Smuzhiyun 
126*4882a593Smuzhiyun /* Bitfields in DWC_PARAMS */
127*4882a593Smuzhiyun #define DWC_PARAMS_MBLK_EN	11		/* multi block transfer */
128*4882a593Smuzhiyun #define DWC_PARAMS_HC_LLP	13		/* set LLP register to zero */
129*4882a593Smuzhiyun #define DWC_PARAMS_MSIZE	16		/* max group transaction size */
130*4882a593Smuzhiyun 
131*4882a593Smuzhiyun /* bursts size */
132*4882a593Smuzhiyun enum dw_dma_msize {
133*4882a593Smuzhiyun 	DW_DMA_MSIZE_1,
134*4882a593Smuzhiyun 	DW_DMA_MSIZE_4,
135*4882a593Smuzhiyun 	DW_DMA_MSIZE_8,
136*4882a593Smuzhiyun 	DW_DMA_MSIZE_16,
137*4882a593Smuzhiyun 	DW_DMA_MSIZE_32,
138*4882a593Smuzhiyun 	DW_DMA_MSIZE_64,
139*4882a593Smuzhiyun 	DW_DMA_MSIZE_128,
140*4882a593Smuzhiyun 	DW_DMA_MSIZE_256,
141*4882a593Smuzhiyun };
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun /* Bitfields in LLP */
144*4882a593Smuzhiyun #define DWC_LLP_LMS(x)		((x) & 3)	/* list master select */
145*4882a593Smuzhiyun #define DWC_LLP_LOC(x)		((x) & ~3)	/* next lli */
146*4882a593Smuzhiyun 
147*4882a593Smuzhiyun /* Bitfields in CTL_LO */
148*4882a593Smuzhiyun #define DWC_CTLL_INT_EN		(1 << 0)	/* irqs enabled? */
149*4882a593Smuzhiyun #define DWC_CTLL_DST_WIDTH(n)	((n)<<1)	/* bytes per element */
150*4882a593Smuzhiyun #define DWC_CTLL_SRC_WIDTH(n)	((n)<<4)
151*4882a593Smuzhiyun #define DWC_CTLL_DST_INC	(0<<7)		/* DAR update/not */
152*4882a593Smuzhiyun #define DWC_CTLL_DST_DEC	(1<<7)
153*4882a593Smuzhiyun #define DWC_CTLL_DST_FIX	(2<<7)
154*4882a593Smuzhiyun #define DWC_CTLL_SRC_INC	(0<<9)		/* SAR update/not */
155*4882a593Smuzhiyun #define DWC_CTLL_SRC_DEC	(1<<9)
156*4882a593Smuzhiyun #define DWC_CTLL_SRC_FIX	(2<<9)
157*4882a593Smuzhiyun #define DWC_CTLL_DST_MSIZE(n)	((n)<<11)	/* burst, #elements */
158*4882a593Smuzhiyun #define DWC_CTLL_SRC_MSIZE(n)	((n)<<14)
159*4882a593Smuzhiyun #define DWC_CTLL_S_GATH_EN	(1 << 17)	/* src gather, !FIX */
160*4882a593Smuzhiyun #define DWC_CTLL_D_SCAT_EN	(1 << 18)	/* dst scatter, !FIX */
161*4882a593Smuzhiyun #define DWC_CTLL_FC(n)		((n) << 20)
162*4882a593Smuzhiyun #define DWC_CTLL_FC_M2M		(0 << 20)	/* mem-to-mem */
163*4882a593Smuzhiyun #define DWC_CTLL_FC_M2P		(1 << 20)	/* mem-to-periph */
164*4882a593Smuzhiyun #define DWC_CTLL_FC_P2M		(2 << 20)	/* periph-to-mem */
165*4882a593Smuzhiyun #define DWC_CTLL_FC_P2P		(3 << 20)	/* periph-to-periph */
166*4882a593Smuzhiyun /* plus 4 transfer types for peripheral-as-flow-controller */
167*4882a593Smuzhiyun #define DWC_CTLL_DMS(n)		((n)<<23)	/* dst master select */
168*4882a593Smuzhiyun #define DWC_CTLL_SMS(n)		((n)<<25)	/* src master select */
169*4882a593Smuzhiyun #define DWC_CTLL_LLP_D_EN	(1 << 27)	/* dest block chain */
170*4882a593Smuzhiyun #define DWC_CTLL_LLP_S_EN	(1 << 28)	/* src block chain */
171*4882a593Smuzhiyun 
172*4882a593Smuzhiyun /* Bitfields in CTL_HI */
173*4882a593Smuzhiyun #define DWC_CTLH_BLOCK_TS_MASK	GENMASK(11, 0)
174*4882a593Smuzhiyun #define DWC_CTLH_BLOCK_TS(x)	((x) & DWC_CTLH_BLOCK_TS_MASK)
175*4882a593Smuzhiyun #define DWC_CTLH_DONE		(1 << 12)
176*4882a593Smuzhiyun 
177*4882a593Smuzhiyun /* Bitfields in CFG_LO */
178*4882a593Smuzhiyun #define DWC_CFGL_CH_PRIOR_MASK	(0x7 << 5)	/* priority mask */
179*4882a593Smuzhiyun #define DWC_CFGL_CH_PRIOR(x)	((x) << 5)	/* priority */
180*4882a593Smuzhiyun #define DWC_CFGL_CH_SUSP	(1 << 8)	/* pause xfer */
181*4882a593Smuzhiyun #define DWC_CFGL_FIFO_EMPTY	(1 << 9)	/* pause xfer */
182*4882a593Smuzhiyun #define DWC_CFGL_HS_DST		(1 << 10)	/* handshake w/dst */
183*4882a593Smuzhiyun #define DWC_CFGL_HS_SRC		(1 << 11)	/* handshake w/src */
184*4882a593Smuzhiyun #define DWC_CFGL_LOCK_CH_XFER	(0 << 12)	/* scope of LOCK_CH */
185*4882a593Smuzhiyun #define DWC_CFGL_LOCK_CH_BLOCK	(1 << 12)
186*4882a593Smuzhiyun #define DWC_CFGL_LOCK_CH_XACT	(2 << 12)
187*4882a593Smuzhiyun #define DWC_CFGL_LOCK_BUS_XFER	(0 << 14)	/* scope of LOCK_BUS */
188*4882a593Smuzhiyun #define DWC_CFGL_LOCK_BUS_BLOCK	(1 << 14)
189*4882a593Smuzhiyun #define DWC_CFGL_LOCK_BUS_XACT	(2 << 14)
190*4882a593Smuzhiyun #define DWC_CFGL_LOCK_CH	(1 << 15)	/* channel lockout */
191*4882a593Smuzhiyun #define DWC_CFGL_LOCK_BUS	(1 << 16)	/* busmaster lockout */
192*4882a593Smuzhiyun #define DWC_CFGL_HS_DST_POL	(1 << 18)	/* dst handshake active low */
193*4882a593Smuzhiyun #define DWC_CFGL_HS_SRC_POL	(1 << 19)	/* src handshake active low */
194*4882a593Smuzhiyun #define DWC_CFGL_MAX_BURST(x)	((x) << 20)
195*4882a593Smuzhiyun #define DWC_CFGL_RELOAD_SAR	(1 << 30)
196*4882a593Smuzhiyun #define DWC_CFGL_RELOAD_DAR	(1 << 31)
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun /* Bitfields in CFG_HI */
199*4882a593Smuzhiyun #define DWC_CFGH_FCMODE		(1 << 0)
200*4882a593Smuzhiyun #define DWC_CFGH_FIFO_MODE	(1 << 1)
201*4882a593Smuzhiyun #define DWC_CFGH_PROTCTL(x)	((x) << 2)
202*4882a593Smuzhiyun #define DWC_CFGH_PROTCTL_DATA	(0 << 2)	/* data access - always set */
203*4882a593Smuzhiyun #define DWC_CFGH_PROTCTL_PRIV	(1 << 2)	/* privileged -> AHB HPROT[1] */
204*4882a593Smuzhiyun #define DWC_CFGH_PROTCTL_BUFFER	(2 << 2)	/* bufferable -> AHB HPROT[2] */
205*4882a593Smuzhiyun #define DWC_CFGH_PROTCTL_CACHE	(4 << 2)	/* cacheable  -> AHB HPROT[3] */
206*4882a593Smuzhiyun #define DWC_CFGH_DS_UPD_EN	(1 << 5)
207*4882a593Smuzhiyun #define DWC_CFGH_SS_UPD_EN	(1 << 6)
208*4882a593Smuzhiyun #define DWC_CFGH_SRC_PER(x)	((x) << 7)
209*4882a593Smuzhiyun #define DWC_CFGH_DST_PER(x)	((x) << 11)
210*4882a593Smuzhiyun 
211*4882a593Smuzhiyun /* Bitfields in SGR */
212*4882a593Smuzhiyun #define DWC_SGR_SGI(x)		((x) << 0)
213*4882a593Smuzhiyun #define DWC_SGR_SGC(x)		((x) << 20)
214*4882a593Smuzhiyun 
215*4882a593Smuzhiyun /* Bitfields in DSR */
216*4882a593Smuzhiyun #define DWC_DSR_DSI(x)		((x) << 0)
217*4882a593Smuzhiyun #define DWC_DSR_DSC(x)		((x) << 20)
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun /* Bitfields in CFG */
220*4882a593Smuzhiyun #define DW_CFG_DMA_EN		(1 << 0)
221*4882a593Smuzhiyun 
222*4882a593Smuzhiyun /* iDMA 32-bit support */
223*4882a593Smuzhiyun 
224*4882a593Smuzhiyun /* bursts size */
225*4882a593Smuzhiyun enum idma32_msize {
226*4882a593Smuzhiyun 	IDMA32_MSIZE_1,
227*4882a593Smuzhiyun 	IDMA32_MSIZE_2,
228*4882a593Smuzhiyun 	IDMA32_MSIZE_4,
229*4882a593Smuzhiyun 	IDMA32_MSIZE_8,
230*4882a593Smuzhiyun 	IDMA32_MSIZE_16,
231*4882a593Smuzhiyun 	IDMA32_MSIZE_32,
232*4882a593Smuzhiyun };
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun /* Bitfields in CTL_HI */
235*4882a593Smuzhiyun #define IDMA32C_CTLH_BLOCK_TS_MASK	GENMASK(16, 0)
236*4882a593Smuzhiyun #define IDMA32C_CTLH_BLOCK_TS(x)	((x) & IDMA32C_CTLH_BLOCK_TS_MASK)
237*4882a593Smuzhiyun #define IDMA32C_CTLH_DONE		(1 << 17)
238*4882a593Smuzhiyun 
239*4882a593Smuzhiyun /* Bitfields in CFG_LO */
240*4882a593Smuzhiyun #define IDMA32C_CFGL_DST_BURST_ALIGN	(1 << 0)	/* dst burst align */
241*4882a593Smuzhiyun #define IDMA32C_CFGL_SRC_BURST_ALIGN	(1 << 1)	/* src burst align */
242*4882a593Smuzhiyun #define IDMA32C_CFGL_CH_DRAIN		(1 << 10)	/* drain FIFO */
243*4882a593Smuzhiyun #define IDMA32C_CFGL_DST_OPT_BL		(1 << 20)	/* optimize dst burst length */
244*4882a593Smuzhiyun #define IDMA32C_CFGL_SRC_OPT_BL		(1 << 21)	/* optimize src burst length */
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun /* Bitfields in CFG_HI */
247*4882a593Smuzhiyun #define IDMA32C_CFGH_SRC_PER(x)		((x) << 0)
248*4882a593Smuzhiyun #define IDMA32C_CFGH_DST_PER(x)		((x) << 4)
249*4882a593Smuzhiyun #define IDMA32C_CFGH_RD_ISSUE_THD(x)	((x) << 8)
250*4882a593Smuzhiyun #define IDMA32C_CFGH_RW_ISSUE_THD(x)	((x) << 18)
251*4882a593Smuzhiyun #define IDMA32C_CFGH_SRC_PER_EXT(x)	((x) << 28)	/* src peripheral extension */
252*4882a593Smuzhiyun #define IDMA32C_CFGH_DST_PER_EXT(x)	((x) << 30)	/* dst peripheral extension */
253*4882a593Smuzhiyun 
254*4882a593Smuzhiyun /* Bitfields in FIFO_PARTITION */
255*4882a593Smuzhiyun #define IDMA32C_FP_PSIZE_CH0(x)		((x) << 0)
256*4882a593Smuzhiyun #define IDMA32C_FP_PSIZE_CH1(x)		((x) << 13)
257*4882a593Smuzhiyun #define IDMA32C_FP_UPDATE		(1 << 26)
258*4882a593Smuzhiyun 
259*4882a593Smuzhiyun enum dw_dmac_flags {
260*4882a593Smuzhiyun 	DW_DMA_IS_CYCLIC = 0,
261*4882a593Smuzhiyun 	DW_DMA_IS_SOFT_LLP = 1,
262*4882a593Smuzhiyun 	DW_DMA_IS_PAUSED = 2,
263*4882a593Smuzhiyun 	DW_DMA_IS_INITIALIZED = 3,
264*4882a593Smuzhiyun };
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun struct dw_dma_chan {
267*4882a593Smuzhiyun 	struct dma_chan			chan;
268*4882a593Smuzhiyun 	void __iomem			*ch_regs;
269*4882a593Smuzhiyun 	u8				mask;
270*4882a593Smuzhiyun 	u8				priority;
271*4882a593Smuzhiyun 	enum dma_transfer_direction	direction;
272*4882a593Smuzhiyun 
273*4882a593Smuzhiyun 	/* software emulation of the LLP transfers */
274*4882a593Smuzhiyun 	struct list_head	*tx_node_active;
275*4882a593Smuzhiyun 
276*4882a593Smuzhiyun 	spinlock_t		lock;
277*4882a593Smuzhiyun 
278*4882a593Smuzhiyun 	/* these other elements are all protected by lock */
279*4882a593Smuzhiyun 	unsigned long		flags;
280*4882a593Smuzhiyun 	struct list_head	active_list;
281*4882a593Smuzhiyun 	struct list_head	queue;
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 	unsigned int		descs_allocated;
284*4882a593Smuzhiyun 
285*4882a593Smuzhiyun 	/* hardware configuration */
286*4882a593Smuzhiyun 	unsigned int		block_size;
287*4882a593Smuzhiyun 	bool			nollp;
288*4882a593Smuzhiyun 	u32			max_burst;
289*4882a593Smuzhiyun 
290*4882a593Smuzhiyun 	/* custom slave configuration */
291*4882a593Smuzhiyun 	struct dw_dma_slave	dws;
292*4882a593Smuzhiyun 
293*4882a593Smuzhiyun 	/* configuration passed via .device_config */
294*4882a593Smuzhiyun 	struct dma_slave_config dma_sconfig;
295*4882a593Smuzhiyun };
296*4882a593Smuzhiyun 
297*4882a593Smuzhiyun static inline struct dw_dma_chan_regs __iomem *
__dwc_regs(struct dw_dma_chan * dwc)298*4882a593Smuzhiyun __dwc_regs(struct dw_dma_chan *dwc)
299*4882a593Smuzhiyun {
300*4882a593Smuzhiyun 	return dwc->ch_regs;
301*4882a593Smuzhiyun }
302*4882a593Smuzhiyun 
303*4882a593Smuzhiyun #define channel_readl(dwc, name) \
304*4882a593Smuzhiyun 	readl(&(__dwc_regs(dwc)->name))
305*4882a593Smuzhiyun #define channel_writel(dwc, name, val) \
306*4882a593Smuzhiyun 	writel((val), &(__dwc_regs(dwc)->name))
307*4882a593Smuzhiyun 
to_dw_dma_chan(struct dma_chan * chan)308*4882a593Smuzhiyun static inline struct dw_dma_chan *to_dw_dma_chan(struct dma_chan *chan)
309*4882a593Smuzhiyun {
310*4882a593Smuzhiyun 	return container_of(chan, struct dw_dma_chan, chan);
311*4882a593Smuzhiyun }
312*4882a593Smuzhiyun 
313*4882a593Smuzhiyun struct dw_dma {
314*4882a593Smuzhiyun 	struct dma_device	dma;
315*4882a593Smuzhiyun 	char			name[20];
316*4882a593Smuzhiyun 	void __iomem		*regs;
317*4882a593Smuzhiyun 	struct dma_pool		*desc_pool;
318*4882a593Smuzhiyun 	struct tasklet_struct	tasklet;
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 	/* channels */
321*4882a593Smuzhiyun 	struct dw_dma_chan	*chan;
322*4882a593Smuzhiyun 	u8			all_chan_mask;
323*4882a593Smuzhiyun 	u8			in_use;
324*4882a593Smuzhiyun 
325*4882a593Smuzhiyun 	/* Channel operations */
326*4882a593Smuzhiyun 	void	(*initialize_chan)(struct dw_dma_chan *dwc);
327*4882a593Smuzhiyun 	void	(*suspend_chan)(struct dw_dma_chan *dwc, bool drain);
328*4882a593Smuzhiyun 	void	(*resume_chan)(struct dw_dma_chan *dwc, bool drain);
329*4882a593Smuzhiyun 	u32	(*prepare_ctllo)(struct dw_dma_chan *dwc);
330*4882a593Smuzhiyun 	void	(*encode_maxburst)(struct dw_dma_chan *dwc, u32 *maxburst);
331*4882a593Smuzhiyun 	u32	(*bytes2block)(struct dw_dma_chan *dwc, size_t bytes,
332*4882a593Smuzhiyun 			       unsigned int width, size_t *len);
333*4882a593Smuzhiyun 	size_t	(*block2bytes)(struct dw_dma_chan *dwc, u32 block, u32 width);
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	/* Device operations */
336*4882a593Smuzhiyun 	void (*set_device_name)(struct dw_dma *dw, int id);
337*4882a593Smuzhiyun 	void (*disable)(struct dw_dma *dw);
338*4882a593Smuzhiyun 	void (*enable)(struct dw_dma *dw);
339*4882a593Smuzhiyun 
340*4882a593Smuzhiyun 	/* platform data */
341*4882a593Smuzhiyun 	struct dw_dma_platform_data	*pdata;
342*4882a593Smuzhiyun };
343*4882a593Smuzhiyun 
__dw_regs(struct dw_dma * dw)344*4882a593Smuzhiyun static inline struct dw_dma_regs __iomem *__dw_regs(struct dw_dma *dw)
345*4882a593Smuzhiyun {
346*4882a593Smuzhiyun 	return dw->regs;
347*4882a593Smuzhiyun }
348*4882a593Smuzhiyun 
349*4882a593Smuzhiyun #define dma_readl(dw, name) \
350*4882a593Smuzhiyun 	readl(&(__dw_regs(dw)->name))
351*4882a593Smuzhiyun #define dma_writel(dw, name, val) \
352*4882a593Smuzhiyun 	writel((val), &(__dw_regs(dw)->name))
353*4882a593Smuzhiyun 
354*4882a593Smuzhiyun #define idma32_readq(dw, name)				\
355*4882a593Smuzhiyun 	hi_lo_readq(&(__dw_regs(dw)->name))
356*4882a593Smuzhiyun #define idma32_writeq(dw, name, val)			\
357*4882a593Smuzhiyun 	hi_lo_writeq((val), &(__dw_regs(dw)->name))
358*4882a593Smuzhiyun 
359*4882a593Smuzhiyun #define channel_set_bit(dw, reg, mask) \
360*4882a593Smuzhiyun 	dma_writel(dw, reg, ((mask) << 8) | (mask))
361*4882a593Smuzhiyun #define channel_clear_bit(dw, reg, mask) \
362*4882a593Smuzhiyun 	dma_writel(dw, reg, ((mask) << 8) | 0)
363*4882a593Smuzhiyun 
to_dw_dma(struct dma_device * ddev)364*4882a593Smuzhiyun static inline struct dw_dma *to_dw_dma(struct dma_device *ddev)
365*4882a593Smuzhiyun {
366*4882a593Smuzhiyun 	return container_of(ddev, struct dw_dma, dma);
367*4882a593Smuzhiyun }
368*4882a593Smuzhiyun 
369*4882a593Smuzhiyun /* LLI == Linked List Item; a.k.a. DMA block descriptor */
370*4882a593Smuzhiyun struct dw_lli {
371*4882a593Smuzhiyun 	/* values that are not changed by hardware */
372*4882a593Smuzhiyun 	__le32		sar;
373*4882a593Smuzhiyun 	__le32		dar;
374*4882a593Smuzhiyun 	__le32		llp;		/* chain to next lli */
375*4882a593Smuzhiyun 	__le32		ctllo;
376*4882a593Smuzhiyun 	/* values that may get written back: */
377*4882a593Smuzhiyun 	__le32		ctlhi;
378*4882a593Smuzhiyun 	/* sstat and dstat can snapshot peripheral register state.
379*4882a593Smuzhiyun 	 * silicon config may discard either or both...
380*4882a593Smuzhiyun 	 */
381*4882a593Smuzhiyun 	__le32		sstat;
382*4882a593Smuzhiyun 	__le32		dstat;
383*4882a593Smuzhiyun };
384*4882a593Smuzhiyun 
385*4882a593Smuzhiyun struct dw_desc {
386*4882a593Smuzhiyun 	/* FIRST values the hardware uses */
387*4882a593Smuzhiyun 	struct dw_lli			lli;
388*4882a593Smuzhiyun 
389*4882a593Smuzhiyun #define lli_set(d, reg, v)		((d)->lli.reg |= cpu_to_le32(v))
390*4882a593Smuzhiyun #define lli_clear(d, reg, v)		((d)->lli.reg &= ~cpu_to_le32(v))
391*4882a593Smuzhiyun #define lli_read(d, reg)		le32_to_cpu((d)->lli.reg)
392*4882a593Smuzhiyun #define lli_write(d, reg, v)		((d)->lli.reg = cpu_to_le32(v))
393*4882a593Smuzhiyun 
394*4882a593Smuzhiyun 	/* THEN values for driver housekeeping */
395*4882a593Smuzhiyun 	struct list_head		desc_node;
396*4882a593Smuzhiyun 	struct list_head		tx_list;
397*4882a593Smuzhiyun 	struct dma_async_tx_descriptor	txd;
398*4882a593Smuzhiyun 	size_t				len;
399*4882a593Smuzhiyun 	size_t				total_len;
400*4882a593Smuzhiyun 	u32				residue;
401*4882a593Smuzhiyun };
402*4882a593Smuzhiyun 
403*4882a593Smuzhiyun #define to_dw_desc(h)	list_entry(h, struct dw_desc, desc_node)
404*4882a593Smuzhiyun 
405*4882a593Smuzhiyun static inline struct dw_desc *
txd_to_dw_desc(struct dma_async_tx_descriptor * txd)406*4882a593Smuzhiyun txd_to_dw_desc(struct dma_async_tx_descriptor *txd)
407*4882a593Smuzhiyun {
408*4882a593Smuzhiyun 	return container_of(txd, struct dw_desc, txd);
409*4882a593Smuzhiyun }
410