xref: /OK3568_Linux_fs/kernel/drivers/spi/spi-lantiq-ssc.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Copyright (C) 2011-2015 Daniel Schwierzeck <daniel.schwierzeck@gmail.com>
4*4882a593Smuzhiyun  * Copyright (C) 2016 Hauke Mehrtens <hauke@hauke-m.de>
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/kernel.h>
8*4882a593Smuzhiyun #include <linux/module.h>
9*4882a593Smuzhiyun #include <linux/of_device.h>
10*4882a593Smuzhiyun #include <linux/clk.h>
11*4882a593Smuzhiyun #include <linux/io.h>
12*4882a593Smuzhiyun #include <linux/delay.h>
13*4882a593Smuzhiyun #include <linux/interrupt.h>
14*4882a593Smuzhiyun #include <linux/sched.h>
15*4882a593Smuzhiyun #include <linux/completion.h>
16*4882a593Smuzhiyun #include <linux/spinlock.h>
17*4882a593Smuzhiyun #include <linux/err.h>
18*4882a593Smuzhiyun #include <linux/pm_runtime.h>
19*4882a593Smuzhiyun #include <linux/spi/spi.h>
20*4882a593Smuzhiyun 
21*4882a593Smuzhiyun #ifdef CONFIG_LANTIQ
22*4882a593Smuzhiyun #include <lantiq_soc.h>
23*4882a593Smuzhiyun #endif
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun #define LTQ_SPI_RX_IRQ_NAME	"spi_rx"
26*4882a593Smuzhiyun #define LTQ_SPI_TX_IRQ_NAME	"spi_tx"
27*4882a593Smuzhiyun #define LTQ_SPI_ERR_IRQ_NAME	"spi_err"
28*4882a593Smuzhiyun #define LTQ_SPI_FRM_IRQ_NAME	"spi_frm"
29*4882a593Smuzhiyun 
30*4882a593Smuzhiyun #define LTQ_SPI_CLC		0x00
31*4882a593Smuzhiyun #define LTQ_SPI_PISEL		0x04
32*4882a593Smuzhiyun #define LTQ_SPI_ID		0x08
33*4882a593Smuzhiyun #define LTQ_SPI_CON		0x10
34*4882a593Smuzhiyun #define LTQ_SPI_STAT		0x14
35*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE	0x18
36*4882a593Smuzhiyun #define LTQ_SPI_TB		0x20
37*4882a593Smuzhiyun #define LTQ_SPI_RB		0x24
38*4882a593Smuzhiyun #define LTQ_SPI_RXFCON		0x30
39*4882a593Smuzhiyun #define LTQ_SPI_TXFCON		0x34
40*4882a593Smuzhiyun #define LTQ_SPI_FSTAT		0x38
41*4882a593Smuzhiyun #define LTQ_SPI_BRT		0x40
42*4882a593Smuzhiyun #define LTQ_SPI_BRSTAT		0x44
43*4882a593Smuzhiyun #define LTQ_SPI_SFCON		0x60
44*4882a593Smuzhiyun #define LTQ_SPI_SFSTAT		0x64
45*4882a593Smuzhiyun #define LTQ_SPI_GPOCON		0x70
46*4882a593Smuzhiyun #define LTQ_SPI_GPOSTAT		0x74
47*4882a593Smuzhiyun #define LTQ_SPI_FPGO		0x78
48*4882a593Smuzhiyun #define LTQ_SPI_RXREQ		0x80
49*4882a593Smuzhiyun #define LTQ_SPI_RXCNT		0x84
50*4882a593Smuzhiyun #define LTQ_SPI_DMACON		0xec
51*4882a593Smuzhiyun #define LTQ_SPI_IRNEN		0xf4
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun #define LTQ_SPI_CLC_SMC_S	16	/* Clock divider for sleep mode */
54*4882a593Smuzhiyun #define LTQ_SPI_CLC_SMC_M	(0xFF << LTQ_SPI_CLC_SMC_S)
55*4882a593Smuzhiyun #define LTQ_SPI_CLC_RMC_S	8	/* Clock divider for normal run mode */
56*4882a593Smuzhiyun #define LTQ_SPI_CLC_RMC_M	(0xFF << LTQ_SPI_CLC_RMC_S)
57*4882a593Smuzhiyun #define LTQ_SPI_CLC_DISS	BIT(1)	/* Disable status bit */
58*4882a593Smuzhiyun #define LTQ_SPI_CLC_DISR	BIT(0)	/* Disable request bit */
59*4882a593Smuzhiyun 
60*4882a593Smuzhiyun #define LTQ_SPI_ID_TXFS_S	24	/* Implemented TX FIFO size */
61*4882a593Smuzhiyun #define LTQ_SPI_ID_RXFS_S	16	/* Implemented RX FIFO size */
62*4882a593Smuzhiyun #define LTQ_SPI_ID_MOD_S	8	/* Module ID */
63*4882a593Smuzhiyun #define LTQ_SPI_ID_MOD_M	(0xff << LTQ_SPI_ID_MOD_S)
64*4882a593Smuzhiyun #define LTQ_SPI_ID_CFG_S	5	/* DMA interface support */
65*4882a593Smuzhiyun #define LTQ_SPI_ID_CFG_M	(1 << LTQ_SPI_ID_CFG_S)
66*4882a593Smuzhiyun #define LTQ_SPI_ID_REV_M	0x1F	/* Hardware revision number */
67*4882a593Smuzhiyun 
68*4882a593Smuzhiyun #define LTQ_SPI_CON_BM_S	16	/* Data width selection */
69*4882a593Smuzhiyun #define LTQ_SPI_CON_BM_M	(0x1F << LTQ_SPI_CON_BM_S)
70*4882a593Smuzhiyun #define LTQ_SPI_CON_EM		BIT(24)	/* Echo mode */
71*4882a593Smuzhiyun #define LTQ_SPI_CON_IDLE	BIT(23)	/* Idle bit value */
72*4882a593Smuzhiyun #define LTQ_SPI_CON_ENBV	BIT(22)	/* Enable byte valid control */
73*4882a593Smuzhiyun #define LTQ_SPI_CON_RUEN	BIT(12)	/* Receive underflow error enable */
74*4882a593Smuzhiyun #define LTQ_SPI_CON_TUEN	BIT(11)	/* Transmit underflow error enable */
75*4882a593Smuzhiyun #define LTQ_SPI_CON_AEN		BIT(10)	/* Abort error enable */
76*4882a593Smuzhiyun #define LTQ_SPI_CON_REN		BIT(9)	/* Receive overflow error enable */
77*4882a593Smuzhiyun #define LTQ_SPI_CON_TEN		BIT(8)	/* Transmit overflow error enable */
78*4882a593Smuzhiyun #define LTQ_SPI_CON_LB		BIT(7)	/* Loopback control */
79*4882a593Smuzhiyun #define LTQ_SPI_CON_PO		BIT(6)	/* Clock polarity control */
80*4882a593Smuzhiyun #define LTQ_SPI_CON_PH		BIT(5)	/* Clock phase control */
81*4882a593Smuzhiyun #define LTQ_SPI_CON_HB		BIT(4)	/* Heading control */
82*4882a593Smuzhiyun #define LTQ_SPI_CON_RXOFF	BIT(1)	/* Switch receiver off */
83*4882a593Smuzhiyun #define LTQ_SPI_CON_TXOFF	BIT(0)	/* Switch transmitter off */
84*4882a593Smuzhiyun 
85*4882a593Smuzhiyun #define LTQ_SPI_STAT_RXBV_S	28
86*4882a593Smuzhiyun #define LTQ_SPI_STAT_RXBV_M	(0x7 << LTQ_SPI_STAT_RXBV_S)
87*4882a593Smuzhiyun #define LTQ_SPI_STAT_BSY	BIT(13)	/* Busy flag */
88*4882a593Smuzhiyun #define LTQ_SPI_STAT_RUE	BIT(12)	/* Receive underflow error flag */
89*4882a593Smuzhiyun #define LTQ_SPI_STAT_TUE	BIT(11)	/* Transmit underflow error flag */
90*4882a593Smuzhiyun #define LTQ_SPI_STAT_AE		BIT(10)	/* Abort error flag */
91*4882a593Smuzhiyun #define LTQ_SPI_STAT_RE		BIT(9)	/* Receive error flag */
92*4882a593Smuzhiyun #define LTQ_SPI_STAT_TE		BIT(8)	/* Transmit error flag */
93*4882a593Smuzhiyun #define LTQ_SPI_STAT_ME		BIT(7)	/* Mode error flag */
94*4882a593Smuzhiyun #define LTQ_SPI_STAT_MS		BIT(1)	/* Master/slave select bit */
95*4882a593Smuzhiyun #define LTQ_SPI_STAT_EN		BIT(0)	/* Enable bit */
96*4882a593Smuzhiyun #define LTQ_SPI_STAT_ERRORS	(LTQ_SPI_STAT_ME | LTQ_SPI_STAT_TE | \
97*4882a593Smuzhiyun 				 LTQ_SPI_STAT_RE | LTQ_SPI_STAT_AE | \
98*4882a593Smuzhiyun 				 LTQ_SPI_STAT_TUE | LTQ_SPI_STAT_RUE)
99*4882a593Smuzhiyun 
100*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETTUE	BIT(15)	/* Set transmit underflow error flag */
101*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETAE	BIT(14)	/* Set abort error flag */
102*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETRE	BIT(13)	/* Set receive error flag */
103*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETTE	BIT(12)	/* Set transmit error flag */
104*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRTUE	BIT(11)	/* Clear transmit underflow error flag */
105*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRAE	BIT(10)	/* Clear abort error flag */
106*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRRE	BIT(9)	/* Clear receive error flag */
107*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRTE	BIT(8)	/* Clear transmit error flag */
108*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETME	BIT(7)	/* Set mode error flag */
109*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRME	BIT(6)	/* Clear mode error flag */
110*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETRUE	BIT(5)	/* Set receive underflow error flag */
111*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRRUE	BIT(4)	/* Clear receive underflow error flag */
112*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETMS	BIT(3)	/* Set master select bit */
113*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLRMS	BIT(2)	/* Clear master select bit */
114*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_SETEN	BIT(1)	/* Set enable bit (operational mode) */
115*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLREN	BIT(0)	/* Clear enable bit (config mode */
116*4882a593Smuzhiyun #define LTQ_SPI_WHBSTATE_CLR_ERRORS	(LTQ_SPI_WHBSTATE_CLRRUE | \
117*4882a593Smuzhiyun 					 LTQ_SPI_WHBSTATE_CLRME | \
118*4882a593Smuzhiyun 					 LTQ_SPI_WHBSTATE_CLRTE | \
119*4882a593Smuzhiyun 					 LTQ_SPI_WHBSTATE_CLRRE | \
120*4882a593Smuzhiyun 					 LTQ_SPI_WHBSTATE_CLRAE | \
121*4882a593Smuzhiyun 					 LTQ_SPI_WHBSTATE_CLRTUE)
122*4882a593Smuzhiyun 
123*4882a593Smuzhiyun #define LTQ_SPI_RXFCON_RXFITL_S	8	/* FIFO interrupt trigger level */
124*4882a593Smuzhiyun #define LTQ_SPI_RXFCON_RXFLU	BIT(1)	/* FIFO flush */
125*4882a593Smuzhiyun #define LTQ_SPI_RXFCON_RXFEN	BIT(0)	/* FIFO enable */
126*4882a593Smuzhiyun 
127*4882a593Smuzhiyun #define LTQ_SPI_TXFCON_TXFITL_S	8	/* FIFO interrupt trigger level */
128*4882a593Smuzhiyun #define LTQ_SPI_TXFCON_TXFLU	BIT(1)	/* FIFO flush */
129*4882a593Smuzhiyun #define LTQ_SPI_TXFCON_TXFEN	BIT(0)	/* FIFO enable */
130*4882a593Smuzhiyun 
131*4882a593Smuzhiyun #define LTQ_SPI_FSTAT_RXFFL_S	0
132*4882a593Smuzhiyun #define LTQ_SPI_FSTAT_TXFFL_S	8
133*4882a593Smuzhiyun 
134*4882a593Smuzhiyun #define LTQ_SPI_GPOCON_ISCSBN_S	8
135*4882a593Smuzhiyun #define LTQ_SPI_GPOCON_INVOUTN_S	0
136*4882a593Smuzhiyun 
137*4882a593Smuzhiyun #define LTQ_SPI_FGPO_SETOUTN_S	8
138*4882a593Smuzhiyun #define LTQ_SPI_FGPO_CLROUTN_S	0
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun #define LTQ_SPI_RXREQ_RXCNT_M	0xFFFF	/* Receive count value */
141*4882a593Smuzhiyun #define LTQ_SPI_RXCNT_TODO_M	0xFFFF	/* Recevie to-do value */
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_TFI	BIT(4)	/* TX finished interrupt */
144*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_F		BIT(3)	/* Frame end interrupt request */
145*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_E		BIT(2)	/* Error end interrupt request */
146*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_T_XWAY	BIT(1)	/* Transmit end interrupt request */
147*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_R_XWAY	BIT(0)	/* Receive end interrupt request */
148*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_R_XRX	BIT(1)	/* Transmit end interrupt request */
149*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_T_XRX	BIT(0)	/* Receive end interrupt request */
150*4882a593Smuzhiyun #define LTQ_SPI_IRNEN_ALL	0x1F
151*4882a593Smuzhiyun 
152*4882a593Smuzhiyun struct lantiq_ssc_spi;
153*4882a593Smuzhiyun 
154*4882a593Smuzhiyun struct lantiq_ssc_hwcfg {
155*4882a593Smuzhiyun 	int (*cfg_irq)(struct platform_device *pdev, struct lantiq_ssc_spi *spi);
156*4882a593Smuzhiyun 	unsigned int	irnen_r;
157*4882a593Smuzhiyun 	unsigned int	irnen_t;
158*4882a593Smuzhiyun 	unsigned int	irncr;
159*4882a593Smuzhiyun 	unsigned int	irnicr;
160*4882a593Smuzhiyun 	bool		irq_ack;
161*4882a593Smuzhiyun 	u32		fifo_size_mask;
162*4882a593Smuzhiyun };
163*4882a593Smuzhiyun 
164*4882a593Smuzhiyun struct lantiq_ssc_spi {
165*4882a593Smuzhiyun 	struct spi_master		*master;
166*4882a593Smuzhiyun 	struct device			*dev;
167*4882a593Smuzhiyun 	void __iomem			*regbase;
168*4882a593Smuzhiyun 	struct clk			*spi_clk;
169*4882a593Smuzhiyun 	struct clk			*fpi_clk;
170*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg	*hwcfg;
171*4882a593Smuzhiyun 
172*4882a593Smuzhiyun 	spinlock_t			lock;
173*4882a593Smuzhiyun 	struct workqueue_struct		*wq;
174*4882a593Smuzhiyun 	struct work_struct		work;
175*4882a593Smuzhiyun 
176*4882a593Smuzhiyun 	const u8			*tx;
177*4882a593Smuzhiyun 	u8				*rx;
178*4882a593Smuzhiyun 	unsigned int			tx_todo;
179*4882a593Smuzhiyun 	unsigned int			rx_todo;
180*4882a593Smuzhiyun 	unsigned int			bits_per_word;
181*4882a593Smuzhiyun 	unsigned int			speed_hz;
182*4882a593Smuzhiyun 	unsigned int			tx_fifo_size;
183*4882a593Smuzhiyun 	unsigned int			rx_fifo_size;
184*4882a593Smuzhiyun 	unsigned int			base_cs;
185*4882a593Smuzhiyun 	unsigned int			fdx_tx_level;
186*4882a593Smuzhiyun };
187*4882a593Smuzhiyun 
lantiq_ssc_readl(const struct lantiq_ssc_spi * spi,u32 reg)188*4882a593Smuzhiyun static u32 lantiq_ssc_readl(const struct lantiq_ssc_spi *spi, u32 reg)
189*4882a593Smuzhiyun {
190*4882a593Smuzhiyun 	return __raw_readl(spi->regbase + reg);
191*4882a593Smuzhiyun }
192*4882a593Smuzhiyun 
lantiq_ssc_writel(const struct lantiq_ssc_spi * spi,u32 val,u32 reg)193*4882a593Smuzhiyun static void lantiq_ssc_writel(const struct lantiq_ssc_spi *spi, u32 val,
194*4882a593Smuzhiyun 			      u32 reg)
195*4882a593Smuzhiyun {
196*4882a593Smuzhiyun 	__raw_writel(val, spi->regbase + reg);
197*4882a593Smuzhiyun }
198*4882a593Smuzhiyun 
lantiq_ssc_maskl(const struct lantiq_ssc_spi * spi,u32 clr,u32 set,u32 reg)199*4882a593Smuzhiyun static void lantiq_ssc_maskl(const struct lantiq_ssc_spi *spi, u32 clr,
200*4882a593Smuzhiyun 			     u32 set, u32 reg)
201*4882a593Smuzhiyun {
202*4882a593Smuzhiyun 	u32 val = __raw_readl(spi->regbase + reg);
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun 	val &= ~clr;
205*4882a593Smuzhiyun 	val |= set;
206*4882a593Smuzhiyun 	__raw_writel(val, spi->regbase + reg);
207*4882a593Smuzhiyun }
208*4882a593Smuzhiyun 
tx_fifo_level(const struct lantiq_ssc_spi * spi)209*4882a593Smuzhiyun static unsigned int tx_fifo_level(const struct lantiq_ssc_spi *spi)
210*4882a593Smuzhiyun {
211*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
212*4882a593Smuzhiyun 	u32 fstat = lantiq_ssc_readl(spi, LTQ_SPI_FSTAT);
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 	return (fstat >> LTQ_SPI_FSTAT_TXFFL_S) & hwcfg->fifo_size_mask;
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun 
rx_fifo_level(const struct lantiq_ssc_spi * spi)217*4882a593Smuzhiyun static unsigned int rx_fifo_level(const struct lantiq_ssc_spi *spi)
218*4882a593Smuzhiyun {
219*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
220*4882a593Smuzhiyun 	u32 fstat = lantiq_ssc_readl(spi, LTQ_SPI_FSTAT);
221*4882a593Smuzhiyun 
222*4882a593Smuzhiyun 	return (fstat >> LTQ_SPI_FSTAT_RXFFL_S) & hwcfg->fifo_size_mask;
223*4882a593Smuzhiyun }
224*4882a593Smuzhiyun 
tx_fifo_free(const struct lantiq_ssc_spi * spi)225*4882a593Smuzhiyun static unsigned int tx_fifo_free(const struct lantiq_ssc_spi *spi)
226*4882a593Smuzhiyun {
227*4882a593Smuzhiyun 	return spi->tx_fifo_size - tx_fifo_level(spi);
228*4882a593Smuzhiyun }
229*4882a593Smuzhiyun 
rx_fifo_reset(const struct lantiq_ssc_spi * spi)230*4882a593Smuzhiyun static void rx_fifo_reset(const struct lantiq_ssc_spi *spi)
231*4882a593Smuzhiyun {
232*4882a593Smuzhiyun 	u32 val = spi->rx_fifo_size << LTQ_SPI_RXFCON_RXFITL_S;
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun 	val |= LTQ_SPI_RXFCON_RXFEN | LTQ_SPI_RXFCON_RXFLU;
235*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, val, LTQ_SPI_RXFCON);
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun 
tx_fifo_reset(const struct lantiq_ssc_spi * spi)238*4882a593Smuzhiyun static void tx_fifo_reset(const struct lantiq_ssc_spi *spi)
239*4882a593Smuzhiyun {
240*4882a593Smuzhiyun 	u32 val = 1 << LTQ_SPI_TXFCON_TXFITL_S;
241*4882a593Smuzhiyun 
242*4882a593Smuzhiyun 	val |= LTQ_SPI_TXFCON_TXFEN | LTQ_SPI_TXFCON_TXFLU;
243*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, val, LTQ_SPI_TXFCON);
244*4882a593Smuzhiyun }
245*4882a593Smuzhiyun 
rx_fifo_flush(const struct lantiq_ssc_spi * spi)246*4882a593Smuzhiyun static void rx_fifo_flush(const struct lantiq_ssc_spi *spi)
247*4882a593Smuzhiyun {
248*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, 0, LTQ_SPI_RXFCON_RXFLU, LTQ_SPI_RXFCON);
249*4882a593Smuzhiyun }
250*4882a593Smuzhiyun 
tx_fifo_flush(const struct lantiq_ssc_spi * spi)251*4882a593Smuzhiyun static void tx_fifo_flush(const struct lantiq_ssc_spi *spi)
252*4882a593Smuzhiyun {
253*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, 0, LTQ_SPI_TXFCON_TXFLU, LTQ_SPI_TXFCON);
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun 
hw_enter_config_mode(const struct lantiq_ssc_spi * spi)256*4882a593Smuzhiyun static void hw_enter_config_mode(const struct lantiq_ssc_spi *spi)
257*4882a593Smuzhiyun {
258*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, LTQ_SPI_WHBSTATE_CLREN, LTQ_SPI_WHBSTATE);
259*4882a593Smuzhiyun }
260*4882a593Smuzhiyun 
hw_enter_active_mode(const struct lantiq_ssc_spi * spi)261*4882a593Smuzhiyun static void hw_enter_active_mode(const struct lantiq_ssc_spi *spi)
262*4882a593Smuzhiyun {
263*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, LTQ_SPI_WHBSTATE_SETEN, LTQ_SPI_WHBSTATE);
264*4882a593Smuzhiyun }
265*4882a593Smuzhiyun 
hw_setup_speed_hz(const struct lantiq_ssc_spi * spi,unsigned int max_speed_hz)266*4882a593Smuzhiyun static void hw_setup_speed_hz(const struct lantiq_ssc_spi *spi,
267*4882a593Smuzhiyun 			      unsigned int max_speed_hz)
268*4882a593Smuzhiyun {
269*4882a593Smuzhiyun 	u32 spi_clk, brt;
270*4882a593Smuzhiyun 
271*4882a593Smuzhiyun 	/*
272*4882a593Smuzhiyun 	 * SPI module clock is derived from FPI bus clock dependent on
273*4882a593Smuzhiyun 	 * divider value in CLC.RMS which is always set to 1.
274*4882a593Smuzhiyun 	 *
275*4882a593Smuzhiyun 	 *                 f_SPI
276*4882a593Smuzhiyun 	 * baudrate = --------------
277*4882a593Smuzhiyun 	 *             2 * (BR + 1)
278*4882a593Smuzhiyun 	 */
279*4882a593Smuzhiyun 	spi_clk = clk_get_rate(spi->fpi_clk) / 2;
280*4882a593Smuzhiyun 
281*4882a593Smuzhiyun 	if (max_speed_hz > spi_clk)
282*4882a593Smuzhiyun 		brt = 0;
283*4882a593Smuzhiyun 	else
284*4882a593Smuzhiyun 		brt = spi_clk / max_speed_hz - 1;
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 	if (brt > 0xFFFF)
287*4882a593Smuzhiyun 		brt = 0xFFFF;
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun 	dev_dbg(spi->dev, "spi_clk %u, max_speed_hz %u, brt %u\n",
290*4882a593Smuzhiyun 		spi_clk, max_speed_hz, brt);
291*4882a593Smuzhiyun 
292*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, brt, LTQ_SPI_BRT);
293*4882a593Smuzhiyun }
294*4882a593Smuzhiyun 
hw_setup_bits_per_word(const struct lantiq_ssc_spi * spi,unsigned int bits_per_word)295*4882a593Smuzhiyun static void hw_setup_bits_per_word(const struct lantiq_ssc_spi *spi,
296*4882a593Smuzhiyun 				   unsigned int bits_per_word)
297*4882a593Smuzhiyun {
298*4882a593Smuzhiyun 	u32 bm;
299*4882a593Smuzhiyun 
300*4882a593Smuzhiyun 	/* CON.BM value = bits_per_word - 1 */
301*4882a593Smuzhiyun 	bm = (bits_per_word - 1) << LTQ_SPI_CON_BM_S;
302*4882a593Smuzhiyun 
303*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, LTQ_SPI_CON_BM_M, bm, LTQ_SPI_CON);
304*4882a593Smuzhiyun }
305*4882a593Smuzhiyun 
hw_setup_clock_mode(const struct lantiq_ssc_spi * spi,unsigned int mode)306*4882a593Smuzhiyun static void hw_setup_clock_mode(const struct lantiq_ssc_spi *spi,
307*4882a593Smuzhiyun 				unsigned int mode)
308*4882a593Smuzhiyun {
309*4882a593Smuzhiyun 	u32 con_set = 0, con_clr = 0;
310*4882a593Smuzhiyun 
311*4882a593Smuzhiyun 	/*
312*4882a593Smuzhiyun 	 * SPI mode mapping in CON register:
313*4882a593Smuzhiyun 	 * Mode CPOL CPHA CON.PO CON.PH
314*4882a593Smuzhiyun 	 *  0    0    0      0      1
315*4882a593Smuzhiyun 	 *  1    0    1      0      0
316*4882a593Smuzhiyun 	 *  2    1    0      1      1
317*4882a593Smuzhiyun 	 *  3    1    1      1      0
318*4882a593Smuzhiyun 	 */
319*4882a593Smuzhiyun 	if (mode & SPI_CPHA)
320*4882a593Smuzhiyun 		con_clr |= LTQ_SPI_CON_PH;
321*4882a593Smuzhiyun 	else
322*4882a593Smuzhiyun 		con_set |= LTQ_SPI_CON_PH;
323*4882a593Smuzhiyun 
324*4882a593Smuzhiyun 	if (mode & SPI_CPOL)
325*4882a593Smuzhiyun 		con_set |= LTQ_SPI_CON_PO | LTQ_SPI_CON_IDLE;
326*4882a593Smuzhiyun 	else
327*4882a593Smuzhiyun 		con_clr |= LTQ_SPI_CON_PO | LTQ_SPI_CON_IDLE;
328*4882a593Smuzhiyun 
329*4882a593Smuzhiyun 	/* Set heading control */
330*4882a593Smuzhiyun 	if (mode & SPI_LSB_FIRST)
331*4882a593Smuzhiyun 		con_clr |= LTQ_SPI_CON_HB;
332*4882a593Smuzhiyun 	else
333*4882a593Smuzhiyun 		con_set |= LTQ_SPI_CON_HB;
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	/* Set loopback mode */
336*4882a593Smuzhiyun 	if (mode & SPI_LOOP)
337*4882a593Smuzhiyun 		con_set |= LTQ_SPI_CON_LB;
338*4882a593Smuzhiyun 	else
339*4882a593Smuzhiyun 		con_clr |= LTQ_SPI_CON_LB;
340*4882a593Smuzhiyun 
341*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, con_clr, con_set, LTQ_SPI_CON);
342*4882a593Smuzhiyun }
343*4882a593Smuzhiyun 
lantiq_ssc_hw_init(const struct lantiq_ssc_spi * spi)344*4882a593Smuzhiyun static void lantiq_ssc_hw_init(const struct lantiq_ssc_spi *spi)
345*4882a593Smuzhiyun {
346*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
347*4882a593Smuzhiyun 
348*4882a593Smuzhiyun 	/*
349*4882a593Smuzhiyun 	 * Set clock divider for run mode to 1 to
350*4882a593Smuzhiyun 	 * run at same frequency as FPI bus
351*4882a593Smuzhiyun 	 */
352*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, 1 << LTQ_SPI_CLC_RMC_S, LTQ_SPI_CLC);
353*4882a593Smuzhiyun 
354*4882a593Smuzhiyun 	/* Put controller into config mode */
355*4882a593Smuzhiyun 	hw_enter_config_mode(spi);
356*4882a593Smuzhiyun 
357*4882a593Smuzhiyun 	/* Clear error flags */
358*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, 0, LTQ_SPI_WHBSTATE_CLR_ERRORS, LTQ_SPI_WHBSTATE);
359*4882a593Smuzhiyun 
360*4882a593Smuzhiyun 	/* Enable error checking, disable TX/RX */
361*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, LTQ_SPI_CON_RUEN | LTQ_SPI_CON_AEN |
362*4882a593Smuzhiyun 		LTQ_SPI_CON_TEN | LTQ_SPI_CON_REN | LTQ_SPI_CON_TXOFF |
363*4882a593Smuzhiyun 		LTQ_SPI_CON_RXOFF, LTQ_SPI_CON);
364*4882a593Smuzhiyun 
365*4882a593Smuzhiyun 	/* Setup default SPI mode */
366*4882a593Smuzhiyun 	hw_setup_bits_per_word(spi, spi->bits_per_word);
367*4882a593Smuzhiyun 	hw_setup_clock_mode(spi, SPI_MODE_0);
368*4882a593Smuzhiyun 
369*4882a593Smuzhiyun 	/* Enable master mode and clear error flags */
370*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, LTQ_SPI_WHBSTATE_SETMS |
371*4882a593Smuzhiyun 			       LTQ_SPI_WHBSTATE_CLR_ERRORS,
372*4882a593Smuzhiyun 			       LTQ_SPI_WHBSTATE);
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun 	/* Reset GPIO/CS registers */
375*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, 0, LTQ_SPI_GPOCON);
376*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, 0xFF00, LTQ_SPI_FPGO);
377*4882a593Smuzhiyun 
378*4882a593Smuzhiyun 	/* Enable and flush FIFOs */
379*4882a593Smuzhiyun 	rx_fifo_reset(spi);
380*4882a593Smuzhiyun 	tx_fifo_reset(spi);
381*4882a593Smuzhiyun 
382*4882a593Smuzhiyun 	/* Enable interrupts */
383*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, hwcfg->irnen_t | hwcfg->irnen_r |
384*4882a593Smuzhiyun 			  LTQ_SPI_IRNEN_E, LTQ_SPI_IRNEN);
385*4882a593Smuzhiyun }
386*4882a593Smuzhiyun 
lantiq_ssc_setup(struct spi_device * spidev)387*4882a593Smuzhiyun static int lantiq_ssc_setup(struct spi_device *spidev)
388*4882a593Smuzhiyun {
389*4882a593Smuzhiyun 	struct spi_master *master = spidev->master;
390*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
391*4882a593Smuzhiyun 	unsigned int cs = spidev->chip_select;
392*4882a593Smuzhiyun 	u32 gpocon;
393*4882a593Smuzhiyun 
394*4882a593Smuzhiyun 	/* GPIOs are used for CS */
395*4882a593Smuzhiyun 	if (spidev->cs_gpiod)
396*4882a593Smuzhiyun 		return 0;
397*4882a593Smuzhiyun 
398*4882a593Smuzhiyun 	dev_dbg(spi->dev, "using internal chipselect %u\n", cs);
399*4882a593Smuzhiyun 
400*4882a593Smuzhiyun 	if (cs < spi->base_cs) {
401*4882a593Smuzhiyun 		dev_err(spi->dev,
402*4882a593Smuzhiyun 			"chipselect %i too small (min %i)\n", cs, spi->base_cs);
403*4882a593Smuzhiyun 		return -EINVAL;
404*4882a593Smuzhiyun 	}
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun 	/* set GPO pin to CS mode */
407*4882a593Smuzhiyun 	gpocon = 1 << ((cs - spi->base_cs) + LTQ_SPI_GPOCON_ISCSBN_S);
408*4882a593Smuzhiyun 
409*4882a593Smuzhiyun 	/* invert GPO pin */
410*4882a593Smuzhiyun 	if (spidev->mode & SPI_CS_HIGH)
411*4882a593Smuzhiyun 		gpocon |= 1 << (cs - spi->base_cs);
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, 0, gpocon, LTQ_SPI_GPOCON);
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun 	return 0;
416*4882a593Smuzhiyun }
417*4882a593Smuzhiyun 
lantiq_ssc_prepare_message(struct spi_master * master,struct spi_message * message)418*4882a593Smuzhiyun static int lantiq_ssc_prepare_message(struct spi_master *master,
419*4882a593Smuzhiyun 				      struct spi_message *message)
420*4882a593Smuzhiyun {
421*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun 	hw_enter_config_mode(spi);
424*4882a593Smuzhiyun 	hw_setup_clock_mode(spi, message->spi->mode);
425*4882a593Smuzhiyun 	hw_enter_active_mode(spi);
426*4882a593Smuzhiyun 
427*4882a593Smuzhiyun 	return 0;
428*4882a593Smuzhiyun }
429*4882a593Smuzhiyun 
hw_setup_transfer(struct lantiq_ssc_spi * spi,struct spi_device * spidev,struct spi_transfer * t)430*4882a593Smuzhiyun static void hw_setup_transfer(struct lantiq_ssc_spi *spi,
431*4882a593Smuzhiyun 			      struct spi_device *spidev, struct spi_transfer *t)
432*4882a593Smuzhiyun {
433*4882a593Smuzhiyun 	unsigned int speed_hz = t->speed_hz;
434*4882a593Smuzhiyun 	unsigned int bits_per_word = t->bits_per_word;
435*4882a593Smuzhiyun 	u32 con;
436*4882a593Smuzhiyun 
437*4882a593Smuzhiyun 	if (bits_per_word != spi->bits_per_word ||
438*4882a593Smuzhiyun 		speed_hz != spi->speed_hz) {
439*4882a593Smuzhiyun 		hw_enter_config_mode(spi);
440*4882a593Smuzhiyun 		hw_setup_speed_hz(spi, speed_hz);
441*4882a593Smuzhiyun 		hw_setup_bits_per_word(spi, bits_per_word);
442*4882a593Smuzhiyun 		hw_enter_active_mode(spi);
443*4882a593Smuzhiyun 
444*4882a593Smuzhiyun 		spi->speed_hz = speed_hz;
445*4882a593Smuzhiyun 		spi->bits_per_word = bits_per_word;
446*4882a593Smuzhiyun 	}
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	/* Configure transmitter and receiver */
449*4882a593Smuzhiyun 	con = lantiq_ssc_readl(spi, LTQ_SPI_CON);
450*4882a593Smuzhiyun 	if (t->tx_buf)
451*4882a593Smuzhiyun 		con &= ~LTQ_SPI_CON_TXOFF;
452*4882a593Smuzhiyun 	else
453*4882a593Smuzhiyun 		con |= LTQ_SPI_CON_TXOFF;
454*4882a593Smuzhiyun 
455*4882a593Smuzhiyun 	if (t->rx_buf)
456*4882a593Smuzhiyun 		con &= ~LTQ_SPI_CON_RXOFF;
457*4882a593Smuzhiyun 	else
458*4882a593Smuzhiyun 		con |= LTQ_SPI_CON_RXOFF;
459*4882a593Smuzhiyun 
460*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, con, LTQ_SPI_CON);
461*4882a593Smuzhiyun }
462*4882a593Smuzhiyun 
lantiq_ssc_unprepare_message(struct spi_master * master,struct spi_message * message)463*4882a593Smuzhiyun static int lantiq_ssc_unprepare_message(struct spi_master *master,
464*4882a593Smuzhiyun 					struct spi_message *message)
465*4882a593Smuzhiyun {
466*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
467*4882a593Smuzhiyun 
468*4882a593Smuzhiyun 	flush_workqueue(spi->wq);
469*4882a593Smuzhiyun 
470*4882a593Smuzhiyun 	/* Disable transmitter and receiver while idle */
471*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, 0, LTQ_SPI_CON_TXOFF | LTQ_SPI_CON_RXOFF,
472*4882a593Smuzhiyun 			 LTQ_SPI_CON);
473*4882a593Smuzhiyun 
474*4882a593Smuzhiyun 	return 0;
475*4882a593Smuzhiyun }
476*4882a593Smuzhiyun 
tx_fifo_write(struct lantiq_ssc_spi * spi)477*4882a593Smuzhiyun static void tx_fifo_write(struct lantiq_ssc_spi *spi)
478*4882a593Smuzhiyun {
479*4882a593Smuzhiyun 	const u8 *tx8;
480*4882a593Smuzhiyun 	const u16 *tx16;
481*4882a593Smuzhiyun 	const u32 *tx32;
482*4882a593Smuzhiyun 	u32 data;
483*4882a593Smuzhiyun 	unsigned int tx_free = tx_fifo_free(spi);
484*4882a593Smuzhiyun 
485*4882a593Smuzhiyun 	spi->fdx_tx_level = 0;
486*4882a593Smuzhiyun 	while (spi->tx_todo && tx_free) {
487*4882a593Smuzhiyun 		switch (spi->bits_per_word) {
488*4882a593Smuzhiyun 		case 2 ... 8:
489*4882a593Smuzhiyun 			tx8 = spi->tx;
490*4882a593Smuzhiyun 			data = *tx8;
491*4882a593Smuzhiyun 			spi->tx_todo--;
492*4882a593Smuzhiyun 			spi->tx++;
493*4882a593Smuzhiyun 			break;
494*4882a593Smuzhiyun 		case 16:
495*4882a593Smuzhiyun 			tx16 = (u16 *) spi->tx;
496*4882a593Smuzhiyun 			data = *tx16;
497*4882a593Smuzhiyun 			spi->tx_todo -= 2;
498*4882a593Smuzhiyun 			spi->tx += 2;
499*4882a593Smuzhiyun 			break;
500*4882a593Smuzhiyun 		case 32:
501*4882a593Smuzhiyun 			tx32 = (u32 *) spi->tx;
502*4882a593Smuzhiyun 			data = *tx32;
503*4882a593Smuzhiyun 			spi->tx_todo -= 4;
504*4882a593Smuzhiyun 			spi->tx += 4;
505*4882a593Smuzhiyun 			break;
506*4882a593Smuzhiyun 		default:
507*4882a593Smuzhiyun 			WARN_ON(1);
508*4882a593Smuzhiyun 			data = 0;
509*4882a593Smuzhiyun 			break;
510*4882a593Smuzhiyun 		}
511*4882a593Smuzhiyun 
512*4882a593Smuzhiyun 		lantiq_ssc_writel(spi, data, LTQ_SPI_TB);
513*4882a593Smuzhiyun 		tx_free--;
514*4882a593Smuzhiyun 		spi->fdx_tx_level++;
515*4882a593Smuzhiyun 	}
516*4882a593Smuzhiyun }
517*4882a593Smuzhiyun 
rx_fifo_read_full_duplex(struct lantiq_ssc_spi * spi)518*4882a593Smuzhiyun static void rx_fifo_read_full_duplex(struct lantiq_ssc_spi *spi)
519*4882a593Smuzhiyun {
520*4882a593Smuzhiyun 	u8 *rx8;
521*4882a593Smuzhiyun 	u16 *rx16;
522*4882a593Smuzhiyun 	u32 *rx32;
523*4882a593Smuzhiyun 	u32 data;
524*4882a593Smuzhiyun 	unsigned int rx_fill = rx_fifo_level(spi);
525*4882a593Smuzhiyun 
526*4882a593Smuzhiyun 	/*
527*4882a593Smuzhiyun 	 * Wait until all expected data to be shifted in.
528*4882a593Smuzhiyun 	 * Otherwise, rx overrun may occur.
529*4882a593Smuzhiyun 	 */
530*4882a593Smuzhiyun 	while (rx_fill != spi->fdx_tx_level)
531*4882a593Smuzhiyun 		rx_fill = rx_fifo_level(spi);
532*4882a593Smuzhiyun 
533*4882a593Smuzhiyun 	while (rx_fill) {
534*4882a593Smuzhiyun 		data = lantiq_ssc_readl(spi, LTQ_SPI_RB);
535*4882a593Smuzhiyun 
536*4882a593Smuzhiyun 		switch (spi->bits_per_word) {
537*4882a593Smuzhiyun 		case 2 ... 8:
538*4882a593Smuzhiyun 			rx8 = spi->rx;
539*4882a593Smuzhiyun 			*rx8 = data;
540*4882a593Smuzhiyun 			spi->rx_todo--;
541*4882a593Smuzhiyun 			spi->rx++;
542*4882a593Smuzhiyun 			break;
543*4882a593Smuzhiyun 		case 16:
544*4882a593Smuzhiyun 			rx16 = (u16 *) spi->rx;
545*4882a593Smuzhiyun 			*rx16 = data;
546*4882a593Smuzhiyun 			spi->rx_todo -= 2;
547*4882a593Smuzhiyun 			spi->rx += 2;
548*4882a593Smuzhiyun 			break;
549*4882a593Smuzhiyun 		case 32:
550*4882a593Smuzhiyun 			rx32 = (u32 *) spi->rx;
551*4882a593Smuzhiyun 			*rx32 = data;
552*4882a593Smuzhiyun 			spi->rx_todo -= 4;
553*4882a593Smuzhiyun 			spi->rx += 4;
554*4882a593Smuzhiyun 			break;
555*4882a593Smuzhiyun 		default:
556*4882a593Smuzhiyun 			WARN_ON(1);
557*4882a593Smuzhiyun 			break;
558*4882a593Smuzhiyun 		}
559*4882a593Smuzhiyun 
560*4882a593Smuzhiyun 		rx_fill--;
561*4882a593Smuzhiyun 	}
562*4882a593Smuzhiyun }
563*4882a593Smuzhiyun 
rx_fifo_read_half_duplex(struct lantiq_ssc_spi * spi)564*4882a593Smuzhiyun static void rx_fifo_read_half_duplex(struct lantiq_ssc_spi *spi)
565*4882a593Smuzhiyun {
566*4882a593Smuzhiyun 	u32 data, *rx32;
567*4882a593Smuzhiyun 	u8 *rx8;
568*4882a593Smuzhiyun 	unsigned int rxbv, shift;
569*4882a593Smuzhiyun 	unsigned int rx_fill = rx_fifo_level(spi);
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun 	/*
572*4882a593Smuzhiyun 	 * In RX-only mode the bits per word value is ignored by HW. A value
573*4882a593Smuzhiyun 	 * of 32 is used instead. Thus all 4 bytes per FIFO must be read.
574*4882a593Smuzhiyun 	 * If remaining RX bytes are less than 4, the FIFO must be read
575*4882a593Smuzhiyun 	 * differently. The amount of received and valid bytes is indicated
576*4882a593Smuzhiyun 	 * by STAT.RXBV register value.
577*4882a593Smuzhiyun 	 */
578*4882a593Smuzhiyun 	while (rx_fill) {
579*4882a593Smuzhiyun 		if (spi->rx_todo < 4)  {
580*4882a593Smuzhiyun 			rxbv = (lantiq_ssc_readl(spi, LTQ_SPI_STAT) &
581*4882a593Smuzhiyun 				LTQ_SPI_STAT_RXBV_M) >> LTQ_SPI_STAT_RXBV_S;
582*4882a593Smuzhiyun 			data = lantiq_ssc_readl(spi, LTQ_SPI_RB);
583*4882a593Smuzhiyun 
584*4882a593Smuzhiyun 			shift = (rxbv - 1) * 8;
585*4882a593Smuzhiyun 			rx8 = spi->rx;
586*4882a593Smuzhiyun 
587*4882a593Smuzhiyun 			while (rxbv) {
588*4882a593Smuzhiyun 				*rx8++ = (data >> shift) & 0xFF;
589*4882a593Smuzhiyun 				rxbv--;
590*4882a593Smuzhiyun 				shift -= 8;
591*4882a593Smuzhiyun 				spi->rx_todo--;
592*4882a593Smuzhiyun 				spi->rx++;
593*4882a593Smuzhiyun 			}
594*4882a593Smuzhiyun 		} else {
595*4882a593Smuzhiyun 			data = lantiq_ssc_readl(spi, LTQ_SPI_RB);
596*4882a593Smuzhiyun 			rx32 = (u32 *) spi->rx;
597*4882a593Smuzhiyun 
598*4882a593Smuzhiyun 			*rx32++ = data;
599*4882a593Smuzhiyun 			spi->rx_todo -= 4;
600*4882a593Smuzhiyun 			spi->rx += 4;
601*4882a593Smuzhiyun 		}
602*4882a593Smuzhiyun 		rx_fill--;
603*4882a593Smuzhiyun 	}
604*4882a593Smuzhiyun }
605*4882a593Smuzhiyun 
rx_request(struct lantiq_ssc_spi * spi)606*4882a593Smuzhiyun static void rx_request(struct lantiq_ssc_spi *spi)
607*4882a593Smuzhiyun {
608*4882a593Smuzhiyun 	unsigned int rxreq, rxreq_max;
609*4882a593Smuzhiyun 
610*4882a593Smuzhiyun 	/*
611*4882a593Smuzhiyun 	 * To avoid receive overflows at high clocks it is better to request
612*4882a593Smuzhiyun 	 * only the amount of bytes that fits into all FIFOs. This value
613*4882a593Smuzhiyun 	 * depends on the FIFO size implemented in hardware.
614*4882a593Smuzhiyun 	 */
615*4882a593Smuzhiyun 	rxreq = spi->rx_todo;
616*4882a593Smuzhiyun 	rxreq_max = spi->rx_fifo_size * 4;
617*4882a593Smuzhiyun 	if (rxreq > rxreq_max)
618*4882a593Smuzhiyun 		rxreq = rxreq_max;
619*4882a593Smuzhiyun 
620*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, rxreq, LTQ_SPI_RXREQ);
621*4882a593Smuzhiyun }
622*4882a593Smuzhiyun 
lantiq_ssc_xmit_interrupt(int irq,void * data)623*4882a593Smuzhiyun static irqreturn_t lantiq_ssc_xmit_interrupt(int irq, void *data)
624*4882a593Smuzhiyun {
625*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = data;
626*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
627*4882a593Smuzhiyun 	u32 val = lantiq_ssc_readl(spi, hwcfg->irncr);
628*4882a593Smuzhiyun 
629*4882a593Smuzhiyun 	spin_lock(&spi->lock);
630*4882a593Smuzhiyun 	if (hwcfg->irq_ack)
631*4882a593Smuzhiyun 		lantiq_ssc_writel(spi, val, hwcfg->irncr);
632*4882a593Smuzhiyun 
633*4882a593Smuzhiyun 	if (spi->tx) {
634*4882a593Smuzhiyun 		if (spi->rx && spi->rx_todo)
635*4882a593Smuzhiyun 			rx_fifo_read_full_duplex(spi);
636*4882a593Smuzhiyun 
637*4882a593Smuzhiyun 		if (spi->tx_todo)
638*4882a593Smuzhiyun 			tx_fifo_write(spi);
639*4882a593Smuzhiyun 		else if (!tx_fifo_level(spi))
640*4882a593Smuzhiyun 			goto completed;
641*4882a593Smuzhiyun 	} else if (spi->rx) {
642*4882a593Smuzhiyun 		if (spi->rx_todo) {
643*4882a593Smuzhiyun 			rx_fifo_read_half_duplex(spi);
644*4882a593Smuzhiyun 
645*4882a593Smuzhiyun 			if (spi->rx_todo)
646*4882a593Smuzhiyun 				rx_request(spi);
647*4882a593Smuzhiyun 			else
648*4882a593Smuzhiyun 				goto completed;
649*4882a593Smuzhiyun 		} else {
650*4882a593Smuzhiyun 			goto completed;
651*4882a593Smuzhiyun 		}
652*4882a593Smuzhiyun 	}
653*4882a593Smuzhiyun 
654*4882a593Smuzhiyun 	spin_unlock(&spi->lock);
655*4882a593Smuzhiyun 	return IRQ_HANDLED;
656*4882a593Smuzhiyun 
657*4882a593Smuzhiyun completed:
658*4882a593Smuzhiyun 	queue_work(spi->wq, &spi->work);
659*4882a593Smuzhiyun 	spin_unlock(&spi->lock);
660*4882a593Smuzhiyun 
661*4882a593Smuzhiyun 	return IRQ_HANDLED;
662*4882a593Smuzhiyun }
663*4882a593Smuzhiyun 
lantiq_ssc_err_interrupt(int irq,void * data)664*4882a593Smuzhiyun static irqreturn_t lantiq_ssc_err_interrupt(int irq, void *data)
665*4882a593Smuzhiyun {
666*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = data;
667*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
668*4882a593Smuzhiyun 	u32 stat = lantiq_ssc_readl(spi, LTQ_SPI_STAT);
669*4882a593Smuzhiyun 	u32 val = lantiq_ssc_readl(spi, hwcfg->irncr);
670*4882a593Smuzhiyun 
671*4882a593Smuzhiyun 	if (!(stat & LTQ_SPI_STAT_ERRORS))
672*4882a593Smuzhiyun 		return IRQ_NONE;
673*4882a593Smuzhiyun 
674*4882a593Smuzhiyun 	spin_lock(&spi->lock);
675*4882a593Smuzhiyun 	if (hwcfg->irq_ack)
676*4882a593Smuzhiyun 		lantiq_ssc_writel(spi, val, hwcfg->irncr);
677*4882a593Smuzhiyun 
678*4882a593Smuzhiyun 	if (stat & LTQ_SPI_STAT_RUE)
679*4882a593Smuzhiyun 		dev_err(spi->dev, "receive underflow error\n");
680*4882a593Smuzhiyun 	if (stat & LTQ_SPI_STAT_TUE)
681*4882a593Smuzhiyun 		dev_err(spi->dev, "transmit underflow error\n");
682*4882a593Smuzhiyun 	if (stat & LTQ_SPI_STAT_AE)
683*4882a593Smuzhiyun 		dev_err(spi->dev, "abort error\n");
684*4882a593Smuzhiyun 	if (stat & LTQ_SPI_STAT_RE)
685*4882a593Smuzhiyun 		dev_err(spi->dev, "receive overflow error\n");
686*4882a593Smuzhiyun 	if (stat & LTQ_SPI_STAT_TE)
687*4882a593Smuzhiyun 		dev_err(spi->dev, "transmit overflow error\n");
688*4882a593Smuzhiyun 	if (stat & LTQ_SPI_STAT_ME)
689*4882a593Smuzhiyun 		dev_err(spi->dev, "mode error\n");
690*4882a593Smuzhiyun 
691*4882a593Smuzhiyun 	/* Clear error flags */
692*4882a593Smuzhiyun 	lantiq_ssc_maskl(spi, 0, LTQ_SPI_WHBSTATE_CLR_ERRORS, LTQ_SPI_WHBSTATE);
693*4882a593Smuzhiyun 
694*4882a593Smuzhiyun 	/* set bad status so it can be retried */
695*4882a593Smuzhiyun 	if (spi->master->cur_msg)
696*4882a593Smuzhiyun 		spi->master->cur_msg->status = -EIO;
697*4882a593Smuzhiyun 	queue_work(spi->wq, &spi->work);
698*4882a593Smuzhiyun 	spin_unlock(&spi->lock);
699*4882a593Smuzhiyun 
700*4882a593Smuzhiyun 	return IRQ_HANDLED;
701*4882a593Smuzhiyun }
702*4882a593Smuzhiyun 
intel_lgm_ssc_isr(int irq,void * data)703*4882a593Smuzhiyun static irqreturn_t intel_lgm_ssc_isr(int irq, void *data)
704*4882a593Smuzhiyun {
705*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = data;
706*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
707*4882a593Smuzhiyun 	u32 val = lantiq_ssc_readl(spi, hwcfg->irncr);
708*4882a593Smuzhiyun 
709*4882a593Smuzhiyun 	if (!(val & LTQ_SPI_IRNEN_ALL))
710*4882a593Smuzhiyun 		return IRQ_NONE;
711*4882a593Smuzhiyun 
712*4882a593Smuzhiyun 	if (val & LTQ_SPI_IRNEN_E)
713*4882a593Smuzhiyun 		return lantiq_ssc_err_interrupt(irq, data);
714*4882a593Smuzhiyun 
715*4882a593Smuzhiyun 	if ((val & hwcfg->irnen_t) || (val & hwcfg->irnen_r))
716*4882a593Smuzhiyun 		return lantiq_ssc_xmit_interrupt(irq, data);
717*4882a593Smuzhiyun 
718*4882a593Smuzhiyun 	return IRQ_HANDLED;
719*4882a593Smuzhiyun }
720*4882a593Smuzhiyun 
transfer_start(struct lantiq_ssc_spi * spi,struct spi_device * spidev,struct spi_transfer * t)721*4882a593Smuzhiyun static int transfer_start(struct lantiq_ssc_spi *spi, struct spi_device *spidev,
722*4882a593Smuzhiyun 			  struct spi_transfer *t)
723*4882a593Smuzhiyun {
724*4882a593Smuzhiyun 	unsigned long flags;
725*4882a593Smuzhiyun 
726*4882a593Smuzhiyun 	spin_lock_irqsave(&spi->lock, flags);
727*4882a593Smuzhiyun 
728*4882a593Smuzhiyun 	spi->tx = t->tx_buf;
729*4882a593Smuzhiyun 	spi->rx = t->rx_buf;
730*4882a593Smuzhiyun 
731*4882a593Smuzhiyun 	if (t->tx_buf) {
732*4882a593Smuzhiyun 		spi->tx_todo = t->len;
733*4882a593Smuzhiyun 
734*4882a593Smuzhiyun 		/* initially fill TX FIFO */
735*4882a593Smuzhiyun 		tx_fifo_write(spi);
736*4882a593Smuzhiyun 	}
737*4882a593Smuzhiyun 
738*4882a593Smuzhiyun 	if (spi->rx) {
739*4882a593Smuzhiyun 		spi->rx_todo = t->len;
740*4882a593Smuzhiyun 
741*4882a593Smuzhiyun 		/* start shift clock in RX-only mode */
742*4882a593Smuzhiyun 		if (!spi->tx)
743*4882a593Smuzhiyun 			rx_request(spi);
744*4882a593Smuzhiyun 	}
745*4882a593Smuzhiyun 
746*4882a593Smuzhiyun 	spin_unlock_irqrestore(&spi->lock, flags);
747*4882a593Smuzhiyun 
748*4882a593Smuzhiyun 	return t->len;
749*4882a593Smuzhiyun }
750*4882a593Smuzhiyun 
751*4882a593Smuzhiyun /*
752*4882a593Smuzhiyun  * The driver only gets an interrupt when the FIFO is empty, but there
753*4882a593Smuzhiyun  * is an additional shift register from which the data is written to
754*4882a593Smuzhiyun  * the wire. We get the last interrupt when the controller starts to
755*4882a593Smuzhiyun  * write the last word to the wire, not when it is finished. Do busy
756*4882a593Smuzhiyun  * waiting till it finishes.
757*4882a593Smuzhiyun  */
lantiq_ssc_bussy_work(struct work_struct * work)758*4882a593Smuzhiyun static void lantiq_ssc_bussy_work(struct work_struct *work)
759*4882a593Smuzhiyun {
760*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi;
761*4882a593Smuzhiyun 	unsigned long long timeout = 8LL * 1000LL;
762*4882a593Smuzhiyun 	unsigned long end;
763*4882a593Smuzhiyun 
764*4882a593Smuzhiyun 	spi = container_of(work, typeof(*spi), work);
765*4882a593Smuzhiyun 
766*4882a593Smuzhiyun 	do_div(timeout, spi->speed_hz);
767*4882a593Smuzhiyun 	timeout += timeout + 100; /* some tolerance */
768*4882a593Smuzhiyun 
769*4882a593Smuzhiyun 	end = jiffies + msecs_to_jiffies(timeout);
770*4882a593Smuzhiyun 	do {
771*4882a593Smuzhiyun 		u32 stat = lantiq_ssc_readl(spi, LTQ_SPI_STAT);
772*4882a593Smuzhiyun 
773*4882a593Smuzhiyun 		if (!(stat & LTQ_SPI_STAT_BSY)) {
774*4882a593Smuzhiyun 			spi_finalize_current_transfer(spi->master);
775*4882a593Smuzhiyun 			return;
776*4882a593Smuzhiyun 		}
777*4882a593Smuzhiyun 
778*4882a593Smuzhiyun 		cond_resched();
779*4882a593Smuzhiyun 	} while (!time_after_eq(jiffies, end));
780*4882a593Smuzhiyun 
781*4882a593Smuzhiyun 	if (spi->master->cur_msg)
782*4882a593Smuzhiyun 		spi->master->cur_msg->status = -EIO;
783*4882a593Smuzhiyun 	spi_finalize_current_transfer(spi->master);
784*4882a593Smuzhiyun }
785*4882a593Smuzhiyun 
lantiq_ssc_handle_err(struct spi_master * master,struct spi_message * message)786*4882a593Smuzhiyun static void lantiq_ssc_handle_err(struct spi_master *master,
787*4882a593Smuzhiyun 				  struct spi_message *message)
788*4882a593Smuzhiyun {
789*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
790*4882a593Smuzhiyun 
791*4882a593Smuzhiyun 	/* flush FIFOs on timeout */
792*4882a593Smuzhiyun 	rx_fifo_flush(spi);
793*4882a593Smuzhiyun 	tx_fifo_flush(spi);
794*4882a593Smuzhiyun }
795*4882a593Smuzhiyun 
lantiq_ssc_set_cs(struct spi_device * spidev,bool enable)796*4882a593Smuzhiyun static void lantiq_ssc_set_cs(struct spi_device *spidev, bool enable)
797*4882a593Smuzhiyun {
798*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = spi_master_get_devdata(spidev->master);
799*4882a593Smuzhiyun 	unsigned int cs = spidev->chip_select;
800*4882a593Smuzhiyun 	u32 fgpo;
801*4882a593Smuzhiyun 
802*4882a593Smuzhiyun 	if (!!(spidev->mode & SPI_CS_HIGH) == enable)
803*4882a593Smuzhiyun 		fgpo = (1 << (cs - spi->base_cs));
804*4882a593Smuzhiyun 	else
805*4882a593Smuzhiyun 		fgpo = (1 << (cs - spi->base_cs + LTQ_SPI_FGPO_SETOUTN_S));
806*4882a593Smuzhiyun 
807*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, fgpo, LTQ_SPI_FPGO);
808*4882a593Smuzhiyun }
809*4882a593Smuzhiyun 
lantiq_ssc_transfer_one(struct spi_master * master,struct spi_device * spidev,struct spi_transfer * t)810*4882a593Smuzhiyun static int lantiq_ssc_transfer_one(struct spi_master *master,
811*4882a593Smuzhiyun 				   struct spi_device *spidev,
812*4882a593Smuzhiyun 				   struct spi_transfer *t)
813*4882a593Smuzhiyun {
814*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
815*4882a593Smuzhiyun 
816*4882a593Smuzhiyun 	hw_setup_transfer(spi, spidev, t);
817*4882a593Smuzhiyun 
818*4882a593Smuzhiyun 	return transfer_start(spi, spidev, t);
819*4882a593Smuzhiyun }
820*4882a593Smuzhiyun 
intel_lgm_cfg_irq(struct platform_device * pdev,struct lantiq_ssc_spi * spi)821*4882a593Smuzhiyun static int intel_lgm_cfg_irq(struct platform_device *pdev, struct lantiq_ssc_spi *spi)
822*4882a593Smuzhiyun {
823*4882a593Smuzhiyun 	int irq;
824*4882a593Smuzhiyun 
825*4882a593Smuzhiyun 	irq = platform_get_irq(pdev, 0);
826*4882a593Smuzhiyun 	if (irq < 0)
827*4882a593Smuzhiyun 		return irq;
828*4882a593Smuzhiyun 
829*4882a593Smuzhiyun 	return devm_request_irq(&pdev->dev, irq, intel_lgm_ssc_isr, 0, "spi", spi);
830*4882a593Smuzhiyun }
831*4882a593Smuzhiyun 
lantiq_cfg_irq(struct platform_device * pdev,struct lantiq_ssc_spi * spi)832*4882a593Smuzhiyun static int lantiq_cfg_irq(struct platform_device *pdev, struct lantiq_ssc_spi *spi)
833*4882a593Smuzhiyun {
834*4882a593Smuzhiyun 	int irq, err;
835*4882a593Smuzhiyun 
836*4882a593Smuzhiyun 	irq = platform_get_irq_byname(pdev, LTQ_SPI_RX_IRQ_NAME);
837*4882a593Smuzhiyun 	if (irq < 0)
838*4882a593Smuzhiyun 		return irq;
839*4882a593Smuzhiyun 
840*4882a593Smuzhiyun 	err = devm_request_irq(&pdev->dev, irq, lantiq_ssc_xmit_interrupt,
841*4882a593Smuzhiyun 			       0, LTQ_SPI_RX_IRQ_NAME, spi);
842*4882a593Smuzhiyun 	if (err)
843*4882a593Smuzhiyun 		return err;
844*4882a593Smuzhiyun 
845*4882a593Smuzhiyun 	irq = platform_get_irq_byname(pdev, LTQ_SPI_TX_IRQ_NAME);
846*4882a593Smuzhiyun 	if (irq < 0)
847*4882a593Smuzhiyun 		return irq;
848*4882a593Smuzhiyun 
849*4882a593Smuzhiyun 	err = devm_request_irq(&pdev->dev, irq, lantiq_ssc_xmit_interrupt,
850*4882a593Smuzhiyun 			       0, LTQ_SPI_TX_IRQ_NAME, spi);
851*4882a593Smuzhiyun 
852*4882a593Smuzhiyun 	if (err)
853*4882a593Smuzhiyun 		return err;
854*4882a593Smuzhiyun 
855*4882a593Smuzhiyun 	irq = platform_get_irq_byname(pdev, LTQ_SPI_ERR_IRQ_NAME);
856*4882a593Smuzhiyun 	if (irq < 0)
857*4882a593Smuzhiyun 		return irq;
858*4882a593Smuzhiyun 
859*4882a593Smuzhiyun 	err = devm_request_irq(&pdev->dev, irq, lantiq_ssc_err_interrupt,
860*4882a593Smuzhiyun 			       0, LTQ_SPI_ERR_IRQ_NAME, spi);
861*4882a593Smuzhiyun 	return err;
862*4882a593Smuzhiyun }
863*4882a593Smuzhiyun 
864*4882a593Smuzhiyun static const struct lantiq_ssc_hwcfg lantiq_ssc_xway = {
865*4882a593Smuzhiyun 	.cfg_irq	= lantiq_cfg_irq,
866*4882a593Smuzhiyun 	.irnen_r	= LTQ_SPI_IRNEN_R_XWAY,
867*4882a593Smuzhiyun 	.irnen_t	= LTQ_SPI_IRNEN_T_XWAY,
868*4882a593Smuzhiyun 	.irnicr		= 0xF8,
869*4882a593Smuzhiyun 	.irncr		= 0xFC,
870*4882a593Smuzhiyun 	.fifo_size_mask	= GENMASK(5, 0),
871*4882a593Smuzhiyun 	.irq_ack	= false,
872*4882a593Smuzhiyun };
873*4882a593Smuzhiyun 
874*4882a593Smuzhiyun static const struct lantiq_ssc_hwcfg lantiq_ssc_xrx = {
875*4882a593Smuzhiyun 	.cfg_irq	= lantiq_cfg_irq,
876*4882a593Smuzhiyun 	.irnen_r	= LTQ_SPI_IRNEN_R_XRX,
877*4882a593Smuzhiyun 	.irnen_t	= LTQ_SPI_IRNEN_T_XRX,
878*4882a593Smuzhiyun 	.irnicr		= 0xF8,
879*4882a593Smuzhiyun 	.irncr		= 0xFC,
880*4882a593Smuzhiyun 	.fifo_size_mask	= GENMASK(5, 0),
881*4882a593Smuzhiyun 	.irq_ack	= false,
882*4882a593Smuzhiyun };
883*4882a593Smuzhiyun 
884*4882a593Smuzhiyun static const struct lantiq_ssc_hwcfg intel_ssc_lgm = {
885*4882a593Smuzhiyun 	.cfg_irq	= intel_lgm_cfg_irq,
886*4882a593Smuzhiyun 	.irnen_r	= LTQ_SPI_IRNEN_R_XRX,
887*4882a593Smuzhiyun 	.irnen_t	= LTQ_SPI_IRNEN_T_XRX,
888*4882a593Smuzhiyun 	.irnicr		= 0xFC,
889*4882a593Smuzhiyun 	.irncr		= 0xF8,
890*4882a593Smuzhiyun 	.fifo_size_mask	= GENMASK(7, 0),
891*4882a593Smuzhiyun 	.irq_ack	= true,
892*4882a593Smuzhiyun };
893*4882a593Smuzhiyun 
894*4882a593Smuzhiyun static const struct of_device_id lantiq_ssc_match[] = {
895*4882a593Smuzhiyun 	{ .compatible = "lantiq,ase-spi", .data = &lantiq_ssc_xway, },
896*4882a593Smuzhiyun 	{ .compatible = "lantiq,falcon-spi", .data = &lantiq_ssc_xrx, },
897*4882a593Smuzhiyun 	{ .compatible = "lantiq,xrx100-spi", .data = &lantiq_ssc_xrx, },
898*4882a593Smuzhiyun 	{ .compatible = "intel,lgm-spi", .data = &intel_ssc_lgm, },
899*4882a593Smuzhiyun 	{},
900*4882a593Smuzhiyun };
901*4882a593Smuzhiyun MODULE_DEVICE_TABLE(of, lantiq_ssc_match);
902*4882a593Smuzhiyun 
lantiq_ssc_probe(struct platform_device * pdev)903*4882a593Smuzhiyun static int lantiq_ssc_probe(struct platform_device *pdev)
904*4882a593Smuzhiyun {
905*4882a593Smuzhiyun 	struct device *dev = &pdev->dev;
906*4882a593Smuzhiyun 	struct spi_master *master;
907*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi;
908*4882a593Smuzhiyun 	const struct lantiq_ssc_hwcfg *hwcfg;
909*4882a593Smuzhiyun 	const struct of_device_id *match;
910*4882a593Smuzhiyun 	u32 id, supports_dma, revision;
911*4882a593Smuzhiyun 	unsigned int num_cs;
912*4882a593Smuzhiyun 	int err;
913*4882a593Smuzhiyun 
914*4882a593Smuzhiyun 	match = of_match_device(lantiq_ssc_match, dev);
915*4882a593Smuzhiyun 	if (!match) {
916*4882a593Smuzhiyun 		dev_err(dev, "no device match\n");
917*4882a593Smuzhiyun 		return -EINVAL;
918*4882a593Smuzhiyun 	}
919*4882a593Smuzhiyun 	hwcfg = match->data;
920*4882a593Smuzhiyun 
921*4882a593Smuzhiyun 	master = spi_alloc_master(dev, sizeof(struct lantiq_ssc_spi));
922*4882a593Smuzhiyun 	if (!master)
923*4882a593Smuzhiyun 		return -ENOMEM;
924*4882a593Smuzhiyun 
925*4882a593Smuzhiyun 	spi = spi_master_get_devdata(master);
926*4882a593Smuzhiyun 	spi->master = master;
927*4882a593Smuzhiyun 	spi->dev = dev;
928*4882a593Smuzhiyun 	spi->hwcfg = hwcfg;
929*4882a593Smuzhiyun 	platform_set_drvdata(pdev, spi);
930*4882a593Smuzhiyun 	spi->regbase = devm_platform_ioremap_resource(pdev, 0);
931*4882a593Smuzhiyun 	if (IS_ERR(spi->regbase)) {
932*4882a593Smuzhiyun 		err = PTR_ERR(spi->regbase);
933*4882a593Smuzhiyun 		goto err_master_put;
934*4882a593Smuzhiyun 	}
935*4882a593Smuzhiyun 
936*4882a593Smuzhiyun 	err = hwcfg->cfg_irq(pdev, spi);
937*4882a593Smuzhiyun 	if (err)
938*4882a593Smuzhiyun 		goto err_master_put;
939*4882a593Smuzhiyun 
940*4882a593Smuzhiyun 	spi->spi_clk = devm_clk_get(dev, "gate");
941*4882a593Smuzhiyun 	if (IS_ERR(spi->spi_clk)) {
942*4882a593Smuzhiyun 		err = PTR_ERR(spi->spi_clk);
943*4882a593Smuzhiyun 		goto err_master_put;
944*4882a593Smuzhiyun 	}
945*4882a593Smuzhiyun 	err = clk_prepare_enable(spi->spi_clk);
946*4882a593Smuzhiyun 	if (err)
947*4882a593Smuzhiyun 		goto err_master_put;
948*4882a593Smuzhiyun 
949*4882a593Smuzhiyun 	/*
950*4882a593Smuzhiyun 	 * Use the old clk_get_fpi() function on Lantiq platform, till it
951*4882a593Smuzhiyun 	 * supports common clk.
952*4882a593Smuzhiyun 	 */
953*4882a593Smuzhiyun #if defined(CONFIG_LANTIQ) && !defined(CONFIG_COMMON_CLK)
954*4882a593Smuzhiyun 	spi->fpi_clk = clk_get_fpi();
955*4882a593Smuzhiyun #else
956*4882a593Smuzhiyun 	spi->fpi_clk = clk_get(dev, "freq");
957*4882a593Smuzhiyun #endif
958*4882a593Smuzhiyun 	if (IS_ERR(spi->fpi_clk)) {
959*4882a593Smuzhiyun 		err = PTR_ERR(spi->fpi_clk);
960*4882a593Smuzhiyun 		goto err_clk_disable;
961*4882a593Smuzhiyun 	}
962*4882a593Smuzhiyun 
963*4882a593Smuzhiyun 	num_cs = 8;
964*4882a593Smuzhiyun 	of_property_read_u32(pdev->dev.of_node, "num-cs", &num_cs);
965*4882a593Smuzhiyun 
966*4882a593Smuzhiyun 	spi->base_cs = 1;
967*4882a593Smuzhiyun 	of_property_read_u32(pdev->dev.of_node, "base-cs", &spi->base_cs);
968*4882a593Smuzhiyun 
969*4882a593Smuzhiyun 	spin_lock_init(&spi->lock);
970*4882a593Smuzhiyun 	spi->bits_per_word = 8;
971*4882a593Smuzhiyun 	spi->speed_hz = 0;
972*4882a593Smuzhiyun 
973*4882a593Smuzhiyun 	master->dev.of_node = pdev->dev.of_node;
974*4882a593Smuzhiyun 	master->num_chipselect = num_cs;
975*4882a593Smuzhiyun 	master->use_gpio_descriptors = true;
976*4882a593Smuzhiyun 	master->setup = lantiq_ssc_setup;
977*4882a593Smuzhiyun 	master->set_cs = lantiq_ssc_set_cs;
978*4882a593Smuzhiyun 	master->handle_err = lantiq_ssc_handle_err;
979*4882a593Smuzhiyun 	master->prepare_message = lantiq_ssc_prepare_message;
980*4882a593Smuzhiyun 	master->unprepare_message = lantiq_ssc_unprepare_message;
981*4882a593Smuzhiyun 	master->transfer_one = lantiq_ssc_transfer_one;
982*4882a593Smuzhiyun 	master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LSB_FIRST | SPI_CS_HIGH |
983*4882a593Smuzhiyun 				SPI_LOOP;
984*4882a593Smuzhiyun 	master->bits_per_word_mask = SPI_BPW_RANGE_MASK(2, 8) |
985*4882a593Smuzhiyun 				     SPI_BPW_MASK(16) | SPI_BPW_MASK(32);
986*4882a593Smuzhiyun 
987*4882a593Smuzhiyun 	spi->wq = alloc_ordered_workqueue(dev_name(dev), WQ_MEM_RECLAIM);
988*4882a593Smuzhiyun 	if (!spi->wq) {
989*4882a593Smuzhiyun 		err = -ENOMEM;
990*4882a593Smuzhiyun 		goto err_clk_put;
991*4882a593Smuzhiyun 	}
992*4882a593Smuzhiyun 	INIT_WORK(&spi->work, lantiq_ssc_bussy_work);
993*4882a593Smuzhiyun 
994*4882a593Smuzhiyun 	id = lantiq_ssc_readl(spi, LTQ_SPI_ID);
995*4882a593Smuzhiyun 	spi->tx_fifo_size = (id >> LTQ_SPI_ID_TXFS_S) & hwcfg->fifo_size_mask;
996*4882a593Smuzhiyun 	spi->rx_fifo_size = (id >> LTQ_SPI_ID_RXFS_S) & hwcfg->fifo_size_mask;
997*4882a593Smuzhiyun 	supports_dma = (id & LTQ_SPI_ID_CFG_M) >> LTQ_SPI_ID_CFG_S;
998*4882a593Smuzhiyun 	revision = id & LTQ_SPI_ID_REV_M;
999*4882a593Smuzhiyun 
1000*4882a593Smuzhiyun 	lantiq_ssc_hw_init(spi);
1001*4882a593Smuzhiyun 
1002*4882a593Smuzhiyun 	dev_info(dev,
1003*4882a593Smuzhiyun 		"Lantiq SSC SPI controller (Rev %i, TXFS %u, RXFS %u, DMA %u)\n",
1004*4882a593Smuzhiyun 		revision, spi->tx_fifo_size, spi->rx_fifo_size, supports_dma);
1005*4882a593Smuzhiyun 
1006*4882a593Smuzhiyun 	err = devm_spi_register_master(dev, master);
1007*4882a593Smuzhiyun 	if (err) {
1008*4882a593Smuzhiyun 		dev_err(dev, "failed to register spi_master\n");
1009*4882a593Smuzhiyun 		goto err_wq_destroy;
1010*4882a593Smuzhiyun 	}
1011*4882a593Smuzhiyun 
1012*4882a593Smuzhiyun 	return 0;
1013*4882a593Smuzhiyun 
1014*4882a593Smuzhiyun err_wq_destroy:
1015*4882a593Smuzhiyun 	destroy_workqueue(spi->wq);
1016*4882a593Smuzhiyun err_clk_put:
1017*4882a593Smuzhiyun 	clk_put(spi->fpi_clk);
1018*4882a593Smuzhiyun err_clk_disable:
1019*4882a593Smuzhiyun 	clk_disable_unprepare(spi->spi_clk);
1020*4882a593Smuzhiyun err_master_put:
1021*4882a593Smuzhiyun 	spi_master_put(master);
1022*4882a593Smuzhiyun 
1023*4882a593Smuzhiyun 	return err;
1024*4882a593Smuzhiyun }
1025*4882a593Smuzhiyun 
lantiq_ssc_remove(struct platform_device * pdev)1026*4882a593Smuzhiyun static int lantiq_ssc_remove(struct platform_device *pdev)
1027*4882a593Smuzhiyun {
1028*4882a593Smuzhiyun 	struct lantiq_ssc_spi *spi = platform_get_drvdata(pdev);
1029*4882a593Smuzhiyun 
1030*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, 0, LTQ_SPI_IRNEN);
1031*4882a593Smuzhiyun 	lantiq_ssc_writel(spi, 0, LTQ_SPI_CLC);
1032*4882a593Smuzhiyun 	rx_fifo_flush(spi);
1033*4882a593Smuzhiyun 	tx_fifo_flush(spi);
1034*4882a593Smuzhiyun 	hw_enter_config_mode(spi);
1035*4882a593Smuzhiyun 
1036*4882a593Smuzhiyun 	destroy_workqueue(spi->wq);
1037*4882a593Smuzhiyun 	clk_disable_unprepare(spi->spi_clk);
1038*4882a593Smuzhiyun 	clk_put(spi->fpi_clk);
1039*4882a593Smuzhiyun 
1040*4882a593Smuzhiyun 	return 0;
1041*4882a593Smuzhiyun }
1042*4882a593Smuzhiyun 
1043*4882a593Smuzhiyun static struct platform_driver lantiq_ssc_driver = {
1044*4882a593Smuzhiyun 	.probe = lantiq_ssc_probe,
1045*4882a593Smuzhiyun 	.remove = lantiq_ssc_remove,
1046*4882a593Smuzhiyun 	.driver = {
1047*4882a593Smuzhiyun 		.name = "spi-lantiq-ssc",
1048*4882a593Smuzhiyun 		.of_match_table = lantiq_ssc_match,
1049*4882a593Smuzhiyun 	},
1050*4882a593Smuzhiyun };
1051*4882a593Smuzhiyun module_platform_driver(lantiq_ssc_driver);
1052*4882a593Smuzhiyun 
1053*4882a593Smuzhiyun MODULE_DESCRIPTION("Lantiq SSC SPI controller driver");
1054*4882a593Smuzhiyun MODULE_AUTHOR("Daniel Schwierzeck <daniel.schwierzeck@gmail.com>");
1055*4882a593Smuzhiyun MODULE_AUTHOR("Hauke Mehrtens <hauke@hauke-m.de>");
1056*4882a593Smuzhiyun MODULE_LICENSE("GPL");
1057*4882a593Smuzhiyun MODULE_ALIAS("platform:spi-lantiq-ssc");
1058