1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun // Copyright (C) 2018 Spreadtrum Communications Inc.
3*4882a593Smuzhiyun
4*4882a593Smuzhiyun #include <linux/clk.h>
5*4882a593Smuzhiyun #include <linux/dmaengine.h>
6*4882a593Smuzhiyun #include <linux/dma-mapping.h>
7*4882a593Smuzhiyun #include <linux/dma/sprd-dma.h>
8*4882a593Smuzhiyun #include <linux/interrupt.h>
9*4882a593Smuzhiyun #include <linux/io.h>
10*4882a593Smuzhiyun #include <linux/iopoll.h>
11*4882a593Smuzhiyun #include <linux/kernel.h>
12*4882a593Smuzhiyun #include <linux/module.h>
13*4882a593Smuzhiyun #include <linux/of.h>
14*4882a593Smuzhiyun #include <linux/of_device.h>
15*4882a593Smuzhiyun #include <linux/of_dma.h>
16*4882a593Smuzhiyun #include <linux/platform_device.h>
17*4882a593Smuzhiyun #include <linux/pm_runtime.h>
18*4882a593Smuzhiyun #include <linux/spi/spi.h>
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun #define SPRD_SPI_TXD 0x0
21*4882a593Smuzhiyun #define SPRD_SPI_CLKD 0x4
22*4882a593Smuzhiyun #define SPRD_SPI_CTL0 0x8
23*4882a593Smuzhiyun #define SPRD_SPI_CTL1 0xc
24*4882a593Smuzhiyun #define SPRD_SPI_CTL2 0x10
25*4882a593Smuzhiyun #define SPRD_SPI_CTL3 0x14
26*4882a593Smuzhiyun #define SPRD_SPI_CTL4 0x18
27*4882a593Smuzhiyun #define SPRD_SPI_CTL5 0x1c
28*4882a593Smuzhiyun #define SPRD_SPI_INT_EN 0x20
29*4882a593Smuzhiyun #define SPRD_SPI_INT_CLR 0x24
30*4882a593Smuzhiyun #define SPRD_SPI_INT_RAW_STS 0x28
31*4882a593Smuzhiyun #define SPRD_SPI_INT_MASK_STS 0x2c
32*4882a593Smuzhiyun #define SPRD_SPI_STS1 0x30
33*4882a593Smuzhiyun #define SPRD_SPI_STS2 0x34
34*4882a593Smuzhiyun #define SPRD_SPI_DSP_WAIT 0x38
35*4882a593Smuzhiyun #define SPRD_SPI_STS3 0x3c
36*4882a593Smuzhiyun #define SPRD_SPI_CTL6 0x40
37*4882a593Smuzhiyun #define SPRD_SPI_STS4 0x44
38*4882a593Smuzhiyun #define SPRD_SPI_FIFO_RST 0x48
39*4882a593Smuzhiyun #define SPRD_SPI_CTL7 0x4c
40*4882a593Smuzhiyun #define SPRD_SPI_STS5 0x50
41*4882a593Smuzhiyun #define SPRD_SPI_CTL8 0x54
42*4882a593Smuzhiyun #define SPRD_SPI_CTL9 0x58
43*4882a593Smuzhiyun #define SPRD_SPI_CTL10 0x5c
44*4882a593Smuzhiyun #define SPRD_SPI_CTL11 0x60
45*4882a593Smuzhiyun #define SPRD_SPI_CTL12 0x64
46*4882a593Smuzhiyun #define SPRD_SPI_STS6 0x68
47*4882a593Smuzhiyun #define SPRD_SPI_STS7 0x6c
48*4882a593Smuzhiyun #define SPRD_SPI_STS8 0x70
49*4882a593Smuzhiyun #define SPRD_SPI_STS9 0x74
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun /* Bits & mask definition for register CTL0 */
52*4882a593Smuzhiyun #define SPRD_SPI_SCK_REV BIT(13)
53*4882a593Smuzhiyun #define SPRD_SPI_NG_TX BIT(1)
54*4882a593Smuzhiyun #define SPRD_SPI_NG_RX BIT(0)
55*4882a593Smuzhiyun #define SPRD_SPI_CHNL_LEN_MASK GENMASK(4, 0)
56*4882a593Smuzhiyun #define SPRD_SPI_CSN_MASK GENMASK(11, 8)
57*4882a593Smuzhiyun #define SPRD_SPI_CS0_VALID BIT(8)
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun /* Bits & mask definition for register SPI_INT_EN */
60*4882a593Smuzhiyun #define SPRD_SPI_TX_END_INT_EN BIT(8)
61*4882a593Smuzhiyun #define SPRD_SPI_RX_END_INT_EN BIT(9)
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun /* Bits & mask definition for register SPI_INT_RAW_STS */
64*4882a593Smuzhiyun #define SPRD_SPI_TX_END_RAW BIT(8)
65*4882a593Smuzhiyun #define SPRD_SPI_RX_END_RAW BIT(9)
66*4882a593Smuzhiyun
67*4882a593Smuzhiyun /* Bits & mask definition for register SPI_INT_CLR */
68*4882a593Smuzhiyun #define SPRD_SPI_TX_END_CLR BIT(8)
69*4882a593Smuzhiyun #define SPRD_SPI_RX_END_CLR BIT(9)
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun /* Bits & mask definition for register INT_MASK_STS */
72*4882a593Smuzhiyun #define SPRD_SPI_MASK_RX_END BIT(9)
73*4882a593Smuzhiyun #define SPRD_SPI_MASK_TX_END BIT(8)
74*4882a593Smuzhiyun
75*4882a593Smuzhiyun /* Bits & mask definition for register STS2 */
76*4882a593Smuzhiyun #define SPRD_SPI_TX_BUSY BIT(8)
77*4882a593Smuzhiyun
78*4882a593Smuzhiyun /* Bits & mask definition for register CTL1 */
79*4882a593Smuzhiyun #define SPRD_SPI_RX_MODE BIT(12)
80*4882a593Smuzhiyun #define SPRD_SPI_TX_MODE BIT(13)
81*4882a593Smuzhiyun #define SPRD_SPI_RTX_MD_MASK GENMASK(13, 12)
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun /* Bits & mask definition for register CTL2 */
84*4882a593Smuzhiyun #define SPRD_SPI_DMA_EN BIT(6)
85*4882a593Smuzhiyun
86*4882a593Smuzhiyun /* Bits & mask definition for register CTL4 */
87*4882a593Smuzhiyun #define SPRD_SPI_START_RX BIT(9)
88*4882a593Smuzhiyun #define SPRD_SPI_ONLY_RECV_MASK GENMASK(8, 0)
89*4882a593Smuzhiyun
90*4882a593Smuzhiyun /* Bits & mask definition for register SPI_INT_CLR */
91*4882a593Smuzhiyun #define SPRD_SPI_RX_END_INT_CLR BIT(9)
92*4882a593Smuzhiyun #define SPRD_SPI_TX_END_INT_CLR BIT(8)
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun /* Bits & mask definition for register SPI_INT_RAW */
95*4882a593Smuzhiyun #define SPRD_SPI_RX_END_IRQ BIT(9)
96*4882a593Smuzhiyun #define SPRD_SPI_TX_END_IRQ BIT(8)
97*4882a593Smuzhiyun
98*4882a593Smuzhiyun /* Bits & mask definition for register CTL12 */
99*4882a593Smuzhiyun #define SPRD_SPI_SW_RX_REQ BIT(0)
100*4882a593Smuzhiyun #define SPRD_SPI_SW_TX_REQ BIT(1)
101*4882a593Smuzhiyun
102*4882a593Smuzhiyun /* Bits & mask definition for register CTL7 */
103*4882a593Smuzhiyun #define SPRD_SPI_DATA_LINE2_EN BIT(15)
104*4882a593Smuzhiyun #define SPRD_SPI_MODE_MASK GENMASK(5, 3)
105*4882a593Smuzhiyun #define SPRD_SPI_MODE_OFFSET 3
106*4882a593Smuzhiyun #define SPRD_SPI_3WIRE_MODE 4
107*4882a593Smuzhiyun #define SPRD_SPI_4WIRE_MODE 0
108*4882a593Smuzhiyun
109*4882a593Smuzhiyun /* Bits & mask definition for register CTL8 */
110*4882a593Smuzhiyun #define SPRD_SPI_TX_MAX_LEN_MASK GENMASK(19, 0)
111*4882a593Smuzhiyun #define SPRD_SPI_TX_LEN_H_MASK GENMASK(3, 0)
112*4882a593Smuzhiyun #define SPRD_SPI_TX_LEN_H_OFFSET 16
113*4882a593Smuzhiyun
114*4882a593Smuzhiyun /* Bits & mask definition for register CTL9 */
115*4882a593Smuzhiyun #define SPRD_SPI_TX_LEN_L_MASK GENMASK(15, 0)
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun /* Bits & mask definition for register CTL10 */
118*4882a593Smuzhiyun #define SPRD_SPI_RX_MAX_LEN_MASK GENMASK(19, 0)
119*4882a593Smuzhiyun #define SPRD_SPI_RX_LEN_H_MASK GENMASK(3, 0)
120*4882a593Smuzhiyun #define SPRD_SPI_RX_LEN_H_OFFSET 16
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun /* Bits & mask definition for register CTL11 */
123*4882a593Smuzhiyun #define SPRD_SPI_RX_LEN_L_MASK GENMASK(15, 0)
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun /* Default & maximum word delay cycles */
126*4882a593Smuzhiyun #define SPRD_SPI_MIN_DELAY_CYCLE 14
127*4882a593Smuzhiyun #define SPRD_SPI_MAX_DELAY_CYCLE 130
128*4882a593Smuzhiyun
129*4882a593Smuzhiyun #define SPRD_SPI_FIFO_SIZE 32
130*4882a593Smuzhiyun #define SPRD_SPI_CHIP_CS_NUM 0x4
131*4882a593Smuzhiyun #define SPRD_SPI_CHNL_LEN 2
132*4882a593Smuzhiyun #define SPRD_SPI_DEFAULT_SOURCE 26000000
133*4882a593Smuzhiyun #define SPRD_SPI_MAX_SPEED_HZ 48000000
134*4882a593Smuzhiyun #define SPRD_SPI_AUTOSUSPEND_DELAY 100
135*4882a593Smuzhiyun #define SPRD_SPI_DMA_STEP 8
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun enum sprd_spi_dma_channel {
138*4882a593Smuzhiyun SPRD_SPI_RX,
139*4882a593Smuzhiyun SPRD_SPI_TX,
140*4882a593Smuzhiyun SPRD_SPI_MAX,
141*4882a593Smuzhiyun };
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun struct sprd_spi_dma {
144*4882a593Smuzhiyun bool enable;
145*4882a593Smuzhiyun struct dma_chan *dma_chan[SPRD_SPI_MAX];
146*4882a593Smuzhiyun enum dma_slave_buswidth width;
147*4882a593Smuzhiyun u32 fragmens_len;
148*4882a593Smuzhiyun u32 rx_len;
149*4882a593Smuzhiyun };
150*4882a593Smuzhiyun
151*4882a593Smuzhiyun struct sprd_spi {
152*4882a593Smuzhiyun void __iomem *base;
153*4882a593Smuzhiyun phys_addr_t phy_base;
154*4882a593Smuzhiyun struct device *dev;
155*4882a593Smuzhiyun struct clk *clk;
156*4882a593Smuzhiyun int irq;
157*4882a593Smuzhiyun u32 src_clk;
158*4882a593Smuzhiyun u32 hw_mode;
159*4882a593Smuzhiyun u32 trans_len;
160*4882a593Smuzhiyun u32 trans_mode;
161*4882a593Smuzhiyun u32 word_delay;
162*4882a593Smuzhiyun u32 hw_speed_hz;
163*4882a593Smuzhiyun u32 len;
164*4882a593Smuzhiyun int status;
165*4882a593Smuzhiyun struct sprd_spi_dma dma;
166*4882a593Smuzhiyun struct completion xfer_completion;
167*4882a593Smuzhiyun const void *tx_buf;
168*4882a593Smuzhiyun void *rx_buf;
169*4882a593Smuzhiyun int (*read_bufs)(struct sprd_spi *ss, u32 len);
170*4882a593Smuzhiyun int (*write_bufs)(struct sprd_spi *ss, u32 len);
171*4882a593Smuzhiyun };
172*4882a593Smuzhiyun
sprd_spi_transfer_max_timeout(struct sprd_spi * ss,struct spi_transfer * t)173*4882a593Smuzhiyun static u32 sprd_spi_transfer_max_timeout(struct sprd_spi *ss,
174*4882a593Smuzhiyun struct spi_transfer *t)
175*4882a593Smuzhiyun {
176*4882a593Smuzhiyun /*
177*4882a593Smuzhiyun * The time spent on transmission of the full FIFO data is the maximum
178*4882a593Smuzhiyun * SPI transmission time.
179*4882a593Smuzhiyun */
180*4882a593Smuzhiyun u32 size = t->bits_per_word * SPRD_SPI_FIFO_SIZE;
181*4882a593Smuzhiyun u32 bit_time_us = DIV_ROUND_UP(USEC_PER_SEC, ss->hw_speed_hz);
182*4882a593Smuzhiyun u32 total_time_us = size * bit_time_us;
183*4882a593Smuzhiyun /*
184*4882a593Smuzhiyun * There is an interval between data and the data in our SPI hardware,
185*4882a593Smuzhiyun * so the total transmission time need add the interval time.
186*4882a593Smuzhiyun */
187*4882a593Smuzhiyun u32 interval_cycle = SPRD_SPI_FIFO_SIZE * ss->word_delay;
188*4882a593Smuzhiyun u32 interval_time_us = DIV_ROUND_UP(interval_cycle * USEC_PER_SEC,
189*4882a593Smuzhiyun ss->src_clk);
190*4882a593Smuzhiyun
191*4882a593Smuzhiyun return total_time_us + interval_time_us;
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun
sprd_spi_wait_for_tx_end(struct sprd_spi * ss,struct spi_transfer * t)194*4882a593Smuzhiyun static int sprd_spi_wait_for_tx_end(struct sprd_spi *ss, struct spi_transfer *t)
195*4882a593Smuzhiyun {
196*4882a593Smuzhiyun u32 val, us;
197*4882a593Smuzhiyun int ret;
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun us = sprd_spi_transfer_max_timeout(ss, t);
200*4882a593Smuzhiyun ret = readl_relaxed_poll_timeout(ss->base + SPRD_SPI_INT_RAW_STS, val,
201*4882a593Smuzhiyun val & SPRD_SPI_TX_END_IRQ, 0, us);
202*4882a593Smuzhiyun if (ret) {
203*4882a593Smuzhiyun dev_err(ss->dev, "SPI error, spi send timeout!\n");
204*4882a593Smuzhiyun return ret;
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun ret = readl_relaxed_poll_timeout(ss->base + SPRD_SPI_STS2, val,
208*4882a593Smuzhiyun !(val & SPRD_SPI_TX_BUSY), 0, us);
209*4882a593Smuzhiyun if (ret) {
210*4882a593Smuzhiyun dev_err(ss->dev, "SPI error, spi busy timeout!\n");
211*4882a593Smuzhiyun return ret;
212*4882a593Smuzhiyun }
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_TX_END_INT_CLR, ss->base + SPRD_SPI_INT_CLR);
215*4882a593Smuzhiyun
216*4882a593Smuzhiyun return 0;
217*4882a593Smuzhiyun }
218*4882a593Smuzhiyun
sprd_spi_wait_for_rx_end(struct sprd_spi * ss,struct spi_transfer * t)219*4882a593Smuzhiyun static int sprd_spi_wait_for_rx_end(struct sprd_spi *ss, struct spi_transfer *t)
220*4882a593Smuzhiyun {
221*4882a593Smuzhiyun u32 val, us;
222*4882a593Smuzhiyun int ret;
223*4882a593Smuzhiyun
224*4882a593Smuzhiyun us = sprd_spi_transfer_max_timeout(ss, t);
225*4882a593Smuzhiyun ret = readl_relaxed_poll_timeout(ss->base + SPRD_SPI_INT_RAW_STS, val,
226*4882a593Smuzhiyun val & SPRD_SPI_RX_END_IRQ, 0, us);
227*4882a593Smuzhiyun if (ret) {
228*4882a593Smuzhiyun dev_err(ss->dev, "SPI error, spi rx timeout!\n");
229*4882a593Smuzhiyun return ret;
230*4882a593Smuzhiyun }
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_RX_END_INT_CLR, ss->base + SPRD_SPI_INT_CLR);
233*4882a593Smuzhiyun
234*4882a593Smuzhiyun return 0;
235*4882a593Smuzhiyun }
236*4882a593Smuzhiyun
sprd_spi_tx_req(struct sprd_spi * ss)237*4882a593Smuzhiyun static void sprd_spi_tx_req(struct sprd_spi *ss)
238*4882a593Smuzhiyun {
239*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_SW_TX_REQ, ss->base + SPRD_SPI_CTL12);
240*4882a593Smuzhiyun }
241*4882a593Smuzhiyun
sprd_spi_rx_req(struct sprd_spi * ss)242*4882a593Smuzhiyun static void sprd_spi_rx_req(struct sprd_spi *ss)
243*4882a593Smuzhiyun {
244*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_SW_RX_REQ, ss->base + SPRD_SPI_CTL12);
245*4882a593Smuzhiyun }
246*4882a593Smuzhiyun
sprd_spi_enter_idle(struct sprd_spi * ss)247*4882a593Smuzhiyun static void sprd_spi_enter_idle(struct sprd_spi *ss)
248*4882a593Smuzhiyun {
249*4882a593Smuzhiyun u32 val = readl_relaxed(ss->base + SPRD_SPI_CTL1);
250*4882a593Smuzhiyun
251*4882a593Smuzhiyun val &= ~SPRD_SPI_RTX_MD_MASK;
252*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL1);
253*4882a593Smuzhiyun }
254*4882a593Smuzhiyun
sprd_spi_set_transfer_bits(struct sprd_spi * ss,u32 bits)255*4882a593Smuzhiyun static void sprd_spi_set_transfer_bits(struct sprd_spi *ss, u32 bits)
256*4882a593Smuzhiyun {
257*4882a593Smuzhiyun u32 val = readl_relaxed(ss->base + SPRD_SPI_CTL0);
258*4882a593Smuzhiyun
259*4882a593Smuzhiyun /* Set the valid bits for every transaction */
260*4882a593Smuzhiyun val &= ~(SPRD_SPI_CHNL_LEN_MASK << SPRD_SPI_CHNL_LEN);
261*4882a593Smuzhiyun val |= bits << SPRD_SPI_CHNL_LEN;
262*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL0);
263*4882a593Smuzhiyun }
264*4882a593Smuzhiyun
sprd_spi_set_tx_length(struct sprd_spi * ss,u32 length)265*4882a593Smuzhiyun static void sprd_spi_set_tx_length(struct sprd_spi *ss, u32 length)
266*4882a593Smuzhiyun {
267*4882a593Smuzhiyun u32 val = readl_relaxed(ss->base + SPRD_SPI_CTL8);
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun length &= SPRD_SPI_TX_MAX_LEN_MASK;
270*4882a593Smuzhiyun val &= ~SPRD_SPI_TX_LEN_H_MASK;
271*4882a593Smuzhiyun val |= length >> SPRD_SPI_TX_LEN_H_OFFSET;
272*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL8);
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun val = length & SPRD_SPI_TX_LEN_L_MASK;
275*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL9);
276*4882a593Smuzhiyun }
277*4882a593Smuzhiyun
sprd_spi_set_rx_length(struct sprd_spi * ss,u32 length)278*4882a593Smuzhiyun static void sprd_spi_set_rx_length(struct sprd_spi *ss, u32 length)
279*4882a593Smuzhiyun {
280*4882a593Smuzhiyun u32 val = readl_relaxed(ss->base + SPRD_SPI_CTL10);
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun length &= SPRD_SPI_RX_MAX_LEN_MASK;
283*4882a593Smuzhiyun val &= ~SPRD_SPI_RX_LEN_H_MASK;
284*4882a593Smuzhiyun val |= length >> SPRD_SPI_RX_LEN_H_OFFSET;
285*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL10);
286*4882a593Smuzhiyun
287*4882a593Smuzhiyun val = length & SPRD_SPI_RX_LEN_L_MASK;
288*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL11);
289*4882a593Smuzhiyun }
290*4882a593Smuzhiyun
sprd_spi_chipselect(struct spi_device * sdev,bool cs)291*4882a593Smuzhiyun static void sprd_spi_chipselect(struct spi_device *sdev, bool cs)
292*4882a593Smuzhiyun {
293*4882a593Smuzhiyun struct spi_controller *sctlr = sdev->controller;
294*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sctlr);
295*4882a593Smuzhiyun u32 val;
296*4882a593Smuzhiyun
297*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL0);
298*4882a593Smuzhiyun /* The SPI controller will pull down CS pin if cs is 0 */
299*4882a593Smuzhiyun if (!cs) {
300*4882a593Smuzhiyun val &= ~SPRD_SPI_CS0_VALID;
301*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL0);
302*4882a593Smuzhiyun } else {
303*4882a593Smuzhiyun val |= SPRD_SPI_CSN_MASK;
304*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL0);
305*4882a593Smuzhiyun }
306*4882a593Smuzhiyun }
307*4882a593Smuzhiyun
sprd_spi_write_only_receive(struct sprd_spi * ss,u32 len)308*4882a593Smuzhiyun static int sprd_spi_write_only_receive(struct sprd_spi *ss, u32 len)
309*4882a593Smuzhiyun {
310*4882a593Smuzhiyun u32 val;
311*4882a593Smuzhiyun
312*4882a593Smuzhiyun /* Clear the start receive bit and reset receive data number */
313*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL4);
314*4882a593Smuzhiyun val &= ~(SPRD_SPI_START_RX | SPRD_SPI_ONLY_RECV_MASK);
315*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL4);
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun /* Set the receive data length */
318*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL4);
319*4882a593Smuzhiyun val |= len & SPRD_SPI_ONLY_RECV_MASK;
320*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL4);
321*4882a593Smuzhiyun
322*4882a593Smuzhiyun /* Trigger to receive data */
323*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL4);
324*4882a593Smuzhiyun val |= SPRD_SPI_START_RX;
325*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL4);
326*4882a593Smuzhiyun
327*4882a593Smuzhiyun return len;
328*4882a593Smuzhiyun }
329*4882a593Smuzhiyun
sprd_spi_write_bufs_u8(struct sprd_spi * ss,u32 len)330*4882a593Smuzhiyun static int sprd_spi_write_bufs_u8(struct sprd_spi *ss, u32 len)
331*4882a593Smuzhiyun {
332*4882a593Smuzhiyun u8 *tx_p = (u8 *)ss->tx_buf;
333*4882a593Smuzhiyun int i;
334*4882a593Smuzhiyun
335*4882a593Smuzhiyun for (i = 0; i < len; i++)
336*4882a593Smuzhiyun writeb_relaxed(tx_p[i], ss->base + SPRD_SPI_TXD);
337*4882a593Smuzhiyun
338*4882a593Smuzhiyun ss->tx_buf += i;
339*4882a593Smuzhiyun return i;
340*4882a593Smuzhiyun }
341*4882a593Smuzhiyun
sprd_spi_write_bufs_u16(struct sprd_spi * ss,u32 len)342*4882a593Smuzhiyun static int sprd_spi_write_bufs_u16(struct sprd_spi *ss, u32 len)
343*4882a593Smuzhiyun {
344*4882a593Smuzhiyun u16 *tx_p = (u16 *)ss->tx_buf;
345*4882a593Smuzhiyun int i;
346*4882a593Smuzhiyun
347*4882a593Smuzhiyun for (i = 0; i < len; i++)
348*4882a593Smuzhiyun writew_relaxed(tx_p[i], ss->base + SPRD_SPI_TXD);
349*4882a593Smuzhiyun
350*4882a593Smuzhiyun ss->tx_buf += i << 1;
351*4882a593Smuzhiyun return i << 1;
352*4882a593Smuzhiyun }
353*4882a593Smuzhiyun
sprd_spi_write_bufs_u32(struct sprd_spi * ss,u32 len)354*4882a593Smuzhiyun static int sprd_spi_write_bufs_u32(struct sprd_spi *ss, u32 len)
355*4882a593Smuzhiyun {
356*4882a593Smuzhiyun u32 *tx_p = (u32 *)ss->tx_buf;
357*4882a593Smuzhiyun int i;
358*4882a593Smuzhiyun
359*4882a593Smuzhiyun for (i = 0; i < len; i++)
360*4882a593Smuzhiyun writel_relaxed(tx_p[i], ss->base + SPRD_SPI_TXD);
361*4882a593Smuzhiyun
362*4882a593Smuzhiyun ss->tx_buf += i << 2;
363*4882a593Smuzhiyun return i << 2;
364*4882a593Smuzhiyun }
365*4882a593Smuzhiyun
sprd_spi_read_bufs_u8(struct sprd_spi * ss,u32 len)366*4882a593Smuzhiyun static int sprd_spi_read_bufs_u8(struct sprd_spi *ss, u32 len)
367*4882a593Smuzhiyun {
368*4882a593Smuzhiyun u8 *rx_p = (u8 *)ss->rx_buf;
369*4882a593Smuzhiyun int i;
370*4882a593Smuzhiyun
371*4882a593Smuzhiyun for (i = 0; i < len; i++)
372*4882a593Smuzhiyun rx_p[i] = readb_relaxed(ss->base + SPRD_SPI_TXD);
373*4882a593Smuzhiyun
374*4882a593Smuzhiyun ss->rx_buf += i;
375*4882a593Smuzhiyun return i;
376*4882a593Smuzhiyun }
377*4882a593Smuzhiyun
sprd_spi_read_bufs_u16(struct sprd_spi * ss,u32 len)378*4882a593Smuzhiyun static int sprd_spi_read_bufs_u16(struct sprd_spi *ss, u32 len)
379*4882a593Smuzhiyun {
380*4882a593Smuzhiyun u16 *rx_p = (u16 *)ss->rx_buf;
381*4882a593Smuzhiyun int i;
382*4882a593Smuzhiyun
383*4882a593Smuzhiyun for (i = 0; i < len; i++)
384*4882a593Smuzhiyun rx_p[i] = readw_relaxed(ss->base + SPRD_SPI_TXD);
385*4882a593Smuzhiyun
386*4882a593Smuzhiyun ss->rx_buf += i << 1;
387*4882a593Smuzhiyun return i << 1;
388*4882a593Smuzhiyun }
389*4882a593Smuzhiyun
sprd_spi_read_bufs_u32(struct sprd_spi * ss,u32 len)390*4882a593Smuzhiyun static int sprd_spi_read_bufs_u32(struct sprd_spi *ss, u32 len)
391*4882a593Smuzhiyun {
392*4882a593Smuzhiyun u32 *rx_p = (u32 *)ss->rx_buf;
393*4882a593Smuzhiyun int i;
394*4882a593Smuzhiyun
395*4882a593Smuzhiyun for (i = 0; i < len; i++)
396*4882a593Smuzhiyun rx_p[i] = readl_relaxed(ss->base + SPRD_SPI_TXD);
397*4882a593Smuzhiyun
398*4882a593Smuzhiyun ss->rx_buf += i << 2;
399*4882a593Smuzhiyun return i << 2;
400*4882a593Smuzhiyun }
401*4882a593Smuzhiyun
sprd_spi_txrx_bufs(struct spi_device * sdev,struct spi_transfer * t)402*4882a593Smuzhiyun static int sprd_spi_txrx_bufs(struct spi_device *sdev, struct spi_transfer *t)
403*4882a593Smuzhiyun {
404*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sdev->controller);
405*4882a593Smuzhiyun u32 trans_len = ss->trans_len, len;
406*4882a593Smuzhiyun int ret, write_size = 0, read_size = 0;
407*4882a593Smuzhiyun
408*4882a593Smuzhiyun while (trans_len) {
409*4882a593Smuzhiyun len = trans_len > SPRD_SPI_FIFO_SIZE ? SPRD_SPI_FIFO_SIZE :
410*4882a593Smuzhiyun trans_len;
411*4882a593Smuzhiyun if (ss->trans_mode & SPRD_SPI_TX_MODE) {
412*4882a593Smuzhiyun sprd_spi_set_tx_length(ss, len);
413*4882a593Smuzhiyun write_size += ss->write_bufs(ss, len);
414*4882a593Smuzhiyun
415*4882a593Smuzhiyun /*
416*4882a593Smuzhiyun * For our 3 wires mode or dual TX line mode, we need
417*4882a593Smuzhiyun * to request the controller to transfer.
418*4882a593Smuzhiyun */
419*4882a593Smuzhiyun if (ss->hw_mode & SPI_3WIRE || ss->hw_mode & SPI_TX_DUAL)
420*4882a593Smuzhiyun sprd_spi_tx_req(ss);
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun ret = sprd_spi_wait_for_tx_end(ss, t);
423*4882a593Smuzhiyun } else {
424*4882a593Smuzhiyun sprd_spi_set_rx_length(ss, len);
425*4882a593Smuzhiyun
426*4882a593Smuzhiyun /*
427*4882a593Smuzhiyun * For our 3 wires mode or dual TX line mode, we need
428*4882a593Smuzhiyun * to request the controller to read.
429*4882a593Smuzhiyun */
430*4882a593Smuzhiyun if (ss->hw_mode & SPI_3WIRE || ss->hw_mode & SPI_TX_DUAL)
431*4882a593Smuzhiyun sprd_spi_rx_req(ss);
432*4882a593Smuzhiyun else
433*4882a593Smuzhiyun write_size += ss->write_bufs(ss, len);
434*4882a593Smuzhiyun
435*4882a593Smuzhiyun ret = sprd_spi_wait_for_rx_end(ss, t);
436*4882a593Smuzhiyun }
437*4882a593Smuzhiyun
438*4882a593Smuzhiyun if (ret)
439*4882a593Smuzhiyun goto complete;
440*4882a593Smuzhiyun
441*4882a593Smuzhiyun if (ss->trans_mode & SPRD_SPI_RX_MODE)
442*4882a593Smuzhiyun read_size += ss->read_bufs(ss, len);
443*4882a593Smuzhiyun
444*4882a593Smuzhiyun trans_len -= len;
445*4882a593Smuzhiyun }
446*4882a593Smuzhiyun
447*4882a593Smuzhiyun if (ss->trans_mode & SPRD_SPI_TX_MODE)
448*4882a593Smuzhiyun ret = write_size;
449*4882a593Smuzhiyun else
450*4882a593Smuzhiyun ret = read_size;
451*4882a593Smuzhiyun complete:
452*4882a593Smuzhiyun sprd_spi_enter_idle(ss);
453*4882a593Smuzhiyun
454*4882a593Smuzhiyun return ret;
455*4882a593Smuzhiyun }
456*4882a593Smuzhiyun
sprd_spi_irq_enable(struct sprd_spi * ss)457*4882a593Smuzhiyun static void sprd_spi_irq_enable(struct sprd_spi *ss)
458*4882a593Smuzhiyun {
459*4882a593Smuzhiyun u32 val;
460*4882a593Smuzhiyun
461*4882a593Smuzhiyun /* Clear interrupt status before enabling interrupt. */
462*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_TX_END_CLR | SPRD_SPI_RX_END_CLR,
463*4882a593Smuzhiyun ss->base + SPRD_SPI_INT_CLR);
464*4882a593Smuzhiyun /* Enable SPI interrupt only in DMA mode. */
465*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_INT_EN);
466*4882a593Smuzhiyun writel_relaxed(val | SPRD_SPI_TX_END_INT_EN |
467*4882a593Smuzhiyun SPRD_SPI_RX_END_INT_EN,
468*4882a593Smuzhiyun ss->base + SPRD_SPI_INT_EN);
469*4882a593Smuzhiyun }
470*4882a593Smuzhiyun
sprd_spi_irq_disable(struct sprd_spi * ss)471*4882a593Smuzhiyun static void sprd_spi_irq_disable(struct sprd_spi *ss)
472*4882a593Smuzhiyun {
473*4882a593Smuzhiyun writel_relaxed(0, ss->base + SPRD_SPI_INT_EN);
474*4882a593Smuzhiyun }
475*4882a593Smuzhiyun
sprd_spi_dma_enable(struct sprd_spi * ss,bool enable)476*4882a593Smuzhiyun static void sprd_spi_dma_enable(struct sprd_spi *ss, bool enable)
477*4882a593Smuzhiyun {
478*4882a593Smuzhiyun u32 val = readl_relaxed(ss->base + SPRD_SPI_CTL2);
479*4882a593Smuzhiyun
480*4882a593Smuzhiyun if (enable)
481*4882a593Smuzhiyun val |= SPRD_SPI_DMA_EN;
482*4882a593Smuzhiyun else
483*4882a593Smuzhiyun val &= ~SPRD_SPI_DMA_EN;
484*4882a593Smuzhiyun
485*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL2);
486*4882a593Smuzhiyun }
487*4882a593Smuzhiyun
sprd_spi_dma_submit(struct dma_chan * dma_chan,struct dma_slave_config * c,struct sg_table * sg,enum dma_transfer_direction dir)488*4882a593Smuzhiyun static int sprd_spi_dma_submit(struct dma_chan *dma_chan,
489*4882a593Smuzhiyun struct dma_slave_config *c,
490*4882a593Smuzhiyun struct sg_table *sg,
491*4882a593Smuzhiyun enum dma_transfer_direction dir)
492*4882a593Smuzhiyun {
493*4882a593Smuzhiyun struct dma_async_tx_descriptor *desc;
494*4882a593Smuzhiyun dma_cookie_t cookie;
495*4882a593Smuzhiyun unsigned long flags;
496*4882a593Smuzhiyun int ret;
497*4882a593Smuzhiyun
498*4882a593Smuzhiyun ret = dmaengine_slave_config(dma_chan, c);
499*4882a593Smuzhiyun if (ret < 0)
500*4882a593Smuzhiyun return ret;
501*4882a593Smuzhiyun
502*4882a593Smuzhiyun flags = SPRD_DMA_FLAGS(SPRD_DMA_CHN_MODE_NONE, SPRD_DMA_NO_TRG,
503*4882a593Smuzhiyun SPRD_DMA_FRAG_REQ, SPRD_DMA_TRANS_INT);
504*4882a593Smuzhiyun desc = dmaengine_prep_slave_sg(dma_chan, sg->sgl, sg->nents, dir, flags);
505*4882a593Smuzhiyun if (!desc)
506*4882a593Smuzhiyun return -ENODEV;
507*4882a593Smuzhiyun
508*4882a593Smuzhiyun cookie = dmaengine_submit(desc);
509*4882a593Smuzhiyun if (dma_submit_error(cookie))
510*4882a593Smuzhiyun return dma_submit_error(cookie);
511*4882a593Smuzhiyun
512*4882a593Smuzhiyun dma_async_issue_pending(dma_chan);
513*4882a593Smuzhiyun
514*4882a593Smuzhiyun return 0;
515*4882a593Smuzhiyun }
516*4882a593Smuzhiyun
sprd_spi_dma_rx_config(struct sprd_spi * ss,struct spi_transfer * t)517*4882a593Smuzhiyun static int sprd_spi_dma_rx_config(struct sprd_spi *ss, struct spi_transfer *t)
518*4882a593Smuzhiyun {
519*4882a593Smuzhiyun struct dma_chan *dma_chan = ss->dma.dma_chan[SPRD_SPI_RX];
520*4882a593Smuzhiyun struct dma_slave_config config = {
521*4882a593Smuzhiyun .src_addr = ss->phy_base,
522*4882a593Smuzhiyun .src_addr_width = ss->dma.width,
523*4882a593Smuzhiyun .dst_addr_width = ss->dma.width,
524*4882a593Smuzhiyun .dst_maxburst = ss->dma.fragmens_len,
525*4882a593Smuzhiyun };
526*4882a593Smuzhiyun int ret;
527*4882a593Smuzhiyun
528*4882a593Smuzhiyun ret = sprd_spi_dma_submit(dma_chan, &config, &t->rx_sg, DMA_DEV_TO_MEM);
529*4882a593Smuzhiyun if (ret)
530*4882a593Smuzhiyun return ret;
531*4882a593Smuzhiyun
532*4882a593Smuzhiyun return ss->dma.rx_len;
533*4882a593Smuzhiyun }
534*4882a593Smuzhiyun
sprd_spi_dma_tx_config(struct sprd_spi * ss,struct spi_transfer * t)535*4882a593Smuzhiyun static int sprd_spi_dma_tx_config(struct sprd_spi *ss, struct spi_transfer *t)
536*4882a593Smuzhiyun {
537*4882a593Smuzhiyun struct dma_chan *dma_chan = ss->dma.dma_chan[SPRD_SPI_TX];
538*4882a593Smuzhiyun struct dma_slave_config config = {
539*4882a593Smuzhiyun .dst_addr = ss->phy_base,
540*4882a593Smuzhiyun .src_addr_width = ss->dma.width,
541*4882a593Smuzhiyun .dst_addr_width = ss->dma.width,
542*4882a593Smuzhiyun .src_maxburst = ss->dma.fragmens_len,
543*4882a593Smuzhiyun };
544*4882a593Smuzhiyun int ret;
545*4882a593Smuzhiyun
546*4882a593Smuzhiyun ret = sprd_spi_dma_submit(dma_chan, &config, &t->tx_sg, DMA_MEM_TO_DEV);
547*4882a593Smuzhiyun if (ret)
548*4882a593Smuzhiyun return ret;
549*4882a593Smuzhiyun
550*4882a593Smuzhiyun return t->len;
551*4882a593Smuzhiyun }
552*4882a593Smuzhiyun
sprd_spi_dma_request(struct sprd_spi * ss)553*4882a593Smuzhiyun static int sprd_spi_dma_request(struct sprd_spi *ss)
554*4882a593Smuzhiyun {
555*4882a593Smuzhiyun ss->dma.dma_chan[SPRD_SPI_RX] = dma_request_chan(ss->dev, "rx_chn");
556*4882a593Smuzhiyun if (IS_ERR_OR_NULL(ss->dma.dma_chan[SPRD_SPI_RX]))
557*4882a593Smuzhiyun return dev_err_probe(ss->dev, PTR_ERR(ss->dma.dma_chan[SPRD_SPI_RX]),
558*4882a593Smuzhiyun "request RX DMA channel failed!\n");
559*4882a593Smuzhiyun
560*4882a593Smuzhiyun ss->dma.dma_chan[SPRD_SPI_TX] = dma_request_chan(ss->dev, "tx_chn");
561*4882a593Smuzhiyun if (IS_ERR_OR_NULL(ss->dma.dma_chan[SPRD_SPI_TX])) {
562*4882a593Smuzhiyun dma_release_channel(ss->dma.dma_chan[SPRD_SPI_RX]);
563*4882a593Smuzhiyun return dev_err_probe(ss->dev, PTR_ERR(ss->dma.dma_chan[SPRD_SPI_TX]),
564*4882a593Smuzhiyun "request TX DMA channel failed!\n");
565*4882a593Smuzhiyun }
566*4882a593Smuzhiyun
567*4882a593Smuzhiyun return 0;
568*4882a593Smuzhiyun }
569*4882a593Smuzhiyun
sprd_spi_dma_release(struct sprd_spi * ss)570*4882a593Smuzhiyun static void sprd_spi_dma_release(struct sprd_spi *ss)
571*4882a593Smuzhiyun {
572*4882a593Smuzhiyun if (ss->dma.dma_chan[SPRD_SPI_RX])
573*4882a593Smuzhiyun dma_release_channel(ss->dma.dma_chan[SPRD_SPI_RX]);
574*4882a593Smuzhiyun
575*4882a593Smuzhiyun if (ss->dma.dma_chan[SPRD_SPI_TX])
576*4882a593Smuzhiyun dma_release_channel(ss->dma.dma_chan[SPRD_SPI_TX]);
577*4882a593Smuzhiyun }
578*4882a593Smuzhiyun
sprd_spi_dma_txrx_bufs(struct spi_device * sdev,struct spi_transfer * t)579*4882a593Smuzhiyun static int sprd_spi_dma_txrx_bufs(struct spi_device *sdev,
580*4882a593Smuzhiyun struct spi_transfer *t)
581*4882a593Smuzhiyun {
582*4882a593Smuzhiyun struct sprd_spi *ss = spi_master_get_devdata(sdev->master);
583*4882a593Smuzhiyun u32 trans_len = ss->trans_len;
584*4882a593Smuzhiyun int ret, write_size = 0;
585*4882a593Smuzhiyun
586*4882a593Smuzhiyun reinit_completion(&ss->xfer_completion);
587*4882a593Smuzhiyun sprd_spi_irq_enable(ss);
588*4882a593Smuzhiyun if (ss->trans_mode & SPRD_SPI_TX_MODE) {
589*4882a593Smuzhiyun write_size = sprd_spi_dma_tx_config(ss, t);
590*4882a593Smuzhiyun sprd_spi_set_tx_length(ss, trans_len);
591*4882a593Smuzhiyun
592*4882a593Smuzhiyun /*
593*4882a593Smuzhiyun * For our 3 wires mode or dual TX line mode, we need
594*4882a593Smuzhiyun * to request the controller to transfer.
595*4882a593Smuzhiyun */
596*4882a593Smuzhiyun if (ss->hw_mode & SPI_3WIRE || ss->hw_mode & SPI_TX_DUAL)
597*4882a593Smuzhiyun sprd_spi_tx_req(ss);
598*4882a593Smuzhiyun } else {
599*4882a593Smuzhiyun sprd_spi_set_rx_length(ss, trans_len);
600*4882a593Smuzhiyun
601*4882a593Smuzhiyun /*
602*4882a593Smuzhiyun * For our 3 wires mode or dual TX line mode, we need
603*4882a593Smuzhiyun * to request the controller to read.
604*4882a593Smuzhiyun */
605*4882a593Smuzhiyun if (ss->hw_mode & SPI_3WIRE || ss->hw_mode & SPI_TX_DUAL)
606*4882a593Smuzhiyun sprd_spi_rx_req(ss);
607*4882a593Smuzhiyun else
608*4882a593Smuzhiyun write_size = ss->write_bufs(ss, trans_len);
609*4882a593Smuzhiyun }
610*4882a593Smuzhiyun
611*4882a593Smuzhiyun if (write_size < 0) {
612*4882a593Smuzhiyun ret = write_size;
613*4882a593Smuzhiyun dev_err(ss->dev, "failed to write, ret = %d\n", ret);
614*4882a593Smuzhiyun goto trans_complete;
615*4882a593Smuzhiyun }
616*4882a593Smuzhiyun
617*4882a593Smuzhiyun if (ss->trans_mode & SPRD_SPI_RX_MODE) {
618*4882a593Smuzhiyun /*
619*4882a593Smuzhiyun * Set up the DMA receive data length, which must be an
620*4882a593Smuzhiyun * integral multiple of fragment length. But when the length
621*4882a593Smuzhiyun * of received data is less than fragment length, DMA can be
622*4882a593Smuzhiyun * configured to receive data according to the actual length
623*4882a593Smuzhiyun * of received data.
624*4882a593Smuzhiyun */
625*4882a593Smuzhiyun ss->dma.rx_len = t->len > ss->dma.fragmens_len ?
626*4882a593Smuzhiyun (t->len - t->len % ss->dma.fragmens_len) :
627*4882a593Smuzhiyun t->len;
628*4882a593Smuzhiyun ret = sprd_spi_dma_rx_config(ss, t);
629*4882a593Smuzhiyun if (ret < 0) {
630*4882a593Smuzhiyun dev_err(&sdev->dev,
631*4882a593Smuzhiyun "failed to configure rx DMA, ret = %d\n", ret);
632*4882a593Smuzhiyun goto trans_complete;
633*4882a593Smuzhiyun }
634*4882a593Smuzhiyun }
635*4882a593Smuzhiyun
636*4882a593Smuzhiyun sprd_spi_dma_enable(ss, true);
637*4882a593Smuzhiyun wait_for_completion(&(ss->xfer_completion));
638*4882a593Smuzhiyun
639*4882a593Smuzhiyun if (ss->trans_mode & SPRD_SPI_TX_MODE)
640*4882a593Smuzhiyun ret = write_size;
641*4882a593Smuzhiyun else
642*4882a593Smuzhiyun ret = ss->dma.rx_len;
643*4882a593Smuzhiyun
644*4882a593Smuzhiyun trans_complete:
645*4882a593Smuzhiyun sprd_spi_dma_enable(ss, false);
646*4882a593Smuzhiyun sprd_spi_enter_idle(ss);
647*4882a593Smuzhiyun sprd_spi_irq_disable(ss);
648*4882a593Smuzhiyun
649*4882a593Smuzhiyun return ret;
650*4882a593Smuzhiyun }
651*4882a593Smuzhiyun
sprd_spi_set_speed(struct sprd_spi * ss,u32 speed_hz)652*4882a593Smuzhiyun static void sprd_spi_set_speed(struct sprd_spi *ss, u32 speed_hz)
653*4882a593Smuzhiyun {
654*4882a593Smuzhiyun /*
655*4882a593Smuzhiyun * From SPI datasheet, the prescale calculation formula:
656*4882a593Smuzhiyun * prescale = SPI source clock / (2 * SPI_freq) - 1;
657*4882a593Smuzhiyun */
658*4882a593Smuzhiyun u32 clk_div = DIV_ROUND_UP(ss->src_clk, speed_hz << 1) - 1;
659*4882a593Smuzhiyun
660*4882a593Smuzhiyun /* Save the real hardware speed */
661*4882a593Smuzhiyun ss->hw_speed_hz = (ss->src_clk >> 1) / (clk_div + 1);
662*4882a593Smuzhiyun writel_relaxed(clk_div, ss->base + SPRD_SPI_CLKD);
663*4882a593Smuzhiyun }
664*4882a593Smuzhiyun
sprd_spi_init_hw(struct sprd_spi * ss,struct spi_transfer * t)665*4882a593Smuzhiyun static int sprd_spi_init_hw(struct sprd_spi *ss, struct spi_transfer *t)
666*4882a593Smuzhiyun {
667*4882a593Smuzhiyun struct spi_delay *d = &t->word_delay;
668*4882a593Smuzhiyun u16 word_delay, interval;
669*4882a593Smuzhiyun u32 val;
670*4882a593Smuzhiyun
671*4882a593Smuzhiyun if (d->unit != SPI_DELAY_UNIT_SCK)
672*4882a593Smuzhiyun return -EINVAL;
673*4882a593Smuzhiyun
674*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL0);
675*4882a593Smuzhiyun val &= ~(SPRD_SPI_SCK_REV | SPRD_SPI_NG_TX | SPRD_SPI_NG_RX);
676*4882a593Smuzhiyun /* Set default chip selection, clock phase and clock polarity */
677*4882a593Smuzhiyun val |= ss->hw_mode & SPI_CPHA ? SPRD_SPI_NG_RX : SPRD_SPI_NG_TX;
678*4882a593Smuzhiyun val |= ss->hw_mode & SPI_CPOL ? SPRD_SPI_SCK_REV : 0;
679*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL0);
680*4882a593Smuzhiyun
681*4882a593Smuzhiyun /*
682*4882a593Smuzhiyun * Set the intervals of two SPI frames, and the inteval calculation
683*4882a593Smuzhiyun * formula as below per datasheet:
684*4882a593Smuzhiyun * interval time (source clock cycles) = interval * 4 + 10.
685*4882a593Smuzhiyun */
686*4882a593Smuzhiyun word_delay = clamp_t(u16, d->value, SPRD_SPI_MIN_DELAY_CYCLE,
687*4882a593Smuzhiyun SPRD_SPI_MAX_DELAY_CYCLE);
688*4882a593Smuzhiyun interval = DIV_ROUND_UP(word_delay - 10, 4);
689*4882a593Smuzhiyun ss->word_delay = interval * 4 + 10;
690*4882a593Smuzhiyun writel_relaxed(interval, ss->base + SPRD_SPI_CTL5);
691*4882a593Smuzhiyun
692*4882a593Smuzhiyun /* Reset SPI fifo */
693*4882a593Smuzhiyun writel_relaxed(1, ss->base + SPRD_SPI_FIFO_RST);
694*4882a593Smuzhiyun writel_relaxed(0, ss->base + SPRD_SPI_FIFO_RST);
695*4882a593Smuzhiyun
696*4882a593Smuzhiyun /* Set SPI work mode */
697*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL7);
698*4882a593Smuzhiyun val &= ~SPRD_SPI_MODE_MASK;
699*4882a593Smuzhiyun
700*4882a593Smuzhiyun if (ss->hw_mode & SPI_3WIRE)
701*4882a593Smuzhiyun val |= SPRD_SPI_3WIRE_MODE << SPRD_SPI_MODE_OFFSET;
702*4882a593Smuzhiyun else
703*4882a593Smuzhiyun val |= SPRD_SPI_4WIRE_MODE << SPRD_SPI_MODE_OFFSET;
704*4882a593Smuzhiyun
705*4882a593Smuzhiyun if (ss->hw_mode & SPI_TX_DUAL)
706*4882a593Smuzhiyun val |= SPRD_SPI_DATA_LINE2_EN;
707*4882a593Smuzhiyun else
708*4882a593Smuzhiyun val &= ~SPRD_SPI_DATA_LINE2_EN;
709*4882a593Smuzhiyun
710*4882a593Smuzhiyun writel_relaxed(val, ss->base + SPRD_SPI_CTL7);
711*4882a593Smuzhiyun
712*4882a593Smuzhiyun return 0;
713*4882a593Smuzhiyun }
714*4882a593Smuzhiyun
sprd_spi_setup_transfer(struct spi_device * sdev,struct spi_transfer * t)715*4882a593Smuzhiyun static int sprd_spi_setup_transfer(struct spi_device *sdev,
716*4882a593Smuzhiyun struct spi_transfer *t)
717*4882a593Smuzhiyun {
718*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sdev->controller);
719*4882a593Smuzhiyun u8 bits_per_word = t->bits_per_word;
720*4882a593Smuzhiyun u32 val, mode = 0;
721*4882a593Smuzhiyun int ret;
722*4882a593Smuzhiyun
723*4882a593Smuzhiyun ss->len = t->len;
724*4882a593Smuzhiyun ss->tx_buf = t->tx_buf;
725*4882a593Smuzhiyun ss->rx_buf = t->rx_buf;
726*4882a593Smuzhiyun
727*4882a593Smuzhiyun ss->hw_mode = sdev->mode;
728*4882a593Smuzhiyun ret = sprd_spi_init_hw(ss, t);
729*4882a593Smuzhiyun if (ret)
730*4882a593Smuzhiyun return ret;
731*4882a593Smuzhiyun
732*4882a593Smuzhiyun /* Set tansfer speed and valid bits */
733*4882a593Smuzhiyun sprd_spi_set_speed(ss, t->speed_hz);
734*4882a593Smuzhiyun sprd_spi_set_transfer_bits(ss, bits_per_word);
735*4882a593Smuzhiyun
736*4882a593Smuzhiyun if (bits_per_word > 16)
737*4882a593Smuzhiyun bits_per_word = round_up(bits_per_word, 16);
738*4882a593Smuzhiyun else
739*4882a593Smuzhiyun bits_per_word = round_up(bits_per_word, 8);
740*4882a593Smuzhiyun
741*4882a593Smuzhiyun switch (bits_per_word) {
742*4882a593Smuzhiyun case 8:
743*4882a593Smuzhiyun ss->trans_len = t->len;
744*4882a593Smuzhiyun ss->read_bufs = sprd_spi_read_bufs_u8;
745*4882a593Smuzhiyun ss->write_bufs = sprd_spi_write_bufs_u8;
746*4882a593Smuzhiyun ss->dma.width = DMA_SLAVE_BUSWIDTH_1_BYTE;
747*4882a593Smuzhiyun ss->dma.fragmens_len = SPRD_SPI_DMA_STEP;
748*4882a593Smuzhiyun break;
749*4882a593Smuzhiyun case 16:
750*4882a593Smuzhiyun ss->trans_len = t->len >> 1;
751*4882a593Smuzhiyun ss->read_bufs = sprd_spi_read_bufs_u16;
752*4882a593Smuzhiyun ss->write_bufs = sprd_spi_write_bufs_u16;
753*4882a593Smuzhiyun ss->dma.width = DMA_SLAVE_BUSWIDTH_2_BYTES;
754*4882a593Smuzhiyun ss->dma.fragmens_len = SPRD_SPI_DMA_STEP << 1;
755*4882a593Smuzhiyun break;
756*4882a593Smuzhiyun case 32:
757*4882a593Smuzhiyun ss->trans_len = t->len >> 2;
758*4882a593Smuzhiyun ss->read_bufs = sprd_spi_read_bufs_u32;
759*4882a593Smuzhiyun ss->write_bufs = sprd_spi_write_bufs_u32;
760*4882a593Smuzhiyun ss->dma.width = DMA_SLAVE_BUSWIDTH_4_BYTES;
761*4882a593Smuzhiyun ss->dma.fragmens_len = SPRD_SPI_DMA_STEP << 2;
762*4882a593Smuzhiyun break;
763*4882a593Smuzhiyun default:
764*4882a593Smuzhiyun return -EINVAL;
765*4882a593Smuzhiyun }
766*4882a593Smuzhiyun
767*4882a593Smuzhiyun /* Set transfer read or write mode */
768*4882a593Smuzhiyun val = readl_relaxed(ss->base + SPRD_SPI_CTL1);
769*4882a593Smuzhiyun val &= ~SPRD_SPI_RTX_MD_MASK;
770*4882a593Smuzhiyun if (t->tx_buf)
771*4882a593Smuzhiyun mode |= SPRD_SPI_TX_MODE;
772*4882a593Smuzhiyun if (t->rx_buf)
773*4882a593Smuzhiyun mode |= SPRD_SPI_RX_MODE;
774*4882a593Smuzhiyun
775*4882a593Smuzhiyun writel_relaxed(val | mode, ss->base + SPRD_SPI_CTL1);
776*4882a593Smuzhiyun
777*4882a593Smuzhiyun ss->trans_mode = mode;
778*4882a593Smuzhiyun
779*4882a593Smuzhiyun /*
780*4882a593Smuzhiyun * If in only receive mode, we need to trigger the SPI controller to
781*4882a593Smuzhiyun * receive data automatically.
782*4882a593Smuzhiyun */
783*4882a593Smuzhiyun if (ss->trans_mode == SPRD_SPI_RX_MODE)
784*4882a593Smuzhiyun ss->write_bufs = sprd_spi_write_only_receive;
785*4882a593Smuzhiyun
786*4882a593Smuzhiyun return 0;
787*4882a593Smuzhiyun }
788*4882a593Smuzhiyun
sprd_spi_transfer_one(struct spi_controller * sctlr,struct spi_device * sdev,struct spi_transfer * t)789*4882a593Smuzhiyun static int sprd_spi_transfer_one(struct spi_controller *sctlr,
790*4882a593Smuzhiyun struct spi_device *sdev,
791*4882a593Smuzhiyun struct spi_transfer *t)
792*4882a593Smuzhiyun {
793*4882a593Smuzhiyun int ret;
794*4882a593Smuzhiyun
795*4882a593Smuzhiyun ret = sprd_spi_setup_transfer(sdev, t);
796*4882a593Smuzhiyun if (ret)
797*4882a593Smuzhiyun goto setup_err;
798*4882a593Smuzhiyun
799*4882a593Smuzhiyun if (sctlr->can_dma(sctlr, sdev, t))
800*4882a593Smuzhiyun ret = sprd_spi_dma_txrx_bufs(sdev, t);
801*4882a593Smuzhiyun else
802*4882a593Smuzhiyun ret = sprd_spi_txrx_bufs(sdev, t);
803*4882a593Smuzhiyun
804*4882a593Smuzhiyun if (ret == t->len)
805*4882a593Smuzhiyun ret = 0;
806*4882a593Smuzhiyun else if (ret >= 0)
807*4882a593Smuzhiyun ret = -EREMOTEIO;
808*4882a593Smuzhiyun
809*4882a593Smuzhiyun setup_err:
810*4882a593Smuzhiyun spi_finalize_current_transfer(sctlr);
811*4882a593Smuzhiyun
812*4882a593Smuzhiyun return ret;
813*4882a593Smuzhiyun }
814*4882a593Smuzhiyun
sprd_spi_handle_irq(int irq,void * data)815*4882a593Smuzhiyun static irqreturn_t sprd_spi_handle_irq(int irq, void *data)
816*4882a593Smuzhiyun {
817*4882a593Smuzhiyun struct sprd_spi *ss = (struct sprd_spi *)data;
818*4882a593Smuzhiyun u32 val = readl_relaxed(ss->base + SPRD_SPI_INT_MASK_STS);
819*4882a593Smuzhiyun
820*4882a593Smuzhiyun if (val & SPRD_SPI_MASK_TX_END) {
821*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_TX_END_CLR, ss->base + SPRD_SPI_INT_CLR);
822*4882a593Smuzhiyun if (!(ss->trans_mode & SPRD_SPI_RX_MODE))
823*4882a593Smuzhiyun complete(&ss->xfer_completion);
824*4882a593Smuzhiyun
825*4882a593Smuzhiyun return IRQ_HANDLED;
826*4882a593Smuzhiyun }
827*4882a593Smuzhiyun
828*4882a593Smuzhiyun if (val & SPRD_SPI_MASK_RX_END) {
829*4882a593Smuzhiyun writel_relaxed(SPRD_SPI_RX_END_CLR, ss->base + SPRD_SPI_INT_CLR);
830*4882a593Smuzhiyun if (ss->dma.rx_len < ss->len) {
831*4882a593Smuzhiyun ss->rx_buf += ss->dma.rx_len;
832*4882a593Smuzhiyun ss->dma.rx_len +=
833*4882a593Smuzhiyun ss->read_bufs(ss, ss->len - ss->dma.rx_len);
834*4882a593Smuzhiyun }
835*4882a593Smuzhiyun complete(&ss->xfer_completion);
836*4882a593Smuzhiyun
837*4882a593Smuzhiyun return IRQ_HANDLED;
838*4882a593Smuzhiyun }
839*4882a593Smuzhiyun
840*4882a593Smuzhiyun return IRQ_NONE;
841*4882a593Smuzhiyun }
842*4882a593Smuzhiyun
sprd_spi_irq_init(struct platform_device * pdev,struct sprd_spi * ss)843*4882a593Smuzhiyun static int sprd_spi_irq_init(struct platform_device *pdev, struct sprd_spi *ss)
844*4882a593Smuzhiyun {
845*4882a593Smuzhiyun int ret;
846*4882a593Smuzhiyun
847*4882a593Smuzhiyun ss->irq = platform_get_irq(pdev, 0);
848*4882a593Smuzhiyun if (ss->irq < 0)
849*4882a593Smuzhiyun return ss->irq;
850*4882a593Smuzhiyun
851*4882a593Smuzhiyun ret = devm_request_irq(&pdev->dev, ss->irq, sprd_spi_handle_irq,
852*4882a593Smuzhiyun 0, pdev->name, ss);
853*4882a593Smuzhiyun if (ret)
854*4882a593Smuzhiyun dev_err(&pdev->dev, "failed to request spi irq %d, ret = %d\n",
855*4882a593Smuzhiyun ss->irq, ret);
856*4882a593Smuzhiyun
857*4882a593Smuzhiyun return ret;
858*4882a593Smuzhiyun }
859*4882a593Smuzhiyun
sprd_spi_clk_init(struct platform_device * pdev,struct sprd_spi * ss)860*4882a593Smuzhiyun static int sprd_spi_clk_init(struct platform_device *pdev, struct sprd_spi *ss)
861*4882a593Smuzhiyun {
862*4882a593Smuzhiyun struct clk *clk_spi, *clk_parent;
863*4882a593Smuzhiyun
864*4882a593Smuzhiyun clk_spi = devm_clk_get(&pdev->dev, "spi");
865*4882a593Smuzhiyun if (IS_ERR(clk_spi)) {
866*4882a593Smuzhiyun dev_warn(&pdev->dev, "can't get the spi clock\n");
867*4882a593Smuzhiyun clk_spi = NULL;
868*4882a593Smuzhiyun }
869*4882a593Smuzhiyun
870*4882a593Smuzhiyun clk_parent = devm_clk_get(&pdev->dev, "source");
871*4882a593Smuzhiyun if (IS_ERR(clk_parent)) {
872*4882a593Smuzhiyun dev_warn(&pdev->dev, "can't get the source clock\n");
873*4882a593Smuzhiyun clk_parent = NULL;
874*4882a593Smuzhiyun }
875*4882a593Smuzhiyun
876*4882a593Smuzhiyun ss->clk = devm_clk_get(&pdev->dev, "enable");
877*4882a593Smuzhiyun if (IS_ERR(ss->clk)) {
878*4882a593Smuzhiyun dev_err(&pdev->dev, "can't get the enable clock\n");
879*4882a593Smuzhiyun return PTR_ERR(ss->clk);
880*4882a593Smuzhiyun }
881*4882a593Smuzhiyun
882*4882a593Smuzhiyun if (!clk_set_parent(clk_spi, clk_parent))
883*4882a593Smuzhiyun ss->src_clk = clk_get_rate(clk_spi);
884*4882a593Smuzhiyun else
885*4882a593Smuzhiyun ss->src_clk = SPRD_SPI_DEFAULT_SOURCE;
886*4882a593Smuzhiyun
887*4882a593Smuzhiyun return 0;
888*4882a593Smuzhiyun }
889*4882a593Smuzhiyun
sprd_spi_can_dma(struct spi_controller * sctlr,struct spi_device * spi,struct spi_transfer * t)890*4882a593Smuzhiyun static bool sprd_spi_can_dma(struct spi_controller *sctlr,
891*4882a593Smuzhiyun struct spi_device *spi, struct spi_transfer *t)
892*4882a593Smuzhiyun {
893*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sctlr);
894*4882a593Smuzhiyun
895*4882a593Smuzhiyun return ss->dma.enable && (t->len > SPRD_SPI_FIFO_SIZE);
896*4882a593Smuzhiyun }
897*4882a593Smuzhiyun
sprd_spi_dma_init(struct platform_device * pdev,struct sprd_spi * ss)898*4882a593Smuzhiyun static int sprd_spi_dma_init(struct platform_device *pdev, struct sprd_spi *ss)
899*4882a593Smuzhiyun {
900*4882a593Smuzhiyun int ret;
901*4882a593Smuzhiyun
902*4882a593Smuzhiyun ret = sprd_spi_dma_request(ss);
903*4882a593Smuzhiyun if (ret) {
904*4882a593Smuzhiyun if (ret == -EPROBE_DEFER)
905*4882a593Smuzhiyun return ret;
906*4882a593Smuzhiyun
907*4882a593Smuzhiyun dev_warn(&pdev->dev,
908*4882a593Smuzhiyun "failed to request dma, enter no dma mode, ret = %d\n",
909*4882a593Smuzhiyun ret);
910*4882a593Smuzhiyun
911*4882a593Smuzhiyun return 0;
912*4882a593Smuzhiyun }
913*4882a593Smuzhiyun
914*4882a593Smuzhiyun ss->dma.enable = true;
915*4882a593Smuzhiyun
916*4882a593Smuzhiyun return 0;
917*4882a593Smuzhiyun }
918*4882a593Smuzhiyun
sprd_spi_probe(struct platform_device * pdev)919*4882a593Smuzhiyun static int sprd_spi_probe(struct platform_device *pdev)
920*4882a593Smuzhiyun {
921*4882a593Smuzhiyun struct spi_controller *sctlr;
922*4882a593Smuzhiyun struct resource *res;
923*4882a593Smuzhiyun struct sprd_spi *ss;
924*4882a593Smuzhiyun int ret;
925*4882a593Smuzhiyun
926*4882a593Smuzhiyun pdev->id = of_alias_get_id(pdev->dev.of_node, "spi");
927*4882a593Smuzhiyun sctlr = spi_alloc_master(&pdev->dev, sizeof(*ss));
928*4882a593Smuzhiyun if (!sctlr)
929*4882a593Smuzhiyun return -ENOMEM;
930*4882a593Smuzhiyun
931*4882a593Smuzhiyun ss = spi_controller_get_devdata(sctlr);
932*4882a593Smuzhiyun res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
933*4882a593Smuzhiyun ss->base = devm_ioremap_resource(&pdev->dev, res);
934*4882a593Smuzhiyun if (IS_ERR(ss->base)) {
935*4882a593Smuzhiyun ret = PTR_ERR(ss->base);
936*4882a593Smuzhiyun goto free_controller;
937*4882a593Smuzhiyun }
938*4882a593Smuzhiyun
939*4882a593Smuzhiyun ss->phy_base = res->start;
940*4882a593Smuzhiyun ss->dev = &pdev->dev;
941*4882a593Smuzhiyun sctlr->dev.of_node = pdev->dev.of_node;
942*4882a593Smuzhiyun sctlr->mode_bits = SPI_CPOL | SPI_CPHA | SPI_3WIRE | SPI_TX_DUAL;
943*4882a593Smuzhiyun sctlr->bus_num = pdev->id;
944*4882a593Smuzhiyun sctlr->set_cs = sprd_spi_chipselect;
945*4882a593Smuzhiyun sctlr->transfer_one = sprd_spi_transfer_one;
946*4882a593Smuzhiyun sctlr->can_dma = sprd_spi_can_dma;
947*4882a593Smuzhiyun sctlr->auto_runtime_pm = true;
948*4882a593Smuzhiyun sctlr->max_speed_hz = min_t(u32, ss->src_clk >> 1,
949*4882a593Smuzhiyun SPRD_SPI_MAX_SPEED_HZ);
950*4882a593Smuzhiyun
951*4882a593Smuzhiyun init_completion(&ss->xfer_completion);
952*4882a593Smuzhiyun platform_set_drvdata(pdev, sctlr);
953*4882a593Smuzhiyun ret = sprd_spi_clk_init(pdev, ss);
954*4882a593Smuzhiyun if (ret)
955*4882a593Smuzhiyun goto free_controller;
956*4882a593Smuzhiyun
957*4882a593Smuzhiyun ret = sprd_spi_irq_init(pdev, ss);
958*4882a593Smuzhiyun if (ret)
959*4882a593Smuzhiyun goto free_controller;
960*4882a593Smuzhiyun
961*4882a593Smuzhiyun ret = sprd_spi_dma_init(pdev, ss);
962*4882a593Smuzhiyun if (ret)
963*4882a593Smuzhiyun goto free_controller;
964*4882a593Smuzhiyun
965*4882a593Smuzhiyun ret = clk_prepare_enable(ss->clk);
966*4882a593Smuzhiyun if (ret)
967*4882a593Smuzhiyun goto release_dma;
968*4882a593Smuzhiyun
969*4882a593Smuzhiyun ret = pm_runtime_set_active(&pdev->dev);
970*4882a593Smuzhiyun if (ret < 0)
971*4882a593Smuzhiyun goto disable_clk;
972*4882a593Smuzhiyun
973*4882a593Smuzhiyun pm_runtime_set_autosuspend_delay(&pdev->dev,
974*4882a593Smuzhiyun SPRD_SPI_AUTOSUSPEND_DELAY);
975*4882a593Smuzhiyun pm_runtime_use_autosuspend(&pdev->dev);
976*4882a593Smuzhiyun pm_runtime_enable(&pdev->dev);
977*4882a593Smuzhiyun ret = pm_runtime_get_sync(&pdev->dev);
978*4882a593Smuzhiyun if (ret < 0) {
979*4882a593Smuzhiyun dev_err(&pdev->dev, "failed to resume SPI controller\n");
980*4882a593Smuzhiyun goto err_rpm_put;
981*4882a593Smuzhiyun }
982*4882a593Smuzhiyun
983*4882a593Smuzhiyun ret = devm_spi_register_controller(&pdev->dev, sctlr);
984*4882a593Smuzhiyun if (ret)
985*4882a593Smuzhiyun goto err_rpm_put;
986*4882a593Smuzhiyun
987*4882a593Smuzhiyun pm_runtime_mark_last_busy(&pdev->dev);
988*4882a593Smuzhiyun pm_runtime_put_autosuspend(&pdev->dev);
989*4882a593Smuzhiyun
990*4882a593Smuzhiyun return 0;
991*4882a593Smuzhiyun
992*4882a593Smuzhiyun err_rpm_put:
993*4882a593Smuzhiyun pm_runtime_put_noidle(&pdev->dev);
994*4882a593Smuzhiyun pm_runtime_disable(&pdev->dev);
995*4882a593Smuzhiyun disable_clk:
996*4882a593Smuzhiyun clk_disable_unprepare(ss->clk);
997*4882a593Smuzhiyun release_dma:
998*4882a593Smuzhiyun sprd_spi_dma_release(ss);
999*4882a593Smuzhiyun free_controller:
1000*4882a593Smuzhiyun spi_controller_put(sctlr);
1001*4882a593Smuzhiyun
1002*4882a593Smuzhiyun return ret;
1003*4882a593Smuzhiyun }
1004*4882a593Smuzhiyun
sprd_spi_remove(struct platform_device * pdev)1005*4882a593Smuzhiyun static int sprd_spi_remove(struct platform_device *pdev)
1006*4882a593Smuzhiyun {
1007*4882a593Smuzhiyun struct spi_controller *sctlr = platform_get_drvdata(pdev);
1008*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sctlr);
1009*4882a593Smuzhiyun int ret;
1010*4882a593Smuzhiyun
1011*4882a593Smuzhiyun ret = pm_runtime_get_sync(ss->dev);
1012*4882a593Smuzhiyun if (ret < 0) {
1013*4882a593Smuzhiyun pm_runtime_put_noidle(ss->dev);
1014*4882a593Smuzhiyun dev_err(ss->dev, "failed to resume SPI controller\n");
1015*4882a593Smuzhiyun return ret;
1016*4882a593Smuzhiyun }
1017*4882a593Smuzhiyun
1018*4882a593Smuzhiyun spi_controller_suspend(sctlr);
1019*4882a593Smuzhiyun
1020*4882a593Smuzhiyun if (ss->dma.enable)
1021*4882a593Smuzhiyun sprd_spi_dma_release(ss);
1022*4882a593Smuzhiyun clk_disable_unprepare(ss->clk);
1023*4882a593Smuzhiyun pm_runtime_put_noidle(&pdev->dev);
1024*4882a593Smuzhiyun pm_runtime_disable(&pdev->dev);
1025*4882a593Smuzhiyun
1026*4882a593Smuzhiyun return 0;
1027*4882a593Smuzhiyun }
1028*4882a593Smuzhiyun
sprd_spi_runtime_suspend(struct device * dev)1029*4882a593Smuzhiyun static int __maybe_unused sprd_spi_runtime_suspend(struct device *dev)
1030*4882a593Smuzhiyun {
1031*4882a593Smuzhiyun struct spi_controller *sctlr = dev_get_drvdata(dev);
1032*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sctlr);
1033*4882a593Smuzhiyun
1034*4882a593Smuzhiyun if (ss->dma.enable)
1035*4882a593Smuzhiyun sprd_spi_dma_release(ss);
1036*4882a593Smuzhiyun
1037*4882a593Smuzhiyun clk_disable_unprepare(ss->clk);
1038*4882a593Smuzhiyun
1039*4882a593Smuzhiyun return 0;
1040*4882a593Smuzhiyun }
1041*4882a593Smuzhiyun
sprd_spi_runtime_resume(struct device * dev)1042*4882a593Smuzhiyun static int __maybe_unused sprd_spi_runtime_resume(struct device *dev)
1043*4882a593Smuzhiyun {
1044*4882a593Smuzhiyun struct spi_controller *sctlr = dev_get_drvdata(dev);
1045*4882a593Smuzhiyun struct sprd_spi *ss = spi_controller_get_devdata(sctlr);
1046*4882a593Smuzhiyun int ret;
1047*4882a593Smuzhiyun
1048*4882a593Smuzhiyun ret = clk_prepare_enable(ss->clk);
1049*4882a593Smuzhiyun if (ret)
1050*4882a593Smuzhiyun return ret;
1051*4882a593Smuzhiyun
1052*4882a593Smuzhiyun if (!ss->dma.enable)
1053*4882a593Smuzhiyun return 0;
1054*4882a593Smuzhiyun
1055*4882a593Smuzhiyun ret = sprd_spi_dma_request(ss);
1056*4882a593Smuzhiyun if (ret)
1057*4882a593Smuzhiyun clk_disable_unprepare(ss->clk);
1058*4882a593Smuzhiyun
1059*4882a593Smuzhiyun return ret;
1060*4882a593Smuzhiyun }
1061*4882a593Smuzhiyun
1062*4882a593Smuzhiyun static const struct dev_pm_ops sprd_spi_pm_ops = {
1063*4882a593Smuzhiyun SET_RUNTIME_PM_OPS(sprd_spi_runtime_suspend,
1064*4882a593Smuzhiyun sprd_spi_runtime_resume, NULL)
1065*4882a593Smuzhiyun };
1066*4882a593Smuzhiyun
1067*4882a593Smuzhiyun static const struct of_device_id sprd_spi_of_match[] = {
1068*4882a593Smuzhiyun { .compatible = "sprd,sc9860-spi", },
1069*4882a593Smuzhiyun { /* sentinel */ }
1070*4882a593Smuzhiyun };
1071*4882a593Smuzhiyun MODULE_DEVICE_TABLE(of, sprd_spi_of_match);
1072*4882a593Smuzhiyun
1073*4882a593Smuzhiyun static struct platform_driver sprd_spi_driver = {
1074*4882a593Smuzhiyun .driver = {
1075*4882a593Smuzhiyun .name = "sprd-spi",
1076*4882a593Smuzhiyun .of_match_table = sprd_spi_of_match,
1077*4882a593Smuzhiyun .pm = &sprd_spi_pm_ops,
1078*4882a593Smuzhiyun },
1079*4882a593Smuzhiyun .probe = sprd_spi_probe,
1080*4882a593Smuzhiyun .remove = sprd_spi_remove,
1081*4882a593Smuzhiyun };
1082*4882a593Smuzhiyun
1083*4882a593Smuzhiyun module_platform_driver(sprd_spi_driver);
1084*4882a593Smuzhiyun
1085*4882a593Smuzhiyun MODULE_DESCRIPTION("Spreadtrum SPI Controller driver");
1086*4882a593Smuzhiyun MODULE_AUTHOR("Lanqing Liu <lanqing.liu@spreadtrum.com>");
1087*4882a593Smuzhiyun MODULE_LICENSE("GPL v2");
1088