xref: /OK3568_Linux_fs/kernel/drivers/slimbus/qcom-ctrl.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Copyright (c) 2011-2017, The Linux Foundation
4*4882a593Smuzhiyun  */
5*4882a593Smuzhiyun 
6*4882a593Smuzhiyun #include <linux/irq.h>
7*4882a593Smuzhiyun #include <linux/kernel.h>
8*4882a593Smuzhiyun #include <linux/init.h>
9*4882a593Smuzhiyun #include <linux/slab.h>
10*4882a593Smuzhiyun #include <linux/io.h>
11*4882a593Smuzhiyun #include <linux/interrupt.h>
12*4882a593Smuzhiyun #include <linux/platform_device.h>
13*4882a593Smuzhiyun #include <linux/delay.h>
14*4882a593Smuzhiyun #include <linux/clk.h>
15*4882a593Smuzhiyun #include <linux/of.h>
16*4882a593Smuzhiyun #include <linux/pm_runtime.h>
17*4882a593Smuzhiyun #include "slimbus.h"
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun /* Manager registers */
20*4882a593Smuzhiyun #define	MGR_CFG		0x200
21*4882a593Smuzhiyun #define	MGR_STATUS	0x204
22*4882a593Smuzhiyun #define	MGR_INT_EN	0x210
23*4882a593Smuzhiyun #define	MGR_INT_STAT	0x214
24*4882a593Smuzhiyun #define	MGR_INT_CLR	0x218
25*4882a593Smuzhiyun #define	MGR_TX_MSG	0x230
26*4882a593Smuzhiyun #define	MGR_RX_MSG	0x270
27*4882a593Smuzhiyun #define	MGR_IE_STAT	0x2F0
28*4882a593Smuzhiyun #define	MGR_VE_STAT	0x300
29*4882a593Smuzhiyun #define	MGR_CFG_ENABLE	1
30*4882a593Smuzhiyun 
31*4882a593Smuzhiyun /* Framer registers */
32*4882a593Smuzhiyun #define	FRM_CFG		0x400
33*4882a593Smuzhiyun #define	FRM_STAT	0x404
34*4882a593Smuzhiyun #define	FRM_INT_EN	0x410
35*4882a593Smuzhiyun #define	FRM_INT_STAT	0x414
36*4882a593Smuzhiyun #define	FRM_INT_CLR	0x418
37*4882a593Smuzhiyun #define	FRM_WAKEUP	0x41C
38*4882a593Smuzhiyun #define	FRM_CLKCTL_DONE	0x420
39*4882a593Smuzhiyun #define	FRM_IE_STAT	0x430
40*4882a593Smuzhiyun #define	FRM_VE_STAT	0x440
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun /* Interface registers */
43*4882a593Smuzhiyun #define	INTF_CFG	0x600
44*4882a593Smuzhiyun #define	INTF_STAT	0x604
45*4882a593Smuzhiyun #define	INTF_INT_EN	0x610
46*4882a593Smuzhiyun #define	INTF_INT_STAT	0x614
47*4882a593Smuzhiyun #define	INTF_INT_CLR	0x618
48*4882a593Smuzhiyun #define	INTF_IE_STAT	0x630
49*4882a593Smuzhiyun #define	INTF_VE_STAT	0x640
50*4882a593Smuzhiyun 
51*4882a593Smuzhiyun /* Interrupt status bits */
52*4882a593Smuzhiyun #define	MGR_INT_TX_NACKED_2	BIT(25)
53*4882a593Smuzhiyun #define	MGR_INT_MSG_BUF_CONTE	BIT(26)
54*4882a593Smuzhiyun #define	MGR_INT_RX_MSG_RCVD	BIT(30)
55*4882a593Smuzhiyun #define	MGR_INT_TX_MSG_SENT	BIT(31)
56*4882a593Smuzhiyun 
57*4882a593Smuzhiyun /* Framer config register settings */
58*4882a593Smuzhiyun #define	FRM_ACTIVE	1
59*4882a593Smuzhiyun #define	CLK_GEAR	7
60*4882a593Smuzhiyun #define	ROOT_FREQ	11
61*4882a593Smuzhiyun #define	REF_CLK_GEAR	15
62*4882a593Smuzhiyun #define	INTR_WAKE	19
63*4882a593Smuzhiyun 
64*4882a593Smuzhiyun #define SLIM_MSG_ASM_FIRST_WORD(l, mt, mc, dt, ad) \
65*4882a593Smuzhiyun 		((l) | ((mt) << 5) | ((mc) << 8) | ((dt) << 15) | ((ad) << 16))
66*4882a593Smuzhiyun 
67*4882a593Smuzhiyun #define SLIM_ROOT_FREQ 24576000
68*4882a593Smuzhiyun #define QCOM_SLIM_AUTOSUSPEND 1000
69*4882a593Smuzhiyun 
70*4882a593Smuzhiyun /* MAX message size over control channel */
71*4882a593Smuzhiyun #define SLIM_MSGQ_BUF_LEN	40
72*4882a593Smuzhiyun #define QCOM_TX_MSGS 2
73*4882a593Smuzhiyun #define QCOM_RX_MSGS	8
74*4882a593Smuzhiyun #define QCOM_BUF_ALLOC_RETRIES	10
75*4882a593Smuzhiyun 
76*4882a593Smuzhiyun #define CFG_PORT(r, v) ((v) ? CFG_PORT_V2(r) : CFG_PORT_V1(r))
77*4882a593Smuzhiyun 
78*4882a593Smuzhiyun /* V2 Component registers */
79*4882a593Smuzhiyun #define CFG_PORT_V2(r) ((r ## _V2))
80*4882a593Smuzhiyun #define	COMP_CFG_V2		4
81*4882a593Smuzhiyun #define	COMP_TRUST_CFG_V2	0x3000
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun /* V1 Component registers */
84*4882a593Smuzhiyun #define CFG_PORT_V1(r) ((r ## _V1))
85*4882a593Smuzhiyun #define	COMP_CFG_V1		0
86*4882a593Smuzhiyun #define	COMP_TRUST_CFG_V1	0x14
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun /* Resource group info for manager, and non-ported generic device-components */
89*4882a593Smuzhiyun #define EE_MGR_RSC_GRP	(1 << 10)
90*4882a593Smuzhiyun #define EE_NGD_2	(2 << 6)
91*4882a593Smuzhiyun #define EE_NGD_1	0
92*4882a593Smuzhiyun 
93*4882a593Smuzhiyun struct slim_ctrl_buf {
94*4882a593Smuzhiyun 	void		*base;
95*4882a593Smuzhiyun 	spinlock_t	lock;
96*4882a593Smuzhiyun 	int		head;
97*4882a593Smuzhiyun 	int		tail;
98*4882a593Smuzhiyun 	int		sl_sz;
99*4882a593Smuzhiyun 	int		n;
100*4882a593Smuzhiyun };
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun struct qcom_slim_ctrl {
103*4882a593Smuzhiyun 	struct slim_controller  ctrl;
104*4882a593Smuzhiyun 	struct slim_framer	framer;
105*4882a593Smuzhiyun 	struct device		*dev;
106*4882a593Smuzhiyun 	void __iomem		*base;
107*4882a593Smuzhiyun 	void __iomem		*slew_reg;
108*4882a593Smuzhiyun 
109*4882a593Smuzhiyun 	struct slim_ctrl_buf	rx;
110*4882a593Smuzhiyun 	struct slim_ctrl_buf	tx;
111*4882a593Smuzhiyun 
112*4882a593Smuzhiyun 	struct completion	**wr_comp;
113*4882a593Smuzhiyun 	int			irq;
114*4882a593Smuzhiyun 	struct workqueue_struct *rxwq;
115*4882a593Smuzhiyun 	struct work_struct	wd;
116*4882a593Smuzhiyun 	struct clk		*rclk;
117*4882a593Smuzhiyun 	struct clk		*hclk;
118*4882a593Smuzhiyun };
119*4882a593Smuzhiyun 
qcom_slim_queue_tx(struct qcom_slim_ctrl * ctrl,void * buf,u8 len,u32 tx_reg)120*4882a593Smuzhiyun static void qcom_slim_queue_tx(struct qcom_slim_ctrl *ctrl, void *buf,
121*4882a593Smuzhiyun 			       u8 len, u32 tx_reg)
122*4882a593Smuzhiyun {
123*4882a593Smuzhiyun 	int count = (len + 3) >> 2;
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 	__iowrite32_copy(ctrl->base + tx_reg, buf, count);
126*4882a593Smuzhiyun 
127*4882a593Smuzhiyun 	/* Ensure Oder of subsequent writes */
128*4882a593Smuzhiyun 	mb();
129*4882a593Smuzhiyun }
130*4882a593Smuzhiyun 
slim_alloc_rxbuf(struct qcom_slim_ctrl * ctrl)131*4882a593Smuzhiyun static void *slim_alloc_rxbuf(struct qcom_slim_ctrl *ctrl)
132*4882a593Smuzhiyun {
133*4882a593Smuzhiyun 	unsigned long flags;
134*4882a593Smuzhiyun 	int idx;
135*4882a593Smuzhiyun 
136*4882a593Smuzhiyun 	spin_lock_irqsave(&ctrl->rx.lock, flags);
137*4882a593Smuzhiyun 	if ((ctrl->rx.tail + 1) % ctrl->rx.n == ctrl->rx.head) {
138*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ctrl->rx.lock, flags);
139*4882a593Smuzhiyun 		dev_err(ctrl->dev, "RX QUEUE full!");
140*4882a593Smuzhiyun 		return NULL;
141*4882a593Smuzhiyun 	}
142*4882a593Smuzhiyun 	idx = ctrl->rx.tail;
143*4882a593Smuzhiyun 	ctrl->rx.tail = (ctrl->rx.tail + 1) % ctrl->rx.n;
144*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ctrl->rx.lock, flags);
145*4882a593Smuzhiyun 
146*4882a593Smuzhiyun 	return ctrl->rx.base + (idx * ctrl->rx.sl_sz);
147*4882a593Smuzhiyun }
148*4882a593Smuzhiyun 
slim_ack_txn(struct qcom_slim_ctrl * ctrl,int err)149*4882a593Smuzhiyun static void slim_ack_txn(struct qcom_slim_ctrl *ctrl, int err)
150*4882a593Smuzhiyun {
151*4882a593Smuzhiyun 	struct completion *comp;
152*4882a593Smuzhiyun 	unsigned long flags;
153*4882a593Smuzhiyun 	int idx;
154*4882a593Smuzhiyun 
155*4882a593Smuzhiyun 	spin_lock_irqsave(&ctrl->tx.lock, flags);
156*4882a593Smuzhiyun 	idx = ctrl->tx.head;
157*4882a593Smuzhiyun 	ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n;
158*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ctrl->tx.lock, flags);
159*4882a593Smuzhiyun 
160*4882a593Smuzhiyun 	comp = ctrl->wr_comp[idx];
161*4882a593Smuzhiyun 	ctrl->wr_comp[idx] = NULL;
162*4882a593Smuzhiyun 
163*4882a593Smuzhiyun 	complete(comp);
164*4882a593Smuzhiyun }
165*4882a593Smuzhiyun 
qcom_slim_handle_tx_irq(struct qcom_slim_ctrl * ctrl,u32 stat)166*4882a593Smuzhiyun static irqreturn_t qcom_slim_handle_tx_irq(struct qcom_slim_ctrl *ctrl,
167*4882a593Smuzhiyun 					   u32 stat)
168*4882a593Smuzhiyun {
169*4882a593Smuzhiyun 	int err = 0;
170*4882a593Smuzhiyun 
171*4882a593Smuzhiyun 	if (stat & MGR_INT_TX_MSG_SENT)
172*4882a593Smuzhiyun 		writel_relaxed(MGR_INT_TX_MSG_SENT,
173*4882a593Smuzhiyun 			       ctrl->base + MGR_INT_CLR);
174*4882a593Smuzhiyun 
175*4882a593Smuzhiyun 	if (stat & MGR_INT_TX_NACKED_2) {
176*4882a593Smuzhiyun 		u32 mgr_stat = readl_relaxed(ctrl->base + MGR_STATUS);
177*4882a593Smuzhiyun 		u32 mgr_ie_stat = readl_relaxed(ctrl->base + MGR_IE_STAT);
178*4882a593Smuzhiyun 		u32 frm_stat = readl_relaxed(ctrl->base + FRM_STAT);
179*4882a593Smuzhiyun 		u32 frm_cfg = readl_relaxed(ctrl->base + FRM_CFG);
180*4882a593Smuzhiyun 		u32 frm_intr_stat = readl_relaxed(ctrl->base + FRM_INT_STAT);
181*4882a593Smuzhiyun 		u32 frm_ie_stat = readl_relaxed(ctrl->base + FRM_IE_STAT);
182*4882a593Smuzhiyun 		u32 intf_stat = readl_relaxed(ctrl->base + INTF_STAT);
183*4882a593Smuzhiyun 		u32 intf_intr_stat = readl_relaxed(ctrl->base + INTF_INT_STAT);
184*4882a593Smuzhiyun 		u32 intf_ie_stat = readl_relaxed(ctrl->base + INTF_IE_STAT);
185*4882a593Smuzhiyun 
186*4882a593Smuzhiyun 		writel_relaxed(MGR_INT_TX_NACKED_2, ctrl->base + MGR_INT_CLR);
187*4882a593Smuzhiyun 
188*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX Nack MGR:int:0x%x, stat:0x%x\n",
189*4882a593Smuzhiyun 			stat, mgr_stat);
190*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX Nack MGR:ie:0x%x\n", mgr_ie_stat);
191*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX Nack FRM:int:0x%x, stat:0x%x\n",
192*4882a593Smuzhiyun 			frm_intr_stat, frm_stat);
193*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX Nack FRM:cfg:0x%x, ie:0x%x\n",
194*4882a593Smuzhiyun 			frm_cfg, frm_ie_stat);
195*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX Nack INTF:intr:0x%x, stat:0x%x\n",
196*4882a593Smuzhiyun 			intf_intr_stat, intf_stat);
197*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX Nack INTF:ie:0x%x\n",
198*4882a593Smuzhiyun 			intf_ie_stat);
199*4882a593Smuzhiyun 		err = -ENOTCONN;
200*4882a593Smuzhiyun 	}
201*4882a593Smuzhiyun 
202*4882a593Smuzhiyun 	slim_ack_txn(ctrl, err);
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun 	return IRQ_HANDLED;
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun 
qcom_slim_handle_rx_irq(struct qcom_slim_ctrl * ctrl,u32 stat)207*4882a593Smuzhiyun static irqreturn_t qcom_slim_handle_rx_irq(struct qcom_slim_ctrl *ctrl,
208*4882a593Smuzhiyun 					   u32 stat)
209*4882a593Smuzhiyun {
210*4882a593Smuzhiyun 	u32 *rx_buf, pkt[10];
211*4882a593Smuzhiyun 	bool q_rx = false;
212*4882a593Smuzhiyun 	u8 mc, mt, len;
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 	pkt[0] = readl_relaxed(ctrl->base + MGR_RX_MSG);
215*4882a593Smuzhiyun 	mt = SLIM_HEADER_GET_MT(pkt[0]);
216*4882a593Smuzhiyun 	len = SLIM_HEADER_GET_RL(pkt[0]);
217*4882a593Smuzhiyun 	mc = SLIM_HEADER_GET_MC(pkt[0]>>8);
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun 	/*
220*4882a593Smuzhiyun 	 * this message cannot be handled by ISR, so
221*4882a593Smuzhiyun 	 * let work-queue handle it
222*4882a593Smuzhiyun 	 */
223*4882a593Smuzhiyun 	if (mt == SLIM_MSG_MT_CORE && mc == SLIM_MSG_MC_REPORT_PRESENT) {
224*4882a593Smuzhiyun 		rx_buf = (u32 *)slim_alloc_rxbuf(ctrl);
225*4882a593Smuzhiyun 		if (!rx_buf) {
226*4882a593Smuzhiyun 			dev_err(ctrl->dev, "dropping RX:0x%x due to RX full\n",
227*4882a593Smuzhiyun 					pkt[0]);
228*4882a593Smuzhiyun 			goto rx_ret_irq;
229*4882a593Smuzhiyun 		}
230*4882a593Smuzhiyun 		rx_buf[0] = pkt[0];
231*4882a593Smuzhiyun 
232*4882a593Smuzhiyun 	} else {
233*4882a593Smuzhiyun 		rx_buf = pkt;
234*4882a593Smuzhiyun 	}
235*4882a593Smuzhiyun 
236*4882a593Smuzhiyun 	__ioread32_copy(rx_buf + 1, ctrl->base + MGR_RX_MSG + 4,
237*4882a593Smuzhiyun 			DIV_ROUND_UP(len, 4));
238*4882a593Smuzhiyun 
239*4882a593Smuzhiyun 	switch (mc) {
240*4882a593Smuzhiyun 
241*4882a593Smuzhiyun 	case SLIM_MSG_MC_REPORT_PRESENT:
242*4882a593Smuzhiyun 		q_rx = true;
243*4882a593Smuzhiyun 		break;
244*4882a593Smuzhiyun 	case SLIM_MSG_MC_REPLY_INFORMATION:
245*4882a593Smuzhiyun 	case SLIM_MSG_MC_REPLY_VALUE:
246*4882a593Smuzhiyun 		slim_msg_response(&ctrl->ctrl, (u8 *)(rx_buf + 1),
247*4882a593Smuzhiyun 				  (u8)(*rx_buf >> 24), (len - 4));
248*4882a593Smuzhiyun 		break;
249*4882a593Smuzhiyun 	default:
250*4882a593Smuzhiyun 		dev_err(ctrl->dev, "unsupported MC,%x MT:%x\n",
251*4882a593Smuzhiyun 			mc, mt);
252*4882a593Smuzhiyun 		break;
253*4882a593Smuzhiyun 	}
254*4882a593Smuzhiyun rx_ret_irq:
255*4882a593Smuzhiyun 	writel(MGR_INT_RX_MSG_RCVD, ctrl->base +
256*4882a593Smuzhiyun 		       MGR_INT_CLR);
257*4882a593Smuzhiyun 	if (q_rx)
258*4882a593Smuzhiyun 		queue_work(ctrl->rxwq, &ctrl->wd);
259*4882a593Smuzhiyun 
260*4882a593Smuzhiyun 	return IRQ_HANDLED;
261*4882a593Smuzhiyun }
262*4882a593Smuzhiyun 
qcom_slim_interrupt(int irq,void * d)263*4882a593Smuzhiyun static irqreturn_t qcom_slim_interrupt(int irq, void *d)
264*4882a593Smuzhiyun {
265*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = d;
266*4882a593Smuzhiyun 	u32 stat = readl_relaxed(ctrl->base + MGR_INT_STAT);
267*4882a593Smuzhiyun 	int ret = IRQ_NONE;
268*4882a593Smuzhiyun 
269*4882a593Smuzhiyun 	if (stat & MGR_INT_TX_MSG_SENT || stat & MGR_INT_TX_NACKED_2)
270*4882a593Smuzhiyun 		ret = qcom_slim_handle_tx_irq(ctrl, stat);
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun 	if (stat & MGR_INT_RX_MSG_RCVD)
273*4882a593Smuzhiyun 		ret = qcom_slim_handle_rx_irq(ctrl, stat);
274*4882a593Smuzhiyun 
275*4882a593Smuzhiyun 	return ret;
276*4882a593Smuzhiyun }
277*4882a593Smuzhiyun 
qcom_clk_pause_wakeup(struct slim_controller * sctrl)278*4882a593Smuzhiyun static int qcom_clk_pause_wakeup(struct slim_controller *sctrl)
279*4882a593Smuzhiyun {
280*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = dev_get_drvdata(sctrl->dev);
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 	clk_prepare_enable(ctrl->hclk);
283*4882a593Smuzhiyun 	clk_prepare_enable(ctrl->rclk);
284*4882a593Smuzhiyun 	enable_irq(ctrl->irq);
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 	writel_relaxed(1, ctrl->base + FRM_WAKEUP);
287*4882a593Smuzhiyun 	/* Make sure framer wakeup write goes through before ISR fires */
288*4882a593Smuzhiyun 	mb();
289*4882a593Smuzhiyun 	/*
290*4882a593Smuzhiyun 	 * HW Workaround: Currently, slave is reporting lost-sync messages
291*4882a593Smuzhiyun 	 * after SLIMbus comes out of clock pause.
292*4882a593Smuzhiyun 	 * Transaction with slave fail before slave reports that message
293*4882a593Smuzhiyun 	 * Give some time for that report to come
294*4882a593Smuzhiyun 	 * SLIMbus wakes up in clock gear 10 at 24.576MHz. With each superframe
295*4882a593Smuzhiyun 	 * being 250 usecs, we wait for 5-10 superframes here to ensure
296*4882a593Smuzhiyun 	 * we get the message
297*4882a593Smuzhiyun 	 */
298*4882a593Smuzhiyun 	usleep_range(1250, 2500);
299*4882a593Smuzhiyun 	return 0;
300*4882a593Smuzhiyun }
301*4882a593Smuzhiyun 
slim_alloc_txbuf(struct qcom_slim_ctrl * ctrl,struct slim_msg_txn * txn,struct completion * done)302*4882a593Smuzhiyun static void *slim_alloc_txbuf(struct qcom_slim_ctrl *ctrl,
303*4882a593Smuzhiyun 			      struct slim_msg_txn *txn,
304*4882a593Smuzhiyun 			      struct completion *done)
305*4882a593Smuzhiyun {
306*4882a593Smuzhiyun 	unsigned long flags;
307*4882a593Smuzhiyun 	int idx;
308*4882a593Smuzhiyun 
309*4882a593Smuzhiyun 	spin_lock_irqsave(&ctrl->tx.lock, flags);
310*4882a593Smuzhiyun 	if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) {
311*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ctrl->tx.lock, flags);
312*4882a593Smuzhiyun 		dev_err(ctrl->dev, "controller TX buf unavailable");
313*4882a593Smuzhiyun 		return NULL;
314*4882a593Smuzhiyun 	}
315*4882a593Smuzhiyun 	idx = ctrl->tx.tail;
316*4882a593Smuzhiyun 	ctrl->wr_comp[idx] = done;
317*4882a593Smuzhiyun 	ctrl->tx.tail = (ctrl->tx.tail + 1) % ctrl->tx.n;
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ctrl->tx.lock, flags);
320*4882a593Smuzhiyun 
321*4882a593Smuzhiyun 	return ctrl->tx.base + (idx * ctrl->tx.sl_sz);
322*4882a593Smuzhiyun }
323*4882a593Smuzhiyun 
324*4882a593Smuzhiyun 
qcom_xfer_msg(struct slim_controller * sctrl,struct slim_msg_txn * txn)325*4882a593Smuzhiyun static int qcom_xfer_msg(struct slim_controller *sctrl,
326*4882a593Smuzhiyun 			 struct slim_msg_txn *txn)
327*4882a593Smuzhiyun {
328*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = dev_get_drvdata(sctrl->dev);
329*4882a593Smuzhiyun 	DECLARE_COMPLETION_ONSTACK(done);
330*4882a593Smuzhiyun 	void *pbuf = slim_alloc_txbuf(ctrl, txn, &done);
331*4882a593Smuzhiyun 	unsigned long ms = txn->rl + HZ;
332*4882a593Smuzhiyun 	u8 *puc;
333*4882a593Smuzhiyun 	int ret = 0, timeout, retries = QCOM_BUF_ALLOC_RETRIES;
334*4882a593Smuzhiyun 	u8 la = txn->la;
335*4882a593Smuzhiyun 	u32 *head;
336*4882a593Smuzhiyun 	/* HW expects length field to be excluded */
337*4882a593Smuzhiyun 	txn->rl--;
338*4882a593Smuzhiyun 
339*4882a593Smuzhiyun 	/* spin till buffer is made available */
340*4882a593Smuzhiyun 	if (!pbuf) {
341*4882a593Smuzhiyun 		while (retries--) {
342*4882a593Smuzhiyun 			usleep_range(10000, 15000);
343*4882a593Smuzhiyun 			pbuf = slim_alloc_txbuf(ctrl, txn, &done);
344*4882a593Smuzhiyun 			if (pbuf)
345*4882a593Smuzhiyun 				break;
346*4882a593Smuzhiyun 		}
347*4882a593Smuzhiyun 	}
348*4882a593Smuzhiyun 
349*4882a593Smuzhiyun 	if (retries < 0 && !pbuf)
350*4882a593Smuzhiyun 		return -ENOMEM;
351*4882a593Smuzhiyun 
352*4882a593Smuzhiyun 	puc = (u8 *)pbuf;
353*4882a593Smuzhiyun 	head = (u32 *)pbuf;
354*4882a593Smuzhiyun 
355*4882a593Smuzhiyun 	if (txn->dt == SLIM_MSG_DEST_LOGICALADDR) {
356*4882a593Smuzhiyun 		*head = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt,
357*4882a593Smuzhiyun 						txn->mc, 0, la);
358*4882a593Smuzhiyun 		puc += 3;
359*4882a593Smuzhiyun 	} else {
360*4882a593Smuzhiyun 		*head = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt,
361*4882a593Smuzhiyun 						txn->mc, 1, la);
362*4882a593Smuzhiyun 		puc += 2;
363*4882a593Smuzhiyun 	}
364*4882a593Smuzhiyun 
365*4882a593Smuzhiyun 	if (slim_tid_txn(txn->mt, txn->mc))
366*4882a593Smuzhiyun 		*(puc++) = txn->tid;
367*4882a593Smuzhiyun 
368*4882a593Smuzhiyun 	if (slim_ec_txn(txn->mt, txn->mc)) {
369*4882a593Smuzhiyun 		*(puc++) = (txn->ec & 0xFF);
370*4882a593Smuzhiyun 		*(puc++) = (txn->ec >> 8) & 0xFF;
371*4882a593Smuzhiyun 	}
372*4882a593Smuzhiyun 
373*4882a593Smuzhiyun 	if (txn->msg && txn->msg->wbuf)
374*4882a593Smuzhiyun 		memcpy(puc, txn->msg->wbuf, txn->msg->num_bytes);
375*4882a593Smuzhiyun 
376*4882a593Smuzhiyun 	qcom_slim_queue_tx(ctrl, head, txn->rl, MGR_TX_MSG);
377*4882a593Smuzhiyun 	timeout = wait_for_completion_timeout(&done, msecs_to_jiffies(ms));
378*4882a593Smuzhiyun 
379*4882a593Smuzhiyun 	if (!timeout) {
380*4882a593Smuzhiyun 		dev_err(ctrl->dev, "TX timed out:MC:0x%x,mt:0x%x", txn->mc,
381*4882a593Smuzhiyun 					txn->mt);
382*4882a593Smuzhiyun 		ret = -ETIMEDOUT;
383*4882a593Smuzhiyun 	}
384*4882a593Smuzhiyun 
385*4882a593Smuzhiyun 	return ret;
386*4882a593Smuzhiyun 
387*4882a593Smuzhiyun }
388*4882a593Smuzhiyun 
qcom_set_laddr(struct slim_controller * sctrl,struct slim_eaddr * ead,u8 laddr)389*4882a593Smuzhiyun static int qcom_set_laddr(struct slim_controller *sctrl,
390*4882a593Smuzhiyun 				struct slim_eaddr *ead, u8 laddr)
391*4882a593Smuzhiyun {
392*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = dev_get_drvdata(sctrl->dev);
393*4882a593Smuzhiyun 	struct {
394*4882a593Smuzhiyun 		__be16 manf_id;
395*4882a593Smuzhiyun 		__be16 prod_code;
396*4882a593Smuzhiyun 		u8 dev_index;
397*4882a593Smuzhiyun 		u8 instance;
398*4882a593Smuzhiyun 		u8 laddr;
399*4882a593Smuzhiyun 	} __packed p;
400*4882a593Smuzhiyun 	struct slim_val_inf msg = {0};
401*4882a593Smuzhiyun 	DEFINE_SLIM_EDEST_TXN(txn, SLIM_MSG_MC_ASSIGN_LOGICAL_ADDRESS,
402*4882a593Smuzhiyun 			      10, laddr, &msg);
403*4882a593Smuzhiyun 	int ret;
404*4882a593Smuzhiyun 
405*4882a593Smuzhiyun 	p.manf_id = cpu_to_be16(ead->manf_id);
406*4882a593Smuzhiyun 	p.prod_code = cpu_to_be16(ead->prod_code);
407*4882a593Smuzhiyun 	p.dev_index = ead->dev_index;
408*4882a593Smuzhiyun 	p.instance = ead->instance;
409*4882a593Smuzhiyun 	p.laddr = laddr;
410*4882a593Smuzhiyun 
411*4882a593Smuzhiyun 	msg.wbuf = (void *)&p;
412*4882a593Smuzhiyun 	msg.num_bytes = 7;
413*4882a593Smuzhiyun 	ret = slim_do_transfer(&ctrl->ctrl, &txn);
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun 	if (ret)
416*4882a593Smuzhiyun 		dev_err(ctrl->dev, "set LA:0x%x failed:ret:%d\n",
417*4882a593Smuzhiyun 				  laddr, ret);
418*4882a593Smuzhiyun 	return ret;
419*4882a593Smuzhiyun }
420*4882a593Smuzhiyun 
slim_get_current_rxbuf(struct qcom_slim_ctrl * ctrl,void * buf)421*4882a593Smuzhiyun static int slim_get_current_rxbuf(struct qcom_slim_ctrl *ctrl, void *buf)
422*4882a593Smuzhiyun {
423*4882a593Smuzhiyun 	unsigned long flags;
424*4882a593Smuzhiyun 
425*4882a593Smuzhiyun 	spin_lock_irqsave(&ctrl->rx.lock, flags);
426*4882a593Smuzhiyun 	if (ctrl->rx.tail == ctrl->rx.head) {
427*4882a593Smuzhiyun 		spin_unlock_irqrestore(&ctrl->rx.lock, flags);
428*4882a593Smuzhiyun 		return -ENODATA;
429*4882a593Smuzhiyun 	}
430*4882a593Smuzhiyun 	memcpy(buf, ctrl->rx.base + (ctrl->rx.head * ctrl->rx.sl_sz),
431*4882a593Smuzhiyun 				ctrl->rx.sl_sz);
432*4882a593Smuzhiyun 
433*4882a593Smuzhiyun 	ctrl->rx.head = (ctrl->rx.head + 1) % ctrl->rx.n;
434*4882a593Smuzhiyun 	spin_unlock_irqrestore(&ctrl->rx.lock, flags);
435*4882a593Smuzhiyun 
436*4882a593Smuzhiyun 	return 0;
437*4882a593Smuzhiyun }
438*4882a593Smuzhiyun 
qcom_slim_rxwq(struct work_struct * work)439*4882a593Smuzhiyun static void qcom_slim_rxwq(struct work_struct *work)
440*4882a593Smuzhiyun {
441*4882a593Smuzhiyun 	u8 buf[SLIM_MSGQ_BUF_LEN];
442*4882a593Smuzhiyun 	u8 mc, mt;
443*4882a593Smuzhiyun 	int ret;
444*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = container_of(work, struct qcom_slim_ctrl,
445*4882a593Smuzhiyun 						 wd);
446*4882a593Smuzhiyun 
447*4882a593Smuzhiyun 	while ((slim_get_current_rxbuf(ctrl, buf)) != -ENODATA) {
448*4882a593Smuzhiyun 		mt = SLIM_HEADER_GET_MT(buf[0]);
449*4882a593Smuzhiyun 		mc = SLIM_HEADER_GET_MC(buf[1]);
450*4882a593Smuzhiyun 		if (mt == SLIM_MSG_MT_CORE &&
451*4882a593Smuzhiyun 			mc == SLIM_MSG_MC_REPORT_PRESENT) {
452*4882a593Smuzhiyun 			struct slim_eaddr ea;
453*4882a593Smuzhiyun 			u8 laddr;
454*4882a593Smuzhiyun 
455*4882a593Smuzhiyun 			ea.manf_id = be16_to_cpup((__be16 *)&buf[2]);
456*4882a593Smuzhiyun 			ea.prod_code = be16_to_cpup((__be16 *)&buf[4]);
457*4882a593Smuzhiyun 			ea.dev_index = buf[6];
458*4882a593Smuzhiyun 			ea.instance = buf[7];
459*4882a593Smuzhiyun 
460*4882a593Smuzhiyun 			ret = slim_device_report_present(&ctrl->ctrl, &ea,
461*4882a593Smuzhiyun 							 &laddr);
462*4882a593Smuzhiyun 			if (ret < 0)
463*4882a593Smuzhiyun 				dev_err(ctrl->dev, "assign laddr failed:%d\n",
464*4882a593Smuzhiyun 					ret);
465*4882a593Smuzhiyun 		} else {
466*4882a593Smuzhiyun 			dev_err(ctrl->dev, "unexpected message:mc:%x, mt:%x\n",
467*4882a593Smuzhiyun 				mc, mt);
468*4882a593Smuzhiyun 		}
469*4882a593Smuzhiyun 	}
470*4882a593Smuzhiyun }
471*4882a593Smuzhiyun 
qcom_slim_prg_slew(struct platform_device * pdev,struct qcom_slim_ctrl * ctrl)472*4882a593Smuzhiyun static void qcom_slim_prg_slew(struct platform_device *pdev,
473*4882a593Smuzhiyun 				struct qcom_slim_ctrl *ctrl)
474*4882a593Smuzhiyun {
475*4882a593Smuzhiyun 	if (!ctrl->slew_reg) {
476*4882a593Smuzhiyun 		/* SLEW RATE register for this SLIMbus */
477*4882a593Smuzhiyun 		ctrl->slew_reg = devm_platform_ioremap_resource_byname(pdev, "slew");
478*4882a593Smuzhiyun 		if (IS_ERR(ctrl->slew_reg))
479*4882a593Smuzhiyun 			return;
480*4882a593Smuzhiyun 	}
481*4882a593Smuzhiyun 
482*4882a593Smuzhiyun 	writel_relaxed(1, ctrl->slew_reg);
483*4882a593Smuzhiyun 	/* Make sure SLIMbus-slew rate enabling goes through */
484*4882a593Smuzhiyun 	wmb();
485*4882a593Smuzhiyun }
486*4882a593Smuzhiyun 
qcom_slim_probe(struct platform_device * pdev)487*4882a593Smuzhiyun static int qcom_slim_probe(struct platform_device *pdev)
488*4882a593Smuzhiyun {
489*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl;
490*4882a593Smuzhiyun 	struct slim_controller *sctrl;
491*4882a593Smuzhiyun 	struct resource *slim_mem;
492*4882a593Smuzhiyun 	int ret, ver;
493*4882a593Smuzhiyun 
494*4882a593Smuzhiyun 	ctrl = devm_kzalloc(&pdev->dev, sizeof(*ctrl), GFP_KERNEL);
495*4882a593Smuzhiyun 	if (!ctrl)
496*4882a593Smuzhiyun 		return -ENOMEM;
497*4882a593Smuzhiyun 
498*4882a593Smuzhiyun 	ctrl->hclk = devm_clk_get(&pdev->dev, "iface");
499*4882a593Smuzhiyun 	if (IS_ERR(ctrl->hclk))
500*4882a593Smuzhiyun 		return PTR_ERR(ctrl->hclk);
501*4882a593Smuzhiyun 
502*4882a593Smuzhiyun 	ctrl->rclk = devm_clk_get(&pdev->dev, "core");
503*4882a593Smuzhiyun 	if (IS_ERR(ctrl->rclk))
504*4882a593Smuzhiyun 		return PTR_ERR(ctrl->rclk);
505*4882a593Smuzhiyun 
506*4882a593Smuzhiyun 	ret = clk_set_rate(ctrl->rclk, SLIM_ROOT_FREQ);
507*4882a593Smuzhiyun 	if (ret) {
508*4882a593Smuzhiyun 		dev_err(&pdev->dev, "ref-clock set-rate failed:%d\n", ret);
509*4882a593Smuzhiyun 		return ret;
510*4882a593Smuzhiyun 	}
511*4882a593Smuzhiyun 
512*4882a593Smuzhiyun 	ctrl->irq = platform_get_irq(pdev, 0);
513*4882a593Smuzhiyun 	if (ctrl->irq < 0) {
514*4882a593Smuzhiyun 		dev_err(&pdev->dev, "no slimbus IRQ\n");
515*4882a593Smuzhiyun 		return ctrl->irq;
516*4882a593Smuzhiyun 	}
517*4882a593Smuzhiyun 
518*4882a593Smuzhiyun 	sctrl = &ctrl->ctrl;
519*4882a593Smuzhiyun 	sctrl->dev = &pdev->dev;
520*4882a593Smuzhiyun 	ctrl->dev = &pdev->dev;
521*4882a593Smuzhiyun 	platform_set_drvdata(pdev, ctrl);
522*4882a593Smuzhiyun 	dev_set_drvdata(ctrl->dev, ctrl);
523*4882a593Smuzhiyun 
524*4882a593Smuzhiyun 	slim_mem = platform_get_resource_byname(pdev, IORESOURCE_MEM, "ctrl");
525*4882a593Smuzhiyun 	ctrl->base = devm_ioremap_resource(ctrl->dev, slim_mem);
526*4882a593Smuzhiyun 	if (IS_ERR(ctrl->base))
527*4882a593Smuzhiyun 		return PTR_ERR(ctrl->base);
528*4882a593Smuzhiyun 
529*4882a593Smuzhiyun 	sctrl->set_laddr = qcom_set_laddr;
530*4882a593Smuzhiyun 	sctrl->xfer_msg = qcom_xfer_msg;
531*4882a593Smuzhiyun 	sctrl->wakeup =  qcom_clk_pause_wakeup;
532*4882a593Smuzhiyun 	ctrl->tx.n = QCOM_TX_MSGS;
533*4882a593Smuzhiyun 	ctrl->tx.sl_sz = SLIM_MSGQ_BUF_LEN;
534*4882a593Smuzhiyun 	ctrl->rx.n = QCOM_RX_MSGS;
535*4882a593Smuzhiyun 	ctrl->rx.sl_sz = SLIM_MSGQ_BUF_LEN;
536*4882a593Smuzhiyun 	ctrl->wr_comp = kcalloc(QCOM_TX_MSGS, sizeof(struct completion *),
537*4882a593Smuzhiyun 				GFP_KERNEL);
538*4882a593Smuzhiyun 	if (!ctrl->wr_comp)
539*4882a593Smuzhiyun 		return -ENOMEM;
540*4882a593Smuzhiyun 
541*4882a593Smuzhiyun 	spin_lock_init(&ctrl->rx.lock);
542*4882a593Smuzhiyun 	spin_lock_init(&ctrl->tx.lock);
543*4882a593Smuzhiyun 	INIT_WORK(&ctrl->wd, qcom_slim_rxwq);
544*4882a593Smuzhiyun 	ctrl->rxwq = create_singlethread_workqueue("qcom_slim_rx");
545*4882a593Smuzhiyun 	if (!ctrl->rxwq) {
546*4882a593Smuzhiyun 		dev_err(ctrl->dev, "Failed to start Rx WQ\n");
547*4882a593Smuzhiyun 		return -ENOMEM;
548*4882a593Smuzhiyun 	}
549*4882a593Smuzhiyun 
550*4882a593Smuzhiyun 	ctrl->framer.rootfreq = SLIM_ROOT_FREQ / 8;
551*4882a593Smuzhiyun 	ctrl->framer.superfreq =
552*4882a593Smuzhiyun 		ctrl->framer.rootfreq / SLIM_CL_PER_SUPERFRAME_DIV8;
553*4882a593Smuzhiyun 	sctrl->a_framer = &ctrl->framer;
554*4882a593Smuzhiyun 	sctrl->clkgear = SLIM_MAX_CLK_GEAR;
555*4882a593Smuzhiyun 
556*4882a593Smuzhiyun 	qcom_slim_prg_slew(pdev, ctrl);
557*4882a593Smuzhiyun 
558*4882a593Smuzhiyun 	ret = devm_request_irq(&pdev->dev, ctrl->irq, qcom_slim_interrupt,
559*4882a593Smuzhiyun 				IRQF_TRIGGER_HIGH, "qcom_slim_irq", ctrl);
560*4882a593Smuzhiyun 	if (ret) {
561*4882a593Smuzhiyun 		dev_err(&pdev->dev, "request IRQ failed\n");
562*4882a593Smuzhiyun 		goto err_request_irq_failed;
563*4882a593Smuzhiyun 	}
564*4882a593Smuzhiyun 
565*4882a593Smuzhiyun 	ret = clk_prepare_enable(ctrl->hclk);
566*4882a593Smuzhiyun 	if (ret)
567*4882a593Smuzhiyun 		goto err_hclk_enable_failed;
568*4882a593Smuzhiyun 
569*4882a593Smuzhiyun 	ret = clk_prepare_enable(ctrl->rclk);
570*4882a593Smuzhiyun 	if (ret)
571*4882a593Smuzhiyun 		goto err_rclk_enable_failed;
572*4882a593Smuzhiyun 
573*4882a593Smuzhiyun 	ctrl->tx.base = devm_kcalloc(&pdev->dev, ctrl->tx.n, ctrl->tx.sl_sz,
574*4882a593Smuzhiyun 				     GFP_KERNEL);
575*4882a593Smuzhiyun 	if (!ctrl->tx.base) {
576*4882a593Smuzhiyun 		ret = -ENOMEM;
577*4882a593Smuzhiyun 		goto err;
578*4882a593Smuzhiyun 	}
579*4882a593Smuzhiyun 
580*4882a593Smuzhiyun 	ctrl->rx.base = devm_kcalloc(&pdev->dev,ctrl->rx.n, ctrl->rx.sl_sz,
581*4882a593Smuzhiyun 				     GFP_KERNEL);
582*4882a593Smuzhiyun 	if (!ctrl->rx.base) {
583*4882a593Smuzhiyun 		ret = -ENOMEM;
584*4882a593Smuzhiyun 		goto err;
585*4882a593Smuzhiyun 	}
586*4882a593Smuzhiyun 
587*4882a593Smuzhiyun 	/* Register with framework before enabling frame, clock */
588*4882a593Smuzhiyun 	ret = slim_register_controller(&ctrl->ctrl);
589*4882a593Smuzhiyun 	if (ret) {
590*4882a593Smuzhiyun 		dev_err(ctrl->dev, "error adding controller\n");
591*4882a593Smuzhiyun 		goto err;
592*4882a593Smuzhiyun 	}
593*4882a593Smuzhiyun 
594*4882a593Smuzhiyun 	ver = readl_relaxed(ctrl->base);
595*4882a593Smuzhiyun 	/* Version info in 16 MSbits */
596*4882a593Smuzhiyun 	ver >>= 16;
597*4882a593Smuzhiyun 	/* Component register initialization */
598*4882a593Smuzhiyun 	writel(1, ctrl->base + CFG_PORT(COMP_CFG, ver));
599*4882a593Smuzhiyun 	writel((EE_MGR_RSC_GRP | EE_NGD_2 | EE_NGD_1),
600*4882a593Smuzhiyun 				ctrl->base + CFG_PORT(COMP_TRUST_CFG, ver));
601*4882a593Smuzhiyun 
602*4882a593Smuzhiyun 	writel((MGR_INT_TX_NACKED_2 |
603*4882a593Smuzhiyun 			MGR_INT_MSG_BUF_CONTE | MGR_INT_RX_MSG_RCVD |
604*4882a593Smuzhiyun 			MGR_INT_TX_MSG_SENT), ctrl->base + MGR_INT_EN);
605*4882a593Smuzhiyun 	writel(1, ctrl->base + MGR_CFG);
606*4882a593Smuzhiyun 	/* Framer register initialization */
607*4882a593Smuzhiyun 	writel((1 << INTR_WAKE) | (0xA << REF_CLK_GEAR) |
608*4882a593Smuzhiyun 		(0xA << CLK_GEAR) | (1 << ROOT_FREQ) | (1 << FRM_ACTIVE) | 1,
609*4882a593Smuzhiyun 		ctrl->base + FRM_CFG);
610*4882a593Smuzhiyun 	writel(MGR_CFG_ENABLE, ctrl->base + MGR_CFG);
611*4882a593Smuzhiyun 	writel(1, ctrl->base + INTF_CFG);
612*4882a593Smuzhiyun 	writel(1, ctrl->base + CFG_PORT(COMP_CFG, ver));
613*4882a593Smuzhiyun 
614*4882a593Smuzhiyun 	pm_runtime_use_autosuspend(&pdev->dev);
615*4882a593Smuzhiyun 	pm_runtime_set_autosuspend_delay(&pdev->dev, QCOM_SLIM_AUTOSUSPEND);
616*4882a593Smuzhiyun 	pm_runtime_set_active(&pdev->dev);
617*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(&pdev->dev);
618*4882a593Smuzhiyun 	pm_runtime_enable(&pdev->dev);
619*4882a593Smuzhiyun 
620*4882a593Smuzhiyun 	dev_dbg(ctrl->dev, "QCOM SB controller is up:ver:0x%x!\n", ver);
621*4882a593Smuzhiyun 	return 0;
622*4882a593Smuzhiyun 
623*4882a593Smuzhiyun err:
624*4882a593Smuzhiyun 	clk_disable_unprepare(ctrl->rclk);
625*4882a593Smuzhiyun err_rclk_enable_failed:
626*4882a593Smuzhiyun 	clk_disable_unprepare(ctrl->hclk);
627*4882a593Smuzhiyun err_hclk_enable_failed:
628*4882a593Smuzhiyun err_request_irq_failed:
629*4882a593Smuzhiyun 	destroy_workqueue(ctrl->rxwq);
630*4882a593Smuzhiyun 	return ret;
631*4882a593Smuzhiyun }
632*4882a593Smuzhiyun 
qcom_slim_remove(struct platform_device * pdev)633*4882a593Smuzhiyun static int qcom_slim_remove(struct platform_device *pdev)
634*4882a593Smuzhiyun {
635*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = platform_get_drvdata(pdev);
636*4882a593Smuzhiyun 
637*4882a593Smuzhiyun 	pm_runtime_disable(&pdev->dev);
638*4882a593Smuzhiyun 	slim_unregister_controller(&ctrl->ctrl);
639*4882a593Smuzhiyun 	clk_disable_unprepare(ctrl->rclk);
640*4882a593Smuzhiyun 	clk_disable_unprepare(ctrl->hclk);
641*4882a593Smuzhiyun 	destroy_workqueue(ctrl->rxwq);
642*4882a593Smuzhiyun 	return 0;
643*4882a593Smuzhiyun }
644*4882a593Smuzhiyun 
645*4882a593Smuzhiyun /*
646*4882a593Smuzhiyun  * If PM_RUNTIME is not defined, these 2 functions become helper
647*4882a593Smuzhiyun  * functions to be called from system suspend/resume.
648*4882a593Smuzhiyun  */
649*4882a593Smuzhiyun #ifdef CONFIG_PM
qcom_slim_runtime_suspend(struct device * device)650*4882a593Smuzhiyun static int qcom_slim_runtime_suspend(struct device *device)
651*4882a593Smuzhiyun {
652*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = dev_get_drvdata(device);
653*4882a593Smuzhiyun 	int ret;
654*4882a593Smuzhiyun 
655*4882a593Smuzhiyun 	dev_dbg(device, "pm_runtime: suspending...\n");
656*4882a593Smuzhiyun 	ret = slim_ctrl_clk_pause(&ctrl->ctrl, false, SLIM_CLK_UNSPECIFIED);
657*4882a593Smuzhiyun 	if (ret) {
658*4882a593Smuzhiyun 		dev_err(device, "clk pause not entered:%d", ret);
659*4882a593Smuzhiyun 	} else {
660*4882a593Smuzhiyun 		disable_irq(ctrl->irq);
661*4882a593Smuzhiyun 		clk_disable_unprepare(ctrl->hclk);
662*4882a593Smuzhiyun 		clk_disable_unprepare(ctrl->rclk);
663*4882a593Smuzhiyun 	}
664*4882a593Smuzhiyun 	return ret;
665*4882a593Smuzhiyun }
666*4882a593Smuzhiyun 
qcom_slim_runtime_resume(struct device * device)667*4882a593Smuzhiyun static int qcom_slim_runtime_resume(struct device *device)
668*4882a593Smuzhiyun {
669*4882a593Smuzhiyun 	struct qcom_slim_ctrl *ctrl = dev_get_drvdata(device);
670*4882a593Smuzhiyun 	int ret = 0;
671*4882a593Smuzhiyun 
672*4882a593Smuzhiyun 	dev_dbg(device, "pm_runtime: resuming...\n");
673*4882a593Smuzhiyun 	ret = slim_ctrl_clk_pause(&ctrl->ctrl, true, 0);
674*4882a593Smuzhiyun 	if (ret)
675*4882a593Smuzhiyun 		dev_err(device, "clk pause not exited:%d", ret);
676*4882a593Smuzhiyun 	return ret;
677*4882a593Smuzhiyun }
678*4882a593Smuzhiyun #endif
679*4882a593Smuzhiyun 
680*4882a593Smuzhiyun #ifdef CONFIG_PM_SLEEP
qcom_slim_suspend(struct device * dev)681*4882a593Smuzhiyun static int qcom_slim_suspend(struct device *dev)
682*4882a593Smuzhiyun {
683*4882a593Smuzhiyun 	int ret = 0;
684*4882a593Smuzhiyun 
685*4882a593Smuzhiyun 	if (!pm_runtime_enabled(dev) ||
686*4882a593Smuzhiyun 		(!pm_runtime_suspended(dev))) {
687*4882a593Smuzhiyun 		dev_dbg(dev, "system suspend");
688*4882a593Smuzhiyun 		ret = qcom_slim_runtime_suspend(dev);
689*4882a593Smuzhiyun 	}
690*4882a593Smuzhiyun 
691*4882a593Smuzhiyun 	return ret;
692*4882a593Smuzhiyun }
693*4882a593Smuzhiyun 
qcom_slim_resume(struct device * dev)694*4882a593Smuzhiyun static int qcom_slim_resume(struct device *dev)
695*4882a593Smuzhiyun {
696*4882a593Smuzhiyun 	if (!pm_runtime_enabled(dev) || !pm_runtime_suspended(dev)) {
697*4882a593Smuzhiyun 		int ret;
698*4882a593Smuzhiyun 
699*4882a593Smuzhiyun 		dev_dbg(dev, "system resume");
700*4882a593Smuzhiyun 		ret = qcom_slim_runtime_resume(dev);
701*4882a593Smuzhiyun 		if (!ret) {
702*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(dev);
703*4882a593Smuzhiyun 			pm_request_autosuspend(dev);
704*4882a593Smuzhiyun 		}
705*4882a593Smuzhiyun 		return ret;
706*4882a593Smuzhiyun 
707*4882a593Smuzhiyun 	}
708*4882a593Smuzhiyun 	return 0;
709*4882a593Smuzhiyun }
710*4882a593Smuzhiyun #endif /* CONFIG_PM_SLEEP */
711*4882a593Smuzhiyun 
712*4882a593Smuzhiyun static const struct dev_pm_ops qcom_slim_dev_pm_ops = {
713*4882a593Smuzhiyun 	SET_SYSTEM_SLEEP_PM_OPS(qcom_slim_suspend, qcom_slim_resume)
714*4882a593Smuzhiyun 	SET_RUNTIME_PM_OPS(
715*4882a593Smuzhiyun 			   qcom_slim_runtime_suspend,
716*4882a593Smuzhiyun 			   qcom_slim_runtime_resume,
717*4882a593Smuzhiyun 			   NULL
718*4882a593Smuzhiyun 	)
719*4882a593Smuzhiyun };
720*4882a593Smuzhiyun 
721*4882a593Smuzhiyun static const struct of_device_id qcom_slim_dt_match[] = {
722*4882a593Smuzhiyun 	{ .compatible = "qcom,slim", },
723*4882a593Smuzhiyun 	{ .compatible = "qcom,apq8064-slim", },
724*4882a593Smuzhiyun 	{}
725*4882a593Smuzhiyun };
726*4882a593Smuzhiyun 
727*4882a593Smuzhiyun static struct platform_driver qcom_slim_driver = {
728*4882a593Smuzhiyun 	.probe = qcom_slim_probe,
729*4882a593Smuzhiyun 	.remove = qcom_slim_remove,
730*4882a593Smuzhiyun 	.driver	= {
731*4882a593Smuzhiyun 		.name = "qcom_slim_ctrl",
732*4882a593Smuzhiyun 		.of_match_table = qcom_slim_dt_match,
733*4882a593Smuzhiyun 		.pm = &qcom_slim_dev_pm_ops,
734*4882a593Smuzhiyun 	},
735*4882a593Smuzhiyun };
736*4882a593Smuzhiyun module_platform_driver(qcom_slim_driver);
737*4882a593Smuzhiyun 
738*4882a593Smuzhiyun MODULE_LICENSE("GPL v2");
739*4882a593Smuzhiyun MODULE_DESCRIPTION("Qualcomm SLIMbus Controller");
740