1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun //
3*4882a593Smuzhiyun // Copyright(c) 2020 Intel Corporation. All rights reserved.
4*4882a593Smuzhiyun //
5*4882a593Smuzhiyun // Author: Cezary Rojewski <cezary.rojewski@intel.com>
6*4882a593Smuzhiyun //
7*4882a593Smuzhiyun
8*4882a593Smuzhiyun #include <linux/devcoredump.h>
9*4882a593Smuzhiyun #include <linux/dma-mapping.h>
10*4882a593Smuzhiyun #include <linux/firmware.h>
11*4882a593Smuzhiyun #include <linux/pci.h>
12*4882a593Smuzhiyun #include <linux/pxa2xx_ssp.h>
13*4882a593Smuzhiyun #include "core.h"
14*4882a593Smuzhiyun #include "messages.h"
15*4882a593Smuzhiyun #include "registers.h"
16*4882a593Smuzhiyun
catpt_dma_filter(struct dma_chan * chan,void * param)17*4882a593Smuzhiyun static bool catpt_dma_filter(struct dma_chan *chan, void *param)
18*4882a593Smuzhiyun {
19*4882a593Smuzhiyun return param == chan->device->dev;
20*4882a593Smuzhiyun }
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun /*
23*4882a593Smuzhiyun * Either engine 0 or 1 can be used for image loading.
24*4882a593Smuzhiyun * Align with Windows driver equivalent and stick to engine 1.
25*4882a593Smuzhiyun */
26*4882a593Smuzhiyun #define CATPT_DMA_DEVID 1
27*4882a593Smuzhiyun #define CATPT_DMA_DSP_ADDR_MASK GENMASK(31, 20)
28*4882a593Smuzhiyun
catpt_dma_request_config_chan(struct catpt_dev * cdev)29*4882a593Smuzhiyun struct dma_chan *catpt_dma_request_config_chan(struct catpt_dev *cdev)
30*4882a593Smuzhiyun {
31*4882a593Smuzhiyun struct dma_slave_config config;
32*4882a593Smuzhiyun struct dma_chan *chan;
33*4882a593Smuzhiyun dma_cap_mask_t mask;
34*4882a593Smuzhiyun int ret;
35*4882a593Smuzhiyun
36*4882a593Smuzhiyun dma_cap_zero(mask);
37*4882a593Smuzhiyun dma_cap_set(DMA_MEMCPY, mask);
38*4882a593Smuzhiyun
39*4882a593Smuzhiyun chan = dma_request_channel(mask, catpt_dma_filter, cdev->dev);
40*4882a593Smuzhiyun if (!chan) {
41*4882a593Smuzhiyun dev_err(cdev->dev, "request channel failed\n");
42*4882a593Smuzhiyun return ERR_PTR(-ENODEV);
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun
45*4882a593Smuzhiyun memset(&config, 0, sizeof(config));
46*4882a593Smuzhiyun config.direction = DMA_MEM_TO_DEV;
47*4882a593Smuzhiyun config.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
48*4882a593Smuzhiyun config.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
49*4882a593Smuzhiyun config.src_maxburst = 16;
50*4882a593Smuzhiyun config.dst_maxburst = 16;
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun ret = dmaengine_slave_config(chan, &config);
53*4882a593Smuzhiyun if (ret) {
54*4882a593Smuzhiyun dev_err(cdev->dev, "slave config failed: %d\n", ret);
55*4882a593Smuzhiyun dma_release_channel(chan);
56*4882a593Smuzhiyun return ERR_PTR(ret);
57*4882a593Smuzhiyun }
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun return chan;
60*4882a593Smuzhiyun }
61*4882a593Smuzhiyun
catpt_dma_memcpy(struct catpt_dev * cdev,struct dma_chan * chan,dma_addr_t dst_addr,dma_addr_t src_addr,size_t size)62*4882a593Smuzhiyun static int catpt_dma_memcpy(struct catpt_dev *cdev, struct dma_chan *chan,
63*4882a593Smuzhiyun dma_addr_t dst_addr, dma_addr_t src_addr,
64*4882a593Smuzhiyun size_t size)
65*4882a593Smuzhiyun {
66*4882a593Smuzhiyun struct dma_async_tx_descriptor *desc;
67*4882a593Smuzhiyun enum dma_status status;
68*4882a593Smuzhiyun int ret;
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun desc = dmaengine_prep_dma_memcpy(chan, dst_addr, src_addr, size,
71*4882a593Smuzhiyun DMA_CTRL_ACK);
72*4882a593Smuzhiyun if (!desc) {
73*4882a593Smuzhiyun dev_err(cdev->dev, "prep dma memcpy failed\n");
74*4882a593Smuzhiyun return -EIO;
75*4882a593Smuzhiyun }
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun /* enable demand mode for dma channel */
78*4882a593Smuzhiyun catpt_updatel_shim(cdev, HMDC,
79*4882a593Smuzhiyun CATPT_HMDC_HDDA(CATPT_DMA_DEVID, chan->chan_id),
80*4882a593Smuzhiyun CATPT_HMDC_HDDA(CATPT_DMA_DEVID, chan->chan_id));
81*4882a593Smuzhiyun
82*4882a593Smuzhiyun ret = dma_submit_error(dmaengine_submit(desc));
83*4882a593Smuzhiyun if (ret) {
84*4882a593Smuzhiyun dev_err(cdev->dev, "submit tx failed: %d\n", ret);
85*4882a593Smuzhiyun goto clear_hdda;
86*4882a593Smuzhiyun }
87*4882a593Smuzhiyun
88*4882a593Smuzhiyun status = dma_wait_for_async_tx(desc);
89*4882a593Smuzhiyun ret = (status == DMA_COMPLETE) ? 0 : -EPROTO;
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun clear_hdda:
92*4882a593Smuzhiyun /* regardless of status, disable access to HOST memory in demand mode */
93*4882a593Smuzhiyun catpt_updatel_shim(cdev, HMDC,
94*4882a593Smuzhiyun CATPT_HMDC_HDDA(CATPT_DMA_DEVID, chan->chan_id), 0);
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun return ret;
97*4882a593Smuzhiyun }
98*4882a593Smuzhiyun
catpt_dma_memcpy_todsp(struct catpt_dev * cdev,struct dma_chan * chan,dma_addr_t dst_addr,dma_addr_t src_addr,size_t size)99*4882a593Smuzhiyun int catpt_dma_memcpy_todsp(struct catpt_dev *cdev, struct dma_chan *chan,
100*4882a593Smuzhiyun dma_addr_t dst_addr, dma_addr_t src_addr,
101*4882a593Smuzhiyun size_t size)
102*4882a593Smuzhiyun {
103*4882a593Smuzhiyun return catpt_dma_memcpy(cdev, chan, dst_addr | CATPT_DMA_DSP_ADDR_MASK,
104*4882a593Smuzhiyun src_addr, size);
105*4882a593Smuzhiyun }
106*4882a593Smuzhiyun
catpt_dma_memcpy_fromdsp(struct catpt_dev * cdev,struct dma_chan * chan,dma_addr_t dst_addr,dma_addr_t src_addr,size_t size)107*4882a593Smuzhiyun int catpt_dma_memcpy_fromdsp(struct catpt_dev *cdev, struct dma_chan *chan,
108*4882a593Smuzhiyun dma_addr_t dst_addr, dma_addr_t src_addr,
109*4882a593Smuzhiyun size_t size)
110*4882a593Smuzhiyun {
111*4882a593Smuzhiyun return catpt_dma_memcpy(cdev, chan, dst_addr,
112*4882a593Smuzhiyun src_addr | CATPT_DMA_DSP_ADDR_MASK, size);
113*4882a593Smuzhiyun }
114*4882a593Smuzhiyun
catpt_dmac_probe(struct catpt_dev * cdev)115*4882a593Smuzhiyun int catpt_dmac_probe(struct catpt_dev *cdev)
116*4882a593Smuzhiyun {
117*4882a593Smuzhiyun struct dw_dma_chip *dmac;
118*4882a593Smuzhiyun int ret;
119*4882a593Smuzhiyun
120*4882a593Smuzhiyun dmac = devm_kzalloc(cdev->dev, sizeof(*dmac), GFP_KERNEL);
121*4882a593Smuzhiyun if (!dmac)
122*4882a593Smuzhiyun return -ENOMEM;
123*4882a593Smuzhiyun
124*4882a593Smuzhiyun dmac->regs = cdev->lpe_ba + cdev->spec->host_dma_offset[CATPT_DMA_DEVID];
125*4882a593Smuzhiyun dmac->dev = cdev->dev;
126*4882a593Smuzhiyun dmac->irq = cdev->irq;
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun ret = dma_coerce_mask_and_coherent(cdev->dev, DMA_BIT_MASK(31));
129*4882a593Smuzhiyun if (ret)
130*4882a593Smuzhiyun return ret;
131*4882a593Smuzhiyun /*
132*4882a593Smuzhiyun * Caller is responsible for putting device in D0 to allow
133*4882a593Smuzhiyun * for I/O and memory access before probing DW.
134*4882a593Smuzhiyun */
135*4882a593Smuzhiyun ret = dw_dma_probe(dmac);
136*4882a593Smuzhiyun if (ret)
137*4882a593Smuzhiyun return ret;
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun cdev->dmac = dmac;
140*4882a593Smuzhiyun return 0;
141*4882a593Smuzhiyun }
142*4882a593Smuzhiyun
catpt_dmac_remove(struct catpt_dev * cdev)143*4882a593Smuzhiyun void catpt_dmac_remove(struct catpt_dev *cdev)
144*4882a593Smuzhiyun {
145*4882a593Smuzhiyun /*
146*4882a593Smuzhiyun * As do_dma_remove() juggles with pm_runtime_get_xxx() and
147*4882a593Smuzhiyun * pm_runtime_put_xxx() while both ADSP and DW 'devices' are part of
148*4882a593Smuzhiyun * the same module, caller makes sure pm_runtime_disable() is invoked
149*4882a593Smuzhiyun * before removing DW to prevent postmortem resume and suspend.
150*4882a593Smuzhiyun */
151*4882a593Smuzhiyun dw_dma_remove(cdev->dmac);
152*4882a593Smuzhiyun }
153*4882a593Smuzhiyun
catpt_dsp_set_srampge(struct catpt_dev * cdev,struct resource * sram,unsigned long mask,unsigned long new)154*4882a593Smuzhiyun static void catpt_dsp_set_srampge(struct catpt_dev *cdev, struct resource *sram,
155*4882a593Smuzhiyun unsigned long mask, unsigned long new)
156*4882a593Smuzhiyun {
157*4882a593Smuzhiyun unsigned long old;
158*4882a593Smuzhiyun u32 off = sram->start;
159*4882a593Smuzhiyun u32 b = __ffs(mask);
160*4882a593Smuzhiyun
161*4882a593Smuzhiyun old = catpt_readl_pci(cdev, VDRTCTL0) & mask;
162*4882a593Smuzhiyun dev_dbg(cdev->dev, "SRAMPGE [0x%08lx] 0x%08lx -> 0x%08lx",
163*4882a593Smuzhiyun mask, old, new);
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun if (old == new)
166*4882a593Smuzhiyun return;
167*4882a593Smuzhiyun
168*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL0, mask, new);
169*4882a593Smuzhiyun /* wait for SRAM power gating to propagate */
170*4882a593Smuzhiyun udelay(60);
171*4882a593Smuzhiyun
172*4882a593Smuzhiyun /*
173*4882a593Smuzhiyun * Dummy read as the very first access after block enable
174*4882a593Smuzhiyun * to prevent byte loss in future operations.
175*4882a593Smuzhiyun */
176*4882a593Smuzhiyun for_each_clear_bit_from(b, &new, fls_long(mask)) {
177*4882a593Smuzhiyun u8 buf[4];
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun /* newly enabled: new bit=0 while old bit=1 */
180*4882a593Smuzhiyun if (test_bit(b, &old)) {
181*4882a593Smuzhiyun dev_dbg(cdev->dev, "sanitize block %ld: off 0x%08x\n",
182*4882a593Smuzhiyun b - __ffs(mask), off);
183*4882a593Smuzhiyun memcpy_fromio(buf, cdev->lpe_ba + off, sizeof(buf));
184*4882a593Smuzhiyun }
185*4882a593Smuzhiyun off += CATPT_MEMBLOCK_SIZE;
186*4882a593Smuzhiyun }
187*4882a593Smuzhiyun }
188*4882a593Smuzhiyun
catpt_dsp_update_srampge(struct catpt_dev * cdev,struct resource * sram,unsigned long mask)189*4882a593Smuzhiyun void catpt_dsp_update_srampge(struct catpt_dev *cdev, struct resource *sram,
190*4882a593Smuzhiyun unsigned long mask)
191*4882a593Smuzhiyun {
192*4882a593Smuzhiyun struct resource *res;
193*4882a593Smuzhiyun unsigned long new = 0;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun /* flag all busy blocks */
196*4882a593Smuzhiyun for (res = sram->child; res; res = res->sibling) {
197*4882a593Smuzhiyun u32 h, l;
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun h = (res->end - sram->start) / CATPT_MEMBLOCK_SIZE;
200*4882a593Smuzhiyun l = (res->start - sram->start) / CATPT_MEMBLOCK_SIZE;
201*4882a593Smuzhiyun new |= GENMASK(h, l);
202*4882a593Smuzhiyun }
203*4882a593Smuzhiyun
204*4882a593Smuzhiyun /* offset value given mask's start and invert it as ON=b0 */
205*4882a593Smuzhiyun new = ~(new << __ffs(mask)) & mask;
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun /* disable core clock gating */
208*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DCLCGE, 0);
209*4882a593Smuzhiyun
210*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, sram, mask, new);
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun /* enable core clock gating */
213*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DCLCGE,
214*4882a593Smuzhiyun CATPT_VDRTCTL2_DCLCGE);
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun
catpt_dsp_stall(struct catpt_dev * cdev,bool stall)217*4882a593Smuzhiyun int catpt_dsp_stall(struct catpt_dev *cdev, bool stall)
218*4882a593Smuzhiyun {
219*4882a593Smuzhiyun u32 reg, val;
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun val = stall ? CATPT_CS_STALL : 0;
222*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1, CATPT_CS_STALL, val);
223*4882a593Smuzhiyun
224*4882a593Smuzhiyun return catpt_readl_poll_shim(cdev, CS1,
225*4882a593Smuzhiyun reg, (reg & CATPT_CS_STALL) == val,
226*4882a593Smuzhiyun 500, 10000);
227*4882a593Smuzhiyun }
228*4882a593Smuzhiyun
catpt_dsp_reset(struct catpt_dev * cdev,bool reset)229*4882a593Smuzhiyun static int catpt_dsp_reset(struct catpt_dev *cdev, bool reset)
230*4882a593Smuzhiyun {
231*4882a593Smuzhiyun u32 reg, val;
232*4882a593Smuzhiyun
233*4882a593Smuzhiyun val = reset ? CATPT_CS_RST : 0;
234*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1, CATPT_CS_RST, val);
235*4882a593Smuzhiyun
236*4882a593Smuzhiyun return catpt_readl_poll_shim(cdev, CS1,
237*4882a593Smuzhiyun reg, (reg & CATPT_CS_RST) == val,
238*4882a593Smuzhiyun 500, 10000);
239*4882a593Smuzhiyun }
240*4882a593Smuzhiyun
lpt_dsp_pll_shutdown(struct catpt_dev * cdev,bool enable)241*4882a593Smuzhiyun void lpt_dsp_pll_shutdown(struct catpt_dev *cdev, bool enable)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun u32 val;
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun val = enable ? LPT_VDRTCTL0_APLLSE : 0;
246*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL0, LPT_VDRTCTL0_APLLSE, val);
247*4882a593Smuzhiyun }
248*4882a593Smuzhiyun
wpt_dsp_pll_shutdown(struct catpt_dev * cdev,bool enable)249*4882a593Smuzhiyun void wpt_dsp_pll_shutdown(struct catpt_dev *cdev, bool enable)
250*4882a593Smuzhiyun {
251*4882a593Smuzhiyun u32 val;
252*4882a593Smuzhiyun
253*4882a593Smuzhiyun val = enable ? WPT_VDRTCTL2_APLLSE : 0;
254*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, WPT_VDRTCTL2_APLLSE, val);
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun
catpt_dsp_select_lpclock(struct catpt_dev * cdev,bool lp,bool waiti)257*4882a593Smuzhiyun static int catpt_dsp_select_lpclock(struct catpt_dev *cdev, bool lp, bool waiti)
258*4882a593Smuzhiyun {
259*4882a593Smuzhiyun u32 mask, reg, val;
260*4882a593Smuzhiyun int ret;
261*4882a593Smuzhiyun
262*4882a593Smuzhiyun mutex_lock(&cdev->clk_mutex);
263*4882a593Smuzhiyun
264*4882a593Smuzhiyun val = lp ? CATPT_CS_LPCS : 0;
265*4882a593Smuzhiyun reg = catpt_readl_shim(cdev, CS1) & CATPT_CS_LPCS;
266*4882a593Smuzhiyun dev_dbg(cdev->dev, "LPCS [0x%08lx] 0x%08x -> 0x%08x",
267*4882a593Smuzhiyun CATPT_CS_LPCS, reg, val);
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun if (reg == val) {
270*4882a593Smuzhiyun mutex_unlock(&cdev->clk_mutex);
271*4882a593Smuzhiyun return 0;
272*4882a593Smuzhiyun }
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun if (waiti) {
275*4882a593Smuzhiyun /* wait for DSP to signal WAIT state */
276*4882a593Smuzhiyun ret = catpt_readl_poll_shim(cdev, ISD,
277*4882a593Smuzhiyun reg, (reg & CATPT_ISD_DCPWM),
278*4882a593Smuzhiyun 500, 10000);
279*4882a593Smuzhiyun if (ret) {
280*4882a593Smuzhiyun dev_warn(cdev->dev, "await WAITI timeout\n");
281*4882a593Smuzhiyun /* no signal - only high clock selection allowed */
282*4882a593Smuzhiyun if (lp) {
283*4882a593Smuzhiyun mutex_unlock(&cdev->clk_mutex);
284*4882a593Smuzhiyun return 0;
285*4882a593Smuzhiyun }
286*4882a593Smuzhiyun }
287*4882a593Smuzhiyun }
288*4882a593Smuzhiyun
289*4882a593Smuzhiyun ret = catpt_readl_poll_shim(cdev, CLKCTL,
290*4882a593Smuzhiyun reg, !(reg & CATPT_CLKCTL_CFCIP),
291*4882a593Smuzhiyun 500, 10000);
292*4882a593Smuzhiyun if (ret)
293*4882a593Smuzhiyun dev_warn(cdev->dev, "clock change still in progress\n");
294*4882a593Smuzhiyun
295*4882a593Smuzhiyun /* default to DSP core & audio fabric high clock */
296*4882a593Smuzhiyun val |= CATPT_CS_DCS_HIGH;
297*4882a593Smuzhiyun mask = CATPT_CS_LPCS | CATPT_CS_DCS;
298*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1, mask, val);
299*4882a593Smuzhiyun
300*4882a593Smuzhiyun ret = catpt_readl_poll_shim(cdev, CLKCTL,
301*4882a593Smuzhiyun reg, !(reg & CATPT_CLKCTL_CFCIP),
302*4882a593Smuzhiyun 500, 10000);
303*4882a593Smuzhiyun if (ret)
304*4882a593Smuzhiyun dev_warn(cdev->dev, "clock change still in progress\n");
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun /* update PLL accordingly */
307*4882a593Smuzhiyun cdev->spec->pll_shutdown(cdev, lp);
308*4882a593Smuzhiyun
309*4882a593Smuzhiyun mutex_unlock(&cdev->clk_mutex);
310*4882a593Smuzhiyun return 0;
311*4882a593Smuzhiyun }
312*4882a593Smuzhiyun
catpt_dsp_update_lpclock(struct catpt_dev * cdev)313*4882a593Smuzhiyun int catpt_dsp_update_lpclock(struct catpt_dev *cdev)
314*4882a593Smuzhiyun {
315*4882a593Smuzhiyun struct catpt_stream_runtime *stream;
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun list_for_each_entry(stream, &cdev->stream_list, node)
318*4882a593Smuzhiyun if (stream->prepared)
319*4882a593Smuzhiyun return catpt_dsp_select_lpclock(cdev, false, true);
320*4882a593Smuzhiyun
321*4882a593Smuzhiyun return catpt_dsp_select_lpclock(cdev, true, true);
322*4882a593Smuzhiyun }
323*4882a593Smuzhiyun
324*4882a593Smuzhiyun /* bring registers to their defaults as HW won't reset itself */
catpt_dsp_set_regs_defaults(struct catpt_dev * cdev)325*4882a593Smuzhiyun static void catpt_dsp_set_regs_defaults(struct catpt_dev *cdev)
326*4882a593Smuzhiyun {
327*4882a593Smuzhiyun int i;
328*4882a593Smuzhiyun
329*4882a593Smuzhiyun catpt_writel_shim(cdev, CS1, CATPT_CS_DEFAULT);
330*4882a593Smuzhiyun catpt_writel_shim(cdev, ISC, CATPT_ISC_DEFAULT);
331*4882a593Smuzhiyun catpt_writel_shim(cdev, ISD, CATPT_ISD_DEFAULT);
332*4882a593Smuzhiyun catpt_writel_shim(cdev, IMC, CATPT_IMC_DEFAULT);
333*4882a593Smuzhiyun catpt_writel_shim(cdev, IMD, CATPT_IMD_DEFAULT);
334*4882a593Smuzhiyun catpt_writel_shim(cdev, IPCC, CATPT_IPCC_DEFAULT);
335*4882a593Smuzhiyun catpt_writel_shim(cdev, IPCD, CATPT_IPCD_DEFAULT);
336*4882a593Smuzhiyun catpt_writel_shim(cdev, CLKCTL, CATPT_CLKCTL_DEFAULT);
337*4882a593Smuzhiyun catpt_writel_shim(cdev, CS2, CATPT_CS2_DEFAULT);
338*4882a593Smuzhiyun catpt_writel_shim(cdev, LTRC, CATPT_LTRC_DEFAULT);
339*4882a593Smuzhiyun catpt_writel_shim(cdev, HMDC, CATPT_HMDC_DEFAULT);
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun for (i = 0; i < CATPT_SSP_COUNT; i++) {
342*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSCR0, CATPT_SSC0_DEFAULT);
343*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSCR1, CATPT_SSC1_DEFAULT);
344*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSSR, CATPT_SSS_DEFAULT);
345*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSITR, CATPT_SSIT_DEFAULT);
346*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSDR, CATPT_SSD_DEFAULT);
347*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSTO, CATPT_SSTO_DEFAULT);
348*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSPSP, CATPT_SSPSP_DEFAULT);
349*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSTSA, CATPT_SSTSA_DEFAULT);
350*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSRSA, CATPT_SSRSA_DEFAULT);
351*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSTSS, CATPT_SSTSS_DEFAULT);
352*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSCR2, CATPT_SSCR2_DEFAULT);
353*4882a593Smuzhiyun catpt_writel_ssp(cdev, i, SSPSP2, CATPT_SSPSP2_DEFAULT);
354*4882a593Smuzhiyun }
355*4882a593Smuzhiyun }
356*4882a593Smuzhiyun
lpt_dsp_power_down(struct catpt_dev * cdev)357*4882a593Smuzhiyun int lpt_dsp_power_down(struct catpt_dev *cdev)
358*4882a593Smuzhiyun {
359*4882a593Smuzhiyun catpt_dsp_reset(cdev, true);
360*4882a593Smuzhiyun
361*4882a593Smuzhiyun /* set 24Mhz clock for both SSPs */
362*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1, CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1),
363*4882a593Smuzhiyun CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1));
364*4882a593Smuzhiyun catpt_dsp_select_lpclock(cdev, true, false);
365*4882a593Smuzhiyun
366*4882a593Smuzhiyun /* DRAM power gating all */
367*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->dram, cdev->spec->dram_mask,
368*4882a593Smuzhiyun cdev->spec->dram_mask);
369*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->iram, cdev->spec->iram_mask,
370*4882a593Smuzhiyun cdev->spec->iram_mask);
371*4882a593Smuzhiyun
372*4882a593Smuzhiyun catpt_updatel_pci(cdev, PMCS, PCI_PM_CTRL_STATE_MASK, PCI_D3hot);
373*4882a593Smuzhiyun /* give hw time to drop off */
374*4882a593Smuzhiyun udelay(50);
375*4882a593Smuzhiyun
376*4882a593Smuzhiyun return 0;
377*4882a593Smuzhiyun }
378*4882a593Smuzhiyun
lpt_dsp_power_up(struct catpt_dev * cdev)379*4882a593Smuzhiyun int lpt_dsp_power_up(struct catpt_dev *cdev)
380*4882a593Smuzhiyun {
381*4882a593Smuzhiyun /* SRAM power gating none */
382*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->dram, cdev->spec->dram_mask, 0);
383*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->iram, cdev->spec->iram_mask, 0);
384*4882a593Smuzhiyun
385*4882a593Smuzhiyun catpt_updatel_pci(cdev, PMCS, PCI_PM_CTRL_STATE_MASK, PCI_D0);
386*4882a593Smuzhiyun /* give hw time to wake up */
387*4882a593Smuzhiyun udelay(100);
388*4882a593Smuzhiyun
389*4882a593Smuzhiyun catpt_dsp_select_lpclock(cdev, false, false);
390*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1,
391*4882a593Smuzhiyun CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1),
392*4882a593Smuzhiyun CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1));
393*4882a593Smuzhiyun /* stagger DSP reset after clock selection */
394*4882a593Smuzhiyun udelay(50);
395*4882a593Smuzhiyun
396*4882a593Smuzhiyun catpt_dsp_reset(cdev, false);
397*4882a593Smuzhiyun /* generate int deassert msg to fix inversed int logic */
398*4882a593Smuzhiyun catpt_updatel_shim(cdev, IMC, CATPT_IMC_IPCDB | CATPT_IMC_IPCCD, 0);
399*4882a593Smuzhiyun
400*4882a593Smuzhiyun return 0;
401*4882a593Smuzhiyun }
402*4882a593Smuzhiyun
wpt_dsp_power_down(struct catpt_dev * cdev)403*4882a593Smuzhiyun int wpt_dsp_power_down(struct catpt_dev *cdev)
404*4882a593Smuzhiyun {
405*4882a593Smuzhiyun u32 mask, val;
406*4882a593Smuzhiyun
407*4882a593Smuzhiyun /* disable core clock gating */
408*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DCLCGE, 0);
409*4882a593Smuzhiyun
410*4882a593Smuzhiyun catpt_dsp_reset(cdev, true);
411*4882a593Smuzhiyun /* set 24Mhz clock for both SSPs */
412*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1, CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1),
413*4882a593Smuzhiyun CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1));
414*4882a593Smuzhiyun catpt_dsp_select_lpclock(cdev, true, false);
415*4882a593Smuzhiyun /* disable MCLK */
416*4882a593Smuzhiyun catpt_updatel_shim(cdev, CLKCTL, CATPT_CLKCTL_SMOS, 0);
417*4882a593Smuzhiyun
418*4882a593Smuzhiyun catpt_dsp_set_regs_defaults(cdev);
419*4882a593Smuzhiyun
420*4882a593Smuzhiyun /* switch clock gating */
421*4882a593Smuzhiyun mask = CATPT_VDRTCTL2_CGEALL & (~CATPT_VDRTCTL2_DCLCGE);
422*4882a593Smuzhiyun val = mask & (~CATPT_VDRTCTL2_DTCGE);
423*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, mask, val);
424*4882a593Smuzhiyun /* enable DTCGE separatelly */
425*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DTCGE,
426*4882a593Smuzhiyun CATPT_VDRTCTL2_DTCGE);
427*4882a593Smuzhiyun
428*4882a593Smuzhiyun /* SRAM power gating all */
429*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->dram, cdev->spec->dram_mask,
430*4882a593Smuzhiyun cdev->spec->dram_mask);
431*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->iram, cdev->spec->iram_mask,
432*4882a593Smuzhiyun cdev->spec->iram_mask);
433*4882a593Smuzhiyun mask = WPT_VDRTCTL0_D3SRAMPGD | WPT_VDRTCTL0_D3PGD;
434*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL0, mask, WPT_VDRTCTL0_D3PGD);
435*4882a593Smuzhiyun
436*4882a593Smuzhiyun catpt_updatel_pci(cdev, PMCS, PCI_PM_CTRL_STATE_MASK, PCI_D3hot);
437*4882a593Smuzhiyun /* give hw time to drop off */
438*4882a593Smuzhiyun udelay(50);
439*4882a593Smuzhiyun
440*4882a593Smuzhiyun /* enable core clock gating */
441*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DCLCGE,
442*4882a593Smuzhiyun CATPT_VDRTCTL2_DCLCGE);
443*4882a593Smuzhiyun udelay(50);
444*4882a593Smuzhiyun
445*4882a593Smuzhiyun return 0;
446*4882a593Smuzhiyun }
447*4882a593Smuzhiyun
wpt_dsp_power_up(struct catpt_dev * cdev)448*4882a593Smuzhiyun int wpt_dsp_power_up(struct catpt_dev *cdev)
449*4882a593Smuzhiyun {
450*4882a593Smuzhiyun u32 mask, val;
451*4882a593Smuzhiyun
452*4882a593Smuzhiyun /* disable core clock gating */
453*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DCLCGE, 0);
454*4882a593Smuzhiyun
455*4882a593Smuzhiyun /* switch clock gating */
456*4882a593Smuzhiyun mask = CATPT_VDRTCTL2_CGEALL & (~CATPT_VDRTCTL2_DCLCGE);
457*4882a593Smuzhiyun val = mask & (~CATPT_VDRTCTL2_DTCGE);
458*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, mask, val);
459*4882a593Smuzhiyun
460*4882a593Smuzhiyun catpt_updatel_pci(cdev, PMCS, PCI_PM_CTRL_STATE_MASK, PCI_D0);
461*4882a593Smuzhiyun
462*4882a593Smuzhiyun /* SRAM power gating none */
463*4882a593Smuzhiyun mask = WPT_VDRTCTL0_D3SRAMPGD | WPT_VDRTCTL0_D3PGD;
464*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL0, mask, mask);
465*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->dram, cdev->spec->dram_mask, 0);
466*4882a593Smuzhiyun catpt_dsp_set_srampge(cdev, &cdev->iram, cdev->spec->iram_mask, 0);
467*4882a593Smuzhiyun
468*4882a593Smuzhiyun catpt_dsp_set_regs_defaults(cdev);
469*4882a593Smuzhiyun
470*4882a593Smuzhiyun /* restore MCLK */
471*4882a593Smuzhiyun catpt_updatel_shim(cdev, CLKCTL, CATPT_CLKCTL_SMOS, CATPT_CLKCTL_SMOS);
472*4882a593Smuzhiyun catpt_dsp_select_lpclock(cdev, false, false);
473*4882a593Smuzhiyun /* set 24Mhz clock for both SSPs */
474*4882a593Smuzhiyun catpt_updatel_shim(cdev, CS1, CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1),
475*4882a593Smuzhiyun CATPT_CS_SBCS(0) | CATPT_CS_SBCS(1));
476*4882a593Smuzhiyun catpt_dsp_reset(cdev, false);
477*4882a593Smuzhiyun
478*4882a593Smuzhiyun /* enable core clock gating */
479*4882a593Smuzhiyun catpt_updatel_pci(cdev, VDRTCTL2, CATPT_VDRTCTL2_DCLCGE,
480*4882a593Smuzhiyun CATPT_VDRTCTL2_DCLCGE);
481*4882a593Smuzhiyun
482*4882a593Smuzhiyun /* generate int deassert msg to fix inversed int logic */
483*4882a593Smuzhiyun catpt_updatel_shim(cdev, IMC, CATPT_IMC_IPCDB | CATPT_IMC_IPCCD, 0);
484*4882a593Smuzhiyun
485*4882a593Smuzhiyun return 0;
486*4882a593Smuzhiyun }
487*4882a593Smuzhiyun
488*4882a593Smuzhiyun #define CATPT_DUMP_MAGIC 0xcd42
489*4882a593Smuzhiyun #define CATPT_DUMP_SECTION_ID_FILE 0x00
490*4882a593Smuzhiyun #define CATPT_DUMP_SECTION_ID_IRAM 0x01
491*4882a593Smuzhiyun #define CATPT_DUMP_SECTION_ID_DRAM 0x02
492*4882a593Smuzhiyun #define CATPT_DUMP_SECTION_ID_REGS 0x03
493*4882a593Smuzhiyun #define CATPT_DUMP_HASH_SIZE 20
494*4882a593Smuzhiyun
495*4882a593Smuzhiyun struct catpt_dump_section_hdr {
496*4882a593Smuzhiyun u16 magic;
497*4882a593Smuzhiyun u8 core_id;
498*4882a593Smuzhiyun u8 section_id;
499*4882a593Smuzhiyun u32 size;
500*4882a593Smuzhiyun };
501*4882a593Smuzhiyun
catpt_coredump(struct catpt_dev * cdev)502*4882a593Smuzhiyun int catpt_coredump(struct catpt_dev *cdev)
503*4882a593Smuzhiyun {
504*4882a593Smuzhiyun struct catpt_dump_section_hdr *hdr;
505*4882a593Smuzhiyun size_t dump_size, regs_size;
506*4882a593Smuzhiyun u8 *dump, *pos;
507*4882a593Smuzhiyun const char *eof;
508*4882a593Smuzhiyun char *info;
509*4882a593Smuzhiyun int i;
510*4882a593Smuzhiyun
511*4882a593Smuzhiyun regs_size = CATPT_SHIM_REGS_SIZE;
512*4882a593Smuzhiyun regs_size += CATPT_DMA_COUNT * CATPT_DMA_REGS_SIZE;
513*4882a593Smuzhiyun regs_size += CATPT_SSP_COUNT * CATPT_SSP_REGS_SIZE;
514*4882a593Smuzhiyun dump_size = resource_size(&cdev->dram);
515*4882a593Smuzhiyun dump_size += resource_size(&cdev->iram);
516*4882a593Smuzhiyun dump_size += regs_size;
517*4882a593Smuzhiyun /* account for header of each section and hash chunk */
518*4882a593Smuzhiyun dump_size += 4 * sizeof(*hdr) + CATPT_DUMP_HASH_SIZE;
519*4882a593Smuzhiyun
520*4882a593Smuzhiyun dump = vzalloc(dump_size);
521*4882a593Smuzhiyun if (!dump)
522*4882a593Smuzhiyun return -ENOMEM;
523*4882a593Smuzhiyun
524*4882a593Smuzhiyun pos = dump;
525*4882a593Smuzhiyun
526*4882a593Smuzhiyun hdr = (struct catpt_dump_section_hdr *)pos;
527*4882a593Smuzhiyun hdr->magic = CATPT_DUMP_MAGIC;
528*4882a593Smuzhiyun hdr->core_id = cdev->spec->core_id;
529*4882a593Smuzhiyun hdr->section_id = CATPT_DUMP_SECTION_ID_FILE;
530*4882a593Smuzhiyun hdr->size = dump_size - sizeof(*hdr);
531*4882a593Smuzhiyun pos += sizeof(*hdr);
532*4882a593Smuzhiyun
533*4882a593Smuzhiyun info = cdev->ipc.config.fw_info;
534*4882a593Smuzhiyun eof = info + FW_INFO_SIZE_MAX;
535*4882a593Smuzhiyun /* navigate to fifth info segment (fw hash) */
536*4882a593Smuzhiyun for (i = 0; i < 4 && info < eof; i++, info++) {
537*4882a593Smuzhiyun /* info segments are separated by space each */
538*4882a593Smuzhiyun info = strnchr(info, eof - info, ' ');
539*4882a593Smuzhiyun if (!info)
540*4882a593Smuzhiyun break;
541*4882a593Smuzhiyun }
542*4882a593Smuzhiyun
543*4882a593Smuzhiyun if (i == 4 && info)
544*4882a593Smuzhiyun memcpy(pos, info, min_t(u32, eof - info, CATPT_DUMP_HASH_SIZE));
545*4882a593Smuzhiyun pos += CATPT_DUMP_HASH_SIZE;
546*4882a593Smuzhiyun
547*4882a593Smuzhiyun hdr = (struct catpt_dump_section_hdr *)pos;
548*4882a593Smuzhiyun hdr->magic = CATPT_DUMP_MAGIC;
549*4882a593Smuzhiyun hdr->core_id = cdev->spec->core_id;
550*4882a593Smuzhiyun hdr->section_id = CATPT_DUMP_SECTION_ID_IRAM;
551*4882a593Smuzhiyun hdr->size = resource_size(&cdev->iram);
552*4882a593Smuzhiyun pos += sizeof(*hdr);
553*4882a593Smuzhiyun
554*4882a593Smuzhiyun memcpy_fromio(pos, cdev->lpe_ba + cdev->iram.start, hdr->size);
555*4882a593Smuzhiyun pos += hdr->size;
556*4882a593Smuzhiyun
557*4882a593Smuzhiyun hdr = (struct catpt_dump_section_hdr *)pos;
558*4882a593Smuzhiyun hdr->magic = CATPT_DUMP_MAGIC;
559*4882a593Smuzhiyun hdr->core_id = cdev->spec->core_id;
560*4882a593Smuzhiyun hdr->section_id = CATPT_DUMP_SECTION_ID_DRAM;
561*4882a593Smuzhiyun hdr->size = resource_size(&cdev->dram);
562*4882a593Smuzhiyun pos += sizeof(*hdr);
563*4882a593Smuzhiyun
564*4882a593Smuzhiyun memcpy_fromio(pos, cdev->lpe_ba + cdev->dram.start, hdr->size);
565*4882a593Smuzhiyun pos += hdr->size;
566*4882a593Smuzhiyun
567*4882a593Smuzhiyun hdr = (struct catpt_dump_section_hdr *)pos;
568*4882a593Smuzhiyun hdr->magic = CATPT_DUMP_MAGIC;
569*4882a593Smuzhiyun hdr->core_id = cdev->spec->core_id;
570*4882a593Smuzhiyun hdr->section_id = CATPT_DUMP_SECTION_ID_REGS;
571*4882a593Smuzhiyun hdr->size = regs_size;
572*4882a593Smuzhiyun pos += sizeof(*hdr);
573*4882a593Smuzhiyun
574*4882a593Smuzhiyun memcpy_fromio(pos, catpt_shim_addr(cdev), CATPT_SHIM_REGS_SIZE);
575*4882a593Smuzhiyun pos += CATPT_SHIM_REGS_SIZE;
576*4882a593Smuzhiyun
577*4882a593Smuzhiyun for (i = 0; i < CATPT_SSP_COUNT; i++) {
578*4882a593Smuzhiyun memcpy_fromio(pos, catpt_ssp_addr(cdev, i),
579*4882a593Smuzhiyun CATPT_SSP_REGS_SIZE);
580*4882a593Smuzhiyun pos += CATPT_SSP_REGS_SIZE;
581*4882a593Smuzhiyun }
582*4882a593Smuzhiyun for (i = 0; i < CATPT_DMA_COUNT; i++) {
583*4882a593Smuzhiyun memcpy_fromio(pos, catpt_dma_addr(cdev, i),
584*4882a593Smuzhiyun CATPT_DMA_REGS_SIZE);
585*4882a593Smuzhiyun pos += CATPT_DMA_REGS_SIZE;
586*4882a593Smuzhiyun }
587*4882a593Smuzhiyun
588*4882a593Smuzhiyun dev_coredumpv(cdev->dev, dump, dump_size, GFP_KERNEL);
589*4882a593Smuzhiyun
590*4882a593Smuzhiyun return 0;
591*4882a593Smuzhiyun }
592