xref: /OK3568_Linux_fs/kernel/drivers/dma/mv_xor.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * offload engine driver for the Marvell XOR engine
4*4882a593Smuzhiyun  * Copyright (C) 2007, 2008, Marvell International Ltd.
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/init.h>
8*4882a593Smuzhiyun #include <linux/slab.h>
9*4882a593Smuzhiyun #include <linux/delay.h>
10*4882a593Smuzhiyun #include <linux/dma-mapping.h>
11*4882a593Smuzhiyun #include <linux/spinlock.h>
12*4882a593Smuzhiyun #include <linux/interrupt.h>
13*4882a593Smuzhiyun #include <linux/of_device.h>
14*4882a593Smuzhiyun #include <linux/platform_device.h>
15*4882a593Smuzhiyun #include <linux/memory.h>
16*4882a593Smuzhiyun #include <linux/clk.h>
17*4882a593Smuzhiyun #include <linux/of.h>
18*4882a593Smuzhiyun #include <linux/of_irq.h>
19*4882a593Smuzhiyun #include <linux/irqdomain.h>
20*4882a593Smuzhiyun #include <linux/cpumask.h>
21*4882a593Smuzhiyun #include <linux/platform_data/dma-mv_xor.h>
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun #include "dmaengine.h"
24*4882a593Smuzhiyun #include "mv_xor.h"
25*4882a593Smuzhiyun 
26*4882a593Smuzhiyun enum mv_xor_type {
27*4882a593Smuzhiyun 	XOR_ORION,
28*4882a593Smuzhiyun 	XOR_ARMADA_38X,
29*4882a593Smuzhiyun 	XOR_ARMADA_37XX,
30*4882a593Smuzhiyun };
31*4882a593Smuzhiyun 
32*4882a593Smuzhiyun enum mv_xor_mode {
33*4882a593Smuzhiyun 	XOR_MODE_IN_REG,
34*4882a593Smuzhiyun 	XOR_MODE_IN_DESC,
35*4882a593Smuzhiyun };
36*4882a593Smuzhiyun 
37*4882a593Smuzhiyun static void mv_xor_issue_pending(struct dma_chan *chan);
38*4882a593Smuzhiyun 
39*4882a593Smuzhiyun #define to_mv_xor_chan(chan)		\
40*4882a593Smuzhiyun 	container_of(chan, struct mv_xor_chan, dmachan)
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun #define to_mv_xor_slot(tx)		\
43*4882a593Smuzhiyun 	container_of(tx, struct mv_xor_desc_slot, async_tx)
44*4882a593Smuzhiyun 
45*4882a593Smuzhiyun #define mv_chan_to_devp(chan)           \
46*4882a593Smuzhiyun 	((chan)->dmadev.dev)
47*4882a593Smuzhiyun 
mv_desc_init(struct mv_xor_desc_slot * desc,dma_addr_t addr,u32 byte_count,enum dma_ctrl_flags flags)48*4882a593Smuzhiyun static void mv_desc_init(struct mv_xor_desc_slot *desc,
49*4882a593Smuzhiyun 			 dma_addr_t addr, u32 byte_count,
50*4882a593Smuzhiyun 			 enum dma_ctrl_flags flags)
51*4882a593Smuzhiyun {
52*4882a593Smuzhiyun 	struct mv_xor_desc *hw_desc = desc->hw_desc;
53*4882a593Smuzhiyun 
54*4882a593Smuzhiyun 	hw_desc->status = XOR_DESC_DMA_OWNED;
55*4882a593Smuzhiyun 	hw_desc->phy_next_desc = 0;
56*4882a593Smuzhiyun 	/* Enable end-of-descriptor interrupts only for DMA_PREP_INTERRUPT */
57*4882a593Smuzhiyun 	hw_desc->desc_command = (flags & DMA_PREP_INTERRUPT) ?
58*4882a593Smuzhiyun 				XOR_DESC_EOD_INT_EN : 0;
59*4882a593Smuzhiyun 	hw_desc->phy_dest_addr = addr;
60*4882a593Smuzhiyun 	hw_desc->byte_count = byte_count;
61*4882a593Smuzhiyun }
62*4882a593Smuzhiyun 
mv_desc_set_mode(struct mv_xor_desc_slot * desc)63*4882a593Smuzhiyun static void mv_desc_set_mode(struct mv_xor_desc_slot *desc)
64*4882a593Smuzhiyun {
65*4882a593Smuzhiyun 	struct mv_xor_desc *hw_desc = desc->hw_desc;
66*4882a593Smuzhiyun 
67*4882a593Smuzhiyun 	switch (desc->type) {
68*4882a593Smuzhiyun 	case DMA_XOR:
69*4882a593Smuzhiyun 	case DMA_INTERRUPT:
70*4882a593Smuzhiyun 		hw_desc->desc_command |= XOR_DESC_OPERATION_XOR;
71*4882a593Smuzhiyun 		break;
72*4882a593Smuzhiyun 	case DMA_MEMCPY:
73*4882a593Smuzhiyun 		hw_desc->desc_command |= XOR_DESC_OPERATION_MEMCPY;
74*4882a593Smuzhiyun 		break;
75*4882a593Smuzhiyun 	default:
76*4882a593Smuzhiyun 		BUG();
77*4882a593Smuzhiyun 		return;
78*4882a593Smuzhiyun 	}
79*4882a593Smuzhiyun }
80*4882a593Smuzhiyun 
mv_desc_set_next_desc(struct mv_xor_desc_slot * desc,u32 next_desc_addr)81*4882a593Smuzhiyun static void mv_desc_set_next_desc(struct mv_xor_desc_slot *desc,
82*4882a593Smuzhiyun 				  u32 next_desc_addr)
83*4882a593Smuzhiyun {
84*4882a593Smuzhiyun 	struct mv_xor_desc *hw_desc = desc->hw_desc;
85*4882a593Smuzhiyun 	BUG_ON(hw_desc->phy_next_desc);
86*4882a593Smuzhiyun 	hw_desc->phy_next_desc = next_desc_addr;
87*4882a593Smuzhiyun }
88*4882a593Smuzhiyun 
mv_desc_set_src_addr(struct mv_xor_desc_slot * desc,int index,dma_addr_t addr)89*4882a593Smuzhiyun static void mv_desc_set_src_addr(struct mv_xor_desc_slot *desc,
90*4882a593Smuzhiyun 				 int index, dma_addr_t addr)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun 	struct mv_xor_desc *hw_desc = desc->hw_desc;
93*4882a593Smuzhiyun 	hw_desc->phy_src_addr[mv_phy_src_idx(index)] = addr;
94*4882a593Smuzhiyun 	if (desc->type == DMA_XOR)
95*4882a593Smuzhiyun 		hw_desc->desc_command |= (1 << index);
96*4882a593Smuzhiyun }
97*4882a593Smuzhiyun 
mv_chan_get_current_desc(struct mv_xor_chan * chan)98*4882a593Smuzhiyun static u32 mv_chan_get_current_desc(struct mv_xor_chan *chan)
99*4882a593Smuzhiyun {
100*4882a593Smuzhiyun 	return readl_relaxed(XOR_CURR_DESC(chan));
101*4882a593Smuzhiyun }
102*4882a593Smuzhiyun 
mv_chan_set_next_descriptor(struct mv_xor_chan * chan,u32 next_desc_addr)103*4882a593Smuzhiyun static void mv_chan_set_next_descriptor(struct mv_xor_chan *chan,
104*4882a593Smuzhiyun 					u32 next_desc_addr)
105*4882a593Smuzhiyun {
106*4882a593Smuzhiyun 	writel_relaxed(next_desc_addr, XOR_NEXT_DESC(chan));
107*4882a593Smuzhiyun }
108*4882a593Smuzhiyun 
mv_chan_unmask_interrupts(struct mv_xor_chan * chan)109*4882a593Smuzhiyun static void mv_chan_unmask_interrupts(struct mv_xor_chan *chan)
110*4882a593Smuzhiyun {
111*4882a593Smuzhiyun 	u32 val = readl_relaxed(XOR_INTR_MASK(chan));
112*4882a593Smuzhiyun 	val |= XOR_INTR_MASK_VALUE << (chan->idx * 16);
113*4882a593Smuzhiyun 	writel_relaxed(val, XOR_INTR_MASK(chan));
114*4882a593Smuzhiyun }
115*4882a593Smuzhiyun 
mv_chan_get_intr_cause(struct mv_xor_chan * chan)116*4882a593Smuzhiyun static u32 mv_chan_get_intr_cause(struct mv_xor_chan *chan)
117*4882a593Smuzhiyun {
118*4882a593Smuzhiyun 	u32 intr_cause = readl_relaxed(XOR_INTR_CAUSE(chan));
119*4882a593Smuzhiyun 	intr_cause = (intr_cause >> (chan->idx * 16)) & 0xFFFF;
120*4882a593Smuzhiyun 	return intr_cause;
121*4882a593Smuzhiyun }
122*4882a593Smuzhiyun 
mv_chan_clear_eoc_cause(struct mv_xor_chan * chan)123*4882a593Smuzhiyun static void mv_chan_clear_eoc_cause(struct mv_xor_chan *chan)
124*4882a593Smuzhiyun {
125*4882a593Smuzhiyun 	u32 val;
126*4882a593Smuzhiyun 
127*4882a593Smuzhiyun 	val = XOR_INT_END_OF_DESC | XOR_INT_END_OF_CHAIN | XOR_INT_STOPPED;
128*4882a593Smuzhiyun 	val = ~(val << (chan->idx * 16));
129*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(chan), "%s, val 0x%08x\n", __func__, val);
130*4882a593Smuzhiyun 	writel_relaxed(val, XOR_INTR_CAUSE(chan));
131*4882a593Smuzhiyun }
132*4882a593Smuzhiyun 
mv_chan_clear_err_status(struct mv_xor_chan * chan)133*4882a593Smuzhiyun static void mv_chan_clear_err_status(struct mv_xor_chan *chan)
134*4882a593Smuzhiyun {
135*4882a593Smuzhiyun 	u32 val = 0xFFFF0000 >> (chan->idx * 16);
136*4882a593Smuzhiyun 	writel_relaxed(val, XOR_INTR_CAUSE(chan));
137*4882a593Smuzhiyun }
138*4882a593Smuzhiyun 
mv_chan_set_mode(struct mv_xor_chan * chan,u32 op_mode)139*4882a593Smuzhiyun static void mv_chan_set_mode(struct mv_xor_chan *chan,
140*4882a593Smuzhiyun 			     u32 op_mode)
141*4882a593Smuzhiyun {
142*4882a593Smuzhiyun 	u32 config = readl_relaxed(XOR_CONFIG(chan));
143*4882a593Smuzhiyun 
144*4882a593Smuzhiyun 	config &= ~0x7;
145*4882a593Smuzhiyun 	config |= op_mode;
146*4882a593Smuzhiyun 
147*4882a593Smuzhiyun #if defined(__BIG_ENDIAN)
148*4882a593Smuzhiyun 	config |= XOR_DESCRIPTOR_SWAP;
149*4882a593Smuzhiyun #else
150*4882a593Smuzhiyun 	config &= ~XOR_DESCRIPTOR_SWAP;
151*4882a593Smuzhiyun #endif
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun 	writel_relaxed(config, XOR_CONFIG(chan));
154*4882a593Smuzhiyun }
155*4882a593Smuzhiyun 
mv_chan_activate(struct mv_xor_chan * chan)156*4882a593Smuzhiyun static void mv_chan_activate(struct mv_xor_chan *chan)
157*4882a593Smuzhiyun {
158*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(chan), " activate chan.\n");
159*4882a593Smuzhiyun 
160*4882a593Smuzhiyun 	/* writel ensures all descriptors are flushed before activation */
161*4882a593Smuzhiyun 	writel(BIT(0), XOR_ACTIVATION(chan));
162*4882a593Smuzhiyun }
163*4882a593Smuzhiyun 
mv_chan_is_busy(struct mv_xor_chan * chan)164*4882a593Smuzhiyun static char mv_chan_is_busy(struct mv_xor_chan *chan)
165*4882a593Smuzhiyun {
166*4882a593Smuzhiyun 	u32 state = readl_relaxed(XOR_ACTIVATION(chan));
167*4882a593Smuzhiyun 
168*4882a593Smuzhiyun 	state = (state >> 4) & 0x3;
169*4882a593Smuzhiyun 
170*4882a593Smuzhiyun 	return (state == 1) ? 1 : 0;
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun /*
174*4882a593Smuzhiyun  * mv_chan_start_new_chain - program the engine to operate on new
175*4882a593Smuzhiyun  * chain headed by sw_desc
176*4882a593Smuzhiyun  * Caller must hold &mv_chan->lock while calling this function
177*4882a593Smuzhiyun  */
mv_chan_start_new_chain(struct mv_xor_chan * mv_chan,struct mv_xor_desc_slot * sw_desc)178*4882a593Smuzhiyun static void mv_chan_start_new_chain(struct mv_xor_chan *mv_chan,
179*4882a593Smuzhiyun 				    struct mv_xor_desc_slot *sw_desc)
180*4882a593Smuzhiyun {
181*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan), "%s %d: sw_desc %p\n",
182*4882a593Smuzhiyun 		__func__, __LINE__, sw_desc);
183*4882a593Smuzhiyun 
184*4882a593Smuzhiyun 	/* set the hardware chain */
185*4882a593Smuzhiyun 	mv_chan_set_next_descriptor(mv_chan, sw_desc->async_tx.phys);
186*4882a593Smuzhiyun 
187*4882a593Smuzhiyun 	mv_chan->pending++;
188*4882a593Smuzhiyun 	mv_xor_issue_pending(&mv_chan->dmachan);
189*4882a593Smuzhiyun }
190*4882a593Smuzhiyun 
191*4882a593Smuzhiyun static dma_cookie_t
mv_desc_run_tx_complete_actions(struct mv_xor_desc_slot * desc,struct mv_xor_chan * mv_chan,dma_cookie_t cookie)192*4882a593Smuzhiyun mv_desc_run_tx_complete_actions(struct mv_xor_desc_slot *desc,
193*4882a593Smuzhiyun 				struct mv_xor_chan *mv_chan,
194*4882a593Smuzhiyun 				dma_cookie_t cookie)
195*4882a593Smuzhiyun {
196*4882a593Smuzhiyun 	BUG_ON(desc->async_tx.cookie < 0);
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun 	if (desc->async_tx.cookie > 0) {
199*4882a593Smuzhiyun 		cookie = desc->async_tx.cookie;
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 		dma_descriptor_unmap(&desc->async_tx);
202*4882a593Smuzhiyun 		/* call the callback (must not sleep or submit new
203*4882a593Smuzhiyun 		 * operations to this channel)
204*4882a593Smuzhiyun 		 */
205*4882a593Smuzhiyun 		dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL);
206*4882a593Smuzhiyun 	}
207*4882a593Smuzhiyun 
208*4882a593Smuzhiyun 	/* run dependent operations */
209*4882a593Smuzhiyun 	dma_run_dependencies(&desc->async_tx);
210*4882a593Smuzhiyun 
211*4882a593Smuzhiyun 	return cookie;
212*4882a593Smuzhiyun }
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun static int
mv_chan_clean_completed_slots(struct mv_xor_chan * mv_chan)215*4882a593Smuzhiyun mv_chan_clean_completed_slots(struct mv_xor_chan *mv_chan)
216*4882a593Smuzhiyun {
217*4882a593Smuzhiyun 	struct mv_xor_desc_slot *iter, *_iter;
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan), "%s %d\n", __func__, __LINE__);
220*4882a593Smuzhiyun 	list_for_each_entry_safe(iter, _iter, &mv_chan->completed_slots,
221*4882a593Smuzhiyun 				 node) {
222*4882a593Smuzhiyun 
223*4882a593Smuzhiyun 		if (async_tx_test_ack(&iter->async_tx)) {
224*4882a593Smuzhiyun 			list_move_tail(&iter->node, &mv_chan->free_slots);
225*4882a593Smuzhiyun 			if (!list_empty(&iter->sg_tx_list)) {
226*4882a593Smuzhiyun 				list_splice_tail_init(&iter->sg_tx_list,
227*4882a593Smuzhiyun 							&mv_chan->free_slots);
228*4882a593Smuzhiyun 			}
229*4882a593Smuzhiyun 		}
230*4882a593Smuzhiyun 	}
231*4882a593Smuzhiyun 	return 0;
232*4882a593Smuzhiyun }
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun static int
mv_desc_clean_slot(struct mv_xor_desc_slot * desc,struct mv_xor_chan * mv_chan)235*4882a593Smuzhiyun mv_desc_clean_slot(struct mv_xor_desc_slot *desc,
236*4882a593Smuzhiyun 		   struct mv_xor_chan *mv_chan)
237*4882a593Smuzhiyun {
238*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan), "%s %d: desc %p flags %d\n",
239*4882a593Smuzhiyun 		__func__, __LINE__, desc, desc->async_tx.flags);
240*4882a593Smuzhiyun 
241*4882a593Smuzhiyun 	/* the client is allowed to attach dependent operations
242*4882a593Smuzhiyun 	 * until 'ack' is set
243*4882a593Smuzhiyun 	 */
244*4882a593Smuzhiyun 	if (!async_tx_test_ack(&desc->async_tx)) {
245*4882a593Smuzhiyun 		/* move this slot to the completed_slots */
246*4882a593Smuzhiyun 		list_move_tail(&desc->node, &mv_chan->completed_slots);
247*4882a593Smuzhiyun 		if (!list_empty(&desc->sg_tx_list)) {
248*4882a593Smuzhiyun 			list_splice_tail_init(&desc->sg_tx_list,
249*4882a593Smuzhiyun 					      &mv_chan->completed_slots);
250*4882a593Smuzhiyun 		}
251*4882a593Smuzhiyun 	} else {
252*4882a593Smuzhiyun 		list_move_tail(&desc->node, &mv_chan->free_slots);
253*4882a593Smuzhiyun 		if (!list_empty(&desc->sg_tx_list)) {
254*4882a593Smuzhiyun 			list_splice_tail_init(&desc->sg_tx_list,
255*4882a593Smuzhiyun 					      &mv_chan->free_slots);
256*4882a593Smuzhiyun 		}
257*4882a593Smuzhiyun 	}
258*4882a593Smuzhiyun 
259*4882a593Smuzhiyun 	return 0;
260*4882a593Smuzhiyun }
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun /* This function must be called with the mv_xor_chan spinlock held */
mv_chan_slot_cleanup(struct mv_xor_chan * mv_chan)263*4882a593Smuzhiyun static void mv_chan_slot_cleanup(struct mv_xor_chan *mv_chan)
264*4882a593Smuzhiyun {
265*4882a593Smuzhiyun 	struct mv_xor_desc_slot *iter, *_iter;
266*4882a593Smuzhiyun 	dma_cookie_t cookie = 0;
267*4882a593Smuzhiyun 	int busy = mv_chan_is_busy(mv_chan);
268*4882a593Smuzhiyun 	u32 current_desc = mv_chan_get_current_desc(mv_chan);
269*4882a593Smuzhiyun 	int current_cleaned = 0;
270*4882a593Smuzhiyun 	struct mv_xor_desc *hw_desc;
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan), "%s %d\n", __func__, __LINE__);
273*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan), "current_desc %x\n", current_desc);
274*4882a593Smuzhiyun 	mv_chan_clean_completed_slots(mv_chan);
275*4882a593Smuzhiyun 
276*4882a593Smuzhiyun 	/* free completed slots from the chain starting with
277*4882a593Smuzhiyun 	 * the oldest descriptor
278*4882a593Smuzhiyun 	 */
279*4882a593Smuzhiyun 
280*4882a593Smuzhiyun 	list_for_each_entry_safe(iter, _iter, &mv_chan->chain,
281*4882a593Smuzhiyun 				 node) {
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 		/* clean finished descriptors */
284*4882a593Smuzhiyun 		hw_desc = iter->hw_desc;
285*4882a593Smuzhiyun 		if (hw_desc->status & XOR_DESC_SUCCESS) {
286*4882a593Smuzhiyun 			cookie = mv_desc_run_tx_complete_actions(iter, mv_chan,
287*4882a593Smuzhiyun 								 cookie);
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun 			/* done processing desc, clean slot */
290*4882a593Smuzhiyun 			mv_desc_clean_slot(iter, mv_chan);
291*4882a593Smuzhiyun 
292*4882a593Smuzhiyun 			/* break if we did cleaned the current */
293*4882a593Smuzhiyun 			if (iter->async_tx.phys == current_desc) {
294*4882a593Smuzhiyun 				current_cleaned = 1;
295*4882a593Smuzhiyun 				break;
296*4882a593Smuzhiyun 			}
297*4882a593Smuzhiyun 		} else {
298*4882a593Smuzhiyun 			if (iter->async_tx.phys == current_desc) {
299*4882a593Smuzhiyun 				current_cleaned = 0;
300*4882a593Smuzhiyun 				break;
301*4882a593Smuzhiyun 			}
302*4882a593Smuzhiyun 		}
303*4882a593Smuzhiyun 	}
304*4882a593Smuzhiyun 
305*4882a593Smuzhiyun 	if ((busy == 0) && !list_empty(&mv_chan->chain)) {
306*4882a593Smuzhiyun 		if (current_cleaned) {
307*4882a593Smuzhiyun 			/*
308*4882a593Smuzhiyun 			 * current descriptor cleaned and removed, run
309*4882a593Smuzhiyun 			 * from list head
310*4882a593Smuzhiyun 			 */
311*4882a593Smuzhiyun 			iter = list_entry(mv_chan->chain.next,
312*4882a593Smuzhiyun 					  struct mv_xor_desc_slot,
313*4882a593Smuzhiyun 					  node);
314*4882a593Smuzhiyun 			mv_chan_start_new_chain(mv_chan, iter);
315*4882a593Smuzhiyun 		} else {
316*4882a593Smuzhiyun 			if (!list_is_last(&iter->node, &mv_chan->chain)) {
317*4882a593Smuzhiyun 				/*
318*4882a593Smuzhiyun 				 * descriptors are still waiting after
319*4882a593Smuzhiyun 				 * current, trigger them
320*4882a593Smuzhiyun 				 */
321*4882a593Smuzhiyun 				iter = list_entry(iter->node.next,
322*4882a593Smuzhiyun 						  struct mv_xor_desc_slot,
323*4882a593Smuzhiyun 						  node);
324*4882a593Smuzhiyun 				mv_chan_start_new_chain(mv_chan, iter);
325*4882a593Smuzhiyun 			} else {
326*4882a593Smuzhiyun 				/*
327*4882a593Smuzhiyun 				 * some descriptors are still waiting
328*4882a593Smuzhiyun 				 * to be cleaned
329*4882a593Smuzhiyun 				 */
330*4882a593Smuzhiyun 				tasklet_schedule(&mv_chan->irq_tasklet);
331*4882a593Smuzhiyun 			}
332*4882a593Smuzhiyun 		}
333*4882a593Smuzhiyun 	}
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	if (cookie > 0)
336*4882a593Smuzhiyun 		mv_chan->dmachan.completed_cookie = cookie;
337*4882a593Smuzhiyun }
338*4882a593Smuzhiyun 
mv_xor_tasklet(struct tasklet_struct * t)339*4882a593Smuzhiyun static void mv_xor_tasklet(struct tasklet_struct *t)
340*4882a593Smuzhiyun {
341*4882a593Smuzhiyun 	struct mv_xor_chan *chan = from_tasklet(chan, t, irq_tasklet);
342*4882a593Smuzhiyun 
343*4882a593Smuzhiyun 	spin_lock(&chan->lock);
344*4882a593Smuzhiyun 	mv_chan_slot_cleanup(chan);
345*4882a593Smuzhiyun 	spin_unlock(&chan->lock);
346*4882a593Smuzhiyun }
347*4882a593Smuzhiyun 
348*4882a593Smuzhiyun static struct mv_xor_desc_slot *
mv_chan_alloc_slot(struct mv_xor_chan * mv_chan)349*4882a593Smuzhiyun mv_chan_alloc_slot(struct mv_xor_chan *mv_chan)
350*4882a593Smuzhiyun {
351*4882a593Smuzhiyun 	struct mv_xor_desc_slot *iter;
352*4882a593Smuzhiyun 
353*4882a593Smuzhiyun 	spin_lock_bh(&mv_chan->lock);
354*4882a593Smuzhiyun 
355*4882a593Smuzhiyun 	if (!list_empty(&mv_chan->free_slots)) {
356*4882a593Smuzhiyun 		iter = list_first_entry(&mv_chan->free_slots,
357*4882a593Smuzhiyun 					struct mv_xor_desc_slot,
358*4882a593Smuzhiyun 					node);
359*4882a593Smuzhiyun 
360*4882a593Smuzhiyun 		list_move_tail(&iter->node, &mv_chan->allocated_slots);
361*4882a593Smuzhiyun 
362*4882a593Smuzhiyun 		spin_unlock_bh(&mv_chan->lock);
363*4882a593Smuzhiyun 
364*4882a593Smuzhiyun 		/* pre-ack descriptor */
365*4882a593Smuzhiyun 		async_tx_ack(&iter->async_tx);
366*4882a593Smuzhiyun 		iter->async_tx.cookie = -EBUSY;
367*4882a593Smuzhiyun 
368*4882a593Smuzhiyun 		return iter;
369*4882a593Smuzhiyun 
370*4882a593Smuzhiyun 	}
371*4882a593Smuzhiyun 
372*4882a593Smuzhiyun 	spin_unlock_bh(&mv_chan->lock);
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun 	/* try to free some slots if the allocation fails */
375*4882a593Smuzhiyun 	tasklet_schedule(&mv_chan->irq_tasklet);
376*4882a593Smuzhiyun 
377*4882a593Smuzhiyun 	return NULL;
378*4882a593Smuzhiyun }
379*4882a593Smuzhiyun 
380*4882a593Smuzhiyun /************************ DMA engine API functions ****************************/
381*4882a593Smuzhiyun static dma_cookie_t
mv_xor_tx_submit(struct dma_async_tx_descriptor * tx)382*4882a593Smuzhiyun mv_xor_tx_submit(struct dma_async_tx_descriptor *tx)
383*4882a593Smuzhiyun {
384*4882a593Smuzhiyun 	struct mv_xor_desc_slot *sw_desc = to_mv_xor_slot(tx);
385*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(tx->chan);
386*4882a593Smuzhiyun 	struct mv_xor_desc_slot *old_chain_tail;
387*4882a593Smuzhiyun 	dma_cookie_t cookie;
388*4882a593Smuzhiyun 	int new_hw_chain = 1;
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan),
391*4882a593Smuzhiyun 		"%s sw_desc %p: async_tx %p\n",
392*4882a593Smuzhiyun 		__func__, sw_desc, &sw_desc->async_tx);
393*4882a593Smuzhiyun 
394*4882a593Smuzhiyun 	spin_lock_bh(&mv_chan->lock);
395*4882a593Smuzhiyun 	cookie = dma_cookie_assign(tx);
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	if (list_empty(&mv_chan->chain))
398*4882a593Smuzhiyun 		list_move_tail(&sw_desc->node, &mv_chan->chain);
399*4882a593Smuzhiyun 	else {
400*4882a593Smuzhiyun 		new_hw_chain = 0;
401*4882a593Smuzhiyun 
402*4882a593Smuzhiyun 		old_chain_tail = list_entry(mv_chan->chain.prev,
403*4882a593Smuzhiyun 					    struct mv_xor_desc_slot,
404*4882a593Smuzhiyun 					    node);
405*4882a593Smuzhiyun 		list_move_tail(&sw_desc->node, &mv_chan->chain);
406*4882a593Smuzhiyun 
407*4882a593Smuzhiyun 		dev_dbg(mv_chan_to_devp(mv_chan), "Append to last desc %pa\n",
408*4882a593Smuzhiyun 			&old_chain_tail->async_tx.phys);
409*4882a593Smuzhiyun 
410*4882a593Smuzhiyun 		/* fix up the hardware chain */
411*4882a593Smuzhiyun 		mv_desc_set_next_desc(old_chain_tail, sw_desc->async_tx.phys);
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 		/* if the channel is not busy */
414*4882a593Smuzhiyun 		if (!mv_chan_is_busy(mv_chan)) {
415*4882a593Smuzhiyun 			u32 current_desc = mv_chan_get_current_desc(mv_chan);
416*4882a593Smuzhiyun 			/*
417*4882a593Smuzhiyun 			 * and the curren desc is the end of the chain before
418*4882a593Smuzhiyun 			 * the append, then we need to start the channel
419*4882a593Smuzhiyun 			 */
420*4882a593Smuzhiyun 			if (current_desc == old_chain_tail->async_tx.phys)
421*4882a593Smuzhiyun 				new_hw_chain = 1;
422*4882a593Smuzhiyun 		}
423*4882a593Smuzhiyun 	}
424*4882a593Smuzhiyun 
425*4882a593Smuzhiyun 	if (new_hw_chain)
426*4882a593Smuzhiyun 		mv_chan_start_new_chain(mv_chan, sw_desc);
427*4882a593Smuzhiyun 
428*4882a593Smuzhiyun 	spin_unlock_bh(&mv_chan->lock);
429*4882a593Smuzhiyun 
430*4882a593Smuzhiyun 	return cookie;
431*4882a593Smuzhiyun }
432*4882a593Smuzhiyun 
433*4882a593Smuzhiyun /* returns the number of allocated descriptors */
mv_xor_alloc_chan_resources(struct dma_chan * chan)434*4882a593Smuzhiyun static int mv_xor_alloc_chan_resources(struct dma_chan *chan)
435*4882a593Smuzhiyun {
436*4882a593Smuzhiyun 	void *virt_desc;
437*4882a593Smuzhiyun 	dma_addr_t dma_desc;
438*4882a593Smuzhiyun 	int idx;
439*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
440*4882a593Smuzhiyun 	struct mv_xor_desc_slot *slot = NULL;
441*4882a593Smuzhiyun 	int num_descs_in_pool = MV_XOR_POOL_SIZE/MV_XOR_SLOT_SIZE;
442*4882a593Smuzhiyun 
443*4882a593Smuzhiyun 	/* Allocate descriptor slots */
444*4882a593Smuzhiyun 	idx = mv_chan->slots_allocated;
445*4882a593Smuzhiyun 	while (idx < num_descs_in_pool) {
446*4882a593Smuzhiyun 		slot = kzalloc(sizeof(*slot), GFP_KERNEL);
447*4882a593Smuzhiyun 		if (!slot) {
448*4882a593Smuzhiyun 			dev_info(mv_chan_to_devp(mv_chan),
449*4882a593Smuzhiyun 				 "channel only initialized %d descriptor slots",
450*4882a593Smuzhiyun 				 idx);
451*4882a593Smuzhiyun 			break;
452*4882a593Smuzhiyun 		}
453*4882a593Smuzhiyun 		virt_desc = mv_chan->dma_desc_pool_virt;
454*4882a593Smuzhiyun 		slot->hw_desc = virt_desc + idx * MV_XOR_SLOT_SIZE;
455*4882a593Smuzhiyun 
456*4882a593Smuzhiyun 		dma_async_tx_descriptor_init(&slot->async_tx, chan);
457*4882a593Smuzhiyun 		slot->async_tx.tx_submit = mv_xor_tx_submit;
458*4882a593Smuzhiyun 		INIT_LIST_HEAD(&slot->node);
459*4882a593Smuzhiyun 		INIT_LIST_HEAD(&slot->sg_tx_list);
460*4882a593Smuzhiyun 		dma_desc = mv_chan->dma_desc_pool;
461*4882a593Smuzhiyun 		slot->async_tx.phys = dma_desc + idx * MV_XOR_SLOT_SIZE;
462*4882a593Smuzhiyun 		slot->idx = idx++;
463*4882a593Smuzhiyun 
464*4882a593Smuzhiyun 		spin_lock_bh(&mv_chan->lock);
465*4882a593Smuzhiyun 		mv_chan->slots_allocated = idx;
466*4882a593Smuzhiyun 		list_add_tail(&slot->node, &mv_chan->free_slots);
467*4882a593Smuzhiyun 		spin_unlock_bh(&mv_chan->lock);
468*4882a593Smuzhiyun 	}
469*4882a593Smuzhiyun 
470*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan),
471*4882a593Smuzhiyun 		"allocated %d descriptor slots\n",
472*4882a593Smuzhiyun 		mv_chan->slots_allocated);
473*4882a593Smuzhiyun 
474*4882a593Smuzhiyun 	return mv_chan->slots_allocated ? : -ENOMEM;
475*4882a593Smuzhiyun }
476*4882a593Smuzhiyun 
477*4882a593Smuzhiyun /*
478*4882a593Smuzhiyun  * Check if source or destination is an PCIe/IO address (non-SDRAM) and add
479*4882a593Smuzhiyun  * a new MBus window if necessary. Use a cache for these check so that
480*4882a593Smuzhiyun  * the MMIO mapped registers don't have to be accessed for this check
481*4882a593Smuzhiyun  * to speed up this process.
482*4882a593Smuzhiyun  */
mv_xor_add_io_win(struct mv_xor_chan * mv_chan,u32 addr)483*4882a593Smuzhiyun static int mv_xor_add_io_win(struct mv_xor_chan *mv_chan, u32 addr)
484*4882a593Smuzhiyun {
485*4882a593Smuzhiyun 	struct mv_xor_device *xordev = mv_chan->xordev;
486*4882a593Smuzhiyun 	void __iomem *base = mv_chan->mmr_high_base;
487*4882a593Smuzhiyun 	u32 win_enable;
488*4882a593Smuzhiyun 	u32 size;
489*4882a593Smuzhiyun 	u8 target, attr;
490*4882a593Smuzhiyun 	int ret;
491*4882a593Smuzhiyun 	int i;
492*4882a593Smuzhiyun 
493*4882a593Smuzhiyun 	/* Nothing needs to get done for the Armada 3700 */
494*4882a593Smuzhiyun 	if (xordev->xor_type == XOR_ARMADA_37XX)
495*4882a593Smuzhiyun 		return 0;
496*4882a593Smuzhiyun 
497*4882a593Smuzhiyun 	/*
498*4882a593Smuzhiyun 	 * Loop over the cached windows to check, if the requested area
499*4882a593Smuzhiyun 	 * is already mapped. If this the case, nothing needs to be done
500*4882a593Smuzhiyun 	 * and we can return.
501*4882a593Smuzhiyun 	 */
502*4882a593Smuzhiyun 	for (i = 0; i < WINDOW_COUNT; i++) {
503*4882a593Smuzhiyun 		if (addr >= xordev->win_start[i] &&
504*4882a593Smuzhiyun 		    addr <= xordev->win_end[i]) {
505*4882a593Smuzhiyun 			/* Window is already mapped */
506*4882a593Smuzhiyun 			return 0;
507*4882a593Smuzhiyun 		}
508*4882a593Smuzhiyun 	}
509*4882a593Smuzhiyun 
510*4882a593Smuzhiyun 	/*
511*4882a593Smuzhiyun 	 * The window is not mapped, so we need to create the new mapping
512*4882a593Smuzhiyun 	 */
513*4882a593Smuzhiyun 
514*4882a593Smuzhiyun 	/* If no IO window is found that addr has to be located in SDRAM */
515*4882a593Smuzhiyun 	ret = mvebu_mbus_get_io_win_info(addr, &size, &target, &attr);
516*4882a593Smuzhiyun 	if (ret < 0)
517*4882a593Smuzhiyun 		return 0;
518*4882a593Smuzhiyun 
519*4882a593Smuzhiyun 	/*
520*4882a593Smuzhiyun 	 * Mask the base addr 'addr' according to 'size' read back from the
521*4882a593Smuzhiyun 	 * MBus window. Otherwise we might end up with an address located
522*4882a593Smuzhiyun 	 * somewhere in the middle of this area here.
523*4882a593Smuzhiyun 	 */
524*4882a593Smuzhiyun 	size -= 1;
525*4882a593Smuzhiyun 	addr &= ~size;
526*4882a593Smuzhiyun 
527*4882a593Smuzhiyun 	/*
528*4882a593Smuzhiyun 	 * Reading one of both enabled register is enough, as they are always
529*4882a593Smuzhiyun 	 * programmed to the identical values
530*4882a593Smuzhiyun 	 */
531*4882a593Smuzhiyun 	win_enable = readl(base + WINDOW_BAR_ENABLE(0));
532*4882a593Smuzhiyun 
533*4882a593Smuzhiyun 	/* Set 'i' to the first free window to write the new values to */
534*4882a593Smuzhiyun 	i = ffs(~win_enable) - 1;
535*4882a593Smuzhiyun 	if (i >= WINDOW_COUNT)
536*4882a593Smuzhiyun 		return -ENOMEM;
537*4882a593Smuzhiyun 
538*4882a593Smuzhiyun 	writel((addr & 0xffff0000) | (attr << 8) | target,
539*4882a593Smuzhiyun 	       base + WINDOW_BASE(i));
540*4882a593Smuzhiyun 	writel(size & 0xffff0000, base + WINDOW_SIZE(i));
541*4882a593Smuzhiyun 
542*4882a593Smuzhiyun 	/* Fill the caching variables for later use */
543*4882a593Smuzhiyun 	xordev->win_start[i] = addr;
544*4882a593Smuzhiyun 	xordev->win_end[i] = addr + size;
545*4882a593Smuzhiyun 
546*4882a593Smuzhiyun 	win_enable |= (1 << i);
547*4882a593Smuzhiyun 	win_enable |= 3 << (16 + (2 * i));
548*4882a593Smuzhiyun 	writel(win_enable, base + WINDOW_BAR_ENABLE(0));
549*4882a593Smuzhiyun 	writel(win_enable, base + WINDOW_BAR_ENABLE(1));
550*4882a593Smuzhiyun 
551*4882a593Smuzhiyun 	return 0;
552*4882a593Smuzhiyun }
553*4882a593Smuzhiyun 
554*4882a593Smuzhiyun static struct dma_async_tx_descriptor *
mv_xor_prep_dma_xor(struct dma_chan * chan,dma_addr_t dest,dma_addr_t * src,unsigned int src_cnt,size_t len,unsigned long flags)555*4882a593Smuzhiyun mv_xor_prep_dma_xor(struct dma_chan *chan, dma_addr_t dest, dma_addr_t *src,
556*4882a593Smuzhiyun 		    unsigned int src_cnt, size_t len, unsigned long flags)
557*4882a593Smuzhiyun {
558*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
559*4882a593Smuzhiyun 	struct mv_xor_desc_slot *sw_desc;
560*4882a593Smuzhiyun 	int ret;
561*4882a593Smuzhiyun 
562*4882a593Smuzhiyun 	if (unlikely(len < MV_XOR_MIN_BYTE_COUNT))
563*4882a593Smuzhiyun 		return NULL;
564*4882a593Smuzhiyun 
565*4882a593Smuzhiyun 	BUG_ON(len > MV_XOR_MAX_BYTE_COUNT);
566*4882a593Smuzhiyun 
567*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan),
568*4882a593Smuzhiyun 		"%s src_cnt: %d len: %zu dest %pad flags: %ld\n",
569*4882a593Smuzhiyun 		__func__, src_cnt, len, &dest, flags);
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun 	/* Check if a new window needs to get added for 'dest' */
572*4882a593Smuzhiyun 	ret = mv_xor_add_io_win(mv_chan, dest);
573*4882a593Smuzhiyun 	if (ret)
574*4882a593Smuzhiyun 		return NULL;
575*4882a593Smuzhiyun 
576*4882a593Smuzhiyun 	sw_desc = mv_chan_alloc_slot(mv_chan);
577*4882a593Smuzhiyun 	if (sw_desc) {
578*4882a593Smuzhiyun 		sw_desc->type = DMA_XOR;
579*4882a593Smuzhiyun 		sw_desc->async_tx.flags = flags;
580*4882a593Smuzhiyun 		mv_desc_init(sw_desc, dest, len, flags);
581*4882a593Smuzhiyun 		if (mv_chan->op_in_desc == XOR_MODE_IN_DESC)
582*4882a593Smuzhiyun 			mv_desc_set_mode(sw_desc);
583*4882a593Smuzhiyun 		while (src_cnt--) {
584*4882a593Smuzhiyun 			/* Check if a new window needs to get added for 'src' */
585*4882a593Smuzhiyun 			ret = mv_xor_add_io_win(mv_chan, src[src_cnt]);
586*4882a593Smuzhiyun 			if (ret)
587*4882a593Smuzhiyun 				return NULL;
588*4882a593Smuzhiyun 			mv_desc_set_src_addr(sw_desc, src_cnt, src[src_cnt]);
589*4882a593Smuzhiyun 		}
590*4882a593Smuzhiyun 	}
591*4882a593Smuzhiyun 
592*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan),
593*4882a593Smuzhiyun 		"%s sw_desc %p async_tx %p \n",
594*4882a593Smuzhiyun 		__func__, sw_desc, &sw_desc->async_tx);
595*4882a593Smuzhiyun 	return sw_desc ? &sw_desc->async_tx : NULL;
596*4882a593Smuzhiyun }
597*4882a593Smuzhiyun 
598*4882a593Smuzhiyun static struct dma_async_tx_descriptor *
mv_xor_prep_dma_memcpy(struct dma_chan * chan,dma_addr_t dest,dma_addr_t src,size_t len,unsigned long flags)599*4882a593Smuzhiyun mv_xor_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src,
600*4882a593Smuzhiyun 		size_t len, unsigned long flags)
601*4882a593Smuzhiyun {
602*4882a593Smuzhiyun 	/*
603*4882a593Smuzhiyun 	 * A MEMCPY operation is identical to an XOR operation with only
604*4882a593Smuzhiyun 	 * a single source address.
605*4882a593Smuzhiyun 	 */
606*4882a593Smuzhiyun 	return mv_xor_prep_dma_xor(chan, dest, &src, 1, len, flags);
607*4882a593Smuzhiyun }
608*4882a593Smuzhiyun 
609*4882a593Smuzhiyun static struct dma_async_tx_descriptor *
mv_xor_prep_dma_interrupt(struct dma_chan * chan,unsigned long flags)610*4882a593Smuzhiyun mv_xor_prep_dma_interrupt(struct dma_chan *chan, unsigned long flags)
611*4882a593Smuzhiyun {
612*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
613*4882a593Smuzhiyun 	dma_addr_t src, dest;
614*4882a593Smuzhiyun 	size_t len;
615*4882a593Smuzhiyun 
616*4882a593Smuzhiyun 	src = mv_chan->dummy_src_addr;
617*4882a593Smuzhiyun 	dest = mv_chan->dummy_dst_addr;
618*4882a593Smuzhiyun 	len = MV_XOR_MIN_BYTE_COUNT;
619*4882a593Smuzhiyun 
620*4882a593Smuzhiyun 	/*
621*4882a593Smuzhiyun 	 * We implement the DMA_INTERRUPT operation as a minimum sized
622*4882a593Smuzhiyun 	 * XOR operation with a single dummy source address.
623*4882a593Smuzhiyun 	 */
624*4882a593Smuzhiyun 	return mv_xor_prep_dma_xor(chan, dest, &src, 1, len, flags);
625*4882a593Smuzhiyun }
626*4882a593Smuzhiyun 
mv_xor_free_chan_resources(struct dma_chan * chan)627*4882a593Smuzhiyun static void mv_xor_free_chan_resources(struct dma_chan *chan)
628*4882a593Smuzhiyun {
629*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
630*4882a593Smuzhiyun 	struct mv_xor_desc_slot *iter, *_iter;
631*4882a593Smuzhiyun 	int in_use_descs = 0;
632*4882a593Smuzhiyun 
633*4882a593Smuzhiyun 	spin_lock_bh(&mv_chan->lock);
634*4882a593Smuzhiyun 
635*4882a593Smuzhiyun 	mv_chan_slot_cleanup(mv_chan);
636*4882a593Smuzhiyun 
637*4882a593Smuzhiyun 	list_for_each_entry_safe(iter, _iter, &mv_chan->chain,
638*4882a593Smuzhiyun 					node) {
639*4882a593Smuzhiyun 		in_use_descs++;
640*4882a593Smuzhiyun 		list_move_tail(&iter->node, &mv_chan->free_slots);
641*4882a593Smuzhiyun 	}
642*4882a593Smuzhiyun 	list_for_each_entry_safe(iter, _iter, &mv_chan->completed_slots,
643*4882a593Smuzhiyun 				 node) {
644*4882a593Smuzhiyun 		in_use_descs++;
645*4882a593Smuzhiyun 		list_move_tail(&iter->node, &mv_chan->free_slots);
646*4882a593Smuzhiyun 	}
647*4882a593Smuzhiyun 	list_for_each_entry_safe(iter, _iter, &mv_chan->allocated_slots,
648*4882a593Smuzhiyun 				 node) {
649*4882a593Smuzhiyun 		in_use_descs++;
650*4882a593Smuzhiyun 		list_move_tail(&iter->node, &mv_chan->free_slots);
651*4882a593Smuzhiyun 	}
652*4882a593Smuzhiyun 	list_for_each_entry_safe_reverse(
653*4882a593Smuzhiyun 		iter, _iter, &mv_chan->free_slots, node) {
654*4882a593Smuzhiyun 		list_del(&iter->node);
655*4882a593Smuzhiyun 		kfree(iter);
656*4882a593Smuzhiyun 		mv_chan->slots_allocated--;
657*4882a593Smuzhiyun 	}
658*4882a593Smuzhiyun 
659*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(mv_chan), "%s slots_allocated %d\n",
660*4882a593Smuzhiyun 		__func__, mv_chan->slots_allocated);
661*4882a593Smuzhiyun 	spin_unlock_bh(&mv_chan->lock);
662*4882a593Smuzhiyun 
663*4882a593Smuzhiyun 	if (in_use_descs)
664*4882a593Smuzhiyun 		dev_err(mv_chan_to_devp(mv_chan),
665*4882a593Smuzhiyun 			"freeing %d in use descriptors!\n", in_use_descs);
666*4882a593Smuzhiyun }
667*4882a593Smuzhiyun 
668*4882a593Smuzhiyun /**
669*4882a593Smuzhiyun  * mv_xor_status - poll the status of an XOR transaction
670*4882a593Smuzhiyun  * @chan: XOR channel handle
671*4882a593Smuzhiyun  * @cookie: XOR transaction identifier
672*4882a593Smuzhiyun  * @txstate: XOR transactions state holder (or NULL)
673*4882a593Smuzhiyun  */
mv_xor_status(struct dma_chan * chan,dma_cookie_t cookie,struct dma_tx_state * txstate)674*4882a593Smuzhiyun static enum dma_status mv_xor_status(struct dma_chan *chan,
675*4882a593Smuzhiyun 					  dma_cookie_t cookie,
676*4882a593Smuzhiyun 					  struct dma_tx_state *txstate)
677*4882a593Smuzhiyun {
678*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
679*4882a593Smuzhiyun 	enum dma_status ret;
680*4882a593Smuzhiyun 
681*4882a593Smuzhiyun 	ret = dma_cookie_status(chan, cookie, txstate);
682*4882a593Smuzhiyun 	if (ret == DMA_COMPLETE)
683*4882a593Smuzhiyun 		return ret;
684*4882a593Smuzhiyun 
685*4882a593Smuzhiyun 	spin_lock_bh(&mv_chan->lock);
686*4882a593Smuzhiyun 	mv_chan_slot_cleanup(mv_chan);
687*4882a593Smuzhiyun 	spin_unlock_bh(&mv_chan->lock);
688*4882a593Smuzhiyun 
689*4882a593Smuzhiyun 	return dma_cookie_status(chan, cookie, txstate);
690*4882a593Smuzhiyun }
691*4882a593Smuzhiyun 
mv_chan_dump_regs(struct mv_xor_chan * chan)692*4882a593Smuzhiyun static void mv_chan_dump_regs(struct mv_xor_chan *chan)
693*4882a593Smuzhiyun {
694*4882a593Smuzhiyun 	u32 val;
695*4882a593Smuzhiyun 
696*4882a593Smuzhiyun 	val = readl_relaxed(XOR_CONFIG(chan));
697*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "config       0x%08x\n", val);
698*4882a593Smuzhiyun 
699*4882a593Smuzhiyun 	val = readl_relaxed(XOR_ACTIVATION(chan));
700*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "activation   0x%08x\n", val);
701*4882a593Smuzhiyun 
702*4882a593Smuzhiyun 	val = readl_relaxed(XOR_INTR_CAUSE(chan));
703*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "intr cause   0x%08x\n", val);
704*4882a593Smuzhiyun 
705*4882a593Smuzhiyun 	val = readl_relaxed(XOR_INTR_MASK(chan));
706*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "intr mask    0x%08x\n", val);
707*4882a593Smuzhiyun 
708*4882a593Smuzhiyun 	val = readl_relaxed(XOR_ERROR_CAUSE(chan));
709*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "error cause  0x%08x\n", val);
710*4882a593Smuzhiyun 
711*4882a593Smuzhiyun 	val = readl_relaxed(XOR_ERROR_ADDR(chan));
712*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "error addr   0x%08x\n", val);
713*4882a593Smuzhiyun }
714*4882a593Smuzhiyun 
mv_chan_err_interrupt_handler(struct mv_xor_chan * chan,u32 intr_cause)715*4882a593Smuzhiyun static void mv_chan_err_interrupt_handler(struct mv_xor_chan *chan,
716*4882a593Smuzhiyun 					  u32 intr_cause)
717*4882a593Smuzhiyun {
718*4882a593Smuzhiyun 	if (intr_cause & XOR_INT_ERR_DECODE) {
719*4882a593Smuzhiyun 		dev_dbg(mv_chan_to_devp(chan), "ignoring address decode error\n");
720*4882a593Smuzhiyun 		return;
721*4882a593Smuzhiyun 	}
722*4882a593Smuzhiyun 
723*4882a593Smuzhiyun 	dev_err(mv_chan_to_devp(chan), "error on chan %d. intr cause 0x%08x\n",
724*4882a593Smuzhiyun 		chan->idx, intr_cause);
725*4882a593Smuzhiyun 
726*4882a593Smuzhiyun 	mv_chan_dump_regs(chan);
727*4882a593Smuzhiyun 	WARN_ON(1);
728*4882a593Smuzhiyun }
729*4882a593Smuzhiyun 
mv_xor_interrupt_handler(int irq,void * data)730*4882a593Smuzhiyun static irqreturn_t mv_xor_interrupt_handler(int irq, void *data)
731*4882a593Smuzhiyun {
732*4882a593Smuzhiyun 	struct mv_xor_chan *chan = data;
733*4882a593Smuzhiyun 	u32 intr_cause = mv_chan_get_intr_cause(chan);
734*4882a593Smuzhiyun 
735*4882a593Smuzhiyun 	dev_dbg(mv_chan_to_devp(chan), "intr cause %x\n", intr_cause);
736*4882a593Smuzhiyun 
737*4882a593Smuzhiyun 	if (intr_cause & XOR_INTR_ERRORS)
738*4882a593Smuzhiyun 		mv_chan_err_interrupt_handler(chan, intr_cause);
739*4882a593Smuzhiyun 
740*4882a593Smuzhiyun 	tasklet_schedule(&chan->irq_tasklet);
741*4882a593Smuzhiyun 
742*4882a593Smuzhiyun 	mv_chan_clear_eoc_cause(chan);
743*4882a593Smuzhiyun 
744*4882a593Smuzhiyun 	return IRQ_HANDLED;
745*4882a593Smuzhiyun }
746*4882a593Smuzhiyun 
mv_xor_issue_pending(struct dma_chan * chan)747*4882a593Smuzhiyun static void mv_xor_issue_pending(struct dma_chan *chan)
748*4882a593Smuzhiyun {
749*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
750*4882a593Smuzhiyun 
751*4882a593Smuzhiyun 	if (mv_chan->pending >= MV_XOR_THRESHOLD) {
752*4882a593Smuzhiyun 		mv_chan->pending = 0;
753*4882a593Smuzhiyun 		mv_chan_activate(mv_chan);
754*4882a593Smuzhiyun 	}
755*4882a593Smuzhiyun }
756*4882a593Smuzhiyun 
757*4882a593Smuzhiyun /*
758*4882a593Smuzhiyun  * Perform a transaction to verify the HW works.
759*4882a593Smuzhiyun  */
760*4882a593Smuzhiyun 
mv_chan_memcpy_self_test(struct mv_xor_chan * mv_chan)761*4882a593Smuzhiyun static int mv_chan_memcpy_self_test(struct mv_xor_chan *mv_chan)
762*4882a593Smuzhiyun {
763*4882a593Smuzhiyun 	int i, ret;
764*4882a593Smuzhiyun 	void *src, *dest;
765*4882a593Smuzhiyun 	dma_addr_t src_dma, dest_dma;
766*4882a593Smuzhiyun 	struct dma_chan *dma_chan;
767*4882a593Smuzhiyun 	dma_cookie_t cookie;
768*4882a593Smuzhiyun 	struct dma_async_tx_descriptor *tx;
769*4882a593Smuzhiyun 	struct dmaengine_unmap_data *unmap;
770*4882a593Smuzhiyun 	int err = 0;
771*4882a593Smuzhiyun 
772*4882a593Smuzhiyun 	src = kmalloc(PAGE_SIZE, GFP_KERNEL);
773*4882a593Smuzhiyun 	if (!src)
774*4882a593Smuzhiyun 		return -ENOMEM;
775*4882a593Smuzhiyun 
776*4882a593Smuzhiyun 	dest = kzalloc(PAGE_SIZE, GFP_KERNEL);
777*4882a593Smuzhiyun 	if (!dest) {
778*4882a593Smuzhiyun 		kfree(src);
779*4882a593Smuzhiyun 		return -ENOMEM;
780*4882a593Smuzhiyun 	}
781*4882a593Smuzhiyun 
782*4882a593Smuzhiyun 	/* Fill in src buffer */
783*4882a593Smuzhiyun 	for (i = 0; i < PAGE_SIZE; i++)
784*4882a593Smuzhiyun 		((u8 *) src)[i] = (u8)i;
785*4882a593Smuzhiyun 
786*4882a593Smuzhiyun 	dma_chan = &mv_chan->dmachan;
787*4882a593Smuzhiyun 	if (mv_xor_alloc_chan_resources(dma_chan) < 1) {
788*4882a593Smuzhiyun 		err = -ENODEV;
789*4882a593Smuzhiyun 		goto out;
790*4882a593Smuzhiyun 	}
791*4882a593Smuzhiyun 
792*4882a593Smuzhiyun 	unmap = dmaengine_get_unmap_data(dma_chan->device->dev, 2, GFP_KERNEL);
793*4882a593Smuzhiyun 	if (!unmap) {
794*4882a593Smuzhiyun 		err = -ENOMEM;
795*4882a593Smuzhiyun 		goto free_resources;
796*4882a593Smuzhiyun 	}
797*4882a593Smuzhiyun 
798*4882a593Smuzhiyun 	src_dma = dma_map_page(dma_chan->device->dev, virt_to_page(src),
799*4882a593Smuzhiyun 			       offset_in_page(src), PAGE_SIZE,
800*4882a593Smuzhiyun 			       DMA_TO_DEVICE);
801*4882a593Smuzhiyun 	unmap->addr[0] = src_dma;
802*4882a593Smuzhiyun 
803*4882a593Smuzhiyun 	ret = dma_mapping_error(dma_chan->device->dev, src_dma);
804*4882a593Smuzhiyun 	if (ret) {
805*4882a593Smuzhiyun 		err = -ENOMEM;
806*4882a593Smuzhiyun 		goto free_resources;
807*4882a593Smuzhiyun 	}
808*4882a593Smuzhiyun 	unmap->to_cnt = 1;
809*4882a593Smuzhiyun 
810*4882a593Smuzhiyun 	dest_dma = dma_map_page(dma_chan->device->dev, virt_to_page(dest),
811*4882a593Smuzhiyun 				offset_in_page(dest), PAGE_SIZE,
812*4882a593Smuzhiyun 				DMA_FROM_DEVICE);
813*4882a593Smuzhiyun 	unmap->addr[1] = dest_dma;
814*4882a593Smuzhiyun 
815*4882a593Smuzhiyun 	ret = dma_mapping_error(dma_chan->device->dev, dest_dma);
816*4882a593Smuzhiyun 	if (ret) {
817*4882a593Smuzhiyun 		err = -ENOMEM;
818*4882a593Smuzhiyun 		goto free_resources;
819*4882a593Smuzhiyun 	}
820*4882a593Smuzhiyun 	unmap->from_cnt = 1;
821*4882a593Smuzhiyun 	unmap->len = PAGE_SIZE;
822*4882a593Smuzhiyun 
823*4882a593Smuzhiyun 	tx = mv_xor_prep_dma_memcpy(dma_chan, dest_dma, src_dma,
824*4882a593Smuzhiyun 				    PAGE_SIZE, 0);
825*4882a593Smuzhiyun 	if (!tx) {
826*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
827*4882a593Smuzhiyun 			"Self-test cannot prepare operation, disabling\n");
828*4882a593Smuzhiyun 		err = -ENODEV;
829*4882a593Smuzhiyun 		goto free_resources;
830*4882a593Smuzhiyun 	}
831*4882a593Smuzhiyun 
832*4882a593Smuzhiyun 	cookie = mv_xor_tx_submit(tx);
833*4882a593Smuzhiyun 	if (dma_submit_error(cookie)) {
834*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
835*4882a593Smuzhiyun 			"Self-test submit error, disabling\n");
836*4882a593Smuzhiyun 		err = -ENODEV;
837*4882a593Smuzhiyun 		goto free_resources;
838*4882a593Smuzhiyun 	}
839*4882a593Smuzhiyun 
840*4882a593Smuzhiyun 	mv_xor_issue_pending(dma_chan);
841*4882a593Smuzhiyun 	async_tx_ack(tx);
842*4882a593Smuzhiyun 	msleep(1);
843*4882a593Smuzhiyun 
844*4882a593Smuzhiyun 	if (mv_xor_status(dma_chan, cookie, NULL) !=
845*4882a593Smuzhiyun 	    DMA_COMPLETE) {
846*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
847*4882a593Smuzhiyun 			"Self-test copy timed out, disabling\n");
848*4882a593Smuzhiyun 		err = -ENODEV;
849*4882a593Smuzhiyun 		goto free_resources;
850*4882a593Smuzhiyun 	}
851*4882a593Smuzhiyun 
852*4882a593Smuzhiyun 	dma_sync_single_for_cpu(dma_chan->device->dev, dest_dma,
853*4882a593Smuzhiyun 				PAGE_SIZE, DMA_FROM_DEVICE);
854*4882a593Smuzhiyun 	if (memcmp(src, dest, PAGE_SIZE)) {
855*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
856*4882a593Smuzhiyun 			"Self-test copy failed compare, disabling\n");
857*4882a593Smuzhiyun 		err = -ENODEV;
858*4882a593Smuzhiyun 		goto free_resources;
859*4882a593Smuzhiyun 	}
860*4882a593Smuzhiyun 
861*4882a593Smuzhiyun free_resources:
862*4882a593Smuzhiyun 	dmaengine_unmap_put(unmap);
863*4882a593Smuzhiyun 	mv_xor_free_chan_resources(dma_chan);
864*4882a593Smuzhiyun out:
865*4882a593Smuzhiyun 	kfree(src);
866*4882a593Smuzhiyun 	kfree(dest);
867*4882a593Smuzhiyun 	return err;
868*4882a593Smuzhiyun }
869*4882a593Smuzhiyun 
870*4882a593Smuzhiyun #define MV_XOR_NUM_SRC_TEST 4 /* must be <= 15 */
871*4882a593Smuzhiyun static int
mv_chan_xor_self_test(struct mv_xor_chan * mv_chan)872*4882a593Smuzhiyun mv_chan_xor_self_test(struct mv_xor_chan *mv_chan)
873*4882a593Smuzhiyun {
874*4882a593Smuzhiyun 	int i, src_idx, ret;
875*4882a593Smuzhiyun 	struct page *dest;
876*4882a593Smuzhiyun 	struct page *xor_srcs[MV_XOR_NUM_SRC_TEST];
877*4882a593Smuzhiyun 	dma_addr_t dma_srcs[MV_XOR_NUM_SRC_TEST];
878*4882a593Smuzhiyun 	dma_addr_t dest_dma;
879*4882a593Smuzhiyun 	struct dma_async_tx_descriptor *tx;
880*4882a593Smuzhiyun 	struct dmaengine_unmap_data *unmap;
881*4882a593Smuzhiyun 	struct dma_chan *dma_chan;
882*4882a593Smuzhiyun 	dma_cookie_t cookie;
883*4882a593Smuzhiyun 	u8 cmp_byte = 0;
884*4882a593Smuzhiyun 	u32 cmp_word;
885*4882a593Smuzhiyun 	int err = 0;
886*4882a593Smuzhiyun 	int src_count = MV_XOR_NUM_SRC_TEST;
887*4882a593Smuzhiyun 
888*4882a593Smuzhiyun 	for (src_idx = 0; src_idx < src_count; src_idx++) {
889*4882a593Smuzhiyun 		xor_srcs[src_idx] = alloc_page(GFP_KERNEL);
890*4882a593Smuzhiyun 		if (!xor_srcs[src_idx]) {
891*4882a593Smuzhiyun 			while (src_idx--)
892*4882a593Smuzhiyun 				__free_page(xor_srcs[src_idx]);
893*4882a593Smuzhiyun 			return -ENOMEM;
894*4882a593Smuzhiyun 		}
895*4882a593Smuzhiyun 	}
896*4882a593Smuzhiyun 
897*4882a593Smuzhiyun 	dest = alloc_page(GFP_KERNEL);
898*4882a593Smuzhiyun 	if (!dest) {
899*4882a593Smuzhiyun 		while (src_idx--)
900*4882a593Smuzhiyun 			__free_page(xor_srcs[src_idx]);
901*4882a593Smuzhiyun 		return -ENOMEM;
902*4882a593Smuzhiyun 	}
903*4882a593Smuzhiyun 
904*4882a593Smuzhiyun 	/* Fill in src buffers */
905*4882a593Smuzhiyun 	for (src_idx = 0; src_idx < src_count; src_idx++) {
906*4882a593Smuzhiyun 		u8 *ptr = page_address(xor_srcs[src_idx]);
907*4882a593Smuzhiyun 		for (i = 0; i < PAGE_SIZE; i++)
908*4882a593Smuzhiyun 			ptr[i] = (1 << src_idx);
909*4882a593Smuzhiyun 	}
910*4882a593Smuzhiyun 
911*4882a593Smuzhiyun 	for (src_idx = 0; src_idx < src_count; src_idx++)
912*4882a593Smuzhiyun 		cmp_byte ^= (u8) (1 << src_idx);
913*4882a593Smuzhiyun 
914*4882a593Smuzhiyun 	cmp_word = (cmp_byte << 24) | (cmp_byte << 16) |
915*4882a593Smuzhiyun 		(cmp_byte << 8) | cmp_byte;
916*4882a593Smuzhiyun 
917*4882a593Smuzhiyun 	memset(page_address(dest), 0, PAGE_SIZE);
918*4882a593Smuzhiyun 
919*4882a593Smuzhiyun 	dma_chan = &mv_chan->dmachan;
920*4882a593Smuzhiyun 	if (mv_xor_alloc_chan_resources(dma_chan) < 1) {
921*4882a593Smuzhiyun 		err = -ENODEV;
922*4882a593Smuzhiyun 		goto out;
923*4882a593Smuzhiyun 	}
924*4882a593Smuzhiyun 
925*4882a593Smuzhiyun 	unmap = dmaengine_get_unmap_data(dma_chan->device->dev, src_count + 1,
926*4882a593Smuzhiyun 					 GFP_KERNEL);
927*4882a593Smuzhiyun 	if (!unmap) {
928*4882a593Smuzhiyun 		err = -ENOMEM;
929*4882a593Smuzhiyun 		goto free_resources;
930*4882a593Smuzhiyun 	}
931*4882a593Smuzhiyun 
932*4882a593Smuzhiyun 	/* test xor */
933*4882a593Smuzhiyun 	for (i = 0; i < src_count; i++) {
934*4882a593Smuzhiyun 		unmap->addr[i] = dma_map_page(dma_chan->device->dev, xor_srcs[i],
935*4882a593Smuzhiyun 					      0, PAGE_SIZE, DMA_TO_DEVICE);
936*4882a593Smuzhiyun 		dma_srcs[i] = unmap->addr[i];
937*4882a593Smuzhiyun 		ret = dma_mapping_error(dma_chan->device->dev, unmap->addr[i]);
938*4882a593Smuzhiyun 		if (ret) {
939*4882a593Smuzhiyun 			err = -ENOMEM;
940*4882a593Smuzhiyun 			goto free_resources;
941*4882a593Smuzhiyun 		}
942*4882a593Smuzhiyun 		unmap->to_cnt++;
943*4882a593Smuzhiyun 	}
944*4882a593Smuzhiyun 
945*4882a593Smuzhiyun 	unmap->addr[src_count] = dma_map_page(dma_chan->device->dev, dest, 0, PAGE_SIZE,
946*4882a593Smuzhiyun 				      DMA_FROM_DEVICE);
947*4882a593Smuzhiyun 	dest_dma = unmap->addr[src_count];
948*4882a593Smuzhiyun 	ret = dma_mapping_error(dma_chan->device->dev, unmap->addr[src_count]);
949*4882a593Smuzhiyun 	if (ret) {
950*4882a593Smuzhiyun 		err = -ENOMEM;
951*4882a593Smuzhiyun 		goto free_resources;
952*4882a593Smuzhiyun 	}
953*4882a593Smuzhiyun 	unmap->from_cnt = 1;
954*4882a593Smuzhiyun 	unmap->len = PAGE_SIZE;
955*4882a593Smuzhiyun 
956*4882a593Smuzhiyun 	tx = mv_xor_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
957*4882a593Smuzhiyun 				 src_count, PAGE_SIZE, 0);
958*4882a593Smuzhiyun 	if (!tx) {
959*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
960*4882a593Smuzhiyun 			"Self-test cannot prepare operation, disabling\n");
961*4882a593Smuzhiyun 		err = -ENODEV;
962*4882a593Smuzhiyun 		goto free_resources;
963*4882a593Smuzhiyun 	}
964*4882a593Smuzhiyun 
965*4882a593Smuzhiyun 	cookie = mv_xor_tx_submit(tx);
966*4882a593Smuzhiyun 	if (dma_submit_error(cookie)) {
967*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
968*4882a593Smuzhiyun 			"Self-test submit error, disabling\n");
969*4882a593Smuzhiyun 		err = -ENODEV;
970*4882a593Smuzhiyun 		goto free_resources;
971*4882a593Smuzhiyun 	}
972*4882a593Smuzhiyun 
973*4882a593Smuzhiyun 	mv_xor_issue_pending(dma_chan);
974*4882a593Smuzhiyun 	async_tx_ack(tx);
975*4882a593Smuzhiyun 	msleep(8);
976*4882a593Smuzhiyun 
977*4882a593Smuzhiyun 	if (mv_xor_status(dma_chan, cookie, NULL) !=
978*4882a593Smuzhiyun 	    DMA_COMPLETE) {
979*4882a593Smuzhiyun 		dev_err(dma_chan->device->dev,
980*4882a593Smuzhiyun 			"Self-test xor timed out, disabling\n");
981*4882a593Smuzhiyun 		err = -ENODEV;
982*4882a593Smuzhiyun 		goto free_resources;
983*4882a593Smuzhiyun 	}
984*4882a593Smuzhiyun 
985*4882a593Smuzhiyun 	dma_sync_single_for_cpu(dma_chan->device->dev, dest_dma,
986*4882a593Smuzhiyun 				PAGE_SIZE, DMA_FROM_DEVICE);
987*4882a593Smuzhiyun 	for (i = 0; i < (PAGE_SIZE / sizeof(u32)); i++) {
988*4882a593Smuzhiyun 		u32 *ptr = page_address(dest);
989*4882a593Smuzhiyun 		if (ptr[i] != cmp_word) {
990*4882a593Smuzhiyun 			dev_err(dma_chan->device->dev,
991*4882a593Smuzhiyun 				"Self-test xor failed compare, disabling. index %d, data %x, expected %x\n",
992*4882a593Smuzhiyun 				i, ptr[i], cmp_word);
993*4882a593Smuzhiyun 			err = -ENODEV;
994*4882a593Smuzhiyun 			goto free_resources;
995*4882a593Smuzhiyun 		}
996*4882a593Smuzhiyun 	}
997*4882a593Smuzhiyun 
998*4882a593Smuzhiyun free_resources:
999*4882a593Smuzhiyun 	dmaengine_unmap_put(unmap);
1000*4882a593Smuzhiyun 	mv_xor_free_chan_resources(dma_chan);
1001*4882a593Smuzhiyun out:
1002*4882a593Smuzhiyun 	src_idx = src_count;
1003*4882a593Smuzhiyun 	while (src_idx--)
1004*4882a593Smuzhiyun 		__free_page(xor_srcs[src_idx]);
1005*4882a593Smuzhiyun 	__free_page(dest);
1006*4882a593Smuzhiyun 	return err;
1007*4882a593Smuzhiyun }
1008*4882a593Smuzhiyun 
mv_xor_channel_remove(struct mv_xor_chan * mv_chan)1009*4882a593Smuzhiyun static int mv_xor_channel_remove(struct mv_xor_chan *mv_chan)
1010*4882a593Smuzhiyun {
1011*4882a593Smuzhiyun 	struct dma_chan *chan, *_chan;
1012*4882a593Smuzhiyun 	struct device *dev = mv_chan->dmadev.dev;
1013*4882a593Smuzhiyun 
1014*4882a593Smuzhiyun 	dma_async_device_unregister(&mv_chan->dmadev);
1015*4882a593Smuzhiyun 
1016*4882a593Smuzhiyun 	dma_free_coherent(dev, MV_XOR_POOL_SIZE,
1017*4882a593Smuzhiyun 			  mv_chan->dma_desc_pool_virt, mv_chan->dma_desc_pool);
1018*4882a593Smuzhiyun 	dma_unmap_single(dev, mv_chan->dummy_src_addr,
1019*4882a593Smuzhiyun 			 MV_XOR_MIN_BYTE_COUNT, DMA_FROM_DEVICE);
1020*4882a593Smuzhiyun 	dma_unmap_single(dev, mv_chan->dummy_dst_addr,
1021*4882a593Smuzhiyun 			 MV_XOR_MIN_BYTE_COUNT, DMA_TO_DEVICE);
1022*4882a593Smuzhiyun 
1023*4882a593Smuzhiyun 	list_for_each_entry_safe(chan, _chan, &mv_chan->dmadev.channels,
1024*4882a593Smuzhiyun 				 device_node) {
1025*4882a593Smuzhiyun 		list_del(&chan->device_node);
1026*4882a593Smuzhiyun 	}
1027*4882a593Smuzhiyun 
1028*4882a593Smuzhiyun 	free_irq(mv_chan->irq, mv_chan);
1029*4882a593Smuzhiyun 
1030*4882a593Smuzhiyun 	return 0;
1031*4882a593Smuzhiyun }
1032*4882a593Smuzhiyun 
1033*4882a593Smuzhiyun static struct mv_xor_chan *
mv_xor_channel_add(struct mv_xor_device * xordev,struct platform_device * pdev,int idx,dma_cap_mask_t cap_mask,int irq)1034*4882a593Smuzhiyun mv_xor_channel_add(struct mv_xor_device *xordev,
1035*4882a593Smuzhiyun 		   struct platform_device *pdev,
1036*4882a593Smuzhiyun 		   int idx, dma_cap_mask_t cap_mask, int irq)
1037*4882a593Smuzhiyun {
1038*4882a593Smuzhiyun 	int ret = 0;
1039*4882a593Smuzhiyun 	struct mv_xor_chan *mv_chan;
1040*4882a593Smuzhiyun 	struct dma_device *dma_dev;
1041*4882a593Smuzhiyun 
1042*4882a593Smuzhiyun 	mv_chan = devm_kzalloc(&pdev->dev, sizeof(*mv_chan), GFP_KERNEL);
1043*4882a593Smuzhiyun 	if (!mv_chan)
1044*4882a593Smuzhiyun 		return ERR_PTR(-ENOMEM);
1045*4882a593Smuzhiyun 
1046*4882a593Smuzhiyun 	mv_chan->idx = idx;
1047*4882a593Smuzhiyun 	mv_chan->irq = irq;
1048*4882a593Smuzhiyun 	if (xordev->xor_type == XOR_ORION)
1049*4882a593Smuzhiyun 		mv_chan->op_in_desc = XOR_MODE_IN_REG;
1050*4882a593Smuzhiyun 	else
1051*4882a593Smuzhiyun 		mv_chan->op_in_desc = XOR_MODE_IN_DESC;
1052*4882a593Smuzhiyun 
1053*4882a593Smuzhiyun 	dma_dev = &mv_chan->dmadev;
1054*4882a593Smuzhiyun 	dma_dev->dev = &pdev->dev;
1055*4882a593Smuzhiyun 	mv_chan->xordev = xordev;
1056*4882a593Smuzhiyun 
1057*4882a593Smuzhiyun 	/*
1058*4882a593Smuzhiyun 	 * These source and destination dummy buffers are used to implement
1059*4882a593Smuzhiyun 	 * a DMA_INTERRUPT operation as a minimum-sized XOR operation.
1060*4882a593Smuzhiyun 	 * Hence, we only need to map the buffers at initialization-time.
1061*4882a593Smuzhiyun 	 */
1062*4882a593Smuzhiyun 	mv_chan->dummy_src_addr = dma_map_single(dma_dev->dev,
1063*4882a593Smuzhiyun 		mv_chan->dummy_src, MV_XOR_MIN_BYTE_COUNT, DMA_FROM_DEVICE);
1064*4882a593Smuzhiyun 	mv_chan->dummy_dst_addr = dma_map_single(dma_dev->dev,
1065*4882a593Smuzhiyun 		mv_chan->dummy_dst, MV_XOR_MIN_BYTE_COUNT, DMA_TO_DEVICE);
1066*4882a593Smuzhiyun 
1067*4882a593Smuzhiyun 	/* allocate coherent memory for hardware descriptors
1068*4882a593Smuzhiyun 	 * note: writecombine gives slightly better performance, but
1069*4882a593Smuzhiyun 	 * requires that we explicitly flush the writes
1070*4882a593Smuzhiyun 	 */
1071*4882a593Smuzhiyun 	mv_chan->dma_desc_pool_virt =
1072*4882a593Smuzhiyun 	  dma_alloc_wc(&pdev->dev, MV_XOR_POOL_SIZE, &mv_chan->dma_desc_pool,
1073*4882a593Smuzhiyun 		       GFP_KERNEL);
1074*4882a593Smuzhiyun 	if (!mv_chan->dma_desc_pool_virt)
1075*4882a593Smuzhiyun 		return ERR_PTR(-ENOMEM);
1076*4882a593Smuzhiyun 
1077*4882a593Smuzhiyun 	/* discover transaction capabilites from the platform data */
1078*4882a593Smuzhiyun 	dma_dev->cap_mask = cap_mask;
1079*4882a593Smuzhiyun 
1080*4882a593Smuzhiyun 	INIT_LIST_HEAD(&dma_dev->channels);
1081*4882a593Smuzhiyun 
1082*4882a593Smuzhiyun 	/* set base routines */
1083*4882a593Smuzhiyun 	dma_dev->device_alloc_chan_resources = mv_xor_alloc_chan_resources;
1084*4882a593Smuzhiyun 	dma_dev->device_free_chan_resources = mv_xor_free_chan_resources;
1085*4882a593Smuzhiyun 	dma_dev->device_tx_status = mv_xor_status;
1086*4882a593Smuzhiyun 	dma_dev->device_issue_pending = mv_xor_issue_pending;
1087*4882a593Smuzhiyun 
1088*4882a593Smuzhiyun 	/* set prep routines based on capability */
1089*4882a593Smuzhiyun 	if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask))
1090*4882a593Smuzhiyun 		dma_dev->device_prep_dma_interrupt = mv_xor_prep_dma_interrupt;
1091*4882a593Smuzhiyun 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask))
1092*4882a593Smuzhiyun 		dma_dev->device_prep_dma_memcpy = mv_xor_prep_dma_memcpy;
1093*4882a593Smuzhiyun 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
1094*4882a593Smuzhiyun 		dma_dev->max_xor = 8;
1095*4882a593Smuzhiyun 		dma_dev->device_prep_dma_xor = mv_xor_prep_dma_xor;
1096*4882a593Smuzhiyun 	}
1097*4882a593Smuzhiyun 
1098*4882a593Smuzhiyun 	mv_chan->mmr_base = xordev->xor_base;
1099*4882a593Smuzhiyun 	mv_chan->mmr_high_base = xordev->xor_high_base;
1100*4882a593Smuzhiyun 	tasklet_setup(&mv_chan->irq_tasklet, mv_xor_tasklet);
1101*4882a593Smuzhiyun 
1102*4882a593Smuzhiyun 	/* clear errors before enabling interrupts */
1103*4882a593Smuzhiyun 	mv_chan_clear_err_status(mv_chan);
1104*4882a593Smuzhiyun 
1105*4882a593Smuzhiyun 	ret = request_irq(mv_chan->irq, mv_xor_interrupt_handler,
1106*4882a593Smuzhiyun 			  0, dev_name(&pdev->dev), mv_chan);
1107*4882a593Smuzhiyun 	if (ret)
1108*4882a593Smuzhiyun 		goto err_free_dma;
1109*4882a593Smuzhiyun 
1110*4882a593Smuzhiyun 	mv_chan_unmask_interrupts(mv_chan);
1111*4882a593Smuzhiyun 
1112*4882a593Smuzhiyun 	if (mv_chan->op_in_desc == XOR_MODE_IN_DESC)
1113*4882a593Smuzhiyun 		mv_chan_set_mode(mv_chan, XOR_OPERATION_MODE_IN_DESC);
1114*4882a593Smuzhiyun 	else
1115*4882a593Smuzhiyun 		mv_chan_set_mode(mv_chan, XOR_OPERATION_MODE_XOR);
1116*4882a593Smuzhiyun 
1117*4882a593Smuzhiyun 	spin_lock_init(&mv_chan->lock);
1118*4882a593Smuzhiyun 	INIT_LIST_HEAD(&mv_chan->chain);
1119*4882a593Smuzhiyun 	INIT_LIST_HEAD(&mv_chan->completed_slots);
1120*4882a593Smuzhiyun 	INIT_LIST_HEAD(&mv_chan->free_slots);
1121*4882a593Smuzhiyun 	INIT_LIST_HEAD(&mv_chan->allocated_slots);
1122*4882a593Smuzhiyun 	mv_chan->dmachan.device = dma_dev;
1123*4882a593Smuzhiyun 	dma_cookie_init(&mv_chan->dmachan);
1124*4882a593Smuzhiyun 
1125*4882a593Smuzhiyun 	list_add_tail(&mv_chan->dmachan.device_node, &dma_dev->channels);
1126*4882a593Smuzhiyun 
1127*4882a593Smuzhiyun 	if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) {
1128*4882a593Smuzhiyun 		ret = mv_chan_memcpy_self_test(mv_chan);
1129*4882a593Smuzhiyun 		dev_dbg(&pdev->dev, "memcpy self test returned %d\n", ret);
1130*4882a593Smuzhiyun 		if (ret)
1131*4882a593Smuzhiyun 			goto err_free_irq;
1132*4882a593Smuzhiyun 	}
1133*4882a593Smuzhiyun 
1134*4882a593Smuzhiyun 	if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
1135*4882a593Smuzhiyun 		ret = mv_chan_xor_self_test(mv_chan);
1136*4882a593Smuzhiyun 		dev_dbg(&pdev->dev, "xor self test returned %d\n", ret);
1137*4882a593Smuzhiyun 		if (ret)
1138*4882a593Smuzhiyun 			goto err_free_irq;
1139*4882a593Smuzhiyun 	}
1140*4882a593Smuzhiyun 
1141*4882a593Smuzhiyun 	dev_info(&pdev->dev, "Marvell XOR (%s): ( %s%s%s)\n",
1142*4882a593Smuzhiyun 		 mv_chan->op_in_desc ? "Descriptor Mode" : "Registers Mode",
1143*4882a593Smuzhiyun 		 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "",
1144*4882a593Smuzhiyun 		 dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "",
1145*4882a593Smuzhiyun 		 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : "");
1146*4882a593Smuzhiyun 
1147*4882a593Smuzhiyun 	ret = dma_async_device_register(dma_dev);
1148*4882a593Smuzhiyun 	if (ret)
1149*4882a593Smuzhiyun 		goto err_free_irq;
1150*4882a593Smuzhiyun 
1151*4882a593Smuzhiyun 	return mv_chan;
1152*4882a593Smuzhiyun 
1153*4882a593Smuzhiyun err_free_irq:
1154*4882a593Smuzhiyun 	free_irq(mv_chan->irq, mv_chan);
1155*4882a593Smuzhiyun err_free_dma:
1156*4882a593Smuzhiyun 	dma_free_coherent(&pdev->dev, MV_XOR_POOL_SIZE,
1157*4882a593Smuzhiyun 			  mv_chan->dma_desc_pool_virt, mv_chan->dma_desc_pool);
1158*4882a593Smuzhiyun 	return ERR_PTR(ret);
1159*4882a593Smuzhiyun }
1160*4882a593Smuzhiyun 
1161*4882a593Smuzhiyun static void
mv_xor_conf_mbus_windows(struct mv_xor_device * xordev,const struct mbus_dram_target_info * dram)1162*4882a593Smuzhiyun mv_xor_conf_mbus_windows(struct mv_xor_device *xordev,
1163*4882a593Smuzhiyun 			 const struct mbus_dram_target_info *dram)
1164*4882a593Smuzhiyun {
1165*4882a593Smuzhiyun 	void __iomem *base = xordev->xor_high_base;
1166*4882a593Smuzhiyun 	u32 win_enable = 0;
1167*4882a593Smuzhiyun 	int i;
1168*4882a593Smuzhiyun 
1169*4882a593Smuzhiyun 	for (i = 0; i < 8; i++) {
1170*4882a593Smuzhiyun 		writel(0, base + WINDOW_BASE(i));
1171*4882a593Smuzhiyun 		writel(0, base + WINDOW_SIZE(i));
1172*4882a593Smuzhiyun 		if (i < 4)
1173*4882a593Smuzhiyun 			writel(0, base + WINDOW_REMAP_HIGH(i));
1174*4882a593Smuzhiyun 	}
1175*4882a593Smuzhiyun 
1176*4882a593Smuzhiyun 	for (i = 0; i < dram->num_cs; i++) {
1177*4882a593Smuzhiyun 		const struct mbus_dram_window *cs = dram->cs + i;
1178*4882a593Smuzhiyun 
1179*4882a593Smuzhiyun 		writel((cs->base & 0xffff0000) |
1180*4882a593Smuzhiyun 		       (cs->mbus_attr << 8) |
1181*4882a593Smuzhiyun 		       dram->mbus_dram_target_id, base + WINDOW_BASE(i));
1182*4882a593Smuzhiyun 		writel((cs->size - 1) & 0xffff0000, base + WINDOW_SIZE(i));
1183*4882a593Smuzhiyun 
1184*4882a593Smuzhiyun 		/* Fill the caching variables for later use */
1185*4882a593Smuzhiyun 		xordev->win_start[i] = cs->base;
1186*4882a593Smuzhiyun 		xordev->win_end[i] = cs->base + cs->size - 1;
1187*4882a593Smuzhiyun 
1188*4882a593Smuzhiyun 		win_enable |= (1 << i);
1189*4882a593Smuzhiyun 		win_enable |= 3 << (16 + (2 * i));
1190*4882a593Smuzhiyun 	}
1191*4882a593Smuzhiyun 
1192*4882a593Smuzhiyun 	writel(win_enable, base + WINDOW_BAR_ENABLE(0));
1193*4882a593Smuzhiyun 	writel(win_enable, base + WINDOW_BAR_ENABLE(1));
1194*4882a593Smuzhiyun 	writel(0, base + WINDOW_OVERRIDE_CTRL(0));
1195*4882a593Smuzhiyun 	writel(0, base + WINDOW_OVERRIDE_CTRL(1));
1196*4882a593Smuzhiyun }
1197*4882a593Smuzhiyun 
1198*4882a593Smuzhiyun static void
mv_xor_conf_mbus_windows_a3700(struct mv_xor_device * xordev)1199*4882a593Smuzhiyun mv_xor_conf_mbus_windows_a3700(struct mv_xor_device *xordev)
1200*4882a593Smuzhiyun {
1201*4882a593Smuzhiyun 	void __iomem *base = xordev->xor_high_base;
1202*4882a593Smuzhiyun 	u32 win_enable = 0;
1203*4882a593Smuzhiyun 	int i;
1204*4882a593Smuzhiyun 
1205*4882a593Smuzhiyun 	for (i = 0; i < 8; i++) {
1206*4882a593Smuzhiyun 		writel(0, base + WINDOW_BASE(i));
1207*4882a593Smuzhiyun 		writel(0, base + WINDOW_SIZE(i));
1208*4882a593Smuzhiyun 		if (i < 4)
1209*4882a593Smuzhiyun 			writel(0, base + WINDOW_REMAP_HIGH(i));
1210*4882a593Smuzhiyun 	}
1211*4882a593Smuzhiyun 	/*
1212*4882a593Smuzhiyun 	 * For Armada3700 open default 4GB Mbus window. The dram
1213*4882a593Smuzhiyun 	 * related configuration are done at AXIS level.
1214*4882a593Smuzhiyun 	 */
1215*4882a593Smuzhiyun 	writel(0xffff0000, base + WINDOW_SIZE(0));
1216*4882a593Smuzhiyun 	win_enable |= 1;
1217*4882a593Smuzhiyun 	win_enable |= 3 << 16;
1218*4882a593Smuzhiyun 
1219*4882a593Smuzhiyun 	writel(win_enable, base + WINDOW_BAR_ENABLE(0));
1220*4882a593Smuzhiyun 	writel(win_enable, base + WINDOW_BAR_ENABLE(1));
1221*4882a593Smuzhiyun 	writel(0, base + WINDOW_OVERRIDE_CTRL(0));
1222*4882a593Smuzhiyun 	writel(0, base + WINDOW_OVERRIDE_CTRL(1));
1223*4882a593Smuzhiyun }
1224*4882a593Smuzhiyun 
1225*4882a593Smuzhiyun /*
1226*4882a593Smuzhiyun  * Since this XOR driver is basically used only for RAID5, we don't
1227*4882a593Smuzhiyun  * need to care about synchronizing ->suspend with DMA activity,
1228*4882a593Smuzhiyun  * because the DMA engine will naturally be quiet due to the block
1229*4882a593Smuzhiyun  * devices being suspended.
1230*4882a593Smuzhiyun  */
mv_xor_suspend(struct platform_device * pdev,pm_message_t state)1231*4882a593Smuzhiyun static int mv_xor_suspend(struct platform_device *pdev, pm_message_t state)
1232*4882a593Smuzhiyun {
1233*4882a593Smuzhiyun 	struct mv_xor_device *xordev = platform_get_drvdata(pdev);
1234*4882a593Smuzhiyun 	int i;
1235*4882a593Smuzhiyun 
1236*4882a593Smuzhiyun 	for (i = 0; i < MV_XOR_MAX_CHANNELS; i++) {
1237*4882a593Smuzhiyun 		struct mv_xor_chan *mv_chan = xordev->channels[i];
1238*4882a593Smuzhiyun 
1239*4882a593Smuzhiyun 		if (!mv_chan)
1240*4882a593Smuzhiyun 			continue;
1241*4882a593Smuzhiyun 
1242*4882a593Smuzhiyun 		mv_chan->saved_config_reg =
1243*4882a593Smuzhiyun 			readl_relaxed(XOR_CONFIG(mv_chan));
1244*4882a593Smuzhiyun 		mv_chan->saved_int_mask_reg =
1245*4882a593Smuzhiyun 			readl_relaxed(XOR_INTR_MASK(mv_chan));
1246*4882a593Smuzhiyun 	}
1247*4882a593Smuzhiyun 
1248*4882a593Smuzhiyun 	return 0;
1249*4882a593Smuzhiyun }
1250*4882a593Smuzhiyun 
mv_xor_resume(struct platform_device * dev)1251*4882a593Smuzhiyun static int mv_xor_resume(struct platform_device *dev)
1252*4882a593Smuzhiyun {
1253*4882a593Smuzhiyun 	struct mv_xor_device *xordev = platform_get_drvdata(dev);
1254*4882a593Smuzhiyun 	const struct mbus_dram_target_info *dram;
1255*4882a593Smuzhiyun 	int i;
1256*4882a593Smuzhiyun 
1257*4882a593Smuzhiyun 	for (i = 0; i < MV_XOR_MAX_CHANNELS; i++) {
1258*4882a593Smuzhiyun 		struct mv_xor_chan *mv_chan = xordev->channels[i];
1259*4882a593Smuzhiyun 
1260*4882a593Smuzhiyun 		if (!mv_chan)
1261*4882a593Smuzhiyun 			continue;
1262*4882a593Smuzhiyun 
1263*4882a593Smuzhiyun 		writel_relaxed(mv_chan->saved_config_reg,
1264*4882a593Smuzhiyun 			       XOR_CONFIG(mv_chan));
1265*4882a593Smuzhiyun 		writel_relaxed(mv_chan->saved_int_mask_reg,
1266*4882a593Smuzhiyun 			       XOR_INTR_MASK(mv_chan));
1267*4882a593Smuzhiyun 	}
1268*4882a593Smuzhiyun 
1269*4882a593Smuzhiyun 	if (xordev->xor_type == XOR_ARMADA_37XX) {
1270*4882a593Smuzhiyun 		mv_xor_conf_mbus_windows_a3700(xordev);
1271*4882a593Smuzhiyun 		return 0;
1272*4882a593Smuzhiyun 	}
1273*4882a593Smuzhiyun 
1274*4882a593Smuzhiyun 	dram = mv_mbus_dram_info();
1275*4882a593Smuzhiyun 	if (dram)
1276*4882a593Smuzhiyun 		mv_xor_conf_mbus_windows(xordev, dram);
1277*4882a593Smuzhiyun 
1278*4882a593Smuzhiyun 	return 0;
1279*4882a593Smuzhiyun }
1280*4882a593Smuzhiyun 
1281*4882a593Smuzhiyun static const struct of_device_id mv_xor_dt_ids[] = {
1282*4882a593Smuzhiyun 	{ .compatible = "marvell,orion-xor", .data = (void *)XOR_ORION },
1283*4882a593Smuzhiyun 	{ .compatible = "marvell,armada-380-xor", .data = (void *)XOR_ARMADA_38X },
1284*4882a593Smuzhiyun 	{ .compatible = "marvell,armada-3700-xor", .data = (void *)XOR_ARMADA_37XX },
1285*4882a593Smuzhiyun 	{},
1286*4882a593Smuzhiyun };
1287*4882a593Smuzhiyun 
1288*4882a593Smuzhiyun static unsigned int mv_xor_engine_count;
1289*4882a593Smuzhiyun 
mv_xor_probe(struct platform_device * pdev)1290*4882a593Smuzhiyun static int mv_xor_probe(struct platform_device *pdev)
1291*4882a593Smuzhiyun {
1292*4882a593Smuzhiyun 	const struct mbus_dram_target_info *dram;
1293*4882a593Smuzhiyun 	struct mv_xor_device *xordev;
1294*4882a593Smuzhiyun 	struct mv_xor_platform_data *pdata = dev_get_platdata(&pdev->dev);
1295*4882a593Smuzhiyun 	struct resource *res;
1296*4882a593Smuzhiyun 	unsigned int max_engines, max_channels;
1297*4882a593Smuzhiyun 	int i, ret;
1298*4882a593Smuzhiyun 
1299*4882a593Smuzhiyun 	dev_notice(&pdev->dev, "Marvell shared XOR driver\n");
1300*4882a593Smuzhiyun 
1301*4882a593Smuzhiyun 	xordev = devm_kzalloc(&pdev->dev, sizeof(*xordev), GFP_KERNEL);
1302*4882a593Smuzhiyun 	if (!xordev)
1303*4882a593Smuzhiyun 		return -ENOMEM;
1304*4882a593Smuzhiyun 
1305*4882a593Smuzhiyun 	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1306*4882a593Smuzhiyun 	if (!res)
1307*4882a593Smuzhiyun 		return -ENODEV;
1308*4882a593Smuzhiyun 
1309*4882a593Smuzhiyun 	xordev->xor_base = devm_ioremap(&pdev->dev, res->start,
1310*4882a593Smuzhiyun 					resource_size(res));
1311*4882a593Smuzhiyun 	if (!xordev->xor_base)
1312*4882a593Smuzhiyun 		return -EBUSY;
1313*4882a593Smuzhiyun 
1314*4882a593Smuzhiyun 	res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
1315*4882a593Smuzhiyun 	if (!res)
1316*4882a593Smuzhiyun 		return -ENODEV;
1317*4882a593Smuzhiyun 
1318*4882a593Smuzhiyun 	xordev->xor_high_base = devm_ioremap(&pdev->dev, res->start,
1319*4882a593Smuzhiyun 					     resource_size(res));
1320*4882a593Smuzhiyun 	if (!xordev->xor_high_base)
1321*4882a593Smuzhiyun 		return -EBUSY;
1322*4882a593Smuzhiyun 
1323*4882a593Smuzhiyun 	platform_set_drvdata(pdev, xordev);
1324*4882a593Smuzhiyun 
1325*4882a593Smuzhiyun 
1326*4882a593Smuzhiyun 	/*
1327*4882a593Smuzhiyun 	 * We need to know which type of XOR device we use before
1328*4882a593Smuzhiyun 	 * setting up. In non-dt case it can only be the legacy one.
1329*4882a593Smuzhiyun 	 */
1330*4882a593Smuzhiyun 	xordev->xor_type = XOR_ORION;
1331*4882a593Smuzhiyun 	if (pdev->dev.of_node) {
1332*4882a593Smuzhiyun 		const struct of_device_id *of_id =
1333*4882a593Smuzhiyun 			of_match_device(mv_xor_dt_ids,
1334*4882a593Smuzhiyun 					&pdev->dev);
1335*4882a593Smuzhiyun 
1336*4882a593Smuzhiyun 		xordev->xor_type = (uintptr_t)of_id->data;
1337*4882a593Smuzhiyun 	}
1338*4882a593Smuzhiyun 
1339*4882a593Smuzhiyun 	/*
1340*4882a593Smuzhiyun 	 * (Re-)program MBUS remapping windows if we are asked to.
1341*4882a593Smuzhiyun 	 */
1342*4882a593Smuzhiyun 	if (xordev->xor_type == XOR_ARMADA_37XX) {
1343*4882a593Smuzhiyun 		mv_xor_conf_mbus_windows_a3700(xordev);
1344*4882a593Smuzhiyun 	} else {
1345*4882a593Smuzhiyun 		dram = mv_mbus_dram_info();
1346*4882a593Smuzhiyun 		if (dram)
1347*4882a593Smuzhiyun 			mv_xor_conf_mbus_windows(xordev, dram);
1348*4882a593Smuzhiyun 	}
1349*4882a593Smuzhiyun 
1350*4882a593Smuzhiyun 	/* Not all platforms can gate the clock, so it is not
1351*4882a593Smuzhiyun 	 * an error if the clock does not exists.
1352*4882a593Smuzhiyun 	 */
1353*4882a593Smuzhiyun 	xordev->clk = clk_get(&pdev->dev, NULL);
1354*4882a593Smuzhiyun 	if (!IS_ERR(xordev->clk))
1355*4882a593Smuzhiyun 		clk_prepare_enable(xordev->clk);
1356*4882a593Smuzhiyun 
1357*4882a593Smuzhiyun 	/*
1358*4882a593Smuzhiyun 	 * We don't want to have more than one channel per CPU in
1359*4882a593Smuzhiyun 	 * order for async_tx to perform well. So we limit the number
1360*4882a593Smuzhiyun 	 * of engines and channels so that we take into account this
1361*4882a593Smuzhiyun 	 * constraint. Note that we also want to use channels from
1362*4882a593Smuzhiyun 	 * separate engines when possible.  For dual-CPU Armada 3700
1363*4882a593Smuzhiyun 	 * SoC with single XOR engine allow using its both channels.
1364*4882a593Smuzhiyun 	 */
1365*4882a593Smuzhiyun 	max_engines = num_present_cpus();
1366*4882a593Smuzhiyun 	if (xordev->xor_type == XOR_ARMADA_37XX)
1367*4882a593Smuzhiyun 		max_channels =	num_present_cpus();
1368*4882a593Smuzhiyun 	else
1369*4882a593Smuzhiyun 		max_channels = min_t(unsigned int,
1370*4882a593Smuzhiyun 				     MV_XOR_MAX_CHANNELS,
1371*4882a593Smuzhiyun 				     DIV_ROUND_UP(num_present_cpus(), 2));
1372*4882a593Smuzhiyun 
1373*4882a593Smuzhiyun 	if (mv_xor_engine_count >= max_engines)
1374*4882a593Smuzhiyun 		return 0;
1375*4882a593Smuzhiyun 
1376*4882a593Smuzhiyun 	if (pdev->dev.of_node) {
1377*4882a593Smuzhiyun 		struct device_node *np;
1378*4882a593Smuzhiyun 		int i = 0;
1379*4882a593Smuzhiyun 
1380*4882a593Smuzhiyun 		for_each_child_of_node(pdev->dev.of_node, np) {
1381*4882a593Smuzhiyun 			struct mv_xor_chan *chan;
1382*4882a593Smuzhiyun 			dma_cap_mask_t cap_mask;
1383*4882a593Smuzhiyun 			int irq;
1384*4882a593Smuzhiyun 
1385*4882a593Smuzhiyun 			if (i >= max_channels)
1386*4882a593Smuzhiyun 				continue;
1387*4882a593Smuzhiyun 
1388*4882a593Smuzhiyun 			dma_cap_zero(cap_mask);
1389*4882a593Smuzhiyun 			dma_cap_set(DMA_MEMCPY, cap_mask);
1390*4882a593Smuzhiyun 			dma_cap_set(DMA_XOR, cap_mask);
1391*4882a593Smuzhiyun 			dma_cap_set(DMA_INTERRUPT, cap_mask);
1392*4882a593Smuzhiyun 
1393*4882a593Smuzhiyun 			irq = irq_of_parse_and_map(np, 0);
1394*4882a593Smuzhiyun 			if (!irq) {
1395*4882a593Smuzhiyun 				ret = -ENODEV;
1396*4882a593Smuzhiyun 				goto err_channel_add;
1397*4882a593Smuzhiyun 			}
1398*4882a593Smuzhiyun 
1399*4882a593Smuzhiyun 			chan = mv_xor_channel_add(xordev, pdev, i,
1400*4882a593Smuzhiyun 						  cap_mask, irq);
1401*4882a593Smuzhiyun 			if (IS_ERR(chan)) {
1402*4882a593Smuzhiyun 				ret = PTR_ERR(chan);
1403*4882a593Smuzhiyun 				irq_dispose_mapping(irq);
1404*4882a593Smuzhiyun 				goto err_channel_add;
1405*4882a593Smuzhiyun 			}
1406*4882a593Smuzhiyun 
1407*4882a593Smuzhiyun 			xordev->channels[i] = chan;
1408*4882a593Smuzhiyun 			i++;
1409*4882a593Smuzhiyun 		}
1410*4882a593Smuzhiyun 	} else if (pdata && pdata->channels) {
1411*4882a593Smuzhiyun 		for (i = 0; i < max_channels; i++) {
1412*4882a593Smuzhiyun 			struct mv_xor_channel_data *cd;
1413*4882a593Smuzhiyun 			struct mv_xor_chan *chan;
1414*4882a593Smuzhiyun 			int irq;
1415*4882a593Smuzhiyun 
1416*4882a593Smuzhiyun 			cd = &pdata->channels[i];
1417*4882a593Smuzhiyun 			irq = platform_get_irq(pdev, i);
1418*4882a593Smuzhiyun 			if (irq < 0) {
1419*4882a593Smuzhiyun 				ret = irq;
1420*4882a593Smuzhiyun 				goto err_channel_add;
1421*4882a593Smuzhiyun 			}
1422*4882a593Smuzhiyun 
1423*4882a593Smuzhiyun 			chan = mv_xor_channel_add(xordev, pdev, i,
1424*4882a593Smuzhiyun 						  cd->cap_mask, irq);
1425*4882a593Smuzhiyun 			if (IS_ERR(chan)) {
1426*4882a593Smuzhiyun 				ret = PTR_ERR(chan);
1427*4882a593Smuzhiyun 				goto err_channel_add;
1428*4882a593Smuzhiyun 			}
1429*4882a593Smuzhiyun 
1430*4882a593Smuzhiyun 			xordev->channels[i] = chan;
1431*4882a593Smuzhiyun 		}
1432*4882a593Smuzhiyun 	}
1433*4882a593Smuzhiyun 
1434*4882a593Smuzhiyun 	return 0;
1435*4882a593Smuzhiyun 
1436*4882a593Smuzhiyun err_channel_add:
1437*4882a593Smuzhiyun 	for (i = 0; i < MV_XOR_MAX_CHANNELS; i++)
1438*4882a593Smuzhiyun 		if (xordev->channels[i]) {
1439*4882a593Smuzhiyun 			mv_xor_channel_remove(xordev->channels[i]);
1440*4882a593Smuzhiyun 			if (pdev->dev.of_node)
1441*4882a593Smuzhiyun 				irq_dispose_mapping(xordev->channels[i]->irq);
1442*4882a593Smuzhiyun 		}
1443*4882a593Smuzhiyun 
1444*4882a593Smuzhiyun 	if (!IS_ERR(xordev->clk)) {
1445*4882a593Smuzhiyun 		clk_disable_unprepare(xordev->clk);
1446*4882a593Smuzhiyun 		clk_put(xordev->clk);
1447*4882a593Smuzhiyun 	}
1448*4882a593Smuzhiyun 
1449*4882a593Smuzhiyun 	return ret;
1450*4882a593Smuzhiyun }
1451*4882a593Smuzhiyun 
1452*4882a593Smuzhiyun static struct platform_driver mv_xor_driver = {
1453*4882a593Smuzhiyun 	.probe		= mv_xor_probe,
1454*4882a593Smuzhiyun 	.suspend        = mv_xor_suspend,
1455*4882a593Smuzhiyun 	.resume         = mv_xor_resume,
1456*4882a593Smuzhiyun 	.driver		= {
1457*4882a593Smuzhiyun 		.name	        = MV_XOR_NAME,
1458*4882a593Smuzhiyun 		.of_match_table = of_match_ptr(mv_xor_dt_ids),
1459*4882a593Smuzhiyun 	},
1460*4882a593Smuzhiyun };
1461*4882a593Smuzhiyun 
1462*4882a593Smuzhiyun builtin_platform_driver(mv_xor_driver);
1463*4882a593Smuzhiyun 
1464*4882a593Smuzhiyun /*
1465*4882a593Smuzhiyun MODULE_AUTHOR("Saeed Bishara <saeed@marvell.com>");
1466*4882a593Smuzhiyun MODULE_DESCRIPTION("DMA engine driver for Marvell's XOR engine");
1467*4882a593Smuzhiyun MODULE_LICENSE("GPL");
1468*4882a593Smuzhiyun */
1469