xref: /OK3568_Linux_fs/kernel/drivers/net/wireless/ralink/rt2x00/rt2800mmio.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*	Copyright (C) 2009 - 2010 Ivo van Doorn <IvDoorn@gmail.com>
3*4882a593Smuzhiyun  *	Copyright (C) 2009 Alban Browaeys <prahal@yahoo.com>
4*4882a593Smuzhiyun  *	Copyright (C) 2009 Felix Fietkau <nbd@openwrt.org>
5*4882a593Smuzhiyun  *	Copyright (C) 2009 Luis Correia <luis.f.correia@gmail.com>
6*4882a593Smuzhiyun  *	Copyright (C) 2009 Mattias Nissler <mattias.nissler@gmx.de>
7*4882a593Smuzhiyun  *	Copyright (C) 2009 Mark Asselstine <asselsm@gmail.com>
8*4882a593Smuzhiyun  *	Copyright (C) 2009 Xose Vazquez Perez <xose.vazquez@gmail.com>
9*4882a593Smuzhiyun  *	Copyright (C) 2009 Bart Zolnierkiewicz <bzolnier@gmail.com>
10*4882a593Smuzhiyun  *	<http://rt2x00.serialmonkey.com>
11*4882a593Smuzhiyun  */
12*4882a593Smuzhiyun 
13*4882a593Smuzhiyun /*	Module: rt2800mmio
14*4882a593Smuzhiyun  *	Abstract: rt2800 MMIO device routines.
15*4882a593Smuzhiyun  */
16*4882a593Smuzhiyun 
17*4882a593Smuzhiyun #include <linux/kernel.h>
18*4882a593Smuzhiyun #include <linux/module.h>
19*4882a593Smuzhiyun #include <linux/export.h>
20*4882a593Smuzhiyun 
21*4882a593Smuzhiyun #include "rt2x00.h"
22*4882a593Smuzhiyun #include "rt2x00mmio.h"
23*4882a593Smuzhiyun #include "rt2800.h"
24*4882a593Smuzhiyun #include "rt2800lib.h"
25*4882a593Smuzhiyun #include "rt2800mmio.h"
26*4882a593Smuzhiyun 
rt2800mmio_get_dma_done(struct data_queue * queue)27*4882a593Smuzhiyun unsigned int rt2800mmio_get_dma_done(struct data_queue *queue)
28*4882a593Smuzhiyun {
29*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
30*4882a593Smuzhiyun 	struct queue_entry *entry;
31*4882a593Smuzhiyun 	int idx, qid;
32*4882a593Smuzhiyun 
33*4882a593Smuzhiyun 	switch (queue->qid) {
34*4882a593Smuzhiyun 	case QID_AC_VO:
35*4882a593Smuzhiyun 	case QID_AC_VI:
36*4882a593Smuzhiyun 	case QID_AC_BE:
37*4882a593Smuzhiyun 	case QID_AC_BK:
38*4882a593Smuzhiyun 		qid = queue->qid;
39*4882a593Smuzhiyun 		idx = rt2x00mmio_register_read(rt2x00dev, TX_DTX_IDX(qid));
40*4882a593Smuzhiyun 		break;
41*4882a593Smuzhiyun 	case QID_MGMT:
42*4882a593Smuzhiyun 		idx = rt2x00mmio_register_read(rt2x00dev, TX_DTX_IDX(5));
43*4882a593Smuzhiyun 		break;
44*4882a593Smuzhiyun 	case QID_RX:
45*4882a593Smuzhiyun 		entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
46*4882a593Smuzhiyun 		idx = entry->entry_idx;
47*4882a593Smuzhiyun 		break;
48*4882a593Smuzhiyun 	default:
49*4882a593Smuzhiyun 		WARN_ON_ONCE(1);
50*4882a593Smuzhiyun 		idx = 0;
51*4882a593Smuzhiyun 		break;
52*4882a593Smuzhiyun 	}
53*4882a593Smuzhiyun 
54*4882a593Smuzhiyun 	return idx;
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_get_dma_done);
57*4882a593Smuzhiyun 
58*4882a593Smuzhiyun /*
59*4882a593Smuzhiyun  * TX descriptor initialization
60*4882a593Smuzhiyun  */
rt2800mmio_get_txwi(struct queue_entry * entry)61*4882a593Smuzhiyun __le32 *rt2800mmio_get_txwi(struct queue_entry *entry)
62*4882a593Smuzhiyun {
63*4882a593Smuzhiyun 	return (__le32 *) entry->skb->data;
64*4882a593Smuzhiyun }
65*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_get_txwi);
66*4882a593Smuzhiyun 
rt2800mmio_write_tx_desc(struct queue_entry * entry,struct txentry_desc * txdesc)67*4882a593Smuzhiyun void rt2800mmio_write_tx_desc(struct queue_entry *entry,
68*4882a593Smuzhiyun 			      struct txentry_desc *txdesc)
69*4882a593Smuzhiyun {
70*4882a593Smuzhiyun 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
71*4882a593Smuzhiyun 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
72*4882a593Smuzhiyun 	__le32 *txd = entry_priv->desc;
73*4882a593Smuzhiyun 	u32 word;
74*4882a593Smuzhiyun 	const unsigned int txwi_size = entry->queue->winfo_size;
75*4882a593Smuzhiyun 
76*4882a593Smuzhiyun 	/*
77*4882a593Smuzhiyun 	 * The buffers pointed by SD_PTR0/SD_LEN0 and SD_PTR1/SD_LEN1
78*4882a593Smuzhiyun 	 * must contains a TXWI structure + 802.11 header + padding + 802.11
79*4882a593Smuzhiyun 	 * data. We choose to have SD_PTR0/SD_LEN0 only contains TXWI and
80*4882a593Smuzhiyun 	 * SD_PTR1/SD_LEN1 contains 802.11 header + padding + 802.11
81*4882a593Smuzhiyun 	 * data. It means that LAST_SEC0 is always 0.
82*4882a593Smuzhiyun 	 */
83*4882a593Smuzhiyun 
84*4882a593Smuzhiyun 	/*
85*4882a593Smuzhiyun 	 * Initialize TX descriptor
86*4882a593Smuzhiyun 	 */
87*4882a593Smuzhiyun 	word = 0;
88*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W0_SD_PTR0, skbdesc->skb_dma);
89*4882a593Smuzhiyun 	rt2x00_desc_write(txd, 0, word);
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 	word = 0;
92*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W1_SD_LEN1, entry->skb->len);
93*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W1_LAST_SEC1,
94*4882a593Smuzhiyun 			   !test_bit(ENTRY_TXD_MORE_FRAG, &txdesc->flags));
95*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W1_BURST,
96*4882a593Smuzhiyun 			   test_bit(ENTRY_TXD_BURST, &txdesc->flags));
97*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W1_SD_LEN0, txwi_size);
98*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W1_LAST_SEC0, 0);
99*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W1_DMA_DONE, 0);
100*4882a593Smuzhiyun 	rt2x00_desc_write(txd, 1, word);
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun 	word = 0;
103*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W2_SD_PTR1,
104*4882a593Smuzhiyun 			   skbdesc->skb_dma + txwi_size);
105*4882a593Smuzhiyun 	rt2x00_desc_write(txd, 2, word);
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun 	word = 0;
108*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W3_WIV,
109*4882a593Smuzhiyun 			   !test_bit(ENTRY_TXD_ENCRYPT_IV, &txdesc->flags));
110*4882a593Smuzhiyun 	rt2x00_set_field32(&word, TXD_W3_QSEL, 2);
111*4882a593Smuzhiyun 	rt2x00_desc_write(txd, 3, word);
112*4882a593Smuzhiyun 
113*4882a593Smuzhiyun 	/*
114*4882a593Smuzhiyun 	 * Register descriptor details in skb frame descriptor.
115*4882a593Smuzhiyun 	 */
116*4882a593Smuzhiyun 	skbdesc->desc = txd;
117*4882a593Smuzhiyun 	skbdesc->desc_len = TXD_DESC_SIZE;
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_write_tx_desc);
120*4882a593Smuzhiyun 
121*4882a593Smuzhiyun /*
122*4882a593Smuzhiyun  * RX control handlers
123*4882a593Smuzhiyun  */
rt2800mmio_fill_rxdone(struct queue_entry * entry,struct rxdone_entry_desc * rxdesc)124*4882a593Smuzhiyun void rt2800mmio_fill_rxdone(struct queue_entry *entry,
125*4882a593Smuzhiyun 			    struct rxdone_entry_desc *rxdesc)
126*4882a593Smuzhiyun {
127*4882a593Smuzhiyun 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
128*4882a593Smuzhiyun 	__le32 *rxd = entry_priv->desc;
129*4882a593Smuzhiyun 	u32 word;
130*4882a593Smuzhiyun 
131*4882a593Smuzhiyun 	word = rt2x00_desc_read(rxd, 3);
132*4882a593Smuzhiyun 
133*4882a593Smuzhiyun 	if (rt2x00_get_field32(word, RXD_W3_CRC_ERROR))
134*4882a593Smuzhiyun 		rxdesc->flags |= RX_FLAG_FAILED_FCS_CRC;
135*4882a593Smuzhiyun 
136*4882a593Smuzhiyun 	/*
137*4882a593Smuzhiyun 	 * Unfortunately we don't know the cipher type used during
138*4882a593Smuzhiyun 	 * decryption. This prevents us from correct providing
139*4882a593Smuzhiyun 	 * correct statistics through debugfs.
140*4882a593Smuzhiyun 	 */
141*4882a593Smuzhiyun 	rxdesc->cipher_status = rt2x00_get_field32(word, RXD_W3_CIPHER_ERROR);
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun 	if (rt2x00_get_field32(word, RXD_W3_DECRYPTED)) {
144*4882a593Smuzhiyun 		/*
145*4882a593Smuzhiyun 		 * Hardware has stripped IV/EIV data from 802.11 frame during
146*4882a593Smuzhiyun 		 * decryption. Unfortunately the descriptor doesn't contain
147*4882a593Smuzhiyun 		 * any fields with the EIV/IV data either, so they can't
148*4882a593Smuzhiyun 		 * be restored by rt2x00lib.
149*4882a593Smuzhiyun 		 */
150*4882a593Smuzhiyun 		rxdesc->flags |= RX_FLAG_IV_STRIPPED;
151*4882a593Smuzhiyun 
152*4882a593Smuzhiyun 		/*
153*4882a593Smuzhiyun 		 * The hardware has already checked the Michael Mic and has
154*4882a593Smuzhiyun 		 * stripped it from the frame. Signal this to mac80211.
155*4882a593Smuzhiyun 		 */
156*4882a593Smuzhiyun 		rxdesc->flags |= RX_FLAG_MMIC_STRIPPED;
157*4882a593Smuzhiyun 
158*4882a593Smuzhiyun 		if (rxdesc->cipher_status == RX_CRYPTO_SUCCESS) {
159*4882a593Smuzhiyun 			rxdesc->flags |= RX_FLAG_DECRYPTED;
160*4882a593Smuzhiyun 		} else if (rxdesc->cipher_status == RX_CRYPTO_FAIL_MIC) {
161*4882a593Smuzhiyun 			/*
162*4882a593Smuzhiyun 			 * In order to check the Michael Mic, the packet must have
163*4882a593Smuzhiyun 			 * been decrypted.  Mac80211 doesnt check the MMIC failure
164*4882a593Smuzhiyun 			 * flag to initiate MMIC countermeasures if the decoded flag
165*4882a593Smuzhiyun 			 * has not been set.
166*4882a593Smuzhiyun 			 */
167*4882a593Smuzhiyun 			rxdesc->flags |= RX_FLAG_DECRYPTED;
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 			rxdesc->flags |= RX_FLAG_MMIC_ERROR;
170*4882a593Smuzhiyun 		}
171*4882a593Smuzhiyun 	}
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun 	if (rt2x00_get_field32(word, RXD_W3_MY_BSS))
174*4882a593Smuzhiyun 		rxdesc->dev_flags |= RXDONE_MY_BSS;
175*4882a593Smuzhiyun 
176*4882a593Smuzhiyun 	if (rt2x00_get_field32(word, RXD_W3_L2PAD))
177*4882a593Smuzhiyun 		rxdesc->dev_flags |= RXDONE_L2PAD;
178*4882a593Smuzhiyun 
179*4882a593Smuzhiyun 	/*
180*4882a593Smuzhiyun 	 * Process the RXWI structure that is at the start of the buffer.
181*4882a593Smuzhiyun 	 */
182*4882a593Smuzhiyun 	rt2800_process_rxwi(entry, rxdesc);
183*4882a593Smuzhiyun }
184*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_fill_rxdone);
185*4882a593Smuzhiyun 
186*4882a593Smuzhiyun /*
187*4882a593Smuzhiyun  * Interrupt functions.
188*4882a593Smuzhiyun  */
rt2800mmio_wakeup(struct rt2x00_dev * rt2x00dev)189*4882a593Smuzhiyun static void rt2800mmio_wakeup(struct rt2x00_dev *rt2x00dev)
190*4882a593Smuzhiyun {
191*4882a593Smuzhiyun 	struct ieee80211_conf conf = { .flags = 0 };
192*4882a593Smuzhiyun 	struct rt2x00lib_conf libconf = { .conf = &conf };
193*4882a593Smuzhiyun 
194*4882a593Smuzhiyun 	rt2800_config(rt2x00dev, &libconf, IEEE80211_CONF_CHANGE_PS);
195*4882a593Smuzhiyun }
196*4882a593Smuzhiyun 
rt2800mmio_enable_interrupt(struct rt2x00_dev * rt2x00dev,struct rt2x00_field32 irq_field)197*4882a593Smuzhiyun static inline void rt2800mmio_enable_interrupt(struct rt2x00_dev *rt2x00dev,
198*4882a593Smuzhiyun 					       struct rt2x00_field32 irq_field)
199*4882a593Smuzhiyun {
200*4882a593Smuzhiyun 	u32 reg;
201*4882a593Smuzhiyun 
202*4882a593Smuzhiyun 	/*
203*4882a593Smuzhiyun 	 * Enable a single interrupt. The interrupt mask register
204*4882a593Smuzhiyun 	 * access needs locking.
205*4882a593Smuzhiyun 	 */
206*4882a593Smuzhiyun 	spin_lock_irq(&rt2x00dev->irqmask_lock);
207*4882a593Smuzhiyun 	reg = rt2x00mmio_register_read(rt2x00dev, INT_MASK_CSR);
208*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, irq_field, 1);
209*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
210*4882a593Smuzhiyun 	spin_unlock_irq(&rt2x00dev->irqmask_lock);
211*4882a593Smuzhiyun }
212*4882a593Smuzhiyun 
rt2800mmio_pretbtt_tasklet(struct tasklet_struct * t)213*4882a593Smuzhiyun void rt2800mmio_pretbtt_tasklet(struct tasklet_struct *t)
214*4882a593Smuzhiyun {
215*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
216*4882a593Smuzhiyun 						    pretbtt_tasklet);
217*4882a593Smuzhiyun 	rt2x00lib_pretbtt(rt2x00dev);
218*4882a593Smuzhiyun 	if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
219*4882a593Smuzhiyun 		rt2800mmio_enable_interrupt(rt2x00dev, INT_MASK_CSR_PRE_TBTT);
220*4882a593Smuzhiyun }
221*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_pretbtt_tasklet);
222*4882a593Smuzhiyun 
rt2800mmio_tbtt_tasklet(struct tasklet_struct * t)223*4882a593Smuzhiyun void rt2800mmio_tbtt_tasklet(struct tasklet_struct *t)
224*4882a593Smuzhiyun {
225*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t, tbtt_tasklet);
226*4882a593Smuzhiyun 	struct rt2800_drv_data *drv_data = rt2x00dev->drv_data;
227*4882a593Smuzhiyun 	u32 reg;
228*4882a593Smuzhiyun 
229*4882a593Smuzhiyun 	rt2x00lib_beacondone(rt2x00dev);
230*4882a593Smuzhiyun 
231*4882a593Smuzhiyun 	if (rt2x00dev->intf_ap_count) {
232*4882a593Smuzhiyun 		/*
233*4882a593Smuzhiyun 		 * The rt2800pci hardware tbtt timer is off by 1us per tbtt
234*4882a593Smuzhiyun 		 * causing beacon skew and as a result causing problems with
235*4882a593Smuzhiyun 		 * some powersaving clients over time. Shorten the beacon
236*4882a593Smuzhiyun 		 * interval every 64 beacons by 64us to mitigate this effect.
237*4882a593Smuzhiyun 		 */
238*4882a593Smuzhiyun 		if (drv_data->tbtt_tick == (BCN_TBTT_OFFSET - 2)) {
239*4882a593Smuzhiyun 			reg = rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG);
240*4882a593Smuzhiyun 			rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_INTERVAL,
241*4882a593Smuzhiyun 					   (rt2x00dev->beacon_int * 16) - 1);
242*4882a593Smuzhiyun 			rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
243*4882a593Smuzhiyun 		} else if (drv_data->tbtt_tick == (BCN_TBTT_OFFSET - 1)) {
244*4882a593Smuzhiyun 			reg = rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG);
245*4882a593Smuzhiyun 			rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_INTERVAL,
246*4882a593Smuzhiyun 					   (rt2x00dev->beacon_int * 16));
247*4882a593Smuzhiyun 			rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
248*4882a593Smuzhiyun 		}
249*4882a593Smuzhiyun 		drv_data->tbtt_tick++;
250*4882a593Smuzhiyun 		drv_data->tbtt_tick %= BCN_TBTT_OFFSET;
251*4882a593Smuzhiyun 	}
252*4882a593Smuzhiyun 
253*4882a593Smuzhiyun 	if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
254*4882a593Smuzhiyun 		rt2800mmio_enable_interrupt(rt2x00dev, INT_MASK_CSR_TBTT);
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_tbtt_tasklet);
257*4882a593Smuzhiyun 
rt2800mmio_rxdone_tasklet(struct tasklet_struct * t)258*4882a593Smuzhiyun void rt2800mmio_rxdone_tasklet(struct tasklet_struct *t)
259*4882a593Smuzhiyun {
260*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
261*4882a593Smuzhiyun 						    rxdone_tasklet);
262*4882a593Smuzhiyun 	if (rt2x00mmio_rxdone(rt2x00dev))
263*4882a593Smuzhiyun 		tasklet_schedule(&rt2x00dev->rxdone_tasklet);
264*4882a593Smuzhiyun 	else if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
265*4882a593Smuzhiyun 		rt2800mmio_enable_interrupt(rt2x00dev, INT_MASK_CSR_RX_DONE);
266*4882a593Smuzhiyun }
267*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_rxdone_tasklet);
268*4882a593Smuzhiyun 
rt2800mmio_autowake_tasklet(struct tasklet_struct * t)269*4882a593Smuzhiyun void rt2800mmio_autowake_tasklet(struct tasklet_struct *t)
270*4882a593Smuzhiyun {
271*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
272*4882a593Smuzhiyun 						    autowake_tasklet);
273*4882a593Smuzhiyun 	rt2800mmio_wakeup(rt2x00dev);
274*4882a593Smuzhiyun 	if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
275*4882a593Smuzhiyun 		rt2800mmio_enable_interrupt(rt2x00dev,
276*4882a593Smuzhiyun 					    INT_MASK_CSR_AUTO_WAKEUP);
277*4882a593Smuzhiyun }
278*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_autowake_tasklet);
279*4882a593Smuzhiyun 
rt2800mmio_fetch_txstatus(struct rt2x00_dev * rt2x00dev)280*4882a593Smuzhiyun static void rt2800mmio_fetch_txstatus(struct rt2x00_dev *rt2x00dev)
281*4882a593Smuzhiyun {
282*4882a593Smuzhiyun 	u32 status;
283*4882a593Smuzhiyun 	unsigned long flags;
284*4882a593Smuzhiyun 
285*4882a593Smuzhiyun 	/*
286*4882a593Smuzhiyun 	 * The TX_FIFO_STATUS interrupt needs special care. We should
287*4882a593Smuzhiyun 	 * read TX_STA_FIFO but we should do it immediately as otherwise
288*4882a593Smuzhiyun 	 * the register can overflow and we would lose status reports.
289*4882a593Smuzhiyun 	 *
290*4882a593Smuzhiyun 	 * Hence, read the TX_STA_FIFO register and copy all tx status
291*4882a593Smuzhiyun 	 * reports into a kernel FIFO which is handled in the txstatus
292*4882a593Smuzhiyun 	 * tasklet. We use a tasklet to process the tx status reports
293*4882a593Smuzhiyun 	 * because we can schedule the tasklet multiple times (when the
294*4882a593Smuzhiyun 	 * interrupt fires again during tx status processing).
295*4882a593Smuzhiyun 	 *
296*4882a593Smuzhiyun 	 * We also read statuses from tx status timeout timer, use
297*4882a593Smuzhiyun 	 * lock to prevent concurent writes to fifo.
298*4882a593Smuzhiyun 	 */
299*4882a593Smuzhiyun 
300*4882a593Smuzhiyun 	spin_lock_irqsave(&rt2x00dev->irqmask_lock, flags);
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun 	while (!kfifo_is_full(&rt2x00dev->txstatus_fifo)) {
303*4882a593Smuzhiyun 		status = rt2x00mmio_register_read(rt2x00dev, TX_STA_FIFO);
304*4882a593Smuzhiyun 		if (!rt2x00_get_field32(status, TX_STA_FIFO_VALID))
305*4882a593Smuzhiyun 			break;
306*4882a593Smuzhiyun 
307*4882a593Smuzhiyun 		kfifo_put(&rt2x00dev->txstatus_fifo, status);
308*4882a593Smuzhiyun 	}
309*4882a593Smuzhiyun 
310*4882a593Smuzhiyun 	spin_unlock_irqrestore(&rt2x00dev->irqmask_lock, flags);
311*4882a593Smuzhiyun }
312*4882a593Smuzhiyun 
rt2800mmio_txstatus_tasklet(struct tasklet_struct * t)313*4882a593Smuzhiyun void rt2800mmio_txstatus_tasklet(struct tasklet_struct *t)
314*4882a593Smuzhiyun {
315*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = from_tasklet(rt2x00dev, t,
316*4882a593Smuzhiyun 						    txstatus_tasklet);
317*4882a593Smuzhiyun 
318*4882a593Smuzhiyun 	rt2800_txdone(rt2x00dev, 16);
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 	if (!kfifo_is_empty(&rt2x00dev->txstatus_fifo))
321*4882a593Smuzhiyun 		tasklet_schedule(&rt2x00dev->txstatus_tasklet);
322*4882a593Smuzhiyun 
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_txstatus_tasklet);
325*4882a593Smuzhiyun 
rt2800mmio_interrupt(int irq,void * dev_instance)326*4882a593Smuzhiyun irqreturn_t rt2800mmio_interrupt(int irq, void *dev_instance)
327*4882a593Smuzhiyun {
328*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = dev_instance;
329*4882a593Smuzhiyun 	u32 reg, mask;
330*4882a593Smuzhiyun 
331*4882a593Smuzhiyun 	/* Read status and ACK all interrupts */
332*4882a593Smuzhiyun 	reg = rt2x00mmio_register_read(rt2x00dev, INT_SOURCE_CSR);
333*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, INT_SOURCE_CSR, reg);
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	if (!reg)
336*4882a593Smuzhiyun 		return IRQ_NONE;
337*4882a593Smuzhiyun 
338*4882a593Smuzhiyun 	if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
339*4882a593Smuzhiyun 		return IRQ_HANDLED;
340*4882a593Smuzhiyun 
341*4882a593Smuzhiyun 	/*
342*4882a593Smuzhiyun 	 * Since INT_MASK_CSR and INT_SOURCE_CSR use the same bits
343*4882a593Smuzhiyun 	 * for interrupts and interrupt masks we can just use the value of
344*4882a593Smuzhiyun 	 * INT_SOURCE_CSR to create the interrupt mask.
345*4882a593Smuzhiyun 	 */
346*4882a593Smuzhiyun 	mask = ~reg;
347*4882a593Smuzhiyun 
348*4882a593Smuzhiyun 	if (rt2x00_get_field32(reg, INT_SOURCE_CSR_TX_FIFO_STATUS)) {
349*4882a593Smuzhiyun 		rt2x00_set_field32(&mask, INT_MASK_CSR_TX_FIFO_STATUS, 1);
350*4882a593Smuzhiyun 		rt2800mmio_fetch_txstatus(rt2x00dev);
351*4882a593Smuzhiyun 		if (!kfifo_is_empty(&rt2x00dev->txstatus_fifo))
352*4882a593Smuzhiyun 			tasklet_schedule(&rt2x00dev->txstatus_tasklet);
353*4882a593Smuzhiyun 	}
354*4882a593Smuzhiyun 
355*4882a593Smuzhiyun 	if (rt2x00_get_field32(reg, INT_SOURCE_CSR_PRE_TBTT))
356*4882a593Smuzhiyun 		tasklet_hi_schedule(&rt2x00dev->pretbtt_tasklet);
357*4882a593Smuzhiyun 
358*4882a593Smuzhiyun 	if (rt2x00_get_field32(reg, INT_SOURCE_CSR_TBTT))
359*4882a593Smuzhiyun 		tasklet_hi_schedule(&rt2x00dev->tbtt_tasklet);
360*4882a593Smuzhiyun 
361*4882a593Smuzhiyun 	if (rt2x00_get_field32(reg, INT_SOURCE_CSR_RX_DONE))
362*4882a593Smuzhiyun 		tasklet_schedule(&rt2x00dev->rxdone_tasklet);
363*4882a593Smuzhiyun 
364*4882a593Smuzhiyun 	if (rt2x00_get_field32(reg, INT_SOURCE_CSR_AUTO_WAKEUP))
365*4882a593Smuzhiyun 		tasklet_schedule(&rt2x00dev->autowake_tasklet);
366*4882a593Smuzhiyun 
367*4882a593Smuzhiyun 	/*
368*4882a593Smuzhiyun 	 * Disable all interrupts for which a tasklet was scheduled right now,
369*4882a593Smuzhiyun 	 * the tasklet will reenable the appropriate interrupts.
370*4882a593Smuzhiyun 	 */
371*4882a593Smuzhiyun 	spin_lock(&rt2x00dev->irqmask_lock);
372*4882a593Smuzhiyun 	reg = rt2x00mmio_register_read(rt2x00dev, INT_MASK_CSR);
373*4882a593Smuzhiyun 	reg &= mask;
374*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
375*4882a593Smuzhiyun 	spin_unlock(&rt2x00dev->irqmask_lock);
376*4882a593Smuzhiyun 
377*4882a593Smuzhiyun 	return IRQ_HANDLED;
378*4882a593Smuzhiyun }
379*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_interrupt);
380*4882a593Smuzhiyun 
rt2800mmio_toggle_irq(struct rt2x00_dev * rt2x00dev,enum dev_state state)381*4882a593Smuzhiyun void rt2800mmio_toggle_irq(struct rt2x00_dev *rt2x00dev,
382*4882a593Smuzhiyun 			   enum dev_state state)
383*4882a593Smuzhiyun {
384*4882a593Smuzhiyun 	u32 reg;
385*4882a593Smuzhiyun 	unsigned long flags;
386*4882a593Smuzhiyun 
387*4882a593Smuzhiyun 	/*
388*4882a593Smuzhiyun 	 * When interrupts are being enabled, the interrupt registers
389*4882a593Smuzhiyun 	 * should clear the register to assure a clean state.
390*4882a593Smuzhiyun 	 */
391*4882a593Smuzhiyun 	if (state == STATE_RADIO_IRQ_ON) {
392*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, INT_SOURCE_CSR);
393*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, INT_SOURCE_CSR, reg);
394*4882a593Smuzhiyun 	}
395*4882a593Smuzhiyun 
396*4882a593Smuzhiyun 	spin_lock_irqsave(&rt2x00dev->irqmask_lock, flags);
397*4882a593Smuzhiyun 	reg = 0;
398*4882a593Smuzhiyun 	if (state == STATE_RADIO_IRQ_ON) {
399*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_MASK_CSR_RX_DONE, 1);
400*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_MASK_CSR_TBTT, 1);
401*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_MASK_CSR_PRE_TBTT, 1);
402*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_MASK_CSR_TX_FIFO_STATUS, 1);
403*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_MASK_CSR_AUTO_WAKEUP, 1);
404*4882a593Smuzhiyun 	}
405*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
406*4882a593Smuzhiyun 	spin_unlock_irqrestore(&rt2x00dev->irqmask_lock, flags);
407*4882a593Smuzhiyun 
408*4882a593Smuzhiyun 	if (state == STATE_RADIO_IRQ_OFF) {
409*4882a593Smuzhiyun 		/*
410*4882a593Smuzhiyun 		 * Wait for possibly running tasklets to finish.
411*4882a593Smuzhiyun 		 */
412*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->txstatus_tasklet);
413*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->rxdone_tasklet);
414*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->autowake_tasklet);
415*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->tbtt_tasklet);
416*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->pretbtt_tasklet);
417*4882a593Smuzhiyun 	}
418*4882a593Smuzhiyun }
419*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_toggle_irq);
420*4882a593Smuzhiyun 
421*4882a593Smuzhiyun /*
422*4882a593Smuzhiyun  * Queue handlers.
423*4882a593Smuzhiyun  */
rt2800mmio_start_queue(struct data_queue * queue)424*4882a593Smuzhiyun void rt2800mmio_start_queue(struct data_queue *queue)
425*4882a593Smuzhiyun {
426*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
427*4882a593Smuzhiyun 	u32 reg;
428*4882a593Smuzhiyun 
429*4882a593Smuzhiyun 	switch (queue->qid) {
430*4882a593Smuzhiyun 	case QID_RX:
431*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, MAC_SYS_CTRL);
432*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, MAC_SYS_CTRL_ENABLE_RX, 1);
433*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
434*4882a593Smuzhiyun 		break;
435*4882a593Smuzhiyun 	case QID_BEACON:
436*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG);
437*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, BCN_TIME_CFG_TSF_TICKING, 1);
438*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, BCN_TIME_CFG_TBTT_ENABLE, 1);
439*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_GEN, 1);
440*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
441*4882a593Smuzhiyun 
442*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, INT_TIMER_EN);
443*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_TIMER_EN_PRE_TBTT_TIMER, 1);
444*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, INT_TIMER_EN, reg);
445*4882a593Smuzhiyun 		break;
446*4882a593Smuzhiyun 	default:
447*4882a593Smuzhiyun 		break;
448*4882a593Smuzhiyun 	}
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_start_queue);
451*4882a593Smuzhiyun 
452*4882a593Smuzhiyun /* 200 ms */
453*4882a593Smuzhiyun #define TXSTATUS_TIMEOUT 200000000
454*4882a593Smuzhiyun 
rt2800mmio_kick_queue(struct data_queue * queue)455*4882a593Smuzhiyun void rt2800mmio_kick_queue(struct data_queue *queue)
456*4882a593Smuzhiyun {
457*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
458*4882a593Smuzhiyun 	struct queue_entry *entry;
459*4882a593Smuzhiyun 
460*4882a593Smuzhiyun 	switch (queue->qid) {
461*4882a593Smuzhiyun 	case QID_AC_VO:
462*4882a593Smuzhiyun 	case QID_AC_VI:
463*4882a593Smuzhiyun 	case QID_AC_BE:
464*4882a593Smuzhiyun 	case QID_AC_BK:
465*4882a593Smuzhiyun 		WARN_ON_ONCE(rt2x00queue_empty(queue));
466*4882a593Smuzhiyun 		entry = rt2x00queue_get_entry(queue, Q_INDEX);
467*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(queue->qid),
468*4882a593Smuzhiyun 					  entry->entry_idx);
469*4882a593Smuzhiyun 		hrtimer_start(&rt2x00dev->txstatus_timer,
470*4882a593Smuzhiyun 			      TXSTATUS_TIMEOUT, HRTIMER_MODE_REL);
471*4882a593Smuzhiyun 		break;
472*4882a593Smuzhiyun 	case QID_MGMT:
473*4882a593Smuzhiyun 		entry = rt2x00queue_get_entry(queue, Q_INDEX);
474*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(5),
475*4882a593Smuzhiyun 					  entry->entry_idx);
476*4882a593Smuzhiyun 		break;
477*4882a593Smuzhiyun 	default:
478*4882a593Smuzhiyun 		break;
479*4882a593Smuzhiyun 	}
480*4882a593Smuzhiyun }
481*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_kick_queue);
482*4882a593Smuzhiyun 
rt2800mmio_flush_queue(struct data_queue * queue,bool drop)483*4882a593Smuzhiyun void rt2800mmio_flush_queue(struct data_queue *queue, bool drop)
484*4882a593Smuzhiyun {
485*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
486*4882a593Smuzhiyun 	bool tx_queue = false;
487*4882a593Smuzhiyun 	unsigned int i;
488*4882a593Smuzhiyun 
489*4882a593Smuzhiyun 	switch (queue->qid) {
490*4882a593Smuzhiyun 	case QID_AC_VO:
491*4882a593Smuzhiyun 	case QID_AC_VI:
492*4882a593Smuzhiyun 	case QID_AC_BE:
493*4882a593Smuzhiyun 	case QID_AC_BK:
494*4882a593Smuzhiyun 		tx_queue = true;
495*4882a593Smuzhiyun 		break;
496*4882a593Smuzhiyun 	case QID_RX:
497*4882a593Smuzhiyun 		break;
498*4882a593Smuzhiyun 	default:
499*4882a593Smuzhiyun 		return;
500*4882a593Smuzhiyun 	}
501*4882a593Smuzhiyun 
502*4882a593Smuzhiyun 	for (i = 0; i < 5; i++) {
503*4882a593Smuzhiyun 		/*
504*4882a593Smuzhiyun 		 * Check if the driver is already done, otherwise we
505*4882a593Smuzhiyun 		 * have to sleep a little while to give the driver/hw
506*4882a593Smuzhiyun 		 * the oppurtunity to complete interrupt process itself.
507*4882a593Smuzhiyun 		 */
508*4882a593Smuzhiyun 		if (rt2x00queue_empty(queue))
509*4882a593Smuzhiyun 			break;
510*4882a593Smuzhiyun 
511*4882a593Smuzhiyun 		/*
512*4882a593Smuzhiyun 		 * For TX queues schedule completion tasklet to catch
513*4882a593Smuzhiyun 		 * tx status timeouts, othewise just wait.
514*4882a593Smuzhiyun 		 */
515*4882a593Smuzhiyun 		if (tx_queue)
516*4882a593Smuzhiyun 			queue_work(rt2x00dev->workqueue, &rt2x00dev->txdone_work);
517*4882a593Smuzhiyun 
518*4882a593Smuzhiyun 		/*
519*4882a593Smuzhiyun 		 * Wait for a little while to give the driver
520*4882a593Smuzhiyun 		 * the oppurtunity to recover itself.
521*4882a593Smuzhiyun 		 */
522*4882a593Smuzhiyun 		msleep(50);
523*4882a593Smuzhiyun 	}
524*4882a593Smuzhiyun }
525*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_flush_queue);
526*4882a593Smuzhiyun 
rt2800mmio_stop_queue(struct data_queue * queue)527*4882a593Smuzhiyun void rt2800mmio_stop_queue(struct data_queue *queue)
528*4882a593Smuzhiyun {
529*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
530*4882a593Smuzhiyun 	u32 reg;
531*4882a593Smuzhiyun 
532*4882a593Smuzhiyun 	switch (queue->qid) {
533*4882a593Smuzhiyun 	case QID_RX:
534*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, MAC_SYS_CTRL);
535*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, MAC_SYS_CTRL_ENABLE_RX, 0);
536*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
537*4882a593Smuzhiyun 		break;
538*4882a593Smuzhiyun 	case QID_BEACON:
539*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG);
540*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, BCN_TIME_CFG_TSF_TICKING, 0);
541*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, BCN_TIME_CFG_TBTT_ENABLE, 0);
542*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_GEN, 0);
543*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
544*4882a593Smuzhiyun 
545*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, INT_TIMER_EN);
546*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, INT_TIMER_EN_PRE_TBTT_TIMER, 0);
547*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, INT_TIMER_EN, reg);
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun 		/*
550*4882a593Smuzhiyun 		 * Wait for current invocation to finish. The tasklet
551*4882a593Smuzhiyun 		 * won't be scheduled anymore afterwards since we disabled
552*4882a593Smuzhiyun 		 * the TBTT and PRE TBTT timer.
553*4882a593Smuzhiyun 		 */
554*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->tbtt_tasklet);
555*4882a593Smuzhiyun 		tasklet_kill(&rt2x00dev->pretbtt_tasklet);
556*4882a593Smuzhiyun 
557*4882a593Smuzhiyun 		break;
558*4882a593Smuzhiyun 	default:
559*4882a593Smuzhiyun 		break;
560*4882a593Smuzhiyun 	}
561*4882a593Smuzhiyun }
562*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_stop_queue);
563*4882a593Smuzhiyun 
rt2800mmio_queue_init(struct data_queue * queue)564*4882a593Smuzhiyun void rt2800mmio_queue_init(struct data_queue *queue)
565*4882a593Smuzhiyun {
566*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
567*4882a593Smuzhiyun 	unsigned short txwi_size, rxwi_size;
568*4882a593Smuzhiyun 
569*4882a593Smuzhiyun 	rt2800_get_txwi_rxwi_size(rt2x00dev, &txwi_size, &rxwi_size);
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun 	switch (queue->qid) {
572*4882a593Smuzhiyun 	case QID_RX:
573*4882a593Smuzhiyun 		queue->limit = 128;
574*4882a593Smuzhiyun 		queue->data_size = AGGREGATION_SIZE;
575*4882a593Smuzhiyun 		queue->desc_size = RXD_DESC_SIZE;
576*4882a593Smuzhiyun 		queue->winfo_size = rxwi_size;
577*4882a593Smuzhiyun 		queue->priv_size = sizeof(struct queue_entry_priv_mmio);
578*4882a593Smuzhiyun 		break;
579*4882a593Smuzhiyun 
580*4882a593Smuzhiyun 	case QID_AC_VO:
581*4882a593Smuzhiyun 	case QID_AC_VI:
582*4882a593Smuzhiyun 	case QID_AC_BE:
583*4882a593Smuzhiyun 	case QID_AC_BK:
584*4882a593Smuzhiyun 		queue->limit = 64;
585*4882a593Smuzhiyun 		queue->data_size = AGGREGATION_SIZE;
586*4882a593Smuzhiyun 		queue->desc_size = TXD_DESC_SIZE;
587*4882a593Smuzhiyun 		queue->winfo_size = txwi_size;
588*4882a593Smuzhiyun 		queue->priv_size = sizeof(struct queue_entry_priv_mmio);
589*4882a593Smuzhiyun 		break;
590*4882a593Smuzhiyun 
591*4882a593Smuzhiyun 	case QID_BEACON:
592*4882a593Smuzhiyun 		queue->limit = 8;
593*4882a593Smuzhiyun 		queue->data_size = 0; /* No DMA required for beacons */
594*4882a593Smuzhiyun 		queue->desc_size = TXD_DESC_SIZE;
595*4882a593Smuzhiyun 		queue->winfo_size = txwi_size;
596*4882a593Smuzhiyun 		queue->priv_size = sizeof(struct queue_entry_priv_mmio);
597*4882a593Smuzhiyun 		break;
598*4882a593Smuzhiyun 
599*4882a593Smuzhiyun 	case QID_ATIM:
600*4882a593Smuzhiyun 	default:
601*4882a593Smuzhiyun 		BUG();
602*4882a593Smuzhiyun 		break;
603*4882a593Smuzhiyun 	}
604*4882a593Smuzhiyun }
605*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_queue_init);
606*4882a593Smuzhiyun 
607*4882a593Smuzhiyun /*
608*4882a593Smuzhiyun  * Initialization functions.
609*4882a593Smuzhiyun  */
rt2800mmio_get_entry_state(struct queue_entry * entry)610*4882a593Smuzhiyun bool rt2800mmio_get_entry_state(struct queue_entry *entry)
611*4882a593Smuzhiyun {
612*4882a593Smuzhiyun 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
613*4882a593Smuzhiyun 	u32 word;
614*4882a593Smuzhiyun 
615*4882a593Smuzhiyun 	if (entry->queue->qid == QID_RX) {
616*4882a593Smuzhiyun 		word = rt2x00_desc_read(entry_priv->desc, 1);
617*4882a593Smuzhiyun 
618*4882a593Smuzhiyun 		return (!rt2x00_get_field32(word, RXD_W1_DMA_DONE));
619*4882a593Smuzhiyun 	} else {
620*4882a593Smuzhiyun 		word = rt2x00_desc_read(entry_priv->desc, 1);
621*4882a593Smuzhiyun 
622*4882a593Smuzhiyun 		return (!rt2x00_get_field32(word, TXD_W1_DMA_DONE));
623*4882a593Smuzhiyun 	}
624*4882a593Smuzhiyun }
625*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_get_entry_state);
626*4882a593Smuzhiyun 
rt2800mmio_clear_entry(struct queue_entry * entry)627*4882a593Smuzhiyun void rt2800mmio_clear_entry(struct queue_entry *entry)
628*4882a593Smuzhiyun {
629*4882a593Smuzhiyun 	struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
630*4882a593Smuzhiyun 	struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
631*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
632*4882a593Smuzhiyun 	u32 word;
633*4882a593Smuzhiyun 
634*4882a593Smuzhiyun 	if (entry->queue->qid == QID_RX) {
635*4882a593Smuzhiyun 		word = rt2x00_desc_read(entry_priv->desc, 0);
636*4882a593Smuzhiyun 		rt2x00_set_field32(&word, RXD_W0_SDP0, skbdesc->skb_dma);
637*4882a593Smuzhiyun 		rt2x00_desc_write(entry_priv->desc, 0, word);
638*4882a593Smuzhiyun 
639*4882a593Smuzhiyun 		word = rt2x00_desc_read(entry_priv->desc, 1);
640*4882a593Smuzhiyun 		rt2x00_set_field32(&word, RXD_W1_DMA_DONE, 0);
641*4882a593Smuzhiyun 		rt2x00_desc_write(entry_priv->desc, 1, word);
642*4882a593Smuzhiyun 
643*4882a593Smuzhiyun 		/*
644*4882a593Smuzhiyun 		 * Set RX IDX in register to inform hardware that we have
645*4882a593Smuzhiyun 		 * handled this entry and it is available for reuse again.
646*4882a593Smuzhiyun 		 */
647*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, RX_CRX_IDX,
648*4882a593Smuzhiyun 					  entry->entry_idx);
649*4882a593Smuzhiyun 	} else {
650*4882a593Smuzhiyun 		word = rt2x00_desc_read(entry_priv->desc, 1);
651*4882a593Smuzhiyun 		rt2x00_set_field32(&word, TXD_W1_DMA_DONE, 1);
652*4882a593Smuzhiyun 		rt2x00_desc_write(entry_priv->desc, 1, word);
653*4882a593Smuzhiyun 
654*4882a593Smuzhiyun 		/* If last entry stop txstatus timer */
655*4882a593Smuzhiyun 		if (entry->queue->length == 1)
656*4882a593Smuzhiyun 			hrtimer_cancel(&rt2x00dev->txstatus_timer);
657*4882a593Smuzhiyun 	}
658*4882a593Smuzhiyun }
659*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_clear_entry);
660*4882a593Smuzhiyun 
rt2800mmio_init_queues(struct rt2x00_dev * rt2x00dev)661*4882a593Smuzhiyun int rt2800mmio_init_queues(struct rt2x00_dev *rt2x00dev)
662*4882a593Smuzhiyun {
663*4882a593Smuzhiyun 	struct queue_entry_priv_mmio *entry_priv;
664*4882a593Smuzhiyun 
665*4882a593Smuzhiyun 	/*
666*4882a593Smuzhiyun 	 * Initialize registers.
667*4882a593Smuzhiyun 	 */
668*4882a593Smuzhiyun 	entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
669*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR0,
670*4882a593Smuzhiyun 				  entry_priv->desc_dma);
671*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT0,
672*4882a593Smuzhiyun 				  rt2x00dev->tx[0].limit);
673*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX0, 0);
674*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX0, 0);
675*4882a593Smuzhiyun 
676*4882a593Smuzhiyun 	entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
677*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR1,
678*4882a593Smuzhiyun 				  entry_priv->desc_dma);
679*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT1,
680*4882a593Smuzhiyun 				  rt2x00dev->tx[1].limit);
681*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX1, 0);
682*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX1, 0);
683*4882a593Smuzhiyun 
684*4882a593Smuzhiyun 	entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
685*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR2,
686*4882a593Smuzhiyun 				  entry_priv->desc_dma);
687*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT2,
688*4882a593Smuzhiyun 				  rt2x00dev->tx[2].limit);
689*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX2, 0);
690*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX2, 0);
691*4882a593Smuzhiyun 
692*4882a593Smuzhiyun 	entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
693*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR3,
694*4882a593Smuzhiyun 				  entry_priv->desc_dma);
695*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT3,
696*4882a593Smuzhiyun 				  rt2x00dev->tx[3].limit);
697*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX3, 0);
698*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX3, 0);
699*4882a593Smuzhiyun 
700*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR4, 0);
701*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT4, 0);
702*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX4, 0);
703*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX4, 0);
704*4882a593Smuzhiyun 
705*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR5, 0);
706*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT5, 0);
707*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX5, 0);
708*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX5, 0);
709*4882a593Smuzhiyun 
710*4882a593Smuzhiyun 	entry_priv = rt2x00dev->rx->entries[0].priv_data;
711*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, RX_BASE_PTR,
712*4882a593Smuzhiyun 				  entry_priv->desc_dma);
713*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, RX_MAX_CNT,
714*4882a593Smuzhiyun 				  rt2x00dev->rx[0].limit);
715*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, RX_CRX_IDX,
716*4882a593Smuzhiyun 				  rt2x00dev->rx[0].limit - 1);
717*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, RX_DRX_IDX, 0);
718*4882a593Smuzhiyun 
719*4882a593Smuzhiyun 	rt2800_disable_wpdma(rt2x00dev);
720*4882a593Smuzhiyun 
721*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, DELAY_INT_CFG, 0);
722*4882a593Smuzhiyun 
723*4882a593Smuzhiyun 	return 0;
724*4882a593Smuzhiyun }
725*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_init_queues);
726*4882a593Smuzhiyun 
rt2800mmio_init_registers(struct rt2x00_dev * rt2x00dev)727*4882a593Smuzhiyun int rt2800mmio_init_registers(struct rt2x00_dev *rt2x00dev)
728*4882a593Smuzhiyun {
729*4882a593Smuzhiyun 	u32 reg;
730*4882a593Smuzhiyun 
731*4882a593Smuzhiyun 	/*
732*4882a593Smuzhiyun 	 * Reset DMA indexes
733*4882a593Smuzhiyun 	 */
734*4882a593Smuzhiyun 	reg = rt2x00mmio_register_read(rt2x00dev, WPDMA_RST_IDX);
735*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX0, 1);
736*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX1, 1);
737*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX2, 1);
738*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX3, 1);
739*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX4, 1);
740*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX5, 1);
741*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, WPDMA_RST_IDX_DRX_IDX0, 1);
742*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, WPDMA_RST_IDX, reg);
743*4882a593Smuzhiyun 
744*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, PBF_SYS_CTRL, 0x00000e1f);
745*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, PBF_SYS_CTRL, 0x00000e00);
746*4882a593Smuzhiyun 
747*4882a593Smuzhiyun 	if (rt2x00_is_pcie(rt2x00dev) &&
748*4882a593Smuzhiyun 	    (rt2x00_rt(rt2x00dev, RT3090) ||
749*4882a593Smuzhiyun 	     rt2x00_rt(rt2x00dev, RT3390) ||
750*4882a593Smuzhiyun 	     rt2x00_rt(rt2x00dev, RT3572) ||
751*4882a593Smuzhiyun 	     rt2x00_rt(rt2x00dev, RT3593) ||
752*4882a593Smuzhiyun 	     rt2x00_rt(rt2x00dev, RT5390) ||
753*4882a593Smuzhiyun 	     rt2x00_rt(rt2x00dev, RT5392) ||
754*4882a593Smuzhiyun 	     rt2x00_rt(rt2x00dev, RT5592))) {
755*4882a593Smuzhiyun 		reg = rt2x00mmio_register_read(rt2x00dev, AUX_CTRL);
756*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, AUX_CTRL_FORCE_PCIE_CLK, 1);
757*4882a593Smuzhiyun 		rt2x00_set_field32(&reg, AUX_CTRL_WAKE_PCIE_EN, 1);
758*4882a593Smuzhiyun 		rt2x00mmio_register_write(rt2x00dev, AUX_CTRL, reg);
759*4882a593Smuzhiyun 	}
760*4882a593Smuzhiyun 
761*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, PWR_PIN_CFG, 0x00000003);
762*4882a593Smuzhiyun 
763*4882a593Smuzhiyun 	reg = 0;
764*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, MAC_SYS_CTRL_RESET_CSR, 1);
765*4882a593Smuzhiyun 	rt2x00_set_field32(&reg, MAC_SYS_CTRL_RESET_BBP, 1);
766*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
767*4882a593Smuzhiyun 
768*4882a593Smuzhiyun 	rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, 0x00000000);
769*4882a593Smuzhiyun 
770*4882a593Smuzhiyun 	return 0;
771*4882a593Smuzhiyun }
772*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_init_registers);
773*4882a593Smuzhiyun 
774*4882a593Smuzhiyun /*
775*4882a593Smuzhiyun  * Device state switch handlers.
776*4882a593Smuzhiyun  */
rt2800mmio_enable_radio(struct rt2x00_dev * rt2x00dev)777*4882a593Smuzhiyun int rt2800mmio_enable_radio(struct rt2x00_dev *rt2x00dev)
778*4882a593Smuzhiyun {
779*4882a593Smuzhiyun 	/* Wait for DMA, ignore error until we initialize queues. */
780*4882a593Smuzhiyun 	rt2800_wait_wpdma_ready(rt2x00dev);
781*4882a593Smuzhiyun 
782*4882a593Smuzhiyun 	if (unlikely(rt2800mmio_init_queues(rt2x00dev)))
783*4882a593Smuzhiyun 		return -EIO;
784*4882a593Smuzhiyun 
785*4882a593Smuzhiyun 	return rt2800_enable_radio(rt2x00dev);
786*4882a593Smuzhiyun }
787*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_enable_radio);
788*4882a593Smuzhiyun 
rt2800mmio_work_txdone(struct work_struct * work)789*4882a593Smuzhiyun static void rt2800mmio_work_txdone(struct work_struct *work)
790*4882a593Smuzhiyun {
791*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev =
792*4882a593Smuzhiyun 	    container_of(work, struct rt2x00_dev, txdone_work);
793*4882a593Smuzhiyun 
794*4882a593Smuzhiyun 	if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
795*4882a593Smuzhiyun 		return;
796*4882a593Smuzhiyun 
797*4882a593Smuzhiyun 	while (!kfifo_is_empty(&rt2x00dev->txstatus_fifo) ||
798*4882a593Smuzhiyun 	       rt2800_txstatus_timeout(rt2x00dev)) {
799*4882a593Smuzhiyun 
800*4882a593Smuzhiyun 		tasklet_disable(&rt2x00dev->txstatus_tasklet);
801*4882a593Smuzhiyun 		rt2800_txdone(rt2x00dev, UINT_MAX);
802*4882a593Smuzhiyun 		rt2800_txdone_nostatus(rt2x00dev);
803*4882a593Smuzhiyun 		tasklet_enable(&rt2x00dev->txstatus_tasklet);
804*4882a593Smuzhiyun 	}
805*4882a593Smuzhiyun 
806*4882a593Smuzhiyun 	if (rt2800_txstatus_pending(rt2x00dev))
807*4882a593Smuzhiyun 		hrtimer_start(&rt2x00dev->txstatus_timer,
808*4882a593Smuzhiyun 			      TXSTATUS_TIMEOUT, HRTIMER_MODE_REL);
809*4882a593Smuzhiyun }
810*4882a593Smuzhiyun 
rt2800mmio_tx_sta_fifo_timeout(struct hrtimer * timer)811*4882a593Smuzhiyun static enum hrtimer_restart rt2800mmio_tx_sta_fifo_timeout(struct hrtimer *timer)
812*4882a593Smuzhiyun {
813*4882a593Smuzhiyun 	struct rt2x00_dev *rt2x00dev =
814*4882a593Smuzhiyun 	    container_of(timer, struct rt2x00_dev, txstatus_timer);
815*4882a593Smuzhiyun 
816*4882a593Smuzhiyun 	if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
817*4882a593Smuzhiyun 		goto out;
818*4882a593Smuzhiyun 
819*4882a593Smuzhiyun 	if (!rt2800_txstatus_pending(rt2x00dev))
820*4882a593Smuzhiyun 		goto out;
821*4882a593Smuzhiyun 
822*4882a593Smuzhiyun 	rt2800mmio_fetch_txstatus(rt2x00dev);
823*4882a593Smuzhiyun 	if (!kfifo_is_empty(&rt2x00dev->txstatus_fifo))
824*4882a593Smuzhiyun 		tasklet_schedule(&rt2x00dev->txstatus_tasklet);
825*4882a593Smuzhiyun 	else
826*4882a593Smuzhiyun 		queue_work(rt2x00dev->workqueue, &rt2x00dev->txdone_work);
827*4882a593Smuzhiyun out:
828*4882a593Smuzhiyun 	return HRTIMER_NORESTART;
829*4882a593Smuzhiyun }
830*4882a593Smuzhiyun 
rt2800mmio_probe_hw(struct rt2x00_dev * rt2x00dev)831*4882a593Smuzhiyun int rt2800mmio_probe_hw(struct rt2x00_dev *rt2x00dev)
832*4882a593Smuzhiyun {
833*4882a593Smuzhiyun 	int retval;
834*4882a593Smuzhiyun 
835*4882a593Smuzhiyun 	retval = rt2800_probe_hw(rt2x00dev);
836*4882a593Smuzhiyun 	if (retval)
837*4882a593Smuzhiyun 		return retval;
838*4882a593Smuzhiyun 
839*4882a593Smuzhiyun 	/*
840*4882a593Smuzhiyun 	 * Set txstatus timer function.
841*4882a593Smuzhiyun 	 */
842*4882a593Smuzhiyun 	rt2x00dev->txstatus_timer.function = rt2800mmio_tx_sta_fifo_timeout;
843*4882a593Smuzhiyun 
844*4882a593Smuzhiyun 	/*
845*4882a593Smuzhiyun 	 * Overwrite TX done handler
846*4882a593Smuzhiyun 	 */
847*4882a593Smuzhiyun 	INIT_WORK(&rt2x00dev->txdone_work, rt2800mmio_work_txdone);
848*4882a593Smuzhiyun 
849*4882a593Smuzhiyun 	return 0;
850*4882a593Smuzhiyun }
851*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(rt2800mmio_probe_hw);
852*4882a593Smuzhiyun 
853*4882a593Smuzhiyun MODULE_AUTHOR(DRV_PROJECT);
854*4882a593Smuzhiyun MODULE_VERSION(DRV_VERSION);
855*4882a593Smuzhiyun MODULE_DESCRIPTION("rt2800 MMIO library");
856*4882a593Smuzhiyun MODULE_LICENSE("GPL");
857