xref: /OK3568_Linux_fs/kernel/drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: (GPL-2.0 OR MIT)
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
4*4882a593Smuzhiyun  * stmmac XGMAC support.
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/iopoll.h>
8*4882a593Smuzhiyun #include "stmmac.h"
9*4882a593Smuzhiyun #include "dwxgmac2.h"
10*4882a593Smuzhiyun 
dwxgmac2_dma_reset(void __iomem * ioaddr)11*4882a593Smuzhiyun static int dwxgmac2_dma_reset(void __iomem *ioaddr)
12*4882a593Smuzhiyun {
13*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_MODE);
14*4882a593Smuzhiyun 
15*4882a593Smuzhiyun 	/* DMA SW reset */
16*4882a593Smuzhiyun 	writel(value | XGMAC_SWR, ioaddr + XGMAC_DMA_MODE);
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun 	return readl_poll_timeout(ioaddr + XGMAC_DMA_MODE, value,
19*4882a593Smuzhiyun 				  !(value & XGMAC_SWR), 0, 100000);
20*4882a593Smuzhiyun }
21*4882a593Smuzhiyun 
dwxgmac2_dma_init(void __iomem * ioaddr,struct stmmac_dma_cfg * dma_cfg,int atds)22*4882a593Smuzhiyun static void dwxgmac2_dma_init(void __iomem *ioaddr,
23*4882a593Smuzhiyun 			      struct stmmac_dma_cfg *dma_cfg, int atds)
24*4882a593Smuzhiyun {
25*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_SYSBUS_MODE);
26*4882a593Smuzhiyun 
27*4882a593Smuzhiyun 	if (dma_cfg->aal)
28*4882a593Smuzhiyun 		value |= XGMAC_AAL;
29*4882a593Smuzhiyun 
30*4882a593Smuzhiyun 	if (dma_cfg->eame)
31*4882a593Smuzhiyun 		value |= XGMAC_EAME;
32*4882a593Smuzhiyun 
33*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_SYSBUS_MODE);
34*4882a593Smuzhiyun }
35*4882a593Smuzhiyun 
dwxgmac2_dma_init_chan(void __iomem * ioaddr,struct stmmac_dma_cfg * dma_cfg,u32 chan)36*4882a593Smuzhiyun static void dwxgmac2_dma_init_chan(void __iomem *ioaddr,
37*4882a593Smuzhiyun 				   struct stmmac_dma_cfg *dma_cfg, u32 chan)
38*4882a593Smuzhiyun {
39*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_CH_CONTROL(chan));
40*4882a593Smuzhiyun 
41*4882a593Smuzhiyun 	if (dma_cfg->pblx8)
42*4882a593Smuzhiyun 		value |= XGMAC_PBLx8;
43*4882a593Smuzhiyun 
44*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_CONTROL(chan));
45*4882a593Smuzhiyun 	writel(XGMAC_DMA_INT_DEFAULT_EN, ioaddr + XGMAC_DMA_CH_INT_EN(chan));
46*4882a593Smuzhiyun }
47*4882a593Smuzhiyun 
dwxgmac2_dma_init_rx_chan(void __iomem * ioaddr,struct stmmac_dma_cfg * dma_cfg,dma_addr_t phy,u32 chan)48*4882a593Smuzhiyun static void dwxgmac2_dma_init_rx_chan(void __iomem *ioaddr,
49*4882a593Smuzhiyun 				      struct stmmac_dma_cfg *dma_cfg,
50*4882a593Smuzhiyun 				      dma_addr_t phy, u32 chan)
51*4882a593Smuzhiyun {
52*4882a593Smuzhiyun 	u32 rxpbl = dma_cfg->rxpbl ?: dma_cfg->pbl;
53*4882a593Smuzhiyun 	u32 value;
54*4882a593Smuzhiyun 
55*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
56*4882a593Smuzhiyun 	value &= ~XGMAC_RxPBL;
57*4882a593Smuzhiyun 	value |= (rxpbl << XGMAC_RxPBL_SHIFT) & XGMAC_RxPBL;
58*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
59*4882a593Smuzhiyun 
60*4882a593Smuzhiyun 	writel(upper_32_bits(phy), ioaddr + XGMAC_DMA_CH_RxDESC_HADDR(chan));
61*4882a593Smuzhiyun 	writel(lower_32_bits(phy), ioaddr + XGMAC_DMA_CH_RxDESC_LADDR(chan));
62*4882a593Smuzhiyun }
63*4882a593Smuzhiyun 
dwxgmac2_dma_init_tx_chan(void __iomem * ioaddr,struct stmmac_dma_cfg * dma_cfg,dma_addr_t phy,u32 chan)64*4882a593Smuzhiyun static void dwxgmac2_dma_init_tx_chan(void __iomem *ioaddr,
65*4882a593Smuzhiyun 				      struct stmmac_dma_cfg *dma_cfg,
66*4882a593Smuzhiyun 				      dma_addr_t phy, u32 chan)
67*4882a593Smuzhiyun {
68*4882a593Smuzhiyun 	u32 txpbl = dma_cfg->txpbl ?: dma_cfg->pbl;
69*4882a593Smuzhiyun 	u32 value;
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
72*4882a593Smuzhiyun 	value &= ~XGMAC_TxPBL;
73*4882a593Smuzhiyun 	value |= (txpbl << XGMAC_TxPBL_SHIFT) & XGMAC_TxPBL;
74*4882a593Smuzhiyun 	value |= XGMAC_OSP;
75*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
76*4882a593Smuzhiyun 
77*4882a593Smuzhiyun 	writel(upper_32_bits(phy), ioaddr + XGMAC_DMA_CH_TxDESC_HADDR(chan));
78*4882a593Smuzhiyun 	writel(lower_32_bits(phy), ioaddr + XGMAC_DMA_CH_TxDESC_LADDR(chan));
79*4882a593Smuzhiyun }
80*4882a593Smuzhiyun 
dwxgmac2_dma_axi(void __iomem * ioaddr,struct stmmac_axi * axi)81*4882a593Smuzhiyun static void dwxgmac2_dma_axi(void __iomem *ioaddr, struct stmmac_axi *axi)
82*4882a593Smuzhiyun {
83*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_SYSBUS_MODE);
84*4882a593Smuzhiyun 	int i;
85*4882a593Smuzhiyun 
86*4882a593Smuzhiyun 	if (axi->axi_lpi_en)
87*4882a593Smuzhiyun 		value |= XGMAC_EN_LPI;
88*4882a593Smuzhiyun 	if (axi->axi_xit_frm)
89*4882a593Smuzhiyun 		value |= XGMAC_LPI_XIT_PKT;
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 	value &= ~XGMAC_WR_OSR_LMT;
92*4882a593Smuzhiyun 	value |= (axi->axi_wr_osr_lmt << XGMAC_WR_OSR_LMT_SHIFT) &
93*4882a593Smuzhiyun 		XGMAC_WR_OSR_LMT;
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun 	value &= ~XGMAC_RD_OSR_LMT;
96*4882a593Smuzhiyun 	value |= (axi->axi_rd_osr_lmt << XGMAC_RD_OSR_LMT_SHIFT) &
97*4882a593Smuzhiyun 		XGMAC_RD_OSR_LMT;
98*4882a593Smuzhiyun 
99*4882a593Smuzhiyun 	if (!axi->axi_fb)
100*4882a593Smuzhiyun 		value |= XGMAC_UNDEF;
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun 	value &= ~XGMAC_BLEN;
103*4882a593Smuzhiyun 	for (i = 0; i < AXI_BLEN; i++) {
104*4882a593Smuzhiyun 		switch (axi->axi_blen[i]) {
105*4882a593Smuzhiyun 		case 256:
106*4882a593Smuzhiyun 			value |= XGMAC_BLEN256;
107*4882a593Smuzhiyun 			break;
108*4882a593Smuzhiyun 		case 128:
109*4882a593Smuzhiyun 			value |= XGMAC_BLEN128;
110*4882a593Smuzhiyun 			break;
111*4882a593Smuzhiyun 		case 64:
112*4882a593Smuzhiyun 			value |= XGMAC_BLEN64;
113*4882a593Smuzhiyun 			break;
114*4882a593Smuzhiyun 		case 32:
115*4882a593Smuzhiyun 			value |= XGMAC_BLEN32;
116*4882a593Smuzhiyun 			break;
117*4882a593Smuzhiyun 		case 16:
118*4882a593Smuzhiyun 			value |= XGMAC_BLEN16;
119*4882a593Smuzhiyun 			break;
120*4882a593Smuzhiyun 		case 8:
121*4882a593Smuzhiyun 			value |= XGMAC_BLEN8;
122*4882a593Smuzhiyun 			break;
123*4882a593Smuzhiyun 		case 4:
124*4882a593Smuzhiyun 			value |= XGMAC_BLEN4;
125*4882a593Smuzhiyun 			break;
126*4882a593Smuzhiyun 		}
127*4882a593Smuzhiyun 	}
128*4882a593Smuzhiyun 
129*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_SYSBUS_MODE);
130*4882a593Smuzhiyun 	writel(XGMAC_TDPS, ioaddr + XGMAC_TX_EDMA_CTRL);
131*4882a593Smuzhiyun 	writel(XGMAC_RDPS, ioaddr + XGMAC_RX_EDMA_CTRL);
132*4882a593Smuzhiyun }
133*4882a593Smuzhiyun 
dwxgmac2_dma_dump_regs(void __iomem * ioaddr,u32 * reg_space)134*4882a593Smuzhiyun static void dwxgmac2_dma_dump_regs(void __iomem *ioaddr, u32 *reg_space)
135*4882a593Smuzhiyun {
136*4882a593Smuzhiyun 	int i;
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun 	for (i = (XGMAC_DMA_MODE / 4); i < XGMAC_REGSIZE; i++)
139*4882a593Smuzhiyun 		reg_space[i] = readl(ioaddr + i * 4);
140*4882a593Smuzhiyun }
141*4882a593Smuzhiyun 
dwxgmac2_dma_rx_mode(void __iomem * ioaddr,int mode,u32 channel,int fifosz,u8 qmode)142*4882a593Smuzhiyun static void dwxgmac2_dma_rx_mode(void __iomem *ioaddr, int mode,
143*4882a593Smuzhiyun 				 u32 channel, int fifosz, u8 qmode)
144*4882a593Smuzhiyun {
145*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_MTL_RXQ_OPMODE(channel));
146*4882a593Smuzhiyun 	unsigned int rqs = fifosz / 256 - 1;
147*4882a593Smuzhiyun 
148*4882a593Smuzhiyun 	if (mode == SF_DMA_MODE) {
149*4882a593Smuzhiyun 		value |= XGMAC_RSF;
150*4882a593Smuzhiyun 	} else {
151*4882a593Smuzhiyun 		value &= ~XGMAC_RSF;
152*4882a593Smuzhiyun 		value &= ~XGMAC_RTC;
153*4882a593Smuzhiyun 
154*4882a593Smuzhiyun 		if (mode <= 64)
155*4882a593Smuzhiyun 			value |= 0x0 << XGMAC_RTC_SHIFT;
156*4882a593Smuzhiyun 		else if (mode <= 96)
157*4882a593Smuzhiyun 			value |= 0x2 << XGMAC_RTC_SHIFT;
158*4882a593Smuzhiyun 		else
159*4882a593Smuzhiyun 			value |= 0x3 << XGMAC_RTC_SHIFT;
160*4882a593Smuzhiyun 	}
161*4882a593Smuzhiyun 
162*4882a593Smuzhiyun 	value &= ~XGMAC_RQS;
163*4882a593Smuzhiyun 	value |= (rqs << XGMAC_RQS_SHIFT) & XGMAC_RQS;
164*4882a593Smuzhiyun 
165*4882a593Smuzhiyun 	if ((fifosz >= 4096) && (qmode != MTL_QUEUE_AVB)) {
166*4882a593Smuzhiyun 		u32 flow = readl(ioaddr + XGMAC_MTL_RXQ_FLOW_CONTROL(channel));
167*4882a593Smuzhiyun 		unsigned int rfd, rfa;
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 		value |= XGMAC_EHFC;
170*4882a593Smuzhiyun 
171*4882a593Smuzhiyun 		/* Set Threshold for Activating Flow Control to min 2 frames,
172*4882a593Smuzhiyun 		 * i.e. 1500 * 2 = 3000 bytes.
173*4882a593Smuzhiyun 		 *
174*4882a593Smuzhiyun 		 * Set Threshold for Deactivating Flow Control to min 1 frame,
175*4882a593Smuzhiyun 		 * i.e. 1500 bytes.
176*4882a593Smuzhiyun 		 */
177*4882a593Smuzhiyun 		switch (fifosz) {
178*4882a593Smuzhiyun 		case 4096:
179*4882a593Smuzhiyun 			/* This violates the above formula because of FIFO size
180*4882a593Smuzhiyun 			 * limit therefore overflow may occur in spite of this.
181*4882a593Smuzhiyun 			 */
182*4882a593Smuzhiyun 			rfd = 0x03; /* Full-2.5K */
183*4882a593Smuzhiyun 			rfa = 0x01; /* Full-1.5K */
184*4882a593Smuzhiyun 			break;
185*4882a593Smuzhiyun 
186*4882a593Smuzhiyun 		default:
187*4882a593Smuzhiyun 			rfd = 0x07; /* Full-4.5K */
188*4882a593Smuzhiyun 			rfa = 0x04; /* Full-3K */
189*4882a593Smuzhiyun 			break;
190*4882a593Smuzhiyun 		}
191*4882a593Smuzhiyun 
192*4882a593Smuzhiyun 		flow &= ~XGMAC_RFD;
193*4882a593Smuzhiyun 		flow |= rfd << XGMAC_RFD_SHIFT;
194*4882a593Smuzhiyun 
195*4882a593Smuzhiyun 		flow &= ~XGMAC_RFA;
196*4882a593Smuzhiyun 		flow |= rfa << XGMAC_RFA_SHIFT;
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun 		writel(flow, ioaddr + XGMAC_MTL_RXQ_FLOW_CONTROL(channel));
199*4882a593Smuzhiyun 	}
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_MTL_RXQ_OPMODE(channel));
202*4882a593Smuzhiyun 
203*4882a593Smuzhiyun 	/* Enable MTL RX overflow */
204*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_MTL_QINTEN(channel));
205*4882a593Smuzhiyun 	writel(value | XGMAC_RXOIE, ioaddr + XGMAC_MTL_QINTEN(channel));
206*4882a593Smuzhiyun }
207*4882a593Smuzhiyun 
dwxgmac2_dma_tx_mode(void __iomem * ioaddr,int mode,u32 channel,int fifosz,u8 qmode)208*4882a593Smuzhiyun static void dwxgmac2_dma_tx_mode(void __iomem *ioaddr, int mode,
209*4882a593Smuzhiyun 				 u32 channel, int fifosz, u8 qmode)
210*4882a593Smuzhiyun {
211*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_MTL_TXQ_OPMODE(channel));
212*4882a593Smuzhiyun 	unsigned int tqs = fifosz / 256 - 1;
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 	if (mode == SF_DMA_MODE) {
215*4882a593Smuzhiyun 		value |= XGMAC_TSF;
216*4882a593Smuzhiyun 	} else {
217*4882a593Smuzhiyun 		value &= ~XGMAC_TSF;
218*4882a593Smuzhiyun 		value &= ~XGMAC_TTC;
219*4882a593Smuzhiyun 
220*4882a593Smuzhiyun 		if (mode <= 64)
221*4882a593Smuzhiyun 			value |= 0x0 << XGMAC_TTC_SHIFT;
222*4882a593Smuzhiyun 		else if (mode <= 96)
223*4882a593Smuzhiyun 			value |= 0x2 << XGMAC_TTC_SHIFT;
224*4882a593Smuzhiyun 		else if (mode <= 128)
225*4882a593Smuzhiyun 			value |= 0x3 << XGMAC_TTC_SHIFT;
226*4882a593Smuzhiyun 		else if (mode <= 192)
227*4882a593Smuzhiyun 			value |= 0x4 << XGMAC_TTC_SHIFT;
228*4882a593Smuzhiyun 		else if (mode <= 256)
229*4882a593Smuzhiyun 			value |= 0x5 << XGMAC_TTC_SHIFT;
230*4882a593Smuzhiyun 		else if (mode <= 384)
231*4882a593Smuzhiyun 			value |= 0x6 << XGMAC_TTC_SHIFT;
232*4882a593Smuzhiyun 		else
233*4882a593Smuzhiyun 			value |= 0x7 << XGMAC_TTC_SHIFT;
234*4882a593Smuzhiyun 	}
235*4882a593Smuzhiyun 
236*4882a593Smuzhiyun 	/* Use static TC to Queue mapping */
237*4882a593Smuzhiyun 	value |= (channel << XGMAC_Q2TCMAP_SHIFT) & XGMAC_Q2TCMAP;
238*4882a593Smuzhiyun 
239*4882a593Smuzhiyun 	value &= ~XGMAC_TXQEN;
240*4882a593Smuzhiyun 	if (qmode != MTL_QUEUE_AVB)
241*4882a593Smuzhiyun 		value |= 0x2 << XGMAC_TXQEN_SHIFT;
242*4882a593Smuzhiyun 	else
243*4882a593Smuzhiyun 		value |= 0x1 << XGMAC_TXQEN_SHIFT;
244*4882a593Smuzhiyun 
245*4882a593Smuzhiyun 	value &= ~XGMAC_TQS;
246*4882a593Smuzhiyun 	value |= (tqs << XGMAC_TQS_SHIFT) & XGMAC_TQS;
247*4882a593Smuzhiyun 
248*4882a593Smuzhiyun 	writel(value, ioaddr +  XGMAC_MTL_TXQ_OPMODE(channel));
249*4882a593Smuzhiyun }
250*4882a593Smuzhiyun 
dwxgmac2_enable_dma_irq(void __iomem * ioaddr,u32 chan,bool rx,bool tx)251*4882a593Smuzhiyun static void dwxgmac2_enable_dma_irq(void __iomem *ioaddr, u32 chan,
252*4882a593Smuzhiyun 				    bool rx, bool tx)
253*4882a593Smuzhiyun {
254*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan));
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun 	if (rx)
257*4882a593Smuzhiyun 		value |= XGMAC_DMA_INT_DEFAULT_RX;
258*4882a593Smuzhiyun 	if (tx)
259*4882a593Smuzhiyun 		value |= XGMAC_DMA_INT_DEFAULT_TX;
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_INT_EN(chan));
262*4882a593Smuzhiyun }
263*4882a593Smuzhiyun 
dwxgmac2_disable_dma_irq(void __iomem * ioaddr,u32 chan,bool rx,bool tx)264*4882a593Smuzhiyun static void dwxgmac2_disable_dma_irq(void __iomem *ioaddr, u32 chan,
265*4882a593Smuzhiyun 				     bool rx, bool tx)
266*4882a593Smuzhiyun {
267*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan));
268*4882a593Smuzhiyun 
269*4882a593Smuzhiyun 	if (rx)
270*4882a593Smuzhiyun 		value &= ~XGMAC_DMA_INT_DEFAULT_RX;
271*4882a593Smuzhiyun 	if (tx)
272*4882a593Smuzhiyun 		value &= ~XGMAC_DMA_INT_DEFAULT_TX;
273*4882a593Smuzhiyun 
274*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_INT_EN(chan));
275*4882a593Smuzhiyun }
276*4882a593Smuzhiyun 
dwxgmac2_dma_start_tx(void __iomem * ioaddr,u32 chan)277*4882a593Smuzhiyun static void dwxgmac2_dma_start_tx(void __iomem *ioaddr, u32 chan)
278*4882a593Smuzhiyun {
279*4882a593Smuzhiyun 	u32 value;
280*4882a593Smuzhiyun 
281*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
282*4882a593Smuzhiyun 	value |= XGMAC_TXST;
283*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
284*4882a593Smuzhiyun 
285*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_TX_CONFIG);
286*4882a593Smuzhiyun 	value |= XGMAC_CONFIG_TE;
287*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_TX_CONFIG);
288*4882a593Smuzhiyun }
289*4882a593Smuzhiyun 
dwxgmac2_dma_stop_tx(void __iomem * ioaddr,u32 chan)290*4882a593Smuzhiyun static void dwxgmac2_dma_stop_tx(void __iomem *ioaddr, u32 chan)
291*4882a593Smuzhiyun {
292*4882a593Smuzhiyun 	u32 value;
293*4882a593Smuzhiyun 
294*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
295*4882a593Smuzhiyun 	value &= ~XGMAC_TXST;
296*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_TX_CONFIG);
299*4882a593Smuzhiyun 	value &= ~XGMAC_CONFIG_TE;
300*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_TX_CONFIG);
301*4882a593Smuzhiyun }
302*4882a593Smuzhiyun 
dwxgmac2_dma_start_rx(void __iomem * ioaddr,u32 chan)303*4882a593Smuzhiyun static void dwxgmac2_dma_start_rx(void __iomem *ioaddr, u32 chan)
304*4882a593Smuzhiyun {
305*4882a593Smuzhiyun 	u32 value;
306*4882a593Smuzhiyun 
307*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
308*4882a593Smuzhiyun 	value |= XGMAC_RXST;
309*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
310*4882a593Smuzhiyun 
311*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_RX_CONFIG);
312*4882a593Smuzhiyun 	value |= XGMAC_CONFIG_RE;
313*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_RX_CONFIG);
314*4882a593Smuzhiyun }
315*4882a593Smuzhiyun 
dwxgmac2_dma_stop_rx(void __iomem * ioaddr,u32 chan)316*4882a593Smuzhiyun static void dwxgmac2_dma_stop_rx(void __iomem *ioaddr, u32 chan)
317*4882a593Smuzhiyun {
318*4882a593Smuzhiyun 	u32 value;
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
321*4882a593Smuzhiyun 	value &= ~XGMAC_RXST;
322*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun 
dwxgmac2_dma_interrupt(void __iomem * ioaddr,struct stmmac_extra_stats * x,u32 chan)325*4882a593Smuzhiyun static int dwxgmac2_dma_interrupt(void __iomem *ioaddr,
326*4882a593Smuzhiyun 				  struct stmmac_extra_stats *x, u32 chan)
327*4882a593Smuzhiyun {
328*4882a593Smuzhiyun 	u32 intr_status = readl(ioaddr + XGMAC_DMA_CH_STATUS(chan));
329*4882a593Smuzhiyun 	u32 intr_en = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan));
330*4882a593Smuzhiyun 	int ret = 0;
331*4882a593Smuzhiyun 
332*4882a593Smuzhiyun 	/* ABNORMAL interrupts */
333*4882a593Smuzhiyun 	if (unlikely(intr_status & XGMAC_AIS)) {
334*4882a593Smuzhiyun 		if (unlikely(intr_status & XGMAC_RBU)) {
335*4882a593Smuzhiyun 			x->rx_buf_unav_irq++;
336*4882a593Smuzhiyun 			ret |= handle_rx;
337*4882a593Smuzhiyun 		}
338*4882a593Smuzhiyun 		if (unlikely(intr_status & XGMAC_TPS)) {
339*4882a593Smuzhiyun 			x->tx_process_stopped_irq++;
340*4882a593Smuzhiyun 			ret |= tx_hard_error;
341*4882a593Smuzhiyun 		}
342*4882a593Smuzhiyun 		if (unlikely(intr_status & XGMAC_FBE)) {
343*4882a593Smuzhiyun 			x->fatal_bus_error_irq++;
344*4882a593Smuzhiyun 			ret |= tx_hard_error;
345*4882a593Smuzhiyun 		}
346*4882a593Smuzhiyun 	}
347*4882a593Smuzhiyun 
348*4882a593Smuzhiyun 	/* TX/RX NORMAL interrupts */
349*4882a593Smuzhiyun 	if (likely(intr_status & XGMAC_NIS)) {
350*4882a593Smuzhiyun 		x->normal_irq_n++;
351*4882a593Smuzhiyun 
352*4882a593Smuzhiyun 		if (likely(intr_status & XGMAC_RI)) {
353*4882a593Smuzhiyun 			x->rx_normal_irq_n++;
354*4882a593Smuzhiyun 			ret |= handle_rx;
355*4882a593Smuzhiyun 		}
356*4882a593Smuzhiyun 		if (likely(intr_status & (XGMAC_TI | XGMAC_TBU))) {
357*4882a593Smuzhiyun 			x->tx_normal_irq_n++;
358*4882a593Smuzhiyun 			ret |= handle_tx;
359*4882a593Smuzhiyun 		}
360*4882a593Smuzhiyun 	}
361*4882a593Smuzhiyun 
362*4882a593Smuzhiyun 	/* Clear interrupts */
363*4882a593Smuzhiyun 	writel(intr_en & intr_status, ioaddr + XGMAC_DMA_CH_STATUS(chan));
364*4882a593Smuzhiyun 
365*4882a593Smuzhiyun 	return ret;
366*4882a593Smuzhiyun }
367*4882a593Smuzhiyun 
dwxgmac2_get_hw_feature(void __iomem * ioaddr,struct dma_features * dma_cap)368*4882a593Smuzhiyun static int dwxgmac2_get_hw_feature(void __iomem *ioaddr,
369*4882a593Smuzhiyun 				   struct dma_features *dma_cap)
370*4882a593Smuzhiyun {
371*4882a593Smuzhiyun 	u32 hw_cap;
372*4882a593Smuzhiyun 
373*4882a593Smuzhiyun 	/*  MAC HW feature 0 */
374*4882a593Smuzhiyun 	hw_cap = readl(ioaddr + XGMAC_HW_FEATURE0);
375*4882a593Smuzhiyun 	dma_cap->vlins = (hw_cap & XGMAC_HWFEAT_SAVLANINS) >> 27;
376*4882a593Smuzhiyun 	dma_cap->rx_coe = (hw_cap & XGMAC_HWFEAT_RXCOESEL) >> 16;
377*4882a593Smuzhiyun 	dma_cap->tx_coe = (hw_cap & XGMAC_HWFEAT_TXCOESEL) >> 14;
378*4882a593Smuzhiyun 	dma_cap->eee = (hw_cap & XGMAC_HWFEAT_EEESEL) >> 13;
379*4882a593Smuzhiyun 	dma_cap->atime_stamp = (hw_cap & XGMAC_HWFEAT_TSSEL) >> 12;
380*4882a593Smuzhiyun 	dma_cap->av = (hw_cap & XGMAC_HWFEAT_AVSEL) >> 11;
381*4882a593Smuzhiyun 	dma_cap->av &= !((hw_cap & XGMAC_HWFEAT_RAVSEL) >> 10);
382*4882a593Smuzhiyun 	dma_cap->arpoffsel = (hw_cap & XGMAC_HWFEAT_ARPOFFSEL) >> 9;
383*4882a593Smuzhiyun 	dma_cap->rmon = (hw_cap & XGMAC_HWFEAT_MMCSEL) >> 8;
384*4882a593Smuzhiyun 	dma_cap->pmt_magic_frame = (hw_cap & XGMAC_HWFEAT_MGKSEL) >> 7;
385*4882a593Smuzhiyun 	dma_cap->pmt_remote_wake_up = (hw_cap & XGMAC_HWFEAT_RWKSEL) >> 6;
386*4882a593Smuzhiyun 	dma_cap->vlhash = (hw_cap & XGMAC_HWFEAT_VLHASH) >> 4;
387*4882a593Smuzhiyun 	dma_cap->mbps_1000 = (hw_cap & XGMAC_HWFEAT_GMIISEL) >> 1;
388*4882a593Smuzhiyun 
389*4882a593Smuzhiyun 	/* MAC HW feature 1 */
390*4882a593Smuzhiyun 	hw_cap = readl(ioaddr + XGMAC_HW_FEATURE1);
391*4882a593Smuzhiyun 	dma_cap->l3l4fnum = (hw_cap & XGMAC_HWFEAT_L3L4FNUM) >> 27;
392*4882a593Smuzhiyun 	dma_cap->hash_tb_sz = (hw_cap & XGMAC_HWFEAT_HASHTBLSZ) >> 24;
393*4882a593Smuzhiyun 	dma_cap->rssen = (hw_cap & XGMAC_HWFEAT_RSSEN) >> 20;
394*4882a593Smuzhiyun 	dma_cap->tsoen = (hw_cap & XGMAC_HWFEAT_TSOEN) >> 18;
395*4882a593Smuzhiyun 	dma_cap->sphen = (hw_cap & XGMAC_HWFEAT_SPHEN) >> 17;
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	dma_cap->addr64 = (hw_cap & XGMAC_HWFEAT_ADDR64) >> 14;
398*4882a593Smuzhiyun 	switch (dma_cap->addr64) {
399*4882a593Smuzhiyun 	case 0:
400*4882a593Smuzhiyun 		dma_cap->addr64 = 32;
401*4882a593Smuzhiyun 		break;
402*4882a593Smuzhiyun 	case 1:
403*4882a593Smuzhiyun 		dma_cap->addr64 = 40;
404*4882a593Smuzhiyun 		break;
405*4882a593Smuzhiyun 	case 2:
406*4882a593Smuzhiyun 		dma_cap->addr64 = 48;
407*4882a593Smuzhiyun 		break;
408*4882a593Smuzhiyun 	default:
409*4882a593Smuzhiyun 		dma_cap->addr64 = 32;
410*4882a593Smuzhiyun 		break;
411*4882a593Smuzhiyun 	}
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 	dma_cap->tx_fifo_size =
414*4882a593Smuzhiyun 		128 << ((hw_cap & XGMAC_HWFEAT_TXFIFOSIZE) >> 6);
415*4882a593Smuzhiyun 	dma_cap->rx_fifo_size =
416*4882a593Smuzhiyun 		128 << ((hw_cap & XGMAC_HWFEAT_RXFIFOSIZE) >> 0);
417*4882a593Smuzhiyun 
418*4882a593Smuzhiyun 	/* MAC HW feature 2 */
419*4882a593Smuzhiyun 	hw_cap = readl(ioaddr + XGMAC_HW_FEATURE2);
420*4882a593Smuzhiyun 	dma_cap->pps_out_num = (hw_cap & XGMAC_HWFEAT_PPSOUTNUM) >> 24;
421*4882a593Smuzhiyun 	dma_cap->number_tx_channel =
422*4882a593Smuzhiyun 		((hw_cap & XGMAC_HWFEAT_TXCHCNT) >> 18) + 1;
423*4882a593Smuzhiyun 	dma_cap->number_rx_channel =
424*4882a593Smuzhiyun 		((hw_cap & XGMAC_HWFEAT_RXCHCNT) >> 12) + 1;
425*4882a593Smuzhiyun 	dma_cap->number_tx_queues =
426*4882a593Smuzhiyun 		((hw_cap & XGMAC_HWFEAT_TXQCNT) >> 6) + 1;
427*4882a593Smuzhiyun 	dma_cap->number_rx_queues =
428*4882a593Smuzhiyun 		((hw_cap & XGMAC_HWFEAT_RXQCNT) >> 0) + 1;
429*4882a593Smuzhiyun 
430*4882a593Smuzhiyun 	/* MAC HW feature 3 */
431*4882a593Smuzhiyun 	hw_cap = readl(ioaddr + XGMAC_HW_FEATURE3);
432*4882a593Smuzhiyun 	dma_cap->tbssel = (hw_cap & XGMAC_HWFEAT_TBSSEL) >> 27;
433*4882a593Smuzhiyun 	dma_cap->fpesel = (hw_cap & XGMAC_HWFEAT_FPESEL) >> 26;
434*4882a593Smuzhiyun 	dma_cap->estwid = (hw_cap & XGMAC_HWFEAT_ESTWID) >> 23;
435*4882a593Smuzhiyun 	dma_cap->estdep = (hw_cap & XGMAC_HWFEAT_ESTDEP) >> 20;
436*4882a593Smuzhiyun 	dma_cap->estsel = (hw_cap & XGMAC_HWFEAT_ESTSEL) >> 19;
437*4882a593Smuzhiyun 	dma_cap->asp = (hw_cap & XGMAC_HWFEAT_ASP) >> 14;
438*4882a593Smuzhiyun 	dma_cap->dvlan = (hw_cap & XGMAC_HWFEAT_DVLAN) >> 13;
439*4882a593Smuzhiyun 	dma_cap->frpes = (hw_cap & XGMAC_HWFEAT_FRPES) >> 11;
440*4882a593Smuzhiyun 	dma_cap->frpbs = (hw_cap & XGMAC_HWFEAT_FRPPB) >> 9;
441*4882a593Smuzhiyun 	dma_cap->frpsel = (hw_cap & XGMAC_HWFEAT_FRPSEL) >> 3;
442*4882a593Smuzhiyun 
443*4882a593Smuzhiyun 	return 0;
444*4882a593Smuzhiyun }
445*4882a593Smuzhiyun 
dwxgmac2_rx_watchdog(void __iomem * ioaddr,u32 riwt,u32 nchan)446*4882a593Smuzhiyun static void dwxgmac2_rx_watchdog(void __iomem *ioaddr, u32 riwt, u32 nchan)
447*4882a593Smuzhiyun {
448*4882a593Smuzhiyun 	u32 i;
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	for (i = 0; i < nchan; i++)
451*4882a593Smuzhiyun 		writel(riwt & XGMAC_RWT, ioaddr + XGMAC_DMA_CH_Rx_WATCHDOG(i));
452*4882a593Smuzhiyun }
453*4882a593Smuzhiyun 
dwxgmac2_set_rx_ring_len(void __iomem * ioaddr,u32 len,u32 chan)454*4882a593Smuzhiyun static void dwxgmac2_set_rx_ring_len(void __iomem *ioaddr, u32 len, u32 chan)
455*4882a593Smuzhiyun {
456*4882a593Smuzhiyun 	writel(len, ioaddr + XGMAC_DMA_CH_RxDESC_RING_LEN(chan));
457*4882a593Smuzhiyun }
458*4882a593Smuzhiyun 
dwxgmac2_set_tx_ring_len(void __iomem * ioaddr,u32 len,u32 chan)459*4882a593Smuzhiyun static void dwxgmac2_set_tx_ring_len(void __iomem *ioaddr, u32 len, u32 chan)
460*4882a593Smuzhiyun {
461*4882a593Smuzhiyun 	writel(len, ioaddr + XGMAC_DMA_CH_TxDESC_RING_LEN(chan));
462*4882a593Smuzhiyun }
463*4882a593Smuzhiyun 
dwxgmac2_set_rx_tail_ptr(void __iomem * ioaddr,u32 ptr,u32 chan)464*4882a593Smuzhiyun static void dwxgmac2_set_rx_tail_ptr(void __iomem *ioaddr, u32 ptr, u32 chan)
465*4882a593Smuzhiyun {
466*4882a593Smuzhiyun 	writel(ptr, ioaddr + XGMAC_DMA_CH_RxDESC_TAIL_LPTR(chan));
467*4882a593Smuzhiyun }
468*4882a593Smuzhiyun 
dwxgmac2_set_tx_tail_ptr(void __iomem * ioaddr,u32 ptr,u32 chan)469*4882a593Smuzhiyun static void dwxgmac2_set_tx_tail_ptr(void __iomem *ioaddr, u32 ptr, u32 chan)
470*4882a593Smuzhiyun {
471*4882a593Smuzhiyun 	writel(ptr, ioaddr + XGMAC_DMA_CH_TxDESC_TAIL_LPTR(chan));
472*4882a593Smuzhiyun }
473*4882a593Smuzhiyun 
dwxgmac2_enable_tso(void __iomem * ioaddr,bool en,u32 chan)474*4882a593Smuzhiyun static void dwxgmac2_enable_tso(void __iomem *ioaddr, bool en, u32 chan)
475*4882a593Smuzhiyun {
476*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
477*4882a593Smuzhiyun 
478*4882a593Smuzhiyun 	if (en)
479*4882a593Smuzhiyun 		value |= XGMAC_TSE;
480*4882a593Smuzhiyun 	else
481*4882a593Smuzhiyun 		value &= ~XGMAC_TSE;
482*4882a593Smuzhiyun 
483*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
484*4882a593Smuzhiyun }
485*4882a593Smuzhiyun 
dwxgmac2_qmode(void __iomem * ioaddr,u32 channel,u8 qmode)486*4882a593Smuzhiyun static void dwxgmac2_qmode(void __iomem *ioaddr, u32 channel, u8 qmode)
487*4882a593Smuzhiyun {
488*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_MTL_TXQ_OPMODE(channel));
489*4882a593Smuzhiyun 	u32 flow = readl(ioaddr + XGMAC_RX_FLOW_CTRL);
490*4882a593Smuzhiyun 
491*4882a593Smuzhiyun 	value &= ~XGMAC_TXQEN;
492*4882a593Smuzhiyun 	if (qmode != MTL_QUEUE_AVB) {
493*4882a593Smuzhiyun 		value |= 0x2 << XGMAC_TXQEN_SHIFT;
494*4882a593Smuzhiyun 		writel(0, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(channel));
495*4882a593Smuzhiyun 	} else {
496*4882a593Smuzhiyun 		value |= 0x1 << XGMAC_TXQEN_SHIFT;
497*4882a593Smuzhiyun 		writel(flow & (~XGMAC_RFE), ioaddr + XGMAC_RX_FLOW_CTRL);
498*4882a593Smuzhiyun 	}
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun 	writel(value, ioaddr +  XGMAC_MTL_TXQ_OPMODE(channel));
501*4882a593Smuzhiyun }
502*4882a593Smuzhiyun 
dwxgmac2_set_bfsize(void __iomem * ioaddr,int bfsize,u32 chan)503*4882a593Smuzhiyun static void dwxgmac2_set_bfsize(void __iomem *ioaddr, int bfsize, u32 chan)
504*4882a593Smuzhiyun {
505*4882a593Smuzhiyun 	u32 value;
506*4882a593Smuzhiyun 
507*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
508*4882a593Smuzhiyun 	value &= ~XGMAC_RBSZ;
509*4882a593Smuzhiyun 	value |= bfsize << XGMAC_RBSZ_SHIFT;
510*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
511*4882a593Smuzhiyun }
512*4882a593Smuzhiyun 
dwxgmac2_enable_sph(void __iomem * ioaddr,bool en,u32 chan)513*4882a593Smuzhiyun static void dwxgmac2_enable_sph(void __iomem *ioaddr, bool en, u32 chan)
514*4882a593Smuzhiyun {
515*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_RX_CONFIG);
516*4882a593Smuzhiyun 
517*4882a593Smuzhiyun 	value &= ~XGMAC_CONFIG_HDSMS;
518*4882a593Smuzhiyun 	value |= XGMAC_CONFIG_HDSMS_256; /* Segment max 256 bytes */
519*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_RX_CONFIG);
520*4882a593Smuzhiyun 
521*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_CONTROL(chan));
522*4882a593Smuzhiyun 	if (en)
523*4882a593Smuzhiyun 		value |= XGMAC_SPH;
524*4882a593Smuzhiyun 	else
525*4882a593Smuzhiyun 		value &= ~XGMAC_SPH;
526*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_CONTROL(chan));
527*4882a593Smuzhiyun }
528*4882a593Smuzhiyun 
dwxgmac2_enable_tbs(void __iomem * ioaddr,bool en,u32 chan)529*4882a593Smuzhiyun static int dwxgmac2_enable_tbs(void __iomem *ioaddr, bool en, u32 chan)
530*4882a593Smuzhiyun {
531*4882a593Smuzhiyun 	u32 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
532*4882a593Smuzhiyun 
533*4882a593Smuzhiyun 	if (en)
534*4882a593Smuzhiyun 		value |= XGMAC_EDSE;
535*4882a593Smuzhiyun 	else
536*4882a593Smuzhiyun 		value &= ~XGMAC_EDSE;
537*4882a593Smuzhiyun 
538*4882a593Smuzhiyun 	writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
539*4882a593Smuzhiyun 
540*4882a593Smuzhiyun 	value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)) & XGMAC_EDSE;
541*4882a593Smuzhiyun 	if (en && !value)
542*4882a593Smuzhiyun 		return -EIO;
543*4882a593Smuzhiyun 
544*4882a593Smuzhiyun 	writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL0);
545*4882a593Smuzhiyun 	writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL1);
546*4882a593Smuzhiyun 	writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL2);
547*4882a593Smuzhiyun 	writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL3);
548*4882a593Smuzhiyun 	return 0;
549*4882a593Smuzhiyun }
550*4882a593Smuzhiyun 
551*4882a593Smuzhiyun const struct stmmac_dma_ops dwxgmac210_dma_ops = {
552*4882a593Smuzhiyun 	.reset = dwxgmac2_dma_reset,
553*4882a593Smuzhiyun 	.init = dwxgmac2_dma_init,
554*4882a593Smuzhiyun 	.init_chan = dwxgmac2_dma_init_chan,
555*4882a593Smuzhiyun 	.init_rx_chan = dwxgmac2_dma_init_rx_chan,
556*4882a593Smuzhiyun 	.init_tx_chan = dwxgmac2_dma_init_tx_chan,
557*4882a593Smuzhiyun 	.axi = dwxgmac2_dma_axi,
558*4882a593Smuzhiyun 	.dump_regs = dwxgmac2_dma_dump_regs,
559*4882a593Smuzhiyun 	.dma_rx_mode = dwxgmac2_dma_rx_mode,
560*4882a593Smuzhiyun 	.dma_tx_mode = dwxgmac2_dma_tx_mode,
561*4882a593Smuzhiyun 	.enable_dma_irq = dwxgmac2_enable_dma_irq,
562*4882a593Smuzhiyun 	.disable_dma_irq = dwxgmac2_disable_dma_irq,
563*4882a593Smuzhiyun 	.start_tx = dwxgmac2_dma_start_tx,
564*4882a593Smuzhiyun 	.stop_tx = dwxgmac2_dma_stop_tx,
565*4882a593Smuzhiyun 	.start_rx = dwxgmac2_dma_start_rx,
566*4882a593Smuzhiyun 	.stop_rx = dwxgmac2_dma_stop_rx,
567*4882a593Smuzhiyun 	.dma_interrupt = dwxgmac2_dma_interrupt,
568*4882a593Smuzhiyun 	.get_hw_feature = dwxgmac2_get_hw_feature,
569*4882a593Smuzhiyun 	.rx_watchdog = dwxgmac2_rx_watchdog,
570*4882a593Smuzhiyun 	.set_rx_ring_len = dwxgmac2_set_rx_ring_len,
571*4882a593Smuzhiyun 	.set_tx_ring_len = dwxgmac2_set_tx_ring_len,
572*4882a593Smuzhiyun 	.set_rx_tail_ptr = dwxgmac2_set_rx_tail_ptr,
573*4882a593Smuzhiyun 	.set_tx_tail_ptr = dwxgmac2_set_tx_tail_ptr,
574*4882a593Smuzhiyun 	.enable_tso = dwxgmac2_enable_tso,
575*4882a593Smuzhiyun 	.qmode = dwxgmac2_qmode,
576*4882a593Smuzhiyun 	.set_bfsize = dwxgmac2_set_bfsize,
577*4882a593Smuzhiyun 	.enable_sph = dwxgmac2_enable_sph,
578*4882a593Smuzhiyun 	.enable_tbs = dwxgmac2_enable_tbs,
579*4882a593Smuzhiyun };
580