xref: /OK3568_Linux_fs/u-boot/drivers/pci/pcie_dw_rockchip.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0+
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Rockchip DesignWare based PCIe host controller driver
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Copyright (c) 2021 Rockchip, Inc.
6*4882a593Smuzhiyun  */
7*4882a593Smuzhiyun 
8*4882a593Smuzhiyun #include <common.h>
9*4882a593Smuzhiyun #include <clk.h>
10*4882a593Smuzhiyun #include <dm.h>
11*4882a593Smuzhiyun #include <generic-phy.h>
12*4882a593Smuzhiyun #include <pci.h>
13*4882a593Smuzhiyun #include <power-domain.h>
14*4882a593Smuzhiyun #include <power/regulator.h>
15*4882a593Smuzhiyun #include <reset.h>
16*4882a593Smuzhiyun #include <syscon.h>
17*4882a593Smuzhiyun #include <asm/io.h>
18*4882a593Smuzhiyun #include <asm-generic/gpio.h>
19*4882a593Smuzhiyun #include <asm/arch-rockchip/clock.h>
20*4882a593Smuzhiyun #include <linux/iopoll.h>
21*4882a593Smuzhiyun #include <linux/ioport.h>
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun DECLARE_GLOBAL_DATA_PTR;
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun #define RK_PCIE_DBG			0
26*4882a593Smuzhiyun 
27*4882a593Smuzhiyun #define __pcie_dev_print_emit(fmt, ...) \
28*4882a593Smuzhiyun ({ \
29*4882a593Smuzhiyun 	printf(fmt, ##__VA_ARGS__); \
30*4882a593Smuzhiyun })
31*4882a593Smuzhiyun 
32*4882a593Smuzhiyun #ifdef dev_err
33*4882a593Smuzhiyun #undef dev_err
34*4882a593Smuzhiyun #define dev_err(dev, fmt, ...) \
35*4882a593Smuzhiyun ({ \
36*4882a593Smuzhiyun 	if (dev) \
37*4882a593Smuzhiyun 		__pcie_dev_print_emit("%s: " fmt, dev->name, \
38*4882a593Smuzhiyun 				##__VA_ARGS__); \
39*4882a593Smuzhiyun })
40*4882a593Smuzhiyun #endif
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun #ifdef dev_info
43*4882a593Smuzhiyun #undef dev_info
44*4882a593Smuzhiyun #define dev_info dev_err
45*4882a593Smuzhiyun #endif
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun #ifdef DEBUG
48*4882a593Smuzhiyun #define dev_dbg dev_err
49*4882a593Smuzhiyun #else
50*4882a593Smuzhiyun #define dev_dbg(dev, fmt, ...)					\
51*4882a593Smuzhiyun ({								\
52*4882a593Smuzhiyun 	if (0)							\
53*4882a593Smuzhiyun 		__dev_printk(7, dev, fmt, ##__VA_ARGS__);	\
54*4882a593Smuzhiyun })
55*4882a593Smuzhiyun #endif
56*4882a593Smuzhiyun 
57*4882a593Smuzhiyun struct rk_pcie {
58*4882a593Smuzhiyun 	struct udevice	*dev;
59*4882a593Smuzhiyun 	struct udevice  *vpcie3v3;
60*4882a593Smuzhiyun 	void		*dbi_base;
61*4882a593Smuzhiyun 	void		*apb_base;
62*4882a593Smuzhiyun 	void		*cfg_base;
63*4882a593Smuzhiyun 	fdt_size_t	cfg_size;
64*4882a593Smuzhiyun 	struct phy	phy;
65*4882a593Smuzhiyun 	struct clk_bulk	clks;
66*4882a593Smuzhiyun 	int		first_busno;
67*4882a593Smuzhiyun 	struct reset_ctl_bulk	rsts;
68*4882a593Smuzhiyun 	struct gpio_desc	rst_gpio;
69*4882a593Smuzhiyun 	struct pci_region	io;
70*4882a593Smuzhiyun 	struct pci_region	mem;
71*4882a593Smuzhiyun 	bool		is_bifurcation;
72*4882a593Smuzhiyun 	u32 gen;
73*4882a593Smuzhiyun };
74*4882a593Smuzhiyun 
75*4882a593Smuzhiyun enum {
76*4882a593Smuzhiyun 	PCIBIOS_SUCCESSFUL = 0x0000,
77*4882a593Smuzhiyun 	PCIBIOS_UNSUPPORTED = -ENODEV,
78*4882a593Smuzhiyun 	PCIBIOS_NODEV = -ENODEV,
79*4882a593Smuzhiyun };
80*4882a593Smuzhiyun 
81*4882a593Smuzhiyun #define msleep(a)		udelay((a) * 1000)
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun /* Parameters for the waiting for iATU enabled routine */
84*4882a593Smuzhiyun #define PCIE_CLIENT_GENERAL_DEBUG	0x104
85*4882a593Smuzhiyun #define PCIE_CLIENT_HOT_RESET_CTRL	0x180
86*4882a593Smuzhiyun #define PCIE_LTSSM_ENABLE_ENHANCE	BIT(4)
87*4882a593Smuzhiyun #define PCIE_CLIENT_LTSSM_STATUS	0x300
88*4882a593Smuzhiyun #define SMLH_LINKUP			BIT(16)
89*4882a593Smuzhiyun #define RDLH_LINKUP			BIT(17)
90*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_FIFO_MODE_CON	0x310
91*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_FIFO_PTN_HIT_D0 0x320
92*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_FIFO_PTN_HIT_D1 0x324
93*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_FIFO_TRN_HIT_D0 0x328
94*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_FIFO_TRN_HIT_D1 0x32c
95*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_FIFO_STATUS	0x350
96*4882a593Smuzhiyun #define PCIE_CLIENT_DBG_TRANSITION_DATA	0xffff0000
97*4882a593Smuzhiyun #define PCIE_CLIENT_DBF_EN		0xffff0003
98*4882a593Smuzhiyun 
99*4882a593Smuzhiyun /* PCI DBICS registers */
100*4882a593Smuzhiyun #define PCIE_LINK_STATUS_REG		0x80
101*4882a593Smuzhiyun #define PCIE_LINK_STATUS_SPEED_OFF	16
102*4882a593Smuzhiyun #define PCIE_LINK_STATUS_SPEED_MASK	(0xf << PCIE_LINK_STATUS_SPEED_OFF)
103*4882a593Smuzhiyun #define PCIE_LINK_STATUS_WIDTH_OFF	20
104*4882a593Smuzhiyun #define PCIE_LINK_STATUS_WIDTH_MASK	(0xf << PCIE_LINK_STATUS_WIDTH_OFF)
105*4882a593Smuzhiyun 
106*4882a593Smuzhiyun #define PCIE_LINK_CAPABILITY		0x7c
107*4882a593Smuzhiyun #define PCIE_LINK_CTL_2			0xa0
108*4882a593Smuzhiyun #define TARGET_LINK_SPEED_MASK		0xf
109*4882a593Smuzhiyun #define LINK_SPEED_GEN_1		0x1
110*4882a593Smuzhiyun #define LINK_SPEED_GEN_2		0x2
111*4882a593Smuzhiyun #define LINK_SPEED_GEN_3		0x3
112*4882a593Smuzhiyun 
113*4882a593Smuzhiyun #define PCIE_MISC_CONTROL_1_OFF		0x8bc
114*4882a593Smuzhiyun #define PCIE_DBI_RO_WR_EN		BIT(0)
115*4882a593Smuzhiyun 
116*4882a593Smuzhiyun #define PCIE_LINK_WIDTH_SPEED_CONTROL	0x80c
117*4882a593Smuzhiyun #define PORT_LOGIC_SPEED_CHANGE		BIT(17)
118*4882a593Smuzhiyun 
119*4882a593Smuzhiyun /*
120*4882a593Smuzhiyun  * iATU Unroll-specific register definitions
121*4882a593Smuzhiyun  * From 4.80 core version the address translation will be made by unroll.
122*4882a593Smuzhiyun  * The registers are offset from atu_base
123*4882a593Smuzhiyun  */
124*4882a593Smuzhiyun #define PCIE_ATU_UNR_REGION_CTRL1	0x00
125*4882a593Smuzhiyun #define PCIE_ATU_UNR_REGION_CTRL2	0x04
126*4882a593Smuzhiyun #define PCIE_ATU_UNR_LOWER_BASE		0x08
127*4882a593Smuzhiyun #define PCIE_ATU_UNR_UPPER_BASE		0x0c
128*4882a593Smuzhiyun #define PCIE_ATU_UNR_LIMIT		0x10
129*4882a593Smuzhiyun #define PCIE_ATU_UNR_LOWER_TARGET	0x14
130*4882a593Smuzhiyun #define PCIE_ATU_UNR_UPPER_TARGET	0x18
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun #define PCIE_ATU_REGION_INDEX1		(0x1 << 0)
133*4882a593Smuzhiyun #define PCIE_ATU_REGION_INDEX0		(0x0 << 0)
134*4882a593Smuzhiyun #define PCIE_ATU_TYPE_MEM		(0x0 << 0)
135*4882a593Smuzhiyun #define PCIE_ATU_TYPE_IO		(0x2 << 0)
136*4882a593Smuzhiyun #define PCIE_ATU_TYPE_CFG0		(0x4 << 0)
137*4882a593Smuzhiyun #define PCIE_ATU_TYPE_CFG1		(0x5 << 0)
138*4882a593Smuzhiyun #define PCIE_ATU_ENABLE			(0x1 << 31)
139*4882a593Smuzhiyun #define PCIE_ATU_BAR_MODE_ENABLE	(0x1 << 30)
140*4882a593Smuzhiyun #define PCIE_ATU_BUS(x)			(((x) & 0xff) << 24)
141*4882a593Smuzhiyun #define PCIE_ATU_DEV(x)			(((x) & 0x1f) << 19)
142*4882a593Smuzhiyun #define PCIE_ATU_FUNC(x)		(((x) & 0x7) << 16)
143*4882a593Smuzhiyun 
144*4882a593Smuzhiyun /* Register address builder */
145*4882a593Smuzhiyun #define PCIE_GET_ATU_OUTB_UNR_REG_OFFSET(region)        \
146*4882a593Smuzhiyun 	((0x3 << 20) | ((region) << 9))
147*4882a593Smuzhiyun #define PCIE_GET_ATU_INB_UNR_REG_OFFSET(region) \
148*4882a593Smuzhiyun 	((0x3 << 20) | ((region) << 9) | (0x1 << 8))
149*4882a593Smuzhiyun 
150*4882a593Smuzhiyun /* Parameters for the waiting for iATU enabled routine */
151*4882a593Smuzhiyun #define LINK_WAIT_MAX_IATU_RETRIES	5
152*4882a593Smuzhiyun #define LINK_WAIT_IATU			10000
153*4882a593Smuzhiyun 
154*4882a593Smuzhiyun #define PCIE_TYPE0_HDR_DBI2_OFFSET      0x100000
155*4882a593Smuzhiyun 
rk_pcie_read(void __iomem * addr,int size,u32 * val)156*4882a593Smuzhiyun static int rk_pcie_read(void __iomem *addr, int size, u32 *val)
157*4882a593Smuzhiyun {
158*4882a593Smuzhiyun 	if ((uintptr_t)addr & (size - 1)) {
159*4882a593Smuzhiyun 		*val = 0;
160*4882a593Smuzhiyun 		return PCIBIOS_UNSUPPORTED;
161*4882a593Smuzhiyun 	}
162*4882a593Smuzhiyun 
163*4882a593Smuzhiyun 	if (size == 4) {
164*4882a593Smuzhiyun 		*val = readl(addr);
165*4882a593Smuzhiyun 	} else if (size == 2) {
166*4882a593Smuzhiyun 		*val = readw(addr);
167*4882a593Smuzhiyun 	} else if (size == 1) {
168*4882a593Smuzhiyun 		*val = readb(addr);
169*4882a593Smuzhiyun 	} else {
170*4882a593Smuzhiyun 		*val = 0;
171*4882a593Smuzhiyun 		return PCIBIOS_NODEV;
172*4882a593Smuzhiyun 	}
173*4882a593Smuzhiyun 
174*4882a593Smuzhiyun 	return PCIBIOS_SUCCESSFUL;
175*4882a593Smuzhiyun }
176*4882a593Smuzhiyun 
rk_pcie_write(void __iomem * addr,int size,u32 val)177*4882a593Smuzhiyun static int rk_pcie_write(void __iomem *addr, int size, u32 val)
178*4882a593Smuzhiyun {
179*4882a593Smuzhiyun 	if ((uintptr_t)addr & (size - 1))
180*4882a593Smuzhiyun 		return PCIBIOS_UNSUPPORTED;
181*4882a593Smuzhiyun 
182*4882a593Smuzhiyun 	if (size == 4)
183*4882a593Smuzhiyun 		writel(val, addr);
184*4882a593Smuzhiyun 	else if (size == 2)
185*4882a593Smuzhiyun 		writew(val, addr);
186*4882a593Smuzhiyun 	else if (size == 1)
187*4882a593Smuzhiyun 		writeb(val, addr);
188*4882a593Smuzhiyun 	else
189*4882a593Smuzhiyun 		return PCIBIOS_NODEV;
190*4882a593Smuzhiyun 
191*4882a593Smuzhiyun 	return PCIBIOS_SUCCESSFUL;
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun 
__rk_pcie_read_apb(struct rk_pcie * rk_pcie,void __iomem * base,u32 reg,size_t size)194*4882a593Smuzhiyun static u32 __rk_pcie_read_apb(struct rk_pcie *rk_pcie, void __iomem *base,
195*4882a593Smuzhiyun 			      u32 reg, size_t size)
196*4882a593Smuzhiyun {
197*4882a593Smuzhiyun 	int ret;
198*4882a593Smuzhiyun 	u32 val;
199*4882a593Smuzhiyun 
200*4882a593Smuzhiyun 	ret = rk_pcie_read(base + reg, size, &val);
201*4882a593Smuzhiyun 	if (ret)
202*4882a593Smuzhiyun 		dev_err(rk_pcie->dev, "Read APB address failed\n");
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun 	return val;
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun 
__rk_pcie_write_apb(struct rk_pcie * rk_pcie,void __iomem * base,u32 reg,size_t size,u32 val)207*4882a593Smuzhiyun static void __rk_pcie_write_apb(struct rk_pcie *rk_pcie, void __iomem *base,
208*4882a593Smuzhiyun 				u32 reg, size_t size, u32 val)
209*4882a593Smuzhiyun {
210*4882a593Smuzhiyun 	int ret;
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun 	ret = rk_pcie_write(base + reg, size, val);
213*4882a593Smuzhiyun 	if (ret)
214*4882a593Smuzhiyun 		dev_err(rk_pcie->dev, "Write APB address failed\n");
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun 
rk_pcie_readl_apb(struct rk_pcie * rk_pcie,u32 reg)217*4882a593Smuzhiyun static inline u32 rk_pcie_readl_apb(struct rk_pcie *rk_pcie, u32 reg)
218*4882a593Smuzhiyun {
219*4882a593Smuzhiyun 	return __rk_pcie_read_apb(rk_pcie, rk_pcie->apb_base, reg, 0x4);
220*4882a593Smuzhiyun }
221*4882a593Smuzhiyun 
rk_pcie_writel_apb(struct rk_pcie * rk_pcie,u32 reg,u32 val)222*4882a593Smuzhiyun static inline void rk_pcie_writel_apb(struct rk_pcie *rk_pcie, u32 reg,
223*4882a593Smuzhiyun 				      u32 val)
224*4882a593Smuzhiyun {
225*4882a593Smuzhiyun 	__rk_pcie_write_apb(rk_pcie, rk_pcie->apb_base, reg, 0x4, val);
226*4882a593Smuzhiyun }
227*4882a593Smuzhiyun 
rk_pcie_get_link_speed(struct rk_pcie * rk_pcie)228*4882a593Smuzhiyun static int rk_pcie_get_link_speed(struct rk_pcie *rk_pcie)
229*4882a593Smuzhiyun {
230*4882a593Smuzhiyun 	return (readl(rk_pcie->dbi_base + PCIE_LINK_STATUS_REG) &
231*4882a593Smuzhiyun 		PCIE_LINK_STATUS_SPEED_MASK) >> PCIE_LINK_STATUS_SPEED_OFF;
232*4882a593Smuzhiyun }
233*4882a593Smuzhiyun 
rk_pcie_get_link_width(struct rk_pcie * rk_pcie)234*4882a593Smuzhiyun static int rk_pcie_get_link_width(struct rk_pcie *rk_pcie)
235*4882a593Smuzhiyun {
236*4882a593Smuzhiyun 	return (readl(rk_pcie->dbi_base + PCIE_LINK_STATUS_REG) &
237*4882a593Smuzhiyun 		PCIE_LINK_STATUS_WIDTH_MASK) >> PCIE_LINK_STATUS_WIDTH_OFF;
238*4882a593Smuzhiyun }
239*4882a593Smuzhiyun 
rk_pcie_writel_ob_unroll(struct rk_pcie * rk_pcie,u32 index,u32 reg,u32 val)240*4882a593Smuzhiyun static void rk_pcie_writel_ob_unroll(struct rk_pcie *rk_pcie, u32 index,
241*4882a593Smuzhiyun 				     u32 reg, u32 val)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun 	u32 offset = PCIE_GET_ATU_OUTB_UNR_REG_OFFSET(index);
244*4882a593Smuzhiyun 	void __iomem *base = rk_pcie->dbi_base;
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun 	writel(val, base + offset + reg);
247*4882a593Smuzhiyun }
248*4882a593Smuzhiyun 
rk_pcie_readl_ob_unroll(struct rk_pcie * rk_pcie,u32 index,u32 reg)249*4882a593Smuzhiyun static u32 rk_pcie_readl_ob_unroll(struct rk_pcie *rk_pcie, u32 index, u32 reg)
250*4882a593Smuzhiyun {
251*4882a593Smuzhiyun 	u32 offset = PCIE_GET_ATU_OUTB_UNR_REG_OFFSET(index);
252*4882a593Smuzhiyun 	void __iomem *base = rk_pcie->dbi_base;
253*4882a593Smuzhiyun 
254*4882a593Smuzhiyun 	return readl(base + offset + reg);
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun 
rk_pcie_dbi_write_enable(struct rk_pcie * rk_pcie,bool en)257*4882a593Smuzhiyun static inline void rk_pcie_dbi_write_enable(struct rk_pcie *rk_pcie, bool en)
258*4882a593Smuzhiyun {
259*4882a593Smuzhiyun 	u32 val;
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 	val = readl(rk_pcie->dbi_base + PCIE_MISC_CONTROL_1_OFF);
262*4882a593Smuzhiyun 
263*4882a593Smuzhiyun 	if (en)
264*4882a593Smuzhiyun 		val |= PCIE_DBI_RO_WR_EN;
265*4882a593Smuzhiyun 	else
266*4882a593Smuzhiyun 		val &= ~PCIE_DBI_RO_WR_EN;
267*4882a593Smuzhiyun 	writel(val, rk_pcie->dbi_base + PCIE_MISC_CONTROL_1_OFF);
268*4882a593Smuzhiyun }
269*4882a593Smuzhiyun 
rk_pcie_setup_host(struct rk_pcie * rk_pcie)270*4882a593Smuzhiyun static void rk_pcie_setup_host(struct rk_pcie *rk_pcie)
271*4882a593Smuzhiyun {
272*4882a593Smuzhiyun 	u32 val;
273*4882a593Smuzhiyun 
274*4882a593Smuzhiyun 	rk_pcie_dbi_write_enable(rk_pcie, true);
275*4882a593Smuzhiyun 
276*4882a593Smuzhiyun 	/* setup RC BARs */
277*4882a593Smuzhiyun 	writel(PCI_BASE_ADDRESS_MEM_TYPE_64,
278*4882a593Smuzhiyun 	       rk_pcie->dbi_base + PCI_BASE_ADDRESS_0);
279*4882a593Smuzhiyun 	writel(0x0, rk_pcie->dbi_base + PCI_BASE_ADDRESS_1);
280*4882a593Smuzhiyun 
281*4882a593Smuzhiyun 	/* setup interrupt pins */
282*4882a593Smuzhiyun 	val = readl(rk_pcie->dbi_base + PCI_INTERRUPT_LINE);
283*4882a593Smuzhiyun 	val &= 0xffff00ff;
284*4882a593Smuzhiyun 	val |= 0x00000100;
285*4882a593Smuzhiyun 	writel(val, rk_pcie->dbi_base + PCI_INTERRUPT_LINE);
286*4882a593Smuzhiyun 
287*4882a593Smuzhiyun 	/* setup bus numbers */
288*4882a593Smuzhiyun 	val = readl(rk_pcie->dbi_base + PCI_PRIMARY_BUS);
289*4882a593Smuzhiyun 	val &= 0xff000000;
290*4882a593Smuzhiyun 	val |= 0x00ff0100;
291*4882a593Smuzhiyun 	writel(val, rk_pcie->dbi_base + PCI_PRIMARY_BUS);
292*4882a593Smuzhiyun 
293*4882a593Smuzhiyun 	val = readl(rk_pcie->dbi_base + PCI_PRIMARY_BUS);
294*4882a593Smuzhiyun 
295*4882a593Smuzhiyun 	/* setup command register */
296*4882a593Smuzhiyun 	val = readl(rk_pcie->dbi_base + PCI_COMMAND);
297*4882a593Smuzhiyun 	val &= 0xffff0000;
298*4882a593Smuzhiyun 	val |= PCI_COMMAND_IO | PCI_COMMAND_MEMORY |
299*4882a593Smuzhiyun 		PCI_COMMAND_MASTER | PCI_COMMAND_SERR;
300*4882a593Smuzhiyun 	writel(val, rk_pcie->dbi_base + PCI_COMMAND);
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun 	/* program correct class for RC */
303*4882a593Smuzhiyun 	writew(PCI_CLASS_BRIDGE_PCI, rk_pcie->dbi_base + PCI_CLASS_DEVICE);
304*4882a593Smuzhiyun 	/* Better disable write permission right after the update */
305*4882a593Smuzhiyun 
306*4882a593Smuzhiyun 	val = readl(rk_pcie->dbi_base + PCIE_LINK_WIDTH_SPEED_CONTROL);
307*4882a593Smuzhiyun 	val |= PORT_LOGIC_SPEED_CHANGE;
308*4882a593Smuzhiyun 	writel(val, rk_pcie->dbi_base + PCIE_LINK_WIDTH_SPEED_CONTROL);
309*4882a593Smuzhiyun 
310*4882a593Smuzhiyun 	/* Disable BAR0 BAR1 */
311*4882a593Smuzhiyun 	writel(0, rk_pcie->dbi_base + PCIE_TYPE0_HDR_DBI2_OFFSET + 0x10 + 0 * 4);
312*4882a593Smuzhiyun 	writel(0, rk_pcie->dbi_base + PCIE_TYPE0_HDR_DBI2_OFFSET + 0x10 + 1 * 4);
313*4882a593Smuzhiyun 
314*4882a593Smuzhiyun 	rk_pcie_dbi_write_enable(rk_pcie, false);
315*4882a593Smuzhiyun }
316*4882a593Smuzhiyun 
rk_pcie_configure(struct rk_pcie * pci,u32 cap_speed)317*4882a593Smuzhiyun static void rk_pcie_configure(struct rk_pcie *pci, u32 cap_speed)
318*4882a593Smuzhiyun {
319*4882a593Smuzhiyun 	u32 val;
320*4882a593Smuzhiyun 
321*4882a593Smuzhiyun 	rk_pcie_dbi_write_enable(pci, true);
322*4882a593Smuzhiyun 
323*4882a593Smuzhiyun 	val = readl(pci->dbi_base + PCIE_LINK_CAPABILITY);
324*4882a593Smuzhiyun 	val &= ~TARGET_LINK_SPEED_MASK;
325*4882a593Smuzhiyun 	val |= cap_speed;
326*4882a593Smuzhiyun 	writel(val, pci->dbi_base + PCIE_LINK_CAPABILITY);
327*4882a593Smuzhiyun 
328*4882a593Smuzhiyun 	val = readl(pci->dbi_base + PCIE_LINK_CTL_2);
329*4882a593Smuzhiyun 	val &= ~TARGET_LINK_SPEED_MASK;
330*4882a593Smuzhiyun 	val |= cap_speed;
331*4882a593Smuzhiyun 	writel(val, pci->dbi_base + PCIE_LINK_CTL_2);
332*4882a593Smuzhiyun 
333*4882a593Smuzhiyun 	rk_pcie_dbi_write_enable(pci, false);
334*4882a593Smuzhiyun }
335*4882a593Smuzhiyun 
rk_pcie_prog_outbound_atu_unroll(struct rk_pcie * pci,int index,int type,u64 cpu_addr,u64 pci_addr,u32 size)336*4882a593Smuzhiyun static void rk_pcie_prog_outbound_atu_unroll(struct rk_pcie *pci, int index,
337*4882a593Smuzhiyun 					     int type, u64 cpu_addr,
338*4882a593Smuzhiyun 					     u64 pci_addr, u32 size)
339*4882a593Smuzhiyun {
340*4882a593Smuzhiyun 	u32 retries, val;
341*4882a593Smuzhiyun 
342*4882a593Smuzhiyun 	dev_dbg(pci->dev, "ATU programmed with: index: %d, type: %d, cpu addr: %8llx, pci addr: %8llx, size: %8x\n",
343*4882a593Smuzhiyun 		index, type, cpu_addr, pci_addr, size);
344*4882a593Smuzhiyun 
345*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_LOWER_BASE,
346*4882a593Smuzhiyun 				 lower_32_bits(cpu_addr));
347*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_UPPER_BASE,
348*4882a593Smuzhiyun 				 upper_32_bits(cpu_addr));
349*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_LIMIT,
350*4882a593Smuzhiyun 				 lower_32_bits(cpu_addr + size - 1));
351*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_LOWER_TARGET,
352*4882a593Smuzhiyun 				 lower_32_bits(pci_addr));
353*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_UPPER_TARGET,
354*4882a593Smuzhiyun 				 upper_32_bits(pci_addr));
355*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL1,
356*4882a593Smuzhiyun 				 type);
357*4882a593Smuzhiyun 	rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL2,
358*4882a593Smuzhiyun 				 PCIE_ATU_ENABLE);
359*4882a593Smuzhiyun 
360*4882a593Smuzhiyun 	/*
361*4882a593Smuzhiyun 	 * Make sure ATU enable takes effect before any subsequent config
362*4882a593Smuzhiyun 	 * and I/O accesses.
363*4882a593Smuzhiyun 	 */
364*4882a593Smuzhiyun 	for (retries = 0; retries < LINK_WAIT_MAX_IATU_RETRIES; retries++) {
365*4882a593Smuzhiyun 		val = rk_pcie_readl_ob_unroll(pci, index,
366*4882a593Smuzhiyun 					      PCIE_ATU_UNR_REGION_CTRL2);
367*4882a593Smuzhiyun 		if (val & PCIE_ATU_ENABLE)
368*4882a593Smuzhiyun 			return;
369*4882a593Smuzhiyun 
370*4882a593Smuzhiyun 		udelay(LINK_WAIT_IATU);
371*4882a593Smuzhiyun 	}
372*4882a593Smuzhiyun 	dev_err(pci->dev, "outbound iATU is not being enabled\n");
373*4882a593Smuzhiyun }
374*4882a593Smuzhiyun 
rk_pcie_addr_valid(pci_dev_t d,int first_busno)375*4882a593Smuzhiyun static int rk_pcie_addr_valid(pci_dev_t d, int first_busno)
376*4882a593Smuzhiyun {
377*4882a593Smuzhiyun 	if ((PCI_BUS(d) == first_busno) && (PCI_DEV(d) > 0))
378*4882a593Smuzhiyun 		return 0;
379*4882a593Smuzhiyun 	if ((PCI_BUS(d) == first_busno + 1) && (PCI_DEV(d) > 0))
380*4882a593Smuzhiyun 		return 0;
381*4882a593Smuzhiyun 
382*4882a593Smuzhiyun 	return 1;
383*4882a593Smuzhiyun }
384*4882a593Smuzhiyun 
set_cfg_address(struct rk_pcie * pcie,pci_dev_t d,uint where)385*4882a593Smuzhiyun static uintptr_t set_cfg_address(struct rk_pcie *pcie,
386*4882a593Smuzhiyun 				 pci_dev_t d, uint where)
387*4882a593Smuzhiyun {
388*4882a593Smuzhiyun 	int bus = PCI_BUS(d) - pcie->first_busno;
389*4882a593Smuzhiyun 	uintptr_t va_address;
390*4882a593Smuzhiyun 	u32 atu_type;
391*4882a593Smuzhiyun 
392*4882a593Smuzhiyun 	/* Use dbi_base for own configuration read and write */
393*4882a593Smuzhiyun 	if (!bus) {
394*4882a593Smuzhiyun 		va_address = (uintptr_t)pcie->dbi_base;
395*4882a593Smuzhiyun 		goto out;
396*4882a593Smuzhiyun 	}
397*4882a593Smuzhiyun 
398*4882a593Smuzhiyun 	if (bus == 1)
399*4882a593Smuzhiyun 		/*
400*4882a593Smuzhiyun 		 * For local bus whose primary bus number is root bridge,
401*4882a593Smuzhiyun 		 * change TLP Type field to 4.
402*4882a593Smuzhiyun 		 */
403*4882a593Smuzhiyun 		atu_type = PCIE_ATU_TYPE_CFG0;
404*4882a593Smuzhiyun 	else
405*4882a593Smuzhiyun 		/* Otherwise, change TLP Type field to 5. */
406*4882a593Smuzhiyun 		atu_type = PCIE_ATU_TYPE_CFG1;
407*4882a593Smuzhiyun 
408*4882a593Smuzhiyun 	/*
409*4882a593Smuzhiyun 	 * Not accessing root port configuration space?
410*4882a593Smuzhiyun 	 * Region #0 is used for Outbound CFG space access.
411*4882a593Smuzhiyun 	 * Direction = Outbound
412*4882a593Smuzhiyun 	 * Region Index = 0
413*4882a593Smuzhiyun 	 */
414*4882a593Smuzhiyun 	d = PCI_MASK_BUS(d);
415*4882a593Smuzhiyun 	d = PCI_ADD_BUS(bus, d);
416*4882a593Smuzhiyun 	rk_pcie_prog_outbound_atu_unroll(pcie, PCIE_ATU_REGION_INDEX1,
417*4882a593Smuzhiyun 					 atu_type, (u64)pcie->cfg_base,
418*4882a593Smuzhiyun 					 d << 8, pcie->cfg_size);
419*4882a593Smuzhiyun 
420*4882a593Smuzhiyun 	va_address = (uintptr_t)pcie->cfg_base;
421*4882a593Smuzhiyun 
422*4882a593Smuzhiyun out:
423*4882a593Smuzhiyun 	va_address += where & ~0x3;
424*4882a593Smuzhiyun 
425*4882a593Smuzhiyun 	return va_address;
426*4882a593Smuzhiyun }
427*4882a593Smuzhiyun 
rockchip_pcie_rd_conf(struct udevice * bus,pci_dev_t bdf,uint offset,ulong * valuep,enum pci_size_t size)428*4882a593Smuzhiyun static int rockchip_pcie_rd_conf(struct udevice *bus, pci_dev_t bdf,
429*4882a593Smuzhiyun 				 uint offset, ulong *valuep,
430*4882a593Smuzhiyun 				 enum pci_size_t size)
431*4882a593Smuzhiyun {
432*4882a593Smuzhiyun 	struct rk_pcie *pcie = dev_get_priv(bus);
433*4882a593Smuzhiyun 	uintptr_t va_address;
434*4882a593Smuzhiyun 	ulong value;
435*4882a593Smuzhiyun 
436*4882a593Smuzhiyun 	debug("PCIE CFG read: bdf=%2x:%2x:%2x\n",
437*4882a593Smuzhiyun 	      PCI_BUS(bdf), PCI_DEV(bdf), PCI_FUNC(bdf));
438*4882a593Smuzhiyun 
439*4882a593Smuzhiyun 	if (!rk_pcie_addr_valid(bdf, pcie->first_busno)) {
440*4882a593Smuzhiyun 		debug("- out of range\n");
441*4882a593Smuzhiyun 		*valuep = pci_get_ff(size);
442*4882a593Smuzhiyun 		return 0;
443*4882a593Smuzhiyun 	}
444*4882a593Smuzhiyun 
445*4882a593Smuzhiyun 	va_address = set_cfg_address(pcie, bdf, offset);
446*4882a593Smuzhiyun 
447*4882a593Smuzhiyun 	value = readl(va_address);
448*4882a593Smuzhiyun 
449*4882a593Smuzhiyun 	debug("(addr,val)=(0x%04x, 0x%08lx)\n", offset, value);
450*4882a593Smuzhiyun 	*valuep = pci_conv_32_to_size(value, offset, size);
451*4882a593Smuzhiyun 
452*4882a593Smuzhiyun 	rk_pcie_prog_outbound_atu_unroll(pcie, PCIE_ATU_REGION_INDEX1,
453*4882a593Smuzhiyun 					 PCIE_ATU_TYPE_IO, pcie->io.phys_start,
454*4882a593Smuzhiyun 					 pcie->io.bus_start, pcie->io.size);
455*4882a593Smuzhiyun 
456*4882a593Smuzhiyun 	return 0;
457*4882a593Smuzhiyun }
458*4882a593Smuzhiyun 
rockchip_pcie_wr_conf(struct udevice * bus,pci_dev_t bdf,uint offset,ulong value,enum pci_size_t size)459*4882a593Smuzhiyun static int rockchip_pcie_wr_conf(struct udevice *bus, pci_dev_t bdf,
460*4882a593Smuzhiyun 				 uint offset, ulong value,
461*4882a593Smuzhiyun 				 enum pci_size_t size)
462*4882a593Smuzhiyun {
463*4882a593Smuzhiyun 	struct rk_pcie *pcie = dev_get_priv(bus);
464*4882a593Smuzhiyun 	uintptr_t va_address;
465*4882a593Smuzhiyun 	ulong old;
466*4882a593Smuzhiyun 
467*4882a593Smuzhiyun 	debug("PCIE CFG write: (b,d,f)=(%2d,%2d,%2d)\n",
468*4882a593Smuzhiyun 	      PCI_BUS(bdf), PCI_DEV(bdf), PCI_FUNC(bdf));
469*4882a593Smuzhiyun 	debug("(addr,val)=(0x%04x, 0x%08lx)\n", offset, value);
470*4882a593Smuzhiyun 
471*4882a593Smuzhiyun 	if (!rk_pcie_addr_valid(bdf, pcie->first_busno)) {
472*4882a593Smuzhiyun 		debug("- out of range\n");
473*4882a593Smuzhiyun 		return 0;
474*4882a593Smuzhiyun 	}
475*4882a593Smuzhiyun 
476*4882a593Smuzhiyun 	va_address = set_cfg_address(pcie, bdf, offset);
477*4882a593Smuzhiyun 
478*4882a593Smuzhiyun 	old = readl(va_address);
479*4882a593Smuzhiyun 	value = pci_conv_size_to_32(old, value, offset, size);
480*4882a593Smuzhiyun 	writel(value, va_address);
481*4882a593Smuzhiyun 
482*4882a593Smuzhiyun 	rk_pcie_prog_outbound_atu_unroll(pcie, PCIE_ATU_REGION_INDEX1,
483*4882a593Smuzhiyun 					 PCIE_ATU_TYPE_IO, pcie->io.phys_start,
484*4882a593Smuzhiyun 					 pcie->io.bus_start, pcie->io.size);
485*4882a593Smuzhiyun 
486*4882a593Smuzhiyun 	return 0;
487*4882a593Smuzhiyun }
488*4882a593Smuzhiyun 
rk_pcie_enable_debug(struct rk_pcie * rk_pcie)489*4882a593Smuzhiyun static void rk_pcie_enable_debug(struct rk_pcie *rk_pcie)
490*4882a593Smuzhiyun {
491*4882a593Smuzhiyun #if RK_PCIE_DBG
492*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_PTN_HIT_D0,
493*4882a593Smuzhiyun 			   PCIE_CLIENT_DBG_TRANSITION_DATA);
494*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_PTN_HIT_D1,
495*4882a593Smuzhiyun 			   PCIE_CLIENT_DBG_TRANSITION_DATA);
496*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_TRN_HIT_D0,
497*4882a593Smuzhiyun 			   PCIE_CLIENT_DBG_TRANSITION_DATA);
498*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_TRN_HIT_D1,
499*4882a593Smuzhiyun 			   PCIE_CLIENT_DBG_TRANSITION_DATA);
500*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_MODE_CON,
501*4882a593Smuzhiyun 			   PCIE_CLIENT_DBF_EN);
502*4882a593Smuzhiyun #endif
503*4882a593Smuzhiyun }
504*4882a593Smuzhiyun 
rk_pcie_debug_dump(struct rk_pcie * rk_pcie)505*4882a593Smuzhiyun static void rk_pcie_debug_dump(struct rk_pcie *rk_pcie)
506*4882a593Smuzhiyun {
507*4882a593Smuzhiyun #if RK_PCIE_DBG
508*4882a593Smuzhiyun 	u32 loop;
509*4882a593Smuzhiyun 
510*4882a593Smuzhiyun 	dev_err(rk_pcie->dev, "ltssm = 0x%x\n",
511*4882a593Smuzhiyun 		 rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_LTSSM_STATUS));
512*4882a593Smuzhiyun 	for (loop = 0; loop < 64; loop++)
513*4882a593Smuzhiyun 		dev_err(rk_pcie->dev, "fifo_status = 0x%x\n",
514*4882a593Smuzhiyun 			 rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_STATUS));
515*4882a593Smuzhiyun #endif
516*4882a593Smuzhiyun }
517*4882a593Smuzhiyun 
rk_pcie_link_status_clear(struct rk_pcie * rk_pcie)518*4882a593Smuzhiyun static inline void rk_pcie_link_status_clear(struct rk_pcie *rk_pcie)
519*4882a593Smuzhiyun {
520*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_GENERAL_DEBUG, 0x0);
521*4882a593Smuzhiyun }
522*4882a593Smuzhiyun 
rk_pcie_disable_ltssm(struct rk_pcie * rk_pcie)523*4882a593Smuzhiyun static inline void rk_pcie_disable_ltssm(struct rk_pcie *rk_pcie)
524*4882a593Smuzhiyun {
525*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, 0x0, 0xc0008);
526*4882a593Smuzhiyun }
527*4882a593Smuzhiyun 
rk_pcie_enable_ltssm(struct rk_pcie * rk_pcie)528*4882a593Smuzhiyun static inline void rk_pcie_enable_ltssm(struct rk_pcie *rk_pcie)
529*4882a593Smuzhiyun {
530*4882a593Smuzhiyun 	rk_pcie_writel_apb(rk_pcie, 0x0, 0xc000c);
531*4882a593Smuzhiyun }
532*4882a593Smuzhiyun 
is_link_up(struct rk_pcie * priv)533*4882a593Smuzhiyun static int is_link_up(struct rk_pcie *priv)
534*4882a593Smuzhiyun {
535*4882a593Smuzhiyun 	u32 val;
536*4882a593Smuzhiyun 
537*4882a593Smuzhiyun 	val = rk_pcie_readl_apb(priv, PCIE_CLIENT_LTSSM_STATUS);
538*4882a593Smuzhiyun 	if ((val & (RDLH_LINKUP | SMLH_LINKUP)) == 0x30000 &&
539*4882a593Smuzhiyun 	    (val & GENMASK(5, 0)) == 0x11)
540*4882a593Smuzhiyun 		return 1;
541*4882a593Smuzhiyun 
542*4882a593Smuzhiyun 	return 0;
543*4882a593Smuzhiyun }
544*4882a593Smuzhiyun 
rk_pcie_link_up(struct rk_pcie * priv,u32 cap_speed)545*4882a593Smuzhiyun static int rk_pcie_link_up(struct rk_pcie *priv, u32 cap_speed)
546*4882a593Smuzhiyun {
547*4882a593Smuzhiyun 	int retries;
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun 	if (is_link_up(priv)) {
550*4882a593Smuzhiyun 		printf("PCI Link already up before configuration!\n");
551*4882a593Smuzhiyun 		return 1;
552*4882a593Smuzhiyun 	}
553*4882a593Smuzhiyun 
554*4882a593Smuzhiyun 	/* DW pre link configurations */
555*4882a593Smuzhiyun 	rk_pcie_configure(priv, cap_speed);
556*4882a593Smuzhiyun 
557*4882a593Smuzhiyun 	/* Release the device */
558*4882a593Smuzhiyun 	if (dm_gpio_is_valid(&priv->rst_gpio)) {
559*4882a593Smuzhiyun 		/*
560*4882a593Smuzhiyun 		 * T_PVPERL (Power stable to PERST# inactive) should be a minimum of 100ms.
561*4882a593Smuzhiyun 		 * We add a 200ms by default for sake of hoping everthings
562*4882a593Smuzhiyun 		 * work fine.
563*4882a593Smuzhiyun 		 */
564*4882a593Smuzhiyun 		msleep(200);
565*4882a593Smuzhiyun 		dm_gpio_set_value(&priv->rst_gpio, 1);
566*4882a593Smuzhiyun 		/*
567*4882a593Smuzhiyun 		 * Add this 20ms delay because we observe link is always up stably after it and
568*4882a593Smuzhiyun 		 * could help us save 20ms for scanning devices.
569*4882a593Smuzhiyun 		 */
570*4882a593Smuzhiyun 		msleep(20);
571*4882a593Smuzhiyun 	}
572*4882a593Smuzhiyun 
573*4882a593Smuzhiyun 	rk_pcie_disable_ltssm(priv);
574*4882a593Smuzhiyun 	rk_pcie_link_status_clear(priv);
575*4882a593Smuzhiyun 	rk_pcie_enable_debug(priv);
576*4882a593Smuzhiyun 
577*4882a593Smuzhiyun 	/* Enable LTSSM */
578*4882a593Smuzhiyun 	rk_pcie_enable_ltssm(priv);
579*4882a593Smuzhiyun 
580*4882a593Smuzhiyun 	for (retries = 0; retries < 50; retries++) {
581*4882a593Smuzhiyun 		if (is_link_up(priv)) {
582*4882a593Smuzhiyun 			dev_info(priv->dev, "PCIe Link up, LTSSM is 0x%x\n",
583*4882a593Smuzhiyun 				 rk_pcie_readl_apb(priv, PCIE_CLIENT_LTSSM_STATUS));
584*4882a593Smuzhiyun 			rk_pcie_debug_dump(priv);
585*4882a593Smuzhiyun 			/* Link maybe in Gen switch recovery but we need to wait more 1s */
586*4882a593Smuzhiyun 			msleep(1000);
587*4882a593Smuzhiyun 			return 0;
588*4882a593Smuzhiyun 		}
589*4882a593Smuzhiyun 
590*4882a593Smuzhiyun 		dev_info(priv->dev, "PCIe Linking... LTSSM is 0x%x\n",
591*4882a593Smuzhiyun 			 rk_pcie_readl_apb(priv, PCIE_CLIENT_LTSSM_STATUS));
592*4882a593Smuzhiyun 		rk_pcie_debug_dump(priv);
593*4882a593Smuzhiyun 		msleep(10);
594*4882a593Smuzhiyun 	}
595*4882a593Smuzhiyun 
596*4882a593Smuzhiyun 	dev_err(priv->dev, "PCIe-%d Link Fail\n", priv->dev->seq);
597*4882a593Smuzhiyun 	return -EINVAL;
598*4882a593Smuzhiyun }
599*4882a593Smuzhiyun 
rockchip_pcie_init_port(struct udevice * dev)600*4882a593Smuzhiyun static int rockchip_pcie_init_port(struct udevice *dev)
601*4882a593Smuzhiyun {
602*4882a593Smuzhiyun 	int ret;
603*4882a593Smuzhiyun 	u32 val;
604*4882a593Smuzhiyun 	struct rk_pcie *priv = dev_get_priv(dev);
605*4882a593Smuzhiyun 	union phy_configure_opts phy_cfg;
606*4882a593Smuzhiyun 
607*4882a593Smuzhiyun 	/* Rest the device */
608*4882a593Smuzhiyun 	if (dm_gpio_is_valid(&priv->rst_gpio))
609*4882a593Smuzhiyun 		dm_gpio_set_value(&priv->rst_gpio, 0);
610*4882a593Smuzhiyun 
611*4882a593Smuzhiyun 	/* Set power and maybe external ref clk input */
612*4882a593Smuzhiyun 	if (priv->vpcie3v3) {
613*4882a593Smuzhiyun 		ret = regulator_set_enable(priv->vpcie3v3, true);
614*4882a593Smuzhiyun 		if (ret) {
615*4882a593Smuzhiyun 			dev_err(priv->dev, "failed to enable vpcie3v3 (ret=%d)\n",
616*4882a593Smuzhiyun 				ret);
617*4882a593Smuzhiyun 			return ret;
618*4882a593Smuzhiyun 		}
619*4882a593Smuzhiyun 	}
620*4882a593Smuzhiyun 
621*4882a593Smuzhiyun 	if (priv->is_bifurcation) {
622*4882a593Smuzhiyun 		phy_cfg.pcie.is_bifurcation = true;
623*4882a593Smuzhiyun 		ret = generic_phy_configure(&priv->phy, &phy_cfg);
624*4882a593Smuzhiyun 		if (ret)
625*4882a593Smuzhiyun 			dev_err(dev, "failed to set bifurcation for phy (ret=%d)\n", ret);
626*4882a593Smuzhiyun 	}
627*4882a593Smuzhiyun 
628*4882a593Smuzhiyun 	ret = generic_phy_init(&priv->phy);
629*4882a593Smuzhiyun 	if (ret) {
630*4882a593Smuzhiyun 		dev_err(dev, "failed to init phy (ret=%d)\n", ret);
631*4882a593Smuzhiyun 		return ret;
632*4882a593Smuzhiyun 	}
633*4882a593Smuzhiyun 
634*4882a593Smuzhiyun 	ret = generic_phy_power_on(&priv->phy);
635*4882a593Smuzhiyun 	if (ret) {
636*4882a593Smuzhiyun 		dev_err(dev, "failed to power on phy (ret=%d)\n", ret);
637*4882a593Smuzhiyun 		goto err_exit_phy;
638*4882a593Smuzhiyun 	}
639*4882a593Smuzhiyun 
640*4882a593Smuzhiyun 	ret = reset_deassert_bulk(&priv->rsts);
641*4882a593Smuzhiyun 	if (ret) {
642*4882a593Smuzhiyun 		dev_err(dev, "failed to deassert resets (ret=%d)\n", ret);
643*4882a593Smuzhiyun 		goto err_power_off_phy;
644*4882a593Smuzhiyun 	}
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun 	ret = clk_enable_bulk(&priv->clks);
647*4882a593Smuzhiyun 	if (ret) {
648*4882a593Smuzhiyun 		dev_err(dev, "failed to enable clks (ret=%d)\n", ret);
649*4882a593Smuzhiyun 		goto err_deassert_bulk;
650*4882a593Smuzhiyun 	}
651*4882a593Smuzhiyun 
652*4882a593Smuzhiyun 	/* LTSSM EN ctrl mode */
653*4882a593Smuzhiyun 	val = rk_pcie_readl_apb(priv, PCIE_CLIENT_HOT_RESET_CTRL);
654*4882a593Smuzhiyun 	val |= PCIE_LTSSM_ENABLE_ENHANCE | (PCIE_LTSSM_ENABLE_ENHANCE << 16);
655*4882a593Smuzhiyun 	rk_pcie_writel_apb(priv, PCIE_CLIENT_HOT_RESET_CTRL, val);
656*4882a593Smuzhiyun 
657*4882a593Smuzhiyun 	/* Set RC mode */
658*4882a593Smuzhiyun 	rk_pcie_writel_apb(priv, 0x0, 0xf00040);
659*4882a593Smuzhiyun 	rk_pcie_setup_host(priv);
660*4882a593Smuzhiyun 
661*4882a593Smuzhiyun 	ret = rk_pcie_link_up(priv, priv->gen);
662*4882a593Smuzhiyun 	if (ret < 0)
663*4882a593Smuzhiyun 		goto err_link_up;
664*4882a593Smuzhiyun 
665*4882a593Smuzhiyun 	return 0;
666*4882a593Smuzhiyun err_link_up:
667*4882a593Smuzhiyun 	clk_disable_bulk(&priv->clks);
668*4882a593Smuzhiyun err_deassert_bulk:
669*4882a593Smuzhiyun 	reset_assert_bulk(&priv->rsts);
670*4882a593Smuzhiyun err_power_off_phy:
671*4882a593Smuzhiyun 	generic_phy_power_off(&priv->phy);
672*4882a593Smuzhiyun err_exit_phy:
673*4882a593Smuzhiyun 	generic_phy_exit(&priv->phy);
674*4882a593Smuzhiyun 	return ret;
675*4882a593Smuzhiyun }
676*4882a593Smuzhiyun 
rockchip_pcie_parse_dt(struct udevice * dev)677*4882a593Smuzhiyun static int rockchip_pcie_parse_dt(struct udevice *dev)
678*4882a593Smuzhiyun {
679*4882a593Smuzhiyun 	struct rk_pcie *priv = dev_get_priv(dev);
680*4882a593Smuzhiyun 	u32 max_link_speed;
681*4882a593Smuzhiyun 	int ret;
682*4882a593Smuzhiyun 	struct resource res;
683*4882a593Smuzhiyun 
684*4882a593Smuzhiyun 	ret = dev_read_resource_byname(dev, "pcie-dbi", &res);
685*4882a593Smuzhiyun 	if (ret)
686*4882a593Smuzhiyun 		return -ENODEV;
687*4882a593Smuzhiyun 	priv->dbi_base = (void *)(res.start);
688*4882a593Smuzhiyun 	dev_dbg(dev, "DBI address is 0x%p\n", priv->dbi_base);
689*4882a593Smuzhiyun 
690*4882a593Smuzhiyun 	ret = dev_read_resource_byname(dev, "pcie-apb", &res);
691*4882a593Smuzhiyun 	if (ret)
692*4882a593Smuzhiyun 		return -ENODEV;
693*4882a593Smuzhiyun 	priv->apb_base = (void *)(res.start);
694*4882a593Smuzhiyun 	dev_dbg(dev, "APB address is 0x%p\n", priv->apb_base);
695*4882a593Smuzhiyun 
696*4882a593Smuzhiyun 	ret = gpio_request_by_name(dev, "reset-gpios", 0,
697*4882a593Smuzhiyun 				   &priv->rst_gpio, GPIOD_IS_OUT);
698*4882a593Smuzhiyun 	if (ret) {
699*4882a593Smuzhiyun 		dev_err(dev, "failed to find reset-gpios property\n");
700*4882a593Smuzhiyun 		return ret;
701*4882a593Smuzhiyun 	}
702*4882a593Smuzhiyun 
703*4882a593Smuzhiyun 	ret = reset_get_bulk(dev, &priv->rsts);
704*4882a593Smuzhiyun 	if (ret) {
705*4882a593Smuzhiyun 		dev_err(dev, "Can't get reset: %d\n", ret);
706*4882a593Smuzhiyun 		return ret;
707*4882a593Smuzhiyun 	}
708*4882a593Smuzhiyun 
709*4882a593Smuzhiyun 	ret = clk_get_bulk(dev, &priv->clks);
710*4882a593Smuzhiyun 	if (ret) {
711*4882a593Smuzhiyun 		dev_err(dev, "Can't get clock: %d\n", ret);
712*4882a593Smuzhiyun 		return ret;
713*4882a593Smuzhiyun 	}
714*4882a593Smuzhiyun 
715*4882a593Smuzhiyun 	ret = device_get_supply_regulator(dev, "vpcie3v3-supply",
716*4882a593Smuzhiyun 					  &priv->vpcie3v3);
717*4882a593Smuzhiyun 	if (ret && ret != -ENOENT) {
718*4882a593Smuzhiyun 		dev_err(dev, "failed to get vpcie3v3 supply (ret=%d)\n", ret);
719*4882a593Smuzhiyun 		return ret;
720*4882a593Smuzhiyun 	}
721*4882a593Smuzhiyun 
722*4882a593Smuzhiyun 	ret = generic_phy_get_by_index(dev, 0, &priv->phy);
723*4882a593Smuzhiyun 	if (ret) {
724*4882a593Smuzhiyun 		dev_err(dev, "failed to get pcie phy (ret=%d)\n", ret);
725*4882a593Smuzhiyun 		return ret;
726*4882a593Smuzhiyun 	}
727*4882a593Smuzhiyun 
728*4882a593Smuzhiyun 	if (dev_read_bool(dev, "rockchip,bifurcation"))
729*4882a593Smuzhiyun 		priv->is_bifurcation = true;
730*4882a593Smuzhiyun 
731*4882a593Smuzhiyun 	ret = ofnode_read_u32(dev->node, "max-link-speed", &max_link_speed);
732*4882a593Smuzhiyun 	if (ret < 0 || max_link_speed > 4)
733*4882a593Smuzhiyun 		priv->gen = 0;
734*4882a593Smuzhiyun 	else
735*4882a593Smuzhiyun 		priv->gen = max_link_speed;
736*4882a593Smuzhiyun 
737*4882a593Smuzhiyun 	return 0;
738*4882a593Smuzhiyun }
739*4882a593Smuzhiyun 
rockchip_pcie_probe(struct udevice * dev)740*4882a593Smuzhiyun static int rockchip_pcie_probe(struct udevice *dev)
741*4882a593Smuzhiyun {
742*4882a593Smuzhiyun 	struct rk_pcie *priv = dev_get_priv(dev);
743*4882a593Smuzhiyun 	struct udevice *ctlr = pci_get_controller(dev);
744*4882a593Smuzhiyun 	struct pci_controller *hose = dev_get_uclass_priv(ctlr);
745*4882a593Smuzhiyun 	int ret;
746*4882a593Smuzhiyun 
747*4882a593Smuzhiyun 	priv->first_busno = dev->seq;
748*4882a593Smuzhiyun 	priv->dev = dev;
749*4882a593Smuzhiyun 
750*4882a593Smuzhiyun 	ret = rockchip_pcie_parse_dt(dev);
751*4882a593Smuzhiyun 	if (ret)
752*4882a593Smuzhiyun 		return ret;
753*4882a593Smuzhiyun 
754*4882a593Smuzhiyun 	ret = rockchip_pcie_init_port(dev);
755*4882a593Smuzhiyun 	if (ret)
756*4882a593Smuzhiyun 		return ret;
757*4882a593Smuzhiyun 
758*4882a593Smuzhiyun 	dev_info(dev, "PCIE-%d: Link up (Gen%d-x%d, Bus%d)\n",
759*4882a593Smuzhiyun 		 dev->seq, rk_pcie_get_link_speed(priv),
760*4882a593Smuzhiyun 		 rk_pcie_get_link_width(priv),
761*4882a593Smuzhiyun 		 hose->first_busno);
762*4882a593Smuzhiyun 
763*4882a593Smuzhiyun 	for (ret = 0; ret < hose->region_count; ret++) {
764*4882a593Smuzhiyun 		if (hose->regions[ret].flags == PCI_REGION_IO) {
765*4882a593Smuzhiyun 			priv->io.phys_start = hose->regions[ret].phys_start; /* IO base */
766*4882a593Smuzhiyun 			priv->io.bus_start  = hose->regions[ret].bus_start;  /* IO_bus_addr */
767*4882a593Smuzhiyun 			priv->io.size       = hose->regions[ret].size;      /* IO size */
768*4882a593Smuzhiyun 		} else if (hose->regions[ret].flags == PCI_REGION_MEM) {
769*4882a593Smuzhiyun 			priv->mem.phys_start = hose->regions[ret].phys_start; /* MEM base */
770*4882a593Smuzhiyun 			priv->mem.bus_start  = hose->regions[ret].bus_start;  /* MEM_bus_addr */
771*4882a593Smuzhiyun 			priv->mem.size	     = hose->regions[ret].size;	    /* MEM size */
772*4882a593Smuzhiyun 		} else if (hose->regions[ret].flags == PCI_REGION_SYS_MEMORY) {
773*4882a593Smuzhiyun 			priv->cfg_base = (void *)(priv->io.phys_start - priv->io.size);
774*4882a593Smuzhiyun 			priv->cfg_size = priv->io.size;
775*4882a593Smuzhiyun 		} else {
776*4882a593Smuzhiyun 			dev_err(dev, "invalid flags type!\n");
777*4882a593Smuzhiyun 		}
778*4882a593Smuzhiyun 	}
779*4882a593Smuzhiyun 
780*4882a593Smuzhiyun 	dev_dbg(dev, "Config space: [0x%p - 0x%p, size 0x%llx]\n",
781*4882a593Smuzhiyun 		priv->cfg_base, priv->cfg_base + priv->cfg_size,
782*4882a593Smuzhiyun 		priv->cfg_size);
783*4882a593Smuzhiyun 
784*4882a593Smuzhiyun 	dev_dbg(dev, "IO space: [0x%llx - 0x%llx, size 0x%x]\n",
785*4882a593Smuzhiyun 		priv->io.phys_start, priv->io.phys_start + priv->io.size,
786*4882a593Smuzhiyun 		priv->io.size);
787*4882a593Smuzhiyun 
788*4882a593Smuzhiyun 	dev_dbg(dev, "IO bus:   [0x%x - 0x%x, size 0x%x]\n",
789*4882a593Smuzhiyun 		priv->io.bus_start, priv->io.bus_start + priv->io.size,
790*4882a593Smuzhiyun 		priv->io.size);
791*4882a593Smuzhiyun 
792*4882a593Smuzhiyun 	dev_dbg(dev, "MEM space: [0x%llx - 0x%llx, size 0x%x]\n",
793*4882a593Smuzhiyun 		priv->mem.phys_start, priv->mem.phys_start + priv->mem.size,
794*4882a593Smuzhiyun 		priv->mem.size);
795*4882a593Smuzhiyun 
796*4882a593Smuzhiyun 	dev_dbg(dev, "MEM bus:   [0x%x - 0x%x, size 0x%x]\n",
797*4882a593Smuzhiyun 		priv->mem.bus_start, priv->mem.bus_start + priv->mem.size,
798*4882a593Smuzhiyun 		priv->mem.size);
799*4882a593Smuzhiyun 
800*4882a593Smuzhiyun 	rk_pcie_prog_outbound_atu_unroll(priv, PCIE_ATU_REGION_INDEX0,
801*4882a593Smuzhiyun 					 PCIE_ATU_TYPE_MEM,
802*4882a593Smuzhiyun 					 priv->mem.phys_start,
803*4882a593Smuzhiyun 					 priv->mem.bus_start, priv->mem.size);
804*4882a593Smuzhiyun 	return 0;
805*4882a593Smuzhiyun }
806*4882a593Smuzhiyun 
807*4882a593Smuzhiyun static const struct dm_pci_ops rockchip_pcie_ops = {
808*4882a593Smuzhiyun 	.read_config	= rockchip_pcie_rd_conf,
809*4882a593Smuzhiyun 	.write_config	= rockchip_pcie_wr_conf,
810*4882a593Smuzhiyun };
811*4882a593Smuzhiyun 
812*4882a593Smuzhiyun static const struct udevice_id rockchip_pcie_ids[] = {
813*4882a593Smuzhiyun 	{ .compatible = "rockchip,rk3528-pcie" },
814*4882a593Smuzhiyun 	{ .compatible = "rockchip,rk3562-pcie" },
815*4882a593Smuzhiyun 	{ .compatible = "rockchip,rk3568-pcie" },
816*4882a593Smuzhiyun 	{ .compatible = "rockchip,rk3588-pcie" },
817*4882a593Smuzhiyun 	{ }
818*4882a593Smuzhiyun };
819*4882a593Smuzhiyun 
820*4882a593Smuzhiyun U_BOOT_DRIVER(rockchip_pcie) = {
821*4882a593Smuzhiyun 	.name			= "pcie_dw_rockchip",
822*4882a593Smuzhiyun 	.id			= UCLASS_PCI,
823*4882a593Smuzhiyun 	.of_match		= rockchip_pcie_ids,
824*4882a593Smuzhiyun 	.ops			= &rockchip_pcie_ops,
825*4882a593Smuzhiyun 	.probe			= rockchip_pcie_probe,
826*4882a593Smuzhiyun 	.priv_auto_alloc_size	= sizeof(struct rk_pcie),
827*4882a593Smuzhiyun };
828