1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * (C) Copyright 2017 Theobroma Systems Design und Consulting GmbH
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * SPDX-License-Identifier: GPL-2.0
5*4882a593Smuzhiyun */
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #include <common.h>
8*4882a593Smuzhiyun #include <clk.h>
9*4882a593Smuzhiyun #include <dm.h>
10*4882a593Smuzhiyun #include <dt-bindings/memory/rk3368-dmc.h>
11*4882a593Smuzhiyun #include <dt-structs.h>
12*4882a593Smuzhiyun #include <ram.h>
13*4882a593Smuzhiyun #include <regmap.h>
14*4882a593Smuzhiyun #include <syscon.h>
15*4882a593Smuzhiyun #include <asm/io.h>
16*4882a593Smuzhiyun #include <asm/arch/clock.h>
17*4882a593Smuzhiyun #include <asm/arch/cru_rk3368.h>
18*4882a593Smuzhiyun #include <asm/arch/grf_rk3368.h>
19*4882a593Smuzhiyun #include <asm/arch/ddr_rk3368.h>
20*4882a593Smuzhiyun #include <asm/arch/sdram_rk3288.h>
21*4882a593Smuzhiyun #include <asm/arch/sdram.h>
22*4882a593Smuzhiyun
23*4882a593Smuzhiyun DECLARE_GLOBAL_DATA_PTR;
24*4882a593Smuzhiyun
25*4882a593Smuzhiyun struct dram_info {
26*4882a593Smuzhiyun struct ram_info info;
27*4882a593Smuzhiyun struct clk ddr_clk;
28*4882a593Smuzhiyun struct rk3368_cru *cru;
29*4882a593Smuzhiyun struct rk3368_grf *grf;
30*4882a593Smuzhiyun struct rk3368_ddr_pctl *pctl;
31*4882a593Smuzhiyun struct rk3368_ddrphy *phy;
32*4882a593Smuzhiyun struct rk3368_pmu_grf *pmugrf;
33*4882a593Smuzhiyun struct rk3368_msch *msch;
34*4882a593Smuzhiyun };
35*4882a593Smuzhiyun
36*4882a593Smuzhiyun struct rk3368_sdram_params {
37*4882a593Smuzhiyun #if CONFIG_IS_ENABLED(OF_PLATDATA)
38*4882a593Smuzhiyun struct dtd_rockchip_rk3368_dmc of_plat;
39*4882a593Smuzhiyun #endif
40*4882a593Smuzhiyun struct rk3288_sdram_pctl_timing pctl_timing;
41*4882a593Smuzhiyun u32 trefi_mem_ddr3;
42*4882a593Smuzhiyun struct rk3288_sdram_channel chan;
43*4882a593Smuzhiyun struct regmap *map;
44*4882a593Smuzhiyun u32 ddr_freq;
45*4882a593Smuzhiyun u32 memory_schedule;
46*4882a593Smuzhiyun u32 ddr_speed_bin;
47*4882a593Smuzhiyun u32 tfaw_mult;
48*4882a593Smuzhiyun };
49*4882a593Smuzhiyun
50*4882a593Smuzhiyun /* PTCL bits */
51*4882a593Smuzhiyun enum {
52*4882a593Smuzhiyun /* PCTL_DFISTCFG0 */
53*4882a593Smuzhiyun DFI_INIT_START = BIT(0),
54*4882a593Smuzhiyun DFI_DATA_BYTE_DISABLE_EN = BIT(2),
55*4882a593Smuzhiyun
56*4882a593Smuzhiyun /* PCTL_DFISTCFG1 */
57*4882a593Smuzhiyun DFI_DRAM_CLK_SR_EN = BIT(0),
58*4882a593Smuzhiyun DFI_DRAM_CLK_DPD_EN = BIT(1),
59*4882a593Smuzhiyun ODT_LEN_BL8_W_SHIFT = 16,
60*4882a593Smuzhiyun
61*4882a593Smuzhiyun /* PCTL_DFISTCFG2 */
62*4882a593Smuzhiyun DFI_PARITY_INTR_EN = BIT(0),
63*4882a593Smuzhiyun DFI_PARITY_EN = BIT(1),
64*4882a593Smuzhiyun
65*4882a593Smuzhiyun /* PCTL_DFILPCFG0 */
66*4882a593Smuzhiyun TLP_RESP_TIME_SHIFT = 16,
67*4882a593Smuzhiyun LP_SR_EN = BIT(8),
68*4882a593Smuzhiyun LP_PD_EN = BIT(0),
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun /* PCTL_DFIODTCFG */
71*4882a593Smuzhiyun RANK0_ODT_WRITE_SEL = BIT(3),
72*4882a593Smuzhiyun RANK1_ODT_WRITE_SEL = BIT(11),
73*4882a593Smuzhiyun
74*4882a593Smuzhiyun /* PCTL_SCFG */
75*4882a593Smuzhiyun HW_LOW_POWER_EN = BIT(0),
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun /* PCTL_MCMD */
78*4882a593Smuzhiyun START_CMD = BIT(31),
79*4882a593Smuzhiyun MCMD_RANK0 = BIT(20),
80*4882a593Smuzhiyun MCMD_RANK1 = BIT(21),
81*4882a593Smuzhiyun DESELECT_CMD = 0,
82*4882a593Smuzhiyun PREA_CMD,
83*4882a593Smuzhiyun REF_CMD,
84*4882a593Smuzhiyun MRS_CMD,
85*4882a593Smuzhiyun ZQCS_CMD,
86*4882a593Smuzhiyun ZQCL_CMD,
87*4882a593Smuzhiyun RSTL_CMD,
88*4882a593Smuzhiyun MRR_CMD = 8,
89*4882a593Smuzhiyun DPDE_CMD,
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun /* PCTL_POWCTL */
92*4882a593Smuzhiyun POWER_UP_START = BIT(0),
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun /* PCTL_POWSTAT */
95*4882a593Smuzhiyun POWER_UP_DONE = BIT(0),
96*4882a593Smuzhiyun
97*4882a593Smuzhiyun /* PCTL_SCTL */
98*4882a593Smuzhiyun INIT_STATE = 0,
99*4882a593Smuzhiyun CFG_STATE,
100*4882a593Smuzhiyun GO_STATE,
101*4882a593Smuzhiyun SLEEP_STATE,
102*4882a593Smuzhiyun WAKEUP_STATE,
103*4882a593Smuzhiyun
104*4882a593Smuzhiyun /* PCTL_STAT */
105*4882a593Smuzhiyun LP_TRIG_SHIFT = 4,
106*4882a593Smuzhiyun LP_TRIG_MASK = 7,
107*4882a593Smuzhiyun PCTL_STAT_MSK = 7,
108*4882a593Smuzhiyun INIT_MEM = 0,
109*4882a593Smuzhiyun CONFIG,
110*4882a593Smuzhiyun CONFIG_REQ,
111*4882a593Smuzhiyun ACCESS,
112*4882a593Smuzhiyun ACCESS_REQ,
113*4882a593Smuzhiyun LOW_POWER,
114*4882a593Smuzhiyun LOW_POWER_ENTRY_REQ,
115*4882a593Smuzhiyun LOW_POWER_EXIT_REQ,
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun /* PCTL_MCFG */
118*4882a593Smuzhiyun DDR2_DDR3_BL_8 = BIT(0),
119*4882a593Smuzhiyun DDR3_EN = BIT(5),
120*4882a593Smuzhiyun TFAW_TRRD_MULT4 = (0 << 18),
121*4882a593Smuzhiyun TFAW_TRRD_MULT5 = (1 << 18),
122*4882a593Smuzhiyun TFAW_TRRD_MULT6 = (2 << 18),
123*4882a593Smuzhiyun };
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun #define DDR3_MR0_WR(n) \
126*4882a593Smuzhiyun ((n <= 8) ? ((n - 4) << 9) : (((n >> 1) & 0x7) << 9))
127*4882a593Smuzhiyun #define DDR3_MR0_CL(n) \
128*4882a593Smuzhiyun ((((n - 4) & 0x7) << 4) | (((n - 4) & 0x8) >> 2))
129*4882a593Smuzhiyun #define DDR3_MR0_BL8 \
130*4882a593Smuzhiyun (0 << 0)
131*4882a593Smuzhiyun #define DDR3_MR0_DLL_RESET \
132*4882a593Smuzhiyun (1 << 8)
133*4882a593Smuzhiyun #define DDR3_MR1_RTT120OHM \
134*4882a593Smuzhiyun ((0 << 9) | (1 << 6) | (0 << 2))
135*4882a593Smuzhiyun #define DDR3_MR2_TWL(n) \
136*4882a593Smuzhiyun (((n - 5) & 0x7) << 3)
137*4882a593Smuzhiyun
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun #ifdef CONFIG_TPL_BUILD
140*4882a593Smuzhiyun
ddr_set_noc_spr_err_stall(struct rk3368_grf * grf,bool enable)141*4882a593Smuzhiyun static void ddr_set_noc_spr_err_stall(struct rk3368_grf *grf, bool enable)
142*4882a593Smuzhiyun {
143*4882a593Smuzhiyun if (enable)
144*4882a593Smuzhiyun rk_setreg(&grf->ddrc0_con0, NOC_RSP_ERR_STALL);
145*4882a593Smuzhiyun else
146*4882a593Smuzhiyun rk_clrreg(&grf->ddrc0_con0, NOC_RSP_ERR_STALL);
147*4882a593Smuzhiyun }
148*4882a593Smuzhiyun
ddr_set_ddr3_mode(struct rk3368_grf * grf,bool ddr3_mode)149*4882a593Smuzhiyun static void ddr_set_ddr3_mode(struct rk3368_grf *grf, bool ddr3_mode)
150*4882a593Smuzhiyun {
151*4882a593Smuzhiyun if (ddr3_mode)
152*4882a593Smuzhiyun rk_setreg(&grf->ddrc0_con0, MSCH0_MAINDDR3_DDR3);
153*4882a593Smuzhiyun else
154*4882a593Smuzhiyun rk_clrreg(&grf->ddrc0_con0, MSCH0_MAINDDR3_DDR3);
155*4882a593Smuzhiyun }
156*4882a593Smuzhiyun
ddrphy_config(struct rk3368_ddrphy * phy,u32 tcl,u32 tal,u32 tcwl)157*4882a593Smuzhiyun static void ddrphy_config(struct rk3368_ddrphy *phy,
158*4882a593Smuzhiyun u32 tcl, u32 tal, u32 tcwl)
159*4882a593Smuzhiyun {
160*4882a593Smuzhiyun int i;
161*4882a593Smuzhiyun
162*4882a593Smuzhiyun /* Set to DDR3 mode */
163*4882a593Smuzhiyun clrsetbits_le32(&phy->reg[1], 0x3, 0x0);
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun /* DDRPHY_REGB: CL, AL */
166*4882a593Smuzhiyun clrsetbits_le32(&phy->reg[0xb], 0xff, tcl << 4 | tal);
167*4882a593Smuzhiyun /* DDRPHY_REGC: CWL */
168*4882a593Smuzhiyun clrsetbits_le32(&phy->reg[0xc], 0x0f, tcwl);
169*4882a593Smuzhiyun
170*4882a593Smuzhiyun /* Update drive-strength */
171*4882a593Smuzhiyun writel(0xcc, &phy->reg[0x11]);
172*4882a593Smuzhiyun writel(0xaa, &phy->reg[0x16]);
173*4882a593Smuzhiyun /*
174*4882a593Smuzhiyun * Update NRCOMP/PRCOMP for all 4 channels (for details of all
175*4882a593Smuzhiyun * affected registers refer to the documentation of DDRPHY_REG20
176*4882a593Smuzhiyun * and DDRPHY_REG21 in the RK3368 TRM.
177*4882a593Smuzhiyun */
178*4882a593Smuzhiyun for (i = 0; i < 4; ++i) {
179*4882a593Smuzhiyun writel(0xcc, &phy->reg[0x20 + i * 0x10]);
180*4882a593Smuzhiyun writel(0x44, &phy->reg[0x21 + i * 0x10]);
181*4882a593Smuzhiyun }
182*4882a593Smuzhiyun
183*4882a593Smuzhiyun /* Enable write-leveling calibration bypass */
184*4882a593Smuzhiyun setbits_le32(&phy->reg[2], BIT(3));
185*4882a593Smuzhiyun }
186*4882a593Smuzhiyun
copy_to_reg(u32 * dest,const u32 * src,u32 n)187*4882a593Smuzhiyun static void copy_to_reg(u32 *dest, const u32 *src, u32 n)
188*4882a593Smuzhiyun {
189*4882a593Smuzhiyun int i;
190*4882a593Smuzhiyun
191*4882a593Smuzhiyun for (i = 0; i < n / sizeof(u32); i++)
192*4882a593Smuzhiyun writel(*src++, dest++);
193*4882a593Smuzhiyun }
194*4882a593Smuzhiyun
send_command(struct rk3368_ddr_pctl * pctl,u32 rank,u32 cmd)195*4882a593Smuzhiyun static void send_command(struct rk3368_ddr_pctl *pctl, u32 rank, u32 cmd)
196*4882a593Smuzhiyun {
197*4882a593Smuzhiyun u32 mcmd = START_CMD | cmd | rank;
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun debug("%s: writing %x to MCMD\n", __func__, mcmd);
200*4882a593Smuzhiyun writel(mcmd, &pctl->mcmd);
201*4882a593Smuzhiyun while (readl(&pctl->mcmd) & START_CMD)
202*4882a593Smuzhiyun /* spin */;
203*4882a593Smuzhiyun }
204*4882a593Smuzhiyun
send_mrs(struct rk3368_ddr_pctl * pctl,u32 rank,u32 mr_num,u32 mr_data)205*4882a593Smuzhiyun static void send_mrs(struct rk3368_ddr_pctl *pctl,
206*4882a593Smuzhiyun u32 rank, u32 mr_num, u32 mr_data)
207*4882a593Smuzhiyun {
208*4882a593Smuzhiyun u32 mcmd = START_CMD | MRS_CMD | rank | (mr_num << 17) | (mr_data << 4);
209*4882a593Smuzhiyun
210*4882a593Smuzhiyun debug("%s: writing %x to MCMD\n", __func__, mcmd);
211*4882a593Smuzhiyun writel(mcmd, &pctl->mcmd);
212*4882a593Smuzhiyun while (readl(&pctl->mcmd) & START_CMD)
213*4882a593Smuzhiyun /* spin */;
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun
memory_init(struct rk3368_ddr_pctl * pctl,struct rk3368_sdram_params * params)216*4882a593Smuzhiyun static int memory_init(struct rk3368_ddr_pctl *pctl,
217*4882a593Smuzhiyun struct rk3368_sdram_params *params)
218*4882a593Smuzhiyun {
219*4882a593Smuzhiyun u32 mr[4];
220*4882a593Smuzhiyun const ulong timeout_ms = 500;
221*4882a593Smuzhiyun ulong tmp;
222*4882a593Smuzhiyun
223*4882a593Smuzhiyun /*
224*4882a593Smuzhiyun * Power up DRAM by DDR_PCTL_POWCTL[0] register of PCTL and
225*4882a593Smuzhiyun * wait power up DRAM finish with DDR_PCTL_POWSTAT[0] register
226*4882a593Smuzhiyun * of PCTL.
227*4882a593Smuzhiyun */
228*4882a593Smuzhiyun writel(POWER_UP_START, &pctl->powctl);
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun tmp = get_timer(0);
231*4882a593Smuzhiyun do {
232*4882a593Smuzhiyun if (get_timer(tmp) > timeout_ms) {
233*4882a593Smuzhiyun pr_err("%s: POWER_UP_START did not complete in %ld ms\n",
234*4882a593Smuzhiyun __func__, timeout_ms);
235*4882a593Smuzhiyun return -ETIME;
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun } while (!(readl(&pctl->powstat) & POWER_UP_DONE));
238*4882a593Smuzhiyun
239*4882a593Smuzhiyun /* Configure MR0 through MR3 */
240*4882a593Smuzhiyun mr[0] = DDR3_MR0_WR(params->pctl_timing.twr) |
241*4882a593Smuzhiyun DDR3_MR0_CL(params->pctl_timing.tcl) |
242*4882a593Smuzhiyun DDR3_MR0_DLL_RESET;
243*4882a593Smuzhiyun mr[1] = DDR3_MR1_RTT120OHM;
244*4882a593Smuzhiyun mr[2] = DDR3_MR2_TWL(params->pctl_timing.tcwl);
245*4882a593Smuzhiyun mr[3] = 0;
246*4882a593Smuzhiyun
247*4882a593Smuzhiyun /*
248*4882a593Smuzhiyun * Also see RK3368 Technical Reference Manual:
249*4882a593Smuzhiyun * "16.6.2 Initialization (DDR3 Initialization Sequence)"
250*4882a593Smuzhiyun */
251*4882a593Smuzhiyun send_command(pctl, MCMD_RANK0 | MCMD_RANK1, DESELECT_CMD);
252*4882a593Smuzhiyun udelay(1);
253*4882a593Smuzhiyun send_command(pctl, MCMD_RANK0 | MCMD_RANK1, PREA_CMD);
254*4882a593Smuzhiyun send_mrs(pctl, MCMD_RANK0 | MCMD_RANK1, 2, mr[2]);
255*4882a593Smuzhiyun send_mrs(pctl, MCMD_RANK0 | MCMD_RANK1, 3, mr[3]);
256*4882a593Smuzhiyun send_mrs(pctl, MCMD_RANK0 | MCMD_RANK1, 1, mr[1]);
257*4882a593Smuzhiyun send_mrs(pctl, MCMD_RANK0 | MCMD_RANK1, 0, mr[0]);
258*4882a593Smuzhiyun send_command(pctl, MCMD_RANK0 | MCMD_RANK1, ZQCL_CMD);
259*4882a593Smuzhiyun
260*4882a593Smuzhiyun return 0;
261*4882a593Smuzhiyun }
262*4882a593Smuzhiyun
move_to_config_state(struct rk3368_ddr_pctl * pctl)263*4882a593Smuzhiyun static void move_to_config_state(struct rk3368_ddr_pctl *pctl)
264*4882a593Smuzhiyun {
265*4882a593Smuzhiyun /*
266*4882a593Smuzhiyun * Also see RK3368 Technical Reference Manual:
267*4882a593Smuzhiyun * "16.6.1 State transition of PCTL (Moving to Config State)"
268*4882a593Smuzhiyun */
269*4882a593Smuzhiyun u32 state = readl(&pctl->stat) & PCTL_STAT_MSK;
270*4882a593Smuzhiyun
271*4882a593Smuzhiyun switch (state) {
272*4882a593Smuzhiyun case LOW_POWER:
273*4882a593Smuzhiyun writel(WAKEUP_STATE, &pctl->sctl);
274*4882a593Smuzhiyun while ((readl(&pctl->stat) & PCTL_STAT_MSK) != ACCESS)
275*4882a593Smuzhiyun /* spin */;
276*4882a593Smuzhiyun
277*4882a593Smuzhiyun /* fall-through */
278*4882a593Smuzhiyun case ACCESS:
279*4882a593Smuzhiyun case INIT_MEM:
280*4882a593Smuzhiyun writel(CFG_STATE, &pctl->sctl);
281*4882a593Smuzhiyun while ((readl(&pctl->stat) & PCTL_STAT_MSK) != CONFIG)
282*4882a593Smuzhiyun /* spin */;
283*4882a593Smuzhiyun break;
284*4882a593Smuzhiyun
285*4882a593Smuzhiyun case CONFIG:
286*4882a593Smuzhiyun return;
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun default:
289*4882a593Smuzhiyun break;
290*4882a593Smuzhiyun }
291*4882a593Smuzhiyun }
292*4882a593Smuzhiyun
move_to_access_state(struct rk3368_ddr_pctl * pctl)293*4882a593Smuzhiyun static void move_to_access_state(struct rk3368_ddr_pctl *pctl)
294*4882a593Smuzhiyun {
295*4882a593Smuzhiyun /*
296*4882a593Smuzhiyun * Also see RK3368 Technical Reference Manual:
297*4882a593Smuzhiyun * "16.6.1 State transition of PCTL (Moving to Access State)"
298*4882a593Smuzhiyun */
299*4882a593Smuzhiyun u32 state = readl(&pctl->stat) & PCTL_STAT_MSK;
300*4882a593Smuzhiyun
301*4882a593Smuzhiyun switch (state) {
302*4882a593Smuzhiyun case LOW_POWER:
303*4882a593Smuzhiyun if (((readl(&pctl->stat) >> LP_TRIG_SHIFT) &
304*4882a593Smuzhiyun LP_TRIG_MASK) == 1)
305*4882a593Smuzhiyun return;
306*4882a593Smuzhiyun
307*4882a593Smuzhiyun writel(WAKEUP_STATE, &pctl->sctl);
308*4882a593Smuzhiyun while ((readl(&pctl->stat) & PCTL_STAT_MSK) != ACCESS)
309*4882a593Smuzhiyun /* spin */;
310*4882a593Smuzhiyun
311*4882a593Smuzhiyun /* fall-through */
312*4882a593Smuzhiyun case INIT_MEM:
313*4882a593Smuzhiyun writel(CFG_STATE, &pctl->sctl);
314*4882a593Smuzhiyun while ((readl(&pctl->stat) & PCTL_STAT_MSK) != CONFIG)
315*4882a593Smuzhiyun /* spin */;
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun /* fall-through */
318*4882a593Smuzhiyun case CONFIG:
319*4882a593Smuzhiyun writel(GO_STATE, &pctl->sctl);
320*4882a593Smuzhiyun while ((readl(&pctl->stat) & PCTL_STAT_MSK) == CONFIG)
321*4882a593Smuzhiyun /* spin */;
322*4882a593Smuzhiyun break;
323*4882a593Smuzhiyun
324*4882a593Smuzhiyun case ACCESS:
325*4882a593Smuzhiyun return;
326*4882a593Smuzhiyun
327*4882a593Smuzhiyun default:
328*4882a593Smuzhiyun break;
329*4882a593Smuzhiyun }
330*4882a593Smuzhiyun }
331*4882a593Smuzhiyun
ddrctl_reset(struct rk3368_cru * cru)332*4882a593Smuzhiyun static void ddrctl_reset(struct rk3368_cru *cru)
333*4882a593Smuzhiyun {
334*4882a593Smuzhiyun const u32 ctl_reset = BIT(3) | BIT(2);
335*4882a593Smuzhiyun const u32 phy_reset = BIT(1) | BIT(0);
336*4882a593Smuzhiyun
337*4882a593Smuzhiyun /*
338*4882a593Smuzhiyun * The PHY reset should be released before the PCTL reset.
339*4882a593Smuzhiyun *
340*4882a593Smuzhiyun * Note that the following sequence (including the number of
341*4882a593Smuzhiyun * us to delay between releasing the PHY and PCTL reset) has
342*4882a593Smuzhiyun * been adapted per feedback received from Rockchips, so do
343*4882a593Smuzhiyun * not try to optimise.
344*4882a593Smuzhiyun */
345*4882a593Smuzhiyun rk_setreg(&cru->softrst_con[10], ctl_reset | phy_reset);
346*4882a593Smuzhiyun udelay(1);
347*4882a593Smuzhiyun rk_clrreg(&cru->softrst_con[10], phy_reset);
348*4882a593Smuzhiyun udelay(5);
349*4882a593Smuzhiyun rk_clrreg(&cru->softrst_con[10], ctl_reset);
350*4882a593Smuzhiyun }
351*4882a593Smuzhiyun
ddrphy_reset(struct rk3368_ddrphy * ddrphy)352*4882a593Smuzhiyun static void ddrphy_reset(struct rk3368_ddrphy *ddrphy)
353*4882a593Smuzhiyun {
354*4882a593Smuzhiyun /*
355*4882a593Smuzhiyun * The analog part of the PHY should be release at least 1000
356*4882a593Smuzhiyun * DRAM cycles before the digital part of the PHY (waiting for
357*4882a593Smuzhiyun * 5us will ensure this for a DRAM clock as low as 200MHz).
358*4882a593Smuzhiyun */
359*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0], BIT(3) | BIT(2));
360*4882a593Smuzhiyun udelay(1);
361*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0], BIT(2));
362*4882a593Smuzhiyun udelay(5);
363*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0], BIT(3));
364*4882a593Smuzhiyun }
365*4882a593Smuzhiyun
ddrphy_config_delays(struct rk3368_ddrphy * ddrphy,u32 freq)366*4882a593Smuzhiyun static void ddrphy_config_delays(struct rk3368_ddrphy *ddrphy, u32 freq)
367*4882a593Smuzhiyun {
368*4882a593Smuzhiyun u32 dqs_dll_delay;
369*4882a593Smuzhiyun
370*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0x13], BIT(4));
371*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0x14], BIT(3));
372*4882a593Smuzhiyun
373*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0x26], BIT(4));
374*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0x27], BIT(3));
375*4882a593Smuzhiyun
376*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0x36], BIT(4));
377*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0x37], BIT(3));
378*4882a593Smuzhiyun
379*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0x46], BIT(4));
380*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0x47], BIT(3));
381*4882a593Smuzhiyun
382*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0x56], BIT(4));
383*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0x57], BIT(3));
384*4882a593Smuzhiyun
385*4882a593Smuzhiyun if (freq <= 400000000)
386*4882a593Smuzhiyun setbits_le32(&ddrphy->reg[0xa4], 0x1f);
387*4882a593Smuzhiyun else
388*4882a593Smuzhiyun clrbits_le32(&ddrphy->reg[0xa4], 0x1f);
389*4882a593Smuzhiyun
390*4882a593Smuzhiyun if (freq < 681000000)
391*4882a593Smuzhiyun dqs_dll_delay = 3; /* 67.5 degree delay */
392*4882a593Smuzhiyun else
393*4882a593Smuzhiyun dqs_dll_delay = 2; /* 45 degree delay */
394*4882a593Smuzhiyun
395*4882a593Smuzhiyun writel(dqs_dll_delay, &ddrphy->reg[0x28]);
396*4882a593Smuzhiyun writel(dqs_dll_delay, &ddrphy->reg[0x38]);
397*4882a593Smuzhiyun writel(dqs_dll_delay, &ddrphy->reg[0x48]);
398*4882a593Smuzhiyun writel(dqs_dll_delay, &ddrphy->reg[0x58]);
399*4882a593Smuzhiyun }
400*4882a593Smuzhiyun
dfi_cfg(struct rk3368_ddr_pctl * pctl)401*4882a593Smuzhiyun static int dfi_cfg(struct rk3368_ddr_pctl *pctl)
402*4882a593Smuzhiyun {
403*4882a593Smuzhiyun const ulong timeout_ms = 200;
404*4882a593Smuzhiyun ulong tmp;
405*4882a593Smuzhiyun
406*4882a593Smuzhiyun writel(DFI_DATA_BYTE_DISABLE_EN, &pctl->dfistcfg0);
407*4882a593Smuzhiyun
408*4882a593Smuzhiyun writel(DFI_DRAM_CLK_SR_EN | DFI_DRAM_CLK_DPD_EN,
409*4882a593Smuzhiyun &pctl->dfistcfg1);
410*4882a593Smuzhiyun writel(DFI_PARITY_INTR_EN | DFI_PARITY_EN, &pctl->dfistcfg2);
411*4882a593Smuzhiyun writel(7 << TLP_RESP_TIME_SHIFT | LP_SR_EN | LP_PD_EN,
412*4882a593Smuzhiyun &pctl->dfilpcfg0);
413*4882a593Smuzhiyun
414*4882a593Smuzhiyun writel(1, &pctl->dfitphyupdtype0);
415*4882a593Smuzhiyun
416*4882a593Smuzhiyun writel(0x1f, &pctl->dfitphyrdlat);
417*4882a593Smuzhiyun writel(0, &pctl->dfitphywrdata);
418*4882a593Smuzhiyun writel(0, &pctl->dfiupdcfg); /* phyupd and ctrlupd disabled */
419*4882a593Smuzhiyun
420*4882a593Smuzhiyun setbits_le32(&pctl->dfistcfg0, DFI_INIT_START);
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun tmp = get_timer(0);
423*4882a593Smuzhiyun do {
424*4882a593Smuzhiyun if (get_timer(tmp) > timeout_ms) {
425*4882a593Smuzhiyun pr_err("%s: DFI init did not complete within %ld ms\n",
426*4882a593Smuzhiyun __func__, timeout_ms);
427*4882a593Smuzhiyun return -ETIME;
428*4882a593Smuzhiyun }
429*4882a593Smuzhiyun } while ((readl(&pctl->dfiststat0) & 1) == 0);
430*4882a593Smuzhiyun
431*4882a593Smuzhiyun return 0;
432*4882a593Smuzhiyun }
433*4882a593Smuzhiyun
ps_to_tCK(const u32 ps,const ulong freq)434*4882a593Smuzhiyun static inline u32 ps_to_tCK(const u32 ps, const ulong freq)
435*4882a593Smuzhiyun {
436*4882a593Smuzhiyun const ulong MHz = 1000000;
437*4882a593Smuzhiyun return DIV_ROUND_UP(ps * freq, 1000000 * MHz);
438*4882a593Smuzhiyun }
439*4882a593Smuzhiyun
ns_to_tCK(const u32 ns,const ulong freq)440*4882a593Smuzhiyun static inline u32 ns_to_tCK(const u32 ns, const ulong freq)
441*4882a593Smuzhiyun {
442*4882a593Smuzhiyun return ps_to_tCK(ns * 1000, freq);
443*4882a593Smuzhiyun }
444*4882a593Smuzhiyun
tCK_to_ps(const ulong tCK,const ulong freq)445*4882a593Smuzhiyun static inline u32 tCK_to_ps(const ulong tCK, const ulong freq)
446*4882a593Smuzhiyun {
447*4882a593Smuzhiyun const ulong MHz = 1000000;
448*4882a593Smuzhiyun return DIV_ROUND_UP(tCK * 1000000 * MHz, freq);
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun
pctl_calc_timings(struct rk3368_sdram_params * params,ulong freq)451*4882a593Smuzhiyun static int pctl_calc_timings(struct rk3368_sdram_params *params,
452*4882a593Smuzhiyun ulong freq)
453*4882a593Smuzhiyun {
454*4882a593Smuzhiyun struct rk3288_sdram_pctl_timing *pctl_timing = ¶ms->pctl_timing;
455*4882a593Smuzhiyun const ulong MHz = 1000000;
456*4882a593Smuzhiyun u32 tccd;
457*4882a593Smuzhiyun u32 tfaw_as_ps;
458*4882a593Smuzhiyun
459*4882a593Smuzhiyun if (params->ddr_speed_bin != DDR3_1600K) {
460*4882a593Smuzhiyun pr_err("%s: unimplemented DDR3 speed bin %d\n",
461*4882a593Smuzhiyun __func__, params->ddr_speed_bin);
462*4882a593Smuzhiyun return -1;
463*4882a593Smuzhiyun }
464*4882a593Smuzhiyun
465*4882a593Smuzhiyun /* PCTL is clocked at 1/2 the DRAM clock; err on the side of caution */
466*4882a593Smuzhiyun pctl_timing->togcnt1u = DIV_ROUND_UP(freq, 2 * MHz);
467*4882a593Smuzhiyun pctl_timing->togcnt100n = DIV_ROUND_UP(freq / 10, 2 * MHz);
468*4882a593Smuzhiyun
469*4882a593Smuzhiyun pctl_timing->tinit = 200; /* 200 usec */
470*4882a593Smuzhiyun pctl_timing->trsth = 500; /* 500 usec */
471*4882a593Smuzhiyun pctl_timing->trefi = 78; /* 7.8usec = 78 * 100ns */
472*4882a593Smuzhiyun params->trefi_mem_ddr3 = ns_to_tCK(pctl_timing->trefi * 100, freq);
473*4882a593Smuzhiyun
474*4882a593Smuzhiyun if (freq <= (400 * MHz)) {
475*4882a593Smuzhiyun pctl_timing->tcl = 6;
476*4882a593Smuzhiyun pctl_timing->tcwl = 10;
477*4882a593Smuzhiyun } else if (freq <= (533 * MHz)) {
478*4882a593Smuzhiyun pctl_timing->tcl = 8;
479*4882a593Smuzhiyun pctl_timing->tcwl = 6;
480*4882a593Smuzhiyun } else if (freq <= (666 * MHz)) {
481*4882a593Smuzhiyun pctl_timing->tcl = 10;
482*4882a593Smuzhiyun pctl_timing->tcwl = 7;
483*4882a593Smuzhiyun } else {
484*4882a593Smuzhiyun pctl_timing->tcl = 11;
485*4882a593Smuzhiyun pctl_timing->tcwl = 8;
486*4882a593Smuzhiyun }
487*4882a593Smuzhiyun
488*4882a593Smuzhiyun pctl_timing->tmrd = 4; /* 4 tCK (all speed bins) */
489*4882a593Smuzhiyun pctl_timing->trfc = ns_to_tCK(350, freq); /* tRFC: 350 (max) @ 8GBit */
490*4882a593Smuzhiyun pctl_timing->trp = max(4u, ps_to_tCK(13750, freq));
491*4882a593Smuzhiyun /*
492*4882a593Smuzhiyun * JESD-79:
493*4882a593Smuzhiyun * READ to WRITE Command Delay = RL + tCCD / 2 + 2tCK - WL
494*4882a593Smuzhiyun */
495*4882a593Smuzhiyun tccd = 4;
496*4882a593Smuzhiyun pctl_timing->trtw = pctl_timing->tcl + tccd/2 + 2 - pctl_timing->tcwl;
497*4882a593Smuzhiyun pctl_timing->tal = 0;
498*4882a593Smuzhiyun pctl_timing->tras = ps_to_tCK(35000, freq);
499*4882a593Smuzhiyun pctl_timing->trc = ps_to_tCK(48750, freq);
500*4882a593Smuzhiyun pctl_timing->trcd = ps_to_tCK(13750, freq);
501*4882a593Smuzhiyun pctl_timing->trrd = max(4u, ps_to_tCK(7500, freq));
502*4882a593Smuzhiyun pctl_timing->trtp = max(4u, ps_to_tCK(7500, freq));
503*4882a593Smuzhiyun pctl_timing->twr = ps_to_tCK(15000, freq);
504*4882a593Smuzhiyun /* The DDR3 mode-register does only support even values for tWR > 8. */
505*4882a593Smuzhiyun if (pctl_timing->twr > 8)
506*4882a593Smuzhiyun pctl_timing->twr = (pctl_timing->twr + 1) & ~1;
507*4882a593Smuzhiyun pctl_timing->twtr = max(4u, ps_to_tCK(7500, freq));
508*4882a593Smuzhiyun pctl_timing->texsr = 512; /* tEXSR(max) is tDLLLK */
509*4882a593Smuzhiyun pctl_timing->txp = max(3u, ps_to_tCK(6000, freq));
510*4882a593Smuzhiyun pctl_timing->txpdll = max(10u, ps_to_tCK(24000, freq));
511*4882a593Smuzhiyun pctl_timing->tzqcs = max(64u, ps_to_tCK(80000, freq));
512*4882a593Smuzhiyun pctl_timing->tzqcsi = 10000; /* as used by Rockchip */
513*4882a593Smuzhiyun pctl_timing->tdqs = 1; /* fixed for DDR3 */
514*4882a593Smuzhiyun pctl_timing->tcksre = max(5u, ps_to_tCK(10000, freq));
515*4882a593Smuzhiyun pctl_timing->tcksrx = max(5u, ps_to_tCK(10000, freq));
516*4882a593Smuzhiyun pctl_timing->tcke = max(3u, ps_to_tCK(5000, freq));
517*4882a593Smuzhiyun pctl_timing->tmod = max(12u, ps_to_tCK(15000, freq));
518*4882a593Smuzhiyun pctl_timing->trstl = ns_to_tCK(100, freq);
519*4882a593Smuzhiyun pctl_timing->tzqcl = max(256u, ps_to_tCK(320000, freq)); /* tZQoper */
520*4882a593Smuzhiyun pctl_timing->tmrr = 0;
521*4882a593Smuzhiyun pctl_timing->tckesr = pctl_timing->tcke + 1; /* JESD-79: tCKE + 1tCK */
522*4882a593Smuzhiyun pctl_timing->tdpd = 0; /* RK3368 TRM: "allowed values for DDR3: 0" */
523*4882a593Smuzhiyun
524*4882a593Smuzhiyun
525*4882a593Smuzhiyun /*
526*4882a593Smuzhiyun * The controller can represent tFAW as 4x, 5x or 6x tRRD only.
527*4882a593Smuzhiyun * We want to use the smallest multiplier that satisfies the tFAW
528*4882a593Smuzhiyun * requirements of the given speed-bin. If necessary, we stretch out
529*4882a593Smuzhiyun * tRRD to allow us to operate on a 6x multiplier for tFAW.
530*4882a593Smuzhiyun */
531*4882a593Smuzhiyun tfaw_as_ps = 40000; /* 40ns: tFAW for DDR3-1600K, 2KB page-size */
532*4882a593Smuzhiyun if (tCK_to_ps(pctl_timing->trrd * 6, freq) < tfaw_as_ps) {
533*4882a593Smuzhiyun /* If tFAW is > 6 x tRRD, we need to stretch tRRD */
534*4882a593Smuzhiyun pctl_timing->trrd = ps_to_tCK(DIV_ROUND_UP(40000, 6), freq);
535*4882a593Smuzhiyun params->tfaw_mult = TFAW_TRRD_MULT6;
536*4882a593Smuzhiyun } else if (tCK_to_ps(pctl_timing->trrd * 5, freq) < tfaw_as_ps) {
537*4882a593Smuzhiyun params->tfaw_mult = TFAW_TRRD_MULT6;
538*4882a593Smuzhiyun } else if (tCK_to_ps(pctl_timing->trrd * 4, freq) < tfaw_as_ps) {
539*4882a593Smuzhiyun params->tfaw_mult = TFAW_TRRD_MULT5;
540*4882a593Smuzhiyun } else {
541*4882a593Smuzhiyun params->tfaw_mult = TFAW_TRRD_MULT4;
542*4882a593Smuzhiyun }
543*4882a593Smuzhiyun
544*4882a593Smuzhiyun return 0;
545*4882a593Smuzhiyun }
546*4882a593Smuzhiyun
pctl_cfg(struct rk3368_ddr_pctl * pctl,struct rk3368_sdram_params * params,struct rk3368_grf * grf)547*4882a593Smuzhiyun static void pctl_cfg(struct rk3368_ddr_pctl *pctl,
548*4882a593Smuzhiyun struct rk3368_sdram_params *params,
549*4882a593Smuzhiyun struct rk3368_grf *grf)
550*4882a593Smuzhiyun {
551*4882a593Smuzhiyun /* Configure PCTL timing registers */
552*4882a593Smuzhiyun params->pctl_timing.trefi |= BIT(31); /* see PCTL_TREFI */
553*4882a593Smuzhiyun copy_to_reg(&pctl->togcnt1u, ¶ms->pctl_timing.togcnt1u,
554*4882a593Smuzhiyun sizeof(params->pctl_timing));
555*4882a593Smuzhiyun writel(params->trefi_mem_ddr3, &pctl->trefi_mem_ddr3);
556*4882a593Smuzhiyun
557*4882a593Smuzhiyun /* Set up ODT write selector and ODT write length */
558*4882a593Smuzhiyun writel((RANK0_ODT_WRITE_SEL | RANK1_ODT_WRITE_SEL), &pctl->dfiodtcfg);
559*4882a593Smuzhiyun writel(7 << ODT_LEN_BL8_W_SHIFT, &pctl->dfiodtcfg1);
560*4882a593Smuzhiyun
561*4882a593Smuzhiyun /* Set up the CL/CWL-dependent timings of DFI */
562*4882a593Smuzhiyun writel((params->pctl_timing.tcl - 1) / 2 - 1, &pctl->dfitrddataen);
563*4882a593Smuzhiyun writel((params->pctl_timing.tcwl - 1) / 2 - 1, &pctl->dfitphywrlat);
564*4882a593Smuzhiyun
565*4882a593Smuzhiyun /* DDR3 */
566*4882a593Smuzhiyun writel(params->tfaw_mult | DDR3_EN | DDR2_DDR3_BL_8, &pctl->mcfg);
567*4882a593Smuzhiyun writel(0x001c0004, &grf->ddrc0_con0);
568*4882a593Smuzhiyun
569*4882a593Smuzhiyun setbits_le32(&pctl->scfg, HW_LOW_POWER_EN);
570*4882a593Smuzhiyun }
571*4882a593Smuzhiyun
ddrphy_data_training(struct rk3368_ddr_pctl * pctl,struct rk3368_ddrphy * ddrphy)572*4882a593Smuzhiyun static int ddrphy_data_training(struct rk3368_ddr_pctl *pctl,
573*4882a593Smuzhiyun struct rk3368_ddrphy *ddrphy)
574*4882a593Smuzhiyun {
575*4882a593Smuzhiyun const u32 trefi = readl(&pctl->trefi);
576*4882a593Smuzhiyun const ulong timeout_ms = 500;
577*4882a593Smuzhiyun ulong tmp;
578*4882a593Smuzhiyun
579*4882a593Smuzhiyun /* disable auto-refresh */
580*4882a593Smuzhiyun writel(0 | BIT(31), &pctl->trefi);
581*4882a593Smuzhiyun
582*4882a593Smuzhiyun clrsetbits_le32(&ddrphy->reg[2], 0x33, 0x20);
583*4882a593Smuzhiyun clrsetbits_le32(&ddrphy->reg[2], 0x33, 0x21);
584*4882a593Smuzhiyun
585*4882a593Smuzhiyun tmp = get_timer(0);
586*4882a593Smuzhiyun do {
587*4882a593Smuzhiyun if (get_timer(tmp) > timeout_ms) {
588*4882a593Smuzhiyun pr_err("%s: did not complete within %ld ms\n",
589*4882a593Smuzhiyun __func__, timeout_ms);
590*4882a593Smuzhiyun return -ETIME;
591*4882a593Smuzhiyun }
592*4882a593Smuzhiyun } while ((readl(&ddrphy->reg[0xff]) & 0xf) != 0xf);
593*4882a593Smuzhiyun
594*4882a593Smuzhiyun send_command(pctl, MCMD_RANK0 | MCMD_RANK1, PREA_CMD);
595*4882a593Smuzhiyun clrsetbits_le32(&ddrphy->reg[2], 0x33, 0x20);
596*4882a593Smuzhiyun /* resume auto-refresh */
597*4882a593Smuzhiyun writel(trefi | BIT(31), &pctl->trefi);
598*4882a593Smuzhiyun
599*4882a593Smuzhiyun return 0;
600*4882a593Smuzhiyun }
601*4882a593Smuzhiyun
sdram_col_row_detect(struct udevice * dev)602*4882a593Smuzhiyun static int sdram_col_row_detect(struct udevice *dev)
603*4882a593Smuzhiyun {
604*4882a593Smuzhiyun struct dram_info *priv = dev_get_priv(dev);
605*4882a593Smuzhiyun struct rk3368_sdram_params *params = dev_get_platdata(dev);
606*4882a593Smuzhiyun struct rk3368_ddr_pctl *pctl = priv->pctl;
607*4882a593Smuzhiyun struct rk3368_msch *msch = priv->msch;
608*4882a593Smuzhiyun const u32 test_pattern = 0x5aa5f00f;
609*4882a593Smuzhiyun int row, col;
610*4882a593Smuzhiyun uintptr_t addr;
611*4882a593Smuzhiyun
612*4882a593Smuzhiyun move_to_config_state(pctl);
613*4882a593Smuzhiyun writel(6, &msch->ddrconf);
614*4882a593Smuzhiyun move_to_access_state(pctl);
615*4882a593Smuzhiyun
616*4882a593Smuzhiyun /* Detect col */
617*4882a593Smuzhiyun for (col = 11; col >= 9; col--) {
618*4882a593Smuzhiyun writel(0, CONFIG_SYS_SDRAM_BASE);
619*4882a593Smuzhiyun addr = CONFIG_SYS_SDRAM_BASE +
620*4882a593Smuzhiyun (1 << (col + params->chan.bw - 1));
621*4882a593Smuzhiyun writel(test_pattern, addr);
622*4882a593Smuzhiyun if ((readl(addr) == test_pattern) &&
623*4882a593Smuzhiyun (readl(CONFIG_SYS_SDRAM_BASE) == 0))
624*4882a593Smuzhiyun break;
625*4882a593Smuzhiyun }
626*4882a593Smuzhiyun
627*4882a593Smuzhiyun if (col == 8) {
628*4882a593Smuzhiyun pr_err("%s: col detect error\n", __func__);
629*4882a593Smuzhiyun return -EINVAL;
630*4882a593Smuzhiyun }
631*4882a593Smuzhiyun
632*4882a593Smuzhiyun move_to_config_state(pctl);
633*4882a593Smuzhiyun writel(15, &msch->ddrconf);
634*4882a593Smuzhiyun move_to_access_state(pctl);
635*4882a593Smuzhiyun
636*4882a593Smuzhiyun /* Detect row*/
637*4882a593Smuzhiyun for (row = 16; row >= 12; row--) {
638*4882a593Smuzhiyun writel(0, CONFIG_SYS_SDRAM_BASE);
639*4882a593Smuzhiyun addr = CONFIG_SYS_SDRAM_BASE + (1 << (row + 15 - 1));
640*4882a593Smuzhiyun writel(test_pattern, addr);
641*4882a593Smuzhiyun if ((readl(addr) == test_pattern) &&
642*4882a593Smuzhiyun (readl(CONFIG_SYS_SDRAM_BASE) == 0))
643*4882a593Smuzhiyun break;
644*4882a593Smuzhiyun }
645*4882a593Smuzhiyun
646*4882a593Smuzhiyun if (row == 11) {
647*4882a593Smuzhiyun pr_err("%s: row detect error\n", __func__);
648*4882a593Smuzhiyun return -EINVAL;
649*4882a593Smuzhiyun }
650*4882a593Smuzhiyun
651*4882a593Smuzhiyun /* Record results */
652*4882a593Smuzhiyun debug("%s: col %d, row %d\n", __func__, col, row);
653*4882a593Smuzhiyun params->chan.col = col;
654*4882a593Smuzhiyun params->chan.cs0_row = row;
655*4882a593Smuzhiyun params->chan.cs1_row = row;
656*4882a593Smuzhiyun params->chan.row_3_4 = 0;
657*4882a593Smuzhiyun
658*4882a593Smuzhiyun return 0;
659*4882a593Smuzhiyun }
660*4882a593Smuzhiyun
msch_niu_config(struct rk3368_msch * msch,struct rk3368_sdram_params * params)661*4882a593Smuzhiyun static int msch_niu_config(struct rk3368_msch *msch,
662*4882a593Smuzhiyun struct rk3368_sdram_params *params)
663*4882a593Smuzhiyun {
664*4882a593Smuzhiyun int i;
665*4882a593Smuzhiyun const u8 cols = params->chan.col - ((params->chan.bw == 2) ? 0 : 1);
666*4882a593Smuzhiyun const u8 rows = params->chan.cs0_row;
667*4882a593Smuzhiyun
668*4882a593Smuzhiyun /*
669*4882a593Smuzhiyun * The DDR address-translation table always assumes a 32bit
670*4882a593Smuzhiyun * bus and the comparison below takes care of adjusting for
671*4882a593Smuzhiyun * a 16bit bus (i.e. one column-address is consumed).
672*4882a593Smuzhiyun */
673*4882a593Smuzhiyun const struct {
674*4882a593Smuzhiyun u8 rows;
675*4882a593Smuzhiyun u8 columns;
676*4882a593Smuzhiyun u8 type;
677*4882a593Smuzhiyun } ddrconf_table[] = {
678*4882a593Smuzhiyun /*
679*4882a593Smuzhiyun * C-B-R-D patterns are first. For these we require an
680*4882a593Smuzhiyun * exact match for the columns and rows (as there's
681*4882a593Smuzhiyun * one entry per possible configuration).
682*4882a593Smuzhiyun */
683*4882a593Smuzhiyun [0] = { .rows = 13, .columns = 10, .type = DMC_MSCH_CBRD },
684*4882a593Smuzhiyun [1] = { .rows = 14, .columns = 10, .type = DMC_MSCH_CBRD },
685*4882a593Smuzhiyun [2] = { .rows = 15, .columns = 10, .type = DMC_MSCH_CBRD },
686*4882a593Smuzhiyun [3] = { .rows = 16, .columns = 10, .type = DMC_MSCH_CBRD },
687*4882a593Smuzhiyun [4] = { .rows = 14, .columns = 11, .type = DMC_MSCH_CBRD },
688*4882a593Smuzhiyun [5] = { .rows = 15, .columns = 11, .type = DMC_MSCH_CBRD },
689*4882a593Smuzhiyun [6] = { .rows = 16, .columns = 11, .type = DMC_MSCH_CBRD },
690*4882a593Smuzhiyun [7] = { .rows = 13, .columns = 9, .type = DMC_MSCH_CBRD },
691*4882a593Smuzhiyun [8] = { .rows = 14, .columns = 9, .type = DMC_MSCH_CBRD },
692*4882a593Smuzhiyun [9] = { .rows = 15, .columns = 9, .type = DMC_MSCH_CBRD },
693*4882a593Smuzhiyun [10] = { .rows = 16, .columns = 9, .type = DMC_MSCH_CBRD },
694*4882a593Smuzhiyun /*
695*4882a593Smuzhiyun * 11 through 13 are C-R-B-D patterns. These are
696*4882a593Smuzhiyun * matched for an exact number of columns and to
697*4882a593Smuzhiyun * ensure that the hardware uses at least as many rows
698*4882a593Smuzhiyun * as the pattern requires (i.e. we make sure that
699*4882a593Smuzhiyun * there's no gaps up until we hit the device/chip-select;
700*4882a593Smuzhiyun * however, these patterns can accept up to 16 rows,
701*4882a593Smuzhiyun * as the row-address continues right after the CS
702*4882a593Smuzhiyun * switching)
703*4882a593Smuzhiyun */
704*4882a593Smuzhiyun [11] = { .rows = 15, .columns = 10, .type = DMC_MSCH_CRBD },
705*4882a593Smuzhiyun [12] = { .rows = 14, .columns = 11, .type = DMC_MSCH_CRBD },
706*4882a593Smuzhiyun [13] = { .rows = 13, .columns = 10, .type = DMC_MSCH_CRBD },
707*4882a593Smuzhiyun /*
708*4882a593Smuzhiyun * 14 and 15 are catch-all variants using a C-B-D-R
709*4882a593Smuzhiyun * scheme (i.e. alternating the chip-select every time
710*4882a593Smuzhiyun * C-B overflows) and stuffing the remaining C-bits
711*4882a593Smuzhiyun * into the top. Matching needs to make sure that the
712*4882a593Smuzhiyun * number of columns is either an exact match (i.e. we
713*4882a593Smuzhiyun * can use less the the maximum number of rows) -or-
714*4882a593Smuzhiyun * that the columns exceed what is given in this table
715*4882a593Smuzhiyun * and the rows are an exact match (in which case the
716*4882a593Smuzhiyun * remaining C-bits will be stuffed onto the top after
717*4882a593Smuzhiyun * the device/chip-select switches).
718*4882a593Smuzhiyun */
719*4882a593Smuzhiyun [14] = { .rows = 16, .columns = 10, .type = DMC_MSCH_CBDR },
720*4882a593Smuzhiyun [15] = { .rows = 16, .columns = 9, .type = DMC_MSCH_CBDR },
721*4882a593Smuzhiyun };
722*4882a593Smuzhiyun
723*4882a593Smuzhiyun /*
724*4882a593Smuzhiyun * For C-B-R-D, we need an exact match (i.e. both for the number of
725*4882a593Smuzhiyun * columns and rows), while for C-B-D-R, only the the number of
726*4882a593Smuzhiyun * columns needs to match.
727*4882a593Smuzhiyun */
728*4882a593Smuzhiyun for (i = 0; i < ARRAY_SIZE(ddrconf_table); i++) {
729*4882a593Smuzhiyun bool match = false;
730*4882a593Smuzhiyun
731*4882a593Smuzhiyun /* If this entry if for a different matcher, then skip it */
732*4882a593Smuzhiyun if (ddrconf_table[i].type != params->memory_schedule)
733*4882a593Smuzhiyun continue;
734*4882a593Smuzhiyun
735*4882a593Smuzhiyun /*
736*4882a593Smuzhiyun * Match according to the rules (exact/inexact/at-least)
737*4882a593Smuzhiyun * documented in the ddrconf_table above.
738*4882a593Smuzhiyun */
739*4882a593Smuzhiyun switch (params->memory_schedule) {
740*4882a593Smuzhiyun case DMC_MSCH_CBRD:
741*4882a593Smuzhiyun match = (ddrconf_table[i].columns == cols) &&
742*4882a593Smuzhiyun (ddrconf_table[i].rows == rows);
743*4882a593Smuzhiyun break;
744*4882a593Smuzhiyun
745*4882a593Smuzhiyun case DMC_MSCH_CRBD:
746*4882a593Smuzhiyun match = (ddrconf_table[i].columns == cols) &&
747*4882a593Smuzhiyun (ddrconf_table[i].rows <= rows);
748*4882a593Smuzhiyun break;
749*4882a593Smuzhiyun
750*4882a593Smuzhiyun case DMC_MSCH_CBDR:
751*4882a593Smuzhiyun match = (ddrconf_table[i].columns == cols) ||
752*4882a593Smuzhiyun ((ddrconf_table[i].columns <= cols) &&
753*4882a593Smuzhiyun (ddrconf_table[i].rows == rows));
754*4882a593Smuzhiyun break;
755*4882a593Smuzhiyun
756*4882a593Smuzhiyun default:
757*4882a593Smuzhiyun break;
758*4882a593Smuzhiyun }
759*4882a593Smuzhiyun
760*4882a593Smuzhiyun if (match) {
761*4882a593Smuzhiyun debug("%s: setting ddrconf 0x%x\n", __func__, i);
762*4882a593Smuzhiyun writel(i, &msch->ddrconf);
763*4882a593Smuzhiyun return 0;
764*4882a593Smuzhiyun }
765*4882a593Smuzhiyun }
766*4882a593Smuzhiyun
767*4882a593Smuzhiyun pr_err("%s: ddrconf (NIU config) not found\n", __func__);
768*4882a593Smuzhiyun return -EINVAL;
769*4882a593Smuzhiyun }
770*4882a593Smuzhiyun
dram_all_config(struct udevice * dev)771*4882a593Smuzhiyun static void dram_all_config(struct udevice *dev)
772*4882a593Smuzhiyun {
773*4882a593Smuzhiyun struct dram_info *priv = dev_get_priv(dev);
774*4882a593Smuzhiyun struct rk3368_pmu_grf *pmugrf = priv->pmugrf;
775*4882a593Smuzhiyun struct rk3368_sdram_params *params = dev_get_platdata(dev);
776*4882a593Smuzhiyun const struct rk3288_sdram_channel *info = ¶ms->chan;
777*4882a593Smuzhiyun u32 sys_reg = 0;
778*4882a593Smuzhiyun const int chan = 0;
779*4882a593Smuzhiyun
780*4882a593Smuzhiyun sys_reg |= DDR3 << SYS_REG_DDRTYPE_SHIFT;
781*4882a593Smuzhiyun sys_reg |= 0 << SYS_REG_NUM_CH_SHIFT;
782*4882a593Smuzhiyun
783*4882a593Smuzhiyun sys_reg |= info->row_3_4 << SYS_REG_ROW_3_4_SHIFT(chan);
784*4882a593Smuzhiyun sys_reg |= 1 << SYS_REG_CHINFO_SHIFT(chan);
785*4882a593Smuzhiyun sys_reg |= (info->rank - 1) << SYS_REG_RANK_SHIFT(chan);
786*4882a593Smuzhiyun sys_reg |= (info->col - 9) << SYS_REG_COL_SHIFT(chan);
787*4882a593Smuzhiyun sys_reg |= info->bk == 3 ? 0 : 1 << SYS_REG_BK_SHIFT(chan);
788*4882a593Smuzhiyun sys_reg |= (info->cs0_row - 13) << SYS_REG_CS0_ROW_SHIFT(chan);
789*4882a593Smuzhiyun sys_reg |= (info->cs1_row - 13) << SYS_REG_CS1_ROW_SHIFT(chan);
790*4882a593Smuzhiyun sys_reg |= (2 >> info->bw) << SYS_REG_BW_SHIFT(chan);
791*4882a593Smuzhiyun sys_reg |= (2 >> info->dbw) << SYS_REG_DBW_SHIFT(chan);
792*4882a593Smuzhiyun
793*4882a593Smuzhiyun writel(sys_reg, &pmugrf->os_reg[2]);
794*4882a593Smuzhiyun }
795*4882a593Smuzhiyun
setup_sdram(struct udevice * dev)796*4882a593Smuzhiyun static int setup_sdram(struct udevice *dev)
797*4882a593Smuzhiyun {
798*4882a593Smuzhiyun struct dram_info *priv = dev_get_priv(dev);
799*4882a593Smuzhiyun struct rk3368_sdram_params *params = dev_get_platdata(dev);
800*4882a593Smuzhiyun
801*4882a593Smuzhiyun struct rk3368_ddr_pctl *pctl = priv->pctl;
802*4882a593Smuzhiyun struct rk3368_ddrphy *ddrphy = priv->phy;
803*4882a593Smuzhiyun struct rk3368_cru *cru = priv->cru;
804*4882a593Smuzhiyun struct rk3368_grf *grf = priv->grf;
805*4882a593Smuzhiyun struct rk3368_msch *msch = priv->msch;
806*4882a593Smuzhiyun
807*4882a593Smuzhiyun int ret;
808*4882a593Smuzhiyun
809*4882a593Smuzhiyun /* The input clock (i.e. DPLL) needs to be 2x the DRAM frequency */
810*4882a593Smuzhiyun ret = clk_set_rate(&priv->ddr_clk, 2 * params->ddr_freq);
811*4882a593Smuzhiyun if (ret < 0) {
812*4882a593Smuzhiyun debug("%s: could not set DDR clock: %d\n", __func__, ret);
813*4882a593Smuzhiyun return ret;
814*4882a593Smuzhiyun }
815*4882a593Smuzhiyun
816*4882a593Smuzhiyun /* Update the read-latency for the RK3368 */
817*4882a593Smuzhiyun writel(0x32, &msch->readlatency);
818*4882a593Smuzhiyun
819*4882a593Smuzhiyun /* Initialise the DDR PCTL and DDR PHY */
820*4882a593Smuzhiyun ddrctl_reset(cru);
821*4882a593Smuzhiyun ddrphy_reset(ddrphy);
822*4882a593Smuzhiyun ddrphy_config_delays(ddrphy, params->ddr_freq);
823*4882a593Smuzhiyun dfi_cfg(pctl);
824*4882a593Smuzhiyun /* Configure relative system information of grf_ddrc0_con0 register */
825*4882a593Smuzhiyun ddr_set_ddr3_mode(grf, true);
826*4882a593Smuzhiyun ddr_set_noc_spr_err_stall(grf, true);
827*4882a593Smuzhiyun /* Calculate timings */
828*4882a593Smuzhiyun pctl_calc_timings(params, params->ddr_freq);
829*4882a593Smuzhiyun /* Initialise the device timings in protocol controller */
830*4882a593Smuzhiyun pctl_cfg(pctl, params, grf);
831*4882a593Smuzhiyun /* Configure AL, CL ... information of PHY registers */
832*4882a593Smuzhiyun ddrphy_config(ddrphy,
833*4882a593Smuzhiyun params->pctl_timing.tcl,
834*4882a593Smuzhiyun params->pctl_timing.tal,
835*4882a593Smuzhiyun params->pctl_timing.tcwl);
836*4882a593Smuzhiyun
837*4882a593Smuzhiyun /* Initialize DRAM and configure with mode-register values */
838*4882a593Smuzhiyun ret = memory_init(pctl, params);
839*4882a593Smuzhiyun if (ret)
840*4882a593Smuzhiyun goto error;
841*4882a593Smuzhiyun
842*4882a593Smuzhiyun move_to_config_state(pctl);
843*4882a593Smuzhiyun /* Perform data-training */
844*4882a593Smuzhiyun ddrphy_data_training(pctl, ddrphy);
845*4882a593Smuzhiyun move_to_access_state(pctl);
846*4882a593Smuzhiyun
847*4882a593Smuzhiyun /* TODO(prt): could detect rank in training... */
848*4882a593Smuzhiyun #ifdef CONFIG_TARGET_EVB_PX5
849*4882a593Smuzhiyun params->chan.rank = 1;
850*4882a593Smuzhiyun #else
851*4882a593Smuzhiyun params->chan.rank = 2;
852*4882a593Smuzhiyun #endif
853*4882a593Smuzhiyun /* TODO(prt): bus width is not auto-detected (yet)... */
854*4882a593Smuzhiyun params->chan.bw = 2; /* 32bit wide bus */
855*4882a593Smuzhiyun params->chan.dbw = params->chan.dbw; /* 32bit wide bus */
856*4882a593Smuzhiyun
857*4882a593Smuzhiyun /* DDR3 is always 8 bank */
858*4882a593Smuzhiyun params->chan.bk = 3;
859*4882a593Smuzhiyun /* Detect col and row number */
860*4882a593Smuzhiyun ret = sdram_col_row_detect(dev);
861*4882a593Smuzhiyun if (ret)
862*4882a593Smuzhiyun goto error;
863*4882a593Smuzhiyun
864*4882a593Smuzhiyun /* Configure NIU DDR configuration */
865*4882a593Smuzhiyun ret = msch_niu_config(msch, params);
866*4882a593Smuzhiyun if (ret)
867*4882a593Smuzhiyun goto error;
868*4882a593Smuzhiyun
869*4882a593Smuzhiyun /* set up OS_REG to communicate w/ next stage and OS */
870*4882a593Smuzhiyun dram_all_config(dev);
871*4882a593Smuzhiyun
872*4882a593Smuzhiyun return 0;
873*4882a593Smuzhiyun
874*4882a593Smuzhiyun error:
875*4882a593Smuzhiyun printf("DRAM init failed!\n");
876*4882a593Smuzhiyun hang();
877*4882a593Smuzhiyun }
878*4882a593Smuzhiyun #endif
879*4882a593Smuzhiyun
rk3368_dmc_ofdata_to_platdata(struct udevice * dev)880*4882a593Smuzhiyun static int rk3368_dmc_ofdata_to_platdata(struct udevice *dev)
881*4882a593Smuzhiyun {
882*4882a593Smuzhiyun int ret = 0;
883*4882a593Smuzhiyun
884*4882a593Smuzhiyun #if !CONFIG_IS_ENABLED(OF_PLATDATA)
885*4882a593Smuzhiyun struct rk3368_sdram_params *plat = dev_get_platdata(dev);
886*4882a593Smuzhiyun
887*4882a593Smuzhiyun ret = regmap_init_mem(dev, &plat->map);
888*4882a593Smuzhiyun if (ret)
889*4882a593Smuzhiyun return ret;
890*4882a593Smuzhiyun #endif
891*4882a593Smuzhiyun
892*4882a593Smuzhiyun return ret;
893*4882a593Smuzhiyun }
894*4882a593Smuzhiyun
895*4882a593Smuzhiyun #if CONFIG_IS_ENABLED(OF_PLATDATA)
conv_of_platdata(struct udevice * dev)896*4882a593Smuzhiyun static int conv_of_platdata(struct udevice *dev)
897*4882a593Smuzhiyun {
898*4882a593Smuzhiyun struct rk3368_sdram_params *plat = dev_get_platdata(dev);
899*4882a593Smuzhiyun struct dtd_rockchip_rk3368_dmc *of_plat = &plat->of_plat;
900*4882a593Smuzhiyun
901*4882a593Smuzhiyun plat->ddr_freq = of_plat->rockchip_ddr_frequency;
902*4882a593Smuzhiyun plat->ddr_speed_bin = of_plat->rockchip_ddr_speed_bin;
903*4882a593Smuzhiyun plat->memory_schedule = of_plat->rockchip_memory_schedule;
904*4882a593Smuzhiyun
905*4882a593Smuzhiyun return 0;
906*4882a593Smuzhiyun }
907*4882a593Smuzhiyun #endif
908*4882a593Smuzhiyun
rk3368_dmc_probe(struct udevice * dev)909*4882a593Smuzhiyun static int rk3368_dmc_probe(struct udevice *dev)
910*4882a593Smuzhiyun {
911*4882a593Smuzhiyun #ifdef CONFIG_TPL_BUILD
912*4882a593Smuzhiyun struct rk3368_sdram_params *plat = dev_get_platdata(dev);
913*4882a593Smuzhiyun struct rk3368_ddr_pctl *pctl;
914*4882a593Smuzhiyun struct rk3368_ddrphy *ddrphy;
915*4882a593Smuzhiyun struct rk3368_cru *cru;
916*4882a593Smuzhiyun struct rk3368_grf *grf;
917*4882a593Smuzhiyun struct rk3368_msch *msch;
918*4882a593Smuzhiyun int ret;
919*4882a593Smuzhiyun struct udevice *dev_clk;
920*4882a593Smuzhiyun #endif
921*4882a593Smuzhiyun struct dram_info *priv = dev_get_priv(dev);
922*4882a593Smuzhiyun
923*4882a593Smuzhiyun #if CONFIG_IS_ENABLED(OF_PLATDATA)
924*4882a593Smuzhiyun ret = conv_of_platdata(dev);
925*4882a593Smuzhiyun if (ret)
926*4882a593Smuzhiyun return ret;
927*4882a593Smuzhiyun #endif
928*4882a593Smuzhiyun
929*4882a593Smuzhiyun priv->pmugrf = syscon_get_first_range(ROCKCHIP_SYSCON_PMUGRF);
930*4882a593Smuzhiyun debug("%s: pmugrf=%p\n", __func__, priv->pmugrf);
931*4882a593Smuzhiyun
932*4882a593Smuzhiyun #ifdef CONFIG_TPL_BUILD
933*4882a593Smuzhiyun pctl = (struct rk3368_ddr_pctl *)plat->of_plat.reg[0];
934*4882a593Smuzhiyun ddrphy = (struct rk3368_ddrphy *)plat->of_plat.reg[2];
935*4882a593Smuzhiyun msch = syscon_get_first_range(ROCKCHIP_SYSCON_MSCH);
936*4882a593Smuzhiyun grf = syscon_get_first_range(ROCKCHIP_SYSCON_GRF);
937*4882a593Smuzhiyun
938*4882a593Smuzhiyun priv->pctl = pctl;
939*4882a593Smuzhiyun priv->phy = ddrphy;
940*4882a593Smuzhiyun priv->msch = msch;
941*4882a593Smuzhiyun priv->grf = grf;
942*4882a593Smuzhiyun
943*4882a593Smuzhiyun ret = rockchip_get_clk(&dev_clk);
944*4882a593Smuzhiyun if (ret)
945*4882a593Smuzhiyun return ret;
946*4882a593Smuzhiyun priv->ddr_clk.id = CLK_DDR;
947*4882a593Smuzhiyun ret = clk_request(dev_clk, &priv->ddr_clk);
948*4882a593Smuzhiyun if (ret)
949*4882a593Smuzhiyun return ret;
950*4882a593Smuzhiyun
951*4882a593Smuzhiyun cru = rockchip_get_cru();
952*4882a593Smuzhiyun priv->cru = cru;
953*4882a593Smuzhiyun if (IS_ERR(priv->cru))
954*4882a593Smuzhiyun return PTR_ERR(priv->cru);
955*4882a593Smuzhiyun
956*4882a593Smuzhiyun ret = setup_sdram(dev);
957*4882a593Smuzhiyun if (ret)
958*4882a593Smuzhiyun return ret;
959*4882a593Smuzhiyun #endif
960*4882a593Smuzhiyun
961*4882a593Smuzhiyun priv->info.base = 0;
962*4882a593Smuzhiyun priv->info.size =
963*4882a593Smuzhiyun rockchip_sdram_size((phys_addr_t)&priv->pmugrf->os_reg[2]);
964*4882a593Smuzhiyun
965*4882a593Smuzhiyun /*
966*4882a593Smuzhiyun * we use the 0x00000000~0xfdffffff space since 0xff000000~0xffffffff
967*4882a593Smuzhiyun * is SoC register space (i.e. reserved), and 0xfe000000~0xfeffffff is
968*4882a593Smuzhiyun * inaccessible for some IP controller.
969*4882a593Smuzhiyun */
970*4882a593Smuzhiyun priv->info.size = min(priv->info.size, (size_t)0xfe000000);
971*4882a593Smuzhiyun
972*4882a593Smuzhiyun return 0;
973*4882a593Smuzhiyun }
974*4882a593Smuzhiyun
rk3368_dmc_get_info(struct udevice * dev,struct ram_info * info)975*4882a593Smuzhiyun static int rk3368_dmc_get_info(struct udevice *dev, struct ram_info *info)
976*4882a593Smuzhiyun {
977*4882a593Smuzhiyun struct dram_info *priv = dev_get_priv(dev);
978*4882a593Smuzhiyun
979*4882a593Smuzhiyun *info = priv->info;
980*4882a593Smuzhiyun return 0;
981*4882a593Smuzhiyun }
982*4882a593Smuzhiyun
983*4882a593Smuzhiyun static struct ram_ops rk3368_dmc_ops = {
984*4882a593Smuzhiyun .get_info = rk3368_dmc_get_info,
985*4882a593Smuzhiyun };
986*4882a593Smuzhiyun
987*4882a593Smuzhiyun
988*4882a593Smuzhiyun static const struct udevice_id rk3368_dmc_ids[] = {
989*4882a593Smuzhiyun { .compatible = "rockchip,rk3368-dmc" },
990*4882a593Smuzhiyun { }
991*4882a593Smuzhiyun };
992*4882a593Smuzhiyun
993*4882a593Smuzhiyun U_BOOT_DRIVER(dmc_rk3368) = {
994*4882a593Smuzhiyun .name = "rockchip_rk3368_dmc",
995*4882a593Smuzhiyun .id = UCLASS_RAM,
996*4882a593Smuzhiyun .of_match = rk3368_dmc_ids,
997*4882a593Smuzhiyun .ops = &rk3368_dmc_ops,
998*4882a593Smuzhiyun .probe = rk3368_dmc_probe,
999*4882a593Smuzhiyun .priv_auto_alloc_size = sizeof(struct dram_info),
1000*4882a593Smuzhiyun .ofdata_to_platdata = rk3368_dmc_ofdata_to_platdata,
1001*4882a593Smuzhiyun .probe = rk3368_dmc_probe,
1002*4882a593Smuzhiyun .priv_auto_alloc_size = sizeof(struct dram_info),
1003*4882a593Smuzhiyun .platdata_auto_alloc_size = sizeof(struct rk3368_sdram_params),
1004*4882a593Smuzhiyun };
1005