xref: /openbmc/u-boot/arch/arm/mach-omap2/am33xx/ddr.c (revision 9ec4a67e)
1 /*
2  * DDR Configuration for AM33xx devices.
3  *
4  * Copyright (C) 2011 Texas Instruments Incorporated - http://www.ti.com/
5  *
6  * SPDX-License-Identifier:	GPL-2.0+
7  */
8 
9 #include <asm/arch/cpu.h>
10 #include <asm/arch/ddr_defs.h>
11 #include <asm/arch/sys_proto.h>
12 #include <asm/io.h>
13 #include <asm/emif.h>
14 
15 /**
16  * Base address for EMIF instances
17  */
18 static struct emif_reg_struct *emif_reg[2] = {
19 				(struct emif_reg_struct *)EMIF4_0_CFG_BASE,
20 				(struct emif_reg_struct *)EMIF4_1_CFG_BASE};
21 
22 /**
23  * Base addresses for DDR PHY cmd/data regs
24  */
25 static struct ddr_cmd_regs *ddr_cmd_reg[2] = {
26 				(struct ddr_cmd_regs *)DDR_PHY_CMD_ADDR,
27 				(struct ddr_cmd_regs *)DDR_PHY_CMD_ADDR2};
28 
29 static struct ddr_data_regs *ddr_data_reg[2] = {
30 				(struct ddr_data_regs *)DDR_PHY_DATA_ADDR,
31 				(struct ddr_data_regs *)DDR_PHY_DATA_ADDR2};
32 
33 /**
34  * Base address for ddr io control instances
35  */
36 static struct ddr_cmdtctrl *ioctrl_reg = {
37 			(struct ddr_cmdtctrl *)DDR_CONTROL_BASE_ADDR};
38 
39 static inline u32 get_mr(int nr, u32 cs, u32 mr_addr)
40 {
41 	u32 mr;
42 
43 	mr_addr |= cs << EMIF_REG_CS_SHIFT;
44 	writel(mr_addr, &emif_reg[nr]->emif_lpddr2_mode_reg_cfg);
45 
46 	mr = readl(&emif_reg[nr]->emif_lpddr2_mode_reg_data);
47 	debug("get_mr: EMIF1 cs %d mr %08x val 0x%x\n", cs, mr_addr, mr);
48 	if (((mr & 0x0000ff00) >>  8) == (mr & 0xff) &&
49 	    ((mr & 0x00ff0000) >> 16) == (mr & 0xff) &&
50 	    ((mr & 0xff000000) >> 24) == (mr & 0xff))
51 		return mr & 0xff;
52 	else
53 		return mr;
54 }
55 
56 static inline void set_mr(int nr, u32 cs, u32 mr_addr, u32 mr_val)
57 {
58 	mr_addr |= cs << EMIF_REG_CS_SHIFT;
59 	writel(mr_addr, &emif_reg[nr]->emif_lpddr2_mode_reg_cfg);
60 	writel(mr_val, &emif_reg[nr]->emif_lpddr2_mode_reg_data);
61 }
62 
63 static void configure_mr(int nr, u32 cs)
64 {
65 	u32 mr_addr;
66 
67 	while (get_mr(nr, cs, LPDDR2_MR0) & LPDDR2_MR0_DAI_MASK)
68 		;
69 	set_mr(nr, cs, LPDDR2_MR10, 0x56);
70 
71 	set_mr(nr, cs, LPDDR2_MR1, 0x43);
72 	set_mr(nr, cs, LPDDR2_MR2, 0x2);
73 
74 	mr_addr = LPDDR2_MR2 | EMIF_REG_REFRESH_EN_MASK;
75 	set_mr(nr, cs, mr_addr, 0x2);
76 }
77 
78 /*
79  * Configure EMIF4D5 registers and MR registers For details about these magic
80  * values please see the EMIF registers section of the TRM.
81  */
82 void config_sdram_emif4d5(const struct emif_regs *regs, int nr)
83 {
84 	writel(0xA0, &emif_reg[nr]->emif_pwr_mgmt_ctrl);
85 	writel(0xA0, &emif_reg[nr]->emif_pwr_mgmt_ctrl_shdw);
86 	writel(regs->zq_config, &emif_reg[nr]->emif_zq_config);
87 
88 	writel(regs->temp_alert_config, &emif_reg[nr]->emif_temp_alert_config);
89 	writel(regs->emif_rd_wr_lvl_rmp_win,
90 	       &emif_reg[nr]->emif_rd_wr_lvl_rmp_win);
91 	writel(regs->emif_rd_wr_lvl_rmp_ctl,
92 	       &emif_reg[nr]->emif_rd_wr_lvl_rmp_ctl);
93 	writel(regs->emif_rd_wr_lvl_ctl, &emif_reg[nr]->emif_rd_wr_lvl_ctl);
94 	writel(regs->emif_rd_wr_exec_thresh,
95 	       &emif_reg[nr]->emif_rd_wr_exec_thresh);
96 
97 	/*
98 	 * for most SOCs these registers won't need to be changed so only
99 	 * write to these registers if someone explicitly has set the
100 	 * register's value.
101 	 */
102 	if(regs->emif_cos_config) {
103 		writel(regs->emif_prio_class_serv_map, &emif_reg[nr]->emif_prio_class_serv_map);
104 		writel(regs->emif_connect_id_serv_1_map, &emif_reg[nr]->emif_connect_id_serv_1_map);
105 		writel(regs->emif_connect_id_serv_2_map, &emif_reg[nr]->emif_connect_id_serv_2_map);
106 		writel(regs->emif_cos_config, &emif_reg[nr]->emif_cos_config);
107 	}
108 
109 	/*
110 	 * Sequence to ensure that the PHY is in a known state prior to
111 	 * startting hardware leveling.  Also acts as to latch some state from
112 	 * the EMIF into the PHY.
113 	 */
114 	writel(0x2011, &emif_reg[nr]->emif_iodft_tlgc);
115 	writel(0x2411, &emif_reg[nr]->emif_iodft_tlgc);
116 	writel(0x2011, &emif_reg[nr]->emif_iodft_tlgc);
117 
118 	clrbits_le32(&emif_reg[nr]->emif_sdram_ref_ctrl,
119 			EMIF_REG_INITREF_DIS_MASK);
120 
121 	writel(regs->sdram_config, &emif_reg[nr]->emif_sdram_config);
122 	writel(regs->sdram_config, &cstat->secure_emif_sdram_config);
123 
124 	/* Wait 1ms because of L3 timeout error */
125 	udelay(1000);
126 
127 	writel(regs->ref_ctrl, &emif_reg[nr]->emif_sdram_ref_ctrl);
128 	writel(regs->ref_ctrl, &emif_reg[nr]->emif_sdram_ref_ctrl_shdw);
129 
130 	/* Perform hardware leveling for DDR3 */
131 	if (emif_sdram_type(regs->sdram_config) == EMIF_SDRAM_TYPE_DDR3) {
132 		writel(readl(&emif_reg[nr]->emif_ddr_ext_phy_ctrl_36) |
133 		       0x100, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_36);
134 		writel(readl(&emif_reg[nr]->emif_ddr_ext_phy_ctrl_36_shdw) |
135 		       0x100, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_36_shdw);
136 
137 		writel(0x80000000, &emif_reg[nr]->emif_rd_wr_lvl_rmp_ctl);
138 
139 		/* Enable read leveling */
140 		writel(0x80000000, &emif_reg[nr]->emif_rd_wr_lvl_ctl);
141 
142 		/*
143 		 * Enable full read and write leveling.  Wait for read and write
144 		 * leveling bit to clear RDWRLVLFULL_START bit 31
145 		 */
146 		while ((readl(&emif_reg[nr]->emif_rd_wr_lvl_ctl) & 0x80000000)
147 		      != 0)
148 			;
149 
150 		/* Check the timeout register to see if leveling is complete */
151 		if ((readl(&emif_reg[nr]->emif_status) & 0x70) != 0)
152 			puts("DDR3 H/W leveling incomplete with errors\n");
153 
154 	} else {
155 		/* DDR2 */
156 		configure_mr(nr, 0);
157 		configure_mr(nr, 1);
158 	}
159 }
160 
161 /**
162  * Configure SDRAM
163  */
164 void config_sdram(const struct emif_regs *regs, int nr)
165 {
166 	if (regs->zq_config) {
167 		writel(regs->zq_config, &emif_reg[nr]->emif_zq_config);
168 		writel(regs->sdram_config, &cstat->secure_emif_sdram_config);
169 		writel(regs->sdram_config, &emif_reg[nr]->emif_sdram_config);
170 
171 		/* Trigger initialization */
172 		writel(0x00003100, &emif_reg[nr]->emif_sdram_ref_ctrl);
173 		/* Wait 1ms because of L3 timeout error */
174 		udelay(1000);
175 
176 		/* Write proper sdram_ref_cref_ctrl value */
177 		writel(regs->ref_ctrl, &emif_reg[nr]->emif_sdram_ref_ctrl);
178 		writel(regs->ref_ctrl, &emif_reg[nr]->emif_sdram_ref_ctrl_shdw);
179 	}
180 	writel(regs->ref_ctrl, &emif_reg[nr]->emif_sdram_ref_ctrl);
181 	writel(regs->ref_ctrl, &emif_reg[nr]->emif_sdram_ref_ctrl_shdw);
182 	writel(regs->sdram_config, &emif_reg[nr]->emif_sdram_config);
183 }
184 
185 /**
186  * Set SDRAM timings
187  */
188 void set_sdram_timings(const struct emif_regs *regs, int nr)
189 {
190 	writel(regs->sdram_tim1, &emif_reg[nr]->emif_sdram_tim_1);
191 	writel(regs->sdram_tim1, &emif_reg[nr]->emif_sdram_tim_1_shdw);
192 	writel(regs->sdram_tim2, &emif_reg[nr]->emif_sdram_tim_2);
193 	writel(regs->sdram_tim2, &emif_reg[nr]->emif_sdram_tim_2_shdw);
194 	writel(regs->sdram_tim3, &emif_reg[nr]->emif_sdram_tim_3);
195 	writel(regs->sdram_tim3, &emif_reg[nr]->emif_sdram_tim_3_shdw);
196 }
197 
198 /*
199  * Configure EXT PHY registers for software leveling
200  */
201 static void ext_phy_settings_swlvl(const struct emif_regs *regs, int nr)
202 {
203 	u32 *ext_phy_ctrl_base = 0;
204 	u32 *emif_ext_phy_ctrl_base = 0;
205 	__maybe_unused const u32 *ext_phy_ctrl_const_regs;
206 	u32 i = 0;
207 	__maybe_unused u32 size;
208 
209 	ext_phy_ctrl_base = (u32 *)&(regs->emif_ddr_ext_phy_ctrl_1);
210 	emif_ext_phy_ctrl_base =
211 			(u32 *)&(emif_reg[nr]->emif_ddr_ext_phy_ctrl_1);
212 
213 	/* Configure external phy control timing registers */
214 	for (i = 0; i < EMIF_EXT_PHY_CTRL_TIMING_REG; i++) {
215 		writel(*ext_phy_ctrl_base, emif_ext_phy_ctrl_base++);
216 		/* Update shadow registers */
217 		writel(*ext_phy_ctrl_base++, emif_ext_phy_ctrl_base++);
218 	}
219 
220 #ifdef CONFIG_AM43XX
221 	/*
222 	 * External phy 6-24 registers do not change with ddr frequency.
223 	 * These only need to be set on DDR2 on AM43xx.
224 	 */
225 	emif_get_ext_phy_ctrl_const_regs(&ext_phy_ctrl_const_regs, &size);
226 
227 	if (!size)
228 		return;
229 
230 	for (i = 0; i < size; i++) {
231 		writel(ext_phy_ctrl_const_regs[i], emif_ext_phy_ctrl_base++);
232 		/* Update shadow registers */
233 		writel(ext_phy_ctrl_const_regs[i], emif_ext_phy_ctrl_base++);
234 	}
235 #endif
236 }
237 
238 /*
239  * Configure EXT PHY registers for hardware leveling
240  */
241 static void ext_phy_settings_hwlvl(const struct emif_regs *regs, int nr)
242 {
243 	/*
244 	 * Enable hardware leveling on the EMIF.  For details about these
245 	 * magic values please see the EMIF registers section of the TRM.
246 	 */
247 	writel(0x08020080, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_1);
248 	writel(0x08020080, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_1_shdw);
249 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_22);
250 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_22_shdw);
251 	writel(0x00600020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_23);
252 	writel(0x00600020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_23_shdw);
253 	writel(0x40010080, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_24);
254 	writel(0x40010080, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_24_shdw);
255 	writel(0x08102040, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_25);
256 	writel(0x08102040, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_25_shdw);
257 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_26);
258 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_26_shdw);
259 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_27);
260 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_27_shdw);
261 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_28);
262 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_28_shdw);
263 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_29);
264 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_29_shdw);
265 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_30);
266 	writel(0x00200020, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_30_shdw);
267 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_31);
268 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_31_shdw);
269 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_32);
270 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_32_shdw);
271 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_33);
272 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_33_shdw);
273 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_34);
274 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_34_shdw);
275 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_35);
276 	writel(0x00000000, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_35_shdw);
277 	writel(0x000000FF, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_36);
278 	writel(0x000000FF, &emif_reg[nr]->emif_ddr_ext_phy_ctrl_36_shdw);
279 
280 	/*
281 	 * Sequence to ensure that the PHY is again in a known state after
282 	 * hardware leveling.
283 	 */
284 	writel(0x2011, &emif_reg[nr]->emif_iodft_tlgc);
285 	writel(0x2411, &emif_reg[nr]->emif_iodft_tlgc);
286 	writel(0x2011, &emif_reg[nr]->emif_iodft_tlgc);
287 }
288 
289 /**
290  * Configure DDR PHY
291  */
292 void config_ddr_phy(const struct emif_regs *regs, int nr)
293 {
294 	/*
295 	 * Disable initialization and refreshes for now until we finish
296 	 * programming EMIF regs and set time between rising edge of
297 	 * DDR_RESET to rising edge of DDR_CKE to > 500us per memory spec.
298 	 * We currently hardcode a value based on a max expected frequency
299 	 * of 400MHz.
300 	 */
301 	writel(EMIF_REG_INITREF_DIS_MASK | 0x3100,
302 		&emif_reg[nr]->emif_sdram_ref_ctrl);
303 
304 	writel(regs->emif_ddr_phy_ctlr_1,
305 		&emif_reg[nr]->emif_ddr_phy_ctrl_1);
306 	writel(regs->emif_ddr_phy_ctlr_1,
307 		&emif_reg[nr]->emif_ddr_phy_ctrl_1_shdw);
308 
309 	if (get_emif_rev((u32)emif_reg[nr]) == EMIF_4D5) {
310 		if (emif_sdram_type(regs->sdram_config) == EMIF_SDRAM_TYPE_DDR3)
311 			ext_phy_settings_hwlvl(regs, nr);
312 		else
313 			ext_phy_settings_swlvl(regs, nr);
314 	}
315 }
316 
317 /**
318  * Configure DDR CMD control registers
319  */
320 void config_cmd_ctrl(const struct cmd_control *cmd, int nr)
321 {
322 	if (!cmd)
323 		return;
324 
325 	writel(cmd->cmd0csratio, &ddr_cmd_reg[nr]->cm0csratio);
326 	writel(cmd->cmd0iclkout, &ddr_cmd_reg[nr]->cm0iclkout);
327 
328 	writel(cmd->cmd1csratio, &ddr_cmd_reg[nr]->cm1csratio);
329 	writel(cmd->cmd1iclkout, &ddr_cmd_reg[nr]->cm1iclkout);
330 
331 	writel(cmd->cmd2csratio, &ddr_cmd_reg[nr]->cm2csratio);
332 	writel(cmd->cmd2iclkout, &ddr_cmd_reg[nr]->cm2iclkout);
333 }
334 
335 /**
336  * Configure DDR DATA registers
337  */
338 void config_ddr_data(const struct ddr_data *data, int nr)
339 {
340 	int i;
341 
342 	if (!data)
343 		return;
344 
345 	for (i = 0; i < DDR_DATA_REGS_NR; i++) {
346 		writel(data->datardsratio0,
347 			&(ddr_data_reg[nr]+i)->dt0rdsratio0);
348 		writel(data->datawdsratio0,
349 			&(ddr_data_reg[nr]+i)->dt0wdsratio0);
350 		writel(data->datawiratio0,
351 			&(ddr_data_reg[nr]+i)->dt0wiratio0);
352 		writel(data->datagiratio0,
353 			&(ddr_data_reg[nr]+i)->dt0giratio0);
354 		writel(data->datafwsratio0,
355 			&(ddr_data_reg[nr]+i)->dt0fwsratio0);
356 		writel(data->datawrsratio0,
357 			&(ddr_data_reg[nr]+i)->dt0wrsratio0);
358 	}
359 }
360 
361 void config_io_ctrl(const struct ctrl_ioregs *ioregs)
362 {
363 	if (!ioregs)
364 		return;
365 
366 	writel(ioregs->cm0ioctl, &ioctrl_reg->cm0ioctl);
367 	writel(ioregs->cm1ioctl, &ioctrl_reg->cm1ioctl);
368 	writel(ioregs->cm2ioctl, &ioctrl_reg->cm2ioctl);
369 	writel(ioregs->dt0ioctl, &ioctrl_reg->dt0ioctl);
370 	writel(ioregs->dt1ioctl, &ioctrl_reg->dt1ioctl);
371 #ifdef CONFIG_AM43XX
372 	writel(ioregs->dt2ioctrl, &ioctrl_reg->dt2ioctrl);
373 	writel(ioregs->dt3ioctrl, &ioctrl_reg->dt3ioctrl);
374 	writel(ioregs->emif_sdram_config_ext,
375 	       &ioctrl_reg->emif_sdram_config_ext);
376 #endif
377 }
378