]> git.ipfire.org Git - people/ms/u-boot.git/blob - drivers/ddr/marvell/axp/ddr3_hw_training.c
arm: mvebu: drivers/ddr: Move Armada XP DDR init code into new directory
[people/ms/u-boot.git] / drivers / ddr / marvell / axp / ddr3_hw_training.c
1 /*
2 * Copyright (C) Marvell International Ltd. and its affiliates
3 *
4 * SPDX-License-Identifier: GPL-2.0
5 */
6
7 #include <common.h>
8 #include <i2c.h>
9 #include <spl.h>
10 #include <asm/io.h>
11 #include <asm/arch/cpu.h>
12 #include <asm/arch/soc.h>
13
14 #include "ddr3_init.h"
15 #include "ddr3_hw_training.h"
16 #include "xor.h"
17
18 #ifdef MV88F78X60
19 #include "ddr3_patterns_64bit.h"
20 #else
21 #include "ddr3_patterns_16bit.h"
22 #if defined(MV88F672X)
23 #include "ddr3_patterns_16bit.h"
24 #endif
25 #endif
26
27 /*
28 * Debug
29 */
30
31 #define DEBUG_MAIN_C(s, d, l) \
32 DEBUG_MAIN_S(s); DEBUG_MAIN_D(d, l); DEBUG_MAIN_S("\n")
33 #define DEBUG_MAIN_FULL_C(s, d, l) \
34 DEBUG_MAIN_FULL_S(s); DEBUG_MAIN_FULL_D(d, l); DEBUG_MAIN_FULL_S("\n")
35
36 #ifdef MV_DEBUG_MAIN
37 #define DEBUG_MAIN_S(s) puts(s)
38 #define DEBUG_MAIN_D(d, l) printf("%x", d)
39 #else
40 #define DEBUG_MAIN_S(s)
41 #define DEBUG_MAIN_D(d, l)
42 #endif
43
44 #ifdef MV_DEBUG_MAIN_FULL
45 #define DEBUG_MAIN_FULL_S(s) puts(s)
46 #define DEBUG_MAIN_FULL_D(d, l) printf("%x", d)
47 #else
48 #define DEBUG_MAIN_FULL_S(s)
49 #define DEBUG_MAIN_FULL_D(d, l)
50 #endif
51
52 #ifdef MV_DEBUG_SUSPEND_RESUME
53 #define DEBUG_SUSPEND_RESUME_S(s) puts(s)
54 #define DEBUG_SUSPEND_RESUME_D(d, l) printf("%x", d)
55 #else
56 #define DEBUG_SUSPEND_RESUME_S(s)
57 #define DEBUG_SUSPEND_RESUME_D(d, l)
58 #endif
59
60 static u32 ddr3_sw_wl_rl_debug;
61 static u32 ddr3_run_pbs = 1;
62
63 void ddr3_print_version(void)
64 {
65 puts("DDR3 Training Sequence - Ver 5.7.");
66 }
67
68 void ddr3_set_sw_wl_rl_debug(u32 val)
69 {
70 ddr3_sw_wl_rl_debug = val;
71 }
72
73 void ddr3_set_pbs(u32 val)
74 {
75 ddr3_run_pbs = val;
76 }
77
78 int ddr3_hw_training(u32 target_freq, u32 ddr_width, int xor_bypass,
79 u32 scrub_offs, u32 scrub_size, int dqs_clk_aligned,
80 int debug_mode, int reg_dimm_skip_wl)
81 {
82 /* A370 has no PBS mechanism */
83 __maybe_unused u32 first_loop_flag = 0;
84 u32 freq, reg;
85 MV_DRAM_INFO dram_info;
86 int ratio_2to1 = 0;
87 int tmp_ratio = 1;
88 int status;
89
90 if (debug_mode)
91 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 1\n");
92
93 memset(&dram_info, 0, sizeof(dram_info));
94 dram_info.num_cs = ddr3_get_cs_num_from_reg();
95 dram_info.cs_ena = ddr3_get_cs_ena_from_reg();
96 dram_info.target_frequency = target_freq;
97 dram_info.ddr_width = ddr_width;
98 dram_info.num_of_std_pups = ddr_width / PUP_SIZE;
99 dram_info.rl400_bug = 0;
100 dram_info.multi_cs_mr_support = 0;
101 #ifdef MV88F67XX
102 dram_info.rl400_bug = 1;
103 #endif
104
105 /* Ignore ECC errors - if ECC is enabled */
106 reg = reg_read(REG_SDRAM_CONFIG_ADDR);
107 if (reg & (1 << REG_SDRAM_CONFIG_ECC_OFFS)) {
108 dram_info.ecc_ena = 1;
109 reg |= (1 << REG_SDRAM_CONFIG_IERR_OFFS);
110 reg_write(REG_SDRAM_CONFIG_ADDR, reg);
111 } else {
112 dram_info.ecc_ena = 0;
113 }
114
115 reg = reg_read(REG_SDRAM_CONFIG_ADDR);
116 if (reg & (1 << REG_SDRAM_CONFIG_REGDIMM_OFFS))
117 dram_info.reg_dimm = 1;
118 else
119 dram_info.reg_dimm = 0;
120
121 dram_info.num_of_total_pups = ddr_width / PUP_SIZE + dram_info.ecc_ena;
122
123 /* Get target 2T value */
124 reg = reg_read(REG_DUNIT_CTRL_LOW_ADDR);
125 dram_info.mode_2t = (reg >> REG_DUNIT_CTRL_LOW_2T_OFFS) &
126 REG_DUNIT_CTRL_LOW_2T_MASK;
127
128 /* Get target CL value */
129 #ifdef MV88F67XX
130 reg = reg_read(REG_DDR3_MR0_ADDR) >> 2;
131 #else
132 reg = reg_read(REG_DDR3_MR0_CS_ADDR) >> 2;
133 #endif
134
135 reg = (((reg >> 1) & 0xE) | (reg & 0x1)) & 0xF;
136 dram_info.cl = ddr3_valid_cl_to_cl(reg);
137
138 /* Get target CWL value */
139 #ifdef MV88F67XX
140 reg = reg_read(REG_DDR3_MR2_ADDR) >> REG_DDR3_MR2_CWL_OFFS;
141 #else
142 reg = reg_read(REG_DDR3_MR2_CS_ADDR) >> REG_DDR3_MR2_CWL_OFFS;
143 #endif
144
145 reg &= REG_DDR3_MR2_CWL_MASK;
146 dram_info.cwl = reg;
147 #if !defined(MV88F67XX)
148 /* A370 has no PBS mechanism */
149 #if defined(MV88F78X60)
150 if ((dram_info.target_frequency > DDR_400) && (ddr3_run_pbs))
151 first_loop_flag = 1;
152 #else
153 /* first_loop_flag = 1; skip mid freq at ALP/A375 */
154 if ((dram_info.target_frequency > DDR_400) && (ddr3_run_pbs) &&
155 (mv_ctrl_revision_get() >= UMC_A0))
156 first_loop_flag = 1;
157 else
158 first_loop_flag = 0;
159 #endif
160 #endif
161
162 freq = dram_info.target_frequency;
163
164 /* Set ODT to always on */
165 ddr3_odt_activate(1);
166
167 /* Init XOR */
168 mv_sys_xor_init(&dram_info);
169
170 /* Get DRAM/HCLK ratio */
171 if (reg_read(REG_DDR_IO_ADDR) & (1 << REG_DDR_IO_CLK_RATIO_OFFS))
172 ratio_2to1 = 1;
173
174 /*
175 * Xor Bypass - ECC support in AXP is currently available for 1:1
176 * modes frequency modes.
177 * Not all frequency modes support the ddr3 training sequence
178 * (Only 1200/300).
179 * Xor Bypass allows using the Xor initializations and scrubbing
180 * inside the ddr3 training sequence without running the training
181 * itself.
182 */
183 if (xor_bypass == 0) {
184 if (ddr3_run_pbs) {
185 DEBUG_MAIN_S("DDR3 Training Sequence - Run with PBS.\n");
186 } else {
187 DEBUG_MAIN_S("DDR3 Training Sequence - Run without PBS.\n");
188 }
189
190 if (dram_info.target_frequency > DFS_MARGIN) {
191 tmp_ratio = 0;
192 freq = DDR_100;
193
194 if (dram_info.reg_dimm == 1)
195 freq = DDR_300;
196
197 if (MV_OK != ddr3_dfs_high_2_low(freq, &dram_info)) {
198 /* Set low - 100Mhz DDR Frequency by HW */
199 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs High2Low)\n");
200 return MV_DDR3_TRAINING_ERR_DFS_H2L;
201 }
202
203 if ((dram_info.reg_dimm == 1) &&
204 (reg_dimm_skip_wl == 0)) {
205 if (MV_OK !=
206 ddr3_write_leveling_hw_reg_dimm(freq,
207 &dram_info))
208 DEBUG_MAIN_S("DDR3 Training Sequence - Registered DIMM Low WL - SKIP\n");
209 }
210
211 if (ddr3_get_log_level() >= MV_LOG_LEVEL_1)
212 ddr3_print_freq(freq);
213
214 if (debug_mode)
215 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 2\n");
216 } else {
217 if (!dqs_clk_aligned) {
218 #ifdef MV88F67XX
219 /*
220 * If running training sequence without DFS,
221 * we must run Write leveling before writing
222 * the patterns
223 */
224
225 /*
226 * ODT - Multi CS system use SW WL,
227 * Single CS System use HW WL
228 */
229 if (dram_info.cs_ena > 1) {
230 if (MV_OK !=
231 ddr3_write_leveling_sw(
232 freq, tmp_ratio,
233 &dram_info)) {
234 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
235 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
236 }
237 } else {
238 if (MV_OK !=
239 ddr3_write_leveling_hw(freq,
240 &dram_info)) {
241 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
242 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
243 }
244 }
245 #else
246 if (MV_OK != ddr3_write_leveling_hw(
247 freq, &dram_info)) {
248 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
249 if (ddr3_sw_wl_rl_debug) {
250 if (MV_OK !=
251 ddr3_write_leveling_sw(
252 freq, tmp_ratio,
253 &dram_info)) {
254 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
255 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
256 }
257 } else {
258 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
259 }
260 }
261 #endif
262 }
263
264 if (debug_mode)
265 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 3\n");
266 }
267
268 if (MV_OK != ddr3_load_patterns(&dram_info, 0)) {
269 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Loading Patterns)\n");
270 return MV_DDR3_TRAINING_ERR_LOAD_PATTERNS;
271 }
272
273 /*
274 * TODO:
275 * The mainline U-Boot port of the bin_hdr DDR training code
276 * needs a delay of minimum 20ms here (10ms is a bit too short
277 * and the CPU hangs). The bin_hdr code doesn't have this delay.
278 * To be save here, lets add a delay of 50ms here.
279 *
280 * Tested on the Marvell DB-MV784MP-GP board
281 */
282 mdelay(50);
283
284 do {
285 freq = dram_info.target_frequency;
286 tmp_ratio = ratio_2to1;
287 DEBUG_MAIN_FULL_S("DDR3 Training Sequence - DEBUG - 4\n");
288
289 #if defined(MV88F78X60)
290 /*
291 * There is a difference on the DFS frequency at the
292 * first iteration of this loop
293 */
294 if (first_loop_flag) {
295 freq = DDR_400;
296 tmp_ratio = 0;
297 }
298 #endif
299
300 if (MV_OK != ddr3_dfs_low_2_high(freq, tmp_ratio,
301 &dram_info)) {
302 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs Low2High)\n");
303 return MV_DDR3_TRAINING_ERR_DFS_H2L;
304 }
305
306 if (ddr3_get_log_level() >= MV_LOG_LEVEL_1) {
307 ddr3_print_freq(freq);
308 }
309
310 if (debug_mode)
311 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 5\n");
312
313 /* Write leveling */
314 if (!dqs_clk_aligned) {
315 #ifdef MV88F67XX
316 /*
317 * ODT - Multi CS system that not support Multi
318 * CS MRS commands must use SW WL
319 */
320 if (dram_info.cs_ena > 1) {
321 if (MV_OK != ddr3_write_leveling_sw(
322 freq, tmp_ratio, &dram_info)) {
323 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
324 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
325 }
326 } else {
327 if (MV_OK != ddr3_write_leveling_hw(
328 freq, &dram_info)) {
329 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
330 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
331 }
332 }
333 #else
334 if ((dram_info.reg_dimm == 1) &&
335 (freq == DDR_400)) {
336 if (reg_dimm_skip_wl == 0) {
337 if (MV_OK != ddr3_write_leveling_hw_reg_dimm(
338 freq, &dram_info))
339 DEBUG_MAIN_S("DDR3 Training Sequence - Registered DIMM WL - SKIP\n");
340 }
341 } else {
342 if (MV_OK != ddr3_write_leveling_hw(
343 freq, &dram_info)) {
344 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
345 if (ddr3_sw_wl_rl_debug) {
346 if (MV_OK != ddr3_write_leveling_sw(
347 freq, tmp_ratio, &dram_info)) {
348 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Sw)\n");
349 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
350 }
351 } else {
352 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
353 }
354 }
355 }
356 #endif
357 if (debug_mode)
358 DEBUG_MAIN_S
359 ("DDR3 Training Sequence - DEBUG - 6\n");
360 }
361
362 /* Read Leveling */
363 /*
364 * Armada 370 - Support for HCLK @ 400MHZ - must use
365 * SW read leveling
366 */
367 if (freq == DDR_400 && dram_info.rl400_bug) {
368 status = ddr3_read_leveling_sw(freq, tmp_ratio,
369 &dram_info);
370 if (MV_OK != status) {
371 DEBUG_MAIN_S
372 ("DDR3 Training Sequence - FAILED (Read Leveling Sw)\n");
373 return status;
374 }
375 } else {
376 if (MV_OK != ddr3_read_leveling_hw(
377 freq, &dram_info)) {
378 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Hw)\n");
379 if (ddr3_sw_wl_rl_debug) {
380 if (MV_OK != ddr3_read_leveling_sw(
381 freq, tmp_ratio,
382 &dram_info)) {
383 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Sw)\n");
384 return MV_DDR3_TRAINING_ERR_WR_LVL_SW;
385 }
386 } else {
387 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
388 }
389 }
390 }
391
392 if (debug_mode)
393 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 7\n");
394
395 if (MV_OK != ddr3_wl_supplement(&dram_info)) {
396 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hi-Freq Sup)\n");
397 return MV_DDR3_TRAINING_ERR_WR_LVL_HI_FREQ;
398 }
399
400 if (debug_mode)
401 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 8\n");
402 #if !defined(MV88F67XX)
403 /* A370 has no PBS mechanism */
404 #if defined(MV88F78X60) || defined(MV88F672X)
405 if (first_loop_flag == 1) {
406 first_loop_flag = 0;
407
408 status = MV_OK;
409 status = ddr3_pbs_rx(&dram_info);
410 if (MV_OK != status) {
411 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (PBS RX)\n");
412 return status;
413 }
414
415 if (debug_mode)
416 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 9\n");
417
418 status = ddr3_pbs_tx(&dram_info);
419 if (MV_OK != status) {
420 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (PBS TX)\n");
421 return status;
422 }
423
424 if (debug_mode)
425 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 10\n");
426 }
427 #endif
428 #endif
429 } while (freq != dram_info.target_frequency);
430
431 status = ddr3_dqs_centralization_rx(&dram_info);
432 if (MV_OK != status) {
433 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (DQS Centralization RX)\n");
434 return status;
435 }
436
437 if (debug_mode)
438 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 11\n");
439
440 status = ddr3_dqs_centralization_tx(&dram_info);
441 if (MV_OK != status) {
442 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (DQS Centralization TX)\n");
443 return status;
444 }
445
446 if (debug_mode)
447 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 12\n");
448 }
449
450 ddr3_set_performance_params(&dram_info);
451
452 if (dram_info.ecc_ena) {
453 /* Need to SCRUB the DRAM memory area to load U-boot */
454 mv_sys_xor_finish();
455 dram_info.num_cs = 1;
456 dram_info.cs_ena = 1;
457 mv_sys_xor_init(&dram_info);
458 mv_xor_mem_init(0, scrub_offs, scrub_size, 0xdeadbeef,
459 0xdeadbeef);
460
461 /* Wait for previous transfer completion */
462 while (mv_xor_state_get(0) != MV_IDLE)
463 ;
464
465 if (debug_mode)
466 DEBUG_MAIN_S("DDR3 Training Sequence - DEBUG - 13\n");
467 }
468
469 /* Return XOR State */
470 mv_sys_xor_finish();
471
472 #if defined(MV88F78X60)
473 /* Save training results in memeory for resume state */
474 ddr3_save_training(&dram_info);
475 #endif
476 /* Clear ODT always on */
477 ddr3_odt_activate(0);
478
479 /* Configure Dynamic read ODT */
480 ddr3_odt_read_dynamic_config(&dram_info);
481
482 return MV_OK;
483 }
484
485 void ddr3_set_performance_params(MV_DRAM_INFO *dram_info)
486 {
487 u32 twr2wr, trd2rd, trd2wr_wr2rd;
488 u32 tmp1, tmp2, reg;
489
490 DEBUG_MAIN_FULL_C("Max WL Phase: ", dram_info->wl_max_phase, 2);
491 DEBUG_MAIN_FULL_C("Min WL Phase: ", dram_info->wl_min_phase, 2);
492 DEBUG_MAIN_FULL_C("Max RL Phase: ", dram_info->rl_max_phase, 2);
493 DEBUG_MAIN_FULL_C("Min RL Phase: ", dram_info->rl_min_phase, 2);
494
495 if (dram_info->wl_max_phase < 2)
496 twr2wr = 0x2;
497 else
498 twr2wr = 0x3;
499
500 trd2rd = 0x1 + (dram_info->rl_max_phase + 1) / 2 +
501 (dram_info->rl_max_phase + 1) % 2;
502
503 tmp1 = (dram_info->rl_max_phase - dram_info->wl_min_phase) / 2 +
504 (((dram_info->rl_max_phase - dram_info->wl_min_phase) % 2) >
505 0 ? 1 : 0);
506 tmp2 = (dram_info->wl_max_phase - dram_info->rl_min_phase) / 2 +
507 ((dram_info->wl_max_phase - dram_info->rl_min_phase) % 2 >
508 0 ? 1 : 0);
509 trd2wr_wr2rd = (tmp1 >= tmp2) ? tmp1 : tmp2;
510
511 trd2wr_wr2rd += 2;
512 trd2rd += 2;
513 twr2wr += 2;
514
515 DEBUG_MAIN_FULL_C("WR 2 WR: ", twr2wr, 2);
516 DEBUG_MAIN_FULL_C("RD 2 RD: ", trd2rd, 2);
517 DEBUG_MAIN_FULL_C("RD 2 WR / WR 2 RD: ", trd2wr_wr2rd, 2);
518
519 reg = reg_read(REG_SDRAM_TIMING_HIGH_ADDR);
520
521 reg &= ~(REG_SDRAM_TIMING_H_W2W_MASK << REG_SDRAM_TIMING_H_W2W_OFFS);
522 reg |= ((twr2wr & REG_SDRAM_TIMING_H_W2W_MASK) <<
523 REG_SDRAM_TIMING_H_W2W_OFFS);
524
525 reg &= ~(REG_SDRAM_TIMING_H_R2R_MASK << REG_SDRAM_TIMING_H_R2R_OFFS);
526 reg &= ~(REG_SDRAM_TIMING_H_R2R_H_MASK <<
527 REG_SDRAM_TIMING_H_R2R_H_OFFS);
528 reg |= ((trd2rd & REG_SDRAM_TIMING_H_R2R_MASK) <<
529 REG_SDRAM_TIMING_H_R2R_OFFS);
530 reg |= (((trd2rd >> 2) & REG_SDRAM_TIMING_H_R2R_H_MASK) <<
531 REG_SDRAM_TIMING_H_R2R_H_OFFS);
532
533 reg &= ~(REG_SDRAM_TIMING_H_R2W_W2R_MASK <<
534 REG_SDRAM_TIMING_H_R2W_W2R_OFFS);
535 reg &= ~(REG_SDRAM_TIMING_H_R2W_W2R_H_MASK <<
536 REG_SDRAM_TIMING_H_R2W_W2R_H_OFFS);
537 reg |= ((trd2wr_wr2rd & REG_SDRAM_TIMING_H_R2W_W2R_MASK) <<
538 REG_SDRAM_TIMING_H_R2W_W2R_OFFS);
539 reg |= (((trd2wr_wr2rd >> 2) & REG_SDRAM_TIMING_H_R2W_W2R_H_MASK) <<
540 REG_SDRAM_TIMING_H_R2W_W2R_H_OFFS);
541
542 reg_write(REG_SDRAM_TIMING_HIGH_ADDR, reg);
543 }
544
545 /*
546 * Perform DDR3 PUP Indirect Write
547 */
548 void ddr3_write_pup_reg(u32 mode, u32 cs, u32 pup, u32 phase, u32 delay)
549 {
550 u32 reg = 0;
551
552 if (pup == PUP_BC)
553 reg |= (1 << REG_PHY_BC_OFFS);
554 else
555 reg |= (pup << REG_PHY_PUP_OFFS);
556
557 reg |= ((0x4 * cs + mode) << REG_PHY_CS_OFFS);
558 reg |= (phase << REG_PHY_PHASE_OFFS) | delay;
559
560 if (mode == PUP_WL_MODE)
561 reg |= ((INIT_WL_DELAY + delay) << REG_PHY_DQS_REF_DLY_OFFS);
562
563 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
564 reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
565 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
566
567 do {
568 reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
569 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
570 } while (reg); /* Wait for '0' to mark the end of the transaction */
571
572 /* If read Leveling mode - need to write to register 3 separetly */
573 if (mode == PUP_RL_MODE) {
574 reg = 0;
575
576 if (pup == PUP_BC)
577 reg |= (1 << REG_PHY_BC_OFFS);
578 else
579 reg |= (pup << REG_PHY_PUP_OFFS);
580
581 reg |= ((0x4 * cs + mode + 1) << REG_PHY_CS_OFFS);
582 reg |= (INIT_RL_DELAY);
583
584 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
585 reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
586 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
587
588 do {
589 reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
590 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
591 } while (reg);
592 }
593 }
594
595 /*
596 * Perform DDR3 PUP Indirect Read
597 */
598 u32 ddr3_read_pup_reg(u32 mode, u32 cs, u32 pup)
599 {
600 u32 reg;
601
602 reg = (pup << REG_PHY_PUP_OFFS) |
603 ((0x4 * cs + mode) << REG_PHY_CS_OFFS);
604 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
605
606 reg |= REG_PHY_REGISTRY_FILE_ACCESS_OP_RD;
607 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, reg); /* 0x16A0 */
608
609 do {
610 reg = reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR) &
611 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
612 } while (reg); /* Wait for '0' to mark the end of the transaction */
613
614 return reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR); /* 0x16A0 */
615 }
616
617 /*
618 * Set training patterns
619 */
620 int ddr3_load_patterns(MV_DRAM_INFO *dram_info, int resume)
621 {
622 u32 reg;
623
624 /* Enable SW override - Required for the ECC Pup */
625 reg = reg_read(REG_DRAM_TRAINING_2_ADDR) |
626 (1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS);
627
628 /* [0] = 1 - Enable SW override */
629 /* 0x15B8 - Training SW 2 Register */
630 reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
631
632 reg = (1 << REG_DRAM_TRAINING_AUTO_OFFS);
633 reg_write(REG_DRAM_TRAINING_ADDR, reg); /* 0x15B0 - Training Register */
634
635 if (resume == 0) {
636 #if defined(MV88F78X60) || defined(MV88F672X)
637 ddr3_load_pbs_patterns(dram_info);
638 #endif
639 ddr3_load_dqs_patterns(dram_info);
640 }
641
642 /* Disable SW override - Must be in a different stage */
643 /* [0]=0 - Enable SW override */
644 reg = reg_read(REG_DRAM_TRAINING_2_ADDR);
645 reg &= ~(1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS);
646 /* 0x15B8 - Training SW 2 Register */
647 reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
648
649 reg = reg_read(REG_DRAM_TRAINING_1_ADDR) |
650 (1 << REG_DRAM_TRAINING_1_TRNBPOINT_OFFS);
651 reg_write(REG_DRAM_TRAINING_1_ADDR, reg);
652
653 /* Set Base Addr */
654 #if defined(MV88F67XX)
655 reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR, 0);
656 #else
657 if (resume == 0)
658 reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR, 0);
659 else
660 reg_write(REG_DRAM_TRAINING_PATTERN_BASE_ADDR,
661 RESUME_RL_PATTERNS_ADDR);
662 #endif
663
664 /* Set Patterns */
665 if (resume == 0) {
666 reg = (dram_info->cs_ena << REG_DRAM_TRAINING_CS_OFFS) |
667 (1 << REG_DRAM_TRAINING_PATTERNS_OFFS);
668 } else {
669 reg = (0x1 << REG_DRAM_TRAINING_CS_OFFS) |
670 (1 << REG_DRAM_TRAINING_PATTERNS_OFFS);
671 }
672
673 reg |= (1 << REG_DRAM_TRAINING_AUTO_OFFS);
674
675 reg_write(REG_DRAM_TRAINING_ADDR, reg);
676
677 udelay(100);
678
679 /* Check if Successful */
680 if (reg_read(REG_DRAM_TRAINING_ADDR) &
681 (1 << REG_DRAM_TRAINING_ERROR_OFFS))
682 return MV_OK;
683 else
684 return MV_FAIL;
685 }
686
687 #if !defined(MV88F67XX)
688 /*
689 * Name: ddr3_save_training(MV_DRAM_INFO *dram_info)
690 * Desc: saves the training results to memeory (RL,WL,PBS,Rx/Tx
691 * Centeralization)
692 * Args: MV_DRAM_INFO *dram_info
693 * Notes:
694 * Returns: None.
695 */
696 void ddr3_save_training(MV_DRAM_INFO *dram_info)
697 {
698 u32 val, pup, tmp_cs, cs, i, dq;
699 u32 crc = 0;
700 u32 regs = 0;
701 u32 *sdram_offset = (u32 *)RESUME_TRAINING_VALUES_ADDR;
702 u32 mode_config[MAX_TRAINING_MODE];
703
704 mode_config[DQS_WR_MODE] = PUP_DQS_WR;
705 mode_config[WL_MODE_] = PUP_WL_MODE;
706 mode_config[RL_MODE_] = PUP_RL_MODE;
707 mode_config[DQS_RD_MODE] = PUP_DQS_RD;
708 mode_config[PBS_TX_DM_MODE] = PUP_PBS_TX_DM;
709 mode_config[PBS_TX_MODE] = PUP_PBS_TX;
710 mode_config[PBS_RX_MODE] = PUP_PBS_RX;
711
712 /* num of training modes */
713 for (i = 0; i < MAX_TRAINING_MODE; i++) {
714 tmp_cs = dram_info->cs_ena;
715 /* num of CS */
716 for (cs = 0; cs < MAX_CS; cs++) {
717 if (tmp_cs & (1 << cs)) {
718 /* num of PUPs */
719 for (pup = 0; pup < dram_info->num_of_total_pups;
720 pup++) {
721 if (pup == dram_info->num_of_std_pups &&
722 dram_info->ecc_ena)
723 pup = ECC_PUP;
724 if (i == PBS_TX_DM_MODE) {
725 /*
726 * Change CS bitmask because
727 * PBS works only with CS0
728 */
729 tmp_cs = 0x1;
730 val = ddr3_read_pup_reg(
731 mode_config[i], CS0, pup);
732 } else if (i == PBS_TX_MODE ||
733 i == PBS_RX_MODE) {
734 /*
735 * Change CS bitmask because
736 * PBS works only with CS0
737 */
738 tmp_cs = 0x1;
739 for (dq = 0; dq <= DQ_NUM;
740 dq++) {
741 val = ddr3_read_pup_reg(
742 mode_config[i] + dq,
743 CS0,
744 pup);
745 (*sdram_offset) = val;
746 crc += *sdram_offset;
747 sdram_offset++;
748 regs++;
749 }
750 continue;
751 } else {
752 val = ddr3_read_pup_reg(
753 mode_config[i], cs, pup);
754 }
755
756 *sdram_offset = val;
757 crc += *sdram_offset;
758 sdram_offset++;
759 regs++;
760 }
761 }
762 }
763 }
764
765 *sdram_offset = reg_read(REG_READ_DATA_SAMPLE_DELAYS_ADDR);
766 crc += *sdram_offset;
767 sdram_offset++;
768 regs++;
769 *sdram_offset = reg_read(REG_READ_DATA_READY_DELAYS_ADDR);
770 crc += *sdram_offset;
771 sdram_offset++;
772 regs++;
773 sdram_offset = (u32 *)NUM_OF_REGISTER_ADDR;
774 *sdram_offset = regs;
775 DEBUG_SUSPEND_RESUME_S("Training Results CheckSum write= ");
776 DEBUG_SUSPEND_RESUME_D(crc, 8);
777 DEBUG_SUSPEND_RESUME_S("\n");
778 sdram_offset = (u32 *)CHECKSUM_RESULT_ADDR;
779 *sdram_offset = crc;
780 }
781
782 /*
783 * Name: ddr3_read_training_results()
784 * Desc: Reads the training results from memeory (RL,WL,PBS,Rx/Tx
785 * Centeralization)
786 * and writes them to the relevant registers
787 * Args: MV_DRAM_INFO *dram_info
788 * Notes:
789 * Returns: None.
790 */
791 int ddr3_read_training_results(void)
792 {
793 u32 val, reg, idx, dqs_wr_idx = 0, crc = 0;
794 u32 *sdram_offset = (u32 *)RESUME_TRAINING_VALUES_ADDR;
795 u32 training_val[RESUME_TRAINING_VALUES_MAX] = { 0 };
796 u32 regs = *((u32 *)NUM_OF_REGISTER_ADDR);
797
798 /*
799 * Read Training results & Dunit registers from memory and write
800 * it to an array
801 */
802 for (idx = 0; idx < regs; idx++) {
803 training_val[idx] = *sdram_offset;
804 crc += *sdram_offset;
805 sdram_offset++;
806 }
807
808 sdram_offset = (u32 *)CHECKSUM_RESULT_ADDR;
809
810 if ((*sdram_offset) == crc) {
811 DEBUG_SUSPEND_RESUME_S("Training Results CheckSum read PASS= ");
812 DEBUG_SUSPEND_RESUME_D(crc, 8);
813 DEBUG_SUSPEND_RESUME_S("\n");
814 } else {
815 DEBUG_MAIN_S("Wrong Training Results CheckSum\n");
816 return MV_FAIL;
817 }
818
819 /*
820 * We iterate through all the registers except for the last 2 since
821 * they are Dunit registers (and not PHY registers)
822 */
823 for (idx = 0; idx < (regs - 2); idx++) {
824 val = training_val[idx];
825 reg = (val >> REG_PHY_CS_OFFS) & 0x3F; /*read the phy address */
826
827 /* Check if the values belongs to the DQS WR */
828 if (reg == PUP_WL_MODE) {
829 /* bit[5:0] in DQS_WR are delay */
830 val = (training_val[dqs_wr_idx++] & 0x3F);
831 /*
832 * bit[15:10] are DQS_WR delay & bit[9:0] are
833 * WL phase & delay
834 */
835 val = (val << REG_PHY_DQS_REF_DLY_OFFS) |
836 (training_val[idx] & 0x3C003FF);
837 /* Add Request pending and write operation bits */
838 val |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
839 } else if (reg == PUP_DQS_WR) {
840 /*
841 * Do nothing since DQS_WR will be done in PUP_WL_MODE
842 */
843 continue;
844 }
845
846 val |= REG_PHY_REGISTRY_FILE_ACCESS_OP_WR;
847 reg_write(REG_PHY_REGISTRY_FILE_ACCESS_ADDR, val);
848 do {
849 val = (reg_read(REG_PHY_REGISTRY_FILE_ACCESS_ADDR)) &
850 REG_PHY_REGISTRY_FILE_ACCESS_OP_DONE;
851 } while (val); /* Wait for '0' to mark the end of the transaction */
852 }
853
854 /* write last 2 Dunit configurations */
855 val = training_val[idx];
856 reg_write(REG_READ_DATA_SAMPLE_DELAYS_ADDR, val); /* reg 0x1538 */
857 val = training_val[idx + 1];
858 reg_write(REG_READ_DATA_READY_DELAYS_ADDR, val); /* reg 0x153c */
859
860 return MV_OK;
861 }
862
863 /*
864 * Name: ddr3_check_if_resume_mode()
865 * Desc: Reads the address (0x3000) of the Resume Magic word (0xDEADB002)
866 * Args: MV_DRAM_INFO *dram_info
867 * Notes:
868 * Returns: return (magic_word == SUSPEND_MAGIC_WORD)
869 */
870 int ddr3_check_if_resume_mode(MV_DRAM_INFO *dram_info, u32 freq)
871 {
872 u32 magic_word;
873 u32 *sdram_offset = (u32 *)BOOT_INFO_ADDR;
874
875 if (dram_info->reg_dimm != 1) {
876 /*
877 * Perform write levleling in order initiate the phy with
878 * low frequency
879 */
880 if (MV_OK != ddr3_write_leveling_hw(freq, dram_info)) {
881 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Write Leveling Hw)\n");
882 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
883 }
884 }
885
886 if (MV_OK != ddr3_load_patterns(dram_info, 1)) {
887 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Loading Patterns)\n");
888 return MV_DDR3_TRAINING_ERR_LOAD_PATTERNS;
889 }
890
891 /* Enable CS0 only for RL */
892 dram_info->cs_ena = 0x1;
893
894 /* Perform Read levleling in order to get stable memory */
895 if (MV_OK != ddr3_read_leveling_hw(freq, dram_info)) {
896 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Read Leveling Hw)\n");
897 return MV_DDR3_TRAINING_ERR_WR_LVL_HW;
898 }
899
900 /* Back to relevant CS */
901 dram_info->cs_ena = ddr3_get_cs_ena_from_reg();
902
903 magic_word = *sdram_offset;
904 return magic_word == SUSPEND_MAGIC_WORD;
905 }
906
907 /*
908 * Name: ddr3_training_suspend_resume()
909 * Desc: Execute the Resume state
910 * Args: MV_DRAM_INFO *dram_info
911 * Notes:
912 * Returns: return (magic_word == SUSPEND_MAGIC_WORD)
913 */
914 int ddr3_training_suspend_resume(MV_DRAM_INFO *dram_info)
915 {
916 u32 freq, reg;
917 int tmp_ratio;
918
919 /* Configure DDR */
920 if (MV_OK != ddr3_read_training_results())
921 return MV_FAIL;
922
923 /* Reset read FIFO */
924 reg = reg_read(REG_DRAM_TRAINING_ADDR);
925
926 /* Start Auto Read Leveling procedure */
927 reg |= (1 << REG_DRAM_TRAINING_RL_OFFS);
928 reg_write(REG_DRAM_TRAINING_ADDR, reg); /* 0x15B0 - Training Register */
929
930 reg = reg_read(REG_DRAM_TRAINING_2_ADDR);
931 reg |= ((1 << REG_DRAM_TRAINING_2_FIFO_RST_OFFS) +
932 (1 << REG_DRAM_TRAINING_2_SW_OVRD_OFFS));
933
934 /* [0] = 1 - Enable SW override, [4] = 1 - FIFO reset */
935 /* 0x15B8 - Training SW 2 Register */
936 reg_write(REG_DRAM_TRAINING_2_ADDR, reg);
937
938 udelay(2);
939
940 reg = reg_read(REG_DRAM_TRAINING_ADDR);
941 /* Clear Auto Read Leveling procedure */
942 reg &= ~(1 << REG_DRAM_TRAINING_RL_OFFS);
943 reg_write(REG_DRAM_TRAINING_ADDR, reg); /* 0x15B0 - Training Register */
944
945 /* Return to target frequency */
946 freq = dram_info->target_frequency;
947 tmp_ratio = 1;
948 if (MV_OK != ddr3_dfs_low_2_high(freq, tmp_ratio, dram_info)) {
949 DEBUG_MAIN_S("DDR3 Training Sequence - FAILED (Dfs Low2High)\n");
950 return MV_DDR3_TRAINING_ERR_DFS_H2L;
951 }
952
953 if (dram_info->ecc_ena) {
954 /* Scabbling the RL area pattern and the training area */
955 mv_sys_xor_finish();
956 dram_info->num_cs = 1;
957 dram_info->cs_ena = 1;
958 mv_sys_xor_init(dram_info);
959 mv_xor_mem_init(0, RESUME_RL_PATTERNS_ADDR,
960 RESUME_RL_PATTERNS_SIZE, 0xFFFFFFFF, 0xFFFFFFFF);
961
962 /* Wait for previous transfer completion */
963
964 while (mv_xor_state_get(0) != MV_IDLE)
965 ;
966
967 /* Return XOR State */
968 mv_sys_xor_finish();
969 }
970
971 return MV_OK;
972 }
973 #endif
974
975 void ddr3_print_freq(u32 freq)
976 {
977 u32 tmp_freq;
978
979 switch (freq) {
980 case 0:
981 tmp_freq = 100;
982 break;
983 case 1:
984 tmp_freq = 300;
985 break;
986 case 2:
987 tmp_freq = 360;
988 break;
989 case 3:
990 tmp_freq = 400;
991 break;
992 case 4:
993 tmp_freq = 444;
994 break;
995 case 5:
996 tmp_freq = 500;
997 break;
998 case 6:
999 tmp_freq = 533;
1000 break;
1001 case 7:
1002 tmp_freq = 600;
1003 break;
1004 case 8:
1005 tmp_freq = 666;
1006 break;
1007 case 9:
1008 tmp_freq = 720;
1009 break;
1010 case 10:
1011 tmp_freq = 800;
1012 break;
1013 default:
1014 tmp_freq = 100;
1015 }
1016
1017 printf("Current frequency is: %dMHz\n", tmp_freq);
1018 }
1019
1020 int ddr3_get_min_max_read_sample_delay(u32 cs_enable, u32 reg, u32 *min,
1021 u32 *max, u32 *cs_max)
1022 {
1023 u32 cs, delay;
1024
1025 *min = 0xFFFFFFFF;
1026 *max = 0x0;
1027
1028 for (cs = 0; cs < MAX_CS; cs++) {
1029 if ((cs_enable & (1 << cs)) == 0)
1030 continue;
1031
1032 delay = ((reg >> (cs * 8)) & 0x1F);
1033
1034 if (delay < *min)
1035 *min = delay;
1036
1037 if (delay > *max) {
1038 *max = delay;
1039 *cs_max = cs;
1040 }
1041 }
1042
1043 return MV_OK;
1044 }
1045
1046 int ddr3_get_min_max_rl_phase(MV_DRAM_INFO *dram_info, u32 *min, u32 *max,
1047 u32 cs)
1048 {
1049 u32 pup, reg, phase;
1050
1051 *min = 0xFFFFFFFF;
1052 *max = 0x0;
1053
1054 for (pup = 0; pup < dram_info->num_of_total_pups; pup++) {
1055 reg = ddr3_read_pup_reg(PUP_RL_MODE, cs, pup);
1056 phase = ((reg >> 8) & 0x7);
1057
1058 if (phase < *min)
1059 *min = phase;
1060
1061 if (phase > *max)
1062 *max = phase;
1063 }
1064
1065 return MV_OK;
1066 }
1067
1068 int ddr3_odt_activate(int activate)
1069 {
1070 u32 reg, mask;
1071
1072 mask = (1 << REG_DUNIT_ODT_CTRL_OVRD_OFFS) |
1073 (1 << REG_DUNIT_ODT_CTRL_OVRD_VAL_OFFS);
1074 /* {0x0000149C} - DDR Dunit ODT Control Register */
1075 reg = reg_read(REG_DUNIT_ODT_CTRL_ADDR);
1076 if (activate)
1077 reg |= mask;
1078 else
1079 reg &= ~mask;
1080
1081 reg_write(REG_DUNIT_ODT_CTRL_ADDR, reg);
1082
1083 return MV_OK;
1084 }
1085
1086 int ddr3_odt_read_dynamic_config(MV_DRAM_INFO *dram_info)
1087 {
1088 u32 min_read_sample_delay, max_read_sample_delay, max_rl_phase;
1089 u32 min, max, cs_max;
1090 u32 cs_ena, reg;
1091
1092 reg = reg_read(REG_READ_DATA_SAMPLE_DELAYS_ADDR);
1093 cs_ena = ddr3_get_cs_ena_from_reg();
1094
1095 /* Get minimum and maximum of read sample delay of all CS */
1096 ddr3_get_min_max_read_sample_delay(cs_ena, reg, &min_read_sample_delay,
1097 &max_read_sample_delay, &cs_max);
1098
1099 /*
1100 * Get minimum and maximum read leveling phase which belongs to the
1101 * maximal read sample delay
1102 */
1103 ddr3_get_min_max_rl_phase(dram_info, &min, &max, cs_max);
1104 max_rl_phase = max;
1105
1106 /* DDR ODT Timing (Low) Register calculation */
1107 reg = reg_read(REG_ODT_TIME_LOW_ADDR);
1108 reg &= ~(0x1FF << REG_ODT_ON_CTL_RD_OFFS);
1109 reg |= (((min_read_sample_delay - 1) & 0xF) << REG_ODT_ON_CTL_RD_OFFS);
1110 reg |= (((max_read_sample_delay + 4 + (((max_rl_phase + 1) / 2) + 1)) &
1111 0x1F) << REG_ODT_OFF_CTL_RD_OFFS);
1112 reg_write(REG_ODT_TIME_LOW_ADDR, reg);
1113
1114 return MV_OK;
1115 }