xref: /rk3399_ARM-atf/bl31/aarch64/bl31_entrypoint.S (revision 88cfd9a604cee29cc6536a43cd136b7eef0b78b9)
14f6ad66aSAchin Gupta/*
2*88cfd9a6SAntonio Nino Diaz * Copyright (c) 2013-2019, ARM Limited and Contributors. All rights reserved.
34f6ad66aSAchin Gupta *
482cb2c1aSdp-arm * SPDX-License-Identifier: BSD-3-Clause
54f6ad66aSAchin Gupta */
64f6ad66aSAchin Gupta
7931f7c61SSoby Mathew#include <platform_def.h>
809d40e0eSAntonio Nino Diaz
909d40e0eSAntonio Nino Diaz#include <arch.h>
1009d40e0eSAntonio Nino Diaz#include <common/bl_common.h>
1109d40e0eSAntonio Nino Diaz#include <el3_common_macros.S>
1209d40e0eSAntonio Nino Diaz#include <lib/pmf/pmf_asm_macros.S>
1309d40e0eSAntonio Nino Diaz#include <lib/runtime_instr.h>
1409d40e0eSAntonio Nino Diaz#include <lib/xlat_tables/xlat_mmu_helpers.h>
154f6ad66aSAchin Gupta
164f6ad66aSAchin Gupta	.globl	bl31_entrypoint
17cf0b1492SSoby Mathew	.globl	bl31_warm_entrypoint
184f6ad66aSAchin Gupta
194f6ad66aSAchin Gupta	/* -----------------------------------------------------
204f6ad66aSAchin Gupta	 * bl31_entrypoint() is the cold boot entrypoint,
214f6ad66aSAchin Gupta	 * executed only by the primary cpu.
224f6ad66aSAchin Gupta	 * -----------------------------------------------------
234f6ad66aSAchin Gupta	 */
244f6ad66aSAchin Gupta
250a30cf54SAndrew Thoelkefunc bl31_entrypoint
2652010cc7SSandrine Bailleux#if !RESET_TO_BL31
274112bfa0SVikram Kanigiri	/* ---------------------------------------------------------------
28a6f340feSSoby Mathew	 * Stash the previous bootloader arguments x0 - x3 for later use.
294112bfa0SVikram Kanigiri	 * ---------------------------------------------------------------
30c10bd2ceSSandrine Bailleux	 */
3129fb905dSVikram Kanigiri	mov	x20, x0
3229fb905dSVikram Kanigiri	mov	x21, x1
33a6f340feSSoby Mathew	mov	x22, x2
34a6f340feSSoby Mathew	mov	x23, x3
35c10bd2ceSSandrine Bailleux
364f603683SHarry Liebel	/* ---------------------------------------------------------------------
3752010cc7SSandrine Bailleux	 * For !RESET_TO_BL31 systems, only the primary CPU ever reaches
3852010cc7SSandrine Bailleux	 * bl31_entrypoint() during the cold boot flow, so the cold/warm boot
3952010cc7SSandrine Bailleux	 * and primary/secondary CPU logic should not be executed in this case.
404f603683SHarry Liebel	 *
4118f2efd6SDavid Cunado	 * Also, assume that the previous bootloader has already initialised the
4218f2efd6SDavid Cunado	 * SCTLR_EL3, including the endianness, and has initialised the memory.
434f603683SHarry Liebel	 * ---------------------------------------------------------------------
444f603683SHarry Liebel	 */
4552010cc7SSandrine Bailleux	el3_entrypoint_common					\
4618f2efd6SDavid Cunado		_init_sctlr=0					\
4752010cc7SSandrine Bailleux		_warm_boot_mailbox=0				\
4852010cc7SSandrine Bailleux		_secondary_cold_boot=0				\
4952010cc7SSandrine Bailleux		_init_memory=0					\
5052010cc7SSandrine Bailleux		_init_c_runtime=1				\
5152010cc7SSandrine Bailleux		_exception_vectors=runtime_exceptions
5252010cc7SSandrine Bailleux#else
53bf031bbaSSandrine Bailleux	/* ---------------------------------------------------------------------
54bf031bbaSSandrine Bailleux	 * For RESET_TO_BL31 systems which have a programmable reset address,
55bf031bbaSSandrine Bailleux	 * bl31_entrypoint() is executed only on the cold boot path so we can
56bf031bbaSSandrine Bailleux	 * skip the warm boot mailbox mechanism.
57bf031bbaSSandrine Bailleux	 * ---------------------------------------------------------------------
58bf031bbaSSandrine Bailleux	 */
5952010cc7SSandrine Bailleux	el3_entrypoint_common					\
6018f2efd6SDavid Cunado		_init_sctlr=1					\
61bf031bbaSSandrine Bailleux		_warm_boot_mailbox=!PROGRAMMABLE_RESET_ADDRESS	\
62a9bec67dSSandrine Bailleux		_secondary_cold_boot=!COLD_BOOT_SINGLE_CPU	\
6352010cc7SSandrine Bailleux		_init_memory=1					\
6452010cc7SSandrine Bailleux		_init_c_runtime=1				\
6552010cc7SSandrine Bailleux		_exception_vectors=runtime_exceptions
664f6ad66aSAchin Gupta
6752010cc7SSandrine Bailleux	/* ---------------------------------------------------------------------
68d178637dSJuan Castillo	 * For RESET_TO_BL31 systems, BL31 is the first bootloader to run so
6952010cc7SSandrine Bailleux	 * there's no argument to relay from a previous bootloader. Zero the
7052010cc7SSandrine Bailleux	 * arguments passed to the platform layer to reflect that.
7152010cc7SSandrine Bailleux	 * ---------------------------------------------------------------------
7265f546a1SSandrine Bailleux	 */
73a6f340feSSoby Mathew	mov	x20, 0
74a6f340feSSoby Mathew	mov	x21, 0
75a6f340feSSoby Mathew	mov	x22, 0
76a6f340feSSoby Mathew	mov	x23, 0
7752010cc7SSandrine Bailleux#endif /* RESET_TO_BL31 */
78931f7c61SSoby Mathew
79931f7c61SSoby Mathew	/* --------------------------------------------------------------------
80931f7c61SSoby Mathew	 * If PIE is enabled, fixup the Global descriptor Table and dynamic
81931f7c61SSoby Mathew	 * relocations
82931f7c61SSoby Mathew	 * --------------------------------------------------------------------
83931f7c61SSoby Mathew	 */
84931f7c61SSoby Mathew#if ENABLE_PIE
85931f7c61SSoby Mathew	mov_imm	x0, BL31_BASE
86931f7c61SSoby Mathew	mov_imm	x1, BL31_LIMIT
87931f7c61SSoby Mathew	bl	fixup_gdt_reloc
88931f7c61SSoby Mathew#endif /* ENABLE_PIE */
89931f7c61SSoby Mathew
90*88cfd9a6SAntonio Nino Diaz	/* --------------------------------------------------------------------
91*88cfd9a6SAntonio Nino Diaz	 * Perform BL31 setup
92*88cfd9a6SAntonio Nino Diaz	 * --------------------------------------------------------------------
934f6ad66aSAchin Gupta	 */
94a6f340feSSoby Mathew	mov	x0, x20
95a6f340feSSoby Mathew	mov	x1, x21
96a6f340feSSoby Mathew	mov	x2, x22
97a6f340feSSoby Mathew	mov	x3, x23
98*88cfd9a6SAntonio Nino Diaz	bl	bl31_setup
994f6ad66aSAchin Gupta
100*88cfd9a6SAntonio Nino Diaz	/* --------------------------------------------------------------------
101*88cfd9a6SAntonio Nino Diaz	 * Enable pointer authentication
102*88cfd9a6SAntonio Nino Diaz	 * --------------------------------------------------------------------
103*88cfd9a6SAntonio Nino Diaz	 */
104*88cfd9a6SAntonio Nino Diaz#if ENABLE_PAUTH
105*88cfd9a6SAntonio Nino Diaz	mrs	x0, sctlr_el3
106*88cfd9a6SAntonio Nino Diaz	orr	x0, x0, #SCTLR_EnIA_BIT
107*88cfd9a6SAntonio Nino Diaz	msr	sctlr_el3, x0
108*88cfd9a6SAntonio Nino Diaz	isb
109*88cfd9a6SAntonio Nino Diaz#endif /* ENABLE_PAUTH */
110*88cfd9a6SAntonio Nino Diaz
111*88cfd9a6SAntonio Nino Diaz	/* --------------------------------------------------------------------
1124f6ad66aSAchin Gupta	 * Jump to main function.
113*88cfd9a6SAntonio Nino Diaz	 * --------------------------------------------------------------------
1144f6ad66aSAchin Gupta	 */
1154f6ad66aSAchin Gupta	bl	bl31_main
1164f6ad66aSAchin Gupta
117*88cfd9a6SAntonio Nino Diaz	/* --------------------------------------------------------------------
11854dc71e7SAchin Gupta	 * Clean the .data & .bss sections to main memory. This ensures
11954dc71e7SAchin Gupta	 * that any global data which was initialised by the primary CPU
12054dc71e7SAchin Gupta	 * is visible to secondary CPUs before they enable their data
12154dc71e7SAchin Gupta	 * caches and participate in coherency.
122*88cfd9a6SAntonio Nino Diaz	 * --------------------------------------------------------------------
12354dc71e7SAchin Gupta	 */
12454dc71e7SAchin Gupta	adr	x0, __DATA_START__
12554dc71e7SAchin Gupta	adr	x1, __DATA_END__
12654dc71e7SAchin Gupta	sub	x1, x1, x0
12754dc71e7SAchin Gupta	bl	clean_dcache_range
12854dc71e7SAchin Gupta
12954dc71e7SAchin Gupta	adr	x0, __BSS_START__
13054dc71e7SAchin Gupta	adr	x1, __BSS_END__
13154dc71e7SAchin Gupta	sub	x1, x1, x0
13254dc71e7SAchin Gupta	bl	clean_dcache_range
13354dc71e7SAchin Gupta
134caa84939SJeenu Viswambharan	b	el3_exit
1358b779620SKévin Petitendfunc bl31_entrypoint
136cf0b1492SSoby Mathew
137cf0b1492SSoby Mathew	/* --------------------------------------------------------------------
138cf0b1492SSoby Mathew	 * This CPU has been physically powered up. It is either resuming from
139cf0b1492SSoby Mathew	 * suspend or has simply been turned on. In both cases, call the BL31
140cf0b1492SSoby Mathew	 * warmboot entrypoint
141cf0b1492SSoby Mathew	 * --------------------------------------------------------------------
142cf0b1492SSoby Mathew	 */
143cf0b1492SSoby Mathewfunc bl31_warm_entrypoint
144872be88aSdp-arm#if ENABLE_RUNTIME_INSTRUMENTATION
145872be88aSdp-arm
146872be88aSdp-arm	/*
147872be88aSdp-arm	 * This timestamp update happens with cache off.  The next
148872be88aSdp-arm	 * timestamp collection will need to do cache maintenance prior
149872be88aSdp-arm	 * to timestamp update.
150872be88aSdp-arm	 */
15181542c00SAntonio Nino Diaz	pmf_calc_timestamp_addr rt_instr_svc, RT_INSTR_EXIT_HW_LOW_PWR
152872be88aSdp-arm	mrs	x1, cntpct_el0
153872be88aSdp-arm	str	x1, [x0]
154872be88aSdp-arm#endif
155872be88aSdp-arm
156cf0b1492SSoby Mathew	/*
157cf0b1492SSoby Mathew	 * On the warm boot path, most of the EL3 initialisations performed by
158cf0b1492SSoby Mathew	 * 'el3_entrypoint_common' must be skipped:
159cf0b1492SSoby Mathew	 *
160cf0b1492SSoby Mathew	 *  - Only when the platform bypasses the BL1/BL31 entrypoint by
16118f2efd6SDavid Cunado	 *    programming the reset address do we need to initialise SCTLR_EL3.
162cf0b1492SSoby Mathew	 *    In other cases, we assume this has been taken care by the
163cf0b1492SSoby Mathew	 *    entrypoint code.
164cf0b1492SSoby Mathew	 *
165cf0b1492SSoby Mathew	 *  - No need to determine the type of boot, we know it is a warm boot.
166cf0b1492SSoby Mathew	 *
167cf0b1492SSoby Mathew	 *  - Do not try to distinguish between primary and secondary CPUs, this
168cf0b1492SSoby Mathew	 *    notion only exists for a cold boot.
169cf0b1492SSoby Mathew	 *
170cf0b1492SSoby Mathew	 *  - No need to initialise the memory or the C runtime environment,
171cf0b1492SSoby Mathew	 *    it has been done once and for all on the cold boot path.
172cf0b1492SSoby Mathew	 */
173cf0b1492SSoby Mathew	el3_entrypoint_common					\
17418f2efd6SDavid Cunado		_init_sctlr=PROGRAMMABLE_RESET_ADDRESS		\
175cf0b1492SSoby Mathew		_warm_boot_mailbox=0				\
176cf0b1492SSoby Mathew		_secondary_cold_boot=0				\
177cf0b1492SSoby Mathew		_init_memory=0					\
178cf0b1492SSoby Mathew		_init_c_runtime=0				\
179cf0b1492SSoby Mathew		_exception_vectors=runtime_exceptions
180cf0b1492SSoby Mathew
18125a93f7cSJeenu Viswambharan	/*
18225a93f7cSJeenu Viswambharan	 * We're about to enable MMU and participate in PSCI state coordination.
18325a93f7cSJeenu Viswambharan	 *
18425a93f7cSJeenu Viswambharan	 * The PSCI implementation invokes platform routines that enable CPUs to
18525a93f7cSJeenu Viswambharan	 * participate in coherency. On a system where CPUs are not
186bcc3c49cSSoby Mathew	 * cache-coherent without appropriate platform specific programming,
187bcc3c49cSSoby Mathew	 * having caches enabled until such time might lead to coherency issues
188bcc3c49cSSoby Mathew	 * (resulting from stale data getting speculatively fetched, among
189bcc3c49cSSoby Mathew	 * others). Therefore we keep data caches disabled even after enabling
190bcc3c49cSSoby Mathew	 * the MMU for such platforms.
19125a93f7cSJeenu Viswambharan	 *
192bcc3c49cSSoby Mathew	 * On systems with hardware-assisted coherency, or on single cluster
193bcc3c49cSSoby Mathew	 * platforms, such platform specific programming is not required to
194bcc3c49cSSoby Mathew	 * enter coherency (as CPUs already are); and there's no reason to have
195bcc3c49cSSoby Mathew	 * caches disabled either.
196cf0b1492SSoby Mathew	 */
197bcc3c49cSSoby Mathew#if HW_ASSISTED_COHERENCY || WARMBOOT_ENABLE_DCACHE_EARLY
19864ee263eSJeenu Viswambharan	mov	x0, xzr
19964ee263eSJeenu Viswambharan#else
20064ee263eSJeenu Viswambharan	mov	x0, #DISABLE_DCACHE
201bcc3c49cSSoby Mathew#endif
20264ee263eSJeenu Viswambharan	bl	bl31_plat_enable_mmu
203bcc3c49cSSoby Mathew
204cf0b1492SSoby Mathew	bl	psci_warmboot_entrypoint
205cf0b1492SSoby Mathew
206872be88aSdp-arm#if ENABLE_RUNTIME_INSTRUMENTATION
20781542c00SAntonio Nino Diaz	pmf_calc_timestamp_addr rt_instr_svc, RT_INSTR_EXIT_PSCI
208872be88aSdp-arm	mov	x19, x0
209872be88aSdp-arm
210872be88aSdp-arm	/*
211872be88aSdp-arm	 * Invalidate before updating timestamp to ensure previous timestamp
212872be88aSdp-arm	 * updates on the same cache line with caches disabled are properly
213872be88aSdp-arm	 * seen by the same core. Without the cache invalidate, the core might
214872be88aSdp-arm	 * write into a stale cache line.
215872be88aSdp-arm	 */
216872be88aSdp-arm	mov	x1, #PMF_TS_SIZE
217872be88aSdp-arm	mov	x20, x30
218872be88aSdp-arm	bl	inv_dcache_range
219872be88aSdp-arm	mov	x30, x20
220872be88aSdp-arm
221872be88aSdp-arm	mrs	x0, cntpct_el0
222872be88aSdp-arm	str	x0, [x19]
223872be88aSdp-arm#endif
224cf0b1492SSoby Mathew	b	el3_exit
225cf0b1492SSoby Mathewendfunc bl31_warm_entrypoint
226