1532ed618SSoby Mathew /* 201cf14ddSMaksims Svecovs * Copyright (c) 2013-2023, Arm Limited and Contributors. All rights reserved. 32b28727eSVarun Wadekar * Copyright (c) 2022, NVIDIA Corporation. All rights reserved. 4532ed618SSoby Mathew * 582cb2c1aSdp-arm * SPDX-License-Identifier: BSD-3-Clause 6532ed618SSoby Mathew */ 7532ed618SSoby Mathew 8532ed618SSoby Mathew #include <assert.h> 940daecc1SAntonio Nino Diaz #include <stdbool.h> 10532ed618SSoby Mathew #include <string.h> 1109d40e0eSAntonio Nino Diaz 1209d40e0eSAntonio Nino Diaz #include <platform_def.h> 1309d40e0eSAntonio Nino Diaz 1409d40e0eSAntonio Nino Diaz #include <arch.h> 1509d40e0eSAntonio Nino Diaz #include <arch_helpers.h> 16b7e398d6SSoby Mathew #include <arch_features.h> 1709d40e0eSAntonio Nino Diaz #include <bl31/interrupt_mgmt.h> 1809d40e0eSAntonio Nino Diaz #include <common/bl_common.h> 19885e2683SClaus Pedersen #include <common/debug.h> 2009d40e0eSAntonio Nino Diaz #include <context.h> 218b95e848SZelalem Aweke #include <drivers/arm/gicv3.h> 2209d40e0eSAntonio Nino Diaz #include <lib/el3_runtime/context_mgmt.h> 2309d40e0eSAntonio Nino Diaz #include <lib/el3_runtime/pubsub_events.h> 2409d40e0eSAntonio Nino Diaz #include <lib/extensions/amu.h> 25744ad974Sjohpow01 #include <lib/extensions/brbe.h> 2609d40e0eSAntonio Nino Diaz #include <lib/extensions/mpam.h> 27dc78e62dSjohpow01 #include <lib/extensions/sme.h> 2809d40e0eSAntonio Nino Diaz #include <lib/extensions/spe.h> 2909d40e0eSAntonio Nino Diaz #include <lib/extensions/sve.h> 30d4582d30SManish V Badarkhe #include <lib/extensions/sys_reg_trace.h> 31813524eaSManish V Badarkhe #include <lib/extensions/trbe.h> 328fcd3d96SManish V Badarkhe #include <lib/extensions/trf.h> 3309d40e0eSAntonio Nino Diaz #include <lib/utils.h> 34532ed618SSoby Mathew 35781d07a4SJayanth Dodderi Chidanand #if ENABLE_FEAT_TWED 36781d07a4SJayanth Dodderi Chidanand /* Make sure delay value fits within the range(0-15) */ 37781d07a4SJayanth Dodderi Chidanand CASSERT(((TWED_DELAY & ~SCR_TWEDEL_MASK) == 0U), assert_twed_delay_value_check); 38781d07a4SJayanth Dodderi Chidanand #endif /* ENABLE_FEAT_TWED */ 39532ed618SSoby Mathew 40781d07a4SJayanth Dodderi Chidanand static void manage_extensions_secure(cpu_context_t *ctx); 41b515f541SZelalem Aweke 42b515f541SZelalem Aweke static void setup_el1_context(cpu_context_t *ctx, const struct entry_point_info *ep) 43b515f541SZelalem Aweke { 44b515f541SZelalem Aweke u_register_t sctlr_elx, actlr_elx; 45b515f541SZelalem Aweke 46b515f541SZelalem Aweke /* 47b515f541SZelalem Aweke * Initialise SCTLR_EL1 to the reset value corresponding to the target 48b515f541SZelalem Aweke * execution state setting all fields rather than relying on the hw. 49b515f541SZelalem Aweke * Some fields have architecturally UNKNOWN reset values and these are 50b515f541SZelalem Aweke * set to zero. 51b515f541SZelalem Aweke * 52b515f541SZelalem Aweke * SCTLR.EE: Endianness is taken from the entrypoint attributes. 53b515f541SZelalem Aweke * 54b515f541SZelalem Aweke * SCTLR.M, SCTLR.C and SCTLR.I: These fields must be zero (as 55b515f541SZelalem Aweke * required by PSCI specification) 56b515f541SZelalem Aweke */ 57b515f541SZelalem Aweke sctlr_elx = (EP_GET_EE(ep->h.attr) != 0U) ? SCTLR_EE_BIT : 0UL; 58b515f541SZelalem Aweke if (GET_RW(ep->spsr) == MODE_RW_64) { 59b515f541SZelalem Aweke sctlr_elx |= SCTLR_EL1_RES1; 60b515f541SZelalem Aweke } else { 61b515f541SZelalem Aweke /* 62b515f541SZelalem Aweke * If the target execution state is AArch32 then the following 63b515f541SZelalem Aweke * fields need to be set. 64b515f541SZelalem Aweke * 65b515f541SZelalem Aweke * SCTRL_EL1.nTWE: Set to one so that EL0 execution of WFE 66b515f541SZelalem Aweke * instructions are not trapped to EL1. 67b515f541SZelalem Aweke * 68b515f541SZelalem Aweke * SCTLR_EL1.nTWI: Set to one so that EL0 execution of WFI 69b515f541SZelalem Aweke * instructions are not trapped to EL1. 70b515f541SZelalem Aweke * 71b515f541SZelalem Aweke * SCTLR_EL1.CP15BEN: Set to one to enable EL0 execution of the 72b515f541SZelalem Aweke * CP15DMB, CP15DSB, and CP15ISB instructions. 73b515f541SZelalem Aweke */ 74b515f541SZelalem Aweke sctlr_elx |= SCTLR_AARCH32_EL1_RES1 | SCTLR_CP15BEN_BIT 75b515f541SZelalem Aweke | SCTLR_NTWI_BIT | SCTLR_NTWE_BIT; 76b515f541SZelalem Aweke } 77b515f541SZelalem Aweke 78b515f541SZelalem Aweke #if ERRATA_A75_764081 79b515f541SZelalem Aweke /* 80b515f541SZelalem Aweke * If workaround of errata 764081 for Cortex-A75 is used then set 81b515f541SZelalem Aweke * SCTLR_EL1.IESB to enable Implicit Error Synchronization Barrier. 82b515f541SZelalem Aweke */ 83b515f541SZelalem Aweke sctlr_elx |= SCTLR_IESB_BIT; 84b515f541SZelalem Aweke #endif 85b515f541SZelalem Aweke /* Store the initialised SCTLR_EL1 value in the cpu_context */ 86b515f541SZelalem Aweke write_ctx_reg(get_el1_sysregs_ctx(ctx), CTX_SCTLR_EL1, sctlr_elx); 87b515f541SZelalem Aweke 88b515f541SZelalem Aweke /* 89b515f541SZelalem Aweke * Base the context ACTLR_EL1 on the current value, as it is 90b515f541SZelalem Aweke * implementation defined. The context restore process will write 91b515f541SZelalem Aweke * the value from the context to the actual register and can cause 92b515f541SZelalem Aweke * problems for processor cores that don't expect certain bits to 93b515f541SZelalem Aweke * be zero. 94b515f541SZelalem Aweke */ 95b515f541SZelalem Aweke actlr_elx = read_actlr_el1(); 96b515f541SZelalem Aweke write_ctx_reg((get_el1_sysregs_ctx(ctx)), (CTX_ACTLR_EL1), (actlr_elx)); 97b515f541SZelalem Aweke } 98b515f541SZelalem Aweke 992bbad1d1SZelalem Aweke /****************************************************************************** 1002bbad1d1SZelalem Aweke * This function performs initializations that are specific to SECURE state 1012bbad1d1SZelalem Aweke * and updates the cpu context specified by 'ctx'. 1022bbad1d1SZelalem Aweke *****************************************************************************/ 1032bbad1d1SZelalem Aweke static void setup_secure_context(cpu_context_t *ctx, const struct entry_point_info *ep) 104532ed618SSoby Mathew { 1052bbad1d1SZelalem Aweke u_register_t scr_el3; 1062bbad1d1SZelalem Aweke el3_state_t *state; 1072bbad1d1SZelalem Aweke 1082bbad1d1SZelalem Aweke state = get_el3state_ctx(ctx); 1092bbad1d1SZelalem Aweke scr_el3 = read_ctx_reg(state, CTX_SCR_EL3); 1102bbad1d1SZelalem Aweke 1112bbad1d1SZelalem Aweke #if defined(IMAGE_BL31) && !defined(SPD_spmd) 112532ed618SSoby Mathew /* 1132bbad1d1SZelalem Aweke * SCR_EL3.IRQ, SCR_EL3.FIQ: Enable the physical FIQ and IRQ routing as 1142bbad1d1SZelalem Aweke * indicated by the interrupt routing model for BL31. 115532ed618SSoby Mathew */ 1162bbad1d1SZelalem Aweke scr_el3 |= get_scr_el3_from_routing_model(SECURE); 1172bbad1d1SZelalem Aweke #endif 1182bbad1d1SZelalem Aweke 1192bbad1d1SZelalem Aweke #if !CTX_INCLUDE_MTE_REGS || ENABLE_ASSERTIONS 1202bbad1d1SZelalem Aweke /* Get Memory Tagging Extension support level */ 1212bbad1d1SZelalem Aweke unsigned int mte = get_armv8_5_mte_support(); 1222bbad1d1SZelalem Aweke #endif 1232bbad1d1SZelalem Aweke /* 1242bbad1d1SZelalem Aweke * Allow access to Allocation Tags when CTX_INCLUDE_MTE_REGS 1252bbad1d1SZelalem Aweke * is set, or when MTE is only implemented at EL0. 1262bbad1d1SZelalem Aweke */ 1272bbad1d1SZelalem Aweke #if CTX_INCLUDE_MTE_REGS 1282bbad1d1SZelalem Aweke assert((mte == MTE_IMPLEMENTED_ELX) || (mte == MTE_IMPLEMENTED_ASY)); 1292bbad1d1SZelalem Aweke scr_el3 |= SCR_ATA_BIT; 1302bbad1d1SZelalem Aweke #else 1312bbad1d1SZelalem Aweke if (mte == MTE_IMPLEMENTED_EL0) { 1322bbad1d1SZelalem Aweke scr_el3 |= SCR_ATA_BIT; 1332bbad1d1SZelalem Aweke } 1342bbad1d1SZelalem Aweke #endif /* CTX_INCLUDE_MTE_REGS */ 1352bbad1d1SZelalem Aweke 1362bbad1d1SZelalem Aweke /* Enable S-EL2 if the next EL is EL2 and S-EL2 is present */ 1372bbad1d1SZelalem Aweke if ((GET_EL(ep->spsr) == MODE_EL2) && is_armv8_4_sel2_present()) { 1382bbad1d1SZelalem Aweke if (GET_RW(ep->spsr) != MODE_RW_64) { 1392bbad1d1SZelalem Aweke ERROR("S-EL2 can not be used in AArch32\n."); 1402bbad1d1SZelalem Aweke panic(); 1412bbad1d1SZelalem Aweke } 1422bbad1d1SZelalem Aweke 1432bbad1d1SZelalem Aweke scr_el3 |= SCR_EEL2_BIT; 1442bbad1d1SZelalem Aweke } 1452bbad1d1SZelalem Aweke 1462bbad1d1SZelalem Aweke write_ctx_reg(state, CTX_SCR_EL3, scr_el3); 1472bbad1d1SZelalem Aweke 148b515f541SZelalem Aweke /* 149b515f541SZelalem Aweke * Initialize EL1 context registers unless SPMC is running 150b515f541SZelalem Aweke * at S-EL2. 151b515f541SZelalem Aweke */ 152b515f541SZelalem Aweke #if !SPMD_SPM_AT_SEL2 153b515f541SZelalem Aweke setup_el1_context(ctx, ep); 154b515f541SZelalem Aweke #endif 155b515f541SZelalem Aweke 1562bbad1d1SZelalem Aweke manage_extensions_secure(ctx); 1572bbad1d1SZelalem Aweke } 1582bbad1d1SZelalem Aweke 1592bbad1d1SZelalem Aweke #if ENABLE_RME 1602bbad1d1SZelalem Aweke /****************************************************************************** 1612bbad1d1SZelalem Aweke * This function performs initializations that are specific to REALM state 1622bbad1d1SZelalem Aweke * and updates the cpu context specified by 'ctx'. 1632bbad1d1SZelalem Aweke *****************************************************************************/ 1642bbad1d1SZelalem Aweke static void setup_realm_context(cpu_context_t *ctx, const struct entry_point_info *ep) 1652bbad1d1SZelalem Aweke { 1662bbad1d1SZelalem Aweke u_register_t scr_el3; 1672bbad1d1SZelalem Aweke el3_state_t *state; 1682bbad1d1SZelalem Aweke 1692bbad1d1SZelalem Aweke state = get_el3state_ctx(ctx); 1702bbad1d1SZelalem Aweke scr_el3 = read_ctx_reg(state, CTX_SCR_EL3); 1712bbad1d1SZelalem Aweke 17201cf14ddSMaksims Svecovs scr_el3 |= SCR_NS_BIT | SCR_NSE_BIT; 17301cf14ddSMaksims Svecovs 17401cf14ddSMaksims Svecovs #if ENABLE_FEAT_CSV2_2 17501cf14ddSMaksims Svecovs /* Enable access to the SCXTNUM_ELx registers. */ 17601cf14ddSMaksims Svecovs scr_el3 |= SCR_EnSCXT_BIT; 17701cf14ddSMaksims Svecovs #endif 1782bbad1d1SZelalem Aweke 1792bbad1d1SZelalem Aweke write_ctx_reg(state, CTX_SCR_EL3, scr_el3); 1802bbad1d1SZelalem Aweke } 1812bbad1d1SZelalem Aweke #endif /* ENABLE_RME */ 1822bbad1d1SZelalem Aweke 1832bbad1d1SZelalem Aweke /****************************************************************************** 1842bbad1d1SZelalem Aweke * This function performs initializations that are specific to NON-SECURE state 1852bbad1d1SZelalem Aweke * and updates the cpu context specified by 'ctx'. 1862bbad1d1SZelalem Aweke *****************************************************************************/ 1872bbad1d1SZelalem Aweke static void setup_ns_context(cpu_context_t *ctx, const struct entry_point_info *ep) 1882bbad1d1SZelalem Aweke { 1892bbad1d1SZelalem Aweke u_register_t scr_el3; 1902bbad1d1SZelalem Aweke el3_state_t *state; 1912bbad1d1SZelalem Aweke 1922bbad1d1SZelalem Aweke state = get_el3state_ctx(ctx); 1932bbad1d1SZelalem Aweke scr_el3 = read_ctx_reg(state, CTX_SCR_EL3); 1942bbad1d1SZelalem Aweke 1952bbad1d1SZelalem Aweke /* SCR_NS: Set the NS bit */ 1962bbad1d1SZelalem Aweke scr_el3 |= SCR_NS_BIT; 1972bbad1d1SZelalem Aweke 1982bbad1d1SZelalem Aweke #if !CTX_INCLUDE_PAUTH_REGS 1992bbad1d1SZelalem Aweke /* 2002bbad1d1SZelalem Aweke * If the pointer authentication registers aren't saved during world 2012bbad1d1SZelalem Aweke * switches the value of the registers can be leaked from the Secure to 2022bbad1d1SZelalem Aweke * the Non-secure world. To prevent this, rather than enabling pointer 2032bbad1d1SZelalem Aweke * authentication everywhere, we only enable it in the Non-secure world. 2042bbad1d1SZelalem Aweke * 2052bbad1d1SZelalem Aweke * If the Secure world wants to use pointer authentication, 2062bbad1d1SZelalem Aweke * CTX_INCLUDE_PAUTH_REGS must be set to 1. 2072bbad1d1SZelalem Aweke */ 2082bbad1d1SZelalem Aweke scr_el3 |= SCR_API_BIT | SCR_APK_BIT; 2092bbad1d1SZelalem Aweke #endif /* !CTX_INCLUDE_PAUTH_REGS */ 2102bbad1d1SZelalem Aweke 2112bbad1d1SZelalem Aweke /* Allow access to Allocation Tags when MTE is implemented. */ 2122bbad1d1SZelalem Aweke scr_el3 |= SCR_ATA_BIT; 2132bbad1d1SZelalem Aweke 21446cc41d5SManish Pandey #if HANDLE_EA_EL3_FIRST_NS 21546cc41d5SManish Pandey /* SCR_EL3.EA: Route External Abort and SError Interrupt to EL3. */ 21646cc41d5SManish Pandey scr_el3 |= SCR_EA_BIT; 21746cc41d5SManish Pandey #endif 21846cc41d5SManish Pandey 21900e8f79cSManish Pandey #if RAS_TRAP_NS_ERR_REC_ACCESS 22000e8f79cSManish Pandey /* 22100e8f79cSManish Pandey * SCR_EL3.TERR: Trap Error record accesses. Accesses to the RAS ERR 22200e8f79cSManish Pandey * and RAS ERX registers from EL1 and EL2(from any security state) 22300e8f79cSManish Pandey * are trapped to EL3. 22400e8f79cSManish Pandey * Set here to trap only for NS EL1/EL2 22500e8f79cSManish Pandey * 22600e8f79cSManish Pandey */ 22700e8f79cSManish Pandey scr_el3 |= SCR_TERR_BIT; 22800e8f79cSManish Pandey #endif 22900e8f79cSManish Pandey 23001cf14ddSMaksims Svecovs #if ENABLE_FEAT_CSV2_2 23101cf14ddSMaksims Svecovs /* Enable access to the SCXTNUM_ELx registers. */ 23201cf14ddSMaksims Svecovs scr_el3 |= SCR_EnSCXT_BIT; 23301cf14ddSMaksims Svecovs #endif 23401cf14ddSMaksims Svecovs 2352bbad1d1SZelalem Aweke #ifdef IMAGE_BL31 2362bbad1d1SZelalem Aweke /* 2372bbad1d1SZelalem Aweke * SCR_EL3.IRQ, SCR_EL3.FIQ: Enable the physical FIQ and IRQ routing as 2382bbad1d1SZelalem Aweke * indicated by the interrupt routing model for BL31. 2392bbad1d1SZelalem Aweke */ 2402bbad1d1SZelalem Aweke scr_el3 |= get_scr_el3_from_routing_model(NON_SECURE); 2412bbad1d1SZelalem Aweke #endif 2422bbad1d1SZelalem Aweke write_ctx_reg(state, CTX_SCR_EL3, scr_el3); 2438b95e848SZelalem Aweke 244b515f541SZelalem Aweke /* Initialize EL1 context registers */ 245b515f541SZelalem Aweke setup_el1_context(ctx, ep); 246b515f541SZelalem Aweke 2478b95e848SZelalem Aweke /* Initialize EL2 context registers */ 2488b95e848SZelalem Aweke #if CTX_INCLUDE_EL2_REGS 2498b95e848SZelalem Aweke 2508b95e848SZelalem Aweke /* 2518b95e848SZelalem Aweke * Initialize SCTLR_EL2 context register using Endianness value 2528b95e848SZelalem Aweke * taken from the entrypoint attribute. 2538b95e848SZelalem Aweke */ 2548b95e848SZelalem Aweke u_register_t sctlr_el2 = (EP_GET_EE(ep->h.attr) != 0U) ? SCTLR_EE_BIT : 0UL; 2558b95e848SZelalem Aweke sctlr_el2 |= SCTLR_EL2_RES1; 2568b95e848SZelalem Aweke write_ctx_reg(get_el2_sysregs_ctx(ctx), CTX_SCTLR_EL2, 2578b95e848SZelalem Aweke sctlr_el2); 2588b95e848SZelalem Aweke 2598b95e848SZelalem Aweke /* 2602b28727eSVarun Wadekar * Program the ICC_SRE_EL2 to make sure the correct bits are set 2612b28727eSVarun Wadekar * when restoring NS context. 2628b95e848SZelalem Aweke */ 2632b28727eSVarun Wadekar u_register_t icc_sre_el2 = ICC_SRE_DIB_BIT | ICC_SRE_DFB_BIT | 2642b28727eSVarun Wadekar ICC_SRE_EN_BIT | ICC_SRE_SRE_BIT; 2658b95e848SZelalem Aweke write_ctx_reg(get_el2_sysregs_ctx(ctx), CTX_ICC_SRE_EL2, 2668b95e848SZelalem Aweke icc_sre_el2); 2677f856198SBoyan Karatotev 2687f856198SBoyan Karatotev /* 2697f856198SBoyan Karatotev * Initialize MDCR_EL2.HPMN to its hardware reset value so we don't 2707f856198SBoyan Karatotev * throw anyone off who expects this to be sensible. 2717f856198SBoyan Karatotev * TODO: A similar thing happens in cm_prepare_el3_exit. They should be 2727f856198SBoyan Karatotev * unified with the proper PMU implementation 2737f856198SBoyan Karatotev */ 2747f856198SBoyan Karatotev u_register_t mdcr_el2 = ((read_pmcr_el0() >> PMCR_EL0_N_SHIFT) & 2757f856198SBoyan Karatotev PMCR_EL0_N_MASK); 2767f856198SBoyan Karatotev write_ctx_reg(get_el2_sysregs_ctx(ctx), CTX_MDCR_EL2, mdcr_el2); 2778b95e848SZelalem Aweke #endif /* CTX_INCLUDE_EL2_REGS */ 278532ed618SSoby Mathew } 279532ed618SSoby Mathew 280532ed618SSoby Mathew /******************************************************************************* 2812bbad1d1SZelalem Aweke * The following function performs initialization of the cpu_context 'ctx' 2822bbad1d1SZelalem Aweke * for first use that is common to all security states, and sets the 2832bbad1d1SZelalem Aweke * initial entrypoint state as specified by the entry_point_info structure. 284532ed618SSoby Mathew * 2858aabea33SPaul Beesley * The EE and ST attributes are used to configure the endianness and secure 286532ed618SSoby Mathew * timer availability for the new execution context. 287532ed618SSoby Mathew ******************************************************************************/ 2882bbad1d1SZelalem Aweke static void setup_context_common(cpu_context_t *ctx, const entry_point_info_t *ep) 289532ed618SSoby Mathew { 290f1be00daSLouis Mayencourt u_register_t scr_el3; 291532ed618SSoby Mathew el3_state_t *state; 292532ed618SSoby Mathew gp_regs_t *gp_regs; 293532ed618SSoby Mathew 294532ed618SSoby Mathew /* Clear any residual register values from the context */ 29532f0d3c6SDouglas Raillard zeromem(ctx, sizeof(*ctx)); 296532ed618SSoby Mathew 297532ed618SSoby Mathew /* 29818f2efd6SDavid Cunado * SCR_EL3 was initialised during reset sequence in macro 29918f2efd6SDavid Cunado * el3_arch_init_common. This code modifies the SCR_EL3 fields that 30018f2efd6SDavid Cunado * affect the next EL. 30118f2efd6SDavid Cunado * 30218f2efd6SDavid Cunado * The following fields are initially set to zero and then updated to 30318f2efd6SDavid Cunado * the required value depending on the state of the SPSR_EL3 and the 30418f2efd6SDavid Cunado * Security state and entrypoint attributes of the next EL. 305532ed618SSoby Mathew */ 306f1be00daSLouis Mayencourt scr_el3 = read_scr(); 30746cc41d5SManish Pandey scr_el3 &= ~(SCR_NS_BIT | SCR_RW_BIT | SCR_EA_BIT | SCR_FIQ_BIT | SCR_IRQ_BIT | 3082bbad1d1SZelalem Aweke SCR_ST_BIT | SCR_HCE_BIT | SCR_NSE_BIT); 309c5ea4f8aSZelalem Aweke 31018f2efd6SDavid Cunado /* 31118f2efd6SDavid Cunado * SCR_EL3.RW: Set the execution state, AArch32 or AArch64, for next 31218f2efd6SDavid Cunado * Exception level as specified by SPSR. 31318f2efd6SDavid Cunado */ 314c5ea4f8aSZelalem Aweke if (GET_RW(ep->spsr) == MODE_RW_64) { 315532ed618SSoby Mathew scr_el3 |= SCR_RW_BIT; 316c5ea4f8aSZelalem Aweke } 3172bbad1d1SZelalem Aweke 31818f2efd6SDavid Cunado /* 31918f2efd6SDavid Cunado * SCR_EL3.ST: Traps Secure EL1 accesses to the Counter-timer Physical 32018f2efd6SDavid Cunado * Secure timer registers to EL3, from AArch64 state only, if specified 321b515f541SZelalem Aweke * by the entrypoint attributes. If SEL2 is present and enabled, the ST 322b515f541SZelalem Aweke * bit always behaves as 1 (i.e. secure physical timer register access 323b515f541SZelalem Aweke * is not trapped) 32418f2efd6SDavid Cunado */ 325c5ea4f8aSZelalem Aweke if (EP_GET_ST(ep->h.attr) != 0U) { 326532ed618SSoby Mathew scr_el3 |= SCR_ST_BIT; 327c5ea4f8aSZelalem Aweke } 328532ed618SSoby Mathew 329cb4ec47bSjohpow01 /* 330cb4ec47bSjohpow01 * If FEAT_HCX is enabled, enable access to HCRX_EL2 by setting 331cb4ec47bSjohpow01 * SCR_EL3.HXEn. 332cb4ec47bSjohpow01 */ 333c5a3ebbdSAndre Przywara if (is_feat_hcx_supported()) { 334cb4ec47bSjohpow01 scr_el3 |= SCR_HXEn_BIT; 335c5a3ebbdSAndre Przywara } 336cb4ec47bSjohpow01 337ff86e0b4SJuan Pablo Conde /* 338ff86e0b4SJuan Pablo Conde * If FEAT_RNG_TRAP is enabled, all reads of the RNDR and RNDRRS 339ff86e0b4SJuan Pablo Conde * registers are trapped to EL3. 340ff86e0b4SJuan Pablo Conde */ 341ff86e0b4SJuan Pablo Conde #if ENABLE_FEAT_RNG_TRAP 342ff86e0b4SJuan Pablo Conde scr_el3 |= SCR_TRNDR_BIT; 343ff86e0b4SJuan Pablo Conde #endif 344ff86e0b4SJuan Pablo Conde 3451a7c1cfeSJeenu Viswambharan #if FAULT_INJECTION_SUPPORT 3461a7c1cfeSJeenu Viswambharan /* Enable fault injection from lower ELs */ 3471a7c1cfeSJeenu Viswambharan scr_el3 |= SCR_FIEN_BIT; 3481a7c1cfeSJeenu Viswambharan #endif 3491a7c1cfeSJeenu Viswambharan 3505283962eSAntonio Nino Diaz /* 351d3331603SMark Brown * SCR_EL3.TCR2EN: Enable access to TCR2_ELx for AArch64 if present. 352d3331603SMark Brown */ 353d3331603SMark Brown if (is_feat_tcr2_supported() && (GET_RW(ep->spsr) == MODE_RW_64)) { 354d3331603SMark Brown scr_el3 |= SCR_TCR2EN_BIT; 355d3331603SMark Brown } 356d3331603SMark Brown 357d3331603SMark Brown /* 3582bbad1d1SZelalem Aweke * CPTR_EL3 was initialized out of reset, copy that value to the 3592bbad1d1SZelalem Aweke * context register. 3605283962eSAntonio Nino Diaz */ 36168ac5ed0SArunachalam Ganapathy write_ctx_reg(get_el3state_ctx(ctx), CTX_CPTR_EL3, read_cptr_el3()); 362532ed618SSoby Mathew 363532ed618SSoby Mathew /* 36418f2efd6SDavid Cunado * SCR_EL3.HCE: Enable HVC instructions if next execution state is 36518f2efd6SDavid Cunado * AArch64 and next EL is EL2, or if next execution state is AArch32 and 36618f2efd6SDavid Cunado * next mode is Hyp. 367110ee433SJimmy Brisson * SCR_EL3.FGTEn: Enable Fine Grained Virtualization Traps under the 368110ee433SJimmy Brisson * same conditions as HVC instructions and when the processor supports 369110ee433SJimmy Brisson * ARMv8.6-FGT. 37029d0ee54SJimmy Brisson * SCR_EL3.ECVEn: Enable Enhanced Counter Virtualization (ECV) 37129d0ee54SJimmy Brisson * CNTPOFF_EL2 register under the same conditions as HVC instructions 37229d0ee54SJimmy Brisson * and when the processor supports ECV. 373532ed618SSoby Mathew */ 374a0fee747SAntonio Nino Diaz if (((GET_RW(ep->spsr) == MODE_RW_64) && (GET_EL(ep->spsr) == MODE_EL2)) 375a0fee747SAntonio Nino Diaz || ((GET_RW(ep->spsr) != MODE_RW_64) 376a0fee747SAntonio Nino Diaz && (GET_M32(ep->spsr) == MODE32_hyp))) { 377532ed618SSoby Mathew scr_el3 |= SCR_HCE_BIT; 378110ee433SJimmy Brisson 379ce485955SAndre Przywara if (is_feat_fgt_supported()) { 380110ee433SJimmy Brisson scr_el3 |= SCR_FGTEN_BIT; 381110ee433SJimmy Brisson } 38229d0ee54SJimmy Brisson 38329d0ee54SJimmy Brisson if (get_armv8_6_ecv_support() 38429d0ee54SJimmy Brisson == ID_AA64MMFR0_EL1_ECV_SELF_SYNCH) { 38529d0ee54SJimmy Brisson scr_el3 |= SCR_ECVEN_BIT; 38629d0ee54SJimmy Brisson } 387532ed618SSoby Mathew } 388532ed618SSoby Mathew 389781d07a4SJayanth Dodderi Chidanand #if ENABLE_FEAT_TWED 3906cac724dSjohpow01 /* Enable WFE trap delay in SCR_EL3 if supported and configured */ 3916cac724dSjohpow01 /* Set delay in SCR_EL3 */ 3926cac724dSjohpow01 scr_el3 &= ~(SCR_TWEDEL_MASK << SCR_TWEDEL_SHIFT); 393781d07a4SJayanth Dodderi Chidanand scr_el3 |= ((TWED_DELAY & SCR_TWEDEL_MASK) 3946cac724dSjohpow01 << SCR_TWEDEL_SHIFT); 3956cac724dSjohpow01 3966cac724dSjohpow01 /* Enable WFE delay */ 3976cac724dSjohpow01 scr_el3 |= SCR_TWEDEn_BIT; 398781d07a4SJayanth Dodderi Chidanand #endif /* ENABLE_FEAT_TWED */ 3996cac724dSjohpow01 40018f2efd6SDavid Cunado /* 401e290a8fcSAlexei Fedorov * Populate EL3 state so that we've the right context 402e290a8fcSAlexei Fedorov * before doing ERET 4033e61b2b5SDavid Cunado */ 404532ed618SSoby Mathew state = get_el3state_ctx(ctx); 405532ed618SSoby Mathew write_ctx_reg(state, CTX_SCR_EL3, scr_el3); 406532ed618SSoby Mathew write_ctx_reg(state, CTX_ELR_EL3, ep->pc); 407532ed618SSoby Mathew write_ctx_reg(state, CTX_SPSR_EL3, ep->spsr); 408532ed618SSoby Mathew 409532ed618SSoby Mathew /* 410532ed618SSoby Mathew * Store the X0-X7 value from the entrypoint into the context 411532ed618SSoby Mathew * Use memcpy as we are in control of the layout of the structures 412532ed618SSoby Mathew */ 413532ed618SSoby Mathew gp_regs = get_gpregs_ctx(ctx); 414532ed618SSoby Mathew memcpy(gp_regs, (void *)&ep->args, sizeof(aapcs64_params_t)); 415532ed618SSoby Mathew } 416532ed618SSoby Mathew 417532ed618SSoby Mathew /******************************************************************************* 4182bbad1d1SZelalem Aweke * Context management library initialization routine. This library is used by 4192bbad1d1SZelalem Aweke * runtime services to share pointers to 'cpu_context' structures for secure 4202bbad1d1SZelalem Aweke * non-secure and realm states. Management of the structures and their associated 4212bbad1d1SZelalem Aweke * memory is not done by the context management library e.g. the PSCI service 4222bbad1d1SZelalem Aweke * manages the cpu context used for entry from and exit to the non-secure state. 4232bbad1d1SZelalem Aweke * The Secure payload dispatcher service manages the context(s) corresponding to 4242bbad1d1SZelalem Aweke * the secure state. It also uses this library to get access to the non-secure 4252bbad1d1SZelalem Aweke * state cpu context pointers. 4262bbad1d1SZelalem Aweke * Lastly, this library provides the API to make SP_EL3 point to the cpu context 4272bbad1d1SZelalem Aweke * which will be used for programming an entry into a lower EL. The same context 4282bbad1d1SZelalem Aweke * will be used to save state upon exception entry from that EL. 4292bbad1d1SZelalem Aweke ******************************************************************************/ 4302bbad1d1SZelalem Aweke void __init cm_init(void) 4312bbad1d1SZelalem Aweke { 4322bbad1d1SZelalem Aweke /* 4332bbad1d1SZelalem Aweke * The context management library has only global data to intialize, but 4342bbad1d1SZelalem Aweke * that will be done when the BSS is zeroed out. 4352bbad1d1SZelalem Aweke */ 4362bbad1d1SZelalem Aweke } 4372bbad1d1SZelalem Aweke 4382bbad1d1SZelalem Aweke /******************************************************************************* 4392bbad1d1SZelalem Aweke * This is the high-level function used to initialize the cpu_context 'ctx' for 4402bbad1d1SZelalem Aweke * first use. It performs initializations that are common to all security states 4412bbad1d1SZelalem Aweke * and initializations specific to the security state specified in 'ep' 4422bbad1d1SZelalem Aweke ******************************************************************************/ 4432bbad1d1SZelalem Aweke void cm_setup_context(cpu_context_t *ctx, const entry_point_info_t *ep) 4442bbad1d1SZelalem Aweke { 4452bbad1d1SZelalem Aweke unsigned int security_state; 4462bbad1d1SZelalem Aweke 4472bbad1d1SZelalem Aweke assert(ctx != NULL); 4482bbad1d1SZelalem Aweke 4492bbad1d1SZelalem Aweke /* 4502bbad1d1SZelalem Aweke * Perform initializations that are common 4512bbad1d1SZelalem Aweke * to all security states 4522bbad1d1SZelalem Aweke */ 4532bbad1d1SZelalem Aweke setup_context_common(ctx, ep); 4542bbad1d1SZelalem Aweke 4552bbad1d1SZelalem Aweke security_state = GET_SECURITY_STATE(ep->h.attr); 4562bbad1d1SZelalem Aweke 4572bbad1d1SZelalem Aweke /* Perform security state specific initializations */ 4582bbad1d1SZelalem Aweke switch (security_state) { 4592bbad1d1SZelalem Aweke case SECURE: 4602bbad1d1SZelalem Aweke setup_secure_context(ctx, ep); 4612bbad1d1SZelalem Aweke break; 4622bbad1d1SZelalem Aweke #if ENABLE_RME 4632bbad1d1SZelalem Aweke case REALM: 4642bbad1d1SZelalem Aweke setup_realm_context(ctx, ep); 4652bbad1d1SZelalem Aweke break; 4662bbad1d1SZelalem Aweke #endif 4672bbad1d1SZelalem Aweke case NON_SECURE: 4682bbad1d1SZelalem Aweke setup_ns_context(ctx, ep); 4692bbad1d1SZelalem Aweke break; 4702bbad1d1SZelalem Aweke default: 4712bbad1d1SZelalem Aweke ERROR("Invalid security state\n"); 4722bbad1d1SZelalem Aweke panic(); 4732bbad1d1SZelalem Aweke break; 4742bbad1d1SZelalem Aweke } 4752bbad1d1SZelalem Aweke } 4762bbad1d1SZelalem Aweke 4772bbad1d1SZelalem Aweke /******************************************************************************* 4780fd0f222SDimitris Papastamos * Enable architecture extensions on first entry to Non-secure world. 4790fd0f222SDimitris Papastamos * When EL2 is implemented but unused `el2_unused` is non-zero, otherwise 4800fd0f222SDimitris Papastamos * it is zero. 4810fd0f222SDimitris Papastamos ******************************************************************************/ 482dc78e62dSjohpow01 static void manage_extensions_nonsecure(bool el2_unused, cpu_context_t *ctx) 4830fd0f222SDimitris Papastamos { 4840fd0f222SDimitris Papastamos #if IMAGE_BL31 4856437a09aSAndre Przywara if (is_feat_spe_supported()) { 486281a08ccSDimitris Papastamos spe_enable(el2_unused); 4876437a09aSAndre Przywara } 488380559c1SDimitris Papastamos 489380559c1SDimitris Papastamos #if ENABLE_AMU 49068ac5ed0SArunachalam Ganapathy amu_enable(el2_unused, ctx); 49168ac5ed0SArunachalam Ganapathy #endif 49268ac5ed0SArunachalam Ganapathy 493dc78e62dSjohpow01 #if ENABLE_SME_FOR_NS 494dc78e62dSjohpow01 /* Enable SME, SVE, and FPU/SIMD for non-secure world. */ 495dc78e62dSjohpow01 sme_enable(ctx); 496dc78e62dSjohpow01 #elif ENABLE_SVE_FOR_NS 497dc78e62dSjohpow01 /* Enable SVE and FPU/SIMD for non-secure world. */ 49868ac5ed0SArunachalam Ganapathy sve_enable(ctx); 499380559c1SDimitris Papastamos #endif 5001a853370SDavid Cunado 501*9448f2b8SAndre Przywara if (is_feat_mpam_supported()) { 5025f835918SJeenu Viswambharan mpam_enable(el2_unused); 503*9448f2b8SAndre Przywara } 504813524eaSManish V Badarkhe 505f5360cfaSAndre Przywara if (is_feat_trbe_supported()) { 506813524eaSManish V Badarkhe trbe_enable(); 507f5360cfaSAndre Przywara } 508813524eaSManish V Badarkhe 509ff491036SAndre Przywara if (is_feat_brbe_supported()) { 510744ad974Sjohpow01 brbe_enable(); 511ff491036SAndre Przywara } 512744ad974Sjohpow01 513d4582d30SManish V Badarkhe #if ENABLE_SYS_REG_TRACE_FOR_NS 514d4582d30SManish V Badarkhe sys_reg_trace_enable(ctx); 515d4582d30SManish V Badarkhe #endif /* ENABLE_SYS_REG_TRACE_FOR_NS */ 516d4582d30SManish V Badarkhe 517fc8d2d39SAndre Przywara if (is_feat_trf_supported()) { 5188fcd3d96SManish V Badarkhe trf_enable(); 519fc8d2d39SAndre Przywara } 5200fd0f222SDimitris Papastamos #endif 5210fd0f222SDimitris Papastamos } 5220fd0f222SDimitris Papastamos 5230fd0f222SDimitris Papastamos /******************************************************************************* 52468ac5ed0SArunachalam Ganapathy * Enable architecture extensions on first entry to Secure world. 52568ac5ed0SArunachalam Ganapathy ******************************************************************************/ 526dc78e62dSjohpow01 static void manage_extensions_secure(cpu_context_t *ctx) 52768ac5ed0SArunachalam Ganapathy { 52868ac5ed0SArunachalam Ganapathy #if IMAGE_BL31 529dc78e62dSjohpow01 #if ENABLE_SME_FOR_NS 530dc78e62dSjohpow01 #if ENABLE_SME_FOR_SWD 531dc78e62dSjohpow01 /* 532dc78e62dSjohpow01 * Enable SME, SVE, FPU/SIMD in secure context, secure manager must 533dc78e62dSjohpow01 * ensure SME, SVE, and FPU/SIMD context properly managed. 534dc78e62dSjohpow01 */ 535dc78e62dSjohpow01 sme_enable(ctx); 536dc78e62dSjohpow01 #else /* ENABLE_SME_FOR_SWD */ 537dc78e62dSjohpow01 /* 538dc78e62dSjohpow01 * Disable SME, SVE, FPU/SIMD in secure context so non-secure world can 539dc78e62dSjohpow01 * safely use the associated registers. 540dc78e62dSjohpow01 */ 541dc78e62dSjohpow01 sme_disable(ctx); 542dc78e62dSjohpow01 #endif /* ENABLE_SME_FOR_SWD */ 543dc78e62dSjohpow01 #elif ENABLE_SVE_FOR_NS 54468ac5ed0SArunachalam Ganapathy #if ENABLE_SVE_FOR_SWD 545dc78e62dSjohpow01 /* 546dc78e62dSjohpow01 * Enable SVE and FPU in secure context, secure manager must ensure that 547dc78e62dSjohpow01 * the SVE and FPU register contexts are properly managed. 548dc78e62dSjohpow01 */ 54968ac5ed0SArunachalam Ganapathy sve_enable(ctx); 550dc78e62dSjohpow01 #else /* ENABLE_SVE_FOR_SWD */ 551dc78e62dSjohpow01 /* 552dc78e62dSjohpow01 * Disable SVE and FPU in secure context so non-secure world can safely 553dc78e62dSjohpow01 * use them. 554dc78e62dSjohpow01 */ 555dc78e62dSjohpow01 sve_disable(ctx); 556dc78e62dSjohpow01 #endif /* ENABLE_SVE_FOR_SWD */ 557dc78e62dSjohpow01 #endif /* ENABLE_SVE_FOR_NS */ 558dc78e62dSjohpow01 #endif /* IMAGE_BL31 */ 55968ac5ed0SArunachalam Ganapathy } 56068ac5ed0SArunachalam Ganapathy 56168ac5ed0SArunachalam Ganapathy /******************************************************************************* 562532ed618SSoby Mathew * The following function initializes the cpu_context for a CPU specified by 563532ed618SSoby Mathew * its `cpu_idx` for first use, and sets the initial entrypoint state as 564532ed618SSoby Mathew * specified by the entry_point_info structure. 565532ed618SSoby Mathew ******************************************************************************/ 566532ed618SSoby Mathew void cm_init_context_by_index(unsigned int cpu_idx, 567532ed618SSoby Mathew const entry_point_info_t *ep) 568532ed618SSoby Mathew { 569532ed618SSoby Mathew cpu_context_t *ctx; 570532ed618SSoby Mathew ctx = cm_get_context_by_index(cpu_idx, GET_SECURITY_STATE(ep->h.attr)); 5711634cae8SAntonio Nino Diaz cm_setup_context(ctx, ep); 572532ed618SSoby Mathew } 573532ed618SSoby Mathew 574532ed618SSoby Mathew /******************************************************************************* 575532ed618SSoby Mathew * The following function initializes the cpu_context for the current CPU 576532ed618SSoby Mathew * for first use, and sets the initial entrypoint state as specified by the 577532ed618SSoby Mathew * entry_point_info structure. 578532ed618SSoby Mathew ******************************************************************************/ 579532ed618SSoby Mathew void cm_init_my_context(const entry_point_info_t *ep) 580532ed618SSoby Mathew { 581532ed618SSoby Mathew cpu_context_t *ctx; 582532ed618SSoby Mathew ctx = cm_get_context(GET_SECURITY_STATE(ep->h.attr)); 5831634cae8SAntonio Nino Diaz cm_setup_context(ctx, ep); 584532ed618SSoby Mathew } 585532ed618SSoby Mathew 586532ed618SSoby Mathew /******************************************************************************* 587c5ea4f8aSZelalem Aweke * Prepare the CPU system registers for first entry into realm, secure, or 588c5ea4f8aSZelalem Aweke * normal world. 589532ed618SSoby Mathew * 590532ed618SSoby Mathew * If execution is requested to EL2 or hyp mode, SCTLR_EL2 is initialized 591532ed618SSoby Mathew * If execution is requested to non-secure EL1 or svc mode, and the CPU supports 592532ed618SSoby Mathew * EL2 then EL2 is disabled by configuring all necessary EL2 registers. 593532ed618SSoby Mathew * For all entries, the EL1 registers are initialized from the cpu_context 594532ed618SSoby Mathew ******************************************************************************/ 595532ed618SSoby Mathew void cm_prepare_el3_exit(uint32_t security_state) 596532ed618SSoby Mathew { 597f1be00daSLouis Mayencourt u_register_t sctlr_elx, scr_el3, mdcr_el2; 598532ed618SSoby Mathew cpu_context_t *ctx = cm_get_context(security_state); 59940daecc1SAntonio Nino Diaz bool el2_unused = false; 600a0fee747SAntonio Nino Diaz uint64_t hcr_el2 = 0U; 601532ed618SSoby Mathew 602a0fee747SAntonio Nino Diaz assert(ctx != NULL); 603532ed618SSoby Mathew 604532ed618SSoby Mathew if (security_state == NON_SECURE) { 605f1be00daSLouis Mayencourt scr_el3 = read_ctx_reg(get_el3state_ctx(ctx), 606a0fee747SAntonio Nino Diaz CTX_SCR_EL3); 607a0fee747SAntonio Nino Diaz if ((scr_el3 & SCR_HCE_BIT) != 0U) { 608532ed618SSoby Mathew /* Use SCTLR_EL1.EE value to initialise sctlr_el2 */ 6092825946eSMax Shvetsov sctlr_elx = read_ctx_reg(get_el1_sysregs_ctx(ctx), 610532ed618SSoby Mathew CTX_SCTLR_EL1); 6112e09d4f8SKen Kuang sctlr_elx &= SCTLR_EE_BIT; 612532ed618SSoby Mathew sctlr_elx |= SCTLR_EL2_RES1; 6135f5d1ed7SLouis Mayencourt #if ERRATA_A75_764081 6145f5d1ed7SLouis Mayencourt /* 6155f5d1ed7SLouis Mayencourt * If workaround of errata 764081 for Cortex-A75 is used 6165f5d1ed7SLouis Mayencourt * then set SCTLR_EL2.IESB to enable Implicit Error 6175f5d1ed7SLouis Mayencourt * Synchronization Barrier. 6185f5d1ed7SLouis Mayencourt */ 6195f5d1ed7SLouis Mayencourt sctlr_elx |= SCTLR_IESB_BIT; 6205f5d1ed7SLouis Mayencourt #endif 621532ed618SSoby Mathew write_sctlr_el2(sctlr_elx); 622a0fee747SAntonio Nino Diaz } else if (el_implemented(2) != EL_IMPL_NONE) { 62340daecc1SAntonio Nino Diaz el2_unused = true; 6240fd0f222SDimitris Papastamos 62518f2efd6SDavid Cunado /* 62618f2efd6SDavid Cunado * EL2 present but unused, need to disable safely. 62718f2efd6SDavid Cunado * SCTLR_EL2 can be ignored in this case. 62818f2efd6SDavid Cunado * 6293ff4aaacSJeenu Viswambharan * Set EL2 register width appropriately: Set HCR_EL2 6303ff4aaacSJeenu Viswambharan * field to match SCR_EL3.RW. 63118f2efd6SDavid Cunado */ 632a0fee747SAntonio Nino Diaz if ((scr_el3 & SCR_RW_BIT) != 0U) 6333ff4aaacSJeenu Viswambharan hcr_el2 |= HCR_RW_BIT; 6343ff4aaacSJeenu Viswambharan 6353ff4aaacSJeenu Viswambharan /* 6363ff4aaacSJeenu Viswambharan * For Armv8.3 pointer authentication feature, disable 6373ff4aaacSJeenu Viswambharan * traps to EL2 when accessing key registers or using 6383ff4aaacSJeenu Viswambharan * pointer authentication instructions from lower ELs. 6393ff4aaacSJeenu Viswambharan */ 6403ff4aaacSJeenu Viswambharan hcr_el2 |= (HCR_API_BIT | HCR_APK_BIT); 6413ff4aaacSJeenu Viswambharan 6423ff4aaacSJeenu Viswambharan write_hcr_el2(hcr_el2); 643532ed618SSoby Mathew 64418f2efd6SDavid Cunado /* 64518f2efd6SDavid Cunado * Initialise CPTR_EL2 setting all fields rather than 64618f2efd6SDavid Cunado * relying on the hw. All fields have architecturally 64718f2efd6SDavid Cunado * UNKNOWN reset values. 64818f2efd6SDavid Cunado * 64918f2efd6SDavid Cunado * CPTR_EL2.TCPAC: Set to zero so that Non-secure EL1 65018f2efd6SDavid Cunado * accesses to the CPACR_EL1 or CPACR from both 65118f2efd6SDavid Cunado * Execution states do not trap to EL2. 65218f2efd6SDavid Cunado * 65318f2efd6SDavid Cunado * CPTR_EL2.TTA: Set to zero so that Non-secure System 65418f2efd6SDavid Cunado * register accesses to the trace registers from both 65518f2efd6SDavid Cunado * Execution states do not trap to EL2. 656d4582d30SManish V Badarkhe * If PE trace unit System registers are not implemented 657d4582d30SManish V Badarkhe * then this bit is reserved, and must be set to zero. 65818f2efd6SDavid Cunado * 65918f2efd6SDavid Cunado * CPTR_EL2.TFP: Set to zero so that Non-secure accesses 66018f2efd6SDavid Cunado * to SIMD and floating-point functionality from both 66118f2efd6SDavid Cunado * Execution states do not trap to EL2. 66218f2efd6SDavid Cunado */ 66318f2efd6SDavid Cunado write_cptr_el2(CPTR_EL2_RESET_VAL & 66418f2efd6SDavid Cunado ~(CPTR_EL2_TCPAC_BIT | CPTR_EL2_TTA_BIT 66518f2efd6SDavid Cunado | CPTR_EL2_TFP_BIT)); 666532ed618SSoby Mathew 66718f2efd6SDavid Cunado /* 6688aabea33SPaul Beesley * Initialise CNTHCTL_EL2. All fields are 66918f2efd6SDavid Cunado * architecturally UNKNOWN on reset and are set to zero 67018f2efd6SDavid Cunado * except for field(s) listed below. 67118f2efd6SDavid Cunado * 672c5ea4f8aSZelalem Aweke * CNTHCTL_EL2.EL1PTEN: Set to one to disable traps to 67318f2efd6SDavid Cunado * Hyp mode of Non-secure EL0 and EL1 accesses to the 67418f2efd6SDavid Cunado * physical timer registers. 67518f2efd6SDavid Cunado * 67618f2efd6SDavid Cunado * CNTHCTL_EL2.EL1PCTEN: Set to one to disable traps to 67718f2efd6SDavid Cunado * Hyp mode of Non-secure EL0 and EL1 accesses to the 67818f2efd6SDavid Cunado * physical counter registers. 67918f2efd6SDavid Cunado */ 68018f2efd6SDavid Cunado write_cnthctl_el2(CNTHCTL_RESET_VAL | 68118f2efd6SDavid Cunado EL1PCEN_BIT | EL1PCTEN_BIT); 682532ed618SSoby Mathew 68318f2efd6SDavid Cunado /* 68418f2efd6SDavid Cunado * Initialise CNTVOFF_EL2 to zero as it resets to an 68518f2efd6SDavid Cunado * architecturally UNKNOWN value. 68618f2efd6SDavid Cunado */ 687532ed618SSoby Mathew write_cntvoff_el2(0); 688532ed618SSoby Mathew 68918f2efd6SDavid Cunado /* 69018f2efd6SDavid Cunado * Set VPIDR_EL2 and VMPIDR_EL2 to match MIDR_EL1 and 69118f2efd6SDavid Cunado * MPIDR_EL1 respectively. 69218f2efd6SDavid Cunado */ 693532ed618SSoby Mathew write_vpidr_el2(read_midr_el1()); 694532ed618SSoby Mathew write_vmpidr_el2(read_mpidr_el1()); 695532ed618SSoby Mathew 696532ed618SSoby Mathew /* 69718f2efd6SDavid Cunado * Initialise VTTBR_EL2. All fields are architecturally 69818f2efd6SDavid Cunado * UNKNOWN on reset. 69918f2efd6SDavid Cunado * 70018f2efd6SDavid Cunado * VTTBR_EL2.VMID: Set to zero. Even though EL1&0 stage 70118f2efd6SDavid Cunado * 2 address translation is disabled, cache maintenance 70218f2efd6SDavid Cunado * operations depend on the VMID. 70318f2efd6SDavid Cunado * 70418f2efd6SDavid Cunado * VTTBR_EL2.BADDR: Set to zero as EL1&0 stage 2 address 70518f2efd6SDavid Cunado * translation is disabled. 706532ed618SSoby Mathew */ 70718f2efd6SDavid Cunado write_vttbr_el2(VTTBR_RESET_VAL & 70818f2efd6SDavid Cunado ~((VTTBR_VMID_MASK << VTTBR_VMID_SHIFT) 70918f2efd6SDavid Cunado | (VTTBR_BADDR_MASK << VTTBR_BADDR_SHIFT))); 71018f2efd6SDavid Cunado 711495f3d3cSDavid Cunado /* 71218f2efd6SDavid Cunado * Initialise MDCR_EL2, setting all fields rather than 71318f2efd6SDavid Cunado * relying on hw. Some fields are architecturally 71418f2efd6SDavid Cunado * UNKNOWN on reset. 71518f2efd6SDavid Cunado * 716e290a8fcSAlexei Fedorov * MDCR_EL2.HLP: Set to one so that event counter 717e290a8fcSAlexei Fedorov * overflow, that is recorded in PMOVSCLR_EL0[0-30], 718e290a8fcSAlexei Fedorov * occurs on the increment that changes 719e290a8fcSAlexei Fedorov * PMEVCNTR<n>_EL0[63] from 1 to 0, when ARMv8.5-PMU is 720e290a8fcSAlexei Fedorov * implemented. This bit is RES0 in versions of the 721e290a8fcSAlexei Fedorov * architecture earlier than ARMv8.5, setting it to 1 722e290a8fcSAlexei Fedorov * doesn't have any effect on them. 723e290a8fcSAlexei Fedorov * 724e290a8fcSAlexei Fedorov * MDCR_EL2.TTRF: Set to zero so that access to Trace 725e290a8fcSAlexei Fedorov * Filter Control register TRFCR_EL1 at EL1 is not 726e290a8fcSAlexei Fedorov * trapped to EL2. This bit is RES0 in versions of 727e290a8fcSAlexei Fedorov * the architecture earlier than ARMv8.4. 728e290a8fcSAlexei Fedorov * 729e290a8fcSAlexei Fedorov * MDCR_EL2.HPMD: Set to one so that event counting is 730e290a8fcSAlexei Fedorov * prohibited at EL2. This bit is RES0 in versions of 731e290a8fcSAlexei Fedorov * the architecture earlier than ARMv8.1, setting it 732e290a8fcSAlexei Fedorov * to 1 doesn't have any effect on them. 733e290a8fcSAlexei Fedorov * 734e290a8fcSAlexei Fedorov * MDCR_EL2.TPMS: Set to zero so that accesses to 735e290a8fcSAlexei Fedorov * Statistical Profiling control registers from EL1 736e290a8fcSAlexei Fedorov * do not trap to EL2. This bit is RES0 when SPE is 737e290a8fcSAlexei Fedorov * not implemented. 738e290a8fcSAlexei Fedorov * 73918f2efd6SDavid Cunado * MDCR_EL2.TDRA: Set to zero so that Non-secure EL0 and 74018f2efd6SDavid Cunado * EL1 System register accesses to the Debug ROM 74118f2efd6SDavid Cunado * registers are not trapped to EL2. 74218f2efd6SDavid Cunado * 74318f2efd6SDavid Cunado * MDCR_EL2.TDOSA: Set to zero so that Non-secure EL1 74418f2efd6SDavid Cunado * System register accesses to the powerdown debug 74518f2efd6SDavid Cunado * registers are not trapped to EL2. 74618f2efd6SDavid Cunado * 74718f2efd6SDavid Cunado * MDCR_EL2.TDA: Set to zero so that System register 74818f2efd6SDavid Cunado * accesses to the debug registers do not trap to EL2. 74918f2efd6SDavid Cunado * 75018f2efd6SDavid Cunado * MDCR_EL2.TDE: Set to zero so that debug exceptions 75118f2efd6SDavid Cunado * are not routed to EL2. 75218f2efd6SDavid Cunado * 75318f2efd6SDavid Cunado * MDCR_EL2.HPME: Set to zero to disable EL2 Performance 75418f2efd6SDavid Cunado * Monitors. 75518f2efd6SDavid Cunado * 75618f2efd6SDavid Cunado * MDCR_EL2.TPM: Set to zero so that Non-secure EL0 and 75718f2efd6SDavid Cunado * EL1 accesses to all Performance Monitors registers 75818f2efd6SDavid Cunado * are not trapped to EL2. 75918f2efd6SDavid Cunado * 76018f2efd6SDavid Cunado * MDCR_EL2.TPMCR: Set to zero so that Non-secure EL0 76118f2efd6SDavid Cunado * and EL1 accesses to the PMCR_EL0 or PMCR are not 76218f2efd6SDavid Cunado * trapped to EL2. 76318f2efd6SDavid Cunado * 76418f2efd6SDavid Cunado * MDCR_EL2.HPMN: Set to value of PMCR_EL0.N which is the 76518f2efd6SDavid Cunado * architecturally-defined reset value. 76640ff9074SManish V Badarkhe * 76740ff9074SManish V Badarkhe * MDCR_EL2.E2TB: Set to zero so that the trace Buffer 76840ff9074SManish V Badarkhe * owning exception level is NS-EL1 and, tracing is 76940ff9074SManish V Badarkhe * prohibited at NS-EL2. These bits are RES0 when 77040ff9074SManish V Badarkhe * FEAT_TRBE is not implemented. 771495f3d3cSDavid Cunado */ 772e290a8fcSAlexei Fedorov mdcr_el2 = ((MDCR_EL2_RESET_VAL | MDCR_EL2_HLP | 773e290a8fcSAlexei Fedorov MDCR_EL2_HPMD) | 77418f2efd6SDavid Cunado ((read_pmcr_el0() & PMCR_EL0_N_BITS) 77518f2efd6SDavid Cunado >> PMCR_EL0_N_SHIFT)) & 776e290a8fcSAlexei Fedorov ~(MDCR_EL2_TTRF | MDCR_EL2_TPMS | 777e290a8fcSAlexei Fedorov MDCR_EL2_TDRA_BIT | MDCR_EL2_TDOSA_BIT | 778e290a8fcSAlexei Fedorov MDCR_EL2_TDA_BIT | MDCR_EL2_TDE_BIT | 779e290a8fcSAlexei Fedorov MDCR_EL2_HPME_BIT | MDCR_EL2_TPM_BIT | 78040ff9074SManish V Badarkhe MDCR_EL2_TPMCR_BIT | 78140ff9074SManish V Badarkhe MDCR_EL2_E2TB(MDCR_EL2_E2TB_EL1)); 782d832aee9Sdp-arm 783d832aee9Sdp-arm write_mdcr_el2(mdcr_el2); 784d832aee9Sdp-arm 785939f66d6SDavid Cunado /* 78618f2efd6SDavid Cunado * Initialise HSTR_EL2. All fields are architecturally 78718f2efd6SDavid Cunado * UNKNOWN on reset. 78818f2efd6SDavid Cunado * 78918f2efd6SDavid Cunado * HSTR_EL2.T<n>: Set all these fields to zero so that 79018f2efd6SDavid Cunado * Non-secure EL0 or EL1 accesses to System registers 79118f2efd6SDavid Cunado * do not trap to EL2. 792939f66d6SDavid Cunado */ 79318f2efd6SDavid Cunado write_hstr_el2(HSTR_EL2_RESET_VAL & ~(HSTR_EL2_T_MASK)); 794939f66d6SDavid Cunado /* 79518f2efd6SDavid Cunado * Initialise CNTHP_CTL_EL2. All fields are 79618f2efd6SDavid Cunado * architecturally UNKNOWN on reset. 79718f2efd6SDavid Cunado * 79818f2efd6SDavid Cunado * CNTHP_CTL_EL2:ENABLE: Set to zero to disable the EL2 79918f2efd6SDavid Cunado * physical timer and prevent timer interrupts. 800939f66d6SDavid Cunado */ 80118f2efd6SDavid Cunado write_cnthp_ctl_el2(CNTHP_CTL_RESET_VAL & 80218f2efd6SDavid Cunado ~(CNTHP_CTL_ENABLE_BIT)); 803532ed618SSoby Mathew } 804dc78e62dSjohpow01 manage_extensions_nonsecure(el2_unused, ctx); 805532ed618SSoby Mathew } 806532ed618SSoby Mathew 80717b4c0ddSDimitris Papastamos cm_el1_sysregs_context_restore(security_state); 80817b4c0ddSDimitris Papastamos cm_set_next_eret_context(security_state); 809532ed618SSoby Mathew } 810532ed618SSoby Mathew 81128f39f02SMax Shvetsov #if CTX_INCLUDE_EL2_REGS 812bb7b85a3SAndre Przywara 813bb7b85a3SAndre Przywara static void el2_sysregs_context_save_fgt(el2_sysregs_t *ctx) 814bb7b85a3SAndre Przywara { 815bb7b85a3SAndre Przywara write_ctx_reg(ctx, CTX_HDFGRTR_EL2, read_hdfgrtr_el2()); 816bb7b85a3SAndre Przywara if (is_feat_amu_supported()) { 817bb7b85a3SAndre Przywara write_ctx_reg(ctx, CTX_HAFGRTR_EL2, read_hafgrtr_el2()); 818bb7b85a3SAndre Przywara } 819bb7b85a3SAndre Przywara write_ctx_reg(ctx, CTX_HDFGWTR_EL2, read_hdfgwtr_el2()); 820bb7b85a3SAndre Przywara write_ctx_reg(ctx, CTX_HFGITR_EL2, read_hfgitr_el2()); 821bb7b85a3SAndre Przywara write_ctx_reg(ctx, CTX_HFGRTR_EL2, read_hfgrtr_el2()); 822bb7b85a3SAndre Przywara write_ctx_reg(ctx, CTX_HFGWTR_EL2, read_hfgwtr_el2()); 823bb7b85a3SAndre Przywara } 824bb7b85a3SAndre Przywara 825bb7b85a3SAndre Przywara static void el2_sysregs_context_restore_fgt(el2_sysregs_t *ctx) 826bb7b85a3SAndre Przywara { 827bb7b85a3SAndre Przywara write_hdfgrtr_el2(read_ctx_reg(ctx, CTX_HDFGRTR_EL2)); 828bb7b85a3SAndre Przywara if (is_feat_amu_supported()) { 829bb7b85a3SAndre Przywara write_hafgrtr_el2(read_ctx_reg(ctx, CTX_HAFGRTR_EL2)); 830bb7b85a3SAndre Przywara } 831bb7b85a3SAndre Przywara write_hdfgwtr_el2(read_ctx_reg(ctx, CTX_HDFGWTR_EL2)); 832bb7b85a3SAndre Przywara write_hfgitr_el2(read_ctx_reg(ctx, CTX_HFGITR_EL2)); 833bb7b85a3SAndre Przywara write_hfgrtr_el2(read_ctx_reg(ctx, CTX_HFGRTR_EL2)); 834bb7b85a3SAndre Przywara write_hfgwtr_el2(read_ctx_reg(ctx, CTX_HFGWTR_EL2)); 835bb7b85a3SAndre Przywara } 836bb7b85a3SAndre Przywara 837*9448f2b8SAndre Przywara static void el2_sysregs_context_save_mpam(el2_sysregs_t *ctx) 838*9448f2b8SAndre Przywara { 839*9448f2b8SAndre Przywara u_register_t mpam_idr = read_mpamidr_el1(); 840*9448f2b8SAndre Przywara 841*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAM2_EL2, read_mpam2_el2()); 842*9448f2b8SAndre Przywara 843*9448f2b8SAndre Przywara /* 844*9448f2b8SAndre Przywara * The context registers that we intend to save would be part of the 845*9448f2b8SAndre Przywara * PE's system register frame only if MPAMIDR_EL1.HAS_HCR == 1. 846*9448f2b8SAndre Przywara */ 847*9448f2b8SAndre Przywara if ((mpam_idr & MPAMIDR_HAS_HCR_BIT) == 0U) { 848*9448f2b8SAndre Przywara return; 849*9448f2b8SAndre Przywara } 850*9448f2b8SAndre Przywara 851*9448f2b8SAndre Przywara /* 852*9448f2b8SAndre Przywara * MPAMHCR_EL2, MPAMVPMV_EL2 and MPAMVPM0_EL2 are always present if 853*9448f2b8SAndre Przywara * MPAMIDR_HAS_HCR_BIT == 1. 854*9448f2b8SAndre Przywara */ 855*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMHCR_EL2, read_mpamhcr_el2()); 856*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM0_EL2, read_mpamvpm0_el2()); 857*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPMV_EL2, read_mpamvpmv_el2()); 858*9448f2b8SAndre Przywara 859*9448f2b8SAndre Przywara /* 860*9448f2b8SAndre Przywara * The number of MPAMVPM registers is implementation defined, their 861*9448f2b8SAndre Przywara * number is stored in the MPAMIDR_EL1 register. 862*9448f2b8SAndre Przywara */ 863*9448f2b8SAndre Przywara switch ((mpam_idr >> MPAMIDR_EL1_VPMR_MAX_SHIFT) & MPAMIDR_EL1_VPMR_MAX_MASK) { 864*9448f2b8SAndre Przywara case 7: 865*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM7_EL2, read_mpamvpm7_el2()); 866*9448f2b8SAndre Przywara __fallthrough; 867*9448f2b8SAndre Przywara case 6: 868*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM6_EL2, read_mpamvpm6_el2()); 869*9448f2b8SAndre Przywara __fallthrough; 870*9448f2b8SAndre Przywara case 5: 871*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM5_EL2, read_mpamvpm5_el2()); 872*9448f2b8SAndre Przywara __fallthrough; 873*9448f2b8SAndre Przywara case 4: 874*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM4_EL2, read_mpamvpm4_el2()); 875*9448f2b8SAndre Przywara __fallthrough; 876*9448f2b8SAndre Przywara case 3: 877*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM3_EL2, read_mpamvpm3_el2()); 878*9448f2b8SAndre Przywara __fallthrough; 879*9448f2b8SAndre Przywara case 2: 880*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM2_EL2, read_mpamvpm2_el2()); 881*9448f2b8SAndre Przywara __fallthrough; 882*9448f2b8SAndre Przywara case 1: 883*9448f2b8SAndre Przywara write_ctx_reg(ctx, CTX_MPAMVPM1_EL2, read_mpamvpm1_el2()); 884*9448f2b8SAndre Przywara break; 885*9448f2b8SAndre Przywara } 886*9448f2b8SAndre Przywara } 887*9448f2b8SAndre Przywara 888*9448f2b8SAndre Przywara static void el2_sysregs_context_restore_mpam(el2_sysregs_t *ctx) 889*9448f2b8SAndre Przywara { 890*9448f2b8SAndre Przywara u_register_t mpam_idr = read_mpamidr_el1(); 891*9448f2b8SAndre Przywara 892*9448f2b8SAndre Przywara write_mpam2_el2(read_ctx_reg(ctx, CTX_MPAM2_EL2)); 893*9448f2b8SAndre Przywara 894*9448f2b8SAndre Przywara if ((mpam_idr & MPAMIDR_HAS_HCR_BIT) == 0U) { 895*9448f2b8SAndre Przywara return; 896*9448f2b8SAndre Przywara } 897*9448f2b8SAndre Przywara 898*9448f2b8SAndre Przywara write_mpamhcr_el2(read_ctx_reg(ctx, CTX_MPAMHCR_EL2)); 899*9448f2b8SAndre Przywara write_mpamvpm0_el2(read_ctx_reg(ctx, CTX_MPAMVPM0_EL2)); 900*9448f2b8SAndre Przywara write_mpamvpmv_el2(read_ctx_reg(ctx, CTX_MPAMVPMV_EL2)); 901*9448f2b8SAndre Przywara 902*9448f2b8SAndre Przywara switch ((mpam_idr >> MPAMIDR_EL1_VPMR_MAX_SHIFT) & MPAMIDR_EL1_VPMR_MAX_MASK) { 903*9448f2b8SAndre Przywara case 7: 904*9448f2b8SAndre Przywara write_mpamvpm7_el2(read_ctx_reg(ctx, CTX_MPAMVPM7_EL2)); 905*9448f2b8SAndre Przywara __fallthrough; 906*9448f2b8SAndre Przywara case 6: 907*9448f2b8SAndre Przywara write_mpamvpm6_el2(read_ctx_reg(ctx, CTX_MPAMVPM6_EL2)); 908*9448f2b8SAndre Przywara __fallthrough; 909*9448f2b8SAndre Przywara case 5: 910*9448f2b8SAndre Przywara write_mpamvpm5_el2(read_ctx_reg(ctx, CTX_MPAMVPM5_EL2)); 911*9448f2b8SAndre Przywara __fallthrough; 912*9448f2b8SAndre Przywara case 4: 913*9448f2b8SAndre Przywara write_mpamvpm4_el2(read_ctx_reg(ctx, CTX_MPAMVPM4_EL2)); 914*9448f2b8SAndre Przywara __fallthrough; 915*9448f2b8SAndre Przywara case 3: 916*9448f2b8SAndre Przywara write_mpamvpm3_el2(read_ctx_reg(ctx, CTX_MPAMVPM3_EL2)); 917*9448f2b8SAndre Przywara __fallthrough; 918*9448f2b8SAndre Przywara case 2: 919*9448f2b8SAndre Przywara write_mpamvpm2_el2(read_ctx_reg(ctx, CTX_MPAMVPM2_EL2)); 920*9448f2b8SAndre Przywara __fallthrough; 921*9448f2b8SAndre Przywara case 1: 922*9448f2b8SAndre Przywara write_mpamvpm1_el2(read_ctx_reg(ctx, CTX_MPAMVPM1_EL2)); 923*9448f2b8SAndre Przywara break; 924*9448f2b8SAndre Przywara } 925*9448f2b8SAndre Przywara } 926*9448f2b8SAndre Przywara 92728f39f02SMax Shvetsov /******************************************************************************* 92828f39f02SMax Shvetsov * Save EL2 sysreg context 92928f39f02SMax Shvetsov ******************************************************************************/ 93028f39f02SMax Shvetsov void cm_el2_sysregs_context_save(uint32_t security_state) 93128f39f02SMax Shvetsov { 93228f39f02SMax Shvetsov u_register_t scr_el3 = read_scr(); 93328f39f02SMax Shvetsov 93428f39f02SMax Shvetsov /* 935c5ea4f8aSZelalem Aweke * Always save the non-secure and realm EL2 context, only save the 93628f39f02SMax Shvetsov * S-EL2 context if S-EL2 is enabled. 93728f39f02SMax Shvetsov */ 938c5ea4f8aSZelalem Aweke if ((security_state != SECURE) || 9396b704da3SRuari Phipps ((security_state == SECURE) && ((scr_el3 & SCR_EEL2_BIT) != 0U))) { 94028f39f02SMax Shvetsov cpu_context_t *ctx; 941d20052f3SZelalem Aweke el2_sysregs_t *el2_sysregs_ctx; 94228f39f02SMax Shvetsov 94328f39f02SMax Shvetsov ctx = cm_get_context(security_state); 94428f39f02SMax Shvetsov assert(ctx != NULL); 94528f39f02SMax Shvetsov 946d20052f3SZelalem Aweke el2_sysregs_ctx = get_el2_sysregs_ctx(ctx); 947d20052f3SZelalem Aweke 948d20052f3SZelalem Aweke el2_sysregs_context_save_common(el2_sysregs_ctx); 949d20052f3SZelalem Aweke #if CTX_INCLUDE_MTE_REGS 950d20052f3SZelalem Aweke el2_sysregs_context_save_mte(el2_sysregs_ctx); 951d20052f3SZelalem Aweke #endif 952*9448f2b8SAndre Przywara if (is_feat_mpam_supported()) { 953d20052f3SZelalem Aweke el2_sysregs_context_save_mpam(el2_sysregs_ctx); 954*9448f2b8SAndre Przywara } 955bb7b85a3SAndre Przywara 956de8c4892SAndre Przywara if (is_feat_fgt_supported()) { 957d20052f3SZelalem Aweke el2_sysregs_context_save_fgt(el2_sysregs_ctx); 958de8c4892SAndre Przywara } 959bb7b85a3SAndre Przywara 960d20052f3SZelalem Aweke #if ENABLE_FEAT_ECV 961d20052f3SZelalem Aweke el2_sysregs_context_save_ecv(el2_sysregs_ctx); 962d20052f3SZelalem Aweke #endif 963d20052f3SZelalem Aweke #if ENABLE_FEAT_VHE 964d20052f3SZelalem Aweke el2_sysregs_context_save_vhe(el2_sysregs_ctx); 965d20052f3SZelalem Aweke #endif 966d20052f3SZelalem Aweke #if RAS_EXTENSION 967d20052f3SZelalem Aweke el2_sysregs_context_save_ras(el2_sysregs_ctx); 968d20052f3SZelalem Aweke #endif 969d20052f3SZelalem Aweke #if CTX_INCLUDE_NEVE_REGS 970d20052f3SZelalem Aweke el2_sysregs_context_save_nv2(el2_sysregs_ctx); 971d20052f3SZelalem Aweke #endif 972fc8d2d39SAndre Przywara if (is_feat_trf_supported()) { 973fc8d2d39SAndre Przywara write_ctx_reg(el2_sysregs_ctx, CTX_TRFCR_EL2, read_trfcr_el2()); 974fc8d2d39SAndre Przywara } 975d20052f3SZelalem Aweke #if ENABLE_FEAT_CSV2_2 976d20052f3SZelalem Aweke el2_sysregs_context_save_csv2(el2_sysregs_ctx); 977d20052f3SZelalem Aweke #endif 978c5a3ebbdSAndre Przywara if (is_feat_hcx_supported()) { 979c5a3ebbdSAndre Przywara write_ctx_reg(el2_sysregs_ctx, CTX_HCRX_EL2, read_hcrx_el2()); 980c5a3ebbdSAndre Przywara } 981d3331603SMark Brown if (is_feat_tcr2_supported()) { 982d3331603SMark Brown write_ctx_reg(el2_sysregs_ctx, CTX_TCR2_EL2, read_tcr2_el2()); 983d3331603SMark Brown } 98428f39f02SMax Shvetsov } 98528f39f02SMax Shvetsov } 98628f39f02SMax Shvetsov 98728f39f02SMax Shvetsov /******************************************************************************* 98828f39f02SMax Shvetsov * Restore EL2 sysreg context 98928f39f02SMax Shvetsov ******************************************************************************/ 99028f39f02SMax Shvetsov void cm_el2_sysregs_context_restore(uint32_t security_state) 99128f39f02SMax Shvetsov { 99228f39f02SMax Shvetsov u_register_t scr_el3 = read_scr(); 99328f39f02SMax Shvetsov 99428f39f02SMax Shvetsov /* 995c5ea4f8aSZelalem Aweke * Always restore the non-secure and realm EL2 context, only restore the 99628f39f02SMax Shvetsov * S-EL2 context if S-EL2 is enabled. 99728f39f02SMax Shvetsov */ 998c5ea4f8aSZelalem Aweke if ((security_state != SECURE) || 9996b704da3SRuari Phipps ((security_state == SECURE) && ((scr_el3 & SCR_EEL2_BIT) != 0U))) { 100028f39f02SMax Shvetsov cpu_context_t *ctx; 1001d20052f3SZelalem Aweke el2_sysregs_t *el2_sysregs_ctx; 100228f39f02SMax Shvetsov 100328f39f02SMax Shvetsov ctx = cm_get_context(security_state); 100428f39f02SMax Shvetsov assert(ctx != NULL); 100528f39f02SMax Shvetsov 1006d20052f3SZelalem Aweke el2_sysregs_ctx = get_el2_sysregs_ctx(ctx); 1007d20052f3SZelalem Aweke 1008d20052f3SZelalem Aweke el2_sysregs_context_restore_common(el2_sysregs_ctx); 1009d20052f3SZelalem Aweke #if CTX_INCLUDE_MTE_REGS 1010d20052f3SZelalem Aweke el2_sysregs_context_restore_mte(el2_sysregs_ctx); 1011d20052f3SZelalem Aweke #endif 1012*9448f2b8SAndre Przywara if (is_feat_mpam_supported()) { 1013d20052f3SZelalem Aweke el2_sysregs_context_restore_mpam(el2_sysregs_ctx); 1014*9448f2b8SAndre Przywara } 1015bb7b85a3SAndre Przywara 1016de8c4892SAndre Przywara if (is_feat_fgt_supported()) { 1017d20052f3SZelalem Aweke el2_sysregs_context_restore_fgt(el2_sysregs_ctx); 1018de8c4892SAndre Przywara } 1019bb7b85a3SAndre Przywara 1020d20052f3SZelalem Aweke #if ENABLE_FEAT_ECV 1021d20052f3SZelalem Aweke el2_sysregs_context_restore_ecv(el2_sysregs_ctx); 1022d20052f3SZelalem Aweke #endif 1023d20052f3SZelalem Aweke #if ENABLE_FEAT_VHE 1024d20052f3SZelalem Aweke el2_sysregs_context_restore_vhe(el2_sysregs_ctx); 1025d20052f3SZelalem Aweke #endif 1026d20052f3SZelalem Aweke #if RAS_EXTENSION 1027d20052f3SZelalem Aweke el2_sysregs_context_restore_ras(el2_sysregs_ctx); 1028d20052f3SZelalem Aweke #endif 1029d20052f3SZelalem Aweke #if CTX_INCLUDE_NEVE_REGS 1030d20052f3SZelalem Aweke el2_sysregs_context_restore_nv2(el2_sysregs_ctx); 1031d20052f3SZelalem Aweke #endif 1032fc8d2d39SAndre Przywara if (is_feat_trf_supported()) { 1033fc8d2d39SAndre Przywara write_trfcr_el2(read_ctx_reg(el2_sysregs_ctx, CTX_TRFCR_EL2)); 1034fc8d2d39SAndre Przywara } 1035d20052f3SZelalem Aweke #if ENABLE_FEAT_CSV2_2 1036d20052f3SZelalem Aweke el2_sysregs_context_restore_csv2(el2_sysregs_ctx); 1037d20052f3SZelalem Aweke #endif 1038c5a3ebbdSAndre Przywara if (is_feat_hcx_supported()) { 1039c5a3ebbdSAndre Przywara write_hcrx_el2(read_ctx_reg(el2_sysregs_ctx, CTX_HCRX_EL2)); 1040c5a3ebbdSAndre Przywara } 1041d3331603SMark Brown if (is_feat_tcr2_supported()) { 1042d3331603SMark Brown write_tcr2_el2(read_ctx_reg(el2_sysregs_ctx, CTX_TCR2_EL2)); 1043d3331603SMark Brown } 104428f39f02SMax Shvetsov } 104528f39f02SMax Shvetsov } 104628f39f02SMax Shvetsov #endif /* CTX_INCLUDE_EL2_REGS */ 104728f39f02SMax Shvetsov 1048532ed618SSoby Mathew /******************************************************************************* 10498b95e848SZelalem Aweke * This function is used to exit to Non-secure world. If CTX_INCLUDE_EL2_REGS 10508b95e848SZelalem Aweke * is enabled, it restores EL1 and EL2 sysreg contexts instead of directly 10518b95e848SZelalem Aweke * updating EL1 and EL2 registers. Otherwise, it calls the generic 10528b95e848SZelalem Aweke * cm_prepare_el3_exit function. 10538b95e848SZelalem Aweke ******************************************************************************/ 10548b95e848SZelalem Aweke void cm_prepare_el3_exit_ns(void) 10558b95e848SZelalem Aweke { 10568b95e848SZelalem Aweke #if CTX_INCLUDE_EL2_REGS 10578b95e848SZelalem Aweke cpu_context_t *ctx = cm_get_context(NON_SECURE); 10588b95e848SZelalem Aweke assert(ctx != NULL); 10598b95e848SZelalem Aweke 1060b515f541SZelalem Aweke /* Assert that EL2 is used. */ 1061b515f541SZelalem Aweke #if ENABLE_ASSERTIONS 1062b515f541SZelalem Aweke el3_state_t *state = get_el3state_ctx(ctx); 1063b515f541SZelalem Aweke u_register_t scr_el3 = read_ctx_reg(state, CTX_SCR_EL3); 1064b515f541SZelalem Aweke #endif 1065b515f541SZelalem Aweke assert(((scr_el3 & SCR_HCE_BIT) != 0UL) && 1066b515f541SZelalem Aweke (el_implemented(2U) != EL_IMPL_NONE)); 1067b515f541SZelalem Aweke 10688b95e848SZelalem Aweke /* 10698b95e848SZelalem Aweke * Currently some extensions are configured using 10708b95e848SZelalem Aweke * direct register updates. Therefore, do this here 10718b95e848SZelalem Aweke * instead of when setting up context. 10728b95e848SZelalem Aweke */ 10738b95e848SZelalem Aweke manage_extensions_nonsecure(0, ctx); 10748b95e848SZelalem Aweke 10758b95e848SZelalem Aweke /* 10768b95e848SZelalem Aweke * Set the NS bit to be able to access the ICC_SRE_EL2 10778b95e848SZelalem Aweke * register when restoring context. 10788b95e848SZelalem Aweke */ 10798b95e848SZelalem Aweke write_scr_el3(read_scr_el3() | SCR_NS_BIT); 10808b95e848SZelalem Aweke 108104825031SOlivier Deprez /* 108204825031SOlivier Deprez * Ensure the NS bit change is committed before the EL2/EL1 108304825031SOlivier Deprez * state restoration. 108404825031SOlivier Deprez */ 108504825031SOlivier Deprez isb(); 108604825031SOlivier Deprez 10878b95e848SZelalem Aweke /* Restore EL2 and EL1 sysreg contexts */ 10888b95e848SZelalem Aweke cm_el2_sysregs_context_restore(NON_SECURE); 10898b95e848SZelalem Aweke cm_el1_sysregs_context_restore(NON_SECURE); 10908b95e848SZelalem Aweke cm_set_next_eret_context(NON_SECURE); 10918b95e848SZelalem Aweke #else 10928b95e848SZelalem Aweke cm_prepare_el3_exit(NON_SECURE); 10938b95e848SZelalem Aweke #endif /* CTX_INCLUDE_EL2_REGS */ 10948b95e848SZelalem Aweke } 10958b95e848SZelalem Aweke 10968b95e848SZelalem Aweke /******************************************************************************* 1097532ed618SSoby Mathew * The next four functions are used by runtime services to save and restore 1098532ed618SSoby Mathew * EL1 context on the 'cpu_context' structure for the specified security 1099532ed618SSoby Mathew * state. 1100532ed618SSoby Mathew ******************************************************************************/ 1101532ed618SSoby Mathew void cm_el1_sysregs_context_save(uint32_t security_state) 1102532ed618SSoby Mathew { 1103532ed618SSoby Mathew cpu_context_t *ctx; 1104532ed618SSoby Mathew 1105532ed618SSoby Mathew ctx = cm_get_context(security_state); 1106a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1107532ed618SSoby Mathew 11082825946eSMax Shvetsov el1_sysregs_context_save(get_el1_sysregs_ctx(ctx)); 110917b4c0ddSDimitris Papastamos 111017b4c0ddSDimitris Papastamos #if IMAGE_BL31 111117b4c0ddSDimitris Papastamos if (security_state == SECURE) 111217b4c0ddSDimitris Papastamos PUBLISH_EVENT(cm_exited_secure_world); 111317b4c0ddSDimitris Papastamos else 111417b4c0ddSDimitris Papastamos PUBLISH_EVENT(cm_exited_normal_world); 111517b4c0ddSDimitris Papastamos #endif 1116532ed618SSoby Mathew } 1117532ed618SSoby Mathew 1118532ed618SSoby Mathew void cm_el1_sysregs_context_restore(uint32_t security_state) 1119532ed618SSoby Mathew { 1120532ed618SSoby Mathew cpu_context_t *ctx; 1121532ed618SSoby Mathew 1122532ed618SSoby Mathew ctx = cm_get_context(security_state); 1123a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1124532ed618SSoby Mathew 11252825946eSMax Shvetsov el1_sysregs_context_restore(get_el1_sysregs_ctx(ctx)); 112617b4c0ddSDimitris Papastamos 112717b4c0ddSDimitris Papastamos #if IMAGE_BL31 112817b4c0ddSDimitris Papastamos if (security_state == SECURE) 112917b4c0ddSDimitris Papastamos PUBLISH_EVENT(cm_entering_secure_world); 113017b4c0ddSDimitris Papastamos else 113117b4c0ddSDimitris Papastamos PUBLISH_EVENT(cm_entering_normal_world); 113217b4c0ddSDimitris Papastamos #endif 1133532ed618SSoby Mathew } 1134532ed618SSoby Mathew 1135532ed618SSoby Mathew /******************************************************************************* 1136532ed618SSoby Mathew * This function populates ELR_EL3 member of 'cpu_context' pertaining to the 1137532ed618SSoby Mathew * given security state with the given entrypoint 1138532ed618SSoby Mathew ******************************************************************************/ 1139532ed618SSoby Mathew void cm_set_elr_el3(uint32_t security_state, uintptr_t entrypoint) 1140532ed618SSoby Mathew { 1141532ed618SSoby Mathew cpu_context_t *ctx; 1142532ed618SSoby Mathew el3_state_t *state; 1143532ed618SSoby Mathew 1144532ed618SSoby Mathew ctx = cm_get_context(security_state); 1145a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1146532ed618SSoby Mathew 1147532ed618SSoby Mathew /* Populate EL3 state so that ERET jumps to the correct entry */ 1148532ed618SSoby Mathew state = get_el3state_ctx(ctx); 1149532ed618SSoby Mathew write_ctx_reg(state, CTX_ELR_EL3, entrypoint); 1150532ed618SSoby Mathew } 1151532ed618SSoby Mathew 1152532ed618SSoby Mathew /******************************************************************************* 1153532ed618SSoby Mathew * This function populates ELR_EL3 and SPSR_EL3 members of 'cpu_context' 1154532ed618SSoby Mathew * pertaining to the given security state 1155532ed618SSoby Mathew ******************************************************************************/ 1156532ed618SSoby Mathew void cm_set_elr_spsr_el3(uint32_t security_state, 1157532ed618SSoby Mathew uintptr_t entrypoint, uint32_t spsr) 1158532ed618SSoby Mathew { 1159532ed618SSoby Mathew cpu_context_t *ctx; 1160532ed618SSoby Mathew el3_state_t *state; 1161532ed618SSoby Mathew 1162532ed618SSoby Mathew ctx = cm_get_context(security_state); 1163a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1164532ed618SSoby Mathew 1165532ed618SSoby Mathew /* Populate EL3 state so that ERET jumps to the correct entry */ 1166532ed618SSoby Mathew state = get_el3state_ctx(ctx); 1167532ed618SSoby Mathew write_ctx_reg(state, CTX_ELR_EL3, entrypoint); 1168532ed618SSoby Mathew write_ctx_reg(state, CTX_SPSR_EL3, spsr); 1169532ed618SSoby Mathew } 1170532ed618SSoby Mathew 1171532ed618SSoby Mathew /******************************************************************************* 1172532ed618SSoby Mathew * This function updates a single bit in the SCR_EL3 member of the 'cpu_context' 1173532ed618SSoby Mathew * pertaining to the given security state using the value and bit position 1174532ed618SSoby Mathew * specified in the parameters. It preserves all other bits. 1175532ed618SSoby Mathew ******************************************************************************/ 1176532ed618SSoby Mathew void cm_write_scr_el3_bit(uint32_t security_state, 1177532ed618SSoby Mathew uint32_t bit_pos, 1178532ed618SSoby Mathew uint32_t value) 1179532ed618SSoby Mathew { 1180532ed618SSoby Mathew cpu_context_t *ctx; 1181532ed618SSoby Mathew el3_state_t *state; 1182f1be00daSLouis Mayencourt u_register_t scr_el3; 1183532ed618SSoby Mathew 1184532ed618SSoby Mathew ctx = cm_get_context(security_state); 1185a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1186532ed618SSoby Mathew 1187532ed618SSoby Mathew /* Ensure that the bit position is a valid one */ 1188d7b5f408SJimmy Brisson assert(((1UL << bit_pos) & SCR_VALID_BIT_MASK) != 0U); 1189532ed618SSoby Mathew 1190532ed618SSoby Mathew /* Ensure that the 'value' is only a bit wide */ 1191a0fee747SAntonio Nino Diaz assert(value <= 1U); 1192532ed618SSoby Mathew 1193532ed618SSoby Mathew /* 1194532ed618SSoby Mathew * Get the SCR_EL3 value from the cpu context, clear the desired bit 1195532ed618SSoby Mathew * and set it to its new value. 1196532ed618SSoby Mathew */ 1197532ed618SSoby Mathew state = get_el3state_ctx(ctx); 1198f1be00daSLouis Mayencourt scr_el3 = read_ctx_reg(state, CTX_SCR_EL3); 1199d7b5f408SJimmy Brisson scr_el3 &= ~(1UL << bit_pos); 1200f1be00daSLouis Mayencourt scr_el3 |= (u_register_t)value << bit_pos; 1201532ed618SSoby Mathew write_ctx_reg(state, CTX_SCR_EL3, scr_el3); 1202532ed618SSoby Mathew } 1203532ed618SSoby Mathew 1204532ed618SSoby Mathew /******************************************************************************* 1205532ed618SSoby Mathew * This function retrieves SCR_EL3 member of 'cpu_context' pertaining to the 1206532ed618SSoby Mathew * given security state. 1207532ed618SSoby Mathew ******************************************************************************/ 1208f1be00daSLouis Mayencourt u_register_t cm_get_scr_el3(uint32_t security_state) 1209532ed618SSoby Mathew { 1210532ed618SSoby Mathew cpu_context_t *ctx; 1211532ed618SSoby Mathew el3_state_t *state; 1212532ed618SSoby Mathew 1213532ed618SSoby Mathew ctx = cm_get_context(security_state); 1214a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1215532ed618SSoby Mathew 1216532ed618SSoby Mathew /* Populate EL3 state so that ERET jumps to the correct entry */ 1217532ed618SSoby Mathew state = get_el3state_ctx(ctx); 1218f1be00daSLouis Mayencourt return read_ctx_reg(state, CTX_SCR_EL3); 1219532ed618SSoby Mathew } 1220532ed618SSoby Mathew 1221532ed618SSoby Mathew /******************************************************************************* 1222532ed618SSoby Mathew * This function is used to program the context that's used for exception 1223532ed618SSoby Mathew * return. This initializes the SP_EL3 to a pointer to a 'cpu_context' set for 1224532ed618SSoby Mathew * the required security state 1225532ed618SSoby Mathew ******************************************************************************/ 1226532ed618SSoby Mathew void cm_set_next_eret_context(uint32_t security_state) 1227532ed618SSoby Mathew { 1228532ed618SSoby Mathew cpu_context_t *ctx; 1229532ed618SSoby Mathew 1230532ed618SSoby Mathew ctx = cm_get_context(security_state); 1231a0fee747SAntonio Nino Diaz assert(ctx != NULL); 1232532ed618SSoby Mathew 1233532ed618SSoby Mathew cm_set_next_context(ctx); 1234532ed618SSoby Mathew } 1235