1 // SPDX-License-Identifier: BSD-2-Clause 2 /* 3 * Copyright (c) 2016-2022, Linaro Limited 4 */ 5 6 #include <gen-asm-defines.h> 7 #include <kernel/boot.h> 8 #include <kernel/thread.h> 9 #include <kernel/thread_private.h> 10 #include <mm/core_mmu_arch.h> 11 #include <sm/pm.h> 12 #include <sm/sm.h> 13 #include <types_ext.h> 14 15 DEFINES 16 { 17 #ifdef ARM32 18 DEFINE(SM_NSEC_CTX_R0, offsetof(struct sm_nsec_ctx, r0)); 19 DEFINE(SM_NSEC_CTX_R8, offsetof(struct sm_nsec_ctx, r8)); 20 DEFINE(SM_SEC_CTX_R0, offsetof(struct sm_sec_ctx, r0)); 21 DEFINE(SM_SEC_CTX_MON_LR, offsetof(struct sm_sec_ctx, mon_lr)); 22 DEFINE(SM_CTX_SEC_SIZE, sizeof(struct sm_sec_ctx)); 23 DEFINE(SM_CTX_SIZE, sizeof(struct sm_ctx)); 24 DEFINE(SM_CTX_NSEC, offsetof(struct sm_ctx, nsec)); 25 DEFINE(SM_CTX_SEC, offsetof(struct sm_ctx, sec)); 26 27 DEFINE(THREAD_SCALL_REG_R0, offsetof(struct thread_scall_regs, r0)); 28 DEFINE(THREAD_SCALL_REG_R5, offsetof(struct thread_scall_regs, r5)); 29 DEFINE(THREAD_SCALL_REG_R6, offsetof(struct thread_scall_regs, r6)); 30 31 /* struct thread_ctx */ 32 DEFINE(THREAD_CTX_STACK_VA_END, offsetof(struct thread_ctx, 33 stack_va_end)); 34 35 /* struct thread_ctx_regs */ 36 DEFINE(THREAD_CTX_REGS_USR_SP, 37 offsetof(struct thread_ctx_regs, usr_sp)); 38 DEFINE(THREAD_CTX_REGS_PC, offsetof(struct thread_ctx_regs, pc)); 39 DEFINE(THREAD_CTX_REGS_CPSR, offsetof(struct thread_ctx_regs, cpsr)); 40 41 /* struct thread_core_local */ 42 DEFINE(THREAD_CORE_LOCAL_R0, offsetof(struct thread_core_local, r[0])); 43 DEFINE(THREAD_CORE_LOCAL_SM_PM_CTX_PHYS, 44 offsetof(struct thread_core_local, sm_pm_ctx_phys)); 45 DEFINE(THREAD_CORE_LOCAL_SIZE, sizeof(struct thread_core_local)); 46 47 DEFINE(SM_PM_CTX_SIZE, sizeof(struct sm_pm_ctx)); 48 #endif /*ARM32*/ 49 50 #ifdef ARM64 51 DEFINE(THREAD_SMC_ARGS_X0, offsetof(struct thread_smc_args, a0)); 52 DEFINE(THREAD_SMC_ARGS_SIZE, sizeof(struct thread_smc_args)); 53 DEFINE(THREAD_SMC_1_2_REGS_SIZE, sizeof(struct thread_smc_1_2_regs)); 54 55 DEFINE(THREAD_SCALL_REG_X0, offsetof(struct thread_scall_regs, x0)); 56 DEFINE(THREAD_SCALL_REG_X2, offsetof(struct thread_scall_regs, x2)); 57 DEFINE(THREAD_SCALL_REG_X5, offsetof(struct thread_scall_regs, x5)); 58 DEFINE(THREAD_SCALL_REG_X6, offsetof(struct thread_scall_regs, x6)); 59 DEFINE(THREAD_SCALL_REG_X30, offsetof(struct thread_scall_regs, x30)); 60 DEFINE(THREAD_SCALL_REG_ELR, offsetof(struct thread_scall_regs, elr)); 61 DEFINE(THREAD_SCALL_REG_SPSR, offsetof(struct thread_scall_regs, spsr)); 62 DEFINE(THREAD_SCALL_REG_SP_EL0, offsetof(struct thread_scall_regs, 63 sp_el0)); 64 #ifdef CFG_TA_PAUTH 65 DEFINE(THREAD_SCALL_REG_APIAKEY_HI, offsetof(struct thread_scall_regs, 66 apiakey_hi)); 67 #endif 68 DEFINE(THREAD_SCALL_REG_SIZE, sizeof(struct thread_scall_regs)); 69 70 /* struct thread_abort_regs */ 71 DEFINE(THREAD_ABT_REG_X0, offsetof(struct thread_abort_regs, x0)); 72 DEFINE(THREAD_ABT_REG_X2, offsetof(struct thread_abort_regs, x2)); 73 DEFINE(THREAD_ABT_REG_X30, offsetof(struct thread_abort_regs, x30)); 74 DEFINE(THREAD_ABT_REG_SPSR, offsetof(struct thread_abort_regs, spsr)); 75 DEFINE(THREAD_ABT_REGS_SIZE, sizeof(struct thread_abort_regs)); 76 #if defined(CFG_TA_PAUTH) || defined(CFG_CORE_PAUTH) 77 DEFINE(THREAD_ABT_REGS_APIAKEY_HI, offsetof(struct thread_abort_regs, 78 apiakey_hi)); 79 #endif 80 81 /* struct thread_ctx */ 82 DEFINE(THREAD_CTX_KERN_SP, offsetof(struct thread_ctx, kern_sp)); 83 DEFINE(THREAD_CTX_STACK_VA_END, offsetof(struct thread_ctx, 84 stack_va_end)); 85 #if defined(CFG_CORE_PAUTH) 86 DEFINE(THREAD_CTX_KEYS, offsetof(struct thread_ctx, keys)); 87 #endif 88 89 /* struct thread_ctx_regs */ 90 DEFINE(THREAD_CTX_REGS_SP, offsetof(struct thread_ctx_regs, sp)); 91 DEFINE(THREAD_CTX_REGS_X0, offsetof(struct thread_ctx_regs, x[0])); 92 DEFINE(THREAD_CTX_REGS_X1, offsetof(struct thread_ctx_regs, x[1])); 93 DEFINE(THREAD_CTX_REGS_X2, offsetof(struct thread_ctx_regs, x[2])); 94 DEFINE(THREAD_CTX_REGS_X4, offsetof(struct thread_ctx_regs, x[4])); 95 DEFINE(THREAD_CTX_REGS_X19, offsetof(struct thread_ctx_regs, x[19])); 96 DEFINE(THREAD_CTX_REGS_TPIDR_EL0, offsetof(struct thread_ctx_regs, 97 tpidr_el0)); 98 #if defined(CFG_TA_PAUTH) || defined(CFG_CORE_PAUTH) 99 DEFINE(THREAD_CTX_REGS_APIAKEY_HI, offsetof(struct thread_ctx_regs, 100 apiakey_hi)); 101 #endif 102 103 /* struct thread_user_mode_rec */ 104 DEFINE(THREAD_USER_MODE_REC_CTX_REGS_PTR, 105 offsetof(struct thread_user_mode_rec, ctx_regs_ptr)); 106 DEFINE(THREAD_USER_MODE_REC_EXIT_STATUS0_PTR, 107 offsetof(struct thread_user_mode_rec, exit_status0_ptr)); 108 DEFINE(THREAD_USER_MODE_REC_X19, 109 offsetof(struct thread_user_mode_rec, x[0])); 110 DEFINE(THREAD_USER_MODE_REC_SIZE, sizeof(struct thread_user_mode_rec)); 111 112 /* struct thread_core_local */ 113 DEFINE(THREAD_CORE_LOCAL_X0, offsetof(struct thread_core_local, x[0])); 114 DEFINE(THREAD_CORE_LOCAL_X2, offsetof(struct thread_core_local, x[2])); 115 DEFINE(THREAD_CORE_LOCAL_KCODE_OFFSET, 116 offsetof(struct thread_core_local, kcode_offset)); 117 #ifdef CFG_CORE_WORKAROUND_SPECTRE_BP_SEC 118 DEFINE(THREAD_CORE_LOCAL_BHB_LOOP_COUNT, 119 offsetof(struct thread_core_local, bhb_loop_count)); 120 #endif 121 #if defined(CFG_CORE_PAUTH) 122 DEFINE(THREAD_CORE_LOCAL_KEYS, 123 offsetof(struct thread_core_local, keys)); 124 #endif 125 #endif /*ARM64*/ 126 127 /* struct thread_ctx */ 128 DEFINE(THREAD_CTX_SIZE, sizeof(struct thread_ctx)); 129 #ifdef CFG_CORE_FFA 130 DEFINE(THREAD_CTX_TSD_RPC_TARGET_INFO, 131 offsetof(struct thread_ctx, tsd.rpc_target_info)) 132 DEFINE(THREAD_CTX_FLAGS, 133 offsetof(struct thread_ctx, flags)) 134 #endif 135 136 /* struct thread_core_local */ 137 DEFINE(THREAD_CORE_LOCAL_TMP_STACK_VA_END, 138 offsetof(struct thread_core_local, tmp_stack_va_end)); 139 DEFINE(THREAD_CORE_LOCAL_CURR_THREAD, 140 offsetof(struct thread_core_local, curr_thread)); 141 DEFINE(THREAD_CORE_LOCAL_FLAGS, 142 offsetof(struct thread_core_local, flags)); 143 DEFINE(THREAD_CORE_LOCAL_ABT_STACK_VA_END, 144 offsetof(struct thread_core_local, abt_stack_va_end)); 145 #if defined(ARM64) && defined(CFG_CORE_FFA) 146 DEFINE(THREAD_CORE_LOCAL_DIRECT_RESP_FID, 147 offsetof(struct thread_core_local, direct_resp_fid)); 148 #endif 149 150 DEFINE(STACK_TMP_GUARD, STACK_CANARY_SIZE / 2 + STACK_TMP_OFFS); 151 152 /* struct core_mmu_config */ 153 DEFINE(CORE_MMU_CONFIG_SIZE, sizeof(struct core_mmu_config)); 154 DEFINE(CORE_MMU_CONFIG_MAP_OFFSET, 155 offsetof(struct core_mmu_config, map_offset)); 156 157 /* struct boot_embdata */ 158 DEFINE(BOOT_EMBDATA_HASHES_OFFSET, 159 offsetof(struct boot_embdata, hashes_offset)); 160 DEFINE(BOOT_EMBDATA_HASHES_LEN, 161 offsetof(struct boot_embdata, hashes_len)); 162 DEFINE(BOOT_EMBDATA_RELOC_OFFSET, 163 offsetof(struct boot_embdata, reloc_offset)); 164 DEFINE(BOOT_EMBDATA_RELOC_LEN, 165 offsetof(struct boot_embdata, reloc_len)); 166 167 #ifdef CORE_MMU_BASE_TABLE_OFFSET 168 /* 169 * This define is too complex to be used as an argument for the 170 * macros add_imm and sub_imm so evaluate it here. 171 */ 172 DEFINE(__CORE_MMU_BASE_TABLE_OFFSET, CORE_MMU_BASE_TABLE_OFFSET); 173 #endif 174 175 } 176