/* * Copyright (c) 2016-2025, Arm Limited and Contributors. All rights reserved. * * SPDX-License-Identifier: BSD-3-Clause */ #include #include #include #include #include #include #include /* * The below function returns the cpu_ops structure matching the * midr of the core. It reads the MIDR and finds the matching * entry in cpu_ops entries. Only the implementation and part number * are used to match the entries. * Return : * r0 - The matching cpu_ops pointer on Success * r0 - 0 on failure. * Clobbers: r0 - r5 */ .macro get_cpu_ops_ptr_impl /* Get the cpu_ops start and end locations */ ldr r4, =(__CPU_OPS_START__ + CPU_MIDR) ldr r5, =(__CPU_OPS_END__ + CPU_MIDR) /* Initialize the return parameter */ mov r0, #0 /* Read the MIDR_EL1 */ ldcopr r2, MIDR ldr r3, =CPU_IMPL_PN_MASK /* Retain only the implementation and part number using mask */ and r2, r2, r3 1: /* Check if we have reached end of list */ cmp r4, r5 bhs error_exit\@ /* load the midr from the cpu_ops */ ldr r1, [r4], #CPU_OPS_SIZE and r1, r1, r3 /* Check if midr matches to midr of this core */ cmp r1, r2 bne 1b /* Subtract the increment and offset to get the cpu-ops pointer */ sub r0, r4, #(CPU_OPS_SIZE + CPU_MIDR) #if ENABLE_ASSERTIONS cmp r0, #0 ASM_ASSERT(ne) #endif error_exit\@: .endm #if defined(IMAGE_BL1) || defined(IMAGE_BL32) || \ (defined(IMAGE_BL2) && RESET_TO_BL2) /* * The reset handler common to all platforms. After a matching * cpu_ops structure entry is found, the correponding reset_handler * in the cpu_ops is invoked. The reset handler is invoked very early * in the boot sequence and it is assumed that we can clobber r0 - r10 * without the need to follow AAPCS. * Clobbers: r0 - r10 */ .globl reset_handler func reset_handler mov r8, lr /* The plat_reset_handler can clobber r0 - r7 */ bl plat_reset_handler /* Get the matching cpu_ops pointer (clobbers: r0 - r5) */ get_cpu_ops_ptr_impl #if ENABLE_ASSERTIONS cmp r0, #0 ASM_ASSERT(ne) #endif /* Get the cpu_ops reset handler */ ldr r1, [r0, #CPU_RESET_FUNC] cmp r1, #0 mov lr, r8 bxne r1 bx lr endfunc reset_handler #endif .globl get_cpu_ops_ptr /* performs an AAPCS compliant call to get_cpu_ops_ptr_impl */ func get_cpu_ops_ptr push {r4 - r5, lr} get_cpu_ops_ptr_impl pop {r4 - r5, pc} endfunc get_cpu_ops_ptr /* * Extract CPU revision and variant, and combine them into a single numeric for * easier comparison. */ .globl cpu_get_rev_var func cpu_get_rev_var ldcopr r1, MIDR /* * Extract the variant[23:20] and revision[3:0] from r1 and pack it in * r0[0:7] as variant[7:4] and revision[3:0]: * * First extract r1[23:16] to r0[7:0] and zero fill the rest. Then * extract r1[3:0] into r0[3:0] retaining other bits. */ ubfx r0, r1, #(MIDR_VAR_SHIFT - MIDR_REV_BITS), #(MIDR_REV_BITS + MIDR_VAR_BITS) bfi r0, r1, #MIDR_REV_SHIFT, #MIDR_REV_BITS bx lr endfunc cpu_get_rev_var /* * Compare the CPU's revision-variant (r0) with a given value (r1), for errata * application purposes. If the revision-variant is less than or same as a given * value, indicates that errata applies; otherwise not. */ .globl cpu_rev_var_ls func cpu_rev_var_ls cmp r0, r1 movls r0, #ERRATA_APPLIES movhi r0, #ERRATA_NOT_APPLIES bx lr endfunc cpu_rev_var_ls /* * Compare the CPU's revision-variant (r0) with a given value (r1), for errata * application purposes. If the revision-variant is higher than or same as a * given value, indicates that errata applies; otherwise not. */ .globl cpu_rev_var_hs func cpu_rev_var_hs cmp r0, r1 movge r0, #ERRATA_APPLIES movlt r0, #ERRATA_NOT_APPLIES bx lr endfunc cpu_rev_var_hs