14f6ad66aSAchin Gupta /* 2*e83b0cadSDan Handley * Copyright (c) 2013-2014, ARM Limited and Contributors. All rights reserved. 34f6ad66aSAchin Gupta * 44f6ad66aSAchin Gupta * Redistribution and use in source and binary forms, with or without 54f6ad66aSAchin Gupta * modification, are permitted provided that the following conditions are met: 64f6ad66aSAchin Gupta * 74f6ad66aSAchin Gupta * Redistributions of source code must retain the above copyright notice, this 84f6ad66aSAchin Gupta * list of conditions and the following disclaimer. 94f6ad66aSAchin Gupta * 104f6ad66aSAchin Gupta * Redistributions in binary form must reproduce the above copyright notice, 114f6ad66aSAchin Gupta * this list of conditions and the following disclaimer in the documentation 124f6ad66aSAchin Gupta * and/or other materials provided with the distribution. 134f6ad66aSAchin Gupta * 144f6ad66aSAchin Gupta * Neither the name of ARM nor the names of its contributors may be used 154f6ad66aSAchin Gupta * to endorse or promote products derived from this software without specific 164f6ad66aSAchin Gupta * prior written permission. 174f6ad66aSAchin Gupta * 184f6ad66aSAchin Gupta * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 194f6ad66aSAchin Gupta * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 204f6ad66aSAchin Gupta * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 214f6ad66aSAchin Gupta * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE 224f6ad66aSAchin Gupta * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 234f6ad66aSAchin Gupta * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 244f6ad66aSAchin Gupta * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 254f6ad66aSAchin Gupta * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 264f6ad66aSAchin Gupta * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 274f6ad66aSAchin Gupta * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 284f6ad66aSAchin Gupta * POSSIBILITY OF SUCH DAMAGE. 294f6ad66aSAchin Gupta */ 304f6ad66aSAchin Gupta 314f6ad66aSAchin Gupta #include <stdio.h> 324f6ad66aSAchin Gupta #include <string.h> 334f6ad66aSAchin Gupta #include <errno.h> 344f6ad66aSAchin Gupta #include <assert.h> 354f6ad66aSAchin Gupta #include <arch_helpers.h> 364f6ad66aSAchin Gupta #include <console.h> 374f6ad66aSAchin Gupta #include <platform.h> 384f6ad66aSAchin Gupta #include <semihosting.h> 394f6ad66aSAchin Gupta #include <bl_common.h> 404f6ad66aSAchin Gupta 414f6ad66aSAchin Gupta /*********************************************************** 424f6ad66aSAchin Gupta * Memory for sharing data while changing exception levels. 434f6ad66aSAchin Gupta * Only used by the primary core. 444f6ad66aSAchin Gupta **********************************************************/ 454f6ad66aSAchin Gupta unsigned char bl2_el_change_mem_ptr[EL_CHANGE_MEM_SIZE]; 464f6ad66aSAchin Gupta 474f6ad66aSAchin Gupta unsigned long *get_el_change_mem_ptr(void) 484f6ad66aSAchin Gupta { 494f6ad66aSAchin Gupta return (unsigned long *) bl2_el_change_mem_ptr; 504f6ad66aSAchin Gupta } 514f6ad66aSAchin Gupta 524f6ad66aSAchin Gupta unsigned long page_align(unsigned long value, unsigned dir) 534f6ad66aSAchin Gupta { 544f6ad66aSAchin Gupta unsigned long page_size = 1 << FOUR_KB_SHIFT; 554f6ad66aSAchin Gupta 564f6ad66aSAchin Gupta /* Round up the limit to the next page boundary */ 574f6ad66aSAchin Gupta if (value & (page_size - 1)) { 584f6ad66aSAchin Gupta value &= ~(page_size - 1); 594f6ad66aSAchin Gupta if (dir == UP) 604f6ad66aSAchin Gupta value += page_size; 614f6ad66aSAchin Gupta } 624f6ad66aSAchin Gupta 634f6ad66aSAchin Gupta return value; 644f6ad66aSAchin Gupta } 654f6ad66aSAchin Gupta 664f6ad66aSAchin Gupta static inline unsigned int is_page_aligned (unsigned long addr) { 674f6ad66aSAchin Gupta const unsigned long page_size = 1 << FOUR_KB_SHIFT; 684f6ad66aSAchin Gupta 694f6ad66aSAchin Gupta return (addr & (page_size - 1)) == 0; 704f6ad66aSAchin Gupta } 714f6ad66aSAchin Gupta 724f6ad66aSAchin Gupta void change_security_state(unsigned int target_security_state) 734f6ad66aSAchin Gupta { 744f6ad66aSAchin Gupta unsigned long scr = read_scr(); 754f6ad66aSAchin Gupta 764f6ad66aSAchin Gupta if (target_security_state == SECURE) 774f6ad66aSAchin Gupta scr &= ~SCR_NS_BIT; 784f6ad66aSAchin Gupta else if (target_security_state == NON_SECURE) 794f6ad66aSAchin Gupta scr |= SCR_NS_BIT; 804f6ad66aSAchin Gupta else 814f6ad66aSAchin Gupta assert(0); 824f6ad66aSAchin Gupta 834f6ad66aSAchin Gupta write_scr(scr); 844f6ad66aSAchin Gupta } 854f6ad66aSAchin Gupta 864f6ad66aSAchin Gupta int drop_el(aapcs64_params *args, 874f6ad66aSAchin Gupta unsigned long spsr, 884f6ad66aSAchin Gupta unsigned long entrypoint) 894f6ad66aSAchin Gupta { 904f6ad66aSAchin Gupta write_spsr(spsr); 914f6ad66aSAchin Gupta write_elr(entrypoint); 924f6ad66aSAchin Gupta eret(args->arg0, 934f6ad66aSAchin Gupta args->arg1, 944f6ad66aSAchin Gupta args->arg2, 954f6ad66aSAchin Gupta args->arg3, 964f6ad66aSAchin Gupta args->arg4, 974f6ad66aSAchin Gupta args->arg5, 984f6ad66aSAchin Gupta args->arg6, 994f6ad66aSAchin Gupta args->arg7); 1004f6ad66aSAchin Gupta return -EINVAL; 1014f6ad66aSAchin Gupta } 1024f6ad66aSAchin Gupta 1034f6ad66aSAchin Gupta long raise_el(aapcs64_params *args) 1044f6ad66aSAchin Gupta { 1054f6ad66aSAchin Gupta return smc(args->arg0, 1064f6ad66aSAchin Gupta args->arg1, 1074f6ad66aSAchin Gupta args->arg2, 1084f6ad66aSAchin Gupta args->arg3, 1094f6ad66aSAchin Gupta args->arg4, 1104f6ad66aSAchin Gupta args->arg5, 1114f6ad66aSAchin Gupta args->arg6, 1124f6ad66aSAchin Gupta args->arg7); 1134f6ad66aSAchin Gupta } 1144f6ad66aSAchin Gupta 1154f6ad66aSAchin Gupta /* 1164f6ad66aSAchin Gupta * TODO: If we are not EL3 then currently we only issue an SMC. 1174f6ad66aSAchin Gupta * Add support for dropping into EL0 etc. Consider adding support 1184f6ad66aSAchin Gupta * for switching from S-EL1 to S-EL0/1 etc. 1194f6ad66aSAchin Gupta */ 1204f6ad66aSAchin Gupta long change_el(el_change_info *info) 1214f6ad66aSAchin Gupta { 1224f6ad66aSAchin Gupta unsigned long current_el = read_current_el(); 1234f6ad66aSAchin Gupta 1244f6ad66aSAchin Gupta if (GET_EL(current_el) == MODE_EL3) { 1254f6ad66aSAchin Gupta /* 1264f6ad66aSAchin Gupta * We can go anywhere from EL3. So find where. 1274f6ad66aSAchin Gupta * TODO: Lots to do if we are going non-secure. 1284f6ad66aSAchin Gupta * Flip the NS bit. Restore NS registers etc. 1294f6ad66aSAchin Gupta * Just doing the bare minimal for now. 1304f6ad66aSAchin Gupta */ 1314f6ad66aSAchin Gupta 1324f6ad66aSAchin Gupta if (info->security_state == NON_SECURE) 1334f6ad66aSAchin Gupta change_security_state(info->security_state); 1344f6ad66aSAchin Gupta 1354f6ad66aSAchin Gupta return drop_el(&info->args, info->spsr, info->entrypoint); 1364f6ad66aSAchin Gupta } else 1374f6ad66aSAchin Gupta return raise_el(&info->args); 1384f6ad66aSAchin Gupta } 1394f6ad66aSAchin Gupta 1404f6ad66aSAchin Gupta /* TODO: add a parameter for DAIF. not needed right now */ 1414f6ad66aSAchin Gupta unsigned long make_spsr(unsigned long target_el, 1424f6ad66aSAchin Gupta unsigned long target_sp, 1434f6ad66aSAchin Gupta unsigned long target_rw) 1444f6ad66aSAchin Gupta { 1454f6ad66aSAchin Gupta unsigned long spsr; 1464f6ad66aSAchin Gupta 1474f6ad66aSAchin Gupta /* Disable all exceptions & setup the EL */ 1484f6ad66aSAchin Gupta spsr = (DAIF_FIQ_BIT | DAIF_IRQ_BIT | DAIF_ABT_BIT | DAIF_DBG_BIT) 1494f6ad66aSAchin Gupta << PSR_DAIF_SHIFT; 1504f6ad66aSAchin Gupta spsr |= PSR_MODE(target_rw, target_el, target_sp); 1514f6ad66aSAchin Gupta 1524f6ad66aSAchin Gupta return spsr; 1534f6ad66aSAchin Gupta } 1544f6ad66aSAchin Gupta 1554f6ad66aSAchin Gupta /******************************************************************************* 1564f6ad66aSAchin Gupta * The next two functions are the weak definitions. Platform specific 1574f6ad66aSAchin Gupta * code can override them if it wishes to. 1584f6ad66aSAchin Gupta ******************************************************************************/ 1594f6ad66aSAchin Gupta 1604f6ad66aSAchin Gupta /******************************************************************************* 1614f6ad66aSAchin Gupta * Function that takes a memory layout into which BL31 has been either top or 1624f6ad66aSAchin Gupta * bottom loaded. Using this information, it populates bl31_mem_layout to tell 1634f6ad66aSAchin Gupta * BL31 how much memory it has access to and how much is available for use. It 1644f6ad66aSAchin Gupta * does not need the address where BL31 has been loaded as BL31 will reclaim 1654f6ad66aSAchin Gupta * all the memory used by BL2. 1664f6ad66aSAchin Gupta * TODO: Revisit if this and init_bl2_mem_layout can be replaced by a single 1674f6ad66aSAchin Gupta * routine. 1684f6ad66aSAchin Gupta ******************************************************************************/ 1694f6ad66aSAchin Gupta void init_bl31_mem_layout(const meminfo *bl2_mem_layout, 1704f6ad66aSAchin Gupta meminfo *bl31_mem_layout, 1714f6ad66aSAchin Gupta unsigned int load_type) 1724f6ad66aSAchin Gupta { 1734f6ad66aSAchin Gupta if (load_type == BOT_LOAD) { 1744f6ad66aSAchin Gupta /* 1754f6ad66aSAchin Gupta * ------------ ^ 1764f6ad66aSAchin Gupta * | BL2 | | 1774f6ad66aSAchin Gupta * |----------| ^ | BL2 1784f6ad66aSAchin Gupta * | | | BL2 free | total 1794f6ad66aSAchin Gupta * | | | size | size 1804f6ad66aSAchin Gupta * |----------| BL2 free base v | 1814f6ad66aSAchin Gupta * | BL31 | | 1824f6ad66aSAchin Gupta * ------------ BL2 total base v 1834f6ad66aSAchin Gupta */ 1844f6ad66aSAchin Gupta unsigned long bl31_size; 1854f6ad66aSAchin Gupta 1864f6ad66aSAchin Gupta bl31_mem_layout->free_base = bl2_mem_layout->free_base; 1874f6ad66aSAchin Gupta 1884f6ad66aSAchin Gupta bl31_size = bl2_mem_layout->free_base - bl2_mem_layout->total_base; 1894f6ad66aSAchin Gupta bl31_mem_layout->free_size = bl2_mem_layout->total_size - bl31_size; 1904f6ad66aSAchin Gupta } else { 1914f6ad66aSAchin Gupta /* 1924f6ad66aSAchin Gupta * ------------ ^ 1934f6ad66aSAchin Gupta * | BL31 | | 1944f6ad66aSAchin Gupta * |----------| ^ | BL2 1954f6ad66aSAchin Gupta * | | | BL2 free | total 1964f6ad66aSAchin Gupta * | | | size | size 1974f6ad66aSAchin Gupta * |----------| BL2 free base v | 1984f6ad66aSAchin Gupta * | BL2 | | 1994f6ad66aSAchin Gupta * ------------ BL2 total base v 2004f6ad66aSAchin Gupta */ 2014f6ad66aSAchin Gupta unsigned long bl2_size; 2024f6ad66aSAchin Gupta 2034f6ad66aSAchin Gupta bl31_mem_layout->free_base = bl2_mem_layout->total_base; 2044f6ad66aSAchin Gupta 2054f6ad66aSAchin Gupta bl2_size = bl2_mem_layout->free_base - bl2_mem_layout->total_base; 2064f6ad66aSAchin Gupta bl31_mem_layout->free_size = bl2_mem_layout->free_size + bl2_size; 2074f6ad66aSAchin Gupta } 2084f6ad66aSAchin Gupta 2094f6ad66aSAchin Gupta bl31_mem_layout->total_base = bl2_mem_layout->total_base; 2104f6ad66aSAchin Gupta bl31_mem_layout->total_size = bl2_mem_layout->total_size; 2114f6ad66aSAchin Gupta bl31_mem_layout->attr = load_type; 2124f6ad66aSAchin Gupta 2134f6ad66aSAchin Gupta flush_dcache_range((unsigned long) bl31_mem_layout, sizeof(meminfo)); 2144f6ad66aSAchin Gupta return; 2154f6ad66aSAchin Gupta } 2164f6ad66aSAchin Gupta 2174f6ad66aSAchin Gupta /******************************************************************************* 2184f6ad66aSAchin Gupta * Function that takes a memory layout into which BL2 has been either top or 2194f6ad66aSAchin Gupta * bottom loaded along with the address where BL2 has been loaded in it. Using 2204f6ad66aSAchin Gupta * this information, it populates bl2_mem_layout to tell BL2 how much memory 2214f6ad66aSAchin Gupta * it has access to and how much is available for use. 2224f6ad66aSAchin Gupta ******************************************************************************/ 2234f6ad66aSAchin Gupta void init_bl2_mem_layout(meminfo *bl1_mem_layout, 2244f6ad66aSAchin Gupta meminfo *bl2_mem_layout, 2254f6ad66aSAchin Gupta unsigned int load_type, 2264f6ad66aSAchin Gupta unsigned long bl2_base) 2274f6ad66aSAchin Gupta { 2284f6ad66aSAchin Gupta unsigned tmp; 2294f6ad66aSAchin Gupta 2304f6ad66aSAchin Gupta if (load_type == BOT_LOAD) { 2314f6ad66aSAchin Gupta bl2_mem_layout->total_base = bl2_base; 2324f6ad66aSAchin Gupta tmp = bl1_mem_layout->free_base - bl2_base; 2334f6ad66aSAchin Gupta bl2_mem_layout->total_size = bl1_mem_layout->free_size + tmp; 2344f6ad66aSAchin Gupta 2354f6ad66aSAchin Gupta } else { 2364f6ad66aSAchin Gupta bl2_mem_layout->total_base = bl1_mem_layout->free_base; 2374f6ad66aSAchin Gupta tmp = bl1_mem_layout->total_base + bl1_mem_layout->total_size; 2384f6ad66aSAchin Gupta bl2_mem_layout->total_size = tmp - bl1_mem_layout->free_base; 2394f6ad66aSAchin Gupta } 2404f6ad66aSAchin Gupta 2414f6ad66aSAchin Gupta bl2_mem_layout->free_base = bl1_mem_layout->free_base; 2424f6ad66aSAchin Gupta bl2_mem_layout->free_size = bl1_mem_layout->free_size; 2434f6ad66aSAchin Gupta bl2_mem_layout->attr = load_type; 2444f6ad66aSAchin Gupta 2454f6ad66aSAchin Gupta flush_dcache_range((unsigned long) bl2_mem_layout, sizeof(meminfo)); 2464f6ad66aSAchin Gupta return; 2474f6ad66aSAchin Gupta } 2484f6ad66aSAchin Gupta 2494f6ad66aSAchin Gupta static void dump_load_info(unsigned long image_load_addr, 2504f6ad66aSAchin Gupta unsigned long image_size, 2514f6ad66aSAchin Gupta const meminfo *mem_layout) 2524f6ad66aSAchin Gupta { 2534f6ad66aSAchin Gupta #if DEBUG 2544f6ad66aSAchin Gupta printf("Trying to load image at address 0x%lx, size = 0x%lx\r\n", 2554f6ad66aSAchin Gupta image_load_addr, image_size); 2564f6ad66aSAchin Gupta printf("Current memory layout:\r\n"); 2574f6ad66aSAchin Gupta printf(" total region = [0x%lx, 0x%lx]\r\n", mem_layout->total_base, 2584f6ad66aSAchin Gupta mem_layout->total_base + mem_layout->total_size); 2594f6ad66aSAchin Gupta printf(" free region = [0x%lx, 0x%lx]\r\n", mem_layout->free_base, 2604f6ad66aSAchin Gupta mem_layout->free_base + mem_layout->free_size); 2614f6ad66aSAchin Gupta #endif 2624f6ad66aSAchin Gupta } 2634f6ad66aSAchin Gupta 2644f6ad66aSAchin Gupta /******************************************************************************* 2654f6ad66aSAchin Gupta * Generic function to load an image into the trusted RAM using semihosting 2664f6ad66aSAchin Gupta * given a name, extents of free memory & whether the image should be loaded at 2674f6ad66aSAchin Gupta * the bottom or top of the free memory. It updates the memory layout if the 2684f6ad66aSAchin Gupta * load is successful. 2694f6ad66aSAchin Gupta ******************************************************************************/ 2704f6ad66aSAchin Gupta unsigned long load_image(meminfo *mem_layout, 2714f6ad66aSAchin Gupta const char *image_name, 2724f6ad66aSAchin Gupta unsigned int load_type, 2734f6ad66aSAchin Gupta unsigned long fixed_addr) 2744f6ad66aSAchin Gupta { 2754f6ad66aSAchin Gupta unsigned long temp_image_base, image_base; 2764f6ad66aSAchin Gupta long offset; 2774f6ad66aSAchin Gupta int image_flen; 2784f6ad66aSAchin Gupta 2794f6ad66aSAchin Gupta /* Find the size of the image */ 2804f6ad66aSAchin Gupta image_flen = semihosting_get_flen(image_name); 2814f6ad66aSAchin Gupta if (image_flen < 0) { 2824f6ad66aSAchin Gupta printf("ERROR: Cannot access '%s' file (%i).\r\n", 2834f6ad66aSAchin Gupta image_name, image_flen); 2844f6ad66aSAchin Gupta return 0; 2854f6ad66aSAchin Gupta } 2864f6ad66aSAchin Gupta 2874f6ad66aSAchin Gupta /* See if we have enough space */ 2884f6ad66aSAchin Gupta if (image_flen > mem_layout->free_size) { 2894f6ad66aSAchin Gupta printf("ERROR: Cannot load '%s' file: Not enough space.\r\n", 2904f6ad66aSAchin Gupta image_name); 2914f6ad66aSAchin Gupta dump_load_info(0, image_flen, mem_layout); 2924f6ad66aSAchin Gupta return 0; 2934f6ad66aSAchin Gupta } 2944f6ad66aSAchin Gupta 2954f6ad66aSAchin Gupta switch (load_type) { 2964f6ad66aSAchin Gupta 2974f6ad66aSAchin Gupta case TOP_LOAD: 2984f6ad66aSAchin Gupta 2994f6ad66aSAchin Gupta /* Load the image in the top of free memory */ 3004f6ad66aSAchin Gupta temp_image_base = mem_layout->free_base + mem_layout->free_size; 3014f6ad66aSAchin Gupta temp_image_base -= image_flen; 3024f6ad66aSAchin Gupta 3034f6ad66aSAchin Gupta /* Page align base address and check whether the image still fits */ 3044f6ad66aSAchin Gupta image_base = page_align(temp_image_base, DOWN); 3054f6ad66aSAchin Gupta assert(image_base <= temp_image_base); 3064f6ad66aSAchin Gupta 3074f6ad66aSAchin Gupta if (image_base < mem_layout->free_base) { 3084f6ad66aSAchin Gupta printf("ERROR: Cannot load '%s' file: Not enough space.\r\n", 3094f6ad66aSAchin Gupta image_name); 3104f6ad66aSAchin Gupta dump_load_info(image_base, image_flen, mem_layout); 3114f6ad66aSAchin Gupta return 0; 3124f6ad66aSAchin Gupta } 3134f6ad66aSAchin Gupta 3144f6ad66aSAchin Gupta /* Calculate the amount of extra memory used due to alignment */ 3154f6ad66aSAchin Gupta offset = temp_image_base - image_base; 3164f6ad66aSAchin Gupta 3174f6ad66aSAchin Gupta break; 3184f6ad66aSAchin Gupta 3194f6ad66aSAchin Gupta case BOT_LOAD: 3204f6ad66aSAchin Gupta 3214f6ad66aSAchin Gupta /* Load the BL2 image in the bottom of free memory */ 3224f6ad66aSAchin Gupta temp_image_base = mem_layout->free_base; 3234f6ad66aSAchin Gupta image_base = page_align(temp_image_base, UP); 3244f6ad66aSAchin Gupta assert(image_base >= temp_image_base); 3254f6ad66aSAchin Gupta 3264f6ad66aSAchin Gupta /* Page align base address and check whether the image still fits */ 3274f6ad66aSAchin Gupta if (image_base + image_flen > 3284f6ad66aSAchin Gupta mem_layout->free_base + mem_layout->free_size) { 3294f6ad66aSAchin Gupta printf("ERROR: Cannot load '%s' file: Not enough space.\r\n", 3304f6ad66aSAchin Gupta image_name); 3314f6ad66aSAchin Gupta dump_load_info(image_base, image_flen, mem_layout); 3324f6ad66aSAchin Gupta return 0; 3334f6ad66aSAchin Gupta } 3344f6ad66aSAchin Gupta 3354f6ad66aSAchin Gupta /* Calculate the amount of extra memory used due to alignment */ 3364f6ad66aSAchin Gupta offset = image_base - temp_image_base; 3374f6ad66aSAchin Gupta 3384f6ad66aSAchin Gupta break; 3394f6ad66aSAchin Gupta 3404f6ad66aSAchin Gupta default: 3414f6ad66aSAchin Gupta assert(0); 3424f6ad66aSAchin Gupta 3434f6ad66aSAchin Gupta } 3444f6ad66aSAchin Gupta 3454f6ad66aSAchin Gupta /* 3464f6ad66aSAchin Gupta * Some images must be loaded at a fixed address, not a dynamic one. 3474f6ad66aSAchin Gupta * 3484f6ad66aSAchin Gupta * This has been implemented as a hack on top of the existing dynamic 3494f6ad66aSAchin Gupta * loading mechanism, for the time being. If the 'fixed_addr' function 3504f6ad66aSAchin Gupta * argument is different from zero, then it will force the load address. 3514f6ad66aSAchin Gupta * So we still have this principle of top/bottom loading but the code 3524f6ad66aSAchin Gupta * determining the load address is bypassed and the load address is 3534f6ad66aSAchin Gupta * forced to the fixed one. 3544f6ad66aSAchin Gupta * 3554f6ad66aSAchin Gupta * This can result in quite a lot of wasted space because we still use 3564f6ad66aSAchin Gupta * 1 sole meminfo structure to represent the extents of free memory, 3574f6ad66aSAchin Gupta * where we should use some sort of linked list. 3584f6ad66aSAchin Gupta * 3594f6ad66aSAchin Gupta * E.g. we want to load BL2 at address 0x04020000, the resulting memory 3604f6ad66aSAchin Gupta * layout should look as follows: 3614f6ad66aSAchin Gupta * ------------ 0x04040000 3624f6ad66aSAchin Gupta * | | <- Free space (1) 3634f6ad66aSAchin Gupta * |----------| 3644f6ad66aSAchin Gupta * | BL2 | 3654f6ad66aSAchin Gupta * |----------| 0x04020000 3664f6ad66aSAchin Gupta * | | <- Free space (2) 3674f6ad66aSAchin Gupta * |----------| 3684f6ad66aSAchin Gupta * | BL1 | 3694f6ad66aSAchin Gupta * ------------ 0x04000000 3704f6ad66aSAchin Gupta * 3714f6ad66aSAchin Gupta * But in the current hacky implementation, we'll need to specify 3724f6ad66aSAchin Gupta * whether BL2 is loaded at the top or bottom of the free memory. 3734f6ad66aSAchin Gupta * E.g. if BL2 is considered as top-loaded, the meminfo structure 3744f6ad66aSAchin Gupta * will give the following view of the memory, hiding the chunk of 3754f6ad66aSAchin Gupta * free memory above BL2: 3764f6ad66aSAchin Gupta * ------------ 0x04040000 3774f6ad66aSAchin Gupta * | | 3784f6ad66aSAchin Gupta * | | 3794f6ad66aSAchin Gupta * | BL2 | 3804f6ad66aSAchin Gupta * |----------| 0x04020000 3814f6ad66aSAchin Gupta * | | <- Free space (2) 3824f6ad66aSAchin Gupta * |----------| 3834f6ad66aSAchin Gupta * | BL1 | 3844f6ad66aSAchin Gupta * ------------ 0x04000000 3854f6ad66aSAchin Gupta */ 3864f6ad66aSAchin Gupta if (fixed_addr != 0) { 3874f6ad66aSAchin Gupta /* Load the image at the given address. */ 3884f6ad66aSAchin Gupta image_base = fixed_addr; 3894f6ad66aSAchin Gupta 3904f6ad66aSAchin Gupta /* Check whether the image fits. */ 3914f6ad66aSAchin Gupta if ((image_base < mem_layout->free_base) || 3924f6ad66aSAchin Gupta (image_base + image_flen > 3934f6ad66aSAchin Gupta mem_layout->free_base + mem_layout->free_size)) { 3944f6ad66aSAchin Gupta printf("ERROR: Cannot load '%s' file: Not enough space.\r\n", 3954f6ad66aSAchin Gupta image_name); 3964f6ad66aSAchin Gupta dump_load_info(image_base, image_flen, mem_layout); 3974f6ad66aSAchin Gupta return 0; 3984f6ad66aSAchin Gupta } 3994f6ad66aSAchin Gupta 4004f6ad66aSAchin Gupta /* Check whether the fixed load address is page-aligned. */ 4014f6ad66aSAchin Gupta if (!is_page_aligned(image_base)) { 4024f6ad66aSAchin Gupta printf("ERROR: Cannot load '%s' file at unaligned address 0x%lx.\r\n", 4034f6ad66aSAchin Gupta image_name, fixed_addr); 4044f6ad66aSAchin Gupta return 0; 4054f6ad66aSAchin Gupta } 4064f6ad66aSAchin Gupta 4074f6ad66aSAchin Gupta /* 4084f6ad66aSAchin Gupta * Calculate the amount of extra memory used due to fixed 4094f6ad66aSAchin Gupta * loading. 4104f6ad66aSAchin Gupta */ 4114f6ad66aSAchin Gupta if (load_type == TOP_LOAD) { 4124f6ad66aSAchin Gupta unsigned long max_addr, space_used; 4134f6ad66aSAchin Gupta /* 4144f6ad66aSAchin Gupta * ------------ max_addr 4154f6ad66aSAchin Gupta * | /wasted/ | | offset 4164f6ad66aSAchin Gupta * |..........|.............................. 4174f6ad66aSAchin Gupta * | image | | image_flen 4184f6ad66aSAchin Gupta * |----------| fixed_addr 4194f6ad66aSAchin Gupta * | | 4204f6ad66aSAchin Gupta * | | 4214f6ad66aSAchin Gupta * ------------ total_base 4224f6ad66aSAchin Gupta */ 4234f6ad66aSAchin Gupta max_addr = mem_layout->total_base + mem_layout->total_size; 4244f6ad66aSAchin Gupta /* 4254f6ad66aSAchin Gupta * Compute the amount of memory used by the image. 4264f6ad66aSAchin Gupta * Corresponds to all space above the image load 4274f6ad66aSAchin Gupta * address. 4284f6ad66aSAchin Gupta */ 4294f6ad66aSAchin Gupta space_used = max_addr - fixed_addr; 4304f6ad66aSAchin Gupta /* 4314f6ad66aSAchin Gupta * Calculate the amount of wasted memory within the 4324f6ad66aSAchin Gupta * amount of memory used by the image. 4334f6ad66aSAchin Gupta */ 4344f6ad66aSAchin Gupta offset = space_used - image_flen; 4354f6ad66aSAchin Gupta } else /* BOT_LOAD */ 4364f6ad66aSAchin Gupta /* 4374f6ad66aSAchin Gupta * ------------ 4384f6ad66aSAchin Gupta * | | 4394f6ad66aSAchin Gupta * | | 4404f6ad66aSAchin Gupta * |----------| 4414f6ad66aSAchin Gupta * | image | 4424f6ad66aSAchin Gupta * |..........| fixed_addr 4434f6ad66aSAchin Gupta * | /wasted/ | | offset 4444f6ad66aSAchin Gupta * ------------ total_base 4454f6ad66aSAchin Gupta */ 4464f6ad66aSAchin Gupta offset = fixed_addr - mem_layout->total_base; 4474f6ad66aSAchin Gupta } 4484f6ad66aSAchin Gupta 4494f6ad66aSAchin Gupta /* We have enough space so load the image now */ 4504f6ad66aSAchin Gupta image_flen = semihosting_download_file(image_name, 4514f6ad66aSAchin Gupta image_flen, 4524f6ad66aSAchin Gupta (void *) image_base); 4534f6ad66aSAchin Gupta if (image_flen <= 0) { 4544f6ad66aSAchin Gupta printf("ERROR: Failed to load '%s' file from semihosting (%i).\r\n", 4554f6ad66aSAchin Gupta image_name, image_flen); 4564f6ad66aSAchin Gupta return 0; 4574f6ad66aSAchin Gupta } 4584f6ad66aSAchin Gupta 4594f6ad66aSAchin Gupta /* 4604f6ad66aSAchin Gupta * File has been successfully loaded. Update the free memory 4614f6ad66aSAchin Gupta * data structure & flush the contents of the TZRAM so that 4624f6ad66aSAchin Gupta * the next EL can see it. 4634f6ad66aSAchin Gupta */ 4644f6ad66aSAchin Gupta /* Update the memory contents */ 4654f6ad66aSAchin Gupta flush_dcache_range(image_base, image_flen); 4664f6ad66aSAchin Gupta 4674f6ad66aSAchin Gupta mem_layout->free_size -= image_flen + offset; 4684f6ad66aSAchin Gupta 4694f6ad66aSAchin Gupta /* Update the base of free memory since its moved up */ 4704f6ad66aSAchin Gupta if (load_type == BOT_LOAD) 4714f6ad66aSAchin Gupta mem_layout->free_base += offset + image_flen; 4724f6ad66aSAchin Gupta 4734f6ad66aSAchin Gupta return image_base; 4744f6ad66aSAchin Gupta } 4754f6ad66aSAchin Gupta 4764f6ad66aSAchin Gupta /******************************************************************************* 4774f6ad66aSAchin Gupta * Run a loaded image from the given entry point. This could result in either 4784f6ad66aSAchin Gupta * dropping into a lower exception level or jumping to a higher exception level. 4794f6ad66aSAchin Gupta * The only way of doing the latter is through an SMC. In either case, setup the 4804f6ad66aSAchin Gupta * parameters for the EL change request correctly. 4814f6ad66aSAchin Gupta ******************************************************************************/ 4824f6ad66aSAchin Gupta int run_image(unsigned long entrypoint, 4834f6ad66aSAchin Gupta unsigned long spsr, 4844f6ad66aSAchin Gupta unsigned long target_security_state, 4854f6ad66aSAchin Gupta meminfo *mem_layout, 4864f6ad66aSAchin Gupta void *data) 4874f6ad66aSAchin Gupta { 4884f6ad66aSAchin Gupta el_change_info run_image_info; 4894f6ad66aSAchin Gupta unsigned long current_el = read_current_el(); 4904f6ad66aSAchin Gupta 4914f6ad66aSAchin Gupta /* Tell next EL what we want done */ 4924f6ad66aSAchin Gupta run_image_info.args.arg0 = RUN_IMAGE; 4934f6ad66aSAchin Gupta run_image_info.entrypoint = entrypoint; 4944f6ad66aSAchin Gupta run_image_info.spsr = spsr; 4954f6ad66aSAchin Gupta run_image_info.security_state = target_security_state; 4964f6ad66aSAchin Gupta run_image_info.next = 0; 4974f6ad66aSAchin Gupta 4984f6ad66aSAchin Gupta /* 4994f6ad66aSAchin Gupta * If we are EL3 then only an eret can take us to the desired 5004f6ad66aSAchin Gupta * exception level. Else for the time being assume that we have 5014f6ad66aSAchin Gupta * to jump to a higher EL and issue an SMC. Contents of argY 5024f6ad66aSAchin Gupta * will go into the general purpose register xY e.g. arg0->x0 5034f6ad66aSAchin Gupta */ 5044f6ad66aSAchin Gupta if (GET_EL(current_el) == MODE_EL3) { 5054f6ad66aSAchin Gupta run_image_info.args.arg1 = (unsigned long) mem_layout; 5064f6ad66aSAchin Gupta run_image_info.args.arg2 = (unsigned long) data; 5074f6ad66aSAchin Gupta } else { 5084f6ad66aSAchin Gupta run_image_info.args.arg1 = entrypoint; 5094f6ad66aSAchin Gupta run_image_info.args.arg2 = spsr; 5104f6ad66aSAchin Gupta run_image_info.args.arg3 = (unsigned long) mem_layout; 5114f6ad66aSAchin Gupta run_image_info.args.arg4 = (unsigned long) data; 5124f6ad66aSAchin Gupta } 5134f6ad66aSAchin Gupta 5144f6ad66aSAchin Gupta return change_el(&run_image_info); 5154f6ad66aSAchin Gupta } 516