14f6ad66aSAchin Gupta /* 2e83b0cadSDan Handley * Copyright (c) 2013-2014, ARM Limited and Contributors. All rights reserved. 34f6ad66aSAchin Gupta * 44f6ad66aSAchin Gupta * Redistribution and use in source and binary forms, with or without 54f6ad66aSAchin Gupta * modification, are permitted provided that the following conditions are met: 64f6ad66aSAchin Gupta * 74f6ad66aSAchin Gupta * Redistributions of source code must retain the above copyright notice, this 84f6ad66aSAchin Gupta * list of conditions and the following disclaimer. 94f6ad66aSAchin Gupta * 104f6ad66aSAchin Gupta * Redistributions in binary form must reproduce the above copyright notice, 114f6ad66aSAchin Gupta * this list of conditions and the following disclaimer in the documentation 124f6ad66aSAchin Gupta * and/or other materials provided with the distribution. 134f6ad66aSAchin Gupta * 144f6ad66aSAchin Gupta * Neither the name of ARM nor the names of its contributors may be used 154f6ad66aSAchin Gupta * to endorse or promote products derived from this software without specific 164f6ad66aSAchin Gupta * prior written permission. 174f6ad66aSAchin Gupta * 184f6ad66aSAchin Gupta * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 194f6ad66aSAchin Gupta * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 204f6ad66aSAchin Gupta * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 214f6ad66aSAchin Gupta * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE 224f6ad66aSAchin Gupta * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 234f6ad66aSAchin Gupta * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 244f6ad66aSAchin Gupta * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 254f6ad66aSAchin Gupta * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 264f6ad66aSAchin Gupta * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 274f6ad66aSAchin Gupta * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 284f6ad66aSAchin Gupta * POSSIBILITY OF SUCH DAMAGE. 294f6ad66aSAchin Gupta */ 304f6ad66aSAchin Gupta 314f6ad66aSAchin Gupta #include <stdio.h> 324f6ad66aSAchin Gupta #include <string.h> 334f6ad66aSAchin Gupta #include <errno.h> 344f6ad66aSAchin Gupta #include <assert.h> 354f6ad66aSAchin Gupta #include <arch_helpers.h> 364f6ad66aSAchin Gupta #include <console.h> 374f6ad66aSAchin Gupta #include <platform.h> 384f6ad66aSAchin Gupta #include <semihosting.h> 394f6ad66aSAchin Gupta #include <bl_common.h> 40*9d72b4eaSJames Morrissey #include "io_storage.h" 41*9d72b4eaSJames Morrissey #include "debug.h" 424f6ad66aSAchin Gupta 434f6ad66aSAchin Gupta /*********************************************************** 444f6ad66aSAchin Gupta * Memory for sharing data while changing exception levels. 454f6ad66aSAchin Gupta * Only used by the primary core. 464f6ad66aSAchin Gupta **********************************************************/ 474f6ad66aSAchin Gupta unsigned char bl2_el_change_mem_ptr[EL_CHANGE_MEM_SIZE]; 484f6ad66aSAchin Gupta 494f6ad66aSAchin Gupta unsigned long *get_el_change_mem_ptr(void) 504f6ad66aSAchin Gupta { 514f6ad66aSAchin Gupta return (unsigned long *) bl2_el_change_mem_ptr; 524f6ad66aSAchin Gupta } 534f6ad66aSAchin Gupta 544f6ad66aSAchin Gupta unsigned long page_align(unsigned long value, unsigned dir) 554f6ad66aSAchin Gupta { 564f6ad66aSAchin Gupta unsigned long page_size = 1 << FOUR_KB_SHIFT; 574f6ad66aSAchin Gupta 584f6ad66aSAchin Gupta /* Round up the limit to the next page boundary */ 594f6ad66aSAchin Gupta if (value & (page_size - 1)) { 604f6ad66aSAchin Gupta value &= ~(page_size - 1); 614f6ad66aSAchin Gupta if (dir == UP) 624f6ad66aSAchin Gupta value += page_size; 634f6ad66aSAchin Gupta } 644f6ad66aSAchin Gupta 654f6ad66aSAchin Gupta return value; 664f6ad66aSAchin Gupta } 674f6ad66aSAchin Gupta 684f6ad66aSAchin Gupta static inline unsigned int is_page_aligned (unsigned long addr) { 694f6ad66aSAchin Gupta const unsigned long page_size = 1 << FOUR_KB_SHIFT; 704f6ad66aSAchin Gupta 714f6ad66aSAchin Gupta return (addr & (page_size - 1)) == 0; 724f6ad66aSAchin Gupta } 734f6ad66aSAchin Gupta 744f6ad66aSAchin Gupta void change_security_state(unsigned int target_security_state) 754f6ad66aSAchin Gupta { 764f6ad66aSAchin Gupta unsigned long scr = read_scr(); 774f6ad66aSAchin Gupta 784f6ad66aSAchin Gupta if (target_security_state == SECURE) 794f6ad66aSAchin Gupta scr &= ~SCR_NS_BIT; 804f6ad66aSAchin Gupta else if (target_security_state == NON_SECURE) 814f6ad66aSAchin Gupta scr |= SCR_NS_BIT; 824f6ad66aSAchin Gupta else 834f6ad66aSAchin Gupta assert(0); 844f6ad66aSAchin Gupta 854f6ad66aSAchin Gupta write_scr(scr); 864f6ad66aSAchin Gupta } 874f6ad66aSAchin Gupta 884f6ad66aSAchin Gupta int drop_el(aapcs64_params *args, 894f6ad66aSAchin Gupta unsigned long spsr, 904f6ad66aSAchin Gupta unsigned long entrypoint) 914f6ad66aSAchin Gupta { 924f6ad66aSAchin Gupta write_spsr(spsr); 934f6ad66aSAchin Gupta write_elr(entrypoint); 944f6ad66aSAchin Gupta eret(args->arg0, 954f6ad66aSAchin Gupta args->arg1, 964f6ad66aSAchin Gupta args->arg2, 974f6ad66aSAchin Gupta args->arg3, 984f6ad66aSAchin Gupta args->arg4, 994f6ad66aSAchin Gupta args->arg5, 1004f6ad66aSAchin Gupta args->arg6, 1014f6ad66aSAchin Gupta args->arg7); 1024f6ad66aSAchin Gupta return -EINVAL; 1034f6ad66aSAchin Gupta } 1044f6ad66aSAchin Gupta 1054f6ad66aSAchin Gupta long raise_el(aapcs64_params *args) 1064f6ad66aSAchin Gupta { 1074f6ad66aSAchin Gupta return smc(args->arg0, 1084f6ad66aSAchin Gupta args->arg1, 1094f6ad66aSAchin Gupta args->arg2, 1104f6ad66aSAchin Gupta args->arg3, 1114f6ad66aSAchin Gupta args->arg4, 1124f6ad66aSAchin Gupta args->arg5, 1134f6ad66aSAchin Gupta args->arg6, 1144f6ad66aSAchin Gupta args->arg7); 1154f6ad66aSAchin Gupta } 1164f6ad66aSAchin Gupta 1174f6ad66aSAchin Gupta /* 1184f6ad66aSAchin Gupta * TODO: If we are not EL3 then currently we only issue an SMC. 1194f6ad66aSAchin Gupta * Add support for dropping into EL0 etc. Consider adding support 1204f6ad66aSAchin Gupta * for switching from S-EL1 to S-EL0/1 etc. 1214f6ad66aSAchin Gupta */ 1224f6ad66aSAchin Gupta long change_el(el_change_info *info) 1234f6ad66aSAchin Gupta { 1244f6ad66aSAchin Gupta unsigned long current_el = read_current_el(); 1254f6ad66aSAchin Gupta 1264f6ad66aSAchin Gupta if (GET_EL(current_el) == MODE_EL3) { 1274f6ad66aSAchin Gupta /* 1284f6ad66aSAchin Gupta * We can go anywhere from EL3. So find where. 1294f6ad66aSAchin Gupta * TODO: Lots to do if we are going non-secure. 1304f6ad66aSAchin Gupta * Flip the NS bit. Restore NS registers etc. 1314f6ad66aSAchin Gupta * Just doing the bare minimal for now. 1324f6ad66aSAchin Gupta */ 1334f6ad66aSAchin Gupta 1344f6ad66aSAchin Gupta if (info->security_state == NON_SECURE) 1354f6ad66aSAchin Gupta change_security_state(info->security_state); 1364f6ad66aSAchin Gupta 1374f6ad66aSAchin Gupta return drop_el(&info->args, info->spsr, info->entrypoint); 1384f6ad66aSAchin Gupta } else 1394f6ad66aSAchin Gupta return raise_el(&info->args); 1404f6ad66aSAchin Gupta } 1414f6ad66aSAchin Gupta 1424f6ad66aSAchin Gupta /* TODO: add a parameter for DAIF. not needed right now */ 1434f6ad66aSAchin Gupta unsigned long make_spsr(unsigned long target_el, 1444f6ad66aSAchin Gupta unsigned long target_sp, 1454f6ad66aSAchin Gupta unsigned long target_rw) 1464f6ad66aSAchin Gupta { 1474f6ad66aSAchin Gupta unsigned long spsr; 1484f6ad66aSAchin Gupta 1494f6ad66aSAchin Gupta /* Disable all exceptions & setup the EL */ 1504f6ad66aSAchin Gupta spsr = (DAIF_FIQ_BIT | DAIF_IRQ_BIT | DAIF_ABT_BIT | DAIF_DBG_BIT) 1514f6ad66aSAchin Gupta << PSR_DAIF_SHIFT; 1524f6ad66aSAchin Gupta spsr |= PSR_MODE(target_rw, target_el, target_sp); 1534f6ad66aSAchin Gupta 1544f6ad66aSAchin Gupta return spsr; 1554f6ad66aSAchin Gupta } 1564f6ad66aSAchin Gupta 1574f6ad66aSAchin Gupta /******************************************************************************* 1584f6ad66aSAchin Gupta * The next two functions are the weak definitions. Platform specific 1594f6ad66aSAchin Gupta * code can override them if it wishes to. 1604f6ad66aSAchin Gupta ******************************************************************************/ 1614f6ad66aSAchin Gupta 1624f6ad66aSAchin Gupta /******************************************************************************* 1634f6ad66aSAchin Gupta * Function that takes a memory layout into which BL31 has been either top or 1644f6ad66aSAchin Gupta * bottom loaded. Using this information, it populates bl31_mem_layout to tell 1654f6ad66aSAchin Gupta * BL31 how much memory it has access to and how much is available for use. It 1664f6ad66aSAchin Gupta * does not need the address where BL31 has been loaded as BL31 will reclaim 1674f6ad66aSAchin Gupta * all the memory used by BL2. 1684f6ad66aSAchin Gupta * TODO: Revisit if this and init_bl2_mem_layout can be replaced by a single 1694f6ad66aSAchin Gupta * routine. 1704f6ad66aSAchin Gupta ******************************************************************************/ 1714f6ad66aSAchin Gupta void init_bl31_mem_layout(const meminfo *bl2_mem_layout, 1724f6ad66aSAchin Gupta meminfo *bl31_mem_layout, 1734f6ad66aSAchin Gupta unsigned int load_type) 1744f6ad66aSAchin Gupta { 1754f6ad66aSAchin Gupta if (load_type == BOT_LOAD) { 1764f6ad66aSAchin Gupta /* 1774f6ad66aSAchin Gupta * ------------ ^ 1784f6ad66aSAchin Gupta * | BL2 | | 1794f6ad66aSAchin Gupta * |----------| ^ | BL2 1804f6ad66aSAchin Gupta * | | | BL2 free | total 1814f6ad66aSAchin Gupta * | | | size | size 1824f6ad66aSAchin Gupta * |----------| BL2 free base v | 1834f6ad66aSAchin Gupta * | BL31 | | 1844f6ad66aSAchin Gupta * ------------ BL2 total base v 1854f6ad66aSAchin Gupta */ 1864f6ad66aSAchin Gupta unsigned long bl31_size; 1874f6ad66aSAchin Gupta 1884f6ad66aSAchin Gupta bl31_mem_layout->free_base = bl2_mem_layout->free_base; 1894f6ad66aSAchin Gupta 1904f6ad66aSAchin Gupta bl31_size = bl2_mem_layout->free_base - bl2_mem_layout->total_base; 1914f6ad66aSAchin Gupta bl31_mem_layout->free_size = bl2_mem_layout->total_size - bl31_size; 1924f6ad66aSAchin Gupta } else { 1934f6ad66aSAchin Gupta /* 1944f6ad66aSAchin Gupta * ------------ ^ 1954f6ad66aSAchin Gupta * | BL31 | | 1964f6ad66aSAchin Gupta * |----------| ^ | BL2 1974f6ad66aSAchin Gupta * | | | BL2 free | total 1984f6ad66aSAchin Gupta * | | | size | size 1994f6ad66aSAchin Gupta * |----------| BL2 free base v | 2004f6ad66aSAchin Gupta * | BL2 | | 2014f6ad66aSAchin Gupta * ------------ BL2 total base v 2024f6ad66aSAchin Gupta */ 2034f6ad66aSAchin Gupta unsigned long bl2_size; 2044f6ad66aSAchin Gupta 2054f6ad66aSAchin Gupta bl31_mem_layout->free_base = bl2_mem_layout->total_base; 2064f6ad66aSAchin Gupta 2074f6ad66aSAchin Gupta bl2_size = bl2_mem_layout->free_base - bl2_mem_layout->total_base; 2084f6ad66aSAchin Gupta bl31_mem_layout->free_size = bl2_mem_layout->free_size + bl2_size; 2094f6ad66aSAchin Gupta } 2104f6ad66aSAchin Gupta 2114f6ad66aSAchin Gupta bl31_mem_layout->total_base = bl2_mem_layout->total_base; 2124f6ad66aSAchin Gupta bl31_mem_layout->total_size = bl2_mem_layout->total_size; 2134f6ad66aSAchin Gupta bl31_mem_layout->attr = load_type; 2144f6ad66aSAchin Gupta 2154f6ad66aSAchin Gupta flush_dcache_range((unsigned long) bl31_mem_layout, sizeof(meminfo)); 2164f6ad66aSAchin Gupta return; 2174f6ad66aSAchin Gupta } 2184f6ad66aSAchin Gupta 2194f6ad66aSAchin Gupta /******************************************************************************* 2204f6ad66aSAchin Gupta * Function that takes a memory layout into which BL2 has been either top or 2214f6ad66aSAchin Gupta * bottom loaded along with the address where BL2 has been loaded in it. Using 2224f6ad66aSAchin Gupta * this information, it populates bl2_mem_layout to tell BL2 how much memory 2234f6ad66aSAchin Gupta * it has access to and how much is available for use. 2244f6ad66aSAchin Gupta ******************************************************************************/ 2254f6ad66aSAchin Gupta void init_bl2_mem_layout(meminfo *bl1_mem_layout, 2264f6ad66aSAchin Gupta meminfo *bl2_mem_layout, 2274f6ad66aSAchin Gupta unsigned int load_type, 2284f6ad66aSAchin Gupta unsigned long bl2_base) 2294f6ad66aSAchin Gupta { 2304f6ad66aSAchin Gupta unsigned tmp; 2314f6ad66aSAchin Gupta 2324f6ad66aSAchin Gupta if (load_type == BOT_LOAD) { 2334f6ad66aSAchin Gupta bl2_mem_layout->total_base = bl2_base; 2344f6ad66aSAchin Gupta tmp = bl1_mem_layout->free_base - bl2_base; 2354f6ad66aSAchin Gupta bl2_mem_layout->total_size = bl1_mem_layout->free_size + tmp; 2364f6ad66aSAchin Gupta 2374f6ad66aSAchin Gupta } else { 2384f6ad66aSAchin Gupta bl2_mem_layout->total_base = bl1_mem_layout->free_base; 2394f6ad66aSAchin Gupta tmp = bl1_mem_layout->total_base + bl1_mem_layout->total_size; 2404f6ad66aSAchin Gupta bl2_mem_layout->total_size = tmp - bl1_mem_layout->free_base; 2414f6ad66aSAchin Gupta } 2424f6ad66aSAchin Gupta 2434f6ad66aSAchin Gupta bl2_mem_layout->free_base = bl1_mem_layout->free_base; 2444f6ad66aSAchin Gupta bl2_mem_layout->free_size = bl1_mem_layout->free_size; 2454f6ad66aSAchin Gupta bl2_mem_layout->attr = load_type; 2464f6ad66aSAchin Gupta 2474f6ad66aSAchin Gupta flush_dcache_range((unsigned long) bl2_mem_layout, sizeof(meminfo)); 2484f6ad66aSAchin Gupta return; 2494f6ad66aSAchin Gupta } 2504f6ad66aSAchin Gupta 2514f6ad66aSAchin Gupta static void dump_load_info(unsigned long image_load_addr, 2524f6ad66aSAchin Gupta unsigned long image_size, 2534f6ad66aSAchin Gupta const meminfo *mem_layout) 2544f6ad66aSAchin Gupta { 2554f6ad66aSAchin Gupta #if DEBUG 2564f6ad66aSAchin Gupta printf("Trying to load image at address 0x%lx, size = 0x%lx\r\n", 2574f6ad66aSAchin Gupta image_load_addr, image_size); 2584f6ad66aSAchin Gupta printf("Current memory layout:\r\n"); 2594f6ad66aSAchin Gupta printf(" total region = [0x%lx, 0x%lx]\r\n", mem_layout->total_base, 2604f6ad66aSAchin Gupta mem_layout->total_base + mem_layout->total_size); 2614f6ad66aSAchin Gupta printf(" free region = [0x%lx, 0x%lx]\r\n", mem_layout->free_base, 2624f6ad66aSAchin Gupta mem_layout->free_base + mem_layout->free_size); 2634f6ad66aSAchin Gupta #endif 2644f6ad66aSAchin Gupta } 2654f6ad66aSAchin Gupta 2664f6ad66aSAchin Gupta /******************************************************************************* 267*9d72b4eaSJames Morrissey * Generic function to load an image into the trusted RAM, 2684f6ad66aSAchin Gupta * given a name, extents of free memory & whether the image should be loaded at 2694f6ad66aSAchin Gupta * the bottom or top of the free memory. It updates the memory layout if the 2704f6ad66aSAchin Gupta * load is successful. 2714f6ad66aSAchin Gupta ******************************************************************************/ 2724f6ad66aSAchin Gupta unsigned long load_image(meminfo *mem_layout, 2734f6ad66aSAchin Gupta const char *image_name, 2744f6ad66aSAchin Gupta unsigned int load_type, 2754f6ad66aSAchin Gupta unsigned long fixed_addr) 2764f6ad66aSAchin Gupta { 277*9d72b4eaSJames Morrissey io_dev_handle dev_handle; 278*9d72b4eaSJames Morrissey io_handle image_handle; 279*9d72b4eaSJames Morrissey void *image_spec; 28040a6f647SJames Morrissey unsigned long temp_image_base = 0; 28140a6f647SJames Morrissey unsigned long image_base = 0; 28240a6f647SJames Morrissey long offset = 0; 283*9d72b4eaSJames Morrissey size_t image_size = 0; 284*9d72b4eaSJames Morrissey size_t bytes_read = 0; 285*9d72b4eaSJames Morrissey int io_result = IO_FAIL; 2864f6ad66aSAchin Gupta 287*9d72b4eaSJames Morrissey assert(mem_layout != NULL); 288*9d72b4eaSJames Morrissey assert(image_name != NULL); 289*9d72b4eaSJames Morrissey 290*9d72b4eaSJames Morrissey /* Obtain a reference to the image by querying the platform layer */ 291*9d72b4eaSJames Morrissey io_result = plat_get_image_source(image_name, &dev_handle, &image_spec); 292*9d72b4eaSJames Morrissey if (io_result != IO_SUCCESS) { 293*9d72b4eaSJames Morrissey ERROR("Failed to obtain reference to image '%s' (%i)\n", 294*9d72b4eaSJames Morrissey image_name, io_result); 2954f6ad66aSAchin Gupta return 0; 2964f6ad66aSAchin Gupta } 2974f6ad66aSAchin Gupta 298*9d72b4eaSJames Morrissey /* Attempt to access the image */ 299*9d72b4eaSJames Morrissey io_result = io_open(dev_handle, image_spec, &image_handle); 300*9d72b4eaSJames Morrissey if (io_result != IO_SUCCESS) { 301*9d72b4eaSJames Morrissey ERROR("Failed to access image '%s' (%i)\n", 302*9d72b4eaSJames Morrissey image_name, io_result); 3034f6ad66aSAchin Gupta return 0; 3044f6ad66aSAchin Gupta } 3054f6ad66aSAchin Gupta 306*9d72b4eaSJames Morrissey /* Find the size of the image */ 307*9d72b4eaSJames Morrissey io_result = io_size(image_handle, &image_size); 308*9d72b4eaSJames Morrissey if ((io_result != IO_SUCCESS) || (image_size == 0)) { 309*9d72b4eaSJames Morrissey ERROR("Failed to determine the size of the image '%s' file (%i)\n", 310*9d72b4eaSJames Morrissey image_name, io_result); 311*9d72b4eaSJames Morrissey goto fail; 312*9d72b4eaSJames Morrissey } 313*9d72b4eaSJames Morrissey 314*9d72b4eaSJames Morrissey /* See if we have enough space */ 315*9d72b4eaSJames Morrissey if (image_size > mem_layout->free_size) { 316*9d72b4eaSJames Morrissey ERROR("ERROR: Cannot load '%s' file: Not enough space.\n", 317*9d72b4eaSJames Morrissey image_name); 318*9d72b4eaSJames Morrissey dump_load_info(0, image_size, mem_layout); 319*9d72b4eaSJames Morrissey goto fail; 320*9d72b4eaSJames Morrissey } 321*9d72b4eaSJames Morrissey 3224f6ad66aSAchin Gupta switch (load_type) { 3234f6ad66aSAchin Gupta 3244f6ad66aSAchin Gupta case TOP_LOAD: 3254f6ad66aSAchin Gupta 3264f6ad66aSAchin Gupta /* Load the image in the top of free memory */ 3274f6ad66aSAchin Gupta temp_image_base = mem_layout->free_base + mem_layout->free_size; 328*9d72b4eaSJames Morrissey temp_image_base -= image_size; 3294f6ad66aSAchin Gupta 3304f6ad66aSAchin Gupta /* Page align base address and check whether the image still fits */ 3314f6ad66aSAchin Gupta image_base = page_align(temp_image_base, DOWN); 3324f6ad66aSAchin Gupta assert(image_base <= temp_image_base); 3334f6ad66aSAchin Gupta 3344f6ad66aSAchin Gupta if (image_base < mem_layout->free_base) { 335*9d72b4eaSJames Morrissey ERROR("Cannot load '%s' file: Not enough space.\n", 3364f6ad66aSAchin Gupta image_name); 337*9d72b4eaSJames Morrissey dump_load_info(image_base, image_size, mem_layout); 338*9d72b4eaSJames Morrissey goto fail; 3394f6ad66aSAchin Gupta } 3404f6ad66aSAchin Gupta 3414f6ad66aSAchin Gupta /* Calculate the amount of extra memory used due to alignment */ 3424f6ad66aSAchin Gupta offset = temp_image_base - image_base; 3434f6ad66aSAchin Gupta 3444f6ad66aSAchin Gupta break; 3454f6ad66aSAchin Gupta 3464f6ad66aSAchin Gupta case BOT_LOAD: 3474f6ad66aSAchin Gupta 3484f6ad66aSAchin Gupta /* Load the BL2 image in the bottom of free memory */ 3494f6ad66aSAchin Gupta temp_image_base = mem_layout->free_base; 3504f6ad66aSAchin Gupta image_base = page_align(temp_image_base, UP); 3514f6ad66aSAchin Gupta assert(image_base >= temp_image_base); 3524f6ad66aSAchin Gupta 3534f6ad66aSAchin Gupta /* Page align base address and check whether the image still fits */ 354*9d72b4eaSJames Morrissey if (image_base + image_size > 3554f6ad66aSAchin Gupta mem_layout->free_base + mem_layout->free_size) { 356*9d72b4eaSJames Morrissey ERROR("Cannot load '%s' file: Not enough space.\n", 3574f6ad66aSAchin Gupta image_name); 358*9d72b4eaSJames Morrissey dump_load_info(image_base, image_size, mem_layout); 359*9d72b4eaSJames Morrissey goto fail; 3604f6ad66aSAchin Gupta } 3614f6ad66aSAchin Gupta 3624f6ad66aSAchin Gupta /* Calculate the amount of extra memory used due to alignment */ 3634f6ad66aSAchin Gupta offset = image_base - temp_image_base; 3644f6ad66aSAchin Gupta 3654f6ad66aSAchin Gupta break; 3664f6ad66aSAchin Gupta 3674f6ad66aSAchin Gupta default: 3684f6ad66aSAchin Gupta assert(0); 3694f6ad66aSAchin Gupta 3704f6ad66aSAchin Gupta } 3714f6ad66aSAchin Gupta 3724f6ad66aSAchin Gupta /* 3734f6ad66aSAchin Gupta * Some images must be loaded at a fixed address, not a dynamic one. 3744f6ad66aSAchin Gupta * 3754f6ad66aSAchin Gupta * This has been implemented as a hack on top of the existing dynamic 3764f6ad66aSAchin Gupta * loading mechanism, for the time being. If the 'fixed_addr' function 3774f6ad66aSAchin Gupta * argument is different from zero, then it will force the load address. 3784f6ad66aSAchin Gupta * So we still have this principle of top/bottom loading but the code 3794f6ad66aSAchin Gupta * determining the load address is bypassed and the load address is 3804f6ad66aSAchin Gupta * forced to the fixed one. 3814f6ad66aSAchin Gupta * 3824f6ad66aSAchin Gupta * This can result in quite a lot of wasted space because we still use 3834f6ad66aSAchin Gupta * 1 sole meminfo structure to represent the extents of free memory, 3844f6ad66aSAchin Gupta * where we should use some sort of linked list. 3854f6ad66aSAchin Gupta * 3864f6ad66aSAchin Gupta * E.g. we want to load BL2 at address 0x04020000, the resulting memory 3874f6ad66aSAchin Gupta * layout should look as follows: 3884f6ad66aSAchin Gupta * ------------ 0x04040000 3894f6ad66aSAchin Gupta * | | <- Free space (1) 3904f6ad66aSAchin Gupta * |----------| 3914f6ad66aSAchin Gupta * | BL2 | 3924f6ad66aSAchin Gupta * |----------| 0x04020000 3934f6ad66aSAchin Gupta * | | <- Free space (2) 3944f6ad66aSAchin Gupta * |----------| 3954f6ad66aSAchin Gupta * | BL1 | 3964f6ad66aSAchin Gupta * ------------ 0x04000000 3974f6ad66aSAchin Gupta * 3984f6ad66aSAchin Gupta * But in the current hacky implementation, we'll need to specify 3994f6ad66aSAchin Gupta * whether BL2 is loaded at the top or bottom of the free memory. 4004f6ad66aSAchin Gupta * E.g. if BL2 is considered as top-loaded, the meminfo structure 4014f6ad66aSAchin Gupta * will give the following view of the memory, hiding the chunk of 4024f6ad66aSAchin Gupta * free memory above BL2: 4034f6ad66aSAchin Gupta * ------------ 0x04040000 4044f6ad66aSAchin Gupta * | | 4054f6ad66aSAchin Gupta * | | 4064f6ad66aSAchin Gupta * | BL2 | 4074f6ad66aSAchin Gupta * |----------| 0x04020000 4084f6ad66aSAchin Gupta * | | <- Free space (2) 4094f6ad66aSAchin Gupta * |----------| 4104f6ad66aSAchin Gupta * | BL1 | 4114f6ad66aSAchin Gupta * ------------ 0x04000000 4124f6ad66aSAchin Gupta */ 4134f6ad66aSAchin Gupta if (fixed_addr != 0) { 4144f6ad66aSAchin Gupta /* Load the image at the given address. */ 4154f6ad66aSAchin Gupta image_base = fixed_addr; 4164f6ad66aSAchin Gupta 4174f6ad66aSAchin Gupta /* Check whether the image fits. */ 4184f6ad66aSAchin Gupta if ((image_base < mem_layout->free_base) || 419*9d72b4eaSJames Morrissey (image_base + image_size > 4204f6ad66aSAchin Gupta mem_layout->free_base + mem_layout->free_size)) { 421*9d72b4eaSJames Morrissey ERROR("Cannot load '%s' file: Not enough space.\n", 4224f6ad66aSAchin Gupta image_name); 423*9d72b4eaSJames Morrissey dump_load_info(image_base, image_size, mem_layout); 424*9d72b4eaSJames Morrissey goto fail; 4254f6ad66aSAchin Gupta } 4264f6ad66aSAchin Gupta 4274f6ad66aSAchin Gupta /* Check whether the fixed load address is page-aligned. */ 4284f6ad66aSAchin Gupta if (!is_page_aligned(image_base)) { 429*9d72b4eaSJames Morrissey ERROR("Cannot load '%s' file at unaligned address 0x%lx\n", 4304f6ad66aSAchin Gupta image_name, fixed_addr); 431*9d72b4eaSJames Morrissey goto fail; 4324f6ad66aSAchin Gupta } 4334f6ad66aSAchin Gupta 4344f6ad66aSAchin Gupta /* 4354f6ad66aSAchin Gupta * Calculate the amount of extra memory used due to fixed 4364f6ad66aSAchin Gupta * loading. 4374f6ad66aSAchin Gupta */ 4384f6ad66aSAchin Gupta if (load_type == TOP_LOAD) { 4394f6ad66aSAchin Gupta unsigned long max_addr, space_used; 4404f6ad66aSAchin Gupta /* 4414f6ad66aSAchin Gupta * ------------ max_addr 4424f6ad66aSAchin Gupta * | /wasted/ | | offset 4434f6ad66aSAchin Gupta * |..........|.............................. 4444f6ad66aSAchin Gupta * | image | | image_flen 4454f6ad66aSAchin Gupta * |----------| fixed_addr 4464f6ad66aSAchin Gupta * | | 4474f6ad66aSAchin Gupta * | | 4484f6ad66aSAchin Gupta * ------------ total_base 4494f6ad66aSAchin Gupta */ 4504f6ad66aSAchin Gupta max_addr = mem_layout->total_base + mem_layout->total_size; 4514f6ad66aSAchin Gupta /* 4524f6ad66aSAchin Gupta * Compute the amount of memory used by the image. 4534f6ad66aSAchin Gupta * Corresponds to all space above the image load 4544f6ad66aSAchin Gupta * address. 4554f6ad66aSAchin Gupta */ 4564f6ad66aSAchin Gupta space_used = max_addr - fixed_addr; 4574f6ad66aSAchin Gupta /* 4584f6ad66aSAchin Gupta * Calculate the amount of wasted memory within the 4594f6ad66aSAchin Gupta * amount of memory used by the image. 4604f6ad66aSAchin Gupta */ 461*9d72b4eaSJames Morrissey offset = space_used - image_size; 4624f6ad66aSAchin Gupta } else /* BOT_LOAD */ 4634f6ad66aSAchin Gupta /* 4644f6ad66aSAchin Gupta * ------------ 4654f6ad66aSAchin Gupta * | | 4664f6ad66aSAchin Gupta * | | 4674f6ad66aSAchin Gupta * |----------| 4684f6ad66aSAchin Gupta * | image | 4694f6ad66aSAchin Gupta * |..........| fixed_addr 4704f6ad66aSAchin Gupta * | /wasted/ | | offset 4714f6ad66aSAchin Gupta * ------------ total_base 4724f6ad66aSAchin Gupta */ 4734f6ad66aSAchin Gupta offset = fixed_addr - mem_layout->total_base; 4744f6ad66aSAchin Gupta } 4754f6ad66aSAchin Gupta 4764f6ad66aSAchin Gupta /* We have enough space so load the image now */ 477*9d72b4eaSJames Morrissey /* TODO: Consider whether to try to recover/retry a partially successful read */ 478*9d72b4eaSJames Morrissey io_result = io_read(image_handle, (void *)image_base, image_size, &bytes_read); 479*9d72b4eaSJames Morrissey if ((io_result != IO_SUCCESS) || (bytes_read < image_size)) { 480*9d72b4eaSJames Morrissey ERROR("Failed to load '%s' file (%i)\n", image_name, io_result); 481*9d72b4eaSJames Morrissey goto fail; 4824f6ad66aSAchin Gupta } 4834f6ad66aSAchin Gupta 4844f6ad66aSAchin Gupta /* 4854f6ad66aSAchin Gupta * File has been successfully loaded. Update the free memory 4864f6ad66aSAchin Gupta * data structure & flush the contents of the TZRAM so that 4874f6ad66aSAchin Gupta * the next EL can see it. 4884f6ad66aSAchin Gupta */ 4894f6ad66aSAchin Gupta /* Update the memory contents */ 490*9d72b4eaSJames Morrissey flush_dcache_range(image_base, image_size); 4914f6ad66aSAchin Gupta 492*9d72b4eaSJames Morrissey mem_layout->free_size -= image_size + offset; 4934f6ad66aSAchin Gupta 4944f6ad66aSAchin Gupta /* Update the base of free memory since its moved up */ 4954f6ad66aSAchin Gupta if (load_type == BOT_LOAD) 496*9d72b4eaSJames Morrissey mem_layout->free_base += offset + image_size; 497*9d72b4eaSJames Morrissey 498*9d72b4eaSJames Morrissey exit: 499*9d72b4eaSJames Morrissey io_result = io_close(image_handle); 500*9d72b4eaSJames Morrissey /* Ignore improbable/unrecoverable error in 'close' */ 501*9d72b4eaSJames Morrissey 502*9d72b4eaSJames Morrissey /* TODO: Consider maintaining open device connection from this bootloader stage */ 503*9d72b4eaSJames Morrissey io_result = io_dev_close(dev_handle); 504*9d72b4eaSJames Morrissey /* Ignore improbable/unrecoverable error in 'dev_close' */ 5054f6ad66aSAchin Gupta 5064f6ad66aSAchin Gupta return image_base; 507*9d72b4eaSJames Morrissey 508*9d72b4eaSJames Morrissey fail: image_base = 0; 509*9d72b4eaSJames Morrissey goto exit; 5104f6ad66aSAchin Gupta } 5114f6ad66aSAchin Gupta 5124f6ad66aSAchin Gupta /******************************************************************************* 5134f6ad66aSAchin Gupta * Run a loaded image from the given entry point. This could result in either 5144f6ad66aSAchin Gupta * dropping into a lower exception level or jumping to a higher exception level. 5154f6ad66aSAchin Gupta * The only way of doing the latter is through an SMC. In either case, setup the 5164f6ad66aSAchin Gupta * parameters for the EL change request correctly. 5174f6ad66aSAchin Gupta ******************************************************************************/ 5184f6ad66aSAchin Gupta int run_image(unsigned long entrypoint, 5194f6ad66aSAchin Gupta unsigned long spsr, 5204f6ad66aSAchin Gupta unsigned long target_security_state, 5214f6ad66aSAchin Gupta meminfo *mem_layout, 5224f6ad66aSAchin Gupta void *data) 5234f6ad66aSAchin Gupta { 5244f6ad66aSAchin Gupta el_change_info run_image_info; 5254f6ad66aSAchin Gupta unsigned long current_el = read_current_el(); 5264f6ad66aSAchin Gupta 5274f6ad66aSAchin Gupta /* Tell next EL what we want done */ 5284f6ad66aSAchin Gupta run_image_info.args.arg0 = RUN_IMAGE; 5294f6ad66aSAchin Gupta run_image_info.entrypoint = entrypoint; 5304f6ad66aSAchin Gupta run_image_info.spsr = spsr; 5314f6ad66aSAchin Gupta run_image_info.security_state = target_security_state; 5324f6ad66aSAchin Gupta run_image_info.next = 0; 5334f6ad66aSAchin Gupta 5344f6ad66aSAchin Gupta /* 5354f6ad66aSAchin Gupta * If we are EL3 then only an eret can take us to the desired 5364f6ad66aSAchin Gupta * exception level. Else for the time being assume that we have 5374f6ad66aSAchin Gupta * to jump to a higher EL and issue an SMC. Contents of argY 5384f6ad66aSAchin Gupta * will go into the general purpose register xY e.g. arg0->x0 5394f6ad66aSAchin Gupta */ 5404f6ad66aSAchin Gupta if (GET_EL(current_el) == MODE_EL3) { 5414f6ad66aSAchin Gupta run_image_info.args.arg1 = (unsigned long) mem_layout; 5424f6ad66aSAchin Gupta run_image_info.args.arg2 = (unsigned long) data; 5434f6ad66aSAchin Gupta } else { 5444f6ad66aSAchin Gupta run_image_info.args.arg1 = entrypoint; 5454f6ad66aSAchin Gupta run_image_info.args.arg2 = spsr; 5464f6ad66aSAchin Gupta run_image_info.args.arg3 = (unsigned long) mem_layout; 5474f6ad66aSAchin Gupta run_image_info.args.arg4 = (unsigned long) data; 5484f6ad66aSAchin Gupta } 5494f6ad66aSAchin Gupta 5504f6ad66aSAchin Gupta return change_el(&run_image_info); 5514f6ad66aSAchin Gupta } 552