1*b7cb133eSJeenu Viswambharan /* 2*b7cb133eSJeenu Viswambharan * Copyright (c) 2017, ARM Limited and Contributors. All rights reserved. 3*b7cb133eSJeenu Viswambharan * 4*b7cb133eSJeenu Viswambharan * SPDX-License-Identifier: BSD-3-Clause 5*b7cb133eSJeenu Viswambharan */ 6*b7cb133eSJeenu Viswambharan 7*b7cb133eSJeenu Viswambharan #include <arch_helpers.h> 8*b7cb133eSJeenu Viswambharan #include <assert.h> 9*b7cb133eSJeenu Viswambharan #include <bl_common.h> 10*b7cb133eSJeenu Viswambharan #include <cassert.h> 11*b7cb133eSJeenu Viswambharan #include <context_mgmt.h> 12*b7cb133eSJeenu Viswambharan #include <debug.h> 13*b7cb133eSJeenu Viswambharan #include <ehf.h> 14*b7cb133eSJeenu Viswambharan #include <interrupt_mgmt.h> 15*b7cb133eSJeenu Viswambharan #include <runtime_svc.h> 16*b7cb133eSJeenu Viswambharan #include <sdei.h> 17*b7cb133eSJeenu Viswambharan #include <string.h> 18*b7cb133eSJeenu Viswambharan #include "sdei_private.h" 19*b7cb133eSJeenu Viswambharan 20*b7cb133eSJeenu Viswambharan #define PE_MASKED 1 21*b7cb133eSJeenu Viswambharan #define PE_NOT_MASKED 0 22*b7cb133eSJeenu Viswambharan 23*b7cb133eSJeenu Viswambharan /* x0-x17 GPREGS context */ 24*b7cb133eSJeenu Viswambharan #define SDEI_SAVED_GPREGS 18 25*b7cb133eSJeenu Viswambharan 26*b7cb133eSJeenu Viswambharan /* Maximum preemption nesting levels: Critical priority and Normal priority */ 27*b7cb133eSJeenu Viswambharan #define MAX_EVENT_NESTING 2 28*b7cb133eSJeenu Viswambharan 29*b7cb133eSJeenu Viswambharan /* Per-CPU SDEI state access macro */ 30*b7cb133eSJeenu Viswambharan #define sdei_get_this_pe_state() (&sdei_cpu_state[plat_my_core_pos()]) 31*b7cb133eSJeenu Viswambharan 32*b7cb133eSJeenu Viswambharan /* Structure to store information about an outstanding dispatch */ 33*b7cb133eSJeenu Viswambharan typedef struct sdei_dispatch_context { 34*b7cb133eSJeenu Viswambharan sdei_ev_map_t *map; 35*b7cb133eSJeenu Viswambharan unsigned int sec_state; 36*b7cb133eSJeenu Viswambharan unsigned int intr_raw; 37*b7cb133eSJeenu Viswambharan uint64_t x[SDEI_SAVED_GPREGS]; 38*b7cb133eSJeenu Viswambharan 39*b7cb133eSJeenu Viswambharan /* Exception state registers */ 40*b7cb133eSJeenu Viswambharan uint64_t elr_el3; 41*b7cb133eSJeenu Viswambharan uint64_t spsr_el3; 42*b7cb133eSJeenu Viswambharan } sdei_dispatch_context_t; 43*b7cb133eSJeenu Viswambharan 44*b7cb133eSJeenu Viswambharan /* Per-CPU SDEI state data */ 45*b7cb133eSJeenu Viswambharan typedef struct sdei_cpu_state { 46*b7cb133eSJeenu Viswambharan sdei_dispatch_context_t dispatch_stack[MAX_EVENT_NESTING]; 47*b7cb133eSJeenu Viswambharan unsigned short stack_top; /* Empty ascending */ 48*b7cb133eSJeenu Viswambharan unsigned int pe_masked:1; 49*b7cb133eSJeenu Viswambharan unsigned int pending_enables:1; 50*b7cb133eSJeenu Viswambharan } sdei_cpu_state_t; 51*b7cb133eSJeenu Viswambharan 52*b7cb133eSJeenu Viswambharan /* SDEI states for all cores in the system */ 53*b7cb133eSJeenu Viswambharan static sdei_cpu_state_t sdei_cpu_state[PLATFORM_CORE_COUNT]; 54*b7cb133eSJeenu Viswambharan 55*b7cb133eSJeenu Viswambharan unsigned int sdei_pe_mask(void) 56*b7cb133eSJeenu Viswambharan { 57*b7cb133eSJeenu Viswambharan unsigned int ret; 58*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state = sdei_get_this_pe_state(); 59*b7cb133eSJeenu Viswambharan 60*b7cb133eSJeenu Viswambharan /* 61*b7cb133eSJeenu Viswambharan * Return value indicates whether this call had any effect in the mask 62*b7cb133eSJeenu Viswambharan * status of this PE. 63*b7cb133eSJeenu Viswambharan */ 64*b7cb133eSJeenu Viswambharan ret = (state->pe_masked ^ PE_MASKED); 65*b7cb133eSJeenu Viswambharan state->pe_masked = PE_MASKED; 66*b7cb133eSJeenu Viswambharan 67*b7cb133eSJeenu Viswambharan return ret; 68*b7cb133eSJeenu Viswambharan } 69*b7cb133eSJeenu Viswambharan 70*b7cb133eSJeenu Viswambharan void sdei_pe_unmask(void) 71*b7cb133eSJeenu Viswambharan { 72*b7cb133eSJeenu Viswambharan int i; 73*b7cb133eSJeenu Viswambharan sdei_ev_map_t *map; 74*b7cb133eSJeenu Viswambharan sdei_entry_t *se; 75*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state = sdei_get_this_pe_state(); 76*b7cb133eSJeenu Viswambharan uint64_t my_mpidr = read_mpidr_el1() & MPIDR_AFFINITY_MASK; 77*b7cb133eSJeenu Viswambharan 78*b7cb133eSJeenu Viswambharan /* 79*b7cb133eSJeenu Viswambharan * If there are pending enables, iterate through the private mappings 80*b7cb133eSJeenu Viswambharan * and enable those bound maps that are in enabled state. Also, iterate 81*b7cb133eSJeenu Viswambharan * through shared mappings and enable interrupts of events that are 82*b7cb133eSJeenu Viswambharan * targeted to this PE. 83*b7cb133eSJeenu Viswambharan */ 84*b7cb133eSJeenu Viswambharan if (state->pending_enables) { 85*b7cb133eSJeenu Viswambharan for_each_private_map(i, map) { 86*b7cb133eSJeenu Viswambharan se = get_event_entry(map); 87*b7cb133eSJeenu Viswambharan if (is_map_bound(map) && GET_EV_STATE(se, ENABLED)) 88*b7cb133eSJeenu Viswambharan plat_ic_enable_interrupt(map->intr); 89*b7cb133eSJeenu Viswambharan } 90*b7cb133eSJeenu Viswambharan 91*b7cb133eSJeenu Viswambharan for_each_shared_map(i, map) { 92*b7cb133eSJeenu Viswambharan se = get_event_entry(map); 93*b7cb133eSJeenu Viswambharan 94*b7cb133eSJeenu Viswambharan sdei_map_lock(map); 95*b7cb133eSJeenu Viswambharan if (is_map_bound(map) && 96*b7cb133eSJeenu Viswambharan GET_EV_STATE(se, ENABLED) && 97*b7cb133eSJeenu Viswambharan (se->reg_flags == SDEI_REGF_RM_PE) && 98*b7cb133eSJeenu Viswambharan (se->affinity == my_mpidr)) { 99*b7cb133eSJeenu Viswambharan plat_ic_enable_interrupt(map->intr); 100*b7cb133eSJeenu Viswambharan } 101*b7cb133eSJeenu Viswambharan sdei_map_unlock(map); 102*b7cb133eSJeenu Viswambharan } 103*b7cb133eSJeenu Viswambharan } 104*b7cb133eSJeenu Viswambharan 105*b7cb133eSJeenu Viswambharan state->pending_enables = 0; 106*b7cb133eSJeenu Viswambharan state->pe_masked = PE_NOT_MASKED; 107*b7cb133eSJeenu Viswambharan } 108*b7cb133eSJeenu Viswambharan 109*b7cb133eSJeenu Viswambharan /* Push a dispatch context to the dispatch stack */ 110*b7cb133eSJeenu Viswambharan static sdei_dispatch_context_t *push_dispatch(void) 111*b7cb133eSJeenu Viswambharan { 112*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state = sdei_get_this_pe_state(); 113*b7cb133eSJeenu Viswambharan sdei_dispatch_context_t *disp_ctx; 114*b7cb133eSJeenu Viswambharan 115*b7cb133eSJeenu Viswambharan /* Cannot have more than max events */ 116*b7cb133eSJeenu Viswambharan assert(state->stack_top < MAX_EVENT_NESTING); 117*b7cb133eSJeenu Viswambharan 118*b7cb133eSJeenu Viswambharan disp_ctx = &state->dispatch_stack[state->stack_top]; 119*b7cb133eSJeenu Viswambharan state->stack_top++; 120*b7cb133eSJeenu Viswambharan 121*b7cb133eSJeenu Viswambharan return disp_ctx; 122*b7cb133eSJeenu Viswambharan } 123*b7cb133eSJeenu Viswambharan 124*b7cb133eSJeenu Viswambharan /* Pop a dispatch context to the dispatch stack */ 125*b7cb133eSJeenu Viswambharan static sdei_dispatch_context_t *pop_dispatch(void) 126*b7cb133eSJeenu Viswambharan { 127*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state = sdei_get_this_pe_state(); 128*b7cb133eSJeenu Viswambharan 129*b7cb133eSJeenu Viswambharan if (state->stack_top == 0) 130*b7cb133eSJeenu Viswambharan return NULL; 131*b7cb133eSJeenu Viswambharan 132*b7cb133eSJeenu Viswambharan assert(state->stack_top <= MAX_EVENT_NESTING); 133*b7cb133eSJeenu Viswambharan 134*b7cb133eSJeenu Viswambharan state->stack_top--; 135*b7cb133eSJeenu Viswambharan 136*b7cb133eSJeenu Viswambharan return &state->dispatch_stack[state->stack_top]; 137*b7cb133eSJeenu Viswambharan } 138*b7cb133eSJeenu Viswambharan 139*b7cb133eSJeenu Viswambharan /* Retrieve the context at the top of dispatch stack */ 140*b7cb133eSJeenu Viswambharan static sdei_dispatch_context_t *get_outstanding_dispatch(void) 141*b7cb133eSJeenu Viswambharan { 142*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state = sdei_get_this_pe_state(); 143*b7cb133eSJeenu Viswambharan 144*b7cb133eSJeenu Viswambharan if (state->stack_top == 0) 145*b7cb133eSJeenu Viswambharan return NULL; 146*b7cb133eSJeenu Viswambharan 147*b7cb133eSJeenu Viswambharan assert(state->stack_top <= MAX_EVENT_NESTING); 148*b7cb133eSJeenu Viswambharan 149*b7cb133eSJeenu Viswambharan return &state->dispatch_stack[state->stack_top - 1]; 150*b7cb133eSJeenu Viswambharan } 151*b7cb133eSJeenu Viswambharan 152*b7cb133eSJeenu Viswambharan static void save_event_ctx(sdei_ev_map_t *map, void *tgt_ctx, int sec_state, 153*b7cb133eSJeenu Viswambharan unsigned int intr_raw) 154*b7cb133eSJeenu Viswambharan { 155*b7cb133eSJeenu Viswambharan sdei_dispatch_context_t *disp_ctx; 156*b7cb133eSJeenu Viswambharan gp_regs_t *tgt_gpregs; 157*b7cb133eSJeenu Viswambharan el3_state_t *tgt_el3; 158*b7cb133eSJeenu Viswambharan 159*b7cb133eSJeenu Viswambharan assert(tgt_ctx); 160*b7cb133eSJeenu Viswambharan tgt_gpregs = get_gpregs_ctx(tgt_ctx); 161*b7cb133eSJeenu Viswambharan tgt_el3 = get_el3state_ctx(tgt_ctx); 162*b7cb133eSJeenu Viswambharan 163*b7cb133eSJeenu Viswambharan disp_ctx = push_dispatch(); 164*b7cb133eSJeenu Viswambharan assert(disp_ctx); 165*b7cb133eSJeenu Viswambharan disp_ctx->sec_state = sec_state; 166*b7cb133eSJeenu Viswambharan disp_ctx->map = map; 167*b7cb133eSJeenu Viswambharan disp_ctx->intr_raw = intr_raw; 168*b7cb133eSJeenu Viswambharan 169*b7cb133eSJeenu Viswambharan /* Save general purpose and exception registers */ 170*b7cb133eSJeenu Viswambharan memcpy(disp_ctx->x, tgt_gpregs, sizeof(disp_ctx->x)); 171*b7cb133eSJeenu Viswambharan disp_ctx->spsr_el3 = read_ctx_reg(tgt_el3, CTX_SPSR_EL3); 172*b7cb133eSJeenu Viswambharan disp_ctx->elr_el3 = read_ctx_reg(tgt_el3, CTX_ELR_EL3); 173*b7cb133eSJeenu Viswambharan } 174*b7cb133eSJeenu Viswambharan 175*b7cb133eSJeenu Viswambharan static void restore_event_ctx(sdei_dispatch_context_t *disp_ctx, void *tgt_ctx) 176*b7cb133eSJeenu Viswambharan { 177*b7cb133eSJeenu Viswambharan gp_regs_t *tgt_gpregs; 178*b7cb133eSJeenu Viswambharan el3_state_t *tgt_el3; 179*b7cb133eSJeenu Viswambharan 180*b7cb133eSJeenu Viswambharan assert(tgt_ctx); 181*b7cb133eSJeenu Viswambharan tgt_gpregs = get_gpregs_ctx(tgt_ctx); 182*b7cb133eSJeenu Viswambharan tgt_el3 = get_el3state_ctx(tgt_ctx); 183*b7cb133eSJeenu Viswambharan 184*b7cb133eSJeenu Viswambharan CASSERT(sizeof(disp_ctx->x) == (SDEI_SAVED_GPREGS * sizeof(uint64_t)), 185*b7cb133eSJeenu Viswambharan foo); 186*b7cb133eSJeenu Viswambharan 187*b7cb133eSJeenu Viswambharan /* Restore general purpose and exception registers */ 188*b7cb133eSJeenu Viswambharan memcpy(tgt_gpregs, disp_ctx->x, sizeof(disp_ctx->x)); 189*b7cb133eSJeenu Viswambharan write_ctx_reg(tgt_el3, CTX_SPSR_EL3, disp_ctx->spsr_el3); 190*b7cb133eSJeenu Viswambharan write_ctx_reg(tgt_el3, CTX_ELR_EL3, disp_ctx->elr_el3); 191*b7cb133eSJeenu Viswambharan } 192*b7cb133eSJeenu Viswambharan 193*b7cb133eSJeenu Viswambharan static void save_secure_context(void) 194*b7cb133eSJeenu Viswambharan { 195*b7cb133eSJeenu Viswambharan cm_el1_sysregs_context_save(SECURE); 196*b7cb133eSJeenu Viswambharan } 197*b7cb133eSJeenu Viswambharan 198*b7cb133eSJeenu Viswambharan /* Restore Secure context and arrange to resume it at the next ERET */ 199*b7cb133eSJeenu Viswambharan static void restore_and_resume_secure_context(void) 200*b7cb133eSJeenu Viswambharan { 201*b7cb133eSJeenu Viswambharan cm_el1_sysregs_context_restore(SECURE); 202*b7cb133eSJeenu Viswambharan cm_set_next_eret_context(SECURE); 203*b7cb133eSJeenu Viswambharan } 204*b7cb133eSJeenu Viswambharan 205*b7cb133eSJeenu Viswambharan /* 206*b7cb133eSJeenu Viswambharan * Restore Non-secure context and arrange to resume it at the next ERET. Return 207*b7cb133eSJeenu Viswambharan * pointer to the Non-secure context. 208*b7cb133eSJeenu Viswambharan */ 209*b7cb133eSJeenu Viswambharan static cpu_context_t *restore_and_resume_ns_context(void) 210*b7cb133eSJeenu Viswambharan { 211*b7cb133eSJeenu Viswambharan cpu_context_t *ns_ctx; 212*b7cb133eSJeenu Viswambharan 213*b7cb133eSJeenu Viswambharan cm_el1_sysregs_context_restore(NON_SECURE); 214*b7cb133eSJeenu Viswambharan cm_set_next_eret_context(NON_SECURE); 215*b7cb133eSJeenu Viswambharan 216*b7cb133eSJeenu Viswambharan ns_ctx = cm_get_context(NON_SECURE); 217*b7cb133eSJeenu Viswambharan assert(ns_ctx); 218*b7cb133eSJeenu Viswambharan 219*b7cb133eSJeenu Viswambharan return ns_ctx; 220*b7cb133eSJeenu Viswambharan } 221*b7cb133eSJeenu Viswambharan 222*b7cb133eSJeenu Viswambharan /* 223*b7cb133eSJeenu Viswambharan * Populate the Non-secure context so that the next ERET will dispatch to the 224*b7cb133eSJeenu Viswambharan * SDEI client. 225*b7cb133eSJeenu Viswambharan */ 226*b7cb133eSJeenu Viswambharan static void setup_ns_dispatch(sdei_ev_map_t *map, sdei_entry_t *se, 227*b7cb133eSJeenu Viswambharan cpu_context_t *ctx, int sec_state_to_resume, 228*b7cb133eSJeenu Viswambharan unsigned int intr_raw) 229*b7cb133eSJeenu Viswambharan { 230*b7cb133eSJeenu Viswambharan el3_state_t *el3_ctx = get_el3state_ctx(ctx); 231*b7cb133eSJeenu Viswambharan 232*b7cb133eSJeenu Viswambharan /* Push the event and context */ 233*b7cb133eSJeenu Viswambharan save_event_ctx(map, ctx, sec_state_to_resume, intr_raw); 234*b7cb133eSJeenu Viswambharan 235*b7cb133eSJeenu Viswambharan /* 236*b7cb133eSJeenu Viswambharan * Setup handler arguments: 237*b7cb133eSJeenu Viswambharan * 238*b7cb133eSJeenu Viswambharan * - x0: Event number 239*b7cb133eSJeenu Viswambharan * - x1: Handler argument supplied at the time of event registration 240*b7cb133eSJeenu Viswambharan * - x2: Interrupted PC 241*b7cb133eSJeenu Viswambharan * - x3: Interrupted SPSR 242*b7cb133eSJeenu Viswambharan */ 243*b7cb133eSJeenu Viswambharan SMC_SET_GP(ctx, CTX_GPREG_X0, map->ev_num); 244*b7cb133eSJeenu Viswambharan SMC_SET_GP(ctx, CTX_GPREG_X1, se->arg); 245*b7cb133eSJeenu Viswambharan SMC_SET_GP(ctx, CTX_GPREG_X2, read_ctx_reg(el3_ctx, CTX_ELR_EL3)); 246*b7cb133eSJeenu Viswambharan SMC_SET_GP(ctx, CTX_GPREG_X3, read_ctx_reg(el3_ctx, CTX_SPSR_EL3)); 247*b7cb133eSJeenu Viswambharan 248*b7cb133eSJeenu Viswambharan /* 249*b7cb133eSJeenu Viswambharan * Prepare for ERET: 250*b7cb133eSJeenu Viswambharan * 251*b7cb133eSJeenu Viswambharan * - Set PC to the registered handler address 252*b7cb133eSJeenu Viswambharan * - Set SPSR to jump to client EL with exceptions masked 253*b7cb133eSJeenu Viswambharan */ 254*b7cb133eSJeenu Viswambharan cm_set_elr_spsr_el3(NON_SECURE, (uintptr_t) se->ep, 255*b7cb133eSJeenu Viswambharan SPSR_64(sdei_client_el(), MODE_SP_ELX, 256*b7cb133eSJeenu Viswambharan DISABLE_ALL_EXCEPTIONS)); 257*b7cb133eSJeenu Viswambharan } 258*b7cb133eSJeenu Viswambharan 259*b7cb133eSJeenu Viswambharan /* Handle a triggered SDEI interrupt while events were masked on this PE */ 260*b7cb133eSJeenu Viswambharan static void handle_masked_trigger(sdei_ev_map_t *map, sdei_entry_t *se, 261*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state, unsigned int intr_raw) 262*b7cb133eSJeenu Viswambharan { 263*b7cb133eSJeenu Viswambharan uint64_t my_mpidr __unused = (read_mpidr_el1() & MPIDR_AFFINITY_MASK); 264*b7cb133eSJeenu Viswambharan int disable = 0; 265*b7cb133eSJeenu Viswambharan 266*b7cb133eSJeenu Viswambharan /* Nothing to do for event 0 */ 267*b7cb133eSJeenu Viswambharan if (map->ev_num == SDEI_EVENT_0) 268*b7cb133eSJeenu Viswambharan return; 269*b7cb133eSJeenu Viswambharan 270*b7cb133eSJeenu Viswambharan /* 271*b7cb133eSJeenu Viswambharan * For a private event, or for a shared event specifically routed to 272*b7cb133eSJeenu Viswambharan * this CPU, we disable interrupt, leave the interrupt pending, and do 273*b7cb133eSJeenu Viswambharan * EOI. 274*b7cb133eSJeenu Viswambharan */ 275*b7cb133eSJeenu Viswambharan if (is_event_private(map)) { 276*b7cb133eSJeenu Viswambharan disable = 1; 277*b7cb133eSJeenu Viswambharan } else if (se->reg_flags == SDEI_REGF_RM_PE) { 278*b7cb133eSJeenu Viswambharan assert(se->affinity == my_mpidr); 279*b7cb133eSJeenu Viswambharan disable = 1; 280*b7cb133eSJeenu Viswambharan } 281*b7cb133eSJeenu Viswambharan 282*b7cb133eSJeenu Viswambharan if (disable) { 283*b7cb133eSJeenu Viswambharan plat_ic_disable_interrupt(map->intr); 284*b7cb133eSJeenu Viswambharan plat_ic_set_interrupt_pending(map->intr); 285*b7cb133eSJeenu Viswambharan plat_ic_end_of_interrupt(intr_raw); 286*b7cb133eSJeenu Viswambharan state->pending_enables = 1; 287*b7cb133eSJeenu Viswambharan 288*b7cb133eSJeenu Viswambharan return; 289*b7cb133eSJeenu Viswambharan } 290*b7cb133eSJeenu Viswambharan 291*b7cb133eSJeenu Viswambharan /* 292*b7cb133eSJeenu Viswambharan * We just received a shared event with routing set to ANY PE. The 293*b7cb133eSJeenu Viswambharan * interrupt can't be delegated on this PE as SDEI events are masked. 294*b7cb133eSJeenu Viswambharan * However, because its routing mode is ANY, it is possible that the 295*b7cb133eSJeenu Viswambharan * event can be delegated on any other PE that hasn't masked events. 296*b7cb133eSJeenu Viswambharan * Therefore, we set the interrupt back pending so as to give other 297*b7cb133eSJeenu Viswambharan * suitable PEs a chance of handling it. 298*b7cb133eSJeenu Viswambharan */ 299*b7cb133eSJeenu Viswambharan assert(plat_ic_is_spi(map->intr)); 300*b7cb133eSJeenu Viswambharan plat_ic_set_interrupt_pending(map->intr); 301*b7cb133eSJeenu Viswambharan 302*b7cb133eSJeenu Viswambharan /* 303*b7cb133eSJeenu Viswambharan * Leaving the same interrupt pending also means that the same interrupt 304*b7cb133eSJeenu Viswambharan * can target this PE again as soon as this PE leaves EL3. Whether and 305*b7cb133eSJeenu Viswambharan * how often that happens depends on the implementation of GIC. 306*b7cb133eSJeenu Viswambharan * 307*b7cb133eSJeenu Viswambharan * We therefore call a platform handler to resolve this situation. 308*b7cb133eSJeenu Viswambharan */ 309*b7cb133eSJeenu Viswambharan plat_sdei_handle_masked_trigger(my_mpidr, map->intr); 310*b7cb133eSJeenu Viswambharan 311*b7cb133eSJeenu Viswambharan /* This PE is masked. We EOI the interrupt, as it can't be delegated */ 312*b7cb133eSJeenu Viswambharan plat_ic_end_of_interrupt(intr_raw); 313*b7cb133eSJeenu Viswambharan } 314*b7cb133eSJeenu Viswambharan 315*b7cb133eSJeenu Viswambharan /* SDEI main interrupt handler */ 316*b7cb133eSJeenu Viswambharan int sdei_intr_handler(uint32_t intr_raw, uint32_t flags, void *handle, 317*b7cb133eSJeenu Viswambharan void *cookie) 318*b7cb133eSJeenu Viswambharan { 319*b7cb133eSJeenu Viswambharan sdei_entry_t *se; 320*b7cb133eSJeenu Viswambharan cpu_context_t *ctx; 321*b7cb133eSJeenu Viswambharan sdei_ev_map_t *map; 322*b7cb133eSJeenu Viswambharan sdei_dispatch_context_t *disp_ctx; 323*b7cb133eSJeenu Viswambharan unsigned int sec_state; 324*b7cb133eSJeenu Viswambharan sdei_cpu_state_t *state; 325*b7cb133eSJeenu Viswambharan uint32_t intr; 326*b7cb133eSJeenu Viswambharan 327*b7cb133eSJeenu Viswambharan /* 328*b7cb133eSJeenu Viswambharan * To handle an event, the following conditions must be true: 329*b7cb133eSJeenu Viswambharan * 330*b7cb133eSJeenu Viswambharan * 1. Event must be signalled 331*b7cb133eSJeenu Viswambharan * 2. Event must be enabled 332*b7cb133eSJeenu Viswambharan * 3. This PE must be a target PE for the event 333*b7cb133eSJeenu Viswambharan * 4. PE must be unmasked for SDEI 334*b7cb133eSJeenu Viswambharan * 5. If this is a normal event, no event must be running 335*b7cb133eSJeenu Viswambharan * 6. If this is a critical event, no critical event must be running 336*b7cb133eSJeenu Viswambharan * 337*b7cb133eSJeenu Viswambharan * (1) and (2) are true when this function is running 338*b7cb133eSJeenu Viswambharan * (3) is enforced in GIC by selecting the appropriate routing option 339*b7cb133eSJeenu Viswambharan * (4) is satisfied by client calling PE_UNMASK 340*b7cb133eSJeenu Viswambharan * (5) and (6) is enforced using interrupt priority, the RPR, in GIC: 341*b7cb133eSJeenu Viswambharan * - Normal SDEI events belong to Normal SDE priority class 342*b7cb133eSJeenu Viswambharan * - Critical SDEI events belong to Critical CSDE priority class 343*b7cb133eSJeenu Viswambharan * 344*b7cb133eSJeenu Viswambharan * The interrupt has already been acknowledged, and therefore is active, 345*b7cb133eSJeenu Viswambharan * so no other PE can handle this event while we are at it. 346*b7cb133eSJeenu Viswambharan * 347*b7cb133eSJeenu Viswambharan * Find if this is an SDEI interrupt. There must be an event mapped to 348*b7cb133eSJeenu Viswambharan * this interrupt 349*b7cb133eSJeenu Viswambharan */ 350*b7cb133eSJeenu Viswambharan intr = plat_ic_get_interrupt_id(intr_raw); 351*b7cb133eSJeenu Viswambharan map = find_event_map_by_intr(intr, plat_ic_is_spi(intr)); 352*b7cb133eSJeenu Viswambharan if (!map) { 353*b7cb133eSJeenu Viswambharan ERROR("No SDEI map for interrupt %u\n", intr); 354*b7cb133eSJeenu Viswambharan panic(); 355*b7cb133eSJeenu Viswambharan } 356*b7cb133eSJeenu Viswambharan 357*b7cb133eSJeenu Viswambharan /* 358*b7cb133eSJeenu Viswambharan * Received interrupt number must either correspond to event 0, or must 359*b7cb133eSJeenu Viswambharan * be bound interrupt. 360*b7cb133eSJeenu Viswambharan */ 361*b7cb133eSJeenu Viswambharan assert((map->ev_num == SDEI_EVENT_0) || is_map_bound(map)); 362*b7cb133eSJeenu Viswambharan 363*b7cb133eSJeenu Viswambharan se = get_event_entry(map); 364*b7cb133eSJeenu Viswambharan state = sdei_get_this_pe_state(); 365*b7cb133eSJeenu Viswambharan 366*b7cb133eSJeenu Viswambharan if (state->pe_masked == PE_MASKED) { 367*b7cb133eSJeenu Viswambharan /* 368*b7cb133eSJeenu Viswambharan * Interrupts received while this PE was masked can't be 369*b7cb133eSJeenu Viswambharan * dispatched. 370*b7cb133eSJeenu Viswambharan */ 371*b7cb133eSJeenu Viswambharan SDEI_LOG("interrupt %u on %lx while PE masked\n", map->intr, 372*b7cb133eSJeenu Viswambharan read_mpidr_el1()); 373*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 374*b7cb133eSJeenu Viswambharan sdei_map_lock(map); 375*b7cb133eSJeenu Viswambharan 376*b7cb133eSJeenu Viswambharan handle_masked_trigger(map, se, state, intr_raw); 377*b7cb133eSJeenu Viswambharan 378*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 379*b7cb133eSJeenu Viswambharan sdei_map_unlock(map); 380*b7cb133eSJeenu Viswambharan 381*b7cb133eSJeenu Viswambharan return 0; 382*b7cb133eSJeenu Viswambharan } 383*b7cb133eSJeenu Viswambharan 384*b7cb133eSJeenu Viswambharan /* Insert load barrier for signalled SDEI event */ 385*b7cb133eSJeenu Viswambharan if (map->ev_num == SDEI_EVENT_0) 386*b7cb133eSJeenu Viswambharan dmbld(); 387*b7cb133eSJeenu Viswambharan 388*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 389*b7cb133eSJeenu Viswambharan sdei_map_lock(map); 390*b7cb133eSJeenu Viswambharan 391*b7cb133eSJeenu Viswambharan /* Assert shared event routed to this PE had been configured so */ 392*b7cb133eSJeenu Viswambharan if (is_event_shared(map) && (se->reg_flags == SDEI_REGF_RM_PE)) { 393*b7cb133eSJeenu Viswambharan assert(se->affinity == 394*b7cb133eSJeenu Viswambharan (read_mpidr_el1() & MPIDR_AFFINITY_MASK)); 395*b7cb133eSJeenu Viswambharan } 396*b7cb133eSJeenu Viswambharan 397*b7cb133eSJeenu Viswambharan if (!can_sdei_state_trans(se, DO_DISPATCH)) { 398*b7cb133eSJeenu Viswambharan SDEI_LOG("SDEI event 0x%x can't be dispatched; state=0x%x\n", 399*b7cb133eSJeenu Viswambharan map->ev_num, se->state); 400*b7cb133eSJeenu Viswambharan 401*b7cb133eSJeenu Viswambharan /* 402*b7cb133eSJeenu Viswambharan * If the event is registered, leave the interrupt pending so 403*b7cb133eSJeenu Viswambharan * that it's delivered when the event is enabled. 404*b7cb133eSJeenu Viswambharan */ 405*b7cb133eSJeenu Viswambharan if (GET_EV_STATE(se, REGISTERED)) 406*b7cb133eSJeenu Viswambharan plat_ic_set_interrupt_pending(map->intr); 407*b7cb133eSJeenu Viswambharan 408*b7cb133eSJeenu Viswambharan /* 409*b7cb133eSJeenu Viswambharan * The interrupt was disabled or unregistered after the handler 410*b7cb133eSJeenu Viswambharan * started to execute, which means now the interrupt is already 411*b7cb133eSJeenu Viswambharan * disabled and we just need to EOI the interrupt. 412*b7cb133eSJeenu Viswambharan */ 413*b7cb133eSJeenu Viswambharan plat_ic_end_of_interrupt(intr_raw); 414*b7cb133eSJeenu Viswambharan 415*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 416*b7cb133eSJeenu Viswambharan sdei_map_unlock(map); 417*b7cb133eSJeenu Viswambharan 418*b7cb133eSJeenu Viswambharan return 0; 419*b7cb133eSJeenu Viswambharan } 420*b7cb133eSJeenu Viswambharan 421*b7cb133eSJeenu Viswambharan disp_ctx = get_outstanding_dispatch(); 422*b7cb133eSJeenu Viswambharan if (is_event_critical(map)) { 423*b7cb133eSJeenu Viswambharan /* 424*b7cb133eSJeenu Viswambharan * If this event is Critical, and if there's an outstanding 425*b7cb133eSJeenu Viswambharan * dispatch, assert the latter is a Normal dispatch. Critical 426*b7cb133eSJeenu Viswambharan * events can preempt an outstanding Normal event dispatch. 427*b7cb133eSJeenu Viswambharan */ 428*b7cb133eSJeenu Viswambharan if (disp_ctx) 429*b7cb133eSJeenu Viswambharan assert(is_event_normal(disp_ctx->map)); 430*b7cb133eSJeenu Viswambharan } else { 431*b7cb133eSJeenu Viswambharan /* 432*b7cb133eSJeenu Viswambharan * If this event is Normal, assert that there are no outstanding 433*b7cb133eSJeenu Viswambharan * dispatches. Normal events can't preempt any outstanding event 434*b7cb133eSJeenu Viswambharan * dispatches. 435*b7cb133eSJeenu Viswambharan */ 436*b7cb133eSJeenu Viswambharan assert(disp_ctx == NULL); 437*b7cb133eSJeenu Viswambharan } 438*b7cb133eSJeenu Viswambharan 439*b7cb133eSJeenu Viswambharan sec_state = get_interrupt_src_ss(flags); 440*b7cb133eSJeenu Viswambharan 441*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 442*b7cb133eSJeenu Viswambharan sdei_map_unlock(map); 443*b7cb133eSJeenu Viswambharan 444*b7cb133eSJeenu Viswambharan SDEI_LOG("ACK %lx, ev:%d ss:%d spsr:%lx ELR:%lx\n", read_mpidr_el1(), 445*b7cb133eSJeenu Viswambharan map->ev_num, sec_state, read_spsr_el3(), 446*b7cb133eSJeenu Viswambharan read_elr_el3()); 447*b7cb133eSJeenu Viswambharan 448*b7cb133eSJeenu Viswambharan ctx = handle; 449*b7cb133eSJeenu Viswambharan 450*b7cb133eSJeenu Viswambharan /* 451*b7cb133eSJeenu Viswambharan * Check if we interrupted secure state. Perform a context switch so 452*b7cb133eSJeenu Viswambharan * that we can delegate to NS. 453*b7cb133eSJeenu Viswambharan */ 454*b7cb133eSJeenu Viswambharan if (sec_state == SECURE) { 455*b7cb133eSJeenu Viswambharan save_secure_context(); 456*b7cb133eSJeenu Viswambharan ctx = restore_and_resume_ns_context(); 457*b7cb133eSJeenu Viswambharan } 458*b7cb133eSJeenu Viswambharan 459*b7cb133eSJeenu Viswambharan setup_ns_dispatch(map, se, ctx, sec_state, intr_raw); 460*b7cb133eSJeenu Viswambharan 461*b7cb133eSJeenu Viswambharan /* 462*b7cb133eSJeenu Viswambharan * End of interrupt is done in sdei_event_complete, when the client 463*b7cb133eSJeenu Viswambharan * signals completion. 464*b7cb133eSJeenu Viswambharan */ 465*b7cb133eSJeenu Viswambharan return 0; 466*b7cb133eSJeenu Viswambharan } 467*b7cb133eSJeenu Viswambharan 468*b7cb133eSJeenu Viswambharan int sdei_event_complete(int resume, uint64_t pc) 469*b7cb133eSJeenu Viswambharan { 470*b7cb133eSJeenu Viswambharan sdei_dispatch_context_t *disp_ctx; 471*b7cb133eSJeenu Viswambharan sdei_entry_t *se; 472*b7cb133eSJeenu Viswambharan sdei_ev_map_t *map; 473*b7cb133eSJeenu Viswambharan cpu_context_t *ctx; 474*b7cb133eSJeenu Viswambharan sdei_action_t act; 475*b7cb133eSJeenu Viswambharan unsigned int client_el = sdei_client_el(); 476*b7cb133eSJeenu Viswambharan 477*b7cb133eSJeenu Viswambharan /* Return error if called without an active event */ 478*b7cb133eSJeenu Viswambharan disp_ctx = pop_dispatch(); 479*b7cb133eSJeenu Viswambharan if (!disp_ctx) 480*b7cb133eSJeenu Viswambharan return SDEI_EDENY; 481*b7cb133eSJeenu Viswambharan 482*b7cb133eSJeenu Viswambharan /* Validate resumption point */ 483*b7cb133eSJeenu Viswambharan if (resume && (plat_sdei_validate_entry_point(pc, client_el) != 0)) 484*b7cb133eSJeenu Viswambharan return SDEI_EDENY; 485*b7cb133eSJeenu Viswambharan 486*b7cb133eSJeenu Viswambharan map = disp_ctx->map; 487*b7cb133eSJeenu Viswambharan assert(map); 488*b7cb133eSJeenu Viswambharan 489*b7cb133eSJeenu Viswambharan se = get_event_entry(map); 490*b7cb133eSJeenu Viswambharan 491*b7cb133eSJeenu Viswambharan SDEI_LOG("EOI:%lx, %d spsr:%lx elr:%lx\n", read_mpidr_el1(), 492*b7cb133eSJeenu Viswambharan map->ev_num, read_spsr_el3(), read_elr_el3()); 493*b7cb133eSJeenu Viswambharan 494*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 495*b7cb133eSJeenu Viswambharan sdei_map_lock(map); 496*b7cb133eSJeenu Viswambharan 497*b7cb133eSJeenu Viswambharan act = resume ? DO_COMPLETE_RESUME : DO_COMPLETE; 498*b7cb133eSJeenu Viswambharan if (!can_sdei_state_trans(se, act)) { 499*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 500*b7cb133eSJeenu Viswambharan sdei_map_unlock(map); 501*b7cb133eSJeenu Viswambharan return SDEI_EDENY; 502*b7cb133eSJeenu Viswambharan } 503*b7cb133eSJeenu Viswambharan 504*b7cb133eSJeenu Viswambharan /* 505*b7cb133eSJeenu Viswambharan * Restore Non-secure to how it was originally interrupted. Once done, 506*b7cb133eSJeenu Viswambharan * it's up-to-date with the saved copy. 507*b7cb133eSJeenu Viswambharan */ 508*b7cb133eSJeenu Viswambharan ctx = cm_get_context(NON_SECURE); 509*b7cb133eSJeenu Viswambharan restore_event_ctx(disp_ctx, ctx); 510*b7cb133eSJeenu Viswambharan 511*b7cb133eSJeenu Viswambharan if (resume) { 512*b7cb133eSJeenu Viswambharan /* 513*b7cb133eSJeenu Viswambharan * Complete-and-resume call. Prepare the Non-secure context 514*b7cb133eSJeenu Viswambharan * (currently active) for complete and resume. 515*b7cb133eSJeenu Viswambharan */ 516*b7cb133eSJeenu Viswambharan cm_set_elr_spsr_el3(NON_SECURE, pc, SPSR_64(client_el, 517*b7cb133eSJeenu Viswambharan MODE_SP_ELX, DISABLE_ALL_EXCEPTIONS)); 518*b7cb133eSJeenu Viswambharan 519*b7cb133eSJeenu Viswambharan /* 520*b7cb133eSJeenu Viswambharan * Make it look as if a synchronous exception were taken at the 521*b7cb133eSJeenu Viswambharan * supplied Non-secure resumption point. Populate SPSR and 522*b7cb133eSJeenu Viswambharan * ELR_ELx so that an ERET from there works as expected. 523*b7cb133eSJeenu Viswambharan * 524*b7cb133eSJeenu Viswambharan * The assumption is that the client, if necessary, would have 525*b7cb133eSJeenu Viswambharan * saved any live content in these registers before making this 526*b7cb133eSJeenu Viswambharan * call. 527*b7cb133eSJeenu Viswambharan */ 528*b7cb133eSJeenu Viswambharan if (client_el == MODE_EL2) { 529*b7cb133eSJeenu Viswambharan write_elr_el2(disp_ctx->elr_el3); 530*b7cb133eSJeenu Viswambharan write_spsr_el2(disp_ctx->spsr_el3); 531*b7cb133eSJeenu Viswambharan } else { 532*b7cb133eSJeenu Viswambharan /* EL1 */ 533*b7cb133eSJeenu Viswambharan write_elr_el1(disp_ctx->elr_el3); 534*b7cb133eSJeenu Viswambharan write_spsr_el1(disp_ctx->spsr_el3); 535*b7cb133eSJeenu Viswambharan } 536*b7cb133eSJeenu Viswambharan } 537*b7cb133eSJeenu Viswambharan 538*b7cb133eSJeenu Viswambharan /* 539*b7cb133eSJeenu Viswambharan * If the cause of dispatch originally interrupted the Secure world, and 540*b7cb133eSJeenu Viswambharan * if Non-secure world wasn't allowed to preempt Secure execution, 541*b7cb133eSJeenu Viswambharan * resume Secure. 542*b7cb133eSJeenu Viswambharan * 543*b7cb133eSJeenu Viswambharan * No need to save the Non-secure context ahead of a world switch: the 544*b7cb133eSJeenu Viswambharan * Non-secure context was fully saved before dispatch, and has been 545*b7cb133eSJeenu Viswambharan * returned to its pre-dispatch state. 546*b7cb133eSJeenu Viswambharan */ 547*b7cb133eSJeenu Viswambharan if ((disp_ctx->sec_state == SECURE) && 548*b7cb133eSJeenu Viswambharan (ehf_is_ns_preemption_allowed() == 0)) { 549*b7cb133eSJeenu Viswambharan restore_and_resume_secure_context(); 550*b7cb133eSJeenu Viswambharan } 551*b7cb133eSJeenu Viswambharan 552*b7cb133eSJeenu Viswambharan if ((map->ev_num == SDEI_EVENT_0) || is_map_bound(map)) { 553*b7cb133eSJeenu Viswambharan /* 554*b7cb133eSJeenu Viswambharan * The event was dispatched after receiving SDEI interrupt. With 555*b7cb133eSJeenu Viswambharan * the event handling completed, EOI the corresponding 556*b7cb133eSJeenu Viswambharan * interrupt. 557*b7cb133eSJeenu Viswambharan */ 558*b7cb133eSJeenu Viswambharan plat_ic_end_of_interrupt(disp_ctx->intr_raw); 559*b7cb133eSJeenu Viswambharan } 560*b7cb133eSJeenu Viswambharan 561*b7cb133eSJeenu Viswambharan if (is_event_shared(map)) 562*b7cb133eSJeenu Viswambharan sdei_map_unlock(map); 563*b7cb133eSJeenu Viswambharan 564*b7cb133eSJeenu Viswambharan return 0; 565*b7cb133eSJeenu Viswambharan } 566*b7cb133eSJeenu Viswambharan 567*b7cb133eSJeenu Viswambharan int sdei_event_context(void *handle, unsigned int param) 568*b7cb133eSJeenu Viswambharan { 569*b7cb133eSJeenu Viswambharan sdei_dispatch_context_t *disp_ctx; 570*b7cb133eSJeenu Viswambharan 571*b7cb133eSJeenu Viswambharan if (param >= SDEI_SAVED_GPREGS) 572*b7cb133eSJeenu Viswambharan return SDEI_EINVAL; 573*b7cb133eSJeenu Viswambharan 574*b7cb133eSJeenu Viswambharan /* Get outstanding dispatch on this CPU */ 575*b7cb133eSJeenu Viswambharan disp_ctx = get_outstanding_dispatch(); 576*b7cb133eSJeenu Viswambharan if (!disp_ctx) 577*b7cb133eSJeenu Viswambharan return SDEI_EDENY; 578*b7cb133eSJeenu Viswambharan 579*b7cb133eSJeenu Viswambharan assert(disp_ctx->map); 580*b7cb133eSJeenu Viswambharan 581*b7cb133eSJeenu Viswambharan if (!can_sdei_state_trans(get_event_entry(disp_ctx->map), DO_CONTEXT)) 582*b7cb133eSJeenu Viswambharan return SDEI_EDENY; 583*b7cb133eSJeenu Viswambharan 584*b7cb133eSJeenu Viswambharan /* 585*b7cb133eSJeenu Viswambharan * No locking is required for the Running status as this is the only CPU 586*b7cb133eSJeenu Viswambharan * which can complete the event 587*b7cb133eSJeenu Viswambharan */ 588*b7cb133eSJeenu Viswambharan 589*b7cb133eSJeenu Viswambharan return disp_ctx->x[param]; 590*b7cb133eSJeenu Viswambharan } 591