11154586bSJeenu Viswambharan /* 270d849c1SVivek Gautam * Copyright (c) 2017-2024, ARM Limited and Contributors. All rights reserved. 31154586bSJeenu Viswambharan * 41154586bSJeenu Viswambharan * SPDX-License-Identifier: BSD-3-Clause 51154586bSJeenu Viswambharan */ 61154586bSJeenu Viswambharan 7ccd4d475SAlexei Fedorov #include <common/debug.h> 8c9263e62SDaniel Boulby #include <cdefs.h> 909d40e0eSAntonio Nino Diaz #include <drivers/arm/smmu_v3.h> 10620dd58bSDeepika Bhavnani #include <drivers/delay_timer.h> 1109d40e0eSAntonio Nino Diaz #include <lib/mmio.h> 1252a314afSOlivier Deprez #include <arch_features.h> 1309d40e0eSAntonio Nino Diaz 14ccd4d475SAlexei Fedorov /* SMMU poll number of retries */ 15620dd58bSDeepika Bhavnani #define SMMU_POLL_TIMEOUT_US U(1000) 161154586bSJeenu Viswambharan 176c5c5320SLucian Paul-Trifu static int smmuv3_poll(uintptr_t smmu_reg, uint32_t mask, 18ccd4d475SAlexei Fedorov uint32_t value) 191154586bSJeenu Viswambharan { 20620dd58bSDeepika Bhavnani uint32_t reg_val; 21620dd58bSDeepika Bhavnani uint64_t timeout; 221154586bSJeenu Viswambharan 23620dd58bSDeepika Bhavnani /* Set 1ms timeout value */ 24620dd58bSDeepika Bhavnani timeout = timeout_init_us(SMMU_POLL_TIMEOUT_US); 25ccd4d475SAlexei Fedorov do { 26ccd4d475SAlexei Fedorov reg_val = mmio_read_32(smmu_reg); 27ccd4d475SAlexei Fedorov if ((reg_val & mask) == value) 28ccd4d475SAlexei Fedorov return 0; 29620dd58bSDeepika Bhavnani } while (!timeout_elapsed(timeout)); 301154586bSJeenu Viswambharan 31620dd58bSDeepika Bhavnani ERROR("Timeout polling SMMUv3 register @%p\n", (void *)smmu_reg); 32ccd4d475SAlexei Fedorov ERROR("Read value 0x%x, expected 0x%x\n", reg_val, 33ccd4d475SAlexei Fedorov value == 0U ? reg_val & ~mask : reg_val | mask); 34ccd4d475SAlexei Fedorov return -1; 356d5f0631SAntonio Nino Diaz } 366d5f0631SAntonio Nino Diaz 371154586bSJeenu Viswambharan /* 381461ad9fSAlexei Fedorov * Abort all incoming transactions in order to implement a default 391461ad9fSAlexei Fedorov * deny policy on reset. 401461ad9fSAlexei Fedorov */ 411461ad9fSAlexei Fedorov int __init smmuv3_security_init(uintptr_t smmu_base) 421461ad9fSAlexei Fedorov { 431461ad9fSAlexei Fedorov /* Attribute update has completed when SMMU_(S)_GBPA.Update bit is 0 */ 441461ad9fSAlexei Fedorov if (smmuv3_poll(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE, 0U) != 0U) 451461ad9fSAlexei Fedorov return -1; 461461ad9fSAlexei Fedorov 471461ad9fSAlexei Fedorov /* 481461ad9fSAlexei Fedorov * SMMU_(S)_CR0 resets to zero with all streams bypassing the SMMU, 491461ad9fSAlexei Fedorov * so just abort all incoming transactions. 501461ad9fSAlexei Fedorov */ 511461ad9fSAlexei Fedorov mmio_setbits_32(smmu_base + SMMU_GBPA, 521461ad9fSAlexei Fedorov SMMU_GBPA_UPDATE | SMMU_GBPA_ABORT); 531461ad9fSAlexei Fedorov 541461ad9fSAlexei Fedorov if (smmuv3_poll(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE, 0U) != 0U) 551461ad9fSAlexei Fedorov return -1; 561461ad9fSAlexei Fedorov 571461ad9fSAlexei Fedorov /* Check if the SMMU supports secure state */ 581461ad9fSAlexei Fedorov if ((mmio_read_32(smmu_base + SMMU_S_IDR1) & 591461ad9fSAlexei Fedorov SMMU_S_IDR1_SECURE_IMPL) == 0U) 601461ad9fSAlexei Fedorov return 0; 611461ad9fSAlexei Fedorov 621461ad9fSAlexei Fedorov /* Abort all incoming secure transactions */ 631461ad9fSAlexei Fedorov if (smmuv3_poll(smmu_base + SMMU_S_GBPA, SMMU_S_GBPA_UPDATE, 0U) != 0U) 641461ad9fSAlexei Fedorov return -1; 651461ad9fSAlexei Fedorov 661461ad9fSAlexei Fedorov mmio_setbits_32(smmu_base + SMMU_S_GBPA, 671461ad9fSAlexei Fedorov SMMU_S_GBPA_UPDATE | SMMU_S_GBPA_ABORT); 681461ad9fSAlexei Fedorov 691461ad9fSAlexei Fedorov return smmuv3_poll(smmu_base + SMMU_S_GBPA, SMMU_S_GBPA_UPDATE, 0U); 701461ad9fSAlexei Fedorov } 711461ad9fSAlexei Fedorov 72a23710b4SVijayenthiran Subramaniam /* Initialize the SMMU by invalidating all secure caches and TLBs. */ 73c9263e62SDaniel Boulby int __init smmuv3_init(uintptr_t smmu_base) 741154586bSJeenu Viswambharan { 7570d849c1SVivek Gautam /* 7670d849c1SVivek Gautam * Initiate invalidation of secure caches and TLBs if the SMMU 7770d849c1SVivek Gautam * supports secure state. If not, it's implementation defined 7870d849c1SVivek Gautam * as to how SMMU_S_INIT register is accessed. 7970d849c1SVivek Gautam * As per Arm SMMUv3 specification the SMMU_S_INIT register in a SMMU 8070d849c1SVivek Gautam * with RME implementation has following properties: 8170d849c1SVivek Gautam * a) all SMMU registers that are specified to be accessible only in 8270d849c1SVivek Gautam * the Secure physical address space are additionally accessible in 8370d849c1SVivek Gautam * Root physical address space. 8470d849c1SVivek Gautam * b) as GPT information is permitted to be cached in a TLB, the 8570d849c1SVivek Gautam * SMMU_S_INIT.INV_ALL operation also invalidates all GPT information 8670d849c1SVivek Gautam * cached in TLBs. 8770d849c1SVivek Gautam * Additionally, it is Root firmware’s responsibility to write to 8870d849c1SVivek Gautam * INV_ALL before enabling SMMU_ROOT_CR0.{ACCESSEN,GPCEN}. 8970d849c1SVivek Gautam */ 9070d849c1SVivek Gautam mmio_write_32(smmu_base + SMMU_S_INIT, SMMU_S_INIT_INV_ALL); 9170d849c1SVivek Gautam 9270d849c1SVivek Gautam /* Wait for global invalidation operation to finish */ 9370d849c1SVivek Gautam if (smmuv3_poll(smmu_base + SMMU_S_INIT, 9470d849c1SVivek Gautam SMMU_S_INIT_INV_ALL, 0U) != 0) { 9570d849c1SVivek Gautam return -1; 9670d849c1SVivek Gautam } 9770d849c1SVivek Gautam 9852a314afSOlivier Deprez #if ENABLE_RME 9952a314afSOlivier Deprez 100aaaf2cc3SSona Mathew if (is_feat_rme_present()) { 10152a314afSOlivier Deprez if ((mmio_read_32(smmu_base + SMMU_ROOT_IDR0) & 10252a314afSOlivier Deprez SMMU_ROOT_IDR0_ROOT_IMPL) == 0U) { 10352a314afSOlivier Deprez WARN("Skip SMMU GPC configuration.\n"); 10452a314afSOlivier Deprez } else { 10552a314afSOlivier Deprez uint64_t gpccr_el3 = read_gpccr_el3(); 10652a314afSOlivier Deprez uint64_t gptbr_el3 = read_gptbr_el3(); 10752a314afSOlivier Deprez 10852a314afSOlivier Deprez /* SMMU_ROOT_GPT_BASE_CFG[16] is RES0. */ 10952a314afSOlivier Deprez gpccr_el3 &= ~(1UL << 16); 11052a314afSOlivier Deprez 11152a314afSOlivier Deprez /* 11252a314afSOlivier Deprez * TODO: SMMU_ROOT_GPT_BASE_CFG is 64b in the spec, 11352a314afSOlivier Deprez * but SMMU model only accepts 32b access. 11452a314afSOlivier Deprez */ 11552a314afSOlivier Deprez mmio_write_32(smmu_base + SMMU_ROOT_GPT_BASE_CFG, 11652a314afSOlivier Deprez gpccr_el3); 11752a314afSOlivier Deprez 11852a314afSOlivier Deprez /* 11952a314afSOlivier Deprez * pa_gpt_table_base[51:12] maps to GPTBR_EL3[39:0] 12052a314afSOlivier Deprez * whereas it maps to SMMU_ROOT_GPT_BASE[51:12] 12152a314afSOlivier Deprez * hence needs a 12 bit left shit. 12252a314afSOlivier Deprez */ 12352a314afSOlivier Deprez mmio_write_64(smmu_base + SMMU_ROOT_GPT_BASE, 12452a314afSOlivier Deprez gptbr_el3 << 12); 12552a314afSOlivier Deprez 12652a314afSOlivier Deprez /* 12752a314afSOlivier Deprez * GPCEN=1: All clients and SMMU-originated accesses, 12852a314afSOlivier Deprez * except GPT-walks, are subject to GPC. 129*8cc97242SOlivier Deprez * 130*8cc97242SOlivier Deprez * It is recommended to set GPCEN and wait for completion 131*8cc97242SOlivier Deprez * prior to setting ACCESSEN. 132*8cc97242SOlivier Deprez */ 133*8cc97242SOlivier Deprez mmio_setbits_32(smmu_base + SMMU_ROOT_CR0, 134*8cc97242SOlivier Deprez SMMU_ROOT_CR0_GPCEN); 135*8cc97242SOlivier Deprez 136*8cc97242SOlivier Deprez /* Poll for GPCEN ack bit. */ 137*8cc97242SOlivier Deprez if (smmuv3_poll(smmu_base + SMMU_ROOT_CR0ACK, 138*8cc97242SOlivier Deprez SMMU_ROOT_CR0_GPCEN, 139*8cc97242SOlivier Deprez SMMU_ROOT_CR0_GPCEN) != 0) { 140*8cc97242SOlivier Deprez WARN("Failed enabling SMMU GPC.\n"); 141*8cc97242SOlivier Deprez } 142*8cc97242SOlivier Deprez 143*8cc97242SOlivier Deprez /* 144*8cc97242SOlivier Deprez * ACCESSEN=1: SMMU- and client-originated accesses are 145*8cc97242SOlivier Deprez * not terminated by this mechanism. 14652a314afSOlivier Deprez */ 14752a314afSOlivier Deprez mmio_setbits_32(smmu_base + SMMU_ROOT_CR0, 14852a314afSOlivier Deprez SMMU_ROOT_CR0_GPCEN | 14952a314afSOlivier Deprez SMMU_ROOT_CR0_ACCESSEN); 15052a314afSOlivier Deprez 151*8cc97242SOlivier Deprez /* Poll for ACCESSEN ack bit. */ 15252a314afSOlivier Deprez if (smmuv3_poll(smmu_base + SMMU_ROOT_CR0ACK, 15352a314afSOlivier Deprez SMMU_ROOT_CR0_ACCESSEN, 15452a314afSOlivier Deprez SMMU_ROOT_CR0_ACCESSEN) != 0) { 155*8cc97242SOlivier Deprez WARN("Failed enabling SMMU ACCESS.\n"); 15652a314afSOlivier Deprez 15752a314afSOlivier Deprez /* 15852a314afSOlivier Deprez * Do not return in error, but fall back to 15952a314afSOlivier Deprez * invalidating all entries through the secure 16052a314afSOlivier Deprez * register file. 16152a314afSOlivier Deprez */ 16252a314afSOlivier Deprez } 16352a314afSOlivier Deprez } 16452a314afSOlivier Deprez } 16552a314afSOlivier Deprez 16652a314afSOlivier Deprez #endif /* ENABLE_RME */ 16752a314afSOlivier Deprez 16870d849c1SVivek Gautam return 0; 1691154586bSJeenu Viswambharan } 1706c5c5320SLucian Paul-Trifu 1716c5c5320SLucian Paul-Trifu int smmuv3_ns_set_abort_all(uintptr_t smmu_base) 1726c5c5320SLucian Paul-Trifu { 1736c5c5320SLucian Paul-Trifu /* Attribute update has completed when SMMU_GBPA.Update bit is 0 */ 1746c5c5320SLucian Paul-Trifu if (smmuv3_poll(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE, 0U) != 0U) { 1756c5c5320SLucian Paul-Trifu return -1; 1766c5c5320SLucian Paul-Trifu } 1776c5c5320SLucian Paul-Trifu 1786c5c5320SLucian Paul-Trifu /* 1796c5c5320SLucian Paul-Trifu * Set GBPA's ABORT bit. Other GBPA fields are presumably ignored then, 1806c5c5320SLucian Paul-Trifu * so simply preserve their value. 1816c5c5320SLucian Paul-Trifu */ 1826c5c5320SLucian Paul-Trifu mmio_setbits_32(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE | SMMU_GBPA_ABORT); 1836c5c5320SLucian Paul-Trifu if (smmuv3_poll(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE, 0U) != 0U) { 1846c5c5320SLucian Paul-Trifu return -1; 1856c5c5320SLucian Paul-Trifu } 1866c5c5320SLucian Paul-Trifu 1876c5c5320SLucian Paul-Trifu /* Disable the SMMU to engage the GBPA fields previously configured. */ 1886c5c5320SLucian Paul-Trifu mmio_clrbits_32(smmu_base + SMMU_CR0, SMMU_CR0_SMMUEN); 1896c5c5320SLucian Paul-Trifu if (smmuv3_poll(smmu_base + SMMU_CR0ACK, SMMU_CR0_SMMUEN, 0U) != 0U) { 1906c5c5320SLucian Paul-Trifu return -1; 1916c5c5320SLucian Paul-Trifu } 1926c5c5320SLucian Paul-Trifu 1936c5c5320SLucian Paul-Trifu return 0; 1946c5c5320SLucian Paul-Trifu } 195