1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun // Miscellaneous Arm SMMU implementation and integration quirks
3*4882a593Smuzhiyun // Copyright (C) 2019 Arm Limited
4*4882a593Smuzhiyun
5*4882a593Smuzhiyun #define pr_fmt(fmt) "arm-smmu: " fmt
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #include <linux/bitfield.h>
8*4882a593Smuzhiyun #include <linux/of.h>
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include "arm-smmu.h"
11*4882a593Smuzhiyun
12*4882a593Smuzhiyun
arm_smmu_gr0_ns(int offset)13*4882a593Smuzhiyun static int arm_smmu_gr0_ns(int offset)
14*4882a593Smuzhiyun {
15*4882a593Smuzhiyun switch(offset) {
16*4882a593Smuzhiyun case ARM_SMMU_GR0_sCR0:
17*4882a593Smuzhiyun case ARM_SMMU_GR0_sACR:
18*4882a593Smuzhiyun case ARM_SMMU_GR0_sGFSR:
19*4882a593Smuzhiyun case ARM_SMMU_GR0_sGFSYNR0:
20*4882a593Smuzhiyun case ARM_SMMU_GR0_sGFSYNR1:
21*4882a593Smuzhiyun case ARM_SMMU_GR0_sGFSYNR2:
22*4882a593Smuzhiyun return offset + 0x400;
23*4882a593Smuzhiyun default:
24*4882a593Smuzhiyun return offset;
25*4882a593Smuzhiyun }
26*4882a593Smuzhiyun }
27*4882a593Smuzhiyun
arm_smmu_read_ns(struct arm_smmu_device * smmu,int page,int offset)28*4882a593Smuzhiyun static u32 arm_smmu_read_ns(struct arm_smmu_device *smmu, int page,
29*4882a593Smuzhiyun int offset)
30*4882a593Smuzhiyun {
31*4882a593Smuzhiyun if (page == ARM_SMMU_GR0)
32*4882a593Smuzhiyun offset = arm_smmu_gr0_ns(offset);
33*4882a593Smuzhiyun return readl_relaxed(arm_smmu_page(smmu, page) + offset);
34*4882a593Smuzhiyun }
35*4882a593Smuzhiyun
arm_smmu_write_ns(struct arm_smmu_device * smmu,int page,int offset,u32 val)36*4882a593Smuzhiyun static void arm_smmu_write_ns(struct arm_smmu_device *smmu, int page,
37*4882a593Smuzhiyun int offset, u32 val)
38*4882a593Smuzhiyun {
39*4882a593Smuzhiyun if (page == ARM_SMMU_GR0)
40*4882a593Smuzhiyun offset = arm_smmu_gr0_ns(offset);
41*4882a593Smuzhiyun writel_relaxed(val, arm_smmu_page(smmu, page) + offset);
42*4882a593Smuzhiyun }
43*4882a593Smuzhiyun
44*4882a593Smuzhiyun /* Since we don't care for sGFAR, we can do without 64-bit accessors */
45*4882a593Smuzhiyun static const struct arm_smmu_impl calxeda_impl = {
46*4882a593Smuzhiyun .read_reg = arm_smmu_read_ns,
47*4882a593Smuzhiyun .write_reg = arm_smmu_write_ns,
48*4882a593Smuzhiyun };
49*4882a593Smuzhiyun
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun struct cavium_smmu {
52*4882a593Smuzhiyun struct arm_smmu_device smmu;
53*4882a593Smuzhiyun u32 id_base;
54*4882a593Smuzhiyun };
55*4882a593Smuzhiyun
cavium_cfg_probe(struct arm_smmu_device * smmu)56*4882a593Smuzhiyun static int cavium_cfg_probe(struct arm_smmu_device *smmu)
57*4882a593Smuzhiyun {
58*4882a593Smuzhiyun static atomic_t context_count = ATOMIC_INIT(0);
59*4882a593Smuzhiyun struct cavium_smmu *cs = container_of(smmu, struct cavium_smmu, smmu);
60*4882a593Smuzhiyun /*
61*4882a593Smuzhiyun * Cavium CN88xx erratum #27704.
62*4882a593Smuzhiyun * Ensure ASID and VMID allocation is unique across all SMMUs in
63*4882a593Smuzhiyun * the system.
64*4882a593Smuzhiyun */
65*4882a593Smuzhiyun cs->id_base = atomic_fetch_add(smmu->num_context_banks, &context_count);
66*4882a593Smuzhiyun dev_notice(smmu->dev, "\tenabling workaround for Cavium erratum 27704\n");
67*4882a593Smuzhiyun
68*4882a593Smuzhiyun return 0;
69*4882a593Smuzhiyun }
70*4882a593Smuzhiyun
cavium_init_context(struct arm_smmu_domain * smmu_domain,struct io_pgtable_cfg * pgtbl_cfg,struct device * dev)71*4882a593Smuzhiyun static int cavium_init_context(struct arm_smmu_domain *smmu_domain,
72*4882a593Smuzhiyun struct io_pgtable_cfg *pgtbl_cfg, struct device *dev)
73*4882a593Smuzhiyun {
74*4882a593Smuzhiyun struct cavium_smmu *cs = container_of(smmu_domain->smmu,
75*4882a593Smuzhiyun struct cavium_smmu, smmu);
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun if (smmu_domain->stage == ARM_SMMU_DOMAIN_S2)
78*4882a593Smuzhiyun smmu_domain->cfg.vmid += cs->id_base;
79*4882a593Smuzhiyun else
80*4882a593Smuzhiyun smmu_domain->cfg.asid += cs->id_base;
81*4882a593Smuzhiyun
82*4882a593Smuzhiyun return 0;
83*4882a593Smuzhiyun }
84*4882a593Smuzhiyun
85*4882a593Smuzhiyun static const struct arm_smmu_impl cavium_impl = {
86*4882a593Smuzhiyun .cfg_probe = cavium_cfg_probe,
87*4882a593Smuzhiyun .init_context = cavium_init_context,
88*4882a593Smuzhiyun };
89*4882a593Smuzhiyun
cavium_smmu_impl_init(struct arm_smmu_device * smmu)90*4882a593Smuzhiyun static struct arm_smmu_device *cavium_smmu_impl_init(struct arm_smmu_device *smmu)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun struct cavium_smmu *cs;
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun cs = devm_kzalloc(smmu->dev, sizeof(*cs), GFP_KERNEL);
95*4882a593Smuzhiyun if (!cs)
96*4882a593Smuzhiyun return ERR_PTR(-ENOMEM);
97*4882a593Smuzhiyun
98*4882a593Smuzhiyun cs->smmu = *smmu;
99*4882a593Smuzhiyun cs->smmu.impl = &cavium_impl;
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun devm_kfree(smmu->dev, smmu);
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun return &cs->smmu;
104*4882a593Smuzhiyun }
105*4882a593Smuzhiyun
106*4882a593Smuzhiyun
107*4882a593Smuzhiyun #define ARM_MMU500_ACTLR_CPRE (1 << 1)
108*4882a593Smuzhiyun
109*4882a593Smuzhiyun #define ARM_MMU500_ACR_CACHE_LOCK (1 << 26)
110*4882a593Smuzhiyun #define ARM_MMU500_ACR_S2CRB_TLBEN (1 << 10)
111*4882a593Smuzhiyun #define ARM_MMU500_ACR_SMTNMB_TLBEN (1 << 8)
112*4882a593Smuzhiyun
arm_mmu500_reset(struct arm_smmu_device * smmu)113*4882a593Smuzhiyun int arm_mmu500_reset(struct arm_smmu_device *smmu)
114*4882a593Smuzhiyun {
115*4882a593Smuzhiyun u32 reg, major;
116*4882a593Smuzhiyun int i;
117*4882a593Smuzhiyun /*
118*4882a593Smuzhiyun * On MMU-500 r2p0 onwards we need to clear ACR.CACHE_LOCK before
119*4882a593Smuzhiyun * writes to the context bank ACTLRs will stick. And we just hope that
120*4882a593Smuzhiyun * Secure has also cleared SACR.CACHE_LOCK for this to take effect...
121*4882a593Smuzhiyun */
122*4882a593Smuzhiyun reg = arm_smmu_gr0_read(smmu, ARM_SMMU_GR0_ID7);
123*4882a593Smuzhiyun major = FIELD_GET(ARM_SMMU_ID7_MAJOR, reg);
124*4882a593Smuzhiyun reg = arm_smmu_gr0_read(smmu, ARM_SMMU_GR0_sACR);
125*4882a593Smuzhiyun if (major >= 2)
126*4882a593Smuzhiyun reg &= ~ARM_MMU500_ACR_CACHE_LOCK;
127*4882a593Smuzhiyun /*
128*4882a593Smuzhiyun * Allow unmatched Stream IDs to allocate bypass
129*4882a593Smuzhiyun * TLB entries for reduced latency.
130*4882a593Smuzhiyun */
131*4882a593Smuzhiyun reg |= ARM_MMU500_ACR_SMTNMB_TLBEN | ARM_MMU500_ACR_S2CRB_TLBEN;
132*4882a593Smuzhiyun arm_smmu_gr0_write(smmu, ARM_SMMU_GR0_sACR, reg);
133*4882a593Smuzhiyun
134*4882a593Smuzhiyun /*
135*4882a593Smuzhiyun * Disable MMU-500's not-particularly-beneficial next-page
136*4882a593Smuzhiyun * prefetcher for the sake of errata #841119 and #826419.
137*4882a593Smuzhiyun */
138*4882a593Smuzhiyun for (i = 0; i < smmu->num_context_banks; ++i) {
139*4882a593Smuzhiyun reg = arm_smmu_cb_read(smmu, i, ARM_SMMU_CB_ACTLR);
140*4882a593Smuzhiyun reg &= ~ARM_MMU500_ACTLR_CPRE;
141*4882a593Smuzhiyun arm_smmu_cb_write(smmu, i, ARM_SMMU_CB_ACTLR, reg);
142*4882a593Smuzhiyun }
143*4882a593Smuzhiyun
144*4882a593Smuzhiyun return 0;
145*4882a593Smuzhiyun }
146*4882a593Smuzhiyun
147*4882a593Smuzhiyun static const struct arm_smmu_impl arm_mmu500_impl = {
148*4882a593Smuzhiyun .reset = arm_mmu500_reset,
149*4882a593Smuzhiyun };
150*4882a593Smuzhiyun
mrvl_mmu500_readq(struct arm_smmu_device * smmu,int page,int off)151*4882a593Smuzhiyun static u64 mrvl_mmu500_readq(struct arm_smmu_device *smmu, int page, int off)
152*4882a593Smuzhiyun {
153*4882a593Smuzhiyun /*
154*4882a593Smuzhiyun * Marvell Armada-AP806 erratum #582743.
155*4882a593Smuzhiyun * Split all the readq to double readl
156*4882a593Smuzhiyun */
157*4882a593Smuzhiyun return hi_lo_readq_relaxed(arm_smmu_page(smmu, page) + off);
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun
mrvl_mmu500_writeq(struct arm_smmu_device * smmu,int page,int off,u64 val)160*4882a593Smuzhiyun static void mrvl_mmu500_writeq(struct arm_smmu_device *smmu, int page, int off,
161*4882a593Smuzhiyun u64 val)
162*4882a593Smuzhiyun {
163*4882a593Smuzhiyun /*
164*4882a593Smuzhiyun * Marvell Armada-AP806 erratum #582743.
165*4882a593Smuzhiyun * Split all the writeq to double writel
166*4882a593Smuzhiyun */
167*4882a593Smuzhiyun hi_lo_writeq_relaxed(val, arm_smmu_page(smmu, page) + off);
168*4882a593Smuzhiyun }
169*4882a593Smuzhiyun
mrvl_mmu500_cfg_probe(struct arm_smmu_device * smmu)170*4882a593Smuzhiyun static int mrvl_mmu500_cfg_probe(struct arm_smmu_device *smmu)
171*4882a593Smuzhiyun {
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun /*
174*4882a593Smuzhiyun * Armada-AP806 erratum #582743.
175*4882a593Smuzhiyun * Hide the SMMU_IDR2.PTFSv8 fields to sidestep the AArch64
176*4882a593Smuzhiyun * formats altogether and allow using 32 bits access on the
177*4882a593Smuzhiyun * interconnect.
178*4882a593Smuzhiyun */
179*4882a593Smuzhiyun smmu->features &= ~(ARM_SMMU_FEAT_FMT_AARCH64_4K |
180*4882a593Smuzhiyun ARM_SMMU_FEAT_FMT_AARCH64_16K |
181*4882a593Smuzhiyun ARM_SMMU_FEAT_FMT_AARCH64_64K);
182*4882a593Smuzhiyun
183*4882a593Smuzhiyun return 0;
184*4882a593Smuzhiyun }
185*4882a593Smuzhiyun
186*4882a593Smuzhiyun static const struct arm_smmu_impl mrvl_mmu500_impl = {
187*4882a593Smuzhiyun .read_reg64 = mrvl_mmu500_readq,
188*4882a593Smuzhiyun .write_reg64 = mrvl_mmu500_writeq,
189*4882a593Smuzhiyun .cfg_probe = mrvl_mmu500_cfg_probe,
190*4882a593Smuzhiyun .reset = arm_mmu500_reset,
191*4882a593Smuzhiyun };
192*4882a593Smuzhiyun
193*4882a593Smuzhiyun
arm_smmu_impl_init(struct arm_smmu_device * smmu)194*4882a593Smuzhiyun struct arm_smmu_device *arm_smmu_impl_init(struct arm_smmu_device *smmu)
195*4882a593Smuzhiyun {
196*4882a593Smuzhiyun const struct device_node *np = smmu->dev->of_node;
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun /*
199*4882a593Smuzhiyun * Set the impl for model-specific implementation quirks first,
200*4882a593Smuzhiyun * such that platform integration quirks can pick it up and
201*4882a593Smuzhiyun * inherit from it if necessary.
202*4882a593Smuzhiyun */
203*4882a593Smuzhiyun switch (smmu->model) {
204*4882a593Smuzhiyun case ARM_MMU500:
205*4882a593Smuzhiyun smmu->impl = &arm_mmu500_impl;
206*4882a593Smuzhiyun break;
207*4882a593Smuzhiyun case CAVIUM_SMMUV2:
208*4882a593Smuzhiyun return cavium_smmu_impl_init(smmu);
209*4882a593Smuzhiyun default:
210*4882a593Smuzhiyun break;
211*4882a593Smuzhiyun }
212*4882a593Smuzhiyun
213*4882a593Smuzhiyun /* This is implicitly MMU-400 */
214*4882a593Smuzhiyun if (of_property_read_bool(np, "calxeda,smmu-secure-config-access"))
215*4882a593Smuzhiyun smmu->impl = &calxeda_impl;
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun if (of_device_is_compatible(np, "nvidia,tegra194-smmu"))
218*4882a593Smuzhiyun return nvidia_smmu_impl_init(smmu);
219*4882a593Smuzhiyun
220*4882a593Smuzhiyun if (of_device_is_compatible(np, "qcom,sdm845-smmu-500") ||
221*4882a593Smuzhiyun of_device_is_compatible(np, "qcom,sc7180-smmu-500") ||
222*4882a593Smuzhiyun of_device_is_compatible(np, "qcom,sm8150-smmu-500") ||
223*4882a593Smuzhiyun of_device_is_compatible(np, "qcom,sm8250-smmu-500"))
224*4882a593Smuzhiyun return qcom_smmu_impl_init(smmu);
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun if (of_device_is_compatible(smmu->dev->of_node, "qcom,adreno-smmu"))
227*4882a593Smuzhiyun return qcom_adreno_smmu_impl_init(smmu);
228*4882a593Smuzhiyun
229*4882a593Smuzhiyun if (of_device_is_compatible(np, "marvell,ap806-smmu-500"))
230*4882a593Smuzhiyun smmu->impl = &mrvl_mmu500_impl;
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun return smmu;
233*4882a593Smuzhiyun }
234