1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0-only */ 2*4882a593Smuzhiyun /* 3*4882a593Smuzhiyun * Copyright (C) 2015, 2016 ARM Ltd. 4*4882a593Smuzhiyun */ 5*4882a593Smuzhiyun #ifndef __KVM_ARM_VGIC_MMIO_H__ 6*4882a593Smuzhiyun #define __KVM_ARM_VGIC_MMIO_H__ 7*4882a593Smuzhiyun 8*4882a593Smuzhiyun struct vgic_register_region { 9*4882a593Smuzhiyun unsigned int reg_offset; 10*4882a593Smuzhiyun unsigned int len; 11*4882a593Smuzhiyun unsigned int bits_per_irq; 12*4882a593Smuzhiyun unsigned int access_flags; 13*4882a593Smuzhiyun union { 14*4882a593Smuzhiyun unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr, 15*4882a593Smuzhiyun unsigned int len); 16*4882a593Smuzhiyun unsigned long (*its_read)(struct kvm *kvm, struct vgic_its *its, 17*4882a593Smuzhiyun gpa_t addr, unsigned int len); 18*4882a593Smuzhiyun }; 19*4882a593Smuzhiyun union { 20*4882a593Smuzhiyun void (*write)(struct kvm_vcpu *vcpu, gpa_t addr, 21*4882a593Smuzhiyun unsigned int len, unsigned long val); 22*4882a593Smuzhiyun void (*its_write)(struct kvm *kvm, struct vgic_its *its, 23*4882a593Smuzhiyun gpa_t addr, unsigned int len, 24*4882a593Smuzhiyun unsigned long val); 25*4882a593Smuzhiyun }; 26*4882a593Smuzhiyun unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr, 27*4882a593Smuzhiyun unsigned int len); 28*4882a593Smuzhiyun union { 29*4882a593Smuzhiyun int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr, 30*4882a593Smuzhiyun unsigned int len, unsigned long val); 31*4882a593Smuzhiyun int (*uaccess_its_write)(struct kvm *kvm, struct vgic_its *its, 32*4882a593Smuzhiyun gpa_t addr, unsigned int len, 33*4882a593Smuzhiyun unsigned long val); 34*4882a593Smuzhiyun }; 35*4882a593Smuzhiyun }; 36*4882a593Smuzhiyun 37*4882a593Smuzhiyun extern struct kvm_io_device_ops kvm_io_gic_ops; 38*4882a593Smuzhiyun 39*4882a593Smuzhiyun #define VGIC_ACCESS_8bit 1 40*4882a593Smuzhiyun #define VGIC_ACCESS_32bit 2 41*4882a593Smuzhiyun #define VGIC_ACCESS_64bit 4 42*4882a593Smuzhiyun 43*4882a593Smuzhiyun /* 44*4882a593Smuzhiyun * Generate a mask that covers the number of bytes required to address 45*4882a593Smuzhiyun * up to 1024 interrupts, each represented by <bits> bits. This assumes 46*4882a593Smuzhiyun * that <bits> is a power of two. 47*4882a593Smuzhiyun */ 48*4882a593Smuzhiyun #define VGIC_ADDR_IRQ_MASK(bits) (((bits) * 1024 / 8) - 1) 49*4882a593Smuzhiyun 50*4882a593Smuzhiyun /* 51*4882a593Smuzhiyun * (addr & mask) gives us the _byte_ offset for the INT ID. 52*4882a593Smuzhiyun * We multiply this by 8 the get the _bit_ offset, then divide this by 53*4882a593Smuzhiyun * the number of bits to learn the actual INT ID. 54*4882a593Smuzhiyun * But instead of a division (which requires a "long long div" implementation), 55*4882a593Smuzhiyun * we shift by the binary logarithm of <bits>. 56*4882a593Smuzhiyun * This assumes that <bits> is a power of two. 57*4882a593Smuzhiyun */ 58*4882a593Smuzhiyun #define VGIC_ADDR_TO_INTID(addr, bits) (((addr) & VGIC_ADDR_IRQ_MASK(bits)) * \ 59*4882a593Smuzhiyun 8 >> ilog2(bits)) 60*4882a593Smuzhiyun 61*4882a593Smuzhiyun /* 62*4882a593Smuzhiyun * Some VGIC registers store per-IRQ information, with a different number 63*4882a593Smuzhiyun * of bits per IRQ. For those registers this macro is used. 64*4882a593Smuzhiyun * The _WITH_LENGTH version instantiates registers with a fixed length 65*4882a593Smuzhiyun * and is mutually exclusive with the _PER_IRQ version. 66*4882a593Smuzhiyun */ 67*4882a593Smuzhiyun #define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc) \ 68*4882a593Smuzhiyun { \ 69*4882a593Smuzhiyun .reg_offset = off, \ 70*4882a593Smuzhiyun .bits_per_irq = bpi, \ 71*4882a593Smuzhiyun .len = bpi * 1024 / 8, \ 72*4882a593Smuzhiyun .access_flags = acc, \ 73*4882a593Smuzhiyun .read = rd, \ 74*4882a593Smuzhiyun .write = wr, \ 75*4882a593Smuzhiyun .uaccess_read = ur, \ 76*4882a593Smuzhiyun .uaccess_write = uw, \ 77*4882a593Smuzhiyun } 78*4882a593Smuzhiyun 79*4882a593Smuzhiyun #define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc) \ 80*4882a593Smuzhiyun { \ 81*4882a593Smuzhiyun .reg_offset = off, \ 82*4882a593Smuzhiyun .bits_per_irq = 0, \ 83*4882a593Smuzhiyun .len = length, \ 84*4882a593Smuzhiyun .access_flags = acc, \ 85*4882a593Smuzhiyun .read = rd, \ 86*4882a593Smuzhiyun .write = wr, \ 87*4882a593Smuzhiyun } 88*4882a593Smuzhiyun 89*4882a593Smuzhiyun #define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc) \ 90*4882a593Smuzhiyun { \ 91*4882a593Smuzhiyun .reg_offset = off, \ 92*4882a593Smuzhiyun .bits_per_irq = 0, \ 93*4882a593Smuzhiyun .len = length, \ 94*4882a593Smuzhiyun .access_flags = acc, \ 95*4882a593Smuzhiyun .read = rd, \ 96*4882a593Smuzhiyun .write = wr, \ 97*4882a593Smuzhiyun .uaccess_read = urd, \ 98*4882a593Smuzhiyun .uaccess_write = uwr, \ 99*4882a593Smuzhiyun } 100*4882a593Smuzhiyun 101*4882a593Smuzhiyun unsigned long vgic_data_mmio_bus_to_host(const void *val, unsigned int len); 102*4882a593Smuzhiyun 103*4882a593Smuzhiyun void vgic_data_host_to_mmio_bus(void *buf, unsigned int len, 104*4882a593Smuzhiyun unsigned long data); 105*4882a593Smuzhiyun 106*4882a593Smuzhiyun unsigned long extract_bytes(u64 data, unsigned int offset, 107*4882a593Smuzhiyun unsigned int num); 108*4882a593Smuzhiyun 109*4882a593Smuzhiyun u64 update_64bit_reg(u64 reg, unsigned int offset, unsigned int len, 110*4882a593Smuzhiyun unsigned long val); 111*4882a593Smuzhiyun 112*4882a593Smuzhiyun unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu, 113*4882a593Smuzhiyun gpa_t addr, unsigned int len); 114*4882a593Smuzhiyun 115*4882a593Smuzhiyun unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu, 116*4882a593Smuzhiyun gpa_t addr, unsigned int len); 117*4882a593Smuzhiyun 118*4882a593Smuzhiyun void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, 119*4882a593Smuzhiyun unsigned int len, unsigned long val); 120*4882a593Smuzhiyun 121*4882a593Smuzhiyun int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, 122*4882a593Smuzhiyun unsigned int len, unsigned long val); 123*4882a593Smuzhiyun 124*4882a593Smuzhiyun unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr, 125*4882a593Smuzhiyun unsigned int len); 126*4882a593Smuzhiyun 127*4882a593Smuzhiyun void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr, 128*4882a593Smuzhiyun unsigned int len, unsigned long val); 129*4882a593Smuzhiyun 130*4882a593Smuzhiyun unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu, 131*4882a593Smuzhiyun gpa_t addr, unsigned int len); 132*4882a593Smuzhiyun 133*4882a593Smuzhiyun void vgic_mmio_write_senable(struct kvm_vcpu *vcpu, 134*4882a593Smuzhiyun gpa_t addr, unsigned int len, 135*4882a593Smuzhiyun unsigned long val); 136*4882a593Smuzhiyun 137*4882a593Smuzhiyun void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu, 138*4882a593Smuzhiyun gpa_t addr, unsigned int len, 139*4882a593Smuzhiyun unsigned long val); 140*4882a593Smuzhiyun 141*4882a593Smuzhiyun int vgic_uaccess_write_senable(struct kvm_vcpu *vcpu, 142*4882a593Smuzhiyun gpa_t addr, unsigned int len, 143*4882a593Smuzhiyun unsigned long val); 144*4882a593Smuzhiyun 145*4882a593Smuzhiyun int vgic_uaccess_write_cenable(struct kvm_vcpu *vcpu, 146*4882a593Smuzhiyun gpa_t addr, unsigned int len, 147*4882a593Smuzhiyun unsigned long val); 148*4882a593Smuzhiyun 149*4882a593Smuzhiyun unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu, 150*4882a593Smuzhiyun gpa_t addr, unsigned int len); 151*4882a593Smuzhiyun 152*4882a593Smuzhiyun unsigned long vgic_uaccess_read_pending(struct kvm_vcpu *vcpu, 153*4882a593Smuzhiyun gpa_t addr, unsigned int len); 154*4882a593Smuzhiyun 155*4882a593Smuzhiyun void vgic_mmio_write_spending(struct kvm_vcpu *vcpu, 156*4882a593Smuzhiyun gpa_t addr, unsigned int len, 157*4882a593Smuzhiyun unsigned long val); 158*4882a593Smuzhiyun 159*4882a593Smuzhiyun void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu, 160*4882a593Smuzhiyun gpa_t addr, unsigned int len, 161*4882a593Smuzhiyun unsigned long val); 162*4882a593Smuzhiyun 163*4882a593Smuzhiyun int vgic_uaccess_write_spending(struct kvm_vcpu *vcpu, 164*4882a593Smuzhiyun gpa_t addr, unsigned int len, 165*4882a593Smuzhiyun unsigned long val); 166*4882a593Smuzhiyun 167*4882a593Smuzhiyun int vgic_uaccess_write_cpending(struct kvm_vcpu *vcpu, 168*4882a593Smuzhiyun gpa_t addr, unsigned int len, 169*4882a593Smuzhiyun unsigned long val); 170*4882a593Smuzhiyun 171*4882a593Smuzhiyun unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu, 172*4882a593Smuzhiyun gpa_t addr, unsigned int len); 173*4882a593Smuzhiyun 174*4882a593Smuzhiyun unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu, 175*4882a593Smuzhiyun gpa_t addr, unsigned int len); 176*4882a593Smuzhiyun 177*4882a593Smuzhiyun void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu, 178*4882a593Smuzhiyun gpa_t addr, unsigned int len, 179*4882a593Smuzhiyun unsigned long val); 180*4882a593Smuzhiyun 181*4882a593Smuzhiyun void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu, 182*4882a593Smuzhiyun gpa_t addr, unsigned int len, 183*4882a593Smuzhiyun unsigned long val); 184*4882a593Smuzhiyun 185*4882a593Smuzhiyun int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu, 186*4882a593Smuzhiyun gpa_t addr, unsigned int len, 187*4882a593Smuzhiyun unsigned long val); 188*4882a593Smuzhiyun 189*4882a593Smuzhiyun int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu, 190*4882a593Smuzhiyun gpa_t addr, unsigned int len, 191*4882a593Smuzhiyun unsigned long val); 192*4882a593Smuzhiyun 193*4882a593Smuzhiyun unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu, 194*4882a593Smuzhiyun gpa_t addr, unsigned int len); 195*4882a593Smuzhiyun 196*4882a593Smuzhiyun void vgic_mmio_write_priority(struct kvm_vcpu *vcpu, 197*4882a593Smuzhiyun gpa_t addr, unsigned int len, 198*4882a593Smuzhiyun unsigned long val); 199*4882a593Smuzhiyun 200*4882a593Smuzhiyun unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu, 201*4882a593Smuzhiyun gpa_t addr, unsigned int len); 202*4882a593Smuzhiyun 203*4882a593Smuzhiyun void vgic_mmio_write_config(struct kvm_vcpu *vcpu, 204*4882a593Smuzhiyun gpa_t addr, unsigned int len, 205*4882a593Smuzhiyun unsigned long val); 206*4882a593Smuzhiyun 207*4882a593Smuzhiyun int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev, 208*4882a593Smuzhiyun bool is_write, int offset, u32 *val); 209*4882a593Smuzhiyun 210*4882a593Smuzhiyun u64 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid); 211*4882a593Smuzhiyun 212*4882a593Smuzhiyun void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid, 213*4882a593Smuzhiyun const u64 val); 214*4882a593Smuzhiyun 215*4882a593Smuzhiyun unsigned int vgic_v2_init_dist_iodev(struct vgic_io_device *dev); 216*4882a593Smuzhiyun 217*4882a593Smuzhiyun unsigned int vgic_v3_init_dist_iodev(struct vgic_io_device *dev); 218*4882a593Smuzhiyun 219*4882a593Smuzhiyun u64 vgic_sanitise_outer_cacheability(u64 reg); 220*4882a593Smuzhiyun u64 vgic_sanitise_inner_cacheability(u64 reg); 221*4882a593Smuzhiyun u64 vgic_sanitise_shareability(u64 reg); 222*4882a593Smuzhiyun u64 vgic_sanitise_field(u64 reg, u64 field_mask, int field_shift, 223*4882a593Smuzhiyun u64 (*sanitise_fn)(u64)); 224*4882a593Smuzhiyun 225*4882a593Smuzhiyun /* Find the proper register handler entry given a certain address offset */ 226*4882a593Smuzhiyun const struct vgic_register_region * 227*4882a593Smuzhiyun vgic_find_mmio_region(const struct vgic_register_region *regions, 228*4882a593Smuzhiyun int nr_regions, unsigned int offset); 229*4882a593Smuzhiyun 230*4882a593Smuzhiyun #endif 231