6 #include <linux/irqchip/arm-gic.h>
8 #include <linux/kvm_host.h>
9 #include <linux/nospec.h>
26 gpa_t addr,
unsigned int len)
28 struct vgic_dist *vgic = &vcpu->kvm->arch.vgic;
31 switch (addr & 0x0c) {
33 value = vgic->
enabled ? GICD_ENABLE : 0;
37 value = (value >> 5) - 1;
38 value |= (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5;
53 gpa_t addr,
unsigned int len,
56 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
57 bool was_enabled = dist->
enabled;
59 switch (addr & 0x0c) {
61 dist->
enabled = val & GICD_ENABLE;
62 if (!was_enabled && dist->
enabled)
73 gpa_t addr,
unsigned int len,
76 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
79 switch (addr & 0x0c) {
82 if ((reg ^ val) & ~GICD_IIDR_REVISION_MASK)
94 reg = FIELD_GET(GICD_IIDR_REVISION_MASK, reg);
98 vcpu->kvm->arch.vgic.v2_groups_user_writable =
true;
111 gpa_t addr,
unsigned int len,
114 if (vcpu->kvm->arch.vgic.v2_groups_user_writable)
121 gpa_t addr,
unsigned int len,
124 int nr_vcpus = atomic_read(&source_vcpu->kvm->online_vcpus);
125 int intid = val & 0xf;
126 int targets = (val >> 16) & 0xff;
127 int mode = (val >> 24) & 0x03;
128 struct kvm_vcpu *vcpu;
129 unsigned long flags, c;
135 targets = (1U << nr_vcpus) - 1;
136 targets &= ~(1U << source_vcpu->vcpu_id);
139 targets = (1U << source_vcpu->vcpu_id);
145 kvm_for_each_vcpu(c, vcpu, source_vcpu->kvm) {
153 raw_spin_lock_irqsave(&irq->
irq_lock, flags);
155 irq->
source |= 1U << source_vcpu->vcpu_id;
163 gpa_t addr,
unsigned int len)
169 for (i = 0; i < len; i++) {
172 val |= (u64)irq->
targets << (i * 8);
181 gpa_t addr,
unsigned int len,
185 u8 cpu_mask = GENMASK(atomic_read(&
vcpu->kvm->online_vcpus) - 1, 0);
193 for (i = 0; i < len; i++) {
197 raw_spin_lock_irqsave(&irq->
irq_lock, flags);
199 irq->
targets = (val >> (i * 8)) & cpu_mask;
203 raw_spin_unlock_irqrestore(&irq->
irq_lock, flags);
209 gpa_t addr,
unsigned int len)
211 u32
intid = addr & 0x0f;
215 for (i = 0; i < len; i++) {
218 val |= (u64)irq->
source << (i * 8);
226 gpa_t addr,
unsigned int len,
229 u32
intid = addr & 0x0f;
233 for (i = 0; i < len; i++) {
236 raw_spin_lock_irqsave(&irq->
irq_lock, flags);
238 irq->
source &= ~((val >> (i * 8)) & 0xff);
242 raw_spin_unlock_irqrestore(&irq->
irq_lock, flags);
248 gpa_t addr,
unsigned int len,
251 u32
intid = addr & 0x0f;
255 for (i = 0; i < len; i++) {
258 raw_spin_lock_irqsave(&irq->
irq_lock, flags);
260 irq->
source |= (val >> (i * 8)) & 0xff;
266 raw_spin_unlock_irqrestore(&irq->
irq_lock, flags);
272 #define GICC_ARCH_VERSION_V2 0x2
276 gpa_t addr,
unsigned int len)
283 switch (addr & 0xff) {
285 val = vmcr.
grpen0 << GIC_CPU_CTRL_EnableGrp0_SHIFT;
286 val |= vmcr.
grpen1 << GIC_CPU_CTRL_EnableGrp1_SHIFT;
287 val |= vmcr.
ackctl << GIC_CPU_CTRL_AckCtl_SHIFT;
288 val |= vmcr.
fiqen << GIC_CPU_CTRL_FIQEn_SHIFT;
289 val |= vmcr.
cbpr << GIC_CPU_CTRL_CBPR_SHIFT;
290 val |= vmcr.
eoim << GIC_CPU_CTRL_EOImodeNS_SHIFT;
293 case GIC_CPU_PRIMASK:
301 val = (vmcr.
pmr & GICV_PMR_PRIORITY_MASK) >>
302 GICV_PMR_PRIORITY_SHIFT;
304 case GIC_CPU_BINPOINT:
307 case GIC_CPU_ALIAS_BINPOINT:
323 gpa_t addr,
unsigned int len,
330 switch (addr & 0xff) {
332 vmcr.
grpen0 = !!(val & GIC_CPU_CTRL_EnableGrp0);
333 vmcr.
grpen1 = !!(val & GIC_CPU_CTRL_EnableGrp1);
334 vmcr.
ackctl = !!(val & GIC_CPU_CTRL_AckCtl);
335 vmcr.
fiqen = !!(val & GIC_CPU_CTRL_FIQEn);
336 vmcr.
cbpr = !!(val & GIC_CPU_CTRL_CBPR);
337 vmcr.
eoim = !!(val & GIC_CPU_CTRL_EOImodeNS);
340 case GIC_CPU_PRIMASK:
348 vmcr.
pmr = (val << GICV_PMR_PRIORITY_SHIFT) &
349 GICV_PMR_PRIORITY_MASK;
351 case GIC_CPU_BINPOINT:
354 case GIC_CPU_ALIAS_BINPOINT:
363 gpa_t addr,
unsigned int len)
367 n = (addr >> 2) & 0x3;
373 return vcpu->arch.vgic_cpu.vgic_v2.vgic_apr;
380 n = array_index_nospec(n, 4);
388 gpa_t addr,
unsigned int len,
393 n = (addr >> 2) & 0x3;
399 vcpu->arch.vgic_cpu.vgic_v2.vgic_apr = val;
406 n = array_index_nospec(n, 4);
502 struct kvm_vcpu *vcpu;
510 vcpu = reg_attr.
vcpu;
511 addr = reg_attr.
addr;
513 switch (attr->group) {
514 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
519 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
540 int offset, u32 *val)
552 int offset, u32 *val)
#define KVM_VGIC_IMP_REV_3
struct vgic_global kvm_vgic_global_state
#define KVM_VGIC_IMP_REV_2
#define VGIC_NR_PRIVATE_IRQS
static void kvm_iodevice_init(struct kvm_io_device *dev, const struct kvm_io_device_ops *ops)
const struct vgic_register_region * regions
struct kvm_vcpu * target_vcpu
int vgic_v2_parse_attr(struct kvm_device *dev, struct kvm_device_attr *attr, struct vgic_reg_attr *reg_attr)
int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write, int offset, u32 *val)
static unsigned long vgic_mmio_read_v2_misc(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
static void vgic_mmio_write_v2_misc(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
static void vgic_mmio_write_sgipends(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
static void vgic_mmio_write_vcpuif(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
int vgic_v2_has_attr_regs(struct kvm_device *dev, struct kvm_device_attr *attr)
int vgic_v2_cpuif_uaccess(struct kvm_vcpu *vcpu, bool is_write, int offset, u32 *val)
static void vgic_mmio_write_sgipendc(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
static void vgic_mmio_write_sgir(struct kvm_vcpu *source_vcpu, gpa_t addr, unsigned int len, unsigned long val)
static void vgic_mmio_write_apr(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
static unsigned long vgic_mmio_read_sgipend(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
#define GICC_ARCH_VERSION_V2
static int vgic_mmio_uaccess_write_v2_misc(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
unsigned int vgic_v2_init_dist_iodev(struct vgic_io_device *dev)
static int vgic_mmio_uaccess_write_v2_group(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
static const struct vgic_register_region vgic_v2_cpu_registers[]
static unsigned long vgic_mmio_read_vcpuif(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
static unsigned long vgic_mmio_read_apr(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
static const struct vgic_register_region vgic_v2_dist_registers[]
static void vgic_mmio_write_target(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
static unsigned long vgic_mmio_read_target(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_mmio_write_senable(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
unsigned long vgic_uaccess_read_pending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
int vgic_uaccess_write_spending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_mmio_write_spending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
void vgic_mmio_write_priority(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
int vgic_uaccess_write_cpending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_mmio_write_config(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len)
int vgic_uaccess_write_cenable(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
int vgic_uaccess_write_senable(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev, bool is_write, int offset, u32 *val)
const struct kvm_io_device_ops kvm_io_gic_ops
void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len, unsigned long val)
const struct vgic_register_region * vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev, gpa_t addr, int len)
#define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc)
#define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc)
#define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc)
#define VGIC_ADDR_TO_INTID(addr, bits)
#define VGIC_ACCESS_32bit
bool vgic_queue_irq_unlock(struct kvm *kvm, struct vgic_irq *irq, unsigned long flags)
struct vgic_irq * vgic_get_irq(struct kvm *kvm, struct kvm_vcpu *vcpu, u32 intid)
void vgic_put_irq(struct kvm *kvm, struct vgic_irq *irq)
void vgic_kick_vcpus(struct kvm *kvm)
static int vgic_v3_max_apr_idx(struct kvm_vcpu *vcpu)