231 lines
7.1 KiB
C
231 lines
7.1 KiB
C
|
/* SPDX-License-Identifier: GPL-2.0-only */
|
||
|
/*
|
||
|
* Copyright (C) 2015, 2016 ARM Ltd.
|
||
|
*/
|
||
|
#ifndef __KVM_ARM_VGIC_MMIO_H__
|
||
|
#define __KVM_ARM_VGIC_MMIO_H__
|
||
|
|
||
|
struct vgic_register_region {
|
||
|
unsigned int reg_offset;
|
||
|
unsigned int len;
|
||
|
unsigned int bits_per_irq;
|
||
|
unsigned int access_flags;
|
||
|
union {
|
||
|
unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len);
|
||
|
unsigned long (*its_read)(struct kvm *kvm, struct vgic_its *its,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
};
|
||
|
union {
|
||
|
void (*write)(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len, unsigned long val);
|
||
|
void (*its_write)(struct kvm *kvm, struct vgic_its *its,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
};
|
||
|
unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len);
|
||
|
union {
|
||
|
int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len, unsigned long val);
|
||
|
int (*uaccess_its_write)(struct kvm *kvm, struct vgic_its *its,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
};
|
||
|
};
|
||
|
|
||
|
extern const struct kvm_io_device_ops kvm_io_gic_ops;
|
||
|
|
||
|
#define VGIC_ACCESS_8bit 1
|
||
|
#define VGIC_ACCESS_32bit 2
|
||
|
#define VGIC_ACCESS_64bit 4
|
||
|
|
||
|
/*
|
||
|
* Generate a mask that covers the number of bytes required to address
|
||
|
* up to 1024 interrupts, each represented by <bits> bits. This assumes
|
||
|
* that <bits> is a power of two.
|
||
|
*/
|
||
|
#define VGIC_ADDR_IRQ_MASK(bits) (((bits) * 1024 / 8) - 1)
|
||
|
|
||
|
/*
|
||
|
* (addr & mask) gives us the _byte_ offset for the INT ID.
|
||
|
* We multiply this by 8 the get the _bit_ offset, then divide this by
|
||
|
* the number of bits to learn the actual INT ID.
|
||
|
* But instead of a division (which requires a "long long div" implementation),
|
||
|
* we shift by the binary logarithm of <bits>.
|
||
|
* This assumes that <bits> is a power of two.
|
||
|
*/
|
||
|
#define VGIC_ADDR_TO_INTID(addr, bits) (((addr) & VGIC_ADDR_IRQ_MASK(bits)) * \
|
||
|
8 >> ilog2(bits))
|
||
|
|
||
|
/*
|
||
|
* Some VGIC registers store per-IRQ information, with a different number
|
||
|
* of bits per IRQ. For those registers this macro is used.
|
||
|
* The _WITH_LENGTH version instantiates registers with a fixed length
|
||
|
* and is mutually exclusive with the _PER_IRQ version.
|
||
|
*/
|
||
|
#define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc) \
|
||
|
{ \
|
||
|
.reg_offset = off, \
|
||
|
.bits_per_irq = bpi, \
|
||
|
.len = bpi * 1024 / 8, \
|
||
|
.access_flags = acc, \
|
||
|
.read = rd, \
|
||
|
.write = wr, \
|
||
|
.uaccess_read = ur, \
|
||
|
.uaccess_write = uw, \
|
||
|
}
|
||
|
|
||
|
#define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc) \
|
||
|
{ \
|
||
|
.reg_offset = off, \
|
||
|
.bits_per_irq = 0, \
|
||
|
.len = length, \
|
||
|
.access_flags = acc, \
|
||
|
.read = rd, \
|
||
|
.write = wr, \
|
||
|
}
|
||
|
|
||
|
#define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc) \
|
||
|
{ \
|
||
|
.reg_offset = off, \
|
||
|
.bits_per_irq = 0, \
|
||
|
.len = length, \
|
||
|
.access_flags = acc, \
|
||
|
.read = rd, \
|
||
|
.write = wr, \
|
||
|
.uaccess_read = urd, \
|
||
|
.uaccess_write = uwr, \
|
||
|
}
|
||
|
|
||
|
unsigned long vgic_data_mmio_bus_to_host(const void *val, unsigned int len);
|
||
|
|
||
|
void vgic_data_host_to_mmio_bus(void *buf, unsigned int len,
|
||
|
unsigned long data);
|
||
|
|
||
|
unsigned long extract_bytes(u64 data, unsigned int offset,
|
||
|
unsigned int num);
|
||
|
|
||
|
u64 update_64bit_reg(u64 reg, unsigned int offset, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len, unsigned long val);
|
||
|
|
||
|
int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len, unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr,
|
||
|
unsigned int len, unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_senable(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_uaccess_write_senable(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_uaccess_write_cenable(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
unsigned long vgic_uaccess_read_pending(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_spending(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_uaccess_write_spending(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_uaccess_write_cpending(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_priority(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len);
|
||
|
|
||
|
void vgic_mmio_write_config(struct kvm_vcpu *vcpu,
|
||
|
gpa_t addr, unsigned int len,
|
||
|
unsigned long val);
|
||
|
|
||
|
int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev,
|
||
|
bool is_write, int offset, u32 *val);
|
||
|
|
||
|
u64 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid);
|
||
|
|
||
|
void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid,
|
||
|
const u64 val);
|
||
|
|
||
|
unsigned int vgic_v2_init_dist_iodev(struct vgic_io_device *dev);
|
||
|
|
||
|
unsigned int vgic_v3_init_dist_iodev(struct vgic_io_device *dev);
|
||
|
|
||
|
u64 vgic_sanitise_outer_cacheability(u64 reg);
|
||
|
u64 vgic_sanitise_inner_cacheability(u64 reg);
|
||
|
u64 vgic_sanitise_shareability(u64 reg);
|
||
|
u64 vgic_sanitise_field(u64 reg, u64 field_mask, int field_shift,
|
||
|
u64 (*sanitise_fn)(u64));
|
||
|
|
||
|
/* Find the proper register handler entry given a certain address offset */
|
||
|
const struct vgic_register_region *
|
||
|
vgic_find_mmio_region(const struct vgic_register_region *regions,
|
||
|
int nr_regions, unsigned int offset);
|
||
|
|
||
|
#endif
|