aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorAnup Patel <anup.patel@wdc.com>2021-11-19 18:15:12 +0530
committerWill Deacon <will@kernel.org>2021-12-14 15:05:30 +0000
commit762224e47cc2f8f5192d014caee3315ad4129fc9 (patch)
tree97ff385814a938919f03f46ec736af802a15b2b4
parent42bfe448c1c38c877cc81ebd403d5fb85bb927a7 (diff)
downloadkvmtool-762224e47cc2f8f5192d014caee3315ad4129fc9.tar.gz
riscv: Add PLIC device emulation
The PLIC (platform level interrupt controller) manages peripheral interrupts in RISC-V world. The per-CPU interrupts are managed using CPU CSRs hence virtualized in-kernel by KVM RISC-V. This patch adds PLIC device emulation for KVMTOOL RISC-V. Signed-off-by: Vincent Chen <vincent.chen@sifive.com> [For PLIC context CLAIM register emulation] Signed-off-by: Anup Patel <anup.patel@wdc.com> Link: https://lore.kernel.org/r/20211119124515.89439-6-anup.patel@wdc.com Signed-off-by: Will Deacon <will@kernel.org>
-rw-r--r--Makefile1
-rw-r--r--riscv/include/kvm/kvm-arch.h2
-rw-r--r--riscv/irq.c4
-rw-r--r--riscv/plic.c521
4 files changed, 526 insertions, 2 deletions
diff --git a/Makefile b/Makefile
index 5caf9b39..0d7c7ec8 100644
--- a/Makefile
+++ b/Makefile
@@ -204,6 +204,7 @@ ifeq ($(ARCH),riscv)
OBJS += riscv/irq.o
OBJS += riscv/kvm.o
OBJS += riscv/kvm-cpu.o
+ OBJS += riscv/plic.o
ifeq ($(RISCV_XLEN),32)
CFLAGS += -mabi=ilp32d -march=rv32gc
endif
diff --git a/riscv/include/kvm/kvm-arch.h b/riscv/include/kvm/kvm-arch.h
index 300538d8..25b80b3f 100644
--- a/riscv/include/kvm/kvm-arch.h
+++ b/riscv/include/kvm/kvm-arch.h
@@ -80,4 +80,6 @@ static inline bool riscv_addr_in_ioport_region(u64 phys_addr)
enum irq_type;
+void plic__irq_trig(struct kvm *kvm, int irq, int level, bool edge);
+
#endif /* KVM__KVM_ARCH_H */
diff --git a/riscv/irq.c b/riscv/irq.c
index 8e605efa..78a582d6 100644
--- a/riscv/irq.c
+++ b/riscv/irq.c
@@ -4,10 +4,10 @@
void kvm__irq_line(struct kvm *kvm, int irq, int level)
{
- /* TODO: */
+ plic__irq_trig(kvm, irq, level, false);
}
void kvm__irq_trigger(struct kvm *kvm, int irq)
{
- /* TODO: */
+ plic__irq_trig(kvm, irq, 1, true);
}
diff --git a/riscv/plic.c b/riscv/plic.c
new file mode 100644
index 00000000..6d64e8b5
--- /dev/null
+++ b/riscv/plic.c
@@ -0,0 +1,521 @@
+
+#include "kvm/devices.h"
+#include "kvm/ioeventfd.h"
+#include "kvm/ioport.h"
+#include "kvm/kvm.h"
+#include "kvm/kvm-cpu.h"
+#include "kvm/irq.h"
+#include "kvm/mutex.h"
+
+#include <linux/byteorder.h>
+#include <linux/kernel.h>
+#include <linux/kvm.h>
+#include <linux/sizes.h>
+
+/*
+ * From the RISC-V Privlidged Spec v1.10:
+ *
+ * Global interrupt sources are assigned small unsigned integer identifiers,
+ * beginning at the value 1. An interrupt ID of 0 is reserved to mean no
+ * interrupt. Interrupt identifiers are also used to break ties when two or
+ * more interrupt sources have the same assigned priority. Smaller values of
+ * interrupt ID take precedence over larger values of interrupt ID.
+ *
+ * While the RISC-V supervisor spec doesn't define the maximum number of
+ * devices supported by the PLIC, the largest number supported by devices
+ * marked as 'riscv,plic0' (which is the only device type this driver supports,
+ * and is the only extant PLIC as of now) is 1024. As mentioned above, device
+ * 0 is defined to be non-existant so this device really only supports 1023
+ * devices.
+ */
+
+#define MAX_DEVICES 1024
+#define MAX_CONTEXTS 15872
+
+/*
+ * The PLIC consists of memory-mapped control registers, with a memory map as
+ * follows:
+ *
+ * base + 0x000000: Reserved (interrupt source 0 does not exist)
+ * base + 0x000004: Interrupt source 1 priority
+ * base + 0x000008: Interrupt source 2 priority
+ * ...
+ * base + 0x000FFC: Interrupt source 1023 priority
+ * base + 0x001000: Pending 0
+ * base + 0x001FFF: Pending
+ * base + 0x002000: Enable bits for sources 0-31 on context 0
+ * base + 0x002004: Enable bits for sources 32-63 on context 0
+ * ...
+ * base + 0x0020FC: Enable bits for sources 992-1023 on context 0
+ * base + 0x002080: Enable bits for sources 0-31 on context 1
+ * ...
+ * base + 0x002100: Enable bits for sources 0-31 on context 2
+ * ...
+ * base + 0x1F1F80: Enable bits for sources 992-1023 on context 15871
+ * base + 0x1F1F84: Reserved
+ * ... (higher context IDs would fit here, but wouldn't fit
+ * inside the per-context priority vector)
+ * base + 0x1FFFFC: Reserved
+ * base + 0x200000: Priority threshold for context 0
+ * base + 0x200004: Claim/complete for context 0
+ * base + 0x200008: Reserved
+ * ...
+ * base + 0x200FFC: Reserved
+ * base + 0x201000: Priority threshold for context 1
+ * base + 0x201004: Claim/complete for context 1
+ * ...
+ * base + 0xFFE000: Priority threshold for context 15871
+ * base + 0xFFE004: Claim/complete for context 15871
+ * base + 0xFFE008: Reserved
+ * ...
+ * base + 0xFFFFFC: Reserved
+ */
+
+/* Each interrupt source has a priority register associated with it. */
+#define PRIORITY_BASE 0
+#define PRIORITY_PER_ID 4
+
+/*
+ * Each hart context has a vector of interupt enable bits associated with it.
+ * There's one bit for each interrupt source.
+ */
+#define ENABLE_BASE 0x2000
+#define ENABLE_PER_HART 0x80
+
+/*
+ * Each hart context has a set of control registers associated with it. Right
+ * now there's only two: a source priority threshold over which the hart will
+ * take an interrupt, and a register to claim interrupts.
+ */
+#define CONTEXT_BASE 0x200000
+#define CONTEXT_PER_HART 0x1000
+#define CONTEXT_THRESHOLD 0
+#define CONTEXT_CLAIM 4
+
+#define REG_SIZE 0x1000000
+
+struct plic_state;
+
+struct plic_context {
+ /* State to which this belongs */
+ struct plic_state *s;
+
+ /* Static Configuration */
+ u32 num;
+ struct kvm_cpu *vcpu;
+
+ /* Local IRQ state */
+ struct mutex irq_lock;
+ u8 irq_priority_threshold;
+ u32 irq_enable[MAX_DEVICES/32];
+ u32 irq_pending[MAX_DEVICES/32];
+ u8 irq_pending_priority[MAX_DEVICES];
+ u32 irq_claimed[MAX_DEVICES/32];
+ u32 irq_autoclear[MAX_DEVICES/32];
+};
+
+struct plic_state {
+ bool ready;
+ struct kvm *kvm;
+ struct device_header dev_hdr;
+
+ /* Static Configuration */
+ u32 num_irq;
+ u32 num_irq_word;
+ u32 max_prio;
+
+ /* Context Array */
+ u32 num_context;
+ struct plic_context *contexts;
+
+ /* Global IRQ state */
+ struct mutex irq_lock;
+ u8 irq_priority[MAX_DEVICES];
+ u32 irq_level[MAX_DEVICES/32];
+};
+
+static struct plic_state plic;
+
+/* Note: Must be called with c->irq_lock held */
+static u32 __plic_context_best_pending_irq(struct plic_state *s,
+ struct plic_context *c)
+{
+ u8 best_irq_prio = 0;
+ u32 i, j, irq, best_irq = 0;
+
+ for (i = 0; i < s->num_irq_word; i++) {
+ if (!c->irq_pending[i])
+ continue;
+
+ for (j = 0; j < 32; j++) {
+ irq = i * 32 + j;
+ if ((s->num_irq <= irq) ||
+ !(c->irq_pending[i] & (1 << j)) ||
+ (c->irq_claimed[i] & (1 << j)))
+ continue;
+
+ if (!best_irq ||
+ (best_irq_prio < c->irq_pending_priority[irq])) {
+ best_irq = irq;
+ best_irq_prio = c->irq_pending_priority[irq];
+ }
+ }
+ }
+
+ return best_irq;
+}
+
+/* Note: Must be called with c->irq_lock held */
+static void __plic_context_irq_update(struct plic_state *s,
+ struct plic_context *c)
+{
+ u32 best_irq = __plic_context_best_pending_irq(s, c);
+ u32 virq = (best_irq) ? KVM_INTERRUPT_SET : KVM_INTERRUPT_UNSET;
+
+ if (ioctl(c->vcpu->vcpu_fd, KVM_INTERRUPT, &virq) < 0)
+ pr_warning("KVM_INTERRUPT failed");
+}
+
+/* Note: Must be called with c->irq_lock held */
+static u32 __plic_context_irq_claim(struct plic_state *s,
+ struct plic_context *c)
+{
+ u32 virq = KVM_INTERRUPT_UNSET;
+ u32 best_irq = __plic_context_best_pending_irq(s, c);
+ u32 best_irq_word = best_irq / 32;
+ u32 best_irq_mask = (1 << (best_irq % 32));
+
+ if (ioctl(c->vcpu->vcpu_fd, KVM_INTERRUPT, &virq) < 0)
+ pr_warning("KVM_INTERRUPT failed");
+
+ if (best_irq) {
+ if (c->irq_autoclear[best_irq_word] & best_irq_mask) {
+ c->irq_pending[best_irq_word] &= ~best_irq_mask;
+ c->irq_pending_priority[best_irq] = 0;
+ c->irq_claimed[best_irq_word] &= ~best_irq_mask;
+ c->irq_autoclear[best_irq_word] &= ~best_irq_mask;
+ } else
+ c->irq_claimed[best_irq_word] |= best_irq_mask;
+ }
+
+ __plic_context_irq_update(s, c);
+
+ return best_irq;
+}
+
+void plic__irq_trig(struct kvm *kvm, int irq, int level, bool edge)
+{
+ bool irq_marked = false;
+ u8 i, irq_prio, irq_word;
+ u32 irq_mask;
+ struct plic_context *c = NULL;
+ struct plic_state *s = &plic;
+
+ if (!s->ready)
+ return;
+
+ if (irq <= 0 || s->num_irq <= (u32)irq)
+ goto done;
+
+ mutex_lock(&s->irq_lock);
+
+ irq_prio = s->irq_priority[irq];
+ irq_word = irq / 32;
+ irq_mask = 1 << (irq % 32);
+
+ if (level)
+ s->irq_level[irq_word] |= irq_mask;
+ else
+ s->irq_level[irq_word] &= ~irq_mask;
+
+ /*
+ * Note: PLIC interrupts are level-triggered. As of now,
+ * there is no notion of edge-triggered interrupts. To
+ * handle this we auto-clear edge-triggered interrupts
+ * when PLIC context CLAIM register is read.
+ */
+ for (i = 0; i < s->num_context; i++) {
+ c = &s->contexts[i];
+
+ mutex_lock(&c->irq_lock);
+ if (c->irq_enable[irq_word] & irq_mask) {
+ if (level) {
+ c->irq_pending[irq_word] |= irq_mask;
+ c->irq_pending_priority[irq] = irq_prio;
+ if (edge)
+ c->irq_autoclear[irq_word] |= irq_mask;
+ } else {
+ c->irq_pending[irq_word] &= ~irq_mask;
+ c->irq_pending_priority[irq] = 0;
+ c->irq_claimed[irq_word] &= ~irq_mask;
+ c->irq_autoclear[irq_word] &= ~irq_mask;
+ }
+ __plic_context_irq_update(s, c);
+ irq_marked = true;
+ }
+ mutex_unlock(&c->irq_lock);
+
+ if (irq_marked)
+ break;
+ }
+
+done:
+ mutex_unlock(&s->irq_lock);
+}
+
+static void plic__priority_read(struct plic_state *s,
+ u64 offset, void *data)
+{
+ u32 irq = (offset >> 2);
+
+ if (irq == 0 || irq >= s->num_irq)
+ return;
+
+ mutex_lock(&s->irq_lock);
+ ioport__write32(data, s->irq_priority[irq]);
+ mutex_unlock(&s->irq_lock);
+}
+
+static void plic__priority_write(struct plic_state *s,
+ u64 offset, void *data)
+{
+ u32 val, irq = (offset >> 2);
+
+ if (irq == 0 || irq >= s->num_irq)
+ return;
+
+ mutex_lock(&s->irq_lock);
+ val = ioport__read32(data);
+ val &= ((1 << PRIORITY_PER_ID) - 1);
+ s->irq_priority[irq] = val;
+ mutex_unlock(&s->irq_lock);
+}
+
+static void plic__context_enable_read(struct plic_state *s,
+ struct plic_context *c,
+ u64 offset, void *data)
+{
+ u32 irq_word = offset >> 2;
+
+ if (s->num_irq_word < irq_word)
+ return;
+
+ mutex_lock(&c->irq_lock);
+ ioport__write32(data, c->irq_enable[irq_word]);
+ mutex_unlock(&c->irq_lock);
+}
+
+static void plic__context_enable_write(struct plic_state *s,
+ struct plic_context *c,
+ u64 offset, void *data)
+{
+ u8 irq_prio;
+ u32 i, irq, irq_mask;
+ u32 irq_word = offset >> 2;
+ u32 old_val, new_val, xor_val;
+
+ if (s->num_irq_word < irq_word)
+ return;
+
+ mutex_lock(&s->irq_lock);
+
+ mutex_lock(&c->irq_lock);
+
+ old_val = c->irq_enable[irq_word];
+ new_val = ioport__read32(data);
+
+ if (irq_word == 0)
+ new_val &= ~0x1;
+
+ c->irq_enable[irq_word] = new_val;
+
+ xor_val = old_val ^ new_val;
+ for (i = 0; i < 32; i++) {
+ irq = irq_word * 32 + i;
+ irq_mask = 1 << i;
+ irq_prio = s->irq_priority[irq];
+ if (!(xor_val & irq_mask))
+ continue;
+ if ((new_val & irq_mask) &&
+ (s->irq_level[irq_word] & irq_mask)) {
+ c->irq_pending[irq_word] |= irq_mask;
+ c->irq_pending_priority[irq] = irq_prio;
+ } else if (!(new_val & irq_mask)) {
+ c->irq_pending[irq_word] &= ~irq_mask;
+ c->irq_pending_priority[irq] = 0;
+ c->irq_claimed[irq_word] &= ~irq_mask;
+ }
+ }
+
+ __plic_context_irq_update(s, c);
+
+ mutex_unlock(&c->irq_lock);
+
+ mutex_unlock(&s->irq_lock);
+}
+
+static void plic__context_read(struct plic_state *s,
+ struct plic_context *c,
+ u64 offset, void *data)
+{
+ mutex_lock(&c->irq_lock);
+
+ switch (offset) {
+ case CONTEXT_THRESHOLD:
+ ioport__write32(data, c->irq_priority_threshold);
+ break;
+ case CONTEXT_CLAIM:
+ ioport__write32(data, __plic_context_irq_claim(s, c));
+ break;
+ default:
+ break;
+ };
+
+ mutex_unlock(&c->irq_lock);
+}
+
+static void plic__context_write(struct plic_state *s,
+ struct plic_context *c,
+ u64 offset, void *data)
+{
+ u32 val, irq_word, irq_mask;
+ bool irq_update = false;
+
+ mutex_lock(&c->irq_lock);
+
+ switch (offset) {
+ case CONTEXT_THRESHOLD:
+ val = ioport__read32(data);
+ val &= ((1 << PRIORITY_PER_ID) - 1);
+ if (val <= s->max_prio)
+ c->irq_priority_threshold = val;
+ else
+ irq_update = true;
+ break;
+ case CONTEXT_CLAIM:
+ val = ioport__read32(data);
+ irq_word = val / 32;
+ irq_mask = 1 << (val % 32);
+ if ((val < plic.num_irq) &&
+ (c->irq_enable[irq_word] & irq_mask)) {
+ c->irq_claimed[irq_word] &= ~irq_mask;
+ irq_update = true;
+ }
+ break;
+ default:
+ irq_update = true;
+ break;
+ };
+
+ if (irq_update)
+ __plic_context_irq_update(s, c);
+
+ mutex_unlock(&c->irq_lock);
+}
+
+static void plic__mmio_callback(struct kvm_cpu *vcpu,
+ u64 addr, u8 *data, u32 len,
+ u8 is_write, void *ptr)
+{
+ u32 cntx;
+ struct plic_state *s = ptr;
+
+ if (len != 4)
+ die("plic: invalid len=%d", len);
+
+ addr &= ~0x3;
+ addr -= RISCV_PLIC;
+
+ if (is_write) {
+ if (PRIORITY_BASE <= addr && addr < ENABLE_BASE) {
+ plic__priority_write(s, addr, data);
+ } else if (ENABLE_BASE <= addr && addr < CONTEXT_BASE) {
+ cntx = (addr - ENABLE_BASE) / ENABLE_PER_HART;
+ addr -= cntx * ENABLE_PER_HART + ENABLE_BASE;
+ if (cntx < s->num_context)
+ plic__context_enable_write(s,
+ &s->contexts[cntx],
+ addr, data);
+ } else if (CONTEXT_BASE <= addr && addr < REG_SIZE) {
+ cntx = (addr - CONTEXT_BASE) / CONTEXT_PER_HART;
+ addr -= cntx * CONTEXT_PER_HART + CONTEXT_BASE;
+ if (cntx < s->num_context)
+ plic__context_write(s, &s->contexts[cntx],
+ addr, data);
+ }
+ } else {
+ if (PRIORITY_BASE <= addr && addr < ENABLE_BASE) {
+ plic__priority_read(s, addr, data);
+ } else if (ENABLE_BASE <= addr && addr < CONTEXT_BASE) {
+ cntx = (addr - ENABLE_BASE) / ENABLE_PER_HART;
+ addr -= cntx * ENABLE_PER_HART + ENABLE_BASE;
+ if (cntx < s->num_context)
+ plic__context_enable_read(s,
+ &s->contexts[cntx],
+ addr, data);
+ } else if (CONTEXT_BASE <= addr && addr < REG_SIZE) {
+ cntx = (addr - CONTEXT_BASE) / CONTEXT_PER_HART;
+ addr -= cntx * CONTEXT_PER_HART + CONTEXT_BASE;
+ if (cntx < s->num_context)
+ plic__context_read(s, &s->contexts[cntx],
+ addr, data);
+ }
+ }
+}
+
+static int plic__init(struct kvm *kvm)
+{
+ u32 i;
+ int ret;
+ struct plic_context *c;
+
+ plic.kvm = kvm;
+ plic.dev_hdr = (struct device_header) {
+ .bus_type = DEVICE_BUS_MMIO,
+ };
+
+ plic.num_irq = MAX_DEVICES;
+ plic.num_irq_word = plic.num_irq / 32;
+ if ((plic.num_irq_word * 32) < plic.num_irq)
+ plic.num_irq_word++;
+ plic.max_prio = (1UL << PRIORITY_PER_ID) - 1;
+
+ plic.num_context = kvm->nrcpus * 2;
+ plic.contexts = calloc(plic.num_context, sizeof(struct plic_context));
+ if (!plic.contexts)
+ return -ENOMEM;
+ for (i = 0; i < plic.num_context; i++) {
+ c = &plic.contexts[i];
+ c->s = &plic;
+ c->num = i;
+ c->vcpu = kvm->cpus[i / 2];
+ mutex_init(&c->irq_lock);
+ }
+
+ mutex_init(&plic.irq_lock);
+
+ ret = kvm__register_mmio(kvm, RISCV_PLIC, RISCV_PLIC_SIZE,
+ false, plic__mmio_callback, &plic);
+ if (ret)
+ return ret;
+
+ ret = device__register(&plic.dev_hdr);
+ if (ret)
+ return ret;
+
+ plic.ready = true;
+
+ return 0;
+
+}
+dev_init(plic__init);
+
+static int plic__exit(struct kvm *kvm)
+{
+ plic.ready = false;
+ kvm__deregister_mmio(kvm, RISCV_PLIC);
+ free(plic.contexts);
+
+ return 0;
+}
+dev_exit(plic__exit);