On 6/21/19 10:38 AM, Marc Zyngier wrote:
> From: Christoffer Dall <[email protected]>
>
> Emulating EL2 also means emulating the EL2 timers. To do so, we expand
> our timer framework to deal with at most 4 timers. At any given time,
> two timers are using the HW timers, and the two others are purely
> emulated.
>
> The role of deciding which is which at any given time is left to a
> mapping function which is called every time we need to make such a
> decision.
>
> Signed-off-by: Christoffer Dall <[email protected]>
> Signed-off-by: Marc Zyngier <[email protected]>
> ---
> arch/arm/include/asm/kvm_emulate.h | 2 +
> include/kvm/arm_arch_timer.h | 5 ++
> include/kvm/arm_vgic.h | 1 +
> virt/kvm/arm/arch_timer.c | 122 ++++++++++++++++++++++++++++-
> virt/kvm/arm/trace.h | 6 +-
> virt/kvm/arm/vgic/vgic.c | 15 ++++
> 6 files changed, 147 insertions(+), 4 deletions(-)
>
> diff --git a/arch/arm/include/asm/kvm_emulate.h
> b/arch/arm/include/asm/kvm_emulate.h
> index 6b7644a383f6..865ce545b465 100644
> --- a/arch/arm/include/asm/kvm_emulate.h
> +++ b/arch/arm/include/asm/kvm_emulate.h
> @@ -333,4 +333,6 @@ static inline unsigned long
> vcpu_data_host_to_guest(struct kvm_vcpu *vcpu,
>
> static inline void vcpu_ptrauth_setup_lazy(struct kvm_vcpu *vcpu) {}
>
> +static inline bool is_hyp_ctxt(struct kvm_vcpu *vcpu) { return false; }
> +
> #endif /* __ARM_KVM_EMULATE_H__ */
> diff --git a/include/kvm/arm_arch_timer.h b/include/kvm/arm_arch_timer.h
> index d120e6c323e7..3a5d9255120e 100644
> --- a/include/kvm/arm_arch_timer.h
> +++ b/include/kvm/arm_arch_timer.h
> @@ -13,6 +13,8 @@
> enum kvm_arch_timers {
> TIMER_PTIMER,
> TIMER_VTIMER,
> + TIMER_HVTIMER,
> + TIMER_HPTIMER,
> NR_KVM_TIMERS
> };
>
> @@ -54,6 +56,7 @@ struct arch_timer_context {
> struct timer_map {
> struct arch_timer_context *direct_vtimer;
> struct arch_timer_context *direct_ptimer;
> + struct arch_timer_context *emul_vtimer;
> struct arch_timer_context *emul_ptimer;
> };
>
> @@ -98,6 +101,8 @@ bool kvm_arch_timer_get_input_level(int vintid);
> #define vcpu_get_timer(v,t) (&vcpu_timer(v)->timers[(t)])
> #define vcpu_vtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_VTIMER])
> #define vcpu_ptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_PTIMER])
> +#define vcpu_hvtimer(v) (&(v)->arch.timer_cpu.timers[TIMER_HVTIMER])
> +#define vcpu_hptimer(v) (&(v)->arch.timer_cpu.timers[TIMER_HPTIMER])
>
> #define arch_timer_ctx_index(ctx) ((ctx) -
> vcpu_timer((ctx)->vcpu)->timers)
>
> diff --git a/include/kvm/arm_vgic.h b/include/kvm/arm_vgic.h
> index c36c86f1ec9a..7fc3b413b3de 100644
> --- a/include/kvm/arm_vgic.h
> +++ b/include/kvm/arm_vgic.h
> @@ -355,6 +355,7 @@ int kvm_vgic_inject_irq(struct kvm *kvm, int cpuid,
> unsigned int intid,
> int kvm_vgic_map_phys_irq(struct kvm_vcpu *vcpu, unsigned int host_irq,
> u32 vintid, bool (*get_input_level)(int vindid));
> int kvm_vgic_unmap_phys_irq(struct kvm_vcpu *vcpu, unsigned int vintid);
> +int kvm_vgic_get_map(struct kvm_vcpu *vcpu, unsigned int vintid);
> bool kvm_vgic_map_is_active(struct kvm_vcpu *vcpu, unsigned int vintid);
>
> int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu);
> diff --git a/virt/kvm/arm/arch_timer.c b/virt/kvm/arm/arch_timer.c
> index 089441a07ed7..3d84c240071d 100644
> --- a/virt/kvm/arm/arch_timer.c
> +++ b/virt/kvm/arm/arch_timer.c
> @@ -15,6 +15,7 @@
> #include <asm/arch_timer.h>
> #include <asm/kvm_emulate.h>
> #include <asm/kvm_hyp.h>
> +#include <asm/kvm_nested.h>
>
> #include <kvm/arm_vgic.h>
> #include <kvm/arm_arch_timer.h>
> @@ -39,6 +40,16 @@ static const struct kvm_irq_level default_vtimer_irq = {
> .level = 1,
> };
>
> +static const struct kvm_irq_level default_hptimer_irq = {
> + .irq = 26,
> + .level = 1,
> +};
> +
> +static const struct kvm_irq_level default_hvtimer_irq = {
> + .irq = 28,
> + .level = 1,
> +};
> +
> static bool kvm_timer_irq_can_fire(struct arch_timer_context *timer_ctx);
> static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level,
> struct arch_timer_context *timer_ctx);
> @@ -58,13 +69,27 @@ u64 kvm_phys_timer_read(void)
>
> static void get_timer_map(struct kvm_vcpu *vcpu, struct timer_map *map)
> {
> - if (has_vhe()) {
> + if (nested_virt_in_use(vcpu)) {
> + if (is_hyp_ctxt(vcpu)) {
> + map->direct_vtimer = vcpu_hvtimer(vcpu);
> + map->direct_ptimer = vcpu_hptimer(vcpu);
> + map->emul_vtimer = vcpu_vtimer(vcpu);
> + map->emul_ptimer = vcpu_ptimer(vcpu);
> + } else {
> + map->direct_vtimer = vcpu_vtimer(vcpu);
> + map->direct_ptimer = vcpu_ptimer(vcpu);
> + map->emul_vtimer = vcpu_hvtimer(vcpu);
> + map->emul_ptimer = vcpu_hptimer(vcpu);
> + }
> + } else if (has_vhe()) {
> map->direct_vtimer = vcpu_vtimer(vcpu);
> map->direct_ptimer = vcpu_ptimer(vcpu);
> + map->emul_vtimer = NULL;
> map->emul_ptimer = NULL;
> } else {
> map->direct_vtimer = vcpu_vtimer(vcpu);
> map->direct_ptimer = NULL;
> + map->emul_vtimer = NULL;
> map->emul_ptimer = vcpu_ptimer(vcpu);
> }
>
> @@ -237,9 +262,11 @@ static bool kvm_timer_should_fire(struct
> arch_timer_context *timer_ctx)
>
> switch (index) {
> case TIMER_VTIMER:
> + case TIMER_HVTIMER:
> cnt_ctl = read_sysreg_el0(SYS_CNTV_CTL);
> break;
> case TIMER_PTIMER:
> + case TIMER_HPTIMER:
> cnt_ctl = read_sysreg_el0(SYS_CNTP_CTL);
> break;
> case NR_KVM_TIMERS:
> @@ -270,6 +297,7 @@ bool kvm_timer_is_pending(struct kvm_vcpu *vcpu)
>
> return kvm_timer_should_fire(map.direct_vtimer) ||
> kvm_timer_should_fire(map.direct_ptimer) ||
> + kvm_timer_should_fire(map.emul_vtimer) ||
> kvm_timer_should_fire(map.emul_ptimer);
> }
>
> @@ -349,6 +377,7 @@ static void timer_save_state(struct arch_timer_context
> *ctx)
>
> switch (index) {
> case TIMER_VTIMER:
> + case TIMER_HVTIMER:
> ctx->cnt_ctl = read_sysreg_el0(SYS_CNTV_CTL);
> ctx->cnt_cval = read_sysreg_el0(SYS_CNTV_CVAL);
>
> @@ -358,6 +387,7 @@ static void timer_save_state(struct arch_timer_context
> *ctx)
>
> break;
> case TIMER_PTIMER:
> + case TIMER_HPTIMER:
> ctx->cnt_ctl = read_sysreg_el0(SYS_CNTP_CTL);
> ctx->cnt_cval = read_sysreg_el0(SYS_CNTP_CVAL);
>
> @@ -395,6 +425,7 @@ static void kvm_timer_blocking(struct kvm_vcpu *vcpu)
> */
> if (!kvm_timer_irq_can_fire(map.direct_vtimer) &&
> !kvm_timer_irq_can_fire(map.direct_ptimer) &&
> + !kvm_timer_irq_can_fire(map.emul_vtimer) &&
> !kvm_timer_irq_can_fire(map.emul_ptimer))
> return;
>
> @@ -428,11 +459,13 @@ static void timer_restore_state(struct
> arch_timer_context *ctx)
>
> switch (index) {
> case TIMER_VTIMER:
> + case TIMER_HVTIMER:
> write_sysreg_el0(ctx->cnt_cval, SYS_CNTV_CVAL);
> isb();
> write_sysreg_el0(ctx->cnt_ctl, SYS_CNTV_CTL);
> break;
> case TIMER_PTIMER:
> + case TIMER_HPTIMER:
> write_sysreg_el0(ctx->cnt_cval, SYS_CNTP_CVAL);
> isb();
> write_sysreg_el0(ctx->cnt_ctl, SYS_CNTP_CTL);
> @@ -519,6 +552,40 @@ static void kvm_timer_vcpu_load_nogic(struct kvm_vcpu
> *vcpu)
> enable_percpu_irq(host_vtimer_irq, host_vtimer_irq_flags);
> }
>
> +static void kvm_timer_vcpu_load_nested_switch(struct kvm_vcpu *vcpu,
> + struct timer_map *map)
> +{
> + int hw, ret;
> +
> + if (!irqchip_in_kernel(vcpu->kvm))
> + return;
> +
> + /*
> + * We only ever unmap the vtimer irq on a VHE system that runs nested
> + * virtualization, in which case we have both a valid emul_vtimer,
> + * emul_ptimer, direct_vtimer, and direct_ptimer.
> + *
> + * Since this is called from kvm_timer_vcpu_load(), a change between
> + * vEL2 and vEL1/0 will have just happened, and the timer_map will
> + * represent this, and therefore we switch the emul/direct mappings
> + * below.
> + */
> + hw = kvm_vgic_get_map(vcpu, map->direct_vtimer->irq.irq);
> + if (hw < 0) {
> + kvm_vgic_unmap_phys_irq(vcpu, map->emul_vtimer->irq.irq);
> + kvm_vgic_unmap_phys_irq(vcpu, map->emul_ptimer->irq.irq);
> +
> + ret = kvm_vgic_map_phys_irq(vcpu,
> + map->direct_vtimer->host_timer_irq,
> + map->direct_vtimer->irq.irq,
> + kvm_arch_timer_get_input_level);
> + ret = kvm_vgic_map_phys_irq(vcpu,
> + map->direct_ptimer->host_timer_irq,
> + map->direct_ptimer->irq.irq,
> + kvm_arch_timer_get_input_level);
> + }
> +}
> +
> void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu)
> {
> struct arch_timer_cpu *timer = vcpu_timer(vcpu);
> @@ -530,6 +597,8 @@ void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu)
> get_timer_map(vcpu, &map);
>
> if (static_branch_likely(&has_gic_active_state)) {
> + kvm_timer_vcpu_load_nested_switch(vcpu, &map);
Would it be possible for this be a conditional call that depends on
nested_virt_in_use(vcpu), since that's the only situation where we change the
direct_vtimer when we switch from vEL1 to vEL2 or viceversa?
Thanks,
Alex
> +
> kvm_timer_vcpu_load_gic(map.direct_vtimer);
> if (map.direct_ptimer)
> kvm_timer_vcpu_load_gic(map.direct_ptimer);
> @@ -545,6 +614,8 @@ void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu)
> if (map.direct_ptimer)
> timer_restore_state(map.direct_ptimer);
>
> + if (map.emul_vtimer)
> + timer_emulate(map.emul_vtimer);
> if (map.emul_ptimer)
> timer_emulate(map.emul_ptimer);
> }
> @@ -589,6 +660,8 @@ void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu)
> * In any case, we re-schedule the hrtimer for the physical timer when
> * coming back to the VCPU thread in kvm_timer_vcpu_load().
> */
> + if (map.emul_vtimer)
> + soft_timer_cancel(&map.emul_vtimer->hrtimer);
> if (map.emul_ptimer)
> soft_timer_cancel(&map.emul_ptimer->hrtimer);
>
> @@ -649,10 +722,14 @@ int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu)
> */
> vcpu_vtimer(vcpu)->cnt_ctl = 0;
> vcpu_ptimer(vcpu)->cnt_ctl = 0;
> + vcpu_hvtimer(vcpu)->cnt_ctl = 0;
> + vcpu_hptimer(vcpu)->cnt_ctl = 0;
>
> if (timer->enabled) {
> kvm_timer_update_irq(vcpu, false, vcpu_vtimer(vcpu));
> kvm_timer_update_irq(vcpu, false, vcpu_ptimer(vcpu));
> + kvm_timer_update_irq(vcpu, false, vcpu_hvtimer(vcpu));
> + kvm_timer_update_irq(vcpu, false, vcpu_hptimer(vcpu));
>
> if (irqchip_in_kernel(vcpu->kvm)) {
> kvm_vgic_reset_mapped_irq(vcpu,
> map.direct_vtimer->irq.irq);
> @@ -661,6 +738,8 @@ int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu)
> }
> }
>
> + if (map.emul_vtimer)
> + soft_timer_cancel(&map.emul_vtimer->hrtimer);
> if (map.emul_ptimer)
> soft_timer_cancel(&map.emul_ptimer->hrtimer);
>
> @@ -691,30 +770,46 @@ void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu)
> struct arch_timer_cpu *timer = vcpu_timer(vcpu);
> struct arch_timer_context *vtimer = vcpu_vtimer(vcpu);
> struct arch_timer_context *ptimer = vcpu_ptimer(vcpu);
> + struct arch_timer_context *hvtimer = vcpu_hvtimer(vcpu);
> + struct arch_timer_context *hptimer = vcpu_hptimer(vcpu);
>
> /* Synchronize cntvoff across all vtimers of a VM. */
> update_vtimer_cntvoff(vcpu, kvm_phys_timer_read());
> ptimer->cntvoff = 0;
> + hvtimer->cntvoff = 0;
> + hptimer->cntvoff = 0;
>
> hrtimer_init(&timer->bg_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
> timer->bg_timer.function = kvm_bg_timer_expire;
>
> hrtimer_init(&vtimer->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
> hrtimer_init(&ptimer->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
> + hrtimer_init(&hvtimer->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
> + hrtimer_init(&hptimer->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
> vtimer->hrtimer.function = kvm_hrtimer_expire;
> ptimer->hrtimer.function = kvm_hrtimer_expire;
> + hvtimer->hrtimer.function = kvm_hrtimer_expire;
> + hptimer->hrtimer.function = kvm_hrtimer_expire;
>
> vtimer->irq.irq = default_vtimer_irq.irq;
> ptimer->irq.irq = default_ptimer_irq.irq;
> + hvtimer->irq.irq = default_hvtimer_irq.irq;
> + hptimer->irq.irq = default_hptimer_irq.irq;
>
> vtimer->host_timer_irq = host_vtimer_irq;
> ptimer->host_timer_irq = host_ptimer_irq;
> + hvtimer->host_timer_irq = host_vtimer_irq;
> + hptimer->host_timer_irq = host_ptimer_irq;
>
> vtimer->host_timer_irq_flags = host_vtimer_irq_flags;
> ptimer->host_timer_irq_flags = host_ptimer_irq_flags;
> + hvtimer->host_timer_irq_flags = host_vtimer_irq_flags;
> + hptimer->host_timer_irq_flags = host_ptimer_irq_flags;
>
> vtimer->vcpu = vcpu;
> ptimer->vcpu = vcpu;
> + hvtimer->vcpu = vcpu;
> + hptimer->vcpu = vcpu;
> }
>
> static void kvm_timer_init_interrupt(void *info)
> @@ -997,7 +1092,7 @@ void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu)
>
> static bool timer_irqs_are_valid(struct kvm_vcpu *vcpu)
> {
> - int vtimer_irq, ptimer_irq;
> + int vtimer_irq, ptimer_irq, hvtimer_irq, hptimer_irq;
> int i, ret;
>
> vtimer_irq = vcpu_vtimer(vcpu)->irq.irq;
> @@ -1010,9 +1105,21 @@ static bool timer_irqs_are_valid(struct kvm_vcpu *vcpu)
> if (ret)
> return false;
>
> + hvtimer_irq = vcpu_hvtimer(vcpu)->irq.irq;
> + ret = kvm_vgic_set_owner(vcpu, hvtimer_irq, vcpu_hvtimer(vcpu));
> + if (ret)
> + return false;
> +
> + hptimer_irq = vcpu_hptimer(vcpu)->irq.irq;
> + ret = kvm_vgic_set_owner(vcpu, hptimer_irq, vcpu_hptimer(vcpu));
> + if (ret)
> + return false;
> +
> kvm_for_each_vcpu(i, vcpu, vcpu->kvm) {
> if (vcpu_vtimer(vcpu)->irq.irq != vtimer_irq ||
> - vcpu_ptimer(vcpu)->irq.irq != ptimer_irq)
> + vcpu_ptimer(vcpu)->irq.irq != ptimer_irq ||
> + vcpu_hvtimer(vcpu)->irq.irq != hvtimer_irq ||
> + vcpu_hptimer(vcpu)->irq.irq != hptimer_irq)
> return false;
> }
>
> @@ -1028,6 +1135,10 @@ bool kvm_arch_timer_get_input_level(int vintid)
> timer = vcpu_vtimer(vcpu);
> else if (vintid == vcpu_ptimer(vcpu)->irq.irq)
> timer = vcpu_ptimer(vcpu);
> + else if (vintid == vcpu_hvtimer(vcpu)->irq.irq)
> + timer = vcpu_hvtimer(vcpu);
> + else if (vintid == vcpu_hptimer(vcpu)->irq.irq)
> + timer = vcpu_hptimer(vcpu);
> else
> BUG();
>
> @@ -1109,6 +1220,7 @@ static void set_timer_irqs(struct kvm *kvm, int
> vtimer_irq, int ptimer_irq)
> kvm_for_each_vcpu(i, vcpu, kvm) {
> vcpu_vtimer(vcpu)->irq.irq = vtimer_irq;
> vcpu_ptimer(vcpu)->irq.irq = ptimer_irq;
> + /* TODO: Add support for hv/hp timers */
> }
> }
>
> @@ -1119,6 +1231,8 @@ int kvm_arm_timer_set_attr(struct kvm_vcpu *vcpu,
> struct kvm_device_attr *attr)
> struct arch_timer_context *ptimer = vcpu_ptimer(vcpu);
> int irq;
>
> + /* TODO: Add support for hv/hp timers */
> +
> if (!irqchip_in_kernel(vcpu->kvm))
> return -EINVAL;
>
> @@ -1151,6 +1265,8 @@ int kvm_arm_timer_get_attr(struct kvm_vcpu *vcpu,
> struct kvm_device_attr *attr)
> struct arch_timer_context *timer;
> int irq;
>
> + /* TODO: Add support for hv/hp timers */
> +
> switch (attr->attr) {
> case KVM_ARM_VCPU_TIMER_IRQ_VTIMER:
> timer = vcpu_vtimer(vcpu);
> diff --git a/virt/kvm/arm/trace.h b/virt/kvm/arm/trace.h
> index 204d210d01c2..3b08cc0376f4 100644
> --- a/virt/kvm/arm/trace.h
> +++ b/virt/kvm/arm/trace.h
> @@ -271,6 +271,7 @@ TRACE_EVENT(kvm_get_timer_map,
> __field( unsigned long, vcpu_id )
> __field( int, direct_vtimer )
> __field( int, direct_ptimer )
> + __field( int, emul_vtimer )
> __field( int, emul_ptimer )
> ),
>
> @@ -279,14 +280,17 @@ TRACE_EVENT(kvm_get_timer_map,
> __entry->direct_vtimer =
> arch_timer_ctx_index(map->direct_vtimer);
> __entry->direct_ptimer =
> (map->direct_ptimer) ?
> arch_timer_ctx_index(map->direct_ptimer) : -1;
> + __entry->emul_vtimer =
> + (map->emul_vtimer) ?
> arch_timer_ctx_index(map->emul_vtimer) : -1;
> __entry->emul_ptimer =
> (map->emul_ptimer) ?
> arch_timer_ctx_index(map->emul_ptimer) : -1;
> ),
>
> - TP_printk("VCPU: %ld, dv: %d, dp: %d, ep: %d",
> + TP_printk("VCPU: %ld, dv: %d, dp: %d, ev: %d, ep: %d",
> __entry->vcpu_id,
> __entry->direct_vtimer,
> __entry->direct_ptimer,
> + __entry->emul_vtimer,
> __entry->emul_ptimer)
> );
>
> diff --git a/virt/kvm/arm/vgic/vgic.c b/virt/kvm/arm/vgic/vgic.c
> index 191deccf60bf..1c5b4dbd33e4 100644
> --- a/virt/kvm/arm/vgic/vgic.c
> +++ b/virt/kvm/arm/vgic/vgic.c
> @@ -569,6 +569,21 @@ int kvm_vgic_unmap_phys_irq(struct kvm_vcpu *vcpu,
> unsigned int vintid)
> return 0;
> }
>
> +int kvm_vgic_get_map(struct kvm_vcpu *vcpu, unsigned int vintid)
> +{
> + struct vgic_irq *irq = vgic_get_irq(vcpu->kvm, vcpu, vintid);
> + unsigned long flags;
> + int ret = -1;
> +
> + raw_spin_lock_irqsave(&irq->irq_lock, flags);
> + if (irq->hw)
> + ret = irq->hwintid;
> + raw_spin_unlock_irqrestore(&irq->irq_lock, flags);
> +
> + vgic_put_irq(vcpu->kvm, irq);
> + return ret;
> +}
> +
> /**
> * kvm_vgic_set_owner - Set the owner of an interrupt for a VM
> *
_______________________________________________
kvmarm mailing list
[email protected]
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm