Hi Ganapatrao, On 5/12/25 12:52 PM, Ganapatrao Kulkarni wrote: > Adding required changes to enable and test HVTIMER and HPTIMER > in vEL2. In default case, PTIMER and VTIMER are validated and with > NV enabled (with argument "-g 1"), HPTIMER and HVTIMER are validated > by injecting respective timer interrupts. > > Signed-off-by: Ganapatrao Kulkarni <gankulka...@os.amperecomputing.com> > --- > tools/testing/selftests/kvm/arch_timer.c | 8 +- > .../testing/selftests/kvm/arm64/arch_timer.c | 118 +++++++++++++++--- > .../selftests/kvm/include/arm64/arch_timer.h | 16 +++ > .../selftests/kvm/include/timer_test.h | 1 + > 4 files changed, 123 insertions(+), 20 deletions(-) > > diff --git a/tools/testing/selftests/kvm/arch_timer.c > b/tools/testing/selftests/kvm/arch_timer.c > index acb2cb596332..5c30bda0462b 100644 > --- a/tools/testing/selftests/kvm/arch_timer.c > +++ b/tools/testing/selftests/kvm/arch_timer.c > @@ -35,6 +35,7 @@ struct test_args test_args = { > .migration_freq_ms = TIMER_TEST_MIGRATION_FREQ_MS, > .timer_err_margin_us = TIMER_TEST_ERR_MARGIN_US, > .reserved = 1, > + .is_nested = false, > }; > > struct kvm_vcpu *vcpus[KVM_MAX_VCPUS]; > @@ -43,6 +44,7 @@ struct test_vcpu_shared_data > vcpu_shared_data[KVM_MAX_VCPUS]; > static pthread_t pt_vcpu_run[KVM_MAX_VCPUS]; > static unsigned long *vcpu_done_map; > static pthread_mutex_t vcpu_done_map_lock; > +bool is_nested; > > static void *test_vcpu_run(void *arg) > { > @@ -193,6 +195,7 @@ static void test_print_help(char *name) > pr_info("\t-o: Counter offset (in counter cycles, default: 0) > [aarch64-only]\n"); > pr_info("\t-e: Interrupt arrival error margin (in us) of the guest > timer (default: %u)\n", > TIMER_TEST_ERR_MARGIN_US); > + pr_info("\t-g: Enable Nested Virtualization, run guest code as guest > hypervisor (default: Disabled)\n"); I would recommend "default: 0" in all enhanced tests to clarify that 0/1 are the values to be used.
Thanks Eric > pr_info("\t-h: print this help screen\n"); > } > > @@ -200,7 +203,7 @@ static bool parse_args(int argc, char *argv[]) > { > int opt; > > - while ((opt = getopt(argc, argv, "hn:i:p:m:o:e:")) != -1) { > + while ((opt = getopt(argc, argv, "hn:i:p:m:o:e:g:")) != -1) { > switch (opt) { > case 'n': > test_args.nr_vcpus = atoi_positive("Number of vCPUs", > optarg); > @@ -226,6 +229,9 @@ static bool parse_args(int argc, char *argv[]) > test_args.counter_offset = strtol(optarg, NULL, 0); > test_args.reserved = 0; > break; > + case 'g': > + test_args.is_nested = atoi_non_negative("Is Nested", > optarg); > + break; > case 'h': > default: > goto err; > diff --git a/tools/testing/selftests/kvm/arm64/arch_timer.c > b/tools/testing/selftests/kvm/arm64/arch_timer.c > index eeba1cc87ff8..50bf05bb6f85 100644 > --- a/tools/testing/selftests/kvm/arm64/arch_timer.c > +++ b/tools/testing/selftests/kvm/arm64/arch_timer.c > @@ -12,16 +12,22 @@ > #include "timer_test.h" > #include "ucall_common.h" > #include "vgic.h" > +#include <nv_util.h> > > enum guest_stage { > GUEST_STAGE_VTIMER_CVAL = 1, > GUEST_STAGE_VTIMER_TVAL, > GUEST_STAGE_PTIMER_CVAL, > GUEST_STAGE_PTIMER_TVAL, > + GUEST_STAGE_HVTIMER_CVAL, > + GUEST_STAGE_HVTIMER_TVAL, > + GUEST_STAGE_HPTIMER_CVAL, > + GUEST_STAGE_HPTIMER_TVAL, > GUEST_STAGE_MAX, > }; > > static int vtimer_irq, ptimer_irq; > +static int hvtimer_irq, hptimer_irq; > > static void > guest_configure_timer_action(struct test_vcpu_shared_data *shared_data) > @@ -47,6 +53,26 @@ guest_configure_timer_action(struct test_vcpu_shared_data > *shared_data) > shared_data->xcnt = timer_get_cntct(PHYSICAL); > timer_set_ctl(PHYSICAL, CTL_ENABLE); > break; > + case GUEST_STAGE_HVTIMER_CVAL: > + timer_set_next_cval_ms(HVIRTUAL, test_args.timer_period_ms); > + shared_data->xcnt = timer_get_cntct(HVIRTUAL); > + timer_set_ctl(HVIRTUAL, CTL_ENABLE); > + break; > + case GUEST_STAGE_HVTIMER_TVAL: > + timer_set_next_tval_ms(HVIRTUAL, test_args.timer_period_ms); > + shared_data->xcnt = timer_get_cntct(HVIRTUAL); > + timer_set_ctl(HVIRTUAL, CTL_ENABLE); > + break; > + case GUEST_STAGE_HPTIMER_CVAL: > + timer_set_next_cval_ms(HPHYSICAL, test_args.timer_period_ms); > + shared_data->xcnt = timer_get_cntct(HPHYSICAL); > + timer_set_ctl(HPHYSICAL, CTL_ENABLE); > + break; > + case GUEST_STAGE_HPTIMER_TVAL: > + timer_set_next_tval_ms(HPHYSICAL, test_args.timer_period_ms); > + shared_data->xcnt = timer_get_cntct(HPHYSICAL); > + timer_set_ctl(HPHYSICAL, CTL_ENABLE); > + break; > default: > GUEST_ASSERT(0); > } > @@ -75,6 +101,16 @@ static void guest_validate_irq(unsigned int intid, > accessor = PHYSICAL; > timer_irq = ptimer_irq; > break; > + case GUEST_STAGE_HVTIMER_CVAL: > + case GUEST_STAGE_HVTIMER_TVAL: > + accessor = HVIRTUAL; > + timer_irq = hvtimer_irq; > + break; > + case GUEST_STAGE_HPTIMER_CVAL: > + case GUEST_STAGE_HPTIMER_TVAL: > + accessor = HPHYSICAL; > + timer_irq = hptimer_irq; > + break; > default: > GUEST_ASSERT(0); > return; > @@ -142,38 +178,79 @@ static void guest_code(void) > { > uint32_t cpu = guest_get_vcpuid(); > struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[cpu]; > + bool is_nested = false; > + enum arch_timer vtimer, ptimer; > + int vtmr_irq, ptmr_irq; > + enum guest_stage stage_vtimer_cval, stage_vtimer_tval; > + enum guest_stage stage_ptimer_cval, stage_ptimer_tval; > > - local_irq_disable(); > + if (read_sysreg(CurrentEL) == CurrentEL_EL2) > + is_nested = true; > > + local_irq_disable(); > gic_init(GIC_V3, test_args.nr_vcpus); > > - timer_set_ctl(VIRTUAL, CTL_IMASK); > - timer_set_ctl(PHYSICAL, CTL_IMASK); > + if (is_nested) { > + > + vtimer = HVIRTUAL; > + ptimer = HPHYSICAL; > + vtmr_irq = hvtimer_irq; > + ptmr_irq = hptimer_irq; > + stage_vtimer_cval = GUEST_STAGE_HVTIMER_CVAL; > + stage_vtimer_tval = GUEST_STAGE_HVTIMER_TVAL; > + stage_ptimer_cval = GUEST_STAGE_HPTIMER_CVAL; > + stage_ptimer_tval = GUEST_STAGE_HPTIMER_TVAL; > + } else { > + vtimer = VIRTUAL; > + ptimer = PHYSICAL; > + vtmr_irq = vtimer_irq; > + ptmr_irq = ptimer_irq; > + stage_vtimer_cval = GUEST_STAGE_VTIMER_CVAL; > + stage_vtimer_tval = GUEST_STAGE_VTIMER_TVAL; > + stage_ptimer_cval = GUEST_STAGE_PTIMER_CVAL; > + stage_ptimer_tval = GUEST_STAGE_PTIMER_TVAL; > + } > + > + timer_set_ctl(vtimer, CTL_IMASK); > + timer_set_ctl(ptimer, CTL_IMASK); > + gic_irq_enable(vtmr_irq); > + gic_irq_enable(ptmr_irq); > > - gic_irq_enable(vtimer_irq); > - gic_irq_enable(ptimer_irq); > local_irq_enable(); > > - guest_run_stage(shared_data, GUEST_STAGE_VTIMER_CVAL); > - guest_run_stage(shared_data, GUEST_STAGE_VTIMER_TVAL); > - guest_run_stage(shared_data, GUEST_STAGE_PTIMER_CVAL); > - guest_run_stage(shared_data, GUEST_STAGE_PTIMER_TVAL); > + guest_run_stage(shared_data, stage_vtimer_cval); > + guest_run_stage(shared_data, stage_vtimer_tval); > + guest_run_stage(shared_data, stage_ptimer_cval); > + guest_run_stage(shared_data, stage_ptimer_tval); > > GUEST_DONE(); > } > > static void test_init_timer_irq(struct kvm_vm *vm) > { > - /* Timer initid should be same for all the vCPUs, so query only vCPU-0 > */ > - vcpu_device_attr_get(vcpus[0], KVM_ARM_VCPU_TIMER_CTRL, > - KVM_ARM_VCPU_TIMER_IRQ_PTIMER, &ptimer_irq); > - vcpu_device_attr_get(vcpus[0], KVM_ARM_VCPU_TIMER_CTRL, > - KVM_ARM_VCPU_TIMER_IRQ_VTIMER, &vtimer_irq); > > - sync_global_to_guest(vm, ptimer_irq); > - sync_global_to_guest(vm, vtimer_irq); > - > - pr_debug("ptimer_irq: %d; vtimer_irq: %d\n", ptimer_irq, vtimer_irq); > + /* Timer initid should be same for all the vCPUs, so query only vCPU-0 > */ > + if (is_vcpu_nested(vcpus[0])) { > + vcpu_device_attr_get(vcpus[0], KVM_ARM_VCPU_TIMER_CTRL, > + KVM_ARM_VCPU_TIMER_IRQ_HPTIMER, &hptimer_irq); > + vcpu_device_attr_get(vcpus[0], KVM_ARM_VCPU_TIMER_CTRL, > + KVM_ARM_VCPU_TIMER_IRQ_HVTIMER, &hvtimer_irq); > + > + sync_global_to_guest(vm, hptimer_irq); > + sync_global_to_guest(vm, hvtimer_irq); > + > + pr_debug("hptimer_irq: %d; hvtimer_irq: %d\n", hptimer_irq, > hvtimer_irq); > + } else { > + vcpu_device_attr_get(vcpus[0], KVM_ARM_VCPU_TIMER_CTRL, > + KVM_ARM_VCPU_TIMER_IRQ_PTIMER, &ptimer_irq); > + vcpu_device_attr_get(vcpus[0], KVM_ARM_VCPU_TIMER_CTRL, > + KVM_ARM_VCPU_TIMER_IRQ_VTIMER, &vtimer_irq); > + > + sync_global_to_guest(vm, ptimer_irq); > + sync_global_to_guest(vm, vtimer_irq); > + > + pr_debug("ptimer_irq: %d; vtimer_irq: %d\n", ptimer_irq, > vtimer_irq); > + } > } > > static int gic_fd; > @@ -184,7 +261,10 @@ struct kvm_vm *test_vm_create(void) > unsigned int i; > int nr_vcpus = test_args.nr_vcpus; > > - vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus); > + if (test_args.is_nested) > + vm = nv_vm_create_with_vcpus_gic(nr_vcpus, vcpus, NULL, > guest_code); > + else > + vm = vm_create_with_vcpus(nr_vcpus, guest_code, vcpus); > > vm_init_descriptor_tables(vm); > vm_install_exception_handler(vm, VECTOR_IRQ_CURRENT, guest_irq_handler); > diff --git a/tools/testing/selftests/kvm/include/arm64/arch_timer.h > b/tools/testing/selftests/kvm/include/arm64/arch_timer.h > index bf461de34785..82ba908fba8b 100644 > --- a/tools/testing/selftests/kvm/include/arm64/arch_timer.h > +++ b/tools/testing/selftests/kvm/include/arm64/arch_timer.h > @@ -11,6 +11,8 @@ > enum arch_timer { > VIRTUAL, > PHYSICAL, > + HVIRTUAL, > + HPHYSICAL, > }; > > #define CTL_ENABLE (1 << 0) > @@ -37,8 +39,10 @@ static inline uint64_t timer_get_cntct(enum arch_timer > timer) > > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > return read_sysreg(cntvct_el0); > case PHYSICAL: > + case HPHYSICAL: > return read_sysreg(cntpct_el0); > default: > GUEST_FAIL("Unexpected timer type = %u", timer); > @@ -52,9 +56,11 @@ static inline void timer_set_cval(enum arch_timer timer, > uint64_t cval) > { > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > write_sysreg(cval, cntv_cval_el0); > break; > case PHYSICAL: > + case HPHYSICAL: > write_sysreg(cval, cntp_cval_el0); > break; > default: > @@ -68,8 +74,10 @@ static inline uint64_t timer_get_cval(enum arch_timer > timer) > { > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > return read_sysreg(cntv_cval_el0); > case PHYSICAL: > + case HPHYSICAL: > return read_sysreg(cntp_cval_el0); > default: > GUEST_FAIL("Unexpected timer type = %u", timer); > @@ -83,9 +91,11 @@ static inline void timer_set_tval(enum arch_timer timer, > int32_t tval) > { > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > write_sysreg(tval, cntv_tval_el0); > break; > case PHYSICAL: > + case HPHYSICAL: > write_sysreg(tval, cntp_tval_el0); > break; > default: > @@ -100,8 +110,10 @@ static inline int32_t timer_get_tval(enum arch_timer > timer) > isb(); > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > return read_sysreg(cntv_tval_el0); > case PHYSICAL: > + case HPHYSICAL: > return read_sysreg(cntp_tval_el0); > default: > GUEST_FAIL("Could not get timer %d\n", timer); > @@ -115,9 +127,11 @@ static inline void timer_set_ctl(enum arch_timer timer, > uint32_t ctl) > { > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > write_sysreg(ctl, cntv_ctl_el0); > break; > case PHYSICAL: > + case HPHYSICAL: > write_sysreg(ctl, cntp_ctl_el0); > break; > default: > @@ -131,8 +145,10 @@ static inline uint32_t timer_get_ctl(enum arch_timer > timer) > { > switch (timer) { > case VIRTUAL: > + case HVIRTUAL: > return read_sysreg(cntv_ctl_el0); > case PHYSICAL: > + case HPHYSICAL: > return read_sysreg(cntp_ctl_el0); > default: > GUEST_FAIL("Unexpected timer type = %u", timer); > diff --git a/tools/testing/selftests/kvm/include/timer_test.h > b/tools/testing/selftests/kvm/include/timer_test.h > index 9b6edaafe6d4..95f61c4a8a80 100644 > --- a/tools/testing/selftests/kvm/include/timer_test.h > +++ b/tools/testing/selftests/kvm/include/timer_test.h > @@ -26,6 +26,7 @@ struct test_args { > /* Members of struct kvm_arm_counter_offset */ > uint64_t counter_offset; > uint64_t reserved; > + bool is_nested; > }; > > /* Shared variables between host and guest */