On Thu, Apr 04, 2019 at 05:07:09PM +0200, Andrew Jones wrote:
> On Fri, Mar 29, 2019 at 01:00:47PM +0000, Dave Martin wrote:
> > Some aspects of vcpu configuration may be too complex to be
> > completed inside KVM_ARM_VCPU_INIT. Thus, there may be a
> > requirement for userspace to do some additional configuration
> > before various other ioctls will work in a consistent way.
> >
> > In particular this will be the case for SVE, where userspace will
> > need to negotiate the set of vector lengths to be made available to
> > the guest before the vcpu becomes fully usable.
> >
> > In order to provide an explicit way for userspace to confirm that
> > it has finished setting up a particular vcpu feature, this patch
> > adds a new ioctl KVM_ARM_VCPU_FINALIZE.
> >
> > When userspace has opted into a feature that requires finalization,
> > typically by means of a feature flag passed to KVM_ARM_VCPU_INIT, a
> > matching call to KVM_ARM_VCPU_FINALIZE is now required before
> > KVM_RUN or KVM_GET_REG_LIST is allowed. Individual features may
> > impose additional restrictions where appropriate.
> >
> > No existing vcpu features are affected by this, so current
> > userspace implementations will continue to work exactly as before,
> > with no need to issue KVM_ARM_VCPU_FINALIZE.
> >
> > As implemented in this patch, KVM_ARM_VCPU_FINALIZE is currently a
> > placeholder: no finalizable features exist yet, so ioctl is not
> > required and will always yield EINVAL. Subsequent patches will add
> > the finalization logic to make use of this ioctl for SVE.
> >
> > No functional change for existing userspace.
> >
> > Signed-off-by: Dave Martin <[email protected]>
> > Reviewed-by: Julien Thierry <[email protected]>
> > Tested-by: zhang.lei <[email protected]>
> >
> > ---
> >
> > Changes since v5:
> >
> > * Commit message, including subject line, rewritten.
> >
> > This patch is a rework of "KVM: arm/arm64: Add hook to finalize the
> > vcpu configuration". The old subject line and commit message no
> > longer accurately described what the patch does. However, the code
> > is an evolution of the previous patch rather than a wholesale
> > rewrite.
> >
> > * Added an explicit KVM_ARM_VCPU_FINALIZE ioctl, rather than just
> > providing internal hooks in the kernel to finalize the vcpu
> > configuration implicitly. This allows userspace to confirm exactly
> > when it has finished configuring the vcpu and is ready to use it.
> >
> > This results in simpler (and hopefully more maintainable) ioctl
> > ordering rules.
> > ---
> > arch/arm/include/asm/kvm_host.h | 4 ++++
> > arch/arm64/include/asm/kvm_host.h | 4 ++++
> > include/uapi/linux/kvm.h | 3 +++
> > virt/kvm/arm/arm.c | 18 ++++++++++++++++++
> > 4 files changed, 29 insertions(+)
> >
> > diff --git a/arch/arm/include/asm/kvm_host.h
> > b/arch/arm/include/asm/kvm_host.h
> > index a49ee01..e80cfc1 100644
> > --- a/arch/arm/include/asm/kvm_host.h
> > +++ b/arch/arm/include/asm/kvm_host.h
> > @@ -19,6 +19,7 @@
> > #ifndef __ARM_KVM_HOST_H__
> > #define __ARM_KVM_HOST_H__
> >
> > +#include <linux/errno.h>
> > #include <linux/types.h>
> > #include <linux/kvm_types.h>
> > #include <asm/cputype.h>
> > @@ -411,4 +412,7 @@ static inline int kvm_arm_setup_stage2(struct kvm *kvm,
> > unsigned long type)
> > return 0;
> > }
> >
> > +#define kvm_arm_vcpu_finalize(vcpu, what) (-EINVAL)
> > +#define kvm_arm_vcpu_is_finalized(vcpu) true
> > +
>
> We usually use inline functions for the stubs.
I guess we could.
The vcpu_has_sve() circular include problem applies here too if we put
the actual function bodies here, which is why I ended up with this. Now
that the bodies (for arm64) are out of line, it actually doesn't matter.
> > #endif /* __ARM_KVM_HOST_H__ */
> > diff --git a/arch/arm64/include/asm/kvm_host.h
> > b/arch/arm64/include/asm/kvm_host.h
> > index 3e89509..98658f7 100644
> > --- a/arch/arm64/include/asm/kvm_host.h
> > +++ b/arch/arm64/include/asm/kvm_host.h
> > @@ -23,6 +23,7 @@
> > #define __ARM64_KVM_HOST_H__
> >
> > #include <linux/bitmap.h>
> > +#include <linux/errno.h>
> > #include <linux/types.h>
> > #include <linux/jump_label.h>
> > #include <linux/kvm_types.h>
> > @@ -625,4 +626,7 @@ void kvm_arch_free_vm(struct kvm *kvm);
> >
> > int kvm_arm_setup_stage2(struct kvm *kvm, unsigned long type);
> >
> > +#define kvm_arm_vcpu_finalize(vcpu, what) (-EINVAL)
> > +#define kvm_arm_vcpu_is_finalized(vcpu) true
>
> Same as above.
Ditto
> > +
> > #endif /* __ARM64_KVM_HOST_H__ */
> > diff --git a/include/uapi/linux/kvm.h b/include/uapi/linux/kvm.h
> > index dc77a5a..c3b8e7a 100644
> > --- a/include/uapi/linux/kvm.h
> > +++ b/include/uapi/linux/kvm.h
> > @@ -1441,6 +1441,9 @@ struct kvm_enc_region {
> > /* Available with KVM_CAP_HYPERV_CPUID */
> > #define KVM_GET_SUPPORTED_HV_CPUID _IOWR(KVMIO, 0xc1, struct kvm_cpuid2)
> >
> > +/* Available with KVM_CAP_ARM_SVE */
> > +#define KVM_ARM_VCPU_FINALIZE _IOW(KVMIO, 0xc2, int)
> > +
> > /* Secure Encrypted Virtualization command */
> > enum sev_cmd_id {
> > /* Guest initialization commands */
> > diff --git a/virt/kvm/arm/arm.c b/virt/kvm/arm/arm.c
> > index c69e137..9edbf0f 100644
> > --- a/virt/kvm/arm/arm.c
> > +++ b/virt/kvm/arm/arm.c
> > @@ -545,6 +545,9 @@ static int kvm_vcpu_first_run_init(struct kvm_vcpu
> > *vcpu)
> > if (likely(vcpu->arch.has_run_once))
> > return 0;
> >
> > + if (!kvm_arm_vcpu_is_finalized(vcpu))
> > + return -EPERM;
> > +
> > vcpu->arch.has_run_once = true;
> >
> > if (likely(irqchip_in_kernel(kvm))) {
> > @@ -1116,6 +1119,10 @@ long kvm_arch_vcpu_ioctl(struct file *filp,
> > if (unlikely(!kvm_vcpu_initialized(vcpu)))
> > break;
> >
> > + r = -EPERM;
> > + if (!kvm_arm_vcpu_is_finalized(vcpu))
> > + break;
> > +
>
> What's the rationale for using EPERM? The finalized concept is very
> similar to the initialized one. So why not also use ENOEXEC for it too?
Hmm, I guess we could equally return ENOEXEC. Initially this felt like
a more distinctive case.
Assuming Marc is happy to take an ABI fix into kvmarm/next, I'm can
change them. We're not absolutely committed until this hits mainline...
> > r = -EFAULT;
> > if (copy_from_user(®_list, user_list, sizeof(reg_list)))
> > break;
> > @@ -1169,6 +1176,17 @@ long kvm_arch_vcpu_ioctl(struct file *filp,
> >
> > return kvm_arm_vcpu_set_events(vcpu, &events);
> > }
> > + case KVM_ARM_VCPU_FINALIZE: {
> > + int what;
> > +
> > + if (!kvm_vcpu_initialized(vcpu))
> > + return -ENOEXEC;
> > +
> > + if (get_user(what, (const int __user *)argp))
> > + return -EFAULT;
> > +
> > + return kvm_arm_vcpu_finalize(vcpu, what);
>
> Almost all the cases use the 'r = ...; break;' type of pattern, leaving
> it to the 'return r' at the end of the function. I guess that's in case
> at some point more stuff is added after the switch. The only cases that
> don't do that are the most recent ones KVM_GET/SET_VCPU_EVENTS, which
> should probably be changed to fit the pattern too, rather than this
> new ioctl following there pattern.
I have no strong opinion on this: it's basically a question of style.
I followed KVM_GET/SET_VCPU_EVENTS, but you're right, the
r = ...; break; style is used for the others.
If there's an intention of putting stuff at the end of the function,
it will make a difference. But this seems unlikely to happen: this
function is really just a dispatcher.
I'm happy to change it (and the others) if there are strong views.
Cheers
---Dave
_______________________________________________
kvmarm mailing list
[email protected]
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm