On Tue, Aug 27, 2019 at 09:10:13PM +0800, Peter Xu wrote:
> Since we've just removed the dependency of vm type in previous patch,
> now we can create the vm much earlier.  Note that to move it earlier
> we used an approximation of number of extra pages but it should be
> fine.
> 
> This prepares for the follow up patches to finally remove the
> duplication of guest mode parsings.
> 
> Signed-off-by: Peter Xu <pet...@redhat.com>
> ---
>  tools/testing/selftests/kvm/dirty_log_test.c | 19 ++++++++++++++++---
>  1 file changed, 16 insertions(+), 3 deletions(-)
> 
> diff --git a/tools/testing/selftests/kvm/dirty_log_test.c 
> b/tools/testing/selftests/kvm/dirty_log_test.c
> index 424efcf8c734..040952f3d4ad 100644
> --- a/tools/testing/selftests/kvm/dirty_log_test.c
> +++ b/tools/testing/selftests/kvm/dirty_log_test.c
> @@ -264,6 +264,9 @@ static struct kvm_vm *create_vm(enum vm_guest_mode mode, 
> uint32_t vcpuid,
>       return vm;
>  }
>  
> +#define DIRTY_MEM_BITS 30 /* 1G */
> +#define PAGE_SHIFT_4K  12
> +
>  static void run_test(enum vm_guest_mode mode, unsigned long iterations,
>                    unsigned long interval, uint64_t phys_offset)
>  {
> @@ -273,6 +276,18 @@ static void run_test(enum vm_guest_mode mode, unsigned 
> long iterations,
>       uint64_t max_gfn;
>       unsigned long *bmap;
>  
> +     /*
> +      * We reserve page table for 2 times of extra dirty mem which
> +      * will definitely cover the original (1G+) test range.  Here
> +      * we do the calculation with 4K page size which is the
> +      * smallest so the page number will be enough for all archs
> +      * (e.g., 64K page size guest will need even less memory for
> +      * page tables).
> +      */
> +     vm = create_vm(mode, VCPU_ID,
> +                    2ul << (DIRTY_MEM_BITS - PAGE_SHIFT_4K),
> +                    guest_code);
> +
>       switch (mode) {
>       case VM_MODE_P52V48_4K:
>               guest_pa_bits = 52;
> @@ -319,7 +334,7 @@ static void run_test(enum vm_guest_mode mode, unsigned 
> long iterations,
>        * A little more than 1G of guest page sized pages.  Cover the
>        * case where the size is not aligned to 64 pages.
>        */
> -     guest_num_pages = (1ul << (30 - guest_page_shift)) + 16;
> +     guest_num_pages = (1ul << (DIRTY_MEM_BITS - guest_page_shift)) + 16;
>  #ifdef __s390x__
>       /* Round up to multiple of 1M (segment size) */
>       guest_num_pages = (guest_num_pages + 0xff) & ~0xffUL;
> @@ -345,8 +360,6 @@ static void run_test(enum vm_guest_mode mode, unsigned 
> long iterations,
>       bmap = bitmap_alloc(host_num_pages);
>       host_bmap_track = bitmap_alloc(host_num_pages);
>  
> -     vm = create_vm(mode, VCPU_ID, guest_num_pages, guest_code);
> -
>  #ifdef USE_CLEAR_DIRTY_LOG
>       struct kvm_enable_cap cap = {};
>  
> -- 
> 2.21.0
>

Reviewed-by: Andrew Jones <drjo...@redhat.com>

Reply via email to