Re: [Qemu-devel] [PATCH untested for-4.2] memory: fix race between TCG and accesses to dirty bitmap

2019-08-08 Thread Alex Bennée


Paolo Bonzini  writes:

> The race is as follows:
>
>   vCPU thread  reader thread
>   ---  ---
>   TLB check -> slow path
> notdirty_mem_write
>   write to RAM
>   set dirty flag
>clear dirty flag
>   TLB check -> fast path
>read memory
> write to RAM
>
> and the second write is missed by the reader.
>
> Fortunately, in order to fix it, no change is required to the
> vCPU thread.  However, the reader thread must delay the read after
> the vCPU thread has finished the write.  This can be approximated
> conservatively by run_on_cpu, which waits for the end of the current
> translation block.
>
> A similar technique is used by KVM, which has to do a synchronous TLB
> flush after doing a test-and-clear of the dirty-page flags.
>
> Reported-by: Dr. David Alan Gilbert 
> Signed-off-by: Paolo Bonzini 

Reviewed-by: Alex Bennée 
Tested-by: Alex Bennée 

> ---
> I tested this some time ago, and enough has changed that I don't
> really trust those old results.  Nevertheless, I am throwing out
> the patch so that it is not forgotten.
>
>  exec.c| 31 +++
>  include/exec/memory.h | 12 
>  memory.c  | 10 +-
>  migration/ram.c   |  1 +
>  4 files changed, 53 insertions(+), 1 deletion(-)
>
> diff --git a/exec.c b/exec.c
> index 3e78de3b8f..ae68f72da4 100644
> --- a/exec.c
> +++ b/exec.c
> @@ -198,6 +198,7 @@ typedef struct subpage_t {
>
>  static void io_mem_init(void);
>  static void memory_map_init(void);
> +static void tcg_log_global_after_sync(MemoryListener *listener);
>  static void tcg_commit(MemoryListener *listener);
>
>  static MemoryRegion io_mem_watch;
> @@ -906,6 +907,7 @@ void cpu_address_space_init(CPUState *cpu, int asidx,
>  newas->cpu = cpu;
>  newas->as = as;
>  if (tcg_enabled()) {
> +newas->tcg_as_listener.log_global_after_sync = 
> tcg_log_global_after_sync;
>  newas->tcg_as_listener.commit = tcg_commit;
>  memory_listener_register(>tcg_as_listener, as);
>  }
> @@ -3143,6 +3145,35 @@ void address_space_dispatch_free(AddressSpaceDispatch 
> *d)
>  g_free(d);
>  }
>
> +static void do_nothing(CPUState *cpu, run_on_cpu_data d)
> +{
> +}
> +
> +static void tcg_log_global_after_sync(MemoryListener *listener)
> +{
> +CPUAddressSpace *cpuas;
> +
> +/* Wait for the CPU to end the current TB.  This avoids the following
> + * incorrect race:
> + *
> + *  vCPU migration
> + *  --   -
> + *  TLB check -> slow path
> + *notdirty_mem_write
> + *  write to RAM
> + *  mark dirty
> + *   clear dirty flag
> + *  TLB check -> fast path
> + *   read memory
> + *write to RAM
> + *
> + * by pushing the migration thread's memory read after the vCPU thread 
> has
> + * written the memory.
> + */
> +cpuas = container_of(listener, CPUAddressSpace, tcg_as_listener);
> +run_on_cpu(cpuas->cpu, do_nothing, RUN_ON_CPU_NULL);
> +}
> +
>  static void tcg_commit(MemoryListener *listener)
>  {
>  CPUAddressSpace *cpuas;
> diff --git a/include/exec/memory.h b/include/exec/memory.h
> index bb0961ddb9..b6bcf31b0a 100644
> --- a/include/exec/memory.h
> +++ b/include/exec/memory.h
> @@ -419,6 +419,7 @@ struct MemoryListener {
>  void (*log_clear)(MemoryListener *listener, MemoryRegionSection 
> *section);
>  void (*log_global_start)(MemoryListener *listener);
>  void (*log_global_stop)(MemoryListener *listener);
> +void (*log_global_after_sync)(MemoryListener *listener);
>  void (*eventfd_add)(MemoryListener *listener, MemoryRegionSection 
> *section,
>  bool match_data, uint64_t data, EventNotifier *e);
>  void (*eventfd_del)(MemoryListener *listener, MemoryRegionSection 
> *section,
> @@ -1681,6 +1682,17 @@ MemoryRegionSection memory_region_find(MemoryRegion 
> *mr,
>   */
>  void memory_global_dirty_log_sync(void);
>
> +/**
> + * memory_global_dirty_log_sync: synchronize the dirty log for all memory
> + *
> + * Synchronizes the vCPUs with a thread that is reading the dirty bitmap.
> + * This function must be called after the dirty log bitmap is cleared, and
> + * before dirty guest memory pages are read.  If you are using
> + * #DirtyBitmapSnapshot, memory_region_snapshot_and_clear_dirty() takes
> + * care of doing this.
> + */
> +void memory_global_after_dirty_log_sync(void);
> +
>  /**
>   * memory_region_transaction_begin: Start a transaction.
>   *
> diff --git a/memory.c b/memory.c
> index e42d63a3a0..edd0c13c38 100644
> --- a/memory.c
> +++ b/memory.c
> @@ -2127,9 +2127,12 @@ DirtyBitmapSnapshot 

Re: [Qemu-devel] [PATCH untested for-4.2] memory: fix race between TCG and accesses to dirty bitmap

2019-08-07 Thread Paolo Bonzini
On 06/08/19 16:23, Peter Maydell wrote:
> On Mon, 29 Jul 2019 at 22:47, Paolo Bonzini  wrote:
>>
>> The race is as follows:
>>
>>   vCPU thread  reader thread
>>   ---  ---
>>   TLB check -> slow path
>> notdirty_mem_write
>>   write to RAM
>>   set dirty flag
>>clear dirty flag
>>   TLB check -> fast path
>>read memory
>> write to RAM
>>
>> and the second write is missed by the reader.
>>
>> Fortunately, in order to fix it, no change is required to the
>> vCPU thread.  However, the reader thread must delay the read after
>> the vCPU thread has finished the write.  This can be approximated
>> conservatively by run_on_cpu, which waits for the end of the current
>> translation block.
>>
>> A similar technique is used by KVM, which has to do a synchronous TLB
>> flush after doing a test-and-clear of the dirty-page flags.
>>
>> Reported-by: Dr. David Alan Gilbert 
>> Signed-off-by: Paolo Bonzini 
>> ---
>> I tested this some time ago, and enough has changed that I don't
>> really trust those old results.  Nevertheless, I am throwing out
>> the patch so that it is not forgotten.
> 
> This patch looks almost the same (maybe identical except for the
> commit message title?) as the patch "memory: introduce
> memory_global_after_dirty_log_sync" which you sent out at almost
> the same time as this one. Which patch should we be reviewing?

Yes, it's the same except for the commit message title.  I forgot a "-1"
after editing the .patch file.

Paolo




Re: [Qemu-devel] [PATCH untested for-4.2] memory: fix race between TCG and accesses to dirty bitmap

2019-08-06 Thread Peter Maydell
On Mon, 29 Jul 2019 at 22:47, Paolo Bonzini  wrote:
>
> The race is as follows:
>
>   vCPU thread  reader thread
>   ---  ---
>   TLB check -> slow path
> notdirty_mem_write
>   write to RAM
>   set dirty flag
>clear dirty flag
>   TLB check -> fast path
>read memory
> write to RAM
>
> and the second write is missed by the reader.
>
> Fortunately, in order to fix it, no change is required to the
> vCPU thread.  However, the reader thread must delay the read after
> the vCPU thread has finished the write.  This can be approximated
> conservatively by run_on_cpu, which waits for the end of the current
> translation block.
>
> A similar technique is used by KVM, which has to do a synchronous TLB
> flush after doing a test-and-clear of the dirty-page flags.
>
> Reported-by: Dr. David Alan Gilbert 
> Signed-off-by: Paolo Bonzini 
> ---
> I tested this some time ago, and enough has changed that I don't
> really trust those old results.  Nevertheless, I am throwing out
> the patch so that it is not forgotten.

This patch looks almost the same (maybe identical except for the
commit message title?) as the patch "memory: introduce
memory_global_after_dirty_log_sync" which you sent out at almost
the same time as this one. Which patch should we be reviewing?

thanks
-- PMM