On 2/18/2020 5:22 PM, Daniel Cho wrote:
Hi Hailiang,
Thanks for your help. If we have any problems we will contact you for your favor.


Hi Zhang,

" If colo-compare got a primary packet without related secondary packet in a certain time , it will automatically trigger checkpoint.  " As you said, the colo-compare will trigger checkpoint, but does it need to limit checkpoint times? There is a problem about doing many checkpoints while we use fio to random write files. Then it will cause low throughput on PVM.
Is this situation is normal on COLO?


Hi Daniel,

The checkpoint time is designed to be user adjustable based on user environment(workload/network status/business conditions...).

In net/colo-compare.c

/* TODO: Should be configurable */
#define REGULAR_PACKET_CHECK_MS 3000

If you need, I can send a patch for this issue. Make users can change the value by QMP and qemu monitor commands.

Thanks

Zhang Chen



Best regards,
Daniel Cho

Zhang, Chen <chen.zh...@intel.com <mailto:chen.zh...@intel.com>> 於 2020年2月17日 週一 下午1:36寫道:


    On 2/15/2020 11:35 AM, Daniel Cho wrote:
    Hi Dave,

    Yes, I agree with you, it does need a timeout.


    Hi Daniel and Dave,

    Current colo-compare already have the timeout mechanism.

    Named packet_check_timer,  It will scan primary packet queue to
    make sure all the primary packet not stay too long time.

    If colo-compare got a primary packet without related secondary
    packet in a certain time , it will automatic trigger checkpoint.

    https://github.com/qemu/qemu/blob/master/net/colo-compare.c#L847


    Thanks

    Zhang Chen



    Hi Hailiang,

    We base on qemu-4.1.0 for using COLO feature, in your patch, we
    found a lot of difference  between your version and ours.
    Could you give us a latest release version which is close your
    developing code?

    Thanks.

    Regards
    Daniel Cho

    Dr. David Alan Gilbert <dgilb...@redhat.com
    <mailto:dgilb...@redhat.com>> 於 2020年2月13日 週四 下午6:38寫道:

        * Daniel Cho (daniel...@qnap.com <mailto:daniel...@qnap.com>)
        wrote:
        > Hi Hailiang,
        >
        > 1.
        >     OK, we will try the patch
        > “0001-COLO-Optimize-memory-back-up-process.patch”,
        > and thanks for your help.
        >
        > 2.
        >     We understand the reason to compare PVM and SVM's
        packet. However, the
        > empty of SVM's packet queue might happened on setting COLO
        feature and SVM
        > broken.
        >
        > On situation 1 ( setting COLO feature ):
        >     We could force do checkpoint after setting COLO feature
        finish, then it
        > will protect the state of PVM and SVM . As the Zhang Chen said.
        >
        > On situation 2 ( SVM broken ):
        >     COLO will do failover for PVM, so it might not cause
        any wrong on PVM.
        >
        > However, those situations are our views, so there might be
        a big difference
        > between reality and our views.
        > If we have any wrong views and opinions, please let us
        know, and correct
        > us.

        It does need a timeout; the SVM being broken or being in a
        state where
        it never sends the corresponding packet (because of a state
        difference)
        can happen and COLO needs to timeout when the packet hasn't
        arrived
        after a while and trigger the checkpoint.

        Dave

        > Thanks.
        >
        > Best regards,
        > Daniel Cho
        >
        > Zhang, Chen <chen.zh...@intel.com
        <mailto:chen.zh...@intel.com>> 於 2020年2月13日 週四
        上午10:17寫道:
        >
        > > Add cc Jason Wang, he is a network expert.
        > >
        > > In case some network things goes wrong.
        > >
        > >
        > >
        > > Thanks
        > >
        > > Zhang Chen
        > >
        > >
        > >
        > > *From:* Zhang, Chen
        > > *Sent:* Thursday, February 13, 2020 10:10 AM
        > > *To:* 'Zhanghailiang' <zhang.zhanghaili...@huawei.com
        <mailto:zhang.zhanghaili...@huawei.com>>; Daniel Cho <
        > > daniel...@qnap.com <mailto:daniel...@qnap.com>>
        > > *Cc:* Dr. David Alan Gilbert <dgilb...@redhat.com
        <mailto:dgilb...@redhat.com>>; qemu-devel@nongnu.org
        <mailto:qemu-devel@nongnu.org>
        > > *Subject:* RE: The issues about architecture of the COLO
        checkpoint
        > >
        > >
        > >
        > > For the issue 2:
        > >
        > >
        > >
        > > COLO need use the network packets to confirm PVM and SVM
        in the same state,
        > >
        > > Generally speaking, we can’t send PVM packets without
        compared with SVM
        > > packets.
        > >
        > > But to prevent jamming, I think COLO can do force
        checkpoint and send the
        > > PVM packets in this case.
        > >
        > >
        > >
        > > Thanks
        > >
        > > Zhang Chen
        > >
        > >
        > >
        > > *From:* Zhanghailiang <zhang.zhanghaili...@huawei.com
        <mailto:zhang.zhanghaili...@huawei.com>>
        > > *Sent:* Thursday, February 13, 2020 9:45 AM
        > > *To:* Daniel Cho <daniel...@qnap.com
        <mailto:daniel...@qnap.com>>
        > > *Cc:* Dr. David Alan Gilbert <dgilb...@redhat.com
        <mailto:dgilb...@redhat.com>>; qemu-devel@nongnu.org
        <mailto:qemu-devel@nongnu.org>;
        > > Zhang, Chen <chen.zh...@intel.com
        <mailto:chen.zh...@intel.com>>
        > > *Subject:* RE: The issues about architecture of the COLO
        checkpoint
        > >
        > >
        > >
        > > Hi,
        > >
        > >
        > >
        > > 1.       After re-walked through the codes, yes, you are
        right, actually,
        > > after the first migration, we will keep dirty log on in
        primary side,
        > >
        > > And only send the dirty pages in PVM to SVM. The ram
        cache in secondary
        > > side is always a backup of PVM, so we don’t have to
        > >
        > > Re-send the none-dirtied pages.
        > >
        > > The reason why the first checkpoint takes longer time is
        we have to backup
        > > the whole VM’s ram into ram cache, that is
        colo_init_ram_cache().
        > >
        > > It is time consuming, but I have optimized in the second
        patch
        > > “0001-COLO-Optimize-memory-back-up-process.patch” which
        you can find in my
        > > previous reply.
        > >
        > >
        > >
        > > Besides, I found that, In my previous reply “We can only
        copy the pages
        > > that dirtied by PVM and SVM in last checkpoint.”,
        > >
        > > We have done this optimization in current upstream codes.
        > >
        > >
        > >
        > > 2.I don’t quite understand this question. For COLO, we
        always need both
        > > network packets of PVM’s and SVM’s to compare before send
        this packets to
        > > client.
        > >
        > > It depends on this to decide whether or not PVM and SVM
        are in same state.
        > >
        > >
        > >
        > > Thanks,
        > >
        > > hailiang
        > >
        > >
        > >
        > > *From:* Daniel Cho [mailto:daniel...@qnap.com
        <mailto:daniel...@qnap.com> <daniel...@qnap.com
        <mailto:daniel...@qnap.com>>]
        > > *Sent:* Wednesday, February 12, 2020 4:37 PM
        > > *To:* Zhang, Chen <chen.zh...@intel.com
        <mailto:chen.zh...@intel.com>>
        > > *Cc:* Zhanghailiang <zhang.zhanghaili...@huawei.com
        <mailto:zhang.zhanghaili...@huawei.com>>; Dr. David Alan
        > > Gilbert <dgilb...@redhat.com
        <mailto:dgilb...@redhat.com>>; qemu-devel@nongnu.org
        <mailto:qemu-devel@nongnu.org>
        > > *Subject:* Re: The issues about architecture of the COLO
        checkpoint
        > >
        > >
        > >
        > > Hi Hailiang,
        > >
        > >
        > >
        > > Thanks for your replaying and explain in detail.
        > >
        > > We will try to use the attachments to enhance memory copy.
        > >
        > >
        > >
        > > However, we have some questions for your replying.
        > >
        > >
        > >
        > > 1.  As you said, "for each checkpoint, we have to send
        the whole PVM's
        > > pages To SVM", why the only first checkpoint will takes
        more pause time?
        > >
        > > In our observing, the first checkpoint will take more
        time for pausing,
        > > then other checkpoints will takes a few time for pausing.
        Does it means
        > > only the first checkpoint will send the whole pages to
        SVM, and the other
        > > checkpoints send the dirty pages to SVM for reloading?
        > >
        > >
        > >
        > > 2. We notice the COLO-COMPARE component will stuck the
        packet until
        > > receive packets from PVM and SVM, as this rule, when we
        add the
        > > COLO-COMPARE to PVM, its network will stuck until SVM
        start. So it is an
        > > other issue to make PVM stuck while setting COLO feature.
        With this issue,
        > > could we let colo-compare to pass the PVM's packet when
        the SVM's packet
        > > queue is empty? Then, the PVM's network won't stock, and
        "if PVM runs
        > > firstly, it still need to wait for The network packets
        from SVM to
        > > compare before send it to client side" won't happened either.
        > >
        > >
        > >
        > > Best regard,
        > >
        > > Daniel Cho
        > >
        > >
        > >
        > > Zhang, Chen <chen.zh...@intel.com
        <mailto:chen.zh...@intel.com>> 於 2020年2月12日 週三
        下午1:45寫道:
        > >
        > >
        > >
        > > > -----Original Message-----
        > > > From: Zhanghailiang <zhang.zhanghaili...@huawei.com
        <mailto:zhang.zhanghaili...@huawei.com>>
        > > > Sent: Wednesday, February 12, 2020 11:18 AM
        > > > To: Dr. David Alan Gilbert <dgilb...@redhat.com
        <mailto:dgilb...@redhat.com>>; Daniel Cho
        > > > <daniel...@qnap.com <mailto:daniel...@qnap.com>>;
        Zhang, Chen <chen.zh...@intel.com <mailto:chen.zh...@intel.com>>
        > > > Cc: qemu-devel@nongnu.org <mailto:qemu-devel@nongnu.org>
        > > > Subject: RE: The issues about architecture of the COLO
        checkpoint
        > > >
        > > > Hi,
        > > >
        > > > Thank you Dave,
        > > >
        > > > I'll reply here directly.
        > > >
        > > > -----Original Message-----
        > > > From: Dr. David Alan Gilbert
        [mailto:dgilb...@redhat.com <mailto:dgilb...@redhat.com>]
        > > > Sent: Wednesday, February 12, 2020 1:48 AM
        > > > To: Daniel Cho <daniel...@qnap.com
        <mailto:daniel...@qnap.com>>; chen.zh...@intel.com
        <mailto:chen.zh...@intel.com>;
        > > > Zhanghailiang <zhang.zhanghaili...@huawei.com
        <mailto:zhang.zhanghaili...@huawei.com>>
        > > > Cc: qemu-devel@nongnu.org <mailto:qemu-devel@nongnu.org>
        > > > Subject: Re: The issues about architecture of the COLO
        checkpoint
        > > >
        > > >
        > > > cc'ing in COLO people:
        > > >
        > > >
        > > > * Daniel Cho (daniel...@qnap.com
        <mailto:daniel...@qnap.com>) wrote:
        > > > > Hi everyone,
        > > > >      We have some issues about setting COLO feature.
        Hope somebody
        > > > > could give us some advice.
        > > > >
        > > > > Issue 1:
        > > > >      We dynamic to set COLO feature for PVM(2 core,
        16G memory),  but
        > > > > the Primary VM will pause a long time(based on memory
        size) for
        > > > > waiting SVM start. Does it have any idea to reduce
        the pause time?
        > > > >
        > > >
        > > > Yes, we do have some ideas to optimize this downtime.
        > > >
        > > > The main problem for current version is, for each
        checkpoint, we have to
        > > > send the whole PVM's pages
        > > > To SVM, and then copy the whole VM's state into SVM
        from ram cache, in
        > > > this process, we need both of them be paused.
        > > > Just as you said, the downtime is based on memory size.
        > > >
        > > > So firstly, we need to reduce the sending data while do
        checkpoint,
        > > actually,
        > > > we can migrate parts of PVM's dirty pages in background
        > > > While both of VMs are running. And then we load these
        pages into ram
        > > > cache (backup memory) in SVM temporarily. While do
        checkpoint,
        > > > We just send the last dirty pages of PVM to slave side
        and then copy the
        > > ram
        > > > cache into SVM. Further on, we don't have
        > > > To send the whole PVM's dirty pages, we can only send
        the pages that
        > > > dirtied by PVM or SVM during two checkpoints. (Because
        > > > If one page is not dirtied by both PVM and SVM, the
        data of this pages
        > > will
        > > > keep same in SVM, PVM, backup memory). This method can
        reduce
        > > > the time that consumed in sending data.
        > > >
        > > > For the second problem, we can reduce the memory copy
        by two methods,
        > > > first one, we don't have to copy the whole pages in ram
        cache,
        > > > We can only copy the pages that dirtied by PVM and SVM
        in last
        > > checkpoint.
        > > > Second, we can use userfault missing function to reduce the
        > > > Time consumed in memory copy. (For the second time, in
        theory, we can
        > > > reduce time consumed in memory into ms level).
        > > >
        > > > You can find the first optimization in attachment, it
        is based on an old
        > > qemu
        > > > version (qemu-2.6), it should not be difficult to rebase it
        > > > Into master or your version. And please feel free to
        send the new
        > > version if
        > > > you want into community ;)
        > > >
        > > >
        > >
        > > Thanks Hailiang!
        > > By the way, Do you have time to push the patches to upstream?
        > > I think this is a better and faster option.
        > >
        > > Thanks
        > > Zhang Chen
        > >
        > > > >
        > > > > Issue 2:
        > > > >      In
        > > > >
        https://github.com/qemu/qemu/blob/master/migration/colo.c#L503,
        > > > > could we move start_vm() before Line 488? Because at
        first checkpoint
        > > > > PVM will wait for SVM's reply, it cause PVM stop for
        a while.
        > > > >
        > > >
        > > > No, that makes no sense, because if PVM runs firstly,
        it still need to
        > > wait for
        > > > The network packets from SVM to compare before send it
        to client side.
        > > >
        > > >
        > > > Thanks,
        > > > Hailiang
        > > >
        > > > >      We set the COLO feature on running VM, so we
        hope the running VM
        > > > > could continuous service for users.
        > > > > Do you have any suggestions for those issues?
        > > > >
        > > > > Best regards,
        > > > > Daniel Cho
        > > > --
        > > > Dr. David Alan Gilbert / dgilb...@redhat.com
        <mailto:dgilb...@redhat.com> / Manchester, UK
        > >
        > >
        --
        Dr. David Alan Gilbert / dgilb...@redhat.com
        <mailto:dgilb...@redhat.com> / Manchester, UK

Reply via email to