does the vm that fails migration have a live snapshot?
if so how many snapshots does the vm have.
I think that there are newer packages of vdsm, libvirt and qemu - can you try to update


On 02/16/2014 12:33 AM, Steve Dainard wrote:
Versions are the same:

[root@ovirt001 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
gpxe-roms-qemu-0.9.7-6.10.el6.noarch
libvirt-0.10.2-29.el6_5.3.x86_64
libvirt-client-0.10.2-29.el6_5.3.x86_64
libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
libvirt-python-0.10.2-29.el6_5.3.x86_64
qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
vdsm-4.13.3-3.el6.x86_64
vdsm-cli-4.13.3-3.el6.noarch
vdsm-gluster-4.13.3-3.el6.noarch
vdsm-python-4.13.3-3.el6.x86_64
vdsm-xmlrpc-4.13.3-3.el6.noarch

[root@ovirt002 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
gpxe-roms-qemu-0.9.7-6.10.el6.noarch
libvirt-0.10.2-29.el6_5.3.x86_64
libvirt-client-0.10.2-29.el6_5.3.x86_64
libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
libvirt-python-0.10.2-29.el6_5.3.x86_64
qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
vdsm-4.13.3-3.el6.x86_64
vdsm-cli-4.13.3-3.el6.noarch
vdsm-gluster-4.13.3-3.el6.noarch
vdsm-python-4.13.3-3.el6.x86_64
vdsm-xmlrpc-4.13.3-3.el6.noarch

Logs attached, thanks.

*Steve Dainard *
IT Infrastructure Manager
Miovision <http://miovision.com/> | /Rethink Traffic/

*Blog <http://miovision.com/blog> | **LinkedIn <https://www.linkedin.com/company/miovision-technologies> | Twitter <https://twitter.com/miovision> | Facebook <https://www.facebook.com/miovision>*
------------------------------------------------------------------------
Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON, Canada | N2C 1L3 This e-mail may contain information that is privileged or confidential. If you are not the intended recipient, please delete the e-mail and any attachments and notify us immediately.


On Sat, Feb 15, 2014 at 6:24 AM, Dafna Ron <d...@redhat.com <mailto:d...@redhat.com>> wrote:

    the migration fails in libvirt:


    Thread-153709::ERROR::2014-02-14
    11:17:40,420::vm::337::vm.Vm::(run)
    vmId=`08434c90-ffa3-4b63-aa8e-5613f7b0e0cd`::Failed to migrate
    Traceback (most recent call last):
      File "/usr/share/vdsm/vm.py", line 323, in run
        self._startUnderlyingMigration()
      File "/usr/share/vdsm/vm.py", line 403, in _startUnderlyingMigration
        None, maxBandwidth)
      File "/usr/share/vdsm/vm.py", line 841, in f
        ret = attr(*args, **kwargs)
      File
    "/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py",
    line 76, in wrapper
        ret = f(*args, **kwargs)
      File "/usr/lib64/python2.6/site-packages/libvirt.py", line 1178,
    in migrateToURI2
        if ret == -1: raise libvirtError ('virDomainMigrateToURI2()
    failed', dom=self)
    libvirtError: Unable to read from monitor: Connection reset by peer
    Thread-54041::DEBUG::2014-02-14
    11:17:41,752::task::579::TaskManager.Task::(_updateState)
    Task=`094c412a-43dc-4c29-a601-d759486469a8`::moving from state
    init -> state preparing
    Thread-54041::INFO::2014-02-14
    11:17:41,753::logUtils::44::dispatcher::(wrapper) Run and protect:
    getVolumeSize(sdUUID='a52938f7-2cf4-4771-acb2-0c78d14999e5',
    spUUID='fcb89071-6cdb-4972-94d1-c9324cebf814',
    imgUUID='97c9108f-a506-415f-ad2
    c-370d707cb130', volUUID='61f82f7f-18e4-4ea8-9db3-71ddd9d4e836',
    options=None)

    Do you have the same libvirt/vdsm/qemu on both your hosts?
    Please attach the libvirt and vm logs from both hosts.

    Thanks,
    Dafna



    On 02/14/2014 04:50 PM, Steve Dainard wrote:

        Quick overview:
        Ovirt 3.3.2 running on CentOS 6.5
        Two hosts: ovirt001, ovirt002
        Migrating two VM's: puppet-agent1, puppet-agent2 from ovirt002
        to ovirt001.

        The first VM puppet-agent1 migrates successfully. The second
        VM puppet-agent2 fails with "Migration failed due to Error:
        Fatal error during migration (VM: puppet-agent2, Source:
        ovirt002, Destination: ovirt001)."

        I've attached the logs if anyone can help me track down the issue.

        Thanks,

        *Steve Dainard *
        IT Infrastructure Manager
        Miovision <http://miovision.com/> | /Rethink Traffic/

        *Blog <http://miovision.com/blog> | **LinkedIn
        <https://www.linkedin.com/company/miovision-technologies>  |
        Twitter <https://twitter.com/miovision>  | Facebook
        <https://www.facebook.com/miovision>*
        ------------------------------------------------------------------------


        Miovision Technologies Inc. | 148 Manitou Drive, Suite 101,
        Kitchener, ON, Canada | N2C 1L3
        This e-mail may contain information that is privileged or
        confidential. If you are not the intended recipient, please
        delete the e-mail and any attachments and notify us immediately.


        _______________________________________________
        Users mailing list
        Users@ovirt.org <mailto:Users@ovirt.org>
        http://lists.ovirt.org/mailman/listinfo/users



-- Dafna Ron




--
Dafna Ron
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to