did you install these vm's from a cd? run it as run-once with a special monitor? try to think if there is anything different in the configuration of these vm's from the other vm's that succeed to migrate?

On 02/17/2014 04:36 PM, Steve Dainard wrote:
Hi Dafna,

No snapshots of either of those VM's have been taken, and there are no updates for any of those packages on EL 6.5.

*Steve Dainard *
IT Infrastructure Manager
Miovision <http://miovision.com/> | /Rethink Traffic/

*Blog <http://miovision.com/blog> | **LinkedIn <https://www.linkedin.com/company/miovision-technologies> | Twitter <https://twitter.com/miovision> | Facebook <https://www.facebook.com/miovision>*
------------------------------------------------------------------------
Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON, Canada | N2C 1L3 This e-mail may contain information that is privileged or confidential. If you are not the intended recipient, please delete the e-mail and any attachments and notify us immediately.


On Sun, Feb 16, 2014 at 7:05 AM, Dafna Ron <[email protected] <mailto:[email protected]>> wrote:

    does the vm that fails migration have a live snapshot?
    if so how many snapshots does the vm have.
    I think that there are newer packages of vdsm, libvirt and qemu -
    can you try to update



    On 02/16/2014 12:33 AM, Steve Dainard wrote:

        Versions are the same:

        [root@ovirt001 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
        gpxe-roms-qemu-0.9.7-6.10.el6.noarch
        libvirt-0.10.2-29.el6_5.3.x86_64
        libvirt-client-0.10.2-29.el6_5.3.x86_64
        libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
        libvirt-python-0.10.2-29.el6_5.3.x86_64
        qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
        qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
        qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
        vdsm-4.13.3-3.el6.x86_64
        vdsm-cli-4.13.3-3.el6.noarch
        vdsm-gluster-4.13.3-3.el6.noarch
        vdsm-python-4.13.3-3.el6.x86_64
        vdsm-xmlrpc-4.13.3-3.el6.noarch

        [root@ovirt002 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
        gpxe-roms-qemu-0.9.7-6.10.el6.noarch
        libvirt-0.10.2-29.el6_5.3.x86_64
        libvirt-client-0.10.2-29.el6_5.3.x86_64
        libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
        libvirt-python-0.10.2-29.el6_5.3.x86_64
        qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
        qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
        qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
        vdsm-4.13.3-3.el6.x86_64
        vdsm-cli-4.13.3-3.el6.noarch
        vdsm-gluster-4.13.3-3.el6.noarch
        vdsm-python-4.13.3-3.el6.x86_64
        vdsm-xmlrpc-4.13.3-3.el6.noarch

        Logs attached, thanks.

        *Steve Dainard *
        IT Infrastructure Manager
        Miovision <http://miovision.com/> | /Rethink Traffic/

        *Blog <http://miovision.com/blog> | **LinkedIn
        <https://www.linkedin.com/company/miovision-technologies>  |
        Twitter <https://twitter.com/miovision>  | Facebook
        <https://www.facebook.com/miovision>*
        ------------------------------------------------------------------------
        Miovision Technologies Inc. | 148 Manitou Drive, Suite 101,
        Kitchener, ON, Canada | N2C 1L3
        This e-mail may contain information that is privileged or
        confidential. If you are not the intended recipient, please
        delete the e-mail and any attachments and notify us immediately.


        On Sat, Feb 15, 2014 at 6:24 AM, Dafna Ron <[email protected]
        <mailto:[email protected]> <mailto:[email protected]
        <mailto:[email protected]>>> wrote:

            the migration fails in libvirt:


            Thread-153709::ERROR::2014-02-14
            11:17:40,420::vm::337::vm.Vm::(run)
            vmId=`08434c90-ffa3-4b63-aa8e-5613f7b0e0cd`::Failed to migrate
            Traceback (most recent call last):
              File "/usr/share/vdsm/vm.py", line 323, in run
                self._startUnderlyingMigration()
              File "/usr/share/vdsm/vm.py", line 403, in
        _startUnderlyingMigration
                None, maxBandwidth)
              File "/usr/share/vdsm/vm.py", line 841, in f
                ret = attr(*args, **kwargs)
              File
"/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py",
            line 76, in wrapper
                ret = f(*args, **kwargs)
              File "/usr/lib64/python2.6/site-packages/libvirt.py",
        line 1178,
            in migrateToURI2
                if ret == -1: raise libvirtError
        ('virDomainMigrateToURI2()
            failed', dom=self)
            libvirtError: Unable to read from monitor: Connection
        reset by peer
            Thread-54041::DEBUG::2014-02-14
            11:17:41,752::task::579::TaskManager.Task::(_updateState)
            Task=`094c412a-43dc-4c29-a601-d759486469a8`::moving from state
            init -> state preparing
            Thread-54041::INFO::2014-02-14
            11:17:41,753::logUtils::44::dispatcher::(wrapper) Run and
        protect:
            getVolumeSize(sdUUID='a52938f7-2cf4-4771-acb2-0c78d14999e5',
            spUUID='fcb89071-6cdb-4972-94d1-c9324cebf814',
            imgUUID='97c9108f-a506-415f-ad2
            c-370d707cb130',
        volUUID='61f82f7f-18e4-4ea8-9db3-71ddd9d4e836',
            options=None)

            Do you have the same libvirt/vdsm/qemu on both your hosts?
            Please attach the libvirt and vm logs from both hosts.

            Thanks,
            Dafna



            On 02/14/2014 04:50 PM, Steve Dainard wrote:

                Quick overview:
                Ovirt 3.3.2 running on CentOS 6.5
                Two hosts: ovirt001, ovirt002
                Migrating two VM's: puppet-agent1, puppet-agent2 from
        ovirt002
                to ovirt001.

                The first VM puppet-agent1 migrates successfully. The
        second
                VM puppet-agent2 fails with "Migration failed due to
        Error:
                Fatal error during migration (VM: puppet-agent2, Source:
                ovirt002, Destination: ovirt001)."

                I've attached the logs if anyone can help me track
        down the issue.

                Thanks,

                *Steve Dainard *
                IT Infrastructure Manager
                Miovision <http://miovision.com/> | /Rethink Traffic/

                *Blog <http://miovision.com/blog> | **LinkedIn
<https://www.linkedin.com/company/miovision-technologies> |
                Twitter <https://twitter.com/miovision>  | Facebook
                <https://www.facebook.com/miovision>*
------------------------------------------------------------------------


                Miovision Technologies Inc. | 148 Manitou Drive, Suite
        101,
                Kitchener, ON, Canada | N2C 1L3
                This e-mail may contain information that is privileged or
                confidential. If you are not the intended recipient,
        please
                delete the e-mail and any attachments and notify us
        immediately.


                _______________________________________________
                Users mailing list
        [email protected] <mailto:[email protected]>
        <mailto:[email protected] <mailto:[email protected]>>

        http://lists.ovirt.org/mailman/listinfo/users



            --     Dafna Ron




-- Dafna Ron




--
Dafna Ron
_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to