Thanks Rene, I'll look into your suggestions. I don't think it's a conflicting MAC as there is only 1 guest, but will check it out.
Would you advise running the engine and nodes on a separate network range to my existing network rather? Thanks. Regards. Neil Wilson. On Fri, Mar 8, 2013 at 1:52 PM, René Koch (ovido) <[email protected]> wrote: > Hi Neil, > > I had a similar issue with my oVirt environment with some vms. > The issue on my side was oVirt and RHEV environment in the same subnet > and conflicting MAC addresses on some vms (as both use the same MAC > range and I didn't change this with engine-config). > > So can you check if this MAC our your vm is in use by an other host/vm > (maybe from a KVM installation)? > > > -- > Best regards > > René Koch > Senior Solution Architect > > ============================================ > ovido gmbh - "Das Linux Systemhaus" > Brünner Straße 163, A-1210 Wien > > Phone: +43 720 / 530 670 > Mobile: +43 660 / 512 21 31 > E-Mail: [email protected] > ============================================ > > > On Fri, 2013-03-08 at 11:27 +0200, Neil wrote: >> Hi guys, >> >> I've got a bit of a strange one, I'm setting up an internal ovirt >> system Centos 6.3 64bit dreyou repo... >> >> and I'm getting lots of packet loss on the guest I've installed, the >> packet loss doesn't happen on the physical hosts, only the VM gets it, >> when communicating from and to it. >> >> 1 node(Centos 6.3 64bit) >> vdsm-4.10.0-0.46.15.el6.x86_64 >> vdsm-cli-4.10.0-0.46.15.el6.noarch >> vdsm-xmlrpc-4.10.0-0.46.15.el6.noarch >> vdsm-python-4.10.0-0.46.15.el6.x86_64 >> >> The engine(also Centos 6.3 64bit(engine has local NFS storage which >> the node connects to) >> ovirt-engine-userportal-3.1.0-3.19.el6.noarch >> ovirt-engine-dbscripts-3.1.0-3.19.el6.noarch >> ovirt-engine-3.1.0-3.19.el6.noarch >> ovirt-engine-backend-3.1.0-3.19.el6.noarch >> ovirt-engine-notification-service-3.1.0-3.19.el6.noarch >> ovirt-image-uploader-3.1.0-16.el6.noarch >> ovirt-engine-genericapi-3.1.0-3.19.el6.noarch >> ovirt-iso-uploader-3.1.0-16.el6.noarch >> ovirt-engine-restapi-3.1.0-3.19.el6.noarch >> ovirt-engine-tools-common-3.1.0-3.19.el6.noarch >> ovirt-engine-sdk-3.2.0.8-1.el6.noarch >> ovirt-engine-webadmin-portal-3.1.0-3.19.el6.noarch >> ovirt-engine-cli-3.2.0.5-1.el6.noarch >> ovirt-log-collector-3.1.0-16.el6.noarch >> ovirt-engine-setup-3.1.0-3.19.el6.noarch >> ovirt-engine-jbossas711-1-0.x86_64 >> ovirt-engine-config-3.1.0-3.19.el6.noarch >> >> Both the node and engine have bonded interfaces all NICS are intel >> 82574L gigabit and the managed switch reflects gigabit on each of the >> ports. >> >> The ifcfg-bond0 is below... >> >> DEVICE=bond0 >> IPADDR=192.168.0.9 >> NETWORK=192.168.0.0 >> NETMASK=255.255.255.0 >> USERCTL=no >> BONDING_OPTS=mode=2 >> BOOTPROTO=none >> MTU=1500 >> ONBOOT=yes >> >> Then the ifcfg-eth0 and eth1 are almost identical... >> DEVICE=eth2 >> USERCTL=no >> ONBOOT=yes >> MASTER=bond0 >> SLAVE=yes >> MTU=1500 >> BOOTPROTO=none >> >> >> These are the network details on the guest, as you can see, there are >> no network errors showing on the guest at all, which is strange.... >> >> eth0 Link encap:Ethernet HWaddr 00:1A:4A:A8:00:00 >> inet addr:192.168.0.12 Bcast:192.168.0.255 Mask:255.255.255.0 >> inet6 addr: fe80::21a:4aff:fea8:0/64 Scope:Link >> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >> RX packets:5050 errors:0 dropped:0 overruns:0 frame:0 >> TX packets:255 errors:0 dropped:0 overruns:0 carrier:0 >> collisions:0 txqueuelen:1000 >> RX bytes:490762 (479.2 KiB) TX bytes:32516 (31.7 KiB) >> >> Ethernet controller: Red Hat, Inc Virtio network device >> >> Has anyone got any ideas? Have I set something up wrong? >> >> Any help or advice is greatly appreciated. >> >> Regards. >> >> Neil Wilson. >> _______________________________________________ >> Users mailing list >> [email protected] >> http://lists.ovirt.org/mailman/listinfo/users > _______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

