Hello,

I'm trying to setup a host as a self-hosted engine as per 
https://www.ovirt.org/documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/
 
<https://www.ovirt.org/documentation/self-hosted/chap-Deploying_Self-Hosted_Engine/>.
The host is configured with two bonded network interfaces:

bond0 = management network for hypervisors, setup as active/passive.
bond1 = network that has VLANs for various network segments for virtual 
machines to use, setup as LACP bond to upstream switches.

On the host, both networks are operational and work as expected.

When setting up the self-hosted engine, bond0 is selected as network to bridge 
with, and a unique IP is given to the self-hosted engine VM.

During the final stages of the self-hosted engine setup, the installer gets 
stuck on ' Waiting for the VDSM host to become operational'.
While it is repeating this every minute or so the host repeats the message 
'bond0: received packet with own address as source address', which is odd to me 
as it's in active/passive bond and I'd only expect to see this kind of message 
on XOR / load balanced interfaces.

Host console screenshot: https://imgur.com/a/a2JLd <https://imgur.com/a/a2JLd>
Host OS: CentOS 7.4
oVirt version: 4.2.0

ip a on host while install is stuck waiting for VDSM:

# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1000
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
       valid_lft forever preferred_lft forever
    inet6 ::1/128 scope host
       valid_lft forever preferred_lft forever
2: enp2s0f0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 9000 qdisc mq master 
bond1 portid 0100000000000000000000373031384833 state UP qlen 1000
    link/ether 78:e3:b5:10:74:88 brd ff:ff:ff:ff:ff:ff
3: enp2s0f1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 9000 qdisc mq master 
bond1 portid 0200000000000000000000373031384833 state UP qlen 1000
    link/ether 78:e3:b5:10:74:88 brd ff:ff:ff:ff:ff:ff
4: ens1f0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master 
bond0 portid 0100000000000000000000474835323543 state UP qlen 1000
    link/ether 00:9c:02:3c:49:90 brd ff:ff:ff:ff:ff:ff
5: ens1f1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master 
bond0 portid 0200000000000000000000474835323543 state UP qlen 1000
    link/ether 00:9c:02:3c:49:90 brd ff:ff:ff:ff:ff:ff
6: bond1: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 9000 qdisc noqueue state 
UP qlen 1000
    link/ether 78:e3:b5:10:74:88 brd ff:ff:ff:ff:ff:ff
    inet6 fe80::7ae3:b5ff:fe10:7488/64 scope link
       valid_lft forever preferred_lft forever
7: storage@bond1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc noqueue 
state UP qlen 1000
    link/ether 78:e3:b5:10:74:88 brd ff:ff:ff:ff:ff:ff
    inet 10.51.40.172/24 brd 10.51.40.255 scope global storage
       valid_lft forever preferred_lft forever
    inet6 fe80::7ae3:b5ff:fe10:7488/64 scope link
       valid_lft forever preferred_lft forever
9: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether ae:c0:02:25:42:24 brd ff:ff:ff:ff:ff:ff
10: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether be:92:5d:c3:28:4d brd ff:ff:ff:ff:ff:ff
30: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue 
master ovirtmgmt state UP qlen 1000
    link/ether 00:9c:02:3c:49:90 brd ff:ff:ff:ff:ff:ff
46: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state 
UP qlen 1000
    link/ether 00:9c:02:3c:49:90 brd ff:ff:ff:ff:ff:ff
    inet 10.51.14.112/24 brd 10.51.14.255 scope global ovirtmgmt
       valid_lft forever preferred_lft forever
47: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
    link/ether 8e:c0:25:88:40:de brd ff:ff:ff:ff:ff:ff


Before the self-hosted engine install was run the following did not exist:

;vdsmdummy;
ovs-system
br-int
ovirtmgmt

and bond0 was not a slave of ovirtmgmt.

I'm now going to kick off a complete reinstall of CentOS 7 on the host as I've 
since tried cleaning up the host using the ovirt-hosted-engine-cleanup command 
and removing the packages which seem to leave the network configuration a mess 
an doesn't actually cleanup files on disk as expected.


--
Sam McLeod (protoporpoise on IRC)
https://smcleod.net
https://twitter.com/s_mcleod

Words are my own opinions and do not necessarily represent those of my employer 
or partners.

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to