Hi oVirt land Hope you are well. Running into this issue, I hope you can help.
Centos7 and it is updated. Ovirt 4.3, latest packages. My network config: [root@mob-r1-d-ovirt-aa-1-01 ~]# ip a 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: ens1f0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master bond0 state UP group default qlen 1000 link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff 3: ens1f1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master bond0 state UP group default qlen 1000 link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff 4: enp11s0f0: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN group default qlen 1000 link/ether 00:90:fa:c2:d2:50 brd ff:ff:ff:ff:ff:ff 5: enp11s0f1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc mq state DOWN group default qlen 1000 link/ether 00:90:fa:c2:d2:54 brd ff:ff:ff:ff:ff:ff 21: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff inet6 fe80::290:faff:fec2:d248/64 scope link valid_lft forever preferred_lft forever 22: bond0.1131@bond0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000 link/ether 00:90:fa:c2:d2:48 brd ff:ff:ff:ff:ff:ff inet 172.18.206.184/23 brd 172.18.207.255 scope global bond0.1131 valid_lft forever preferred_lft forever inet6 fe80::290:faff:fec2:d248/64 scope link valid_lft forever preferred_lft forever [root@mob-r1-d-ovirt-aa-1-01 network-scripts]# cat ifcfg-bond0 BONDING_OPTS='mode=1 miimon=100' TYPE=Bond BONDING_MASTER=yes PROXY_METHOD=none BROWSER_ONLY=no IPV6INIT=no NAME=bond0 UUID=c11ef6ef-794f-4683-a068-d6338e5c19b6 DEVICE=bond0 ONBOOT=yes [root@mob-r1-d-ovirt-aa-1-01 network-scripts]# cat ifcfg-bond0.1131 DEVICE=bond0.1131 VLAN=yes ONBOOT=yes MTU=1500 IPADDR=172.18.206.184 NETMASK=255.255.254.0 GATEWAY=172.18.206.1 BOOTPROTO=none MTU=1500 DEFROUTE=yes NM_CONTROLLED=no IPV6INIT=yes DNS1=172.20.150.10 DNS2=172.20.150.11 I get the following error: [ INFO ] TASK [ovirt.hosted_engine_setup : Generate output list] [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Validate selected bridge interface if management bridge does not exists] [ INFO ] skipping: [localhost] Please indicate a nic to set ovirtmgmt bridge on: (bond0, bond0.1131) [bond0.1131]: Please specify which way the network connectivity should be checked (ping, dns, tcp, none) [dns]: .. .. .. .. .. [ INFO ] ok: [localhost] [ INFO ] TASK [ovirt.hosted_engine_setup : Validate selected bridge interface if management bridge does not exists] [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "The selected network interface is not valid"} [ ERROR ] Failed to execute stage 'Closing up': Failed executing ansible-playbook [ INFO ] Stage: Clean up And if I create the ifcfg-ovirtmgmt as a bridge it fails later. What is the correct network setup for my bond configuration to do a self hosted-engine setup ? Regards Nar
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/NRU4CI6GLGL77RDY7O7EGZOD67TJOPJ3/