On Thu, Feb 15, 2018 at 3:54 PM, Kuko Armas <[email protected]> wrote:
> > I'm not sure if I should submit a bug report about this, so I ask around > here first... > I've found a bug that "seems" related but I think it's not ( > https://bugzilla.redhat.com/show_bug.cgi?id=1523661) > > This is the problem: > > - I'm trying to do a clean HE deploy with oVirt 4.2.1 on a clean CentOS > 7.4 host > - I have a LACP bond (bond0) and I need my management network to be on > vlan 1005, si I have created interface bond0.1005 on the host and > everything works > - I run hosted-engine deploy, and it always fails with > > [ ERROR ] fatal: [localhost]: FAILED! => {"attempts": 50, "changed": true, > "cmd": "ip rule list | grep ovirtmgmt | sed s/\\\\[.*\\\\]\\ //g | awk '{ > print $9 }'", "delta": "0:00:00.006473", "end": "2018-02-15 > 13:57:11.132359", "rc": 0, "start": "2018-02-15 13:57:11.125886", "stderr": > "", "stderr_lines": [], "stdout": "", "stdout_lines": []} > [ ERROR ] Failed to execute stage 'Closing up': Failed executing > ansible-playbook > > - Looking at the ansible playbook, I see it's trying to look for an ip > rule using a custom routing table, but I have no such rule > > [root@ovirt1 ~]# ip rule > 0: from all lookup local > 32766: from all lookup main > 32767: from all lookup default > > - I also find that I have no "ovirtmgmt" bridge > > bridge name bridge id STP enabled interfaces > ;vdsmdummy; 8000.000000000000 no > virbr0 8000.525400e6ca97 yes virbr0-nic > vnet0 > > - But I haven't found any reference in the ansible playbook to this > network creation. > > - The HE VM gets created and I can connect with SSH, so I tried to find > out if the ovirtmgmt network is created via vdsm from the engine > - Looking at the engine.log I found this: > > 2018-02-15 13:49:26,850Z INFO > [org.ovirt.engine.core.bll.host.HostConnectivityChecker] > (EE-ManagedThreadFactory-engine-Thread-1) [15c7e33a] Engine managed to > communicate wi > th VDSM agent on host 'ovirt1' with address 'ovirt1' > ('06651b32-4ef8-4b5d-ab2d-c38e84c2d790') > 2018-02-15 13:49:30,302Z ERROR [org.ovirt.engine.core.dal. > dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedThreadFactory-engine-Thread-1) > [15c7e33a] EVENT_ID: VLAN_ID_ > MISMATCH_FOR_MANAGEMENT_NETWORK_CONFIGURATION(1,119), Failed to configure > management network on host ovirt1. Host ovirt1 has an interface bond0.1005 > for the management netwo > rk configuration with VLAN-ID (1005), which is different from data-center > definition (none). > 2018-02-15 13:49:30,302Z ERROR > [org.ovirt.engine.core.bll.hostdeploy.InstallVdsInternalCommand] > (EE-ManagedThreadFactory-engine-Thread-1) [15c7e33a] Exception: > org.ovirt.eng > ine.core.bll.network.NetworkConfigurator$NetworkConfiguratorException: > Failed to configure management network > > - So I guess that the engine tried to create the ovirtmgmt bridge on the > host via vdsm, but it failed because "Host ovirt1 has an interface > bond0.1005 for the management netwo > rk configuration with VLAN-ID (1005), which is different from data-center > definition (none)" > - Of course I haven't had the opportunity to setup the management > network's vlan in the datacenter yet, because I'm still trying to deploy > the Hosted Engine > > Is this a supported configuration? Is there a way I can tell the > datacenter that the management network is on vlan 1005? Should I file a bug > report? > Yes, please. > Is there a workaround? > You can pass --noansible and fallback to the previous flow, sorry. > > Salu2! > -- > Miguel Armas > CanaryTek Consultoria y Sistemas SL > http://www.canarytek.com/ > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

