Hi Dominik, I have added LACP bond network to all hosts and renamed the Hosted Engine using "/usr/share/ovirt-engine/setup/bin/ovirt-engine-rename."
However, I am still missing the option to assign Migration and Management network roles to this new bond. Can you advise where I can find this option? Thanks, Douglas Duckworth, MSc, LFCS HPC System Administrator Scientific Computing Unit<https://scu.med.cornell.edu> Weill Cornell Medicine 1300 York Avenue New York, NY 10065 E: [email protected]<mailto:[email protected]> O: 212-746-6305 F: 212-746-8690 On Fri, Aug 24, 2018 at 11:52 AM Dominik Holler <[email protected]<mailto:[email protected]>> wrote: On Fri, 24 Aug 2018 09:46:25 -0400 Douglas Duckworth <[email protected]<mailto:[email protected]>> wrote: > Sorry, I mean "migration network" for moving live migration traffic. > You have to create a new logical network in "Network > Networks > New" and assign this to ib0 in "Compute > Hosts > hostname > Network Interfaces > Setup Host Networks". After this you can assign a role to this network in "Compute > Clusters > Clustername > Logical Networks > Manage Networks" > FDR infiniband much faster than 1Gb network which currently acts as > migration network, vm network, display network, mgmt network, etc. > > Thanks, > > Douglas Duckworth, MSc, LFCS > HPC System Administrator > Scientific Computing Unit > Weill Cornell Medicine > 1300 York - LC-502 > E: [email protected]<mailto:[email protected]> > O: 212-746-6305 > F: 212-746-8690 > > > On Fri, Aug 24, 2018 at 9:36 AM, Dominik Holler > <[email protected]<mailto:[email protected]>> > wrote: > > > On Thu, 23 Aug 2018 13:51:39 -0400 > > Douglas Duckworth <[email protected]<mailto:[email protected]>> > > wrote: > > > > > THANKS! > > > > > > ib0 now up with NFS storage back on this hypervisor > > > > > > > Thanks for letting us know. > > > > > Though how do I make it a transfer network? I don't see an > > > option. > > > > I do not understand the meaning of "transfer network". > > The network interface to use for NFS results from the routing > > tables of the host. > > In "Compute > Clusters > Clustername > Logical Networks > Manage > > Networks" network roles for some kind of loads can be assigned, but > > not for NFS access. > > > > > > > Thanks, > > > > > > Douglas Duckworth, MSc, LFCS > > > HPC System Administrator > > > Scientific Computing Unit > > > Weill Cornell Medicine > > > 1300 York - LC-502 > > > E: [email protected]<mailto:[email protected]> > > > O: 212-746-6305 > > > F: 212-746-8690 > > > > > > > > > On Thu, Aug 23, 2018 at 11:12 AM, Douglas Duckworth > > > <[email protected]<mailto:[email protected]> > > > > wrote: > > > > > > > Hi Dominik > > > > > > > > Yes, the network-script was created by our Ansible role that > > > > deploys CentOS hosts. It pulls the IP from DNS then templates > > > > the script and copies to host. > > > > > > > > I will try this oVirt step then see if it works! > > > > > > > > Thanks, > > > > > > > > Douglas Duckworth, MSc, LFCS > > > > HPC System Administrator > > > > Scientific Computing Unit > > > > Weill Cornell Medicine > > > > 1300 York - LC-502 > > > > E: [email protected]<mailto:[email protected]> > > > > O: 212-746-6305 > > > > F: 212-746-8690 > > > > > > > > > > > > On Thu, Aug 23, 2018 at 11:09 AM, Dominik Holler > > > > <[email protected]<mailto:[email protected]>> wrote: > > > > > > > >> Is ifcfg-ib0 created before adding the host? > > > >> Can ib0 be reconfigured using engine, e.g. by > > > >> "Compute > Hosts > hostx > Network Interfaces > Setup Host > > > >> Networks"? If this some kind of self-hosted engine? > > > >> > > > >> On Thu, 23 Aug 2018 09:30:59 -0400 > > > >> Douglas Duckworth > > > >> <[email protected]<mailto:[email protected]>> wrote: > > > >> > > > >> > Here's a link to the files: > > > >> > > > > >> > https://urldefense.proofpoint.com/v2/url?u=https-3A__bit.ly_ > > > >> 2wjZ6Vo&d=DwICAg&c=lb62iw4YL4RFalcE2hQUQealT9-RXrryqt9KZX2qu > > > >> 2s&r=2Fzhh_78OGspKQpl_e-CbhH6xUjnRkaqPFUS2wTJ2cw&m=Y25- > > > >> OOvgu58jlC82-fzBeNIpQ7ZscoHznffUhqE6EBM&s=QQXlC9Tisa60TvimyS > > > >> 3BnFDCaDF7VPD8eCzT-Fke-p0&e= > > > >> > > > > >> > Thank you! > > > >> > > > > >> > Thanks, > > > >> > > > > >> > Douglas Duckworth, MSc, LFCS > > > >> > HPC System Administrator > > > >> > Scientific Computing Unit > > > >> > Weill Cornell Medicine > > > >> > 1300 York - LC-502 > > > >> > E: [email protected]<mailto:[email protected]> > > > >> > O: 212-746-6305 > > > >> > F: 212-746-8690 > > > >> > > > > >> > > > > >> > On Thu, Aug 23, 2018 at 6:51 AM, Dominik Holler > > > >> > <[email protected]<mailto:[email protected]>> wrote: > > > >> > > > > >> > > Would you please share the vdsm.log and the supervdsm.log > > > >> > > from this host? > > > >> > > > > > >> > > On Wed, 22 Aug 2018 11:36:09 -0400 > > > >> > > Douglas Duckworth > > > >> > > <[email protected]<mailto:[email protected]>> wrote: > > > >> > > > > > >> > > > Hi > > > >> > > > > > > >> > > > I keep losing ib0 connection on hypervisor after adding > > > >> > > > host to engine. This makes the host not really work > > > >> > > > since NFS will be mounted over ib0. > > > >> > > > > > > >> > > > I don't really understand why this occurs. > > > >> > > > > > > >> > > > OS: > > > >> > > > > > > >> > > > [root@ovirt-hv2 ~]# cat /etc/redhat-release > > > >> > > > CentOS Linux release 7.5.1804 (Core) > > > >> > > > > > > >> > > > Here's the network script: > > > >> > > > > > > >> > > > [root@ovirt-hv2 ~]# > > > >> > > > cat /etc/sysconfig/network-scripts/ifcfg-ib0 DEVICE=ib0 > > > >> > > > BOOTPROTO=static > > > >> > > > IPADDR=172.16.0.207 > > > >> > > > NETMASK=255.255.255.0 > > > >> > > > ONBOOT=yes > > > >> > > > ZONE=public > > > >> > > > > > > >> > > > When I try "ifup" > > > >> > > > > > > >> > > > [root@ovirt-hv2 ~]# ifup ib0 > > > >> > > > Error: Connection activation failed: No suitable device > > > >> > > > found for this connection. > > > >> > > > > > > >> > > > The error in syslog: > > > >> > > > > > > >> > > > Aug 22 11:31:50 ovirt-hv2 kernel: IPv4: martian source > > > >> > > > 172.16.0.87 from 172.16.0.49, on dev ib0 > > > >> > > > Aug 22 11:31:53 ovirt-hv2 NetworkManager[1070]: <info> > > > >> > > > [1534951913.7486] audit: op="connection-activate" > > > >> > > > uuid="2ab4abde-b8a5-6cbc-19b1-2bfb193e4e89" name="System > > > >> > > > ib0" result="fail" reason="No suitable device found for > > > >> > > > this connection. > > > >> > > > > > > >> > > > As you can see media state up: > > > >> > > > > > > >> > > > [root@ovirt-hv2 ~]# ip a > > > >> > > > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue > > > >> > > > state UNKNOWN group default qlen 1000 > > > >> > > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > > >> > > > inet 127.0.0.1/8<http://127.0.0.1/8> scope host lo > > > >> > > > valid_lft forever preferred_lft forever > > > >> > > > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > > > >> > > > mq master ovirtmgmt state UP group default qlen 1000 > > > >> > > > link/ether 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff > > > >> > > > 3: em2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > >> > > > qdisc mq state DOWN group default qlen 1000 > > > >> > > > link/ether 50:9a:4c:89:d3:82 brd ff:ff:ff:ff:ff:ff > > > >> > > > 4: p1p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > >> > > > qdisc mq state DOWN group default qlen 1000 > > > >> > > > link/ether b4:96:91:13:ea:68 brd ff:ff:ff:ff:ff:ff > > > >> > > > 5: p1p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 > > > >> > > > qdisc mq state DOWN group default qlen 1000 > > > >> > > > link/ether b4:96:91:13:ea:6a brd ff:ff:ff:ff:ff:ff > > > >> > > > 6: idrac: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 > > > >> > > > qdisc pfifo_fast state UNKNOWN group default qlen 1000 > > > >> > > > link/ether 50:9a:4c:89:d3:84 brd ff:ff:ff:ff:ff:ff > > > >> > > > inet 169.254.0.2/16<http://169.254.0.2/16> brd > > > >> > > > 169.254.255.255 scope global > > > >> > > > idrac valid_lft forever preferred_lft forever > > > >> > > > 7: ib0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 2044 qdisc > > > >> > > > mq state UP group default qlen 256 > > > >> > > > link/infiniband > > > >> > > > a0:00:02:08:fe:80:00:00:00:00:00:00:ec:0d:9a:03:00:1d:13:41 > > > >> > > > brd > > > >> > > > 00:ff:ff:ff:ff:12:40:1b:ff:ff:00:00:00:00:00:00:ff:ff:ff:ff > > > >> > > > 8: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop > > > >> > > > state DOWN group default qlen 1000 link/ether > > > >> > > > 12:b4:30:22:39:5b brd ff:ff:ff:ff:ff:ff 9: br-int: > > > >> > > > <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > > > >> > > > group default qlen 1000 link/ether 3e:32:e6:66:98:49 brd > > > >> > > > ff:ff:ff:ff:ff:ff 25: ovirtmgmt: > > > >> > > > <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue > > > >> > > > state UP group default qlen 1000 link/ether > > > >> > > > 50:9a:4c:89:d3:81 brd ff:ff:ff:ff:ff:ff inet > > > >> > > > 10.0.0.183/16<http://10.0.0.183/16> brd 10.0.255.255 scope > > > >> > > > global ovirtmgmt > > > >> > > > valid_lft forever preferred_lft forever 26: > > > >> > > > genev_sys_6081: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu > > > >> > > > 65000 qdisc noqueue master ovs-system state UNKNOWN > > > >> > > > group default qlen 1000 link/ether aa:32:82:1b:01:d9 brd > > > >> > > > ff:ff:ff:ff:ff:ff 27: ;vdsmdummy;: <BROADCAST,MULTICAST> > > > >> > > > mtu 1500 qdisc noop state DOWN group default qlen 1000 > > > >> > > > link/ether 32:ff:5d:b8:c2:b4 brd ff:ff:ff:ff:ff:ff > > > >> > > > > > > >> > > > The card is FDR: > > > >> > > > > > > >> > > > [root@ovirt-hv2 ~]# lspci -v | grep Mellanox > > > >> > > > 01:00.0 Network controller: Mellanox Technologies MT27500 > > > >> > > > Family [ConnectX-3] > > > >> > > > Subsystem: Mellanox Technologies Device 0051 > > > >> > > > > > > >> > > > Latest OFED driver: > > > >> > > > > > > >> > > > [root@ovirt-hv2 ~]# /etc/init.d/openibd status > > > >> > > > > > > >> > > > HCA driver loaded > > > >> > > > > > > >> > > > Configured IPoIB devices: > > > >> > > > ib0 > > > >> > > > > > > >> > > > Currently active IPoIB devices: > > > >> > > > ib0 > > > >> > > > Configured Mellanox EN devices: > > > >> > > > > > > >> > > > Currently active Mellanox devices: > > > >> > > > ib0 > > > >> > > > > > > >> > > > The following OFED modules are loaded: > > > >> > > > > > > >> > > > rdma_ucm > > > >> > > > rdma_cm > > > >> > > > ib_ipoib > > > >> > > > mlx4_core > > > >> > > > mlx4_ib > > > >> > > > mlx4_en > > > >> > > > mlx5_core > > > >> > > > mlx5_ib > > > >> > > > ib_uverbs > > > >> > > > ib_umad > > > >> > > > ib_ucm > > > >> > > > ib_cm > > > >> > > > ib_core > > > >> > > > mlxfw > > > >> > > > mlx5_fpga_tools > > > >> > > > > > > >> > > > I can add an IP to ib0 using "ip addr" though I need > > > >> > > > Network Manager to work with ib0. > > > >> > > > > > > >> > > > > > > >> > > > Thanks, > > > >> > > > > > > >> > > > Douglas Duckworth, MSc, LFCS > > > >> > > > HPC System Administrator > > > >> > > > Scientific Computing Unit > > > >> > > > Weill Cornell Medicine > > > >> > > > 1300 York - LC-502 > > > >> > > > E: [email protected]<mailto:[email protected]> > > > >> > > > O: 212-746-6305 > > > >> > > > F: 212-746-8690 > > > >> > > > > > >> > > > > > >> > > > >> > > > > > > > >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/GJTWCJE5DC3RXRI3SNH2NJ2AZ2DDWO4L/

