Edi, danken, this is again the malfunctioning heuristics in the client for handling ipv6. Is there a bug on this issue ?
On Jul 27, 2016 8:57 AM, <aleksey.maksi...@it-kb.ru> wrote: > I enabled ipv6 for "lo" and "ovirtmgmt" interfaces and deployment process > overt completed successfully. > > # cat /etc/sysctl.conf > > net.ipv6.conf.all.disable_ipv6 = 1 > net.ipv6.conf.default.disable_ipv6 = 1 > net.ipv6.conf.lo.disable_ipv6 = 0 > net.ipv6.conf.ovirtmgmt.disable_ipv6 = 0 > > Thank you for your Help! > > 26.07.2016, 09:13, "Edward Haas" <eh...@redhat.com>: > > > > On Mon, Jul 25, 2016 at 4:02 PM, Simone Tiraboschi <stira...@redhat.com> > wrote: > > On Mon, Jul 25, 2016 at 2:38 PM, <aleksey.maksi...@it-kb.ru> wrote: > > telnet kom-ad01-vm31.holding.com 54321 = success connection > > > > telnet ::1 54321 > > Trying ::1... > > telnet: connect to address ::1: Network is unreachable > > > > (ipv6 on my server disabled) > > Ok, so the issue seams here: now by default vdsm binds on :: and its > heuristc can end up using ipv6. > See this one: https://bugzilla.redhat.com/show_bug.cgi?id=1350883 > > Can you please try enabling ipv6 on your host or setting > management_ip = 0.0.0.0 > under the [address] section in /etc/vdsm/vdsm.conf > and then restarting vdsm. > > > > Could you please also add the 'ip addr' output? Just interested to see how > IPv6 was > disabled on the host. > It will be even better if you could apply the patch ( > https://gerrit.ovirt.org/#/c/60020) and check. > > > > > > > 25.07.2016, 15:35, "Simone Tiraboschi" <stira...@redhat.com>: > >> On Mon, Jul 25, 2016 at 2:15 PM, <aleksey.maksi...@it-kb.ru> wrote: > >>> # ss -plutn > >>> > >>> Netid State Recv-Q Send-Q Local Address:Port Peer Address:Port > >>> > >>> udp UNCONN 0 0 *:111 *:* users:(("rpcbind",pid=827,fd=6)) > >>> udp UNCONN 0 0 *:161 *:* users:(("snmpd",pid=1609,fd=6)) > >>> udp UNCONN 0 0 127.0.0.1:323 *:* users:(("chronyd",pid=795,fd=1)) > >>> udp UNCONN 0 0 *:959 *:* users:(("rpcbind",pid=827,fd=7)) > >>> udp UNCONN 0 0 127.0.0.1:25375 *:* users:(("snmpd",pid=1609,fd=8)) > >>> udp UNCONN 0 0 127.0.0.1:25376 *:* users:(("cmapeerd",pid=2056,fd=5)) > >>> udp UNCONN 0 0 127.0.0.1:25393 *:* users:(("cmanicd",pid=2278,fd=3)) > >>> udp UNCONN 0 0 :::111 :::* users:(("rpcbind",pid=827,fd=9)) > >>> udp UNCONN 0 0 :::959 :::* users:(("rpcbind",pid=827,fd=10)) > >>> tcp LISTEN 0 128 *:2381 *:* > users:(("hpsmhd",pid=3903,fd=4),("hpsmhd",pid=3901,fd=4),("hpsmhd",pid=3900,fd=4),("hpsmhd",pid=3899,fd=4),("hpsmhd",pid=3898,fd=4),("hpsmhd",pid=3893,fd=4)) > >>> tcp LISTEN 0 128 *:111 *:* users:(("rpcbind",pid=827,fd=8)) > >>> tcp LISTEN 0 5 *:54322 *:* users:(("ovirt-imageio-d",pid=753,fd=3)) > >>> tcp LISTEN 0 128 *:22 *:* users:(("sshd",pid=1606,fd=3)) > >>> tcp LISTEN 0 100 127.0.0.1:25 *:* users:(("master",pid=1948,fd=13)) > >>> tcp LISTEN 0 128 *:2301 *:* > users:(("hpsmhd",pid=3903,fd=3),("hpsmhd",pid=3901,fd=3),("hpsmhd",pid=3900,fd=3),("hpsmhd",pid=3899,fd=3),("hpsmhd",pid=3898,fd=3),("hpsmhd",pid=3893,fd=3)) > >>> tcp LISTEN 0 30 *:16514 *:* users:(("libvirtd",pid=10688,fd=13)) > >>> tcp LISTEN 0 128 127.0.0.1:199 *:* users:(("snmpd",pid=1609,fd=9)) > >>> tcp LISTEN 0 128 :::111 :::* users:(("rpcbind",pid=827,fd=11)) > >>> tcp LISTEN 0 5 :::54321 :::* users:(("vdsm",pid=11077,fd=23)) > >> > >> vdsm is properly bind over ipv6. > >> > >> Can you please check if you can connect to vdsm with: > >> telnet kom-ad01-vm31.holding.com 54321 > >> and with > >> telnet ::1 54321 > >> ? > >> > >>> tcp LISTEN 0 30 :::16514 :::* users:(("libvirtd",pid=10688,fd=14)) > >>> > >>> 25.07.2016, 15:11, "Simone Tiraboschi" <stira...@redhat.com>: > >>>> On Mon, Jul 25, 2016 at 2:03 PM, <aleksey.maksi...@it-kb.ru> wrote: > >>>>> Yes. > >>>>> > >>>>> # ping $(python -c 'import socket; print(socket.gethostname())') > >>>>> > >>>>> PING KOM-AD01-VM31.holding.com (10.1.0.231) 56(84) bytes of data. > >>>>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=1 > ttl=64 time=0.030 ms > >>>>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=2 > ttl=64 time=0.015 ms > >>>>> 64 bytes from kom-ad01-vm31.holding.com (10.1.0.231): icmp_seq=3 > ttl=64 time=0.011 ms > >>>>> ^C > >>>>> --- KOM-AD01-VM31.holding.com ping statistics --- > >>>>> 3 packets transmitted, 3 received, 0% packet loss, time 2001ms > >>>>> rtt min/avg/max/mdev = 0.011/0.018/0.030/0.009 ms > >>>>> > >>>>> but... > >>>> > >>>> and the output of > >>>> ss -plutn > >>>> > >>>>> # vdsClient -s 0 getVdsCaps > >>>>> > >>>>> Traceback (most recent call last): > >>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in <module> > >>>>> code, message = commands[command][0](commandArgs) > >>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in do_getCap > >>>>> return self.ExecAndExit(self.s.getVdsCapabilities()) > >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1233, in __call__ > >>>>> return self.__send(self.__name, args) > >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1587, in __request > >>>>> verbose=self.__verbose > >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1273, in request > >>>>> return self.single_request(host, handler, request_body, > verbose) > >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1301, in > single_request > >>>>> self.send_content(h, request_body) > >>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line 1448, in > send_content > >>>>> connection.endheaders(request_body) > >>>>> File "/usr/lib64/python2.7/httplib.py", line 975, in endheaders > >>>>> self._send_output(message_body) > >>>>> File "/usr/lib64/python2.7/httplib.py", line 835, in _send_output > >>>>> self.send(msg) > >>>>> File "/usr/lib64/python2.7/httplib.py", line 797, in send > >>>>> self.connect() > >>>>> File "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line > 203, in connect > >>>>> sock = socket.create_connection((self.host, self.port), > self.timeout) > >>>>> File "/usr/lib64/python2.7/socket.py", line 571, in > create_connection > >>>>> raise err > >>>>> error: [Errno 101] Network is unreachable > >>>>> > >>>>> 25.07.2016, 14:58, "aleksey.maksi...@it-kb.ru" < > aleksey.maksi...@it-kb.ru>: > >>>>>> Ok. > >>>>>> > >>>>>> 1) I stopped and disabled the service NetworkManager > >>>>>> # systemctl stop NetworkManager > >>>>>> # systemctl disable NetworkManager > >>>>>> > >>>>>> 2) I filled /etc/resolv.conf and remove DNS1,DNS2 and add > PEERDNS=no in ifcfg-* file. > >>>>>> > >>>>>> 3) Reboot server > >>>>>> > >>>>>> 4) Try deploy oVirt HE 4 and I get the same error > >>>>>> > >>>>>> [ INFO ] Creating Volume Group > >>>>>> [ INFO ] Creating Storage Domain > >>>>>> [ INFO ] Creating Storage Pool > >>>>>> [ INFO ] Connecting Storage Pool > >>>>>> [ INFO ] Verifying sanlock lockspace initialization > >>>>>> [ ERROR ] Failed to execute stage 'Misc configuration': [Errno > 101] Network is unreachable > >>>>>> [ INFO ] Stage: Clean up > >>>>>> [ INFO ] Generating answer file > '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160725143420.conf' > >>>>>> [ INFO ] Stage: Pre-termination > >>>>>> [ INFO ] Stage: Termination > >>>>>> [ ERROR ] Hosted Engine deployment failed: this system is not > reliable, please check the issue, fix and redeploy > >>>>>> Log file is located at > /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup- > 20160725142534-t81kwf.log > >>>>>> > >>>>>> What ideas further? > >>>>>> > >>>>>> 25.07.2016, 13:06, "Simone Tiraboschi" <stira...@redhat.com>: > >>>>>>> On Mon, Jul 25, 2016 at 11:54 AM, <aleksey.maksi...@it-kb.ru> > wrote: > >>>>>>>> What am I supposed to do for successfully deploy ovirt 4 ? > >>>>>>>> Any ideas ? > >>>>>>> > >>>>>>> Can you please try to explicitly configure your DNS with > nameserver > >>>>>>> under /etc/resolv.conf and remove DNS1 and DNS2 and set > PEERDNS=no for > >>>>>>> the interface you are going to use? > >>>>>>> > >>>>>>>> 25.07.2016, 12:47, "aleksey.maksi...@it-kb.ru" < > aleksey.maksi...@it-kb.ru>: > >>>>>>>>> "Aleksey, was your DNS configured with DNS1 and DNS2 just on > the interface you used to create the management bridge on?" > >>>>>>>>> > >>>>>>>>> Yes. Of course > >>>>>>>>> > >>>>>>>>> 25.07.2016, 12:27, "Simone Tiraboschi" <stira...@redhat.com > >: > >>>>>>>>>> On Mon, Jul 25, 2016 at 10:56 AM, Piotr Kliczewski > >>>>>>>>>> <piotr.kliczew...@gmail.com> wrote: > >>>>>>>>>>> This could be the issue here as well as for BZ #1358530 > >>>>>>>>>>> > >>>>>>>>>>> On Mon, Jul 25, 2016 at 10:53 AM, < > aleksey.maksi...@it-kb.ru> wrote: > >>>>>>>>>>>> Could this be due to the fact that the ovirt installer > has changed network configuration files (ifcfg-*, resolv.conf) ? > >>>>>>>>>>>> After the error in ovirt installation process I see > from resolv.conf disappeared on my DNS servers entry and now the server is > unable to resolve names. > >>>>>>>>>> > >>>>>>>>>> So it could be related to > https://bugzilla.redhat.com/show_bug.cgi?id=1160423 > >>>>>>>>>> > >>>>>>>>>> Aleksey, was your DNS configured with DNS1 and DNS2 just > on the > >>>>>>>>>> interface you used to create the management bridge on? > >>>>>>>>>> Can you please try the workaround described here > >>>>>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=1160423#c25 ? > >>>>>>>>>> > >>>>>>>>>>>> 25.07.2016, 11:26, "Simone Tiraboschi" < > stira...@redhat.com>: > >>>>>>>>>>>>> On Mon, Jul 25, 2016 at 10:22 AM, < > aleksey.maksi...@it-kb.ru> wrote: > >>>>>>>>>>>>>> # vdsClient -s 0 getVdsCaps > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Traceback (most recent call last): > >>>>>>>>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 2980, in > <module> > >>>>>>>>>>>>>> code, message = commands[command][0](commandArgs) > >>>>>>>>>>>>>> File "/usr/share/vdsm/vdsClient.py", line 543, in > do_getCap > >>>>>>>>>>>>>> return > self.ExecAndExit(self.s.getVdsCapabilities()) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line > 1233, in __call__ > >>>>>>>>>>>>>> return self.__send(self.__name, args) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line > 1587, in __request > >>>>>>>>>>>>>> verbose=self.__verbose > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line > 1273, in request > >>>>>>>>>>>>>> return self.single_request(host, handler, > request_body, verbose) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line > 1301, in single_request > >>>>>>>>>>>>>> self.send_content(h, request_body) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/xmlrpclib.py", line > 1448, in send_content > >>>>>>>>>>>>>> connection.endheaders(request_body) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 975, > in endheaders > >>>>>>>>>>>>>> self._send_output(message_body) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 835, > in _send_output > >>>>>>>>>>>>>> self.send(msg) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/httplib.py", line 797, > in send > >>>>>>>>>>>>>> self.connect() > >>>>>>>>>>>>>> File > "/usr/lib/python2.7/site-packages/vdsm/m2cutils.py", line 203, in connect > >>>>>>>>>>>>>> sock = socket.create_connection((self.host, > self.port), self.timeout) > >>>>>>>>>>>>>> File "/usr/lib64/python2.7/socket.py", line 571, > in create_connection > >>>>>>>>>>>>>> raise err > >>>>>>>>>>>>>> error: [Errno 101] Network is unreachable > >>>>>>>>>>>>> > >>>>>>>>>>>>> Yaniv, can you please take also a look to this one? > >>>>>>>>>>>>> it's exactly the opposite of > https://bugzilla.redhat.com/1358530 > >>>>>>>>>>>>> Here the jsonrpcclient works but not the xmlrpc one. > >>>>>>>>>>>>> > >>>>>>>>>>>>>> 25.07.2016, 11:17, "Simone Tiraboschi" < > stira...@redhat.com>: > >>>>>>>>>>>>>>> On Mon, Jul 25, 2016 at 7:51 AM, < > aleksey.maksi...@it-kb.ru> wrote: > >>>>>>>>>>>>>>>> Simone, there is something interesting in the > vdsm.log? > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> For what I saw the issue is not related to the > storage but to the network. > >>>>>>>>>>>>>>> ovirt-hosted-engine-setup uses the jsonrpc client, > instead the code > >>>>>>>>>>>>>>> from ovirt-hosted-engine-ha still uses the xmlrpc > client somewhere and > >>>>>>>>>>>>>>> this happens also when the setup asks to create the > lockspace volume. > >>>>>>>>>>>>>>> It seams that in your case the xmlrpc client could > not connect vdsm on > >>>>>>>>>>>>>>> the localhost. > >>>>>>>>>>>>>>> It could be somehow related to: > >>>>>>>>>>>>>>> https://bugzilla.redhat.com/1358530 > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> Can you please try executing > >>>>>>>>>>>>>>> sudo vdsClient -s 0 getVdsCaps > >>>>>>>>>>>>>>> on that host? > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> 22.07.2016, 19:36, "aleksey.maksi...@it-kb.ru" < > aleksey.maksi...@it-kb.ru>: > >>>>>>>>>>>>>>>>> Simone, thanks for link. > >>>>>>>>>>>>>>>>> vdsm.log attached > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> 22.07.2016, 19:28, "Simone Tiraboschi" < > stira...@redhat.com>: > >>>>>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 5:59 PM, < > aleksey.maksi...@it-kb.ru> wrote: > >>>>>>>>>>>>>>>>>>> Thank you for your response, Simone. > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> Log attached. > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> It seams it comes from VDSM, can you please > attach also vdsm.log? > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> I don't use ovirt-engine-appliance because I > have not found "how-to" for ovirt-engine-appliance deployment in hosted > engine configuration. > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> yum install ovirt-engine-appliance > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Then follow the instruction here: > >>>>>>>>>>>>>>>>>> > http://www.ovirt.org/develop/release-management/features/heapplianceflow/ > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> 22.07.2016, 17:09, "Simone Tiraboschi" < > stira...@redhat.com>: > >>>>>>>>>>>>>>>>>>>> Hi Aleksey, > >>>>>>>>>>>>>>>>>>>> Can you please attach hosted-engine-setup > logs? > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> On Fri, Jul 22, 2016 at 3:46 PM, < > aleksey.maksi...@it-kb.ru> wrote: > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Hello oVirt guru`s ! > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> I have problem with initial deploy of > ovirt 4.0 hosted engine. > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> My environment : > >>>>>>>>>>>>>>>>>>>>> ============================ > >>>>>>>>>>>>>>>>>>>>> * Two servers HP ProLiant DL 360 G5 with > Qlogic FC HBA connected (with > >>>>>>>>>>>>>>>>>>>>> multipathd) to storage HP 3PAR 7200 > >>>>>>>>>>>>>>>>>>>>> * On each server installed CentOS 7.2 > Linux (3.10.0-327.22.2.el7.x86_64) > >>>>>>>>>>>>>>>>>>>>> * On 3PAR storage I created 2 LUNs for > oVirt. > >>>>>>>>>>>>>>>>>>>>> - First LUN for oVirt Hosted Engine VM > (60GB) > >>>>>>>>>>>>>>>>>>>>> - Second LUN for all other VMs (2TB) > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> # multipath -ll > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> 3par-vv1 (360002ac0000000000000001b0000cec9) > dm-0 3PARdata,VV > >>>>>>>>>>>>>>>>>>>>> size=60G features='1 queue_if_no_path' > hwhandler='1 alua' wp=rw > >>>>>>>>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 > status=active > >>>>>>>>>>>>>>>>>>>>> |- 2:0:1:1 sdd 8:48 active ready running > >>>>>>>>>>>>>>>>>>>>> |- 3:0:0:1 sdf 8:80 active ready running > >>>>>>>>>>>>>>>>>>>>> |- 2:0:0:1 sdb 8:16 active ready running > >>>>>>>>>>>>>>>>>>>>> `- 3:0:1:1 sdh 8:112 active ready > running > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> 3par-vv2 (360002ac000000000000000160000cec9) > dm-1 3PARdata,VV > >>>>>>>>>>>>>>>>>>>>> size=2.0T features='1 queue_if_no_path' > hwhandler='1 alua' wp=rw > >>>>>>>>>>>>>>>>>>>>> `-+- policy='round-robin 0' prio=50 > status=active > >>>>>>>>>>>>>>>>>>>>> |- 2:0:0:0 sda 8:0 active ready running > >>>>>>>>>>>>>>>>>>>>> |- 3:0:0:0 sde 8:64 active ready running > >>>>>>>>>>>>>>>>>>>>> |- 2:0:1:0 sdc 8:32 active ready running > >>>>>>>>>>>>>>>>>>>>> `- 3:0:1:0 sdg 8:96 active ready running > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> My steps on first server (initial deploy > of ovirt 4.0 hosted engine): > >>>>>>>>>>>>>>>>>>>>> ============================ > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> # systemctl stop NetworkManager > >>>>>>>>>>>>>>>>>>>>> # systemctl disable NetworkManager > >>>>>>>>>>>>>>>>>>>>> # yum -y install > http://resources.ovirt.org/pub/yum-repo/ovirt-release40.rpm > >>>>>>>>>>>>>>>>>>>>> # yum -y install epel-release > >>>>>>>>>>>>>>>>>>>>> # wget > >>>>>>>>>>>>>>>>>>>>> > http://mirror.yandex.ru/centos/7/isos/x86_64/CentOS-7-x86_64-NetInstall-1511.iso > >>>>>>>>>>>>>>>>>>>>> -P /tmp/ > >>>>>>>>>>>>>>>>>>>>> # yum install ovirt-hosted-engine-setup > >>>>>>>>>>>>>>>>>>>>> # yum install screen > >>>>>>>>>>>>>>>>>>>>> # screen -RD > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> ...in screen session : > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> # hosted-engine --deploy > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> ... > >>>>>>>>>>>>>>>>>>>>> in configuration process I chose "fc" as > storage type for oVirt hosted > >>>>>>>>>>>>>>>>>>>>> engine vm and select 60GB LUN... > >>>>>>>>>>>>>>>>>>>>> ... > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> --== CONFIGURATION PREVIEW ==-- > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> ... > >>>>>>>>>>>>>>>>>>>>> Firewall manager : iptables > >>>>>>>>>>>>>>>>>>>>> Gateway address : 10.1.0.1 > >>>>>>>>>>>>>>>>>>>>> Host name for web application : > KOM-AD01-OVIRT1 > >>>>>>>>>>>>>>>>>>>>> Storage Domain type : fc > >>>>>>>>>>>>>>>>>>>>> Host ID : 1 > >>>>>>>>>>>>>>>>>>>>> LUN ID : > >>>>>>>>>>>>>>>>>>>>> 360002ac0000000000000001b0000cec9 > >>>>>>>>>>>>>>>>>>>>> Image size GB : 40 > >>>>>>>>>>>>>>>>>>>>> Console type : vnc > >>>>>>>>>>>>>>>>>>>>> Memory size MB : 4096 > >>>>>>>>>>>>>>>>>>>>> MAC address : 00:16:3e:77:1d:07 > >>>>>>>>>>>>>>>>>>>>> Boot type : cdrom > >>>>>>>>>>>>>>>>>>>>> Number of CPUs : 2 > >>>>>>>>>>>>>>>>>>>>> ISO image (cdrom > boot/cloud-init) : > >>>>>>>>>>>>>>>>>>>>> /tmp/CentOS-7-x86_64-NetInstall-1511.iso > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> Can I ask why you prefer/need to manually > create a VM installing from > >>>>>>>>>>>>>>>>>>>> a CD instead of using the ready-to-use > ovirt-engine-appliance? > >>>>>>>>>>>>>>>>>>>> Using the appliance makes the setup process > a lot shorted and more comfortable. > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> CPU Type : model_Penryn > >>>>>>>>>>>>>>>>>>>>> ... > >>>>>>>>>>>>>>>>>>>>> and get error after step "Verifying > sanlock lockspace initialization" > >>>>>>>>>>>>>>>>>>>>> ... > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace > initialization > >>>>>>>>>>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc > configuration': [Errno 101] Network > >>>>>>>>>>>>>>>>>>>>> is unreachable > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Clean up > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Generating answer file > >>>>>>>>>>>>>>>>>>>>> > '/var/lib/ovirt-hosted-engine-setup/answers/answers-20160722124133.conf' > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Termination > >>>>>>>>>>>>>>>>>>>>> [ ERROR ] Hosted Engine deployment > failed: this system is not reliable, > >>>>>>>>>>>>>>>>>>>>> please check the issue, fix and redeploy > >>>>>>>>>>>>>>>>>>>>> Log file is located at > >>>>>>>>>>>>>>>>>>>>> > /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup- > 20160722123404-t26vw0.log > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Interestingly > >>>>>>>>>>>>>>>>>>>>> ============================ > >>>>>>>>>>>>>>>>>>>>> If I try to deploy hosted-engine v3.6, > everything goes well in the same > >>>>>>>>>>>>>>>>>>>>> configuration !! : > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> .... > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction setup > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Package installation > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Misc configuration > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring libvirt > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring VDSM > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Starting vdsmd > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Waiting for VDSM hardware info > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring the management bridge > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Volume Group > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Storage Domain > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Storage Pool > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Connecting Storage Pool > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Verifying sanlock lockspace > initialization > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Image for > 'hosted-engine.lockspace' ... > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Image for > 'hosted-engine.lockspace' created successfully > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating Image for > 'hosted-engine.metadata' ... > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Image for > 'hosted-engine.metadata' created successfully > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating VM Image > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Destroying Storage Pool > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Start monitoring domain > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Configuring VM > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Updating hosted-engine > configuration > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Transaction commit > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Stage: Closing up > >>>>>>>>>>>>>>>>>>>>> [ INFO ] Creating VM > >>>>>>>>>>>>>>>>>>>>> You can now connect to the VM > with the following command: > >>>>>>>>>>>>>>>>>>>>> /bin/remote-viewer > vnc://localhost:5900 > >>>>>>>>>>>>>>>>>>>>> ... > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> What could be the problem? > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> > _______________________________________________ > >>>>>>>>>>>>>>>>>>>>> Users mailing list > >>>>>>>>>>>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>>>>>>>>>>> > http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>>>>> _______________________________________________ > >>>>>>>>>>>> Users mailing list > >>>>>>>>>>>> Users@ovirt.org > >>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>>>>> > >>>>>>>>> _______________________________________________ > >>>>>>>>> Users mailing list > >>>>>>>>> Users@ovirt.org > >>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>>> > >>>>>> _______________________________________________ > >>>>>> Users mailing list > >>>>>> Users@ovirt.org > >>>>>> http://lists.ovirt.org/mailman/listinfo/users > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users