On Wed, Dec 2, 2015 at 12:19 PM, Budur Nagaraju <nbud...@gmail.com> wrote:
> I have installed KVM in the nested environment in ESXi6.x version is that > recommended ? > I often use KVM over KVM in nested environment but honestly I never tried to run KVM over ESXi but I suspect that all of your issues comes from there. > apart from Hosted engine is there any other alternate way to configure > Engine HA cluster ? > Nothing else from the project. You can use two external VMs in cluster with pacemaker but it's completely up to you. > > > -Nagaraju > > > On Wed, Dec 2, 2015 at 4:11 PM, Simone Tiraboschi <stira...@redhat.com> > wrote: > >> >> >> On Wed, Dec 2, 2015 at 11:25 AM, Budur Nagaraju <nbud...@gmail.com> >> wrote: >> >>> pls fine the logs from the below mentioned URL, >>> >>> http://pastebin.com/ZeKyyFbN >>> >> >> OK, the issue is here: >> >> Thread-88::ERROR::2015-12-02 >> 15:06:27,735::vm::2358::vm.Vm::(_startUnderlyingVm) >> vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::The vm start process failed >> Traceback (most recent call last): >> File "/usr/share/vdsm/virt/vm.py", line 2298, in _startUnderlyingVm >> self._run() >> File "/usr/share/vdsm/virt/vm.py", line 3363, in _run >> self._connection.createXML(domxml, flags), >> File "/usr/lib/python2.6/site-packages/vdsm/libvirtconnection.py", line >> 119, in wrapper >> ret = f(*args, **kwargs) >> File "/usr/lib64/python2.6/site-packages/libvirt.py", line 2709, in >> createXML >> if ret is None:raise libvirtError('virDomainCreateXML() failed', >> conn=self) >> libvirtError: unsupported configuration: Domain requires KVM, but it is >> not available. Check that virtualization is enabled in the host BIOS, and >> host configuration is setup to load the kvm modules. >> Thread-88::DEBUG::2015-12-02 >> 15:06:27,751::vm::2813::vm.Vm::(setDownStatus) >> vmId=`93db4369-285f-48bc-bc68-181d9de41a3c`::Changed state to Down: >> unsupported configuration: Domain requires KVM, but it is not available. >> Check that virtualization is enabled in the host BIOS, and host >> configuration is setup to load the kvm modules. (code=1) >> >> but it's pretty strange cause hosted-engine-setup already explicitly >> check for visualization support and just exits with a clear error if not. >> Did you played with the kvm module while hosted-engine-setup was running? >> >> Can you please hosted-engine-setup logs? >> >> >>> >>> On Fri, Nov 27, 2015 at 6:39 PM, Simone Tiraboschi <stira...@redhat.com> >>> wrote: >>> >>>> >>>> >>>> On Fri, Nov 27, 2015 at 12:42 PM, Maxim Kovgan <kovg...@gmail.com> >>>> wrote: >>>> >>>>> Maybe even makes sense to open a bugzilla ticket already. Better safe >>>>> than sorry. >>>>> >>>> >>>> We still need at least one log file to understand what happened. >>>> >>>> >>>>> On Nov 27, 2015 11:35 AM, "Simone Tiraboschi" <stira...@redhat.com> >>>>> wrote: >>>>> >>>>>> >>>>>> On Fri, Nov 27, 2015 at 10:10 AM, Budur Nagaraju <nbud...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> I do not know what logs you are expecting ? the logs which I got is >>>>>>> pasted in the mail if you require in pastebin let me know I will upload >>>>>>> there . >>>>>>> >>>>>> >>>>>> >>>>>> Please run sosreport utility and share the resulting archive where >>>>>> you prefer. >>>>>> You can follow this guide: >>>>>> http://www.linuxtechi.com/how-to-create-sosreport-in-linux/ >>>>>> >>>>>>> >>>>>>> >>>>>>> On Fri, Nov 27, 2015 at 1:58 PM, Sandro Bonazzola < >>>>>>> sbona...@redhat.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Fri, Nov 27, 2015 at 8:34 AM, Budur Nagaraju <nbud...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> I got only 10lines to in the vdsm logs and are below , >>>>>>>>> >>>>>>>>> >>>>>>>> Can you please provide full sos report? >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> [root@he /]# tail -f /var/log/vdsm/vdsm.log >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,360::resourceManager::616::Storage.ResourceManager::(releaseResource) >>>>>>>>> Trying to release resource 'Storage.HsmDomainMonitorLock' >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,360::resourceManager::635::Storage.ResourceManager::(releaseResource) >>>>>>>>> Released resource 'Storage.HsmDomainMonitorLock' (0 active users) >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,360::resourceManager::641::Storage.ResourceManager::(releaseResource) >>>>>>>>> Resource 'Storage.HsmDomainMonitorLock' is free, finding out if >>>>>>>>> anyone is >>>>>>>>> waiting for it. >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,360::resourceManager::649::Storage.ResourceManager::(releaseResource) >>>>>>>>> No one is waiting for resource 'Storage.HsmDomainMonitorLock', >>>>>>>>> Clearing >>>>>>>>> records. >>>>>>>>> Thread-100::INFO::2015-11-27 >>>>>>>>> 12:58:57,360::logUtils::47::dispatcher::(wrapper) Run and protect: >>>>>>>>> stopMonitoringDomain, Return response: None >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,361::task::1191::Storage.TaskManager.Task::(prepare) >>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::finished: None >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,361::task::595::Storage.TaskManager.Task::(_updateState) >>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::moving from state >>>>>>>>> preparing -> >>>>>>>>> state finished >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,361::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) >>>>>>>>> Owner.releaseAll requests {} resources {} >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,361::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) >>>>>>>>> Owner.cancelAll requests {} >>>>>>>>> Thread-100::DEBUG::2015-11-27 >>>>>>>>> 12:58:57,361::task::993::Storage.TaskManager.Task::(_decref) >>>>>>>>> Task=`0128b179-fdb3-474b-a196-8cc81a72a837`::ref 0 aborting False >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Nov 26, 2015 at 4:20 PM, Simone Tiraboschi < >>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Thu, Nov 26, 2015 at 11:05 AM, Budur Nagaraju < >>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> *Below are the entire logs* >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> Sorry, with the entire log I mean if you can attach or share >>>>>>>>>> somewhere the whole /var/log/vdsm/vdsm.log cause the latest ten >>>>>>>>>> lines are >>>>>>>>>> not enough to point out the issue. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/vdsm.log * >>>>>>>>>>> >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>> Detected protocol xml from 127.0.0.1:50944 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>> over >>>>>>>>>>> http detected from ('127.0.0.1', 50944) >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>> Adding connection from 127.0.0.1:50945 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>> Connection removed from 127.0.0.1:50945 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>> Detected protocol xml from 127.0.0.1:50945 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>> over >>>>>>>>>>> http detected from ('127.0.0.1', 50945) >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>> Adding connection from 127.0.0.1:50946 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>> Connection removed from 127.0.0.1:50946 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>> Detected protocol xml from 127.0.0.1:50946 >>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>> 15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>> over >>>>>>>>>>> http detected from ('127.0.0.1', 50946) >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/supervdsm.log * >>>>>>>>>>> >>>>>>>>>>> MainProcess::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:30,234::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> call readMultipathConf with () {} >>>>>>>>>>> MainProcess::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:30,234::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> return readMultipathConf with ['# RHEV REVISION 1.1', '', 'defaults >>>>>>>>>>> {', >>>>>>>>>>> ' polling_interval 5', ' getuid_callout >>>>>>>>>>> "/lib/udev/scsi_id --whitelisted --replace-whitespace >>>>>>>>>>> --device=/dev/%n"', >>>>>>>>>>> ' no_path_retry fail', ' user_friendly_names >>>>>>>>>>> no', ' >>>>>>>>>>> flush_on_last_del yes', ' fast_io_fail_tmo 5', ' >>>>>>>>>>> dev_loss_tmo 30', ' max_fds 4096', >>>>>>>>>>> '}', '', >>>>>>>>>>> 'devices {', 'device {', ' vendor "HITACHI"', ' >>>>>>>>>>> product "DF.*"', ' getuid_callout >>>>>>>>>>> "/lib/udev/scsi_id --whitelisted --replace-whitespace >>>>>>>>>>> --device=/dev/%n"', >>>>>>>>>>> '}', 'device {', ' vendor "COMPELNT"', ' >>>>>>>>>>> product "Compellent Vol"', ' no_path_retry >>>>>>>>>>> fail', '}', 'device {', ' # multipath.conf.default', ' >>>>>>>>>>> vendor "DGC"', ' product ".*"', >>>>>>>>>>> ' >>>>>>>>>>> product_blacklist "LUNZ"', ' path_grouping_policy >>>>>>>>>>> "group_by_prio"', ' path_checker "emc_clariion"', ' >>>>>>>>>>> hardware_handler "1 emc"', ' prio >>>>>>>>>>> "emc"', ' >>>>>>>>>>> failback immediate', ' rr_weight >>>>>>>>>>> "uniform"', ' # vdsm required configuration', ' >>>>>>>>>>> getuid_callout "/lib/udev/scsi_id --whitelisted >>>>>>>>>>> --replace-whitespace --device=/dev/%n"', ' features >>>>>>>>>>> "0"', >>>>>>>>>>> ' no_path_retry fail', '}', '}'] >>>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:31,365::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> call getHardwareInfo with () {} >>>>>>>>>>> MainProcess|Thread-13::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:31,397::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> return getHardwareInfo with {'systemProductName': 'KVM', >>>>>>>>>>> 'systemUUID': >>>>>>>>>>> 'f91632f2-7a17-4ddb-9631-742f82a77480', 'systemFamily': 'Red Hat >>>>>>>>>>> Enterprise >>>>>>>>>>> Linux', 'systemVersion': 'RHEL 7.0.0 PC (i440FX + PIIX, 1996)', >>>>>>>>>>> 'systemManufacturer': 'Red Hat'} >>>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:35,393::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> call validateAccess with ('qemu', ('qemu', 'kvm'), >>>>>>>>>>> '/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {} >>>>>>>>>>> MainProcess|Thread-21::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:35,395::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> return validateAccess with None >>>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:36,067::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> call validateAccess with ('qemu', ('qemu', 'kvm'), >>>>>>>>>>> '/rhev/data-center/mnt/10.204.207.152:_home_vms', 5) {} >>>>>>>>>>> MainProcess|Thread-22::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:36,069::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> return validateAccess with None >>>>>>>>>>> MainProcess|PolicyEngine::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:40,619::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> call ksmTune with ({'run': 0},) {} >>>>>>>>>>> MainProcess|PolicyEngine::DEBUG::2015-11-26 >>>>>>>>>>> 15:13:40,619::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper) >>>>>>>>>>> return ksmTune with None >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/connectivity.log * >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2015-11-26 15:02:02,632:DEBUG:recent_client:False >>>>>>>>>>> 2015-11-26 15:04:44,975:DEBUG:recent_client:True >>>>>>>>>>> 2015-11-26 15:05:15,039:DEBUG:recent_client:False >>>>>>>>>>> 2015-11-26 15:07:23,311:DEBUG:recent_client:True >>>>>>>>>>> 2015-11-26 15:08:25,774:DEBUG:recent_client:True, >>>>>>>>>>> ovirtmgmt:(operstate:up speed:0 duplex:unknown), lo:(operstate:up >>>>>>>>>>> speed:0 >>>>>>>>>>> duplex:unknown), ;vdsmdummy;:(operstate:down speed:0 >>>>>>>>>>> duplex:unknown), >>>>>>>>>>> bond0:(operstate:down speed:0 duplex:unknown), eth0:(operstate:up >>>>>>>>>>> speed:100 >>>>>>>>>>> duplex:full) >>>>>>>>>>> 2015-11-26 15:08:55,845:DEBUG:recent_client:False >>>>>>>>>>> 2015-11-26 15:08:59,859:DEBUG:recent_client:True >>>>>>>>>>> 2015-11-26 15:09:29,929:DEBUG:recent_client:False >>>>>>>>>>> 2015-11-26 15:13:32,292:DEBUG:recent_client:True, >>>>>>>>>>> ovirtmgmt:(operstate:up speed:0 duplex:unknown), lo:(operstate:up >>>>>>>>>>> speed:0 >>>>>>>>>>> duplex:unknown), ;vdsmdummy;:(operstate:down speed:0 >>>>>>>>>>> duplex:unknown), >>>>>>>>>>> bond0:(operstate:down speed:0 duplex:unknown), eth0:(operstate:up >>>>>>>>>>> speed:100 >>>>>>>>>>> duplex:full) >>>>>>>>>>> 2015-11-26 15:14:02,363:DEBUG:recent_client:False >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> *[root@he ~]# tail -f /var/log/vdsm/mom.log * >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2015-11-26 15:13:30,581 - mom.Policy - INFO - Loaded policy >>>>>>>>>>> '04-cputune' >>>>>>>>>>> 2015-11-26 15:13:30,581 - mom.PolicyEngine - INFO - Policy >>>>>>>>>>> Engine starting >>>>>>>>>>> 2015-11-26 15:13:30,582 - mom.RPCServer - INFO - RPC Server is >>>>>>>>>>> disabled >>>>>>>>>>> 2015-11-26 15:13:40,618 - mom.Controllers.KSM - INFO - Updating >>>>>>>>>>> KSM configuration: pages_to_scan:0 merge_across_nodes:8 run:0 >>>>>>>>>>> sleep_millisecs:0 >>>>>>>>>>> 2015-11-26 15:14:51,492 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>>> 2015-11-26 15:14:56,962 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>>> 2015-11-26 15:15:02,451 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>>> 2015-11-26 15:15:07,777 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>>> 2015-11-26 15:15:13,267 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>>> 2015-11-26 15:15:18,765 - mom.RPCServer - INFO - getStatistics() >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Nov 26, 2015 at 3:28 PM, Budur Nagaraju < >>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Below are the logs, >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> [root@he ~]# tail -f /var/log/vdsm/vdsm.log >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:05,622::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>> Detected protocol xml from 127.0.0.1:50944 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:05,623::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>>> over >>>>>>>>>>>> http detected from ('127.0.0.1', 50944) >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:05,703::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>>> Adding connection from 127.0.0.1:50945 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,101::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>>> Connection removed from 127.0.0.1:50945 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,101::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>> Detected protocol xml from 127.0.0.1:50945 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,101::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>>> over >>>>>>>>>>>> http detected from ('127.0.0.1', 50945) >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,182::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>>> Adding connection from 127.0.0.1:50946 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,710::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>>> Connection removed from 127.0.0.1:50946 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,711::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>> Detected protocol xml from 127.0.0.1:50946 >>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>> 15:16:06,711::BindingXMLRPC::1173::XmlDetector::(handleSocket) xml >>>>>>>>>>>> over >>>>>>>>>>>> http detected from ('127.0.0.1', 50946) >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Nov 26, 2015 at 3:06 PM, Simone Tiraboschi < >>>>>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Nov 26, 2015 at 10:33 AM, Budur Nagaraju < >>>>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> I have done a fresh installation and now am getting the below >>>>>>>>>>>>>> error, >>>>>>>>>>>>>> >>>>>>>>>>>>>> [ INFO ] Updating hosted-engine configuration >>>>>>>>>>>>>> [ INFO ] Stage: Transaction commit >>>>>>>>>>>>>> [ INFO ] Stage: Closing up >>>>>>>>>>>>>> The following network ports should be opened: >>>>>>>>>>>>>> tcp:5900 >>>>>>>>>>>>>> tcp:5901 >>>>>>>>>>>>>> udp:5900 >>>>>>>>>>>>>> udp:5901 >>>>>>>>>>>>>> An example of the required configuration for >>>>>>>>>>>>>> iptables can be found at: >>>>>>>>>>>>>> /etc/ovirt-hosted-engine/iptables.example >>>>>>>>>>>>>> In order to configure firewalld, copy the files from >>>>>>>>>>>>>> /etc/ovirt-hosted-engine/firewalld to >>>>>>>>>>>>>> /etc/firewalld/services >>>>>>>>>>>>>> and execute the following commands: >>>>>>>>>>>>>> firewall-cmd -service hosted-console >>>>>>>>>>>>>> [ INFO ] Creating VM >>>>>>>>>>>>>> [ ERROR ] Failed to execute stage 'Closing up': Cannot set >>>>>>>>>>>>>> temporary password for console connection. The VM may not have >>>>>>>>>>>>>> been >>>>>>>>>>>>>> created: please check VDSM logs >>>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126145701.conf' >>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> [root@he ovirt]# tail -f /var/log/vdsm/ >>>>>>>>>>>>>> backup/ connectivity.log mom.log >>>>>>>>>>>>>> supervdsm.log vdsm.log >>>>>>>>>>>>>> [root@he ovirt]# tail -f /var/log/vdsm/vdsm.log >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:07,564::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>>>> Detected protocol xml from 127.0.0.1:42741 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:07,564::BindingXMLRPC::1173::XmlDetector::(handleSocket) >>>>>>>>>>>>>> xml over >>>>>>>>>>>>>> http detected from ('127.0.0.1', 42741) >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:07,644::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>>>>> Adding connection from 127.0.0.1:42742 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,088::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>>>>> Connection removed from 127.0.0.1:42742 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,088::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>>>> Detected protocol xml from 127.0.0.1:42742 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,088::BindingXMLRPC::1173::XmlDetector::(handleSocket) >>>>>>>>>>>>>> xml over >>>>>>>>>>>>>> http detected from ('127.0.0.1', 42742) >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,171::protocoldetector::187::vds.MultiProtocolAcceptor::(_add_connection) >>>>>>>>>>>>>> Adding connection from 127.0.0.1:42743 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,572::protocoldetector::201::vds.MultiProtocolAcceptor::(_remove_connection) >>>>>>>>>>>>>> Connection removed from 127.0.0.1:42743 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,573::protocoldetector::247::vds.MultiProtocolAcceptor::(_handle_connection_read) >>>>>>>>>>>>>> Detected protocol xml from 127.0.0.1:42743 >>>>>>>>>>>>>> Detector thread::DEBUG::2015-11-26 >>>>>>>>>>>>>> 14:57:08,573::BindingXMLRPC::1173::XmlDetector::(handleSocket) >>>>>>>>>>>>>> xml over >>>>>>>>>>>>>> http detected from ('127.0.0.1', 42743) >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> It failed before, can you please attach the whole VDSM logs? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 2:01 PM, Simone Tiraboschi < >>>>>>>>>>>>>> stira...@redhat.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 7:30 AM, Budur Nagaraju < >>>>>>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Its a fresh setup ,I have deleted all the vms ,still am >>>>>>>>>>>>>>>> facing same issues . >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Can you please paste the output of >>>>>>>>>>>>>>> vdsClient -s 0 list >>>>>>>>>>>>>>> ? >>>>>>>>>>>>>>> thanks >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 11:56 AM, Oved Ourfali < >>>>>>>>>>>>>>>> oourf...@redhat.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Seems like you have existing VMs running on the host (you >>>>>>>>>>>>>>>>> can check that by looking for qemu processes on your host). >>>>>>>>>>>>>>>>> Is that a clean deployment, or was the host used before >>>>>>>>>>>>>>>>> for running VMs? >>>>>>>>>>>>>>>>> Perhaps you already ran the hosted engine setup, and the >>>>>>>>>>>>>>>>> VM was left there? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> CC-ing Sandro who is more familiar in that than me. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>> Oved >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Nov 26, 2015 at 7:07 AM, Budur Nagaraju < >>>>>>>>>>>>>>>>> nbud...@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> HI >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Getting below error while configuring Hosted engine, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> root@he ~]# hosted-engine --deploy >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Initializing >>>>>>>>>>>>>>>>>> [ INFO ] Generating a temporary VNC password. >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Environment setup >>>>>>>>>>>>>>>>>> Continuing will configure this host for serving >>>>>>>>>>>>>>>>>> as hypervisor and create a VM where you have to install >>>>>>>>>>>>>>>>>> oVirt Engine >>>>>>>>>>>>>>>>>> afterwards. >>>>>>>>>>>>>>>>>> Are you sure you want to continue? (Yes, >>>>>>>>>>>>>>>>>> No)[Yes]: yes >>>>>>>>>>>>>>>>>> Configuration files: [] >>>>>>>>>>>>>>>>>> Log file: >>>>>>>>>>>>>>>>>> /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20151126102302-bkozgk.log >>>>>>>>>>>>>>>>>> Version: otopi-1.3.2 (otopi-1.3.2-1.el6) >>>>>>>>>>>>>>>>>> It has been detected that this program is >>>>>>>>>>>>>>>>>> executed through an SSH connection without using screen. >>>>>>>>>>>>>>>>>> Continuing with the installation may lead to >>>>>>>>>>>>>>>>>> broken installation if the network connection fails. >>>>>>>>>>>>>>>>>> It is highly recommended to abort the >>>>>>>>>>>>>>>>>> installation and run it inside a screen session using >>>>>>>>>>>>>>>>>> command "screen". >>>>>>>>>>>>>>>>>> Do you want to continue anyway? (Yes, No)[No]: >>>>>>>>>>>>>>>>>> yes >>>>>>>>>>>>>>>>>> [WARNING] Cannot detect if hardware supports >>>>>>>>>>>>>>>>>> virtualization >>>>>>>>>>>>>>>>>> [ INFO ] Bridge ovirtmgmt already created >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Environment packages setup >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Programs detection >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Environment setup >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> *[ ERROR ] The following VMs has been found: >>>>>>>>>>>>>>>>>> 2b8d6d91-d838-44f6-ae3b-c92cda014280[ ERROR ] Failed to >>>>>>>>>>>>>>>>>> execute stage >>>>>>>>>>>>>>>>>> 'Environment setup': Cannot setup Hosted Engine with other >>>>>>>>>>>>>>>>>> VMs running* >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Clean up >>>>>>>>>>>>>>>>>> [ INFO ] Generating answer file >>>>>>>>>>>>>>>>>> '/var/lib/ovirt-hosted-engine-setup/answers/answers-20151126102310.conf' >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Pre-termination >>>>>>>>>>>>>>>>>> [ INFO ] Stage: Termination >>>>>>>>>>>>>>>>>> [root@he ~]# >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Users mailing list >>>>>>>>> Users@ovirt.org >>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Sandro Bonazzola >>>>>>>> Better technology. Faster innovation. Powered by community >>>>>>>> collaboration. >>>>>>>> See how it works at redhat.com >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> Users@ovirt.org >>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>> >>>>>> >>>> >>> >> >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users