On Tue, Jun 6, 2017 at 2:56 PM, Brendan Hartzell <mrr...@gmail.com> wrote:
> Upon login to the server, to watch terminal output, I noticed that the > node status is degraded. > > [root@node-1 ~]# nodectl check > Status: WARN > Bootloader ... OK > Layer boot entries ... OK > Valid boot entries ... OK > Mount points ... OK > Separate /var ... OK > Discard is used ... OK > Basic storage ... OK > Initialized VG ... OK > Initialized Thin Pool ... OK > Initialized LVs ... OK > Thin storage ... OK > Checking available space in thinpool ... OK > Checking thinpool auto-extend ... OK > vdsmd ... BAD > Yuval, can you help here? > > > Pressing forward with the retry using the web-UI. > > After resetting my iSCSI storage (on the storage server side), Install > started. > > Status in the web-UI: > Creating Storage Domain > Creating Storage Pool > Connecting Storage Pool > Verifying sanlock lockspace initialization > Creating Image for 'hosted-engine.lockspace' ... > Image for 'hosted-engine.lockspace' created successfully > Creating Image for 'hosted-engine.metadata' ... > Image for 'hosted-engine.metadata' created successfully > Creating VM Image > Extracting disk image from OVF archive (could take a few minutes depending > on archive size) > Validating pre-allocated volume size > > Output from the terminal: > [45863.076979]watchdog watchdog0: watchdog did not stop! > > System restarted. > > Attaching ovirt-hosted-engine-setup log. > > I'm running an SOS report, but it's too big for the users list. I can > email it directly to you upon request. > > On Tue, Jun 6, 2017 at 12:12 AM, Simone Tiraboschi <stira...@redhat.com> > wrote: > >> >> >> On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell <mrr...@gmail.com> >> wrote: >> >>> As requested, >>> >> >> It seams fine, there are no pending locks now. >> Could you please retry? >> >> >>> >>> The output of ovirt-hosted-engine-cleanup >>> >>> [root@node-1 ~]# ovirt-hosted-engine-cleanup >>> This will de-configure the host to run ovirt-hosted-engine-setup from >>> scratch. >>> Caution, this operation should be used with care. >>> >>> Are you sure you want to proceed? [y/n] >>> y >>> -=== Destroy hosted-engine VM ===- >>> You must run deploy first >>> -=== Stop HA services ===- >>> -=== Shutdown sanlock ===- >>> shutdown force 1 wait 0 >>> shutdown done 0 >>> -=== Disconnecting the hosted-engine storage domain ===- >>> You must run deploy first >>> -=== De-configure VDSM networks ===- >>> -=== Stop other services ===- >>> -=== De-configure external daemons ===- >>> -=== Removing configuration files ===- >>> ? /etc/init/libvirtd.conf already missing >>> - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml >>> ? /etc/ovirt-hosted-engine/answers.conf already missing >>> ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing >>> - removing /etc/vdsm/vdsm.conf >>> - removing /etc/pki/vdsm/certs/cacert.pem >>> - removing /etc/pki/vdsm/certs/vdsmcert.pem >>> - removing /etc/pki/vdsm/keys/vdsmkey.pem >>> - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem >>> - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem >>> - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem >>> - removing /etc/pki/vdsm/libvirt-spice/server-key.pem >>> ? /etc/pki/CA/cacert.pem already missing >>> ? /etc/pki/libvirt/*.pem already missing >>> ? /etc/pki/libvirt/private/*.pem already missing >>> ? /etc/pki/ovirt-vmconsole/*.pem already missing >>> - removing /var/cache/libvirt/qemu >>> ? /var/run/ovirt-hosted-engine-ha/* already missing >>> [root@node-1 ~]# >>> >>> Output of sanlock client status: >>> [root@node-1 ~]# sanlock client status >>> [root@node-1 ~]# >>> >>> Thank you for your help! >>> >>> On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi <stira...@redhat.com> >>> wrote: >>> >>>> >>>> >>>> On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell <mrr...@gmail.com> >>>> wrote: >>>> >>>>> After letting this sit for a few days, does anyone have any ideas as >>>>> to how to deal with my situation? Would anyone like me to send the SOS >>>>> report directly to them? It's a 9MB file. >>>>> >>>>> If nothing comes up, I'm going to try and sift through the SOS report >>>>> tonight, but I won't know what I'm trying to find. >>>>> >>>>> Thank you for any and all help. >>>>> >>>>> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola <sbona...@redhat.com> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell <mrr...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Ran the 4 commands listed above, no errors on the screen. >>>>>>> >>>>>>> Started the hosted-engine standard setup from the web-UI. >>>>>>> >>>>>>> Using iSCSI for the storage. >>>>>>> >>>>>>> Using mostly default options, I got these errors in the web-UI. >>>>>>> >>>>>>> Error creating Volume Group: Failed to initialize physical device: >>>>>>> ("[u'/dev/mapper/36589cfc000000de7482638fcfcebbbb4']",) >>>>>>> Failed to execute stage 'Misc configuration': Failed to initialize >>>>>>> physical device: ("[u'/dev/mapper/36589cfc00000 >>>>>>> 0de7482638fcfcebbbb4']",) >>>>>>> Hosted Engine deployment failed: this system is not reliable, please >>>>>>> check the issue,fix and redeploy >>>>>>> >>>>>>> I rebuilt my iSCSI (I don't think I cleaned it up from a previous >>>>>>> install). >>>>>>> Re-ran the above 4 commands. >>>>>>> Restarted hosted engine standard setup from web-UI. >>>>>>> Install moved past "Connecting Storage Pool" so I believe the above >>>>>>> was my fault. >>>>>>> >>>>>>> These are the last messages displayed on the web-UI. >>>>>>> Creating Storage Pool >>>>>>> Connecting Storage Pool >>>>>>> Verifying sanlock lockspace initialization >>>>>>> Creating Image for 'hosted-engine.lockspace' ... >>>>>>> Image for 'hosted-engine.lockspace' created successfully >>>>>>> Creating Image for 'hosted-engine.metadata' ... >>>>>>> Image for 'hosted-engine.metadata' created successfully >>>>>>> Creating VM Image >>>>>>> Extracting disk image from OVF archive (could take a few minutes >>>>>>> depending on archive size) >>>>>>> Validating pre-allocated volume size >>>>>>> Uploading volume to data domain (could take a few minutes depending >>>>>>> on archive size) >>>>>>> >>>>>>> At the host terminal, I got the error "watchdog watchdog0: watchdog >>>>>>> did not stop!" >>>>>>> Then the host restarted. >>>>>>> >>>>>> >>>>>> Simone, can you help here? >>>>>> >>>>>> >>>> Ok, sorry for the delay. >>>> The second installation attempt seams fine but it seams that >>>> ovirt-hosted-engine-cleanup failed stopping sanlock and so the >>>> watchdog kick in rebooting your system in the middle of deployment attempt. >>>> >>>> could you please post the output of >>>> ovirt-hosted-engine-cleanup >>>> sanlock client status >>>> ? >>>> >>>> >>>> >>>>> >>>>>> >>>>>> >>>>>>> >>>>>>> This is as far as I've gotten in previous attempts. >>>>>>> >>>>>>> Attaching the hosted-engine-setup log. >>>>>>> >>>>>>> The SOS report is 9MB and the ovirt users group will drop the email. >>>>>>> >>>>>>> On Wed, May 31, 2017 at 6:59 AM, Sandro Bonazzola < >>>>>>> sbona...@redhat.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, May 31, 2017 at 3:10 PM, Brendan Hartzell <mrr...@gmail.com >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> Now that you have identified the problem, should I run the >>>>>>>>> following commands and send you another SOS? >>>>>>>>> >>>>>>>>> ovirt-hosted-engine-cleanup >>>>>>>>> vdsm-tool configure --force >>>>>>>>> systemctl restart libvirtd >>>>>>>>> systemctl restart vdsm >>>>>>>>> >>>>>>>>> Or is there a different plan in mind? >>>>>>>>> >>>>>>>> >>>>>>>> I would have expected someone from virt team to follow up for >>>>>>>> further investigations :-) >>>>>>>> above commands should work. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> Thank you, >>>>>>>>> >>>>>>>>> Brendan >>>>>>>>> >>>>>>>>> On Tue, May 30, 2017 at 11:42 PM, Sandro Bonazzola < >>>>>>>>> sbona...@redhat.com> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, May 31, 2017 at 4:45 AM, Brendan Hartzell < >>>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Can you please elaborate about the failure you see here and how >>>>>>>>>>> are you trying to manually partition the host? >>>>>>>>>>> >>>>>>>>>>> Sure, I will start from the beginning. >>>>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>>>>> <(201)%20705-2604>.iso >>>>>>>>>>> - During installation I setup one of the two interfaces and >>>>>>>>>>> check the box to automatically use the connection. >>>>>>>>>>> - I'm currently providing a host name of node-1.test.net until >>>>>>>>>>> I have a successful process. >>>>>>>>>>> - I configure date and time for my timezone and to use an >>>>>>>>>>> internal NTP server. >>>>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>>>>>>>>>> flash drive, check the box that I would like to make additional >>>>>>>>>>> space, and >>>>>>>>>>> click done. In the reclaim disk space window, I click delete all, >>>>>>>>>>> and then >>>>>>>>>>> reclaim space. I go back into the Installation Destination, select >>>>>>>>>>> that I >>>>>>>>>>> will configure partitioning, and click done. The Manual >>>>>>>>>>> Partitioning >>>>>>>>>>> window opens, I use the option to automatically create mount points. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> In this screen, please change partitioning scheme from LVM to LVM >>>>>>>>>> Thin Provisioning: it should solve your following error. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> At this point, /boot is 1024MB, /var is 15GB, / is 88.11 GB, >>>>>>>>>>> and swap is 11.57GB. I then change / to 23.11 GB, update settings, >>>>>>>>>>> change >>>>>>>>>>> /var to 80GB, update settings again, and click done. I accept the >>>>>>>>>>> changes >>>>>>>>>>> and begin installation. >>>>>>>>>>> >>>>>>>>>>> I tried these changes based on this article: >>>>>>>>>>> http://www.ovirt.org/documentation/self-hosted/chap >>>>>>>>>>> -Deploying_Self-Hosted_Engine/ >>>>>>>>>>> >>>>>>>>>>> The article does say that you can specify a different directory >>>>>>>>>>> than /var/tmp, but I don't recall seeing that option. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> If the setup detects not enough space in /var/tmp for extracting >>>>>>>>>> the appliance it will ask about a different directory. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> After some time, I get the following error: >>>>>>>>>>> There was an error running the kickstart script at line 7. This >>>>>>>>>>> is a fatal error and installation will be aborted. The details of >>>>>>>>>>> this >>>>>>>>>>> error are: >>>>>>>>>>> >>>>>>>>>>> [INFO] Trying to create a manageable base from '/' >>>>>>>>>>> [ERROR] LVM Thin Provisioning partitioning scheme is required. >>>>>>>>>>> For autoinstall via Kickstart with LVM Thin Provisioning check >>>>>>>>>>> options >>>>>>>>>>> --thinpool and --grow. Please consult documentation for details. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> ^^ this one should be solved by the LVM Thin Provisioning scheme >>>>>>>>>> mentioned above.. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Traceback (most recent call last): >>>>>>>>>>> File "/usr/lib64/python2.7/runpy.py", line 162, in >>>>>>>>>>> _run_module_as_main "__main__", fname, loader, pkg_name) >>>>>>>>>>> File "/usr/lib64/python2.7runpy.py", line 72, in _run_code exec >>>>>>>>>>> code in run_globals >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__main__.py", >>>>>>>>>>> line 51, in <module> CliApplication() >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/__init__.py", >>>>>>>>>>> line 82, in CliApplication() >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/hooks.py", line >>>>>>>>>>> 120, in emit cb(self.context, *args) >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>>>>> line 169, in post_argparse layout.initialize(args.source, >>>>>>>>>>> args.init_nvr) >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/plugins/core.py", >>>>>>>>>>> line 216, in initialize self.app.imgbase.init_layout_from(source, >>>>>>>>>>> init_nvr) >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", >>>>>>>>>>> line 271, in init_layout_from self.init_tags_on(existing_lv) >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/imgbase.py", >>>>>>>>>>> line 243, in init_tags_on pool = lv.thinpool() >>>>>>>>>>> File "/usr/lib/python2.7/site-packages/imgbased/lvm.py", line >>>>>>>>>>> 250, in thinpool raise MissingLvmThinPool() >>>>>>>>>>> imgbased.lvm.MissingLvmThinPoo >>>>>>>>>>> l >>>>>>>>>>> >>>>>>>>>>> At this point, the only option is to exit the installer. >>>>>>>>>>> >>>>>>>>>>> **************************** >>>>>>>>>>> >>>>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>>>> supported anymore. >>>>>>>>>>> >>>>>>>>>>> Not a problem. >>>>>>>>>>> >>>>>>>>>>> **************************** >>>>>>>>>>> >>>>>>>>>>> Can you please provide hosted engine setup logs or better a full >>>>>>>>>>> sos report? (sosreport -a) >>>>>>>>>>> >>>>>>>>>>> Again, the process I'm following: >>>>>>>>>>> - Using: ovirt-node-ng-installer-ovirt-4.1-2017052604 >>>>>>>>>>> <(201)%20705-2604>.iso >>>>>>>>>>> - During installation I setup one of the two interfaces and >>>>>>>>>>> check the box to automatically use the connection. >>>>>>>>>>> - I'm currently providing a host name of node-1.test.net until >>>>>>>>>>> I have a successful process. >>>>>>>>>>> - I configure date and time for my timezone and to use an >>>>>>>>>>> internal NTP server. >>>>>>>>>>> - On Installation Destination, I pick my 128GB USB3.0 SanDisk >>>>>>>>>>> flash drive, check the box that I would like to make additional >>>>>>>>>>> space, and >>>>>>>>>>> click done. In the reclaim disk space window, I click delete all, >>>>>>>>>>> and then >>>>>>>>>>> reclaim space. >>>>>>>>>>> - Begin Installation and set a root password. >>>>>>>>>>> - Perform a yum update - no packages marked for update (as >>>>>>>>>>> expected) >>>>>>>>>>> - Use vi to update /etc/hosts with a reference for >>>>>>>>>>> node-1.test.net and engine.test.net >>>>>>>>>>> - First attempt at hosted-engine from web-UI >>>>>>>>>>> - Setup downloads and installs ovirt-engine-applianc >>>>>>>>>>> e-4.1-20170523.1.el7.centos.noarch.rpm >>>>>>>>>>> *Failed to execute stage 'Environment setup': Failed to >>>>>>>>>>> reconfigure libvirt for VDSM >>>>>>>>>>> *Hosted Engine deployment failed >>>>>>>>>>> - Attached SOS report >>>>>>>>>>> The checksum is: aa56097edc0b63c49caaf1a1fde021bc >>>>>>>>>>> >>>>>>>>>>> At this point, I would run ovirt-hosted-engine-cleanup and I >>>>>>>>>>> would get further along in the install process. However, because >>>>>>>>>>> this is a >>>>>>>>>>> fresh install, I'm going to leave things here for now so you can >>>>>>>>>>> review the >>>>>>>>>>> SOS. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thanks for the SOS report! >>>>>>>>>> Hosted Engine setup fails on: >>>>>>>>>> >>>>>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>>>>> plugin.execute:921 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>>>>> '--force') stdout: >>>>>>>>>> >>>>>>>>>> Checking configuration status... >>>>>>>>>> >>>>>>>>>> Current revision of multipath.conf detected, preserving >>>>>>>>>> lvm is configured for vdsm >>>>>>>>>> libvirt is already configured for vdsm >>>>>>>>>> SUCCESS: ssl configured to true. No conflicts >>>>>>>>>> >>>>>>>>>> Running configure... >>>>>>>>>> Reconfiguration of libvirt is done. >>>>>>>>>> >>>>>>>>>> 2017-05-30 19:24:39 DEBUG otopi.plugins.gr_he_setup.system.vdsmenv >>>>>>>>>> plugin.execute:926 execute-output: ('/bin/vdsm-tool', 'configure', >>>>>>>>>> '--force') stderr: >>>>>>>>>> Error: ServiceOperationError: _systemctlStart failed >>>>>>>>>> Job for libvirtd.service failed because the control process >>>>>>>>>> exited with error code. See "systemctl status libvirtd.service" and >>>>>>>>>> "journalctl -xe" for details. >>>>>>>>>> >>>>>>>>>> At the same time journalctl shows: >>>>>>>>>> >>>>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: libvirt >>>>>>>>>> version: 2.0.0, package: 10.el7_3.5 (CentOS BuildSystem < >>>>>>>>>> http://bugs.centos.org>, 2017-03-03-02:09:45, >>>>>>>>>> c1bm.rdu2.centos.org) >>>>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: hostname: >>>>>>>>>> node-1.test.net >>>>>>>>>> May 30 19:24:39 node-1.test.net libvirtd[20954]: The server >>>>>>>>>> certificate /etc/pki/vdsm/certs/vdsmcert.pem is not yet active >>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service: >>>>>>>>>> main process exited, code=exited, status=6/NOTCONFIGURED >>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Failed to start >>>>>>>>>> Virtualization daemon. >>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: Unit >>>>>>>>>> libvirtd.service entered failed state. >>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>>>>> failed. >>>>>>>>>> May 30 19:24:39 node-1.test.net systemd[1]: libvirtd.service >>>>>>>>>> holdoff time over, scheduling restart. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> **************************** >>>>>>>>>>> >>>>>>>>>>> I'd like to understand the issues you faced before suggesting to >>>>>>>>>>> restart from scratch. >>>>>>>>>>> >>>>>>>>>>> Too late... I did two re-installs to get a more accurate account >>>>>>>>>>> of my install process for above. >>>>>>>>>>> >>>>>>>>>>> **************************** >>>>>>>>>>> >>>>>>>>>>> Thank you for your help! >>>>>>>>>>> >>>>>>>>>>> Brendan >>>>>>>>>>> >>>>>>>>>>> On Tue, May 30, 2017 at 12:17 AM, Sandro Bonazzola < >>>>>>>>>>> sbona...@redhat.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Tue, May 30, 2017 at 6:49 AM, Brendan Hartzell < >>>>>>>>>>>> mrr...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> oVirt users list, >>>>>>>>>>>>> >>>>>>>>>>>>> Long story short, I've been spending weeks on this project for >>>>>>>>>>>>> my home lab with no success. >>>>>>>>>>>>> >>>>>>>>>>>>> I would like to successfully install two nodes that host a >>>>>>>>>>>>> highly available engine with an iSCSI storage back-end. >>>>>>>>>>>>> >>>>>>>>>>>>> I have read through most, if not all, of the guides on >>>>>>>>>>>>> ovirt.org with no substantial help. >>>>>>>>>>>>> >>>>>>>>>>>>> Successfully, I have done the following: >>>>>>>>>>>>> Install oVirt Engine on a bare metal system, added a node, and >>>>>>>>>>>>> started exploring - not desired. >>>>>>>>>>>>> Install oVirt Node 4.0.6 on a bare metal system - fails if >>>>>>>>>>>>> partitions are not done automatically. >>>>>>>>>>>>> Install oVirt Node 4.1.2 on a bare metal system - fails if >>>>>>>>>>>>> partitions are not done automatically. >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Can you please elaborate about the failure you see here and how >>>>>>>>>>>> are you trying to manually partition the host? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> My process after installing a Node: >>>>>>>>>>>>> Run a yum update - just to be sure, but I am using latest iso >>>>>>>>>>>>> images from downloads section. >>>>>>>>>>>>> Edit /etc/hosts for local name resolution - the goal is to >>>>>>>>>>>>> host DNS as a virtual machine, eventually. >>>>>>>>>>>>> On 4.1 if I install ovirt-engine-appliance from yum, it does >>>>>>>>>>>>> simplify one step in the hosted engine setup. If I do this on >>>>>>>>>>>>> 4.0 it >>>>>>>>>>>>> discards the image and uses the default. >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Being this a new install, please use 4.1. oVirt 4.0 is not >>>>>>>>>>>> supported anymore. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> On 4.1 the hosted engine setup fails immediately unless I run >>>>>>>>>>>>> the hosted engine cleanup from the shell. >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Can you please provide hosted engine setup logs or better a >>>>>>>>>>>> full sos report? (sosreport -a) >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> If I do this, I can typically get to the point of installing. >>>>>>>>>>>>> >>>>>>>>>>>>> When I do get to the installation phase, I get to a point just >>>>>>>>>>>>> after extracting the OVA that I get a message on the shell saying >>>>>>>>>>>>> something >>>>>>>>>>>>> about the watchdog running the whole time and then the node >>>>>>>>>>>>> reboots. >>>>>>>>>>>>> >>>>>>>>>>>>> I found one email thread that sounded like my issue and >>>>>>>>>>>>> suggested the following commands: >>>>>>>>>>>>> vdsm-tool configure --force >>>>>>>>>>>>> systemctl restart libvirtd >>>>>>>>>>>>> systemctl restart vdsmd >>>>>>>>>>>>> >>>>>>>>>>>>> Unfortunately, these commands did not help my situation like >>>>>>>>>>>>> the other individual. >>>>>>>>>>>>> >>>>>>>>>>>>> What log file would everyone like to see first? Given that I >>>>>>>>>>>>> still consider myself relatively new to Linux, please identify >>>>>>>>>>>>> the path for >>>>>>>>>>>>> the log file requested. >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> See above >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Also, because I plan on performing a clean install for this >>>>>>>>>>>>> thread using my process from above (I'm not expecting my outcome >>>>>>>>>>>>> to be any >>>>>>>>>>>>> different), are there any tips and tricks that might result in a >>>>>>>>>>>>> success? >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> I'd like to understand the issues you faced before suggesting >>>>>>>>>>>> to restart from scratch. >>>>>>>>>>>> Adding some people who may help as well. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Thank you for any and all help, >>>>>>>>>>>>> Brendan >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>> Users@ovirt.org >>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> >>>>>>>>>>>> SANDRO BONAZZOLA >>>>>>>>>>>> >>>>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG >>>>>>>>>>>> VIRTUALIZATION R&D >>>>>>>>>>>> >>>>>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>>>>> <https://red.ht/sig> >>>>>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> SANDRO BONAZZOLA >>>>>>>>>> >>>>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION >>>>>>>>>> R&D >>>>>>>>>> >>>>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>>>> <https://red.ht/sig> >>>>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> SANDRO BONAZZOLA >>>>>>>> >>>>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>>>>> >>>>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>>>> <https://red.ht/sig> >>>>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> SANDRO BONAZZOLA >>>>>> >>>>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D >>>>>> >>>>>> Red Hat EMEA <https://www.redhat.com/> >>>>>> <https://red.ht/sig> >>>>>> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted> >>>>>> >>>>> >>>>> >>>> >>> >> > -- SANDRO BONAZZOLA ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R&D Red Hat EMEA <https://www.redhat.com/> <https://red.ht/sig> TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users