On 3-6-2020 15:29, Gianluca Cecchi wrote:
> On Mon, May 25, 2020 at 8:50 AM Yedidyah Bar David <d...@redhat.com
> <mailto:d...@redhat.com>> wrote:
>
>     On Sun, May 24, 2020 at 9:36 PM Gianluca Cecchi
>     <gianluca.cec...@gmail.com <mailto:gianluca.cec...@gmail.com>> wrote:
>     >
>     > On Sun, May 24, 2020 at 11:47 AM Yedidyah Bar David
>     <d...@redhat.com <mailto:d...@redhat.com>> wrote:
>     >>
>     >>
>     >>
>     >> Hi, Gianluca. Replying to your email on "4.4 HCI Install Failure -
>     >> Missing /etc/pki/CA/cacert.pem":
>     >>
>     >> On Sun, May 24, 2020 at 12:28 PM Gianluca Cecchi
>     >> <gianluca.cec...@gmail.com <mailto:gianluca.cec...@gmail.com>>
>     wrote:
>     >> >
>     >> > I I remember correctly it happened to me during the beta
>     cycle and the only "strange" character I used for the admin
>     password was the @
>     >> > Donna if it related with what you reported for the % character
>     >>
>     >> Did you open a bug?
>     >>
>     >> In any case, my above patch is not supposed to fix '@', only
>     '%' (I think).
>     >>
>     >> Thanks and best regards,
>     >>
>     >
>     > No, I didn't open a bug, because I scratched the system and
>     installed again this time without the error, but I don't remember
>     if I used the same password with the @ character or not....
>     > I will put attention in case of future 4.4 new installations
>
>     Very well, thanks :-)
>     -- 
>     Didi
>
>
>
> Just to avoid opening a bug for a different thing, today I tried a
> single host HCI setup with the wizard and it failed.
> Installed from ovirt-node-ng final 4.4 iso.
> I see I have no /etc/pki/CA directory on the host at the moment, but I
> don't know if the install workflow had not arrived there yet or what.
> Last lines in my wizard window are these ones below.
I have had this too and saw that during the test week is was also reported.
The 'solution' was to run /usr/sbin/ovirt-hosted-engine-cleanup and let
the wizard stay where it is and afterwards click Redeploy VM and then it
will run to its conclusion. Don't know what precisely it fixes. Had a
look and couldn't find it.

Joop


> Password used contains only letters, numbers and the "_" character in
> this attempt
> I'm in the "Prepare VM" stage.
>
> [ INFO ] TASK [ovirt.hosted_engine_setup : Stop libvirt service]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Drop vdsm config statements]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Restore initial abrt config
> files]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Restart abrtd service]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Drop libvirt sasl2
> configuration by vdsm]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Stop and disable services]
> [ INFO ] ok: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Restore initial libvirt
> default network configuration]
> [ INFO ] changed: [localhost]
> [ INFO ] TASK [ovirt.hosted_engine_setup : Start libvirt]
> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
> "Unable to start service libvirtd: Job for libvirtd.service failed
> because the control process exited with error code.\nSee \"systemctl
> status libvirtd.service\" and \"journalctl -xe\" for details.\n"}
>
> Status of libvirtd service is this one:
>
> [root@ovirt01 g.cecchi]# systemctl status libvirtd -l --no-pager
> ● libvirtd.service - Virtualization daemon
>    Loaded: loaded (/usr/lib/systemd/system/libvirtd.service; enabled;
> vendor preset: enabled)
>   Drop-In: /etc/systemd/system/libvirtd.service.d
>            └─unlimited-core.conf
>    Active: failed (Result: exit-code) since Wed 2020-06-03 15:13:35
> CEST; 7min ago
>      Docs: man:libvirtd(8)
>            https://libvirt.org
>   Process: 20001 ExecStart=/usr/sbin/libvirtd $LIBVIRTD_ARGS
> (code=exited, status=6)
>  Main PID: 20001 (code=exited, status=6)
>     Tasks: 2 (limit: 32768)
>    Memory: 70.1M
>    CGroup: /system.slice/libvirtd.service
>            ├─3926 /usr/sbin/dnsmasq
> --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro
> --dhcp-script=/usr/libexec/libvirt_leaseshelper
>            └─3927 /usr/sbin/dnsmasq
> --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro
> --dhcp-script=/usr/libexec/libvirt_leaseshelper
>
> Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service:
> Service RestartSec=100ms expired, scheduling restart.
> Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service:
> Scheduled restart job, restart counter is at 5.
> Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: Stopped
> Virtualization daemon.
> Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service:
> Start request repeated too quickly.
> Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service:
> Failed with result 'exit-code'.
> Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: Failed to start
> Virtualization daemon.
> [root@ovirt01 g.cecchi]#
>
> Let me know what files do you want to analyze the problem
> Under /var/log/ovirt-hosted-engine-setup I have:
>
> [root@ovirt01 ovirt-hosted-engine-setup]# ls -lrt
> total 632
> -rw-r--r--. 1 root root 123814 Jun  3 15:07
> ovirt-hosted-engine-setup-ansible-get_network_interfaces-20205315737-ooohyb.log
> -rw-r--r--. 1 root root 126674 Jun  3 15:08
> ovirt-hosted-engine-setup-ansible-validate_hostnames-20205315737-oqixuw.log
> -rw-r--r--. 1 root root 127548 Jun  3 15:10
> ovirt-hosted-engine-setup-ansible-validate_hostnames-202053151022-yls4qo.log
> -rw-r--r--. 1 root root 261482 Jun  3 15:13
> ovirt-hosted-engine-setup-ansible-initial_clean-20205315123-7x25zv.log
> [root@ovirt01 ovirt-hosted-engine-setup]#
>
> In this environment no dns but entry in /etc/hosts of the server.
> Host is on 192.168.1.x on eno1 and a vlan 100 on the same interface,
> used for "simulated" storage network
>
> 2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel
> state UP group default qlen 1000
>     link/ether b8:ae:ed:7f:17:11 brd ff:ff:ff:ff:ff:ff
>     inet 192.168.1.211/24 <http://192.168.1.211/24> brd 192.168.1.255
> scope global noprefixroute eno1
>        valid_lft forever preferred_lft forever
>     inet6 fe80::315c:2e9f:d009:8c0c/64 scope link noprefixroute
>        valid_lft forever preferred_lft forever
>
> 7: eno1.100@eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc
> noqueue state UP group default qlen 1000
>     link/ether b8:ae:ed:7f:17:11 brd ff:ff:ff:ff:ff:ff
>     inet 192.168.100.211/24 <http://192.168.100.211/24> brd
> 192.168.100.255 scope global noprefixroute eno1.100
>        valid_lft forever preferred_lft forever
>     inet6 fe80::be0f:c4be:cf7a:9da5/64 scope link noprefixroute
>        valid_lft forever preferred_lft forever
>
> Thanks,
> Gianluca
>
>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/IPDND2OXUV2DXLAJ24KRA4AXDFA532C2/

_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YLC2XIFQYVDJCWITH6IGQNIWSJOS6IPK/

Reply via email to