On 3-6-2020 15:29, Gianluca Cecchi wrote: > On Mon, May 25, 2020 at 8:50 AM Yedidyah Bar David <d...@redhat.com > <mailto:d...@redhat.com>> wrote: > > On Sun, May 24, 2020 at 9:36 PM Gianluca Cecchi > <gianluca.cec...@gmail.com <mailto:gianluca.cec...@gmail.com>> wrote: > > > > On Sun, May 24, 2020 at 11:47 AM Yedidyah Bar David > <d...@redhat.com <mailto:d...@redhat.com>> wrote: > >> > >> > >> > >> Hi, Gianluca. Replying to your email on "4.4 HCI Install Failure - > >> Missing /etc/pki/CA/cacert.pem": > >> > >> On Sun, May 24, 2020 at 12:28 PM Gianluca Cecchi > >> <gianluca.cec...@gmail.com <mailto:gianluca.cec...@gmail.com>> > wrote: > >> > > >> > I I remember correctly it happened to me during the beta > cycle and the only "strange" character I used for the admin > password was the @ > >> > Donna if it related with what you reported for the % character > >> > >> Did you open a bug? > >> > >> In any case, my above patch is not supposed to fix '@', only > '%' (I think). > >> > >> Thanks and best regards, > >> > > > > No, I didn't open a bug, because I scratched the system and > installed again this time without the error, but I don't remember > if I used the same password with the @ character or not.... > > I will put attention in case of future 4.4 new installations > > Very well, thanks :-) > -- > Didi > > > > Just to avoid opening a bug for a different thing, today I tried a > single host HCI setup with the wizard and it failed. > Installed from ovirt-node-ng final 4.4 iso. > I see I have no /etc/pki/CA directory on the host at the moment, but I > don't know if the install workflow had not arrived there yet or what. > Last lines in my wizard window are these ones below. I have had this too and saw that during the test week is was also reported. The 'solution' was to run /usr/sbin/ovirt-hosted-engine-cleanup and let the wizard stay where it is and afterwards click Redeploy VM and then it will run to its conclusion. Don't know what precisely it fixes. Had a look and couldn't find it.
Joop > Password used contains only letters, numbers and the "_" character in > this attempt > I'm in the "Prepare VM" stage. > > [ INFO ] TASK [ovirt.hosted_engine_setup : Stop libvirt service] > [ INFO ] changed: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Drop vdsm config statements] > [ INFO ] changed: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Restore initial abrt config > files] > [ INFO ] changed: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Restart abrtd service] > [ INFO ] changed: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Drop libvirt sasl2 > configuration by vdsm] > [ INFO ] changed: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Stop and disable services] > [ INFO ] ok: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Restore initial libvirt > default network configuration] > [ INFO ] changed: [localhost] > [ INFO ] TASK [ovirt.hosted_engine_setup : Start libvirt] > [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": > "Unable to start service libvirtd: Job for libvirtd.service failed > because the control process exited with error code.\nSee \"systemctl > status libvirtd.service\" and \"journalctl -xe\" for details.\n"} > > Status of libvirtd service is this one: > > [root@ovirt01 g.cecchi]# systemctl status libvirtd -l --no-pager > ● libvirtd.service - Virtualization daemon > Loaded: loaded (/usr/lib/systemd/system/libvirtd.service; enabled; > vendor preset: enabled) > Drop-In: /etc/systemd/system/libvirtd.service.d > └─unlimited-core.conf > Active: failed (Result: exit-code) since Wed 2020-06-03 15:13:35 > CEST; 7min ago > Docs: man:libvirtd(8) > https://libvirt.org > Process: 20001 ExecStart=/usr/sbin/libvirtd $LIBVIRTD_ARGS > (code=exited, status=6) > Main PID: 20001 (code=exited, status=6) > Tasks: 2 (limit: 32768) > Memory: 70.1M > CGroup: /system.slice/libvirtd.service > ├─3926 /usr/sbin/dnsmasq > --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro > --dhcp-script=/usr/libexec/libvirt_leaseshelper > └─3927 /usr/sbin/dnsmasq > --conf-file=/var/lib/libvirt/dnsmasq/default.conf --leasefile-ro > --dhcp-script=/usr/libexec/libvirt_leaseshelper > > Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: > Service RestartSec=100ms expired, scheduling restart. > Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: > Scheduled restart job, restart counter is at 5. > Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: Stopped > Virtualization daemon. > Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: > Start request repeated too quickly. > Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: libvirtd.service: > Failed with result 'exit-code'. > Jun 03 15:13:35 ovirt01.mydomain.local systemd[1]: Failed to start > Virtualization daemon. > [root@ovirt01 g.cecchi]# > > Let me know what files do you want to analyze the problem > Under /var/log/ovirt-hosted-engine-setup I have: > > [root@ovirt01 ovirt-hosted-engine-setup]# ls -lrt > total 632 > -rw-r--r--. 1 root root 123814 Jun 3 15:07 > ovirt-hosted-engine-setup-ansible-get_network_interfaces-20205315737-ooohyb.log > -rw-r--r--. 1 root root 126674 Jun 3 15:08 > ovirt-hosted-engine-setup-ansible-validate_hostnames-20205315737-oqixuw.log > -rw-r--r--. 1 root root 127548 Jun 3 15:10 > ovirt-hosted-engine-setup-ansible-validate_hostnames-202053151022-yls4qo.log > -rw-r--r--. 1 root root 261482 Jun 3 15:13 > ovirt-hosted-engine-setup-ansible-initial_clean-20205315123-7x25zv.log > [root@ovirt01 ovirt-hosted-engine-setup]# > > In this environment no dns but entry in /etc/hosts of the server. > Host is on 192.168.1.x on eno1 and a vlan 100 on the same interface, > used for "simulated" storage network > > 2: eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel > state UP group default qlen 1000 > link/ether b8:ae:ed:7f:17:11 brd ff:ff:ff:ff:ff:ff > inet 192.168.1.211/24 <http://192.168.1.211/24> brd 192.168.1.255 > scope global noprefixroute eno1 > valid_lft forever preferred_lft forever > inet6 fe80::315c:2e9f:d009:8c0c/64 scope link noprefixroute > valid_lft forever preferred_lft forever > > 7: eno1.100@eno1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > noqueue state UP group default qlen 1000 > link/ether b8:ae:ed:7f:17:11 brd ff:ff:ff:ff:ff:ff > inet 192.168.100.211/24 <http://192.168.100.211/24> brd > 192.168.100.255 scope global noprefixroute eno1.100 > valid_lft forever preferred_lft forever > inet6 fe80::be0f:c4be:cf7a:9da5/64 scope link noprefixroute > valid_lft forever preferred_lft forever > > Thanks, > Gianluca > > > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/IPDND2OXUV2DXLAJ24KRA4AXDFA532C2/
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/YLC2XIFQYVDJCWITH6IGQNIWSJOS6IPK/