Hi Sahina, Yes the glusterd daemon was not running. I have started it and is able to add a glusterfs storage domain. Thank you so much for your help.
Oops! I allocated 50GiB for this storage domain and it requires 60GiB. On Wed, Jul 11, 2018 at 11:47 AM, Sahina Bose <sab...@redhat.com> wrote: > Is glusterd running on the server: goku.sanren.** > There's an error > Failed to get volume info: Command execution failed > error: Connection failed. Please check if gluster daemon is operational > > Please check the volume status using "gluster volume status engine" > > and if all looks ok, attach the mount logs from /var/log/glusterfs > > On Wed, Jul 11, 2018 at 1:57 PM, Sakhi Hadebe <sa...@sanren.ac.za> wrote: > >> Hi, >> >> I have managed to fix the error by enabling the DMA Virtualisation in >> BIOS. I am now hit with a new error: It's failing to add a glusterfs >> storage domain: >> >> [ INFO ] TASK [Add glusterfs storage domain] >> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is >> "[Problem while trying to mount target]". HTTP response code is 400. >> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault >> reason is \"Operation Failed\". Fault detail is \"[Problem while trying to >> mount target]\". HTTP response code is 400."} >> Please specify the storage you would like to use (glusterfs, >> iscsi, fc, nfs)[nfs]: >> >> Attached are vdsm and engine log files. >> >> >> >> >> >> On Wed, Jul 11, 2018 at 9:57 AM, Sakhi Hadebe <sa...@sanren.ac.za> wrote: >> >>> >>> >>> On Wed, Jul 11, 2018 at 9:33 AM, Sakhi Hadebe <sa...@sanren.ac.za> >>> wrote: >>> >>>> Hi, >>>> >>>> Below are the versions of packages installed. Please find the logs >>>> attached. >>>> Qemu: >>>> ipxe-roms-qemu-20170123-1.git4e85b27.el7_4.1.noarch >>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64 >>>> qemu-img-ev-2.10.0-21.el7_5.4.1.x86_64 >>>> qemu-kvm-ev-2.10.0-21.el7_5.4.1.x86_64 >>>> qemu-kvm-common-ev-2.10.0-21.el7_5.4.1.x86_64 >>>> >>>> Libvirt installed packages: >>>> libvirt-daemon-driver-storage-disk-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-config-nwfilter-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-iscsi-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-network-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-libs-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-secret-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-core-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-gluster-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-python-3.9.0-1.el7.x86_64 >>>> libvirt-daemon-driver-nodedev-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-rbd-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-scsi-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-config-network-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-client-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-kvm-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-logical-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-interface-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-lock-sanlock-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-storage-mpath-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-lxc-3.9.0-14.el7_5.6.x86_64 >>>> libvirt-daemon-driver-nwfilter-3.9.0-14.el7_5.6.x86_64 >>>> >>>> Virt-manager: >>>> virt-manager-common-1.4.3-3.el7.noarch >>>> >>>> oVirt: >>>> [root@localhost network-scripts]# rpm -qa | grep ovirt >>>> ovirt-setup-lib-1.1.4-1.el7.centos.noarch >>>> cockpit-ovirt-dashboard-0.11.28-1.el7.noarch >>>> ovirt-imageio-common-1.3.1.2-0.el7.centos.noarch >>>> ovirt-vmconsole-host-1.0.5-4.el7.centos.noarch >>>> ovirt-host-dependencies-4.2.3-1.el7.x86_64 >>>> ovirt-engine-sdk-python-3.6.9.1-1.el7.noarch >>>> ovirt-imageio-daemon-1.3.1.2-0.el7.centos.noarch >>>> ovirt-host-4.2.3-1.el7.x86_64 >>>> python-ovirt-engine-sdk4-4.2.7-2.el7.x86_64 >>>> ovirt-host-deploy-1.7.4-1.el7.noarch >>>> cockpit-machines-ovirt-169-1.el7.noarch >>>> ovirt-hosted-engine-ha-2.2.14-1.el7.noarch >>>> ovirt-vmconsole-1.0.5-4.el7.centos.noarch >>>> ovirt-provider-ovn-driver-1.2.11-1.el7.noarch >>>> ovirt-engine-appliance-4.2-20180626.1.el7.noarch >>>> ovirt-release42-4.2.4-1.el7.noarch >>>> ovirt-hosted-engine-setup-2.2.22.1-1.el7.noarch >>>> >>>> >>>> >>>> >>>> >>>> >>>> On Wed, Jul 11, 2018 at 6:48 AM, Yedidyah Bar David <d...@redhat.com> >>>> wrote: >>>> >>>>> On Tue, Jul 10, 2018 at 11:32 PM, Sakhi Hadebe <sa...@sanren.ac.za> >>>>> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> I did not select any CPU architecture. It doenst gove me the option >>>>>> to select one. It only states the number of virtual CPUs and the memory >>>>>> for >>>>>> the engine VM. >>>>>> >>>>>> Looking at the documentation of installing ovirt-release36.rpm....it >>>>>> does allow you to select te CPU, but not when installing >>>>>> ovirt-release42.rpm >>>>>> >>>>>> On Tuesday, July 10, 2018, Alastair Neil <ajneil.t...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> what did you select as your CPU architecture when you created the >>>>>>> cluster? It looks like the VM is trying to use a CPU type of "Custom", >>>>>>> how >>>>>>> many nodes in your cluster? I suggest you specify the lowest common >>>>>>> denominator of CPU architecture (e.g. Sandybridge) of the nodes as the >>>>>>> CPU >>>>>>> architecture of the cluster.. >>>>>>> >>>>>>> On Tue, 10 Jul 2018 at 12:01, Sakhi Hadebe <sa...@sanren.ac.za> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> I have just re-installed centOS 7 in 3 servers and have configured >>>>>>>> gluster volumes following this documentation: >>>>>>>> https://www.ovirt.org/blog/2016/03/up-and-running-with-ovirt-3-6/, >>>>>>>> But I have installed >>>>>>>> >>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release42.rpm >>>>>>>> >>>>>>>> package. >>>>>>>> Hosted-engine --deploy is failing with this error: >>>>>>>> >>>>>>>> "rhel7", "--virt-type", "kvm", "--memory", "16384", "--vcpus", >>>>>>>> "4", "--network", "network=default,mac=00:16:3e:09:5e:5d,model=virtio", >>>>>>>> "--disk", "/var/tmp/localvm0nnJH9/images >>>>>>>> /eacac30d-0304-4c77-8753-6965e4b8c2e7/d494577e-027a-4209-895b-6132e6fc6b9a", >>>>>>>> "--import", "--disk", >>>>>>>> "path=/var/tmp/localvm0nnJH9/seed.iso,device=cdrom", >>>>>>>> "--noautoconsole", "--rng", "/dev/random", "--graphics", "vnc", >>>>>>>> "--video", >>>>>>>> "vga", "--sound", "none", "--controller", "usb,model=none", >>>>>>>> "--memballoon", >>>>>>>> "none", "--boot", "hd,menu=off", "--clock", "kvmclock_present=yes"], >>>>>>>> "delta": "0:00:00.979003", "end": "2018-07-10 17:55:11.308555", "msg": >>>>>>>> "non-zero return code", "rc": 1, "start": "2018-07-10 17:55:10.329552", >>>>>>>> "stderr": "ERROR unsupported configuration: CPU mode 'custom' for >>>>>>>> x86_64 >>>>>>>> kvm domain on x86_64 host is not supported by hypervisor\nDomain >>>>>>>> installation does not appear to have been successful.\nIf it was, you >>>>>>>> can >>>>>>>> restart your domain by running:\n virsh --connect qemu:///system start >>>>>>>> HostedEngineLocal\notherwise, please restart your installation.", >>>>>>>> "stderr_lines": ["ERROR unsupported configuration: CPU mode >>>>>>>> 'custom' for >>>>>>>> x86_64 kvm domain on x86_64 host is not supported by hypervisor", >>>>>>>> "Domain >>>>>>>> installation does not appear to have been successful.", "If it was, >>>>>>>> you can >>>>>>>> restart your domain by running:", " virsh --connect qemu:///system >>>>>>>> start >>>>>>>> HostedEngineLocal", "otherwise, please restart your installation."], >>>>>>>> "stdout": "\nStarting install...", "stdout_lines": ["", "Starting >>>>>>>> install..."]} >>>>>>>> >>>>>>> >>>>> This seems to be in the phase where we create a local vm for the >>>>> engine. We do this with plain virt-install, nothing fancy. Searching the >>>>> net for "unsupported configuration: CPU mode 'custom'" finds other >>>>> relevant >>>>> reports, you might want to check them. You can see the command in >>>>> bootstrap_local_vm.yml . >>>>> >>>>> Please check/share versions of relevant packages (libvirt*, qemu*, >>>>> etc) and relevant logs (libvirt). >>>>> >>>>> Also updating the subject line and adding Simone. >>>>> >>>>> Best regards, >>>>> -- >>>>> Didi >>>>> >>>> >>>> >>>> >>>> -- >>>> Regards, >>>> Sakhi Hadebe >>>> >>>> Engineer: South African National Research Network (SANReN)Competency Area, >>>> Meraka, CSIR >>>> >>>> Tel: +27 12 841 2308 <+27128414213> >>>> Fax: +27 12 841 4223 <+27128414223> >>>> Cell: +27 71 331 9622 <+27823034657> >>>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>>> >>>> >>> >>> >>> -- >>> Regards, >>> Sakhi Hadebe >>> >>> Engineer: South African National Research Network (SANReN)Competency Area, >>> Meraka, CSIR >>> >>> Tel: +27 12 841 2308 <+27128414213> >>> Fax: +27 12 841 4223 <+27128414223> >>> Cell: +27 71 331 9622 <+27823034657> >>> Email: sa...@sanren.ac.za <shad...@csir.co.za> >>> >>> >> >> >> -- >> Regards, >> Sakhi Hadebe >> >> Engineer: South African National Research Network (SANReN)Competency Area, >> Meraka, CSIR >> >> Tel: +27 12 841 2308 <+27128414213> >> Fax: +27 12 841 4223 <+27128414223> >> Cell: +27 71 331 9622 <+27823034657> >> Email: sa...@sanren.ac.za <shad...@csir.co.za> >> >> >> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: https://www.ovirt.org/communit >> y/about/community-guidelines/ >> List Archives: https://lists.ovirt.org/archiv >> es/list/users@ovirt.org/message/YHKUKW22QLRVS56XZBXEWOGORFWFEGIA/ >> >> > -- Regards, Sakhi Hadebe Engineer: South African National Research Network (SANReN)Competency Area, Meraka, CSIR Tel: +27 12 841 2308 <+27128414213> Fax: +27 12 841 4223 <+27128414223> Cell: +27 71 331 9622 <+27823034657> Email: sa...@sanren.ac.za <shad...@csir.co.za>
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/XPYN7BIFCGY2LVVWOPEQCWQUHNN3JAJE/