Thank you all for your help, I have managed to deploy the engine
successfully. It was a quiet a lesson.

On Wed, Jul 11, 2018 at 11:55 AM, Sakhi Hadebe <[email protected]> wrote:

> ​Hi Sahina,
>
> Yes the glusterd daemon was not running. I have started it and is able to
> add a glusterfs storage domain. Thank you so much for your help.
>
> Oops! I allocated 50GiB for this storage domain and it requires 60GiB.
>
> On Wed, Jul 11, 2018 at 11:47 AM, Sahina Bose <[email protected]> wrote:
>
>> Is glusterd running on the server: goku.sanren.**
>> There's an error
>> Failed to get volume info: Command execution failed
>> error: Connection failed. Please check if gluster daemon is operational
>>
>> Please check the volume status using "gluster volume status engine"
>>
>> and if all looks ok, attach the mount logs from /var/log/glusterfs
>>
>> On Wed, Jul 11, 2018 at 1:57 PM, Sakhi Hadebe <[email protected]> wrote:
>>
>>> Hi,
>>>
>>> I have managed to fix the error by enabling the DMA Virtualisation in
>>> BIOS. I am now hit with a new error: It's failing to add a glusterfs
>>> storage domain:
>>>
>>> [ INFO  ] TASK [Add glusterfs storage domain]
>>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
>>> "[Problem while trying to mount target]". HTTP response code is 400.
>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
>>> "Fault reason is \"Operation Failed\". Fault detail is \"[Problem while
>>> trying to mount target]\". HTTP response code is 400."}
>>>           Please specify the storage you would like to use (glusterfs,
>>> iscsi, fc, nfs)[nfs]:
>>>
>>> Attached are vdsm and engine log files.
>>>
>>>
>>>
>>>
>>>
>>> On Wed, Jul 11, 2018 at 9:57 AM, Sakhi Hadebe <[email protected]>
>>> wrote:
>>>
>>>>
>>>>
>>>> On Wed, Jul 11, 2018 at 9:33 AM, Sakhi Hadebe <[email protected]>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Below are the versions of packages installed. Please find the logs
>>>>> attached.
>>>>> Qemu:
>>>>> ipxe-roms-qemu-20170123-1.git4e85b27.el7_4.1.noarch
>>>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64
>>>>> qemu-img-ev-2.10.0-21.el7_5.4.1.x86_64
>>>>> qemu-kvm-ev-2.10.0-21.el7_5.4.1.x86_64
>>>>> qemu-kvm-common-ev-2.10.0-21.el7_5.4.1.x86_64
>>>>>
>>>>> Libvirt installed packages:
>>>>> libvirt-daemon-driver-storage-disk-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-config-nwfilter-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-iscsi-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-network-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-libs-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-secret-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-core-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-gluster-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-python-3.9.0-1.el7.x86_64
>>>>> libvirt-daemon-driver-nodedev-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-rbd-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-scsi-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-config-network-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-client-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-kvm-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-logical-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-interface-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-lock-sanlock-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-storage-mpath-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-lxc-3.9.0-14.el7_5.6.x86_64
>>>>> libvirt-daemon-driver-nwfilter-3.9.0-14.el7_5.6.x86_64
>>>>>
>>>>> Virt-manager:
>>>>> virt-manager-common-1.4.3-3.el7.noarch
>>>>>
>>>>> oVirt:
>>>>> [root@localhost network-scripts]# rpm -qa | grep ovirt
>>>>> ovirt-setup-lib-1.1.4-1.el7.centos.noarch
>>>>> cockpit-ovirt-dashboard-0.11.28-1.el7.noarch
>>>>> ovirt-imageio-common-1.3.1.2-0.el7.centos.noarch
>>>>> ovirt-vmconsole-host-1.0.5-4.el7.centos.noarch
>>>>> ovirt-host-dependencies-4.2.3-1.el7.x86_64
>>>>> ovirt-engine-sdk-python-3.6.9.1-1.el7.noarch
>>>>> ovirt-imageio-daemon-1.3.1.2-0.el7.centos.noarch
>>>>> ovirt-host-4.2.3-1.el7.x86_64
>>>>> python-ovirt-engine-sdk4-4.2.7-2.el7.x86_64
>>>>> ovirt-host-deploy-1.7.4-1.el7.noarch
>>>>> cockpit-machines-ovirt-169-1.el7.noarch
>>>>> ovirt-hosted-engine-ha-2.2.14-1.el7.noarch
>>>>> ovirt-vmconsole-1.0.5-4.el7.centos.noarch
>>>>> ovirt-provider-ovn-driver-1.2.11-1.el7.noarch
>>>>> ovirt-engine-appliance-4.2-20180626.1.el7.noarch
>>>>> ovirt-release42-4.2.4-1.el7.noarch
>>>>> ovirt-hosted-engine-setup-2.2.22.1-1.el7.noarch
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Jul 11, 2018 at 6:48 AM, Yedidyah Bar David <[email protected]>
>>>>> wrote:
>>>>>
>>>>>> On Tue, Jul 10, 2018 at 11:32 PM, Sakhi Hadebe <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> I did not select any CPU architecture. It doenst gove me the option
>>>>>>> to select one. It only states the number of virtual CPUs and the memory 
>>>>>>> for
>>>>>>> the engine VM.
>>>>>>>
>>>>>>> Looking at the documentation of installing ovirt-release36.rpm....it
>>>>>>> does allow you to select te CPU, but not when installing 
>>>>>>> ovirt-release42.rpm
>>>>>>>
>>>>>>> On Tuesday, July 10, 2018, Alastair Neil <[email protected]>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> what did you select as your CPU architecture when you created the
>>>>>>>> cluster?  It looks like the VM is trying to use a CPU type of 
>>>>>>>> "Custom", how
>>>>>>>> many nodes in your cluster?  I suggest you specify the lowest common
>>>>>>>> denominator of CPU architecture (e.g. Sandybridge) of the nodes as the 
>>>>>>>> CPU
>>>>>>>> architecture of the cluster..
>>>>>>>>
>>>>>>>> On Tue, 10 Jul 2018 at 12:01, Sakhi Hadebe <[email protected]>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> I have just re-installed centOS 7 in 3 servers and have configured
>>>>>>>>> gluster volumes following this documentation:
>>>>>>>>> https://www.ovirt.org/blog/2016/03/up-and-running-with-ovirt-3-6/,
>>>>>>>>> But I have installed
>>>>>>>>>
>>>>>>>>> http://resources.ovirt.org/pub/yum-repo/ovirt-release42.rpm
>>>>>>>>>
>>>>>>>>> ​package.
>>>>>>>>> Hosted-engine --deploy is failing with this error:
>>>>>>>>>
>>>>>>>>>  "rhel7", "--virt-type", "kvm", "--memory", "16384", "--vcpus",
>>>>>>>>> "4", "--network", 
>>>>>>>>> "network=default,mac=00:16:3e:09:5e:5d,model=virtio",
>>>>>>>>> "--disk", "/var/tmp/localvm0nnJH9/images
>>>>>>>>> /eacac30d-0304-4c77-8753-6965e4b8c2e7/d494577e-027a-4209-895b-6132e6fc6b9a",
>>>>>>>>> "--import", "--disk", 
>>>>>>>>> "path=/var/tmp/localvm0nnJH9/seed.iso,device=cdrom",
>>>>>>>>> "--noautoconsole", "--rng", "/dev/random", "--graphics", "vnc", 
>>>>>>>>> "--video",
>>>>>>>>> "vga", "--sound", "none", "--controller", "usb,model=none", 
>>>>>>>>> "--memballoon",
>>>>>>>>> "none", "--boot", "hd,menu=off", "--clock", "kvmclock_present=yes"],
>>>>>>>>> "delta": "0:00:00.979003", "end": "2018-07-10 17:55:11.308555", "msg":
>>>>>>>>> "non-zero return code", "rc": 1, "start": "2018-07-10 
>>>>>>>>> 17:55:10.329552",
>>>>>>>>> "stderr": "ERROR    unsupported configuration: CPU mode 'custom' for 
>>>>>>>>> x86_64
>>>>>>>>> kvm domain on x86_64 host is not supported by hypervisor\nDomain
>>>>>>>>> installation does not appear to have been successful.\nIf it was, you 
>>>>>>>>> can
>>>>>>>>> restart your domain by running:\n  virsh --connect qemu:///system 
>>>>>>>>> start
>>>>>>>>> HostedEngineLocal\notherwise, please restart your installation.",
>>>>>>>>> "stderr_lines": ["ERROR    unsupported configuration: CPU mode 
>>>>>>>>> 'custom' for
>>>>>>>>> x86_64 kvm domain on x86_64 host is not supported by hypervisor", 
>>>>>>>>> "Domain
>>>>>>>>> installation does not appear to have been successful.", "If it was, 
>>>>>>>>> you can
>>>>>>>>> restart your domain by running:", "  virsh --connect qemu:///system 
>>>>>>>>> start
>>>>>>>>> HostedEngineLocal", "otherwise, please restart your installation."],
>>>>>>>>> "stdout": "\nStarting install...", "stdout_lines": ["", "Starting
>>>>>>>>> install..."]}
>>>>>>>>>
>>>>>>>>
>>>>>> This seems to be in the phase where we create a local vm for the
>>>>>> engine. We do this with plain virt-install, nothing fancy. Searching the
>>>>>> net for "unsupported configuration: CPU mode 'custom'" finds other 
>>>>>> relevant
>>>>>> reports, you might want to check them. You can see the command in
>>>>>> bootstrap_local_vm.yml .
>>>>>>
>>>>>> Please check/share versions of relevant packages (libvirt*, qemu*,
>>>>>> etc) and relevant logs (libvirt).
>>>>>>
>>>>>> Also updating the subject line and adding Simone.
>>>>>>
>>>>>> Best regards,
>>>>>> --
>>>>>> Didi
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Regards,
>>>>> Sakhi Hadebe
>>>>>
>>>>> Engineer: South African National Research Network (SANReN)Competency 
>>>>> Area, Meraka, CSIR
>>>>>
>>>>> Tel:   +27 12 841 2308 <+27128414213>
>>>>> Fax:   +27 12 841 4223 <+27128414223>
>>>>> Cell:  +27 71 331 9622 <+27823034657>
>>>>> Email: [email protected] <[email protected]>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Regards,
>>>> Sakhi Hadebe
>>>>
>>>> Engineer: South African National Research Network (SANReN)Competency Area, 
>>>> Meraka, CSIR
>>>>
>>>> Tel:   +27 12 841 2308 <+27128414213>
>>>> Fax:   +27 12 841 4223 <+27128414223>
>>>> Cell:  +27 71 331 9622 <+27823034657>
>>>> Email: [email protected] <[email protected]>
>>>>
>>>>
>>>
>>>
>>> --
>>> Regards,
>>> Sakhi Hadebe
>>>
>>> Engineer: South African National Research Network (SANReN)Competency Area, 
>>> Meraka, CSIR
>>>
>>> Tel:   +27 12 841 2308 <+27128414213>
>>> Fax:   +27 12 841 4223 <+27128414223>
>>> Cell:  +27 71 331 9622 <+27823034657>
>>> Email: [email protected] <[email protected]>
>>>
>>>
>>> _______________________________________________
>>> Users mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct: https://www.ovirt.org/communit
>>> y/about/community-guidelines/
>>> List Archives: https://lists.ovirt.org/archiv
>>> es/list/[email protected]/message/YHKUKW22QLRVS56XZBXEWOGORFWFEGIA/
>>>
>>>
>>
>
>
> --
> Regards,
> Sakhi Hadebe
>
> Engineer: South African National Research Network (SANReN)Competency Area, 
> Meraka, CSIR
>
> Tel:   +27 12 841 2308 <+27128414213>
> Fax:   +27 12 841 4223 <+27128414223>
> Cell:  +27 71 331 9622 <+27823034657>
> Email: [email protected] <[email protected]>
>
>


-- 
Regards,
Sakhi Hadebe

Engineer: South African National Research Network (SANReN)Competency
Area, Meraka, CSIR

Tel:   +27 12 841 2308 <+27128414213>
Fax:   +27 12 841 4223 <+27128414223>
Cell:  +27 71 331 9622 <+27823034657>
Email: [email protected] <[email protected]>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/3NDRNYG5JMSJBZE33SOA2MORTTE6IXHN/

Reply via email to