Il 12/07/2017 10:59, knarra ha scritto:
On 07/12/2017 01:43 PM, Simone Marchioni wrote:
Il 11/07/2017 11:23, knarra ha scritto:
Hi,
reply here to both Gianluca and Kasturi.
Gianluca: I had ovirt-4.1-dependencies.repo enabled, and gluster 3.8
packages, but glusterfs-server was missing in my "yum install"
command, so added glusterfs-server to my installation.
Kasturi: packages ovirt-hosted-engine-setup, gdeploy and
cockpit-ovirt-dashboard already installed and updated. vdsm-gluster
was missing, so added to my installation.
okay, cool.
:-)
Rerun deployment and IT WORKED! I can read the message "Succesfully
deployed Gluster" with the blue button "Continue to Hosted Engine
Deployment". There's a minor glitch in the window: the green "V" in
the circle is missing, like there's a missing image (or a wrong path,
as I had to remove "ansible" from the grafton-sanity-check.sh path...)
There is a bug for this and it will be fixed soon. Here is the bug id
for your reference. https://bugzilla.redhat.com/show_bug.cgi?id=1462082
Ok, thank you!
Although the deployment worked, and the firewalld and gluterfs errors
are gone, a couple of errors remains:
AFTER VG/LV CREATION, START/STOP/RELOAD/GLUSTER AND FIREWALLD HANDLING:
PLAY [gluster_servers]
*********************************************************
TASK [Run a shell script]
******************************************************
fatal: [ha1.domain.it]: FAILED! => {"failed": true, "msg": "The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
fatal: [ha2.domain.it]: FAILED! => {"failed": true, "msg": "The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
fatal: [ha3.domain.it]: FAILED! => {"failed": true, "msg": "The
conditional check 'result.rc != 0' failed. The error was: error while
evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
to retry, use: --limit @/tmp/tmpJnz4g3/run-script.retry
May be you missed to change the path of the script
"/usr/share/ansible/gdeploy/scripts/disable-gluster-hooks.sh" . That
is why this failure.
You're right: changed the path and now it's ok.
PLAY RECAP
*********************************************************************
ha1.domain.it : ok=0 changed=0 unreachable=0 failed=1
ha2.domain.it : ok=0 changed=0 unreachable=0 failed=1
ha3.domain.it : ok=0 changed=0 unreachable=0 failed=1
PLAY [gluster_servers]
*********************************************************
TASK [Run a command in the shell]
**********************************************
failed: [ha1.domain.it] (item=usermod -a -G gluster qemu) =>
{"changed": true, "cmd": "usermod -a -G gluster qemu", "delta":
"0:00:00.003144", "end": "2017-07-12 00:22:46.836832", "failed":
true, "item": "usermod -a -G gluster qemu", "rc": 6, "start":
"2017-07-12 00:22:46.833688", "stderr": "usermod: group 'gluster'
does not exist", "stderr_lines": ["usermod: group 'gluster' does not
exist"], "stdout": "", "stdout_lines": []}
failed: [ha2.domain.it] (item=usermod -a -G gluster qemu) =>
{"changed": true, "cmd": "usermod -a -G gluster qemu", "delta":
"0:00:00.003647", "end": "2017-07-12 00:22:46.895964", "failed":
true, "item": "usermod -a -G gluster qemu", "rc": 6, "start":
"2017-07-12 00:22:46.892317", "stderr": "usermod: group 'gluster'
does not exist", "stderr_lines": ["usermod: group 'gluster' does not
exist"], "stdout": "", "stdout_lines": []}
failed: [ha3.domain.it] (item=usermod -a -G gluster qemu) =>
{"changed": true, "cmd": "usermod -a -G gluster qemu", "delta":
"0:00:00.007008", "end": "2017-07-12 00:22:47.016600", "failed":
true, "item": "usermod -a -G gluster qemu", "rc": 6, "start":
"2017-07-12 00:22:47.009592", "stderr": "usermod: group 'gluster'
does not exist", "stderr_lines": ["usermod: group 'gluster' does not
exist"], "stdout": "", "stdout_lines": []}
to retry, use: --limit @/tmp/tmpJnz4g3/shell_cmd.retry
PLAY RECAP
*********************************************************************
ha1.domain.it : ok=0 changed=0 unreachable=0 failed=1
ha2.domain.it : ok=0 changed=0 unreachable=0 failed=1
ha3.domain.it : ok=0 changed=0 unreachable=0 failed=1
This error can be safely ignored.
Ok
These are a problem for my installation or can I ignore them?
You can just manually run the script to disable hooks on all the
nodes. Other error you can ignore.
Done it
By the way, I'm writing and documenting this process and can prepare
a tutorial if someone is interested.
Thank you again for your support: now I'll proceed with the Hosted
Engine Deployment.
Good to know that you can now start with Hosted Engine Deployment.
Started the Hosted Engine Deployment, but I have a different problem now.
As the installer asked, I specified some parameters, in particular a
pingable gateway address. Specified the host1 gateway.
Proceeding with the installer, it requires The Engine VM IP address
(DHCP or Static). I selected static and specified an IP Address, but the
IP *IS NOT* in the same subnet as the host1. The VMs IP addresses are
all on a different subnet.
The installer shows a red message:
The Engine VM (aa.bb.cc.dd/SM) and the default gateway (ww.xx.yy.zz)
will not be in the same IP subnet. Static routing configuration are not
supported on automatic VM configuration.
I'm starting to think that BOTH the hosts IPs and the VM IPs MUST BE ON
THE SAME SUBNET.
Is this a requirement or there's a way to deal with this configuration?
It's related only to "automatic VM configuration" or to oVirt in
general? Once installed oVirt Engine can I have VMs on different subnet?
Hi
Simone
_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users