---------- Forwarded message ---------
From: Leo David <leoa...@gmail.com>
Date: Tue, Apr 2, 2019, 15:10
Subject: Re: [ovirt-users] Re: HE - engine gluster volume - not mounted
To: Sahina Bose <sab...@redhat.com>


I have deleted everything in the engine gluster mount path, unmounted the
engine gluster volume ( not deleted the volume ) ,  and started the wizard
with "Use already configured storage". I have pointed to use this gluster
volume,  volume gets mounted under the correct path, but installation still
fails:

[ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[]".
HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault
reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code
is 400."}

On the node's vdsm.log I can continuously see:
2019-04-02 13:02:18,832+0100 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC
call Host.getStats succeeded in 0.03 seconds (__init__:312)
2019-04-02 13:02:19,906+0100 INFO  (vmrecovery) [vdsm.api] START
getConnectedStoragePoolsList(options=None) from=internal,
task_id=769c3983-5160-44e4-b1d8-7ab4e41ddd34 (api:48)
2019-04-02 13:02:19,907+0100 INFO  (vmrecovery) [vdsm.api] FINISH
getConnectedStoragePoolsList return={'poollist': []} from=internal,
task_id=769c3983-5160-44e4-b1d8-7ab4e41ddd34 (api:54)
2019-04-02 13:02:19,907+0100 INFO  (vmrecovery) [vds] recovery: waiting for
storage pool to go up (clientIF:709)
2019-04-02 13:02:21,737+0100 INFO  (periodic/2) [vdsm.api] START
repoStats(domains=()) from=internal,
task_id=ba12fbc1-0170-41a2-82e6-8ccb05ae9e09 (api:48)
2019-04-02 13:02:21,738+0100 INFO  (periodic/2) [vdsm.api] FINISH repoStats
return={} from=internal, task_id=ba12fbc1-0170-41a2-82e6-8ccb05ae9e09
(api:54)

Should I perform an "engine-cleanup",  delete lvms from Cockpit and start
it all over ?
Did anyone successfully used this particular iso image
"ovirt-node-ng-installer-4.3.2-2019031908.el7.iso" for a single node
installation ?
Thank you !
Leo


On Tue, Apr 2, 2019 at 1:45 PM Sahina Bose <sab...@redhat.com> wrote:

> Is it possible you have not cleared the gluster volume between installs?
>
> What's the corresponding error in vdsm.log?
>
>
> On Tue, Apr 2, 2019 at 4:07 PM Leo David <leoa...@gmail.com> wrote:
> >
> > And there it is the last lines on the ansible_create_storage_domain log:
> >
> > 2019-04-02 10:53:49,139+0100 DEBUG var changed: host "localhost" var
> "otopi_storage_domain_details" type "<type 'dict'>" value: "{
> >     "changed": false,
> >     "exception": "Traceback (most recent call last):\n  File
> \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 664,
> in main\n    storage_domains_module.post_create_check(sd_id)\n  File
> \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 526,
> in post_create_check\n    id=storage_domain.id,\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in
> add\n    return self._internal_add(storage_domain, headers, query, wait)\n
> File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232,
> in _internal_add\n    return future.wait() if wait else future\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in
> wait\n    return self._code(response)\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in
> callback\n    self._check_fault(response)\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in
> _check_fault\n    self._raise_error(response, body)\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in
> _raise_error\n    raise error\nError: Fault reason is \"Operation Failed\".
> Fault detail is \"[]\". HTTP response code is 400.\n",
> >     "failed": true,
> >     "msg": "Fault reason is \"Operation Failed\". Fault detail is
> \"[]\". HTTP response code is 400."
> > }"
> > 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var
> "ansible_play_hosts" type "<type 'list'>" value: "[]"
> > 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var
> "play_hosts" type "<type 'list'>" value: "[]"
> > 2019-04-02 10:53:49,142+0100 DEBUG var changed: host "localhost" var
> "ansible_play_batch" type "<type 'list'>" value: "[]"
> > 2019-04-02 10:53:49,142+0100 ERROR ansible failed {'status': 'FAILED',
> 'ansible_type': 'task', 'ansible_task': u'Activate storage domain',
> 'ansible_result': u'type: <type \'dict\'>\nstr: {\'_ansible_parsed\': True,
> u\'exception\': u\'Traceback (most recent call last):\\n  File
> "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 664,
> in main\\n    storage_domains_module.post_create_check(sd_id)\\n  File
> "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 526',
> 'task_duration': 9, 'ansible_host': u'localhost', 'ansible_playbook':
> u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml'}
> > 2019-04-02 10:53:49,143+0100 DEBUG ansible on_any args
> <ansible.executor.task_result.TaskResult object at 0x7f03fd025e50> kwargs
> ignore_errors:None
> > 2019-04-02 10:53:49,148+0100 INFO ansible stats {
> >     "ansible_playbook":
> "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
> >     "ansible_playbook_duration": "01:15 Minutes",
> >     "ansible_result": "type: <type 'dict'>\nstr: {u'localhost':
> {'unreachable': 0, 'skipped': 6, 'ok': 23, 'changed': 0, 'failures': 1}}",
> >     "ansible_type": "finish",
> >     "status": "FAILED"
> > }
> > 2019-04-02 10:53:49,149+0100 INFO SUMMARY:
> > Duration        Task Name
> > --------        --------
> > [ < 1 sec ]     Execute just a specific set of steps
> > [  00:02  ]     Force facts gathering
> > [  00:02  ]     Check local VM dir stat
> > [  00:02  ]     Obtain SSO token using username/password credentials
> > [  00:02  ]     Fetch host facts
> > [  00:01  ]     Fetch cluster ID
> > [  00:02  ]     Fetch cluster facts
> > [  00:02  ]     Fetch Datacenter facts
> > [  00:01  ]     Fetch Datacenter ID
> > [  00:01  ]     Fetch Datacenter name
> > [  00:02  ]     Add glusterfs storage domain
> > [  00:02  ]     Get storage domain details
> > [  00:02  ]     Find the appliance OVF
> > [  00:02  ]     Parse OVF
> > [  00:02  ]     Get required size
> > [ FAILED  ]     Activate storage domain
> >
> > Any ideea on how to escalate this issue ?
> > It just does not make sense to not be able to install from scratch a
> fresh node...
> >
> > Have a nice day  !
> >
> > Leo
> >
> >
> > On Tue, Apr 2, 2019 at 12:11 PM Gobinda Das <go...@redhat.com> wrote:
> >>
> >> Hi Leo,
> >>  Can you please paste "df -Th" and "gluster v status" out put ?
> >> Wanted to make sure engine mounted and volumes and bricks are up.
> >> What does vdsm log say?
> >>
> >> On Tue, Apr 2, 2019 at 2:06 PM Leo David <leoa...@gmail.com> wrote:
> >>>
> >>> Thank you very much !
> >>> I have just installed a new fresh node,   and triggered the single
> instance hyperconverged setup. It seems it fails at the hosted engine final
> steps of deployment:
> >>>  INFO ] TASK [ovirt.hosted_engine_setup : Get required size]
> >>> [ INFO ] ok: [localhost]
> >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove unsuitable storage
> domain]
> >>> [ INFO ] skipping: [localhost]
> >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Check storage domain free
> space]
> >>> [ INFO ] skipping: [localhost]
> >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain]
> >>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
> "[Cannot attach Storage. There is no active Host in the Data Center.]".
> HTTP response code is 409.
> >>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
> "Fault reason is \"Operation Failed\". Fault detail is \"[Cannot attach
> Storage. There is no active Host in the Data Center.]\". HTTP response code
> is 409."}
> >>> Also,  the
> ovirt-hosted-engine-setup-ansible-create_storage_domain-201932112413-xkq6nb.log
> throws the following:
> >>>
> >>> 2019-04-02 09:25:40,420+0100 DEBUG var changed: host "localhost" var
> "otopi_storage_domain_details" type "<type 'dict'>" value: "{
> >>>     "changed": false,
> >>>     "exception": "Traceback (most recent call last):\n  File
> \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 664,
> in main\n    storage_domains_module.post_create_check(sd_id)\n  File
> \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 526,
> in post_create_check\n    id=storage_domain.id,\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in
> add\n    return self._internal_add(storage_domain, headers, query, wait)\n
> File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232,
> in _internal_add\n    return future.wait() if wait else future\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in
> wait\n    return self._code(response)\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in
> callback\n    self._check_fault(response)\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in
> _check_fault\n    self._raise_error(response, body)\n  File
> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in
> _raise_error\n    raise error\nError: Fault reason is \"Operation Failed\".
> Fault detail is \"[Cannot attach Storage. There is no active Host in the
> Data Center.]\". HTTP response code is 409.\n",
> >>>     "failed": true,
> >>>     "msg": "Fault reason is \"Operation Failed\". Fault detail is
> \"[Cannot attach Storage. There is no active Host in the Data Center.]\".
> HTTP response code is 409."
> >>> }"
> >>>
> >>> I have used the ovirt-node-ng-installer-4.3.2-2019031908.el7.iso. So
> far,  I am unable to deploy oVirt single node Hyperconverged...
> >>> Any thoughts ?
> >>>
> >>>
> >>>
> >>> On Mon, Apr 1, 2019 at 9:46 PM Simone Tiraboschi <stira...@redhat.com>
> wrote:
> >>>>
> >>>>
> >>>>
> >>>> On Mon, Apr 1, 2019 at 6:14 PM Leo David <leoa...@gmail.com> wrote:
> >>>>>
> >>>>> Thank you Simone.
> >>>>> I've decides to go for a new fresh install from iso, and i'll keep
> posted if any troubles arise. But I am still trying to understand what are
> the services that mount the lvms and volumes after configuration. There is
> nothing related in fstab, so I assume there are a couple of .mount files
> somewhere in the filesystem.
> >>>>> Im just trying to understand node's underneath workflow.
> >>>>
> >>>>
> >>>> hosted-engine configuration is stored in
> /etc/ovirt-hosted-engine/hosted-engine.conf; ovirt-ha-broker will mount the
> hosted-engine storage domain according to that and so ovirt-ha-agent will
> be able to start the engine VM.
> >>>> Everything else is just in the engine DB.
> >>>>
> >>>>>
> >>>>>
> >>>>> On Mon, Apr 1, 2019, 10:16 Simone Tiraboschi <stira...@redhat.com>
> wrote:
> >>>>>>
> >>>>>> Hi,
> >>>>>> to understand what's failing I'd suggest to start attaching setup
> logs.
> >>>>>>
> >>>>>> On Sun, Mar 31, 2019 at 5:06 PM Leo David <leoa...@gmail.com>
> wrote:
> >>>>>>>
> >>>>>>> Hello Everyone,
> >>>>>>> Using 4.3.2 installation, and after running through HyperConverged
> Setup,  at the last stage it fails. It seems that the previously created
> "engine" volume is not mounted under "/rhev" path, therefore the setup
> cannot finish the deployment.
> >>>>>>> Any ideea which are the services responsible of mounting the
> volumes on oVirt Node distribution ? I'm thinking that maybe this
> particularly one failed to start for some reason...
> >>>>>>> Thank you very much !
> >>>>>>>
> >>>>>>> --
> >>>>>>> Best regards, Leo David
> >>>>>>> _______________________________________________
> >>>>>>> Users mailing list -- users@ovirt.org
> >>>>>>> To unsubscribe send an email to users-le...@ovirt.org
> >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >>>>>>> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >>>>>>> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/PUXDAQHVNZWF4TIXZ3GIBZHSJ7IC2VHC/
> >>>
> >>>
> >>>
> >>> --
> >>> Best regards, Leo David
> >>> _______________________________________________
> >>> Users mailing list -- users@ovirt.org
> >>> To unsubscribe send an email to users-le...@ovirt.org
> >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> >>> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> >>> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/NROCMKIFJDEIWD7PELNTBKDQMPMFLDRN/
> >>
> >>
> >>
> >> --
> >>
> >>
> >> Thanks,
> >> Gobinda
> >
> >
> >
> > --
> > Best regards, Leo David
> > _______________________________________________
> > Users mailing list -- users@ovirt.org
> > To unsubscribe send an email to users-le...@ovirt.org
> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> > oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/XDJNBS6EOXCMJKQCSSUHL4YRIYWRKDLS/
>


-- 
Best regards, Leo David
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AS4RTPBCZDGAVYMAAW7XCPC34X25L2JV/

Reply via email to