---------- Forwarded message --------- From: Leo David <leoa...@gmail.com> Date: Tue, Apr 2, 2019, 15:10 Subject: Re: [ovirt-users] Re: HE - engine gluster volume - not mounted To: Sahina Bose <sab...@redhat.com>
I have deleted everything in the engine gluster mount path, unmounted the engine gluster volume ( not deleted the volume ) , and started the wizard with "Use already configured storage". I have pointed to use this gluster volume, volume gets mounted under the correct path, but installation still fails: [ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain] [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[]". HTTP response code is 400. [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 400."} On the node's vdsm.log I can continuously see: 2019-04-02 13:02:18,832+0100 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC call Host.getStats succeeded in 0.03 seconds (__init__:312) 2019-04-02 13:02:19,906+0100 INFO (vmrecovery) [vdsm.api] START getConnectedStoragePoolsList(options=None) from=internal, task_id=769c3983-5160-44e4-b1d8-7ab4e41ddd34 (api:48) 2019-04-02 13:02:19,907+0100 INFO (vmrecovery) [vdsm.api] FINISH getConnectedStoragePoolsList return={'poollist': []} from=internal, task_id=769c3983-5160-44e4-b1d8-7ab4e41ddd34 (api:54) 2019-04-02 13:02:19,907+0100 INFO (vmrecovery) [vds] recovery: waiting for storage pool to go up (clientIF:709) 2019-04-02 13:02:21,737+0100 INFO (periodic/2) [vdsm.api] START repoStats(domains=()) from=internal, task_id=ba12fbc1-0170-41a2-82e6-8ccb05ae9e09 (api:48) 2019-04-02 13:02:21,738+0100 INFO (periodic/2) [vdsm.api] FINISH repoStats return={} from=internal, task_id=ba12fbc1-0170-41a2-82e6-8ccb05ae9e09 (api:54) Should I perform an "engine-cleanup", delete lvms from Cockpit and start it all over ? Did anyone successfully used this particular iso image "ovirt-node-ng-installer-4.3.2-2019031908.el7.iso" for a single node installation ? Thank you ! Leo On Tue, Apr 2, 2019 at 1:45 PM Sahina Bose <sab...@redhat.com> wrote: > Is it possible you have not cleared the gluster volume between installs? > > What's the corresponding error in vdsm.log? > > > On Tue, Apr 2, 2019 at 4:07 PM Leo David <leoa...@gmail.com> wrote: > > > > And there it is the last lines on the ansible_create_storage_domain log: > > > > 2019-04-02 10:53:49,139+0100 DEBUG var changed: host "localhost" var > "otopi_storage_domain_details" type "<type 'dict'>" value: "{ > > "changed": false, > > "exception": "Traceback (most recent call last):\n File > \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 664, > in main\n storage_domains_module.post_create_check(sd_id)\n File > \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 526, > in post_create_check\n id=storage_domain.id,\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in > add\n return self._internal_add(storage_domain, headers, query, wait)\n > File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, > in _internal_add\n return future.wait() if wait else future\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in > wait\n return self._code(response)\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in > callback\n self._check_fault(response)\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in > _check_fault\n self._raise_error(response, body)\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in > _raise_error\n raise error\nError: Fault reason is \"Operation Failed\". > Fault detail is \"[]\". HTTP response code is 400.\n", > > "failed": true, > > "msg": "Fault reason is \"Operation Failed\". Fault detail is > \"[]\". HTTP response code is 400." > > }" > > 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var > "ansible_play_hosts" type "<type 'list'>" value: "[]" > > 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var > "play_hosts" type "<type 'list'>" value: "[]" > > 2019-04-02 10:53:49,142+0100 DEBUG var changed: host "localhost" var > "ansible_play_batch" type "<type 'list'>" value: "[]" > > 2019-04-02 10:53:49,142+0100 ERROR ansible failed {'status': 'FAILED', > 'ansible_type': 'task', 'ansible_task': u'Activate storage domain', > 'ansible_result': u'type: <type \'dict\'>\nstr: {\'_ansible_parsed\': True, > u\'exception\': u\'Traceback (most recent call last):\\n File > "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 664, > in main\\n storage_domains_module.post_create_check(sd_id)\\n File > "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 526', > 'task_duration': 9, 'ansible_host': u'localhost', 'ansible_playbook': > u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml'} > > 2019-04-02 10:53:49,143+0100 DEBUG ansible on_any args > <ansible.executor.task_result.TaskResult object at 0x7f03fd025e50> kwargs > ignore_errors:None > > 2019-04-02 10:53:49,148+0100 INFO ansible stats { > > "ansible_playbook": > "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", > > "ansible_playbook_duration": "01:15 Minutes", > > "ansible_result": "type: <type 'dict'>\nstr: {u'localhost': > {'unreachable': 0, 'skipped': 6, 'ok': 23, 'changed': 0, 'failures': 1}}", > > "ansible_type": "finish", > > "status": "FAILED" > > } > > 2019-04-02 10:53:49,149+0100 INFO SUMMARY: > > Duration Task Name > > -------- -------- > > [ < 1 sec ] Execute just a specific set of steps > > [ 00:02 ] Force facts gathering > > [ 00:02 ] Check local VM dir stat > > [ 00:02 ] Obtain SSO token using username/password credentials > > [ 00:02 ] Fetch host facts > > [ 00:01 ] Fetch cluster ID > > [ 00:02 ] Fetch cluster facts > > [ 00:02 ] Fetch Datacenter facts > > [ 00:01 ] Fetch Datacenter ID > > [ 00:01 ] Fetch Datacenter name > > [ 00:02 ] Add glusterfs storage domain > > [ 00:02 ] Get storage domain details > > [ 00:02 ] Find the appliance OVF > > [ 00:02 ] Parse OVF > > [ 00:02 ] Get required size > > [ FAILED ] Activate storage domain > > > > Any ideea on how to escalate this issue ? > > It just does not make sense to not be able to install from scratch a > fresh node... > > > > Have a nice day ! > > > > Leo > > > > > > On Tue, Apr 2, 2019 at 12:11 PM Gobinda Das <go...@redhat.com> wrote: > >> > >> Hi Leo, > >> Can you please paste "df -Th" and "gluster v status" out put ? > >> Wanted to make sure engine mounted and volumes and bricks are up. > >> What does vdsm log say? > >> > >> On Tue, Apr 2, 2019 at 2:06 PM Leo David <leoa...@gmail.com> wrote: > >>> > >>> Thank you very much ! > >>> I have just installed a new fresh node, and triggered the single > instance hyperconverged setup. It seems it fails at the hosted engine final > steps of deployment: > >>> INFO ] TASK [ovirt.hosted_engine_setup : Get required size] > >>> [ INFO ] ok: [localhost] > >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove unsuitable storage > domain] > >>> [ INFO ] skipping: [localhost] > >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Check storage domain free > space] > >>> [ INFO ] skipping: [localhost] > >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain] > >>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is > "[Cannot attach Storage. There is no active Host in the Data Center.]". > HTTP response code is 409. > >>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": > "Fault reason is \"Operation Failed\". Fault detail is \"[Cannot attach > Storage. There is no active Host in the Data Center.]\". HTTP response code > is 409."} > >>> Also, the > ovirt-hosted-engine-setup-ansible-create_storage_domain-201932112413-xkq6nb.log > throws the following: > >>> > >>> 2019-04-02 09:25:40,420+0100 DEBUG var changed: host "localhost" var > "otopi_storage_domain_details" type "<type 'dict'>" value: "{ > >>> "changed": false, > >>> "exception": "Traceback (most recent call last):\n File > \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 664, > in main\n storage_domains_module.post_create_check(sd_id)\n File > \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 526, > in post_create_check\n id=storage_domain.id,\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, in > add\n return self._internal_add(storage_domain, headers, query, wait)\n > File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, > in _internal_add\n return future.wait() if wait else future\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in > wait\n return self._code(response)\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in > callback\n self._check_fault(response)\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in > _check_fault\n self._raise_error(response, body)\n File > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in > _raise_error\n raise error\nError: Fault reason is \"Operation Failed\". > Fault detail is \"[Cannot attach Storage. There is no active Host in the > Data Center.]\". HTTP response code is 409.\n", > >>> "failed": true, > >>> "msg": "Fault reason is \"Operation Failed\". Fault detail is > \"[Cannot attach Storage. There is no active Host in the Data Center.]\". > HTTP response code is 409." > >>> }" > >>> > >>> I have used the ovirt-node-ng-installer-4.3.2-2019031908.el7.iso. So > far, I am unable to deploy oVirt single node Hyperconverged... > >>> Any thoughts ? > >>> > >>> > >>> > >>> On Mon, Apr 1, 2019 at 9:46 PM Simone Tiraboschi <stira...@redhat.com> > wrote: > >>>> > >>>> > >>>> > >>>> On Mon, Apr 1, 2019 at 6:14 PM Leo David <leoa...@gmail.com> wrote: > >>>>> > >>>>> Thank you Simone. > >>>>> I've decides to go for a new fresh install from iso, and i'll keep > posted if any troubles arise. But I am still trying to understand what are > the services that mount the lvms and volumes after configuration. There is > nothing related in fstab, so I assume there are a couple of .mount files > somewhere in the filesystem. > >>>>> Im just trying to understand node's underneath workflow. > >>>> > >>>> > >>>> hosted-engine configuration is stored in > /etc/ovirt-hosted-engine/hosted-engine.conf; ovirt-ha-broker will mount the > hosted-engine storage domain according to that and so ovirt-ha-agent will > be able to start the engine VM. > >>>> Everything else is just in the engine DB. > >>>> > >>>>> > >>>>> > >>>>> On Mon, Apr 1, 2019, 10:16 Simone Tiraboschi <stira...@redhat.com> > wrote: > >>>>>> > >>>>>> Hi, > >>>>>> to understand what's failing I'd suggest to start attaching setup > logs. > >>>>>> > >>>>>> On Sun, Mar 31, 2019 at 5:06 PM Leo David <leoa...@gmail.com> > wrote: > >>>>>>> > >>>>>>> Hello Everyone, > >>>>>>> Using 4.3.2 installation, and after running through HyperConverged > Setup, at the last stage it fails. It seems that the previously created > "engine" volume is not mounted under "/rhev" path, therefore the setup > cannot finish the deployment. > >>>>>>> Any ideea which are the services responsible of mounting the > volumes on oVirt Node distribution ? I'm thinking that maybe this > particularly one failed to start for some reason... > >>>>>>> Thank you very much ! > >>>>>>> > >>>>>>> -- > >>>>>>> Best regards, Leo David > >>>>>>> _______________________________________________ > >>>>>>> Users mailing list -- users@ovirt.org > >>>>>>> To unsubscribe send an email to users-le...@ovirt.org > >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > >>>>>>> oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > >>>>>>> List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/PUXDAQHVNZWF4TIXZ3GIBZHSJ7IC2VHC/ > >>> > >>> > >>> > >>> -- > >>> Best regards, Leo David > >>> _______________________________________________ > >>> Users mailing list -- users@ovirt.org > >>> To unsubscribe send an email to users-le...@ovirt.org > >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > >>> oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > >>> List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/NROCMKIFJDEIWD7PELNTBKDQMPMFLDRN/ > >> > >> > >> > >> -- > >> > >> > >> Thanks, > >> Gobinda > > > > > > > > -- > > Best regards, Leo David > > _______________________________________________ > > Users mailing list -- users@ovirt.org > > To unsubscribe send an email to users-le...@ovirt.org > > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/XDJNBS6EOXCMJKQCSSUHL4YRIYWRKDLS/ > -- Best regards, Leo David
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/AS4RTPBCZDGAVYMAAW7XCPC34X25L2JV/