On Tue, Apr 2, 2019 at 8:14 PM Leo David <[email protected]> wrote:
>
> Just to loop in,  i've forgot to hit "Reply all"
>
> I have deleted everything in the engine gluster mount path, unmounted the 
> engine gluster volume ( not deleted the volume ) ,  and started the wizard 
> with "Use already configured storage". I have pointed to use this gluster 
> volume,  volume gets mounted under the correct path, but installation still 
> fails:
>
> [ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain]
> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is "[]". 
> HTTP response code is 400.
> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault 
> reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is 
> 400."}

And I guess we don't have the engine logs to look at this?
Is there any way you can access the engine console to check?

>
> On the node's vdsm.log I can continuously see:
> 2019-04-02 13:02:18,832+0100 INFO  (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC 
> call Host.getStats succeeded in 0.03 seconds (__init__:312)
> 2019-04-02 13:02:19,906+0100 INFO  (vmrecovery) [vdsm.api] START 
> getConnectedStoragePoolsList(options=None) from=internal, 
> task_id=769c3983-5160-44e4-b1d8-7ab4e41ddd34 (api:48)
> 2019-04-02 13:02:19,907+0100 INFO  (vmrecovery) [vdsm.api] FINISH 
> getConnectedStoragePoolsList return={'poollist': []} from=internal, 
> task_id=769c3983-5160-44e4-b1d8-7ab4e41ddd34 (api:54)
> 2019-04-02 13:02:19,907+0100 INFO  (vmrecovery) [vds] recovery: waiting for 
> storage pool to go up (clientIF:709)
> 2019-04-02 13:02:21,737+0100 INFO  (periodic/2) [vdsm.api] START 
> repoStats(domains=()) from=internal, 
> task_id=ba12fbc1-0170-41a2-82e6-8ccb05ae9e09 (api:48)
> 2019-04-02 13:02:21,738+0100 INFO  (periodic/2) [vdsm.api] FINISH repoStats 
> return={} from=internal, task_id=ba12fbc1-0170-41a2-82e6-8ccb05ae9e09 (api:54)
>

Any calls to "START connectStorageServer" in vdsm.log?

> Should I perform an "engine-cleanup",  delete lvms from Cockpit and start it 
> all over ?

I doubt if that would resolve issue since you did clean up files from the mount.

> Did anyone successfully used this particular iso image 
> "ovirt-node-ng-installer-4.3.2-2019031908.el7.iso" for a single node 
> installation ?
Sorry, don't know.

> Thank you !
> Leo
>
> On Tue, Apr 2, 2019 at 1:45 PM Sahina Bose <[email protected]> wrote:
>>
>> Is it possible you have not cleared the gluster volume between installs?
>>
>> What's the corresponding error in vdsm.log?
>>
>>
>> On Tue, Apr 2, 2019 at 4:07 PM Leo David <[email protected]> wrote:
>> >
>> > And there it is the last lines on the ansible_create_storage_domain log:
>> >
>> > 2019-04-02 10:53:49,139+0100 DEBUG var changed: host "localhost" var 
>> > "otopi_storage_domain_details" type "<type 'dict'>" value: "{
>> >     "changed": false,
>> >     "exception": "Traceback (most recent call last):\n  File 
>> > \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 
>> > 664, in main\n    storage_domains_module.post_create_check(sd_id)\n  File 
>> > \"/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py\", line 
>> > 526, in post_create_check\n    id=storage_domain.id,\n  File 
>> > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, 
>> > in add\n    return self._internal_add(storage_domain, headers, query, 
>> > wait)\n  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", 
>> > line 232, in _internal_add\n    return future.wait() if wait else future\n 
>> >  File \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 
>> > 55, in wait\n    return self._code(response)\n  File 
>> > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, in 
>> > callback\n    self._check_fault(response)\n  File 
>> > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, in 
>> > _check_fault\n    self._raise_error(response, body)\n  File 
>> > \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, in 
>> > _raise_error\n    raise error\nError: Fault reason is \"Operation 
>> > Failed\". Fault detail is \"[]\". HTTP response code is 400.\n",
>> >     "failed": true,
>> >     "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[]\". 
>> > HTTP response code is 400."
>> > }"
>> > 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var 
>> > "ansible_play_hosts" type "<type 'list'>" value: "[]"
>> > 2019-04-02 10:53:49,141+0100 DEBUG var changed: host "localhost" var 
>> > "play_hosts" type "<type 'list'>" value: "[]"
>> > 2019-04-02 10:53:49,142+0100 DEBUG var changed: host "localhost" var 
>> > "ansible_play_batch" type "<type 'list'>" value: "[]"
>> > 2019-04-02 10:53:49,142+0100 ERROR ansible failed {'status': 'FAILED', 
>> > 'ansible_type': 'task', 'ansible_task': u'Activate storage domain', 
>> > 'ansible_result': u'type: <type \'dict\'>\nstr: {\'_ansible_parsed\': 
>> > True, u\'exception\': u\'Traceback (most recent call last):\\n  File 
>> > "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 664, 
>> > in main\\n    storage_domains_module.post_create_check(sd_id)\\n  File 
>> > "/tmp/ansible_ovirt_storage_domain_payload_6Jxg5v/__main__.py", line 526', 
>> > 'task_duration': 9, 'ansible_host': u'localhost', 'ansible_playbook': 
>> > u'/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml'}
>> > 2019-04-02 10:53:49,143+0100 DEBUG ansible on_any args 
>> > <ansible.executor.task_result.TaskResult object at 0x7f03fd025e50> kwargs 
>> > ignore_errors:None
>> > 2019-04-02 10:53:49,148+0100 INFO ansible stats {
>> >     "ansible_playbook": 
>> > "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml",
>> >     "ansible_playbook_duration": "01:15 Minutes",
>> >     "ansible_result": "type: <type 'dict'>\nstr: {u'localhost': 
>> > {'unreachable': 0, 'skipped': 6, 'ok': 23, 'changed': 0, 'failures': 1}}",
>> >     "ansible_type": "finish",
>> >     "status": "FAILED"
>> > }
>> > 2019-04-02 10:53:49,149+0100 INFO SUMMARY:
>> > Duration        Task Name
>> > --------        --------
>> > [ < 1 sec ]     Execute just a specific set of steps
>> > [  00:02  ]     Force facts gathering
>> > [  00:02  ]     Check local VM dir stat
>> > [  00:02  ]     Obtain SSO token using username/password credentials
>> > [  00:02  ]     Fetch host facts
>> > [  00:01  ]     Fetch cluster ID
>> > [  00:02  ]     Fetch cluster facts
>> > [  00:02  ]     Fetch Datacenter facts
>> > [  00:01  ]     Fetch Datacenter ID
>> > [  00:01  ]     Fetch Datacenter name
>> > [  00:02  ]     Add glusterfs storage domain
>> > [  00:02  ]     Get storage domain details
>> > [  00:02  ]     Find the appliance OVF
>> > [  00:02  ]     Parse OVF
>> > [  00:02  ]     Get required size
>> > [ FAILED  ]     Activate storage domain
>> >
>> > Any ideea on how to escalate this issue ?
>> > It just does not make sense to not be able to install from scratch a fresh 
>> > node...
>> >
>> > Have a nice day  !
>> >
>> > Leo
>> >
>> >
>> > On Tue, Apr 2, 2019 at 12:11 PM Gobinda Das <[email protected]> wrote:
>> >>
>> >> Hi Leo,
>> >>  Can you please paste "df -Th" and "gluster v status" out put ?
>> >> Wanted to make sure engine mounted and volumes and bricks are up.
>> >> What does vdsm log say?
>> >>
>> >> On Tue, Apr 2, 2019 at 2:06 PM Leo David <[email protected]> wrote:
>> >>>
>> >>> Thank you very much !
>> >>> I have just installed a new fresh node,   and triggered the single 
>> >>> instance hyperconverged setup. It seems it fails at the hosted engine 
>> >>> final steps of deployment:
>> >>>  INFO ] TASK [ovirt.hosted_engine_setup : Get required size]
>> >>> [ INFO ] ok: [localhost]
>> >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Remove unsuitable storage 
>> >>> domain]
>> >>> [ INFO ] skipping: [localhost]
>> >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Check storage domain free 
>> >>> space]
>> >>> [ INFO ] skipping: [localhost]
>> >>> [ INFO ] TASK [ovirt.hosted_engine_setup : Activate storage domain]
>> >>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is 
>> >>> "[Cannot attach Storage. There is no active Host in the Data Center.]". 
>> >>> HTTP response code is 409.
>> >>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": 
>> >>> "Fault reason is \"Operation Failed\". Fault detail is \"[Cannot attach 
>> >>> Storage. There is no active Host in the Data Center.]\". HTTP response 
>> >>> code is 409."}
>> >>> Also,  the 
>> >>> ovirt-hosted-engine-setup-ansible-create_storage_domain-201932112413-xkq6nb.log
>> >>>  throws the following:
>> >>>
>> >>> 2019-04-02 09:25:40,420+0100 DEBUG var changed: host "localhost" var 
>> >>> "otopi_storage_domain_details" type "<type 'dict'>" value: "{
>> >>>     "changed": false,
>> >>>     "exception": "Traceback (most recent call last):\n  File 
>> >>> \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", line 
>> >>> 664, in main\n    storage_domains_module.post_create_check(sd_id)\n  
>> >>> File \"/tmp/ansible_ovirt_storage_domain_payload_87MSyY/__main__.py\", 
>> >>> line 526, in post_create_check\n    id=storage_domain.id,\n  File 
>> >>> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\", line 3053, 
>> >>> in add\n    return self._internal_add(storage_domain, headers, query, 
>> >>> wait)\n  File 
>> >>> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 232, 
>> >>> in _internal_add\n    return future.wait() if wait else future\n  File 
>> >>> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 55, in 
>> >>> wait\n    return self._code(response)\n  File 
>> >>> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 229, 
>> >>> in callback\n    self._check_fault(response)\n  File 
>> >>> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 132, 
>> >>> in _check_fault\n    self._raise_error(response, body)\n  File 
>> >>> \"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\", line 118, 
>> >>> in _raise_error\n    raise error\nError: Fault reason is \"Operation 
>> >>> Failed\". Fault detail is \"[Cannot attach Storage. There is no active 
>> >>> Host in the Data Center.]\". HTTP response code is 409.\n",
>> >>>     "failed": true,
>> >>>     "msg": "Fault reason is \"Operation Failed\". Fault detail is 
>> >>> \"[Cannot attach Storage. There is no active Host in the Data 
>> >>> Center.]\". HTTP response code is 409."
>> >>> }"
>> >>>
>> >>> I have used the ovirt-node-ng-installer-4.3.2-2019031908.el7.iso. So 
>> >>> far,  I am unable to deploy oVirt single node Hyperconverged...
>> >>> Any thoughts ?
>> >>>
>> >>>
>> >>>
>> >>> On Mon, Apr 1, 2019 at 9:46 PM Simone Tiraboschi <[email protected]> 
>> >>> wrote:
>> >>>>
>> >>>>
>> >>>>
>> >>>> On Mon, Apr 1, 2019 at 6:14 PM Leo David <[email protected]> wrote:
>> >>>>>
>> >>>>> Thank you Simone.
>> >>>>> I've decides to go for a new fresh install from iso, and i'll keep 
>> >>>>> posted if any troubles arise. But I am still trying to understand what 
>> >>>>> are the services that mount the lvms and volumes after configuration. 
>> >>>>> There is nothing related in fstab, so I assume there are a couple of 
>> >>>>> .mount files somewhere in the filesystem.
>> >>>>> Im just trying to understand node's underneath workflow.
>> >>>>
>> >>>>
>> >>>> hosted-engine configuration is stored in 
>> >>>> /etc/ovirt-hosted-engine/hosted-engine.conf; ovirt-ha-broker will mount 
>> >>>> the hosted-engine storage domain according to that and so 
>> >>>> ovirt-ha-agent will be able to start the engine VM.
>> >>>> Everything else is just in the engine DB.
>> >>>>
>> >>>>>
>> >>>>>
>> >>>>> On Mon, Apr 1, 2019, 10:16 Simone Tiraboschi <[email protected]> 
>> >>>>> wrote:
>> >>>>>>
>> >>>>>> Hi,
>> >>>>>> to understand what's failing I'd suggest to start attaching setup 
>> >>>>>> logs.
>> >>>>>>
>> >>>>>> On Sun, Mar 31, 2019 at 5:06 PM Leo David <[email protected]> wrote:
>> >>>>>>>
>> >>>>>>> Hello Everyone,
>> >>>>>>> Using 4.3.2 installation, and after running through HyperConverged 
>> >>>>>>> Setup,  at the last stage it fails. It seems that the previously 
>> >>>>>>> created "engine" volume is not mounted under "/rhev" path, therefore 
>> >>>>>>> the setup cannot finish the deployment.
>> >>>>>>> Any ideea which are the services responsible of mounting the volumes 
>> >>>>>>> on oVirt Node distribution ? I'm thinking that maybe this 
>> >>>>>>> particularly one failed to start for some reason...
>> >>>>>>> Thank you very much !
>> >>>>>>>
>> >>>>>>> --
>> >>>>>>> Best regards, Leo David
>> >>>>>>> _______________________________________________
>> >>>>>>> Users mailing list -- [email protected]
>> >>>>>>> To unsubscribe send an email to [email protected]
>> >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> >>>>>>> oVirt Code of Conduct: 
>> >>>>>>> https://www.ovirt.org/community/about/community-guidelines/
>> >>>>>>> List Archives: 
>> >>>>>>> https://lists.ovirt.org/archives/list/[email protected]/message/PUXDAQHVNZWF4TIXZ3GIBZHSJ7IC2VHC/
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> Best regards, Leo David
>> >>> _______________________________________________
>> >>> Users mailing list -- [email protected]
>> >>> To unsubscribe send an email to [email protected]
>> >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> >>> oVirt Code of Conduct: 
>> >>> https://www.ovirt.org/community/about/community-guidelines/
>> >>> List Archives: 
>> >>> https://lists.ovirt.org/archives/list/[email protected]/message/NROCMKIFJDEIWD7PELNTBKDQMPMFLDRN/
>> >>
>> >>
>> >>
>> >> --
>> >>
>> >>
>> >> Thanks,
>> >> Gobinda
>> >
>> >
>> >
>> > --
>> > Best regards, Leo David
>> > _______________________________________________
>> > Users mailing list -- [email protected]
>> > To unsubscribe send an email to [email protected]
>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> > oVirt Code of Conduct: 
>> > https://www.ovirt.org/community/about/community-guidelines/
>> > List Archives: 
>> > https://lists.ovirt.org/archives/list/[email protected]/message/XDJNBS6EOXCMJKQCSSUHL4YRIYWRKDLS/
>
>
>
> --
> Best regards, Leo David
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/IB3NQXD4RGII6M6UUMTFXZ4M5NFXJJG3/

Reply via email to