Il giorno mar 15 gen 2019 alle ore 21:27 Nir Soffer <[email protected]> ha scritto:
> On Tue, Jan 15, 2019 at 4:32 PM Gianluca Cecchi <[email protected]> > wrote: > >> The mail was partly scrambled in its contents so I put some clarification >> here: >> >> On Tue, Jan 15, 2019 at 2:38 PM Gianluca Cecchi < >> [email protected]> wrote: >> >>> >>>> >>> So after starting from scratch and using also the info as detailed on >>> thread: >>> https://www.mail-archive.com/[email protected]/msg52879.html >>> >>> the steps now have been: >>> >>> - install from ovirt-node-ng-installer-4.3.0-2019011010.el7.iso and >>> reboot >>> >>> - connect to cockpit and open terminal >>> >> >> This step is related to ssh daemon >> cd /etc/ssh >> chmod 600 *key >> systemctl restart sshd >> >> The step below is related to ovirt-imageio-daemon >> >> >>> mkdir /var/run/vdsm >>> chmod 755 /var/run/vdsm >>> chown vdsm.kvm /var/run/vdsm >>> mkdir /var/run/vdsm/dhclientmon >>> chmod 755 /var/run/vdsm/dhclientmon/ >>> chown vdsm.kvm /var/run/vdsm/dhclientmon/ >>> mkdir /var/run/vdsm/trackedInterfaces >>> chmod 755 /var/run/vdsm/trackedInterfaces/ >>> chown vdsm.kvm /var/run/vdsm/trackedInterfaces/ >>> mkdir /var/run/vdsm/v2v >>> chmod 700 /var/run/vdsm/v2v >>> chown vdsm.kvm /var/run/vdsm/v2v/ >>> mkdir /var/run/vdsm/vhostuser >>> chmod 755 /var/run/vdsm/vhostuser/ >>> chown vdsm.kvm /var/run/vdsm/vhostuser/ >>> mkdir /var/run/vdsm/payload >>> chmod 755 /var/run/vdsm/payload/ >>> chown vdsm.kvm /var/run/vdsm/payload/ >>> >>> systemctl restart sshd >>> >> >> Actually: >> >> systemctl restart ovirt-imageio-daemon >> >> >>> >>> - put in the newer version of vdsm-api.pickle >>> from vdsm-api-4.30.5-2.gitf824ec2.el7.noarch.rpm >>> in /usr/lib/python2.7/site-packages/vdsm/rpc/vdsm-api.pickle >>> >> >> download of vdsm-api.pickle can be directly done here eventually: >> >> https://drive.google.com/file/d/1AhakKhm_dzx-Gxt-Y1OojzRUwHs75kot/view?usp=sharing >> >> >>> >>> - run the wizard for the gluster+he setup (the right positioned option) >>> inside the gdeploy text window click edit and add >>> " >>> [diskcount] >>> 1 >>> >>> " >>> under the section >>> " >>> [disktype] >>> jbod >>> " >>> >>> In my case with single disk I choose JBOD options >> >> >>> - first 2 steps ok >>> >>> - last step fails in finish part >>> >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Fetch Datacenter name] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Add NFS storage domain] >>> [ INFO ] skipping: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Add glusterfs storage domain] >>> [ INFO ] changed: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Add iSCSI storage domain] >>> [ INFO ] skipping: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Add Fibre Channel storage >>> domain] >>> [ INFO ] skipping: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Get storage domain details] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : debug] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Find the appliance OVF] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : debug] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Parse OVF] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Get required size] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : debug] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Remove unsuitable storage >>> domain] >>> [ INFO ] skipping: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : debug] >>> [ INFO ] ok: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Check storage domain free >>> space] >>> [ INFO ] skipping: [localhost] >>> [ INFO ] TASK [oVirt.hosted-engine-setup : Activate storage domain] >>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is >>> "[]". HTTP response code is 400. >>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": >>> "Fault reason is \"Operation Failed\". Fault detail is \"[]\". HTTP >>> response code is 400."} >>> >>> On engine.log I see >>> >>> 2019-01-15 13:50:35,317+01 INFO >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] >>> (default task-2) [51725212] START, CreateStoragePoolVDSCommand(HostName = >>> ov4301.localdomain.lo >>> cal, >>> CreateStoragePoolVDSCommandParameters:{hostId='e8f105f1-37ed-4ac4-bfc3-b1e55ed3027f', >>> storagePoolId='96a31a7e-18bb-11e9-9a34-00163e6196f3', >>> storagePoolName='Default', masterDomainId='14ec2fc7-8c2 >>> b-487c-8f4f-428644650928', >>> domainsIdList='[14ec2fc7-8c2b-487c-8f4f-428644650928]', >>> masterVersion='1'}), log id: 4baccd53 >>> 2019-01-15 13:50:36,345+01 ERROR >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] >>> (default task-2) [51725212] Failed in 'CreateStoragePoolVDS' method >>> 2019-01-15 13:50:36,354+01 ERROR >>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>> (default task-2) [51725212] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), >>> VDSM ov4301.localdomain.local command CreateStoragePoolVDS failed: Cannot >>> acquire host id: (u'14ec2fc7-8c2b-487c-8f4f-428644650928', >>> SanlockException(-203, 'Sanlock lockspace add failure', 'Watchdog device >>> error')) >>> 2019-01-15 13:50:36,354+01 INFO >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] >>> (default task-2) [51725212] Command >>> 'org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand' >>> return value 'StatusOnlyReturn [status=Status [code=661, message=Cannot >>> acquire host id: (u'14ec2fc7-8c2b-487c-8f4f-428644650928', >>> SanlockException(-203, 'Sanlock lockspace add failure', 'Watchdog device >>> error'))]]' >>> 2019-01-15 13:50:36,354+01 INFO >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] >>> (default task-2) [51725212] HostName = ov4301.localdomain.local >>> 2019-01-15 13:50:36,355+01 ERROR >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] >>> (default task-2) [51725212] Command 'CreateStoragePoolVDSCommand(HostName = >>> ov4301.localdomain.local, >>> CreateStoragePoolVDSCommandParameters:{hostId='e8f105f1-37ed-4ac4-bfc3-b1e55ed3027f', >>> storagePoolId='96a31a7e-18bb-11e9-9a34-00163e6196f3', >>> storagePoolName='Default', >>> masterDomainId='14ec2fc7-8c2b-487c-8f4f-428644650928', >>> domainsIdList='[14ec2fc7-8c2b-487c-8f4f-428644650928]', >>> masterVersion='1'})' execution failed: VDSGenericException: >>> VDSErrorException: Failed to CreateStoragePoolVDS, error = Cannot acquire >>> host id: (u'14ec2fc7-8c2b-487c-8f4f-428644650928', SanlockException(-203, >>> 'Sanlock lockspace add failure', 'Watchdog device error')), code = 661 >>> 2019-01-15 13:50:36,355+01 INFO >>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] >>> (default task-2) [51725212] FINISH, CreateStoragePoolVDSCommand, return: , >>> log id: 4baccd53 >>> 2019-01-15 13:50:36,355+01 ERROR >>> [org.ovirt.engine.core.bll.storage.pool.AddStoragePoolWithStoragesCommand] >>> (default task-2) [51725212] Command >>> 'org.ovirt.engine.core.bll.storage.pool.AddStoragePoolWithStoragesCommand' >>> failed: EngineException: >>> org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException: >>> VDSGenericException: VDSErrorException: Failed to CreateStoragePoolVDS, >>> error = Cannot acquire host id: (u'14ec2fc7-8c2b-487c-8f4f-428644650928', >>> SanlockException(-203, 'Sanlock lockspace add failure', 'Watchdog device >>> error')), code = 661 (Failed with error AcquireHostIdFailure and code 661) >>> 2019-01-15 13:50:36,379+01 INFO >>> [org.ovirt.engine.core.bll.CommandCompensator] (default task-2) [51725212] >>> Command [id=c55d9962-368e-4e0c-8fee-bd06e7570062]: Compensating >>> DELETED_OR_UPDATED_ENTITY of >>> org.ovirt.engine.core.common.businessentities.StoragePool; snapshot: >>> id=96a31a7e-18bb-11e9-9a34-00163e6196f3. >>> >>> On host: >>> >>> [root@ov4301 log]# cat /etc/hosts >>> 192.168.124.50 ov43eng.localdomain.local # temporary entry added by >>> hosted-engine-setup for the bootstrap VM >>> 127.0.0.1 localhost localhost.localdomain localhost4 >>> localhost4.localdomain4 >>> ::1 localhost localhost.localdomain localhost6 >>> localhost6.localdomain6 >>> 192.168.122.210 ov43eng.localdomain.local ov43eng >>> 192.168.122.211 ov4301.localdomain.local ov4301 >>> [root@ov4301 log]# >>> >>> [root@ov4301 log]# df -h | grep gluster >>> /dev/mapper/gluster_vg_sdb-gluster_lv_engine 64G >>> 36M 64G 1% /gluster_bricks/engine >>> /dev/mapper/gluster_vg_sdb-gluster_lv_data 30G >>> 34M 30G 1% /gluster_bricks/data >>> /dev/mapper/gluster_vg_sdb-gluster_lv_vmstore 20G >>> 34M 20G 1% /gluster_bricks/vmstore >>> 192.168.123.211:/engine 64G >>> 691M 64G 2% /rhev/data-center/mnt/glusterSD/192.168.123.211:_engine >>> [root@ov4301 log]# >>> >>> >>> and in its messages: >>> >>> Jan 15 13:35:49 ov4301 dnsmasq-dhcp[22934]: DHCPREQUEST(virbr0) >>> 192.168.124.50 00:16:3e:61:96:f3 >>> Jan 15 13:35:49 ov4301 dnsmasq-dhcp[22934]: DHCPACK(virbr0) >>> 192.168.124.50 00:16:3e:61:96:f3 ov43eng >>> Jan 15 13:35:49 ov4301 dnsmasq-dhcp[22934]: not giving name ov43eng to >>> the DHCP lease of 192.168.124.50 because the name exists in /etc/hosts with >>> address 192.168.122.210 >>> Jan 15 13:40:01 ov4301 systemd: Started Session 38 of user root. >>> Jan 15 13:47:12 ov4301 vdsm[29591]: WARN MOM not available. >>> Jan 15 13:47:12 ov4301 vdsm[29591]: WARN MOM not available, KSM stats >>> will be missing. >>> Jan 15 13:49:05 ov4301 python: ansible-setup Invoked with filter=* >>> gather_subset=['all'] fact_path=/etc/ansible/facts.d gather_timeout=10 >>> Jan 15 13:49:19 ov4301 python: ansible-stat Invoked with >>> checksum_algorithm=sha1 get_checksum=True follow=False >>> path=/var/tmp/localvmOIXI_W get_md5=None get_mime=True get_attributes=True >>> Jan 15 13:49:24 ov4301 python: ansible-ovirt_auth Invoked with >>> username=None kerberos=False timeout=0 url=None insecure=True hostname=None >>> compress=True state=present headers=None token=None ovirt_auth=None >>> ca_file=None password=NOT_LOGGING_PARAMETER >>> Jan 15 13:49:29 ov4301 python: ansible-ovirt_host_facts Invoked with >>> all_content=False pattern=name=ov4301.localdomain.local fetch_nested=False >>> nested_attributes=[] auth={'timeout': 0, 'url': ' >>> https://ov43eng.localdomain.local/ovirt-engine/api', 'insecure': True, >>> 'kerberos': False, 'compress': True, 'headers': None, 'token': >>> 'Q8qt0Z9DmHJRdg3wk7YxNOAs0JPpBMxxstVx3I8skbulwRWp1SsVXuZYq4DUuPWeEnUZ2bD8TAuwCzJ3qlFYlw', >>> 'ca_file': None} >>> Jan 15 13:49:35 ov4301 python: ansible-ovirt_cluster_facts Invoked with >>> pattern= fetch_nested=False nested_attributes=[] auth={'timeout': 0, 'url': >>> 'https://ov43eng.localdomain.local/ovirt-engine/api', 'insecure': True, >>> 'kerberos': False, 'compress': True, 'headers': None, 'token': >>> 'Q8qt0Z9DmHJRdg3wk7YxNOAs0JPpBMxxstVx3I8skbulwRWp1SsVXuZYq4DUuPWeEnUZ2bD8TAuwCzJ3qlFYlw', >>> 'ca_file': None} >>> Jan 15 13:49:43 ov4301 python: ansible-ovirt_datacenter_facts Invoked >>> with pattern= fetch_nested=False nested_attributes=[] auth={'timeout': 0, >>> 'url': 'https://ov43eng.localdomain.local/ovirt-engine/api', >>> 'insecure': True, 'kerberos': False, 'compress': True, 'headers': None, >>> 'token': >>> 'Q8qt0Z9DmHJRdg3wk7YxNOAs0JPpBMxxstVx3I8skbulwRWp1SsVXuZYq4bD8TAuwCzJ3qlFYlw', >>> 'ca_file': None} host=ov4301.localdomain.local nested_attributes=[] >>> wait=True domain_function=data name=hosted_storage >>> critical_space_action_blocker=None posixfs=None poll_interval=3 >>> glusterfs=None nfs=None timeout=180 backup=None discard_after_delete=None >>> Jan 15 13:50:34 ov4301 systemd: Started Session c24 of user root. >>> Jan 15 13:50:34 ov4301 sanlock[26802]: 2019-01-15 13:50:34 11549 >>> [21290]: s1 wdmd_connect failed -13 >>> Jan 15 13:50:34 ov4301 sanlock[26802]: 2019-01-15 13:50:34 11549 >>> [21290]: s1 connect_watchdog failed -1 >>> Jan 15 13:50:35 ov4301 sanlock[26802]: 2019-01-15 13:50:35 11550 >>> [26810]: s1 add_lockspace fail result -203 >>> Jan 15 13:56:48 ov4301 dnsmasq-dhcp[22934]: DHCPREQUEST(virbr0) >>> 192.168.124.50 00:16:3e:61:96:f3 >>> Jan 15 13:56:48 ov4301 dnsmasq-dhcp[22934]: DHCPACK(virbr0) >>> 192.168.124.50 00:16:3e:61:96:f3 ov43eng >>> Jan 15 13:56:48 ov4301 dnsmasq-dhcp[22934]: not giving name ov43eng to >>> the DHCP lease of 192.168.124.50 because the name exists in /etc/hosts with >>> address 192.168.122.210 >>> Jan 15 13:59:34 ov4301 chronyd[26447]: Source 212.45.144.206 replaced >>> with 80.211.52.109 >>> Jan 15 13:59:44 ov4301 vdsm[29591]: WARN MOM not available. >>> Jan 15 13:59:44 ov4301 vdsm[29591]: WARN MOM not available, KSM stats >>> will be missing. >>> DUuPWeEnUZ2bD8TAuwCzJ3qlFYlw', 'ca_file': None} >>> Jan 15 13:49:54 ov4301 python: ansible-ovirt_storage_domain Invoked with >>> comment=None warning_low_space=None fetch_nested=False localfs=None >>> data_center=Default id=None iscsi=None state=unattached >>> wipe_after_delete=None destroy=None fcp=None description=None format=None >>> auth={'username':********@internal', 'url': ' >>> https://ov43eng.localdomain.local/ovirt-engine/api', 'insecure': True, >>> 'password': 'passw0rd'} host=ov4301.localdomain.local nested_attributes=[] >>> wait=True domain_function=data name=hosted_storage >>> critical_space_action_blocker=None posixfs=None poll_interval=3 >>> glusterfs={'path': '/engine', 'mount_options': None, 'address': >>> '192.168.123.211'} nfs=None timeout=180 backup=None >>> discard_after_delete=None >>> Jan 15 13:50:01 ov4301 systemd: Started Session 39 of user root. >>> Jan 15 13:50:01 ov4301 systemd: Created slice vdsm-glusterfs.slice. >>> Jan 15 13:50:01 ov4301 systemd: Started /usr/bin/mount -t glusterfs >>> 192.168.123.211:/engine /rhev/data-center/mnt/glusterSD/192.168.123.211: >>> _engine. >>> Jan 15 13:50:01 ov4301 kernel: fuse init (API version 7.22) >>> Jan 15 13:50:01 ov4301 systemd: Mounting FUSE Control File System... >>> Jan 15 13:50:02 ov4301 systemd: Mounted FUSE Control File System. >>> Jan 15 13:50:02 ov4301 systemd: Started Session c20 of user root. >>> Jan 15 13:50:02 ov4301 systemd: Started Session c21 of user root. >>> Jan 15 13:50:03 ov4301 systemd: Started Session c22 of user root. >>> Jan 15 13:50:03 ov4301 systemd: Started Session c23 of user root. >>> Jan 15 13:50:12 ov4301 python: ansible-ovirt_storage_domain_facts >>> Invoked with pattern=name=hosted_storage fetch_nested=False >>> nested_attributes=[] auth={'timeout': 0, 'url': ' >>> https://ov43eng.localdomain.local/ovirt-engine/api', 'insecure': True, >>> 'kerberos': False, 'compress': True, 'headers': None, 'token': >>> 'Q8qt0Z9DmHJRdg3wk7YxNOAs0JPpBMxxstVx3I8skbulwRWp1SsVXuZYq4DUuPWeEnUZ2bD8TAuwCzJ3qlFYlw', >>> 'ca_file': None} >>> Jan 15 13:50:18 ov4301 python: ansible-find Invoked with excludes=None >>> paths=['/var/tmp/localvmOIXI_W/master'] file_type=file age=None >>> contains=None recurse=True age_stamp=mtime patterns=['^.*.(?<!meta).ovf$'] >>> depth=None get_checksum=False use_regex=True follow=False hidden=False >>> size=None >>> Jan 15 13:50:21 ov4301 python: ansible-xml Invoked with >>> xpath=/ovf:Envelope/Section/Disk count=False set_children=None >>> xmlstring=None strip_cdata_tags=False attribute=size pretty_print=False >>> add_children=None value=None content=attribute state=present >>> namespaces={'vssd': ' >>> http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_VirtualSystemSettingData', >>> 'rasd': ' >>> http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_ResourceAllocationSettingData', >>> 'xsi': 'http://www.w3.org/2001/XMLSchema-instance', 'ovf': ' >>> http://schemas.dmtf.org/ovf/envelope/1/'} input_type=yaml >>> print_match=False >>> path=/var/tmp/localvmOIXI_W/master/vms/c99e3e6b-db14-446f-aaee-48a056d3dd93/c99e3e6b-db14-446f-aaee-48a056d3dd93.ovf >>> backup=False >>> Jan 15 13:50:30 ov4301 python: ansible-ovirt_storage_domain Invoked with >>> comment=None warning_low_space=None fetch_nested=False localfs=None >>> data_center=Default id=None iscsi=None state=present wipe_after_delete=None >>> destroy=None fcp=None description=None format=None auth={'timeout': 0, >>> 'url': 'https://ov43eng.localdomain.local/ovirt-engine/api', >>> 'insecure': True, 'kerberos': False, 'compress': True, 'headers': None, >>> 'token': >>> 'Q8qt0Z9DmHJRdg3wk7YxNOAs0JPpBMxxstVx3I8skbulwRWp1SsVXuZYq4DUuPWeEnUZ2bD8TAuwCzJ3qlFYlw', >>> 'ca_file': None} host=ov4301.localdomain.local nested_attributes=[] >>> wait=True domain_function=data name=hosted_storage >>> critical_space_action_blocker=None posixfs=None poll_interval=3 >>> glusterfs=None nfs=None timeout=180 backup=None discard_after_delete=None >>> Jan 15 13:50:34 ov4301 systemd: Started Session c24 of user root. >>> Jan 15 13:50:34 ov4301 sanlock[26802]: 2019-01-15 13:50:34 11549 >>> [21290]: s1 wdmd_connect failed -13 >>> Jan 15 13:50:34 ov4301 sanlock[26802]: 2019-01-15 13:50:34 11549 >>> [21290]: s1 connect_watchdog failed -1 >>> Jan 15 13:50:35 ov4301 sanlock[26802]: 2019-01-15 13:50:35 11550 >>> [26810]: s1 add_lockspace fail result -203 >>> Jan 15 13:56:48 ov4301 dnsmasq-dhcp[22934]: DHCPREQUEST(virbr0) >>> 192.168.124.50 00:16:3e:61:96:f3 >>> Jan 15 13:56:48 ov4301 dnsmasq-dhcp[22934]: DHCPACK(virbr0) >>> 192.168.124.50 00:16:3e:61:96:f3 ov43eng >>> Jan 15 13:56:48 ov4301 dnsmasq-dhcp[22934]: not giving name ov43eng to >>> the DHCP lease of 192.168.124.50 because the name exists in /etc/hosts with >>> address 192.168.122.210 >>> Jan 15 13:59:34 ov4301 chronyd[26447]: Source 212.45.144.206 replaced >>> with 80.211.52.109 >>> Jan 15 13:59:44 ov4301 vdsm[29591]: WARN MOM not available. >>> Jan 15 13:59:44 ov4301 vdsm[29591]: WARN MOM not available, KSM stats >>> will be missing. >>> >>> In vdsm.log : >>> >>> 2019-01-15 13:50:34,980+0100 INFO (jsonrpc/2) [storage.StorageDomain] >>> sdUUID=14ec2fc7-8c2b-487c-8f4f-428644650928 (fileSD:533) >>> 2019-01-15 13:50:34,984+0100 INFO (jsonrpc/2) [storage.StoragePool] >>> Creating pool directory >>> '/rhev/data-center/96a31a7e-18bb-11e9-9a34-00163e6196f3' (sp:634) >>> 2019-01-15 13:50:34,984+0100 INFO (jsonrpc/2) [storage.fileUtils] >>> Creating directory: /rhev/data-center/96a31a7e-18bb-11e9-9a34-00163e6196f3 >>> mode: None (fileUtils:199) >>> 2019-01-15 13:50:34,985+0100 INFO (jsonrpc/2) [storage.SANLock] >>> Acquiring host id for domain 14ec2fc7-8c2b-487c-8f4f-428644650928 (id=250, >>> async=False) (clusterlock:294) >>> 2019-01-15 13:50:35,987+0100 INFO (jsonrpc/2) [vdsm.api] FINISH >>> createStoragePool error=Cannot acquire host id: >>> (u'14ec2fc7-8c2b-487c-8f4f-428644650928', SanlockException(-203, 'Sanlock >>> lockspace add failure', 'Watchdog device error')) >>> from=::ffff:192.168.124.50,42356, flow_id=51725212, >>> task_id=7cbd7c09-e934-4396-bd9d-61e9f0e00bd3 (api:52) >>> 2019-01-15 13:50:35,988+0100 ERROR (jsonrpc/2) >>> [storage.TaskManager.Task] (Task='7cbd7c09-e934-4396-bd9d-61e9f0e00bd3') >>> Unexpected error (task:875) >>> Traceback (most recent call last): >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line >>> 882, in _run >>> return fn(*args, **kargs) >>> File "<string>", line 2, in createStoragePool >>> File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, >>> in method >>> ret = func(*args, **kwargs) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line >>> 1003, in createStoragePool >>> leaseParams) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 636, >>> in create >>> self._acquireTemporaryClusterLock(msdUUID, leaseParams) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 567, >>> in _acquireTemporaryClusterLock >>> msd.acquireHostId(self.id) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/sd.py", line 860, >>> in acquireHostId >>> self._manifest.acquireHostId(hostId, async) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/sd.py", line 467, >>> in acquireHostId >>> self._domainLock.acquireHostId(hostId, async) >>> File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", >>> line 325, in acquireHostId >>> raise se.AcquireHostIdFailure(self._sdUUID, e) >>> AcquireHostIdFailure: Cannot acquire host id: >>> (u'14ec2fc7-8c2b-487c-8f4f-428644650928', SanlockException(-203, 'Sanlock >>> lockspace add failure', 'Watchdog device error')) >>> 2019-01-15 13:50:35,988+0100 INFO (jsonrpc/2) >>> [storage.TaskManager.Task] (Task='7cbd7c09-e934-4396-bd9d-61e9f0e00bd3') >>> aborting: Task is aborted: "Cannot acquire host id: >>> (u'14ec2fc7-8c2b-487c-8f4f-428644650928', SanlockException(-203, 'Sanlock >>> lockspace add failure', 'Watchdog device error'))" - code 661 (task:1181) >>> 2019-01-15 13:50:35,989+0100 ERROR (jsonrpc/2) [storage.Dispatcher] >>> FINISH createStoragePool error=Cannot acquire host id: >>> (u'14ec2fc7-8c2b-487c-8f4f-428644650928', SanlockException(-203, 'Sanlock >>> lockspace add failure', 'Watchdog device error')) (dispatcher:81) >>> 2019-01-15 13:50:35,990+0100 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] >>> RPC call StoragePool.create failed (error 661) in 1.01 seconds >>> (__init__:312) >>> 2019-01-15 13:50:38,109+0100 INFO (vmrecovery) [vdsm.api] START >>> getConnectedStoragePoolsList(options=None) from=internal, >>> task_id=e69af7e1-c456-4822-ae06-7b309263257d (api:48) >>> 2019-01-15 13:50:38,109+0100 INFO (vmrecovery) [vdsm.api] FINISH >>> getConnectedStoragePoolsList return={'poollist': []} from=internal, >>> task_id=e69af7e1-c456-4822-ae06-7b309263257d (api:54) >>> 2019-01-15 13:50:38,110+0100 INFO (vmrecovery) [vds] recovery: waiting >>> for storage pool to go up (clientIF:705) >>> 2019-01-15 13:50:39,802+0100 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] >>> RPC call GlusterHost.list succeeded in 0.33 seconds (__init__:312) >>> 2019-01-15 13:50:39,996+0100 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] >>> RPC call GlusterVolume.list succeeded in 0.18 seconds (__init__:312) >>> 2019-01-15 13:50:43,115+0100 INFO (vmrecovery) [vdsm.api] START >>> getConnectedStoragePoolsList(options=None) from=internal, >>> task_id=729c054e-021a-4d2e-b36c-edfb186a1210 (api:48) >>> 2019-01-15 13:50:43,116+0100 INFO (vmrecovery) [vdsm.api] FINISH >>> getConnectedStoragePoolsList return={'poollist': []} from=internal, >>> task_id=729c054e-021a-4d2e-b36c-edfb186a1210 (api:54) >>> 2019-01-15 13:50:43,116+0100 INFO (vmrecovery) [vds] recovery: waiting >>> for storage pool to go up (clientIF:705) >>> 2019-01-15 13:50:43,611+0100 INFO (jsonrpc/5) [api.host] START >>> getStats() from=::ffff:192.168.124.50,42356 (api:48) >>> 2019-01-15 13:50:43,628+0100 INFO (jsonrpc/5) [vdsm.api] START >>> repoStats(domains=()) from=::ffff:192.168.124.50,42356, >>> task_id=868d05f4-5535-4cb3-b283-92d3c1595bb3 (api:48) >>> 2019-01-15 13:50:43,628+0100 INFO (jsonrpc/5) [vdsm.api] FINISH >>> repoStats return={} from=::ffff:192.168.124.50,42356, >>> task_id=868d05f4-5535-4cb3-b283-92d3c1595bb3 (api:54) >>> 2019-01-15 13:50:43,628+0100 INFO (jsonrpc/5) [vdsm.api] START >>> multipath_health() from=::ffff:192.168.124.50,42356, >>> task_id=7f4fdad0-2b2e-4dcf-88b8-b7cf2689d4d9 (api:48) >>> >>> >>> Gianluca >>> >>> >> >> regarding sanlock daemon and watchdog multiple daemon, the latter seems >> to have no log file but I can only see status of service: >> >> [root@ov4301 ~]# systemctl status wdmd >> ● wdmd.service - Watchdog Multiplexing Daemon >> Loaded: loaded (/usr/lib/systemd/system/wdmd.service; disabled; vendor >> preset: disabled) >> Active: active (running) since Tue 2019-01-15 10:38:13 CET; 4h 32min >> ago >> Main PID: 3763 (wdmd) >> Tasks: 1 >> CGroup: /system.slice/wdmd.service >> └─3763 /usr/sbin/wdmd >> >> Jan 15 10:38:12 ov4301.localdomain.local systemd[1]: Starting Watchdog >> Multiplexing Daemon... >> Jan 15 10:38:13 ov4301.localdomain.local systemd-wdmd[3707]: Loading the >> softdog kernel module: ...] >> Jan 15 10:38:13 ov4301.localdomain.local wdmd[3759]: group 'sanlock' not >> found, using socket gid: 0 >> > > Looks like the host is not configured properly. > > Running vdsm-tool configure --force should fix this, and must be part of > the install process. > Has this been fixed? If not, is this tracked in a BZ? > > >> Jan 15 10:38:13 ov4301.localdomain.local wdmd[3763]: wdmd started S0 H1 G0 >> Jan 15 10:38:13 ov4301.localdomain.local wdmd[3763]: /dev/watchdog0 armed >> with fire_timeout 60 >> Jan 15 10:38:13 ov4301.localdomain.local systemd[1]: Started Watchdog >> Multiplexing Daemon. >> Hint: Some lines were ellipsized, use -l to show in full. >> [root@ov4301 ~]# >> >> while for sanlock: >> >> [root@ov4301 ~]# cat /var/log/sanlock.log >> 2019-01-15 10:38:13 6 [3721]: sanlock daemon started 3.6.0 host >> 8fd6d41c-99e8-4c3a-8212-68dd1856927c.ov4301.loc >> 2019-01-15 10:38:18 11 [3721]: helper pid 3725 dead wait 0 >> 2019-01-15 12:54:56 8211 [26802]: sanlock daemon started 3.6.0 host >> 4dc13694-53a6-41dc-93b2-8ce371e903f5.ov4301.loc >> 2019-01-15 12:54:57 8211 [26802]: set scheduler RR|RESET_ON_FORK priority >> 99 failed: Operation not permitted >> 2019-01-15 13:50:34 11549 [26810]: s1 lockspace >> 14ec2fc7-8c2b-487c-8f4f-428644650928:250:/rhev/data-center/mnt/glusterSD/192.168.123.211: >> _engine/14ec2fc7-8c2b-487c-8f4f-428644650928/dom_md/ids:0 >> 2019-01-15 13:50:34 11549 [21290]: s1 wdmd_connect failed -13 >> 2019-01-15 13:50:34 11549 [21290]: s1 connect_watchdog failed -1 >> 2019-01-15 13:50:35 11550 [26810]: s1 add_lockspace fail result -203 >> [root@ov4301 ~]# >> >> The question is: why these daemons start (with error) upon the node boot >> after install? >> Or is it a bug that the groups are not in place yet and I can restart in >> the inital step (as for ovirt-imageio-daemon) other daemons too? >> >> >> >> _______________________________________________ >> Users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> https://lists.ovirt.org/archives/list/[email protected]/message/ZHLDIPDFQ5PBCQMD2A7IAAAPG3S3SJ2G/ >> > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/RDYYVHJ6EZCYLQATTNQ7RDPJFHHRGM65/ > -- SANDRO BONAZZOLA MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV Red Hat EMEA <https://www.redhat.com/> [email protected] <https://red.ht/sig>
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/A3CY2BIVS3RU2X3D3V4EROZHA7R6OZKM/

