Hi Maor, My disk are of 4K block size and from this bug seems that gluster replica needs 512B block size. Is there a way to make gluster function with 4K drives?
Thank you! On Sun, Jun 4, 2017 at 2:34 PM, Maor Lipchuk <[email protected]> wrote: > Hi Alex, > > I saw a bug that might be related to the issue you encountered at > https://bugzilla.redhat.com/show_bug.cgi?id=1386443 > > Sahina, maybe you have any advise? Do you think that BZ1386443is related? > > Regards, > Maor > > On Sat, Jun 3, 2017 at 8:45 PM, Abi Askushi <[email protected]> > wrote: > > Hi All, > > > > I have installed successfully several times oVirt (version 4.1) with 3 > nodes > > on top glusterfs. > > > > This time, when trying to configure the same setup, I am facing the > > following issue which doesn't seem to go away. During installation i get > the > > error: > > > > Failed to execute stage 'Misc configuration': Cannot acquire host id: > > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock > > lockspace add failure', 'Invalid argument')) > > > > The only different in this setup is that instead of standard > partitioning i > > have GPT partitioning and the disks have 4K block size instead of 512. > > > > The /var/log/sanlock.log has the following lines: > > > > 2017-06-03 19:21:15+0200 23450 [943]: s9 lockspace > > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:250:/rhev/data- > center/mnt/_var_lib_ovirt-hosted-engin-setup_tmptjkIDI/ > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/ids:0 > > 2017-06-03 19:21:36+0200 23471 [944]: s9:r5 resource > > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047:SDM:/rhev/data- > center/mnt/_var_lib_ovirt-hosted-engine-setup_tmptjkIDI/ > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047/dom_md/leases:1048576 > > for 2,9,23040 > > 2017-06-03 19:21:36+0200 23471 [943]: s10 lockspace > > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922:250:/rhev/data- > center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f- > 4838-8e26-c8b4d5e5e922/dom_md/ids:0 > > 2017-06-03 19:21:36+0200 23471 [23522]: a5a6b0e7 aio collect RD > > 0x7f59b00008c0:0x7f59b00008d0:0x7f59b0101000 result -22:0 match res > > 2017-06-03 19:21:36+0200 23471 [23522]: read_sectors delta_leader offset > > 127488 rv -22 > > /rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/a5a6b0e7-fc3f-4838- > 8e26-c8b4d5e5e922/dom_md/ids > > 2017-06-03 19:21:37+0200 23472 [930]: s9 host 250 1 23450 > > 88c2244c-a782-40ed-9560-6cfa4d46f853.v0.neptune > > 2017-06-03 19:21:37+0200 23472 [943]: s10 add_lockspace fail result -22 > > > > And /var/log/vdsm/vdsm.log says: > > > > 2017-06-03 19:19:38,176+0200 WARN (jsonrpc/3) > > [storage.StorageServer.MountConnection] Using user specified > > backup-volfile-servers option (storageServer:253) > > 2017-06-03 19:21:12,379+0200 WARN (periodic/1) [throttled] MOM not > > available. (throttledlog:105) > > 2017-06-03 19:21:12,380+0200 WARN (periodic/1) [throttled] MOM not > > available, KSM stats will be missing. (throttledlog:105) > > 2017-06-03 19:21:14,714+0200 WARN (jsonrpc/1) > > [storage.StorageServer.MountConnection] Using user specified > > backup-volfile-servers option (storageServer:253) > > 2017-06-03 19:21:15,515+0200 ERROR (jsonrpc/4) [storage.initSANLock] > Cannot > > initialize SANLock for domain a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922 > > (clusterlock:238) > > Traceback (most recent call last): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", > line > > 234, in initSANLock > > sanlock.init_lockspace(sdUUID, idsPath) > > SanlockException: (107, 'Sanlock lockspace init failure', 'Transport > > endpoint is not connected') > > 2017-06-03 19:21:15,515+0200 WARN (jsonrpc/4) > > [storage.StorageDomainManifest] lease did not initialize successfully > > (sd:557) > > Traceback (most recent call last): > > File "/usr/share/vdsm/storage/sd.py", line 552, in initDomainLock > > self._domainLock.initLock(self.getDomainLease()) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", > line > > 271, in initLock > > initSANLock(self._sdUUID, self._idsPath, lease) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", > line > > 239, in initSANLock > > raise se.ClusterLockInitError() > > ClusterLockInitError: Could not initialize cluster lock: () > > 2017-06-03 19:21:37,867+0200 ERROR (jsonrpc/2) [storage.StoragePool] > Create > > pool hosted_datacenter canceled (sp:655) > > Traceback (most recent call last): > > File "/usr/share/vdsm/storage/sp.py", line 652, in create > > self.attachSD(sdUUID) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", > line > > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/share/vdsm/storage/sp.py", line 971, in attachSD > > dom.acquireHostId(self.id) > > File "/usr/share/vdsm/storage/sd.py", line 790, in acquireHostId > > self._manifest.acquireHostId(hostId, async) > > File "/usr/share/vdsm/storage/sd.py", line 449, in acquireHostId > > self._domainLock.acquireHostId(hostId, async) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", > line > > 297, in acquireHostId > > raise se.AcquireHostIdFailure(self._sdUUID, e) > > AcquireHostIdFailure: Cannot acquire host id: > > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock > > lockspace add failure', 'Invalid argument')) > > 2017-06-03 19:21:37,870+0200 ERROR (jsonrpc/2) [storage.StoragePool] > Domain > > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 detach from MSD > > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 Ver 1 failed. (sp:528) > > Traceback (most recent call last): > > File "/usr/share/vdsm/storage/sp.py", line 525, in __cleanupDomains > > self.detachSD(sdUUID) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", > line > > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/share/vdsm/storage/sp.py", line 1046, in detachSD > > raise se.CannotDetachMasterStorageDomain(sdUUID) > > CannotDetachMasterStorageDomain: Illegal action: > > (u'ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047',) > > 2017-06-03 19:21:37,872+0200 ERROR (jsonrpc/2) [storage.StoragePool] > Domain > > a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922 detach from MSD > > ba6bd862-c2b8-46e7-b2c8-91e4a5bb2047 Ver 1 failed. (sp:528) > > Traceback (most recent call last): > > File "/usr/share/vdsm/storage/sp.py", line 525, in __cleanupDomains > > self.detachSD(sdUUID) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", > line > > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/share/vdsm/storage/sp.py", line 1043, in detachSD > > self.validateAttachedDomain(dom) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", > line > > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/share/vdsm/storage/sp.py", line 542, in > validateAttachedDomain > > self.validatePoolSD(dom.sdUUID) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", > line > > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/share/vdsm/storage/sp.py", line 535, in validatePoolSD > > raise se.StorageDomainNotMemberOfPool(self.spUUID, sdUUID) > > StorageDomainNotMemberOfPool: Domain is not member in pool: > > u'pool=a1e7e9dd-0cf4-41ae-ba13-36297ed66309, > > domain=a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922' > > 2017-06-03 19:21:40,063+0200 ERROR (jsonrpc/2) [storage.TaskManager.Task] > > (Task='a2476a33-26f8-4ebd-876d-02fe5d13ef78') Unexpected error > (task:870) > > Traceback (most recent call last): > > File "/usr/share/vdsm/storage/task.py", line 877, in _run > > return fn(*args, **kargs) > > File "/usr/lib/python2.7/site-packages/vdsm/logUtils.py", line 52, in > > wrapper > > res = f(*args, **kwargs) > > File "/usr/share/vdsm/storage/hsm.py", line 959, in createStoragePool > > leaseParams) > > File "/usr/share/vdsm/storage/sp.py", line 652, in create > > self.attachSD(sdUUID) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", > line > > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/share/vdsm/storage/sp.py", line 971, in attachSD > > dom.acquireHostId(self.id) > > File "/usr/share/vdsm/storage/sd.py", line 790, in acquireHostId > > self._manifest.acquireHostId(hostId, async) > > File "/usr/share/vdsm/storage/sd.py", line 449, in acquireHostId > > self._domainLock.acquireHostId(hostId, async) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/clusterlock.py", > line > > 297, in acquireHostId > > raise se.AcquireHostIdFailure(self._sdUUID, e) > > AcquireHostIdFailure: Cannot acquire host id: > > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock > > lockspace add failure', 'Invalid argument')) > > 2017-06-03 19:21:40,067+0200 ERROR (jsonrpc/2) [storage.Dispatcher] > > {'status': {'message': "Cannot acquire host id: > > (u'a5a6b0e7-fc3f-4838-8e26-c8b4d5e5e922', SanlockException(22, 'Sanlock > > lockspace add failure', 'Invalid argument'))", 'code': 661}} > (dispatcher:77) > > > > The gluster volume prepared for engine storage is online and no split > brain > > is reported. I don't understand what needs to be done to overcome this. > Any > > idea will be appreciated. > > > > Thank you, > > Alex > > > > _______________________________________________ > > Users mailing list > > [email protected] > > http://lists.ovirt.org/mailman/listinfo/users > > >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

