The root cause seems to be a SANlock issue: Thread-32382::ERROR:: 2013-09-20 13:16:34,126::clusterlock::145::initSANLock: initSANLock) Cannot initialize SANLock for domain 17d21ac7-5859-4f25-8de7-2a9433d50c11 Traceback (most recent call last): File "/usr/share/vdsm/storage/clusterlock.py", line 140, in initSANLock sanlock.init_lockspace(sdUUID, idsPath) SanlockException: (22, 'Sanlock lockspace init failure', 'Invalid argument') Thread-32382::WARNING:: 2013-09-20 13:16:34,127::sd::428::Storage.StorageDomain: initSPMlease) lease did not initialize successfully Traceback (most recent call last): File "/usr/share/vdsm/storage/sd.py", line 423, in initSPMlease self._clusterLock.initLock() File "/usr/share/vdsm/storage/clusterlock.py", line 163, in initLock initSANLock(self._sdUUID, self._idsPath, self._leasesPath) File "/usr/share/vdsm/storage/clusterlock.py", line 146, in initSANLock raise se.ClusterLockInitError() ClusterLockInitError: Could not initialize cluster lock: ()
Can you include /var/log/sanlock.log and /var/log/messages please? ----- Original Message ----- > From: "Steve Dainard" <[email protected]> > To: "Deepak C Shetty" <[email protected]> > Cc: "users" <[email protected]> > Sent: Friday, September 20, 2013 8:30:59 PM > Subject: Re: [Users] Ovirt 3.3 Fedora 19 add gluster storage permissions > error > Awesome, thanks guys. Its weird that that article tells you to set with > 'key=value' rather than 'key value' must be some legacy stuff. > Once those changes are in place I hit a different error. Deepak, maybe you've > seen this one on new storage domain add: > [root@ovirt-manager2 ~]# tail -f /var/log/ovirt-engine/engine.log > 2013-09-20 13:16:36,226 ERROR > [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] > (ajp--127.0.0.1-8702-9) Command CreateStoragePoolVDS execution failed. > Exception: VDSErrorException: VDSGenericException: VDSErrorException: Failed > to CreateStoragePoolVDS, error = Cannot acquire host id: > ('17d21ac7-5859-4f25-8de7-2a9433d50c11', SanlockException(22, 'Sanlock > lockspace add failure', 'Invalid argument')) > 2013-09-20 13:16:36,229 INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand] > (ajp--127.0.0.1-8702-9) FINISH, CreateStoragePoolVDSCommand, log id: > 672635cc > 2013-09-20 13:16:36,231 ERROR > [org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand] > (ajp--127.0.0.1-8702-9) Command > org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand throw > Vdc Bll exception. With error message VdcBLLException: > org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException: > VDSGenericException: VDSErrorException: Failed to CreateStoragePoolVDS, > error = Cannot acquire host id: ('17d21ac7-5859-4f25-8de7-2a9433d50c11', > SanlockException(22, 'Sanlock lockspace add failure', 'Invalid argument')) > (Failed with VDSM error AcquireHostIdFailure and code 661) > 2013-09-20 13:16:36,296 INFO > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (ajp--127.0.0.1-8702-9) Correlation ID: 11070337, Call Stack: null, Custom > Event ID: -1, Message: Failed to attach Storage Domains to Data Center > Default. (User: admin@internal) > 2013-09-20 13:16:36,299 INFO > [org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand] > (ajp--127.0.0.1-8702-9) Lock freed to object EngineLock [exclusiveLocks= > key: 5849b030-626e-47cb-ad90-3ce782d831b3 value: POOL > , sharedLocks= ] > 2013-09-20 13:16:36,387 INFO > [org.ovirt.engine.core.bll.storage.AttachStorageDomainToPoolCommand] > (ajp--127.0.0.1-8702-9) Command [id=293a1e97-e949-4c17-92c6-c01f2221204e]: > Compensating CHANGED_ENTITY of > org.ovirt.engine.core.common.businessentities.StoragePool; snapshot: > id=5849b030-626e-47cb-ad90-3ce782d831b3. > 2013-09-20 13:16:36,398 INFO > [org.ovirt.engine.core.bll.storage.AttachStorageDomainToPoolCommand] > (ajp--127.0.0.1-8702-9) Command [id=293a1e97-e949-4c17-92c6-c01f2221204e]: > Compensating NEW_ENTITY_ID of > org.ovirt.engine.core.common.businessentities.StoragePoolIsoMap; snapshot: > storagePoolId = 5849b030-626e-47cb-ad90-3ce782d831b3, storageId = > 17d21ac7-5859-4f25-8de7-2a9433d50c11. > 2013-09-20 13:16:36,425 INFO > [org.ovirt.engine.core.bll.storage.AttachStorageDomainToPoolCommand] > (ajp--127.0.0.1-8702-9) Command [id=293a1e97-e949-4c17-92c6-c01f2221204e]: > Compensating CHANGED_ENTITY of > org.ovirt.engine.core.common.businessentities.StorageDomainStatic; snapshot: > id=17d21ac7-5859-4f25-8de7-2a9433d50c11. > 2013-09-20 13:16:36,464 INFO > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (ajp--127.0.0.1-8702-9) Correlation ID: 302ae6eb, Job ID: > 014ec59b-e6d7-4e5e-b588-4fb0dfa8f1c8, Call Stack: null, Custom Event ID: -1, > Message: Failed to attach Storage Domain rep2-virt to Data Center Default. > (User: admin@internal) > [root@ovirt001 ~]# tail -f /var/log/vdsm/vdsm.log > Thread-32374::DEBUG::2013-09-20 > 13:16:18,107::task::579::TaskManager.Task::(_updateState) > Task=`f4cab975-d5fa-463a-990e-ab32686c6806`::moving from state init -> state > preparing > Thread-32374::INFO::2013-09-20 > 13:16:18,107::logUtils::44::dispatcher::(wrapper) Run and protect: > repoStats(options=None) > Thread-32374::INFO::2013-09-20 > 13:16:18,107::logUtils::47::dispatcher::(wrapper) Run and protect: > repoStats, Return response: {} > Thread-32374::DEBUG::2013-09-20 > 13:16:18,108::task::1168::TaskManager.Task::(prepare) > Task=`f4cab975-d5fa-463a-990e-ab32686c6806`::finished: {} > Thread-32374::DEBUG::2013-09-20 > 13:16:18,108::task::579::TaskManager.Task::(_updateState) > Task=`f4cab975-d5fa-463a-990e-ab32686c6806`::moving from state preparing -> > state finished > Thread-32374::DEBUG::2013-09-20 > 13:16:18,108::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-32374::DEBUG::2013-09-20 > 13:16:18,108::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32374::DEBUG::2013-09-20 > 13:16:18,108::task::974::TaskManager.Task::(_decref) > Task=`f4cab975-d5fa-463a-990e-ab32686c6806`::ref 0 aborting False > Thread-32379::DEBUG::2013-09-20 > 13:16:29,509::BindingXMLRPC::177::vds::(wrapper) client [10.0.0.34] > Thread-32379::DEBUG::2013-09-20 > 13:16:29,510::task::579::TaskManager.Task::(_updateState) > Task=`1ad55ba1-afaa-4524-b3fa-3d55a421e8bc`::moving from state init -> state > preparing > Thread-32379::INFO::2013-09-20 > 13:16:29,510::logUtils::44::dispatcher::(wrapper) Run and protect: > connectStorageServer(domType=7, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', > 'connection': '192.168.1.1:rep2-virt', 'iqn': '', 'portal': '', 'user': '', > 'vfs_type': 'glusterfs', 'password': '******', 'id': > '00000000-0000-0000-0000-000000000000'}], options=None) > Thread-32379::DEBUG::2013-09-20 > 13:16:29,516::hsm::2333::Storage.HSM::(__prefetchDomains) glusterDomPath: > glusterSD/* > Thread-32379::DEBUG::2013-09-20 > 13:16:29,523::hsm::2345::Storage.HSM::(__prefetchDomains) Found SD uuids: () > Thread-32379::DEBUG::2013-09-20 > 13:16:29,523::hsm::2396::Storage.HSM::(connectStorageServer) knownSDs: > {b72b61d1-e11c-496d-ad3a-6f566a1f0ad1: storage.glusterSD.findDomain, > 983c4aa1-7b00-4d3b-b6ad-1fd2cf9297ce: storage.glusterSD.findDomain, > b91afb39-f96e-4eb3-bc6c-9f08fa16869c: storage.glusterSD.findDomain} > Thread-32379::INFO::2013-09-20 > 13:16:29,523::logUtils::47::dispatcher::(wrapper) Run and protect: > connectStorageServer, Return response: {'statuslist': [{'status': 0, 'id': > '00000000-0000-0000-0000-000000000000'}]} > Thread-32379::DEBUG::2013-09-20 > 13:16:29,523::task::1168::TaskManager.Task::(prepare) > Task=`1ad55ba1-afaa-4524-b3fa-3d55a421e8bc`::finished: {'statuslist': > [{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]} > Thread-32379::DEBUG::2013-09-20 > 13:16:29,524::task::579::TaskManager.Task::(_updateState) > Task=`1ad55ba1-afaa-4524-b3fa-3d55a421e8bc`::moving from state preparing -> > state finished > Thread-32379::DEBUG::2013-09-20 > 13:16:29,524::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-32379::DEBUG::2013-09-20 > 13:16:29,524::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32379::DEBUG::2013-09-20 > 13:16:29,524::task::974::TaskManager.Task::(_decref) > Task=`1ad55ba1-afaa-4524-b3fa-3d55a421e8bc`::ref 0 aborting False > Thread-32382::DEBUG::2013-09-20 > 13:16:29,888::BindingXMLRPC::177::vds::(wrapper) client [10.0.0.34] > Thread-32382::DEBUG::2013-09-20 > 13:16:29,888::task::579::TaskManager.Task::(_updateState) > Task=`a3ba925b-65ee-42a6-8506-927a06f63995`::moving from state init -> state > preparing > Thread-32382::INFO::2013-09-20 > 13:16:29,889::logUtils::44::dispatcher::(wrapper) Run and protect: > createStorageDomain(storageType=7, > sdUUID='17d21ac7-5859-4f25-8de7-2a9433d50c11', domainName='rep2-virt', > typeSpecificArg='192.168.1.1:rep2-virt', domClass=1, domVersion='3', > options=None) > Thread-32382::DEBUG::2013-09-20 > 13:16:29,889::misc::807::SamplingMethod::(__call__) Trying to enter sampling > method (storage.sdc.refreshStorage) > Thread-32382::DEBUG::2013-09-20 > 13:16:29,889::misc::809::SamplingMethod::(__call__) Got in to sampling > method > Thread-32382::DEBUG::2013-09-20 > 13:16:29,889::misc::807::SamplingMethod::(__call__) Trying to enter sampling > method (storage.iscsi.rescan) > Thread-32382::DEBUG::2013-09-20 > 13:16:29,889::misc::809::SamplingMethod::(__call__) Got in to sampling > method > Thread-32382::DEBUG::2013-09-20 > 13:16:29,889::iscsiadm::91::Storage.Misc.excCmd::(_runCmd) '/usr/bin/sudo -n > /sbin/iscsiadm -m session -R' (cwd None) > Thread-32382::DEBUG::2013-09-20 > 13:16:29,904::iscsiadm::91::Storage.Misc.excCmd::(_runCmd) FAILED: <err> = > 'iscsiadm: No session found.\n'; <rc> = 21 > Thread-32382::DEBUG::2013-09-20 > 13:16:29,904::misc::817::SamplingMethod::(__call__) Returning last result > Thread-32382::DEBUG::2013-09-20 > 13:16:32,931::multipath::111::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo > -n /sbin/multipath' (cwd None) > Thread-32382::DEBUG::2013-09-20 > 13:16:33,255::multipath::111::Storage.Misc.excCmd::(rescan) SUCCESS: <err> = > ''; <rc> = 0 > Thread-32382::DEBUG::2013-09-20 > 13:16:33,256::lvm::483::OperationMutex::(_invalidateAllPvs) Operation 'lvm > invalidate operation' got the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,256::lvm::485::OperationMutex::(_invalidateAllPvs) Operation 'lvm > invalidate operation' released the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,256::lvm::494::OperationMutex::(_invalidateAllVgs) Operation 'lvm > invalidate operation' got the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,256::lvm::496::OperationMutex::(_invalidateAllVgs) Operation 'lvm > invalidate operation' released the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,257::lvm::514::OperationMutex::(_invalidateAllLvs) Operation 'lvm > invalidate operation' got the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,257::lvm::516::OperationMutex::(_invalidateAllLvs) Operation 'lvm > invalidate operation' released the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,257::misc::817::SamplingMethod::(__call__) Returning last result > Thread-32382::ERROR::2013-09-20 > 13:16:33,257::sdc::137::Storage.StorageDomainCache::(_findDomain) looking > for unfetched domain 17d21ac7-5859-4f25-8de7-2a9433d50c11 > Thread-32382::ERROR::2013-09-20 > 13:16:33,257::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDomain) > looking for domain 17d21ac7-5859-4f25-8de7-2a9433d50c11 > Thread-32382::DEBUG::2013-09-20 > 13:16:33,258::lvm::374::OperationMutex::(_reloadvgs) Operation 'lvm reload > operation' got the operation mutex > Thread-32382::DEBUG::2013-09-20 > 13:16:33,259::lvm::311::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n > /sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"] > ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3 > filter = [ \'r|.*|\' ] } global { locking_type=1 prioritise_write_locks=1 > wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } " --noheadings > --units b --nosuffix --separator | -o > uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free > 17d21ac7-5859-4f25-8de7-2a9433d50c11' (cwd None) > Thread-32382::DEBUG::2013-09-20 > 13:16:33,285::lvm::311::Storage.Misc.excCmd::(cmd) FAILED: <err> = ' Volume > group "17d21ac7-5859-4f25-8de7-2a9433d50c11" not found\n'; <rc> = 5 > Thread-32382::WARNING::2013-09-20 > 13:16:33,286::lvm::379::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] [' > Volume group "17d21ac7-5859-4f25-8de7-2a9433d50c11" not found'] > Thread-32382::DEBUG::2013-09-20 > 13:16:33,286::lvm::403::OperationMutex::(_reloadvgs) Operation 'lvm reload > operation' released the operation mutex > Thread-32382::ERROR::2013-09-20 > 13:16:33,295::sdc::143::Storage.StorageDomainCache::(_findDomain) domain > 17d21ac7-5859-4f25-8de7-2a9433d50c11 not found > Traceback (most recent call last): > File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain > dom = findMethod(sdUUID) > File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain > raise se.StorageDomainDoesNotExist(sdUUID) > StorageDomainDoesNotExist: Storage domain does not exist: > ('17d21ac7-5859-4f25-8de7-2a9433d50c11',) > Thread-32382::INFO::2013-09-20 > 13:16:33,295::nfsSD::69::Storage.StorageDomain::(create) > sdUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11 domainName=rep2-virt > remotePath=192.168.1.1:rep2-virt domClass=1 > Thread-32382::DEBUG::2013-09-20 > 13:16:33,430::persistentDict::192::Storage.PersistentDict::(__init__) > Created a persistent dict with FileMetadataRW backend > Thread-32382::DEBUG::2013-09-20 > 13:16:33,445::persistentDict::234::Storage.PersistentDict::(refresh) read > lines (FileMetadataRW)=[] > Thread-32382::WARNING::2013-09-20 > 13:16:33,446::persistentDict::256::Storage.PersistentDict::(refresh) data > has no embedded checksum - trust it as it is > Thread-32382::DEBUG::2013-09-20 > 13:16:33,447::persistentDict::167::Storage.PersistentDict::(transaction) > Starting transaction > Thread-32382::DEBUG::2013-09-20 > 13:16:33,448::persistentDict::173::Storage.PersistentDict::(transaction) > Flushing changes > Thread-32382::DEBUG::2013-09-20 > 13:16:33,449::persistentDict::299::Storage.PersistentDict::(flush) about to > write lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=rep2-virt', > 'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5', 'LOCKPOLICY=', > 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', > 'REMOTE_PATH=192.168.1.1:rep2-virt', 'ROLE=Regular', > 'SDUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11', 'TYPE=GLUSTERFS', > 'VERSION=3', '_SHA_CKSUM=2b07fbc8c65e20eef5180ab785016bde543c6746'] > Thread-32382::DEBUG::2013-09-20 > 13:16:33,454::persistentDict::175::Storage.PersistentDict::(transaction) > Finished transaction > Thread-32382::DEBUG::2013-09-20 > 13:16:33,457::fileSD::153::Storage.StorageDomain::(__init__) Reading domain > in path > /rhev/data-center/mnt/glusterSD/192.168.1.1:rep2-virt/17d21ac7-5859-4f25-8de7-2a9433d50c11 > Thread-32382::DEBUG::2013-09-20 > 13:16:33,457::persistentDict::192::Storage.PersistentDict::(__init__) > Created a persistent dict with FileMetadataRW backend > Thread-32382::DEBUG::2013-09-20 > 13:16:33,469::persistentDict::234::Storage.PersistentDict::(refresh) read > lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=rep2-virt', > 'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5', 'LOCKPOLICY=', > 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', > 'REMOTE_PATH=192.168.1.1:rep2-virt', 'ROLE=Regular', > 'SDUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11', 'TYPE=GLUSTERFS', > 'VERSION=3', '_SHA_CKSUM=2b07fbc8c65e20eef5180ab785016bde543c6746'] > Thread-32382::DEBUG::2013-09-20 > 13:16:33,472::fileSD::535::Storage.StorageDomain::(imageGarbageCollector) > Removing remnants of deleted images [] > Thread-32382::DEBUG::2013-09-20 > 13:16:33,472::resourceManager::420::ResourceManager::(registerNamespace) > Registering namespace '17d21ac7-5859-4f25-8de7-2a9433d50c11_imageNS' > Thread-32382::DEBUG::2013-09-20 > 13:16:33,473::resourceManager::420::ResourceManager::(registerNamespace) > Registering namespace '17d21ac7-5859-4f25-8de7-2a9433d50c11_volumeNS' > Thread-32382::DEBUG::2013-09-20 > 13:16:33,473::clusterlock::137::initSANLock::(initSANLock) Initializing > SANLock for domain 17d21ac7-5859-4f25-8de7-2a9433d50c11 > Thread-32387::DEBUG::2013-09-20 > 13:16:33,717::task::579::TaskManager.Task::(_updateState) > Task=`0e11c8e5-4e40-4d28-9eaf-129db67b2f4d`::moving from state init -> state > preparing > Thread-32387::INFO::2013-09-20 > 13:16:33,718::logUtils::44::dispatcher::(wrapper) Run and protect: > repoStats(options=None) > Thread-32387::INFO::2013-09-20 > 13:16:33,718::logUtils::47::dispatcher::(wrapper) Run and protect: > repoStats, Return response: {} > Thread-32387::DEBUG::2013-09-20 > 13:16:33,718::task::1168::TaskManager.Task::(prepare) > Task=`0e11c8e5-4e40-4d28-9eaf-129db67b2f4d`::finished: {} > Thread-32387::DEBUG::2013-09-20 > 13:16:33,718::task::579::TaskManager.Task::(_updateState) > Task=`0e11c8e5-4e40-4d28-9eaf-129db67b2f4d`::moving from state preparing -> > state finished > Thread-32387::DEBUG::2013-09-20 > 13:16:33,718::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-32387::DEBUG::2013-09-20 > 13:16:33,718::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32387::DEBUG::2013-09-20 > 13:16:33,719::task::974::TaskManager.Task::(_decref) > Task=`0e11c8e5-4e40-4d28-9eaf-129db67b2f4d`::ref 0 aborting False > Thread-32382::ERROR::2013-09-20 > 13:16:34,126::clusterlock::145::initSANLock::(initSANLock) Cannot initialize > SANLock for domain 17d21ac7-5859-4f25-8de7-2a9433d50c11 > Traceback (most recent call last): > File "/usr/share/vdsm/storage/clusterlock.py", line 140, in initSANLock > sanlock.init_lockspace(sdUUID, idsPath) > SanlockException: (22, 'Sanlock lockspace init failure', 'Invalid argument') > Thread-32382::WARNING::2013-09-20 > 13:16:34,127::sd::428::Storage.StorageDomain::(initSPMlease) lease did not > initialize successfully > Traceback (most recent call last): > File "/usr/share/vdsm/storage/sd.py", line 423, in initSPMlease > self._clusterLock.initLock() > File "/usr/share/vdsm/storage/clusterlock.py", line 163, in initLock > initSANLock(self._sdUUID, self._idsPath, self._leasesPath) > File "/usr/share/vdsm/storage/clusterlock.py", line 146, in initSANLock > raise se.ClusterLockInitError() > ClusterLockInitError: Could not initialize cluster lock: () > Thread-32382::DEBUG::2013-09-20 > 13:16:34,127::hsm::2624::Storage.HSM::(createStorageDomain) knownSDs: > {b72b61d1-e11c-496d-ad3a-6f566a1f0ad1: storage.glusterSD.findDomain, > 983c4aa1-7b00-4d3b-b6ad-1fd2cf9297ce: storage.glusterSD.findDomain, > b91afb39-f96e-4eb3-bc6c-9f08fa16869c: storage.glusterSD.findDomain, > 17d21ac7-5859-4f25-8de7-2a9433d50c11: storage.glusterSD.findDomain} > Thread-32382::INFO::2013-09-20 > 13:16:34,128::logUtils::47::dispatcher::(wrapper) Run and protect: > createStorageDomain, Return response: None > Thread-32382::DEBUG::2013-09-20 > 13:16:34,128::task::1168::TaskManager.Task::(prepare) > Task=`a3ba925b-65ee-42a6-8506-927a06f63995`::finished: None > Thread-32382::DEBUG::2013-09-20 > 13:16:34,128::task::579::TaskManager.Task::(_updateState) > Task=`a3ba925b-65ee-42a6-8506-927a06f63995`::moving from state preparing -> > state finished > Thread-32382::DEBUG::2013-09-20 > 13:16:34,128::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-32382::DEBUG::2013-09-20 > 13:16:34,128::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32382::DEBUG::2013-09-20 > 13:16:34,129::task::974::TaskManager.Task::(_decref) > Task=`a3ba925b-65ee-42a6-8506-927a06f63995`::ref 0 aborting False > Thread-32389::DEBUG::2013-09-20 > 13:16:34,219::BindingXMLRPC::177::vds::(wrapper) client [10.0.0.34] > Thread-32389::DEBUG::2013-09-20 > 13:16:34,219::task::579::TaskManager.Task::(_updateState) > Task=`a0d5d4d6-dcb7-4293-bf8a-cf1e2204f586`::moving from state init -> state > preparing > Thread-32389::INFO::2013-09-20 > 13:16:34,220::logUtils::44::dispatcher::(wrapper) Run and protect: > getStorageDomainStats(sdUUID='17d21ac7-5859-4f25-8de7-2a9433d50c11', > options=None) > Thread-32389::DEBUG::2013-09-20 > 13:16:34,220::resourceManager::197::ResourceManager.Request::(__init__) > ResName=`Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11`ReqID=`9f37d808-9ad2-4c06-99ef-449b43049e80`::Request > was made in '/usr/share/vdsm/storage/hsm.py' line '2784' at > 'getStorageDomainStats' > Thread-32389::DEBUG::2013-09-20 > 13:16:34,220::resourceManager::541::ResourceManager::(registerResource) > Trying to register resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' > for lock type 'shared' > Thread-32389::DEBUG::2013-09-20 > 13:16:34,221::resourceManager::600::ResourceManager::(registerResource) > Resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' is free. Now locking > as 'shared' (1 active user) > Thread-32389::DEBUG::2013-09-20 > 13:16:34,221::resourceManager::237::ResourceManager.Request::(grant) > ResName=`Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11`ReqID=`9f37d808-9ad2-4c06-99ef-449b43049e80`::Granted > request > Thread-32389::DEBUG::2013-09-20 > 13:16:34,221::task::811::TaskManager.Task::(resourceAcquired) > Task=`a0d5d4d6-dcb7-4293-bf8a-cf1e2204f586`::_resourcesAcquired: > Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11 (shared) > Thread-32389::DEBUG::2013-09-20 > 13:16:34,221::task::974::TaskManager.Task::(_decref) > Task=`a0d5d4d6-dcb7-4293-bf8a-cf1e2204f586`::ref 1 aborting False > Thread-32389::INFO::2013-09-20 > 13:16:34,223::logUtils::47::dispatcher::(wrapper) Run and protect: > getStorageDomainStats, Return response: {'stats': {'mdasize': 0, > 'mdathreshold': True, 'mdavalid': True, 'diskfree': '14182986809344', > 'disktotal': '14199600185344', 'mdafree': 0}} > Thread-32389::DEBUG::2013-09-20 > 13:16:34,223::task::1168::TaskManager.Task::(prepare) > Task=`a0d5d4d6-dcb7-4293-bf8a-cf1e2204f586`::finished: {'stats': {'mdasize': > 0, 'mdathreshold': True, 'mdavalid': True, 'diskfree': '14182986809344', > 'disktotal': '14199600185344', 'mdafree': 0}} > Thread-32389::DEBUG::2013-09-20 > 13:16:34,223::task::579::TaskManager.Task::(_updateState) > Task=`a0d5d4d6-dcb7-4293-bf8a-cf1e2204f586`::moving from state preparing -> > state finished > Thread-32389::DEBUG::2013-09-20 > 13:16:34,223::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources > {'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11': < ResourceRef > 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11', isValid: 'True' obj: > 'None'>} > Thread-32389::DEBUG::2013-09-20 > 13:16:34,224::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32389::DEBUG::2013-09-20 > 13:16:34,224::resourceManager::615::ResourceManager::(releaseResource) > Trying to release resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' > Thread-32389::DEBUG::2013-09-20 > 13:16:34,224::resourceManager::634::ResourceManager::(releaseResource) > Released resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' (0 active > users) > Thread-32389::DEBUG::2013-09-20 > 13:16:34,224::resourceManager::640::ResourceManager::(releaseResource) > Resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' is free, finding out > if anyone is waiting for it. > Thread-32389::DEBUG::2013-09-20 > 13:16:34,224::resourceManager::648::ResourceManager::(releaseResource) No > one is waiting for resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11', > Clearing records. > Thread-32389::DEBUG::2013-09-20 > 13:16:34,225::task::974::TaskManager.Task::(_decref) > Task=`a0d5d4d6-dcb7-4293-bf8a-cf1e2204f586`::ref 0 aborting False > Thread-32390::DEBUG::2013-09-20 > 13:16:35,099::BindingXMLRPC::177::vds::(wrapper) client [10.0.0.34] > Thread-32390::DEBUG::2013-09-20 > 13:16:35,099::task::579::TaskManager.Task::(_updateState) > Task=`089c5f71-9cbb-4626-8f2d-3ed3547a98cd`::moving from state init -> state > preparing > Thread-32390::INFO::2013-09-20 > 13:16:35,099::logUtils::44::dispatcher::(wrapper) Run and protect: > connectStorageServer(domType=7, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', > 'connection': '192.168.1.1:rep2-virt', 'iqn': '', 'portal': '', 'user': '', > 'vfs_type': 'glusterfs', 'password': '******', 'id': > 'cecee482-87e1-4ecc-8bda-0e0ec84d7792'}], options=None) > Thread-32390::DEBUG::2013-09-20 > 13:16:35,105::hsm::2333::Storage.HSM::(__prefetchDomains) glusterDomPath: > glusterSD/* > Thread-32390::DEBUG::2013-09-20 > 13:16:35,112::hsm::2345::Storage.HSM::(__prefetchDomains) Found SD uuids: > ('17d21ac7-5859-4f25-8de7-2a9433d50c11', > '17d21ac7-5859-4f25-8de7-2a9433d50c11') > Thread-32390::DEBUG::2013-09-20 > 13:16:35,113::hsm::2396::Storage.HSM::(connectStorageServer) knownSDs: > {b72b61d1-e11c-496d-ad3a-6f566a1f0ad1: storage.glusterSD.findDomain, > 983c4aa1-7b00-4d3b-b6ad-1fd2cf9297ce: storage.glusterSD.findDomain, > b91afb39-f96e-4eb3-bc6c-9f08fa16869c: storage.glusterSD.findDomain, > 17d21ac7-5859-4f25-8de7-2a9433d50c11: storage.glusterSD.findDomain} > Thread-32390::INFO::2013-09-20 > 13:16:35,113::logUtils::47::dispatcher::(wrapper) Run and protect: > connectStorageServer, Return response: {'statuslist': [{'status': 0, 'id': > 'cecee482-87e1-4ecc-8bda-0e0ec84d7792'}]} > Thread-32390::DEBUG::2013-09-20 > 13:16:35,113::task::1168::TaskManager.Task::(prepare) > Task=`089c5f71-9cbb-4626-8f2d-3ed3547a98cd`::finished: {'statuslist': > [{'status': 0, 'id': 'cecee482-87e1-4ecc-8bda-0e0ec84d7792'}]} > Thread-32390::DEBUG::2013-09-20 > 13:16:35,113::task::579::TaskManager.Task::(_updateState) > Task=`089c5f71-9cbb-4626-8f2d-3ed3547a98cd`::moving from state preparing -> > state finished > Thread-32390::DEBUG::2013-09-20 > 13:16:35,113::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-32390::DEBUG::2013-09-20 > 13:16:35,114::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32390::DEBUG::2013-09-20 > 13:16:35,114::task::974::TaskManager.Task::(_decref) > Task=`089c5f71-9cbb-4626-8f2d-3ed3547a98cd`::ref 0 aborting False > Thread-32393::DEBUG::2013-09-20 > 13:16:35,148::BindingXMLRPC::177::vds::(wrapper) client [10.0.0.34] > Thread-32393::DEBUG::2013-09-20 > 13:16:35,148::task::579::TaskManager.Task::(_updateState) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::moving from state init -> state > preparing > Thread-32393::INFO::2013-09-20 > 13:16:35,148::logUtils::44::dispatcher::(wrapper) Run and protect: > createStoragePool(poolType=None, > spUUID='5849b030-626e-47cb-ad90-3ce782d831b3', poolName='Default', > masterDom='17d21ac7-5859-4f25-8de7-2a9433d50c11', > domList=['17d21ac7-5859-4f25-8de7-2a9433d50c11'], masterVersion=9, > lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, > ioOpTimeoutSec=10, leaseRetries=3, options=None) > Thread-32393::INFO::2013-09-20 > 13:16:35,149::fileSD::315::Storage.StorageDomain::(validate) > sdUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11 > Thread-32393::DEBUG::2013-09-20 > 13:16:35,161::persistentDict::234::Storage.PersistentDict::(refresh) read > lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=rep2-virt', > 'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5', 'LOCKPOLICY=', > 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', > 'REMOTE_PATH=192.168.1.1:rep2-virt', 'ROLE=Regular', > 'SDUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11', 'TYPE=GLUSTERFS', > 'VERSION=3', '_SHA_CKSUM=2b07fbc8c65e20eef5180ab785016bde543c6746'] > Thread-32393::DEBUG::2013-09-20 > 13:16:35,162::resourceManager::197::ResourceManager.Request::(__init__) > ResName=`Storage.5849b030-626e-47cb-ad90-3ce782d831b3`ReqID=`31060ad0-6633-4bbf-a859-b3f0c27af760`::Request > was made in '/usr/share/vdsm/storage/hsm.py' line '954' at > 'createStoragePool' > Thread-32393::DEBUG::2013-09-20 > 13:16:35,162::resourceManager::541::ResourceManager::(registerResource) > Trying to register resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' > for lock type 'exclusive' > Thread-32393::DEBUG::2013-09-20 > 13:16:35,162::resourceManager::600::ResourceManager::(registerResource) > Resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' is free. Now locking > as 'exclusive' (1 active user) > Thread-32393::DEBUG::2013-09-20 > 13:16:35,162::resourceManager::237::ResourceManager.Request::(grant) > ResName=`Storage.5849b030-626e-47cb-ad90-3ce782d831b3`ReqID=`31060ad0-6633-4bbf-a859-b3f0c27af760`::Granted > request > Thread-32393::DEBUG::2013-09-20 > 13:16:35,163::task::811::TaskManager.Task::(resourceAcquired) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::_resourcesAcquired: > Storage.5849b030-626e-47cb-ad90-3ce782d831b3 (exclusive) > Thread-32393::DEBUG::2013-09-20 > 13:16:35,163::task::974::TaskManager.Task::(_decref) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::ref 1 aborting False > Thread-32393::DEBUG::2013-09-20 > 13:16:35,163::resourceManager::197::ResourceManager.Request::(__init__) > ResName=`Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11`ReqID=`db5f52a0-d455-419c-b8a5-86fc6b695571`::Request > was made in '/usr/share/vdsm/storage/hsm.py' line '956' at > 'createStoragePool' > Thread-32393::DEBUG::2013-09-20 > 13:16:35,164::resourceManager::541::ResourceManager::(registerResource) > Trying to register resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' > for lock type 'exclusive' > Thread-32393::DEBUG::2013-09-20 > 13:16:35,164::resourceManager::600::ResourceManager::(registerResource) > Resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' is free. Now locking > as 'exclusive' (1 active user) > Thread-32393::DEBUG::2013-09-20 > 13:16:35,164::resourceManager::237::ResourceManager.Request::(grant) > ResName=`Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11`ReqID=`db5f52a0-d455-419c-b8a5-86fc6b695571`::Granted > request > Thread-32393::DEBUG::2013-09-20 > 13:16:35,165::task::811::TaskManager.Task::(resourceAcquired) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::_resourcesAcquired: > Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11 (exclusive) > Thread-32393::DEBUG::2013-09-20 > 13:16:35,165::task::974::TaskManager.Task::(_decref) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::ref 1 aborting False > Thread-32393::INFO::2013-09-20 > 13:16:35,166::sp::592::Storage.StoragePool::(create) > spUUID=5849b030-626e-47cb-ad90-3ce782d831b3 poolName=Default > master_sd=17d21ac7-5859-4f25-8de7-2a9433d50c11 > domList=['17d21ac7-5859-4f25-8de7-2a9433d50c11'] masterVersion=9 > {'LEASETIMESEC': 60, 'IOOPTIMEOUTSEC': 10, 'LEASERETRIES': 3, > 'LOCKRENEWALINTERVALSEC': 5} > Thread-32393::INFO::2013-09-20 > 13:16:35,166::fileSD::315::Storage.StorageDomain::(validate) > sdUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11 > Thread-32393::DEBUG::2013-09-20 > 13:16:35,177::persistentDict::234::Storage.PersistentDict::(refresh) read > lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=rep2-virt', > 'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5', 'LOCKPOLICY=', > 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', > 'REMOTE_PATH=192.168.1.1:rep2-virt', 'ROLE=Regular', > 'SDUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11', 'TYPE=GLUSTERFS', > 'VERSION=3', '_SHA_CKSUM=2b07fbc8c65e20eef5180ab785016bde543c6746'] > Thread-32393::DEBUG::2013-09-20 > 13:16:35,188::persistentDict::234::Storage.PersistentDict::(refresh) read > lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=rep2-virt', > 'IOOPTIMEOUTSEC=1', 'LEASERETRIES=3', 'LEASETIMESEC=5', 'LOCKPOLICY=', > 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', > 'REMOTE_PATH=192.168.1.1:rep2-virt', 'ROLE=Regular', > 'SDUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11', 'TYPE=GLUSTERFS', > 'VERSION=3', '_SHA_CKSUM=2b07fbc8c65e20eef5180ab785016bde543c6746'] > Thread-32393::WARNING::2013-09-20 > 13:16:35,189::fileUtils::167::Storage.fileUtils::(createdir) Dir > /rhev/data-center/5849b030-626e-47cb-ad90-3ce782d831b3 already exists > Thread-32393::DEBUG::2013-09-20 > 13:16:35,189::persistentDict::167::Storage.PersistentDict::(transaction) > Starting transaction > Thread-32393::DEBUG::2013-09-20 > 13:16:35,189::persistentDict::173::Storage.PersistentDict::(transaction) > Flushing changes > Thread-32393::DEBUG::2013-09-20 > 13:16:35,189::persistentDict::299::Storage.PersistentDict::(flush) about to > write lines (FileMetadataRW)=['CLASS=Data', 'DESCRIPTION=rep2-virt', > 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60', 'LOCKPOLICY=', > 'LOCKRENEWALINTERVALSEC=5', 'POOL_UUID=', > 'REMOTE_PATH=192.168.1.1:rep2-virt', 'ROLE=Regular', > 'SDUUID=17d21ac7-5859-4f25-8de7-2a9433d50c11', 'TYPE=GLUSTERFS', > 'VERSION=3', '_SHA_CKSUM=61b814a65ea3ede1f0ae1d58e139adc06bf9eda5'] > Thread-32393::DEBUG::2013-09-20 > 13:16:35,194::persistentDict::175::Storage.PersistentDict::(transaction) > Finished transaction > Thread-32393::INFO::2013-09-20 > 13:16:35,194::clusterlock::174::SANLock::(acquireHostId) Acquiring host id > for domain 17d21ac7-5859-4f25-8de7-2a9433d50c11 (id: 250) > Thread-32393::ERROR::2013-09-20 > 13:16:36,196::task::850::TaskManager.Task::(_setError) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::Unexpected error > Traceback (most recent call last): > File "/usr/share/vdsm/storage/task.py", line 857, in _run > return fn(*args, **kargs) > File "/usr/share/vdsm/logUtils.py", line 45, in wrapper > res = f(*args, **kwargs) > File "/usr/share/vdsm/storage/hsm.py", line 960, in createStoragePool > masterVersion, leaseParams) > File "/usr/share/vdsm/storage/sp.py", line 617, in create > self._acquireTemporaryClusterLock(msdUUID, leaseParams) > File "/usr/share/vdsm/storage/sp.py", line 559, in > _acquireTemporaryClusterLock > msd.acquireHostId( self.id ) > File "/usr/share/vdsm/storage/sd.py", line 458, in acquireHostId > self._clusterLock.acquireHostId(hostId, async) > File "/usr/share/vdsm/storage/clusterlock.py", line 189, in acquireHostId > raise se.AcquireHostIdFailure(self._sdUUID, e) > AcquireHostIdFailure: Cannot acquire host id: > ('17d21ac7-5859-4f25-8de7-2a9433d50c11', SanlockException(22, 'Sanlock > lockspace add failure', 'Invalid argument')) > Thread-32393::DEBUG::2013-09-20 > 13:16:36,196::task::869::TaskManager.Task::(_run) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::Task._run: > 72348d40-8442-4dbf-bc66-1d354da5fc31 (None, > '5849b030-626e-47cb-ad90-3ce782d831b3', 'Default', > '17d21ac7-5859-4f25-8de7-2a9433d50c11', > ['17d21ac7-5859-4f25-8de7-2a9433d50c11'], 9, None, 5, 60, 10, 3) {} failed - > stopping task > Thread-32393::DEBUG::2013-09-20 > 13:16:36,197::task::1194::TaskManager.Task::(stop) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::stopping in state preparing > (force False) > Thread-32393::DEBUG::2013-09-20 > 13:16:36,197::task::974::TaskManager.Task::(_decref) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::ref 1 aborting True > Thread-32393::INFO::2013-09-20 > 13:16:36,197::task::1151::TaskManager.Task::(prepare) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::aborting: Task is aborted: > 'Cannot acquire host id' - code 661 > Thread-32393::DEBUG::2013-09-20 > 13:16:36,197::task::1156::TaskManager.Task::(prepare) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::Prepare: aborted: Cannot > acquire host id > Thread-32393::DEBUG::2013-09-20 > 13:16:36,197::task::974::TaskManager.Task::(_decref) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::ref 0 aborting True > Thread-32393::DEBUG::2013-09-20 > 13:16:36,197::task::909::TaskManager.Task::(_doAbort) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::Task._doAbort: force False > Thread-32393::DEBUG::2013-09-20 > 13:16:36,198::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32393::DEBUG::2013-09-20 > 13:16:36,198::task::579::TaskManager.Task::(_updateState) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::moving from state preparing -> > state aborting > Thread-32393::DEBUG::2013-09-20 > 13:16:36,198::task::534::TaskManager.Task::(__state_aborting) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::_aborting: recover policy none > Thread-32393::DEBUG::2013-09-20 > 13:16:36,198::task::579::TaskManager.Task::(_updateState) > Task=`72348d40-8442-4dbf-bc66-1d354da5fc31`::moving from state aborting -> > state failed > Thread-32393::DEBUG::2013-09-20 > 13:16:36,198::resourceManager::939::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources > {'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11': < ResourceRef > 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11', isValid: 'True' obj: > 'None'>, 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3': < ResourceRef > 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3', isValid: 'True' obj: > 'None'>} > Thread-32393::DEBUG::2013-09-20 > 13:16:36,198::resourceManager::976::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-32393::DEBUG::2013-09-20 > 13:16:36,199::resourceManager::615::ResourceManager::(releaseResource) > Trying to release resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' > Thread-32393::DEBUG::2013-09-20 > 13:16:36,199::resourceManager::634::ResourceManager::(releaseResource) > Released resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' (0 active > users) > Thread-32393::DEBUG::2013-09-20 > 13:16:36,199::resourceManager::640::ResourceManager::(releaseResource) > Resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11' is free, finding out > if anyone is waiting for it. > Thread-32393::DEBUG::2013-09-20 > 13:16:36,199::resourceManager::648::ResourceManager::(releaseResource) No > one is waiting for resource 'Storage.17d21ac7-5859-4f25-8de7-2a9433d50c11', > Clearing records. > Thread-32393::DEBUG::2013-09-20 > 13:16:36,199::resourceManager::615::ResourceManager::(releaseResource) > Trying to release resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' > Thread-32393::DEBUG::2013-09-20 > 13:16:36,200::resourceManager::634::ResourceManager::(releaseResource) > Released resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' (0 active > users) > Thread-32393::DEBUG::2013-09-20 > 13:16:36,200::resourceManager::640::ResourceManager::(releaseResource) > Resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3' is free, finding out > if anyone is waiting for it. > Thread-32393::DEBUG::2013-09-20 > 13:16:36,200::resourceManager::648::ResourceManager::(releaseResource) No > one is waiting for resource 'Storage.5849b030-626e-47cb-ad90-3ce782d831b3', > Clearing records. > Thread-32393::ERROR::2013-09-20 > 13:16:36,200::dispatcher::67::Storage.Dispatcher.Protect::(run) {'status': > {'message': "Cannot acquire host id: > ('17d21ac7-5859-4f25-8de7-2a9433d50c11', SanlockException(22, 'Sanlock > lockspace add failure', 'Invalid argument'))", 'code': 661}} > Thread-32398::DEBUG::2013-09-20 > 13:16:48,921::task::579::TaskManager.Task::(_updateState) > Task=`a5bce432-622b-499b-a216-d9a1f876e3ca`::moving from state init -> state > preparing > Thread-32398::INFO::2013-09-20 > 13:16:48,922::logUtils::44::dispatcher::(wrapper) Run and protect: > repoStats(options=None) > Thread-32398::INFO::2013-09-20 > 13:16:48,922::logUtils::47::dispatcher::(wrapper) Run and protect: > repoStats, Return response: {} > Thread-32398::DEBUG::2013-09-20 > 13:16:48,922::task::1168::TaskManager.Task::(prepare) > Task=`a5bce432-622b-499b-a216-d9a1f876e3ca`::finished: {} > Thread-32398::DEBUG::2013-09-20 > 13:16:48,922::task::579::TaskManager.Task::(_updateState) > Task=`a5bce432-622b-499b-a216-d9a1f876e3ca`::moving from state preparing -> > state finished > Steve Dainard > Infrastructure Manager > Miovision | Rethink Traffic > 519-513-2407 ex.250 > 877-646-8476 (toll-free) > Blog | LinkedIn | Twitter | Facebook > Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON, > Canada | N2C 1L3 > This e-mail may contain information that is privileged or confidential. If > you are not the intended recipient, please delete the e-mail and any > attachments and notify us immediately. > On Fri, Sep 20, 2013 at 12:23 PM, Deepak C Shetty < > [email protected] > wrote: > > Either you can use the volume set .. option as mentioned in the wikipage > > > --or -- > > > If the Gluster volume is added / managed to the oVirt UI.. go to "Volumes" > > tab, select the gluster volume and > > > click on "Optimize for virt. store". That should also set the volume > > options > > in addition to few other things > > > thanx, > > > deepak > > > ______________________________ _________________ > > > Users mailing list > > > [email protected] > > > http://lists.ovirt.org/ mailman/listinfo/users > > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users _______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

