[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action
First of all i have check permission on my storage. And all permission is right: group kvm (36) and user vdsm (36), chown 36:36 and chmod 0755. After i made 0777 permissions and updating all working fine! My NFS storage on Synology NAS, i now i trying to find a way to make anonuid=36,anongid=36 in the export by right way, not by editing of exports file. After i find it, i will try to use it and 755 permission. Thanks to ALL! BR, Alex ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RRSDPOE4YR4JQKMXAOKHLSMEQ62RYI7A/
[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action
Ok, i will try to set 777 permissoin on NFS storage. But, why this issue starting from updating 4.30.32-1 to 4.30.33-1? Withowt any another changes. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ICWRJ75Q7DIZDDNYRP757YHDEN4N537V/
[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action
Working fine with 4.30.32-1 of course, sorry. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/PABXTFHRCZE5AOQ2KUWQ7JEVXJRLZVW3/
[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action
Htis issue can be resolve by downgrading of the packages: Installing : vdsm-api-4.30.32-1.el7.noarch 1/26 Installing : vdsm-common-4.30.32-1.el7.noarch 2/26 Installing : vdsm-yajsonrpc-4.30.32-1.el7.noarch 3/26 Installing : vdsm-network-4.30.32-1.el7.x86_64 4/26 Installing : vdsm-python-4.30.32-1.el7.noarch 5/26 Installing : vdsm-jsonrpc-4.30.32-1.el7.noarch 6/26 Installing : vdsm-http-4.30.32-1.el7.noarch 7/26 Installing : vdsm-hook-vmfex-dev-4.30.32-1.el7.noarch 8/26 Installing : vdsm-4.30.32-1.el7.x86_64 9/26 Installing : vdsm-gluster-4.30.32-1.el7.x86_64 10/26 Installing : vdsm-hook-ethtool-options-4.30.32-1.el7.noarch 11/26 Installing : vdsm-hook-fcoe-4.30.32-1.el7.noarch 12/26 Installing : vdsm-client-4.30.32-1.el7.noarch 13/26 Cleanup: vdsm-client-4.30.33-1.el7.noarch 14/26 Cleanup: vdsm-hook-ethtool-options-4.30.33-1.el7.noarch 15/26 Cleanup: vdsm-gluster-4.30.33-1.el7.x86_64 16/26 Cleanup: vdsm-hook-fcoe-4.30.33-1.el7.noarch 17/26 Cleanup: vdsm-hook-vmfex-dev-4.30.33-1.el7.noarch 18/26 Cleanup: vdsm-4.30.33-1.el7.x86_64 19/26 Cleanup: vdsm-jsonrpc-4.30.33-1.el7.noarch 20/26 Cleanup: vdsm-http-4.30.33-1.el7.noarch 21/26 Cleanup: vdsm-python-4.30.33-1.el7.noarch 22/26 Cleanup: vdsm-network-4.30.33-1.el7.x86_64 23/26 Cleanup: vdsm-common-4.30.33-1.el7.noarch 24/26 Cleanup: vdsm-api-4.30.33-1.el7.noarch
[ovirt-users] Re: Ovirt-engine-ha cannot to see live status of Hosted Engine
Hi! You was right. The problem was due to an error in the hosts file. The FQDN of engine has another IP in this file on this host from previous instalation. Thank you very muth. Please help me with my another question. I know that you can help me. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/7QIY5HJK3PRM4EXBZZVJJ7WZWLAOJUPI/
[ovirt-users] Re: Gluster Heal Issue
Hi! I did it with working Gluster. Just copy missing files from one of the host and start hesl volume after this. But the main - i dont understand, why this is happening with this issue. I saw this many times after maintanance of one host, for example. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/O5UALK2HXLQZDDBLXYDZJUCHC6LLS7J3/
[ovirt-users] Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action
Hi! I trying to upgrade my hosts and have problem with it. After uprgading one host i see that this one NonOperational. All was fine with vdsm-4.30.24-1.el7 but after upgrading with new version vdsm-4.30.40-1.el7.x86_64 and some others i have errors. Firtst of all i see in ovirt Events: Host srv02 cannot access the Storage Domain(s) attached to the Data Center Default. Setting Host state to Non-Operational. My Default storage domain with HE VM data on NFS storage. In messages log of host: srv02 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR Traceback (most recent call last):#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/a gent.py", line 131, in _run_agent#012return action(he)#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 55, in action_proper#012return he.start_monitoring ()#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py", line 432, in start_monitoring#012self._initialize_broker()#012 File "/usr/lib/python2.7/site-packages/ ovirt_hosted_engine_ha/agent/hosted_engine.py", line 556, in _initialize_broker#012m.get('options', {}))#012 File "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", line 8 9, in start_monitor#012).format(t=type, o=options, e=e)#012RequestError: brokerlink - failed to start monitor via ovirt-ha-broker: [Errno 2] No such file or directory, [monitor: 'network', options: {'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr': '192.168.2.248'}] Feb 1 15:41:42 srv02 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR Trying to restart agent In broker log: MainThread::WARNING::2020-02-01 15:43:35,167::storage_broker::97::ovirt_hosted_engine_ha.broker.storage_broker.StorageBroker::(__init__) Can't connect vdsm storage: Command StorageDomain.getInfo with ar gs {'storagedomainID': 'bbdddea7-9cd6-41e7-ace5-fb9a6795caa8'} failed: (code=350, message=Error in storage domain action: (u'sdUUID=bbdddea7-9cd6-41e7-ace5-fb9a6795caa8',)) In vdsm.lod 2020-02-01 15:44:19,930+0600 INFO (jsonrpc/0) [vdsm.api] FINISH getStorageDomainInfo error=[Errno 1] Operation not permitted from=::1,57528, task_id=40683f67-d7b0-4105-aab8-6338deb54b00 (api:52) 2020-02-01 15:44:19,930+0600 ERROR (jsonrpc/0) [storage.TaskManager.Task] (Task='40683f67-d7b0-4105-aab8-6338deb54b00') Unexpected error (task:875) Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _run return fn(*args, **kargs) File "", line 2, in getStorageDomainInfo File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 2753, in getStorageDomainInfo dom = self.validateSdUUID(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 305, in validateSdUUID sdDom = sdCache.produce(sdUUID=sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 110, in produce domain.getRealDomain() File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in getRealDomain return self._cache._realProduce(self._sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in _realProduce domain = self._findDomain(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in _findDomain return findMethod(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/nfsSD.py", line 145, in findDomain return NfsStorageDomain(NfsStorageDomain.findDomainPath(sdUUID)) File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 378, in __init__ manifest.sdUUID, manifest.mountpoint) File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 853, in _detect_block_size block_size = iop.probe_block_size(mountpoint) File "/usr/lib/python2.7/site-packages/vdsm/storage/outOfProcess.py", line 384, in probe_block_size return self._ioproc.probe_block_size(dir_path) File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 602, in probe_block_size "probe_block_size", {"dir": dir_path}, self.timeout) File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 448, in _sendCommand raise OSError(errcode, errstr) OSError: [Errno 1] Operation not permitted 2020-02-01 15:44:19,930+0600 INFO (jsonrpc/0) [storage.TaskManager.Task] (Task='40683f67-d7b0-4105-aab8-6338deb54b00') aborting: Task is aborted: u'[Errno 1] Operation not permitted' - code 100 (task:1 181) 2020-02-01 15:44:19,930+0600 ERROR (jsonrpc/0) [storage.Dispatcher] FINISH getStorageDomainInfo error=[Errno 1] Operation not permitted (dispatcher:87) But i see that this domain is mounted (by mount command): storage:/volume3/ovirt-hosted on /rhev/data-center/mnt/storage:_volume3_ovirt-hosted type nfs4
[ovirt-users] Re: Ovirt-engine-ha cannot to see live status of Hosted Engine
Hi! How i can to check the network? All the same on two nodes except IP address. Pinges working fine, and others. Here also broker log srom srv00. You cal see moment when HostedEngine migrated to another host: Thread-2::INFO::2018-09-18 12:48:07,531::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-3::INFO::2018-09-18 12:48:07,767::mem_free::51::mem_free.MemFree::(action) memFree: 12774 Thread-1::INFO::2018-09-18 12:48:07,901::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-2::INFO::2018-09-18 12:48:17,555::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-3::INFO::2018-09-18 12:48:17,812::mem_free::51::mem_free.MemFree::(action) memFree: 12766 Thread-3::INFO::2018-09-18 12:48:26,852::mem_free::51::mem_free.MemFree::(action) memFree: 12757 Thread-1::INFO::2018-09-18 12:48:27,453::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-2::INFO::2018-09-18 12:48:27,587::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-5::WARNING::2018-09-18 12:48:30,495::engine_health::233::engine_health.EngineHealth::(_result_from_stats) bad health status: Hosted Engine is not up! Thread-3::INFO::2018-09-18 12:48:36,894::mem_free::51::mem_free.MemFree::(action) memFree: 12759 Thread-2::INFO::2018-09-18 12:48:37,619::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-1::INFO::2018-09-18 12:48:37,727::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-3::INFO::2018-09-18 12:48:46,944::mem_free::51::mem_free.MemFree::(action) memFree: 12762 Thread-2::INFO::2018-09-18 12:48:47,651::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-1::INFO::2018-09-18 12:48:48,006::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-5::WARNING::2018-09-18 12:48:50,603::engine_health::233::engine_health.EngineHealth::(_result_from_stats) bad health status: Hosted Engine is not up! Thread-3::INFO::2018-09-18 12:48:57,021::mem_free::51::mem_free.MemFree::(action) memFree: 12736 Thread-1::INFO::2018-09-18 12:48:57,285::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-2::INFO::2018-09-18 12:48:57,679::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-4::INFO::2018-09-18 12:49:04,920::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load) System load total=0.0401, engine=0.0030, non-e ngine=0.0371 Thread-3::INFO::2018-09-18 12:49:07,064::mem_free::51::mem_free.MemFree::(action) memFree: 12740 Thread-1::INFO::2018-09-18 12:49:07,561::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-2::INFO::2018-09-18 12:49:07,760::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-5::WARNING::2018-09-18 12:49:10,715::engine_health::233::engine_health.EngineHealth::(_result_from_stats) bad health status: Hosted Engine is not up! Thread-5::WARNING::2018-09-18 12:49:10,823::engine_health::233::engine_health.EngineHealth::(_result_from_stats) bad health status: Hosted Engine is not up! Thread-7::WARNING::2018-09-18 12:49:12,961::engine_health::233::engine_health.EngineHealth::(_result_from_stats) bad health status: Hosted Engine is not up! Thread-3::INFO::2018-09-18 12:49:17,114::mem_free::51::mem_free.MemFree::(action) memFree: 12739 Thread-1::INFO::2018-09-18 12:49:17,817::ping::60::ping.Ping::(action) Successfully pinged 192.168.2.248 Thread-2::INFO::2018-09-18 12:49:17,888::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge ovirtmgmt with ports Thread-5::WARNING::2018-09-18 12:49:19,945::engine_health::233::engine_health.EngineHealth::(_result_from_stats) bad health status: Hosted Engine is not up! Thread-7::INFO::2018-09-18 12:49:25,650::engine_health::191::engine_health.EngineHealth::(_result_from_stats) VM successfully migrated away from this host. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZHGU3EYDUUGRNZQPLK6KHLU7AETCOVTR/
[ovirt-users] Ovirt-engine-ha cannot to see live status of Hosted Engine
Good day for all. I have some issues with Ovirt 4.2.6. But now the main this of it: I have two Centos 7 Nodes with same config and last Ovirt 4.2.6 with Hostedengine with disk on NFS storage. Also some of virtual machines working good. But, when HostedEngine running on one node (srv02.local) everything is fine. After migrating to another node (srv00.local), i see that agent cannot to check livelinness of HostedEngine. After few minutes HostedEngine going to reboot and after some time i see some situation. After migration to another node (srv00.local) all looks OK. hosted-engine --vm-status commang when HosterEngine on srv00 node: --== Host 1 status ==-- conf_on_shared_storage : True Status up-to-date : True Hostname : srv02.local Host ID: 1 Engine status : {"reason": "vm not running on this host", "health": "bad", "vm": "down_unexpected", "detail": "unknown"} Score : 0 stopped: False Local maintenance : False crc32 : ecc7ad2d local_conf_timestamp : 78328 Host timestamp : 78328 Extra metadata (valid at timestamp): metadata_parse_version=1 metadata_feature_version=1 timestamp=78328 (Tue Sep 18 12:44:18 2018) host-id=1 score=0 vm_conf_refresh_time=78328 (Tue Sep 18 12:44:18 2018) conf_on_shared_storage=True maintenance=False state=EngineUnexpectedlyDown stopped=False timeout=Fri Jan 2 03:49:58 1970 --== Host 2 status ==-- conf_on_shared_storage : True Status up-to-date : True Hostname : srv00.local Host ID: 2 Engine status : {"reason": "failed liveliness check", "health": "bad", "vm": "up", "detail": "Up"} Score : 3400 stopped: False Local maintenance : False crc32 : 1d62b106 local_conf_timestamp : 326288 Host timestamp : 326288 Extra metadata (valid at timestamp): metadata_parse_version=1 metadata_feature_version=1 timestamp=326288 (Tue Sep 18 12:44:21 2018) host-id=2 score=3400 vm_conf_refresh_time=326288 (Tue Sep 18 12:44:21 2018) conf_on_shared_storage=True maintenance=False state=EngineStarting stopped=False Log agent.log from srv00.local: MainThread::INFO::2018-09-18 12:40:51,749::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedE ngine::(consume) VM is powering up.. MainThread::INFO::2018-09-18 12:40:52,052::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine. HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 3400) MainThread::INFO::2018-09-18 12:41:01,066::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedE ngine::(consume) VM is powering up.. MainThread::INFO::2018-09-18 12:41:01,374::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine. HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 3400) MainThread::INFO::2018-09-18 12:41:11,393::state_machine::169::ovirt_hosted_engine_ha.agent.hosted_engine. HostedEngine::(refresh) Global metadata: {'maintenance': False} MainThread::INFO::2018-09-18 12:41:11,393::state_machine::174::ovirt_hosted_engine_ha.agent.hosted_engine. HostedEngine::(refresh) Host srv02.local.pioner.kz (id 1): {'conf_on_shared_storage': True, 'extra': 'meta data_parse_version=1\nmetadata_feature_version=1\ntimestamp=78128 (Tue Sep 18 12:40:58 2018)\nhost-id=1\ns core=0\nvm_conf_refresh_time=78128 (Tue Sep 18 12:40:58 2018)\nconf_on_shared_storage=True\nmaintenance=Fa lse\nstate=EngineUnexpectedlyDown\nstopped=False\ntimeout=Fri Jan 2 03:49:58 1970\n', 'hostname': 'srv02. local.pioner.kz', 'alive': True, 'host-id': 1, 'engine-status': {'reason': 'vm not running on this host', 'health': 'bad', 'vm': 'down_unexpected', 'detail': 'unknown'}, 'score': 0, 'stopped': False, 'maintenance ': False, 'crc32': 'e18e3f22', 'local_conf_timestamp': 78128, 'host-ts': 78128} MainThread::INFO::2018-09-18 12:41:11,393::state_machine::177::ovirt_hosted_engine_ha.agent.hosted_engine. HostedEngine::(refresh) Local (id 2): {'engine-health': {'reason': 'failed liveliness check', 'health': 'b ad', 'vm': 'up', 'detail': 'Up'}, 'bridge': True, 'mem-free': 12763.0, 'maintenance': False, 'cpu-load': 0 .0364, 'gateway': 1.0, 'storage-domain': True} MainThread::INFO::2018-09-18 12:41:11,393::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedE ngine::(consume) VM is powering up.. MainThread::INFO::2018-09-18 12:41:11,703::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine. HostedEngine::(_monitoring_loop) Current state EngineStarting