[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action

2020-02-01 Thread asm
First of all i have check permission on my storage. And all permission is 
right: group kvm (36) and user vdsm (36), chown 36:36 and  chmod 0755. 
After i made 0777 permissions and updating all working fine!
My NFS storage on Synology NAS, i now i trying to find a way to make 
anonuid=36,anongid=36 in the export by right way, not by editing of exports 
file. After i find it, i will try to use it and 755 permission.
Thanks to ALL!
BR,
Alex
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RRSDPOE4YR4JQKMXAOKHLSMEQ62RYI7A/


[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action

2020-02-01 Thread asm
Ok, i will try to set 777 permissoin on NFS storage. But, why this issue 
starting from updating  4.30.32-1 to  4.30.33-1? Withowt any another changes.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ICWRJ75Q7DIZDDNYRP757YHDEN4N537V/


[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action

2020-02-01 Thread asm
Working fine with 4.30.32-1 of course, sorry.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PABXTFHRCZE5AOQ2KUWQ7JEVXJRLZVW3/


[ovirt-users] Re: Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action

2020-02-01 Thread asm
Htis issue can be resolve by downgrading of the packages:
  Installing : vdsm-api-4.30.32-1.el7.noarch

 1/26
  Installing : vdsm-common-4.30.32-1.el7.noarch 

 2/26
  Installing : vdsm-yajsonrpc-4.30.32-1.el7.noarch  

 3/26
  Installing : vdsm-network-4.30.32-1.el7.x86_64

 4/26
  Installing : vdsm-python-4.30.32-1.el7.noarch 

 5/26
  Installing : vdsm-jsonrpc-4.30.32-1.el7.noarch

 6/26
  Installing : vdsm-http-4.30.32-1.el7.noarch   

 7/26
  Installing : vdsm-hook-vmfex-dev-4.30.32-1.el7.noarch 

 8/26
  Installing : vdsm-4.30.32-1.el7.x86_64

 9/26
  Installing : vdsm-gluster-4.30.32-1.el7.x86_64

10/26
  Installing : vdsm-hook-ethtool-options-4.30.32-1.el7.noarch   

11/26
  Installing : vdsm-hook-fcoe-4.30.32-1.el7.noarch  

12/26
  Installing : vdsm-client-4.30.32-1.el7.noarch 

13/26
  Cleanup: vdsm-client-4.30.33-1.el7.noarch 

14/26
  Cleanup: vdsm-hook-ethtool-options-4.30.33-1.el7.noarch   

15/26
  Cleanup: vdsm-gluster-4.30.33-1.el7.x86_64

16/26
  Cleanup: vdsm-hook-fcoe-4.30.33-1.el7.noarch  

17/26
  Cleanup: vdsm-hook-vmfex-dev-4.30.33-1.el7.noarch 

18/26
  Cleanup: vdsm-4.30.33-1.el7.x86_64

19/26
  Cleanup: vdsm-jsonrpc-4.30.33-1.el7.noarch

20/26
  Cleanup: vdsm-http-4.30.33-1.el7.noarch   

21/26
  Cleanup: vdsm-python-4.30.33-1.el7.noarch 

22/26
  Cleanup: vdsm-network-4.30.33-1.el7.x86_64

23/26
  Cleanup: vdsm-common-4.30.33-1.el7.noarch 

24/26
  Cleanup: vdsm-api-4.30.33-1.el7.noarch 

[ovirt-users] Re: Ovirt-engine-ha cannot to see live status of Hosted Engine

2020-02-01 Thread asm
Hi! You was right. 
The problem was due to an error in the hosts file. The FQDN of engine has 
another IP in this file on this host from previous instalation.
Thank you very muth. 
Please help me with my another question. I know that you can help me.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7QIY5HJK3PRM4EXBZZVJJ7WZWLAOJUPI/


[ovirt-users] Re: Gluster Heal Issue

2020-02-01 Thread asm
Hi! 
I did it with working Gluster. Just copy missing files from one of the host and 
start hesl volume after this.
But the main - i dont understand, why this is happening with this issue. I saw 
this many times after maintanance of one host, for example.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/O5UALK2HXLQZDDBLXYDZJUCHC6LLS7J3/


[ovirt-users] Can't connect vdsm storage: Command StorageDomain.getInfo with args failed: (code=350, message=Error in storage domain action

2020-02-01 Thread asm
Hi! I trying to upgrade my hosts and have problem with it. After uprgading one 
host i see that this one NonOperational. All was fine with vdsm-4.30.24-1.el7 
but after upgrading with new version vdsm-4.30.40-1.el7.x86_64 and some others 
i have errors. 
Firtst of all i see in ovirt Events: Host srv02 cannot access the Storage 
Domain(s)  attached to the Data Center Default. Setting Host state to 
Non-Operational. My Default storage domain with HE VM data on NFS storage.

In messages log of host:  
srv02 journal: ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR 
Traceback (most recent call last):#012  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/a
gent.py", line 131, in _run_agent#012return action(he)#012  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
55, in action_proper#012return he.start_monitoring
()#012  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 432, in start_monitoring#012self._initialize_broker()#012  File 
"/usr/lib/python2.7/site-packages/
ovirt_hosted_engine_ha/agent/hosted_engine.py", line 556, in 
_initialize_broker#012m.get('options', {}))#012  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", 
line 8
9, in start_monitor#012).format(t=type, o=options, e=e)#012RequestError: 
brokerlink - failed to start monitor via ovirt-ha-broker: [Errno 2] No such 
file or directory, [monitor: 'network', options:
{'tcp_t_address': None, 'network_test': None, 'tcp_t_port': None, 'addr': 
'192.168.2.248'}]
Feb  1 15:41:42 srv02 journal: ovirt-ha-agent 
ovirt_hosted_engine_ha.agent.agent.Agent ERROR Trying to restart agent

In broker log: 
MainThread::WARNING::2020-02-01 
15:43:35,167::storage_broker::97::ovirt_hosted_engine_ha.broker.storage_broker.StorageBroker::(__init__)
 Can't connect vdsm storage: Command StorageDomain.getInfo with ar
gs {'storagedomainID': 'bbdddea7-9cd6-41e7-ace5-fb9a6795caa8'} failed:
(code=350, message=Error in storage domain action: 
(u'sdUUID=bbdddea7-9cd6-41e7-ace5-fb9a6795caa8',))

In vdsm.lod
2020-02-01 15:44:19,930+0600 INFO  (jsonrpc/0) [vdsm.api] FINISH 
getStorageDomainInfo error=[Errno 1] Operation not permitted from=::1,57528, 
task_id=40683f67-d7b0-4105-aab8-6338deb54b00 (api:52)
2020-02-01 15:44:19,930+0600 ERROR (jsonrpc/0) [storage.TaskManager.Task] 
(Task='40683f67-d7b0-4105-aab8-6338deb54b00') Unexpected error (task:875)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in 
_run
return fn(*args, **kargs)
  File "", line 2, in getStorageDomainInfo
  File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method
ret = func(*args, **kwargs)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 2753, in 
getStorageDomainInfo
dom = self.validateSdUUID(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 305, in 
validateSdUUID
sdDom = sdCache.produce(sdUUID=sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 110, in 
produce
domain.getRealDomain()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in 
getRealDomain
return self._cache._realProduce(self._sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in 
_realProduce
domain = self._findDomain(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in 
_findDomain
return findMethod(sdUUID)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/nfsSD.py", line 145, in 
findDomain
return NfsStorageDomain(NfsStorageDomain.findDomainPath(sdUUID))
  File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 378, in 
__init__
manifest.sdUUID, manifest.mountpoint)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/fileSD.py", line 853, in 
_detect_block_size
block_size = iop.probe_block_size(mountpoint)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/outOfProcess.py", line 
384, in probe_block_size
return self._ioproc.probe_block_size(dir_path)
  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 602, in 
probe_block_size
"probe_block_size", {"dir": dir_path}, self.timeout)
  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 448, in 
_sendCommand
raise OSError(errcode, errstr)
OSError: [Errno 1] Operation not permitted
2020-02-01 15:44:19,930+0600 INFO  (jsonrpc/0) [storage.TaskManager.Task] 
(Task='40683f67-d7b0-4105-aab8-6338deb54b00') aborting: Task is aborted: 
u'[Errno 1] Operation not permitted' - code 100 (task:1
181)
2020-02-01 15:44:19,930+0600 ERROR (jsonrpc/0) [storage.Dispatcher] FINISH 
getStorageDomainInfo error=[Errno 1] Operation not permitted (dispatcher:87)

But i see that this domain is mounted (by mount command):
storage:/volume3/ovirt-hosted on 
/rhev/data-center/mnt/storage:_volume3_ovirt-hosted type nfs4 

[ovirt-users] Re: Ovirt-engine-ha cannot to see live status of Hosted Engine

2018-09-18 Thread asm
Hi! How i can to check the network? All the same on two nodes except IP 
address. Pinges working fine, and others.
Here also broker log srom srv00. You cal see moment when HostedEngine migrated 
to another host:
Thread-2::INFO::2018-09-18 
12:48:07,531::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-3::INFO::2018-09-18 
12:48:07,767::mem_free::51::mem_free.MemFree::(action) memFree: 12774
Thread-1::INFO::2018-09-18 12:48:07,901::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-2::INFO::2018-09-18 
12:48:17,555::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-3::INFO::2018-09-18 
12:48:17,812::mem_free::51::mem_free.MemFree::(action) memFree: 12766
Thread-3::INFO::2018-09-18 
12:48:26,852::mem_free::51::mem_free.MemFree::(action) memFree: 12757
Thread-1::INFO::2018-09-18 12:48:27,453::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-2::INFO::2018-09-18 
12:48:27,587::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-5::WARNING::2018-09-18 
12:48:30,495::engine_health::233::engine_health.EngineHealth::(_result_from_stats)
 bad health status: Hosted Engine is not up!

Thread-3::INFO::2018-09-18 
12:48:36,894::mem_free::51::mem_free.MemFree::(action) memFree: 12759
Thread-2::INFO::2018-09-18 
12:48:37,619::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-1::INFO::2018-09-18 12:48:37,727::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-3::INFO::2018-09-18 
12:48:46,944::mem_free::51::mem_free.MemFree::(action) memFree: 12762
Thread-2::INFO::2018-09-18 
12:48:47,651::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-1::INFO::2018-09-18 12:48:48,006::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-5::WARNING::2018-09-18 
12:48:50,603::engine_health::233::engine_health.EngineHealth::(_result_from_stats)
 bad health status: Hosted Engine is not up!

Thread-3::INFO::2018-09-18 
12:48:57,021::mem_free::51::mem_free.MemFree::(action) memFree: 12736
Thread-1::INFO::2018-09-18 12:48:57,285::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-2::INFO::2018-09-18 
12:48:57,679::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-4::INFO::2018-09-18 
12:49:04,920::cpu_load_no_engine::126::cpu_load_no_engine.CpuLoadNoEngine::(calculate_load)
 System load total=0.0401, engine=0.0030, non-e
ngine=0.0371
Thread-3::INFO::2018-09-18 
12:49:07,064::mem_free::51::mem_free.MemFree::(action) memFree: 12740
Thread-1::INFO::2018-09-18 12:49:07,561::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-2::INFO::2018-09-18 
12:49:07,760::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-5::WARNING::2018-09-18 
12:49:10,715::engine_health::233::engine_health.EngineHealth::(_result_from_stats)
 bad health status: Hosted Engine is not up!

Thread-5::WARNING::2018-09-18 
12:49:10,823::engine_health::233::engine_health.EngineHealth::(_result_from_stats)
 bad health status: Hosted Engine is not up!

Thread-7::WARNING::2018-09-18 
12:49:12,961::engine_health::233::engine_health.EngineHealth::(_result_from_stats)
 bad health status: Hosted Engine is not up!

Thread-3::INFO::2018-09-18 
12:49:17,114::mem_free::51::mem_free.MemFree::(action) memFree: 12739
Thread-1::INFO::2018-09-18 12:49:17,817::ping::60::ping.Ping::(action) 
Successfully pinged 192.168.2.248
Thread-2::INFO::2018-09-18 
12:49:17,888::mgmt_bridge::62::mgmt_bridge.MgmtBridge::(action) Found bridge 
ovirtmgmt with ports
Thread-5::WARNING::2018-09-18 
12:49:19,945::engine_health::233::engine_health.EngineHealth::(_result_from_stats)
 bad health status: Hosted Engine is not up!

Thread-7::INFO::2018-09-18 
12:49:25,650::engine_health::191::engine_health.EngineHealth::(_result_from_stats)
 VM successfully migrated away from this host.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZHGU3EYDUUGRNZQPLK6KHLU7AETCOVTR/


[ovirt-users] Ovirt-engine-ha cannot to see live status of Hosted Engine

2018-09-18 Thread asm
Good day for all.
I have some issues with Ovirt 4.2.6. But now the main this of it:
I have two Centos 7 Nodes with same config and last Ovirt 4.2.6 with 
Hostedengine with disk on NFS storage.
Also some of virtual machines working good. 
But, when HostedEngine running on one node (srv02.local) everything is fine. 
After migrating to another node (srv00.local), i see that agent cannot to check 
livelinness of HostedEngine. After few minutes HostedEngine going to reboot and 
after some time i see some situation. After migration to another node 
(srv00.local) all looks OK.

hosted-engine --vm-status commang when HosterEngine on srv00 node:
--== Host 1 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : srv02.local
Host ID: 1
Engine status  : {"reason": "vm not running on this host", 
"health": "bad", "vm": "down_unexpected", "detail": "unknown"}
Score  : 0
stopped: False
Local maintenance  : False
crc32  : ecc7ad2d
local_conf_timestamp   : 78328
Host timestamp : 78328
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=78328 (Tue Sep 18 12:44:18 2018)
host-id=1
score=0
vm_conf_refresh_time=78328 (Tue Sep 18 12:44:18 2018)
conf_on_shared_storage=True
maintenance=False
state=EngineUnexpectedlyDown
stopped=False
timeout=Fri Jan  2 03:49:58 1970


--== Host 2 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : srv00.local
Host ID: 2
Engine status  : {"reason": "failed liveliness check", 
"health": "bad", "vm": "up", "detail": "Up"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 1d62b106
local_conf_timestamp   : 326288
Host timestamp : 326288
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=326288 (Tue Sep 18 12:44:21 2018)
host-id=2
score=3400
vm_conf_refresh_time=326288 (Tue Sep 18 12:44:21 2018)
conf_on_shared_storage=True
maintenance=False
state=EngineStarting
stopped=False

Log agent.log from srv00.local:

MainThread::INFO::2018-09-18 
12:40:51,749::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedE
ngine::(consume) VM is powering up..
MainThread::INFO::2018-09-18 
12:40:52,052::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 3400)
MainThread::INFO::2018-09-18 
12:41:01,066::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedE
ngine::(consume) VM is powering up..
MainThread::INFO::2018-09-18 
12:41:01,374::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(_monitoring_loop) Current state EngineStarting (score: 3400)
MainThread::INFO::2018-09-18 
12:41:11,393::state_machine::169::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(refresh) Global metadata: {'maintenance': False}
MainThread::INFO::2018-09-18 
12:41:11,393::state_machine::174::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(refresh) Host srv02.local.pioner.kz (id 1): 
{'conf_on_shared_storage': True, 'extra': 'meta
data_parse_version=1\nmetadata_feature_version=1\ntimestamp=78128 (Tue Sep 18 
12:40:58 2018)\nhost-id=1\ns
core=0\nvm_conf_refresh_time=78128 (Tue Sep 18 12:40:58 
2018)\nconf_on_shared_storage=True\nmaintenance=Fa
lse\nstate=EngineUnexpectedlyDown\nstopped=False\ntimeout=Fri Jan  2 03:49:58 
1970\n', 'hostname': 'srv02.
local.pioner.kz', 'alive': True, 'host-id': 1, 'engine-status': {'reason': 'vm 
not running on this host',
'health': 'bad', 'vm': 'down_unexpected', 'detail': 'unknown'}, 'score': 0, 
'stopped': False, 'maintenance
': False, 'crc32': 'e18e3f22', 'local_conf_timestamp': 78128, 'host-ts': 78128}
MainThread::INFO::2018-09-18 
12:41:11,393::state_machine::177::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(refresh) Local (id 2): {'engine-health': {'reason': 'failed 
liveliness check', 'health': 'b
ad', 'vm': 'up', 'detail': 'Up'}, 'bridge': True, 'mem-free': 12763.0, 
'maintenance': False, 'cpu-load': 0
.0364, 'gateway': 1.0, 'storage-domain': True}
MainThread::INFO::2018-09-18 
12:41:11,393::states::779::ovirt_hosted_engine_ha.agent.hosted_engine.HostedE
ngine::(consume) VM is powering up..
MainThread::INFO::2018-09-18 
12:41:11,703::hosted_engine::491::ovirt_hosted_engine_ha.agent.hosted_engine.
HostedEngine::(_monitoring_loop) Current state EngineStarting