[ovirt-users] Re: Periodic fall host

2018-10-17 Thread Spickiy Nikita
:36,995+0300 ERROR (periodic/14) [Executor] Unhandled exception 
in  timeout=30.0, duration=60 at 0x7f5b26cfce90> (executor:317)
2018-10-17 13:44:36,995+0300 ERROR (periodic/14) [storage.Dispatcher] FINISH 
getVolumeSize error=Connection timed out (dispatcher:86)
2018-10-17 13:44:36,995+0300 ERROR (periodic/14) [storage.TaskManager.Task] 
(Task='b198455f-2847-4c8e-821b-6c269bed2411') Unexpected error (task:875)
2018-10-17 13:44:36,996+0300 ERROR (periodic/16) [Executor] Unhandled exception 
in  timeout=30.0, duration=60 at 0x7f5b26cfc790> (executor:317)
2018-10-17 13:44:36,996+0300 ERROR (periodic/16) [storage.Dispatcher] FINISH 
getVolumeSize error=Connection timed out (dispatcher:86)
2018-10-17 13:44:36,996+0300 ERROR (periodic/16) [storage.TaskManager.Task] 
(Task='60d13a5b-0b10-4a60-b9ee-05820a4e47a7') Unexpected error (task:875)
2018-10-17 13:44:37,001+0300 ERROR (periodic/17) [storage.TaskManager.Task] 
(Task='6c827331-3373-4ed0-9630-0cea0d989a06') Unexpected error (task:875)
2018-10-17 13:44:37,002+0300 ERROR (periodic/17) [Executor] Unhandled exception 
in  timeout=30.0, duration=60 at 0x7f5b26fc9f90> (executor:317)
2018-10-17 13:44:37,002+0300 ERROR (periodic/17) [storage.Dispatcher] FINISH 
getVolumeSize error=Connection timed out (dispatcher:86)
2018-10-17 14:56:02,788+0300 ERROR (migmon/f52e8e78) [root] Unhandled exception 
(logutils:412)
2018-10-17 14:56:02,884+0300 ERROR (migmon/f52e8e78) [root] FINISH thread 
 failed 
(concurrent:201)

On 17 Oct 2018, at 14:42, Sahina Bose 
mailto:sab...@redhat.com>> wrote:



On Tue, Oct 16, 2018 at 11:39 PM Spickiy Nikita 
mailto:n.spic...@outlook.com>> wrote:
Hi, i have oVirt instance (4.2.1.6-1.el7.centos). So, i have cluster with 
gluster. Hosts periodically non response and VM's is not responding. Usually it 
happens after get message "command GetGlusterVolumeHealInfoVDS failed: Message 
timeout which can be caused by communication issues".

Will solve the trouble if an increase timeout for get heat status? And how to 
do it?

I attach part log below:

https://paste.fedoraproject.org/paste/8TTzwjMbYk32d7wd7Ix0Pw/raw

2018-10-15 14:44:22,582+03 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(DefaultQuartzScheduler6) [70cfd553] EVENT_ID: 
VDS_BROKER_COMMAND_FAILURE(10,802), VDSM 
ovirt3.example.org<http://ovirt3.example.org/> command 
GetGlusterVolumeHealInfoVDS failed: Message timeout which can be caused by 
communication issues
2018-10-15 14:44:22,584+03 ERROR 
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeHealInfoVDSCommand] 
(DefaultQuartzScheduler6) [70cfd553] Command 
'GetGlusterVolumeHealInfoVDSCommand(HostName = 
ovirt3.example.org<http://ovirt3.example.org/>, 
GlusterVolumeVDSParameters:{hostId='39215015-2537-4329-921f-c11256f99e04', 
volumeName='domain1'})' execution failed: VDSGenericException: 
VDSNetworkException: Message timeout which can be caused by communication issues
2018-10-15 14:44:22,584+03 WARN  [org.ovirt.engine.core.vdsbroker.VdsManager] 
(EE-ManagedThreadFactory-engine-Thread-7) [70cfd553] Host 
'ovirt3.example.org<http://ovirt3.example.org/>' is not responding. It will 
stay in Connecting state for a grace period of 77 seconds and after that an 
attempt to fence the host will be issued.
2018-10-15 14:44:22,591+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engine-Thread-7) [70cfd553] EVENT_ID: 
VDS_HOST_NOT_RESPONDING_CONNECTING(9,008), Host 
ovirt3.example.org<http://ovirt3.example.org/> is not responding. It will stay 
in Connecting state for a grace period of 77 seconds and after that an attempt 
to fence the host will be issued.
2018-10-15 14:44:54,620+03 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engine-Thread-13) [] EVENT_ID: 
VDS_STORAGE_VDS_STATS_FAILED(189), Host 
ovirt3.example.org<http://ovirt3.example.org/> reports about one of the Active 
Storage Domains as Problematic.
2018-10-15 14:44:54,827+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-46) [6d9504d1] EVENT_ID: 
VDS_SET_NONOPERATIONAL_DOMAIN(522), Host 
ovirt3.example.org<http://ovirt3.example.org/> cannot access the Storage 
Domain(s) DOMAIN1 attached to the Data Center Default. Setting Host state to 
Non-Operational.
2018-10-15 14:44:54,840+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-46) [6d9504d1] EVENT_ID: 
CONNECT_STORAGE_POOL_FAILED(995), Failed to connect Host 
ovirt3.example.org<http://ovirt3.example.org/> to Storage Pool Default
2018-10-15 14:45:28,698+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-87) [] EVENT_ID: 
VM_NOT_RESPONDING(126), VM HostedEngine is

[ovirt-users] Periodic fall host

2018-10-16 Thread Spickiy Nikita
Hi, i have oVirt instance (4.2.1.6-1.el7.centos). So, i have cluster with 
gluster. Hosts periodically non response and VM's is not responding. Usually it 
happens after get message "command GetGlusterVolumeHealInfoVDS failed: Message 
timeout which can be caused by communication issues".

Will solve the trouble if an increase timeout for get heat status? And how to 
do it?

I attach part log below:

https://paste.fedoraproject.org/paste/8TTzwjMbYk32d7wd7Ix0Pw/raw

2018-10-15 14:44:22,582+03 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(DefaultQuartzScheduler6) [70cfd553] EVENT_ID: 
VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ovirt3.example.org command 
GetGlusterVolumeHealInfoVDS failed: Message timeout which can be caused by 
communication issues
2018-10-15 14:44:22,584+03 ERROR 
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeHealInfoVDSCommand] 
(DefaultQuartzScheduler6) [70cfd553] Command 
'GetGlusterVolumeHealInfoVDSCommand(HostName = ovirt3.example.org, 
GlusterVolumeVDSParameters:{hostId='39215015-2537-4329-921f-c11256f99e04', 
volumeName='domain1'})' execution failed: VDSGenericException: 
VDSNetworkException: Message timeout which can be caused by communication issues
2018-10-15 14:44:22,584+03 WARN  [org.ovirt.engine.core.vdsbroker.VdsManager] 
(EE-ManagedThreadFactory-engine-Thread-7) [70cfd553] Host 'ovirt3.example.org' 
is not responding. It will stay in Connecting state for a grace period of 77 
seconds and after that an attempt to fence the host will be issued.
2018-10-15 14:44:22,591+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engine-Thread-7) [70cfd553] EVENT_ID: 
VDS_HOST_NOT_RESPONDING_CONNECTING(9,008), Host ovirt3.example.org is not 
responding. It will stay in Connecting state for a grace period of 77 seconds 
and after that an attempt to fence the host will be issued.
2018-10-15 14:44:54,620+03 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engine-Thread-13) [] EVENT_ID: 
VDS_STORAGE_VDS_STATS_FAILED(189), Host ovirt3.example.org reports about one of 
the Active Storage Domains as Problematic.
2018-10-15 14:44:54,827+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-46) [6d9504d1] EVENT_ID: 
VDS_SET_NONOPERATIONAL_DOMAIN(522), Host ovirt3.example.org cannot access the 
Storage Domain(s) DOMAIN1 attached to the Data Center Default. Setting Host 
state to Non-Operational.
2018-10-15 14:44:54,840+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-46) [6d9504d1] EVENT_ID: 
CONNECT_STORAGE_POOL_FAILED(995), Failed to connect Host ovirt3.example.org to 
Storage Pool Default
2018-10-15 14:45:28,698+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-87) [] EVENT_ID: 
VM_NOT_RESPONDING(126), VM HostedEngine is not responding.
2018-10-15 14:45:30,296+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-72) [] EVENT_ID: 
VM_NOT_RESPONDING(126), VM vm2 is not responding.
2018-10-15 14:45:30,362+03 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engineScheduled-Thread-72) [] EVENT_ID: 
VM_NOT_RESPONDING(126), VM vm3 is not responding.


___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XK7YX6FINFOKA7WGK2ST7KGTCICS6M25/


[ovirt-users] oVirt node not installed

2018-07-16 Thread Spickiy Nikita
Hi, i try install oVirt node, but getting errors on step post-install script. 
Screen error in attachment.
 [cid:944CCD96-B8FF-4ABA-AF17-FA6A7DC184BB@digital-ecosystems.ru] 
[cid:8D77635B-AF62-4894-A9E2-F57E9F290DE5@digital-ecosystems.ru]
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2NAYW6CR72YP6ADQMCID7ANH2D756WZT/


[ovirt-users] Re: Trying to update a host. (ovirt4.1)

2018-06-20 Thread Spickiy Nikita
Disable EPEL repo and repeat update

On 19 Jun 2018, at 22:26, Jacob Green 
mailto:jgr...@aasteel.com>> wrote:


I just did not know where to look for the errors, I now see that it is telling 
me it is failing on this package "collectd"



So when I go to my host and I run yum list collectd I see that collectd is 
available to install via EPEL repos. Note: I did not setup this cluster not 
sure if epel is normal.




So looks like my problem here has to do with the epel package being available 
and being newer?


Thank you all.


On 06/19/2018 09:40 AM, Jacob Green wrote:
When I right click an empty host, and select upgrade and confirm that I 
want to upgrade. It simply comes back with install failed after a minute or 
two. I have no idea why its failing, there also does not appear to be anything 
in /var/log/yum.log so I am not sure where else to look to figure out why it 
cannot upgrade. Also to be clear, the wording in Ovirt uses the term upgrade, 
however I am under the impression it simply means update, not actually upgrade 
to 4.2.



Thank you all.


___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to 
users-le...@ovirt.org
Privacy Statement: 
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZFVQXUEYBCHORCRKYRBU4JTLQZPFACK3/


--
Jacob Green

Systems Admin

American Alloy Steel

713-300-5690

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to 
users-le...@ovirt.org
Privacy Statement: 
https://nam03.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.ovirt.org%2Fsite%2Fprivacy-policy%2F=02%7C01%7C%7C7e6324f27cec4c61e53008d5d5f96b0b%7C84df9e7fe9f640afb435%7C1%7C0%7C636650189570671175=f21XohOzUZFSHeh1DXADEONFkZy9hVk8OVB0m9lCaKE%3D=0
oVirt Code of Conduct: 
https://nam03.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.ovirt.org%2Fcommunity%2Fabout%2Fcommunity-guidelines%2F=02%7C01%7C%7C7e6324f27cec4c61e53008d5d5f96b0b%7C84df9e7fe9f640afb435%7C1%7C0%7C636650189570671175=3iX5GXTxhszloLbTm1MWkOBx%2BQ6diX1oeZkDFoApG%2Bg%3D=0
List Archives: 
https://nam03.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.ovirt.org%2Farchives%2Flist%2Fusers%40ovirt.org%2Fmessage%2F3D3TFBVEV4I7KKPP4JYCVMWZCOGFIY4K%2F=02%7C01%7C%7C7e6324f27cec4c61e53008d5d5f96b0b%7C84df9e7fe9f640afb435%7C1%7C0%7C636650189570671175=%2BfR8yruTYRkY3Ji1asW1V8yiD8D0s9w%2FvL7j537lHP8%3D=0

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VJT7O3UUBNVQYORC4FSBSHNPFKBXBD6W/


[ovirt-users] Error UI in engine

2018-06-14 Thread Spickiy Nikita
Hi, i use engine 4.2.3.8-1.el7 version. When i try opening setting Storage > 
Volumes > MyVolume i get error at the top of the screen.
In log UI https://paste.fedoraproject.org/paste/---B4tGrLbAVYpl1TvR0sg/raw.

How to fix it?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TUUUBCHX2SDLB2NJ2EOMPZCS4ZMWU4H7/


[ovirt-users] Error starting VM after update Ovirt node

2018-06-11 Thread Spickiy Nikita
Hi, i upgraded oVirt node to 4.2.3.1 version. After upgrade I can not migrate 
VM or run VM  on the node. After viewing logs file i find several errors. One 
of them:

2018-06-07 15:52:20,536+07 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(ForkJoinPool-1-worker-3) [] EVENT_ID: VM_DOWN_ERROR(119), VM 
ansible.example.com is down with error. Exit message: internal error: Process 
exited prior to exec: libvirt: Lock Driver error : Failed to ope
n socket to sanlock daemon: Permission denied.

I run setenforce 0 and VM can running on the node.


audit2allow -a


#= setfiles_t ==

# WARNING: 'bin_t' is a base type.
allow setfiles_t bin_t:file entrypoint;
allow setfiles_t self:capability sys_chroot;

#= svirt_t ==

# The file 
'/rhev/data-center/mnt/10.10.10.100:_var_nfs-ovirt-iso-share_files/dbf9ff5d-2cd4-408e-8a9d-08d39fa9a58a/images/----/CentOS-7-x86_64-Minimal-1708.iso'
 is mislabeled on your system.
# Fix with $ restorecon -R -v 
/rhev/data-center/mnt/10.10.10.100:_var_nfs-ovirt-iso-share_files/dbf9ff5d-2cd4-408e-8a9d-08d39fa9a58a/images/----/CentOS-7-x86_64-Minimal-1708.iso
# This avc can be allowed using the boolean 'virt_use_nfs'
allow svirt_t nfs_t:file open;

# The file '/run/sanlock/sanlock.sock' is mislabeled on your system.
# Fix with $ restorecon -R -v /run/sanlock/sanlock.sock
# This avc can be allowed using the boolean 'virt_use_sanlock'
allow svirt_t sanlock_t:unix_stream_socket connectto;

I run restorecon -R -v /run/sanlock/sanlock.sock, but it did not help
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VJCIWULSAP5EHFZPIFQHRECDYS3YZQM5/


[ovirt-users] Off hosted-engine destroy virtualization

2017-08-22 Thread Spickiy Nikita
Hi. I try setting hosted-engine. I make cluster of two node. I run installation 
step by step as written in ovirt documentation. I add storage, iso domain, etc. 
Then i create VM. Everything working, all right. Now i shutdown hosted-engine 
and my VM also not avaliable. It normal behavior? Or i something doing at wrong?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] How to delay start ovirt service?

2017-08-17 Thread Spickiy Nikita
Hi, i try install oVirt hosted-engine. For storage i use iscsi target 
(Infotrend DS1012RE). I make install, but have error to service 
ovirt-ha-broker.service and vdsmd.service:

systemctl -l status ovirt-ha-broker.service
● ovirt-ha-broker.service - oVirt Hosted Engine High Availability 
Communications Broker

▽
   Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-broker.service; enabled; 
vendor preset: disabled)

▽
   Active: active (running) since Thu 2017-08-17 15:28:57 +07; 48s ago

▽
 Main PID: 1926 (ovirt-ha-broker)
   CGroup: /system.slice/ovirt-ha-broker.service
   └─1926 /usr/bin/python 
/usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker --no-daemon

Aug 17 15:28:57 vnode10.pi.local systemd[1]: Started oVirt Hosted Engine High 
Availability Communications Broker.
Aug 17 15:28:57 vnode10.pi.local systemd[1]: Starting oVirt Hosted Engine High 
Availability Communications Broker...
Aug 17 15:29:08 vnode10.pi.local ovirt-ha-broker[1926]: ovirt-ha-broker 
ovirt_hosted_engine_ha.broker.listener.ConnectionHandler ERROR Error handling 
request, data: 'set-storage-domain FilesystemBackend dom_type=iscsi 
sd_uuid=817a5b54-26c1-4d7d-a711-dafdb1ea477d'
Traceback (most recent 
call last):
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/listener.py", 
line 166, in handle
data)
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/listener.py", 
line 299, in _dispatch

.set_storage_domain(client, sd_type, **options)
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/storage_broker.py",
 line 66, in set_storage_domain

self._backends[client].connect()
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/storage_backends.py",
 line 470, in connect
for lv in 
os.listdir(os.path.join("/dev", uuid)):
OSError: [Errno 2] No 
such file or directory: '/dev/817a5b54-26c1-4d7d-a711-dafdb1ea477d’

systemctl -l status vdsmd.service
● vdsmd.service - Virtual Desktop Server Manager
   Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled; vendor 
preset: enabled)
   Active: active (running) since Thu 2017-08-17 15:29:07 +07; 1min 30s ago
  Process: 3686 ExecStartPre=/usr/libexec/vdsm/vdsmd_init_common.sh --pre-start 
(code=exited, status=0/SUCCESS)
 Main PID: 3759 (vdsm)
   CGroup: /system.slice/vdsmd.service
   └─3759 /usr/bin/python2 /usr/share/vdsm/vdsm

Aug 17 15:29:06 vnode10.pi.local vdsmd_init_common.sh[3686]: vdsm: Running 
syslog_available
Aug 17 15:29:06 vnode10.pi.local vdsmd_init_common.sh[3686]: vdsm: Running 
nwfilter
Aug 17 15:29:07 vnode10.pi.local vdsmd_init_common.sh[3686]: vdsm: Running 
dummybr
Aug 17 15:29:07 vnode10.pi.local vdsmd_init_common.sh[3686]: vdsm: Running 
tune_system
Aug 17 15:29:07 vnode10.pi.local vdsmd_init_common.sh[3686]: vdsm: Running 
test_space
Aug 17 15:29:07 vnode10.pi.local vdsmd_init_common.sh[3686]: vdsm: Running 
test_lo
Aug 17 15:29:07 vnode10.pi.local systemd[1]: Started Virtual Desktop Server 
Manager.
Aug 17 15:29:08 vnode10.pi.local vdsm[3759]: vdsm throttled WARN MOM not 
available.
Aug 17 15:29:08 vnode10.pi.local vdsm[3759]: vdsm throttled WARN MOM not 
available, KSM stats will be missing.
Aug 17 15:29:08 vnode10.pi.local vdsm[3759]: vdsm root ERROR failed to retrieve 
Hosted Engine HA info
 Traceback (most recent call last):
   File 
"/usr/lib/python2.7/site-packages/vdsm/host/api.py", line 231, in _getHaInfo
 stats = 
instance.get_all_stats()
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", 
line 103, in get_all_stats
 
self._configure_broker_conn(broker)
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/client/client.py", 
line 180, in _configure_broker_conn
 dom_type=dom_type)
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", 
line 177, in set_storage_domain
 .format(sd_type, options, e))
 RequestError: Failed to set 
storage domain FilesystemBackend, options 

Re: [ovirt-users] Hosted-engine is not migrate to other hosts

2017-08-15 Thread Spickiy Nikita
Hi, thanks for the answer. I decided to full re-insall ovirt, as i did errors 
in the install. As i have error vdsmd service:

● vdsmd.service - Virtual Desktop Server Manager
   Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled; vendor 
preset: enabled)
   Active: active (running) since Mon 2017-08-14 16:12:04 +07; 55min ago
 Main PID: 3901 (vdsm)
   CGroup: /system.slice/vdsmd.service
   └─3901 /usr/bin/python2 /usr/share/vdsm/vdsm

Aug 14 16:12:03 vnode10.pi.local vdsmd_init_common.sh[3829]: vdsm: Running 
syslog_available
Aug 14 16:12:03 vnode10.pi.local vdsmd_init_common.sh[3829]: vdsm: Running 
nwfilter
Aug 14 16:12:04 vnode10.pi.local vdsmd_init_common.sh[3829]: vdsm: Running 
dummybr
Aug 14 16:12:04 vnode10.pi.local vdsmd_init_common.sh[3829]: vdsm: Running 
tune_system
Aug 14 16:12:04 vnode10.pi.local vdsmd_init_common.sh[3829]: vdsm: Running 
test_space
Aug 14 16:12:04 vnode10.pi.local vdsmd_init_common.sh[3829]: vdsm: Running 
test_lo
Aug 14 16:12:04 vnode10.pi.local systemd[1]: Started Virtual Desktop Server 
Manager.
Aug 14 16:12:05 vnode10.pi.local vdsm[3901]: vdsm throttled WARN MOM not 
available.
Aug 14 16:12:05 vnode10.pi.local vdsm[3901]: vdsm throttled WARN MOM not 
available, KSM stats will be missing.
Aug 14 16:12:05 vnode10.pi.local vdsm[3901]: vdsm root ERROR failed to retrieve 
Hosted Engine HA info
 Traceback (most recent call last):
   File 
"/usr/lib/python2.7/site-packages/vdsm/host/api.py", line 231, in _getHaInfo…


But when i removed 1 host for re-install, then service get status is ok:

● vdsmd.service - Virtual Desktop Server Manager
   Loaded: loaded (/usr/lib/systemd/system/vdsmd.service; enabled; vendor 
preset: enabled)
   Active: active (running) since Mon 2017-08-14 13:15:43 +07; 1 day 1h ago
 Main PID: 29945 (vdsm)
   CGroup: /system.slice/vdsmd.service
   ├─  647 /usr/bin/dd 
if=/rhev/data-center/mnt/10.10.20.25:_var_nfs-ovirt-iso-share_files/78c7fad8-4c1a-4f08-9b04-df660e1479c0/dom_md/metadata
 of=/dev/null bs=4096 count=1 iflag=direct
   ├─ 1776 /usr/libexec/ioprocess --read-pipe-fd 59 --write-pipe-fd 53 
--max-threads 10 --max-queued-requests 10
   ├─ 7806 /usr/libexec/ioprocess --read-pipe-fd 65 --write-pipe-fd 64 
--max-threads 10 --max-queued-requests 10
   └─29945 /usr/bin/python2 /usr/share/vdsm/vdsm

Aug 15 13:26:05 vnode11.pi.local vdsm[29945]: vdsm vds.dispatcher WARN 
unhandled close event
Aug 15 13:26:05 vnode11.pi.local vdsm[29945]: vdsm vds.dispatcher WARN 
unhandled close event
Aug 15 13:28:00 vnode11.pi.local vdsm[29945]: vdsm vds.dispatcher WARN 
unhandled close event
Aug 15 13:28:00 vnode11.pi.local vdsm[29945]: vdsm vds.dispatcher WARN 
unhandled close event
Aug 15 13:28:00 vnode11.pi.local vdsm[29945]: vdsm vds.dispatcher WARN 
unhandled close event
Aug 15 13:28:00 vnode11.pi.local vdsm[29945]: vdsm vds.dispatcher WARN 
unhandled close event
Aug 15 14:46:42 vnode11.pi.local vdsm[29945]: vdsm root WARN File: 
/var/lib/libvirt/qemu/channels/aa47c2e6-73ad-4c37-a641-adf2b127fd03.com.redhat.rhevm.vdsm
 already removed
Aug 15 14:46:42 vnode11.pi.local vdsm[29945]: vdsm root WARN File: 
/var/lib/libvirt/qemu/channels/aa47c2e6-73ad-4c37-a641-adf2b127fd03.org.qemu.guest_agent.0
 already removed
Aug 15 14:48:03 vnode11.pi.local vdsm[29945]: vdsm root WARN File: 
/var/lib/libvirt/qemu/channels/aa47c2e6-73ad-4c37-a641-adf2b127fd03.com.redhat.rhevm.vdsm
 already removed
Aug 15 14:48:03 vnode11.pi.local vdsm[29945]: vdsm root WARN File: 
/var/lib/libvirt/qemu/channels/aa47c2e6-73ad-4c37-a641-adf2b127fd03.org.qemu.guest_agent.0
 already removed

But when i shutdown host, on which run hosted-engine, he is not up to the 
available host. I waited 30 minutes, biut not up. I re-install ovirt without 
errors and watch behavior. If situation repeat i writing about it. Thanks for 
help!

On 15 Aug 2017, at 14:35, Yedidyah Bar David 
<d...@redhat.com<mailto:d...@redhat.com>> wrote:

On Mon, Aug 14, 2017 at 12:04 PM, Spickiy Nikita 
<n.spic...@outlook.com<mailto:n.spic...@outlook.com>> wrote:
Hi, in advance sorry, i beginner. I try setting HA on oVirt. I install 
hosted-engine on host and after adding two hosts in web portal. When i install 
hosts then select DEPLOY in hosted-engine. But when i shutdown node, on which 
run hosted-engine, and hosted-engine is not start on other hosts automaticlly.

How long did you wait? It does take up to something like 10 minutes
normally, IIRC.

I guess that trouble in state host(he is state=EngineDown),

No, this is ok.

but not found information how to fix it. Maybe it do because i don't setting 
power management?

No. Power management is required for HA inside the engine, for other VMs.
For hosted-engine it's not required.

If you still have problems, please check/share
/var/log/ovirt-hosted-engine-ha/agent.log
on all 3 hosts.

Best,


My

[ovirt-users] Hosted-engine is not migrate to other hosts

2017-08-14 Thread Spickiy Nikita
Hi, in advance sorry, i beginner. I try setting HA on oVirt. I install 
hosted-engine on host and after adding two hosts in web portal. When i install 
hosts then select DEPLOY in hosted-engine. But when i shutdown node, on which 
run hosted-engine, and hosted-engine is not start on other hosts automaticlly. 
I guess that trouble in state host(he is state=EngineDown), but not found 
information how to fix it. Maybe it do because i don't setting power management?

My configuration:
--== Host 1 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : vnode10
Host ID: 1
Engine status  : {"health": "good", "vm": "up", "detail": 
"up"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 68e523f8
local_conf_timestamp   : 178768
Host timestamp : 178753
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=178753 (Mon Aug 14 15:22:49 2017)
host-id=1
score=3400
vm_conf_refresh_time=178768 (Mon Aug 14 15:23:04 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False


--== Host 2 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : vnode11
Host ID: 2
Engine status  : {"reason": "vm not running on this host", 
"health": "bad", "vm": "down", "detail": "unknown"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 1204d5c0
local_conf_timestamp   : 9290
Host timestamp : 9274
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=9274 (Mon Aug 14 15:22:53 2017)
host-id=2
score=3400
vm_conf_refresh_time=9290 (Mon Aug 14 15:23:10 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False


--== Host 3 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : vnode13
Host ID: 3
Engine status  : {"reason": "vm not running on this host", 
"health": "bad", "vm": "down", "detail": "unknown"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 14e97435
local_conf_timestamp   : 188749
Host timestamp : 188732
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=188732 (Mon Aug 14 15:22:55 2017)
host-id=3
score=3400
vm_conf_refresh_time=188749 (Mon Aug 14 15:23:11 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False

Thank you in advance!
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users