[ovirt-users]Отн: HyperConverged Self-Hosted deployment fails

2019-01-19 Thread Strahil Nikolov
Hi Again,
it seems that sanlock error -223 indicated sanlock lockspace error.I have 
somehow reinitialize the lockspace and the engine is up and running, but I have 
2 VMs defined :1. The engine itself 2. A VM called "External-HostedEngineLocal"
I'm pretty sure that there are some tasks that the wizard completes after 
successfull power-on of the engine , which should clean up the situation and in 
my case - is not actually working.
Could someone advise how to get rid of that VM and what should I do in order to 
complete the deployment.
Thanks in advance for all who read this thread.
Best Regards,Strahil Nikolov

  От: Strahil Nikolov 
 До: Simone Tiraboschi  
Копие: users 
 Изпратен: събота, 19 януари 2019 г. 23:34
 Тема: Отн: [ovirt-users] HyperConverged Self-Hosted deployment fails
   
Hello All,
it seems that the ovirt-ha-broker has some problems:Thread-8::DEBUG::2019-01-19 
19:30:16,048::stompreactor::479::jsonrpc.AsyncoreClient::(send) Sending response
...skipping...
smtp-server = localhost
smtp-port = 25
source-email = root@localhost
destination-emails = root@localhost

[notify]
state_transition = maintenance|start|stop|migrate|up|down

Listener::DEBUG::2019-01-19 
19:30:31,741::heconflib::95::ovirt_hosted_engine_ha.broker.notifications.Notifications.config.broker::(_dd_pipe_tar)
 stderr
: 
Thread-3::DEBUG::2019-01-19 
19:30:31,747::stompreactor::479::jsonrpc.AsyncoreClient::(send) Sending response
StatusStorageThread::ERROR::2019-01-19 
19:30:31,751::status_broker::90::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(run)
 Failed t
o update state.
Traceback (most recent call last):
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py",
 line 82, in run
if (self._status_broker._inquire_whiteboard_lock() or
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py",
 line 190, in _inquire_whiteboard_lock
self.host_id, self._lease_file)
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py",
 line 128, in host_id
raise ex.HostIdNotLockedError("Host id is not set")
HostIdNotLockedError: Host id is not set
StatusStorageThread::ERROR::2019-01-19 
19:30:31,751::status_broker::70::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(trigger_restart)
 Trying to restart the brokerAnd most probably the issue is within the sanlock:
2019-01-19 19:29:57 4739 [4602]: worker0 aio collect WR 
0x7f92a8c0:0x7f92a8d0:0x7f92acc7 result 1048576:0 other free
2019-01-19 19:30:01 4744 [4603]: s8 lockspace 
hosted-engine:1:/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde:0
2019-01-19 19:30:01 4744 [2779]: verify_leader 1 wrong magic 0 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
2019-01-19 19:30:01 4744 [2779]: leader1 delta_acquire_begin error -223 
lockspace hosted-engine host_id 1
2019-01-19 19:30:01 4744 [2779]: leader2 path 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
 offset 0
2019-01-19 19:30:01 4744 [2779]: leader3 m 0 v 30003 ss 512 nh 0 mh 1 oi 0 og 0 
lv 0
2019-01-19 19:30:01 4744 [2779]: leader4 sn hosted-engine rn  ts 0 cs 60346c59
2019-01-19 19:30:02 4745 [4603]: s8 add_lockspace fail result -223
2019-01-19 19:30:07 4750 [4603]: s9 lockspace 
hosted-engine:1:/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde:0
2019-01-19 19:30:07 4750 [2837]: verify_leader 1 wrong magic 0 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
2019-01-19 19:30:07 4750 [2837]: leader1 delta_acquire_begin error -223 
lockspace hosted-engine host_id 1
2019-01-19 19:30:07 4750 [2837]: leader2 path 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
 offset 0
2019-01-19 19:30:07 4750 [2837]: leader3 m 0 v 30003 ss 512 nh 0 mh 1 oi 0 og 0 
lv 0
2019-01-19 19:30:07 4750 [2837]: leader4 sn hosted-engine rn  ts 0 cs 60346c59
2019-01-19 19:30:08 4751 [4603]: s9 add_lockspace fail result -223
Can someone guide me how to go further ? Can debug be enabled for sanlock ?
Best Regards,Strahil Nikolov

  От: Strahil Nikolov 
 До: Simone Tiraboschi  
Копие: users 
 Изпратен: събота, 19 януари 2019 г. 17:54
 Тема: Отн: [ovirt-users] HyperConverged Self-Hosted deployment fails
  
Thanks Simone,
I will check the broker.I didn't specify the layout correctly - it's 'replica 3 
arbiter 1' which was OK last time I used this layout.
Best Regards,Strahil Nikolov

  От: Simone Tiraboschi 
 До: hunter86bg  
Копие: users 
 Изпратен: събота, 19 януари 2019 г. 17:42
 Тема: Re: [ovirt-users] HyperCon

[ovirt-users]Отн: HyperConverged Self-Hosted deployment fails

2019-01-19 Thread Strahil Nikolov
Hello All,
it seems that the ovirt-ha-broker has some problems:Thread-8::DEBUG::2019-01-19 
19:30:16,048::stompreactor::479::jsonrpc.AsyncoreClient::(send) Sending response
...skipping...
smtp-server = localhost
smtp-port = 25
source-email = root@localhost
destination-emails = root@localhost

[notify]
state_transition = maintenance|start|stop|migrate|up|down

Listener::DEBUG::2019-01-19 
19:30:31,741::heconflib::95::ovirt_hosted_engine_ha.broker.notifications.Notifications.config.broker::(_dd_pipe_tar)
 stderr
: 
Thread-3::DEBUG::2019-01-19 
19:30:31,747::stompreactor::479::jsonrpc.AsyncoreClient::(send) Sending response
StatusStorageThread::ERROR::2019-01-19 
19:30:31,751::status_broker::90::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(run)
 Failed t
o update state.
Traceback (most recent call last):
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py",
 line 82, in run
if (self._status_broker._inquire_whiteboard_lock() or
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py",
 line 190, in _inquire_whiteboard_lock
self.host_id, self._lease_file)
  File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/broker/status_broker.py",
 line 128, in host_id
raise ex.HostIdNotLockedError("Host id is not set")
HostIdNotLockedError: Host id is not set
StatusStorageThread::ERROR::2019-01-19 
19:30:31,751::status_broker::70::ovirt_hosted_engine_ha.broker.status_broker.StatusBroker.Update::(trigger_restart)
 Trying to restart the brokerAnd most probably the issue is within the sanlock:
2019-01-19 19:29:57 4739 [4602]: worker0 aio collect WR 
0x7f92a8c0:0x7f92a8d0:0x7f92acc7 result 1048576:0 other free
2019-01-19 19:30:01 4744 [4603]: s8 lockspace 
hosted-engine:1:/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde:0
2019-01-19 19:30:01 4744 [2779]: verify_leader 1 wrong magic 0 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
2019-01-19 19:30:01 4744 [2779]: leader1 delta_acquire_begin error -223 
lockspace hosted-engine host_id 1
2019-01-19 19:30:01 4744 [2779]: leader2 path 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
 offset 0
2019-01-19 19:30:01 4744 [2779]: leader3 m 0 v 30003 ss 512 nh 0 mh 1 oi 0 og 0 
lv 0
2019-01-19 19:30:01 4744 [2779]: leader4 sn hosted-engine rn  ts 0 cs 60346c59
2019-01-19 19:30:02 4745 [4603]: s8 add_lockspace fail result -223
2019-01-19 19:30:07 4750 [4603]: s9 lockspace 
hosted-engine:1:/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde:0
2019-01-19 19:30:07 4750 [2837]: verify_leader 1 wrong magic 0 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
2019-01-19 19:30:07 4750 [2837]: leader1 delta_acquire_begin error -223 
lockspace hosted-engine host_id 1
2019-01-19 19:30:07 4750 [2837]: leader2 path 
/var/run/vdsm/storage/b388324b-eaf4-4158-8e1b-0b7c9b861002/5a849a03-ecbc-4b3f-b558-ec2ebbc42c6f/dd663799-36ed-47de-8208-d357f803efde
 offset 0
2019-01-19 19:30:07 4750 [2837]: leader3 m 0 v 30003 ss 512 nh 0 mh 1 oi 0 og 0 
lv 0
2019-01-19 19:30:07 4750 [2837]: leader4 sn hosted-engine rn  ts 0 cs 60346c59
2019-01-19 19:30:08 4751 [4603]: s9 add_lockspace fail result -223
Can someone guide me how to go further ? Can debug be enabled for sanlock ?
Best Regards,Strahil Nikolov

  От: Strahil Nikolov 
 До: Simone Tiraboschi  
Копие: users 
 Изпратен: събота, 19 януари 2019 г. 17:54
 Тема: Отн: [ovirt-users] HyperConverged Self-Hosted deployment fails
   
Thanks Simone,
I will check the broker.I didn't specify the layout correctly - it's 'replica 3 
arbiter 1' which was OK last time I used this layout.
Best Regards,Strahil Nikolov

  От: Simone Tiraboschi 
 До: hunter86bg  
Копие: users 
 Изпратен: събота, 19 януари 2019 г. 17:42
 Тема: Re: [ovirt-users] HyperConverged Self-Hosted deployment fails
  


On Sat, Jan 19, 2019 at 1:07 PM  wrote:

Hello Community,

recently I managed somehow to deploy a 2 node cluster on GlusterFS , but after 
a serious engine failiure - I have decided to start from scratch.


2 node hyperconverged gluster is definitively a bad idea since it's not going 
to protect you from split brains.Please choose 1 or 3 but not 2. 
What I have done so far:
1. Inctall CentOS7 from scratch
2. Add ovirt repositories, vdo,cockpit for ovirt
3. Deployed the gluster cluster using cockpit 
4. Trying to deploy the hosted-engine , which has failed several times.


Without any logs it's difficult to guess what really happened but I think that 
it could be related to the two nodes approach which is explicitly prevented. 

Up to now I have detecte

[ovirt-users] Re: Host non-responsive after yum update CentOS7/Ovirt3.6

2019-01-19 Thread Jason Herring
If I try to run "vdsm-tool restore-nets", which is what starting the 
vdsm-network.service seems to do first, I get the following:


( a large number of lines of the first error)

..

libvirt: XML-RPC error : authentication failed: authentication failed
libvirt: XML-RPC error : authentication failed: authentication failed
libvirt: XML-RPC error : authentication failed: authentication failed
libvirt: XML-RPC error : authentication failed: authentication failed
Traceback (most recent call last):
  File "/usr/share/vdsm/vdsm-restore-net-config", line 476, in 
    restore(args)
  File "/usr/share/vdsm/vdsm-restore-net-config", line 434, in restore
    _restore_sriov_numvfs()
  File "/usr/share/vdsm/vdsm-restore-net-config", line 84, in 
_restore_sriov_numvfs

    sriov_devices = _get_sriov_devices()
  File "/usr/share/vdsm/vdsm-restore-net-config", line 56, in 
_get_sriov_devices

    devices = hostdev.list_by_caps()
  File "/usr/share/vdsm/hostdev.py", line 219, in list_by_caps
    libvirt_devices = _get_devices_from_libvirt()
  File "/usr/share/vdsm/hostdev.py", line 204, in _get_devices_from_libvirt
    for device in libvirtconnection.get().listAllDevices(0))
  File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", 
line 164, in get

    password)
  File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", 
line 100, in open_connection

    return utils.retry(libvirtOpen, timeout=10, sleep=0.2)
  File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 959, in retry
    return func()
  File "/usr/lib64/python2.7/site-packages/libvirt.py", line 104, in 
openAuth

    if ret is None:raise libvirtError('virConnectOpenAuth() failed')
libvirt.libvirtError: authentication failed: authentication failed
Traceback (most recent call last):
  File "/usr/bin/vdsm-tool", line 219, in main
    return tool_command[cmd]["command"](*args)
  File "/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", 
line 41, in restore_command

    exec_restore(cmd)
  File "/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", 
line 54, in exec_restore

    raise EnvironmentError('Failed to restore the persisted networks')
EnvironmentError: Failed to restore the persisted networks
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/N5JHE4RECU56UJHFXNSOOWQLJS23HZCN/


[ovirt-users] Re: Host non-responsive after yum update CentOS7/Ovirt3.6

2019-01-19 Thread Jason Herring
So, I try to manually start vdsm-network.service and see this, 
suggesting I look in upgrade.log:



# systemctl status vdsm-network.service
● vdsm-network.service - Virtual Desktop Server Manager network restoration
   Loaded: loaded (/usr/lib/systemd/system/vdsm-network.service; 
enabled; vendor preset: enabled)
   Active: failed (Result: exit-code) since Sat 2019-01-19 10:25:02 
PST; 8s ago
  Process: 6845 ExecStart=/usr/bin/vdsm-tool restore-nets (code=exited, 
status=1/FAILURE)
  Process: 6837 ExecStartPre=/usr/bin/vdsm-tool --vvverbose --append 
--logfile=/var/log/vdsm/upgrade.log upgrade-unified-persistence 
(code=exited, status=0/SUCCESS)

 Main PID: 6845 (code=exited, status=1/FAILURE)

Jan 19 10:25:02 vmserver2 vdsm-tool[6845]: return 
tool_command[cmd]["command"](*args)
Jan 19 10:25:02 vmserver2 vdsm-tool[6845]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 41, 
in restore_command

Jan 19 10:25:02 vmserver2 vdsm-tool[6845]: exec_restore(cmd)
Jan 19 10:25:02 vmserver2 vdsm-tool[6845]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 54, 
in exec_restore
Jan 19 10:25:02 vmserver2 vdsm-tool[6845]: raise 
EnvironmentError('Failed to restore the persisted networks')
Jan 19 10:25:02 vmserver2 vdsm-tool[6845]: EnvironmentError: Failed to 
restore the persisted networks
Jan 19 10:25:02 vmserver2 systemd[1]: vdsm-network.service: main process 
exited, code=exited, status=1/FAILURE
Jan 19 10:25:02 vmserver2 systemd[1]: Failed to start Virtual Desktop 
Server Manager network restoration.
Jan 19 10:25:02 vmserver2 systemd[1]: Unit vdsm-network.service entered 
failed state.

Jan 19 10:25:02 vmserver2 systemd[1]: vdsm-network.service failed.


In upgrade.log I see this:


# cat upgrade.log
MainThread::DEBUG::2016-01-10 
10:54:47,615::upgrade::90::upgrade::(apply_upgrade) Running upgrade 
upgrade-unified-persistence
MainThread::DEBUG::2016-01-10 
10:54:47,623::libvirtconnection::160::root::(get) trying to connect libvirt
MainThread::DEBUG::2016-01-10 
10:54:47,639::netinfo::714::root::(_get_gateway) The gateway 192.168.1.1 
is duplicated for the device em1
MainThread::DEBUG::2016-01-10 10:54:47,647::utils::669::root::(execCmd) 
/sbin/ip route show to 0.0.0.0/0 table main (cwd None)
MainThread::DEBUG::2016-01-10 10:54:47,650::utils::687::root::(execCmd) 
SUCCESS:  = '';  = 0
MainThread::DEBUG::2016-01-10 
10:54:47,651::unified_persistence::46::root::(run) 
upgrade-unified-persistence upgrade persisting networks {} and bondings {}
MainThread::INFO::2016-01-10 
10:54:47,651::netconfpersistence::179::root::(_clearDisk) Clearing 
/var/run/vdsm/netconf/nets/ and /var/run/vdsm/netconf/bonds/
MainThread::DEBUG::2016-01-10 
10:54:47,651::netconfpersistence::187::root::(_clearDisk) No existent 
config to clear.
MainThread::INFO::2016-01-10 
10:54:47,652::netconfpersistence::179::root::(_clearDisk) Clearing 
/var/run/vdsm/netconf/nets/ and /var/run/vdsm/netconf/bonds/
MainThread::DEBUG::2016-01-10 
10:54:47,652::netconfpersistence::187::root::(_clearDisk) No existent 
config to clear.
MainThread::INFO::2016-01-10 
10:54:47,652::netconfpersistence::129::root::(save) Saved new config 
RunningConfig({}, {}) to /var/run/vdsm/netconf/nets/ and 
/var/run/vdsm/netconf/bonds/
MainThread::DEBUG::2016-01-10 10:54:47,652::utils::669::root::(execCmd) 
/usr/share/vdsm/vdsm-store-net-config unified (cwd None)
MainThread::DEBUG::2016-01-10 10:54:47,672::utils::687::root::(execCmd) 
SUCCESS:  = 'cp: cannot stat 
\xe2\x80\x98/var/run/vdsm/netconf\xe2\x80\x99: No such file or 
directory\n';  = 0
MainThread::DEBUG::2016-01-10 
10:54:47,672::upgrade::51::upgrade::(_upgrade_seal) Upgrade 
upgrade-unified-persistence successfully performed


I see some references to allowing duplicate gateways if identical on the 
mailing list...

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CXRDO2A2IL6G2OJPPABWRVSQHLJ2DBMG/


[ovirt-users] Re: Host non-responsive after yum update CentOS7/Ovirt3.6

2019-01-19 Thread Jason Herring
I assumed that if ovirt3.6 required older versions of CentOS7, etc, it 
would not allow the yum update due to dependency.  I guess that's a bad 
assumption.


vdsm does not start in fact.  Here's the journal when attempting to 
start it.



# journalctl -xe
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: libvirt.libvirtError: 
authentication failed: authentication failed
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: Traceback (most recent call 
last):
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: File "/usr/bin/vdsm-tool", 
line 219, in main
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: return 
tool_command[cmd]["command"](*args)
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 41, 
in restore_command

Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: exec_restore(cmd)
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: File 
"/usr/lib/python2.7/site-packages/vdsm/tool/restore_nets.py", line 54, 
in exec_restore
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: raise 
EnvironmentError('Failed to restore the persisted networks')
Jan 19 10:13:36 vmserver2 vdsm-tool[6794]: EnvironmentError: Failed to 
restore the persisted networks
Jan 19 10:13:36 vmserver2 systemd[1]: vdsm-network.service: main process 
exited, code=exited, status=1/FAILURE
Jan 19 10:13:36 vmserver2 systemd[1]: Failed to start Virtual Desktop 
Server Manager network restoration.

-- Subject: Unit vdsm-network.service has failed
-- Defined-By: systemd
-- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel
--
-- Unit vdsm-network.service has failed.
--
-- The result is failed.
Jan 19 10:13:36 vmserver2 systemd[1]: Dependency failed for Virtual 
Desktop Server Manager.

-- Subject: Unit vdsmd.service has failed
-- Defined-By: systemd
-- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel
--
-- Unit vdsmd.service has failed.
--
-- The result is dependency.
Jan 19 10:13:36 vmserver2 systemd[1]: Dependency failed for MOM instance 
configured for VDSM purposes.

-- Subject: Unit mom-vdsm.service has failed
-- Defined-By: systemd
-- Support: http://lists.freedesktop.org/mailman/listinfo/systemd-devel
--
-- Unit mom-vdsm.service has failed.
--
-- The result is dependency.
Jan 19 10:13:36 vmserver2 systemd[1]: Job mom-vdsm.service/start failed 
with result 'dependency'.
Jan 19 10:13:36 vmserver2 systemd[1]: Job vdsmd.service/start failed 
with result 'dependency'.
Jan 19 10:13:36 vmserver2 systemd[1]: Unit vdsm-network.service entered 
failed state.

Jan 19 10:13:36 vmserver2 systemd[1]: vdsm-network.service failed.
Jan 19 10:13:36 vmserver2 polkitd[4789]: Unregistered Authentication 
Agent for unix-process:6781:161167654 (system bus name :1.953, object 
path /org/freedesktop/Policy

lines 2643-2681/2681 (END)
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SRRWD3VRKHC5JGCQ7IGKGW4Q5ELX2QDZ/


[ovirt-users]Отн: HyperConverged Self-Hosted deployment fails

2019-01-19 Thread Strahil Nikolov
Thanks Simone,
I will check the broker.I didn't specify the layout correctly - it's 'replica 3 
arbiter 1' which was OK last time I used this layout.
Best Regards,Strahil Nikolov

  От: Simone Tiraboschi 
 До: hunter86bg  
Копие: users 
 Изпратен: събота, 19 януари 2019 г. 17:42
 Тема: Re: [ovirt-users] HyperConverged Self-Hosted deployment fails
   


On Sat, Jan 19, 2019 at 1:07 PM  wrote:

Hello Community,

recently I managed somehow to deploy a 2 node cluster on GlusterFS , but after 
a serious engine failiure - I have decided to start from scratch.


2 node hyperconverged gluster is definitively a bad idea since it's not going 
to protect you from split brains.Please choose 1 or 3 but not 2. 
What I have done so far:
1. Inctall CentOS7 from scratch
2. Add ovirt repositories, vdo,cockpit for ovirt
3. Deployed the gluster cluster using cockpit 
4. Trying to deploy the hosted-engine , which has failed several times.


Without any logs it's difficult to guess what really happened but I think that 
it could be related to the two nodes approach which is explicitly prevented. 

Up to now I have detected that ovirt-ha-agent is giving:

яну 19 13:54:57 ovirt1.localdomain ovirt-ha-agent[16992]: ovirt-ha-agent 
ovirt_hosted_engine_ha.agent.agent.Agent ERROR Traceback (most recent call 
last):
                                                               File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
131, in _run_agent
                                                                 return 
action(he)
                                                               File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
55, in action_proper
                                                                 return 
he.start_monitoring()
                                                               File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 413, in start_monitoring
                                                                 
self._initialize_broker()
                                                               File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 535, in _initialize_broker
                                                                 
m.get('options', {}))
                                                               File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", 
line 83, in start_monitor
                                                                 .format(type, 
options, e ))
                                                             RequestError: 
Failed to start monitor ping, options {'addr': '192.168.1.1'}: [Errno 2] No 
such file or directory


This simply means that ovirt-ha-agents fails to communicate (in order to send a 
ping to check network connectivity) with ovirt-ha-broker over a unix domain 
socket.
'[Errno 2] No such file or directory' means that the socket is closed on 
ovirt-ha-broker side: you can probably see why checking 
/var/log/ovirt-hosted-engine-ha/broker.log but if didn't successfully completed 
the setup this is not surprising me and I strongly suggest to correctly 
complete the deployment before trying anything else.  

According to https://access.redhat.com/solutions/3353391 , the 
/etc/ovirt-hosted-engine/hosted-engine.conf should be empty , but it's OK:

[root@ovirt1 tmp]# cat /etc/ovirt-hosted-engine/hosted-engine.conf
fqdn=engine.localdomain
vm_disk_id=bb0a9839-a05d-4d0a-998c-74da539a9574
vm_disk_vol_id=c1fc3c59-bc6e-4b74-a624-557a1a62a34f
vmid=d0e695da-ec1a-4d6f-b094-44a8cac5f5cd
storage=ovirt1.localdomain:/engine
nfs_version=
mnt_options=backup-volfile-servers=ovirt2.localdomain:ovirt3.localdomain
conf=/var/run/ovirt-hosted-engine-ha/vm.conf
host_id=1
console=vnc
domainType=glusterfs
spUUID=----
sdUUID=444e524e-9008-48f8-b842-1ce7b95bf248
connectionUUID=e29cf818-5ee5-46e1-85c1-8aeefa33e95d
ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
ca_subject="C=EN, L=Test, O=Test, CN=Test"
vdsm_use_ssl=true
gateway=192.168.1.1
bridge=ovirtmgmt
metadata_volume_UUID=a3be2390-017f-485b-8f42-716fb6094692
metadata_image_UUID=368fb8dc-6049-4ef0-8cf8-9d3c4d772d59
lockspace_volume_UUID=41762f85-5d00-488f-bcd0-3de49ec39e8b
lockspace_image_UUID=de100b9b-07ac-4986-9d86-603475572510
conf_volume_UUID=4306f6d6-7fe9-499d-81a5-6b354e8ecb79
conf_image_UUID=d090dd3f-fc62-442a-9710-29eeb56b0019

# The following are used only for iSCSI storage
iqn=
portal=
user=
password=
port=

Ovirt-ha-agent version is:
ovirt-hosted-engine-ha-2.2.18-1.el7.noarch

Can you guide me in order to resolve this issue and to deploy the self-hosted 
engine ?
Where should I start from ?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy

[ovirt-users] Re: HyperConverged Self-Hosted deployment fails

2019-01-19 Thread Simone Tiraboschi
On Sat, Jan 19, 2019 at 1:07 PM  wrote:

> Hello Community,
>
> recently I managed somehow to deploy a 2 node cluster on GlusterFS , but
> after a serious engine failiure - I have decided to start from scratch.
>

2 node hyperconverged gluster is definitively a bad idea since it's not
going to protect you from split brains.
Please choose 1 or 3 but not 2.


> What I have done so far:
> 1. Inctall CentOS7 from scratch
> 2. Add ovirt repositories, vdo,cockpit for ovirt
> 3. Deployed the gluster cluster using cockpit
> 4. Trying to deploy the hosted-engine , which has failed several times.
>

Without any logs it's difficult to guess what really happened but I think
that it could be related to the two nodes approach which is explicitly
prevented.


>
> Up to now I have detected that ovirt-ha-agent is giving:
>
> яну 19 13:54:57 ovirt1.localdomain ovirt-ha-agent[16992]: ovirt-ha-agent
> ovirt_hosted_engine_ha.agent.agent.Agent ERROR Traceback (most recent call
> last):
>File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 131, in _run_agent
>  return
> action(he)
>File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 55, in action_proper
>  return
> he.start_monitoring()
>File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 413, in start_monitoring
>
>  self._initialize_broker()
>File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 535, in _initialize_broker
>
>  m.get('options', {}))
>File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py",
> line 83, in start_monitor
>
>  .format(type, options, e ))
>  RequestError:
> Failed to start monitor ping, options {'addr': '192.168.1.1'}: [Errno 2] No
> such file or directory
>

This simply means that ovirt-ha-agents fails to communicate (in order to
send a ping to check network connectivity) with ovirt-ha-broker over a unix
domain socket.
'[Errno 2] No such file or directory' means that the socket is closed on
ovirt-ha-broker side: you can probably see why checking
/var/log/ovirt-hosted-engine-ha/broker.log but if didn't successfully
completed the setup this is not surprising me and I strongly suggest to
correctly complete the deployment before trying anything else.


>
> According to https://access.redhat.com/solutions/3353391 , the
> /etc/ovirt-hosted-engine/hosted-engine.conf should be empty , but it's OK:
>
> [root@ovirt1 tmp]# cat /etc/ovirt-hosted-engine/hosted-engine.conf
> fqdn=engine.localdomain
> vm_disk_id=bb0a9839-a05d-4d0a-998c-74da539a9574
> vm_disk_vol_id=c1fc3c59-bc6e-4b74-a624-557a1a62a34f
> vmid=d0e695da-ec1a-4d6f-b094-44a8cac5f5cd
> storage=ovirt1.localdomain:/engine
> nfs_version=
> mnt_options=backup-volfile-servers=ovirt2.localdomain:ovirt3.localdomain
> conf=/var/run/ovirt-hosted-engine-ha/vm.conf
> host_id=1
> console=vnc
> domainType=glusterfs
> spUUID=----
> sdUUID=444e524e-9008-48f8-b842-1ce7b95bf248
> connectionUUID=e29cf818-5ee5-46e1-85c1-8aeefa33e95d
> ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
> ca_subject="C=EN, L=Test, O=Test, CN=Test"
> vdsm_use_ssl=true
> gateway=192.168.1.1
> bridge=ovirtmgmt
> metadata_volume_UUID=a3be2390-017f-485b-8f42-716fb6094692
> metadata_image_UUID=368fb8dc-6049-4ef0-8cf8-9d3c4d772d59
> lockspace_volume_UUID=41762f85-5d00-488f-bcd0-3de49ec39e8b
> lockspace_image_UUID=de100b9b-07ac-4986-9d86-603475572510
> conf_volume_UUID=4306f6d6-7fe9-499d-81a5-6b354e8ecb79
> conf_image_UUID=d090dd3f-fc62-442a-9710-29eeb56b0019
>
> # The following are used only for iSCSI storage
> iqn=
> portal=
> user=
> password=
> port=
>
> Ovirt-ha-agent version is:
> ovirt-hosted-engine-ha-2.2.18-1.el7.noarch
>
> Can you guide me in order to resolve this issue and to deploy the
> self-hosted engine ?
> Where should I start from ?
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/D72UQFMNOEJJOPGDDAAUTM73ADLGOBR2/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/pri

[ovirt-users] HyperConverged Self-Hosted deployment fails

2019-01-19 Thread hunter86_bg
Hello Community,

recently I managed somehow to deploy a 2 node cluster on GlusterFS , but after 
a serious engine failiure - I have decided to start from scratch.
What I have done so far:
1. Inctall CentOS7 from scratch
2. Add ovirt repositories, vdo,cockpit for ovirt
3. Deployed the gluster cluster using cockpit 
4. Trying to deploy the hosted-engine , which has failed several times.

Up to now I have detected that ovirt-ha-agent is giving:

яну 19 13:54:57 ovirt1.localdomain ovirt-ha-agent[16992]: ovirt-ha-agent 
ovirt_hosted_engine_ha.agent.agent.Agent ERROR Traceback (most recent call 
last):
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
131, in _run_agent
 return 
action(he)
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py", line 
55, in action_proper
 return 
he.start_monitoring()
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 413, in start_monitoring
 
self._initialize_broker()
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
 line 535, in _initialize_broker
 
m.get('options', {}))
   File 
"/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/lib/brokerlink.py", 
line 83, in start_monitor
 .format(type, 
options, e ))
 RequestError: 
Failed to start monitor ping, options {'addr': '192.168.1.1'}: [Errno 2] No 
such file or directory

According to https://access.redhat.com/solutions/3353391 , the 
/etc/ovirt-hosted-engine/hosted-engine.conf should be empty , but it's OK:

[root@ovirt1 tmp]# cat /etc/ovirt-hosted-engine/hosted-engine.conf
fqdn=engine.localdomain
vm_disk_id=bb0a9839-a05d-4d0a-998c-74da539a9574
vm_disk_vol_id=c1fc3c59-bc6e-4b74-a624-557a1a62a34f
vmid=d0e695da-ec1a-4d6f-b094-44a8cac5f5cd
storage=ovirt1.localdomain:/engine
nfs_version=
mnt_options=backup-volfile-servers=ovirt2.localdomain:ovirt3.localdomain
conf=/var/run/ovirt-hosted-engine-ha/vm.conf
host_id=1
console=vnc
domainType=glusterfs
spUUID=----
sdUUID=444e524e-9008-48f8-b842-1ce7b95bf248
connectionUUID=e29cf818-5ee5-46e1-85c1-8aeefa33e95d
ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
ca_subject="C=EN, L=Test, O=Test, CN=Test"
vdsm_use_ssl=true
gateway=192.168.1.1
bridge=ovirtmgmt
metadata_volume_UUID=a3be2390-017f-485b-8f42-716fb6094692
metadata_image_UUID=368fb8dc-6049-4ef0-8cf8-9d3c4d772d59
lockspace_volume_UUID=41762f85-5d00-488f-bcd0-3de49ec39e8b
lockspace_image_UUID=de100b9b-07ac-4986-9d86-603475572510
conf_volume_UUID=4306f6d6-7fe9-499d-81a5-6b354e8ecb79
conf_image_UUID=d090dd3f-fc62-442a-9710-29eeb56b0019

# The following are used only for iSCSI storage
iqn=
portal=
user=
password=
port=

Ovirt-ha-agent version is:
ovirt-hosted-engine-ha-2.2.18-1.el7.noarch

Can you guide me in order to resolve this issue and to deploy the self-hosted 
engine ?
Where should I start from ?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/D72UQFMNOEJJOPGDDAAUTM73ADLGOBR2/


[ovirt-users] Re: EXTERNAL: Re: Single Node Deployment - Self-Hosted-Engine

2019-01-19 Thread Simone Tiraboschi
On Fri, Jan 18, 2019 at 8:50 PM Albrecht, Thomas C <
thomas.c.albre...@lmco.com> wrote:

> While i realize it’s not ideal, I created an NFS export on the single node
> for demonstration purposes. Very simple to do. At some point, we’ll be
> moving the storage to a NAS.
>
> Can anyone tell me why a single node building using a locally hosted NFS
> server isn’t good for this use case?
>

AFAIK NFS in loopback can still cause deadlock under heavy memory
consumption so you can try it for a test or demo while I'd discourage it
for production environments.


>
> Tom A.
>
> Sent from my iPhone
>
> On Jan 18, 2019, at 1:16 PM, Brian Wilson (briwils2)  > wrote:
>
> Thanks Simone,
>
> I am currently trying to go through that as I found a wiki page that
> talked about single node HCI,
>
> Is there anyway to leave Gluster off the table and just use a local on
> host Storage Domain for simplicity?
>
>
> Thanks
> Brian
>
>
> From: Simone Tiraboschi mailto:stira...@redhat.com>>
> Date: Friday, January 18, 2019 at 11:42 AM
> To: "Brian Wilson (briwils2)"  briwi...@cisco.com>>
> Cc: users mailto:users@ovirt.org>>
> Subject: Re: [ovirt-users] Single Node Deployment - Self-Hosted-Engine
>
>
>
> On Fri, Jan 18, 2019 at 6:18 PM Brian Wilson  briwi...@cisco.com>> wrote:
> I have a question regarding installing a single node and then running the
> self-hosted engine on it.
>
> I would like to keep this as simple as possible and am running into an
> issue with where to tell the installer to place the engine.  It asks me to
> choose a shared storage technology(gfs,nfs, iscsi) and i dont want to rely
> on anything outside of this single box to provide storage.  I would like to
> add another volume (LVM) as a local storage domain to be used for this as
> well as all other VMs to be run on this all in one box.
>
> Is there some secret sauce to getting this installer to choose a local
> location rather than a shared one?
>
> Yes, a replica 1 hyper-converged gluster volume.
>
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org users-le...@ovirt.org>
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/PQQSNOA7JLKKJ53RHN3MHNMXATRKUEOA/
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org users-le...@ovirt.org>
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/TYX7IR7LHI4LOAXUW46DRBUXJDR3DFXQ/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KHPL46TBRIYZE5IW5DRQCYI5KMK5VUDL/