Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 08:29, aleksey.maksi...@it-kb.ru wrote:
> 
> There are more ideas?
> 
> 15.09.2016, 14:40, "aleksey.maksi...@it-kb.ru" :
>> Martin, I physically turned off the server through the iLO2. See screenshots.
>> I did not touch Virtual Machine (KOM-AD01-PBX02) at the same time.
>> The virtual machine has been turned on at the time when the host shut down.
>> 
>> 15.09.2016, 14:27, "Martin Perina" :
>>>  Hi,
>>> 
>>>  I found out this in the log:
>>> 
>>>  2016-09-15 12:02:04,661 INFO  
>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
>>> (ForkJoinPool-1-worker-6) [] VM 
>>> '660bafca-e9c3-4191-99b4-295ff8553488'(KOM-AD01-PBX02) moved from 'Up' --> 
>>> 'Down'
>>>  2016-09-15 12:02:04,788 INFO  
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
>>> (ForkJoinPool-1-worker-6) [] Correlation ID: null, Call Stack: null, Custom 
>>> Event ID: -1, Message: VM KOM-AD01-PBX02 is down. Exit message: User shut 
>>> down from within the guest

since it shut down cleanly, can you please check the guest's logs to see what 
triggered the shutdown? In such cases it is considered a user requested 
shutdown and such VMs are not restarted automatically
We are aware of a similar issue on specific hw - 
https://bugzilla.redhat.com/show_bug.cgi?id=1341106

>>> 
>>>  If I'm not mistaken, this means that VM was properly shutted down from 
>>> within itself and in that case it's not restarted automatically. So I'm 
>>> curious what actions have you made to make host KOM-AD01-VM31 
>>> non-responsive?
>>> 
>>>  If you want to test fencing properly, then I suggest you to either block 
>>> connection between host and engine on host side and forcibly stop ovirtmgmt 
>>> network interface on host and watch fencing is applied.
>>> 
>>>  Martin
>>> 
>>>  On Thu, Sep 15, 2016 at 1:16 PM,  wrote:
  engine.log for this period.
 
  15.09.2016, 14:01, "Martin Perina" :
>  On Thu, Sep 15, 2016 at 12:47 PM,  wrote:
>>  Hi Martin.
>>  I have a stupid question. Use Watchdog device mandatory to 
>> automatically start a virtual machine in host Fencing process?
> 
>  ​AFAIK it's not, but I'm not na expert, adding Arik.
> 
>  You need correct power management setup for the hosts and VM has to be 
> marked as highly available​ for sure.
> 
>>  15.09.2016, 13:43, "Martin Perina" :
>>>  Hi,
>>> 
>>>  could you please share whole engine.log?
>>> 
>>>  Thanks
>>> 
>>>  Martin Perina
>>> 
>>>  On Thu, Sep 15, 2016 at 12:01 PM,  wrote:
  Hello oVirt guru`s !
 
  I have oVirt Hosted Engine 4.0.3-1.el7.centos on two CentOS 7.2 hosts 
 (HP ProLiant DL 360 G5) connected to shared FC SAN Storage.
 
  1. I configured Power Management for the Hosts (successfully added 
 Fencing Agent for iLO2 from my hosts)
 
  2. I created new VM (KOM-AD01-PBX02) and installed Guest OS (Ubuntu 
 Server 16.04 LTS) and oVirt Guest Agent
  (As described herein 
 https://blog.it-kb.ru/2016/09/14/install-ovirt-4-0-part-2-about-data-center-iso-domain-logical-network-vlan-vm-settings-console-guest-agent-live-migration/)
 In VM settings on "High Availability" I turned on the option 
 "Highly Available" and change "Priority" to "High"
 
  3. Now I'm trying to check Hard-Fencing and power off my first host 
 (KOM-AD01-VM31) from his iLO (KOM-AD01-ILO31).
 
  Fencing successfully works and server is automatically turned on, but 
 my HA VM not started on second host (KOM-AD01-VM32).
 
  These events I see in the oVirt web console:
 
  Sep 15, 2016 12:08:13 PMHost KOM-AD01-VM31 power management 
 was verified successfully.
  Sep 15, 2016 12:08:13 PMStatus of host KOM-AD01-VM31 was set 
 to Up.
  Sep 15, 2016 12:08:05 PMExecuting power management status on 
 Host KOM-AD01-VM31 using Proxy Host KOM-AD01-VM32 and Fence Agent 
 ilo:KOM-AD01-ILO31.holding.com.
  Sep 15, 2016 12:05:48 PMHost KOM-AD01-VM31 is rebooting.
  Sep 15, 2016 12:05:48 PMHost KOM-AD01-VM31 was started by 
 SYSTEM.
  Sep 15, 2016 12:05:48 PMPower management start of Host 
 KOM-AD01-VM31 succeeded.
  Sep 15, 2016 12:05:41 PMExecuting power management status on 
 Host KOM-AD01-VM31 using Proxy Host KOM-AD01-VM32 and Fence Agent 
 ilo:KOM-AD01-ILO31.holding.com.
  Sep 15, 2016 12:05:19 PMExecuting power management start on 
 Host KOM-AD01-VM31 using Proxy Host KOM-AD01-VM32 and Fence Agent 
 ilo:KOM-AD01-ILO31.holding.com.
  Sep 15, 2016 12:05:19 PMPower management start of Host 
 KOM-AD01-VM31 initiated.
  Sep 15, 2016 12:05:19 PMAuto fence for host KOM-AD0

Re: [ovirt-users] Support on oVirt

2016-09-16 Thread Yaniv Dary
Remote viewer for spice.

Yaniv Dary
Technical Product Manager
Red Hat Israel Ltd.
34 Jerusalem Road
Building A, 4th floor
Ra'anana, Israel 4350109

Tel : +972 (9) 7692306
8272306
Email: yd...@redhat.com
IRC : ydary

On Sep 16, 2016 10:00, "Johann van der Westhuizen" 
wrote:

> Hi,
>
>
>
> Thanks for the reply.
>
>
>
> I actually did sort out this issue.
>
>
>
> But I have another issue, when I click on console the .vv file is
> downloaded, but when I open the .vv file it asks to open with a program,
> what program should I use to open the file that I can access the console?
>
>
>
> Thank you.
>
>
>
> Regards / Groete,
>
>
>
> *Johann van der Westhuizen*
>
> *Systems Administrator*
>
> [image: cid:image001.png@01D1DB97.0D86FCB0]
>
> *joh...@sharenet.co.za * | +27 (021) 700 4827 |
> www.sharenet.co.za
>
>
>
> [image: cid:image002.png@01D1DE78.2EE61F10]
>  [image:
> cid:image003.jpg@01D1DE78.2EE61F10] 
>
>
>
> *From:* Yaniv Dary [mailto:yd...@redhat.com]
> *Sent:* 15 September 2016 23:07
> *To:* Johann van der Westhuizen 
> *Cc:* users 
> *Subject:* Re: [ovirt-users] Support on oVirt
>
>
>
> Can you share logs? Sos report?
>
> Yaniv Dary
> Technical Product Manager
> Red Hat Israel Ltd.
> 34 Jerusalem Road
> Building A, 4th floor
> Ra'anana, Israel 4350109
>
> Tel : +972 (9) 7692306
> 8272306
> Email: yd...@redhat.com
> IRC : ydary
>
>
>
> On Sep 15, 2016 22:28, "Johann van der Westhuizen" 
> wrote:
>
> Hi,
>
>
>
> Wonder if you could help me with oVirt, having some issues starting a new
> VM, please see below:
>
>
>
>
>
> Getting this error just as I try to start the VM, the green play button.
> What can I do to resolve this issue?
>
>
>
> Regards / Groete,
>
>
>
> *Johann van der Westhuizen*
>
> *Systems Administrator*
>
> [image: cid:image001.png@01D1DB97.0D86FCB0]
>
> *joh...@sharenet.co.za * | +27 (021) 700 4827 |
> www.sharenet.co.za
>
>
>
> [image: cid:image002.png@01D1DE78.2EE61F10]
> [image:
> cid:image003.jpg@01D1DE78.2EE61F10] 
>
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] can't import vm from KVM host

2016-09-16 Thread Nelson Lameiras
I upgraded to oVirt 4.0.4 RC1 to test this issue (engine and hosts) and I still 
have the same problem.

Migration GUI still shows zero disks when the source VM had block disk devices.

Can you confirm that the issue 1362525 has been integrated in this RC?
Should I do some additional actions in order to see/make it work?

Thanks,
Nelson

- Original Message -
From: "Shahar Havivi" 
To: "Nelson Lameiras" , "Nisim Simsolo" 

Cc: users@ovirt.org
Sent: Sunday, August 28, 2016 8:45:48 AM
Subject: Re: [ovirt-users] can't import vm from KVM host

On 25.08.16 11:32, Nelson Lameiras wrote:
> Thanks for your answer. I do not think that I'm experiencing that same bug 
> exactly.
> 
> The issue 1362525 (below) concerns a KVM VM with iSCSI disk being importend 
> to RHEV that does not perform well.
> 
> In my case, when oVirt GUI proposes me the list of machines on KVM host to 
> import, each machine as a "disk count" next to it.
> - This "disk count" is correct if the VM has it's storage on a raw/qcow file 
> (local) - and import goes well
> - But "disk count" is 0 (always) if the VM has it's storage on a block device 
> (ie iscsi) - and import is not possible at all!
> 
> It seems to me that current oVirt migration tool is unable to read/analyse 
> distant VM machines with block device storage... Maybe I'm analysing this 
> issue wrongly.
Yes,
This was the issue as well - seeing zero disks when the source VM had block 
disk devices.

Adding Nisim to verify it.

> 
> Thanks,
> nelson
> - Original Message -
> From: "Shahar Havivi" 
> To: "Nelson Lameiras" 
> Cc: users@ovirt.org
> Sent: Tuesday, August 23, 2016 12:19:13 PM
> Subject: Re: [ovirt-users] can't import vm from KVM host
> 
> On 23.08.16 11:24, Nelson Lameiras wrote:
> > Hello,
> Hi,
> 
> We do have a bug on that issue:
> https://bugzilla.redhat.com/show_bug.cgi?id=1362525
> and its handled.
> 
> Thanks,
>  Shahar.
> > 
> > I'm trying to import virtual machines from a KVM host (centos 7.2) to an 
> > oVirt 4.0.2 Cluster using the "import" feature on GUI.
> > 
> > If the original VM is using RAW/QCOW2 files as storage, everything works 
> > fine. 
> > 
> > But if the original VM is using a special block special device as storage 
> > (like a LVM or SAN volume), it's simply not recognized. 
> > The VM does appear in the import list of the KVM host, but it's disk count 
> > is 0!
> > 
> > Is this a known technical obstacle or am I doing something wrong ?
> > 
> > below is the storage part of the xml describing the original VM :
> > 
> >   
> >   
> >   
> >> function='0x0'/>
> > 
> > 
> >   
> >   
> >   
> >> function='0x0'/>
> > 
> > 
> > We have hundreds of virtual machines in production with this type of 
> > configuration... How can we migrate them safely to oVirt?
> > 
> > thanks
> > 
> > Nelson
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Nested oVirt : Unofficial network recommendations wanted

2016-09-16 Thread Nicolas Ecarnot

Le 15/09/2016 à 23:27, Edward Haas a écrit :



On Thu, Sep 15, 2016 at 3:43 PM, Nicolas Ecarnot > wrote:


Hello,

I'm trying to setup a nested oVirt for the first time, and
according to what I read around and experience, some special
network settings have to be chosen.

For this first try, the bare-metal host is a Debian, running KVM,
and the virtual NICs are setup as macvtap in VEPA mode.


On what device you attached it? Bridge? the physical nic?


On my physical host, ifconfig is showing the following devices :
- lo, obviously
- eth0, primary used
- wlan0, not used, WIFI...
- virbr0
- virbr0-nic

The last two devices were created when installing and playing with the 
KVM Virtual Machine Manager.


When trying to assign one of them to a VM, the only choices are :
- NAT
- eth0: macvtap
- wlan0: macvtap
- custom

and the source mode can be chosen between :
- Bridge
- VEPA
- Private
- Passthrough

My main goal is simplicity, and I'd rather use simple bridging, no NAT, 
simple.


What would be the simplest choice?



I'm not up to date with macvtap, but I think I understood that one
of its limit was that no packet could be exchanged between the
host and the guests. So far, this is leading me to access my own
local VMs from another host. Too bad.

I'm also witnessing frequent loss of packets.

So far, I'm also seeing that guests can not ping each others, so
I'm not going further before having solved these basic issues.


By definition: http://virt.kernelnewbies.org/MacVTap

I'm remembering the good old times of lots of bridges where my VMs
could be reached by anyone (this was desired), but virt manager is
not offering me this choice. I also would like to avoid NAT for
other reasons.

To you all (4) people who are playing with nested oVirt :
- which is your preferred bare metal OS?
- which is your preferred guest (first virt level) OS?
- which network setups and modes are working best?

Thank you.

-- 
Nicolas ECARNOT

___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users






--
Nicolas ECARNOT

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Martin Perina
On Fri, Sep 16, 2016 at 9:26 AM, Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
> > On 16 Sep 2016, at 08:29, aleksey.maksi...@it-kb.ru wrote:
> >
> > There are more ideas?
> >
> > 15.09.2016, 14:40, "aleksey.maksi...@it-kb.ru" <
> aleksey.maksi...@it-kb.ru>:
> >> Martin, I physically turned off the server through the iLO2. See
> screenshots.
> >> I did not touch Virtual Machine (KOM-AD01-PBX02) at the same time.
> >> The virtual machine has been turned on at the time when the host shut
> down.
> >>
> >> 15.09.2016, 14:27, "Martin Perina" :
> >>>  Hi,
> >>>
> >>>  I found out this in the log:
> >>>
> >>>  2016-09-15 12:02:04,661 INFO  [org.ovirt.engine.core.
> vdsbroker.monitoring.VmAnalyzer] (ForkJoinPool-1-worker-6) [] VM
> '660bafca-e9c3-4191-99b4-295ff8553488'(KOM-AD01-PBX02) moved from 'Up'
> --> 'Down'
> >>>  2016-09-15 12:02:04,788 INFO  [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-6) []
> Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: VM
> KOM-AD01-PBX02 is down. Exit message: User shut down from within the guest
>
> since it shut down cleanly, can you please check the guest's logs to see
> what triggered the shutdown? In such cases it is considered a user
> requested shutdown and such VMs are not restarted automatically
>

​That's exactly what I meant by my response. From the log it's obvious that
VM was shutdown properly, so engine will not restart it on a different.
host. Also on most modern hosts if you execute power management off action,
a signal is sent to OS to execute ​

​regular shutdown so VMs are also shutted down properly.
​

> We are aware of a similar issue on specific hw -
> https://bugzilla.redhat.com/show_bug.cgi?id=1341106
>
> >>>
> >>>  If I'm not mistaken, this means that VM was properly shutted down
> from within itself and in that case it's not restarted automatically. So
> I'm curious what actions have you made to make host KOM-AD01-VM31
> non-responsive?
> >>>
> >>>  If you want to test fencing properly, then I suggest you to either
> block connection between host and engine on host side and forcibly stop
> ovirtmgmt network interface on host and watch fencing is applied.
>

​Try above if you want to test fencing. Of course you can always configure
firewall rule to drop all packets between engine and host or unplug host
network cable​.

>>>
> >>>  Martin
> >>>
> >>>  On Thu, Sep 15, 2016 at 1:16 PM,  wrote:
>   engine.log for this period.
> 
>   15.09.2016, 14:01, "Martin Perina" :
> >  On Thu, Sep 15, 2016 at 12:47 PM, 
> wrote:
> >>  Hi Martin.
> >>  I have a stupid question. Use Watchdog device mandatory to
> automatically start a virtual machine in host Fencing process?
> >
> >  ​AFAIK it's not, but I'm not na expert, adding Arik.
> >
> >  You need correct power management setup for the hosts and VM has to
> be marked as highly available​ for sure.
> >
> >>  15.09.2016, 13:43, "Martin Perina" :
> >>>  Hi,
> >>>
> >>>  could you please share whole engine.log?
> >>>
> >>>  Thanks
> >>>
> >>>  Martin Perina
> >>>
> >>>  On Thu, Sep 15, 2016 at 12:01 PM, 
> wrote:
>   Hello oVirt guru`s !
> 
>   I have oVirt Hosted Engine 4.0.3-1.el7.centos on two CentOS 7.2
> hosts (HP ProLiant DL 360 G5) connected to shared FC SAN Storage.
> 
>   1. I configured Power Management for the Hosts (successfully
> added Fencing Agent for iLO2 from my hosts)
> 
>   2. I created new VM (KOM-AD01-PBX02) and installed Guest OS
> (Ubuntu Server 16.04 LTS) and oVirt Guest Agent
>   (As described herein https://blog.it-kb.ru/2016/09/
> 14/install-ovirt-4-0-part-2-about-data-center-iso-domain-
> logical-network-vlan-vm-settings-console-guest-agent-live-migration/)
>  In VM settings on "High Availability" I turned on the option
> "Highly Available" and change "Priority" to "High"
> 
>   3. Now I'm trying to check Hard-Fencing and power off my first
> host (KOM-AD01-VM31) from his iLO (KOM-AD01-ILO31).
> 
>   Fencing successfully works and server is automatically turned
> on, but my HA VM not started on second host (KOM-AD01-VM32).
> 
>   These events I see in the oVirt web console:
> 
>   Sep 15, 2016 12:08:13 PMHost KOM-AD01-VM31 power
> management was verified successfully.
>   Sep 15, 2016 12:08:13 PMStatus of host KOM-AD01-VM31 was
> set to Up.
>   Sep 15, 2016 12:08:05 PMExecuting power management
> status on Host KOM-AD01-VM31 using Proxy Host KOM-AD01-VM32 and Fence Agent
> ilo:KOM-AD01-ILO31.holding.com.
>   Sep 15, 2016 12:05:48 PMHost KOM-AD01-VM31 is rebooting.
>   Sep 15, 2016 12:05:48 PMHost KOM-AD01-VM31 was started
> by SYSTEM.
>   Sep 15, 2016 12:05:48 PMPower management start of Host
> KOM-AD01-V

Re: [ovirt-users] Fail to upgrade to 4.1 from 4.0

2016-09-16 Thread Martin Perina
Hi,

I'm glad that you new attempt to upgrade 3.6 to 4.0 were successfull and
4.0 is now working OK for you.

More answers to questions inline:

On Thu, Sep 15, 2016 at 9:50 PM, Marcin M. Jessa  wrote:

> On 15/09/2016 12:58, Martin Perina wrote:
>
>> ​So something went wrong on that upgrade from 3.6 to 4.0 and  it was not
>> a good idea to do another upgrade when previous one went wrong.
>>
>
> The upgrade process went fine. The upgrade script never showed any errors.
>

​Hmm, that's strange, because from latest setup log you sent me I'd say
that 3.6 to 4.0 upgrade was not successful. But to confirm that I'd need to
investigate all setup logs from 3.6 upgrade ..
​


>
> Anyway to confirm that could you please share with us also upgrade log
>> between
>> 3.6 and 4.0? Also how did you do the upgrade (especially what OS have
>> you used on 3.6 and if upgrade was done using engine-backup or in-place)?
>>
>
> I backed up the installation but then I ran in-place upgrade. I was told
> on the irc I don't need to restore from backups when I'm already on CentOS
> 7. Actually failing to restore from backup brought me to the IRC in the
> first place.
>

​Yes, when doing in-place upgrade you don't need to do restore, that's why
I asked how you performed your upgrade
​


>
> I'm on CentOS Linux release 7.2.1511 (Core)
> Upgrading I followed exactly the same doc you pointed to below.
> Except when updating to 4.1 failed, I uninstalled everything that had to
> do with ovirt and installed the 4.0 release.
>
> More details about upgrade options from 3.6 to 4.0 can be found at:
>> http://www.ovirt.org/documentation/migration-engine-3.6-to-4.0/
>>
> Because of that doc I installed 4.1 instead of the stable 4.0 ...
>

​Yes, that's a shame for our project to have outdated upgrade doc several
month after GA release :-(
Thanks for pointing this out, here's PR which fixes wrong repository URL:

https://github.com/oVirt/ovirt-site/pull/467


>
> ​So from current log I can see that upgrade from 3.6 to 4.0 was not
>> successfully finished and that's why you had issues during upgrade on
>> master.
>>
> Out of curiousity, where did you see that 3.6 to 4.0 update failed?
>

​In your setup log I can see following:

  Dropping materialized views...
  Skipping upgrade script
/usr/share/ovirt-engine/dbscripts/upgrade/03_06__set_version.sql, its
version 0306 is <= current version
04000860
  Skipping upgrade script
/usr/share/ovirt-engine/dbscripts/upgrade/03_06_0010_update_haswell_vds_to_new_name.sql,
its version 03060010 is <= current version 04000860​


​If 3.6 was properly upgraded to 4.0, you wouldn't see any records like
this, you'd see applying only 4.1 upgrade scripts, for example:

  ​Dropping materialized views...
  Running upgrade sql script
'/usr/share/ovirt-engine/share/ovirt-engine/dbscripts/upgrade/04_01_0010_add_mac_pool_id_to_vds_group.sql'...
  Running upgrade sql script
'/usr/share/ovirt-engine/share/ovirt-engine/dbscripts/upgrade/04_01_0020_empty_current_cd_to_null.sql'...




> Personaly I'd recommend you to get back into 3.6 (either by restoring
>> from backup (if you performed in-place upgrade) or shutting down new
>> host and starte pervious 3.6 host (if you performed upgrade using
>> engine-backup) and do upgrade to 4.0 again​. And if upgrade from 3.6 to
>> 4.0 is not successful, please share the logs again.
>>
>
> I'll try from scratch again, uninstalling everything, installing 3.6,
> restoring backups and upgrading to 4.0
>
> I don't want to be a party popper but to be honest you need a testing
> engineer. The upgrade process is way too fragile and breaks too easily...
>
> --
>
> Marcin M. Jessa
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Simone Tiraboschi
On Fri, Sep 16, 2016 at 12:50 PM, Martin Perina  wrote:

>
>
> On Fri, Sep 16, 2016 at 9:26 AM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
>
>>
>> > On 16 Sep 2016, at 08:29, aleksey.maksi...@it-kb.ru wrote:
>> >
>> > There are more ideas?
>> >
>> > 15.09.2016, 14:40, "aleksey.maksi...@it-kb.ru" <
>> aleksey.maksi...@it-kb.ru>:
>> >> Martin, I physically turned off the server through the iLO2. See
>> screenshots.
>> >> I did not touch Virtual Machine (KOM-AD01-PBX02) at the same time.
>> >> The virtual machine has been turned on at the time when the host shut
>> down.
>> >>
>> >> 15.09.2016, 14:27, "Martin Perina" :
>> >>>  Hi,
>> >>>
>> >>>  I found out this in the log:
>> >>>
>> >>>  2016-09-15 12:02:04,661 INFO  
>> >>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>> (ForkJoinPool-1-worker-6) [] VM 
>> '660bafca-e9c3-4191-99b4-295ff8553488'(KOM-AD01-PBX02)
>> moved from 'Up' --> 'Down'
>> >>>  2016-09-15 12:02:04,788 INFO  [org.ovirt.engine.core.dal.dbb
>> roker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-6) []
>> Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: VM
>> KOM-AD01-PBX02 is down. Exit message: User shut down from within the guest
>>
>> since it shut down cleanly, can you please check the guest's logs to see
>> what triggered the shutdown? In such cases it is considered a user
>> requested shutdown and such VMs are not restarted automatically
>>
>
> ​That's exactly what I meant by my response. From the log it's obvious
> that VM was shutdown properly, so engine will not restart it on a
> different. host. Also on most modern hosts if you execute power management
> off action, a signal is sent to OS to execute ​
>
> ​regular shutdown so VMs are also shutted down properly.
>

I understand the reason, but is it really what the user expects?

I mean, if I set HA mode on a VM I'd expect the that the engine cares to
keep it up of restart if needed regardless of shutdown reasons.
For instance, on hosted-engine the HA agent, if not in global maintenance
mode, will restart the engine VM regardless of who or why it went off.



> ​
>
>> We are aware of a similar issue on specific hw -
>> https://bugzilla.redhat.com/show_bug.cgi?id=1341106
>>
>> >>>
>> >>>  If I'm not mistaken, this means that VM was properly shutted down
>> from within itself and in that case it's not restarted automatically. So
>> I'm curious what actions have you made to make host KOM-AD01-VM31
>> non-responsive?
>> >>>
>> >>>  If you want to test fencing properly, then I suggest you to either
>> block connection between host and engine on host side and forcibly stop
>> ovirtmgmt network interface on host and watch fencing is applied.
>>
>
> ​Try above if you want to test fencing. Of course you can always configure
> firewall rule to drop all packets between engine and host or unplug host
> network cable​.
>
> >>>
>> >>>  Martin
>> >>>
>> >>>  On Thu, Sep 15, 2016 at 1:16 PM,  wrote:
>>   engine.log for this period.
>> 
>>   15.09.2016, 14:01, "Martin Perina" :
>> >  On Thu, Sep 15, 2016 at 12:47 PM, 
>> wrote:
>> >>  Hi Martin.
>> >>  I have a stupid question. Use Watchdog device mandatory to
>> automatically start a virtual machine in host Fencing process?
>> >
>> >  ​AFAIK it's not, but I'm not na expert, adding Arik.
>> >
>> >  You need correct power management setup for the hosts and VM has
>> to be marked as highly available​ for sure.
>> >
>> >>  15.09.2016, 13:43, "Martin Perina" :
>> >>>  Hi,
>> >>>
>> >>>  could you please share whole engine.log?
>> >>>
>> >>>  Thanks
>> >>>
>> >>>  Martin Perina
>> >>>
>> >>>  On Thu, Sep 15, 2016 at 12:01 PM, 
>> wrote:
>>   Hello oVirt guru`s !
>> 
>>   I have oVirt Hosted Engine 4.0.3-1.el7.centos on two CentOS 7.2
>> hosts (HP ProLiant DL 360 G5) connected to shared FC SAN Storage.
>> 
>>   1. I configured Power Management for the Hosts (successfully
>> added Fencing Agent for iLO2 from my hosts)
>> 
>>   2. I created new VM (KOM-AD01-PBX02) and installed Guest OS
>> (Ubuntu Server 16.04 LTS) and oVirt Guest Agent
>>   (As described herein https://blog.it-kb.ru/2016/09/
>> 14/install-ovirt-4-0-part-2-about-data-center-iso-domain-log
>> ical-network-vlan-vm-settings-console-guest-agent-live-migration/)
>>  In VM settings on "High Availability" I turned on the option
>> "Highly Available" and change "Priority" to "High"
>> 
>>   3. Now I'm trying to check Hard-Fencing and power off my first
>> host (KOM-AD01-VM31) from his iLO (KOM-AD01-ILO31).
>> 
>>   Fencing successfully works and server is automatically turned
>> on, but my HA VM not started on second host (KOM-AD01-VM32).
>> 
>>   These events I see in the oVirt web console:
>> 
>>   Sep 15, 2016 12:08:13 PMHost KOM-AD01-VM31 power
>> management w

[ovirt-users] ovirt 3.6: hosted engine migration to glusterfs

2016-09-16 Thread Fedele Stabile
Hello,
I have just upgraded my ovirt cluster from 3.5 to 3.6 and now I have
HostedEngine on the LocalStorage of the node.
I have executed the command hosted-engine --deploy on other nodes, 
actually I have the HE image on a nfs filesystem mounted on 5 nodes 
but I would migrate the HE image to glusterfs.
What are the steps to begin this operation?
Is it possible just move the disk on the glusterfs?
.. alternatively, I have to install HE again but on glusterfs,
restoring the db as if I have to restore from a corrupted storage.
In that case I suppose that a VM named HostedEngine will appear in a
down state and is not possible to delete it by WebUI because is not
handled by the new engine: the only possibility is to delete it from
psql
working on engine-db.

Can you suggest me the easiest way?

Fedele Stabile

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Martin Perina
On Fri, Sep 16, 2016 at 1:54 PM, Simone Tiraboschi 
wrote:

>
>
> On Fri, Sep 16, 2016 at 12:50 PM, Martin Perina 
> wrote:
>
>>
>>
>> On Fri, Sep 16, 2016 at 9:26 AM, Michal Skrivanek <
>> michal.skriva...@redhat.com> wrote:
>>
>>>
>>> > On 16 Sep 2016, at 08:29, aleksey.maksi...@it-kb.ru wrote:
>>> >
>>> > There are more ideas?
>>> >
>>> > 15.09.2016, 14:40, "aleksey.maksi...@it-kb.ru" <
>>> aleksey.maksi...@it-kb.ru>:
>>> >> Martin, I physically turned off the server through the iLO2. See
>>> screenshots.
>>> >> I did not touch Virtual Machine (KOM-AD01-PBX02) at the same time.
>>> >> The virtual machine has been turned on at the time when the host shut
>>> down.
>>> >>
>>> >> 15.09.2016, 14:27, "Martin Perina" :
>>> >>>  Hi,
>>> >>>
>>> >>>  I found out this in the log:
>>> >>>
>>> >>>  2016-09-15 12:02:04,661 INFO  
>>> >>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>>> (ForkJoinPool-1-worker-6) [] VM 
>>> '660bafca-e9c3-4191-99b4-295ff8553488'(KOM-AD01-PBX02)
>>> moved from 'Up' --> 'Down'
>>> >>>  2016-09-15 12:02:04,788 INFO  [org.ovirt.engine.core.dal.dbb
>>> roker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-6) []
>>> Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: VM
>>> KOM-AD01-PBX02 is down. Exit message: User shut down from within the guest
>>>
>>> since it shut down cleanly, can you please check the guest's logs to see
>>> what triggered the shutdown? In such cases it is considered a user
>>> requested shutdown and such VMs are not restarted automatically
>>>
>>
>> ​That's exactly what I meant by my response. From the log it's obvious
>> that VM was shutdown properly, so engine will not restart it on a
>> different. host. Also on most modern hosts if you execute power management
>> off action, a signal is sent to OS to execute ​
>>
>> ​regular shutdown so VMs are also shutted down properly.
>>
>
> I understand the reason, but is it really what the user expects?
>
> I mean, if I set HA mode on a VM I'd expect the that the engine cares to
> keep it up of restart if needed regardless of shutdown reasons.
>

​AFAIK that's correct, we need to be able ​
​shutdown HA VM​
​
​ without being it immediately restarted on different host. We want to
restart HA VM only if host, where HA VM is running, is non-responsive.

For instance, on hosted-engine the HA agent, if not in global maintenance
> mode, will restart the engine VM regardless of who or why it went off.
>

​Well, HE VM is definitely not a standard HA VM :-)
​


>
>
>
>> ​
>>
>>> We are aware of a similar issue on specific hw -
>>> https://bugzilla.redhat.com/show_bug.cgi?id=1341106
>>>
>>> >>>
>>> >>>  If I'm not mistaken, this means that VM was properly shutted down
>>> from within itself and in that case it's not restarted automatically. So
>>> I'm curious what actions have you made to make host KOM-AD01-VM31
>>> non-responsive?
>>> >>>
>>> >>>  If you want to test fencing properly, then I suggest you to either
>>> block connection between host and engine on host side and forcibly stop
>>> ovirtmgmt network interface on host and watch fencing is applied.
>>>
>>
>> ​Try above if you want to test fencing. Of course you can always
>> configure firewall rule to drop all packets between engine and host or
>> unplug host network cable​.
>>
>> >>>
>>> >>>  Martin
>>> >>>
>>> >>>  On Thu, Sep 15, 2016 at 1:16 PM,  wrote:
>>>   engine.log for this period.
>>> 
>>>   15.09.2016, 14:01, "Martin Perina" :
>>> >  On Thu, Sep 15, 2016 at 12:47 PM, 
>>> wrote:
>>> >>  Hi Martin.
>>> >>  I have a stupid question. Use Watchdog device mandatory to
>>> automatically start a virtual machine in host Fencing process?
>>> >
>>> >  ​AFAIK it's not, but I'm not na expert, adding Arik.
>>> >
>>> >  You need correct power management setup for the hosts and VM has
>>> to be marked as highly available​ for sure.
>>> >
>>> >>  15.09.2016, 13:43, "Martin Perina" :
>>> >>>  Hi,
>>> >>>
>>> >>>  could you please share whole engine.log?
>>> >>>
>>> >>>  Thanks
>>> >>>
>>> >>>  Martin Perina
>>> >>>
>>> >>>  On Thu, Sep 15, 2016 at 12:01 PM, 
>>> wrote:
>>>   Hello oVirt guru`s !
>>> 
>>>   I have oVirt Hosted Engine 4.0.3-1.el7.centos on two CentOS
>>> 7.2 hosts (HP ProLiant DL 360 G5) connected to shared FC SAN Storage.
>>> 
>>>   1. I configured Power Management for the Hosts (successfully
>>> added Fencing Agent for iLO2 from my hosts)
>>> 
>>>   2. I created new VM (KOM-AD01-PBX02) and installed Guest OS
>>> (Ubuntu Server 16.04 LTS) and oVirt Guest Agent
>>>   (As described herein https://blog.it-kb.ru/2016/09/
>>> 14/install-ovirt-4-0-part-2-about-data-center-iso-domain-log
>>> ical-network-vlan-vm-settings-console-guest-agent-live-migration/)
>>>  In VM settings on "High Availability" I turned on the
>>> option "Highly Available" and change "Priority" to "High"
>>

Re: [ovirt-users] live migration with openvswitch

2016-09-16 Thread Michal Skrivanek

> On 15 Sep 2016, at 21:46, Edward Haas  wrote:
> 
> 
> 
> On Thu, Sep 15, 2016 at 1:30 PM, Michal Skrivanek 
> mailto:michal.skriva...@redhat.com>> wrote:
> 
> > On 15 Sep 2016, at 10:11, Dan Kenigsberg  > > wrote:
> >
> > On Wed, Sep 14, 2016 at 03:04:14PM +0200, Michal Skrivanek wrote:
> >>
> >>> On 09 Sep 2016, at 13:09, Edward Haas  >>> > wrote:
> >>>
> >>>
> >>>
> >>> On Thu, Sep 8, 2016 at 11:27 AM, Pavel Levshin  >>>   >>> >> wrote:
> >>> Hi.
> >>>
> >>> I'm trying to learn Ovirt 4 and have a problem with it.
> >>>
> >>> My cluster consists of 3 nodes. I use Openvswitch for network 
> >>> connectivity. I have a HostedEngine and one additional VM in the cluster.
> >>>
> >>> When I try to migrate the VM to another node, it fails. From vdsm and 
> >>> libvirtd logs I see that proper network interface on destination node 
> >>> cannot be found. Libvirt tries to find Openvswitch bridge with name like 
> >>> "vdsmbr_AOYiPtcT". It exists on source node, but it is unique on every 
> >>> node, because it contains random part. Additionally, it changes on every 
> >>> reboot.
> >>>
> >>> How this is supposed to work?
> >>>
> >>> --
> >>> Pavel Levshin
> >>>
> >>>
> >>>
> >>> Hi Pavel,
> >>>
> >>> VM migration is supported on the master branch, however it has not been 
> >>> ported to 4.0 yet.
> >>
> >>> You can either build VDSM from source (from master branch) or try to 
> >>> apply this patch on what you have:
> >>> https://gerrit.ovirt.org/#/c/59645  
> >>> >
> >>
> >> That’s quite a horrible solution right now. I certainly would not like to 
> >> see it in 4.0 (given the hacks around display).
> 
> What is horrible exactly?
> It's not too late to propose other solutions.

if OVS is the next great feature it should fit into the code accordingly. I.e. 
using hooks only when it’s absolutely necessary and as a temporary measure only 
until the respective proper RFEs are implemented and available. E.g. when there 
is a libvirt support missing we can add a qemu command line parameter ourselves 
bypassing libvirt but we always should have a clear plan (i.e. a bug) to move 
away from there as soon as the support is there(requested back then when we 
went with the hack)

Such things should be reviewed as soon as we get to a similar area, so while 
modifying libvirt-hook.sh we can see the original reason for the hook is not 
valid anymore as everything is addressed and the hacky code should have been 
removed
It was easy to see that because there is a clear comment about dependent bugs 
and issues (though missed by all the reviewers, unfortunately!)
Your new code doesn’t have anything like that and I have no idea what kind of 
API or behavior we actually need, whether appropriate requests has been filed 
on e.g. libvirt. That makes it very hard to revisit in the future by the next 
random person.

> 
> Display uses libvirt to resolve a network name to an IP address for it to 
> bound to. But that works only for linux bridges.
> That is limiting, especially now that we do not have a Linux bridge, but 
> something else.

that’s ok, whatever needs to be done. But then please make sure you’re not 
breaking existing features, at least again not without a plan(==bug) to fix it.

> 
> >> Do we have a bug/plan to improve it?
> >
> > We have Bug 1362495 - [OVS] - Add support for live migration
> > to track that.
> >
> > I'm afraid that we are not yet ready to backport it to 4.0 - we found
> > out that as it is, it break migration for vmfex and external network
> > providers; it also breaks when a buggy Engine db does not send a
> > displayNetwork. But we plan to fix these issues quite soon.

which “buggy” engine? There were changes in parameters, most of these issues 
are not relevant anymore since we ditched <3.6 though.
Again it’s ok as long as it is clearly mentioned like "3.6 engine sends it in 
such and such parameter, we can drop it once we support 4.0+"

> >
> > The hacks arround display are an actual imporovement. For "legacy"
> > switchType, we maintain an on-host libvirt-side database of all networks
> > only to keep libvirt happy. Having a database copy has all the known
> > troubles of mismatches and being out of sync. For "ovs" switchType, we
> > do not (we don't use a bridge, but a port group so there's no natural
> > way to define our network in libvirt). Modifying the listening address
> > on destination is the flexible and quick way to do it - I wish we had
> > the libvirt migrate hook years ago.
> 
> doesn’t it prevent seamless virti-viewer console connection?
> 
> The end result is the same, we listen on the address of a specific network.
> Previously it contained a network name and libvirt converted it to the 
> correct IP it should bind to, now vdsm resolves it.

so did we 

Re: [ovirt-users] live migration with openvswitch

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 14:36, Michal Skrivanek  
> wrote:
> 
> 
>> On 15 Sep 2016, at 21:46, Edward Haas > > wrote:
>> 
>> 
>> 
>> On Thu, Sep 15, 2016 at 1:30 PM, Michal Skrivanek 
>> mailto:michal.skriva...@redhat.com>> wrote:
>> 
>> > On 15 Sep 2016, at 10:11, Dan Kenigsberg > > > wrote:
>> >
>> > On Wed, Sep 14, 2016 at 03:04:14PM +0200, Michal Skrivanek wrote:
>> >>
>> >>> On 09 Sep 2016, at 13:09, Edward Haas > >>> > wrote:
>> >>>
>> >>>
>> >>>
>> >>> On Thu, Sep 8, 2016 at 11:27 AM, Pavel Levshin > >>>  > >>> >> wrote:
>> >>> Hi.
>> >>>
>> >>> I'm trying to learn Ovirt 4 and have a problem with it.
>> >>>
>> >>> My cluster consists of 3 nodes. I use Openvswitch for network 
>> >>> connectivity. I have a HostedEngine and one additional VM in the cluster.
>> >>>
>> >>> When I try to migrate the VM to another node, it fails. From vdsm and 
>> >>> libvirtd logs I see that proper network interface on destination node 
>> >>> cannot be found. Libvirt tries to find Openvswitch bridge with name like 
>> >>> "vdsmbr_AOYiPtcT". It exists on source node, but it is unique on every 
>> >>> node, because it contains random part. Additionally, it changes on every 
>> >>> reboot.
>> >>>
>> >>> How this is supposed to work?
>> >>>
>> >>> --
>> >>> Pavel Levshin
>> >>>
>> >>>
>> >>>
>> >>> Hi Pavel,
>> >>>
>> >>> VM migration is supported on the master branch, however it has not been 
>> >>> ported to 4.0 yet.
>> >>
>> >>> You can either build VDSM from source (from master branch) or try to 
>> >>> apply this patch on what you have:
>> >>> https://gerrit.ovirt.org/#/c/59645  
>> >>> >
>> >>
>> >> That’s quite a horrible solution right now. I certainly would not like to 
>> >> see it in 4.0 (given the hacks around display).
>> 
>> What is horrible exactly?
>> It's not too late to propose other solutions.
> 
> if OVS is the next great feature it should fit into the code accordingly. 
> I.e. using hooks only when it’s absolutely necessary and as a temporary 
> measure only until the respective proper RFEs are implemented and available. 
> E.g. when there is a libvirt support missing we can add a qemu command line 
> parameter ourselves bypassing libvirt but we always should have a clear plan 
> (i.e. a bug) to move away from there as soon as the support is 
> there(requested back then when we went with the hack)
> 
> Such things should be reviewed as soon as we get to a similar area, so while 
> modifying libvirt-hook.sh we can see the original reason for the hook is not 
> valid anymore as everything is addressed and the hacky code should have been 
> removed
> It was easy to see that because there is a clear comment about dependent bugs 
> and issues (though missed by all the reviewers, unfortunately!)
> Your new code doesn’t have anything like that and I have no idea what kind of 
> API or behavior we actually need, whether appropriate requests has been filed 
> on e.g. libvirt. That makes it very hard to revisit in the future by the next 
> random person.
> 
>> 
>> Display uses libvirt to resolve a network name to an IP address for it to 
>> bound to. But that works only for linux bridges.
>> That is limiting, especially now that we do not have a Linux bridge, but 
>> something else.
> 
> that’s ok, whatever needs to be done. But then please make sure you’re not 
> breaking existing features, at least again not without a plan(==bug) to fix 
> it.
> 
>> 
>> >> Do we have a bug/plan to improve it?
>> >
>> > We have Bug 1362495 - [OVS] - Add support for live migration
>> > to track that.

oh, and yes, that’s exactly the tracking I wanted to make sure exists. There’s 
just no link in the gerrit commit itself so I didn’t find it (but I wasn’t 
really looking hard either;-)

Thanks,
michal

>> >
>> > I'm afraid that we are not yet ready to backport it to 4.0 - we found
>> > out that as it is, it break migration for vmfex and external network
>> > providers; it also breaks when a buggy Engine db does not send a
>> > displayNetwork. But we plan to fix these issues quite soon.
> 
> which “buggy” engine? There were changes in parameters, most of these issues 
> are not relevant anymore since we ditched <3.6 though.
> Again it’s ok as long as it is clearly mentioned like "3.6 engine sends it in 
> such and such parameter, we can drop it once we support 4.0+"
> 
>> >
>> > The hacks arround display are an actual imporovement. For "legacy"
>> > switchType, we maintain an on-host libvirt-side database of all networks
>> > only to keep libvirt happy. Having a database copy has all the known
>> > troubles of mismatches and being out of sync. For "ovs" switchType, we
>> > do not (we don't use a bridge, but a port group so there's no natural
>> > way to define our network in libvirt

Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 14:23, Martin Perina  wrote:
> 
> 
> 
> On Fri, Sep 16, 2016 at 1:54 PM, Simone Tiraboschi  > wrote:
> 
> 
> On Fri, Sep 16, 2016 at 12:50 PM, Martin Perina  > wrote:
> 
> 
> On Fri, Sep 16, 2016 at 9:26 AM, Michal Skrivanek 
> mailto:michal.skriva...@redhat.com>> wrote:
> 
> > On 16 Sep 2016, at 08:29, aleksey.maksi...@it-kb.ru 
> >  wrote:
> >
> > There are more ideas?
> >
> > 15.09.2016, 14:40, "aleksey.maksi...@it-kb.ru 
> > "  > >:
> >> Martin, I physically turned off the server through the iLO2. See 
> >> screenshots.
> >> I did not touch Virtual Machine (KOM-AD01-PBX02) at the same time.
> >> The virtual machine has been turned on at the time when the host shut down.
> >>
> >> 15.09.2016, 14:27, "Martin Perina"  >> >:
> >>>  Hi,
> >>>
> >>>  I found out this in the log:
> >>>
> >>>  2016-09-15 12:02:04,661 INFO  
> >>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
> >>> (ForkJoinPool-1-worker-6) [] VM 
> >>> '660bafca-e9c3-4191-99b4-295ff8553488'(KOM-AD01-PBX02) moved from 'Up' 
> >>> --> 'Down'
> >>>  2016-09-15 12:02:04,788 INFO  
> >>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
> >>> (ForkJoinPool-1-worker-6) [] Correlation ID: null, Call Stack: null, 
> >>> Custom Event ID: -1, Message: VM KOM-AD01-PBX02 is down. Exit message: 
> >>> User shut down from within the guest
> 
> since it shut down cleanly, can you please check the guest's logs to see what 
> triggered the shutdown? In such cases it is considered a user requested 
> shutdown and such VMs are not restarted automatically
> 
> ​That's exactly what I meant by my response. From the log it's obvious that 
> VM was shutdown properly, so engine will not restart it on a different. host. 
> Also on most modern hosts if you execute power management off action, a 
> signal is sent to OS to execute ​ ​regular shutdown so VMs are also shutted 
> down properly.
> 
> I understand the reason, but is it really what the user expects?
> 
> I mean, if I set HA mode on a VM I'd expect the that the engine cares to keep 
> it up of restart if needed regardless of shutdown reasons.

no, that’s not how HA works today. When you log into a guest and issue 
“shutdown” we do not restart the VM under your hands. We can argue how it 
should or may work, but this is the defined behavior since the dawn of oVirt.

> 
> ​AFAIK that's correct, we need to be able ​​shutdown HA VM​​​ without being 
> it immediately restarted on different host. We want to restart HA VM only if 
> host, where HA VM is running, is non-responsive.

we try to restart it in all other cases other than user initiated shutdown, 
e.g. a QEMU process crash on an otherwise-healthy host

> 
> For instance, on hosted-engine the HA agent, if not in global maintenance 
> mode, will restart the engine VM regardless of who or why it went off.
> 
> ​Well, HE VM is definitely not a standard HA VM :-)
> ​ 
> 
>  
> ​
> We are aware of a similar issue on specific hw - 
> https://bugzilla.redhat.com/show_bug.cgi?id=1341106 
> 
> 
> >>>
> >>>  If I'm not mistaken, this means that VM was properly shutted down from 
> >>> within itself and in that case it's not restarted automatically. So I'm 
> >>> curious what actions have you made to make host KOM-AD01-VM31 
> >>> non-responsive?
> >>>
> >>>  If you want to test fencing properly, then I suggest you to either block 
> >>> connection between host and engine on host side and forcibly stop 
> >>> ovirtmgmt network interface on host and watch fencing is applied.
> 
> ​Try above if you want to test fencing. Of course you can always configure 
> firewall rule to drop all packets between engine and host or unplug host 
> network cable​.
> 
> >>>
> >>>  Martin
> >>>
> >>>  On Thu, Sep 15, 2016 at 1:16 PM,  >>> > wrote:
>   engine.log for this period.
> 
>   15.09.2016, 14:01, "Martin Perina"   >:
> >  On Thu, Sep 15, 2016 at 12:47 PM,  > > wrote:
> >>  Hi Martin.
> >>  I have a stupid question. Use Watchdog device mandatory to 
> >> automatically start a virtual machine in host Fencing process?
> >
> >  ​AFAIK it's not, but I'm not na expert, adding Arik.
> >
> >  You need correct power management setup for the hosts and VM has to be 
> > marked as highly available​ for sure.
> >
> >>  15.09.2016, 13:43, "Martin Perina"  >> >:
> >>>  Hi,
> >>>
> >>>  could you please share whole engine.log?
> >>>
> >>>  Thanks
> >>>
> >>>  Martin Perina
> >>>
> >>>  On Thu, Sep 15, 2016 at 12:01 PM,  >>> > wrote:
>   Hello oVi

[ovirt-users] Changing Hosted Engine destEmail

2016-09-16 Thread Simon Hallam
Hi,

I'm curious how we can change the email addresses that the hosted-engine-ha 
system sends it's notifications to (this is on oVirt 3.6)?

Cheers,

Simon Hallam
Network Development & Linux Support Officer



Please visit our new website at www.pml.ac.uk and follow us on Twitter 
@PlymouthMarine

Winner of the Environment & Conservation category, the Charity Awards 2014.

Plymouth Marine Laboratory (PML) is a company limited by guarantee registered 
in England & Wales, company number 4178503. Registered Charity No. 1091222. 
Registered Office: Prospect Place, The Hoe, Plymouth  PL1 3DH, UK. 

This message is private and confidential. If you have received this message in 
error, please notify the sender and remove it from your system. You are 
reminded that e-mail communications are not secure and may contain viruses; PML 
accepts no liability for any loss or damage which may be caused by viruses.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Gianluca Cecchi
On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
> no, that’s not how HA works today. When you log into a guest and issue
> “shutdown” we do not restart the VM under your hands. We can argue how it
> should or may work, but this is the defined behavior since the dawn of
> oVirt.
>
>
> ​AFAIK that's correct, we need to be able ​
> ​shutdown HA VM​
> ​
> ​ without being it immediately restarted on different host. We want to
> restart HA VM only if host, where HA VM is running, is non-responsive.
>
>
> we try to restart it in all other cases other than user initiated
> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
>
>
Hi, just another question in case HA is not configured at all.
If I run the "shutdown -h now" command on an host where some VMs are
running, what is the expected behavior?
Clean VM shutdown (with or without timeout in case it doesn't complete?) or
crash of their related QEMU processes?

Thanks,
Gianluca
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 15:05, Gianluca Cecchi  wrote:
> 
> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek 
> mailto:michal.skriva...@redhat.com>> wrote:
> 
> no, that’s not how HA works today. When you log into a guest and issue 
> “shutdown” we do not restart the VM under your hands. We can argue how it 
> should or may work, but this is the defined behavior since the dawn of oVirt.
> 
>> 
>> ​AFAIK that's correct, we need to be able ​​shutdown HA VM​​​ without being 
>> it immediately restarted on different host. We want to restart HA VM only if 
>> host, where HA VM is running, is non-responsive.
> 
> we try to restart it in all other cases other than user initiated shutdown, 
> e.g. a QEMU process crash on an otherwise-healthy host
> 
> 
> Hi, just another question in case HA is not configured at all.

by “HA configured” I expect you’re referring to the “Highly Available” checkbox 
in Edit VM dialog.

> If I run the "shutdown -h now" command on an host where some VMs are running, 
> what is the expected behavior?
> Clean VM shutdown (with or without timeout in case it doesn't complete?) or 
> crash of their related QEMU processes?

expectation is that you won’t do that. That’s why there is the Maintenance host 
state.
But if you do that regardless, with VMs running, all the processes will be 
terminated in a regular system way, i.e. all QEMU processes get SIGTERM. From 
the perspective of each guest this is not a clean shutdown and it would just 
get killed 

Thanks,
michal
> 
> Thanks,
> Gianluca
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Gianluca Cecchi
On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
> On 16 Sep 2016, at 15:05, Gianluca Cecchi 
> wrote:
>
> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
>
>>
>> no, that’s not how HA works today. When you log into a guest and issue
>> “shutdown” we do not restart the VM under your hands. We can argue how it
>> should or may work, but this is the defined behavior since the dawn of
>> oVirt.
>>
>>
>> ​AFAIK that's correct, we need to be able ​
>> ​shutdown HA VM​
>> ​
>> ​ without being it immediately restarted on different host. We want to
>> restart HA VM only if host, where HA VM is running, is non-responsive.
>>
>>
>> we try to restart it in all other cases other than user initiated
>> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
>>
>>
> Hi, just another question in case HA is not configured at all.
>
>
> by “HA configured” I expect you’re referring to the “Highly Available”
> checkbox in Edit VM dialog.
>

Yes


>
> If I run the "shutdown -h now" command on an host where some VMs are
> running, what is the expected behavior?
> Clean VM shutdown (with or without timeout in case it doesn't complete?)
> or crash of their related QEMU processes?
>
>
> expectation is that you won’t do that. That’s why there is the Maintenance
> host state.
> But if you do that regardless, with VMs running, all the processes will be
> terminated in a regular system way, i.e. all QEMU processes get SIGTERM.
> From the perspective of each guest this is not a clean shutdown and it
> would just get killed
>
>
Yes, I was thinking about the scenario of one guy issuing the command (or
pressing the button) by mistake.
Thanks,
Gianluca
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Simone Tiraboschi
On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
> On 16 Sep 2016, at 15:05, Gianluca Cecchi 
> wrote:
>
> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
>
>>
>> no, that’s not how HA works today. When you log into a guest and issue
>> “shutdown” we do not restart the VM under your hands. We can argue how it
>> should or may work, but this is the defined behavior since the dawn of
>> oVirt.
>>
>>
>> ​AFAIK that's correct, we need to be able ​
>> ​shutdown HA VM​
>> ​
>> ​ without being it immediately restarted on different host. We want to
>> restart HA VM only if host, where HA VM is running, is non-responsive.
>>
>>
>> we try to restart it in all other cases other than user initiated
>> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
>>
>>
> Hi, just another question in case HA is not configured at all.
>
>
> by “HA configured” I expect you’re referring to the “Highly Available”
> checkbox in Edit VM dialog.
>
> If I run the "shutdown -h now" command on an host where some VMs are
> running, what is the expected behavior?
> Clean VM shutdown (with or without timeout in case it doesn't complete?)
> or crash of their related QEMU processes?
>
>
> expectation is that you won’t do that. That’s why there is the Maintenance
> host state.
> But if you do that regardless, with VMs running, all the processes will be
> terminated in a regular system way, i.e. all QEMU processes get SIGTERM.
> From the perspective of each guest this is not a clean shutdown and it
> would just get killed
>
>
Aleksey is reporting that he started a shutdown on his host by power
management and the VM processes didn't get roughly killed but smoothly shut
down and so they didn't restarted regardless of their HA flag and so this
thread.


> Thanks,
> michal
>
>
> Thanks,
> Gianluca
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread aleksey . maksimov
Hi Simone.Exactly.Now I'll put the journald on the guest and try to understand how the guest off. 16.09.2016, 16:25, "Simone Tiraboschi" :  On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek  wrote: On 16 Sep 2016, at 15:05, Gianluca Cecchi  wrote: On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek  wrote: no, that’s not how HA works today. When you log into a guest and issue “shutdown” we do not restart the VM under your hands. We can argue how it should or may work, but this is the defined behavior since the dawn of oVirt.  ​AFAIK that's correct, we need to be able ​​shutdown HA VM​​​ without being it immediately restarted on different host. We want to restart HA VM only if host, where HA VM is running, is non-responsive. we try to restart it in all other cases other than user initiated shutdown, e.g. a QEMU process crash on an otherwise-healthy host Hi, just another question in case HA is not configured at all. by “HA configured” I expect you’re referring to the “Highly Available” checkbox in Edit VM dialog. If I run the "shutdown -h now" command on an host where some VMs are running, what is the expected behavior?Clean VM shutdown (with or without timeout in case it doesn't complete?) or crash of their related QEMU processes? expectation is that you won’t do that. That’s why there is the Maintenance host state.But if you do that regardless, with VMs running, all the processes will be terminated in a regular system way, i.e. all QEMU processes get SIGTERM. From the perspective of each guest this is not a clean shutdown and it would just get killed   Aleksey is reporting that he started a shutdown on his host by power management and the VM processes didn't get roughly killed but smoothly shut down and so they didn't restarted regardless of their HA flag and so this thread.   Thanks,michal Thanks,Gianluca___Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users___Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users ___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Simone Tiraboschi
On Fri, Sep 16, 2016 at 3:34 PM, Michal Skrivanek <
michal.skriva...@redhat.com> wrote:

>
> On 16 Sep 2016, at 15:31, aleksey.maksi...@it-kb.ru wrote:
>
> Hi Simone.
> Exactly.
> Now I'll put the journald on the guest and try to understand how the guest
> off.
>
>
> great. thanks
>
>
> 16.09.2016, 16:25, "Simone Tiraboschi" :
>
>
>
> On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
>
>
>
> On 16 Sep 2016, at 15:05, Gianluca Cecchi 
> wrote:
>
> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
>
>
> no, that’s not how HA works today. When you log into a guest and issue
> “shutdown” we do not restart the VM under your hands. We can argue how it
> should or may work, but this is the defined behavior since the dawn of
> oVirt.
>
>
>
> ​AFAIK that's correct, we need to be able ​
> ​shutdown HA VM​
> ​
> ​ without being it immediately restarted on different host. We want to
> restart HA VM only if host, where HA VM is running, is non-responsive.
>
>
> we try to restart it in all other cases other than user initiated
> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
>
>
> Hi, just another question in case HA is not configured at all.
>
>
> by “HA configured” I expect you’re referring to the “Highly Available”
> checkbox in Edit VM dialog.
>
>
> If I run the "shutdown -h now" command on an host where some VMs are
> running, what is the expected behavior?
> Clean VM shutdown (with or without timeout in case it doesn't complete?)
> or crash of their related QEMU processes?
>
>
> expectation is that you won’t do that. That’s why there is the Maintenance
> host state.
> But if you do that regardless, with VMs running, all the processes will be
> terminated in a regular system way, i.e. all QEMU processes get SIGTERM.
> From the perspective of each guest this is not a clean shutdown and it
> would just get killed
>
>
>
> Aleksey is reporting that he started a shutdown on his host by power
> management and the VM processes didn't get roughly killed but smoothly shut
> down and so they didn't restarted regardless of their HA flag and so this
> thread.
>
>
> Gianluca talks about “shutdown -h now”, you talk about power management
> action, those are two different things. The current idea is that systemd or
> some other component just propagates the action to the guest and if that
> guest is configured to handle it as a shutdown it starts it itself as well
> so it looks like a user-initiated one. Even though this mostly makes sense
> it is not ok for current HA logic
>
>
Aleksey, can you please also test this scenario?

>
>
>
> Thanks,
> michal
>
>
> Thanks,
> Gianluca
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 15:31, aleksey.maksi...@it-kb.ru wrote:
> 
> Hi Simone.
> Exactly.
> Now I'll put the journald on the guest and try to understand how the guest 
> off.

great. thanks

>  
> 16.09.2016, 16:25, "Simone Tiraboschi" :
>>  
>>  
>> On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek 
>> mailto:michal.skriva...@redhat.com>> wrote:
>>  
>>> On 16 Sep 2016, at 15:05, Gianluca Cecchi >> > wrote:
>>>  
>>> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek 
>>> mailto:michal.skriva...@redhat.com>> wrote:
>>>  
>>> no, that’s not how HA works today. When you log into a guest and issue 
>>> “shutdown” we do not restart the VM under your hands. We can argue how it 
>>> should or may work, but this is the defined behavior since the dawn of 
>>> oVirt.
>>>  
 
  ​AFAIK that's correct, we need to be able ​​shutdown HA VM​​​ without 
 being it immediately restarted on different host. We want to restart HA VM 
 only if host, where HA VM is running, is non-responsive.
>>>  
>>> we try to restart it in all other cases other than user initiated shutdown, 
>>> e.g. a QEMU process crash on an otherwise-healthy host
>>>  
>>> Hi, just another question in case HA is not configured at all.
>>  
>> by “HA configured” I expect you’re referring to the “Highly Available” 
>> checkbox in Edit VM dialog.
>>  
>>> 
>>> If I run the "shutdown -h now" command on an host where some VMs are 
>>> running, what is the expected behavior?
>>> Clean VM shutdown (with or without timeout in case it doesn't complete?) or 
>>> crash of their related QEMU processes?
>>  
>> expectation is that you won’t do that. That’s why there is the Maintenance 
>> host state.
>> But if you do that regardless, with VMs running, all the processes will be 
>> terminated in a regular system way, i.e. all QEMU processes get SIGTERM. 
>> From the perspective of each guest this is not a clean shutdown and it would 
>> just get killed 
>>  
>>  
>> Aleksey is reporting that he started a shutdown on his host by power 
>> management and the VM processes didn't get roughly killed but smoothly shut 
>> down and so they didn't restarted regardless of their HA flag and so this 
>> thread. 

Gianluca talks about “shutdown -h now”, you talk about power management action, 
those are two different things. The current idea is that systemd or some other 
component just propagates the action to the guest and if that guest is 
configured to handle it as a shutdown it starts it itself as well so it looks 
like a user-initiated one. Even though this mostly makes sense it is not ok for 
current HA logic

>>  
>>  
>> Thanks,
>> michal
>>> 
>>>  
>>> Thanks,
>>> Gianluca
>>> ___
>>> Users mailing list
>>> Users@ovirt.org 
>>> http://lists.ovirt.org/mailman/listinfo/users 
>>> 
>> ___
>> Users mailing list
>> Users@ovirt.org 
>> http://lists.ovirt.org/mailman/listinfo/users 
>> 
>>  
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 16:02, aleksey.maksi...@it-kb.ru wrote:
> 
> So, colleagues. 
> I again tested the Fencing and now I think that my host-server power-button 
> (physically or through ILO) sends a KILL-command to the host OS (and as a 
> result to VM)

thanks for confirmation, then it is indeed 
https://bugzilla.redhat.com/show_bug.cgi?id=1341106

I’m not sure if there is any good workaround. You can always 
reconfigure(disable) ACPI in the guest, then HA logic would work ok but it also 
means there is no graceful shutdown and your VM would be killed uncleanly. 

> This journald log in my guest OS when I press the power-button on the host:
> 
> ..
> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping ACPI event daemon...
> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping User Manager for UID 
> 1000...
> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Starting Unattended Upgrades 
> Shutdown...
> Sep 16 16:19:27 KOM-AD01-PBX02 snapd[2583]: 2016/09/16 16:19:27.289063 
> main.go:67: Exiting on terminated signal.
> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2940]: pam_unix(sshd:session): session 
> closed for user user
> Sep 16 16:19:27 KOM-AD01-PBX02 su[3015]: pam_unix(su:session): session closed 
> for user root
> Sep 16 16:19:27 KOM-AD01-PBX02 spice-vdagentd[2638]: vdagentd quiting, 
> returning status 0
> Sep 16 16:19:27 KOM-AD01-PBX02 sudo[3014]: pam_unix(sudo:session): session 
> closed for user root
> Sep 16 16:19:27 KOM-AD01-PBX02 /usr/lib/snapd/snapd[2583]: main.go:67: 
> Exiting on terminated signal.
> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2812]: Received signal 15; terminating.
> ..
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Unmount All 
> Filesystems.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped target Local File Systems 
> (Pre).
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopping Monitoring of LVM2 
> mirrors, snapshots etc. using dmeventd or progress polling...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Remount Root and Kernel 
> File Systems.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Create Static Device Nodes 
> in /dev.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Shutdown.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Final Step.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Starting Reboot...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Monitoring of LVM2 
> mirrors, snapshots etc. using dmeventd or progress polling.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Shutting down.
> Sep 16 16:19:28 KOM-AD01-PBX02 kernel: [drm:qxl_enc_commit [qxl]] *ERROR* 
> head number too large or missing monitors config: c984a000, 
> 0systemd-shutdown[1]: Sending SIGTERM to remaining processes...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd-journald[3342]: Journal stopped
> -- Reboot --
> 
> Perhaps this feature of HP ProLiant DL 360 G5. I dont know.
> 
> If I test the unavailability of a host other ways that everything is going 
> well.
> 
> I described my experience testing Fencing on practical examples on my blog 
> for everyone in Russian.
> https://blog.it-kb.ru/2016/09/16/install-ovirt-4-0-part-4-about-ssh-soft-fencing-and-hard-fencing-over-hp-proliant-ilo2-power-managment-agent-and-test-of-high-availability/
> 
> 
> Thank you all very much for your participation and support.
> 
> Michal, what kind of scenario are you talking about?
> 
> 
> PS: Excuse me for my bad English :)
> 
> 
> 16.09.2016, 16:37, "Simone Tiraboschi" :
>> On Fri, Sep 16, 2016 at 3:34 PM, Michal Skrivanek 
>>  wrote:
 On 16 Sep 2016, at 15:31, aleksey.maksi...@it-kb.ru wrote:
 
 Hi Simone.
 Exactly.
 Now I'll put the journald on the guest and try to understand how the guest 
 off.
>>> 
>>> great. thanks
>>> 
 16.09.2016, 16:25, "Simone Tiraboschi" :
> On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek 
>  wrote:
>>> On 16 Sep 2016, at 15:05, Gianluca Cecchi  
>>> wrote:
>>> 
>>> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek 
>>>  wrote:
 no, that’s not how HA works today. When you log into a guest and issue 
 “shutdown” we do not restart the VM under your hands. We can argue how 
 it should or may work, but this is the defined behavior since the dawn 
 of oVirt.
 
> ​AFAIK that's correct, we need to be able ​
> ​shutdown HA VM​
> ​
> ​ without being it immediately restarted on different host. We want 
> to restart HA VM only if host, where HA VM is running, is 
> non-responsive.
 
 we try to restart it in all other cases other than user initiated 
 shutdown, e.g. a QEMU process crash on an otherwise-healthy host
>>> Hi, just another question in case HA is not configured at all.
>> 
>> by “HA configured” I expect you’re referring to the “Highly Available” 
>> checkbox in Edit VM dialog.
>> 
>>> If I run the "shutdown -h now" command on an host where some VMs are

Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread aleksey . maksimov
So, colleagues. 
I again tested the Fencing and now I think that my host-server power-button 
(physically or through ILO) sends a KILL-command to the host OS (and as a 
result to VM)
This journald log in my guest OS when I press the power-button on the host:

...
Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping ACPI event daemon...
Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping User Manager for UID 1000...
Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Starting Unattended Upgrades 
Shutdown...
Sep 16 16:19:27 KOM-AD01-PBX02 snapd[2583]: 2016/09/16 16:19:27.289063 
main.go:67: Exiting on terminated signal.
Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2940]: pam_unix(sshd:session): session 
closed for user user
Sep 16 16:19:27 KOM-AD01-PBX02 su[3015]: pam_unix(su:session): session closed 
for user root
Sep 16 16:19:27 KOM-AD01-PBX02 spice-vdagentd[2638]: vdagentd quiting, 
returning status 0
Sep 16 16:19:27 KOM-AD01-PBX02 sudo[3014]: pam_unix(sudo:session): session 
closed for user root
Sep 16 16:19:27 KOM-AD01-PBX02 /usr/lib/snapd/snapd[2583]: main.go:67: Exiting 
on terminated signal.
Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2812]: Received signal 15; terminating.
...
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Unmount All 
Filesystems.
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped target Local File Systems 
(Pre).
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopping Monitoring of LVM2 mirrors, 
snapshots etc. using dmeventd or progress polling...
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Remount Root and Kernel File 
Systems.
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Create Static Device Nodes 
in /dev.
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Shutdown.
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Final Step.
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Starting Reboot...
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Monitoring of LVM2 mirrors, 
snapshots etc. using dmeventd or progress polling.
Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Shutting down.
Sep 16 16:19:28 KOM-AD01-PBX02 kernel: [drm:qxl_enc_commit [qxl]] *ERROR* head 
number too large or missing monitors config: c984a000, 
0systemd-shutdown[1]: Sending SIGTERM to remaining processes...
Sep 16 16:19:28 KOM-AD01-PBX02 systemd-journald[3342]: Journal stopped
-- Reboot --

Perhaps this feature of HP ProLiant DL 360 G5. I dont know.

If I test the unavailability of a host other ways that everything is going well.

I described my experience testing Fencing on practical examples on my blog for 
everyone in Russian.
https://blog.it-kb.ru/2016/09/16/install-ovirt-4-0-part-4-about-ssh-soft-fencing-and-hard-fencing-over-hp-proliant-ilo2-power-managment-agent-and-test-of-high-availability/


Thank you all very much for your participation and support.

Michal, what kind of scenario are you talking about?


PS: Excuse me for my bad English :)


16.09.2016, 16:37, "Simone Tiraboschi" :
> On Fri, Sep 16, 2016 at 3:34 PM, Michal Skrivanek 
>  wrote:
>>> On 16 Sep 2016, at 15:31, aleksey.maksi...@it-kb.ru wrote:
>>>
>>> Hi Simone.
>>> Exactly.
>>> Now I'll put the journald on the guest and try to understand how the guest 
>>> off.
>>
>> great. thanks
>>
>>> 16.09.2016, 16:25, "Simone Tiraboschi" :
 On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek 
  wrote:
>> On 16 Sep 2016, at 15:05, Gianluca Cecchi  
>> wrote:
>>
>> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek 
>>  wrote:
>>> no, that’s not how HA works today. When you log into a guest and issue 
>>> “shutdown” we do not restart the VM under your hands. We can argue how 
>>> it should or may work, but this is the defined behavior since the dawn 
>>> of oVirt.
>>>
 ​AFAIK that's correct, we need to be able ​
 ​shutdown HA VM​
 ​
 ​ without being it immediately restarted on different host. We want to 
 restart HA VM only if host, where HA VM is running, is non-responsive.
>>>
>>> we try to restart it in all other cases other than user initiated 
>>> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
>> Hi, just another question in case HA is not configured at all.
>
> by “HA configured” I expect you’re referring to the “Highly Available” 
> checkbox in Edit VM dialog.
>
>> If I run the "shutdown -h now" command on an host where some VMs are 
>> running, what is the expected behavior?
>> Clean VM shutdown (with or without timeout in case it doesn't complete?) 
>> or crash of their related QEMU processes?
>
> expectation is that you won’t do that. That’s why there is the 
> Maintenance host state.
> But if you do that regardless, with VMs running, all the processes will 
> be terminated in a regular system way, i.e. all QEMU processes get 
> SIGTERM. From the perspective of each guest this is not a clean shutdown 
> and it would just get killed

 Aleksey is 

Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Simone Tiraboschi
On Fri, Sep 16, 2016 at 4:02 PM,  wrote:

> So, colleagues.
> I again tested the Fencing and now I think that my host-server
> power-button (physically or through ILO) sends a KILL-command to the host
> OS (and as a result to VM)
> This journald log in my guest OS when I press the power-button on the host:
>
> ...
> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping ACPI event daemon...
> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping User Manager for UID
> 1000...
> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Starting Unattended Upgrades
> Shutdown...
> Sep 16 16:19:27 KOM-AD01-PBX02 snapd[2583]: 2016/09/16 16:19:27.289063
> main.go:67: Exiting on terminated signal.
> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2940]: pam_unix(sshd:session): session
> closed for user user
> Sep 16 16:19:27 KOM-AD01-PBX02 su[3015]: pam_unix(su:session): session
> closed for user root
> Sep 16 16:19:27 KOM-AD01-PBX02 spice-vdagentd[2638]: vdagentd quiting,
> returning status 0
> Sep 16 16:19:27 KOM-AD01-PBX02 sudo[3014]: pam_unix(sudo:session): session
> closed for user root
> Sep 16 16:19:27 KOM-AD01-PBX02 /usr/lib/snapd/snapd[2583]: main.go:67:
> Exiting on terminated signal.
> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2812]: Received signal 15; terminating.
> ...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Unmount All
> Filesystems.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped target Local File
> Systems (Pre).
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopping Monitoring of LVM2
> mirrors, snapshots etc. using dmeventd or progress polling...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Remount Root and Kernel
> File Systems.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Create Static Device
> Nodes in /dev.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Shutdown.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Final Step.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Starting Reboot...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Monitoring of LVM2
> mirrors, snapshots etc. using dmeventd or progress polling.
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Shutting down.
> Sep 16 16:19:28 KOM-AD01-PBX02 kernel: [drm:qxl_enc_commit [qxl]] *ERROR*
> head number too large or missing monitors config: c984a000,
> 0systemd-shutdown[1]: Sending SIGTERM to remaining processes...
> Sep 16 16:19:28 KOM-AD01-PBX02 systemd-journald[3342]: Journal stopped
> -- Reboot --
>
> Perhaps this feature of HP ProLiant DL 360 G5. I dont know.
>
> If I test the unavailability of a host other ways that everything is going
> well.
>
> I described my experience testing Fencing on practical examples on my blog
> for everyone in Russian.
> https://blog.it-kb.ru/2016/09/16/install-ovirt-4-0-part-4-
> about-ssh-soft-fencing-and-hard-fencing-over-hp-proliant-
> ilo2-power-managment-agent-and-test-of-high-availability/
>
>
> Thank you all very much for your participation and support.
>
> Michal, what kind of scenario are you talking about?
>

Basically what you just did,
the question is what happens when you run 'shutdown -h now' (or press the
physical button if configured to trigger a soft shutdown); is it going to
propagate somehow the shutdown action to the VMs or to brutally kill them?

In the first case the VMs will not restart regardless of their HA flags.


>
>
> PS: Excuse me for my bad English :)
>
>
> 16.09.2016, 16:37, "Simone Tiraboschi" :
> > On Fri, Sep 16, 2016 at 3:34 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
> >>> On 16 Sep 2016, at 15:31, aleksey.maksi...@it-kb.ru wrote:
> >>>
> >>> Hi Simone.
> >>> Exactly.
> >>> Now I'll put the journald on the guest and try to understand how the
> guest off.
> >>
> >> great. thanks
> >>
> >>> 16.09.2016, 16:25, "Simone Tiraboschi" :
>  On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
> >> On 16 Sep 2016, at 15:05, Gianluca Cecchi <
> gianluca.cec...@gmail.com> wrote:
> >>
> >> On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek <
> michal.skriva...@redhat.com> wrote:
> >>> no, that’s not how HA works today. When you log into a guest and
> issue “shutdown” we do not restart the VM under your hands. We can argue
> how it should or may work, but this is the defined behavior since the dawn
> of oVirt.
> >>>
>  ​AFAIK that's correct, we need to be able ​
>  ​shutdown HA VM​
>  ​
>  ​ without being it immediately restarted on different host. We
> want to restart HA VM only if host, where HA VM is running, is
> non-responsive.
> >>>
> >>> we try to restart it in all other cases other than user initiated
> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
> >> Hi, just another question in case HA is not configured at all.
> >
> > by “HA configured” I expect you’re referring to the “Highly
> Available” checkbox in Edit VM dialog.
> >
> >> If I run the "shutdown -h now" command on an h

Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread aleksey . maksimov
Тested.

If I run 'shutdown -h now' on host with running HA VM (not HostedEngine VM)...

in oVirt web-console appears event:

Sep 16, 2016 5:13:18 PM VM KOM-AD01-PBX02 is down. Exit message: User shut down 
from within the guest

HA VM is turned off and will not start on another host.

This journald log from HA VM guest OS:

...
Sep 16 17:06:48 KOM-AD01-PBX02 python[2637]: [100B blob data]
Sep 16 17:06:53 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
reply from 91.189.91.157:123 (ntp.ubuntu.com).
Sep 16 17:07:03 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
reply from 91.189.89.199:123 (ntp.ubuntu.com).
Sep 16 17:07:13 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
reply from 91.189.89.198:123 (ntp.ubuntu.com).
Sep 16 17:07:23 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
reply from 91.189.94.4:123 (ntp.ubuntu.com).
Sep 16 17:08:48 KOM-AD01-PBX02 python[2637]: [90B blob data]
Sep 16 17:08:49 KOM-AD01-PBX02 python[2637]: [155B blob data]
Sep 16 17:08:49 KOM-AD01-PBX02 python[2637]: [100B blob data]
Sep 16 17:10:49 KOM-AD01-PBX02 python[2637]: [90B blob data]
Sep 16 17:10:50 KOM-AD01-PBX02 python[2637]: [155B blob data]
Sep 16 17:10:50 KOM-AD01-PBX02 python[2637]: [100B blob data]
-- Reboot --
...

Before shutting down in the log no termination procedures.
It looks like a rough poweroff the VM

16.09.2016, 17:08, "Simone Tiraboschi" :
> On Fri, Sep 16, 2016 at 4:02 PM,  wrote:
>> So, colleagues.
>> I again tested the Fencing and now I think that my host-server power-button 
>> (physically or through ILO) sends a KILL-command to the host OS (and as a 
>> result to VM)
>> This journald log in my guest OS when I press the power-button on the host:
>>
>> ...
>> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping ACPI event daemon...
>> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping User Manager for UID 
>> 1000...
>> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Starting Unattended Upgrades 
>> Shutdown...
>> Sep 16 16:19:27 KOM-AD01-PBX02 snapd[2583]: 2016/09/16 16:19:27.289063 
>> main.go:67: Exiting on terminated signal.
>> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2940]: pam_unix(sshd:session): session 
>> closed for user user
>> Sep 16 16:19:27 KOM-AD01-PBX02 su[3015]: pam_unix(su:session): session 
>> closed for user root
>> Sep 16 16:19:27 KOM-AD01-PBX02 spice-vdagentd[2638]: vdagentd quiting, 
>> returning status 0
>> Sep 16 16:19:27 KOM-AD01-PBX02 sudo[3014]: pam_unix(sudo:session): session 
>> closed for user root
>> Sep 16 16:19:27 KOM-AD01-PBX02 /usr/lib/snapd/snapd[2583]: main.go:67: 
>> Exiting on terminated signal.
>> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2812]: Received signal 15; terminating.
>> ...
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Unmount All 
>> Filesystems.
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped target Local File Systems 
>> (Pre).
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopping Monitoring of LVM2 
>> mirrors, snapshots etc. using dmeventd or progress polling...
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Remount Root and Kernel 
>> File Systems.
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Create Static Device 
>> Nodes in /dev.
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Shutdown.
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Final Step.
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Starting Reboot...
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Monitoring of LVM2 
>> mirrors, snapshots etc. using dmeventd or progress polling.
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Shutting down.
>> Sep 16 16:19:28 KOM-AD01-PBX02 kernel: [drm:qxl_enc_commit [qxl]] *ERROR* 
>> head number too large or missing monitors config: c984a000, 
>> 0systemd-shutdown[1]: Sending SIGTERM to remaining processes...
>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd-journald[3342]: Journal stopped
>> -- Reboot --
>>
>> Perhaps this feature of HP ProLiant DL 360 G5. I dont know.
>>
>> If I test the unavailability of a host other ways that everything is going 
>> well.
>>
>> I described my experience testing Fencing on practical examples on my blog 
>> for everyone in Russian.
>> https://blog.it-kb.ru/2016/09/16/install-ovirt-4-0-part-4-about-ssh-soft-fencing-and-hard-fencing-over-hp-proliant-ilo2-power-managment-agent-and-test-of-high-availability/
>>
>> Thank you all very much for your participation and support.
>>
>> Michal, what kind of scenario are you talking about?
>
> Basically what you just did,
> the question is what happens when you run 'shutdown -h now' (or press the 
> physical button if configured to trigger a soft shutdown); is it going to 
> propagate somehow the shutdown action to the VMs or to brutally kill them?
>
> In the first case the VMs will not restart regardless of their HA flags.
>
>> PS: Excuse me for my bad English :)
>>
>> 16.09.2016, 16:37, "Simone Tiraboschi" :
>>> On Fri, Sep 16, 2016 at 3:34 PM, Michal Skrivanek 
>>>

Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread aleksey . maksimov
"your VM would be killed uncleanly."

This is not a good idea, I think


16.09.2016, 17:14, "Michal Skrivanek" :
>>  On 16 Sep 2016, at 16:02, aleksey.maksi...@it-kb.ru wrote:
>>
>>  So, colleagues.
>>  I again tested the Fencing and now I think that my host-server power-button 
>> (physically or through ILO) sends a KILL-command to the host OS (and as a 
>> result to VM)
>
> thanks for confirmation, then it is indeed 
> https://bugzilla.redhat.com/show_bug.cgi?id=1341106
>
> I’m not sure if there is any good workaround. You can always 
> reconfigure(disable) ACPI in the guest, then HA logic would work ok but it 
> also means there is no graceful shutdown and your VM would be killed 
> uncleanly.
>
>>  This journald log in my guest OS when I press the power-button on the host:
>>
>>  ..
>>  Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping ACPI event daemon...
>>  Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping User Manager for UID 
>> 1000...
>>  Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Starting Unattended Upgrades 
>> Shutdown...
>>  Sep 16 16:19:27 KOM-AD01-PBX02 snapd[2583]: 2016/09/16 16:19:27.289063 
>> main.go:67: Exiting on terminated signal.
>>  Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2940]: pam_unix(sshd:session): session 
>> closed for user user
>>  Sep 16 16:19:27 KOM-AD01-PBX02 su[3015]: pam_unix(su:session): session 
>> closed for user root
>>  Sep 16 16:19:27 KOM-AD01-PBX02 spice-vdagentd[2638]: vdagentd quiting, 
>> returning status 0
>>  Sep 16 16:19:27 KOM-AD01-PBX02 sudo[3014]: pam_unix(sudo:session): session 
>> closed for user root
>>  Sep 16 16:19:27 KOM-AD01-PBX02 /usr/lib/snapd/snapd[2583]: main.go:67: 
>> Exiting on terminated signal.
>>  Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2812]: Received signal 15; terminating.
>>  ..
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Unmount All 
>> Filesystems.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped target Local File 
>> Systems (Pre).
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopping Monitoring of LVM2 
>> mirrors, snapshots etc. using dmeventd or progress polling...
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Remount Root and Kernel 
>> File Systems.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Create Static Device 
>> Nodes in /dev.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Shutdown.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Final Step.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Starting Reboot...
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Monitoring of LVM2 
>> mirrors, snapshots etc. using dmeventd or progress polling.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Shutting down.
>>  Sep 16 16:19:28 KOM-AD01-PBX02 kernel: [drm:qxl_enc_commit [qxl]] *ERROR* 
>> head number too large or missing monitors config: c984a000, 
>> 0systemd-shutdown[1]: Sending SIGTERM to remaining processes...
>>  Sep 16 16:19:28 KOM-AD01-PBX02 systemd-journald[3342]: Journal stopped
>>  -- Reboot --
>>
>>  Perhaps this feature of HP ProLiant DL 360 G5. I dont know.
>>
>>  If I test the unavailability of a host other ways that everything is going 
>> well.
>>
>>  I described my experience testing Fencing on practical examples on my blog 
>> for everyone in Russian.
>>  
>> https://blog.it-kb.ru/2016/09/16/install-ovirt-4-0-part-4-about-ssh-soft-fencing-and-hard-fencing-over-hp-proliant-ilo2-power-managment-agent-and-test-of-high-availability/
>>
>>  Thank you all very much for your participation and support.
>>
>>  Michal, what kind of scenario are you talking about?
>>
>>  PS: Excuse me for my bad English :)
>>
>>  16.09.2016, 16:37, "Simone Tiraboschi" :
>>>  On Fri, Sep 16, 2016 at 3:34 PM, Michal Skrivanek 
>>>  wrote:
>  On 16 Sep 2016, at 15:31, aleksey.maksi...@it-kb.ru wrote:
>
>  Hi Simone.
>  Exactly.
>  Now I'll put the journald on the guest and try to understand how the 
> guest off.

  great. thanks

>  16.09.2016, 16:25, "Simone Tiraboschi" :
>>  On Fri, Sep 16, 2016 at 3:13 PM, Michal Skrivanek 
>>  wrote:
  On 16 Sep 2016, at 15:05, Gianluca Cecchi  
 wrote:

  On Fri, Sep 16, 2016 at 2:50 PM, Michal Skrivanek 
  wrote:
>  no, that’s not how HA works today. When you log into a guest and 
> issue “shutdown” we do not restart the VM under your hands. We can 
> argue how it should or may work, but this is the defined behavior 
> since the dawn of oVirt.
>
>>  ​AFAIK that's correct, we need to be able ​
>>  ​shutdown HA VM​
>>  ​
>>  ​ without being it immediately restarted on different host. We want 
>> to restart HA VM only if host, where HA VM is running, is 
>> non-responsive.
>
>  we try to restart it in all other cases other than user initiated 
> shutdown, e.g. a QEMU process crash on an otherwise-healthy host
  Hi, just 

Re: [ovirt-users] oVirt 4.0.3 (Hosted Engine) - High Availability VM not restart after auto-fencing of host.

2016-09-16 Thread Michal Skrivanek

> On 16 Sep 2016, at 16:34, aleksey.maksi...@it-kb.ru wrote:
> 
> Тested.
> 
> If I run 'shutdown -h now' on host with running HA VM (not HostedEngine VM)...
> 
> in oVirt web-console appears event:
> 
> Sep 16, 2016 5:13:18 PM VM KOM-AD01-PBX02 is down. Exit message: User shut 
> down from within the guest

that would be another bug. It should be recognized properly as a “kill”. Can 
you please share host logs from this attempt as well?

> 
> HA VM is turned off and will not start on another host.
> 
> This journald log from HA VM guest OS:
> 
> ...
> Sep 16 17:06:48 KOM-AD01-PBX02 python[2637]: [100B blob data]
> Sep 16 17:06:53 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
> reply from 91.189.91.157:123 (ntp.ubuntu.com).
> Sep 16 17:07:03 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
> reply from 91.189.89.199:123 (ntp.ubuntu.com).
> Sep 16 17:07:13 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
> reply from 91.189.89.198:123 (ntp.ubuntu.com).
> Sep 16 17:07:23 KOM-AD01-PBX02 systemd-timesyncd[1739]: Timed out waiting for 
> reply from 91.189.94.4:123 (ntp.ubuntu.com).
> Sep 16 17:08:48 KOM-AD01-PBX02 python[2637]: [90B blob data]
> Sep 16 17:08:49 KOM-AD01-PBX02 python[2637]: [155B blob data]
> Sep 16 17:08:49 KOM-AD01-PBX02 python[2637]: [100B blob data]
> Sep 16 17:10:49 KOM-AD01-PBX02 python[2637]: [90B blob data]
> Sep 16 17:10:50 KOM-AD01-PBX02 python[2637]: [155B blob data]
> Sep 16 17:10:50 KOM-AD01-PBX02 python[2637]: [100B blob data]
> -- Reboot --
> ...
> 
> Before shutting down in the log no termination procedures.
> It looks like a rough poweroff the VM

yep, that is expected. But it should be properly detected as such and HE VM 
should restart. Somehow vdsm misidentifies the reason for the shutdown.

> 
> 16.09.2016, 17:08, "Simone Tiraboschi" :
>> On Fri, Sep 16, 2016 at 4:02 PM,  wrote:
>>> So, colleagues.
>>> I again tested the Fencing and now I think that my host-server power-button 
>>> (physically or through ILO) sends a KILL-command to the host OS (and as a 
>>> result to VM)
>>> This journald log in my guest OS when I press the power-button on the host:
>>> 
>>> ...
>>> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping ACPI event daemon...
>>> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Stopping User Manager for UID 
>>> 1000...
>>> Sep 16 16:19:27 KOM-AD01-PBX02 systemd[1]: Starting Unattended Upgrades 
>>> Shutdown...
>>> Sep 16 16:19:27 KOM-AD01-PBX02 snapd[2583]: 2016/09/16 16:19:27.289063 
>>> main.go:67: Exiting on terminated signal.
>>> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2940]: pam_unix(sshd:session): session 
>>> closed for user user
>>> Sep 16 16:19:27 KOM-AD01-PBX02 su[3015]: pam_unix(su:session): session 
>>> closed for user root
>>> Sep 16 16:19:27 KOM-AD01-PBX02 spice-vdagentd[2638]: vdagentd quiting, 
>>> returning status 0
>>> Sep 16 16:19:27 KOM-AD01-PBX02 sudo[3014]: pam_unix(sudo:session): session 
>>> closed for user root
>>> Sep 16 16:19:27 KOM-AD01-PBX02 /usr/lib/snapd/snapd[2583]: main.go:67: 
>>> Exiting on terminated signal.
>>> Sep 16 16:19:27 KOM-AD01-PBX02 sshd[2812]: Received signal 15; terminating.
>>> ...
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Unmount All 
>>> Filesystems.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped target Local File 
>>> Systems (Pre).
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopping Monitoring of LVM2 
>>> mirrors, snapshots etc. using dmeventd or progress polling...
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Remount Root and Kernel 
>>> File Systems.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Create Static Device 
>>> Nodes in /dev.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Shutdown.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Reached target Final Step.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Starting Reboot...
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Stopped Monitoring of LVM2 
>>> mirrors, snapshots etc. using dmeventd or progress polling.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd[1]: Shutting down.
>>> Sep 16 16:19:28 KOM-AD01-PBX02 kernel: [drm:qxl_enc_commit [qxl]] *ERROR* 
>>> head number too large or missing monitors config: c984a000, 
>>> 0systemd-shutdown[1]: Sending SIGTERM to remaining processes...
>>> Sep 16 16:19:28 KOM-AD01-PBX02 systemd-journald[3342]: Journal stopped
>>> -- Reboot --
>>> 
>>> Perhaps this feature of HP ProLiant DL 360 G5. I dont know.
>>> 
>>> If I test the unavailability of a host other ways that everything is going 
>>> well.
>>> 
>>> I described my experience testing Fencing on practical examples on my blog 
>>> for everyone in Russian.
>>> https://blog.it-kb.ru/2016/09/16/install-ovirt-4-0-part-4-about-ssh-soft-fencing-and-hard-fencing-over-hp-proliant-ilo2-power-managment-agent-and-test-of-high-availability/
>>> 
>>> Thank you all very much for your participation and support.
>>> 
>>> Michal, what kind of scenario are you talking about

Re: [ovirt-users] Nested oVirt : Unofficial network recommendations wanted

2016-09-16 Thread Edward Haas
On Fri, Sep 16, 2016 at 1:32 PM, Nicolas Ecarnot 
wrote:

> Le 15/09/2016 à 23:27, Edward Haas a écrit :
>
>
>
> On Thu, Sep 15, 2016 at 3:43 PM, Nicolas Ecarnot 
> wrote:
>
>> Hello,
>>
>> I'm trying to setup a nested oVirt for the first time, and according to
>> what I read around and experience, some special network settings have to be
>> chosen.
>>
>> For this first try, the bare-metal host is a Debian, running KVM, and the
>> virtual NICs are setup as macvtap in VEPA mode.
>>
>
> On what device you attached it? Bridge? the physical nic?
>
>
> On my physical host, ifconfig is showing the following devices :
> - lo, obviously
> - eth0, primary used
> - wlan0, not used, WIFI...
> - virbr0
> - virbr0-nic
>
> The last two devices were created when installing and playing with the KVM
> Virtual Machine Manager.
>
> When trying to assign one of them to a VM, the only choices are :
> - NAT
> - eth0: macvtap
> - wlan0: macvtap
> - custom
>
> and the source mode can be chosen between :
> - Bridge
> - VEPA
> - Private
> - Passthrough
>
> My main goal is simplicity, and I'd rather use simple bridging, no NAT,
> simple.
>
> What would be the simplest choice?
>

With Virtual Machine Manager you can edit-connection details and create
bridges on the host that you can connect to.
I usually use an isolated network, that does not connect to the outside
world and is limited to the host. If I need it to access the outside world,
I can either connect one of the nics to the bridge created on the host or
just create another vnic that uses nat/macvtap.
Note that VMM is mainly focused on simple VM connectivity, using it for
nested virtualization is not simple and will require from you additional
effort.


>
>
>> I'm not up to date with macvtap, but I think I understood that one of its
>> limit was that no packet could be exchanged between the host and the
>> guests. So far, this is leading me to access my own local VMs from another
>> host. Too bad.
>>
>
> I'm also witnessing frequent loss of packets.
>>
>> So far, I'm also seeing that guests can not ping each others, so I'm not
>> going further before having solved these basic issues.
>>
>
> By definition: http://virt.kernelnewbies.org/MacVTap
>
>
>>
>
> I'm remembering the good old times of lots of bridges where my VMs could
>> be reached by anyone (this was desired), but virt manager is not offering
>> me this choice. I also would like to avoid NAT for other reasons.
>>
>> To you all (4) people who are playing with nested oVirt :
>> - which is your preferred bare metal OS?
>> - which is your preferred guest (first virt level) OS?
>> - which network setups and modes are working best?
>>
>> Thank you.
>>
>> --
>> Nicolas ECARNOT
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>
>
> --
> Nicolas ECARNOT
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Support on oVirt

2016-09-16 Thread Johann van der Westhuizen
Hi,

 

Thanks for the reply.

 

I actually did sort out this issue.

 

But I have another issue, when I click on console the .vv file is downloaded, 
but when I open the .vv file it asks to open with a program, what program 
should I use to open the file that I can access the console?

 

Thank you.

 

Regards / Groete,

 

Johann van der Westhuizen

Systems Administrator



  joh...@sharenet.co.za | +27 (021) 700 4827 |   
 www.sharenet.co.za

 

    
 

 

From: Yaniv Dary [mailto:yd...@redhat.com] 
Sent: 15 September 2016 23:07
To: Johann van der Westhuizen 
Cc: users 
Subject: Re: [ovirt-users] Support on oVirt

 

Can you share logs? Sos report? 

Yaniv Dary
Technical Product Manager
Red Hat Israel Ltd.
34 Jerusalem Road
Building A, 4th floor
Ra'anana, Israel 4350109

Tel : +972 (9) 7692306
8272306
Email: yd...@redhat.com  
IRC : ydary

 

On Sep 15, 2016 22:28, "Johann van der Westhuizen" mailto:joh...@sharenet.co.za> > wrote:

Hi,

 

Wonder if you could help me with oVirt, having some issues starting a new VM, 
please see below:

 



 

Getting this error just as I try to start the VM, the green play button. What 
can I do to resolve this issue?

 

Regards / Groete,

 

Johann van der Westhuizen

Systems Administrator



joh...@sharenet.co.za   | +27 
  (021) 700 4827 |   
 www.sharenet.co.za

 

   
 

 


___
Users mailing list
Users@ovirt.org  
http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Support on oVirt

2016-09-16 Thread Johann van der Westhuizen
Hi,

 

And for VNC?

 

Regards / Groete,

 

Johann van der Westhuizen

Systems Administrator



  joh...@sharenet.co.za | +27 (021) 700 4827 |   
 www.sharenet.co.za

 

    
 

 

From: Yaniv Dary [mailto:yd...@redhat.com] 
Sent: 16 September 2016 10:11
To: Johann van der Westhuizen 
Cc: users 
Subject: RE: [ovirt-users] Support on oVirt

 

Remote viewer for spice. 

Yaniv Dary
Technical Product Manager
Red Hat Israel Ltd.
34 Jerusalem Road
Building A, 4th floor
Ra'anana, Israel 4350109

Tel : +972 (9) 7692306
8272306
Email: yd...@redhat.com  
IRC : ydary

 

On Sep 16, 2016 10:00, "Johann van der Westhuizen" mailto:joh...@sharenet.co.za> > wrote:

Hi,

 

Thanks for the reply.

 

I actually did sort out this issue.

 

But I have another issue, when I click on console the .vv file is downloaded, 
but when I open the .vv file it asks to open with a program, what program 
should I use to open the file that I can access the console?

 

Thank you.

 

Regards / Groete,

 

Johann van der Westhuizen

Systems Administrator



  joh...@sharenet.co.za | +27 (021) 700 4827 |   
 www.sharenet.co.za

 

   
 

 

From: Yaniv Dary [mailto:yd...@redhat.com  ] 
Sent: 15 September 2016 23:07
To: Johann van der Westhuizen mailto:joh...@sharenet.co.za> >
Cc: users mailto:users@ovirt.org> >
Subject: Re: [ovirt-users] Support on oVirt

 

Can you share logs? Sos report? 

Yaniv Dary
Technical Product Manager
Red Hat Israel Ltd.
34 Jerusalem Road
Building A, 4th floor
Ra'anana, Israel 4350109

Tel : +972 (9) 7692306  
8272306
Email: yd...@redhat.com  
IRC : ydary

 

On Sep 15, 2016 22:28, "Johann van der Westhuizen" mailto:joh...@sharenet.co.za> > wrote:

Hi,

 

Wonder if you could help me with oVirt, having some issues starting a new VM, 
please see below:

 



 

Getting this error just as I try to start the VM, the green play button. What 
can I do to resolve this issue?

 

Regards / Groete,

 

Johann van der Westhuizen

Systems Administrator



joh...@sharenet.co.za   | +27 
  (021) 700 4827 |   
 www.sharenet.co.za

 

   
 

 


___
Users mailing list
Users@ovirt.org  
http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Migrating from KVM to oVirt

2016-09-16 Thread Logan Kuhn
Our current infrastructure is as follows: 

KVM managed with NFS exported raw files. 

We are planning on moving to ovirt, but we want to move our disk images into 
Ceph via Cinder is this possible and if so, how? 

Regards, 
Logan 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Migrating from KVM to oVirt

2016-09-16 Thread Nir Soffer
On Sat, Sep 17, 2016 at 12:19 AM, Logan Kuhn  wrote:

> Our current infrastructure is as follows:
>
> KVM managed with NFS exported raw files.
>
> We are planning on moving to ovirt, but we want to move our disk images
> into Ceph via Cinder is this possible and if so, how?
>

Importing vms from kvm is very easy in 4.0, you just import
the vm, ovirt will do everything for you in few clicks.

We don't support yet copying disks to ceph, only creating new disks.

This should work:

1. Create vm with ceph disk

2. Find the ceph disk name in cinder using the disk name in engine

3. Copy the image manually to ceph using ceph command line tools
The machine copying the disks should have cephx auth keys
deployed.

Note that ovirt hosts do not deploy ceph auth keys on the host, we
deploy the keys dynamically using libvirt secrets, so only vms that
need access to ceph disk can access it.

Adding Daniel to add more details on how to find the cinder disk.

There are some ceph users on this list, maybe they can add more
details.

Nir
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users