[ovirt-users] active directory

2017-06-06 Thread qinglong.d...@horebdata.cn
Hi all,
I used "engine-manage-domains" to add AD to ovirt in earlier version. 
What should I do in ovirt 4.1? Hope someone can help. Thanks!
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Building oVirt engine on Debian

2017-06-06 Thread Leni Kadali Mutungi
On 6/6/17, Yedidyah Bar David  wrote:
> On Tue, Jun 6, 2017 at 6:55 AM, Leni Kadali Mutungi
>  wrote:
> So, did the engine also start successfully?
Successfully ran
`/home/user/ovirt_engine/share/ovirt-engine/services/ovirt-engine/ovirt-engine.py
start`
Unfortunately all I get is the home page. Trying to access the user
portal and administration portal leaves me with a blank page. I guess
this is the consequence of not building the GWT bits?

-- 
- Warm regards
Leni Kadali Mutungi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Doug Ingham
Hey Matthew,
 I think it's VDSM that handles the pausing & resuming of the VMs.

An analogous small-scale scenario...the Gluster layer for one of our
smaller oVirt clusters temporarily lost quorum the other week, locking all
I/O for about 30 minutes. The VMs all went into pause & then resumed
automatically when quorum was restored.

To my surprise/relief, not a single one of the 10 odd VMs reported any
errors.

YMMV

Doug

On 6 June 2017 at 13:45, Matthew Trent 
wrote:

> Thanks for the replies, all!
>
> Yep, Chris is right. TrueNAS HA is active/passive and there isn't a way
> around that when failing between heads.
>
> Sven: In my experience with iX support, they have directed me to reboot
> the active node to initiate failover. There's "hactl takeover" and "hactl
> giveback" commends, but reboot seems to be their preferred method.
>
> VMs going into a paused state and resuming when storage is back online
> sounds great. As long as oVirt's pause/resume isn't significantly slower
> than the 30-or-so seconds the TrueNAS takes to complete its failover,
> that's a pretty tolerable interruption for my needs. So my next questions
> are:
>
> 1) Assuming the SAN failover DOES work correctly, can anyone comment on
> their experience with oVirt pausing/thawing VMs in an NFS-based
> active/passive SAN failover scenario? Does it work reliably without
> intervention? Is it reasonably fast?
>
> 2) Is there anything else in the oVirt stack that might cause it to "freak
> out" rather than gracefully pause/unpause VMs?
>
> 2a) Particularly: I'm running hosted engine on the same TrueNAS storage.
> Does that change anything WRT to timeouts and oVirt's HA and fencing and
> sanlock and such?
>
> 2b) Is there a limit to how long oVirt will wait for storage before doing
> something more drastic than just pausing VMs?
>
> --
> Matthew Trent
> Network Engineer
> Lewis County IT Services
> 360.740.1247 - Helpdesk
> 360.740.3343 - Direct line
>
> 
> From: users-boun...@ovirt.org  on behalf of
> Chris Adams 
> Sent: Tuesday, June 6, 2017 7:21 AM
> To: users@ovirt.org
> Subject: Re: [ovirt-users] Seamless SAN HA failovers with oVirt?
>
> Once upon a time, Juan Pablo  said:
> > Chris, if you have active-active with multipath: you upgrade one system,
> > reboot it, check it came active again, then upgrade the other.
>
> Yes, but that's still not how a TrueNAS (and most other low- to
> mid-range SANs) works, so is not relevant.  The TrueNAS only has a
> single active node talking to the hard drives at a time, because having
> two nodes talking to the same storage at the same time is a hard problem
> to solve (typically requires custom hardware with active cache coherency
> and such).
>
> You can (and should) use multipath between servers and a TrueNAS, and
> that protects against NIC, cable, and switch failures, but does not help
> with a controller failure/reboot/upgrade.  Multipath is also used to
> provide better bandwidth sharing between links than ethernet LAGs.
>
> --
> Chris Adams 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>



-- 
Doug
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine not responding

2017-06-06 Thread Maton, Brett
I managed to solve the problem with guestfish and the help of this page
http://manpages.ubuntu.com/manpages/precise/man1/guestfish.1.html



On 6 June 2017 at 22:07, Maton, Brett  wrote:

> I'm pretty sure that I've got a failing (to boot) kernel on my Self hosted
> engine, as this has happened a few times on 'regular' guest VM's
>
>   I've tried getting it to boot from (several) CD Image, but there is no
> VNC console and hosted-engine --console just hangs at Escape character is ^]
>
> Any suggestions ?
>
> I know which LVM is the hosted engine disk but I don't know how to mount
> it from the physical host, If I could mount it I could probably switch the
> kernel back to an earlier one.
>
> All else fails I have a back up of the engine taken this morning, but I
> seem to recall having problems re-deploying hosted-engine when one already
> exists in the cluster (dead or not)
>
> Regards,
> Brett
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt install error

2017-06-06 Thread ovirt

I finally figured out what the error was all about

The default location for the gdeploy script is:
/usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh

The oVirt node installer 
"ovirt-node-ng-installer-ovirt-4.1-2017060504.iso" installed it in a 
different location:

/usr/share/gdeploy/scripts/grafton-sanity-check.sh

I copied the "gdeploy" folder to the default location & the error went 
away.


**btw, I installed oVirt from scratch twice & both times got the same 
error**


On 2017-06-06 13:01, ov...@fateknollogee.com wrote:

Ok, I will re-check a few things based on this:
https://bugzilla.redhat.com/show_bug.cgi?id=1405447

On 2017-06-06 12:58, ov...@fateknollogee.com wrote:

How do I check that?

Today, I'm re-installing but getting this error message:


PLAY [gluster_servers] 
*


TASK [Run a shell script] 
**

fatal: [ovirt-N1-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg":
"The conditional check 'result.rc != 0' failed. The error was: error
while evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
fatal: [ovirt-N3-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg":
"The conditional check 'result.rc != 0' failed. The error was: error
while evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
fatal: [ovirt-N2-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg":
"The conditional check 'result.rc != 0' failed. The error was: error
while evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
to retry, use: --limit @/tmp/tmpEzKSy6/run-script.retry

PLAY RECAP 
*
ovirt-N1-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1
ovirt-N2-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1
ovirt-N3-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1



On 2017-06-01 00:08, knarra wrote:

On 06/01/2017 01:19 AM, ov...@fateknollogee.com wrote:

Any ideas what this is:

TASK [Run a shell script] 
**
fatal: [ovirt-node1.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error 
while evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-node3.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error 
while evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-node2.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error 
while evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}

to retry, use: --limit @/tmp/tmpaOHOtY/run-script.retry
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Hi,

Can you see if the script which is getting execute is present in 
that path ?


Thanks

kasturi

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Hosted Engine not responding

2017-06-06 Thread Maton, Brett
I'm pretty sure that I've got a failing (to boot) kernel on my Self hosted
engine, as this has happened a few times on 'regular' guest VM's

  I've tried getting it to boot from (several) CD Image, but there is no
VNC console and hosted-engine --console just hangs at Escape character is ^]

Any suggestions ?

I know which LVM is the hosted engine disk but I don't know how to mount it
from the physical host, If I could mount it I could probably switch the
kernel back to an earlier one.

All else fails I have a back up of the engine taken this morning, but I
seem to recall having problems re-deploying hosted-engine when one already
exists in the cluster (dead or not)

Regards,
Brett
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt install error

2017-06-06 Thread ovirt
Ok, I will re-check a few things based on this: 
https://bugzilla.redhat.com/show_bug.cgi?id=1405447


On 2017-06-06 12:58, ov...@fateknollogee.com wrote:

How do I check that?

Today, I'm re-installing but getting this error message:


PLAY [gluster_servers] 
*


TASK [Run a shell script] 
**

fatal: [ovirt-N1-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg":
"The conditional check 'result.rc != 0' failed. The error was: error
while evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
fatal: [ovirt-N3-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg":
"The conditional check 'result.rc != 0' failed. The error was: error
while evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
fatal: [ovirt-N2-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg":
"The conditional check 'result.rc != 0' failed. The error was: error
while evaluating conditional (result.rc != 0): 'dict object' has no
attribute 'rc'"}
to retry, use: --limit @/tmp/tmpEzKSy6/run-script.retry

PLAY RECAP 
*
ovirt-N1-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1
ovirt-N2-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1
ovirt-N3-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1



On 2017-06-01 00:08, knarra wrote:

On 06/01/2017 01:19 AM, ov...@fateknollogee.com wrote:

Any ideas what this is:

TASK [Run a shell script] 
**
fatal: [ovirt-node1.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-node3.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-node2.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}

to retry, use: --limit @/tmp/tmpaOHOtY/run-script.retry
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Hi,

Can you see if the script which is getting execute is present in 
that path ?


Thanks

kasturi

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt install error

2017-06-06 Thread ovirt

How do I check that?

Today, I'm re-installing but getting this error message:


PLAY [gluster_servers] 
*


TASK [Run a shell script] 
**
fatal: [ovirt-N1-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg": 
"The conditional check 'result.rc != 0' failed. The error was: error 
while evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-N3-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg": 
"The conditional check 'result.rc != 0' failed. The error was: error 
while evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-N2-f25.fatek-dc.lab]: FAILED! => {"failed": true, "msg": 
"The conditional check 'result.rc != 0' failed. The error was: error 
while evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}

to retry, use: --limit @/tmp/tmpEzKSy6/run-script.retry

PLAY RECAP 
*
ovirt-N1-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1
ovirt-N2-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1
ovirt-N3-f25.fatek-dc.lab  : ok=0changed=0unreachable=0
failed=1



On 2017-06-01 00:08, knarra wrote:

On 06/01/2017 01:19 AM, ov...@fateknollogee.com wrote:

Any ideas what this is:

TASK [Run a shell script] 
**
fatal: [ovirt-node1.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-node3.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}
fatal: [ovirt-node2.lab]: FAILED! => {"failed": true, "msg": "The 
conditional check 'result.rc != 0' failed. The error was: error while 
evaluating conditional (result.rc != 0): 'dict object' has no 
attribute 'rc'"}

to retry, use: --limit @/tmp/tmpaOHOtY/run-script.retry
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Hi,

Can you see if the script which is getting execute is present in 
that path ?


Thanks

kasturi

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Migrate Hosted Engine to a new storage

2017-06-06 Thread Ling Ho
Hello,

I'm trying to migrate my Hosted Engine from a old NFS storage domain to
a new NFS storage domain. I am running 4.1 now.

I've searched around and found this reference
http://lists.ovirt.org/pipermail/users/2017-January/078739.html. Is it
possible now to migrate Host Engine storage using the GUI under 4.1? I
tried moving the disk using the GUI but it says the VM is not managed by
the engine.

Or do I need to follow the instructions listed in the post?

Thanks,
...
ling
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Alex Crow
I use Open-E in production on standard Intel (Supermicro) hardware. It 
can work in A/A (only in respect of ovirt, ie one LUN normally active on 
one server, the other LUN normally stays on the the other node) or A/P 
mode with multipath. Even in A/P mode it fails over quick enough to 
avoid VM pauses, using virtual IPs that float between the nodes. These 
modes are supported for both iSCSI or NFS.


I've also successfully implemented the same kind of rapid failover using 
standard linux HA tools (pacemaker and corosync). I've had migration 
times under 2s.


NFS has the added complications of filesystem locking. Maybe some of the 
docs on the CTDB site will help, as they ensure that NFS will be running 
on the same ports on each host and locking DBs will be shared between 
the two hosts. I have no idea if TrueNAS supports CTDB or similar 
distributed locking mechanisms.


Caveat: this is with iSCSI resources. I've not really run VMs in oVirt 
in anger against any kind of NFS storage yet. My boss wants to try 
Tintri, so I'll see how that works.


Cheers

Alex

On 06/06/17 18:45, Matthew Trent wrote:

Thanks for the replies, all!

Yep, Chris is right. TrueNAS HA is active/passive and there isn't a way around 
that when failing between heads.

Sven: In my experience with iX support, they have directed me to reboot the active node to initiate 
failover. There's "hactl takeover" and "hactl giveback" commends, but reboot 
seems to be their preferred method.

VMs going into a paused state and resuming when storage is back online sounds 
great. As long as oVirt's pause/resume isn't significantly slower than the 
30-or-so seconds the TrueNAS takes to complete its failover, that's a pretty 
tolerable interruption for my needs. So my next questions are:

1) Assuming the SAN failover DOES work correctly, can anyone comment on their 
experience with oVirt pausing/thawing VMs in an NFS-based active/passive SAN 
failover scenario? Does it work reliably without intervention? Is it reasonably 
fast?

2) Is there anything else in the oVirt stack that might cause it to "freak out" 
rather than gracefully pause/unpause VMs?

2a) Particularly: I'm running hosted engine on the same TrueNAS storage. Does 
that change anything WRT to timeouts and oVirt's HA and fencing and sanlock and 
such?

2b) Is there a limit to how long oVirt will wait for storage before doing 
something more drastic than just pausing VMs?

--
Matthew Trent
Network Engineer
Lewis County IT Services
360.740.1247 - Helpdesk
360.740.3343 - Direct line


From: users-boun...@ovirt.org  on behalf of Chris Adams 

Sent: Tuesday, June 6, 2017 7:21 AM
To: users@ovirt.org
Subject: Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

Once upon a time, Juan Pablo  said:

Chris, if you have active-active with multipath: you upgrade one system,
reboot it, check it came active again, then upgrade the other.

Yes, but that's still not how a TrueNAS (and most other low- to
mid-range SANs) works, so is not relevant.  The TrueNAS only has a
single active node talking to the hard drives at a time, because having
two nodes talking to the same storage at the same time is a hard problem
to solve (typically requires custom hardware with active cache coherency
and such).

You can (and should) use multipath between servers and a TrueNAS, and
that protects against NIC, cable, and switch failures, but does not help
with a controller failure/reboot/upgrade.  Multipath is also used to
provide better bandwidth sharing between links than ethernet LAGs.

--
Chris Adams 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


--
This message is intended only for the addressee and may contain
confidential information. Unless you are that person, you may not
disclose its contents or use it in any way and are requested to delete
the message along with any attachments and notify us immediately.
This email is not intended to, nor should it be taken to, constitute advice.
The information provided is correct to our knowledge & belief and must not
be used as a substitute for obtaining tax, regulatory, investment, legal or
any other appropriate advice.

"Transact" is operated by Integrated Financial Arrangements Ltd.
29 Clement's Lane, London EC4N 7AE. Tel: (020) 7608 4900 Fax: (020) 7608 5300.
(Registered office: as above; Registered in England and Wales under
number: 3727592). Authorised and regulated by the Financial Conduct
Authority (entered on the Financial Services Register; no. 190856).
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Matthew Trent
Thanks for the replies, all!

Yep, Chris is right. TrueNAS HA is active/passive and there isn't a way around 
that when failing between heads.

Sven: In my experience with iX support, they have directed me to reboot the 
active node to initiate failover. There's "hactl takeover" and "hactl giveback" 
commends, but reboot seems to be their preferred method.

VMs going into a paused state and resuming when storage is back online sounds 
great. As long as oVirt's pause/resume isn't significantly slower than the 
30-or-so seconds the TrueNAS takes to complete its failover, that's a pretty 
tolerable interruption for my needs. So my next questions are:

1) Assuming the SAN failover DOES work correctly, can anyone comment on their 
experience with oVirt pausing/thawing VMs in an NFS-based active/passive SAN 
failover scenario? Does it work reliably without intervention? Is it reasonably 
fast?

2) Is there anything else in the oVirt stack that might cause it to "freak out" 
rather than gracefully pause/unpause VMs?

2a) Particularly: I'm running hosted engine on the same TrueNAS storage. Does 
that change anything WRT to timeouts and oVirt's HA and fencing and sanlock and 
such?

2b) Is there a limit to how long oVirt will wait for storage before doing 
something more drastic than just pausing VMs?

--
Matthew Trent
Network Engineer
Lewis County IT Services
360.740.1247 - Helpdesk
360.740.3343 - Direct line


From: users-boun...@ovirt.org  on behalf of Chris 
Adams 
Sent: Tuesday, June 6, 2017 7:21 AM
To: users@ovirt.org
Subject: Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

Once upon a time, Juan Pablo  said:
> Chris, if you have active-active with multipath: you upgrade one system,
> reboot it, check it came active again, then upgrade the other.

Yes, but that's still not how a TrueNAS (and most other low- to
mid-range SANs) works, so is not relevant.  The TrueNAS only has a
single active node talking to the hard drives at a time, because having
two nodes talking to the same storage at the same time is a hard problem
to solve (typically requires custom hardware with active cache coherency
and such).

You can (and should) use multipath between servers and a TrueNAS, and
that protects against NIC, cable, and switch failures, but does not help
with a controller failure/reboot/upgrade.  Multipath is also used to
provide better bandwidth sharing between links than ethernet LAGs.

--
Chris Adams 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [EXTERNAL] Re: oVirt: Host Removal Question

2017-06-06 Thread Alexander Wels
On Tuesday, June 6, 2017 11:18:35 AM EDT Anthony. Fillmore wrote:
> Hey Alexander,
> 
> I did those exact steps roughly two days ago...the host is still stuck in
> preparing for maintenance mode.  Confirming the host has been rebooted
> seems to have no effect.
> 
> Any other ideas?  Some way to hit the Ovirt Database and manipulate the
> value for what state the host is in?  Remove the host entirely from the
> back-end database?
> 

That should have worked, but if all else fails you can always set the host in 
maintaince in the database doing this in the engine database:

update vds_dynamic set status=2 where vds_id = 

you can figure out the id of the host by doing this

select vds_id from vds_static where vds_name=''

If you are feeling brave you can combine the two into a single statement, but 
I would be sure you have the right host before you do that.

Then once you have updated the status, you can refresh the webadmin and the 
host should be in maintenaince mode and you should be able to remove it 
normally. I would not attempt more manual manipulation of the database than 
setting the status to 2 (which is maintenaince).

> -Original Message-
> From: Alexander Wels [mailto:aw...@redhat.com]
> Sent: Tuesday, June 6, 2017 9:29 AM
> To: users@ovirt.org
> Cc: Brandon.Markgraf ; Anthony.Fillmore
>  Subject: [EXTERNAL] Re: [ovirt-users] oVirt:
> Host Removal Question
> On Monday, June 5, 2017 4:10:54 PM EDT Brandon. Markgraf wrote:
> > Hello oVirt Users,
> > We have a cluster that has been decommissioned and we are trying to
> > remove the hosts from the oVirt Engine but one host is stuck in
> > "Preparing for Maintenance".  It's preventing me from removing that
> > host and the associated cluster.
> > 
> > The physical server has been shut down and is no longer accessible.
> > What's the best way to take this server in oVirt Engine out of this
> > status so we can remove the host and the cluster from oVirt Engine?
> > 
> > Thanks so much in advance and please let me know if there are any
> > questions.
> > 
> > Brandon Markgraf | *Target |
> 
> IIRC when I had that issue (I removed the host physically before removing it
> from oVirt) I had to do the following:
> 
> - Right click in the grid on the host, and select confirm host has been
> rebooted (this is basically you telling oVirt the host has been fenced). -
> After a while the host should end up as down (or unknown state). - You
> should be able to select maintaince at that point, this might take a while
> (timeouts etc). - This should put the host in maintaince and you should be
> able to remove it then.


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt: Host Removal Question

2017-06-06 Thread Alexander Wels
On Monday, June 5, 2017 4:10:54 PM EDT Brandon. Markgraf wrote:
> Hello oVirt Users,
> We have a cluster that has been decommissioned and we are trying to remove
> the hosts from the oVirt Engine but one host is stuck in "Preparing for
> Maintenance".  It's preventing me from removing that host and the
> associated cluster.
> 
> The physical server has been shut down and is no longer accessible.  What's
> the best way to take this server in oVirt Engine out of this status so we
> can remove the host and the cluster from oVirt Engine?
> 
> Thanks so much in advance and please let me know if there are any questions.
> 
> Brandon Markgraf | *Target |

IIRC when I had that issue (I removed the host physically before removing it 
from oVirt) I had to do the following:

- Right click in the grid on the host, and select confirm host has been 
rebooted (this is basically you telling oVirt the host has been fenced).
- After a while the host should end up as down (or unknown state).
- You should be able to select maintaince at that point, this might take a 
while (timeouts etc).
- This should put the host in maintaince and you should be able to remove it 
then.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Chris Adams
Once upon a time, Juan Pablo  said:
> Chris, if you have active-active with multipath: you upgrade one system,
> reboot it, check it came active again, then upgrade the other.

Yes, but that's still not how a TrueNAS (and most other low- to
mid-range SANs) works, so is not relevant.  The TrueNAS only has a
single active node talking to the hard drives at a time, because having
two nodes talking to the same storage at the same time is a hard problem
to solve (typically requires custom hardware with active cache coherency
and such).

You can (and should) use multipath between servers and a TrueNAS, and
that protects against NIC, cable, and switch failures, but does not help
with a controller failure/reboot/upgrade.  Multipath is also used to
provide better bandwidth sharing between links than ethernet LAGs.

-- 
Chris Adams 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Juan Pablo
Chris, if you have active-active with multipath: you upgrade one system,
reboot it, check it came active again, then upgrade the other.
-seamless.
-no service interruption.
-not locked to any storage solution.

multipath was designed exactly for that.


2017-06-06 11:03 GMT-03:00 Chris Adams :

> Once upon a time, Juan Pablo  said:
> > Im saying you can do it with multipath and not rely on truenas/freenas.
> > with an active/active configuration on the virt side...instead of
> > active/passive on the storage side.
>
> But there's still only one active system (the active TrueNAS node)
> connected to the hard drives, and the only way to upgrade is to reboot
> it.  Multipath doesn't bypass that.
>
> --
> Chris Adams 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Chris Adams
Once upon a time, Juan Pablo  said:
> Im saying you can do it with multipath and not rely on truenas/freenas.
> with an active/active configuration on the virt side...instead of
> active/passive on the storage side.

But there's still only one active system (the active TrueNAS node)
connected to the hard drives, and the only way to upgrade is to reboot
it.  Multipath doesn't bypass that.

-- 
Chris Adams 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Juan Pablo
Im saying you can do it with multipath and not rely on truenas/freenas.
with an active/active configuration on the virt side...instead of
active/passive on the storage side.



2017-06-06 10:44 GMT-03:00 Chris Adams :

> Once upon a time, Juan Pablo  said:
> > I think its not related to something on the trueNAS side. if you are
> using
> > iscsi multipath you should be using round-robin
>
> TrueNAS HA is active/standby, so multipath has nothing to do with
> rebooting/upgrading a TrueNAS.
>
> --
> Chris Adams 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt: Host Removal Question

2017-06-06 Thread Brandon . Markgraf
Hello oVirt Users,
We have a cluster that has been decommissioned and we are trying to remove the 
hosts from the oVirt Engine but one host is stuck in "Preparing for 
Maintenance".  It's preventing me from removing that host and the associated 
cluster.

The physical server has been shut down and is no longer accessible.  What's the 
best way to take this server in oVirt Engine out of this status so we can 
remove the host and the cluster from oVirt Engine?

Thanks so much in advance and please let me know if there are any questions.

Brandon Markgraf | *Target |

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Chris Adams
Once upon a time, Juan Pablo  said:
> I think its not related to something on the trueNAS side. if you are using
> iscsi multipath you should be using round-robin

TrueNAS HA is active/standby, so multipath has nothing to do with
rebooting/upgrading a TrueNAS.

-- 
Chris Adams 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Juan Pablo
I think its not related to something on the trueNAS side. if you are using
iscsi multipath you should be using round-robin , if one of the paths goes
down you still have the other path with your information., so no sanlock .
unfortunately if you want iscsi mpath on ovirt, its prefered to edit the
config by hand and test. also, with multipath, you can tell the os to 'stop
using' one of the paths( represented as a disk).
 so, for example:
multipath -ll should* be looking like this:
36589cfc0034968eacf965e3c dm-17 FreeNAS ,iSCSI Disk
size=50G features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=0 status=enabled
  |- 31:0:0:5  sdl 8:176 failed faulty running
  |- 32:0:0:5  sdm 8:192 failed faulty running
  |- 35:0:0:5  sdo 8:224 failed faulty running
  `- 34:0:0:5  sdn 8:208 failed faulty running

and working correctly like this:
36589cfc00ee205ed6757fa724bac dm-2 FreeNAS ,iSCSI Disk
size=5.5T features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
  |- 13:0:0:10 sdi 8:128 active ready  running
  |- 15:0:0:10 sdk 8:160 active ready  running
  |- 28:0:0:10 sdg 8:96  active ready  running
  `- 29:0:0:10 sdj 8:144 active ready  running
(yes, they are different ones, I wont disconnect one path just to show an
example =) )

hope I clarified a bit.

cant tell how it would work on nfs or if it works at all.


2017-06-06 10:06 GMT-03:00 Chris Adams :

> Once upon a time, Sven Achtelik  said:
> > I was failing over by rebooting one of the TrueNas nodes and this took
> some time for the other node to take over. I was thinking about asking the
> TN guys if there is a command or procedure to speed up the failover.
>
> That's the way TrueNAS failover works; there is no "graceful" failover,
> you just reboot the active node.
>
> --
> Chris Adams 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Chris Adams
Once upon a time, Sven Achtelik  said:
> I was failing over by rebooting one of the TrueNas nodes and this took some 
> time for the other node to take over. I was thinking about asking the TN guys 
> if there is a command or procedure to speed up the failover.

That's the way TrueNAS failover works; there is no "graceful" failover,
you just reboot the active node.

-- 
Chris Adams 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Nested KVM for oVirt 4.1.2

2017-06-06 Thread ovirt

Well you can always run the engine on a physical node directly too.
The question is why would you want that when hosted engine gives you
fail-over and reliability features.


Failover & reliability are definitely worth having


So I would install Node on all four hosts, enable all of them for
hosted engine and enable three of them for Gluster if hyperconverged
is what your want.


Hyperconverged is what I want.
There does not seem to be a good reason to have a 4th node, so I'll take 
your advice & go 3 nodes + hosted engine.


On 2017-06-06 01:34, Martin Sivak wrote:

Hi,


Real hardware: what I meant to say was I have 4 hosts (not vm's).
If I understand you correctly, I should install oVirt Node (using the 
iso)

on 3 of my hosts & the hosted engine runs as a vm on "Host 1"?


Hosted engine runs as VM on one of the hosts. But not necessarily on
the first one (it can even move if it decides so).


Why can't the hosted engine run on its own host?


Well you can always run the engine on a physical node directly too.
The question is why would you want that when hosted engine gives you
fail-over and reliability features.

So I would install Node on all four hosts, enable all of them for
hosted engine and enable three of them for Gluster if hyperconverged
is what your want.

Best regards

--
Martin Sivak
SLA / oVirt

On Tue, Jun 6, 2017 at 10:09 AM,   wrote:

Sandro, thx for the reply.
Once I get comfortable with oVirt + CentOS, then I'll go & use Fedora 
25/26

and contribute!

Real hardware: what I meant to say was I have 4 hosts (not vm's).
If I understand you correctly, I should install oVirt Node (using the 
iso)

on 3 of my hosts & the hosted engine runs as a vm on "Host 1"?

Why can't the hosted engine run on its own host?


On 2017-06-06 03:29, Sandro Bonazzola wrote:


On Mon, Jun 5, 2017 at 10:13 PM,  wrote:


I want to test oVirt with real hardware, no more nested VMs.
3 hosts, each vm will be Fedora (maybe CentOS, I prefer Fedora)



Please note Fedora support within oVirt project is a best-effort 
task.

There's no testing of oVirt on Fedora done by the core development
team.
That said, I would be happy if you contribute fedora testing :-)

I'm not sure to understand what you're looking into now since you 
talk

about real hardware and VMs in the same sentence.
Suggested minimal deployment is 3 real hardware hosts as hypervisor
nodes and hosted engine with hyperconverged deployment.
You can use oVirt Node iso in order to do the whole deployment using
cockpit.


What is the install process?

On 2017-05-30 02:42, Sandro Bonazzola wrote:

On Tue, May 30, 2017 at 5:18 AM,  wrote:

Sandro,
If & when one decides to "graduate" & use real hardware, what is the
install process?

The install process depends on how you want to design your lab .
You can have a small deployment with just 3 hosts in hyperconverged
setup or a large datacenter with 200 hypervisors and one or more
dedicated SAN for the storage.
If you go with an hyperconverged setup, you can install oVirt Node
on
3 hosts and then on one of them use cockpit to deploy gluster and
hosted engine on top of it in hyperconverged mode.

You can find an installation guide here:
http://www.ovirt.org/documentation/install-guide/Installation_Guide/
[1]

Is the gluster part still automated or that has to be done
manually?

If you go with hyperconverged mode it's now automated. You can find
more info here:



http://www.ovirt.org/develop/release-management/features/gluster/gdeploy-cockpit-integration/


[2]
Sahina, please ensure above link is updated. I see it shows the
feature still in WIP while it's been released.

Another question, what type of use cases & jobs is oVirt being
deployed in & how are people getting tech support?

About use cases for oVirt you can find some examples here:
http://www.ovirt.org/community/user-stories/users-and-providers/ [3]
If you want dedicated support, I would recommend to get a Red Hat
Virtualization (which is oVirt with technical support and some
additions) subscription getting Red Hat support.
Another place to get support if you stay with oVirt is the
community:
this mailing list, the IRC channel and social media, have a look
here
for other contacts: http://www.ovirt.org/community/ [4]

On 2017-05-29 04:33, Sandro Bonazzola wrote:
On Mon, May 29, 2017 at 10:21 AM,  wrote:

I assume people are using oVirt in production?

Sure, I was just wondering why you were running in nested
virtualization :-)
Being your use case a "playground" environment, I can suggest you to
have a look at Lago http://lago.readthedocs.io/en/stable/ [5] [1]
and at Lago demo at https://github.com/lago-project/lago-demo [6]
[2]
to help you preparing an isolated test environment for your
learning.

On 2017-05-29 04:13, Sandro Bonazzola wrote:
On Mon, May 29, 2017 at 12:12 AM,  wrote:





Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-06 Thread Abi Askushi
Just to note that the mentioned logs below are from the dd with bs=512,
which were failing.
Attached the full logs from mount and brick.

Alex

On Tue, Jun 6, 2017 at 3:18 PM, Abi Askushi  wrote:

> Hi Krutika,
>
> My comments inline.
>
> Also attached the strace of:
> strace -y -ff -o /root/512-trace-on-root.log dd if=/dev/zero
> of=/mnt/test2.img oflag=direct bs=512 count=1
>
> and of:
> strace -y -ff -o /root/4096-trace-on-root.log dd if=/dev/zero
> of=/mnt/test2.img oflag=direct bs=4096 count=16
>
> I have mounted gluster volume at /mnt.
> The dd with bs=4096 is successful.
>
> The gluster mount log gives only the following:
> [2017-06-06 12:04:54.102576] W [MSGID: 114031] 
> [client-rpc-fops.c:854:client3_3_writev_cbk]
> 0-engine-client-0: remote operation failed [Invalid argument]
> [2017-06-06 12:04:54.102591] W [MSGID: 114031] 
> [client-rpc-fops.c:854:client3_3_writev_cbk]
> 0-engine-client-1: remote operation failed [Invalid argument]
> [2017-06-06 12:04:54.103355] W [fuse-bridge.c:2312:fuse_writev_cbk]
> 0-glusterfs-fuse: 205: WRITE => -1 gfid=075ab3a5-0274-4f07-a075-2748c3b4d394
> fd=0x7faf1d08706c (Transport endpoint is not connected)
>
> The gluster brick log gives:
> [2017-06-06 12:07:03.793080] E [MSGID: 113072] [posix.c:3453:posix_writev]
> 0-engine-posix: write failed: offset 0, [Invalid argument]
> [2017-06-06 12:07:03.793172] E [MSGID: 115067] 
> [server-rpc-fops.c:1346:server_writev_cbk]
> 0-engine-server: 291: WRITEV 0 (075ab3a5-0274-4f07-a075-2748c3b4d394) ==>
> (Invalid argument) [Invalid argument]
>
>
>
> On Tue, Jun 6, 2017 at 12:50 PM, Krutika Dhananjay 
> wrote:
>
>> OK.
>>
>> So for the 'Transport endpoint is not connected' issue, could you share
>> the mount and brick logs?
>>
>> Hmmm.. 'Invalid argument' error even on the root partition. What if you
>> change bs to 4096 and run?
>>
> If I use bs=4096 the dd is successful on /root and at gluster mounted
> volume.
>
>>
>> The logs I showed in my earlier mail shows that gluster is merely
>> returning the error it got from the disk file system where the
>> brick is hosted. But you're right about the fact that the offset 127488
>> is not 4K-aligned.
>>
>> If the dd on /root worked for you with bs=4096, could you try the same
>> directly on gluster mount point on a dummy file and capture the strace
>> output of dd?
>> You can perhaps reuse your existing gluster volume by mounting it at
>> another location and doing the dd.
>> Here's what you need to execute:
>>
>> strace -ff -T -p  -o 
>> `
>>
>> FWIW, here's something I found in man(2) open:
>>
>>
>>
>>
>> *Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
>> buffer and the file offset must all be multiples of the logical block size
>> of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
>> of the   underlying storage (typically 512 bytes) suffices.  The
>> logical block size can be determined using the ioctl(2) BLKSSZGET operation
>> or from the shell using the command:   blockdev --getss*
>>
> Please note also that the physical disks are of 4K sector size (native).
> Thus OS is having 4096/4096 local/physical sector size.
> [root@v0 ~]# blockdev --getss /dev/sda
> 4096
> [root@v0 ~]# blockdev --getpbsz /dev/sda
> 4096
>
>>
>>
>> -Krutika
>>
>>
>> On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi 
>> wrote:
>>
>>> Also when testing with dd i get the following:
>>>
>>> *Testing on the gluster mount: *
>>> dd if=/dev/zero of=/rhev/data-center/mnt/glusterSD/10.100.100.1:
>>> _engine/test2.img oflag=direct bs=512 count=1
>>> dd: error writing β/rhev/data-center/mnt/glusterSD/10.100.100.1:
>>> _engine/test2.imgβ: *Transport endpoint is not connected*
>>> 1+0 records in
>>> 0+0 records out
>>> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>>>
>>> *Testing on the /root directory (XFS): *
>>> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
>>> dd: error writing β/test2.imgβ:* Invalid argument*
>>> 1+0 records in
>>> 0+0 records out
>>> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>>>
>>> Seems that the gluster is trying to do the same and fails.
>>>
>>>
>>>
>>> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi 
>>> wrote:
>>>
 The question that rises is what is needed to make gluster aware of the
 4K physical sectors presented to it (the logical sector is also 4K). The
 offset (127488) at the log does not seem aligned at 4K.

 Alex

 On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi 
 wrote:

> Hi Krutika,
>
> I am saying that I am facing this issue with 4k drives. I never
> encountered this issue with 512 drives.
>
> Alex
>
> On Jun 5, 2017 14:26, "Krutika Dhananjay"  wrote:
>
>> This seems like a case of O_DIRECT reads and writes gone wrong,
>> judging by the 'Invalid argument' errors.
>>
>> The two operations 

Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-06 Thread Abi Askushi
Hi Krutika,

My comments inline.

Also attached the strace of:
strace -y -ff -o /root/512-trace-on-root.log dd if=/dev/zero
of=/mnt/test2.img oflag=direct bs=512 count=1

and of:
strace -y -ff -o /root/4096-trace-on-root.log dd if=/dev/zero
of=/mnt/test2.img oflag=direct bs=4096 count=16

I have mounted gluster volume at /mnt.
The dd with bs=4096 is successful.

The gluster mount log gives only the following:
[2017-06-06 12:04:54.102576] W [MSGID: 114031]
[client-rpc-fops.c:854:client3_3_writev_cbk] 0-engine-client-0: remote
operation failed [Invalid argument]
[2017-06-06 12:04:54.102591] W [MSGID: 114031]
[client-rpc-fops.c:854:client3_3_writev_cbk] 0-engine-client-1: remote
operation failed [Invalid argument]
[2017-06-06 12:04:54.103355] W [fuse-bridge.c:2312:fuse_writev_cbk]
0-glusterfs-fuse: 205: WRITE => -1
gfid=075ab3a5-0274-4f07-a075-2748c3b4d394 fd=0x7faf1d08706c (Transport
endpoint is not connected)

The gluster brick log gives:
[2017-06-06 12:07:03.793080] E [MSGID: 113072] [posix.c:3453:posix_writev]
0-engine-posix: write failed: offset 0, [Invalid argument]
[2017-06-06 12:07:03.793172] E [MSGID: 115067]
[server-rpc-fops.c:1346:server_writev_cbk] 0-engine-server: 291: WRITEV 0
(075ab3a5-0274-4f07-a075-2748c3b4d394) ==> (Invalid argument) [Invalid
argument]



On Tue, Jun 6, 2017 at 12:50 PM, Krutika Dhananjay 
wrote:

> OK.
>
> So for the 'Transport endpoint is not connected' issue, could you share
> the mount and brick logs?
>
> Hmmm.. 'Invalid argument' error even on the root partition. What if you
> change bs to 4096 and run?
>
If I use bs=4096 the dd is successful on /root and at gluster mounted
volume.

>
> The logs I showed in my earlier mail shows that gluster is merely
> returning the error it got from the disk file system where the
> brick is hosted. But you're right about the fact that the offset 127488 is
> not 4K-aligned.
>
> If the dd on /root worked for you with bs=4096, could you try the same
> directly on gluster mount point on a dummy file and capture the strace
> output of dd?
> You can perhaps reuse your existing gluster volume by mounting it at
> another location and doing the dd.
> Here's what you need to execute:
>
> strace -ff -T -p  -o 
> `
>
> FWIW, here's something I found in man(2) open:
>
>
>
>
> *Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
> buffer and the file offset must all be multiples of the logical block size
> of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
> of the   underlying storage (typically 512 bytes) suffices.  The
> logical block size can be determined using the ioctl(2) BLKSSZGET operation
> or from the shell using the command:   blockdev --getss*
>
Please note also that the physical disks are of 4K sector size (native).
Thus OS is having 4096/4096 local/physical sector size.
[root@v0 ~]# blockdev --getss /dev/sda
4096
[root@v0 ~]# blockdev --getpbsz /dev/sda
4096

>
>
> -Krutika
>
>
> On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi 
> wrote:
>
>> Also when testing with dd i get the following:
>>
>> *Testing on the gluster mount: *
>> dd if=/dev/zero 
>> of=/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.img
>> oflag=direct bs=512 count=1
>> dd: error writing β/rhev/data-center/mnt/glusterSD/10.100.100.1:
>> _engine/test2.imgβ: *Transport endpoint is not connected*
>> 1+0 records in
>> 0+0 records out
>> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>>
>> *Testing on the /root directory (XFS): *
>> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
>> dd: error writing β/test2.imgβ:* Invalid argument*
>> 1+0 records in
>> 0+0 records out
>> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>>
>> Seems that the gluster is trying to do the same and fails.
>>
>>
>>
>> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi 
>> wrote:
>>
>>> The question that rises is what is needed to make gluster aware of the
>>> 4K physical sectors presented to it (the logical sector is also 4K). The
>>> offset (127488) at the log does not seem aligned at 4K.
>>>
>>> Alex
>>>
>>> On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi 
>>> wrote:
>>>
 Hi Krutika,

 I am saying that I am facing this issue with 4k drives. I never
 encountered this issue with 512 drives.

 Alex

 On Jun 5, 2017 14:26, "Krutika Dhananjay"  wrote:

> This seems like a case of O_DIRECT reads and writes gone wrong,
> judging by the 'Invalid argument' errors.
>
> The two operations that have failed on gluster bricks are:
>
> [2017-06-05 09:40:39.428979] E [MSGID: 113072]
> [posix.c:3453:posix_writev] 0-engine-posix: write failed: offset 0,
> [Invalid argument]
> [2017-06-05 09:41:00.865760] E [MSGID: 113040]
> [posix.c:3178:posix_readv] 0-engine-posix: read failed on
> gfid=8c94f658-ac3c-4e3a-b368-8c038513a914, fd=0x7f408584c06c,
> offset=127488 

Re: [ovirt-users] virt-viewer disabling rhel6

2017-06-06 Thread cmc
Thanks for the update Lev.

On Sun, Jun 4, 2017 at 11:48 AM, Lev Veyde  wrote:

> Hi Cam,
>
> The reason why it works in RHEL 6.7 clients is due to the fact that
> version of virt-viewer that is supplied with it, doesn't support the
> mechanism to check for the minimum required version.
>
> Wasn't aware we can modify the versions we require through
> RemoteViewerSupportedVersions config. Michal - thanks for the hint.
>
> Thanks in advance,
>
> On Fri, Jun 2, 2017 at 4:00 PM, cmc  wrote:
>
>> Thanks Michal, that is a huge help. We're busy building an image for EL7
>> but it isn't yet fully finished, so we're still on 6.x for now. We're
>> updating to 6.8 and then 6.9 in the meantime. Interesting that it still
>> works for 6.7 though - I can't explain that. I have updated
>> RemoteViewerSupportedVersions and restarted the engine and it works like a
>> charm.
>>
>> Cheers,
>>
>> Cam
>>
>> On Thu, Jun 1, 2017 at 7:22 PM, Michal Skrivanek <
>> michal.skriva...@redhat.com> wrote:
>>
>>>
>>> On 1 Jun 2017, at 15:10, Lev Veyde  wrote:
>>>
>>> Hi Cam,
>>>
>>> Unfotunately RHEL 6 clients are no longer supported in the oVirt 4.1 due
>>> to the new functions that were added which require a more recent version of
>>> the virt-viewer, thus the issue.
>>>
>>> You should use a more recent version e.g. to use RHEL 7 as the client to
>>> resolve the issue.
>>>
>>>
>>> That said, using engine-config you can change the version check and
>>> allow to launch it . You’ll miss some features but it may not really be
>>> that important if all you need is to see the screen and cannot update
>>> clients to EL7
>>> See https://bugzilla.redhat.com/show_bug.cgi?id=1285883
>>>
>>> Thanks,
>>> michal
>>>
>>>
>>> Thanks in advance,
>>>
>>> On Wed, May 31, 2017 at 4:50 PM, cmc  wrote:
>>>
 Hi,

 virt-viewer no longer appears to work when trying to launch a console
 with EL 6.8 and oVirt 4.1. The error is:

 "At least Remote Viewer version 99.0-1 is required to setup this
 connection"

 When I ran remote-viewer in debug mode, it seems that it is
 deliberately disabling rhel6 by setting the version to a non-existent
 version:

 (remote-viewer:23829): remote-viewer-DEBUG: Minimum version '2.0-160'
 for OS id 'rhev-win64'
 (remote-viewer:23829): remote-viewer-DEBUG: Minimum version '2.0-160'
 for OS id 'rhev-win32'
 (remote-viewer:23829): remote-viewer-DEBUG: Minimum version '2.0-6'
 for OS id 'rhel7'
 (remote-viewer:23829): remote-viewer-DEBUG: Minimum version '99.0-1'
 for OS id 'rhel6'

 rhel 6.7 (and presumably brfore) works fine. I contacted the
 maintainers of virt-viewer and they said that this is an ovirt issue.
 Is this somehow disabled in 4.1? Can someone tell me why this is the
 case?

 Thanks in advance for any insights,

 Cam
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

>>>
>>>
>>>
>>> --
>>>
>>> Lev Veyde
>>>
>>> Software Engineer, RHCE | RHCVA | MCITP
>>> Red Hat Israel
>>>
>>> 
>>>
>>> l...@redhat.com | lve...@redhat.com
>>> 
>>> TRIED. TESTED. TRUSTED. 
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
> --
>
> Lev Veyde
>
> Software Engineer, RHCE | RHCVA | MCITP
>
> Red Hat Israel
>
> 
>
> l...@redhat.com | lve...@redhat.com
> 
> TRIED. TESTED. TRUSTED. 
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-06 Thread Krutika Dhananjay
I stand corrected.

Just realised the strace command I gave was wrong.

Here's what you would actually need to execute:

strace -y -ff -o  

-Krutika

On Tue, Jun 6, 2017 at 3:20 PM, Krutika Dhananjay 
wrote:

> OK.
>
> So for the 'Transport endpoint is not connected' issue, could you share
> the mount and brick logs?
>
> Hmmm.. 'Invalid argument' error even on the root partition. What if you
> change bs to 4096 and run?
>
> The logs I showed in my earlier mail shows that gluster is merely
> returning the error it got from the disk file system where the
> brick is hosted. But you're right about the fact that the offset 127488 is
> not 4K-aligned.
>
> If the dd on /root worked for you with bs=4096, could you try the same
> directly on gluster mount point on a dummy file and capture the strace
> output of dd?
> You can perhaps reuse your existing gluster volume by mounting it at
> another location and doing the dd.
> Here's what you need to execute:
>
> strace -ff -T -p  -o 
> `
>
> FWIW, here's something I found in man(2) open:
>
>
>
>
> *Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
> buffer and the file offset must all be multiples of the logical block size
> of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
> of the   underlying storage (typically 512 bytes) suffices.  The
> logical block size can be determined using the ioctl(2) BLKSSZGET operation
> or from the shell using the command:   blockdev --getss*
>
>
> -Krutika
>
>
> On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi 
> wrote:
>
>> Also when testing with dd i get the following:
>>
>> *Testing on the gluster mount: *
>> dd if=/dev/zero 
>> of=/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.img
>> oflag=direct bs=512 count=1
>> dd: error writing β/rhev/data-center/mnt/glusterSD/10.100.100.1:
>> _engine/test2.imgβ: *Transport endpoint is not connected*
>> 1+0 records in
>> 0+0 records out
>> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>>
>> *Testing on the /root directory (XFS): *
>> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
>> dd: error writing β/test2.imgβ:* Invalid argument*
>> 1+0 records in
>> 0+0 records out
>> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>>
>> Seems that the gluster is trying to do the same and fails.
>>
>>
>>
>> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi 
>> wrote:
>>
>>> The question that rises is what is needed to make gluster aware of the
>>> 4K physical sectors presented to it (the logical sector is also 4K). The
>>> offset (127488) at the log does not seem aligned at 4K.
>>>
>>> Alex
>>>
>>> On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi 
>>> wrote:
>>>
 Hi Krutika,

 I am saying that I am facing this issue with 4k drives. I never
 encountered this issue with 512 drives.

 Alex

 On Jun 5, 2017 14:26, "Krutika Dhananjay"  wrote:

> This seems like a case of O_DIRECT reads and writes gone wrong,
> judging by the 'Invalid argument' errors.
>
> The two operations that have failed on gluster bricks are:
>
> [2017-06-05 09:40:39.428979] E [MSGID: 113072]
> [posix.c:3453:posix_writev] 0-engine-posix: write failed: offset 0,
> [Invalid argument]
> [2017-06-05 09:41:00.865760] E [MSGID: 113040]
> [posix.c:3178:posix_readv] 0-engine-posix: read failed on
> gfid=8c94f658-ac3c-4e3a-b368-8c038513a914, fd=0x7f408584c06c,
> offset=127488 size=512, buf=0x7f4083c0b000 [Invalid argument]
>
> But then, both the write and the read have 512byte-aligned offset,
> size and buf address (which is correct).
>
> Are you saying you don't see this issue with 4K block-size?
>
> -Krutika
>
> On Mon, Jun 5, 2017 at 3:21 PM, Abi Askushi 
> wrote:
>
>> Hi Sahina,
>>
>> Attached are the logs. Let me know if sth else is needed.
>>
>> I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K
>> stripe size at the moment.
>> I have prepared the storage as below:
>>
>> pvcreate --dataalignment 256K /dev/sda4
>> vgcreate --physicalextentsize 256K gluster /dev/sda4
>>
>> lvcreate -n engine --size 120G gluster
>> mkfs.xfs -f -i size=512 /dev/gluster/engine
>>
>> Thanx,
>> Alex
>>
>> On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose 
>> wrote:
>>
>>> Can we have the gluster mount logs and brick logs to check if it's
>>> the same issue?
>>>
>>> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi <
>>> rightkickt...@gmail.com> wrote:
>>>
 I clean installed everything and ran into the same.
 I then ran gdeploy and encountered the same issue when deploying
 engine.
 Seems that gluster (?) doesn't like 4K sector drives. I am not sure
 if it has to do with 

[ovirt-users] Live migration error in 4.1.2 (next attempt)

2017-06-06 Thread Vadim

Can anybody help me to solve this.

I'm having trouble in live migration. Migration always finished with error. May 
be it is essential, on dashboard cluster status always N/A. VM can run on both 
hosts. After turn on debug log for libvirt i'v got such errors

2017-06-06 09:41:04.842+: 1302: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:42:04.847+: 1305: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:43:04.850+: 1304: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:44:04.841+: 1301: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:44:25.373+: 10320: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:44:55.373+: 10320: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:45:04.851+: 1303: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:45:25.373+: 10320: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:46:04.852+: 1302: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:47:04.858+: 1305: error : qemuDomainObjBeginJobInternal:3107 
: Timed out during operation: cannot acquire state change lock (held by 
remoteDispatchDomainMigratePrepare3Params)
2017-06-06 09:47:19.950+: 1263: error : qemuMonitorIO:695 : internal error: 
End of file from monitor
2017-06-06 09:47:19.951+: 1263: error : qemuProcessReportLogError:1810 : 
internal error: qemu unexpectedly closed the monitor: 
2017-06-06T09:40:26.681446Z qemu-kvm: warning: CPU(s) not present in any NUMA 
nodes: 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15


Ovirt 4.1.1 clean install ugraded to 4.1.2

I tried different migration policies but all of them ended by error.

libvirt debug log attached.

#rpm- qa | grep -e libvirt -e qemu | sort

centos-release-qemu-ev-1.0-1.el7.noarch
ipxe-roms-qemu-20160127-5.git6366fa7a.el7.noarch
libvirt-client-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-config-nwfilter-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-interface-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-network-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-nodedev-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-nwfilter-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-qemu-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-secret-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-driver-storage-2.0.0-10.el7_3.9.x86_64
libvirt-daemon-kvm-2.0.0-10.el7_3.9.x86_64
libvirt-lock-sanlock-2.0.0-10.el7_3.9.x86_64
libvirt-python-2.0.0-2.el7.x86_64
qemu-guest-agent-2.5.0-3.el7.x86_64
qemu-img-ev-2.6.0-28.el7_3.9.1.x86_64
qemu-kvm-common-ev-2.6.0-28.el7_3.9.1.x86_64
qemu-kvm-ev-2.6.0-28.el7_3.9.1.x86_64
qemu-kvm-tools-ev-2.6.0-28.el7_3.9.1.x86_64
 

--
Thanks,
Vadim





migration.tar.bz2
Description: application/bzip
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt gluster sanlock issue

2017-06-06 Thread Krutika Dhananjay
OK.

So for the 'Transport endpoint is not connected' issue, could you share the
mount and brick logs?

Hmmm.. 'Invalid argument' error even on the root partition. What if you
change bs to 4096 and run?

The logs I showed in my earlier mail shows that gluster is merely returning
the error it got from the disk file system where the
brick is hosted. But you're right about the fact that the offset 127488 is
not 4K-aligned.

If the dd on /root worked for you with bs=4096, could you try the same
directly on gluster mount point on a dummy file and capture the strace
output of dd?
You can perhaps reuse your existing gluster volume by mounting it at
another location and doing the dd.
Here's what you need to execute:

strace -ff -T -p  -o
`

FWIW, here's something I found in man(2) open:




*Under  Linux  2.4,  transfer  sizes,  and  the alignment of the user
buffer and the file offset must all be multiples of the logical block size
of the filesystem.  Since Linux 2.6.0, alignment to the logical block size
of the   underlying storage (typically 512 bytes) suffices.  The
logical block size can be determined using the ioctl(2) BLKSSZGET operation
or from the shell using the command:   blockdev --getss*


-Krutika


On Tue, Jun 6, 2017 at 1:18 AM, Abi Askushi  wrote:

> Also when testing with dd i get the following:
>
> *Testing on the gluster mount: *
> dd if=/dev/zero 
> of=/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.img
> oflag=direct bs=512 count=1
> dd: error writing 
> β/rhev/data-center/mnt/glusterSD/10.100.100.1:_engine/test2.imgβ:
> *Transport endpoint is not connected*
> 1+0 records in
> 0+0 records out
> 0 bytes (0 B) copied, 0.00336755 s, 0.0 kB/s
>
> *Testing on the /root directory (XFS): *
> dd if=/dev/zero of=/test2.img oflag=direct bs=512 count=1
> dd: error writing β/test2.imgβ:* Invalid argument*
> 1+0 records in
> 0+0 records out
> 0 bytes (0 B) copied, 0.000321239 s, 0.0 kB/s
>
> Seems that the gluster is trying to do the same and fails.
>
>
>
> On Mon, Jun 5, 2017 at 10:10 PM, Abi Askushi 
> wrote:
>
>> The question that rises is what is needed to make gluster aware of the 4K
>> physical sectors presented to it (the logical sector is also 4K). The
>> offset (127488) at the log does not seem aligned at 4K.
>>
>> Alex
>>
>> On Mon, Jun 5, 2017 at 2:47 PM, Abi Askushi 
>> wrote:
>>
>>> Hi Krutika,
>>>
>>> I am saying that I am facing this issue with 4k drives. I never
>>> encountered this issue with 512 drives.
>>>
>>> Alex
>>>
>>> On Jun 5, 2017 14:26, "Krutika Dhananjay"  wrote:
>>>
 This seems like a case of O_DIRECT reads and writes gone wrong, judging
 by the 'Invalid argument' errors.

 The two operations that have failed on gluster bricks are:

 [2017-06-05 09:40:39.428979] E [MSGID: 113072]
 [posix.c:3453:posix_writev] 0-engine-posix: write failed: offset 0,
 [Invalid argument]
 [2017-06-05 09:41:00.865760] E [MSGID: 113040]
 [posix.c:3178:posix_readv] 0-engine-posix: read failed on
 gfid=8c94f658-ac3c-4e3a-b368-8c038513a914, fd=0x7f408584c06c,
 offset=127488 size=512, buf=0x7f4083c0b000 [Invalid argument]

 But then, both the write and the read have 512byte-aligned offset, size
 and buf address (which is correct).

 Are you saying you don't see this issue with 4K block-size?

 -Krutika

 On Mon, Jun 5, 2017 at 3:21 PM, Abi Askushi 
 wrote:

> Hi Sahina,
>
> Attached are the logs. Let me know if sth else is needed.
>
> I have 5 disks (with 4K physical sector) in RAID5. The RAID has 64K
> stripe size at the moment.
> I have prepared the storage as below:
>
> pvcreate --dataalignment 256K /dev/sda4
> vgcreate --physicalextentsize 256K gluster /dev/sda4
>
> lvcreate -n engine --size 120G gluster
> mkfs.xfs -f -i size=512 /dev/gluster/engine
>
> Thanx,
> Alex
>
> On Mon, Jun 5, 2017 at 12:14 PM, Sahina Bose 
> wrote:
>
>> Can we have the gluster mount logs and brick logs to check if it's
>> the same issue?
>>
>> On Sun, Jun 4, 2017 at 11:21 PM, Abi Askushi > > wrote:
>>
>>> I clean installed everything and ran into the same.
>>> I then ran gdeploy and encountered the same issue when deploying
>>> engine.
>>> Seems that gluster (?) doesn't like 4K sector drives. I am not sure
>>> if it has to do with alignment. The weird thing is that gluster volumes 
>>> are
>>> all ok, replicating normally and no split brain is reported.
>>>
>>> The solution to the mentioned bug (1386443
>>> ) was to
>>> format with 512 sector size, which for my case is not an option:
>>>
>>> mkfs.xfs -f -i size=512 -s size=512 

Re: [ovirt-users] Nested KVM for oVirt 4.1.2

2017-06-06 Thread Martin Sivak
Hi,

> Real hardware: what I meant to say was I have 4 hosts (not vm's).
> If I understand you correctly, I should install oVirt Node (using the iso)
> on 3 of my hosts & the hosted engine runs as a vm on "Host 1"?

Hosted engine runs as VM on one of the hosts. But not necessarily on
the first one (it can even move if it decides so).

> Why can't the hosted engine run on its own host?

Well you can always run the engine on a physical node directly too.
The question is why would you want that when hosted engine gives you
fail-over and reliability features.

So I would install Node on all four hosts, enable all of them for
hosted engine and enable three of them for Gluster if hyperconverged
is what your want.

Best regards

--
Martin Sivak
SLA / oVirt

On Tue, Jun 6, 2017 at 10:09 AM,   wrote:
> Sandro, thx for the reply.
> Once I get comfortable with oVirt + CentOS, then I'll go & use Fedora 25/26
> and contribute!
>
> Real hardware: what I meant to say was I have 4 hosts (not vm's).
> If I understand you correctly, I should install oVirt Node (using the iso)
> on 3 of my hosts & the hosted engine runs as a vm on "Host 1"?
>
> Why can't the hosted engine run on its own host?
>
>
> On 2017-06-06 03:29, Sandro Bonazzola wrote:
>>
>> On Mon, Jun 5, 2017 at 10:13 PM,  wrote:
>>
>>> I want to test oVirt with real hardware, no more nested VMs.
>>> 3 hosts, each vm will be Fedora (maybe CentOS, I prefer Fedora)
>>
>>
>> Please note Fedora support within oVirt project is a best-effort task.
>> There's no testing of oVirt on Fedora done by the core development
>> team.
>> That said, I would be happy if you contribute fedora testing :-)
>>
>> I'm not sure to understand what you're looking into now since you talk
>> about real hardware and VMs in the same sentence.
>> Suggested minimal deployment is 3 real hardware hosts as hypervisor
>> nodes and hosted engine with hyperconverged deployment.
>> You can use oVirt Node iso in order to do the whole deployment using
>> cockpit.
>>
>>> What is the install process?
>>>
>>> On 2017-05-30 02:42, Sandro Bonazzola wrote:
>>>
>>> On Tue, May 30, 2017 at 5:18 AM,  wrote:
>>>
>>> Sandro,
>>> If & when one decides to "graduate" & use real hardware, what is the
>>> install process?
>>>
>>> The install process depends on how you want to design your lab .
>>> You can have a small deployment with just 3 hosts in hyperconverged
>>> setup or a large datacenter with 200 hypervisors and one or more
>>> dedicated SAN for the storage.
>>> If you go with an hyperconverged setup, you can install oVirt Node
>>> on
>>> 3 hosts and then on one of them use cockpit to deploy gluster and
>>> hosted engine on top of it in hyperconverged mode.
>>>
>>> You can find an installation guide here:
>>> http://www.ovirt.org/documentation/install-guide/Installation_Guide/
>>> [1]
>>>
>>> Is the gluster part still automated or that has to be done
>>> manually?
>>>
>>> If you go with hyperconverged mode it's now automated. You can find
>>> more info here:
>>>
>>
>> http://www.ovirt.org/develop/release-management/features/gluster/gdeploy-cockpit-integration/
>>>
>>> [2]
>>> Sahina, please ensure above link is updated. I see it shows the
>>> feature still in WIP while it's been released.
>>>
>>> Another question, what type of use cases & jobs is oVirt being
>>> deployed in & how are people getting tech support?
>>>
>>> About use cases for oVirt you can find some examples here:
>>> http://www.ovirt.org/community/user-stories/users-and-providers/ [3]
>>> If you want dedicated support, I would recommend to get a Red Hat
>>> Virtualization (which is oVirt with technical support and some
>>> additions) subscription getting Red Hat support.
>>> Another place to get support if you stay with oVirt is the
>>> community:
>>> this mailing list, the IRC channel and social media, have a look
>>> here
>>> for other contacts: http://www.ovirt.org/community/ [4]
>>>
>>> On 2017-05-29 04:33, Sandro Bonazzola wrote:
>>> On Mon, May 29, 2017 at 10:21 AM,  wrote:
>>>
>>> I assume people are using oVirt in production?
>>>
>>> Sure, I was just wondering why you were running in nested
>>> virtualization :-)
>>> Being your use case a "playground" environment, I can suggest you to
>>> have a look at Lago http://lago.readthedocs.io/en/stable/ [5] [1]
>>> and at Lago demo at https://github.com/lago-project/lago-demo [6]
>>> [2]
>>> to help you preparing an isolated test environment for your
>>> learning.
>>>
>>> On 2017-05-29 04:13, Sandro Bonazzola wrote:
>>> On Mon, May 29, 2017 at 12:12 AM,  wrote:
>>>
>>>
>>
>> http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-with-nested-kvm/
>>>
>>> [7]
>>> [3]
>>> [1]
>>> [1]
>>>
>>> I have one CentOS7 host (physical) & 3x oVirt nodes 4.1.2 (these are
>>> vm's).
>>>
>>> Hi, can you please share the use case for this setup?
>>>
>>> I have installed vdsm-hook-nestedvm 

Re: [ovirt-users] Nested KVM for oVirt 4.1.2

2017-06-06 Thread ovirt

Sandro, thx for the reply.
Once I get comfortable with oVirt + CentOS, then I'll go & use Fedora 
25/26 and contribute!


Real hardware: what I meant to say was I have 4 hosts (not vm's).
If I understand you correctly, I should install oVirt Node (using the 
iso) on 3 of my hosts & the hosted engine runs as a vm on "Host 1"?


Why can't the hosted engine run on its own host?

On 2017-06-06 03:29, Sandro Bonazzola wrote:

On Mon, Jun 5, 2017 at 10:13 PM,  wrote:


I want to test oVirt with real hardware, no more nested VMs.
3 hosts, each vm will be Fedora (maybe CentOS, I prefer Fedora)


Please note Fedora support within oVirt project is a best-effort task.
There's no testing of oVirt on Fedora done by the core development
team.
That said, I would be happy if you contribute fedora testing :-)

I'm not sure to understand what you're looking into now since you talk
about real hardware and VMs in the same sentence.
Suggested minimal deployment is 3 real hardware hosts as hypervisor
nodes and hosted engine with hyperconverged deployment.
You can use oVirt Node iso in order to do the whole deployment using
cockpit.


What is the install process?

On 2017-05-30 02:42, Sandro Bonazzola wrote:

On Tue, May 30, 2017 at 5:18 AM,  wrote:

Sandro,
If & when one decides to "graduate" & use real hardware, what is the
install process?

The install process depends on how you want to design your lab .
You can have a small deployment with just 3 hosts in hyperconverged
setup or a large datacenter with 200 hypervisors and one or more
dedicated SAN for the storage.
If you go with an hyperconverged setup, you can install oVirt Node
on
3 hosts and then on one of them use cockpit to deploy gluster and
hosted engine on top of it in hyperconverged mode.

You can find an installation guide here:
http://www.ovirt.org/documentation/install-guide/Installation_Guide/
[1]

Is the gluster part still automated or that has to be done
manually?

If you go with hyperconverged mode it's now automated. You can find
more info here:


http://www.ovirt.org/develop/release-management/features/gluster/gdeploy-cockpit-integration/

[2]
Sahina, please ensure above link is updated. I see it shows the
feature still in WIP while it's been released.

Another question, what type of use cases & jobs is oVirt being
deployed in & how are people getting tech support?

About use cases for oVirt you can find some examples here:
http://www.ovirt.org/community/user-stories/users-and-providers/ [3]
If you want dedicated support, I would recommend to get a Red Hat
Virtualization (which is oVirt with technical support and some
additions) subscription getting Red Hat support.
Another place to get support if you stay with oVirt is the
community:
this mailing list, the IRC channel and social media, have a look
here
for other contacts: http://www.ovirt.org/community/ [4]

On 2017-05-29 04:33, Sandro Bonazzola wrote:
On Mon, May 29, 2017 at 10:21 AM,  wrote:

I assume people are using oVirt in production?

Sure, I was just wondering why you were running in nested
virtualization :-)
Being your use case a "playground" environment, I can suggest you to
have a look at Lago http://lago.readthedocs.io/en/stable/ [5] [1]
and at Lago demo at https://github.com/lago-project/lago-demo [6]
[2]
to help you preparing an isolated test environment for your
learning.

On 2017-05-29 04:13, Sandro Bonazzola wrote:
On Mon, May 29, 2017 at 12:12 AM,  wrote:



http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-with-nested-kvm/

[7]
[3]
[1]
[1]

I have one CentOS7 host (physical) & 3x oVirt nodes 4.1.2 (these are
vm's).

Hi, can you please share the use case for this setup?

I have installed vdsm-hook-nestedvm on the host.

Should I install vdsm-hook-macspoof on the 3x node vm's?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users [8] [4] [2] [2]

--

SANDRO BONAZZOLA

ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

Red Hat EMEA [3]

[4]

TRIED. TESTED. TRUSTED. [5]

Links:
--
[1]



http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-with-nested-kvm/

[7]
[3]
[1]
[2] http://lists.ovirt.org/mailman/listinfo/users [8] [4] [2]
[3] https://www.redhat.com/
[4] https://red.ht/sig
[5] https://redhat.com/trusted

--

SANDRO BONAZZOLA

ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

Red Hat EMEA [3]

[4]

TRIED. TESTED. TRUSTED. [5]

Links:
--
[1]



http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-with-nested-kvm/

[7]
[3]
[2] http://lists.ovirt.org/mailman/listinfo/users [8] [4]
[3] https://www.redhat.com/
[4] https://red.ht/sig
[5] https://redhat.com/trusted

--

SANDRO BONAZZOLA

ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

Red Hat EMEA [5]

[6]

TRIED. TESTED. TRUSTED. [7]

Links:
--
[1] 

Re: [ovirt-users] hosted-engine network

2017-06-06 Thread Arsène Gschwind

Ive forgot to mention, i'm runing the latest oVirt 4.1.2 version.

Thanks


On 06/06/2017 09:50 AM, Arsène Gschwind wrote:


Hi,

I've migrated our oVirt engine to hosted-engine located on a FC 
storage LUN, so far so good.
For some reason I'm not able to start the hosted-engine VM, after 
digging in the log files i could figured out the reason. The Network 
device was set to "None" as follow:

devices={nicModel:pv,macAddr:00:16:3e:3a:6b:60,linkActive:true,network:*None*,deviceId:56cb4d71-13ff-42a8-bb83-7faef99fd3ea,address:{slot:0x03,bus:0x00,domain:0x,type:pci,function:0x0},device:bridge,type:interface}

I've created a new config file 
/var/run/ovirt-hosted-engine-ha/vm.conf.new and set the nic device to 
ovirtmgnt, the I could start the hosted-engine using :
hosted-engine --vm-start 
--vm-conf=var/run/ovirt-hosted-engine-ha/vm.conf.new


The nic  device line in vm.conf.new looks like:
devices={nicModel:pv,macAddr:00:16:3e:3a:6b:60,linkActive:true,network:*ovirtmgmt*,deviceId:56cb4d71-13ff-42a8-bb83-7faef99fd3ea,address:{slot:0x03,bus:0x00,domain:0x,type:pci,function:0x0},device:bridge,type:interface}

I tried to find out a way to change this setting on vm.conf but it is 
constantly rewritten, even when using the webui is say's:

HostedEngine:

  * There was an attempt to change Hosted Engine VM values that are
locked.

Is there a way to modify/correct the hosted-engine vm.conf file so it 
will stay and not be overwritten?


Thanks a lot for any hint/help

rgds,
arsène

--

*Arsène Gschwind*
Fa. Sapify AG im Auftrag der Universität Basel
IT Services
Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
Tel. +41 79 449 25 63  | http://its.unibas.ch 
ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


--

*Arsène Gschwind*
Fa. Sapify AG im Auftrag der Universität Basel
IT Services
Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
Tel. +41 79 449 25 63  | http://its.unibas.ch 
ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] hosted-engine network

2017-06-06 Thread Arsène Gschwind

Hi,

I've migrated our oVirt engine to hosted-engine located on a FC storage 
LUN, so far so good.
For some reason I'm not able to start the hosted-engine VM, after 
digging in the log files i could figured out the reason. The Network 
device was set to "None" as follow:

devices={nicModel:pv,macAddr:00:16:3e:3a:6b:60,linkActive:true,network:*None*,deviceId:56cb4d71-13ff-42a8-bb83-7faef99fd3ea,address:{slot:0x03,bus:0x00,domain:0x,type:pci,function:0x0},device:bridge,type:interface}

I've created a new config file 
/var/run/ovirt-hosted-engine-ha/vm.conf.new and set the nic device to 
ovirtmgnt, the I could start the hosted-engine using :
hosted-engine --vm-start 
--vm-conf=var/run/ovirt-hosted-engine-ha/vm.conf.new


The nic  device line in vm.conf.new looks like:
devices={nicModel:pv,macAddr:00:16:3e:3a:6b:60,linkActive:true,network:*ovirtmgmt*,deviceId:56cb4d71-13ff-42a8-bb83-7faef99fd3ea,address:{slot:0x03,bus:0x00,domain:0x,type:pci,function:0x0},device:bridge,type:interface}

I tried to find out a way to change this setting on vm.conf but it is 
constantly rewritten, even when using the webui is say's:

HostedEngine:

 * There was an attempt to change Hosted Engine VM values that are locked.

Is there a way to modify/correct the hosted-engine vm.conf file so it 
will stay and not be overwritten?


Thanks a lot for any hint/help

rgds,
arsène

--

*Arsène Gschwind*
Fa. Sapify AG im Auftrag der Universität Basel
IT Services
Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
Tel. +41 79 449 25 63  | http://its.unibas.ch 
ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Nested KVM for oVirt 4.1.2

2017-06-06 Thread Sandro Bonazzola
On Mon, Jun 5, 2017 at 10:13 PM,  wrote:

> I want to test oVirt with real hardware, no more nested VMs.
> 3 hosts, each vm will be Fedora (maybe CentOS, I prefer Fedora)
>

Please note Fedora support within oVirt project is a best-effort task.
There's no testing of oVirt on Fedora done by the core development team.
That said, I would be happy if you contribute fedora testing :-)

I'm not sure to understand what you're looking into now since you talk
about real hardware and VMs in the same sentence.
Suggested minimal deployment is 3 real hardware hosts as hypervisor nodes
and hosted engine with hyperconverged deployment.
You can use oVirt Node iso in order to do the whole deployment using
cockpit.



> What is the install process?
>
>
> On 2017-05-30 02:42, Sandro Bonazzola wrote:
>
>> On Tue, May 30, 2017 at 5:18 AM,  wrote:
>>
>> Sandro,
>>> If & when one decides to "graduate" & use real hardware, what is the
>>> install process?
>>>
>>
>> The install process depends on how you want to design your lab .
>> You can have a small deployment with just 3 hosts in hyperconverged
>> setup or a large datacenter with 200 hypervisors and one or more
>> dedicated SAN for the storage.
>> If you go with an hyperconverged setup, you can install oVirt Node on
>> 3 hosts and then on one of them use cockpit to deploy gluster and
>> hosted engine on top of it in hyperconverged mode.
>>
>> You can find an installation guide here:
>> http://www.ovirt.org/documentation/install-guide/Installation_Guide/
>>
>> Is the gluster part still automated or that has to be done manually?
>>>
>>
>> If you go with hyperconverged mode it's now automated. You can find
>> more info here:
>> http://www.ovirt.org/develop/release-management/features/glu
>> ster/gdeploy-cockpit-integration/
>> Sahina, please ensure above link is updated. I see it shows the
>> feature still in WIP while it's been released.
>>
>> Another question, what type of use cases & jobs is oVirt being
>>> deployed in & how are people getting tech support?
>>>
>>
>> About use cases for oVirt you can find some examples here:
>> http://www.ovirt.org/community/user-stories/users-and-providers/
>> If you want dedicated support, I would recommend to get a Red Hat
>> Virtualization (which is oVirt with technical support and some
>> additions) subscription getting Red Hat support.
>> Another place to get support if you stay with oVirt is the community:
>> this mailing list, the IRC channel and social media, have a look here
>> for other contacts: http://www.ovirt.org/community/
>>
>> On 2017-05-29 04:33, Sandro Bonazzola wrote:
>>> On Mon, May 29, 2017 at 10:21 AM,  wrote:
>>>
>>> I assume people are using oVirt in production?
>>>
>>> Sure, I was just wondering why you were running in nested
>>> virtualization :-)
>>> Being your use case a "playground" environment, I can suggest you to
>>> have a look at Lago http://lago.readthedocs.io/en/stable/ [1]
>>> and at Lago demo at https://github.com/lago-project/lago-demo [2]
>>> to help you preparing an isolated test environment for your
>>> learning.
>>>
>>> On 2017-05-29 04:13, Sandro Bonazzola wrote:
>>> On Mon, May 29, 2017 at 12:12 AM,  wrote:
>>>
>>>
>>> http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-
>> with-nested-kvm/
>>
>>> [3]
>>> [1]
>>> [1]
>>>
>>> I have one CentOS7 host (physical) & 3x oVirt nodes 4.1.2 (these are
>>> vm's).
>>>
>>> Hi, can you please share the use case for this setup?
>>>
>>> I have installed vdsm-hook-nestedvm on the host.
>>>
>>> Should I install vdsm-hook-macspoof on the 3x node vm's?
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users [4] [2] [2]
>>>
>>> --
>>>
>>> SANDRO BONAZZOLA
>>>
>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R
>>>
>>> Red Hat EMEA [3]
>>>
>>> [4]
>>>
>>> TRIED. TESTED. TRUSTED. [5]
>>>
>>> Links:
>>> --
>>> [1]
>>>
>>>
>>> http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-
>> with-nested-kvm/
>>
>>> [3]
>>> [1]
>>> [2] http://lists.ovirt.org/mailman/listinfo/users [4] [2]
>>> [3] https://www.redhat.com/
>>> [4] https://red.ht/sig
>>> [5] https://redhat.com/trusted
>>>
>>> --
>>>
>>> SANDRO BONAZZOLA
>>>
>>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R
>>>
>>> Red Hat EMEA [3]
>>>
>>> [4]
>>>
>>> TRIED. TESTED. TRUSTED. [5]
>>>
>>> Links:
>>> --
>>> [1]
>>>
>>> http://community.redhat.com/blog/2013/08/testing-ovirt-3-3-
>> with-nested-kvm/
>>
>>> [3]
>>> [2] http://lists.ovirt.org/mailman/listinfo/users [4]
>>> [3] https://www.redhat.com/
>>> [4] https://red.ht/sig
>>> [5] https://redhat.com/trusted
>>>
>>
>> --
>>
>> SANDRO BONAZZOLA
>>
>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R
>>
>> Red Hat EMEA [5]
>>
>>  [6]
>>
>> TRIED. TESTED. TRUSTED. [7]
>>
>>
>>
>> Links:
>> --

Re: [ovirt-users] Hosted-Engine Deploy - Error Creating a Storage Domain

2017-06-06 Thread Sandro Bonazzola
On Tue, Jun 6, 2017 at 4:24 AM, Langley, Robert 
wrote:

> FYI: My miss. Firewall port for VDSM needed to be added to my zone(s).
> Yay! The host is now in GREEN status within the Default Cluster.
>

Happy to see you solved the issue!




>
> Sent using OWA for iPhone
> --
> *From:* Langley, Robert
> *Sent:* Monday, June 5, 2017 5:16:51 PM
> *To:* Sandro Bonazzola; Sahina Bose
> *Cc:* Simone Tiraboschi; Nir Soffer; Allon Mureinik; Tal Nisan; users
> *Subject:* RE: [ovirt-users] Hosted-Engine Deploy - Error Creating a
> Storage Domain
>
>
> I have good news. With the IOProcess connection, I was suspecting maybe it
> would be something with the system (BIOS related).
>
> The Dell PE R730 was at BIOS version 2.2.5 and there was an update in
> April to version 2.4.3 (with another inbetween these two). There were a
> couple of fixes between the version the server had and the latest, which
> had me wondering if they may be related to the IOProcess issue I was
> experiencing.
>
> After applying this update the Hosted Engine deployment went further.
>
> It could explain why I could not add this host in a previous installation.
>
>
>
> I’ll start another thread for this next bump I’m running into, unless I
> can figure it out. Has to do with the VDSM host. Engine says it cannot
> communicate with this host its running on. So, the setup timed out waiting
> for the VDSM host to start.
>
>
>
> *From:* Sandro Bonazzola [mailto:sbona...@redhat.com]
> *Sent:* Monday, June 5, 2017 9:00 AM
> *To:* Langley, Robert ; Sahina Bose <
> sab...@redhat.com>
> *Cc:* Simone Tiraboschi ; Nir Soffer <
> nsof...@redhat.com>; Allon Mureinik ; Tal Nisan <
> tni...@redhat.com>; users 
> *Subject:* Re: [ovirt-users] Hosted-Engine Deploy - Error Creating a
> Storage Domain
>
>
>
>
>
>
>
> On Fri, Jun 2, 2017 at 6:16 PM, Langley, Robert <
> robert.lang...@ventura.org> wrote:
>
> Any progress?
>
>
>
> Nir? Allon? Tal?
>
>
>
>
>
> One thing that has been going through my mind is whether oVirt allows a
> GlusterFS storage domain to work with multiple CPU types?
>
>
>
> Sahina?
>
>
>
> The two dedicated GlusterFS storage servers are AMD Opteron. And the third
> server for the replica 3, which I am hoping I can also use as an oVirt host
> is a new Intel Xeon (Dell PE R830). I know GlusterFS allows for mixed
> hardware, but I’m not sure about when oVirt manages GlusterFS, if mixed
> hardware and mixed use has been accounted for?
>
>
>
> *From:* Simone Tiraboschi [mailto:stira...@redhat.com]
> *Sent:* Wednesday, May 31, 2017 8:41 AM
> *To:* Langley, Robert 
> *Cc:* Sandro Bonazzola ; Nir Soffer <
> nsof...@redhat.com>; Allon Mureinik ; Tal Nisan <
> tni...@redhat.com>; users 
>
>
> *Subject:* Re: [ovirt-users] Hosted-Engine Deploy - Error Creating a
> Storage Domain
>
>
>
> It seams something related to IOProcess connection.
>
>
>
> 2017-05-25 20:54:40,362-0700 INFO  (jsonrpc/4) [IOProcessClient] Starting
> client ioprocess-3 (__init__:330)
>
> 2017-05-25 20:54:40,370-0700 INFO  (ioprocess/31239) [IOProcess] Starting
> ioprocess (__init__:452)
>
> 2017-05-25 20:54:40,407-0700 ERROR (jsonrpc/4) [storage.TaskManager.Task]
> (Task='89dd17d2-8a38-4825-9ba2-f231f1aff9f5') Unexpected error (task:870)
>
> Traceback (most recent call last):
>
>   File "/usr/share/vdsm/storage/task.py", line 877, in _run
>
> return fn(*args, **kargs)
>
>   File "/usr/lib/python2.7/site-packages/vdsm/logUtils.py", line 52, in
> wrapper
>
> res = f(*args, **kwargs)
>
>   File "/usr/share/vdsm/storage/hsm.py", line 2581, in createStorageDomain
>
> storageType, domVersion)
>
>   File "/usr/share/vdsm/storage/nfsSD.py", line 87, in create
>
> remotePath, storageType, version)
>
>   File "/usr/share/vdsm/storage/fileSD.py", line 421, in _prepareMetadata
>
> procPool.fileUtils.createdir(metadataDir, 0o775)
>
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/outOfProcess.py",
> line 166, in createdir
>
> self._iop.mkdir(tmpPath, mode)
>
>   File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line
> 524, in mkdir
>
> self.timeout)
>
>   File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line
> 466, in _sendCommand
>
> raise OSError(errcode, errstr)
>
> OSError: [Errno 61] No data available
>
> 2017-05-25 20:54:40,409-0700 INFO  (jsonrpc/4) [storage.TaskManager.Task]
> (Task='89dd17d2-8a38-4825-9ba2-f231f1aff9f5') aborting: Task is aborted:
> u'[Errno 61] No data available' - code 100 (task:1175)
>
> 2017-05-25 20:54:40,409-0700 ERROR (jsonrpc/4) [storage.Dispatcher] [Errno
> 61] No data available (dispatcher:80)
>
> Traceback (most recent call last):
>
>   File "/usr/share/vdsm/storage/dispatcher.py", line 72, in wrapper
>
> result = ctask.prepare(func, *args, **kwargs)
>
>   File 

Re: [ovirt-users] unsuccessful hosted engine install

2017-06-06 Thread Simone Tiraboschi
On Tue, Jun 6, 2017 at 2:10 AM, Brendan Hartzell  wrote:

> As requested,
>

It seams fine, there are no pending locks now.
Could you please retry?


>
> The output of ovirt-hosted-engine-cleanup
>
> [root@node-1 ~]# ovirt-hosted-engine-cleanup
> This will de-configure the host to run ovirt-hosted-engine-setup from
> scratch.
> Caution, this operation should be used with care.
>
> Are you sure you want to proceed? [y/n]
> y
>  -=== Destroy hosted-engine VM ===-
> You must run deploy first
>  -=== Stop HA services ===-
>  -=== Shutdown sanlock ===-
> shutdown force 1 wait 0
> shutdown done 0
>  -=== Disconnecting the hosted-engine storage domain ===-
> You must run deploy first
>  -=== De-configure VDSM networks ===-
>  -=== Stop other services ===-
>  -=== De-configure external daemons ===-
>  -=== Removing configuration files ===-
> ? /etc/init/libvirtd.conf already missing
> - removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml
> ? /etc/ovirt-hosted-engine/answers.conf already missing
> ? /etc/ovirt-hosted-engine/hosted-engine.conf already missing
> - removing /etc/vdsm/vdsm.conf
> - removing /etc/pki/vdsm/certs/cacert.pem
> - removing /etc/pki/vdsm/certs/vdsmcert.pem
> - removing /etc/pki/vdsm/keys/vdsmkey.pem
> - removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem
> - removing /etc/pki/vdsm/libvirt-spice/ca-key.pem
> - removing /etc/pki/vdsm/libvirt-spice/server-cert.pem
> - removing /etc/pki/vdsm/libvirt-spice/server-key.pem
> ? /etc/pki/CA/cacert.pem already missing
> ? /etc/pki/libvirt/*.pem already missing
> ? /etc/pki/libvirt/private/*.pem already missing
> ? /etc/pki/ovirt-vmconsole/*.pem already missing
> - removing /var/cache/libvirt/qemu
> ? /var/run/ovirt-hosted-engine-ha/* already missing
> [root@node-1 ~]#
>
> Output of sanlock client status:
> [root@node-1 ~]# sanlock client status
> [root@node-1 ~]#
>
> Thank you for your help!
>
> On Mon, Jun 5, 2017 at 7:25 AM, Simone Tiraboschi 
> wrote:
>
>>
>>
>> On Mon, Jun 5, 2017 at 3:57 PM, Brendan Hartzell 
>> wrote:
>>
>>> After letting this sit for a few days, does anyone have any ideas as to
>>> how to deal with my situation?  Would anyone like me to send the SOS report
>>> directly to them?  It's a 9MB file.
>>>
>>> If nothing comes up, I'm going to try and sift through the SOS report
>>> tonight, but I won't know what I'm trying to find.
>>>
>>> Thank you for any and all help.
>>>
>>> On Thu, Jun 1, 2017 at 1:15 AM, Sandro Bonazzola 
>>> wrote:
>>>


 On Thu, Jun 1, 2017 at 6:36 AM, Brendan Hartzell 
 wrote:

> Ran the 4 commands listed above, no errors on the screen.
>
> Started the hosted-engine standard setup from the web-UI.
>
> Using iSCSI for the storage.
>
> Using mostly default options, I got these errors in the web-UI.
>
>  Error creating Volume Group: Failed to initialize physical device:
> ("[u'/dev/mapper/36589cfc00de7482638fcfce4']",)
> Failed to execute stage 'Misc configuration': Failed to initialize
> physical device: ("[u'/dev/mapper/36589cfc0
> 0de7482638fcfce4']",)
> Hosted Engine deployment failed: this system is not reliable, please
> check the issue,fix and redeploy
>
> I rebuilt my iSCSI (I don't think I cleaned it up from a previous
> install).
> Re-ran the above 4 commands.
> Restarted hosted engine standard setup from web-UI.
> Install moved past "Connecting Storage Pool" so I believe the above
> was my fault.
>
> These are the last messages displayed on the web-UI.
>  Creating Storage Pool
> Connecting Storage Pool
> Verifying sanlock lockspace initialization
> Creating Image for 'hosted-engine.lockspace' ...
> Image for 'hosted-engine.lockspace' created successfully
> Creating Image for 'hosted-engine.metadata' ...
> Image for 'hosted-engine.metadata' created successfully
> Creating VM Image
> Extracting disk image from OVF archive (could take a few minutes
> depending on archive size)
> Validating pre-allocated volume size
> Uploading volume to data domain (could take a few minutes depending on
> archive size)
>
> At the host terminal, I got the error "watchdog watchdog0: watchdog
> did not stop!"
> Then the host restarted.
>

 Simone, can you help here?


>> Ok, sorry for the delay.
>> The second installation attempt seams fine but it seams that
>> ovirt-hosted-engine-cleanup failed stopping sanlock and so the
>> watchdog kick in rebooting your system in the middle of deployment attempt.
>>
>> could you please post the output of
>>ovirt-hosted-engine-cleanup
>>sanlock client status
>> ?
>>
>>
>>
>>>


>
> This is as far as I've gotten in previous attempts.
>
> Attaching the hosted-engine-setup log.
>
> The SOS report is 9MB and the ovirt users group will drop 

Re: [ovirt-users] Seamless SAN HA failovers with oVirt?

2017-06-06 Thread Sven Achtelik
Hi Matthew, 

I'm also using a HA TrueNAS as the storage. I have NFS as well as iscsi shares 
and did do some in place upgrade. The failover went more or less smooth, it was 
more of an issue on the TrueNas side where the different vlans didn't come up. 
This caused the engine to take down the storage domain and things took some 
time until everything was up again. The VMs in ovirt did go into paused mode 
and started to work again as soon as the failover was done. I was failing over 
by rebooting one of the TrueNas nodes and this took some time for the other 
node to take over. I was thinking about asking the TN guys if there is a 
command or procedure to speed up the failover. In all I didn't stop any VMs 
although the VMs paused. Depending on the critically of the VMs you might want 
to move to another storage. 

Sven 

-Ursprüngliche Nachricht-
Von: users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] Im Auftrag von 
Matthew Trent
Gesendet: Montag, 5. Juni 2017 23:48
An: users 
Betreff: [ovirt-users] Seamless SAN HA failovers with oVirt?

I'm using two TrueNAS HA SANs (FreeBSD-based ZFS) to provide storage via NFS to 
7 oVirt boxes and about 25 VMs.

For SAN system upgrades I've always scheduled a maintenance window, shut down 
all the oVirt stuff, upgraded the SANs, and spun everything back up. It's 
pretty disruptive, but I assumed that was the thing to do.

However, in talking with the TrueNAS vendor they said the majority of their 
customers are using VMWare and they almost always do TrueNAS updates in 
production. They just upgrade one head of the TrueNAS HA pair then failover to 
the other head and upgrade it too. There's a 30-ish second pause in I/O while 
the disk arrays are taken over by the other HA head, but VMWare just tolerates 
it and continues without skipping a beat. They say this is standard procedure 
in the SAN world and virtualization systems should tolerate 30-60 seconds of 
I/O pause for HA failovers seamlessly.

It sounds great to me, but I wanted to pick this lists' brain -- is anyone 
doing this with oVirt? Are you able to failover your HA SAN with 30-60 seconds 
of no I/O without oVirt freaking out?

If not, are there any tunables relating to this? I see the default NFS mount 
options look fairly tolerant (proto=tcp,timeo=600,retrans=6), but are there 
VDSM or sanlock or some other oVirt timeouts that will kick in and start 
putting storage domains into error states, fencing hosts or something before 
that? I've never timed anything, but I want to say my past experience is that 
ovirt hosted engine started showing errors almost immediately when we've had 
SAN issues in the past.

Thanks!

--
Matthew Trent
Network Engineer
Lewis County IT Services
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Building oVirt engine on Debian

2017-06-06 Thread Yedidyah Bar David
On Tue, Jun 6, 2017 at 6:55 AM, Leni Kadali Mutungi
 wrote:
> Setup was successful. Attached is the message I received. I didn't
> mind the firewalld bits since I don't have that installed. However all
> the ovn-* commands didn't work. I tried locating their equivalents,
> thinking that they could be in the ovirt_engine folder or something
> along those lines. The `sed` and `keytool` commands worked. However I
> don't have OpenSwitch installed so I'll add that if it's necessary.

You can also ignore ovn for now - just reply 'No' when asked, and even
if you replied Yes I do not think you'll have many problems if you do
not actually use it.

So, did the engine also start successfully?

Best,

>
> --
> - Warm regards
> Leni Kadali Mutungi



-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users