Re: [ovirt-users] New user intro & some questions

2015-02-11 Thread Federico Simoncelli
- Original Message -
> From: "Sandro Bonazzola" 
> To: "Sven Kieske" , users@ovirt.org, "Federico 
> Simoncelli" 
> Sent: Tuesday, February 10, 2015 4:56:05 PM
> Subject: Re: [ovirt-users] New user intro & some questions
> 
> Il 10/02/2015 16:49, Sven Kieske ha scritto:
> > 
> > 
> > On 10/02/15 03:02, Jason Brooks wrote:
> >> The meaning of "support" is important here -- support from whom? It's
> >> true that there's no gluster+virt SKU of the RHEV downstream project.
> >> All configurations of ovirt proper are self-supported, or community-
> >> supported, and what we choose to support is up to us individuals.
> >>
> >> However, gluster + virt on the same nodes does work -- even w/ management
> >> through the engine. I do use gluster on my virt nodes, but I don't manage
> >> them w/ the engine, because, afaik, there isn't a way to have gluster and
> >> virt on separate networks this way, so I just manage gluster from the
> >> gluster cli.
> >>
> >> It's true, oVirt is happiest w/ separate machines for everything, and
> >> a rock-solid san of some sort, etc., but that's not the only way, and
> >> as you point out, hardware isn't free.
> > 
> > Well you might be interested in these upcoming features:
> > 
> > http://www.ovirt.org/Features/Self_Hosted_Engine_Hyper_Converged_Gluster_Support
> > 
> > sadly the slides from this talk are not online:
> > https://fosdem.org/2015/schedule/event/hyperconvergence/
> > 
> > maybe brian(cc'ed) can put them somewhere?
> 
> Or Federico (CCed) :-)

The slides of my sessions are available at:

http://www.ovirt.org/images/6/6c/2015-ovirt-glusterfs-hyperconvergence.pdf
http://www.ovirt.org/images/9/97/2015-docker-ovirt-iaas.pdf

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted engine: sending ioctl 5401 to a partition!

2014-11-27 Thread Federico Simoncelli
- Original Message -
> From: "Chris Adams" 
> To: users@ovirt.org
> Sent: Friday, November 21, 2014 10:28:28 PM
> Subject: [ovirt-users] Hosted engine: sending ioctl 5401 to a partition!
> 
> I have set up oVirt with hosted engine, on an iSCSI volume.  On both
> nodes, the kernel logs the following about every 10 seconds:
> 
> Nov 21 15:27:49 node8 kernel: ovirt-ha-broker: sending ioctl 5401 to a
> partition!
> 
> Is this a known bug, something that I need to address, etc.?

Is this on centos or fedora?

We may have to do some testing to identify where that's coming
from.

Feel free to ping me: fsimonce (#ovirt on OFTC) so we can check
what's going on.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Reclaim/Trim/issue_discards : how to inform my LUNs about disk space waste?

2014-11-27 Thread Federico Simoncelli
Hi Nicolas,
 you can find more information on this at:

https://bugzilla.redhat.com/show_bug.cgi?id=981626

First of all an important note (that was already mentioned): vdsm
is not using lvm.conf, so whatever change you make there it won't
affect vdsm behavior.

Anyway long story short, enabling issue_discards in lvm would lead
to lvm commands starvation when the lv that you're removing is large
and the granularity is small.

The correct solution is to use blkdiscard on the lv and I happened
to submit a patch series for that yesterday:

http://gerrit.ovirt.org/#/q/status:open+project:vdsm+branch:master+topic:block-discard,n,z
(very much experimental)

The approach is to begin with issuing blkdiscard when "wipe after
delete" is selected on the disk.

That is because blkdiscard in the majority of the cases will wipe
the lv data and I know that someone in the past has been brave enough
to try and recover data from a mistakenly removed lv that wasn't
post-zeroed.

Anyway extending the support to non post-zero is trivial and it's
just a matter of agreement and expectations.

With regard to the legitimate question of why both post-zero and
block discard, the answer is that after discussing it with storage
array experts it seems that blkdiscard has no contract in
guaranteeing that the data will be blanked out and it could later
on show up even on a completely different LUN of the same storage.

-- 
Federico

- Original Message -
> From: "Nicolas Ecarnot" 
> To: "Federico Simoncelli" 
> Cc: "users" 
> Sent: Thursday, November 27, 2014 10:43:06 AM
> Subject: Re: [ovirt-users] Reclaim/Trim/issue_discards : how to inform my 
> LUNs about disk space waste?
> 
> Le 22/07/2014 14:23, Federico Simoncelli a écrit :
> > - Original Message -
> >> From: "Nicolas Ecarnot" 
> >> To: "users" 
> >> Sent: Thursday, July 3, 2014 10:54:57 AM
> >> Subject: [ovirt-users] Reclaim/Trim/issue_discards : how to inform my LUNs
> >> about disk space waste?
> >>
> >> Hi,
> >>
> >> In my hosts, I see that /etc/lvm/lvm.conf shows :
> >>
> >> issue_discards = 0
> >>
> >> Can I enable it to 1 ?
> >>
> >> Thank you?
> >
> > You can change it but it's not going to affect the lvm behavior in VDSM
> > since we don't use the host lvm config file.
> 
> Frederico,
> 
> May you describe a little more how it's done, explain the principle, or
> point us to a place where we can learn more about how LVM is used in
> oVirt, amongst the manager and the hosts.
> 
> Thank you.
> 
> >
> > This will be probably addressed as part of bz1017284 as we're considering
> > to extend discard also to vdsm images (and not direct luns only).
> >
> > https://bugzilla.redhat.com/show_bug.cgi?id=1017284
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed Glance Image Import Task Cannot Be Removed

2014-11-26 Thread Federico Simoncelli
Hi Chris, can you share the link/code of the task removal script that
you used?

Thanks,
-- 
Federico

- Original Message -
> From: "Chris Routh" 
> To: users@ovirt.org
> Sent: Tuesday, November 25, 2014 2:25:16 PM
> Subject: Re: [ovirt-users] Failed Glance Image Import Task Cannot Be Removed
> 
> No ideas from anyone?
> 
> On Sun, 2014-11-23 at 08:13 -0500, Chris Routh wrote:
> > Greetings fellow mailing listers,
> > 
> > I've been using oVirt since September now, and using version 3.5
> > currently. When I first setup my small cluster I decided to import a
> > CentOS image from the glance image repository. However the first time
> > I tried it the import failed to complete. The problem is that oVirt
> > has not clued in that this import failed even 2 months and several
> > reboots of the entire cluster later. I ended up re-attempting the
> > import a day later (this is back in September) and it worked, however
> > this invalid task is still in my task list.
> > 
> > I tried using a few tricks from this mailing list when I encountered
> > this, one was to run a task removal script on the hosted-engine.
> > However while that removed other stale tasks like drive copying.. it
> > never seems to work for this import task. This task disappears when I
> > run the script, but reappears once the hosted-engine is restarted.
> > It's stuck.. somewhere. I had settled on giving up on cleaning up this
> > nuisance a month ago, and had planned to forget about it; until today.
> > 
> > Today I am upgrading my storage pool from a simple NFS share on an old
> > linux server to a proper NFS volume on a new FreeNAS device.
> > Everything was smooth until I moved everything off the old NFS share
> > and wanted to put it into maintenance mode so I could cleanly remove
> > it from the datacenter. I cannot remove it. Attempting to remove it
> > produces the error:
> > 
> > 
> > 
> > Operation Canceled
> > Error while executing action: Cannot deactivate Data Domain while
> > there are running tasks on this data domain.
> > -Please wait until tasks will finish and try again.
> > 
> > So now I am forced to find the solution to removing this task. I know
> > the task's ID is 765780ce but how do I rid myself of this ghost once
> > and for all? Also important to mention I have noticed there is a
> > 'Locked' disk associated with this task.
> > 
> > Regards,
> > 
> > 
> > Chris Routh
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> 
> 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Status libgfapi support in oVirt

2014-11-21 Thread Federico Simoncelli
- Original Message -
> From: "noc" 
> To: users@ovirt.org
> Sent: Friday, November 21, 2014 10:01:30 AM
> Subject: Re: [ovirt-users] Status libgfapi support in oVirt
> 
> On 21-11-2014 9:47, noc wrote:
> > The VM doesn't start but that can be caused by my L2 virt OR that the
> >  is wrong. Shouldn't there be a port in
> > the 24007 or 49152 range?
> >
> >
> Sorry, forgot to install vdsm-gluster. Starting the VM on an el6 host
> now works and still the same line with the port=0 so that doesn't seem
> to matter.
> 
> OK, back to setting up a el6 host except when you generate a el7 version
> too which would be awesome 8-) .

I updated the packages (rebasing on a newer master) and I provided an
el7 build as well:

https://fsimonce.fedorapeople.org/vdsm-libgfapi/

These rpms are less tested than the previous ones but the rebase was
straight forward.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Status libgfapi support in oVirt

2014-11-20 Thread Federico Simoncelli
- Original Message -
> From: "noc" 
> To: users@ovirt.org
> Sent: Thursday, November 20, 2014 8:46:01 AM
> Subject: Re: [ovirt-users] Status libgfapi support in oVirt
> 
> On 19-11-2014 23:44, Darrell Budic wrote:
> > Is there an el7 build of this available too?
> >
> That would be nice too. Forgot that I updated my test env to el7 to see
> if that helped. Can test on F20 @home tonight if needed. Gonna try
> something else first and will let you all know how that went.

I've prepared a fedora 20 build as well:

https://fsimonce.fedorapeople.org/vdsm-libgfapi/fc20/

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Status libgfapi support in oVirt

2014-11-19 Thread Federico Simoncelli
- Original Message -
> From: "noc" 
> To: users@ovirt.org
> Sent: Wednesday, November 19, 2014 9:36:28 AM
> Subject: Re: [ovirt-users] Status libgfapi support in oVirt
> 
> On 18-11-2014 20:57, Christopher Young wrote:
> 
> I'm replying to 'up' this as well as I'm most interested in this. I actually
> thought this was implemented and working too.
> 
> On Mon, Nov 17, 2014 at 10:01 AM, Daniel Helgenberger <
> daniel.helgenber...@m-box.de > wrote:
> 
> 
> Hello Joop,
> 
> thanks for raising the issue as it is one of the things I assumed are
> already implemented and working.
> 
> Sadly I cannot provide any answer ...
> 
> On 17.11.2014 09:39, Joop wrote:
> > I have been trying to use libgfapi glusterfs support in oVirt but can't
> > get it to work. After talks on IRC it seems I should apply a patch
> > ( http://gerrit.ovirt.org/33768 ) to enable libgf BUT I can't get it to
> > work. Systems used:
> > - hosts Centos7 or Fedora20 (so upto date qemu/libvirt/oVirt(3.5))
> > - glusterfs-3.6.1
> > - vdsm-4.16.0-524.gitbc618a4.el7.x86_64 (snapshot master 14-nov)
> > - vdsm-4.16.7-1.gitdb83943.el7.x86_64 (official ovirt-3.5 vdsm, seems
> > newer than master snapshot?? )
> > 
> > Just adding the patch to vdsm-4.16.7-1.gitdb83943.el7.x86_64 doesn't
> > work, vdsm doesn't start anymore due to an error in virt/vm.py.
> > 
> > Q1: what is de exact status of libgf and oVirt.
> > Q2: how do I test that patch?
> 
> I experimented a little more and found that if I create a VM in oVirt on a
> glusterfs storage domain and start it, it won't use libgfapi, BUT if I use
> virsh on the host where the VM runs and then add a disk the libgfapi way the
> VM will see the disk and can use it. So the underlying infra is capable of
> using libgf but oVirt isn't using it. Thats where the patch comes in I think
> but I can't get it to work.

Correct. oVirt up until now didn't use libgfapi because of missing features
(e.g. live snapshot). It seems that now all those gaps have been fixed and
we're trying to re-enable libgfapi.

I just mentioned that I uploaded an el6 build here:

https://fsimonce.fedorapeople.org/vdsm-libgfapi/

and sources here (to rebuild on fedora):

https://fsimonce.fedorapeople.org/vdsm-libgfapi/source/

Let me know if the most of you are using fedora and I'll make a build on
fedora as well.

Please let me know how it goes. Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Status libgfapi support in oVirt

2014-11-19 Thread Federico Simoncelli
- Original Message -
> From: "Joop" 
> To: users@ovirt.org
> Sent: Monday, November 17, 2014 9:39:36 AM
> Subject: [ovirt-users] Status libgfapi support in oVirt
> 
> I have been trying to use libgfapi glusterfs support in oVirt but can't
> get it to work. After talks on IRC it seems I should apply a patch
> (http://gerrit.ovirt.org/33768) to enable libgf BUT I can't get it to
> work. Systems used:
> - hosts Centos7 or Fedora20 (so upto date qemu/libvirt/oVirt(3.5))
> - glusterfs-3.6.1
> - vdsm-4.16.0-524.gitbc618a4.el7.x86_64 (snapshot master 14-nov)
> - vdsm-4.16.7-1.gitdb83943.el7.x86_64 (official ovirt-3.5 vdsm, seems
> newer than master snapshot?? )
> 
> Just adding the patch to vdsm-4.16.7-1.gitdb83943.el7.x86_64 doesn't
> work, vdsm doesn't start anymore due to an error in virt/vm.py.
> 
> Q1: what is de exact status of libgf and oVirt.
> Q2: how do I test that patch?

Rebasing and applying patches could be tricky sometimes and if you
got an error in virt/vm.py it is most likely because the patch
didn't apply cleanly.

I prepared a build (el6) here:

https://fsimonce.fedorapeople.org/vdsm-libgfapi/

In case you want to try it on fedora you just need to get the
source rpm here:

https://fsimonce.fedorapeople.org/vdsm-libgfapi/source/

and rebuild it on fedora.

Let me know if you have any problem.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [QE] oVirt 3.5.1 status

2014-10-26 Thread Federico Simoncelli
- Original Message -
> From: "Sven Kieske" 
> To: users@ovirt.org
> Sent: Thursday, October 23, 2014 6:42:11 PM
> Subject: Re: [ovirt-users] [QE] oVirt 3.5.1 status
> 
> -BEGIN PGP SIGNED MESSAGE-
> Hash: SHA1
> 
> Hi,
> 
> please consider:
> https://bugzilla.redhat.com/show_bug.cgi?id=1156115
> as a backport to 3.5.1
> 
> I don't know if you plan to release a new vdsm version
> though and I also don't know if this patch is already
> matured enough, if I can test or help make this patch
> better, please let me know, as I'm very interested in
> getting it into this release.

Hi Sven, first of all we need to merge the patch in master and yes,
as you suggested having some help would speed up the process.

The patch affects the move/copy of images with (one or more)
snapshots.

I already briefly tested the patch with qemu-img from rhel 6 so
we need to cover other platforms (fedora and centos) and test:

- cold/live move of disks from one storage to another (both
  nfs/iscsi and cold move from nfs to iscsi and backward)

- export vms (with snapshots) to export domain and re-import it
  (both nfs and iscsi)

Once the patch is in master and we have a feedback on how stable
it is we may consider it for backporting (maybe 3.5.2 or 3.5.3
if ever).

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)

2014-10-08 Thread Federico Simoncelli
- Original Message -
> From: "Ricardo Esteves" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Wednesday, October 8, 2014 1:32:51 AM
> Subject: Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> 
> Hi, here it goes:
> 
> ethtool -i eth3
> 
> driver: bnx2
> version: 2.2.4g
> firmware-version: bc 5.2.3
> bus-info: :06:00.1
> supports-statistics: yes
> supports-test: yes
> supports-eeprom-access: yes
> supports-register-dump: yes
> supports-priv-flags: no

Thanks, can you also add the output of:

 # lspci -nn

I'd have expected the driver to be bnx2i (bnx2 is a regular ethernet driver,
no offloading).

Can you also check if you have the bnx2i driver loaded?

 # lsmod | grep bnx2

and eventually if there's any bnx2 related message in /var/log/messages

Check also the adapter bios (at boot time) if by any chance you have to enable
the offloading there first (check the specific manual too).

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)

2014-10-07 Thread Federico Simoncelli
- Original Message -
> From: "Ricardo Esteves" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Tuesday, October 7, 2014 8:44:19 PM
> Subject: Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> 
> cat /var/lib/iscsi/ifaces/eth3
> # BEGIN RECORD 6.2.0-873.10.el6
> iface.iscsi_ifacename = eth3
> iface.transport_name = tcp
> iface.vlan_id = 0
> iface.vlan_priority = 0
> iface.iface_num = 0
> iface.mtu = 0
> iface.port = 0
> # END RECORD
> 
> Is there anyway to tell ovirt to use bnx2i instead of tcp?

Hi Ricardo, can you paste the output of:

 # ethtool -i eth3

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [ovirt-devel] Building vdsm within Fedora

2014-09-26 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Sandro Bonazzola" 
> Cc: crobi...@redhat.com, "users" , de...@ovirt.org
> Sent: Thursday, September 25, 2014 3:06:01 PM
> Subject: Re: [ovirt-devel] [ovirt-users] Building vdsm within Fedora
> 
> On Wed, Sep 24, 2014 at 10:57:21AM +0200, Sandro Bonazzola wrote:
> > Il 24/09/2014 09:44, Sven Kieske ha scritto:
> > > 
> > > 
> > > On 24/09/14 09:13, Federico Simoncelli wrote:
> > >> You probably missed the first part "we were using qemu-kvm/qemu-img in
> > >> the spec file". In that case you won't fail in any requirement.
> > >>
> > >> Basically the question is: was there any problem on centos6 before
> > >> committing http://gerrit.ovirt.org/31214 ?
> > 
> > Federico: as we checked a few minutes ago, it seems there's no problem in
> > requiring qemu-kvm/qemu-img in the spec file.
> > Only issue is that if non rhev version is installed a manual "yum update"
> > is required for moving to the rhevm version.
> 
> Right. Without the patch, RPM does not enforce qemu-kvm-rhev. So our
> code has to check for qemu-kvm-rhev functionality, instead of knowing
> that it is there.  Furthermore, we had several reports of users finding
> themselves without qemu-kvm-rhev on their node, and not understanding
> why they do not have live merge.

Live merge? The biggest problem with live merge is libvirt not qemu.

Anyway the qemu-kvm/qemu-kvm-rhev problem is relevant only for centos
and centos has a specific way to address these special needs:

http://www.centos.org/variants/

"""
A CentOS variant is a special edition of CentOS Linux that starts with
the core distribution, then replaces or supplements a specific subset of
packages. This may include replacing everything down to the kernel,
networking, and other subsystems.
"""

I think the plan was to have our own centos variant (shipping qemu-kvm-rhev).
I remember Doron participated to the centos meetings but I don't remember
the outcome.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [ovirt-devel] Building vdsm within Fedora

2014-09-24 Thread Federico Simoncelli
- Original Message -
> From: "Sandro Bonazzola" 
> To: "Federico Simoncelli" 
> Cc: de...@ovirt.org, "users" , "Sven Kieske" 
> 
> Sent: Wednesday, September 24, 2014 11:01:35 AM
> Subject: Re: [ovirt-devel] Building vdsm within Fedora
> 
> Il 24/09/2014 10:35, Federico Simoncelli ha scritto:
> > - Original Message -
> >> From: "Sven Kieske" 
> >> To: de...@ovirt.org, "users" 
> >> Sent: Wednesday, September 24, 2014 9:44:17 AM
> >> Subject: Re: [ovirt-devel] Building vdsm within Fedora
> >>
> >> On 24/09/14 09:13, Federico Simoncelli wrote:
> >>> You probably missed the first part "we were using qemu-kvm/qemu-img in
> >>> the spec file". In that case you won't fail in any requirement.
> >>>
> >>> Basically the question is: was there any problem on centos6 before
> >>> committing http://gerrit.ovirt.org/31214 ?
> >>
> >> Of course there was a problem, please follow the link in this very
> >> commit to the according bugzilla:
> >>
> >> https://bugzilla.redhat.com/show_bug.cgi?id=1127763
> >>
> >> In short: you can not use live snapshots without this updated spec file.
> >>
> >> And it's a PITA to install this package by hand, you must track
> >> it's versions yourself etc pp. you basically lose all the stuff
> >> a proper spec file gives you.
> > 
> > As soon as you have the ovirt repository installed there shouldn't be any
> > reason for you to have any of these problems.
> > 
> > Sandro, is there any reason why the rpm available here:
> > 
> > http://resources.ovirt.org/pub/ovirt-3.4/rpm/el6/x86_64/
> > 
> > are not published here?
> > 
> > http://resources.ovirt.org/releases/3.4/rpm/el6/x86_64/
> 
> this second link points to the previous layout, abandoned since we moved from
> /releases to /pub.
> /releases is still around for historical purpose, I think we should consider
> to drop it at some point avoinding confusion or renaming it to something
> that make it clear that it shouldn't be used anymore.

Sven can you let us know if you still have any problem using:

http://resources.ovirt.org/pub/yum-repo/ovirt-release34.rpm
(which should contain the correct ovirt.repo)

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [ovirt-devel] Building vdsm within Fedora

2014-09-24 Thread Federico Simoncelli
- Original Message -
> From: "Sven Kieske" 
> To: de...@ovirt.org, "users" 
> Sent: Wednesday, September 24, 2014 9:44:17 AM
> Subject: Re: [ovirt-devel] Building vdsm within Fedora
> 
> On 24/09/14 09:13, Federico Simoncelli wrote:
> > You probably missed the first part "we were using qemu-kvm/qemu-img in
> > the spec file". In that case you won't fail in any requirement.
> > 
> > Basically the question is: was there any problem on centos6 before
> > committing http://gerrit.ovirt.org/31214 ?
> 
> Of course there was a problem, please follow the link in this very
> commit to the according bugzilla:
> 
> https://bugzilla.redhat.com/show_bug.cgi?id=1127763
> 
> In short: you can not use live snapshots without this updated spec file.
> 
> And it's a PITA to install this package by hand, you must track
> it's versions yourself etc pp. you basically lose all the stuff
> a proper spec file gives you.

As soon as you have the ovirt repository installed there shouldn't be any
reason for you to have any of these problems.

Sandro, is there any reason why the rpm available here:

http://resources.ovirt.org/pub/ovirt-3.4/rpm/el6/x86_64/

are not published here?

http://resources.ovirt.org/releases/3.4/rpm/el6/x86_64/

Is there any additional repository (that provides qemu-*-rhev) that we are
missing from the ovirt.repo file?

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Building vdsm within Fedora

2014-09-24 Thread Federico Simoncelli
- Original Message -
> From: "Sandro Bonazzola" 
> To: "Federico Simoncelli" , "Dan Kenigsberg" 
> 
> Cc: de...@ovirt.org, dougsl...@redhat.com, "Sven Kieske" 
> , "users" 
> Sent: Wednesday, September 24, 2014 9:00:20 AM
> Subject: Re: Building vdsm within Fedora
> 
> Il 24/09/2014 08:53, Federico Simoncelli ha scritto:
> > - Original Message -
> >> From: "Dan Kenigsberg" 
> >> To: "Sandro Bonazzola" , de...@ovirt.org,
> >> fsimo...@redhat.com, dougsl...@redhat.com
> >> Cc: "Sven Kieske" , "users" 
> >> Sent: Tuesday, September 23, 2014 11:21:18 PM
> >> Subject: Building vdsm within Fedora
> >>
> >> Since Vdsm was open-sourced, it was built and deployed via
> >> Fedora.
> >>
> >> Recently [http://gerrit.ovirt.org/31214] vdsm introduced a spec-file
> >> dependency onf qemu-kvm-rhev, and considered to backport it to the
> >> ovirt-3.4 brach.
> >>
> >> Requiring qemu-kvm-rhev, which is not part of Fedora's EPEL6 branch,
> >> violates Fedora's standards.
> >>
> >> So basically we have two options:
> >>
> >> 1. Revert the qemu-kvm-rhev dependency.
> >> 2. Drop vdsm from EPEL6 (or completely from Fedora); ship Vdsm only
> >>within the oVirt repositories.
> >>
> >> A third option would be to have one rpm, with qemu-kvm-rhev, shipped in
> >> ovirt, and another without it - shipped in Fedora. I find this overly
> >> complex and confusing.
> > 
> > I think that until now (centos6) we were using qemu-kvm/qemu-img in the
> > spec file and then the ovirt repository was distributing qemu-*-rhev
> > from:
> > 
> >  http://resources.ovirt.org/pub/ovirt-3.4-snapshot/rpm/el6/x86_64/
> > 
> > It this not possible with centos7? Any problem with that?
> 
> We're shipping qemu-kvm-rhev on 3.4, 3.5 and master for EL6 and EL7.
> The issue is that if you don't enable ovirt, epel fails repository closure.

You probably missed the first part "we were using qemu-kvm/qemu-img in
the spec file". In that case you won't fail in any requirement.

Basically the question is: was there any problem on centos6 before
committing http://gerrit.ovirt.org/31214 ?

If not, why don't we do the same for centos7?

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Building vdsm within Fedora

2014-09-23 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Sandro Bonazzola" , de...@ovirt.org, 
> fsimo...@redhat.com, dougsl...@redhat.com
> Cc: "Sven Kieske" , "users" 
> Sent: Tuesday, September 23, 2014 11:21:18 PM
> Subject: Building vdsm within Fedora
> 
> Since Vdsm was open-sourced, it was built and deployed via
> Fedora.
> 
> Recently [http://gerrit.ovirt.org/31214] vdsm introduced a spec-file
> dependency onf qemu-kvm-rhev, and considered to backport it to the
> ovirt-3.4 brach.
> 
> Requiring qemu-kvm-rhev, which is not part of Fedora's EPEL6 branch,
> violates Fedora's standards.
> 
> So basically we have two options:
> 
> 1. Revert the qemu-kvm-rhev dependency.
> 2. Drop vdsm from EPEL6 (or completely from Fedora); ship Vdsm only
>within the oVirt repositories.
> 
> A third option would be to have one rpm, with qemu-kvm-rhev, shipped in
> ovirt, and another without it - shipped in Fedora. I find this overly
> complex and confusing.

I think that until now (centos6) we were using qemu-kvm/qemu-img in the
spec file and then the ovirt repository was distributing qemu-*-rhev
from:

 http://resources.ovirt.org/pub/ovirt-3.4-snapshot/rpm/el6/x86_64/

It this not possible with centos7? Any problem with that?

I find being in fedora a way to keep the spec file and the rpm updated
and as clean as possible.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fedora 21 Test Day

2014-09-23 Thread Federico Simoncelli
FYI, in a couple of days it will be the Fedora 21 virtualization
test day:

http://fedoramagazine.org/5tftw-2014-09-02/

https://fedoraproject.org/wiki/Test_Day:2014-09-25_Virtualization

it's a good opportunity for us to check for regressions with vdsm
(qemu, libvirt, virt-tools), get more attention from the fedora
community, have quicker feedback on issues and get them fixed before
release.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Problem Refreshing/Using ovirt-image-repository / 3.5 RC2

2014-09-23 Thread Federico Simoncelli
- Original Message -
> From: "Oved Ourfali" 
> To: j...@internetx.com, "Federico Simoncelli" 
> Cc: users@ovirt.org, "Allon Mureinik" 
> Sent: Tuesday, September 23, 2014 9:56:28 AM
> Subject: Re: [ovirt-users] Problem Refreshing/Using ovirt-image-repository / 
> 3.5  RC2
> 
> - Original Message -
> > From: "InterNetX - Juergen Gotteswinter" 
> > To: users@ovirt.org
> > Sent: Tuesday, September 23, 2014 10:41:41 AM
> > Subject: Re: [ovirt-users] Problem Refreshing/Using ovirt-image-repository
> > / 3.5   RC2
> > 
> > Am 23.09.2014 um 09:32 schrieb Oved Ourfali:
> > > 
> > > 
> > > - Original Message -
> > >> From: "InterNetX - Juergen Gotteswinter" 
> > >> To: users@ovirt.org
> > >> Sent: Tuesday, September 23, 2014 10:29:07 AM
> > >> Subject: [ovirt-users] Problem Refreshing/Using ovirt-image-repository /
> > >> 3.5  RC2
> > >>
> > >> Hi,
> > >>
> > >> when trying to refresh the ovirt glance repository i get a 500 Error
> > >> Message
> > >>
> > >>
> > >> Operation Canceled
> > >>
> > >> Error while executing action: A Request to the Server failed with the
> > >> following Status Code: 500
> > >>
> > >>
> > >> engine.log says:
> > >>
> > >> 2014-09-23 09:23:08,960 INFO
> > >> [org.ovirt.engine.core.bll.provider.TestProviderConnectivityCommand]
> > >> (ajp--127.0.0.1-8702-10) [7fffb4bd] Running command:
> > >> TestProviderConnectivityCommand internal: false. Entities affected :
> > >> ID: aaa0----123456789aaa Type: SystemAction group
> > >> CREATE_STORAGE_POOL with role type ADMIN
> > >> 2014-09-23 09:23:08,975 INFO
> > >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > >> (ajp--127.0.0.1-8702-10) [7fffb4bd] Correlation ID: 7fffb4bd, Call
> > >> Stack: null, Custom Event ID: -1, Message: Unrecognized audit log type
> > >> has been used.
> > >> 2014-09-23 09:23:20,173 INFO
> > >> [org.ovirt.engine.core.bll.aaa.LogoutUserCommand]
> > >> (ajp--127.0.0.1-8702-11) [712895c3] Running command: LogoutUserCommand
> > >> internal: false.
> > >> 2014-09-23 09:23:20,184 INFO
> > >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> > >> (ajp--127.0.0.1-8702-11) [712895c3] Correlation ID: 712895c3, Call
> > >> Stack: null, Custom Event ID: -1, Message: User admin logged out.
> > >> 2014-09-23 09:23:20,262 INFO
> > >> [org.ovirt.engine.core.bll.aaa.LoginAdminUserCommand]
> > >> (ajp--127.0.0.1-8702-6) Running command: LoginAdminUserCommand internal:
> > >> false.
> > >>
> > > 
> > > All these message are good... no error here.
> > > Can you attach the full engine log?
> > > 
> > 
> > imho there is nothing else related to this :/ i attached the log
> > starting from today. except firing up a test vm nothing else happened
> > yet (and several tries refreshing the image repo)
> > 
> 
> I don't see a refresh attempt in the log, but i'm not familiar enough with
> that.
> Federico - can you have a look?

I don't see any reference to glance or error 500 in the logs. My impression
is that the error 500 is between the ui and the engine... have you tried to
force-refresh the ovirt webadmin page?

You can try and use the rest-api to check if the listing is working there.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Issues with Storage domain

2014-09-09 Thread Federico Simoncelli
- Original Message -
> From: "Carlos A. Diaz" 
> To: "Nir Soffer" 
> Cc: users@ovirt.org
> Sent: Monday, September 8, 2014 7:27:20 PM
> Subject: Re: [ovirt-users] Issues with Storage domain
> 
> This is the content of:
> 
> [root@accda05 dom_md]# cat metadata
> CLASS=Data
> DESCRIPTION=Prod
> IOOPTIMEOUTSEC=1
> LEASERETRIES=3
> LEASETIMESEC=5
> LOCKPOLICY=
> LOCKRENEWALINTERVALSEC=5
> MASTER_VERSION=0
> POOL_UUID=a705b31a-1a6e-4207-bcd8-abf95b0a586c
> REMOTE_PATH=10.244.44.200:/exports/ovirt
> ROLE=Regular
> SDUUID=ef295295-163b-4c67-b5c1-bf51c86ce59d
> TYPE=NFS
> VERSION=3
> _SHA_CKSUM=2d911132dac3ad18997eca42a1c495c1e1a34584


_SHA_CKSUM is valid but you can try to remove it (you'll just get warnings
about the missing checksum).

I expect vdsm to try and mount your domain on:

 /rhev/data-center/mnt/10.244.44.200:_exports_ovirt

If that's successful please report the output of these commands:

 # mount
 # ls -lR /rhev/data-center/mnt/10.244.44.200:_exports_ovirt


If instead the share is not mounted please report the mount error that you
find in the vdsm logs.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Can I use qcow2?

2014-09-03 Thread Federico Simoncelli
- Original Message -
> From: "Itamar Heim" 
> To: "Demeter Tibor" 
> Cc: users@ovirt.org, "Allon Mureinik" , "Federico 
> Simoncelli" 
> Sent: Wednesday, September 3, 2014 12:50:30 PM
> Subject: Re: [ovirt-users] Can I use qcow2?
> 
> On 09/03/2014 11:14 AM, Demeter Tibor wrote:
> > On shared glusterfs.
> >
> 
> Allon/Federico - I remember on NFS, qcow2 isn't used by default, since
> raw is sparse by default.
> (but i don't remember if it "won't work", or just not enabled by default).
> 
> can one create a qcow2 disk for a VM with gluster storage?

Yes through rest-api. From the ovirt-shell you could run:

 $ add disk \
   --vm-identifier  \
   --provisioned_size  \
   --interface virtio \
   --name  \
   --format cow \
   --sparse true \
   --storage_domains-storage_domain 
storage_domain.name=

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] feature review - ReportGuestDisksLogicalDeviceName

2014-09-02 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Liron Aravot" 
> Cc: users@ovirt.org, de...@ovirt.org, smizr...@redhat.com, 
> fsimo...@redhat.com, "Michal Skrivanek"
> , "Vinzenz Feenstra" , "Allon 
> Mureinik" 
> Sent: Monday, September 1, 2014 11:23:45 PM
> Subject: Re: feature review - ReportGuestDisksLogicalDeviceName
> 
> On Sun, Aug 31, 2014 at 07:20:04AM -0400, Liron Aravot wrote:
> > Feel free to review the the following feature.
> > 
> > http://www.ovirt.org/Features/ReportGuestDisksLogicalDeviceName
> 
> Thanks for posting this feature page. Two things worry me about this
> feature. The first is timing. It is not reasonable to suggest an API
> change, and expect it to get to ovirt-3.5.0. We are two late anyway.
> 
> The other one is the suggested API. You suggest placing volatile and
> optional infomation in getVMList. It won't be the first time that we
> have it (guestIPs, guestFQDN, clientIP, and displayIP are there) but
> it's foreign to the notion of "conf" reported by getVMList() - the set
> of parameters needed to recreate the VM.

At first sight this seems something belonging to getVmStats (which
is reporting already other guest agent information).

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] SPM in oVirt 3.6

2014-08-05 Thread Federico Simoncelli
- Original Message -
> From: "Nir Soffer" 
> To: "Daniel Helgenberger" 
> Cc: users@ovirt.org, "Federico Simoncelli" 
> Sent: Monday, July 28, 2014 6:43:30 PM
> Subject: Re: [ovirt-users] SPM in oVirt 3.6
> 
> - Original Message -
> > From: "Daniel Helgenberger" 
> > To: users@ovirt.org
> > Sent: Friday, July 25, 2014 7:51:33 PM
> > Subject: [ovirt-users] SPM in oVirt 3.6
> > 
> > just out of pure curiosity: In a BZ [1] Allon mentions SPM will go away
> > in ovirt 3.6.
> > 
> > This seems like a major change for me. I assume this will replace
> > sanlock as well? What will SPM be replaced with?
> 
> No, sanlock is not going anywhere.
> 
> The change is that we will not have an SPM node, but any node that need to
> make meta data changes, will take a lock using sanlock while it make the
> changes.
> 
> Federico: can you describe in more details how it is going to work?

Most of the information can be found on the feature page:

http://www.ovirt.org/Features/Decommission_Master_Domain_and_SPM

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] 答复: Error after changing IP of Node (FQDN is still the same)

2014-08-05 Thread Federico Simoncelli
What's the version of the vdsm and sanlock packages?

Can you please share the logs on the host side? We need vdsm.log
and sanlock.log containing the relevant errors (Cannot acquire host
id).

Thanks,
-- 
Federico


- Original Message -
> From: "ml ml" 
> To: d...@redhat.com
> Cc: "users@ovirt.org" 
> Sent: Sunday, August 3, 2014 8:57:18 PM
> Subject: Re: [ovirt-users]答复: Error after changing IP of Node (FQDN is 
> still the same)
> 
> ok, i now removed the nodes and added them again. Same FQDN. I still get
> "Start SPM Task failed - result: cleanSuccess, message: VDSGenericException:
> VDSErrorException: Failed to HSMGetTaskStatusVDS, error = Cannot acquire
> host id, code = 661"
> 
> I also got some error.
> 
> Now whats the deal with that host id?
> 
> Can somone please point me to the way how to debug this instead of pressing
> some remove and add buttons?
> 
> Does someone really know how ovirt works under the hood?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Reclaim/Trim/issue_discards : how to inform my LUNs about disk space waste?

2014-07-22 Thread Federico Simoncelli
- Original Message -
> From: "Nicolas Ecarnot" 
> To: "users" 
> Sent: Thursday, July 3, 2014 10:54:57 AM
> Subject: [ovirt-users] Reclaim/Trim/issue_discards : how to inform my LUNs 
> about disk space waste?
> 
> Hi,
> 
> In my hosts, I see that /etc/lvm/lvm.conf shows :
> 
> issue_discards = 0
> 
> Can I enable it to 1 ?
> 
> Thank you?

You can change it but it's not going to affect the lvm behavior in VDSM
since we don't use the host lvm config file.

This will be probably addressed as part of bz1017284 as we're considering
to extend discard also to vdsm images (and not direct luns only).

https://bugzilla.redhat.com/show_bug.cgi?id=1017284

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Can HA Agent control NFS Mount?

2014-06-18 Thread Federico Simoncelli
- Original Message -
> From: "Bob Doolittle" 
> To: "Doron Fediuck" , "Andrew Lau" 
> 
> Cc: "users" , "Federico Simoncelli" 
> Sent: Saturday, June 14, 2014 1:29:54 AM
> Subject: Re: [ovirt-users] Can HA Agent control NFS Mount?
>
> 
> But there may be more going on. Even if I stop vdsmd, the HA services,
> and libvirtd, and sleep 60 seconds, I still see a lock held on the
> Engine VM storage:
> 
> daemon 6f3af037-d05e-4ad8-a53c-61627e0c2464.xion2.smar
> p -1 helper
> p -1 listener
> p -1 status
> s 
> 003510e8-966a-47e6-a5eb-3b5c8a6070a9:1:/rhev/data-center/mnt/xion2.smartcity.net\:_export_VM__NewDataDomain/003510e8-966a-47e6-a5eb-3b5c8a6070a9/dom_md/ids:0
> s 
> hosted-engine:1:/rhev/data-center/mnt/xion2\:_export_vm_he1/18eeab54-e482-497f-b096-11f8a43f94f4/ha_agent/hosted-engine.lockspace:0

This output shows that the lockspaces are still acquired. When you put 
hosted-engine
in maintenance they must be released.
One by directly using rem_lockspace (since it's the hosted-engine one) and the 
other
one by stopMonitoringDomain.

I quickly looked at the ovirt-hosted-engine* projects and I haven't found 
anything
related to that.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] sanlock + gluster recovery -- RFE

2014-05-21 Thread Federico Simoncelli
- Original Message -
> From: "Giuseppe Ragusa" 
> To: fsimo...@redhat.com
> Cc: users@ovirt.org
> Sent: Wednesday, May 21, 2014 5:15:30 PM
> Subject: sanlock + gluster recovery -- RFE
> 
> Hi,
> 
> > - Original Message -
> > > From: "Ted Miller" 
> > > To: "users" 
> > > Sent: Tuesday, May 20, 2014 11:31:42 PM
> > > Subject: [ovirt-users] sanlock + gluster recovery -- RFE
> > > 
> > > As you are aware, there is an ongoing split-brain problem with running
> > > sanlock on replicated gluster storage. Personally, I believe that this is
> > > the 5th time that I have been bitten by this sanlock+gluster problem.
> > > 
> > > I believe that the following are true (if not, my entire request is
> > > probably
> > > off base).
> > > 
> > > 
> > > * ovirt uses sanlock in such a way that when the sanlock storage is
> > > on a
> > > replicated gluster file system, very small storage disruptions can
> > > result in a gluster split-brain on the sanlock space
> > 
> > Although this is possible (at the moment) we are working hard to avoid it.
> > The hardest part here is to ensure that the gluster volume is properly
> > configured.
> > 
> > The suggested configuration for a volume to be used with ovirt is:
> > 
> > Volume Name: (...)
> > Type: Replicate
> > Volume ID: (...)
> > Status: Started
> > Number of Bricks: 1 x 3 = 3
> > Transport-type: tcp
> > Bricks:
> > (...three bricks...)
> > Options Reconfigured:
> > network.ping-timeout: 10
> > cluster.quorum-type: auto
> > 
> > The two options ping-timeout and quorum-type are really important.
> > 
> > You would also need a build where this bug is fixed in order to avoid any
> > chance of a split-brain:
> > 
> > https://bugzilla.redhat.com/show_bug.cgi?id=1066996
> 
> It seems that the aforementioned bug is peculiar to 3-bricks setups.
> 
> I understand that a 3-bricks setup can allow proper quorum formation without
> resorting to "first-configured-brick-has-more-weight" convention used with
> only 2 bricks and quorum "auto" (which makes one node "special", so not
> properly any-single-fault tolerant).

Correct.

> But, since we are on ovirt-users, is there a similar suggested configuration
> for a 2-hosts setup oVirt+GlusterFS with oVirt-side power management
> properly configured and tested-working?
> I mean a configuration where "any" host can go south and oVirt (through the
> other one) fences it (forcibly powering it off with confirmation from IPMI
> or similar) then restarts HA-marked vms that were running there, all the
> while keeping the underlying GlusterFS-based storage domains responsive and
> readable/writeable (maybe apart from a lapse between detected other-node
> unresposiveness and confirmed fencing)?

We already had a discussion with gluster asking if it was possible to
add fencing to the replica 2 quorum/consistency mechanism.

The idea is that as soon as you can't replicate a write you have to
freeze all IO until either the connection is re-established or you
know that the other host has been killed.

Adding Vijay.
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] sanlock + gluster recovery -- RFE

2014-05-21 Thread Federico Simoncelli
- Original Message -
> From: "Ted Miller" 
> To: "users" 
> Sent: Tuesday, May 20, 2014 11:31:42 PM
> Subject: [ovirt-users] sanlock + gluster recovery -- RFE
> 
> As you are aware, there is an ongoing split-brain problem with running
> sanlock on replicated gluster storage. Personally, I believe that this is
> the 5th time that I have been bitten by this sanlock+gluster problem.
> 
> I believe that the following are true (if not, my entire request is probably
> off base).
> 
> 
> * ovirt uses sanlock in such a way that when the sanlock storage is on a
> replicated gluster file system, very small storage disruptions can
> result in a gluster split-brain on the sanlock space

Although this is possible (at the moment) we are working hard to avoid it.
The hardest part here is to ensure that the gluster volume is properly
configured.

The suggested configuration for a volume to be used with ovirt is:

Volume Name: (...)
Type: Replicate
Volume ID: (...)
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
(...three bricks...)
Options Reconfigured:
network.ping-timeout: 10
cluster.quorum-type: auto

The two options ping-timeout and quorum-type are really important.

You would also need a build where this bug is fixed in order to avoid any
chance of a split-brain:

https://bugzilla.redhat.com/show_bug.cgi?id=1066996

> How did I get into this mess?
> 
> ...
> 
> What I would like to see in ovirt to help me (and others like me). Alternates
> listed in order from most desirable (automatic) to least desirable (set of
> commands to type, with lots of variables to figure out).

The real solution is to avoid the split-brain altogether. At the moment it
seems that using the suggested configurations and the bug fix we shouldn't
hit a split-brain.

> 1. automagic recovery
> 
> 2. recovery subcommand
> 
> 3. script
> 
> 4. commands

I think that the commands to resolve a split-brain should be documented.
I just started a page here:

http://www.ovirt.org/Gluster_Storage_Domain_Reference

Could you add your documentation there? Thanks!

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)

2014-05-14 Thread Federico Simoncelli
- Original Message -
> From: "Federico Simoncelli" 
> To: "Ricardo Esteves" 
> Cc: users@ovirt.org
> Sent: Wednesday, May 14, 2014 3:47:58 PM
> Subject: Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> 
> - Original Message -----
> > From: "Ricardo Esteves" 
> > To: "Federico Simoncelli" 
> > Cc: users@ovirt.org
> > Sent: Wednesday, May 14, 2014 1:45:53 AM
> > Subject: RE: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> > 
> > In attachment follows the defaults and the modified versions of the nodes
> > files.
> 
> If selecting the relevant host in "Hosts" you see the bnx2i interface in the
> "Network Interfaces" subtab, then you can try to:

Sorry I just noticed that you mentioned in the subject that you're using
oVirt 3.2. What I suggested is available only since oVirt 3.4.

-- 
Federico

> 1. create a new network in the "Network" tab, "VM network" checkbox should
>be disabled
> 2. select the relevant host in "Hosts" and use "Setup Host Networks" in the
>"Network Interfaces" subtab
> 3. configure the bnx2i interface and assign it to the new network you just
>created
> 4. in "iSCSI Multipathing" subtab of tab "Data Center" add a new entry where
>you bind the iscsi connection to the new network you created
> 
> Ping me on IRC if you need more help. My nick is fsimonce on #ovirt
> 
> --
> Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)

2014-05-14 Thread Federico Simoncelli
- Original Message -
> From: "Ricardo Esteves" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Wednesday, May 14, 2014 1:45:53 AM
> Subject: RE: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> 
> In attachment follows the defaults and the modified versions of the nodes
> files.

If selecting the relevant host in "Hosts" you see the bnx2i interface in the
"Network Interfaces" subtab, then you can try to:

1. create a new network in the "Network" tab, "VM network" checkbox should
   be disabled
2. select the relevant host in "Hosts" and use "Setup Host Networks" in the
   "Network Interfaces" subtab
3. configure the bnx2i interface and assign it to the new network you just
   created
4. in "iSCSI Multipathing" subtab of tab "Data Center" add a new entry where
   you bind the iscsi connection to the new network you created

Ping me on IRC if you need more help. My nick is fsimonce on #ovirt

-- 
Federico

> -Original Message-
> From: Federico Simoncelli [mailto:fsimo...@redhat.com]
> Sent: terça-feira, 13 de Maio de 2014 08:58
> To: Ricardo Esteves
> Cc: users@ovirt.org
> Subject: Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> 
> - Original Message -
> > From: "Ricardo Esteves" 
> > To: users@ovirt.org
> > Sent: Friday, April 11, 2014 1:07:31 AM
> > Subject: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> > 
> > Hi,
> > 
> > I've put my host on maintenance, then i configured iscsi offload for
> > my broadcom cards changing target's file's (192.168.12.2,3260 and
> > 192.168.12.4,3260) in my node
> > iqn.1986-03.com.hp:storage.msa2324i.1226151a6 to use interface
> > bnx2i.d8:d3:85:67:e3:bb, but after activating the host, configurations
> > are back to default.
> 
> Can you share the changes you made? Thanks.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Snapshot removal

2014-05-13 Thread Federico Simoncelli
- Original Message -
> From: "Dafna Ron" 
> To: users@ovirt.org
> Sent: Thursday, April 17, 2014 12:52:04 PM
> Subject: Re: [ovirt-users] Snapshot removal
> 
> NFS is wipe_after_delete=true always.
> so delete of snapshot will merge the data to upper level image + zero in
> on the data which is why this is taking a long time.

I double checked this and I was very much surprised by this finding!

We need a bz right away, Dafna can file it? Thanks!

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)

2014-05-13 Thread Federico Simoncelli
- Original Message -
> From: "Ricardo Esteves" 
> To: users@ovirt.org
> Sent: Friday, April 11, 2014 1:07:31 AM
> Subject: [ovirt-users] oVirt 3.2 - iSCSI offload (broadcom - bnx2i)
> 
> Hi,
> 
> I've put my host on maintenance, then i configured iscsi offload for my
> broadcom cards changing target's file's (192.168.12.2,3260 and
> 192.168.12.4,3260) in my node
> iqn.1986-03.com.hp:storage.msa2324i.1226151a6 to use interface
> bnx2i.d8:d3:85:67:e3:bb, but after activating the host, configurations
> are back to default.

Can you share the changes you made? Thanks.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Sanlock log entry

2014-05-12 Thread Federico Simoncelli
- Original Message -
> From: "Maurice James" 
> To: users@ovirt.org
> Sent: Saturday, April 12, 2014 3:06:37 PM
> Subject: [ovirt-users] Sanlock log entry
> 
> I'm seeing this about every 20 seconds in my sanlock.log.What does it mean?
> 
> 
> s1:r1694 resource
> a033c2ac-0d01-490c-9552-99ca53d6a64a:SDM:/rhev/data-center/mnt/ashtivh02.suprtekstic.com:_var_lib_exports_storage/a033c2ac-0d01-490c-9552-99ca53d6a64a/dom_md/leases:1048576
> for 3,14,3960

Can you provide more info? Anything weird in the VDSM logs? (errors/warnings)
Anything in the engine logs?

That message may be related to acquiring the SPM role but it shouldn't happen
so often.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Error creating Disks

2014-04-16 Thread Federico Simoncelli
- Original Message -
> From: "Maurice James" 
> To: d...@redhat.com
> Cc: users@ovirt.org
> Sent: Tuesday, April 15, 2014 6:54:11 PM
> Subject: Re: [ovirt-users] Error creating Disks
> 
> Logs are attached.
> 
> Live Migration failed

In the engine logs I see:

2014-04-15 12:51:07,420 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.SnapshotVDSCommand] 
(org.ovirt.thread.pool-6-thread-42) START, SnapshotVDSCommand(HostName = 
vhost3, HostId = bc9c25e6-714e-4eac-8af0-860ac76fd195, 
vmId=ba49605b-fb7e-4a70-a380-6286d3903e50), log id: e953f9a
...
2014-04-15 12:51:07,496 ERROR 
[org.ovirt.engine.core.vdsbroker.vdsbroker.SnapshotVDSCommand] 
(org.ovirt.thread.pool-6-thread-42) Command SnapshotVDSCommand(HostName = 
vhost3, HostId = bc9c25e6-714e-4eac-8af0-860ac76fd195, 
vmId=ba49605b-fb7e-4a70-a380-6286d3903e50) execution failed. Exception: 
VDSErrorException: VDSGenericException: VDSErrorException: Failed to 
SnapshotVDS, error = Snapshot failed, code = 48

but I can't find anything related to that in the vdsm log. Are you sure
you attached the correct vdsm log?

Are the hosts running centos or fedora?

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt 3.5 planning

2014-03-20 Thread Federico Simoncelli
- Original Message -
> From: "Itamar Heim" 
> To: "Federico Alberto Sayd" , users@ovirt.org, "Federico 
> Simoncelli" 
> Sent: Thursday, March 20, 2014 6:47:14 PM
> Subject: Re: [Users] oVirt 3.5 planning
> 
> On 03/20/2014 07:30 PM, Federico Alberto Sayd wrote:
> > 3 - Another question, when you convert a VM to template I see that it is
> > created with preallocated disk even if the original VM had thinly
> > provisioned disk. Is there no way to make the template with the same
> > type of disk (thinly provisioned)??
> 
> on NFS - it doesn't matter.
> on block storage - i don't remember why. maybe federico remembers.

It was for performance reasons since on block it would result in a qcow2
image. We wanted the access to the template to be as fast as possible.

The same question was brought up few months ago when we discussed importing
images from glance as templates. Anyway in that case we decided (for simplicity)
to allow the import as template also for qcow2 images.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] GSoC 14 Idea Discussion - virt-sparsify integration

2014-03-19 Thread Federico Simoncelli
- Original Message -
> From: "Utkarsh Singh" 
> To: users@ovirt.org
> Cc: fsimo...@redhat.com
> Sent: Friday, March 7, 2014 6:16:46 PM
> Subject: GSoC 14 Idea Discussion - virt-sparsify integration
> 
> Hello,
> 
> I am Utkarsh, a 4th year undergrad from IIT Delhi and a GSoC-14
> aspirant. I have been lately involved in an ongoing project Baadal
> Cloud Computing Platform in my institute, which has got me interested
> in oVirt for a potential GSoC project.
> 
> I was going through the virt-sparsify integration project idea. I have
> gone through the architecture documentation on the oVirt website. As
> far as I understand, the virt-sparsify integration needs to be done on
> the VDSM daemon, and it's control is either going to be completely
> independent of ovirt-engine (for example running it once every 24
> hours), or it's something that is centrally controlled by the
> ovirt-engine through XML/RPC calls. The details are not specified in
> the project ideas page. I would like to ask -

The request to sparsifying the image is controlled by ovirt-engine.
The user will pick one (or eventually more) disk(s) that are not in use
(vm down) and he'll request to sparsify it/them.

> 1. What would be the proposed ideal implementation? (Central-Control
> or Independent-Control)

Central-Control

> 2. Is virt-sparsify usage going to be automated or
> administrator-triggered, or a combination of both?

administrator-triggered

> There are some aspects of the idea, which I would like to discuss
> before I start working on a proposal.
> 
> It's not necessary that an automated usage of virt-sparsify is limited
> to any simple idea. Architecture documentation states that
> ovirt-engine has features like Monitoring that would allow
> administrators (and possibly users) to be aware of vm-guest
> performance as well as vm-host performance. I am not very sure about
> how this data is collected, Is it done through MoM, or Is this
> directly done by VDSM, or is someone else doing this (for hosts). It
> would be great if someone can explain that to me. This information
> about vm-guest usage and vm-host health can help in determining how
> virt-sparsify is to be used.

The vm/hosts statistics are gathered and provided by VDSM.
Anyway I would leave this part out at the moment. The virt-sparsify
command is a long running task and in the current architecture it
can be only an SPM task.
There is some ongoing work to remove the pool and the SPM (making
virt-sparsify operable by any host) but I wouldn't block on that.

> I am also not very clear about the Shared Storage component in the
> architecture. Does oVirt make any assumptions about the Shared
> Storage. For example, the performance difference between running
> virt-sparsify on NFS as compared to running it (if possible) directly
> on storage hardware. If the Storage solution is necessarily a NAS
> instance, then virt-sparsify on NFS mount is the only option.

The storage connections are already managed by vdsm and the volume
chains are maintained transparently in /rhev/data-center/...
There are few differences between image files on NFS/Gluster and images
stored on LVs but with regard to how to reach them it is transparent
(similar path).

> Right now, I am in the process of setting up oVirt on my system, and
> getting more familiar with the architecture. Regarding my experience.
> I am acquainted with both Java and Python. I have little experience
> with JBoss, but I have worked on some other Web Application Servers
> like web2py and Play Framework. My involvement in Baadal Platform has
> got me acquainted with libvirt/QEMU, the details of which I have
> mentioned below (if anyone is interested).

Depending on the amount of time that you can dedicate to this project
it seems that you could tackle both the vdsm and ovirt-engine parts.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] "Volume Group does not exist". Blame device-mapper ?

2014-02-17 Thread Federico Simoncelli
- Original Message -
> From: "Nicolas Ecarnot" 
> To: "Federico Simoncelli" 
> Cc: "users" 
> Sent: Monday, February 17, 2014 10:14:56 AM
> Subject: Re: [Users] "Volume Group does not exist". Blame device-mapper ?
> 
> Le 14/02/2014 15:39, Federico Simoncelli a écrit :
> > Hi Nicolas,
> >   are you still able to reproduce this issue? Are you using fedora or
> > centos?
> >
> > If providing the logs is problematic for you could you try to ping me
> > on irc (fsimonce on #ovirt OFTC) so that we can work on the issue
> > together?
> >
> > Thanks,
> >
> 
> Hi Frederico,
> 
> Since I haven't changed anything related to the SAN or the network, I'm
> pretty sure I'll be able to reproduce the bug.
> We are using CentOS.
> I can provide the logs, no issue.
> 
> This week, our oVirt setup will be strongly used, so this is not the
> better time to play with it. I'm very thankful you took the time to
> answer, but may I delay my answer about this bug to next week?

Ok, no problem. Feel free to contact me on IRC when you start testing.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] "Volume Group does not exist". Blame device-mapper ?

2014-02-14 Thread Federico Simoncelli
Hi Nicolas,
 are you still able to reproduce this issue? Are you using fedora or
centos?

If providing the logs is problematic for you could you try to ping me
on irc (fsimonce on #ovirt OFTC) so that we can work on the issue
together?

Thanks,
-- 
Federico

- Original Message -
> From: "Nicolas Ecarnot" 
> To: "users" 
> Sent: Monday, January 20, 2014 11:06:21 AM
> Subject: [Users] "Volume Group does not exist". Blame device-mapper ?
> 
> Hi,
> 
> oVirt 3.3, no big issue since the recent snapshot joke, but all in all
> running fine.
> 
> All my VM are stored in a iSCSI SAN. The VM usually are using only one
> or two disks (1: system, 2: data) and it is OK.
> 
> Friday, I created a new LUN. Inside a VM, I linked to it via iscsiadm
> and successfully login to the Lun (session, automatic attach on boot,
> read, write) : nice.
> 
> Then after detaching it and shuting down the MV, and for the first time,
> I tried to make use of the feature "direct attach" to attach the disk
> directly from oVirt, login the session via oVirt.
> I connected nice and I saw the disk appear in my VM as /dev/sda or
> whatever. I was able to mount it, read and write.
> 
> Then disaster stoke all this : many nodes suddenly began to become
> unresponsive, quickly migrating their VM to the remaining nodes.
> Hopefully, the migrations ran fine and I lost no VM nor downtime, but I
> had to reboot every concerned node (other actions failed).
> 
> In the failing nodes, /var/log/messages showed the log you can read in
> the end of this message.
> I first get device-mapper warnings, then the host unable to collaborate
> with the logical volumes.
> 
> The 3 volumes are the three main storage domains, perfectly up and
> running where I store my oVirt VMs.
> 
> My reflexions :
> - I'm not sure device-mapper is to blame. I frequently see device mapper
> complaining and nothing is getting worse (not oVirt specifically)
> - I have not change my network settings for months (bonding, linking...)
> The only new factor is the usage of direct attach LUN.
> - This morning I was able to reproduce the bug, just by trying again
> this attachement, and booting the VM. No mounting of the LUN, just VM
> booting, waiting, and this is enough to crash oVirt.
> - when the disaster happens, usually, amongst the nodes, only three
> nodes gets stroke, the only one that run VMs. Obviously, after
> migration, different nodes are hosting the VMs, and those new nodes are
> the one that then get stroke.
> 
> This is quite reproductible.
> 
> And frightening.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Google Summer of Code 2014

2014-02-13 Thread Federico Simoncelli
Hi everyone, I started a wiki page to list ideas for Google Summer
of Code 2014:

http://www.ovirt.org/Summer_of_Code

The deadline for the submission is really soon (14th of Feb) but
please feel free to try and add any idea that you may have.

For more information about Google Summer of Code please refer to:

https://developers.google.com/open-source/soc/

If you can't edit the wiki page please follow up to this thread
with your proposals.

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt 3.4 test day - Template Versions

2014-01-27 Thread Federico Simoncelli
- Original Message -
> From: "Omer Frenkel" 
> To: "Federico Simoncelli" 
> Cc: "oVirt Users List" , "Itamar Heim" 
> Sent: Monday, January 27, 2014 4:31:56 PM
> Subject: Re: oVirt 3.4 test day - Template Versions
> 
> Thanks for the feedback! much appreciated.
> 
> - Original Message -
> > From: "Federico Simoncelli" 
> > To: "oVirt Users List" 
> > Cc: "Omer Frenkel" , "Itamar Heim" 
> > Sent: Monday, January 27, 2014 5:12:38 PM
> > Subject: oVirt 3.4 test day - Template Versions
> > 
> > Feature tested:
> > 
> > http://www.ovirt.org/Features/Template_Versions
> > 
> > - create a new vm vm1 and make a template template1 from it
> > - create a new vm vm2 based on template1 and make some changes
> > - upgrade to 3.4
> > - create a new template template1.1 from vm2
> > - create a new vm vm3 from template1 (clone) - content ok
> > - create a new vm vm4 from template1.1 (thin) - content ok
> > - create a new vm vm5 from template1 last (thin) - content ok (same as 1.1)
> > - try to remove template1 (failed as template1.1 is still present)
> > - try to remove template1.1 (failed as vm5 is still present)
> > - create a new vm vm6 and make a template blank1.1 as new version of the
> >   blank template (succeeded)
> > - create a vm pool vmpool1 with the "latest" template from template1
> > - create a vm pool vmpool2 with the "template1.1" (last) template from
> > template1
> > - start vmpool1 and vmpool2 and verify that the content is the same
> > - create a new template template1.2
> > - start vmpool1 and verify that the content is the same as latest
> > (template1.2)
> > - start vmpool2 and verify that the content is the same as template1.1
> > 
> > Suggestions:
> > 
> > - the template blank is special, I am not sure if allowing versioning may
> >   be confusing (for example is not even editable)
> 
> right, i also thought about this, and my thought was not to block the user
> from doing this,
> but if it was confusing we better block it.
> 
> > - as far as I can see the "Sub Version Name" is not editable anymore (after
> >   picking it)
> 
> thanks, i see its missing in the UI, do you care to open a bug on that?

https://bugzilla.redhat.com/show_bug.cgi?id=1058501 

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Data Center stuck between "Non Responsive" and "Contending"

2014-01-27 Thread Federico Simoncelli
- Original Message -
> From: "Ted Miller" 
> To: "Federico Simoncelli" , "Itamar Heim" 
> 
> Cc: users@ovirt.org
> Sent: Monday, January 27, 2014 7:16:14 PM
> Subject: Re: [Users] Data Center stuck between "Non Responsive" and 
> "Contending"
> 
> 
> On 1/27/2014 3:47 AM, Federico Simoncelli wrote:
> > Maybe someone from gluster can identify easily what happened. Meanwhile if
> > you just want to repair your data-center you could try with:
> >
> >   $ cd
> >   
> > /rhev/data-center/mnt/glusterSD/10.41.65.2\:VM2/0322a407-2b16-40dc-ac67-13d387c6eb4c/dom_md/
> >   $ touch ids
> >   $ sanlock direct init -s
> >   0322a407-2b16-40dc-ac67-13d387c6eb4c:0:ids:1048576
> Federico,
> 
> I won't be able to do anything to the ovirt setup for another 5 hours or so
> (it is a trial system I am working on  at home, I am at work), but I will try
> your repair script and report back.
> 
> In bugzilla 862975 they suggested turning off write-behind caching and "eager
> locking" on the gluster volume to avoid/reduce the problems that come from
> many different computers all writing to the same file(s) on a very frequent
> basis.  If I interpret the comment in the bug correctly, it did seem to help
> in that situation.  My situation is a little different.  My gluster setup is
> replicate only, replica 3 (though there are only two hosts).  I was not
> stress-testing it, I was just using it, trying to figure out how I can import
> some old VMWare VMs without an ESXi server to run them on.

Have you done anything similar to what is described here in comment 21?

https://bugzilla.redhat.com/show_bug.cgi?id=859589#c21

When did you realize that you weren't able to use the data-center anymore?
Can you describe exactly what you did and what happened, for example:

1. I created the data center (up and running)
2. I tried to import some VMs from VMWare
3. During the import (or after it) the data-center went in the contending state
...

Did something special happened? I don't know, power loss, split-brain?
For example also an excessive load on one of the servers could have triggered
a timeout somewhere (forcing the data-center to go back in the contending
state).

Could you check if any host was fenced? (Forcibly rebooted)

> I am guessing that what makes cluster storage have the (Master) designation
> is that this is the one that actually contains the sanlocks?  If so, would it
> make sense to set up a gluster volume to be (Master), but not use it for VM
> storage, just for storing the sanlock info?  Separate gluster volume(s) could
> then have the VMs on it(them), and would not need the optimizations turned
> off.

Any domain must be able to become the master at any time. Without a master
the data center is unusable (at the present time), that's why we migrate (or
reconstruct) it on another domain when necessary.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] oVirt 3.4 test day - Template Versions

2014-01-27 Thread Federico Simoncelli
Feature tested:

http://www.ovirt.org/Features/Template_Versions

- create a new vm vm1 and make a template template1 from it
- create a new vm vm2 based on template1 and make some changes
- upgrade to 3.4
- create a new template template1.1 from vm2
- create a new vm vm3 from template1 (clone) - content ok
- create a new vm vm4 from template1.1 (thin) - content ok
- create a new vm vm5 from template1 last (thin) - content ok (same as 1.1)
- try to remove template1 (failed as template1.1 is still present)
- try to remove template1.1 (failed as vm5 is still present)
- create a new vm vm6 and make a template blank1.1 as new version of the
  blank template (succeeded)
- create a vm pool vmpool1 with the "latest" template from template1
- create a vm pool vmpool2 with the "template1.1" (last) template from template1
- start vmpool1 and vmpool2 and verify that the content is the same
- create a new template template1.2
- start vmpool1 and verify that the content is the same as latest (template1.2)
- start vmpool2 and verify that the content is the same as template1.1

Suggestions:

- the template blank is special, I am not sure if allowing versioning may
  be confusing (for example is not even editable)
- as far as I can see the "Sub Version Name" is not editable anymore (after
  picking it)

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Data Center stuck between "Non Responsive" and "Contending"

2014-01-27 Thread Federico Simoncelli
- Original Message -
> From: "Itamar Heim" 
> To: "Ted Miller" , users@ovirt.org, "Federico Simoncelli" 
> 
> Cc: "Allon Mureinik" 
> Sent: Sunday, January 26, 2014 11:17:04 PM
> Subject: Re: [Users] Data Center stuck between "Non Responsive" and 
> "Contending"
> 
> On 01/27/2014 12:00 AM, Ted Miller wrote:
> >
> > On 1/26/2014 4:00 PM, Itamar Heim wrote:
> >> On 01/26/2014 10:51 PM, Ted Miller wrote:
> >>>
> >>> On 1/26/2014 3:10 PM, Itamar Heim wrote:
> >>>> On 01/26/2014 10:08 PM, Ted Miller wrote:
> >>>> is this gluster storage (guessing sunce you mentioned a 'volume')
> >>> yes (mentioned under "setup" above)
> >>>> does it have a quorum?
> >>> Volume Name: VM2
> >>> Type: Replicate
> >>> Volume ID: 7bea8d3b-ec2a-4939-8da8-a82e6bda841e
> >>> Status: Started
> >>> Number of Bricks: 1 x 3 = 3
> >>> Transport-type: tcp
> >>> Bricks:
> >>> Brick1: 10.41.65.2:/bricks/01/VM2
> >>> Brick2: 10.41.65.4:/bricks/01/VM2
> >>> Brick3: 10.41.65.4:/bricks/101/VM2
> >>> Options Reconfigured:
> >>> cluster.server-quorum-type: server
> >>> storage.owner-gid: 36
> >>> storage.owner-uid: 36
> >>> auth.allow: *
> >>> user.cifs: off
> >>> nfs.disa
> >>>> (there were reports of split brain on the domain metadata before when
> >>>> no quorum exist for gluster)
> >>> after full heal:
> >>>
> >>> [root@office4a ~]$ gluster volume heal VM2 info
> >>> Gathering Heal info on volume VM2 has been successful
> >>>
> >>> Brick 10.41.65.2:/bricks/01/VM2
> >>> Number of entries: 0
> >>>
> >>> Brick 10.41.65.4:/bricks/01/VM2
> >>> Number of entries: 0
> >>>
> >>> Brick 10.41.65.4:/bricks/101/VM2
> >>> Number of entries: 0
> >>> [root@office4a ~]$ gluster volume heal VM2 info split-brain
> >>> Gathering Heal info on volume VM2 has been successful
> >>>
> >>> Brick 10.41.65.2:/bricks/01/VM2
> >>> Number of entries: 0
> >>>
> >>> Brick 10.41.65.4:/bricks/01/VM2
> >>> Number of entries: 0
> >>>
> >>> Brick 10.41.65.4:/bricks/101/VM2
> >>> Number of entries: 0
> >>>
> >>> noticed this in host /var/log/messages (while looking for something
> >>> else).  Loop seems to repeat over and over.
> >>>
> >>> Jan 26 15:35:52 office4a sanlock[3763]: 2014-01-26 15:35:52-0500 14678
> >>> [30419]: read_sectors delta_leader offset 512 rv -90
> >>> /rhev/data-center/mnt/glusterSD/10.41.65.2:VM2/0322a407-2b16-40dc-ac67-13d387c6eb4c/dom_md/ids
> >>>
> >>>
> >>> Jan 26 15:35:53 office4a sanlock[3763]: 2014-01-26 15:35:53-0500 14679
> >>> [3771]: s1997 add_lockspace fail result -90
> >>> Jan 26 15:35:58 office4a vdsm TaskManager.Task ERROR
> >>> Task=`89885661-88eb-4ea3-8793-00438735e4ab`::Unexpected
> >>> error#012Traceback
> >>> (most recent call last):#012  File "/usr/share/vdsm/storage/task.py",
> >>> line
> >>> 857, in _run#012 return fn(*args, **kargs)#012  File
> >>> "/usr/share/vdsm/logUtils.py", line 45, in wrapper#012res = f(*args,
> >>> **kwargs)#012  File "/usr/share/vdsm/storage/hsm.py", line 2111, in
> >>> getAllTasksStatuses#012allTasksStatus = sp.getAllTasksStatuses()#012
> >>> File "/usr/share/vdsm/storage/securable.py", line 66, in wrapper#012
> >>> raise
> >>> SecureError()#012SecureError
> >>> Jan 26 15:35:59 office4a sanlock[3763]: 2014-01-26 15:35:59-0500 14686
> >>> [30495]: read_sectors delta_leader offset 512 rv -90
> >>> /rhev/data-center/mnt/glusterSD/10.41.65.2:VM2/0322a407-2b16-40dc-ac67-13d387c6eb4c/dom_md/ids
> >>>
> >>>
> >>> Jan 26 15:36:00 office4a sanlock[3763]: 2014-01-26 15:36:00-0500 14687
> >>> [3772]: s1998 add_lockspace fail result -90
> >>> Jan 26 15:36:00 office4a vdsm TaskManager.Task ERROR
> >>> Task=`8db9ff1a-2894-407a-915a-279f6a7eb205`::Unexpected
> >>> error#012Traceback
> >>> (most recent call last):#012  File "/usr/share/vdsm/storage/task.py",
> >>> line
> >>> 857, in _run#012 return fn(*args, **kargs)#

Re: [Users] wiki site out of date?

2014-01-13 Thread Federico Simoncelli
- Original Message -
> From: "Sven Kieske" 
> To: "Users@ovirt.org List" 
> Sent: Monday, January 13, 2014 12:07:20 PM
> Subject: [Users] wiki site out of date?
> 
> Hi,
> 
> is this feature page up to date?
> 
> http://www.ovirt.org/Features/Online_Virtual_Drive_Resize
> 
> specifically the point:
> 
> QEMU-GA
> 
> support for notifying the guest and updating the size of the visible
> disk: To be integrated

That actually meant triggering the partition/lvm/filesystem resize
automatically (in the guest), so that you don't have to do it manually.

> because we used online drive resize and it does show up in ubuntu
> based vm, without qemu-ga installed?

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Broken Snapshots

2014-01-02 Thread Federico Simoncelli
- Original Message -
> From: "Leonid Natapov" 
> To: "Federico Simoncelli" 
> Cc: "Maurice James" , d...@redhat.com, "eduardo 
> Warszawski" , "Liron
> Aravot" , users@ovirt.org
> Sent: Thursday, January 2, 2014 2:09:13 PM
> Subject: Re: [Users] Broken Snapshots
> 
> Fede,there is an open BZ about that. See if it's the same
> https://bugzilla.redhat.com/show_bug.cgi?id=996945

That's the second part (deleting the broken snapshot), I'm trying to
understand why the snapshot failed.

-- 
Federico

> - Original Message -
> From: "Federico Simoncelli" 
> To: "Maurice James" 
> Cc: d...@redhat.com, "Leonid Natapov" , "eduardo
> Warszawski" , "Liron Aravot" ,
> users@ovirt.org
> Sent: Thursday, January 2, 2014 3:07:53 PM
> Subject: Re: [Users] Broken Snapshots
> 
> - Original Message -
> > From: "Maurice James" 
> > To: d...@redhat.com, "Leonid Natapov" 
> > Cc: "eduardo Warszawski" , "Federico Simoncelli"
> > , "Liron Aravot"
> > , users@ovirt.org
> > Sent: Thursday, January 2, 2014 2:04:06 PM
> > Subject: RE: [Users] Broken Snapshots
> > 
> > When I get home from work. I will attempt to delete the snapshot again then
> > send the vdsm.log to you guys. Thanks
> > 
> > > Date: Thu, 2 Jan 2014 12:58:47 +
> > > From: d...@redhat.com
> > > To: lnata...@redhat.com
> > > CC: midnightst...@msn.com; ewars...@redhat.com; fsimo...@redhat.com;
> > > lara...@redhat.com; users@ovirt.org
> > > Subject: Re: [Users] Broken Snapshots
> > > 
> > > Leo, there are two issues here:
> > > 1. I want to know what happened in Maurice's environment in the first
> > > place (why is the snapshot broken).
> > > 2. help with a workaround so that Maurice can delete the broken
> > > snapshots and continue working.
> 
> At least for me this is hard to track, can we open a bug and attach the logs
> there?
> 
> Are you using fedora or centos? On centos there's a known issue:
> 
> https://bugzilla.redhat.com/show_bug.cgi?id=1009100
> 
> --
> Federico

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Broken Snapshots

2014-01-02 Thread Federico Simoncelli
- Original Message -
> From: "Maurice James" 
> To: d...@redhat.com, "Leonid Natapov" 
> Cc: "eduardo Warszawski" , "Federico Simoncelli" 
> , "Liron Aravot"
> , users@ovirt.org
> Sent: Thursday, January 2, 2014 2:04:06 PM
> Subject: RE: [Users] Broken Snapshots
> 
> When I get home from work. I will attempt to delete the snapshot again then
> send the vdsm.log to you guys. Thanks
> 
> > Date: Thu, 2 Jan 2014 12:58:47 +
> > From: d...@redhat.com
> > To: lnata...@redhat.com
> > CC: midnightst...@msn.com; ewars...@redhat.com; fsimo...@redhat.com;
> > lara...@redhat.com; users@ovirt.org
> > Subject: Re: [Users] Broken Snapshots
> > 
> > Leo, there are two issues here:
> > 1. I want to know what happened in Maurice's environment in the first
> > place (why is the snapshot broken).
> > 2. help with a workaround so that Maurice can delete the broken
> > snapshots and continue working.

At least for me this is hard to track, can we open a bug and attach the logs
there?

Are you using fedora or centos? On centos there's a known issue:

https://bugzilla.redhat.com/show_bug.cgi?id=1009100

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-23 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: users@ovirt.org
> Sent: Monday, December 23, 2013 5:34:55 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> Here is the message I get on the console : Error while executing action
> Attach Storage Domain: AcquireHostIdFailure
> The software seems to go pretty far : it reaches the locked stated before
> failing.
> 
> In engine.log
> 2013-12-23 16:56:49,497 ERROR
> [org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand]
> (ajp--127.0.0.1-8702-2) Command
> org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand throw
> Vdc Bll exception. With error message VdcBLLException:
> org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException:
> VDSGenericException: VDSErrorException: Failed to CreateStoragePoolVDS,
> error = Cannot acquire host id: ('8c626a3f-5846-434e-83d8-6238e1ff9e03',
> SanlockException(-203, 'Sanlock lockspace add failure', 'Sanlock
> exception')) (Failed with error AcquireHostIdFailure and code 661)
> 
> Can this help ?

What's the logs in VDSM? Is this the same host where wdmd was up and
running or another one? If you restarted your laptop and you didn't
persist the module loading (following the instruction in one of my
previous emails) you'll end up in the same problem every time.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-23 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: users@ovirt.org
> Sent: Monday, December 23, 2013 4:33:23 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> Ok.
> What I am doing is just adding a new NFS domain that fails : "Failed to add
> Storage Domain DataLab2. (User: admin@internal)"
> And I thought that the "/dev/watchdog0 failed to set timeout" msg was
> signalling an error.

No that's just the attempt to use the laptop watchdog but then it
fallbacks to the softdog one:

Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog1 armed with fire_timeout 60
Dec 20 23:44:11 lab2 systemd[1]: Started Watchdog Multiplexing Daemon.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-23 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Saturday, December 21, 2013 12:37:51 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> Nope. Nothing more in logs.
> My guess is that the timeout problem generates the error.
> However, in reality if you run "mount", you have the target partitions
> mounted

If you still have a problem and something is not working there's an error
somewhere and we only have to find it. Look in the engine logs and in the
vdsm logs for any error (not only "Traceback" but also "ERROR").

Try to describe with more details what you're trying to do, what you expect
to happen and what is happening instead.

> Therefore, I guess the problem is to understand why "dev/watchdog0 failed
> to set timeout"

The watchdog provided by your laptop is not working properly or it's
not able to set the timeout we need.

You inserted the softdog module and wdmd is now up and running as you
reported with:

> 2013/12/21 Federico Simoncelli 
> 
> > - Original Message -
> > > From: "Pascal Jakobi" 
> > > To: "Federico Simoncelli" , users@ovirt.org
> > > Sent: Friday, December 20, 2013 11:54:21 PM
> > > Subject: Re: [Users] AcquireHostId problem
> > >
> > > Dec 20 23:43:59 lab2 kernel: [183033.639261] softdog: Software Watchdog
> > > Timer: 0.08 initialized. soft_noboot=0 soft_margin=60 sec soft_panic=0
> > > (nowayout=0)
> > > Dec 20 23:44:11 lab2 systemd[1]: Starting Watchdog Multiplexing Daemon...
> > > Dec 20 23:44:11 lab2 wdmd[25072]: wdmd started S0 H1 G179
> > > Dec 20 23:44:11 lab2 systemd-wdmd[25066]: Starting wdmd: [  OK  ]
> > > Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog0 failed to set timeout
> > > Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog0 disarmed
> > > Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog1 armed with fire_timeout 
> > > 60
> > > Dec 20 23:44:11 lab2 systemd[1]: Started Watchdog Multiplexing Daemon.

So as far as I can see this part is working.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-20 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: "Federico Simoncelli" , users@ovirt.org
> Sent: Friday, December 20, 2013 11:54:21 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> Dec 20 23:43:59 lab2 kernel: [183033.639261] softdog: Software Watchdog
> Timer: 0.08 initialized. soft_noboot=0 soft_margin=60 sec soft_panic=0
> (nowayout=0)
> Dec 20 23:44:11 lab2 systemd[1]: Starting Watchdog Multiplexing Daemon...
> Dec 20 23:44:11 lab2 wdmd[25072]: wdmd started S0 H1 G179
> Dec 20 23:44:11 lab2 systemd-wdmd[25066]: Starting wdmd: [  OK  ]
> Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog0 failed to set timeout
> Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog0 disarmed
> Dec 20 23:44:11 lab2 wdmd[25072]: /dev/watchdog1 armed with fire_timeout 60
> Dec 20 23:44:11 lab2 systemd[1]: Started Watchdog Multiplexing Daemon.
> Dec 20 23:45:33 lab2 rpc.mountd[2819]: authenticated mount request from
> 192.168.1.41:994 for /home/vdsm/data (/home/vdsm/data)
> Dec 20 23:45:39 lab2 rpc.mountd[2819]: authenticated mount request from
> 192.168.1.41:954 for /home/vdsm/data (/home/vdsm/data)
> 
> Seems to work a bit.
> However I still get "unable to attach storage" when creating a domain

It is probably a different error now. Anything interesting in vdsm.log?

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-20 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org, "David Teigland" 
> Sent: Friday, December 20, 2013 7:19:52 AM
> Subject: Re: [Users] AcquireHostId problem
> 
> Here you go !
> I am running F19 on a Lenovo S30.
> Thxs

Thanks, can you open a bug on this issue?
(Attach also the files to the bug).

I suppose it will be later split into different ones, one for the failing
watchdog device and maybe an RFE to wdmd to automatically load the softdog
if there are no usable watchdog devices.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-19 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Thursday, December 19, 2013 5:17:59 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> sanlock-2.8-1.fc19.x86_64
> 
> Dec 18 16:46:05 lab2 systemd-wdmd[818]: Starting wdmd: [  OK  ]
> Dec 18 16:46:05 lab2 systemd[1]: Unit wdmd.service entered failed state.
> Dec 18 20:52:12 lab2 wdmd[887]: wdmd started S0 H1 G179
> Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 failed to set timeout
> Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog0 disarmed
> Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog failed to set timeout
> Dec 18 20:52:12 lab2 wdmd[887]: /dev/watchdog disarmed
> Dec 18 20:52:12 lab2 wdmd[887]: no watchdog device, load a watchdog driver
> Dec 18 20:52:12 lab2 systemd[1]: wdmd.service: main process exited,
> code=exited, status=255/n/a
> 
> I do not really understand all this. Furthermore, I could not find
> reference to it in the documentation. Am I missing sthing ? What does "load
> a wathdog driver" means in concrete ?

It seems that you have a watchdog device (probably provided by the
motherboard of your pc/server) that is not currently usable by wdmd.

The temporary workaround to get your system to work is to modprobe
the softdog kernel module and restart wdmd:

 # modprobe softdog
 # service wdmd restart

Depending on the distribution you're using there are different ways
to automatically load the module at boot, for example on fedora you
can use:

http://www.freedesktop.org/software/systemd/man/modules-load.d.html

and for centos/rhel 6 you can look at the content of:

/etc/sysconfig/modules

for an example of how to do the same (e.g. kvm.modules).

That said, in order to get your specific watchdog device to work
properly I'd need some additional information. Can you provide me
the output of these two commands?

 # lspci -vvv
 # lsmod

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-19 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Thursday, December 19, 2013 4:05:07 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> Federico
> 
> 
> On may suspect wdmd isn't running as wdmd_connect failed (see sanlock.log).
> I have ran a "ps" command - no wdmd
> 
> Any idea why wdmd_connect might fail ?

Are you using fedora? What is the version of sanlock?
Do you see any information about wdmd in /var/log/messages?
Is the wdmd service started?

# service wdmd status
# systemctl status wdmd.service

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] AcquireHostId problem

2013-12-19 Thread Federico Simoncelli
- Original Message -
> From: "Pascal Jakobi" 
> To: users@ovirt.org
> Sent: Wednesday, December 18, 2013 9:32:00 PM
> Subject: Re: [Users] AcquireHostId problem
> 
> sanlock.log :
> 
> 2013-12-18 21:23:32+0100 1900 [867]: s1 lockspace
> b2d69b22-a8b8-466c-bf1f-b6e565228238:250:/rhev/data-center/mnt/lab2.home:_home_vdsm_data/b2d69b22-a8b8-466c-bf1f-b6e565228238/dom_md/ids:0
> 2013-12-18 21:23:52+0100 1920 [4238]: s1 wdmd_connect failed -111
> 2013-12-18 21:23:52+0100 1920 [4238]: s1 create_watchdog failed -1
> 2013-12-18 21:23:53+0100 1921 [867]: s1 add_lockspace fail result -203

Hi Pascal,
 is wdmd up and running?

# ps aux | grep wdmd
root  1650  0.0  0.2  13552  3320 ?SLs  03:49   0:00 wdmd -G sanlock

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Cinder Integration

2013-11-19 Thread Federico Simoncelli
- Original Message -
> From: "Itamar Heim" 
> To: "Udaya Kiran P" , users@ovirt.org, "Oved Ourfalli" 
> , "Federico
> Simoncelli" 
> Sent: Tuesday, November 19, 2013 12:42:05 PM
> Subject: Re: [Users] Cinder Integration
> 
> On 11/19/2013 06:00 AM, Udaya Kiran P wrote:
> > Hi All,
> >
> > I want to consume the oVirt Storage Domains in OpenStack Cinder.
> >
> > Is this driver available or are there any resources pointing on how this
> > can be done?
> 
> Federico - was that your sample driver or Oved's?

It was Oved's. I have the feeling that I did some research in that area
as well but then I moved to glance.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Live storage migration snapshot removal (fails)

2013-11-08 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Sander Grendelman" 
> Cc: users@ovirt.org, fsimo...@redhat.com, ykap...@redhat.com
> Sent: Friday, November 8, 2013 4:06:53 PM
> Subject: Re: [Users] Live storage migration snapshot removal (fails)
> 
> On Fri, Nov 08, 2013 at 02:20:39PM +0100, Sander Grendelman wrote:
> 
> 
> 
> > 9d4e8a43-4851-42ff-a684-f3d802527cf7/c512267d-ebba-4907-a782-fec9b6c95116
> > 52178458-1764-4317-b85b-71843054aae9::WARNING::2013-11-08
> > 14:02:53,772::image::1164::Storage.Image::(merge) Auto shrink after
> > merge failed
> > Traceback (most recent call last):
> >   File "/usr/share/vdsm/storage/image.py", line 1162, in merge
> > srcVol.shrinkToOptimalSize()
> >   File "/usr/share/vdsm/storage/blockVolume.py", line 320, in
> > shrinkToOptimalSize
> > qemuImg.FORMAT.QCOW2)
> >   File "/usr/lib64/python2.6/site-packages/vdsm/qemuImg.py", line 109, in
> >   check
> > raise QImgError(rc, out, err, "unable to parse qemu-img check output")
> > QImgError: ecode=0, stdout=['No errors were found on the image.'],
> > stderr=[], message=unable to parse qemu-img check output
> 
> 
> I'm not sure that it's the only problem in this flow, but there's a
> clear bug in lib/vdsm/qemuImg.py's check() function: it fails to parse
> the output of qemu-img.
> 
> Would you open a bug on that? I found no open one.

I remember that this was discussed and the agreement was that if the offset
is not reported by qemu-img we should have used the old method to calculate
the new volume size.

We'll probably need to verify it. Sander can you open a bug on this?

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Urgent: Export NFS Migration issue oVirt 3.0 -> 3.2.1

2013-11-08 Thread Federico Simoncelli
- Original Message -
> From: "Sven Knohsalla" 
> To: users@ovirt.org
> Sent: Friday, November 8, 2013 10:32:32 AM
> Subject: Re: [Users] Urgent: Export NFS Migration issue oVirt 3.0 -> 3.2.1
> 
> Hi,
> 
> I could eliminate this issue to our oVirt 3.0 instance, as the pool_uuid &
> SHA checksum in metadata on NFS Export wasn't cleared properly from engine
> 3.0.
> (/NFSmountpoint//dom_md/metadata)

Hi Sven,
 can you send the original metadata and the relevant vdsm logs?

If I read your engine logs correctly we need the vdsm logs from the host
deovn-a01 (vds id 66b546c2-ae62-11e1-b734-5254005cbe44) around the same
time this was issued:

2013-11-08 08:12:49,075 INFO  
[org.ovirt.engine.core.bll.storage.ConnectStorageToVdsCommand] 
(pool-5-thread-39) Running command: ConnectStorageToVdsCommand internal: true. 
Entities affected :  ID: aaa0----123456789aaa Type: System
2013-11-08 08:12:49,079 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand] 
(pool-5-thread-39) START, ConnectStorageServerVDSCommand(vdsId = 
66b546c2-ae62-11e1-b734-5254005cbe44, storagePoolId = 
----, storageType = NFS, connectionList = [{ 
id: 2a84acc3-1700-45c4-bbf7-a3305b338f83, connection: 172.16.101.95:/ovirtmig02 
};]), log id: 2482b112
2013-11-08 08:12:52,092 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand] 
(pool-5-thread-50) FINISH, ConnectStorageServerVDSCommand, return: 
{2a84acc3-1700-45c4-bbf7-a3305b338f83=451}, log id: 7dcfb51f
2013-11-08 08:12:52,099 ERROR 
[org.ovirt.engine.core.bll.storage.NFSStorageHelper] (pool-5-thread-50) The 
connection with details 172.16.101.95:/ovirtmig02 failed because of error code 
451 and error message is: error storage server connection

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt Weekly Meeting Minutes -- 2013-10-09

2013-10-10 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Mike Burns" 
> Cc: bo...@ovirt.org, "users" 
> Sent: Wednesday, October 9, 2013 5:45:23 PM
> Subject: Re: [Users] oVirt Weekly Meeting Minutes -- 2013-10-09
> 
> On Wed, Oct 09, 2013 at 11:15:41AM -0400, Mike Burns wrote:
> > Minutes:
> > http://ovirt.org/meetings/ovirt/2013/ovirt.2013-10-09-14.06.html
> > Minutes (text):
> > http://ovirt.org/meetings/ovirt/2013/ovirt.2013-10-09-14.06.txt
> > Log:
> > http://ovirt.org/meetings/ovirt/2013/ovirt.2013-10-09-14.06.log.html
> > 
> > =
> > #ovirt: oVirt Weekly sync
> > =
> > 
> > 
> > Meeting started by mburns at 14:06:41 UTC. The full logs are available
> > at http://ovirt.org/meetings/ovirt/2013/ovirt.2013-10-09-14.06.log.html
> > .
> > 
> > Meeting summary
> > ---
> > * agenda and roll call  (mburns, 14:07:00)
> >   * 3.3 updates  (mburns, 14:07:17)
> >   * 3.4 planning  (mburns, 14:07:24)
> >   * conferences and workshops  (mburns, 14:07:31)
> >   * infra update  (mburns, 14:07:34)
> > 
> > * 3.3 updates  (mburns, 14:08:42)
> >   * 3.3.0.1 vdsm packages are posted to updates-testing  (mburns,
> > 14:09:04)
> >   * LINK: https://bugzilla.redhat.com/show_bug.cgi?id=1009100
> > (sbonazzo, 14:10:33)
> >   * 2 open bugs blocking 3.3.0.1  (mburns, 14:29:35)
> >   * 1 is deferred due to qemu-kvm feature set in el6  (mburns, 14:29:49)
> >   * other is allowed versions for vdsm  (mburns, 14:30:01)
> >   * vdsm version bug will be backported to 3.3.0.1 today  (mburns,
> > 14:30:13)
> >   * ACTION: sbonazzo to build engine 3.3.0.1 tomorrow  (mburns,
> > 14:30:22)
> >   * ACTION: mburns to post 3.3.0.1 to ovirt.org tomorrow  (mburns,
> > 14:30:32)
> >   * expected release:  next week  (mburns, 14:30:46)
> >   * ACTION: danken and sbonazzo to provide release notes for 3.3.0.1
> > (mburns, 14:37:56)
> 
> """
> A vdsm bug (BZ#1007980) made it impossible to migrate or re-run a VM
> with a glusterfs-backed virtual disk if the VM was originally started
> with an empty cdrom.
> 
> If you have encountered this bug, you would have to manually find the
> affected VMs with
> 
> psql -U engine -d engine -c "select distinct vm_name from vm_static,
> vm_device where vm_guid=vm_id and device='cdrom' and address ilike
> '%pci%';"
> 
> and remove their junk cdrom address with
> 
> psql -U engine -d engine -c "update vm_device set address='' where
> device='cdrom' and address ilike '%pci%';"
> """

We are currently building vdsm-4.12.1-4 that is carrying a new critical
fix related to drive resize (extension). Release notes:

"""
A vdsm bug introduced in a specific case of disk resize (raw on nfs)
accidentally wipes the content of the virtual disk.
The issue was masked on the master (and ovirt-3.3) branch by an unrelated
change that happened to fix the problem leaving only vdsm-4.12 affected.
It is of critical importance to update all your machines to the new vdsm
release.
"""

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] VM wont restart after some NFS snapshot restore.

2013-10-04 Thread Federico Simoncelli
Usman,
 the only thing that comes to my mind is something related to:

http://gerrit.ovirt.org/13529

which means that in some way the restored volumes are either inaccessible
(permissions?) or their metadata is corrupted (but it doesn't seem so).

There is probably another Traceback in the logs that should give us more
information.

Could you post somewhere the entire vdsm log?

Thanks.
-- 
Federico

- Original Message -
> From: "Usman Aslam" 
> To: "Federico Simoncelli" 
> Cc: users@ovirt.org
> Sent: Friday, October 4, 2013 4:49:08 PM
> Subject: Re: [Users] VM wont restart after some NFS snapshot restore.
> 
> Federico,
> 
> The files reside on this mount on the hypervisor
> /rhev/data-center/mnt/xyz-02.tufts.edu:
> _vol_tusk__vm_tusk__vm/fa3279ec-2912-45ac-b7bc-9fe89151ed99/images/79ccd989-3033-4e6a-80da-ba210c94225a
> 
> and are symlinked as described below
> 
> [root@xyz-02 430cd986-6488-403b-8d46-29abbc3eba38]# pwd
> /rhev/data-center/430cd986-6488-403b-8d46-29abbc3eba38
> [root@xyz-02 430cd986-6488-403b-8d46-29abbc3eba38]# ll
> total 12
> lrwxrwxrwx 1 vdsm kvm 120 Oct  3 12:35 ee2ae498-6e45-448d-8f91-0efca377dcf6
> -> /rhev/data-center/mnt/xyz-02.tufts.edu:
> _vol_tusk__iso_tusk__iso/ee2ae498-6e45-448d-8f91-0efca377dcf6
> lrwxrwxrwx 1 vdsm kvm 118 Oct  3 12:35 fa3279ec-2912-45ac-b7bc-9fe89151ed99
> -> /rhev/data-center/mnt/xyz-02.tufts.edu:
> _vol_tusk__vm_tusk__vm/fa3279ec-2912-45ac-b7bc-9fe89151ed99
> lrwxrwxrwx 1 vdsm kvm 118 Oct  3 12:35 mastersd ->
> /rhev/data-center/mnt/xyz-02.tufts.edu:
> _vol_tusk__vm_tusk__vm/fa3279ec-2912-45ac-b7bc-9fe89151ed99
> 
> I did a diff and the contents of the the *Original *meta file (that works
> and VM starts but have bad file system) and the *Backup *meta file (the
> files being restored from nfs snapshot) *are exactly the same*.
> Contents are listed blow. Also the files sizes for all six related files
> are exactly the same.
> 
> [root@xyz-02 images]# cat
> 79ccd989-3033-4e6a-80da-ba210c94225a/039a8482-c267-4051-b1e6-1c1dee49b3d7.meta
> DOMAIN=fa3279ec-2912-45ac-b7bc-9fe89151ed99
> VOLTYPE=SHARED
> CTIME=1368457020
> FORMAT=RAW
> IMAGE=59b6a429-bd11-40c6-a218-78df840725c6
> DISKTYPE=2
> PUUID=----
> LEGALITY=LEGAL
> MTIME=1368457020
> POOL_UUID=
> DESCRIPTION=Active VM
> TYPE=SPARSE
> SIZE=104857600
> EOF
> [root@tss-tusk-ovirt-02 images]# cat
> 79ccd989-3033-4e6a-80da-ba210c94225a/8d48505d-846d-49a7-8b50-d972ee051145.meta
> DOMAIN=fa3279ec-2912-45ac-b7bc-9fe89151ed99
> CTIME=1370303194
> FORMAT=COW
> DISKTYPE=2
> LEGALITY=LEGAL
> SIZE=104857600
> VOLTYPE=LEAF
> DESCRIPTION=
> IMAGE=79ccd989-3033-4e6a-80da-ba210c94225a
> PUUID=039a8482-c267-4051-b1e6-1c1dee49b3d7
> MTIME=1370303194
> POOL_UUID=
> TYPE=SPARSE
> EOF
> 
> Any help would be greatly appreciated!
> 
> Thanks,
> Usman
> 
> 
> On Fri, Oct 4, 2013 at 9:50 AM, Federico Simoncelli
> wrote:
> 
> > Hi Usman,
> >  can you paste somewhere the content of the meta files?
> >
> >  $ cat 039a8482-c267-4051-b1e6-1c1dee49b3d7.meta
> > 8d48505d-846d-49a7-8b50-d972ee051145.meta
> >
> > could you also provide the absolute path to those files? (in the vdsm host)
> >
> > Thanks,
> > --
> > Federico
> >
> > - Original Message -
> > > From: "Usman Aslam" 
> > > To: users@ovirt.org
> > > Sent: Thursday, October 3, 2013 4:29:43 AM
> > > Subject: [Users] VM wont restart after some NFS snapshot restore.
> > >
> > > I have some VM's that live on NFS share. Basically, I had to revert the
> > VM
> > > disk to a backup from a few days ago. So I powered the VM down, copied
> > over
> > > the following files
> > >
> > > 039a8482-c267-4051-b1e6-1c1dee49b3d7
> > > 039a8482-c267-4051-b1e6-1c1dee49b3d7.lease
> > > 039a8482-c267-4051-b1e6-1c1dee49b3d7.meta
> > > 8d48505d-846d-49a7-8b50-d972ee051145
> > > 8d48505d-846d-49a7-8b50-d972ee051145.lease
> > > 8d48505d-846d-49a7-8b50-d972ee051145.meta
> > >
> > > and now when I try to power the VM, it complains
> > >
> > > 2013-Oct-02, 22:02:38
> > > Failed to run VM zabbix-prod-01 (User: admin@internal).
> > > 2013-Oct-02, 22:02:38
> > > Failed to run VM zabbix-prod-01 on Host
> > > tss-tusk-ovirt-01-ovirtmgmt.tusk.tufts.edu .
> > > 2013-Oct-02, 22:02:38
> > > VM zabbix-prod-01 is down. Exit message: 'truesize'.
> > >
> > > Any ideas on how I could resolve this? Perhaps a better way of
> > approaching
> > > t

Re: [Users] VM wont restart after some NFS snapshot restore.

2013-10-04 Thread Federico Simoncelli
Hi Usman,
 can you paste somewhere the content of the meta files?

 $ cat 039a8482-c267-4051-b1e6-1c1dee49b3d7.meta 
8d48505d-846d-49a7-8b50-d972ee051145.meta

could you also provide the absolute path to those files? (in the vdsm host)

Thanks,
-- 
Federico

- Original Message -
> From: "Usman Aslam" 
> To: users@ovirt.org
> Sent: Thursday, October 3, 2013 4:29:43 AM
> Subject: [Users] VM wont restart after some NFS snapshot restore.
> 
> I have some VM's that live on NFS share. Basically, I had to revert the VM
> disk to a backup from a few days ago. So I powered the VM down, copied over
> the following files
> 
> 039a8482-c267-4051-b1e6-1c1dee49b3d7
> 039a8482-c267-4051-b1e6-1c1dee49b3d7.lease
> 039a8482-c267-4051-b1e6-1c1dee49b3d7.meta
> 8d48505d-846d-49a7-8b50-d972ee051145
> 8d48505d-846d-49a7-8b50-d972ee051145.lease
> 8d48505d-846d-49a7-8b50-d972ee051145.meta
> 
> and now when I try to power the VM, it complains
> 
> 2013-Oct-02, 22:02:38
> Failed to run VM zabbix-prod-01 (User: admin@internal).
> 2013-Oct-02, 22:02:38
> Failed to run VM zabbix-prod-01 on Host
> tss-tusk-ovirt-01-ovirtmgmt.tusk.tufts.edu .
> 2013-Oct-02, 22:02:38
> VM zabbix-prod-01 is down. Exit message: 'truesize'.
> 
> Any ideas on how I could resolve this? Perhaps a better way of approaching
> the restore on a filesystem level?
> 
> I see the following the vsdm.log
> 
> Thread-7843::ERROR::2013-10-02
> 22:02:37,548::vm::716::vm.Vm::(_startUnderlyingVm)
> vmId=`8e8764ad-6b4c-48d8-9a19-fa5cf77208ef`::The vm start process failed
> Traceback (most recent call last):
> File "/usr/share/vdsm/vm.py", line 678, in _startUnderlyingVm
> self._run()
> File "/usr/share/vdsm/libvirtvm.py", line 1467, in _run
> devices = self.buildConfDevices()
> File "/usr/share/vdsm/vm.py", line 515, in buildConfDevices
> self._normalizeVdsmImg(drv)
> File "/usr/share/vdsm/vm.py", line 408, in _normalizeVdsmImg
> drv['truesize'] = res['truesize']
> KeyError: 'truesize'
> Thread-7843::DEBUG::2013-10-02 22:02:37,553::vm::1065::vm.Vm::(setDownStatus)
> vmId=`8e8764ad-6b4c-48d8-9a19-fa5cf77208ef`::Changed state to Down:
> 'truesize'
> 
> 
> Any help would be really nice, thanks!
> --
> Usman
> 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Glance with oVirt

2013-10-02 Thread Federico Simoncelli
- Original Message -
> From: "Itamar Heim" 
> To: "Riccardo Brunetti" 
> Cc: "Jason Brooks" , users@ovirt.org, "Federico 
> Simoncelli" 
> Sent: Wednesday, October 2, 2013 9:18:57 PM
> Subject: Re: [Users] Glance with oVirt
> 
> On 09/25/2013 05:03 PM, Riccardo Brunetti wrote:
> > On 09/24/2013 08:24 PM, Itamar Heim wrote:
> >> On 09/24/2013 06:06 PM, Jason Brooks wrote:
> >>
> > Dear all.
> > Unfortunately I did manage to work out only the first step of the
> > procedure.
> > I can successfully define the glance external provider and from the
> > storage tab I can list the images available in glance, but when I try to
> > import an image (ie. a Fedora19 qcow2 image which is working inside
> > OpenStack) the disk is stuck in "Illegal" state and in the logs
> > (ovirt-engine) I see messages like:
> >
> > ...
> >
> > Can you help me?
> >
> > Thanks a lot
> > Riccardo
> >
> 
> was this resolved?

Several improvements have been introduced lately (both in engine and in vdsm).
We might have two issues involved here (qcow2 case):

https://bugzilla.redhat.com/show_bug.cgi?id=1013643
http://gerrit.ovirt.org/#/c/19222/

as far as I know the ovirt-engine-3.3 branch is in a good state now with
regard to the glance integration (bz1013643 is still pending but it will
be merged soon) and I expect to have a much more stable import for qcow2
in the upcoming 3.3.1.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Resizing disks destroys contents

2013-10-01 Thread Federico Simoncelli
Hi Martijn,
 can you post somewhere the relevant vdsm logs (from the spm host)?

Thanks,
-- 
Federico


- Original Message -
> From: "Martijn Grendelman" 
> To: users@ovirt.org
> Sent: Tuesday, October 1, 2013 5:01:39 PM
> Subject: [Users] Resizing disks destroys contents
> 
> Hi,
> 
> I just tried out another feature of oVirt and again, I am shocked by the
> results.
> 
> I did the following:
> - create new VM based on an earlier created template, with 20 GB disk
> - Run the VM -> boots fine
> - Shut down the VM
> - Via Disks -> Edit -> Extend size by(GB) add 20 GB to the disk
> - Run the VM
> 
> Result: no bootable device. Linux installation gone.
> 
> Just to be sure, I booted the VM with a gparted live iso, and gparted
> reports the entire 40 GB as unallocated space.
> 
> Where's my data? What's wrong with my oVirt installation? What am I
> doing wrong?
> 
> Regards,
> Martijn.
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] vdsm live migration errors in latest master

2013-09-26 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Federico Simoncelli" 
> Cc: "Dead Horse" , "users" , 
> vdsm-de...@fedorahosted.org,
> aba...@redhat.com
> Sent: Thursday, September 26, 2013 2:09:15 PM
> Subject: Re: [Users] vdsm live migration errors in latest master
> 
> On Thu, Sep 26, 2013 at 05:35:46AM -0400, Federico Simoncelli wrote:
> > - Original Message -
> > > From: "Dan Kenigsberg" 
> > > To: "Federico Simoncelli" 
> > > Cc: "Dead Horse" , "users"
> > > , vdsm-de...@fedorahosted.org,
> > > aba...@redhat.com
> > > Sent: Thursday, September 26, 2013 1:38:15 AM
> > > Subject: Re: [Users] vdsm live migration errors in latest master
> > > 
> > > On Tue, Sep 24, 2013 at 12:04:14PM -0400, Federico Simoncelli wrote:
> > > > - Original Message -
> > > > > From: "Dan Kenigsberg" 
> > > > > To: "Dead Horse" 
> > > > > Cc: "" ,
> > > > > vdsm-de...@fedorahosted.org,
> > > > > fsimo...@redhat.com, aba...@redhat.com
> > > > > Sent: Tuesday, September 24, 2013 11:44:48 AM
> > > > > Subject: Re: [Users] vdsm live migration errors in latest master
> > > > > 
> > > > > On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
> > > > > > Seeing failed live migrations and these errors in the vdsm logs
> > > > > > with
> > > > > > latest
> > > > > > VDSM/Engine master.
> > > > > > Hosts are EL6.4
> > > > > 
> > > > > Thanks for posting this report.
> > > > > 
> > > > > The log is from the source of migration, right?
> > > > > Could you trace the history of the hosts of this VM? Could it be that
> > > > > it
> > > > > was started on an older version of vdsm (say ovirt-3.3.0) and then
> > > > > (due
> > > > > to migration or vdsm upgrade) got into a host with a much newer vdsm?
> > > > > 
> > > > > Would you share the vmCreate (or vmMigrationCreate) line for this Vm
> > > > > in
> > > > > your log? I smells like an unintended regression of
> > > > > http://gerrit.ovirt.org/17714
> > > > > vm: extend shared property to support locking
> > > > > 
> > > > > solving it may not be trivial, as we should not call
> > > > > _normalizeDriveSharedAttribute() automatically on migration
> > > > > destination,
> > > > > as it may well still be apart of a 3.3 clusterLevel.
> > > > > 
> > > > > Also, migration from vdsm with extended shared property, to an ovirt
> > > > > 3.3
> > > > > vdsm is going to explode (in a different way), since the destination
> > > > > does not expect the extended values.
> > > > > 
> > > > > Federico, do we have a choice but to revert that patch, and use
> > > > > something like "shared3" property instead?
> > > > 
> > > > I filed a bug at:
> > > > 
> > > > https://bugzilla.redhat.com/show_bug.cgi?id=1011608
> > > > 
> > > > A possible fix could be:
> > > > 
> > > > http://gerrit.ovirt.org/#/c/19509
> > > 
> > > Beyond this, we must make sure that on Engine side, the extended shared
> > > values would be used only for clusterLevel 3.4 and above.
> > > 
> > > Are the extended shared values already used by Engine?
> > 
> > Yes. That's the idea. Actually to be fair, the second case you mentioned
> > (migrating from extended shared property to old vdsm) it wouldn't have been
> > possible I suppose (the issue here is that Dead Horse has one or more
> > hosts running on master instead of 3.3). The extended shared property would
> > have appeared only in 3.4 and to allow the migration you would have had to
> > upgrade all the nodes.
> > 
> > But anyway since we were also talking about a new 3.3.1 barnch I just went
> > ahead and covered all cases.
> 
> I do not see how the 3.3.1 branch is relevant to the discussion, as its
> Vdsm is NOT going to support clusterLevel 3.4.

That is what I was referring to.

If 3.3.1 was 3.3.0 + backported patches then we just wouldn't backport the
extended shared attributes patch and that's it. But from what I understood
3.3.1 will be rebased on master (where instead we have the extended shared
attributes) and that is why we have to cover both migration direction cases
(instead of just the simple getattr one).

> Pardon my slowliness, but would you confirm that this feature is to be
> used only on clusterLevel 3.4 and above? If so, I'm +2ing your patch.

Yes, the extended attributes will be used in the hosted engine and cluster
level 3.4.
But what the engine does is not relevant to +2ing correct vdsm patches.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] vdsm live migration errors in latest master

2013-09-26 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Federico Simoncelli" 
> Cc: "Dead Horse" , "users" , 
> vdsm-de...@fedorahosted.org,
> aba...@redhat.com
> Sent: Thursday, September 26, 2013 1:38:15 AM
> Subject: Re: [Users] vdsm live migration errors in latest master
> 
> On Tue, Sep 24, 2013 at 12:04:14PM -0400, Federico Simoncelli wrote:
> > - Original Message -
> > > From: "Dan Kenigsberg" 
> > > To: "Dead Horse" 
> > > Cc: "" , vdsm-de...@fedorahosted.org,
> > > fsimo...@redhat.com, aba...@redhat.com
> > > Sent: Tuesday, September 24, 2013 11:44:48 AM
> > > Subject: Re: [Users] vdsm live migration errors in latest master
> > > 
> > > On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
> > > > Seeing failed live migrations and these errors in the vdsm logs with
> > > > latest
> > > > VDSM/Engine master.
> > > > Hosts are EL6.4
> > > 
> > > Thanks for posting this report.
> > > 
> > > The log is from the source of migration, right?
> > > Could you trace the history of the hosts of this VM? Could it be that it
> > > was started on an older version of vdsm (say ovirt-3.3.0) and then (due
> > > to migration or vdsm upgrade) got into a host with a much newer vdsm?
> > > 
> > > Would you share the vmCreate (or vmMigrationCreate) line for this Vm in
> > > your log? I smells like an unintended regression of
> > > http://gerrit.ovirt.org/17714
> > > vm: extend shared property to support locking
> > > 
> > > solving it may not be trivial, as we should not call
> > > _normalizeDriveSharedAttribute() automatically on migration destination,
> > > as it may well still be apart of a 3.3 clusterLevel.
> > > 
> > > Also, migration from vdsm with extended shared property, to an ovirt 3.3
> > > vdsm is going to explode (in a different way), since the destination
> > > does not expect the extended values.
> > > 
> > > Federico, do we have a choice but to revert that patch, and use
> > > something like "shared3" property instead?
> > 
> > I filed a bug at:
> > 
> > https://bugzilla.redhat.com/show_bug.cgi?id=1011608
> > 
> > A possible fix could be:
> > 
> > http://gerrit.ovirt.org/#/c/19509
> 
> Beyond this, we must make sure that on Engine side, the extended shared
> values would be used only for clusterLevel 3.4 and above.
> 
> Are the extended shared values already used by Engine?

Yes. That's the idea. Actually to be fair, the second case you mentioned
(migrating from extended shared property to old vdsm) it wouldn't have been
possible I suppose (the issue here is that Dead Horse has one or more
hosts running on master instead of 3.3). The extended shared property would
have appeared only in 3.4 and to allow the migration you would have had to
upgrade all the nodes.

But anyway since we were also talking about a new 3.3.1 barnch I just went
ahead and covered all cases.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] vdsm live migration errors in latest master

2013-09-24 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Dead Horse" 
> Cc: "" , vdsm-de...@fedorahosted.org, 
> fsimo...@redhat.com, aba...@redhat.com
> Sent: Tuesday, September 24, 2013 11:44:48 AM
> Subject: Re: [Users] vdsm live migration errors in latest master
> 
> On Mon, Sep 23, 2013 at 04:05:34PM -0500, Dead Horse wrote:
> > Seeing failed live migrations and these errors in the vdsm logs with latest
> > VDSM/Engine master.
> > Hosts are EL6.4
> 
> Thanks for posting this report.
> 
> The log is from the source of migration, right?
> Could you trace the history of the hosts of this VM? Could it be that it
> was started on an older version of vdsm (say ovirt-3.3.0) and then (due
> to migration or vdsm upgrade) got into a host with a much newer vdsm?
> 
> Would you share the vmCreate (or vmMigrationCreate) line for this Vm in
> your log? I smells like an unintended regression of
> http://gerrit.ovirt.org/17714
> vm: extend shared property to support locking
> 
> solving it may not be trivial, as we should not call
> _normalizeDriveSharedAttribute() automatically on migration destination,
> as it may well still be apart of a 3.3 clusterLevel.
> 
> Also, migration from vdsm with extended shared property, to an ovirt 3.3
> vdsm is going to explode (in a different way), since the destination
> does not expect the extended values.
> 
> Federico, do we have a choice but to revert that patch, and use
> something like "shared3" property instead?

I filed a bug at:

https://bugzilla.redhat.com/show_bug.cgi?id=1011608

A possible fix could be:

http://gerrit.ovirt.org/#/c/19509

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt 3.3.0-4/F19 - Extending VM disk gives correct size but appears to wipe the drive contents

2013-09-23 Thread Federico Simoncelli
Hi Chris,
 can you post the vdsm logs (spm host) somewhere?
Thanks.

-- 
Federico

- Original Message -
> From: "Chris SULLIVAN (WGK)" 
> To: users@ovirt.org
> Sent: Monday, September 23, 2013 9:08:26 PM
> Subject: [Users] oVirt 3.3.0-4/F19 - Extending VM disk gives correct size but 
> appears to wipe the drive contents
> 
> Hi,
> 
> I had a number of Windows VMs running in oVirt 3.3 that required their
> preallocated OS disks to be extended. Each OS disk had a single partition
> taking up the entire drive. As per
> http://www.ovirt.org/Features/Online_Virtual_Drive_Resize I shut down all
> the VMs, extended each OS disk by 10GB (total 25GB) via the web interface,
> then clicked OK. The tasks appeared to complete successfully and each of the
> OS disks had the expected real size on the Gluster storage volume.
> 
> 
> 
> On startup however none of the VMs would recognize their OS disk as being a
> bootable device. Checking one of the OS disks via TestDisk (both quick and
> deep scans) revealed no partitions and the error ‘Partition sector doesn’t
> have the endmark 0xAA55’. It appears that each OS disk was wiped as part of
> the extension process although I’m really hoping that this isn’t the case!
> 
> 
> 
> Are there any other approaches I could use to attempt to recover the OS disk
> data or at least verify whether the original disk partitions are
> recoverable?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Unable to attach to storage domain (Ovirt 3.2)

2013-09-22 Thread Federico Simoncelli
Hi Dan, it looks like one of the domains is missing:

6cf7e7e9-3ae5-4645-a29c-fb17ecb38a50

Is there any target missing? (disconnected or somehow faulty or
unreachable)

-- 
Federico

- Original Message -
> From: "Dan Ferris" 
> To: users@ovirt.org
> Sent: Friday, September 20, 2013 4:01:06 AM
> Subject: [Users] Unable to attach to storage domain (Ovirt 3.2)
> 
> Hi,
> 
> This is my first post to the list.  I am happy to say that we have been
> using Ovirt for 6 months with a few bumps, but it's mostly been ok.
> 
> Until tonight that is...
> 
> I had to do a maintenance that required rebooting both of our Hypervisor
> nodes.  Both of them run Fedora Core 18 and have been happy for months.
>   After rebooting them tonight, they will not attach to the storage.  If
> it matters, the storage is a server running LIO with a Fibre Channel target.
> 
> Vdsm log:
> 
> Thread-22::DEBUG::2013-09-19
> 21:57:09,392::misc::84::Storage.Misc.excCmd::() '/usr/bin/dd
> iflag=direct if=/dev/b358e46b-635b-4c0e-8e73-0a494602e21d/metadata
> bs=4096 count=1' (cwd None)
> Thread-22::DEBUG::2013-09-19
> 21:57:09,400::misc::84::Storage.Misc.excCmd::() SUCCESS:  =
> '1+0 records in\n1+0 records out\n4096 bytes (4.1 kB) copied,
> 0.000547161 s, 7.5 MB/s\n';  = 0
> Thread-23::DEBUG::2013-09-19
> 21:57:16,587::lvm::368::OperationMutex::(_reloadvgs) Operation 'lvm
> reload operation' got the operation mutex
> Thread-23::DEBUG::2013-09-19
> 21:57:16,587::misc::84::Storage.Misc.excCmd::() u'/usr/bin/sudo
> -n /sbin/lvm vgs --config " devices { preferred_names =
> [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0
> disable_after_error_count=3 filter = [
> \\"a%360014055193f840cb3743f9befef7aa3%\\", \\"r%.*%\\" ] }  global {
> locking_type=1  prioritise_write_locks=1  wait_for_locks=1 }  backup {
> retain_min = 50  retain_days = 0 } " --noheadings --units b --nosuffix
> --separator | -o
> uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free
> 6cf7e7e9-3ae5-4645-a29c-fb17ecb38a50' (cwd None)
> Thread-23::DEBUG::2013-09-19
> 21:57:16,643::misc::84::Storage.Misc.excCmd::() FAILED:  =
> '  Volume group "6cf7e7e9-3ae5-4645-a29c-fb17ecb38a50" not found\n';
>  = 5
> Thread-23::WARNING::2013-09-19
> 21:57:16,649::lvm::373::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 []
> ['  Volume group "6cf7e7e9-3ae5-4645-a29c-fb17ecb38a50" not found']
> Thread-23::DEBUG::2013-09-19
> 21:57:16,649::lvm::397::OperationMutex::(_reloadvgs) Operation 'lvm
> reload operation' released the operation mutex
> Thread-23::ERROR::2013-09-19
> 21:57:16,650::domainMonitor::208::Storage.DomainMonitorThread::(_monitorDomain)
> Error while collecting domain 6cf7e7e9-3ae5-4645-a29c-fb17ecb38a50
> monitoring information
> Traceback (most recent call last):
>File "/usr/share/vdsm/storage/domainMonitor.py", line 182, in
> _monitorDomain
>  self.domain = sdCache.produce(self.sdUUID)
>File "/usr/share/vdsm/storage/sdc.py", line 97, in produce
>  domain.getRealDomain()
>File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
>  return self._cache._realProduce(self._sdUUID)
>File "/usr/share/vdsm/storage/sdc.py", line 121, in _realProduce
>  domain = self._findDomain(sdUUID)
>File "/usr/share/vdsm/storage/sdc.py", line 152, in _findDomain
>  raise se.StorageDomainDoesNotExist(sdUUID)
> StorageDomainDoesNotExist: Storage domain does not exist:
> (u'6cf7e7e9-3ae5-4645-a29c-fb17ecb38a50',)
> 
> vgs output (Note that I don't know what the device
> (Wy3Ymi-J7bJ-hVxg-sg3L-F5Gv-MQmz-Utwv7z is) :
> 
> [root@node01 vdsm]# vgs
>Couldn't find device with uuid Wy3Ymi-J7bJ-hVxg-sg3L-F5Gv-MQmz-Utwv7z.
>VG   #PV #LV #SN Attr   VSize   VFree
>b358e46b-635b-4c0e-8e73-0a494602e21d   1  39   0 wz--n-   8.19t  5.88t
>build  2   2   0 wz-pn- 299.75g 16.00m
>fedora 1   3   0 wz--n- 557.88g 0
> 
> lvs output:
> 
> [root@node01 vdsm]# lvs
>Couldn't find device with uuid Wy3Ymi-J7bJ-hVxg-sg3L-F5Gv-MQmz-Utwv7z.
>LV   VG
>   Attr  LSizePool Origin Data%  Move Log Copy%  Convert
>0b8cca47-313f-48da-84f2-154810790d5a
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a   40.00g
>0f6f7572-8797-4d84-831b-87dbc4e1aa48
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a  100.00g
>19a1473f-c375-411f-9a02-c6054b9a28d2
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a   50.00g
>221144dc-51dc-46ae-9399-c0b8e030f38a
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a   40.00g
>2386932f-5f68-46e1-99a4-e96c944ac21b
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a   40.00g
>3e027010-931b-43d6-9c9f-eeeabbdcd47a
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a2.00g
>4257ccc2-94d5-4d71-b21a-c188acbf7ca1
> b358e46b-635b-4c0e-8e73-0a494602e21d -wi-a  200.00g
>4979b2a4-04aa-46a1-be0d-f10be0a1f587
> b358e46b-635b-4c0e-8e73-0

Re: [Users] NFS troubleshooting page

2013-09-11 Thread Federico Simoncelli
- Original Message -
> From: "Itamar Heim" 
> To: "Markus Stockhausen" 
> Cc: "Ayal Baron" , "Federico Simoncelli" 
> , users@ovirt.org, "Allon Mureinik"
> 
> Sent: Monday, September 9, 2013 3:45:56 PM
> Subject: Re: AW: [Users] NFS troubleshooting page
> 
> On 09/09/2013 04:42 PM, Markus Stockhausen wrote:
> >> ayal/federico - thoughts on how we can make things better? warn? etc?
> >
> > Got my account and already modified the wiki.
> >
> > Thanks for the help.
> >
> > Markus
> >
> 
> thanks markus. my question to ayal/federico is on how to fix the
> original issue to at least warn about the issue.

Good question. I don't have an answer yet, but I'll look into it.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt - Glance Integration Deep Dive Session

2013-08-05 Thread Federico Simoncelli
- Original Message -
> From: "Federico Simoncelli" 
> To: "oVirt Users" , "engine-devel" 
> Sent: Monday, July 29, 2013 11:44:21 AM
> Subject: oVirt - Glance Integration Deep Dive Session
> 
> Hi everyone,
>  on Tuesday at 3pm (CEST) I will be presenting the recent work done in
> integrating OpenStack Glance into oVirt 3.3.
> 
> The presentation will include both a high level overview (usage in webadmin)
> and a deep dive about the low level implementation details.
> 
> When:
> Tue 30 Jul 2013 15:00 - 16:00 (CEST)
> 
> Where:
> https://sas.elluminate.com/m.jnlp?sid=819&password=M.9E565882E4EA0288E3479F3D2141BD
> 
> Bridge: 8425973915#
> Phone numbers: http://www.ovirt.org/Intercall

Hi everyone,
  here you are the slides and the recordings:

Slides:
 http://www.ovirt.org/File:Ovirt-2013-glance-integration-deep-dive.pdf

Youtube video:
 http://www.youtube.com/watch?v=_Nyi1xyiQnY

Elluminate recording:
 
https://sas.elluminate.com/p.jnlp?psid=2013-07-30.0631.M.46676CB153495B16DF1807973906F0.vcr&sid=819

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] oVirt - Glance Integration Deep Dive Session

2013-07-29 Thread Federico Simoncelli
BEGIN:VCALENDAR
PRODID:Zimbra-Calendar-Provider
VERSION:2.0
METHOD:REQUEST
BEGIN:VTIMEZONE
TZID:Europe/Berlin
BEGIN:STANDARD
DTSTART:16010101T03
TZOFFSETTO:+0100
TZOFFSETFROM:+0200
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=10;BYDAY=-1SU
TZNAME:CET
END:STANDARD
BEGIN:DAYLIGHT
DTSTART:16010101T02
TZOFFSETTO:+0200
TZOFFSETFROM:+0100
RRULE:FREQ=YEARLY;WKST=MO;INTERVAL=1;BYMONTH=3;BYDAY=-1SU
TZNAME:CEST
END:DAYLIGHT
END:VTIMEZONE
BEGIN:VEVENT
UID:8ea8a413-e9de-46be-be2d-8f83ca22df3a
SUMMARY:oVirt - Glance Integration Deep Dive Session
ATTENDEE;CN=oVirt Users;ROLE=REQ-PARTICIPANT;PARTSTAT=NEEDS-ACTION;RSVP=TRUE
 :mailto:users@ovirt.org
ATTENDEE;CN=engine-devel;ROLE=REQ-PARTICIPANT;PARTSTAT=NEEDS-ACTION;RSVP=TRU
 E:mailto:engine-de...@ovirt.org
ORGANIZER;CN=Federico Simoncelli:mailto:fsimo...@redhat.com
DTSTART;TZID="Europe/Berlin":20130730T15
DTEND;TZID="Europe/Berlin":20130730T16
STATUS:CONFIRMED
CLASS:PUBLIC
X-MICROSOFT-CDO-INTENDEDSTATUS:BUSY
TRANSP:OPAQUE
LAST-MODIFIED:20130729T094421Z
DTSTAMP:20130729T094421Z
SEQUENCE:0
DESCRIPTION:The following is a new meeting request:\n\nSubject: oVirt - Glan
 ce Integration Deep Dive Session \nOrganizer: "Federico Simoncelli"  \n\nTime: Tuesday\, July 30\, 2013\, 3:00:00 PM - 4:00:00 PM G
 MT +01:00 Amsterdam\, Berlin\, Bern\, Rome\, Stockholm\, Vienna\n \nInvitees
 : users@ovirt.org\; engine-de...@ovirt.org \n\n\n*~*~*~*~*~*~*~*~*~*\n\nHi e
 veryone\,\n on Tuesday at 3pm (CEST) I will be presenting the recent work do
 ne in\nintegrating OpenStack Glance into oVirt 3.3.\n\nThe presentation will
  include both a high level overview (usage in webadmin)\nand a deep dive abo
 ut the low level implementation details.\n\nWhen:\nTue 30 Jul 2013 15:00 - 1
 6:00 (CEST)\n\nWhere:\nhttps://sas.elluminate.com/m.jnlp?sid=819&password=M.
 9E565882E4EA0288E3479F3D2141BD\n\nBridge: 8425973915#\nPhone numbers: http:/
 /www.ovirt.org/Intercall\n\n-- \nFederico\n
BEGIN:VALARM
ACTION:DISPLAY
TRIGGER;RELATED=START:-PT5M
DESCRIPTION:Reminder
END:VALARM
END:VEVENT
END:VCALENDAR___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] VM crashes and doesn't recover

2013-04-09 Thread Federico Simoncelli
- Original Message -
> From: "Yuval M" 
> To: "Dan Kenigsberg" 
> Cc: users@ovirt.org, "Nezer Zaidenberg" 
> Sent: Friday, March 29, 2013 2:19:43 PM
> Subject: Re: [Users] VM crashes and doesn't recover
> 
> Any ideas on what can cause that storage crash?
> could it be related to using a SSD?

What the logs say is that the IO on the storage domain are failing (both
the oop timeouts and the sanlock log) and this triggers the VDSM restart.

> On Sun, Mar 24, 2013 at 09:50:02PM +0200, Yuval M wrote:
> > I am running vdsm from packages as my interest is in developing for the
> > I noticed that when the storage domain crashes I can't even do "df -h"
> > (hangs)

This is also consistent with the unreachable domain.

The dmesg log that you attached doesn't contain timestamps so it's hard to
correlate with the rest.

If you want you can try to reproduce the issue and resubmit the logs:

/var/log/vdsm/vdsm.log
/var/log/sanlock.log
/var/log/messages

(Maybe stating also the exact time when the issue begins to appear)

In the logs I noticed that you're using only one NFS domain, and I think that
the SSD (on the storage side) shouldn't be a problem. When you experience such
failure are you able to read/write from/to the SSD on machine that is serving
the share? (If it's the same machine check that using the "real" path where
it's mounted, not the nfs share)

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Failing to attach NFS data storage domain (Ovirt 3.2)

2013-03-20 Thread Federico Simoncelli
- Original Message -
> From: "Limor Gavish" 
> To: "Federico Simoncelli" 
> Cc: "Yuval M" , users@ovirt.org, "Nezer Zaidenberg" 
> , "Eli Mesika"
> , "Maor Lipchuk" 
> Sent: Wednesday, March 20, 2013 9:02:35 PM
> Subject: Re: [Users] Failing to attach NFS data storage domain (Ovirt 3.2)
> 
> Thank you very much for your response.
> 
> 
> Attached VDSM logs as you requested (The VDSM logs where the NFS
> domain was created were missing so we had to recreate the NFS
> domain, therefore the sdUUID has changed).
> Here is the rest of the commands you asked:
> 
> 
> [root@bufferoverflow wil]# mount -t nfs
> bufferoverflow:/home/BO_Ovirt_Storage /mnt/tmp
> [root@bufferoverflow wil]# cd
> /mnt/tmp/1083422e-a5db-41b6-b667-b9ef1ef244f0/dom_md/
> [root@bufferoverflow dom_md]# ls -l
> total 2052
> -rw-rw 1 vdsm kvm 1048576 Mar 20 21:46 ids
> -rw-rw 1 vdsm kvm 0 Mar 20 21:45 inbox
> -rw-rw 1 vdsm kvm 2097152 Mar 20 21:45 leases
> -rw-r--r-- 1 vdsm kvm 311 Mar 20 21:45 metadata
> -rw-rw 1 vdsm kvm 0 Mar 20 21:45 outbox
> [root@bufferoverflow dom_md]# sanlock direct dump ids

Sorry I should have mentioned that if you use root_squash for your nfs
share you have to switch to the vdsm user:

(root)# su -s /bin/sh vdsm
(vdsm)$ cd /mnt/tmp//dom_md/

(vdsm)$ sanlock direct dump ids
(and now you should be able to see the output)

If the output is still empty then used hexdump -C to inspect it
(and eventually post it here compressed).

Another important thing that you should check is:

# ps fax | grep sanlock

If the output doesn't look like the following:

 1966 ?SLs0:00 wdmd -G sanlock
 2036 ?SLsl   0:00 sanlock daemon -U sanlock -G sanlock
 2037 ?S  0:00  \_ sanlock daemon -U sanlock -G sanlock

Then I suggest you to update sanlock to the latest build:

http://koji.fedoraproject.org/koji/buildinfo?buildID=377815
(sanlock-2.6-7.fc18)

And eventually if after rebooting the problem persists, please post
also the sanlock log (/var/log/sanlock.log)

> Please note, the VDSM is running as a system service (it was
> installed from a package) while ovirt-engine was built from sources
> and thus is not running as root. Is this an issue?

It shouldn't be.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Failing to attach NFS data storage domain (Ovirt 3.2)

2013-03-20 Thread Federico Simoncelli
- Original Message -
> From: "Limor Gavish" 
> To: "Eli Mesika" 
> Cc: "Yuval M" , users@ovirt.org, "Nezer Zaidenberg" 
> 
> Sent: Wednesday, March 20, 2013 11:47:49 AM
> Subject: Re: [Users] Failing to attach NFS data storage domain (Ovirt 3.2)
> 
> Thank you very much for your reply.
> I attached the vdsm.log

Hi Limor, can you please inspect the status of the NFS mount?

# mkdir /mnt/tmp
# mount -t nfs  /mnt/tmp
# cd /mnt/tmp/1902354b-4c39-4707-ac6c-3637aaf1943b/dom_md

And please report the output of:

# ls -l

# sanlock direct dump ids

Can you also include more vdsm logs? More specifically the ones where the
NFS domain has been created?
(createStorageDomain with sdUUID='1902354b-4c39-4707-ac6c-3637aaf1943b')

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt 3.2 on CentOS with Gluster 3.3

2013-03-13 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Balamurugan Arumugam" , "Federico Simoncelli" 
> , "Mike Burns"
> 
> Cc: "Rob Zwissler" , users@ovirt.org, a...@ovirt.org, 
> "Aravinda VK" , "Ayal
> Baron" 
> Sent: Wednesday, March 13, 2013 9:03:39 PM
> Subject: Re: [Users] oVirt 3.2 on CentOS with Gluster 3.3
> 
> On Mon, Mar 11, 2013 at 12:34:51PM +0200, Dan Kenigsberg wrote:
> > On Mon, Mar 11, 2013 at 06:09:56AM -0400, Balamurugan Arumugam
> > wrote:
> > > > > >>>Rob,
> > > > > >>>
> > > > > >>>It seems that a bug in vdsm code is hiding the real issue.
> > > > > >>>Could you do a
> > > > > >>>
> > > > > >>> sed -i s/ParseError/ElementTree.ParseError
> > > > > >>> /usr/share/vdsm/gluster/cli.py
> > > > > >>>
> > > > > >>>restart vdsmd, and retry?
> > > > > >>>
> > > > > >>>Bala, would you send a patch fixing the ParseError issue
> > > > > >>>(and
> > > > > >>>adding a
> 
> Ok, both issues have fixes which are in the ovirt-3.2 git branch.
> I believe this deserves a respin of vdsm, as having an undeclated
> requirement is impolite.
> 
> Federico, Mike, would you take care for that?

Since we're at it... I have the feeling that this might be important
enough to be backported to 3.2 too:

http://gerrit.ovirt.org/#/c/12178/

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] F18 iSCSI/FC and latest systemd/udev

2013-02-16 Thread Federico Simoncelli
- Original Message -
> From: "Federico Simoncelli" 
> To: "Jeff Bailey" 
> Cc: users@ovirt.org
> Sent: Saturday, February 16, 2013 9:16:07 AM
> Subject: Re: [Users] F18 iSCSI/FC and latest systemd/udev
> 
> - Original Message -
> > From: "Jeff Bailey" 
> > To: users@ovirt.org
> > Sent: Saturday, February 16, 2013 5:17:49 AM
> > Subject: [Users] F18 iSCSI/FC and latest systemd/udev
> > 
> > While not an actual problem with oVirt, the latest systemd/udev
> > packages for F18 (197-1) break permissions on LVM volumes and stop
> > vdsm/qemu/etc from accessing them.  I just downgraded them and
> > everything seems OK but I thought I'd let people know (easier to just
> > avoid rather than repair :) ).  There's a bugzilla for it from a week
> > or two ago but since 3.2 came out I figured a lot more people might be
> > installing it on new F18 installations with all the updates and running
> > into problems.
> 
> Please test and give karma to:
> 
> https://admin.fedoraproject.org/updates/FEDORA-2013-1775/vdsm-4.10.3-7.fc18

Actually the correct package to give karma to is:

https://admin.fedoraproject.org/updates/FEDORA-2013-2581/vdsm-4.10.3-7.fc18

as it obsoleted FEDORA-2013-1775.

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] F18 iSCSI/FC and latest systemd/udev

2013-02-16 Thread Federico Simoncelli
- Original Message -
> From: "Jeff Bailey" 
> To: users@ovirt.org
> Sent: Saturday, February 16, 2013 5:17:49 AM
> Subject: [Users] F18 iSCSI/FC and latest systemd/udev
> 
> While not an actual problem with oVirt, the latest systemd/udev
> packages for F18 (197-1) break permissions on LVM volumes and stop
> vdsm/qemu/etc from accessing them.  I just downgraded them and everything
> seems OK but I thought I'd let people know (easier to just avoid rather
> than repair :) ).  There's a bugzilla for it from a week or two ago but
> since 3.2 came out I figured a lot more people might be installing it on
> new F18 installations with all the updates and running into problems.

Please test and give karma to:

https://admin.fedoraproject.org/updates/FEDORA-2013-1775/vdsm-4.10.3-7.fc18

which is requiring the correct systemd package.
If we reach 3 points of karma the package will be released.

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] oVirt at FOSDEM 2013

2013-02-06 Thread Federico Simoncelli
oVirt recently had a strong presence at FOSDEM 2013, 2nd-3rd of
February in Brussels. The amount of people attending FOSDEM has
been estimated in 5000+ "geeks" from all over the world and the
exposure that the oVirt project had there has been outstanding.

To read (and see) more about oVirt at FOSDEM 2013:

http://simon3z.blogspot.it/2013/02/ovirt-at-fosdem-2013.html

Thank you all and see you soon at the next event!
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] 3.2 beta and f18 host on dell R815 problem

2013-01-31 Thread Federico Simoncelli
Hi Gianluca can you post/attach/provide the output of cpuid?

# cpuid

In case it's not installed it's provided by the rpm:

cpuid-20120601-2.fc18.x86_64

Thanks,
-- 
Federico

- Original Message -
> From: "Gianluca Cecchi" 
> To: "users" 
> Sent: Thursday, January 31, 2013 12:24:41 PM
> Subject: [Users] 3.2 beta and f18 host on dell R815 problem
> 
> during install of server I get this
> Host installation failed. Fix installation issues and try to
> Re-Install
> 
> In deploy log
> 
> 
> 2013-01-31 12:17:30 DEBUG
> otopi.plugins.ovirt_host_deploy.vdsm.hardware
> hardware._isVirtualizationEnabled:144 virtualization support
> GenuineIntel (cpu: False, bios: True)
> 2013-01-31 12:17:30 DEBUG otopi.context context._executeMethod:127
> method exception
> Traceback (most recent call last):
>   File "/tmp/ovirt-SfEARpd3h4/pythonlib/otopi/context.py", line 117,
> in _executeMethod
> method['method']()
>   File
>   "/tmp/ovirt-SfEARpd3h4/otopi-plugins/ovirt-host-deploy/vdsm/hardware.py",
> line 170, in _validate_virtualization
> _('Hardware does not support virtualization')
> RuntimeError: Hardware does not support virtualization
> 2013-01-31 12:17:30 ERROR otopi.context context._executeMethod:136
> Failed to execute stage 'Setup validation': Hardware does not support
> virtualization
> 
> note the GenuineIntel above... ??
> But actually it is AMD
> 
> [root@f18ovn03 ~]# lsmod|grep kvm
> kvm_amd59623  0
> kvm   431794  1 kvm_amd
> 
> cat /proc/cpuinfo
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Using the chrome/chromium browser with oVirt

2012-12-03 Thread Federico Simoncelli
I recently did some work to get the chrome/chromium browser to work
with oVirt to connect to the VMs with spice.

Few patches are needed for the spice-xpi component:

http://lists.freedesktop.org/archives/spice-devel/2012-December/011729.html
http://lists.freedesktop.org/archives/spice-devel/2012-December/011730.html
http://lists.freedesktop.org/archives/spice-devel/2012-December/011731.html

For convenience, a fedora 17 build can be found here:

http://fsimonce.fedorapeople.org/spice-xpi-chromium-f17/

And an engine patch (wip) has also been proposed here:

http://gerrit.ovirt.org/#/c/9679/

At the moment the combination of the spice-xpi build and the webadmin patch
has been successfully tested with both:

chromium-21.0.1180.89-1.fc17.x86_64
firefox-17.0-1.fc17.x86_64

Any feedback is welcome,
thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] could not add local storage domain

2012-11-16 Thread Federico Simoncelli
- Original Message -
> From: "Cristian Falcas" 
> To: "Federico Simoncelli" 
> Cc: "Jorick Astrego" , users@ovirt.org
> Sent: Friday, November 16, 2012 6:47:50 PM
> Subject: Re: [Users] could not add local storage domain
> 
> it's working for me with the latest files.
> 
> Current issues:
> - You need to create the db user as superuser
> - disable selinux.

Can you grab the relevant AVC errors and report them in a bug?

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Testing oVirt nightly

2012-11-16 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Itamar Heim" , "Federico Simoncelli" 
> 
> Cc: "Joop" , users@ovirt.org, "Moran Goldboim" 
> , "Ofer Schreiber"
> 
> Sent: Thursday, November 15, 2012 10:52:39 PM
> Subject: Re: [Users] Testing oVirt nightly
> 
> On Thu, Nov 15, 2012 at 08:03:51PM +0200, Itamar Heim wrote:
> > On 11/15/2012 04:53 PM, Joop wrote:
> > >
> > >>>What I would like is to get a working config consisting of 2
> > >>>storage
> > >>>servers, 2 hosts and a management server. Storage connected to a
> > >>>pair of
> > >>>10G switches and the public side of the servers and VMs
> > >>>connected to a
> > >>>pair of access switches. For that I need:
> > >>>- bonding
> > >>>- separate networks for storage and ovirtmgmt
> > >>>- storage is using gluster
> > >>>Ideal would be to do all configuration from the webui.
> > >>>
> > >>>Items 1 and 2 need a version of the DC/Cluster of 3.2 but then
> > >>>I'm
> > >>>stopped
> > >>>from going any further because the version of vdsmd on the
> > >>>storage
> > >>>server
> > >>>isn't compatible with the DC/Cluster version.
> > >>
> > >>can you please explain what's the issue of compatibility between
> > >>the two
> > >Message in the webui console is
> > >Host st01 is compatible with versions (3.0,3.1) and cannot join
> > >Cluster
> > >Default which is set to version 3.2
> > >
> > >>(vdsm nightly should give a 3.2 compatibility)?
> > >Where does that nightly come from?
> > >On the management server I have:
> > >[root@mgmt01 /]# rpm -aq | grep vdsm
> > >vdsm-bootstrap-4.10.1-0.129.git2c2c228.fc17.noarch
> > >Which is newer than the vdsm on the hosts which is:
> > >[root@st01 ~]# rpm -aq | grep vdsm
> > >vdsm-python-4.10.0-10.fc17.x86_64
> > >vdsm-xmlrpc-4.10.0-10.fc17.noarch
> > >vdsm-gluster-4.10.0-10.fc17.noarch
> > >vdsm-cli-4.10.0-10.fc17.noarch
> > >vdsm-4.10.0-10.fc17.x86_64
> > >vdsm-rest-4.10.0-10.fc17.noarch
> > >
> > >But I just put st01 into maintanance and did a reinstall but no
> > >version
> > >update of vdsm. I still have the vdsm-bootstrap logs in /tmp if
> > >needed.
> > 
> > danken, which version of vdsm upstream is 3.2 compatible?
> > where can users get it?
> 
> We haven't had an ovirt-3.2 release yet.
> Users can build it themselves from master branch's
> vdsm-4.10.2-something.
> Federico, could you build a test-only package for fedora 18, based on
> master?

It looks like Joop is running vdsm on fc17 (from the rpm versions). I just
did a test-only build of vdsm-4.10.1 for fc17 that can be found here:   

http://fsimonce.fedorapeople.org/vdsm-3.2-f17/x86_64/

For fc18 there's already a build here (vdsm-4.10.1-1.gitgf2f6683.fc18):

http://koji.fedoraproject.org/koji/packageinfo?packageID=12944

I intend to do a newer build for fc18 in the next days (if that's what you
were asking).
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] could not add local storage domain

2012-11-15 Thread Federico Simoncelli
Hi Jorick and Cristian,
 the error you posted here looks like an issue we recently fixed (vdsm
on nfs with kernel 3.6).
Anyway it's quite difficult to make a comprehensive list of things to
report and tests to execute. For this particular issue (and not as a
general rule) I suggest you to contact me on IRC (fsimonce on #ovirt
OFTC) so that we can sort out the issue together.

We will report back to the ML our findings so that it will be helpful
for everyone else.

-- 
Federico

- Original Message -
> From: "Jorick Astrego" 
> To: users@ovirt.org
> Sent: Wednesday, November 14, 2012 1:45:21 PM
> Subject: Re: [Users] could not add local storage domain
> 
> - I'm not the original submitter of this issue, but I have exactly
> the same problem with the latest nightly all-in-one installation.
> 
> We don't use public key auth for sshd on this machine so that's not
> the problem. This is what I see in the vdsm.log:
> 

[...]

> Thread-17:: INFO::2012-11-14
> 12:46:14,129::logUtils::37::dispatcher::(wrapper) Run and protect:
> connectStorageServer(domType=4,
> spUUID='----',
> conList=[{'connection': '/data', 'iqn': '', 'portal': '', 'user':
> '', 'password': '**', 'id':
> '----', 'port': ''}], options=None)
> Thread-17::ERROR::2012-11-14
> 12:46:14,212::hsm::2057::Storage.HSM::(connectStorageServer) Could
> not connect to storageServer
> Traceback (most recent call last):
> File "/usr/share/vdsm/storage/hsm.py", line 2054, in
> connectStorageServer
> conObj.connect()
> File "/usr/share/vdsm/storage/storageServer.py", line 462, in connect
> if not self.checkTarget():
> File "/usr/share/vdsm/storage/storageServer.py", line 449, in
> checkTarget
> fileSD.validateDirAccess(self._path))
> File "/usr/share/vdsm/storage/fileSD.py", line 51, in
> validateDirAccess
> getProcPool().fileUtils.validateAccess(dirPath)
> File "/usr/share/vdsm/storage/remoteFileHandler.py", line 274, in
> callCrabRPCFunction
> *args, **kwargs)
> File "/usr/share/vdsm/storage/remoteFileHandler.py", line 180, in
> callCrabRPCFunction
> rawLength = self._recvAll(LENGTH_STRUCT_LENGTH, timeout)
> File "/usr/share/vdsm/storage/remoteFileHandler.py", line 149, in
> _recvAll
> timeLeft):
> File "/usr/lib64/python2.7/contextlib.py", line 84, in helper
> return GeneratorContextManager(func(*args, **kwds))
> File "/usr/share/vdsm/storage/remoteFileHandler.py", line 136, in
> _poll
> raise Timeout()
> Timeout
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] SELinux policy issue with oVirt/sanlock

2012-10-24 Thread Federico Simoncelli
- Original Message -
> From: "Haim Ateya" 
> To: "Brian Vetter" 
> Cc: users@ovirt.org, seli...@lists.fedoraproject.org
> Sent: Wednesday, October 24, 2012 7:03:39 PM
> Subject: Re: [Users] SELinux policy issue with oVirt/sanlock
> 
> - Original Message -
> > From: "Brian Vetter" 
> > To: "Haim Ateya" 
> > Cc: users@ovirt.org, seli...@lists.fedoraproject.org
> > Sent: Wednesday, October 24, 2012 6:24:31 PM
> > Subject: Re: [Users] SELinux policy issue with oVirt/sanlock
> > 
> > I removed lock_manager=sanlock from the settings file, restarted
> > the
> > daemons, and all works fine right now. I'm guessing that means
> > there
> > is no locking of the VMs (the default?).
> 
> that's right, i'm glad it works for you, but it just a workaround
> since we expect this configuration to work, it would be much
> appreciated if you
> could open a bug on that issue so we can track and resolve when
> possible.
> please attach all required logs such as: vdsm.log, libvirtd.log,
> qemu.log (under /var/log/libvirt/qemu/), audit.log, sanlock.log and
> /var/log/messages.

What's the bug number? To clarify/recap:

- the lock_manager=sanlock configuration is correct (and it shouldn't
  be removed)
- you should set setenforce 0 (with lock_manager=sanlock) and try to
  start a VM; all the avc errors that you find in /var/log/messages
  and in /var/log/audit/audit.log should be used to open a selinux
  policy bug

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Error creating the first storage domain (NFS)

2012-10-23 Thread Federico Simoncelli
- Original Message -
> From: "Brian Vetter" 
> To: "Federico Simoncelli" 
> Cc: "Vered Volansky" , users@ovirt.org, "David Teigland" 
> 
> Sent: Wednesday, October 24, 2012 5:48:21 AM
> Subject: Re: [Users] Error creating the first storage domain (NFS)
> 
> Ugh. Spoke a little too soon. While I got past my problem creating a
> storage domain, I ran into a new sanlock issue.
> 
> When trying to run a VM (the first one so I can create a template), I
> get an error in the admin UI:
> 
> VM DCC4.0 is down. Exit message: Failed to acquire lock: Permission
> denied.
> 
> On a lark, I turned off selinux enforcement and tried it again. It
> worked just fine.
> 
> So what selinux option do I need to enable to get it to work? The
> only other sanlock specific settings I saw are:
> 
> sanlock_use_fusefs --> off
> sanlock_use_nfs --> on
> sanlock_use_samba --> off
> 
> Do I turn these all on or is there some other setting I need to
> enable?

No for nfs you just need sanlock_use_nfs.
I'd say that if you could verify the scratch build that I prepared at:

http://koji.fedoraproject.org/koji/taskinfo?taskID=4620480

(up until starting a vm), then all the new selinux errors/messages that
you see in the audit log (/var/log/audit/audit.log) are issues that
should be reported to the selinux-policy package.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Error creating the first storage domain (NFS)

2012-10-23 Thread Federico Simoncelli
- Original Message -
> From: "Brian Vetter" 
> To: "Federico Simoncelli" 
> Cc: "Vered Volansky" , users@ovirt.org, "David Teigland" 
> 
> Sent: Wednesday, October 24, 2012 4:54:11 AM
> Subject: Re: [Users] Error creating the first storage domain (NFS)
> 
> That was the problem. I checked the sanlock_use_nfs boolean and it
> was off. I set it and then created and attached the storage and it
> all works.

Thanks for testing.
Do you have a way of verifying a scratch build?

http://koji.fedoraproject.org/koji/taskinfo?taskID=4620480

This should fix your problem (on a brand new installation).

-- 
Federico

> On Oct 23, 2012, at 8:55 PM, Federico Simoncelli wrote:
> 
> > Hi Brian,
> > I hate progressing by guesses but could you try to disable selinux:
> > 
> > # setenforce 0
> > 
> > If that works you could go on, re-enable it and try something more
> > specific:
> > 
> > # setenforce 1
> > # setsebool sanlock_use_nfs on
> > 
> > I have the feeling that the vdsm patch setting the sanlock_use_nfs
> > sebool flag didn't made it to fedora 17 yet.
> > --
> > Federico
> > 
> > - Original Message -
> >> From: "Brian Vetter" 
> >> To: "Federico Simoncelli" 
> >> Cc: "Vered Volansky" , users@ovirt.org, "David
> >> Teigland" 
> >> Sent: Tuesday, October 23, 2012 6:10:36 PM
> >> Subject: Re: [Users] Error creating the first storage domain (NFS)
> >> 
> >> Ok. Here's four log files:
> >> 
> >> engine.log from my ovirt engine server.
> >> vdsm.log from my host
> >> sanlock.log from my host
> >> messages from my host
> >> 
> >> The errors occur around the 20:17:57 time frame. You might see
> >> other
> >> errors from either previous attempts or for the time after when I
> >> tried to attach the storage domain. It looks like everything
> >> starts
> >> with an error -13 in sanlock. If the -13 maps to 13/EPERM in
> >> errno.h, then it is likely be some kind of permission or other
> >> access error. I saw things that were related to the nfs
> >> directories
> >> not being owned by vdsm:kvm, but that is not the case here.
> >> 
> >> I did see a note online about some issues with sanlock and F17
> >> (which
> >> I am running), but those bugs were related to sanlock crashing.
> >> 
> >> Brian
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Error creating the first storage domain (NFS)

2012-10-23 Thread Federico Simoncelli
Hi Brian,
 I hate progressing by guesses but could you try to disable selinux:

 # setenforce 0

If that works you could go on, re-enable it and try something more
specific:

 # setenforce 1
 # setsebool sanlock_use_nfs on

I have the feeling that the vdsm patch setting the sanlock_use_nfs
sebool flag didn't made it to fedora 17 yet.
-- 
Federico

- Original Message -
> From: "Brian Vetter" 
> To: "Federico Simoncelli" 
> Cc: "Vered Volansky" , users@ovirt.org, "David Teigland" 
> 
> Sent: Tuesday, October 23, 2012 6:10:36 PM
> Subject: Re: [Users] Error creating the first storage domain (NFS)
> 
> Ok. Here's four log files:
> 
> engine.log from my ovirt engine server.
> vdsm.log from my host
> sanlock.log from my host
> messages from my host
> 
> The errors occur around the 20:17:57 time frame. You might see other
> errors from either previous attempts or for the time after when I
> tried to attach the storage domain. It looks like everything starts
> with an error -13 in sanlock. If the -13 maps to 13/EPERM in
> errno.h, then it is likely be some kind of permission or other
> access error. I saw things that were related to the nfs directories
> not being owned by vdsm:kvm, but that is not the case here.
> 
> I did see a note online about some issues with sanlock and F17 (which
> I am running), but those bugs were related to sanlock crashing.
> 
> Brian
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Error creating the first storage domain (NFS)

2012-10-23 Thread Federico Simoncelli
- Original Message -
> From: "Vered Volansky" 
> To: "Brian Vetter" 
> Cc: users@ovirt.org
> Sent: Tuesday, October 23, 2012 11:38:42 AM
> Subject: Re: [Users] Error creating the first storage domain (NFS)
> 
> Hi Brian,
> 
> We'll need your engine & host (full) logs at the very least to look
> into the problem.
> Can you try it with nfs3 and tell us if it works?
> 
> Note, more comments in the email body.
> 
> Regards,
> Vered

Hi Brian,
 we also need the sanlock logs (/var/log/sanlock.log).

Adding David to the thread as he might be able to help us debugging
your problem (sanlock-2.4-2.fc17).

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Can't start a VM - sanlock permission denied

2012-10-15 Thread Federico Simoncelli
- Original Message -
> From: "Dan Kenigsberg" 
> To: "Mike Burns" 
> Cc: "Federico Simoncelli" , users@ovirt.org
> Sent: Monday, October 15, 2012 11:02:45 AM
> Subject: Re: [Users] Can't start a VM - sanlock permission denied
> 
> On Sun, Oct 14, 2012 at 09:53:51PM -0400, Mike Burns wrote:
> > On Sun, 2012-10-14 at 19:11 -0400, Federico Simoncelli wrote:
> > > - Original Message -
> > > > From: "Alexandre Santos" 
> > > > To: "Dan Kenigsberg" 
> > > > Cc: "Haim Ateya" , users@ovirt.org,
> > > > "Federico Simoncelli" 
> > > > Sent: Sunday, October 14, 2012 7:23:36 PM
> > > > Subject: Re: [Users] Can't start a VM - sanlock permission
> > > > denied
> > > > 
> > > > 2012/10/13 Dan Kenigsberg < dan...@redhat.com >
> > > > 
> > > > On Sat, Oct 13, 2012 at 11:25:37AM +0100, Alexandre Santos
> > > > wrote:
> > > > > Hi,
> > > > > after getting to the oVirt Node console (F2) I figured out
> > > > > that
> > > > > selinux
> > > > > wasn't allowing the sanlock, so I entered the setsebool
> > > > > virt_use_sanlock 1
> > > > > and the problem is fixed.
> > > > 
> > > > Which version of vdsm is istalled on your node? and which
> > > > selinux-policy? sanlock should work out-of-the-box.
> > > > 
> > > > 
> > > > vdsm-4.10.0-10.fc17
> > > > 
> > > > on /etc/sysconfig/selinux
> > > > SELINUX=enforcing
> > > > SELINUXTYPE=targeted
> > > 
> > > As far as I understand the selinux policies for the ovirt-node
> > > are set
> > > by recipe/common-post.ks (in the ovirt-node repo):
> > > 
> > > semanage  boolean -m -S targeted -F /dev/stdin  << \EOF_semanage
> > > allow_execstack=0
> > > virt_use_nfs=1
> > > EOF_semanage
> > > 
> > > We should update it with what vdsm is currently setting:
> > > 
> > > virt_use_sanlock=1
> > > sanlock_use_nfs=1
> > > 
> > 
> > Shouldn't vdsm be setting these if they're needed?
> 
> It should - I'd like to know which vdsm version was it, and why this
> was skipped.

The version was 4.10.0-10.fc17 and what I thought (but I didn't test yesterday
night) is that the ovirt-node was overriding what we were setting.
Anyway this is not the case.

> > I can certainly set
> > the values, but IMO, if vdsm needs it, vdsm should set it.
> 
> virt_use_nfs=1 made it into the node. Maybe there was a good reason
> for it that applies to virt_use_sanlock as well. (I really hate to
> persist the policy files, and dislike the idea of setting virt_use_sanlock
> every time vdsmd starts - it's slow).

We set them when we install vdsm (not when the service starts) so they should
be good to go in the iso. It might be a glitch during the vdsm package
installation, it could be something like semanage taking the boolean from the
host where the iso is built rather than the root where the package is installed.

Do we have the iso build logs?

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Can't start a VM - sanlock permission denied

2012-10-14 Thread Federico Simoncelli
- Original Message -
> From: "Alexandre Santos" 
> To: "Dan Kenigsberg" 
> Cc: "Haim Ateya" , users@ovirt.org, "Federico Simoncelli" 
> 
> Sent: Sunday, October 14, 2012 7:23:36 PM
> Subject: Re: [Users] Can't start a VM - sanlock permission denied
> 
> 2012/10/13 Dan Kenigsberg < dan...@redhat.com >
> 
> On Sat, Oct 13, 2012 at 11:25:37AM +0100, Alexandre Santos wrote:
> > Hi,
> > after getting to the oVirt Node console (F2) I figured out that
> > selinux
> > wasn't allowing the sanlock, so I entered the setsebool
> > virt_use_sanlock 1
> > and the problem is fixed.
> 
> Which version of vdsm is istalled on your node? and which
> selinux-policy? sanlock should work out-of-the-box.
> 
> 
> vdsm-4.10.0-10.fc17
> 
> on /etc/sysconfig/selinux
> SELINUX=enforcing
> SELINUXTYPE=targeted

As far as I understand the selinux policies for the ovirt-node are set
by recipe/common-post.ks (in the ovirt-node repo):

semanage  boolean -m -S targeted -F /dev/stdin  << \EOF_semanage
allow_execstack=0
virt_use_nfs=1
EOF_semanage

We should update it with what vdsm is currently setting:

virt_use_sanlock=1
sanlock_use_nfs=1

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Fwd: Re: oVirt live snapshot problem

2012-06-13 Thread Federico Simoncelli
- Original Message -
> From: "Haim Ateya" 
> To: "Neil" , "Federico Simoncelli" 
> , "Kiril Nesenko" 
> Cc: users@ovirt.org
> Sent: Wednesday, June 13, 2012 7:00:34 AM
> Subject: Re: [Users] Fwd: Re: oVirt live snapshot problem
> 
> Federico\Kiril,
> 
> is this problem known to you ?

I can't say without the logs. Please look at the relevant log parts that
I quote in my bug comment and check if they are the same:

https://bugzilla.redhat.com/show_bug.cgi?id=829645#c3

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] image ownership

2012-05-10 Thread Federico Simoncelli
- Original Message -
> From: "Jacob Wyatt" 
> To: users@ovirt.org
> Sent: Wednesday, May 9, 2012 8:08:54 PM
> Subject: [Users] image ownership
> 
> Greetings all,
> 
> I've set up a new oVirt installation and it's behaving strangely with
> regard to virtual machine image files on the NFS storage.  Whenever
> I shut down a machine it's changing the owner of the image to
> root:root (0:0) instead of vdsm:kvm (36:36).  After that it can't
> start or do anything with that image again until I manually change
> the ownership back.  Everything works fine again until I shut the
> machine down.  I assume this is some mistake I've made in
> installation.  I did not have this problem in the test environment,
> but I'm stumped as to what went wrong.
> 
> -Jacob

Hi Jacob,
 could you check the dynamic_ownership in /etc/libvirt/qemu.conf:

# grep dynamic_ownership /etc/libvirt/qemu.conf 
#dynamic_ownership = 1
dynamic_ownership=0 # by vdsm

Thanks,
-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] error in multipath.py

2012-04-24 Thread Federico Simoncelli
Could you check if this fixes your problem?

http://gerrit.ovirt.org/3863

Thanks,
-- 
Federico

- Original Message -
> From: ov...@qip.ru
> To: users@ovirt.org
> Sent: Tuesday, April 24, 2012 8:54:58 AM
> Subject: [Users] error in multipath.py
> 
> 
> i tried the latest releases of vdsm from jenkins . ovirt .org, but
> found that they did'n work. V dsm d cycles and corrupted my
> multipath . conf .
> 
> error is in
> 
> / usr /share/ vdsm /storage/ multipath .py
> 
> this is the diff with my py
> 
> # diff multipath .py.bad multipath .py
> 88,89c88,89
> < first = mpath conf [0]
> < second = mpath conf [1]
> ---
> > first = mpath conf .split('\n', 1)[0]
> > second = mpath conf .split('\n', 1)[1]
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Can't start vm

2012-03-12 Thread Federico Simoncelli


- Original Message -
> From: "kumar shantanu" 
> To: "users" 
> Sent: Monday, March 12, 2012 9:16:34 AM
> Subject: [Users] Can't start vm
> 
> 
> Hi all,
> 
> I created host from ovirt manager but when trying to run it's failing
> with the error,
> 
> ==> vdsm.log <==
> Thread-180651::DEBUG::2012-03-12
> 13:42:12,482::vm::577::vm.Vm::(_startUnderlyingVm)
> vmId=`c13c4c09-f696-47e1-b8cd-8d499242e151`::_ongoingCreations
> released
> Thread-180651::ERROR::2012-03-12
> 13:42:12,482::vm::601::vm.Vm::(_startUnderlyingVm)
> vmId=`c13c4c09-f696-47e1-b8cd-8d499242e151`::The vm start process
> failed
> Traceback (most recent call last):
> File "/usr/share/vdsm/vm.py", line 567, in _startUnderlyingVm
> self._run()
> File "/usr/share/vdsm/libvirtvm.py", line 1306, in _run
> self._connection.createXML(domxml, flags),
> File "/usr/share/vdsm/libvirtconnection.py", line 82, in wrapper
> ret = f(*args, **kwargs)
> File "/usr/lib64/python2.6/site-packages/libvirt.py", line 2087, in
> createXML
> if ret is None:raise libvirtError('virDomainCreateXML() failed',
> conn=self)
> libvirtError: internal error Process exited while reading console log
> output: Supported machines are:
> pc RHEL 6.2.0 PC (alias of rhel6.2.0)
> 
> 
> 
> Pythong version running is
> 
> [root@ovirt ~]# python -V
> Python 2.7
> 
> Can anyone please suggest .

Hi Kumar,
 when the engine starts a VM it also specifies a machine type.
The machine types supported by an host depend on the system (RHEL/Fedora)
and you can get the list with:

 # vdsClient 0 getVdsCaps | grep emulatedMachines
emulatedMachines = ['pc-1.1', 'pc', 'pc-1.0', 'pc-0.15', ...

Once you discovered the types supported by your hosts you can configure
the engine with the correct value:

http://www.ovirt.org/wiki/Engine_Node_Integration

psql -U postgres engine -c "update vdc_options set option_value='pc-0.14' 
 where option_name='EmulatedMachine' and version='3.0';"

I assume that you ran the command above but your VDSM hosts are rhel6,
so you would need to use the "rhel6.2.0" value instead.

I believe that the value "pc" is an alias that works both for RHEL and
Fedora and it might be handy for testing, but in general I really discourage
its use because it would allow a mixed cluster of RHEL and Fedora hosts
which could be problematic in case of live migrations.

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] FOSDEM 2012: VDSM - The oVirt Node Management Agent

2012-02-07 Thread Federico Simoncelli
Here you can find my slides presented at FOSDEM 2012:

http://people.redhat.com/fsimonce/fosdem-2012/

-- 
Federico
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users