Re: [ovirt-users] Ovirt Hypervisor vdsm.Scheduler logs fill partition

2016-10-23 Thread Steve Dainard
Do you know when .34 will be released?

http://mirror.centos.org/centos/7/virt/x86_64/ovirt-3.6/
Latest version is:
vdsm-cli-4.17.32-1.el7.noarch.rpm 08-Aug-2016 17:36

On Fri, Oct 14, 2016 at 1:11 AM, Francesco Romani <from...@redhat.com>
wrote:

>
> - Original Message -
> > From: "Simone Tiraboschi" <stira...@redhat.com>
> > To: "Steve Dainard" <sdain...@spd1.com>, "Francesco Romani" <
> from...@redhat.com>
> > Cc: "users" <users@ovirt.org>
> > Sent: Friday, October 14, 2016 9:59:49 AM
> > Subject: Re: [ovirt-users] Ovirt Hypervisor vdsm.Scheduler logs fill
> partition
> >
> > On Fri, Oct 14, 2016 at 1:12 AM, Steve Dainard <sdain...@spd1.com>
> wrote:
> >
> > > Hello,
> > >
> > > I had a hypervisor semi-crash this week, 4 of ~10 VM's continued to
> run,
> > > but the others were killed off somehow and all VM's running on this
> host
> > > had '?' status in the ovirt UI.
> > >
> > > This appears to have been caused by vdsm logs filling up disk space on
> the
> > > logging partition.
> > >
> > > I've attached the log file vdsm.log.27.xz which shows this error:
> > >
> > > vdsm.Scheduler::DEBUG::2016-10-11
> > > 16:42:09,318::executor::216::Executor::(_discard)
> > > Worker discarded:  > > action= > > 'virt.periodic.DriveWatermarkMonitor'>
> > > at 0x7f8e90021210> at 0x7f8e90021250> discarded at 0x7f8dd123e850>
> > >
> > > which happens more and more frequently throughout the log.
> > >
> > > It was a bit difficult to understand what caused the failure, but the
> logs
> > > were getting really large, then being xz'd which compressed 11G+ into
> a few
> > > MB. Once this happened the disk space would be freed, and nagios
> wouldn't
> > > hit the 3rd check to throw a warning, until pretty much right at the
> crash.
> > >
> > > I was able to restart vdsmd to resolve the issue, but I still need to
> know
> > > why these logs started to stack up so I can avoid this issue in the
> future.
> > >
> >
> > We had this one: https://bugzilla.redhat.com/show_bug.cgi?id=1383259
> > but in your case the logs are rotating.
> > Francesco?
>
> Hi,
>
> yes, it is a different issue. Here the log messages are caused by the
> Worker threads
> of the periodic subsystem, which are leaking[1].
> This was a bug in Vdsm (insufficient protection against rogue domains),
> but the
> real problem is that some of your domain are being unresponsive at
> hypervisor level.
> The most likely cause is in turn unresponsive storages.
>
> Fixes are been committed and shipped with Vdsm 4.17.34.
>
> See: ttps://bugzilla.redhat.com/1364925
>
> HTH,
>
> +++
>
> [1] actually, they are replaced too quickly, leading to unbound growth.
> So those aren't actually "leaking", Vdsm is just overzealous handling one
> error condition,
> making things worse than before.
> Still serious issue, no doubt, but quite different cause.
>
> --
> Francesco Romani
> Red Hat Engineering Virtualization R & D
> Phone: 8261328
> IRC: fromani
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Ovirt 3.5 host gluster storage connection failure

2015-12-22 Thread Steve Dainard
I have two hosts, only one of them was running VM's at the time of
this crash so I can't tell if this is a node specific problem.

rpm -qa | egrep -i 'gluster|vdsm|libvirt' |sort
glusterfs-3.6.7-1.el7.x86_64
glusterfs-api-3.6.7-1.el7.x86_64
glusterfs-cli-3.6.7-1.el7.x86_64
glusterfs-fuse-3.6.7-1.el7.x86_64
glusterfs-libs-3.6.7-1.el7.x86_64
glusterfs-rdma-3.6.7-1.el7.x86_64
libvirt-client-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-config-nwfilter-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-interface-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-network-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-nodedev-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-nwfilter-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-qemu-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-secret-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-driver-storage-1.2.8-16.el7_1.5.x86_64
libvirt-daemon-kvm-1.2.8-16.el7_1.5.x86_64
libvirt-lock-sanlock-1.2.8-16.el7_1.5.x86_64
libvirt-python-1.2.8-7.el7_1.1.x86_64
vdsm-4.16.30-0.el7.centos.x86_64
vdsm-cli-4.16.30-0.el7.centos.noarch
vdsm-jsonrpc-4.16.30-0.el7.centos.noarch
vdsm-python-4.16.30-0.el7.centos.noarch
vdsm-python-zombiereaper-4.16.30-0.el7.centos.noarch
vdsm-xmlrpc-4.16.30-0.el7.centos.noarch
vdsm-yajsonrpc-4.16.30-0.el7.centos.noarch

VM's were in a paused state, with errors in UI:

2015-Dec-22, 15:06
VM pcic-apps has paused due to unknown storage error.
2015-Dec-22, 15:06
Host compute2 is not responding. It will stay in Connecting state for
a grace period of 82 seconds and after that an attempt to fence the
host will be issued.
2015-Dec-22, 15:03
Invalid status on Data Center EDC2. Setting Data Center status to Non
Responsive (On host compute2, Error: General Exception).
2015-Dec-22, 15:03
VM pcic-storage has paused due to unknown storage error.
2015-Dec-22, 15:03
VM docker1 has paused due to unknown storage error.

VDSM logs look normal until:
Dummy-99::DEBUG::2015-12-22
23:03:58,949::storage_mailbox::731::Storage.Misc.excCmd::(_checkForMail)
dd 
if=/rhev/data-center/f72ec125-69a1-4c1b-a5e1-313fcb70b6ff/mastersd/dom_md/inbox
iflag=direct,fullblock count=1 bs=1024000 (cwd None)
Dummy-99::DEBUG::2015-12-22
23:03:58,963::storage_mailbox::731::Storage.Misc.excCmd::(_checkForMail)
SUCCESS:  = '1+0 records in\n1+0 records out\n1024000 bytes (1.0
MB) copied, 0.00350501 s, 292 MB/s\n';  = 0
VM Channels Listener::INFO::2015-12-22
23:03:59,527::guestagent::180::vm.Vm::(_handleAPIVersion)
vmId=`7067679e-43aa-43c0-b263-b0a711ade2e2`::Guest API version changed
from 2 to 1
Thread-245428::DEBUG::2015-12-22
23:03:59,718::libvirtconnection::151::root::(wrapper) Unknown
libvirterror: ecode: 80 edom: 20 level: 2 message: metadata not found:
Requested metadata element is not present
libvirtEventLoop::INFO::2015-12-22
23:04:00,447::vm::4982::vm.Vm::(_onIOError)
vmId=`376e98b7-7798-46e8-be03-5dddf6cfb54f`::abnormal vm stop device
virtio-disk0 error eother
libvirtEventLoop::DEBUG::2015-12-22
23:04:00,447::vm::5666::vm.Vm::(_onLibvirtLifecycleEvent)
vmId=`376e98b7-7798-46e8-be03-5dddf6cfb54f`::event Suspended detail 2
opaque None
libvirtEventLoop::INFO::2015-12-22
23:04:00,447::vm::4982::vm.Vm::(_onIOError)
vmId=`376e98b7-7798-46e8-be03-5dddf6cfb54f`::abnormal vm stop device
virtio-disk0 error eother
...
libvirtEventLoop::INFO::2015-12-22
23:04:00,843::vm::4982::vm.Vm::(_onIOError)
vmId=`97fbbf97-944b-4b77-b0bf-6a831f9090d8`::abnormal vm stop device
virtio-disk1 error eother
libvirtEventLoop::DEBUG::2015-12-22
23:04:00,844::vm::5666::vm.Vm::(_onLibvirtLifecycleEvent)
vmId=`97fbbf97-944b-4b77-b0bf-6a831f9090d8`::event Suspended detail 2
opaque None
Dummy-99::DEBUG::2015-12-22
23:04:00,973::storage_mailbox::731::Storage.Misc.excCmd::(_checkForMail)
dd 
if=/rhev/data-center/f72ec125-69a1-4c1b-a5e1-313fcb70b6ff/mastersd/dom_md/inbox
iflag=direct,fullblock count=1 bs=1024000 (cwd None)
Dummy-99::DEBUG::2015-12-22
23:04:00,983::storage_mailbox::731::Storage.Misc.excCmd::(_checkForMail)
FAILED:  = "dd: failed to open
'/rhev/data-center/f72ec125-69a1-4c1b-a5e1-313fcb70b6ff/mastersd/dom_md/inbox':
Transport endpoint is not connected\n";  = 1
Dummy-99::ERROR::2015-12-22
23:04:00,983::storage_mailbox::787::Storage.MailBox.SpmMailMonitor::(run)
Error checking for mail
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/storage_mailbox.py", line 785, in run
self._checkForMail()
  File "/usr/share/vdsm/storage/storage_mailbox.py", line 734, in _checkForMail
"Could not read mailbox: %s" % self._inbox)
IOError: [Errno 5] _handleRequests._checkForMail - Could not read
mailbox: 
/rhev/data-center/f72ec125-69a1-4c1b-a5e1-313fcb70b6ff/mastersd/dom_md/inbox
Dummy-99::DEBUG::2015-12-22
23:04:02,987::storage_mailbox::731::Storage.Misc.excCmd::(_checkForMail)
dd 
if=/rhev/data-center/f72ec125-69a1-4c1b-a5e1-313fcb70b6ff/mastersd/dom_md/inbox
iflag=direct,fullblock count=1 bs=1024000 (cwd None)
Dummy-99::DEBUG::2015-12-22

[ovirt-users] LDAP authentication with TLS

2015-10-06 Thread Steve Dainard
Hello,

Trying to configure Ovirt 3.5.3.1-1.el7.centos for LDAP authentication.

I've configured the appropriate aaa profile but I'm getting TLS errors
 when I search for users to add via ovirt:

The connection reader was unable to successfully complete TLS
negotiation: javax_net_ssl_SSLHandshakeException:
sun_security_validator_ValidatorException: No trusted certificate
found caused by sun_security_validator_ValidatorException: No trusted
certificate found

I added the external CA certificate using keytool as per
https://github.com/oVirt/ovirt-engine-extension-aaa-ldap with
appropriate adjustments of course:

keytool -importcert -noprompt -trustcacerts -alias myrootca \
   -file myrootca.pem -keystore myrootca.jks -storepass changeit

I know this certificate works, and can connect to LDAP with TLS as I'm
using the same LDAP configuration/certificate with SSSD.

Can anyone clarify whether I should be adding the external CA
certificate or the LDAP host certificate with keytool or any other
suggestions?

Thanks,
Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Feature request: add stateless option in VM pool configuration

2015-08-21 Thread Steve Dainard
I'd like to request a setting in VM pools which forces the VM to be stateless.

I see from the docs this will occur if selected under run once, or
started from the user portal, but it would be nice to set this for a
normal admin start of the VM(s) as well.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Feature request: add stateless option in VM pool configuration

2015-08-21 Thread Steve Dainard
I should also mention that I can't start multiple pool VM's at the
same time with run-once from the admin portal, so I'd have to
individually run once each VM which is a member of the pool.

On Fri, Aug 21, 2015 at 3:57 PM, Steve Dainard sdain...@spd1.com wrote:
 I'd like to request a setting in VM pools which forces the VM to be stateless.

 I see from the docs this will occur if selected under run once, or
 started from the user portal, but it would be nice to set this for a
 normal admin start of the VM(s) as well.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hostusb hook - VM device errors in Windows VM

2014-07-25 Thread Steve Dainard
Any other ideas here? Is there a specific driver I should load instead of
the Windows default one?

Thanks,
Steve


On Tue, Jul 22, 2014 at 10:23 AM, Steve Dainard sdain...@miovision.com
wrote:

 I just saw the your device can perform faster warning again in Windows
 and decided to check it out.

 Should the USB device be showing under an Intel controller? Is the RH
 controller only for spice?


 ​


 On Tue, Jul 22, 2014 at 10:12 AM, Dan Kenigsberg dan...@redhat.com
 wrote:

 On Tue, Jul 22, 2014 at 03:50:59PM +0200, Michal Skrivanek wrote:
 
  On Jul 22, 2014, at 15:49 , Steve Dainard sdain...@miovision.com
 wrote:
 
   Hi Michal,
  
   How can I generate libvirt xml from rhevm?
 
  virsh -r dumpxml domain on the host

 Or dig into vdsm.log (in case the VM is no longer there)



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hostusb hook - VM device errors in Windows VM

2014-07-22 Thread Steve Dainard
Hi Michal,

How can I generate libvirt xml from rhevm?

Thanks,
Steve


On Tue, Jul 22, 2014 at 4:12 AM, Michal Skrivanek 
michal.skriva...@redhat.com wrote:


 On 21 Jul 2014, at 20:54, Steve Dainard wrote:

 I should mention I can mount this usb drive in a CentOS 6.5 VM without any
 problems.


 Hi,
 there should be no difference configuration-wise. well, please compare
 libvirt's xml to be sure and confirm
 If it's the case then it might be a problem of qemu/kvm and/or windows

 Thanks,
 michal



 On Mon, Jul 21, 2014 at 2:11 PM, Steve Dainard sdain...@miovision.com
 wrote:

 I'm using the hostusb hook on RHEV 3.4 trial.

 The usb device is passed through to the VM, but I'm getting errors in a
 Windows VM when the device driver is loaded.

 I started with a simple usb drive, on the host it is listed as:

 Bus 002 Device 010: ID 05dc:c75c Lexar Media, Inc.

 Which I added as 0x05dc:0xc75c to the Windows 7 x64 VM.

 In Windows I get an error in device manager:
 USB Mass Storage Device This device cannot start. (Code 10)
 Properties/General Tab: Device type: Universal Serial Bus Controllers,
 Manufacturer: Compatible USB storage device, Location: Port_#0001.Hub_#0001

 Under hardware Ids:
 USB\VID_05DCPID_C75CREV_0102
 USB\VID_05DCPID_C75C

 So it looks like the proper USB device ID is passed to the VM.

 I don't see any error messages in event viewer, and I don't see anything
 in VDSM logs either.

 Any help is appreciated.

 Steve


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hostusb hook - VM device errors in Windows VM

2014-07-22 Thread Steve Dainard
The USB section is the same on both guests:

controller type='usb' index='0'
  alias name='usb0'/
  address type='pci' domain='0x' bus='0x00' slot='0x01'
function='0x2'/
/controller

hostdev mode='subsystem' type='usb' managed='no'
  source
vendor id='0x05dc'/
product id='0xc75c'/
address bus='2' device='10'/
  /source
  alias name='hostdev0'/
/hostdev

I thought it might be a Windows issue, I've tried on both Win7 64 and
2008R2 64 with the same result. At one point there was a notification
bubble that the USB device can run faster - which I thought might mean the
Windows guest was provided a USB v1 controller.

I've attached both the xml's if you think there might be something I'm not
seeing.

Thanks,
Steve


On Tue, Jul 22, 2014 at 9:50 AM, Michal Skrivanek 
michal.skriva...@redhat.com wrote:


 On Jul 22, 2014, at 15:49 , Steve Dainard sdain...@miovision.com wrote:

  Hi Michal,
 
  How can I generate libvirt xml from rhevm?

 virsh -r dumpxml domain on the host

 Thanks,
 michal

 
  Thanks,
  Steve
 
 
  On Tue, Jul 22, 2014 at 4:12 AM, Michal Skrivanek 
 michal.skriva...@redhat.com wrote:
 
  On 21 Jul 2014, at 20:54, Steve Dainard wrote:
 
  I should mention I can mount this usb drive in a CentOS 6.5 VM without
 any problems.
 
  Hi,
  there should be no difference configuration-wise. well, please compare
 libvirt's xml to be sure and confirm
  If it's the case then it might be a problem of qemu/kvm and/or windows
 
  Thanks,
  michal
 
 
 
  On Mon, Jul 21, 2014 at 2:11 PM, Steve Dainard sdain...@miovision.com
 wrote:
  I'm using the hostusb hook on RHEV 3.4 trial.
 
  The usb device is passed through to the VM, but I'm getting errors in a
 Windows VM when the device driver is loaded.
 
  I started with a simple usb drive, on the host it is listed as:
 
  Bus 002 Device 010: ID 05dc:c75c Lexar Media, Inc.
 
  Which I added as 0x05dc:0xc75c to the Windows 7 x64 VM.
 
  In Windows I get an error in device manager:
  USB Mass Storage Device This device cannot start. (Code 10)
  Properties/General Tab: Device type: Universal Serial Bus Controllers,
 Manufacturer: Compatible USB storage device, Location: Port_#0001.Hub_#0001
 
  Under hardware Ids:
  USB\VID_05DCPID_C75CREV_0102
  USB\VID_05DCPID_C75C
 
  So it looks like the proper USB device ID is passed to the VM.
 
  I don't see any error messages in event viewer, and I don't see
 anything in VDSM logs either.
 
  Any help is appreciated.
 
  Steve
 
  ___
  Users mailing list
  Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
 
 


CentOS 6.5 VM

domain type='kvm' id='52'
  namesteve-test/name
  uuiddeabe801-bc66-43c9-9d99-ecd852c3ac6e/uuid
  memory unit='KiB'2097152/memory
  currentMemory unit='KiB'2097152/currentMemory
  memtune
min_guarantee unit='KiB'2097152/min_guarantee
  /memtune
  vcpu placement='static' current='2'160/vcpu
  cputune
shares1020/shares
  /cputune
  sysinfo type='smbios'
system
  entry name='manufacturer'Red Hat/entry
  entry name='product'RHEV Hypervisor/entry
  entry name='version'6Server-6.5.0.1.el6/entry
  entry name='serial'----002590DE3912/entry
  entry name='uuid'deabe801-bc66-43c9-9d99-ecd852c3ac6e/entry
/system
  /sysinfo
  os
type arch='x86_64' machine='rhel6.5.0'hvm/type
smbios mode='sysinfo'/
  /os
  features
acpi/
  /features
  cpu mode='custom' match='exact'
model fallback='allow'SandyBridge/model
topology sockets='160' cores='1' threads='1'/
  /cpu
  clock offset='variable' adjustment='0' basis='utc'
timer name='rtc' tickpolicy='catchup'/
  /clock
  on_poweroffdestroy/on_poweroff
  on_rebootrestart/on_reboot
  on_crashdestroy/on_crash
  devices
emulator/usr/libexec/qemu-kvm/emulator
disk type='file' device='cdrom'
  driver name='qemu' type='raw'/
  source file='/rhev/data-center/mnt/rhev-storage-vip.miovision.corp:_vol_vol__rhev__storage_ISO/3b8b7896-cc20-4923-bf7d-f5a66b49fae7/images/----/CentOS-6.4-x86_64-bin-DVD1.iso' startupPolicy='optional'
seclabel model='selinux' relabel='no'/
  /source
  target dev='hdc' bus='ide'/
  readonly/
  serial/serial
  boot order='2'/
  alias name='ide0-1-0'/
  address type='drive' controller='0' bus='1' target='0' unit='0'/
/disk
disk type='file' device='disk' snapshot='no'
  driver name='qemu' type='qcow2' cache='none' error_policy='stop' io='threads'/
  source file='/rhev/data-center/5849b030-626e-47cb-ad90-3ce782d831b3/b9db3498-6f0d-4e95-a2a8-c6df24bda991/images/6edab781-7f9b-4c4d-beb2-1bf197f325d1/91a6de98-80e2-49c6-8ab1-c93e52c28537'
seclabel model='selinux' relabel='no'/
  /source
  target dev='vda' bus='virtio'/
  serial6edab781-7f9b-4c4d-beb2-1bf197f325d1/serial
  boot order='1'/
  alias name='virtio-disk0'/
  address type='pci' domain

Re: [ovirt-users] hostusb hook - VM device errors in Windows VM

2014-07-22 Thread Steve Dainard
I just saw the your device can perform faster warning again in Windows
and decided to check it out.

Should the USB device be showing under an Intel controller? Is the RH
controller only for spice?


​


On Tue, Jul 22, 2014 at 10:12 AM, Dan Kenigsberg dan...@redhat.com wrote:

 On Tue, Jul 22, 2014 at 03:50:59PM +0200, Michal Skrivanek wrote:
 
  On Jul 22, 2014, at 15:49 , Steve Dainard sdain...@miovision.com
 wrote:
 
   Hi Michal,
  
   How can I generate libvirt xml from rhevm?
 
  virsh -r dumpxml domain on the host

 Or dig into vdsm.log (in case the VM is no longer there)

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] RHEV 3.4 trial hosted-engine either host wants to take ownership

2014-07-21 Thread Steve Dainard
I added a hook to rhevm, and then restarted the engine service which
triggered a hosted-engine VM shutdown (likely because of the failed
liveliness check).

Once the hosted-engine VM shutdown it did not restart on the other host.

On both hosts configured for hosted-engine I'm seeing logs from ha-agent
where each host thinks the other host has a better score. Is there supposed
to be a mechanism for a tie breaker here? I do notice that the log mentions
best REMOTE host, so perhaps I'm interpreting this message incorrectly.

ha-agent logs:

Host 001:

MainThread::INFO::2014-07-21
11:51:57,396::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1405957917.4 type=state_transition
detail=EngineDown-EngineDown hostname='rhev001.miovision.corp'
MainThread::INFO::2014-07-21
11:51:57,397::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Success, was notification of state_transition (EngineDown-EngineDown) sent?
ignored
MainThread::INFO::2014-07-21
11:51:57,924::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Current state EngineDown (score: 2400)
MainThread::INFO::2014-07-21
11:51:57,924::hosted_engine::328::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Best remote host rhev002.miovision.corp (id: 2, score: 2400)
MainThread::INFO::2014-07-21
11:52:07,961::states::454::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(consume)
Engine down, local host does not have best score
MainThread::INFO::2014-07-21
11:52:07,975::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1405957927.98 type=state_transition
detail=EngineDown-EngineDown hostname='rhev001.miovision.corp'

Host 002:

MainThread::INFO::2014-07-21
11:51:47,405::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1405957907.41 type=state_transition
detail=EngineDown-EngineDown hostname='rhev002.miovision.corp'
MainThread::INFO::2014-07-21
11:51:47,406::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Success, was notification of state_transition (EngineDown-EngineDown) sent?
ignored
MainThread::INFO::2014-07-21
11:51:47,834::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Current state EngineDown (score: 2400)
MainThread::INFO::2014-07-21
11:51:47,835::hosted_engine::328::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Best remote host rhev001.miovision.corp (id: 1, score: 2400)
MainThread::INFO::2014-07-21
11:51:57,870::states::454::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(consume)
Engine down, local host does not have best score
MainThread::INFO::2014-07-21
11:51:57,883::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1405957917.88 type=state_transition
detail=EngineDown-EngineDown hostname='rhev002.miovision.corp'

This went on for 20 minutes about an hour ago, and I decided to --vm-start
on one of the hosts. The manager VM runs for a few minutes with the engine
ui accessible, before shutting itself down again.

I then put host 002 into local maintenance mode, and host 001 auto started
the hosted-engine VM. The logging still references host 002 as the 'best
remote host' even though the calculated score is now 0:

MainThread::INFO::2014-07-21
12:03:24,011::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1405958604.01 type=state_transition
detail=EngineUp-EngineUp hostname='rhev001.miovision.corp'
MainThread::INFO::2014-07-21
12:03:24,013::brokerlink::117::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Success, was notification of state_transition (EngineUp-EngineUp) sent?
ignored
MainThread::INFO::2014-07-21
12:03:24,515::hosted_engine::323::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Current state EngineUp (score: 2400)
MainThread::INFO::2014-07-21
12:03:24,516::hosted_engine::328::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(start_monitoring)
Best remote host rhev002.miovision.corp (id: 2, score: 0)
MainThread::INFO::2014-07-21
12:03:34,567::brokerlink::108::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(notify)
Trying: notify time=1405958614.57 type=state_transition
detail=EngineUp-EngineUp hostname='rhev001.miovision.corp'

Once the hosted-engine VM was up for about 5 minutes I took host 002 out of
local maintenance mode and the VM has not since shutdown.

Is this expected behaviour? Is this the normal recovery process when two
hosts both hosting hosted-engine are started at the same time? I would have
expected once hosted-engine VM was detected as bad (liveliness check from
when I restarted the engine service) and the VM was shutdown, that it would
spin back up on the next available host.

Thanks,
Steve

[ovirt-users] hostusb hook - VM device errors in Windows VM

2014-07-21 Thread Steve Dainard
I'm using the hostusb hook on RHEV 3.4 trial.

The usb device is passed through to the VM, but I'm getting errors in a
Windows VM when the device driver is loaded.

I started with a simple usb drive, on the host it is listed as:

Bus 002 Device 010: ID 05dc:c75c Lexar Media, Inc.

Which I added as 0x05dc:0xc75c to the Windows 7 x64 VM.

In Windows I get an error in device manager:
USB Mass Storage Device This device cannot start. (Code 10)
Properties/General Tab: Device type: Universal Serial Bus Controllers,
Manufacturer: Compatible USB storage device, Location: Port_#0001.Hub_#0001

Under hardware Ids:
USB\VID_05DCPID_C75CREV_0102
USB\VID_05DCPID_C75C

So it looks like the proper USB device ID is passed to the VM.

I don't see any error messages in event viewer, and I don't see anything in
VDSM logs either.

Any help is appreciated.

Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hostusb hook - VM device errors in Windows VM

2014-07-21 Thread Steve Dainard
I should mention I can mount this usb drive in a CentOS 6.5 VM without any
problems.


On Mon, Jul 21, 2014 at 2:11 PM, Steve Dainard sdain...@miovision.com
wrote:

 I'm using the hostusb hook on RHEV 3.4 trial.

 The usb device is passed through to the VM, but I'm getting errors in a
 Windows VM when the device driver is loaded.

 I started with a simple usb drive, on the host it is listed as:

 Bus 002 Device 010: ID 05dc:c75c Lexar Media, Inc.

 Which I added as 0x05dc:0xc75c to the Windows 7 x64 VM.

 In Windows I get an error in device manager:
 USB Mass Storage Device This device cannot start. (Code 10)
 Properties/General Tab: Device type: Universal Serial Bus Controllers,
 Manufacturer: Compatible USB storage device, Location: Port_#0001.Hub_#0001

 Under hardware Ids:
 USB\VID_05DCPID_C75CREV_0102
 USB\VID_05DCPID_C75C

 So it looks like the proper USB device ID is passed to the VM.

 I don't see any error messages in event viewer, and I don't see anything
 in VDSM logs either.

 Any help is appreciated.

 Steve

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Relationship bw storage domain uuid/images/children and VM's

2014-07-18 Thread Steve Dainard
:



 - Original Message -
  From: Steve Dainard sdain...@miovision.com
  To: users users@ovirt.org
  Sent: Thursday, July 17, 2014 7:51:31 PM
  Subject: [ovirt-users] Relationship bw storage domain
 uuid/images/childrenand VM's
 
  Hello,
 
  I'd like to get an understanding of the relationship between VM's using a
  storage domain, and the child directories listed under .../storage
 domain
  name/storage domain uuid/images/.
 
  Running through some backup scenarios I'm noticing a significant
 difference
  between the number of provisioned VM's using a storage domain (21) +
  templates (6) versus the number of child directories under images/ (107).

 Can you please elaborate (if possible) on the number of images per VM that
 you're having in your setup?

 
  Running RHEV 3.4 trial.
 
  Thanks,
  Steve
 
  ___
  Users mailing list
  Users@ovirt.org
  http://lists.ovirt.org/mailman/listinfo/users
 

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Online backup options

2014-07-18 Thread Steve Dainard
Anyone?


On Fri, Jun 20, 2014 at 9:35 AM, Steve Dainard sdain...@miovision.com
wrote:

 Hello Ovirt team,

 Reading this bulletin: https://access.redhat.com/site/solutions/117763
 there is a reference to 'private Red Hat Bug # 523354' covering online
 backups of VM's.

 Can someone comment on this feature, and rough timeline? Is this a native
 backup solution that will be included with Ovirt/RHEV?

 Is this Ovirt feature where the work is being done?
 http://www.ovirt.org/Features/Backup-Restore_API_Integration It seems
 like this may be a different feature specifically for 3rd party backup
 options.

 Thanks,
 Steve

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Relationship bw storage domain uuid/images/children and VM's

2014-07-17 Thread Steve Dainard
Hello,

I'd like to get an understanding of the relationship between VM's using a
storage domain, and the child directories listed under .../storage domain
name/storage domain uuid/images/.

Running through some backup scenarios I'm noticing a significant difference
between the number of provisioned VM's using a storage domain (21) +
templates (6) versus the number of child directories under images/ (107).

Running RHEV 3.4 trial.

Thanks,
Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Master storage domain no longer exists, how to recover?

2014-07-14 Thread Steve Dainard
I'm running a trial of RHEV 3.4 and decided to run through our DR plan.

I have two storage domains.
vm-store
aws_storage_gateway

vm-store is the domain which contains all local VM's which would be
considered important. I must have put vm-store into maintenance mode before
aws_storage_gateway which caused aws_storage_gateway to become the master
domain. It also happens to be non-existant now after recovery, as the DR
process is to re-create this SD with new disks as all the data is on aws
anyways.

Is there a method of changing the master storage domain when all domains
are in maintenance mode?

If I attempt to destroy it I get a UI error:

Error while executing action: Cannot destroy the master Storage Domain from
the Data Center without another active Storage Domain to take its place.
-Either activate another Storage Domain in the Data Center, or remove the
Data Center.
-If you have problems with the master Data Domain, consider following the
recovery process described in the documentation, or contact your system
administrator.

I tried to fake the storage domain by creating another one, copy in the
file structure, and renaming the folder uuid and updating the meta file to
match the missing SD without luck.

Can I update the db storage pool reference with the vm-store SD uuid?

engine=# select * from storage_pool;
  id  |  name   |   description   |
storage_p
ool_type | storage_pool_format_type | status | master_domain_version |
spm_vds_id | c
ompatibility_version | _create_date  | _update_date

| quota_enforcement_type | free_text_comment | is_local
--+-+-+--
-+--++---++--
-+---+---
++---+--
 *5849b030-626e-47cb-ad90-3ce782d831b3* | Default | The default Data Center
|
   1 | 3|  2 |19 |
   | 3
.3   | 2014-05-29 19:46:50.199772-04 | 2014-07-10
09:24:22.790738-04
|  0 |   | f
(1 row)




Thanks,
Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Master storage domain no longer exists, how to recover?

2014-07-14 Thread Steve Dainard
And 5 minutes later I found...

https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Virtualization/3.4/html-single/Administration_Guide/index.html
3.5.6. Re-Initializing a Data Center: Recovery Procedure

Which worked quite well.

Steve


On Mon, Jul 14, 2014 at 10:27 AM, Steve Dainard sdain...@miovision.com
wrote:

 I'm running a trial of RHEV 3.4 and decided to run through our DR plan.

 I have two storage domains.
 vm-store
 aws_storage_gateway

 vm-store is the domain which contains all local VM's which would be
 considered important. I must have put vm-store into maintenance mode before
 aws_storage_gateway which caused aws_storage_gateway to become the master
 domain. It also happens to be non-existant now after recovery, as the DR
 process is to re-create this SD with new disks as all the data is on aws
 anyways.

 Is there a method of changing the master storage domain when all domains
 are in maintenance mode?

 If I attempt to destroy it I get a UI error:

 Error while executing action: Cannot destroy the master Storage Domain
 from the Data Center without another active Storage Domain to take its
 place.
 -Either activate another Storage Domain in the Data Center, or remove the
 Data Center.
 -If you have problems with the master Data Domain, consider following the
 recovery process described in the documentation, or contact your system
 administrator.

 I tried to fake the storage domain by creating another one, copy in the
 file structure, and renaming the folder uuid and updating the meta file to
 match the missing SD without luck.

 Can I update the db storage pool reference with the vm-store SD uuid?

 engine=# select * from storage_pool;
   id  |  name   |   description
 | storage_p
 ool_type | storage_pool_format_type | status | master_domain_version |
 spm_vds_id | c
 ompatibility_version | _create_date  |
 _update_date
 | quota_enforcement_type | free_text_comment | is_local

 --+-+-+--

 -+--++---++--

 -+---+---
 ++---+--
  *5849b030-626e-47cb-ad90-3ce782d831b3* | Default | The default Data
 Center |
1 | 3|  2 |19 |
| 3
 .3   | 2014-05-29 19:46:50.199772-04 | 2014-07-10
 09:24:22.790738-04
 |  0 |   | f
 (1 row)




 Thanks,
 Steve

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Online backup options

2014-06-20 Thread Steve Dainard
Hello Ovirt team,

Reading this bulletin: https://access.redhat.com/site/solutions/117763
there is a reference to 'private Red Hat Bug # 523354' covering online
backups of VM's.

Can someone comment on this feature, and rough timeline? Is this a native
backup solution that will be included with Ovirt/RHEV?

Is this Ovirt feature where the work is being done?
http://www.ovirt.org/Features/Backup-Restore_API_Integration It seems like
this may be a different feature specifically for 3rd party backup options.

Thanks,
Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-25 Thread Steve Dainard
Restarting vdsm and hosts didn't do anything helpful.

I was able to clone from latest snapshot, then live snapshot the new cloned
VM.

After upgrading engine to 3.4 and upgrading my hosts, I can now live
snapshot this VM.


*Steve *


On Thu, Apr 24, 2014 at 1:48 AM, Itamar Heim ih...@redhat.com wrote:

 On 04/23/2014 09:57 PM, R P Herrold wrote:

 On Wed, 23 Apr 2014, Steve Dainard wrote:

  I have other VM's with the same amount of snapshots without this problem.
 No conclusion jumping going on. More interested in what the best practice
 is for VM's that accumulate snapshots over time.


 For some real world context, we seem to accumulate snapshots
 using our local approach, and are not that focused on, or
 attentive about removing them.  The 'highwater mark' of 39, on
 a machine that has been around since it was provisioned:
 2010-01-05

 [root@xxx backups]# ./count-snapshots.sh | sort -n | tail -3
 38 vm_64099
 38 vm_98036
 39 vm_06359

 Accumulating large numbers of snapshots seems more the
 function of pets, than ephemeral 'cattle'

 I wrote the first paragraph without looking up the 'owners' of
 the images. As I dereference the VM id's, all of the top ten
 in that list turn out to be mailservers, radius servers, name
 servers, and such, where the business unit owners chose not
 (or neglect) to 'winnow' their herd.  There are no ephemeral
 use units in the top ten

 -- Russ herrold
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


 please note there is a recommended limit of having no more than 500
 snapshots per block storage domain due to some LVM performance issues with
 high number of LVs. each disk/snapshot is an LV.
 NFS doesn't have this limitation.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Anyone using gluster storage domain with WAN geo-rep?

2014-04-23 Thread Steve Dainard
I'm currently using a two node combined virt/storage setup with Ovirt 3.3.4
and Gluster 3.4.2 (replica 2, glusterfs storage domain). I'll call this
pair PROD.

I'm then geo-replicating to another gluster replica pair on the local net,
btrfs underlying storage, and volume snapshots so I can recover my storage
domain from different points in time if necessary. Its also local so
restore time is much better than off-site. I'll call this pair BACKUP.

I'm planning on setting up geo-replication from BACKUP to an EC2 gluster
target. I'll call this host EC2HOST.

PROD ---geo-rep-lan--- BACKUP ---geo-rep-wan--- EC2HOST

I'd like to avoid saturating my WAN link during office hours. I have some
ideas (or combination of):

1. limit bandwidth during certain hours to the offsite hosts. But
realistically the bandwidth I would allocate is so low I don't see the
purpose of this. Also with 8 guests running, I'm noticing quite a bit of
data transfer to the local backup nodes (avg 6-8MB/s), and I'm thinking
there is a lot of thrashing going on which isn't useful to backup offsite
anyways.

2. stop WAN geo-replication during office hours, and restart for
overnight/weekend hours.

3. Not use geo-rep between BACKUP --- EC2HOST, use rsync on one of the
btrfs volume snapshots so we avoid the thrashing. In this case I could
limit WAN speed to 1MB/s which should be fine for most differences
throughout the day.

So my question is, how do you off-site your storage domains, what
constraints have you identified and how have you dealt with them? And of
course how would you deal with the scenario I've oulined above?

Thanks,




*Steve*
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Steve Dainard
No alert in web ui, I restarted the VM yesterday just in case, no change. I
also restored an earlier snapshot and tried to re-snapshot, same result.


*Steve *


On Tue, Apr 22, 2014 at 10:57 AM, Dafna Ron d...@redhat.com wrote:

 This is the actual problem:

 bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::DEBUG::2014-04-22
 10:21:49,374::volume::1058::Storage.Misc.excCmd::(createVolume) FAILED:
 err = '/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
 66d9ae9-e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb:
 error while creating qcow2: No such file or directory\n'; rc = 1

 from that you see the actual failure:

 bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
 10:21:49,392::volume::286::Storage.Volume::(clone) Volume.clone: can't
 clone: /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d
 9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7 to
 /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-
 e46a-46f8-9f4b-964d8af0675b/87efa937-b31f-4bb1-aee1-0ee1
 4a0dc6fb
 bf025a73-eeeb-4ac5-b8a9-32afa4ae482e::ERROR::2014-04-22
 10:21:49,392::volume::508::Storage.Volume::(create) Unexpected error
 Traceback (most recent call last):
   File /usr/share/vdsm/storage/volume.py, line 466, in create
 srcVolUUID, imgPath, volPath)
   File /usr/share/vdsm/storage/fileVolume.py, line 160, in _create
 volParent.clone(imgPath, volUUID, volFormat, preallocate)
   File /usr/share/vdsm/storage/volume.py, line 287, in clone
 raise se.CannotCloneVolume(self.volumePath, dst_path, str(e))
 CannotCloneVolume: Cannot clone volume: 'src=/rhev/data-center/
 9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-
 4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-
 964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7, dst=/rhev/data-cen
 ter/9497ef2c-8368-4c92-8d61-7f318a90748f/95b9d922-4df7-
 4d3b-9bca-467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-
 964d8af0675b/87efa937-b31f-4bb1-aee1-0ee14a0dc6fb: Error creating a new
 volume: ([Formatting \'/rhev/data-center/9497ef2c-8368-
 4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-
 467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b/
 87efa937-b31f-4bb1-aee1-0ee14a0dc6fb\', fmt=qcow2 size=21474836480
 backing_file=\'../466d9ae9-e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa
 1c-4436-baca-ca55726d54d7\' backing_fmt=\'qcow2\' encryption=off
 cluster_size=65536 ],)'


 do you have any alert in the webadmin to restart the vm?

 Dafna


 On 04/22/2014 03:31 PM, Steve Dainard wrote:

 Sorry for the confusion.

 I attempted to take a live snapshot of a running VM. After that failed, I
 migrated the VM to another host, and attempted the live snapshot again
 without success, eliminating a single host as the cause of failure.

 Ovirt is 3.3.4, storage domain is gluster 3.4.2.1, OS is CentOS 6.5.

 Package versions:
 libvirt-0.10.2-29.el6_5.5.x86_64
 libvirt-lock-sanlock-0.10.2-29.el6_5.5.x86_64
 qemu-img-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
 qemu-kvm-rhev-0.12.1.2-2.415.el6.nux.3.x86_64
 qemu-kvm-rhev-tools-0.12.1.2-2.415.el6.nux.3.x86_64
 vdsm-4.13.3-4.el6.x86_64
 vdsm-gluster-4.13.3-4.el6.noarch


 I made another live snapshot attempt at 10:21 EST today, full vdsm.log
 attached, and a truncated engine.log.

 Thanks,

 *Steve
 *



 On Tue, Apr 22, 2014 at 9:48 AM, Dafna Ron d...@redhat.com mailto:
 d...@redhat.com wrote:

 please explain the flow of what you are trying to do, are you
 trying to live migrate the disk (from one storage to another), are
 you trying to migrate the vm and after vm migration is finished
 you try to take a live snapshot of the vm? or are you trying to
 take a live snapshot of the vm during a vm migration from host1 to
 host2?

 Please attach full vdsm logs from any host you are using (if you
 are trying to migrate the vm from host1 to host2) + please attach
 engine log.

 Also, what is the vdsm, libvirt and qemu versions, what ovirt
 version are you using and what is the storage you are using?

 Thanks,

 Dafna




 On 04/22/2014 02:12 PM, Steve Dainard wrote:

 I've attempted migrating the vm to another host and taking a
 snapshot, but I get this error:

 6efd33f4-984c-4513-b5e6-fffdca2e983b::ERROR::2014-04-22
 01:09:37,296::volume::286::Storage.Volume::(clone)
 Volume.clone: can't clone:
 /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-
 e46a-46f8-9f4b-964d8af0675b/1a67de4b-aa1c-4436-baca-ca55726d54d7
 to
 /rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/466d9ae9-
 e46a-46f8-9f4b-964d8af0675b/b230596f-97bc-4532-ba57-5654fa9c6c51

 A bit more of the vdsm log is attached.

 Other vm's are snapshotting without issue

Re: [ovirt-users] Ovirt snapshot failing on one VM

2014-04-22 Thread Steve Dainard
 272 Apr 10 14:34
3ece1489-9bff-4223-ab97-e45135106222.meta
-rw-rw. 1 vdsm kvm22413312 Apr 10 14:29
dcee2e8a-8803-44e2-80e8-82c882af83ef
-rw-rw. 1 vdsm kvm 1048576 Apr 10 14:28
3ece1489-9bff-4223-ab97-e45135106222.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 14:28
dcee2e8a-8803-44e2-80e8-82c882af83ef.meta
-rw-rw. 1 vdsm kvm54460416 Apr 10 14:26
57066786-613a-46ff-b2f9-06d84678975b
-rw-rw. 1 vdsm kvm 1048576 Apr 10 14:26
dcee2e8a-8803-44e2-80e8-82c882af83ef.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 14:26
57066786-613a-46ff-b2f9-06d84678975b.meta
-rw-rw. 1 vdsm kvm15728640 Apr 10 13:31
121ae509-d2b2-4df2-a56f-dfdba4b8d21c
-rw-rw. 1 vdsm kvm 1048576 Apr 10 13:30
57066786-613a-46ff-b2f9-06d84678975b.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 13:30
121ae509-d2b2-4df2-a56f-dfdba4b8d21c.meta
-rw-rw. 1 vdsm kvm 5767168 Apr 10 13:18
1d95a9d2-e4ba-4bcc-ba71-5d493a838dcc
-rw-rw. 1 vdsm kvm 1048576 Apr 10 13:17
121ae509-d2b2-4df2-a56f-dfdba4b8d21c.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 13:17
1d95a9d2-e4ba-4bcc-ba71-5d493a838dcc.meta
-rw-rw. 1 vdsm kvm 5373952 Apr 10 13:13
3ce8936a-38f5-43a9-a4e0-820094fbeb04
-rw-rw. 1 vdsm kvm 1048576 Apr 10 13:13
1d95a9d2-e4ba-4bcc-ba71-5d493a838dcc.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 13:12
3ce8936a-38f5-43a9-a4e0-820094fbeb04.meta
-rw-rw. 1 vdsm kvm  3815243776 Apr 10 13:11
7211d323-c398-4c1c-8524-a1047f9d5ec9
-rw-rw. 1 vdsm kvm 1048576 Apr 10 13:11
3ce8936a-38f5-43a9-a4e0-820094fbeb04.lease
-rw-r--r--. 1 vdsm kvm 272 Apr 10 13:11
7211d323-c398-4c1c-8524-a1047f9d5ec9.meta
-rw-r--r--. 1 vdsm kvm 272 Mar 19 10:35
af94adc4-fad4-42f5-a004-689670311d66.meta
-rw-rw. 1 vdsm kvm 21474836480 Mar 19 10:22
af94adc4-fad4-42f5-a004-689670311d66
-rw-rw. 1 vdsm kvm 1048576 Mar 19 09:39
7211d323-c398-4c1c-8524-a1047f9d5ec9.lease
-rw-rw. 1 vdsm kvm 1048576 Mar 19 09:39
af94adc4-fad4-42f5-a004-689670311d66.lease

Its just very odd that I can snapshot any other VM except this one.

I just cloned a new VM from the last snapshot on this VM and it created
without issue. I was also able to snapshot the new VM without a problem.


*Steve *


On Tue, Apr 22, 2014 at 12:51 PM, Dafna Ron d...@redhat.com wrote:

 it's the same error:

 c1d7c4e-392b-4a62-9836-3add1360a46d::DEBUG::2014-04-22
 12:13:44,340::volume::1058::Storage.Misc.excCmd::(createVolume) FAILED:
 err = '/rhev/data-center/9497ef2c-8368-4c92-8d61-7f318a90748f/
 95b9d922-4df7-4d3b-9bca-467e2fd9d573/images/4
 66d9ae9-e46a-46f8-9f4b-964d8af0675b/0b2d15e5-bf4f-4eaf-90e2-f1bd51a3a936:
 error while creating qcow2: No such file or directory\n'; rc = 1


 were these 23 snapshots created any way each time we fail to create the
 snapshot or are these older snapshots which you actually created before the
 failure?

 at this point my main theory is that somewhere along the line you had some
 sort of failure in your storage and from that time each snapshot you create
 will fail.
 if the snapshots are created during the failure can you please delete the
 snapshots you do not need and try again?

 There should not be a limit on how many snapshots you can have since it's
 only a link changing the image the vm should boot from.
 Having said that, it's not ideal to have that many snapshots and can
 probably lead to unexpected results so I would not recommend having that
 many snapshots on a single vm :)

 for example, my second theory would be that because we have so many
 snapshots we have some sort of race where part of the createVolume command
 expects some result from a query run before the create itself and because
 there are so many snapshots there is no such file on the volume because
 it's too far up the list.

 can you also run: ls -l /rhev/data-center/9497ef2c-
 8368-4c92-8d61-7f318a90748f/95b9d922-4df7-4d3b-9bca-
 467e2fd9d573/images/466d9ae9-e46a-46f8-9f4b-964d8af0675b

 lets see what images are listed under that vm.

 btw, you know that your export domain is getting StorageDomainDoesNotExist
 in the vdsm log? is that domain in up state? can you try to deactivate the
 export domain?

 Thanks,

 Dafna





 On 04/22/2014 05:20 PM, Steve Dainard wrote:

 Ominous..

 23 snapshots. Is there an upper limit?

 Offline snapshot fails as well. Both logs attached again (snapshot
 attempted at 12:13 EST).

 *Steve *


 On Tue, Apr 22, 2014 at 11:20 AM, Dafna Ron d...@redhat.com mailto:
 d...@redhat.com wrote:

 are you able to take an offline snapshot? (while the vm is down)
 how many snapshots do you have on this vm?


 On 04/22/2014 04:19 PM, Steve Dainard wrote:

 No alert in web ui, I restarted the VM yesterday just in case,
 no change. I also restored an earlier snapshot and tried to
 re-snapshot, same result.

 *Steve
 *



 On Tue, Apr 22, 2014 at 10:57 AM, Dafna Ron d...@redhat.com
  mailto:d...@redhat.com mailto:d

Re: [ovirt-users] New VM from template, vm disk creation time very high

2014-04-21 Thread Steve Dainard
This may have been lost over the weekend...


*Steve*


On Thu, Apr 17, 2014 at 1:21 PM, Steve  wrote:

 When I create a new vm from a template (centos 6, 20gig thin disk) it
 takes close to 40 minutes to clone the disk image.

 Are there any settings that control how much bandwidth the cloning process
 can use, or how its prioritized? This is brutally slow, and although a
 quick clone isn't needed often, occasionally when someone is asking for a
 resource it would be nice to expedite the process.

 Storage: Gluster replica 2 volume
 Network: 10gig
 Ovirt: 3.3.4, Centos 6.5

 Thanks,


 *Steve *

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] New VM from template, vm disk creation time very high

2014-04-21 Thread Steve Dainard
Hi

Yes I'm using glusterfs as storage domain, and glusterd is running on the
ovirt nodes.



*Steve *


On Mon, Apr 21, 2014 at 11:46 AM, Ovirt User ldrt8...@gmail.com wrote:

 hi steve,

 are you using bluster fs as storage domain ? if yes, inside compute node
 or you dedicated 2 or more node for GLUSTER ?

 Il giorno 21/apr/2014, alle ore 16:00, Steve Dainard 
 sdain...@miovision.com ha scritto:

 This may have been lost over the weekend...


 *Steve*


 On Thu, Apr 17, 2014 at 1:21 PM, Steve  wrote:

 When I create a new vm from a template (centos 6, 20gig thin disk) it
 takes close to 40 minutes to clone the disk image.

 Are there any settings that control how much bandwidth the cloning
 process can use, or how its prioritized? This is brutally slow, and
 although a quick clone isn't needed often, occasionally when someone is
 asking for a resource it would be nice to expedite the process.

 Storage: Gluster replica 2 volume
 Network: 10gig
 Ovirt: 3.3.4, Centos 6.5

 Thanks,


 *Steve *


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] New VM from template, vm disk creation time very high

2014-04-17 Thread Steve Dainard
When I create a new vm from a template (centos 6, 20gig thin disk) it takes
close to 40 minutes to clone the disk image.

Are there any settings that control how much bandwidth the cloning process
can use, or how its prioritized? This is brutally slow, and although a
quick clone isn't needed often, occasionally when someone is asking for a
resource it would be nice to expedite the process.

Storage: Gluster replica 2 volume
Network: 10gig
Ovirt: 3.3.4, Centos 6.5

Thanks,


*Steve*
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Opinions needed: 3 node gluster replica 3 | NFS async | snapshots for consistency

2014-02-23 Thread Steve Dainard
On Sun, Feb 23, 2014 at 4:27 AM, Ayal Baron aba...@redhat.com wrote:



 - Original Message -
  I'm looking for some opinions on this configuration in an effort to
 increase
  write performance:
 
  3 storage nodes using glusterfs in replica 3, quorum.

 gluster doesn't support replica 3 yet, so I'm not sure how heavily I'd
 rely on this.


Glusterfs or RHSS doesn't support rep 3? How could I create a quorum
without 3+ hosts?



  Ovirt storage domain via NFS

 why NFS and not gluster?


Gluster via posix SD doesn't have any performance gains over NFS, maybe the
opposite.

Gluster 'native' SD's are broken on EL6.5 so I have been unable to test
performance. I have heard performance can be upwards of 3x NFS for raw
write.

Gluster doesn't have an async write option, so its doubtful it will ever be
close to NFS async speeds.



  Volume set nfs.trusted-sync on
  On Ovirt, taking snapshots often enough to recover from a storage crash

 Note that this would have negative write performance impact


The difference between NFS sync (50MB/s) and async (300MB/s on 10g) write
speeds should more than compensate for the performance hit of taking
snapshots more often. And that's just raw speed. If we take into
consideration IOPS (guest small writes) async is leaps and bounds ahead.


If we assume the site has backup UPS and generator power and we can build a
highly available storage system with 3 nodes in quorum, are there any
potential issues other than a write performance hit?

The issue I thought might be most prevalent is if an ovirt host goes down
and the VM's are automatically brought back up on another host, they could
incur disk corruption and need to be brought back down and restored to the
last snapshot state. This basically means the HA feature should be disabled.

Even worse, if the gluster node with CTDB NFS IP goes down, it may not have
written out and replicated to its peers.  -- I think I may have just
answered my own question.


Thanks,
Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Opinions needed: 3 node gluster replica 3 | NFS async | snapshots for consistency

2014-02-23 Thread Steve Dainard
On Sun, Feb 23, 2014 at 3:20 PM, Ayal Baron aba...@redhat.com wrote:



 - Original Message -
  On Sun, Feb 23, 2014 at 4:27 AM, Ayal Baron aba...@redhat.com wrote:
 
  
  
   - Original Message -
I'm looking for some opinions on this configuration in an effort to
   increase
write performance:
   
3 storage nodes using glusterfs in replica 3, quorum.
  
   gluster doesn't support replica 3 yet, so I'm not sure how heavily I'd
   rely on this.
  
 
  Glusterfs or RHSS doesn't support rep 3? How could I create a quorum
  without 3+ hosts?

 glusterfs has the capability but it hasn't been widely tested with oVirt
 yet and we've already found a couple of issues there.
 afaiu gluster has the ability to define a tie breaker (a third node which
 is part of the quorum but does not provide a third replica of the data).


Good to know, I'll dig into this.



 
 
  
Ovirt storage domain via NFS
  
   why NFS and not gluster?
  
 
  Gluster via posix SD doesn't have any performance gains over NFS, maybe
 the
  opposite.

 gluster via posix is mounting it using the gluster fuse client which
 should provide better performance + availability than NFS.


Availability for sure, but performance is seriously questionable. I've run
in both scenarios and haven't seen a performance improvement, the general
consensus seems to be fuse is adding overhead and therefore decreasing
performance vs. NFS.



 
  Gluster 'native' SD's are broken on EL6.5 so I have been unable to test
  performance. I have heard performance can be upwards of 3x NFS for raw
  write.

 Broken how?


Ongoing issues, libgfapi support wasn't available, and then was disabled
because snapshot support wasn't built into the kvm packages which was a
dependency. There are a few threads in reference to this, and some effort
to get CentOS builds to enable snapshot support in kvm.

I have installed rebuilt qemu packages with the RHEV snapshot flag enabled,
and was just able to create a native gluster SD, maybe I missed something
during a previous attempt. I'll test performance and see if its close to
what I'm looking for.



 
  Gluster doesn't have an async write option, so its doubtful it will ever
 be
  close to NFS async speeds.t
 
 
  
Volume set nfs.trusted-sync on
On Ovirt, taking snapshots often enough to recover from a storage
 crash
  
   Note that this would have negative write performance impact
  
 
  The difference between NFS sync (50MB/s) and async (300MB/s on 10g)
 write
  speeds should more than compensate for the performance hit of taking
  snapshots more often. And that's just raw speed. If we take into
  consideration IOPS (guest small writes) async is leaps and bounds ahead.

 I would test this, since qemu is already doing async I/O (using threads
 when native AIO is not supported) and oVirt runs it with cache=none (direct
 I/O) so sync ops should not happen that often (depends on guest).  You may
 be still enjoying performance boost, but I've seen UPS systems fail before
 bringing down multiple nodes at once.
 In addition, if you do not guarantee your data is safe when you create a
 snapshot (and it doesn't seem like you are) then I see no reason to think
 your snapshots are any better off than latest state on disk.


My logic here was if a snapshot is run, then the disk and system state
should be consistent at time of snapshot once its been written to storage.
If the host failed during snapshot then the snapshot would be incomplete,
and the last complete snapshot would need to be used for recovery.



 
 
  If we assume the site has backup UPS and generator power and we can
 build a
  highly available storage system with 3 nodes in quorum, are there any
  potential issues other than a write performance hit?
 
  The issue I thought might be most prevalent is if an ovirt host goes down
  and the VM's are automatically brought back up on another host, they
 could
  incur disk corruption and need to be brought back down and restored to
 the
  last snapshot state. This basically means the HA feature should be
 disabled.

 I'm not sure I understand what your concern is here, what would cause the
 data corruption? if your node crashed then there is no I/O in flight.  So
 starting up the VM should be perfectly safe.


Good point, that makes sense.



 
  Even worse, if the gluster node with CTDB NFS IP goes down, it may not
 have
  written out and replicated to its peers.  -- I think I may have just
  answered my own question.

 If 'trusted-sync' means that the CTDB NFS node acks the I/O before it
 reached quorum then I'd say that's a gluster bug.


http://gluster.org/community/documentation/index.php/Gluster_3.2:_Setting_Volume_Options#nfs.trusted-syncSpecifically
mentions data won't be guaranteed to be on disk, but doesn't
mention if data would be replicated in memory between gluster nodes.
Technically async breaks the NFS protocol standard anyways but this seems
like a question for the gluster guys, I'll 

Re: [Users] Why choose glusterfs over iscsi for oVirt storage?

2014-02-20 Thread Steve Dainard
On Mon, Feb 17, 2014 at 3:50 AM, Justin Clacherty jus...@redfish.com.auwrote:

 Hi,



 I'm just setting up some storage for use with oVirt and am wondering why I
 might choose glusterfs rather than just exporting a raid array as iscsi.
 The oVirt team seem to be pushing gluster (though that could just be
 because it's a Red Hat product).  Can anyone answer this one?


If you're contemplating software iscsi then you probably aren't working on
a production setup that requires high availability or data consistency. If
you are, you should take a deep dive on what your single point of failures
are. If you're actually talking about a SAN lun, then I would imagine you
aren't the storage admin, and can leave the particulars in their hands.

Its funny, in RH's current supported version of RHEV, native gluster
storage isn't even an option. This is a pretty new project, and if you
follow the bread crumbs on distributed file systems like glusterfs or ceph
you'll start to see the future advantages. You can also figure that seeing
as both projects are spearheaded by RH dev's that there is likely a lot of
cross-talk and feature requests making both projects better, and who
wouldn't promote a better solution?




 What I have come up with is as follows.



 For:


There is too much to cover in any of these advantages, you're going to need
to do a lot of research on each of these 'features' if you want to use them
successfully.

 -  easy to expand

 -  redundancy across storage devices/easy replication

 -  high availablility

 -  performance

 -  it's kind of cool J

 -  maintenance?



 Against (these are guesses):

 -  performance? (multiple layers of filesystems everywhere - fs
 in vm + image on gluster + gluster + block filesystem)

Its worse than just a ton of software layers. NAS protocols seem to be
focused on data consistency, which means they don't write async to storage.
iscsi is typically async and has much better throughput, but also a greater
chance for data loss or corruption. Technically you can achieve the same
level of performance as iscsi using NFS (backed by glusterfs if you like)
but you need to set options on the volume to allow async writes.


 -  complexity

If you're doing storage properly, the underpinnings are always complex
unless you're paying someone else to do it (read: SAN / managed service
provider). Research multipath and HA on software iscsi and you'll see what
I mean.


 -  maintenance?



 Any help here is appreciated.  Also, does the underlying block level
 filesystem matter here?  VMs running under ovirt would be typical business
 applications - file serving (samba4), email, databases, web servers, etc.


There is a lot to answer here and I don't have all the answers. Take a look
at the gluster docs for underlying file system requirements. Any block
device should work. Specifically I'll mention that the glusterfs team
doesn't suggest hosting db's on glusterfs - many small reads/writes are not
one of glusters strong points.




 Cheers,

 Justin.







 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Opinions needed: 3 node gluster replica 3 | NFS async | snapshots for consistency

2014-02-20 Thread Steve Dainard
I'm looking for some opinions on this configuration in an effort to
increase write performance:

3 storage nodes using glusterfs in replica 3, quorum.
Ovirt storage domain via NFS
Volume set nfs.trusted-sync on
On Ovirt, taking snapshots often enough to recover from a storage crash
Using CTDB to manage NFS storage domain IP, moving it to another storage
node when necessary

Something along the lines of EC2's data consistency model, where only
snapshots can be considered reliable. The Ovirt added advantage would be
memory consistency at time of snapshot as well.

Feedback appreciated, including 'you are insane for thinking this is a good
idea' (and some supported reasoning would be great).

Thanks,



*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage domain

2014-02-19 Thread Steve Dainard
Hi Nir,

I have a thread open on the gluster side about heal-failed operations, so
I'll wait for a response on that side.

Agreed on two node quorum, I'm waiting for a 3rd node right now :) But in
the meantime or for anyone who reads this thread, if you only have 2
storage nodes you have to weigh the risks of 2 nodes in quorum ensuring
storage consistency, or 2 nodes no quorum with an extra shot at uptime.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Wed, Feb 19, 2014 at 4:13 AM, Nir Soffer nsof...@redhat.com wrote:

 - Original Message -
  From: Steve Dainard sdain...@miovision.com
  To: Nir Soffer nsof...@redhat.com
  Cc: users users@ovirt.org
  Sent: Tuesday, February 11, 2014 7:42:37 PM
  Subject: Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage
 domain
 
  Enabled logging, logs attached.

 According to sanlock and gluster log:

 1. on the host, sanlock is failing write to the ids volume
 2. on the gluster side we see failure to heal the ids file.

 This looks like glusterfs issue, and should be handled by glusterfs folks.

 You probably should configure sanlock log level back to the default by
 commenting
 out the configuration I suggested in the previous mail.

 According to gluster configuration in this log, this looks like 2 replicas
 with auto quorum.
 This setup is not recommended because both machines must be up all the
 time.
 When one machine is down, your entire storage is down.

 Check this post explaining this issue:
 http://lists.ovirt.org/pipermail/users/2014-February/021541.html

 Thanks,
 Nir

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Live migration of VM's occasionally fails

2014-02-17 Thread Steve Dainard
Hi Dafna,

No snapshots of either of those VM's have been taken, and there are no
updates for any of those packages on EL 6.5.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Sun, Feb 16, 2014 at 7:05 AM, Dafna Ron d...@redhat.com wrote:

 does the vm that fails migration have a live snapshot?
 if so how many snapshots does the vm have.
 I think that there are newer packages of vdsm, libvirt and qemu - can you
 try to update



 On 02/16/2014 12:33 AM, Steve Dainard wrote:

 Versions are the same:

 [root@ovirt001 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
 gpxe-roms-qemu-0.9.7-6.10.el6.noarch
 libvirt-0.10.2-29.el6_5.3.x86_64
 libvirt-client-0.10.2-29.el6_5.3.x86_64
 libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
 libvirt-python-0.10.2-29.el6_5.3.x86_64
 qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
 vdsm-4.13.3-3.el6.x86_64
 vdsm-cli-4.13.3-3.el6.noarch
 vdsm-gluster-4.13.3-3.el6.noarch
 vdsm-python-4.13.3-3.el6.x86_64
 vdsm-xmlrpc-4.13.3-3.el6.noarch

 [root@ovirt002 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
 gpxe-roms-qemu-0.9.7-6.10.el6.noarch
 libvirt-0.10.2-29.el6_5.3.x86_64
 libvirt-client-0.10.2-29.el6_5.3.x86_64
 libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
 libvirt-python-0.10.2-29.el6_5.3.x86_64
 qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
 vdsm-4.13.3-3.el6.x86_64
 vdsm-cli-4.13.3-3.el6.noarch
 vdsm-gluster-4.13.3-3.el6.noarch
 vdsm-python-4.13.3-3.el6.x86_64
 vdsm-xmlrpc-4.13.3-3.el6.noarch

 Logs attached, thanks.

 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/

 *Blog http://miovision.com/blog | **LinkedIn https://www.linkedin.com/
 company/miovision-technologies  | Twitter https://twitter.com/miovision
  | Facebook https://www.facebook.com/miovision*
 
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
 ON, Canada | N2C 1L3
 This e-mail may contain information that is privileged or confidential.
 If you are not the intended recipient, please delete the e-mail and any
 attachments and notify us immediately.


 On Sat, Feb 15, 2014 at 6:24 AM, Dafna Ron d...@redhat.com mailto:
 d...@redhat.com wrote:

 the migration fails in libvirt:


 Thread-153709::ERROR::2014-02-14
 11:17:40,420::vm::337::vm.Vm::(run)
 vmId=`08434c90-ffa3-4b63-aa8e-5613f7b0e0cd`::Failed to migrate
 Traceback (most recent call last):
   File /usr/share/vdsm/vm.py, line 323, in run
 self._startUnderlyingMigration()
   File /usr/share/vdsm/vm.py, line 403, in _startUnderlyingMigration
 None, maxBandwidth)
   File /usr/share/vdsm/vm.py, line 841, in f
 ret = attr(*args, **kwargs)
   File
 /usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py,
 line 76, in wrapper
 ret = f(*args, **kwargs)
   File /usr/lib64/python2.6/site-packages/libvirt.py, line 1178,
 in migrateToURI2
 if ret == -1: raise libvirtError ('virDomainMigrateToURI2()
 failed', dom=self)
 libvirtError: Unable to read from monitor: Connection reset by peer
 Thread-54041::DEBUG::2014-02-14
 11:17:41,752::task::579::TaskManager.Task::(_updateState)
 Task=`094c412a-43dc-4c29-a601-d759486469a8`::moving from state
 init - state preparing
 Thread-54041::INFO::2014-02-14
 11:17:41,753::logUtils::44::dispatcher::(wrapper) Run and protect:
 getVolumeSize(sdUUID='a52938f7-2cf4-4771-acb2-0c78d14999e5',
 spUUID='fcb89071-6cdb-4972-94d1-c9324cebf814',
 imgUUID='97c9108f-a506-415f-ad2
 c-370d707cb130', volUUID='61f82f7f-18e4-4ea8-9db3-71ddd9d4e836',
 options=None)

 Do you have the same libvirt/vdsm/qemu on both your hosts?
 Please attach the libvirt and vm logs from both hosts.

 Thanks,
 Dafna



 On 02/14/2014 04:50 PM, Steve Dainard wrote:

 Quick overview:
 Ovirt 3.3.2 running on CentOS 6.5
 Two hosts: ovirt001, ovirt002
 Migrating two VM's: puppet-agent1, puppet-agent2 from ovirt002
 to ovirt001.

 The first VM puppet-agent1 migrates successfully. The second
 VM puppet-agent2 fails with Migration failed due to Error:
 Fatal error during migration (VM: puppet-agent2

Re: [Users] Live migration of VM's occasionally fails

2014-02-17 Thread Steve Dainard
Failed live migration is wider spread than these two VM's, but they are a
good example because they were both built from the same template and have
no modifications after they were created. They were also migrated one after
the other, with one successfully migrating and the other not.

Are there any increased logging levels that might help determine what the
issue is?

Thanks,

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Mon, Feb 17, 2014 at 11:47 AM, Dafna Ron d...@redhat.com wrote:

 did you install these vm's from a cd? run it as run-once with a special
 monitor?
 try to think if there is anything different in the configuration of these
 vm's from the other vm's that succeed to migrate?


 On 02/17/2014 04:36 PM, Steve Dainard wrote:

 Hi Dafna,

 No snapshots of either of those VM's have been taken, and there are no
 updates for any of those packages on EL 6.5.

 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/

 *Blog http://miovision.com/blog | **LinkedIn https://www.linkedin.com/
 company/miovision-technologies  | Twitter https://twitter.com/miovision
  | Facebook https://www.facebook.com/miovision*
 
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
 ON, Canada | N2C 1L3
 This e-mail may contain information that is privileged or confidential.
 If you are not the intended recipient, please delete the e-mail and any
 attachments and notify us immediately.


 On Sun, Feb 16, 2014 at 7:05 AM, Dafna Ron d...@redhat.com mailto:
 d...@redhat.com wrote:

 does the vm that fails migration have a live snapshot?
 if so how many snapshots does the vm have.
 I think that there are newer packages of vdsm, libvirt and qemu -
 can you try to update



 On 02/16/2014 12:33 AM, Steve Dainard wrote:

 Versions are the same:

 [root@ovirt001 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
 gpxe-roms-qemu-0.9.7-6.10.el6.noarch
 libvirt-0.10.2-29.el6_5.3.x86_64
 libvirt-client-0.10.2-29.el6_5.3.x86_64
 libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
 libvirt-python-0.10.2-29.el6_5.3.x86_64
 qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
 vdsm-4.13.3-3.el6.x86_64
 vdsm-cli-4.13.3-3.el6.noarch
 vdsm-gluster-4.13.3-3.el6.noarch
 vdsm-python-4.13.3-3.el6.x86_64
 vdsm-xmlrpc-4.13.3-3.el6.noarch

 [root@ovirt002 ~]# rpm -qa | egrep 'libvirt|vdsm|qemu' | sort
 gpxe-roms-qemu-0.9.7-6.10.el6.noarch
 libvirt-0.10.2-29.el6_5.3.x86_64
 libvirt-client-0.10.2-29.el6_5.3.x86_64
 libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
 libvirt-python-0.10.2-29.el6_5.3.x86_64
 qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
 qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
 vdsm-4.13.3-3.el6.x86_64
 vdsm-cli-4.13.3-3.el6.noarch
 vdsm-gluster-4.13.3-3.el6.noarch
 vdsm-python-4.13.3-3.el6.x86_64
 vdsm-xmlrpc-4.13.3-3.el6.noarch

 Logs attached, thanks.

 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/

 *Blog http://miovision.com/blog | **LinkedIn
 https://www.linkedin.com/company/miovision-technologies  |
 Twitter https://twitter.com/miovision  | Facebook
 https://www.facebook.com/miovision*
 
 
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101,
 Kitchener, ON, Canada | N2C 1L3
 This e-mail may contain information that is privileged or
 confidential. If you are not the intended recipient, please
 delete the e-mail and any attachments and notify us immediately.


 On Sat, Feb 15, 2014 at 6:24 AM, Dafna Ron d...@redhat.com
 mailto:d...@redhat.com mailto:d...@redhat.com

 mailto:d...@redhat.com wrote:

 the migration fails in libvirt:


 Thread-153709::ERROR::2014-02-14
 11:17:40,420::vm::337::vm.Vm::(run)
 vmId=`08434c90-ffa3-4b63-aa8e-5613f7b0e0cd`::Failed to
 migrate

Re: [Users] Live migration of VM's occasionally fails

2014-02-17 Thread Steve Dainard
VM's are identical, same template, same cpu/mem/nic. Server type, thin
provisioned on NFS (backend is glusterfs 3.4).

Does monitor = spice console? I don't believe either of them had a spice
connection.

I don't see anything in the ovirt001 sanlock.log:

2014-02-14 11:16:05-0500 255246 [5111]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:16:05-0500 255246 [5111]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:16:15-0500 255256 [5110]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:16:15-0500 255256 [5110]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:16:25-0500 255266 [5111]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:16:25-0500 255266 [5111]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:16:36-0500 255276 [5110]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:16:36-0500 255276 [5110]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:16:46-0500 255286 [5111]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:16:46-0500 255286 [5111]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:16:56-0500 255296 [5110]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:16:56-0500 255296 [5110]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:17:06-0500 255306 [5111]: cmd_inq_lockspace 4,14
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:17:06-0500 255306 [5111]: cmd_inq_lockspace 4,14 done 0
2014-02-14 11:17:06-0500 255307 [5105]: cmd_register ci 4 fd 14 pid 31132
2014-02-14 11:17:06-0500 255307 [5105]: cmd_restrict ci 4 fd 14 pid 31132
flags 1
2014-02-14 11:17:16-0500 255316 [5110]: cmd_inq_lockspace 5,15
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:17:16-0500 255316 [5110]: cmd_inq_lockspace 5,15 done 0
2014-02-14 11:17:26-0500 255326 [5111]: cmd_inq_lockspace 5,15
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:17:26-0500 255326 [5111]: cmd_inq_lockspace 5,15 done 0
2014-02-14 11:17:26-0500 255326 [5110]: cmd_acquire 4,14,31132 ci_in 5 fd
15 count 0
2014-02-14 11:17:26-0500 255326 [5110]: cmd_acquire 4,14,31132 result 0
pid_dead 0
2014-02-14 11:17:26-0500 255326 [5111]: cmd_acquire 4,14,31132 ci_in 6 fd
16 count 0
2014-02-14 11:17:26-0500 255326 [5111]: cmd_acquire 4,14,31132 result 0
pid_dead 0
2014-02-14 11:17:36-0500 255336 [5110]: cmd_inq_lockspace 5,15
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:17:36-0500 255336 [5110]: cmd_inq_lockspace 5,15 done 0
2014-02-14 11:17:39-0500 255340 [5105]: cmd_register ci 5 fd 15 pid 31319
2014-02-14 11:17:39-0500 255340 [5105]: cmd_restrict ci 5 fd 15 pid 31319
flags 1
2014-02-14 11:17:39-0500 255340 [5105]: client_pid_dead 5,15,31319
cmd_active 0 suspend 0
2014-02-14 11:17:46-0500 255346 [5111]: cmd_inq_lockspace 5,15
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:17:46-0500 255346 [5111]: cmd_inq_lockspace 5,15 done 0
2014-02-14 11:17:56-0500 255356 [5110]: cmd_inq_lockspace 5,15
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0
2014-02-14 11:17:56-0500 255356 [5110]: cmd_inq_lockspace 5,15 done 0
2014-02-14 11:18:06-0500 255366 [5111]: cmd_inq_lockspace 5,15
a52938f7-2cf4-4771-acb2-0c78d14999e5:1:/rhev/data-center/mnt/gluster-store-vip:_rep1/a52938f7-2cf4-4771-acb2-0c78d14999e5/dom_md/ids:0
flags 0

ovirt002 sanlock.log has on entries during that time frame.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e

Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage domain

2014-02-09 Thread Steve Dainard
vdsm can write to it:

[root@ovirt001 rep2-mount]# sudo -u vdsm dd if=/dev/zero of=__test__ bs=1M
count=1 oflag=direct
1+0 records in
1+0 records out
1048576 bytes (1.0 MB) copied, 0.12691 s, 8.3 MB/s

[root@ovirt001 rep2-mount]# pwd
/mnt/storage/rep2-mount

[root@ovirt001 rep2-mount]# mount
/dev/md125p2 on / type ext4 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw,rootcontext=system_u:object_r:tmpfs_t:s0)
/dev/md125p1 on /boot type ext4 (rw)
/dev/mapper/gluster-storage--domain on /mnt/storage/lv-storage-domain type
xfs (rw)
/dev/mapper/gluster-iso--domain on /mnt/storage/lv-iso-domain type xfs (rw)
/dev/mapper/gluster-vm--domain on /mnt/storage/lv-vm-domain type xfs (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
10.0.10.2:/iso-store on /mnt/storage/iso-mount type fuse.glusterfs
(rw,default_permissions,allow_other,max_read=131072)
10.0.10.2:/rep2 on /mnt/storage/rep2-mount type fuse.glusterfs
(rw,default_permissions,allow_other,max_read=131072)

[root@ovirt001 rep2-mount]# gluster volume info rep2
Volume Name: rep2
Type: Replicate
Volume ID: b89a21bb-5ad1-493f-b197-8f990ab3ba77
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 10.0.10.2:/mnt/storage/lv-vm-domain/rep2
Brick2: 10.0.10.3:/mnt/storage/lv-vm-domain/rep2
Options Reconfigured:
storage.owner-gid: 36
storage.owner-uid: 36
server.allow-insecure: on
cluster.quorum-type: auto

Thanks,

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Sun, Feb 9, 2014 at 1:30 AM, Nir Soffer nsof...@redhat.com wrote:

 - Original Message -
  From: Steve Dainard sdain...@miovision.com
  To: Nir Soffer nsof...@redhat.com
  Cc: users users@ovirt.org
  Sent: Sunday, February 9, 2014 3:51:03 AM
  Subject: Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage
 domain
 
  Hi Nir,
 
  [root@ovirt001 storage]# mount -t glusterfs 10.0.10.2:/rep2 rep2-mount/
  [root@ovirt001 storage]# ls -lh rep2-mount/
  total 0
  -rwxr-xr-x. 1 vdsm kvm  0 Feb  5 10:36 __DIRECT_IO_TEST__
  drwxr-xr-x. 4 vdsm kvm 32 Feb  5 10:36
 ff0e0521-a8fa-4c10-8372-7b67ac3fca31
  [root@ovirt001 storage]# ls -lh
  total 0
  drwxr-xr-x. 4 vdsm kvm  91 Jan 30 17:34 iso-mount
  drwxr-xr-x. 3 root root 23 Jan 30 17:31 lv-iso-domain
  drwxr-xr-x. 3 vdsm kvm  35 Jan 29 17:43 lv-storage-domain
  drwxr-xr-x. 3 vdsm kvm  17 Feb  4 15:43 lv-vm-domain
  drwxr-xr-x. 4 vdsm kvm  91 Feb  5 10:36 rep2-mount

 Looks good.

 Can you write into rep2-mount?

 Please try:

 sudo -u vdsm dd if=/dev/zero of=rep2-mount/__test__ bs=1M count=1
 oflag=direct
 rm rep2-mount/__test__

 Thanks,
 Nir

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage domain

2014-02-08 Thread Steve Dainard
Hi Nir,

[root@ovirt001 storage]# mount -t glusterfs 10.0.10.2:/rep2 rep2-mount/
[root@ovirt001 storage]# ls -lh rep2-mount/
total 0
-rwxr-xr-x. 1 vdsm kvm  0 Feb  5 10:36 __DIRECT_IO_TEST__
drwxr-xr-x. 4 vdsm kvm 32 Feb  5 10:36 ff0e0521-a8fa-4c10-8372-7b67ac3fca31
[root@ovirt001 storage]# ls -lh
total 0
drwxr-xr-x. 4 vdsm kvm  91 Jan 30 17:34 iso-mount
drwxr-xr-x. 3 root root 23 Jan 30 17:31 lv-iso-domain
drwxr-xr-x. 3 vdsm kvm  35 Jan 29 17:43 lv-storage-domain
drwxr-xr-x. 3 vdsm kvm  17 Feb  4 15:43 lv-vm-domain
drwxr-xr-x. 4 vdsm kvm  91 Feb  5 10:36 rep2-mount


*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Sat, Feb 8, 2014 at 5:28 PM, Nir Soffer nsof...@redhat.com wrote:

 - Original Message -
  From: Steve Dainard sdain...@miovision.com
  To: Nir Soffer nsof...@redhat.com, users users@ovirt.org
  Sent: Friday, February 7, 2014 6:27:04 PM
  Subject: Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage
 domain
 
  Hi Nir,
 
  Do you need any more info from me? I missed the 'sanlock client
 host_status
  -D' request, info below:

 Hi Steve,

 It looks like your glusterfs mount is not writable by vdsm or sanlock.
 This can
 happen when permissions or owner of the directory is not correct.

 Can you try to mount the glusterfs volume manually, and share the output
 of ls -lh?

 sudo mkdir /tmp/gluster
 sudo mount -t glusterfs 10.0.10.2:/rep2 /tmp/gluster
 ls -lh /tmp/gluster/
 sudo umount /tmp/gluster

 Thanks,
 Nir

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage domain

2014-02-04 Thread Steve Dainard
::TaskManager.Task::(_decref)
Task=`66924dbf-5a1c-473e-a158-d038aae38dc3`::ref 0 aborting True
Thread-31::DEBUG::2014-02-04
09:54:08,827::task::909::TaskManager.Task::(_doAbort)
Task=`66924dbf-5a1c-473e-a158-d038aae38dc3`::Task._doAbort: force False
Thread-31::DEBUG::2014-02-04
09:54:08,827::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-31::DEBUG::2014-02-04
09:54:08,827::task::579::TaskManager.Task::(_updateState)
Task=`66924dbf-5a1c-473e-a158-d038aae38dc3`::moving from state preparing -
state aborting
Thread-31::DEBUG::2014-02-04
09:54:08,827::task::534::TaskManager.Task::(__state_aborting)
Task=`66924dbf-5a1c-473e-a158-d038aae38dc3`::_aborting: recover policy none
Thread-31::DEBUG::2014-02-04
09:54:08,827::task::579::TaskManager.Task::(_updateState)
Task=`66924dbf-5a1c-473e-a158-d038aae38dc3`::moving from state aborting -
state failed
Thread-31::DEBUG::2014-02-04
09:54:08,827::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources
{'Storage.471487ed-2946-4dfc-8ec3-96546006be12':  ResourceRef
'Storage.471487ed-2946-4dfc-8ec3-96546006be12', isValid: 'True' obj:
'None', 'Storage.8c4e8898-c91a-4d49-98e8-b6467791a9cc':  ResourceRef
'Storage.8c4e8898-c91a-4d49-98e8-b6467791a9cc', isValid: 'True' obj:
'None'}
Thread-31::DEBUG::2014-02-04
09:54:08,828::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-31::DEBUG::2014-02-04
09:54:08,828::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource 'Storage.471487ed-2946-4dfc-8ec3-96546006be12'
Thread-31::DEBUG::2014-02-04
09:54:08,828::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.471487ed-2946-4dfc-8ec3-96546006be12' (0 active
users)
Thread-31::DEBUG::2014-02-04
09:54:08,828::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.471487ed-2946-4dfc-8ec3-96546006be12' is free, finding
out if anyone is waiting for it.
Thread-31::DEBUG::2014-02-04
09:54:08,828::resourceManager::648::ResourceManager::(releaseResource) No
one is waiting for resource 'Storage.471487ed-2946-4dfc-8ec3-96546006be12',
Clearing records.
Thread-31::DEBUG::2014-02-04
09:54:08,828::resourceManager::615::ResourceManager::(releaseResource)
Trying to release resource 'Storage.8c4e8898-c91a-4d49-98e8-b6467791a9cc'
Thread-31::DEBUG::2014-02-04
09:54:08,829::resourceManager::634::ResourceManager::(releaseResource)
Released resource 'Storage.8c4e8898-c91a-4d49-98e8-b6467791a9cc' (0 active
users)
Thread-31::DEBUG::2014-02-04
09:54:08,829::resourceManager::640::ResourceManager::(releaseResource)
Resource 'Storage.8c4e8898-c91a-4d49-98e8-b6467791a9cc' is free, finding
out if anyone is waiting for it.
Thread-31::DEBUG::2014-02-04
09:54:08,829::resourceManager::648::ResourceManager::(releaseResource) No
one is waiting for resource 'Storage.8c4e8898-c91a-4d49-98e8-b6467791a9cc',
Clearing records.
Thread-31::ERROR::2014-02-04
09:54:08,829::dispatcher::67::Storage.Dispatcher.Protect::(run) {'status':
{'message': Cannot acquire host id:
('471487ed-2946-4dfc-8ec3-96546006be12', SanlockException(22, 'Sanlock
lockspace add failure', 'Invalid argument')), 'code': 661}}


*Storage domain metadata file:*

CLASS=Data
DESCRIPTION=gluster-store-rep2
IOOPTIMEOUTSEC=10
LEASERETRIES=3
LEASETIMESEC=60
LOCKPOLICY=
LOCKRENEWALINTERVALSEC=5
POOL_UUID=
REMOTE_PATH=10.0.10.3:/rep2
ROLE=Regular
SDUUID=471487ed-2946-4dfc-8ec3-96546006be12
TYPE=POSIXFS
VERSION=3
_SHA_CKSUM=469191aac3fb8ef504b6a4d301b6d8be6fffece1



*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage domain

2014-02-04 Thread Steve Dainard
I should be able to provide any logs required, I've reverted to my NFS
storage domain but can move a host over to POSIX whenever necessary.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Tue, Feb 4, 2014 at 10:20 AM, Elad Ben Aharon ebena...@redhat.comwrote:

 Nir,
 Can you take a look? the user gets the same Sanlock exception as reported
 here:  https://bugzilla.redhat.com/show_bug.cgi?id=1046430

 - Original Message -
 From: Steve Dainard sdain...@miovision.com
 To: users users@ovirt.org
 Sent: Tuesday, February 4, 2014 5:09:43 PM
 Subject: [Users] Ovirt 3.3.2 Cannot attach POSIX (gluster) storage domain

 I can successfully create a POSIX storage domain backed by gluster, but at
 the end of creation I get an error message failed to acquire host id.

 Note that I have successfully created/activated NFS DC/SD on the same
 ovirt/hosts.

 I have some logs when I tried to attach to the DC after failure:

 engine.log

 2014-02-04 09:54:04,324 INFO
 [org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand]
 (ajp--127.0.0.1-8702-3) [1dd40406] Lock Acquired to object EngineLock [ex
 clusiveLocks= key: 8c4e8898-c91a-4d49-98e8-b6467791a9cc value: POOL
 , sharedLocks= ]
 2014-02-04 09:54:04,473 INFO
 [org.ovirt.engine.core.bll.storage.AddStoragePoolWithStoragesCommand]
 (pool-6-thread-42) [1dd40406] Running command: AddStoragePoolWithStorages
 Command internal: false. Entities affected : ID:
 8c4e8898-c91a-4d49-98e8-b6467791a9cc Type: StoragePool
 2014-02-04 09:54:04,673 INFO
 [org.ovirt.engine.core.bll.storage.ConnectStorageToVdsCommand]
 (pool-6-thread-42) [3f86c31b] Running command: ConnectStorageToVdsCommand
 intern
 al: true. Entities affected : ID: aaa0----123456789aaa
 Type: System
 2014-02-04 09:54:04,682 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand]
 (pool-6-thread-42) [3f86c31b] START, ConnectStorageServerVDSCommand(
 HostName = ovirt001, HostId = 48f13d47-8346-4ff6-81ca-4f4324069db3,
 storagePoolId = ----, storageType =
 POSIXFS, connectionList = [{ id: 87f9
 ff74-93c4-4fe5-9a56-ed5338290af9, connection: 10.0.10.3:/rep2, iqn: null,
 vfsType: glusterfs, mountOptions: null, nfsVersion: null, nfsRetrans: null,
 nfsTimeo: null };]), lo
 g id: 332ff091
 2014-02-04 09:54:05,089 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.ConnectStorageServerVDSCommand]
 (pool-6-thread-42) [3f86c31b] FINISH, ConnectStorageServerVDSCommand
 , return: {87f9ff74-93c4-4fe5-9a56-ed5338290af9=0}, log id: 332ff091
 2014-02-04 09:54:05,093 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand]
 (pool-6-thread-42) [3f86c31b] START, CreateStoragePoolVDSCommand(HostNa
 me = ovirt001, HostId = 48f13d47-8346-4ff6-81ca-4f4324069db3,
 storagePoolId=8c4e8898-c91a-4d49-98e8-b6467791a9cc, storageType=POSIXFS,
 storagePoolName=IT, masterDomainId=471
 487ed-2946-4dfc-8ec3-96546006be12,
 domainsIdList=[471487ed-2946-4dfc-8ec3-96546006be12], masterVersion=3), log
 id: 1be84579
 2014-02-04 09:54:08,833 ERROR
 [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStorag
 ePoolVDSCommand] (pool-6-thread-42) [3f86c31b] Failed in
 CreateStoragePoolVDS method
 2014-02-04 09:54:08,834 ERROR
 [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand]
 (pool-6-thread-42) [3f86c31b] Error code AcquireHostIdFailure and error
 message VDSGenericException: VDSErrorException: Failed to
 CreateStoragePoolVDS, error = Cannot acquire host id:
 ('471487ed-2946-4dfc-8ec3-96546006be12', SanlockException(22, 'Sanlock
 lockspace add failure', 'Invalid argument'))
 2014-02-04 09:54:08,835 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand]
 (pool-6-thread-42) [3f86c31b] Command
 org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand
 return value
 StatusOnlyReturnForXmlRpc [mStatus=StatusForXmlRpc [mCode=661,
 mMessage=Cannot acquire host id: ('471487ed-2946-4dfc-8ec3-96546006be12',
 SanlockException(22, 'Sanlock lockspace add failure', 'Invalid argument'))]]
 2014-02-04 09:54:08,836 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand]
 (pool-6-thread-42) [3f86c31b] HostName = ovirt001
 2014-02-04 09:54:08,840 ERROR
 [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateStoragePoolVDSCommand]
 (pool-6-thread-42) [3f86c31b] Command CreateStoragePoolVDS execution
 failed. Exception

Re: [Users] oVirt 3.3.3 RC EL6 Live Snapshot

2014-02-03 Thread Steve Dainard
[root@ovirt002 ~]# vdsClient -s 0 getStorageDomainInfo
a52938f7-2cf4-4771-acb2-0c78d14999e5
uuid = a52938f7-2cf4-4771-acb2-0c78d14999e5
pool = ['fcb89071-6cdb-4972-94d1-c9324cebf814']
lver = 5
version = 3
role = Master
remotePath = gluster-store-vip:/rep1
spm_id = 2
type = NFS
class = Data
master_ver = 1
name = gluster-store-rep1


*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Sun, Feb 2, 2014 at 2:55 PM, Dafna Ron d...@redhat.com wrote:

 please run vdsClient -s 0 getStorageDomainInfo a52938f7-2cf4-4771-acb2-
 0c78d14999e5

 Thanks,

 Dafna



 On 02/02/2014 03:02 PM, Steve Dainard wrote:

 Logs attached with VM running on qemu-kvm-rhev packages installed.

 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/
 519-513-2407 ex.250

 877-646-8476 (toll-free)

 *Blog http://miovision.com/blog | **LinkedIn https://www.linkedin.com/
 company/miovision-technologies  | Twitter https://twitter.com/miovision
  | Facebook https://www.facebook.com/miovision*
 
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
 ON, Canada | N2C 1L3
 This e-mail may contain information that is privileged or confidential.
 If you are not the intended recipient, please delete the e-mail and any
 attachments and notify us immediately.


 On Sun, Feb 2, 2014 at 5:05 AM, Dafna Ron d...@redhat.com mailto:
 d...@redhat.com wrote:

 can you please upload full engine, vdsm, libvirt and vm's qemu logs?


 On 02/02/2014 02:08 AM, Steve Dainard wrote:

 I have two CentOS 6.5 Ovirt hosts (ovirt001, ovirt002)

 I've installed the applicable qemu-kvm-rhev packages from this
 site: http://www.dreyou.org/ovirt/vdsm32/Packages/ on ovirt002.

 On ovirt001 if I take a live snapshot:

 Snapshot 'test qemu-kvm' creation for VM 'snapshot-test' was
 initiated by admin@internal.
 The VM is paused
 Failed to create live snapshot 'test qemu-kvm' for VM
 'snapshot-test'. VM restart is recommended.
 Failed to complete snapshot 'test qemu-kvm' creation for VM
 'snapshot-test'.
 The VM is then started, and the status for the snapshot
 changes to OK.

 On ovirt002 (with the packages from dreyou) I don't get any
 messages about a snapshot failing, but my VM is still paused
 to complete the snapshot. Is there something else other than
 the qemu-kvm-rhev packages that would enable this functionality?

 I've looked for some information on when the packages would be
 built as required in the CentOS repos, but I don't see
 anything definitive.

 http://lists.ovirt.org/pipermail/users/2013-December/019126.html
 Looks like one of the maintainers is waiting for someone to
 tell him what flags need to be set.

 Also, another thread here:
 http://comments.gmane.org/gmane.comp.emulators.ovirt.arch/1618
 same maintainer, mentioning that he hasn't seen anything in
 the bug tracker.

 There is a bug here:
 https://bugzilla.redhat.com/show_bug.cgi?id=1009100 that seems
 to have ended in finding a way for qemu to expose whether it
 supports live snapshots, rather than figuring out how to get
 the CentOS team the info they need to build the packages with
 the proper flags set.

 I have bcc'd both dreyou (packaged the qemu-kvm-rhev packages
 listed above) and Russ (CentOS maintainer mentioned in the
 other threads) if they wish to chime in and perhaps
 collaborate on which flags, if any, should be set for the
 qemu-kvm builds so we can get a CentOS bug report going and
 hammer this out.

 Thanks everyone.

 **crosses fingers and hopes for live snapshots soon**



 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/
 519-513-2407 tel:519-513-2407 tel:519-513-2407
 tel:519-513-2407 ex.250
 877-646-8476 tel:877-646-8476 tel:877-646-8476

 tel:877-646-8476 (toll-free)

 *Blog http://miovision.com/blog | **LinkedIn
 https://www.linkedin.com/company/miovision-technologies  |
 Twitter https://twitter.com

Re: [Users] oVirt 3.3.3 RC EL6 Live Snapshot

2014-02-03 Thread Steve Dainard
[root@ovirt002 ~]# rpm -qa | egrep 'qemu|vdsm|libvirt' | sort
gpxe-roms-qemu-0.9.7-6.10.el6.noarch
libvirt-0.10.2-29.el6_5.3.x86_64
libvirt-client-0.10.2-29.el6_5.3.x86_64
libvirt-lock-sanlock-0.10.2-29.el6_5.3.x86_64
libvirt-python-0.10.2-29.el6_5.3.x86_64
qemu-img-rhev-0.12.1.2-2.355.el6.5.x86_64
qemu-kvm-rhev-0.12.1.2-2.355.el6.5.x86_64
qemu-kvm-rhev-tools-0.12.1.2-2.355.el6.5.x86_64
vdsm-4.13.3-2.el6.x86_64
vdsm-cli-4.13.3-2.el6.noarch
vdsm-python-4.13.3-2.el6.x86_64
vdsm-xmlrpc-4.13.3-2.el6.noarch


*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Mon, Feb 3, 2014 at 11:54 AM, Dafna Ron d...@redhat.com wrote:

 Can you also put the vdsm, libvirt and qemu packages?

 Thanks,
 Dafna



 On 02/03/2014 04:49 PM, Steve Dainard wrote:

 FYI I'm running version 3.3.2, not the 3.3.3 beta.

 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/
 519-513-2407 ex.250
 877-646-8476 (toll-free)

 *Blog http://miovision.com/blog | **LinkedIn https://www.linkedin.com/
 company/miovision-technologies  | Twitter https://twitter.com/miovision
  | Facebook https://www.facebook.com/miovision*
 
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
 ON, Canada | N2C 1L3
 This e-mail may contain information that is privileged or confidential.
 If you are not the intended recipient, please delete the e-mail and any
 attachments and notify us immediately.


 On Mon, Feb 3, 2014 at 11:24 AM, Dafna Ron d...@redhat.com mailto:
 d...@redhat.com wrote:

 Thanks Steve.

 from the logs I can see that the create snapshot succeeds and that
 the vm is resumed.
 the vm moves to pause as part of libvirt flows:

 2014-02-02 14:41:20.872+: 5843: debug :
 qemuProcessHandleStop:728 : Transitioned guest snapshot-test to
 paused state
 2014-02-02 14:41:30.031+: 5843: debug :
 qemuProcessHandleResume:776 : Transitioned guest snapshot-test out
 of paused into resumed state

 There are bugs here but I am not sure yet if this is libvirt
 regression or engine.

 I'm adding Elad and Maor since in engine logs I can't see anything
 calling for live snapshot (only for snapshot) - Maor, shouldn't
 live snapshot command be logged somewhere in the logs?
 Is it possible that engine is calling to create snapshot and not
 create live snapshot which is why the vm pauses?

 Elad, if engine is not logging live snapshot anywhere I would open
 a bug for engine (to print that in the logs).
 Also, there is a bug in vdsm log for sdc where the below is logged
 as ERROR and not INFO:

 Thread-23::ERROR::2014-02-02
 09:51:19,497::sdc::137::Storage.StorageDomainCache::(_findDomain)
 looking for unfetched domain a52938f7-2cf4-4771-acb2-0c78d14999e5
 Thread-23::ERROR::2014-02-02
 09:51:19,497::sdc::154::Storage.StorageDomainCache::(_
 findUnfetchedDomain)
 looking for domain a52938f7-2cf4-4771-acb2-0c78d14999e5

 If the engine was sending live snapshot or if there is no
 difference in the two commands in engine side than I would open a
 bug for libvirt for pausing the vm during live snapshot.

 Dafna


 On 02/03/2014 02:41 PM, Steve Dainard wrote:

 [root@ovirt002 ~]# vdsClient -s 0 getStorageDomainInfo
 a52938f7-2cf4-4771-acb2-0c78d14999e5
 uuid = a52938f7-2cf4-4771-acb2-0c78d14999e5
 pool = ['fcb89071-6cdb-4972-94d1-c9324cebf814']
 lver = 5
 version = 3
 role = Master
 remotePath = gluster-store-vip:/rep1
 spm_id = 2
 type = NFS
 class = Data
 master_ver = 1
 name = gluster-store-rep1


 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | /Rethink Traffic/
 519-513-2407 tel:519-513-2407 ex.250
 877-646-8476 tel:877-646-8476 (toll-free)

 *Blog http://miovision.com/blog | **LinkedIn
 https://www.linkedin.com/company/miovision-technologies  |
 Twitter https://twitter.com/miovision  | Facebook
 https://www.facebook.com/miovision*
 
 
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101,
 Kitchener, ON, Canada | N2C 1L3

Re: [Users] oVirt 3.3.3 RC EL6 Live Snapshot

2014-02-01 Thread Steve Dainard
I have two CentOS 6.5 Ovirt hosts (ovirt001, ovirt002)

I've installed the applicable qemu-kvm-rhev packages from this site:
http://www.dreyou.org/ovirt/vdsm32/Packages/ on ovirt002.

On ovirt001 if I take a live snapshot:

Snapshot 'test qemu-kvm' creation for VM 'snapshot-test' was initiated by
admin@internal.
The VM is paused
Failed to create live snapshot 'test qemu-kvm' for VM 'snapshot-test'. VM
restart is recommended.
Failed to complete snapshot 'test qemu-kvm' creation for VM 'snapshot-test'.
The VM is then started, and the status for the snapshot changes to OK.

On ovirt002 (with the packages from dreyou) I don't get any messages about
a snapshot failing, but my VM is still paused to complete the snapshot. Is
there something else other than the qemu-kvm-rhev packages that would
enable this functionality?

I've looked for some information on when the packages would be built as
required in the CentOS repos, but I don't see anything definitive.

http://lists.ovirt.org/pipermail/users/2013-December/019126.html Looks like
one of the maintainers is waiting for someone to tell him what flags need
to be set.

Also, another thread here:
http://comments.gmane.org/gmane.comp.emulators.ovirt.arch/1618 same
maintainer, mentioning that he hasn't seen anything in the bug tracker.

There is a bug here: https://bugzilla.redhat.com/show_bug.cgi?id=1009100 that
seems to have ended in finding a way for qemu to expose whether it supports
live snapshots, rather than figuring out how to get the CentOS team the
info they need to build the packages with the proper flags set.

I have bcc'd both dreyou (packaged the qemu-kvm-rhev packages listed above)
and Russ (CentOS maintainer mentioned in the other threads) if they wish to
chime in and perhaps collaborate on which flags, if any, should be set for
the qemu-kvm builds so we can get a CentOS bug report going and hammer this
out.

Thanks everyone.

**crosses fingers and hopes for live snapshots soon**



*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Fri, Jan 31, 2014 at 1:26 PM, Steve Dainard sdain...@miovision.comwrote:


 How would you developers, speaking for the oVirt-community, propose to
 solve this for CentOS _now_ ?

 I would imagine that the easiest way is that you build and host this one
 package(qemu-kvm-rhev), since you´ve basically already have the source
 and recipe (since you´re already providing it for RHEV anyway). Then,
 once that´s in place, it´s more a question of where to host the
 packages, in what repository. Be it your own, or some other repo set up
 for the SIG.

 This is my view, how I as a user view this issue.



 I think this is a pretty valid view.

 What would it take to get the correct qemu package hosted in the ovirt
 repo?



  --

 Med Vänliga Hälsningar


 ---
 Karli Sjöberg
 Swedish University of Agricultural Sciences Box 7079 (Visiting Address
 Kronåsvägen 8)
 S-750 07 Uppsala, Sweden
 Phone:  +46-(0)18-67 15 66
 karli.sjob...@slu.se
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-31 Thread Steve Dainard
I've reconfigured my setup (good succes below, but need clarity on gluster
option):

Two nodes total, both running virt and glusterfs storage (2 node replica,
quorum).

I've created an NFS storage domain, pointed at the first nodes IP address.
I've launched a 2008 R2 SP1 install with a virtio-scsi disk, and the SCSI
pass-through driver on the same node as the NFS domain is pointing at.

Windows guest install has been running for roughly 1.5 hours, still
Expanding Windows files (55%) ...

top is showing:
  PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND

 3609 root  20   0 1380m  33m 2604 S 35.4  0.1 231:39.75 glusterfsd

21444 qemu  20   0 6362m 4.1g 6592 S 10.3  8.7  10:11.53 qemu-kvm


This is a 2 socket, 6 core xeon machine with 48GB of RAM, and 6x 7200rpm
enterprise sata disks in RAID5 so I don't think we're hitting hardware
limitations.

dd on xfs (no gluster)

time dd if=/dev/zero of=test bs=1M count=2048
2048+0 records in
2048+0 records out
2147483648 bytes (2.1 GB) copied, 4.15787 s, 516 MB/s

real 0m4.351s
user 0m0.000s
sys 0m1.661s


time dd if=/dev/zero of=test bs=1k count=200
200+0 records in
200+0 records out
204800 bytes (2.0 GB) copied, 4.06949 s, 503 MB/s

real 0m4.260s
user 0m0.176s
sys 0m3.991s


I've enabled nfs.trusted-sync (
http://gluster.org/community/documentation/index.php/Gluster_3.2:_Setting_Volume_Options#nfs.trusted-sync
) on the gluster volume, and the speed difference is immeasurable . Can
anyone explain what this option does, and what the risks are with a 2 node
gluster replica volume with quorum enabled?

Thanks,
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-31 Thread Steve Dainard


 I've enabled nfs.trusted-sync (
 http://gluster.org/community/documentation/index.php/Gluster_3.2:_Setting_Volume_Options#nfs.trusted-sync
 ) on the gluster volume, and the speed difference is immeasurable . Can
 anyone explain what this option does, and what the risks are with a 2
 node gluster replica volume with quorum enabled?


Sorry, I understand async, I meant option nfs.trusted-write, and if it
would help this situation.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] oVirt 3.3.3 RC EL6 Live Snapshot

2014-01-31 Thread Steve Dainard


 How would you developers, speaking for the oVirt-community, propose to
 solve this for CentOS _now_ ?

 I would imagine that the easiest way is that you build and host this one
 package(qemu-kvm-rhev), since you´ve basically already have the source
 and recipe (since you´re already providing it for RHEV anyway). Then,
 once that´s in place, it´s more a question of where to host the
 packages, in what repository. Be it your own, or some other repo set up
 for the SIG.

 This is my view, how I as a user view this issue.



I think this is a pretty valid view.

What would it take to get the correct qemu package hosted in the ovirt repo?



 --

 Med Vänliga Hälsningar


 ---
 Karli Sjöberg
 Swedish University of Agricultural Sciences Box 7079 (Visiting Address
 Kronåsvägen 8)
 S-750 07 Uppsala, Sweden
 Phone:  +46-(0)18-67 15 66
 karli.sjob...@slu.se
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-31 Thread Steve Dainard
IDE is just as slow. Just over 2 hours for 2008R2 install.

Is this what you mean by kvm?
lsmod | grep kvm
kvm_intel  54285  3
kvm   332980  1 kvm_intel


*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Fri, Jan 31, 2014 at 8:20 PM, Vadim Rozenfeld vroze...@redhat.comwrote:

 On Fri, 2014-01-31 at 11:37 -0500, Steve Dainard wrote:
  I've reconfigured my setup (good succes below, but need clarity on
  gluster option):
 
 
  Two nodes total, both running virt and glusterfs storage (2 node
  replica, quorum).
 
 
  I've created an NFS storage domain, pointed at the first nodes IP
  address. I've launched a 2008 R2 SP1 install with a virtio-scsi disk,
  and the SCSI pass-through driver on the same node as the NFS domain is
  pointing at.
 
 
  Windows guest install has been running for roughly 1.5 hours, still
  Expanding Windows files (55%) ...

 [VR]
 Does it work faster with IDE?
 Do you have kvm enabled?
 Thanks,
 Vadim.

 
 
  top is showing:
PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND
 
   3609 root  20   0 1380m  33m 2604 S 35.4  0.1 231:39.75
  glusterfsd
  21444 qemu  20   0 6362m 4.1g 6592 S 10.3  8.7  10:11.53 qemu-kvm
 
 
 
  This is a 2 socket, 6 core xeon machine with 48GB of RAM, and 6x
  7200rpm enterprise sata disks in RAID5 so I don't think we're hitting
  hardware limitations.
 
 
  dd on xfs (no gluster)
 
 
  time dd if=/dev/zero of=test bs=1M count=2048
  2048+0 records in
  2048+0 records out
  2147483648 bytes (2.1 GB) copied, 4.15787 s, 516 MB/s
 
 
  real 0m4.351s
  user 0m0.000s
  sys 0m1.661s
 
 
 
 
  time dd if=/dev/zero of=test bs=1k count=200
  200+0 records in
  200+0 records out
  204800 bytes (2.0 GB) copied, 4.06949 s, 503 MB/s
 
 
  real 0m4.260s
  user 0m0.176s
  sys 0m3.991s
 
 
 
 
  I've enabled nfs.trusted-sync
  (
 http://gluster.org/community/documentation/index.php/Gluster_3.2:_Setting_Volume_Options#nfs.trusted-sync)
 on the gluster volume, and the speed difference is immeasurable . Can
 anyone explain what this option does, and what the risks are with a 2 node
 gluster replica volume with quorum enabled?
 
 
  Thanks,



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] engine-backup restore how to

2014-01-30 Thread Steve Dainard
Is this file supposed to exist:
2014-01-30 10:24:18,990 WARN  [org.ovirt.engine.core.utils.LocalConfig]
(MSC service thread 1-23) The file /etc/ovirt-engine/engine.conf doesn't
exist or isn't readable. Will return an empty set of properties.

I can't find it anywhere on the system.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Thu, Jan 30, 2014 at 10:32 AM, Steve Dainard sdain...@miovision.comwrote:

 I can connect to the db with the 'engine' user.

 Initially 'engine' wasn't a member of any roles, I added it to 'postgres'.
 engine= \du
 List of roles
  Role name | Attributes  | Member of
 ---+-+
  engine| | {postgres}
  postgres  | Superuser   | {}
: Create role
: Create DB

 If you meant file permissions they are:
 # ll
 total 24
 -rw---. 1 ovirt ovirt 380 Jan 29 18:35 10-setup-database.conf
 -rw---. 1 ovirt ovirt 378 Jan 15 15:58
 10-setup-database.conf.20140129183539
 -rw-r--r--. 1 root  root   33 Jan 15 15:58 10-setup-jboss.conf
 -rw---. 1 ovirt ovirt 384 Jan 15 15:59 10-setup-pki.conf
 -rw-r--r--. 1 root  root  259 Jan 15 15:58 10-setup-protocols.conf
 -rw-r--r--. 1 root  root  204 Dec 13 03:22 README


 On ovirt-engine restart (engine.log):
 i
 2014-01-30 10:24:18,988 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Loaded file
 /usr/share/ovirt-engine/services/ovirt-engine/ovirt-engine.conf.
 2014-01-30 10:24:18,990 WARN  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) The file /etc/ovirt-engine/engine.conf doesn't
 exist or isn't readable. Will return an empty set of properties.
 2014-01-30 10:24:18,991 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Loaded file
 /etc/ovirt-engine/engine.conf.d/10-setup-database.conf.
 2014-01-30 10:24:18,992 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Loaded file
 /etc/ovirt-engine/engine.conf.d/10-setup-jboss.conf.
 2014-01-30 10:24:18,994 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Loaded file
 /etc/ovirt-engine/engine.conf.d/10-setup-pki.conf.
 2014-01-30 10:24:18,994 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Loaded file
 /etc/ovirt-engine/engine.conf.d/10-setup-protocols.conf.
 2014-01-30 10:24:18,995 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_AJP_ENABLED is true.
 2014-01-30 10:24:18,996 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_AJP_PORT is 8702.
 2014-01-30 10:24:18,997 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_APPS is engine.ear.
 2014-01-30 10:24:18,997 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_CACHE is
 /var/cache/ovirt-engine.
 2014-01-30 10:24:18,998 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_CHECK_INTERVAL is
 1000.
 2014-01-30 10:24:18,998 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_CONNECTION_TIMEOUT
 is 30.
 2014-01-30 10:24:18,999 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_DATABASE is
 engine.
 2014-01-30 10:24:19,000 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_DRIVER is
 org.postgresql.Driver.
 2014-01-30 10:24:19,000 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_HOST is localhost.
 2014-01-30 10:24:19,001 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_MAX_CONNECTIONS is
 100.
 2014-01-30 10:24:19,001 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_MIN_CONNECTIONS is
 1.
 2014-01-30 10:24:19,002 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_PASSWORD is ***.
 2014-01-30 10:24:19,002 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value of property ENGINE_DB_PORT is 5432.
 2014-01-30 10:24:19,003 INFO  [org.ovirt.engine.core.utils.LocalConfig]
 (MSC service thread 1-23) Value

Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-29 Thread Steve Dainard
On Wed, Jan 29, 2014 at 5:11 AM, Vadim Rozenfeld vroze...@redhat.comwrote:

 On Wed, 2014-01-29 at 11:30 +0200, Ronen Hod wrote:
  Adding the virtio-scsi developers.
  Anyhow, virtio-scsi is newer and less established than viostor (the
  block device), so you might want to try it out.

 [VR]
 Was it SCSI Controller or SCSI pass-through controller?
 If it's SCSI Controller then it will be viostor (virtio-blk) device
 driver.


SCSI Controller is listed in device manager.

Hardware ID's:
PCI\VEN_1AF4DEV_1004SUBSYS_00081AF4REV_00
PCI\VEN_1AF4DEV_1004SUBSYS_00081AF4
PCI\VEN_1AF4DEV_1004CC_01
PCI\VEN_1AF4DEV_1004CC_0100




  A disclaimer: There are time and patches gaps between RHEL and other
  versions.
 
  Ronen.
 
  On 01/28/2014 10:39 PM, Steve Dainard wrote:
 
   I've had a bit of luck here.
  
  
   Overall IO performance is very poor during Windows updates, but a
   contributing factor seems to be the SCSI Controller device in the
   guest. This last install I didn't install a driver for that device,

 [VR]
 Does it mean that your system disk is IDE and the data disk (virtio-blk)
 is not accessible?


In Ovirt 3.3.2-1.el6 I do not have an option to add a virtio-blk device:
Screenshot here:
https://dl.dropboxusercontent.com/u/21916057/Screenshot%20from%202014-01-29%2010%3A04%3A57.png

VM disk drive is Red Hat VirtIO SCSI Disk Device, storage controller is
listed as Red Hat VirtIO SCSI Controller as shown in device manager.
Screenshot here:
https://dl.dropboxusercontent.com/u/21916057/Screenshot%20from%202014-01-29%2009%3A57%3A24.png

In Ovirt manager the disk interface is listed as VirtIO.
Screenshot here:
https://dl.dropboxusercontent.com/u/21916057/Screenshot%20from%202014-01-29%2009%3A58%3A35.png



and my performance is much better. Updates still chug along quite
   slowly, but I seem to have more than the  100KB/s write speeds I
   was seeing previously.
  
  
   Does anyone know what this device is for? I have the Red Hat VirtIO
   SCSI Controller listed under storage controllers.

 [VR]
 It's a virtio-blk device. OS cannot see this volume unless you have
 viostor.sys driver installed on it.


Interesting that my VM's can see the controller, but I can't add a disk for
that controller in Ovirt. Is there a package I have missed on install?

rpm -qa | grep ovirt
ovirt-host-deploy-java-1.1.3-1.el6.noarch
ovirt-engine-backend-3.3.2-1.el6.noarch
ovirt-engine-lib-3.3.2-1.el6.noarch
ovirt-engine-restapi-3.3.2-1.el6.noarch
ovirt-engine-sdk-python-3.3.0.8-1.el6.noarch
ovirt-log-collector-3.3.2-2.el6.noarch
ovirt-engine-dbscripts-3.3.2-1.el6.noarch
ovirt-engine-webadmin-portal-3.3.2-1.el6.noarch
ovirt-host-deploy-1.1.3-1.el6.noarch
ovirt-image-uploader-3.3.2-1.el6.noarch
ovirt-engine-websocket-proxy-3.3.2-1.el6.noarch
ovirt-engine-userportal-3.3.2-1.el6.noarch
ovirt-engine-setup-3.3.2-1.el6.noarch
ovirt-iso-uploader-3.3.2-1.el6.noarch
ovirt-engine-cli-3.3.0.6-1.el6.noarch
ovirt-engine-3.3.2-1.el6.noarch
ovirt-engine-tools-3.3.2-1.el6.noarch


  
   I've setup a NFS storage domain on my desktops SSD.
   I've re-installed
   win 2008 r2 and initially it was running smoother.
  
   Disk performance peaks at 100MB/s.
  
   If I copy a 250MB file from a share into the Windows
   VM, it writes out
 [VR]
 Do you copy it with Explorer or any other copy program?


Windows Explorer only.


 Do you have HPET enabled?


I can't find it in the guest 'system devices'. On the hosts the current
clock source is 'tsc', although 'hpet' is an available option.


 How does it work with if you copy from/to local (non-NFS) storage?


Not sure, this is a royal pain to setup. Can I use my ISO domain in two
different data centers at the same time? I don't have an option to create
an ISO / NFS domain in the local storage DC.

When I use the import option with the default DC's ISO domain, I get an
error There is no storage domain under the specified path. Check event log
for more details. VDMS logs show Resource namespace
0e90e574-b003-4a62-867d-cf274b17e6b1_imageNS already registered so I'm
guessing the answer is no.

I tried to deploy with WDS, but the 64bit drivers apparently aren't signed,
and on x86 I get an error about the NIC not being supported even with the
drivers added to WDS.



 What is your virtio-win drivers package origin and version?


virtio-win-0.1-74.iso -
http://alt.fedoraproject.org/pub/alt/virtio-win/latest/images/



 Thanks,
 Vadim.



Appreciate it,
Steve
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] engine-backup restore how to

2014-01-29 Thread Steve Dainard
There doesn't seem to be any solid documentation on how to use the
engine-backup restore function, and I'm not able to restore a backup.

The best I've come up with is:
1. Install engine on new host
2. Stop engine
3. run engine-backup --mode=restore --file=filename --log=logfile

Fail.

Log shows:
psql: FATAL:  password authentication failed for user engine
2014-01-29 18:20:30 10285: FATAL: Can't connect to the database

4. engine-backup --mode=restore --file=engine.bak --log=logfile
--change-db-credentials --db-host=localhost --db-user=engine
--db-name=engine --db-password='newpassword'

Fails with same error.

5. change user to postgres, drop the old db, create a new db named engine,
set password for engine user same as 'newpassword'

6. engine-backup --mode=restore --file=engine.bak --log=logfile
--change-db-credentials --db-host=localhost --db-user=engine
--db-name=engine --db-password='newpassword'

Restoring...
Rewriting /etc/ovirt-engine/engine.conf.d/10-setup-database.conf
Note: you might need to manually fix:
- iptables/firewalld configuration
- autostart of ovirt-engine service
You can now start the engine service and then restart httpd
Done.

7. start ovirt-engine, restart httpd, browse to web ui

Blank page, no content.

8. stop firewall, browse to web ui

Blank page, no content

9. Engine log contains:

2014-01-29 18:35:56,973 INFO  [org.ovirt.engine.core.utils.LocalConfig]
(MSC service
thread 1-40) Value of property SENSITIVE_KEYS is
,ENGINE_DB_PASSWORD,ENGINE_PKI_TR
UST_STORE_PASSWORD,ENGINE_PKI_ENGINE_STORE_PASSWORD.
2014-01-29 18:35:57,330 ERROR [org.ovirt.engine.core.bll.Backend] (MSC
service thread
 1-25) Error in getting DB connection. The database is inaccessible.
Original exception is: BadSqlGrammarException: CallableStatementCallback;
bad SQL grammar [{call checkdbconnection()}]; nested exception is
org.postgresql.util.PSQLException: ERROR: function checkdbconnection() does
not exist
  Hint: No function matches the given name and argument types. You might
need to add explicit type casts.
  Position: 15
2014-01-29 18:35:58,336 ERROR [org.ovirt.engine.core.bll.Backend] (MSC
service thread 1-25) Error in getting DB connection. The database is
inaccessible. Original exception is: UncategorizedSQLException:
CallableStatementCallback; uncategorized SQLException for SQL [{call
checkdbconnection()}]; SQL state [25P02]; error code [0]; ERROR: current
transaction is aborted, commands ignored until end of transaction block;
nested exception is org.postgresql.util.PSQLException: ERROR: current
transaction is aborted, commands ignored until end of transaction block


*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] engine-backup restore how to

2014-01-29 Thread Steve Dainard
I also see this error in engine.log which repeats every second if I am
trying to access the web ui.

2014-01-29 18:59:47,531 ERROR [org.ovirt.engine.core.bll.Backend]
(ajp--127.0.0.1-8702-4) Error in getting DB connection. The database is
inaccessible. Original exception is: UncategorizedSQLException:
CallableStatementCallback; uncategorized SQLException for SQL [{call
checkdbconnection()}]; SQL state [25P02]; error code [0]; ERROR: current
transaction is aborted, commands ignored until end of transaction block;
nested exception is org.postgresql.util.PSQLException: ERROR: current
transaction is aborted, commands ignored until end of transaction block

It looks like the db inserted correctly, I took a quick look through some
tables and can see the valid admin user, and snapshots. But I can't say for
certain.

The IP address of the new server does not match the IP of the old (backup
file) server, would this have any impact? I would think not as its a local
db.

When I changed the password for the psql engine user, is there any config
file this is referenced in that may not have been updated?

Thanks,

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Wed, Jan 29, 2014 at 7:06 PM, Alon Bar-Lev alo...@redhat.com wrote:



 - Original Message -
  From: Steve Dainard sdain...@miovision.com
  To: users users@ovirt.org
  Sent: Thursday, January 30, 2014 1:59:08 AM
  Subject: [Users] engine-backup restore how to
 
  There doesn't seem to be any solid documentation on how to use the
  engine-backup restore function, and I'm not able to restore a backup.
 
  The best I've come up with is:
  1. Install engine on new host
  2. Stop engine
  3. run engine-backup --mode=restore --file=filename --log=logfile
 
  Fail.
 
  Log shows:
  psql: FATAL: password authentication failed for user engine
  2014-01-29 18:20:30 10285: FATAL: Can't connect to the database
 
  4. engine-backup --mode=restore --file=engine.bak --log=logfile
  --change-db-credentials --db-host=localhost --db-user=engine
  --db-name=engine --db-password='newpassword'
 
  Fails with same error.

 the --db-password must match the user's actual password within database,
 --change-db-credentials does not change the password in database but the
 host/port/user/password that are used by enigne.

 
  5. change user to postgres, drop the old db, create a new db named
 engine,
  set password for engine user same as 'newpassword'
 
  6. engine-backup --mode=restore --file=engine.bak --log=logfile
  --change-db-credentials --db-host=localhost --db-user=engine
  --db-name=engine --db-password='newpassword'

 Ok, this is correct now.

  Restoring...
  Rewriting /etc/ovirt-engine/engine.conf.d/10-setup-database.conf
  Note: you might need to manually fix:
  - iptables/firewalld configuration
  - autostart of ovirt-engine service
  You can now start the engine service and then restart httpd
  Done.
 
  7. start ovirt-engine, restart httpd, browse to web ui
 
  Blank page, no content.
 
  8. stop firewall, browse to web ui
 
  Blank page, no content
 
  9. Engine log contains:
 
  2014-01-29 18:35:56,973 INFO [org.ovirt.engine.core.utils.LocalConfig]
 (MSC
  service
  thread 1-40) Value of property SENSITIVE_KEYS is
  ,ENGINE_DB_PASSWORD,ENGINE_PKI_TR
  UST_STORE_PASSWORD,ENGINE_PKI_ENGINE_STORE_PASSWORD.
  2014-01-29 18:35:57,330 ERROR [org.ovirt.engine.core.bll.Backend] (MSC
  service thread
  1-25) Error in getting DB connection. The database is inaccessible.
 Original
  exception is: BadSqlGrammarException: CallableStatementCallback; bad SQL
  grammar [{call checkdbconnection()}]; nested exception is
  org.postgresql.util.PSQLException: ERROR: function checkdbconnection()
 does
  not exist
  Hint: No function matches the given name and argument types. You might
 need
  to add explicit type casts.
  Position: 15
  2014-01-29 18:35:58,336 ERROR [org.ovirt.engine.core.bll.Backend] (MSC
  service thread 1-25) Error in getting DB connection. The database is
  inaccessible. Original exception is: UncategorizedSQLException:
  CallableStatementCallback; uncategorized SQLException for SQL [{call
  checkdbconnection()}]; SQL state [25P02]; error code [0]; ERROR: current
  transaction is aborted, commands ignored until end of transaction block;
  nested exception is org.postgresql.util.PSQLException: ERROR: current
  transaction is aborted, commands ignored until end of transaction

Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-28 Thread Steve Dainard
I've had a bit of luck here.

Overall IO performance is very poor during Windows updates, but a
contributing factor seems to be the SCSI Controller device in the guest.
This last install I didn't install a driver for that device, and my
performance is much better. Updates still chug along quite slowly, but I
seem to have more than the  100KB/s write speeds I was seeing previously.

Does anyone know what this device is for? I have the Red Hat VirtIO SCSI
Controller listed under storage controllers.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Sun, Jan 26, 2014 at 2:33 AM, Itamar Heim ih...@redhat.com wrote:

 On 01/26/2014 02:37 AM, Steve Dainard wrote:

 Thanks for the responses everyone, really appreciate it.

 I've condensed the other questions into this reply.


 Steve,
 What is the CPU load of the GlusterFS host when comparing the raw
 brick test to the gluster mount point test? Give it 30 seconds and
 see what top reports. You'll probably have to significantly increase
 the count on the test so that it runs that long.

 - Nick



 Gluster mount point:

 *4K* on GLUSTER host
 [root@gluster1 rep2]# dd if=/dev/zero of=/mnt/rep2/test1 bs=4k
 count=50
 50+0 records in
 50+0 records out
 204800 tel:204800 bytes (2.0 GB) copied, 100.076 s, 20.5 MB/s


 Top reported this right away:
 PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND
   1826 root  20   0  294m  33m 2540 S 27.2  0.4   0:04.31 glusterfs
   2126 root  20   0 1391m  31m 2336 S 22.6  0.4  11:25.48 glusterfsd

 Then at about 20+ seconds top reports this:
PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND
   1826 root  20   0  294m  35m 2660 R 141.7  0.5   1:14.94 glusterfs
   2126 root  20   0 1392m  31m 2344 S 33.7  0.4  11:46.56 glusterfsd

 *4K* Directly on the brick:
 dd if=/dev/zero of=test1 bs=4k count=50
 50+0 records in
 50+0 records out
 204800 tel:204800 bytes (2.0 GB) copied, 4.99367 s, 410 MB/s


   7750 root  20   0  102m  648  544 R 50.3  0.0   0:01.52 dd
   7719 root  20   0 000 D  1.0  0.0   0:01.50 flush-253:2

 Same test, gluster mount point on OVIRT host:
 dd if=/dev/zero of=/mnt/rep2/test1 bs=4k count=50
 50+0 records in
 50+0 records out
 204800 tel:204800 bytes (2.0 GB) copied, 42.4518 s, 48.2 MB/s


PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND
   2126 root  20   0 1396m  31m 2360 S 40.5  0.4  13:28.89 glusterfsd


 Same test, on OVIRT host but against NFS mount point:
 dd if=/dev/zero of=/mnt/rep2-nfs/test1 bs=4k count=50
 50+0 records in
 50+0 records out
 204800 tel:204800 bytes (2.0 GB) copied, 18.8911 s, 108 MB/s


 PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND
   2141 root  20   0  550m 184m 2840 R 84.6  2.3  16:43.10 glusterfs
   2126 root  20   0 1407m  30m 2368 S 49.8  0.4  13:49.07 glusterfsd

 Interesting - It looks like if I use a NFS mount point, I incur a cpu
 hit on two processes instead of just the daemon. I also get much better
 performance if I'm not running dd (fuse) on the GLUSTER host.


 The storage servers are a bit older, but are both dual socket
 quad core

 opterons with 4x 7200rpm drives.


 A block size of 4k is quite small so that the context switch
 overhead involved with fuse would be more perceivable.

 Would it be possible to increase the block size for dd and test?



 I'm in the process of setting up a share from my desktop and
 I'll see if

 I can bench between the two systems. Not sure if my ssd will
 impact the

 tests, I've heard there isn't an advantage using ssd storage for
 glusterfs.


 Do you have any pointers to this source of information? Typically
 glusterfs performance for virtualization work loads is bound by the
 slowest element in the entire stack. Usually storage/disks happen to
 be the bottleneck and ssd storage does benefit glusterfs.

 -Vijay


 I had a couple technical calls with RH (re: RHSS), and when I asked if
 SSD's could add any benefit I was told no. The context may have been in
 a product comparison to other storage vendors, where they use SSD's for
 read/write caching, versus having an all SSD storage domain (which I'm
 not proposing

Re: [Users] Ovirt Gluster problems

2014-01-28 Thread Steve Dainard
Not sure if this is exactly your issue, but this post here:
http://comments.gmane.org/gmane.comp.emulators.ovirt.user/12200 might lead
you in the right direction.

one note - if you back it up while its attached to an engine, you will
need to edit its meta data file to remove the association to allow the
other engine to connect it to the new pool for restore.


*Steve Dainard *



On Tue, Jan 28, 2014 at 12:41 PM, Juan Pablo Lorier jplor...@gmail.comwrote:

 Hi,

 I had some issues with a gluster cluster and after some time trying to
 get the storage domain up or delete it (I opened a BZ about a deadlock
 in the process of removing the domain) I gave up and destroyed the DC.
 The thing is that I want to add the hosts that where part of the DC and
 now I get that I can't as they have the volume. I try to stop the volume
 but I can't as no host is running in the deleted cluster and for some
 reason, ovirt needs that.
 I can't delete the hosts either as they have the volume... so  I'm back
 in another chicken and egg problem.
 Any hints??

 PD: I can't nuke the hole ovirt plataform as I have another DC in
 production otherwise I would :-)

 Regards,


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-25 Thread Steve Dainard
 lock
when I attempt a gluster storage domain.




I've setup a NFS storage domain on my desktops SSD. I've re-installed win
2008 r2 and initially it was running smoother.

Disk performance peaks at 100MB/s.

If I copy a 250MB file from a share into the Windows VM, it writes out
quickly, less than 5 seconds.

If I copy 20 files, ranging in file sizes from 4k to 200MB, totaling in
650MB from the share - windows becomes unresponsive, in top the desktop's
nfs daemon is barely being touched at all, and then eventually is not hit.
I can still interact with the VM's windows through the spice console.
Eventually the file transfer will start and rocket through the transfer.

I've opened a 271MB zip file with 4454 files and started the extract
process but the progress windows will sit on 'calculating...' after a
significant period of time the decompression starts and runs at
200KB/second. Windows is guesstimating 1HR completion time. Eventually
even this freezes up, and my spice console mouse won't grab. I can still
see the resource monitor in the Windows VM doing its thing but have to
poweroff the VM as its no longer usable.

The windows update process is the same. It seems like when the guest needs
quick large writes its fine, but lots of io causes serious hanging,
unresponsiveness, spice mouse cursor freeze, and eventually poweroff/reboot
is the only way to get it back.

Also, during window 2008 r2 install the 'expanding windows files' task is
quite slow, roughly 1% progress every 20 seconds (~30 mins to complete).
The GLUSTER host shows these stats pretty consistently:

PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND

 8139 root  20   0 1380m  28m 2476 R 83.1  0.4   8:35.78 glusterfsd

 8295 root  20   0  550m 186m 2980 S  4.3  2.4   1:52.56 glusterfs

bwm-ng v0.6 (probing every 2.000s), press 'h' for help
input: /proc/net/dev type: rate
  \ iface   Rx   Tx
 Total

==
   lo:3719.31 KB/s 3719.31 KB/s 7438.62
KB/s
 eth0:3405.12 KB/s 3903.28 KB/s 7308.40
KB/s


I've copied the same zip file to an nfs mount point on the OVIRT host
(gluster backend) and get about 25 - 600 KB/s during unzip. The same test
on NFS mount point (desktop SSD ext4 backend) averaged a network transfer
speed of 5MB/s and completed in about 40 seconds.

I have a RHEL 6.5 guest running on the NFS/gluster backend storage domain,
and just did the same test. Extracting the file took 22.3 seconds (faster
than the fuse mount point on the host !?!?).

GLUSTER host top reported this while the RHEL guest was decompressing the
zip file:
  PID USER  PR  NI  VIRT  RES  SHR S %CPU %MEMTIME+  COMMAND

 2141 root  20   0  555m 187m 2844 S  4.0  2.4  18:17.00 glusterfs

 2122 root  20   0 1380m  31m 2396 S  2.3  0.4  83:19.40 glusterfsd





*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-24 Thread Steve Dainard
Not sure what a good method to bench this would be, but:

An NFS mount point on virt host:
[root@ovirt001 iso-store]# dd if=/dev/zero of=test1 bs=4k count=10
10+0 records in
10+0 records out
40960 bytes (410 MB) copied, 3.95399 s, 104 MB/s

Raw brick performance on gluster server (yes, I know I shouldn't write
directly to the brick):
[root@gluster1 iso-store]# dd if=/dev/zero of=test bs=4k count=10
10+0 records in
10+0 records out
40960 bytes (410 MB) copied, 3.06743 s, 134 MB/s

Gluster mount point on gluster server:
[root@gluster1 iso-store]# dd if=/dev/zero of=test bs=4k count=10
10+0 records in
10+0 records out
40960 bytes (410 MB) copied, 19.5766 s, 20.9 MB/s

The storage servers are a bit older, but are both dual socket quad core
opterons with 4x 7200rpm drives.

I'm in the process of setting up a share from my desktop and I'll see if I
can bench between the two systems. Not sure if my ssd will impact the
tests, I've heard there isn't an advantage using ssd storage for glusterfs.

Does anyone have a hardware reference design for glusterfs as a backend for
virt? Or is there a benchmark utility?

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Thu, Jan 23, 2014 at 7:18 PM, Andrew Cathrow acath...@redhat.com wrote:

 Are we sure that the issue is the guest I/O - what's the raw performance
 on the host accessing the gluster storage?

 --

 *From: *Steve Dainard sdain...@miovision.com
 *To: *Itamar Heim ih...@redhat.com
 *Cc: *Ronen Hod r...@redhat.com, users users@ovirt.org, Sanjay
 Rao s...@redhat.com
 *Sent: *Thursday, January 23, 2014 4:56:58 PM
 *Subject: *Re: [Users] Extremely poor disk access speeds in Windows guest


 I have two options, virtio and virtio-scsi.

 I was using virtio, and have also attempted virtio-scsi on another Windows
 guest with the same results.

 Using the newest drivers, virtio-win-0.1-74.iso.

 *Steve Dainard *
 IT Infrastructure Manager
 Miovision http://miovision.com/ | *Rethink Traffic*
 519-513-2407 ex.250
 877-646-8476 (toll-free)

 *Blog http://miovision.com/blog  |  **LinkedIn
 https://www.linkedin.com/company/miovision-technologies  |  Twitter
 https://twitter.com/miovision  |  Facebook
 https://www.facebook.com/miovision*
 --
  Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener,
 ON, Canada | N2C 1L3
 This e-mail may contain information that is privileged or confidential. If
 you are not the intended recipient, please delete the e-mail and any
 attachments and notify us immediately.


 On Thu, Jan 23, 2014 at 4:24 PM, Itamar Heim ih...@redhat.com wrote:

 On 01/23/2014 07:46 PM, Steve Dainard wrote:

 Backing Storage: Gluster Replica
 Storage Domain: NFS
 Ovirt Hosts: CentOS 6.5
 Ovirt version: 3.3.2
 Network: GigE
 # of VM's: 3 - two Linux guests are idle, one Windows guest is
 installing updates.

 I've installed a Windows 2008 R2 guest with virtio disk, and all the
 drivers from the latest virtio iso. I've also installed the spice agent
 drivers.

 Guest disk access is horribly slow, Resource monitor during Windows
 updates shows Disk peaking at 1MB/sec (scale never increases) and Disk
 Queue Length Peaking at 5 and looks to be sitting at that level 99% of
 the time. 113 updates in Windows has been running solidly for about 2.5
 hours and is at 89/113 updates complete.


 virtio-block or virtio-scsi?
 which windows guest driver version for that?


 I can't say my Linux guests are blisteringly fast, but updating a guest
 from RHEL 6.3 fresh install to 6.5 took about 25 minutes.

 If anyone has any ideas, please let me know - I haven't found any tuning
 docs for Windows guests that could explain this issue.

 Thanks,


 *Steve Dainard *



 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users




 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Extremely poor disk access speeds in Windows guest

2014-01-23 Thread Steve Dainard
Backing Storage: Gluster Replica
Storage Domain: NFS
Ovirt Hosts: CentOS 6.5
Ovirt version: 3.3.2
Network: GigE
# of VM's: 3 - two Linux guests are idle, one Windows guest is installing
updates.

I've installed a Windows 2008 R2 guest with virtio disk, and all the
drivers from the latest virtio iso. I've also installed the spice agent
drivers.

Guest disk access is horribly slow, Resource monitor during Windows updates
shows Disk peaking at 1MB/sec (scale never increases) and Disk Queue Length
Peaking at 5 and looks to be sitting at that level 99% of the time. 113
updates in Windows has been running solidly for about 2.5 hours and is at
89/113 updates complete.

I can't say my Linux guests are blisteringly fast, but updating a guest
from RHEL 6.3 fresh install to 6.5 took about 25 minutes.

If anyone has any ideas, please let me know - I haven't found any tuning
docs for Windows guests that could explain this issue.

Thanks,


*Steve Dainard *
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Extremely poor disk access speeds in Windows guest

2014-01-23 Thread Steve Dainard
I have two options, virtio and virtio-scsi.

I was using virtio, and have also attempted virtio-scsi on another Windows
guest with the same results.

Using the newest drivers, virtio-win-0.1-74.iso.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |  **LinkedIn
https://www.linkedin.com/company/miovision-technologies  |  Twitter
https://twitter.com/miovision  |  Facebook
https://www.facebook.com/miovision*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Thu, Jan 23, 2014 at 4:24 PM, Itamar Heim ih...@redhat.com wrote:

 On 01/23/2014 07:46 PM, Steve Dainard wrote:

 Backing Storage: Gluster Replica
 Storage Domain: NFS
 Ovirt Hosts: CentOS 6.5
 Ovirt version: 3.3.2
 Network: GigE
 # of VM's: 3 - two Linux guests are idle, one Windows guest is
 installing updates.

 I've installed a Windows 2008 R2 guest with virtio disk, and all the
 drivers from the latest virtio iso. I've also installed the spice agent
 drivers.

 Guest disk access is horribly slow, Resource monitor during Windows
 updates shows Disk peaking at 1MB/sec (scale never increases) and Disk
 Queue Length Peaking at 5 and looks to be sitting at that level 99% of
 the time. 113 updates in Windows has been running solidly for about 2.5
 hours and is at 89/113 updates complete.


 virtio-block or virtio-scsi?
 which windows guest driver version for that?


 I can't say my Linux guests are blisteringly fast, but updating a guest
 from RHEL 6.3 fresh install to 6.5 took about 25 minutes.

 If anyone has any ideas, please let me know - I haven't found any tuning
 docs for Windows guests that could explain this issue.

 Thanks,


 *Steve Dainard *



 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Bottleneck writing to a VM w/ mounted GlusterFS

2013-09-28 Thread Steve Dainard
Are you duplicating the traffic over the same physical network by relaying
through the VM, rather than writing directly to network storage, thereby
halving the write performance?

Assuming you're on a GigE network, are all the network devices running in
full duplex?

Just some guesses based on the fact that the throughput is almost exactly
half.

*Steve Dainard *
IT Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |
**LinkedInhttps://www.linkedin.com/company/miovision-technologies  |
 Twitter https://twitter.com/miovision  |
Facebookhttps://www.facebook.com/miovision
*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Thu, Sep 26, 2013 at 5:23 AM, Stefano Stagnaro
stefa...@prisma-eng.comwrote:

 Hello,

 I'm testing oVirt 3.3 with GlusterFS libgfapi back-end. I'm using a node
 for engine and one for VDSM. From the VMs I'm mounting a second GlusterFS
 volume on a third storage server.

 I'm experiencing very bad transfer rates (38MB/s) writing from a client to
 a VM on the mounted GlusterFS. On the other hand, from the VM itself I can
 move a big file from the root vda (libgfapi) to the mounted GlusterFS at
 70MB/s.

 I can't really figure out where the bottleneck could be. I'm using only
 the default ovirtmgmt network.

 Thank you for your help, any hint will be appreciated.

 Regards,
 --
 Stefano Stagnaro
 IT Manager

 Prisma Engineering S.r.l.
 Via Petrocchi, 4
 20127 Milano – Italy

 Tel. 02 26113507 int 339
 e-mail: stefa...@prisma-eng.com
 skype: stefano.stagnaro




 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Ovirt 3.3 Fedora 19 add gluster storage permissions error

2013-09-19 Thread Steve Dainard
Hello,

New Ovirt 3.3 install on Fedora 19.

When I try to add a gluster storage domain I get the following:

*UI error:*
*Error while executing action Add Storage Connection: Permission settings
on the specified path do not allow access to the storage.*
*Verify permission settings on the specified storage path.*

*VDSM logs contain:*
Thread-393::DEBUG::2013-09-19
11:59:42,399::BindingXMLRPC::177::vds::(wrapper) client [10.0.0.34]
Thread-393::DEBUG::2013-09-19
11:59:42,399::task::579::TaskManager.Task::(_updateState)
Task=`12c38fec-0072-4974-a8e3-9125b3908246`::moving from state init -
state preparing
Thread-393::INFO::2013-09-19
11:59:42,400::logUtils::44::dispatcher::(wrapper) Run and protect:
connectStorageServer(domType=7,
spUUID='----', conList=[{'port': '',
'connection': '192.168.1.1:/rep2-virt', 'iqn': '', 'portal': '', 'user':
'', 'vfs_type': 'glusterfs', 'password': '**', 'id':
'----'}], options=None)
Thread-393::DEBUG::2013-09-19
11:59:42,405::mount::226::Storage.Misc.excCmd::(_runcmd) '/usr/bin/sudo -n
/usr/bin/mount -t glusterfs 192.168.1.1:/rep2-virt
/rhev/data-center/mnt/glusterSD/192.168.1.1:_rep2-virt' (cwd None)
Thread-393::DEBUG::2013-09-19
11:59:42,490::mount::226::Storage.Misc.excCmd::(_runcmd) '/usr/bin/sudo -n
/usr/bin/umount -f -l /rhev/data-center/mnt/glusterSD/192.168.1.1:_rep2-virt'
(cwd None)
Thread-393::ERROR::2013-09-19
11:59:42,505::hsm::2382::Storage.HSM::(connectStorageServer) Could not
connect to storageServer
Traceback (most recent call last):
  File /usr/share/vdsm/storage/hsm.py, line 2379, in connectStorageServer
conObj.connect()
  File /usr/share/vdsm/storage/storageServer.py, line 227, in connect
raise e
StorageServerAccessPermissionError: Permission settings on the specified
path do not allow access to the storage. Verify permission settings on the
specified storage path.: 'path =
/rhev/data-center/mnt/glusterSD/192.168.1.1:_rep2-virt'
Thread-393::DEBUG::2013-09-19
11:59:42,506::hsm::2396::Storage.HSM::(connectStorageServer) knownSDs: {}
Thread-393::INFO::2013-09-19
11:59:42,506::logUtils::47::dispatcher::(wrapper) Run and protect:
connectStorageServer, Return response: {'statuslist': [{'status': 469,
'id': '----'}]}
Thread-393::DEBUG::2013-09-19
11:59:42,506::task::1168::TaskManager.Task::(prepare)
Task=`12c38fec-0072-4974-a8e3-9125b3908246`::finished: {'statuslist':
[{'status': 469, 'id': '----'}]}
Thread-393::DEBUG::2013-09-19
11:59:42,506::task::579::TaskManager.Task::(_updateState)
Task=`12c38fec-0072-4974-a8e3-9125b3908246`::moving from state preparing -
state finished
Thread-393::DEBUG::2013-09-19
11:59:42,506::resourceManager::939::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-393::DEBUG::2013-09-19
11:59:42,507::resourceManager::976::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-393::DEBUG::2013-09-19
11:59:42,507::task::974::TaskManager.Task::(_decref)
Task=`12c38fec-0072-4974-a8e3-9125b3908246`::ref 0 aborting False

*Other info:*
- I have two nodes, ovirt001, ovirt002 they are both Fedora 19.
- The gluster bricks are replicated and located on the nodes.
(ovirt001:rep2-virt, ovirt002:rep2-virt)
- Local directory for the mount, I changed permissions on glusterSD to 777,
it was 755, and there is nothing in that directory:
[root@ovirt001 mnt]# pwd
/rhev/data-center/mnt
[root@ovirt001 mnt]# ll
total 4
drwxrwxrwx. 2 vdsm kvm 4096 Sep 19 12:18 glusterSD

I find it odd that the UUID's listed in the vdsm logs are zero's..

Appreciate any help,


*Steve
*
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Announcement: oVirt 3.3.0 release!

2013-09-18 Thread Steve Dainard
Awesome.

Dave can you have someone edit the installation page and specifically
mention that EL hypervisors don't have native Gluster (fuse only) support -
It would be very frustrating to think you can have all the features, and OS
stability as well, only to find out EL doesn't support a key feature.

I've seen lots of requests on the board for a back-port so this might
preempt more questions / problems.

*Steve *
*
*
On Tue, Sep 17, 2013 at 8:31 AM, Dave Neary dne...@redhat.com wrote:


 The oVirt development team is very happy to announce the general
 availability of oVirt 3.3.0 as of September 16th 2013. This release
 solidifies oVirt as a leading KVM management application, and open
 source alternative to VMware vSphere.

 oVirt is available now for Fedora 19 and Red Hat Enterprise Linux 6.4
 (or similar).

 Get started with oVirt now! http://www.ovirt.org/Download

 Chief among the many new features in the release are:

 * Tight integration with Gluster - take advantage of native GlusterFS
 support, or use oVirt to manage your Gluster bricks
 * Integration with OpenStack - share images stored in Glance, and take
 advantage of Neutron integration for network topology definition
 * Extensibility and control - with improvements in VM scheduling, an
 increased array of hooks and APIs, oVirt gives you absolute control over
 your virtual datacenter

 Read more about the oVirt 3.3 release on the Red Hat community blog:

 http://community.redhat.com/ovirt-3-3-spices-up-the-software-defined-datacenter-with-openstack-and-gluster-integration/

 The full release announcement:
 http://www.ovirt.org/OVirt_3.3_release_announcement

 oVirt 3.3 release notes: http://www.ovirt.org/OVirt_3.3_release_notes


 Regards,
 Dave Neary.

 --
 Dave Neary - Community Action and Impact
 Open Source and Standards, Red Hat - http://community.redhat.com
 Ph: +33 9 50 71 55 62 / Cell: +33 6 77 01 92 13
 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] so, what do you want next in oVirt?

2013-08-29 Thread Steve Dainard

 importing any data storage domain i assume?

 Yes.


  Possibility of direct use of HW by VM.


 such as?

 Telephone modem. PCI-express cards. Graphic cards


+ USB devices, we have hardware license keys for some software. In kvm/qemu
I can expose the license key directly to a VM.





  and the absolutely fantastic feature would be to create clusters from
 Intel and AMD processors together!


 well, you can do that today if you want to via a config change. the only
 thing is live migration won't work (you should probably use -cpu host to
 get best performance, since live migration won't be used anyway)
 (well, in theory we could live migrate only between hosts of same cpu
 vendor, but not sure interesting enough use case to make cluster and
 scheduling more complex). though you can do that part on your own with the
 new pluggable scheduler, or use -cpu host to get max performance if you
 don't care about live migration

  From my point of view it is better to have slower cpu performance and
 possibility to use all of our servers in cluster. I would like to have live
 migration available from intel to amd. The problem is only in cpu
 instruction sets? If so, I can use only common sets.


 Another feature which I forgot is network between VMs and mirroring
 traffic. Both configurable from WUI.


 Thank you ;-)


 __**_
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/**mailman/listinfo/usershttp://lists.ovirt.org/mailman/listinfo/users



 __**_
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/**mailman/listinfo/usershttp://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Ovirt 3.3 nightly, Gluster 3.4 stable, cannot launch VM with gluster storage domain backed disk

2013-07-17 Thread Steve Dainard
, FullListVdsCommand(HostName =
ovirt001, HostId = d07967ab-3764-47ff-8755-bc539a7feb3b,
vds=Host[ovirt001], vmIds=[8e2c9057-deee-48a6-8314-a34530fc53cb]), log id:
119758a
2013-07-17 11:12:39,453 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVdsCommand]
(DefaultQuartzScheduler_Worker-2) FINISH, FullListVdsCommand, return:
[Ljava.util.HashMap;@2e73b796, log id: 119758a
2013-07-17 11:12:39,478 ERROR
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-2) Rerun vm
8e2c9057-deee-48a6-8314-a34530fc53cb. Called from vds ovirt001
2013-07-17 11:12:39,574 INFO  [org.ovirt.engine.core.bll.RunVmCommand]
(pool-6-thread-49) Lock Acquired to object EngineLock [exclusiveLocks= key:
8e2c9057-deee-48a6-8314-a34530fc53cb value: VM
, sharedLocks= ]
2013-07-17 11:12:39,603 INFO
 [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(pool-6-thread-49) START, IsVmDuringInitiatingVDSCommand( vmId =
8e2c9057-deee-48a6-8314-a34530fc53cb), log id: 497e83ec
2013-07-17 11:12:39,606 INFO
 [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(pool-6-thread-49) FINISH, IsVmDuringInitiatingVDSCommand, return: false,
log id: 497e83ec
2013-07-17 11:12:39,661 INFO
 [org.ovirt.engine.core.bll.scheduling.VdsSelector] (pool-6-thread-49)  VDS
ovirt001 d07967ab-3764-47ff-8755-bc539a7feb3b have failed running this VM
in the current selection cycle
2013-07-17 11:12:39,663 WARN  [org.ovirt.engine.core.bll.RunVmCommand]
(pool-6-thread-49) CanDoAction of action RunVm failed.
Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,ACTION_TYPE_FAILED_VDS_VM_CLUSTER
2013-07-17 11:12:39,664 INFO  [org.ovirt.engine.core.bll.RunVmCommand]
(pool-6-thread-49) Lock freed to object EngineLock [exclusiveLocks= key:
8e2c9057-deee-48a6-8314-a34530fc53cb value: VM
, sharedLocks= ]
2013-07-17 11:13:49,097 INFO  [org.ovirt.engine.core.bll.AsyncTaskManager]
(DefaultQuartzScheduler_Worker-42) Setting new tasks map. The map contains
now 0 tasks
2013-07-17 11:13:49,099 INFO  [org.ovirt.engine.core.bll.AsyncTaskManager]
(DefaultQuartzScheduler_Worker-42) Cleared all tasks of pool
5849b030-626e-47cb-ad90-3ce782d831b3.



*Steve Dainard *
Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |
**LinkedInhttps://www.linkedin.com/company/miovision-technologies  |
 Twitter https://twitter.com/miovision  |
Facebookhttps://www.facebook.com/miovision
*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Ovirt 3.3 nightly, Gluster 3.4 stable, cannot launch VM with gluster storage domain backed disk

2013-07-17 Thread Steve Dainard
-
8314-a34530fc53cb value: VM
, sharedLocks= ]
2013-07-17 12:39:51,459 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
(DefaultQuartzScheduler_Worker-85) START, DestroyVDSCommand(HostName =
ovirt001,
HostId = d07967ab-3764-47ff-8755-bc539a7feb3b,
vmId=8e2c9057-deee-48a6-8314-a34530fc53cb, force=false, secondsToWait=0,
gracefully=false), log id: 60626686
2013-07-17 12:39:51,548 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
(DefaultQuartzScheduler_Worker-85) FINISH, DestroyVDSCommand, log id:
60626686
2013-07-17 12:39:51,635 INFO
 [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-85) Running on vds during rerun failed vm:
null
2013-07-17 12:39:51,641 INFO
 [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-85) vm VM1 running in db and not running in
vds - add to
rerun treatment. vds ovirt001
2013-07-17 12:39:51,660 ERROR
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
(DefaultQuartzScheduler_Worker-85) Rerun vm
8e2c9057-deee-48a6-8314-a34530fc53cb. Called
 from vds ovirt001
2013-07-17 12:39:51,729 INFO  [org.ovirt.engine.core.bll.RunVmCommand]
(pool-6-thread-50) Lock Acquired to object EngineLock [exclusiveLocks= key:
8e2c9057-deee-48a6-8314-a3
4530fc53cb value: VM
, sharedLocks= ]
2013-07-17 12:39:51,753 INFO
 [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(pool-6-thread-50) START, IsVmDuringInitiatingVDSCommand( vmId =
8e2c9057-deee
-48a6-8314-a34530fc53cb), log id: 7647c7d4
2013-07-17 12:39:51,753 INFO
 [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand]
(pool-6-thread-50) FINISH, IsVmDuringInitiatingVDSCommand, return: false,
log
id: 7647c7d4
2013-07-17 12:39:51,794 INFO
 [org.ovirt.engine.core.bll.scheduling.VdsSelector] (pool-6-thread-50)  VDS
ovirt001 d07967ab-3764-47ff-8755-bc539a7feb3b have failed running th
is VM in the current selection cycle
2013-07-17 12:39:51,794 WARN  [org.ovirt.engine.core.bll.RunVmCommand]
(pool-6-thread-50) CanDoAction of action RunVm failed.
Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,ACTION_TYPE_FAILED_VDS_VM_CLUSTER
2013-07-17 12:39:51,795 INFO  [org.ovirt.engine.core.bll.RunVmCommand]
(pool-6-thread-50) Lock freed to object EngineLock [exclusiveLocks= key:
8e2c9057-deee-48a6-8314-a34530fc53cb value: VM
, sharedLocks= ]



*Steve Dainard *
Infrastructure Manager
Miovision http://miovision.com/ | *Rethink Traffic*
519-513-2407 ex.250
877-646-8476 (toll-free)

*Blog http://miovision.com/blog  |
**LinkedInhttps://www.linkedin.com/company/miovision-technologies  |
 Twitter https://twitter.com/miovision  |
Facebookhttps://www.facebook.com/miovision
*
--
 Miovision Technologies Inc. | 148 Manitou Drive, Suite 101, Kitchener, ON,
Canada | N2C 1L3
This e-mail may contain information that is privileged or confidential. If
you are not the intended recipient, please delete the e-mail and any
attachments and notify us immediately.


On Wed, Jul 17, 2013 at 12:21 PM, Vijay Bellur vbel...@redhat.com wrote:

 On 07/17/2013 09:04 PM, Steve Dainard wrote:




 *Web-UI displays:*

 VM VM1 is down. Exit message: internal error process exited while
 connecting to monitor: qemu-system-x86_64: -drive
 file=gluster://ovirt001/vol1/**a87a7ef6-2c74-4d8e-a6e0-**
 a392d0f791cf/images/238cc6cf-**070c-4483-b686-c0de7ddf0dfa/**
 ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a,if=none,id=drive-**
 virtio-disk0,format=raw,**serial=238cc6cf-070c-4483-**
 b686-c0de7ddf0dfa,cache=none,**werror=stop,rerror=stop,aio=**threads:
 could not open disk image
 gluster://ovirt001/vol1/**a87a7ef6-2c74-4d8e-a6e0-**
 a392d0f791cf/images/238cc6cf-**070c-4483-b686-c0de7ddf0dfa/**
 ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a:
 No such file or directory .
 VM VM1 was started by admin@internal (Host: ovirt001).
 The disk VM1_Disk1 was successfully added to VM VM1.

 *I can see the image on the gluster machine, and it looks to have the
 correct permissions:*

 [root@ovirt001 238cc6cf-070c-4483-b686-**c0de7ddf0dfa]# pwd
 /mnt/storage1/vol1/a87a7ef6-**2c74-4d8e-a6e0-a392d0f791cf/**
 images/238cc6cf-070c-4483-**b686-c0de7ddf0dfa
 [root@ovirt001 238cc6cf-070c-4483-b686-**c0de7ddf0dfa]# ll
 total 1028
 -rw-rw. 2 vdsm kvm 32212254720 Jul 17 11:11
 ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a
 -rw-rw. 2 vdsm kvm 1048576 Jul 17 11:11
 ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a.lease
 -rw-r--r--. 2 vdsm kvm 268 Jul 17 11:11
 ff2bca2d-4ed1-46c6-93c8-**22a39bb1626a.meta


 Can you please try after doing these changes:

 1) gluster volume set volname server.allow-insecure on

 2) Edit /etc/glusterfs/glusterd.vol to contain this line:
 option rpc-auth-allow-insecure on

 Post 2), restarting glusterd would be necessary.

 Thanks,
 Vijay

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] attaching glusterfs storage domain -- method glusterServicesGet is not supported

2013-07-01 Thread Steve Dainard
Creating /var/lib/glusterd/groups/virt on each node and adding parameters
found here:
https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.0/html/Quick_Start_Guide/chap-Quick_Start_Guide-Virtual_Preparation.html
solved
this issue.




Steve Dainard
Infrastructure Manager
Miovision Technologies Inc.
Phone: 519-513-2407 x250


On Mon, Jul 1, 2013 at 4:54 PM, Steve Dainard sdain...@miovision.comwrote:

 I'm using Ovirt nightly on Fedora 18 to determine if ovirt + gluster is
 something that will work for my organization (or at least when RHEV is
 released with this functionality). I'm attempting to use the nodes for both
 virt and gluster storage.

 I've successfully created gluster bricks on two hosts, ovirt001 
 ovirt002, and started volume vol1 through engine web ui. I've created a
 gluster storage domain, but cannot attach to data center.

 *Engine UI error:*
 Failed to attach Storage Domains to Data Center Default. (User:
 admin@internal)

 *Engine /var/log/ovirt-engine/engine.log errors:*
 2013-07-01 16:29:02,650 ERROR
 [org.ovirt.engine.core.vdsbroker.gluster.GlusterServicesListVDSCommand]
 (pool-6-thread-49) Command GlusterServicesListVDS execution failed.
 Exception: VDSNetworkException: org.apache.xmlrpc.XmlRpcException: type
 'exceptions.Exception':method glusterServicesGet is not supported
 I've attached the much larger log file

 *Host
 /var/log/glusterfs/rhev-data-center-mnt-glusterSD-ovirt001\:vol1.log when
 attaching:*
 [2013-07-01 20:40:33.718871] I [afr-self-heal-data.c:655:afr_sh_data_fix]
 0-vol1-replicate-0: no active sinks for performing self-heal on file
 /b2076340-84de-45ff-9d4b-d0d48b935fca/dom_md/ids
 [2013-07-01 20:40:33.721059] W
 [client-rpc-fops.c:873:client3_3_writev_cbk] 0-vol1-client-0: remote
 operation failed: Invalid argument
 [2013-07-01 20:40:33.721104] W
 [client-rpc-fops.c:873:client3_3_writev_cbk] 0-vol1-client-1: remote
 operation failed: Invalid argument
 [2013-07-01 20:40:33.721130] W [fuse-bridge.c:2127:fuse_writev_cbk]
 0-glusterfs-fuse: 304: WRITE = -1 (Invalid argument)

 *Engine repos:*
 [root@ovirt-manager2 yum.repos.d]# ll
 total 16
 -rw-r--r--. 1 root root 1145 Dec 20  2012 fedora.repo
 -rw-r--r--. 1 root root 1105 Dec 20  2012 fedora-updates.repo
 -rw-r--r--. 1 root root 1163 Dec 20  2012 fedora-updates-testing.repo
 -rw-r--r--. 1 root root  144 Jun 21 18:34 ovirt-nightly.repo

 [root@ovirt-manager2 yum.repos.d]# cat *
 [fedora]
 name=Fedora $releasever - $basearch
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/releases/$releasever/Everything/$basearch/os/
 mirrorlist=
 https://mirrors.fedoraproject.org/metalink?repo=fedora-$releaseverarch=$basearch
 enabled=1
 #metadata_expire=7d
 gpgcheck=1
 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-$basearch

 [fedora-debuginfo]
 name=Fedora $releasever - $basearch - Debug
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/releases/$releasever/Everything/$basearch/debug/
 mirrorlist=
 https://mirrors.fedoraproject.org/metalink?repo=fedora-debug-$releaseverarch=$basearch
 enabled=0
 metadata_expire=7d
 gpgcheck=1
 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-$basearch

 [fedora-source]
 name=Fedora $releasever - Source
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/releases/$releasever/Everything/source/SRPMS/
 mirrorlist=
 https://mirrors.fedoraproject.org/metalink?repo=fedora-source-$releaseverarch=$basearch
 enabled=0
 metadata_expire=7d
 gpgcheck=1
 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-$basearch
 [updates]
 name=Fedora $releasever - $basearch - Updates
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/updates/$releasever/$basearch/
 mirrorlist=
 https://mirrors.fedoraproject.org/metalink?repo=updates-released-f$releaseverarch=$basearch
 enabled=1
 gpgcheck=1
 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-$basearch

 [updates-debuginfo]
 name=Fedora $releasever - $basearch - Updates - Debug
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/updates/$releasever/$basearch/debug/
 mirrorlist=
 https://mirrors.fedoraproject.org/metalink?repo=updates-released-debug-f$releaseverarch=$basearch
 enabled=0
 gpgcheck=1
 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-$basearch

 [updates-source]
 name=Fedora $releasever - Updates Source
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/updates/$releasever/SRPMS/
 mirrorlist=
 https://mirrors.fedoraproject.org/metalink?repo=updates-released-source-f$releaseverarch=$basearch
 enabled=0
 gpgcheck=1
 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-$basearch
 [updates-testing]
 name=Fedora $releasever - $basearch - Test Updates
 failovermethod=priority
 #baseurl=
 http://download.fedoraproject.org/pub/fedora/linux/updates/testing/$releasever/$basearch/
 mirrorlist=
 https://mirrors.fedoraproject.org