[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-12-04 Thread Albl, Oliver
Hi Amit,



  is it inactive, but not in maintenance mode.



Thank you,

Oliver



Von: Amit Bawer 
Gesendet: Mittwoch, 4. Dezember 2019 16:36
An: Albl, Oliver 
Cc: users@ovirt.org; Nir Soffer 
Betreff: Re: [ovirt-users] Re: Cannot activate/deactivate storage domain



in check we had here, we got similar warnings for using the ignore OVF updates 
checks, but the SD was set inactive at end of process.

what is the SD status in your case after this try?





On Wed, Dec 4, 2019 at 4:49 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:

   Yes.

   Am 04.12.2019 um 15:47 schrieb Amit Bawer 
mailto:aba...@redhat.com><mailto:aba...@redhat.com<mailto:aba...@redhat.com>>>:



   On Wed, Dec 4, 2019 at 4:42 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com><mailto:oliver.a...@fabasoft.com<mailto:oliver.a...@fabasoft.com>>>
 wrote:
   Hi Amit,

 unfortunately no success.

   Dec 4, 2019, 3:41:36 PM
   Storage Domain HOST_LUN_219 (Data Center xxx) was deactivated by system 
because it's not visible by any of the hosts.

   Dec 4, 2019, 3:35:09 PM
   Failed to update VMs/Templates OVF data for Storage Domain HOST_LUN_219 in 
Data Center Production.

   Dec 4, 2019, 3:35:09 PM
   Failed to update OVF disks 77c64b39-fe50-4d05-b77f-8131ad1f95f9, OVF data 
isn't updated on those OVF stores (Data Center Production, Storage Domain 
HOST_LUN_219).

   Have you selected the checkbox for "Ignore OVF update failure" before 
putting into maintenance?


   All the best,
   Oliver

   Von: Amit Bawer 
mailto:aba...@redhat.com><mailto:aba...@redhat.com<mailto:aba...@redhat.com>>>
   Gesendet: Mittwoch, 4. Dezember 2019 15:20
   An: Albl, Oliver 
mailto:oliver.a...@fabasoft.com><mailto:oliver.a...@fabasoft.com<mailto:oliver.a...@fabasoft.com>>>
   Cc: 
users@ovirt.org<mailto:users@ovirt.org><mailto:users@ovirt.org<mailto:users@ovirt.org>>;
 Nir Soffer 
mailto:nsof...@redhat.com><mailto:nsof...@redhat.com<mailto:nsof...@redhat.com>>>
   Betreff: Re: [ovirt-users] Re: Cannot activate/deactivate storage domain

   Hi Oliver,

   For deactivating the unresponsive storage domains, you can use the Compute 
-> Data Centers -> Maintenance option with "Ignore OVF update failure" checked.
   This will force deactivation of the SD.

   Will provide further details about the issue in the ticket.


   On Tue, Dec 3, 2019 at 12:02 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com><mailto:oliver.a...@fabasoft.com<mailto:oliver.a...@fabasoft.com>>>
 wrote:
   Hi,

 does anybody have an advice how to activate or safely remove that storage 
domain?

   Thank you!
   Oliver
   -Ursprüngliche Nachricht-
   Von: Oliver Albl 
mailto:oliver.a...@fabasoft.com><mailto:oliver.a...@fabasoft.com<mailto:oliver.a...@fabasoft.com>>>
   Gesendet: Dienstag, 5. November 2019 11:20
   An: 
users@ovirt.org<mailto:users@ovirt.org><mailto:users@ovirt.org<mailto:users@ovirt.org>>
   Betreff: [ovirt-users] Re: Cannot activate/deactivate storage domain

   > On Mon, Nov 4, 2019 at 9:18 PM Albl, Oliver 
http://fabasoft.com><http://fabasoft.com> wrote:
   >
   > What was the last change in the system? upgrade? network change? storage 
change?
   >

   Last change was four weeks ago ovirt upgrade from 4.3.3 to 4.3.6.7 
(including CentOS hosts to 7.7 1908)

   >
   > This is expected if some domain is not accessible on all hosts.
   >
   >
   > This means sanlock timed out renewing the lockspace
   >
   >
   > If a host cannot access all storage domain in the DC, the system set
   > it to non-operational, and will probably try to reconnect it later.
   >
   >
   > This means reading 4k from start of the metadata lv took 9.6 seconds.
   > Something in
   > the way to storage is bad (kernel, network, storage).
   >
   >
   > We 20 seconds (4 retires, 5 seconds per retry) gracetime in multipath
   > when there are no active paths, before I/O fails, pausing the VM. We
   > also resume paused VMs when storage monitoring works again, so maybe
   > the VM were paused and resumed.
   >
   > However for storage monitoring we have strict 10 seconds timeout. If
   > reading from the metadata lv times out or fail and does not operated
   > normally after
   > 5 minutes, the
   > domain will become inactive.
   >
   >
   > This can explain the read timeouts.
   >
   >
   > This looks the right way to troubleshoot this.
   >
   >
   > We need vdsm logs to understand this failure.
   >
   >
   > This does not mean OVF is corrupted, only that we could not store new
   > data. The older data on the other OVFSTORE disk is probably fine.
   > Hopefuly the system will not try to write to the other OVFSTORE disk
   > overwriting the last g

[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-12-04 Thread Albl, Oliver


smime.p7m
Description: S/MIME encrypted message
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4Q4ATO7KOMUF4PKRVPHA3BRMTLWSUWEU/


[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-12-04 Thread Albl, Oliver
Yes.

Am 04.12.2019 um 15:47 schrieb Amit Bawer 
mailto:aba...@redhat.com>>:



On Wed, Dec 4, 2019 at 4:42 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:
Hi Amit,

  unfortunately no success.

Dec 4, 2019, 3:41:36 PM
Storage Domain HOST_LUN_219 (Data Center xxx) was deactivated by system because 
it's not visible by any of the hosts.

Dec 4, 2019, 3:35:09 PM
Failed to update VMs/Templates OVF data for Storage Domain HOST_LUN_219 in Data 
Center Production.

Dec 4, 2019, 3:35:09 PM
Failed to update OVF disks 77c64b39-fe50-4d05-b77f-8131ad1f95f9, OVF data isn't 
updated on those OVF stores (Data Center Production, Storage Domain 
HOST_LUN_219).

Have you selected the checkbox for "Ignore OVF update failure" before putting 
into maintenance?


All the best,
Oliver

Von: Amit Bawer mailto:aba...@redhat.com>>
Gesendet: Mittwoch, 4. Dezember 2019 15:20
An: Albl, Oliver mailto:oliver.a...@fabasoft.com>>
Cc: users@ovirt.org<mailto:users@ovirt.org>; Nir Soffer 
mailto:nsof...@redhat.com>>
Betreff: Re: [ovirt-users] Re: Cannot activate/deactivate storage domain

Hi Oliver,

For deactivating the unresponsive storage domains, you can use the Compute -> 
Data Centers -> Maintenance option with "Ignore OVF update failure" checked.
This will force deactivation of the SD.

Will provide further details about the issue in the ticket.


On Tue, Dec 3, 2019 at 12:02 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:
Hi,

  does anybody have an advice how to activate or safely remove that storage 
domain?

Thank you!
Oliver
-Ursprüngliche Nachricht-
Von: Oliver Albl mailto:oliver.a...@fabasoft.com>>
Gesendet: Dienstag, 5. November 2019 11:20
An: users@ovirt.org<mailto:users@ovirt.org>
Betreff: [ovirt-users] Re: Cannot activate/deactivate storage domain

> On Mon, Nov 4, 2019 at 9:18 PM Albl, Oliver 
> http://fabasoft.com> wrote:
>
> What was the last change in the system? upgrade? network change? storage 
> change?
>

Last change was four weeks ago ovirt upgrade from 4.3.3 to 4.3.6.7 (including 
CentOS hosts to 7.7 1908)

>
> This is expected if some domain is not accessible on all hosts.
>
>
> This means sanlock timed out renewing the lockspace
>
>
> If a host cannot access all storage domain in the DC, the system set
> it to non-operational, and will probably try to reconnect it later.
>
>
> This means reading 4k from start of the metadata lv took 9.6 seconds.
> Something in
> the way to storage is bad (kernel, network, storage).
>
>
> We 20 seconds (4 retires, 5 seconds per retry) gracetime in multipath
> when there are no active paths, before I/O fails, pausing the VM. We
> also resume paused VMs when storage monitoring works again, so maybe
> the VM were paused and resumed.
>
> However for storage monitoring we have strict 10 seconds timeout. If
> reading from the metadata lv times out or fail and does not operated
> normally after
> 5 minutes, the
> domain will become inactive.
>
>
> This can explain the read timeouts.
>
>
> This looks the right way to troubleshoot this.
>
>
> We need vdsm logs to understand this failure.
>
>
> This does not mean OVF is corrupted, only that we could not store new
> data. The older data on the other OVFSTORE disk is probably fine.
> Hopefuly the system will not try to write to the other OVFSTORE disk
> overwriting the last good version.
>
>
> This is normal, the first 2048 bytes are always zeroes. This area was
> used for domain metadata in older versions.
>
>
> Please share more details:
>
> - output of "lsblk"
> - output of "multipath -ll"
> - output of "/usr/libexec/vdsm/fc-scan -v"
> - output of "vgs -o +tags problem-domain-id"
> - output of "lvs -o +tags problem-domain-id"
> - contents of /etc/multipath.conf
> - contents of /etc/multipath.conf.d/*.conf
> - /var/log/messages since the issue started
> - /var/log/vdsm/vdsm.log* since the issue started on one of the hosts
>
> A bug is probably the best place to keep these logs and make it easy to trac.

Please see https://bugzilla.redhat.com/show_bug.cgi?id=1768821

>
> Thanks,
> Nir

Thank you!
Oliver
___
Users mailing list -- users@ovirt.org<mailto:users@ovirt.org>
To unsubscribe send an email to 
users-le...@ovirt.org<mailto:users-le...@ovirt.org> Privacy Statement: 
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QZ5ZN2S7N54JYVV3RWOYOHTEAWFQ23Q7/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/H5MDS2RZXPE65CMQEOF6WN7ZVWGCDETO/


[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-12-04 Thread Albl, Oliver
Hi Amit,



  unfortunately no success.



Dec 4, 2019, 3:41:36 PM

Storage Domain HOST_LUN_219 (Data Center xxx) was deactivated by system because 
it's not visible by any of the hosts.



Dec 4, 2019, 3:35:09 PM

Failed to update VMs/Templates OVF data for Storage Domain HOST_LUN_219 in Data 
Center Production.



Dec 4, 2019, 3:35:09 PM

Failed to update OVF disks 77c64b39-fe50-4d05-b77f-8131ad1f95f9, OVF data isn't 
updated on those OVF stores (Data Center Production, Storage Domain 
HOST_LUN_219).



All the best,

Oliver



Von: Amit Bawer 
Gesendet: Mittwoch, 4. Dezember 2019 15:20
An: Albl, Oliver 
Cc: users@ovirt.org; Nir Soffer 
Betreff: Re: [ovirt-users] Re: Cannot activate/deactivate storage domain



Hi Oliver,



For deactivating the unresponsive storage domains, you can use the Compute -> 
Data Centers -> Maintenance option with "Ignore OVF update failure" checked.

This will force deactivation of the SD.



Will provide further details about the issue in the ticket.





On Tue, Dec 3, 2019 at 12:02 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:

   Hi,

 does anybody have an advice how to activate or safely remove that storage 
domain?

   Thank you!
   Oliver
   -Ursprüngliche Nachricht-
   Von: Oliver Albl mailto:oliver.a...@fabasoft.com>>
   Gesendet: Dienstag, 5. November 2019 11:20
   An: users@ovirt.org<mailto:users@ovirt.org>
   Betreff: [ovirt-users] Re: Cannot activate/deactivate storage domain

   > On Mon, Nov 4, 2019 at 9:18 PM Albl, Oliver 
http://fabasoft.com> wrote:
   >
   > What was the last change in the system? upgrade? network change? storage 
change?
   >

   Last change was four weeks ago ovirt upgrade from 4.3.3 to 4.3.6.7 
(including CentOS hosts to 7.7 1908)

   >
   > This is expected if some domain is not accessible on all hosts.
   >
   >
   > This means sanlock timed out renewing the lockspace
   >
   >
   > If a host cannot access all storage domain in the DC, the system set
   > it to non-operational, and will probably try to reconnect it later.
   >
   >
   > This means reading 4k from start of the metadata lv took 9.6 seconds.
   > Something in
   > the way to storage is bad (kernel, network, storage).
   >
   >
   > We 20 seconds (4 retires, 5 seconds per retry) gracetime in multipath
   > when there are no active paths, before I/O fails, pausing the VM. We
   > also resume paused VMs when storage monitoring works again, so maybe
   > the VM were paused and resumed.
   >
   > However for storage monitoring we have strict 10 seconds timeout. If
   > reading from the metadata lv times out or fail and does not operated
   > normally after
   > 5 minutes, the
   > domain will become inactive.
   >
   >
   > This can explain the read timeouts.
   >
   >
   > This looks the right way to troubleshoot this.
   >
   >
   > We need vdsm logs to understand this failure.
   >
   >
   > This does not mean OVF is corrupted, only that we could not store new
   > data. The older data on the other OVFSTORE disk is probably fine.
   > Hopefuly the system will not try to write to the other OVFSTORE disk
   > overwriting the last good version.
   >
   >
   > This is normal, the first 2048 bytes are always zeroes. This area was
   > used for domain metadata in older versions.
   >
   >
   > Please share more details:
   >
   > - output of "lsblk"
   > - output of "multipath -ll"
   > - output of "/usr/libexec/vdsm/fc-scan -v"
   > - output of "vgs -o +tags problem-domain-id"
   > - output of "lvs -o +tags problem-domain-id"
   > - contents of /etc/multipath.conf
   > - contents of /etc/multipath.conf.d/*.conf
   > - /var/log/messages since the issue started
   > - /var/log/vdsm/vdsm.log* since the issue started on one of the hosts
   >
   > A bug is probably the best place to keep these logs and make it easy to 
trac.

   Please see https://bugzilla.redhat.com/show_bug.cgi?id=1768821

   >
   > Thanks,
   > Nir

   Thank you!
   Oliver
   ___
   Users mailing list -- users@ovirt.org<mailto:users@ovirt.org>
   To unsubscribe send an email to 
users-le...@ovirt.org<mailto:users-le...@ovirt.org> Privacy Statement: 
https://www.ovirt.org/site/privacy-policy/
   oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
   List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QZ5ZN2S7N54JYVV3RWOYOHTEAWFQ23Q7/

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UVYHPVKPV5575BQ4XUYOFGZV4KZ2IF2H/


[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-12-04 Thread Albl, Oliver


smime.p7m
Description: S/MIME encrypted message
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BWQ6JWCEA2SCQX4YSL3Y5Z5IHONQ7ZH3/


[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-12-03 Thread Albl, Oliver
Hi,

  does anybody have an advice how to activate or safely remove that storage 
domain? 

Thank you!
Oliver
-Ursprüngliche Nachricht-
Von: Oliver Albl  
Gesendet: Dienstag, 5. November 2019 11:20
An: users@ovirt.org
Betreff: [ovirt-users] Re: Cannot activate/deactivate storage domain

> On Mon, Nov 4, 2019 at 9:18 PM Albl, Oliver  wrote:
> 
> What was the last change in the system? upgrade? network change? storage 
> change?
> 

Last change was four weeks ago ovirt upgrade from 4.3.3 to 4.3.6.7 (including 
CentOS hosts to 7.7 1908)

> 
> This is expected if some domain is not accessible on all hosts.
> 
> 
> This means sanlock timed out renewing the lockspace
> 
> 
> If a host cannot access all storage domain in the DC, the system set 
> it to non-operational, and will probably try to reconnect it later.
> 
> 
> This means reading 4k from start of the metadata lv took 9.6 seconds.
> Something in
> the way to storage is bad (kernel, network, storage).
> 
> 
> We 20 seconds (4 retires, 5 seconds per retry) gracetime in multipath 
> when there are no active paths, before I/O fails, pausing the VM. We 
> also resume paused VMs when storage monitoring works again, so maybe 
> the VM were paused and resumed.
> 
> However for storage monitoring we have strict 10 seconds timeout. If 
> reading from the metadata lv times out or fail and does not operated 
> normally after
> 5 minutes, the
> domain will become inactive.
> 
> 
> This can explain the read timeouts.
> 
> 
> This looks the right way to troubleshoot this.
> 
> 
> We need vdsm logs to understand this failure.
> 
> 
> This does not mean OVF is corrupted, only that we could not store new 
> data. The older data on the other OVFSTORE disk is probably fine. 
> Hopefuly the system will not try to write to the other OVFSTORE disk 
> overwriting the last good version.
> 
> 
> This is normal, the first 2048 bytes are always zeroes. This area was 
> used for domain metadata in older versions.
> 
> 
> Please share more details:
> 
> - output of "lsblk"
> - output of "multipath -ll"
> - output of "/usr/libexec/vdsm/fc-scan -v"
> - output of "vgs -o +tags problem-domain-id"
> - output of "lvs -o +tags problem-domain-id"
> - contents of /etc/multipath.conf
> - contents of /etc/multipath.conf.d/*.conf
> - /var/log/messages since the issue started
> - /var/log/vdsm/vdsm.log* since the issue started on one of the hosts
> 
> A bug is probably the best place to keep these logs and make it easy to trac.

Please see https://bugzilla.redhat.com/show_bug.cgi?id=1768821

> 
> Thanks,
> Nir

Thank you!
Oliver
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: 
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QZ5ZN2S7N54JYVV3RWOYOHTEAWFQ23Q7/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PQOG4UN7RCUMI2XGURKAQYOJ6Y72MEWH/


[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-11-28 Thread Albl, Oliver
Hi,

  any ideas if or how I can recover the storage domain? I will need to destroy 
it, as the ongoing scsi scans are becoming an impediment.

Thank you and all the best,
Oliver

-Ursprüngliche Nachricht-
Von: Oliver Albl 
Gesendet: Dienstag, 5. November 2019 11:20
An: users@ovirt.org
Betreff: [ovirt-users] Re: Cannot activate/deactivate storage domain

> On Mon, Nov 4, 2019 at 9:18 PM Albl, Oliver  wrote:
>
> What was the last change in the system? upgrade? network change? storage
> change?
>

Last change was four weeks ago ovirt upgrade from 4.3.3 to 4.3.6.7 (including
CentOS hosts to 7.7 1908)

>
> This is expected if some domain is not accessible on all hosts.
>
>
> This means sanlock timed out renewing the lockspace
>
>
> If a host cannot access all storage domain in the DC, the system set
> it to non-operational, and will probably try to reconnect it later.
>
>
> This means reading 4k from start of the metadata lv took 9.6 seconds.
> Something in
> the way to storage is bad (kernel, network, storage).
>
>
> We 20 seconds (4 retires, 5 seconds per retry) gracetime in multipath
> when there are no active paths, before I/O fails, pausing the VM. We
> also resume paused VMs when storage monitoring works again, so maybe
> the VM were paused and resumed.
>
> However for storage monitoring we have strict 10 seconds timeout. If
> reading from the metadata lv times out or fail and does not operated
> normally after
> 5 minutes, the
> domain will become inactive.
>
>
> This can explain the read timeouts.
>
>
> This looks the right way to troubleshoot this.
>
>
> We need vdsm logs to understand this failure.
>
>
> This does not mean OVF is corrupted, only that we could not store new
> data. The older data on the other OVFSTORE disk is probably fine.
> Hopefuly the system will not try to write to the other OVFSTORE disk
> overwriting the last good version.
>
>
> This is normal, the first 2048 bytes are always zeroes. This area was
> used for domain metadata in older versions.
>
>
> Please share more details:
>
> - output of "lsblk"
> - output of "multipath -ll"
> - output of "/usr/libexec/vdsm/fc-scan -v"
> - output of "vgs -o +tags problem-domain-id"
> - output of "lvs -o +tags problem-domain-id"
> - contents of /etc/multipath.conf
> - contents of /etc/multipath.conf.d/*.conf
> - /var/log/messages since the issue started
> - /var/log/vdsm/vdsm.log* since the issue started on one of the hosts
>
> A bug is probably the best place to keep these logs and make it easy to
> trac.

Please see https://bugzilla.redhat.com/show_bug.cgi?id=1768821

>
> Thanks,
> Nir

Thank you!
Oliver
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QZ5ZN2S7N54JYVV3RWOYOHTEAWFQ23Q7/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZF2JJRFTP43XZNLFYXQIAOJKVDGYKAHL/


[ovirt-users] Re: Cannot activate/deactivate storage domain

2019-11-28 Thread Albl, Oliver


smime.p7m
Description: S/MIME encrypted message
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GSBKAWBDS432YCFCI76AOE4TKZDK72F6/


[ovirt-users] Cannot activate/deactivate storage domain

2019-11-04 Thread Albl, Oliver
Hi all,

  I run an oVirt 4.3.6.7-1.el7 installation (50+ hosts, 40+ FC storage domains 
on two all-flash arrays) and experienced a problem accessing single storage 
domains.

As a result, hosts were taken "not operational" because they could not see all 
storage domains, SPM started to move around the hosts.

oVirt messages start with:
2019-11-04 15:10:22.739+01 | VDSM HOST082 command SpmStatusVDS failed: (-202, 
'Sanlock resource read failure', 'IO timeout')
2019-11-04 15:10:22.781+01 | Invalid status on Data Center . Setting Data 
Center status to Non Responsive (On host HOST82, Error: General Exception).
...
2019-11-04 15:13:58.836+01 | Host HOST017 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:58.85+01  | Host HOST005 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:58.85+01  | Host HOST012 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:58.851+01 | Host HOST002 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:58.851+01 | Host HOST010 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:58.851+01 | Host HOST011 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:58.852+01 | Host HOST004 cannot access the Storage Domain(s) 
HOST_LUN_204 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.011+01 | Host HOST017 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.238+01 | Host HOST004 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.249+01 | Host HOST005 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.255+01 | Host HOST012 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.273+01 | Host HOST002 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.279+01 | Host HOST010 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:13:59.386+01 | Host HOST011 cannot access the Storage Domain(s) 
 attached to the Data Center . Setting Host state to 
Non-Operational.
2019-11-04 15:15:14.145+01 | Storage domain HOST_LUN_221 experienced a high 
latency of 9.60953 seconds from host HOST038. This may cause performance and 
functional issues. Please consult your Storage Administrator.

The problem mainly affected two storage domains (on the same array) but I also 
saw single messages for other storage domains (one the other array as well).

Storage domains stayed available to the hosts, all VMs continued to run.

When constantly reading from the storage domains (/bin/dd iflag=direct 
if=  bs=4096 count=1 of=/dev/null) we got expected 20+ MBytes/s on 
all but some storage domains. One of them showed "transfer rates" around 200 
Bytes/s, but went up to normal performance from time to time. Transfer rate to 
this domain was different between the hosts.

/var/log/messages contain qla2xxx abort messages on almost all hosts. There are 
no errors on SAN switches or storage array (but vendor is still investigating). 
I did not see high load on the storage array.

The system seemed to stabilize when I stopped all VMs on the affected storage 
domain and this storage domain became "inactive". Currently, this storage 
domain still is inactive and we cannot place it in maintenance mode ("Failed to 
deactivate Storage Domain") nor activate it. OVF Metadata seems to be corrupt 
as well (failed to update OVF disks , OVF data isn't updated on those OVF 
stores). The first six 512 byte blocks of /dev//metadata seem to contain 
only zeros.

Any advice on how to proceed here?
Is there a way to recover this storage domain?

All the best,
Oliver

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XIFBJG4BYAXT4KKRDGYGHBXKD36E2I52/


[ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

2019-06-18 Thread Albl, Oliver
Hi,

  I filed https://bugzilla.redhat.com/show_bug.cgi?id=1721481

Stopping ovirt-engine-dwhd does not terminate the long running queries.

All the best,
Oliver

Von: Shirly Radco 
Gesendet: Montag, 17. Juni 2019 12:08
An: Albl, Oliver 
Cc: slev...@redhat.com; users@ovirt.org; emes...@redhat.com; d...@redhat.com
Betreff: Re: [ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

Hi,

Please open a bug with the details you added here.
Please attach setup logs (all of them since 4.2, including) and 
ovirt-engine-dwh log so we can investigate this issue.

If you stop ovirt-engine-dwhd process, does pg calm down?

Thank you

--

Shirly Radco

BI Senior Software Engineer

Red Hat<https://www.redhat.com/>
[https://marketing-outfit-prod-images.s3-us-west-2.amazonaws.com/f5445ae0c9ddafd5b2f1836854d7416a/Logo-RedHat-Email.png]<https://www.redhat.com/>




On Sun, Jun 16, 2019 at 12:07 PM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:
Hi,

ovirt_engine_history=# SELECT id, version, script, checksum, installed_by, 
started_at, ended_at, state, current, comment
FROM schema_version
ORDER BY version DESC limit 10  ;
id | version  |  script 
  | checksum | installed_by | 
started_at |  ended_at  |   state   | current |
comment
+--+---+--+--+++---+-+---

50 | 04030020 | upgrade/04_03_0020_update_rx_tx_rate_percition.sql  
  | c480b070cc69681cf62fb853f3d7139f | ovirt_engine_history | 2019-06-14 
18:14:44.960503 | 2019-06-14 18:14:44.991593 | SKIPPED   | t   | Installed
already by 04020040
49 | 04030010 | upgrade/04_03_0010_update_network_name_length.sql   
  | a1a0d75560575cdc60c0bbaad2cda773 | ovirt_engine_history | 2019-06-14 
18:14:44.904426 | 2019-06-14 18:14:44.935725 | SKIPPED   | f   | Installed
already by 04020030
48 | 04020040 | upgrade/04_02_0040_update_rx_tx_rate_percition.sql  
  | c480b070cc69681cf62fb853f3d7139f | ovirt_engine_history | 2018-07-07 
14:34:42.505446 | 2018-07-07 14:36:31.662577 | INSTALLED | f   |
47 | 04020030 | upgrade/04_02_0030_update_network_name_length.sql   
  | a1a0d75560575cdc60c0bbaad2cda773 | ovirt_engine_history | 2018-07-07 
14:34:42.438056 | 2018-07-07 14:34:42.482705 | INSTALLED | f   |
46 | 04020020 | 
upgrade/04_02_0020__updated_vm_interface_history_id_to_bigint.sql | 
58a8afa29fc720dc87f37b7f9c9e0151 | ovirt_engine_history | 2018-04-18 
17:17:04.908667 | 2018-04-18 17:17:39.111339 | INSTALLED | f   |
45 | 04020010 | upgrade/04_02_0010_updated_vm_template_name_length.sql  
  | 4b5391f40e8787e3b1033635aafe18a1 | ovirt_engine_history | 2018-01-05 
09:56:39.213757 | 2018-01-05 09:56:39.238775 | SKIPPED   | f   | Installed
already by 04010020
44 | 04010020 | upgrade/04_01_0020_updated_vm_template_name_lentgh.sql  
  | 4b5391f40e8787e3b1033635aafe18a1 | ovirt_engine_history | 2017-10-05 
13:53:04.225474 | 2017-10-05 13:53:04.269508 | INSTALLED | f   |
43 | 04010010 | upgrade/04_01_0010_added_seconds_in_status_to_sample_tables.sql 
  | be7a1b2fc7f03d263b45a613d5bced03 | ovirt_engine_history | 2017-02-03 
13:16:18.29672  | 2017-02-03 13:16:18.320728 | SKIPPED   | f   | Installed
already by 0450
42 | 0450 | upgrade/04_00_0050_added_seconds_in_status_to_sample_tables.sql 
  | be7a1b2fc7f03d263b45a613d5bced03 | ovirt_engine_history | 2016-10-03 
15:13:33.856501 | 2016-10-03 15:13:34.010135 | INSTALLED | f   |
41 | 0440 | upgrade/04_00_0040_drop_all_history_db_foreign_keys.sql 
  | ed8b2c02bea97d0ee21f737614a2d5e3 | ovirt_engine_history | 2016-10-03 
15:13:33.763905 | 2016-10-03 15:13:33.839532 | INSTALLED | f   |
(10 rows)

All the best,
Oliver

Von: Shirly Radco mailto:sra...@redhat.com>>
Gesendet: Sonntag, 16. Juni 2019 11:04
An: Albl, Oliver mailto:oliver.a...@fabasoft.com>>
Cc: slev...@redhat.com<mailto:slev...@redhat.com>; 
users@ovirt.org<mailto:users@ovirt.org>; Eli Mesika 
mailto:emes...@redhat.com>>; Yedidyah Bar David 
mailto:d...@redhat.com>>
Betreff: Re: [ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

Hi,

Please attach here the result of the following query from the 
ovirt_engine_history db:

SELECT id, version, script, checksum, installed_by, started_at, ended_at, 
state, current, comment
FROM schema_version
ORDER BY version DESC limit 10  ;

Best regards,

--

Shirly Radco

BI Senior Software Engineer

Red Hat<https://www.redhat.com/>
[Das Bild wurde vom Absender entfernt.]<https://www.redhat.com/>



-- Forwarded message -
From: Albl, Oliver mailto:oliver.a...@fabasoft.com>>
D

[ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

2019-06-16 Thread Albl, Oliver
Hi,



ovirt_engine_history=# SELECT id, version, script, checksum, installed_by, 
started_at, ended_at, state, current, comment

FROM schema_version

ORDER BY version DESC limit 10  ;

id | version  |  script 
  | checksum | installed_by | 
started_at |  ended_at  |   state   | current |

comment

+--+---+--+--+++---+-+---



50 | 04030020 | upgrade/04_03_0020_update_rx_tx_rate_percition.sql  
  | c480b070cc69681cf62fb853f3d7139f | ovirt_engine_history | 2019-06-14 
18:14:44.960503 | 2019-06-14 18:14:44.991593 | SKIPPED   | t   | Installed

already by 04020040

49 | 04030010 | upgrade/04_03_0010_update_network_name_length.sql   
  | a1a0d75560575cdc60c0bbaad2cda773 | ovirt_engine_history | 2019-06-14 
18:14:44.904426 | 2019-06-14 18:14:44.935725 | SKIPPED   | f   | Installed

already by 04020030

48 | 04020040 | upgrade/04_02_0040_update_rx_tx_rate_percition.sql  
  | c480b070cc69681cf62fb853f3d7139f | ovirt_engine_history | 2018-07-07 
14:34:42.505446 | 2018-07-07 14:36:31.662577 | INSTALLED | f   |

47 | 04020030 | upgrade/04_02_0030_update_network_name_length.sql   
  | a1a0d75560575cdc60c0bbaad2cda773 | ovirt_engine_history | 2018-07-07 
14:34:42.438056 | 2018-07-07 14:34:42.482705 | INSTALLED | f   |

46 | 04020020 | 
upgrade/04_02_0020__updated_vm_interface_history_id_to_bigint.sql | 
58a8afa29fc720dc87f37b7f9c9e0151 | ovirt_engine_history | 2018-04-18 
17:17:04.908667 | 2018-04-18 17:17:39.111339 | INSTALLED | f   |

45 | 04020010 | upgrade/04_02_0010_updated_vm_template_name_length.sql  
  | 4b5391f40e8787e3b1033635aafe18a1 | ovirt_engine_history | 2018-01-05 
09:56:39.213757 | 2018-01-05 09:56:39.238775 | SKIPPED   | f   | Installed

already by 04010020

44 | 04010020 | upgrade/04_01_0020_updated_vm_template_name_lentgh.sql  
  | 4b5391f40e8787e3b1033635aafe18a1 | ovirt_engine_history | 2017-10-05 
13:53:04.225474 | 2017-10-05 13:53:04.269508 | INSTALLED | f   |

43 | 04010010 | upgrade/04_01_0010_added_seconds_in_status_to_sample_tables.sql 
  | be7a1b2fc7f03d263b45a613d5bced03 | ovirt_engine_history | 2017-02-03 
13:16:18.29672  | 2017-02-03 13:16:18.320728 | SKIPPED   | f   | Installed

already by 0450

42 | 0450 | upgrade/04_00_0050_added_seconds_in_status_to_sample_tables.sql 
  | be7a1b2fc7f03d263b45a613d5bced03 | ovirt_engine_history | 2016-10-03 
15:13:33.856501 | 2016-10-03 15:13:34.010135 | INSTALLED | f   |

41 | 0440 | upgrade/04_00_0040_drop_all_history_db_foreign_keys.sql 
  | ed8b2c02bea97d0ee21f737614a2d5e3 | ovirt_engine_history | 2016-10-03 
15:13:33.763905 | 2016-10-03 15:13:33.839532 | INSTALLED | f   |

(10 rows)



All the best,

Oliver



Von: Shirly Radco 
Gesendet: Sonntag, 16. Juni 2019 11:04
An: Albl, Oliver 
Cc: slev...@redhat.com; users@ovirt.org; Eli Mesika ; 
Yedidyah Bar David 
Betreff: Re: [ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4



Hi,



Please attach here the result of the following query from the 
ovirt_engine_history db:



SELECT id, version, script, checksum, installed_by, started_at, ended_at, 
state, current, comment

FROM schema_version

ORDER BY version DESC limit 10  ;



Best regards,

--

Shirly Radco

BI Senior Software Engineer

Red Hat<https://www.redhat.com/>

<https://www.redhat.com/>





-- Forwarded message -
From: Albl, Oliver mailto:oliver.a...@fabasoft.com>>
Date: Sun, Jun 16, 2019 at 11:42 AM
Subject: AW: [ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4
To: sra...@redhat.com<mailto:sra...@redhat.com> 
mailto:sra...@redhat.com>>
Cc: slev...@redhat.com<mailto:slev...@redhat.com> 
mailto:slev...@redhat.com>>, 
users@ovirt.org<mailto:users@ovirt.org> 
mailto:users@ovirt.org>>



Hi,



  rebooted oVirt engine VM (no hosted engine), same result. DHW packages are:



ovirt-engine-dwh.noarch   4.3.0-1.el7@ovirt-4.3

ovirt-engine-dwh-setup.noarch 4.3.0-1.el7@ovirt-4.3



After reboot the following two queries start running again (they were running 
for more than 24 hours before this reboot):



SELECT name, AVG(total) AS total, AVG(used) AS used, MAX(previous_used) AS 
previous_used, MAX(previous_total) AS previous_total, 
MAX(vm_disk_usage_percent) AS us

age_percent FROM (SELECT history_datetime, vms.vm_name AS name, 
SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) AS used, 
COALESCE(MAX(previous_actual_size), 0) AS previous_used, COALESCE(MAX(previous_

total_size), 0) AS previous_total, CAS

[ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

2019-06-16 Thread Albl, Oliver


smime.p7m
Description: S/MIME encrypted message
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2CJYQSHMHWQLFWM2FXYQV5KW3XVJYDDO/


[ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

2019-06-16 Thread Albl, Oliver
Hi,



  rebooted oVirt engine VM (no hosted engine), same result. DHW packages are:



ovirt-engine-dwh.noarch   4.3.0-1.el7@ovirt-4.3

ovirt-engine-dwh-setup.noarch 4.3.0-1.el7@ovirt-4.3



After reboot the following two queries start running again (they were running 
for more than 24 hours before this reboot):



SELECT name, AVG(total) AS total, AVG(used) AS used, MAX(previous_used) AS 
previous_used, MAX(previous_total) AS previous_total, 
MAX(vm_disk_usage_percent) AS us

age_percent FROM (SELECT history_datetime, vms.vm_name AS name, 
SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) AS used, 
COALESCE(MAX(previous_actual_size), 0) AS previous_used, COALESCE(MAX(previous_

total_size), 0) AS previous_total, CASE WHEN MAX(vm_disk_size_mb) = 0 THEN 0 
ELSE (CAST(MAX(vm_disk_actual_size_mb) AS DECIMAL)/ MAX(vm_disk_size_mb) * 100) 
END AS vm_disk_usage_percent FROM v4_2_statistics_vms_

disks_resources_usage_samples samples INNER JOIN 
v4_2_latest_configuration_vms_devices devices ON devices.device_id = vm_disk_id 
INNER JOIN v4_2_latest_configuration_vms vms ON vms.vm_id = devices.vm_id INNER 
JO

IN v4_2_latest_configuration_vms_disks disks ON disks.vm_disk_id = 
samples.vm_disk_id LEFT OUTER JOIN (SELECT vm_name, AVG(previous_actual_size) 
AS previous_actual_size, AVG(previous_total_size) AS previous_tota

l_size FROM (SELEC





All the best,

Oliver



Von: Shirly Radco 
Gesendet: Sonntag, 16. Juni 2019 10:34
An: Albl, Oliver 
Cc: slev...@redhat.com; users@ovirt.org
Betreff: Re: [ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4



Hi,



Please try to restart postgres an ovirt-engine-dwhd.

It looks that this is a 4.2 query.

What version are the dwh packages?



Best,


--

Shirly Radco

BI Senior Software Engineer

Red Hat<https://www.redhat.com/>

<https://www.redhat.com/>







On Sun, Jun 16, 2019 at 11:10 AM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:

   Hi,



 I am not using hosted engine but a dedicated installation. The upgrade was 
performed with following steps:



   yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release43.rpm

   yum update "ovirt-*-setup*"

   engine-setup

   yum update



   All the best,

   Oliver



   Von: Shani Leviim mailto:slev...@redhat.com>>
   Gesendet: Sonntag, 16. Juni 2019 09:03
   An: Albl, Oliver mailto:oliver.a...@fabasoft.com>>
   Cc: users@ovirt.org<mailto:users@ovirt.org>
   Betreff: Re: [ovirt-users] Cannot access dashboard after upgrading to 4.3.4



   Hi,

   Did you run 'engine-setup' and 'yum update' after upgrading ovirt?



   Taken from here as a reference: 
https://ovirt.org/documentation/self-hosted/chap-upgrading_the_self-hosted_engine.html



   Regards,

   Shani Leviim





   On Sat, Jun 15, 2019 at 10:26 AM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:

  Hi all,



I upgraded our installation from 4.2.8 to 4.3.4 and I am unable to 
access the dashboard in UI as there is only “Loading data…” displayed. There 
are postgres processes running on the engine taking 100% cpu. One is running 
for more than 14 hours:



  SELECT name, AVG(total) AS total, AVG(used) AS used, MAX(previous_used) 
AS previous_used, MAX(previous_total) AS previous_total, MAX(vm_disk_

  usage_percent) AS usage_percent FROM (SELECT history_datetime, 
vms.vm_name AS name, SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) 
AS used, COALESCE(MAX(previous_actual_size), 0)

  AS previous_used, COALESCE(MAX(previous_total_size), 0) AS 
previous_total, CASE WHEN MAX(vm_disk_size_mb) = 0 THEN 0 ELSE 
(CAST(MAX(vm_disk_actual_size_mb) AS DECIMAL)/ MAX(vm_disk_size_mb) *

  100) END AS vm_disk_usage_percent FROM 
v4_2_statistics_vms_disks_resources_usage_samples samples INNER JOIN 
v4_2_latest_configuration_vms_devices devices ON devices.device_id = vm_disk_id 
IN

  NER JOIN v4_2_latest_configuration_vms vms ON vms.vm_id = devices.vm_id 
INNER JOIN v4_2_latest_configuration_vms_disks disks ON disks.vm_disk_id = 
samples.vm_disk_id LEFT OUTER JOIN (SELECT v

  m_name, AVG(previous_actual_size) AS previous_actual_size, 
AVG(previous_total_size) AS previous_total_size FROM (SELEC

  13783 | 2019-06-15 08:41:52.349769+02 | active | SELECT name, AVG(total) 
AS total, AVG(used) AS used, MAX(previous_used) AS previous_used, 
MAX(previous_total) AS previous_total, MAX(vm_disk_

  usage_percent) AS usage_percent FROM (SELECT history_datetime, 
vms.vm_name AS name, SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) 
AS used, COALESCE(MAX(previous_actual_size), 0)

  AS previous_used, COALESCE(MAX(previous_total_size), 0) AS 
previous_total, CASE WHEN MAX(vm_disk_size_mb) = 0 THEN 0 ELSE 
(CAST(MAX(vm_disk_actual_size_mb) AS DECIMAL)/ MAX(vm_disk_size_mb) *

  100) END AS vm_disk_usage_percent FROM 
v4_2_statistics_

[ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

2019-06-16 Thread Albl, Oliver


smime.p7m
Description: S/MIME encrypted message
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4QXJQ7Y2RNS3RMHOAW2QWBJAL5I7MCRC/


[ovirt-users] Re: Cannot access dashboard after upgrading to 4.3.4

2019-06-16 Thread Albl, Oliver
Hi,



  I am not using hosted engine but a dedicated installation. The upgrade was 
performed with following steps:



yum install http://resources.ovirt.org/pub/yum-repo/ovirt-release43.rpm

yum update "ovirt-*-setup*"

engine-setup

yum update



All the best,

Oliver



Von: Shani Leviim 
Gesendet: Sonntag, 16. Juni 2019 09:03
An: Albl, Oliver 
Cc: users@ovirt.org
Betreff: Re: [ovirt-users] Cannot access dashboard after upgrading to 4.3.4



Hi,

Did you run 'engine-setup' and 'yum update' after upgrading ovirt?



Taken from here as a reference: 
https://ovirt.org/documentation/self-hosted/chap-upgrading_the_self-hosted_engine.html



Regards,

Shani Leviim





On Sat, Jun 15, 2019 at 10:26 AM Albl, Oliver 
mailto:oliver.a...@fabasoft.com>> wrote:

   Hi all,



 I upgraded our installation from 4.2.8 to 4.3.4 and I am unable to access 
the dashboard in UI as there is only “Loading data…” displayed. There are 
postgres processes running on the engine taking 100% cpu. One is running for 
more than 14 hours:



   SELECT name, AVG(total) AS total, AVG(used) AS used, MAX(previous_used) AS 
previous_used, MAX(previous_total) AS previous_total, MAX(vm_disk_

   usage_percent) AS usage_percent FROM (SELECT history_datetime, vms.vm_name 
AS name, SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) AS used, 
COALESCE(MAX(previous_actual_size), 0)

   AS previous_used, COALESCE(MAX(previous_total_size), 0) AS previous_total, 
CASE WHEN MAX(vm_disk_size_mb) = 0 THEN 0 ELSE 
(CAST(MAX(vm_disk_actual_size_mb) AS DECIMAL)/ MAX(vm_disk_size_mb) *

   100) END AS vm_disk_usage_percent FROM 
v4_2_statistics_vms_disks_resources_usage_samples samples INNER JOIN 
v4_2_latest_configuration_vms_devices devices ON devices.device_id = vm_disk_id 
IN

   NER JOIN v4_2_latest_configuration_vms vms ON vms.vm_id = devices.vm_id 
INNER JOIN v4_2_latest_configuration_vms_disks disks ON disks.vm_disk_id = 
samples.vm_disk_id LEFT OUTER JOIN (SELECT v

   m_name, AVG(previous_actual_size) AS previous_actual_size, 
AVG(previous_total_size) AS previous_total_size FROM (SELEC

   13783 | 2019-06-15 08:41:52.349769+02 | active | SELECT name, AVG(total) AS 
total, AVG(used) AS used, MAX(previous_used) AS previous_used, 
MAX(previous_total) AS previous_total, MAX(vm_disk_

   usage_percent) AS usage_percent FROM (SELECT history_datetime, vms.vm_name 
AS name, SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) AS used, 
COALESCE(MAX(previous_actual_size), 0)

   AS previous_used, COALESCE(MAX(previous_total_size), 0) AS previous_total, 
CASE WHEN MAX(vm_disk_size_mb) = 0 THEN 0 ELSE 
(CAST(MAX(vm_disk_actual_size_mb) AS DECIMAL)/ MAX(vm_disk_size_mb) *

   100) END AS vm_disk_usage_percent FROM 
v4_2_statistics_vms_disks_resources_usage_samples samples INNER JOIN 
v4_2_latest_configuration_vms_devices devices ON devices.device_id = vm_disk_id 
IN

   NER JOIN v4_2_latest_configuration_vms vms ON vms.vm_id = devices.vm_id 
INNER JOIN v4_2_latest_configuration_vms_disks disks ON disks.vm_disk_id = 
samples.vm_disk_id LEFT OUTER JOIN (SELECT v

   m_name, AVG(previous_actual_size) AS previous_actual_size, 
AVG(previous_total_size) AS previous_total_size FROM (SELEC…



   oVirt Engine is configured with 128 GB memory and 24 cores.



   Please advice how to adress this.



   Thank you and all the best,

   Oliver



   ___
   Users mailing list -- users@ovirt.org<mailto:users@ovirt.org>
   To unsubscribe send an email to 
users-le...@ovirt.org<mailto:users-le...@ovirt.org>
   Privacy Statement: https://www.ovirt.org/site/privacy-policy/
   oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
   List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JJWRQKGCIHRG765YJCEG5CV6BJTFPFAO/

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BW6GWQ43ACOPMXBZZKLRPF6TUTVOU7WD/


[ovirt-users] Cannot access dashboard after upgrading to 4.3.4

2019-06-15 Thread Albl, Oliver
Hi all,



  I upgraded our installation from 4.2.8 to 4.3.4 and I am unable to access the 
dashboard in UI as there is only "Loading data..." displayed. There are 
postgres processes running on the engine taking 100% cpu. One is running for 
more than 14 hours:



SELECT name, AVG(total) AS total, AVG(used) AS used, MAX(previous_used) AS 
previous_used, MAX(previous_total) AS previous_total, MAX(vm_disk_

usage_percent) AS usage_percent FROM (SELECT history_datetime, vms.vm_name AS 
name, SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) AS used, 
COALESCE(MAX(previous_actual_size), 0)

AS previous_used, COALESCE(MAX(previous_total_size), 0) AS previous_total, CASE 
WHEN MAX(vm_disk_size_mb) = 0 THEN 0 ELSE (CAST(MAX(vm_disk_actual_size_mb) AS 
DECIMAL)/ MAX(vm_disk_size_mb) *

100) END AS vm_disk_usage_percent FROM 
v4_2_statistics_vms_disks_resources_usage_samples samples INNER JOIN 
v4_2_latest_configuration_vms_devices devices ON devices.device_id = vm_disk_id 
IN

NER JOIN v4_2_latest_configuration_vms vms ON vms.vm_id = devices.vm_id INNER 
JOIN v4_2_latest_configuration_vms_disks disks ON disks.vm_disk_id = 
samples.vm_disk_id LEFT OUTER JOIN (SELECT v

m_name, AVG(previous_actual_size) AS previous_actual_size, 
AVG(previous_total_size) AS previous_total_size FROM (SELEC

13783 | 2019-06-15 08:41:52.349769+02 | active | SELECT name, AVG(total) AS 
total, AVG(used) AS used, MAX(previous_used) AS previous_used, 
MAX(previous_total) AS previous_total, MAX(vm_disk_

usage_percent) AS usage_percent FROM (SELECT history_datetime, vms.vm_name AS 
name, SUM(vm_disk_size_mb) AS total, SUM(vm_disk_actual_size_mb) AS used, 
COALESCE(MAX(previous_actual_size), 0)

AS previous_used, COALESCE(MAX(previous_total_size), 0) AS previous_total, CASE 
WHEN MAX(vm_disk_size_mb) = 0 THEN 0 ELSE (CAST(MAX(vm_disk_actual_size_mb) AS 
DECIMAL)/ MAX(vm_disk_size_mb) *

100) END AS vm_disk_usage_percent FROM 
v4_2_statistics_vms_disks_resources_usage_samples samples INNER JOIN 
v4_2_latest_configuration_vms_devices devices ON devices.device_id = vm_disk_id 
IN

NER JOIN v4_2_latest_configuration_vms vms ON vms.vm_id = devices.vm_id INNER 
JOIN v4_2_latest_configuration_vms_disks disks ON disks.vm_disk_id = 
samples.vm_disk_id LEFT OUTER JOIN (SELECT v

m_name, AVG(previous_actual_size) AS previous_actual_size, 
AVG(previous_total_size) AS previous_total_size FROM (SELEC...



oVirt Engine is configured with 128 GB memory and 24 cores.



Please advice how to adress this.



Thank you and all the best,

Oliver



___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JJWRQKGCIHRG765YJCEG5CV6BJTFPFAO/


[ovirt-users] Re: Cannot create VMs

2018-10-18 Thread Albl, Oliver
Hi all,

  resending this message as it came in empty:

I run a small oVirt installation where I automatically create VMs, run some 
workload and delete them again. Today I ran into a situation where I could not 
create VMs anymore. Messages were:

Failed to create VM  (User: …).
Failed to complete VM  creation.

Any ideas about the cause for this? I filed a BZ 
(https://bugzilla.redhat.com/show_bug.cgi?id=1640797) containing the engine 
logs.

I am on oVirt 4.2.6.4-1.el7 on CentOS 7.5

Viele Grüße,
Oliver


-Ursprüngliche Nachricht-
Von: Albl, Oliver 
Gesendet: Donnerstag, 18. Oktober 2018 20:49
An: users@ovirt.org
Betreff: [ovirt-users] Cannot create VMs

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: 
https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WHIFHWA2SMKUAM5PQI3BROMQ2ASR3AAV/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CKGP3OOWKH6NO4MN5BR2IONVQZ5ZY57C/


[ovirt-users] Re: Cannot create VMs

2018-10-18 Thread Albl, Oliver


smime.p7m
Description: S/MIME encrypted message
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NBTNWBLQJ26AJJY5U3OPLNIR53FPEEZZ/


[ovirt-users] Re: Failed to update VMs/Templates OVF data, cannot change SPM

2018-06-20 Thread Albl, Oliver
200 WARN  (jsonrpc/7) [storage.ResourceManager] 
Resource factory failed to create resource 
'01_img_3ad1987a-8b7d-426d-9d51-4a78cb0a888f.0ebefe5e-9053-4bf1-bdfd-fdb26579c179'.
 Canceling request. (resourceManager:543)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", line 
539, in registerResource
obj = namespaceObj.factory.createResource(name, lockType)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", 
line 193, in createResource
lockType)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", 
line 122, in __getResourceCandidatesList
imgUUID=resourceName)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/image.py", line 206, in 
getChain
if len(uuidlist) == 1 and srcVol.isShared():
  File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 1434, in 
isShared
return self._manifest.isShared()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 141, in 
isShared
return self.getVolType() == sc.type2name(sc.SHARED_VOL)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 134, in 
getVolType
self.voltype = self.getMetaParam(sc.VOLTYPE)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line 118, in 
getMetaParam
meta = self.getMetadata()
  File "/usr/lib/python2.7/site-packages/vdsm/storage/blockVolume.py", line 
112, in getMetadata
md = VolumeMetadata.from_lines(lines)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/volumemetadata.py", line 
103, in from_lines
"Missing metadata key: %s: found: %s" % (e, md))
MetaDataKeyNotFoundError: Meta Data key not found error: ("Missing metadata 
key: 'DOMAIN': found: {}",)
2018-06-20 09:28:28,072+0200 WARN  (jsonrpc/7) 
[storage.ResourceManager.Request] 
(ResName='01_img_3ad1987a-8b7d-426d-9d51-4a78cb0a888f.0ebefe5e-9053-4bf1-bdfd-fdb26579c179',
 ReqID='10c95223-f349-4ac3-ab2f-7a5f3d1c7749') Tried to cancel a processed 
request (resourceManager:187)
2018-06-20 09:28:28,073+0200 INFO  (jsonrpc/7) [vdsm.api] FINISH 
setVolumeDescription error=Could not acquire resource. Probably resource 
factory threw an exception.: () from=:::,51790, flow_id=7e4edb74, 
task_id=5f1fda67-a073-419a-bba5-9bf680c0e5d5 (api:50)
2018-06-20 09:28:28,073+0200 ERROR (jsonrpc/7) [storage.TaskManager.Task] 
(Task='5f1fda67-a073-419a-bba5-9bf680c0e5d5') Unexpected error (task:875)
Traceback (most recent call last):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in 
_run
return fn(*args, **kargs)
  File "", line 2, in setVolumeDescription
  File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in method
ret = func(*args, **kwargs)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 1294, in 
setVolumeDescription
pool.setVolumeDescription(sdUUID, imgUUID, volUUID, description)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line 79, 
in wrapper
return method(self, *args, **kwargs)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 2011, in 
setVolumeDescription
with rm.acquireResource(img_ns, imgUUID, rm.EXCLUSIVE):
  File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", line 
1025, in acquireResource
return _manager.acquireResource(namespace, name, lockType, timeout=timeout)
  File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", line 
475, in acquireResource
raise se.ResourceAcqusitionFailed()
ResourceAcqusitionFailed: Could not acquire resource. Probably resource factory 
threw an exception.: ()
2018-06-20 09:28:28,073+0200 INFO  (jsonrpc/7) [storage.TaskManager.Task] 
(Task='5f1fda67-a073-419a-bba5-9bf680c0e5d5') aborting: Task is aborted: 
u'Could not acquire resource. Probably resource factory threw an exception.: 
()' - code 100 (task:1181)
2018-06-20 09:28:28,073+0200 ERROR (jsonrpc/7) [storage.Dispatcher] FINISH 
setVolumeDescription error=Could not acquire resource. Probably resource 
factory threw an exception.: () (dispatcher:82)

I worked around the SPM problem by running engine-setup (neither vds-client 
Host/StoragePool *Task*, async_task_entries or taskcleaner.sh showed any tasks).

All the best,
Oliver

Von: Nir Soffer mailto:nsof...@redhat.com>>
Gesendet: Dienstag, 19. Juni 2018 23:39
An: Elad Ben Aharon mailto:ebena...@redhat.com>>
Cc: users@ovirt.org<mailto:users@ovirt.org>; Albl, Oliver 
mailto:oliver.a...@fabasoft.com>>
Betreff: [ovirt-users] Re: Failed to update VMs/Templates OVF data, cannot 
change SPM

On Wed, Jun 20, 2018 at 12:11 AM Elad Ben Aharon 
mailto:ebena...@redhat.com>> wrote:
Can you please attach engine.log and vdsm.log?

On Tue, Jun 19, 2018 at 6:58 PM, 

[ovirt-users] Re: Failed to update VMs/Templates OVF data, cannot change SPM

2018-06-19 Thread Albl, Oliver
Hi all,

  does anybody have an idea how to address this? There is also a side effect 
that I cannot change SPM ("Error while executing action: Cannot force select 
SPM. The Storage Pool has running tasks.").

All the best,
Oliver

Von: Albl, Oliver
Gesendet: Mittwoch, 13. Juni 2018 12:32
An: users@ovirt.org
Betreff: Failed to update VMs/Templates OVF data

Hi,

  I have a FC storage domain reporting the following messages every hour:

VDSM command SetVolumeDescriptionVDS failed: Could not acquire resource. 
Probably resource factory threw an exception.: ()
Failed to update OVF disks cb04b55c-10fb-46fe-b9de-3c133a94e6a5, OVF data isn't 
updated on those OVF stores (Data Center VMTEST, Storage Domain VMHOST_LUN_62).
Failed to update VMs/Templates OVF data for Storage Domain VMHOST_LUN_62 in 
Data Center VMTEST.

Trying to manually update OVF results in "Error while executing action 
UpdateOvfStoreForStorageDomain: Internal Engine Error"

I run oVirt 4.2.3.5-1.el7.centos on CentOS 7.5 (3.10.0-862.3.2.el7.x86_64) with 
vdsm-4.20.27.1-1.el7.centos.x86_64

Engine log:

2018-06-13 12:15:35,649+02 WARN  
[org.ovirt.engine.core.dal.job.ExecutionMessageDirector] (default task-12) 
[092d8f27-c0a4-4d78-a8cb-f4738aff71e6] The message key 
'UpdateOvfStoreForStorageDomain' is missing from 'bundles/ExecutionMessages'
2018-06-13 12:15:35,655+02 INFO  
[org.ovirt.engine.core.bll.storage.domain.UpdateOvfStoreForStorageDomainCommand]
 (default task-12) [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] Lock Acquired to 
object 
'EngineLock:{exclusiveLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE, 
373efd46-8aea-4d0e-96cc-1da0debf72d0=STORAGE]', sharedLocks=''}'
2018-06-13 12:15:35,660+02 INFO  
[org.ovirt.engine.core.bll.storage.domain.UpdateOvfStoreForStorageDomainCommand]
 (default task-12) [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] Running command: 
UpdateOvfStoreForStorageDomainCommand internal: false. Entities affected :  ID: 
373efd46-8aea-4d0e-96cc-1da0debf72d0 Type: StorageAction group 
MANIPULATE_STORAGE_DOMAIN with role type ADMIN
2018-06-13 12:15:35,670+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Running command: 
ProcessOvfUpdateForStoragePoolCommand internal: true. Entities affected :  ID: 
3edb5295-3797-4cd0-9b43-f46ec1ee7b14 Type: StoragePool
2018-06-13 12:15:35,674+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Attempting to update VM OVFs in Data Center 
'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Successfully updated VM OVFs in Data Center 
'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Attempting to update template OVFs in Data Center 
'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Successfully updated templates OVFs in Data 
Center 'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Attempting to remove unneeded template/vm OVFs in 
Data Center 'VMTEST'
2018-06-13 12:15:35,680+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Successfully removed unneeded template/vm OVFs in 
Data Center 'VMTEST'
2018-06-13 12:15:35,684+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Lock freed to object 
'EngineLock:{exclusiveLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE, 
373efd46-8aea-4d0e-96cc-1da0debf72d0=STORAGE]', sharedLocks=''}'
2018-06-13 12:15:35,704+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand]
 (default task-12) [24485c23] Lock Acquired to object 
'EngineLock:{exclusiveLocks='[]', 
sharedLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE]'}'
2018-06-13 12:15:35,714+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand]
 (default task-12) [24485c23] Running command: 
ProcessOvfUpdateForStorageDomainCommand internal: true. Entities affected :  
ID: 373efd46-8aea-4d0e-96cc-1da0debf72d0 Type: StorageAction group 
MANIPULATE_STORAGE_DOMAIN with role type ADMIN
2018-06-13 12:15:35,724+02 INFO  
[org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] 
(default task-12) [24485c23] START, SetVolumeDescriptionVDSCommand( 
SetVolumeDescriptionVDSCommandParameters:{storagePoolId='3edb5295-3797-4cd0-9b43-f46ec1ee7b14',
 ignoreFailoverLimit='false', 
storageDomainId='373efd46-8aea-4d0e-96cc-1da0debf72d0', 
imageGroupId='cb04b55c-10fb-46fe-b9de-3c133a94e6a5', 
im

[ovirt-users] Moving from thin to preallocated storage domains

2018-06-13 Thread Albl, Oliver
Hi all,

  I have to move some FC storage domains from thin to preallocated. I would set 
the storage domain to maintenance, convert the LUN from thin to preallocated on 
the array, remove "Discard After Delete" from the advanced settings of the 
storage domain and active it again. Is there anything else I need to take care 
of?

All the best,
Oliver

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VUEQY5DHUC633US5HZQO3N2IQ2TVCZPX/


[ovirt-users] Failed to update VMs/Templates OVF data

2018-06-13 Thread Albl, Oliver
Hi,

  I have a FC storage domain reporting the following messages every hour:

VDSM command SetVolumeDescriptionVDS failed: Could not acquire resource. 
Probably resource factory threw an exception.: ()
Failed to update OVF disks cb04b55c-10fb-46fe-b9de-3c133a94e6a5, OVF data isn't 
updated on those OVF stores (Data Center VMTEST, Storage Domain VMHOST_LUN_62).
Failed to update VMs/Templates OVF data for Storage Domain VMHOST_LUN_62 in 
Data Center VMTEST.

Trying to manually update OVF results in "Error while executing action 
UpdateOvfStoreForStorageDomain: Internal Engine Error"

I run oVirt 4.2.3.5-1.el7.centos on CentOS 7.5 (3.10.0-862.3.2.el7.x86_64) with 
vdsm-4.20.27.1-1.el7.centos.x86_64

Engine log:

2018-06-13 12:15:35,649+02 WARN  
[org.ovirt.engine.core.dal.job.ExecutionMessageDirector] (default task-12) 
[092d8f27-c0a4-4d78-a8cb-f4738aff71e6] The message key 
'UpdateOvfStoreForStorageDomain' is missing from 'bundles/ExecutionMessages'
2018-06-13 12:15:35,655+02 INFO  
[org.ovirt.engine.core.bll.storage.domain.UpdateOvfStoreForStorageDomainCommand]
 (default task-12) [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] Lock Acquired to 
object 
'EngineLock:{exclusiveLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE, 
373efd46-8aea-4d0e-96cc-1da0debf72d0=STORAGE]', sharedLocks=''}'
2018-06-13 12:15:35,660+02 INFO  
[org.ovirt.engine.core.bll.storage.domain.UpdateOvfStoreForStorageDomainCommand]
 (default task-12) [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] Running command: 
UpdateOvfStoreForStorageDomainCommand internal: false. Entities affected :  ID: 
373efd46-8aea-4d0e-96cc-1da0debf72d0 Type: StorageAction group 
MANIPULATE_STORAGE_DOMAIN with role type ADMIN
2018-06-13 12:15:35,670+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Running command: 
ProcessOvfUpdateForStoragePoolCommand internal: true. Entities affected :  ID: 
3edb5295-3797-4cd0-9b43-f46ec1ee7b14 Type: StoragePool
2018-06-13 12:15:35,674+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Attempting to update VM OVFs in Data Center 
'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Successfully updated VM OVFs in Data Center 
'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Attempting to update template OVFs in Data Center 
'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Successfully updated templates OVFs in Data 
Center 'VMTEST'
2018-06-13 12:15:35,678+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Attempting to remove unneeded template/vm OVFs in 
Data Center 'VMTEST'
2018-06-13 12:15:35,680+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Successfully removed unneeded template/vm OVFs in 
Data Center 'VMTEST'
2018-06-13 12:15:35,684+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand]
 (default task-12) [4fd5b59a] Lock freed to object 
'EngineLock:{exclusiveLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE, 
373efd46-8aea-4d0e-96cc-1da0debf72d0=STORAGE]', sharedLocks=''}'
2018-06-13 12:15:35,704+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand]
 (default task-12) [24485c23] Lock Acquired to object 
'EngineLock:{exclusiveLocks='[]', 
sharedLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE]'}'
2018-06-13 12:15:35,714+02 INFO  
[org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand]
 (default task-12) [24485c23] Running command: 
ProcessOvfUpdateForStorageDomainCommand internal: true. Entities affected :  
ID: 373efd46-8aea-4d0e-96cc-1da0debf72d0 Type: StorageAction group 
MANIPULATE_STORAGE_DOMAIN with role type ADMIN
2018-06-13 12:15:35,724+02 INFO  
[org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] 
(default task-12) [24485c23] START, SetVolumeDescriptionVDSCommand( 
SetVolumeDescriptionVDSCommandParameters:{storagePoolId='3edb5295-3797-4cd0-9b43-f46ec1ee7b14',
 ignoreFailoverLimit='false', 
storageDomainId='373efd46-8aea-4d0e-96cc-1da0debf72d0', 
imageGroupId='cb04b55c-10fb-46fe-b9de-3c133a94e6a5', 
imageId='a1e7554d-530c-4c07-a4b5-459a1c509e39'}), log id: 747d674f
2018-06-13 12:15:35,724+02 INFO  
[org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] 
(default task-12) [24485c23] -- executeIrsBrokerCommand: calling 
'setVolumeDescription', parameters:
2018-06-13 12:15:35,724+02 INFO  

Re: [ovirt-users] Cannot change Cluster Compatibility Version when a VM is active

2016-09-23 Thread Albl, Oliver
Michal,

  thank you for your quick response! I am running 300+ VMs so any other (safe) 
option would be very welcome…

All the best,
Simone

Von: users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] Im Auftrag von 
Michal Skrivanek
Gesendet: Freitag, 23. September 2016 16:57
An: Bruckner, Simone 
Cc: users@ovirt.org
Betreff: Re: [ovirt-users] Cannot change Cluster Compatibility Version when a 
VM is active


On 23 Sep 2016, at 16:49, Bruckner, Simone 
> wrote:

Hi all,

  I am trying to upgrade an oVirt installation (3.6.7.5-1.el6) to 4.0. My 
datacenters and clusters have 3.5 compatibility settings.

I followed the instructions from 
http://www.ovirt.org/documentation/migration-engine-3.6-to-4.0/ but cannot 
proceed in engine-setup as 3.5 compatibility is not supported.

When trying to change cluster compatibility from 3.5 to 3.6 I receive “Cannot 
change Cluster Compatibility Version when a VM is active. Please shutdown all 
VMs in the Cluster.” According 
tohttps://bugzilla.redhat.com/show_bug.cgi?id=1341023 this should be fixed in 
3.6.7. Any ideas?

this bug is blocking it, later bugs
(linked from there) allows it, though there are other issues…so if you have an 
option to shut them down please do so.
note those are RHEV bugs, not oVirt bugs so the exact build may differ

Thanks,
michal



Best Regards,
Simone Bruckner

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
Hi,

  I am starting with oVirt 3.3.2 and I have an issue adding a host to a cluster.

I am using oVirt Engine Version 3.3.2-1.el6
There is a cluster with one host (installed with oVirt Node - 3.0.3 - 1.1.fc19 
ISO image) up and running.
I installed a second host using the same ISO image.
I approved the host in the cluster.

When I try to activate the second host, I receive the following messages in the 
events pane:

State was set to Up for host host02.
Host host02 reports about one of the Active Storage Domains as Problematic.
Host host02 cannot access one of the Storage Domains attached to the Data 
Center Test303. Stetting Host state to Non-Operational.
Failed to connect Host host02 to Storage Pool Test303

There are 3 FC Storage Domains configured and visible to both hosts.
multipath -ll shows all LUNs on both hosts.

The engine.log reports the following about every five minutes:

2014-01-03 13:50:15,408 ERROR 
[org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] (pool-6-thread-44) 
Domain 7841a1c0-181a-4d43-9a25-b707accb5c4b: LUN_105 check timeot 69.7 is too 
big
2014-01-03 13:50:15,409 ERROR 
[org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] (pool-6-thread-44) 
Domain 52cf84ce-6eda-4337-8c94-491d94f5a18d: LUN_103 check timeot 59.6 is too 
big
2014-01-03 13:50:15,410 ERROR [org.ovirt.engine.core.bll.InitVdsOnUpCommand] 
(pool-6-thread-44) Storage Domain LUN_105 of pool Test303 is in problem in host 
host02
2014-01-03 13:50:15,411 ERROR [org.ovirt.engine.core.bll.InitVdsOnUpCommand] 
(pool-6-thread-44) Storage Domain LUN_103 of pool Test030 is in problem in host 
host02

Please let me know if there are any log files I should attach.

Thank you for your help!

All the best,
Oliver Albl

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
Dafna,

  you were right, it seems to be a caching issue. Rebooting the host did the 
job:

Before Reboot:

[root@host01 log]# vdsClient -s 0 getStorageDomainsList
52cf84ce-6eda-4337-8c94-491d94f5a18d
f404398a-97f9-474c-af2c-e8887f53f688
7841a1c0-181a-4d43-9a25-b707accb5c4b

[root@host02 log]# vdsClient -s 0 getStorageDomainsList
52cf84ce-6eda-4337-8c94-491d94f5a18d
f404398a-97f9-474c-af2c-e8887f53f688
7841a1c0-181a-4d43-9a25-b707accb5c4b
925ee53a-69b5-440f-b145-138ada5b452e

After Reboot:

[root@host02 admin]# vdsClient -s 0 getStorageDomainsList
52cf84ce-6eda-4337-8c94-491d94f5a18d
f404398a-97f9-474c-af2c-e8887f53f688
7841a1c0-181a-4d43-9a25-b707accb5c4b

So now I have both hosts up and running but when I try to start a VM on the 
second host, I receive the following messages in the events pane:

VM TEST2 was started by oliver.albl (Host: host02)
VM TEST2 is down. Exit message: Child quit during startup handshake: 
Input/output error.

Thanks again for your help!
Oliver

-Ursprüngliche Nachricht-
Von: Dafna Ron [mailto:d...@redhat.com] 
Gesendet: Freitag, 03. Jänner 2014 14:22
An: Albl, Oliver
Cc: users@ovirt.org
Betreff: Re: [Users] Host cannot access storage domains

yes, please attach the vdsm log
also, can you run vdsClient 0 getStorageDomainsList and vdsClient 0 
getDeviceList on both hosts?

It might be a cache issue, so can you please restart the host and if it helps 
attach output before and after the reboot?

Thanks,

Dafna


On 01/03/2014 01:12 PM, Albl, Oliver wrote:

 Hi,

 I am starting with oVirt 3.3.2 and I have an issue adding a host to a 
 cluster.

 I am using oVirt Engine Version 3.3.2-1.el6

 There is a cluster with one host (installed with oVirt Node - 3.0.3 -
 1.1.fc19 ISO image) up and running.

 I installed a second host using the same ISO image.

 I approved the host in the cluster.

 When I try to activate the second host, I receive the following 
 messages in the events pane:

 State was set to Up for host host02.

 Host host02 reports about one of the Active Storage Domains as 
 Problematic.

 Host host02 cannot access one of the Storage Domains attached to the 
 Data Center Test303. Stetting Host state to Non-Operational.

 Failed to connect Host host02 to Storage Pool Test303

 There are 3 FC Storage Domains configured and visible to both hosts.

 multipath -ll shows all LUNs on both hosts.

 The engine.log reports the following about every five minutes:

 2014-01-03 13:50:15,408 ERROR
 [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
 (pool-6-thread-44) Domain 7841a1c0-181a-4d43-9a25-b707accb5c4b: 
 LUN_105 check timeot 69.7 is too big

 2014-01-03 13:50:15,409 ERROR
 [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand]
 (pool-6-thread-44) Domain 52cf84ce-6eda-4337-8c94-491d94f5a18d: 
 LUN_103 check timeot 59.6 is too big

 2014-01-03 13:50:15,410 ERROR
 [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44) 
 Storage Domain LUN_105 of pool Test303 is in problem in host host02

 2014-01-03 13:50:15,411 ERROR
 [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44) 
 Storage Domain LUN_103 of pool Test030 is in problem in host host02

 Please let me know if there are any log files I should attach.

 Thank you for your help!

 All the best,

 Oliver Albl



 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


--
Dafna Ron
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
  
[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] 
(DefaultQuartzScheduler_Worker-7) [24696b3e] START, DestroyVDSCommand(HostName 
= host02, HostId = 6dc7fac6-149e-4445-ace1-3c334a24d52a, 
vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, force=false, secondsToWait=0, 
gracefully=false), log id: 6a95ffd5
2014-01-03 14:52:15,783 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] 
(DefaultQuartzScheduler_Worker-7) [24696b3e] FINISH, DestroyVDSCommand, log id: 
6a95ffd5
2014-01-03 14:52:15,804 INFO  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(DefaultQuartzScheduler_Worker-7) [24696b3e] Correlation ID: null, Call Stack: 
null, Custom Event ID: -1, Message: VM TEST2 is down. Exit message: Child quit 
during startup handshake: Input/output error.
2014-01-03 14:52:15,805 INFO  
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] 
(DefaultQuartzScheduler_Worker-7) [24696b3e] Running on vds during rerun failed 
vm: null
2014-01-03 14:52:15,805 INFO  
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] 
(DefaultQuartzScheduler_Worker-7) [24696b3e] vm TEST2 running in db and not 
running in vds - add to rerun treatment. vds host02
2014-01-03 14:52:15,808 ERROR 
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] 
(DefaultQuartzScheduler_Worker-7) [24696b3e] Rerun vm 
d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6. Called from vds host02
2014-01-03 14:52:15,810 INFO  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(pool-6-thread-40) [24696b3e] Correlation ID: 2ab5cd2, Job ID: 
2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event ID: -1, 
Message: Failed to run VM TEST2 on Host host02.
2014-01-03 14:52:15,823 INFO  
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] 
(pool-6-thread-40) [24696b3e] START, IsVmDuringInitiatingVDSCommand( vmId = 
d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 35e1eec
2014-01-03 14:52:15,824 INFO  
[org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] 
(pool-6-thread-40) [24696b3e] FINISH, IsVmDuringInitiatingVDSCommand, return: 
false, log id: 35e1eec
2014-01-03 14:52:15,858 WARN  [org.ovirt.engine.core.bll.RunVmOnceCommand] 
(pool-6-thread-40) [24696b3e] CanDoAction of action RunVmOnce failed. 
Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__VM,SCHEDULING_ALL_HOSTS_FILTERED_OUT
2014-01-03 14:52:15,862 INFO  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(pool-6-thread-40) [24696b3e] Correlation ID: 2ab5cd2, Job ID: 
2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event ID: -1, 
Message: Failed to run VM TEST2 (User: oliver.albl).

Thanks,
Oliver
-Ursprüngliche Nachricht-
Von: Dafna Ron [mailto:d...@redhat.com] 
Gesendet: Freitag, 03. Jänner 2014 14:51
An: Albl, Oliver
Cc: users@ovirt.org
Betreff: Re: AW: [Users] Host cannot access storage domains

Thanks for reporting the issue :)

As for the vm, can you please find the error in vdsm.log and in engine and 
paste it?

Thanks,

Dafna


On 01/03/2014 01:49 PM, Albl, Oliver wrote:
 Dafna,

you were right, it seems to be a caching issue. Rebooting the host did the 
 job:

 Before Reboot:

 [root@host01 log]# vdsClient -s 0 getStorageDomainsList 
 52cf84ce-6eda-4337-8c94-491d94f5a18d
 f404398a-97f9-474c-af2c-e8887f53f688
 7841a1c0-181a-4d43-9a25-b707accb5c4b

 [root@host02 log]# vdsClient -s 0 getStorageDomainsList 
 52cf84ce-6eda-4337-8c94-491d94f5a18d
 f404398a-97f9-474c-af2c-e8887f53f688
 7841a1c0-181a-4d43-9a25-b707accb5c4b
 925ee53a-69b5-440f-b145-138ada5b452e

 After Reboot:

 [root@host02 admin]# vdsClient -s 0 getStorageDomainsList 
 52cf84ce-6eda-4337-8c94-491d94f5a18d
 f404398a-97f9-474c-af2c-e8887f53f688
 7841a1c0-181a-4d43-9a25-b707accb5c4b

 So now I have both hosts up and running but when I try to start a VM on the 
 second host, I receive the following messages in the events pane:

 VM TEST2 was started by oliver.albl (Host: host02) VM TEST2 is down. 
 Exit message: Child quit during startup handshake: Input/output error.

 Thanks again for your help!
 Oliver

 -Ursprüngliche Nachricht-
 Von: Dafna Ron [mailto:d...@redhat.com]
 Gesendet: Freitag, 03. Jänner 2014 14:22
 An: Albl, Oliver
 Cc: users@ovirt.org
 Betreff: Re: [Users] Host cannot access storage domains

 yes, please attach the vdsm log
 also, can you run vdsClient 0 getStorageDomainsList and vdsClient 0 
 getDeviceList on both hosts?

 It might be a cache issue, so can you please restart the host and if it helps 
 attach output before and after the reboot?

 Thanks,

 Dafna


 On 01/03/2014 01:12 PM, Albl, Oliver wrote:
 Hi,

 I am starting with oVirt 3.3.2 and I have an issue adding a host to a 
 cluster.

 I am using oVirt Engine Version 3.3.2-1.el6

 There is a cluster with one host (installed with oVirt Node - 3.0.3 -
 1.1.fc19 ISO image) up and running.

 I installed a second host using the same ISO image.

 I approved the host in the cluster.

 When I try to activate the second

Re: [Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
Dafna,

  Libvirtd.log shows no errors, but VM log shows the following:

2014-01-03 13:52:11.296+: starting up
LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin 
QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name OATEST2 -S -machine 
pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m 1024 -realtime mlock=off -smp 
1,sockets=1,cores=1,threads=1 -uuid d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 
-smbios type=1,manufacturer=oVirt,product=oVirt 
Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C384C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6
 -no-user-config -nodefaults -chardev 
socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,server,nowait 
-mon chardev=charmonitor,id=monitor,mode=control -rtc 
base=2014-01-03T13:52:11,driftfix=slew -no-shutdown -device 
piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device 
virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device 
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive 
if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device 
ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive 
file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5c4b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-8387-ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca992-b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=native
 -device 
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
 -chardev 
socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait
 -device 
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=com.redhat.rhevm.vdsm
 -chardev 
socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait
 -device 
virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=channel1,name=org.qemu.guest_agent.0
 -chardev spicevmc,id=charchannel2,name=vdagent -device 
virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=channel2,name=com.redhat.spice.0
 -spice 
tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channel=main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-channel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usbredir,seamless-migration=on
 -k en-us -device 
qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,addr=0x2 
-device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
libvirt: Lock Driver error : unsupported configuration: Read/write, exclusive 
access, disks were present, but no leases specified
2014-01-03 13:52:11.306+: shutting down

Not sure what you mean with this 
http://forums.opensuse.org/english/get-technical-help-here/virtualization/492483-cannot-start-libvert-kvm-guests-after-update-tumbleweed.html.
 Do you want me to update libvirt with these repos on the oVirt-Node based 
installation?

Thanks,
Oliver
-Ursprüngliche Nachricht-
Von: Dafna Ron [mailto:d...@redhat.com] 
Gesendet: Freitag, 03. Jänner 2014 15:10
An: Albl, Oliver
Cc: users@ovirt.org
Betreff: Re: AW: AW: [Users] Host cannot access storage domains

actually, looking at this again, it's a libvirt error and it can be related to 
selinux or sasl.
can you also, look at libvirt log and the vm log under /var/log/libvirt?

On 01/03/2014 02:00 PM, Albl, Oliver wrote:
 Dafna,

please find the logs below:

 ERRORs in vdsm.log on host02:

 Thread-61::ERROR::2014-01-03 
 13:51:48,956::sdc::137::Storage.StorageDomainCache::(_findDomain) 
 looking for unfetched domain f404398a-97f9-474c-af2c-e8887f53f688
 Thread-61::ERROR::2014-01-03 
 13:51:48,959::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDom
 ain) looking for domain f404398a-97f9-474c-af2c-e8887f53f688
 Thread-323::ERROR::2014-01-03 
 13:52:11,527::vm::2132::vm.Vm::(_startUnderlyingVm) 
 vmId=`d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6`::The vm start process failed 
 Traceback (most recent call last):
File /usr/share/vdsm/vm.py, line 2092, in _startUnderlyingVm
  self._run()
File /usr/share/vdsm/vm.py, line 2959, in _run
  self._connection.createXML(domxml, flags),
File /usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py, line 
 76, in wrapper
  ret = f(*args, **kwargs)
File /usr/lib64/python2.7/site-packages/libvirt.py, line 2920, in 
 createXML
 libvirtError: Child quit during startup handshake: Input/output error
 Thread-60::ERROR::2014-01-03 
 13:52:23,111::sdc::137::Storage.StorageDomainCache::(_findDomain) 
 looking for unfetched domain 52cf84ce-6eda-4337-8c94-491d94f5a18d
 Thread-60::ERROR::2014-01-03 
 13:52:23,111::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDom
 ain) looking for domain 52cf84ce-6eda-4337-8c94-491d94f5a18d
 Thread-62::ERROR::2014-01-03 
 13:52:26,353::sdc::137::Storage.StorageDomainCache::(_findDomain) 
 looking for unfetched domain 7841a1c0-181a-4d43-9a25

Re: [Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
I installed both hosts using the oVirt Node ISO image:

OS Version: oVirt Node - 3.0.3 - 1.1.fc19
Kernel Version: 3.11.9 - 200.fc19.x86_64
KVM Version: 1.6.1 - 2.fc19
LIBVIRT Version: libvirt-1.1.3.1-2.fc19
VDSM Version: vdsm-4.13.0-11.fc19

Thanks,
Oliver
-Ursprüngliche Nachricht-
Von: Dafna Ron [mailto:d...@redhat.com] 
Gesendet: Freitag, 03. Jänner 2014 15:24
An: Albl, Oliver
Cc: users@ovirt.org
Betreff: Re: AW: AW: AW: [Users] Host cannot access storage domains

ignore the link :)

so searching for this error I hit an old bug and it seemed to be an issue 
between libvirt/sanlock.

https://bugzilla.redhat.com/show_bug.cgi?id=828633

are you using latest packages?




On 01/03/2014 02:15 PM, Albl, Oliver wrote:
 Dafna,

Libvirtd.log shows no errors, but VM log shows the following:

 2014-01-03 13:52:11.296+: starting up LC_ALL=C 
 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin 
 QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name OATEST2 -S -machine 
 pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m 1024 -realtime mlock=off 
 -smp 1,sockets=1,cores=1,threads=1 -uuid 
 d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -smbios 
 type=1,manufacturer=oVirt,product=oVirt 
 Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C384
 C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -no-user-config 
 -nodefaults -chardev 
 socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,serve
 r,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc 
 base=2014-01-03T13:52:11,driftfix=slew -no-shutdown -device 
 piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device 
 virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device 
 virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive 
 if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device 
 ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive 
 file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5c4
 b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-8387-
 ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca992-
 b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=nat
 ive -device 
 virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,id
 =virtio-disk0,bootindex=1 -chardev 
 socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a2
 c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait -device 
 virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=chann
 el0,name=com.redhat.rhevm.vdsm -chardev 
 socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a2
 c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait -device 
 virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=chann
 el1,name=org.qemu.guest_agent.0 -chardev 
 spicevmc,id=charchannel2,name=vdagent -device 
 virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=chann
 el2,name=com.redhat.spice.0 -spice 
 tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channel=
 main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-cha
 nnel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=usb
 redir,seamless-migration=on -k en-us -device 
 qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,addr=
 0x2 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
 libvirt: Lock Driver error : unsupported configuration: Read/write, 
 exclusive access, disks were present, but no leases specified
 2014-01-03 13:52:11.306+: shutting down

 Not sure what you mean with this 
 http://forums.opensuse.org/english/get-technical-help-here/virtualization/492483-cannot-start-libvert-kvm-guests-after-update-tumbleweed.html.
  Do you want me to update libvirt with these repos on the oVirt-Node based 
 installation?

 Thanks,
 Oliver
 -Ursprüngliche Nachricht-
 Von: Dafna Ron [mailto:d...@redhat.com]
 Gesendet: Freitag, 03. Jänner 2014 15:10
 An: Albl, Oliver
 Cc: users@ovirt.org
 Betreff: Re: AW: AW: [Users] Host cannot access storage domains

 actually, looking at this again, it's a libvirt error and it can be related 
 to selinux or sasl.
 can you also, look at libvirt log and the vm log under /var/log/libvirt?

 On 01/03/2014 02:00 PM, Albl, Oliver wrote:
 Dafna,

 please find the logs below:

 ERRORs in vdsm.log on host02:

 Thread-61::ERROR::2014-01-03
 13:51:48,956::sdc::137::Storage.StorageDomainCache::(_findDomain)
 looking for unfetched domain f404398a-97f9-474c-af2c-e8887f53f688
 Thread-61::ERROR::2014-01-03
 13:51:48,959::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDo
 m
 ain) looking for domain f404398a-97f9-474c-af2c-e8887f53f688
 Thread-323::ERROR::2014-01-03
 13:52:11,527::vm::2132::vm.Vm::(_startUnderlyingVm) 
 vmId=`d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6`::The vm start process failed 
 Traceback (most recent call last):
 File /usr/share/vdsm/vm.py, line 2092, in _startUnderlyingVm
   self._run()
 File /usr/share/vdsm/vm.py, line 2959, in _run
   self

Re: [Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
Dafna,

  yes, the VM starts on the first node, the issues are on the second node only.

/etc/libvirt/qemu-sanlock.conf is identical on on both nodes:

auto_disk_leases=0
require_lease_for_disks=0

yum updates reports Using yum is not supported...

Thanks,
Oliver

-Ursprüngliche Nachricht-
Von: Dafna Ron [mailto:d...@redhat.com] 
Gesendet: Freitag, 03. Jänner 2014 15:39
An: Albl, Oliver
Cc: users@ovirt.org
Betreff: Re: AW: AW: AW: AW: [Users] Host cannot access storage domains

ok, let's try to zoom in on the issue...
can you run vm's on the first host or do you have issues only on the second 
host you added?
can you run on both hosts?
# egrep -v ^# /etc/libvirt/qemu-sanlock.conf

can you run yum update on one of the hosts and see if there are newer packages?

Thanks,

Dafna

On 01/03/2014 02:30 PM, Albl, Oliver wrote:
 I installed both hosts using the oVirt Node ISO image:

 OS Version: oVirt Node - 3.0.3 - 1.1.fc19 Kernel Version: 3.11.9 - 
 200.fc19.x86_64 KVM Version: 1.6.1 - 2.fc19 LIBVIRT Version: 
 libvirt-1.1.3.1-2.fc19 VDSM Version: vdsm-4.13.0-11.fc19

 Thanks,
 Oliver
 -Ursprüngliche Nachricht-
 Von: Dafna Ron [mailto:d...@redhat.com]
 Gesendet: Freitag, 03. Jänner 2014 15:24
 An: Albl, Oliver
 Cc: users@ovirt.org
 Betreff: Re: AW: AW: AW: [Users] Host cannot access storage domains

 ignore the link :)

 so searching for this error I hit an old bug and it seemed to be an issue 
 between libvirt/sanlock.

 https://bugzilla.redhat.com/show_bug.cgi?id=828633

 are you using latest packages?




 On 01/03/2014 02:15 PM, Albl, Oliver wrote:
 Dafna,

 Libvirtd.log shows no errors, but VM log shows the following:

 2014-01-03 13:52:11.296+: starting up LC_ALL=C 
 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
 QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name OATEST2 -S -machine 
 pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m 1024 -realtime mlock=off 
 -smp 1,sockets=1,cores=1,threads=1 -uuid
 d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -smbios 
 type=1,manufacturer=oVirt,product=oVirt
 Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C38
 4
 C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -no-user-config 
 -nodefaults -chardev 
 socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,serv
 e r,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc 
 base=2014-01-03T13:52:11,driftfix=slew -no-shutdown -device
 piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
 virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
 virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive 
 if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device
 ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive
 file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5c
 4
 b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-8387
 -
 ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca992
 - 
 b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=na
 t
 ive -device
 virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,i
 d
 =virtio-disk0,bootindex=1 -chardev
 socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a
 2 c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait 
 -device 
 virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=chan
 n el0,name=com.redhat.rhevm.vdsm -chardev
 socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb-a
 2 c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait 
 -device 
 virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=chan
 n
 el1,name=org.qemu.guest_agent.0 -chardev 
 spicevmc,id=charchannel2,name=vdagent -device 
 virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=chan
 n
 el2,name=com.redhat.spice.0 -spice
 tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channel
 = 
 main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-ch
 a 
 nnel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=us
 b redir,seamless-migration=on -k en-us -device 
 qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,addr
 =
 0x2 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
 libvirt: Lock Driver error : unsupported configuration: Read/write, 
 exclusive access, disks were present, but no leases specified
 2014-01-03 13:52:11.306+: shutting down

 Not sure what you mean with this 
 http://forums.opensuse.org/english/get-technical-help-here/virtualization/492483-cannot-start-libvert-kvm-guests-after-update-tumbleweed.html.
  Do you want me to update libvirt with these repos on the oVirt-Node based 
 installation?

 Thanks,
 Oliver
 -Ursprüngliche Nachricht-
 Von: Dafna Ron [mailto:d...@redhat.com]
 Gesendet: Freitag, 03. Jänner 2014 15:10
 An: Albl, Oliver
 Cc: users@ovirt.org
 Betreff: Re: AW: AW: [Users] Host cannot access storage domains

 actually, looking at this again, it's a libvirt error and it can be related 
 to selinux

Re: [Users] Host cannot access storage domains

2014-01-03 Thread Albl, Oliver
Redirecting to /bin/systemctl reconfigure  vdsmd.service
Unknown operation 'reconfigure'.

... seems to me, I should get rid of the ovirt-node iso installation and move 
to a rpm based install?

Thanks,
Oliver
-Ursprüngliche Nachricht-
Von: Dafna Ron [mailto:d...@redhat.com] 
Gesendet: Freitag, 03. Jänner 2014 15:51
An: Albl, Oliver
Cc: users@ovirt.org
Betreff: Re: AW: AW: AW: AW: AW: [Users] Host cannot access storage domains

can you run:
service vdsmd reconfigure on the second host?

On 01/03/2014 02:43 PM, Albl, Oliver wrote:
 Dafna,

yes, the VM starts on the first node, the issues are on the second node 
 only.

 /etc/libvirt/qemu-sanlock.conf is identical on on both nodes:

 auto_disk_leases=0
 require_lease_for_disks=0

 yum updates reports Using yum is not supported...

 Thanks,
 Oliver

 -Ursprüngliche Nachricht-
 Von: Dafna Ron [mailto:d...@redhat.com]
 Gesendet: Freitag, 03. Jänner 2014 15:39
 An: Albl, Oliver
 Cc: users@ovirt.org
 Betreff: Re: AW: AW: AW: AW: [Users] Host cannot access storage 
 domains

 ok, let's try to zoom in on the issue...
 can you run vm's on the first host or do you have issues only on the second 
 host you added?
 can you run on both hosts?
 # egrep -v ^# /etc/libvirt/qemu-sanlock.conf

 can you run yum update on one of the hosts and see if there are newer 
 packages?

 Thanks,

 Dafna

 On 01/03/2014 02:30 PM, Albl, Oliver wrote:
 I installed both hosts using the oVirt Node ISO image:

 OS Version: oVirt Node - 3.0.3 - 1.1.fc19 Kernel Version: 3.11.9 -
 200.fc19.x86_64 KVM Version: 1.6.1 - 2.fc19 LIBVIRT Version:
 libvirt-1.1.3.1-2.fc19 VDSM Version: vdsm-4.13.0-11.fc19

 Thanks,
 Oliver
 -Ursprüngliche Nachricht-
 Von: Dafna Ron [mailto:d...@redhat.com]
 Gesendet: Freitag, 03. Jänner 2014 15:24
 An: Albl, Oliver
 Cc: users@ovirt.org
 Betreff: Re: AW: AW: AW: [Users] Host cannot access storage domains

 ignore the link :)

 so searching for this error I hit an old bug and it seemed to be an issue 
 between libvirt/sanlock.

 https://bugzilla.redhat.com/show_bug.cgi?id=828633

 are you using latest packages?




 On 01/03/2014 02:15 PM, Albl, Oliver wrote:
 Dafna,

  Libvirtd.log shows no errors, but VM log shows the following:

 2014-01-03 13:52:11.296+: starting up LC_ALL=C 
 PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin
 QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name OATEST2 -S -machine 
 pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m 1024 -realtime 
 mlock=off -smp 1,sockets=1,cores=1,threads=1 -uuid
 d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -smbios 
 type=1,manufacturer=oVirt,product=oVirt
 Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C3
 8
 4
 C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -no-user-config 
 -nodefaults -chardev 
 socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,ser
 v e r,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc 
 base=2014-01-03T13:52:11,driftfix=slew -no-shutdown -device
 piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device
 virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device
 virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive 
 if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device
 ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive 
 file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5
 c
 4
 b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-838
 7
 -
 ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca99
 2
 -
 b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=n
 a
 t
 ive -device
 virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0,
 i
 d
 =virtio-disk0,bootindex=1 -chardev
 socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb-
 a
 2 c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait
 -device
 virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=cha
 n n el0,name=com.redhat.rhevm.vdsm -chardev 
 socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb-
 a
 2 c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait
 -device
 virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=cha
 n
 n
 el1,name=org.qemu.guest_agent.0 -chardev 
 spicevmc,id=charchannel2,name=vdagent -device 
 virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=cha
 n
 n
 el2,name=com.redhat.spice.0 -spice
 tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channe
 l
 =
 main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-c
 h
 a
 nnel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=u
 s b redir,seamless-migration=on -k en-us -device 
 qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,add
 r
 =
 0x2 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
 libvirt: Lock Driver error : unsupported configuration: Read/write, 
 exclusive access, disks were present, but no leases specified
 2014-01-03 13:52:11.306+: shutting down

Re: [Users] all hosts are Non responsive

2014-01-03 Thread Albl, Oliver
Zach,

  I had a similar problem. I temporarily disabled SELinux on the hosts using 
setenforce 0. Let me know if this helps.

All the best,
Oliver

Von: users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] Im Auftrag von 
Zach Underwood
Gesendet: Freitag, 03. Jänner 2014 19:32
An: users@ovirt.org
Betreff: [Users] all hosts are Non responsive

When I try to activate a host to the engine I get the error non responsive. I 
can ping the hosts  but I cant connect with ssh.

oVirt Engine Version: 3.3.2-1.el6

ovirt-node-iso-3.0.3-1.1.vdsm.fc19




--
Zach Underwood (RHCE,RHCSA,RHCT,UACA)
My websitehttp://zachunderwood.me
My photoshttp://zunder1990.openphoto.me
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users