[ovirt-users] Re: Local storage formatting

2018-09-04 Thread Matt Simonsen

On 09/04/2018 02:22 PM, Nir Soffer wrote:
Maybe you have lvm filter set, which is highly recommend for an oVirt 
hypervisor.



Indeed, I do.   I am not sure I have the right filter however, so I 
appreciate the help.


This is the filter setup initially:

  filter = [ "a|^/dev/mapper/3600508b1001c7e172160824d7b204c3b2$|", 
"r|.*|" ]


Just to be clear, my intent isn't to add /dev/sdb to the main volume 
group, but to make a new volume group to setup a local ext4 mount point.



I changed it to:

  filter = [ "a|^/dev/sdb|", 
"a|^/dev/mapper/3600508b1001c7e172160824d7b204c3b2$|", "r|.*|" ]


Following this and a reboot, I was able to create a PV, VG, and LV.

# pvcreate /dev/sdb
# vgcreate data /dev/sdb
# lvcreate -L800g /dev/data --name local_images
# mkfs.ext4 /dev/data/local_images
-- adjust fstab
# mount -a

It seems to function as expected now the the filter has been adjusted.  
But is the filter doing what it is "supposed" to?


When I run the command "vdsm-tool config-lvm-filter" what I see is:

[root@node4-g8-h4 ~]# vdsm-tool config-lvm-filter
Analyzing host...
LVM filter is already configured for Vdsm

Thanks for the help and confirming how this should work.

Matt




To add /dev/sdb, you need to add it to the lvm filter in 
/etc/lvm/lvm.conf.


After you configure the device properly, you can generate lvm filter
for the current setup using:

    vdsm-tool config-lvm-filter

Here is example run on unconfigued oVirt host:

#  vdsm-tool config-lvm-filter
Analyzing host...
Found these mounted logical volumes on this host:

  logical volume:  /dev/mapper/fedora_voodoo1-root
  mountpoint:      /
  devices:         /dev/vda2

  logical volume:  /dev/mapper/fedora_voodoo1-swap
  mountpoint:      [SWAP]
  devices:         /dev/vda2

This is the recommended LVM filter for this host:

  filter = [ "a|^/dev/vda2$|", "r|.*|" ]

This filter allows LVM to access the local devices used by the
hypervisor, but not shared storage owned by Vdsm. If you add a new
device to the volume group, you will need to edit the filter manually.


Nir
On 09/04/2018 01:23 PM, Matt Simonsen wrote:

> Hello,
>
> I'm running oVirt with several data centers, some with NFS
storage and
> some with local storage.
>
> I had problems in the past with a large pool and local storage. The
> problem was nodectl showed the pool being too full (I think
>80%), but
> it was only the images that made the pool "full" -- and this
storage
> was carefully setup such that there was no chance it would actually
> fill.  The LVs for oVirt itself were all under 20%, yet nodectl
still
> reported the pool was too full.
>
> My solution so far has been to use our RAID card tools, so that
sda is
> the oVirt node install, and sdb is for images.  There are probably
> other good reasons for me to handle it this way, for example being
> able to use different RAID levels, but I'm hoping someone can
confirm
> my partitioning below doesn't have some risk I'm now yet aware of.
>
> I setup a new volume group for images, as below:
>
>
> [root@node4-g8-h4 multipath]# pvs
>   PV VG Fmt  Attr PSize
> PFree
>   /dev/mapper/3600508b1001c7e172160824d7b204c3b2 onn_node4-g8-h4
lvm2
> a--  <119.00g  <22.85g
>   /dev/sdb1  data lvm2 a--
1.13t
> <361.30g
>
> [root@node4-g8-h4 multipath]# vgs
>   VG  #PV #LV #SN Attr   VSize    VFree
>   data  1   1   0 wz--n-    1.13t <361.30g
>   onn_node4-g8-h4   1  13   0 wz--n- <119.00g <22.85g
>
> [root@node4-g8-h4 multipath]# lvs
>   LV   VG Attr LSize
> Pool   Origin Data%  Meta% Move Log
> Cpy%Sync Convert
>   images_main  data -wi-ao 800.00g
>   home onn_node4-g8-h4 Vwi-aotz--
> 1.00g pool00 4.79
>   ovirt-node-ng-4.2.5.1-0.20180816.0   onn_node4-g8-h4 Vwi---tz-k
> 64.10g pool00 root
>   ovirt-node-ng-4.2.5.1-0.20180816.0+1 onn_node4-g8-h4 Vwi---tz--
> 64.10g pool00 ovirt-node-ng-4.2.5.1-0.20180816.0
>   ovirt-node-ng-4.2.6-0.20180903.0 onn_node4-g8-h4 Vri---tz-k
> 64.10g pool00
>   ovirt-node-ng-4.2.6-0.20180903.0+1   onn_node4-g8-h4 Vwi-aotz--
> 64.10g pool00 ovirt-node-ng-4.2.6-0.20180903.0 4.83
>   pool00   onn_node4-g8-h4 twi-aotz--
> 91.10g   8.94 0.49
>   root onn_node4-g8-h4 Vwi---tz--
> 64.10g pool00
>   swap onn_node4-g8-h4
-wi-ao 4.00g
>   tmp  onn_node4-g8-h4 Vwi-aotz--
> 1.00g pool00 4.87
>   var  onn_node4-g8-h4 Vwi-aotz--
> 15.00g pool00 3.31

[ovirt-users] Re: Local storage formatting

2018-09-04 Thread Nir Soffer
On Wed, Sep 5, 2018 at 12:01 AM Matt Simonsen  wrote:

> Hello all,
>
> Following this report below, I did a reboot. Now I have a real question.
>
> I added the VG, LV and mount point to this node using the port 9090 web
> interface.
>
> Now the volume group isn't active and will not mount, causing the boot
> to hang.
>
> I am able to do "vgchange -ay data" and then a manual mount in rescue mode.
>
> Any feedback on the best way to add a new volume group to an empty
> partition (sdb) would be appreciated. Prior to using the web interface,
> I was having failures using the manual tools to /dev/sdb with an error
> "device /dev/sdb excluded by filter" which I suspect is related.
>

Maybe you have lvm filter set, which is highly recommend for an oVirt
hypervisor.

To add /dev/sdb, you need to add it to the lvm filter in /etc/lvm/lvm.conf.

After you configure the device properly, you can generate lvm filter
for the current setup using:

vdsm-tool config-lvm-filter

Here is example run on unconfigued oVirt host:

#  vdsm-tool config-lvm-filter
Analyzing host...
Found these mounted logical volumes on this host:

  logical volume:  /dev/mapper/fedora_voodoo1-root
  mountpoint:  /
  devices: /dev/vda2

  logical volume:  /dev/mapper/fedora_voodoo1-swap
  mountpoint:  [SWAP]
  devices: /dev/vda2

This is the recommended LVM filter for this host:

  filter = [ "a|^/dev/vda2$|", "r|.*|" ]

This filter allows LVM to access the local devices used by the
hypervisor, but not shared storage owned by Vdsm. If you add a new
device to the volume group, you will need to edit the filter manually.


Nir

On 09/04/2018 01:23 PM, Matt Simonsen wrote:

> > Hello,
> >
> > I'm running oVirt with several data centers, some with NFS storage and
> > some with local storage.
> >
> > I had problems in the past with a large pool and local storage. The
> > problem was nodectl showed the pool being too full (I think >80%), but
> > it was only the images that made the pool "full" -- and this storage
> > was carefully setup such that there was no chance it would actually
> > fill.  The LVs for oVirt itself were all under 20%, yet nodectl still
> > reported the pool was too full.
> >
> > My solution so far has been to use our RAID card tools, so that sda is
> > the oVirt node install, and sdb is for images.  There are probably
> > other good reasons for me to handle it this way, for example being
> > able to use different RAID levels, but I'm hoping someone can confirm
> > my partitioning below doesn't have some risk I'm now yet aware of.
> >
> > I setup a new volume group for images, as below:
> >
> >
> > [root@node4-g8-h4 multipath]# pvs
> >   PV VG Fmt  Attr PSize
> > PFree
> >   /dev/mapper/3600508b1001c7e172160824d7b204c3b2 onn_node4-g8-h4 lvm2
> > a--  <119.00g  <22.85g
> >   /dev/sdb1  data lvm2 a-- 1.13t
> > <361.30g
> >
> > [root@node4-g8-h4 multipath]# vgs
> >   VG  #PV #LV #SN Attr   VSizeVFree
> >   data  1   1   0 wz--n-1.13t <361.30g
> >   onn_node4-g8-h4   1  13   0 wz--n- <119.00g  <22.85g
> >
> > [root@node4-g8-h4 multipath]# lvs
> >   LV   VG  Attr LSize
> > Pool   Origin Data%  Meta% Move Log
> > Cpy%Sync Convert
> >   images_main  data-wi-ao 800.00g
> >   home onn_node4-g8-h4 Vwi-aotz--
> > 1.00g pool00 4.79
> >   ovirt-node-ng-4.2.5.1-0.20180816.0   onn_node4-g8-h4 Vwi---tz-k
> > 64.10g pool00 root
> >   ovirt-node-ng-4.2.5.1-0.20180816.0+1 onn_node4-g8-h4 Vwi---tz--
> > 64.10g pool00 ovirt-node-ng-4.2.5.1-0.20180816.0
> >   ovirt-node-ng-4.2.6-0.20180903.0 onn_node4-g8-h4 Vri---tz-k
> > 64.10g pool00
> >   ovirt-node-ng-4.2.6-0.20180903.0+1   onn_node4-g8-h4 Vwi-aotz--
> > 64.10g pool00 ovirt-node-ng-4.2.6-0.20180903.0 4.83
> >   pool00   onn_node4-g8-h4 twi-aotz--
> > 91.10g   8.94 0.49
> >   root onn_node4-g8-h4 Vwi---tz--
> > 64.10g pool00
> >   swap onn_node4-g8-h4 -wi-ao 4.00g
> >   tmp  onn_node4-g8-h4 Vwi-aotz--
> > 1.00g pool00 4.87
> >   var  onn_node4-g8-h4 Vwi-aotz--
> > 15.00g pool00 3.31
> >   var_crashonn_node4-g8-h4 Vwi-aotz--
> > 10.00g pool00 2.86
> >   var_log  onn_node4-g8-h4 Vwi-aotz--
> > 8.00g pool00 3.57
> >   var_log_auditonn_node4-g8-h4 Vwi-aotz--
> > 2.00g pool004.89
> >
> >
> >
> > The images_main is setup as "Block device for filesystems" with ext4.
> > Is there any reason I should consider pool for thinly provisioned
> > volumes?  I don't need to over-allocate storage and it seems to me
> > like a fix

[ovirt-users] Re: Local storage formatting

2018-09-04 Thread Matt Simonsen

Hello all,

Following this report below, I did a reboot. Now I have a real question.

I added the VG, LV and mount point to this node using the port 9090 web 
interface.


Now the volume group isn't active and will not mount, causing the boot 
to hang.


I am able to do "vgchange -ay data" and then a manual mount in rescue mode.

Any feedback on the best way to add a new volume group to an empty 
partition (sdb) would be appreciated. Prior to using the web interface, 
I was having failures using the manual tools to /dev/sdb with an error 
"device /dev/sdb excluded by filter" which I suspect is related.


Thanks

Matt





On 09/04/2018 01:23 PM, Matt Simonsen wrote:

Hello,

I'm running oVirt with several data centers, some with NFS storage and 
some with local storage.


I had problems in the past with a large pool and local storage. The 
problem was nodectl showed the pool being too full (I think >80%), but 
it was only the images that made the pool "full" -- and this storage 
was carefully setup such that there was no chance it would actually 
fill.  The LVs for oVirt itself were all under 20%, yet nodectl still 
reported the pool was too full.


My solution so far has been to use our RAID card tools, so that sda is 
the oVirt node install, and sdb is for images.  There are probably 
other good reasons for me to handle it this way, for example being 
able to use different RAID levels, but I'm hoping someone can confirm 
my partitioning below doesn't have some risk I'm now yet aware of.


I setup a new volume group for images, as below:


[root@node4-g8-h4 multipath]# pvs
  PV VG Fmt  Attr PSize    
PFree
  /dev/mapper/3600508b1001c7e172160824d7b204c3b2 onn_node4-g8-h4 lvm2 
a--  <119.00g  <22.85g
  /dev/sdb1  data lvm2 a-- 1.13t 
<361.30g


[root@node4-g8-h4 multipath]# vgs
  VG  #PV #LV #SN Attr   VSize    VFree
  data  1   1   0 wz--n-    1.13t <361.30g
  onn_node4-g8-h4   1  13   0 wz--n- <119.00g  <22.85g

[root@node4-g8-h4 multipath]# lvs
  LV   VG  Attr LSize   
Pool   Origin Data%  Meta% Move Log 
Cpy%Sync Convert

  images_main  data    -wi-ao 800.00g
  home onn_node4-g8-h4 Vwi-aotz--   
1.00g pool00 4.79
  ovirt-node-ng-4.2.5.1-0.20180816.0   onn_node4-g8-h4 Vwi---tz-k 
64.10g pool00 root
  ovirt-node-ng-4.2.5.1-0.20180816.0+1 onn_node4-g8-h4 Vwi---tz-- 
64.10g pool00 ovirt-node-ng-4.2.5.1-0.20180816.0
  ovirt-node-ng-4.2.6-0.20180903.0 onn_node4-g8-h4 Vri---tz-k 
64.10g pool00
  ovirt-node-ng-4.2.6-0.20180903.0+1   onn_node4-g8-h4 Vwi-aotz-- 
64.10g pool00 ovirt-node-ng-4.2.6-0.20180903.0 4.83
  pool00   onn_node4-g8-h4 twi-aotz-- 
91.10g   8.94 0.49
  root onn_node4-g8-h4 Vwi---tz-- 
64.10g pool00

  swap onn_node4-g8-h4 -wi-ao 4.00g
  tmp  onn_node4-g8-h4 Vwi-aotz--   
1.00g pool00 4.87
  var  onn_node4-g8-h4 Vwi-aotz-- 
15.00g pool00 3.31
  var_crash    onn_node4-g8-h4 Vwi-aotz-- 
10.00g pool00 2.86
  var_log  onn_node4-g8-h4 Vwi-aotz--   
8.00g pool00 3.57
  var_log_audit    onn_node4-g8-h4 Vwi-aotz--   
2.00g pool00    4.89




The images_main is setup as "Block device for filesystems" with ext4. 
Is there any reason I should consider pool for thinly provisioned 
volumes?  I don't need to over-allocate storage and it seems to me 
like a fixed partition is ideal. Please confirm or let me know if 
there's anything else I should consider.



Thanks

Matt
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/I7N547X6DC7KHHVCDGKXQGNJV6TG7E3U/

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LJINANK6PAGVV22H5OTYTJ3M4WIWPTMV/


[ovirt-users] Re: problem with multipath.conf

2018-09-04 Thread Nir Soffer
On Tue, Sep 4, 2018 at 11:12 PM Nir Soffer  wrote:

> On Tue, Sep 4, 2018 at 11:03 PM  wrote:
>
>> after upgrading to 4.20.39-1.el7 sas multipath stopped being detected
>
>
>> I did a diff to the two files
>>
>> Is the current behaviour the correct one or the previous one ?
>> I think sas multipath should also be detected no ?
>>
>
> It should be detect it this is shared storage that you want to use in oVirt
> It should not be detect if you have local disk using sas.
>
> So we really cannot know.
>
> [root@g1-car0136 etc]# diff multipath.conf multipath.conf.201809031555
>> 1c1
>> < # VDSM REVISION 1.6
>> ---
>> > # VDSM REVISION 1.5
>> 101,109d100
>> < }
>> <
>> < # Whitelist FCP and iSCSI devices.
>> < blacklist {
>> < protocol ".*"
>> < }
>> <
>> < blacklist_exceptions {
>> < protocol "(scsi:fcp|scsi:iscsi)"
>>
>
> oVirt does not officially support sas, but this should work:
>
> protocol "(scsi:fcp|scsi:iscsi|scsi:sas)"
>
> The best way to configure it is to add a dropin configuration file:
>
> $ cat /etc/multipath/conf.d/my.conf
>
> blacklist_exceptions {
> protocol "(scsi:fcp|scsi:iscsi|scsi:sas)"
> }
>

Actually, to enable sas support, you can do this:

# cat /etc/multipath/conf.d/enable-sas.conf
blacklist_exceptions {
protocol "scsi:sas"
}


>
> See also this thread:
>
> https://lists.ovirt.org/archives/list/users@ovirt.org/thread/RFEQ5JK2RZM3Q7U3RDARIV7ZPDMHSPW2/
>
> I suggest you file a bug for this here:
> https://bugzilla.redhat.com/enter_bug.cgi?product=vdsm
>
> Nir
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3EKG7OOGVCWK2WMJF5FD4BICKY3DFWNP/


[ovirt-users] Local storage formatting

2018-09-04 Thread Matt Simonsen

Hello,

I'm running oVirt with several data centers, some with NFS storage and 
some with local storage.


I had problems in the past with a large pool and local storage. The 
problem was nodectl showed the pool being too full (I think >80%), but 
it was only the images that made the pool "full" -- and this storage was 
carefully setup such that there was no chance it would actually fill.  
The LVs for oVirt itself were all under 20%, yet nodectl still reported 
the pool was too full.


My solution so far has been to use our RAID card tools, so that sda is 
the oVirt node install, and sdb is for images.  There are probably other 
good reasons for me to handle it this way, for example being able to use 
different RAID levels, but I'm hoping someone can confirm my 
partitioning below doesn't have some risk I'm now yet aware of.


I setup a new volume group for images, as below:


[root@node4-g8-h4 multipath]# pvs
  PV VG Fmt  Attr PSize    
PFree
  /dev/mapper/3600508b1001c7e172160824d7b204c3b2 onn_node4-g8-h4 lvm2 
a--  <119.00g  <22.85g
  /dev/sdb1  data lvm2 a-- 
1.13t <361.30g


[root@node4-g8-h4 multipath]# vgs
  VG  #PV #LV #SN Attr   VSize    VFree
  data  1   1   0 wz--n-    1.13t <361.30g
  onn_node4-g8-h4   1  13   0 wz--n- <119.00g  <22.85g

[root@node4-g8-h4 multipath]# lvs
  LV   VG  Attr LSize   
Pool   Origin Data%  Meta% Move Log Cpy%Sync 
Convert

  images_main  data    -wi-ao 800.00g
  home onn_node4-g8-h4 Vwi-aotz--   
1.00g pool00 4.79
  ovirt-node-ng-4.2.5.1-0.20180816.0   onn_node4-g8-h4 Vwi---tz-k 
64.10g pool00 root
  ovirt-node-ng-4.2.5.1-0.20180816.0+1 onn_node4-g8-h4 Vwi---tz-- 
64.10g pool00 ovirt-node-ng-4.2.5.1-0.20180816.0
  ovirt-node-ng-4.2.6-0.20180903.0 onn_node4-g8-h4 Vri---tz-k 
64.10g pool00
  ovirt-node-ng-4.2.6-0.20180903.0+1   onn_node4-g8-h4 Vwi-aotz-- 
64.10g pool00 ovirt-node-ng-4.2.6-0.20180903.0 4.83
  pool00   onn_node4-g8-h4 twi-aotz-- 
91.10g   8.94 0.49
  root onn_node4-g8-h4 Vwi---tz-- 
64.10g pool00

  swap onn_node4-g8-h4 -wi-ao 4.00g
  tmp  onn_node4-g8-h4 Vwi-aotz--   
1.00g pool00 4.87
  var  onn_node4-g8-h4 Vwi-aotz-- 
15.00g pool00 3.31
  var_crash    onn_node4-g8-h4 Vwi-aotz-- 
10.00g pool00 2.86
  var_log  onn_node4-g8-h4 Vwi-aotz--   
8.00g pool00 3.57
  var_log_audit    onn_node4-g8-h4 Vwi-aotz--   
2.00g pool00    4.89




The images_main is setup as "Block device for filesystems" with ext4. Is 
there any reason I should consider pool for thinly provisioned volumes?  
I don't need to over-allocate storage and it seems to me like a fixed 
partition is ideal. Please confirm or let me know if there's anything 
else I should consider.



Thanks

Matt
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/I7N547X6DC7KHHVCDGKXQGNJV6TG7E3U/


[ovirt-users] Re: problem with multipath.conf

2018-09-04 Thread Nir Soffer
On Tue, Sep 4, 2018 at 11:03 PM  wrote:

> after upgrading to 4.20.39-1.el7 sas multipath stopped being detected


> I did a diff to the two files
>
> Is the current behaviour the correct one or the previous one ?
> I think sas multipath should also be detected no ?
>

It should be detect it this is shared storage that you want to use in oVirt
It should not be detect if you have local disk using sas.

So we really cannot know.

[root@g1-car0136 etc]# diff multipath.conf multipath.conf.201809031555
> 1c1
> < # VDSM REVISION 1.6
> ---
> > # VDSM REVISION 1.5
> 101,109d100
> < }
> <
> < # Whitelist FCP and iSCSI devices.
> < blacklist {
> < protocol ".*"
> < }
> <
> < blacklist_exceptions {
> < protocol "(scsi:fcp|scsi:iscsi)"
>

oVirt does not officially support sas, but this should work:

protocol "(scsi:fcp|scsi:iscsi|scsi:sas)"

The best way to configure it is to add a dropin configuration file:

$ cat /etc/multipath/conf.d/my.conf

blacklist_exceptions {
protocol "(scsi:fcp|scsi:iscsi|scsi:sas)"
}

See also this thread:
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/RFEQ5JK2RZM3Q7U3RDARIV7ZPDMHSPW2/

I suggest you file a bug for this here:
https://bugzilla.redhat.com/enter_bug.cgi?product=vdsm

Nir
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QGACLGVCVRK66XT2DB3PMXZ22EVZNF7K/


[ovirt-users] problem with multipath.conf

2018-09-04 Thread g . vasilopoulos
after upgrading to 4.20.39-1.el7 sas multipath stopped being detected

I did a diff to the two files

Is the current behaviour the correct one or the previous one ?
I think sas multipath should also be detected no ?

[root@g1-car0136 etc]# diff multipath.conf multipath.conf.201809031555 
1c1
< # VDSM REVISION 1.6
---
> # VDSM REVISION 1.5
101,109d100
< }
< 
< # Whitelist FCP and iSCSI devices.
< blacklist {
< protocol ".*"
< }
< 
< blacklist_exceptions {
< protocol "(scsi:fcp|scsi:iscsi)"
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ICS4JWPSTGNHI23TJFNHKK5GUBGAY4LD/


[ovirt-users] Re: [ANN] oVirt 4.2.6 is now generally available

2018-09-04 Thread Nir Soffer
On Tue, Sep 4, 2018 at 9:51 PM Nir Soffer  wrote:

> On Tue, Sep 4, 2018 at 11:30 AM Fabrice Bacchella <
> fabrice.bacche...@orange.fr> wrote:
>
>>
>>
>> > Le 3 sept. 2018 à 19:15, Nir Soffer  a écrit :
>>
>> Thank you for you help, but I'm still not out of trouble.
>>
>> >
>> > On Mon, Sep 3, 2018 at 8:01 PM Fabrice Bacchella <
>> fabrice.bacche...@orange.fr> wrote:
>> >
>> >> Le 3 sept. 2018 à 18:31, Nir Soffer  a écrit :
>> >>
>> >> On Mon, Sep 3, 2018 at 5:07 PM Fabrice Bacchella <
>> fabrice.bacche...@orange.fr> wrote:
>> >> In the release notes, I see:
>> >>
>> >> • BZ 1622700 [downstream clone - 4.2.6] [RFE][Dalton] - Blacklist all
>> local disk in multipath on RHEL / RHEV Host (RHEL 7.5)
>> >> Feature:
>> >> Blacklist local devices in multipath.
>> >>
>> >> Reason:
>> >> multipath repeatedly logs irrelevant errors for local devices.
>> >>
>> >> Result:
>> >> Local devices are blacklisted, and no irrelevant errors are logged
>> anymore.
>> >>
>> >> What defines a local disk ? I'm using a SAN on SAS. For many peoples,
>> SAS is only for local disks, but that's not the case. Will other 4.2.6 will
>> detect that ?
>> >>
>> >> We don't have any support for SAS.
>> >>
>> >> If you SAS drives are attached to the host using FC or iSCSI, you are
>> fine.
>> >
>> > Nope, they are attached using SAS.
>> >
>> > I guess oVirt see them as FCP devices?
>>
>> yes, in ovirt UI, I've configured my storage to be on FCP, and everything
>> worked well since 3.6.
>>
>> >
>> > Are these disks connected to multiple hosts?
>>
>> Yes, that's a real SAN, multi-attached to HPE's blades
>> >
>> > Please share the output of:
>> >
>> > vdsm-client Host getDeviceList
>>
>> Things are strange:
>>
>> {
>> "status": "used",
>> "vendorID": "HP iLO",
>> "GUID": "HP_iLO_LUN_01_Media_0_02660A01-0:1",
>> "capacity": "1073741824",
>> "fwrev": "2.10",
>> "discard_zeroes_data": 0,
>> "vgUUID": "",
>> "pathlist": [],
>> "pvsize": "",
>> "discard_max_bytes": 0,
>> "pathstatus": [
>> {
>> "capacity": "1073741824",
>> "physdev": "sddj",
>> "type": "FCP",
>> "state": "active",
>> "lun": "1"
>> }
>> ],
>> "devtype": "FCP",
>> "physicalblocksize": "512",
>> "pvUUID": "",
>> "serial": "",
>> "logicalblocksize": "512",
>> "productID": "LUN 01 Media 0"
>> },
>> ...
>> {
>> "status": "used",
>> "vendorID": "HP",
>> "GUID": "3600c0ff0002631c42168f1560100",
>> "capacity": "1198996324352",
>> "fwrev": "G22x",
>> "discard_zeroes_data": 0,
>> "vgUUID": "xGCmpC-DhHe-3v6v-6LJw-iS24-ExCE-0Hv48U",
>> "pathlist": [],
>> "pvsize": "1198698528768",
>> "discard_max_bytes": 0,
>> "pathstatus": [
>> {
>> "capacity": "1198996324352",
>> "physdev": "sdc",
>> "type": "FCP",
>> "state": "active",
>> "lun": "16"
>> },
>> {
>> "capacity": "1198996324352",
>> "physdev": "sds",
>> "type": "FCP",
>> "state": "active",
>> "lun": "16"
>> },
>>
>>
>> ...
>>
>> The first one is an embedded flash drive:
>> lrwxrwxrwx 1 root root 10 Jul 12 17:11
>> /dev/disk/by-id/usb-HP_iLO_LUN_01_Media_0_02660A01-0:1 -> ../../sddj
>> lrwxrwxrwx 1 root root 10 Jul 12 17:11
>> /dev/disk/by-path/pci-:00:14.0-usb-0:3.1:1.0-scsi-0:0:0:1 -> ../../sddj
>>
>> So why "type": "FCP",  ?
>>
>
> "FCP" actually means "not iSCSI". This why your sas storage works while
> oVirt does
> know anything about sas.
>
> This is why the blacklist by protocol feature was introduced in 7.5, to
> multipath can grab
> only shared storage, and avoid grabbing local devices like your SSD.
> See https://bugzilla.redhat.com/show_bug.cgi?id=1593459
>
> According to this bug:
> https://bugzilla.redhat.com/show_bug.cgi?id=1607749
>
> The fix is available in:
> device-mapper-multipath-0.4.9-119.el7_5.1.x86_64
>
> Which device-mapper-multipath package are you using?
>
>
>
>> The second is indeed a SAS drives behind a SAS SAN (a MSA 2040 SAS from
>> HPE).
>>
>>
>> >  ...
>> > Where do I find the protocol multipath thinks the drives are using ?
>> >
>> > multipath.conf(5) says:
>> >
>> >The protocol strings that multipath recognizes are scsi:fcp,
>> scsi:spi, scsi:ssa, scsi:sbp,
>> >scsi:srp, scsi:iscsi, scsi:sas, scsi:adt, scsi:ata, scsi:unspec,
>> ccw, cciss, nvme,  and
>> >undef.  The protocol that a path is using can be viewed by
>> running multipathd show
>> >paths format "%d %P"
>>
>> I have a centos 7.5:
>>
>> lsb_release -a
>> LSB Version::core-4.1-amd64:core-4.1-noarch
>> Distributor ID: CentOS
>> Description:CentOS Linux 

[ovirt-users] Re: [ANN] oVirt 4.2.6 is now generally available

2018-09-04 Thread Nir Soffer
On Tue, Sep 4, 2018 at 11:30 AM Fabrice Bacchella <
fabrice.bacche...@orange.fr> wrote:

>
>
> > Le 3 sept. 2018 à 19:15, Nir Soffer  a écrit :
>
> Thank you for you help, but I'm still not out of trouble.
>
> >
> > On Mon, Sep 3, 2018 at 8:01 PM Fabrice Bacchella <
> fabrice.bacche...@orange.fr> wrote:
> >
> >> Le 3 sept. 2018 à 18:31, Nir Soffer  a écrit :
> >>
> >> On Mon, Sep 3, 2018 at 5:07 PM Fabrice Bacchella <
> fabrice.bacche...@orange.fr> wrote:
> >> In the release notes, I see:
> >>
> >> • BZ 1622700 [downstream clone - 4.2.6] [RFE][Dalton] - Blacklist all
> local disk in multipath on RHEL / RHEV Host (RHEL 7.5)
> >> Feature:
> >> Blacklist local devices in multipath.
> >>
> >> Reason:
> >> multipath repeatedly logs irrelevant errors for local devices.
> >>
> >> Result:
> >> Local devices are blacklisted, and no irrelevant errors are logged
> anymore.
> >>
> >> What defines a local disk ? I'm using a SAN on SAS. For many peoples,
> SAS is only for local disks, but that's not the case. Will other 4.2.6 will
> detect that ?
> >>
> >> We don't have any support for SAS.
> >>
> >> If you SAS drives are attached to the host using FC or iSCSI, you are
> fine.
> >
> > Nope, they are attached using SAS.
> >
> > I guess oVirt see them as FCP devices?
>
> yes, in ovirt UI, I've configured my storage to be on FCP, and everything
> worked well since 3.6.
>
> >
> > Are these disks connected to multiple hosts?
>
> Yes, that's a real SAN, multi-attached to HPE's blades
> >
> > Please share the output of:
> >
> > vdsm-client Host getDeviceList
>
> Things are strange:
>
> {
> "status": "used",
> "vendorID": "HP iLO",
> "GUID": "HP_iLO_LUN_01_Media_0_02660A01-0:1",
> "capacity": "1073741824",
> "fwrev": "2.10",
> "discard_zeroes_data": 0,
> "vgUUID": "",
> "pathlist": [],
> "pvsize": "",
> "discard_max_bytes": 0,
> "pathstatus": [
> {
> "capacity": "1073741824",
> "physdev": "sddj",
> "type": "FCP",
> "state": "active",
> "lun": "1"
> }
> ],
> "devtype": "FCP",
> "physicalblocksize": "512",
> "pvUUID": "",
> "serial": "",
> "logicalblocksize": "512",
> "productID": "LUN 01 Media 0"
> },
> ...
> {
> "status": "used",
> "vendorID": "HP",
> "GUID": "3600c0ff0002631c42168f1560100",
> "capacity": "1198996324352",
> "fwrev": "G22x",
> "discard_zeroes_data": 0,
> "vgUUID": "xGCmpC-DhHe-3v6v-6LJw-iS24-ExCE-0Hv48U",
> "pathlist": [],
> "pvsize": "1198698528768",
> "discard_max_bytes": 0,
> "pathstatus": [
> {
> "capacity": "1198996324352",
> "physdev": "sdc",
> "type": "FCP",
> "state": "active",
> "lun": "16"
> },
> {
> "capacity": "1198996324352",
> "physdev": "sds",
> "type": "FCP",
> "state": "active",
> "lun": "16"
> },
>
>
> ...
>
> The first one is an embedded flash drive:
> lrwxrwxrwx 1 root root 10 Jul 12 17:11
> /dev/disk/by-id/usb-HP_iLO_LUN_01_Media_0_02660A01-0:1 -> ../../sddj
> lrwxrwxrwx 1 root root 10 Jul 12 17:11
> /dev/disk/by-path/pci-:00:14.0-usb-0:3.1:1.0-scsi-0:0:0:1 -> ../../sddj
>
> So why "type": "FCP",  ?
>

"FCP" actually means "not iSCSI". This why your sas storage works while
oVirt does
know anything about sas.

This is why the blacklist by protocol feature was introduced in 7.5, to
multipath can grab
only shared storage, and avoid grabbing local devices like your SSD.
See https://bugzilla.redhat.com/show_bug.cgi?id=1593459

According to this bug:
https://bugzilla.redhat.com/show_bug.cgi?id=1607749

The fix is available in:
device-mapper-multipath-0.4.9-119.el7_5.1.x86_64

Which device-mapper-multipath package are you using?


> The second is indeed a SAS drives behind a SAS SAN (a MSA 2040 SAS from
> HPE).
>
>
> >  ...
> > Where do I find the protocol multipath thinks the drives are using ?
> >
> > multipath.conf(5) says:
> >
> >The protocol strings that multipath recognizes are scsi:fcp,
> scsi:spi, scsi:ssa, scsi:sbp,
> >scsi:srp, scsi:iscsi, scsi:sas, scsi:adt, scsi:ata, scsi:unspec,
> ccw, cciss, nvme,  and
> >undef.  The protocol that a path is using can be viewed by
> running multipathd show
> >paths format "%d %P"
>
> I have a centos 7.5:
>
> lsb_release -a
> LSB Version::core-4.1-amd64:core-4.1-noarch
> Distributor ID: CentOS
> Description:CentOS Linux release 7.5.1804 (Core)
> Release:7.5.1804
> Codename:   Core
>
> and I don't have this in multipath.conf(5). But blacklist_exceptions
> exists.
>
> The given command don't works:
> multipathd show paths format "%d %P"
> 

[ovirt-users] Re: self-hosted engine lost password

2018-09-04 Thread Rob Epping
On Thu, 2018-08-30 at 17:34 +0200, Simone Tiraboschi wrote:
> Hi,

Hi Simone,

> I'd suggest to
> # set global maintenance mode
> hosted-engine --set-maintenance --mode=global
> # shutdown the engine VM
> hosted-engine --vm-shutdown
> # start the engine VM in paused mode with:
> hosted-engine --vm-start-paused
> # set VNC password
> hosted-engine --add-console-password
> # connect to the VM via VNC
> # resume HE VM
> /usr/bin/virsh -c
> qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf
> resume
> HostedEngine
> # quick go back to grub menu on the engine VM and follow this:
> https://wiki.centos.org/TipsAndTricks/ResetRootPassword

Thank you, this worked great with the caveat that in centos 7 the boot
paramters "rd.break enforcing=0" need to be directly after the kernel,
not at the end of the "linux/linux16/linuxefi" line.

Our hosted engine is now up to date.

GRTNX,
RobJE

> On Thu, Aug 30, 2018 at 3:55 PM Rob Epping 
> wrote:
> 
> > Hello list,
> > 
> > I'm trying to update our ovirt self-hosted engine to the latest
> > version
> > using the docs at
> > https://www.ovirt.org/documentation/upgrade-guide/chap-Updates_betw
> > een_
> > Minor_Releases/#updating-the-ovirt-self-hosted-engine-and-
> > underlying-
> > virtualization-hosts
> >  > ween_Minor_Releases/#updating-the-ovirt-self-hosted-engine-and-
> > underlying-virtualization-hosts>
> > 
> > For this I need to logon to the engine VM, but none of the
> > passwords we
> > used seems to work.
> > I can not seem to find information on how te reset passwords of the
> > hosted engine, so I'm stuck.
> > 
> > Am I doing something wrong or is access to the self-hosted engine
> > required. When access is required, is there a way to reset the
> > password?
> > 
> > THNX && GRTNX,
> > RobJE___
> > Users mailing list -- users@ovirt.org
> > To unsubscribe send an email to users-le...@ovirt.org
> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> > oVirt Code of Conduct:
> > https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> > https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZZQIU
> > AQIXI5D6ZDD4V6LFVG4WWYB7YEW/
> > 

-- 
Met vriendelijke groet,
Rob Epping | NOC | Afdeling: +31 88 773 25 86


+31 88 77 325 87 | www.speakup.nl
Institutenweg 6 7521 PK Enschede | Postbus 1330 7500 BH Enschede

signature.asc
Description: This is a digitally signed message part
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JHMT7GXO5PRSQOXGQ2S5OEKSAOG25MJ7/


[ovirt-users] Re: Engine Setup Error

2018-09-04 Thread Simone Tiraboschi
On Tue, Sep 4, 2018 at 6:07 PM Sakhi Hadebe  wrote:

> Hi Sahina,
>
> I am sorry I can't reproduce the error nor access the logs since I did a
> fresh installed pn nodes. However now I can't even react that far because
> the engine deployment fails to start the host up:
>
>
> [ INFO ] TASK [Wait for the host to be up]
> [ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": {"ovirt_hosts":
> [{"address": "goku.sanren.ac.za", "affinity_labels": [],
> "auto_numa_status": "unknown", "certificate": {"organization": "
> sanren.ac.za", "subject": "O=sanren.ac.za,CN=goku.sanren.ac.za"},
> "cluster": {"href": 
> "/ovirt-engine/api/clusters/1ca368cc-b052-11e8-b7de-00163e008187",
> "id": "1ca368cc-b052-11e8-b7de-00163e008187"}, "comment": "", "cpu":
> {"speed": 0.0, "topology": {}}, "device_passthrough": {"enabled": false},
> "devices": [], "external_network_provider_configurations": [],
> "external_status": "ok", "hardware_information": {"supported_rng_sources":
> []}, "hooks": [], "href": "/ovirt-engine/api/hosts/
> 1c575995-70b1-43f7-b348-4a9788e070cd", "id": 
> "1c575995-70b1-43f7-b348-4a9788e070cd",
> "katello_errata": [], "kdump_status": "unknown", "ksm": {"enabled": false},
> "max_scheduling_memory": 0, "memory": 0, "name": "goku.sanren.ac.za",
> "network_attachments": [], "nics": [], "numa_nodes": [], "numa_supported":
> false, "os": {"custom_kernel_cmdline": ""}, "permissions": [], "port":
> 54321, "power_management": {"automatic_pm_enabled": true, "enabled": false,
> "kdump_detection": true, "pm_proxies": []}, "protocol": "stomp",
> "se_linux": {}, "spm": {"priority": 5, "status": "none"}, "ssh":
> {"fingerprint": "SHA256:B3/PDH551EFid93fm6PoRryi6/cXuVE8yNgiiiROh84",
> "port": 22}, "statistics": [], "status": "install_failed",
> "storage_connection_extensions": [], "summary": {"total": 0}, "tags": [],
> "transparent_huge_pages": {"enabled": false}, "type": "ovirt_node",
> "unmanaged_networks": [], "update_available": false}]}, "attempts": 120,
> "changed": false}
>
> "status": "install_failed"

You have to check host-deploy logs to get a details error message.


>
> Please help.
>
> On Mon, Sep 3, 2018 at 1:34 PM, Sahina Bose  wrote:
>
>>
>>
>> On Wed, Aug 29, 2018 at 8:39 PM, Sakhi Hadebe  wrote:
>>
>>> Hi,
>>>
>>> I am sorry to bother you again.
>>>
>>> I am trying to deploy an oVirt engine for oVirtNode-4.2.5.1. I get the
>>> same error I encountered before:
>>>
>>> [ INFO  ] TASK [Add glusterfs storage domain]
>>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
>>> "[Problem while trying to mount target]". HTTP response code is 400.
>>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
>>> "Fault reason is \"Operation Failed\". Fault detail is \"[Problem while
>>> trying to mount target]\". HTTP response code is 400."}
>>>   Please specify the storage you would like to use (glusterfs,
>>> iscsi, fc, nfs)[nfs]:
>>>
>>> The glusterd daemon is running.
>>>
>>
>> mounting 172.16.4.18:/engine at
>> /rhev/data-center/mnt/glusterSD/172.16.4.18:_engine (mount:204)
>> 2018-08-29 16:47:28,846+0200 ERROR (jsonrpc/3) [storage.HSM] Could not
>> connect to storageServer (hsm:2398)
>>
>> Can you try to see if you are able to mount 172.16.4.18:/engine on the
>> server you're deploying Hosted Engine using "mount -t glusterfs 
>> 172.16.4.18:/engine
>> /mnt/test"
>>
>>
>>> During the deployment of the engine it sets the engine entry in the
>>> /etc/hosts file with the IP Address of 192.168.124.* which it gets form the
>>> virbr0 bridge interface. I stopped the bridge and deleted it, but still
>>> giving the same error. Not sure what causes it to use that interface.
>>> Please help!
>>>
>>> But I give the engine an IP of 192.168.1.10 same subnet as my gateway
>>> and my ovirtmgmt bridge. Attached is the ifconfig output of my Node,
>>> engine.log and vdsm.log.
>>>
>>> Your assistance is always appreciated
>>>
>>>
>>>
>>>
>>>
>>> On Wed, Jul 11, 2018 at 11:47 AM, Sahina Bose  wrote:
>>>
 Is glusterd running on the server: goku.sanren.**
 There's an error
 Failed to get volume info: Command execution failed
 error: Connection failed. Please check if gluster daemon is operational

 Please check the volume status using "gluster volume status engine"

 and if all looks ok, attach the mount logs from /var/log/glusterfs

 On Wed, Jul 11, 2018 at 1:57 PM, Sakhi Hadebe 
 wrote:

> Hi,
>
> I have managed to fix the error by enabling the DMA Virtualisation in
> BIOS. I am now hit with a new error: It's failing to add a glusterfs
> storage domain:
>
> [ INFO  ] TASK [Add glusterfs storage domain]
> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
> "[Problem while trying to mount target]". HTTP response code is 400.
> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
> "Fault reason is \"Operation Failed\". Fault detail is \"[Problem while
> trying to mo

[ovirt-users] Re: Engine Setup Error

2018-09-04 Thread Sakhi Hadebe
Hi Sahina,

I am sorry I can't reproduce the error nor access the logs since I did a
fresh installed pn nodes. However now I can't even react that far because
the engine deployment fails to start the host up:


[ INFO ] TASK [Wait for the host to be up]
[ ERROR ] fatal: [localhost]: FAILED! => {"ansible_facts": {"ovirt_hosts":
[{"address": "goku.sanren.ac.za", "affinity_labels": [],
"auto_numa_status": "unknown", "certificate": {"organization": "sanren.ac.za",
"subject": "O=sanren.ac.za,CN=goku.sanren.ac.za"}, "cluster": {"href":
"/ovirt-engine/api/clusters/1ca368cc-b052-11e8-b7de-00163e008187", "id":
"1ca368cc-b052-11e8-b7de-00163e008187"}, "comment": "", "cpu": {"speed":
0.0, "topology": {}}, "device_passthrough": {"enabled": false}, "devices":
[], "external_network_provider_configurations": [], "external_status":
"ok", "hardware_information": {"supported_rng_sources": []}, "hooks": [],
"href": "/ovirt-engine/api/hosts/1c575995-70b1-43f7-b348-4a9788e070cd",
"id": "1c575995-70b1-43f7-b348-4a9788e070cd", "katello_errata": [],
"kdump_status": "unknown", "ksm": {"enabled": false},
"max_scheduling_memory": 0, "memory": 0, "name": "goku.sanren.ac.za",
"network_attachments": [], "nics": [], "numa_nodes": [], "numa_supported":
false, "os": {"custom_kernel_cmdline": ""}, "permissions": [], "port":
54321, "power_management": {"automatic_pm_enabled": true, "enabled": false,
"kdump_detection": true, "pm_proxies": []}, "protocol": "stomp",
"se_linux": {}, "spm": {"priority": 5, "status": "none"}, "ssh":
{"fingerprint": "SHA256:B3/PDH551EFid93fm6PoRryi6/cXuVE8yNgiiiROh84",
"port": 22}, "statistics": [], "status": "install_failed",
"storage_connection_extensions": [], "summary": {"total": 0}, "tags": [],
"transparent_huge_pages": {"enabled": false}, "type": "ovirt_node",
"unmanaged_networks": [], "update_available": false}]}, "attempts": 120,
"changed": false}


Please help.

On Mon, Sep 3, 2018 at 1:34 PM, Sahina Bose  wrote:

>
>
> On Wed, Aug 29, 2018 at 8:39 PM, Sakhi Hadebe  wrote:
>
>> Hi,
>>
>> I am sorry to bother you again.
>>
>> I am trying to deploy an oVirt engine for oVirtNode-4.2.5.1. I get the
>> same error I encountered before:
>>
>> [ INFO  ] TASK [Add glusterfs storage domain]
>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
>> "[Problem while trying to mount target]". HTTP response code is 400.
>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault
>> reason is \"Operation Failed\". Fault detail is \"[Problem while trying
>> to mount target]\". HTTP response code is 400."}
>>   Please specify the storage you would like to use (glusterfs,
>> iscsi, fc, nfs)[nfs]:
>>
>> The glusterd daemon is running.
>>
>
> mounting 172.16.4.18:/engine at /rhev/data-center/mnt/
> glusterSD/172.16.4.18:_engine (mount:204)
> 2018-08-29 16:47:28,846+0200 ERROR (jsonrpc/3) [storage.HSM] Could not
> connect to storageServer (hsm:2398)
>
> Can you try to see if you are able to mount 172.16.4.18:/engine on the
> server you're deploying Hosted Engine using "mount -t glusterfs 
> 172.16.4.18:/engine
> /mnt/test"
>
>
>> During the deployment of the engine it sets the engine entry in the
>> /etc/hosts file with the IP Address of 192.168.124.* which it gets form the
>> virbr0 bridge interface. I stopped the bridge and deleted it, but still
>> giving the same error. Not sure what causes it to use that interface.
>> Please help!
>>
>> But I give the engine an IP of 192.168.1.10 same subnet as my gateway and
>> my ovirtmgmt bridge. Attached is the ifconfig output of my Node, engine.log
>> and vdsm.log.
>>
>> Your assistance is always appreciated
>>
>>
>>
>>
>>
>> On Wed, Jul 11, 2018 at 11:47 AM, Sahina Bose  wrote:
>>
>>> Is glusterd running on the server: goku.sanren.**
>>> There's an error
>>> Failed to get volume info: Command execution failed
>>> error: Connection failed. Please check if gluster daemon is operational
>>>
>>> Please check the volume status using "gluster volume status engine"
>>>
>>> and if all looks ok, attach the mount logs from /var/log/glusterfs
>>>
>>> On Wed, Jul 11, 2018 at 1:57 PM, Sakhi Hadebe 
>>> wrote:
>>>
 Hi,

 I have managed to fix the error by enabling the DMA Virtualisation in
 BIOS. I am now hit with a new error: It's failing to add a glusterfs
 storage domain:

 [ INFO  ] TASK [Add glusterfs storage domain]
 [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
 "[Problem while trying to mount target]". HTTP response code is 400.
 [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg":
 "Fault reason is \"Operation Failed\". Fault detail is \"[Problem while
 trying to mount target]\". HTTP response code is 400."}
   Please specify the storage you would like to use (glusterfs,
 iscsi, fc, nfs)[nfs]:

 Attached are vdsm and engine log files.





 On Wed, Jul 11, 2018 at 9:57 AM, Sakhi Hadebe 
 wrote:

>
>
>

[ovirt-users] Re: Upgraded host, engine now won't boot

2018-09-04 Thread Darrell Budic
Glad you got into it and got it working. Not sure why it keeps unpausing, could 
open a bug if you wanted.

Yep, the engine-setup is on the engine vm itself, not the hosts. You want 
https://www.ovirt.org/documentation/self-hosted/chap-Maintenance_and_Upgrading_Resources/
 
,
 although if you aren’t running the appliance in the first place, I don’t think 
it all applies. Just log in and run yum update or yum update ovirt* and then 
engine-setup again.


> From: Jim Kusznir 
> Subject: Re: [ovirt-users] Upgraded host, engine now won't boot
> Date: September 3, 2018 at 10:21:01 PM CDT
> To: Darrell Budic
> Cc: users
> 
> Ok, finally got it...Had to get a terminal ready with the virsh command and 
> guess what the instance number was, and then run suspend right after starting 
> with --vm-start-paused.  Got it to really be paused, got into the console, 
> booted the old kernel, and have now been repairing a bad yum transactionI 
> *think* I've finished that.
> 
> So, if I understand correctly, after the yum update, I should run 
> engine-setup?  Do I run that inside the engine vm, or on the host its running 
> on?
> 
> BTW: I did look up upgrade procedures on the documentation for the release.  
> It links to two or three levels of other documents, then ends in an error 404.
> 
> --Jim
> 
> On Mon, Sep 3, 2018 at 6:39 PM, Jim Kusznir  > wrote:
> global maintence mode is already on.  hosted-engine --vm-start-paused results 
> in a non-paused VM being started.  Of course, this is executed after 
> hosted-engine --vm-poweroff and suitable time left to let things shut down.
> 
> I just ran another test, and did in fact see the engine was briefly paused, 
> but then was quickly put in the running state.  I don't know by what, though. 
>  Global maintence mode is definitely enabled, every run of the hosted-engine 
> command reminds me!
> 
> 
> 
> 
> 
> On Mon, Sep 3, 2018 at 11:12 AM, Darrell Budic  > wrote:
> Don’t know if there’s anything special, it’s been a while since I’ve needed 
> to start it in paused mode. Try putting it in HA maintenance mode from the 
> CLI and then start it in paused mode maybe?
> 
>> From: Jim Kusznir mailto:j...@palousetech.com>>
>> Subject: Re: [ovirt-users] Upgraded host, engine now won't boot
>> Date: September 3, 2018 at 1:08:27 PM CDT
>> 
>> To: Darrell Budic
>> Cc: users
>> 
>> Unfortunately, I seem unable to get connected to the console early enough to 
>> actually see a kernel list.
>> 
>> I've tried the hosted-engine --start-vm-paused command, but it just starts 
>> it (running mode, not paused).  By the time I can get vnc connected, I have 
>> just that last line.  ctrl-alt-del doesn't do anything with it, either.  
>> sending a reset through virsh seems to just kill the VM (it doesn't respawn).
>> 
>> ha seems to have some trouble with this too...Originally I allowed ha to 
>> start it, and it would take it a good long while before it gave up on the 
>> engine and reset it.  It instantly booted to the same crashed state, and 
>> again waited a "good long while" (sorry, never timed it, but I know it was 
>> >5 min).
>> 
>> My current thought is that I need to get the engine started in paused mode, 
>> connect vnc, then unpause it with virsh to catch what is happening.  Is 
>> there any magic to getting it started in paused mode?
>> 
>> On Mon, Sep 3, 2018 at 11:03 AM, Darrell Budic > > wrote:
>> Send it a ctl-alt-delete and see what happens. Possibly try an older kernel 
>> at the grub boot menu. Could also try stopping it with hosted-engine 
>> —vm-stop and let HA reboot it, see if it boots or get onto the console 
>> quickly and try and watch more of the boot.
>> 
>> Ssh and yum upgrade is fine for the OS, although it’s a good idea to enable 
>> Global HA Maintenance first so the HA watchdogs don’t reboot it in the 
>> middle of that. After that, run “engine-setup” again, at least if there are 
>> new ovirt engine updates to be done. Then disable Global HA Maintenance, and 
>> run "shutdown -h now” to stop the Engine VM (rebooting seems to cause it to 
>> exit anyway, HA seems to run it as a single execution VM. Or at least in the 
>> past, it seems to quit anyway on me and shutdown triggered HA faster). Wait 
>> a few minutes, and HA will respawn it on a new instance and you can log into 
>> your engine again.
>>> From: Jim Kusznir mailto:j...@palousetech.com>>
>>> Subject: Re: [ovirt-users] Upgraded host, engine now won't boot
>>> Date: September 3, 2018 at 12:45:22 PM CDT
>>> To: Darrell Budic
>>> Cc: users
>>> 
>>> 
>>> Thanks to Jayme who pointed me to the --add-console-password hosted-engine 
>>> command to set a password for vnc.  Using that, I see only the single line:
>>> 
>>> Probing EDD (edd=off to disable)... ok
>>> 
>>> --Jim
>>> 
>>> On Mon, Sep 3, 2018 at 10:26 AM, Jim Kuszni

[ovirt-users] Re: ovirt4 api search storage domain

2018-09-04 Thread Simone Tiraboschi
On Tue, Sep 4, 2018 at 3:10 PM David David  wrote:

> hi
>
> need to lookup storage name by vm disk id.
> how to do it in python or in java?
>

#!/usr/bin/env python
# -*- coding: utf-8 -*-
import logging
import ovirtsdk4 as sdk

logging.basicConfig(level=logging.DEBUG, filename='example.log')

# set variables
fqdn = 'engine.example.com'
password = 'yourpwd'
disk_id = '3810cf0f-f559-4fa8-a441-935e4ecbc794'

# Create the connection to the server:
connection = sdk.Connection(
url='https://{fqdn}/ovirt-engine/api'.format(fqdn=fqdn),
username='admin@internal',
password=password,
insecure=True,
debug=True,
log=logging.getLogger(),
)

disk =
connection.system_service().disks_service().disk_service(disk_id).get()
sd_id = disk.storage_domains[0].id
sd = connection.system_service(
).storage_domains_service().storage_domain_service(sd_id).get()

print("storage_domain: {sdname}".format(sdname=sd.name))

# Close the connection to the server:
connection.close()




> thanks
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/YC5UZA6TYQQFC2P2WW6TRVFKXV7X5YVI/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SOARWCQTMSUIFBSANSDARX5MKVCEXYLR/


[ovirt-users] OVN and impact of missing connectivity with OVN Provider/Central

2018-09-04 Thread Gianluca Cecchi
Hello,
I have VM1 and VM2 with their vnics on OVN.
They are running on the same host.
Suppose this host (and so its OVN Controller) looses connectivity with the
OVN Provider (that in my case runs on oVirt engine, that is an external
server).
Is it correct/expected that VM1 looses connectivity with VM2 until fixed?

So, in other words, is the OVN Provider a sort of single point of failure
(eg if I restart enigine in my case)?

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ESBKINTLKTWNQ2PXCCFSPQBHNK5ECSNO/


[ovirt-users] ovirt4 api search storage domain

2018-09-04 Thread David David
hi

need to lookup storage name by vm disk id.
how to do it in python or in java?
thanks
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YC5UZA6TYQQFC2P2WW6TRVFKXV7X5YVI/


[ovirt-users] Re: Managing multiple oVirt installs?

2018-09-04 Thread femi adegoke
Just an FYI:
The Glance registry does not have the latest current stable release which is 
Gaprindashvili-4
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/V4C73LOMHDJR3MPA4E4CXTBNVQXHGNP4/


[ovirt-users] Re: IPoIB broken with ovirt 4.2.6

2018-09-04 Thread Sandro Bonazzola
2018-09-03 16:22 GMT+02:00 Giulio Casella :

> Hi,
> latest ovirt node stable (4.2.6 today) introduced a bug in kernel: IP
> over infiniband is not workingh anymore after an upgrade, due to kernel
> 3.10.0-862.11.6.el7.x86_64.
>
> You can find some detail here:
>
> https://bugs.centos.org/view.php?id=15193
>
> dmesg is full of "failed to modify QP to RTR: -22", and the networking
> stack (in my case used to connect to storage) is broken. The interface
> can obtain an address via DHCP, but also a simple ICMP ping fails.
>
> Does someone have news about a fix for this issue?
>

Thanks for reporting, I wasn't aware of this issue.
We'll issue an async respin as soon as a new kernel will be available.
Adding this to release notes.




>
> TIA,
> Giulio
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-
> guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/
> message/5ZJNK4ZENUJSBREZIW5OM55YLMI45B34/
>



-- 

SANDRO BONAZZOLA

MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV

Red Hat EMEA 

sbona...@redhat.com


___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NRPOEVTJ33P35NWEWS452TY43NNVUGZX/


[ovirt-users] Re: Managing multiple oVirt installs?

2018-09-04 Thread Donny Davis
Go-to the glance registry that ships with ovirt. There is a manageiq image
in there. It ships as an appliance, so there is no install required.

You can also get an image from the manageiq site for ovirt and upload it

On Tue, Sep 4, 2018, 12:06 AM femi adegoke  wrote:

> Donny,
>
> Have you installed manageiq in oVirt?
> Can you share any useful tips/blog posts/notes etc.?
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/FODXM24E7OGYH46RP5DFG62UHTU53LPP/
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/M3MDJUNWI4NGSA6VWVRNV7DWAIPHQWDN/


[ovirt-users] Re: Wrong network threshold limit warnings on 4.2.5

2018-09-04 Thread Edward Haas
On Tue, Sep 4, 2018 at 10:42 AM, Gianluca Cecchi 
wrote:

> On Tue, Sep 4, 2018 at 9:02 AM Edward Haas  wrote:
>
>> Hello Florian,
>>
>> Thanks for checking the patch and posting the bug.
>>
>> You need to restart vdsmd and supervdsmd.
>> It should not affect running VM/s, but you always have a risk that
>> something unexpected can happen. Perhaps try it on a host and then proceed
>> with others.
>>
>> Thanks,
>> Edy.
>>
>
> I'm having similar problem in a 3 hosts oVirt test cluster with these
> notifications every day on 1Gbit adapters.
> I have bond0 on em1 and em2 and then bondo.65, bond0.68, bond0.167 vlans
> defined for the VMs
> I get these warnings
>
> Message:Host ov300 has network interface which exceeded the defined threshold 
> [95%] (em1: transmit rate[98%], receive rate [0%])
>
> when actually I think the 3 VMs running on this host generate few MB/s of
> traffic
> I applied the changes to the 3 hosts.
>
> I notice that due to dependencies it is sufficient to restart supervdsmd
> and then also vdsmd will be automatically restarted, correct?
>
> In my case for each of the 3 hosts, after restarting supervdsmd I got
> messages like these, but without impacts on runnign VMs
>
> VDSM ov300 command GetStatsAsyncVDS failed: Broken pipe 9/4/18 9:07:52 AM
> Host ov300 is not responding. It will stay in Connecting state for a grace
> period of 61 seconds and after that an attempt to fence the host will be
> issued. 9/4/18 9:07:52 AM
> No faulty multipath paths on host ov300 9/4/18 9:07:58 AM
> Executing power management status on Host ov300 using Proxy Host ov200 and
> Fence Agent ipmilan:10.10.193.103. 9/4/18 9:07:58 AM
> Status of host ov300 was set to Up. 9/4/18 9:07:58 AM
> Host ov300 power management was verified successfully. 9/4/18 9:07:58 AM
>
> Please note that when doing on SPM host you could also get these:
>
> VDSM ov301 command SpmStatusVDS failed: Broken pipe 9/4/18 9:10:00 AM
> Host ov301 is not responding. It will stay in Connecting state for a grace
> period of 81 seconds and after that an attempt to fence the host will be
> issued. 9/4/18 9:10:00 AM
> Invalid status on Data Center MYDC. Setting Data Center status to Non
> Responsive (On host ov301, Error: Network error during communication with
> the Host.). 9/4/18 9:10:00 AM
>
> with reassignment of SPM role:
> VDSM command GetStoragePoolInfoVDS failed: Heartbeat exceeded 9/4/18
> 9:10:12 AM
> Storage Pool Manager runs on Host ov200 (Address: ov200), Data Center
> MYDC. 9/4/18 9:10:14 AM
>
> Probably safer to manually move the SPM before restarting supervdsmd on
> that host.
>
> Let's see this evening if I will get any message about thresholds.
>
> BTW: one question. I see in the code iface.Type.NIC and now
> also iface.Type.BOND. Don't you think that you should manage also the
> network teaming option available in RH EL 7, as described here:
> https://access.redhat.com/documentation/en-us/red_hat_
> enterprise_linux/7/html/networking_guide/ch-configure_network_teaming
> ?
> This only if it is supported to use the new network teaming implementation
> in oVirt, and I'm not sure about it...
>

There are no immediate plans to support it in VDSM.
We are evaluating the options to change the way we interact with the host
networking, that may open the door for team and others to get in.


> Thanks,
> Gianluca
>
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/A7NDZZ5QLNJBBMFGQAFX3CGP4LKAZVNZ/


[ovirt-users] Re: [ANN] oVirt 4.2.6 is now generally available

2018-09-04 Thread Fabrice Bacchella


> Le 3 sept. 2018 à 19:15, Nir Soffer  a écrit :

Thank you for you help, but I'm still not out of trouble.

> 
> On Mon, Sep 3, 2018 at 8:01 PM Fabrice Bacchella 
>  wrote:
> 
>> Le 3 sept. 2018 à 18:31, Nir Soffer  a écrit :
>> 
>> On Mon, Sep 3, 2018 at 5:07 PM Fabrice Bacchella 
>>  wrote:
>> In the release notes, I see:
>> 
>> • BZ 1622700 [downstream clone - 4.2.6] [RFE][Dalton] - Blacklist all local 
>> disk in multipath on RHEL / RHEV Host (RHEL 7.5)
>> Feature:
>> Blacklist local devices in multipath. 
>> 
>> Reason: 
>> multipath repeatedly logs irrelevant errors for local devices.
>> 
>> Result: 
>> Local devices are blacklisted, and no irrelevant errors are logged anymore.
>> 
>> What defines a local disk ? I'm using a SAN on SAS. For many peoples, SAS is 
>> only for local disks, but that's not the case. Will other 4.2.6 will detect 
>> that ?
>> 
>> We don't have any support for SAS.
>> 
>> If you SAS drives are attached to the host using FC or iSCSI, you are fine.
> 
> Nope, they are attached using SAS.
> 
> I guess oVirt see them as FCP devices?

yes, in ovirt UI, I've configured my storage to be on FCP, and everything 
worked well since 3.6.

> 
> Are these disks connected to multiple hosts?

Yes, that's a real SAN, multi-attached to HPE's blades
> 
> Please share the output of:
> 
> vdsm-client Host getDeviceList

Things are strange:

{
"status": "used", 
"vendorID": "HP iLO", 
"GUID": "HP_iLO_LUN_01_Media_0_02660A01-0:1", 
"capacity": "1073741824", 
"fwrev": "2.10", 
"discard_zeroes_data": 0, 
"vgUUID": "", 
"pathlist": [], 
"pvsize": "", 
"discard_max_bytes": 0, 
"pathstatus": [
{
"capacity": "1073741824", 
"physdev": "sddj", 
"type": "FCP", 
"state": "active", 
"lun": "1"
}
], 
"devtype": "FCP", 
"physicalblocksize": "512", 
"pvUUID": "", 
"serial": "", 
"logicalblocksize": "512", 
"productID": "LUN 01 Media 0"
},
...
{
"status": "used", 
"vendorID": "HP", 
"GUID": "3600c0ff0002631c42168f1560100", 
"capacity": "1198996324352", 
"fwrev": "G22x", 
"discard_zeroes_data": 0, 
"vgUUID": "xGCmpC-DhHe-3v6v-6LJw-iS24-ExCE-0Hv48U", 
"pathlist": [], 
"pvsize": "1198698528768", 
"discard_max_bytes": 0, 
"pathstatus": [
{
"capacity": "1198996324352", 
"physdev": "sdc", 
"type": "FCP", 
"state": "active", 
"lun": "16"
}, 
{
"capacity": "1198996324352", 
"physdev": "sds", 
"type": "FCP", 
"state": "active", 
"lun": "16"
}, 


...

The first one is an embedded flash drive:
lrwxrwxrwx 1 root root 10 Jul 12 17:11 
/dev/disk/by-id/usb-HP_iLO_LUN_01_Media_0_02660A01-0:1 -> ../../sddj
lrwxrwxrwx 1 root root 10 Jul 12 17:11 
/dev/disk/by-path/pci-:00:14.0-usb-0:3.1:1.0-scsi-0:0:0:1 -> ../../sddj

So why "type": "FCP",  ?

The second is indeed a SAS drives behind a SAS SAN (a MSA 2040 SAS from HPE).


>  ...
> Where do I find the protocol multipath thinks the drives are using ?
> 
> multipath.conf(5) says:
> 
>The protocol strings that multipath recognizes are scsi:fcp, scsi:spi, 
> scsi:ssa, scsi:sbp,
>scsi:srp, scsi:iscsi, scsi:sas, scsi:adt, scsi:ata, scsi:unspec, ccw, 
> cciss, nvme,  and
>undef.  The protocol that a path is using can be viewed by running 
> multipathd show
>paths format "%d %P"

I have a centos 7.5:

lsb_release -a
LSB Version::core-4.1-amd64:core-4.1-noarch
Distributor ID: CentOS
Description:CentOS Linux release 7.5.1804 (Core) 
Release:7.5.1804
Codename:   Core

and I don't have this in multipath.conf(5). But blacklist_exceptions exists.

The given command don't works:
multipathd show paths format "%d %P"
dev  
sddi 
sddj 
sda  
...


> 
> So this should work:
> 
> blacklist_exceptions {
> protocol "(scsi:fcp|scsi:iscsi|scsi:sas)" 
>   
>
> }
> 
> The best way to make this change is to create a dropin conf file, 
> and not touch /etc/multipath.conf, so vdsm will be able to update later.
> 
> $cat /etc/multipath/conf.d/local.conf
> blacklist_exceptions {
> protocol "(scsi:fcp|scsi:iscsi|scsi:sas)" 
>   
>
> }

The header in /etc/multipath.conf says:

# The recommended way to add configuration for your storage is to add a
# drop-in configuration file in "/etc/mu

[ovirt-users] Re: Wrong network threshold limit warnings on 4.2.5

2018-09-04 Thread Gianluca Cecchi
On Tue, Sep 4, 2018 at 9:02 AM Edward Haas  wrote:

> Hello Florian,
>
> Thanks for checking the patch and posting the bug.
>
> You need to restart vdsmd and supervdsmd.
> It should not affect running VM/s, but you always have a risk that
> something unexpected can happen. Perhaps try it on a host and then proceed
> with others.
>
> Thanks,
> Edy.
>

I'm having similar problem in a 3 hosts oVirt test cluster with these
notifications every day on 1Gbit adapters.
I have bond0 on em1 and em2 and then bondo.65, bond0.68, bond0.167 vlans
defined for the VMs
I get these warnings

Message:Host ov300 has network interface which exceeded the defined
threshold [95%] (em1: transmit rate[98%], receive rate [0%])

when actually I think the 3 VMs running on this host generate few MB/s of
traffic
I applied the changes to the 3 hosts.

I notice that due to dependencies it is sufficient to restart supervdsmd
and then also vdsmd will be automatically restarted, correct?

In my case for each of the 3 hosts, after restarting supervdsmd I got
messages like these, but without impacts on runnign VMs

VDSM ov300 command GetStatsAsyncVDS failed: Broken pipe 9/4/18 9:07:52 AM
Host ov300 is not responding. It will stay in Connecting state for a grace
period of 61 seconds and after that an attempt to fence the host will be
issued. 9/4/18 9:07:52 AM
No faulty multipath paths on host ov300 9/4/18 9:07:58 AM
Executing power management status on Host ov300 using Proxy Host ov200 and
Fence Agent ipmilan:10.10.193.103. 9/4/18 9:07:58 AM
Status of host ov300 was set to Up. 9/4/18 9:07:58 AM
Host ov300 power management was verified successfully. 9/4/18 9:07:58 AM

Please note that when doing on SPM host you could also get these:

VDSM ov301 command SpmStatusVDS failed: Broken pipe 9/4/18 9:10:00 AM
Host ov301 is not responding. It will stay in Connecting state for a grace
period of 81 seconds and after that an attempt to fence the host will be
issued. 9/4/18 9:10:00 AM
Invalid status on Data Center MYDC. Setting Data Center status to Non
Responsive (On host ov301, Error: Network error during communication with
the Host.). 9/4/18 9:10:00 AM

with reassignment of SPM role:
VDSM command GetStoragePoolInfoVDS failed: Heartbeat exceeded 9/4/18
9:10:12 AM
Storage Pool Manager runs on Host ov200 (Address: ov200), Data Center MYDC.
9/4/18 9:10:14 AM

Probably safer to manually move the SPM before restarting supervdsmd on
that host.

Let's see this evening if I will get any message about thresholds.

BTW: one question. I see in the code iface.Type.NIC and now
also iface.Type.BOND. Don't you think that you should manage also the
network teaming option available in RH EL 7, as described here:
https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/7/html/networking_guide/ch-configure_network_teaming
?
This only if it is supported to use the new network teaming implementation
in oVirt, and I'm not sure about it...

Thanks,
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/EXK6HG5JLSCDLV77FGSZNTCPITTBG7IY/


[ovirt-users] Re: Wrong network threshold limit warnings on 4.2.5

2018-09-04 Thread Edward Haas
Hello Florian,

Thanks for checking the patch and posting the bug.

You need to restart vdsmd and supervdsmd.
It should not affect running VM/s, but you always have a risk that
something unexpected can happen. Perhaps try it on a host and then proceed
with others.

Thanks,
Edy.


On Tue, Sep 4, 2018 at 9:45 AM, Florian Schmid  wrote:

> Hello Edward,
>
> I have applied the patch and it looks very good!
> vdsm-client Host getStats ->
> ...
> "enp9s0.88": {
> "rxErrors": "0",
> "name": "enp9s0.88",
> "tx": "1226",
> "txDropped": "0",
> "sampleTime": 1536043097.701361,
> "rx": "98642",
> "txErrors": "0",
> "state": "up",
> "speed": "1",
> "rxDropped": "0"
> },
> ...
>
> Bridge devices still have only 1000 configured:
> "vm-int-dev": {
> "rxErrors": "0",
> "name": "vm-int-dev",
> "tx": "578",
> "txDropped": "0",
> "sampleTime": 1536043097.701361,
> "rx": "27843284",
> "txErrors": "0",
> "state": "up",
> "speed": "1000",
> "rxDropped": "0"
> },
>
> One important question:
> I want to apply this patch without upgrading all hosts, because this is a
> huge task.
> When I apply that patch only to this particular file, which service do I
> need to restart?
> I have restarted now all three vdsm services, but I think, I can't do that
> while VMs are running on the hosts, do I?
>
> LG Florian
>
> --
> *Von: *"Florian Schmid" 
> *An: *"edwardh" 
> *CC: *"users" 
> *Gesendet: *Dienstag, 4. September 2018 08:32:09
> *Betreff: *[ovirt-users] Re: Wrong network threshold limit warnings on
> 4.2.5
>
> Hello Edward,
>
> raised a bug report: https://bugzilla.redhat.com/show_bug.cgi?id=1625098
>
> I will try your patch.
>
> LG Florian
>
> --
> *Von: *"Edward Haas" 
> *An: *"p staniforth" , "Florian Schmid" <
> fsch...@ubimet.com>
> *CC: *"users" 
> *Gesendet: *Montag, 3. September 2018 16:42:25
> *Betreff: *Re: Wrong network threshold limit warnings on 4.2.5
>
> Indeed looks like a nasty bug.
> Could you please open a bug on this? https://tinyurl.com/ya7crjhf
>
> If you can, could you also verify the fix? https://gerrit.ovirt.org/#/c/
> 94132/
>
> Thanks,
> Edy.
>
>
>
> On Mon, Sep 3, 2018 at 2:32 PM, Staniforth, Paul <
> p.stanifo...@leedsbeckett.ac.uk> wrote:
>
>> Hello Edward,
>>
>>I am also seeing this problem, it's on our
>> ovirtmgmt.
>>
>>
>> cat /sys/class/net/eno49/speed
>> 1
>>
>>
>>
>> cat /sys/class/net/eno49.20/speed
>> 1
>>
>>
>>
>> cat /sys/class/net/ovirtmgmt/speed
>> cat: /sys/class/net/ovirtmgmt/speed: Invalid argument
>>
>>
>> vdsm-client Host getStats ->
>> ...
>>
>> "eno49": {
>> "rxErrors": "0",
>> "name": "eno49",
>> "tx": "3456777",
>> "txDropped": "0",
>> "sampleTime": 1535974190.687987,
>> "rx": "121362321",
>> "txErrors": "0",
>> "state": "up",
>> "speed": "1",
>> "rxDropped": "2"
>> },
>>
>> "eno49.20": {
>> "rxErrors": "0",
>> "name": "eno49.20",
>> "tx": "3384452",
>> "txDropped": "0",
>> "sampleTime": 1535974190.687987,
>> "rx": "115884579",
>> "txErrors": "0",
>> "state": "up",
>> "speed": "1000",
>> "rxDropped": "0"
>> },
>>
>> "ovirtmgmt": {
>> "rxErrors": "0",
>> "name": "ovirtmgmt",
>> "tx": "3383804",
>> "txDropped": "0",
>> "sampleTime": 1535974190.687987,
>> "rx": "115710919",
>> "txErrors": "0",
>> "state": "up",
>> "speed": "1000",
>> "rxDropped": "0"
>> },
>>
>>
>> Regards,
>>
>>Paul S.
>>
>> --
>> *From:* Florian Schmid 
>> *Sent:* 03 September 2018 11:44
>> *To:* edwa...@redhat.com
>> *Cc:* users
>> *Subject:* [ovirt-users] Re: Wrong network threshold limit warnings on
>> 4.2.5
>>
>> Hi Edward,
>>
>> I got some alarms today from a server and I have checked your command
>> there. (not at the time the issue happened!!)
>> Hosts are on latest patch level CentOS 7.5 and oVirt 4.2.5
>>
>> Example:
>> cat /sys/class/net/enp9s0/speed
>> 1
>>
>> cat /sys/class/net/enp9s0.80/speed
>> 1
>>
>> cat /sys/class/net/vm-int-nfs/speed
>> cat: /sys/class/net/vm-int-nfs/speed: invalid argument <- this is
>> the bridge for the VMs
>>
>> vdsm-client Host getStats ->
>> ...
>> "enp9s0": {
>> "rxErrors": "0",
>> "name": "enp9s0",
>> "tx": "3335325754762",
>> "txDropped": "0",
>> "sampleTime": 1535970960.602359,
>>