[ovirt-users] gluster on iSCSI devices in ovirt environment

2024-01-18 Thread p...@email.cz

hello dears,
can anybody explain  me HOWTO realize 2 nodes + aribiter gluster from 
two (three)  locations on block iSCSI devices ?


Something like this:
gluster volume create  TEST replica 3 arbiter 1 iSCSI target >    < 
location-three-host3 - /dev/sda5 e.g. >  - ALL applied on multinode 
ovirt cluster


thx a lot for any help

regs.
Pa.___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VBP7TKZNWLOCY7IAQNEAHWBQXRSQBPE5/


[ovirt-users] Gluster: Ideas for migration

2023-09-14 Thread jonas
Hello

I have to migrate the Gluster volumes from an old oVirt cluster to a newly 
built one. I looked into migration strategies, but everything that Red Hat 
recommends is related to replacing old bricks. In a testing environment I 
created two clusters and wanted to migrate one volume after the other. 
Unfortunately that fails because a node cannot be part of two clusters at the 
same time.

The next thing I see, is to recreate the volumes on the new cluster, then 
constantly rsync the files from the old cluster to the new one and at a 
specified point in time make the cut over where I stop the applicaiton, do a 
final rsync and remount the new volume under the old path.

Is there any other, nicer way I could accomplish migrating a volume from one 
Gluster cluster to another?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/N7USOAG2AZB3DBXPU6PWF4FJBTUAQFHE/


[ovirt-users] Gluster Geo-Replication session not visible in oVirt Manager UI

2023-03-15 Thread simon
I have previously created Gluster geo-replicated sessions from the command line 
on 4.4.5. These sessions appear in the 'Geo-Replication' Tab of the Gluster 
Volume on the Master.

I've just built 2 new oVirt clusters 4.5.3.2 and configured a geo-replicated 
session. The session doesn't appear in the WebUI but appears to be working when 
checked from the command line.

Any ideas anyone please?

Regards

Simon
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MZ5E5JRQX6UF2UKCS6AE72KHI4F2HDHF/


[ovirt-users] Gluster setup for oVirt

2022-09-12 Thread Jonas

Hello all

I tried to setup Gluster volumes in cockpit using the wizard. Based on 
Red Hat's recommendations I wanted to put the Volume for the oVirt 
Engine on a thick provisioned logical volume [1] and therefore removed 
the line thinpoolname and corresponding configuration from the yml file 
(see below). Unfortunately, this approach was not successful. My 
solution is now to only create a data volume and manually create a thick 
provisioned gluster volume manually. What would you recommend doing?


Thanks your any input :)

Regards,
Jonas

[1]: 
https://access.redhat.com/documentation/en-us/red_hat_hyperconverged_infrastructure_for_virtualization/1.8/html/deploying_red_hat_hyperconverged_infrastructure_for_virtualization/rhhi-requirements#rhhi-req-lv


hc_nodes:
  hosts:
    server-005.storage.int.rabe.ch:
  gluster_infra_volume_groups:
    - vgname: vg_tier1_01
  pvname: /dev/md/raid_tier1_gluster
  gluster_infra_mount_devices:
    - path: /gluster_bricks/tier1-ovirt-engine-01/gb-01
  lvname: lv_tier1_ovirt_engine_01
  vgname: vg_tier1_01
    - path: /gluster_bricks/tier1-ovirt-data-01/gb-01
  lvname: lv_tier1_ovirt_data_01
  vgname: vg_tier1_01
  blacklist_mpath_devices:
    - raid_tier1_gluster
  gluster_infra_thinpools:
    - vgname: vg_tier1_01
  thinpoolname: lv_tier1_ovirt_data_01_tp
  poolmetadatasize: 16G
  gluster_infra_lv_logicalvols:
    - vgname: vg_tier1_01
  lvname: lv_tier1_ovirt_engine_01
  lvsize: 100G
    - vgname: vg_tier1_01
  thinpool: lv_tier1_ovirt_data_01_tp
  lvname: lv_tier1_ovirt_data_01
  lvsize: 16000G
    server-006.storage.int.rabe.ch:
  gluster_infra_volume_groups:
    - vgname: vg_tier1_01
  pvname: /dev/md/raid_tier1_gluster
  gluster_infra_mount_devices:
    - path: /gluster_bricks/tier1-ovirt-engine-01/gb-01
  lvname: lv_tier1_ovirt_engine_01
  vgname: vg_tier1_01
    - path: /gluster_bricks/tier1-ovirt-data-01/gb-01
  lvname: lv_tier1_ovirt_data_01
  vgname: vg_tier1_01
  blacklist_mpath_devices:
    - raid_tier1_gluster
  gluster_infra_thinpools:
    - vgname: vg_tier1_01
  thinpoolname: lv_tier1_ovirt_data_01_tp
  poolmetadatasize: 16G
  gluster_infra_lv_logicalvols:
    - vgname: vg_tier1_01
  lvname: lv_tier1_ovirt_engine_01
  lvsize: 100G
    - vgname: vg_tier1_01
  thinpool: lv_tier1_ovirt_data_01_tp
  lvname: lv_tier1_ovirt_data_01
  lvsize: 16000G
    server-007.storage.int.rabe.ch:
  gluster_infra_volume_groups:
    - vgname: vg_tier0_01
  pvname: /dev/md/raid_tier0_gluster
  gluster_infra_mount_devices:
    - path: /gluster_bricks/tier1-ovirt-engine-01/gb-01
  lvname: lv_tier1_ovirt_engine_01
  vgname: vg_tier0_01
    - path: /gluster_bricks/tier1-ovirt-data-01/gb-01
  lvname: lv_tier1_ovirt_data_01
  vgname: vg_tier0_01
  blacklist_mpath_devices:
    - raid_tier0_gluster
  gluster_infra_thinpools:
    - vgname: vg_tier0_01
  thinpoolname: lv_tier1_ovirt_data_01_tp
  poolmetadatasize: 1G
  gluster_infra_lv_logicalvols:
    - vgname: vg_tier0_01
  lvname: lv_tier1_ovirt_engine_01
  lvsize: 20G
    - vgname: vg_tier0_01
  thinpool: lv_tier1_ovirt_data_01_tp
  lvname: lv_tier1_ovirt_data_01
  lvsize: 32G
  vars:
    gluster_infra_disktype: JBOD
    gluster_infra_daling: 1024K
    gluster_set_selinux_labels: true
    gluster_infra_fw_ports:
  - 2049/tcp
  - 54321/tcp
  - 5900/tcp
  - 5900-6923/tcp
  - 5666/tcp
  - 16514/tcp
    gluster_infra_fw_permanent: true
    gluster_infra_fw_state: enabled
    gluster_infra_fw_zone: public
    gluster_infra_fw_services:
  - glusterfs
    gluster_features_force_varlogsizecheck: false
    cluster_nodes:
  - server-005.storage.int.rabe.ch
  - server-006.storage.int.rabe.ch
  - server-007.storage.int.rabe.ch
    gluster_features_hci_cluster: '{{ cluster_nodes }}'
    gluster_features_hci_volumes:
  - volname: tier1-ovirt-engine-01
    brick: /gluster_bricks/tier1-ovirt-engine-01/gb-01
    arbiter: 1
  - volname: tier1-ovirt-data-01
    brick: /gluster_bricks/tier1-ovirt-data-01/gb-01
    arbiter: 1
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WX2CIKP43KOCG76UUWIOVI4QPXYNKQNX/


[ovirt-users] Gluster network - associate brick

2022-08-05 Thread r greg
hi all,

*** new to oVirt and still learning ***

Sorry for the long thread...

I have a 3x node hyperconverged setup on v4.5.1. 

4x 1G NICS

NIC0 
> ovirtmgmt (Hosted-Engine VM)
> vmnetwork vlan102 (all VMs are placed on this network)
NIC1
> migration
NIC2 - NIC3 > bond0
> storage

Logical Networks:
ovirtmgmt - role: VM network | management | display | default route
vmnetwork - role: VM network
migrate - role: migration network
storage - role: gluster network

During deployment, I overlooked a setting and on node2 the host was deployed 
with Name: node2.mydomain.lab --- Hostname/IP: 172.16.20.X/24 (WebUI > Compute 
> Hosts)

I suspect because of this I see the following entries on 
/var/log/ovirt-engine/engine.log (only for node2)

2022-08-04 12:00:15,460Z WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-16) [] 
Could not associate brick 'node2.mydomain.lab:/gluster_bricks/vmstore/vmstore' 
of volume '1ca6a01a-9230-4bb1-844e-8064f3eadb53' with correct network as no 
gluster network found in cluster '1770ade4-0f6f-11ed-b8f6-00163e6faae8'

Is this something I need to be worried about or correct somehow?

From node1

gluster> peer status
Number of Peers: 2

Hostname: node2.mydomain.lab
Uuid: a4468bb0-a3b3-42bc-9070-769da5a13427
State: Peer in Cluster (Connected)
Other names:
172.16.20.X

Hostname: node3.mydomain.lab
Uuid: 2b1273a4-667e-4925-af5e-00904988595a
State: Peer in Cluster (Connected)
Other names:
172.16.20.Z


volume status (same output Online Y --- for volumes vmstore and engine )
Status of volume: data
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick node1.mydomain.lab:/gluster_brick
s/data/data 58734 0  Y   31586
Brick node2.mydomain.lab:/gluster_brick
s/data/data 55148 0  Y   4317 
Brick node3.mydomain.lab:/gluster_brick
s/data/data 57021 0  Y   5242 
Self-heal Daemon on localhost   N/A   N/AY   63170
Self-heal Daemon on node2.mydomain.lab  N/A   N/AY   4365 
Self-heal Daemon on node3.mydomain.lab  N/A   N/AY   5385
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Z5FXYQR5FDMICJTHP7FQ5X4MO4VNND4A/


[ovirt-users] Gluster volume "deleted" by accident --- Is it possible to recover?

2022-07-17 Thread itforums51
hi everyone,

I have a 3x node ovirt 4.4.6 cluster in HC setup.

Today I was intending to extend the data and vmstore volume adding another 
brick each; then by accident I pressed the "cleanup" button. Basically it looks 
that the volume were deleted.

I am wondering whether there is a process of trying to recover these volumes 
and therefore all VMs (including the Hosted-Engine).

```
lvs
  LV VG  Attr   LSize   Pool
 Origin   Data%  Meta%  Move 
Log Cpy%Sync Convert
  gluster_lv_datagluster_vg_sda4 Vwi---t--- 500.00g 
gluster_thinpool_gluster_vg_sda4
 
  gluster_lv_data-brick1 gluster_vg_sda4 Vwi-aot--- 500.00g 
gluster_thinpool_gluster_vg_sda4  0.45  
 
  gluster_lv_engine  gluster_vg_sda4 -wi-a- 100.00g 

 
  gluster_lv_vmstore gluster_vg_sda4 Vwi---t--- 500.00g 
gluster_thinpool_gluster_vg_sda4
 
  gluster_lv_vmstore-brick1  gluster_vg_sda4 Vwi-aot--- 500.00g 
gluster_thinpool_gluster_vg_sda4  0.33  
 
  gluster_thinpool_gluster_vg_sda4   gluster_vg_sda4 twi-aot---  <7.07t 
  11.46  0.89
```  
I would appreciate any advice. 

TIA
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ITF5IYLWGG2MPAPG2JBD2GWA5QZDPVSA/


[ovirt-users] gluster service on the cluster is unchecked on hci cluster

2022-07-10 Thread Jiří Sléžka

Hi,

I would like to change CPU Type in my oVirt 4.4.10 HCI cluster (based on 
3 glusterfs/virt hosts). When I try to I got this error


Error while executing action: Cannot disable gluster service on the 
cluster as it contains volumes.


As I remember I had Gluster Service enabled on this cluster but now both 
(Enable Virt Services and Enable Gluster Service) checkboxes are grayed 
out and Gluster Service is unchecked.


Also Storage / Volumes displays my volumes... well, displays one brick 
on particular host in unknown state (? mark) which is new situation. As 
I can see from command line all bricks are online, no healing in 
progress, all looks good...


I am not sure if the second issue is relevant to first one so main 
question is how can I (re)enable gluster service in my cluster?


Thanks in advance,

Jiri


smime.p7s
Description: Elektronicky podpis S/MIME
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/S4NVCQ33ZSJSHR7P7K7OICSA5F253BVA/


[ovirt-users] Gluster Volume cannot be activated Ovirt 4.5 Centos 8 Stream

2022-06-20 Thread m . rohweder
Hi,

i convertet my ovirt 4.5 tu ovirt 4.5 hyperconverged.
(activate on cluster the gluster service and reinstalled all hosts)

I can create bricks on hosts and i'm able to create a Volume. All with the 
Ovirt GUI.

But if i want to activate the Volume, i get the error mesage that no host with 
running gluster is found in my cluster.
All hosts show message that ovirt think glusterd ist not running, but its 
running on all  hosts.

What can i du to use gluster with the lokal storage on each host?

Greatings Michael
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/G255YU5ZDBS66QFF3GKX2QHP3NYY55ZR/


[ovirt-users] gluster heal success but a directory doesn't heal

2022-06-20 Thread Diego Ercolani
Hello list, I have a problem derived from some hangs in ovirt during upgrade 
procedures, I have a gluster based self hosted engine deploy with "glen" as the 
gluster based hosted engine volume:
This is the situation I'm facing:
[root@ovirt-node3 master]# gluster volume heal glen info
Brick ovirt-node2.ovirt:/brickhe/glen
/3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks 
Status: Connected
Number of entries: 1

Brick ovirt-node3.ovirt:/brickhe/glen
/3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks 
Status: Connected
Number of entries: 1

Brick ovirt-node4.ovirt:/dati/glen <- arbiter
/3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks 
Status: Connected
Number of entries: 1

so, as manual suggest, I issued a heal operation:
[root@ovirt-node3 master]# gluster volume heal glen
Launching heal operation to perform index self heal on volume glen has been 
successful 
Use heal info commands to check status.

The Heal operation produces no results as the successive heal info report the 
same from wich I started.
But this is the situation in the log files:
[root@ovirt-node2 ~]# less /var/log/glusterfs/glfsheal-glen.log <- no errors

[root@ovirt-node3 ~]# less /var/log/glusterfs/glfsheal-glen.log <- inside the 
log I have error entries:
[2022-06-20 07:33:05.891367 +] W [MSGID: 114031] 
[client-rpc-fops_v2.c:2620:client4_0_lookup_cbk] 0-glen-client-2: remote 
operation failed. [{path=}, 
{gfid=44d74dba-19e8-47a3-89e8-f4a6cb37d5ec}, {errno=2}, {error=No such file or 
directory}] 
[root@ovirt-node4 ~]# less /var/log/glusterfs/glfsheal-glen.log <- same kind of 
errors
[2022-06-20 07:27:10.486822 +] W [MSGID: 114031] 
[client-rpc-fops_v2.c:2620:client4_0_lookup_cbk] 0-glen-client-1: remote 
operation failed. [{path=}, 
{gfid=b7b1fec5-8246-46eb-afde-ba06f52897d2}, {errno=2}, {error=No such file or 
directory}] 

On the nodes the glen volume is correctly mounted and:

[root@ovirt-node2 localhost:_glen]# ls -l 
3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks/
[root@ovirt-node3 localhost:_glen]# ls -l 
3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks/
[root@ovirt-node4 localhost:_glen]# ls -l 
3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks/
all return no files

and issuing a ls on the brick source:
[root@ovirt-node2 glen]# ls -l 
3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks/
total 0
[root@ovirt-node3 glen]# ls -l 
3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks/
total 0
drwxr-xr-x. 2 vdsm kvm 156 Jun  9 17:20 ccb6fd19-1b67-42b9-a032-31e12d62ed0e
[root@ovirt-node4 glen]# ls -l 
3577c21e-f757-4405-97d1-0f827c9b4e22/master/tasks/
total 0

so it turn out there is some difference between bricks...

Can you please help address this issue?
Thank you



 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/26DOSMM7SKLAFG6CLPXVLL44RPRUE7ZD/


[ovirt-users] Gluster storage and TRIM VDO

2022-03-29 Thread Oleh Horbachov
Hello everyone. I have a Gluster distributed replication cluster deployed. The 
cluster - store for ovirt. For bricks - VDO over a raw disk. When discarding 
via 'fstrim -av' the storage hangs for a few seconds and the connection is 
lost. Does anyone know the best practices for using TRIM with VDO in the 
context of ovirt?
ovirt - v4.4.10
gluster - v8.6
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UCTN2ZIG3EDVUU5COPXLMOH2T6WHTPBB/


[ovirt-users] gluster in ovirt-node in 4.5

2022-03-24 Thread Yedidyah Bar David
Hi all,

In relation to a recent question here (thread "[ovirt-devel] [ANN]
Schedule for oVirt 4.5.0"), we are now blocked with the following
chain of changes/dependencies:

1. ovirt-ansible-collection recently moved from ansible-2.9 to
ansible-core 2.12.
2. ovirt-hosted-engine-setup followed it.
3. ovirt-release-host-node (the package including dependencies for
ovirt-node) requires gluster-ansible-roles.
4. gluster-ansible-roles right now requires 'ansible >= 2.9' (not
core), and I only checked one of its dependencies,
gluster-ansible-infra, and this one requires 'ansible >= 2.5'.
5. ansible-core does not 'Provide: ansible', IIUC intentionally.

So we should do one of:

1. Fix gluster-ansible* packages to work with ansible-core 2.12.

2. Only patch gluster-ansible* packages to require ansible-core,
without making sure they actually work with it. This will satisfy all
deps (I guess), make the thing installable, but will likely break when
actually used. Not sure it's such a good option, but nonetheless
relevant. Might make sense if someone is going to work on (1.) soon
but not immediately. This is what would have happened in practice, if
ansible-core would have 'Provide:'-ed ansible.

3. Patch ovirt-release-host-node to not require gluster-ansible*
anymore. This means it will not be included in ovirt-node. Users that
will want to use it will have to install the dependencies manually,
somehow, presumably after (1.) is done independently.

Our team (RHV integration) does not have capacity for (1.). I intend
to do (3.) very soon, unless we get volunteers for doing (1.) or
strong voices for (2.).

Best regards,
-- 
Didi
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GRHTPVIVTOFNONE32AIG4TXNXDFH535L/


[ovirt-users] Gluster issue with brick going down

2022-03-21 Thread Chris Adams
I have a hyper-converged cluster running oVirt 4.4.10 and Gluster 8.6.
Periodically, one brick of one volume will drop out, but it's seemingly
random as to which volume and brick is affected.  All I see in the brick
log is:

[2022-03-19 13:27:36.360727] W [MSGID: 113075] 
[posix-helpers.c:2135:posix_fs_health_check] 0-vmstore-posix: 
aio_read_cmp_buf() on /gluster_bricks/vmstore/vmstore/.glusterfs/health_check 
returned ret is -1 error is Structure needs cleaning 
[2022-03-19 13:27:36.361160] M [MSGID: 113075] 
[posix-helpers.c:2214:posix_health_check_thread_proc] 0-vmstore-posix: 
health-check failed, going down 
[2022-03-19 13:27:36.361395] M [MSGID: 113075] 
[posix-helpers.c:2232:posix_health_check_thread_proc] 0-vmstore-posix: still 
alive! -> SIGTERM

Searching around, I see references to similar issues, but no real
solutions.  I see a suggestion that changing the health-check-interval
from 10 to 30 seconds helps, but it looks like 30 seconds is the default
with this version of Gluster (and I don't see it explicitly set for any
of my volumes).

While "Structure needs cleaning" appears to be an XFS filesystem error,
I don't see any XFS errors from the kernel.

This is a low I/O cluster - the storage network is on two 10 gig
switches with a two-port LAG to each server, but typically is only
seeing a few tens of megabits per second.

-- 
Chris Adams 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ACE5G25RRGOE4MADK4MYJJFAIDP5BZCJ/


[ovirt-users] Gluster Performance issues

2022-02-23 Thread Alex Morrison
Hello All,

We have 3 servers with a raid 50 array each, we are having extreme
performance issues with our gluster, writes on gluster seem to take at
least 3 times longer than on the raid directly. Can this be improved? I've
read through several other performance issues threads but have been unable
to make any improvements

"gluster volume info" and "gluster volume profile vmstore info" is below

=

-Inside Gluster - test took 35+ hours:
[root@ovirt1 1801ed24-5b55-4431-9813-496143367f66]# bonnie++ -d . -s 600G
-n 0 -m TEST -f -b -u root
Using uid:0, gid:0.
Writing intelligently...done
Rewriting...done
Reading intelligently...done
start 'em...done...done...done...done...done...
Version  1.98   --Sequential Output-- --Sequential Input-
--Random-
-Per Chr- --Block-- -Rewrite- -Per Chr- --Block--
--Seeks--
Name:Size etc/sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec
%CP
TEST   600G   35.7m  17 5824k   7112m  13 182.7
  6
Latency5466ms   12754ms  3499ms
 1589ms

1.98,1.98,TEST,1,1644359706,600G,,8192,5,,,36598,17,5824,7,,,114950,13,182.7,6,,,5466ms,12754ms,,3499ms,1589ms,,

=

-Outside Gluster - test took 18 minutes:
[root@ovirt1 1801ed24-5b55-4431-9813-496143367f66]# bonnie++ -d . -s 600G
-n 0 -m TEST -f -b -u root
Using uid:0, gid:0.
Writing intelligently...done
Rewriting...done
Reading intelligently...done
start 'em...done...done...done...done...done...
Version  1.98   --Sequential Output-- --Sequential Input-
--Random-
-Per Chr- --Block-- -Rewrite- -Per Chr- --Block--
--Seeks--
Name:Size etc/sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec
%CP
TEST   600G567m  78  149m  30307m  37  83.0
 57
Latency 205ms4630ms  1450ms
679ms

1.98,1.98,TEST,1,1648288012,600G,,8192,5,,,580384,78,152597,30,,,314533,37,83.0,57,,,205ms,4630ms,,1450ms,679ms,,

=

[root@ovirt1 1801ed24-5b55-4431-9813-496143367f66]# gluster volume info
Volume Name: engine
Type: Replicate
Volume ID: 7ed15c5a-f054-450c-bac9-3ad1b4e5931b
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: ovirt1-storage.dgi:/gluster_bricks/engine/engine
Brick2: ovirt2-storage.dgi:/gluster_bricks/engine/engine
Brick3: ovirt3-storage.dgi:/gluster_bricks/engine/engine
Options Reconfigured:
cluster.granular-entry-heal: enable
performance.strict-o-direct: on
network.ping-timeout: 30
storage.owner-gid: 36
storage.owner-uid: 36
server.event-threads: 4
client.event-threads: 4
cluster.choose-local: off
user.cifs: off
features.shard: on
cluster.shd-wait-qlength: 1
cluster.shd-max-threads: 8
cluster.locking-scheme: granular
cluster.data-self-heal-algorithm: full
cluster.server-quorum-type: server
cluster.quorum-type: auto
cluster.eager-lock: enable
network.remote-dio: off
performance.low-prio-threads: 32
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
transport.address-family: inet
storage.fips-mode-rchecksum: on
nfs.disable: on
performance.client-io-threads: on
diagnostics.latency-measurement: on
diagnostics.count-fop-hits: on

Volume Name: vmstore
Type: Replicate
Volume ID: 2670ff29-8d43-4610-a437-c6ec2c235753
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: ovirt1-storage.dgi:/gluster_bricks/vmstore/vmstore
Brick2: ovirt2-storage.dgi:/gluster_bricks/vmstore/vmstore
Brick3: ovirt3-storage.dgi:/gluster_bricks/vmstore/vmstore
Options Reconfigured:
cluster.granular-entry-heal: enable
performance.strict-o-direct: on
network.ping-timeout: 20
storage.owner-gid: 36
storage.owner-uid: 36
server.event-threads: 4
client.event-threads: 4
cluster.choose-local: off
user.cifs: off
features.shard: on
cluster.shd-wait-qlength: 1
cluster.shd-max-threads: 8
cluster.locking-scheme: granular
cluster.data-self-heal-algorithm: full
cluster.server-quorum-type: server
cluster.quorum-type: auto
cluster.eager-lock: enable
network.remote-dio: off
performance.low-prio-threads: 32
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
transport.address-family: inet
storage.fips-mode-rchecksum: on
nfs.disable: on
performance.client-io-threads: on
diagnostics.latency-measurement: on
diagnostics.count-fop-hits: on
server.tcp-user-timeout: 20
server.keepalive-time: 10
server.keepalive-interval: 2
server.keepalive-count: 5
cluster.lookup-optimize: off


[ovirt-users] gluster and virtualization

2022-02-02 Thread eevans
My setup is 3 ovirt nodes that run gluster independently of the engine server, 
even thought the engine still controls it. So 4 nodes, one engine and 3 
clustered nodes.
This has been and running with no issues except this:
But now my arbiter node will not load the gluster drive when virtualization is 
enable in the BIOS. I've been scratching my head on this and need some 
direction.
I am attaching the error.

https://1drv.ms/u/s!AvgvEzKKSZHbhMRQmUHDvv_Xv7dkhw?e=QGdfYR

Keep in mind, this error does not occur is VT is turned off..it boots normally. 

Thanks in advance.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HXRDM6W3IRTSUK46FYZZR4JRR766B2AX/


[ovirt-users] Gluster Hook differences between fresh and old clusters

2022-01-09 Thread Strahil Nikolov via Users
Hi All,

recently I have migrated from 4.3.10 to 4.4.9 and it seems something odd is 
happening.

Symptoms:
- A lot of warnings for Gluster hook discrepancies
- Trying to refresh the hooks via the sync button fails (engine error: 
https://justpaste.it/827zo )
- Existing "Default" cluster tracks more hooks than a fresh new cluster 
New cluster hooks: http://i.imgur.com/FEL2Z1D.png
Migrated cluster: https://i.imgur.com/L8dWYZY.png

What can I do to resolve the issue ? I've tried to resync the hooks, move away 
/var/lib/glusterd/hooks/1/ and reinstall gluster packages, try to resolve via 
the "Resolve Conflicts" in the UI and nothing helped so far.


Best Regards,
Strahil Nikolov
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RYSNQTAGXEAX2O677ELEAYRXDAUX52IQ/


[ovirt-users] Gluster Install Fail again :(

2021-10-29 Thread admin
I have been working on getting this up and running for about a week now and I 
am totally frustrated.  I am not sure even where to begin.  Here is the error I 
get when it fails,

TASK [gluster.features/roles/gluster_hci : Create the GlusterFS volumes] ***

An exception occurred during task execution. To see the full traceback, use 
-vvv. The error was: NoneType: None
failed: [storage1.private.net] (item={'volname': 'engine', 'brick': 
'/gluster_bricks/engine/engine', 'arbiter': 0}) => {"ansible_loop_var": "item", 
"changed": false, "item": {"arbiter": 0, "brick": 
"/gluster_bricks/engine/engine", "volname": "engine"}, "msg": "error running 
gluster (/usr/sbin/gluster --mode=script volume create engine replica 3 
transport tcp storage1.private.net:/gluster_bricks/engine/engine 
storage2.private.net:/gluster_bricks/engine/engine 
storage3.private.net:/gluster_bricks/engine/engine force) command (rc=1): 
volume create: engine: failed: Staging failed on storage3.private.net. Error: 
Host storage1.private.net not connected\nStaging failed on 
storage2.private.net. Error: Host storage1.private.net not connected\n"}

An exception occurred during task execution. To see the full traceback, use 
-vvv. The error was: NoneType: None
failed: [storage1.private.net] (item={'volname': 'data', 'brick': 
'/gluster_bricks/data/data', 'arbiter': 0}) => {"ansible_loop_var": "item", 
"changed": false, "item": {"arbiter": 0, "brick": "/gluster_bricks/data/data", 
"volname": "data"}, "msg": "error running gluster (/usr/sbin/gluster 
--mode=script volume create data replica 3 transport tcp 
storage1.private.net:/gluster_bricks/data/data 
storage2.private.net:/gluster_bricks/data/data 
storage3.private.net:/gluster_bricks/data/data force) command (rc=1): volume 
create: data: failed: Staging failed on storage2.private.net. Error: Host 
storage1.private.net not connected\nStaging failed on storage3.private.net. 
Error: Host storage1.private.net not connected\n"}

An exception occurred during task execution. To see the full traceback, use 
-vvv. The error was: NoneType: None
failed: [storage1.private.net] (item={'volname': 'vmstore', 'brick': 
'/gluster_bricks/vmstore/vmstore', 'arbiter': 0}) => {"ansible_loop_var": 
"item", "changed": false, "item": {"arbiter": 0, "brick": 
"/gluster_bricks/vmstore/vmstore", "volname": "vmstore"}, "msg": "error running 
gluster (/usr/sbin/gluster --mode=script volume create vmstore replica 3 
transport tcp storage1.private.net:/gluster_bricks/vmstore/vmstore 
storage2.private.net:/gluster_bricks/vmstore/vmstore 
storage3.private.net:/gluster_bricks/vmstore/vmstore force) command (rc=1): 
volume create: vmstore: failed: Staging failed on storage3.private.net. Error: 
Host storage1.private.net not connected\nStaging failed on 
storage2.private.net. Error: Host storage1.private.net is not in 'Peer in 
Cluster' state\n"}

Here are the facts.

using 4.4.9 of ovirt.
using ovirtnode os 
partion for gluster:  /dev/vda4  > 4T in unformatted space.

able to ssh into each host on the private.net and known hosts and fqdn passes 
fine.

On the volume  page:
all default settings.

On the bricks page:
JBOD / Blacklist true / storage host  storage1.private.net / default lvm except 
the device is /dev/sda4 

I really need to get this setup.  The first failure was the filter error, so I 
edited the /etc/lvm/lvm.conf  to comment out the filter line.  Then without 
doing a clean up I reran the deployment and got the above error.  

Thanks in advance
Brad

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SAYZ3STV3ILDE42T6JUXLKVHSIX7LRI5/


[ovirt-users] gluster 5834 Unsynced entries present

2021-10-01 Thread Dominique D
yesterday I had a glich and my second ovnode2 server restarted 

here are some errors in the events :

VDSM ovnode3.telecom.lan command SpmStatusVDS failed: Connection timeout for 
host 'ovnode3.telecom.lan', last response arrived 2455 ms ago.
Host ovnode3.telecom.lan is not responding. It will stay in Connecting state 
for a grace period of 86 seconds and after that an attempt to fence the host 
will be issued.
Invalid status on Data Center Default. Setting Data Center status to Non 
Responsive (On host ovnode3.telecom.lan, Error: Network error during 
communication with the Host.).
Executing power management status on Host ovnode3.telecom.lan using Proxy Host 
ovnode1.telecom.lan and Fence Agent ipmilan:10.5.1.16.

Now my 3 bricks have errors from my gluster volume 


[root@ovnode2 ~]# gluster volume status
Status of volume: datassd
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick ovnode1s.telecom.lan:/gluster_bricks/
datassd/datassd 49152 0  Y   4027
Brick ovnode2s.telecom.lan:/gluster_bricks/
datassd/datassd 49153 0  Y   2393
Brick ovnode3s.telecom.lan:/gluster_bricks/
datassd/datassd 49152 0  Y   2347
Self-heal Daemon on localhost   N/A   N/AY   2405
Self-heal Daemon on ovnode3s.telecom.lanN/A   N/AY   2366
Self-heal Daemon on 172.16.70.91N/A   N/AY   4043

Task Status of Volume datassd
--
There are no active volume tasks


gluster volume heal datassd info | grep -i "Number of entries:" | grep -v 
"entries: 0"
Number of entries: 5759

in the webadmin all the bricks are green with comments for two : 

ovnode1 Up, 5834 Unsynced entries present
ovnode2 Up,
ovnode3 Up, 5820 Unsynced entries present

I tried this without success 

gluster volume heal datassd 
Launching heal operation to perform index self heal on volume datassd has been 
unsuccessful:
Glusterd Syncop Mgmt brick op 'Heal' failed. Please check glustershd log file 
for details.

What are the next steps ? 

Thank you
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QRI2K34O2X3NEEYLWTZJYG26EYH6CJQU/


[ovirt-users] Gluster bricks error

2021-07-21 Thread Patrick Lomakin
For a long time I have been seeing the same error, which cannot be 
corrected.After restarting the host which has a Volume of one or more Bricks, 
the Volume starts with the status "Online", but the Bricks remain "Offline". 
This leads to having to manually restart Volume, the ovirt-ha and ovirt-broker 
services, and run the hosted-engine --connect-storage command. And only after 
that I can start the hosted engine back to normal. I tried this on different 
server hardware and different operating systems for the host, but the result is 
the same. This is a very serious flaw that nullifies the high availability in 
HCI using GlusterFS.Regards!
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/O6LA5WAZPQDVDVKBQM4EYA7MHBEZDGZ6/


[ovirt-users] Gluster deploy error!

2021-07-09 Thread Patrick Lomakin
Hello! I have tried to deploy single node vith gluster, but if select 
"Compression and dedublication" I get an error:

TASK [gluster.infra/roles/backend_setup : Create thick logical volume] *
failed: [host01] (item={'vgname': 'gluster_vg_sda4', 'lvname': 
'gluster_lv_engine', 'size': '1970G'}) => {"ansible_index_var": "index", 
"ansible_loop_var": "item", "changed": false, "err": "  Volume group 
\"gluster_vg_sda4\" has insufficient free space (504319 extents): 504320 
required.\n", "index": 0, "item": {"lvname": "gluster_lv_engine", "size": 
"1970G", "vgname": "gluster_vg_sda4"}, "msg": "Creating logical volume 
'gluster_lv_engine' failed", "rc": 5}
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4WV4AYMIY53LFO7WRBM6LA3TVMTZW25C/


[ovirt-users] Gluster Geo-Replication Fails

2021-05-17 Thread simon
Hi All,

I have to sites each with 2 x 3 Node clusters. oVirt 4.4.5 with GlusterFS 8.4.
On Site A I have a Replica 3 + Arbiter volume that I want to geo-replicate to 
Site B.
I've setup a Replica 3 + Arbiter volume on Site B.
All password-less ssh requirements are in place and the Geo-Replication Session 
has been created successfully.
I issue the 'Start' command successfully but the status switches between 
'Initializing', 'Active' & 'Failed' within 20 seconds constantly.
Can anyone point me in the direction to get this up and running as it is our 
proposed DR solution.

Kind Regards

Shimme
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MHWPJRPDPOQMUFCNIRTU4XSEDXUWCOVW/


[ovirt-users] Gluster volumes not healing (perhaps after host maintenance?)

2021-04-24 Thread David White via Users
I discovered that the servers I purchased did not come with 10Gbps network 
cards, like I thought they did. So my storage network has been running on a 
1Gbps connection for the past week, since I deployed the servers into the 
datacenter a little over a week ago. I purchased 10Gbps cards, and put one of 
my hosts into maintenance mode yesterday, prior to replacing the daughter card. 
It is now back online running fine on the 10Gbps card.

All VMs seem to be working, even when I migrate them onto cha2, which is the 
host I did maintenance on yesterday morning.
The other two hosts are still running on the 1Gbps connection, but I plan to do 
maintenance on them next week.

The oVirt manager shows that all 3 hosts are up, and that all of my volumes - 
and all of my bricks - are up. However, every time I look at the storage, it 
appears that the self-heal info for 1 of the volumes is 10 minutes, and the 
self-heal info for another volume is 50+ minutes.

This morning is the first time in the last couple of days that I've paid close 
attention to the numbers, but I don't see them going down.

When I log into each of the hosts, I do see everything is connected in gluster.
It is interesting to me, in this particular case, though that gluster on cha3 
notices the hostname of 10.1.0.10 to be the IP address, and not the hostname 
(cha1).
The host that I did the maintenance on is cha2.

[root@cha3-storage dwhite]# gluster peer statusNumber of Peers: 2Hostname: 
10.1.0.10Uuid: 87a4f344-321a-48b9-adfb-e3d2b56b8e7bState: Peer in Cluster 
(Connected)Hostname: cha2-storage.mgt.barredowlweb.comUuid: 
93e12dee-c37d-43aa-a9e9-f4740b9cab14State: Peer in Cluster (Connected)

When I run `gluster volume heal data`, I see the following:
[root@cha3-storage dwhite]# gluster volume heal data
Launching heal operation to perform index self heal on volume data has been 
unsuccessful:
Commit failed on cha2-storage.mgt.barredowlweb.com. Please check log file for 
details.

I get the same results if I run the command on cha2, for any volume:
[root@cha2-storage dwhite]# gluster volume heal data
Launching heal operation to perform index self heal on volume data has been 
unsuccessful:
Glusterd Syncop Mgmt brick op 'Heal' failed. Please check glustershd log file 
for details.
[root@cha2-storage dwhite]# gluster volume heal vmstore
Launching heal operation to perform index self heal on volume vmstore has been 
unsuccessful:
Glusterd Syncop Mgmt brick op 'Heal' failed. Please check glustershd log file 
for details.

I see a lot of stuff like this on cha2 /var/log/glusterfs/glustershd.log:
[2021-04-24 11:33:01.319888] I [rpc-clnt.c:1975:rpc_clnt_reconfig] 
2-engine-client-0: changing port to 49153 (from 0)[2021-04-24 11:33:01.329463] 
I [MSGID: 114057] [client-handshake.c:1128:select_server_supported_programs] 
2-engine-client-0: Using Program [{Program-name=GlusterFS 4.x v1}, 
{Num=1298437}, {Version=400}][2021-04-24 11:33:01.330075] W [MSGID: 114043] 
[client-handshake.c:727:client_setvolume_cbk] 2-engine-client-0: failed to set 
the volume [{errno=2}, {error=No such file or directory}][2021-04-24 
11:33:01.330116] W [MSGID: 114007] 
[client-handshake.c:752:client_setvolume_cbk] 2-engine-client-0: failed to get 
from reply dict [{process-uuid}, {errno=22}, {error=Invalid 
argument}][2021-04-24 11:33:01.330140] E [MSGID: 114044] 
[client-handshake.c:757:client_setvolume_cbk] 2-engine-client-0: SETVOLUME on 
remote-host failed [{remote-error=Brick not found}, {errno=2}, {error=No such 
file or directory}][2021-04-24 11:33:01.330155] I [MSGID: 114051] 
[client-handshake.c:879:client_setvolume_cbk] 2-engine-client-0: sending 
CHILD_CONNECTING event [][2021-04-24 11:33:01.640480] I 
[rpc-clnt.c:1975:rpc_clnt_reconfig] 3-vmstore-client-0: changing port to 49154 
(from 0)The message "W [MSGID: 114007] 
[client-handshake.c:752:client_setvolume_cbk] 3-vmstore-client-0: failed to get 
from reply dict [{process-uuid}, {errno=22}, {error=Invalid argument}]" 
repeated 4 times between [2021-04-24 11:32:49.602164] and [2021-04-24 
11:33:01.649850][2021-04-24 11:33:01.649867] E [MSGID: 114044] 
[client-handshake.c:757:client_setvolume_cbk] 3-vmstore-client-0: SETVOLUME on 
remote-host failed [{remote-error=Brick not found}, {errno=2}, {error=No such 
file or directory}][2021-04-24 11:33:01.649969] I [MSGID: 114051] 
[client-handshake.c:879:client_setvolume_cbk] 3-vmstore-client-0: sending 
CHILD_CONNECTING event [][2021-04-24 11:33:01.650095] I [MSGID: 114018] 
[client.c:2225:client_rpc_notify] 3-vmstore-client-0: disconnected from client, 
process will keep trying to connect glusterd until brick's port is available 
[{conn-name=vmstore-client-0}]

How do I further troubleshoot?

Sent with ProtonMail Secure Email.

publickey - dmwhite823@protonmail.com - 0x320CD582.asc
Description: application/pgp-keys


signature.asc
Description: OpenPGP digital signature
___
Users mailing list -- users@ovirt.org
To unsubscribe 

[ovirt-users] Gluster version upgrades

2021-03-06 Thread Sketch
Is the gluster version on an oVirt host tied to the oVirt version, or 
would it be safe to upgrade to newer versions of gluster?


I have noticed gluster is often updated to new major versions on oVirt 
point release upgrades.  We have some compute+storage hosts on 4.3.6 which 
can't be upgraded easily at the moment, but we are having some gluster 
issues that appear to be due to bugs that I wonder if upgrading might 
help.  Would an in-place upgrade of gluster be a bad idea without also 
updating oVirt?

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XR47MTPOQ6XTPT7TOH6LGEWYCH2YKRS2/


[ovirt-users] Gluster volume engine stuck in healing with 1 unsynched entry & HostedEngine paused

2021-02-27 Thread souvaliotimaria

Hello everyone,

Any help would be greatly appreciated in the following problem.

In my lab, the day before yesterday, we had power issues, with a UPS going 
off-line and following the power outage of the NFS/DNS server I have set up to 
serve ovirt with isos and as a DNS server (our other DNS servers are located as 
VMs within the oVirt environment). We found a broadcast storm on the switch 
(due to a faulty NIC on the aformentioned UPS) that the ovirt nodes are 
connected and later on had to re-establish several of the virtual connections 
as well. The above led to one of the hosts becoming NonResponsive, two machines 
becoming unresponsive and three VMs shuting down. 

The oVirt environment, version 4.3.5.2, is a replica 2 + arbiter 1 environment 
and runs GlusterFS with the recommended volumes of data, engine and vmstore.

So far, the times there was some kind of a problem, usually oVirt was able to 
solve it by its own.

This time, however, after we recovered from the above state, the volumes of 
data and vmstore successfully healing , the volume engine became stuck to the 
healing process (Up, unsynched entries, needs healing), and from the web GUI I 
see that the VM HostedEngine is paused due to a storage I/O error while the 
output of virsh list --all command shows that the HostedEngine is running.. How 
is that happening?

I tried to manually trigger the healing process for the volume but nothing with 
gluster volume heal engine

The command 
gluster volume heal engine info 
shows the following 

[root@ov-no3 ~]# gluster volume heal engine info
Brick ov-no1.ariadne-t.local:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0

Brick ov-no2.ariadne-t.local:/gluster_bricks/engine/engine
/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
 
Status: Connected
Number of entries: 1

Brick ov-no3.ariadne-t.local:/gluster_bricks/engine/engine
/80f6e393-9718-4738-a14a-64cf43c3d8c2/images/d5de54b6-9f8e-4fba-819b-ebf6780757d2/a48555f4-be23-4467-8a54-400ae7baf9d7
 
Status: Connected
Number of entries: 1

This morning I came upon this Reddit post 
https://www.reddit.com/r/gluster/comments/fl3yb7/entries_stuck_in_heal_pending/ 
where it seems that after a graceful reboot one of the ovirt hosts, the gluster 
came back online after it completed the appropriate healing processes. The 
thing is from what I have read that when there are unsynched entries in the 
gluster a host cannot be put into maintenance mode so that it can be rebooted, 
correct?

Should I try to restart the glusterd service.

Could someone tell me what I should do?

Thank you all for your time and help,
Maria Souvalioti
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BSOF7BXAMVJ4IMYUEB3OBU4T64FGYA2J/


[ovirt-users] Gluster Storage

2021-01-25 Thread dkiplagat
Hi, Am new using oVirt and i would like to know if i could deploy oVirt and be 
able to use it to deploy and manage Gluster storage.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NVEUHPXAUR4O4N363DEMW2LE6GRBBNKJ/


[ovirt-users] Gluster Hyperconverged fails with single disk partitioned

2021-01-20 Thread Shantur Rathore
Hi,

I am trying to setup a single host Self-Hosted hyperconverged setup with
GlusterFS.
I have a custom partitioning where I provide 100G for oVirt and its
partitions and rest 800G to a physical partition (/dev/sda4).

When I try to create gluster deployment with the wizard, it fails

TASK [gluster.infra/roles/backend_setup : Create volume groups]

failed: [ovirt-macpro-16.lab.ced.bskyb.com] (item={'key':
'gluster_vg_sda4', 'value': [{'vgname': 'gluster_vg_sda4', 'pvname':
'/dev/sda4'}]}) => {"ansible_loop_var": "item", "changed": false, "err": "
 Device /dev/sda4 excluded by a filter.\n", "item": {"key":
"gluster_vg_sda4", "value": [{"pvname": "/dev/sda4", "vgname":
"gluster_vg_sda4"}]}, "msg": "Creating physical volume '/dev/sda4' failed",
"rc": 5}

I checked and /etc/lvm/lvm.conf filter doesn't allow /dev/sda4. It only
allows PV for onn VG.
Once I manually allow /dev/sda4 to lvm filter, it works fine and gluster
deployment completes.

Fdisk :

# fdisk -l /dev/sda
Disk /dev/sda: 931.9 GiB, 100081440 bytes, 1954210120 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 4096 bytes
I/O size (minimum/optimal): 4096 bytes / 4096 bytes
Disklabel type: gpt
Disk identifier: FE209000-85B5-489A-8A86-4CF0C91B2E7D

Device StartEndSectors   Size Type
/dev/sda1   204812308471228800   600M EFI System
/dev/sda2123084833279992097152 1G Linux filesystem
/dev/sda33328000  213043199  209715200   100G Linux LVM
/dev/sda4  213043200 1954209791 1741166592 830.3G Linux filesystem

LVS

# lvs
  LV VG  Attr   LSize  Pool  Origin
  Data%  Meta%  Move Log Cpy%Sync Convert
  home   onn Vwi-aotz-- 10.00g pool0
   0.11
  ovirt-node-ng-4.4.4-0.20201221.0   onn Vwi---tz-k 10.00g pool0 root
  ovirt-node-ng-4.4.4-0.20201221.0+1 onn Vwi-aotz-- 10.00g pool0
ovirt-node-ng-4.4.4-0.20201221.0 25.26
  pool0  onn twi-aotz-- 95.89g
   2.95   14.39
  root   onn Vri---tz-k 10.00g pool0
  swap   onn -wi-ao  4.00g
  tmponn Vwi-aotz-- 10.00g pool0
   0.12
  varonn Vwi-aotz-- 20.00g pool0
   0.92
  var_crash  onn Vwi-aotz-- 10.00g pool0
   0.11
  var_logonn Vwi-aotz-- 10.00g pool0
   0.13
  var_log_audit  onn Vwi-aotz--  4.00g pool0
   0.27



# grep filter /etc/lvm/lvm.conf
filter =
["a|^/dev/disk/by-id/lvm-pv-uuid-QrvErF-eaS9-PxbI-wCBV-3OxJ-V600-NG7raZ$|",
"r|.*|"]

Am I doing something which oVirt isn't expecting?
Is there anyway to provide tell gluster deployment to add it to lvm config.

Thanks,
Shantur
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BP7BQWG3O7IFRLU4W6ZNV4J6PHR4DUZF/


[ovirt-users] Gluster release and oVirt 4.4

2021-01-11 Thread Simon Coter
Hi,

is there any plan to introduce Gluster-8 for hyper-converged architecture with 
oVirt 4.4 ?
Just wondering because I can see Gluster-7 is declared EOL on Dec 11, 2020 
(https://www.gluster.org/release-schedule/ 
)
Thanks

Simon___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/CL533FADCSACZBJ64VR6X5JSQ2PJWPQY/


[ovirt-users] Gluster volume slower then raid1 zpool speed

2020-11-23 Thread Harry O
Hi,
Can anyone help me with the performance on my 3 node gluster on zfs (it is 
setup with one arbiter)
The performance on the single vm I have on it (with engine) is 50% worse then a 
single bare metal disk, on the writes.
I have enabled "Optimize for virt store"
I run 1Gbps 1500MTU network, could this be the write performance killer?
Is this to be expected from a 2xHDD zfs raid one on each node, with 3xNode 
arbiter setup?
Maybe I should move to raid 5 or 6?
Maybe I should add SSD cache to raid1 zfs zpools?
What are your thoughts? What to do for optimize this setup?
I would like to run zfs with gluster and I can deal with a little performance 
loss, but not that much.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SYP4I4MQDKLCIFMUSXVYCUOFNC25LNDR/


[ovirt-users] Gluster volume not responding

2020-10-08 Thread simon
Hi,

Every Monday and Wednesday morning there are gluster connectivity timeouts but 
all checks of the network and network configs are ok.


Description of problem:

The following entries were found in the engine.log following VMs becoming 
unresponsive and hosts fencing. This issue has been causing issues since the 
beginning of September and no amount of reading logs is helping. This issue 
occurs every Wednesday morning at exactly the same time.

WARN  [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] 
(EE-ManagedThreadFactory-engine-Thread-974045) [] domain 
'bc482086-598b-46b1-9189-0146fa03447c:pltfm_data03' in problem 'PROBLEMATIC'. 
vds: 'bdtpltfmovt02'
WARN  [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] 
(EE-ManagedThreadFactory-engine-Thread-974069) [] domain 
'bf807836-b64e-4913-ab41-cfe04ca9abab:pltfm_data01' in problem 'PROBLEMATIC'. 
vds: 'bdtpltfmovt02'
WARN  [org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] 
(EE-ManagedThreadFactory-engine-Thread-974121) [] domain 
'bc482086-598b-46b1-9189-0146fa03447c:pltfm_data03' in problem 'PROBLEMATIC'. 
vds: 'bdtpltfmovt03'


INFO  [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
(EE-ManagedThreadFactory-engineScheduled-Thread-78) [] VM 
'00a082de-c827-4c97-9846-ec32d1ddbfa6'(bdtfmnpproddb03) moved from 'Up' --> 
'NotResponding'
INFO  [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
(EE-ManagedThreadFactory-engineScheduled-Thread-78) [] VM 
'f5457f04-054e-4684-9702-40ed4a3e4bdb'(bdtk8shaproxy02) moved from 'Up' --> 
'NotResponding'
INFO  [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
(EE-ManagedThreadFactory-engineScheduled-Thread-78) [] VM 
'bea85b27-18e7-4936-9871-cdb987baebdd'(bdtdepjump) moved from 'Up' --> 
'NotResponding'
INFO  [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
(EE-ManagedThreadFactory-engineScheduled-Thread-78) [] VM 
'ba1d4fe2-97e7-491a-9485-8319281e7784'(bdtcmgmtnfs01) moved from 'Up' --> 
'NotResponding'
INFO  [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
(EE-ManagedThreadFactory-engineScheduled-Thread-78) [] VM 
'9e253848-7153-43e8-8126-dba2d7f2d214'(bdtdepnfs01) moved from 'Up' --> 
'NotResponding'
INFO  [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] 
(EE-ManagedThreadFactory-engineScheduled-Thread-78) [] VM 
'1e58d106-4b65-4296-8d11-2142abb7808e'(bdtionjump) moved from 'Up' --> 
'NotResponding'

VDSM Log from one of the Gluster Peers:

[2020-10-05 03:03:25.038883] W [MSGID: 114031] 
[client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-pltfm_data02-client-0: 
remote operation failed. Path: / (----0001) 
[Transport endpoint is not connected]
[2020-10-07 05:37:39.582138] C 
[rpc-clnt-ping.c:162:rpc_clnt_ping_timer_expired] 0-pltfm_data02-client-0: 
server x.x.x.x:49153 has not responded in the last 30 seconds, disconnecting.
[2020-10-07 05:37:39.583217] I [MSGID: 114018] 
[client.c:2288:client_rpc_notify] 0-pltfm_data02-client-0: disconnected from 
pltfm_data02-client-0. Client process will keep trying to connect to glusterd 
until brick's port is available
[2020-10-07 05:37:39.584213] E [rpc-clnt.c:346:saved_frames_unwind] (--> 
/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7fe83aa96fbb] (--> 
/lib64/libgfrpc.so.0(+0xce11)[0x7fe83a85fe11] (--> 
/lib64/libgfrpc.so.0(+0xcf2e)[0x7fe83a85ff2e] (--> 
/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x91)[0x7fe83a861521] (--> 
/lib64/libgfrpc.so.0(+0xf0c8)[0x7fe83a8620c8] ) 0-pltfm_data02-client-0: 
forced unwinding frame type(GlusterFS 4.x v1) op(LOOKUP(27)) called at 
2020-10-07 05:37:09.003907 (xid=0x7e6a8c)

Current Version: 4.3.4.3-1.el7 - Although we are keen to upgrade, we need 
stability for this production environment before doing so.

This Data Center has 2 x 3 Node clusters (Admin & Platform) which each have a 3 
Replica Gluster configuration which is not managed by the self hosted ovirt 
engine.

Any assistance is appreciated.

Regards

Shimme
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DKR32J3XTHH5TJ6YFMKB2DJJDOG5PM2X/


[ovirt-users] Gluster Volumes - Correct Peer Connection

2020-09-23 Thread Jeremey Wise
I just noticed when HCI setup bult the gluster engine / data / vmstore
volumes... it did use correctly the definition of 10Gb "back end"
interfaces / hosts.

But..  oVirt Engine is NOT referencing this.
it lists bricks as 1Gb "managment / host" interfaces.  Is this a GUI
issue?  I doubt this and how do I correct it?
###  Data Volume Example
Name:
data
Volume ID:
0ae7b487-8b87-4192-bd30-621d445902fe
Volume Type:
Replicate
Replica Count:
3
Number of Bricks:
3
Transport Types:
TCP
Maximum no of snapshots:
256
Capacity:
999.51 GiB total, 269.02 GiB used, 730.49 GiB free, 297.91 GiB Guaranteed
free, 78 Deduplication/Compression savings (%)


medusa.penguinpages.local
medusa.penguinpages.local:/gluster_bricks/data/data
25%
OK
odin.penguinpages.local
odin.penguinpages.local:/gluster_bricks/data/data
25%
OK
thor.penguinpages.local
thor.penguinpages.local:/gluster_bricks/data/data
25%
OK


# I have storage back end of 172.16.101.x  which is 10Gb dedicated for
replication. Peers reflect this
[root@odin c4918f28-00ce-49f9-91c8-224796a158b9]# gluster peer status
Number of Peers: 2

Hostname: thorst.penguinpages.local
Uuid: 7726b514-e7c3-4705-bbc9-5a90c8a966c9
State: Peer in Cluster (Connected)

Hostname: medusast.penguinpages.local
Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
State: Peer in Cluster (Connected)
[root@odin c4918f28-00ce-49f9-91c8-224796a158b9]#



-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ULE66KK5UEGM5GTRG6IMWZLUEI6JLHVI/


[ovirt-users] Gluster Domain Storage full

2020-09-21 Thread suporte
Hello, 

I'm running oVirt Version 4.3.4.3-1.el7. 
I have a small GlusterFS Domain storage brick on a dedicated filesystem serving 
only one VM. 
The VM filled all the Domain storage. 
The Linux filesystem has 4.1G available and 100% used, the mounted brick has 
0GB available and 100% used 

I can not do anything with this disk, for example, if I try to move it to 
another Gluster Domain Storage get the message: 

Error while executing action: Cannot move Virtual Disk. Low disk space on 
Storage Domain 

Any idea? 

Thanks 

-- 

Jose Ferradeira 
http://www.logicworks.pt 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WFN2VOQZPPVCGXAIFEYVIDEVJEUCSWY7/


[ovirt-users] Gluster quorum issue on 3-node HCI with extra 5-nodes as compute and storage nodes

2020-09-14 Thread thomas
Yes, I've also posted this on the Gluster Slack. But I am using Gluster mostly 
because it's part of oVirt HCI, so don't just send me away, please!

Problem: GlusterD refusing to start due to quorum issues for volumes where it 
isn’t contributing any brick

(I've had this before on a different farm, but there it was transitory. Now I 
have it in a more observable manner, that's why I open a new topic)

In a test farm with recycled servers, I started running Gluster via oVirt 
3node-HCI, because I got 3 machines originally.
They were set up as group A in a 2:1 (replica:arbiter) oVirt HCI setup with 
'engine', 'vmstore' and 'data' volumes, one brick on each node.

I then got another five machines with hardware specs that were rather different 
to group A, so I set those up as group B to mostly act as compute nodes, but 
also to provide extra storage, mostly to be used externally as GlusterFS 
shares. It took a bit of fiddling with Ansible but I got these 5 nodes to serve 
two more Gluster volumes 'tape' and 'scratch' using dispersed bricks (4 
disperse:1 redundancy), RAID5 in my mind.

The two groups are in one Gluster, not because they serve bricks to the same 
volumes, but because oVirt doesn't like nodes to be in different Glusters (or 
actually, to already be in a Gluster when you add them as host node). But the 
two groups provide bricks to distinct volumes, there is no overlap.

After setup things have been running fine for weeks, but now I needed to 
restart a machine from group B, which has ‘tape’ and ‘scratch’ bricks, but none 
from original oVirt ‘engine’, ‘vmstore’ and ‘data’ in group A. Yet the gluster 
daemon refuses to start, citing a loss of quorum for these three volumes, even 
if it has no bricks in them… which makes no sense to me.

I am afraid the source of the issue is concept issues: I clearly don't really 
understand some design assumptions of Gluster.
And I'm afraid the design assumptions of Gluster and of oVirt (even with HCI), 
are not as related as one might assume from the marketing materials on the 
oVirt home-page.

But most of all I'd like to know: How do I fix this now?

I can't heal 'tape' and 'scratch', which are growing ever more apart while the 
glusterd on this machine in group B refuses to come online for lack of a quorum 
on volumes where it is not contributing bricks.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives:


[ovirt-users] Gluster Name too long

2020-09-12 Thread Jeremey Wise
Deployment on three node cluster using oVirt HCI wizard.

I think this is a bug where it needs to do either a pre-flight name length
validation, or increase valid field length.


I avoid using /dev/sd#   as those can change.  And the wizard allows for
this change to a more explicit devices Ex:
/dev/mapper/Samsung_SSD_850_PRO_512GB_S250NXAGA15787L


Error:
TASK [gluster.infra/roles/backend_setup : Create a LV thinpool for similar
device types] ***
task path:
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thin_pool_create.yml:239
failed: [thorst.penguinpages.local] (item={'vgname':
'gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L', 'thinpoolname':
'gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L',
'poolmetadatasize': '3G'}) => {"ansible_loop_var": "item", "changed":
false, "err": "  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tmeta\"
is too long.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tmeta\"
length 130 is not supported.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
length 130 is not supported.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tmeta\"
length 130 is not supported.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
length 130 is not supported.\n", "item": {"poolmetadatasize": "3G",
"thinpoolname":
"gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L",
"vgname": "gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L"}, "msg":
"Creating logical volume 'None' failed", "rc": 5}
failed: [medusast.penguinpages.local] (item={'vgname':
'gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306', 'thinpoolname':
'gluster_thinpool_gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306',
'poolmetadatasize': '3G'}) => {"ansible_loop_var": "item", "changed":
false, "err": "  Internal error: LV name
\"gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306/gluster_thinpool_gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306\"
length 130 is not supported.\n", "item": {"poolmetadatasize": "3G",
"thinpoolname":
"gluster_thinpool_gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306",
"vgname": "gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306"},
"msg": "Creating logical volume 'None' failed", "rc": 5}
changed: [odinst.penguinpages.local] => (item={'vgname':
'gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137', 'thinpoolname':
'gluster_thinpool_gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137',
'poolmetadatasize': '3G'}) => {"ansible_loop_var": "item", "changed": true,
"item": {"poolmetadatasize": "3G", "thinpoolname":
"gluster_thinpool_gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137",
"vgname": "gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137"}, "msg": ""}


I will revert back to  /dev/sd# for now... but this should be cleaned up.

Attached is YAML file for deployment of cluster

-- 
p enguinpages
hc_nodes:
  hosts:
thorst.penguinpages.local:
  gluster_infra_volume_groups:
- vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  pvname: /dev/mapper/vdo_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  gluster_infra_mount_devices:
- path: /gluster_bricks/engine
  lvname: gluster_lv_engine
  vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
- path: /gluster_bricks/data
  lvname: gluster_lv_data
  vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
- path: /gluster_bricks/vmstore
  lvname: gluster_lv_vmstore
  vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  gluster_infra_vdo:
- name: vdo_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  device: 

[ovirt-users] Gluster Volume Type Distributed

2020-08-27 Thread Dominique Deschênes


Hi Everyone,

I would like to use Distrbuted Volume type but the volume type is Gray out. I 
can only use the replicate type. 


It's normal ?


3 ovirt Servers 4.4.1-2020080418

Can I configure a replicate volume for the engine domain and Distributed for 
the data domain?





Thank you


Dominique Deschênes
Ingénieur chargé de projets, Responsable TI
816, boulevard Guimond, Longueuil J4G 1T5
 450 670-8383 x105  450 670-2259


          


___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7TMGYCS4EF3KDKED46BVIL7JH3H4EKSH/


[ovirt-users] Gluster remount?

2020-06-11 Thread Christoph Köhler

Hey,

it happens in our ovirt 4.3.9 occasionally that after reboot one of the 
dedicated gluster nodes (replica3, arbiter 1) some client connections 
are missing on server side:


# gluster volume status gluvol3 clients


Client connections for volume gluvol3
--
Brick : glusrv01:/gluster/p1/brick1
Clients connected : 17
Hostname
.
.
.

Brick : glusrv02:/gluster/p1/brick1
Clients connected : 15 < two connections missing
Hostname
.
.
.
--
Brick : glusrv03:/gluster/p1/brick1
Clients connected : 17
Hostname
.
.
.
--

I don't know why it happes, the log files don't really help.

Now: is there a possibility on the missing nodes to reinitiate the 
gluster mount like 'mount -o remount' or so? The only way I actually 
know is to set the concerned node to maintenance and reactivate it. But 
that is really long-moving with many VMs...


Any ideas?

Thanks
Chris
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6GR6XODCVILK5B6TQ7KTL3527GM2YM6A/


[ovirt-users] Gluster error in server log.

2020-06-05 Thread Krist van Besien
Hello all,

On my ovirt HC cluster I constantly get the following kinds of errors:

From /var/log/ovirt-engine/engine.log

2020-06-05 10:38:36,652+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-15) [] 
Could not associate brick 'on1.ws.kri.st:/gluster_bricks/vmstore/vmstore' of 
volume 'dab47af2-16fc-461d-956e-daab00c8489e' with correct network as no 
gluster network found in cluster 'a8a38ffe-a499-11ea-9471-00163e5ffe63'

Normally you get these errors if you forgot to define and assign a storage 
network. However I do have a storage network, have assigned it to the interface 
used by gluster on each host, and have set it as the “gluster” network in the 
default cluster.
So why am I still getting this error?

Krist

Vriendelijke Groet | Best Regards | Freundliche Grüße | Cordialement
Krist van Besien
krist.vanbes...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WGKRR6TZCJCZB4GB23JWS4QPIX6Q34T4/


[ovirt-users] Gluster deployment fails with missing UUID

2020-04-28 Thread Shareef Jalloq
Hi,

I'm running the gluster deployment flow and am trying to use a second drive
as the gluster volume.  It's /dev/sdb on each node and I'm using the JBOD
mode.

I'm seeing the following gluster ansible task fail and a google search
doesn't bring up much.

TASK [gluster.infra/roles/backend_setup : Create volume groups]


failed: [ovirt-gluster-01.jalloq.co.uk] (item={u'vgname':
u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) => {"ansible_loop_var": "item",
"changed": false, "err": "  Couldn't find device with uuid
Y8FVs8-LP6w-R6CR-Yosh-c40j-17XP-ttP3Np.\n  Couldn't find device with uuid
tA4lpO-hM9f-S8ci-BdPh-lTve-0Rh1-3Bcsfy.\n  Couldn't find device with uuid
RG3w6j-yrxn-2iMw-ngd0-HgMS-i5dP-CGjaRk.\n  Couldn't find device with uuid
lQV02e-TUZE-PXCd-GWEd-eGqe-c2xC-pauHG7.\n  Device /dev/sdb excluded by a
filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"},
"msg": "Creating physical volume '/dev/sdb' failed", "rc": 5}
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5U3K3IPYCFOLUFJ56FGJI3TYWT6NOLAZ/


[ovirt-users] Gluster problems with new disk and device name change and overlap

2020-04-07 Thread Gianluca Cecchi
Hi,
I have configured a single host HCI environment through the GUI wizard in
4.3.9.
Initial setup has thai layout of disks, as seen by the operating system:
/dev/sda -> for ovirt-node-ng OS
/dev/nvme0n1 --> for gluster, engine and data volumes
/dev/nvme1n1 --> for gluster, vmstore volume

So far so good and all is ok.
I notice that, even with single path internal disks, at the end oVirt
configures the gluster disks as multipath devices and LVM2 PV structure on
top of the multipath devices.
Is this for "code optimization" at low level or what is the rationale for
that, as with Gluster normally you do use local disks and so single path?
Multipath structure generated:

[root@ovirt ~]# multipath -l
nvme.8086-50484b53373530353031325233373541474e-494e54454c205353 dm-5
NVME,INTEL SSDPED1K375GA
size=349G features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=0 status=active
  `- 0:0:1:0 nvme0n1 259:0 active undef running
eui.010001005cd2e4b5e7db4d51 dm-6 NVME,INTEL SSDPEDKX040T7

size=932G features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=0 status=active
  `- 2:0:1:0 nvme1n1 259:2 active undef running
[root@ovirt ~]#

Anyway, on top of the multipah devices

On /dev/nvme0n1:
gluster_vg_nvme0n1 volume group with gluster_lv_data and gluster_lv_engine

On /dev/nvme1n1
gluster_vg_nvme1n1 volume group with gluster_lv_vmstore logical volume

The problem arises when I add another nvme disk, that, occupying a PCI
slot, it seems has always higher priority of the previous /dev/nvme1n1 disk
and so takes its name.

After booting the node:

old nvme0n1 --> unmodified name
old nvme1n1 --> becomes nvme2n1
new disk --> gets name nvme1n1

>From a funcional point of view I have no problems apart LVM warnings I send
below and also because the xfs entries in fstab are with UUID:

UUID=fa5dd3cb-aeef-470e-b982-432ac896d87a /gluster_bricks/engine xfs
inode64,noatime,nodiratime 0 0
UUID=43bed7de-66b1-491d-8055-5b4ef9b0482f /gluster_bricks/data xfs
inode64,noatime,nodiratime 0 0
UUID=b81a491c-0a4c-4c11-89d8-9db7fe82888e /gluster_bricks/vmstore xfs
inode64,noatime,nodiratime 0 0

lvs commands get:

[root@ovirt ~]# lvs
  WARNING: Not using device /dev/nvme0n1 for PV
eYfuXw-yaPd-cMUE-0dnA-tVON-uZ9g-5x4BDp.
  WARNING: Not using device /dev/nvme2n1 for PV
O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl.
  WARNING: PV eYfuXw-yaPd-cMUE-0dnA-tVON-uZ9g-5x4BDp prefers device
/dev/mapper/nvme.8086-50484b53373530353031325233373541474e-494e54454c20535344504544314b3337354741-0001
because device is used by LV.
  WARNING: PV O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl prefers device
/dev/mapper/eui.010001005cd2e4e359284f51 because device is used by
LV.
  LV  VG Attr   LSize
PoolOrigin
...

Or, for the old nvme1n1 disk, now nvme2n1 multipath device:

[root@ovirt ~]# pvdisplay /dev/mapper/eui.010001005cd2e4e359284f51
  WARNING: Not using device /dev/nvme0n1 for PV
eYfuXw-yaPd-cMUE-0dnA-tVON-uZ9g-5x4BDp.
  WARNING: Not using device /dev/nvme2n1 for PV
O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl.
  WARNING: PV eYfuXw-yaPd-cMUE-0dnA-tVON-uZ9g-5x4BDp prefers device
/dev/mapper/nvme.8086-50484b53373530353031325233373541474e-494e54454c20535344504544314b3337354741-0001
because device is used by LV.
  WARNING: PV O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl prefers device
/dev/mapper/eui.010001005cd2e4e359284f51 because device is used by
LV.
  --- Physical volume ---
  PV Name   /dev/mapper/eui.010001005cd2e4e359284f51
  VG Name   gluster_vg_nvme1n1
  PV Size   931.51 GiB / not usable 1.71 MiB
  Allocatable   yes (but full)
  PE Size   4.00 MiB
  Total PE  238467
  Free PE   0
  Allocated PE  238467
  PV UUID   O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl

[root@ovirt ~]#

I'm able to create PV on top of the new multipath device detected by the
system (see the nvme1n1 of the underlying disk):

eui.010001005cd2e4b5e7db4d51 dm-6 NVME,INTEL SSDPEDKX040T7

size=3.6T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=0 status=active
  `- 1:0:1:0 nvme1n1 259:1 active undef running

[root@ovirt ~]# pvcreate --dataalignment 256K
/dev/mapper/eui.010001005cd2e4b5e7db4d51
  WARNING: Not using device /dev/nvme0n1 for PV
eYfuXw-yaPd-cMUE-0dnA-tVON-uZ9g-5x4BDp.
  WARNING: Not using device /dev/nvme2n1 for PV
O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl.
  WARNING: PV eYfuXw-yaPd-cMUE-0dnA-tVON-uZ9g-5x4BDp prefers device
/dev/mapper/nvme.8086-50484b53373530353031325233373541474e-494e54454c20535344504544314b3337354741-0001
because device is used by LV.
  WARNING: PV O43LFq-46Gc-RRgS-Sk1F-5mFZ-Qw4n-oxXgJl prefers device
/dev/mapper/eui.010001005cd2e4e359284f51 because device is used by
LV.
  Physical volume "/dev/mapper/eui.010001005cd2e4b5e7db4d51"

[ovirt-users] Gluster permissions HCI

2020-03-25 Thread Strahil Nikolov
Hello All,
can someone assist me with some issue.

Could you check the ownership of some folders for me ?

1. ls -l /rhev/data-center/mnt/glusterSD
2. ls -l /rhev/data-center/mnt/glusterSD/_
3. ls -l /rhev/data-center/mnt/glusterSD/_//images
4. ls -l /rhev/data-center/mnt/glusterSD/_
5. ls -l /rhev/data-center/mnt/glusterSD/_//images

Also mention your gluster version.

Thanks in advance.

Best Regards,
Strahil Nikolov
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HQ4Z3PVFWMEPGMNHKN542YLIVGPIRSZL/


[ovirt-users] Gluster Settings

2020-03-19 Thread Christian Reiss

Hey folks,

quick question. For running Gluster / oVirt I found several places, some 
outdated (ovirt docs), gluster Mailinglists, oVirt Mailinglists etc that 
recommend different things.


Here is what I found out/configured:

features.barrier: disable
features.show-snapshot-directory: on
features.uss: enable
cluster.data-self-heal-algorithm: full
cluster.entry-self-heal: on
cluster.data-self-heal: on
cluster.metadata-self-heal: on
cluster.readdir-optimize: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
nfs.disable: on
performance.client-io-threads: off
network.remote-dio: off
performance.strict-o-direct: on
client.event-threads: 16
cluster.choose-local: true
snap-activate-on-create: enable
auto-delete: enable

Would you agree or change anything (usual vm workload).

Thanks! o/
And keep healthy.

--
with kind regards,
mit freundlichen Gruessen,

Christian Reiss
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ABTBEHQG7A3F45F7TS2EB3KAGVHGUC5N/


[ovirt-users] Gluster Heal Issue

2020-01-31 Thread Christian Reiss

Hey folks,

in our production setup with 3 nodes (HCI) we took one host down 
(maintenance, stop gluster, poweroff via ssh/ovirt engine). Once it was 
up the gluster hat 2k healing entries that went down in a matter on 10 
minutes to 2.


Those two give me a headache:

[root@node03:~] # gluster vol heal ssd_storage info
Brick node01:/gluster_bricks/ssd_storage/ssd_storage


Status: Connected
Number of entries: 2

Brick node02:/gluster_bricks/ssd_storage/ssd_storage
Status: Connected
Number of entries: 0

Brick node03:/gluster_bricks/ssd_storage/ssd_storage


Status: Connected
Number of entries: 2

No paths, only gfid. We took down node2, so it does not have the file:

[root@node01:~] # md5sum 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6
75c4941683b7eabc223fc9d5f022a77c 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6


[root@node02:~] # md5sum 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6
md5sum: 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6: 
No such file or directory


[root@node03:~] # md5sum 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6
75c4941683b7eabc223fc9d5f022a77c 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6


The other two files are md5-identical.

These flags are identical, too:

[root@node01:~] # getfattr -d -m . -e hex 
/gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6

getfattr: Removing leading '/' from absolute path names
# file: 
gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6

security.selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f743a733000
trusted.afr.dirty=0x
trusted.afr.ssd_storage-client-1=0x004f0001
trusted.gfid=0xa121e4fb09844e4194d78f0c4f87f4b6
trusted.gfid2path.d4cf876a215b173f=0x62653331383633382d653861302d346336642d393737642d3761393337616138343830362f38366461303238392d663734662d343230302d393238342d3637386537626437363139352e31323030
trusted.glusterfs.mdata=0x015e349b1e1139aa2a5e349b1e1139aa2a5e349949304a5eb2

getfattr: Removing leading '/' from absolute path names
# file: 
gluster_bricks/ssd_storage/ssd_storage/.glusterfs/a1/21/a121e4fb-0984-4e41-94d7-8f0c4f87f4b6

security.selinux=0x73797374656d5f753a6f626a6563745f723a676c7573746572645f627269636b5f743a733000
trusted.afr.dirty=0x
trusted.afr.ssd_storage-client-1=0x004f0001
trusted.gfid=0xa121e4fb09844e4194d78f0c4f87f4b6
trusted.gfid2path.d4cf876a215b173f=0x62653331383633382d653861302d346336642d393737642d3761393337616138343830362f38366461303238392d663734662d343230302d393238342d3637386537626437363139352e31323030
trusted.glusterfs.mdata=0x015e349b1e1139aa2a5e349b1e1139aa2a5e349949304a5eb2

Now, I dont dare simply proceeding withouth some advice.
Anyone got a clue on who to resolve this issue? File #2 is identical to 
this one, from a problem point of view.


Have a great weekend!
-Chris.

--
with kind regards,
mit freundlichen Gruessen,

Christian Reiss
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/FGIQFIRC6QYN4AYB3NRPM42KX4ENIF2A/


[ovirt-users] gluster shard size

2020-01-24 Thread Alex McWhirter

building a new gluster volume this weekend, trying to optimize it fully
for virt. RHGS states that it supports only a 512mb shard size, so i ask
why is the default for ovirt 64mb?___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5FT6HLJDQCQKQ3BQG7CD2IF2OC6F544T/


[ovirt-users] Gluster storage options

2020-01-23 Thread Shareef Jalloq
Hi there,

I'm wanting to build a 3 node Gluster hyperconverged setup but am
struggling to find documentation and examples of the storage setup.

There seems to be a dead link to an old blog post on the Gluster section of
the documentation:
https://www.ovirt.org/blog/2018/02/up-and-running-with-ovirt-4-2-and-gluster-storage/

Is the flow to install the oVirt Node image on a boot drive and then add
disks for Gluster? Or is Gluster setup first with ovirt installed on top?

Thanks.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VBUO7APZDQJB2JF3ECBLR2JEUHDWO2IW/


[ovirt-users] Gluster: a lof of Number of ntries in heal pending

2020-01-20 Thread Stefan Wolf
Hello to all,

I ve a problem with gluster

[root@kvm10 ~]# gluster volume heal data info summary
Brick kvm10:/gluster_bricks/data
Status: Connected
Total Number of entries: 868
Number of entries in heal pending: 868
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm320.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 1
Number of entries in heal pending: 1
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm360.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 867
Number of entries in heal pending: 867
Number of entries in split-brain: 0
Number of entries possibly healing: 0

Brick kvm380.durchhalten.intern:/gluster_bricks/data
Status: Connected
Total Number of entries: 868
Number of entries in heal pending: 868
Number of entries in split-brain: 0
Number of entries possibly healing: 0

[root@kvm10 ~]# gluster volume heal data info split-brain
Brick kvm10:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm320.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm360.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

Brick kvm380.durchhalten.intern:/gluster_bricks/data
Status: Connected
Number of entries in split-brain: 0

As I understand there is no split-brain but 868 files ar in state heal pending.
I ve restarted every node.

I ve also tried:
[root@kvm10 ~]# gluster volume heal data full
Launching heal operation to perform full self heal on volume data has been 
successful
Use heal info commands to check status.

but even after a week there is no really change ( I started with 912 Number of 
entries in heal pending)

can somebody tell what exactly is the problem and how can I solve it.

thank you very much
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/PN63LC3OBQOM7IQY763ZS5V6VZDUFPNP/


[ovirt-users] gluster shards not healing

2020-01-13 Thread Jayme
I have a 3-way replica HCI setup.  I recently placed one host in
maintenance to perform work on it.  When I re-activated it I've noticed
that many of my gluster volumes are not completing the heal process.

heal info shows shard files in heal pending.  I looked up the files and it
appears that they exist on the other two hosts (the ones that remained
active) but do not exist on the host that was in maintenance.

I tried to run a manual heal on one of the volumes and then a full heal as
well but there are still unhealed shards.  The shard files also still do
not exist on the maintenance host.  Here is an example from one of my
volumes:

# gluster volume heal prod_a info
Brick gluster0:/gluster_bricks/prod_a/prod_a
Status: Connected
Number of entries: 0

Brick gluster1:/gluster_bricks/prod_a/prod_a
/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177
/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.178
Status: Connected
Number of entries: 2

Brick gluster2:/gluster_bricks/prod_a/prod_a
/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177
/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.178
Status: Connected
Number of entries: 2


host0:

# ls -al
/gluster_bricks/prod_a/prod_a/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177
ls: cannot access
/gluster_bricks/prod_a/prod_a/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177:
No such file or directory

host1:

# ls -al
/gluster_bricks/prod_a/prod_a/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177
-rw-rw. 2 root root 67108864 Jan 13 16:57
/gluster_bricks/prod_a/prod_a/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177

host2:

# ls -al
/gluster_bricks/prod_a/prod_a/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177
-rw-rw. 2 root root 67108864 Jan 13 16:57
/gluster_bricks/prod_a/prod_a/.shard/a746f8d2-5044-4d20-b525-24456e6f6f16.177


How can I heal these volumes?

Thanks!
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IGG5PR6DCOTWDZJWOVCSTTKOD74VZKUA/


[ovirt-users] Gluster FS & hosted engine fails to set up

2019-12-08 Thread rob . downer
I have set up 3 new servers and as you can see Gluster is working well, however 
the hosted engine deployment fails 

can anyone suggest a reason ?

I have wiped and set up all three servers again and set up Gluster first.
This is the gluster congig I have used for the setup.

Please review the configuration. Once you click the 'Finish Deployment' button, 
the management VM will be transferred to the configured storage and the 
configuration of your hosted engine cluster will be finalized. You will be able 
to use your hosted engine once this step finishes.
* StorageStorage Type:glusterfs
Storage Domain 
Connection:gfs3.gluster.private:/engine
Mount 
Options:backup-volfile-servers=gfs2.gluster.private:gfs1.gluster.private
Disk 
Size (GiB):58


[ INFO ] TASK [ovirt.hosted_engine_setup : Add glusterfs storage domain]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is 
"[Unexpected exception]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault 
reason is \"Operation Failed\". Fault detail is \"[Unexpected exception]\". 
HTTP response code is 400."}


root@ovirt3 ~]# gluster volume status 
Status of volume: data
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs3.gluster.private:/gluster_bricks/
data/data   49152 0  Y   3756 
Brick gfs2.gluster.private:/gluster_bricks/
data/data   49153 0  Y   3181 
Brick gfs1.gluster.private:/gluster_bricks/
data/data   49152 0  Y   15548
Self-heal Daemon on localhost   N/A   N/AY   17602
Self-heal Daemon on gfs1.gluster.privateN/A   N/AY   15706
Self-heal Daemon on gfs2.gluster.privateN/A   N/AY   3348 
 
Task Status of Volume data
--
There are no active volume tasks
 
Status of volume: engine
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs3.gluster.private:/gluster_bricks/
engine/engine   49153 0  Y   3769 
Brick gfs2.gluster.private:/gluster_bricks/
engine/engine   49154 0  Y   3194 
Brick gfs1.gluster.private:/gluster_bricks/
engine/engine   49153 0  Y   15559
Self-heal Daemon on localhost   N/A   N/AY   17602
Self-heal Daemon on gfs1.gluster.privateN/A   N/AY   15706
Self-heal Daemon on gfs2.gluster.privateN/A   N/AY   3348 
 
Task Status of Volume engine
--
There are no active volume tasks
 
Status of volume: vmstore
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs3.gluster.private:/gluster_bricks/
vmstore/vmstore 49154 0  Y   3786 
Brick gfs2.gluster.private:/gluster_bricks/
vmstore/vmstore 49152 0  Y   2901 
Brick gfs1.gluster.private:/gluster_bricks/
vmstore/vmstore 49154 0  Y   15568
Self-heal Daemon on localhost   N/A   N/AY   17602
Self-heal Daemon on gfs1.gluster.privateN/A   N/AY   15706
Self-heal Daemon on gfs2.gluster.privateN/A   N/AY   3348 
 
Task Status of Volume vmstore
--
There are no active volume tasks
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/T4TP2PHEHFU7QFLE7RXWGCGNJLSFTQ2N/


[ovirt-users] Gluster mount still fails on Engine deployment - any suggestions...

2019-12-08 Thread rob . downer
Hi Engine deployment fails here...

[ INFO ] TASK [ovirt.hosted_engine_setup : Add glusterfs storage domain]
[ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is 
"[Unexpected exception]". HTTP response code is 400.
[ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault 
reason is \"Operation Failed\". Fault detail is \"[Unexpected exception]\". 
HTTP response code is 400."}

However Gluster looks good...

I have reinstalled all nodes from scratch.

root@ovirt3 ~]# gluster volume status 
Status of volume: data
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs3.gluster.private:/gluster_bricks/
data/data   49152 0  Y   3756 
Brick gfs2.gluster.private:/gluster_bricks/
data/data   49153 0  Y   3181 
Brick gfs1.gluster.private:/gluster_bricks/
data/data   49152 0  Y   15548
Self-heal Daemon on localhost   N/A   N/AY   17602
Self-heal Daemon on gfs1.gluster.privateN/A   N/AY   15706
Self-heal Daemon on gfs2.gluster.privateN/A   N/AY   3348 
 
Task Status of Volume data
--
There are no active volume tasks
 
Status of volume: engine
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs3.gluster.private:/gluster_bricks/
engine/engine   49153 0  Y   3769 
Brick gfs2.gluster.private:/gluster_bricks/
engine/engine   49154 0  Y   3194 
Brick gfs1.gluster.private:/gluster_bricks/
engine/engine   49153 0  Y   15559
Self-heal Daemon on localhost   N/A   N/AY   17602
Self-heal Daemon on gfs1.gluster.privateN/A   N/AY   15706
Self-heal Daemon on gfs2.gluster.privateN/A   N/AY   3348 
 
Task Status of Volume engine
--
There are no active volume tasks
 
Status of volume: vmstore
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs3.gluster.private:/gluster_bricks/
vmstore/vmstore 49154 0  Y   3786 
Brick gfs2.gluster.private:/gluster_bricks/
vmstore/vmstore 49152 0  Y   2901 
Brick gfs1.gluster.private:/gluster_bricks/
vmstore/vmstore 49154 0  Y   15568
Self-heal Daemon on localhost   N/A   N/AY   17602
Self-heal Daemon on gfs1.gluster.privateN/A   N/AY   15706
Self-heal Daemon on gfs2.gluster.privateN/A   N/AY   3348 
 
Task Status of Volume vmstore
--
There are no active volume tasks
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ACVH5XGIYYTEXNT4RLU47JAE3FASTKYM/


[ovirt-users] Gluster set up fails - Nearly there I think...

2019-11-23 Thread rob . downer
Gluster fails with
vdo: ERROR - Device /dev/sdb excluded by a filter.\n", 

however I have run

[root@ovirt1 ~]# vdo create --name=vdo1 --device=/dev/sdb --force
Creating VDO vdo1
Starting VDO vdo1
Starting compression on VDO vdo1
VDO instance 1 volume is ready at /dev/mapper/vdo1
[root@ovirt1 ~]# 

there are no filters in lvm.conf

I have run

wipefs -a /dev/sdb —force

on all hosts before start
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RPVWGWIP35QWFNCAGABMF4GC24IEZWX5/


[ovirt-users] Gluster setup 3 Node - Now only showing single node setup in setup Wizard

2019-11-23 Thread rob . downer
I have set up 3 Nodes with a separate volume for Gluster, I have set up the two 
networks and DNS works fine SSH has been set up for Gluster and you can login 
via ssh to the other two hosts from the host used to set up.

When going to Virtualisation > Setup Gluster and Hosted Engine only single node 
shows up.

I have restarted all 3 machines.

All nodes machines show up in Dashboard etc...

I have set this up before but it all worked, I erased everything and set up 
again with the separate volume to be used for Gluster storage 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4BVLMM6L6I7LEVKXVKQCQIT4DLSYRADX/


[ovirt-users] Gluster & Hyper Converged setup

2019-11-18 Thread rob . downer
Hi,

Gluster will not set up and fails... can anyone see why ?

/etc/hosts set up for both backend Gluster network and front end, also LAN DNS 
set up on the subnet for the front end.


TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] **
task path: 
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml:17
skipping: [gfs2.gluster.private] => {"changed": false,
"skip_reason": "Conditional result was False"}
skipping: [gfs1.gluster.private] => {"changed": false,
"skip_reason": "Conditional result was False"}
skipping: [gfs3.gluster.private] => {"changed": false,
"skip_reason": "Conditional result was False"}

TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] **
task path: 
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml:24
ok: [gfs2.gluster.private] => {"ansible_facts": {"pv_dataalign":
"3072K\n"}, "changed": false}
ok: [gfs1.gluster.private] => {"ansible_facts": {"pv_dataalign":
"3072K\n"}, "changed": false}
ok: [gfs3.gluster.private] => {"ansible_facts": {"pv_dataalign":
"3072K\n"}, "changed": false}

TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for RAID] 
***
task path: 
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml:34
ok: [gfs2.gluster.private] => {"ansible_facts": {"vg_pesize":
"3072K\n"}, "changed": false}
ok: [gfs1.gluster.private] => {"ansible_facts": {"vg_pesize":
"3072K\n"}, "changed": false}
ok: [gfs3.gluster.private] => {"ansible_facts": {"vg_pesize":
"3072K\n"}, "changed": false}

TASK [gluster.infra/roles/backend_setup : Create volume groups] 
task path: 
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/vg_create.yml:46
failed: [gfs1.gluster.private] (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'}) => {"ansible_loop_var":
"item", "changed": false, "item": {"pvname":
"/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg":
"Device /dev/sdb not found."}
failed: [gfs3.gluster.private] (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'}) => {"ansible_loop_var":
"item", "changed": false, "item": {"pvname":
"/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg":
"Device /dev/sdb not found."}
failed: [gfs2.gluster.private] (item={u'vgname': u'gluster_vg_sdb',
u'pvname': u'/dev/sdb'}) => {"ansible_loop_var":
"item", "changed": false, "item": {"pvname":
"/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg":
"Device /dev/sdb not found."}

NO MORE HOSTS LEFT *

NO MORE HOSTS LEFT *

PLAY RECAP *
gfs1.gluster.private : ok=10 changed=0 unreachable=0 failed=1 skipped=16
rescued=0 ignored=0
gfs2.gluster.private : ok=11 changed=1 unreachable=0 failed=1 skipped=16
rescued=0 ignored=0
gfs3.gluster.private : ok=10 changed=0 unreachable=0 failed=1 skipped=16
rescued=0 ignored=0
 0 /  0
 Reply
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WN7U626EMZCWOTXOOMHGPD3S2X5IA6SJ/


[ovirt-users] Gluster setup

2019-11-15 Thread rob . downer
I have set up a 3 node system.

Gluster has its own backend network and I have tried entering the FQDN hosts 
via ssh as follows...
gfs1.gluster.private10.10.45.11
gfs2.gluster.private10.10.45.12 
gfs3.gluster.private10.10.45.13

I entered at /etc/hosts

127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
::1 localhost localhost.localdomain localhost6 localhost6.localdomain6
gfs1.gluster.private10.10.45.11
gfs2.gluster.private10.10.45.12
gfs3.gluster.private10.10.45.13

but on the CLI 

host gfs1.gluster.private

returns 

[root@ovirt1 etc]# host gfs1.gluster.private
Host gfs1.gluster.private not found: 3(NXDOMAIN)
[root@ovirt1 etc]# 

I guess this is the wrong hosts file, resolver.conf lists files first for 
lookup...
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ILABGNZFOH5BP6JW7VZIEA4GIZE7DXUT/


[ovirt-users] Gluster Questions

2019-11-15 Thread Christian Reiss

Hey folks,

Running a 3 node HCI cluster (in testing stages) I would love to hear 
your input. All nodes are exactly identical and have a local storage of 
8tb in SSDs made out of a RAID6.


Gluster was setup to match this (Raid6, cluster of 256k).

There is the option of compression & dedup, coming from ZFS this is a 
memory hog and kind of insane. What are your thoughts on compression & 
dedup at this time?


The effective size of the compressed drive is suggested 10-fold of the 
original sizes. Seems a big crazy high; any suggestions here?


Thanks for your input!
-Chris.

--
 Christian Reiss - em...@christian-reiss.de /"\  ASCII Ribbon
   supp...@alpha-labs.net   \ /Campaign
 X   against HTML
 WEB alpha-labs.net / \   in eMails

 GPG Retrieval https://gpg.christian-reiss.de
 GPG ID ABCD43C5, 0x44E29126ABCD43C5
 GPG fingerprint = 9549 F537 2596 86BA 733C  A4ED 44E2 9126 ABCD 43C5

 "It's better to reign in hell than to serve in heaven.",
  John Milton, Paradise lost.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SWQSEXHX5DLNE6Y3SBBPIEP3YYPNRDWS/


[ovirt-users] Gluster Network Issues....

2019-11-14 Thread rob . downer
I have set up 3 SuperMicro's with Ovirt Node and all pretty sweet.

FQDN set up for LAN and also after setup I have enabled a second NIC with FQDN 
for a Gluster network.

The issue is the second ports seem to be unavailable for network access by ping 
or login if you login on root the system says that the ports are available 
for login on the bash shell after login and node check comes back fine.

I have IPMI set up on the systems as well for access.

am I missing something ?

I realise Gluster should be on a seperate LAN and will put it on a 10BGe 
network but I'm just testing.

I have the lastest stable build.

Any help would be appreciated.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IWBYBLLUXOM7XMDSKZ7EDZRJVKVUQ47V/


[ovirt-users] Gluster: Bricks remove failed

2019-09-11 Thread toslavik
Hi.
There is an ovirt-hosted-engine on gluster volume engine 
Replicate replica count 3

Migrate to other drives.
I do:
gluster volume add-brick engine clemens:/gluster-bricks/engine 
tiberius:/gluster-bricks/engine octavius:/gluster-bricks/engine force
volume add-brick: success

gluster volume remove-brick engine tiberius:/engine/datastore 
clemens:/engine/datastore octavius:/engine/datastore start
volume remove-brick start: success
ID: dd9453d3-b688-4ed8-ad37-ba901615046c

gluster volume remove-brick engine octavius:/engine/datastore status
 Node Rebalanced-files  size   scanned  failures   
skipped   status  run time in h:m:s
-  ---   ---   ---   ---   
---  --
localhost750.0GB34 1
 0completed0:00:02
clemens   1121.0MB31 0 
0completed0:00:02
 tiberius   1225.0MB36 0
 0completed0:00:02

gluster volume status engine
Status of volume: engine
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick octavius:/engine/datastore49156 0  Y   15669
Brick tiberius:/engine/datastore49156 0  Y   15930
Brick clemens:/engine/datastore 49156 0  Y   16193
Brick clemens:/gluster-bricks/engine49159 0  Y   6168 
Brick tiberius:/gluster-bricks/engine   49163 0  Y   29524
Brick octavius:/gluster-bricks/engine   49159 0  Y   50056
Self-heal Daemon on localhost   N/A   N/AY   50087
Self-heal Daemon on clemens N/A   N/AY   6263 
Self-heal Daemon on tiberiusN/A   N/AY   29583
 
Task Status of Volume engine
--
Task : Remove brick
ID   : dd9453d3-b688-4ed8-ad37-ba901615046c
Removed bricks: 
tiberius:/engine/datastore
clemens:/engine/datastore
octavius:/engine/datastore
Status   : completed  


But the data did not migrate

du -hs /gluster-bricks/engine/ /engine/datastore/
49M /gluster-bricks/engine/
20G /engine/datastore/

Can you give some advice?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Q2KHWTOXBWO6FQKKTM36OIEMN2B6GYQK/


[ovirt-users] gluster

2019-09-09 Thread mailing-ovirt
Hi, 

 

I see options to deploy ovirt with gluster during the initial rollout,
however I can't seem to find information as to how I can add it following a
non gluster initial setup: 

 

GlusterFS Version:

[N/A]

 

Thanks

 

Simon

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KNRQN76L3XGNLYXYEIRYVDCJGU5BBGL5/


[ovirt-users] Gluster 3.12 vs Engine 4.3.5

2019-06-19 Thread Николаев Алексей
Hi community! Is it possible to continue using independent gluster 3.12 as a data domain with engine 4.3.5?___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VXDVKO4V7JNO2NRJ66JGC54Z6HM6GGDT/


[ovirt-users] Gluster Deployment Failed - No Medium Found

2019-06-03 Thread smcintosh
Good Morning, 

I'm completely new to this and I'm testing setting up a Gluster environment 
with Ovirt. However, my deployment keeps fails and I don't understand what it 
means. Any assistance would be much appreciated. Please see error below...

Error Message

TASK [gluster.infra/roles/backend_setup : Create volume groups] 
failed: [ov1.test1.lan] (item={u'vgname': u'gluster_vg_sdb', u'pvname': 
u'/dev/sdb'}) => {"changed": false, "err": "  /dev/sdb: open failed: No medium 
found\n  Device /dev/sdb excluded by a filter.\n", "item": {"pvname": 
"/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume 
'/dev/sdb' failed", "rc": 5}
failed: [ov3.test1.lan] (item={u'vgname': u'gluster_vg_sdb', u'pvname': 
u'/dev/sdb'}) => {"changed": false, "err": "  /dev/sdb: open failed: No medium 
found\n  Device /dev/sdb excluded by a filter.\n", "item": {"pvname": 
"/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume 
'/dev/sdb' failed", "rc": 5}
failed: [ov2.test1.lan] (item={u'vgname': u'gluster_vg_sdb', u'pvname': 
u'/dev/sdb'}) => {"changed": false, "err": "  /dev/sdb: open failed: No medium 
found\n  Device /dev/sdb excluded by a filter.\n", "item": {"pvname": 
"/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume 
'/dev/sdb' failed", "rc": 5}

NO MORE HOSTS LEFT *

NO MORE HOSTS LEFT *
to retry, use: --limit 
@/usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.retry

PLAY RECAP *
ov1.test1.lan: ok=10   changed=0unreachable=0failed=1   
ov2.test1.lan: ok=10   changed=0unreachable=0failed=1   
ov3.test1.lan: ok=10   changed=0unreachable=0failed=1  


Any assistance would be much appreciated. 

Regards
Stephen
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MC5POSKKAVFDSMGHPOBUGBPWHW3LOOUD/


[ovirt-users] Gluster rebuild: request suggestions (poor IO performance)

2019-05-21 Thread Jim Kusznir
Hi:

I've been having one heck of a time for nearly the entire time I've been
running ovirt with disk IO performance.  I've tried a variety of things,
I've posted to this list for help several times, and it sounds like in most
cases the problems are due to design decisions and such.

My cluster has been devolving into nearly unusable performance, and I
believe its mostly disk IO related.  I'm currently using FreeNAS as my
primary VM storage (via NFS), but now it too is performing slowly (it
started out reasonable, but slowly devolved for unknown reasons).

I'm ready to switch back to gluster if I can get specific recommendations
as to what I need to do to make it work.  I feel like I've been trying
random things, and sinking money into this to try and make it work, but
nothing has really fixed the problem.

I have 3 Dell R610 servers with 750GB SSDs as their primary drive.  I had
used some Seagate SSHDs, but the internal Dell DRAC raid controller (which
had been configured to pass them through as a single disk volume, but still
wasn't really JBOD), but it started silently failing them, and causing
major issues for gluster.  I think the DRAC just doesn't like those HDDs.

I can put some real spinning disks in; perhaps a RAID-1 pair of 2TB?  These
servers only take 2.5" hdd's, so that greatly limits my options.

I'm sure others out there are using Dell R610 servers...what do  you use
for storage?  How does it perform?  What do I need to do to get this
cluster actually usable again?  Are PERC-6i storage controllers usable?
I'm not even sure where to go troubleshooting now...everything is so
slw.

BTW: I had a small data volume on the SSDs, and the gluster performance on
those was pretty poor.  performance of the hosted engine is pretty poor
still, and it is still on the SSDs.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/IGR3RDAKQYXSPGAQCHWS5SGKOYA4QKJY/


[ovirt-users] Gluster service failure

2019-05-15 Thread Koen Vanoppen
Dear all,

One little issue. I have 1 hypervisor in my datacenter that keeps having
it's gluster status disconnected in the GUI. But if I look on the server,
the service is running. I added the logs after I clicked on the "Restart
gluster service"

Kind regards,

Koen

--
IMPORTANT!
This message has been scanned for viruses and phishing links.
However, it is your responsibility to evaluate the links and attachments you 
choose to click.
If you are uncertain, we always try to help.
Greetings helpd...@actnet.se



--
IMPORTANT!
This message has been scanned for viruses and phishing links.
However, it is your responsibility to evaluate the links and attachments you 
choose to click.
If you are uncertain, we always try to help.
Greetings helpd...@actnet.se


[2016-10-06 06:11:06.659540] W [rpc-transport.c:291:rpc_transport_load] 
0-rpc-transport: volume 'rdma.management': transport-type 'rdma' is not valid 
or not found on this machine
[2016-10-06 06:11:06.659572] W [rpcsvc.c:1623:rpcsvc_create_listener] 
0-rpc-service: cannot create listener, initing the transport failed
[2016-10-06 06:11:06.659663] E [MSGID: 106243] [glusterd.c:1652:init] 
0-management: creation of 1 listeners failed, continuing with succeeded 
transport
[2016-10-06 06:11:09.639316] I [MSGID: 106513] 
[glusterd-store.c:2098:glusterd_restore_op_version] 0-glusterd: retrieved 
op-version: 30712
[2016-10-06 06:11:09.639470] I [MSGID: 106194] 
[glusterd-store.c:3628:glusterd_store_retrieve_missed_snaps_list] 0-management: 
No missed snaps list.
Final graph:
+--+
  1: volume management
  2: type mgmt/glusterd
  3: option rpc-auth.auth-glusterfs on
  4: option rpc-auth.auth-unix on
  5: option rpc-auth.auth-null on
  6: option rpc-auth-allow-insecure on
  7: option transport.socket.listen-backlog 128
  8: option event-threads 1
  9: option ping-timeout 0
 10: option transport.socket.read-fail-log off
 11: option transport.socket.keepalive-interval 2
 12: option transport.socket.keepalive-time 10
 13: option transport-type rdma
 14: option working-directory /var/lib/glusterd
 15: end-volume
 16:  
+--+
[2016-10-06 06:11:09.640454] I [MSGID: 101190] 
[event-epoll.c:628:event_dispatch_epoll_worker] 0-epoll: Started thread with 
index 1
[2016-10-06 06:11:17.393310] I [MSGID: 106163] 
[glusterd-handshake.c:1271:__glusterd_mgmt_hndsk_versions_ack] 0-management: 
using the op-version 30712
[2016-10-06 06:11:17.393436] I [MSGID: 106544] 
[glusterd.c:155:glusterd_uuid_init] 0-management: retrieved UUID: 
48ec9c09-c98c-47fe-9892-7addf50789a8
[2016-10-06 06:11:17.395099] I [MSGID: 106490] 
[glusterd-handler.c:2608:__glusterd_handle_incoming_friend_req] 0-glusterd: 
Received probe from uuid: 561c450b-2ed7-4ceb-8e39-e9a1227e48ae
[2016-10-06 06:11:17.396915] I [MSGID: 106493] 
[glusterd-handler.c:3852:glusterd_xfer_friend_add_resp] 0-glusterd: Responded 
to ovirthyp01dev.example.com (24007), ret: 0, op_ret: -1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DP4XXZTUSI3LOPC5R6EM4KAEXICIVAKO/


[ovirt-users] Gluster volume heals and after 5 seconds has /dom_md/ids dirty again

2019-05-13 Thread Andreas Elvers
I restored my engine to a gluster volume named :/engine on a three node 
hyperconverged oVirt 4.3.3.1 cluster. Before restoring I was checking the 
status of the volumes. They were clean. No heal entries. All peers connected. 
gluster volume status looked good. Then I restored. This went well. The engine 
is up. But the engine gluster volume shows entries on node02 and node03. The 
engine was installed to node01. I have to deploy the engine to the other two 
hosts to reach full HA, but I bet maintenance is not possible until the volume 
is healed. 

I tried "gluster volume heal engine" also with added "full". The heal entries 
will disappear for a few seconds and then /dom_md/ids will pop up again. The 
__DIRECT_IO_TEST__ will join later. The split-brain info has no entries. Is 
this some kind of hidden split brain? Maybe there is data on node01 brick which 
got not synced to the other two nodes? I can only speculate. Gluster docs say: 
this should heal. But it doesn't.  I have two other volumes. Those are fine. 
One of them containing 3 VMs that are running. I also tried to shut down the 
engine, so no-one was using the volume. Then heal. Same effect. Those two files 
will always show up. But none other. Heal can always be started successfully 
from any of the participating nodes.

Reset the volume bricks one by one and cross fingers? 

[root@node03 ~]#  gluster volume heal engine info
Brick node01.infra.solutions.work:/gluster_bricks/engine/engine
Status: Connected
Number of entries: 0

Brick node02.infra.solutions.work:/gluster_bricks/engine/engine
/9f4d5ae9-e01d-4b73-8b6d-e349279e9782/dom_md/ids
/__DIRECT_IO_TEST__
Status: Connected
Number of entries: 2

Brick node03.infra.solutions.work:/gluster_bricks/engine/engine
/9f4d5ae9-e01d-4b73-8b6d-e349279e9782/dom_md/ids
/__DIRECT_IO_TEST__
Status: Connected
Number of entries: 2
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/L3YCRPRAGPUMBZIBFOPT6L4B7H4M6HLS/


[ovirt-users] Gluster Snapshot Datepicker Not Working?

2019-05-10 Thread Alex McWhirter
Updated to 4.3.3.7, the date picker for gluster snapshot appears to not 
be working? It wont register clicks, and manually typing in times 
doesn't work.


Can anyone else confirm?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/BV5WG6HDZ5Q4VYTZYWLDTRF4TAFAYAQ6/


[ovirt-users] Gluster and few iSCSI Datastores in one Data Center

2019-04-23 Thread toslavik
I decided to add another cluster to the existing data center (Enable Virt 
Service + Enable Gluster Service).
Three nodes. But after installing the nodes (without errors) Ovirt Engine loses 
them cyclically. From the logs you can see that this happens when you try to 
interrogate the connected block devices.
These block devices are LVM for VM. I use several Datastores submitted via 
iSCSI. About 2000 virtual machines and more than 2300 LVMs. All in the neighbor 
cluster (one data center).
Logs in https://drive.google.com/open?id=1ja7Usxx5YCFDgjD2X51z9tzn_ycPoC2g

Why is this happening and what can be done in this case?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NNPWWJWJEIZTHRY7R3GMFANOO6JDVAPA/


[ovirt-users] Gluster suggestions

2019-04-16 Thread Magnus Isaksson
Hello

I would like some suggestions on what type of solution with Gluster i should 
use.

I have 4 hosts with 3 disks each, i want to user as much space as possible but 
also some redundancy, like raid5 or 6
The 4 hosts are running oVirt on CentOS 7
I have 2 clusters due to some licensing issues, so i cant use ovirts setup of 
gluster

Today i have set it up as follows
Type: Disperse
Number of Bricks: 1 x (8 + 4) = 12

So i have 3 bricks per host (one brick per disk)

But this setup is not working very well, as soon as i get some traffic on the 
volume this start to fail on the ovirt nodes, loosing connection etc.

All hosts are connected via 10G interface

So any suggestions on how i would set this up as best is much appreciated.

Regards
 Magnus Isaksson



___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7R2FLPNTC7PYZQHSHDRU6CX6VYAWONRI/


[ovirt-users] Gluster suggestions

2019-04-16 Thread magnus
 Hello

I would like some suggestions on what type of solution with Gluster i should 
use.

I have 4 hosts with 3 disks each, i want to user as much space as possible but 
also some redundancy, like raid5 or 6
The 4 hosts are running oVirt on CentOS 7
I have 2 clusters due to some licensing issues, so i cant use ovirts setup of 
gluster

Today i have set it up as follows
Type: Disperse
Number of Bricks: 1 x (8 + 4) = 12

So i have 3 bricks per host (one brick per disk)

But this setup is not working very well, as soon as i get some traffic on the 
volume this start to fail on the ovirt nodes, loosing connection etc.

All hosts are connected via 10G interface

So any suggestions on how i would set this up as best is much appreciated.

Regards
 Magnus Isaksson
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AIEPMRANHFB7J2FH6IHTFJ2GUXVB37ZP/


[ovirt-users] Gluster arbiter volume storage domain - change

2019-04-16 Thread Leo David
Hi Everyone,
I have wrongly configured the main gluster volume ( 12 identical 1tb ssd
disks, replica 3 distributed-replicated, across 6 nodes - 2 per node ) with
arbiter one.
Oviously I am wasting storage space in this scenario with the arbiter
bricks, and I would like to convert the volume to non-arbitrated one, so
having all the data evenly spreaded across all the disks.
Considering the the storage is being used by about 40 vms in production,
what would it be the steps, or is there any chance to change the volume
type to non-arbitrated on the fly and then rebalance ?
Thank you very much !
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UBEZWN35M365IKCIE3U6TRHDDX7TS75T/


[ovirt-users] Gluster VM image Resync Time

2019-03-27 Thread Indivar Nair
Hi All,

We are planning a 2 + 1 arbitrated mirrored Gluster setup.
We would have around 50 - 60 VMs, with an average 500GB disk size.

Now in case one of the Gluster Nodes go completely out of sync, roughly,
how long would it take to resync? (as per your experience)
Will it impact the working of VMs in any way?
Is there anything to be taken care of, in advance, to prepare for such a
situation?

Regards,


Indivar Nair
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HEFHHQYVMXGMEKI3UVWAVMV3IKY7SMR2/


[ovirt-users] Gluster event logging

2019-03-25 Thread Ron Jerome
I'm running oVirt 4.3.2, and just upgraded gluster to 5.5.  I see that the 
gluster event daemon now works, however the events are not being processed by 
the ovirt engine.  On the engine side I'm seeing: 

"engine.log:2019-03-25 17:14:04,707-04 ERROR 
[org.ovirt.engine.core.services.GlusterEventsWebHookServlet] (default task-198) 
[] Error processing event data" in the engine log. 

On the gluster servers in event.log, I'm seeing: 

"[2019-03-25 17:14:04,704] WARNING [utils - 272:publish_to_webhook] - Event 
push failed to URL: http://ovirt-engine:80/ovirt-engine/services/glusterevents, 
Event: {"event": "CLIENT_CONNECT", "message": {"brick_path": 
"/gluster_bricks/ISO/ISO", "client_identifier": "10.10.15.12:49120", 
"client_uid": 
"CTX_ID:d988b64d-2996-44c4-88ed-ee768c9e0c21-GRAPH_ID:0-PID:43711-HOST:ovirt2-PC_NAME:ISO-client-2-RECON_NO:-0",
 "server_identifier": "10.10.15.12:49155", "subdir_mount": "(null)"}, "nodeid": 
"9a6ea54e-c692-45bd-889b-ed20c6fb1145", "ts": 1553548444}, Status Code: 500"

How can I determine why these events are not getting processed?

Thanks,
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/D574QVY4X2DK4KAXQGHXE7PYILJT5MDA/


[ovirt-users] "gluster-ansible-roles is not installed on Host" error on Cockpit

2019-03-09 Thread Hesham Ahmed
On a new 4.3.1 oVirt Node installation, when trying to deploy HCI
(also when trying adding a new gluster volume to existing clusters)
using Cockpit, an error is displayed "gluster-ansible-roles is not
installed on Host. To continue deployment, please install
gluster-ansible-roles on Host and try again". There is no package
named gluster-ansible-roles in the repositories:

[root@localhost ~]# yum install gluster-ansible-roles
Loaded plugins: enabled_repos_upload, fastestmirror, imgbased-persist,
package_upload, product-id, search-disabled-repos,
subscription-manager, vdsmupgrade
This system is not registered with an entitlement server. You can use
subscription-manager to register.
Loading mirror speeds from cached hostfile
 * ovirt-4.3-epel: mirror.horizon.vn
No package gluster-ansible-roles available.
Error: Nothing to do
Uploading Enabled Repositories Report
Cannot upload enabled repos report, is this client registered?

This is due to check introduced here:
https://gerrit.ovirt.org/#/c/98023/1/dashboard/src/helpers/AnsibleUtil.js

Changing the line from:
[ "rpm", "-qa", "gluster-ansible-roles" ], { "superuser":"require" }
to
[ "rpm", "-qa", "gluster-ansible" ], { "superuser":"require" }
resolves the issue. The above code snippet is installed at
/usr/share/cockpit/ovirt-dashboard/app.js on oVirt node and can be
patched by running "sed -i 's/gluster-ansible-roles/gluster-ansible/g'
/usr/share/cockpit/ovirt-dashboard/app.js && systemctl restart
cockpit"
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/243QJOXO2KTWYU5CDH3OC7WJ6Z2EL4CG/


[ovirt-users] Gluster messages after upgrade to 4.3.1

2019-03-01 Thread Stefano Danzi

Hello,

I've just upgrade to version 4.3.1 and I can see this message in gluster 
log of all my host (running oVirt Node):


The message "E [MSGID: 101191] 
[event-epoll.c:671:event_dispatch_epoll_worker] 0-epoll: Failed to 
dispatch handler" repeated 59 times between [2019-03-01 10:21:42.099983] 
and [2019-03-01 10:23:38.340971


Another strange thing:

A Vm was running. I shutted down it fo mistake. I was no more able to 
run this vm. The error was: "Bad volume specification ".
After a little investigation I notice that disk image was no more owned 
by vdsm.kvm but root.root. I changed back to correct value and vm 
started fine.

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/S5VIYA7QAO5QMWMAMB3OAULMKRLBYHVS/


[ovirt-users] Gluster - performance.strict-o-direct and other performance tuning in different storage backends

2019-02-25 Thread Leo David
Hello Everyone,
As per some previous posts,  this "performance.strict-o-direct=on" setting
caused trouble or poor vm iops.  I've noticed that this option is still
part of default setup or automatically configured with
"Optimize for virt. store" button.
In the end... is this setting a good or a bad practice for setting the vm
storage volume ?
Does it depends ( like maybe other gluster performance options ) on the
storage backend:
- raid type /  jbod
- raid controller cache size
I am usually using jbod disks attached to lsi hba card ( no cache ). Any
gluster recommendations regarding this setup ?
Is there any documentation for best practices on configurating ovirt's
gluster for different types of storage backends ?
Thank you very much !

Have a great week,

Leo

-- 
Best regards, Leo David
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7FKL42JSHIKPMKLLMDPKYM4XT4V5GT4W/


[ovirt-users] Gluster setup Problem

2019-02-22 Thread Matthew Roth
I have 3 servers,  Node 1 is 3tb /dev/sda, Node 2, 3tb /dev/sdb,  node3 3tb 
/dev/sdb

I start the process for gluster deployment. I change node 1 to sda and all the 
other ones to sdb. I get no errors however, 

when I get to 
Creating physical Volume all it does is spin forever . doesnt get any further. 
I can leave it there for 5 hours and doesn't go anywhere. 

#gdeploy configuration generated by cockpit-gluster plugin
[hosts]
cmdnode1.cmd911.com
cmdnode2.cmd911.com
cmdnode3.cmd911.com

[script1:cmdnode1.cmd911.com]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sda -h 
cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com

[script1:cmdnode2.cmd911.com]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h 
cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com

[script1:cmdnode3.cmd911.com]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h 
cmdnode1.cmd911.com, cmdnode2.cmd911.com, cmdnode3.cmd911.com

[disktype]
raid6

[diskcount]
12

[stripesize]
256

[service1]
action=enable
service=chronyd

[service2]
action=restart
service=chronyd

[shell2]
action=execute
command=vdsm-tool configure --force

[script3]
action=execute
file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh
ignore_script_errors=no

[pv1:cmdnode1.cmd911.com]
action=create
devices=sda
ignore_pv_errors=no

[pv1:cmdnode2.cmd911.com]
action=create
devices=sdb
ignore_pv_errors=no

[pv1:cmdnode3.cmd911.com]
action=create
devices=sdb
ignore_pv_errors=no

[vg1:cmdnode1.cmd911.com]
action=create
vgname=gluster_vg_sda
pvname=sda
ignore_vg_errors=no

[vg1:cmdnode2.cmd911.com]
action=create
vgname=gluster_vg_sdb
pvname=sdb
ignore_vg_errors=no

[vg1:cmdnode3.cmd911.com]
action=create
vgname=gluster_vg_sdb
pvname=sdb
ignore_vg_errors=no

[lv1:cmdnode1.cmd911.com]
action=create
poolname=gluster_thinpool_sda
ignore_lv_errors=no
vgname=gluster_vg_sda
lvtype=thinpool
size=1005GB
poolmetadatasize=5GB

[lv2:cmdnode2.cmd911.com]
action=create
poolname=gluster_thinpool_sdb
ignore_lv_errors=no
vgname=gluster_vg_sdb
lvtype=thinpool
size=1005GB
poolmetadatasize=5GB

[lv3:cmdnode3.cmd911.com]
action=create
poolname=gluster_thinpool_sdb
ignore_lv_errors=no
vgname=gluster_vg_sdb
lvtype=thinpool
size=41GB
poolmetadatasize=1GB

[lv4:cmdnode1.cmd911.com]
action=create
lvname=gluster_lv_engine
ignore_lv_errors=no
vgname=gluster_vg_sda
mount=/gluster_bricks/engine
size=100GB
lvtype=thick

[lv5:cmdnode1.cmd911.com]
action=create
lvname=gluster_lv_data
ignore_lv_errors=no
vgname=gluster_vg_sda
mount=/gluster_bricks/data
lvtype=thinlv
poolname=gluster_thinpool_sda
virtualsize=500GB

[lv6:cmdnode1.cmd911.com]
action=create
lvname=gluster_lv_vmstore
ignore_lv_errors=no
vgname=gluster_vg_sda
mount=/gluster_bricks/vmstore
lvtype=thinlv
poolname=gluster_thinpool_sda
virtualsize=500GB

[lv7:cmdnode2.cmd911.com]
action=create
lvname=gluster_lv_engine
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/engine
size=100GB
lvtype=thick

[lv8:cmdnode2.cmd911.com]
action=create
lvname=gluster_lv_data
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/data
lvtype=thinlv
poolname=gluster_thinpool_sdb
virtualsize=500GB

[lv9:cmdnode2.cmd911.com]
action=create
lvname=gluster_lv_vmstore
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/vmstore
lvtype=thinlv
poolname=gluster_thinpool_sdb
virtualsize=500GB

[lv10:cmdnode3.cmd911.com]
action=create
lvname=gluster_lv_engine
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/engine
size=20GB
lvtype=thick

[lv11:cmdnode3.cmd911.com]
action=create
lvname=gluster_lv_data
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/data
lvtype=thinlv
poolname=gluster_thinpool_sdb
virtualsize=20GB

[lv12:cmdnode3.cmd911.com]
action=create
lvname=gluster_lv_vmstore
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/vmstore
lvtype=thinlv
poolname=gluster_thinpool_sdb
virtualsize=20GB

[selinux]
yes

[service3]
action=restart
service=glusterd
slice_setup=yes

[firewalld]
action=add
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
services=glusterfs

[script2]
action=execute
file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh

[shell3]
action=execute
command=usermod -a -G gluster qemu

[volume1]
action=create
volname=engine
transport=tcp
replica=yes
replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable
brick_dirs=cmdnode1.cmd911.com:/gluster_bricks/engine/engine,cmdnode2.cmd911.com:/gluster_bricks/engine/engine,cmdnode3.cmd911.com:/gluster_bricks/engine/engine
ignore_volume_errors=no
arbiter_count=1

[volume2]
action=create
volname=data
transport=tcp
replica=yes
replica_count=3

[ovirt-users] Gluster version on oVirt 4.3.0

2019-02-07 Thread suporte
Hi, 

What Glusterfs version should I use with oVirt 4.3.0 ? 

Thanks 

-- 

Jose Ferradeira 
http://www.logicworks.pt 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/Z7ZZO5DES3HZEAYYUVEJHNTVVMSTOUPB/


[ovirt-users] Gluster Disk Full

2018-12-17 Thread suporte
Hi, 

I have a gluster volume with disk full. It's a simple oVirt all in one. 
The VM with the disk on that volume stopped. 

The error message: 
Failed to update OVF disks 1252828a-3265-4d5c-9440-40b2b85c1feb, OVF data isn't 
updated on those OVF stores (Data Center Default, Storage Domain gfs1). 
VDSM command SetVolumeDescriptionVDS failed: Error while updating volume meta 
data: 
("(u'/rhev/data-center/mnt/glusterSD/gfs1.growtrade.pt:_gv0/0f21e2bf-44d9-4afd-8800-28e99d617a4b/images/1252828a-3265-4d5c-9440-40b2b85c1feb/f8b62638-ba48-4ed8-8101-3027c124d20c',)[Errno
 28] No space left on device: 
u'/rhev/data-center/mnt/glusterSD/gfs1.growtrade.pt:_gv0/0f21e2bf-44d9-4afd-8800-28e99d617a4b/images/1252828a-3265-4d5c-9440-40b2b85c1feb/f8b62638-ba48-4ed8-8101-3027c124d20c.meta.new'",)
 


Any idea what can I do? 

Thanks 

-- 

Jose Ferradeira 
http://www.logicworks.pt 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LLVOIBFXYI2T7T67IRKPZ2VNW7GRI5ZL/


[ovirt-users] Gluster with two ovirt nodes

2018-12-12 Thread Stefan Wolf
Hello,

i like to set up glusterfs with two ovirt nodes and on more "normal" node
is this possible?
i 've setup glusterfs in cli on two ovirt nodes and 3rd network storage.
glusterfs is up and running.
But now i like to get something like VIP with ctdb for example. is there
any possibility to set this up with ovirt?
Or do i *ve to setup ovirt manually in centos to install ctdb?
Or are there any other ideas?

thank you stefan
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4JKFSI4XIMXCKQQEQ7W4ZPWNASYZ52TL/


[ovirt-users] Gluster hooks

2018-12-10 Thread Alex K
Hi all,

I see the following gluster hooks at ovirt cluster enabled:

[image: image.png]

What is their purpose? Do they need to be all enabled?

Thanx,
Alex
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/53CPEDLD4GR32OV4TRJNHZDLFWO2RTHN/


[ovirt-users] Gluster snapshots on hyperconverged with VDO ?

2018-10-08 Thread guillaume . pavese
I evaluate deploying ovirt 4.2 on centos 7.5 with VDO dedup in production.

When deploying gluster with cockpit, after selecting deduplication/compression, 
the option to have the brick thin provisioned is grayed out.

As a result, gluster snapshots are not possible. :
 - Neither in CLI : "snapshot create: failed: Snapshot is supported only for 
thin provisioned LV. Ensure that all bricks of data are thinly provisioned LV.
"
- Nor through the scheduler in volume management : I see the icon "Volume 
snapshots are scheduled" but I do not see any being created. There are no 
snapshot creation  failure message in the manager.

So I have two questions : 
- At deployment, is changing the gdeploy option manually from thick to thin 
supported for a LV on VDO device(or even a good idea)?
- Should I open a bug report?
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/7SP44X3BOIWAWMAK7I3SSJ5WQWSVX4VA/


[ovirt-users] Gluster JSON-RPC errors

2018-10-04 Thread Maton, Brett
I'm seeing the following errors appear in the event log every 10 minutes
for each participating host in the gluster cluster

GetGlusterVolumeHealInfoVDS failed: Internal JSON-RPC error: {'reason':
"'bool' object has no attribute 'getiterator'"}

Gluster brick health is good

Any ideas ?

oVirt 4.2.7.2-1.el7
CentOS 7
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2KLFU2C5UHDFLTH3XUHZ5DGF7WVNGNJZ/


[ovirt-users] Gluster deployment - Arbiter volume - different disk size - volumes sizes

2018-09-12 Thread florentl

Hello,
I have one more question for the gluster deployment.
I have three servers with different disk sizes :
- server1 : 3 x SSD configured in raid 5 mode for a total capacity of 900 GB
- server2 : 6 x SSD configurerd in raid 6 mode for a total capacity of 
2,2 TB

- server3 : same as server2 (2,2TB)
Because my server1 has less capacity than the others I will use it as an 
arbiter volume (during the deployment I will declare it as the last host 
: host3).

Is-it correct ?

So I will create 4 volumes :
- engine : 50 GB type replicated without arbiter
- data : ??
- vmstore : ???
- iso : 15 GB replicated with arbiter volume

Can someone tell me what are the differences between data and vmstore 
volume ? In the earlier version there was only information about data 
volume and export volume. I found nothing about vmstore volume.


If the wmstore is, as its name suggests, the place where all vm files 
will be stored including vm disks images, I think I have to set most of 
the capacity for this volume.

So at the end my 4 volumes will be :
- engine : 50 GB type replicated without arbiter
- data : 500 TB replicated with arbiter volume
- vmstore : 1,5 TB replicated with arbiter volume
- iso : 15 GB replicated with arbiter volume

What do you think of this ?

Thanks,

Florent

___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/S6Q7D734NCFDS3CU736ZI5SM3GSWKAUC/


[ovirt-users] Gluster Issues

2018-09-11 Thread Endre Karlson
Hi, we are seeing some issues where our hosts oom kill glusterd after a
while but there's plenty of memory?

Running Centos 7,4.x and Ovirt 4.2.x
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/U4SRQK74C7DPAB3GWCJQTZZZWU346CAW/


[ovirt-users] Gluster clients intermittently hang until first gluster server in a Replica 1 Arbiter 1 cluster is rebooted, server error: 0-management: Unlocking failed & client error: bailing out fra

2018-09-02 Thread Sam McLeod
We've got an odd problem where clients are blocked from writing to Gluster 
volumes until the first node of the Gluster cluster is rebooted.

I suspect I've either configured something incorrectly with the arbiter / 
replica configuration of the volumes, or there is some sort of bug in the 
gluster client-server connection that we're triggering.

I was wondering if anyone has seen this or could point me in the right 
direction?


Environment:
Typology: 3 node cluster, replica 2, arbiter 1 (third node is metadata only).
Version: Client and Servers both running 4.1.3, both on CentOS 7, kernel 
4.18.x, (Xen) VMs with relatively fast networked SSD storage backing them, XFS.
Client: Native Gluster FUSE client mounting via the kubernetes provider

Problem:
Seemingly randomly some clients will be blocked / are unable to write to what 
should be a highly available gluster volume.
The client gluster logs show it failing to do new file operations across 
various volumes and all three nodes of the gluster.
The server gluster (or OS) logs do not show any warnings or errors.
The client recovers and is able to write to volumes again after the first node 
of the gluster cluster is rebooted.
Until the first node of the gluster cluster is rebooted, the client fails to 
write to the volume that is (or should be) available on the second node (a 
replica) and third node (an arbiter only node).

What 'fixes' the issue:
Although the clients (kubernetes hosts) connect to all 3 nodes of the Gluster 
cluster - restarting the first gluster node always unblocks the IO and allows 
the client to continue writing.
Stopping and starting the glusterd service on the gluster server is not enough 
to fix the issue, nor is restarting its networking.
This suggests to me that the volume is unavailable for writing for some reason 
and restarting the first node in the cluster either clears some sort of TCP 
sessions between the client-server or between the server-server replication.

Expected behaviour:

If the first gluster node / server had failed or was blocked from performing 
operations for some reason (which it doesn't seem it is), I'd expect the 
clients to access data from the second gluster node and write metadata to the 
third gluster node as well as it's an arbiter / metadata only node.
If for some reason the a gluster node was not able to serve connections to 
clients, I'd expect to see errors in the volume, glusterd or brick log files 
(there are none on the first gluster node).
If the first gluster node was for some reason blocking IO on a volume, I'd 
expect that node either to show as unhealthy or unavailable in the gluster peer 
status or gluster volume status.


Client gluster errors:

staging_static in this example is a volume name.
You can see the client trying to connect to the second and third nodes of the 
gluster cluster and failing (unsure as to why?)
The server side logs on the first gluster node do not show any errors or 
problems, but the second / third node show errors in the glusterd.log when 
trying to 'unlock' the 0-management volume on the first node.


On a gluster client (a kubernetes host using the kubernetes connector which 
uses the native fuse client) when its blocked from writing but the gluster 
appears healthy (other than the errors mentioned later):

[2018-09-02 15:33:22.750874] E [rpc-clnt.c:184:call_bail] 
0-staging_static-client-2: bailing out frame type(GlusterFS 4.x v1) 
op(INODELK(29)) xid = 0x1cce sent = 2018-09-02 15:03:22.417773. timeout = 1800 
for :49154
[2018-09-02 15:33:22.750989] E [MSGID: 114031] 
[client-rpc-fops_v2.c:1306:client4_0_inodelk_cbk] 0-staging_static-client-2: 
remote operation failed [Transport endpoint is not connected]
[2018-09-02 16:03:23.097905] E [rpc-clnt.c:184:call_bail] 
0-staging_static-client-1: bailing out frame type(GlusterFS 4.x v1) 
op(INODELK(29)) xid = 0x2e21 sent = 2018-09-02 15:33:22.765751. timeout = 1800 
for :49154
[2018-09-02 16:03:23.097988] E [MSGID: 114031] 
[client-rpc-fops_v2.c:1306:client4_0_inodelk_cbk] 0-staging_static-client-1: 
remote operation failed [Transport endpoint is not connected]
[2018-09-02 16:33:23.439172] E [rpc-clnt.c:184:call_bail] 
0-staging_static-client-2: bailing out frame type(GlusterFS 4.x v1) 
op(INODELK(29)) xid = 0x1d4b sent = 2018-09-02 16:03:23.098133. timeout = 1800 
for :49154
[2018-09-02 16:33:23.439282] E [MSGID: 114031] 
[client-rpc-fops_v2.c:1306:client4_0_inodelk_cbk] 0-staging_static-client-2: 
remote operation failed [Transport endpoint is not connected]
[2018-09-02 17:03:23.786858] E [rpc-clnt.c:184:call_bail] 
0-staging_static-client-1: bailing out frame type(GlusterFS 4.x v1) 
op(INODELK(29)) xid = 0x2ee7 sent = 2018-09-02 16:33:23.455171. timeout = 1800 
for :49154
[2018-09-02 17:03:23.786971] E [MSGID: 114031] 
[client-rpc-fops_v2.c:1306:client4_0_inodelk_cbk] 0-staging_static-client-1: 
remote operation failed [Transport endpoint is not connected]
[2018-09-02 17:33:24.160607] E 

[ovirt-users] Gluster Deployment hangs on enabling or disabling chronyd service

2018-07-17 Thread Sakhi Hadebe
Hi,

Why is gluster deployment hangs on enabling or disabling the chronyd
service? I have enabled passwordless ssh to access itself and other two
nodes.

What would be the solution to get it pass this stage?

On Fri, Jul 13, 2018 at 10:41 AM, Sakhi Hadebe  wrote:

> Hi,
>
> We are running the following setup:
>
> ovirt-engine:
> - CentOS Linux release 7.5.1804 (Core)
> - ovirt-engine-4.2.4.5-1.el7.noarch
>
> node (trying to add):
> - CentOS Linux release 7.5.1804 (Core)
> - vdsm-4.20.32-1.el7.x86_64
>
> - ovirt-release42-4.2.4-1.el7.noarch
>
>
> We successfully added the second node to the Cluster. Two nodes that have 
> been successfully added are running ovirt-node-4.2.4.
>
> ovirt-node-ng-image-update-placeholder-4.2.4-1.el7.noarch
> ovirt-node-ng-nodectl-4.2.0-0.20180626.0.el7.noarch
> ovirt-provider-ovn-driver-1.2.11-1.el7.noarch
> ovirt-release42-4.2.4-1.el7.noarch
> ovirt-release-host-node-4.2.4-1.el7.noarch
>
> While adding a new node to our cluster the installations fail.
>
> I have attached a piece of engine.log and a full ovirt-host-deploy logs of 
> the failing node from the engine.
> Help would be very much appreciated.
>
>
> --
> Regards,
> Sakhi Hadebe
>
>


-- 
Regards,
Sakhi Hadebe

Engineer: South African National Research Network (SANReN)Competency
Area, Meraka, CSIR

Tel:   +27 12 841 2308 <+27128414213>
Fax:   +27 12 841 4223 <+27128414223>
Cell:  +27 71 331 9622 <+27823034657>
Email: sa...@sanren.ac.za 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KBMRRUJWHZZGWBKFLOC5MXYQWGNUATSN/


[ovirt-users] Gluster not synicng changes between nodes for engine

2018-06-16 Thread Hanson Turner

Hi Guys,

I've got 60 some odd files for each of the nodes in the cluster, they 
don't seem to be syncing.


Running a volume heal engine full, reports successful. Running volume 
heal engine info reports the same files, and doesn't seem to be syncing.


Running a volume heal engine info split-brain, there's nothing listed in 
split-brain.


Peers show as connected. Gluster volumes are started/up.

Hosted-engine --vm-status reports :
The hosted engine configuration has not been retrieved from shared 
storage. Please ensure that ovirt-ha-agent is running and the storage 
server is reachable.


This is leaving the cluster in an engine down with all vm's down state...

Thanks,
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YPNWM222K2U7NX32CIME7KINWPCLBSCR/


[ovirt-users] Gluster Deployment: Repositories, Packages

2018-06-14 Thread femi adegoke
In Step 2 of the HE deployment what should be filled in here?
Repositories: ??
Packages: ??
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DW64XCCI55OIYVHPT5WURGT5LIBMDLT5/


[ovirt-users] Gluster problems, cluster performance issues

2018-05-28 Thread Jim Kusznir
Hello:

I've been having some cluster and gluster performance issues lately.  I
also found that my cluster was out of date, and was trying to apply updates
(hoping to fix some of these), and discovered the ovirt 4.1 repos were
taken completely offline.  So, I was forced to begin an upgrade to 4.2.
According to docs I found/read, I needed only add the new repo, do a yum
update, reboot, and be good on my hosts (did the yum update, the
engine-setup on my hosted engine).  Things seemed to work relatively well,
except for a gluster sync issue that showed up.

My cluster is a 3 node hyperconverged cluster.  I upgraded the hosted
engine first, then engine 3.  When engine 3 came back up, for some reason
one of my gluster volumes would not sync.  Here's sample output:

[root@ovirt3 ~]# gluster volume heal data-hdd info
Brick 172.172.1.11:/gluster/brick3/data-hdd
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-446b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-44f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-42a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2
Status: Connected
Number of entries: 8

Brick 172.172.1.12:/gluster/brick3/data-hdd
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-446b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-44f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-42a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b
Status: Connected
Number of entries: 8

Brick 172.172.1.13:/gluster/brick3/data-hdd
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-42a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-446b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-44f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625
Status: Connected
Number of entries: 8

-
Its been in this state for a couple days now, and bandwidth monitoring
shows no appreciable data moving.  I've tried repeatedly commanding a full
heal from all three clusters in the node.  Its always the same files that
need healing.

When running gluster volume heal data-hdd statistics, I see sometimes
different information, but always some number of "heal failed" entries.  It
shows 0 for split brain.

I'm not quite sure what to do.  I suspect it may be due to nodes 1 and 2
still being on the older ovirt/gluster release, but I'm afraid to upgrade
and reboot them until I have a good gluster sync (don't need to create a
split brain issue).  How do I proceed with this?

Second issue: I've been experiencing VERY POOR performance on most of my
VMs.  To the tune that logging into a windows 10 vm via remote desktop can
take 5 minutes, launching quickbooks inside said vm can easily take 10
minutes.  On some linux VMs, I get random messages like this:
Message from syslogd@unifi at May 28 20:39:23 ...
 kernel:[6171996.308904] NMI watchdog: BUG: soft 

[ovirt-users] Gluster quorum

2018-05-10 Thread Demeter Tibor
Dear Ovirt Users, 
I've followed up the self-hosted-engine upgrade documentation, I upgraded my 
4.1 system to 4.2.3. 
I upgaded the first node with yum upgrade, it seems working now fine. But since 
upgrade, the gluster informations seems to displayed incorrect on the admin 
panel. The volume yellow, and there are red bricks from that node. 
I've checked in console, I think my gluster is not degraded: 

root@n1 ~]# gluster volume list 
volume1 
volume2 
[root@n1 ~]# gluster volume info 
Volume Name: volume1 
Type: Distributed-Replicate 
Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 
Status: Started 
Snapshot Count: 0 
Number of Bricks: 3 x 3 = 9 
Transport-type: tcp 
Bricks: 
Brick1: 10.104.0.1:/gluster/brick/brick1 
Brick2: 10.104.0.2:/gluster/brick/brick1 
Brick3: 10.104.0.3:/gluster/brick/brick1 
Brick4: 10.104.0.1:/gluster/brick/brick2 
Brick5: 10.104.0.2:/gluster/brick/brick2 
Brick6: 10.104.0.3:/gluster/brick/brick2 
Brick7: 10.104.0.1:/gluster/brick/brick3 
Brick8: 10.104.0.2:/gluster/brick/brick3 
Brick9: 10.104.0.3:/gluster/brick/brick3 
Options Reconfigured: 
transport.address-family: inet 
performance.readdir-ahead: on 
nfs.disable: on 
storage.owner-uid: 36 
storage.owner-gid: 36 
performance.quick-read: off 
performance.read-ahead: off 
performance.io-cache: off 
performance.stat-prefetch: off 
performance.low-prio-threads: 32 
network.remote-dio: enable 
cluster.eager-lock: enable 
cluster.quorum-type: auto 
cluster.server-quorum-type: server 
cluster.data-self-heal-algorithm: full 
cluster.locking-scheme: granular 
cluster.shd-max-threads: 8 
cluster.shd-wait-qlength: 1 
features.shard: on 
user.cifs: off 
server.allow-insecure: on 
Volume Name: volume2 
Type: Distributed-Replicate 
Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 
Status: Started 
Snapshot Count: 0 
Number of Bricks: 3 x 3 = 9 
Transport-type: tcp 
Bricks: 
Brick1: 10.104.0.1:/gluster2/brick/brick1 
Brick2: 10.104.0.2:/gluster2/brick/brick1 
Brick3: 10.104.0.3:/gluster2/brick/brick1 
Brick4: 10.104.0.1:/gluster2/brick/brick2 
Brick5: 10.104.0.2:/gluster2/brick/brick2 
Brick6: 10.104.0.3:/gluster2/brick/brick2 
Brick7: 10.104.0.1:/gluster2/brick/brick3 
Brick8: 10.104.0.2:/gluster2/brick/brick3 
Brick9: 10.104.0.3:/gluster2/brick/brick3 
Options Reconfigured: 
nfs.disable: on 
performance.readdir-ahead: on 
transport.address-family: inet 
cluster.quorum-type: auto 
network.ping-timeout: 10 
auth.allow: * 
performance.quick-read: off 
performance.read-ahead: off 
performance.io-cache: off 
performance.stat-prefetch: off 
performance.low-prio-threads: 32 
network.remote-dio: enable 
cluster.eager-lock: enable 
cluster.server-quorum-type: server 
cluster.data-self-heal-algorithm: full 
cluster.locking-scheme: granular 
cluster.shd-max-threads: 8 
cluster.shd-wait-qlength: 1 
features.shard: on 
user.cifs: off 
storage.owner-uid: 36 
storage.owner-gid: 36 
server.allow-insecure: on 
[root@n1 ~]# gluster volume status 
Status of volume: volume1 
Gluster process TCP Port RDMA Port Online Pid 
-- 
Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464 
Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937 
Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506 
Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457 
Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943 
Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514 
Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465 
Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949 
Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520 
Self-heal Daemon on localhost N/A N/A Y 54356 
Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 
Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 
Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 
Task Status of Volume volume1 
-- 
There are no active volume tasks 
Status of volume: volume2 
Gluster process TCP Port RDMA Port Online Pid 
-- 
Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852 
Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955 
Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527 
Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851 
Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961 
Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533 
Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883 
Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968 
Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541 
Self-heal Daemon on localhost N/A N/A Y 54356 
Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 
Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 
Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 
Task Status of Volume volume2 
-- 
There are no active volume tasks 
I think ovirt can't read valid informations about 

[ovirt-users] gluster dispersed volume provisioning

2018-05-09 Thread josip
Hi, have a quick question regarding ovirt UI and provisioning of gluster 
volumes.
I've found an old thread - 
https://lists.ovirt.org/pipermail/users/2015-February/064602.html - where it's 
said that creating dispersed volumes is not supported but that it will be in 
the next release.
Is there any update on when will it be supported?

Thanks
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org


Re: [ovirt-users] [Gluster-users] GlusterFS performance with only one drive per host?

2018-03-25 Thread Vlad Kopylov
The bottleneck is definitely not the disk speed with glusterFS, no
point of using SSD for bricks what so ever

-v

On Thu, Mar 22, 2018 at 6:01 AM, Sahina Bose  wrote:
>
>
> On Mon, Mar 19, 2018 at 5:57 PM, Jayme  wrote:
>>
>> I'm spec'ing a new oVirt build using three Dell R720's w/ 256GB.  I'm
>> considering storage options.  I don't have a requirement for high amounts of
>> storage, I have a little over 1TB to store but want some overhead so I'm
>> thinking 2TB of usable space would be sufficient.
>>
>> I've been doing some research on Micron 1100 2TB ssd's and they seem to
>> offer a lot of value for the money.  I'm considering using smaller cheaper
>> SSDs for boot drives and using one 2TB micron SSD in each host for a
>> glusterFS replica 3 setup (on the fence about using an arbiter, I like the
>> extra redundancy replicate 3 will give me).
>>
>> My question is, would I see a performance hit using only one drive in each
>> host with glusterFS or should I try to add more physical disks.  Such as 6
>> 1TB drives instead of 3 2TB drives?
>
>
> [Adding gluster-users for inputs here]
>
>>
>> Also one other question.  I've read that gluster can only be done in
>> groups of three.  Meaning you need 3, 6, or 9 hosts.  Is this true?  If I
>> had an operational replicate 3 glusterFS setup and wanted to add more
>> capacity I would have to add 3 more hosts, or is it possible for me to add a
>> 4th host in to the mix for extra processing power down the road?
>
>
> In oVirt, we support replica 3 or replica 3 with arbiter (where one of the 3
> bricks is a low storage arbiter brick). To expand storage, you would need to
> add in multiples of 3 bricks. However if you only want to expand compute
> capacity in your HC environment, you can add a 4th node.
>
>>
>> Thanks!
>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>
> ___
> Gluster-users mailing list
> gluster-us...@gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] GlusterFS performance with only one drive per host?

2018-03-25 Thread Manoj Pillai
My take is that unless you have loads of data and are trying to optimize
for cost/TB, HDDs are probably not the right choice. This is particularly
true for random I/O workloads for which HDDs are really quite bad.

I'd recommend a recent gluster release, and some tuning because the default
settings are not optimized for performance. Some options to consider:
client.event-threads
server.event-threads
cluster.choose-local
performance.client-io-threads

You can toggle the last two and see what works for you. You'd probably need
to set event-threads to 4 or more. Ideally you'd tune some of the thread
pools based on observed bottlenecks in collected stats. top (top -bHd 10 >
top_threads.out.txt) is great for this. Using 6 small drives/bricks instead
of 3 is also a good idea to reduce likelihood of rpc bottlenecks.

There has been an effort to improve gluster performance over fast SSDs.
Hence the recommendation to try with a recent release. You can also check
in on some of the issues being worked on:
https://github.com/gluster/glusterfs/issues/412
https://github.com/gluster/glusterfs/issues/410

-- Manoj

On Sat, Mar 24, 2018 at 4:14 AM, Jayme  wrote:

> Do you feel that SSDs are worth the extra cost or am I better off using
> regular HDDs?  I'm looking for the best performance I can get with glusterFS
>
> On Fri, Mar 23, 2018 at 12:03 AM, Manoj Pillai  wrote:
>
>>
>>
>> On Thu, Mar 22, 2018 at 3:31 PM, Sahina Bose  wrote:
>>
>>>
>>>
>>> On Mon, Mar 19, 2018 at 5:57 PM, Jayme  wrote:
>>>
 I'm spec'ing a new oVirt build using three Dell R720's w/ 256GB.  I'm
 considering storage options.  I don't have a requirement for high amounts
 of storage, I have a little over 1TB to store but want some overhead so I'm
 thinking 2TB of usable space would be sufficient.

 I've been doing some research on Micron 1100 2TB ssd's and they seem to
 offer a lot of value for the money.  I'm considering using smaller cheaper
 SSDs for boot drives and using one 2TB micron SSD in each host for a
 glusterFS replica 3 setup (on the fence about using an arbiter, I like the
 extra redundancy replicate 3 will give me).

 My question is, would I see a performance hit using only one drive in
 each host with glusterFS or should I try to add more physical disks.  Such
 as 6 1TB drives instead of 3 2TB drives?

>>>
>> It is possible. With SSDs the rpc layer can become the bottleneck with
>> some workloads, especially if there are not enough connections out to the
>> server side. We had experimented with a multi-connection model for this
>> reason:  https://review.gluster.org/#/c/19133/.
>>
>> -- Manoj
>>
>>>
>>> [Adding gluster-users for inputs here]
>>>
>>>
 Also one other question.  I've read that gluster can only be done in
 groups of three.  Meaning you need 3, 6, or 9 hosts.  Is this true?  If I
 had an operational replicate 3 glusterFS setup and wanted to add more
 capacity I would have to add 3 more hosts, or is it possible for me to add
 a 4th host in to the mix for extra processing power down the road?

>>>
>>> In oVirt, we support replica 3 or replica 3 with arbiter (where one of
>>> the 3 bricks is a low storage arbiter brick). To expand storage, you would
>>> need to add in multiples of 3 bricks. However if you only want to expand
>>> compute capacity in your HC environment, you can add a 4th node.
>>>
>>>
 Thanks!


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


>>>
>>> ___
>>> Gluster-users mailing list
>>> gluster-us...@gluster.org
>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] GlusterFS performance with only one drive per host?

2018-03-25 Thread Manoj Pillai
On Thu, Mar 22, 2018 at 3:31 PM, Sahina Bose  wrote:

>
>
> On Mon, Mar 19, 2018 at 5:57 PM, Jayme  wrote:
>
>> I'm spec'ing a new oVirt build using three Dell R720's w/ 256GB.  I'm
>> considering storage options.  I don't have a requirement for high amounts
>> of storage, I have a little over 1TB to store but want some overhead so I'm
>> thinking 2TB of usable space would be sufficient.
>>
>> I've been doing some research on Micron 1100 2TB ssd's and they seem to
>> offer a lot of value for the money.  I'm considering using smaller cheaper
>> SSDs for boot drives and using one 2TB micron SSD in each host for a
>> glusterFS replica 3 setup (on the fence about using an arbiter, I like the
>> extra redundancy replicate 3 will give me).
>>
>> My question is, would I see a performance hit using only one drive in
>> each host with glusterFS or should I try to add more physical disks.  Such
>> as 6 1TB drives instead of 3 2TB drives?
>>
>
It is possible. With SSDs the rpc layer can become the bottleneck with some
workloads, especially if there are not enough connections out to the server
side. We had experimented with a multi-connection model for this reason:
https://review.gluster.org/#/c/19133/.

-- Manoj

>
> [Adding gluster-users for inputs here]
>
>
>> Also one other question.  I've read that gluster can only be done in
>> groups of three.  Meaning you need 3, 6, or 9 hosts.  Is this true?  If I
>> had an operational replicate 3 glusterFS setup and wanted to add more
>> capacity I would have to add 3 more hosts, or is it possible for me to add
>> a 4th host in to the mix for extra processing power down the road?
>>
>
> In oVirt, we support replica 3 or replica 3 with arbiter (where one of the
> 3 bricks is a low storage arbiter brick). To expand storage, you would need
> to add in multiples of 3 bricks. However if you only want to expand compute
> capacity in your HC environment, you can add a 4th node.
>
>
>> Thanks!
>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
> ___
> Gluster-users mailing list
> gluster-us...@gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] GlusterFS performance with only one drive per host?

2018-03-24 Thread Alex K
I would go with at least 4 HDDs per host in RAID 10. Then focus on network
performance where bottleneck usualy is for gluster.



On Sat, Mar 24, 2018, 00:44 Jayme  wrote:

> Do you feel that SSDs are worth the extra cost or am I better off using
> regular HDDs?  I'm looking for the best performance I can get with glusterFS
>
> On Fri, Mar 23, 2018 at 12:03 AM, Manoj Pillai  wrote:
>
>>
>>
>> On Thu, Mar 22, 2018 at 3:31 PM, Sahina Bose  wrote:
>>
>>>
>>>
>>> On Mon, Mar 19, 2018 at 5:57 PM, Jayme  wrote:
>>>
 I'm spec'ing a new oVirt build using three Dell R720's w/ 256GB.  I'm
 considering storage options.  I don't have a requirement for high amounts
 of storage, I have a little over 1TB to store but want some overhead so I'm
 thinking 2TB of usable space would be sufficient.

 I've been doing some research on Micron 1100 2TB ssd's and they seem to
 offer a lot of value for the money.  I'm considering using smaller cheaper
 SSDs for boot drives and using one 2TB micron SSD in each host for a
 glusterFS replica 3 setup (on the fence about using an arbiter, I like the
 extra redundancy replicate 3 will give me).

 My question is, would I see a performance hit using only one drive in
 each host with glusterFS or should I try to add more physical disks.  Such
 as 6 1TB drives instead of 3 2TB drives?

>>>
>> It is possible. With SSDs the rpc layer can become the bottleneck with
>> some workloads, especially if there are not enough connections out to the
>> server side. We had experimented with a multi-connection model for this
>> reason:  https://review.gluster.org/#/c/19133/.
>>
>> -- Manoj
>>
>>>
>>> [Adding gluster-users for inputs here]
>>>
>>>
 Also one other question.  I've read that gluster can only be done in
 groups of three.  Meaning you need 3, 6, or 9 hosts.  Is this true?  If I
 had an operational replicate 3 glusterFS setup and wanted to add more
 capacity I would have to add 3 more hosts, or is it possible for me to add
 a 4th host in to the mix for extra processing power down the road?

>>>
>>> In oVirt, we support replica 3 or replica 3 with arbiter (where one of
>>> the 3 bricks is a low storage arbiter brick). To expand storage, you would
>>> need to add in multiples of 3 bricks. However if you only want to expand
>>> compute capacity in your HC environment, you can add a 4th node.
>>>
>>>
 Thanks!


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


>>>
>>> ___
>>> Gluster-users mailing list
>>> gluster-us...@gluster.org
>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] GlusterFS performance with only one drive per host?

2018-03-23 Thread Jayme
Do you feel that SSDs are worth the extra cost or am I better off using
regular HDDs?  I'm looking for the best performance I can get with glusterFS

On Fri, Mar 23, 2018 at 12:03 AM, Manoj Pillai  wrote:

>
>
> On Thu, Mar 22, 2018 at 3:31 PM, Sahina Bose  wrote:
>
>>
>>
>> On Mon, Mar 19, 2018 at 5:57 PM, Jayme  wrote:
>>
>>> I'm spec'ing a new oVirt build using three Dell R720's w/ 256GB.  I'm
>>> considering storage options.  I don't have a requirement for high amounts
>>> of storage, I have a little over 1TB to store but want some overhead so I'm
>>> thinking 2TB of usable space would be sufficient.
>>>
>>> I've been doing some research on Micron 1100 2TB ssd's and they seem to
>>> offer a lot of value for the money.  I'm considering using smaller cheaper
>>> SSDs for boot drives and using one 2TB micron SSD in each host for a
>>> glusterFS replica 3 setup (on the fence about using an arbiter, I like the
>>> extra redundancy replicate 3 will give me).
>>>
>>> My question is, would I see a performance hit using only one drive in
>>> each host with glusterFS or should I try to add more physical disks.  Such
>>> as 6 1TB drives instead of 3 2TB drives?
>>>
>>
> It is possible. With SSDs the rpc layer can become the bottleneck with
> some workloads, especially if there are not enough connections out to the
> server side. We had experimented with a multi-connection model for this
> reason:  https://review.gluster.org/#/c/19133/.
>
> -- Manoj
>
>>
>> [Adding gluster-users for inputs here]
>>
>>
>>> Also one other question.  I've read that gluster can only be done in
>>> groups of three.  Meaning you need 3, 6, or 9 hosts.  Is this true?  If I
>>> had an operational replicate 3 glusterFS setup and wanted to add more
>>> capacity I would have to add 3 more hosts, or is it possible for me to add
>>> a 4th host in to the mix for extra processing power down the road?
>>>
>>
>> In oVirt, we support replica 3 or replica 3 with arbiter (where one of
>> the 3 bricks is a low storage arbiter brick). To expand storage, you would
>> need to add in multiples of 3 bricks. However if you only want to expand
>> compute capacity in your HC environment, you can add a 4th node.
>>
>>
>>> Thanks!
>>>
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>> ___
>> Gluster-users mailing list
>> gluster-us...@gluster.org
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] gluster self-heal takes cluster offline

2018-03-23 Thread Darrell Budic
What version of ovirt and gluster? Sounds like something I just saw with 
gluster 3.12.x, are you using libgfapi or just fuse mounts?

> From: Sahina Bose <sab...@redhat.com>
> Subject: Re: [ovirt-users] gluster self-heal takes cluster offline
> Date: March 23, 2018 at 1:26:01 AM CDT
> To: Jim Kusznir
> Cc: Ravishankar Narayanankutty; users
> 
> 
> 
> On Fri, Mar 16, 2018 at 2:45 AM, Jim Kusznir <j...@palousetech.com 
> <mailto:j...@palousetech.com>> wrote:
> Hi all:
> 
> I'm trying to understand why/how (and most importantly, how to fix) an 
> substantial issue I had last night.  This happened one other time, but I 
> didn't know/understand all the parts associated with it until last night.
> 
> I have a 3 node hyperconverged (self-hosted engine, Gluster on each node) 
> cluster.  Gluster is Replica 2 + arbitrar.  Current network configuration is 
> 2x GigE on load balance ("LAG Group" on switch), plus one GigE from each 
> server on a separate vlan, intended for Gluster (but not used).  Server 
> hardware is Dell R610's, each server as an SSD in it.  Server 1 and 2 have 
> the full replica, server 3 is the arbitrar.
> 
> I put server 2 into maintence so I can work on the hardware, including turn 
> it off and such.  In the course of the work, I found that I needed to 
> reconfigure the SSD's partitioning somewhat, and it resulted in wiping the 
> data partition (storing VM images).  I figure, its no big deal, gluster will 
> rebuild that in short order.  I did take care of the extended attr settings 
> and the like, and when I booted it up, gluster came up as expected and began 
> rebuilding the disk.
> 
> How big was the data on this partition? What was the shard size set on the 
> gluster volume?
> Out of curiosity, how long did it take to heal and come back to operational?
> 
> 
> The problem is that suddenly my entire cluster got very sluggish.  The entine 
> was marking nodes and VMs failed and unfaling them throughout the system, 
> fairly randomly.  It didn't matter what node the engine or VM was on.  At one 
> point, it power cycled server 1 for "non-responsive" (even though everything 
> was running on it, and the gluster rebuild was working on it).  As a result 
> of this, about 6 VMs were killed and my entire gluster system went down hard 
> (suspending all remaining VMs and the engine), as there were no remaining 
> full copies of the data.  After several minutes (these are Dell servers, 
> after all...), server 1 came back up, and gluster resumed the rebuild, and 
> came online on the cluster.  I had to manually (virtsh command) unpause the 
> engine, and then struggle through trying to get critical VMs back up.  
> Everything was super slow, and load averages on the servers were often seen 
> in excess of 80 (these are 8 core / 16 thread boxes).  Actual CPU usage 
> (reported by top) was rarely above 40% (inclusive of all CPUs) for any one 
> server. Glusterfs was often seen using 180%-350% of a CPU on server 1 and 2.  
> 
> I ended up putting the cluster in global HA maintence mode and disabling 
> power fencing on the nodes until the process finished.  It appeared on at 
> least two occasions a functional node was marked bad and had the fencing not 
> been disabled, a node would have rebooted, just further exacerbating the 
> problem.  
> 
> Its clear that the gluster rebuild overloaded things and caused the problem.  
> I don't know why the load was so high (even IOWait was low), but load 
> averages were definately tied to the glusterfs cpu utilization %.   At no 
> point did I have any problems pinging any machine (host or VM) unless the 
> engine decided it was dead and killed it.
> 
> Why did my system bite it so hard with the rebuild?  I baby'ed it along until 
> the rebuild was complete, after which it returned to normal operation.
> 
> As of this event, all networking (host/engine management, gluster, and VM 
> network) were on the same vlan.  I'd love to move things off, but so far any 
> attempt to do so breaks my cluster.  How can I move my management interfaces 
> to a separate VLAN/IP Space?  I also want to move Gluster to its own private 
> space, but it seems if I change anything in the peers file, the entire 
> gluster cluster goes down.  The dedicated gluster network is listed as a 
> secondary hostname for all peers already.
> 
> Will the above network reconfigurations be enough?  I got the impression that 
> the issue may not have been purely network based, but possibly server IO 
> overload.  Is this likely / right?
> 
> I appreciate input.  I don't think gluster's recovery is supposed to do as 
> much damage as it did the last two or three times any healing was required.
> 
> Thanks!

Re: [ovirt-users] Gluster: VM disk stuck in transfer; georep gone wonky

2018-03-20 Thread Sahina Bose
On Mon, Mar 19, 2018 at 10:08 PM, Jim Kusznir  wrote:

> Hi all:
>
> Sorry for yet another semi-related message to the list.  In my attempts to
> troubleshoot and verify some suspicions on the nature of the performance
> problems I posted under "Major Performance Issues with gluster", I
> attempted to move one of my problem VM's back to the original storage
> (SSD-backed).  It appeared to be moving fine, but last night froze at 84%.
> This morning (8hrs later), its still at 84%.
>
> I need to get that VM back up and running, but I don't know how...It seems
> to be stuck in limbo.
>
> The only thing I explicitly did last night as well that may have caused an
> issue is finally set up and activated georep to an offsite backup machine.
> That too seems to have gone a bit wonky.  On the ovirt server side, it
> shows normal with all but data-hdd show a last sync'ed time of 3am (which
> matches my bandwidth graphs for the WAN connections involved).  data-hdd
> (the new disk-backed storage with most of my data in it) shows not yet
> synced, but I'm also not currently seeing bandwidth usage anymore.
>
> I logged into the georep destination box, and found system load a bit
> high, a bunch of gluster and rsync processes running, and both data and
> data-hdd using MORE disk space than the origional (data-hdd using 4x more
> disk space than is on the master node).  Not sure what to do about this; I
> paused the replication from the cluster, but that hasn't seem to had an
> effect on the georep destination.
>

For the geo-rep gone wonky - can you provide some more information to debug
this. The logs are at /var/log/glusterfs/geo-replication. Please provide
the logs from the master and slave.


> I promise I'll stop trying things until I get guidance from the list!
> Please do help; I need the VM HDD unstuck so I can start it.
>
> Thanks!
> --Jim
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Gluster: VM disk stuck in transfer; georep gone wonky

2018-03-20 Thread Jim Kusznir
Thank you for the replies.

While waiting, I found one more google responce that said to run
engine-setup.  I did that, and it fixed the issue.  the VM is now running
again.

As to checking the logs, I'm not sure which ones to check...there are so
many in so many different places.

I was not able to detach the disk, as "an operation is currently in
process"  No matter what i did to the disk, it was essentially still
locked, even though it no longer said "locked" after I removed it with the
unlock script.

So, it appears running engine-setup can really fix a bunch of stuff!  An
important tip to remember...

--Jim

On Mon, Mar 19, 2018 at 11:55 PM, Tony Brian Albers  wrote:

> I read somewhere about clearing out wrong stuff from the UI by manually
> editing the database, maybe you can try searching for something like that.
>
> With regards to the VM, I'd probably just delete it, edit the DB and
> remove all sorts of references to it and then recover it from backup.
>
> Is there nothing about all this in the ovirt logs on the engine and the
> host? It might point you in the right direction.
>
> HTH
>
> /tony
>
>
> On 20/03/18 07:48, Jim Kusznir wrote:
> > Unfortunately, I came under heavy pressure to get this vm back up.  So,
> > i did more googling and attempted to recover myself.  I've gotten
> > closer, but still not quite.
> >
> > I found this post:
> >
> > http://lists.ovirt.org/pipermail/users/2015-November/035686.html
> >
> > Which gave me the unlock tool, which was successful in unlocking the
> > disk.  Unfortunately, it did not delete the task, nor did ovirt do so on
> > its own after the disk was unlocked.
> >
> > So I found the taskcleaner.sh in the same directory and attempted to
> > clean the task outexcept it doesn't seem to see the task (none of
> > the show tasks options seemed to work or the delete all options).  I did
> > still have the task uuid from the gui, so i attempted to use that, but
> > all I got back was a "t" on one line and a "0" on the next, so I have no
> > idea what that was supposed to mean.  In any case, the web UI still
> > shows the task, still won't let me start the VM and appears convinced
> > its still copying.  I've tried restarting the engine and vdsm on the
> > SPM, neither have helped.  I can't find any evidence of the task on the
> > command line; only in the UI.
> >
> > I'd create a new VM if i could rescue the image, but I'm not sure I can
> > manage to get this image accepted in another VM
> >
> > How do i recover now?
> >
> > --Jim
> >
> > On Mon, Mar 19, 2018 at 9:38 AM, Jim Kusznir  > > wrote:
> >
> > Hi all:
> >
> > Sorry for yet another semi-related message to the list.  In my
> > attempts to troubleshoot and verify some suspicions on the nature of
> > the performance problems I posted under "Major Performance Issues
> > with gluster", I attempted to move one of my problem VM's back to
> > the original storage (SSD-backed).  It appeared to be moving fine,
> > but last night froze at 84%.  This morning (8hrs later), its still
> > at 84%.
> >
> > I need to get that VM back up and running, but I don't know how...It
> > seems to be stuck in limbo.
> >
> > The only thing I explicitly did last night as well that may have
> > caused an issue is finally set up and activated georep to an offsite
> > backup machine.  That too seems to have gone a bit wonky.  On the
> > ovirt server side, it shows normal with all but data-hdd show a last
> > sync'ed time of 3am (which matches my bandwidth graphs for the WAN
> > connections involved).  data-hdd (the new disk-backed storage with
> > most of my data in it) shows not yet synced, but I'm also not
> > currently seeing bandwidth usage anymore.
> >
> > I logged into the georep destination box, and found system load a
> > bit high, a bunch of gluster and rsync processes running, and both
> > data and data-hdd using MORE disk space than the origional (data-hdd
> > using 4x more disk space than is on the master node).  Not sure what
> > to do about this; I paused the replication from the cluster, but
> > that hasn't seem to had an effect on the georep destination.
> >
> > I promise I'll stop trying things until I get guidance from the
> > list!  Please do help; I need the VM HDD unstuck so I can start it.
> >
> > Thanks!
> > --Jim
> >
> >
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
>
>
> --
> Tony Albers
> Systems administrator, IT-development
> Royal Danish Library, Victor Albecks Vej 1, 8000 Aarhus C, Denmark.
> Tel: +45 2566 2383 / +45 8946 2316
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users 

Re: [ovirt-users] Gluster: VM disk stuck in transfer; georep gone wonky

2018-03-20 Thread Eyal Shenitzky
Can you please check if you can detach the disk from the VM and attach it
to the created VM?

On Tue, Mar 20, 2018 at 8:48 AM, Jim Kusznir  wrote:

> Unfortunately, I came under heavy pressure to get this vm back up.  So, i
> did more googling and attempted to recover myself.  I've gotten closer, but
> still not quite.
>
> I found this post:
>
> http://lists.ovirt.org/pipermail/users/2015-November/035686.html
>
> Which gave me the unlock tool, which was successful in unlocking the
> disk.  Unfortunately, it did not delete the task, nor did ovirt do so on
> its own after the disk was unlocked.
>
> So I found the taskcleaner.sh in the same directory and attempted to clean
> the task outexcept it doesn't seem to see the task (none of the show
> tasks options seemed to work or the delete all options).  I did still have
> the task uuid from the gui, so i attempted to use that, but all I got back
> was a "t" on one line and a "0" on the next, so I have no idea what that
> was supposed to mean.  In any case, the web UI still shows the task, still
> won't let me start the VM and appears convinced its still copying.  I've
> tried restarting the engine and vdsm on the SPM, neither have helped.  I
> can't find any evidence of the task on the command line; only in the UI.
>
> I'd create a new VM if i could rescue the image, but I'm not sure I can
> manage to get this image accepted in another VM
>
> How do i recover now?
>
> --Jim
>
> On Mon, Mar 19, 2018 at 9:38 AM, Jim Kusznir  wrote:
>
>> Hi all:
>>
>> Sorry for yet another semi-related message to the list.  In my attempts
>> to troubleshoot and verify some suspicions on the nature of the performance
>> problems I posted under "Major Performance Issues with gluster", I
>> attempted to move one of my problem VM's back to the original storage
>> (SSD-backed).  It appeared to be moving fine, but last night froze at 84%.
>> This morning (8hrs later), its still at 84%.
>>
>> I need to get that VM back up and running, but I don't know how...It
>> seems to be stuck in limbo.
>>
>> The only thing I explicitly did last night as well that may have caused
>> an issue is finally set up and activated georep to an offsite backup
>> machine.  That too seems to have gone a bit wonky.  On the ovirt server
>> side, it shows normal with all but data-hdd show a last sync'ed time of 3am
>> (which matches my bandwidth graphs for the WAN connections involved).
>> data-hdd (the new disk-backed storage with most of my data in it) shows not
>> yet synced, but I'm also not currently seeing bandwidth usage anymore.
>>
>> I logged into the georep destination box, and found system load a bit
>> high, a bunch of gluster and rsync processes running, and both data and
>> data-hdd using MORE disk space than the origional (data-hdd using 4x more
>> disk space than is on the master node).  Not sure what to do about this; I
>> paused the replication from the cluster, but that hasn't seem to had an
>> effect on the georep destination.
>>
>> I promise I'll stop trying things until I get guidance from the list!
>> Please do help; I need the VM HDD unstuck so I can start it.
>>
>> Thanks!
>> --Jim
>>
>>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>


-- 
Regards,
Eyal Shenitzky
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Gluster: VM disk stuck in transfer; georep gone wonky

2018-03-20 Thread Eyal Shenitzky
Hi,

Can you please send the VDSM and Engine log?

Thanks

On Mon, Mar 19, 2018 at 6:38 PM, Jim Kusznir  wrote:

> Hi all:
>
> Sorry for yet another semi-related message to the list.  In my attempts to
> troubleshoot and verify some suspicions on the nature of the performance
> problems I posted under "Major Performance Issues with gluster", I
> attempted to move one of my problem VM's back to the original storage
> (SSD-backed).  It appeared to be moving fine, but last night froze at 84%.
> This morning (8hrs later), its still at 84%.
>
> I need to get that VM back up and running, but I don't know how...It seems
> to be stuck in limbo.
>
> The only thing I explicitly did last night as well that may have caused an
> issue is finally set up and activated georep to an offsite backup machine.
> That too seems to have gone a bit wonky.  On the ovirt server side, it
> shows normal with all but data-hdd show a last sync'ed time of 3am (which
> matches my bandwidth graphs for the WAN connections involved).  data-hdd
> (the new disk-backed storage with most of my data in it) shows not yet
> synced, but I'm also not currently seeing bandwidth usage anymore.
>
> I logged into the georep destination box, and found system load a bit
> high, a bunch of gluster and rsync processes running, and both data and
> data-hdd using MORE disk space than the origional (data-hdd using 4x more
> disk space than is on the master node).  Not sure what to do about this; I
> paused the replication from the cluster, but that hasn't seem to had an
> effect on the georep destination.
>
> I promise I'll stop trying things until I get guidance from the list!
> Please do help; I need the VM HDD unstuck so I can start it.
>
> Thanks!
> --Jim
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>


-- 
Regards,
Eyal Shenitzky
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Gluster: VM disk stuck in transfer; georep gone wonky

2018-03-20 Thread Jim Kusznir
Unfortunately, I came under heavy pressure to get this vm back up.  So, i
did more googling and attempted to recover myself.  I've gotten closer, but
still not quite.

I found this post:

http://lists.ovirt.org/pipermail/users/2015-November/035686.html

Which gave me the unlock tool, which was successful in unlocking the disk.
Unfortunately, it did not delete the task, nor did ovirt do so on its own
after the disk was unlocked.

So I found the taskcleaner.sh in the same directory and attempted to clean
the task outexcept it doesn't seem to see the task (none of the show
tasks options seemed to work or the delete all options).  I did still have
the task uuid from the gui, so i attempted to use that, but all I got back
was a "t" on one line and a "0" on the next, so I have no idea what that
was supposed to mean.  In any case, the web UI still shows the task, still
won't let me start the VM and appears convinced its still copying.  I've
tried restarting the engine and vdsm on the SPM, neither have helped.  I
can't find any evidence of the task on the command line; only in the UI.

I'd create a new VM if i could rescue the image, but I'm not sure I can
manage to get this image accepted in another VM

How do i recover now?

--Jim

On Mon, Mar 19, 2018 at 9:38 AM, Jim Kusznir  wrote:

> Hi all:
>
> Sorry for yet another semi-related message to the list.  In my attempts to
> troubleshoot and verify some suspicions on the nature of the performance
> problems I posted under "Major Performance Issues with gluster", I
> attempted to move one of my problem VM's back to the original storage
> (SSD-backed).  It appeared to be moving fine, but last night froze at 84%.
> This morning (8hrs later), its still at 84%.
>
> I need to get that VM back up and running, but I don't know how...It seems
> to be stuck in limbo.
>
> The only thing I explicitly did last night as well that may have caused an
> issue is finally set up and activated georep to an offsite backup machine.
> That too seems to have gone a bit wonky.  On the ovirt server side, it
> shows normal with all but data-hdd show a last sync'ed time of 3am (which
> matches my bandwidth graphs for the WAN connections involved).  data-hdd
> (the new disk-backed storage with most of my data in it) shows not yet
> synced, but I'm also not currently seeing bandwidth usage anymore.
>
> I logged into the georep destination box, and found system load a bit
> high, a bunch of gluster and rsync processes running, and both data and
> data-hdd using MORE disk space than the origional (data-hdd using 4x more
> disk space than is on the master node).  Not sure what to do about this; I
> paused the replication from the cluster, but that hasn't seem to had an
> effect on the georep destination.
>
> I promise I'll stop trying things until I get guidance from the list!
> Please do help; I need the VM HDD unstuck so I can start it.
>
> Thanks!
> --Jim
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] gluster self-heal takes cluster offline

2018-03-15 Thread Jim Kusznir
Hi all:

I'm trying to understand why/how (and most importantly, how to fix) an
substantial issue I had last night.  This happened one other time, but I
didn't know/understand all the parts associated with it until last night.

I have a 3 node hyperconverged (self-hosted engine, Gluster on each node)
cluster.  Gluster is Replica 2 + arbitrar.  Current network configuration
is 2x GigE on load balance ("LAG Group" on switch), plus one GigE from each
server on a separate vlan, intended for Gluster (but not used).  Server
hardware is Dell R610's, each server as an SSD in it.  Server 1 and 2 have
the full replica, server 3 is the arbitrar.

I put server 2 into maintence so I can work on the hardware, including turn
it off and such.  In the course of the work, I found that I needed to
reconfigure the SSD's partitioning somewhat, and it resulted in wiping the
data partition (storing VM images).  I figure, its no big deal, gluster
will rebuild that in short order.  I did take care of the extended attr
settings and the like, and when I booted it up, gluster came up as expected
and began rebuilding the disk.

The problem is that suddenly my entire cluster got very sluggish.  The
entine was marking nodes and VMs failed and unfaling them throughout the
system, fairly randomly.  It didn't matter what node the engine or VM was
on.  At one point, it power cycled server 1 for "non-responsive" (even
though everything was running on it, and the gluster rebuild was working on
it).  As a result of this, about 6 VMs were killed and my entire gluster
system went down hard (suspending all remaining VMs and the engine), as
there were no remaining full copies of the data.  After several minutes
(these are Dell servers, after all...), server 1 came back up, and gluster
resumed the rebuild, and came online on the cluster.  I had to manually
(virtsh command) unpause the engine, and then struggle through trying to
get critical VMs back up.  Everything was super slow, and load averages on
the servers were often seen in excess of 80 (these are 8 core / 16 thread
boxes).  Actual CPU usage (reported by top) was rarely above 40% (inclusive
of all CPUs) for any one server. Glusterfs was often seen using 180%-350%
of a CPU on server 1 and 2.

I ended up putting the cluster in global HA maintence mode and disabling
power fencing on the nodes until the process finished.  It appeared on at
least two occasions a functional node was marked bad and had the fencing
not been disabled, a node would have rebooted, just further exacerbating
the problem.

Its clear that the gluster rebuild overloaded things and caused the
problem.  I don't know why the load was so high (even IOWait was low), but
load averages were definately tied to the glusterfs cpu utilization %.   At
no point did I have any problems pinging any machine (host or VM) unless
the engine decided it was dead and killed it.

Why did my system bite it so hard with the rebuild?  I baby'ed it along
until the rebuild was complete, after which it returned to normal operation.

As of this event, all networking (host/engine management, gluster, and VM
network) were on the same vlan.  I'd love to move things off, but so far
any attempt to do so breaks my cluster.  How can I move my management
interfaces to a separate VLAN/IP Space?  I also want to move Gluster to its
own private space, but it seems if I change anything in the peers file, the
entire gluster cluster goes down.  The dedicated gluster network is listed
as a secondary hostname for all peers already.

Will the above network reconfigurations be enough?  I got the impression
that the issue may not have been purely network based, but possibly server
IO overload.  Is this likely / right?

I appreciate input.  I don't think gluster's recovery is supposed to do as
much damage as it did the last two or three times any healing was required.

Thanks!
--Jim
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Gluster Snapshot Schedule Failing on 4.2.1

2018-03-08 Thread Hesham Ahmed
Log file attached to the bug. Do let me know if you need anything else.

On Thu, Mar 8, 2018, 4:32 PM Sahina Bose  wrote:

> Thanks for your report, we will take a look. Could you attach the
> engine.log to the bug?
>
> On Wed, Mar 7, 2018 at 11:20 PM, Hesham Ahmed  wrote:
>
>> I am having issues with the Gluster Snapshot UI since upgrade to 4.2 and
>> now with 4.2.1. The UI doesn't appear as I explained in the bug report:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1530186
>>
>> I can now see the UI when I clear the cookies and try the snapshots UI
>> from within the volume details screen, however scheduled snapshots are not
>> being created. The engine log shows a single error:
>>
>> 2018-03-07 20:00:00,051+03 ERROR
>> [org.ovirt.engine.core.utils.timer.JobWrapper] (QuartzOvirtDBScheduler1)
>> [12237b15] Failed to invoke scheduled method onTimer: null
>>
>> Anyone scheduling snapshots successfully wtih 4.2?
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Gluster Snapshot Schedule Failing on 4.2.1

2018-03-08 Thread Sahina Bose
Thanks for your report, we will take a look. Could you attach the
engine.log to the bug?

On Wed, Mar 7, 2018 at 11:20 PM, Hesham Ahmed  wrote:

> I am having issues with the Gluster Snapshot UI since upgrade to 4.2 and
> now with 4.2.1. The UI doesn't appear as I explained in the bug report:
> https://bugzilla.redhat.com/show_bug.cgi?id=1530186
>
> I can now see the UI when I clear the cookies and try the snapshots UI
> from within the volume details screen, however scheduled snapshots are not
> being created. The engine log shows a single error:
>
> 2018-03-07 20:00:00,051+03 ERROR 
> [org.ovirt.engine.core.utils.timer.JobWrapper]
> (QuartzOvirtDBScheduler1) [12237b15] Failed to invoke scheduled method
> onTimer: null
>
> Anyone scheduling snapshots successfully wtih 4.2?
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Gluster Snapshot Schedule Failing on 4.2.1

2018-03-07 Thread Hesham Ahmed
I am having issues with the Gluster Snapshot UI since upgrade to 4.2 and
now with 4.2.1. The UI doesn't appear as I explained in the bug report:
https://bugzilla.redhat.com/show_bug.cgi?id=1530186

I can now see the UI when I clear the cookies and try the snapshots UI from
within the volume details screen, however scheduled snapshots are not being
created. The engine log shows a single error:

2018-03-07 20:00:00,051+03 ERROR
[org.ovirt.engine.core.utils.timer.JobWrapper] (QuartzOvirtDBScheduler1)
[12237b15] Failed to invoke scheduled method onTimer: null

Anyone scheduling snapshots successfully wtih 4.2?
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


  1   2   3   4   5   >