I encountered this after upgrading clients to 3.12.9 as well. It’s not present 
in 3.12.8 or 3.12.6. I’ve added some data I had to that bug, can produce more 
if needed. Forgot to mention my server cluster is at 3.12.9, and is not showing 
any problems, it’s just the clients.

A test cluster on 3.12.11 also shows it, just slower because it’s got fewer 
clients on it.


> From: Sahina Bose <[email protected]>
> Subject: [ovirt-users] Re: Ovirt cluster unstable; gluster to blame (again)
> Date: July 9, 2018 at 10:42:15 AM CDT
> To: Edward Clay; Jim Kusznir
> Cc: users
> 
> see response about bug at 
> https://lists.ovirt.org/archives/list/[email protected]/thread/WRYEBOLNHJZGKKJUNF77TJ7WMBS66ZYK/
>  
> <https://lists.ovirt.org/archives/list/[email protected]/thread/WRYEBOLNHJZGKKJUNF77TJ7WMBS66ZYK/>
>  which seems to indicate the referenced bug is fixed at 3.12.2 and higher.
> 
> Could you attach the statedump of the process to the bug 
> https://bugzilla.redhat.com/show_bug.cgi?id=1593826 
> <https://bugzilla.redhat.com/show_bug.cgi?id=1593826> as requested?
> 
> 
> 
> On Mon, Jul 9, 2018 at 8:38 PM, Edward Clay <[email protected] 
> <mailto:[email protected]>> wrote:
> Just to add my .02 here.  I've opened a bug on this issue where HV/host 
> connected to clusterfs volumes are running out of ram.  This seemed to be a 
> bug fixed in gluster 3.13 but that patch doesn't seem to be avaiable any 
> longer and 3.12 is what ovirt is using.  For example I have a host that was 
> showing 72% of memory consumption with 3 VMs running on it.  If I migrate 
> those VMs to another Host memory consumption drops to 52%.  If i put this 
> host into maintenance and then activate it it drops down to 2% or so.  Since 
> I ran into this issue I've been manually watching memory consumption on each 
> host and migrating VMs from it to others to keep things from dying.  I'm 
> hoping with the announcement of gluster 3.12 end of life and the move to 
> gluster 4.1 that this will get fixed or that the patch from 3.13 can get 
> backported so this problem will go away.
> 
> https://bugzilla.redhat.com/show_bug.cgi?id=1593826 
> <https://bugzilla.redhat.com/show_bug.cgi?id=1593826>
> 
> On 07/07/2018 11:49 AM, Jim Kusznir wrote:
>> **Security Notice - This external email is NOT from The Hut Group** 
>> 
>> This host has NO VMs running on it, only 3 running cluster-wide (including 
>> the engine, which is on its own storage):
>> 
>> top - 10:44:41 up 1 day, 17:10,  1 user,  load average: 15.86, 14.33, 13.39
>> Tasks: 381 total,   1 running, 379 sleeping,   1 stopped,   0 zombie
>> %Cpu(s):  2.7 us,  2.1 sy,  0.0 ni, 89.0 id,  6.1 wa,  0.0 hi,  0.2 si,  0.0 
>> st
>> KiB Mem : 32764284 total,   338232 free,   842324 used, 31583728 buff/cache
>> KiB Swap: 12582908 total, 12258660 free,   324248 used. 31076748 avail Mem 
>> 
>>   PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND  
>>                                                                              
>>                            
>> 13279 root      20   0 2380708  37628   4396 S  51.7  0.1   3768:03 
>> glusterfsd                                                                   
>>                                      
>> 13273 root      20   0 2233212  20460   4380 S  17.2  0.1 105:50.44 
>> glusterfsd                                                                   
>>                                     
>> 13287 root      20   0 2233212  20608   4340 S   4.3  0.1  34:27.20 
>> glusterfsd                                                                   
>>                                     
>> 16205 vdsm       0 -20 5048672  88940  13364 S   1.3  0.3   0:32.69 vdsmd    
>>                                                                              
>>                            
>> 16300 vdsm      20   0  608488  25096   5404 S   1.3  0.1   0:05.78 python   
>>                                                                              
>>                            
>>  1109 vdsm      20   0 3127696  44228   8552 S   0.7  0.1  18:49.76 
>> ovirt-ha-broker                                                              
>>                                      
>> 25555 root      20   0       0      0      0 S   0.7  0.0   0:00.13 
>> kworker/u64:3                                                                
>>                                      
>>    10 root      20   0       0      0      0 S   0.3  0.0   4:22.36 
>> rcu_sched                                                                    
>>                                      
>>   572 root       0 -20       0      0      0 S   0.3  0.0   0:12.02 
>> kworker/1:1H                                                                 
>>                                      
>>   797 root      20   0       0      0      0 S   0.3  0.0   1:59.59 
>> kdmwork-253:2                                                                
>>                                      
>>   877 root       0 -20       0      0      0 S   0.3  0.0   0:11.34 
>> kworker/3:1H                                                                 
>>                                      
>>  1028 root      20   0       0      0      0 S   0.3  0.0   0:35.35 
>> xfsaild/dm-10                                                                
>>                                      
>>  1869 root      20   0 1496472  10540   6564 S   0.3  0.0   2:15.46 python   
>>                                                                              
>>                            
>>  3747 root      20   0       0      0      0 D   0.3  0.0   0:01.21 
>> kworker/u64:1                                                                
>>                                      
>> 10979 root      15  -5  723504  15644   3920 S   0.3  0.0  22:46.27 
>> glusterfs                                                                    
>>                                     
>> 15085 root      20   0  680884  10792   4328 S   0.3  0.0   0:01.13 glusterd 
>>                                                                              
>>                            
>> 16102 root      15  -5 1204216  44948  11160 S   0.3  0.1   0:18.61 
>> supervdsmd                                     
>> 
>> At the moment, the engine is barely usable, my other VMs appear to be 
>> unresponsive.  Two on one host, one on another, and none on the third.
>> 
>> 
>> 
>> On Sat, Jul 7, 2018 at 10:38 AM, Jim Kusznir <[email protected] 
>> <mailto:[email protected]>> wrote:
>> I run 4-7 VMs, and most of them are 2GB ram.  I have 2 VMs with 4GB.
>> 
>> Ram hasn't been an issue until recent ovirt/gluster upgrades.  Storage has 
>> always been slow, especially with these drives.  However, even watching 
>> network utilization on my switch, the gig-e links never max out.
>> 
>> The loadavg issues and unresponsive behavior started with yesterday's ovirt 
>> updates.  I now have one VM with low I/O that lives on a separate storage 
>> volume (data, fully SSD backed instead of data-hdd, which was having the 
>> issues).  I moved it to a ovirt host with no other VMs on it, and that had 
>> reshly been rebooted.  Before it had this one VM on it, loadavg was >0.5.  
>> Now its up in the 20's, with only one low Disk I/O, 4GB ram VM on the host.
>> 
>> This to me says there's now a new problem separate from Gluster.  I don't 
>> have any non-gluster storage available to test with.  I did notice that the 
>> last update included a new kernel, and it appears its the qemu-kvm processes 
>> that are consuming way more CPU than they used to now.
>> 
>> Are there any known issues?  I'm going to reboot into my previous kernel to 
>> see if its kernel-caused.
>> 
>> --Jim
>> 
>> 
>> 
>> On Fri, Jul 6, 2018 at 11:07 PM, Johan Bernhardsson <[email protected] 
>> <mailto:[email protected]>> wrote:
>> That is a single sata drive that is slow on random I/O and that has to be 
>> synced with 2 other servers. Gluster works syncronous so one write has to be 
>> written and acknowledged on all the three nodes.
>> 
>> So you have a bottle neck in io on drives and one on network and depending 
>> on how many virtual servers you have and how much ram they take you might 
>> have memory.
>> 
>> Load spikes when you have a wait somewhere and are overusing capacity. But 
>> it's now only CPU that load is counted on. It is waiting for resources so it 
>> can be memory or Network or drives.
>> 
>> How many virtual server do you run and how much ram do they consume?
>> 
>> On July 7, 2018 09:51:42 Jim Kusznir <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>>> In case it matters, the data-hdd gluster volume uses these hard drives:
>>> 
>>> https://www.amazon.com/gp/product/B01M1NHCZT/ref=oh_aui_detailpage_o05_s00?ie=UTF8&psc=1
>>>  
>>> <https://www.amazon.com/gp/product/B01M1NHCZT/ref=oh_aui_detailpage_o05_s00?ie=UTF8&psc=1>
>>> 
>>> This is in a Dell R610 with PERC6/i (one drive per server, configured as a 
>>> single drive volume to pass it through as its own /dev/sd* device).  Inside 
>>> the OS, its partitioned with lvm_thin, then an lvm volume formatted with 
>>> XFS and mounted as /gluster/brick3, with the data-hdd volume created inside 
>>> that.
>>> 
>>> --Jim
>>> 
>>> On Fri, Jul 6, 2018 at 10:45 PM, Jim Kusznir <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> So, I'm still at a loss...It sounds like its either insufficient ram/swap, 
>>> or insufficient network.  It seems to be neither now.  At this point, it 
>>> appears that gluster is just "broke" and killing my systems for no 
>>> descernable reason.  Here's detals, all from the same system (currently 
>>> running 3 VMs):
>>> 
>>> [root@ovirt3 ~]# w
>>>  22:26:53 up 36 days,  4:34,  1 user,  load average: 42.78, 55.98, 53.31
>>> USER     TTY      FROM             LOGIN@   IDLE   JCPU   PCPU WHAT
>>> root     pts/0    192.168.8.90     22:26    2.00s  0.12s  0.11s w
>>> 
>>> bwm-ng reports the highest data usage was about 6MB/s during this test (and 
>>> that was combined; I have two different gig networks.  One gluster network 
>>> (primary VM storage) runs on one, the other network handles everything 
>>> else).
>>> 
>>> [root@ovirt3 ~]# free -m
>>>               total        used        free      shared  buff/cache   
>>> available
>>> Mem:          31996       13236         232          18       18526       
>>> 18195
>>> Swap:         16383        1475       14908
>>> 
>>> top - 22:32:56 up 36 days,  4:41,  1 user,  load average: 17.99, 39.69, 
>>> 47.66
>>> Tasks: 407 total,   1 running, 405 sleeping,   1 stopped,   0 zombie
>>> %Cpu(s):  8.6 us,  2.1 sy,  0.0 ni, 87.6 id,  1.6 wa,  0.0 hi,  0.1 si,  
>>> 0.0 st
>>> KiB Mem : 32764284 total,   228296 free, 13541952 used, 18994036 buff/cache
>>> KiB Swap: 16777212 total, 15246200 free,  1531012 used. 18643960 avail Mem 
>>> 
>>>   PID USER      PR  NI    VIRT    RES    SHR S  %CPU %MEM     TIME+ COMMAND 
>>>                                                                             
>>>                              
>>> 30036 qemu      20   0 6872324   5.2g  13532 S 144.6 16.5 216:14.55 
>>> /usr/libexec/qemu-kvm -name guest=BillingWin,debug-threads=on -S -object 
>>> secret,id=masterKey0,format=raw,file=/v+ 
>>> 28501 qemu      20   0 5034968   3.6g  12880 S  16.2 11.7  73:44.99 
>>> /usr/libexec/qemu-kvm -name guest=FusionPBX,debug-threads=on -S -object 
>>> secret,id=masterKey0,format=raw,file=/va+ 
>>>  2694 root      20   0 2169224  12164   3108 S   5.0  0.0   3290:42 
>>> /usr/sbin/glusterfsd -s ovirt3.nwfiber.com <http://ovirt3.nwfiber.com/> 
>>> --volfile-iddata.ovirt3.nwfiber.com 
>>> <http://data.ovirt3.nwfiber.com/>.gluster-brick2-data -p /var/run/+ 
>>> 14293 root      15  -5  944700  13356   4436 S   4.0  0.0  16:32.15 
>>> /usr/sbin/glusterfs --volfile-server=192.168.8.11 
>>> --volfile-server=192.168.8.12 --volfile-server=192.168.8.13 --+ 
>>> 25100 vdsm       0 -20 6747440 107868  12836 S   2.3  0.3  21:35.20 
>>> /usr/bin/python2 /usr/share/vdsm/vdsmd                                      
>>>                                       
>>> 28971 qemu      20   0 2842592   1.5g  13548 S   1.7  4.7 241:46.49 
>>> /usr/libexec/qemu-kvm -name guest=unifi.palousetech.com 
>>> <http://unifi.palousetech.com/>,debug-threads=on -S -object 
>>> secret,id=masterKey0,format=+ 
>>> 12095 root      20   0  162276   2836   1868 R   1.3  0.0   0:00.25 top     
>>>                                                                             
>>>                               
>>>  2708 root      20   0 1906040  12404   3080 S   1.0  0.0   1083:33 
>>> /usr/sbin/glusterfsd -s ovirt3.nwfiber.com <http://ovirt3.nwfiber.com/> 
>>> --volfile-idengine.ovirt3.nwfiber.com 
>>> <http://engine.ovirt3.nwfiber.com/>.gluster-brick1-engine -p /var/+ 
>>> 28623 qemu      20   0 4749536   1.7g  12896 S   0.7  5.5   4:30.64 
>>> /usr/libexec/qemu-kvm -name guest=billing.nwfiber.com 
>>> <http://billing.nwfiber.com/>,debug-threads=on -S -object 
>>> secret,id=masterKey0,format=ra+ 
>>>    10 root      20   0       0      0      0 S   0.3  0.0 215:54.72 
>>> [rcu_sched]                                                                 
>>>                                      
>>>  1030 sanlock   rt   0  773804  27908   2744 S   0.3  0.1  35:55.61 
>>> /usr/sbin/sanlock daemon                                                    
>>>                                       
>>>  1890 zabbix    20   0   83904   1696   1612 S   0.3  0.0  24:30.63 
>>> /usr/sbin/zabbix_agentd: collector [idle 1 sec]                             
>>>                                      
>>>  2722 root      20   0 1298004   6148   2580 S   0.3  0.0  38:10.82 
>>> /usr/sbin/glusterfsd -s ovirt3.nwfiber.com <http://ovirt3.nwfiber.com/> 
>>> --volfile-idiso.ovirt3.nwfiber.com 
>>> <http://iso.ovirt3.nwfiber.com/>.gluster-brick4-iso -p /var/run/gl+ 
>>>  6340 root      20   0       0      0      0 S   0.3  0.0   0:04.30 
>>> [kworker/7:0]                                                               
>>>                                      
>>> 10652 root      20   0       0      0      0 S   0.3  0.0   0:00.23 
>>> [kworker/u64:2]                                                             
>>>                                       
>>> 14724 root      20   0 1076344  17400   3200 S   0.3  0.1  10:04.13 
>>> /usr/sbin/glusterfs -s localhost --volfile-id gluster/glustershd -p 
>>> /var/run/gluster/glustershd/glustershd.pid <http://ustershd.pid/> -+ 
>>> 22011 root      20   0       0      0      0 S   0.3  0.0   0:05.04 
>>> [kworker/10:1]                                                              
>>>                                       
>>> 
>>> Not sure why the system load dropped other than I was trying to take a 
>>> picture of it :)
>>> 
>>> In any case, it appears that at this time, I have plenty of swap, ram, and 
>>> network capacity, and yet things are still running very sluggish; I'm still 
>>> getting e-mails from servers complaining about loss of communication with 
>>> something or another; I still get e-mails from the engine about bad engine 
>>> status, then recovery, etc.
>>> 
>>> I've shut down 2/3 of my VMs, too....just trying to keep the critical ones 
>>> operating.
>>> 
>>> At this point, I don't believe the problem is the memory leak, but it seems 
>>> to be triggered by the memory leak, as in all my problems started when I 
>>> got low ram warnings from one of my 3 nodes and began recovery efforts from 
>>> that.
>>> 
>>> I do really like the idea / concept behind glusterfs, but I really have to 
>>> figure out why its been so poor performing from day one, and its caused 95% 
>>> of my outages (including several large ones lately).  If I can get it 
>>> stable, reliable, and well performing, then I'd love to keep it.  If I 
>>> can't, then perhaps NFS is the way to go?  I don't like the single point of 
>>> failure aspect of it, but my other NAS boxes I run for clients (central 
>>> storage for windows boxes) have been very solid; If I could get that kind 
>>> of reliability for my ovirt stack, it would be a substantial improvement.  
>>> Currently, it seems about every other month I have a gluster-induced outage.
>>> 
>>> Sometimes I wonder if its just hyperconverged is the issue, but my 
>>> infrastructure doesn't justify three servers at the same location...I might 
>>> be able to do two, but even that seems like its pushing it.
>>> 
>>> Looks like I can upgrade to 10G for about $900.  I can order a dual-Xeon 
>>> supermicro 12-disk server, loaded with 2TB WD Enterprise disks and a pair 
>>> of SSDs for the os, 32GB ram, 2.67Ghz CPUs for about $720 delivered.  I've 
>>> got to do something to improve my reliability; I can't keep going the way I 
>>> have been....
>>> 
>>> --Jim
>>> 
>>> 
>>> On Fri, Jul 6, 2018 at 9:13 PM, Johan Bernhardsson <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> Load like that is mostly io based either the machine is swapping or network 
>>> is to slow. Check I/o wait in top.
>>> 
>>> And the problem where you get oom killer to kill off gluster. That means 
>>> that you don't monitor ram usage on the servers? Either it's eating all 
>>> your ram and swap gets really io intensive and then is killed off. Or you 
>>> have the wrong swap settings in sysctl.conf (there are tons of broken 
>>> guides that recommends swappines to 0 but that disables swap on newer 
>>> kernels. The proper swappines for only swapping when nesseary is 1 or a 
>>> sufficiently low number like 10 default is 60)
>>> 
>>> 
>>> Moving to nfs will not improve things. You will get more memory since 
>>> gluster isn't running and that is good. But you will have a single node 
>>> that can fail with all your storage and it would still be on 1 gigabit only 
>>> and your three node cluster would easily saturate that link.
>>> 
>>> On July 7, 2018 04:13:13 Jim Kusznir <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> 
>>>> So far it does not appear to be helping much. I'm still getting VM's 
>>>> locking up and all kinds of notices from overt engine about non-responsive 
>>>> hosts.  I'm still seeing load averages in the 20-30 range.
>>>> 
>>>> Jim
>>>> 
>>>> On Fri, Jul 6, 2018, 3:13 PM Jim Kusznir <[email protected] 
>>>> <mailto:[email protected]>> wrote:
>>>> Thank you for the advice and help
>>>> 
>>>> I do plan on going 10Gbps networking; haven't quite jumped off that cliff 
>>>> yet, though.
>>>> 
>>>> I did put my data-hdd (main VM storage volume) onto a dedicated 1Gbps 
>>>> network, and I've watched throughput on that and never seen more than 
>>>> 60GB/s achieved (as reported by bwm-ng).  I have a separate 1Gbps network 
>>>> for communication and ovirt migration, but I wanted to break that up 
>>>> further (separate out VM traffice from migration/mgmt traffic).  My three 
>>>> SSD-backed gluster volumes run the main network too, as I haven't been 
>>>> able to get them to move to the new network (which I was trying to use as 
>>>> all gluster).  I tried bonding, but that seamed to reduce performance 
>>>> rather than improve it.
>>>> 
>>>> --Jim
>>>> 
>>>> On Fri, Jul 6, 2018 at 2:52 PM, Jamie Lawrence <[email protected] 
>>>> <mailto:[email protected]>> wrote:
>>>> Hi Jim,
>>>> 
>>>> I don't have any targeted suggestions, because there isn't much to latch 
>>>> on to. I can say Gluster replica three  (no arbiters) on dedicated servers 
>>>> serving a couple Ovirt VM clusters here have not had these sorts of 
>>>> issues. 
>>>> 
>>>> I suspect your long heal times (and the resultant long periods of high 
>>>> load) are at least partly related to 1G networking. That is just a matter 
>>>> of IO - heals of VMs involve moving a lot of bits. My cluster uses 10G 
>>>> bonded NICs on the gluster and ovirt boxes for storage traffic and 
>>>> separate bonded 1G for ovirtmgmt and communication with other 
>>>> machines/people, and we're occasionally hitting the bandwidth ceiling on 
>>>> the storage network. I'm starting to think about 40/100G, different ways 
>>>> of splitting up intensive systems, and considering iSCSI for specific 
>>>> volumes, although I really don't want to go there.
>>>> 
>>>> I don't run FreeNAS[1], but I do run FreeBSD as storage servers for their 
>>>> excellent ZFS implementation, mostly for backups. ZFS will make your 
>>>> `heal` problem go away, but not your bandwidth problems, which become 
>>>> worse (because of fewer NICS pushing traffic). 10G hardware is not exactly 
>>>> in the impulse-buy territory, but if you can, I'd recommend doing some 
>>>> testing using it. I think at least some of your problems are related.
>>>> 
>>>> If that's not possible, my next stops would be optimizing everything I 
>>>> could about sharding, healing and optimizing for serving the shard size to 
>>>> squeeze as much performance out of 1G as I could, but that will only go so 
>>>> far.
>>>> 
>>>> -j
>>>> 
>>>> [1] FreeNAS is just a storage-tuned FreeBSD with a GUI.
>>>> 
>>>> > On Jul 6, 2018, at 1:19 PM, Jim Kusznir <[email protected] 
>>>> > <mailto:[email protected]>> wrote:
>>>> > 
>>>> > hi all:
>>>> > 
>>>> > Once again my production ovirt cluster is collapsing in on itself.  My 
>>>> > servers are intermittently unavailable or degrading, customers are 
>>>> > noticing and calling in.  This seems to be yet another gluster failure 
>>>> > that I haven't been able to pin down.
>>>> > 
>>>> > I posted about this a while ago, but didn't get anywhere (no replies 
>>>> > that I found).  The problem started out as a glusterfsd process 
>>>> > consuming large amounts of ram (up to the point where ram and swap were 
>>>> > exhausted and the kernel OOM killer killed off the glusterfsd process).  
>>>> > For reasons not clear to me at this time, that resulted in any VMs 
>>>> > running on that host and that gluster volume to be paused with I/O error 
>>>> > (the glusterfs process is usually unharmed; why it didn't continue I/O 
>>>> > with other servers is confusing to me).
>>>> > 
>>>> > I have 3 servers and a total of 4 gluster volumes (engine, iso, data, 
>>>> > and data-hdd).  The first 3 are replica 2+arb; the 4th (data-hdd) is 
>>>> > replica 3.  The first 3 are backed by an LVM partition (some thin 
>>>> > provisioned) on an SSD; the 4th is on a seagate hybrid disk (hdd + some 
>>>> > internal flash for acceleration).  data-hdd is the only thing on the 
>>>> > disk.  Servers are Dell R610 with the PERC/6i raid card, with the disks 
>>>> > individually passed through to the OS (no raid enabled).
>>>> > 
>>>> > The above RAM usage issue came from the data-hdd volume.  Yesterday, I 
>>>> > cought one of the glusterfsd high ram usage before the OOM-Killer had to 
>>>> > run.  I was able to migrate the VMs off the machine and for good 
>>>> > measure, reboot the entire machine (after taking this opportunity to run 
>>>> > the software updates that ovirt said were pending).  Upon booting back 
>>>> > up, the necessary volume healing began.  However, this time, the healing 
>>>> > caused all three servers to go to very, very high load averages (I saw 
>>>> > just under 200 on one server; typically they've been 40-70) with top 
>>>> > reporting IO Wait at 7-20%.  Network for this volume is a dedicated gig 
>>>> > network.  According to bwm-ng, initially the network bandwidth would hit 
>>>> > 50MB/s (yes, bytes), but tailed off to mostly in the kB/s for a while.  
>>>> > All machines' load averages were still 40+ and gluster volume heal 
>>>> > data-hdd info reported 5 items needing healing.  Server's were 
>>>> > intermittently experiencing IO issues, even on the 3 gluster volumes 
>>>> > that appeared largely unaffected.  Even the OS activities on the hosts 
>>>> > itself (logging in, running commands) would often be very delayed.  The 
>>>> > ovirt engine was seemingly randomly throwing engine down / engine up / 
>>>> > engine failed notifications.  Responsiveness on ANY VM was horrific most 
>>>> > of the time, with random VMs being inaccessible.
>>>> > 
>>>> > I let the gluster heal run overnight.  By morning, there were still 5 
>>>> > items needing healing, all three servers were still experiencing high 
>>>> > load, and servers were still largely unstable.
>>>> > 
>>>> > I've noticed that all of my ovirt outages (and I've had a lot, way more 
>>>> > than is acceptable for a production cluster) have come from gluster.  I 
>>>> > still have 3 VMs who's hard disk images have become corrupted by my last 
>>>> > gluster crash that I haven't had time to repair / rebuild yet (I believe 
>>>> > this crash was caused by the OOM issue previously mentioned, but I 
>>>> > didn't know it at the time).
>>>> > 
>>>> > Is gluster really ready for production yet?  It seems so unstable to 
>>>> > me....  I'm looking at replacing gluster with a dedicated NFS server 
>>>> > likely FreeNAS.  Any suggestions?  What is the "right" way to do 
>>>> > production storage on this (3 node cluster)?  Can I get this gluster 
>>>> > volume stable enough to get my VMs to run reliably again until I can 
>>>> > deploy another storage solution?
>>>> > 
>>>> > --Jim
>>>> > _______________________________________________
>>>> > Users mailing list -- [email protected] <mailto:[email protected]>
>>>> > To unsubscribe send an email to [email protected] 
>>>> > <mailto:[email protected]>
>>>> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
>>>> > <https://www.ovirt.org/site/privacy-policy/>
>>>> > oVirt Code of Conduct: 
>>>> > https://www.ovirt.org/community/about/community-guidelines/ 
>>>> > <https://www.ovirt.org/community/about/community-guidelines/>
>>>> > List Archives: 
>>>> > https://lists.ovirt.org/archives/list/[email protected]/message/YQX3LQFQQPW4JTCB7B6FY2LLR6NA2CB3/
>>>> >  
>>>> > <https://lists.ovirt.org/archives/list/[email protected]/message/YQX3LQFQQPW4JTCB7B6FY2LLR6NA2CB3/>
>>>> 
>>>> 
>>>> _______________________________________________
>>>> Users mailing list -- [email protected] <mailto:users%40ovirt.org>
>>>> To unsubscribe send an email to [email protected] 
>>>> <mailto:users-leave%40ovirt.org>
>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
>>>> <https://www.ovirt.org/site/privacy-policy/>
>>>> oVirt Code of Conduct: 
>>>> https://www.ovirt.org/community/about/community-guidelines/ 
>>>> <https://www.ovirt.org/community/about/community-guidelines/>
>>>> List Archives: 
>>>> https://lists.ovirt.org/archives/list/[email protected]/message/O2HIECLFMYGKH3KSZHHSMDUVGOEBI7GQ/
>>>>  
>>>> <https://lists.ovirt.org/archives/list/[email protected]/message/O2HIECLFMYGKH3KSZHHSMDUVGOEBI7GQ/>
>>> 
>>> 
>>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> _______________________________________________
>> Users mailing list -- [email protected] <mailto:[email protected]>
>> To unsubscribe send an email to [email protected] 
>> <mailto:[email protected]>
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
>> <https://www.ovirt.org/site/privacy-policy/>
>> oVirt Code of Conduct: 
>> https://www.ovirt.org/community/about/community-guidelines/ 
>> <https://www.ovirt.org/community/about/community-guidelines/>
>> List Archives: 
>> https://lists.ovirt.org/archives/list/[email protected]/message/T2M4J3Z7RPSGPEHNC33WFC2HUYOVL6FB/
>>  
>> <https://lists.ovirt.org/archives/list/[email protected]/message/T2M4J3Z7RPSGPEHNC33WFC2HUYOVL6FB/>
>> 
> 
> Edward Clay 
> Systems Administrator
> The Hut Group <http://www.thehutgroup.com/> 
> 
> Tel: 
> Email: [email protected] <mailto:[email protected]>
> 
> For the purposes of this email, the "company" means The Hut Group Limited, a 
> company registered in England and Wales (company number 6539496) whose 
> registered office is at Fifth Floor, Voyager House, Chicago Avenue, 
> Manchester Airport, M90 3DQ and/or any of its respective subsidiaries. 
> 
> Confidentiality Notice 
> This e-mail is confidential and intended for the use of the named recipient 
> only. If you are not the intended recipient please notify us by telephone 
> immediately on +44(0)1606 811888 or return it to us by e-mail. Please then 
> delete it from your system and note that any use, dissemination, forwarding, 
> printing or copying is strictly prohibited. Any views or opinions are solely 
> those of the author and do not necessarily represent those of the company. 
> 
> Encryptions and Viruses 
> Please note that this e-mail and any attachments have not been encrypted. 
> They may therefore be liable to be compromised. Please also note that it is 
> your responsibility to scan this e-mail and any attachments for viruses. We 
> do not, to the extent permitted by law, accept any liability (whether in 
> contract, negligence or otherwise) for any virus infection and/or external 
> compromise of security and/or confidentiality in relation to transmissions 
> sent by e-mail. 
> 
> Monitoring 
> Activity and use of the company's systems is monitored to secure its 
> effective use and operation and for other lawful business purposes. 
> Communications using these systems will also be monitored and may be recorded 
> to secure effective use and operation and for other lawful business purposes.
> 
> hgvyjuv
> 
> _______________________________________________
> Users mailing list -- [email protected] <mailto:[email protected]>
> To unsubscribe send an email to [email protected] 
> <mailto:[email protected]>
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ 
> <https://www.ovirt.org/site/privacy-policy/>
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/ 
> <https://www.ovirt.org/community/about/community-guidelines/>
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/Y2ZFGU2XDAXPMNLPQVHRDTNJQDFVWGCL/
>  
> <https://lists.ovirt.org/archives/list/[email protected]/message/Y2ZFGU2XDAXPMNLPQVHRDTNJQDFVWGCL/>
> 
> 
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/[email protected]/message/5YDMKKRWPEACWOOGVQPF2KGK6ZWUJVQY/

_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/4LCSECPBDN4HAUCGHVQGB3DPBZVURZ6K/

Reply via email to