Re: [Gluster-users] Memory recommendations Gluster setup with two bricks

2015-10-08 Thread Thibault Godouet
I am not aware of any guideline, and I can't pretend to be a Gluster
expert, but there are multiple processes and threads so I'd be tempted to
have more than 1 vCPUs: maybe 2 to 4?
As for the RAM, 1GB seems pretty small too.  I think that you will benefit
from caching if you have more memory. Maybe 4-8GB would be a better
starting point?

Obviously it depends on how much data you have, the performance you are
after, etc.
Dear list,

What would be the minimal recommended amount of memory for each Gluster
server in a Gluster set-up with two bricks, one datavolume and two
application servers mounting this datavolume? All servers are running
Ubuntu 14.04.
At the moment we're using VMWare VM's with 1 VCPU and 1GB of memory for
the bricks but in some cases this seems to be insufficient. If someone
could point me to some guidelines I'd be grateful. Thanks in advance!

Best,

Jeremy


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster-Nagios

2015-10-08 Thread Punit Dambiwal
Hi,

I am getting the following error :-


[root@monitor-001 yum.repos.d]#
/usr/lib64/nagios/plugins/gluster/discovery.py -c ssd -H stor1
Traceback (most recent call last):
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 510, in

clusterdata = discoverCluster(args.hostip, args.cluster, args.timeout)
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 88, in
discoverCluster
componentlist = discoverVolumes(hostip, timeout)
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 56, in
discoverVolumes
timeout=timeout)
  File "/usr/lib64/nagios/plugins/gluster/server_utils.py", line 107, in
execNRPECommand
resultDict = json.loads(outputStr)
  File "/usr/lib64/python2.6/json/__init__.py", line 307, in loads
return _default_decoder.decode(s)
  File "/usr/lib64/python2.6/json/decoder.py", line 319, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
  File "/usr/lib64/python2.6/json/decoder.py", line 336, in raw_decode
obj, end = self._scanner.iterscan(s, **kw).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 217, in JSONArray
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 155, in JSONString
return scanstring(match.string, match.end(), encoding, strict)
ValueError: ('Invalid control character at: line 1 column 1023 (char
1023)', '{"ssd": {"name": "ssd", "disperseCount": "0", "bricks":
[{"brickpath": "/bricks/b/vol1", "brickaddress": "stor1", "hostUuid":
"5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": "/bricks/b/vol1",
"brickaddress": "stor2", "hostUuid":
"b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": "/bricks/b/vol1",
"brickaddress": "stor3", "hostUuid":
"40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": "/bricks/b/vol1",
"brickaddress": "stor4", "hostUuid":
"5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": "/bricks/c/vol1",
"brickaddress": "stor1", "hostUuid":
"5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": "/bricks/c/vol1",
"brickaddress": "stor2", "hostUuid":
"b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": "/bricks/c/vol1",
"brickaddress": "stor3", "hostUuid":
"40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": "/bricks/c/vol1",
"brickaddress": "stor4", "hostUuid":
"5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": "/bricks/d/vol1",
"brickaddress": "stor1", "hostUuid": "5fcb5150-f0a5-4a\n')
[root@monitor-001 yum.repos.d]#
-

--
[root@monitor-001 yum.repos.d]# /usr/lib64/nagios/plugins/check_nrpe -H
stor1 -c discover_volume_list
{"ssd": {"type": "DISTRIBUTED_REPLICATE", "name": "ssd"}, "lockvol":
{"type": "REPLICATE", "name": "lockvol"}}
[root@monitor-001 yum.repos.d]#
--

Please help me to solve this issue...

Thanks,
Punit

On Fri, Oct 2, 2015 at 12:15 AM, Sahina Bose  wrote:

> The gluster-nagios packages have not been tested on Ubuntu
>
> Looking at the error below, it looks like the rpm has not updated the
> nrpe.cfg correctly. You may need to edit the spec file for the config file
> paths on Ubuntu and rebuild.
>
>
> On 10/01/2015 05:45 PM, Amudhan P wrote:
>
> OSError: [Errno 2] No such file or directory is now sorted out by by
> changing NRPE_PATH  in "constants.py".
>
> now if i run discovery.py
>
> testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$ sudo python
> discovery.py -c vm-gfs -H 192.168.1.11
> Failed to execute NRPE command 'discover_volume_list' in host
> '192.168.1.11'
> Error : NRPE: Command 'discover_volume_list' not defined
> Make sure NRPE server in host '192.168.1.11' is configured to accept
> requests from Nagios server
>
>
> testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$
> /usr/lib/nagios/plugins/check_nrpe -H 192.168.1.11 -c discover_volume_list
> NRPE: Command 'discover_volume_list' not defined
>
>
> My client is responding to other nrpe command.
> testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$
> /usr/lib/nagios/plugins/check_nrpe -H 192.168.1.11 -c check_load
> OK - 

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Humble Devassy Chirammal
The steps for replacing the brick is documented and available @
http://gluster.readthedocs.org/en/latest/Administrator%20Guide/Managing%20Volumes/.
Hope it helps.
On Oct 7, 2015 12:36 PM, "Lindsay Mathieson" 
wrote:

> First up - one of the things that concerns me re gluster is the incoherent
> state of documentation. The only docs linked on the main webpage are for
> 3.2 and there is almost nothing on how to handle failure modes such as dead
> disks/bricks etc, which is one of glusters primary functions.
>
> My problem - I have a replica 2 volume, 2 nodes, 2 bricks (zfs datasets).
>
> As a test, I destroyed one brick (zfs destroy the dataset).
>
>
> Can't start the datastore1:
>
>   volume start: datastore1: failed: Failed to find brick directory
> /glusterdata/datastore1 for volume datastore1. Reason : No such file or
> directory
>
> A bit disturbing, I was hoping it would work off the remaining brick.
>
> Can't replace the brick:
>
>   gluster volume replace-brick datastore1
> vnb.proxmox.softlog:/glusterdata/datastore1
> vnb.proxmox.softlog:/glusterdata/datastore1-2 commit force
>
> because the store is not running.
>
> After a lot of googling I found list messages referencing the remove brick
> command:
> gluster volume remove-brick datastore1 replica 2
> vnb.proxmox.softlog:/glusterdata/datastore1c commit force
>
> Fails with the unhelpful error:
>
> wrong brick type: commit, use :
> Usage: volume remove-brick  [replica ]  ...
> 
>
> In the end I destroyed and recreated the volume so I could resume testing,
> but I have no idea how I would handle a real failed brick in the future
>
> --
> Lindsay
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster-Nagios

2015-10-08 Thread Ramesh Nachimuthu
Looks like you are hitting the NRPE Payload issue. Standard NRPE 
packages from epel/fedora has 1024 bytes payload limit. We have to 
increment this to 8192 to fix this. You can see more info at 
http://serverfault.com/questions/613288/truncating-return-data-as-it-is-bigger-then-nrpe-allows. 



Let me know if u need any more info.

Regards,
Ramesh

On 10/08/2015 02:48 PM, Punit Dambiwal wrote:

Hi,

I am getting the following error :-


[root@monitor-001 yum.repos.d]# 
/usr/lib64/nagios/plugins/gluster/discovery.py -c ssd -H stor1

Traceback (most recent call last):
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 510, in 


clusterdata = discoverCluster(args.hostip, args.cluster, args.timeout)
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 88, in 
discoverCluster

componentlist = discoverVolumes(hostip, timeout)
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 56, in 
discoverVolumes

timeout=timeout)
  File "/usr/lib64/nagios/plugins/gluster/server_utils.py", line 107, 
in execNRPECommand

resultDict = json.loads(outputStr)
  File "/usr/lib64/python2.6/json/__init__.py", line 307, in loads
return _default_decoder.decode(s)
  File "/usr/lib64/python2.6/json/decoder.py", line 319, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
  File "/usr/lib64/python2.6/json/decoder.py", line 336, in raw_decode
obj, end = self._scanner.iterscan(s, **kw).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 217, in JSONArray
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 155, in JSONString
return scanstring(match.string, match.end(), encoding, strict)
ValueError: ('Invalid control character at: line 1 column 1023 (char 
1023)', '{"ssd": {"name": "ssd", "disperseCount": "0", "bricks": 
[{"brickpath": "/bricks/b/vol1", "brickaddress": "stor1", "hostUuid": 
"5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": 
"/bricks/b/vol1", "brickaddress": "stor2", "hostUuid": 
"b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": 
"/bricks/b/vol1", "brickaddress": "stor3", "hostUuid": 
"40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": 
"/bricks/b/vol1", "brickaddress": "stor4", "hostUuid": 
"5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": 
"/bricks/c/vol1", "brickaddress": "stor1", "hostUuid": 
"5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": 
"/bricks/c/vol1", "brickaddress": "stor2", "hostUuid": 
"b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": 
"/bricks/c/vol1", "brickaddress": "stor3", "hostUuid": 
"40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": 
"/bricks/c/vol1", "brickaddress": "stor4", "hostUuid": 
"5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": 
"/bricks/d/vol1", "brickaddress": "stor1", "hostUuid": 
"5fcb5150-f0a5-4a\n')

[root@monitor-001 yum.repos.d]#
-

--
[root@monitor-001 yum.repos.d]# /usr/lib64/nagios/plugins/check_nrpe 
-H stor1 -c discover_volume_list
{"ssd": {"type": "DISTRIBUTED_REPLICATE", "name": "ssd"}, "lockvol": 
{"type": "REPLICATE", "name": "lockvol"}}

[root@monitor-001 yum.repos.d]#
--

Please help me to solve this issue...

Thanks,
Punit

On Fri, Oct 2, 2015 at 12:15 AM, Sahina Bose > wrote:


The gluster-nagios packages have not been tested on Ubuntu

Looking at the error below, it looks like the rpm has not updated
the nrpe.cfg correctly. You may need to edit the spec file for the
config file paths on Ubuntu and rebuild.


On 10/01/2015 05:45 PM, Amudhan P wrote:

OSError: [Errno 2] No such file or directory is now sorted out by
by changing NRPE_PATH  in "constants.py".

now if i run discovery.py

testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$ sudo
python discovery.py -c vm-gfs -H 192.168.1.11
Failed to execute NRPE command 'discover_volume_list' in host
'192.168.1.11'
Error : NRPE: Command 'discover_volume_list' not 

[Gluster-users] .glusterfs folder has taken huge disk space

2015-10-08 Thread Sreedhara S N
Dear gluster Ants,

Recently we have configured two servers with GlusterFS and we started
syncing from one server to another. We found that the metadata cache folder
has taken huge space and it has many huge files with more than 5GB files.
This folder is eating most of the space of volume. Could you please let us
know, can we remove those files from metada folder if the sync is perfect.

Kindly let us provide the solution at the earliest.

Thanks in Advance,

Regards,
Sreedhara Shanbhog.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Writing to distributed (non-replicated) volume with failed nodes

2015-10-08 Thread Susant Palai
Hi,
  If the file creation hashes to the brick which is down, then it fails with 
ENOENT.

Susant

- Original Message -
From: "Leonid Isaev" 
To: gluster-users@gluster.org
Sent: Thursday, 8 October, 2015 7:54:07 AM
Subject: [Gluster-users] Writing to distributed (non-replicated) volume with
failed nodes

Hi,

I have an 8-node trusted pool with a distributed, non-replicated volume
The bricks are located only on 2 machines (2 bricks per node), so there are 6
dummy" nodes. Everything is working great until one of the brick-arrying nodes
experiences a power outage.
In this case, I can still mount the volume after a timeout (there is
plenty of servers to ask for metadata, after all) and read files from there,
but whenever I try to create a random-named file (e.g. running touch
/mnt/.lock-${RANDOM}${RANDOM}) this succeeds only sometimes, but often fails
with "no such file or directory". I understand that error if I were touching
files that already exist on the offline node (but invisible with the degraded
volume), but these are new random files which never existed before.
So, why does writing to the online bricks fail, and what can I do to
enable it? The machines run fully up-to-date Fedora 22 and ArchLinux with
gluster 3.7.4. I tried to look for similar problems on this ML, but haven't
found anything related, sorry if I missed something.

Thanks!
L.

-- 
Leonid Isaev
GPG fingerprints: DA92 034D B4A8 EC51 7EA6  20DF 9291 EE8A 043C B8C4
  C0DF 20D0 C075 C3F1 E1BE  775A A7AE F6CB 164B 5A6D
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Understanding management traffic and TLS

2015-10-08 Thread Игорь Бирюлин
Hello,
I've installed ssl/tls by this documentation:
https://access.redhat.com/documentation/en-US/Red_Hat_Storage/3.1/html/Administration_Guide/ch08s03.html

http://gluster.readthedocs.org/en/latest/Administrator%20Guide/SSL/




2015-10-07 19:23 GMT+03:00 John Weir :

> I am trying to understand the nature of the Gluster management traffic.
>
> This would be the traffic that can be secured with the presence of the
> ‘secure-access’ file.
>
> There appears to be a bug when ‘secure-access’ is enabled and I am
> evaluating the security implications of having TLS management traffic
> disabled.
>
> https://bugzilla.redhat.com/show_bug.cgi?id=1258931
>
> Can someone point me to the documentation on this traffic or provide me
> with details?
>
> Thank you very much - John
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Suggested method for replacing an entire node

2015-10-08 Thread Alastair Neil
I think you should back up /var/lib/glusterd and then restore it after the
reinstall and installation of glusterfs packages.  Assuming the node will
have the same hostname and ip addresses and you are installing the same
version gluster bits, I think it should be fine.  I am assuming you are not
using ssl for the connections if so you will need to back up the keys for
that too.

-Alastair

On 8 October 2015 at 00:12, Atin Mukherjee  wrote:

>
>
> On 10/07/2015 10:28 PM, Gene Liverman wrote:
> > I want to replace my existing CentOS 6 nodes with CentOS 7 ones. Is
> > there a recommended way to go about this from the perspective of
> > Gluster? I am running a 3 node replicated cluster (3 servers each with 1
> > brick). In case it makes a difference, my bricks are on separate drives
> > formatted as XFS so it is possible that I can do my OS reinstall without
> > wiping out the data on two nodes (the third had a hardware failure so it
> > will be fresh from the ground up).
> That's possible. You could do the re-installation one at a time. Once
> the node comes back online self heal daemon will take care of healing
> the data. AFR team can correct me if I am wrong.
>
> Thanks,
> Atin
> >
> >
> >
> >
> > Thanks,
> > *Gene Liverman*
> > Systems Integration Architect
> > Information Technology Services
> > University of West Georgia
> > glive...@westga.edu 
> >
> > ITS: Making Technology Work for You!
> >
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-users
> >
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster-Nagios

2015-10-08 Thread Amudhan P
can you pass any link or help to use this source tar file to work in
Ubuntu. editing spec file and setting path doesn't help.

On Thu, Oct 1, 2015 at 9:45 PM, Sahina Bose  wrote:

> The gluster-nagios packages have not been tested on Ubuntu
>
> Looking at the error below, it looks like the rpm has not updated the
> nrpe.cfg correctly. You may need to edit the spec file for the config file
> paths on Ubuntu and rebuild.
>
>
> On 10/01/2015 05:45 PM, Amudhan P wrote:
>
> OSError: [Errno 2] No such file or directory is now sorted out by by
> changing NRPE_PATH  in "constants.py".
>
> now if i run discovery.py
>
> testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$ sudo python
> discovery.py -c vm-gfs -H 192.168.1.11
> Failed to execute NRPE command 'discover_volume_list' in host
> '192.168.1.11'
> Error : NRPE: Command 'discover_volume_list' not defined
> Make sure NRPE server in host '192.168.1.11' is configured to accept
> requests from Nagios server
>
>
> testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$
> /usr/lib/nagios/plugins/check_nrpe -H 192.168.1.11 -c discover_volume_list
> NRPE: Command 'discover_volume_list' not defined
>
>
> My client is responding to other nrpe command.
> testusr@gfsovirt:/usr/local/lib/nagios/plugins/gluster$
> /usr/lib/nagios/plugins/check_nrpe -H 192.168.1.11 -c check_load
> OK - load average: 0.01, 0.03, 0.10|load1=0.010;15.000;30.000;0;
> load5=0.030;10.000;25.000;0; load15=0.100;5.000;20.000;0;
>
>
>
> On Thu, Oct 1, 2015 at 5:20 PM, Sahina Bose  wrote:
>
>> Looks like a conflict in versions of python and python-cpopen.
>> Can you give us the version of these packages?
>>
>> Also, what's the output of
>>  /usr/lib64/nagios/plugins/check_nrpe -H 192.168.1.11 -c
>> discover_volume_list
>>
>>
>>
>>
>> On 10/01/2015 04:10 PM, Amudhan P wrote:
>>
>> Hi,
>>
>> I am getting a error when i run discovery.py.
>>
>> discovery.py -c vm-gfs -H 192.168.1.11
>>
>> Traceback (most recent call last):
>>   File "discovery.py", line 541, in 
>> clusterdata = discoverCluster(args.hostip, args.cluster, args.timeout)
>>   File "discovery.py", line 90, in discoverCluster
>> componentlist = discoverVolumes(hostip, timeout)
>>   File "discovery.py", line 53, in discoverVolumes
>> timeout=timeout)
>>   File "/usr/local/lib/nagios/plugins/gluster/server_utils.py", line 114,
>> in execNRPECommand
>> (returncode, outputStr, err) = utils.execCmd(nrpeCmd, raw=True)
>>   File "/usr/lib/python2.7/dist-packages/glusternagios/utils.py", line
>> 403, in execCmd
>> deathSignal=deathSignal, childUmask=childUmask)
>>   File "/usr/local/lib/python2.7/dist-packages/cpopen/__init__.py", line
>> 63, in __init__
>> **kw)
>>   File "/usr/lib/python2.7/subprocess.py", line 710, in __init__
>> errread, errwrite)
>>   File "/usr/local/lib/python2.7/dist-packages/cpopen/__init__.py", line
>> 82, in _execute_child_v276
>> restore_sigpipe=restore_sigpipe
>>   File "/usr/local/lib/python2.7/dist-packages/cpopen/__init__.py", line
>> 107, in _execute_child_v275
>> restore_sigpipe
>> OSError: [Errno 2] No such file or directory
>>
>> Gluster version : 3.7.4
>> OS : Ubuntu 14.04
>> Complied from source tar file.
>>
>>
>> regards
>> Amudhan
>>
>>
>>
>>
>> On Wed, Sep 30, 2015 at 6:21 PM, Humble Devassy Chirammal <
>> humble.deva...@gmail.com> wrote:
>>
>>> The EL7 rpms of gluster-nagios are available @
>>> http://download.gluster.org/pub/gluster/glusterfs-nagios/1.1.0/
>>>
>>> Hope it helps!
>>>
>>> --Humble
>>>
>>>
>>> On Tue, Sep 29, 2015 at 10:56 AM, Sahina Bose < 
>>> sab...@redhat.com> wrote:
>>>
 We will publish the EL7 builds soon.

 The source tarballs are now available at -
 
 http://download.gluster.org/pub/gluster/glusterfs-nagios/

 thanks
 sahina


 On 09/25/2015 12:55 PM, Humble Devassy Chirammal wrote:

 HI Michael,

 Yes, only el6 packages are available @
 
 http://download.gluster.org/pub/gluster/glusterfs-nagios/ . I am
 looping nagios project team leads to this thread. Lets wait for them to
 revert.

 --Humble


 On Sun, Sep 20, 2015 at 2:32 PM, Prof. Dr. Michael Schefczyk <
 mich...@schefczyk.net> wrote:

> Dear All,
>
> In June 2014, the gluster-nagios team (thanks!) published the
> availability of gluster-nagios-common and gluster-nagios-addons on this
> list. As far as I can tell, this quite extensive gluster nagios monitoring
> tool is available for el6 only. Are there known plans to make this
> available for el7 outside the RHEL-repos (
> 
> http://ftp.redhat.de/pub/redhat/linux/enterprise/7Server/en/RHS/SRPMS/),
> e.g. for use with oVirt / Centos 7 

Re: [Gluster-users] Suggested method for replacing an entire node

2015-10-08 Thread Alastair Neil
Ahh that is good to know.

On 8 October 2015 at 09:50, Atin Mukherjee 
wrote:

> -Atin
> Sent from one plus one
> On Oct 8, 2015 7:17 PM, "Alastair Neil"  wrote:
> >
> > I think you should back up /var/lib/glusterd and then restore it after
> the reinstall and installation of glusterfs packages.  Assuming the node
> will have the same hostname and ip addresses and you are installing the
> same version gluster bits, I think it should be fine.  I am assuming you
> are not using ssl for the connections if so you will need to back up the
> keys for that too.
> If the same machine is used with out hostname/ IP change, backing up
> glusterd configuration *is not* needed as syncing the configuration will be
> taken care peer handshaking.
>
> >
> > -Alastair
> >
> > On 8 October 2015 at 00:12, Atin Mukherjee  wrote:
> >>
> >>
> >>
> >> On 10/07/2015 10:28 PM, Gene Liverman wrote:
> >> > I want to replace my existing CentOS 6 nodes with CentOS 7 ones. Is
> >> > there a recommended way to go about this from the perspective of
> >> > Gluster? I am running a 3 node replicated cluster (3 servers each
> with 1
> >> > brick). In case it makes a difference, my bricks are on separate
> drives
> >> > formatted as XFS so it is possible that I can do my OS reinstall
> without
> >> > wiping out the data on two nodes (the third had a hardware failure so
> it
> >> > will be fresh from the ground up).
> >> That's possible. You could do the re-installation one at a time. Once
> >> the node comes back online self heal daemon will take care of healing
> >> the data. AFR team can correct me if I am wrong.
> >>
> >> Thanks,
> >> Atin
> >> >
> >> >
> >> >
> >> >
> >> > Thanks,
> >> > *Gene Liverman*
> >> > Systems Integration Architect
> >> > Information Technology Services
> >> > University of West Georgia
> >> > glive...@westga.edu 
> >> >
> >> > ITS: Making Technology Work for You!
> >> >
> >> >
> >> > ___
> >> > Gluster-users mailing list
> >> > Gluster-users@gluster.org
> >> > http://www.gluster.org/mailman/listinfo/gluster-users
> >> >
> >> ___
> >> Gluster-users mailing list
> >> Gluster-users@gluster.org
> >> http://www.gluster.org/mailman/listinfo/gluster-users
> >
> >
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-users
>
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Suggested method for replacing an entire node

2015-10-08 Thread Atin Mukherjee
-Atin
Sent from one plus one
On Oct 8, 2015 7:17 PM, "Alastair Neil"  wrote:
>
> I think you should back up /var/lib/glusterd and then restore it after
the reinstall and installation of glusterfs packages.  Assuming the node
will have the same hostname and ip addresses and you are installing the
same version gluster bits, I think it should be fine.  I am assuming you
are not using ssl for the connections if so you will need to back up the
keys for that too.
If the same machine is used with out hostname/ IP change, backing up
glusterd configuration *is not* needed as syncing the configuration will be
taken care peer handshaking.
>
> -Alastair
>
> On 8 October 2015 at 00:12, Atin Mukherjee  wrote:
>>
>>
>>
>> On 10/07/2015 10:28 PM, Gene Liverman wrote:
>> > I want to replace my existing CentOS 6 nodes with CentOS 7 ones. Is
>> > there a recommended way to go about this from the perspective of
>> > Gluster? I am running a 3 node replicated cluster (3 servers each with
1
>> > brick). In case it makes a difference, my bricks are on separate drives
>> > formatted as XFS so it is possible that I can do my OS reinstall
without
>> > wiping out the data on two nodes (the third had a hardware failure so
it
>> > will be fresh from the ground up).
>> That's possible. You could do the re-installation one at a time. Once
>> the node comes back online self heal daemon will take care of healing
>> the data. AFR team can correct me if I am wrong.
>>
>> Thanks,
>> Atin
>> >
>> >
>> >
>> >
>> > Thanks,
>> > *Gene Liverman*
>> > Systems Integration Architect
>> > Information Technology Services
>> > University of West Georgia
>> > glive...@westga.edu 
>> >
>> > ITS: Making Technology Work for You!
>> >
>> >
>> > ___
>> > Gluster-users mailing list
>> > Gluster-users@gluster.org
>> > http://www.gluster.org/mailman/listinfo/gluster-users
>> >
>> ___
>> Gluster-users mailing list
>> Gluster-users@gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] glusterd crashing

2015-10-08 Thread Vijay Bellur

On Wednesday 07 October 2015 09:20 PM, Atin Mukherjee wrote:

This looks like a glibc corruption to me. Which distribution platform
are you running Gluster on?



A crash in glibc would mostly be due to memory corruption caused by the 
application. Can we please open a tracking bug if not done yet?


Thanks,
Vijay

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Suggested method for replacing an entire node

2015-10-08 Thread Gene Liverman
 Thanks for all the replies! Just to make sure I have this right, the
following should work for *both* machines with and machines without a
currently populated brick if the name and IP stay the same:

   - reinstall os
   - reinstall gluster software
   - start gluster

Do I need to do any peer probing or anything else? Do I need to do any
brick removal / adding (I'm thinking no but want to make sure)?




Thanks,
*Gene Liverman*
Systems Integration Architect
Information Technology Services
University of West Georgia
glive...@westga.edu

ITS: Making Technology Work for You!



On Thu, Oct 8, 2015 at 9:52 AM, Alastair Neil  wrote:

> Ahh that is good to know.
>
> On 8 October 2015 at 09:50, Atin Mukherjee 
> wrote:
>
>> -Atin
>> Sent from one plus one
>> On Oct 8, 2015 7:17 PM, "Alastair Neil"  wrote:
>> >
>> > I think you should back up /var/lib/glusterd and then restore it after
>> the reinstall and installation of glusterfs packages.  Assuming the node
>> will have the same hostname and ip addresses and you are installing the
>> same version gluster bits, I think it should be fine.  I am assuming you
>> are not using ssl for the connections if so you will need to back up the
>> keys for that too.
>> If the same machine is used with out hostname/ IP change, backing up
>> glusterd configuration *is not* needed as syncing the configuration will be
>> taken care peer handshaking.
>>
>> >
>> > -Alastair
>> >
>> > On 8 October 2015 at 00:12, Atin Mukherjee  wrote:
>> >>
>> >>
>> >>
>> >> On 10/07/2015 10:28 PM, Gene Liverman wrote:
>> >> > I want to replace my existing CentOS 6 nodes with CentOS 7 ones. Is
>> >> > there a recommended way to go about this from the perspective of
>> >> > Gluster? I am running a 3 node replicated cluster (3 servers each
>> with 1
>> >> > brick). In case it makes a difference, my bricks are on separate
>> drives
>> >> > formatted as XFS so it is possible that I can do my OS reinstall
>> without
>> >> > wiping out the data on two nodes (the third had a hardware failure
>> so it
>> >> > will be fresh from the ground up).
>> >> That's possible. You could do the re-installation one at a time. Once
>> >> the node comes back online self heal daemon will take care of healing
>> >> the data. AFR team can correct me if I am wrong.
>> >>
>> >> Thanks,
>> >> Atin
>> >> >
>> >> >
>> >> >
>> >> >
>> >> > Thanks,
>> >> > *Gene Liverman*
>> >> > Systems Integration Architect
>> >> > Information Technology Services
>> >> > University of West Georgia
>> >> > glive...@westga.edu 
>> >> >
>> >> > ITS: Making Technology Work for You!
>> >> >
>> >> >
>> >> > ___
>> >> > Gluster-users mailing list
>> >> > Gluster-users@gluster.org
>> >> > http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >
>> >> ___
>> >> Gluster-users mailing list
>> >> Gluster-users@gluster.org
>> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >
>> >
>> >
>> > ___
>> > Gluster-users mailing list
>> > Gluster-users@gluster.org
>> > http://www.gluster.org/mailman/listinfo/gluster-users
>>
>>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Please help in tuning Gluster FS in 2 node application system

2015-10-08 Thread M.Tarkeshwar Rao
Hi,

I have created Volume for Gluster FS in distributed replicated mode-

 gluster volume info gvol0

Volume Name: gvol0
Type: Replicate
Volume ID: 550c1759-f708-4199-8273-4aa2e6b1ce56
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glust2:/var/lib/gvol0/mediation
Brick2: glust3:/var/lib/gvol0/mediation
Options Reconfigured:
diagnostics.latency-measurement: off
performance.write-behind: off
nfs.rpc-auth-allow: 10.61.*
nfs.export-volumes: off
nfs.addr-namelookup: off
nfs.disable: on
auth.allow: 10.61.*
performance.readdir-ahead: on


But Sync time is too much between 2 nodes thats why my processing
performance is very bad.

My Second node is getting same data from first node very late thats why
second node is not able to process and remain in idle mode.

Please provide tuning/syncing parameters.

Please provide good values for these parameters for small and big size
files processing.

Please provide response ASAP. (need urgently)

Thank you

Regards
Tarkesh
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] glusterd crashing

2015-10-08 Thread Vijay Bellur

On Thursday 08 October 2015 11:01 PM, Gene Liverman wrote:

Happy to do so... what all info should go in the bug report?



Guidelines for logging a bug are available at [1]. Please try to provide 
relevant data requested in the Package Information, Cluster Information, 
Volume Information and Logs section of the guidelines page. Information 
pertaining to bricks, volume details & statedump can be skipped as it is 
the management daemon that is crashing here. Attaching the entire 
glusterd log would be helpful.


Regards,
Vijay

[1] 
http://www.gluster.org/community/documentation/index.php/Bug_reporting_guidelines


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Fwd: Gluster Instance Crash v3.7.4

2015-10-08 Thread Vijay Bellur

On Thursday 08 October 2015 10:41 PM, Ivan Rossi wrote:

some more data about the volume:

Volume Name: Volume-xxx
Type: Distributed-Replicate
Volume ID: b19cc9e2-071e-4f68-95e3-7c3e26d263a8
Status: Started
Number of Bricks: 2 x 3 = 6
Transport-type: tcp
Bricks:
Brick1: gluster1cr:/opt/gluster/bricks/MintVelvet-brick1/brick1
Brick2: gluster2cr:/opt/gluster/bricks/MintVelvet-brick1/brick1
Brick3: gluster3cr:/opt/gluster/bricks/MintVelvet-brick1/brick1
Brick4: gluster1cr:/opt/gluster/bricks/MintVelvet-brick2/brick2
Brick5: gluster2cr:/opt/gluster/bricks/MintVelvet-brick2/brick2
Brick6: gluster3cr:/opt/gluster/bricks/MintVelvet-brick2/brick2
Options Reconfigured:
features.barrier: disable
performance.readdir-ahead: on




Thanks for the additional details. Would it be possible to share the log 
file of the gluster process that crashed?


Regards,
Vijay

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] glusterd crashing

2015-10-08 Thread Gene Liverman
Happy to do so... what all info should go in the bug report?





--
*Gene Liverman*
Systems Integration Architect
Information Technology Services
University of West Georgia
glive...@westga.edu
678.839.5492

ITS: Making Technology Work for You!




On Thu, Oct 8, 2015 at 1:04 PM, Vijay Bellur  wrote:

> On Wednesday 07 October 2015 09:20 PM, Atin Mukherjee wrote:
>
>> This looks like a glibc corruption to me. Which distribution platform
>> are you running Gluster on?
>>
>>
> A crash in glibc would mostly be due to memory corruption caused by the
> application. Can we please open a tracking bug if not done yet?
>
> Thanks,
> Vijay
>
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Gene Liverman
So... this kinda applies to me too and I want to get some clarification: I
have the following setup

# gluster volume info

Volume Name: gv0
Type: Replicate
Volume ID: fc50d049-cebe-4a3f-82a6-748847226099
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: eapps-gluster01:/export/sdb1/gv0
Brick2: eapps-gluster02:/export/sdb1/gv0
Brick3: eapps-gluster03:/export/sdb1/gv0
Options Reconfigured:
diagnostics.count-fop-hits: on
diagnostics.latency-measurement: on
nfs.drc: off

eapps-gluster03 had a hard drive failure so I replaced it, formatted the
drive and now need gluster to be happy again. Gluster put a .glusterfs
folder in /export/sdb1/gv0 but nothing else has shown up and the brick is
offline. I read the docs on replacing a brick but seem to be missing
something and would appreciate some help. Thanks!





--
*Gene Liverman*
Systems Integration Architect
Information Technology Services
University of West Georgia
glive...@westga.edu

ITS: Making Technology Work for You!



On Thu, Oct 8, 2015 at 2:46 PM, Pranith Kumar Karampuri  wrote:

> On 3.7.4, all you need to do is execute "gluster volume replace-brick
>  commit force" and rest will be taken care by afr. We are in the
> process of coming up with new commands like "gluster volume reset-brick
>  start/commit" for wiping/re-formatting of the disk. So wait just
> a little longer :-).
>
> Pranith
>
>
> On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:
>
>
> On 8 October 2015 at 07:19, Joe Julian  wrote:
>
>> I documented this on my blog at
>> https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/ which is
>> still accurate for the latest version.
>>
>> The bug report I filed for this was closed without resolution. I assume
>> there's no plans for ever making this easy for administrators.
>> https://bugzilla.redhat.com/show_bug.cgi?id=991084
>>
>
>
> Yes, its the sort of workaround one can never remember in an emergency,
> you'd have to google it up ...
>
> In the case I was working with, probably easier and quicker to do a
> remove-brick/add-brick.
>
> thanks,
>
>
> --
> Lindsay
>
>
> ___
> Gluster-users mailing 
> listGluster-users@gluster.orghttp://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Writing to distributed (non-replicated) volume with failed nodes

2015-10-08 Thread Leonid Isaev
Hi,

On Thu, Oct 08, 2015 at 08:15:42AM -0400, Susant Palai wrote:
> If the file creation hashes to the brick which is down, then it fails with
> ENOENT.

OK, thanks for the explanation. But is there any way (some dht configuration,
perhaps?) to force this hashing to include only online bricks?

Otherwise the situation is rather odd: no matter how big a distributed volume
is, it's only as robust as the weakest node, because some file operations will
always fail...

Thanks,
L.

> - Original Message -
> From: "Leonid Isaev" 
> To: gluster-users@gluster.org
> Sent: Thursday, 8 October, 2015 7:54:07 AM
> Subject: [Gluster-users] Writing to distributed (non-replicated) volume with  
> failed nodes
> 
> Hi,
> 
>   I have an 8-node trusted pool with a distributed, non-replicated volume
> The bricks are located only on 2 machines (2 bricks per node), so there are 6
> dummy" nodes. Everything is working great until one of the brick-arrying nodes
> experiences a power outage.
>   In this case, I can still mount the volume after a timeout (there is
> plenty of servers to ask for metadata, after all) and read files from there,
> but whenever I try to create a random-named file (e.g. running touch
> /mnt/.lock-${RANDOM}${RANDOM}) this succeeds only sometimes, but often fails
> with "no such file or directory". I understand that error if I were touching
> files that already exist on the offline node (but invisible with the degraded
> volume), but these are new random files which never existed before.
>   So, why does writing to the online bricks fail, and what can I do to
> enable it? The machines run fully up-to-date Fedora 22 and ArchLinux with
> gluster 3.7.4. I tried to look for similar problems on this ML, but haven't
> found anything related, sorry if I missed something.
> 
> Thanks!
> L.
> 
> -- 
> Leonid Isaev
> GPG fingerprints: DA92 034D B4A8 EC51 7EA6  20DF 9291 EE8A 043C B8C4
>   C0DF 20D0 C075 C3F1 E1BE  775A A7AE F6CB 164B 5A6D
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users

-- 
Leonid Isaev
GPG fingerprints: DA92 034D B4A8 EC51 7EA6  20DF 9291 EE8A 043C B8C4
  C0DF 20D0 C075 C3F1 E1BE  775A A7AE F6CB 164B 5A6D
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Lindsay Mathieson
Very nice!

Any chance of a wheezy repo? ... 

Sent from Mail for Windows 10



From: Pranith Kumar Karampuri
Sent: Friday, 9 October 2015 4:46 AM
To: Lindsay Mathieson;Joe Julian
Cc: gluster-users
Subject: Re: [Gluster-users] How to replace a dead brick? (3.6.5)


On 3.7.4, all you need to do is execute "gluster volume replace-brick  
commit force" and rest will be taken care by afr. We are in the process of 
coming up with new commands like "gluster volume reset-brick  
start/commit" for wiping/re-formatting of the disk. So wait just a little 
longer :-).

Pranith
On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:

On 8 October 2015 at 07:19, Joe Julian  wrote:
I documented this on my blog at 
https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/ which is still 
accurate for the latest version.

The bug report I filed for this was closed without resolution. I assume there's 
no plans for ever making this easy for administrators.
https://bugzilla.redhat.com/show_bug.cgi?id=991084

Yes, its the sort of workaround one can never remember in an emergency, you'd 
have to google it up ...
In the case I was working with, probably easier and quicker to do a 
remove-brick/add-brick.
thanks,


-- 
Lindsay



___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users



___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Lindsay Mathieson
Did you try erasing the brick first?

Sent from Mail for Windows 10



From: Joe Julian
Sent: Friday, 9 October 2015 6:02 AM
To: Pranith Kumar Karampuri;Lindsay Mathieson
Cc: gluster-users
Subject: Re: [Gluster-users] How to replace a dead brick? (3.6.5)



On 10/08/2015 11:46 AM, Pranith Kumar Karampuri wrote:
On 3.7.4, all you need to do is execute "gluster volume replace-brick  
commit force" and rest will be taken care by afr. We are in the process of 
coming up with new commands like "gluster volume reset-brick  
start/commit" for wiping/re-formatting of the disk. So wait just a little 
longer :-).

Pranith

Nope. 
Volume Name: test
Type: Replicate
Volume ID: 426a1719-7cc2-4dac-97b4-67491679e00e
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: questor:/tmp/foo1.1
Brick2: questor:/tmp/foo1.2


Status of volume: test
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick questor:/tmp/foo1.1   49162 0  Y   20825
Brick questor:/tmp/foo1.2   49163 0  Y   20859
NFS Server on localhost N/A   N/A    N   N/A  
Self-heal Daemon on localhost   N/A   N/A    Y   20887


[root@questor]# kill 20825
[root@questor]# rm -rf /tmp/foo1.1
[root@questor]# mkdir /tmp/foo1.1
[root@questor]# gluster volume replace-brick test commit force
Usage: volume replace-brick{commit force}
[root@questor]# gluster volume replace-brick test questor:/tmp/foo1.1 
questor:/tmp/foo1.1 commit force
volume replace-brick: failed: Brick: questor:/tmp/foo1.1 not available. Brick 
may be containing or be contained by an existing brick



On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:

On 8 October 2015 at 07:19, Joe Julian  wrote:
I documented this on my blog at 
https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/ which is still 
accurate for the latest version.

The bug report I filed for this was closed without resolution. I assume there's 
no plans for ever making this easy for administrators.
https://bugzilla.redhat.com/show_bug.cgi?id=991084

Yes, its the sort of workaround one can never remember in an emergency, you'd 
have to google it up ...
In the case I was working with, probably easier and quicker to do a 
remove-brick/add-brick.
thanks,


-- 
Lindsay



___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users




___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Lindsay Mathieson
On 9 October 2015 at 05:20, Gene Liverman  wrote:

> eapps-gluster03 had a hard drive failure so I replaced it, formatted the
> drive and now need gluster to be happy again. Gluster put a .glusterfs
> folder in /export/sdb1/gv0 but nothing else has shown up and the brick is
> offline. I read the docs on replacing a brick but seem to be missing
> something and would appreciate some help. Thanks!
>

In my testing here, remove-brick/add-brick did the trick.

volume remove-brick gv0 replica 2 eapps-gluster03:/export/sdb1/gv0 force
volume add-brick gv0 replica 3 eapps-gluster03:/export/sdb1/gv0 [force]



-- 
Lindsay
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Please help in tuning Gluster FS in 2 nodeapplication system

2015-10-08 Thread Lindsay Mathieson
More details needed.
- Hardware on each Node, CPU, RAM, Disks
- Op. Sys. & Version
- Details of Disk setup, Filing System type, Raid Levels, Cache
- Network Setuip, speed, Bonding 
- What benchmarks have you run? What read/write and IOPS are yo seeing?



Sent from Mail for Windows 10



From: M.Tarkeshwar Rao
Sent: Friday, 9 October 2015 3:43 AM
To: gluster-users@gluster.org
Cc: Priyanka Agrawal
Subject: [Gluster-users] Please help in tuning Gluster FS in 2 nodeapplication 
system


Hi,

I have created Volume for Gluster FS in distributed replicated mode-

 gluster volume info gvol0

Volume Name: gvol0
Type: Replicate
Volume ID: 550c1759-f708-4199-8273-4aa2e6b1ce56
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: glust2:/var/lib/gvol0/mediation
Brick2: glust3:/var/lib/gvol0/mediation
Options Reconfigured:
diagnostics.latency-measurement: off
performance.write-behind: off
nfs.rpc-auth-allow: 10.61.*
nfs.export-volumes: off
nfs.addr-namelookup: off
nfs.disable: on
auth.allow: 10.61.*
performance.readdir-ahead: on


But Sync time is too much between 2 nodes thats why my processing performance 
is very bad.

My Second node is getting same data from first node very late thats why second 
node is not able to process and remain in idle mode.

Please provide tuning/syncing parameters.

Please provide good values for these parameters for small and big size files 
processing.

Please provide response ASAP. (need urgently)

Thank you

Regards
Tarkesh


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster-Nagios

2015-10-08 Thread Punit Dambiwal
Hi Ramesh,

Even after recompile nrpe with increased value still the same issue...

Thanks,
Punit

On Fri, Oct 9, 2015 at 9:21 AM, Punit Dambiwal  wrote:

> Hi Ramesh,
>
> Thanks for the update...as i have install nagios and nrpe via yum,should i
> need to remove nrpe and reinstall through source package ??
>
> Thanks,
> Punit
>
> On Thu, Oct 8, 2015 at 6:49 PM, Ramesh Nachimuthu 
> wrote:
>
>> Looks like you are hitting the NRPE Payload issue. Standard NRPE packages
>> from epel/fedora has 1024 bytes payload limit. We have to increment this to
>> 8192 to fix this. You can see more info at
>> 
>> http://serverfault.com/questions/613288/truncating-return-data-as-it-is-bigger-then-nrpe-allows.
>>
>>
>> Let me know if u need any more info.
>>
>> Regards,
>> Ramesh
>>
>>
>> On 10/08/2015 02:48 PM, Punit Dambiwal wrote:
>>
>> Hi,
>>
>> I am getting the following error :-
>>
>> 
>> [root@monitor-001 yum.repos.d]#
>> /usr/lib64/nagios/plugins/gluster/discovery.py -c ssd -H stor1
>> Traceback (most recent call last):
>>   File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 510, in
>> 
>> clusterdata = discoverCluster(args.hostip, args.cluster, args.timeout)
>>   File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 88, in
>> discoverCluster
>> componentlist = discoverVolumes(hostip, timeout)
>>   File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 56, in
>> discoverVolumes
>> timeout=timeout)
>>   File "/usr/lib64/nagios/plugins/gluster/server_utils.py", line 107, in
>> execNRPECommand
>> resultDict = json.loads(outputStr)
>>   File "/usr/lib64/python2.6/json/__init__.py", line 307, in loads
>> return _default_decoder.decode(s)
>>   File "/usr/lib64/python2.6/json/decoder.py", line 319, in decode
>> obj, end = self.raw_decode(s, idx=_w(s, 0).end())
>>   File "/usr/lib64/python2.6/json/decoder.py", line 336, in raw_decode
>> obj, end = self._scanner.iterscan(s, **kw).next()
>>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
>> rval, next_pos = action(m, context)
>>   File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
>> value, end = iterscan(s, idx=end, context=context).next()
>>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
>> rval, next_pos = action(m, context)
>>   File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
>> value, end = iterscan(s, idx=end, context=context).next()
>>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
>> rval, next_pos = action(m, context)
>>   File "/usr/lib64/python2.6/json/decoder.py", line 217, in JSONArray
>> value, end = iterscan(s, idx=end, context=context).next()
>>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
>> rval, next_pos = action(m, context)
>>   File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
>> value, end = iterscan(s, idx=end, context=context).next()
>>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
>> rval, next_pos = action(m, context)
>>   File "/usr/lib64/python2.6/json/decoder.py", line 155, in JSONString
>> return scanstring(match.string, match.end(), encoding, strict)
>> ValueError: ('Invalid control character at: line 1 column 1023 (char
>> 1023)', '{"ssd": {"name": "ssd", "disperseCount": "0", "bricks":
>> [{"brickpath": "/bricks/b/vol1", "brickaddress": "stor1", "hostUuid":
>> "5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": "/bricks/b/vol1",
>> "brickaddress": "stor2", "hostUuid":
>> "b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": "/bricks/b/vol1",
>> "brickaddress": "stor3", "hostUuid":
>> "40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": "/bricks/b/vol1",
>> "brickaddress": "stor4", "hostUuid":
>> "5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": "/bricks/c/vol1",
>> "brickaddress": "stor1", "hostUuid":
>> "5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": "/bricks/c/vol1",
>> "brickaddress": "stor2", "hostUuid":
>> "b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": "/bricks/c/vol1",
>> "brickaddress": "stor3", "hostUuid":
>> "40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": "/bricks/c/vol1",
>> "brickaddress": "stor4", "hostUuid":
>> "5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": "/bricks/d/vol1",
>> "brickaddress": "stor1", "hostUuid": "5fcb5150-f0a5-4a\n')
>> [root@monitor-001 yum.repos.d]#
>> -
>>
>> --
>> [root@monitor-001 yum.repos.d]# /usr/lib64/nagios/plugins/check_nrpe -H
>> stor1 -c discover_volume_list
>> {"ssd": {"type": "DISTRIBUTED_REPLICATE", "name": "ssd"}, "lockvol":
>> {"type": "REPLICATE", "name": "lockvol"}}
>> [root@monitor-001 yum.repos.d]#
>> --
>>
>> Please help me to solve this issue...
>>
>> Thanks,
>> Punit
>>
>> On Fri, Oct 2, 2015 at 12:15 

[Gluster-users] 3.6.6 Release Notes

2015-10-08 Thread Lindsay Mathieson
Are they anywhere? couldn't find them.

thanks,

-- 
Lindsay
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster-Nagios

2015-10-08 Thread Punit Dambiwal
Hi Ramesh,

Thanks for the update...as i have install nagios and nrpe via yum,should i
need to remove nrpe and reinstall through source package ??

Thanks,
Punit

On Thu, Oct 8, 2015 at 6:49 PM, Ramesh Nachimuthu 
wrote:

> Looks like you are hitting the NRPE Payload issue. Standard NRPE packages
> from epel/fedora has 1024 bytes payload limit. We have to increment this to
> 8192 to fix this. You can see more info at
> 
> http://serverfault.com/questions/613288/truncating-return-data-as-it-is-bigger-then-nrpe-allows.
>
>
> Let me know if u need any more info.
>
> Regards,
> Ramesh
>
>
> On 10/08/2015 02:48 PM, Punit Dambiwal wrote:
>
> Hi,
>
> I am getting the following error :-
>
> 
> [root@monitor-001 yum.repos.d]#
> /usr/lib64/nagios/plugins/gluster/discovery.py -c ssd -H stor1
> Traceback (most recent call last):
>   File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 510, in
> 
> clusterdata = discoverCluster(args.hostip, args.cluster, args.timeout)
>   File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 88, in
> discoverCluster
> componentlist = discoverVolumes(hostip, timeout)
>   File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 56, in
> discoverVolumes
> timeout=timeout)
>   File "/usr/lib64/nagios/plugins/gluster/server_utils.py", line 107, in
> execNRPECommand
> resultDict = json.loads(outputStr)
>   File "/usr/lib64/python2.6/json/__init__.py", line 307, in loads
> return _default_decoder.decode(s)
>   File "/usr/lib64/python2.6/json/decoder.py", line 319, in decode
> obj, end = self.raw_decode(s, idx=_w(s, 0).end())
>   File "/usr/lib64/python2.6/json/decoder.py", line 336, in raw_decode
> obj, end = self._scanner.iterscan(s, **kw).next()
>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
>   File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
> value, end = iterscan(s, idx=end, context=context).next()
>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
>   File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
> value, end = iterscan(s, idx=end, context=context).next()
>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
>   File "/usr/lib64/python2.6/json/decoder.py", line 217, in JSONArray
> value, end = iterscan(s, idx=end, context=context).next()
>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
>   File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
> value, end = iterscan(s, idx=end, context=context).next()
>   File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
>   File "/usr/lib64/python2.6/json/decoder.py", line 155, in JSONString
> return scanstring(match.string, match.end(), encoding, strict)
> ValueError: ('Invalid control character at: line 1 column 1023 (char
> 1023)', '{"ssd": {"name": "ssd", "disperseCount": "0", "bricks":
> [{"brickpath": "/bricks/b/vol1", "brickaddress": "stor1", "hostUuid":
> "5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": "/bricks/b/vol1",
> "brickaddress": "stor2", "hostUuid":
> "b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": "/bricks/b/vol1",
> "brickaddress": "stor3", "hostUuid":
> "40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": "/bricks/b/vol1",
> "brickaddress": "stor4", "hostUuid":
> "5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": "/bricks/c/vol1",
> "brickaddress": "stor1", "hostUuid":
> "5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath": "/bricks/c/vol1",
> "brickaddress": "stor2", "hostUuid":
> "b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath": "/bricks/c/vol1",
> "brickaddress": "stor3", "hostUuid":
> "40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath": "/bricks/c/vol1",
> "brickaddress": "stor4", "hostUuid":
> "5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath": "/bricks/d/vol1",
> "brickaddress": "stor1", "hostUuid": "5fcb5150-f0a5-4a\n')
> [root@monitor-001 yum.repos.d]#
> -
>
> --
> [root@monitor-001 yum.repos.d]# /usr/lib64/nagios/plugins/check_nrpe -H
> stor1 -c discover_volume_list
> {"ssd": {"type": "DISTRIBUTED_REPLICATE", "name": "ssd"}, "lockvol":
> {"type": "REPLICATE", "name": "lockvol"}}
> [root@monitor-001 yum.repos.d]#
> --
>
> Please help me to solve this issue...
>
> Thanks,
> Punit
>
> On Fri, Oct 2, 2015 at 12:15 AM, Sahina Bose  wrote:
>
>> The gluster-nagios packages have not been tested on Ubuntu
>>
>> Looking at the error below, it looks like the rpm has not updated the
>> nrpe.cfg correctly. You may need to edit the spec file for the config file
>> paths on Ubuntu and 

Re: [Gluster-users] Suggested method for replacing an entire node

2015-10-08 Thread Atin Mukherjee


On 10/08/2015 09:52 PM, Gene Liverman wrote:
>  Thanks for all the replies! Just to make sure I have this right, the
> following should work for *both* machines with and machines without a
> currently populated brick if the name and IP stay the same:
> 
>   * reinstall os
>   * reinstall gluster software
>   * start gluster
> 
> Do I need to do any peer probing or anything else? Do I need to do any
> brick removal / adding (I'm thinking no but want to make sure)?
No you don't.
> 
> 
> 
> 
> Thanks,
> *Gene Liverman*
> Systems Integration Architect
> Information Technology Services
> University of West Georgia
> glive...@westga.edu 
> 
> ITS: Making Technology Work for You!
> 
> 
> 
> On Thu, Oct 8, 2015 at 9:52 AM, Alastair Neil  > wrote:
> 
> Ahh that is good to know.
> 
> On 8 October 2015 at 09:50, Atin Mukherjee
> > wrote:
> 
> -Atin
> Sent from one plus one
> On Oct 8, 2015 7:17 PM, "Alastair Neil"  > wrote:
> >
> > I think you should back up /var/lib/glusterd and then restore it 
> after the reinstall and installation of glusterfs packages.  Assuming the 
> node will have the same hostname and ip addresses and you are installing the 
> same version gluster bits, I think it should be fine.  I am assuming you are 
> not using ssl for the connections if so you will need to back up the keys for 
> that too.
> If the same machine is used with out hostname/ IP change,
> backing up glusterd configuration *is not* needed as syncing the
> configuration will be taken care peer handshaking.
> 
> 
> >
> > -Alastair
> >
> > On 8 October 2015 at 00:12, Atin Mukherjee
> > wrote:
> >>
> >>
> >>
> >> On 10/07/2015 10:28 PM, Gene Liverman wrote:
> >> > I want to replace my existing CentOS 6 nodes with CentOS 7
> ones. Is
> >> > there a recommended way to go about this from the
> perspective of
> >> > Gluster? I am running a 3 node replicated cluster (3
> servers each with 1
> >> > brick). In case it makes a difference, my bricks are on
> separate drives
> >> > formatted as XFS so it is possible that I can do my OS
> reinstall without
> >> > wiping out the data on two nodes (the third had a hardware
> failure so it
> >> > will be fresh from the ground up).
> >> That's possible. You could do the re-installation one at a
> time. Once
> >> the node comes back online self heal daemon will take care of
> healing
> >> the data. AFR team can correct me if I am wrong.
> >>
> >> Thanks,
> >> Atin
> >> >
> >> >
> >> >
> >> >
> >> > Thanks,
> >> > *Gene Liverman*
> >> > Systems Integration Architect
> >> > Information Technology Services
> >> > University of West Georgia
> >> > glive...@westga.edu 
> >
> >> >
> >> > ITS: Making Technology Work for You!
> >> >
> >> >
> >> > ___
> >> > Gluster-users mailing list
> >> > Gluster-users@gluster.org 
> >> > http://www.gluster.org/mailman/listinfo/gluster-users
> >> >
> >> ___
> >> Gluster-users mailing list
> >> Gluster-users@gluster.org 
> >> http://www.gluster.org/mailman/listinfo/gluster-users
> >
> >
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org 
> > http://www.gluster.org/mailman/listinfo/gluster-users
> 
> 
> 
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org 
> http://www.gluster.org/mailman/listinfo/gluster-users
> 
> 
> 
> 
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
> 
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster-Nagios

2015-10-08 Thread Sahina Bose

You can update the packages with the ones built from source.
You will need to update both the client and server nrpe packages with 
the modified payload limit to resolve this

- nagios-plugins-nrpe
- nrpe

Have you done that?

On 10/09/2015 07:17 AM, Punit Dambiwal wrote:

Hi Ramesh,

Even after recompile nrpe with increased value still the same issue...

Thanks,
Punit

On Fri, Oct 9, 2015 at 9:21 AM, Punit Dambiwal > wrote:


Hi Ramesh,

Thanks for the update...as i have install nagios and nrpe via
yum,should i need to remove nrpe and reinstall through source
package ??

Thanks,
Punit

On Thu, Oct 8, 2015 at 6:49 PM, Ramesh Nachimuthu
> wrote:

Looks like you are hitting the NRPE Payload issue. Standard
NRPE packages from epel/fedora has 1024 bytes payload limit.
We have to increment this to 8192 to fix this. You can see
more info at

http://serverfault.com/questions/613288/truncating-return-data-as-it-is-bigger-then-nrpe-allows.


Let me know if u need any more info.

Regards,
Ramesh


On 10/08/2015 02:48 PM, Punit Dambiwal wrote:

Hi,

I am getting the following error :-


[root@monitor-001 yum.repos.d]#
/usr/lib64/nagios/plugins/gluster/discovery.py -c ssd -H stor1
Traceback (most recent call last):
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line
510, in 
clusterdata = discoverCluster(args.hostip, args.cluster,
args.timeout)
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line
88, in discoverCluster
componentlist = discoverVolumes(hostip, timeout)
  File "/usr/lib64/nagios/plugins/gluster/discovery.py", line
56, in discoverVolumes
timeout=timeout)
  File "/usr/lib64/nagios/plugins/gluster/server_utils.py",
line 107, in execNRPECommand
resultDict = json.loads(outputStr)
  File "/usr/lib64/python2.6/json/__init__.py", line 307, in
loads
return _default_decoder.decode(s)
  File "/usr/lib64/python2.6/json/decoder.py", line 319, in
decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
  File "/usr/lib64/python2.6/json/decoder.py", line 336, in
raw_decode
obj, end = self._scanner.iterscan(s, **kw).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in
iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in
JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in
iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in
JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in
iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 217, in
JSONArray
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in
iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 183, in
JSONObject
value, end = iterscan(s, idx=end, context=context).next()
  File "/usr/lib64/python2.6/json/scanner.py", line 55, in
iterscan
rval, next_pos = action(m, context)
  File "/usr/lib64/python2.6/json/decoder.py", line 155, in
JSONString
return scanstring(match.string, match.end(), encoding,
strict)
ValueError: ('Invalid control character at: line 1 column
1023 (char 1023)', '{"ssd": {"name": "ssd", "disperseCount":
"0", "bricks": [{"brickpath": "/bricks/b/vol1",
"brickaddress": "stor1", "hostUuid":
"5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath":
"/bricks/b/vol1", "brickaddress": "stor2", "hostUuid":
"b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath":
"/bricks/b/vol1", "brickaddress": "stor3", "hostUuid":
"40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath":
"/bricks/b/vol1", "brickaddress": "stor4", "hostUuid":
"5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath":
"/bricks/c/vol1", "brickaddress": "stor1", "hostUuid":
"5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath":
"/bricks/c/vol1", "brickaddress": "stor2", "hostUuid":
"b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath":
"/bricks/c/vol1", "brickaddress": "stor3", "hostUuid":

[Gluster-users] Test results and Performance Tuning efforts ...

2015-10-08 Thread Lindsay Mathieson
Morning, hope the folllowing ramble is ok, just examining the results of some 
extensive (and destructive ) testing of gluster 3.6.4 on some disks I had 
spare. Cluster purpose is solely for hosting qemu vm’s via Proxmox 3.4

Setup: 3 Nodes, well spec’d
-  64 GB RAM
-   VNB & VNG
* CPU : E5-2620
-   VNA
* CPU’s : Dual E5-2660
-  Already in use as a Proxmox and Ceph Cluster running 30 Windows VM’s

Gluster Bricks.
-  All bricks on ZFS with 4 GB RAM ZIL, 1GB SSD SLOG and 10GB SSD Cache
-  LZ4 Compression
-  Sync disabled

Brick 1:
-  6 Velocitoraptors in a RAID10+ (3 Mirrors)
-  High performance
-  Already hosting 8 VM’s

Bricks 2 & 3:
-  Spare external USB 1TB Toshiba Drive attached via USB3
-  Crap performance  About 50/100 MB/s R/W


Overall impressions – pretty good. Installation is easy and now I’ve been 
pointed to up to date docs and got the hang of the commands, I’m happy with the 
administration – vastly simpler than Ceph. The ability to access the files on 
the native filesystem is good for peace of mind and enables some interesting 
benchmark comparisons. I simulated drive failure by killing all the gluster 
processes on a node and it seemed to cope ok. 

I would like to see better status information such as “Heal % progress”, 
“Rebalance % progress”

NB: Pulling a USB external drive is a *bad* idea as it has no TLER support and 
this killed an entire node, had to hard reset it. In production I would use  
something like WD Red NAS drives.


Despite all the abuse I threw at it I had no problems with split brain etc and 
the integration with proxmox is excellent. When running write tests I was very 
pleased to see it max out my bonded 2x1GB connections, something ceph has never 
been able to do. I consistently got 110+ MB/s raw write results inside VM’s

Currently running 4 VM’s off the Gluster datastore with no issues.

Benchmark results – done using Crystal DiskMark inside a Windows 7 VM, with 
VIRTIO drivers and writeback enabled. I tested a Gluster replica 3 setup, 
replica 1 and direct off the disk (ZFS). Multpile tests were run to get a feel 
for average results.

Node VNB
- Replica 3
- Local Brick: External USB Toshiba Drive
- ---
- CrystalDiskMark 3.0.3 x64 (C) 2007-2013 hiyohiyo
-Crystal Dew World : http://crystalmark.info/
- ---
- * MB/s = 1,000,000 byte/s [SATA/300 = 300,000,000 byte/s]
- 
-Sequential Read :   738.642 MB/s
-   Sequential Write :   114.461 MB/s
-  Random Read 512KB :   720.623 MB/s
- Random Write 512KB :   115.084 MB/s
- Random Read 4KB (QD=1) : 9.684 MB/s [  2364.3 IOPS]
-Random Write 4KB (QD=1) : 2.511 MB/s [   613.0 IOPS]
-Random Read 4KB (QD=32) :24.264 MB/s [  5923.7 IOPS]
-   Random Write 4KB (QD=32) : 5.685 MB/s [  1387.8 IOPS]
- 
-   Test : 1000 MB [C: 70.1% (44.8/63.9 GB)] (x5)
-   Date : 2015/10/09 9:30:37
- OS : Windows 7 Professional N SP1 [6.1 Build 7601] (x64)


Node VNA
- Replica 1 (So no writing over ethernet)
- Local Brick: High performance Velocipraptors in RAID10
-Sequential Read :   735.224 MB/s
-   Sequential Write :   718.203 MB/s
-  Random Read 512KB :   888.090 MB/s
- Random Write 512KB :   453.174 MB/s
- Random Read 4KB (QD=1) :11.808 MB/s [  2882.9 IOPS]
-Random Write 4KB (QD=1) : 4.249 MB/s [  1037.4 IOPS]
-Random Read 4KB (QD=32) :34.787 MB/s [  8492.8 IOPS]
-   Random Write 4KB (QD=32) : 5.487 MB/s [  1339.5 IOPS]


Node VNA
- Direct on ZFS (No Gluster)
-Sequential Read :  2841.216 MB/s
-   Sequential Write :  1568.681 MB/s
-  Random Read 512KB :  1753.746 MB/s
- Random Write 512KB :  1219.437 MB/s
- Random Read 4KB (QD=1) :26.852 MB/s [  6555.6 IOPS]
-Random Write 4KB (QD=1) :20.930 MB/s [  5109.8 IOPS]
-Random Read 4KB (QD=32) :58.515 MB/s [ 14286.0 IOPS]
-   Random Write 4KB (QD=32) :46.303 MB/s [ 11304.3 IOPS]



Performance:
Raw read performance is excellent, averaging 700Mb/s – I’d say the ZFS & 
Cluster caches are working well.
As  mentioned raw write maxed out at 110 MB/s, near the max ethernet speed.
Random I/O is pretty average, it could be the Toshba drives bring things down, 
though even when I took them out of the equation it wasn’t much improved.

Direct off the disk was more than double the replica 1 brick in all areas, but 
I don’t find that surprising. I expected a fair amount of overhead with a 
cluster fs, and a 1-brick setup is not a real world usage. I was fairly 
impressed that adding two bricks to replica 3 made no real difference to the 
read results and the write results were obviously limuted by network speed. If 
only I could afford 10GB cards and a switch 

Re: [Gluster-users] [IMPORTANT, PLEASE READ] replace-brick problem with all releases till now

2015-10-08 Thread Pranith Kumar Karampuri



On 10/03/2015 05:31 AM, Steve Dainard wrote:

On Thu, Oct 1, 2015 at 2:24 AM, Pranith Kumar Karampuri
 wrote:

hi,
  In releases till now from day-1 with replication, there is a corner
case bug which can wipe out the all the bricks in that replica set  when the
disk/brick(s) are  replaced.

Here are the steps that could lead to that situation:
0) Clients are operating on the volume and are actively pumping data.
1) Execute replace-brick command (OR) take down the brick that needs the
disk replaced or re-formatted. and bring the brick back up.

So the better course of action would be to remove-brick  replica
 start, replace the disk, and then add-brick  replica 
? Perhaps it would be wise to un-peer the host before adding the brick
back?

Is there any chance that adding a 3rd replica to a 2 replica cluster
with active client writes could cause the same issue?
Yes there is that chance. On 3.7.3 replace-brick issue is already fixed. 
This mail applies mostly for version before it, i.e. from v3.1.x. till 
v3.7.2 We are also trying to fix the add-brick issue and providing new 
command called 'reset-brick' for reformatting of brick usecase. Let us 
see if we can deliver them by 3.7.6.


For older versions, these are the steps for performing replace-brick: 
http://gluster.readthedocs.org/en/latest/Administrator%20Guide/Managing%20Volumes/#replace-brick


Pranith


On 3.7.3 I recently lost 2 of 3 bricks all the way down to the XFS
filesystem being corrupted, but I blamed that on the disk controller
which was doing a raid0 pass-through on 2 hosts, but not the new 3rd
host. This occurred after some time though, and client writes were
being blocked while the 3rd brick was being added.


2) Client creates a file/directory just on the root of the brick which
succeeds on the new brick but fails on the bricks that have been online
(May be because the file existed on the bricks that are good copies)
3) Now when self-heal is triggered from the Client/self-heal-daemon it
thinks the just replaced brick is the correct directory and deletes the
files/directories from the bricks that have the actual data.

I have been working on afr for almost 4 years now and never saw any user
complain about this problem. We were working on a document for an  official
way to replace brick/disk but it never occurred to us that this  could
happen until recently. I am going to get a proper document by end of this
week on replacing the bricks/disks in a safe way. And will keep you  guys
posted about fixes to prevent this from happening entirely.

Pranith
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] 3.6.6 Release Notes

2015-10-08 Thread Pranith Kumar Karampuri

Adding 3.6.x release maintainer, Raghavendra Bhat.

Pranith
On 10/09/2015 06:13 AM, Lindsay Mathieson wrote:

Are they anywhere? couldn't find them.

thanks,

--
Lindsay


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Pranith Kumar Karampuri

Follow the steps at:
http://gluster.readthedocs.org/en/latest/Administrator%20Guide/Managing%20Volumes/#replace-brick

Read the steps at section:
"Replacing brick in Replicate/Distributed Replicate volumes".

We are working on making all the extra steps vanish and just one command 
will take care of everything going forward. Will update gluster-users 
once that happens.


Pranith

On 10/09/2015 12:50 AM, Gene Liverman wrote:
So... this kinda applies to me too and I want to get some 
clarification: I have the following setup


# gluster volume info

Volume Name: gv0
Type: Replicate
Volume ID: fc50d049-cebe-4a3f-82a6-748847226099
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: eapps-gluster01:/export/sdb1/gv0
Brick2: eapps-gluster02:/export/sdb1/gv0
Brick3: eapps-gluster03:/export/sdb1/gv0
Options Reconfigured:
diagnostics.count-fop-hits: on
diagnostics.latency-measurement: on
nfs.drc: off

eapps-gluster03 had a hard drive failure so I replaced it, formatted 
the drive and now need gluster to be happy again. Gluster put a 
.glusterfs folder in /export/sdb1/gv0 but nothing else has shown up 
and the brick is offline. I read the docs on replacing a brick but 
seem to be missing something and would appreciate some help. Thanks!






--
*Gene Liverman*
Systems Integration Architect
Information Technology Services
University of West Georgia
glive...@westga.edu 

ITS: Making Technology Work for You!



On Thu, Oct 8, 2015 at 2:46 PM, Pranith Kumar Karampuri 
> wrote:


On 3.7.4, all you need to do is execute "gluster volume
replace-brick  commit force" and rest will be taken care
by afr. We are in the process of coming up with new commands like
"gluster volume reset-brick  start/commit" for
wiping/re-formatting of the disk. So wait just a little longer :-).

Pranith


On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:


On 8 October 2015 at 07:19, Joe Julian > wrote:

I documented this on my blog at
https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/
which is still accurate for the latest version.

The bug report I filed for this was closed without
resolution. I assume there's no plans for ever making this
easy for administrators.
https://bugzilla.redhat.com/show_bug.cgi?id=991084



Yes, its the sort of workaround one can never remember in an
emergency, you'd have to google it up ...

In the case I was working with, probably easier and quicker to do
a remove-brick/add-brick.

thanks,


-- 
Lindsay



___
Gluster-users mailing list
Gluster-users@gluster.org  
http://www.gluster.org/mailman/listinfo/gluster-users



___
Gluster-users mailing list
Gluster-users@gluster.org 
http://www.gluster.org/mailman/listinfo/gluster-users




___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Pranith Kumar Karampuri



On 10/09/2015 01:32 AM, Joe Julian wrote:


On 10/08/2015 11:46 AM, Pranith Kumar Karampuri wrote:
On 3.7.4, all you need to do is execute "gluster volume replace-brick 
 commit force" and rest will be taken care by afr. We are in 
the process of coming up with new commands like "gluster volume 
reset-brick  start/commit" for wiping/re-formatting of the 
disk. So wait just a little longer :-).


Pranith



Nope.

Volume Name: test
Type: Replicate
Volume ID: 426a1719-7cc2-4dac-97b4-67491679e00e
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: questor:/tmp/foo1.1
Brick2: questor:/tmp/foo1.2


Status of volume: test
Gluster process TCP Port  RDMA Port
Online  Pid

--
Brick questor:/tmp/foo1.1   49162 0 Y   20825
Brick questor:/tmp/foo1.2   49163 0 Y   20859
NFS Server on localhost N/A   N/A N   N/A
Self-heal Daemon on localhost   N/A   N/A Y  
20887




[root@questor]# kill 20825
[root@questor]# rm -rf /tmp/foo1.1
[root@questor]# mkdir /tmp/foo1.1
[root@questor]# gluster volume replace-brick test commit force
Usage: volume replace-brick
{commit force}
[root@questor]# gluster volume replace-brick test questor:/tmp/foo1.1 
questor:/tmp/foo1.1 commit force
volume replace-brick: failed: Brick: questor:/tmp/foo1.1 not 
available. Brick may be containing or be contained by an existing brick
This is exactly the case that will be covered with "gluster volume 
reset-brick  start/commit"


Pranith




On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:


On 8 October 2015 at 07:19, Joe Julian > wrote:


I documented this on my blog at
https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/
which is still accurate for the latest version.

The bug report I filed for this was closed without resolution. I
assume there's no plans for ever making this easy for
administrators.
https://bugzilla.redhat.com/show_bug.cgi?id=991084



Yes, its the sort of workaround one can never remember in an 
emergency, you'd have to google it up ...


In the case I was working with, probably easier and quicker to do a 
remove-brick/add-brick.


thanks,


--
Lindsay


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users






___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Pranith Kumar Karampuri
On 3.7.4, all you need to do is execute "gluster volume replace-brick 
 commit force" and rest will be taken care by afr. We are in 
the process of coming up with new commands like "gluster volume 
reset-brick  start/commit" for wiping/re-formatting of the 
disk. So wait just a little longer :-).


Pranith

On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:


On 8 October 2015 at 07:19, Joe Julian > wrote:


I documented this on my blog at
https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/
which is still accurate for the latest version.

The bug report I filed for this was closed without resolution. I
assume there's no plans for ever making this easy for administrators.
https://bugzilla.redhat.com/show_bug.cgi?id=991084



Yes, its the sort of workaround one can never remember in an 
emergency, you'd have to google it up ...


In the case I was working with, probably easier and quicker to do a 
remove-brick/add-brick.


thanks,


--
Lindsay


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] How to replace a dead brick? (3.6.5)

2015-10-08 Thread Joe Julian


On 10/08/2015 11:46 AM, Pranith Kumar Karampuri wrote:
On 3.7.4, all you need to do is execute "gluster volume replace-brick 
 commit force" and rest will be taken care by afr. We are in 
the process of coming up with new commands like "gluster volume 
reset-brick  start/commit" for wiping/re-formatting of the 
disk. So wait just a little longer :-).


Pranith



Nope.

   Volume Name: test
   Type: Replicate
   Volume ID: 426a1719-7cc2-4dac-97b4-67491679e00e
   Status: Started
   Number of Bricks: 1 x 2 = 2
   Transport-type: tcp
   Bricks:
   Brick1: questor:/tmp/foo1.1
   Brick2: questor:/tmp/foo1.2


   Status of volume: test
   Gluster process TCP Port  RDMA Port
   Online  Pid
   
--
   Brick questor:/tmp/foo1.1   49162 0 Y   20825
   Brick questor:/tmp/foo1.2   49163 0 Y   20859
   NFS Server on localhost N/A   N/A N   N/A
   Self-heal Daemon on localhost   N/A   N/A Y   20887



[root@questor]# kill 20825
[root@questor]# rm -rf /tmp/foo1.1
[root@questor]# mkdir /tmp/foo1.1
[root@questor]# gluster volume replace-brick test commit force
Usage: volume replace-brick{commit 
force}
[root@questor]# gluster volume replace-brick test questor:/tmp/foo1.1 
questor:/tmp/foo1.1 commit force
volume replace-brick: failed: Brick: questor:/tmp/foo1.1 not available. 
Brick may be containing or be contained by an existing brick




On 10/08/2015 11:26 AM, Lindsay Mathieson wrote:


On 8 October 2015 at 07:19, Joe Julian > wrote:


I documented this on my blog at
https://joejulian.name/blog/replacing-a-brick-on-glusterfs-340/
which is still accurate for the latest version.

The bug report I filed for this was closed without resolution. I
assume there's no plans for ever making this easy for administrators.
https://bugzilla.redhat.com/show_bug.cgi?id=991084



Yes, its the sort of workaround one can never remember in an 
emergency, you'd have to google it up ...


In the case I was working with, probably easier and quicker to do a 
remove-brick/add-brick.


thanks,


--
Lindsay


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users




___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users