craig,
A) stat on server that generated the file:
j...@riff:/mnt/gluster$ stat y.out
File: `y.out'
Size: 214574579712 Blocks: 291441832 IO Block: 65536 regular file
Device: 16h/22d Inode: 14213316644377695875 Links: 1
Access: (0777/-rwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root)
Access: 2010-11-09 10:32:16.000000000 -0800
Modify: 2010-11-09 13:57:05.000000000 -0800
Change: 2010-11-09 13:57:05.000000000 -0800
B) stat on gluster brick (brick2) that is housing the file:
[r...@brick2 ~]# stat /exp2/y.out
File: `/exp2/y.out'
Size: 214574579712 Blocks: 291441832 IO Block: 4096 regular file
Device: fd00h/64768d Inode: 655412 Links: 1
Access: (0777/-rwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root)
Access: 2010-11-09 10:32:16.000000000 -0800
Modify: 2010-11-09 13:57:05.000000000 -0800
Change: 2010-11-09 13:57:05.000000000 -0800
c) df on brick 2
[r...@brick2 ~]# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/VolGroup00-LogVol00
143G 143G 0 100% /
/dev/hda1 99M 13M 82M 13% /boot
tmpfs 470M 0 470M 0% /dev/shm
172.16.1.76:/gs-test 283G 158G 119G 58% /mnt/gluster
interesting that block size is different depending on who you ask.
[r...@brick2 ~]# dumpe2fs /dev/hda1 | grep -i 'Block size'
dumpe2fs 1.39 (29-May-2006)
Block size: 1024
-Matt
On Nov 11, 2010, at 9:44 PM, Craig Carl wrote:
Matt,
Based on your Gluster servers configs that file is bigger than the
available disk space, obviously that isn't right.
Can you send us the output of `stat y.out` taken from the Gluster
mount point and from the back end of the server Gluster created the
file on
I'm also going to try and reproduce the problem here on 3.1 and
3.1.1qa5.
Thanks,
Craig
-->
Craig Carl
Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Gtalk - [email protected]
From: "Matt Hodson" <[email protected]>
To: "Craig Carl" <[email protected]>
Cc: "Jeff Kozlowski" <[email protected]>, [email protected]
Sent: Wednesday, November 10, 2010 9:21:40 AM
Subject: Re: [Gluster-users] filling gluster cluster with large file
doesn't crash the system?!
Craig,
inline...
On Nov 10, 2010, at 7:17 AM, Craig Carl wrote:
Matt -
A couple of questions -
What is your volume config? (`gluster volume info all`)
gluster> volume info all
Volume Name: gs-test
Type: Distribute
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: 172.16.1.76:/exp1
Brick2: 172.16.2.117:/exp2
What is the hardware config for each storage server?
brick 1 = 141GB
brick 2 = 143GB
What command did you run to create the test data?
#perl -e 'print rand while 1' > y.out &
What process is still writing to the file?
same one as above.
Thanks,
Craig
-->
Craig Carl
Gluster, Inc.
Cell - (408) 829-9953 (California, USA)
Gtalk - [email protected]
From: "Matt Hodson" <[email protected]>
To: [email protected]
Cc: "Jeff Kozlowski" <[email protected]>
Sent: Tuesday, November 9, 2010 10:46:04 AM
Subject: Re: [Gluster-users] filling gluster cluster with large file
doesn't crash the system?!
I should also note that on this non-production test rig the block size
on both bricks is 1KB (1024) so the theoretical file size limit is
16GB. so how then did i get a file of 200GB?
-matt
On Nov 9, 2010, at 10:34 AM, Matt Hodson wrote:
> craig et al,
>
> I have a 2 brick distributed 283GB gluster cluster on CentoOS 5. we
> nfs mounted the cluster from a 3rd machine and wrote random junk to
> a file. i watched the file grow to 200GB on the cluster when it
> appeared to stop. however the machine writing to the file still
> lists the file as growing. it's now at over 320GB. what's going on?
>
> -matt
>
> -------
> Matt Hodson
> Scientific Customer Support, Geospiza
> (206) 633-4403, Ext. 111
> http://www.geospiza.com
>
>
>
>
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users