Matt, 
Based on your Gluster servers configs that file is bigger than the available 
disk space, obviously that isn't right. 

Can you send us the output of `stat y.out` taken from the Gluster mount point 
and from the back end of the server Gluster created the file on? 

I'm also going to try and reproduce the problem here on 3.1 and 3.1.1qa5. 





Thanks, 
Craig 

--> 
Craig Carl 



Gluster, Inc. 
Cell - (408) 829-9953 (California, USA) 
Gtalk - [email protected] 


From: "Matt Hodson" <[email protected]> 
To: "Craig Carl" <[email protected]> 
Cc: "Jeff Kozlowski" <[email protected]>, [email protected] 
Sent: Wednesday, November 10, 2010 9:21:40 AM 
Subject: Re: [Gluster-users] filling gluster cluster with large file doesn't 
crash the system?! 

Craig, 
inline... 



On Nov 10, 2010, at 7:17 AM, Craig Carl wrote: 




Matt - 
A couple of questions - 

What is your volume config? (`gluster volume info all`) 



gluster> volume info all 


Volume Name: gs-test 
Type: Distribute 
Status: Started 
Number of Bricks: 2 
Transport-type: tcp 
Bricks: 
Brick1: 172.16.1.76:/exp1 
Brick2: 172.16.2.117:/exp2 





What is the hardware config for each storage server? 



brick 1 = 141GB 
brick 2 = 143GB 




What command did you run to create the test data? 


#perl -e 'print rand while 1' > y.out & 





What process is still writing to the file? 



same one as above. 








Thanks, 
Craig 

--> 
Craig Carl 



Gluster, Inc. 
Cell - (408) 829-9953 (California, USA) 
Gtalk - [email protected] 



From: "Matt Hodson" < [email protected] > 
To: [email protected] 
Cc: "Jeff Kozlowski" < [email protected] > 
Sent: Tuesday, November 9, 2010 10:46:04 AM 
Subject: Re: [Gluster-users] filling gluster cluster with large file doesn't 
crash the system?! 

I should also note that on this non-production test rig the block size 
on both bricks is 1KB (1024) so the theoretical file size limit is 
16GB. so how then did i get a file of 200GB? 
-matt 

On Nov 9, 2010, at 10:34 AM, Matt Hodson wrote: 

> craig et al, 
> 
> I have a 2 brick distributed 283GB gluster cluster on CentoOS 5. we 
> nfs mounted the cluster from a 3rd machine and wrote random junk to 
> a file. i watched the file grow to 200GB on the cluster when it 
> appeared to stop. however the machine writing to the file still 
> lists the file as growing. it's now at over 320GB. what's going on? 
> 
> -matt 
> 
> ------- 
> Matt Hodson 
> Scientific Customer Support, Geospiza 
> (206) 633-4403, Ext. 111 
> http://www.geospiza.com 
> 
> 
> 
> 


_______________________________________________ 
Gluster-users mailing list 
[email protected] 
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users 

_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to