Here is the output of volume profile info after 1.1Gb file is copied
over mounted glusterfs on client:
[root@gfs1 data]# gluster volume profile test info
Brick: 93.123.32.41:/data
-------------------------
Cumulative Stats:
Block Size: 131072b+
No. of Reads: 0
No. of Writes: 8192
%-latency Avg-latency Min-Latency Max-Latency No. of
calls Fop
--------- ----------- ----------- -----------
------------ ----
0.00 0.00 us 0.00 us 0.00 us 1 RELEASE
0.02 114.00 us 35.00 us 208.00 us 5 LOOKUP
0.26 9306.00 us 9306.00 us 9306.00 us 1 CREATE
0.75 26260.00 us 26260.00 us 26260.00 us 1 FLUSH
98.97 424.91 us 165.00 us 14056.00 us 8192 WRITE
Duration: 18352 seconds
Data Read: 0 bytes
Data Written: 1073741824 bytes
Interval 14 Stats:
Block Size: 131072b+
No. of Reads: 0
No. of Writes: 5353
%-latency Avg-latency Min-Latency Max-Latency No. of
calls Fop
--------- ----------- ----------- -----------
------------ ----
0.00 0.00 us 0.00 us 0.00 us 1 RELEASE
1.22 26260.00 us 26260.00 us 26260.00 us 1 FLUSH
98.78 397.60 us 169.00 us 13423.00 us 5353 WRITE
Duration: 172 seconds
Data Read: 0 bytes
Data Written: 701628416 bytes
[root@gfs1 data]#
---
Find out about our new Cloud service - Cloudware.bg
<http://cloudware.bg/?utm_source=email&utm_medium=signature&utm_content=link&utm_campaign=newwebsite>
Access anywhere. Manage it yourself. Pay as you go.
------------------------------------------------------------------------
*Yavor Marinov*
System Administrator
Neterra Ltd.
Telephone: +359 2 975 16 16
Fax: +359 2 975 34 36
Mobile: +359 888 610 048
www.neterra.net <http://www.neterra.net>
On 05/23/2013 04:54 PM, Michael Brown wrote:
That's only a single data point (one LOOKUP call) - this tells you
nothing.
You need to check after it's been running (and processing traffic) for
a while.
For example, here's the stats off one of my bricks (an SSD):
Brick: fearless2:/export/bricks/500117310007a84c/glusterdata
------------------------------------------------------------
Cumulative Stats:
Block Size: 32b+ 64b+ 128b+
No. of Reads: 0 0 1
No. of Writes: 1 5634 4252
Block Size: 256b+ 512b+ 2048b+
No. of Reads: 0 1 0
No. of Writes: 343 24 1
Block Size: 4096b+ 8192b+ 16384b+
No. of Reads: 8 7 10
No. of Writes: 4 0 0
Block Size: 32768b+ 65536b+ 131072b+
No. of Reads: 25 165 436
No. of Writes: 2 7 36
%-latency Avg-latency Min-Latency Max-Latency No. of
calls Fop
--------- ----------- ----------- -----------
------------ ----
0.00 0.00 us 0.00 us 0.00 us
120 FORGET
0.00 0.00 us 0.00 us 0.00 us
2940 RELEASE
0.00 0.00 us 0.00 us 0.00 us 4554
RELEASEDIR
0.00 104.00 us 104.00 us 104.00 us 1
TRUNCATE
0.00 125.00 us 110.00 us 140.00 us
2 XATTROP
0.01 12.62 us 8.00 us 25.00 us
146 ACCESS
0.04 81.77 us 65.00 us 112.00 us 60
SETXATTR
0.05 30.99 us 25.00 us 57.00 us
212 SETATTR
0.06 12.73 us 8.00 us 59.00 us
574 INODELK
0.09 188.43 us 140.00 us 244.00 us
60 CREATE
0.11 25.24 us 15.00 us 149.00 us
533 STATFS
0.12 260.72 us 206.00 us 430.00 us
60 MKDIR
0.26 11.34 us 5.00 us 127.00 us
2925 FLUSH
0.27 15.14 us 7.00 us 90.00 us
2274 ENTRYLK
0.36 102.52 us 81.00 us 161.00 us
442 RMDIR
0.63 27.61 us 17.00 us 606.00 us
2880 OPEN
0.76 171.58 us 91.00 us 5691.00 us
555 UNLINK
0.87 22.66 us 8.00 us 469.00 us
4812 READDIR
0.87 24.37 us 10.00 us 1302.00 us
4506 STAT
0.94 61.67 us 16.00 us 194.00 us 1917
GETXATTR
1.06 51.20 us 10.00 us 224.00 us
2600 FSTAT
1.14 31.46 us 18.00 us 1016.00 us
4554 OPENDIR
2.56 31.19 us 18.00 us 4373.00 us
10304 WRITE
2.58 417.28 us 15.00 us 1860.00 us
776 READ
3.64 17.26 us 6.00 us 4824.00 us 26507
FINODELK
24.03 146.42 us 49.00 us 9854.00 us 20622
FXATTROP
26.67 652.70 us 42.00 us 89705.00 us 5134
READDIRP
32.86 128.19 us 9.00 us 4617.00 us
32204 LOOKUP
On 13-05-23 09:03 AM, Yavor Marinov wrote:
I've just enabled profiling of the volume and this is the information
from the profile info printed:
[root@gfs1 ~]# gluster volume profile test info
Brick: 93.123.32.41:/data
-------------------------
Cumulative Stats:
%-latency Avg-latency Min-Latency Max-Latency No. of
calls Fop
--------- ----------- ----------- -----------
------------ ----
100.00 148.00 us 148.00 us 148.00 us
1 LOOKUP
Duration: 13950 seconds
Data Read: 0 bytes
Data Written: 0 bytes
Interval 4 Stats:
Duration: 7910 seconds
Data Read: 0 bytes
Data Written: 0 bytes
[root@gfs1 ~]#
Anything here that might be useful ?
---
Find out about our new Cloud service - Cloudware.bg
<http://cloudware.bg/?utm_source=email&utm_medium=signature&utm_content=link&utm_campaign=newwebsite>
Access anywhere. Manage it yourself. Pay as you go.
------------------------------------------------------------------------
*Yavor Marinov*
System Administrator
Neterra Ltd.
Telephone: +359 2 975 16 16
Fax: +359 2 975 34 36
Mobile: +359 888 610 048
www.neterra.net <http://www.neterra.net>
On 05/23/2013 01:10 PM, Явор Маринов wrote:
I've made a mistake we are using 30Mbit connectivity on all of the
nodes. Below is a iperf test between the node and the client
[root@gfs4 ~]# iperf -c 93.123.32.41
------------------------------------------------------------
Client connecting to 93.123.32.41, TCP port 5001
TCP window size: 23.2 KByte (default)
------------------------------------------------------------
[ 3] local 93.123.32.44 port 49838 connected with 93.123.32.41 port
5001
[ ID] Interval Transfer Bandwidth
[ 3] 0.0-10.1 sec 49.9 MBytes 41.5 Mbits/sec
[root@gfs4 ~]#
But when trying to copy a 1Gb file on the client's mounted volume
the speed between the client and the node is ~500kb/s
---
Find out about our new Cloud service - Cloudware.bg
<http://cloudware.bg/?utm_source=email&utm_medium=signature&utm_content=link&utm_campaign=newwebsite>
Access anywhere. Manage it yourself. Pay as you go.
------------------------------------------------------------------------
*Yavor Marinov*
System Administrator
Neterra Ltd.
Telephone: +359 2 975 16 16
Fax: +359 2 975 34 36
Mobile: +359 888 610 048
www.neterra.net <http://www.neterra.net>
On 05/23/2013 12:16 PM, Nux! wrote:
On 23.05.2013 09:41, Явор Маринов wrote:
Thanks for your reply.
No matter how many nodes (currently the volume is only with its own
node) the speed is really slow. For testing purposes, i made a volume
with only one node, without any replication - however the speed is
still ~500kb/s. The cloud servers are limited to 30Gbit/s but still
the traffic when writing to the node is ~500kb/s
i'm using 3.3.1 glusterfsd with kernel 2.6.18-348.el5xen and i need
to know if the the problem is within the kernel.
I don't think it is a problem with gluster; I never used el5 for
this, but I doubt there's an inherent problem with it either. That
speed limit looks odd to me and I think it's somewhere in your setup.
Have you done any actual speed tests in the VMs?
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users
--
Michael Brown | `One of the main causes of the fall of
Systems Consultant | the Roman Empire was that, lacking zero,
Net Direct Inc. | they had no way to indicate successful
☎: +1 519 883 1172 x5106 | termination of their C programs.' - Firth
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users