Hi,
I am using Lustre 1.6.0.1 with one OST and 20 clients in an
HPC cluster.
The OST/MDT/MGS has a 16 channel 3ware 9650 using raid6. I currently
have another lustre installation
(version 1.4.5) and it has been working trouble free for over an year.
The OS is CentOS 4. There are 4 network
ports in the storage server in adaptive load balanced mode and aggregate
network throughout is great (with 4 x netperf/iperf from clients)
in an ideal situation when clients pick up different mac addresses of
the different interfaces in their arp table.
I have a few questions about Lustre and hope someone can help me.
* I had to re-export the lustre volume via nfs on the new 1.6.0.1 setup
to other infrastructure boxes.
After the export, i get the following error messages in the OSS -
Jun 21 09:31:11 lustre-3ware kernel: Lustre:
4946:0:(lustre_fsfilt.h:205:fsfilt_start_log()) scratch-OST0000: slow
journal start 33s
Jun 21 09:31:11 lustre-3ware kernel: Lustre:
4946:0:(lustre_fsfilt.h:205:fsfilt_start_log()) Skipped 22 previous
similar messages
Jun 21 09:31:11 lustre-3ware kernel: Lustre:
4874:0:(filter.c:1139:filter_parent_lock()) scratch-OST0000: slow parent
lock 33s
Jun 21 09:31:11 lustre-3ware kernel: Lustre:
4874:0:(filter.c:1139:filter_parent_lock()) Skipped 6 previous similar
messages
Also is the NFS re-export option stable in version 1.6? I read some
posts before in the list reporting kernel panics on Lustre 1.4.
*I was evaluating GFS for the past few weeks with GNBD and the
performance was amazing (at least for my purpose with one storage
server). It was very fast, especially for small files.
But i had to dump it because of stability reasons. The problems were
these - has 6 daemons that need to come up in a particular order. If
some of
the kernel modules crash on heavy load on a node, the whole cluster
freezes. It had the issue of quorum, which is beneficial on a HA setup,
may be not for HPC.
In some cases, i have to keep just one server running that re-exports
the volume via nfs even if the hpc nodes are down. Like during a power
failure for example. Quorum is a
problem in that case. But it was mostly stability that made me not go
with GFS + GNBD.
*Now the problem - Lustre performance dips a lot when it comes to small
files. Please see the following fileop -f 5 test comparing NFS and Lustre -
Lustre -
Fileop: File size is 1, Output is in Ops/sec. (A=Avg, B=Best, W=Worst)
. mkdir rmdir create read write close stat access chmod
readdir link unlink delete Total_files
A 5 1654 691 132 14228 719 4874 1987 32737 1718
2506 1262 1340 1608 125
NFS -
Fileop: File size is 1, Output is in Ops/sec. (A=Avg, B=Best, W=Worst)
. mkdir rmdir create read write close stat access chmod
readdir link unlink delete Total_files
A 5 177 594 459 380747 137392 2282 1219 444312 502
1274 306 513 464 125
Could you please recommend any tunables to get a bit more
performance out of Lustre with lots of small files? Lots of small files
was bad in GFS too, but
it was better than NFS though.
*Also the read performance of Lustre seems to be a little behind NFS. I
had /opt which has all the software for users moved to Lustre in the new
setup. But
software like Matlab, Splus etc takes almost a minute to come up. The
second time is very fast though, maybe due to caching. So i am thinking
of putting /opt
back to NFS. Is it possible to boost the read performance of Lustre a bit?
*Is there a way to make disk quotas activate at startup automatically on
a Lustre client? The lfs quotaon <mount point> works sometimes. But
sometimes it gives an a resource busy error message.
*One last question. In the older Lustre setup (version 1.4.5), i have 5
scsi drives one each as an OST for a single volume. The volume became
full. But df still reported
that there is 27GB free. There doesn't seem to be an lfs df option in
that version of Lustre. So i couldn't see the individual utilization of
each of the 5 OST. Is this a striping
problem?
I know it's a lot of questions. Hope some of them are
solvable. Thanks very much.
Best Regards
Balagopal Pillai
_______________________________________________
Lustre-discuss mailing list
[email protected]
https://mail.clusterfs.com/mailman/listinfo/lustre-discuss