Hi, We recently switched from a NetApp FAS250 with NFS to a dual-head FAS270c with iSCSI + Lustre. We are running Lustre 1.4.7.3 both on the Lustre OSS/MDS and on about 20 client machines, all RHEL 4ES Up4 boxes.
Our setup is pretty straightforward, with both NetApp heads active, each serving one LUN over iSCSI. We have a pair of Lustre boxes, each seeing both LUN's as the same devices and in the same order (using the multipath driver). Each Lustre box serves as an OSS for one of the LUN's and a failover OSS for the other. For a failover MDS setup, we've put the MDS partition on the iSCSI fabric and not on local storage. One Lustre box acts as both and OSS and the cluster FS MDS. We've been trying to improve Lustre performance for our Web serving environment (millions of small files). Naturally, there are mostly read ops, with the ocassional write-intensive cron jobs (rsyncs for cold backups, publishing processes, etc.). After going through the 2006 mailing list archives and Bugzilla: https://bugzilla.clusterfs.com/show_bug.cgi?id=10265 https://bugzilla.clusterfs.com/show_bug.cgi?id=6252 I've come up with the following to run on our OST's/MDS and clients, as appropriate: echo 0 > /proc/sys/portals/debug for LRU in /proc/fs/lustre/ldlm/namespaces/*/lru_size; do case LRU in MDC*) echo 2500 > $LRU ;; OSC*) echo 1000 > $LRU ;; esac; done for i in `find /proc/fs/lustre -name max_read_ahead_mb`; do echo 4 > $i; done for i in `find /proc/fs/lustre -name max_read_ahead_whole_mb`; do echo 1 > $i; done Does this seem appropriate for a web serving environment? Are there tunables that I am not using correctly? Comments would be much appreciated. Thank you & best, Zlatin Zlatin Ivanov Systems Administrator New York Magazine 444 Madison Ave, 4th Fl New York, NY 10022 212.508.0521 _______________________________________________ Lustre-discuss mailing list [email protected] https://mail.clusterfs.com/mailman/listinfo/lustre-discuss
