Hi,

We recently switched from a NetApp FAS250 with NFS to a dual-head
FAS270c with iSCSI + Lustre. We are running Lustre 1.4.7.3 both on the
Lustre OSS/MDS and on about 20 client machines, all RHEL 4ES Up4 boxes.

Our setup is pretty straightforward, with both NetApp heads active, each
serving one LUN over iSCSI. We have a pair of Lustre boxes, each seeing
both LUN's as the same devices and in the same order (using the
multipath driver). Each Lustre box serves as an OSS for one of the LUN's
and a failover OSS for the other. For a failover MDS setup, we've put
the MDS partition on the iSCSI fabric and not on local storage. One
Lustre box acts as both and OSS and the cluster FS MDS.

We've been trying to improve Lustre performance for our Web serving
environment (millions of small files). Naturally, there are mostly read
ops, with the ocassional write-intensive cron jobs (rsyncs for cold
backups, publishing processes, etc.).

After going through the 2006 mailing list archives and Bugzilla:

https://bugzilla.clusterfs.com/show_bug.cgi?id=10265
https://bugzilla.clusterfs.com/show_bug.cgi?id=6252

I've come up with the following to run on our OST's/MDS and clients, as
appropriate:

echo 0 > /proc/sys/portals/debug
for LRU in /proc/fs/lustre/ldlm/namespaces/*/lru_size; do
  case LRU in
    MDC*)
      echo 2500 > $LRU ;;
    OSC*)
      echo 1000 > $LRU ;;
  esac;
done
for i in `find /proc/fs/lustre -name max_read_ahead_mb`; do
  echo 4 > $i;
done
for i in `find /proc/fs/lustre -name max_read_ahead_whole_mb`; do
  echo 1 > $i;
done

Does this seem appropriate for a web serving environment? Are there
tunables that I am not using correctly? Comments would be much
appreciated.

Thank you & best, Zlatin

Zlatin Ivanov
Systems Administrator
New York Magazine
444 Madison Ave, 4th Fl
New York, NY 10022
212.508.0521

_______________________________________________
Lustre-discuss mailing list
[email protected]
https://mail.clusterfs.com/mailman/listinfo/lustre-discuss

Reply via email to