Hi,

I have been evaluating the parallel read performance of PVFS2 and wonder
how I can get better performance. 

Hardware configuration:

- 100 nodes, each with two quad-core processors and 16 GB memory
- 1 GigE connection out of each node and connecting to a 1GigE switch in
the rack (3 racks in total)
- Each rack switch has 4 10-GigE connections to a backbone
full-bandwidth 10-GigE switch
- Software (md) RAID0 on 4 SATA disks, with a capacity of 500 GB per
node
- Raw RAID0 bulk data transfer rate around 200 MB/s  (dd a 4GB file
after dropping linux vfs cache)

Software configuration:

- 2.6.26-10smp kernel
- PVFS 2.8.1 
- All 100 nodes running both data servers and metadata servers
- All 100 nodes running application programs (MPI codes) and accessing
data via PVFS Linux kernel interface 
- PVFS mounted on each node via the local host (since all of them are
also metadata servers)
- Data striping across all the nodes

Experiment setup:

- Four datasets consisting of files of 1 GB, 256 MB, 64 MB, and 2 MB,
respectively
- Each dataset has 1.6 terabyte data (that is, 1600 1GB files, 6400
256MB files, etc.)
- A simple parallel (MPI) program that reads all the files of each
dataset
     * Read tasks split among the MPI process evenly (standard block
decomposition algorithm)
     * Used all 100 nodes 
     * Varied the number of cores used (1, 2, 4)
     * MPI_Barrier() called before and after the reading of all the
files assigned to a process
- Experiments conducted with no interference from other applications

Below is the sustained read performance I measured. During the
experiments, the network traffic peaked at around 2 GB/s as shown by
Ganglia running on the cluster.

                                1 reader per node     2 readers per node
4 readers per node

1 GB file dataset         660 MB/s               1218 MB/s
1106 MB/s
256 MB file dataset     663 MB/s               1205 MB/s
1153 MB/s
64 MB file dataset       677 MB/s               1302 MB/s
1377 MB/s
2 MB file dataset         502 MB/s               549 MB/s
587 MB/s


Given the hardware/software configuration,  are these reasonable
performance results or shall I expect better outcome? Thanks. 

Tiankai


_______________________________________________
Pvfs2-users mailing list
[email protected]
http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users

Reply via email to