Hmmm, now that I think about it, what really surprises me is 16MB/s with
4MB block size.  Have you tried different file sizes?  I assume the
default max request size in src/io/dev/pint-dev-shared.h is still

#define PVFS2_BUFMAP_DEFAULT_DESC_SIZE  (4 * (1024 * 1024))
#define PVFS2_BUFMAP_DEFAULT_DESC_SHIFT 22 /* NOTE: 2^22 == 4MB */

Maybe someone else has some thoughts?
Dean

On Tue, 6 Dec 2005, Praveen KJ wrote:

> Dean,
>         Here is the snippet from the output.
>
>         Command line used: /opt/iozone/bin/iozone -r 128k -w -f file1 -s
> 4g -w -i 1
>                                                             random
> random    bkwd  record stride
>               KB  reclen   write rewrite    read    reread    read
> write    read rewrite   read   fwrite frewrite   fread  freread
>          4194304     128                    63600
> 63100
>
>
> -Praveen
>
> Dean Hildebrand wrote:
>
> >65MB/s for 128K is really suprising.  My experience with pvfs2 is that
> >performance is worse for smaller block size.  (although the graph is
> >sometimes wavy with an upward trend).  I wonder if iozone is behaving
> >properly.......  What is your iozone command line?  Instead of
> >iozone -aec -i 0 -i 1 -+n -f /path/to/pvfs2/file
> >
> >have you tried specific file sizes and ranges
> >iozone -ec -i 0 -i 1 -+n -r 128K -s 100M -f /path/to/pvfs2/file
> >iozone -ec -i 0 -i 1 -+n -r 1M -s 100M -f /path/to/pvfs2/file
> >iozone -ec -i 0 -i 1 -+n -r 4M -s 100M -f /path/to/pvfs2/file
> >
> >Dean
> >
> >On Tue, 6 Dec 2005, Praveen KJ wrote:
> >
> >
> >
> >>Rob/Murali/Dean,
> >>
> >>                     The read throughput seems to be variable
> >>considering the iozone record size.
> >>The experimental setup was same as described before.  The variation of
> >>record size and throughput for a sample  ..
> >>---------------------------------------------------------------------------------------------
> >>record_size         throughput       comments
> >>---------------------------------------------------------------------------------------------
> >>                                               ( Steady increment of
> >>throughput till 128k)
> >>128k -                65 MB/s        ( Steady decrement of throughput
> >>till 1M)
> >>1M   -                6 MB/s           ( Steady increment till - )
> >>4 M   -               16 MB/s
> >>
> >>                     A graph plot would look wavy.
> >>
> >>                     Iozone is not our final application :)  We just
> >>wanted to observe the performance space and also make sure installation
> >>was done right.  So your inputs were definitely helpful.
> >>
> >>Thanks,
> >>Praveen
> >>
> >>Murali Vilayannur wrote:
> >>
> >>
> >>
> >>>Thanks Dean!
> >>>I will take a look at the VFS code paths/performance analysis.
> >>>Murali
> >>>
> >>>On Wed, 30 Nov 2005, Dean Hildebrand wrote:
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>>I've seen this through VFS.  Here are some graphs (ignore the pNFS lines)
> >>>>with 4MB block size and the standard read/write interfaces.
> >>>>Overview:
> >>>>http://www.citi.umich.edu/projects/asci/pnfs/darkstar-exp/
> >>>>
> >>>>Single file:
> >>>>http://www.citi.umich.edu/projects/asci/pnfs/darkstar-exp/experiments-5_13095_image001.gif
> >>>>
> >>>>Separate files:
> >>>>http://www.citi.umich.edu/projects/asci/pnfs/darkstar-exp/experiments-5_15651_image001.gif
> >>>>
> >>>>As the # of data servers increases, aggregate read performance increases
> >>>>but individual client read performance suffers a little.
> >>>>
> >>>>Dean
> >>>>
> >>>>On Wed, 30 Nov 2005, Murali Vilayannur wrote:
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>>Hi Praveen,
> >>>>>Sorry for the late response. Unfortunately, we have not seen this kind of
> >>>>>behavior on many of our setups thru the MPI-I/O/ADIO-PVFS2 interfaces.
> >>>>>It is conceivable that performance thru the VFS do not scale with
> >>>>>number of servers.
> >>>>>
> >>>>>Could you tell us how the file is being read (i.e using what interfaces,
> >>>>>what block sizes etc)?
> >>>>>There still aren't that many tweakable settings just yet, but we are
> >>>>>working on some I/O performance tuning and we will hopefully have a 
> >>>>>better
> >>>>>handle on what settings affect performance.
> >>>>>Thanks for the reports,
> >>>>>Murali
> >>>>>
> >>>>>On Mon, 21 Nov 2005, Praveen KJ wrote:
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>>Hi,
> >>>>>>    We have a pvfs2 client reading a 5 GB file sequentially from
> >>>>>>multiple pvfs2 data servers.
> >>>>>>    There is just one metadata server. All the nodes are connected via
> >>>>>>GigE network.
> >>>>>>
> >>>>>>    The problem we are seeing is that the single-client read throughput
> >>>>>>decreases as the number of servers increase.
> >>>>>>    With 24 servers we see around 5 MB/s read compared to 50 MB/s read
> >>>>>>with just 2 servers.
> >>>>>>    The setup between each experiment is exactly same, except the
> >>>>>>number of servers.
> >>>>>>
> >>>>>>    We have tried a couple of pvfs2 versions including the latest 1.3.1.
> >>>>>>    Is there any setting we have to tweak ?  We have taken only pvfs2
> >>>>>>defaults till now.
> >>>>>>    Some of the relevant options are pasted below.
> >>>>>>
> >>>>>>Thanks,
> >>>>>>Praveen
> >>>>>>
> >>>>>>------------------------------------------
> >>>>>>pvfs2-fs.conf ( except hostnames )
> >>>>>>------------------------------------------
> >>>>>><Defaults>
> >>>>>>      UnexpectedRequests 50
> >>>>>>      LogFile /tmp/pvfs2-server.log
> >>>>>>      EventLogging none
> >>>>>>      LogStamp usec
> >>>>>>      BMIModules bmi_tcp
> >>>>>>      FlowModules flowproto_multiqueue
> >>>>>>      PerfUpdateInterval 1000
> >>>>>>      ServerJobBMITimeoutSecs 30
> >>>>>>      ServerJobFlowTimeoutSecs 30
> >>>>>>      ClientJobBMITimeoutSecs 300
> >>>>>>      ClientJobFlowTimeoutSecs 300
> >>>>>>      ClientRetryLimit 5
> >>>>>>      ClientRetryDelayMilliSecs 2000
> >>>>>></Defaults>
> >>>>>>
> >>>>>><Filesystem>
> >>>>>>      Name pvfs2-fs
> >>>>>>      ID 1175641402
> >>>>>>      RootHandle 1048576
> >>>>>>      <StorageHints>
> >>>>>>              TroveSyncMeta yes
> >>>>>>              TroveSyncData no
> >>>>>>              AttrCacheKeywords datafile_handles,metafile_dist
> >>>>>>              AttrCacheKeywords dir_ent, symlink_target
> >>>>>>              AttrCacheSize 4093
> >>>>>>              AttrCacheMaxNumElems 32768
> >>>>>>      </StorageHints>
> >>>>>></Filesystem>
> >>>>>>_______________________________________________
> >>>>>>PVFS2-users mailing list
> >>>>>>[email protected]
> >>>>>>http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>_______________________________________________
> >>>>>PVFS2-users mailing list
> >>>>>[email protected]
> >>>>>http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>
> >>>>
>
_______________________________________________
PVFS2-users mailing list
[email protected]
http://www.beowulf-underground.org/mailman/listinfo/pvfs2-users

Reply via email to