There is tool called bonnie++ for file system
benchmarking 
http://www.coker.com.au/bonnie++/

aslo there some results of benchmarking  at reiserfs
 site www.namesys.com 

Regards

UW







--- "Linda W." <[EMAIL PROTECTED]> wrote:
> Henrik Nordstrom wrote:
> 
> >On Sat, 25 Oct 2003, Linda W. wrote:
> >
> >  
> >
> >>Was about to move my squid directory off onto it's
> own partition and was
> >>wondering what filesystem to use, in the sense of
> is there a linux (x86)  
> >>filesystem that performs best for squid?  Any
> special params for block
> >>size?  It's just a single SCSI disk.
> >>    
> >>
> >
> >The general consensus is that reiserfs mounted with
> noatime is currently
> >the best performer for a Linux Squid cache. This
> conclusion arrived after
> >countless benchmarks in different configurations,
> mostly thanks to Joe 
> >Cooper.
> >  
> >
> ---
>     I'm slightly confused -- do you mean reiserfs is
> best out of the 
> journalled
> fs's, or best including non-journaled async (ext2?
> fat32?) fs's.
> 
> >But you can always set up your own benchmarks to
> see what runs best on 
> >your hardware. For benchmarking I highly recommends
> the polygraph 
> >benchmark program with polymix-4 workload.
> >
> >Only problem with benchmarking is that you need at
> least two extra
> >computers to run the benchmark (one acting as
> client, one acting as
> >server), and that it takes some tries before one is
> used to how to run the
> >benchmarks..
> >  
> >
> ---
>     Doing benchmarks right is fairly difficult.  So
> many variables.  So 
> many parameters
> can affect things.  Like just choice of fs's default
> allocation unit.  
> If a format prog has
> defaults of a 512-byte allocation block, it might
> make a big difference 
> in a test where
> another sets up for 16Kb blocks.  Defaults could
> explain a difference in 
> performance
> if most read/writes are >512 bytes and <16Kb.
> 
>     Do you know off hand what Reiserfs's default
> alloc size is?
> 
> > > I'm guessing but a journaling fs might slow it
> down?
> >
> >Depends.
> >
> >A journalled filesystem can be a lot faster than a
> syncronous
> >non-journaled filesystem and also gives a better
> level of fault tolerance.
> >
> >A simple asyncronous non-journalled filesystem is
> almost faster than a 
> >journalled filesystem, but is at the same time very
> sensitive to errors.
> >  
> >
> ---
>    Aren't ext2 and fat32, ufs, etc....all pretty
> much 
> async/non-journaled?  Weren't they
> (and in many cases, still are) used for decades
> without being 
> "sensitive"?  Yes, a system
> crash required an fsck/chkdsk, but if the OS doesn't
> crash that often, 
> is it really
> "sensitive". 
> 
>     FAT32 and ext2 only mis-behave during system
> failures (a common 
> event pre win2000),
> but win2k and xp don't keel over and die unexpected
> as often and only 
> rarely do I have 
> uncontrolled shutdowns -- and my linux system? My
> average system uptime 
> over the
> past 2 months has been about a week
> (installing new HW).  Half of those were crashes --
> I use a journaling 
> fs (xfs).   Before
> that, last wtmp averaged out at a 29 day uptime. 
> 
> Bugs happen in journaling fs's too -- all of the
> files I'd modified in 
> the previous day had '0's
> written throughout them.  Yep -- somehow the journal
> replayed all of the 
> file transactions
> going back about 36 hours with all binary zeros. 
> The backup job files 
> that were created
> during the morning were erased (on separate hard
> disk).  The backup from 
> the morning
> before was intact.  Never had a 'sensitive' file
> system do such 
> comprehensive 'forgetting'
> of every file that had been touched in previous 24
> hours.  Log files 
> were reset as though
> the 24th never happened.  Trippy.  Very
> rare/fluke...but also, 
> unfortunately, a possibility.
> 
> >>I recently ran a Sandra filesystem benchmark on
> FAT32/NTFS and found
> >>NTFS was around 3-4x slower than FAT32.  Suprised
> me since NTFS is
> >>supposed to be MS's state-of-the-fart file system,
> but I wondered if the
> >>journaling was slowing it.
> >>    
> >>
> >
> >NTFS is a cool filesystem design, but yes,
> jornaling slows it down a 
> >little. It is not really fair to compare NTFS to
> FAT32 on NT as the FAT32 
> >is completely asyncronous with absolutely no fault
> tolerance.
> >  
> >
> ----
>     What other windows file system would one compare
> NTFS to?  BTW,  at 
> one point, I thought
> I remember fat32 being syncronous on linux. 
> Theoretically, with no 
> support for access rights,
> file owner and limited time field accuracy, FAT32
> should run faster than 
> ext2. 
> 
>     But -- for a 'temporary internet cache', how
> much fault tolerance 
> does one need?  I could
> see (if  memory was cheap enough) of running squid
> on a RAM disk.  If 
> your server stays up
> for a month at a time, I think the effects of losing
> the cache once a 
> month would be negligible
> compared to the benefit of zero ms disk access...
> 
> >>I wonder...if one stuck a mySQL database on the
> back end of squid for a
> >>FS driver and ran mySQL using one big 'file' that
> was namd
> >>/dev/sdAx...or weirder /dev/raw/sdAx (or whatever
> the syntax would be).
> >>    
> >>
> >
> >Juck.. why would one want to do so?
> >  
> >
> ---
>     I dunno...the algorithms to store and retrieve
> data in a database 
> might have been given
> more research bucks to be optimized for speed than
> the the squid 
> database on top of a
> file system delay.  It's a WAG (Wild Ass
> Guess)...but databases place 
> heavy emphasis on
> getting high TPS -- something similar to what squid
> (and apache).  But 
> for squid, it's just
> retrieval, not a great deal of processing...I think
> retrieving static 
> data from a database should
> be something a database would have to have
> excelllent performance on to 
> be successful.
> 
=== message truncated ===


__________________________________
Do you Yahoo!?
Exclusive Video Premiere - Britney Spears
http://launch.yahoo.com/promos/britneyspears/

Reply via email to