[zfs-discuss] Checksum error in single device ZFS pool mysterious resilvering after reboot

2008-02-09 Thread Joe Peterson
I have recently been experimenting with ZFS (under FreeBSD 7.0), and I am
excited about the possibilities.  I have seen, however, two puzzling issues,
and I am hoping someone here might have some insight or info.  Note that the
original zpool that I am discussing is gone now - I had to recreate a new one
to make space for a larger root partition, etc...

Note that I had one ZFS pool with only a single device/partition (i.e. no
mirroring or raidz).  Anyway, the two issues are:

1) During a routine test scrub, a checksum error was found.  There were also
some other errors reported around this time, but not fatal (note that I am
investigating potential FreeBSD DMA messages in my log - could be related; the
hardware seems to check out fine, BTW).  This checksum error was persistent
due to the fact that there was no mirror.  One file was affected, and upon
further investigation, a 64K block (1/2 of a ZFS block?) appeared to be data
from some other file.

2) After my first reboot from the above situation, I found that zpool status
reported that resilvering completed  Also, all error counts were zeroed
(not unusual), but the pool now reported that no known errors existed.
Rescrubbing revealed the checksum error again, but this time no subsequent
reboots ever caused resilvering or hiding of the file error again.  The
puzzling thing is why would ZFS resilver anything, given that there was only
one device in the pool?  Plus, the fact that the pool now thought it was OK
was disturbing.

I am, of course, unsure if 64K of the original file was corrupted at the time
of the original file wrote, some time after that (well, at least it appeared
to be text data that would have been either from another file, specifically a
Mozilla mork DB file, or perhaps on its way to another file and mistakenly
written into this file (although no other files had checksum issues)), or if
the strange resilver caused the block to become this way.  The history of
that file is that I had copied my entire home dir to the ZFS pool with no
errors reported, and I had not touched that file before the scrub found the
problem.

I did some forensics later by temporarily disabling the checksum error in the
ZFS code in order to read the whole file (this is how I revealed the contents
of the bad block).  The bad 64K was only a small part of the file, starting at
byte 655360 and going for exactly 64K.  The rest of the file, up to 3MB or so,
was fine.

Anyway, does any of the above ring any bells?

Thanks, Joe
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS Performance Issue

2008-02-09 Thread Henk Langeveld
William Fretts-Saxton wrote:
 Unfortunately, I don't know the record size of the writes.  Is it as
 simple as looking @ the size of a file, before and after a client
 request, and noting the difference in size?

and

 The I/O is actually done by RRD4J, [...] a Java version of 'rrdtool'

If it behaves like rrdtool, it will limit the size of the file, by
consolidating older data.  After every n samples, older data will be
replaced by an aggregate, freeing space for new samples.  To me that
implies random I/O.  You really need a tool like dtrace (or old
fashioned truss) to see the sample rate and size.

Cheers,
Henk
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] OpenSolaris, ZFS and Hardware RAID, a recipe for success?

2008-02-09 Thread Nick
I have been tasked with putting together a storage solution for use in a 
virtualization setup, serving NFS, CIFS, and iSCSI, over GigE. I've inherited a 
few components to work with:

x86 dual core server , 512MB LSI-ELP RAID card
12 x 300GB 15Krpm SAS disks  array
2GB Flash to IDE disk/adaptor.

The system will be serving virtual hard disks to a range of vmware systems 
connected by GigE, running enterprise workloads that are impossible to predict 
at this point.

Using the RAID cards capability for RAID6 sounds attractive?
Using the Flash RAM for the ZIL?
Using zfs for general storage management?

Has anyone built a similar system, what is the true path to success?
What are the pitfalls?
What should I have on my reading list for starters?

All advice most gratefully received.

MM.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OpenSolaris, ZFS and Hardware RAID, a recipe for success?

2008-02-09 Thread Richard Elling
Nick wrote:
 I have been tasked with putting together a storage solution for use in a 
 virtualization setup, serving NFS, CIFS, and iSCSI, over GigE. I've inherited 
 a few components to work with:

   x86 dual core server , 512MB LSI-ELP RAID card
   12 x 300GB 15Krpm SAS disks  array
   2GB Flash to IDE disk/adaptor.

 The system will be serving virtual hard disks to a range of vmware systems 
 connected by GigE, running enterprise workloads that are impossible to 
 predict at this point.

   Using the RAID cards capability for RAID6 sounds attractive?
   

Assuming the card works well with Solaris, this sounds like a
reasonable solution.

   Using the Flash RAM for the ZIL?
   

I'm not sure why you would want to do this.  Just carve off a
LUN or slice on the RAID card and use its NVRAM cache. 
A consumer class flash disk will be slower.

   Using zfs for general storage management?

   

cool.

 Has anyone built a similar system, what is the true path to success?
   

Success is at the summit, but there are several paths up the mountain.

 What are the pitfalls?
 What should I have on my reading list for starters?
   

Start with the ZFS system admin guide on opensolaris.org.
We try to keep the solarisinternals.com wikis up to date also.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OpenSolaris, ZFS and Hardware RAID, a recipe for success?

2008-02-09 Thread Jonathan Loran



Richard Elling wrote:

Nick wrote:
  


Using the RAID cards capability for RAID6 sounds attractive?
  



Assuming the card works well with Solaris, this sounds like a
reasonable solution.

  
Careful here.  If your workload is unpredictable, RAID 6 (and RAID 5) 
for that matter will break down under highly randomized write loads.  
There's a lot of trickery done with hardware RAID cards that can do some 
read-ahead caching magic, improving the read-paritycalc-paritycalc-write 
cycle, but you can't beat out the laws of physics.  If you do *know* 
you'll be streaming more than writing random small number of blocks, 
RAID 6 hardware can work.  But with transaction like loads, performance 
will suck. 


Jon

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss