Re: [zfs-discuss] marvell88sx2 driver build126

2009-11-10 Thread Tim Cook
On Mon, Nov 9, 2009 at 2:51 PM, Cindy Swearingen
cindy.swearin...@sun.comwrote:

 Hi,

 I can't find any bug-related issues with marvell88sx2 in b126.

 I looked over Dave Hollister's shoulder while he searched for
 marvell in his webrevs of this putback and nothing came up:

  driver change with build 126?
 not for the SATA framework, but for HBAs there is:
 http://hub.opensolaris.org/bin/view/Community+Group+on/2009093001

 I will find a thumper, load build 125, create a raidz pool, and
 upgrade to b126.

 I'll also send the error messages that Tim provided to someone who
 works in the driver group.

 Thanks,

 Cindy


I tried the build 125 driver and it didn't make a difference.  The odd part
I've just noticed is that it's port 4 on both cards that have been giving me
issues.  I guess it's possible it's just a coincidence/bad luck.

I've grabbed the b125 ISO from genunix and am going to try booting off the
livecd to see if it produces different results.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] marvell88sx2 driver build126

2009-11-10 Thread Orvar Korvar
Does this mean that there are no driver changes in marvell88sx2, between b125 
and b126? If no driver changes, then it means that we both had extreme unluck 
with our drives, because we both had checksum errors? And my discs were brand 
new. 

How probable is this? Something is weird here. What is your opinion on this? 
Should we agree that there was a hardware error, and it was just a coincidence?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS and oracle on SAN disks

2009-11-10 Thread Ian Garbutt
I believe the best practice is to use seperate disks/zpool for oracle database 
files as the record size needs to be set the same as the db block size - when 
using a jbod or internal disks.

If the server is using a large SAN LUN can anybody see any issues if there is 
only one zpool and the dataset is set to have the record size set rather than 
at the zpool level?

Thanks

Ian
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] FreeNAS 0.7 with zfs out

2009-11-10 Thread Eugen Leitl

Apparently went live on 6th November. This isn't FreeBSD 8.x zfs,
but at least raidz2 is there.

http://www.freenas.org/

FreeNAS 0.7 (Khasadar)
Sunday, 21 June 2009
Majors changes:

* Add ability to configure the login shell for a user.
* Upgrade Samba to 3.0.37.
* Upgrade transmission to 1.72.
* Local users must join the group 'ftp' to be able to login via FTP if 
'Local users only' in 'Services|FTP' is enabled.
* Upgrade lighttpd to 1.4.23.
* Add a user portal. This allows a local user to login and change it's 
password. The user must have access permissions to login to the user portal. 
This can be configured in the user configuration WebGUI. Please note that the 
administrator needs to apply changes done by the local users because there have 
to be restarted several services (which is not allowed to be done by a user).
* Upgrade ProFTPD to 1.3.2a.
* Upgrade iSCSI initiator to 2.2.3.
* Upgrade fusefs-ntfs/ntfs-3g to 2009.4.4.
* Announce AFP shares using Bonjour/ZeroConf (FR 2839592). Thanks to Morton 
Jonuschat.
* Add AFP FP_SYNCFORK command support (FR 2836955). Thanks to Morton 
Jonuschat for the patch.
* Upgrade e2fsprogs to 1.41.8.
* Add Adaptec AACRAID 32/64-bit driver to v5.2.0 Build 17517.
* Upgrade inadyn-mt to 02.14.10.
* Upgrade fuppes to SVN-0.640.


Minors changes:

* Set transmission umask to 0002 per default. This can be customize via the 
'Services|BitTorrent' WebGUI or the rc.conf variable 'transmission_umask' (FR 
2813791).
* Add ixgbe driver to i386 kernel.
* Add ixgb driver to AMD64 kernel (BR 2813759).
* Add support for Blowfish 448 bits encryption (FR 2816028).
* Add configuration option in 'Services|BitTorrent' to enable/disable usage 
of distributed hash table (DHT).

* Add /usr/bin/getopt command (FR 2824548).
* Add extra options for S.M.A.R.T. in 'Disks|Management|Disk|Edit' (FR 
2824730).
* Add RAID1 balance algorithm 'prefer' (FR 2833989).
* Add latvian language support. Thanks to the translators.
* Update Quixplorer russian translation (BR 2841900). Thanks to Alexey 
Sannikov.
* Add 'Max. login attempts' to 'Services|FTP' (FR 2844193).
* Get AFP dbd cnid scheme working (BR 2844900).
* Set 'dir-listing.encoding = utf-8' for the webserver directory listing 
(FR 2872624).
* Display volume serial number in 'Disks|Management' (FR 2881880).
* Now it is possible to configure iSCSI-Targets for export: removable media 
(static  dynamic size), pass-through devices. Thanks to Vasily Chalykh.


Bug fixes:

* Prohibit user 'transmission' to login via FTP.
* ZPool disk space usage isn't displayed correctly (BR 2810584).
* Improved Unison WebGUI to be able to configure ZFS shares as working 
directory (BR 2795084).
* Synchronizing ZFS configuration fails (BR 2814324).
* Restrict bittorrent administrative WebGUI port to [1024,65535] (BR 
2835342).
* The 'Unmount disk/partition' checkbox on 'Disks|Mount Point|Fsck' was 
ignored (BR 2860297).

 
FreeNAS 0.7RC1 (Sardaukar)
Sunday, 21 June 2009
Majors changes:

* Upgrade to FreeBSD 7.2.
* Include ZFS support. Thanks to Nelson Silva for doing core coding and 
Falk Menzel for testing and giving some tipps and ideas.
* Upgrade iSCSI initiator to 2.1.1.
* Replace iSCSI target by istgt. Thanks to Daisuke Aoyama for the WebGUI 
adaption. Please note, if you have used devices with the previous iSCSI target 
software you have to recreate your target.
* Add WOL patch for nVidia(nfe(4)) and 3Com(xl(4)). Thanks to Tobias Reber.
* Upgrade mt-daapd/firefly to svn-1696.
* Refactor 'Diagnostics|Log' WebGUI.
* Add kernel patches to get ARTiGO A2000 hardware working. Thanks to David 
Davis for the patches.
* Respect the modified log file location (via rc.conf for syslog, fuppes, 
mt-daapd, ...) in the WebGUI (FR 2778803/2791772).
* Upgrade transmission to 1.61. Add 'Watch directory' and 'Extra options' 
fields to 'Services|BitTorrent' WebGUI.
* Add entry 'FTP - Ban module' to the list of log files in 
'Diagnostics|Log' if the module is enabled (FR 2797652).
* Add 'iperf', a tool to measure maximum TCP and UDP bandwidth (FR 2785038).
* Add 'bsnmp-ucd' module that implements parts of UCD-SNMP-MIB.
* Add SNMP client tools: bsnmpget, bsnmpset and bsnmpwalk
* Add 'Auxiliary parameters' to 'Services|SNMP' that will be added to the 
end of the snmpd.config file.
* Upgrade e2fsprogs to 1.41.5.
* Upgrade rsync to 3.0.6.
* Upgrade tftp-hpa to 0.49.


Bug fixes:

* Hardening WebGUI to prevent cross-site request forgery attacks (JPCERT/CC 
JVN#15267895).

 
FreeNAS 0.69.2 (Muad'Dib)
Monday, 20 April 2009
Majors changes:

* Add another WOL patch. It is tested for nfe(4) und xl(4). Thanks to 
Tobias Reber.
* Add switch in 'System|Advanced' WebGUI to enable the console screensaver 
(FR 2777301).
* Upgrade Adaptec SCSI RAID administration tool to 

Re: [zfs-discuss] Finding SATA cards for ZFS; was Lundman home NAS

2009-11-10 Thread Robin Bowes
On 01/09/09 08:26, James Andrewartha wrote:
 Jorgen Lundman wrote:
 The mv8 is a marvell based chipset, and it appears there are no
 Solaris drivers for it.  There doesn't appear to be any movement from
 Sun or marvell to provide any either.

 Do you mean specifically Marvell 6480 drivers? I use both DAC-SATA-MV8
 and AOC-SAT2-MV8, which use Marvell MV88SX and works very well in
 Solaris. (Package SUNWmv88sx).
 
 They're PCI-X SATA cards, the AOC-SASLP-MV8 is a PCIe SAS card and has
 no (Open)Solaris driver.
 

Shame, I was just thinking that this was a nice looking card that could
replace my AOC-SAT2-MV8s.

Ah well...

R.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] marvell88sx2 driver build126

2009-11-10 Thread Cindy Swearingen

Hi Orvar,

Correct, I don't see any marvell8sx2 driver changes between b125-126.

So far, only you and Tim are reporting these issues.

Generally, we see bugs filed by the internal test teams if they see
similar problems.

I will try to reproduce the RAIDZ checksum errors separately from the
marvell88sx2 issue.

Thanks,

Cindy


On 11/10/09 02:25, Orvar Korvar wrote:
Does this mean that there are no driver changes in marvell88sx2, between b125 and b126? If no driver changes, then it means that we both had extreme unluck with our drives, because we both had checksum errors? And my discs were brand new. 


How probable is this? Something is weird here. What is your opinion on this? 
Should we agree that there was a hardware error, and it was just a coincidence?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Odd sparing problem

2009-11-10 Thread Tim Cook
So, I currently have a pool with 12 disks raid-z2 (12+2).  As you may have
seen in the other thread, I've been having on and off issues with b126
randomly dropping drives.  Well, I think after changing several cables, and
doing about 20 reboots plugging one drive in at a time (I only booted to the
marvell bios, not the whole way into the OS), I've gotten the marvell cards
to settle down.  The problem is, I'm now seeing this in a zpool output:

  pool: fserv
 state: ONLINE
 scrub: resilver completed after 0h0m with 0 errors on Tue Nov 10 09:15:12
2009
config:

NAME  STATE READ WRITE CKSUM
fserv ONLINE   0 0 0
  raidz2-0ONLINE   0 0 0
c8t0d0ONLINE   0 0 0
c8t1d0ONLINE   0 0 0
c8t2d0ONLINE   0 0 0
c8t3d0ONLINE   0 0 0
c8t4d0ONLINE   0 0 0
c8t5d0ONLINE   0 0 0
c7t0d0ONLINE   0 0 0
c7t1d0ONLINE   0 0 0
c7t2d0ONLINE   0 0 0
c7t3d0ONLINE   0 0 0
c7t4d0ONLINE   0 0 0
spare-11  ONLINE   0 0 5
  c7t5d0  ONLINE   0 0 0  30K resilvered
  c7t6d0  ONLINE   0 0 0
spares
  c7t6d0  INUSE currently in use


Anyone have any thoughts?  I'm trying to figure out how to get c7t6d0 back
to being a hotspare since c7t5d0 is installed, there, and happy.  It's
almost as if it's using both disks for spare-11 right now.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ..and now ZFS send dedupe

2009-11-10 Thread Roman Naumenko

James C. McPherson wrote, On 09-11-09 04:40 PM:

Roman Naumenko wrote:
  

Interesting stuff.

By the way, is there  a place to watch lated news like this on zfs/opensolaris?
rss maybe?


You could subscribe to onnv-not...@opensolaris.org...

James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
  

Thanks, James.

What is the subscription process? Just to send email?

--
Roman Naumenko
ro...@frontline.ca
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] RAID-Z and virtualization

2009-11-10 Thread Joe Auty
Toby Thain wrote:
 On 8-Nov-09, at 12:20 PM, Joe Auty wrote:

 Tim Cook wrote:
 On Sun, Nov 8, 2009 at 2:03 AM, besson3c j...@netmusician.org
 mailto:j...@netmusician.org wrote:

 ...


 Why not just convert the VM's to run in virtualbox and run Solaris
 directly on the hardware?


 That's another possibility, but it depends on how Virtualbox stacks
 up against VMWare Server. At this point a lot of planning would be
 necessary to switch to something else, although this is possibility.

 How would Virtualbox stack up against VMWare Server? Last I checked
 it doesn't have a remote console of any sort, which would be a deal
 breaker. Can I disable allocating virtual memory to Virtualbox VMs?
 Can I get my VMs to auto boot in a specific order at runlevel 3? Can
 I control my VMs via the command line? 

 Yes you certainly can. Works well, even for GUI based guests, as there
 is vm-level VRDP (VNC/Remote Desktop) access as well as whatever
 remote access the guest provides.



 I thought Virtualbox was GUI only, designed for Desktop use primarily? 

 Not at all. Read up on VBoxHeadless.


I take it that Virtualbox, being Qemu/KVM based will support 64 bit
versions of FreeBSD guests, unlike Xen based solutions?



 --Toby

 This switch will only make sense if all of this points to a net positive.



 --Tim


 -- 
 Joe Auty
 NetMusician: web publishing software for musicians
 http://www.netmusician.org
 j...@netmusician.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org mailto:zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


-- 
Joe Auty
NetMusician: web publishing software for musicians
http://www.netmusician.org
j...@netmusician.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS and oracle on SAN disks

2009-11-10 Thread Richard Elling

On Nov 10, 2009, at 5:32 AM, Ian Garbutt wrote:

I believe the best practice is to use seperate disks/zpool for  
oracle database files as the record size needs to be set the same as  
the db block size - when using a jbod or internal disks.


recordsize is not a pool property, it is a dataset (zvol or file  
system) property.
You can have different recordsize or volblocksize property settings  
for each
dataset in the pool.  You can also have different cache policies,  
which will

also be useful. Roch reports a 40% performance improvement by adjusting
the logbias property.
http://blogs.sun.com/roch/entry/synchronous_write_bias_property

However, in that case, a large SAN LUN was a hybrid storage pool.
If you are using a traditional RAID array, then the nonvolatile write  
cache

can also improve performance.
 -- richard



If the server is using a large SAN LUN can anybody see any issues if  
there is only one zpool and the dataset is set to have the record  
size set rather than at the zpool level?


Thanks

Ian
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] marvell88sx2 driver build126

2009-11-10 Thread Richard Elling


On Nov 10, 2009, at 1:25 AM, Orvar Korvar wrote:

Does this mean that there are no driver changes in marvell88sx2,  
between b125 and b126? If no driver changes, then it means that we  
both had extreme unluck with our drives, because we both had  
checksum errors? And my discs were brand new.


There are other drivers in the software stack that may have changed.
 -- richard



How probable is this? Something is weird here. What is your opinion  
on this? Should we agree that there was a hardware error, and it was  
just a coincidence?

--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + fsck

2009-11-10 Thread Joerg Moellenkamp
Hi, 

 *everybody* is interested in the flag days page. Including me.
 Asking me to raise the priority is not helpful.
 
 From my perspective, it's a surprise that 'everybody' is interested, as I'm
 not seeing a lot of people complaining that the flag day page is not updating.
 Only a couple of people on this list, and one of those is me!
 Perhaps I'm looking in the wrong places.

I used this page frequently, too. But now i'm just using the twitter account 
feeded by onnv-notify . You can look to it at  http://twitter.com/codenews 

Regards
 Joerg

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs inotify?

2009-11-10 Thread Jeremy Kitchen

On Nov 10, 2009, at 10:23 AM, Andrew Daugherity wrote:


For example:
rsync -avn --delete-before /export/ims/.zfs/snapshot/zfs-auto- 
snap_daily-2009-11-09-1900/ /export/ims/.zfs/snapshot/zfs-auto- 
snap_daily-2009-11-08-1900/


[...]



If you cared to see changes within files (I don't),


toss the -W flag on there then and it should run even faster,  
methinks, as it won't compare file contents at all.


-Jeremy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs inotify?

2009-11-10 Thread Andrew Daugherity
Thanks for info, although the audit system seems a lot more complex than what I 
need.  Would still be nice if they fixed bart to work on large filesystems, 
though.

Turns out the solution was right under my nose -- rsync in dry-run mode works 
quite well as a snapshot diff tool.  I'll share this with the list, in case 
it helps anyone else.

For example:
rsync -avn --delete-before 
/export/ims/.zfs/snapshot/zfs-auto-snap_daily-2009-11-09-1900/ 
/export/ims/.zfs/snapshot/zfs-auto-snap_daily-2009-11-08-1900/

This makes a list of what files have been changed, added, or deleted between 
these two snapshots, and runs in about 10 seconds.  If you cared to see changes 
within files (I don't), it would be trivial to add a loop along the lines of 
'$rsync_cmd | while read file; do diff $snap1/$file $snap2/file; done'.

Note the trailing slashes (otherwise rsync works one directory higher and 
considers the snapshot directory name, which we don't want) and that the newer 
snapshot is the source and the older snapshot, the destination.  I'm 
[ab]using rsync to have it tell me exactly how it would make the destination 
be a replica of the source.

FWIW, I'm using rsync 3.0.6 from opencsw.  Older rsync should work fine but may 
take longer to run.


-Andrew

 Richard Elling richard.ell...@gmail.com 11/9/2009 7:33 PM  
Seems to me that you really want auditing.  You can configure the audit
system to only record the events you are interested in.
http://docs.sun.com/app/docs/doc/816-4557/auditov-1?l=ena=view
   -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PSARC recover files?

2009-11-10 Thread BJ Quinn
Say I end up with a handful of unrecoverable bad blocks that just so happen to 
be referenced by ALL of my snapshots (in some file that's been around forever). 
 Say I don't care about the file or two in which the bad blocks exist.  Is 
there any way to purge those blocks from the pool (and all snapshots) without 
having to restore the whole pool from backup?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PSARC recover files?

2009-11-10 Thread Tim Cook
On Tue, Nov 10, 2009 at 2:40 PM, BJ Quinn bjqu...@seidal.com wrote:

 Say I end up with a handful of unrecoverable bad blocks that just so happen
 to be referenced by ALL of my snapshots (in some file that's been around
 forever).  Say I don't care about the file or two in which the bad blocks
 exist.  Is there any way to purge those blocks from the pool (and all
 snapshots) without having to restore the whole pool from backup?



No.  The whole point of a snapshot is to keep a consistent on-disk state
from a certain point in time.  I'm not entirely sure how you managed to
corrupt blocks that are part of an existing snapshot though, as they'd be
read-only.  The only way that should even be able to happen is if you took a
snapshot after the blocks were already corrupted.  Any new writes would be
allocated from new blocks.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PSARC recover files?

2009-11-10 Thread A Darren Dunham
On Tue, Nov 10, 2009 at 03:04:24PM -0600, Tim Cook wrote:
 No.  The whole point of a snapshot is to keep a consistent on-disk state
 from a certain point in time.  I'm not entirely sure how you managed to
 corrupt blocks that are part of an existing snapshot though, as they'd be
 read-only.

Physical corruption of the media
Something outside of ZFS diddling bits on storage

 The only way that should even be able to happen is if you took a
 snapshot after the blocks were already corrupted.  Any new writes would be
 allocated from new blocks.

It can be corrupted while it sits on disk.  Since it's read-only, you
can't force it to allocate anything and clean itself up.

-- 
Darren
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PSARC recover files?

2009-11-10 Thread Tim Cook
On Tue, Nov 10, 2009 at 3:19 PM, A Darren Dunham ddun...@taos.com wrote:

 On Tue, Nov 10, 2009 at 03:04:24PM -0600, Tim Cook wrote:
  No.  The whole point of a snapshot is to keep a consistent on-disk state
  from a certain point in time.  I'm not entirely sure how you managed to
  corrupt blocks that are part of an existing snapshot though, as they'd be
  read-only.

 Physical corruption of the media
 Something outside of ZFS diddling bits on storage

  The only way that should even be able to happen is if you took a
  snapshot after the blocks were already corrupted.  Any new writes would
 be
  allocated from new blocks.

 It can be corrupted while it sits on disk.  Since it's read-only, you
 can't force it to allocate anything and clean itself up.



You're telling me a scrub won't actively clean up corruption in snapshots?
That sounds absolutely absurd to me.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PSARC recover files?

2009-11-10 Thread Nicolas Williams
On Tue, Nov 10, 2009 at 03:33:22PM -0600, Tim Cook wrote:
 You're telling me a scrub won't actively clean up corruption in snapshots?
 That sounds absolutely absurd to me.

Depends on how much redundancy you have in your pool.  If you have no
mirrors, no RAID-Z, and no ditto blocks for data, well, you have no
redundancy, and ZFS won't be able to recover affected files.

Nico
-- 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Rebooting while Scrubing in snv_126

2009-11-10 Thread Francois Laagel

Greetings folks,

Something funny happened to my amd64 box last night. I shut it down while a
a scrub was running on rpool. This was not a fast reboot or anything like that.

Since then, the system does not come up any more. I still can boot in single
user mode but /sbin/zfs mount -va hangs while displaying:

Reading ZFS config: *

I went a bit further in single user mode: truss -f ksh 
/lib/svc/method/fs-local's
last sign of life is a memcntl, shortly after a ZFS_IOC_POOL_CONFIGS
successful ioctl.

Any word of wisdom here? I really hope this recoverable since this was my 
primary
system (no serial console though, so I can't really capture the essence of
the drama  :)  )

I think rpool was version 19 but I'm not 100% positive here.

Regards.

   Francois

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zle compression ?

2009-11-10 Thread roland
by some posting on zfs-fuse mailinglist, i came across zle compression which 
seems to be part of the dedupe-commit some days ago:

http://hg.genunix.org/onnv-gate.hg/diff/e2081f502306/usr/src/uts/common/fs/zfs/zle.c

--snipp
31 + * Zero-length encoding.  This is a fast and simple algorithm to eliminate
  32 + * runs of zeroes.  Each chunk of compressed data begins with a 
length byte, b.
  33 + * If b  n (where n is the compression parameter) then the next b + 
1 bytes
  34 + * are literal values.  If b = n then the next (256 - b + 1) bytes 
are zero.
--snipp

i`m curious - what does that mean?

does zfs have another compression scheme named zle now ? 
if yes, why ?

wasn´t zero-length encoding already there and just a builtin feature ?

maybe that builtin has now become an option ?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] This is the scrub that never ends...

2009-11-10 Thread Bill Sommerfeld
On Fri, 2009-09-11 at 13:51 -0400, Will Murnane wrote:
 On Thu, Sep 10, 2009 at 13:06, Will Murnane will.murn...@gmail.com wrote:
  On Wed, Sep 9, 2009 at 21:29, Bill Sommerfeld sommerf...@sun.com wrote:
  Any suggestions?
 
  Let it run for another day.
  I'll let it keep running as long as it wants this time.
  scrub: scrub completed after 42h32m with 0 errors on Thu Sep 10 17:20:19 2009
 
 And the people rejoiced.  So I guess the issue is more scrubs may
 report ETA very inaccurately than scrubs never finish.  Thanks for
 the suggestions and support.

One of my pools routinely does this -- the scrub gets to 100% after
about 50 hours but keeps going for another day or more after that.

It turns out that zpool reports number of blocks visited vs number of
blocks allocated, but clamps the ratio at 100%.

If there is substantial turnover in the pool, it appears you may end up
needing to visit more blocks than are actually allocated at any one
point in time.

I made a modified version of the zpool command and this is what it
prints for me:

...
 scrub: scrub in progress for 74h25m, 119.90% done, 0h0m to go
 5428197411840 blocks examined, 4527262118912 blocks allocated
...

This is the (trivial) source change I made to see what's going on under
the covers:

diff -r 12fb4fb507d6 usr/src/cmd/zpool/zpool_main.c
--- a/usr/src/cmd/zpool/zpool_main.cMon Oct 26 22:25:39 2009 -0700
+++ b/usr/src/cmd/zpool/zpool_main.cTue Nov 10 17:07:59 2009 -0500
@@ -2941,12 +2941,15 @@
 
if (examined == 0)
examined = 1;
-   if (examined  total)
-   total = examined;
 
fraction_done = (double)examined / total;
-   minutes_left = (uint64_t)((now - start) *
-   (1 - fraction_done) / fraction_done / 60);
+   if (fraction_done  1) {
+   minutes_left = (uint64_t)((now - start) *
+   (1 - fraction_done) / fraction_done / 60);
+   } else {
+   minutes_left = 0;
+   }
+
minutes_taken = (uint64_t)((now - start) / 60);
 
(void) printf(gettext(%s in progress for %lluh%um, %.2f%% done,

@@ -2954,6 +2957,9 @@
scrub_type, (u_longlong_t)(minutes_taken / 60),
(uint_t)(minutes_taken % 60), 100 * fraction_done,
(u_longlong_t)(minutes_left / 60), (uint_t)(minutes_left %
60));
+   (void) printf(gettext(\t %lld blocks examined, %lld blocks
allocated\n),
+   examined,
+   total);
 }
 
 static void

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Odd sparing problem

2009-11-10 Thread Cindy Swearingen

Hi Tim,

I'm not sure I understand this output completely, but have you
tried detaching the spare?

Cindy

On 11/10/09 09:21, Tim Cook wrote:
So, I currently have a pool with 12 disks raid-z2 (12+2).  As you may 
have seen in the other thread, I've been having on and off issues with 
b126 randomly dropping drives.  Well, I think after changing several 
cables, and doing about 20 reboots plugging one drive in at a time (I 
only booted to the marvell bios, not the whole way into the OS), I've 
gotten the marvell cards to settle down.  The problem is, I'm now seeing 
this in a zpool output:


  pool: fserv
 state: ONLINE
 scrub: resilver completed after 0h0m with 0 errors on Tue Nov 10 
09:15:12 2009

config:

NAME  STATE READ WRITE CKSUM
fserv ONLINE   0 0 0
  raidz2-0ONLINE   0 0 0
c8t0d0ONLINE   0 0 0
c8t1d0ONLINE   0 0 0
c8t2d0ONLINE   0 0 0
c8t3d0ONLINE   0 0 0
c8t4d0ONLINE   0 0 0
c8t5d0ONLINE   0 0 0
c7t0d0ONLINE   0 0 0
c7t1d0ONLINE   0 0 0
c7t2d0ONLINE   0 0 0
c7t3d0ONLINE   0 0 0
c7t4d0ONLINE   0 0 0
spare-11  ONLINE   0 0 5
  c7t5d0  ONLINE   0 0 0  30K resilvered
  c7t6d0  ONLINE   0 0 0
spares
  c7t6d0  INUSE currently in use


Anyone have any thoughts?  I'm trying to figure out how to get c7t6d0 
back to being a hotspare since c7t5d0 is installed, there, and happy.  
It's almost as if it's using both disks for spare-11 right now.


--Tim




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Odd sparing problem

2009-11-10 Thread Tim Cook
On Tue, Nov 10, 2009 at 4:38 PM, Cindy Swearingen
cindy.swearin...@sun.comwrote:

 Hi Tim,

 I'm not sure I understand this output completely, but have you
 tried detaching the spare?

 Cindy


Hey Cindy,

Detaching did in fact solve the issue.  During my previous issues when the
spare kicked in, it actually automatically detached itself once I replaced
the failed drive, so I didn't understand what was going on this time around.

Thanks!
--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] marvell88sx2 driver build126

2009-11-10 Thread Tim Cook
On Tue, Nov 10, 2009 at 10:55 AM, Richard Elling
richard.ell...@gmail.comwrote:


 On Nov 10, 2009, at 1:25 AM, Orvar Korvar wrote:

  Does this mean that there are no driver changes in marvell88sx2, between
 b125 and b126? If no driver changes, then it means that we both had extreme
 unluck with our drives, because we both had checksum errors? And my discs
 were brand new.


 There are other drivers in the software stack that may have changed.
  -- richard



 How probable is this? Something is weird here. What is your opinion on
 this? Should we agree that there was a hardware error, and it was just a
 coincidence?



So... I do appear to have reached somewhat of a truce with the system and
b126 at the moment.  I'm now going through and replacing the last of my old
maxtor 300GB drives with brand new hitachi 1TB drives.  One thing I'm
noticing is a lot of checksum errors being generated during the resilver.
Is this normal?  Furthermore, since I see no known data errors, is it safe
to assume it's all being corrected, and I'm not losing any data?  I still do
have a separate copy of this data on a box at work that should be completely
consistent... but I will need to re-purpose that storage soon, and will be
without a known good backup for a while (I know, I know).  I'd rather do a
fresh zfs send/receive than find out 6 months from now I lost something.

  pool: fserv
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
 scrub: resilver in progress for 0h8m, 0.89% done, 15h14m to go
config:

NAMESTATE READ WRITE CKSUM
fserv   DEGRADED 0 0 0
  raidz2-0  DEGRADED 0 0 0
c8t0d0  ONLINE   0 0 0
c8t1d0  ONLINE   0 0 0
c8t2d0  ONLINE   0 0 0
c8t3d0  ONLINE   0 0 0
c8t4d0  ONLINE   0 0 0
c8t5d0  ONLINE   0 0 0
c7t0d0  ONLINE   0 0 0
c7t1d0  ONLINE   0 0 0
c7t2d0  ONLINE   0 0 0
replacing-9 DEGRADED 0 0  161K
  14274451003165180679  FAULTED  0 0 0  was
/dev/dsk/c7t3d0s0/old
  c7t3d0ONLINE   0 0 0  2.05G
resilvered
c7t4d0  ONLINE   0 0 0
c7t5d0  ONLINE   0 0 0
spares
  c7t6d0AVAIL

errors: No known data errors


--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PSARC recover files?

2009-11-10 Thread BJ Quinn
I believe it was physical corruption of the media.  Strange thing is last time 
it happened to me it also managed to replicate the bad blocks over to my backup 
server replicated with SNDR...

And yes, it IS read only, and a scrub will NOT actively clean up corruption in 
snapshots.  It will DETECT corruption, but if it's unrecoverable, that's that.  
It's unrecoverable.  If there's not enough redundancy in the pool, I'm ok with 
the data not being recoverable.  But wouldn't there be a way to purge out the 
bad blocks if for example it was only in a single bad file out of millions of 
files, and I didn't care about the file in question?  I don't want to recover 
the file, I want to have a working version of my pool+snapshots minus the tiny 
bit that was obviously corrupt.

Barring another solution, I'd have to take the pool in question, delete the bad 
file, and delete ALL the snapshots.  Then restore the old snapshots from backup 
to another pool, and copy over the current data from the pool that had a 
problem over to the new pool.  I can get most of my snapshots back that way, 
with the best known current data sitting on top as the active data set.  
Problem is with hundreds of snapshots plus compression, zfs send/recv takes 
over 24 hours to restore a full backup like that to a new storage device.  Last 
time this happened to me, I just had to say goodbye to all my snapshots and 
deal with it, all over a couple of kilobytes of temp files.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Fwd: [ilugb] Does ZFS support Hole Punching/Discard

2009-11-10 Thread George Janczuk
I've been following the use of SSD with ZFS and HSPs for some time now, and I 
am working (in an architectural capacity) with one of our IT guys to set up our 
own ZFS HSP (using a J4200 connected to an X2270).

The best practice seems to be to use an Intel X25-M for the L2ARC (Readzilla) 
and an Intel X25-E for the ZIL/SLOG (Logzilla).

However, whilst being a BIG thing in the Windows 7 world - I have pretty much 
heard nothing about Intel's G2 devices and updated firmware when Intel's SSDs 
are used in a ZFS HSP. In particular, does ZFS use or support the TRIM command? 
Is it even relevant or useful in a hierarchical (vs. primary) storage context?

Any comment would be appreciated. Some comment from the Fishworks guys in 
particular would be great!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs eradication

2009-11-10 Thread Brian Kolaci

Hi,

I was discussing the common practice of disk eradication used by many firms for 
security.  I was thinking this may be a useful feature of ZFS to have an option 
to eradicate data as its removed, meaning after the last reference/snapshot is 
done and a block is freed, then write the eradication patterns back to the 
removed blocks.

By any chance, has this been discussed or considered before?

Thanks,

Brian
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Fwd: [ilugb] Does ZFS support Hole Punching/Discard

2009-11-10 Thread Tim Cook
On Tue, Nov 10, 2009 at 6:51 PM, George Janczuk 
geor...@objectconsulting.com.au wrote:

 I've been following the use of SSD with ZFS and HSPs for some time now, and
 I am working (in an architectural capacity) with one of our IT guys to set
 up our own ZFS HSP (using a J4200 connected to an X2270).

 The best practice seems to be to use an Intel X25-M for the L2ARC
 (Readzilla) and an Intel X25-E for the ZIL/SLOG (Logzilla).

 However, whilst being a BIG thing in the Windows 7 world - I have pretty
 much heard nothing about Intel's G2 devices and updated firmware when
 Intel's SSDs are used in a ZFS HSP. In particular, does ZFS use or support
 the TRIM command? Is it even relevant or useful in a hierarchical (vs.
 primary) storage context?

 Any comment would be appreciated. Some comment from the Fishworks guys in
 particular would be great!



My personal thought would be that it doesn't really make sense to even have
it, at least for readzilla.  In theory, you always want the SSD to be full,
or nearly full, as it's a cache.  The whole point of TRIM, from my
understanding, is to speed up the drive by zeroing out unused blocks so they
next time you try to write to them, they don't have to be cleared, then
written to.  When dealing with a cache, there shouldn't (again in theory) be
any free blocks, a warmed cache should be full of data.

Logzilla is kind of in the same boat, it should constantly be filling and
emptying as new data comes in.  I'd imagine the TRIM would just add
unnecessary overhead.  It could in theory help there by zeroing out blocks
ahead of time before a new batch of writes come in if you have a period of
little I/O.  My thought is it would be far more work than it's worth, but
I'll let the coders decide that one.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs eradication

2009-11-10 Thread Mark A. Carlson

Typically this is called Sanitization and could be done as part of
an evacuation of data from the disk in preparation for removal.

You would want to specify the patterns to write and the number of
passes.

-- mark

Brian Kolaci wrote:

Hi,

I was discussing the common practice of disk eradication used by many 
firms for security.  I was thinking this may be a useful feature of 
ZFS to have an option to eradicate data as its removed, meaning after 
the last reference/snapshot is done and a block is freed, then write 
the eradication patterns back to the removed blocks.


By any chance, has this been discussed or considered before?

Thanks,

Brian
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--
http://www.sun.com  * Mark A. Carlson *
Sr. Architect

*Systems Group*
Phone x69559 / 303-223-6139
Email mark.carl...@sun.com



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Zpool hosed during testing

2009-11-10 Thread Ron Mexico
This didn't occur on a production server, but I thought I'd post this anyway 
because it might be interesting.

I'm currently testing a ZFS NAS machine consisting of a Dell R710 with two Dell 
5/E SAS HBAs. Right now I'm in the middle of torture testing the system, 
simulating drive failures, exporting the storage pool, rearranging the disks in 
different slots, and what have you. Up until now, everything has been going 
swimmingly.

Here was my original zpool configuration: 

NAME STATE READ WRITE CKSUM
tank ONLINE   0 0 0
  raidz2 ONLINE   0 0 0
c1t1d0   ONLINE   0 0 0
c1t2d0   ONLINE   0 0 0
c1t3d0   ONLINE   0 0 0
c1t4d0   ONLINE   0 0 0
c1t5d0   ONLINE   0 0 0
c1t6d0   ONLINE   0 0 0
c1t7d0   ONLINE   0 0 0
c1t8d0   ONLINE   0 0 0
c1t9d0   ONLINE   0 0 0
c1t10d0  ONLINE   0 0 0
c1t11d0  ONLINE   0 0 0
c1t12d0  ONLINE   0 0 0
  raidz2 ONLINE   0 0 0
c2t25d0  ONLINE   0 0 0
c2t26d0  ONLINE   0 0 0
c2t27d0  ONLINE   0 0 0
c2t28d0  ONLINE   0 0 0
c2t29d0  ONLINE   0 0 0
c2t30d0  ONLINE   0 0 0
c2t31d0  ONLINE   0 0 0
c2t32d0  ONLINE   0 0 0
c2t33d0  ONLINE   0 0 0
c2t34d0  ONLINE   0 0 0
c2t35d0  ONLINE   0 0 0
c2t36d0  ONLINE   0 0 0

I exported the tank zpool, and rearranged drives in the chassis and reimported 
it - I ended up with this:

NAME STATE READ WRITE CKSUM
tank ONLINE   0 0 0
  raidz2 ONLINE   0 0 0
c2t31d0  ONLINE   0 0 0
c1t2d0   ONLINE   0 0 0
c1t3d0   ONLINE   0 0 0
c1t1d0   ONLINE   0 0 0
c1t12d0  ONLINE   0 0 0
c1t6d0   ONLINE   0 0 0
c1t7d0   ONLINE   0 0 0
c1t8d0   ONLINE   0 0 0
c1t9d0   ONLINE   0 0 0
c1t5d0   ONLINE   0 0 0
c1t11d0  ONLINE   0 0 0
c2t25d0  ONLINE   0 0 0
  raidz2 ONLINE   0 0 0
c1t4d0   ONLINE   0 0 0
c2t26d0  ONLINE   0 0 0
c2t27d0  ONLINE   0 0 0
c2t28d0  ONLINE   0 0 0
c2t29d0  ONLINE   0 0 0
c2t30d0  ONLINE   0 0 0
c1t10d0  ONLINE   0 0 0
c2t32d0  ONLINE   0 0 0
c2t33d0  ONLINE   0 0 0
c2t34d0  ONLINE   0 0 0
c2t35d0  ONLINE   0 0 0
c2t48d0  ONLINE   0 0 0

Great. No problems.

Next, I took c2t48d0 offline and then unconfigured it with cfgadm.

# zpool offline tank c2t48d0
# cfgadm -c unconfigure c2::dsk/c2t48d0

I checked the status next.

# zpool status tank
 pool: tank
state: DEGRADED
   status: One or more devices has been taken offline by the administrator.
  Sufficient replicas exist for the pool to continue functioning in 
a
  degraded state.
action: Online the device using 'zpool online' or replace the device with
'zpool replace'.
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM
tank DEGRADED 0 0 0
  raidz2 ONLINE   0 0 0
c2t31d0  ONLINE   0 0 0
c1t2d0   ONLINE   0 0 0
c1t3d0   ONLINE   0 0 0
c1t1d0   ONLINE   0 0 0
c1t12d0  ONLINE   0 0 0
c1t6d0   ONLINE   0 0 0
c1t7d0   ONLINE   0 0 0
c1t8d0   ONLINE   0 0 0
c1t9d0   ONLINE   0 0 0
c1t5d0   ONLINE   0 0 0
c1t11d0  ONLINE   0 0 0
c2t25d0  ONLINE   0 0 0
  raidz2 DEGRADED 0 0 0
c1t4d0   ONLINE   0 0 0
c2t26d0  ONLINE   0 0 0
c2t27d0  ONLINE   0 0 0
c2t28d0  ONLINE   0 0 0
c2t29d0  ONLINE   0 0 0
c2t30d0  ONLINE   0 0 0
c1t10d0  ONLINE   0 0 0
c2t32d0  ONLINE   0 0 0

Re: [zfs-discuss] zfs eradication

2009-11-10 Thread David Magda

On Nov 10, 2009, at 20:55, Mark A. Carlson wrote:


Typically this is called Sanitization and could be done as part of
an evacuation of data from the disk in preparation for removal.

You would want to specify the patterns to write and the number of
passes.


See also remanence:

http://en.wikipedia.org/wiki/Data_remanence

(S)ATA actually has a protocol command (secure erase) that will  
cause the disk to over write all of its sectors, and not be usable  
until its done. This doesn't exist in SCSI / SAS / FC as far as I know.


Generally speaking one over write is sufficient to prevent data from  
being accessible, but various government standards specify anywhere  
between one and four passes:


http://en.wikipedia.org/wiki/Data_erasure

Degaussing or complete destruction is usually necessary for the top  
secret stuff. DBAN is a useful (open-source) utility that I tend to  
recommend for regular folk:


http://www.dban.org/

While it could be useful, there are penalties in various jurisdictions  
for leaking data (especially with government-related stuff), so I'm  
not sure if Sun would want to potentially expose itself to  
inappropriate use that doesn't clean everything properly.


With ZFS encryption coming up, it could be sufficient to have your  
data sets encrypted and then simply forget the key. The data is  
still technically there, but (theoretically) completely inaccessible.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs eradication

2009-11-10 Thread Trevor Pretty





Excuse me for mentioning it but why not just use the format command?


format(1M)

  analyze
  
Run read, write, compare tests, and data
purge. The data purge
function implements the National Computer Security Center Guide to
Understanding Data Remnance (NCSC-TG-025 version 2) Overwriting
Algorithm. See NOTES.
  
  
The NCSC-TG-025 algorithm for overwriting
meets the DoD 5200.28-M (ADP
Security Manual) Eraser Procedures specification. The NIST Guidelines
for Media Sanitization (NIST SP 800-88)
also reference this algorithm..
  


And if the disk is buggered (a very technical term). A great big hammer!


Mark A. Carlson wrote:

  
  Typically this is called "Sanitization" and could be
done as part of 
an evacuation of data from the disk in preparation for removal.
  
You would want to specify the patterns to write and the number of
passes.
  
-- mark
  
Brian Kolaci wrote:
  Hi, 

I was discussing the common practice of disk eradication used by many
firms for security. I was thinking this may be a useful feature of ZFS
to have an option to eradicate data as its removed, meaning after the
last reference/snapshot is done and a block is freed, then write the
eradication patterns back to the removed blocks. 

By any chance, has this been discussed or considered before? 

Thanks, 

Brian 
___ 
zfs-discuss mailing list 
zfs-discuss@opensolaris.org

http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

  
  
  -- 
  
  

  

 Mark A. Carlson 
Sr. Architect

Systems Group
Phone x69559 / 303-223-6139
Email mark.carl...@sun.com


  

  
  
  
  




www.eagle.co.nz
This email is confidential and may be legally 
privileged. If received in error please destroy and immediately notify 
us.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] CIFS crashes when accessed with Adobe Photoshop Elements 6.0 via Vista

2009-11-10 Thread scott smallie
upgrade to the latest dev release fixed the problem for me.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] marvell88sx2 driver build126

2009-11-10 Thread Tim Cook
On Tue, Nov 10, 2009 at 5:15 PM, Tim Cook t...@cook.ms wrote:



 On Tue, Nov 10, 2009 at 10:55 AM, Richard Elling richard.ell...@gmail.com
  wrote:


 On Nov 10, 2009, at 1:25 AM, Orvar Korvar wrote:

  Does this mean that there are no driver changes in marvell88sx2, between
 b125 and b126? If no driver changes, then it means that we both had extreme
 unluck with our drives, because we both had checksum errors? And my discs
 were brand new.


 There are other drivers in the software stack that may have changed.
  -- richard



 How probable is this? Something is weird here. What is your opinion on
 this? Should we agree that there was a hardware error, and it was just a
 coincidence?



 So... I do appear to have reached somewhat of a truce with the system and
 b126 at the moment.  I'm now going through and replacing the last of my old
 maxtor 300GB drives with brand new hitachi 1TB drives.  One thing I'm
 noticing is a lot of checksum errors being generated during the resilver.
 Is this normal?  Furthermore, since I see no known data errors, is it safe
 to assume it's all being corrected, and I'm not losing any data?  I still do
 have a separate copy of this data on a box at work that should be completely
 consistent... but I will need to re-purpose that storage soon, and will be
 without a known good backup for a while (I know, I know).  I'd rather do a
 fresh zfs send/receive than find out 6 months from now I lost something.

   pool: fserv
  state: DEGRADED
 status: One or more devices is currently being resilvered.  The pool will
 continue to function, possibly in a degraded state.
 action: Wait for the resilver to complete.
  scrub: resilver in progress for 0h8m, 0.89% done, 15h14m to go

 config:

 NAMESTATE READ WRITE CKSUM
 fserv   DEGRADED 0 0 0
   raidz2-0  DEGRADED 0 0 0
 c8t0d0  ONLINE   0 0 0
 c8t1d0  ONLINE   0 0 0
 c8t2d0  ONLINE   0 0 0

 c8t3d0  ONLINE   0 0 0
 c8t4d0  ONLINE   0 0 0
 c8t5d0  ONLINE   0 0 0
 c7t0d0  ONLINE   0 0 0
 c7t1d0  ONLINE   0 0 0
 c7t2d0  ONLINE   0 0 0
 replacing-9 DEGRADED 0 0  161K
   14274451003165180679  FAULTED  0 0 0  was
 /dev/dsk/c7t3d0s0/old
   c7t3d0ONLINE   0 0 0  2.05G
 resilvered

 c7t4d0  ONLINE   0 0 0
 c7t5d0  ONLINE   0 0 0
 spares
   c7t6d0AVAIL


 errors: No known data errors


 --Tim



Anyone?  It's up to 7.35M checksum errors and it's rebuilding extremely
slowly (as evidenced by the 10 hour time).  The errors are only showing on
the replacing-9 line, not the individual drive.


  pool: fserv
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
 scrub: resilver in progress for 6h56m, 39.61% done, 10h34m to go
config:

NAMESTATE READ WRITE CKSUM
fserv   DEGRADED 0 0 0
  raidz2-0  DEGRADED 0 0 0
c8t0d0  ONLINE   0 0 0
c8t1d0  ONLINE   0 0 0
c8t2d0  ONLINE   0 0 0
c8t3d0  ONLINE   0 0 0
c8t4d0  ONLINE   0 0 0
c8t5d0  ONLINE   0 0 0
c7t0d0  ONLINE   0 0 0
c7t1d0  ONLINE   0 0 0
c7t2d0  ONLINE   0 0 0
replacing-9 DEGRADED 0 0 7.35M
  14274451003165180679  FAULTED  0 0 0  was
/dev/dsk/c7t3d0s0/old
  c7t3d0ONLINE   0 0 0  91.9G
resilvered
c7t4d0  ONLINE   0 0 0
c7t5d0  ONLINE   0 0 0
spares
  c7t6d0AVAIL

errors: No known data errors



--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss