[zfs-discuss] All drives intact but vdev UNAVAIL in raidz1

2011-09-06 Thread Tyler Benster
Hi all,

I'm running FreeNAS 7 (based on FreeBSD 7.2) and using ZFS v.13. I had two 2TB 
drives and one 1TB drive in a raidz1. I recently bought another 2TB drive to 
replace the 1TB drive. Unfortunately, I ran into some serious issues which are 
described in the steps that I did below:

First, I scrubbed the disks and found no errors. Next I turned off the 
computer, unplugged the 1TB drive and plugged in the new 2TB drive into the 
same SATA port. I turned on the machine, and executed 'zpool status'. The 
system hung and ignored all interrupt and kill signals, and terminated any 
shutdown sequence, so I hard-shutdown the machine. Next I unplugged the new 2TB 
drive and plugged in the old 1TB drive. When I enter 'zpool status' I see that 
the pool is UNAVAIL and the vdev is CORRUPTED, and all three drives are online. 
I tried to export  import the pool, but was unable to import as there 'are not 
enough redundant copies of the data.'

It seems quite likely that all of the data is intact, and that something 
different is preventing me from accessing the pool. What can I do to recover 
the pool? I have downloaded the Solaris 11 express livecd if that would be of 
any use.

Thanks in advance,
Tyler
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] All drives intact but vdev UNAVAIL in raidz1

2011-09-06 Thread Mark J Musante

On Tue, 6 Sep 2011, Tyler Benster wrote:

It seems quite likely that all of the data is intact, and that something 
different is preventing me from accessing the pool. What can I do to 
recover the pool? I have downloaded the Solaris 11 express livecd if 
that would be of any use.


Try running zdb -l on the disk and see if the labels are still there. 
Also, could you show us the output of 'zpools status'? Normally zfs would 
not hang if one disk of a raidz group is missing, but it might do that if 
one toplevel is missing.


If the zdb command shows all four labels to be correct, then you can try a 
zpool scrub and see if that resilvers the data for you.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] BAD WD drives - defective by design?

2011-09-06 Thread Richard Elling
On Aug 29, 2011, at 2:07 PM, Roy Sigurd Karlsbakk wrote:

 Hi all
 
 It seems recent WD drives that aren't Raid edition can cause rather a lot 
 of problems on RAID systems. We have a few machines with LSI controllers 
 (6801/6081/9201) and we're seeing massive errors occuring. The usual pattern 
 is a drive failing or even a resilver/scrub starting and then, suddenly, most 
 drives on the whole backplane report errors. These are usually No Device (as 
 reported by iostat -En), but the result is that we may see data corruption at 
 the end. We also have a system setup with Hitachi Deskstars, which has been 
 running for almost a year without issues. One system with a mixture of WD 
 Blacks and greens showed the same errors as described, but has been working 
 well after the WD drives were replaced by deskstars.

Sounds familiar.

 
 Now, it seems WD has changed their firmware to inhibit people from using them 
 for other things than toys (read: PCs etc). Since we've seen this issue on 
 different controllers and different drives, and can't reproduce it with 
 Hitachi Deskstars, I would guess the firmware upgrade from WD is the issue.

Likely.

 
 Would it be possible to fix this in ZFS somehow?

No, the error is 1-2 layers below ZFS.

 The drives seem to work well except for those No Device errors….

'nuff said.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] BAD WD drives - defective by design?

2011-09-06 Thread John Martin

http://wdc.custhelp.com/app/answers/detail/a_id/1397/~/difference-between-desktop-edition-and-raid-%28enterprise%29-edition-drives
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs send and dedupe

2011-09-06 Thread Freddie Cash
Just curious if anyone has looked into the relationship between zpool
dedupe, zfs zend dedupe, memory use, and network throughput.

For example, does 'zfs send -D' use the same DDT as the pool? Or does it
require more memory for it's own DDT, thus impacting performance of both?

If you have a deduped pool on both ends of the send, does -D make any
difference?

If neither pool is deduped, does -D make a difference?

We're waiting on a replacement backplane for our newest zfs-based storage
box, so won't be able to look into this ourselves until next week at the
earliest. Thought i'd check if anyone else has already done some comparisons
or benchmarks.

Cheers,
Freddie
fjwc...@gmail.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send and dedupe

2011-09-06 Thread Richard Elling
On Sep 6, 2011, at 9:01 PM, Freddie Cash wrote:
 Just curious if anyone has looked into the relationship between zpool dedupe, 
 zfs zend dedupe, memory use, and network throughput.
 

Yes.

 For example, does 'zfs send -D' use the same DDT as the pool?
 

No.

 Or does it require more memory for it's own DDT, thus impacting performance 
 of both?
 

Yes, no.

 If you have a deduped pool on both ends of the send, does -D make any 
 difference?
 

Yes, if the data is deduplicable.

 If neither pool is deduped, does -D make a difference?
 

Yes, if the data is deduplicable.

 We're waiting on a replacement backplane for our newest zfs-based storage 
 box, so won't be able to look into this ourselves until next week at the 
 earliest. Thought i'd check if anyone else has already done some comparisons 
 or benchmarks.
 

I'm not aware of any benchmarks, and I'd be surprised if they could be applied 
to real-world
cases. zfs send deduplication is very, very, very dependent on the data being 
sent. It is also
dependent on the release, since it is broken in many OpenSolaris and derived 
builds. Fixes
have recently been submitted into the illumos source tree. Recent Nexenta 
distributions also
have the fixes.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send and dedupe

2011-09-06 Thread Daniel Carosone
On Tue, Sep 06, 2011 at 10:05:54PM -0700, Richard Elling wrote:
 On Sep 6, 2011, at 9:01 PM, Freddie Cash wrote:
 
  For example, does 'zfs send -D' use the same DDT as the pool?
 
 No.

My understanding was that 'zfs send -D' would use the pool's DDT in 
building its own, if present. If blocks were known by the filesystem
to be duplicate, it would use that knowledge to skip some work seeding
its own ddt and stream back-references. This doesn't change the stream
contents vs what it would have generated without these hints, so No
still works as a short answer :) 

That understanding was based on discussions and blog posts at the
time, not looking at code. At least in theory, it should help avoid
reading and checksumming extra data blocks if this knowledge can be
used, so less work regardless of measurable impact on send throughput.
(It's more about diminished impact to other concurrent activities)

The point has mostly been moot in practice, though, because I've found
zfs send -D just plain doesn't work and often generates invalid
streams, as you note. Good to know there are fixes.

--
Dan.

pgpjZ1t9mVCs0.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss