Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-18 Thread martyscholes
Well said. -original message- Subject: Re: [zfs-discuss] Proposed idea for enhancement - damage control From: Bob Friesenhahn bfrie...@simple.dallas.tx.us Date: 02/17/2010 11:10 On Wed, 17 Feb 2010, Marty Scholes wrote: Bob, the vast majority of your post I agree with. At the same time, I

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-18 Thread Miles Nordin
dc == Daniel Carosone d...@geek.com.au writes: dc single-disk laptops are a pretty common use-case. It does not help this case. It helps the case where a single laptop disk fails and you recover it with dd conv=noerror,sync. This case is uncommon because few people know how to do it, or

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-18 Thread Adam Leventhal
Hey Bob, My own conclusions (supported by Adam Leventhal's excellent paper) are that - maximum device size should be constrained based on its time to resilver. - devices are growing too large and it is about time to transition to the next smaller physical size. I don't disagree with

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Christo Kutrovsky
Dan, loose was a typo. I meant lose. Interesting how a typo (write error) can cause a lot of confusion on what exactly I mean :) Resulting in corrupted interpretation. Note that my idea/proposal is targeted for a growing number of home users. To those, value for money usually is a much more

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Bob Friesenhahn
On Wed, 17 Feb 2010, Daniel Carosone wrote: These small numbers just tell you to be more worried about defending against the other stuff. Let's not forget that the most common cause of data loss is human error! Bob -- Bob Friesenhahn bfrie...@simple.dallas.tx.us,

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Bob Friesenhahn
On Wed, 17 Feb 2010, Marty Scholes wrote: Bob, the vast majority of your post I agree with. At the same time, I might disagree with a couple of things. I don't really care how long a resilver takes (hours, days, months) given a couple things: * Sufficient protection exists on the degraded

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Miles Nordin
ck == Christo Kutrovsky kutrov...@pythian.com writes: ck I could always put copies=2 (or more) to my important ck datasets and take some risk and tolerate such a failure. copies=2 has proven to be mostly useless in practice. If there were a real-world device that tended to randomly

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Frank Middleton
On 02/17/10 02:38 PM, Miles Nordin wrote: copies=2 has proven to be mostly useless in practice. Not true. Take an ancient PC with a mirrored root pool, no bus error checking and non-ECC memory, that flawlessly passes every known diagnostic (SMC included). Reboot with copies=1 and the same

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread David Magda
On Feb 17, 2010, at 12:34, Richard Elling wrote: I'm not sure how to connect those into the system (USB 3?), but when you build it, let us know how it works out. FireWire 3200 preferably. Anyone know if USB 3 sucks as much CPU as previous versions? If I'm burning CPU on I/O I'd rather

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Casper . Dik
If there were a real-world device that tended to randomly flip bits, or randomly replace swaths of LBA's with zeroes, but otherwise behave normally (not return any errors, not slow down retrying reads, not fail to attach), then copies=2 would be really valuable, but so far it seems no such

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Casper . Dik
If there were a real-world device that tended to randomly flip bits, or randomly replace swaths of LBA's with zeroes, but otherwise behave normally (not return any errors, not slow down retrying reads, not fail to attach), then copies=2 would be really valuable, but so far it seems no such

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Daniel Carosone
On Wed, Feb 17, 2010 at 02:38:04PM -0500, Miles Nordin wrote: copies=2 has proven to be mostly useless in practice. I disagree. Perhaps my cases fit under the weasel-word mostly, but single-disk laptops are a pretty common use-case. If there were a real-world device that tended to randomly

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-17 Thread Christo Kutrovsky
Dan, Exactly what I meant. An allocation policy, that will help in distributing the data in a way that when one disk is lost (entire mirror) than some data remains fully accessible as opposed to not been able to access pieces all over the storage pool. -- This message posted from

[zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Christo Kutrovsky
Just finished reading the following excellent post: http://queue.acm.org/detail.cfm?id=1670144 And started thinking what would be the best long term setup for a home server, given limited number of disk slots (say 10). I considered something like simply do a 2way mirror. What are the chances

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread James Dickens
On Tue, Feb 16, 2010 at 6:47 PM, Christo Kutrovsky kutrov...@pythian.comwrote: Just finished reading the following excellent post: http://queue.acm.org/detail.cfm?id=1670144 And started thinking what would be the best long term setup for a home server, given limited number of disk slots

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Christo Kutrovsky
Thanks for your feedback James, but that's not the direction where I wanted this discussion to go. The goal was not how to create a better solution for an enterprise. The goal was to do damage control in a disk failure scenario involving data loss. Back to the original question/idea. Which

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Bob Friesenhahn
On Tue, 16 Feb 2010, Christo Kutrovsky wrote: Just finished reading the following excellent post: http://queue.acm.org/detail.cfm?id=1670144 A nice article, even if I don't agree with all of its surmises and conclusions. :-) In fact, I would reach a different conclusion. I considered

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Bob Friesenhahn
On Tue, 16 Feb 2010, Christo Kutrovsky wrote: The goal was to do damage control in a disk failure scenario involving data loss. Back to the original question/idea. Which would you prefer, loose a couple of datasets, or loose a little bit of every file in every dataset. This ignores the

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Richard Elling
On Feb 16, 2010, at 4:47 PM, Christo Kutrovsky wrote: Just finished reading the following excellent post: http://queue.acm.org/detail.cfm?id=1670144 And started thinking what would be the best long term setup for a home server, given limited number of disk slots (say 10). I considered

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Daniel Carosone
On Tue, Feb 16, 2010 at 06:28:05PM -0800, Richard Elling wrote: The problem is that MTBF measurements are only one part of the picture. Murphy's Law says something will go wrong, so also plan on backups. +n Imagine this scenario: You lost 2 disks, and unfortunately you lost the 2 sides of

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Daniel Carosone
On Tue, Feb 16, 2010 at 04:47:11PM -0800, Christo Kutrovsky wrote: One of the ideas that sparkled is have a max devices property for each data set, and limit how many mirrored devices a given data set can be spread on. I mean if you don't need the performance, you can limit (minimize) the

Re: [zfs-discuss] Proposed idea for enhancement - damage control

2010-02-16 Thread Christo Kutrovsky
Bob, Using a separate pool would dictate other limitations, such as not been able to use more space than what's allocated in the pool. You could add space as needed, but you can't remove (move) devices freely. By using a shared pool with a hint of desired vdev/space allocation policy, you