Re: [zfs-discuss] Archiving and Restoring Snapshots
I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Hey folk, The are two problems with RAID-Z in builds snv_120 through snv_123 that will both be resolved in build snv_124. The problems are as follows: 1. Data corruption on a RAID-Z system of any sort (raidz1, raidz2, raidz3) can lead to spurious checksum errors being reported on devices that were not used as part of the reconstruction. These errors are harmless and can be cleared safely (zpool clear pool). 2. There is a far more serious problem with single-parity RAID-Z that can lead to data corruption. This data corruption is recoverable as long as no additional data corruption or drive failure occurs. That is to say, data is fine provided there is not an additional problem. The problem is present on all raidz1 configurations that use an odd number of children (disks) e.g. 4+1, or 6+1. Note that raidz1 configurations with an even number of children (e.g. 3+1), raidz2, and raidz3 are unaffected. The recommended course of action is to roll back to build snv_119 or earlier. If for some reason this is impossible, please email me PRIVATELY, and we can discuss the best course of action for you. After rolling back initiate a scrub. ZFS will identify and correct these errors, but if enough accumulate it will (incorrectly) identify drives as faulty (which they likely aren't). You can clear these failures (zpool clear pool). Without rolling back, repeated scrubs will eventually remove all traces of the data corruption. You may need to clear checksum failures as they're identified to ensure that enough drives remain online. For reference here's the bug: 6869090 filebench on thumper with ZFS (snv_120) raidz causes checksum errors from all drives Apologies for the bug and for any inconvenience this caused. Below is a technical description of the two issues. This is for interest only and does not contain additional discussion of symptoms or prescriptive action. Adam ---8--- 1. In situations where a block read from a RAID-Z vdev fails to checksum but there were no errors from any of the child vdevs (e.g. hard drives) we must enter combinatorial reconstruction in which we attempt every combination of data and parity until we find the correct data. The logic was modified to scale to triple-parity RAID-Z and in doing so I introduced a bug in which spurious errors reports may in some circumstances be generated for vdevs that were not used as part of the data reconstruction. These do not represent actual corruption or problems with the underlying devices and can be ignored and cleared. 2. This one is far subtler and requires an understanding of how RAID-Z writes work. For that I strongly recommend the following blog post from Jeff Bonwick: http://blogs.sun.com/bonwick/entry/raid_z Basically, RAID-Z writes full stripes every time; note that without careful accounting it would be possible to effectively fragment the vdev such that single sectors were free but useless since single-parity RAID-Z requires two adjacent sectors to store data (one for data, one for parity). To address this, RAID-Z rounds up its allocation to the next (nparity + 1). This ensures that all space is accounted for. RAID-Z will thus skip sectors that are unused based on this rounding. For example, under raidz1 a write of 1024 bytes would result in 512 bytes of parity, 512 bytes of data on two devices and 512 bytes skipped. To improve performance, ZFS aggregates multiple adjacent IOs into a single large IO. Further, hard drives themselves can perform aggregation of adjacent IOs. We noted that these skipped sectors were inhibiting performance so added optional IOs that could be used to improve aggregation. This yielded a significant performance boost for all RAID-Z configurations. Another nuance of single-parity RAID-Z is that while it normally lays down stripes as P D D (parity, data, data, ...), it will switch every megabyte to move the parity into the second position (data, parity, data, ...). This was ostensibly to effect the same improvement as between RAID-4 and RAID-5 -- distributed parity. However, to implement RAID-5 actually requires full distribution of parity AND RAID-Z already distributes parity by virtue of the skipped sectors and variable width stripes. In other words, this was not a particularly valid optimization. It was accordingly discarded for double- and tripe-parity RAID-Z. They contain no such swapping. The implementation of this swapping was not taken into account for the optional IOs so rather than writing the optional IO into the skipped sector, the optional IO overwrote the first sector of the subsequent stripe with zeros. The aggregation does not always happen so the corruption is ususally not pervasive. Futher, raidz1 vdevs with odd numbers of children are more likely to encounter the problem. Let's say we have a raidz1 vdev with three children. Two writes of 1K each would look like this: disks 0 1 2
Re: [zfs-discuss] zfs performance cliff when over 80% util, still occuring when pool in 6
So I have poked and prodded the disks and they both seem fine. Any yet my rpool is still slow. Any ideas on what do do now. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Motherboard for home zfs/solaris file server
Hello, your (open)solaris for Ecc support (which seems to have been dropped from 200906) is misunderstanding. OS 2009.06 also supports ECC as 2005 did. Just install it and use my updated ecccheck.pl script to get informed about errors. Also you might verify that Solaris' memory scrubber is really running if you are that curious: http://developmentonsolaris.wordpress.com/2009/03/06/how-to-make-sure-memory-scrubber-is-running/ Karel -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Read about ZFS backup - Still confused
I am just a simple home user. When I was using linux, I backed up my home directory (which contained all my critical data) using tar. I backed up my linux partition using partimage. These backups were put on dvd's. That way I could restore (and have) even if the hard drive completely went belly up. I would like to duplicate this scheme using zfs commands. I know I can copy a snapshot to a dvd but can I recover using just the snapshot or does it rely on the zfs file system on my hard drive being ok? Cork -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Alan, Thanks for the detailed explanation. The rollback successfully fixed my 5-disk RAID-Z errors. I'll hold off another upgrade attempt until 124 rolls out. Fortunately, I didn't do a zfs upgrade right away after installing 121. For those that did, this could be very painful. Gary -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Hey folk, The are two problems with RAID-Z in builds snv_120 through snv_123 that will both be resolved in build snv_124. The problems are as follows: Thanks for letting us know. Is there is a way to get promt updates on such issues for Opensolaris? (except like reading a discussion list). Paid support maybe is the answer? Is there any? -- Roman -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] How ZFS prefetch issues impact the use of Zvols are remote LUNs
Hi. As the subject indicates, I'm trying to understand the impact of the ZFS prefetch issues and if they only impact a local zfs filesystem versus say a zvol that remotely using the lun via iSCSI or Fibrechannel. Can anyone comment on this? The specific issues are consolidated here: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-July/030169.html Thanks in advance. This email and any files transmitted with it are confidential and are intended solely for the use of the individual or entity to whom they are addressed. This communication may contain material protected by the attorney-client privilege. If you are not the intended recipient, be advised that any use, dissemination, forwarding, printing or copying is strictly prohibited. If you have received this email in error, please contact the sender and delete all copies. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] [storage-discuss] How ZFS prefetch issues impact the use of Zvols are remote LUNs
Quoting en...@businessgrade.com: Hi. As the subject indicates, I'm trying to understand the impact of the ZFS prefetch issues and if they only impact a local zfs filesystem versus say a zvol that remotely using the lun via iSCSI or Fibrechannel. Can anyone comment on this? The specific issues are consolidated here: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-July/030169.html Thanks in advance. Sorry the question is not formatted properly. I mean to say, .. versus say a zvol that a remote host is using as a lun via iSCSI or FC This email and any files transmitted with it are confidential and are intended solely for the use of the individual or entity to whom they are addressed. This communication may contain material protected by the attorney-client privilege. If you are not the intended recipient, be advised that any use, dissemination, forwarding, printing or copying is strictly prohibited. If you have received this email in error, please contact the sender and delete all copies. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
I agree and Cindy Swearingen and I are talking to marketing to get this fixed. Thanks to all for bringing this to our attention. lori On 09/03/09 00:55, Ross wrote: I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
And a question here how to control number of dev version to install? -- Roman -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
On Sep 2, 2009, at 11:55 PM, Ross wrote: I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. Some points to help clarify the situation: 1. There is no other way to archive a dataset than using a snapshot 2. You cannot build a zpool on a tape 3. The stability of the protocol is only a problem if it becomes impossible to run some version of OpenSolaris on hardware that is needed to receive the snapshot. Given the ubiquity of virtualization and the x86 legacy, I don't think this is a problem for at least the expected lifetime of the storage medium. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. Data integrity is a problem for all archiving systems, not just ZFS. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Roman Naumenko wrote: Hey folk, The are two problems with RAID-Z in builds snv_120 through snv_123 that will both be resolved in build snv_124. The problems are as follows: Thanks for letting us know. Is there is a way to get promt updates on such issues for Opensolaris? (except like reading a discussion list). Paid support maybe is the answer? Is there any? You can learn about support options for OpenSolaris 2009.06 at http://www.sun.com/service/opensolaris/index.jsp?intcmp=2166 . However, AFAIK OpenSolaris 2009.06 does not have the problem being discussed. The snv_ builds are developer builds, and support contracts are not available for them. So if you want the newest supportable features, choose OpenSolaris (the distro). If you want to test out new features fresh out of the oven you can test the snv_ builds. --JeffV ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
yes to all the comments below. Those are all mitigating factors. But I also agree with Ross and Mike and others that we should be more clear about when send/recv is appropriate and when it's not the best choice. We're looking into it. Lori On 09/03/09 10:06, Richard Elling wrote: On Sep 2, 2009, at 11:55 PM, Ross wrote: I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. Some points to help clarify the situation: 1. There is no other way to archive a dataset than using a snapshot 2. You cannot build a zpool on a tape 3. The stability of the protocol is only a problem if it becomes impossible to run some version of OpenSolaris on hardware that is needed to receive the snapshot. Given the ubiquity of virtualization and the x86 legacy, I don't think this is a problem for at least the expected lifetime of the storage medium. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. Data integrity is a problem for all archiving systems, not just ZFS. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Adam Leventhal Sent: Thursday, September 03, 2009 2:08 AM To: zfs-discuss@opensolaris.org discuss Subject: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123 Hey folk, The are two problems with RAID-Z in builds snv_120 through snv_123 that will both be resolved in build snv_124. The problems are as follows: 1. Data corruption on a RAID-Z system of any sort (raidz1, raidz2, raidz3) can lead to spurious checksum errors being reported on devices that were not used as part of the reconstruction. These errors are harmless and can be cleared safely (zpool clear pool). 2. There is a far more serious problem with single-parity RAID-Z that can lead to data corruption. This data corruption is recoverable as long as no additional data corruption or drive failure occurs. That is to say, data is fine provided there is not an additional problem. The problem is present on all raidz1 configurations that use an odd number of children (disks) e.g. 4+1, or 6+1. Note that raidz1 configurations with an even number of children (e.g. 3+1), raidz2, and raidz3 are unaffected. The recommended course of action is to roll back to build snv_119 or earlier. If for some reason this is impossible, please email me PRIVATELY, and we can discuss the best course of action for you. After rolling back initiate a scrub. ZFS will identify and correct these errors, but if enough accumulate it will (incorrectly) identify drives as faulty (which they likely aren't). You can clear these failures (zpool clear pool). Without rolling back, repeated scrubs will eventually remove all traces of the data corruption. You may need to clear checksum failures as they're identified to ensure that enough drives remain online. For reference here's the bug: 6869090 filebench on thumper with ZFS (snv_120) raidz causes checksum errors from all drives Apologies for the bug and for any inconvenience this caused. Below is a technical description of the two issues. This is for interest only and does not contain additional discussion of symptoms or prescriptive action. Adam ---8--- 1. In situations where a block read from a RAID-Z vdev fails to checksum but there were no errors from any of the child vdevs (e.g. hard drives) we must enter combinatorial reconstruction in which we attempt every combination of data and parity until we find the correct data. The logic was modified to scale to triple-parity RAID-Z and in doing so I introduced a bug in which spurious errors reports may in some circumstances be generated for vdevs that were not used as part of the data reconstruction. These do not represent actual corruption or problems with the underlying devices and can be ignored and cleared. 2. This one is far subtler and requires an understanding of how RAID-Z writes work. For that I strongly recommend the following blog post from Jeff Bonwick: http://blogs.sun.com/bonwick/entry/raid_z Basically, RAID-Z writes full stripes every time; note that without careful accounting it would be possible to effectively fragment the vdev such that single sectors were free but useless since single-parity RAID-Z requires two adjacent sectors to store data (one for data, one for parity). To address this, RAID-Z rounds up its allocation to the next (nparity + 1). This ensures that all space is accounted for. RAID-Z will thus skip sectors that are unused based on this rounding. For example, under raidz1 a write of 1024 bytes would result in 512 bytes of parity, 512 bytes of data on two devices and 512 bytes skipped. To improve performance, ZFS aggregates multiple adjacent IOs into a single large IO. Further, hard drives themselves can perform aggregation of adjacent IOs. We noted that these skipped sectors were inhibiting performance so added optional IOs that could be used to improve aggregation. This yielded a significant performance boost for all RAID-Z configurations. Another nuance of single-parity RAID-Z is that while it normally lays down stripes as P D D (parity, data, data, ...), it will switch every megabyte to move the parity into the second position (data, parity, data, ...). This was ostensibly to effect the same improvement as between RAID-4 and RAID-5 -- distributed parity. However, to implement RAID-5 actually requires full distribution of parity AND RAID-Z already distributes parity by virtue of the skipped sectors and variable width stripes. In other words, this was not a particularly valid optimization. It was accordingly discarded for double- and tripe-parity RAID-Z. They contain no such swapping. The implementation of this swapping was not taken into account for the optional IOs so rather than writing the optional IO into the skipped sector, the optional IO overwrote the first sector of the subsequent stripe with zeros. The
Re: [zfs-discuss] zfs performance cliff when over 80% util, still occuring when pool in 6
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of John-Paul Drawneek Sent: Thursday, September 03, 2009 2:13 AM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] zfs performance cliff when over 80% util, still occuring when pool in 6 So I have poked and prodded the disks and they both seem fine. Any yet my rpool is still slow. Any ideas on what do do now. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] [storage-discuss] How ZFS prefetch issues impact the use of Zvols are remote LUNs
Please unsubscribe me COLLIER -Original Message- From: storage-discuss-boun...@opensolaris.org [mailto:storage-discuss-boun...@opensolaris.org] On Behalf Of en...@businessgrade.com Sent: Thursday, September 03, 2009 8:52 AM To: en...@businessgrade.com Cc: zfs-discuss@opensolaris.org; storage-disc...@opensolaris.org Subject: Re: [storage-discuss] How ZFS prefetch issues impact the use of Zvols are remote LUNs Quoting en...@businessgrade.com: Hi. As the subject indicates, I'm trying to understand the impact of the ZFS prefetch issues and if they only impact a local zfs filesystem versus say a zvol that remotely using the lun via iSCSI or Fibrechannel. Can anyone comment on this? The specific issues are consolidated here: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-July/030169.html Thanks in advance. Sorry the question is not formatted properly. I mean to say, .. versus say a zvol that a remote host is using as a lun via iSCSI or FC This email and any files transmitted with it are confidential and are intended solely for the use of the individual or entity to whom they are addressed. This communication may contain material protected by the attorney-client privilege. If you are not the intended recipient, be advised that any use, dissemination, forwarding, printing or copying is strictly prohibited. If you have received this email in error, please contact the sender and delete all copies. ___ storage-discuss mailing list storage-disc...@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/storage-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Read about ZFS backup - Still confused
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Cork Smith Sent: Thursday, September 03, 2009 4:43 AM To: zfs-discuss@opensolaris.org Subject: [zfs-discuss] Read about ZFS backup - Still confused I am just a simple home user. When I was using linux, I backed up my home directory (which contained all my critical data) using tar. I backed up my linux partition using partimage. These backups were put on dvd's. That way I could restore (and have) even if the hard drive completely went belly up. I would like to duplicate this scheme using zfs commands. I know I can copy a snapshot to a dvd but can I recover using just the snapshot or does it rely on the zfs file system on my hard drive being ok? Cork -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Lori Alt Sent: Thursday, September 03, 2009 8:59 AM To: Ross Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Archiving and Restoring Snapshots I agree and Cindy Swearingen and I are talking to marketing to get this fixed. Thanks to all for bringing this to our attention. lori On 09/03/09 00:55, Ross wrote: I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Gary Gendel Sent: Thursday, September 03, 2009 7:45 AM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123 Alan, Thanks for the detailed explanation. The rollback successfully fixed my 5-disk RAID-Z errors. I'll hold off another upgrade attempt until 124 rolls out. Fortunately, I didn't do a zfs upgrade right away after installing 121. For those that did, this could be very painful. Gary -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Richard Elling Sent: Thursday, September 03, 2009 9:06 AM To: Ross Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Archiving and Restoring Snapshots On Sep 2, 2009, at 11:55 PM, Ross wrote: I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. Some points to help clarify the situation: 1. There is no other way to archive a dataset than using a snapshot 2. You cannot build a zpool on a tape 3. The stability of the protocol is only a problem if it becomes impossible to run some version of OpenSolaris on hardware that is needed to receive the snapshot. Given the ubiquity of virtualization and the x86 legacy, I don't think this is a problem for at least the expected lifetime of the storage medium. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. Data integrity is a problem for all archiving systems, not just ZFS. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Roman Naumenko Sent: Thursday, September 03, 2009 9:03 AM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123 And a question here how to control number of dev version to install? -- Roman -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Jeff Victor Sent: Thursday, September 03, 2009 9:10 AM To: Roman Naumenko Cc: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123 Roman Naumenko wrote: Hey folk, The are two problems with RAID-Z in builds snv_120 through snv_123 that will both be resolved in build snv_124. The problems are as follows: Thanks for letting us know. Is there is a way to get promt updates on such issues for Opensolaris? (except like reading a discussion list). Paid support maybe is the answer? Is there any? You can learn about support options for OpenSolaris 2009.06 at http://www.sun.com/service/opensolaris/index.jsp?intcmp=2166 . However, AFAIK OpenSolaris 2009.06 does not have the problem being discussed. The snv_ builds are developer builds, and support contracts are not available for them. So if you want the newest supportable features, choose OpenSolaris (the distro). If you want to test out new features fresh out of the oven you can test the snv_ builds. --JeffV ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Roman Naumenko Sent: Thursday, September 03, 2009 8:39 AM To: zfs-discuss@opensolaris.org Subject: Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123 Hey folk, The are two problems with RAID-Z in builds snv_120 through snv_123 that will both be resolved in build snv_124. The problems are as follows: Thanks for letting us know. Is there is a way to get promt updates on such issues for Opensolaris? (except like reading a discussion list). Paid support maybe is the answer? Is there any? -- Roman -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] [storage-discuss] How ZFS prefetch issues impact the use of Zvols are remote LUNs
Please unsubscribe me COLLIER -Original Message- From: storage-discuss-boun...@opensolaris.org [mailto:storage-discuss-boun...@opensolaris.org] On Behalf Of en...@businessgrade.com Sent: Thursday, September 03, 2009 8:49 AM To: zfs-discuss@opensolaris.org; storage-disc...@opensolaris.org Subject: [storage-discuss] How ZFS prefetch issues impact the use of Zvols are remote LUNs Hi. As the subject indicates, I'm trying to understand the impact of the ZFS prefetch issues and if they only impact a local zfs filesystem versus say a zvol that remotely using the lun via iSCSI or Fibrechannel. Can anyone comment on this? The specific issues are consolidated here: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-July/030169.html Thanks in advance. This email and any files transmitted with it are confidential and are intended solely for the use of the individual or entity to whom they are addressed. This communication may contain material protected by the attorney-client privilege. If you are not the intended recipient, be advised that any use, dissemination, forwarding, printing or copying is strictly prohibited. If you have received this email in error, please contact the sender and delete all copies. ___ storage-discuss mailing list storage-disc...@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/storage-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
Please unsubscribe me COLLIER -Original Message- From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Lori Alt Sent: Thursday, September 03, 2009 9:14 AM To: Richard Elling Cc: zfs-discuss@opensolaris.org; Ross Subject: Re: [zfs-discuss] Archiving and Restoring Snapshots yes to all the comments below. Those are all mitigating factors. But I also agree with Ross and Mike and others that we should be more clear about when send/recv is appropriate and when it's not the best choice. We're looking into it. Lori On 09/03/09 10:06, Richard Elling wrote: On Sep 2, 2009, at 11:55 PM, Ross wrote: I agree, mailing that to all Sun customers is something I think is likely to turn around and bite you. Some points to help clarify the situation: 1. There is no other way to archive a dataset than using a snapshot 2. You cannot build a zpool on a tape 3. The stability of the protocol is only a problem if it becomes impossible to run some version of OpenSolaris on hardware that is needed to receive the snapshot. Given the ubiquity of virtualization and the x86 legacy, I don't think this is a problem for at least the expected lifetime of the storage medium. A lot of people are now going to use that to archive their data, and some of them are not going to be happy when months or years down the line they try to restore it and find that the 'zfs receive' just fails, with no hope of recovering their data. And they're not going to blame the device that's corrupted a bit, they're going to come here blaming Sun and ZFS. Archiving to a format with such a high risk of loosing everything sounds like incredibly bad advice. Especially when the article actually uses the term create archives for long-term storage. Long term storage doesn't mean storage that could break with the next upgrade, or storage that will fail if even a single bit has been corrupted anywhere in the process. Sun really need to decide what they are doing with zfs send/receive because we're getting very mixed messages now. Data integrity is a problem for all archiving systems, not just ZFS. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
Some points to help clarify the situation: 1. There is no other way to archive a dataset than n using a snapshot Other than tar, star, and numerous other archive utilities. The fact that ZFS doesn't have any alternative built in doesn't mean that dumping a send to a file suddenly becomes a good idea. Especially when there is no step in those instructions to verify that the archived file is actually a valid zfs send stream. 2. You cannot build a zpool on a tape So you need a different solution for tape, possibly streaming the files to tape, with the zpool configuration also documented and backed up somehow if necessary. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Yeah, I wouldn't mind knowing that too. With the old snv builds I just downloaded the appropriate image, with OpenSolaris and the development repository, is there any way to pick a particular build? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] nuke lots of snapshots
Sorry if this is a faq, but I just got a time sensitive dictim from the higherups to disable and remove all remnants of rolling snapshots on our DR filer. Is there a way for me to nuke all snapshots with a single command, or to I have to manually destroy all 600+ snapshots with zfs destroy? osol 2008.11 thx jake ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nuke lots of snapshots
Le 3 sept. 09 à 19:57, Jacob Ritorto a écrit : Sorry if this is a faq, but I just got a time sensitive dictim from the higherups to disable and remove all remnants of rolling snapshots on our DR filer. Is there a way for me to nuke all snapshots with a single command, or to I have to manually destroy all 600+ snapshots with zfs destroy? zfs list -r -t snapshot -o name -H pool | xargs -tl zfs destroy should destroy all the snapshots in a pool Gaëtan -- Gaëtan Lehmann Biologie du Développement et de la Reproduction INRA de Jouy-en-Josas (France) tel: +33 1 34 65 29 66fax: 01 34 65 29 09 http://voxel.jouy.inra.fr http://www.itk.org http://www.mandriva.org http://www.bepo.fr PGP.sig Description: Ceci est une signature électronique PGP ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] nuke lots of snapshots
Gaëtan Lehmann wrote: zfs list -r -t snapshot -o name -H pool | xargs -tl zfs destroy should destroy all the snapshots in a pool Thanks Gaëtan. I added 'grep auto' to filter on just the rolling snaps and found that xargs wouldn't let me put both flags on the same dash, so: zfs list -r -t snapshot -o name -H poolName | grep auto | xargs -t -l zfs destroy worked for me. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
On Sep 3, 2009, at 1:25 PM, Ross myxi...@googlemail.com wrote: Yeah, I wouldn't mind knowing that too. With the old snv builds I just downloaded the appropriate image, with OpenSolaris and the development repository, is there any way to pick a particular build? I just do a 'pkg list entire' and then install the build I want with a 'pkg install entire-build' -Ross ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] (no subject)
___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
On Sep 3, 2009, at 1:25 PM, Ross myxi...@googlemail.com wrote: Yeah, I wouldn't mind knowing that too. With the old snv builds I just downloaded the appropriate image, with OpenSolaris and the development repository, is there any way to pick a particular build? I just do a 'pkg list entire' and then install the build I want with a 'pkg install entire-build' Ross, can you provide details? Doesn't it show the latest? uname -a SunOS zsan00 5.11 snv_118 i86pc i386 i86pc Solaris r...@zsan00:~# pkg list entire NAME (PUBLISHER) VERSION STATE UFIX entire0.5.11-0.118installed u--- r...@zsan00:~# pkg refresh r...@zsan00:~# pkg list entire NAME (PUBLISHER) VERSION STATE UFIX entire0.5.11-0.118installed u--- -- Roman Naumenko -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Hey, web-admins, you see what happens when mailling list is screwed up from the beginning? -- Roman -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Hi Adam, Thanks for the info on this. Some people, including myself, reported seeing checksum errors within mirrors too. Is it considered that these checksum errors within mirrors could also be related to this bug, or is there another bug related to checksum errors within mirrors that I should take a look at? Search for 'mirror' here: http://opensolaris.org/jive/thread.jspa?threadID=111316tstart=0 Cheers, Simon And good luck with the fix for build 124. Are talking days or weeks for the fix to be available, do you think? :) -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] snv_110 - snv_121 produces checksum errors on Raid-Z pool
So what's the consensus on checksum errors appearing within mirror vdevs? Is it caused the same bug announced by Adam, or is something else causing it? If so, what's the bug id? Cheers, Simon -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] check a zfs rcvd file
On Wed, 2 Sep 2009 13:06:35 -0500 (CDT) Bob Friesenhahn bfrie...@simple.dallas.tx.us wrote: Nothing prevents validating the self-verifying archive file via this zfs recv -vn technique. Does this verify the ZFS format/integrity of the stream? Or is the only way to do that to zfs recv the stream into ZFS? -- Dick Hoogendijk -- PGP/GnuPG key: 01D2433D + http://nagual.nl/ | SunOS 10u7 5/09 | OpenSolaris 2010.02 B121 + All that's really worth doing is what we do for others (Lewis Carrol) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
Hey Simon, Thanks for the info on this. Some people, including myself, reported seeing checksum errors within mirrors too. Is it considered that these checksum errors within mirrors could also be related to this bug, or is there another bug related to checksum errors within mirrors that I should take a look at? Absolutely not. That is an unrelated issue. This problem is isolated to RAID-Z. And good luck with the fix for build 124. Are talking days or weeks for the fix to be available, do you think? :) -- Days or hours. Adam -- Adam Leventhal, Fishworks http://blogs.sun.com/ahl ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] check a zfs rcvd file
On 09/03/09 14:21, dick hoogendijk wrote: On Wed, 2 Sep 2009 13:06:35 -0500 (CDT) Bob Friesenhahn bfrie...@simple.dallas.tx.us wrote: Nothing prevents validating the self-verifying archive file via this zfs recv -vn technique. Does this verify the ZFS format/integrity of the stream? Or is the only way to do that to zfs recv the stream into ZFS? The -n option does some verification. It verifies that the record headers distributed throughout the stream are syntactically valid. Since each record header contains a length field which allows the next header to be found, one bad header will cause the processing of the stream to abort. But it doesn't verify the content of the data associated with each record. We might want to implement an option to enhance zfs recv -n to calculate a checksum of each dataset's records as it's reading the stream and then verify the checksum when the dataset's END record is seen. I'm looking at integrating a utility which allows the metadata in a stream to be dumped for debugging purposes (zstreamdump). It also verifies that the data in the stream agrees with the checksum. lori ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Problem with RAID-Z in builds snv_120 - snv_123
OK, thanks Adam. I'll look elsewhere for the mirror checksum error issue. In fact there's already a response here, which I shall check up on: http://opensolaris.org/jive/thread.jspa?messageID=413169#413169 Thanks again, and I look forward to grabbing 124 soon. Cheers, Simon -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] [osol-discuss] zfs data loss after crash
[cc'ing zfs-discuss] Scott Feldstein wrote: Hi, I had an opensolaris instance (b118) completely crash the other day (entirely my fault) and I took one of my mirrored zfs drives and imported into another system in order to retrieve the data. I imported the data via zpool import -f id pool and I just noticed that the data is way out of date. Most of the files look like I have lost 2 weeks worth of data. So, my questions are: 1) Is this expected? 2) Is there anyway to get my recent data back? Someone on zfs-discuss is likely to be able to help you better. Cheers, -- Shawn Walker ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Read about ZFS backup - Still confused
Let me try rephrasing this. I would like the ability to restore so my system mirrors its state at the time when I backed it up given the old hard drive is now a door stop. Cork -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] snv_110 - snv_121 produces checksum errors on Raid-Z pool
Thanks Gaëtan. What's the bug id for this iommu bug on Intel platforms? In my case, I have an AMD processor with ECC RAM, so probably not related to the Intel iommu bug. I'm seeing the checksum errors in a mirrored rpool using SSDs so maybe it could be something like cosmic rays causing occasional random bits to flip? After clearing the errors and scrubbing the pool a couple of times until the errors were fixed, I have not seen any new checksum errors, and I'm using 121 at the moment, though I should probably drop back to 117 to avoid the RAID-Z bug, although I have a RAID-Z2 vedev and not a RAID-Z1 vdev so I should not encounter the more serious problem mentioned. After the errors reported during the scrub on snv 121, I run a scrub on snv 118 and find the same amount of error, all on rpool/dump. I dropped that zvol, rerun the scrub again still on snv 118 without any error. After a reboot on snv 121 and a new scrub, no checksum error are reported. You did #zfs destroy rpool/dump ? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] snv_110 - snv_121 produces checksum errors on Raid-Z pool
It was someone from Sun that recently asked me to repost here about the checksum problem on mirrored drives. I was reluctant to do so because you and Bob might start flames again, and you did! You both sound very defensive, but of course I would never make an unsubstantiated speculation that you might have vulnerable hardware :-). But in case you do, please don't shoot the messenger... Instead of being negative, how about some conjectures of your own about this?. here's a summary of what is happening: An old machine with mirrored drives and a suspect mobo (maybe not checking PCI parity) gets checksum errors on reboot and scrub. With copies=1 it fails to repair them. With copies=2 it apparently fixes them, but zcksummon shows quite clearly that on a scrub, zfs finds and repairs them again on every scrub, even though scrub shows no errors. Typically these files are system libraries and unless you actually replace them, they are never truly repaired. Although I really don't think this is caused by cosmic rays, are you also saying that PCs without ECC on memory and/or buses will *never* experience a glitch? You obviously don't play the lottery :-) [ZFS errors due to memory hits seem far more likely than winning a 6 ball lottery for typical retail consumer loads] On 09/02/09 06:54 PM, Tim Cook wrote: Define more systems. How many people do you think are on 121? And of Absolutely no idea. Enough, though. those, how many are on the zfs mailing list? And of those, how many Probably - all of them (yes, this is an unsubstantiated speculation). have done a scrub recently to see the checksum errors? Do you have some proof to validate your beliefs? If you had read the thread carefully, you would note that a scrub actually clears the errors (but zcksummon shows that they really aren't cleared). And doesn't the guide tell us to run scrubs frequently? I am sure we all dutifully do so :-). I'd be quite happy to send you the proof. REGARDLESS, had you read all the posts to this thread, you'd know you've already been proven wrong: Wrong about what? Reading posts before they are posted? I have read every post most carefully. Having experienced checksum failures on mirrored drives for 4 months now (and there's a CR against snv115 for a similar problem), what exactly do you think I am trying to prove, or what beliefs? After 4 months of hearing the hardware being blamed for the checksum problem (which is easy to reproduce against snv111b), all I'm doing is agreeing that it is likely triggered by some kind of soft hardware glitch, we just don't know what the glitch might be. The SPoFs on this machine are the disk controller, the PCI bus, and memory, (and cpu, of course). Take your pick. FWIW it always picks on SUNWcsl (libdlpi.so.1) - 3 or 4 times now, and more recently, /usr/share/doc/SUNWmusicbrainz/COPYING.bz2. I am skeptical that the disk controller is picking on certain files, so that leaves memory and the bus. Take your pick. New files get added to the list quite infrequently. But it could also be a pure software bug - some kind of race condition, perhaps. On Wed, Sep 2, 2009 at 11:15 AM, Brent Jones br...@servuhome.net mailto:br...@servuhome.net wrote: I see this issue on each of my X4540's, 64GB of ECC memory, 1TB drives. Rolling back to snv_118 does not reveal any checksum errors, only snc_121 So, the commodity hardware here doesn't hold up, unless Sun isn't validating their equipment (not likely, as these servers have had no hardware issues prior to this build) Exactly. My whole point. Glad to hear that Sun hardware is as reliable as ever! I hope Richard's new and improved zcksummon will shed more light on this... Cheers -- Frank ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Read about ZFS backup - Still confused
Cork To answer your question just use tar for everything. It's about the best we've got. :-( When the disk turns into a doorstop re-install OpenSolaris/Solaris and then tar back all your data. I keep a complete list of EVER change I make on any OS (including the Redmond one) so I can re-create the machine. And, IMHO - And I know I will get shot at for saying it, but One reason why I would not use ZFS root in a real live production environment, is not having the equivalent of ufsdump/ufsrestore so I can do a bare metal restore. ZFS root works great on my laptop, but I know lots who still rely on ufsdump to a local tape drive for quick bare metal restores. The only good news in UNIX is much more tidy then Windows and there is very little that is not in /home (or /export/home) that gets changed throughout the OSes life. Unless somebody know better Cork Smith wrote: Let me try rephrasing this. I would like the ability to restore so my system mirrors its state at the time when I backed it up given the old hard drive is now a door stop. Cork www.eagle.co.nz This email is confidential and may be legally privileged. If received in error please destroy and immediately notify us. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Archiving and Restoring Snapshots
On Fri, Sep 4, 2009 at 12:17 AM, Ross myxi...@googlemail.com wrote: Hi Richard, Actually, reading your reply has made me realise I was overlooking something when I talked about tar, star, etc... How do you backup a ZFS volume? That's something traditional tools can't do. Are snapshots the only way to create a backup or archive of those? Personally I'm quite happy with snapshots - we have a ZFS system at work that's replicating all of it's data to an offsite ZFS store using snapshots. Using ZFS as a backup store is something I'm quite happy with, it's just storing just a snapshot file that makes me nervous. The correct answer is ndmp. Whether Sun will ever add it to opensolaris is another subject entirely though. --Tim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss