Re: [zfs-discuss] Thoughts on ZFS Pool Backup Strategies

2010-03-22 Thread Constantin Gonzalez
Hi, I agree 100% with Chris. Notice the on their own part of the original post. Yes, nobody wants to run zfs send or (s)tar by hand. That's why Chris's script is so useful: You set it up and forget and get the job done for 80% of home users. On another note, I was positively surprised by the

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Andrew Gabriel
Robert Milkowski wrote: To add my 0.2 cents... I think starting/stopping scrub belongs to cron, smf, etc. and not to zfs itself. However what would be nice to have is an ability to freeze/resume a scrub and also limit its rate of scrubbing. One of the reason is that when working in SAN

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Svein Skogen
On 22.03.2010 02:13, Edward Ned Harvey wrote: Actually ... Why should there be a ZFS property to share NFS, when you can already do that with share and dfstab? And still the zfs property exists. Probably because it is easy to create new filesystems and clone them; as NFS only works per

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Svein Skogen
On 21.03.2010 01:25, Robert Milkowski wrote: To add my 0.2 cents... I think starting/stopping scrub belongs to cron, smf, etc. and not to zfs itself. However what would be nice to have is an ability to freeze/resume a scrub and also limit its rate of scrubbing. One of the reason is that when

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Robert Milkowski
On 22/03/2010 01:13, Edward Ned Harvey wrote: Actually ... Why should there be a ZFS property to share NFS, when you can already do that with share and dfstab? And still the zfs property exists. Probably because it is easy to create new filesystems and clone them; as

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Robert Milkowski
On 22/03/2010 08:49, Andrew Gabriel wrote: Robert Milkowski wrote: To add my 0.2 cents... I think starting/stopping scrub belongs to cron, smf, etc. and not to zfs itself. However what would be nice to have is an ability to freeze/resume a scrub and also limit its rate of scrubbing. One

[zfs-discuss] David Plaunt is currently away.

2010-03-22 Thread D Plaunt
I will be out of the office starting 22/03/2010 and will not return until 06/04/2010. Hello, I am currently working on a project and out of the office. I will be checking my message twice a day but may be unavailable to follow up on your requests. If the matter requires immediate attention

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Edward Ned Harvey
Does cron happen to know how many other scrubs are running, bogging down your IO system? If the scrub scheduling was integrated into zfs itself, It doesn't need to. Crontab entry: /root/bin/scruball.sh /root/bin/scruball.sh: #!/usr/bin/bash for filesystem in filesystem1 filesystem2

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Svein Skogen
On 22.03.2010 13:35, Edward Ned Harvey wrote: Does cron happen to know how many other scrubs are running, bogging down your IO system? If the scrub scheduling was integrated into zfs itself, It doesn't need to. Crontab entry: /root/bin/scruball.sh /root/bin/scruball.sh: #!/usr/bin/bash for

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Edward Ned Harvey
no, it is not a subdirectory it is a filesystem mounted on top of the subdirectory. So unless you use NFSv4 with mirror mounts or an automounter other NFS version will show you contents of a directory and not a filesystem. It doesn't matter if it is a zfs or not. Ok, I learned something

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Edward Ned Harvey
IIRC it's zpool scrub, and last time I checked, the zpool command exited (with status 0) as soon as it had started the scrub. Your command would start _ALL_ scrubs in paralell as a result. You're right. I did that wrong. Sorry 'bout that. So either way, if there's a zfs property for scrub,

Re: [zfs-discuss] ZFS+CIFS: Volume Shadow Services, or Simple Symlink?

2010-03-22 Thread Edward Ned Harvey
Not being a CIFS user, could you clarify/confirm for me.. is this just a presentation issue, ie making a directory icon appear in a gooey windows explorer (or mac or whatever equivalent) view for people to click on? The windows client could access the .zfs/snapshot dir via typed pathname if

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Robert Milkowski
On 22/03/2010 12:50, Edward Ned Harvey wrote: no, it is not a subdirectory it is a filesystem mounted on top of the subdirectory. So unless you use NFSv4 with mirror mounts or an automounter other NFS version will show you contents of a directory and not a filesystem. It doesn't matter if it is

Re: [zfs-discuss] ZFS Performance on SATA Deive

2010-03-22 Thread Kashif Mumtaz
hi, Thanks for all the reply. I have found the real culprit. Hard disk was faulty. I changed the hard disk.And now ZFS performance is much better. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Svein Skogen
On 22.03.2010 13:54, Edward Ned Harvey wrote: IIRC it's zpool scrub, and last time I checked, the zpool command exited (with status 0) as soon as it had started the scrub. Your command would start _ALL_ scrubs in paralell as a result. You're right. I did that wrong. Sorry 'bout that. So

Re: [zfs-discuss] Intel SASUC8I - worth every penny

2010-03-22 Thread Cooper Hubbell
I've moved to 7200RPM 2.5 laptop drives over 3.5 drives, for a combination of reasons: lower-power, better performance than a comparable sized 3.5 drives, and generally lower-capacities meaning resilver times are smaller. They're a bit more $/GB, but not a lot. If you can stomach the

Re: [zfs-discuss] Intel SASUC8I - worth every penny

2010-03-22 Thread Erik Trimble
Cooper Hubbell wrote: Regarding the 2.5 laptop drives, do the inherent error detection properties of ZFS subdue any concerns over a laptop drive's higher bit error rate or rated MTBF? I've been reading about OpenSolaris and ZFS for several months now and am incredibly intrigued, but have yet

Re: [zfs-discuss] Intel SASUC8I - worth every penny

2010-03-22 Thread Svein Skogen
On 22.03.2010 16:24, Cooper Hubbell wrote: I've moved to 7200RPM 2.5 laptop drives over 3.5 drives, for a combination of reasons: lower-power, better performance than a comparable sized 3.5 drives, and generally lower-capacities meaning resilver times are smaller. They're a bit more $/GB, but

Re: [zfs-discuss] Thoughts on ZFS Pool Backup Strategies

2010-03-22 Thread David Dyer-Bennet
On Sat, March 20, 2010 07:31, Chris Gerhard wrote: Up to a point. zfs send | zfs receive does make a very good back up scheme for the home user with a moderate amount of storage. Especially when the entire back up will fit on a single drive which I think would cover the majority of home

Re: [zfs-discuss] Rethinking my zpool

2010-03-22 Thread Chris Dunbar
Thank you to all who responded. This response in particular was very helpful and I think I will stick with my current zpool configuration (choice a if you're reading below). I primarily host VMware virtual machines over NFS from this server's predecessor and this server will be doing the same

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Richard Elling
On Mar 22, 2010, at 7:30 AM, Svein Skogen wrote: On 22.03.2010 13:54, Edward Ned Harvey wrote: IIRC it's zpool scrub, and last time I checked, the zpool command exited (with status 0) as soon as it had started the scrub. Your command would start _ALL_ scrubs in paralell as a result.

Re: [zfs-discuss] ZFS send and receive corruption across a WAN link?

2010-03-22 Thread Richard Elling
On Mar 19, 2010, at 1:28 PM, Richard Jahnel wrote: They way we do this here is: zfs snapshot voln...@snapnow [i]#code to break on error and email not shown.[/i] zfs send -i voln...@snapbefore voln...@snapnow | pigz -p4 -1 file [i]#code to break on error and email not shown.[/i] scp

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Svein Skogen
On 22.03.2010 18:10, Richard Elling wrote: On Mar 22, 2010, at 7:30 AM, Svein Skogen wrote: On 22.03.2010 13:54, Edward Ned Harvey wrote: IIRC it's zpool scrub, and last time I checked, the zpool command exited (with status 0) as soon as it had started the scrub. Your command would start

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Richard Elling
On Mar 22, 2010, at 10:36 AM, Svein Skogen wrote: On 22.03.2010 18:10, Richard Elling wrote: On Mar 22, 2010, at 7:30 AM, Svein Skogen wrote: On 22.03.2010 13:54, Edward Ned Harvey wrote: IIRC it's zpool scrub, and last time I checked, the zpool command exited (with status 0) as soon as it

[zfs-discuss] David Plaunt is currently away.

2010-03-22 Thread D Plaunt
I will be out of the office starting 22/03/2010 and will not return until 06/04/2010. Hello, I am currently working on a project and out of the office. I will be checking my message twice a day but may be unavailable to follow up on your requests. If the matter requires immediate attention

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Bill Sommerfeld
On 03/22/10 11:02, Richard Elling wrote: Scrub tends to be a random workload dominated by IOPS, not bandwidth. you may want to look at this again post build 128; the addition of metadata prefetch to scrub/resilver in that build appears to have dramatically changed how it performs (largely for

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Richard Elling
On Mar 22, 2010, at 11:33 AM, Bill Sommerfeld wrote: On 03/22/10 11:02, Richard Elling wrote: Scrub tends to be a random workload dominated by IOPS, not bandwidth. you may want to look at this again post build 128; the addition of metadata prefetch to scrub/resilver in that build appears to

[zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Harry Putnam
This may be a bit dimwitted since I don't really understand how snapshots work. I mean the part concerning COW (copy on right) and how it takes so little room. But here I'm not asking about that. It appears to me that the default snapshot setup shares some aspects of a vcs (version control

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Matt Cowger
This is totally doable, and a reasonable use of zfs snapshots - we do some similar things. You can easily determine if the snapshot has changed by checking the output of zfs list for the snapshot. --M -Original Message- From: zfs-discuss-boun...@opensolaris.org

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Ian Collins
On 03/23/10 09:34 AM, Harry Putnam wrote: This may be a bit dimwitted since I don't really understand how snapshots work. I mean the part concerning COW (copy on right) and how it takes so little room. But here I'm not asking about that. It appears to me that the default snapshot setup shares

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Edward Ned Harvey
This may be a bit dimwitted since I don't really understand how snapshots work. I mean the part concerning COW (copy on right) and how it takes so little room. COW and snapshots are very simple to explain. Suppose you're chugging along using your filesystem, and then one moment, you tell the

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Edward Ned Harvey
In other words, there is no case where multiple scrubs compete for the resources of a single disk because a single disk only participates in one pool. Excellent point. However, the problem scenario was described as SAN. I can easily imagine a scenario where some SAN administrator created

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Harry Putnam
Matt Cowger mcow...@salesforce.com writes: This is totally doable, and a reasonable use of zfs snapshots - we do some similar things. Good, thanks for the input. You can easily determine if the snapshot has changed by checking the output of zfs list for the snapshot. Do you mean to just

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Matt Cowger
zfs list | grep '@' zpool/f...@1154758324G - 461G - zpool/f...@1208482 6.94G - 338G - zpool/f...@daily.netbackup 1.07G - 344G - zpool/f...@11547581.77G - 242G -

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Brandon High
On Mon, Mar 22, 2010 at 1:58 PM, Ian Collins i...@ianshome.com wrote: On 03/23/10 09:34 AM, Harry Putnam wrote: Oh, something I meant to ask... is there some standard way to tell before calling for a snapshot, if the directory structure has changed at all, other than aging I mean. Is there

Re: [zfs-discuss] Proposition of a new zpool property.

2010-03-22 Thread Brandon High
On Mon, Mar 22, 2010 at 12:21 PM, Richard Elling richard.ell...@gmail.comwrote: Yes, it is better. But still nowhere near platter speed. All it takes is one little seek... True, dat. I find that scrubs start very slow ( 20MB/s) with the disks at near-100% utilization. Towards the end of

Re: [zfs-discuss] ZFS send and receive corruption across a WAN link?

2010-03-22 Thread Brandon High
On Mon, Mar 22, 2010 at 10:26 AM, Richard Elling richard.ell...@gmail.comwrote: NB. deduped streams should further reduce the snapshot size. I haven't seen a lot of discussion on the list regarding send dedup, but I understand it'll use the DDT if you have dedup enabled on your dataset. What's

Re: [zfs-discuss] CR 6880994 and pkg fix

2010-03-22 Thread Frank Middleton
On 03/21/10 03:24 PM, Richard Elling wrote: I feel confident we are not seeing a b0rken drive here. But something is clearly amiss and we cannot rule out the processor, memory, or controller. Absolutely no question of that, otherwise this list would be flooded :-). However, the purpose of

Re: [zfs-discuss] ZFS send and receive corruption across a WAN link?

2010-03-22 Thread Lori Alt
On 03/22/10 05:04 PM, Brandon High wrote: On Mon, Mar 22, 2010 at 10:26 AM, Richard Elling richard.ell...@gmail.com mailto:richard.ell...@gmail.com wrote: NB. deduped streams should further reduce the snapshot size. I haven't seen a lot of discussion on the list regarding send dedup,

Re: [zfs-discuss] ZFS send and receive corruption across a WAN link?

2010-03-22 Thread Nicolas Williams
On Thu, Mar 18, 2010 at 10:38:00PM -0700, Rob wrote: Can a ZFS send stream become corrupt when piped between two hosts across a WAN link using 'ssh'? No. SSHv2 uses HMAC-MD5 and/or HMAC-SHA-1, depending on what gets negotiated, for integrity protection. The chances of random on the wire

Re: [zfs-discuss] snapshots as versioning tool

2010-03-22 Thread Edward Ned Harvey
You can easily determine if the snapshot has changed by checking the output of zfs list for the snapshot. Do you mean to just grep it out of the output of zfs list -t snapshot I think the point is: You can easily tell how many MB changed in a snapshot, and therefore you can easily

[zfs-discuss] zfs send/receive and file system properties

2010-03-22 Thread Len Zaifman
I am trying to coordinate properties and data between 2 file servers. on file server 1 I have: zfs get all zfs52/export/os/sles10sp2 NAME PROPERTY VALUE SOURCE zfs52/export/os/sles10sp2 type filesystem

[zfs-discuss] LSISAS2004 support

2010-03-22 Thread Bart Nabbe
All, I did some digging and I was under the impression that the mr_sas driver was to support the LSISAS2004 HBA controller from LSI. I did add the pci id to the driver alias for mr_sas, but then the driver still showed up as unattached (see below). Did I miss something, or was my assumption

Re: [zfs-discuss] CR 6880994 and pkg fix

2010-03-22 Thread Richard Elling
On Mar 22, 2010, at 4:21 PM, Frank Middleton wrote: On 03/21/10 03:24 PM, Richard Elling wrote: I feel confident we are not seeing a b0rken drive here. But something is clearly amiss and we cannot rule out the processor, memory, or controller. Absolutely no question of that, otherwise

[zfs-discuss] pool use from network poor performance

2010-03-22 Thread homerun
Hi i have now two pools rpool 2-way mirror ( pata ) data 4-way raidz2 ( sata ) if i access to datapool from network , smb , nfs , ftp , sftp , jne... i get only max 200 KB/s speeds compared to rpool that give XX MB/S speeds to and from network it is slow. Any ideas what reasons might be and