Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread James Van Artsdalen
On Thu, 8 Jul 2010, Edward Ned Harvey wrote: Yep. Provided it supported ZFS, a Mac Mini makes for a compelling SOHO server. Warning: a Mac Mini does not have eSATA ports for external storage. It's dangerous to use USB for external storage since many (most? all?) USB-SATA chips discard SYNC

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Alex Blewitt
On 9 Jul 2010, at 08:55, James Van Artsdalen james-opensola...@jrv.org wrote: On Thu, 8 Jul 2010, Edward Ned Harvey wrote: Yep. Provided it supported ZFS, a Mac Mini makes for a compelling SOHO server. Warning: a Mac Mini does not have eSATA ports for external storage. It's dangerous to

Re: [zfs-discuss] Consequences of resilvering failure

2010-07-09 Thread Brandon High
On Tue, Jul 6, 2010 at 10:05 AM, Roy Sigurd Karlsbakk r...@karlsbakk.netwrote: The pool will remain available, but you will have data corruption. The simple way to avoid this, is to use a raidz2, where the chances are far lower for data corruption. It's also possible to replace a drive while

Re: [zfs-discuss] dedup accounting anomaly / dedup experiments

2010-07-09 Thread Brandon High
On Thu, Jul 1, 2010 at 1:33 AM, Lutz Schumann presa...@storageconcepts.dewrote: Anyone knowing why the dedup factor is wrong ? Any insights on what has actually been written (compressed meta data, deduped meta data .. etc.) would be greatly appreshiated. Metadata and ditto blocks. Even with

Re: [zfs-discuss] SATA 6G controller for OSOL

2010-07-09 Thread Vladimir Kotal
On 07/ 9/10 09:58 AM, Brandon High wrote: On Fri, Jul 9, 2010 at 12:42 AM, James Van Artsdalen james-opensola...@jrv.org wrote: If these 6 Gb/s controllers are based on the Marvell part I would test them thoroughly before deployment - those chips have been problematic. The Marvell 88SE9123

Re: [zfs-discuss] ZFS recovery tools

2010-07-09 Thread R. Eulenberg
So, just I took the right command. For me the output is very cryptic and I cannot get any information helping me. I uploaded the output to a filehoster. http://ifile.it/vzwn50s/Output.txt I hope you can tell me what it means. Regards ron -- This message posted from opensolaris.org

[zfs-discuss] spreading data after adding devices to pool

2010-07-09 Thread George Helyar
I use ZFS (on FreeBSD) for my home NAS. I started on 4 drives then added 4 and have now added another 4, bringing the total up to 12 drives on 3 raidzs in 1 pool. I was just wondering if there was any advantage or disadvantage to spreading the data across the 3 raidz, as two are currently full

Re: [zfs-discuss] Should i enable Write-Cache ?

2010-07-09 Thread Richard Elling
On Jul 8, 2010, at 4:37 PM, Edward Ned Harvey wrote: From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- boun...@opensolaris.org] On Behalf Of Philippe Schwarz 3Ware cards Any drawback (except that without BBU, i've got a pb in case of power loss) in enabling the WC with ZFS

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Edward Ned Harvey
From: Rich Teer [mailto:rich.t...@rite-group.com] Sent: Thursday, July 08, 2010 7:43 PM Yep. Provided it supported ZFS, a Mac Mini makes for a compelling SOHO server. The lack of ZFS is the main thing holding me back here... I don't really want to go into much detail here (it's a zfs

Re: [zfs-discuss] SATA 6G controller for OSOL

2010-07-09 Thread Matt Urbanowski
My advice would be to NOT use the AD2SA6GPX1 HBA for building an opensolaris storage box. Although the AHCI driver will load, the drives are not visible to the OS, and device configuration fails according to 'cfgadm -al'. I have a couple of them that are now residing in a linux box as I was

[zfs-discuss] Hash functions (was Re: Hashing files rapidly on ZFS)

2010-07-09 Thread Nicolas Williams
On Thu, Jul 08, 2010 at 08:42:33PM -0700, Garrett D'Amore wrote: On Fri, 2010-07-09 at 10:23 +1000, Peter Jeremy wrote: In theory, collisions happen. In practice, given a cryptographic hash, if you can find two different blocks or files that produce the same output, please publicise it

Re: [zfs-discuss] spreading data after adding devices to pool

2010-07-09 Thread Matt Urbanowski
You could move the data elsewhere using zfs send and recv, destroy the original datasets and then recreate them. This would stripe the data across the vdevs. Of course, when BP-rewrite becomes available it should be possible to simply redistribute blocks amongst the various vdevs without having

Re: [zfs-discuss] zpool spares listed twice, as both AVAIL and FAULTED

2010-07-09 Thread Ryan Schwartz
Hi Cindy, Not sure exactly when the drives went into this state, but it is likely that it happened when I added a second pool, added the same spares to the second pool, then later destroyed the second pool. There have been no controller or any other hardware changes to this system - it is all

[zfs-discuss] BP rewrite? (Was Re: spreading data after adding devices to pool)

2010-07-09 Thread Roy Sigurd Karlsbakk
Anyone knows where in the pipeline BP rewrite is, or how long this pipeline is? You could move the data elsewhere using zfs send and recv, destroy the original datasets and then recreate them. This would stripe the data across the vdevs. Of course, when BP-rewrite becomes available it

Re: [zfs-discuss] zpool spares listed twice, as both AVAIL and FAULTED

2010-07-09 Thread Cindy Swearingen
Hi Ryan, Which Solaris release is this? Thanks, Cindy On 07/09/10 10:38, Ryan Schwartz wrote: Hi Cindy, Not sure exactly when the drives went into this state, but it is likely that it happened when I added a second pool, added the same spares to the second pool, then later destroyed the

Re: [zfs-discuss] zpool spares listed twice, as both AVAIL and FAULTED

2010-07-09 Thread Ryan Schwartz
Ok, so after removing the spares marked as AVAIL and re-adding them again, I put myself back in the you're effed, dude boat. What I should have done at that point is a zpool export/import at that point which would have resolved it. So what I did was recreate the steps that got me into the state

Re: [zfs-discuss] zpool spares listed twice, as both AVAIL and FAULTED

2010-07-09 Thread Ryan Schwartz
Cindy, [IDGSUN02:/] root# cat /etc/release Solaris 10 10/08 s10x_u6wos_07b X86 Copyright 2008 Sun Microsystems, Inc. All Rights Reserved. Use is subject to license terms. Assembled 27 October 2008 But as

Re: [zfs-discuss] zpool spares listed twice, as both AVAIL and FAULTED

2010-07-09 Thread Cindy Swearingen
I was going to suggest the export/import step next. :-) I'm glad you were able to resolve it. We are working on making spare behavior more robust. In the meantime, my advice is keep life simple and do not share spares, logs, caches, or even disks between pools. Thanks, Cindy On 07/09/10

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Miles Nordin
ab == Alex Blewitt alex.blew...@gmail.com writes: ab All Mac Minis have FireWire - the new ones have FW800. I tried attaching just two disks to a ZFS host using firewire, and it worked very badly for me. I found: 1. The solaris firewire stack isn't as good as the Mac OS one. 2. Solaris

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Garrett D'Amore
On Fri, 2010-07-09 at 15:02 -0400, Miles Nordin wrote: ab == Alex Blewitt alex.blew...@gmail.com writes: ab All Mac Minis have FireWire - the new ones have FW800. I tried attaching just two disks to a ZFS host using firewire, and it worked very badly for me. I found: 1. The

Re: [zfs-discuss] snapshot out of space

2010-07-09 Thread Brandon High
On Fri, Jul 9, 2010 at 8:04 AM, Tony MacDoodle tpsdoo...@gmail.com wrote: datapool/pluto refreservation 70G local This means that every snapshot will require 70G of free space? No. Could you provide the information requested? -B -- Brandon High : bh...@freaks.com

Re: [zfs-discuss] SATA 6G controller for OSOL

2010-07-09 Thread Brandon High
On Fri, Jul 9, 2010 at 2:40 AM, Vladimir Kotal vladimir.ko...@sun.comwrote: Could you be more specific about the problems with 88SE9123, especially with SATA ? I am in the process of setting up a system with AD2SA6GPX1 HBA based on this chipset (at least according to the product pages [*]).

Re: [zfs-discuss] SATA 6G controller for OSOL

2010-07-09 Thread Matt Urbanowski
Agreed! I'm not sure why Addonics is selling them, given the history of problems. At any rate, I'm glad that I didn't pay anything for the 3 three that I have. On Fri, Jul 9, 2010 at 1:56 PM, Brandon High bh...@freaks.com wrote: On Fri, Jul 9, 2010 at 2:40 AM, Vladimir Kotal

Re: [zfs-discuss] zfs snapshot revert

2010-07-09 Thread zfsnoob4
I'm not trying to fix anything in particular, I'm just curious. In case I rollback a filesystem and then realize, I wanted a file from the original file system (before rollback). I read the section on clones here: http://docs.sun.com/app/docs/doc/819-5461/gavvx?a=view but I'm still not sure

[zfs-discuss] resilver of older root pool disk

2010-07-09 Thread Frank Middleton
This is a hypothetical question that could actually happen: Suppose a root pool is a mirror of c0t0d0s0 and c0t1d0s0 and for some reason c0t0d0s0 goes off line, but comes back on line after a shutdown. The primary boot disk would then be c0t0d0s0 which would have much older data than c0t1d0s0.

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Alex Blewitt
On 9 Jul 2010, at 20:38, Garrett D'Amore wrote: On Fri, 2010-07-09 at 15:02 -0400, Miles Nordin wrote: ab == Alex Blewitt alex.blew...@gmail.com writes: ab All Mac Minis have FireWire - the new ones have FW800. I tried attaching just two disks to a ZFS host using firewire, and it

Re: [zfs-discuss] BP rewrite? (Was Re: spreading data after adding devices to pool)

2010-07-09 Thread Lasse Osterild
+1 I badly need this. On 09/07/2010, at 19.40, Roy Sigurd Karlsbakk wrote: Anyone knows where in the pipeline BP rewrite is, or how long this pipeline is? You could move the data elsewhere using zfs send and recv, destroy the original datasets and then recreate them. This would stripe

Re: [zfs-discuss] zfs snapshot revert

2010-07-09 Thread Ian Collins
On 07/10/10 08:10 AM, zfsnoob4 wrote: I'm not trying to fix anything in particular, I'm just curious. In case I rollback a filesystem and then realize, I wanted a file from the original file system (before rollback). I read the section on clones here:

[zfs-discuss] zfs send/recv hanging in 2009.06

2010-07-09 Thread BJ Quinn
I have a couple of systems running 2009.06 that hang on relatively large zfs send/recv jobs. With the -v option, I see the snapshots coming across, and at some point the process just pauses, IO and CPU usage go to zero, and it takes a hard reboot to get back to normal. The same script running

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Peter Taps
Folks, I would appreciate it if you can create a separate thread for Mac Mini. Back to the original subject. NetApp has deep pockets. A few companies have already backed out of zfs as they cannot afford to go through a lawsuit. I am in a stealth startup company and we rely on zfs for our

[zfs-discuss] ZFS, IPS (IBM ServeRAID) driver, and a kernel panic...

2010-07-09 Thread Kyle McDonald
Hi, I have been trying out the latest NextentaCore and NexentaStor Community ed. builds (they have the driver I need built in) on the hardware I have with this controller. The only difference between the 2 machines is that the 'Core' machine has 16GB of RAM and the 'Stor' one has 12GB. On both

Re: [zfs-discuss] ZFS, IPS (IBM ServeRAID) driver, and a kernel panic...

2010-07-09 Thread Garrett D'Amore
First off, you need to test 3.0.3 if you're using dedup. Earlier versions had an unduly large number of issues when used with dedup. Hopefully with 3.0.3 we've got the bulk of the problems resolved. ;-) Secondly, from your stack backtrace, yes, it appears ips is implicated. If I had source for

Re: [zfs-discuss] SATA 6G controller for OSOL

2010-07-09 Thread Graham McArdle
This thread from Marc Bevand and his blog linked therein might have some useful alternative suggestions. http://opensolaris.org/jive/thread.jspa?messageID=480925 I've bookmarked it because it's quite a handy summary and I hope he keeps updating it with new info -- This message posted from

Re: [zfs-discuss] zfs send/recv hanging in 2009.06

2010-07-09 Thread Ian Collins
On 07/10/10 09:49 AM, BJ Quinn wrote: I have a couple of systems running 2009.06 that hang on relatively large zfs send/recv jobs. With the -v option, I see the snapshots coming across, and at some point the process just pauses, IO and CPU usage go to zero, and it takes a hard reboot to get

Re: [zfs-discuss] zfs send/recv hanging in 2009.06

2010-07-09 Thread Giovanni Tirloni
On Fri, Jul 9, 2010 at 6:49 PM, BJ Quinn bjqu...@seidal.com wrote: I have a couple of systems running 2009.06 that hang on relatively large zfs send/recv jobs.  With the -v option, I see the snapshots coming across, and at some point the process just pauses, IO and CPU usage go to zero, and

[zfs-discuss] Debunking the dedup memory myth

2010-07-09 Thread Edward Ned Harvey
Whenever somebody asks the question, How much memory do I need to dedup X terabytes filesystem, the standard answer is as much as you can afford to buy. This is true and correct, but I don't believe it's the best we can do. Because as much as you can buy is a true assessment for memory in *any*

Re: [zfs-discuss] Debunking the dedup memory myth

2010-07-09 Thread Brandon High
On Fri, Jul 9, 2010 at 5:00 PM, Edward Ned Harvey solar...@nedharvey.comwrote: The default ZFS block size is 128K. If you have a filesystem with 128G used, that means you are consuming 1,048,576 blocks, each of which must be checksummed. ZFS uses adler32 and sha256, which means 4bytes and

[zfs-discuss] block align SSD for use as a l2arc cache

2010-07-09 Thread Geoff Nordli
I have an Intel X25-M 80GB SSD. For optimum performance, I need to block align the SSD device, but I am not sure exactly how I should to it. If I run the format - fdisk it allows me to partition based on a cylinder, but I don't think that is sufficient enough. Can someone tell me how

Re: [zfs-discuss] Legality and the future of zfs...

2010-07-09 Thread Erik Trimble
On 7/9/2010 2:55 PM, Peter Taps wrote: Folks, I would appreciate it if you can create a separate thread for Mac Mini. Back to the original subject. NetApp has deep pockets. A few companies have already backed out of zfs as they cannot afford to go through a lawsuit. I am in a stealth startup

Re: [zfs-discuss] Debunking the dedup memory myth

2010-07-09 Thread Erik Trimble
On 7/9/2010 5:18 PM, Brandon High wrote: On Fri, Jul 9, 2010 at 5:00 PM, Edward Ned Harvey solar...@nedharvey.com mailto:solar...@nedharvey.com wrote: The default ZFS block size is 128K. If you have a filesystem with 128G used, that means you are consuming 1,048,576 blocks, each of

Re: [zfs-discuss] block align SSD for use as a l2arc cache

2010-07-09 Thread Erik Trimble
On 7/9/2010 5:55 PM, Geoff Nordli wrote: I have an Intel X25-M 80GB SSD. For optimum performance, I need to block align the SSD device, but I am not sure exactly how I should to it. If I run the format - fdisk it allows me to partition based on a cylinder, but I don't think that is

Re: [zfs-discuss] Debunking the dedup memory myth

2010-07-09 Thread Neil Perrin
On 07/09/10 19:40, Erik Trimble wrote: On 7/9/2010 5:18 PM, Brandon High wrote: On Fri, Jul 9, 2010 at 5:00 PM, Edward Ned Harvey solar...@nedharvey.com mailto:solar...@nedharvey.com wrote: The default ZFS block size is 128K. If you have a filesystem with 128G used, that means you

Re: [zfs-discuss] block align SSD for use as a l2arc cache

2010-07-09 Thread Geoff Nordli
-Original Message- From: Erik Trimble Sent: Friday, July 09, 2010 6:45 PM Subject: Re: [zfs-discuss] block align SSD for use as a l2arc cache On 7/9/2010 5:55 PM, Geoff Nordli wrote: I have an Intel X25-M 80GB SSD. For optimum performance, I need to block align the SSD

[zfs-discuss] Scrub extremely slow?

2010-07-09 Thread Hernan F
Hello, I'm trying to figure out why I'm getting about 10MB/s scrubs, on a pool where I can easily get 100MB/s. It's 4x 1TB SATA2 (nv_sata), raidz. Athlon64 with 8GB RAM. Here's the output while I cat an 8GB file to /dev/null r...@solaris:~# zpool iostat 20 capacity