Re: [zfs-discuss] Data transfer taking a longer time than expected (Possibly dedup related)

2010-10-04 Thread Ville Ojamo
The article would probably be correct. In my experience, and looking at the
archives for other posts, dedup really needs the RAM and preferably L2ARC
device as well. As someone else put it, home servers need not apply.

I would point out the very slow dataset destroy caveat depending on which
build you are using. At least until and including b134 destroying a dataset
that had at some point dedup turned on, with low memory, results in _very_
lengthy operation. And prepare to give it a long time (days?). If you
absolutely must shut down the system during it, the next restart will be
painfully slow (think week or so). There has been few posts about this,
including mine. I might be mistaken but someone explained it might have been
due to the dataset destroy operation running along with resilver.


-V


From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Tom
Sent: 25 September 2010 04:40
To: David Blasingame Oracle
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Data transfer taking a longer time than expected
(Possibly dedup related)

Thanks a lot for that. I'm not experienced in reading the output of dtrace,
but I'm pretty sure that dedup was the cause here, as I disabling it during
the transfer, immediately raised the transfer speed to ~100MB/s.

Thanks for the article you linked to — it seems my system would need about
16GB RAM for dedup to work smoothly in my case...


On Fri, Sep 24, 2010 at 11:10 PM, David Blasingame Oracle
david.blasing...@oracle.com wrote:
How do you know it is dedup causing the problem?

You can check to see how much is by looking at the threads (look for ddt)

mdb -k 

::threadlist -v

or dtrace it.

fbt:zfs:ddt*:entry

You can disable dedup.  I believe current dedup data stays until it gets
over written.  I'm not sure what send would do, but I would assume the new
filesystem if dedup is not enabled would not have dedup'd data.  

You might also want to read.

http://blogs.sun.com/roch/entry/dedup_performance_considerations1

As far as the impact of ctrl-c on a move operation, When I do a test to
move a file from one file system to another an ctrl-c the operation, the
file is intact on the original filesystem and on the new filesystem it is
partial.  So you would have to be careful about which data has already been
copied.

Dave

On 09/24/10 14:34, Thomas S. wrote: 
Hi all

I'm currently moving a fairly big dataset (~2TB) within the same zpool. Data
is being moved from a dataset to another, which has dedup enabled.

The transfer started at quite a slow transfer speed — maybe 12MB/s. But it
is now crawling to a near halt. Only 800GB has been moved in 48 hours.

I looked for similar problems on the forums and other places, and it seems
dedup needs a much bigger amount of RAM than the server currently has (3GB),
to perform smoothly for such an operation.

My question is, how can I gracefully stop the ongoing operation? What I did
was simply mv temp/* new/ in an ssh session (which is still open).

Can I disable dedup on the dataset while the transfer is going on? Can I
simply Ctrl-C the procress to stop it? Shoul I be careful of anything?

Help would be appreciated
  

-- 





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Stephan Budach
Hi,

once I created a zpool of single vdevs not using mirroring of any kind. Now I 
wonder if it's possible to add vdevs and mirror the currently existing ones.

Thanks,
budy
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Any way for snapshot zpool with RAIDZ or for independent devices

2010-10-04 Thread sridhar surampudi
Hi,
With recent additions, using zpool split I could split a mirrored zpool and 
create new pool with the given name. 

Is there any direct or indirect mechanism where I can create snapshot of 
devices under a existing zpool where new devices are created so that I can 
recreate stack (new zpool and all file systems) without modifying the data.

So that with new stack i can able to access consistent (snapshot ) data (which 
is same as data present on original when backup took) with the new stack 
created.


Thanks  Regards,
sridhar.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] hot spare remains in use

2010-10-04 Thread Brian Kolaci
Hi,

I had a hot spare used to replace a failed drive, but then the drive appears to 
be fine anyway.
After clearing the error it shows that the drive was resilvered, but keeps the 
spare in use.

zpool status pool2
 pool: pool2
state: ONLINE
scrub: none requested
config:

   NAMESTATE READ WRITE CKSUM
   pool2   ONLINE   0 0 0
 raidz2ONLINE   0 0 0
   c10t8d0 ONLINE   0 0 0
   c10t9d0 ONLINE   0 0 0
   c10t10d0ONLINE   0 0 0
   spare   ONLINE   0 0 0
 c10t11d0  ONLINE   0 0 0
 c10t22d0  ONLINE   0 0 0
   c10t12d0ONLINE   0 0 0
   c10t13d0ONLINE   0 0 0
   c10t14d0ONLINE   0 0 0
   spares
 c10t22d0  INUSE currently in use

errors: No known data errors

How can I get the spare out of the pool?

Thanks,

Brian

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] hot spare remains in use

2010-10-04 Thread Cindy Swearingen

Hi Brian,

You could manually detach the spare, like this:

# zpool detach pool2 c10t22d0

Sometimes, you might need to clear the pool error but I don't
see any residual errors in this output:

# zpool clear pool2

I would use fmdump -eV to see what's going with c10t11d0.

Thanks,

Cindy

On 10/04/10 07:47, Brian Kolaci wrote:

Hi,

I had a hot spare used to replace a failed drive, but then the drive appears to 
be fine anyway.
After clearing the error it shows that the drive was resilvered, but keeps the 
spare in use.

zpool status pool2
 pool: pool2
state: ONLINE
scrub: none requested
config:

   NAMESTATE READ WRITE CKSUM
   pool2   ONLINE   0 0 0
 raidz2ONLINE   0 0 0
   c10t8d0 ONLINE   0 0 0
   c10t9d0 ONLINE   0 0 0
   c10t10d0ONLINE   0 0 0
   spare   ONLINE   0 0 0
 c10t11d0  ONLINE   0 0 0
 c10t22d0  ONLINE   0 0 0
   c10t12d0ONLINE   0 0 0
   c10t13d0ONLINE   0 0 0
   c10t14d0ONLINE   0 0 0
   spares
 c10t22d0  INUSE currently in use

errors: No known data errors

How can I get the spare out of the pool?

Thanks,

Brian

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Darren J Moffat

On 04/10/2010 15:24, Stephan Budach wrote:

once I created a zpool of single vdevs not using mirroring of any kind. Now I 
wonder if it's possible to add vdevs and mirror the currently existing ones.


Yes.

zpool attach pool existing_device new_device

Do that for each of the vdev devices you are mirroring.

--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] tagged ACL groups: let's just keep digging until we come out the other side

2010-10-04 Thread Nicolas Williams
On Thu, Sep 30, 2010 at 08:14:24PM -0400, Miles Nordin wrote:
  Can the user in (3) fix the permissions from Windows?
 
 no, not under my proposal.

Let's give it a whirld anyways:

 but it sounds like currently people cannot ``fix'' permissions through
 the quirky autotranslation anyway, certainly not to the point where
 neither unix nor windows users are confused: windows users are always
 confused, and unix users don't get to see all the permissions.

No, that's not right.  Today you can fix permissions from any NFSv4
client that exports an NFSv4-style ACL interface to users.  You can fix
permissions from Windows.  You can fix permissions a local Solaris
shell.  You can also fix permissions from NFSv3 clients (but you get
POSIX Draft - ZFS translated ACLs, which are confusing because they
tend to result in DENY ACEs being scattered all over).  You can also
chmod, but you lose your ACL if you do that.

  Now what?
 
 set the unix perms to 777 as a sign to the unix people to either (a)
 leave it alone, or (b) learn to use 'chmod A...'.  This will actually
 work: it's not a hand-waving hypothetical that just doesn't play out.

I would think that 777 would invite chmods.  I think you are handwaving.

 What I provide, which we don't have now, is a way to make:
 
   /tub/dataset/a subtree
 
 -rwxrwxrwxin old unix
 [working, changeable permissions] in windows
 
   /tub/dataset/b subtree
 
 -rw-r--r--in old unix
 [everything: everyone]in windows, but unix permissions 
   still enforced
 
 this means:
 
  * unix writers and windows writers can cooperate even within a single
dataset
 
  * an intuitive warning sign when non-native permissions are in effect, 
 
  * fewer leaked-data surprises

I don't understand what exactly you're proposing.  You've not said
anything about how chmod is to be handled.

 If you accept that the autotranslation between the two permissions
 regimes is total shit, which it is, then what I offer is the best oyu
 can hope for.

If I could understand what you're proposing I might agree, who knows.
But I do think there's other possibilities, some probably better than
what you propose (whatever that is).

Here's a crazy alternative that might work (or not): allow users to
pre-configure named ACLs where the names are {owner, group, mode}.
E.g., we could have:

.zfs/ACLs/user/[group:][d|-]permissions[.inherit]
^   ^^^  ^
||   |
+-- owned by |   |
user   +-- applies to  |
 directory   |
 or other|
 objects |
 |
see below

When chmod()ing an object... ZFS would search for the most specific
matching file in .zfs/ACLs/ and, if found, would replace the chmod()ed
object's ACL with that of the .zfs/ACLs/... file found.  The .inherit
suffix would indicate that if the chmod() target's parent directory has
inherittable ACEs then they will be groupmasked and added to the ACEs
from the .zfs/ACLs/... file to produce a final ACL.

E.g., a chmod(0644) of /a/b/c/foo (say, a file owned by 'joe' with group
'staff', with /, /a, /a/b, and /a/b/c all being datasets), where c has
inherittable ACEs would cause ZFS to search for
.zfs/ACLs/joe/staff:-rw-r--r--.inherit, .zfs/ACLs/joe/-rw-r--r--.inherit, 
zfs/ACLs/joe/staff:-rw-r--r--, and .zfs/ACLs/joe/-rw-r--r--, first in
/a/b/c, then /a/b, then /a, then /.

I said this is crazy.  Is it?  I think it probably is.  This would
almost certainly prove to be a hard-to-use design.  Users would need to
be educated in order to not be surprised...  OTOH, it puts much more
control in the hands of the user.  These named ACLs could be inheritted
from parent datasets as a way to avoid having to set them up too many
times.  And with the .inherit twist it probably has enough granularity
of control to be useful (particularly if users are dataset-happy).
Finally, these could even be managed remotely.

I see zero chance of such a design being adopted.

It'd be better, IMO, to go for non-POSIX-equivalent groupmasking and
translations of POSIX mode_t and POSIX Draft ACLs to ZFS ACLs.  For
example: take the current translations, remove all owner@ and group DENY
ACEs, then sort any remaining user DENY ACEs to be first, and any
everyone@ DENY ACEs to be last.  The results would surely be surprising
to some users, but the kinds of mode_t and POSIX Draft ACLs where
surprise is likely are rare.

That's two alternatives right there.

Nico
-- 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org

Re: [zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Stephan Budach
Hi Darren,

gee, thanks. Of course the would be a resilver due for each vdev, but that 
shouldn't harm, although the vdevs are quite big.

Thanks,
budy
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Cindy Swearingen

Hi--

Yes, you would use the zpool attach command to convert a
non-redundant configuration into a mirrored pool configuration.

http://docs.sun.com/app/docs/doc/819-5461/gcfhe?l=ena=view

See:

Example 4–6 Converting a Nonredundant ZFS Storage Pool to a Mirrored ZFS 
Storage Pool


If you have more than one device in the pool, you would continue to
attach a new disk to each existing device, like this:

# zpool status test
  pool: test
 state: ONLINE
 scan: resilvered 85.5K in 0h0m with 0 errors on Mon Oct  4 08:44:35 2010
config:

NAMESTATE READ WRITE CKSUM
testONLINE   0 0 0
  mirror-0  ONLINE   0 0 0
c3t1d0  ONLINE   0 0 0
c3t2d0  ONLINE   0 0 0
c3t3d0  ONLINE   0 0 0

# zpool attach test c3t1d0 c4t1d0
# zpool attach test c3t2d0 c4t2d0
# zpool attach test c3t3d0 c4t3d0

This would create a mirrored pool with 3 two-way mirrors.

I would suggest attaching one disk at a time, letting it
resilver and then run a scrub to ensure that each new disk is
functional.

Thanks,

Cindy

On 10/04/10 08:24, Stephan Budach wrote:/dev/dsk/c2t5d0s2

Hi,

once I created a zpool of single vdevs not using mirroring of any kind. Now I 
wonder if it's possible to add vdevs and mirror the currently existing ones.

Thanks,
budy

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Darren J Moffat

On 04/10/2010 15:48, Stephan Budach wrote:

gee, thanks. Of course the would be a resilver due for each vdev, but that 
shouldn't harm, although the vdevs are quite big.


Of course there will be a resilver, otherwise the mirror won't get the 
existing data and it wouldn't be a mirror.


--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Cindy Swearingen

Duh. Yeah, its Monday morning.

I didn't have 6 devices to play with so I tried to fake it.
It should look like this:

# zpool status test
  pool: test
  state: ONLINE
  scan: resilvered 85.5K in 0h0m with 0 errors on Mon Oct  4 08:54:06
2010
 config:

  NAME  STATE READ WRITE CKSUM
  test  ONLINE   0 0 0
c3t1d0  ONLINE   0 0 0
c3t2d0  ONLINE   0 0 0
c3t3d0  ONLINE   0 0 0
# zpool attach test c3t1d0 c4t1d0
# zpool attach test c3t2d0 c4t2d0
# zpool attach test c3t3d0 c4t3d0

Cindy

On 10/04/10 09:05, Stephan Budach wrote:

Hi Cindy,

very well - thanks.

I noticed that either the pool you're using and the zpool that is 
described inb the docs already show a mirror-0 configuration, which 
isn't the case for my zpool:


zpool status obelixData
  pool: obelixData
 state: ONLINE
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM
obelixData   ONLINE   0 0 0
  c4t21D023038FA8d0  ONLINE   0 0 0
  c4t21D02305FF42d0  ONLINE   0 0 0

errors: No known data errors

Actually, this zpool consists of two FC raids and I think I created it 
simply by adding these two devs to the pool.

Does this disqualify my zpool for upgrading?

Thanks,
budy


Am 04.10.10 16:48, schrieb Cindy Swearingen:

Hi--

Yes, you would use the zpool attach command to convert a
non-redundant configuration into a mirrored pool configuration.

http://docs.sun.com/app/docs/doc/819-5461/gcfhe?l=ena=view

See:

Example 4–6 Converting a Nonredundant ZFS Storage Pool to a Mirrored 
ZFS Storage Pool


If you have more than one device in the pool, you would continue to
attach a new disk to each existing device, like this:

# zpool status test
  pool: test
 state: ONLINE
 scan: resilvered 85.5K in 0h0m with 0 errors on Mon Oct  4 08:44:35 2010
config:

NAMESTATE READ WRITE CKSUM
testONLINE   0 0 0
  mirror-0  ONLINE   0 0 0
c3t1d0  ONLINE   0 0 0
c3t2d0  ONLINE   0 0 0
c3t3d0  ONLINE   0 0 0

# zpool attach test c3t1d0 c4t1d0
# zpool attach test c3t2d0 c4t2d0
# zpool attach test c3t3d0 c4t3d0

This would create a mirrored pool with 3 two-way mirrors.

I would suggest attaching one disk at a time, letting it
resilver and then run a scrub to ensure that each new disk is
functional.

Thanks,

Cindy

On 10/04/10 08:24, Stephan Budach wrote:/dev/dsk/c2t5d0s2

Hi,

once I created a zpool of single vdevs not using mirroring of any 
kind. Now I wonder if it's possible to add vdevs and mirror the 
currently existing ones.


Thanks,
budy




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs send|recv and inherited recordsize

2010-10-04 Thread Robert Milkowski

Hi,

I thought that if I use zfs send snap | zfs recv  if on a receiving 
side the recordsize property is set to different value it will be 
honored. But it doesn't seem to be the case, at least on snv_130.



 $ zfs get recordsize test/m1
NAME PROPERTYVALUESOURCE
test/m1  recordsize  128K default

 $ ls -nil /test/m1/f1
 5 -rw-r--r--   1 011048576 Oct  4 10:31 
/test/m1/f1


 $ zdb -vv test/m1 5
Dataset test/m1 [ZPL], ID 1082, cr_txg 33413, 1.02M, 5 objects

Object  lvl   iblk   dblk  dsize  lsize   %full  type
 5216K   128K  1.00M 1M  100.00  ZFS plain file

 $ zfs snapshot test/m...@s1
 $ zfs create -o recordsize=32k test/m2
 $ zfs send test/m...@s1 | zfs recv test/m2/m1



 $ zfs get recordsize test/m2/m1
NAMEPROPERTYVALUESOURCE
test/m2/m1  recordsize  32K  inherited from test/m2

 $ ls -lni /test/m2/m1/f1
 5 -rw-r--r--   1 011048576 Oct  4 10:31 
/test/m2/m1/f1


 $ zdb -vv test/m2/m1 5
Dataset test/m2/m1 [ZPL], ID 1110, cr_txg 33537, 1.02M, 5 objects

Object  lvl   iblk   dblk  dsize  lsize   %full  type
 5216K   128K  1.00M 1M  100.00  ZFS plain file


Well, dblk is 128KB - I would expect it to be 32K.
Lets see what happens if I use cp instead:


 $ cp /test/m2/m1/f1 /test/m2/m1/f2
 $ ls -lni /test/m2/m1/f2
 6 -rw-r--r--   1 011048576 Oct  4 11:15 
/test/m2/m1/f2


 $ zdb -vv test/m2/m1 6
Dataset test/m2/m1 [ZPL], ID 1110, cr_txg 33537, 2.03M, 6 objects

Object  lvl   iblk   dblk  dsize  lsize   %full  type
 6216K32K  1.00M 1M  100.00  ZFS plain file


Now it is fine.

--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating to an aclmode-less world

2010-10-04 Thread Simon Breden
Any ideas anyone?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can I upgrade a striped pool of vdevs to mirrored vdevs?

2010-10-04 Thread Cindy Swearingen

To answer your other questions,

I'm not sure I'm following your FC raid description:

Are you saying you created two LUNs from a FC RAID array
and added them to the pool?

If so, then yes, you can still attach more LUNs from the
array to create a mirrored pool.

A best practice is to mirror across controllers for better
reliability, but ZFS doesn't check if the disks to attach
are from the same array, if that's what you mean.

Thanks,

Cindy

On 10/04/10 09:05, Stephan Budach wrote:

Hi Cindy,

very well - thanks.

I noticed that either the pool you're using and the zpool that is 
described inb the docs already show a mirror-0 configuration, which 
isn't the case for my zpool:


zpool status obelixData
  pool: obelixData
 state: ONLINE
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM
obelixData   ONLINE   0 0 0
  c4t21D023038FA8d0  ONLINE   0 0 0
  c4t21D02305FF42d0  ONLINE   0 0 0

errors: No known data errors

Actually, this zpool consists of two FC raids and I think I created it 
simply by adding these two devs to the pool.

Does this disqualify my zpool for upgrading?

Thanks,
budy


Am 04.10.10 16:48, schrieb Cindy Swearingen:

Hi--

Yes, you would use the zpool attach command to convert a
non-redundant configuration into a mirrored pool configuration.

http://docs.sun.com/app/docs/doc/819-5461/gcfhe?l=ena=view

See:

Example 4–6 Converting a Nonredundant ZFS Storage Pool to a Mirrored 
ZFS Storage Pool


If you have more than one device in the pool, you would continue to
attach a new disk to each existing device, like this:

# zpool status test
  pool: test
 state: ONLINE
 scan: resilvered 85.5K in 0h0m with 0 errors on Mon Oct  4 08:44:35 2010
config:

NAMESTATE READ WRITE CKSUM
testONLINE   0 0 0
  mirror-0  ONLINE   0 0 0
c3t1d0  ONLINE   0 0 0
c3t2d0  ONLINE   0 0 0
c3t3d0  ONLINE   0 0 0

# zpool attach test c3t1d0 c4t1d0
# zpool attach test c3t2d0 c4t2d0
# zpool attach test c3t3d0 c4t3d0

This would create a mirrored pool with 3 two-way mirrors.

I would suggest attaching one disk at a time, letting it
resilver and then run a scrub to ensure that each new disk is
functional.

Thanks,

Cindy

On 10/04/10 08:24, Stephan Budach wrote:/dev/dsk/c2t5d0s2

Hi,

once I created a zpool of single vdevs not using mirroring of any 
kind. Now I wonder if it's possible to add vdevs and mirror the 
currently existing ones.


Thanks,
budy




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs volume snapshot

2010-10-04 Thread Wei Li
Hi All,

If a ZFS volume is presented to LDOM guest domain as whole disk (used as root 
disk), does anyone know how to snapshot it?  It is something like how to 
snapshot zfs raw volume (NOTE, no ufs file system directly created on ZFS 
volume in above case)).  

Thanks!
Wei
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Unexpected ZFS space consumption

2010-10-04 Thread Glenn Lagasse
* JR Dalrymple (j...@jrssite.com) wrote:
 I'm pretty new to ZFS and OpenSolaris as a whole. I am an experienced
 storage administrator, however my storage equipment has typically been
 NetApp or EMC branded. I administer NetApp FAS2000 and FAS3000 series
 boxes to host a VMware only virtual infrastructure so I am versed on a
 pretty high level at storage provisioning for a virtual environment.
 
 My problem is unexpected disk usage on deduplicated datasets holding
 little more than VMDKs. I experimented with deduplication on ZFS and
 compared it to deduplication on NetApp and found basically identical
 returns on a mix of backup data and user data. I was pretty excited to
 put some VMDKs of my own on to a system of my own. I have been
 disappointed with the actual results however :(
 
 Upon building VMs on this storage I found the data to consume an as
 expected OS only amount of disk space. As time went on the VMDKs
 filled out to consume their entire allocated disk space. I was hoping
 I could recover the lost physical disk space by using sdelete on the
 guests to zero out unused space on the disks, however this didn't
 happen as per du or df on the storage host. After zeroing unused disk
 space I was really hoping that the VMDKs would only consume the amount
 of disk actually filled by the guest as they did when the VMs were
 fresh. I have properly aligned VMDKs so I don't think that the problem
 lies there.
 
 I'm not sure what information to offer that might be helpful except
 the following (nfs0 is the dataset I'm working with primarily):
 
 jrdal...@yac-stor1:~$ uname -a SunOS yac-stor1 5.11 snv_134 i86pc i386
 i86pc Solaris jrdal...@yac-stor1:~$ zpool list NAMESIZE  ALLOC
 FREECAP  DEDUP  HEALTH  ALTROOT rpool   540G   228G   312G42%
 1.26x  ONLINE  - jrdal...@yac-stor1:~$ zfs list NAME
 USED  AVAIL  REFER  MOUNTPOINT rpool329G
 238G  88.5K  /rpool rpool/ROOT  7.97G   238G19K
 legacy rpool/ROOT/opensolaris  8.41M   238G  2.85G  /
 rpool/ROOT/opensolaris-143.5M   238G  3.88G  /
 rpool/ROOT/opensolaris-27.92G   238G  5.52G  / rpool/dump
 2.00G   238G  2.00G  - rpool/export1.03G   238G23K
 /export rpool/export/home   1.03G   238G23K  /export/home
 rpool/export/home/jrdalrym  1.03G   238G  1.03G  /export/home/jrdalrym
 rpool/iscsi  103G   238G21K  /rpool/iscsi
 rpool/iscsi/iscsi0   103G   301G  40.5G  - rpool/nfs0
 153G  87.3G   153G  /rpool/nfs0 rpool/nfs1  49.6G
 238G  40.5G  /rpool/nfs1 rpool/nfs2  9.99G  50.0G
 9.94G  /rpool/nfs2 rpool/swap  2.00G   240G   100M  -
 jrdal...@yac-stor1:~$ zfs get all rpool/nfs0 NAMEPROPERTY
 VALUE   SOURCE rpool/nfs0  type
 filesystem  - rpool/nfs0  creation
 Wed Aug 25 20:28 2010   - rpool/nfs0  used
 153G- rpool/nfs0  available
 87.3G   - rpool/nfs0  referenced
 153G- rpool/nfs0  compressratio
 1.00x   - rpool/nfs0  mounted
 yes - rpool/nfs0  quota
 240Glocal rpool/nfs0  reservation
 nonedefault rpool/nfs0  recordsize
 128Kdefault rpool/nfs0  mountpoint
 /rpool/nfs0 default rpool/nfs0  sharenfs
 ro...@192.168.10.0/24   local rpool/nfs0  checksum
 on  default rpool/nfs0  compression
 off default rpool/nfs0  atime
 on  default rpool/nfs0  devices
 on  default rpool/nfs0  exec
 on  default rpool/nfs0  setuid
 on  default rpool/nfs0  readonly
 off default rpool/nfs0  zoned
 off default rpool/nfs0  snapdir
 hidden  default rpool/nfs0  aclmode
 groupmask   default rpool/nfs0  aclinherit
 restricted  default rpool/nfs0  canmount
 on  default rpool/nfs0  shareiscsi
 off default rpool/nfs0  xattr
 on  default rpool/nfs0  copies
 1   default rpool/nfs0  version
 4   - rpool/nfs0  utf8only
 off - rpool/nfs0  normalization
 none- rpool/nfs0  casesensitivity
 sensitive   - rpool/nfs0  vscan
 off default rpool/nfs0  nbmand
 off default rpool/nfs0  sharesmb
 off default rpool/nfs0  refquota
 nonedefault rpool/nfs0  refreservation
 nonedefault rpool/nfs0  primarycache
 all  

[zfs-discuss] When is it okay to turn off the verify option.

2010-10-04 Thread Peter Taps
Folks,

As I understand, the hash generated by sha256 is almost guaranteed not to 
collide. I am thinking it is okay to turn off verify property on the zpool. 
However, if there is indeed a collision, we lose data. Scrub cannot recover 
such lost data.

I am wondering in real life when is it okay to turn off verify option? I 
guess for storing business critical data (HR, finance, etc.), you cannot afford 
to turn this option off. 

Thank you in advance for your help.

Regards,
Peter
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] hot spare remains in use

2010-10-04 Thread Brian Kolaci
Thanks, that did it.  I thought detach was only for mirrors and I have a 
raidz2, so I didn't think to use that there.  I tried replace/remove.
I guess the spare is actually a mirror of the disk and the spare disk and is 
treated as such.

Thanks again,

Brian

On Oct 4, 2010, at 10:27 AM, Cindy Swearingen wrote:

 Hi Brian,
 
 You could manually detach the spare, like this:
 
 # zpool detach pool2 c10t22d0
 
 Sometimes, you might need to clear the pool error but I don't
 see any residual errors in this output:
 
 # zpool clear pool2
 
 I would use fmdump -eV to see what's going with c10t11d0.
 
 Thanks,
 
 Cindy
 
 On 10/04/10 07:47, Brian Kolaci wrote:
 Hi,
 I had a hot spare used to replace a failed drive, but then the drive appears 
 to be fine anyway.
 After clearing the error it shows that the drive was resilvered, but keeps 
 the spare in use.
 zpool status pool2
 pool: pool2
 state: ONLINE
 scrub: none requested
 config:
   NAMESTATE READ WRITE CKSUM
   pool2   ONLINE   0 0 0
 raidz2ONLINE   0 0 0
   c10t8d0 ONLINE   0 0 0
   c10t9d0 ONLINE   0 0 0
   c10t10d0ONLINE   0 0 0
   spare   ONLINE   0 0 0
 c10t11d0  ONLINE   0 0 0
 c10t22d0  ONLINE   0 0 0
   c10t12d0ONLINE   0 0 0
   c10t13d0ONLINE   0 0 0
   c10t14d0ONLINE   0 0 0
   spares
 c10t22d0  INUSE currently in use
 errors: No known data errors
 How can I get the spare out of the pool?
 Thanks,
 Brian
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] hot spare remains in use

2010-10-04 Thread Cindy Swearingen

Hi Brian,

Yes, the current wording around detaching spares is kind of confusing
for RAIDZ configurations. I will fix that shortly.

Thanks,

Cindy

On 10/04/10 11:43, Brian Kolaci wrote:

Thanks, that did it.  I thought detach was only for mirrors and I have a 
raidz2, so I didn't think to use that there.  I tried replace/remove.
I guess the spare is actually a mirror of the disk and the spare disk and is 
treated as such.

Thanks again,

Brian

On Oct 4, 2010, at 10:27 AM, Cindy Swearingen wrote:


Hi Brian,

You could manually detach the spare, like this:

# zpool detach pool2 c10t22d0

Sometimes, you might need to clear the pool error but I don't
see any residual errors in this output:

# zpool clear pool2

I would use fmdump -eV to see what's going with c10t11d0.

Thanks,

Cindy

On 10/04/10 07:47, Brian Kolaci wrote:

Hi,
I had a hot spare used to replace a failed drive, but then the drive appears to 
be fine anyway.
After clearing the error it shows that the drive was resilvered, but keeps the 
spare in use.
zpool status pool2
pool: pool2
state: ONLINE
scrub: none requested
config:
  NAMESTATE READ WRITE CKSUM
  pool2   ONLINE   0 0 0
raidz2ONLINE   0 0 0
  c10t8d0 ONLINE   0 0 0
  c10t9d0 ONLINE   0 0 0
  c10t10d0ONLINE   0 0 0
  spare   ONLINE   0 0 0
c10t11d0  ONLINE   0 0 0
c10t22d0  ONLINE   0 0 0
  c10t12d0ONLINE   0 0 0
  c10t13d0ONLINE   0 0 0
  c10t14d0ONLINE   0 0 0
  spares
c10t22d0  INUSE currently in use
errors: No known data errors
How can I get the spare out of the pool?
Thanks,
Brian
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Please warn a home user against OpenSolaris under VirtualBox under WinXP ; )

2010-10-04 Thread Nils
Well at the risk of being repetetive too: 
or another box.

So yes I am considering it, but that is probably the option that requires less 
guidance in this thread.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send|recv and inherited recordsize

2010-10-04 Thread Matthew Ahrens
That's correct.

This behavior is because the send|recv operates on the DMU objects,
whereas the recordsize property is interpreted by the ZPL.  The ZPL
checks the recordsize property when a file grows.  But the recv
doesn't grow any files, it just dumps data into the underlying
objects.

--matt

On Mon, Oct 4, 2010 at 11:20 AM, Robert Milkowski mi...@task.gda.pl wrote:
 Hi,

 I thought that if I use zfs send snap | zfs recv  if on a receiving side
 the recordsize property is set to different value it will be honored. But it
 doesn't seem to be the case, at least on snv_130.


  $ zfs get recordsize test/m1
 NAME     PROPERTY    VALUE    SOURCE
 test/m1  recordsize  128K     default

  $ ls -nil /test/m1/f1
         5 -rw-r--r--   1 0        1        1048576 Oct  4 10:31 /test/m1/f1

  $ zdb -vv test/m1 5
 Dataset test/m1 [ZPL], ID 1082, cr_txg 33413, 1.02M, 5 objects

    Object  lvl   iblk   dblk  dsize  lsize   %full  type
         5    2    16K   128K  1.00M     1M  100.00  ZFS plain file

  $ zfs snapshot test/m...@s1
  $ zfs create -o recordsize=32k test/m2
  $ zfs send test/m...@s1 | zfs recv test/m2/m1



  $ zfs get recordsize test/m2/m1
 NAME        PROPERTY    VALUE    SOURCE
 test/m2/m1  recordsize  32K      inherited from test/m2

  $ ls -lni /test/m2/m1/f1
         5 -rw-r--r--   1 0        1        1048576 Oct  4 10:31
 /test/m2/m1/f1

  $ zdb -vv test/m2/m1 5
 Dataset test/m2/m1 [ZPL], ID 1110, cr_txg 33537, 1.02M, 5 objects

    Object  lvl   iblk   dblk  dsize  lsize   %full  type
         5    2    16K   128K  1.00M     1M  100.00  ZFS plain file


 Well, dblk is 128KB - I would expect it to be 32K.
 Lets see what happens if I use cp instead:


  $ cp /test/m2/m1/f1 /test/m2/m1/f2
  $ ls -lni /test/m2/m1/f2
         6 -rw-r--r--   1 0        1        1048576 Oct  4 11:15
 /test/m2/m1/f2

  $ zdb -vv test/m2/m1 6
 Dataset test/m2/m1 [ZPL], ID 1110, cr_txg 33537, 2.03M, 6 objects

    Object  lvl   iblk   dblk  dsize  lsize   %full  type
         6    2    16K    32K  1.00M     1M  100.00  ZFS plain file


 Now it is fine.

 --
 Robert Milkowski
 http://milek.blogspot.com

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs send|recv and inherited recordsize

2010-10-04 Thread Robert Milkowski


thank you.

On 04/10/2010 19:55, Matthew Ahrens wrote:

That's correct.

This behavior is because the send|recv operates on the DMU objects,
whereas the recordsize property is interpreted by the ZPL.  The ZPL
checks the recordsize property when a file grows.  But the recv
doesn't grow any files, it just dumps data into the underlying
objects.

--matt

On Mon, Oct 4, 2010 at 11:20 AM, Robert Milkowskimi...@task.gda.pl  wrote:
   

Hi,

I thought that if I use zfs send snap | zfs recv  if on a receiving side
the recordsize property is set to different value it will be honored. But it
doesn't seem to be the case, at least on snv_130.


  $ zfs get recordsize test/m1
NAME PROPERTYVALUESOURCE
test/m1  recordsize  128K default

  $ ls -nil /test/m1/f1
 5 -rw-r--r--   1 011048576 Oct  4 10:31 /test/m1/f1

  $ zdb -vv test/m1 5
Dataset test/m1 [ZPL], ID 1082, cr_txg 33413, 1.02M, 5 objects

Object  lvl   iblk   dblk  dsize  lsize   %full  type
 5216K   128K  1.00M 1M  100.00  ZFS plain file

  $ zfs snapshot test/m...@s1
  $ zfs create -o recordsize=32k test/m2
  $ zfs send test/m...@s1 | zfs recv test/m2/m1



  $ zfs get recordsize test/m2/m1
NAMEPROPERTYVALUESOURCE
test/m2/m1  recordsize  32K  inherited from test/m2

  $ ls -lni /test/m2/m1/f1
 5 -rw-r--r--   1 011048576 Oct  4 10:31
/test/m2/m1/f1

  $ zdb -vv test/m2/m1 5
Dataset test/m2/m1 [ZPL], ID 1110, cr_txg 33537, 1.02M, 5 objects

Object  lvl   iblk   dblk  dsize  lsize   %full  type
 5216K   128K  1.00M 1M  100.00  ZFS plain file


Well, dblk is 128KB - I would expect it to be 32K.
Lets see what happens if I use cp instead:


  $ cp /test/m2/m1/f1 /test/m2/m1/f2
  $ ls -lni /test/m2/m1/f2
 6 -rw-r--r--   1 011048576 Oct  4 11:15
/test/m2/m1/f2

  $ zdb -vv test/m2/m1 6
Dataset test/m2/m1 [ZPL], ID 1110, cr_txg 33537, 2.03M, 6 objects

Object  lvl   iblk   dblk  dsize  lsize   %full  type
 6216K32K  1.00M 1M  100.00  ZFS plain file


Now it is fine.

--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 
   


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] When is it okay to turn off the verify option.

2010-10-04 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Peter Taps
 
 As I understand, the hash generated by sha256 is almost guaranteed
 not to collide. I am thinking it is okay to turn off verify property
 on the zpool. However, if there is indeed a collision, we lose data.
 Scrub cannot recover such lost data.
 
 I am wondering in real life when is it okay to turn off verify
 option? I guess for storing business critical data (HR, finance, etc.),
 you cannot afford to turn this option off.

Right on all points.  It's a calculated risk.  If you have a hash collision,
you will lose data undetected, and backups won't save you unless *you* are
the backup.  That is, if the good data, before it got corrupted by your
system, happens to be saved somewhere else before it reached your system.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] When is it okay to turn off the verify option.

2010-10-04 Thread Scott Meilicke
Why do you want to turn verify off? If performance is the reason, is it 
significant, on and off?

On Oct 4, 2010, at 2:28 PM, Edward Ned Harvey wrote:

 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Peter Taps
 
 As I understand, the hash generated by sha256 is almost guaranteed
 not to collide. I am thinking it is okay to turn off verify property
 on the zpool. However, if there is indeed a collision, we lose data.
 Scrub cannot recover such lost data.
 
 I am wondering in real life when is it okay to turn off verify
 option? I guess for storing business critical data (HR, finance, etc.),
 you cannot afford to turn this option off.
 
 Right on all points.  It's a calculated risk.  If you have a hash collision,
 you will lose data undetected, and backups won't save you unless *you* are
 the backup.  That is, if the good data, before it got corrupted by your
 system, happens to be saved somewhere else before it reached your system.
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Scott Meilicke



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS crypto bug status change

2010-10-04 Thread David Magda
Seems that the bug for ZFS data set encryption is now in a state of  
10-Fix Delivered:


http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=4854202

Via:

http://sparcv9.blogspot.com/2010/10/zfs-crypto-integrated.html

Thank you Mr. Moffat et al. Hopefully the rest of us will be able to  
bang on this at some point. :)


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating to an aclmode-less world

2010-10-04 Thread Cindy Swearingen

Hi Simon,

I don't think you will see much difference for these reasons:

1. The CIFS server ignores the aclinherit/aclmode properties.

2. Your aclinherit=passthrough setting overrides the aclmode
property anyway.

3. The only difference is that if you use chmod on these files
to manually change the permissions, you will lose the ACL values.

Thanks,

Cindy

On 09/29/10 13:09, Simon Breden wrote:
Currently I'm still using OpenSolaris b134 and I had used the 'aclmode' property on my file systems. However, the aclmode property has been dropped now: http://arc.opensolaris.org/caselog/PSARC/2010/029/20100126_mark.shellenbaum 


I'm wondering what will happen to the ACLs on these files and directories if I 
upgrade to a newer Solaris version (OpenIndiana b147 perhaps).

I'm sharing the file systems using CIFS.

I was using very simple ACLs like below for easy inheritance of ACLs, which 
worked OK for my needs.

# zfs set aclinherit=passthrough tank/home/fred/projects
# zfs set aclmode=passthrough tank/home/fred/projects
# chmod A=\
owner@:rwxpdDaARWcCos:fd-:allow,\
group@:rwxpdDaARWcCos:fd-:allow,\
everyone@:rwxpdDaARWcCos:fd-:deny \
/tank/home/fred/projects
# chown fred:fred /tank/home/fred/projects
# zfs set sharesmb=name=projects tank/home/fred/projects

Cheers,
Simon

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] When is it okay to turn off the verify option.

2010-10-04 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Scott Meilicke
 
 Why do you want to turn verify off? If performance is the reason, is it
 significant, on and off?

Under most circumstances, verify won't hurt performance.  It won't hurt
reads of any kind, and it won't hurt writes when you're writing unique data,
or if you're writing duplicate data which is warm in the read cache.  

It will basically hurt write performance if you are writing duplicate data,
which was not read recently.  This might be the case, for example, if this
machine is the target for some remote machine to backup onto.

The problem doesn't exist if you're copying local data, because you first
read the data (now it's warm in cache) before writing it.  So the verify
operation is essentially zero time in that case.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss