Re: [zfs-discuss] Solaris 10u9 with zpool version 22, but no DEDUP (version 21 reserved)

2010-09-12 Thread Roy Sigurd Karlsbakk
 If not, will any possible problems be avoided if I remove (transfer
 data away from) any filesystems with dedup=on ?

I would think re-copying data from a deduped dataset to a non-deduped dataset 
will fix it, yes. But then, who knows, perhaps Oracle will fix dedup and make 
it usable one day...

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Solaris 10u9 with zpool version 22, but no DEDUP (version 21 reserved)

2010-09-12 Thread Richard Elling
On Sep 12, 2010, at 3:47 AM, Roy Sigurd Karlsbakk r...@karlsbakk.net wrote:

 If not, will any possible problems be avoided if I remove (transfer
 data away from) any filesystems with dedup=on ?
 
 I would think re-copying data from a deduped dataset to a non-deduped dataset 
 will fix it, yes. But then, who knows, perhaps Oracle will fix dedup and make 
 it usable one day...

Dedup changes large I/Os into small I/Os. If you expect dedup to change
large I/Os into no I/Os, then you will be unhappy. Thus, the best solution for
dedup is one which can handle large amounts of small I/Os.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Failed zfs send invalid backup stream.............

2010-09-12 Thread Humberto Ramirez
I'm trying to replicate a 300 GB pool with this command

zfs send al...@3 | zfs receive -F omega

about 2 hours in to the process it fails with this error

cannot receive new filesystem stream: invalid backup stream

I have tried setting the target read only  (zfs set readonly=on omega)
also disable Timeslider thinking it might have something to do with it.

What could be causing the error ? if the target is a new hard drive can I use 
this   zfs send al...@3  /dev/c10t0d0 ?

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Failed zfs send invalid backup stream.............

2010-09-12 Thread Andrew Gabriel

Humberto Ramirez wrote:

I'm trying to replicate a 300 GB pool with this command

zfs send al...@3 | zfs receive -F omega

about 2 hours in to the process it fails with this error

cannot receive new filesystem stream: invalid backup stream

I have tried setting the target read only  (zfs set readonly=on omega)
also disable Timeslider thinking it might have something to do with it.

What could be causing the error ?


Could be zfs filesystem version too old on the sending side (I have one 
such case).

What are their versions, and what release/build of the OS are you using?

if the target is a new hard drive can I use 
this   zfs send al...@3  /dev/c10t0d0 ?
  


That command doesn't make much sense for the purpose of doing anything 
useful.


--
Andrew Gabriel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Failed zfs send invalid backup stream.............

2010-09-12 Thread Humberto Ramirez
Thanks for the reply Andrew, there both 22 (I checked on that prior to post)

Thanks.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to migrate to 4KB sector drives?

2010-09-12 Thread Orvar Korvar
No replies. Does this mean that you should avoid large drives with 4KB sectors, 
that is, new drives? ZFS does not handle new drives?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Hang on zpool import (dedup related)

2010-09-12 Thread Chris Murray
Another hang on zpool import thread, I'm afraid, because I don't seem to have 
observed any great successes in the others and I hope there's a way of saving 
my data ...

In March, using OpenSolaris build 134, I created a zpool, some zfs filesystems, 
enabled dedup on them, moved content into them and promptly discovered how slow 
it was because I only have 4GB RAM. Even with 30GB L2ARC, the performance was 
unacceptable. The trouble started when the machine hung one day. Ever since, 
I've been unable to import my pool without it hanging again. At the time I saw 
posts from others who had run into similar problems, so I thought it best that 
I wait until a later build, on the assumption that some ZFS dedup bug would be 
fixed and I could see my data again. I've been waiting ever since, and only 
just had a chance to try build 147, thanks to illumos and a schillix live CD.

However, the pool still won't import, so I'd much appreciate any 
troubleshooting hints and tips to help me on my way.

schillix b147i

My process is:
1. boot the live CD.
2. on the console session, run  vmstat 1
3. from another machine, SSH in with multiple sessions and:
vmstat 60
vmstat 1
zpool import -f zp
zpool iostat zp 1
zpool iostat zp -v 5
4. wait until it all stops

What I observe is that the zpool import command never finishes, there will be a 
lengthy period of read activity made up of very small reads which then stops 
before an even longer period of what looks like no disk activity.

zp   512G  1.31T  0  0  0  0

The box will be responsive for quite some time, seemingly doing not a great 
deal:

 kthr  memorypagedisk  faults  cpu
 r b w   swap  free  re  mf pi po fr de sr cd cd rm s0   in   sy   cs us sy id
 0 0 0 2749064 3122988 0  7  0  0  0  0  0  0  1  0  0  365  218  714  0  1 99

Then after a matter of hours it'll hang. SSH sessions are no longer responsive. 
On the console I can press return which creates a new line, but vmstat will 
have stopped updating.

Interestingly, what I observed in b134 was the same thing, however the free 
memory would slowly decrease over the course of hours, before a sudden 
nose-dive right before the lock up. Now it appears to hang without that same 
effect.

While the import appears to be working, I can cd to /zp and look at content of 
the filesystems of 5 of the 9 esx* directories.
Coincidence or not, it's the last four which appear to be empty - esx_prod 
onward.

# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
zp 905G  1.28T23K  /zp
zp/nfs 889G  1.28T32K  /zp/nfs
zp/nfs/esx_dev 264G  1.28T   264G  /zp/nfs/esx_dev
zp/nfs/esx_hedgehog   25.8G  1.28T  25.8G  /zp/nfs/esx_hedgehog
zp/nfs/esx_meerkat 223G  1.28T   223G  /zp/nfs/esx_meerkat
zp/nfs/esx_meerkat_dedup   938M  1.28T   938M  /zp/nfs/esx_meerkat_dedup
zp/nfs/esx_page   8.90G  1.28T  8.90G  /zp/nfs/esx_page
zp/nfs/esx_prod306G  1.28T   306G  /zp/nfs/esx_prod
zp/nfs/esx_skunk21K  1.28T21K  /zp/nfs/esx_skunk
zp/nfs/esx_temp   45.5G  1.28T  45.5G  /zp/nfs/esx_temp
zp/nfs/esx_template   15.2G  1.28T  15.2G  /zp/nfs/esx_template

Any help would be appreciated. What could be going wrong here? Is it getting 
progressively closer to becoming imported each time I try this, or will it be 
starting from scratch? Feels to me like there's an action in the 
/zp/nfs/esx_prod filesystem it's trying to replay and never getting to the end 
of, for some reason. In case it was getting in a muddle with the l2arc, I 
removed the cache device a matter of minutes into this run. It hasn't hung yet, 
vmstat is still updating, but I tried a 'zpool import' in one of the windows to 
see if I could even see a pool on another disk, and that hasn't returned me 
back to the prompt yet. Also tried to SSH in with another session, and that 
hasn't produced the login prompt.

Thanks in advance,
Chris
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to migrate to 4KB sector drives?

2010-09-12 Thread Brandon High
On Sun, Sep 12, 2010 at 10:07 AM, Orvar Korvar
knatte_fnatte_tja...@yahoo.com wrote:
 No replies. Does this mean that you should avoid large drives with 4KB 
 sectors, that is, new drives? ZFS does not handle new drives?

Solaris 10u9 handles 4k sectors, so it might be in a post-b134 release of osol.

-B

-- 
Brandon High : bh...@freaks.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Hang on zpool import (dedup related)

2010-09-12 Thread George Wilson

Chris Murray wrote:

Another hang on zpool import thread, I'm afraid, because I don't seem to have 
observed any great successes in the others and I hope there's a way of saving 
my data ...

In March, using OpenSolaris build 134, I created a zpool, some zfs filesystems, 
enabled dedup on them, moved content into them and promptly discovered how slow 
it was because I only have 4GB RAM. Even with 30GB L2ARC, the performance was 
unacceptable. The trouble started when the machine hung one day. Ever since, 
I've been unable to import my pool without it hanging again. At the time I saw 
posts from others who had run into similar problems, so I thought it best that 
I wait until a later build, on the assumption that some ZFS dedup bug would be 
fixed and I could see my data again. I've been waiting ever since, and only 
just had a chance to try build 147, thanks to illumos and a schillix live CD.

However, the pool still won't import, so I'd much appreciate any 
troubleshooting hints and tips to help me on my way.

schillix b147i

My process is:
1. boot the live CD.
2. on the console session, run  vmstat 1
3. from another machine, SSH in with multiple sessions and:
vmstat 60
vmstat 1
zpool import -f zp
zpool iostat zp 1
zpool iostat zp -v 5
4. wait until it all stops

What I observe is that the zpool import command never finishes, there will be a 
lengthy period of read activity made up of very small reads which then stops 
before an even longer period of what looks like no disk activity.

zp   512G  1.31T  0  0  0  0

The box will be responsive for quite some time, seemingly doing not a great 
deal:

 kthr  memorypagedisk  faults  cpu
 r b w   swap  free  re  mf pi po fr de sr cd cd rm s0   in   sy   cs us sy id
 0 0 0 2749064 3122988 0  7  0  0  0  0  0  0  1  0  0  365  218  714  0  1 99

Then after a matter of hours it'll hang. SSH sessions are no longer responsive. 
On the console I can press return which creates a new line, but vmstat will 
have stopped updating.

Interestingly, what I observed in b134 was the same thing, however the free 
memory would slowly decrease over the course of hours, before a sudden 
nose-dive right before the lock up. Now it appears to hang without that same 
effect.

While the import appears to be working, I can cd to /zp and look at content of the 
filesystems of 5 of the 9 esx* directories.
Coincidence or not, it's the last four which appear to be empty - esx_prod 
onward.

# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
zp 905G  1.28T23K  /zp
zp/nfs 889G  1.28T32K  /zp/nfs
zp/nfs/esx_dev 264G  1.28T   264G  /zp/nfs/esx_dev
zp/nfs/esx_hedgehog   25.8G  1.28T  25.8G  /zp/nfs/esx_hedgehog
zp/nfs/esx_meerkat 223G  1.28T   223G  /zp/nfs/esx_meerkat
zp/nfs/esx_meerkat_dedup   938M  1.28T   938M  /zp/nfs/esx_meerkat_dedup
zp/nfs/esx_page   8.90G  1.28T  8.90G  /zp/nfs/esx_page
zp/nfs/esx_prod306G  1.28T   306G  /zp/nfs/esx_prod
zp/nfs/esx_skunk21K  1.28T21K  /zp/nfs/esx_skunk
zp/nfs/esx_temp   45.5G  1.28T  45.5G  /zp/nfs/esx_temp
zp/nfs/esx_template   15.2G  1.28T  15.2G  /zp/nfs/esx_template

Any help would be appreciated. What could be going wrong here? Is it getting 
progressively closer to becoming imported each time I try this, or will it be 
starting from scratch? Feels to me like there's an action in the 
/zp/nfs/esx_prod filesystem it's trying to replay and never getting to the end 
of, for some reason. In case it was getting in a muddle with the l2arc, I 
removed the cache device a matter of minutes into this run. It hasn't hung yet, 
vmstat is still updating, but I tried a 'zpool import' in one of the windows to 
see if I could even see a pool on another disk, and that hasn't returned me 
back to the prompt yet. Also tried to SSH in with another session, and that 
hasn't produced the login prompt.

Thanks in advance,
Chris


It looks like you may be past the import phase and into the mounting 
phase. What I would recommend is that you 'zpool import -N zp' so that 
none of the datasets get mounted and only the import happens. Then one 
by one you can mount the datasets in order (starting with 'zp') so you 
can find out which one maybe hanging.


- George
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Has anyone seen zpool corruption with VirtualBox shared folders?

2010-09-12 Thread Warren Strange
I posted the following to the VirtualBox forum. I would be interested in 
finding out if anyone else has ever seen zpool corruption with VirtualBox as a 
host on OpenSolaris:

-
I am running OpenSolaris b134 as a VirtualBox host, with a Linux guest.

I have experienced 6-7 instances of my zpool getting corrupted.  I am wondering 
if anyone else has ever seen this before. 

This is on a mirrored zpool - using drives from two different manufacturers 
(i.e. it is very unlikely both drives would fail at the same time, with the 
same blocks going bad). I initially thought I might have a memory problem - 
which could explain the simultaneous disk failures. After running memory 
diagnostics for 24 hours with no errors reported, I am beginning to suspect it 
might be something else.

I am using shared folders from the guest - mounted at guest boot up time. 

Is it possible that the Solaris vboxsf shared folder kernel driver is causing 
corruption? Being in the kernel, would it allow bypassing of the normal zfs 
integrity mechanisms? Or is it possible there is some locking issue or race 
condition that triggers the corruption?

Anecdotally, when I see the corruption the sequence of events seems to be:

- dmesg reports various vbox drivers being loaded (normal - just loading the 
drivers)
- Guest boots - gets just pass grub boot screen to the initial redhat boot 
screen. 
- The Guest hangs and never boots. 
- zpool status -v  reports corrupted files. The files are on the zpool 
containing the shared folders and the VirtualBox images


Thoughts?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Hang on zpool import (dedup related)

2010-09-12 Thread Chris Murray
Absolutely spot on George. The import with -N took seconds.

Working on the assumption that esx_prod is the one with the problem, I bumped 
that to the bottom of the list. Each mount was done in a second:

# zfs mount zp
# zfs mount zp/nfs
# zfs mount zp/nfs/esx_dev
# zfs mount zp/nfs/esx_hedgehog
# zfs mount zp/nfs/esx_meerkat
# zfs mount zp/nfs/esx_meerkat_dedup
# zfs mount zp/nfs/esx_page
# zfs mount zp/nfs/esx_skunk
# zfs mount zp/nfs/esx_temp
# zfs mount zp/nfs/esx_template

And those directories have the content in them that I'd expect. Good!

So now I try to mount esx_prod, and the influx of reads has started in   zpool 
iostat zp 1

This is the filesystem with the issue, but what can I do now?

Thanks again.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone seen zpool corruption with VirtualBox shared folders?

2010-09-12 Thread Jeff Savit

Hi Warren,

This may not help much, except perhaps as a way to eliminate possible 
causes, but I ran b134 with VirtualBox and guests on ZFS for quite a 
long time without any such symptoms. My pool is a simple, unmirrored 
one, so the difference may be there. I used shared folders without 
incident. Guests include Linux (several distros, including RH), Windows, 
Solaris, BSD.


--Jeff

On 09/12/10 11:05 AM, Warren Strange wrote:

I posted the following to the VirtualBox forum. I would be interested in 
finding out if anyone else has ever seen zpool corruption with VirtualBox as a 
host on OpenSolaris:

-
I am running OpenSolaris b134 as a VirtualBox host, with a Linux guest.

I have experienced 6-7 instances of my zpool getting corrupted.  I am wondering 
if anyone else has ever seen this before.

This is on a mirrored zpool - using drives from two different manufacturers 
(i.e. it is very unlikely both drives would fail at the same time, with the 
same blocks going bad). I initially thought I might have a memory problem - 
which could explain the simultaneous disk failures. After running memory 
diagnostics for 24 hours with no errors reported, I am beginning to suspect it 
might be something else.

I am using shared folders from the guest - mounted at guest boot up time.

Is it possible that the Solaris vboxsf shared folder kernel driver is causing 
corruption? Being in the kernel, would it allow bypassing of the normal zfs 
integrity mechanisms? Or is it possible there is some locking issue or race 
condition that triggers the corruption?

Anecdotally, when I see the corruption the sequence of events seems to be:

- dmesg reports various vbox drivers being loaded (normal - just loading the 
drivers)
- Guest boots - gets just pass grub boot screen to the initial redhat boot 
screen.
- The Guest hangs and never boots.
- zpool status -v  reports corrupted files. The files are on the zpool 
containing the shared folders and the VirtualBox images


Thoughts?
   



--


Jeff Savit | Principal Sales Consultant
Phone: 602.824.6275
Email: jeff.sa...@oracle.com | Blog: http://blogs.sun.com/jsavit
Oracle North America Commercial Hardware
Operating Environments  Infrastructure S/W Pillar
2355 E Camelback Rd | Phoenix, AZ 85016



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Failed zfs send invalid backup stream.............

2010-09-12 Thread Richard Elling
On Sep 12, 2010, at 8:24 AM, Humberto Ramirez wrote:

 I'm trying to replicate a 300 GB pool with this command
 
 zfs send al...@3 | zfs receive -F omega
 
 about 2 hours in to the process it fails with this error
 
 cannot receive new filesystem stream: invalid backup stream
 
 I have tried setting the target read only  (zfs set readonly=on omega)
 also disable Timeslider thinking it might have something to do with it.
 
 What could be causing the error ? if the target is a new hard drive can I use 
 this   zfs send al...@3  /dev/c10t0d0 ?

Network problems will cause this.  Try sending through ssh to take advantage
of extra data stream protection.
 -- richard

-- 
OpenStorage Summit, October 25-27, Palo Alto, CA
http://nexenta-summit2010.eventbrite.com

Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone seen zpool corruption with VirtualBox shared folders?

2010-09-12 Thread Richard Elling
On Sep 12, 2010, at 11:05 AM, Warren Strange wrote:

 I posted the following to the VirtualBox forum. I would be interested in 
 finding out if anyone else has ever seen zpool corruption with VirtualBox as 
 a host on OpenSolaris:
 
 -
 I am running OpenSolaris b134 as a VirtualBox host, with a Linux guest.
 
 I have experienced 6-7 instances of my zpool getting corrupted.  I am 
 wondering if anyone else has ever seen this before. 
 
 This is on a mirrored zpool - using drives from two different manufacturers 
 (i.e. it is very unlikely both drives would fail at the same time, with the 
 same blocks going bad). I initially thought I might have a memory problem - 
 which could explain the simultaneous disk failures. After running memory 
 diagnostics for 24 hours with no errors reported, I am beginning to suspect 
 it might be something else.

So we are clear, you are running VirtualBox on ZFS, rather than ZFS on 
VirtualBox?

 I am using shared folders from the guest - mounted at guest boot up time. 
 
 Is it possible that the Solaris vboxsf shared folder kernel driver is causing 
 corruption? Being in the kernel, would it allow bypassing of the normal zfs 
 integrity mechanisms? Or is it possible there is some locking issue or race 
 condition that triggers the corruption?
 
 Anecdotally, when I see the corruption the sequence of events seems to be:
 
 - dmesg reports various vbox drivers being loaded (normal - just loading the 
 drivers)
 - Guest boots - gets just pass grub boot screen to the initial redhat boot 
 screen. 
 - The Guest hangs and never boots. 
 - zpool status -v  reports corrupted files. The files are on the zpool 
 containing the shared folders and the VirtualBox images
 
 
 Thoughts?

Bad power supply, HBA, cables, or other common cause.
To help you determine the sort of corruption, for mirrored pools FMA will record
the nature of the discrepancies.
fmdump -eV
will show a checksum error and the associated bitmap comparisons.
 -- richard

-- 
OpenStorage Summit, October 25-27, Palo Alto, CA
http://nexenta-summit2010.eventbrite.com
ZFS and performance consulting
http://www.RichardElling.com












___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone seen zpool corruption with VirtualBox shared folders?

2010-09-12 Thread Warren Strange
 So we are clear, you are running VirtualBox on ZFS,
 rather than ZFS on VirtualBox?
 

Correct


 
 Bad power supply, HBA, cables, or other common cause.
 To help you determine the sort of corruption, for
 mirrored pools FMA will record
 the nature of the discrepancies.
   fmdump -eV
 will show a checksum error and the associated bitmap
 comparisons.

Below are the errors reported from the two disks. Not sure if anything looks 
suspicious (other than the obvious checksum error)




Sep 10 2010 12:49:42.315641690 ereport.fs.zfs.checksum
nvlist version: 0
class = ereport.fs.zfs.checksum
ena = 0x95816e82e2900401
detector = (embedded nvlist)
nvlist version: 0
version = 0x0
scheme = zfs
pool = 0xf3cb5e110f2c88ec
vdev = 0x961d9b28c1440020
(end detector)

pool = tank
pool_guid = 0xf3cb5e110f2c88ec
pool_context = 0
pool_failmode = wait
vdev_guid = 0x961d9b28c1440020
vdev_type = disk
vdev_path = /dev/dsk/c8t5d0s0
vdev_devid = id1,s...@sata_wdc_wd15eads-00p_wd-wcavu0351361/a
parent_guid = 0xdae51838a62627b9
parent_type = mirror
zio_err = 50
zio_offset = 0x1ef6813a00
zio_size = 0x2
zio_objset = 0x10
zio_object = 0x1402f
zio_level = 0
zio_blkid = 0x76f
cksum_expected = 0x405288851d24 0x100655c808fa2072 0xa89d11a403482052 
0xf1041fd6f838c6eb
cksum_actual = 0x40528884fd24 0x100655c803286072 0xa89d111c8af30052 
0xf0fbe93b4f02c6eb
cksum_algorithm = fletcher4
__ttl = 0x1
__tod = 0x4c8a7dc6 0x12d04f5a

Sep 10 2010 12:49:42.315641636 ereport.fs.zfs.checksum
nvlist version: 0
class = ereport.fs.zfs.checksum
ena = 0x95816e82e2900401
detector = (embedded nvlist)
nvlist version: 0
version = 0x0
scheme = zfs
pool = 0xf3cb5e110f2c88ec
vdev = 0x969570b704d5bff1
(end detector)

pool = tank
pool_guid = 0xf3cb5e110f2c88ec
pool_context = 0
pool_failmode = wait
vdev_guid = 0x969570b704d5bff1
vdev_type = disk
vdev_path = /dev/dsk/c8t4d0s0
vdev_devid = id1,s...@sata_st31500341as9vs3b4cp/a
parent_guid = 0xdae51838a62627b9
parent_type = mirror
zio_err = 50
zio_offset = 0x1ef6813a00
zio_size = 0x2
zio_objset = 0x10
zio_object = 0x1402f
zio_level = 0
zio_blkid = 0x76f
cksum_expected = 0x405288851d24 0x100655c808fa2072 0xa89d11a403482052 
0xf1041fd6f838c6eb
cksum_actual = 0x40528884fd24 0x100655c803286072 0xa89d111c8af30052 
0xf0fbe93b4f02c6eb
cksum_algorithm = fletcher4
__ttl = 0x1
__tod = 0x4c8a7dc6 0x12d04f24

Message was edited by: wstrange
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone seen zpool corruption with VirtualBox shared folders?

2010-09-12 Thread Richard Elling
Comments below...

On Sep 12, 2010, at 2:56 PM, Warren Strange wrote:
 So we are clear, you are running VirtualBox on ZFS,
 rather than ZFS on VirtualBox?
 
 
 Correct
 
 
 Bad power supply, HBA, cables, or other common cause.
 To help you determine the sort of corruption, for
 mirrored pools FMA will record
 the nature of the discrepancies.
  fmdump -eV
 will show a checksum error and the associated bitmap
 comparisons.
 
 Below are the errors reported from the two disks. Not sure if anything looks 
 suspicious (other than the obvious checksum error)
 
 Sep 10 2010 12:49:42.315641690 ereport.fs.zfs.checksum
 nvlist version: 0
   class = ereport.fs.zfs.checksum
   ena = 0x95816e82e2900401
   detector = (embedded nvlist)
   nvlist version: 0
   version = 0x0
   scheme = zfs
   pool = 0xf3cb5e110f2c88ec
   vdev = 0x961d9b28c1440020
   (end detector)
 
   pool = tank
   pool_guid = 0xf3cb5e110f2c88ec
   pool_context = 0
   pool_failmode = wait
   vdev_guid = 0x961d9b28c1440020
   vdev_type = disk
   vdev_path = /dev/dsk/c8t5d0s0
   vdev_devid = id1,s...@sata_wdc_wd15eads-00p_wd-wcavu0351361/a
   parent_guid = 0xdae51838a62627b9
   parent_type = mirror
   zio_err = 50
   zio_offset = 0x1ef6813a00
   zio_size = 0x2
   zio_objset = 0x10
   zio_object = 0x1402f
   zio_level = 0
   zio_blkid = 0x76f
   cksum_expected = 0x405288851d24 0x100655c808fa2072 0xa89d11a403482052 
 0xf1041fd6f838c6eb
   cksum_actual = 0x40528884fd24 0x100655c803286072 0xa89d111c8af30052 
 0xf0fbe93b4f02c6eb
   cksum_algorithm = fletcher4
   __ttl = 0x1
   __tod = 0x4c8a7dc6 0x12d04f5a
 
 Sep 10 2010 12:49:42.315641636 ereport.fs.zfs.checksum
 nvlist version: 0
   class = ereport.fs.zfs.checksum
   ena = 0x95816e82e2900401
   detector = (embedded nvlist)
   nvlist version: 0
   version = 0x0
   scheme = zfs
   pool = 0xf3cb5e110f2c88ec
   vdev = 0x969570b704d5bff1
   (end detector)
 
   pool = tank
   pool_guid = 0xf3cb5e110f2c88ec
   pool_context = 0
   pool_failmode = wait
   vdev_guid = 0x969570b704d5bff1
   vdev_type = disk
   vdev_path = /dev/dsk/c8t4d0s0
   vdev_devid = id1,s...@sata_st31500341as9vs3b4cp/a
   parent_guid = 0xdae51838a62627b9
   parent_type = mirror
   zio_err = 50
   zio_offset = 0x1ef6813a00
   zio_size = 0x2
   zio_objset = 0x10
   zio_object = 0x1402f
   zio_level = 0
   zio_blkid = 0x76f
   cksum_expected = 0x405288851d24 0x100655c808fa2072 0xa89d11a403482052 
 0xf1041fd6f838c6eb
   cksum_actual = 0x40528884fd24 0x100655c803286072 0xa89d111c8af30052 
 0xf0fbe93b4f02c6eb
   cksum_algorithm = fletcher4
   __ttl = 0x1
   __tod = 0x4c8a7dc6 0x12d04f24

In the case where one side of the mirror is corrupted and the other is correct, 
then
you will be shown the difference between the two, in the form of an abbreviated 
bitmap.

In this case, the data on each side of the mirror is the same, with a large 
degree of
confidence. So the source of the corruption is likely to be the same -- some 
common 
component: CPU, RAM, HBA, I/O path, etc. You can rule out the disks as suspects.
With some additional experiments you can determine if the corruption occurred 
during
the write or the read.
 -- richard

-- 
OpenStorage Summit, October 25-27, Palo Alto, CA
http://nexenta-summit2010.eventbrite.com

Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to migrate to 4KB sector drives?

2010-09-12 Thread Richard Elling
On Sep 12, 2010, at 10:11 AM, Brandon High wrote:

 On Sun, Sep 12, 2010 at 10:07 AM, Orvar Korvar
 knatte_fnatte_tja...@yahoo.com wrote:
 No replies. Does this mean that you should avoid large drives with 4KB 
 sectors, that is, new drives? ZFS does not handle new drives?
 
 Solaris 10u9 handles 4k sectors, so it might be in a post-b134 release of 
 osol.

OSol source yes, binaries no :-(  You will need another distro besides 
OpenSolaris.
The needed support in sd was added around the b137 timeframe.
 -- richard

-- 
OpenStorage Summit, October 25-27, Palo Alto, CA
http://nexenta-summit2010.eventbrite.com

Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to migrate to 4KB sector drives?

2010-09-12 Thread Mike Gerdts
On Sun, Sep 12, 2010 at 5:42 PM, Richard Elling rich...@nexenta.com wrote:
 On Sep 12, 2010, at 10:11 AM, Brandon High wrote:

 On Sun, Sep 12, 2010 at 10:07 AM, Orvar Korvar
 knatte_fnatte_tja...@yahoo.com wrote:
 No replies. Does this mean that you should avoid large drives with 4KB 
 sectors, that is, new drives? ZFS does not handle new drives?

 Solaris 10u9 handles 4k sectors, so it might be in a post-b134 release of 
 osol.

 OSol source yes, binaries no :-(  You will need another distro besides 
 OpenSolaris.
 The needed support in sd was added around the b137 timeframe.

OpenIndiana, to be released on Tuesday, is based on b146 or later.

-- 
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to migrate to 4KB sector drives?

2010-09-12 Thread Brandon High
On Sun, Sep 12, 2010 at 3:42 PM, Richard Elling rich...@nexenta.com wrote:
 OSol source yes, binaries no :-(  You will need another distro besides 
 OpenSolaris.
 The needed support in sd was added around the b137 timeframe.

Do you know if it's been backported to Nexenta Core? There doesn't
seem to be a list of backported changes available on nexenta.org.

-B

-- 
Brandon High : bh...@freaks.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OCZ Vertex 2 Pro performance numbers

2010-09-12 Thread Marc Bevand
(I am aware I am replying to an old post...)

Arne Jansen sensille at gmx.net writes:
 
 Now the test for the Vertex 2 Pro. This was fun.
 For more explanation please see the thread Crucial RealSSD C300 and cache
 flush?
 This time I made sure the device is attached via 3GBit SATA. This is also
 only a short test. I'll retest after some weeks of usage.
 
 cache enabled, 32 buffers, 64k blocks
 linear write, random data: 96 MB/s
 linear read, random data: 206 MB/s
 linear write, zero data: 234 MB/s
 linear read, zero data: 255 MB/s

This discrepancy between tests with random data and zero data is puzzling
to me. Does this suggest that the SSD does transparent compression between
its Sandforce SF-1500 controller and the NAND flash chips?

 cache enabled, 32 buffers, 4k blocks
 random write, random data: 41 MB/s (10300 ops/s)
 random read, random data: 76 MB/s (19000 ops/s)
 random write, zero data: 54 MB/s (13800 ops/s)
 random read, zero data: 91 MB/s (22800 ops/s)

These IOPS numbers are significantly below the 5 IOPS announced
by OCZ. But I supposed this is due to your benchmark tool not aligning
the ops to 4K boundaries, correct?

-mrb

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] file recovery on lost RAIDZ array

2010-09-12 Thread Michael Eskowitz
I recently lost all of the data on my single parity raid z array.  Each of the 
drives was encrypted with the zfs array built within the encrypted volumes.

I am not exactly sure what happened.  The files were there and accessible and 
then they were all gone.  The server apparently crashed and rebooted and 
everything was lost.  After the crash I remounted the encrypted drives and the 
zpool was still reporting that roughly 3TB of the 7TB array were used, but I 
could not see any of the files through the array's mount point.  I unmounted 
the zpool and then remounted it and suddenly zpool was reporting 0TB were used. 
 I did not remap the virtual device.  The only thing of note that I saw was 
that the name of storage pool had changed.  Originally it was Movies and then 
it became Movita.  I am guessing that the file system became corrupted some 
how.  (zpool status did not report any errors)

So, my questions are these... 

Is there anyway to undelete data from a lost raidz array?  If I build a new 
virtual device on top of the old one and the drive topology remains the same, 
can we scan the drives for files from old arrays?

Also, is there any way to repair a corrupted storage pool?  Is it possible to 
backup the file table or whatever partition index zfs maintains?


I imagine that you all are going to suggest that I scrub the array, but that is 
not an option at this point.  I had a backup of all of the data lost as I am 
moving between file servers so at a certain point I gave up and decided to 
start fresh.  This doesn't give me a warm fuzzy feeling about zfs, though.

Thanks,
-Mike
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resilver = defrag?

2010-09-12 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Orvar Korvar
 
 I am not really worried about fragmentation. I was just wondering if I
 attach new drives and zfs send recieve to a new zpool, would count as
 defrag. But apparently, not.

Apparently not in all situations would be more appropriate.

The understanding I had was:  If you send a single zfs send | receive, then
it does effectively get defragmented, because the receiving filesystem is
going to re-layout the received filesystem, and there is nothing
pre-existing to make the receiving filesystem dance around...  But if you're
sending some initial, plus incrementals, then you're actually repeating the
same operations that probably caused the original filesystem to become
fragmented in the first place.  And in fact, it seems unavoidable...

Suppose you have a large file, which is all sequential on disk.  You make a
snapshot of it.  Which means all the individual blocks must not be
overwritten.  And then you overwrite a few bytes scattered randomly in the
middle of the file.  The nature of copy on write is such that of course, the
latest version of the filesystem is impossible to remain contiguous.  Your
only choices are:  To read  write copies of the whole file, including
multiple copies of what didn't change, or you leave the existing data in
place where it is on disk, and you instead write your new random bytes to
other non-contiguous locations on disk.  Hence fragmentation.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] resilver = defrag?

2010-09-12 Thread Richard Elling
On Sep 12, 2010, at 8:27 PM, Edward Ned Harvey wrote:

 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Orvar Korvar
 
 I am not really worried about fragmentation. I was just wondering if I
 attach new drives and zfs send recieve to a new zpool, would count as
 defrag. But apparently, not.
 
 Apparently not in all situations would be more appropriate.
 
 The understanding I had was:  If you send a single zfs send | receive, then
 it does effectively get defragmented, because the receiving filesystem is
 going to re-layout the received filesystem, and there is nothing
 pre-existing to make the receiving filesystem dance around...  But if you're
 sending some initial, plus incrementals, then you're actually repeating the
 same operations that probably caused the original filesystem to become
 fragmented in the first place.  And in fact, it seems unavoidable...
 
 Suppose you have a large file, which is all sequential on disk.  You make a
 snapshot of it.  Which means all the individual blocks must not be
 overwritten.  And then you overwrite a few bytes scattered randomly in the
 middle of the file.  The nature of copy on write is such that of course, the
 latest version of the filesystem is impossible to remain contiguous.  Your
 only choices are:  To read  write copies of the whole file, including
 multiple copies of what didn't change, or you leave the existing data in
 place where it is on disk, and you instead write your new random bytes to
 other non-contiguous locations on disk.  Hence fragmentation.

This operational definition of fragmentation comes from the single-user,
single-tasking world (PeeCees). In that world, only one thread writes files
from one application at one time. In those cases, there is a reasonable
expectation that a single file's blocks might be contiguous on a single disk.
That isn't the world we live in, where have RAID, multi-user, or multi-threaded 
environments. 
 -- richard

-- 
OpenStorage Summit, October 25-27, Palo Alto, CA
http://nexenta-summit2010.eventbrite.com

Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to migrate to 4KB sector drives?

2010-09-12 Thread Nikola M
Orvar Korvar wrote:
 ZFS does not handle 4K sector drives well, you need to create a new zpool 
 with 4K property (ashift) set. 
 http://www.solarismen.de/archives/5-Solaris-and-the-new-4K-Sector-Disks-e.g.-WDxxEARS-Part-2.html

 Are there plans to allow resilver to handle 4K sector drives?
   
Not sure about resilvering to 4K but as manual for Solaris under link
you provided is describing,
it can make new zpools aligned to 4K.

Untill OpenIndiana.org come to life,
maybe in meantime you can try to build Illumos OS/Net testing part, have
it in separate BE on top of 134b Opensolaris from genunix.org and try to
_for testing_ 4K sector drives rpool instructions you found, on it.
http://www.illumos.org/projects/illumos-gate/wiki/How_To_Build_illumos

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OCZ Vertex 2 Pro performance numbers

2010-09-12 Thread Marc Bevand
Marc Bevand m.bevand at gmail.com writes:
 
 This discrepancy between tests with random data and zero data is puzzling
 to me. Does this suggest that the SSD does transparent compression between
 its Sandforce SF-1500 controller and the NAND flash chips?

Replying to myself: yes, SF-1500 does transparent deduplication
and compression to reduce write-amplification. Wow.

http://www.semiaccurate.com/2010/05/03/sandforce-ssds-break-tpc-c-records/

-mrb


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss