Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Tonmaus
Hi Bob,
 
 It is necessary to look at all the factors which
 might result in data 
 loss before deciding what the most effective steps
 are to minimize 
 the probability of loss.
 
 Bob

I am under the impression that exactly those were the considerations for both 
the ZFS designers to implement a scrub function to ZFS and the author of Best 
Practises to recommend performing this function frequently. I am hearing you 
are coming to a different conclusion and I would be interested in learning what 
could possibly be so highly interpretable in this.

Regards,

Tonmaus
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Reverse lookup: inode to name lookup

2010-05-02 Thread Casper . Dik


You can do in the kernel by calling vnodetopath(). I don't know if it
is exposed to user space.

Yes, in /proc/*/path  (kinda).

But that could be slow if you have large directories so you have to
think about where you would use it.

The kernel caches file names; however, it cannot be use for files that 
aren't in use.

It is certainly possible to create a .zfs/snapshot_byinode but it is not 
clear when it helps but it can be used for finding the earlier copy of a 
directory (netapp/.snapshot)

Casper

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Victor Latushkin
On May 2, 2010, at 8:47 AM, Steve Staples wrote:

 Hi there!
 
 I am new to the list, and to OpenSolaris, as well as ZPS.
 
 I am creating a zpool/zfs to use on my NAS server, and basically I want some
 redundancy for my files/media.   What I am looking to do, is get a bunch of
 2TB drives, and mount them mirrored, and in a zpool so that I don't have to
 worry about running out of room. (I know, pretty typical I guess).
 
 My problem is, is that not all 2TB hard drives are the same size (even
 though they should be 2 trillion bytes, there is still sometimes a +/- (I've
 only noticed this 2x so far) ) and if I create them mirrored, and one fails,
 and then I replace the drive, and for some reason, it is 1byte smaller, it
 will not work.
 
 How would I go about fixing this problem?

This problem is already fixed for you in ZFS. For disk sizes in 2TB it may 
tolerate difference in size up to approximately a little bit less than half a 
metaslab size which is currently likely to be 16GB, thus it may tolerate 
difference in size of up to, say, 7.5GB. 

I think that in most cases difference in sizes is below that figure.

You can see it for yourself:

bash-4.0# mkfile -n 2 d0
bash-4.0# zpool create pool `pwd`/d0
bash-4.0# mkfile -n 1992869543936 d1
bash-4.0# zpool attach pool `pwd`/d0 `pwd`/d1
bash-4.0# zpool status pool
  pool: pool
 state: ONLINE
 scrub: resilver completed after 0h0m with 0 errors on Sun May  2 15:25:24 2010
config:

NAME STATE READ WRITE CKSUM
pool ONLINE   0 0 0
  mirror-0   ONLINE   0 0 0
/var/tmp/d0  ONLINE   0 0 0
/var/tmp/d1  ONLINE   0 0 0  83.5K resilvered

errors: No known data errors
bash-4.0# zpool detach pool `pwd`/d1

So you can see that even though difference in size between d0 and d1 is 
7130456064 (~6.6GB), it can still be attached just fine. Let's now detach d1 
and make it 1 byte smaller:

bash-4.0# mkfile -n 1992869543935 d1
bash-4.0# zpool attach pool `pwd`/d0 `pwd`/d1
cannot attach /var/tmp/d1 to /var/tmp/d0: device is too small
bash-4.0# 

This time is is no longer possible to attach it, because size is not enough to 
fit the same number (116) of 16G metaslabs;

 THIS is just a thought, I am looking for thoughts and opinions on doing
 this... it prolly would be a bad idea, but hey, does it hurt to ask?
 
 I have been thinking, and would it be a good idea, to have on the 2TB
 drives, say 1TB or 500GB files and then mount them as mirrored?   So
 basically, have a 2TB hard drive, set up like:
 
 (where drive1 and drive2 are the paths to the mount points)
 Mkfile 465gb /drive1/drive1part1
 Mkfile 465gb /drive1/drive1part2
 Mkfile 465gb /drive1/drive1part3
 Mkfile 465gb /drive1/drive1part4
 
 Mkfile 465gb /drive2/drive2part1
 Mkfile 465gb /drive2/drive2part2
 Mkfile 465gb /drive2/drive2part3
 Mkfile 465gb /drive2/drive2part4
 
 (I use 465gb, as 2TB = 2trillion bytes, / 4 = 465.66 gb)
 
 And then add them to the zpool
 Zpool add medianas mirror /drive1/drive1part1 /drive2/drive2/part1
 Zpool add medianas mirror /drive1/drive1part2 /drive2/drive2/part2
 Zpool add medianas mirror /drive1/drive1part3 /drive2/drive2/part3
 Zpool add medianas mirror /drive1/drive1part4 /drive2/drive2/part4

This is not a good idea


regards
victor

 And then, if a drive goes and I only have a 500gb and a 1.5tb drives, they
 could be replaced that way?
 
 I am sure there are performance issues in doing this, but would the
 performance outweigh the possibility of hard drive failure and replacing
 drives?
 
 Sorry for posting a novel, but I am just concerned about failure on bigger
 drives, and putting my media/files into basically what consists of a JBOD
 type array (on steroids).
 
 Steve
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] rpool gone after zpool detach

2010-05-02 Thread Jan Riechers
I am using a mirrored system pool on 2 80G drives - however I was only using
40G since I thought I might use the rest for something else. ZFS Time Slider
was complaining the pool was filled for 90% and I decided to increase pool
size.
What I did was a zpool detach of one of the mirrored hdds and increased the
partition size to 100% with fdisk. When I wanted to reattach the hdd, system
complained about an IO error and it hang.
Now the rpool is gone on both drives (I also tried to find it via zpool
import booting from USB - without success). Is there any chance I can
recover the lost rpool? And what did I do wrong (except for not having a
backup first)?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] rpool gone after zpool detach

2010-05-02 Thread Roy Sigurd Karlsbakk
- Jan Riechers jan.riech...@googlemail.com skrev: 


I am using a mirrored system pool on 2 80G drives - however I was only using 
40G since I thought I might use the rest for something else. ZFS Time Slider 
was complaining the pool was filled for 90% and I decided to increase pool 
size. 
What I did was a zpool detach of one of the mirrored hdds and increased the 
partition size to 100% with fdisk. When I wanted to reattach the hdd, system 
complained about an IO error and it hang. 
Now the rpool is gone on both drives (I also tried to find it via zpool import 
booting from USB - without success). Is there any chance I can recover the lost 
rpool? And what did I do wrong (except for not having a backup first)? 

___ 
zfs-discuss mailing list 
zfs-discuss@opensolaris.org 
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss 
What error messages do you get? Please give more info 

roy 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] First Setup

2010-05-02 Thread Ian D

Hi!  We're building our first dedicated ZFS-based NAS/SAN (probably using 
Nexenta) and I'd like to run the specs by you all to see if you have any 
recommendations.  All of it is already bought, but it's not too late to add to 
it.  
Dell PowerEdge R9102x Intel X7550 2GHz, 8 cores each plus HyperThreading256GB 
RAM2x 4GB DDRDrive X1 (PCIe) in mirror for the ZIL8x 100GB Samsung SS805 SSDs 
for the L2ARC (on an onboard PERC H700 controller)2x PERC H800 and 2x PERC6 
controllers for the JBODs below11x Dell MD1000 JBODs with: 45x 750GB SATA 
HDDs 30x 1000GB SATA HDDs 60X 300GB SAS 15K HDDs 30x 600GB SAS 15K 
HDDs2x 10GbE ports
We plan to connect about 30 servers to it.  About 8 will be currently I/O-bound 
MySQL databases with a 60/40 read bias, the rest will have much lighter usage.  
Most connections will be through iSCSI.
Is there anything that seems out of proportion?  Where do you think the 
bottleneck will be?  If I'm going to use the SAS 15K drives for databases and 
the SATA drives for NFS/backups, how should I setup the pools? 
Thank you for any advice!
  ___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool rename?

2010-05-02 Thread Richard L. Hamilton
One can rename a zpool on import

zpool import -f pool_or_id newname

Is there any way to rename it (back again, perhaps)
on export?

(I had to rename rpool in an old disk image to access
some stuff in it, and I'd like to put it back the way it
was so it's properly usable if I ever want to boot off of it.)

But I suppose there must be other scenarios where that would
be useful too...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool rename?

2010-05-02 Thread Roy Sigurd Karlsbakk
- Richard L. Hamilton rlha...@smart.net skrev:

 One can rename a zpool on import
 
 zpool import -f pool_or_id newname
 
 Is there any way to rename it (back again, perhaps)
 on export?
 
 (I had to rename rpool in an old disk image to access
 some stuff in it, and I'd like to put it back the way it
 was so it's properly usable if I ever want to boot off of it.)
 
 But I suppose there must be other scenarios where that would
 be useful too...

just export it and reimportit with the old name

roy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Reverse lookup: inode to name lookup

2010-05-02 Thread Edward Ned Harvey
 From: cas...@holland.sun.com [mailto:cas...@holland.sun.com] On Behalf
 Of casper@sun.com
 
 It is certainly possible to create a .zfs/snapshot_byinode but it is
 not
 clear when it helps but it can be used for finding the earlier copy of
 a
 directory (netapp/.snapshot)

Do you happen to have any idea how easy/difficult it would be, to create
something like .zfs/snapshot_by_inode?  And although I hypothetically
described the behavior of such a thing once before, I don't recall seeing
any response to that.  So I don't know if there's any agreement on the
proposed behavior.

That suggestion, again, was:

Inside the .zfs directory, there is presently only one subdirectory,
snapshot.  Let there be more:
.zfs/snapshot
.zfs/name_from_inode
.zfs/inode_from_name

Inside the .zfs/snapshot directory, there's a list of snapshots.  The same
is true for the name_from_inode, and inode_from_name directories too.
However, inside the .zfs/snapshot/snapname directory, there is an actual
snapshot of all the files and directories of the filesystem.  The other two
directories behave as follows:

-- #1 -- 
.zfs/name_from_inode:

The .zfs/name_from_inode directory provides a mechanism to perform reverse
inode--name lookup.  

If a user does ls .zfs/name_from_inode/snapname then they will see
nothing.  (The system will not generate a complete list of all the inodes in
the filesystem; that would be crazy).  But if they explicitly cat
.zfs/name_from_inode/snapname/12345 then the system does an inode--name
reverse lookup, and if the result is accessible with the user's permission
level, then the result is just a text output of the pathname of the object.
(Presumably a directory, because there is currently no facility to reverse
lookup a file.  Directories do have a reference to their parent, via '..'
entries, but files have no such thing.)

Thus, if a user wants to find all the old snapshots of a directory in the
present filesystem, even if the name or location of that directory may have
changed, they could do this:
ls -di /tank/path/to/some/dirname
(result inode number 12345)
cat /tank/.zfs/name_from_inode/snapname/12345
(result path/to/previous/old-dirname
ls /tank/.zfs/snapshot/snapname/path/to/previous/old-dirname
And I am slowly working on scripts now, to simplify all the above into a
single command:
zhist ls /tank/path/to/some/dirname
would display all the former snapnames for that object.

It is possible for the OS, in between two snapshots, to recycle an inode
number.  So it is possible to mistakenly identify some completely unrelated
object as a former snapshot of a present object.  As far as I know, this is
unavoidable, but also unlikely.

-- #2 --
.zfs/inode_from_name:

The .zfs/inode_from_name directory provides a mechanism to find the inode
number of an object, when ls -di is not possible, because CIFS doesn't
support inodes.

So a CIFS client would do this:
cat /tank/.zfs/inode_from_name/path/to/some/dirname
(result inode number 12345)

The rest of the process would be the same as above.
cat /tank/.zfs/name_from_inode/snapname/12345
(result path/to/previous/old-dirname
ls /tank/.zfs/snapshot/snapname/path/to/previous/old-dirname

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Steve Staples
 
 My problem is, is that not all 2TB hard drives are the same size (even
 though they should be 2 trillion bytes, there is still sometimes a +/-
 (I've
 only noticed this 2x so far) ) and if I create them mirrored, and one
 fails,
 and then I replace the drive, and for some reason, it is 1byte smaller,
 it
 will not work.

If you use the latest developer release from genunix.org, it's already
fixed.  This is what I recommend, in general.
http://genunix.org/dist/indiana/   It looks like the latest one is
osol-dev-134

If you use the latest opensolaris release (2009.06) I'm not sure if it's
already fixed.

If you use Solaris 10, even fully updated, you are wise to ask this
question.  Because the fix is not present.

If you don't already have the fix, the recommendation would be to partition
the disks first.  Throw away something like 1% of the disk intentionally.
Yes this is extra work and a little wasteful, but it's preventive work for a
real problem.  http://docs.sun.com/app/docs/doc/806-4073/6jd67r9hu

(Quoting Richard Elling)
CR 6844090, zfs should be able to mirror to a smaller disk
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6844090 
b117, June 2009

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Steve Staples
 On May 2, 2010, at 8:47 AM, Steve Staples wrote:
 
  Hi there!
 
  I am new to the list, and to OpenSolaris, as well as ZPS.
 
  I am creating a zpool/zfs to use on my NAS server, and basically I want
some
  redundancy for my files/media.   What I am looking to do, is get a bunch
of
  2TB drives, and mount them mirrored, and in a zpool so that I don't have
to
  worry about running out of room. (I know, pretty typical I guess).
 
  My problem is, is that not all 2TB hard drives are the same size (even
  though they should be 2 trillion bytes, there is still sometimes a +/-
(I've
  only noticed this 2x so far) ) and if I create them mirrored, and one
fails,
  and then I replace the drive, and for some reason, it is 1byte smaller,
it
  will not work.
 
  How would I go about fixing this problem?
 
 This problem is already fixed for you in ZFS. For disk sizes in 2TB it may
 tolerate difference in size up to approximately a little bit less than
half a
 metaslab size which is currently likely to be 16GB, thus it may tolerate
 difference in size of up to, say, 7.5GB.
 
 I think that in most cases difference in sizes is below that figure.
 
 You can see it for yourself:
 
 bash-4.0# mkfile -n 2 d0
 bash-4.0# zpool create pool `pwd`/d0
 bash-4.0# mkfile -n 1992869543936 d1
 bash-4.0# zpool attach pool `pwd`/d0 `pwd`/d1
 bash-4.0# zpool status pool
   pool: pool
  state: ONLINE
  scrub: resilver completed after 0h0m with 0 errors on Sun May  2 15:25:24
 2010
 config:
 
   NAME STATE READ WRITE CKSUM
   pool ONLINE   0 0 0
 mirror-0   ONLINE   0 0 0
   /var/tmp/d0  ONLINE   0 0 0
   /var/tmp/d1  ONLINE   0 0 0  83.5K resilvered
 
 errors: No known data errors
 bash-4.0# zpool detach pool `pwd`/d1
 
 So you can see that even though difference in size between d0 and d1 is
 7130456064 (~6.6GB), it can still be attached just fine. Let's now detach
d1
 and make it 1 byte smaller:

I've done this in testing already, and mirrored a 10gb drive, with an 80gb
drive, and yes, the mirror assumes the 10gb size.  But, if I remove the 10gb
drive, and replace it with another 80gb, if the other 80gb drive is smaller
by even 1 byte, you cannot attach it.
 
 bash-4.0# mkfile -n 1992869543935 d1
 bash-4.0# zpool attach pool `pwd`/d0 `pwd`/d1
 cannot attach /var/tmp/d1 to /var/tmp/d0: device is too small
 bash-4.0#
 
 This time is is no longer possible to attach it, because size is not
enough to
 fit the same number (116) of 16G metaslabs;

And yes, I've gotten the exact same results when using file based disks
(mkfile -n 1992869543935 d1) of 1byte less.
Replacing with a bigger drive, is no problem, replacing with the same drive,
that is 1byte smaller... cannot do.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Steve Staples
  From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
  boun...@opensolaris.org] On Behalf Of Steve Staples
 
  My problem is, is that not all 2TB hard drives are the same size (even
  though they should be 2 trillion bytes, there is still sometimes a +/-
  (I've
  only noticed this 2x so far) ) and if I create them mirrored, and one
  fails,
  and then I replace the drive, and for some reason, it is 1byte smaller,
  it
  will not work.
 
 If you use the latest developer release from genunix.org, it's already
 fixed.  This is what I recommend, in general.
 http://genunix.org/dist/indiana/   It looks like the latest one is
 osol-dev-134
 
 If you use the latest opensolaris release (2009.06) I'm not sure if it's
 already fixed.

I am currently using OpenSolaris 2009.06
If I was to upgrade to the current developer version, forgive my ignorance
(since I am new to *solaris), but how would I do this?
Also, are there docs for this yet on what has changed?  I am looking in the
directory there, and I don't see any notes.   Any gentle nudge in the
right direction would be appreciated :)

 If you use Solaris 10, even fully updated, you are wise to ask this
 question.  Because the fix is not present.
 
 If you don't already have the fix, the recommendation would be to
partition
 the disks first.  Throw away something like 1% of the disk intentionally.
 Yes this is extra work and a little wasteful, but it's preventive work for
a
 real problem.  http://docs.sun.com/app/docs/doc/806-4073/6jd67r9hu

If I was to take this route, I was under the impression that ZFS takes the
drive as a whole, regardless of partitions, and wipes the data?  Apparently
I am wrong?
I don't have any problems with losing a few gig if needed, to save future
problems.  If I have to partition a drive of 2tb to 1.8tb (2 trillion bytes
= 1.862g), then oh well. 

 (Quoting Richard Elling)
 CR 6844090, zfs should be able to mirror to a smaller disk
 http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6844090
 b117, June 2009


Steve

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs testing / problems

2010-05-02 Thread Roy Sigurd Karlsbakk
Hi all

Testing variable size 'disks' in mirror, I followed Victor Latushkin's example

bash-4.0# mkfile -n 2 d0
bash-4.0# zpool create pool $PWD/d0
bash-4.0# mkfile -n 1992869543936 d1
bash-4.0# zpool attach pool $PWD/d0 $PWD/d1

and so on - this works well. Now, to try to mess with ZFS a litt (or a lot), I 
tried corrupting parts of both sides of the mirror to see what ZFS would do 
about it

r...@mime:/testpool/testdisks# dd if=/dev/urandom of=d1 bs=100k count=1 skip=30
1+0 records in
1+0 records out
102400 bytes (102 kB) copied, 0.00208301 s, 49.2 MB/s
r...@mime:/testpool/testdisks# dd if=/dev/urandom of=d0 bs=100k count=1 skip=50
1+0 records in
1+0 records out
102400 bytes (102 kB) copied, 0.00205321 s, 49.9 MB/s

This resulted in a panic - see below for the info from the kernel log. I'd 
forgotten to turn on dumps after last reinstall, but it should be easy to 
reproduce it. 

I think I read somewhere that it was normal for zfs/zpool to panic if it lost 
contact with a pool - is this what happened here? if so, is it possible to 
change this behaviour? If osol looses contact with a pool, I'd rather try to 
debug it and reboot myself if I want to, rather than having the system 
automatically panic.


May  2 17:42:09 mime unix: [ID 836849 kern.notice] May  2 17:42:09 mime 
^Mpanic[cpu1]/thread=ff00044d0c60: 
May  2 17:42:09 mime genunix: [ID 603766 kern.notice] assertion failed: 0 == 
zap_increment_int(os, (-1ULL), user, delta, tx) (0x0 == 0x32), file: ../.
./common/fs/zfs/dmu_objset.c, line: 1086
May  2 17:42:09 mime unix: [ID 10 kern.notice] 
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d09c0 
genunix:assfail3+c1 ()
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d0a20 
zfs:do_userquota_callback+11f ()
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d0a70 
zfs:dmu_objset_do_userquota_callbacks+a9 ()
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d0ae0 
zfs:dsl_pool_sync+f0 ()
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d0ba0 
zfs:spa_sync+3a9 ()
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d0c40 
zfs:txg_sync_thread+24a ()
May  2 17:42:09 mime genunix: [ID 655072 kern.notice] ff00044d0c50 
unix:thread_start+8 ()


Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Roy Sigurd Karlsbakk
 I am currently using OpenSolaris 2009.06
 If I was to upgrade to the current developer version, forgive my
 ignorance
 (since I am new to *solaris), but how would I do this?

# pkg set-publisher -O http://pkg.opensolaris.org/dev opensolaris.org
# pkg image-update

That'll take you to snv_134 or whatever the latest is

roy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] First Setup

2010-05-02 Thread Roy Sigurd Karlsbakk
- Ian D rewar...@hotmail.com skrev: 


Hi! We're building our first dedicated ZFS-based NAS/SAN (probably using 
Nexenta) and I'd like to run the specs by you all to see if you have any 
recommendations. All of it is already bought, but it's not too late to add to 
it. 


Dell PowerEdge R910 
2x Intel X7550 2GHz, 8 cores each plus HyperThreading 
256GB RAM 
2x 4GB DDRDrive X1 (PCIe) in mirror for the ZIL 
8x 100GB Samsung SS805 SSDs for the L2ARC (on an onboard PERC H700 controller) 
2x PERC H800 and 2x PERC6 controllers for the JBODs below 
11x Dell MD1000 JBODs with: 
45x 750GB SATA HDDs 
30x 1000GB SATA HDDs 
60X 300GB SAS 15K HDDs 
30x 600GB SAS 15K HDDs 
2x 10GbE ports 


We plan to connect about 30 servers to it. About 8 will be currently I/O-bound 
MySQL databases with a 60/40 read bias, the rest will have much lighter usage. 
Most connections will be through iSCSI. 


Is there anything that seems out of proportion? Where do you think the 
bottleneck will be? If I'm going to use the SAS 15K drives for databases and 
the SATA drives for NFS/backups, how should I setup the pools? 
For the DB pools, I'd say use RAID10. For the backup areas, RAIDz2 groups with 
something like 8-10 drives in each should be ok. For high write, 4GB ZIL might 
be a little low, and you have probably way too much CPU for just NAS/SAN. 
Remember to leave a few disks for spares (1-2 of each size?). Apart from that, 
it looks like you may save some rack space by getting new (consumer) 2TB 
drives, but that's your choice :) 

Best regards 

roy 
-- 
Roy Sigurd Karlsbakk 
(+47) 97542685 
r...@karlsbakk.net 
http://blogg.karlsbakk.net/ 
-- 
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk. 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Bob Friesenhahn

On Sun, 2 May 2010, Tonmaus wrote:


I am under the impression that exactly those were the considerations 
for both the ZFS designers to implement a scrub function to ZFS and 
the author of Best Practises to recommend performing this function 
frequently. I am hearing you are coming to a different conclusion 
and I would be interested in learning what could possibly be so 
highly interpretable in this.


The value of periodic scrub is subject to opinion.  There are some 
highly respected folks on this list who put less faith in scrub 
because they believe more in MTTDL statistical models and less in the 
value of early detection (scrub == early detection).  With a single 
level of redundancy, early detection is more useful since there is 
just one opportunity to correct the error and correcting the error 
early decreases the chance of a later uncorrectable error.  Scrub will 
help repair the results of transient wrong hardware operation, or 
partial media failures, but will not keep a whole disk from failing.


Once the computed MTTDL for the storage configuration is sufficiently 
high, then other factors such as the reliability of ECC memory, kernel 
bugs, and hardware design flaws, become dominant.  The human factor is 
often the most dominant factor when it comes to data loss since most 
data loss is still due to human error.  Most data loss problems we see 
reported here are due to human error or hardware design flaws.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Spare in use althought disk is healthy ?

2010-05-02 Thread Lutz Schumann
Hello, 

thanks for the feedback and sorry for the delay in answering.

I checked the log and the fmadm. It seems the log does not show changes, 
however fmadm shows: 

Apr 23 2010 18:32:26.363495457 ereport.io.scsi.cmd.disk.dev.rqs.derr
Apr 23 2010 18:32:26.363482031 ereport.io.scsi.cmd.disk.recovered

Same thing for the other disk: 

Apr 21 2010 15:02:24.117303285 ereport.io.scsi.cmd.disk.dev.rqs.derr
Apr 21 2010 15:02:24.117300448 ereport.io.scsi.cmd.disk.recovered

It seems there is a VERY short temp error. 

I will try to detach this. 

Is this a Bug ? 
Robert
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] rpool gone after zpool detach

2010-05-02 Thread Jan Riechers
On Sun, May 2, 2010 at 3:51 PM, Jan Riechers jan.riech...@googlemail.comwrote:



 On Sun, May 2, 2010 at 6:06 AM, Roy Sigurd Karlsbakk 
 r...@karlsbakk.netwrote:

 - Jan Riechers jan.riech...@googlemail.com skrev:

 I am using a mirrored system pool on 2 80G drives - however I was only
 using 40G since I thought I might use the rest for something else. ZFS Time
 Slider was complaining the pool was filled for 90% and I decided to increase
 pool size.
 What I did was a zpool detach of one of the mirrored hdds and increased
 the partition size to 100% with fdisk. When I wanted to reattach the hdd,
 system complained about an IO error and it hang.
 Now the rpool is gone on both drives (I also tried to find it via zpool
 import booting from USB - without success). Is there any chance I can
 recover the lost rpool? And what did I do wrong (except for not having a
 backup first)?

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 What error messages do you get? Please give more info


 Hi roy I'm not sure what the error message said in detail - however I
 cannot boot anymore since grub won't find the pool. However looking with zdb
 -l It finds at least on of the hdds 2 labels (the other which I didn't touch
 shows no label):
 pfexec zdb -l /dev/dsk/c6t5d0s2
 Is there a chance to fix this pool?



I managed to put back the old partition table/slice setup - since zdb showed
some outdated pool on s2 - when I got back the correct slice setup - I could
simply use c6t5d0s0 as rpool root.
The mirrored disk was never really bootable and the mirror somehow broken.
Is there an official howto for setting up a mirror for the rpool?
-- 
Jan
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] rpool gone after zpool detach

2010-05-02 Thread Richard Elling
On May 2, 2010, at 10:27 AM, Jan Riechers wrote:

 On Sun, May 2, 2010 at 3:51 PM, Jan Riechers jan.riech...@googlemail.com 
 wrote:
 
 
 On Sun, May 2, 2010 at 6:06 AM, Roy Sigurd Karlsbakk r...@karlsbakk.net 
 wrote:
 - Jan Riechers jan.riech...@googlemail.com skrev: 
 I am using a mirrored system pool on 2 80G drives - however I was only using 
 40G since I thought I might use the rest for something else. ZFS Time Slider 
 was complaining the pool was filled for 90% and I decided to increase pool 
 size.
 What I did was a zpool detach of one of the mirrored hdds and increased the 
 partition size to 100% with fdisk. When I wanted to reattach the hdd, system 
 complained about an IO error and it hang.
 Now the rpool is gone on both drives (I also tried to find it via zpool 
 import booting from USB - without success). Is there any chance I can recover 
 the lost rpool? And what did I do wrong (except for not having a backup 
 first)?
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
 What error messages do you get? Please give more info
 
 
 Hi roy I'm not sure what the error message said in detail - however I cannot 
 boot anymore since grub won't find the pool. However looking with zdb -l It 
 finds at least on of the hdds 2 labels (the other which I didn't touch shows 
 no label):
 pfexec zdb -l /dev/dsk/c6t5d0s2
 Is there a chance to fix this pool?
  
 
 I managed to put back the old partition table/slice setup - since zdb showed 
 some outdated pool on s2 - when I got back the correct slice setup - I could 
 simply use c6t5d0s0 as rpool root.
 The mirrored disk was never really bootable and the mirror somehow broken. Is 
 there an official howto for setting up a mirror for the rpool?

ZFS Administration Guide.
http://hub.opensolaris.org/bin/view/Community+Group+zfs/docs
 -- richard

ZFS storage and performance consulting at http://www.RichardElling.com





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Richard Elling
On May 1, 2010, at 1:56 PM, Bob Friesenhahn wrote:
 On Fri, 30 Apr 2010, Freddie Cash wrote:
 Without a periodic scrub that touches every single bit of data in the pool, 
 how can you be sure
 that 10-year files that haven't been opened in 5 years are still intact?
 
 You don't.  But it seems that having two or three extra copies of the data on 
 different disks should instill considerable confidence.  With sufficient 
 redundancy, chances are that the computer will explode before it loses data 
 due to media corruption.  The calculated time before data loss becomes longer 
 than even the pyramids in Egypt could withstand.

These calculations are based on fixed MTBF.  But disk MTBF decreases with 
age. Most disks are only rated at 3-5 years of expected lifetime. Hence, 
archivists
use solutions with longer lifetimes (high quality tape = 30 years) and plans 
for 
migrating the data to newer media before the expected media lifetime is 
reached.  
In short, if you don't expect to read your 5-year lifetime rated disk for 
another 5 years, 
then your solution is uhmm... shall we say... in need of improvement.

 
 The situation becomes similar to having a house with a heavy front door with 
 three deadbolt locks, and many glass windows.  The front door with its three 
 locks is no longer a concern when you are evaluating your home for its 
 security against burglary or home invasion because the glass windows are so 
 fragile and easily broken.
 
 It is necessary to look at all the factors which might result in data loss 
 before deciding what the most effective steps are to minimize the probability 
 of loss.

Yep... and manage the data over time.  There is a good reason why library 
scientists
will never worry about the future of their profession :-)
http://en.wikipedia.org/wiki/Library_science

 -- richard


ZFS storage and performance consulting at http://www.RichardElling.com




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Roy Sigurd Karlsbakk
- Roy Sigurd Karlsbakk r...@karlsbakk.net skrev:

 Hi all
 
 I have a test system with snv134 and 8x2TB drives in RAIDz2 and
 currently no Zil or L2ARC. I noticed the I/O speed to NFS shares on
 the testpool drops to something hardly usable while scrubbing the
 pool.
 
 How can I address this? Will adding Zil or L2ARC help? Is it possible
 to tune down scrub's priority somehow?

Further testing shows NFS speeds are acceptable after adding Zil and L2ARC (my 
test system has two SSDs for the root, so I detached on of them and split it 
into a 4GB slice for Zil and the rest for L2ARC).

Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread Giovanni
Hi guys

I am new to Opensolaris and ZFS world, I have 6x2TB SATA hdds on my system, I 
picked a single 2TB disk and installed opensolaris (therefore zpool was created 
by the installer)

I went ahead and created a new pool gpool with raidz (the kind of redundancy 
I want. Here's the output:

@server:/# zfs list
NAME USED  AVAIL  REFER  MOUNTPOINT
gpool119K  7.13T  30.4K  /gpool
rpool   7.78G  1.78T78K  /rpool
rpool/ROOT  3.30G  1.78T19K  legacy
rpool/ROOT/opensolaris  3.30G  1.78T  3.15G  /
rpool/dump  2.00G  1.78T  2.00G  -
rpool/export 491M  1.78T21K  /export
rpool/export/home491M  1.78T21K  /export/home
rpool/export/home/G   491M  1.78T   491M  /export/home/G
rpool/swap  2.00G  1.78T   101M  -
@server:/# 

@server:/# zpool status
  pool: gpool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
gpool   ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c8t1d0  ONLINE   0 0 0
c8t2d0  ONLINE   0 0 0
c8t3d0  ONLINE   0 0 0
c8t4d0  ONLINE   0 0 0
c8t5d0  ONLINE   0 0 0

errors: No known data errors

  pool: rpool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
rpool   ONLINE   0 0 0
  c8t0d0s0  ONLINE   0 0 0

errors: No known data errors
@server:/# 


 Now, I want to get rid of rpool in its entirely, I want to migrate all 
settings, boot records, files from that rpool to gpool and then add the 
member of rpool c8t0d0s0 to my existing gpool so that I have a RAIDZ of 6x 
drives. 

Any guidance on how to do it? I tried to do zfs snapshot

# zfs snapshot rp...@move


But I don't see the snapshow anywhere on rpool/.zfs (there is no .zfs folder)

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread mark.musa...@oracle.com
You can't get rid of rpool. That's the pool you're booting from. Root  
pools can only be single disks or n-way mirrors.


As to your other question, you can view the snapshots by using the  
command zfs list -t all, or turn on the listsnaps property for the  
pool. See  http://docs.sun.com/app/docs/doc/817-2271/ghbxt?a=view for  
more info.




Regards,
Mark

On 2 May 2010, at 15:58, Giovanni g...@csu.fullerton.edu wrote:


Hi guys

I am new to Opensolaris and ZFS world, I have 6x2TB SATA hdds on my  
system, I picked a single 2TB disk and installed opensolaris  
(therefore zpool was created by the installer)


I went ahead and created a new pool gpool with raidz (the kind of  
redundancy I want. Here's the output:


@server:/# zfs list
NAME USED  AVAIL  REFER  MOUNTPOINT
gpool119K  7.13T  30.4K  /gpool
rpool   7.78G  1.78T78K  /rpool
rpool/ROOT  3.30G  1.78T19K  legacy
rpool/ROOT/opensolaris  3.30G  1.78T  3.15G  /
rpool/dump  2.00G  1.78T  2.00G  -
rpool/export 491M  1.78T21K  /export
rpool/export/home491M  1.78T21K  /export/home
rpool/export/home/G   491M  1.78T   491M  /export/home/G
rpool/swap  2.00G  1.78T   101M  -
@server:/#

@server:/# zpool status
 pool: gpool
state: ONLINE
scrub: none requested
config:

   NAMESTATE READ WRITE CKSUM
   gpool   ONLINE   0 0 0
 raidz1ONLINE   0 0 0
   c8t1d0  ONLINE   0 0 0
   c8t2d0  ONLINE   0 0 0
   c8t3d0  ONLINE   0 0 0
   c8t4d0  ONLINE   0 0 0
   c8t5d0  ONLINE   0 0 0

errors: No known data errors

 pool: rpool
state: ONLINE
scrub: none requested
config:

   NAMESTATE READ WRITE CKSUM
   rpool   ONLINE   0 0 0
 c8t0d0s0  ONLINE   0 0 0

errors: No known data errors
@server:/#


Now, I want to get rid of rpool in its entirely, I want to migrate  
all settings, boot records, files from that rpool to gpool and  
then add the member of rpool c8t0d0s0 to my existing gpool so that  
I have a RAIDZ of 6x drives.


Any guidance on how to do it? I tried to do zfs snapshot

# zfs snapshot rp...@move


But I don't see the snapshow anywhere on rpool/.zfs (there is  
no .zfs folder)


Thanks
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread Roy Sigurd Karlsbakk
- Giovanni g...@csu.fullerton.edu skrev:

 Hi guys
 
 I am new to Opensolaris and ZFS world, I have 6x2TB SATA hdds on my
 system, I picked a single 2TB disk and installed opensolaris
 (therefore zpool was created by the installer)
 
 I went ahead and created a new pool gpool with raidz (the kind of
 redundancy I want. Here's the output:
  Now, I want to get rid of rpool in its entirely, I want to migrate
 all settings, boot records, files from that rpool to gpool and then
 add the member of rpool c8t0d0s0 to my existing gpool so that I have
 a RAIDZ of 6x drives. 
 
 Any guidance on how to do it? I tried to do zfs snapshot

You can't boot off raidz. That's for data only. Get a couple of cheap drives or 
SSDs for the root and use the large drives for data

Vennlige hilsener

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Bob Friesenhahn

On Sun, 2 May 2010, Richard Elling wrote:


These calculations are based on fixed MTBF.  But disk MTBF decreases with
age. Most disks are only rated at 3-5 years of expected lifetime. Hence, 
archivists
use solutions with longer lifetimes (high quality tape = 30 years) and plans for
migrating the data to newer media before the expected media lifetime is reached.
In short, if you don't expect to read your 5-year lifetime rated disk for 
another 5 years,
then your solution is uhmm... shall we say... in need of improvement.


Yes, the hardware does not last forever.  It only needs to last while 
it is still being used and should only be used during its expected 
service life.  Your point is a good one.


On the flip-side, using 'zfs scrub' puts more stress on the system 
which may make it more likely to fail.  It increases load on the power 
supplies, CPUs, interfaces, and disks.  A system which might work fine 
under normal load may be stressed and misbehave under scrub.  Using 
scrub on a weak system could actually increase the chance of data 
loss.



ZFS storage and performance consulting at http://www.RichardElling.com


Please send $$$ to the above address in return for wisdom.

Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Dave Pooser
On 5/2/10 3:12 PM, Bob Friesenhahn bfrie...@simple.dallas.tx.us wrote:

 On the flip-side, using 'zfs scrub' puts more stress on the system
 which may make it more likely to fail.  It increases load on the power
 supplies, CPUs, interfaces, and disks.  A system which might work fine
 under normal load may be stressed and misbehave under scrub.  Using
 scrub on a weak system could actually increase the chance of data
 loss.

If my system is going to fail under the stress of a scrub, it's going to
fail under the stress of a resilver. From my perspective, I'm not as scared
of data corruption as I am of data corruption *that I don't know about.* I
only keep backups for a finite amount of time. If I scrub every week, and my
zpool dies during a scrub, then I know it's time to pull out last week's
backup, where I know (thanks to scrubbing) the data was not corrupt. I've
lived the experience where a user comes to me because he tried to open a
seven-year-old file and it was corrupt. Not a blankety-blank thing I could
do, because we only retain backup tapes for four years and the four-year-old
tape had a backup of the file post-corruption.

Data loss may be unavoidable, but that's why we keep backups. It's the
invisible data loss that makes life suboptimal.
-- 
Dave Pooser, ACSA
Manager of Information Services
Alford Media  http://www.alfordmedia.com


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread Bob Friesenhahn

On Sun, 2 May 2010, Roy Sigurd Karlsbakk wrote:


Any guidance on how to do it? I tried to do zfs snapshot


You can't boot off raidz. That's for data only.


Unless you use FreeBSD ...

Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread Giovanni
Thank you. I was not aware that root pools could not be moved.

But here's the kicker, what if I have a single drive for root pool, and its 
failing... I connect a new HDD to replace the boot drive thats dying, ZFS has 
no way of migrating to a new drive?

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Bob Friesenhahn

On Sun, 2 May 2010, Dave Pooser wrote:


If my system is going to fail under the stress of a scrub, it's going to
fail under the stress of a resilver. From my perspective, I'm not as scared


I don't disagree with any of the opinions you stated except to point 
out that resilver will usually hit the (old) hardware less severely 
than scrub.  Resilver does not have to access any of the redundant 
copies of data or metadata, unless they are the only remaining good 
copy.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread Bob Friesenhahn

On Sun, 2 May 2010, Giovanni wrote:


Thank you. I was not aware that root pools could not be moved.

But here's the kicker, what if I have a single drive for root pool, 
and its failing... I connect a new HDD to replace the boot drive 
thats dying, ZFS has no way of migrating to a new drive?


There is a way.  The way would be to add a drive as a mirror of the 
old one.  The new drive should be in a position where the BIOS is able 
to boot from it.  The new drive would have its contents resilved based 
on the existing drive. However, if the old one is already failing, 
that could be somewhat problematic.  Once the new drive is functional, 
you can detatch the failing one.  Make sure that GRUB will boot from 
the new drive.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Virtual to physical migration

2010-05-02 Thread Orvar Korvar
You do know that OpenSolaris + VirtualBox can trash your ZFS raid? You can 
loose your data. There is a post about write cache and ZFS and VirtualbBox, I 
think you need to disable it?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Default 'zpool' I want to move it to my new raidz pool 'gpool' how?

2010-05-02 Thread Brandon High
On Sun, May 2, 2010 at 1:55 PM, Giovanni g...@csu.fullerton.edu wrote:
 But here's the kicker, what if I have a single drive for root pool, and its 
 failing... I connect a new HDD to replace the boot drive thats dying, ZFS has 
 no way of migrating to a new drive?

You can move root pools, I did it yesterday from a single 160GB drive
to mirrored 120GB drives.

You can't move them to anything other than a single disk or N-way
mirror. No raidz, no striped sets.

If a disk is failing, you can connect a new disk to the system and do
a zpool replace, which will resilver the new drive with the data. Or
you can attach a new drive to the pool to make a mirror.

-B

-- 
Brandon High : bh...@freaks.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance drop during scrub?

2010-05-02 Thread Richard Elling
On May 2, 2010, at 12:05 PM, Roy Sigurd Karlsbakk wrote:
 - Roy Sigurd Karlsbakk r...@karlsbakk.net skrev:
 
 Hi all
 
 I have a test system with snv134 and 8x2TB drives in RAIDz2 and
 currently no Zil or L2ARC. I noticed the I/O speed to NFS shares on
 the testpool drops to something hardly usable while scrubbing the
 pool.
 
 How can I address this? Will adding Zil or L2ARC help? Is it possible
 to tune down scrub's priority somehow?
 
 Further testing shows NFS speeds are acceptable after adding Zil and L2ARC 
 (my test system has two SSDs for the root, so I detached on of them and split 
 it into a 4GB slice for Zil and the rest for L2ARC).

Ok, this makes sense.  If you are using a pool configuration which is not
so good for high IOPS workloads (raidz*) and you give it a latency-sensitive,
synchronous IOPS workload (NFS) along with another high IOPS workload 
(scrub), then the latency-sensitive workload will notice.  Adding the SSD as a
separate log is a good idea.
 -- richard

ZFS storage and performance consulting at http://www.RichardElling.com




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Single-disk pool corrupted after controller failure

2010-05-02 Thread Peter Jeremy
On 2010-May-02 04:06:41 +0800, Diogo Franco diogomfra...@gmail.com wrote:
regular data corruption and then the box locked up. I had also
converted the pool to v14 a few days before, so the freebsd v13 tools
couldn't do anything to help.

Note that ZFS v14 was imported to FreeBSD 8-stable in mid-January.
I can't comment whether it would be able to recover your data.

On 2010-May-02 05:07:17 +0800, Bill Sommerfeld bill.sommerf...@oracle.com 
wrote:
  2) the labels are not at the start of what solaris sees as p1, and 
thus are somewhere else on the disk.  I'd look more closely at how 
freebsd computes the start of the partition or slice '/dev/ad6s1d'
that contains the pool.

I think #2 is somewhat more likely.

This is almost certainly the problem.  ad6s1 may be the same as c5d0p1
but OpenSolaris isn't going to understand the FreeBSD partition label
on that slice.  All I can suggest is to (temporarily) change the disk
slicing so that there is a fdisk slice that matches ad6s1d.

-- 
Peter Jeremy


pgpuiR7yDRv37.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs testing / problems

2010-05-02 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Roy Sigurd Karlsbakk
 
 bash-4.0# mkfile -n 2 d0
 bash-4.0# zpool create pool $PWD/d0
 bash-4.0# mkfile -n 1992869543936 d1
 bash-4.0# zpool attach pool $PWD/d0 $PWD/d1

As long as you're just doing this for testing, great.  I wouldn't suggest a 
configuration like that for any sort of permanent configuration.  Filesystems 
inside of files ... In general, not a great idea.  With only rare exceptions.

Also, you know you can do it all on one line, right?
zpool create mypool mirror $PWD/d0 $PWD/d1

Do a zpool status next.  I think the supposed resilver will be basically 
instantaneous, because the pool is empty.  But you should just check, and make 
sure the pool is healthy and not resilvering before you start your abuse.


 r...@mime:/testpool/testdisks# dd if=/dev/urandom of=d1 bs=100k count=1
 skip=30
 1+0 records in
 1+0 records out
 102400 bytes (102 kB) copied, 0.00208301 s, 49.2 MB/s
 r...@mime:/testpool/testdisks# dd if=/dev/urandom of=d0 bs=100k count=1
 skip=50
 1+0 records in
 1+0 records out
 102400 bytes (102 kB) copied, 0.00205321 s, 49.9 MB/s
 
 This resulted in a panic - see below for the info from the kernel log.

That looks perfect to me, except one thing.  You should zpool export before 
doing those dd's.  While ZFS will correctly identify the faulty blocks (and in 
your case, correct them because you have a mirror) ... If you're doing that to 
the device while it's mounted, I think that's worse than unknown randomness 
happening on disks that aren't reporting the randomness.  I think what you're 
doing (writing to the file, thus perhaps eliminating the ZFS open file handle 
to the device) is causing the devices to be swept out from under ZFS's 
feet.  I think this technique is more like a simulation of unplugging disks, 
and less like a simulation of random undetected errors happening on disks.

Also, if you're writing the randomness to blocks that happen to be unoccupied 
by anything, the system, I believe, won't even notice, because it'll never read 
the empty space that you've intentionally corrupted.

To ensure a good result, I would recommend:  
Create the filesystem as you've done.
Fill up the filesystem.  Thus, when you later do a scrub it will have to 
inspect all blocks.
zpool export.
Perform the dd's from urandom as you've done.
zpool import.
Check:  zpool status (it will probably say no errors)
Then zpool scrub.
After some time, zpool status will probably show that it found and corrected 
errors.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Edward Ned Harvey
 From: Roy Sigurd Karlsbakk [mailto:r...@karlsbakk.net]
 Sent: Sunday, May 02, 2010 11:55 AM
 
  I am currently using OpenSolaris 2009.06
  If I was to upgrade to the current developer version, forgive my
  ignorance
  (since I am new to *solaris), but how would I do this?
 
 # pkg set-publisher -O http://pkg.opensolaris.org/dev opensolaris.org
 # pkg image-update
 
 That'll take you to snv_134 or whatever the latest is

This may be perfectly correct, but my personal experience is to say that 
installing from scratch with the later ISO is much more reliable than applying 
this upgrade path.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool mirror (dumb question)

2010-05-02 Thread Edward Ned Harvey
 From: Steve Staples [mailto:thestapler...@gmail.com]
 
 I am currently using OpenSolaris 2009.06
 If I was to upgrade to the current developer version, forgive my
 ignorance
 (since I am new to *solaris), but how would I do this?

If you go to genunix.org (using the URL in my previous email) you can simply
download the 134 iso file there.  And install the OS from it.

If you ever need to remember this, there's a ridiculously simple way to
figure it out each time.  Go to http://opensolaris.org (not to be confused
with opensolaris.com) Go to the download page, and start reading.  Before
too long, you'll see some info about downloading and installing, or
upgrading to developer releases.  Including a reference to genunix.


  Zpool using partitions instead of whole devices
 
 If I was to take this route, I was under the impression that ZFS takes
 the
 drive as a whole, regardless of partitions, and wipes the data?

This is true, if you specify the whole device.
For example, if you format  partition c0t0d0, and then you zpool create
tank c0t0d0, then you're going to obliterate your format and partitions.
However, if you zpool create tank c0t0d0p1 or zpool create tank c0t0d0s1
then it will use the slice (partition) instead of the whole device.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss