[zfs-discuss] Can this be done?

2009-03-28 Thread Michael Shadle
I currently have a 7x1.5tb raidz1.

I want to add phase 2 which is another 7x1.5tb raidz1

Can I add the second phase to the first phase and basically have two
raid5's striped (in raid terms?)

Yes, I probably should upgrade the zpool format too. Currently running
snv_104. Also should upgrade to 110.

If that is possible, would anyone happen to have the simple command
lines to do it quick? I assume I'd be creating another raidz1 and then
somehow growing the tank zpool?

Does this make sense, or is this stupid from a performance
perspective? Should I just have two separate zpools? Ideally I would
like to have one massive data storage target. I'd be fine with somehow
changing this into a raidz2 as well, I suppose, since I had planned on
it being another raidz1 anyway.

Or, perhaps I could add tank #2 as a raidz2, and then move all the
data off tank #1, and then add disks individually from tank #1 until I
have all 14 disks in a single raidz2? Performance is not an absolute
must - I can deal with a little bit of overhead.

Thanks in advance.


[r...@nas01 ~]# zpool status
  pool: rpool
 state: ONLINE
status: The pool is formatted using an older on-disk format.  The pool can
still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
pool will no longer be accessible on older software versions.
 scrub: none requested
config:

NAME  STATE READ WRITE CKSUM
rpool ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c0t0d0s0  ONLINE   0 0 0
c0t1d0s0  ONLINE   0 0 0

errors: No known data errors

  pool: tank
 state: ONLINE
status: The pool is formatted using an older on-disk format.  The pool can
still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
pool will no longer be accessible on older software versions.
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
tankONLINE   0 0 0
  raidz1ONLINE   0 0 0
c2t0d0  ONLINE   0 0 0
c0t2d0  ONLINE   0 0 0
c0t3d0  ONLINE   0 0 0
c0t4d0  ONLINE   0 0 0
c0t5d0  ONLINE   0 0 0
c0t6d0  ONLINE   0 0 0
c0t7d0  ONLINE   0 0 0

errors: No known data errors
[r...@nas01 ~]# zpool list
NAMESIZE   USED  AVAILCAP  HEALTH  ALTROOT
rpool   149G  11.4G   138G 7%  ONLINE  -
tank   9.50T  9.34T   159G98%  ONLINE  -
[r...@nas01 ~]#
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Peter Tribble
On Sat, Mar 28, 2009 at 8:12 AM, Michael Shadle mike...@gmail.com wrote:
 I currently have a 7x1.5tb raidz1.

 I want to add phase 2 which is another 7x1.5tb raidz1

 Can I add the second phase to the first phase and basically have two
 raid5's striped (in raid terms?)

Yes, that's how it's done.

 Yes, I probably should upgrade the zpool format too. Currently running
 snv_104. Also should upgrade to 110.

 If that is possible, would anyone happen to have the simple command
 lines to do it quick? I assume I'd be creating another raidz1 and then
 somehow growing the tank zpool?

zpool add tank raidz1 disk_1 disk_2 disk_3 ...

(The syntax is just like creating a pool, only with add instead of create.)

 Does this make sense, or is this stupid from a performance
 perspective? Should I just have two separate zpools? Ideally I would
 like to have one massive data storage target.

It makes perfect sense. My thumpers have a number of raidz vdevs combined
into a single pool. Your performance scales with the number of vdevs, and
its better to combine them into a single pool as you combine the performance.
Generally, unless you want different behaviour from different pools, it's easier
to combine them.

-- 
-Peter Tribble
http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Peter Tribble
On Sat, Mar 28, 2009 at 11:06 AM, Michael Shadle mike...@gmail.com wrote:
 On Sat, Mar 28, 2009 at 1:37 AM, Peter Tribble peter.trib...@gmail.com 
 wrote:

 zpool add tank raidz1 disk_1 disk_2 disk_3 ...

 (The syntax is just like creating a pool, only with add instead of create.)

 so I can add individual disks to the existing tank zpool anytime i want?

Yes, but you wouldn't want to do that. (And zpool might not like it.)

If you just add a disk, it just gets added as a new device. So you have
unprotected storage.

In particular, you can't grow the existing raidz. What you're doing
here is adding
a second raidz1 vdev. That's good because the 2nd phase of your storage is
just like the first phase.

 It makes perfect sense. My thumpers have a number of raidz vdevs combined
 into a single pool. Your performance scales with the number of vdevs, and
 its better to combine them into a single pool as you combine the performance.
 Generally, unless you want different behaviour from different pools, it's 
 easier
 to combine them.

 so essentially you're tleling me to keep it at raidz1 (not raidz2 as
 many people usually stress when getting up to a certain # of disks,
 like 8 or so most people start bringing it up a lot)

The choice of raidz1 versus raidz2 is another matter. Given that you've already
got raidz1, and you can't (yet) grow that or expand it to raidz2, then
there doesn't
seem to be much point to having the second half of your storage being more
protected.

If you were starting from scratch, then you have a choice between a single
raidz2 vdev and a pair of raidz1 vdevs. (Lots of other choices too,
but that is really
what you're asking here I guess.)

With 1.5T drives, I would want a second layer of protection. If you didn't have
backups (by which I mean an independent copy of your important data) then
raidz2 is strongly indicated.

I have a thumper that's a primary fileserver. It has a single pool
that is made up
of a number of raidz2 vdevs. I'm replacing it with a pair of machines
(this gives
me system redundancy as well); each one will have a number of raidz1 vdevs
because I can always get the data back off the other machine if something
goes wrong.

-- 
-Peter Tribble
http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam

Harry wrote:

 However I hit a snag right away.  Pulled a 200gb added a 750 but on
 boot when I get to that screen I see:

Press F3 to enter configuration utility
Primary channel:  WDC WD200-blah-blah   190782 MB
  Primary secondary:  WDC WD750-blah-blah

 Note the size is missing.  And there the boot hangs.. forever.
 Pressing F3 has no effect whatever.


Ian Collins i...@ianshome.com replied:
   
 It looks like the card BOIS has written to the dives and expects the
 same drive to be there.

Harry wrote:

 So it appears I'm stuck.  I'm thinking the problem is that the adaptec
 for some reason needs to be able to report a size and that is whats
 hanging things up.

 So I wondered if it might get me past this hang if I were to plug that
 new 750 into a windows/or linux machine and  format it, either as ntfs
 on windows or just fdisk a partition of the entire drive on linux.


Ian replied:
   
 Unlikely, ZFS will re-label the drive an the card should be able to
 work with new drives.

How would this relabeling occur.  Somehow I need to get past that
screen right?

 Check the BIOS documentation for the card.

I mentioned that pressing F3 doesn't do anything.  That is, I have
now way to enter the configuration tool.

I'm not sure what you mean then, except to maybe learn if this card
will not work for these drives.



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Casper . Dik


I mentioned that pressing F3 doesn't do anything.  That is, I have
now way to enter the configuration tool.


Does it work when you first remove the 200GB drive, reboot, press F3
and see what you can do in that configuration tool.

It is possible that it first needs to forget the 200GB drive.

Casper

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Michael Shadle
On Sat, Mar 28, 2009 at 4:30 AM, Peter Tribble peter.trib...@gmail.com wrote:

 so I can add individual disks to the existing tank zpool anytime i want?

 Yes, but you wouldn't want to do that. (And zpool might not like it.)

 If you just add a disk, it just gets added as a new device. So you have
 unprotected storage.

so you're saying i should add 7 disks to match the existing setup (or
at least 2 disks so it has some sort of redundancy)

and i would run zpool add tank raidz1 disk1 disk2 disk3 disk4 disk5
disk6 disk7 ... if my goal is to use 7 disks. this would allow it to
become part of one large storage pool with two identical types of
redundancy setups (separate from each other, like two physically
different raidsets combined, though, which is fine)

 In particular, you can't grow the existing raidz. What you're doing
 here is adding
 a second raidz1 vdev. That's good because the 2nd phase of your storage is
 just like the first phase.

i guess this is redundant, but would i be able to see these as one
large storage pool, or would i essentially have tank and tank2?

is there a way to combine them? just the command above?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
casper@sun.com writes:

 Does it work when you first remove the 200GB drive, reboot, press F3
 and see what you can do in that configuration tool.



 It is possible that it first needs to forget the 200GB drive.

Might be the way.  Seems likely given the way you've shown how zfs can
remember the real size of a mirror when dealing with different size
drives in a mirror.

I'll try that next.

I've also been told there are sometimes a jumper necessary.  The
adaptec specs show this controller at 1.5 transfer rate and the drives
are sata II 3.gb.  I found some info about it suggesting jumpering
pins 5-6 would set it to 1.5gb so tried that, but got the same hang
at PCI sata card screen as before.  And as before pressing F3 does
nothing.



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
casper@sun.com writes:

I mentioned that pressing F3 doesn't do anything.  That is, I have
now way to enter the configuration tool.


 Does it work when you first remove the 200GB drive, reboot, press F3
 and see what you can do in that configuration tool.

 It is possible that it first needs to forget the 200GB drive.

OK, booted with no sata drives attached.  At the press F3 screen,
again pressing F3 appears to have no effect.  But after a moment two
messages saying `no drives found' print out, and then boot proceeds.

Once booted up I see the recurring message where I should see a login
prompt (I'm setup to boot into console mode).

  ata_id_common Busy Status 0xfe error 0x0  

Repeated 4 times, then after maybe a 2-3 minute wait the regular login
prompt appears.

There is something more going on that is dragging the machine down so
that typing at the prompt is delayed by huge pauses.  

I'm not really able to find out what it is yet but will post back
later after a reboot.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Blake
Have you checked the specs of the 1205 to see what maximum drive size
it supports?  That's an older card, IIRC, so it might top out at 500gb
or something.

On Sat, Mar 28, 2009 at 10:33 AM, Harry Putnam rea...@newsguy.com wrote:
 casper@sun.com writes:

I mentioned that pressing F3 doesn't do anything.  That is, I have
now way to enter the configuration tool.


 Does it work when you first remove the 200GB drive, reboot, press F3
 and see what you can do in that configuration tool.

 It is possible that it first needs to forget the 200GB drive.

 OK, booted with no sata drives attached.  At the press F3 screen,
 again pressing F3 appears to have no effect.  But after a moment two
 messages saying `no drives found' print out, and then boot proceeds.

 Once booted up I see the recurring message where I should see a login
 prompt (I'm setup to boot into console mode).

  ata_id_common Busy Status 0xfe error 0x0

 Repeated 4 times, then after maybe a 2-3 minute wait the regular login
 prompt appears.

 There is something more going on that is dragging the machine down so
 that typing at the prompt is delayed by huge pauses.

 I'm not really able to find out what it is yet but will post back
 later after a reboot.


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Bob Friesenhahn

On Sat, 28 Mar 2009, Peter Tribble wrote:


The choice of raidz1 versus raidz2 is another matter. Given that 
you've already got raidz1, and you can't (yet) grow that or expand 
it to raidz2, then there doesn't seem to be much point to having the 
second half of your storage being more protected.


There is no harm from using a raidz2 vdev even if an existing vdev is 
only raidz1.  If raidz2 is an available option then it is wise to 
choose it.  Of course starting out with raidz2 would have been even 
better.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] j4200 drive carriers

2009-03-28 Thread Blake
This is true.  Unfortunately, in my experience, controller quality is
still very important.  ZFS can preserve data all day long, but that
doesn't help much if the controller misbehaves (you may have good data
that can't be retrieved or manipulated properly - it's happened to me
with whitebox hardware).

If anyone buys whitebox hardware for ZFS in production, make sure the
vendor will give you support/warranty for OpenSolaris/ZFS.



On Fri, Mar 27, 2009 at 7:33 PM, jpdrawneek j...@drawneek.demon.co.uk wrote:
 Mertol Ozyoney wrote:

 But the whole point of zfs is that you can use inexpensive drives and with
 enough in RAID to make it reliable.

 Best regards
 Mertol

 Mertol Ozyoney Storage Practice - Sales Manager

 Sun Microsystems, TR
 Istanbul TR
 Phone +902123352200
 Mobile +905339310752
 Fax +90212335
 Email mertol.ozyo...@sun.com



 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Richard Elling
 Sent: Sunday, February 01, 2009 10:24 PM
 To: John-Paul Drawneek
 Cc: zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] j4200 drive carriers

 John-Paul Drawneek wrote:


 the J series is far to new to be hitting ebay yet.

 Any alot of people will not be buying the J series for obvious reasons


 The obvious reason is that Sun cannot service random disk
 drives you buy from Fry's (or elsewhere). People who value data
 tend to value service contracts for disk drives.
 -- richard

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] is zpool export/import | faster than rsync or cp

2009-03-28 Thread Blake
zfs send/recv *is* faster (especially since b105) than rsync,
especially when you are dealing with lots of small files.  rsync has
to check each file, which can take a long time - zfs send/recv just
moves blocks.

2009/3/27 Ahmed Kamal email.ahmedka...@googlemail.com:
 ZFS replication basics at http://cuddletech.com/blog/pivot/entry.php?id=984
 Regards

 On Sat, Mar 28, 2009 at 1:57 AM, Harry Putnam rea...@newsguy.com wrote:

 [...]

 Harry wrote:
  Now I'm wondering if the export/import sub commands might not be a
  good bit faster.
 
 Ian Collins i...@ianshome.com answered:
  I think you are thinking of zfs send/receive.
 
  I've never done a direct comparison, but zfs send/receive would be my
  preferred way to move data between pools.

 Why is that?  I'm too new to know what all it encompasses (and a bit
 dense to boot)

 Fajar A. Nugraha fa...@fajar.net writes:

  On Sat, Mar 28, 2009 at 5:05 AM, Harry Putnam rea...@newsguy.com
  wrote:
  Now I'm wondering if the export/import sub commands might not be a
  good bit faster.
 
  I believe the greatest advantage of zfs send/receive over rsync is not
  about speed, but rather it's on zfs send -R, which would (from man
  page)
 
               Generate a replication stream  package,  which  will
               replicate  the specified filesystem, and all descen-
               dant file systems, up to the  named  snapshot.  When
               received, all properties, snapshots, descendent file
               systems, and clones are preserved.
 
  pretty much allows you to clone a complete pool preserving its
  structure.
  As usual, compressing the backup stream (whether rsync or zfs) might
  help reduce transfer time a lot. My favorite is lzop (since it's very
  fast), but gzip should work as well.
 

 Nice... good reasons it appears.


 Robert Milkowski mi...@task.gda.pl writes:

  Hello Harry,

 [...]

  As Ian pointed you want zfs send|receive and not import/export.
  For a first full copy zfs send not necessarily will be noticeably
  faster than rsync but it depends on data. If for example you have
  milions of small files zfs send could be much faster then rsync.
  But it shouldn't be slower in any case.
 
  zfs send|receive really shines when it comes to sending incremental
  changes.

 Now that would be something to make it stand out.  Can you tell me a
 bit more about that would work..I mean would you just keep receiving
 only changes at one end and how do they appear on the filesystem.

 There is a backup tool called `rsnapshot' that uses rsync but creates
 hard links to all unchanged files and moves only changes to changed
 files.  This is all put in a serial directory system and ends up
 taking a tiny fraction of the space that full backups would take, yet
 retains a way to get to unchanged files right in the same directory
 (the hard link).

 Is what your talking about similar in some way.

 =     *     =     *     =     *     =

 To all posters... many thanks for the input.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Tim
On Sat, Mar 28, 2009 at 10:29 AM, Bob Friesenhahn 
bfrie...@simple.dallas.tx.us wrote:

 On Sat, 28 Mar 2009, Peter Tribble wrote:


 The choice of raidz1 versus raidz2 is another matter. Given that you've
 already got raidz1, and you can't (yet) grow that or expand it to raidz2,
 then there doesn't seem to be much point to having the second half of your
 storage being more protected.


 There is no harm from using a raidz2 vdev even if an existing vdev is only
 raidz1.  If raidz2 is an available option then it is wise to choose it.  Of
 course starting out with raidz2 would have been even better.

 Bob


#1: yes, there is harm as he may very well run into inconsistent performance
which is a complete PITA to track down when you've got differing raidtypes
underlying a volume.
#2: raidz2 isn't always wise to choose.  It's a matter of performance,
space, security requirements.  7+1 is fine for raidz1.  If he was pushing 10
data disks that'd be another story.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Jonathan
Michael Shadle wrote:
 On Sat, Mar 28, 2009 at 1:37 AM, Peter Tribble
peter.trib...@gmail.com wrote:

 zpool add tank raidz1 disk_1 disk_2 disk_3 ...

 (The syntax is just like creating a pool, only with add instead of
create.)

 so I can add individual disks to the existing tank zpool anytime i want?

Using the command above that Peter gave you would get you a result
similar to this

NAMESTATE READ WRITE CKSUM
storage2ONLINE   0 0 0
  raidz1ONLINE   0 0 0
ad16ONLINE   0 0 0
ad14ONLINE   0 0 0
ad10ONLINE   0 0 0
ad12ONLINE   0 0 0
  raidz1ONLINE   0 0 0
da2 ONLINE   0 0 0
da0 ONLINE   0 0 0
da1 ONLINE   0 0 0
da3 ONLINE   0 0 0

The actual setup is a RAIDZ1 of 1.5TB drives and a RAIDZ1 of 500GB
drives with the data striped across the two RAIDZs.  In your case it
would be 7 drives in each RAIDZ based on what you said before but I
don't have *that* much money for my home file server.

 so essentially you're tleling me to keep it at raidz1 (not raidz2 as
 many people usually stress when getting up to a certain # of disks,
 like 8 or so most people start bringing it up a lot)

This really depends on how valuable your data is.  Richard Elling has a
lot of great information about MTTDL here
http://blogs.sun.com/relling/tags/mttdl
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
Blake blake.ir...@gmail.com writes:

 Have you checked the specs of the 1205 to see what maximum drive size
 it supports?  That's an older card, IIRC, so it might top out at 500gb
 or something.

I did yes, all it really says is that is supports drives larger than
137gb.

   http://www.adaptec.com/en-US/support/sata/sata_host/ASH-1205SA/

But it also says its top end transfer rate is 1.5 gb.
The new drives are 3.0gb.  

I'm told there is a possibility of setting a jumper to make sata II
work as sata I.  I've tried one jumper position I found some info on
here:
http://www.google.com/search?hl=enrlz=1G1GGLQ_ENUS273q=WD+Sata+II+jumpersbtnG=Search

The direct URL is too long to post here but the one above will get you
there.  Its the top one on the page there.  And then Figure 4 in the
pdf.

I tried jumping 5-6 but still got the same behavior when the Sata
controller screen (Press F3 to configure) showed up.  It sees the
drive by name but not the size and there it hangs.

Casper has suggested I pull both old sata drives and reboot, possibly
allowing the hardware to `forget' those drives.

I'm doing that now as reported earlier in the thread.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Bob Friesenhahn

On Sat, 28 Mar 2009, Tim wrote:


#1: yes, there is harm as he may very well run into inconsistent performance
which is a complete PITA to track down when you've got differing raidtypes
underlying a volume.


Inconsistent performance can come from many things, including a single 
balky disk drive.  The small difference between RAID types does not 
seem like enough to worry about.  If it was a mix between raidz2 and 
mirrors then there is more cause for concern.


It is true that if the performance of the vdevs is not well balanced, 
then some vdevs could fill up faster than others when the system is 
under extremely heavy write loads.



#2: raidz2 isn't always wise to choose.  It's a matter of performance,
space, security requirements.  7+1 is fine for raidz1.  If he was pushing 10
data disks that'd be another story.


Many in the industry have already declared RAID5 to be unsafe at any 
speed with today's huge SATA disk drives.  The data recovery model 
for raidz1 is similar to RAID5.  If the user can afford it, then 
raidz2 offers considerably more peace of mind.


If you are using 750GB+ SATA drives then your 7+1 is fine for raidz1 
notion does not seem so bright.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
Harry Putnam rea...@newsguy.com writes:

 Once booted up I see the recurring message where I should see a login
 prompt (I'm setup to boot into console mode).

   ata_id_common Busy Status 0xfe error 0x0  

 Repeated 4 times, then after maybe a 2-3 minute wait the regular login
 prompt appears.

 There is something more going on that is dragging the machine down so
 that typing at the prompt is delayed by huge pauses.  

 I'm not really able to find out what it is yet but will post back
 later after a reboot.

Those 4 lines continue to appear on ever boot now.. (still with old
200gb sata drives out.

Seems when the boot prompt does appear there is something dragging
down the os making logging in really sluggish.

It appears to be the fmd daemon running periodically and driving
resource drain on cpu up to 99 percent.

That seems to continue to occur for at least several cycles watching
it in `top'.  They last maybe close to 1 minute then disappear for a
while.  Several minutes at least.  The cylces seem to keep coming
after being booted up for 10 minutes now.

I'm not sure which logs to look in to see whats happing.  But earlier
today these showed up in /var/adm/messages:

Mar 28 09:38:25 zfs fmd: [ID 441519 daemon.error] SUNW-MSG-ID:
ZFS-8000-D3, TYPE: Fault, VER: 1, SEVERITY: Major
Mar 28 09:38:25 zfs IMPACT: Fault tolerance of the pool may be
compromised.

The console login continues to be nearly useless with the delays and
pauses while typing.

However ssh in and the terminal I get seem to be less effected, or
even un-effected so I can do things in some kind or reasonable way.

I'm really not sure what to do though.  I did 
  `zpool destroy thatpool' on the pool that was on the sata drives.

That appears to have worked, but didn't help with resource drain
coming from `fmd'

The tail of the log pointed to by svcs -l system/fmd shows:
(/var/svc/log/system-fmd:default.log)
[...]
[ Mar 28 09:20:31 Enabled. ]
[ Mar 28 09:25:11 Executing start method (/usr/lib/fm/fmd/fmd). ]
[ Mar 28 09:28:24 Method start exited with status 0. ]
[ Mar 28 09:53:50 Enabled. ]
[ Mar 28 09:54:50 Executing start method (/usr/lib/fm/fmd/fmd). ]
[ Mar 28 09:55:44 Method start exited with status 0. ]
[ Mar 28 09:59:21 Stopping because service disabled. ]
[ Mar 28 09:59:21 Executing stop method (:kill). ]
[ Mar 28 10:01:09 Enabled. ]
[ Mar 28 10:02:17 Executing start method (/usr/lib/fm/fmd/fmd). ]
[ Mar 28 10:02:49 Method start exited with status 0. ]

I have no idea is that is normal or what.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Bob Friesenhahn

On Sat, 28 Mar 2009, Jonathan wrote:


This really depends on how valuable your data is.  Richard Elling has a
lot of great information about MTTDL here
http://blogs.sun.com/relling/tags/mttdl


Almost any data with a grade higher than disposable junk becomes 
pretty valuable once you consider the time and effort to recover it. 
Multiple terrabytes of data takes quite a long time to recover at 
perhaps a few tens of megabytes per second and critical backup 
resources are consumed in the mean time.  Meanwhile, business goes on.


Do the math in advance and decide if you are really willing to put 
yourself in the middle of a long recovery situation.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Michael Shadle
2009/3/28 Tim t...@tcsac.net:

 There is no harm from using a raidz2 vdev even if an existing vdev is only
 raidz1.  If raidz2 is an available option then it is wise to choose it.  Of
 course starting out with raidz2 would have been even better.

 #2: raidz2 isn't always wise to choose.  It's a matter of performance,
 space, security requirements.  7+1 is fine for raidz1.  If he was pushing 10
 data disks that'd be another story.

if i went raidz2 i'd want the entire 14 disk array in it i think.

i'd rather not do a raidz2 with less than 100% of the disks and then a
second raidz1 (or 2) because i'd wind up losing much more disk space.
essentially, i am willing to give up 2 of 14 disks (roughly of course)
to parity.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Bob Friesenhahn

On Sat, 28 Mar 2009, Michael Shadle wrote:

if i went raidz2 i'd want the entire 14 disk array in it i think.

i'd rather not do a raidz2 with less than 100% of the disks and then a
second raidz1 (or 2) because i'd wind up losing much more disk space.
essentially, i am willing to give up 2 of 14 disks (roughly of course)
to parity.


Hopefully you consider all of the costs before making this sort of 
decision.  If you are a lousy tipper you can't expect very good 
service the next time you come to visit. :-)


If 14 disks cost a lot, then you should carefully balance the cost 
of the wasted disk against the cost of lost performance or the cost 
of lost availability.  In many business environments, the potential 
for lost availability more than justifies purchasing more wasted 
disk.  In many business environments, the potential for lousy 
performance more than justifies purchasing more wasted disk.  Any 
good businessman should be able to specify a dollars per hour cost 
to the business if the storage is not available, or unable to provide 
sufficient performance to meet business needs.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Blake
what's the output of 'fmadm faulty'?

On Sat, Mar 28, 2009 at 12:22 PM, Harry Putnam rea...@newsguy.com wrote:
 Harry Putnam rea...@newsguy.com writes:

 Once booted up I see the recurring message where I should see a login
 prompt (I'm setup to boot into console mode).

   ata_id_common Busy Status 0xfe error 0x0

 Repeated 4 times, then after maybe a 2-3 minute wait the regular login
 prompt appears.

 There is something more going on that is dragging the machine down so
 that typing at the prompt is delayed by huge pauses.

 I'm not really able to find out what it is yet but will post back
 later after a reboot.

 Those 4 lines continue to appear on ever boot now.. (still with old
 200gb sata drives out.

 Seems when the boot prompt does appear there is something dragging
 down the os making logging in really sluggish.

 It appears to be the fmd daemon running periodically and driving
 resource drain on cpu up to 99 percent.

 That seems to continue to occur for at least several cycles watching
 it in `top'.  They last maybe close to 1 minute then disappear for a
 while.  Several minutes at least.  The cylces seem to keep coming
 after being booted up for 10 minutes now.

 I'm not sure which logs to look in to see whats happing.  But earlier
 today these showed up in /var/adm/messages:

 Mar 28 09:38:25 zfs fmd: [ID 441519 daemon.error] SUNW-MSG-ID:
 ZFS-8000-D3, TYPE: Fault, VER: 1, SEVERITY: Major
 Mar 28 09:38:25 zfs IMPACT: Fault tolerance of the pool may be
 compromised.

 The console login continues to be nearly useless with the delays and
 pauses while typing.

 However ssh in and the terminal I get seem to be less effected, or
 even un-effected so I can do things in some kind or reasonable way.

 I'm really not sure what to do though.  I did
  `zpool destroy thatpool' on the pool that was on the sata drives.

 That appears to have worked, but didn't help with resource drain
 coming from `fmd'

 The tail of the log pointed to by svcs -l system/fmd shows:
 (/var/svc/log/system-fmd:default.log)
 [...]
 [ Mar 28 09:20:31 Enabled. ]
 [ Mar 28 09:25:11 Executing start method (/usr/lib/fm/fmd/fmd). ]
 [ Mar 28 09:28:24 Method start exited with status 0. ]
 [ Mar 28 09:53:50 Enabled. ]
 [ Mar 28 09:54:50 Executing start method (/usr/lib/fm/fmd/fmd). ]
 [ Mar 28 09:55:44 Method start exited with status 0. ]
 [ Mar 28 09:59:21 Stopping because service disabled. ]
 [ Mar 28 09:59:21 Executing stop method (:kill). ]
 [ Mar 28 10:01:09 Enabled. ]
 [ Mar 28 10:02:17 Executing start method (/usr/lib/fm/fmd/fmd). ]
 [ Mar 28 10:02:49 Method start exited with status 0. ]

 I have no idea is that is normal or what.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs scheduled replication script?

2009-03-28 Thread Fajar A. Nugraha
I have a backup system using zfs send/receive (I know there are pros
and cons to that, but it's suitable for what I need).

What I have now is a script which runs daily, do zfs send, compress
and write it to a file, then transfer it with ftp to a remote host. It
does full backup every 1st, and do incremental (with 1st as reference)
after that. It works, but not quite resource-effective (for example,
the full backup every month, and the big size of incremental backup on
30th).

I'm thinking of changing it to a script which can automate replication
of a zfs pool or filesystem via zfs send/receive to a remote host (via
ssh or whatever). It should be smart enough to choose between full and
incremental, and choose which snapshot to base the incremental stream
from (in case a scheduled incremental is missed), preferably able to
use snapshots created by zfs/auto-snapshot smf service.

To prevent re-inventing the wheel, does such script exists already?
I prever not to use AVS as I can't use on existing zfs pool.

Regards,

Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
Blake blake.ir...@gmail.com writes:

 what's the output of 'fmadm faulty'?

It hangs for maybe 30 seconds, then returns the prompt.
And again now on this boot, before the console prompt appears I see:
   ata_id_common: Busy Status 0xfe error 0x0

Repeated 4 times.

And again, the console prompt is barely usable.

 login type username return

A full minute or so later, the passwd part comes up

Enter passwd and another minute or so to login.

I don't see fmd making huge cpu resource drain so not sure what is
dragging down the login prompt.

`top' shows 99.6 % idle even 99.8 but still the login prompt is very
sluggish. 

I can't really even work with it, even after finally getting logged in
its just too aggravating to wait that minute or so between typing
something and seeing it appear on the cmd line.

Fortunately ssh works and that terminal seems un-affected. 

But it appears I have done major damage somehow by pulling out those 2
drives.  Or something else I've unwittingly done.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] j4200 drive carriers

2009-03-28 Thread Miles Nordin
 b == Blake  blake.ir...@gmail.com writes:

 b ZFS can preserve data all day long, but that doesn't help much
 b if the controller misbehaves

the most common kind of controller, or rather driver, misbehavior is
to time out commands to failed drives while ZFS waits inappropriately
long, or to freeze access to drives other than the drive that failed,
or to fail to support hot-swap properly.  These problems affect
availability (no interruptions in data access), not reliability (once
put there, stuff stays there, but maybe only readable after manual
intervention).  If the box is doing a ~throwaway job like tape
emulation for backup, good reliability with crap availability may be
more acceptable than the same situation on a production box.

 b it's happened to me with whitebox hardware).

and also with X4500, taking 1+ years to not completely fix, and
closed-source driver so you cannot try to fix it yourself.  though
there seem to be fewer complaints with the new LSI-based controllers.
Paying a premium for good integration makes more sense if the
integration was actually good in the past.  I'd also pay a premium for
something like VA Linux which bundled hardware with stable revisions
of good-quality open-source drivers, but I guess that didn't work out
well for VA.

It makes some sense but less sense to pay a premium for a situation in
which the integration could theoretically be good even though it
hasn't been in the past, because there may be FUD problems that were
solved (both presently and in the past) without your hearing about
them so you are actually getting something for the premium you've
paid, but $vendor cannot tell you exactly what you are getting because
knowing what problems were fixed would make it easier for you to fix
them yourself without paying.  That's a pretty damn cynical outlook on
the situation, but it seems to be a realistic/common one.

That said why fuss about drive cost?  If the drive-to-chassis lock-in
is effective, then just accept the tying as done and compare cost/TB
with Xtore or Dell FC or whatever.  The real interop/tying thing I see
needing untangling in ZFS is for FC and iSCSI to work well---if this
were done then ZFS could be combined with a variety of competing
physical storage, and it's reasonable to expect it be well-integrated
with non-Sun FC/iSCSI---I think the claim ``well there is FUD so you
have to buy Sun FC targets'' will not fly so well in that case because
you are already paying an integration premium to both vendors and
might reasonably expect some combinations of them to claim and deliver
good interoperability.


pgp2XIRwL36hQ.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Tim
On Sat, Mar 28, 2009 at 11:12 AM, Bob Friesenhahn 
bfrie...@simple.dallas.tx.us wrote:

 On Sat, 28 Mar 2009, Tim wrote:


 #1: yes, there is harm as he may very well run into inconsistent
 performance
 which is a complete PITA to track down when you've got differing raidtypes
 underlying a volume.


 Inconsistent performance can come from many things, including a single
 balky disk drive.  The small difference between RAID types does not seem
 like enough to worry about.  If it was a mix between raidz2 and mirrors then
 there is more cause for concern.

 It is true that if the performance of the vdevs is not well balanced, then
 some vdevs could fill up faster than others when the system is under
 extremely heavy write loads.


An extra parity disk is hardly a small difference.  You pay your penalty
at some point for the extra parity, and it will come back to bite you in the
ass.  It's why NOBODY including Sun, supports it on enterprise arrays.




  #2: raidz2 isn't always wise to choose.  It's a matter of performance,
 space, security requirements.  7+1 is fine for raidz1.  If he was pushing
 10
 data disks that'd be another story.


 Many in the industry have already declared RAID5 to be unsafe at any
 speed with today's huge SATA disk drives.  The data recovery model for
 raidz1 is similar to RAID5.  If the user can afford it, then raidz2 offers
 considerably more peace of mind.

 If you are using 750GB+ SATA drives then your 7+1 is fine for raidz1
 notion does not seem so bright.


Many in the industry make their money selling you disk drives, of course
they're going to declare you need to buy more.  The math you ask people to
do points towards a 7+1 being more than acceptable if you have a hot-spare.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread David Champion
* On 28 Mar 2009, Peter Tribble wrote: 
 The choice of raidz1 versus raidz2 is another matter. Given that
 you've already got raidz1, and you can't (yet) grow that or expand
 it to raidz2, then there doesn't seem to be much point to having the
 second half of your storage being more protected.

 If you were starting from scratch, then you have a choice between a
 single raidz2 vdev and a pair of raidz1 vdevs. (Lots of other choices
 too, but that is really what you're asking here I guess.)

I've had too many joint failures in my life to put much faith in raidz1,
especially with 7 disks that likely come from the same manufacturing
batch and might exhibit the same flaws.  A single-redundancy system of 7
disks (gross) has too high a MTTDL for my taste.

If you can sell yourself on raidz2 and the loss of two more disks' worth
of data -- a loss which IMO is more than made up for by the gain in
security -- consider this technique:

1. build a new zpool of a single raidz2;
2. migrate your data from the old zpool to the new one;
3. destroy the old zpool, releasing its volumes;
4. use 'zpool add' to add those old volumes to the new zpool as a
   second raidz2 vdev (see Richard Elling's previous post).

Now you have a single zpool consisting of two raidz2 vdevs.

The migration in step 2 can be done either by 'zfs send'ing each zfs
in the zpool, or by constructing analogous zfs in the new zpool and
rsyncing the files across in one go.

-- 
 -D.d...@uchicago.eduNSITUniversity of Chicago
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs scheduled replication script?

2009-03-28 Thread Brent Jones
On Sat, Mar 28, 2009 at 11:20 AM, Fajar A. Nugraha fa...@fajar.net wrote:
 I have a backup system using zfs send/receive (I know there are pros
 and cons to that, but it's suitable for what I need).

 What I have now is a script which runs daily, do zfs send, compress
 and write it to a file, then transfer it with ftp to a remote host. It
 does full backup every 1st, and do incremental (with 1st as reference)
 after that. It works, but not quite resource-effective (for example,
 the full backup every month, and the big size of incremental backup on
 30th).

 I'm thinking of changing it to a script which can automate replication
 of a zfs pool or filesystem via zfs send/receive to a remote host (via
 ssh or whatever). It should be smart enough to choose between full and
 incremental, and choose which snapshot to base the incremental stream
 from (in case a scheduled incremental is missed), preferably able to
 use snapshots created by zfs/auto-snapshot smf service.

 To prevent re-inventing the wheel, does such script exists already?
 I prever not to use AVS as I can't use on existing zfs pool.

 Regards,

 Fajar
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


The ZFS automatic snapshot tools have the ability to execute any
command after a snapshot has taken place, such as ZFS send.
However, I ran into some issues, as it wasn't strictly designed for
that and didn't handle errors very well.

I have since modified some scripts out there, and rolled them into my
own, you can see it here at pastebin.com:

http://pastebin.com/m3871e478

Inspiration:
http://blogs.sun.com/constantin/entry/zfs_replicator_script_new_edition
http://blogs.sun.com/timf/en_IE/entry/zfs_automatic_snapshots_in_nv

Those are some good resources, from that, you can make something work
that is tailored to your environment.

-- 
Brent Jones
br...@servuhome.net
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
casper@sun.com writes:

I mentioned that pressing F3 doesn't do anything.  That is, I have
now way to enter the configuration tool.


 Does it work when you first remove the 200GB drive, reboot, press F3
 and see what you can do in that configuration tool.

 It is possible that it first needs to forget the 200GB drive.

It was worth a try... but as posted further along in the thread I ran
into a bunch of mess going that route.  Booting without the drives
causes several kinds of things to happen that I don't understand.

bootup passes the `Press F3 screen' Ok after a brief pause.

But when I get to the consol login I see a line like:

  ata_id_common Busy Status 0xfe error 0x0

Repeated 4 times then after a 30 second or so pause comes the login
prompt.  But the prompt is very unresponsive.  I mean if I type into
it takes nearly a full minute or more to show up

So after taking a couple minutes getting logged in, any activity at
the command line remains very sluggish.

Fortunately though, I can ssh into the box and that terminal seems to
be un-effected.

Watching with `top' I see the fmd daemon running in a recurring cycle
and each time it comes on it drives the resource usage up into the
high 90s for a minute or two.  It appears to continue to cycle for at
least 15 minutes at which point I rebooted again.

It appears that if I boot without the drives installed I will continue
to have this happen.  But re-insert the 2 drives in the original order
and it all goes away. 

Some of this may be repititious since I've been dicussing the sata
card on another (non-osol [gentoo.user]) list too and have kind of
lost track of what all got posted where.

But first, both of you have really been patient and taken a fair bit
of time to explain things... and I really appreciate it... so thanks. 


Blake blake.ir...@gmail.com writes:

 what's the output of 'fmadm faulty'?

You may have already seen my answer to that, but wanted to let you
both know that I've learned that at least one other computer user has
posted information to the effect that my card the Adaptec 1205sa which
apparently has an sil3112a chip is incompatable with Seagate 750 sata
drives ... mine are WD but its beginning to look like this card may
not support the newer sata drives.

I'm starting to think I may have just got the wrong card and should
just throw in the towel and get something known to work with my
hardware.

I did attempt to learn that before buying and did post at least
twice on the opensolaris groups trying to find out what would work for
sure on my hardware.  Either I didn't get replies or no one knew, I
don't remember which now.

Readers may have thought I was just too lazy to use the HCL but the
truth is there is precious little on it for my hardware.

But on the off chance either of you or maybe someone following along
will know of a sata card known to work with my hardware... I'm listing
the hardware here :

Athlon64 2.2ghz 3400+ - Aopen AK86-L mobo - Topped out at 3gb CORSAIR
ram

 (The onboard sata controller is not recognized by osol - hence
   the need for sata PCI card.).

Incidentally, I did find that Adaptec 1205sa on the HCL, listed by one
user who was able to boot with it installed and have drives
recognized.  The hardware was NOT a real match for mine but I took a
chance on it anyway... and it does work, at least with older/smaller
sata drives.

The specs on the 1205sa say that it will handle drives larger than
137gb and that its top transfer rate is 1.5gb.

So its not really clear what its' upper limit, if any, is far as size.
But it is an older card and may predate sata II.




 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Ian Collins

Blake wrote:

Have you checked the specs of the 1205 to see what maximum drive size
it supports?  That's an older card, IIRC, so it might top out at 500gb
or something.

  
Please stop top-posting to threads where everyone else is 
normal-posting, it mucks up the flow of the thread.


Thanks,

--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Michael Shadle
Well this is for a home storage array for my dvds and such. If I have  
to turn it off to swap a failed disk it's fine. It does not need to be  
highly available and I do not need extreme performance like a database  
for example. 45mb/sec would even be acceptable.


On Mar 28, 2009, at 10:47 AM, Bob Friesenhahn bfrie...@simple.dallas.tx.us 
 wrote:



On Sat, 28 Mar 2009, Michael Shadle wrote:

if i went raidz2 i'd want the entire 14 disk array in it i think.

i'd rather not do a raidz2 with less than 100% of the disks and  
then a

second raidz1 (or 2) because i'd wind up losing much more disk space.
essentially, i am willing to give up 2 of 14 disks (roughly of  
course)

to parity.


Hopefully you consider all of the costs before making this sort of  
decision.  If you are a lousy tipper you can't expect very good  
service the next time you come to visit. :-)


If 14 disks cost a lot, then you should carefully balance the cost  
of the wasted disk against the cost of lost performance or the  
cost of lost availability.  In many business environments, the  
potential for lost availability more than justifies purchasing more  
wasted disk.  In many business environments, the potential for  
lousy performance more than justifies purchasing more wasted  
disk.  Any good businessman should be able to specify a dollars per  
hour cost to the business if the storage is not available, or  
unable to provide sufficient performance to meet business needs.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Ian Collins

Harry Putnam wrote:


I'm starting to think I may have just got the wrong card and should
just throw in the towel and get something known to work with my
hardware.

  

That sounds like a good plan!

Using older SATA controllers with SATA II drives may work, but there 
aren't any guarantees.


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Harry Putnam
Ian Collins i...@ianshome.com writes:

 Harry Putnam wrote:

 I'm starting to think I may have just got the wrong card and should
 just throw in the towel and get something known to work with my
 hardware.

   
 That sounds like a good plan!

 Using older SATA controllers with SATA II drives may work, but there
 aren't any guarantees.

Finding a sataII card is proving to be very difficult.  The reason is
that I only have PCI no PCI express.  I haven't see a single one
listed as SATAII compatible and have spent a bit time googling.

I'd settle for a PCI card that someone can vouch for.  I did find a
message on an Australian overclockers list that indicated that fellow
had used sil3114 (mine is 3112a) card to run SATAII at 640gb.  But of
course those guys are all talking about `Windows' of one strain or
another so may or may not mean much.

I don't supposed you'd have a tip about a PCI sata card capable of
handling newer SATAII would you? 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Bob Friesenhahn

On Sat, 28 Mar 2009, Michael Shadle wrote:

Well this is for a home storage array for my dvds and such. If I have to turn 
it off to swap a failed disk it's fine. It does not need to be highly 
available and I do not need extreme performance like a database for example. 
45mb/sec would even be acceptable.


I can see that 14 disks costs a lot for a home storage array but to 
you the data on your home storage array may be just as important as 
data on some businesses enterprise storage array.  In fact, it may be 
even more critical since it seems unlikely that you will have an 
effective backup system in place like large businesses do.


The main problem with raidz1 is that if a disk fails and you replace 
it, that if a second disk substantially fails during resilvering 
(which needs to successfully read all data on remaining disks) then 
your ZFS pool (or at least part of the files) may be toast.  The more 
data which must be read during resilvering, the higher the probability 
that there will be a failure.  If 12TB of data needs to be read to 
resilver a 1TB disk, then that is a lot of successful reading which 
needs to go on.


In order to lessen risk, you can schedule a periodic zfs scrub via a 
cron job so that there is less probabily of encountering data which 
can not be read.  This will not save you from entirely failed disk 
drives though.


As far as Tim's post that NOBODY recommends using better than RAID5, I 
hardly consider companies like IBM and NetApp to be NOBODY.  Only 
Sun RAID hardware seems to lack RAID6, but Sun offers ZFS's raidz2 so 
it does not matter.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs scheduled replication script?

2009-03-28 Thread Fajar A. Nugraha
On Sun, Mar 29, 2009 at 3:40 AM, Brent Jones br...@servuhome.net wrote:
 I have since modified some scripts out there, and rolled them into my
 own, you can see it here at pastebin.com:

 http://pastebin.com/m3871e478

Thanks Brent.

Your script seems to handle failed replication and locking pretty well.
It doesn't seem to log WHY the replication failed though, so I think
there should be something that captures stderr on line 91.

One more question, is there anything on that script that requires ksh?
A quick glance seems to indicate that it will work with bash as well.

Regards,

Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zfs using java

2009-03-28 Thread Howard Huntley






Howard Huntley wrote:

  
  
I once installed ZFS on my home Sun Blade 100 and it worked fine on the
sun blade 100 running solaris 10. I later reinstalled Solaris 10 09
version
and created zpool which is not visible using the the java control
panel.
When I attempt to run the Java control panel to manage the zfs system I
receive an error message stating "!Launch Error, No
application is
registered with this sun Java or I have no rights to use any
applications that are registered, see my sys admin." Can any
one tell
me how to get this straightened out. I have been fooling around with it
for some time now.
  
Is any one is Jacksonville, Florida??
  -- 
  
  signature1
  Howard Huntley Jr. MCP, MCSE
Micro-Computer Systems Specialist
  


-- 

signature1
Howard Huntley Jr. MCP, MCSE
Micro-Computer Systems Specialist



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Frank Middleton

On 03/28/09 20:01, Harry Putnam wrote:


Finding a sataII card is proving to be very difficult.  The reason is
that I only have PCI no PCI express.  I haven't see a single one
listed as SATAII compatible and have spent a bit time googling.


It's even worse if you have an old SPARC system. We've had great results
with some LSI LOGIC SAS3041XL-S cards we got on E-Bay in conjunction
with 3x1.5TB Seagate drives, for 2.7TiB of raidz. The combination proved
faster than mirrored 10,000 RPM SCSI disks using UFS, in an unscientific
benchmark (bonnie). I don't think this LSI controller is SATA II, but
it has no problems with the 1.5TB Seagates; they are Sun branded cards
and worked right of of the box in 3.3V 66MHz PCI slots. 2.7TiB of raidz
for around $400. Amazing... And ZFS is just plain incredible - makes
every other file system look so antiquated :-)

Hope this helps -- Frank
 
 
___

zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Michael Shadle



On Mar 28, 2009, at 5:22 PM, Bob Friesenhahn bfrie...@simple.dallas.tx.us 
 wrote:



On Sat, 28 Mar 2009, Michael Shadle wrote:

Well this is for a home storage array for my dvds and such. If I  
have to turn it off to swap a failed disk it's fine. It does not  
need to be highly available and I do not need extreme performance  
like a database for example. 45mb/sec would even be acceptable.


I can see that 14 disks costs a lot for a home storage array but to  
you the data on your home storage array may be just as important as  
data on some businesses enterprise storage array.  In fact, it may  
be even more critical since it seems unlikely that you will have an  
effective backup system in place like large businesses do.




Well I might back up the more important stuff offsite. But in theory  
it's all replaceable. Just would be a pain.


Could I setup a raidz2 on the new zdev then destroy the old one and  
then raidz2 that technically if I want? Then both sets would have  
double redundancy, if I was feeling paranoid. But raid5 has served  
people well for a long time... Is resilvering speed roughly the same  
as a raid5 controller rebuild?


The main problem with raidz1 is that if a disk fails and you replace  
it, that if a second disk substantially fails during resilvering  
(which needs to successfully read all data on remaining disks) then  
your ZFS pool (or at least part of the files) may be toast.  The  
more data which must be read during resilvering, the higher the  
probability that there will be a failure.  If 12TB of data needs to  
be read to resilver a 1TB disk, then that


This is good info to know. I guess I'm willing to take the risk of a  
resilver. It's got a dedicated quad core proc doing nothing else than  
exporting samba and zfs... I wonder how long it would take.





In order to lessen risk, you can schedule a periodic zfs scrub via a  
cron job so that there is less probabily of encountering data which  
can not be read.  This will not save you from entirely failed disk  
drives though.




I do a weekly scrub and an fmadm faulty every 5 or 10 mins to email me  
if anything comes up...

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can this be done?

2009-03-28 Thread Scott Lawson



Bob Friesenhahn wrote:

On Sat, 28 Mar 2009, Michael Shadle wrote:

Well this is for a home storage array for my dvds and such. If I have 
to turn it off to swap a failed disk it's fine. It does not need to 
be highly available and I do not need extreme performance like a 
database for example. 45mb/sec would even be acceptable.


I can see that 14 disks costs a lot for a home storage array but to 
you the data on your home storage array may be just as important as 
data on some businesses enterprise storage array.  In fact, it may be 
even more critical since it seems unlikely that you will have an 
effective backup system in place like large businesses do.


The main problem with raidz1 is that if a disk fails and you replace 
it, that if a second disk substantially fails during resilvering 
(which needs to successfully read all data on remaining disks) then 
your ZFS pool (or at least part of the files) may be toast.  The more 
data which must be read during resilvering, the higher the probability 
that there will be a failure.  If 12TB of data needs to be read to 
resilver a 1TB disk, then that is a lot of successful reading which 
needs to go on.
This is a very good point for anyone following this and wondering why 
RAIDZ2 is a good idea. I have seen over the years several large RAID 5 
hardware arrays go belly
up as a 2nd drive fails during a rebuild with the end result of the 
entire RAID set being rendered useless. If you can afford it then you 
should use it. RAID6 or RAIDZ2 was
made for big SATA drives. If you do use it though, one should make sure 
that you have reasonable CPU as it does require a bit more grunt to run 
over RAIDZ.


The bigger the disks and the bigger the stripe the more likely you are 
to encounter a issue during a rebuild of a failed drive. plain and simple.


In order to lessen risk, you can schedule a periodic zfs scrub via a 
cron job so that there is less probabily of encountering data which 
can not be read.  This will not save you from entirely failed disk 
drives though.


As far as Tim's post that NOBODY recommends using better than RAID5, I 
hardly consider companies like IBM and NetApp to be NOBODY.  Only 
Sun RAID hardware seems to lack RAID6, but Sun offers ZFS's raidz2 so 
it does not matter.
Plenty of Sun hardware comes with RAID6 support out of the box these 
days Bob. Certainly all of the 4140, 4150, 4240 and 4250 2 socket x86 
/x64 systems have
hardware controllers for this. Also all of the 6140's, 6540 and 6780's 
disk arrays do also have RAID 6 if they have Crystal firmware and of 
course the Open Storage 7000 series

machines do as well being that they are Opensolaris and ZFS based.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, 
http://www.simplesystems.org/users/bfriesen/

GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--
_

Scott Lawson
Systems Architect
Information Communication Technology Services

Manukau Institute of Technology
Private Bag 94006
South Auckland Mail Centre
Manukau 2240
Auckland
New Zealand

Phone  : +64 09 968 7611
Fax: +64 09 968 7641
Mobile : +64 27 568 7611

mailto:sc...@manukau.ac.nz

http://www.manukau.ac.nz

__

perl -e 'print $i=pack(c5,(41*2),sqrt(7056),(unpack(c,H)-2),oct(115),10);'

__



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss