Re: [zfs-discuss] how to destroy a pool by id?

2009-06-21 Thread dick hoogendijk
On Sun, 21 Jun 2009 13:56:52 +1000 (EST)
Andre van Eyssen an...@purplecow.org wrote:

 On Sat, 20 Jun 2009, Cindy Swearingen wrote:
 
  I wish we had a zpool destroy option like this:
 
  # zpool destroy -really_dead tank2
 
 Cindy,
 
 The moment we implemented such a thing, there would be a rash of
 requests saying:
 
 a) I just destroyed my pool with -really_dead - how can I get my data 
 back??!
 b) I was able to recover my data from -really_dead - can we have 
 -ultra-nuke please?

Following your logic there shouldn't have existed a rm -f * option
too.

-- 
Dick Hoogendijk -- PGP/GnuPG key: 01D2433D
+ http://nagual.nl/ | nevada / OpenSolaris 2009.06 release
+ All that's really worth doing is what we do for others (Lewis Carrol)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mobo SATA migration to AOC-SAT2-MV8 SATA card

2009-06-21 Thread Simon Breden
There are probably too many questions in my last post, so I will post the 
questions as separate forum threads.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Simon Breden
Hi, I'm trying to find out which controller card people here recommend that can 
drive 8 SATA hard drives and that would work with my Asus M2N-SLI Deluxe 
motherboard, which has following expansion slots:
2 x PCI Express x16 slot at x16, x8 speed (PCIe)

The main requirements I have are:
- drive 8 SATA drives
- rock solid reliability with x86 OpenSolaris 2009.06 or SXCE
- easy to identify failed drives and replace them (hot swap is not necessary 
but a bonus if supported) 
- I must be able to move disks with data from one controller to another of 
different brands (and back!), only doing zpool export and import, which implies 
the HBA must be able to run in JBOD-mode without storing or modify anything on 
the disks. And preferably, the drives must show up with the format command.
- should support staggered spinup of drives preferably

From limited research I can see that at least the following 3 main 
possibilities exist:

1. Supermicro AOC-SAT2-MV8 (PCI-X interface) (pure SATA) (~$100)
2. Supermicro AOC-USAS-L8i / AOC-USASLP-L8i (PCIe interface) (miniSAS to SATA 
cables) (~$100)
3. LSI SAS 3081E-R or other LSI 'MegaRAID' cards (PCIe interface) (miniSAS to 
SATA cables) (~$200+)

1. AOC-SAT2-MV8 :
Again, from reading a bit, I can see that although the M2N-SLI Deluxe 
motherboard does not have a PCI-X slot, apparently it could take the 
AOC-SAT2-MV8 card in one of the PCIe slots, although the card would only run in 
32-bit mode, instead of 64-bit mode, and would therefore run slower.

2. AOC-USAS-L8i :
The AOC-USAS-L8i card looks possible too, again running in the PCIe slot, but 
the old threads I saw on this seem to talk about some device numbering issue 
which could make determining the right failed drive to pull out, a difficult 
task -- see here for more details:
http://www.opensolaris.org/jive/thread.jspa?messageID=271751
http://www.opensolaris.org/jive/thread.jspa?threadID=46982tstart=90

3. LSI SAS 3081E-R or other LSI 'MegaRAID' cards :
http://www.lsi.com/storage_home/products_home/host_bus_adapters/sas_hbas/lsisas3081er/index.html?remote=1locale
http://www.newegg.com/Product/Product.aspx?Item=N82E16816118100

This forum thread from DEC 2007 doesn't sound too good regarding drive 
numbering (for identifying failed drives etc), but the thread is 18 months old, 
and perhaps the issues may have been resolved now?
Also I noticed an extra '-R' in the model number I found, but this might be an 
omission of the original forum poster -- see here:
http://www.opensolaris.org/jive/thread.jspa?threadID=46982tstart=90

I saw Ben Rockwood saying good things about the LSI MegaRAID cards, although 
the model he references supports only 4 internal and 4 external drives so is 
not what I want -- see here:
http://opensolaris.org/jive/message.jspa?messageID=368445#368445

Perhaps there are better LSI MegaRAID cards that people know of and can 
recommend? Preferably not too expensive though, as it's for a home system :)

If anyone can throw some light on these topics, I would be pleased to hear from 
you. Thanks a lot.

Simon

http://breden.org.uk/2008/03/02/a-home-fileserver-using-zfs/
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread dick hoogendijk
On Sun, 21 Jun 2009 06:35:50 PDT
Simon Breden no-re...@opensolaris.org wrote:

 If anyone can throw some light on these topics, I would be pleased to
 hear from you. Thanks a lot.

I follow this thread with much interest.
Curious to see what'll come out of it.

-- 
Dick Hoogendijk -- PGP/GnuPG key: 01D2433D
+ http://nagual.nl/ | nevada / OpenSolaris 2009.06 release
+ All that's really worth doing is what we do for others (Lewis Carrol)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Simon Breden
After checking some more sources, it seems that if I used the AOC-SAT2-MV8 with 
this motherboard, I would need to run it on the standard PCI slot. Here is the 
full listing of the motherboard's expansion slots:

2 x PCI Express x16 slot at x16, x8 speed 
2 x PCI Express x1 
3 x PCI 2.2   --- this one
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Things I Like About ZFS

2009-06-21 Thread Peter Tribble
On Sat, Jun 20, 2009 at 3:30 AM, Dave Ringkorno-re...@opensolaris.org wrote:
 I'll start:

 - The commands are easy to remember -- all two of them.  Which is easier, SVM 
 or ZFS, to mirror your disks?  I've been using SVM for years and still have 
 to break out the manual to use metadb, metainit, metastat, metattach, 
 metadetach, etc.  I hardly ever have to break out the ZFS manual.  I can 
 actually remember the commands and options to do things.  Don't even start me 
 on VxVM.

Hehe. The simplicity is interesting. I'm actually starting to get confused by
those two commands, and start to wish it went down to one.

 - Boasting to the unconverted.  We still have a lot of VxVM and SVM on 
 Solaris, and LVM on AIX, in the office.  The other admins are always having 
 issues with storage migrations, full filesystems, Live Upgrade, corrupted 
 root filesystems, etc.  I love being able to offer solutions to their 
 immediate problems, and follow it up with, You know, if your box was on ZFS 
 this wouldn't be an issue.

Out of inodes. Huh? The very concept is so antiquated.

The great success of ZFS, to me, is the fact that it rapidly became essentially
invisible. It just does its job and you soon forget that it's there
(until you have to
deal with one of the alternatives, which throws it into sharp relief).

-- 
-Peter Tribble
http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] x4500 resilvering spare taking forever?

2009-06-21 Thread Joe Kearney
UPDATE: It's now back down to 0.9% complete.  Does anyone have a clue as to 
whats happening here or where I can look for problems?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] x4500 resilvering spare taking forever?

2009-06-21 Thread Joe Kearney
 Are you taking snapshots periodically? If so, you're
 using a build old
 enough to restart resilver/scrub whenever a snapshot
 is taken.

Actually yes, I take snapshots once an hour of various things.  I'll try 
disabling them for the time being and see how far along it gets.

Thanks!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] x4500 resilvering spare taking forever?

2009-06-21 Thread Joe Kearney
 Also, b57 is about 2 years old and misses the
 improvements in performance,
 especially in scrub performance.

Yep, I know.  I'll upgrade them at some point down the road, but they've been 
serving our needs nicely so far.

Thanks!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Orvar Korvar
I use the AOC-SAT2-MV8 in a ordinary PCI slot. The PCI slot maxes at 150MB/sec 
or so. That is the fastest you will get. That card works very good with 
Solaris/OpenSolaris. Detects automatically, etc. Ive heard though that it does 
not work with hot swapping discs - avoid this.

However, In a PCI-X it will max at 1GB/sec. I have been thinkin about buying a 
server mobo (they have PCI-X) to get 1GB/sec. Or should I buy a PCIe card 
instead? I dont know.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] x4500 resilvering spare taking forever?

2009-06-21 Thread Richard Elling

Joe Kearney wrote:

Also, b57 is about 2 years old and misses the
improvements in performance,
especially in scrub performance.



Yep, I know.  I'll upgrade them at some point down the road, but they've been 
serving our needs nicely so far.

  


Yep, it also suffers from the bug that restarts resilvers when you take a
snapshot. This was fixed in b94.
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6343667
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread roland
just a side-question:

I folthis thread with much interest.

what are these * for ?

why is followed turned into fol* on this board?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread dick hoogendijk
On Sun, 21 Jun 2009 14:07:49 PDT
roland no-re...@opensolaris.org wrote:

 just a side-question:
 
 I folthis thread with much interest.
 
 what are these * for ?
 
 why is followed turned into fol* on this board?

The text of my original message was:

On Sun, 21 Jun 2009 06:35:50 PDT
Simon Breden no-re...@opensolaris.org wrote:

 If anyone can throw some light on these topics, I would be pleased to
 hear from you. Thanks a lot.

I follow this thread with much interest.
Curious to see what'll come out of it.

Does the change occur again?

-- 
Dick Hoogendijk -- PGP/GnuPG key: 01D2433D
+ http://nagual.nl/ | nevada / OpenSolaris 2009.06 release
+ All that's really worth doing is what we do for others (Lewis Carrol)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Simon Breden
Hey Kebabber, long time no hear! :)

It's great to hear that you've had good experiences with the card. It's a great 
pity to have throughput drop from a potential 1GB/s to 150MB/s, but as most of 
my use of the NAS is across the network, and not local intra-NAS transfers, 
this should not be a problem. Of course, with a single GbE connection speeds 
are normally limited to around 50MB/s or so anyway...

Tell me, have you had any drives fail and had to figure out how to identify the 
failed drive and replace it  resilver using the AOC-SAT2-MV8, or have you 
tried any experiments to test resilvering ? I'm just curious as to how easy it 
is to do this with this controller card.

Like yourself, I was toying with the idea of upgrading and buying a shiny new 
mobo with dual 64-bit PCI-X slots and socket LGA1366 for Xeon 5500 series 
(Nehalem) processors -- the SuperMicro X8SAX here: 
http://www.supermicro.com/products/motherboard/Xeon3000/X58/X8SAX.cfm

Then I added up the price of all the components and decided to try to make do 
with the existing kit and just do an upgrade.

So I narrowed down possible SATA controllers to the above choices and I'm 
interested in people's experiences of using these controllers to help me decide.

Seems like the cheapest and simplest choice will be the AOC-SAT2-MV8, and I 
just take a hit on the reduced speed -- but that won't be a big problem.

However, as I have 2 x PCIe x16 slots available, if the AOC-USAS-L8i is 
reliable and doesn't have issues now with identifying drive ids, and supports 
JBOD mode, then it looks like the better choice. It is uses the more modern PCI 
Express (PCIe) interface, rather than the ageing PCI-X interface, fine as I'm 
sure it is.

Simon
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Ian Collins

 On Mon 22/06/09 02:07 , roland no-re...@opensolaris.org sent:
 just a side-question:
 
 I folthis thread with much
 interest.
 what are these * for ?
 
 why is followed turned into fol* on this
 board?

It isn't a board, it's a mail list.  All the forum does is bugger up the 
formatiing and threading of emails!

-- 
Ian
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Things I Like About ZFS

2009-06-21 Thread Simon Breden
OK, my turn:

- combining file system + volume manager + RAID + pool + scrub + resilvering + 
snapshots + rollback + end-to-end integrity + 256-but block checksums + 
on-the-fly healing of blocks with checksum errors on read
- one liners that are mostly remembered, and simple to guess if forgotten
- with RAID-Z2, superb protection + hot spares
- easy sharing via CIFS and NFS
- iSCSI as a block device target
- open source software RAID  no proprietary hardware RAID card required
- free
- did I forget something? :)
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] x4500 resilvering spare taking forever?

2009-06-21 Thread Andrew Gabriel

Joe Kearney wrote:

UPDATE: It's now back down to 0.9% complete.  Does anyone have a clue as to 
whats happening here or where I can look for problems?


There was also a bug which restarted resilvers each time you issue a
zpool status command as a privilaged user. Make sure to check the
progress by issuing zpool status commands as a non-privilaged user.

--
Andrew
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs find list of properties per dataset

2009-06-21 Thread Harry Putnam
How can I list out all the properties available for a filesystem?

Banging away in man zfs seems to be going nowhere pretty fast.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs find list of properties per dataset

2009-06-21 Thread James C. McPherson
On Sun, 21 Jun 2009 18:28:56 -0500
Harry Putnam rea...@newsguy.com wrote:

 How can I list out all the properties available for a filesystem?
 
 Banging away in man zfs seems to be going nowhere pretty fast.

Is zfs get all datasetname not helping?


$ zfs get all rpool
NAME   PROPERTYVALUE   SOURCE
rpool  typefilesystem  -
rpool  creationWed May  6 21:51 2009   -
rpool  used85.9G   -
rpool  available   115G-
rpool  referenced  81.5K   -
rpool  compressratio   1.00x   -
rpool  mounted yes -
rpool  quota   nonedefault
rpool  reservation nonedefault
rpool  recordsize  128Kdefault
rpool  mountpoint  /rpool  default
rpool  sharenfsoff default
rpool  checksumon  default
rpool  compression off default
rpool  atime   on  default
rpool  devices on  default
rpool  execon  default
rpool  setuid  on  default
rpool  readonlyoff default
rpool  zoned   off default
rpool  snapdir hidden  default
rpool  aclmode groupmask   default
rpool  aclinherit  restricted  default
rpool  canmounton  default
rpool  shareiscsi  off default
rpool  xattr   on  default
rpool  copies  1   default
rpool  version 3   -
rpool  utf8onlyoff -
rpool  normalization   none-
rpool  casesensitivity sensitive   -
rpool  vscan   off default
rpool  nbmand  off default
rpool  sharesmboff default
rpool  refquotanonedefault
rpool  refreservation  nonedefault
rpool  primarycacheall default
rpool  secondarycache  all default
rpool  usedbysnapshots 0   -
rpool  usedbydataset   81.5K   -
rpool  usedbychildren  85.9G   -
rpool  usedbyrefreservation0   -
rpool  org.opensolaris.caiman:install  ready   local









James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
Kernel Conference Australia - http://au.sun.com/sunnews/events/2009/kernel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs find list of properties per dataset

2009-06-21 Thread Harry Putnam
James C. McPherson james.mcpher...@sun.com writes:

[...]

 Is zfs get all datasetname not helping?

[...]

Gack... quite the reverse

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs find list of properties per dataset

2009-06-21 Thread James C. McPherson
On Sun, 21 Jun 2009 18:47:30 -0500
Harry Putnam rea...@newsguy.com wrote:

 James C. McPherson james.mcpher...@sun.com writes:
 
 [...]
 
  Is zfs get all datasetname not helping?
 
 [...]
 
 Gack... quite the reverse

your original question was somewhat vague - could you clarify
what it is you need to find out?


James
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
Kernel Conference Australia - http://au.sun.com/sunnews/events/2009/kernel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs find list of properties per dataset

2009-06-21 Thread Harry Putnam
James C. McPherson james.mcpher...@sun.com writes:


 your original question was somewhat vague - could you clarify
 what it is you need to find out?


I wanted to see a list of all the properties available to zfs on a
filesystem.

Your answer was very helpful... thanks My quick reply may not have
conveyed that I was happy with your answer and that it helps a lot.

The actual exact thing I was looking into was where snapdir was hidden
and where visible.  But I didn't know the name of the option
(snapdir).

Once I saw the full list... of course it was obvious.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] two pools on boot disk?

2009-06-21 Thread Fajar A. Nugraha
On Sat, Jun 20, 2009 at 7:02 PM, Michael
Sullivanmichael.p.sulli...@mac.com wrote:
 One really interesting bit is how easily it is to make the disk in a pool
 bigger by doing a zpool replace on the device.  It couldn't have been any
 easier with ZFS.

It's interesting how you achieved that, although it'd be much easier
if the installer supports that from the GUI instead of having to use
zpool replace as a workaround. I believe using export-import as
described in 
http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#ZFS_Root_Pool_Recovery
should also work.

-- 
Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Carson Gaspar
I'll chime in as a happy owner of the LSI SAS 3081E-R PCI-E board. It 
works just fine. You need to get lsiutil from the LSI web site to 
fully access all the functionality, and they cleverly hide the download 
link only under their FC HBAs on their support site, even though it 
works for everything.



As for identifying disks, you can just use lsiutil:

root:gandalf 0 # lsiutil -p 1 42

LSI Logic MPT Configuration Utility, Version 1.62, January 14, 2009

1 MPT Port found

 Port Name Chip Vendor/Type/RevMPT Rev  Firmware Rev  IOC
 1.  mpt0  LSI Logic SAS1068E B3 105  011a 0

mpt0 is /dev/cfg/c6

 B___T___L  Type   Operating System Device Name
 0   0   0  Disk   /dev/rdsk/c6t0d0s2
 0   1   0  Disk   /dev/rdsk/c6t1d0s2
 0   2   0  Disk   /dev/rdsk/c6t2d0s2
 0   3   0  Disk   /dev/rdsk/c6t3d0s2
 0   4   0  Disk   /dev/rdsk/c6t4d0s2
 0   5   0  Disk   /dev/rdsk/c6t5d0s2
 0   6   0  Disk   /dev/rdsk/c6t6d0s2
 0   7   0  Disk   /dev/rdsk/c6t7d0s2



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs IO scheduler

2009-06-21 Thread tester
Hello,

Trying to understand the ZFS IO scheduler, because of the async nature it is 
not very apparent, can someone give a short explanation for each of these stack 
traces and for  their frequency

this is the command

dd if=/dev/zero of=/test/test1/trash count=1 bs=1024k;sync

no other IO is happening to the test pool. OS is on a zfs pool (rpool)


I don't see any zio_vdev_io_start in any of the function stacks, any idea why?

dtrace -n 'io:::start { @a[stack()] = count(); }'

dtrace: description 'io:::start ' matched 6 probes


 genunix`bdev_strategy+0x44
  zfs`vdev_disk_io_start+0x2a8
  zfs`zio_execute+0x74
  genunix`taskq_thread+0x1a4
  unix`thread_start+0x4
   20

  genunix`bdev_strategy+0x44
  zfs`vdev_disk_io_start+0x2a8
  zfs`zio_execute+0x74
  zfs`vdev_queue_io_done+0x84
  zfs`vdev_disk_io_done+0x4
  zfs`zio_execute+0x74
  genunix`taskq_thread+0x1a4
  unix`thread_start+0x4
   31

  genunix`bdev_strategy+0x44
  zfs`vdev_disk_io_start+0x2a8
  zfs`zio_execute+0x74
  zfs`vdev_mirror_io_start+0x1b4
  zfs`zio_execute+0x74
  zfs`vdev_mirror_io_start+0x1b4
  zfs`zio_execute+0x74
  genunix`taskq_thread+0x1a4
  unix`thread_start+0x4
   34

  genunix`bdev_strategy+0x44
  zfs`vdev_disk_io_start+0x2a8
  zfs`zio_execute+0x74
  zfs`vdev_mirror_io_start+0x1b4
  zfs`zio_execute+0x74
  genunix`taskq_thread+0x1a4
  unix`thread_start+0x4
   45

  genunix`bdev_strategy+0x44
  zfs`vdev_disk_io_start+0x2a8
  zfs`zio_execute+0x74
  zfs`vdev_queue_io_done+0x9c
  zfs`vdev_disk_io_done+0x4
  zfs`zio_execute+0x74
  genunix`taskq_thread+0x1a4
  unix`thread_start+0x4
   53
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread James C. McPherson
On Sun, 21 Jun 2009 19:01:31 -0700
Carson Gaspar car...@taltos.org wrote:

 I'll chime in as a happy owner of the LSI SAS 3081E-R PCI-E board. It 
 works just fine. You need to get lsiutil from the LSI web site to 
 fully access all the functionality, and they cleverly hide the download 
 link only under their FC HBAs on their support site, even though it 
 works for everything.

As a member of the team which works on mpt(7d), I'm disappointed that\
you believe you need to use lsiutil to fully access all the functionality
of the board.

What gaps have you found in mpt(7d) and the standard OpenSolaris
tools that lsiutil fixes for you?

What is the full functionality that you believe is missing?



 As for identifying disks, you can just use lsiutil:

... or use cfgadm(1m) which has had this ability for many years.

 
 root:gandalf 0 # lsiutil -p 1 42
 
 LSI Logic MPT Configuration Utility, Version 1.62, January 14, 2009
 
 1 MPT Port found
 
   Port Name Chip Vendor/Type/RevMPT Rev  Firmware Rev  IOC
   1.  mpt0  LSI Logic SAS1068E B3 105  011a 0
 
 mpt0 is /dev/cfg/c6
 
   B___T___L  Type   Operating System Device Name
   0   0   0  Disk   /dev/rdsk/c6t0d0s2
   0   1   0  Disk   /dev/rdsk/c6t1d0s2
   0   2   0  Disk   /dev/rdsk/c6t2d0s2
   0   3   0  Disk   /dev/rdsk/c6t3d0s2
   0   4   0  Disk   /dev/rdsk/c6t4d0s2
   0   5   0  Disk   /dev/rdsk/c6t5d0s2
   0   6   0  Disk   /dev/rdsk/c6t6d0s2
   0   7   0  Disk   /dev/rdsk/c6t7d0s2

You can get that information from use of cfgadm(1m).



James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
Kernel Conference Australia - http://au.sun.com/sunnews/events/2009/kernel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Andre van Eyssen

On Sun, 21 Jun 2009, Carson Gaspar wrote:

I'll chime in as a happy owner of the LSI SAS 3081E-R PCI-E board. It works 
just fine. You need to get lsiutil from the LSI web site to fully access 
all the functionality, and they cleverly hide the download link only under 
their FC HBAs on their support site, even though it works for everything.


I'll add another vote for the LSI products. I have a four port PCI-X card 
in my V880, and the performance is good and the product is well behaved. 
The only caveats:


1. Make sure you upgrade the firmware ASAP
2. You may need to use lsiutil to fiddle the target mappings

Andre.


--
Andre van Eyssen.
mail: an...@purplecow.org  jabber: an...@interact.purplecow.org
purplecow.org: UNIX for the masses http://www2.purplecow.org
purplecow.org: PCOWpix http://pix.purplecow.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Jorgen Lundman


I only have a 32bit PCI bus in the Intel Atom 330 board, so I have no 
choice than to be slower, but I can confirm that the Supermicro 
dac-sata-mv8 (SATA-1) card works just fine, and does display in cfgadm. 
(Hot-swapping is possible).


I have been told aoc-sat2-mv8 does as well (SATA-II) but I have not 
personally tried it.


Lund

--
Jorgen Lundman   | lund...@lundman.net
Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
Shibuya-ku, Tokyo| +81 (0)90-5578-8500  (cell)
Japan| +81 (0)3 -3375-1767  (home)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] two pools on boot disk?

2009-06-21 Thread Michael Sullivan

Fajar,

Yes, you could probably do send/receive from one pool to another, but  
that would be somewhat more time consuming and you'd have to make sure  
everything was right in your GRUB menu.lst as well as boot blocks, not  
to mention the potential for namespace collisions when dealing with a  
root pool.  But this is missing my point.


The thing I found more interesting was that a pool could be increased  
in space by doing a zpool replace with a larger disk. This means if  
say, you have a pool of 100GB disks and you want to increase the size,  
you can replace them with bigger disks effectively growing the pool.   
Not sure how this works out with configurations other than in RAID 0  
and RAID 1, but I thought it was a pretty nice feature knowing I can  
put bigger disks in really easily.


I also agree the installer should have an expert mode for  
configuring disks.  The all-or-nothing approach is easy for people  
who have never been exposed to Solaris or OpenSolaris, but leaves  
people out in the cold if you wish to have different configuration for  
your disks.


The Automated Installer, is supposed to give this sort of flexibility,  
but I haven't tried it out yet.


Regards,

Mike

---
Michael Sullivan
michael.p.sulli...@me.com
http://www.kamiogi.net/
Japan Mobile: +81-80-3202-2599
US Phone: +1-561-283-2034

On 22 Jun 2009, at 11:00 , Fajar A. Nugraha wrote:


On Sat, Jun 20, 2009 at 7:02 PM, Michael
Sullivanmichael.p.sulli...@mac.com wrote:
One really interesting bit is how easily it is to make the disk in  
a pool
bigger by doing a zpool replace on the device.  It couldn't have  
been any

easier with ZFS.


It's interesting how you achieved that, although it'd be much easier
if the installer supports that from the GUI instead of having to use
zpool replace as a workaround. I believe using export-import as
described in 
http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#ZFS_Root_Pool_Recovery
should also work.

--
Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Erik Trimble

Jorgen Lundman wrote:


I only have a 32bit PCI bus in the Intel Atom 330 board, so I have no 
choice than to be slower, but I can confirm that the Supermicro 
dac-sata-mv8 (SATA-1) card works just fine, and does display in 
cfgadm. (Hot-swapping is possible).


I have been told aoc-sat2-mv8 does as well (SATA-II) but I have not 
personally tried it.


Lund



I have an AOC-SAT2-MV8 in an older Opteron-based system.  It's a 
2-socket, Opteron 252 system with 8GB of RAM, and PCI-X slots.  It's one 
of the newer AOC cards with the latest Marvell chipset, and it works 
like a champ - very well, smooth, and I don't see any problems.  Simple, 
out-of-the-box installation and works with no tinkering at all (with OS 
2008.11 and 2009.05).


That said, there's a couple of things you want to be aware of about the AOC:

(1) it uses normal sata cables. This is really nice in terms of 
availability (you can get any length you want easily at any computer 
store), but it's a bit messy compared to the nice multi-lane ones.


(2) It's a PCI-X card, and will run at 133Mhz.  I have a second gigabit 
ethernet card in my motherboard, which limits the two PCI-X cards to 
100Mhz.The down side is that with 8 drives and 2 gigabit ethernet 
interfaces, it's not hard to flood the PCI-X bus (which can pump 100Mhz 
x 64bit = 6400 Mbps max, but about 50% of that under real usage)


(3) as a PCI-X card, it's a two-third's length, low-profile card. It 
will fit in any PCI-X slot you have. However, if you are trying to put 
it in a 32-bit PCI slot, be aware that it extends about 2 inches (50mm) 
beyond the back of the PCI slot. Make sure you have the proper 
clearances for such a card.  Also, it's a 3.3v card (won't work in 5v 
slots).  None of this should be a problem in any modern motherboard/case 
setup, only in really old stuff.


(4) All the SATA connectors are on the end of the card, which means 
you'll need _at least_ another 1 inch (25mm) clearance to plug the 
cables in.



As much as I like the card, these days, I'd chose the LSI PCI-E model.  
The PCI-E bus is just superior to PCI-X - you get much less bus 
contention which means it's easier to get full throughput from each card



One more thing:   I've found that the newest MLC-based SSDs with the 
newer barefoot SATA controller and 64MB or more of cache are more than 
suitable for use as Read cache, and they actually do OK as write-cache, 
too.  Particularly for small business server machine (those that have 
have 8-12 data drives, total).


And, these days, there's nice little funky dual-2.5 drives in a floppy 
form-factor things.


http://www.addonics.com/products/mobile_rack/doubledrive.asp

Example new SSD for Readzilla/Logzilla :

http://www.ocztechnology.com/products/flash_drives/ocz_summit_series_sata_ii_2_5-ssd


--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] cutting up a SSD for read/log use...

2009-06-21 Thread Erik Trimble
I just looked at pricing for the higher-end MLC devices, and it looks 
like I'm better off getting a single drive of 2X capacity than two with 
X capacity.


Leaving aside the issue that by using 2 drives I get  2 x 3.0Gbps SATA 
performance instead of 1 x 3.0Gbps,  are there problems with using two 
slices instead of whole-drives?  That is, one slice for Read and the 
other for ZIL?


My main concern is exactly how the on-drive cache would be used in a 
two-slices configuration. In order to get decent performance, I really 
need the on-drive cache to be used properly.



--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Speeding up resilver on x4500

2009-06-21 Thread Stuart Anderson

It is currently taking ~1 week to resilver an x4500 running S10U6,
recently patched with~170M small files on ~170 datasets after a
disk failure/replacement, i.e.,

 scrub: resilver in progress for 53h47m, 30.72% done, 121h19m to go

Is there anything that can be tuned to improve this performance, e.g.,
adding a faster cache device for reading and/or writing?



I am also curious if anyone has a prediction on when the
snapshot-restarting-resilvering bug will be patched in Solaris 10?

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6343667


Thanks.

--
Stuart Anderson  ander...@ligo.caltech.edu
http://www.ligo.caltech.edu/~anderson



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cutting up a SSD for read/log use...

2009-06-21 Thread James Lever

Hi Erik,

On 22/06/2009, at 1:15 PM, Erik Trimble wrote:

I just looked at pricing for the higher-end MLC devices, and it  
looks like I'm better off getting a single drive of 2X capacity than  
two with X capacity.


Leaving aside the issue that by using 2 drives I get  2 x 3.0Gbps  
SATA performance instead of 1 x 3.0Gbps,  are there problems with  
using two slices instead of whole-drives?  That is, one slice for  
Read and the other for ZIL?


The benefit you will get using 2 drives instead of 1 will be doubling  
your IOPS which will improve your overall performance, especially when  
using those drives as ZILs.


Are you planning on using these drives as primary data storage and ZIL  
for the same volumes or as primary storage for (say) your rpool and  
ZIL for a data pool on spinning metal?


cheers,
James

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs find list of properties per dataset

2009-06-21 Thread Richard Elling

Harry Putnam wrote:

James C. McPherson james.mcpher...@sun.com writes:

  

your original question was somewhat vague - could you clarify
what it is you need to find out?




I wanted to see a list of all the properties available to zfs on a
filesystem.
  


NB, the user and group quotas are stored as properties, but get all does
not return them.  So it is not a true statement that get all returns 
all of

the properties.
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Speeding up resilver on x4500

2009-06-21 Thread Richard Elling

Stuart Anderson wrote:

It is currently taking ~1 week to resilver an x4500 running S10U6,
recently patched with~170M small files on ~170 datasets after a
disk failure/replacement, i.e.,


wow, that is impressive.  There is zero chance of doing that with a
manageable number of UFS file systems.



 scrub: resilver in progress for 53h47m, 30.72% done, 121h19m to go

Is there anything that can be tuned to improve this performance, e.g.,
adding a faster cache device for reading and/or writing?


Resilver tends to be bound by one of two limits:

1. sequential write performance of the resilvering device

2. random I/O performance of the non-resilvering devices

A while back, I was doing some characterization of this, but the
funding disappeared :-(  So, it is unclear whether or how caching
might help.
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cutting up a SSD for read/log use...

2009-06-21 Thread Richard Elling

Erik Trimble wrote:
I just looked at pricing for the higher-end MLC devices, and it looks 
like I'm better off getting a single drive of 2X capacity than two 
with X capacity.


Leaving aside the issue that by using 2 drives I get  2 x 3.0Gbps SATA 
performance instead of 1 x 3.0Gbps,  are there problems with using two 
slices instead of whole-drives?  That is, one slice for Read and the 
other for ZIL?


My main concern is exactly how the on-drive cache would be used in a 
two-slices configuration. In order to get decent performance, I really 
need the on-drive cache to be used properly.


Is the on-disk cache volatile?  For most SSDs I'm familiar with, the on-disk
cache is non-volatile, so all of the rules pertaining to whole disks with
volatile write buffers are nullified.
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Speeding up resilver on x4500

2009-06-21 Thread Stuart Anderson


On Jun 21, 2009, at 8:57 PM, Richard Elling wrote:


Stuart Anderson wrote:

It is currently taking ~1 week to resilver an x4500 running S10U6,
recently patched with~170M small files on ~170 datasets after a
disk failure/replacement, i.e.,


wow, that is impressive.  There is zero chance of doing that with a
manageable number of UFS file systems.


However, it is a bit disconcerting to have to run with reduced data
protection for an entire week. While I am certainly not going back to
UFS, it seems like it should be at least theoretically possible to do  
this

several orders of magnitude faster, e.g., what if every block on the
replacement disk had its RAIDZ2 data recomputed from the degraded
array regardless of whether the pool was using it or not. In that case
I would expect it to be able to sequentially reconstruct in the same few
hours it would take a HW RAID controller to do the same RAID6 job.

Perhaps there needs to be an option to re-order the loops for
resilvering on pools with lots of small files to resilver in device
order rather than filesystem order?





scrub: resilver in progress for 53h47m, 30.72% done, 121h19m to go

Is there anything that can be tuned to improve this performance,  
e.g.,

adding a faster cache device for reading and/or writing?


Resilver tends to be bound by one of two limits:

1. sequential write performance of the resilvering device

2. random I/O performance of the non-resilvering devices



A quick look at iostat leads me to conjecture that the vdev rebuilding  
is

taking a very low priority compared to ongoing application I/O (NFSD
in this case). Are there any ZFS knobs that control the relative  
priority of

resilvering to other disk I/O tasks?

Thanks.

--
Stuart Anderson  ander...@ligo.caltech.edu
http://www.ligo.caltech.edu/~anderson



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Backup schemes invoving rsync Linux_rmt = Osol

2009-06-21 Thread Harry Putnam
There is a lot more about snapshots and backup schemes that I don't
know than there is that I do know...  My needs are pretty simple and
small compared to some of the heavy users here, but I'd like to run
just the rough idea by the group.

One backup I need to make is from to linux online servers to a zpool.

Whats involved is two websites I use sometimes quite a bit but more
often not so much and sometimes not at all for quite a stretch.

I thought I'd rsync  the base directory to a zpool using the `--inplace'
flag to rsync that has been discussed here a time or two.

It appears the snapshot setup was including the specific zpool I plan
to use but at a frequency level way above what seems I'd need.  But them
I'm not really sure how to go at this.

I see mnthly weekly hourly and frequent which appear to be every 15
min. 

I thought it might be better to turn that off completely.  Then use a
cron job to run the rsync once per week.   The same little script
would run a snapshot following the completion of the rsync run.

I'd probably use something like this for rsync:

Cron calls  my script and runs:
  rsync -avvz --stats --delete --exclude-from='exclude_file'
--delete-excluded --inplace 
 m...@remote:/home/me/public_html/ /www/remote_www/

It would be run with the `authorized_keys' mechanism
  
Then runs a snapshot of zfs z3/www

I'm thinking that snapshot would catch every thing since the last run.
   
Is that a reasonable plan?  

Sorry to ruminate on here... but would like any ideas you may have.

I'd probably do something very similar to the www plan with the home
lan machines... Two linux boxes are no problem, but I haven't really
figured out the two windows XP machines yet.  It seems there's always
been bugs in Cifs implementation preventing reliable
connections.. Currently I can only see top level of any windows shares
from zfs host.

A while back I had `retrospect' (a windows version) connecting to zfs
machine, but that fell through so often... it was not working.

Windows based backup tools (at least some of them) expect the admin to
be able to navigate to appropriate shares or at least use UNK
addressing.   But I have yet to get the zfs machine to show up in
windows network places.  Even when I can force things by typing an UNK
address on windows explorer there is a very good chance after
either machine is rebooted ... the connection would not work thereafter.

Seems like initiating the connections from zfs host is a better way to
go.  And I'm currently looking in to using Bacula.

Using rsync against the windows machines seems to mostly be a
non-starter due to something in ntfs or the OS that makes files appear
to be all new every so often causing rsync to be fooled into pulling
the same files across again.

That, by itself is bad enough but not sure what impact it might have
on zfs snapshots. .. maybe none... long as --inplace is being used.

I've been pretty lazy about it and have continued to backup the
windows machines with things like ghost... making complete drive images.

But since it seems like nothing but headaches trying to connect to the
zfs host reliably...(From windows), I run the backup onto an internal
drive and then have to move the monster file again to get it on a zfs
filesystem.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] bug access

2009-06-21 Thread Jens Elkner
Hi,

 This CR has been marked as incomplete by User 1-UM-1502
 for the reason Need More Info.  Please update the CR
 providing the information requested in the Evaluation and/or Comments
 field.

hmmm - wondering, how to find out, what 'more info' means and how to
provide this info. There is no URL in the bug-response and it even
seems to be impossible, to obtain the current state via 
http://bugs.opensolaris.org/view_bug.do?bug_id= 
(Bug Database Search is even more bogus - in general it doesn't find
any bugs by ID).

Should I continue to ignore these responses and mark those bugs internally
as 'gets probably never fixed'?

Regards.
jel.
-- 
Otto-von-Guericke University http://www.cs.uni-magdeburg.de/
Department of Computer Science   Geb. 29 R 027, Universitaetsplatz 2
39106 Magdeburg, Germany Tel: +49 391 67 12768
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cutting up a SSD for read/log use...

2009-06-21 Thread Erik Trimble

Richard Elling wrote:

Erik Trimble wrote:
I just looked at pricing for the higher-end MLC devices, and it looks 
like I'm better off getting a single drive of 2X capacity than two 
with X capacity.


Leaving aside the issue that by using 2 drives I get  2 x 3.0Gbps 
SATA performance instead of 1 x 3.0Gbps,  are there problems with 
using two slices instead of whole-drives?  That is, one slice for 
Read and the other for ZIL?


My main concern is exactly how the on-drive cache would be used in a 
two-slices configuration. In order to get decent performance, I 
really need the on-drive cache to be used properly.


Is the on-disk cache volatile?  For most SSDs I'm familiar with, the 
on-disk

cache is non-volatile, so all of the rules pertaining to whole disks with
volatile write buffers are nullified.
-- richard


I'm pretty sure it is volatile.  It's a single DRAM chip.

http://benchmarkreviews.com/index.php?option=com_contenttask=viewid=333Itemid=60limit=1limitstart=3

(this review is specific to the OCZ Summit, but as the parts are pretty 
much Samsung-standard, it's identical to several other brand's versions)



Are you sure most other SSDs have nvram as cache?  I'm looking around, 
and they _seem_ to be using standard DRAM, just like Hard drives...  Or 
maybe the SLC-based ones use nvram, and the MLC-based ones dram...



--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cutting up a SSD for read/log use...

2009-06-21 Thread Erik Trimble

James Lever wrote:

Hi Erik,

On 22/06/2009, at 1:15 PM, Erik Trimble wrote:

I just looked at pricing for the higher-end MLC devices, and it looks 
like I'm better off getting a single drive of 2X capacity than two 
with X capacity.


Leaving aside the issue that by using 2 drives I get  2 x 3.0Gbps 
SATA performance instead of 1 x 3.0Gbps,  are there problems with 
using two slices instead of whole-drives?  That is, one slice for 
Read and the other for ZIL?


The benefit you will get using 2 drives instead of 1 will be doubling 
your IOPS which will improve your overall performance, especially when 
using those drives as ZILs.


Are you planning on using these drives as primary data storage and ZIL 
for the same volumes or as primary storage for (say) your rpool and 
ZIL for a data pool on spinning metal?


cheers,
James


ZIL and Read cache for a data pool of HDs.

i.e.

zpool create tank mirror c1t0d0 c1t1d0 [...] log c2t0d0 cache  c3t0d0

or

zpool create tank mirror c1t0d0 c1t1d0 [...] log c2t0d0s0 cache c2t0d0s1




--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best controller card for 8 SATA drives ?

2009-06-21 Thread Eric D. Mudama

On Mon, Jun 22 at 12:05, Andre van Eyssen wrote:
I'll add another vote for the LSI products. I have a four port PCI-X card 
in my V880, and the performance is good and the product is well behaved.  
The only caveats:


1. Make sure you upgrade the firmware ASAP
2. You may need to use lsiutil to fiddle the target mappings


We bought a Dell T610 as a fileserver, and it comes with an LSI 1068E
based board (PERC6/i SAS).  Worked out of the box, no special drivers
or anything to install, everything autodetected just fine.

Hotplug works great too, I've yanked drives (Came with WD RE3 SASA
devices) while the box was under load without issues, took ~5 seconds
to timeout the device and give me full interactivity at the console.
They then show right back up when hot plugged back in, and I can
resilver without problems.

--eric

--
Eric D. Mudama
edmud...@mail.bounceswoosh.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] PicoLCD Was: Best controller card for 8 SATA drives ?

2009-06-21 Thread Jorgen Lundman


I hesitate to post this question here, since the relation to ZFS is 
tenuous at best (zfs to sata controller to LCD panel).


But maybe someone has already been down this path before me. Looking at 
building a RAID, with osol and zfs, I naturally want a front-panel. I 
was looking at something like;


http://www.mini-box.com/picoLCD-256x64-Sideshow-CDROM-Bay

Since it appears to come with OpenSource drivers. Based on lcd4linux, 
which I can compile with marginal massaging. Has anyone run this 
successfully with osol?


It appears to handle mrtg directly, so I should be able to graph a whole 
load of ZFS data. Has someone already been down this road too?




--
Jorgen Lundman   | lund...@lundman.net
Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
Shibuya-ku, Tokyo| +81 (0)90-5578-8500  (cell)
Japan| +81 (0)3 -3375-1767  (home)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Why Oracle process open(2)/ioctl(2) /dev/dtrace/helper?

2009-06-21 Thread Sang-Thong Chang


Hi all,


Under what circumstance would an exiting process open(2) and
ioctl(2) on /dev/dtrace/helper?


I have an issue here where CU was complaining that their
Oracle processes were taking a long time in ioctl(2) on
/dev/dtrace/helper during shutdown. Removing group/world
readable bit on /dev/dtrace/helper avoided the long shutdown
time.

So, the puzzling question is why /dev/dtrace/helper was being
ioctl(2)ed on?

I doubt this is a binary with USDT.  truss(1M) shown the following:


Base time stamp:  1244728105.2977  [ Thu Jun 11 22:48:25 KST 2009 ]
16003/1:psargs: oraclePOM1 (LOCAL=NO)
16003/1:read(31, 0x1069F3796, 2064) (sleeping...)
...
16003/1:1468.91560.0005 open(/dev/dtrace/helper, 
O_RDWR)  = 8
16003/1:1477.37738.4617 ioctl(8, 
(('d'24)|('t'16)|('h'8)|2), 0x) = 0
16003/1:1477.37810.0008 Received signal #14, SIGALRM 
[caught]
16003/1:1477.37830.0002 lwp_sigmask(SIG_SETMASK, 
0x9FBEF057, 0xFFF7) = 0xFFBFFEFF [0x]


   1277  */
   1278 #define DTRACEHIOC  (('d'  24) | ('t'  16) | 
('h'  8))
   1279 #define DTRACEHIOC_ADD  (DTRACEHIOC | 1)/* add 
helper */
   1280 #define DTRACEHIOC_REMOVE   (DTRACEHIOC | 2)/* 
remove helper */




ustack() from dtrace(1M) shown this was from within the
exit hanlder in exit(2).

 open(/dev/dtrace/helper)

  libc.so.1`open
  libCrun.so.1`0x7a50aed8
  libCrun.so.1`0x7a50b0f4
  ld.so.1`call_fini+0xd0
  ld.so.1`atexit_fini+0x80
  libc.so.1`_exithandle+0x48
  libc.so.1`exit+0x4
  oracle`_start+0x184

It would be great if anyone can help extend my understanding
of dtrace. Thanks.

Regards
Sang Thong


-

1788:   oraclePOM1 (LOCAL=NO)
data model = _LP64  flags = ORPHAN|MSACCT|MSFORK
 /1:flags = ASLEEP  read(0x11,0x106346b16,0x810)

1788:   oracleEOM1 (LOCAL=NO)
/lib/sparcv9/libumem.so.1
/oraom/app/DBMS/lib/libskgxp10.so
/oraom/app/DBMS/lib/libhasgen10.so
/opt/ORCLcluster/lib/libskgxn2.so
/oraom/app/DBMS/lib/libocr10.so
/oraom/app/DBMS/lib/libocrb10.so
/oraom/app/DBMS/lib/libocrutl10.so
/oraom/app/DBMS/lib/libjox10.so
/oraom/app/DBMS/lib/libclsra10.so
/oraom/app/DBMS/lib/libdbcfg10.so
/oraom/app/DBMS/lib/libnnz10.so
/lib/sparcv9/libkstat.so.1
/lib/sparcv9/libnsl.so.1
/lib/sparcv9/libsocket.so.1
/lib/sparcv9/libgen.so.1
/lib/sparcv9/libdl.so.1
/usr/lib/sparcv9/libsched.so.1
/lib/sparcv9/libc.so.1
/lib/sparcv9/libaio.so.1
/lib/sparcv9/librt.so.1
/lib/sparcv9/libm.so.2
/opt/SUNWcluster/lib/sparcv9/libudlm.so
/usr/cluster/lib/sparcv9/libhaops.so.1
/lib/sparcv9/libmd.so.1
/usr/cluster/lib/sparcv9/libscha.so.1
/usr/cluster/lib/sparcv9/libsecurity.so.1
/usr/cluster/lib/sparcv9/libclos.so.1
/lib/sparcv9/libdoor.so.1
/usr/lib/sparcv9/libzonecfg.so.1
/lib/sparcv9/libsecdb.so.1
/usr/lib/sparcv9/libCstd.so.1
/usr/lib/sparcv9/libCrun.so.1
/lib/sparcv9/libuuid.so.1
/lib/sparcv9/libnvpair.so.1
/lib/sparcv9/libsysevent.so.1
/lib/sparcv9/libsec.so.1
/usr/lib/sparcv9/libbrand.so.1
/usr/lib/sparcv9/libpool.so.1
/lib/sparcv9/libscf.so.1
/lib/sparcv9/libproc.so.1
/lib/sparcv9/libuutil.so.1
/usr/lib/sparcv9/libxml2.so.2
/lib/sparcv9/libcmd.so.1
/lib/sparcv9/libavl.so.1
/usr/lib/sparcv9/libexacct.so.1
/lib/sparcv9/librtld_db.so.1
/lib/sparcv9/libelf.so.1
/lib/sparcv9/libctf.so.1
/lib/sparcv9/libpthread.so.1
/usr/lib/sparcv9/libz.so.1
/platform/sun4u-us3/lib/sparcv9/libc_psr.so.1

  Mon, 22 Jun 2009 09:06:09 KST
 open(/dev/dtrace/helper)

  libc.so.1`open
  libCrun.so.1`0x7a50aed8
  libCrun.so.1`0x7a50b0f4
  ld.so.1`call_fini+0xd0
  ld.so.1`atexit_fini+0x80
  libc.so.1`_exithandle+0x48
  libc.so.1`exit+0x4
  oracle`_start+0x184

***
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Sorry, wrong alias Re: Why Oracle process open(2)/ioctl(2) /dev/dtrace/helper?

2009-06-21 Thread Sang-Thong Chang


My apologies... sent to wrong alias.


Sang-Thong Chang wrote:


Hi all,


Under what circumstance would an exiting process open(2) and
ioctl(2) on /dev/dtrace/helper?


I have an issue here where CU was complaining that their
Oracle processes were taking a long time in ioctl(2) on
/dev/dtrace/helper during shutdown. Removing group/world
readable bit on /dev/dtrace/helper avoided the long shutdown
time.

So, the puzzling question is why /dev/dtrace/helper was being
ioctl(2)ed on?

I doubt this is a binary with USDT.  truss(1M) shown the following:


Base time stamp:  1244728105.2977  [ Thu Jun 11 22:48:25 KST 2009 ]
16003/1:psargs: oraclePOM1 (LOCAL=NO)
16003/1:read(31, 0x1069F3796, 2064) (sleeping...)
...
16003/1:1468.91560.0005 open(/dev/dtrace/helper, 
O_RDWR)  = 8
16003/1:1477.37738.4617 ioctl(8, 
(('d'24)|('t'16)|('h'8)|2), 0x) = 0
16003/1:1477.37810.0008 Received signal #14, SIGALRM 
[caught]
16003/1:1477.37830.0002 lwp_sigmask(SIG_SETMASK, 
0x9FBEF057, 0xFFF7) = 0xFFBFFEFF [0x]


   1277  */
   1278 #define DTRACEHIOC  (('d'  24) | ('t'  16) | 
('h'  8))
   1279 #define DTRACEHIOC_ADD  (DTRACEHIOC | 1)/* add 
helper */
   1280 #define DTRACEHIOC_REMOVE   (DTRACEHIOC | 2)/* 
remove helper */




ustack() from dtrace(1M) shown this was from within the
exit hanlder in exit(2).

 open(/dev/dtrace/helper)

  libc.so.1`open
  libCrun.so.1`0x7a50aed8
  libCrun.so.1`0x7a50b0f4
  ld.so.1`call_fini+0xd0
  ld.so.1`atexit_fini+0x80
  libc.so.1`_exithandle+0x48
  libc.so.1`exit+0x4
  oracle`_start+0x184

It would be great if anyone can help extend my understanding
of dtrace. Thanks.

Regards
Sang Thong


-

1788:   oraclePOM1 (LOCAL=NO)
data model = _LP64  flags = ORPHAN|MSACCT|MSFORK
 /1:flags = ASLEEP  read(0x11,0x106346b16,0x810)

1788:   oracleEOM1 (LOCAL=NO)
/lib/sparcv9/libumem.so.1
/oraom/app/DBMS/lib/libskgxp10.so
/oraom/app/DBMS/lib/libhasgen10.so
/opt/ORCLcluster/lib/libskgxn2.so
/oraom/app/DBMS/lib/libocr10.so
/oraom/app/DBMS/lib/libocrb10.so
/oraom/app/DBMS/lib/libocrutl10.so
/oraom/app/DBMS/lib/libjox10.so
/oraom/app/DBMS/lib/libclsra10.so
/oraom/app/DBMS/lib/libdbcfg10.so
/oraom/app/DBMS/lib/libnnz10.so
/lib/sparcv9/libkstat.so.1
/lib/sparcv9/libnsl.so.1
/lib/sparcv9/libsocket.so.1
/lib/sparcv9/libgen.so.1
/lib/sparcv9/libdl.so.1
/usr/lib/sparcv9/libsched.so.1
/lib/sparcv9/libc.so.1
/lib/sparcv9/libaio.so.1
/lib/sparcv9/librt.so.1
/lib/sparcv9/libm.so.2
/opt/SUNWcluster/lib/sparcv9/libudlm.so
/usr/cluster/lib/sparcv9/libhaops.so.1
/lib/sparcv9/libmd.so.1
/usr/cluster/lib/sparcv9/libscha.so.1
/usr/cluster/lib/sparcv9/libsecurity.so.1
/usr/cluster/lib/sparcv9/libclos.so.1
/lib/sparcv9/libdoor.so.1
/usr/lib/sparcv9/libzonecfg.so.1
/lib/sparcv9/libsecdb.so.1
/usr/lib/sparcv9/libCstd.so.1
/usr/lib/sparcv9/libCrun.so.1
/lib/sparcv9/libuuid.so.1
/lib/sparcv9/libnvpair.so.1
/lib/sparcv9/libsysevent.so.1
/lib/sparcv9/libsec.so.1
/usr/lib/sparcv9/libbrand.so.1
/usr/lib/sparcv9/libpool.so.1
/lib/sparcv9/libscf.so.1
/lib/sparcv9/libproc.so.1
/lib/sparcv9/libuutil.so.1
/usr/lib/sparcv9/libxml2.so.2
/lib/sparcv9/libcmd.so.1
/lib/sparcv9/libavl.so.1
/usr/lib/sparcv9/libexacct.so.1
/lib/sparcv9/librtld_db.so.1
/lib/sparcv9/libelf.so.1
/lib/sparcv9/libctf.so.1
/lib/sparcv9/libpthread.so.1
/usr/lib/sparcv9/libz.so.1
/platform/sun4u-us3/lib/sparcv9/libc_psr.so.1

   Mon, 22 Jun 2009 09:06:09 KST
 open(/dev/dtrace/helper)

  libc.so.1`open
  libCrun.so.1`0x7a50aed8
  libCrun.so.1`0x7a50b0f4
  ld.so.1`call_fini+0xd0
  ld.so.1`atexit_fini+0x80
  libc.so.1`_exithandle+0x48
  libc.so.1`exit+0x4
  oracle`_start+0x184

***



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Speeding up resilver on x4500

2009-06-21 Thread Nicholas Lee
On Mon, Jun 22, 2009 at 4:24 PM, Stuart Anderson
ander...@ligo.caltech.eduwrote:


 However, it is a bit disconcerting to have to run with reduced data
 protection for an entire week. While I am certainly not going back to
 UFS, it seems like it should be at least theoretically possible to do this
 several orders of magnitude faster, e.g., what if every block on the
 replacement disk had its RAIDZ2 data recomputed from the degraded


Maybe this is also saying - that for large disk sets a single RAIDZ2
provides a false sense of security.

Nicholas
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss