Re: [zfs-discuss] If you have ZFS in production, willing to share some details (with me)?

2009-09-19 Thread Blake
On Fri, Sep 18, 2009 at 1:51 PM, Steffen Weiberle
steffen.weibe...@sun.com wrote:
 I am trying to compile some deployment scenarios of ZFS.

 # of systems
3

 amount of storage
10 TB on storage server (can scale to 30)

 application profile(s)
NFS and CIFS

 type of workload (low, high; random, sequential; read-only, read-write,
 write-only)
Boot drives, Nearline backup, Postgres DB (OpenNMS)

 storage type(s)
SATA

 industry
Software

 whether it is private or I can share in a summary
 anything else that might be of interest
You can share my info :)


 Thanks in advance!!

 Steffen
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] deduplication

2009-09-18 Thread Blake
Thanks James!  I look forward to these - we could really use dedup in my org.

Blake

On Thu, Sep 17, 2009 at 6:02 PM, James C. McPherson
james.mcpher...@sun.com wrote:
 On Thu, 17 Sep 2009 11:50:17 -0500
 Tim Cook t...@cook.ms wrote:

 On Thu, Sep 17, 2009 at 5:27 AM, Thomas Burgess wonsl...@gmail.com wrote:

 
  I think you're right, and i also think we'll still see a new post asking
  about it once or twice a week.
 [snip]
 As we should.  Did the video of the talks about dedup ever even get posted
 to Sun's site?  I never saw it.  I remember being told we were all idiots
 when pointing out that it had mysteriously not been posted...

 Hi Tim,
 I certainly do not recall calling anybody an idiot for asking
 about the video or slideware.


 I definitely _do_ recall asking for people to be patient because

 (1) we had lighting problems with the auditorium which interfered
    with recording video

 (2) we have been getting the videos professionally edited so that
    when we can put them up on an appropriate site (which I imagine
    will be slx.sun.com), then the vids will adhere to the high
    standards which you have come to expect.

 (3) professional editing of videos takes time and money. We are
    getting this done as fast as we can.


 I asked Deirdre about the videos yesterday, she said that they
 are almost ready. Rest assured that when they are ready I will
 announce their availability as soon as I possibly can.


 James C. McPherson
 --
 Senior Kernel Software Engineer, Solaris
 Sun Microsystems
 http://blogs.sun.com/jmcp       http://www.jmcp.homeunix.com/blog
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Expanding a raidz pool?

2009-09-02 Thread Blake
On Fri, Aug 28, 2009 at 11:22 PM, Ty
Newtonty.new...@copperchipgames.com wrote:
 Hi,
 I've read a few articles about the lack of 'simple' raidz pool expansion 
 capability in ZFS.  I am interested in having a go at developing this 
 functionality.  Is anyone working on this at the moment?

 I'll explain what I am proposing.  As mentioned in many forums, the concept 
 is really simple: allow a raidz pool to grow by adding one or more disks to 
 an existing pool.  My intended user group is the consumer market, as opposed 
 to the enterprise, so I expect I'll put some rather strict limitations on 
 how/when this functionality will operate: to make the first implementation 
 more achievable.

 The use case I will try and solve first is, what I see as, the simplest.  I 
 have a raidz pool configured with 1 file system on top; no snapshots.  I want 
 to add an additional disk (must be at least the same size as the rest of the 
 disks in the pool).  I don't mind if there is some downtime.  I want all my 
 data to take advantage of the additional disk.

Have you looked at the 'add' section of the zpool manpage?  You can
add another vdev, provided it provides similar parity, something like:

zpool add data raidz2 c4t14d0 c4t15d0 c5t12d0 c5t13d0 c5t14d0

which I did a few weeks ago.  Here, I had a raidz2 pool called 'data'
made up of 5 disks.  I added another 5 disks also configured as raidz2
with this command.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Status/priority of 6761786

2009-08-27 Thread Blake
I think the value of auto-snapshotting zvols is debatable.  At least,
there are not many folks who need to do this.

What I'd rather see is a default property of 'auto-snapshot=off' for zvols.

Blake

On Thu, Aug 27, 2009 at 4:29 PM, Tim Cookt...@cook.ms wrote:


 On Thu, Aug 27, 2009 at 3:24 PM, Remco Lengers re...@lengers.com wrote:

 Dave,

 Its logged as an RFE (Request for Enhancement) not as a CR (bug).

 The status is 3-Accepted/  P1  RFE

 RFE's are generally looked at in a much different way then a CR.

 ..Remco


 Seriously?  It's considered works as designed for a system to take 5+
 hours to boot?  Wow.

 --Tim

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs+nfs: scary nfs log entries?

2009-08-19 Thread Blake Irvin
I have a zfs dataset that I use for network home directories.  The box is 
running 2008.11 with the auto-snapshot service enabled.  To help debug some 
mysterious file deletion issues, I've enabled nfs logging (all my clients are 
NFSv3 Linux boxes).

I keep seeing lines like this in the nfslog:
br
br
pre
Wed Aug 19 10:20:48 2009 0 host.name.domain.com 1168 
zfs-auto-snap.hourly-2009-08-17-09.00/username/incoming.file b _ i r 0 nfs 0 *
/pre
br

Why is my path showing up with the name of a snapshot?  This scares me since 
snapshots get rolled off automatically...on the other hand, I know that the 
snapshots are read-only.  Any insights?
br
br

Blake
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] missing disk space

2009-08-03 Thread Blake
On Mon, Aug 3, 2009 at 12:35 PM, David E. Andersondanders...@gmail.com wrote:
 I am new to ZFS, so please bear with me...

 I created a raidz1 pool from three 1.5TB disks on OpenSolaris 2009.6.  I see
 less than 1TB useable space.  What did I do wrong?

 $ zpool list
 NAME  SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
 rpool 464G  42.2G   422G 9%  ONLINE  -
 storage  1.36T   143K  1.36T 0%  ONLINE  -

 $ df -h /storage
 Filesystem    Size  Used Avail Use% Mounted on
 storage   913G   26K  913G   1% /storage

 $ zpool status
   pool: rpool
  state: ONLINE
  scrub: none requested
 config:

     NAME  STATE READ WRITE CKSUM
     rpool ONLINE   0 0 0
       mirror  ONLINE   0 0 0
         c8t1d0s0  ONLINE   0 0 0
         c10d0s0   ONLINE   0 0 0

 errors: No known data errors

   pool: storage
  state: ONLINE
  scrub: none requested
 config:

     NAME STATE READ WRITE CKSUM
     storage  ONLINE   0 0 0
       raidz1 ONLINE   0 0 0
         c9d0s2   ONLINE   0 0 0
         c10d0s2  ONLINE   0 0 0
         c10d1s2  ONLINE   0 0 0

 errors: No known data errors

 --
 David


can you post the output of 'zfs get all storage' ?

blake
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] missing disk space

2009-08-03 Thread Blake
On Mon, Aug 3, 2009 at 1:41 PM, David E. Andersondanders...@gmail.com wrote:
 $ zfs get all storage
 NAME PROPERTY  VALUE  SOURCE
 storage  type  filesystem -
 storage  creation  Fri Jul 10 21:19 2009  -
 storage  used  89.2K  -
 storage  available 913G   -
 storage  referenced    25.3K  -
 storage  compressratio 1.00x  -
 storage  mounted   yes    -
 storage  quota none   default
 storage  reservation   none   default
 storage  recordsize    128K   default
 storage  mountpoint    /storage   default
 storage  sharenfs  off    default
 storage  checksum  on default
 storage  compression   off    default
 storage  atime on default
 storage  devices   on default
 storage  exec  on default
 storage  setuid    on default
 storage  readonly  off    default
 storage  zoned off    default
 storage  snapdir   hidden default
 storage  aclmode   groupmask  default
 storage  aclinherit    restricted default
 storage  canmount  on default
 storage  shareiscsi    off    default
 storage  xattr on default
 storage  copies    1  default
 storage  version   3  -
 storage  utf8only  off    -
 storage  normalization none   -
 storage  casesensitivity   sensitive  -
 storage  vscan off    default
 storage  nbmand    off    default
 storage  sharesmb  off    default
 storage  refquota  none   default
 storage  refreservation    none   default
 storage  primarycache  all    default
 storage  secondarycache    all    default
 storage  usedbysnapshots   0  -
 storage  usedbydataset 25.3K  -
 storage  usedbychildren    63.9K  -
 storage  usedbyrefreservation  0

 On Mon, Aug 3, 2009 at 10:34 AM, Blake blake.ir...@gmail.com wrote:

 On Mon, Aug 3, 2009 at 12:35 PM, David E. Andersondanders...@gmail.com
 wrote:
  I am new to ZFS, so please bear with me...
 
  I created a raidz1 pool from three 1.5TB disks on OpenSolaris 2009.6.  I
  see
  less than 1TB useable space.  What did I do wrong?
 snip


 can you post the output of 'zfs get all storage' ?

 blake



 --
 David


Looking back at your 'zpool status' output, I think you might have
accidentally made put one of your big storage disks in your rpool, and
one of your little rpool disks in your big storage pool.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] why is zpool import still hanging in opensolaris 2009.06 ??? no fix yet ???

2009-07-24 Thread Blake
This sounds like a bug I hit - if you have zvols on your pool, and
automatic snapshots enabled, the thousands of resultant snapshots have
to be polled by devfsadm during boot, which take a long time - several
seconds per zvol.

I remove the auto-snapshot property from my zvols and the slow boot stopped.

Blake



On Thu, Jul 23, 2009 at 5:53 AM, Luc De Meyerno-re...@opensolaris.org wrote:
 Follow-up : happy end ...

 It took quite some thinkering but... i have my data back...

 I ended up starting without the troublesome zfs storage array, de-installed 
 the iscsitartget software and re-installed it...just to have solaris boot 
 without complaining about missing modules...

 That left me with a system that would boot as long as the storage was 
 disconnected... Reconnecting it made the boot stop at the hostname. Then the 
 disk activity light would flash every second or so forever... I then rebooted 
 using milestone=none. That worked also with the storage attached! So now I 
 was sure that some software process was causing a hangup (or what appeared to 
 be a hangup.) I could now in milestone none verify that the pool was intact: 
 and so it was... fortunately I had not broken the pool itself... all online 
 with no errors to report.
 I then went to milestone-all which again made the system hang with the disk 
 activity every second forever. I think the task doing this was devfsadm. I 
 then assumed on a gut feeling that somehow the system was scanning or 
 checking the pool. I left the system overnight in a desperate attempt because 
 I calculated the 500GB checking to take about 8 hrs if the system would 
 *really* scan everything. (I copied a 1 TB drive last week which took nearly 
 20 hrs, so I learned that sometimes I need to wait... copying these big disks 
 takes a *lot* of time!)

 This morning I switched on the monitor and lo and behold : a login screen 
 The store was there!

 Lesson for myself and others: you MUST wait at the hostname line: the system 
 WILL eventually come online... but don't ask how long it takes... I hate to 
 think how long it would take if I had a 10TB system. (but then again, a 
 file-system-check on an ext2 disk also takes forever...)

 I re-enabled the iscsitgtd and did a list : it saw one of the two targets ! 
 (which was ok because I remembered that I had turned off the shareiscsi flag 
 on the second share.
 I then went ahead and connected the system back into the network and 
 repaired the iscsi-target on the virtual mainframe : WORKED ! Copied over 
 the virtual disks to local store so I can at least start up these servers 
 asap again.
 Then set the iscsishare on the second and most important share: OK! Listed 
 the targets: THERE, BOTH! Repaired it's connection too: WORKED...!

 I am copying everything away from the ZFS pools now, but my data is 
 recovered... fortunately.

 I now have mixed feelings about the ordeal: yes Sun Solaris kept its promise: 
 I did not loose my data. But the time and trouble it took to recover in this 
 case (just to restart a system for example taking an overnight wait!) is 
 something that a few of my customers would *seriously* dislike...

 But: a happy end after all... most important data rescued and 2nd important : 
 I learned a lot in the process...

 Bye
 Luc De Meyer
 Belgium
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Things I Like About ZFS

2009-06-19 Thread Blake
On Fri, Jun 19, 2009 at 10:30 PM, Dave Ringkorno-re...@opensolaris.org wrote:
 I'll start:

 - The commands are easy to remember -- all two of them.  Which is easier, SVM 
 or ZFS, to mirror your disks?  I've been using SVM for years and still have 
 to break out the manual to use metadb, metainit, metastat, metattach, 
 metadetach, etc.  I hardly ever have to break out the ZFS manual.  I can 
 actually remember the commands and options to do things.  Don't even start me 
 on VxVM.

 - Boasting to the unconverted.  We still have a lot of VxVM and SVM on 
 Solaris, and LVM on AIX, in the office.  The other admins are always having 
 issues with storage migrations, full filesystems, Live Upgrade, corrupted 
 root filesystems, etc.  I love being able to offer solutions to their 
 immediate problems, and follow it up with, You know, if your box was on ZFS 
 this wouldn't be an issue.

Interesting.  Usually the problems make their way to this list more
than the successes.  Glad to hear it!

BTW, ZFS just saved my skin tonight after I botched an OpenNMS upgrade
and was able to go back to my auto-snapshots :)

And there was a power failure earlier that took down a bunch of hosts
that rely on our multi-terabyte ZFS filer, as well as the filer itself
- no waiting around for fsck, thanks!

Blake
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] replicating a root pool

2009-05-22 Thread Blake
On Fri, May 22, 2009 at 12:40 AM, Ian Collins i...@ianshome.com wrote:

 Mark J Musante wrote:

 On Thu, 21 May 2009, Ian Collins wrote:

  I'm trying to use zfs send/receive to replicate the root pool of a system
 and I can't think of a way to stop the received copy attempting to mount the
 filesystem over the root of the destination pool.


 If you're using build 107 or later, there's a hidden -u option available
 for zfs receive to tell it not to mount the dataset.  See
 http://tinyurl.com/crgog8 for more details.

  Thanks for the tip Mark, unfortunately I'm stuck with Solaris 10 on this
 system.


I just did this the old way, and it wasn't that hard.  I didn't even script
it (yet), but it seems like it should be easy to do if you use the
 solarisinternals recipe.

Blake
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs reliability under xen

2009-05-22 Thread Blake
On Fri, May 22, 2009 at 2:44 PM, Ahmed Kamal 
email.ahmedka...@googlemail.com wrote:

 However, if you need to decide, whether to use Xen, test your setup
 before going into production and ask your boss, whether he can live with
 innovative ... solutions ;-)


 Thanks a lot for the informative reply. It has been definitely helpful
 I am however interested in the reliability of running the ZFS stack as Xen
 domU (and not dom0). For instance, I am worried that the emulated disk
 controller would not obey flushes, or write ordering thus stabbing zfs in
 the back.

 Regards


 I've gotten very good performance numbers for I/O out of a 2008.11 PV domU
with a zfs zvol as the storage device/install disk

Blake
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Compression/copies on root pool RFE

2009-05-06 Thread Blake
On Wed, May 6, 2009 at 11:14 AM, Rich Teer rich.t...@rite-group.com wrote:
 On Wed, 6 May 2009, Richard Elling wrote:

 popular interactive installers much more simplified.  I agree that
 interactive installation needs to remain as simple as possible.

 How about offering a choice an installation time: Custom or default??

 Those that don't want/need the interactive flexibility can pick default
 whereas others who want more flexibility (but still want or need an
 interactive installation) can pick the custom option.  Just a thought...

If you do propose this on caiman-discuss, I'd suggest an option to
mirror 2 boot devices as well.  Doing the slice attach/installgrub is
nontrivial for, say, a user who's primarily a dev.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz vdev size... again.

2009-04-28 Thread Blake
On Tue, Apr 28, 2009 at 10:08 AM, Tim t...@tcsac.net wrote:


 On Mon, Apr 27, 2009 at 8:25 PM, Richard Elling richard.ell...@gmail.com
 wrote:

 I do not believe you can achieve five 9s with current consumer disk
 drives for an extended period, say 1 year.

 Just to pipe up, while very few vendors can pull it off, we've seen five 9's
 with Hitachi gear using SATA.

Can you specify the hardware?

I've recently switched to LSI SAS1068E controllers and am swimmingly
happy.  (That's my $.02 - controllers (not surprisingly) affect the
niceness of a software RAID solution like ZFS quite a bit - maybe even
more than the actual drives...?)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [storage-discuss] Supermicro AOC-SASLP-MV8

2009-04-22 Thread Blake
I'm quite happy so far with my LSI cards, which replaced a couple of
the Supermicro Marvell cards:

# scanpci
...
pci bus 0x0007 cardnum 0x00 function 0x00: vendor 0x1000 device 0x0058
 LSI Logic / Symbios Logic SAS1068E PCI-Express Fusion-MPT SAS


On Wed, Apr 22, 2009 at 2:45 AM, James Andrewartha jam...@daa.com.au wrote:
 myxi...@googlemail.com wrote:
 Bouncing a thread from the device drivers list:
 http://opensolaris.org/jive/thread.jspa?messageID=357176

 Does anybody know if OpenSolaris will support this new Supermicro card,
 based on the Marvell 88SE6480 chipset? It's a true PCI Express 8 port
 JBOD SAS/SATA controller with pricing apparently around $125.

 If it works with OpenSolaris it sounds pretty much perfect.

 The Linux support for the 6480 builds on the 6440 mvsas support, so I don't
 think marvell88sx would work, and there doesn't seem to be a Marvell SAS
 driver for Solaris at all, so I'd say it's not supported.
 http://www.hardforum.com/showthread.php?t=1397855 has a fair few people
 testing it out, but mostly under Windows.

 --
 James Andrewartha
 ___
 storage-discuss mailing list
 storage-disc...@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/storage-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] What can I do to shorten the long awkward names of snapshots?

2009-04-16 Thread Blake
The cool thing about the way Tim has built the service is that you can
edit the variable values in the method script to make snapshot titles
pretty much whatever you want.  I think he made a good compromise
choice between simplicity and clarity in the current titling system.

Remember that the Time Slider snapshot viewer essentially makes this
transparent to a the end user.  And you can make use of Time Slider
remotely using ssh -X hostname and then nautilus --no-desktop.



On Wed, Apr 15, 2009 at 8:32 PM, Andre van Eyssen an...@purplecow.org wrote:
 On Wed, 15 Apr 2009, Harry Putnam wrote:


 Would become:
  a:freq-041509_1630

 Can I suggest perhaps something inspired by the old convention for DNS
 serials, along the lines of fmmddtt? Like:

 a:f200904151630

 This makes things easier to sort and lines up in a tidy manner.


 --
 Andre van Eyssen.
 mail: an...@purplecow.org          jabber: an...@interact.purplecow.org
 purplecow.org: UNIX for the masses http://www2.purplecow.org
 purplecow.org: PCOWpix             http://pix.purplecow.org

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [storage-discuss] Supermicro SAS/SATA controllers?

2009-04-15 Thread Blake Irvin


On Apr 15, 2009, at 8:28 AM, Nicholas Lee emptysa...@gmail.com wrote:




On Tue, Apr 14, 2009 at 5:57 AM, Will Murnane  
will.murn...@gmail.com wrote:


 Has anyone done any specific testing with SSD devices and solaris  
other than

 the FISHWORKS stuff?  Which is better for what - SLC and MLC?
My impression is that the flash controllers make a much bigger
difference than the type of flash inside.  You should take a look at
AnandTech's review of the new OCZ Vertex drives [1], which has a
fairly comprehensive set of benchmarks.  I don't think any of the
products they review are really optimal choices, though; the Intel
X25-E drives look good until you see the price tag, and even they only
do 30-odd MB/s random writes.


Couple excellent articles about SSD from adandtech last month:
http://www.anandtech.com/showdoc.aspx?i=3532 - SSD versus Enterprise  
SAS and SATA disks (20/3/09)
http://www.anandtech.com/storage/showdoc.aspx?i=3531 - The SSD  
Anthology: Understanding SSDs and New Drives from OCZ (18/3/09)


And it looks like the Intel fragmentation issue is fixed as well: 
http://techreport.com/discussions.x/16739

It's a shame the Sun Writezilla devices are almost 10k USD - seems  
there are the only units on the market that (apart from cost) work  
well all around as slog devices - form factor, interface/drivers  
and performance.
What about the new flash drives Andy was showing off in Vegas?  Those  
looked small (capacity) - perhaps cheap too?






How much of an issue does the random write bandwidth limit have on a  
slog device? What about latency?  I would have thought the write  
traffic pattern for slog io was more sequential and bursty.



Nicholas
___
storage-discuss mailing list
storage-disc...@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/storage-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How recoverable is an 'unrecoverable error'?

2009-04-15 Thread Blake
On Wed, Apr 15, 2009 at 11:49 AM, Uwe Dippel udip...@gmail.com wrote:
 Bob Friesenhahn wrote:

 Since it was not reported that user data was impacted, it seems likely
 that there was a read failure (or bad checksum) for ZFS metadata which is
 redundantly stored.

 (Maybe I am too much of a linguist to not stumble over the wording here.) If
 it is 'redundant', it is 'recoverable', am I right? Why, if this is the
 case, does scrub not recover it, and scrub even fails to correct the CKSUM
 error as long as it is flagged 'unrecoverable', but can do exactly that
 after the 'clear' command?


 Ubuntu Linux is unlikely to notice data problems unless the drive reports
 hard errors.  ZFS is much better at checking for errors.

 No doubt. But ext3 also seems to need much less attention, very much fewer
 commands. Which leaves it as a viable alternative. I still hope that one day
 ZFS will be maintainable as simple as ext3; respectively do all that
 maintenance on its own.  :)

 Uwe

You only need to decide what you want here.  Yes, ext3 requires less
maintenance, because it can't tell you if a block becomes corrupt
(though fsck-in when that *does* happen can require hours, compared to
zfs replacing with a good block from the other half of your mirror).

ZFS can *fully* do it's job only when it has several copies of blocks
to choose from.  Since you have only one disk here, ZFS can only say
'hey, your checksum for this block is bad - sorry'.  ext3 might do the
same thing, though only if you tried to use the block with an
application that knew what the block was supposed to look like.

That said, I think your comments raise a valid point that ZFS could be
a little easier for individuals to use.  I totally understand why Sun
doesn't focus on end-user management tools (not their market) - on the
other hand, the code is out there, so if you see a problem, get some
people together to write some management tools! :)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [Fwd: ZFS user/group quotas space accounting [PSARC/2009/204 FastTrack timeout 04/08/2009]]

2009-03-31 Thread Blake
much cheering ensues!

2009/3/31 Matthew Ahrens matthew.ahr...@sun.com:
 FYI, I filed this PSARC case yesterday, and expect to integrate into
 OpenSolaris in April.  Your comments are welcome.

 http://arc.opensolaris.org/caselog/PSARC/2009/204/

 --matt


 -- Forwarded message --
 From: Matthew Ahrens ahr...@dm-eng-01.sfbay.sun.com
 To: psarc-...@sun.com
 Date: Mon, 30 Mar 2009 20:39:05 -0700 (PDT)
 Subject: ZFS user/group quotas  space accounting [PSARC/2009/204 FastTrack
 timeout 04/08/2009]

 Template Version: @(#)sac_nextcase 1.68 02/23/09 SMI
 This information is Copyright 2009 Sun Microsystems
 1. Introduction
    1.1. Project/Component Working Name:
         ZFS user/group quotas  space accounting
    1.2. Name of Document Author/Supplier:
         Author:  Matthew Ahrens
    1.3  Date of This Document:
        30 March, 2009
 4. Technical Description
 ZFS user/group space accounting

 A. SUMMARY

 This case adds support to ZFS for user/group quotas  per-uid/gid space
 tracking.

 B. PROBLEM

 Enterprise customers often want to know who is using space, based on
 what uid and gid owns each file.

 Education customers often want to apply per-user quotas to hundreds of
 thousands of users.  In these situations, the number of users and/or
 existing infrastructure prohibits using one filesystem per user and
 setting filesystem-wide quotas.

 C. PROPOSED SOLUTION

 1. Overview

 Each filesystem keeps track of how much space inside it is owned by each
 user (uid) and group (gid).  This is the amount of space referenced,
 so relationships between filesystems, descendents, clones, and snapshots
 are ignored, and each tracks their user used and group used
 independently.  This is the same policy behind the referenced,
 refquota, and refreservation properties.  The amount of space
 charged is the amount of space reported by struct stat's st_blocks and
 du(1).

 Both POSIX ids (uid  gid) and untranslated SIDs are supported (eg, when
 sharing filesystems over SMB without a name service translation set up).

 ZFS will now enforce quotas on the amount of space referenced by files
 owned by particular users and groups.  Enforcement may be delayed by
 several seconds.  In other words, users may go a bit over their quota
 before the system notices that they are over quota and begins to refuse
 additional writes with EDQUOT.  This decision was made to get the
 feature to market in a reasonable time, with a minimum of engineering
 resources expended.  The design and implementation do not preclude
 implementing strict enforcement at a later date.

 User space accounting and quotas stick with each dataset (snapshot,
 filesystem, and clone).  This means that user quotas (and space
 accounting) are not inherited.  They will be copied to a new snapshot,
 and keep the values they had at the time the snapshot was taken.
 Likewise, user quotas will be copied to a clone (from its origin
 snapshot), and they will be copied with zfs send (even without -R).
 (User accounting and quota information is not actually copied to
 snapshots and clones, just referenced and copied-on-write like other
 filesystem contents.)

 The user space accounting and quotas is reported by the new
 userused@user, groupused@group, userquota@user, and
 groupquota@group properties, and by the new zfs userspace and zfs
 groupspace subcommands, which are detailed below.

 2. Version Compatibility

 To use these features, the pool must be upgraded to a new on-disk
 version (15). Old filesystems must have their space accounting
 information initialized by running zfs userspace fs or upgrading the
 old filesystem to a new on-disk version (4).  To set user quotas, the
 pool and filesystem must both be upgraded.

 3. Permissions

 Setting or changing user quotas are administrative actions, subject to
 the same privilege requirements as other zfs subcommands.  There are new
 userquota and groupquota permissions which can be granted with zfs
 allow, to allow those properties to be viewed and changed.

 Unprivileged users can only view their own userquota and userused, and
 the groupquota and groupused of any groups they belong to.  The new
 userused and groupused permissions can be granted with zfs allow
 to permit users to view these properties.

 The existing version permission (granted with zfs allow) permits the
 accounting information to be initialized by zfs userspace.

 4. New Properties

 user/group space accounting information and quotas can be manipulated
 with 4 new properties:

 zfs get userused@user fs|snap
 zfs get groupused@group fs|snap

 zfs get userquota@user fs|snap
 zfs get groupquota@group fs|snap

 zfs set userquota@user=quota fs
 zfs set groupquota@user=quota fs

 The user or group is specified using one of the following forms:
 posix name (eg. ahrens)
 posix numeric id (eg. 126829)
 sid name (eg. ahr...@sun)
 sid numeric id (eg. S-1-12345-12423-125829)

 For zfs set, if a nonexistent name is specified, an error is
 

Re: [zfs-discuss] Data corruption during resilver operation

2009-03-30 Thread Blake
You are seeing snapshots from Time-Slider's automatic snapshot service.

If you have a copy of each of these 58 files elsewhere, I suppose you
could re-copy them to the mirror and then do 'zpool clear [poolname]'
to reset the error counter.



On Sun, Mar 29, 2009 at 10:28 PM, Harry Putnam rea...@newsguy.com wrote:
 I'm in well over my head with this report from zpool status saying:

 root # zpool status z3
  pool: z3
  state: DEGRADED
 status: One or more devices has experienced an error resulting in data
        corruption.  Applications may be affected.
 action: Restore the file in question if possible.  Otherwise restore the
        entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
  scrub: resilver completed after 0h7m with 38 errors on Sun Mar 29 18:37:28 
 2009
 config:

        NAME        STATE     READ WRITE CKSUM
        z3          DEGRADED     0     0    40
          mirror    DEGRADED     0     0    80
            c5d0    DEGRADED     0     0    80  too many errors
            c6d0    DEGRADED     0     0    80  too many errors

 This is that last thing and apparently the result of a series of steps
 I've taken to increase a zpool mirrors size.

 There was quite a lot of huffing and puffing with the sata controller
 that holds this mirror but the short version is:

 zpool z3 created as mirror on 2 older 200gb SATAI disks.  On an
 adaptec 1205sa PCI controller.

 After deciding I wanted increase the size of this pool, I detached 1
 disk, then pulled it out.  I replaced it with a newer bigger sata II
 wd750 gb disk.  When I attempted to startup and attach this disk, I
 didn't get by the boot process, and discovered my sata controller
 could not handle the newer SATAII disk..  No boot was possible.

 I finally got the sata contoller in shape to work by flashing the 2
 part BIOS with latest bios for that card. (Sil 3112a chip).

 Restarted with 1 original 200gb disk and 1 new 750gb disk.
 It booted and I was abble to attach the new larger drive and begin the
 resilvering process.

 I went on to other things, but when I checked back I found the error
 report cited above.

 I stared looking through the data but didn't really see much wrong.  I
 check the byte size with `du -sb' on the zpool and the source of the
 data on a remote linux host.  They were not the same but quite close.
 I didn't think that meant much since it was on different filesystems.
 zfs and reiserfs.

 I went to the web page cited in the report to see what I could learn.
 To summarize it said this was serious business.  That data might not
 even be able to be removed but that for sure it needed to be replaced
 from clean backup.

 Using zpool status -v z3  I learned there were 51 files said to be
 corrupt.  But when I looked at the files they were not part of the
 original data.

 The original data was put there by an rsync process from a remote
 host. and contained none of the named files.  There files are of the
 form (wrapped for mail):

  z3/www/rea...@zfs-auto-snap:frequent-2009-03-29-18:55:\
    /www/localhost/htdocs/lcweb/TrainingVids/VegasTraining/\
       VegasTraiiningTransitions.avi

  (All on one line)

 I'm not at all clear on what this is.  The part after the colon is
 what was rsynced over.  The files that turned up in the report are all
 *.mov *.avi, *.mpg or *.pdf.

 I didn't make any snapshots, nor did I set anything to have them made
 automatically... so not sure where this snapshot came from or really
 even if it is in fact a snapshot.

 Is it somehow a product of the resilvering?

 When I go to the root of this filesystem (/www) and run a find command
 like:
  find . -name 'VegasTraiiningTransitions.avi'

 The file is found.  I haven't been able to test if they play yet but
 wondering what this snapshot stuff means.  And what I should do about
 it.

 The warning clearly suggests they must be replaced with good copies.

 That wouldn't be too big a deal, but I do still have the other new
 disk to insert and resilver.

 So what is the smart move here?... Replace the data before continuing
 with the enlargement of the pool? Or something else?


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-30 Thread Blake
On Sat, Mar 28, 2009 at 6:57 PM, Ian Collins i...@ianshome.com wrote:
 Please stop top-posting to threads where everyone else is normal-posting, it
 mucks up the flow of the thread.

 Thanks,

 --
 Ian.

Apologies - top-posting seems to be the Gmail default (or I set it so
long ago that I forgot it was there).
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Timeslider causing errors..?

2009-03-30 Thread Blake
Do you have more than one Boot Environment?

pfexec beadm list



On Mon, Mar 30, 2009 at 1:33 PM, Harry Putnam rea...@newsguy.com wrote:
 After messing around with Timeslider... I started getting errors and
 the frequent and hourly services were failing, causing the service to
 be put into maintenance status.

 Not really being sure what to do with it I first tried
 `svcadm restart' on them.  But they  went right back into maintenance
 mode.

 I did save the logs and have posted them here:

 Is it the failure to open crontab that's causing all the fuss?

  www.jtan.com/~reader/slider/disp.cgi


 I then turned the services for frequent and hourly clear off with
 `svcadm disable'

 I thought that would allow me to access the timeslider applet which
 had been showing only the same error as svcs -xv showed.

 But it didn't .  the dialog still shows:
  Snapshot manager service dependency error

 but now of course:
  disabled    svc:/system/filesystem/zfs/auto-snapshot:frequent
  disabled    svc:/system/filesystem/zfs/auto-snapshot:hourly

 Before it showed some of the same stuff as `svcs -xv'

  (see the logs a url above)

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Data corruption during resilver operation

2009-03-30 Thread Blake
Sounds like the best way - I was about to suggest that anyway :)

On Mon, Mar 30, 2009 at 3:03 PM, Harry Putnam rea...@newsguy.com wrote:
 Blake blake.ir...@gmail.com writes:
 You are seeing snapshots from Time-Slider's automatic snapshot service.

 If you have a copy of each of these 58 files elsewhere, I suppose you
 could re-copy them to the mirror and then do 'zpool clear [poolname]'
 to reset the error counter.

 Thanks... I did try coping from the source to replace those but it
 didn't appear to make any difference... still got the errors.

 I finally just assumed I'd done something untoward during all the
 shuffle of upgrading a 200gb mirror to a 750 gb mirror and flashing
 the bios of the PCI sata controller card in the middle.

 So resorted to zpool destroy badpool

 Finished the switch from 200gb to 750gb with no zpool on either.

 Created the mirror using the 2 750gb disks.  And finally rsynced the
 data across from a linux machine to the new zpool as before.


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] j4200 drive carriers

2009-03-30 Thread Blake
no idea how many of these there are:

http://www.google.com/products?q=570-1182hl=enshow=li

2009/3/30 Tim t...@tcsac.net:


 On Mon, Mar 30, 2009 at 3:56 AM, Mike Futerko m...@maytech.net wrote:

 Hello

  1) Dual IO module option
  2) Multipath support
  3) Zone support [multi host connecting to same JBOD or same set of
  JBOD's
  connected in series. ]

 This sounds interesting - where I can read more about connecting two
 hosts to same J4200 etc?


 Thanks
 Mike

 FWIW, it looks like someone at Sun saw the complaints in this thread and or
 (more likely) had enough customer complaints.  It appears you can buy the
 tray independently now.  Although, it's $500 (so they're apparently made
 entirely of diamond and platinum).  In Sun marketing's defense, that was a
 great way of making the drive prices seem reasonable.

 http://sunsolve.sun.com/handbook_pub/validateUser.do?target=Systems/J4200/components

 --Tim
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] About snapshots or versioned backups

2009-03-30 Thread Blake
you need zfs list -t snapshot

by default, snapshots aren't shown in zfs list anymore, hence the -t option



On Mon, Mar 30, 2009 at 11:41 AM, Harry Putnam rea...@newsguy.com wrote:
 Richard Elling richard.ell...@gmail.com writes:

 It can go very fine, though you'll need to set the parameters yourself,
 if you want to use different settings.  A few weeks ago, I posted a way
 to see the settings, which the time slider admin tool won't show.  There
 is a diminishing return for exposing such complexity, but you might try
 an RFE if you feel strongly about it.
 http://opensolaris.org/jive/thread.jspa?messageID=353761

 I meant in terms of per directory not frequency.



 If I wanted to be able to go back in time with just /etc for example.
 Is it also possible?


 Possible? Yes, to some degree.  But that is probably not worth the
 complexity
 involved. The contents of /etc just doesn't change very
 often. Snapshots are
 done on a per-file system basis and /etc doesn't really warrant a
 separate file
 system -- and I'm not sure you can separate /etc from /, since it is
 required
 early in the boot sequence.

 The part about etc not changing may be true after you've established a
 good setup.  But while getting there, I've always found it a good
 choice for frequent backup.  That has been on linux, not solaris.

 Maybe Osol doesn't use etc as much as linux systems.

 But then other directories may need more frequent backup than the
 filesystem they are on.  I guess one could create a filesystem for
 such a directory.

 And I think I may be getting confused between filesystem snapshots and
 BE snapshots. The etc directory must be included in a BE.  So would BE
 snapshots cover all of rpool?  Or just `/' or are they even different.

 The whole scheme I see with gnu/bin/df is kind of confusing too.  Its
 not really even clear if /etc is part of rpool.

 zfs list -r rpool doesn't show etc, just '/'so I guess not.


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zfs using java

2009-03-30 Thread Blake
Can you list the exact command you used to launch the control panel?
I'm not sure what tool you are referring to.



2009/3/25 Howard Huntley hhuntle...@comcast.net:
 I once installed ZFS on my home Sun Blade 100 and it worked fine on the sun
 blade 100 running solaris 10. I reinstalled Solaris 10 09 version and
 created a zpool which is not visible using the java control panel. When I
 attempt to run the Java control panel to manage the zfs system I receive an
 error message stating !Launch Error, No application is registered with this
 sun Java or I have no rights to use any applications that are registered,
 see my sys admin. Can any one tell me how to get this straightened out. I
 have been fooling around with it for some time now.

 Is any one is Jacksonville, Florida??
 --

 Howard Huntley Jr. MCP, MCSE
 Micro-Computer Systems Specialist

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] About snapshots or versioned backups

2009-03-30 Thread Blake
There is a bug where the automatic snapshot service dies if there are
multiple boot environments.  Do you have these?  I think you can check
with Update Manager.



On Mon, Mar 30, 2009 at 7:20 PM, Harry Putnam rea...@newsguy.com wrote:
 you need zfs list -t snapshot

 by default, snapshots aren't shown in zfs list anymore, hence the -t option


 Yikes, I've got dozens of the things... I monkeyed around a bit with
 timeslider but thought I canceled out whatever settings I'd messed
 with.

 Frequent and hourly both are way to often for most of my data.

 I think I've kind of painted myself into a corner.  I apparently
 turned on timeslider... but one pool had some kind of corruption
 problem that I fixed by zpool destroy entire pool.

 But I keep getting errors from timeslider that would put frequent and
 hourly into maintenance mode.  Which meant I couldn't do anything with
 the timeslider applet.  It seems a little light on robustness.. not
 able to be used if there is any problem.

 Finally I disabled both frequent and hourly... and of course then the
 timeslider  I unusable because the services are off.

 I tried restarting them again after getting the offending pool rebuilt
 involving at least 2 reboots.  But now restarting them, and
 immediately they go to maintenance mode.  And of course the timeslider
 applet is useless.

 Looking at the log output its the same as what I posted earlier... in
 a different thread.

   www.jtan.com/~reader/slider/disp.cgi

 It appears to be related to not being able to open a crontab file.

 Doesn't say which but I see several in /var/pool/cron/crontabs

  ls -l /var/spool/cron/crontabs
 total 9
 -rw--- 1 root sys    1004 2008-11-19 18:13 adm
 -r 1 root root   1365 2008-11-19 18:30 lp
 -rw--- 1 root root   1241 2009-03-30 17:15 root
 -rw--- 1 root sys    1122 2008-11-19 18:33 sys
 -rw--- 1 root daemon  394 2009-03-30 18:06 zfssnap

 So I'm not sure what the problem is.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Blake
Have you checked the specs of the 1205 to see what maximum drive size
it supports?  That's an older card, IIRC, so it might top out at 500gb
or something.

On Sat, Mar 28, 2009 at 10:33 AM, Harry Putnam rea...@newsguy.com wrote:
 casper@sun.com writes:

I mentioned that pressing F3 doesn't do anything.  That is, I have
now way to enter the configuration tool.


 Does it work when you first remove the 200GB drive, reboot, press F3
 and see what you can do in that configuration tool.

 It is possible that it first needs to forget the 200GB drive.

 OK, booted with no sata drives attached.  At the press F3 screen,
 again pressing F3 appears to have no effect.  But after a moment two
 messages saying `no drives found' print out, and then boot proceeds.

 Once booted up I see the recurring message where I should see a login
 prompt (I'm setup to boot into console mode).

  ata_id_common Busy Status 0xfe error 0x0

 Repeated 4 times, then after maybe a 2-3 minute wait the regular login
 prompt appears.

 There is something more going on that is dragging the machine down so
 that typing at the prompt is delayed by huge pauses.

 I'm not really able to find out what it is yet but will post back
 later after a reboot.


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] j4200 drive carriers

2009-03-28 Thread Blake
This is true.  Unfortunately, in my experience, controller quality is
still very important.  ZFS can preserve data all day long, but that
doesn't help much if the controller misbehaves (you may have good data
that can't be retrieved or manipulated properly - it's happened to me
with whitebox hardware).

If anyone buys whitebox hardware for ZFS in production, make sure the
vendor will give you support/warranty for OpenSolaris/ZFS.



On Fri, Mar 27, 2009 at 7:33 PM, jpdrawneek j...@drawneek.demon.co.uk wrote:
 Mertol Ozyoney wrote:

 But the whole point of zfs is that you can use inexpensive drives and with
 enough in RAID to make it reliable.

 Best regards
 Mertol

 Mertol Ozyoney Storage Practice - Sales Manager

 Sun Microsystems, TR
 Istanbul TR
 Phone +902123352200
 Mobile +905339310752
 Fax +90212335
 Email mertol.ozyo...@sun.com



 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Richard Elling
 Sent: Sunday, February 01, 2009 10:24 PM
 To: John-Paul Drawneek
 Cc: zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] j4200 drive carriers

 John-Paul Drawneek wrote:


 the J series is far to new to be hitting ebay yet.

 Any alot of people will not be buying the J series for obvious reasons


 The obvious reason is that Sun cannot service random disk
 drives you buy from Fry's (or elsewhere). People who value data
 tend to value service contracts for disk drives.
 -- richard

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] is zpool export/import | faster than rsync or cp

2009-03-28 Thread Blake
zfs send/recv *is* faster (especially since b105) than rsync,
especially when you are dealing with lots of small files.  rsync has
to check each file, which can take a long time - zfs send/recv just
moves blocks.

2009/3/27 Ahmed Kamal email.ahmedka...@googlemail.com:
 ZFS replication basics at http://cuddletech.com/blog/pivot/entry.php?id=984
 Regards

 On Sat, Mar 28, 2009 at 1:57 AM, Harry Putnam rea...@newsguy.com wrote:

 [...]

 Harry wrote:
  Now I'm wondering if the export/import sub commands might not be a
  good bit faster.
 
 Ian Collins i...@ianshome.com answered:
  I think you are thinking of zfs send/receive.
 
  I've never done a direct comparison, but zfs send/receive would be my
  preferred way to move data between pools.

 Why is that?  I'm too new to know what all it encompasses (and a bit
 dense to boot)

 Fajar A. Nugraha fa...@fajar.net writes:

  On Sat, Mar 28, 2009 at 5:05 AM, Harry Putnam rea...@newsguy.com
  wrote:
  Now I'm wondering if the export/import sub commands might not be a
  good bit faster.
 
  I believe the greatest advantage of zfs send/receive over rsync is not
  about speed, but rather it's on zfs send -R, which would (from man
  page)
 
               Generate a replication stream  package,  which  will
               replicate  the specified filesystem, and all descen-
               dant file systems, up to the  named  snapshot.  When
               received, all properties, snapshots, descendent file
               systems, and clones are preserved.
 
  pretty much allows you to clone a complete pool preserving its
  structure.
  As usual, compressing the backup stream (whether rsync or zfs) might
  help reduce transfer time a lot. My favorite is lzop (since it's very
  fast), but gzip should work as well.
 

 Nice... good reasons it appears.


 Robert Milkowski mi...@task.gda.pl writes:

  Hello Harry,

 [...]

  As Ian pointed you want zfs send|receive and not import/export.
  For a first full copy zfs send not necessarily will be noticeably
  faster than rsync but it depends on data. If for example you have
  milions of small files zfs send could be much faster then rsync.
  But it shouldn't be slower in any case.
 
  zfs send|receive really shines when it comes to sending incremental
  changes.

 Now that would be something to make it stand out.  Can you tell me a
 bit more about that would work..I mean would you just keep receiving
 only changes at one end and how do they appear on the filesystem.

 There is a backup tool called `rsnapshot' that uses rsync but creates
 hard links to all unchanged files and moves only changes to changed
 files.  This is all put in a serial directory system and ends up
 taking a tiny fraction of the space that full backups would take, yet
 retains a way to get to unchanged files right in the same directory
 (the hard link).

 Is what your talking about similar in some way.

 =     *     =     *     =     *     =

 To all posters... many thanks for the input.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Growing a zpool mirror breaks on Adaptec 1205sa PCI

2009-03-28 Thread Blake
what's the output of 'fmadm faulty'?

On Sat, Mar 28, 2009 at 12:22 PM, Harry Putnam rea...@newsguy.com wrote:
 Harry Putnam rea...@newsguy.com writes:

 Once booted up I see the recurring message where I should see a login
 prompt (I'm setup to boot into console mode).

   ata_id_common Busy Status 0xfe error 0x0

 Repeated 4 times, then after maybe a 2-3 minute wait the regular login
 prompt appears.

 There is something more going on that is dragging the machine down so
 that typing at the prompt is delayed by huge pauses.

 I'm not really able to find out what it is yet but will post back
 later after a reboot.

 Those 4 lines continue to appear on ever boot now.. (still with old
 200gb sata drives out.

 Seems when the boot prompt does appear there is something dragging
 down the os making logging in really sluggish.

 It appears to be the fmd daemon running periodically and driving
 resource drain on cpu up to 99 percent.

 That seems to continue to occur for at least several cycles watching
 it in `top'.  They last maybe close to 1 minute then disappear for a
 while.  Several minutes at least.  The cylces seem to keep coming
 after being booted up for 10 minutes now.

 I'm not sure which logs to look in to see whats happing.  But earlier
 today these showed up in /var/adm/messages:

 Mar 28 09:38:25 zfs fmd: [ID 441519 daemon.error] SUNW-MSG-ID:
 ZFS-8000-D3, TYPE: Fault, VER: 1, SEVERITY: Major
 Mar 28 09:38:25 zfs IMPACT: Fault tolerance of the pool may be
 compromised.

 The console login continues to be nearly useless with the delays and
 pauses while typing.

 However ssh in and the terminal I get seem to be less effected, or
 even un-effected so I can do things in some kind or reasonable way.

 I'm really not sure what to do though.  I did
  `zpool destroy thatpool' on the pool that was on the sata drives.

 That appears to have worked, but didn't help with resource drain
 coming from `fmd'

 The tail of the log pointed to by svcs -l system/fmd shows:
 (/var/svc/log/system-fmd:default.log)
 [...]
 [ Mar 28 09:20:31 Enabled. ]
 [ Mar 28 09:25:11 Executing start method (/usr/lib/fm/fmd/fmd). ]
 [ Mar 28 09:28:24 Method start exited with status 0. ]
 [ Mar 28 09:53:50 Enabled. ]
 [ Mar 28 09:54:50 Executing start method (/usr/lib/fm/fmd/fmd). ]
 [ Mar 28 09:55:44 Method start exited with status 0. ]
 [ Mar 28 09:59:21 Stopping because service disabled. ]
 [ Mar 28 09:59:21 Executing stop method (:kill). ]
 [ Mar 28 10:01:09 Enabled. ]
 [ Mar 28 10:02:17 Executing start method (/usr/lib/fm/fmd/fmd). ]
 [ Mar 28 10:02:49 Method start exited with status 0. ]

 I have no idea is that is normal or what.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to increase rpool size in a VM?

2009-03-25 Thread Blake
You need to use 'installgrub' to get the right boot pits in place on
your new disk.

The manpage for installgrub is pretty helpful.

On Wed, Mar 25, 2009 at 6:18 PM, Bob Doolittle robert.doolit...@sun.com wrote:
 Hi,

 I have a build 109 system installed in a VM, and my rpool capacity is
 getting close to full.

 Since it's a VM, I can easily increase the size of the disk, or add another,
 larger disk to the VM.

 What's the easiest strategy for increasing my capacity?

 I tried adding a 2nd larger disk, did a zpool attach, waited for resilvering
 to complete, did a zpool detach of the 1st disk, but then it seemed it
 couldn't find my grub menu... I couldn't figure out a way to simply add a
 2nd disk to the rpool, it seems like it's limited to a single device.

 Suggestions?

 Please keep me on the reply list, I'm not subscribed to this list currently.

 Thanks,
  Bob

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zpools on USB zpool.cache zpool import

2009-03-24 Thread Blake
+1

On Mon, Mar 23, 2009 at 8:43 PM, Damon Atkins damon.atk...@yahoo.com.au wrote:

 PS it would be nice to have a zpool diskinfo devicepath reports  if the
 device belongs to a zpool imported or not, and all the details about any
 zpool it can find on the disk. e.g. file-systems (zdb is only for ZFS
 engineers says the man page). 'zpool import' needs an option to list the
 file systems of a pool which is not yet imported and its properties so you
 can have more information about it before importing it.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Size discrepancy (beyond expected amount?)

2009-03-20 Thread Blake
Replies inline (I really would recommend reading the whole ZFS Best
Practices guide a few times - many of your questions are answered in
that document):

On Fri, Mar 20, 2009 at 3:15 PM, Harry Putnam rea...@newsguy.com wrote:


 I didn't make it clear.  1 disk, the one with rpool on it is 60gb.
 The other 3 are 500GB.  Using a 500gb to mirror a 60gb would be
 something of a waste .. eh?
In the near-term, yes, but it would work.


 And is a mirror of rpool really important?  I assumed there would be
 some way to backup rpool in a format that could be written onto a new
 disk and booted in the event of rpool disk failure.  With the backup
 not kept on rpool disk.
You say you want a storage server you can forget about - that sounds
like zfs self-healing, which requires a mirror at least.


 Someone had suggested even creating an rpool mirror and putting the
 bootmanager bits on its mbr but then keeping it in storage instead of
 on the machine (freeing a controller port).
This is not a replacement for live redundancy.


 It has the downside of having to be re mirrored every now and then.
Actually, the point of a backup is to have a known-good copy of data
somewhere.  Re-mirroring would be a mistake, as it destroys your old
data state.


 But could be safe enough if nothing of great importance was kept on
 the rpool... Just OS and config changes, some BE's.  But nothing that
 couldn't be lost.

 Then in the event of disk failure... You'd have to just install the
 spare, boot it, and bring it up to date.

 Something kind of like what people do with ghost images on windows
 OS.

 Start with two mirrored pools of two disks each. In the future,
 you will be able to add two or more disks to your non-root pool.
 You can't do that with a RAIDZ pool.

 Well one thing there... if I use 5 500gb disks (no counting rpool disk
 - 6 total), by the time my raidz fills up, I'll need a whole new
 machine really since I'll be out of controller ports and its getting
 hard to find controllers that are not PCI express already. (My
 hardware is plain PCI only and even then the onboard sata is not
 recognized and I'm addding a PCI sata controller already)
If the hardware is old/partially supported/flaky, all the more reason
to use mirrors.  Any single disk from a mirror can be used standalone.
 Big disks are cheap: http://tinyurl.com/5tzguf


 Also some of the older data will have outlived its usefulness so what
 needs transferring to a new setup may not be really hard to
 accommodate or insurmountable.

 And finally, I'm 65 yrs old... Its kind of hard to imagine my wife and
 I filling up the nearly 2tb of spc the above mentioned raidz1 would
 afford before we go before the old grim reaper.

 Even with lots of pictures and video projects thrown in.

 I'm really thinking now to go to 5 500gb disks in raidz1, and one
 hotswap (Plus the rpool on 1 60gb disk). I would be clear out of both
 sata and IDE controller ports then, so I'm hoping I can add a hot swap
 by pulling one of the raidz disks long enough to add the
 hotswap... then take it back out and replace the missing raidz disk.
See the zfs docs for more about hot spares.  The 'hot' part means the
disk is in the chassis and spinning all the time, ready to replace a
failed drive automatically.  Not something easy work out the hardware
for in a situation like yours.  If you don't have room for yet another
disk in the chassis, you won't be able to use a hot spare.


 I could do this by getting 3 more 500gb disks 2 more for the raidz and
 1 for hotswap.  No other hardware would be needed. All the while
 assuming I can mix 3 500GB IDE and 2 500GB SATA with no problems.

 If you need to, you can even detach one side of the mirror
 of each pool. You can't do that with a RAIDZ pool. If you need
 larger pools you can replace all the disks in both pools with
 larger disks. You can do that with a RAIDZ pool, but more
 flexibility exists with mirrored pools.

 1. Yes, sensible.
 2. Saving space isn't always the best configuration.
 3. I don't know.
 4. Yes, with more disks, you can identify hot spares to
 be used in the case of a disk failure.

 Nice thanks (To Bob F as well).  And I'm not being hard headed about
 using a mirror config.  Its just that I have limited controller ports
 (4 ide 2 sata), limited budget, and kind of wanted to get this backup
 machine setup to where I could basically just leave it alone and let
 the backups run.

 On 3) Mixing IDE and SATA on same zpool
 I'd really like to hear from someone who has done that.
In my experience, zfs doesn't care what kind of block device you give it.


 About 4).. so if all controllers are already full with either a zpool
 or rpool.  Do you pull out one of the raidz1 disks to add a hotswap
 then remove the hotswap and put the pulled disk from the raidz back?

 If so, does that cause some kind of resilvering or does some other
 thing happen when a machine is booted with a raidz1 disk misssing, and
 then rebooted with it back in place?

Re: [zfs-discuss] Size discrepancy (beyond expected amount?)

2009-03-19 Thread Blake
IIRC, that's about right.  If you look at the zfs best practices wiki
(genunix.org I think?), there should be some space calculations linked
in there somewhere.

On Thu, Mar 19, 2009 at 6:50 PM, Harry Putnam rea...@newsguy.com wrote:
 I'm finally getting close to the setup I wanted, after quite a bit of
 experimentation and bugging these lists endlessly.

 So first, thanks for your tolerance and patience.

 My setup consists of 4 disks.  One holds the OS (rpool) and 3 more all
 the same model and brand, all 500gb.

 I've created a zpool in raidz1 configuration with:

  zpool create  zbk raidz1  c3d0 c4d0 c4d1

 No errors showed up and zpool status shows no problems with those
 three:
   pool: zbk
  state: ONLINE
  scrub: none requested
  config:

        NAME        STATE     READ WRITE CKSUM
        zbk         ONLINE       0     0     0
          raidz1    ONLINE       0     0     0
            c3d0    ONLINE       0     0     0
            c4d0    ONLINE       0     0     0
            c4d1    ONLINE       0     0     0


 However, I appear to have lost an awfull lot of space... even above
 what I expercted.

  df -h
 [...]
  zbk                   913G   26K  913G   1% /zbk

 It appears something like 1 entire disk is gobbled up by raidz1.

 The same disks configured in zpool with no raidz1  shows  1.4tb with df.

 I was under the impression raidz1 would take something like 20%.. but
 this is more like 33.33%.

 So, is this to be expected or is something wrong here?

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Size discrepancy (beyond expected amount?)

2009-03-19 Thread Blake
This verifies my guess:

http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide#RAID-Z_Configuration_Requirements_and_Recommendations

On Thu, Mar 19, 2009 at 6:57 PM, Blake blake.ir...@gmail.com wrote:
 IIRC, that's about right.  If you look at the zfs best practices wiki
 (genunix.org I think?), there should be some space calculations linked
 in there somewhere.

 On Thu, Mar 19, 2009 at 6:50 PM, Harry Putnam rea...@newsguy.com wrote:
 I'm finally getting close to the setup I wanted, after quite a bit of
 experimentation and bugging these lists endlessly.

 So first, thanks for your tolerance and patience.

 My setup consists of 4 disks.  One holds the OS (rpool) and 3 more all
 the same model and brand, all 500gb.

 I've created a zpool in raidz1 configuration with:

  zpool create  zbk raidz1  c3d0 c4d0 c4d1

 No errors showed up and zpool status shows no problems with those
 three:
   pool: zbk
  state: ONLINE
  scrub: none requested
  config:

        NAME        STATE     READ WRITE CKSUM
        zbk         ONLINE       0     0     0
          raidz1    ONLINE       0     0     0
            c3d0    ONLINE       0     0     0
            c4d0    ONLINE       0     0     0
            c4d1    ONLINE       0     0     0


 However, I appear to have lost an awfull lot of space... even above
 what I expercted.

  df -h
 [...]
  zbk                   913G   26K  913G   1% /zbk

 It appears something like 1 entire disk is gobbled up by raidz1.

 The same disks configured in zpool with no raidz1  shows  1.4tb with df.

 I was under the impression raidz1 would take something like 20%.. but
 this is more like 33.33%.

 So, is this to be expected or is something wrong here?

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Size discrepancy (beyond expected amount?)

2009-03-19 Thread Blake
I'd be careful about raidz unless you have either:

1 - automatic notification of failure set up using fmadm

2 - at least one hot spare

Because raidz is parity-based (it does some math-magic to give you
redundancy), replacing a disk that's failed can take a very long time
compared to mirror resilvering (the zfs term for rebuilding
redundancy).

You can get a nice 1000gb SATA drive on newegg or a similar site for
about $90 - well worth the extra money ($120) for the convenience of
mirroring.  Mirrors are probably also faster for any kind of video
playback (like the video projects you mention).

I use raidz2 with 2 hot spares at my company, yes, but only for data
warehousing.  User data (windows home dirs) I put on mirrors.



On Thu, Mar 19, 2009 at 10:00 PM, Harry Putnam rea...@newsguy.com wrote:
 Bob Friesenhahn bfrie...@simple.dallas.tx.us writes:

 On Thu, 19 Mar 2009, Harry Putnam wrote:
 I've created a zpool in raidz1 configuration with:

  zpool create  zbk raidz1  c3d0 c4d0 c4d1

 This is not a very useful configuration.  With this number of disks,
 it is best to use two of them to build a mirror, and save the other
 disk for something else (e.g. to build a mirrored root pool). Mirrors
 perform better, are more fault tolerant, and are easier to administer.

 Ok, I was going by comments on a site called Simons' blog that tells
 you how to set things up with zfs.  Of course it is just one guys
 opinion.

 Let me just say a couple of words about my intended usage.

 I'm building a home NAS server for my home lan using opensolaris and
 zfs.

 I will be backing up 4 Windows XP boxes, 2 of which are used primarily
 for processing video.  And eventully I'd be storing finished video
 projects too.  Often these projects run to 50gb or so.  But I don't do
 so many.  Maybe 9-12 in a yr.

 So I'd want to put ghosted images of each machine OS (several apiece)
 and running backups (incremental) that would stretch back a few weeks.
 And the projects mentioned above.  As well as what is becoming quite a
 large photo collection.

 Also 2 linux boxes will getting backed up there.

 I thought the kind of redundancy offered by raidz1 would be enough for
 my needs and would allow me to get a little more storage room out of
 my disks than mirrored setup.

 I suspect as well, that the access times in raidz1 or mirrored would
 be vastly higher than what the low end consumer grade NASs that are
 availabe offer.

 I did try one out.. a WD Worldbook (about $200 US) that advertises
 gigabit access but in use cannot even come close to what a 10/100
 connect can handle.

 So raidz1 would probably be adequate for me... I wouldn't be putting
 it to the test like a commercial operation might.

 You mentioned admistration was a bigger problem with raidz1, can you
 be more specific there?... I have really know idea what to expect in
 that regard with either technique.

 With five disks, raidz1 becomes useful.

 The three 500gb I have now are all one brand and model number and IDE ata.
 If I were to expand to 5, those 2 would need to be sata or else I'd
 also need to add a PCI IDE controller.

 With that in mind would it be problematic to make up the 5 by adding 2
 sata 500gb to the mix?

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Freezing OpenSolaris with ZFS

2009-03-15 Thread Blake
This sounds quite like the problems I've been having with a spotty
sata controller and/or motherboard.  See my thread from last week
about copying large amounts of data that forced a reboot.  Lots of
good info from engineers and users in that thread.



On Sun, Mar 15, 2009 at 1:17 PM, Markus Denhoff denh...@net-bite.net wrote:
 Hi there,

 we set up an OpenSolaris/ZFS based storage server with two zpools: rpool is
 a mirror for the operating system. tank is a raidz for data storage.

 The system is used to store large video files and has attached 12x1GB
 SATA-drives (2 mirrored for the system). Everytime large files are copied
 around the system hangs without apparent reason, 50% kernel CPU usage (so
 one core is occupied totally) and about 2GB of free RAM (8GB installed). On
 idle nothing crashes. Furthermore every scrub on tank hangs the system up
 below 1% finished. Neither the /var/adm/messages nor the /var/log/syslog
 file contains any errors or warnings. We limited the ZFS ARC cache to 4GB
 with an entry in /etc/system.

 Does anyone has an idea what's happening there and how to solve the problem?

 Below some outputs which may help.

 Thanks and greetings from germany,

 Markus Denhoff,
 Sebastian Friederichs

 # zpool status tank
  pool: tank
  state: ONLINE
  scrub: none requested
 config:

        NAME         STATE     READ WRITE CKSUM
        tank         ONLINE       0     0     0
          raidz1     ONLINE       0     0     0
            c6t2d0   ONLINE       0     0     0
            c6t3d0   ONLINE       0     0     0
            c6t4d0   ONLINE       0     0     0
            c6t5d0   ONLINE       0     0     0
            c6t6d0   ONLINE       0     0     0
            c6t7d0   ONLINE       0     0     0
            c6t8d0   ONLINE       0     0     0
            c6t9d0   ONLINE       0     0     0
            c6t10d0  ONLINE       0     0     0
            c6t11d0  ONLINE       0     0     0

 errors: No known data errors

 # zpool iostat
               capacity     operations    bandwidth
 pool         used  avail   read  write   read  write
 --  -  -  -  -  -  -
 rpool       37.8G   890G      3      2  94.7K  17.4K
 tank        2.03T  7.03T    112      0  4.62M    906
 --  -  -  -  -  -  -

 # zfs list
 NAME                       USED  AVAIL  REFER  MOUNTPOINT
 rpool                     39.8G   874G    72K  /rpool
 rpool/ROOT                35.7G   874G    18K  legacy
 rpool/ROOT/opensolaris    35.6G   874G  35.3G  /
 rpool/ROOT/opensolaris-1  89.9M   874G  2.47G  /tmp/tmp8CN5TR
 rpool/dump                2.00G   874G  2.00G  -
 rpool/export               172M   874G    19K  /export
 rpool/export/home          172M   874G    21K  /export/home
 rpool/swap                2.00G   876G    24K  -
 tank                      1.81T  6.17T  32.2K  /tank
 tank/data                 1.81T  6.17T  1.77T  /data
 tank/public-share         34.9K  6.17T  34.9K  /public-share
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS GSoC ideas page rough draft

2009-03-14 Thread Blake
I just thought of an enhancement to zfs that would be very helpful in
disaster recovery situations - having zfs cache device serial/model
numbers - the information we see in cfgadm -v.

I'm feeling the pain of this now as I try to figure out which disks on
my failed filer belonged to my raidz2 pool - zpool status tells me the
pool is faulted (I don't have enough working SATA ports to connect all
the drives from the pool), but doesn't tell me which individual
devices were in that pool (just the devids of the devices).


On Mon, Mar 9, 2009 at 9:30 AM, C. cbergst...@netsyncro.com wrote:

 Here's my rough draft of GSoC ideas

 http://www.osunix.org/docs/DOC-1022

 Also want to thank everyone for their feedback.

 Please keep in mind that for creating a stronger application we only have a
 few days.

 We still need to :

 1) Find more mentors.  (Please add your name to the doc or confirm via email
 and which idea you're most interested in)
 2) Add contacts from each organization that may be interested (OpenSolaris,
 FreeBSD...)
 3) Finalize the application, student checklist, mentor checklist and
 template
 4) Start to give ideas for very accurate project descriptions/details (We
 have some time for this)

 Thanks

 ./Christopher

 ---
 Community driven OpenSolaris Technology - http://www.osunix.org
 blog: http://www.codestrom.com

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS GSoC ideas page rough draft

2009-03-14 Thread Blake
That's excellent information Richard.

I have just exported the faulted pool (the disks aren't attached), but
what you are suggesting is verified by looking at the cache as it is -
I can see this information for my rpool mirror disks.

One other note - I see that using the old prtvtoc | fmthard trick to
create my 2nd mirror disk for the rpool created a generic label, and
it looks like zdb only stores the disk label:

# format

AVAILABLE DISK SELECTIONS:
   0. c4t0d0 ATA-ST3250310NS-SN04 cyl 30398 alt 2 hd 255 sec 63
  /p...@0,0/pci15d9,1...@5/d...@0,0
   1. c4t1d0 DEFAULT cyl 30398 alt 2 hd 255 sec 63
  /p...@0,0/pci15d9,1...@5/d...@1,0

# zdb

...

children[0]
type='disk'
id=0
guid=9660743866801172401
path='/dev/dsk/c4t0d0s0'
devid='id1,s...@f00439638498493aa00037d740001/a'
phys_path='/p...@0,0/pci15d9,1...@5/d...@0,0:a'
whole_disk=0
DTL=101
children[1]
type='disk'
id=1
guid=5364095963942871931
path='/dev/dsk/c4t1d0s0'
devid='id1,s...@ast3250310ns=9sf06bc8/a'
phys_path='/p...@0,0/pci15d9,1...@5/d...@1,0:a'
whole_disk=0
DTL=99


Cfgadm, however, returns the full serial number and model number for
each disk.  So I think storing the information cfgadm collects in
zpool.cache would be preferable to reading disk labels.




On Sat, Mar 14, 2009 at 12:47 PM, Richard Elling
richard.ell...@gmail.com wrote:
 Toby Thain wrote:

 On 14-Mar-09, at 12:09 PM, Blake wrote:

 I just thought of an enhancement to zfs that would be very helpful in
 disaster recovery situations - having zfs cache device serial/model
 numbers - the information we see in cfgadm -v.

 +1  I haven't needed this but it sounds very sensible. I can imagine it
 could help a lot in some drive replacement situations.

 What do your devids look like?  Mine are something like:
       devid='id1,s...@sata_st3500320as_9qm3fwft/g'

 Where ST3500320AS is the model number and 9QM3FWFT is the serial
 number.  These are stored verbatim in the zpool.cache.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] What to do with a disk partition

2009-03-14 Thread Blake
I think you will be helped by looking at this document:

http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#ZFS_Root_Pool_Recommendations_and_Requirements

It addresses many of your questions.

I think the easiest way to back up your OS might be to attach a disk
to the rpool as a mirror, use 'installgrub' to get the grub boot
blocks onto the new mirror disk, then detach this disk and put it in
storage.

There are also more advanced backup methods covered in the document above.



On Sat, Mar 14, 2009 at 11:51 PM, Harry Putnam rea...@newsguy.com wrote:
 I'm still learning and haven't setup anything I can't destroy if it
 can be better put together some other way.  I would like to save any
 updates and customizations so far if possible.  But that isn't
 critical either.

 Originally I installed osol-11 101b (now at 108) on a 60gb disk
 partitioned 25/45 with the OS on the 25.  These were just fdisk
 partitions, no slices.

 I've since installed 2 more drives and have 2 more to install once I
 get a working sata controller.

 The purpose of this server once I quit tinkering so much and decide I know
 enough to put it into service is just a home backup server.

 So currently I have 3 disks to work with.
 Disk 0 (60gb) as mentioned holds the OS on a 25gb fdisk partition c3d0p1.
     And another zpool on the remaining 45gb at c3d0p2
 Disk 1 (250gb) 1 fdisk partition (full disk)
 Disk 2 ( 500gb 1 fdisk partition (full disk)

 I've done nothing so far with the last 2, just did get them installed
 but I had planned to put them into a raidz1 configuration, which would
 give me 750GB in that zpool, and use them to backup 4 windows machines
 and one linux machine.  That is, assuming that isn't a nitwit thing to
 do and some other configuration would be better.
 Any opinions/suggestions .. welcome.

 I'd like some opinions on how best to utilize the remaining 45gb
 partition on disk 0.  The one with the OS on the other partition.

 I originally thought I'd use it to backup the OS itself, but after
 listening here a while I wonder if that would be better served if I
 had made rpool encompass that whole disk.  Instead of having rpool as
 the OS of 25gb on c3d0p1 and rpool.bk as a zpool on the other 45gb
 partition at c3d0p2.

 I guess I'm asking if its wise to have more than 1 zpool on a disk?
 Especially if one of them holds the OS.

 Or is it wiser to reinstall on a single partition of the whole disk?

 How do people backup there OS, in general.  Just keep a herd of
 snapshots letting the oldest be destroyed over time...?

 Is Keeping those on the same zpool and disk as the OS wise?

 Its a home setup so I don't need to get too fancy like having backups
 offsite or whatever.. nobody elses data will be at risk. Of course an
 irate wife with lost data might not be too pretty of picture...

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Blake Irvin
This is really great information, though most of the controllers  
mentioned aren't on the OpenSolaris HCL.  Seems like that should be  
corrected :)


My thanks to the community for their support.

On Mar 12, 2009, at 10:42 PM, James C. McPherson james.mcpher...@sun.com 
 wrote:



On Thu, 12 Mar 2009 22:24:12 -0400
Miles Nordin car...@ivy.net wrote:


wm == Will Murnane will.murn...@gmail.com writes:



* SR = Software RAID IT = Integrate. Target mode. IR mode
is not supported.

   wm Integrated target mode lets you export some storage attached
   wm to the host system (through another adapter, presumably) as a
   wm storage device.  IR mode is almost certainly Internal RAID,
   wm which that card doesn't have support for.

no, the supermicro page for AOC-USAS-L8i does claim support for all
three, and supermicro has an ``IR driver'' available for download for
Linux and Windows, or at least a link to one.

I'm trying to figure out what's involved in determining and switching
modes, why you'd want to switch them, what cards support which modes,
which solaris drivers support which modes, u.s.w.

The answer may be very simple, like ``the driver supports only IR.
Most cards support IR, and cards that don't support IR won't work.   
IR

can run in single-LUN mode.  Some IR cards support RAID5, others
support only RAID 0, 1, 10.''  Or it could be ``the driver supports
only SR.  The driver is what determines the mode, and it does this by
loading firmware into the card, and the first step in initializing  
the

card is always for the driver to load in a firmware blob.  All
currently-produced cards support SR.''  so...actually, now that I say
it, I guess the answer cannot be very simple.  It's going to have to
be a little complicated.
Anyway, I can guess, too.  I was hoping someone would know for sure
off-hand.



Hi Miles,
the mpt(7D) driver supports that card. mpt(7D) supports both
IT and IR firmware variants. You can find out the specifics
for what RAID volume levels are supported by reading the
raidctl(1M) manpage. I don't think you can switch between IT
and IR firmware, but not having needed to know this before,
I haven't tried it.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcphttp://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Blake
I start the cp, and then, with prstat -a, watch the cpu load for the
cp process climb to 25% on a 4-core machine.

Load, measured for example with 'uptime', climbs steadily until the reboot.

Note that the machine does not dump properly, panic or hang - rather,
it reboots.

I attached a screenshot earlier in this thread of the little bit of
error message I could see on the console.  The machine is trying to
dump to the dump zvol, but fails to do so.  Only sometimes do I see an
error on the machine's local console - mos times, it simply reboots.



On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:
 Hm -

 Crashes, or hangs? Moreover - how do you know a CPU is pegged?

 Seems like we could do a little more discovery on what the actual problem
 here is, as I can read it about 4 different ways.

 By this last piece of information, I'm guessing the system does not crash,
 but goes really really slow??

 Crash == panic == we see stack dump on console and try to take a dump
 hang == nothing works == no response - might be worth looking at mdb -K
        or booting with a -k on the boot line.

 So - are we crashing, hanging, or something different?

 It might simply be that you are eating up all your memory, and your physical
 backing storage is taking a while to catch up?

 Nathan.

 Blake wrote:

 My dump device is already on a different controller - the motherboards
 built-in nVidia SATA controller.

 The raidz2 vdev is the one I'm having trouble with (copying the same
 files to the mirrored rpool on the nVidia controller work nicely).  I
 do notice that, when using cp to copy the files to the raidz2 pool,
 load on the machine climbs steadily until the crash, and one proc core
 pegs at 100%.

 Frustrating, yes.

 On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
 maidakalexand...@johndeere.com wrote:

 If you're having issues with a disk contoller or disk IO driver its
 highly likely that a savecore to disk after the panic will fail.  I'm not
 sure how to work around this, maybe a dedicated dump device not on a
 controller that uses a different driver then the one that you're having
 issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore to
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out of
 space on the dump device:

 r...@host:~# dumpadm
     Dump content: kernel pages
      Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the system
 load was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
 richard.ell...@gmail.com wrote:

 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.


 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for
 swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 --
 //
 // Nathan Kroenert              nathan.kroen...@sun.com         //
 // Systems Engineer             Phone:  +61 3 9869-6255         //
 // Sun Microsystems             Fax:    +61 3 9869-6288         //
 // Level 7, 476 St. Kilda Road  Mobile: 0419 305 456            //
 // Melbourne 3004   Victoria    Australia                       //
 //

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Blake
So, if I boot with the -k boot flags (to load the kernel debugger?)
what do I need to look for?  I'm no expert at kernel debugging.

I think this is a pci error judging by the console output, or at least
is i/o related...

thanks for your feedback,
Blake

On Thu, Mar 12, 2009 at 2:18 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:
 definitely time to bust out some mdb -K or boot -k and see what it's moaning
 about.

 I did not see the screenshot earlier... sorry about that.

 Nathan.

 Blake wrote:

 I start the cp, and then, with prstat -a, watch the cpu load for the
 cp process climb to 25% on a 4-core machine.

 Load, measured for example with 'uptime', climbs steadily until the
 reboot.

 Note that the machine does not dump properly, panic or hang - rather,
 it reboots.

 I attached a screenshot earlier in this thread of the little bit of
 error message I could see on the console.  The machine is trying to
 dump to the dump zvol, but fails to do so.  Only sometimes do I see an
 error on the machine's local console - mos times, it simply reboots.



 On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
 nathan.kroen...@sun.com wrote:

 Hm -

 Crashes, or hangs? Moreover - how do you know a CPU is pegged?

 Seems like we could do a little more discovery on what the actual problem
 here is, as I can read it about 4 different ways.

 By this last piece of information, I'm guessing the system does not
 crash,
 but goes really really slow??

 Crash == panic == we see stack dump on console and try to take a dump
 hang == nothing works == no response - might be worth looking at mdb -K
       or booting with a -k on the boot line.

 So - are we crashing, hanging, or something different?

 It might simply be that you are eating up all your memory, and your
 physical
 backing storage is taking a while to catch up?

 Nathan.

 Blake wrote:

 My dump device is already on a different controller - the motherboards
 built-in nVidia SATA controller.

 The raidz2 vdev is the one I'm having trouble with (copying the same
 files to the mirrored rpool on the nVidia controller work nicely).  I
 do notice that, when using cp to copy the files to the raidz2 pool,
 load on the machine climbs steadily until the crash, and one proc core
 pegs at 100%.

 Frustrating, yes.

 On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
 maidakalexand...@johndeere.com wrote:

 If you're having issues with a disk contoller or disk IO driver its
 highly likely that a savecore to disk after the panic will fail.  I'm
 not
 sure how to work around this, maybe a dedicated dump device not on a
 controller that uses a different driver then the one that you're having
 issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore
 to
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out
 of
 space on the dump device:

 r...@host:~# dumpadm
    Dump content: kernel pages
     Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the
 system
 load was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
 richard.ell...@gmail.com wrote:

 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com
 wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.


 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for
 swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 --
 //
 // Nathan Kroenert              nathan.kroen...@sun.com         //
 // Systems Engineer             Phone:  +61 3 9869-6255         //
 // Sun Microsystems             Fax:    +61 3 9869-6288         //
 // Level 7, 476 St. Kilda Road  Mobile: 0419 305 456            //
 // Melbourne 3004   Victoria    Australia

Re: [zfs-discuss] User quota design discussion..

2009-03-12 Thread Blake
That is pretty freaking cool.

On Thu, Mar 12, 2009 at 11:38 AM, Eric Schrock eric.schr...@sun.com wrote:
 Note that:

 6501037 want user/group quotas on ZFS

 Is already committed to be fixed in build 113 (i.e. in the next month).

 - Eric

 On Thu, Mar 12, 2009 at 12:04:04PM +0900, Jorgen Lundman wrote:

 In the style of a discussion over a beverage, and talking about
 user-quotas on ZFS, I recently pondered a design for implementing user
 quotas on ZFS after having far too little sleep.

 It is probably nothing new, but I would be curious what you experts
 think of the feasibility of implementing such a system and/or whether or
 not it would even realistically work.

 I'm not suggesting that someone should do the work, or even that I will,
 but rather in the interest of chatting about it.

 Feel free to ridicule me as required! :)

 Thoughts:

 Here at work we would like to have user quotas based on uid (and
 presumably gid) to be able to fully replace the NetApps we run. Current
 ZFS are not good enough for our situation. We simply can not mount
 500,000 file-systems on all the NFS clients. Nor do all servers we run
 support mirror-mounts. Nor do auto-mount see newly created directories
 without a full remount.

 Current UFS-style-user-quotas are very exact. To the byte even. We do
 not need this precision. If a user has 50MB of quota, and they are able
 to reach 51MB usage, then that is acceptable to us. Especially since
 they have to go under 50MB to be able to write new data, anyway.

 Instead of having complicated code in the kernel layer, slowing down the
 file-system with locking and semaphores (and perhaps avoiding learning
 indepth ZFS code?), I was wondering if a more simplistic setup could be
 designed, that would still be acceptable. I will use the word
 'acceptable' a lot. Sorry.

 My thoughts are that the ZFS file-system will simply write a
 'transaction log' on a pipe. By transaction log I mean uid, gid and
 'byte count changed'. And by pipe I don't necessarily mean pipe(2), but
 it could be a fifo, pipe or socket. But currently I'm thinking
 '/dev/quota' style.

 User-land will then have a daemon, whether or not it is one daemon per
 file-system or really just one daemon does not matter. This process will
 open '/dev/quota' and empty the transaction log entries constantly. Take
 the uid,gid entries and update the byte-count in its database. How we
 store this database is up to us, but since it is in user-land it should
 have more flexibility, and is not as critical to be fast as it would
 have to be in kernel.

 The daemon process can also grow in number of threads as demand increases.

 Once a user's quota reaches the limit (note here that /the/ call to
 write() that goes over the limit will succeed, and probably a couple
 more after. This is acceptable) the process will blacklist the uid in
 kernel. Future calls to creat/open(CREAT)/write/(insert list of calls)
 will be denied. Naturally calls to unlink/read etc should still succeed.
 If the uid goes under the limit, the uid black-listing will be removed.

 If the user-land process crashes or dies, for whatever reason, the
 buffer of the pipe will grow in the kernel. If the daemon is restarted
 sufficiently quickly, all is well, it merely needs to catch up. If the
 pipe does ever get full and items have to be discarded, a full-scan will
 be required of the file-system. Since even with UFS quotas we need to
 occasionally run 'quotacheck', it would seem this too, is acceptable (if
 undesirable).

 If you have no daemon process running at all, you have no quotas at all.
 But the same can be said about quite a few daemons. The administrators
 need to adjust their usage.

 I can see a complication with doing a rescan. How could this be done
 efficiently? I don't know if there is a neat way to make this happen
 internally to ZFS, but from a user-land only point of view, perhaps a
 snapshot could be created (synchronised with the /dev/quota pipe
 reading?) and start a scan on the snapshot, while still processing
 kernel log. Once the scan is complete, merge the two sets.

 Advantages are that only small hooks are required in ZFS. The byte
 updates, and the blacklist with checks for being blacklisted.

 Disadvantages are that it is loss of precision, and possibly slower
 rescans? Sanity?

 But I do not really know the internals of ZFS, so I might be completely
 wrong, and everyone is laughing already.

 Discuss?

 Lund

 --
 Jorgen Lundman       | lund...@lundman.net
 Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
 Shibuya-ku, Tokyo    | +81 (0)90-5578-8500          (cell)
 Japan                | +81 (0)3 -3375-1767          (home)
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 --
 Eric Schrock, Fishworks                        http://blogs.sun.com/eschrock
 ___
 

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Blake
I've managed to get the data transfer to work by rearranging my disks
so that all of them sit on the integrated SATA controller.

So, I feel pretty certain that this is either an issue with the
Supermicro aoc-sat2-mv8 card, or with PCI-X on the motherboard (though
I would think that the integrated SATA would also be using the PCI
bus?).

The motherboard, for those interested, is an HD8ME-2 (not, I now find
after buying this box from Silicon Mechanics, a board that's on the
Solaris HCL...)

http://www.supermicro.com/Aplus/motherboard/Opteron2000/MCP55/h8dme-2.cfm

So I'm not considering one of LSI's HBA's - what do list members think
about this device:

http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm



On Thu, Mar 12, 2009 at 2:18 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:
 definitely time to bust out some mdb -K or boot -k and see what it's moaning
 about.

 I did not see the screenshot earlier... sorry about that.

 Nathan.

 Blake wrote:

 I start the cp, and then, with prstat -a, watch the cpu load for the
 cp process climb to 25% on a 4-core machine.

 Load, measured for example with 'uptime', climbs steadily until the
 reboot.

 Note that the machine does not dump properly, panic or hang - rather,
 it reboots.

 I attached a screenshot earlier in this thread of the little bit of
 error message I could see on the console.  The machine is trying to
 dump to the dump zvol, but fails to do so.  Only sometimes do I see an
 error on the machine's local console - mos times, it simply reboots.



 On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
 nathan.kroen...@sun.com wrote:

 Hm -

 Crashes, or hangs? Moreover - how do you know a CPU is pegged?

 Seems like we could do a little more discovery on what the actual problem
 here is, as I can read it about 4 different ways.

 By this last piece of information, I'm guessing the system does not
 crash,
 but goes really really slow??

 Crash == panic == we see stack dump on console and try to take a dump
 hang == nothing works == no response - might be worth looking at mdb -K
       or booting with a -k on the boot line.

 So - are we crashing, hanging, or something different?

 It might simply be that you are eating up all your memory, and your
 physical
 backing storage is taking a while to catch up?

 Nathan.

 Blake wrote:

 My dump device is already on a different controller - the motherboards
 built-in nVidia SATA controller.

 The raidz2 vdev is the one I'm having trouble with (copying the same
 files to the mirrored rpool on the nVidia controller work nicely).  I
 do notice that, when using cp to copy the files to the raidz2 pool,
 load on the machine climbs steadily until the crash, and one proc core
 pegs at 100%.

 Frustrating, yes.

 On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
 maidakalexand...@johndeere.com wrote:

 If you're having issues with a disk contoller or disk IO driver its
 highly likely that a savecore to disk after the panic will fail.  I'm
 not
 sure how to work around this, maybe a dedicated dump device not on a
 controller that uses a different driver then the one that you're having
 issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore
 to
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out
 of
 space on the dump device:

 r...@host:~# dumpadm
    Dump content: kernel pages
     Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the
 system
 load was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
 richard.ell...@gmail.com wrote:

 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com
 wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.


 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for
 swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs

[zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I have a H8DM8-2 motherboard with a pair of AOC-SAT2-MV8 SATA
controller cards in a 16-disk Supermicro chassis.

I'm running OpenSolaris 2008.11, and the machine performs very well
unless I start to copy a large amount of data to the ZFS (software
raid) array that's on the Supermicro SATA controllers.  If I do this,
the machine inevitably reboots.

What can I do to troubleshoot?  The BIOS of the motherboard and the
SATA card firmware are fully updated.  I'm running the latest stable
OpenSolaris, and see nothing amiss in the system logs when this
happens.  I've enabled savecore and debug-level syslog, but am getting
no indicators from Solaris as to what's wrong.

Interestingly, I can push the same amount of data to the mirror boot
disks, which are on the board's built-in nVidia SATA controller
without issue.

The vdev I'm pushing to is a 5-disk raidz2 with 2 hot spares.

Help!  :)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:
 The copy operation will make all the disks start seeking at the same time and
 will make your CPU activity jump to a significant percentage to compute the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above theory
 is unlikely because spinning up consumes much more power than when seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Export ZFS via ISCSI to Linux - Is it stable for production use now?

2009-03-11 Thread Blake
I blogged this a while ago:

http://blog.clockworm.com/2007/10/connecting-linux-centos-5-to-solaris.html

On Wed, Mar 11, 2009 at 1:02 PM, howard chen howac...@gmail.com wrote:
 Hello,

 On Wed, Mar 11, 2009 at 10:20 PM, Darren J Moffat
 darr...@opensolaris.org wrote:

 1. Is this setup suitable for mission critical use now?

 Yes, why wouldn't it be ?


 Because I just wonder why some other people are using zfs/fuse on Linux, e.g.
 http://www.drwetter.org/blog/zfs_under_linux.en.html
 http://www.wizy.org/wiki/ZFS_on_FUSE


 Also seems hard to find a complete tutorial on accessing ZFS from
 Linux using ISCSI. This should be attractive, isn't?

 So I don't know if it is experimental.

 Howard
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
fmdump is not helping much:

r...@host:~# fmdump -eV
TIME   CLASS
fmdump: /var/fm/fmd/errlog is empty


comparing that screenshot to the output of cfgadm is interesting -
looks like the controller(s):

r...@host:~# cfgadm -v
Ap_Id  Receptacle   Occupant Condition  Information
When Type Busy Phys_Id
sata4/0::dsk/c4t0d0connectedconfigured   ok
Mod: ST3250310NS FRev: SN06 SN: 9SF06CZZ
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5:0
sata4/1::dsk/c4t1d0connectedconfigured   ok
Mod: ST3250310NS FRev: SN06 SN: 9SF06BC8
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5:1
sata5/0emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,1:0
sata5/1::dsk/c7t1d0connectedconfigured   ok
Mod: WDC WD10EACS-00D6B0 FRev: 01.01A01 SN: WD-WCAU40244615
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5,1:1
sata6/0emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,2:0
sata6/1emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,2:1
sata7/0emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:0
sata7/1emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:1
sata7/2::dsk/c5t2d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0376631
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:2
sata7/3::dsk/c5t3d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0350798
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:3
sata7/4::dsk/c5t4d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0403574
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:4
sata7/5::dsk/c5t5d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0312592
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:5
sata7/6::dsk/c5t6d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0399779
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:6
sata7/7::dsk/c5t7d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0441660
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:7
sata8/0::dsk/c6t0d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT1000344
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:0
sata8/1emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:1
sata8/2emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:2
sata8/3emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:3
sata8/4emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:4
sata8/5emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:5
sata8/6emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:6
sata8/7emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:7


On Wed, Mar 11, 2009 at 2:40 PM, Blake blake.ir...@gmail.com wrote:
 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:
 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:
 The copy operation will make all the disks

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I think that TMC Research is the company that designed the
Supermicro-branded controller card that has the Marvell SATA
controller chip on it.  Googling around I see connections between
Supermicro and TMC.

This is the card:

http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:
 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 


 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
Could the problem be related to this bug:

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

I'm testing setting the maximum payload size as a workaround, as noted
in the bug notes.



On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:
 I think that TMC Research is the company that designed the
 Supermicro-branded controller card that has the Marvell SATA
 controller chip on it.  Googling around I see connections between
 Supermicro and TMC.

 This is the card:

 http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

 On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:
 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 


 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
Any chance this could be the motherboard?  I suspect the controller.

The boot disks are on the built-in nVidia controller.

On Wed, Mar 11, 2009 at 3:41 PM, Remco Lengers re...@lengers.com wrote:
 - Upgrade FW of controller to highest or known working level
I think I have the latest controller firmware.

 - Upgrade driver or OS level.
I'm going to try to go from 101b to 108 or whatever the current dev release is.

 - Try another controller (may be its broken and barfs under stress ?)
In the works.

 - Analyze the crash dump (if any is saved)
Crash dump is not saving properly.

 - It may be its a know Solaris or driver bug and somebody has heard of it
 before.
Any takers on this?  :)


 hth,
Thanks!


 ..Remco

 Blake wrote:

 Could the problem be related to this bug:

 http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

 I'm testing setting the maximum payload size as a workaround, as noted
 in the bug notes.



 On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:

 I think that TMC Research is the company that designed the
 Supermicro-branded controller card that has the Marvell SATA
 controller chip on it.  Googling around I see connections between
 Supermicro and TMC.

 This is the card:

 http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

 On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:

 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com
 wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to
 compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your
 PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 



 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I guess I didn't make it clear that I had already tried using savecore
to retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out
of space on the dump device:

r...@host:~# dumpadm
  Dump content: kernel pages
   Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated)
Savecore directory: /var/crash/host
  Savecore enabled: yes

I was using the -L option only to try to get some idea of why the
system load was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
richard.ell...@gmail.com wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the core
 will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:
 If you're having issues with a disk contoller or disk IO driver its highly 
 likely that a savecore to disk after the panic will fail.  I'm not sure how 
 to work around this, maybe a dedicated dump device not on a controller that 
 uses a different driver then the one that you're having issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org 
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore to 
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out of 
 space on the dump device:

 r...@host:~# dumpadm
      Dump content: kernel pages
       Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore 
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the system load 
 was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
 wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Is there a limit to snapshotting?

2009-03-08 Thread Blake
I think it's filesystems, not snapshots, that take a long time to
enumerate.  (If I'm wrong, somebody correct me :)



On Sun, Mar 8, 2009 at 10:10 PM, mike mike...@gmail.com wrote:
 I do a daily snapshot of two filesystems, and over the past few months
 it's obviously grown to a bunch.

 zfs list shows me all of those.

 I can change it to use the -t flag to not show them, so that's good.
 However, I'm worried about boot times and other things.

 Will it get to a point with 1000's of snapshots that it takes a long
 time to boot, or do any sort of sync or scrub activities?

 Thanks :)
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] large file copy bug?

2009-03-06 Thread Blake
I have savecore enabled, but it doesn't look like the machine is
dumping core as it should - that is, I don't think it's a panic - I
suspect interrupt handling.

Speaking of which, does OpenSolaris support Plug'n'Play IRQ assignment?



On Thu, Mar 5, 2009 at 3:12 PM, Mark J Musante mmusa...@east.sun.com wrote:
 On Thu, 5 Mar 2009, Blake wrote:

 I had a 2008.11 machine crash while moving a 700gb file from one machine
 to another using cp.  I looked for an existing bug for this, but found
 nothing.

 Has anyone else seen behavior like this?  I wanted to check before filing
 a bug.

 Have you got a copy of the stack dump?  That would make it easier to track
 down.


 Regards,
 markm

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] large file copy bug?

2009-03-06 Thread Blake
I have savecore enabled, but nothing in /var/crash:

r...@filer:~# savecore -v
savecore: dump already processed
r...@filer:~# ls /var/crash/filer/
r...@filer:~#



On Fri, Mar 6, 2009 at 4:21 PM, Mark J Musante mmusa...@east.sun.com wrote:
 On Fri, 6 Mar 2009, Blake wrote:

 I have savecore enabled, but it doesn't look like the machine is dumping
 core as it should - that is, I don't think it's a panic - I suspect
 interrupt handling.

 Then when you say you had a machine crash, what did you mean?

 Did you look in /var/crash/* to see if there's something there?  If not,
 it's possible your dump zvol just needs to have its core retrieved, which
 means you can just run 'savecore' at any time and get it.


 Regards,
 markm

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] large file copy bug?

2009-03-06 Thread Blake
These are fair questions, answered inline below :)

On Fri, Mar 6, 2009 at 4:45 PM, Mark J Musante mmusa...@east.sun.com wrote:
 On Fri, 6 Mar 2009, Blake wrote:
 OK, just to ask the dumb questions: is dumpadm configured for
 /var/crash/filer?  Is the dump zvol big enough?  How do you know the whole
 machine crashed instead of, for example, the X server just coincidentally
 died?
r...@filer:~# dumpadm
  Dump content: kernel pages
   Dump device: /dev/zvol/dsk/rpool/dump (dedicated)
Savecore directory: /var/crash/filer
  Savecore enabled: yes

I walked over to the console to see the machine rebooting.  gdm is
disabled on this box.

One point of possible interest is that the target fs is on a zpool
with 'compression=6' enabled - that is, gzip level 6 compression.

The machine has 4gb ram, which means the 2gb dump zvol should be big
enough, IIRC.


 I'm not trying to be difficult, just trying to narrow down where the problem
 is.  If you file a bug that just says 'cp killed my machine', it's likely to
 be marked incomplete until there's a reliable way of reproducing it.


 Regards,
 markm

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-05 Thread Blake
How I do recursive, selective snapshot destroys:

http://blog.clockworm.com/2008/03/remove-old-zfs-snapshots.html





 Saturday, February 28, 2009, 10:14:20 PM, you wrote:

 TW I would really add : make insane zfs destroy -r| poolname  as
 TW harmless as zpool destroy poolname (recoverable)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] large file copy bug?

2009-03-05 Thread Blake
I had a 2008.11 machine crash while moving a 700gb file from one
machine to another using cp.  I looked for an existing bug for this,
but found nothing.

Has anyone else seen behavior like this?  I wanted to check before filing a bug.

cheers,
Blake
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Blake
When I go here:
http://opensolaris.org/os/project/isns/bui

I get an error.  Where are you getting BUI from?


On Tue, Mar 3, 2009 at 5:16 PM, Richard Elling richard.ell...@gmail.comwrote:


 FWIW, I just took at look at the BUI in b108 and it seems to have
 garnered some love since the last time I looked at it (a year ago?)
 I encourage folks to take a fresh look at it.
   https://localhost:6789

 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Blake
That's what I thought you meant, and I got excited thinking that you were
talking about OpenSolaris :)
I'll see about getting the new packages and trying them out.



On Tue, Mar 3, 2009 at 8:36 PM, Richard Elling richard.ell...@gmail.comwrote:

 Blake wrote:

 When I go here:

 http://opensolaris.org/os/project/isns/bui

 I get an error. Â Where are you getting BUI from?


 The BUI is in webconsole which is available on your local machine at
 port 6789
   https://localhost:6798

 If you want to access it remotely, you'll need to change the configuration
 as documented
   http://docs.sun.com/app/docs/doc/817-1985/gdhgt?a=view

 -- richard


 On Tue, Mar 3, 2009 at 5:16 PM, Richard Elling 
 richard.ell...@gmail.commailto:
 richard.ell...@gmail.com wrote:


FWIW, I just took at look at the BUI in b108 and it seems to have
garnered some love since the last time I looked at it (a year ago?)
I encourage folks to take a fresh look at it.
  https://localhost:6789

-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS volume corrupted?

2009-03-02 Thread Blake
It looks like you only have one physical device in this pool.  Is that correct?



On Mon, Mar 2, 2009 at 9:01 AM, Lars-Gunnar Persson
lars-gunnar.pers...@nersc.no wrote:
 Hey to everyone on this mailing list (since this is my first post)!

 We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after some system
 work this weekend we have a problem with only one ZFS volume.

 We have a pool called /Data with many file systems and two volumes. The
 status of my zpool is:

 -bash-3.00$ zpool status
  pool: Data
  state: ONLINE
  scrub: scrub in progress, 5.99% done, 13h38m to go
 config:

        NAME                     STATE     READ WRITE CKSUM
        Data                     ONLINE       0     0     0
          c4t5000402001FC442Cd0  ONLINE       0     0     0

 errors: No known data errors


 Yesterday I started the scrub process because I read that was a smart thing
 to do after a zpool export and zpool import procedure. I did this because I
 wanted to move the zpool to another OS installation but changed my mind and
 did a zpool import on the same OS as I did an export.

 After checking as much information as I could find on the web, I was advised
 to to run the zpool scrub after an import.

 Well, the problem now is that one volume in this zpool is not working. I've
 shared it via iscsi to a Linux host (all of this was working on Friday). The
 Linux host reports that it can't find a partition table. Here is the log
 from the Linux host:

 Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr sectors
 (268435 MB)
 Mar  2 11:09:36 eva kernel: SCSI device sdb: drive cache: write through
 Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr sectors
 (268435 MB)
 Mar  2 11:09:37 eva kernel: SCSI device sdb: drive cache: write through
 Mar  2 11:09:37 eva kernel:  sdb: unknown partition table
 Mar  2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28, channel 0, id
 0, lun 0


 So I checked the status on my Solaris server and I found this information a
 bit strange;:

 -bash-3.00$ zfs list Data/subversion1
 NAME               USED  AVAIL  REFER  MOUNTPOINT
 Data/subversion1  22.5K   519G  22.5K  -

 How  can it bed 519GB available on a volume that is 250GB in size? Here are
 more details:

 -bash-3.00$ zfs get all Data/subversion1
 NAME              PROPERTY       VALUE                  SOURCE
 Data/subversion1  type           volume                 -
 Data/subversion1  creation       Wed Apr  2  9:06 2008  -
 Data/subversion1  used           22.5K                  -
 Data/subversion1  available      519G                   -
 Data/subversion1  referenced     22.5K                  -
 Data/subversion1  compressratio  1.00x                  -
 Data/subversion1  reservation    250G                   local
 Data/subversion1  volsize        250G                   -
 Data/subversion1  volblocksize   8K                     -
 Data/subversion1  checksum       on                     default
 Data/subversion1  compression    off                    default
 Data/subversion1  readonly       off                    default
 Data/subversion1  shareiscsi     off                    local


 Will this be fixed after the scrub process is finished tomorrow or is this
 volume lost forever?

 Hoping for some quick answers as the data is quite important for us.

 Regards,

 Lars-Gunnar Persson

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Virutal zfs server vs hardware zfs server

2009-03-02 Thread Blake
yes, most nvidia hardware will give you much better performance on
OpenSolaris (provided the card is fairly recent)

On Mon, Mar 2, 2009 at 6:18 AM, Juergen Nickelsen n...@jnickelsen.de wrote:
 Juergen Nickelsen n...@jnickelsen.de writes:

 Solaris Bundled Driver: * vgatext/ ** radeon
 Video
 ATI Technologies Inc
 R360 NJ [Radeon 9800 XT]

 I *think* this is the same driver used with my work laptop (which I
 don't have at hand to check, unfortunately), also with ATI graphics
 hardware.

 Confirmed.
 Regards, Juergen.

 --
 What you won was the obligation to pay more for something than
 anybody else thought it was worth.
               -- Delainey and Rasmussen's Betty about eBay
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] invalid vdev configuration after power failure

2009-03-02 Thread Blake
that link suggests that this is a problem with a dirty export:

http://www.sun.com/msg/ZFS-8000-EY

maybe try importing on system A again, doing a 'zpool export', waiting
for completion, then moving to system B to import?

On Sun, Mar 1, 2009 at 2:29 PM, Kyle Kakligian small...@gmail.com wrote:
 What does it mean for a vdev to have an invalid configuration and how
 can it be fixed or reset? As you can see, the following pool can no
 longer be imported: (Note that the last accessed by another system
 warning is because I moved these drives to my test workstation.)

 ~$ zpool import -f pool0
 cannot import 'pool0': invalid vdev configuration

 ~$ zpool import
  pool: pool0
    id: 5915552147942272438
  state: UNAVAIL
 status: The pool was last accessed by another system.
 action: The pool cannot be imported due to damaged devices or data.
   see: http://www.sun.com/msg/ZFS-8000-EY
 config:

        pool0       UNAVAIL  insufficient replicas
          raidz1    UNAVAIL  corrupted data
            c5d1p0  ONLINE
            c4d0p0  ONLINE
            c4d1p0  ONLINE
            c6d0p0  ONLINE
            c5d0p0  ONLINE
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Blake
excellent!  i wasn't sure if that was the case, though i had heard rumors.


On Mon, Mar 2, 2009 at 12:36 PM, Matthew Ahrens matthew.ahr...@sun.com wrote:
 Blake wrote:

 zfs send is great for moving a filesystem with lots of tiny files,
 since it just handles the blocks :)



 I'd like to see:

 pool-shrinking (and an option to shrink disk A when i want disk B to
 become a mirror, but A is a few blocks bigger)

 I'm working on it.

 install to mirror from the liveCD gui

 zfs recovery tools (sometimes bad things happen)

 automated installgrub when mirroring an rpool

 --matt

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can VirtualBox run a 64 bit guests on 32 bit host

2009-02-28 Thread Blake Irvin

Check out http://www.sun.com/bigadmin/hcl/data/os


Sent from my iPhone

On Feb 28, 2009, at 2:20 AM, Harry Putnam rea...@newsguy.com wrote:


Brian Hechinger wo...@4amlunch.net writes:

[...]


I think it would be better to answer this question that it would to
attempt to answer the VirtualBox question (I run it on a 64-bit OS,
so I can't really answer that anyway).


Thanks yes and appreciated here


The benefit to running ZFS on a 64-bit OS is if you have a large
amount of RAM.  I don't know what the breaking point is, but I can
definitely tell you that a 32-bit kernel and 4GB ram doesn't mix
well.  If all you are doing is testing ZFS on VMs you probably
aren't all that worried about performance so it really shouldn't be
an issue for you to run 32-bit.  I'd say keep your RAM allocations
down, and I wish I knew what to tell you to keep it under.
Hopefully someone who has a better grasp of all that can chime in.

Once you put it on real hardware, however, you really want a 64-bit
CPU and as much RAM as you can toss at the machine.



Sounds sensible, thanks for common sense input.

Just the little I've tinkered with zfs so far I'm in love already. zfs
is much more responive to some kinds of things I'm used to waiting for
on linux reiserfs.

Commands like du, mv, rm etc on hefty amounts of data are always slow
as molasses on linux/reiserfs (and reiserfs is faster than ext3).  I
have'nt tried ext4 but have been told it is no faster.

Whereas zfs gets those jobs done in short order... very noticably
faster but I am just going by feel but at least on very similar
hardware (cpu wise). (The linux is on Intel 3.06 celeron 2gb ram)

I guess there is something called btrfs (nicknamed butter fs) that is
supposed to be linux answer to zfs but it isn't ready for primetime
yet and I can say it will have a ways to go to compare to zfs.

My usage and skill level is probably the lowest on this list easily
but even I see some real nice features with zfs.  It seams taylor made
for semi-ambitious home NAS.

So Brian, If you can bear with my windyness a bit more, one of the
things flopping around in the back of my mind is something already
mentioned here too.. change out the mobo instead of dinking around
with addon pci sata controller..

I have 64 bit hardware... but am a bit scared of having lots of
trouble getting opensol to run peacefully on it.  Its a (somewhat old
fashioned now) athlon64 2.2 ghz +3400/Aopen AK86-L mobo. (socket 754)

The little jave tool that tests the hardware says my sata controller
wont work (the testing tool saw it as a VIA raid controller) and
suggests I turn off RAID in the bios.

After a carefull look in the bios menus I'm not finding any way to
turn it off so guessing the sata ports will be useless unless I
install a pci addon sata controller.

So thinking of justs changing out the mobo for something with stuff
that is known to work.

The machine came with an Asus mobo that I ruined myself by dicking
aournd installing RAM... somehow shorted out something, then mobo
became useless.

But I'm thinking of turning to Asus again and making sure there is
onboard SATA with at least 4 prts  and preferebly 6.

So cutting to the chase here... would you happen to have a
recommendation from your own experience, or something you've heard
will work and that can stand more ram... my current setup tops out at
3gb.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Blake Irvin

Shrinking pools would also solve the right-sizing dilemma.

Sent from my iPhone

On Feb 28, 2009, at 3:37 AM, Joe Esposito j...@j-espo.com wrote:


I'm using opensolaris and zfs at my house for my photography storage
as well as for an offsite backup location for my employer and several
side web projects.

I have an 80g drive as my root drive.  I recently took posesion of 2
74g 10k drives which I'd love to add as a mirror to replace the 80 g
drive.

From what I gather it is only possible if I zfs export my storage
array and reinstall solaris on the new disks.

So I guess I'm hoping zfs shrink and grow commands show up sooner or  
later.


Just a data point.

Joe Esposito
www.j-espo.com

On 2/28/09, C. Bergström cbergst...@netsyncro.com wrote:

Blake wrote:

Gnome GUI for desktop ZFS administration



On Fri, Feb 27, 2009 at 9:13 PM, Blake blake.ir...@gmail.com  
wrote:



zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B  
to

become a mirror, but A is a few blocks bigger)

This may be interesting... I'm not sure how often you need to  
shrink a
pool though?  Could this be classified more as a Home or SME level  
feature?

install to mirror from the liveCD gui


I'm not working on OpenSolaris at all, but for when my projects
installer is more ready /we/ can certainly do this..

zfs recovery tools (sometimes bad things happen)

Agreed.. part of what I think keeps zfs so stable though is the  
complete
lack of dependence on any recovery tools..  It forces customers to  
bring

up the issue instead of dirty hack and nobody knows.

automated installgrub when mirroring an rpool


This goes back to an installer option?

./C

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Virutal zfs server vs hardware zfs server

2009-02-27 Thread Blake
On Fri, Feb 27, 2009 at 12:31 PM, Harry Putnam rea...@newsguy.com wrote:
 Are you talking about the official Opensol-11 install iso or something
 else?
The official 2008.11 LiveCD has the tool on the default desktop as an icon.


 A big issue with running a VM is that ZFS prefers direct access to storage.

 What effect does this preferance have?  Does it perform badly when it
 does not have direct access to storage?  Are the virtual disks
 supplied by vmware less functional in some way?
I would expect pretty bad performance adding VMWare as a layer in
between ZFS and your block devices.  ZFS documentation specifically
advises against abstracting block devices whenever possible.  Since
ZFS is trying to checksum blocks, the fewer abstraction layers you
have in between ZFS and spinning rust, the less points of
error/failure.



 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Virutal zfs server vs hardware zfs server

2009-02-27 Thread Blake
I meant that the more layers you remove, the less layers there are
that can tell ZFS something that's not true.  I guess ZFS would still
catch those errors in most cases - it would still be a pain to deal
with needless errors.  Also I like to do what the manual says, and the
manual says avoid abstraction layers :)

Harry, Richard is probably right.  There are plenty of boards with
nVidia or Intel SATA that should work fine.  Search for 'opensolaris
hcl' (hardware compatibility list) - there are about 400+ mobos listed
there that are reported to work.



On Fri, Feb 27, 2009 at 1:48 PM, Bob Friesenhahn
bfrie...@simple.dallas.tx.us wrote:
 On Fri, 27 Feb 2009, Blake wrote:

 SinceZFS is trying to checksum blocks, the fewer abstraction layers
 youhave in between ZFS and spinning rust, the less points oferror/failure.

 Are you saying that ZFS checksums are responsible for the failure?

 In what way does more layers of abstraction cause particular problems for
 ZFS which won't also occur with some other filesystem?

 Bob
 --
 Bob Friesenhahn
 bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
 GraphicsMagick Maintainer,    http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread Blake
zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B to
become a mirror, but A is a few blocks bigger)

install to mirror from the liveCD gui

zfs recovery tools (sometimes bad things happen)

automated installgrub when mirroring an rpool




On Fri, Feb 27, 2009 at 8:02 PM, Richard Elling
richard.ell...@gmail.com wrote:
 David Magda wrote:

 On Feb 27, 2009, at 18:23, C. Bergström wrote:

 Blake wrote:

 Care to share any of those in advance?  It might be cool to see input
 from listees and generally get some wheels turning...

 raidz boot support in grub 2 is pretty high on my list to be honest..

 Which brings up another question of where is the raidz stuff mostly?

 usr/src/uts/common/fs/zfs/vdev_raidz.c ?

 Any high level summary, docs or blog entries of what the process would
 look like for a raidz boot support is also appreciated.

 Given the threads that have appeared on this list lately, how about
 codifying / standardizing the output of zfs send so that it can be backed
 up to tape? :)

 It wouldn't help.  zfs send is a data stream which contains parts of files,
 not files (in the usual sense), so there is no real way to take a send
 stream and extract a file, other than by doing a receive.

 At the risk of repeating the Best Practices Guide (again):
 The zfs send and receive commands do not provide an enterprise-level backup
 solution.
 -- richard

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Virutal zfs server vs hardware zfs server

2009-02-27 Thread Blake
Brandon makes a good point.  I think that's an option to pursue if you
don't want to risk messing up your Windows install.

If you can, dedicate entire disks, rather that partitions, to ZFS.
It's easier to manage.  ZFS is managed by the VMs processor in this
case, so you will take a bigger performance hit than running on bare
metal.  That said, my filer exporting ZFS over NFS to 10 busy CentOS
clients barely breaks a sweat.



On Fri, Feb 27, 2009 at 7:51 PM, Harry Putnam rea...@newsguy.com wrote:
 Brandon High bh...@freaks.com writes:

 On Thu, Feb 26, 2009 at 8:35 AM, Blake blake.ir...@gmail.com wrote:
   A big issue with running a VM is that ZFS prefers direct access to 
 storage.

 VMWare can give VMs direct access to the actual disks. This should
 avoid the overhead of using virtual disks.

 Can you say if it makes a noticeable difference to zfs.  I'd noticed
 that option but didn't connect it to this conversation.  Also, if I
 recall there is some warning about being an advanced user to use that
 option or something similar.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread Blake
Gnome GUI for desktop ZFS administration



On Fri, Feb 27, 2009 at 9:13 PM, Blake blake.ir...@gmail.com wrote:
 zfs send is great for moving a filesystem with lots of tiny files,
 since it just handles the blocks :)



 I'd like to see:

 pool-shrinking (and an option to shrink disk A when i want disk B to
 become a mirror, but A is a few blocks bigger)

 install to mirror from the liveCD gui

 zfs recovery tools (sometimes bad things happen)

 automated installgrub when mirroring an rpool




 On Fri, Feb 27, 2009 at 8:02 PM, Richard Elling
 richard.ell...@gmail.com wrote:
 David Magda wrote:

 On Feb 27, 2009, at 18:23, C. Bergström wrote:

 Blake wrote:

 Care to share any of those in advance?  It might be cool to see input
 from listees and generally get some wheels turning...

 raidz boot support in grub 2 is pretty high on my list to be honest..

 Which brings up another question of where is the raidz stuff mostly?

 usr/src/uts/common/fs/zfs/vdev_raidz.c ?

 Any high level summary, docs or blog entries of what the process would
 look like for a raidz boot support is also appreciated.

 Given the threads that have appeared on this list lately, how about
 codifying / standardizing the output of zfs send so that it can be backed
 up to tape? :)

 It wouldn't help.  zfs send is a data stream which contains parts of files,
 not files (in the usual sense), so there is no real way to take a send
 stream and extract a file, other than by doing a receive.

 At the risk of repeating the Best Practices Guide (again):
 The zfs send and receive commands do not provide an enterprise-level backup
 solution.
 -- richard

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Can VirtualBox run a 64 bit guests on 32 bit host

2009-02-26 Thread Blake
The changelog says 64-bit guest on 32-bit host support was added in 2.1:

http://www.virtualbox.org/wiki/Changelog

On Thu, Feb 26, 2009 at 10:48 AM, Brian Hechinger wo...@4amlunch.net wrote:
 On Wed, Feb 25, 2009 at 07:14:14PM -0600, Harry Putnam wrote:

 My whole purpose is to experiment with zfs... would I see much
 difference if opensol was installed 64 bit as compared to 32 bit?

 I noticed the Simon blogs that describe how to setup a home zfs server
 ( http://breden.org.uk/2008/03/02/a-home-fileserver-using-zfs/ )
 mention it is best setup 64 bit, but no real reason is given.

 I think it would be better to answer this question that it would to attempt
 to answer the VirtualBox question (I run it on a 64-bit OS, so I can't really
 answer that anyway).

 The benefit to running ZFS on a 64-bit OS is if you have a large amount of 
 RAM.
 I don't know what the breaking point is, but I can definitely tell you that a
 32-bit kernel and 4GB ram doesn't mix well.  If all you are doing is testing 
 ZFS
 on VMs you probably aren't all that worried about performance so it really 
 shouldn't
 be an issue for you to run 32-bit.  I'd say keep your RAM allocations down, 
 and I
 wish I knew what to tell you to keep it under.  Hopefully someone who has a 
 better
 grasp of all that can chime in.

 Once you put it on real hardware, however, you really want a 64-bit CPU and 
 as much
 RAM as you can toss at the machine.

 -brian
 --
 Coding in C is like sending a 3 year old to do groceries. You gotta
 tell them exactly what you want or you'll end up with a cupboard full of
 pop tarts and pancake mix. -- IRC User (http://www.bash.org/?841435)
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating to ZFS

2009-02-26 Thread Blake
Rafael,
   If you are talking just about moving a bunch of data, take a look
at rsync.  I think it will work nicely for moving files from one
volume to another, preserving attributes.  It comes bundled with
2008.11 and up.



On Thu, Feb 26, 2009 at 10:41 AM,  cindy.swearin...@sun.com wrote:
 Hi Rafael,

 The information on that site looks very out-of-date. I will attempt to
 resolve this problem.

 Other than using Live Upgrade to migrate a UFS root file system to a ZFS
 root file system, you can use ufsdump and ufsrestore to migrate UFS data to
 ZFS file system.

 Other data migration processes are mostly manual.

 If something else is in the works, we'll let you know.

 Cindy

 Rafael Friedlander wrote:

 Hi,
 According to
 http://www.sun.com/emrkt/campaign_docs/expertexchange/knowledge/solaris_zfs_install.html#2
 we will have a tool for data migration to ZFS (not a root file system).

 Do you know when this is happening?

 Thanks, Rafael.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Virutal zfs server vs hardware zfs server

2009-02-26 Thread Blake
Harry,
   The LiveCD for OpenSolaris has a driver detection tool on it - this
will let you see if your hardware is supported without touching the
installed XP system.

   A big issue with running a VM is that ZFS prefers direct access to storage.


On Wed, Feb 25, 2009 at 11:48 PM, Harry Putnam rea...@newsguy.com wrote:
 I'm experimenting with a zfs home server.  Running Opensol-11 by
 way of vmware on WinXP.

 It seems one way to avoid all the hardware problems one might run into
 trying to install opensol on available or spare hardware.

 Are there some bad gotchas about running opensol/zfs through vmware and
 never going to real hardware?

 One thing comes to mind is the overhead of two OSs on one processor.
 An Athlon64 2.2 +3400 running 32bit Windows XP and opensol in VMware.

 But if I lay off the windows OS... like not really working it with
 transcibing video or compressing masses of data or the like. Is this
 likely to be a problem?

 Also I'm loosing out on going 64 bit since its not likely this machine
 supports the AMD V extensions... and I'm short on SATA connections.  I
 only have two onboard, but plan to install a pci style sata controller
 to squeeze in some more discs.

 Its a  big old ANTEC case so I don't think getting the discs in there
 will be much of a problem.  But have wondered if a PCI sata controller
 is likely to be a big problem.

 So, are there things I need to know about that will make running a zfs
 home server from vmware a bad idea?

 The server will be serving as backup destination for 5 home machines
 and most likely would see service only about 2-3 days a week far as
 any kind of heavy usage like ghosted disc images and other large
 chunks of data + a regular 3 day a week backup running from windows
 using `retrospect' to backup user directories and changed files in
 C:\.

 A 6th (linux) machine may eventually start using the server but for
 now its pretty selfcontained and has lots of disc space.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-26 Thread Blake
Care to share any of those in advance?  It might be cool to see input
from listees and generally get some wheels turning...

On Wed, Feb 25, 2009 at 4:39 AM, C. Bergström
cbergst...@netsyncro.com wrote:

 Hi everyone.

 I've got a couple ideas for good zfs GSoC projects, but wanted to stir some
 interest.  Anyone interested to help mentor?  The deadline is around the
 corner so if planning hasn't happened yet it should start soon.  If there is
 interest who would the org administrator be?

 Thanks

 ./Christopher
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Motherboard for home zfs/solaris file server -- ECC claims

2009-02-26 Thread Blake
IIRC, the AMD board I have at my office has hardware ECC scrub.  I
have no idea if Solaris knows about this or makes any use of it (or
needs to?)

On Tue, Feb 24, 2009 at 2:50 PM, Miles Nordin car...@ivy.net wrote:
 rl == Rob Logan r...@logan.com writes:

    rl that's why this X58 MB claims ECC support:

 the claim is worth something.  People always say ``AMD supports ECC
 because the memory controller is in the CPU so they all support it, it
 cannot be taken away from you by lying idiot motherboard manufacturers
 or greedy marketers trying to segment users into different demand
 groups'' but you still need some motherboard BIOS to flip the ECC
 switch to ``wings stay on'' mode before you start down the runway.

 Here is a rather outdated and Linux-specific workaround for cheapo AMD
 desktop boards that don't have an ECC option in their BIOS:

  http://newsgroups.derkeiler.com/Archive/Alt/alt.comp.periphs.mainboard.asus/2005-10/msg00365.html
  http://hyvatti.iki.fi/~jaakko/sw/

 The discussion about ECC-only vs scrub-and-fix, about how to read from
 PCI if ECC errors are happening (though not necessarily which stick),
 and his 10-ohm testing method, is also interesting.  I still don't
 understand what chip-kill means.

 I remember something about a memory scrubbing kernel thread in
 Solaris.  This sounds like the AMD chips have a hardware scrubber?
 Also how are ECC errors reported in Solaris?  I guess this is getting
 OT though.

 Anyway ECC is not just a feature bullet to gather up and feel good.
 You have to finish the job and actually interact with it.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Migrating to ZFS

2009-02-26 Thread Blake
Ah - I think I was getting confused by my experience with the modified
rsync on OS X.


On Thu, Feb 26, 2009 at 1:54 PM, Ian Collins i...@ianshome.com wrote:
 Blake wrote:

 Rafael,
   If you are talking just about moving a bunch of data, take a look
 at rsync.  I think it will work nicely for moving files from one
 volume to another, preserving attributes.  It comes bundled with
 2008.11 and up.




 But not ACLs.

 --
 Ian.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backing up ZFS snapshots

2009-02-25 Thread Blake
I'm sure that's true.  My point was that, given the choice between a
zfs send/recv from one set of devices to another, where the target is
another pool, and sending a zfs stream to a tarball, I'd sooner choose
a solution that's all live filesystems.

If backups are *really* important, then it's certainly better to use a
product with commercial support.  I think Amanda is zfs-aware now?


On Mon, Feb 23, 2009 at 12:16 PM, Miles Nordin car...@ivy.net wrote:
 b == Blake  blake.ir...@gmail.com writes:

     c There are other problems besides the versioning.

     b Agreed - I don't think that archiving simply the send stream
     b is a smart idea (yet, until the stream format is stabilized

 *there* *are* *other* *problems* *besides* *the* *versioning*!

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Confused about zfs recv -d, apparently

2009-02-22 Thread Blake
I'm actually working on this for an application at my org.  I'll try
to post my work somewhere when done (hopefully this week).

Are you keeping in mind the fact that the '-i' option needs a pair of
snapshots (original and current) to work properly?



On Sun, Feb 22, 2009 at 2:14 PM, David Dyer-Bennet d...@dd-b.net wrote:

 On Sun, February 22, 2009 00:15, David Dyer-Bennet wrote:
 First, it fails because the destination directory doesn't exist.  Then it
 fails because it DOES exist.  I really expected one of those to work.  So,
 what am I confused about now?  (Running 2008.11)

 # zpool import -R /backups/bup-ruin bup-ruin
 # zfs send -R z...@bup-20090222-054457utc | zfs receive -dv
 bup-ruin/fsfs/zp1
 cannot receive: specified fs (bup-ruin/fsfs/zp1) does not exist
 # zfs create bup-ruin/fsfs/zp1
 # zfs send -R z...@bup-20090222-054457utc | zfs receive -dv
 bup-ruin/fsfs/zp1
 cannot receive new filesystem stream: destination 'bup-ruin/fsfs/zp1'
 exists
 must specify -F to overwrite it

 I've tried some more things.  Isn't this the same operations as above,
 but with different (and more reasonable) results?

 But isn't this the same, with different results?

 # zfs list -r bup-ruin
 NAME   USED  AVAIL  REFER  MOUNTPOINT
 bup-ruin  79.5K   913G18K  /backups/bup-ruin
 # zfs create bup-ruin/fsfs
 # zfs send -R rpool/export/h...@bup-2009\
 0216-044512UTC | zfs recv -d bup-ruin/fsfs/rpool
 cannot receive: specified fs (bup-ruin/fsfs/rpool) does not exist
 r...@fsfs:/export/home/localddb/src/bup2# zfs create bup-ruin/fsfs/rpool
 r...@fsfs:/export/home/localddb/src/bup2# zfs send -R
 rpool/export/h...@bup-2009\
 0216-044512UTC | zfs recv -d bup-ruin/fsfs/rpool

 These second results are what I expected, after reading the error messages
 and the manual.  But the first example is what I actually got, originally.
  (Different pools, slightly).

 Here's what I'm trying to do:

 I'm trying to store backups of multiple pools (which are on disks mounted
 in the desktop chassis) on external pools consisting of a single external
 USB drive.

 My concept is to do a send -R for the initial backup of each, and then to
 do send -i with suitable params for the later backups.  This should keep
 the external filesystems in synch with the internal filesystems up to the
 snapshot most recently synced.  But I can't find commands to make this
 work.

 (Note that I need to back up two pools, rpool and zp1, from the destkop on
 the the single external pool bup-ruin.  I'm importing bup-ruin with
 altroot to avoid the mountoints of the backed-up filesystems on it
 conflicting with each other or with stuff already mounted on the desktop.)


 --
 David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/
 Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/
 Photos: http://dd-b.net/photography/gallery/
 Dragaera: http://dragaera.info

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backing up ZFS snapshots

2009-02-22 Thread Blake
I thinks that's legitimate so long as you don't change ZFS versions.

Personally, I'm more comfortable doing a 'zfs send | zfs recv' than I
am storing the send stream itself.  The problem I have with the stream
is that I may not be able to receive it in a future version of ZFS,
while I'm pretty sure that I can upgrade an actual pool/fs pretty
easily.



On Sun, Feb 22, 2009 at 4:48 PM, David Abrahams d...@boostpro.com wrote:

 on Wed Feb 18 2009, Frank Cusack fcusack-AT-fcusack.com wrote:

 On February 17, 2009 3:57:34 PM -0800 Joe S js.li...@gmail.com wrote:
 On Tue, Feb 17, 2009 at 3:35 PM, David Magda dma...@ee.ryerson.ca wrote:
 If you want to do back ups of your file system use a documented utility
 (tar, cpio, pax, zip, etc.).

 I'm going to try to use Amanda and backup my data (not snapshots).

 You missed the point, which is not to avoid snapshots, but to avoid
 saving the stream as a backup.  Backing up a snapshot is typically
 preferred to backing up a live filesystem.

 Has anyone here noticed that
 http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide
 suggests in several places that zfs send streams be stored for backup?

 --
 Dave Abrahams
 BoostPro Computing
 http://www.boostpro.com

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backing up ZFS snapshots

2009-02-22 Thread Blake
Agreed - I don't think that archiving simply the send stream is a
smart idea (yet, until the stream format is stabilized in some way).
I'd much rather archive to a normal ZFS filesystem.  With ZFS's
enormous pool capacities, it's probably the closest thing we have
right now to a future-proof filesystem.

On Sun, Feb 22, 2009 at 6:58 PM, Miles Nordin car...@ivy.net wrote:

 well fine, but there's certainly not a consensus on that, which makes
 it not a ``best practice.''  There are other problems besides the
 versioning.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Oracle doc: Unable to run Forms applications when using Solaris Zettabyte File System (ZFS)

2009-02-20 Thread Blake
If this happens if ZFS is in use anywhere in the system, I'm not sure
of a solution.

If you just need Oracle files and activity to be on something other
than ZFS, could you try creating a ZFS block device and formatting it
UFS?

(disclaimer: I'm not an Oracle user)

On Fri, Feb 20, 2009 at 12:12 PM, Jose Gregores jose.grego...@sun.com wrote:
 Hi all,
 We have a Suncluster/ZFS  customer complaining about this issue.

 Subject: Unable to run Forms applications when using Solaris Zettabyte
 File System (ZFS)
   Doc ID: 730691.1 Type: PROBLEM
   Modified Date: 24-SEP-2008 Status: PUBLISHED

 In this Document
   Symptoms
   Changes
   Cause
   Solution
   References

 Applies to:
 Oracle Application Server 10g Enterprise Edition - Version: 1.0.2.0.0 to
 10.1.3.0.0
 Oracle Forms - Version: 6.0 to 10.1.2
 Solaris Operating System (SPARC 64-bit)
 Solaris Operating System (SPARC 32-bit)
 Symptoms

 A common symptom will include Oracle Forms applications intermittently or
 consistently failing at startup or while in use. The error displayed on the
 client will be FRM-92101.
 Changes
 This problem can be exposed when enabling Solaris Zettabyte File System
 (ZFS).
 Cause

 Oracle Fusion Middleware and all of it subcomponents (i.e. Forms, Reports,
 etc) have not been certified to be used in a Solaris ZFS environment or any
 other specific file systems.  This includes all Application Server versions
 1.0 - 10.1.3.  Therefore, Oracle technical support will be limited for
 installations configured in this type of environment.

 Usually, Oracle Fusion Middleware software can be installed on file systems
 provided by the OS vendor. If there is a compatibility issue specific to an
 OS version and/or file system that is being used,  contact the OS vendors
 directly.

 Internal Bug 7308848 - CLARIFICATION ON AS CERTIFICATION USING ZFS
 Solution

 Because Solaris ZFS has not been certified and technical support will be
 limited for this configuration, it is strongly recommended that ZFS not be
 used without extensive testing prior to moving to production.  Problems
 reproducible only when using a specific file system type should be directed
 toward the OS vendor and not Oracle.

 It is not expected that ZFS will be supported for use with Application
 Server 10.1.2 or 10.1.3 before these product versions reach their desupport
 dates.  Information about Fusion Middleware support and support dates can be
 found in a brochure titled, Lifetime Support Policy: Oracle Technology
 Products found on the Oracle web site.

 http://www.oracle.com/support/library/brochure/lifetime-support-technology.pdf

 Please refer to the Application Server Certification documentation for the
 latest updates regarding this issue.

 http://www.oracle.com/technology/software/products/ias/files/as_certification_r2_101202.html
 References


 Do we have a solution or suggestion for this problem ?
 Thanks.


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Oracle doc: Unable to run Forms applications when using Solaris Zettabyte File System (ZFS)

2009-02-20 Thread Blake
It looks like this is a bigger Oracle environment than any database
environment that I've ever worked in :)

Anyway - when you create a zfs filesystem, you can do it with the '-V'
option, which makes a block device of a fixed size that's accessible
through the usual /dev device path.  You can then put a UFS filesystem
on this ZFS-backed block device.

See 'man zfs' for details.



On Fri, Feb 20, 2009 at 1:43 PM, Jose Gregores jose.grego...@sun.com wrote:
 Blake escreveu:

 If this happens if ZFS is in use anywhere in the system, I'm not sure
 of a solution.


 Yes, this happens on our customer that has a 3 node Suncluster with 150
 zpools and 50 zones and 15 Oracle databases.

 If you just need Oracle files and activity to be on something other
 than ZFS, could you try creating a ZFS block device and formatting it
 UFS?



 How can   we do that ?

 (disclaimer: I'm not an Oracle user)

 On Fri, Feb 20, 2009 at 12:12 PM, Jose Gregores jose.grego...@sun.com
 wrote:


 Hi all,
 We have a Suncluster/ZFS  customer complaining about this issue.

 Subject: Unable to run Forms applications when using Solaris Zettabyte
 File System (ZFS)
   Doc ID: 730691.1 Type: PROBLEM
   Modified Date: 24-SEP-2008 Status: PUBLISHED

 In this Document
   Symptoms
   Changes
   Cause
   Solution
   References

 Applies to:
 Oracle Application Server 10g Enterprise Edition - Version: 1.0.2.0.0 to
 10.1.3.0.0
 Oracle Forms - Version: 6.0 to 10.1.2
 Solaris Operating System (SPARC 64-bit)
 Solaris Operating System (SPARC 32-bit)
 Symptoms

 A common symptom will include Oracle Forms applications intermittently or
 consistently failing at startup or while in use. The error displayed on the
 client will be FRM-92101.
 Changes
 This problem can be exposed when enabling Solaris Zettabyte File System
 (ZFS).
 Cause

 Oracle Fusion Middleware and all of it subcomponents (i.e. Forms, Reports,
 etc) have not been certified to be used in a Solaris ZFS environment or any
 other specific file systems.  This includes all Application Server versions
 1.0 - 10.1.3.  Therefore, Oracle technical support will be limited for
 installations configured in this type of environment.

 Usually, Oracle Fusion Middleware software can be installed on file systems
 provided by the OS vendor. If there is a compatibility issue specific to an
 OS version and/or file system that is being used,  contact the OS vendors
 directly.

 Internal Bug 7308848 - CLARIFICATION ON AS CERTIFICATION USING ZFS
 Solution

 Because Solaris ZFS has not been certified and technical support will be
 limited for this configuration, it is strongly recommended that ZFS not be
 used without extensive testing prior to moving to production.  Problems
 reproducible only when using a specific file system type should be directed
 toward the OS vendor and not Oracle.

 It is not expected that ZFS will be supported for use with Application
 Server 10.1.2 or 10.1.3 before these product versions reach their desupport
 dates.  Information about Fusion Middleware support and support dates can be
 found in a brochure titled, Lifetime Support Policy: Oracle Technology
 Products found on the Oracle web site.

 http://www.oracle.com/support/library/brochure/lifetime-support-technology.pdf

 Please refer to the Application Server Certification documentation for the
 latest updates regarding this issue.

 http://www.oracle.com/technology/software/products/ias/files/as_certification_r2_101202.html
 References


 Do we have a solution or suggestion for this problem ?
 Thanks.


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Confused about prerequisites for ZFS to work

2009-02-18 Thread Blake
You definitely need SUNWsmbskr - the cifs server provided with
OpenSolaris is tied to the kernel at some low level.

I found this entry helpful:

http://blogs.sun.com/timthomas/entry/solaris_cifs_in_workgroup_mode



On Wed, Feb 18, 2009 at 1:03 PM, Harry Putnam rea...@newsguy.com wrote:
 Ian Collins wrote:

 Harry Putnam wrote:

 [...]

 Still when I look again... its still in maintenance mode.

 What does tail /var/svc/log/network-smb-server:default.log show?

 The log file for a service listed as part of the long listing (svcs -l
 smb/server).


 Following these two commands:
 svcadm disable sbm/server
 svcadm enable -r smb/server

 [...]

 [ Feb 18 11:53:42 Executing start method (/usr/lib/smbsrv/smbd start). ]
 smbd: NetBIOS services started
 smbd: kernel bind error: No such file or directory
 smbd: daemon initialization failed
 [ Feb 18 11:53:43 Method start exited with status 95. ]

 I wondered about the kernel in previous post:

 I noticed a
 special kernel package right next to the smb server called. SUNWsmbskr
 (smb/server kernel).

 So does this require a special kernel?

 Does that mean I need to do something to the kernel..?  Replace it with
 SUNWsmbskr?

 Or is it something less ambitious?

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zpool scrub in cron hangs u3/u4 server, stumps tech support.

2009-02-18 Thread Blake
Bob is correct to praise LiveUpgrade.  It's pretty much risk-free when
used properly, provided you have some spare slices/disks.

At the same time, I'd say that this is probably an appropriate time to
escalate the bug with support - the answers you are getting aren't
satisfactory.

I would also consider creating a user/role with zfs admin privileges
only, and trying to run the scrub command from cron as this user - I
had a similar problem with an old ZFS version which I worked around by
issuing commands as a user other than root.



On Wed, Feb 18, 2009 at 12:44 PM, Bob Friesenhahn
bfrie...@simple.dallas.tx.us wrote:
 On Wed, 18 Feb 2009, Elizabeth Schwartz wrote:

 It's an old version but it's a *supported* version and we have a
 five-figure support contract. That used to matter.

 I can understand your frustration.  ZFS in Solaris 10U3 was a bit rough
 around the edges.  It is definitely improved in later releases.

 I've never used Live Upgrade; I want to try it out but not on my
 production file server, and I want to know that this particular bug is
 fixed first, something more definite than many improvements

 As long as you have spare bootable partitions, Live Upgrade is exceedingly
 useful.  It allows you to create a new boot environment with the newer
 Solaris installed, and with all of your local changes applied.  You can
 double-check to make sure that everything is ready to go via a mount to the
 new boot evironment.  Switching to the new boot environment is as simple as
 'luactivate' followed by a reboot. It is likely to work first time, but if
 it does not, you can reboot to your previous boot environment for minimal
 server down time.  If you are using Grub, then each boot environment is
 listed in the Grub boot menu.

 With proper care, using Live Upgrade is safer (and faster) for production
 systems than applying large numbers of patches spanning many Solaris 10
 generations.  You can also use multiple boot environments to apply patches,
 in order to minimize risk and minimize down time.

 If you are able to install Solaris 10U6 with ZFS boot, then subsequent Live
 Upgrades should be far easier since boot evironments are directories in the
 root pool ('rpool') rather than in dedicated partitions.

 Bob
 --
 Bob Friesenhahn
 bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
 GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Confused about prerequisites for ZFS to work

2009-02-18 Thread Blake
have you made sure that samba is *disabled*?

svcs samba

?

On Wed, Feb 18, 2009 at 4:14 PM, Harry Putnam rea...@newsguy.com wrote:
 Blake wrote:

 You definitely need SUNWsmbskr - the cifs server provided with
 OpenSolaris is tied to the kernel at some low level.

 I found this entry helpful:

 http://blogs.sun.com/timthomas/entry/solaris_cifs_in_workgroup_mode

 Looks like it will be immensely so..
 However it appears from the comments there that I need to have the smb
 server online before the directions there are very usefull.

 I've installed both SUNWsmbs and SUNWsmbskr  and rebooted as directed.
 pkg shows them installed.
   pkg list|grep smbs
  SUNWsmbs  0.5.11-0.101installed
  SUNWsmbskr0.5.11-0.101installed

 However I still get the same error when trying to start smb server.
  [ Feb 18 14:36:36 Enabled. ]
  [ Feb 18 14:36:36 Executing start method (/usr/lib/smbsrv/smbd
  start). ]
  smbd: NetBIOS services started
  smbd: kernel bind error: No such file or directory
  smbd: daemon initialization failed
  [ Feb 18 14:36:37 Method start exited with status 95. ]

 I'm not finding many clues with google as to what to do, but as always its
 not doubt that my search strings are lacking, since I don't have enough
 experience yet to know what to look for exactly.  Something like this:
how to install SUNWsmbskr
 Turns up man messges about the need to reboot after installing but not any I
 see about how to deal with any problems with the install itself.

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] scrub on snv-b107

2009-02-17 Thread Blake
Do you have more data on the 107 pool than on the sol10 pool?

On Tue, Feb 17, 2009 at 6:11 AM, dick hoogendijk d...@nagual.nl wrote:
 scrub completed after 1h9m with 0 errors on Tue Feb 17 12:09:31 2009

 This is about twice as slow as the same srub on a solaris 10 box with a
 mirrored zfs root pool. Has scrub become that much slower? And if so,
 why?

 --
 Dick Hoogendijk -- PGP/GnuPG key: 01D2433D
 + http://nagual.nl/ | SunOS sxce snv107 ++
 + All that's really worth doing is what we do for others (Lewis Carrol)
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs destroy hanging

2009-02-14 Thread Blake
I think you can kill the destroy command process using traditional methods.

Perhaps your slowness issue is because the pool is an older format.
I've not had these problems since upgrading to the zfs version that
comes default with 2008.11


On Fri, Feb 13, 2009 at 4:14 PM, David Dyer-Bennet d...@dd-b.net wrote:
 This shouldn't be taking anywhere *near* half an hour.  The snapshots
 differ trivially, by one or two files and less than 10k of data (they're
 test results from working on my backup script).  But so far, it's still
 sitting there after more than half an hour.

 local...@fsfs:~/src/bup2# zfs destroy ruin/export
 cannot destroy 'ruin/export': filesystem has children
 use '-r' to destroy the following datasets:
 ruin/export/h...@bup-20090210-202557utc
 ruin/export/h...@20090210-213902utc
 ruin/export/home/local...@first
 ruin/export/home/local...@second
 ruin/export/home/local...@bup-20090210-202557utc
 ruin/export/home/local...@20090210-213902utc
 ruin/export/home/localddb
 ruin/export/home
 local...@fsfs:~/src/bup2# zfs destroy -r ruin/export

 It's still hung.

 Ah, here's zfs list output from shortly before I started the destroy:

 ruin 474G   440G   431G  /backups/ruin
 ruin/export 35.0M   440G18K  /backups/ruin/export
 ruin/export/home35.0M   440G19K  /export/home
 ruin/export/home/localddb 35M   440G  27.8M  /export/home/localddb

 As you can see, the ruin/export/home filesystem (and subs) is NOT large.

 iostat shows no activity on pool ruin over a minute.

 local...@fsfs:~$ pfexec zpool iostat ruin 10
   capacity operationsbandwidth
 pool used  avail   read  write   read  write
 --  -  -  -  -  -  -
 ruin 474G   454G 10  0  1.13M840
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0
 ruin 474G   454G  0  0  0  0

 The pool still thinks it is healthy.

 local...@fsfs:~$ zpool status -v ruin
  pool: ruin
  state: ONLINE
 status: The pool is formatted using an older on-disk format.  The pool can
still be used, but some features are unavailable.
 action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
pool will no longer be accessible on older software versions.
  scrub: scrub completed after 4h42m with 0 errors on Mon Feb  9 19:10:49 2009
 config:

NAMESTATE READ WRITE CKSUM
ruinONLINE   0 0 0
  c7t0d0ONLINE   0 0 0

 errors: No known data errors

 There is still a process out there trying to run that destroy.  It doesn't
 appear to be using much cpu time.

 local...@fsfs:~$ ps -ef | grep zfs
 localddb  7291  7228   0 15:10:56 pts/4   0:00 grep zfs
root  7223  7101   0 14:18:27 pts/3   0:00 zfs destroy -r ruin/export

 Running 2008.11.

 local...@fsfs:~$ uname -a
 SunOS fsfs 5.11 snv_101b i86pc i386 i86pc Solaris

 Any suggestions?  Eventually I'll kill the process by the gentlest way
 that works, I suppose (if it doesn't complete).
 --
 David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/
 Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/
 Photos: http://dd-b.net/photography/gallery/
 Dragaera: http://dragaera.info

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-12 Thread Blake
That does look like the issue being discussed.

It's a little alarming that the bug was reported against snv54 and is
still not fixed :(

Does anyone know how to push for resolution on this?  USB is pretty
common, like it or not for storage purposes - especially amongst the
laptop-using dev crowd that OpenSolaris apparently targets.



On Thu, Feb 12, 2009 at 4:44 PM, bdebel...@intelesyscorp.com
bdebel...@intelesyscorp.com wrote:
 Is this the crux of the problem?

 http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6424510

 'For usb devices, the driver currently ignores DKIOCFLUSHWRITECACHE.
 This can cause catastrophic data corruption in the event of power loss,
 even for filesystems like ZFS that are designed to survive it.
 Dropping a flush-cache command is just as bad as dropping a write.
 It violates the interface that software relies on to use the device.'
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] strange 'too many errors' msg

2009-02-12 Thread Blake
I think you could try clearing the pool - however, consulting the
fault management tools (fmdump and it's kin) might be smart first.
It's possible this is an error in the controller.

The output of 'cfgadm' might be of use also.



On Wed, Feb 11, 2009 at 7:12 PM, Jens Elkner
jel+...@cs.uni-magdeburg.de wrote:
 Hi,

 just found on a X4500 with S10u6:

 fmd: [ID 441519 daemon.error] SUNW-MSG-ID: ZFS-8000-GH, TYPE: Fault, VER: 1, 
 SEVERITY: Major
 EVENT-TIME: Wed Feb 11 16:03:26 CET 2009
 PLATFORM: Sun Fire X4500, CSN: 00:14:4F:20:E0:2C , HOSTNAME: peng
 SOURCE: zfs-diagnosis, REV: 1.0
 EVENT-ID: 74e6f0ec-b1e7-e49b-8d71-dc1c9b68ad2b
 DESC: The number of checksum errors associated with a ZFS device exceeded 
 acceptable levels.  Refer to http://sun.com/msg/ZFS-8000-GH for more 
 information.
 AUTO-RESPONSE: The device has been marked as degraded.  An attempt will be 
 made to activate a hot spare if available.
 IMPACT: Fault tolerance of the pool may be compromised.
 REC-ACTION: Run 'zpool status -x' and replace the bad device.

 zpool status -x
 ...
  mirror  DEGRADED 0 0 0
spare DEGRADED 0 0 0
  c6t6d0  DEGRADED 0 0 0  too many errors
  c4t0d0  ONLINE   0 0 0
c7t6d0ONLINE   0 0 0
 ...
spares
  c4t0d0  INUSE currently in use
  c4t4d0  AVAIL

 Strange thing is, that for more than 3 month there was no single error
 logged with any drive. IIRC, before u4 I've seen occasionaly a bad
 checksum error message, but this was obviously the result from the
 wellknown race condition of the marvell driver when havy writes took place.

 So I tend to interprete it as an false alarm and think about
 'zpool ... clear c6t6d0'.

 What do you think. Is this a good idea?

 Regards,
 jel.

 BTW: zpool status -x  msg refers to http://www.sun.com/msg/ZFS-8000-9P,
 the event to http://sun.com/msg/ZFS-8000-GH - little bit
 inconsistent I think.
 --
 Otto-von-Guericke University http://www.cs.uni-magdeburg.de/
 Department of Computer Science   Geb. 29 R 027, Universitaetsplatz 2
 39106 Magdeburg, Germany Tel: +49 391 67 12768
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS: unreliable for professional usage?

2009-02-12 Thread Blake
I'm sure it's very hard to write good error handling code for hardware
events like this.

I think, after skimming this thread (a pretty wild ride), we can at
least decide that there is an RFE for a recovery tool for zfs -
something to allow us to try to pull data from a failed pool.  That
seems like a reasonable tool to request/work on, no?


On Thu, Feb 12, 2009 at 6:03 PM, Toby Thain t...@telegraphics.com.au wrote:

 On 12-Feb-09, at 3:02 PM, Tim wrote:


 On Thu, Feb 12, 2009 at 11:31 AM, David Dyer-Bennet d...@dd-b.net wrote:

 On Thu, February 12, 2009 10:10, Ross wrote:

  Of course, that does assume that devices are being truthful when they
  say
  that data has been committed, but a little data loss from badly designed
  hardware is I feel acceptable, so long as ZFS can have a go at
  recovering
  corrupted pools when it does happen, instead of giving up completely
  like
  it does now.

 Well; not acceptable as such.  But I'd agree it's outside ZFS's purview.
  The blame for data lost due to hardware actively lying and not working to
 spec goes to the hardware vendor, not to ZFS.

 If ZFS could easily and reliably warn about such hardware I'd want it to,
 but the consensus seems to be that we don't have a reliable qualification
 procedure.  In terms of upselling people to a Sun storage solution, having
 ZFS diagnose problems with their cheap hardware early is clearly desirable
 :-).



 Right, well I can't imagine it's impossible to write a small app that can
 test whether or not drives are honoring correctly by issuing a commit and
 immediately reading back to see if it was indeed committed or not.

 You do realise that this is not as easy as it looks? :) For one thing, the
 drive will simply serve the read from cache.
 It's hard to imagine a test that doesn't involve literally pulling plugs;
 even better, a purpose built hardware test harness.
 Nonetheless I hope that someone comes up with a brilliant test. But if the
 ZFS team hasn't found one yet... it looks grim :)
 --Toby

 Like a zfs test cXtX.  Of course, then you can't just blame the hardware
 everytime something in zfs breaks ;)

 --Tim

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS snapshot splitting joinin

2009-02-09 Thread Blake
I believe Tim Foster's zfs backup service (very beta atm) has support
for splitting zfs send backups.  Might want to check that out and see
about modifying it for your needs.

On Thu, Feb 5, 2009 at 3:15 PM, Michael McKnight
michael_mcknigh...@yahoo.com wrote:
 Hi everyone,

 I appreciate the discussion on the practicality of archiving ZFS sends, but 
 right now I don't know of any other options.  I'm a home user, so 
 Enterprise-level solutions aren't available and as far as I know, tar, cpio, 
 etc. don't capture ACL's and other low-level filesystem attributes.   Plus, 
 they are all susceptible to corruption while in storage, making recovery no 
 more likely than with a zfs send.

 The checksumming capability is a key factor to me.  I would rather not be 
 able to restore the data than to unknowingly restore bad data.  This is the 
 biggest reason I started using ZFS to start with.  Too many cases of 
 invisible file corruption.  Admittedly, it would be nicer if zfs recv 
 would flag individual files with checksum problems rather than completely 
 failing the restore.

 What I need is a complete snapshot of the filesystem (ie. ufsdump) and, 
 correct me if I'm wrong, but zfs send/recv is the closest (only) thing we 
 have.  And I need to be able to break up this complete snapshot into pieces 
 small enough to fit onto a DVD-DL.

 So far, using ZFS send/recv works great as long as the files aren't split.  I 
 have seen suggestions on using something like 7z (?) instead of split as an 
 option.  Does anyone else have any other ideas on how to successfully break 
 up a send file and join it back together?

 Thanks again,
 Michael
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Introducing zilstat

2009-01-31 Thread Blake
I'm already using it.  This could be really useful for my Windows
roaming-profile application of ZFS/NFS/SMB

On Fri, Jan 30, 2009 at 9:35 PM, Richard Elling
richard.ell...@gmail.com wrote:
 For those who didn't follow down the thread this afternoon,
 I have posted a tool call zilstat which will help you to answer
 the question of whether a separate log might help your
 workload.  Details start here:
 http://richardelling.blogspot.com/2009/01/zilstat.html

 Enjoy!
  -- richard

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool status -x strangeness

2009-01-30 Thread Blake
Maybe ZFS hasn't seen an error in a long enough time that it considers
the pool healthy?  You could try clearing the pool and then observing.



On Wed, Jan 28, 2009 at 9:40 AM, Ben Miller mil...@eecis.udel.edu wrote:
 # zpool status -xv
 all pools are healthy

 Ben

 What does 'zpool status -xv' show?

 On Tue, Jan 27, 2009 at 8:01 AM, Ben Miller
 mil...@eecis.udel.edu wrote:
  I forgot the pool that's having problems was
 recreated recently so it's already at zfs version 3.
 I just did a 'zfs upgrade -a' for another pool, but
 some of those filesystems failed since they are busy
  and couldn't be unmounted.

  # zfs upgrade -a
  cannot unmount '/var/mysql': Device busy
  cannot unmount '/var/postfix': Device busy
  
  6 filesystems upgraded
  821 filesystems already at this version
 
  Ben
 
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting assistance.

2009-01-27 Thread Blake
I guess you could try 'zpool import -f'.  This is a pretty odd status,
I think.  I'm pretty sure raidz1 should survive a single disk failure.

Perhaps a more knowledgeable list member can explain.

On Sat, Jan 24, 2009 at 12:48 PM, Brad Hill b...@thosehills.com wrote:
 I've seen reports of a recent Seagate firmware update
 bricking drives again.

 What's the output of 'zpool import' from the LiveCD?
  It sounds like
 ore than 1 drive is dropping off.


 r...@opensolaris:~# zpool import
  pool: tank
id: 16342816386332636568
  state: FAULTED
 status: The pool was last accessed by another system.
 action: The pool cannot be imported due to damaged devices or data.
The pool may be active on another system, but can be imported using
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-EY
 config:

tankFAULTED  corrupted data
  raidz1DEGRADED
c6t0d0  ONLINE
c6t1d0  ONLINE
c6t2d0  ONLINE
c6t3d0  UNAVAIL  cannot open
c6t4d0  ONLINE

  pool: rpool
id: 9891756864015178061
  state: ONLINE
 status: The pool was last accessed by another system.
 action: The pool can be imported using its name or numeric identifier and
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-EY
 config:

rpool   ONLINE
  c3d0s0ONLINE
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Unable to destory a pool

2009-01-27 Thread Blake
Can you share the output of 'uname -a' and the disk controller you are using?

On Sun, Jan 25, 2009 at 6:24 PM, Ramesh Mudradi rameshm.ku...@gmail.com wrote:
 # zpool list
 NAME SIZE   USED  AVAILCAP  HEALTH  ALTROOT
 jira-app-zpool   272G   330K   272G 0%  ONLINE  -

 The following command hangs forever. If I reboot the box , zpool list shows 
 online as I mentioned the output above.

 # zpool destroy -f jira-app-zpool

 How can get rid of this pool and any reference to it.

 bash-3.00# zpool status
  pool: jira-app-zpool
  state: UNAVAIL
 status: One or more devices are faultd in response to IO failures.
 action: Make sure the affected devices are connected, then run 'zpool clear'.
   see: http://www.sun.com/msg/ZFS-8000-HC
  scrub: none requested
 config:

NAMESTATE READ WRITE CKSUM
jira-app-zpool  UNAVAIL  0 0 4  insufficient replicas
  c3t0d3FAULTED  0 0 4  experienced I/O failures

 errors: 2 data errors, use '-v' for a list
 bash-3.00#
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing HDD in x4500

2009-01-27 Thread Blake
I'm not an authority, but on my 'vanilla' filer, using the same
controller chipset as the thumper, I've been in really good shape
since moving to zfs boot in 10/08 and doing 'zpool upgrade' and 'zfs
upgrade' to all my mirrors (3 3-way).  I'd been having similar
troubles to yours in the past.

My system is pretty puny next to yours, but it's been reliable now for
slightly over a month.


On Tue, Jan 27, 2009 at 12:19 AM, Jorgen Lundman lund...@gmo.jp wrote:

 The vendor wanted to come in and replace an HDD in the 2nd X4500, as it
 was constantly busy, and since our x4500 has always died miserably in
 the past when a HDD dies, they wanted to replace it before the HDD
 actually died.

 The usual was done, HDD replaced, resilvering started and ran for about
 50 minutes. Then the system hung, same as always, all ZFS related
 commands would just hang and do nothing. System is otherwise fine and
 completely idle.

 The vendor for some reason decided to fsck root-fs, not sure why as it
 is mounted with logging, and also decided it would be best to do so
 from a CDRom boot.

 Anyway, that was 12 hours ago and the x4500 is still down. I think they
 have it at single-user prompt resilvering again. (I also noticed they'd
 decided to break the mirror of the root disks for some very strange
 reason). It still shows:

   raidz1  DEGRADED 0 0 0
 c0t1d0ONLINE   0 0 0
 replacing UNAVAIL  0 0 0  insufficient replicas
   c1t1d0s0/o  OFFLINE  0 0 0
   c1t1d0  UNAVAIL  0 0 0  cannot open

 So I am pretty sure it'll hang again sometime soon. What is interesting
 though is that this is on x4500-02, and all our previous troubles mailed
 to the list was regarding our first x4500. The hardware is all
 different, but identical. Solaris 10 5/08.

 Anyway, I think they want to boot CDrom to fsck root again for some
 reason, but since customers have been without their mail for 12 hours,
 they can go a little longer, I guess.

 What I was really wondering, has there been any progress or patches
 regarding the system always hanging whenever a HDD dies (or is replaced
 it seems). It really is rather frustrating.

 Lund

 --
 Jorgen Lundman   | lund...@lundman.net
 Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
 Shibuya-ku, Tokyo| +81 (0)90-5578-8500  (cell)
 Japan| +81 (0)3 -3375-1767  (home)
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool status -x strangeness

2009-01-27 Thread Blake
What does 'zpool status -xv' show?

On Tue, Jan 27, 2009 at 8:01 AM, Ben Miller mil...@eecis.udel.edu wrote:
 I forgot the pool that's having problems was recreated recently so it's 
 already at zfs version 3.  I just did a 'zfs upgrade -a' for another pool, 
 but some of those filesystems failed since they are busy and couldn't be 
 unmounted.

 # zfs upgrade -a
 cannot unmount '/var/mysql': Device busy
 cannot unmount '/var/postfix': Device busy
 
 6 filesystems upgraded
 821 filesystems already at this version

 Ben

 You can upgrade live.  'zfs upgrade' with no
 arguments shows you the
 zfs version status of filesystems present without
 upgrading.



 On Jan 24, 2009, at 10:19 AM, Ben Miller
 mil...@eecis.udel.edu wrote:

  We haven't done 'zfs upgrade ...' any.  I'll give
 that a try the
  next time the system can be taken down.
 
  Ben
 
  A little gotcha that I found in my 10u6 update
  process was that 'zpool
  upgrade [poolname]' is not the same as 'zfs
 upgrade
  [poolname]/[filesystem(s)]'
 
  What does 'zfs upgrade' say?  I'm not saying this
 is
  the source of
  your problem, but it's a detail that seemed to
 affect
  stability for
  me.
 
 
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


  1   2   >