Re: [zfs-discuss] reboot when copying large amounts of data

2009-12-20 Thread arnaud

Hi folks,
 I was trying to load a large file in /tmp so that a process that 
parses it wouldn't be limited by a disk throughput bottleneck. My rig 
here has only 12GB of RAM and the file I copied is about 12GB as well.

Before the copied finished, my system restarted.
I'm pretty up to date, the system has b129 running.

Googling about this I stumbled upon this thread dating back to March ...
http://mail.opensolaris.org/pipermail/zfs-discuss/2009-March/027264.html

Since this seems to be a known issue, and pretty serious in my book, is 
there a fix pending or is that not investigated ?


thanks in advance and happy holidays to all
-=arnaud=-


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-12-20 Thread Ian Collins

arnaud wrote:

Hi folks,
 I was trying to load a large file in /tmp so that a process that 
parses it wouldn't be limited by a disk throughput bottleneck. My rig 
here has only 12GB of RAM and the file I copied is about 12GB as well.

Before the copied finished, my system restarted.
I'm pretty up to date, the system has b129 running.

12GB of ECC RAM? 

This kind of spontaneous reboot with crash dumps enabled (do you have 
them enabled?) tend to be hardware related.  The most common bit of bad 
hardware is RAM...


--
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-14 Thread James C. McPherson
On Fri, 13 Mar 2009 22:46:19 -0400
Miles Nordin car...@ivy.net wrote:

  jcm == James C McPherson james.mcpher...@sun.com writes:
 
jcm We haven't got mega_sas on SPARC at this point either.
 
 The card Blake found:
 
  http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm
  
 http://www.lsi.com/storage_home/products_home/host_bus_adapters/sas_hbas/lsisas3080xr/index.html
 
 any idea if that's got a shot of working on SPARC (with mpt i presume,
 or is there a fourth lsi driver)?  or is it impossible to tell the
 chip/firmware of 3080X?  or is it not going to happen period without
 Forth firmware?

The lsi.com has a link to their driver for that card, which
is itmpt. That *generally* means that it'll work fine with 
the mpt driver - on either sparc or x86/x64.

I've not come across that particular model before, so I can't
say for sure. 

As for will it work on sparc - yes, I would expect so. BUT
without fcode it won't be bootable. That's what you are really
asking, isn't it? It turns out you can change between IT and
IR firmware using LSI's sasflash utility, and if I'm reading
other parts of 
http://www.lsi.com/DistributionSystem/AssetDocument/support/downloads/SASFlash_Readme.txt
 correctly, then you should be able to flash
fcode onto the card if it isn't there already. Have a look at
the other downloads on the LSI url you mentioned.

But no, sorry, no guarantees.

James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-14 Thread Miles Nordin
 jcm == James C McPherson james.mcpher...@sun.com writes:

   jcm As for will it work on sparc - yes, I would expect so. BUT
   jcm without fcode it won't be bootable. That's what you are really
   jcm asking, isn't it?

just if it will work.  I want to add a slog.

   jcm It turns out you can change between IT and IR firmware using
   jcm LSI's sasflash utility,

is there a reason to use one mode or the other?  itmpt sounds like it
only works with IT but that's a WAG and probably wrong.  Is there a
way to tell which mode mpt is using right now?  Does mpt behave
differently depending on mode?  

Thanks for the pointer to raidctl---I got this out of it:

-8-
ja...@dharavi:~$ pfexec raidctl -l -g 0.2.0 2
DiskVendor  Product FirmwareCapacityStatus  HSP

0.2.0   ATA WDC WD10EADS-00 1A01931.5G  GOODN/A
GUID:50014ee202720bc0
ja...@dharavi:~$ pfexec raidctl -l 2
Controller  TypeVersion

c2  LSI_1068E   1.22.01.00
ja...@dharavi:~$ 
-8-

I was hoping to see an -IT or -IR suffix on the version number but
there isn't one.  In the README for sasflash, it looks like SASFLASH
will not tell you what kind of firmware you currently have, either,
just let you choose what kind to load.

In the ^C peecee boot widget, it shows my card's firmware version
ending in -IT, so I think I have IT firmware.  If I do, then Will's
speculation that IT means the card acts as a SATA target is an
incorrect guess because it's acting like a normal card right now.

but on SPARC I'd have no way of seeing this -IT suffix, and I don't
know if I trust it anyway without understanding what the different
versions are supposed to do differently.  so, is there any way to tell
from mpt in what mode the card's running, or what the consequence of
that is?

It sounds like it's possible one could buy a 1068E card from another
vendor and end up with -IR firmware instead, without changing chip
number or necessarily even PCI ID, so if we are trying to predict
which cards behave how, it's worth trying to figure out.

I did try to RTFM.  On OpenSolaris I get this loveliness:

ja...@dharavi:~$ man mpt
No manual entry for mpt.

on SXCE I do at least have a man page, but there's no mention of IT,
IR, SR inside it.

   jcm Have a look at the other downloads on the LSI url you
   jcm mentioned.

oops.  yeah, a proprietary sparc driver is there!  That's not exactly,
pow, we're definitely done, because their driver is for Sol10 not
nevada.  But hopefully that means the card will not need some
firmware-provided ``initialization'' and might work with the sparc mpt
included in Nevada!

As for the closed-source LSI tool 'lsiutil' mentioned in this
thread:

 http://www.opensolaris.org/jive/message.jspa?messageID=184811#184811

  A hidden undocumented option 15 - option 12 in lsiutil, to make
  drive mapping based on port number rather than some
  LSI-GUID/drive-serial as it supposedly is by default.  My
  AOC-USAS-L8i card is already behaving port-based, though: Solaris
  device names corresponding to physical ports even when I move disks
  around.  I guess it's other LSI cards that have the problem
  mentioned in the thread.

there is a lsiutil binary inside the SPARC itmpt_sparc_5.07.04.zip
file, so hopefully it's the same one the thread says is hard to find.

I guess I'm a little closer to understanding the LSI cards, though I
still don't know for sure what chips are on what card models (one can
sort of guess), nor what IT, IR, SR mean.  I'm still kind of pissed to
be passing around warez too.  Option 15, 12 wouldn't be secret if we
had source to lsiutil.  We would have less problems with whether the
driver is for Sol10 or Nevada.  The firmware on this card is way too
big, and there does not seem to be any SAS card for less than
$100/port with an open-source driver that works well.


pgpcpqGMhNZwQ.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-14 Thread Tim
On Sat, Mar 14, 2009 at 12:57 PM, Miles Nordin car...@ivy.net wrote:

  jcm == James C McPherson james.mcpher...@sun.com writes:

jcm As for will it work on sparc - yes, I would expect so. BUT
   jcm without fcode it won't be bootable. That's what you are really
   jcm asking, isn't it?

 just if it will work.  I want to add a slog.

   jcm It turns out you can change between IT and IR firmware using
   jcm LSI's sasflash utility,

 is there a reason to use one mode or the other?  itmpt sounds like it
 only works with IT but that's a WAG and probably wrong.  Is there a
 way to tell which mode mpt is using right now?  Does mpt behave
 differently depending on mode?

 Thanks for the pointer to raidctl---I got this out of it:

 -8-
 ja...@dharavi:~$ pfexec raidctl -l -g 0.2.0 2
 DiskVendor  Product FirmwareCapacityStatus  HSP

 
 0.2.0   ATA WDC WD10EADS-00 1A01931.5G  GOODN/A
 GUID:50014ee202720bc0
 ja...@dharavi:~$ pfexec raidctl -l 2
 Controller  TypeVersion
 
 c2  LSI_1068E   1.22.01.00
 ja...@dharavi:~$
 -8-

 I was hoping to see an -IT or -IR suffix on the version number but
 there isn't one.  In the README for sasflash, it looks like SASFLASH
 will not tell you what kind of firmware you currently have, either,
 just let you choose what kind to load.

 In the ^C peecee boot widget, it shows my card's firmware version
 ending in -IT, so I think I have IT firmware.  If I do, then Will's
 speculation that IT means the card acts as a SATA target is an
 incorrect guess because it's acting like a normal card right now.

 but on SPARC I'd have no way of seeing this -IT suffix, and I don't
 know if I trust it anyway without understanding what the different
 versions are supposed to do differently.  so, is there any way to tell
 from mpt in what mode the card's running, or what the consequence of
 that is?

 It sounds like it's possible one could buy a 1068E card from another
 vendor and end up with -IR firmware instead, without changing chip
 number or necessarily even PCI ID, so if we are trying to predict
 which cards behave how, it's worth trying to figure out.

 I did try to RTFM.  On OpenSolaris I get this loveliness:

 ja...@dharavi:~$ man mpt
 No manual entry for mpt.

 on SXCE I do at least have a man page, but there's no mention of IT,
 IR, SR inside it.


IR==integrated RAID.  This is a stripped down version of raid.  It only does
1/0/1E.  No 5 or 6.  See:
http://www.lsi.com/storage_home/products_home/host_bus_adapters/software_solutions/integrated_raid/

IT==initiator. In other words it should function as a regular non-RAID
adapter or as a target.  Initiator vs. target depends entirely on the driver
bound to the card.  See:
http://www.supermicro.com/manuals/other/LSI%20MegaRAID_Configuration_for_the_LSI_1068_Controller.pdf

SR==software RAID.  I believe this is nearly the same as IR mode, but
Supermicro for instance has an add-in card that will allow it to do RAID-5
as well.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Blake Irvin
This is really great information, though most of the controllers  
mentioned aren't on the OpenSolaris HCL.  Seems like that should be  
corrected :)


My thanks to the community for their support.

On Mar 12, 2009, at 10:42 PM, James C. McPherson james.mcpher...@sun.com 
 wrote:



On Thu, 12 Mar 2009 22:24:12 -0400
Miles Nordin car...@ivy.net wrote:


wm == Will Murnane will.murn...@gmail.com writes:



* SR = Software RAID IT = Integrate. Target mode. IR mode
is not supported.

   wm Integrated target mode lets you export some storage attached
   wm to the host system (through another adapter, presumably) as a
   wm storage device.  IR mode is almost certainly Internal RAID,
   wm which that card doesn't have support for.

no, the supermicro page for AOC-USAS-L8i does claim support for all
three, and supermicro has an ``IR driver'' available for download for
Linux and Windows, or at least a link to one.

I'm trying to figure out what's involved in determining and switching
modes, why you'd want to switch them, what cards support which modes,
which solaris drivers support which modes, u.s.w.

The answer may be very simple, like ``the driver supports only IR.
Most cards support IR, and cards that don't support IR won't work.   
IR

can run in single-LUN mode.  Some IR cards support RAID5, others
support only RAID 0, 1, 10.''  Or it could be ``the driver supports
only SR.  The driver is what determines the mode, and it does this by
loading firmware into the card, and the first step in initializing  
the

card is always for the driver to load in a firmware blob.  All
currently-produced cards support SR.''  so...actually, now that I say
it, I guess the answer cannot be very simple.  It's going to have to
be a little complicated.
Anyway, I can guess, too.  I was hoping someone would know for sure
off-hand.



Hi Miles,
the mpt(7D) driver supports that card. mpt(7D) supports both
IT and IR firmware variants. You can find out the specifics
for what RAID volume levels are supported by reading the
raidctl(1M) manpage. I don't think you can switch between IT
and IR firmware, but not having needed to know this before,
I haven't tried it.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcphttp://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Miles Nordin
 jcm == James C McPherson james.mcpher...@sun.com writes:

   jcm the mpt(7D) driver supports that card.

Then I am apparently stuck with a closed-source driver again, and
again by surprise.  I bought it because I thought you said 1068E was
supported by mega_sas:

 http://www.osnews.com/thread?317113

   jcm The driver for LSI's MegaRAID SAS card is mega_sas which was
   jcm integrated into snv_88. It's planned for backporting to a
   jcm Solaris 10 update.

but apparently it was wishful thinking on my part:

   jcm There are several LSI cards which use the 1068 and 1068E chip.
   jcm Some of these use mpt(7d), some use mega_sas(7d). It all
   jcm depends on the firmware of the card, basically. You could also
   jcm have a look at the PCI IDs database at
   jcm http://pciids.sourceforge.net to see what the name to pci
   jcm vid/did mapping is. That provides a fairly good indicator of
   jcm whether you'll need mpt(7d) or mega_sas(7d).

I downloaded the pci.ids file from sourceforge, but I do not
understand how to tell which cards have the proprietary closed-source
driver, even if I somehow know the PCI ID of the card before I buy it?

(pci.ids lists only chip number in the description, and (a) LSI does
not disclose the chip-to-card model association, and apparently the
same chip can have different drivers so chip number isn't enough).  

Is there some file in OpenSolaris against which I can cross-reference
this?  or...really, just use instead of pci.ids, since only the PCI ID
not the description is enough to uniquely identify the card.

Is it possible to change the firmware and thus the driver binding
without changing the PCI ID so that I have to worry about
manufacturers doing that, or can I really count on the PCI ID alone to
tell me which driver will run the card?  I'm worried that for example
adding an iButton to unlock RAID5 on a supermicro card will change the
driver attachment.

Also does anyone know which cards work on SPARC?  None?  All?  I know
the SATA framework is x86 only, but AIUI none of the LSI cards
actually use the SATA framework (so, presumably things like
hd/smartmontools will not work, but at least the card has got a better
shot of working on SPARC).  I cannot test the AOC-USAS-L8i card I have
in SPARC because it is PCIe, and I have only a PCI-X SPARC.

Does anyone have an affordable card which is working well with the
open-source mega_sas driver?  It seems we are still without any
open-source SATA card that works well.


pgpSCd24rqLXT.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Miles Nordin
 c == Miles Nordin car...@ivy.net writes:

 c Is there some file in OpenSolaris against which I can
 c cross-reference this?  or...really, just use instead of
 c pci.ids, since only the PCI ID not the description is enough

I found these:

 
http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/os/driver_aliases
 
http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/sparc/os/driver_aliases

but they don't seem to be complete.  mega_sas is not listed in either.


pgpCbCV2kLAON.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Tim
On Fri, Mar 13, 2009 at 3:33 PM, Miles Nordin car...@ivy.net wrote:

  c == Miles Nordin car...@ivy.net writes:

 c Is there some file in OpenSolaris against which I can
 c cross-reference this?  or...really, just use instead of
 c pci.ids, since only the PCI ID not the description is enough

 I found these:


 http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/os/driver_aliases

 http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/sparc/os/driver_aliases

 but they don't seem to be complete.  mega_sas is not listed in either.


You can manually create an entry in /etc/driver_aliases to force a driver to
bind to a specific card, regardless of firmware.  Whether it will work or
not isn't apparent.  What is apparent is that it won't be supported by
anyone though :)

ANYWAYS, a bit of research goes a long ways :)

You'll find the various pci device ID's towards the bottom.
http://src.opensolaris.org/source/xref/zfs-crypto/phase2/usr/src/pkgdefs/SUNWmegasas/postinstall?r=6542

Looks like there's a TON of supported cards.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Tim
On Fri, Mar 13, 2009 at 3:33 PM, Miles Nordin car...@ivy.net wrote:

  c == Miles Nordin car...@ivy.net writes:

 c Is there some file in OpenSolaris against which I can
 c cross-reference this?  or...really, just use instead of
 c pci.ids, since only the PCI ID not the description is enough

 I found these:


 http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/os/driver_aliases

 http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/sparc/os/driver_aliases

 but they don't seem to be complete.  mega_sas is not listed in either.


Oh, and for people too lazy to google, here's a start:
http://pci-ids.ucw.cz/read/PC/1000/0060
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Miles Nordin
 t == Tim  t...@tcsac.net writes:

 t 
http://src.opensolaris.org/source/xref/zfs-crypto/phase2/usr/src/pkgdefs/SUNWmegasas/postinstall?r=6542

thanks.

 t Looks like there's a TON of supported cards.

They are all 1078 cards though.  James mentioned mega_sas supports
some 1068E cards depending on the firmware?

The 1078 seem to have rather large built-in write caches and thus be
expensive and unnecessary for people who want to do raidz/slog.  For
example a lot of the cardsd are PERC with 4 ports for $500---costs
more than the drives you plug it into.  Are there any cheap cards for
mega_sas?

any word on which cards work on SPARC?

 t http://pci-ids.ucw.cz/read/PC/1000/0060

or else the whole list at pciids.sourceforge.net in the mail i quoted


pgpppwfuNOQQF.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread James C. McPherson
On Fri, 13 Mar 2009 18:00:04 -0400
Miles Nordin car...@ivy.net wrote:

  t == Tim  t...@tcsac.net writes:
 
  t 
 http://src.opensolaris.org/source/xref/zfs-crypto/phase2/usr/src/pkgdefs/SUNWmegasas/postinstall?r=6542
 
 thanks.
 
  t Looks like there's a TON of supported cards.
 
 They are all 1078 cards though.  James mentioned mega_sas supports
 some 1068E cards depending on the firmware?

Yes, that's correct. Not very convenient, I'm afraid.
 
 The 1078 seem to have rather large built-in write caches and thus be
 expensive and unnecessary for people who want to do raidz/slog.  For
 example a lot of the cardsd are PERC with 4 ports for $500---costs
 more than the drives you plug it into.  Are there any cheap cards for
 mega_sas?

I don't know, sorry. I've only ever seen Sun and Dell-branded
MegaRAID SAS cards, and I have no idea about pricing.

 any word on which cards work on SPARC?
 
  t http://pci-ids.ucw.cz/read/PC/1000/0060
 
 or else the whole list at pciids.sourceforge.net in the mail i quoted

We haven't got mega_sas on SPARC at this point either.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-13 Thread Miles Nordin
 jcm == James C McPherson james.mcpher...@sun.com writes:

   jcm We haven't got mega_sas on SPARC at this point either.

The card Blake found:

 http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm
 
http://www.lsi.com/storage_home/products_home/host_bus_adapters/sas_hbas/lsisas3080xr/index.html

any idea if that's got a shot of working on SPARC (with mpt i presume,
or is there a fourth lsi driver)?  or is it impossible to tell the
chip/firmware of 3080X?  or is it not going to happen period without
Forth firmware?


pgp2DuxsyGlCt.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Blake
I start the cp, and then, with prstat -a, watch the cpu load for the
cp process climb to 25% on a 4-core machine.

Load, measured for example with 'uptime', climbs steadily until the reboot.

Note that the machine does not dump properly, panic or hang - rather,
it reboots.

I attached a screenshot earlier in this thread of the little bit of
error message I could see on the console.  The machine is trying to
dump to the dump zvol, but fails to do so.  Only sometimes do I see an
error on the machine's local console - mos times, it simply reboots.



On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:
 Hm -

 Crashes, or hangs? Moreover - how do you know a CPU is pegged?

 Seems like we could do a little more discovery on what the actual problem
 here is, as I can read it about 4 different ways.

 By this last piece of information, I'm guessing the system does not crash,
 but goes really really slow??

 Crash == panic == we see stack dump on console and try to take a dump
 hang == nothing works == no response - might be worth looking at mdb -K
        or booting with a -k on the boot line.

 So - are we crashing, hanging, or something different?

 It might simply be that you are eating up all your memory, and your physical
 backing storage is taking a while to catch up?

 Nathan.

 Blake wrote:

 My dump device is already on a different controller - the motherboards
 built-in nVidia SATA controller.

 The raidz2 vdev is the one I'm having trouble with (copying the same
 files to the mirrored rpool on the nVidia controller work nicely).  I
 do notice that, when using cp to copy the files to the raidz2 pool,
 load on the machine climbs steadily until the crash, and one proc core
 pegs at 100%.

 Frustrating, yes.

 On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
 maidakalexand...@johndeere.com wrote:

 If you're having issues with a disk contoller or disk IO driver its
 highly likely that a savecore to disk after the panic will fail.  I'm not
 sure how to work around this, maybe a dedicated dump device not on a
 controller that uses a different driver then the one that you're having
 issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore to
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out of
 space on the dump device:

 r...@host:~# dumpadm
     Dump content: kernel pages
      Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the system
 load was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
 richard.ell...@gmail.com wrote:

 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.


 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for
 swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 --
 //
 // Nathan Kroenert              nathan.kroen...@sun.com         //
 // Systems Engineer             Phone:  +61 3 9869-6255         //
 // Sun Microsystems             Fax:    +61 3 9869-6288         //
 // Level 7, 476 St. Kilda Road  Mobile: 0419 305 456            //
 // Melbourne 3004   Victoria    Australia                       //
 //

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Nathan Kroenert

definitely time to bust out some mdb -k and see what it's moaning about.

I did not see the screenshot earlier... sorry about that.

Nathan.

Blake wrote:

I start the cp, and then, with prstat -a, watch the cpu load for the
cp process climb to 25% on a 4-core machine.

Load, measured for example with 'uptime', climbs steadily until the reboot.

Note that the machine does not dump properly, panic or hang - rather,
it reboots.

I attached a screenshot earlier in this thread of the little bit of
error message I could see on the console.  The machine is trying to
dump to the dump zvol, but fails to do so.  Only sometimes do I see an
error on the machine's local console - mos times, it simply reboots.



On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:

Hm -

Crashes, or hangs? Moreover - how do you know a CPU is pegged?

Seems like we could do a little more discovery on what the actual problem
here is, as I can read it about 4 different ways.

By this last piece of information, I'm guessing the system does not crash,
but goes really really slow??

Crash == panic == we see stack dump on console and try to take a dump
hang == nothing works == no response - might be worth looking at mdb -K
   or booting with a -k on the boot line.

So - are we crashing, hanging, or something different?

It might simply be that you are eating up all your memory, and your physical
backing storage is taking a while to catch up?

Nathan.

Blake wrote:

My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:

If you're having issues with a disk contoller or disk IO driver its
highly likely that a savecore to disk after the panic will fail.  I'm not
sure how to work around this, maybe a dedicated dump device not on a
controller that uses a different driver then the one that you're having
issues with?

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
Sent: Wednesday, March 11, 2009 4:45 PM
To: Richard Elling
Cc: Marc Bevand; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] reboot when copying large amounts of data

I guess I didn't make it clear that I had already tried using savecore to
retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out of
space on the dump device:

r...@host:~# dumpadm
Dump content: kernel pages
 Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
directory: /var/crash/host
 Savecore enabled: yes

I was using the -L option only to try to get some idea of why the system
load was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
richard.ell...@gmail.com wrote:

Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


I'm working on testing this some more by doing a savecore -L right
after I start the copy.



savecore -L is not what you want.

By default, for OpenSolaris, savecore on boot is disabled.  But the
core will have been dumped into the dump slice, which is not used for
swap.
So you should be able to run savecore at a later time to collect the
core from the last dump.
-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

--
//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Systems Engineer Phone:  +61 3 9869-6255 //
// Sun Microsystems Fax:+61 3 9869-6288 //
// Level 7, 476 St. Kilda Road  Mobile: 0419 305 456//
// Melbourne 3004   VictoriaAustralia   //
//



--
//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Systems Engineer Phone:  +61 3 9869-6255 //
// Sun Microsystems Fax:+61 3 9869-6288 //
// Level 7, 476 St. Kilda Road  Mobile: 0419 305 456//
// Melbourne 3004   VictoriaAustralia

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Nathan Kroenert
definitely time to bust out some mdb -K or boot -k and see what it's 
moaning about.


I did not see the screenshot earlier... sorry about that.

Nathan.

Blake wrote:

I start the cp, and then, with prstat -a, watch the cpu load for the
cp process climb to 25% on a 4-core machine.

Load, measured for example with 'uptime', climbs steadily until the reboot.

Note that the machine does not dump properly, panic or hang - rather,
it reboots.

I attached a screenshot earlier in this thread of the little bit of
error message I could see on the console.  The machine is trying to
dump to the dump zvol, but fails to do so.  Only sometimes do I see an
error on the machine's local console - mos times, it simply reboots.



On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:

Hm -

Crashes, or hangs? Moreover - how do you know a CPU is pegged?

Seems like we could do a little more discovery on what the actual problem
here is, as I can read it about 4 different ways.

By this last piece of information, I'm guessing the system does not crash,
but goes really really slow??

Crash == panic == we see stack dump on console and try to take a dump
hang == nothing works == no response - might be worth looking at mdb -K
   or booting with a -k on the boot line.

So - are we crashing, hanging, or something different?

It might simply be that you are eating up all your memory, and your physical
backing storage is taking a while to catch up?

Nathan.

Blake wrote:

My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:

If you're having issues with a disk contoller or disk IO driver its
highly likely that a savecore to disk after the panic will fail.  I'm not
sure how to work around this, maybe a dedicated dump device not on a
controller that uses a different driver then the one that you're having
issues with?

-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
Sent: Wednesday, March 11, 2009 4:45 PM
To: Richard Elling
Cc: Marc Bevand; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] reboot when copying large amounts of data

I guess I didn't make it clear that I had already tried using savecore to
retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out of
space on the dump device:

r...@host:~# dumpadm
Dump content: kernel pages
 Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
directory: /var/crash/host
 Savecore enabled: yes

I was using the -L option only to try to get some idea of why the system
load was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
richard.ell...@gmail.com wrote:

Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


I'm working on testing this some more by doing a savecore -L right
after I start the copy.



savecore -L is not what you want.

By default, for OpenSolaris, savecore on boot is disabled.  But the
core will have been dumped into the dump slice, which is not used for
swap.
So you should be able to run savecore at a later time to collect the
core from the last dump.
-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

--
//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Systems Engineer Phone:  +61 3 9869-6255 //
// Sun Microsystems Fax:+61 3 9869-6288 //
// Level 7, 476 St. Kilda Road  Mobile: 0419 305 456//
// Melbourne 3004   VictoriaAustralia   //
//



--
//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Systems Engineer Phone:  +61 3 9869-6255 //
// Sun Microsystems Fax:+61 3 9869-6288 //
// Level 7, 476 St. Kilda Road  Mobile: 0419 305 456//
// Melbourne 3004   VictoriaAustralia

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Blake
So, if I boot with the -k boot flags (to load the kernel debugger?)
what do I need to look for?  I'm no expert at kernel debugging.

I think this is a pci error judging by the console output, or at least
is i/o related...

thanks for your feedback,
Blake

On Thu, Mar 12, 2009 at 2:18 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:
 definitely time to bust out some mdb -K or boot -k and see what it's moaning
 about.

 I did not see the screenshot earlier... sorry about that.

 Nathan.

 Blake wrote:

 I start the cp, and then, with prstat -a, watch the cpu load for the
 cp process climb to 25% on a 4-core machine.

 Load, measured for example with 'uptime', climbs steadily until the
 reboot.

 Note that the machine does not dump properly, panic or hang - rather,
 it reboots.

 I attached a screenshot earlier in this thread of the little bit of
 error message I could see on the console.  The machine is trying to
 dump to the dump zvol, but fails to do so.  Only sometimes do I see an
 error on the machine's local console - mos times, it simply reboots.



 On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
 nathan.kroen...@sun.com wrote:

 Hm -

 Crashes, or hangs? Moreover - how do you know a CPU is pegged?

 Seems like we could do a little more discovery on what the actual problem
 here is, as I can read it about 4 different ways.

 By this last piece of information, I'm guessing the system does not
 crash,
 but goes really really slow??

 Crash == panic == we see stack dump on console and try to take a dump
 hang == nothing works == no response - might be worth looking at mdb -K
       or booting with a -k on the boot line.

 So - are we crashing, hanging, or something different?

 It might simply be that you are eating up all your memory, and your
 physical
 backing storage is taking a while to catch up?

 Nathan.

 Blake wrote:

 My dump device is already on a different controller - the motherboards
 built-in nVidia SATA controller.

 The raidz2 vdev is the one I'm having trouble with (copying the same
 files to the mirrored rpool on the nVidia controller work nicely).  I
 do notice that, when using cp to copy the files to the raidz2 pool,
 load on the machine climbs steadily until the crash, and one proc core
 pegs at 100%.

 Frustrating, yes.

 On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
 maidakalexand...@johndeere.com wrote:

 If you're having issues with a disk contoller or disk IO driver its
 highly likely that a savecore to disk after the panic will fail.  I'm
 not
 sure how to work around this, maybe a dedicated dump device not on a
 controller that uses a different driver then the one that you're having
 issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore
 to
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out
 of
 space on the dump device:

 r...@host:~# dumpadm
    Dump content: kernel pages
     Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the
 system
 load was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
 richard.ell...@gmail.com wrote:

 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com
 wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.


 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for
 swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 --
 //
 // Nathan Kroenert              nathan.kroen...@sun.com         //
 // Systems Engineer             Phone:  +61 3 9869-6255         //
 // Sun Microsystems             Fax:    +61 3 9869-6288         //
 // Level 7, 476 St. Kilda Road  Mobile: 0419 305 456            //
 // Melbourne 3004   Victoria    Australia

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Miles Nordin
 maj == Maidak Alexander J maidakalexand...@johndeere.com writes:

   maj If you're having issues with a disk contoller or disk IO
   maj driver its highly likely that a savecore to disk after the
   maj panic will fail.  I'm not sure how to work around this

not in Solaris, but as a concept for solving the problem:

 
http://git.kernel.org/?p=linux/kernel/git/torvalds/linux-2.6.git;a=blob;f=Documentation/kdump/kdump.txt;h=3f4bc840da8b7c068076dd057216e846e098db9f;hb=4a6908a3a050aacc9c3a2f36b276b46c0629ad91

They load a second kernel into a reserved spot of RAM, like 64MB or
so, and forget about it.  After a crash, they boot the second kernel.
The second kernel runs using the reserved area of RAM as its working
space, not touching any other memory, as if you were running on a very
old machine with tiny RAM.  It reprobes all the hardware, and then
performs the dump.  I don't know if it actually works, but the
approach is appropriate if you are trying to debug the storage stack.
You could even have a main kernel which crashes while taking an
ordinary coredump, and then use the backup dumping-kernel to coredump
the main kernel in mid-coredump---a dump of a dumping kernel.

I think some Solaris developers were discussing putting coredump
features into Xen, so the host could take the dump (or, maybe even
something better than a dump---for example, if you built host/target
debugging features into Xen for debugging running kernels, then you
could just force a breakpoint in the guest instead of panic.  Since
Xen can hibernate domU's onto disk (it can, right?), you can treat the
hibernated Xen-specific representation of the domU as the-dump,
groveling through the ``dump'' with the same host/target tools you
could use on a running kernel without any special dump support in the
debugger itself).  IIRC NetBSD developers discussed the same idea
years ago but neither implementation exists.


pgpsmSOamFWH7.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Blake
I've managed to get the data transfer to work by rearranging my disks
so that all of them sit on the integrated SATA controller.

So, I feel pretty certain that this is either an issue with the
Supermicro aoc-sat2-mv8 card, or with PCI-X on the motherboard (though
I would think that the integrated SATA would also be using the PCI
bus?).

The motherboard, for those interested, is an HD8ME-2 (not, I now find
after buying this box from Silicon Mechanics, a board that's on the
Solaris HCL...)

http://www.supermicro.com/Aplus/motherboard/Opteron2000/MCP55/h8dme-2.cfm

So I'm not considering one of LSI's HBA's - what do list members think
about this device:

http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm



On Thu, Mar 12, 2009 at 2:18 AM, Nathan Kroenert
nathan.kroen...@sun.com wrote:
 definitely time to bust out some mdb -K or boot -k and see what it's moaning
 about.

 I did not see the screenshot earlier... sorry about that.

 Nathan.

 Blake wrote:

 I start the cp, and then, with prstat -a, watch the cpu load for the
 cp process climb to 25% on a 4-core machine.

 Load, measured for example with 'uptime', climbs steadily until the
 reboot.

 Note that the machine does not dump properly, panic or hang - rather,
 it reboots.

 I attached a screenshot earlier in this thread of the little bit of
 error message I could see on the console.  The machine is trying to
 dump to the dump zvol, but fails to do so.  Only sometimes do I see an
 error on the machine's local console - mos times, it simply reboots.



 On Thu, Mar 12, 2009 at 1:55 AM, Nathan Kroenert
 nathan.kroen...@sun.com wrote:

 Hm -

 Crashes, or hangs? Moreover - how do you know a CPU is pegged?

 Seems like we could do a little more discovery on what the actual problem
 here is, as I can read it about 4 different ways.

 By this last piece of information, I'm guessing the system does not
 crash,
 but goes really really slow??

 Crash == panic == we see stack dump on console and try to take a dump
 hang == nothing works == no response - might be worth looking at mdb -K
       or booting with a -k on the boot line.

 So - are we crashing, hanging, or something different?

 It might simply be that you are eating up all your memory, and your
 physical
 backing storage is taking a while to catch up?

 Nathan.

 Blake wrote:

 My dump device is already on a different controller - the motherboards
 built-in nVidia SATA controller.

 The raidz2 vdev is the one I'm having trouble with (copying the same
 files to the mirrored rpool on the nVidia controller work nicely).  I
 do notice that, when using cp to copy the files to the raidz2 pool,
 load on the machine climbs steadily until the crash, and one proc core
 pegs at 100%.

 Frustrating, yes.

 On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
 maidakalexand...@johndeere.com wrote:

 If you're having issues with a disk contoller or disk IO driver its
 highly likely that a savecore to disk after the panic will fail.  I'm
 not
 sure how to work around this, maybe a dedicated dump device not on a
 controller that uses a different driver then the one that you're having
 issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore
 to
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out
 of
 space on the dump device:

 r...@host:~# dumpadm
    Dump content: kernel pages
     Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the
 system
 load was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
 richard.ell...@gmail.com wrote:

 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com
 wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.


 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for
 swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Tim
On Thu, Mar 12, 2009 at 2:22 PM, Blake blake.ir...@gmail.com wrote:

 I've managed to get the data transfer to work by rearranging my disks
 so that all of them sit on the integrated SATA controller.

 So, I feel pretty certain that this is either an issue with the
 Supermicro aoc-sat2-mv8 card, or with PCI-X on the motherboard (though
 I would think that the integrated SATA would also be using the PCI
 bus?).

 The motherboard, for those interested, is an HD8ME-2 (not, I now find
 after buying this box from Silicon Mechanics, a board that's on the
 Solaris HCL...)

 http://www.supermicro.com/Aplus/motherboard/Opteron2000/MCP55/h8dme-2.cfm
 

 So I'm not considering one of LSI's HBA's - what do list members think
 about this device:

 http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htmhttp://www.provantage.com/lsi-logic-lsi00117%7E7LSIG03X.htm
 



I believe the MCP55's SATA controllers are actually PCI-E based.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Dave



Tim wrote:



On Thu, Mar 12, 2009 at 2:22 PM, Blake blake.ir...@gmail.com 
mailto:blake.ir...@gmail.com wrote:


I've managed to get the data transfer to work by rearranging my disks
so that all of them sit on the integrated SATA controller.

So, I feel pretty certain that this is either an issue with the
Supermicro aoc-sat2-mv8 card, or with PCI-X on the motherboard (though
I would think that the integrated SATA would also be using the PCI
bus?).

The motherboard, for those interested, is an HD8ME-2 (not, I now find
after buying this box from Silicon Mechanics, a board that's on the
Solaris HCL...)

http://www.supermicro.com/Aplus/motherboard/Opteron2000/MCP55/h8dme-2.cfm

So I'm not considering one of LSI's HBA's - what do list members think
about this device:

http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm
http://www.provantage.com/lsi-logic-lsi00117%7E7LSIG03X.htm



I believe the MCP55's SATA controllers are actually PCI-E based.


I use Tyan 2927 motherboards. They have on-board nVidia MCP55 chipsets, 
which is the same chipset at the X4500 (IIRC). I wouldn't trust the 
MCP55 chipset in OpenSolaris. I had random disk hangs even while the 
machine was mostly idle.


In Feb 2008 I bought AOC-SAT2-MV8 cards and moved all my drives to these 
add-in cards. I haven't had any issues with drive hanging since. There 
does not seem to be any problems with the SAT2-MV8 under heavy load in 
my servers from what I've seen.


When the SuperMicro AOC-USAS-L8i came out later last year, I started 
using them instead. They work better than the SAT2-MV8s.


This card needs a 3U or bigger case:
http://www.supermicro.com/products/accessories/addon/AOC-USAS-L8i.cfm

This is the low profile card that will fit in a 2U:
http://www.supermicro.com/products/accessories/addon/AOC-USASLP-L8i.cfm

They both work in normal PCI-E slots on my Tyan 2927 mobos.

Finding good non-Sun hardware that works very well under OpenSolaris is 
frustrating to say the least. Good luck.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Miles Nordin
 b == Blake  blake.ir...@gmail.com writes:

 b http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm

I'm having trouble matching up chips, cards, drivers, platforms, and
modes with the LSI stuff.  The more I look at it the mroe confused I
get.

Platforms:
 x86
 SPARC

Drivers:
 mpt
 mega_sas
 mfi

Chips:
 1068   (SAS, PCI-X)
 1068E  (SAS, PCIe)
 1078   ???  
   -- from supermicro, seems to be SAS, PCIe, with support for 256 -
  512MB RAM instead of the 16 - 32MB RAM on the others
 1030   (parallel scsi)

Cards:
  LSI cards 
http://www.lsi.com/storage_home/products_home/host_bus_adapters/sas_hbas/index.html
  I love the way they use the numbers 3800 and 3080, so you are
  constantly transposing them thus leaving google littered with all
  this confusingly wrong information.

LSISAS3800X(PCI-X, external ports)
LSISAS3080X-R  (PCI-X, internal ports)

LSISAS3801X(PCI-X, external ports)

LSISAS3801E(PCIe, external ports)
LSISAS3081E-R  (PCIe, internal ports)

  I would have thought -R meant ``suports RAID'' but all I can really
  glean through the foggy marketing-glass behind which all the
  information is hidden, is -R means ``all the ports are internal''.

  Supermicro cards http://www.supermicro.com/products/accessories/index.cfm
wow, this is even more of a mess.
 These are all UIO cards so I assume they have the PCIe bracket on backwards
AOC-USAS-L4i(PCIe, 4 internal 4 external)
AOC-USAS-L8i, AOC-USASLP-L8i(PCIe, internal ports)
   based on 1068E
   sounds similar to LSISAS3081E.  Is that also 1068E?
   supports RAID0, RAID1, RAID10
   AOC-USAS-L4iR
   identical to the above, but ``includes iButton''
   which is an old type of smartcard-like device with 
   sometimes crypto and javacard support.
   apparently some kind of license key to 
   unlock RAID5?  no L8iR exists though, only L4iR.
   I have the L8i, and it does have an iButton socket
   with no button in it.
AOC-USAS-H4iR
AOC-USAS-H8iR, AOC-USASLP-H8iR  (PCIe, internal ports)
   based on 1078
   low-profile version has more memory than fullsize version?!

but here is the most fun thing about the supermicro cards.  All
cards have one driver *EXCEPT* the L8i, which has three drivers
for three modes: IT, IR, and SR.  When I google for this I find
notes on some of their integrated motherboards like:

 * The onboard LSI 1068E supported SR and IT mode but not IR mode.

I also found this:

 * SR = Software RAID IT = Integrate. Target mode. IR mode is not supported.

but no idea what the three modes are.  searching for SAS SR IT IR
doesn't work either, so it's not some SAS thing.  What *is* it?

also there seem to be two different kinds of quad-SATA connector on
these SAS cards so there are two different kinds of octopus cable.

Questions:

 * which chips are used by each of the LSI boards?  I can guess, but
   in particular LSISAS3800X and LSISAS3801X seem to be different
   chips, while from the list of chips I'd have no choice but to guess
   they are both 1068.

 * which drivers work on x86 and which SPARC?  I know some LSI cards
   work in SPARC but maybe not all---do the drivers support the same
   set of cards on both platforms?  Or will normal cards not work in
   SPARC for lack of Forth firmware to perform some LSI-proprietary
   ``initialization'' ritual?

 * which chips go with which drivers?  Is it even that simple---will
   adding an iButton RAID5 license to a SuperMicro board make the same
   card change from mega_sas to mpt attachment, or something similar?

   For example there is a bug here about a 1068E card which doesn't
   work, even though most 1068E cards do work:

http://bugs.opensolaris.org/view_bug.do?bug_id=6736187

   Maybe the Solaris driver needs IR mode and won't work with the
   onboard supermicro chip which supports only ``software raid''
   whatever that means, which is maybe denoted by SR?  What does the
   iButton unlock, then, features of IR mode which are abstracted from
   the OS driver?

 * What are SR, IT, and IR mode?  Which modes do the Solaris drivers
   use, or does it matter?

 * Has someone found the tool mentioned here by some above-the-table
   means, or only by request from LSI?:

http://www.opensolaris.org/jive/message.jspa?messageID=184811#184811

   The mention that a SPARC version of the tool exists is encouraging.
   The procedure to clear persistent mappings through the BIOS
   obviously won't work on SPARC.

Here are the notes I have so far:

-8-
 The driver for LSI's MegaRAID SAS card is mega_sas which
 was integrated into snv_88. It's planned for backporting to
 a Solaris 10 update.
There is also a BSD-licensed driver for that hardware, called
mfi. It's available from
http://www.itee.uq.edu.au/~dlg/mfi

 a scsi_vhci
 sort of driver for the LSI card in the Ultra {20,25}
Well yes, that's mpt(7d) as delivered 

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Nathan Kroenert
For what it's worth, I have been running Nevada (so, same kernel as 
opensolaris) for ages (at least 18 months) on a Gigabyte board with the 
MCP55 chipset and it's been flawless.


I liked it so much, I bought it's newer brother, based on the nvidia 
750SLI chipset...   M750SLI-DS4


Cheers!

Nathan.


On 13/03/09 09:21 AM, Dave wrote:



Tim wrote:



On Thu, Mar 12, 2009 at 2:22 PM, Blake blake.ir...@gmail.com 
mailto:blake.ir...@gmail.com wrote:


I've managed to get the data transfer to work by rearranging my disks
so that all of them sit on the integrated SATA controller.

So, I feel pretty certain that this is either an issue with the
Supermicro aoc-sat2-mv8 card, or with PCI-X on the motherboard 
(though

I would think that the integrated SATA would also be using the PCI
bus?).

The motherboard, for those interested, is an HD8ME-2 (not, I now find
after buying this box from Silicon Mechanics, a board that's on the
Solaris HCL...)


http://www.supermicro.com/Aplus/motherboard/Opteron2000/MCP55/h8dme-2.cfm 



So I'm not considering one of LSI's HBA's - what do list members 
think

about this device:

http://www.provantage.com/lsi-logic-lsi00117~7LSIG03X.htm
http://www.provantage.com/lsi-logic-lsi00117%7E7LSIG03X.htm



I believe the MCP55's SATA controllers are actually PCI-E based.


I use Tyan 2927 motherboards. They have on-board nVidia MCP55 chipsets, 
which is the same chipset at the X4500 (IIRC). I wouldn't trust the 
MCP55 chipset in OpenSolaris. I had random disk hangs even while the 
machine was mostly idle.


In Feb 2008 I bought AOC-SAT2-MV8 cards and moved all my drives to these 
add-in cards. I haven't had any issues with drive hanging since. There 
does not seem to be any problems with the SAT2-MV8 under heavy load in 
my servers from what I've seen.


When the SuperMicro AOC-USAS-L8i came out later last year, I started 
using them instead. They work better than the SAT2-MV8s.


This card needs a 3U or bigger case:
http://www.supermicro.com/products/accessories/addon/AOC-USAS-L8i.cfm

This is the low profile card that will fit in a 2U:
http://www.supermicro.com/products/accessories/addon/AOC-USASLP-L8i.cfm

They both work in normal PCI-E slots on my Tyan 2927 mobos.

Finding good non-Sun hardware that works very well under OpenSolaris is 
frustrating to say the least. Good luck.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--


//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Senior Systems Engineer  Phone:  +61 3 9869 6255 //
// Global Systems Engineering   Fax:+61 3 9869 6288 //
// Level 7, 476 St. Kilda Road  //
// Melbourne 3004   VictoriaAustralia   //
//
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Will Murnane
On Thu, Mar 12, 2009 at 18:30, Miles Nordin car...@ivy.net wrote:
  I love the way they use the numbers 3800 and 3080, so you are
  constantly transposing them thus leaving google littered with all
  this confusingly wrong information.
Think of the middle two digits as (number of external ports, number of
internal ports).  For example, I have a 3442E-R which has 4 internal
and 4 external ports, the 3800 has 8 external ports and 0 internal,
and so forth.  One place this breaks down is with cards like the ;
it has a total of 8 ports, any group of 4 of which can be mapped to
internal or external ports.

   AOC-USAS-L4iR
       identical to the above, but ``includes iButton''
       which is an old type of smartcard-like device with
       sometimes crypto and javacard support.
       apparently some kind of license key to
       unlock RAID5?  no L8iR exists though, only L4iR.
       I have the L8i, and it does have an iButton socket
       with no button in it.
I think the iButton is just used as an unlock code for the builtin
RAID 5 functionality.  Nothing the end user cares about, unless they
want RAID and have to spend the extra money.

     * SR = Software RAID IT = Integrate. Target mode. IR mode is not 
 supported.
Integrated target mode lets you export some storage attached to the
host system (through another adapter, presumably) as a storage device.
 IR mode is almost certainly Internal RAID, which that card doesn't
have support for.

 also there seem to be two different kinds of quad-SATA connector on
 these SAS cards so there are two different kinds of octopus cable.
Yes---SFF-8484 and SFF-8087 are the key words.

 SATA disks will always show up when attached to a SAS HBA,
 because that's one of the requirements of the SAS specification.
I'm not sure what you mean by this.  SAS controllers can control SATA
disks, and interact with them.  They don't just show up; they're
first-class citizens.

Will
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread Miles Nordin
 wm == Will Murnane will.murn...@gmail.com writes:

     * SR = Software RAID IT = Integrate. Target mode. IR mode
 is not supported.
wm Integrated target mode lets you export some storage attached
wm to the host system (through another adapter, presumably) as a
wm storage device.  IR mode is almost certainly Internal RAID,
wm which that card doesn't have support for.

no, the supermicro page for AOC-USAS-L8i does claim support for all
three, and supermicro has an ``IR driver'' available for download for
Linux and Windows, or at least a link to one.

I'm trying to figure out what's involved in determining and switching
modes, why you'd want to switch them, what cards support which modes,
which solaris drivers support which modes, u.s.w.

The answer may be very simple, like ``the driver supports only IR.
Most cards support IR, and cards that don't support IR won't work.  IR
can run in single-LUN mode.  Some IR cards support RAID5, others
support only RAID 0, 1, 10.''  Or it could be ``the driver supports
only SR.  The driver is what determines the mode, and it does this by
loading firmware into the card, and the first step in initializing the
card is always for the driver to load in a firmware blob.  All
currently-produced cards support SR.''  so...actually, now that I say
it, I guess the answer cannot be very simple.  It's going to have to
be a little complicated.

Anyway, I can guess, too.  I was hoping someone would know for sure
off-hand.


pgpv7SB8wKna7.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-12 Thread James C. McPherson
On Thu, 12 Mar 2009 22:24:12 -0400
Miles Nordin car...@ivy.net wrote:

  wm == Will Murnane will.murn...@gmail.com writes:
 
      * SR = Software RAID IT = Integrate. Target mode. IR mode
  is not supported.
 wm Integrated target mode lets you export some storage attached
 wm to the host system (through another adapter, presumably) as a
 wm storage device.  IR mode is almost certainly Internal RAID,
 wm which that card doesn't have support for.
 
 no, the supermicro page for AOC-USAS-L8i does claim support for all
 three, and supermicro has an ``IR driver'' available for download for
 Linux and Windows, or at least a link to one.
 
 I'm trying to figure out what's involved in determining and switching
 modes, why you'd want to switch them, what cards support which modes,
 which solaris drivers support which modes, u.s.w.
 
 The answer may be very simple, like ``the driver supports only IR.
 Most cards support IR, and cards that don't support IR won't work.  IR
 can run in single-LUN mode.  Some IR cards support RAID5, others
 support only RAID 0, 1, 10.''  Or it could be ``the driver supports
 only SR.  The driver is what determines the mode, and it does this by
 loading firmware into the card, and the first step in initializing the
 card is always for the driver to load in a firmware blob.  All
 currently-produced cards support SR.''  so...actually, now that I say
 it, I guess the answer cannot be very simple.  It's going to have to
 be a little complicated.
 Anyway, I can guess, too.  I was hoping someone would know for sure
 off-hand.


Hi Miles,
the mpt(7D) driver supports that card. mpt(7D) supports both
IT and IR firmware variants. You can find out the specifics
for what RAID volume levels are supported by reading the 
raidctl(1M) manpage. I don't think you can switch between IT
and IR firmware, but not having needed to know this before,
I haven't tried it.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I have a H8DM8-2 motherboard with a pair of AOC-SAT2-MV8 SATA
controller cards in a 16-disk Supermicro chassis.

I'm running OpenSolaris 2008.11, and the machine performs very well
unless I start to copy a large amount of data to the ZFS (software
raid) array that's on the Supermicro SATA controllers.  If I do this,
the machine inevitably reboots.

What can I do to troubleshoot?  The BIOS of the motherboard and the
SATA card firmware are fully updated.  I'm running the latest stable
OpenSolaris, and see nothing amiss in the system logs when this
happens.  I've enabled savecore and debug-level syslog, but am getting
no indicators from Solaris as to what's wrong.

Interestingly, I can push the same amount of data to the mirror boot
disks, which are on the board's built-in nVidia SATA controller
without issue.

The vdev I'm pushing to is a 5-disk raidz2 with 2 hot spares.

Help!  :)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Marc Bevand
The copy operation will make all the disks start seeking at the same time and 
will make your CPU activity jump to a significant percentage to compute the 
ZFS checksum and RAIDZ parity. I think you could be overloading your PSU 
because of the sudden increase in power consumption...

However if you are *not* using SATA staggered spin-up, then the above theory 
is unlikely because spinning up consumes much more power than when seeking. 
So, in a sense, a successful boot proves your PSU is powerful enough.

Trying reproducing the problem by copying data on a smaller number of disks. 
You tried 2 and 16. Try 8.

-marc

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:
 The copy operation will make all the disks start seeking at the same time and
 will make your CPU activity jump to a significant percentage to compute the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above theory
 is unlikely because spinning up consumes much more power than when seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Remco Lengers

Something is not right in the IO space. The messages talk about

vendor ID = 11AB

0x11AB  Marvell Semiconductor

TMC Research

Vendor Id: 0x1030
Short Name: TMC

Does fmdump -eV give any clue when the box comes back up?

..Remco


Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

The copy operation will make all the disks start seeking at the same time and
will make your CPU activity jump to a significant percentage to compute the
ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
because of the sudden increase in power consumption...

However if you are *not* using SATA staggered spin-up, then the above theory
is unlikely because spinning up consumes much more power than when seeking.
So, in a sense, a successful boot proves your PSU is powerful enough.

Trying reproducing the problem by copying data on a smaller number of disks.
You tried 2 and 16. Try 8.

-marc

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss








___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
fmdump is not helping much:

r...@host:~# fmdump -eV
TIME   CLASS
fmdump: /var/fm/fmd/errlog is empty


comparing that screenshot to the output of cfgadm is interesting -
looks like the controller(s):

r...@host:~# cfgadm -v
Ap_Id  Receptacle   Occupant Condition  Information
When Type Busy Phys_Id
sata4/0::dsk/c4t0d0connectedconfigured   ok
Mod: ST3250310NS FRev: SN06 SN: 9SF06CZZ
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5:0
sata4/1::dsk/c4t1d0connectedconfigured   ok
Mod: ST3250310NS FRev: SN06 SN: 9SF06BC8
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5:1
sata5/0emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,1:0
sata5/1::dsk/c7t1d0connectedconfigured   ok
Mod: WDC WD10EACS-00D6B0 FRev: 01.01A01 SN: WD-WCAU40244615
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5,1:1
sata6/0emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,2:0
sata6/1emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,2:1
sata7/0emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:0
sata7/1emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:1
sata7/2::dsk/c5t2d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0376631
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:2
sata7/3::dsk/c5t3d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0350798
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:3
sata7/4::dsk/c5t4d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0403574
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:4
sata7/5::dsk/c5t5d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0312592
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:5
sata7/6::dsk/c5t6d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0399779
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:6
sata7/7::dsk/c5t7d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0441660
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:7
sata8/0::dsk/c6t0d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT1000344
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:0
sata8/1emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:1
sata8/2emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:2
sata8/3emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:3
sata8/4emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:4
sata8/5emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:5
sata8/6emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:6
sata8/7emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:7


On Wed, Mar 11, 2009 at 2:40 PM, Blake blake.ir...@gmail.com wrote:
 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:
 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:
 The copy operation will make all the disks 

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I think that TMC Research is the company that designed the
Supermicro-branded controller card that has the Marvell SATA
controller chip on it.  Googling around I see connections between
Supermicro and TMC.

This is the card:

http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:
 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 


 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
Could the problem be related to this bug:

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

I'm testing setting the maximum payload size as a workaround, as noted
in the bug notes.



On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:
 I think that TMC Research is the company that designed the
 Supermicro-branded controller card that has the Marvell SATA
 controller chip on it.  Googling around I see connections between
 Supermicro and TMC.

 This is the card:

 http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

 On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:
 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 


 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Remco Lengers

looks worth a go otherwise:

if the boot disk is also off that controller it may be too hosed to 
write anything to the boot disk hence FMA doesn't see any issue when it 
comes up. Possible further actions:


- Upgrade FW of controller to highest or known working level
- Upgrade driver or OS level.
- Try another controller (may be its broken and barfs under stress ?)
- Analyze the crash dump (if any is saved)
- It may be its a know Solaris or driver bug and somebody has heard of 
it before.


hth,

..Remco

Blake wrote:

Could the problem be related to this bug:

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

I'm testing setting the maximum payload size as a workaround, as noted
in the bug notes.



On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:

I think that TMC Research is the company that designed the
Supermicro-branded controller card that has the Marvell SATA
controller chip on it.  Googling around I see connections between
Supermicro and TMC.

This is the card:

http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:

Something is not right in the IO space. The messages talk about

vendor ID = 11AB

0x11AB  Marvell Semiconductor

TMC Research

Vendor Id: 0x1030
Short Name: TMC

Does fmdump -eV give any clue when the box comes back up?

..Remco


Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

The copy operation will make all the disks start seeking at the same
time and
will make your CPU activity jump to a significant percentage to compute
the
ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
because of the sudden increase in power consumption...

However if you are *not* using SATA staggered spin-up, then the above
theory
is unlikely because spinning up consumes much more power than when
seeking.
So, in a sense, a successful boot proves your PSU is powerful enough.

Trying reproducing the problem by copying data on a smaller number of
disks.
You tried 2 and 16. Try 8.

-marc

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss







___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
Any chance this could be the motherboard?  I suspect the controller.

The boot disks are on the built-in nVidia controller.

On Wed, Mar 11, 2009 at 3:41 PM, Remco Lengers re...@lengers.com wrote:
 - Upgrade FW of controller to highest or known working level
I think I have the latest controller firmware.

 - Upgrade driver or OS level.
I'm going to try to go from 101b to 108 or whatever the current dev release is.

 - Try another controller (may be its broken and barfs under stress ?)
In the works.

 - Analyze the crash dump (if any is saved)
Crash dump is not saving properly.

 - It may be its a know Solaris or driver bug and somebody has heard of it
 before.
Any takers on this?  :)


 hth,
Thanks!


 ..Remco

 Blake wrote:

 Could the problem be related to this bug:

 http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

 I'm testing setting the maximum payload size as a workaround, as noted
 in the bug notes.



 On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:

 I think that TMC Research is the company that designed the
 Supermicro-branded controller card that has the Marvell SATA
 controller chip on it.  Googling around I see connections between
 Supermicro and TMC.

 This is the card:

 http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

 On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:

 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com
 wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to
 compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your
 PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 



 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Richard Elling

Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:
  

I'm working on testing this some more by doing a savecore -L right
after I start the copy.




savecore -L is not what you want.

By default, for OpenSolaris, savecore on boot is disabled.  But the core
will have been dumped into the dump slice, which is not used for swap.
So you should be able to run savecore at a later time to collect the
core from the last dump.
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I guess I didn't make it clear that I had already tried using savecore
to retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out
of space on the dump device:

r...@host:~# dumpadm
  Dump content: kernel pages
   Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated)
Savecore directory: /var/crash/host
  Savecore enabled: yes

I was using the -L option only to try to get some idea of why the
system load was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
richard.ell...@gmail.com wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the core
 will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Maidak Alexander J
If you're having issues with a disk contoller or disk IO driver its highly 
likely that a savecore to disk after the panic will fail.  I'm not sure how to 
work around this, maybe a dedicated dump device not on a controller that uses a 
different driver then the one that you're having issues with?

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
Sent: Wednesday, March 11, 2009 4:45 PM
To: Richard Elling
Cc: Marc Bevand; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] reboot when copying large amounts of data

I guess I didn't make it clear that I had already tried using savecore to 
retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out of space 
on the dump device:

r...@host:~# dumpadm
  Dump content: kernel pages
   Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore directory: 
/var/crash/host
  Savecore enabled: yes

I was using the -L option only to try to get some idea of why the system load 
was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The 
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right 
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the 
 core will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the 
 core from the last dump.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:
 If you're having issues with a disk contoller or disk IO driver its highly 
 likely that a savecore to disk after the panic will fail.  I'm not sure how 
 to work around this, maybe a dedicated dump device not on a controller that 
 uses a different driver then the one that you're having issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org 
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore to 
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out of 
 space on the dump device:

 r...@host:~# dumpadm
      Dump content: kernel pages
       Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore 
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the system load 
 was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
 wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Nathan Kroenert

Hm -

Crashes, or hangs? Moreover - how do you know a CPU is pegged?

Seems like we could do a little more discovery on what the actual 
problem here is, as I can read it about 4 different ways.


By this last piece of information, I'm guessing the system does not 
crash, but goes really really slow??


Crash == panic == we see stack dump on console and try to take a dump
hang == nothing works == no response - might be worth looking at mdb -K
or booting with a -k on the boot line.

So - are we crashing, hanging, or something different?

It might simply be that you are eating up all your memory, and your 
physical backing storage is taking a while to catch up?


Nathan.

Blake wrote:

My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:

If you're having issues with a disk contoller or disk IO driver its highly 
likely that a savecore to disk after the panic will fail.  I'm not sure how to 
work around this, maybe a dedicated dump device not on a controller that uses a 
different driver then the one that you're having issues with?

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
Sent: Wednesday, March 11, 2009 4:45 PM
To: Richard Elling
Cc: Marc Bevand; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] reboot when copying large amounts of data

I guess I didn't make it clear that I had already tried using savecore to 
retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out of space 
on the dump device:

r...@host:~# dumpadm
 Dump content: kernel pages
  Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore directory: 
/var/crash/host
 Savecore enabled: yes

I was using the -L option only to try to get some idea of why the system load 
was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
wrote:

Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


I'm working on testing this some more by doing a savecore -L right
after I start the copy.



savecore -L is not what you want.

By default, for OpenSolaris, savecore on boot is disabled.  But the
core will have been dumped into the dump slice, which is not used for swap.
So you should be able to run savecore at a later time to collect the
core from the last dump.
-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--
//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Systems Engineer Phone:  +61 3 9869-6255 //
// Sun Microsystems Fax:+61 3 9869-6288 //
// Level 7, 476 St. Kilda Road  Mobile: 0419 305 456//
// Melbourne 3004   VictoriaAustralia   //
//
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss