Re: [zfs-discuss] Nexsan SATABeast and ZFS

2009-03-11 Thread Lars-Gunnar Persson

I would like to go back to my question for a second:

I checked with my Nexsan supplier and they confirmed that access to  
every single disk in SATABeast is not possible. The smallest entities  
I can create on the SATABeast are RAID 0 or 1 arrays. With RAID 1 I'll  
loose too much disk space and I believe that leaves me with RAID 0 as  
the only reasonable option. But with this unsecure RAID format I'll  
need higher redundancy in the ZFS configuration. I think I'll go with  
the following configuration:


On the Nexsan SATABeast:
* 14 disks configured in 7 RAID arrays with RAID level 0 (each disk is  
1 TB which gives me a total of 14 TB raw disk space).

* Each RAID 0 array configured as one volume.

On the Sun Fire X4100 M2 with Solaris 10:
* Add all 7 volumes to one zpool configured in on raidz2 (gives me  
approx. 8,8 TB available disk space)


Any comments or suggestions?

Best regards, Lars-Gunnar Persson

On 11. mars. 2009, at 02.39, Bob Friesenhahn wrote:


On Tue, 10 Mar 2009, A Darren Dunham wrote:


What part isn't true?  ZFS has a independent checksum for the data
block.  But if the data block is spread over multiple disks, then  
each

of the disks have to be read to verify the checksum.


I interpreted what you said to imply that RAID6 type algorithms were  
being used to validate the data, rather than to correct wrong data.   
I agree that it is necessary to read a full ZFS block in order to  
use the ZFS block checksum.  I also agree that a raidz2 vdev has  
IOPS behavior which is similar to a single disk.



From what I understand, a raidz2 with a very large number of disks
won't use all of the disks to store one ZFS block.  There is a  
maximum number of disks in a stripe which can be supported by the  
ZFS block size.


--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ext4 bug zfs handling of the very same situation

2009-03-11 Thread Thomas Maier-Komor
Hi,

there was recently a bug reported against EXT4 that gets triggered by
KDE: https://bugs.edge.launchpad.net/ubuntu/+source/linux/+bug/317781

Now I'd like to verify that my understanding of ZFS behavior and
implementations is correct, and ZFS is unaffected from this kind of
issue. Maybe somebody would like to comment on this.

The underlying problem with ext4 is that some kde executables do
something like this:
1a) open and read data from file x, close file x
1b) open and truncate file x
1c) write data to file x
1d) close file x

or

2a) open and read data from file x, close file x
2b) open and truncate file x.new
2c) write data to file x.new
2d) close file x.new
2e) rename file x.new to file x

Concerning case 1) I think ZFS may lose data if power is lost right
after 1b) and open(xxx,O_WRONLY|O_TRUNC|O_CREAT) is issued in a
transaction group separately from the one containing 1c/1d.

Concerning case 2) I cannot see ZFS losing any data, because of
copy-on-write and transaction grouping.

Theodore Ts'o (ext4 developer) commented that both cases are flawed and
cannot be supported correctly, because of a lacking fsync() before
close. Is this correct? His comment is over here:
https://bugs.edge.launchpad.net/ubuntu/+source/linux/+bug/317781/comments/54

Any thoughts or comments?

TIA,
Thomas
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ext4 bug zfs handling of the very same situation

2009-03-11 Thread Casper . Dik


The underlying problem with ext4 is that some kde executables do
something like this:
1a) open and read data from file x, close file x
1b) open and truncate file x
1c) write data to file x
1d) close file x

or

2a) open and read data from file x, close file x
2b) open and truncate file x.new
2c) write data to file x.new
2d) close file x.new
2e) rename file x.new to file x

Concerning case 1) I think ZFS may lose data if power is lost right
after 1b) and open(xxx,O_WRONLY|O_TRUNC|O_CREAT) is issued in a
transaction group separately from the one containing 1c/1d.

Yes, I would assume that is possible but the change for it happening is
small.  Other filesystems prefer to write the meta data prompt, but
ZFS will easily wait until the file is completely written.

And UFS has the extra problem that it can change the file size and reading
will show garbage in the file.  (It changes the inode, possibly because 
it's in the log, but it hasn't written the data).   We've seen that problem
with UFS and the /etc/*_* driver files.  Precisely because we didn't
flush/fsync.  (And in some cases we used fsync(fileno(file)), but the
new content was still in the stdio buffer)

Only versioned filesystems can make the first sequence work.

Concerning case 2) I cannot see ZFS losing any data, because of
copy-on-write and transaction grouping.

Theodore Ts'o (ext4 developer) commented that both cases are flawed and
cannot be supported correctly, because of a lacking fsync() before
close. Is this correct? His comment is over here:
https://bugs.edge.launchpad.net/ubuntu/+source/linux/+bug/317781/comments/54

Perhaps we should Theodore Ts'o that ZFS gets this right :-)

I'm assuming that all transactions in group N all happened before
group N+1 at least when it comes to the partial order in which the
transactions happen.

Casper

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [osol-discuss] ext4 bug zfs handling of the very same situation

2009-03-11 Thread Joerg Schilling
Thomas Maier-Komor tho...@maier-komor.de wrote:

 there was recently a bug reported against EXT4 that gets triggered by
 KDE: https://bugs.edge.launchpad.net/ubuntu/+source/linux/+bug/317781

 Now I'd like to verify that my understanding of ZFS behavior and
 implementations is correct, and ZFS is unaffected from this kind of
 issue. Maybe somebody would like to comment on this.

 The underlying problem with ext4 is that some kde executables do
 something like this:
 1a) open and read data from file x, close file x
 1b) open and truncate file x
 1c) write data to file x
 1d) close file x

 or

 2a) open and read data from file x, close file x
 2b) open and truncate file x.new
 2c) write data to file x.new
 2d) close file x.new
 2e) rename file x.new to file x

 Concerning case 1) I think ZFS may lose data if power is lost right
 after 1b) and open(xxx,O_WRONLY|O_TRUNC|O_CREAT) is issued in a
 transaction group separately from the one containing 1c/1d.

It depends on what you call loses data.

If you truncate an existing file, the content of the file is expected 
to be lost if the file system behaves correctly.

 Concerning case 2) I cannot see ZFS losing any data, because of
 copy-on-write and transaction grouping.

 Theodore Ts'o (ext4 developer) commented that both cases are flawed and
 cannot be supported correctly, because of a lacking fsync() before
 close. Is this correct? His comment is over here:
 https://bugs.edge.launchpad.net/ubuntu/+source/linux/+bug/317781/comments/54

An application cannot be sure on whether the data has been written to disk
if there was no fsync() call. Star for this reason by default calls fsync() 
before closing a file in order to be able to include information about problems
in the star exit code.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I have a H8DM8-2 motherboard with a pair of AOC-SAT2-MV8 SATA
controller cards in a 16-disk Supermicro chassis.

I'm running OpenSolaris 2008.11, and the machine performs very well
unless I start to copy a large amount of data to the ZFS (software
raid) array that's on the Supermicro SATA controllers.  If I do this,
the machine inevitably reboots.

What can I do to troubleshoot?  The BIOS of the motherboard and the
SATA card firmware are fully updated.  I'm running the latest stable
OpenSolaris, and see nothing amiss in the system logs when this
happens.  I've enabled savecore and debug-level syslog, but am getting
no indicators from Solaris as to what's wrong.

Interestingly, I can push the same amount of data to the mirror boot
disks, which are on the board's built-in nVidia SATA controller
without issue.

The vdev I'm pushing to is a 5-disk raidz2 with 2 hot spares.

Help!  :)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Export ZFS via ISCSI to Linux - Is it stable for production use now?

2009-03-11 Thread howard chen
Hello,

I want to setup an opensolaris for centralized storage server, using
ZFS as the underlying FS, on a RAID 10 SATA disks.

I will export the storage blocks using ISCSI to RHEL 5 (less than 10
clients, and I will format the partition as EXT3)

I want to ask...

1. Is this setup suitable for mission critical use now?
2. Can I use LVM with this setup?


Currently we are using NFS as the centralized storage solutions.


Thanks.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Export ZFS via ISCSI to Linux - Is it stable for production use now?

2009-03-11 Thread Darren J Moffat

howard chen wrote:

Hello,

I want to setup an opensolaris for centralized storage server, using
ZFS as the underlying FS, on a RAID 10 SATA disks.

I will export the storage blocks using ISCSI to RHEL 5 (less than 10
clients, and I will format the partition as EXT3)

I want to ask...

1. Is this setup suitable for mission critical use now?


Yes, why wouldn't it be ?

It is even officially supported as part of the Sun Unified Storage 7000 
series appliances built from OpenSolaris and ZFS technology.



2. Can I use LVM with this setup?


Yes you can.

--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Marc Bevand
The copy operation will make all the disks start seeking at the same time and 
will make your CPU activity jump to a significant percentage to compute the 
ZFS checksum and RAIDZ parity. I think you could be overloading your PSU 
because of the sudden increase in power consumption...

However if you are *not* using SATA staggered spin-up, then the above theory 
is unlikely because spinning up consumes much more power than when seeking. 
So, in a sense, a successful boot proves your PSU is powerful enough.

Trying reproducing the problem by copying data on a smaller number of disks. 
You tried 2 and 16. Try 8.

-marc

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:
 The copy operation will make all the disks start seeking at the same time and
 will make your CPU activity jump to a significant percentage to compute the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above theory
 is unlikely because spinning up consumes much more power than when seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Import Error - Please Help

2009-03-11 Thread Michael Dupree jr
Okay so for some reason my Opensolaris 106 installation quit booting
properly (completely seperate problem). At this point I was still able
to boot into safe mode, but not into a normal desktop session.

First a disk failed, I replaced this disk with another. The array
completely resilvered. I was able to scrub the array, etc.

Second, I added 2 more disks, issuing the replace command to use these
2 new disks instead of 2 of the old disks (so I replaced 2 good IDE
disks with 2 good SATA disks, all 4 disks were attached at the same
time, thus the reason for multiple replace's at one time).

The resilver completed successfully. I was able to scrub the array. I
exported the array, rebooted, imported the array, and scrubbed it
again - no problems.

Now I removed the 2 IDE drives, rebooted again, was able to import the
array, scrub it, etc no problems.

(Until this point everything has been done in safe mode, and the
replacing stuff never went away although the scrub's completed fully,
and all the origional drives (that had been replaced) have been
removed).

SO I then proceed to install opensolaris 109 (in order to get a fully
booting/working machine again). Now when I zpool import I get the
following:

# zpool import
 pool: tank
   id: 15417328101188042571
 state: UNAVAIL
action: The pool cannot be imported due to damaged devices or data.
config:

   tankUNAVAIL  insufficient replicas
 raidz1ONLINE
   c1t1d0  ONLINE
   c3t0d0  ONLINE
   c2t0d0  ONLINE
   c4t1d0  ONLINE
   c1t0d0p2ONLINE
 raidz1UNAVAIL  corrupted data
   c4t7d0  ONLINE
   c4t6d0  ONLINE
   replacing   ONLINE
 13069118156582972341  ONLINE
 c2t1d0ONLINE
   replacing   DEGRADED
 11144084490198638692  UNAVAIL  cannot open
 c5t0d0ONLINE
   replacing   ONLINE
 3239645952837551668   ONLINE
 c5t1d0ONLINE
   c4t0d0  ONLINE
   c4t5d0  ONLINE
   c4t2d0  ONLINE
   c4t4d0  ONLINE
   c4t3d0  ONLINE
#


Everything shows as online except for one vdev (that is currently
replacing), so I don't understand why it says insufficient replicas
instead of DEGRADED

Any help appreciated.

Thanks in advance,
-mike


-- 
Michael Dupree jr.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Nexsan SATABeast and ZFS

2009-03-11 Thread Moore, Joe
Lars-Gunnar Persson wrote:
 I would like to go back to my question for a second:
 
 I checked with my Nexsan supplier and they confirmed that access to
 every single disk in SATABeast is not possible. The smallest entities
 I can create on the SATABeast are RAID 0 or 1 arrays. With RAID 1 I'll
 loose too much disk space and I believe that leaves me with RAID 0 as
 the only reasonable option. But with this unsecure RAID format I'll
 need higher redundancy in the ZFS configuration. I think I'll go with
 the following configuration:
 
 On the Nexsan SATABeast:
 * 14 disks configured in 7 RAID arrays with RAID level 0 (each disk is
 1 TB which gives me a total of 14 TB raw disk space).
 * Each RAID 0 array configured as one volume.

So what the front end will see is 7 disks, 2TB each disk.

 
 On the Sun Fire X4100 M2 with Solaris 10:
 * Add all 7 volumes to one zpool configured in on raidz2 (gives me
 approx. 8,8 TB available disk space)

You'll get 5 LUNs worth of space in this config, or 10TB of usable space.

 
 Any comments or suggestions?

Given the hardware constraints (no single-disk volumes allowed) this is a good 
configuration for most purposes.

The advantages/disadvantages are:
. 10TB of usable disk space, out of 14TB purchased.
. At least three hard disk failures are required to lose the ZFS pool.
. Random non-cached read performance will be about 300 IO/sec.
. Sequential reads and writes of the whole ZFS blocksize will be fast (up to 
2000 IO/sec).
. One hard drive failure will cause the used blocks of the 2TB LUN (raid0 pair) 
to be resilvered, even though the other half of the pair is not damaged.  The 
other half of the pair is more likely to fail during the ZFS resilvering 
operation because of increased load.

You'll want to pay special attention to the cache settings on the Nexsan.  You 
earlier showed that the write cache is enabled, but IIRC the array doesn't have 
a nonvolatile (battery-backed) cache.  If that's the case, MAKE SURE it's 
hooked up to a UPS that can support it for the 30 second cache flush timeout on 
the array.  And make sure you don't power it down hard.  I think you want to 
uncheck the ignore FUA setting, so that FUA requests are respected.  My guess 
is that this will cause the array to properly handle the cache_flush requests 
that ZFS uses to ensure data consistancy.

--Joe
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Export ZFS via ISCSI to Linux - Is it stable for production use now?

2009-03-11 Thread howard chen
Hello,

On Wed, Mar 11, 2009 at 10:20 PM, Darren J Moffat
darr...@opensolaris.org wrote:

 1. Is this setup suitable for mission critical use now?

 Yes, why wouldn't it be ?


Because I just wonder why some other people are using zfs/fuse on Linux, e.g.
http://www.drwetter.org/blog/zfs_under_linux.en.html
http://www.wizy.org/wiki/ZFS_on_FUSE


Also seems hard to find a complete tutorial on accessing ZFS from
Linux using ISCSI. This should be attractive, isn't?

So I don't know if it is experimental.

Howard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Export ZFS via ISCSI to Linux - Is it stable for production use now?

2009-03-11 Thread Darren J Moffat

howard chen wrote:

Hello,

On Wed, Mar 11, 2009 at 10:20 PM, Darren J Moffat
darr...@opensolaris.org wrote:

1. Is this setup suitable for mission critical use now?

Yes, why wouldn't it be ?



Because I just wonder why some other people are using zfs/fuse on Linux, e.g.
http://www.drwetter.org/blog/zfs_under_linux.en.html
http://www.wizy.org/wiki/ZFS_on_FUSE


ZFS/fuse on Linux gives you something very different to exporting ZFS 
ZVOLs over iSCSI.  They solve completely different problems.


As for why people play with zfs/fuse on Linux rather than running ZFS on 
OpenSolaris - well because they can, or because their religion means 
they want to use Linux or just to test zfs/fuse.  Same as why people 
play with ZFS on FreeBSD or MacOS X.  Because they can and for what ever 
reasons they have they want to use ZFS and an OS other than OpenSolaris.


None of that implies that sharing ZVOLs over iSCSI isn't stable.


Also seems hard to find a complete tutorial on accessing ZFS from
Linux using ISCSI. This should be attractive, isn't?


On the OpenSolaris side:

Create your pool:

# zpool create tank your list of disks and raid config goes here

Turn on the iSCSI service

# svcadm enable iscsitgtd
[ or if you are on a release with COMSTART make that iscsi/target ]

Create the ZVOL for the iSCSI lun and share it out:

# zfs create -V 10g -o shareiscsi=on tank/lun1

If you want more control over the sharing options see iscsitadm(1M)

That is all that has to be done on the OpenSolaris side to make a 10g 
lun available over iSCSI.  The rest of it is all how Linux sets up its 
iSCSI client side which I don't know but I know on Solaris it is very 
easy using iscsiadm(1M).


--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Export ZFS via ISCSI to Linux - Is it stable for production use now?

2009-03-11 Thread Blake
I blogged this a while ago:

http://blog.clockworm.com/2007/10/connecting-linux-centos-5-to-solaris.html

On Wed, Mar 11, 2009 at 1:02 PM, howard chen howac...@gmail.com wrote:
 Hello,

 On Wed, Mar 11, 2009 at 10:20 PM, Darren J Moffat
 darr...@opensolaris.org wrote:

 1. Is this setup suitable for mission critical use now?

 Yes, why wouldn't it be ?


 Because I just wonder why some other people are using zfs/fuse on Linux, e.g.
 http://www.drwetter.org/blog/zfs_under_linux.en.html
 http://www.wizy.org/wiki/ZFS_on_FUSE


 Also seems hard to find a complete tutorial on accessing ZFS from
 Linux using ISCSI. This should be attractive, isn't?

 So I don't know if it is experimental.

 Howard
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Remco Lengers

Something is not right in the IO space. The messages talk about

vendor ID = 11AB

0x11AB  Marvell Semiconductor

TMC Research

Vendor Id: 0x1030
Short Name: TMC

Does fmdump -eV give any clue when the box comes back up?

..Remco


Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

The copy operation will make all the disks start seeking at the same time and
will make your CPU activity jump to a significant percentage to compute the
ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
because of the sudden increase in power consumption...

However if you are *not* using SATA staggered spin-up, then the above theory
is unlikely because spinning up consumes much more power than when seeking.
So, in a sense, a successful boot proves your PSU is powerful enough.

Trying reproducing the problem by copying data on a smaller number of disks.
You tried 2 and 16. Try 8.

-marc

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss








___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
fmdump is not helping much:

r...@host:~# fmdump -eV
TIME   CLASS
fmdump: /var/fm/fmd/errlog is empty


comparing that screenshot to the output of cfgadm is interesting -
looks like the controller(s):

r...@host:~# cfgadm -v
Ap_Id  Receptacle   Occupant Condition  Information
When Type Busy Phys_Id
sata4/0::dsk/c4t0d0connectedconfigured   ok
Mod: ST3250310NS FRev: SN06 SN: 9SF06CZZ
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5:0
sata4/1::dsk/c4t1d0connectedconfigured   ok
Mod: ST3250310NS FRev: SN06 SN: 9SF06BC8
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5:1
sata5/0emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,1:0
sata5/1::dsk/c7t1d0connectedconfigured   ok
Mod: WDC WD10EACS-00D6B0 FRev: 01.01A01 SN: WD-WCAU40244615
unavailable  disk n/devices/p...@0,0/pci15d9,1...@5,1:1
sata6/0emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,2:0
sata6/1emptyunconfigured ok
unavailable  sata-portn/devices/p...@0,0/pci15d9,1...@5,2:1
sata7/0emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:0
sata7/1emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:1
sata7/2::dsk/c5t2d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0376631
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:2
sata7/3::dsk/c5t3d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0350798
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:3
sata7/4::dsk/c5t4d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0403574
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:4
sata7/5::dsk/c5t5d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0312592
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:5
sata7/6::dsk/c5t6d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0399779
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:6
sata7/7::dsk/c5t7d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT0441660
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@4:7
sata8/0::dsk/c6t0d0connectedconfigured   ok
Mod: WDC WD7500AYYS-01RCA0 FRev: 30.04G30 SN: WD-WCAPT1000344
unavailable  disk n
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:0
sata8/1emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:1
sata8/2emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:2
sata8/3emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:3
sata8/4emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:4
sata8/5emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:5
sata8/6emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:6
sata8/7emptyunconfigured ok
unavailable  sata-portn
/devices/p...@0,0/pci10de,3...@a/pci1033,1...@0/pci11ab,1...@6:7


On Wed, Mar 11, 2009 at 2:40 PM, Blake blake.ir...@gmail.com wrote:
 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:
 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:
 The copy operation will make all the disks 

Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I think that TMC Research is the company that designed the
Supermicro-branded controller card that has the Marvell SATA
controller chip on it.  Googling around I see connections between
Supermicro and TMC.

This is the card:

http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:
 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 


 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
Could the problem be related to this bug:

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

I'm testing setting the maximum payload size as a workaround, as noted
in the bug notes.



On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:
 I think that TMC Research is the company that designed the
 Supermicro-branded controller card that has the Marvell SATA
 controller chip on it.  Googling around I see connections between
 Supermicro and TMC.

 This is the card:

 http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

 On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:
 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 


 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Remco Lengers

looks worth a go otherwise:

if the boot disk is also off that controller it may be too hosed to 
write anything to the boot disk hence FMA doesn't see any issue when it 
comes up. Possible further actions:


- Upgrade FW of controller to highest or known working level
- Upgrade driver or OS level.
- Try another controller (may be its broken and barfs under stress ?)
- Analyze the crash dump (if any is saved)
- It may be its a know Solaris or driver bug and somebody has heard of 
it before.


hth,

..Remco

Blake wrote:

Could the problem be related to this bug:

http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

I'm testing setting the maximum payload size as a workaround, as noted
in the bug notes.



On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:

I think that TMC Research is the company that designed the
Supermicro-branded controller card that has the Marvell SATA
controller chip on it.  Googling around I see connections between
Supermicro and TMC.

This is the card:

http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:

Something is not right in the IO space. The messages talk about

vendor ID = 11AB

0x11AB  Marvell Semiconductor

TMC Research

Vendor Id: 0x1030
Short Name: TMC

Does fmdump -eV give any clue when the box comes back up?

..Remco


Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

I'm working on testing this some more by doing a savecore -L right
after I start the copy.

BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
supports 16, but isn't fully populated).

So far as I know, there is no spinup happening - these are not RAID
controllers, just dumb SATA JBOD controllers, so I don't think they
control drive spin in any particular way.  Correct me if I'm wrong, of
course.



On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com wrote:

The copy operation will make all the disks start seeking at the same
time and
will make your CPU activity jump to a significant percentage to compute
the
ZFS checksum and RAIDZ parity. I think you could be overloading your PSU
because of the sudden increase in power consumption...

However if you are *not* using SATA staggered spin-up, then the above
theory
is unlikely because spinning up consumes much more power than when
seeking.
So, in a sense, a successful boot proves your PSU is powerful enough.

Trying reproducing the problem by copying data on a smaller number of
disks.
You tried 2 and 16. Try 8.

-marc

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss







___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
Any chance this could be the motherboard?  I suspect the controller.

The boot disks are on the built-in nVidia controller.

On Wed, Mar 11, 2009 at 3:41 PM, Remco Lengers re...@lengers.com wrote:
 - Upgrade FW of controller to highest or known working level
I think I have the latest controller firmware.

 - Upgrade driver or OS level.
I'm going to try to go from 101b to 108 or whatever the current dev release is.

 - Try another controller (may be its broken and barfs under stress ?)
In the works.

 - Analyze the crash dump (if any is saved)
Crash dump is not saving properly.

 - It may be its a know Solaris or driver bug and somebody has heard of it
 before.
Any takers on this?  :)


 hth,
Thanks!


 ..Remco

 Blake wrote:

 Could the problem be related to this bug:

 http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6793353

 I'm testing setting the maximum payload size as a workaround, as noted
 in the bug notes.



 On Wed, Mar 11, 2009 at 3:14 PM, Blake blake.ir...@gmail.com wrote:

 I think that TMC Research is the company that designed the
 Supermicro-branded controller card that has the Marvell SATA
 controller chip on it.  Googling around I see connections between
 Supermicro and TMC.

 This is the card:

 http://www.supermicro.com/products/accessories/addon/AOC-SAT2-MV8.cfm

 On Wed, Mar 11, 2009 at 3:08 PM, Remco Lengers re...@lengers.com wrote:

 Something is not right in the IO space. The messages talk about

 vendor ID = 11AB

 0x11AB  Marvell Semiconductor

 TMC Research

 Vendor Id: 0x1030
 Short Name: TMC

 Does fmdump -eV give any clue when the box comes back up?

 ..Remco


 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:

 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.

 BTW, I'm copying to a raidz2 of only 5 disks, not 16 (the chassis
 supports 16, but isn't fully populated).

 So far as I know, there is no spinup happening - these are not RAID
 controllers, just dumb SATA JBOD controllers, so I don't think they
 control drive spin in any particular way.  Correct me if I'm wrong, of
 course.



 On Wed, Mar 11, 2009 at 11:23 AM, Marc Bevand m.bev...@gmail.com
 wrote:

 The copy operation will make all the disks start seeking at the same
 time and
 will make your CPU activity jump to a significant percentage to
 compute
 the
 ZFS checksum and RAIDZ parity. I think you could be overloading your
 PSU
 because of the sudden increase in power consumption...

 However if you are *not* using SATA staggered spin-up, then the above
 theory
 is unlikely because spinning up consumes much more power than when
 seeking.
 So, in a sense, a successful boot proves your PSU is powerful enough.

 Trying reproducing the problem by copying data on a smaller number of
 disks.
 You tried 2 and 16. Try 8.

 -marc

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


 



 

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS on a SAN

2009-03-11 Thread Grant Lowe

Hi All,

I'm new on ZFS, so I hope this isn't too basic a question.  I have a host where 
I setup ZFS.  The Oracle DBAs did their thing and I know have a number of ZFS 
datasets with their respective clones and snapshots on serverA.  I want to 
export some of the clones to serverB.  Do I need to zone serverB to see the 
same LUNs as serverA?  Or does it have to have preexisting, empty LUNs to 
import the clones?  Please help.  Thanks.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS on a SAN

2009-03-11 Thread Erik Trimble
I'm not 100% sure what your question here is, but let me give you a
(hopefully) complete answer:

(1) ZFS is NOT a clustered file system, in the sense that it is NOT
possible for two hosts to have the same LUN mounted at the same time,
even if both are hooked to a SAN and can normally see that LUN.

(2) ZFS can do failover, however.  If you have a LUN from a SAN on
hostA, create a ZFS pool in it, and use as normal.  Should you with to
failover the LUN to hostB, you need to do a 'zpool export zpool' on
hostA, then 'zpool import zpool' on hostB.  If hostA has been lost
completely (hung/died/etc) and you are unable to do an 'export' on it,
you can force the import on hostB via 'zpool import -f zpool'


ZFS requires that you import/export entire POOLS, not just filesystems.
So, given what you seem to want, I'd recommend this:

On the SAN, create (2) LUNs - one for your primary data, and one for
your snapshots/backups.

On hostA, create a zpool on the primary data LUN (call it zpool A), and
another zpool on the backup LUN (zpool B).  Take snapshots on A, then
use 'zfs send' and 'zfs receive' to copy the clone/snapshot over to
zpool B. then 'zpool export B'

On hostB, import the snapshot pool:  'zfs import B'



It might just be as easy to have two independent zpools on each host,
and just do a 'zfs send' on hostA, and 'zfs receive' on hostB to copy
the snapshot/clone over the wire.

-Erik



On Wed, 2009-03-11 at 13:18 -0700, Grant Lowe wrote:
 Hi All,
 
 I'm new on ZFS, so I hope this isn't too basic a question.  I have a host 
 where I setup ZFS.  The Oracle DBAs did their thing and I know have a number 
 of ZFS datasets with their respective clones and snapshots on serverA.  I 
 want to export some of the clones to serverB.  Do I need to zone serverB to 
 see the same LUNs as serverA?  Or does it have to have preexisting, empty 
 LUNs to import the clones?  Please help.  Thanks.
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
-- 
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS on a SAN

2009-03-11 Thread Grant Lowe

Hi Eric,

Thanks for the quick response.  Then on hostB, the new LUN will need the same 
amount of disk space for the pool, as on hostA, if I'm understanding you 
correctly.  Correct?  Thanks!



- Original Message 
From: Erik Trimble erik.trim...@sun.com
To: Grant Lowe gl...@sbcglobal.net
Cc: zfs-discuss@opensolaris.org
Sent: Wednesday, March 11, 2009 1:42:06 PM
Subject: Re: [zfs-discuss] ZFS on a SAN

I'm not 100% sure what your question here is, but let me give you a
(hopefully) complete answer:

(1) ZFS is NOT a clustered file system, in the sense that it is NOT
possible for two hosts to have the same LUN mounted at the same time,
even if both are hooked to a SAN and can normally see that LUN.

(2) ZFS can do failover, however.  If you have a LUN from a SAN on
hostA, create a ZFS pool in it, and use as normal.  Should you with to
failover the LUN to hostB, you need to do a 'zpool export zpool' on
hostA, then 'zpool import zpool' on hostB.  If hostA has been lost
completely (hung/died/etc) and you are unable to do an 'export' on it,
you can force the import on hostB via 'zpool import -f zpool'


ZFS requires that you import/export entire POOLS, not just filesystems.
So, given what you seem to want, I'd recommend this:

On the SAN, create (2) LUNs - one for your primary data, and one for
your snapshots/backups.

On hostA, create a zpool on the primary data LUN (call it zpool A), and
another zpool on the backup LUN (zpool B).  Take snapshots on A, then
use 'zfs send' and 'zfs receive' to copy the clone/snapshot over to
zpool B. then 'zpool export B'

On hostB, import the snapshot pool:  'zfs import B'



It might just be as easy to have two independent zpools on each host,
and just do a 'zfs send' on hostA, and 'zfs receive' on hostB to copy
the snapshot/clone over the wire.

-Erik



On Wed, 2009-03-11 at 13:18 -0700, Grant Lowe wrote:
 Hi All,
 
 I'm new on ZFS, so I hope this isn't too basic a question.  I have a host 
 where I setup ZFS.  The Oracle DBAs did their thing and I know have a number 
 of ZFS datasets with their respective clones and snapshots on serverA.  I 
 want to export some of the clones to serverB.  Do I need to zone serverB to 
 see the same LUNs as serverA?  Or does it have to have preexisting, empty 
 LUNs to import the clones?  Please help.  Thanks.
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
-- 
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS on a SAN

2009-03-11 Thread Richard Elling

Erik Trimble wrote:

I'm not 100% sure what your question here is, but let me give you a
(hopefully) complete answer:

(1) ZFS is NOT a clustered file system, in the sense that it is NOT
possible for two hosts to have the same LUN mounted at the same time,
even if both are hooked to a SAN and can normally see that LUN.
  


Need to be clear on the terminology here.  Yes, it is possible for two
systems to have access to a single LUN and have ZFS file systems.
The ZFS limitation is at the vdev level (partition or slice), which is
below the LUN level.



(2) ZFS can do failover, however.  If you have a LUN from a SAN on
hostA, create a ZFS pool in it, and use as normal.  Should you with to
failover the LUN to hostB, you need to do a 'zpool export zpool' on
hostA, then 'zpool import zpool' on hostB.  If hostA has been lost
completely (hung/died/etc) and you are unable to do an 'export' on it,
you can force the import on hostB via 'zpool import -f zpool'
  


LUN masking or reservations occur at the LUN level, which is why it is
often better (safer) to design with the expectation that a LUN will be
only available to one host at a time.  Or, to say it differently, if you 
think

of one vdev/LUN, then you can innoculate yourself from later grief :-)



ZFS requires that you import/export entire POOLS, not just filesystems.
So, given what you seem to want, I'd recommend this:

On the SAN, create (2) LUNs - one for your primary data, and one for
your snapshots/backups.

On hostA, create a zpool on the primary data LUN (call it zpool A), and
another zpool on the backup LUN (zpool B).  Take snapshots on A, then
use 'zfs send' and 'zfs receive' to copy the clone/snapshot over to
zpool B. then 'zpool export B'

On hostB, import the snapshot pool:  'zfs import B'



It might just be as easy to have two independent zpools on each host,
and just do a 'zfs send' on hostA, and 'zfs receive' on hostB to copy
the snapshot/clone over the wire.
  


+1
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Richard Elling

Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:
  

I'm working on testing this some more by doing a savecore -L right
after I start the copy.




savecore -L is not what you want.

By default, for OpenSolaris, savecore on boot is disabled.  But the core
will have been dumped into the dump slice, which is not used for swap.
So you should be able to run savecore at a later time to collect the
core from the last dump.
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS on a SAN

2009-03-11 Thread Erik Trimble
On Wed, 2009-03-11 at 13:50 -0700, Grant Lowe wrote:
 Hi Eric,
 
 Thanks for the quick response.  Then on hostB, the new LUN will need the same 
 amount of disk space for the pool, as on hostA, if I'm understanding you 
 correctly.  Correct?  Thanks!

I'm assuming you're referring to my second case, where you don't bother
to 'import'/'export' and single LUN, and just use the 'send'/'receive'
option. HostB's zpool only needs to have enough space to handle the
snapshot size.  You can of course delete a snapshot on either host after
it's taken, so you may not need to keep all snapshots around on both
hosts.

So, let's say this happens:  pool on hostA is 2TB, with 1TB of active
data, and a typical snapshot of 100GB.  Which means you can keep around
10 snapshots on hostA. On hostB, you need a zpool of 1TB for the 'live'
data, plus 100GB extra to store an incoming snapshot, but once you've
received the snapshot, you can delete any older snapshot, so you really
only need 1.1TB on hostB. Unless, of course, you want to keep the older
snapshots.


For old timer's sake, compare 'zfs send' and 'zfs'receive' to 'ufsdump'
and 'ufsrestore' (respectively) for their utility.

-Erik





 
 
 - Original Message 
 From: Erik Trimble erik.trim...@sun.com
 To: Grant Lowe gl...@sbcglobal.net
 Cc: zfs-discuss@opensolaris.org
 Sent: Wednesday, March 11, 2009 1:42:06 PM
 Subject: Re: [zfs-discuss] ZFS on a SAN
 
 I'm not 100% sure what your question here is, but let me give you a
 (hopefully) complete answer:
 
 (1) ZFS is NOT a clustered file system, in the sense that it is NOT
 possible for two hosts to have the same LUN mounted at the same time,
 even if both are hooked to a SAN and can normally see that LUN.
 
 (2) ZFS can do failover, however.  If you have a LUN from a SAN on
 hostA, create a ZFS pool in it, and use as normal.  Should you with to
 failover the LUN to hostB, you need to do a 'zpool export zpool' on
 hostA, then 'zpool import zpool' on hostB.  If hostA has been lost
 completely (hung/died/etc) and you are unable to do an 'export' on it,
 you can force the import on hostB via 'zpool import -f zpool'
 
 
 ZFS requires that you import/export entire POOLS, not just filesystems.
 So, given what you seem to want, I'd recommend this:
 
 On the SAN, create (2) LUNs - one for your primary data, and one for
 your snapshots/backups.
 
 On hostA, create a zpool on the primary data LUN (call it zpool A), and
 another zpool on the backup LUN (zpool B).  Take snapshots on A, then
 use 'zfs send' and 'zfs receive' to copy the clone/snapshot over to
 zpool B. then 'zpool export B'
 
 On hostB, import the snapshot pool:  'zfs import B'
 
 
 
 It might just be as easy to have two independent zpools on each host,
 and just do a 'zfs send' on hostA, and 'zfs receive' on hostB to copy
 the snapshot/clone over the wire.
 
 -Erik
 
 
 
 On Wed, 2009-03-11 at 13:18 -0700, Grant Lowe wrote:
  Hi All,
  
  I'm new on ZFS, so I hope this isn't too basic a question.  I have a host 
  where I setup ZFS.  The Oracle DBAs did their thing and I know have a 
  number of ZFS datasets with their respective clones and snapshots on 
  serverA.  I want to export some of the clones to serverB.  Do I need to 
  zone serverB to see the same LUNs as serverA?  Or does it have to have 
  preexisting, empty LUNs to import the clones?  Please help.  Thanks.
  
  ___
  zfs-discuss mailing list
  zfs-discuss@opensolaris.org
  http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
-- 
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
I guess I didn't make it clear that I had already tried using savecore
to retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out
of space on the dump device:

r...@host:~# dumpadm
  Dump content: kernel pages
   Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated)
Savecore directory: /var/crash/host
  Savecore enabled: yes

I was using the -L option only to try to get some idea of why the
system load was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling
richard.ell...@gmail.com wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the core
 will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Trying to determine if this box will be compatible with Opensolaris or Solaris

2009-03-11 Thread mike
http://www.intel.com/products/server/storage-systems/ssr212mc2/ssr212mc2-overview.htm
http://www.intel.com/support/motherboards/server/ssr212mc2/index.htm

It's hard to use the HAL sometimes.

I am trying to locate chipset info but having a hard time...
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Why did my zvol shrink ?

2009-03-11 Thread Brian H. Nelson

I'm doing a little testing and I hit a strange point. Here is a zvol (clone)

pool1/volclone  type volume -
pool1/volclone  origin   pool1/v...@diff1   -
pool1/volclone  reservation  none   default
pool1/volclone  volsize  191G   -
pool1/volclone  volblocksize 8K -

The zvol has UFS on it. It has always been 191G and we've never 
attempted to resize it. However, if I just try to grow it, it gives me 
an error:


-bash-3.00# growfs 
/dev/zvol/rdsk/pool1/volclone
400555998 sectors  current size of 400556032 sectors


Is the zvol is somehow smaller than it was originally? How/why?

It fsck ok, so UFS doesn't seem to notice.

This is solaris 10 u6 currently, the machine (and zpool) have gone 
through a few update releases since creation.


Thanks for any input,
-Brian

--
---
Brian H. Nelson Youngstown State University
System Administrator   Media and Academic Computing
 bnelson[at]cis.ysu.edu
---

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Trying to determine if this box will be compatible with Opensolaris or Solaris

2009-03-11 Thread mike
Looks like I may have slightly answered my question.

I found that it uses the server board S5000PSL, which is in the HCL:
http://www.sun.com/bigadmin/hcl/data/systems/details/2944.html

It's not listed in the OpenSolaris one, only the Solaris one. I would
think something this old (dated 2007) would be forward compatible...


On Wed, Mar 11, 2009 at 5:14 PM, mike mike...@gmail.com wrote:
 http://www.intel.com/products/server/storage-systems/ssr212mc2/ssr212mc2-overview.htm
 http://www.intel.com/support/motherboards/server/ssr212mc2/index.htm

 It's hard to use the HAL sometimes.

 I am trying to locate chipset info but having a hard time...

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS on a SAN

2009-03-11 Thread Grant Lowe

Hi Eric,

Thanks.  That scenario makes sense.  I have a better of how to set things up 
now. It's a three-step process, which I didn't realize.

grant



- Original Message 
From: Erik Trimble erik.trim...@sun.com
To: Grant Lowe gl...@sbcglobal.net
Cc: zfs-discuss@opensolaris.org
Sent: Wednesday, March 11, 2009 1:42:06 PM
Subject: Re: [zfs-discuss] ZFS on a SAN

I'm not 100% sure what your question here is, but let me give you a
(hopefully) complete answer:

(1) ZFS is NOT a clustered file system, in the sense that it is NOT
possible for two hosts to have the same LUN mounted at the same time,
even if both are hooked to a SAN and can normally see that LUN.

(2) ZFS can do failover, however.  If you have a LUN from a SAN on
hostA, create a ZFS pool in it, and use as normal.  Should you with to
failover the LUN to hostB, you need to do a 'zpool export zpool' on
hostA, then 'zpool import zpool' on hostB.  If hostA has been lost
completely (hung/died/etc) and you are unable to do an 'export' on it,
you can force the import on hostB via 'zpool import -f zpool'


ZFS requires that you import/export entire POOLS, not just filesystems.
So, given what you seem to want, I'd recommend this:

On the SAN, create (2) LUNs - one for your primary data, and one for
your snapshots/backups.

On hostA, create a zpool on the primary data LUN (call it zpool A), and
another zpool on the backup LUN (zpool B).  Take snapshots on A, then
use 'zfs send' and 'zfs receive' to copy the clone/snapshot over to
zpool B. then 'zpool export B'

On hostB, import the snapshot pool:  'zfs import B'



It might just be as easy to have two independent zpools on each host,
and just do a 'zfs send' on hostA, and 'zfs receive' on hostB to copy
the snapshot/clone over the wire.

-Erik



On Wed, 2009-03-11 at 13:18 -0700, Grant Lowe wrote:
 Hi All,
 
 I'm new on ZFS, so I hope this isn't too basic a question.  I have a host 
 where I setup ZFS.  The Oracle DBAs did their thing and I know have a number 
 of ZFS datasets with their respective clones and snapshots on serverA.  I 
 want to export some of the clones to serverB.  Do I need to zone serverB to 
 see the same LUNs as serverA?  Or does it have to have preexisting, empty 
 LUNs to import the clones?  Please help.  Thanks.
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
-- 
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS on a SAN

2009-03-11 Thread Grant Lowe

Hey Richard,

That explanation help clarify things.  I have another question, but maybe it'll 
be a new topic.  Basically I would like to export some stuff of vxfs file 
system on a different host, and import to zfs.  Is there a way to do that?

grant



- Original Message 
From: Richard Elling richard.ell...@gmail.com
To: Erik Trimble erik.trim...@sun.com
Cc: Grant Lowe gl...@sbcglobal.net; zfs-discuss@opensolaris.org
Sent: Wednesday, March 11, 2009 1:52:31 PM
Subject: Re: [zfs-discuss] ZFS on a SAN

Erik Trimble wrote:
 I'm not 100% sure what your question here is, but let me give you a
 (hopefully) complete answer:

 (1) ZFS is NOT a clustered file system, in the sense that it is NOT
 possible for two hosts to have the same LUN mounted at the same time,
 even if both are hooked to a SAN and can normally see that LUN.
  

Need to be clear on the terminology here.  Yes, it is possible for two
systems to have access to a single LUN and have ZFS file systems.
The ZFS limitation is at the vdev level (partition or slice), which is
below the LUN level.


 (2) ZFS can do failover, however.  If you have a LUN from a SAN on
 hostA, create a ZFS pool in it, and use as normal.  Should you with to
 failover the LUN to hostB, you need to do a 'zpool export zpool' on
 hostA, then 'zpool import zpool' on hostB.  If hostA has been lost
 completely (hung/died/etc) and you are unable to do an 'export' on it,
 you can force the import on hostB via 'zpool import -f zpool'
  

LUN masking or reservations occur at the LUN level, which is why it is
often better (safer) to design with the expectation that a LUN will be
only available to one host at a time.  Or, to say it differently, if you 
think
of one vdev/LUN, then you can innoculate yourself from later grief :-)


 ZFS requires that you import/export entire POOLS, not just filesystems.
 So, given what you seem to want, I'd recommend this:

 On the SAN, create (2) LUNs - one for your primary data, and one for
 your snapshots/backups.

 On hostA, create a zpool on the primary data LUN (call it zpool A), and
 another zpool on the backup LUN (zpool B).  Take snapshots on A, then
 use 'zfs send' and 'zfs receive' to copy the clone/snapshot over to
 zpool B. then 'zpool export B'

 On hostB, import the snapshot pool:  'zfs import B'



 It might just be as easy to have two independent zpools on each host,
 and just do a 'zfs send' on hostA, and 'zfs receive' on hostB to copy
 the snapshot/clone over the wire.
  

+1
-- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Trying to determine if this box will be compatible with Opensolaris or Solaris

2009-03-11 Thread James C. McPherson
On Wed, 11 Mar 2009 17:55:23 -0700
mike mike...@gmail.com wrote:

 Looks like I may have slightly answered my question.
 
 I found that it uses the server board S5000PSL, which is in the HCL:
 http://www.sun.com/bigadmin/hcl/data/systems/details/2944.html
 
 It's not listed in the OpenSolaris one, only the Solaris one. I would
 think something this old (dated 2007) would be forward compatible...

Just because it hasn't been specifically mentioned in the
OpenSolaris or SX hcl, doesn't mean that it isn't supported.
Neither does it necessarily mean that support was dropped.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Trying to determine if this box will be compatible with Opensolaris or Solaris

2009-03-11 Thread David Magda


On Mar 11, 2009, at 20:14, mike wrote:


http://www.intel.com/products/server/storage-systems/ssr212mc2/ssr212mc2-overview.htm
http://www.intel.com/support/motherboards/server/ssr212mc2/index.htm

It's hard to use the HAL sometimes.

I am trying to locate chipset info but having a hard time...


If you know someone who already has the hardware, you can ask them to  
run the Sun Device Detection Tool:


http://www.sun.com/bigadmin/hcl/hcts/device_detect.jsp

It runs under other operating system (Windows, Linux, BSD) AFAIK, so a  
re-install or reboot isn't necessary to see what it comes up with.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Trying to determine if this box will be compatible with Opensolaris or Solaris

2009-03-11 Thread mike
doesnt it require java and x11?

On Wed, Mar 11, 2009 at 6:53 PM, David Magda dma...@ee.ryerson.ca wrote:

 On Mar 11, 2009, at 20:14, mike wrote:


 http://www.intel.com/products/server/storage-systems/ssr212mc2/ssr212mc2-overview.htm
 http://www.intel.com/support/motherboards/server/ssr212mc2/index.htm

 It's hard to use the HAL sometimes.

 I am trying to locate chipset info but having a hard time...

 If you know someone who already has the hardware, you can ask them to run
 the Sun Device Detection Tool:

 http://www.sun.com/bigadmin/hcl/hcts/device_detect.jsp

 It runs under other operating system (Windows, Linux, BSD) AFAIK, so a
 re-install or reboot isn't necessary to see what it comes up with.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Trying to determine if this box will be compatible with Opensolaris or Solaris

2009-03-11 Thread David Magda


On Mar 11, 2009, at 21:59, mike wrote:

On Wed, Mar 11, 2009 at 6:53 PM, David Magda dma...@ee.ryerson.ca  
wrote:




If you know someone who already has the hardware, you can ask them  
to run

the Sun Device Detection Tool:

http://www.sun.com/bigadmin/hcl/hcts/device_detect.jsp

It runs under other operating system (Windows, Linux, BSD) AFAIK,  
so a

re-install or reboot isn't necessary to see what it comes up with.



doesnt it require java and x11?


Yes, it requires Java 1.5+; a GUI is needed, but I don't think X11 is  
specifically required (X is the GUI on Unix-y systems of course). Java  
doesn't specifically need X, it simply uses whatever the OS has.


Looking at the page a bit more, you can run commands on the system and  
save the output  to a file that can be processed by the tool on  
another system:


Apart from testing the current system on which Sun Device Detection  
Tool is invoked, you can also test the device data files that are  
generated from the external systems. To test the external device  
data files, print the PCI configuration of the external systems to a  
text file by using the following commands:


• prtconf -pv on Solaris OS.
• lspci -vv -n on Linux OS.
• reg query hklm\system\currentcontrolset\enum\pci /s on Windows OS.



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] User quota design discussion..

2009-03-11 Thread Jorgen Lundman


In the style of a discussion over a beverage, and talking about 
user-quotas on ZFS, I recently pondered a design for implementing user 
quotas on ZFS after having far too little sleep.


It is probably nothing new, but I would be curious what you experts 
think of the feasibility of implementing such a system and/or whether or 
not it would even realistically work.


I'm not suggesting that someone should do the work, or even that I will, 
but rather in the interest of chatting about it.


Feel free to ridicule me as required! :)

Thoughts:

Here at work we would like to have user quotas based on uid (and 
presumably gid) to be able to fully replace the NetApps we run. Current 
ZFS are not good enough for our situation. We simply can not mount 
500,000 file-systems on all the NFS clients. Nor do all servers we run 
support mirror-mounts. Nor do auto-mount see newly created directories 
without a full remount.


Current UFS-style-user-quotas are very exact. To the byte even. We do 
not need this precision. If a user has 50MB of quota, and they are able 
to reach 51MB usage, then that is acceptable to us. Especially since 
they have to go under 50MB to be able to write new data, anyway.


Instead of having complicated code in the kernel layer, slowing down the 
file-system with locking and semaphores (and perhaps avoiding learning 
indepth ZFS code?), I was wondering if a more simplistic setup could be 
designed, that would still be acceptable. I will use the word 
'acceptable' a lot. Sorry.


My thoughts are that the ZFS file-system will simply write a 
'transaction log' on a pipe. By transaction log I mean uid, gid and 
'byte count changed'. And by pipe I don't necessarily mean pipe(2), but 
it could be a fifo, pipe or socket. But currently I'm thinking 
'/dev/quota' style.


User-land will then have a daemon, whether or not it is one daemon per 
file-system or really just one daemon does not matter. This process will 
open '/dev/quota' and empty the transaction log entries constantly. Take 
the uid,gid entries and update the byte-count in its database. How we 
store this database is up to us, but since it is in user-land it should 
have more flexibility, and is not as critical to be fast as it would 
have to be in kernel.


The daemon process can also grow in number of threads as demand increases.

Once a user's quota reaches the limit (note here that /the/ call to 
write() that goes over the limit will succeed, and probably a couple 
more after. This is acceptable) the process will blacklist the uid in 
kernel. Future calls to creat/open(CREAT)/write/(insert list of calls) 
will be denied. Naturally calls to unlink/read etc should still succeed. 
If the uid goes under the limit, the uid black-listing will be removed.


If the user-land process crashes or dies, for whatever reason, the 
buffer of the pipe will grow in the kernel. If the daemon is restarted 
sufficiently quickly, all is well, it merely needs to catch up. If the 
pipe does ever get full and items have to be discarded, a full-scan will 
be required of the file-system. Since even with UFS quotas we need to 
occasionally run 'quotacheck', it would seem this too, is acceptable (if 
undesirable).


If you have no daemon process running at all, you have no quotas at all. 
But the same can be said about quite a few daemons. The administrators 
need to adjust their usage.


I can see a complication with doing a rescan. How could this be done 
efficiently? I don't know if there is a neat way to make this happen 
internally to ZFS, but from a user-land only point of view, perhaps a 
snapshot could be created (synchronised with the /dev/quota pipe 
reading?) and start a scan on the snapshot, while still processing 
kernel log. Once the scan is complete, merge the two sets.


Advantages are that only small hooks are required in ZFS. The byte 
updates, and the blacklist with checks for being blacklisted.


Disadvantages are that it is loss of precision, and possibly slower 
rescans? Sanity?


But I do not really know the internals of ZFS, so I might be completely 
wrong, and everyone is laughing already.


Discuss?

Lund

--
Jorgen Lundman   | lund...@lundman.net
Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
Shibuya-ku, Tokyo| +81 (0)90-5578-8500  (cell)
Japan| +81 (0)3 -3375-1767  (home)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Maidak Alexander J
If you're having issues with a disk contoller or disk IO driver its highly 
likely that a savecore to disk after the panic will fail.  I'm not sure how to 
work around this, maybe a dedicated dump device not on a controller that uses a 
different driver then the one that you're having issues with?

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
Sent: Wednesday, March 11, 2009 4:45 PM
To: Richard Elling
Cc: Marc Bevand; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] reboot when copying large amounts of data

I guess I didn't make it clear that I had already tried using savecore to 
retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out of space 
on the dump device:

r...@host:~# dumpadm
  Dump content: kernel pages
   Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore directory: 
/var/crash/host
  Savecore enabled: yes

I was using the -L option only to try to get some idea of why the system load 
was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The 
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right 
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the 
 core will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the 
 core from the last dump.
 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Why did my zvol shrink ?

2009-03-11 Thread Gavin Maltby



Brian H. Nelson wrote:
I'm doing a little testing and I hit a strange point. Here is a zvol 
(clone)


pool1/volclone  type volume -
pool1/volclone  origin   pool1/v...@diff1   -
pool1/volclone  reservation  none   default
pool1/volclone  volsize  191G   -
pool1/volclone  volblocksize 8K -

The zvol has UFS on it. It has always been 191G and we've never 
attempted to resize it. However, if I just try to grow it, it gives me 
an error:


-bash-3.00# growfs 
/dev/zvol/rdsk/pool1/volclone400555998 
sectors  current size of 400556032 sectors


Is the zvol is somehow smaller than it was originally? How/why?


I think ufs requires an integer number of cylinder groups and I'm
guessing the volume size you have presented it with is somewhere
in between - so it has rounded down to the largest cylinder
group boundary less than or equal to the device size.

Gavin

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Blake
My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:
 If you're having issues with a disk contoller or disk IO driver its highly 
 likely that a savecore to disk after the panic will fail.  I'm not sure how 
 to work around this, maybe a dedicated dump device not on a controller that 
 uses a different driver then the one that you're having issues with?

 -Original Message-
 From: zfs-discuss-boun...@opensolaris.org 
 [mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
 Sent: Wednesday, March 11, 2009 4:45 PM
 To: Richard Elling
 Cc: Marc Bevand; zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] reboot when copying large amounts of data

 I guess I didn't make it clear that I had already tried using savecore to 
 retrieve the core from the dump device.

 I added a larger zvol for dump, to make sure that I wasn't running out of 
 space on the dump device:

 r...@host:~# dumpadm
      Dump content: kernel pages
       Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore 
 directory: /var/crash/host
  Savecore enabled: yes

 I was using the -L option only to try to get some idea of why the system load 
 was climbing to 1 during a simple file copy.



 On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
 wrote:
 Blake wrote:

 I'm attaching a screenshot of the console just before reboot.  The
 dump doesn't seem to be working, or savecore isn't working.

 On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


 I'm working on testing this some more by doing a savecore -L right
 after I start the copy.



 savecore -L is not what you want.

 By default, for OpenSolaris, savecore on boot is disabled.  But the
 core will have been dumped into the dump slice, which is not used for swap.
 So you should be able to run savecore at a later time to collect the
 core from the last dump.
 -- richard


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] reboot when copying large amounts of data

2009-03-11 Thread Nathan Kroenert

Hm -

Crashes, or hangs? Moreover - how do you know a CPU is pegged?

Seems like we could do a little more discovery on what the actual 
problem here is, as I can read it about 4 different ways.


By this last piece of information, I'm guessing the system does not 
crash, but goes really really slow??


Crash == panic == we see stack dump on console and try to take a dump
hang == nothing works == no response - might be worth looking at mdb -K
or booting with a -k on the boot line.

So - are we crashing, hanging, or something different?

It might simply be that you are eating up all your memory, and your 
physical backing storage is taking a while to catch up?


Nathan.

Blake wrote:

My dump device is already on a different controller - the motherboards
built-in nVidia SATA controller.

The raidz2 vdev is the one I'm having trouble with (copying the same
files to the mirrored rpool on the nVidia controller work nicely).  I
do notice that, when using cp to copy the files to the raidz2 pool,
load on the machine climbs steadily until the crash, and one proc core
pegs at 100%.

Frustrating, yes.

On Thu, Mar 12, 2009 at 12:31 AM, Maidak Alexander J
maidakalexand...@johndeere.com wrote:

If you're having issues with a disk contoller or disk IO driver its highly 
likely that a savecore to disk after the panic will fail.  I'm not sure how to 
work around this, maybe a dedicated dump device not on a controller that uses a 
different driver then the one that you're having issues with?

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Blake
Sent: Wednesday, March 11, 2009 4:45 PM
To: Richard Elling
Cc: Marc Bevand; zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] reboot when copying large amounts of data

I guess I didn't make it clear that I had already tried using savecore to 
retrieve the core from the dump device.

I added a larger zvol for dump, to make sure that I wasn't running out of space 
on the dump device:

r...@host:~# dumpadm
 Dump content: kernel pages
  Dump device: /dev/zvol/dsk/rpool/bigdump (dedicated) Savecore directory: 
/var/crash/host
 Savecore enabled: yes

I was using the -L option only to try to get some idea of why the system load 
was climbing to 1 during a simple file copy.



On Wed, Mar 11, 2009 at 4:58 PM, Richard Elling richard.ell...@gmail.com 
wrote:

Blake wrote:

I'm attaching a screenshot of the console just before reboot.  The
dump doesn't seem to be working, or savecore isn't working.

On Wed, Mar 11, 2009 at 11:33 AM, Blake blake.ir...@gmail.com wrote:


I'm working on testing this some more by doing a savecore -L right
after I start the copy.



savecore -L is not what you want.

By default, for OpenSolaris, savecore on boot is disabled.  But the
core will have been dumped into the dump slice, which is not used for swap.
So you should be able to run savecore at a later time to collect the
core from the last dump.
-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


--
//
// Nathan Kroenert  nathan.kroen...@sun.com //
// Systems Engineer Phone:  +61 3 9869-6255 //
// Sun Microsystems Fax:+61 3 9869-6288 //
// Level 7, 476 St. Kilda Road  Mobile: 0419 305 456//
// Melbourne 3004   VictoriaAustralia   //
//
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss