Re: [zfs-discuss] Mount External USB cdrom on zfs

2009-01-27 Thread Ian Collins
iman habibi wrote:
 Dear support
 when i connect my external usb dvdrom to the sparc machine which has
 installed solaris 10u6 based zfs file system,,it return this error:

Your ZFS question is?

DVDs uue the HSFS filesystem.

One good place for general Solaris questions is comp.unix.solaris.

-- 
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Mount External USB cdrom on zfs

2009-01-27 Thread Johan Hartzenberg
On Tue, Jan 27, 2009 at 9:49 AM, iman habibi iman.hab...@gmail.com wrote:

 Dear support
 when i connect my external usb dvdrom to the sparc machine which has
 installed solaris 10u6 based zfs file system,,it return this error:

 bash-3.00# mount /dev/dsk/c1t0d0s0 /dvd/
 Jan 27 11:08:41 global ufs: NOTICE: mount: not a UFS magic number (0x0)
 mount: /dev/dsk/c1t0d0s0 is not this fstype


On Solaris, by default mount assumes that the file system type to be mounted
is UFS.

Basically, when mounting anything other than UFS, you need to specify what
it is.  The two exceptions are:
a) When the vfstab can give information about what file system type to
expect, or
b) When using zfs mount (which only mounts zfs file systems)

So essentially you need to specify the file system type on the mount
command, like this:

mount -F hsfs -r /dev/dsk/c1t0d0s0 /dvd/

The -r is for read-only.

You can also (optionally) add a line to your /etc/vfstab file, like this:
/dev/dsk/c1t0d0s0 - /dvd hsfs - no ro

With this in place you can then mount the disk using:

mount /dvd

(It will learn the device, read-only flag, and the file system type from
/etc/vfstab automatically)

Of course I am wondering why you don't use the auto-mounter.

There are of course other things you could do.  You could change the
default file system type in /etc/default/fs, but that is not recommended.
You could write a little script to mount disks.  etc etc etc.

For more info, read man mount and man vfstab

Cheers
  _Johan


-- 
Any sufficiently advanced technology is indistinguishable from magic.
   Arthur C. Clarke

My blog: http://initialprogramload.blogspot.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] freeze/ thaw zfs file system

2009-01-27 Thread ajit jain
Hi All,

Can we freeze and thaw a zfs file-system either from user land (lockfs
for ufs) or from the kernel space or through ioctl?

Thanks in advance for your help.

Best regards,
ajit
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] freeze/ thaw zfs file system

2009-01-27 Thread Andrew Gabriel
ajit jain wrote:
 Hi All,

 Can we freeze and thaw a zfs file-system either from user land (lockfs
 for ufs) or from the kernel space or through ioctl?
   

Can you step back a level and explain what you're trying to achieve?

Freezing UFS is to get round ufs-specific issues which don't apply to 
ZFS, so you maybe asking the wrong question.

-- 
Andrew
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Problems using ZFS on Smart Array P400

2009-01-27 Thread Alex
I am having trouble getting ZFS to behave as I would expect.

I am using the HP driver (cpqary3) for the Smart Array P400 (in a HP Proliant 
DL385 G2) with 10k 2.5 146GB SAS drives. The drives appear correctly, however 
due to the controller not offering JBOD functionality I had to configure each 
drive as a RAID0 logical drive.

Everything appears to work fine, the drives are detected and I created a mirror 
for the OS to install to and an additional raidz2 array with the remaining 6 
discs.

But when I remove a disc and then reinsert it I cannot get ZFS to accept it 
back into the array see bellow for the details.

I thought it might be a problem with using the whole discs eg: c1t*d0 so I 
created a single partition on each and used that, but had the same results. The 
module seem to detect the drive has been reinserted successfully but the OS 
doesn't seem to want to write to it.

Any help would be most appreciated as I would much prefer to use ZFS's software 
capabilities rather than the hardware card in the machine.

When rebooting the system the Array BIOS also displays some interesting 
behavior.

### BIOS Output

1792-Slot 1 Drive Array - Valid Data Found in the Array Accelerator
Data will automatically be written to the drive array
1779-Slot 1 Drive Array - Replacement drive(s) detected OR previously failed 
drives(s) no appear to be operational
POrt 2I: Box1: Bay3
Logical drives(s) disabled due to possible data loss.
Select F1 to continue with logical drive(s) disabled
Select F2 to accept data loss and to re-enable logical drive(s)

 Terminal output

bash-3.00# zpool status test

pool: test
state: DEGRADED
status: One or more devices could not be opened. Sufficient replicas exist for
the pool to continue functioning in a degraded state.
action: Attach the missing device and online it using 'zpool online'.
see: http://www.sun.com/msg/ZFS-8000-2Q
scrub: resilver completed after 0h0m with 0 errors on Tue Jan 27 03:30:16 2009
config:

NAME STATE READ WRITE CKSUM
test DEGRADED 0 0 0
raidz2 DEGRADED 0 0 0
c1t2d0p0 ONLINE 0 0 0
c1t3d0p0 ONLINE 0 0 0
c1t4d0p0 ONLINE 0 0 0
c1t5d0p0 UNAVAIL 0 0 0 cannot open
c1t6d0p0 ONLINE 0 0 0
c1t8d0p0 ONLINE 0 0 0

errors: No known data errors
bash-3.00# zpool online test c1t5d0p0
warning: device 'c1t5d0p0' onlined, but remains in faulted state
use 'zpool replace' to replace devices that are no longer present

bash-3.00# dmesg

Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] NOTICE: Smart Array 
P400 Controller
Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] Hot-plug drive 
inserted, Port: 2I Box: 1 Bay: 3
Jan 27 03:27:40 unknown cpqary3: [ID 479030 kern.notice] Configured Drive ? 
... YES
Jan 27 03:27:40 unknown cpqary3: [ID 10 kern.notice]
Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] NOTICE: Smart Array 
P400 Controller
Jan 27 03:27:40 unknown cpqary3: [ID 834734 kern.notice] Media exchange 
detected, logical drive 6
Jan 27 03:27:40 unknown cpqary3: [ID 10 kern.notice]
...
Jan 27 03:36:24 unknown scsi: [ID 107833 kern.warning] WARNING: 
/p...@38,0/pci1166,1...@10/pci103c,3...@0/s...@5,0 (sd6):
Jan 27 03:36:24 unknown SYNCHRONIZE CACHE command failed (5)
...
Jan 27 03:47:58 unknown scsi: [ID 107833 kern.warning] WARNING: 
/p...@38,0/pci1166,1...@10/pci103c,3...@0/s...@5,0 (sd6):
Jan 27 03:47:58 unknown drive offline
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problems using ZFS on Smart Array P400

2009-01-27 Thread Fajar A. Nugraha
On Tue, Jan 27, 2009 at 7:16 PM, Alex a...@pancentric.com wrote:
 I am using the HP driver (cpqary3) for the Smart Array P400 (in a HP Proliant 
 DL385 G2) with 10k 2.5 146GB SAS drives. The drives appear correctly, 
 however due to the controller not offering JBOD functionality I had to 
 configure each drive as a RAID0 logical drive.

Ouch. Short comment, it might not worth it. Seriously.

Does the P400 have a battery-backed cache? If yes, it will be MUCH
easier to simply let it handle RAID5/10 and use stripe config for zfs.
HW controllers with battery-backed cache will reduce the possibility
of raid5 write hole. Depending on what your goal is, it might be the
best choice.


 But when I remove a disc and then reinsert it I cannot get ZFS to accept it 
 back into the array see bellow for the details.

I don't think it's zfs' fault.

 Any help would be most appreciated as I would much prefer to use ZFS's 
 software capabilities rather than the hardware card in the machine.

Yet you're stuck with a hardware that practically does not allow you
to do just that.

 1779-Slot 1 Drive Array - Replacement drive(s) detected OR previously failed 
 drives(s) no appear to be operational
 POrt 2I: Box1: Bay3
 Logical drives(s) disabled due to possible data loss.
 Select F1 to continue with logical drive(s) disabled
 Select F2 to accept data loss and to re-enable logical drive(s)

From this output, the HW controller refuses to enable the removed
disc, so that the OS (and thus zfs) can't see it yet. You can enable
it from BIOS during booting (which kinda defeat the whole hot-swap
thing), or you might be able to find a way (using ILO perhaps, or with
some HP-supplied software installed on the OS that can talk to P400)
to enable it without having to reboot.

Regards,

Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problems using ZFS on Smart Array P400

2009-01-27 Thread Edmund White
I'm testing the same thing on a DL380 G5 with P400 controller. I set
individual RAID 0 logical drives for each disk. I ended up with the same
result upon drive removal. I'm looking into whether the hpacucli array
command line utility will let me re-enable a logical drive from its
interface.

-- 
Edmund William White
ewwh...@mac.com



 From: Alex a...@pancentric.com
 Date: Tue, 27 Jan 2009 04:16:56 -0800 (PST)
 To: zfs-discuss@opensolaris.org
 Subject: [zfs-discuss] Problems using ZFS on Smart Array P400
 
 I am having trouble getting ZFS to behave as I would expect.
 
 I am using the HP driver (cpqary3) for the Smart Array P400 (in a HP Proliant
 DL385 G2) with 10k 2.5 146GB SAS drives. The drives appear correctly, however
 due to the controller not offering JBOD functionality I had to configure each
 drive as a RAID0 logical drive.
 
 Everything appears to work fine, the drives are detected and I created a
 mirror for the OS to install to and an additional raidz2 array with the
 remaining 6 discs.
 
 But when I remove a disc and then reinsert it I cannot get ZFS to accept it
 back into the array see bellow for the details.
 
 I thought it might be a problem with using the whole discs eg: c1t*d0 so I
 created a single partition on each and used that, but had the same results.
 The module seem to detect the drive has been reinserted successfully but the
 OS doesn't seem to want to write to it.
 
 Any help would be most appreciated as I would much prefer to use ZFS's
 software capabilities rather than the hardware card in the machine.
 
 When rebooting the system the Array BIOS also displays some interesting
 behavior.
 
 ### BIOS Output
 
 1792-Slot 1 Drive Array - Valid Data Found in the Array Accelerator
 Data will automatically be written to the drive array
 1779-Slot 1 Drive Array - Replacement drive(s) detected OR previously failed
 drives(s) no appear to be operational
 POrt 2I: Box1: Bay3
 Logical drives(s) disabled due to possible data loss.
 Select F1 to continue with logical drive(s) disabled
 Select F2 to accept data loss and to re-enable logical drive(s)
 
  Terminal output
 
 bash-3.00# zpool status test
 
 pool: test
 state: DEGRADED
 status: One or more devices could not be opened. Sufficient replicas exist for
 the pool to continue functioning in a degraded state.
 action: Attach the missing device and online it using 'zpool online'.
 see: http://www.sun.com/msg/ZFS-8000-2Q
 scrub: resilver completed after 0h0m with 0 errors on Tue Jan 27 03:30:16 2009
 config:
 
 NAME STATE READ WRITE CKSUM
 test DEGRADED 0 0 0
 raidz2 DEGRADED 0 0 0
 c1t2d0p0 ONLINE 0 0 0
 c1t3d0p0 ONLINE 0 0 0
 c1t4d0p0 ONLINE 0 0 0
 c1t5d0p0 UNAVAIL 0 0 0 cannot open
 c1t6d0p0 ONLINE 0 0 0
 c1t8d0p0 ONLINE 0 0 0
 
 errors: No known data errors
 bash-3.00# zpool online test c1t5d0p0
 warning: device 'c1t5d0p0' onlined, but remains in faulted state
 use 'zpool replace' to replace devices that are no longer present
 
 bash-3.00# dmesg
 
 Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] NOTICE: Smart Array
 P400 Controller
 Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] Hot-plug drive
 inserted, Port: 2I Box: 1 Bay: 3
 Jan 27 03:27:40 unknown cpqary3: [ID 479030 kern.notice] Configured Drive ?
 ... YES
 Jan 27 03:27:40 unknown cpqary3: [ID 10 kern.notice]
 Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] NOTICE: Smart Array
 P400 Controller
 Jan 27 03:27:40 unknown cpqary3: [ID 834734 kern.notice] Media exchange
 detected, logical drive 6
 Jan 27 03:27:40 unknown cpqary3: [ID 10 kern.notice]
 ...
 Jan 27 03:36:24 unknown scsi: [ID 107833 kern.warning] WARNING:
 /p...@38,0/pci1166,1...@10/pci103c,3...@0/s...@5,0 (sd6):
 Jan 27 03:36:24 unknown SYNCHRONIZE CACHE command failed (5)
 ...
 Jan 27 03:47:58 unknown scsi: [ID 107833 kern.warning] WARNING:
 /p...@38,0/pci1166,1...@10/pci103c,3...@0/s...@5,0 (sd6):
 Jan 27 03:47:58 unknown drive offline
 -- 
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool status -x strangeness

2009-01-27 Thread Ben Miller
I forgot the pool that's having problems was recreated recently so it's already 
at zfs version 3.  I just did a 'zfs upgrade -a' for another pool, but some of 
those filesystems failed since they are busy and couldn't be unmounted.

# zfs upgrade -a
cannot unmount '/var/mysql': Device busy
cannot unmount '/var/postfix': Device busy

6 filesystems upgraded
821 filesystems already at this version

Ben

 You can upgrade live.  'zfs upgrade' with no
 arguments shows you the  
 zfs version status of filesystems present without
 upgrading.
 
 
 
 On Jan 24, 2009, at 10:19 AM, Ben Miller
 mil...@eecis.udel.edu wrote:
 
  We haven't done 'zfs upgrade ...' any.  I'll give
 that a try the  
  next time the system can be taken down.
 
  Ben
 
  A little gotcha that I found in my 10u6 update
  process was that 'zpool
  upgrade [poolname]' is not the same as 'zfs
 upgrade
  [poolname]/[filesystem(s)]'
 
  What does 'zfs upgrade' say?  I'm not saying this
 is
  the source of
  your problem, but it's a detail that seemed to
 affect
  stability for
  me.
 
 
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] freeze/ thaw zfs file system

2009-01-27 Thread ajit jain
Hi Andrew,

I am writing a filtering device which tracks the write to the
file-system. I am doing it for ufs, vxfs and for zfs. Sometime for
consistent point I need to freeze the file-system which flushes dirty
block to the disk and block every IO on the top level. So, for ufs and
vxfs I got lockfs and VX_FREEZE/VX_THAW, but with zfs I didn't get the
luck.

Thanks for your response.

Best regards,
ajit

On Tue, Jan 27, 2009 at 5:34 PM, Andrew Gabriel
agabr...@opensolaris.org wrote:
 ajit jain wrote:

 Hi All,

 Can we freeze and thaw a zfs file-system either from user land (lockfs
 for ufs) or from the kernel space or through ioctl?


 Can you step back a level and explain what you're trying to achieve?

 Freezing UFS is to get round ufs-specific issues which don't apply to ZFS,
 so you maybe asking the wrong question.

 --
 Andrew

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problems using ZFS on Smart Array P400

2009-01-27 Thread Craig Morgan
You need to step back and appreciate that the manner in which you are  
presenting Solaris with disks is the problem and not necessarily ZFS.

As your storage system is incapable of JBOD operation, you have  
decided to present each disk as a 'simple' RAID0 volume. Whilst this  
looks like a 'pass-thru' access method to the disk and its contents,  
it is far from it. The HW RAID sub-system is creating a logical volume  
based on this single spindle (in exactly the same way it would be for  
multiple spindles, aka a stripe), metadata is recorded by the RAID  
system with regard to the make-up of said volume.

The important issue here is that you have a non-redundant RAID (!)  
config, hence a single failure (in this case your single spindle  
failure) causes the RAID sub-system to declare the volume (and hence  
its operational status) as failed, this in turn is declared to the OS  
as a failed volume. At this juncture, intervention is normally  
necessary to re-destroy/re-create a volume (remember no redundancy--- 
so this is manual!) and hence re-present it to the OS (which will find  
a new UID for the volume and treat it as a new device). On occasions  
it may be possible to intervene and resurrect a volume by manually  
overriding the status of the RAID0 volume, but in many HW RAID systems  
this is not to be recommended.

In short, you've got more abstractions (layers) in place than you need/ 
desire and that is fundamentally the cause of your problem ... either  
plump for a simpler array or swallow some loss of transparency in the  
ZFS layer and present redundant RAID sets from your array, but live  
with the consequences of increased admin and complexity and some loss  
of transparency/protection---but hopefully the RAID sub-system will be  
capable of automated recovery in most circumstances of simple failures.

Craig

On 27 Jan 2009, at 13:00, Edmund White wrote:

 I'm testing the same thing on a DL380 G5 with P400 controller. I set
 individual RAID 0 logical drives for each disk. I ended up with the  
 same
 result upon drive removal. I'm looking into whether the hpacucli array
 command line utility will let me re-enable a logical drive from its
 interface.

 -- 
 Edmund William White
 ewwh...@mac.com



 From: Alex a...@pancentric.com
 Date: Tue, 27 Jan 2009 04:16:56 -0800 (PST)
 To: zfs-discuss@opensolaris.org
 Subject: [zfs-discuss] Problems using ZFS on Smart Array P400

 I am having trouble getting ZFS to behave as I would expect.

 I am using the HP driver (cpqary3) for the Smart Array P400 (in a  
 HP Proliant
 DL385 G2) with 10k 2.5 146GB SAS drives. The drives appear  
 correctly, however
 due to the controller not offering JBOD functionality I had to  
 configure each
 drive as a RAID0 logical drive.

 Everything appears to work fine, the drives are detected and I  
 created a
 mirror for the OS to install to and an additional raidz2 array with  
 the
 remaining 6 discs.

 But when I remove a disc and then reinsert it I cannot get ZFS to  
 accept it
 back into the array see bellow for the details.

 I thought it might be a problem with using the whole discs eg:  
 c1t*d0 so I
 created a single partition on each and used that, but had the same  
 results.
 The module seem to detect the drive has been reinserted  
 successfully but the
 OS doesn't seem to want to write to it.

 Any help would be most appreciated as I would much prefer to use  
 ZFS's
 software capabilities rather than the hardware card in the machine.

 When rebooting the system the Array BIOS also displays some  
 interesting
 behavior.

 ### BIOS Output

 1792-Slot 1 Drive Array - Valid Data Found in the Array Accelerator
 Data will automatically be written to the drive array
 1779-Slot 1 Drive Array - Replacement drive(s) detected OR  
 previously failed
 drives(s) no appear to be operational
 POrt 2I: Box1: Bay3
 Logical drives(s) disabled due to possible data loss.
 Select F1 to continue with logical drive(s) disabled
 Select F2 to accept data loss and to re-enable logical drive(s)

  Terminal output

 bash-3.00# zpool status test

 pool: test
 state: DEGRADED
 status: One or more devices could not be opened. Sufficient  
 replicas exist for
 the pool to continue functioning in a degraded state.
 action: Attach the missing device and online it using 'zpool online'.
 see: http://www.sun.com/msg/ZFS-8000-2Q
 scrub: resilver completed after 0h0m with 0 errors on Tue Jan 27  
 03:30:16 2009
 config:

 NAME STATE READ WRITE CKSUM
 test DEGRADED 0 0 0
 raidz2 DEGRADED 0 0 0
 c1t2d0p0 ONLINE 0 0 0
 c1t3d0p0 ONLINE 0 0 0
 c1t4d0p0 ONLINE 0 0 0
 c1t5d0p0 UNAVAIL 0 0 0 cannot open
 c1t6d0p0 ONLINE 0 0 0
 c1t8d0p0 ONLINE 0 0 0

 errors: No known data errors
 bash-3.00# zpool online test c1t5d0p0
 warning: device 'c1t5d0p0' onlined, but remains in faulted state
 use 'zpool replace' to replace devices that are no longer present

 bash-3.00# dmesg

 Jan 27 03:27:40 unknown cpqary3: [ID 823470 kern.notice] NOTICE:  
 Smart 

[zfs-discuss] Crazy Problem with

2009-01-27 Thread Henri Meddox
Hi Folks,
call me a lernen ;-)

I got a crazy Problem with zpool list and the size of my pool:

created zpool create raidz2 hdd1 hdd2 hdd3 - each hdd is about 1GB.
 
zpool list shows me a size of 2.95GB - shouldn't this bis online 1GB?

After creating a file about 500MB - Capacity is shown as 50 % - The right 
value?

Is this a known bug / feature?

THX
Henri
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Crazy Problem with

2009-01-27 Thread Mika Borner
Henri Meddox wrote:
 Hi Folks,
 call me a lernen ;-)

 I got a crazy Problem with zpool list and the size of my pool:

 created zpool create raidz2 hdd1 hdd2 hdd3 - each hdd is about 1GB.
  
 zpool list shows me a size of 2.95GB - shouldn't this bis online 1GB?

 After creating a file about 500MB - Capacity is shown as 50 % - The right 
 value?

 Is this a known bug / feature?

   

You're lucky. Ben just wrote about it :-)

http://www.cuddletech.com/blog/pivot/entry.php?id=1013



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Crazy Problem with

2009-01-27 Thread Mika Borner
Mika Borner wrote:

 You're lucky. Ben just wrote about it :-)

 http://www.cuddletech.com/blog/pivot/entry.php?id=1013


   
Oops, should have read your message completly :-) Anyway you can 
lernen something from it...
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problems using ZFS on Smart Array P400

2009-01-27 Thread Edmund White
Given that I have lots of ProLiant equipment, are there any recommended
controllers that would work in this situation? Is this an issue unique to
the Smart Array controllers? If I do choose to use some level of hardware
RAID on the existing Smart Array P400, what's the best way to use it with
ZFS (assume 8 disks with an emphasis on capacity)?

-- 
Edmund William White
ewwh...@mac.com



 From: Craig Morgan craig.mor...@sun.com
 Date: Tue, 27 Jan 2009 13:54:46 +
 To: Edmund White ewwh...@mac.com
 Cc: Alex a...@pancentric.com, zfs-discuss@opensolaris.org
 Subject: Re: [zfs-discuss] Problems using ZFS on Smart Array P400
 
 You need to step back and appreciate that the manner in which you are
 presenting Solaris with disks is the problem and not necessarily ZFS.
 
 As your storage system is incapable of JBOD operation, you have
 decided to present each disk as a 'simple' RAID0 volume. Whilst this
 looks like a 'pass-thru' access method to the disk and its contents,
 it is far from it. The HW RAID sub-system is creating a logical volume
 based on this single spindle (in exactly the same way it would be for
 multiple spindles, aka a stripe), metadata is recorded by the RAID
 system with regard to the make-up of said volume.
 
 The important issue here is that you have a non-redundant RAID (!)
 config, hence a single failure (in this case your single spindle
 failure) causes the RAID sub-system to declare the volume (and hence
 its operational status) as failed, this in turn is declared to the OS
 as a failed volume. At this juncture, intervention is normally
 necessary to re-destroy/re-create a volume (remember no redundancy---
 so this is manual!) and hence re-present it to the OS (which will find
 a new UID for the volume and treat it as a new device). On occasions
 it may be possible to intervene and resurrect a volume by manually
 overriding the status of the RAID0 volume, but in many HW RAID systems
 this is not to be recommended.
 
 In short, you've got more abstractions (layers) in place than you need/
 desire and that is fundamentally the cause of your problem ... either
 plump for a simpler array or swallow some loss of transparency in the
 ZFS layer and present redundant RAID sets from your array, but live
 with the consequences of increased admin and complexity and some loss
 of transparency/protection---but hopefully the RAID sub-system will be
 capable of automated recovery in most circumstances of simple failures.
 
 Craig
 
 On 27 Jan 2009, at 13:00, Edmund White wrote:
 
 
 -- 
 Craig
 
 Craig Morgan
 t: +44 (0)791 338 3190
 f: +44 (0)870 705 1726
 e: craig.mor...@sun.com
 
 ~ 
 
   NOTICE:  This email message is for the sole use of the intended
   recipient(s) and may contain confidential and privileged information.
   Any unauthorized review, use, disclosure or distribution is
 prohibited.
   If you are not the intended recipient, please contact the sender by
   reply email and destroy all copies of the original message.
 ~ 
 
 
 


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] how to fix zpool with corrupted disk?

2009-01-27 Thread Toby Thain

On 26-Jan-09, at 8:15 PM, Miles Nordin wrote:

 js == Jakov Sosic jso...@gmail.com writes:
 tt == Toby Thain t...@telegraphics.com.au writes:

 js Yes but that will do the complete resilvering, and I just want
 js to fix the corrupted blocks... :)

 tt What you are asking for is impossible, since ZFS cannot know
 tt which blocks are corrupted without actually checking them

 yeah of course you have to read every (occupied) block, but he's still
 not asking for something completely nonsensical.  What if the good
 drive has a latent sector error in one of the blocks that hasn't been
 scribbled over on the bad drive?  scrub could heal the error if not
 for the ``too many errors'' fault, while 'zpool replace' could not
 heal it.

Yes he's asking for a scrub. I was just pointing out that resilver  
is something else entirely :)

--Toby

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ? Changing storage pool serial number

2009-01-27 Thread Tim Thomas
Hi

I took a look at the archives and I have seen a few threads about using 
array block level snapshots with ZFS and how we face the old issue 
that we used to see with logical volumes and unique IDs (quite 
correctly) stopping the same volume being presented twice to the same 
server.

IHAC wanting to do this with a ZFS storage pool and an LSI array. All 
works well when they present the array snapshot to hosts other than the 
one where the original zpool resides but (as expected) we cannot import 
the zpool on the array snapshot to the original host..though no error 
message is issues, it just does not work.

So, I am assuming that this is because the zpools have the same serial 
number (happy to be corrected).

My question is are there any RFEs that will enable a zpool's serial 
number to be changed (if that is what is required to make this work).

It may be that this is an awful idea..in which case I am happy to hear 
that as well and will feed that back to the customer.

Thanks

Tim

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ? Changing storage pool serial number

2009-01-27 Thread Jim Dunham
Hi Tim,

 I took a look at the archives and I have seen a few threads about  
 using
 array block level snapshots with ZFS and how we face the old issue
 that we used to see with logical volumes and unique IDs (quite
 correctly) stopping the same volume being presented twice to the same
 server.

 IHAC wanting to do this with a ZFS storage pool and an LSI array. All
 works well when they present the array snapshot to hosts other than  
 the
 one where the original zpool resides but (as expected) we cannot  
 import
 the zpool on the array snapshot to the original host..though no error
 message is issues, it just does not work.

 So, I am assuming that this is because the zpools have the same serial
 number (happy to be corrected).

 My question is are there any RFEs that will enable a zpool's serial
 number to be changed (if that is what is required to make this work).

http://bugs.opensolaris.org/view_bug.do?bug_id=5097228

 It may be that this is an awful idea..in which case I am happy to hear
 that as well and will feed that back to the customer.

For both controller-based  host-based snapshots, replicas, even iSCSI  
LUs, this would be an awful [good] idea.  :-)

- Jim



 Thanks

 Tim

 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] freeze/ thaw zfs file system

2009-01-27 Thread Richard Elling
ajit jain wrote:
 Hi Andrew,

 I am writing a filtering device which tracks the write to the
 file-system. I am doing it for ufs, vxfs and for zfs. Sometime for
 consistent point I need to freeze the file-system which flushes dirty
 block to the disk and block every IO on the top level. So, for ufs and
 vxfs I got lockfs and VX_FREEZE/VX_THAW, but with zfs I didn't get the
 luck.
   

I see no need to freeze the file system, or maybe I would say, I see
no benefit to freezing a file system.  Perhaps your needs will be met
by the snapshot feature which will also cause a sync(2).
 -- richard

 Thanks for your response.

 Best regards,
 ajit

 On Tue, Jan 27, 2009 at 5:34 PM, Andrew Gabriel
 agabr...@opensolaris.org wrote:
   
 ajit jain wrote:
 
 Hi All,

 Can we freeze and thaw a zfs file-system either from user land (lockfs
 for ufs) or from the kernel space or through ioctl?

   
 Can you step back a level and explain what you're trying to achieve?

 Freezing UFS is to get round ufs-specific issues which don't apply to ZFS,
 so you maybe asking the wrong question.

 --
 Andrew

 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Crazy Problem with

2009-01-27 Thread Richard Elling
Henri Meddox wrote:
 Hi Folks,
 call me a lernen ;-)

 I got a crazy Problem with zpool list and the size of my pool:

 created zpool create raidz2 hdd1 hdd2 hdd3 - each hdd is about 1GB.
  
 zpool list shows me a size of 2.95GB - shouldn't this bis online 1GB?

 After creating a file about 500MB - Capacity is shown as 50 % - The right 
 value?

 Is this a known bug / feature?
   

[Didn't I answer this just the other day?  I need to improve my search 
skills :-)]
 From the man page for zpool(1m)
 These space usage properties report  actual  physical  space
 available  to  the  storage  pool. The physical space can be
 different from the total amount of space that any  contained
 datasets  can  actually  use.  The amount of space used in a
 raidz configuration depends on the  characteristics  of  the
 data being written. In addition, ZFS reserves some space for
 internal accounting that  the  zfs(1M)  command  takes  into
 account,  but the zpool command does not. For non-full pools
 of a reasonable size, these effects should be invisible. For
 small  pools,  or  pools  that are close to being completely
 full, these discrepancies may become more noticeable.

 From the man page for zfs(1m)
 The amount of space available to the dataset and all its
 children,  assuming  that  there is no other activity in
 the pool. Because space is shared within a pool, availa-
 bility  can be limited by any number of factors, includ-
 ing physical pool size, quotas, reservations,  or  other
 datasets within the pool.

 -- richard


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting assistan

2009-01-27 Thread Brad Hill
Any ideas on this? It looks like a potential bug to me, or there is something 
that I'm not seeing.

Thanks again!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] freeze/ thaw zfs file system

2009-01-27 Thread Chris Ridd

On 27 Jan 2009, at 17:59, Richard Elling wrote:

 ajit jain wrote:
 Hi Andrew,

 I am writing a filtering device which tracks the write to the
 file-system. I am doing it for ufs, vxfs and for zfs. Sometime for
 consistent point I need to freeze the file-system which flushes dirty
 block to the disk and block every IO on the top level. So, for ufs  
 and
 vxfs I got lockfs and VX_FREEZE/VX_THAW, but with zfs I didn't get  
 the
 luck.


 I see no need to freeze the file system, or maybe I would say, I see
 no benefit to freezing a file system.  Perhaps your needs will be met
 by the snapshot feature which will also cause a sync(2).

It sounds like the OP could take advantage of the vscan service 
http://www.opensolaris.org/os/project/vscan/ 
 

Cheers,

Chris
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS and Amanda

2009-01-27 Thread Marcelo Leal
Hello all,
 There is some project here to integrate amanda on opensolaris, or some howto 
for integration with ZFS? Some use case (using the opensource version)?
 The amanda site there is a few instructions, but i think here we can create 
something more specific to OS.
 Thanks.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Import Issues Btw 101 104

2009-01-27 Thread Daniel Templeton
Hi!

I have a system with S10, b101, and b104 installed in the same partition 
on disk 1.  On disks 1 and 2 in different partitions, I also created ZFS 
pools from S10 to be imported by b101 and b104.  Pool 1 is mirrored.  
Pool 2 is not.  About every three builds, I replace the oldest build 
with the latest available and switch to that as the default OS.  Up 
through 101 everything was fine.  I just installed 104, however, and 
when I do the zpool import, the mirrored pool is picked up just fine, 
but the non-mirrored pool shows up as corrupted.  zpool import -D shows 
me that zpool on b104 thinks that the pool is on c1t1d0p2, whereas S10, 
b101, and fdisk agree that it's actually on c1t1d0p3.  How do I convince 
zpool on b104 where my non-mirrored pool really is?  I'm a bit afraid to 
do an import -f because that pool is my home directory, and I'd really 
rather not screw it up.  And I don't see where the -f will change 
zpool's mind about where the pool actually lives.  Maybe import -c from 
the default location?  Where is the default location?  Any thoughts or 
suggestions?

Thanks!
Daniel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting assistan

2009-01-27 Thread Chris Du
Do you know 7200.11 has firmware bugs? 

Go to seagate website to check.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] how to fix zpool with corrupted disk?

2009-01-27 Thread Jakov Sosic
 Jakov Sosic wrote:
 Be happy, the data is already fixed. The DEGRADED
 state is used
 when too many errors were found in a short period of
 time, which
 one would use as an idicator of a failing device.
  However, since the
 evice is not actually failed, it is of no practical
 use in your test case.

Well yes, now I do get it!

The thing is, after scrubbing, I rebooted the machine and dd-ed the 
/dev/urandom on the other device in the mirror pool. After that, I rebooted 
again, and checked the md5sum of a file that occupies entire drive. The md5 is 
same as at the beggining of the test, so that's it. Problem solved, it seems 
that scrub really solves the problem, but after it solves it I have to zpool 
clear so that errors go away :) Pool seems in degraded state altough it's 
fully corrected...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting assistance.

2009-01-27 Thread Blake
I guess you could try 'zpool import -f'.  This is a pretty odd status,
I think.  I'm pretty sure raidz1 should survive a single disk failure.

Perhaps a more knowledgeable list member can explain.

On Sat, Jan 24, 2009 at 12:48 PM, Brad Hill b...@thosehills.com wrote:
 I've seen reports of a recent Seagate firmware update
 bricking drives again.

 What's the output of 'zpool import' from the LiveCD?
  It sounds like
 ore than 1 drive is dropping off.


 r...@opensolaris:~# zpool import
  pool: tank
id: 16342816386332636568
  state: FAULTED
 status: The pool was last accessed by another system.
 action: The pool cannot be imported due to damaged devices or data.
The pool may be active on another system, but can be imported using
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-EY
 config:

tankFAULTED  corrupted data
  raidz1DEGRADED
c6t0d0  ONLINE
c6t1d0  ONLINE
c6t2d0  ONLINE
c6t3d0  UNAVAIL  cannot open
c6t4d0  ONLINE

  pool: rpool
id: 9891756864015178061
  state: ONLINE
 status: The pool was last accessed by another system.
 action: The pool can be imported using its name or numeric identifier and
the '-f' flag.
   see: http://www.sun.com/msg/ZFS-8000-EY
 config:

rpool   ONLINE
  c3d0s0ONLINE
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Unable to destory a pool

2009-01-27 Thread Blake
Can you share the output of 'uname -a' and the disk controller you are using?

On Sun, Jan 25, 2009 at 6:24 PM, Ramesh Mudradi rameshm.ku...@gmail.com wrote:
 # zpool list
 NAME SIZE   USED  AVAILCAP  HEALTH  ALTROOT
 jira-app-zpool   272G   330K   272G 0%  ONLINE  -

 The following command hangs forever. If I reboot the box , zpool list shows 
 online as I mentioned the output above.

 # zpool destroy -f jira-app-zpool

 How can get rid of this pool and any reference to it.

 bash-3.00# zpool status
  pool: jira-app-zpool
  state: UNAVAIL
 status: One or more devices are faultd in response to IO failures.
 action: Make sure the affected devices are connected, then run 'zpool clear'.
   see: http://www.sun.com/msg/ZFS-8000-HC
  scrub: none requested
 config:

NAMESTATE READ WRITE CKSUM
jira-app-zpool  UNAVAIL  0 0 4  insufficient replicas
  c3t0d3FAULTED  0 0 4  experienced I/O failures

 errors: 2 data errors, use '-v' for a list
 bash-3.00#
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Can't add mirror half back to mirror after moving controllers

2009-01-27 Thread Brian Hechinger
In the previous config I had two RAID0 hardware stripes on an LSI1068
that were then mirrored together with ZFS.

I then got a PERC 6/i card (aka LSI1078) to stick in the box and so I
moved the one stripe over to that (and had to re-create the stripe of
course).

The problem is that once the stripe is moved to the PERC it's a handful
of sectors smaller.

The tricky part is that this is my ZFS Root OS pool.  What are my best
options for getting it all moved over to the new vdev so I can rebuild
the primary stripe on the PERC?

Also, what's a good stripe size to configure on the PERC that would work
best with ZFS?

Thanks!

-brian
-- 
Coding in C is like sending a 3 year old to do groceries. You gotta
tell them exactly what you want or you'll end up with a cupboard full of
pop tarts and pancake mix. -- IRC User (http://www.bash.org/?841435)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Unusual CIFS write bursts

2009-01-27 Thread Richard Elling
comment far below...

Brent Jones wrote:
 On Mon, Jan 26, 2009 at 10:40 PM, Brent Jones br...@servuhome.net wrote:
   
 While doing some performance testing on a pair of X4540's running
 snv_105, I noticed some odd behavior while using CIFS.
 I am copying a 6TB database file (yes, a single file) over our GigE
 network to the X4540, then snapshotting that data to the secondary
 X4540.
 Writing said 6TB file can peak our gigabit network, with about
 95-100MB/sec going over the wire (can't ask for any more, really).

 However, the disk IO on the X4540 appears unusual. I would expect the
 disks to be constantly writing 95-100MB/sec, but it appears it buffers
 about 1GB worth of data before committing to disk. This is in contrast
 to NFS write behavior, where as I write a 1GB file to the NFS server
 from an NFS client, traffic on the wire correlates concisely to the
 disk writes. For example, 60MB/sec on the wire via NFS will trigger
 60MB/sec on disk. This is a single file on both cases.

 I wouldn't have a problem with this buffer, it seems to be a rolling
 10-second buffer, if I am copying several small files at lower speeds,
 the disk buffer still seems to purge after roughly 10 seconds, not
 when a certain size is reached. The larger the amount of data that
 goes into the buffer is what causes a problem, writing 1GB to disk can
 cause the system to slow down substantially, all network traffic
 pauses or drops to mere kilobytes a second while it writes this
 buffer.

 I would like to see a smoother handling of this buffer, or a tuneable
 to make the buffer write more often or fill quicker.

 This is a 48TB unit, 64GB ram, and the arcstat perl script reports my
 ARC is 55GB in size, with near 0% miss on reads.

 Has anyone seen something similar, or know of any un-documented
 tuneables to reduce the effects of this?


 Here is 'zpool iostat' output, in 1 second intervals while this write
 storm occurs.


 # zpool iostat pdxfilu01 1
   capacity operationsbandwidth
 pool used  avail   read  write   read  write
 --  -  -  -  -  -  -
 pdxfilu01   2.09T  36.0T  1 61   143K  7.30M
 pdxfilu01   2.09T  36.0T  0  0  0  0
 pdxfilu01   2.09T  36.0T  0  0  0  0
 pdxfilu01   2.09T  36.0T  0  0  0  0
 pdxfilu01   2.09T  36.0T  0 60  0  7.55M
 pdxfilu01   2.09T  36.0T  0  1.70K  0   211M
 pdxfilu01   2.09T  36.0T  0  2.56K  0   323M
 pdxfilu01   2.09T  36.0T  0  2.97K  0   375M
 pdxfilu01   2.09T  36.0T  0  3.15K  0   399M
 pdxfilu01   2.09T  36.0T  0  2.22K  0   244M
 pdxfilu01   2.09T  36.0T  0  0  0  0
 pdxfilu01   2.09T  36.0T  0  0  0  0
 pdxfilu01   2.09T  36.0T  0  0  0  0
 pdxfilu01   2.09T  36.0T  0  0  0  0


 Here is my 'zpool status' output.

 # zpool status
  pool: pdxfilu01
  state: ONLINE
  scrub: none requested
 config:

NAMESTATE READ WRITE CKSUM
pdxfilu01   ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c5t0d0  ONLINE   0 0 0
c6t0d0  ONLINE   0 0 0
c7t0d0  ONLINE   0 0 0
c8t0d0  ONLINE   0 0 0
c9t0d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c4t1d0  ONLINE   0 0 0
c6t1d0  ONLINE   0 0 0
c7t1d0  ONLINE   0 0 0
c8t1d0  ONLINE   0 0 0
c9t1d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c4t2d0  ONLINE   0 0 0
c5t2d0  ONLINE   0 0 0
c7t2d0  ONLINE   0 0 0
c8t2d0  ONLINE   0 0 0
c9t2d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c4t3d0  ONLINE   0 0 0
c5t3d0  ONLINE   0 0 0
c6t3d0  ONLINE   0 0 0
c8t3d0  ONLINE   0 0 0
c9t3d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c4t4d0  ONLINE   0 0 0
c5t4d0  ONLINE   0 0 0
c6t4d0  ONLINE   0 0 0
c7t4d0  ONLINE   0 0 0
c9t4d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c4t5d0  ONLINE   0 0 0
c5t5d0  ONLINE   0 0 0
c6t5d0  ONLINE   0 0 0
c7t5d0  ONLINE   0 0 0
c8t5d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c4t6d0  ONLINE   0 0 0
c5t6d0  ONLINE   0 0 0
c6t6d0  ONLINE   0 0 0
c7t6d0  

Re: [zfs-discuss] Unusual CIFS write bursts

2009-01-27 Thread Brent Jones
On Tue, Jan 27, 2009 at 5:47 PM, Richard Elling
richard.ell...@gmail.com wrote:
 comment far below...

 Brent Jones wrote:

 On Mon, Jan 26, 2009 at 10:40 PM, Brent Jones br...@servuhome.net wrote:





 --
 Brent Jones
 br...@servuhome.net



 I found some insight to the behavior I found at this Sun blog by Roch
 Bourbonnais : http://blogs.sun.com/roch/date/20080514

 Excerpt from the section that I seem to have encountered:

 The new code keeps track of the amount of data accepted in a TXG and
 the time it takes to sync. It dynamically adjusts that amount so that
 each TXG sync takes about 5 seconds (txg_time variable). It also
 clamps the limit to no more than 1/8th of physical memory. 

 So, when I fill up that transaction group buffer, that is when I see
 that 4-5 second I/O burst of several hundred megabytes per second.
 He also documents that the buffer flush can, and does issue delays to
 the writing threads, which is why I'm seeing those momentary drops in
 throughput and sluggish system performance while that write buffer is
 flushed to disk.


 Yes, this tends to be more efficient. You can tune it by setting
 zfs_txg_synctime, which is 5 by default.  It is rare that we've seen
 this be a win, which is why we don't mention it in the Evil Tuning
 Guide.

 Wish there was a better way to handle that, but at the speed I'm
 writing (and I'll be getting a 10GigE link soon), I don't see any
 other graceful methods of handling that much data in a buffer


 I think your workload might change dramatically when you get a
 faster pipe.  So unless you really feel compelled to change it, I
 wouldn't suggest changing it.
 -- richard

 Loving these X4540's so far though...





Are there any additional tuneables, such as opening a new txg buffer
before the previous one is flushed? Or otherwise allow writes to
continue without the tick delay? My workload will be pretty
consistent, it is going to serve a few roles, which I hope to
accomplish in the same units:
- large scale backups
- cifs share for window app servers
- nfs server for unix app servers

GigE quickly became the bottleneck, and I imagine 10GigE will add
further stress to those write buffers.

-- 
Brent Jones
br...@servuhome.net
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] firewire card?

2009-01-27 Thread Frank Cusack
just installed s10_u6 with a root pool.  i'm blown away.  so now i want
to attach my external storage via firewire.  i could use usb2 but i prefer
firewire as i won't need an external hub.

what firewire cards are supported for x86?  the HCL doesn't list any that
i could find.  i searched for any of the terms 'firewire', '1394', 'ohci',
'uhci', 'ehci'.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] firewire card?

2009-01-27 Thread Bob Friesenhahn
On Tue, 27 Jan 2009, Frank Cusack wrote:

 what firewire cards are supported for x86?  the HCL doesn't list any that
 i could find.  i searched for any of the terms 'firewire', '1394', 'ohci',
 'uhci', 'ehci'.

The Sun Ultra-40 (recently discontinued) comes with dual 400Mbit 
firefire ports.  See if you can figure out what hardware it uses.  I 
would be surprised if it is not the same hardware that was used for 
SPARC systems like the Blade 2500 and Ultra-45, with a commercially 
available adaptor card still available as recent as a couple of years 
ago (when I last checked).  The firewire support in the Ultra-40 is 
from the motherboard rather than an adaptor card.

However, be aware that bugs causing problems with ZFS were reported 
against the Firewire support, and it is not clear if anyone got around 
to fixing them even though the fix was apparently simple.

I am using USB2 with two external drives without problem.

Bob
==
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] firewire card?

2009-01-27 Thread James C. McPherson
On Tue, 27 Jan 2009 21:01:55 -0600 (CST)
Bob Friesenhahn bfrie...@simple.dallas.tx.us wrote:

 On Tue, 27 Jan 2009, Frank Cusack wrote:
 
  what firewire cards are supported for x86?  the HCL doesn't list
  any that i could find.  i searched for any of the terms 'firewire',
  '1394', 'ohci', 'uhci', 'ehci'.
 
 The Sun Ultra-40 (recently discontinued) comes with dual 400Mbit 
 firefire ports.  See if you can figure out what hardware it uses.  I 
 would be surprised if it is not the same hardware that was used for 
 SPARC systems like the Blade 2500 and Ultra-45, with a commercially 
 available adaptor card still available as recent as a couple of years 
 ago (when I last checked).  The firewire support in the Ultra-40 is 
 from the motherboard rather than an adaptor card.
 
 However, be aware that bugs causing problems with ZFS were reported 
 against the Firewire support, and it is not clear if anyone got
 around to fixing them even though the fix was apparently simple.
 
 I am using USB2 with two external drives without problem.

Hi Frank,
This is what lspci reports on my u40m2 system:


01:08.0 FireWire (IEEE 1394): Texas Instruments TSB43AB22/A IEEE-1394a-2000 
Controller (PHY/Link) (prog-if 10 [OHCI])
Subsystem: Sun Microsystems Computer Corp.: Unknown device 6676
Control: I/O- Mem+ BusMaster+ SpecCycle- MemWINV- VGASnoop- ParErr- 
Stepping- SERR- FastB2B-
Status: Cap+ 66Mhz- UDF- FastB2B- ParErr- DEVSEL=medium TAbort- 
TAbort- MAbort- SERR- PERR-
Latency: 64 (500ns min, 1000ns max), cache line size 10
Interrupt: pin A routed to IRQ 11
Region 0: Memory at a0104800 (32-bit, non-prefetchable)
Region 1: Memory at a010 (32-bit, non-prefetchable)
Capabilities: [44] Power Management version 2
Flags: PMEClk- DSI- D1+ D2+ AuxCurrent=0mA 
PME(D0+,D1+,D2+,D3hot+,D3cold-)
Status: D0 PME-Enable- DSel=0 DScale=0 PME+


prtconf reports that the full pci vid/did information is
pci104c,8023.108e.6676.0

As Bob mentioned, this is on the motherboard rather than
a plugin card.


James C. McPherson
--
Senior Kernel Software Engineer, Solaris
Sun Microsystems
http://blogs.sun.com/jmcp   http://www.jmcp.homeunix.com/blog
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing HDD in x4500

2009-01-27 Thread Blake
I'm not an authority, but on my 'vanilla' filer, using the same
controller chipset as the thumper, I've been in really good shape
since moving to zfs boot in 10/08 and doing 'zpool upgrade' and 'zfs
upgrade' to all my mirrors (3 3-way).  I'd been having similar
troubles to yours in the past.

My system is pretty puny next to yours, but it's been reliable now for
slightly over a month.


On Tue, Jan 27, 2009 at 12:19 AM, Jorgen Lundman lund...@gmo.jp wrote:

 The vendor wanted to come in and replace an HDD in the 2nd X4500, as it
 was constantly busy, and since our x4500 has always died miserably in
 the past when a HDD dies, they wanted to replace it before the HDD
 actually died.

 The usual was done, HDD replaced, resilvering started and ran for about
 50 minutes. Then the system hung, same as always, all ZFS related
 commands would just hang and do nothing. System is otherwise fine and
 completely idle.

 The vendor for some reason decided to fsck root-fs, not sure why as it
 is mounted with logging, and also decided it would be best to do so
 from a CDRom boot.

 Anyway, that was 12 hours ago and the x4500 is still down. I think they
 have it at single-user prompt resilvering again. (I also noticed they'd
 decided to break the mirror of the root disks for some very strange
 reason). It still shows:

   raidz1  DEGRADED 0 0 0
 c0t1d0ONLINE   0 0 0
 replacing UNAVAIL  0 0 0  insufficient replicas
   c1t1d0s0/o  OFFLINE  0 0 0
   c1t1d0  UNAVAIL  0 0 0  cannot open

 So I am pretty sure it'll hang again sometime soon. What is interesting
 though is that this is on x4500-02, and all our previous troubles mailed
 to the list was regarding our first x4500. The hardware is all
 different, but identical. Solaris 10 5/08.

 Anyway, I think they want to boot CDrom to fsck root again for some
 reason, but since customers have been without their mail for 12 hours,
 they can go a little longer, I guess.

 What I was really wondering, has there been any progress or patches
 regarding the system always hanging whenever a HDD dies (or is replaced
 it seems). It really is rather frustrating.

 Lund

 --
 Jorgen Lundman   | lund...@lundman.net
 Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
 Shibuya-ku, Tokyo| +81 (0)90-5578-8500  (cell)
 Japan| +81 (0)3 -3375-1767  (home)
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing HDD in x4500

2009-01-27 Thread Jorgen Lundman

Thanks for your reply,

While the savecore is working its way up the chain to (hopefully) Sun, 
the vendor asked us not to use it, so we moved x4500-02 to use x4500-04 
and x4500-05. But perhaps moving to Sol 10 10/08 on x4500-02 when fixed 
is the way to go.

The savecore had the usual info, that everything is blocked waiting on 
locks:


   601*  threads trying to get a mutex (598 user, 3 kernel)
   longest sleeping 10 minutes 13.52 seconds earlier
   115*  threads trying to get an rwlock (115 user, 0 kernel)

1678   total threads in allthreads list (1231 user, 447 kernel)
10   thread_reapcnt
 0   lwp_reapcnt
1688   nthread

   thread pri pctcpu   idle   PID  wchan 
command
   0xfe8000137c80  60  0.000  -9m44.88s 0 0xfe84d816cdc8 
sched
   0xfe800092cc80  60  0.000  -9m44.52s 0 0xc03c6538 
sched
   0xfe8527458b40  59  0.005  -1m41.38s  1217 0xb02339e0 
/usr/lib/nfs/rquotad
   0xfe8527b534e0  60  0.000   -5m4.79s   402 0xfe84d816cdc8 
/usr/lib/nfs/lockd
   0xfe852578f460  60  0.000  -4m59.79s   402 0xc0633fc8 
/usr/lib/nfs/lockd
   0xfe8532ad47a0  60  0.000  -10m4.40s   623 0xfe84bde48598 
/usr/lib/nfs/nfsd
   0xfe8532ad3d80  60  0.000  -10m9.10s   623 0xfe84d816ced8 
/usr/lib/nfs/nfsd
   0xfe8532ad3360  60  0.000  -10m3.77s   623 0xfe84d816cde0 
/usr/lib/nfs/nfsd
   0xfe85341e9100  60  0.000  -10m6.85s   623 0xfe84bde48428 
/usr/lib/nfs/nfsd
   0xfe85341e8a40  60  0.000  -10m4.76s   623 0xfe84d816ced8 
/usr/lib/nfs/nfsd

SolarisCAT(vmcore.0/10X) tlist sobj locks | grep nfsd | wc -l
  680

scl_writer = 0xfe8000185c80  - locking thread



thread 0xfe8000185c80
 kernel thread: 0xfe8000185c80  PID: 0 
cmd: sched
t_wchan: 0xfbc8200a  sobj: condition var (from genunix:bflush+0x4d)
t_procp: 0xfbc22dc0(proc_sched)
   p_as: 0xfbc24a20(kas)
   zone: global
t_stk: 0xfe8000185c80  sp: 0xfe8000185aa0  t_stkbase: 
0xfe8000181000
t_pri: 99(SYS)  pctcpu: 0.00
t_lwp: 0x0  psrset: 0  last CPU: 0
idle: 44943 ticks (7 minutes 29.43 seconds)
start: Tue Jan 27 23:44:21 2009
age: 674 seconds (11 minutes 14 seconds)
tstate: TS_SLEEP - awaiting an event
tflg:   T_TALLOCSTK - thread structure allocated from stk
tpflg:  none set
tsched: TS_LOAD - thread is in memory
 TS_DONT_SWAP - thread/LWP should not be swapped
pflag:  SSYS - system resident process

pc:  0xfb83616f unix:_resume_from_idle+0xf8 resume_return
startpc: 0xeff889e0 zfs:spa_async_thread+0x0

unix:_resume_from_idle+0xf8 resume_return()
unix:swtch+0x12a()
genunix:cv_wait+0x68()
genunix:bflush+0x4d()
genunix:ldi_close+0xbe()
zfs:vdev_disk_close+0x6a()
zfs:vdev_close+0x13()
zfs:vdev_raidz_close+0x26()
zfs:vdev_close+0x13()
zfs:vdev_reopen+0x1d()
zfs:spa_async_reopen+0x5f()
zfs:spa_async_thread+0xc8()
unix:thread_start+0x8()
-- end of kernel thread's stack --




Blake wrote:
 I'm not an authority, but on my 'vanilla' filer, using the same
 controller chipset as the thumper, I've been in really good shape
 since moving to zfs boot in 10/08 and doing 'zpool upgrade' and 'zfs
 upgrade' to all my mirrors (3 3-way).  I'd been having similar
 troubles to yours in the past.
 
 My system is pretty puny next to yours, but it's been reliable now for
 slightly over a month.
 
 
 On Tue, Jan 27, 2009 at 12:19 AM, Jorgen Lundman lund...@gmo.jp wrote:
 The vendor wanted to come in and replace an HDD in the 2nd X4500, as it
 was constantly busy, and since our x4500 has always died miserably in
 the past when a HDD dies, they wanted to replace it before the HDD
 actually died.

 The usual was done, HDD replaced, resilvering started and ran for about
 50 minutes. Then the system hung, same as always, all ZFS related
 commands would just hang and do nothing. System is otherwise fine and
 completely idle.

 The vendor for some reason decided to fsck root-fs, not sure why as it
 is mounted with logging, and also decided it would be best to do so
 from a CDRom boot.

 Anyway, that was 12 hours ago and the x4500 is still down. I think they
 have it at single-user prompt resilvering again. (I also noticed they'd
 decided to break the mirror of the root disks for some very strange
 reason). It still shows:

   raidz1  DEGRADED 0 0 0
 c0t1d0ONLINE   0 0 0
 replacing UNAVAIL  0 0 0  insufficient replicas
   c1t1d0s0/o  OFFLINE  0 0 0
   c1t1d0  UNAVAIL  0 0 0  cannot open

 So I am pretty sure it'll hang again sometime soon. What is interesting
 though is that this is on x4500-02, and all our previous troubles mailed
 to the list was regarding our first x4500. The hardware is all
 different, but identical. Solaris 10 5/08.

 Anyway, I think they want to boot CDrom to 

Re: [zfs-discuss] zpool status -x strangeness

2009-01-27 Thread Blake
What does 'zpool status -xv' show?

On Tue, Jan 27, 2009 at 8:01 AM, Ben Miller mil...@eecis.udel.edu wrote:
 I forgot the pool that's having problems was recreated recently so it's 
 already at zfs version 3.  I just did a 'zfs upgrade -a' for another pool, 
 but some of those filesystems failed since they are busy and couldn't be 
 unmounted.

 # zfs upgrade -a
 cannot unmount '/var/mysql': Device busy
 cannot unmount '/var/postfix': Device busy
 
 6 filesystems upgraded
 821 filesystems already at this version

 Ben

 You can upgrade live.  'zfs upgrade' with no
 arguments shows you the
 zfs version status of filesystems present without
 upgrading.



 On Jan 24, 2009, at 10:19 AM, Ben Miller
 mil...@eecis.udel.edu wrote:

  We haven't done 'zfs upgrade ...' any.  I'll give
 that a try the
  next time the system can be taken down.
 
  Ben
 
  A little gotcha that I found in my 10u6 update
  process was that 'zpool
  upgrade [poolname]' is not the same as 'zfs
 upgrade
  [poolname]/[filesystem(s)]'
 
  What does 'zfs upgrade' say?  I'm not saying this
 is
  the source of
  your problem, but it's a detail that seemed to
 affect
  stability for
  me.
 
 
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting

2009-01-27 Thread Brad Hill
r...@opensolaris:~# zpool import -f tank
internal error: Bad exchange descriptor
Abort (core dumped)

Hoping someone has seen that before... the Google is seriously letting me down 
on that one.

 I guess you could try 'zpool import -f'.  This is a
 pretty odd status,
 I think.  I'm pretty sure raidz1 should survive a
 single disk failure.
 
 Perhaps a more knowledgeable list member can explain.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing HDD in x4500

2009-01-27 Thread Tim
On Tue, Jan 27, 2009 at 9:28 PM, Jorgen Lundman lund...@gmo.jp wrote:


 Thanks for your reply,

 While the savecore is working its way up the chain to (hopefully) Sun,
 the vendor asked us not to use it, so we moved x4500-02 to use x4500-04
 and x4500-05. But perhaps moving to Sol 10 10/08 on x4500-02 when fixed
 is the way to go.

 The savecore had the usual info, that everything is blocked waiting on
 locks:


I assume you've changed the failmode to continue already?

http://prefetch.net/blog/index.php/2008/03/01/configuring-zfs-to-gracefully-deal-with-failures/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] firewire card?

2009-01-27 Thread Robert Thurlow
Frank Cusack wrote:
 just installed s10_u6 with a root pool.  i'm blown away.  so now i want
 to attach my external storage via firewire.

I was able to use this cheap thing with good initial results:

http://www.newegg.com/Product/Product.aspx?Item=N82E16815124002

However, I ran into a frequent hang[1] and declined to put it in
the HWCL.  The hang appears to be a bug in the scsa1394 kernel
module, and are not yet fixed.  I've moved the disks our of
their Firewire case and inside a machine with SATA-to-IDE
converters (ick).

[1] http://bugs.opensolaris.org/view_bug.do?bug_id=6539587

Rob T
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting

2009-01-27 Thread Blake
This is outside the scope of my knowledge/experience.  Maybe there is
now a core file you can examine?  That might help you at least see
what's going on?

On Tue, Jan 27, 2009 at 10:32 PM, Brad Hill b...@thosehills.com wrote:
 r...@opensolaris:~# zpool import -f tank
 internal error: Bad exchange descriptor
 Abort (core dumped)

 Hoping someone has seen that before... the Google is seriously letting me 
 down on that one.

 I guess you could try 'zpool import -f'.  This is a
 pretty odd status,
 I think.  I'm pretty sure raidz1 should survive a
 single disk failure.

 Perhaps a more knowledgeable list member can explain.
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] firewire card?

2009-01-27 Thread Frank Cusack
thanks for all the feedback.  i guess i'll stick with usb2.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing HDD in x4500

2009-01-27 Thread Jorgen Lundman
 
 I assume you've changed the failmode to continue already?
 
 http://prefetch.net/blog/index.php/2008/03/01/configuring-zfs-to-gracefully-deal-with-failures/
  

This appears to be new to 10/08, so that is another vote to upgrade. 
Also interesting that the default is wait, since it almost behaves 
like it. Not sure why it would block zpool, zfs and df commands as 
well though?


Lund


-- 
Jorgen Lundman   | lund...@lundman.net
Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
Shibuya-ku, Tokyo| +81 (0)90-5578-8500  (cell)
Japan| +81 (0)3 -3375-1767  (home)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting assistan

2009-01-27 Thread Brad Hill
I do, thank you. The disk that went out sounds like it had a head crash or some 
such - loud clicking shortly after spin-up then it spins down and gives me 
nothing. BIOS doesn't even detect it properly to do a firmware update.


 Do you know 7200.11 has firmware bugs? 
 
 Go to seagate website to check.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Raidz1 faulted with single bad disk. Requesting

2009-01-27 Thread Ross
Just a thought, but have you physically disconnected the bad disk?  It's not 
unheard of for a bad disk to cause problems with others.

Failing that, it's the corrupted data bit that's worrying me, it sounds like 
you may have other corruption on the pool (always a risk with single parity 
raid), but I'm worried that it's not giving you any more details as to what's 
wrong.

Also, what version of OpenSolaris are you running?  Could you maybe try booting 
off a CD of the latest build?  There are often improvements in the way ZFS 
copes with errors, so it's worth a try.  I don't think it's likely to help, but 
I wouldn't discount it.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] mount race condition?

2009-01-27 Thread Frank Cusack
i was wondering if you have a zfs filesystem that mounts in a subdir
in another zfs filesystem, is there any problem with zfs finding
them in the wrong order and then failing to mount correctly?

say you have pool1/data which mounts on /data and pool2/foo which
mounts on /data/subdir/foo, what if at boot time, pool2 is imported
first, what happens?  /data would exist but /data/subdir wouldn't
exist since pool1/data hasn't been mounted yet.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss