Re: [zfs-discuss] Strange problem, possibly zfs/zpool

2009-09-14 Thread Marc Emmerson
OK problem solved.

I had incorrectly assumed that the server wasn't booting, the longest I had 
left it was over night and there was still no logon prompt in the morning!  The 
reality is that it was just taking a very long time due to an excessive amount 
of automatically created snapshots by Time Slider.

It did occur to me earlier that the amount of snapshots might cause delays 
mounting the pool as I had previously left Time Slider on for all filesystems 
and volumes.  Back then I used Time Slider to erase the snapshots, again, 
incorrectly assuming it would also erase the volume snapshots, I can now 
confirm that it does not!

So, to resolve, I reinstalled the OS (installing another instance to another 
disk would work), imported the array, the import command then seems to get 
stuck (obviously, in hindsight), although the core filesystems are present and 
accessible.  Now, before rebooting, I then run this command to erase all 
automatically created snapshots:

zfs list -r -t snapshot -o name -H array/iscsi | grep zfs-auto-snap | xargs -t 
-l zfs destroy

Obviously, this takes while ;-)

After this completes, I reboot and all is well again!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Intel X25-E SSD in x4500 followup

2009-09-14 Thread Peter Eriksson
I can confirm that on an X4240 with the LSI (mpt) controller:

X25-M G1 with 8820 still returns invalid selftest data
X25-E G1 with 8850 now returns correct selftest data
(I haven't got any X25-M G2)

Going to replace an X25-E with the old firmware in one of our X4500s
soon and we'll see if things work right there)

I still see heavy write load-induced bus resets with the 8850-firmware X25-Es 
on the X4240 though.
(Unless I wrap the X25-E inside a DiskSuite SVM metadevice for some strange 
reason).
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Intel X25-E SSD in x4500 followup

2009-09-14 Thread Peter Eriksson
Now tested a firmware 8850 X25-E in one of our X4500:s and things look better:

 # /ifm/bin/smartctl -d scsi -l selftest /dev/rdsk/c5t7d0s0
 smartctl version 5.38 [i386-pc-solaris2.10] Copyright (C) 2002-8 Bruce Allen
 Home page is http://smartmontools.sourceforge.net/
 
 No self-tests have been logged

No scsi console errors so far.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] raidz replace issue

2009-09-14 Thread Mark J Musante

On Sat, 12 Sep 2009, Jeremy Kister wrote:


scrub: resilver in progress, 0.12% done, 108h42m to go
 [...]
  raidz1  DEGRADED 0 0 0
   c3t8d0ONLINE   0 0 0
   c5t8d0ONLINE   0 0 0
   c3t9d0ONLINE   0 0 0
   replacing DEGRADED 0 0 0
 c5t9d0s0/o  UNAVAIL  0 0 0  cannot open
 c5t9d0  ONLINE   0 0 0

woohoo!  i've never had to use either s0 or s0/o, but hey, i'm happy.


Glad to see it's working.  I opened CR 6881631 to track this issue.


Regards,
markm
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Reboot seems to mess up all rpools

2009-09-14 Thread Frank Middleton

[Originally posted to indiana-discuss]

On certain X86 machines there's a hardware/software glitch
that causes odd transient checksum failures that always seem
to affect the same files even if you replace them. This has
been submitted as a bug:

Bug 11201 -  Checksum failures on mirrored drives - now
CR 6880994 P4 kernel/zfs Checksum failures on mirrored drives

We have SPARC based ZFS servers where we keep a copy of this
rpool so we can more easily replace the damaged files (usually
system libraries). In addition, to check the validity of the
zfs send stream of the ZFS server rpool, there's a copy of that
as well. For good reasons there might be several rpools in
this data pool at any given time.

When the ZFS server is rebooted, it tries to update the boot
archive of every rpool it can find, including the X86 archive,
which fails because it's the wrong architecture.

The ZFS server is currently at snv103, but the backup server has
an additional disk with snv111b on it, which was recently updated
to snv122. However, if you boot snv103 and then reboot, it will
also update the snv122 boot archive, rendering snv122 unbootable.
All versions up to and including snv122 exhibit this behavior.

I'm not sure why updating the boot archive would do this, but surely
this is a bug. Reboot should only update it's own archive, and not
any ZFS archives at all if it is running from UFS. Before submitting
a bug report, I thought I'd check here to see if a) if this is has
already been reported, and b) if I have the terminology right.

Thanks -- Frank

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] USB WD Passport 500GB zfs mirror bug

2009-09-14 Thread Louis-Frédéric Feuillette
On Sun, 2009-09-13 at 11:01 -0700, Stefan Parvu wrote:
 5. Disconnecting the other disk. Problems occur:
 # zpool status zones
   pool: zones
  state: ONLINE
 status: One or more devices has experienced an unrecoverable error.
 An
 attempt was made to correct the error.  Applications are
 unaffected.
 action: Determine if the device needs to be replaced, and clear the
 errors
 using 'zpool clear' or replace the device with 'zpool
 replace'.
see: http://www.sun.com/msg/ZFS-8000-9P
  scrub: resilver completed after 0h0m with 0 errors on Sun Sep 13
 20:58:02 2009
 config:
 
 NAME  STATE READ WRITE CKSUM
 zones ONLINE   0 0 0
   mirror  ONLINE   0 0 0
 c7t0d0p0  ONLINE   0   167 0  294K resilvered
 c7t0d0p0  ONLINE   0 0 0  208K resilvered
 
 errors: No known data errors
 
 
 # zpool status zones
   pool: zones
  state: DEGRADED
 status: One or more devices could not be used because the label is
 missing or
 invalid.  Sufficient replicas exist for the pool to continue
 functioning in a degraded state.
 action: Replace the device using 'zpool replace'.
see: http://www.sun.com/msg/ZFS-8000-4J
  scrub: resilver completed after 0h0m with 0 errors on Sun Sep 13
 20:58:02 2009
 config:
 
 NAME  STATE READ WRITE CKSUM
 zones DEGRADED 0 0 0
   mirror  DEGRADED 0 0 0
 c7t0d0p0  ONLINE   0   167 0  294K resilvered
 c7t0d0p0  FAULTED  0   113 0  corrupted data
 
 errors: No known data errors
 
 
 I have disconnected c8t0d0p0 but zfs reports that c7t0d0p0 has been
 faulty !?

Both disks read c7t0d0p0, not c7t0d0p0 and c8t0d0p0 as you have in #1-4.
Typo?

-- 
Louis-Frédéric Feuillette jeb...@gmail.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool status OK but zfs filesystem seems hung

2009-09-14 Thread Dave
Thanks for the reply but this seems to be a bit different.  

a couple of things I failed to mention;
1) this is a secondary pool and not the root pool. 
2) the snapshot are trimmed to only keep 80 or so.

The system boots and runs fine.   It's just an issue for this secondary pool 
and filesystem.It seems to be directly related to I/O intensive operations 
as the (full) backup seems to trigger it, never seen it happen with incremental 
backups...


Thanks.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool status OK but zfs filesystem seems hung

2009-09-14 Thread Dave
Hello all,
  I have a situation where zpool status shows no known data errors but all 
processes on a specific filesystem are hung.  This has happened 2 times before 
since we installed Opensolaris 2009.06 snv_111b. For instance there are two 
files systems in this pool 'zfs get all' on one filesystem returns with out 
issue when ran on the other filesystem it hangs.  Also a 'df -h' hangs, etc.

This file system has many different operation running on it;
1) It receives incremental snapshot every 30 minutes continuously.
2) every night a clone is made from one of the received snapshot streams then a 
filesystem backup is taken on that clone (the backup is a directory traversal) 
once the backup completes the clone is destroyed. 

We tried to upgrade to the latest build but ran in to the current 'check sum' 
issue in build snv_122 so we rolled back.

# uname -a
SunOS lahar2 5.11 snv_111b i86pc i386 i86pc

# zpool status zdisk1
  pool: zdisk1
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
zdisk1  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c7t1d0  ONLINE   0 0 0
c7t2d0  ONLINE   0 0 0
c7t3d0  ONLINE   0 0 0
c7t4d0  ONLINE   0 0 0
c7t5d0  ONLINE   0 0 0
spares
  c7t6d0AVAIL

errors: No known data errors


The filesystem is currently in this 'hung' state,  is there any commands I can 
run to help debug the issue?

TIA
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS flar image.

2009-09-14 Thread RB
Is it possible to create flar image of ZFS root filesystem to install it to 
other macines?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS LDOMs Jumpstart virtual disk issue

2009-09-14 Thread Enda O'Connor

RB wrote:
I have zfs on my base T5210 box installed with LDOMS (v.1.0.3).  Every time I try to jumpstart my Guest machine, I get the following error. 



ERROR: One or more disks are found, but one of the following problems exists:
- Hardware failure
- The disk(s) available on this system cannot be used to install 
Solaris Software. They do not have a valid label. If you want to use the 
disk(s) for the install, use format(1M) to label the disk and restart the 
installation.
Solaris installation program exited.


If I try to label the disk using format, I get the following error
format label
Ready to label disk, continue? y

Warning: error writing EFI.
Label failed.

Any help would be appreciated.

run format -e then label and select SMI,
this will erase any data on said disks by the way.

Enda
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS flar image.

2009-09-14 Thread Enda O'Connor

RB wrote:

Is it possible to create flar image of ZFS root filesystem to install it to 
other macines?


yes but it needs solaris update 7 or later to install a zfs flar
see

http://www.opensolaris.org/os/community/zfs/boot/flash/;jsessionid=AB24EEFB6955AD505F19A152CDEC84A8

isn't supported on opensolaris by the way.

Enda
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS flar image.

2009-09-14 Thread Cindy . Swearingen

Hi RB,

We have a draft of the ZFS/flar image support here:

http://opensolaris.org/os/community/zfs/boot/flash/

Make sure you review the Solaris OS requirements.

Thanks,

Cindy

On 09/14/09 11:45, RB wrote:

Is it possible to create flar image of ZFS root filesystem to install it to 
other macines?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Reboot seems to mess up all rpools

2009-09-14 Thread Frank Middleton

Absent any replies to the list, submitted as a bug:

http://defect.opensolaris.org/bz/show_bug.cgi?id=11358

Cheers -- Frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS and quota/refqoutoa question

2009-09-14 Thread Peter Wilk

All,


IHAC that is asking the following...reviewing the following document

http://docs.sun.com/app/docs/doc/817-2271/6mhupg6nc

it appears it may not for the parent setting will transcend downwards to 
the child


Can anyone elaborate if this is correct or not..

Thanks

Peter


the question is regarding to filesystem hierarchy and quota and refquota 
usage.


tank/appswill be mounted as /apps  -- need to be set with 10G
tank/apps/data1  will need to be mount as /apps/data1, need to be set 
with 20G alone.


The question is:
If refquota is being used to set the filesystem sizes on /apps and 
/apps/data1. /apps/data1 will not be inheriting the quota from /apps. In 
other words, /apps/data1 will have full 20G usable disk space and /apps 
will have 10G usable disk space. /apps/data1 will not be inherting disk 
space from /apps.


--
Peter Wilk
Technical Support Engineer
Sun Microsystems, Inc.
Operating Systems Technology Service Center

Email: peter.w...@sun.com
Phone: 1-800-USA-4SUN
My Working Hours : 8am-4pm EST, Monday 
		   9am-5pm EST, Tuesday-Friday

My Manager's Email: dawn.b...@sun.com
===
TO REACH THE NEXT AVAILABLE ENGINEER:
1. Call 1-800-USA-4SUN choose opt 2 and enter your case number.
2. Wait for my voice mail message to begin.
3. Press 0 during my message to reach the next available engineer.
4. You will hear hold music until the next engineer answers.

Submit, check and update tickets at http://www.sun.com/osc

This email may contain confidential and privileged material for the sole
use of the intended recipient. Any review or distribution by others is
strictly prohibited. If you are not the intended recipient please
contact the sender and delete all copies.
 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss