Re: [zfs-discuss] [storage-discuss] Disk Issues

2010-02-20 Thread Brian McKerr
Thanks everyone who has tried to help. this has gotten a bit crazier, I 
removed the 'faulty' drive and let the pool run in degraded mode. It would 
appear that now another drive has decided to play up;



de-bash-4.0# zpool status
  pool: data
 state: DEGRADED
status: One or more devices has been taken offline by the administrator.
Sufficient replicas exist for the pool to continue functioning in a
degraded state.
action: Online the device using 'zpool online' or replace the device with
'zpool replace'.
 scrub: resilver completed after 2h35m with 0 errors on Wed Feb 17 13:48:16 2010
config:

NAMESTATE READ WRITE CKSUM
dataDEGRADED 0 0 0
  raidz2-0  DEGRADED 0 0 0
c6t0d0  ONLINE   0 0 0
c6t1d0  ONLINE   0 0 0
c6t2d0  ONLINE   0 0 0
c6t3d0  ONLINE   0 0 0
c6t4d0  OFFLINE  0 0 0  366G resilvered
c6t5d0  ONLINE   0 0 0

errors: No known data errors

Now I'm transferring some data to the pool
vicer/sw/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot 
sd0   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd2  18.50.02.10.0  0.2  0.0   13.5   3   4   0   0   0   0 
sd3  18.50.02.10.0  0.2  0.0   14.8   4   5   0   0   0   0 
sd4   1.00.00.00.0  9.0  1.0 .9 100 100   0   0   0   0 
sd5  19.50.02.10.0  0.2  0.0   11.9   3   4   0   0   0   0 
sd6   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd7  18.50.02.10.0  0.3  0.1   22.7   8   8   0   0   0   0 
sd8   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
sd9   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
 extended device statistics  errors --- 
devicer/sw/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot 
sd0   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd2   0.50.00.00.0  0.0  0.0   15.6   0   1   0   0   0   0 
sd3   0.50.00.00.0  0.0  0.0   33.5   0   2   0   0   0   0 
sd4   0.50.00.00.0  9.0  1.0 1.9 100 100   0   0   0   0 
sd5   0.50.00.00.0  0.0  0.0   21.4   0   1   0   0   0   0 
sd6   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd7   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd8   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
sd9   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
 extended device statistics  errors --- 
devicer/sw/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot 
sd0   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd2   0.50.50.00.0  0.0  0.05.9   0   1   0   0   0   0 
sd3   0.50.50.00.0  0.0  0.0   10.3   0   1   0   0   0   0 
sd4   0.50.00.00.0  9.0  1.0 1.8 100 100   0   0   0   0 
sd5   0.50.50.00.0  0.0  0.0   11.1   0   1   0   0   0   0 
sd6   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd7   0.50.50.00.0  0.0  0.08.2   0   1   0   0   0   0 
sd8   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
sd9   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
 extended device statistics  errors --- 
devicer/sw/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot 
sd0   0.50.00.00.0  0.0  0.01.7   0   0   0   0   0   0 
sd2   6.5   16.00.00.7  0.3  0.1   15.5   5   6   0   0   0   0 
sd3   6.07.50.00.7  0.4  0.1   33.8   8   8   0   0   0   0 
sd4   0.50.00.00.0  9.0  1.0 1.9 100 100   0   0   0   0 
sd5   5.5   17.50.00.7  0.2  0.09.9   4   5   0   0   0   0 
sd6   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd7   6.5   17.50.00.7  0.4  0.1   18.0   6   6   0   0   0   0 
sd8   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
sd9   0.00.00.00.0  0.0  0.00.0   0   0   0  24  24  48 
 extended device statistics  errors --- 
devicer/sw/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot 
sd0   0.00.00.00.0  0.0  0.00.0   0   0   0   0   0   0 
sd2   2.00.00.10.0  0.0  0.0   16.4   1   2   0   0   0   0 
sd3   2.00.00.10.0  0.0  0.0   29.4   1   3   0   0   0   0 
sd4   1.00.00.00.0  9.0  1.0 .9 100 100   0   0   0   0 
sd5   2.00.00.10.0  0.0  0.0   28.4   1   4   0   0   0   0 
sd6   

Re: [zfs-discuss] Disk Issues

2010-02-08 Thread Brian McKerr
Ok, I changed the cable and also tried swapping the port on the motherboard. 
The drive continued to have huge asvc_t and also started to have huge wsvc_t. I 
unplugged it and the 'pool' is now operating as per expected performance wise.

See the 'storage' forum for any further updates as I am now convinced this has 
nothing to do with ZFS or my attempt to disable the ZIL. 8-)
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Disk Issues

2010-02-07 Thread Brian McKerr
While not strictly a ZFS issue as such I thought I'd post here as this and the 
storage forums are my best bet in terms of getting some help.

I have a machine that I recently set up with b130, b131 and b132. With each 
build I have been playing around with ZFS raidz2 and mirroring to do a little 
performance testing. This is a 6 SATA port ICH10 Intel motherboard, running in 
AHCI mode. OS is on a USB flash. Suffice to say I have noticed that 1 
particular drive out of 6 seems to have very high asvc_t practically all the 
time. This is an excerpt from 'iostat -xnM c6t2d0 2';

 [i]   r/sw/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
   70.5  502.00.04.1  0.0  1.30.02.2   0  54 c6t2d0
   50.5  137.50.03.0  0.0  0.70.03.9   0  47 c6t2d0
   71.0  163.50.04.8  0.0  0.80.03.4   0  61 c6t2d0
   13.5   29.50.01.0  0.0  2.60.0   61.4   0  88 c6t2d0
1.00.50.00.0  0.0  3.60.0 2406.2   0 100 c6t2d0
1.01.00.00.0  0.0  4.00.0 1993.4   0 100 c6t2d0
1.01.50.00.0  0.0  4.00.0 1593.8   0 100 c6t2d0
2.03.00.00.1  0.0  4.00.0  791.6   0 100 c6t2d0
1.02.00.00.1  0.0  4.00.0 1320.3   0 100 c6t2d0
1.05.00.00.3  0.0  3.60.0  595.1   0 100 c6t2d0
[/i]

and here is the drive shown with the other in the raidz2 pool

extended device statistics
r/sw/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
0.01.50.00.0  0.0  0.00.00.5   0   0 c6t0d0
0.01.50.00.0  0.0  0.00.00.3   0   0 c6t1d0
1.01.00.00.0  0.0  4.00.0 1994.8   0 100 c6t2d0
1.01.50.00.0  0.0  0.00.05.2   0   1 c6t3d0
1.01.50.00.0  0.0  0.00.06.9   0   1 c6t4d0
1.01.50.00.0  0.0  0.00.0   10.1   0   2 c6t5d0
extended device statistics
r/sw/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
1.05.50.00.2  0.0  0.00.01.6   0   1 c6t0d0
1.05.50.00.2  0.0  0.00.01.5   0   1 c6t1d0
2.03.50.00.1  0.0  4.00.0  721.8   0 100 c6t2d0
1.05.50.00.2  0.0  0.00.01.9   0   1 c6t3d0
1.05.50.00.2  0.0  0.00.01.6   0   1 c6t4d0
2.05.50.00.2  0.0  0.00.03.1   0   2 c6t5d0
extended device statistics
r/sw/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
0.03.50.00.1  0.0  0.00.00.4   0   0 c6t0d0
0.03.50.00.1  0.0  0.00.01.8   0   0 c6t1d0
1.02.00.00.1  0.0  4.00.0 1327.1   0 100 c6t2d0
1.03.50.00.1  0.0  0.00.04.9   0   1 c6t3d0
1.03.50.00.1  0.0  0.00.03.9   0   1 c6t4d0
1.03.50.00.1  0.0  0.00.02.0   0   1 c6t5d0

I have seen asvc_t as high as 2.

There do not appear to be any errors hardware wise as 'iostat -e' shows

 [i]   errors ---
device  s/w h/w trn tot
sd0   0   0   0   0
sd2   0   0   0   0
sd3   0   0   0   0
sd4   0   0   0   0
sd5   0   0   0   0
sd6   0   0   0   0
sd7   0   0   0   0
[/i]



'zpool iostat -v 2' pauses for anywhere between 3 and 10 seconds before it 
prints the stats for that particular drive in the pool;

   capacity operationsbandwidth
poolalloc   free   read  write   read  write
--  -  -  -  -  -  -
data 185G  5.26T  3115  8.96K  2.48M
  raidz2 185G  5.26T  3115  8.96K  2.48M
c6t0d0  -  -  2 26  2.70K   643K
c6t1d0  -  -  2 26  2.49K   643K

* INSERT ~10 SECOND PAUSE*

c6t2d0  -  -  2 24  2.81K   643K
c6t3d0  -  -  2 26  2.75K   643K
c6t4d0  -  -  2 26  2.45K   643K
c6t5d0  -  -  2 26  2.71K   643K
--  -  -  -  -  -  -
rpool   3.50G  3.94G  0  0  9.99K   1010
  c5t0d0s0  3.50G  3.94G  0  0  9.99K   1010
--  -  -  -  -  -  -
swpool   102K  3.69G  0  0 19  0
  /dev/rdsk/c7t0d0   102K  3.69G  0  0 19  0
--  -  -  -  -  -  -

I have booted up a linux rescue CD that has S.M.A.R.T support (system rescue 
CD) and performed the 'long' test on each drive. All drives pass the 'test'. 
There also appears to be no system errors with the drives under linux either.

Can anyone shed any light on this issue, or suggest what I could try next ? I 
am sort of discounting hardware problems given that I do not see errors from 
the live linux CD. Maybe I 

Re: [zfs-discuss] Disk Issues

2010-02-07 Thread Brian McKerr
 
 
 I'd say your easiest two options are swap ports and
 see if the problem
 follows the drive. If it does, swap the drive out.
 
 
 --Tim
 ___


Yep, that sounds like a plan.

Thanks for your suggestion.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Separate Zil on HDD ?

2009-12-02 Thread Brian McKerr
Hi all,

I have a home server based on SNV_127 with 8 disks;

2 x 500GB mirrored root pool
6 x 1TB raidz2 data pool

This server performs a few functions;

NFS : for several 'lab' ESX virtual machines
NFS : mythtv storage (videos, music, recordings etc)
Samba : for home directories for all networked PCs

I backup the important data to external USB hdd each day.


I previously had a linux NFS server that I had mounted 'ASYNC' and, as one 
would expect, NFS performance was pretty good getting close to 900gb/s. Now 
that I have moved to opensolaris,  NFS performance is not very good, I'm 
guessing mainly due to the 'SYNC' nature of NFS.  I've seen various threads and 
most point at 2 options;

1. Disable the ZIL
2. Add independent log device/s

I happen to have 2 x 250GB Western Digital RE3 7200rpm (Raid edition, rated for 
24x7 usage etc) hard drives sitting doing nothing and was wondering whether it 
might speed up NFS, and possibly general filesystem usage, by adding these 
devices as log devices to the data pool.  I understand that an SSD is 
considered ideal for log devices but I'm thinking that these 2 drives should at 
least be better than having the ZIL 'inside' the zpool.

If adding these devices, should I add them as mirrored or individual to get 
some sort of load balancing (according to zpool manpage) and perhaps a little 
bit more performance ?

I'm running ZFS version 19 which 'zpool upgrade -v' shows me as having 'log 
device removal' support. Can I easily remove these devices if I find that they 
have resulted in little/no performance improvements ?

Any help/tips greatly appreciated.

Cheers.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Backing up ZVOLs

2009-11-14 Thread Brian McKerr
Hello all,

Are there any best practices / recommendations for ways of doing this ?

In this case the ZVOLs would be iSCSI LUNS containing ESX VMs .I am aware 
of the of the need for the VMs to be quiesced for the backups to be useful.

Cheers.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backing up ZVOLs

2009-11-14 Thread Brian McKerr
Thanks for the help.

I was curious whether the zfs send|receive was considered suitable given a few 
things I've read which said somethings along the lines of don't count on being 
able to restore this stuff. Ideally that is what I would use with the 
'incremental' option so as to only backup changed blocks on subsequent backups.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss