Re: [zfs-discuss] zfs sata mirror slower than single disk

2012-07-17 Thread Michael Hase

sorry to insist, but still no real answer...

On Mon, 16 Jul 2012, Bob Friesenhahn wrote:


On Tue, 17 Jul 2012, Michael Hase wrote:


So only one thing left: mirror should read 2x


I don't think that mirror should necessarily read 2x faster even though the 
potential is there to do so.  Last I heard, zfs did not include a special 
read scheduler for sequential reads from a mirrored pair.  As a result, 50% 
of the time, a read will be scheduled for a device which already has a read 
scheduled.  If this is indeed true, the typical performance would be 150%. 
There may be some other scheduling factor (e.g. estimate of busyness) which 
might still allow zfs to select the right side and do better than that.


If you were to add a second vdev (i.e. stripe) then you should see very close 
to 200% due to the default round-robin scheduling of the writes.


My expectation would be  200%, as 4 disks are involved. It may not be the 
perfect 4x scaling, but imho it should be (and is for a scsi system) more 
than half of the theoretical throughput. This is solaris or a solaris 
derivative, not linux ;-)




It is really difficult to measure zfs read performance due to caching 
effects.  One way to do it is to write a large file (containing random data 
such as returned from /dev/urandom) to a zfs filesystem, unmount the 
filesystem, remount the filesystem, and then time how long it takes to read 
the file once.  The reason why this works is because remounting the 
filesystem restarts the filesystem cache.


Ok, did a zpool export/import cycle between the dd read and write test.
This really empties the arc, checked this with arc_summary.pl. the test 
even uses two processes in parallel (doesn't make a difference). Result is 
still the same:


dd write:  2x 58 MB/sec  -- perfect, each disk does  110 MB/sec
dd read:   2x 68 MB/sec  -- imho too slow, about 68 MB/sec per disk

For writes each disk gets 900 128k io requests/sec with asvc_t in the 8-9 
msec range. For reads each disk only gets 500 io requests/sec, asvc_t 
18-20 msec with the default zfs_vdev_maxpending=10. When reducing 
zfs_vdev_maxpending the asvc_t drops accordingly, the i/o rate remains at 
500/sec per disk, throughput also the same. I think iostat values should 
be reliable here. These high iops numbers make sense as we work on empty 
pools so there aren't very high seek times.


All benchmarks (dd, bonnie, will try iozone) lead to the same result: on 
the sata mirror pair read performance is in the range of a single disk. 
For the sas disks (only two available for testing) and for the scsi system 
there is quite good throughput scaling.


Here for comparison a table for 1-4 36gb 15k u320 scsi disks on an old 
sxde box (nevada b130):


seq write  factor   seq read  factor
MB/sec  MB/sec
single821 78   1
mirror791137   1.75
2x mirror1201.5  251   3.2

This is exactly what's imho to be expected from mirrors and striped 
mirrors. It just doesn't happen for my sata pool. Still have no reference 
numbers for other sata pools, just one with the 4k/512bytes sector problem 
which is even slower than mine. It seems the zfs performance people just 
use sas disks and be done.


Michael



Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
old ibm dual opteron intellistation with external hp msa30, 36gb 15k u320 scsi 
disks



  pool: scsi1
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
scsi1   ONLINE   0 0 0
  c3t4d0ONLINE   0 0 0

errors: No known data errors

Version  1.96   --Sequential Output-- --Sequential Input- --Random-
Concurrency   1 -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks--
MachineSize K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec %CP
zfssingle   16G   137  99 82739  20 39453   9   314  99 78251   7 856.9   8
Latency   160ms4799ms5292ms   43210us3274ms2069ms
Version  1.96   --Sequential Create-- Random Create
zfssingle   -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete--
  files  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP
 16  8819  34 + +++ 26318  68 20390  73 + +++ 26846  72
Latency 16413us 108us 231us   12206us  46us 124us
1.96,1.96,zfssingle,1,1342514790,16G,,137,99,82739,20,39453,9,314,99,78251,7,856.9,8,16,8819,34,+,+++,26318,68,20390,73,+,+++,26846,72,160ms,4799ms,5292ms,43210us,3274ms,2069ms,16413us,108us,231us,12206us,46us,124us

##

  pool: scsi1
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
scsi1   ONLINE   0 

Re: [zfs-discuss] zfs sata mirror slower than single disk

2012-07-17 Thread Bob Friesenhahn

On Tue, 17 Jul 2012, Michael Hase wrote:


If you were to add a second vdev (i.e. stripe) then you should see very 
close to 200% due to the default round-robin scheduling of the writes.


My expectation would be  200%, as 4 disks are involved. It may not be the 
perfect 4x scaling, but imho it should be (and is for a scsi system) more 
than half of the theoretical throughput. This is solaris or a solaris 
derivative, not linux ;-)


Here are some results from my own machine based on the 'virgin mount' 
test approach.  The results show less boost than is reported by a 
benchmark tool like 'iozone' which sees benefits from caching.


I get an initial sequential read speed of 657 MB/s on my new pool 
which has 1200 MB/s of raw bandwidth (if mirrors could produce 100% 
boost).  Reading the file a second time reports 6.9 GB/s.


The below is with a 2.6 GB test file but with a 26 GB test file (just 
add another zero to 'count' and wait longer) I see an initial read 
rate of 618 MB/s and a re-read rate of 8.2 GB/s.  The raw disk can 
transfer 150 MB/s.


% zpool status
   pool: tank
  state: ONLINE
status: The pool is formatted using an older on-disk format.  The pool can
 still be used, but some features are unavailable.
action: Upgrade the pool using 'zpool upgrade'.  Once this is done, the
 pool will no longer be accessible on older software versions.
   scan: scrub repaired 0 in 0h10m with 0 errors on Mon Jul 16 04:30:48 2012
config:

 NAME  STATE READ WRITE CKSUM
 tank  ONLINE   0 0 0
   mirror-0ONLINE   0 0 0
 c7t5393E8CA21FAd0p0   ONLINE   0 0 0
 c11t5393D8CA34B2d0p0  ONLINE   0 0 0
   mirror-1ONLINE   0 0 0
 c8t5393E8CA2066d0p0   ONLINE   0 0 0
 c12t5393E8CA2196d0p0  ONLINE   0 0 0
   mirror-2ONLINE   0 0 0
 c9t5393D8CA82A2d0p0   ONLINE   0 0 0
 c13t5393E8CA2116d0p0  ONLINE   0 0 0
   mirror-3ONLINE   0 0 0
 c10t5393D8CA59C2d0p0  ONLINE   0 0 0
 c14t5393D8CA828Ed0p0  ONLINE   0 0 0

errors: No known data errors
% pfexec zfs create tank/zfstest
% pfexec zfs create tank/zfstest/defaults
% cd /tank/zfstest/defaults
% pfexec dd if=/dev/urandom of=random.dat bs=128k count=2
2+0 records in
2+0 records out
262144 bytes (2.6 GB) copied, 36.8133 s, 71.2 MB/s
% cd ..
% pfexec zfs umount tank/zfstest/defaults
% pfexec zfs mount tank/zfstest/defaults
% cd defaults
% dd if=random.dat of=/dev/null bs=128k count=2
2+0 records in
2+0 records out
262144 bytes (2.6 GB) copied, 3.99229 s, 657 MB/s
% pfexec dd if=/dev/rdsk/c7t5393E8CA21FAd0p0 of=/dev/null bs=128k count=2000
2000+0 records in
2000+0 records out
262144000 bytes (262 MB) copied, 1.74532 s, 150 MB/s
% bc
scale=8
657/150
4.3800

It is very difficult to benchmark with a cache which works so well:

% dd if=random.dat of=/dev/null bs=128k count=2
2+0 records in
2+0 records out
262144 bytes (2.6 GB) copied, 0.379147 s, 6.9 GB/s

Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Problem: Disconnected command timeout for Target X

2012-07-17 Thread Roberto Scudeller
Hi all,

I'm using Opensolaris snv_134 with LSI Controllers and a motherboard
supermicro, with 20 sata disks, zfs in raid-10 conf. I mounted this
zfs_storage with NFS.
I'm not opensolaris specialist. What're the commands to show hardware
information? Like 'lshw' in linux but for opensolaris.

The storage stopped working, but ping responds. SSH and NFS is out. When I
open the console showing this messages:

Jul  2 13:00:27 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:00:27 storageDisconnected command timeout for Target 4
Jul  2 13:01:28 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:01:28 storageDisconnected command timeout for Target 3
Jul  2 13:02:28 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:02:28 storageDisconnected command timeout for Target 2
Jul  2 13:03:29 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:03:29 storageDisconnected command timeout for Target 1
Jul  2 13:04:29 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:04:29 storageDisconnected command timeout for Target 0
Jul  2 13:05:40 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:05:40 storageDisconnected command timeout for Target 6
Jul  2 13:06:40 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:06:40 storageDisconnected command timeout for Target 5

Any ideas? Could help me?

-- 
Roberto Scudeller
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problem: Disconnected command timeout for Target X

2012-07-17 Thread Bob Friesenhahn

On Tue, 17 Jul 2012, Roberto Scudeller wrote:


Hi all,

I'm using Opensolaris snv_134 with LSI Controllers and a motherboard 
supermicro, with 20 sata disks, zfs in raid-10 conf. I mounted this zfs_storage 
with
NFS.
I'm not opensolaris specialist. What're the commands to show hardware 
information? Like 'lshw' in linux but for opensolaris.


cfgadm, prtconf, prtpicl, prtdiag

zpool status

fmadm faulty

It sounds like you may have a broken cable or power supply failure to 
some disks.


Bob



The storage stopped working, but ping responds. SSH and NFS is out. When I open 
the console showing this messages:

Jul  2 13:00:27 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:00:27 storage    Disconnected command timeout for Target 4
Jul  2 13:01:28 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:01:28 storage    Disconnected command timeout for Target 3
Jul  2 13:02:28 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:02:28 storage    Disconnected command timeout for Target 2
Jul  2 13:03:29 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:03:29 storage    Disconnected command timeout for Target 1
Jul  2 13:04:29 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:04:29 storage    Disconnected command timeout for Target 0
Jul  2 13:05:40 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:05:40 storage    Disconnected command timeout for Target 6
Jul  2 13:06:40 storage scsi: [ID 107833 kern.warning] WARNING: 
/pci@0,0/pci8086,340a@3/pci1000,3140@0 (mpt2):
Jul  2 13:06:40 storage    Disconnected command timeout for Target 5

Any ideas? Could help me?

--
Roberto Scudeller






--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sata mirror slower than single disk

2012-07-17 Thread Michael Hase

On Tue, 17 Jul 2012, Bob Friesenhahn wrote:


On Tue, 17 Jul 2012, Michael Hase wrote:


If you were to add a second vdev (i.e. stripe) then you should see very 
close to 200% due to the default round-robin scheduling of the writes.


My expectation would be  200%, as 4 disks are involved. It may not be the 
perfect 4x scaling, but imho it should be (and is for a scsi system) more 
than half of the theoretical throughput. This is solaris or a solaris 
derivative, not linux ;-)


Here are some results from my own machine based on the 'virgin mount' test 
approach.  The results show less boost than is reported by a benchmark tool 
like 'iozone' which sees benefits from caching.


I get an initial sequential read speed of 657 MB/s on my new pool which has 
1200 MB/s of raw bandwidth (if mirrors could produce 100% boost).  Reading 
the file a second time reports 6.9 GB/s.


The below is with a 2.6 GB test file but with a 26 GB test file (just add 
another zero to 'count' and wait longer) I see an initial read rate of 618 
MB/s and a re-read rate of 8.2 GB/s.  The raw disk can transfer 150 MB/s.


To work around these caching effects just use a file  2 times the size 
of ram, iostat then shows the numbers really coming from disk. I always 
test like this. a re-read rate of 8.2 GB/s is really just memory 
bandwidth, but quite impressive ;-)



% pfexec zfs create tank/zfstest/defaults
% cd /tank/zfstest/defaults
% pfexec dd if=/dev/urandom of=random.dat bs=128k count=2
2+0 records in
2+0 records out
262144 bytes (2.6 GB) copied, 36.8133 s, 71.2 MB/s
% cd ..
% pfexec zfs umount tank/zfstest/defaults
% pfexec zfs mount tank/zfstest/defaults
% cd defaults
% dd if=random.dat of=/dev/null bs=128k count=2
2+0 records in
2+0 records out
262144 bytes (2.6 GB) copied, 3.99229 s, 657 MB/s
% pfexec dd if=/dev/rdsk/c7t5393E8CA21FAd0p0 of=/dev/null bs=128k 
count=2000

2000+0 records in
2000+0 records out
262144000 bytes (262 MB) copied, 1.74532 s, 150 MB/s
% bc
scale=8
657/150
4.3800

It is very difficult to benchmark with a cache which works so well:

% dd if=random.dat of=/dev/null bs=128k count=2
2+0 records in
2+0 records out
262144 bytes (2.6 GB) copied, 0.379147 s, 6.9 GB/s


This is not my point, I'm pretty sure I did not measure any arc effects - 
maybe with the one exception of the raid0 test on the scsi array. Don't 
know why the arc had this effect, filesize was 2x of ram. The point is: 
I'm searching for an explanation for the relative slowness of a mirror 
pair of sata disks, or some tuning knobs, or something like the disks are 
plain crap, or maybe: zfs throttles sata disks in general (don't know the 
internals).


In the range of  600 MB/s other issues may show up (pcie bus contention, 
hba contention, cpu load). And performance at this level could be just 
good enough, not requiring any further tuning. Could you recheck with only 
4 disks (2 mirror pairs)? If you just get some 350 MB/s it could be the 
same problem as with my boxes. All sata disks?


Michael



Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sata mirror slower than single disk

2012-07-17 Thread Bob Friesenhahn

On Tue, 17 Jul 2012, Michael Hase wrote:


The below is with a 2.6 GB test file but with a 26 GB test file (just add 
another zero to 'count' and wait longer) I see an initial read rate of 618 
MB/s and a re-read rate of 8.2 GB/s.  The raw disk can transfer 150 MB/s.


To work around these caching effects just use a file  2 times the size of 
ram, iostat then shows the numbers really coming from disk. I always test 
like this. a re-read rate of 8.2 GB/s is really just memory bandwidth, but 
quite impressive ;-)


Yes, in the past I have done benchmarking with file size 2X the size 
of memory.  This does not necessary erase all caching because the ARC 
is smart enough not to toss everything.


At the moment I have an iozone benchark run up from 8 GB to 256 GB 
file size.  I see that it has started the 256 GB size now.  It may be 
a while.  Maybe a day.


In the range of  600 MB/s other issues may show up (pcie bus contention, hba 
contention, cpu load). And performance at this level could be just good 
enough, not requiring any further tuning. Could you recheck with only 4 disks 
(2 mirror pairs)? If you just get some 350 MB/s it could be the same problem 
as with my boxes. All sata disks?


Unfortunately, I already put my pool into use and can not conveniently 
destroy it now.


The disks I am using are SAS (7200 RPM, 1 GB) but return similar 
per-disk data rates as the SATA disks I use for the boot pool.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Largest zpool / Most Vdevs

2012-07-17 Thread Matt Hardy
While I'm aware of the published maximums of zfs, I was wondering if 
anyone could share information on the largest zpool (both total usable 
capacity and vdevs) they've seen/deployed?


I know that generally speaking more vdevs = more iops, but I would think 
that there is a point of diminishing returns. Any opinions on this?


Cheers,
Matt Hardy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sata mirror slower than single disk

2012-07-17 Thread Bob Friesenhahn

On Tue, 17 Jul 2012, Michael Hase wrote:

To work around these caching effects just use a file  2 times the size of 
ram, iostat then shows the numbers really coming from disk. I always test 
like this. a re-read rate of 8.2 GB/s is really just memory bandwidth, but 
quite impressive ;-)


Ok, the iozone benchmark finally completed.  The results do suggest 
that reading from mirrors substantially improves the throughput. 
This is interesting since the results differ (better than) from my 
'virgin mount' test approach:


Command line used: iozone -a -i 0 -i 1 -y 64 -q 512 -n 8G -g 256G

  KB  reclen   write rewritereadreread
 8388608  64  572933 1008668  6945355  7509762
 8388608 128 2753805 2388803  6482464  7041942
 8388608 256 2508358 2331419  2969764  3045430
 8388608 512 2407497 2131829  3021579  3086763
16777216  64  671365  879080  6323844  6608806
16777216 128 1279401 2286287  6409733  6739226
16777216 256 2382223 2211097  2957624  3021704
16777216 512 2237742 2179611  3048039  3085978
33554432  64  933712  699966  6418428  6604694
33554432 128  459896  431640  6443848  6546043
33554432 256  90  430989  2997615  3026246
33554432 512  427158  430891  3042620  3100287
67108864  64  426720  427167  6628750  6738623
67108864 128  419328  422581  153  6743711
67108864 256  419441  419129  3044352  3056615
67108864 512  431053  417203  3090652  3112296
   134217728  64  417668   55434   759351   760994
   134217728 128  409383  400433   759161   765120
   134217728 256  408193  405868   763892   766184
   134217728 512  408114  403473   761683   766615
   268435456  64  418910   55239   768042   768498
   268435456 128  408990  399732   763279   766882
   268435456 256  413919  399386   760800   764468
   268435456 512  410246  403019   766627   768739

Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problem: Disconnected command timeout for Target X

2012-07-17 Thread Roberto Scudeller
Hi Bob,

Thanks for the answers.

How do I test your theory?

In this case, I use common disks SATA 2, not Nearline SAS (NL SATA) or SAS.
Do you think the disks SATA are the problem?

Cheers,


2012/7/17 Bob Friesenhahn bfrie...@simple.dallas.tx.us

 On Tue, 17 Jul 2012, Roberto Scudeller wrote:

  Hi all,

 I'm using Opensolaris snv_134 with LSI Controllers and a motherboard
 supermicro, with 20 sata disks, zfs in raid-10 conf. I mounted this
 zfs_storage with
 NFS.
 I'm not opensolaris specialist. What're the commands to show hardware
 information? Like 'lshw' in linux but for opensolaris.


 cfgadm, prtconf, prtpicl, prtdiag

 zpool status

 fmadm faulty

 It sounds like you may have a broken cable or power supply failure to some
 disks.

 Bob



 The storage stopped working, but ping responds. SSH and NFS is out. When
 I open the console showing this messages:

 Jul  2 13:00:27 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:00:27 storageDisconnected command timeout for Target 4
 Jul  2 13:01:28 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:01:28 storageDisconnected command timeout for Target 3
 Jul  2 13:02:28 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:02:28 storageDisconnected command timeout for Target 2
 Jul  2 13:03:29 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:03:29 storageDisconnected command timeout for Target 1
 Jul  2 13:04:29 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:04:29 storageDisconnected command timeout for Target 0
 Jul  2 13:05:40 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:05:40 storageDisconnected command timeout for Target 6
 Jul  2 13:06:40 storage scsi: [ID 107833 kern.warning] WARNING: /pci@0
 ,0/pci8086,340a@3/**pci1000,3140@0 (mpt2):
 Jul  2 13:06:40 storageDisconnected command timeout for Target 5

 Any ideas? Could help me?

 --
 Roberto Scudeller





 --
 Bob Friesenhahn
 bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/**
 users/bfriesen/ http://www.simplesystems.org/users/bfriesen/
 GraphicsMagick Maintainer,http://www.GraphicsMagick.org/




-- 
Roberto Scudeller
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Problem: Disconnected command timeout for Target X

2012-07-17 Thread Bob Friesenhahn

On Tue, 17 Jul 2012, Roberto Scudeller wrote:


Hi Bob,

Thanks for the answers.

How do I test your theory?


I would use 'dd' to see if it is possible to transfer data from one of 
the problem devices.  Gain physical access to the system and check the 
signal and power cables to these devices closely.


Use 'iostat -xe' to see what error counts have accumulated.  Also 
'iostat -E'.



In this case, I use common disks SATA 2, not Nearline SAS (NL SATA) or SAS. Do 
you think the disks SATA are the problem?


There have been reports of congestion leading to timeouts and resets 
when SATA disks are on expanders.  There have also been reports that 
one failing disk can cause problems when on expanders.  Regardless, if 
this system has been previously operating fine for some time, these 
errors would indicate a change in the hardware shared by all these 
devices.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Has anyone switched from IR - IT firmware on the fly ? (existing zpool on LSI 9211-8i)

2012-07-17 Thread Jason Usher
We have a running zpool with a 12 disk raidz3 vdev in it ... we gave ZFS the 
full, raw disks ... all is well.

However, we built it on two LSI 9211-8i cards and we forgot to change from IR 
firmware to IT firmware.

Is there any danger in shutting down the OS, flashing the cards to IT firmware, 
and then booting back up ?

We did not create any raid configuration - as far as we know, the LSI cards are 
just passing through the disks to ZFS ... but maybe not ?

I'd like to hear of someone else doing this successfully before we try it ...


We created the zpool with raw disks:

zpool create -m /mount/point MYPOOL raidz3 da{0,1,2,3,4,5,6,7,8,9,10,11}

and diskinfo tells us that each disk is:

da1 512 3000592982016   5860533168

The physical label (the sticker) on the disk also says 5860533168 sectors ... 
so that seems to line up ...


Someone else in the world has made this change while inflight and can confirm 
?

Thanks.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone switched from IR - IT firmware on the fly ? (existing zpool on LSI 9211-8i)

2012-07-17 Thread Damon Pollard
Hi Jason,

I have done this in the past. (3x LSI 1068E - IBM BR10i).

Your pool has no tie with the hardware used to host it (including your
HBA). You could change all your hardware, and still import your pool
correctly.

If you really want to be on the safe side; you can export your pool before
the firmware change and then import when your satisfied the firmware
change is complete.

Export: http://docs.oracle.com/cd/E19082-01/817-2271/gazqr/index.html
Import: http://docs.oracle.com/cd/E19082-01/817-2271/gazuf/index.html

Damon Pollard

On Wed, Jul 18, 2012 at 6:14 AM, Jason Usher jushe...@yahoo.com wrote:

 We have a running zpool with a 12 disk raidz3 vdev in it ... we gave ZFS
 the full, raw disks ... all is well.

 However, we built it on two LSI 9211-8i cards and we forgot to change from
 IR firmware to IT firmware.

 Is there any danger in shutting down the OS, flashing the cards to IT
 firmware, and then booting back up ?

 We did not create any raid configuration - as far as we know, the LSI
 cards are just passing through the disks to ZFS ... but maybe not ?

 I'd like to hear of someone else doing this successfully before we try it
 ...


 We created the zpool with raw disks:

 zpool create -m /mount/point MYPOOL raidz3 da{0,1,2,3,4,5,6,7,8,9,10,11}

 and diskinfo tells us that each disk is:

 da1 512 3000592982016   5860533168

 The physical label (the sticker) on the disk also says 5860533168 sectors
 ... so that seems to line up ...


 Someone else in the world has made this change while inflight and can
 confirm ?

 Thanks.
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone switched from IR - IT firmware on the fly ? (existing zpool on LSI 9211-8i)

2012-07-17 Thread Jason Usher

Ok, and your LSI 1068E also had alternate IR and IT firmwares, and you went 
from IR - IT ?

Is that correct ?

Thanks.


--- On Tue, 7/17/12, Damon Pollard damon.poll...@birchmangroup.com wrote:

From: Damon Pollard damon.poll...@birchmangroup.com
Subject: Re: [zfs-discuss] Has anyone switched from IR - IT firmware on the 
fly ? (existing zpool on LSI 9211-8i)
To: Jason Usher jushe...@yahoo.com
Cc: zfs-discuss@opensolaris.org
Date: Tuesday, July 17, 2012, 5:05 PM

Hi Jason,
I have done this in the past. (3x LSI 1068E - IBM BR10i).
Your pool has no tie with the hardware used to host it (including your HBA). 
You could change all your hardware, and still import your pool correctly.

If you really want to be on the safe side; you can export your pool before the 
firmware change and then import when your satisfied the firmware change is 
complete.
Export: http://docs.oracle.com/cd/E19082-01/817-2271/gazqr/index.html
Import: http://docs.oracle.com/cd/E19082-01/817-2271/gazuf/index.html
Damon Pollard


On Wed, Jul 18, 2012 at 6:14 AM, Jason Usher jushe...@yahoo.com wrote:

We have a running zpool with a 12 disk raidz3 vdev in it ... we gave ZFS the 
full, raw disks ... all is well.



However, we built it on two LSI 9211-8i cards and we forgot to change from IR 
firmware to IT firmware.



Is there any danger in shutting down the OS, flashing the cards to IT firmware, 
and then booting back up ?



We did not create any raid configuration - as far as we know, the LSI cards are 
just passing through the disks to ZFS ... but maybe not ?



I'd like to hear of someone else doing this successfully before we try it ...





We created the zpool with raw disks:



zpool create -m /mount/point MYPOOL raidz3 da{0,1,2,3,4,5,6,7,8,9,10,11}



and diskinfo tells us that each disk is:



da1     512     3000592982016   5860533168



The physical label (the sticker) on the disk also says 5860533168 sectors ... 
so that seems to line up ...





Someone else in the world has made this change while inflight and can confirm 
?



Thanks.

___

zfs-discuss mailing list

zfs-discuss@opensolaris.org

http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Has anyone switched from IR - IT firmware on the fly ? (existing zpool on LSI 9211-8i)

2012-07-17 Thread Damon Pollard
Correct.

LSI 1068E has IR and IT firmwares + I have gone from IR - IT and IT - IR
without hassle.

Damon Pollard


On Wed, Jul 18, 2012 at 8:13 AM, Jason Usher jushe...@yahoo.com wrote:


 Ok, and your LSI 1068E also had alternate IR and IT firmwares, and you
 went from IR - IT ?

 Is that correct ?

 Thanks.


 --- On Tue, 7/17/12, Damon Pollard damon.poll...@birchmangroup.com
 wrote:

 From: Damon Pollard damon.poll...@birchmangroup.com
 Subject: Re: [zfs-discuss] Has anyone switched from IR - IT firmware on
 the fly ? (existing zpool on LSI 9211-8i)
 To: Jason Usher jushe...@yahoo.com
 Cc: zfs-discuss@opensolaris.org
 Date: Tuesday, July 17, 2012, 5:05 PM

 Hi Jason,
 I have done this in the past. (3x LSI 1068E - IBM BR10i).
 Your pool has no tie with the hardware used to host it (including your
 HBA). You could change all your hardware, and still import your pool
 correctly.

 If you really want to be on the safe side; you can export your pool before
 the firmware change and then import when your satisfied the firmware
 change is complete.
 Export: http://docs.oracle.com/cd/E19082-01/817-2271/gazqr/index.html
 Import: http://docs.oracle.com/cd/E19082-01/817-2271/gazuf/index.html
 Damon Pollard


 On Wed, Jul 18, 2012 at 6:14 AM, Jason Usher jushe...@yahoo.com wrote:

 We have a running zpool with a 12 disk raidz3 vdev in it ... we gave ZFS
 the full, raw disks ... all is well.



 However, we built it on two LSI 9211-8i cards and we forgot to change from
 IR firmware to IT firmware.



 Is there any danger in shutting down the OS, flashing the cards to IT
 firmware, and then booting back up ?



 We did not create any raid configuration - as far as we know, the LSI
 cards are just passing through the disks to ZFS ... but maybe not ?



 I'd like to hear of someone else doing this successfully before we try it
 ...





 We created the zpool with raw disks:



 zpool create -m /mount/point MYPOOL raidz3 da{0,1,2,3,4,5,6,7,8,9,10,11}



 and diskinfo tells us that each disk is:



 da1 512 3000592982016   5860533168



 The physical label (the sticker) on the disk also says 5860533168 sectors
 ... so that seems to line up ...





 Someone else in the world has made this change while inflight and can
 confirm ?



 Thanks.

 ___

 zfs-discuss mailing list

 zfs-discuss@opensolaris.org

 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss