Re: [zfs-discuss] drive speeds etc

2010-09-28 Thread Fei Xu
I have both EVDS and EARS 2TB green drive.  And I have to say they are not good 
to build storage servers.

EVDS has compatibility issue with my supermicro appliance.  it will hang when 
doing huge data send or copy.  from IOSTAT I can see the data throughput is 
stuck on green disks with extremely high wait time.

for EARS drives, they are ok running opensolaris but veryvery poor performance 
handling small files.  I'm doing SVN via NFS share and it takes triple time CO 
a repository compare to NEtapp FAS960.  but the issue could be resolved by 
adding SSD as log device.  notice that, green disk seek time is 15ms and normal 
7200RPM disk is around 8.5ms.
I'll try the link provided by Christian to see if it can help the performance.

anyway, I've decided to use seagate 7200.11 which is big enough and fast.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-10 Thread Fei Xu
  by the way, in HDtune, I saw C7: Ultra DMA CRC
 error count is a little high which indicates a
 potential connection issue.  Maybe all are caused by
 the enclosure?
 
 Bingo!


You are right, I've done a lot of tests and the defect is narrorw down the  
problem hardware.  The two pool works fine in one chassis but after moved to 
original enclosure, it just failed when CP or ZFS send.  I also noticed when 
the machine bootup reading ZFS configure, there is a warning message.

 REading ZFS config: *Warning /p...@0,0/pci8086,3...@8/pci15d9,1...@0(mpt0):
Discovery in progress, can't verify IO unit config.

I did search a lot but cannot find more details.  
my 2 server configuration:
1. PRoblem chassis   supermicro SuperChassis847e2.  Tysonberg MB with onboard 
LSI 1068e (IT mode, which direct expose HD to system without RAID), Single 
Xeon5520.
2. Good Chassis:Self-developed chassis by other department.  S5000WB MB, 
single E5504, 2 PCIe-4x LSI 3081 HBA card.

Seems the SAS cable are all connecting right.  I suspect the issue of onboard 
1068e and moving the LSI3081 card to the problem server to test.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-09 Thread Fei Xu
 
 Service times here are crap. Disks are malfunctioning
 in some way. If
 your source disks can take seconds (or 10+ seconds)
 to reply, then of
 course your copy will be slow. Disk is probably
 having a hard time
 reading the data or something.
 


Yeah, that should not go over 15ms.  I just cannot understand why it starts ok 
with hundred GB files transfered and then suddenly fall to sleep.
by the way,  WDIDLE time is already disabled which might cause some issue.  
I've changed to another system to test ZFS send between 8*1TB pool and 4*1TB 
pool.  hope everythings OK on this case.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-09 Thread Fei Xu
Just to update the status and findings.

I've checked TLER settings and they are off by default.

I moved the source pool to another chassis and do the 3.8TB send again.  this 
time, not any problems!  the difference is 
1. New chassis
2. BIGGER memory.  32GB v.s 12GB
3. although wdidle time is disabled by default, I've change the HD mode from 
silent to performance in HDtune.  this is what I once heard from some website 
that might also fix the disk head park/unpark issue (aka, C1).

seems TLER is not the root cause or at least, set to off is ok.

my next step will be 
1. move back HD to see if it's the performance mode fix the issue
2. if not, add more memory and try again.

by the way, in HDtune, I saw C7: Ultra DMA CRC error count is a little high 
which indicates a potential connection issue.  Maybe all are caused by the 
enclosure?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] performance leakage when copy huge data

2010-09-08 Thread Fei Xu
Hi all:

I'm a new guy who is just started ZFS for half a year.  We are using 
Nexenta in corporate pilot environment.  these days, when I was trying to move 
around 4TB data from an old pool(4*2TB raidz) to new pool (11*2TB raidz2), it 
seems will never end up successfully.
1. I used CP first.  the initial speed is great @ 250MB/s.  but it will 
suddenly become s...@400kb/s when around 400-600GB data is copied.  I've tried 
several times.
2. Use Rsync.  Speed is crab, around 15MB/s.  it will end up good but 
performance seems dropped even after Rsync finished.
3. use ZFS send|recv.  Speed is close to CP.  but still suddenly go down to 
xxxKB/s after around 1.5TB migrated.
4. I've tried CP to remote server which has 1Gb nic connection.  It can be 
finished but I can see a significant system slow down.

what's the root cause of that?
is it because the memory is too low?  all my ZFS system has 12GB memory.  I'm 
considering adding at least 12GB more to each of them.

thanks
fei
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-08 Thread Fei Xu
thank you Ian.  I've re-build the pool to 9*2TB Raidz2 and start the ZFS send 
command.  result will come out after about 3 hours.

thanks
fei
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-08 Thread Fei Xu
now it gets extremly slow at around 400G sent.  

first iostat result is captured when the send operation starts.

capacity operationsbandwidth
pool alloc   free   read  write   read  write
---  -  -  -  -  -  -
sh001a   37.6G  16.2T  0  1.17K 82   146M
  raidz2 37.6G  16.2T  0  1.17K 82   146M
c0t10d0  -  -  0201974  21.0M
c0t11d0  -  -  0201974  21.1M
c0t23d0  -  -  0201  1.56K  21.0M
c0t24d0  -  -  0201  1.26K  21.0M
c0t25d0  -  -  0201662  21.1M
c0t26d0  -  -  0201  1.26K  21.1M
c0t2d0   -  -  0202974  21.1M
c0t5d0   -  -  0200662  20.9M
c0t6d0   -  -  0200  1.26K  21.0M
---  -  -  -  -  -  -
syspool  10.6G   137G 11 13   668K   137K
  c3d0s0 10.6G   137G 11 13   668K   137K
---  -  -  -  -  -  -
vol1 5.40T  1.85T621  5  76.9M  12.4K
  raidz1 5.40T  1.85T621  5  76.9M  12.4K
c0t22d0  -  -280  3  19.5M  14.2K
c0t3d0   -  -279  3  19.5M  13.9K
c0t20d0  -  -280  3  19.5M  14.2K
c0t21d0  -  -280  3  19.5M  13.9K
---  -  -  -  -  -  -

---
below result is when ZFS send stuck @ 397G.  Seems the HD I/O is quite normal.  
then, where is the data... notice that, IOstat command response very slow.

capacity operationsbandwidth
pool alloc   free   read  write   read  write
---  -  -  -  -  -  -
sh001a397G  15.9T  0  1.08K490   136M
  raidz2  397G  15.9T  0  1.08K490   136M
c0t10d0  -  -  0185  1.68K  19.4M
c0t11d0  -  -  0185  1.71K  19.4M
c0t23d0  -  -  0185  1.99K  19.4M
c0t24d0  -  -  0185  1.79K  19.4M
c0t25d0  -  -  0185  2.10K  19.4M
c0t26d0  -  -  0185  2.07K  19.4M
c0t2d0   -  -  0185  1.99K  19.4M
c0t5d0   -  -  0185  2.12K  19.4M
c0t6d0   -  -  0185  2.23K  19.4M
---  -  -  -  -  -  -
syspool  10.6G   137G  2  6   131K  48.0K
  c3d0s0 10.6G   137G  2  6   131K  48.0K
---  -  -  -  -  -  -
vol1 5.40T  1.85T   1009  1   125M  2.85K
  raidz1 5.40T  1.85T   1009  1   125M  2.85K
c0t22d0  -  -453  0  31.6M  2.64K
c0t3d0   -  -452  0  31.6M  2.58K
c0t20d0  -  -453  0  31.6M  2.64K
c0t21d0  -  -453  0  31.6M  2.56K
---  -  -  -  -  -  -
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-08 Thread Fei Xu

 ve you get dedup enabled?  Note the read bandwith is
 much higher.
 
 -- 
 Ian.
 

no, dedup is not enabled since it's still not stable enough even for test 
environment.
here is a JPG of Read/Write indicator.  RED line is read and GREEN line is 
write.
you can see, because destination pool has more disks, more throughput ability, 
so, most of the time, it's just waiting for source pool reading.

http://bbs.manmi.com/UploadFile/2010-9/20109911304144848.jpg
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] performance leakage when copy huge data

2010-09-08 Thread Fei Xu
I dig deeper into it and might find some useful information.
I attached an X25 SSD for ZIL to see if it helps.  but no luck.
I run IOstate -xnz for more details and got interesting result as below.(maybe 
too long)
some explaination:
1. c2d0 is SSD for ZIL
2. c0t3d0, c0t20d0, c0t21d0, c0t22d0 is source pool.
3. you can see the first result is the same as zpool iostate -v, and I've ran 
-xnz several times, the first result were always good.
4. then, 30s later comes up the second one, and then 3rd, 4th.  you can see my 
source pool disks are extremely slow and busy.  One by One.  asvc_t is high, %b 
is high.  that's abnormal.
5. I think it might because those 4 drives are the old WD green disks with 
defect of disk head park/unpark and TLER.  other 12 2TB disks are the latest 
one which already fixed.
6. Let me try to test on 1TB disk and see if any issue.



r...@sh1slcs001:/volumes# iostat -xnz 30
extended device statistics  
r/sw/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
0.40.03.00.3  0.0  0.00.00.2   0   0 c2d0
1.48.9   57.8   40.3  0.0  0.12.55.8   1   5 c3d0
0.00.00.00.0  0.0  0.00.0  986.5   0   0 fd0
2.8  154.8  141.5 16517.7  0.0  2.40.1   15.0   2  25 c0t2d0
  408.40.3 29140.40.9  0.0  4.10.1   10.1   2  55 c0t3d0
2.8  154.9  136.7 16519.6  0.0  2.40.1   14.9   2  25 c0t5d0
2.8  155.0  139.4 16520.9  0.0  2.40.1   15.0   2  26 c0t6d0
0.00.00.50.0  0.0  0.00.00.1   0   0 c0t7d0
0.00.00.50.0  0.0  0.00.03.4   0   0 c0t8d0
0.00.00.40.0  0.0  0.00.00.1   0   0 c0t9d0
2.8  154.9  139.2 16523.9  0.0  2.40.1   15.2   2  26 c0t10d0
2.8  155.0  136.1 16525.4  0.0  2.50.1   15.7   2  26 c0t11d0
  409.90.3 29152.10.9  0.0  3.90.19.4   2  52 c0t20d0
  409.50.3 29152.00.9  0.0  3.90.19.6   2  52 c0t21d0
  409.20.3 29155.70.9  0.0  4.10.1   10.0   2  55 c0t22d0
2.8  155.0  138.6 16527.6  0.0  2.40.1   15.2   2  26 c0t23d0
2.9  154.9  138.0 16528.8  0.0  2.40.1   15.1   2  26 c0t24d0
2.9  155.0  136.1 16530.2  0.0  2.50.1   15.8   2  27 c0t25d0
2.8  155.1  139.7 16531.6  0.0  2.40.1   15.5   2  26 c0t26d0
extended device statistics  
r/sw/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
0.30.01.20.0  0.0  0.00.00.1   0   0 c2d0
0.1   17.70.1   51.7  0.0  0.10.24.1   0   7 c3d0
0.12.10.0   79.8  0.0  0.00.14.0   0   0 c0t2d0
0.20.07.10.0  0.1  2.3  278.5 11365.1   1  46 c0t3d0
0.12.20.0   79.9  0.0  0.00.13.7   0   0 c0t5d0
0.12.30.0   80.0  0.0  0.00.19.2   0   0 c0t6d0
0.12.50.0   80.1  0.0  0.00.13.8   0   0 c0t10d0
0.12.40.0   80.0  0.0  0.00.19.5   0   0 c0t11d0
1.90.0  133.00.0  0.1  2.8   60.2 1520.6   2  51 c0t20d0
1.60.0  110.90.0  0.0  0.00.07.6   0   0 c0t21d0
1.60.0  109.50.0  0.0  0.00.0   11.5   0   0 c0t22d0
0.12.40.0   80.0  0.0  0.00.18.6   0   0 c0t23d0
0.12.60.0   80.0  0.0  0.00.12.9   0   0 c0t24d0
0.12.40.0   79.9  0.0  0.00.16.2   0   0 c0t25d0
0.12.30.0   79.9  0.0  0.00.17.8   0   0 c0t26d0
extended device statistics  
r/sw/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
0.10.00.60.0  0.0  0.00.00.1   0   0 c2d0
0.0   21.30.0   82.3  0.0  0.10.44.0   0   8 c3d0
0.02.10.0   68.2  0.0  0.00.23.8   0   0 c0t2d0
0.70.0   39.10.0  0.0  0.6   64.0  884.1   1  10 c0t3d0
0.02.20.0   68.2  0.0  0.00.13.5   0   0 c0t5d0
0.02.30.0   68.3  0.0  0.00.13.5   0   0 c0t6d0
0.02.30.0   68.3  0.0  0.00.13.2   0   0 c0t10d0
0.02.20.0   68.3  0.0  0.00.13.7   0   0 c0t11d0
2.00.0  133.70.0  0.0  0.00.09.4   0   0 c0t20d0
2.10.0  135.80.0  0.1  5.2   67.8 2498.1   3  88 c0t21d0
2.80.0  134.40.0  0.0  0.00.03.2   0   0 c0t22d0
0.02.10.0   68.2  0.0  0.00.13.6   0   0 c0t23d0
0.02.40.0   68.4  0.0  0.00.13.1   0   0 c0t24d0
0.02.40.0   68.4  0.0  0.00.13.4   0   0 c0t25d0
0.02.20.0   68.3  0.0  0.00.13.3   0   0 c0t26d0
extended device statistics  
r/sw/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
0.10.00.60.0  0.0  0.00.00.1   0   0 c2d0
0.1   16.50.1   44.5  0.0  0.10.24.8   0   7 c3d0
0.02.30.0   73.8  0.0  0.00.13.8   0   0 c0t2d0
3.5