[zfs-discuss] NFS performance?

2010-07-23 Thread Sigbjorn Lie
Hi,

I've been searching around on the Internet to fine some help with this, but 
have been
unsuccessfull so far.

I have some performance issues with my file server. I have an OpenSolaris 
server with a Pentium D
3GHz CPU, 4GB of memory, and a RAIDZ1 over 4 x Seagate (ST31500341AS) 1,5TB 
SATA drives.

If I compile or even just unpack a tar.gz archive with source code (or any 
archive with lots of
small files), on my Linux client onto a NFS mounted disk to the OpenSolaris 
server, it's extremely
slow compared to unpacking this archive on the locally on the server. A 22MB 
.tar.gz file
containng 7360 files takes 9 minutes and 12seconds to unpack over NFS.

Unpacking the same file locally on the server is just under 2 seconds. Between 
the server and
client I have a gigabit network, which at the time of testing had no other 
significant load. My
NFS mount options are: rw,hard,intr,nfsvers=3,tcp,sec=sys.

Any suggestions to why this is?


Regards,
Sigbjorn


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Confused about consumer drives and zfs can someone help?

2010-07-23 Thread tomwaters
There is alot there to reply to...but I will try and help...

Re. TLER. Do not worry about TLER when using ZFS. ZFS will handle it either way 
and will NOT time out and drop the drive...it may wait a long time, but it will 
not time out and drop the drive - nor will it have an issue if you do enable 
TLER-ON (which sets time out to 7 seonds).  I run both with TLER-ON (disks from 
an old mdadm raid array) and without (TLER-OFF) 1.5TB WDEADS.

I can not speak for the WD EARS, but the WDEADS are fine in my home nas. I also 
run 1.5TB Samsung Green/Silencer series and Seagate 11's. Others swear by 
Hitachi. I would recommend the Samsung or Hitachi and not the new WD EARS which 
have that 4k sectors or whatever it is.

Re the CPU, do not go low power Atom etc, go a newish Core2 duo...the power 
differential at idle is bugger all and when you want to use the nas, ZFS will 
make good use of the CPU. Honestly, sit down and do the calculations on power 
savings of a low power cpu and you'll see it's better to just not have that 5th 
beer on a friday - you'll save more money that way and be MUCH happier with 
your nas performance.

re. cards...I use and recommend these 8-Port SUPERMICRO AOC-USASLP-L8I UIO SAS. 
They are cheap on e-bay, just work and are fast. Use them.

You do want alot of ram I use 8GB, but you can use 4. Ram is cheap, ZFS loves 
ram, just buy 8.

IMHO (and that of the best practice guide) - you should mirror the rpool (o/s 
disk). Just buy 2 cheap laptop drives and when installing choose to mirror 
them.

I hope that helps.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Phil Harman
That's because NFS adds synchronous writes to the mix (e.g. the client needs to 
know certain transactions made it to nonvolatile storage in case the server 
restarts etc). The simplest safe solution, although not cheap, is to add an SSD 
log device to the pool.

On 23 Jul 2010, at 08:11, Sigbjorn Lie sigbj...@nixtra.com wrote:

 Hi,
 
 I've been searching around on the Internet to fine some help with this, but 
 have been
 unsuccessfull so far.
 
 I have some performance issues with my file server. I have an OpenSolaris 
 server with a Pentium D
 3GHz CPU, 4GB of memory, and a RAIDZ1 over 4 x Seagate (ST31500341AS) 1,5TB 
 SATA drives.
 
 If I compile or even just unpack a tar.gz archive with source code (or any 
 archive with lots of
 small files), on my Linux client onto a NFS mounted disk to the OpenSolaris 
 server, it's extremely
 slow compared to unpacking this archive on the locally on the server. A 22MB 
 .tar.gz file
 containng 7360 files takes 9 minutes and 12seconds to unpack over NFS.
 
 Unpacking the same file locally on the server is just under 2 seconds. 
 Between the server and
 client I have a gigabit network, which at the time of testing had no other 
 significant load. My
 NFS mount options are: rw,hard,intr,nfsvers=3,tcp,sec=sys.
 
 Any suggestions to why this is?
 
 
 Regards,
 Sigbjorn
 
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Confused about consumer drives and zfs can someone help?

2010-07-23 Thread Thomas Burgess
I've found the Seagate 7200.12 1tb drives and Hitachi 7k2000 2TB drives to
be by far the best.

I've read lots of horror stories about any WD drive with 4k
sectorsit'sbest to stay away from them.

I've also read plenty of people say that the green drives are terrible.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] L2ARC and ZIL on same SSD?

2010-07-23 Thread Thomas Burgess
On Wed, Jul 21, 2010 at 12:42 PM, Orvar Korvar 
knatte_fnatte_tja...@yahoo.com wrote:

 Are there any drawbacks to partition a SSD in two parts and use L2ARC on
 one partition, and ZIL on the other? Any thoughts?
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



It's not going to be as good as having separate but i can tell you that i
did this on my home system and it was WELL worth it.

I used one of the sandforce 1500 based SSD's 50 gb

i used 9 gb for ZIL, and the rest for L2ARC.   adding the zil gave me about
400-500% nfs write performance.   Seeing as you can't ever use more than
half your ram for ZIL anyways, the only real downside to doing this is that
i/o becomes split between zil and L2arc but realistically it depends on your
workloadfor mine, i noticed a HUGE benefit from doing this.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Thomas Burgess
On Fri, Jul 23, 2010 at 3:11 AM, Sigbjorn Lie sigbj...@nixtra.com wrote:

 Hi,

 I've been searching around on the Internet to fine some help with this, but
 have been
 unsuccessfull so far.

 I have some performance issues with my file server. I have an OpenSolaris
 server with a Pentium D
 3GHz CPU, 4GB of memory, and a RAIDZ1 over 4 x Seagate (ST31500341AS) 1,5TB
 SATA drives.

 If I compile or even just unpack a tar.gz archive with source code (or any
 archive with lots of
 small files), on my Linux client onto a NFS mounted disk to the OpenSolaris
 server, it's extremely
 slow compared to unpacking this archive on the locally on the server. A
 22MB .tar.gz file
 containng 7360 files takes 9 minutes and 12seconds to unpack over NFS.

 Unpacking the same file locally on the server is just under 2 seconds.
 Between the server and
 client I have a gigabit network, which at the time of testing had no other
 significant load. My
 NFS mount options are: rw,hard,intr,nfsvers=3,tcp,sec=sys.

 Any suggestions to why this is?


 Regards,
 Sigbjorn


 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




as someone else said, adding an ssd log device can help hugely.  I saw about
a 500% nfs write increase by doing this.
I've heard of people getting even more.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Andrew Gabriel

Thomas Burgess wrote:



On Fri, Jul 23, 2010 at 3:11 AM, Sigbjorn Lie sigbj...@nixtra.com 
mailto:sigbj...@nixtra.com wrote:


Hi,

I've been searching around on the Internet to fine some help with
this, but have been
unsuccessfull so far.

I have some performance issues with my file server. I have an
OpenSolaris server with a Pentium D
3GHz CPU, 4GB of memory, and a RAIDZ1 over 4 x Seagate
(ST31500341AS) 1,5TB SATA drives.

If I compile or even just unpack a tar.gz archive with source code
(or any archive with lots of
small files), on my Linux client onto a NFS mounted disk to the
OpenSolaris server, it's extremely
slow compared to unpacking this archive on the locally on the
server. A 22MB .tar.gz file
containng 7360 files takes 9 minutes and 12seconds to unpack over NFS.

Unpacking the same file locally on the server is just under 2
seconds. Between the server and
client I have a gigabit network, which at the time of testing had
no other significant load. My
NFS mount options are: rw,hard,intr,nfsvers=3,tcp,sec=sys.

Any suggestions to why this is?


Regards,
Sigbjorn


as someone else said, adding an ssd log device can help hugely.  I saw 
about a 500% nfs write increase by doing this.

I've heard of people getting even more.


Another option if you don't care quite so much about data security in 
the event of an unexpected system outage would be to use Robert 
Milkowski and Neil Perrin's zil synchronicity [PSARC/2010/108] changes 
with sync=disabled, when the changes work their way into an available 
build. The risk is that if the file server goes down unexpectedly, it 
might come back up having lost some seconds worth of changes which it 
told the client (lied) that it had committed to disk, when it hadn't, 
and this violates the NFS protocol. That might be OK if you are using it 
to hold source that's being built, where you can kick off a build again 
if the server did go down in the middle of it. Wouldn't be a good idea 
for some other applications though (although Linux ran this way for many 
years, seemingly without many complaints). Note that there's no 
increased risk of the zpool going bad - it's just that after the reboot, 
filesystems with sync=disabled will look like they were rewound by some 
seconds (possibly up to 30 seconds).


--
Andrew Gabriel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread tomwaters
I agree, I get apalling NFS speeds compared to CIFS/Samba..ie. CIFS/Samba of 
95-105MB and NFS of 5-20MB.

Not the thread hijack, but I assume a SSD ZIL will similarly improve an iSCSI 
target...as I am getting 2-5MB on that too.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Phil Harman
On 23 Jul 2010, at 09:18, Andrew Gabriel andrew.gabr...@oracle.com wrote:

 Thomas Burgess wrote:
 
 On Fri, Jul 23, 2010 at 3:11 AM, Sigbjorn Lie sigbj...@nixtra.com 
 mailto:sigbj...@nixtra.com wrote:
 
Hi,
 
I've been searching around on the Internet to fine some help with
this, but have been
unsuccessfull so far.
 
I have some performance issues with my file server. I have an
OpenSolaris server with a Pentium D
3GHz CPU, 4GB of memory, and a RAIDZ1 over 4 x Seagate
(ST31500341AS) 1,5TB SATA drives.
 
If I compile or even just unpack a tar.gz archive with source code
(or any archive with lots of
small files), on my Linux client onto a NFS mounted disk to the
OpenSolaris server, it's extremely
slow compared to unpacking this archive on the locally on the
server. A 22MB .tar.gz file
containng 7360 files takes 9 minutes and 12seconds to unpack over NFS.
 
Unpacking the same file locally on the server is just under 2
seconds. Between the server and
client I have a gigabit network, which at the time of testing had
no other significant load. My
NFS mount options are: rw,hard,intr,nfsvers=3,tcp,sec=sys.
 
Any suggestions to why this is?
 
 
Regards,
Sigbjorn
 
 
 as someone else said, adding an ssd log device can help hugely.  I saw about 
 a 500% nfs write increase by doing this.
 I've heard of people getting even more.
 
 Another option if you don't care quite so much about data security in the 
 event of an unexpected system outage would be to use Robert Milkowski and 
 Neil Perrin's zil synchronicity [PSARC/2010/108] changes with sync=disabled, 
 when the changes work their way into an available build. The risk is that if 
 the file server goes down unexpectedly, it might come back up having lost 
 some seconds worth of changes which it told the client (lied) that it had 
 committed to disk, when it hadn't, and this violates the NFS protocol. That 
 might be OK if you are using it to hold source that's being built, where you 
 can kick off a build again if the server did go down in the middle of it. 
 Wouldn't be a good idea for some other applications though (although Linux 
 ran this way for many years, seemingly without many complaints). Note that 
 there's no increased risk of the zpool going bad - it's just that after the 
 reboot, filesystems with sync=disabled will look like they were rewound by 
 some secon
 ds (possibly up to 30 seconds).

That's assuming you know it happened and that you need  to restart the build 
(ideally with a make clean). All the NFS client knows is that the NFS server 
went away for some time. It still assumes nothing was lost. I can imagine cases 
where the build might continue to completion but with partially corrupted 
files. It's unlikely, but conceivable. Of course, databases like dbm, MySQL or 
Oracle would go blithely on up the swanee with silent data corruption.

The fact that people run unsafe systems seemingly without complaint for years 
assumes that they know silent data corruption when they see^H^H^Hhear it ... 
which, of course, they didn't ... because it is silent ... or having 
encountered corrupted data, that they have the faintest idea where it came 
from. In my day to day work I still find many people that have been 
(apparently) very lucky.

Feel free to play fast and loose with your own data, but I won't with mine, 
thanks! ;)
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Sigbjorn Lie
I see I have already received several replies, thanks to all!

I would not like to risk losing any data, so I believe a ZIL device would be 
the way for me. I see
these exists in different prices. Any reason why I would not buy a cheap one? 
Like the Intel X25-V
SSD 40GB 2,5?

What size of ZIL device would be recommened for my pool consisting for 4 x 
1,5TB drives? Any
brands I should stay away from?



Regards,
Sigbjorn





On Fri, July 23, 2010 09:48, Phil Harman wrote:
 That's because NFS adds synchronous writes to the mix (e.g. the client needs 
 to know certain
 transactions made it to nonvolatile storage in case the server restarts etc). 
 The simplest safe
 solution, although not cheap, is to add an SSD log device to the pool.

 On 23 Jul 2010, at 08:11, Sigbjorn Lie sigbj...@nixtra.com wrote:


 Hi,


 I've been searching around on the Internet to fine some help with this, but 
 have been
 unsuccessfull so far.

 I have some performance issues with my file server. I have an OpenSolaris 
 server with a Pentium
 D
 3GHz CPU, 4GB of memory, and a RAIDZ1 over 4 x Seagate (ST31500341AS) 1,5TB 
 SATA drives.


 If I compile or even just unpack a tar.gz archive with source code (or any 
 archive with lots of
  small files), on my Linux client onto a NFS mounted disk to the OpenSolaris 
 server, it's
 extremely slow compared to unpacking this archive on the locally on the 
 server. A 22MB .tar.gz
 file containng 7360 files takes 9 minutes and 12seconds to unpack over NFS.

 Unpacking the same file locally on the server is just under 2 seconds. 
 Between the server and
 client I have a gigabit network, which at the time of testing had no other 
 significant load. My
 NFS mount options are: rw,hard,intr,nfsvers=3,tcp,sec=sys.


 Any suggestions to why this is?




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Sigbjorn Lie
On Fri, July 23, 2010 10:42, tomwaters wrote:
 I agree, I get apalling NFS speeds compared to CIFS/Samba..ie. CIFS/Samba of 
 95-105MB and NFS of
 5-20MB.


 Not the thread hijack, but I assume a SSD ZIL will similarly improve an iSCSI 
 target...as I am
 getting 2-5MB on that too. --
 This message posted from opensolaris.org


This is exactly the numbers I'm getting as well.

What's the reason for such low rate when using iSCSI?




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Phil Harman


Sent from my iPhone

On 23 Jul 2010, at 09:42, tomwaters tomwat...@chadmail.com wrote:

 I agree, I get apalling NFS speeds compared to CIFS/Samba..ie. CIFS/Samba of 
 95-105MB and NFS of 5-20MB.
 
 Not the thread hijack, but I assume a SSD ZIL will similarly improve an iSCSI 
 target...as I am getting 2-5MB on that too.

Yes, it generally will. I've seen some huge improvements with iSCSI, but YMMV 
depending on your config, application and workload.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Phil Harman

On 23/07/2010 10:02, Sigbjorn Lie wrote:

On Fri, July 23, 2010 10:42, tomwaters wrote:
   

I agree, I get apalling NFS speeds compared to CIFS/Samba..ie. CIFS/Samba of 
95-105MB and NFS of
5-20MB.


Not the thread hijack, but I assume a SSD ZIL will similarly improve an iSCSI 
target...as I am
getting 2-5MB on that too. --
This message posted from opensolaris.org
 

This is exactly the numbers I'm getting as well.

What's the reason for such low rate when using iSCSI?
   


The filesystem or application using the iSCSI target may be requesting 
regular cache flushes. These will require synchronous writes to disk. An 
SSD doesn't remove the sync writes, it just makes them a lot faster. 
Other sensible storage servers typically use NVRAM caches to solve this 
problem. Others just play fast and loose with your data.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Thomas Burgess
On Fri, Jul 23, 2010 at 5:00 AM, Sigbjorn Lie sigbj...@nixtra.com wrote:

 I see I have already received several replies, thanks to all!

 I would not like to risk losing any data, so I believe a ZIL device would
 be the way for me. I see
 these exists in different prices. Any reason why I would not buy a cheap
 one? Like the Intel X25-V
 SSD 40GB 2,5?

 What size of ZIL device would be recommened for my pool consisting for 4 x
 1,5TB drives? Any
 brands I should stay away from?



 Regards,
 Sigbjorn

 Like i said, i bought a 50 gb OCZ Vertex Limited Edition...it's like 200
dollars, up to 15,000 random iops (iops is what you want for fast zil)


I've gotten excelent performance out of it.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] physically removed a pool - how to I tell it to forget the pool?

2010-07-23 Thread tomwaters
Hi guys, I physically removed disks from a pool without offlining the pool 
first...(yes I know) anyway I now want to delete/destroy the pool...but zpool 
destroy -f  dvr says can not open 'dvr':no such pool

I can not offline it or delete it!

I want to reuse the name dvr but how do I do this?

ie. how do I totally delete a pool who's disks have been physically removed?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] physically removed a pool - how to I tell it to forget the pool?

2010-07-23 Thread Francois Napoleoni

If all disks were actually removed, renaming /etc/zfs/zpool.cache and rebooting 
should do the trick.

I am not sure but you may have to import root pool at next reboot.

F.



tomwaters wrote:

Hi guys, I physically removed disks from a pool without offlining the pool 
first...(yes I know) anyway I now want to delete/destroy the pool...but zpool 
destroy -f  dvr says can not open 'dvr':no such pool

I can not offline it or delete it!

I want to reuse the name dvr but how do I do this?

ie. how do I totally delete a pool who's disks have been physically removed?


--
Oracle Global Customer Services
Francois Napoleoni / TSC Engineer Solaris  Networking, Global Systems Support
Email : francois.napole...@oracle.com
Phone : +33 (0)1 34 03 17 07
Fax   : +33 (0)1 34 03 11 14
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Increase resilver priority

2010-07-23 Thread Giovanni Tirloni
Hello,

 We've seen some resilvers on idle servers that are taking ages. Is it
possible to speed up resilver operations somehow?

 Eg. iostat shows 5MB/s writes on the replaced disks.

 I'm thinking a small performance degradation would be sometimes
better than the increased risk window (where a vdev is degraded).

Thank you,

-- 
Giovanni Tirloni
gtirl...@sysdroid.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Sigbjorn Lie

On Fri, July 23, 2010 11:21, Thomas Burgess wrote:
 On Fri, Jul 23, 2010 at 5:00 AM, Sigbjorn Lie sigbj...@nixtra.com wrote:


 I see I have already received several replies, thanks to all!


 I would not like to risk losing any data, so I believe a ZIL device would
 be the way for me. I see these exists in different prices. Any reason why I 
 would not buy a cheap
  one? Like the Intel X25-V SSD 40GB 2,5?


 What size of ZIL device would be recommened for my pool consisting for 4 x
 1,5TB drives? Any
 brands I should stay away from?



 Regards,
 Sigbjorn


 Like i said, i bought a 50 gb OCZ Vertex Limited Edition...it's like 200

 dollars, up to 15,000 random iops (iops is what you want for fast zil)


 I've gotten excelent performance out of it.



The X25-V has up to 25k random read iops and up to 2.5k random write iops per 
second, so that
would seem okay for approx $80. :)

What about mirroring? Do I need mirrored ZIL devices in case of a power outage?


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Darren J Moffat

On 23/07/2010 10:53, Sigbjorn Lie wrote:

The X25-V has up to 25k random read iops and up to 2.5k random write iops per 
second, so that
would seem okay for approx $80. :)

What about mirroring? Do I need mirrored ZIL devices in case of a power outage?


Note there is not a ZIL device, there is a slog device.  Every pool has 
one or more ZIL it may or may not have a slog device used to hold ZIL 
contents, wither a ZIL is on the slog or not depends on a lot of factors 
including the logbias property.


You don't need to mirror the slog device to protect against a power 
outage. You need to mirror the slog if you want to protect against 
loosing synchronous writes (but not pool consistency on disk) on power 
outage *and* failure of your slog device at the same time (ie a double 
fault).


--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] physically removed a pool - how to I tell it to forget the pool?

2010-07-23 Thread tomwaters
It was fine on the reboot...so even though zfs destroy threw up the errors, it 
did remove them...just needed a reboot to refresh/remove
 it in the zpool list.

thanks.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs raidz1 and traditional raid 5 perfomrance comparision

2010-07-23 Thread Edward Ned Harvey
 From: Robert Milkowski [mailto:mi...@task.gda.pl]

 [In raidz] The issue is that each zfs filesystem block is basically 
 spread across
 n-1 devices.
 So every time you want to read back a single fs block you need to wait
 for all n-1 devices to provide you with a part of it - and keep in mind
 in zfs you can't get a partial block even if that's what you are asking
 for as zfs has to check checksum of entire fs block.

Can anyone else confirm or deny the correctness of this statement?

If you read a small file from a raidz volume, do you have to wait for every
single disk to return a small chunk of the blocksize?  I know this is true
for large files which require more than one block, obviously, but even a
small file gets spread out across multiple disks?

This may be the way it's currently implemented, but it's not a mathematical
requirement.  It is possible, if desired, to implement raid parity and still
allow small files to be written entirely on a single disk, without losing
redundancy.  Thus providing the redundancy, the large file performance,
(both of which are already present in raidz), and also optimizing small file
random operations, which may not already be optimized in raidz.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Phil Harman

 Milkowski and Neil Perrin's zil synchronicity [PSARC/2010/108] changes
 with sync=disabled, when the changes work their way into an available
 
 The fact that people run unsafe systems seemingly without complaint for
 years assumes that they know silent data corruption when they
 see^H^H^Hhear it ... which, of course, they didn't ... because it is
 silent ... or having encountered corrupted data, that they have the
 faintest idea where it came from. In my day to day work I still find
 many people that have been (apparently) very lucky.

Running with sync disabled, or ZIL disabled, you could call unsafe if you
want to use a generalization and a stereotype.  

Just like people say writeback is unsafe.  If you apply a little more
intelligence, you'll know, it's safe in some conditions, and not in other
conditions.  Like ... If you have a BBU, you can use your writeback safely.
And if you're not sharing stuff across the network, you're guaranteed the
disabled ZIL is safe.  But even when you are sharing stuff across the
network, the disabled ZIL can still be safe under the following conditions:

If you are only doing file sharing (NFS, CIFS) and you are willing to
reboot/remount from all your clients after an ungraceful shutdown of your
server, then it's safe to run with ZIL disabled.

If you're unsure, then adding SSD nonvolatile log device, as people have
said, is the way to go.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Sigbjorn Lie
 
 What size of ZIL device would be recommened for my pool consisting for

Get the smallest one.  Even an unrealistic high performance scenario cannot
come close to using 32G.  I am sure you'll never reach even 4G usage.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Sigbjorn Lie
 
 What about mirroring? Do I need mirrored ZIL devices in case of a power
 outage?

You don't need mirroring for the sake of *power outage* but you *do* need
mirroring for the sake of preventing data loss when one of the SSD devices
fails.  There is some gray area here:

If you have zpool  19, then you do not have log device removal which
means you lose your whole zpool in the event of a failed unmirrored log
device.  (Techniques exist to recover, but it's not always easy.)

If you have zpool = 19, then the danger is much smaller.  If you have a
failed unmirrored log device, and the failure is detected, then the log
device is simply marked failed and the system slows down, and everything
is fine.  But if you have an undetected failure, *and* an ungraceful reboot
(which is more likely than it seems) then you risk up to 30 sec of data that
was intended to be written, immediately before the crash.

None of that is a concern, if you have a mirrored log device.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Andrew Gabriel

Edward Ned Harvey wrote:

From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
boun...@opensolaris.org] On Behalf Of Phil Harman

Milkowski and Neil Perrin's zil synchronicity [PSARC/2010/108] changes
with sync=disabled, when the changes work their way into an available

The fact that people run unsafe systems seemingly without complaint for
years assumes that they know silent data corruption when they
see^H^H^Hhear it ... which, of course, they didn't ... because it is
silent ... or having encountered corrupted data, that they have the
faintest idea where it came from. In my day to day work I still find
many people that have been (apparently) very lucky.



Running with sync disabled, or ZIL disabled, you could call unsafe if you
want to use a generalization and a stereotype.  


Just like people say writeback is unsafe.  If you apply a little more
intelligence, you'll know, it's safe in some conditions, and not in other
conditions.  Like ... If you have a BBU, you can use your writeback safely.
And if you're not sharing stuff across the network, you're guaranteed the
disabled ZIL is safe.  But even when you are sharing stuff across the
network, the disabled ZIL can still be safe under the following conditions:

If you are only doing file sharing (NFS, CIFS) and you are willing to
reboot/remount from all your clients after an ungraceful shutdown of your
server, then it's safe to run with ZIL disabled.
  


No, that's not safe. The client can still lose up to 30 seconds of data, 
which could be, for example, an email message which is received and 
foldered on the server, and is then lost. It's probably /*safe enough*/ 
for most home users, but you should be fully aware of the potential 
implications before embarking on this route.


(As I said before, the zpool itself is not at any additional risk of 
corruption, it's just that you might find the zfs filesystems with 
sync=disabled appear to have been rewound by up to 30 seconds.)



If you're unsure, then adding SSD nonvolatile log device, as people have
said, is the way to go.
  


--
Andrew Gabriel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs raidz1 and traditional raid 5 perfomrance comparision

2010-07-23 Thread Arne Jansen

Edward Ned Harvey wrote:

From: Robert Milkowski [mailto:mi...@task.gda.pl]

[In raidz] The issue is that each zfs filesystem block is basically 
spread across

n-1 devices.
So every time you want to read back a single fs block you need to wait
for all n-1 devices to provide you with a part of it - and keep in mind
in zfs you can't get a partial block even if that's what you are asking
for as zfs has to check checksum of entire fs block.


Can anyone else confirm or deny the correctness of this statement?

If you read a small file from a raidz volume, do you have to wait for every
single disk to return a small chunk of the blocksize?  I know this is true
for large files which require more than one block, obviously, but even a
small file gets spread out across multiple disks?

This may be the way it's currently implemented, but it's not a mathematical
requirement.  It is possible, if desired, to implement raid parity and still
allow small files to be written entirely on a single disk, without losing
redundancy.  Thus providing the redundancy, the large file performance,
(both of which are already present in raidz), and also optimizing small file
random operations, which may not already be optimized in raidz.


As I understand it that's the whole point of raidz. Each block is its own
stripe. If necessary the block gets broken down into 512 byte chunks to spread
it as wide as possible. Each block gets its own parity added. So if the array
is too wide for the block to be spread to all disks, you also lose space because
the stripe is not full and parity gets added to that small stripe. That means
if you only write 512 byte blocks, each write writes 3 blocks to disk, so the
net capacity goes down to one third, regardless how many disks you have in your
raid group.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs add -r output

2010-07-23 Thread Cindy Swearingen

Hi Ryan,

You are seeing this CR:

http://bugs.opensolaris.org/view_bug.do?bug_id=6916574

zpool add -n displays incorrect structure

This is a display problem only.

Thanks,

Cindy

On 07/22/10 15:54, Ryan Schwartz wrote:

I've got a system running s10x_u7wos_08 with only half of the disks 
provisioned. When performing a dry run of a zpool add, I'm seeing some strange 
output:

root# zpool add -n vst raidz2 c0t2d0 c5t1d0 c4t1d0 c4t5d0 c7t1d0 c7t5d0 c6t1d0 
c6t5d0 c1t1d0 c1t5d0 c0t1d0 raidz2 c0t5d0 c0t5d0 c4t4d0 c7t0d0 c7t4d0 c6t0d0 
c6t4d0 c1t0d0 c1t4d0 c0t0d0 c0t4d0
would update 'vst' to the following configuration:
vst
  raidz2
c5t3d0
c5t7d0
c4t3d0
c4t7d0
c7t3d0
c7t7d0
c6t3d0
c6t7d0
c1t3d0
c1t7d0
c0t3d0
  raidz2
c0t7d0
c5t2d0
c5t6d0
c4t2d0
c4t6d0
c7t2d0
c7t6d0
c6t2d0
c6t6d0
c1t2d0
c1t6d0
  raidz2
c0t2
c5t1
c4t1
c4t5
c7t1
c7t5
c6t1
c6t5
c1t1
c1t5
c0t1
  raidz2
c0t5
c0t5
c4t4
c7t0
c7t4
c6t0
c6t4
c1t0
c1t4
c0t0
c0t4

I would expect the new configuration would use the full shorthand device names, 
including the d0. Is this normal/expected output for a system on s10x_u7wos_08?

Thanks,

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Linder, Doug
Phil Harmon wrote:

  Not the thread hijack, but I assume a SSD ZIL will similarly improve
 an iSCSI target...as I am getting 2-5MB on that too.
 
 Yes, it generally will. I've seen some huge improvements with iSCSI,
 but YMMV depending on your config, application and workload.


Sorry this isn't completely ZFS-related, but with all this expert storage 
knowledge here...

On a related note - all other things being equal, is there any reason to choose 
NFS over ISCI, or vice-versa?  I'm currently looking at this decision.  We have 
a NetApp (I wish it were a ZFS-based appliance!) and need to remotely mount a 
filesystem from it.  It will share the filesystem either as NFS or ICSI.  Some 
of my colleagues say it would be better to use NFS.  Their reasoning is 
basically: That's the way it's always been done.  I'm leaning towards ISCSI.  
My reasoning is that it removes a whole extra layer of complexity - as I 
understand it, the remote client just treats the remote mount like any other 
physical device.  And I've had MAJOR headaches over the years fixing/tweaking 
NFS.  Even though version 4 seems better, I'd still rather bypass it 
completely.  I believe in the Keep it simple, stupid philosophy.

I do realize that NFS is probably better for remote filesystems that have 
multiple simultaneous users, but we won't be doing that in this case.

Any major arguments for/against one over the other?

Thanks for any suggestions.

Doug Linder
--
Learn more about Merchant Link at www.merchantlink.com.

THIS MESSAGE IS CONFIDENTIAL.  This e-mail message and any attachments are 
proprietary and confidential information intended only for the use of the 
recipient(s) named above.  If you are not the intended recipient, you may not 
print, distribute, or copy this message or any attachments.  If you have 
received this communication in error, please notify the sender by return e-mail 
and delete this message and any attachments from your computer.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Increase resilver priority

2010-07-23 Thread Richard Elling
On Jul 23, 2010, at 2:31 AM, Giovanni Tirloni wrote:

 Hello,
 
 We've seen some resilvers on idle servers that are taking ages. Is it
 possible to speed up resilver operations somehow?
 
 Eg. iostat shows 5MB/s writes on the replaced disks.

This is lower than I expect, but It may be IOPS bound. What does
iostat say about the IOPS and asvc_t ?
 -- richard

-- 
Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Increase resilver priority

2010-07-23 Thread Bill Sommerfeld

On 07/23/10 02:31, Giovanni Tirloni wrote:

  We've seen some resilvers on idle servers that are taking ages. Is it
possible to speed up resilver operations somehow?

  Eg. iostat shows5MB/s writes on the replaced disks.


What build of opensolaris are you running?  There were some recent 
improvements (notably the addition of prefetch to the pool traverse used 
by scrub and resilver) which sped this up significantly for my systems.


Also: if there are large numbers of snapshots, pools seem to take longer 
to resilver, particularly when there's a lot of metadata divergence 
between snapshots.  Turning off atime updates (if you and your 
applications can cope with this) may also help going forward.


- Bill


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Increase resilver priority

2010-07-23 Thread Giovanni Tirloni
On Fri, Jul 23, 2010 at 11:59 AM, Richard Elling rich...@nexenta.com wrote:
 On Jul 23, 2010, at 2:31 AM, Giovanni Tirloni wrote:

 Hello,

 We've seen some resilvers on idle servers that are taking ages. Is it
 possible to speed up resilver operations somehow?

 Eg. iostat shows 5MB/s writes on the replaced disks.

 This is lower than I expect, but It may be IOPS bound. What does
 iostat say about the IOPS and asvc_t ?
  -- richard

It seems to have improved a bit.

 scrub: resilver in progress for 7h19m, 75.37% done, 2h23m to go
config:

NAME  STATE READ WRITE CKSUM
storage   DEGRADED 0 1 0
  mirror  DEGRADED 0 0 0
c3t2d0ONLINE   0 0 0
replacing DEGRADED 1.29M 0 0
  c3t3d0s0/o  FAULTED  0 0 0  corrupted data
  c3t3d0  DEGRADED 0 0 1.29M  too many errors
  mirror  ONLINE   0 0 0
c3t4d0ONLINE   0 0 0
c3t5d0ONLINE   0 0 0
  mirror  DEGRADED 0 0 0
c3t6d0ONLINE   0 0 0
c3t7d0REMOVED  0 0 0
  mirror  ONLINE   0 0 0
c3t8d0ONLINE   0 0 0
c3t9d0ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c3t10d0   ONLINE   0 0 0
c3t11d0   ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c3t12d0   ONLINE   0 0 0
c3t13d0   ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c3t14d0   ONLINE   0 0 0
c3t15d0   ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c3t16d0   ONLINE   0 0 0
c3t17d0   ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c3t18d0   ONLINE   0 0 0
c3t19d0   ONLINE   0 0 0
  mirror  ONLINE   0 0 0
c3t20d0   ONLINE   0 0 0
c3t21d0   ONLINE   0 0 0
logs  DEGRADED 0 1 0
  mirror  ONLINE   0 0 0
c3t1d0ONLINE   0 0 0
c3t22d0   ONLINE   0 0 0


extended device statistics
r/sw/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
  582.2  864.3 68925.2 37511.2  0.0 52.70.0   36.4   0 610 c3
0.0  201.70.0  806.7  0.0  0.00.00.1   0   2 c3t0d0
0.0  268.20.0 10531.2  0.0  0.10.00.4   0  10 c3t1d0
  144.10.0 18375.90.0  0.0  9.50.0   65.7   0 100 c3t2d0
   79.5  125.2 10109.9 15634.3  0.0 35.00.0  171.0   0 100 c3t3d0
   10.90.0 1181.30.0  0.0  0.10.0   13.3   0  10 c3t4d0
   19.90.0 2120.60.0  0.0  0.30.0   15.6   0  19 c3t5d0
   35.80.0 3819.50.0  0.0  0.60.0   18.1   0  28 c3t6d0
0.00.00.00.0  0.0  0.00.00.0   0   0 c3t7d0
   22.90.0 2506.60.0  0.0  0.50.0   22.0   0  22 c3t8d0
   15.90.0 1639.80.0  0.0  0.30.0   20.5   0  15 c3t9d0
   23.80.0 2889.60.0  0.0  0.50.0   19.8   0  27 c3t10d0
   21.90.0 2558.30.0  0.0  0.60.0   28.9   0  19 c3t11d0
   32.80.0 3151.90.0  0.0  1.20.0   37.4   0  25 c3t12d0
   25.80.0 2707.80.0  0.0  0.50.0   18.8   0  26 c3t13d0
   19.90.0 2281.10.0  0.0  0.30.0   17.5   0  24 c3t14d0
   23.80.0 2782.30.0  0.0  0.30.0   14.6   0  20 c3t15d0
   18.90.0 2249.80.0  0.0  0.40.0   19.7   0  23 c3t16d0
   21.90.0 2519.50.0  0.0  0.50.0   22.6   0  27 c3t17d0
   12.90.0 1653.20.0  0.0  0.20.0   16.8   0  18 c3t18d0
   26.80.0 3262.70.0  0.0  0.80.0   28.4   0  29 c3t19d0
9.90.0 1271.70.0  0.0  0.10.0   14.3   0  13 c3t20d0
   14.90.0 1843.90.0  0.0  0.30.0   20.5   0  19 c3t21d0
0.0  269.20.0 10539.0  0.0  0.40.01.3   0  33 c3t22d0
extended device statistics
r/sw/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
  405.1  745.2 51057.2 29893.8  0.0 53.30.0   46.4   0 457 c3
0.0  252.10.0 1008.3  0.0  0.00.00.1   0   2 c3t0d0
0.0  177.10.0 5485.8  0.0  0.00.00.3   0   4 c3t1d0
  145.00.0 18438.00.0  0.0 15.10.0  104.0   0 100 c3t2d0
   80.0  140.0 10147.8 17925.9  0.0 35.00.0  159.0   0 100 c3t3d0
8.00.0 1024.30.0  0.0  0.20.0   19.9   0  12 c3t4d0
7.00.0  768.30.0  0.0  0.10.0   15.5   0  11 c3t5d0
   24.00.0 3009.00.0  0.0  0.50.0   

Re: [zfs-discuss] Increase resilver priority

2010-07-23 Thread Giovanni Tirloni
On Fri, Jul 23, 2010 at 12:50 PM, Bill Sommerfeld
bill.sommerf...@oracle.com wrote:
 On 07/23/10 02:31, Giovanni Tirloni wrote:

  We've seen some resilvers on idle servers that are taking ages. Is it
 possible to speed up resilver operations somehow?

  Eg. iostat shows5MB/s writes on the replaced disks.

 What build of opensolaris are you running?  There were some recent
 improvements (notably the addition of prefetch to the pool traverse used by
 scrub and resilver) which sped this up significantly for my systems.

b111. Thanks for the heads up regarding these improvements, I'll try
that in b134.

 Also: if there are large numbers of snapshots, pools seem to take longer to
 resilver, particularly when there's a lot of metadata divergence between
 snapshots.  Turning off atime updates (if you and your applications can cope
 with this) may also help going forward.

There are 7 snapshots and atime is disabled.

-- 
Giovanni Tirloni
gtirl...@sysdroid.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Random system hang build 134

2010-07-23 Thread Som Pathak
Symptoms:

1. System remains pingable
2. When trying to ssh in terminal hangs after entering pass
3. At the console terminal hangs after entering pass
4. Problem persists after disabling snapshots/compression/dedup

Solution:

Hard reboot (A+F1 does not work)

Configuration:

Supermicro Mobo
24 x 2TB WD Black Drives (2 x 12 drive in RAIDZ2)
Mirrored OCZ 32GB SSD drives for OS (ZFS mirrored rpool)
Mirrored Intel extreme 64GB drives (ZIL)
Quad Port Intel Gb NIC - aggr1
12GB DDR3


Does anyone have any idea what the problem could be or how I can track this 
down? Just so that you know I have 2 of these system with identical setup and 
hardware so I have excluded the possibility that this could be a bad piece of 
hardware.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS volume turned into a socket - any way to restore data?

2010-07-23 Thread Richard Elling
On Jul 23, 2010, at 9:10 AM, Ruslan Sivak wrote:

 I have recently upgraded from NexentaStor 2 to NexentaStor 3 and somehow one 
 of my volumes got corrupted.  Its showing up as a socket.  Has anyone seen 
 this before?  Is there a way to get my data back?  It seems like it's still 
 there, but not recognized as a folder.  I ran zpool scrub, but it came back 
 clean.  

very strange indeed.

 
 Attached is the output of #zdb data/rt
 
 2.0K sr-xr-xr-x 17 root root 17 Jul 21 17:43 rt
 
 # stat rt
  File: `rt'
  Size: 17  Blocks: 4  IO Block: 1536   socket
 Device: 2d90015h/47775765d  Inode: 3   Links: 17
 Access: (0555/sr-xr-xr-x)  Uid: (0/root)   Gid: (0/root)
 Access: 2010-07-21 13:48:09.485116161 -0400
 Modify: 2010-07-21 17:43:13.019153434 -0400
 Change: 2010-07-22 15:50:06.222143953 -0400
 
 # cd rt
 bash: cd: rt: Not a directory
 
 # cat rt
 cat: rt: Operation not supported on transport endpoint
 
 a# zfs list
 NAME USED  AVAIL  REFER  MOUNTPOINT
 data 992G   828G  47.8K  /volumes/data
 ...
 data/rt  800G   828G   800G  /volumes/data/rt
 data/rt20   828G   800G  /volumes/data/rt2
 syspool 6.31G  23.0G  35.5K  legacy
 syspool/dump2.79G  23.0G  2.79G  -
 syspool/rootfs-nmu-000   115M  23.0G  1.32G  legacy
 syspool/rootfs-nmu-001  1.59G  23.0G  1.17G  legacy
 syspool/rootfs-nmu-002   793M  23.0G  1.17G  legacy
 syspool/swap1.03G  24.0G16K  -
 
 # zfs umount data/rt
 #ls -lahs
 ...
 2.0K drwxr-xr-x  2 root root  2 Jul 22 16:48 rt
 2.0K srwxr-xr-x 17 root root 17 Jul 21 17:43 rt2

If at this point data/rt is unmounted, then:
mv rt2 rt2-bad
zfs mount data/rt


 -- richard

 ...
 -- 
 This message posted from 
 opensolaris.orgzdb.zip___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

-- 
Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Confused about consumer drives and zfs can someone help?

2010-07-23 Thread JavaWebDev

 On 7/23/2010 3:39 AM, tomwaters wrote:

There is alot there to reply to...but I will try and help...

Re. TLER. Do not worry about TLER when using ZFS. ZFS will handle it either way and will 
NOT time out and drop the drive...it may wait a long time, but it will not time out and 
drop the drive - nor will it have an issue if you do enable TLER-ON (which sets time out 
to 7 seonds).  I run both with TLER-ON (disks from an old mdadm raid array) 
and without (TLER-OFF) 1.5TB WDEADS.


That's reassuring. I guess the only time the issue will come up is when 
the drive starts to develop errors.



I can not speak for the WD EARS, but the WDEADS are fine in my home nas. I also 
run 1.5TB Samsung Green/Silencer series and Seagate 11's. Others swear by 
Hitachi. I would recommend the Samsung or Hitachi and not the new WD EARS which 
have that 4k sectors or whatever it is.


I've been digging through the archives more and I'm starting to lean 
towards laptop drives. I found this discussion 
http://opensolaris.org/jive/message.jspa?messageID=468269#468269 and 
contacted someone else who has used 3 WD Scorpio Black drives in a 3 way 
mirror for a year without any problems.


This is mostly going to be a backup server so I'm not too worried about 
performance.



Re the CPU, do not go low power Atom etc, go a newish Core2 duo...the power 
differential at idle is bugger all and when you want to use the nas, ZFS will 
make good use of the CPU. Honestly, sit down and do the calculations on power 
savings of a low power cpu and you'll see it's better to just not have that 5th 
beer on a friday - you'll save more money that way and be MUCH happier with 
your nas performance.

re. cards...I use and recommend these 8-Port SUPERMICRO AOC-USASLP-L8I UIO SAS. 
They are cheap on e-bay, just work and are fast. Use them.

You do want alot of ram I use 8GB, but you can use 4. Ram is cheap, ZFS loves 
ram, just buy 8.

IMHO (and that of the best practice guide) - you should mirror the rpool (o/s disk). Just 
buy 2 cheap laptop drives and when installing choose to mirror them.


I can find some motherboards with 4-6 onboard sata ports. If I go with 2 
USB flash drives for a mirrored rpool do you think that would be ok? 
Performance seems to be about the same as 5400 rpm laptop drives.



I hope that helps.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] snapshot .zfs folder can only be seen in the top of a file system?

2010-07-23 Thread Fred Liu
Hi,

Is it true? Any way to find it in every hierarchy?

Thanks.

Fred
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs raidz1 and traditional raid 5 perfomrance comparision

2010-07-23 Thread Edward Ned Harvey
 From: Arne Jansen [mailto:sensi...@gmx.net]
 
  Can anyone else confirm or deny the correctness of this statement?
 
 As I understand it that's the whole point of raidz. Each block is its
 own
 stripe. 

Nope, that doesn't count for confirmation.  It is at least theoretically
possible to implement raidz using techniques that would (a) unintelligently
stripe all blocks (even small ones) across multiple disks, thus hurting
performance on small operations, or (b) implement raidz such that striping
of blocks behaves differently for small operations (plus parity).  So the
confirmation I'm looking for would be somebody who knows the actual source
code, and the actual architecture that was chosen to implement raidz in this
case.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Linder, Doug
 
 On a related note - all other things being equal, is there any reason
 to choose NFS over ISCI, or vice-versa?  I'm currently looking at this

iscsi and NFS are completely different technologies.  If you use iscsi, then
all the initiators (clients) are the things which format and control the
filesystem.  So the limitations of the filesystem are determined by
whichever clustering filesystem you've chosen to implement.  It probably
won't do snapshots and so forth.  Although the ZFS filesystem could make a
snapshot, it wouldn't be automatically mounted or made available without the
clients doing explicit mounts...

With NFS, the filesystem is formatted and controlled by the server.  Both
WAFL and ZFS do some pretty good things with snapshotting, and making
snapshots available to users without any effort.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snapshot .zfs folder can only be seen in the top of a file system?

2010-07-23 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Fred Liu
 
 Is it true? Any way to find it in every hierarchy?

Yup.  Nope.

If you use ZFS, you make a filesystem at whatever level you need it, in
order for the .zfs directory to be available to whatever clients will be
needing it...

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snapshot .zfs folder can only be seen in the top of a file system?

2010-07-23 Thread Fred Liu
Thanks.
But too many file systems may be an issue for management and also normal user 
cannot create file system.
I think it should go like what NetApp's snapshot does.
It is a pity.

Thanks.

Fred

 -Original Message-
 From: Edward Ned Harvey [mailto:sh...@nedharvey.com]
 Sent: 星期六, 七月 24, 2010 10:22
 To: Fred Liu; zfs-discuss@opensolaris.org
 Subject: RE: [zfs-discuss] snapshot .zfs folder can only be seen in the
 top of a file system?
 
  From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
  boun...@opensolaris.org] On Behalf Of Fred Liu
 
  Is it true? Any way to find it in every hierarchy?
 
 Yup.  Nope.
 
 If you use ZFS, you make a filesystem at whatever level you need it, in
 order for the .zfs directory to be available to whatever clients will
 be
 needing it...
 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] NFS performance?

2010-07-23 Thread Garrett D'Amore
Fundamentally, my recommendation is to choose NFS if your clients can
use it.  You'll get a lot of potential advantages in the NFS/zfs
integration, so better performance.  Plus you can serve multiple
clients, etc.

The only reason to use iSCSI is when you don't have a choice, IMO.  You
should only use iSCSI with a single initiator at any point in time
unless you have some higher level contention management in place.

- Garrett


On Fri, 2010-07-23 at 22:20 -0400, Edward Ned Harvey wrote:
  From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
  boun...@opensolaris.org] On Behalf Of Linder, Doug
  
  On a related note - all other things being equal, is there any reason
  to choose NFS over ISCI, or vice-versa?  I'm currently looking at this
 
 iscsi and NFS are completely different technologies.  If you use iscsi, then
 all the initiators (clients) are the things which format and control the
 filesystem.  So the limitations of the filesystem are determined by
 whichever clustering filesystem you've chosen to implement.  It probably
 won't do snapshots and so forth.  Although the ZFS filesystem could make a
 snapshot, it wouldn't be automatically mounted or made available without the
 clients doing explicit mounts...
 
 With NFS, the filesystem is formatted and controlled by the server.  Both
 WAFL and ZFS do some pretty good things with snapshotting, and making
 snapshots available to users without any effort.
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
 


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snapshot .zfs folder can only be seen in the top of a file system?

2010-07-23 Thread Richard Elling


On Jul 23, 2010, at 7:33 PM, Fred Liu fred_...@issi.com wrote:

 Thanks.
 But too many file systems may be an issue for management and also normal user 
 cannot create file system.

The ability to create or snapshot a file system can easily be delegated to a 
user.

 I think it should go like what NetApp's snapshot does.

There was a long thread on this topic earlier this year. Please see the
archives for details.

 It is a pity.

Disagree. The benefits are not as great as advertised.
  -- Richard

 
 Thanks.
 
 Fred
 
 -Original Message-
 From: Edward Ned Harvey [mailto:sh...@nedharvey.com]
 Sent: 星期六, 七月 24, 2010 10:22
 To: Fred Liu; zfs-discuss@opensolaris.org
 Subject: RE: [zfs-discuss] snapshot .zfs folder can only be seen in the
 top of a file system?
 
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Fred Liu
 
 Is it true? Any way to find it in every hierarchy?
 
 Yup.  Nope.
 
 If you use ZFS, you make a filesystem at whatever level you need it, in
 order for the .zfs directory to be available to whatever clients will
 be
 needing it...
 
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snapshot .zfs folder can only be seen in the top of a file system?

2010-07-23 Thread Sam Fourman Jr.
  I think it should go like what NetApp's snapshot does.

 There was a long thread on this topic earlier this year. Please see the
 archives for details.

 Do you have the URL? I don't have a long subscription


I too do not have a long subscription, and I would be interested in
the subject line
so I can search and do further reading.

-- 

Sam Fourman Jr.
Fourman Networks
http://www.fourmannetworks.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss