Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-26 Thread Peter Bridge
You have not described your requirements (low-power ??, low-cost ??).
But I'll contribute some pointers anyway! :)

Well for a home NAS I'm looking at noise as a big factor.  Also for a 24x7 box, 
power consumption, that's why the northbridge is putting me off slightly.

So far the other solutions mentioned here, and what I've found on various blogs 
are going to be a lot more noisy and power hungry.  Although the cost also 
makes this option very attractive, I'd be happy to pay a bit more for something 
with more grunt so long as it's quiet and energy efficient. Although I don't 
really mind if the box is slow, with ZFS what I'm looking for is secure data 
storage.  But with only two SATA ports I don't see how this board will quite 
acheive that.  Unless I boot from USB and use 2 SATA and 2 IDE drives.

btw I'm seeing mixed reports if the D945GCLF2 is seen as 64 bit by solaris.

Anyway thanks all for the great feedback!
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-26 Thread mike
On Sun, Oct 26, 2008 at 12:47 AM, Peter Bridge [EMAIL PROTECTED] wrote:

 Well for a home NAS I'm looking at noise as a big factor.  Also for a 24x7 
 box, power consumption, that's why the northbridge is putting me off slightly.

That's why I built a full-sized tower using a Lian-Li case with noise
dampening on the sides, robber grommets and such in the drive bays,
etc..
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-26 Thread Peter Bridge
Same case and same idea :)  I have 2 dampered drives already installed from a 
previous project.  Another 2 I pulled out to install into a qnap 209.  Ideally 
I'd return the disks and replace the qnap with this new single ZFS NAS, 
although I'm quite fond of the qnap bt client, and can't use all four SATA with 
this board anyway...
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Verify files' checksums

2008-10-26 Thread Peter Schuller
 1) When I notice an error in a file that I've copied from a ZFS disk I
 want to know whether that error is also in the original file on my ZFS
 disk or if it's only in the copy.

This was already addressed but let me do so slightly differently: One
of the major points of ZFS checksumming is that, in the abscence of
software bugs or hardware memory corruption issues when the file is
read on the host, successfully reading a file is supposed to mean that
you got the correct version of the file (either from physical disk or
from cache, having previously been read from physical disk).

A scrub is still required if you want to make sure the file is okay
*ON DISK*, unless you can satisfy yourself that no relevant data is
cached somewhere (or unless someone can inform me of a way to nuke a
particular file and related resources from cache).

 Up to now I've been storing md5sums for all files, but keeping the
 files and their md5sums synchronized is a burden I could do without.

FWIW I wanted to mention here that if you care a lot about this, I'd
recommend something like par2[1] instead. It uses forward error
correction[2], allowing you to not only detect corruption, but also
correct it. You can choose your desired level of redundancy expressed
as a percentage of the file size.

[1] http://en.wikipedia.org/wiki/Parchive
[2] http://en.wikipedia.org/wiki/Forward_error_correction

-- 
/ Peter Schuller

PGP userID: 0xE9758B7D or 'Peter Schuller [EMAIL PROTECTED]'
Key retrieval: Send an E-Mail to [EMAIL PROTECTED]
E-Mail: [EMAIL PROTECTED] Web: http://www.scode.org



pgpTBfuHa8mpF.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Verify files' checksums

2008-10-26 Thread Chris Gerhard
A slight nit.  

Using cat(1) to read the file to /dev/null will not actually cause the data to 
be read thanks to the magic that is mmap().  If you use dd(1) to read the file 
then yes you will either get the data and thus know it's blocks match  their 
checksums or dd will give you an error if you have no redundancy.

--chris
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] recommendations on adding vdev to raidz zpool

2008-10-26 Thread Mattias Pantzare
On Sun, Oct 26, 2008 at 5:31 AM, Peter Baumgartner [EMAIL PROTECTED] wrote:
 I have a 7x150GB drive (+1 spare) raidz pool that I need to expand.
 There are 6 open drive bays, so I bought 6 300GB drives and went to
 add them as a raidz vdev to the existing zpool, but I didn't realize
 the raidz vdevs needed to have the same number of drives. (why is
 that?)

They do not have to have the same number of drivers, you can even mix
raidz and plain
disks. That is more a recommendation. Add -f to the command.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Boot from mirror

2008-10-26 Thread dick hoogendijk
Is it or isn't it possible to boot off two mirrored ZFS disks and if
yes, can this be done in the upcoming solaris 10 10/08 too?

-- 
Dick Hoogendijk -- PGP/GnuPG key: 01D2433D
++ http://nagual.nl/ + SunOS sxce snv99 ++
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Verify files' checksums

2008-10-26 Thread James Litchfield
A nit on the nit...

cat does not use mmap for files = 32K in size. For those files
it's a simple read() into a buffer and write() it out.

Jim
---
Chris Gerhard wrote:
 A slight nit.  

 Using cat(1) to read the file to /dev/null will not actually cause the data 
 to be read thanks to the magic that is mmap().  If you use dd(1) to read the 
 file then yes you will either get the data and thus know it's blocks match  
 their checksums or dd will give you an error if you have no redundancy.

 --chris
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] recommendations on adding vdev to raidz zpool

2008-10-26 Thread Peter Baumgartner
On Sun, Oct 26, 2008 at 4:02 AM, Mattias Pantzare [EMAIL PROTECTED] wrote:
 On Sun, Oct 26, 2008 at 5:31 AM, Peter Baumgartner [EMAIL PROTECTED] wrote:
 I have a 7x150GB drive (+1 spare) raidz pool that I need to expand.
 There are 6 open drive bays, so I bought 6 300GB drives and went to
 add them as a raidz vdev to the existing zpool, but I didn't realize
 the raidz vdevs needed to have the same number of drives. (why is
 that?)

 They do not have to have the same number of drivers, you can even mix
 raidz and plain
 disks. That is more a recommendation. Add -f to the command.

What is the risk of creating a pool consisting of two raidz vdevs that
don't have the same number of disks?
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-26 Thread Matt Harrison
On Sat, Oct 25, 2008 at 06:50:46PM -0700, Nigel Smith wrote:
 Hi Matt
 What chipset is your PCI network card?
 (obviously, it not Intel, but what is it?)
 Do you know which driver the card is using?

I believe it's some sort of Realtek (8139 probably). It's coming up as rtls0

 You say '..The system was fine for a couple of weeks..'.
 At that point did you change any software - do any updates or upgrades?
 For instance, did you upgrade to a new build of OpenSolaris?

No, since the original problem with the onboard NICs it hasn't been upgraded
or anything.

 If not, then I would guess it's some sort of hardware problem.
 Can you try different cables and a different switch - anything
 in the path between client  server is suspect.

Have tried different cables and switch ports, I will try a different switch
as soon as I can get some space on one of the others.

 A mismatch of Ethernet duplex settings can cause problems - are
 you sure this is Ok.

Not 100% sure, but I will check as best I can.

 To get an idea of how the network is running try this:
 
 On the Solaris box, do an Ethernet capture with 'snoop' to a file.
 http://docs.sun.com/app/docs/doc/819-2240/snoop-1m?a=view
 
  # snoop -d {device} -o {filename}
 
 .. then while capturing, try to play your video file through the network.
 Control-C to stop the capture.
 
 You can then use Ethereal or WireShark to analyze the capture file.
 On the 'Analyze' menu, select 'Expert Info'.
 This will look through all the packets and will report
 any warning or errors it sees.

It's coming up with a huge number of TCP Bad Checksum errors, a few
Previous Segment Lost and a few Fast retransmission.

Thanks

Matt


pgpodoJIT4jdS.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] recommendations on adding vdev to raidz zpool

2008-10-26 Thread Mattias Pantzare
On Sun, Oct 26, 2008 at 3:00 PM, Peter Baumgartner [EMAIL PROTECTED] wrote:
 On Sun, Oct 26, 2008 at 4:02 AM, Mattias Pantzare [EMAIL PROTECTED] wrote:
 On Sun, Oct 26, 2008 at 5:31 AM, Peter Baumgartner [EMAIL PROTECTED] wrote:
 I have a 7x150GB drive (+1 spare) raidz pool that I need to expand.
 There are 6 open drive bays, so I bought 6 300GB drives and went to
 add them as a raidz vdev to the existing zpool, but I didn't realize
 the raidz vdevs needed to have the same number of drives. (why is
 that?)

 They do not have to have the same number of drivers, you can even mix
 raidz and plain
 disks. That is more a recommendation. Add -f to the command.

 What is the risk of creating a pool consisting of two raidz vdevs that
 don't have the same number of disks?

Slightly different reliability and performance on different parts of
the pool. Nothing to worry about in your case.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] recommendations on adding vdev to raidz zpool

2008-10-26 Thread Peter Tribble
On Sun, Oct 26, 2008 at 2:00 PM, Peter Baumgartner [EMAIL PROTECTED] wrote:

 They do not have to have the same number of drivers, you can even mix
 raidz and plain
 disks. That is more a recommendation. Add -f to the command.

 What is the risk of creating a pool consisting of two raidz vdevs that
 don't have the same number of disks?

One risk is that you mistyped the command, when you actually meant
to specify a balanced configuration.

-- 
-Peter Tribble
http://www.petertribble.co.uk/ - http://ptribble.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Cannot remove slog device from zpool

2008-10-26 Thread Ethan Erchinger

Hello,

I've looked quickly through the archives and haven't found mention of 
this issue.  I'm running SXCE (snv_99), which I believe uses zfs version 
13.  I had an existing zpool:
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Cannot remove slog device from zpool

2008-10-26 Thread Ethan Erchinger
Sorry for the first incomplete send,  stupid Ctrl-Enter. :-)

Hello,

I've looked quickly through the archives and haven't found mention of 
this issue.  I'm running SXCE (snv_99), which uses zfs version 13.  I 
had an existing zpool:
--
[EMAIL PROTECTED] ~]$ zpool status -v data
  pool: data
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ WRITE CKSUM
data   ONLINE   0 0 0
  mirror   ONLINE   0 0 0
c4t1d0p0   ONLINE   0 0 0
c4t9d0p0   ONLINE   0 0 0
  ...
cache
  c4t15d0p0ONLINE   0 0 0

errors: No known data errors

--

The cache device (c4t15d0p0) is an Intel SSD.  To test zil, I removed 
the cache device, and added it as a log device:
--
[EMAIL PROTECTED] ~]$ pfexec zpool remove data c4t15d0p0
[EMAIL PROTECTED] ~]$ pfexec zpool add data log c4t15d0p0
[EMAIL PROTECTED] ~]$ zpool status -v data
  pool: data
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ WRITE CKSUM
data   ONLINE   0 0 0
  mirror   ONLINE   0 0 0
c4t1d0p0   ONLINE   0 0 0
c4t9d0p0   ONLINE   0 0 0
  ...
logs   ONLINE   0 0 0
  c4t15d0p0ONLINE   0 0 0

errors: No known data errors
--

The device is working fine.  I then said, that was fun, time to remove 
and add as cache device.  But that doesn't seem possible:
--
[EMAIL PROTECTED] ~]$ pfexec zpool remove data c4t15d0p0
cannot remove c4t15d0p0: only inactive hot spares or cache devices can 
be removed
--

I've also tried using detach, offline, each failing in other more 
obvious ways.  The manpage does say that those devices should be 
removable/replaceable.  At this point the only way to reclaim my SSD 
device is to destroy the zpool.

Just in-case you are wondering about versions:
--
[EMAIL PROTECTED] ~]$ zpool upgrade data
This system is currently running ZFS pool version 13.

Pool 'data' is already formatted using the current version.
[EMAIL PROTECTED] ~]$ uname -a
SunOS opensolaris 5.11 snv_99 i86pc i386 i86pc
--

Any ideas?

Thanks,
Ethan
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
I've done some more research, but would still greatly appreciate someone 
helping me understand this.

It seems that writes to only the home directory of the person logged in to the 
console suffers from degraded performance.  If I write to a subdirectory 
beneath my home, or to any other directory on the system, performance is great. 
 But if I have a session on the console, no matter where else I test from 
(Gnome or a remote shell), writes ONLY to my home suffer.  If I log out of the 
console and then SSH in from another system, writes to the home directory no 
longer suffer from degraded performance.

This has proven true on every OpenSolaris system I've tried--all of which are 
using ZFS.  So what is it about logging into the console that slows write 
performance to ONLY the top level home directory of the username on the same 
console?
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread Ethan Erchinger

William Bauer wrote:

I've done some more research, but would still greatly appreciate someone 
helping me understand this.

It seems that writes to only the home directory of the person logged in to the 
console suffers from degraded performance.  If I write to a subdirectory 
beneath my home, or to any other directory on the system, performance is great. 
 But if I have a session on the console, no matter where else I test from 
(Gnome or a remote shell), writes ONLY to my home suffer.  If I log out of the 
console and then SSH in from another system, writes to the home directory no 
longer suffer from degraded performance.

This has proven true on every OpenSolaris system I've tried--all of which are 
using ZFS.  So what is it about logging into the console that slows write 
performance to ONLY the top level home directory of the username on the same 
console


Maybe something to do with autofs?  What happens if you move your home dir?

Ethan
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-26 Thread Daniel Leaberry
I just built a homeserver that pulls 62 watts from the plug at idle for
~$700. I had some of the parts lying around but even if you bought
everything at frys you should be able to set yourself up for under 1K for
the next 3-5 years.

Seasonic 80 plus 300 watt power supply
Intel DP35DP motherboard (onboard nic is an intel, has 6 sata from ich9)
E5200 Wolfdale processor (5 watts at idle, 30 watts at load)
4 x 1TB Samsung EcoGreen 5400 rpm harddrives
4GB ram
80GB laptop harddrive I had lying around for the system disk
Ati PCI 9200 video card I had lying around (pulls 2-3 watts)
Cheapest mid tower case off newegg with free shipping

The samsung drives are great, they don't ever spindown. I had problems with
some seagates a year or so back where the drive firmware would spin them
down after 90 seconds of inactivity. The samsungs just chug along at 4-5
watts each.

I use raidz on the 4 1TB drives and set copies=2 on the system drive. The
real power savings is from the cpu. Any dual core 45nm intel chip idles at
unbelivably low wattage. Add the 5400 rpm drives and you've got a nice cool
running, quiet, low power system.

If I only needed 1 TB I'd probably use the atom board with the two ports and
a usb jump drive for the system disk, but I needed more space.

Daniel

On Sun, Oct 26, 2008 at 2:51 AM, Peter Bridge [EMAIL PROTECTED]wrote:

 Same case and same idea :)  I have 2 dampered drives already installed from
 a previous project.  Another 2 I pulled out to install into a qnap 209.
  Ideally I'd return the disks and replace the qnap with this new single ZFS
 NAS, although I'm quite fond of the qnap bt client, and can't use all four
 SATA with this board anyway...
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
If that were the case, why would it matter if I was logged into the console, 
and why would subdirectories of my home exhibit better write performance than 
the top level home directory?  A write to /export/home/username is slower than 
to /export/home/username/blah, but ONLY if that user is logged into the console.

I'm not saying it's not an odd autofs interaction, because I've seen things 
just as strange.  It just seems like the console login is the tripwire for this 
performance issue, and autofs would come to play at every login and not just a 
login to the console.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] recommendations on adding vdev to raidz zpool

2008-10-26 Thread Bob Friesenhahn
On Sun, 26 Oct 2008, Peter Baumgartner wrote:

 What is the risk of creating a pool consisting of two raidz vdevs that
 don't have the same number of disks?

Virtually no risk.  The only matter of concern would be if the vdevs 
have substantially different I/O performance and latencies since ZFS's 
load share mechanism will put more data on the vdev which is more 
responsive when the vdevs are under heavy write load.  However, if the 
faster vdev is built with drives of newer design and with higher 
capacity, this behavior may be a good thing.

Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Boot from mirror

2008-10-26 Thread Richard Elling
dick hoogendijk wrote:
 Is it or isn't it possible to boot off two mirrored ZFS disks and if
 yes, can this be done in the upcoming solaris 10 10/08 too?

   
Yes. Yes.  For details, please consult the ZFS Administration Guide.
http://www.opensolaris.org/os/community/zfs/docs/zfsadmin.pdf
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Cannot remove slog device from zpool

2008-10-26 Thread Richard Elling
CR 6574286 removing a slog doesn't work
http://bugs.opensolaris.org/view_bug.do?bug_id=6574286
 -- richard

Ethan Erchinger wrote:
 Sorry for the first incomplete send,  stupid Ctrl-Enter. :-)
 
 Hello,

 I've looked quickly through the archives and haven't found mention of 
 this issue.  I'm running SXCE (snv_99), which uses zfs version 13.  I 
 had an existing zpool:
 --
 [EMAIL PROTECTED] ~]$ zpool status -v data
   pool: data
  state: ONLINE
  scrub: none requested
 config:

 NAME   STATE READ WRITE CKSUM
 data   ONLINE   0 0 0
   mirror   ONLINE   0 0 0
 c4t1d0p0   ONLINE   0 0 0
 c4t9d0p0   ONLINE   0 0 0
   ...
 cache
   c4t15d0p0ONLINE   0 0 0

 errors: No known data errors

 --

 The cache device (c4t15d0p0) is an Intel SSD.  To test zil, I removed 
 the cache device, and added it as a log device:
 --
 [EMAIL PROTECTED] ~]$ pfexec zpool remove data c4t15d0p0
 [EMAIL PROTECTED] ~]$ pfexec zpool add data log c4t15d0p0
 [EMAIL PROTECTED] ~]$ zpool status -v data
   pool: data
  state: ONLINE
  scrub: none requested
 config:

 NAME   STATE READ WRITE CKSUM
 data   ONLINE   0 0 0
   mirror   ONLINE   0 0 0
 c4t1d0p0   ONLINE   0 0 0
 c4t9d0p0   ONLINE   0 0 0
   ...
 logs   ONLINE   0 0 0
   c4t15d0p0ONLINE   0 0 0

 errors: No known data errors
 --

 The device is working fine.  I then said, that was fun, time to remove 
 and add as cache device.  But that doesn't seem possible:
 --
 [EMAIL PROTECTED] ~]$ pfexec zpool remove data c4t15d0p0
 cannot remove c4t15d0p0: only inactive hot spares or cache devices can 
 be removed
 --

 I've also tried using detach, offline, each failing in other more 
 obvious ways.  The manpage does say that those devices should be 
 removable/replaceable.  At this point the only way to reclaim my SSD 
 device is to destroy the zpool.

 Just in-case you are wondering about versions:
 --
 [EMAIL PROTECTED] ~]$ zpool upgrade data
 This system is currently running ZFS pool version 13.

 Pool 'data' is already formatted using the current version.
 [EMAIL PROTECTED] ~]$ uname -a
 SunOS opensolaris 5.11 snv_99 i86pc i386 i86pc
 --

 Any ideas?

 Thanks,
 Ethan
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-26 Thread Orvar Korvar
For those of you who wants to build a NAS, this is mandatory reading I think. 
Read all comments too.

http://breden.org.uk/2008/03/02/a-home-fileserver-using-zfs/



I use a P45 mobo, Intel Q9450, ATI4850 and 4 GB RAM. AOC SATA card with 8 Sata 
slots. 4 Samsung 500GB drives. Works excellent in a P182 antec.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread Bob Friesenhahn
On Sun, 26 Oct 2008, William Bauer wrote:
 This has proven true on every OpenSolaris system I've tried--all of 
 which are using ZFS.  So what is it about logging into the console 
 that slows write performance to ONLY the top level home directory of 
 the username on the same console?

Recently OpenSolaris included a 'tracker' daemon and supportive 
framework which responds to any filesystem change in the user's home 
directory and opens the changed file so that it can be indexed. 
There were probably already Gnome features which depend on this file 
change detection (e.g. so that an icon may be generated). Perhaps this 
explains the problem?

Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread Volker A. Brandt
 If that were the case, why would it matter if I was logged into the console,
 and why would subdirectories of my home exhibit better write performance
 than the top level home directory?  A write to /export/home/username is
 slower than to /export/home/username/blah, but ONLY if that user is logged
 into the console. 

This smells of name resolution delays somewhere.  Do you have
a shell prompt that gets some host name or user name from
name services?  Is your /home directory owned by a non-existing
user or group?  Do you accidentally have something enabled
in /etc/nsswitch.conf that does not exist (ldap, nis, nis+)?
Maybe the first miss gets cached and all other misses get resolved
from the cache?

Or your nscd is disabled/confused/broken disabled altogether?

You have an interesting problem there...


Good luck -- Volker
-- 

Volker A. Brandt  Consulting and Support for Sun Solaris
Brandt  Brandt Computer GmbH   WWW: http://www.bb-c.de/
Am Wiesenpfad 6, 53340 Meckenheim Email: [EMAIL PROTECTED]
Handelsregister: Amtsgericht Bonn, HRB 10513  Schuhgröße: 45
Geschäftsführer: Rainer J. H. Brandt und Volker A. Brandt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool import problem

2008-10-26 Thread Terry Heatlie
Folks,
I have a zpool with a raidz2 configuration which I've been switching between
two machines - an old one with a hardware problem and a new one, which
doesn't have hardware issues, but has a different configuration .   I've
been trying to import the pool on the new machine, so I can back up the
data, because the old (broken) machine resets (I don't think it's panicking,
because there are no logged messages) every time I try to tar off the data
from the ZFS.

 Unfortunately, the first time I tried to import the pool on the new
machine, I didn't have the right five drives in it, so it didn't work.
 After I figured out that I was confused about which was the boot drive, I
did get the five drives into the new machine and asked it to import the
pool.  It said that the pool could not be imported due to damaged devices or
data.   Which is slightly odd, since it had been mounting the pool fine on
the broken machine before.

I then moved the drives back into the old machine, figuring I'd at least
copy some small stuff onto a USB stick (it only dies reading large files,
apparently), but now the old machine can't mount the pool either, and asking
it to import gives the same message.   It shows all five drives online, but
says the pool is UNAVAIL due to insufficient replicas, and the  raidz2 is
UNAVAIL due to corrupted data.

Must I resign myself to having lost this pool due to the hardware problems
I've had, and restore such backups as I have on the new machine, or is there
something that can be done to get the pool back online at least in degraded
mode?

Thanks in advance,

--Terry.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-26 Thread Nigel Smith
Ok on the answers to all my questions.
There's nothing that really stands out as being obviously wrong.
Just out of interest, what build of OpenSolaris are you using?

One thing you could try on the Ethernet capture file, is to set
the WireShark 'Time' column like this:
View  Time Display Format  Seconds Since Previous Displayed Packet

Then look down the time column for any unusual high time delays
between packets. Any unusually high delays during
a data transfer phase, may indicate a problem.


Another thing you could try is measuring network performance
with a utility called 'iperf'.
It's not part of Solaris, so you would need to compile it.
Download the source from here:
http://sourceforge.net/projects/iperf/

I've just compiled the latest version 2.0.4 on snv_93
without problem, using the normal configure, make, make install.

If you want to run 'iperf' on a windows box, you can
download a '.exe' of an older version here:
http://www.noc.ucf.edu/Tools/Iperf/

You can find tutorials on how to use it at these links:
http://www.openmaniak.com/iperf.php
http://www.enterprisenetworkingplanet.com/netos/article.php/3657236

I've just tried 'iperf' between my OpenSolaris pc  an old
Windows pc, both with low-cost realtek gigabit cards and
linked via a low-cost NetGear switch. I measured a TCP
bandwidth of 196 Mbit/sec in one direction and
145 Mbit/sec in the opposite direction.
(On OpenSolaris, Iperf was not able to increase
the default TCP window size of 48K bytes.)
Regards
Nigel Smith
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] HELP! zfs with iscsitadm (on poweredge1900) and VMWare!

2008-10-26 Thread Nigel Smith
I asked Tano to use the 'snoop' command to capture the Ethernet
packets to a file, while he attempted VMware's 'VMotion'.

 # snoop -d {device} -o {filename} tcp port 3260

This file was made available to me on Tano's web server.
The file size was nearly 85 Mbytes, capturing over 100,000 packets.
I have downloaded the capture file, and been looking at it with 
Ethereal and WireShark.

I do not have a corresponding 'iscsisnoop.d' file, but from the pattern
of activity that I see, I can well imagine that it would show the same
pattern of that we saw from Eugene, which I reported on here:
http://mail.opensolaris.org/pipermail/storage-discuss/2008-October/006444.html

(So here I'm looking at what's happening at the lower TCP level,
rather than at the iScsi level.)

In the Ethernet capture file, I can see the pattern of bursts of
writes from the initiator. The Target can accept so many of these,
and then needs to slow things down by reducing the TCP window size.
Eventually the target says the TCP Window size is zero, effectively
asking the initiator to stop.

Now to start with, the target only leaves the 'TCP ZeroWindow', in
place for a fraction of a second. Then it opens things up again
by sending a 'TCP Window Update', restoring the window to 65160 bytes,
and transfer resumes. This is normal and expected.

But eventually we get to a stage where the target sets the TCP 'ZeroWindow'
and leaves it there for an extended period of time.  I talking about seconds 
here.
The initiator starts to send 'TCP ZeroWindowProbe' packets every 5 seconds.
The target promptly responds with a 'TCP ZeroWindowProbeAck' packet.
(Presumably, this is the initiator just confirming that the target is still 
alive.)
This cycle of Probes  Ack's repeats for 50 seconds.
During this period the target shows no sign of wanting to accept any more data.
Then the initiator seems to decide it has had enough, and just cannot
be bothered to wait any longer, and it [RST,ACK]'s the TCP session, and
then starts a fresh iscsi login.
(And then we go around the whole cycle of the pattern again.)

The question is why has the target refused to accept any more data for over 50 
seconds!

The obvious conclusion would be that the OpenSolaris box is so busy that
it does not have any time left to empty the network stack buffers.
But this then just leads you to another  question - why?

So the mystery deepens, and I am running out of ideas!

Tano, maybe you could check the network performance, with the 'iperf'
programs, as mentioned here:
http://mail.opensolaris.org/pipermail/zfs-discuss/2008-October/052136.html

Does the OpenSolaris box give any indication of being busy with other things?
Try running 'prstat' to see if it gives any clues.

Presumably you are using ZFS as the backing store for iScsi, in
which case, maybe try with a UFS formatted disk to see if that is a factor.
Regards
Nigel Smith
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-26 Thread Matt Harrison
Nigel Smith wrote:
 Ok on the answers to all my questions.
 There's nothing that really stands out as being obviously wrong.
 Just out of interest, what build of OpenSolaris are you using?
 
 One thing you could try on the Ethernet capture file, is to set
 the WireShark 'Time' column like this:
 View  Time Display Format  Seconds Since Previous Displayed Packet
 
 Then look down the time column for any unusual high time delays
 between packets. Any unusually high delays during
 a data transfer phase, may indicate a problem.

Along with the errors that I noted previously, some of the packets to
seem to be taking a rather long time (0.5s).

I've taken a cap file from wireshark in the hope it clears up some
information. The capture is less than a minute of playing a video over
the cifs share.

It's a little too large to send in a mail so I've posted it at

http://distfiles.genestate.com/_1_20081027010354.zip

 Another thing you could try is measuring network performance
 with a utility called 'iperf'.

Thanks for pointing this program out, I've just run it to the gentoo
firewall we've got, and it's reporting good speeds for the network.

Thanks

Matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-26 Thread Matt Harrison
Nigel Smith wrote:
 Ok on the answers to all my questions.
 There's nothing that really stands out as being obviously wrong.
 Just out of interest, what build of OpenSolaris are you using?

Damn forgot to add that, I'm running SXCE snv_97.

Thanks

Matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
 This smells of name resolution delays somewhere.  Do
 you have
 a shell prompt that gets some host name or user name
 from
 name services?  Is your /home directory owned by a
 non-existing
 user or group?  Do you accidentally have something
 enabled
 in /etc/nsswitch.conf that does not exist (ldap, nis,
 nis+)?
 Maybe the first miss gets cached and all other misses
 get resolved
 from the cache?
 
 Or your nscd is disabled/confused/broken disabled
 altogether?


I think you've misunderstood something here, perhaps in the way I've tried to 
explain it.  I'm having trouble figuring out what that might be.  Perhaps your 
response was to the wrong discussion?
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
This sounds plausible I suppose  Being unfamiliar with this tracker daemon, 
I can blindly accept it as a maybe!
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
For clarity, here's how you can reproduce what I'm asking about:

This is for local file systems on build 86 and not about NFS or
any remote mounts.  You can repeat these 100 times and always get
the same result, whether you reboot between trials or leave the
system running.

1.  Log into the console of your OpenSolaris system.

2.  Open a terminal in your home directory.

3.  Create a file as in the command below and note the write
speed it returns.  I know this method is not a true test of write
performance, but I'm interested in relative and not absolute numbers.
This method seems quite consistent in the results it returns.

dd if=/dev/zero of=outfile bs=1024 count=100

4.  Enter a subdirectory of your home and do the same as above.

5.  Log out of the console and perform the same tests via a remote
login using SSH or your favorite method.  See that now there is no
difference in performance.

I've found the difference between steps 3 and 4 to be typically 25%.
This happens even with the most vanilla new install.  On VirtualBox,
the differences can be more like 5x (i.e. 500%!).

Note that if you do step 5 when still logged into the console,
you get the same results as in step 3-4.

That's it.  Maybe not hugely important, but I'm trying to understand
why this happens.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import problem

2008-10-26 Thread Richard Elling
Terry Heatlie wrote:
 Folks,

 I have a zpool with a raidz2 configuration which I've been switching 
 between two machines - an old one with a hardware problem and a new 
 one, which doesn't have hardware issues, but has a different 
 configuration .   I've been trying to import the pool on the new 
 machine, so I can back up the data, because the old (broken) machine 
 resets (I don't think it's panicking, because there are no logged 
 messages) every time I try to tar off the data from the ZFS.  

  Unfortunately, the first time I tried to import the pool on the new 
 machine, I didn't have the right five drives in it, so it didn't work. 
  After I figured out that I was confused about which was the boot 
 drive, I did get the five drives into the new machine and asked it to 
 import the pool.  It said that the pool could not be imported due to 
 damaged devices or data.   Which is slightly odd, since it had been 
 mounting the pool fine on the broken machine before.   

 I then moved the drives back into the old machine, figuring I'd at 
 least copy some small stuff onto a USB stick (it only dies reading 
 large files, apparently), but now the old machine can't mount the pool 
 either, and asking it to import gives the same message.   It shows all 
 five drives online, but says the pool is UNAVAIL due to insufficient 
 replicas, and the  raidz2 is UNAVAIL due to corrupted data.

 Must I resign myself to having lost this pool due to the hardware 
 problems I've had, and restore such backups as I have on the new 
 machine, or is there something that can be done to get the pool back 
 online at least in degraded mode?

Note: we're also working on a troubleshooting wiki... need more days in the
hour...

You should try to read the labels from each device.
zdb -l /dev/rdsk/...

You should see 4 labels for each proper device.

Here is my hypothesis:

If you see a device which has only label 0 and 1, then it may
be the case that the label has overlapping partitions.  Why does
this matter?  Because under normal circumstances, the actual
devices used for creating or importing the pool are stored in the
/etc/zfs/zpool.cache file.  When the system boots, it looks there
first and will import the pools listed therein.

When you export the pool, the zpool.cache entries for the pool
are removed.

If the pool is not in zpool.cache, then zpool import scans all of
the devices found in /dev/dsk for valid pools.  If you have overlapping
partitions or slices, then a partially exposed vdev may be found.
But since it won't be complete, due to perhaps not being able to
see the end of the device which is where labels 2  3 are located,
then it will be marked as bad.  The solution would be to reconcile
the partions/slices using format.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread Richard Elling
I cannot recreate this on b101. There is no significant difference between
the two on my system.
 -- richard

William Bauer wrote:
 For clarity, here's how you can reproduce what I'm asking about:

 This is for local file systems on build 86 and not about NFS or
 any remote mounts.  You can repeat these 100 times and always get
 the same result, whether you reboot between trials or leave the
 system running.

 1.  Log into the console of your OpenSolaris system.

 2.  Open a terminal in your home directory.

 3.  Create a file as in the command below and note the write
 speed it returns.  I know this method is not a true test of write
 performance, but I'm interested in relative and not absolute numbers.
 This method seems quite consistent in the results it returns.

 dd if=/dev/zero of=outfile bs=1024 count=100

 4.  Enter a subdirectory of your home and do the same as above.

 5.  Log out of the console and perform the same tests via a remote
 login using SSH or your favorite method.  See that now there is no
 difference in performance.

 I've found the difference between steps 3 and 4 to be typically 25%.
 This happens even with the most vanilla new install.  On VirtualBox,
 the differences can be more like 5x (i.e. 500%!).

 Note that if you do step 5 when still logged into the console,
 you get the same results as in step 3-4.

 That's it.  Maybe not hugely important, but I'm trying to understand
 why this happens.
 --
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
 I cannot recreate this on b101. There is no significant difference between
 the two on my system.

That's encouraging...unless no one can reproduce it on 86, then I'm forgetting 
something.  I've done this a dozen times on several systems, so maybe ZFS 
performance has been improved.

What numbers did you get (out of curiosity)?
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
Bingo!  I just updated a system from 86 to 99 and the problem is gone.  Even 
better, it was a VB guest, and the ZFS performance on the guest increased 5x in 
this test, as I mentioned earlier.  Granted, a VB guest may not be the best 
test and it only applies to top level home directories, but it was consistent 
and now returns the same higher numbers no matter what.

Thank you for testing b101.  So how close ARE we to the official November 
release?  Early, mid, or late in the month...!
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS write performance on boot disk

2008-10-26 Thread William Bauer
After a zpool upgrade, this simple test's write speed jumped up yet another 
20%.  Looks like ZFS is getting better.  As one would hope  expect.
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss