Re: [zfs-discuss] linux versus sol10

2006-11-08 Thread Paul van der Zwan


On 7 Nov 2006, at 21:02, Michael Schuster wrote:


listman wrote:
hi, i found a comment comparing linux and solaris but wasn't sure  
which version of solaris was being referred. can the list confirm  
that this issue isn't a problem with solaris10/zfs??
Linux also supports asynchronous directory updates which can make  
a significant performance improvement when branching. On Solaris  
machines, inode creation is very slow and can result in very long  
iowait states.


I think this cannot be commented on in a useful fashion without  
more information this supposed issue. AFAIK, neither ufs nor zfs  
create inodes (at run time), so this is somewhat hard to put into  
context.


get a complete description of what this is about, then maybe we can  
give you a useful answer.




This could be related to Linux trading reliability for speed by doing  
async metadata updates.
If your system crashes before your metadata is flushed to disk your  
filesystem might be hosed and a restore

from backups may be needed.

Paul



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re[2]: [zfs-discuss] linux versus sol10

2006-11-08 Thread Robert Milkowski
Hello Paul,

Wednesday, November 8, 2006, 3:23:35 PM, you wrote:

PvdZ On 7 Nov 2006, at 21:02, Michael Schuster wrote:

 listman wrote:
 hi, i found a comment comparing linux and solaris but wasn't sure  
 which version of solaris was being referred. can the list confirm  
 that this issue isn't a problem with solaris10/zfs??
 Linux also supports asynchronous directory updates which can make  
 a significant performance improvement when branching. On Solaris  
 machines, inode creation is very slow and can result in very long  
 iowait states.

 I think this cannot be commented on in a useful fashion without  
 more information this supposed issue. AFAIK, neither ufs nor zfs  
 create inodes (at run time), so this is somewhat hard to put into  
 context.

 get a complete description of what this is about, then maybe we can  
 give you a useful answer.


PvdZ This could be related to Linux trading reliability for speed by doing
PvdZ async metadata updates.
PvdZ If your system crashes before your metadata is flushed to disk your  
PvdZ filesystem might be hosed and a restore
PvdZ from backups may be needed.

you can achieve something similar with fastfs on ufs file systems and
setting zil_disable to 1 on ZFS.



-- 
Best regards,
 Robertmailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: Re[2]: [zfs-discuss] linux versus sol10

2006-11-08 Thread Paul van der Zwan


On 8 Nov 2006, at 16:16, Robert Milkowski wrote:


Hello Paul,

Wednesday, November 8, 2006, 3:23:35 PM, you wrote:

PvdZ On 7 Nov 2006, at 21:02, Michael Schuster wrote:


listman wrote:

hi, i found a comment comparing linux and solaris but wasn't sure
which version of solaris was being referred. can the list confirm
that this issue isn't a problem with solaris10/zfs??
Linux also supports asynchronous directory updates which can make
a significant performance improvement when branching. On Solaris
machines, inode creation is very slow and can result in very long
iowait states.


I think this cannot be commented on in a useful fashion without
more information this supposed issue. AFAIK, neither ufs nor zfs
create inodes (at run time), so this is somewhat hard to put into
context.

get a complete description of what this is about, then maybe we can
give you a useful answer.



PvdZ This could be related to Linux trading reliability for speed  
by doing

PvdZ async metadata updates.
PvdZ If your system crashes before your metadata is flushed to  
disk your

PvdZ filesystem might be hosed and a restore
PvdZ from backups may be needed.

you can achieve something similar with fastfs on ufs file systems and
setting zil_disable to 1 on ZFS.



Sure UFS and ZFS can be faster, but having fast, but possibly  
dangerous, defaults

gives you nice benchmark figures ;-)
In real life I prefer the safe, but a bit slower, defaults, as should  
anybody

who values his data.

Paul

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: Re[2]: [zfs-discuss] linux versus sol10

2006-11-08 Thread Joerg Schilling
Paul van der Zwan [EMAIL PROTECTED] wrote:

 Sure UFS and ZFS can be faster, but having fast, but possibly  
 dangerous, defaults
 gives you nice benchmark figures ;-)
 In real life I prefer the safe, but a bit slower, defaults, as should  
 anybody
 who values his data.

There is another point besides having dangerous defaults on Linux:

People don't know what they benchmark. This is caused by the fact that
people usually meter the time for a tar xf bla call, while a lot of the
data is only inside the RAM when tar is finished and you would need to wait
until the data is on disk. 

Solaris starts earlier with copying the RAM cache to disk than Linux does and
Solaris usually gives you bettter I/O bandwidth than Linux. 

Jörg

-- 
 EMail:[EMAIL PROTECTED] (home) Jörg Schilling D-13353 Berlin
   [EMAIL PROTECTED](uni)  
   [EMAIL PROTECTED] (work) Blog: http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/old/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] linux versus sol10

2006-11-08 Thread Neil Perrin



Robert Milkowski wrote On 11/08/06 08:16,:

Hello Paul,

Wednesday, November 8, 2006, 3:23:35 PM, you wrote:

PvdZ On 7 Nov 2006, at 21:02, Michael Schuster wrote:



listman wrote:

hi, i found a comment comparing linux and solaris but wasn't sure  
which version of solaris was being referred. can the list confirm  
that this issue isn't a problem with solaris10/zfs??
Linux also supports asynchronous directory updates which can make  
a significant performance improvement when branching. On Solaris  
machines, inode creation is very slow and can result in very long  
iowait states.


I think this cannot be commented on in a useful fashion without  
more information this supposed issue. AFAIK, neither ufs nor zfs  
create inodes (at run time), so this is somewhat hard to put into  
context.


get a complete description of what this is about, then maybe we can  
give you a useful answer.





PvdZ This could be related to Linux trading reliability for speed by doing
PvdZ async metadata updates.
PvdZ If your system crashes before your metadata is flushed to disk your  
PvdZ filesystem might be hosed and a restore

PvdZ from backups may be needed.

you can achieve something similar with fastfs on ufs file systems and
setting zil_disable to 1 on ZFS.


There's a difference for both of these.

UFS now has logging (journalling) as the default, and so any crashes/power fails
will keep the integrity of the metadata intact (ie no fsck/restore).

ZFS has no problem either as its fully transacts both data and meta data
and should never see corruption with intent log disabled or enabled.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] linux versus sol10

2006-11-08 Thread Matthew Ahrens

Robert Milkowski wrote:

PvdZ This could be related to Linux trading reliability for speed by doing
PvdZ async metadata updates.
PvdZ If your system crashes before your metadata is flushed to disk your  
PvdZ filesystem might be hosed and a restore

PvdZ from backups may be needed.

you can achieve something similar with fastfs on ufs file systems and
setting zil_disable to 1 on ZFS.


No, zil_disable does not trade off consistency for performance the way 
'fastfs' on ufs or async metadata updates on EXT do!


Setting zil_disable causes ZFS to not push synchronous operations (eg, 
fsync(), O_DSYNC, NFS ops) to disk immediately, but it does not 
compromise filesystem integrity in any way.  Unlike these other 
filesystems fast modes, ZFS (even with zil_disable=1) will not corrupt 
itself and send you to backup tapes.


To state it another way, setting 'zil_disable=1' on ZFS will at worst 
cause some operations which requested synchronous semantics to not 
actually be on disk in the event of a crash, whereas other filesystems 
can corrupt themselves and lose all your data.


All that said, 'zil_disable' is a completely unsupported hack, and 
subject to change at any time.  It will probably eventually be replaced 
by 6280630 zil synchronicity.


--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re[2]: [zfs-discuss] linux versus sol10

2006-11-08 Thread Robert Milkowski
Hello Matthew,

Wednesday, November 8, 2006, 5:31:28 PM, you wrote:

MA Robert Milkowski wrote:
 PvdZ This could be related to Linux trading reliability for speed by doing
 PvdZ async metadata updates.
 PvdZ If your system crashes before your metadata is flushed to disk your  
 PvdZ filesystem might be hosed and a restore
 PvdZ from backups may be needed.
 
 you can achieve something similar with fastfs on ufs file systems and
 setting zil_disable to 1 on ZFS.

MA No, zil_disable does not trade off consistency for performance the way
MA 'fastfs' on ufs or async metadata updates on EXT do!

I know that. But from user perspective setting zil_disable can make
many applications much faster.

-- 
Best regards,
 Robertmailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Thoughts on patching + zfs root

2006-11-08 Thread Jason King
Anxiously anticipating the ability to boot off zfs, I know there's been some 
talk about leveraging some of the snapshotting/cloning features in conjunction 
with upgrades and patches.

What I am really hoping for is the ability to clone /, patch the clone, then 
boot off the clone (by doing a clone swap).  This would minimize the downtime 
needed to patch (as compared to today) since the install could be done while 
the system is still up and running.  I suspect to do this might require some 
interaction with things like patchadd, etc., so I am curious if such a feature 
is already in the works, or if perhaps an RFE should be filed.

Assuming the plans are already there for this, I'd like to anticipate any 
gotchas with our standard install procedure, and was just wondering if any 
thought had been put in as to what the requirements would be.  Just off the top 
of my head, I would guess it'd mostly revolve around making sure stuff is 
separated properly into different filesystems, but was curious if anyone else 
has thought about this.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] OpenSolaris vs. Solaris 10 11/06 (S10u3) for NFS ZFS Server

2006-11-08 Thread Wes Williams
I'm in the process of building a Solaris NFS server with ZFS and was wondering 
if any gurus here have any comments as to choosing the upcoming Solairs 10 
11/06 [presumably] or OpenSolaris bXX/Solairs Express for this use.  Even with 
my use of OpenSolaris I maintain a service contract to show my support, so bug 
fixes in a static supported version shouldn't be an issue in picking a 
version.

So, the short question is are there any super-cool must-have ZFS/NFS features 
in OpenSolaris now that S10u3 won't have right away?

Thanks!
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OpenSolaris vs. Solaris 10 11/06 (S10u3) for NFS ZFS Server

2006-11-08 Thread Al Hopper
On Wed, 8 Nov 2006, Wes Williams wrote:

[  reformatted ... ]
 I'm in the process of building a Solaris NFS server with ZFS and was
 wondering if any gurus here have any comments as to choosing the
 upcoming Solairs 10 11/06 [presumably] or OpenSolaris bXX/Solairs
 Express for this use.  Even with my use of OpenSolaris I maintain a
 service contract to show my support, so bug fixes in a static
 supported version shouldn't be an issue in picking a version.

 So, the short question is are there any super-cool must-have ZFS/NFS
 features in OpenSolaris now that S10u3 won't have right away?

[ Hi Wes from [EMAIL PROTECTED] ]

And the smart-ask answer is:

By Definition: the OpenSolaris/Solaris Express feature that is *your*
must-have feature, probably won't be in Update 3! :)

Since U3 is a cherry picked subset of OpenSolaris, it's difficult to
predict which features made it and which did not.  I know that it's a Sun
decision made from a basis of (Sun) solid logic - it's just difficult to
figure out this mental process, from the outside looking in.

In my case, the ability to clone a zone from a zfs snapshot did'nt make
it! :(

I don't think that there are any ZFS NFS related fixes that did'nt make it
into U3 (others will correct me if I'm wrong here).  It's a tough choice
to make...  The huge upside is ... that there is a choice!  :)

Regards,

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
   Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
 OpenSolaris Governing Board (OGB) Member - Feb 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: OpenSolaris vs. Solaris 10 11/06 (S10u3) for NFS

2006-11-08 Thread Wes Williams
 [ Hi Wes from [EMAIL PROTECTED] ]
 
 And the smart-ask answer is:
 
 By Definition: the OpenSolaris/Solaris Express
 feature that is *your*
 must-have feature, probably won't be in Update 3!
 :)
 
Exactly, that's why I used quotes as I'm sure I'd be happy with S10u3, assuming 
ignorance is bliss.  Then again, I've never believed ignorance is bliss.  ;)

I simply don't know enough about the difference between the two OS versions yet 
to make an educated guess as to which venue to pursue presently - especially 
since S10u3 isn't released at the moment and I haven't followed ZFS closely 
since ZFSv1 was current.

 snip
 I don't think that there are any ZFS NFS related
 fixes that did'nt make it
 into U3 (others will correct me if I'm wrong here).
  It's a tough choice
 o make...  The huge upside is ... that there is a
 choice!  :)

Indeed, knowing there are no known NFS fixes omitted is great news, not that I 
was aware of anything in NFS needing fixing.  To me this is just reassuring 
since my goal is to build this box and nearly forget about it while it's hidden 
away and 'it just works.'

Since I didn't hear anything like ZFSv3 vs. ZFSv2 or some such, I'll probably 
just go w/ S10u3 for this project.

Thanks for your feedback Al.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OpenSolaris vs. Solaris 10 11/06 (S10u3) for NFS ZFS Server

2006-11-08 Thread Mike Gerdts

On 11/8/06, Al Hopper [EMAIL PROTECTED] wrote:

In my case, the ability to clone a zone from a zfs snapshot did'nt make
it! :(


Yeah, but if you read the man page in the U3 beta, you would see that
the man page changes made it over.  Personally, I would have preferred
the code instead of the man pages.  :)

Per a discussion I had with Jerry Jelinek after I brought this up as
part of the beta:

Jerry said...

backporting this to S10 until we have the ability to upgrade
zones that reside on zfs. At this time we do not recommend
customers place their zones on zfs. Obviously this works and
is usually ok as long as you do not care about being able to
upgrade your system, but this is not something we can
generally support until we have an upgrade solution.


So long as upgrades are not something that I am concerned about, it
seems as though the following is an approach that would work out OK.
Please let me know if you concur.

### Initial setup of /zones fs

# zfs create local/zones
# zfs set mountpoint=/zones local/zones

### Create my full-root template zone

# zonecfg -z template-full
create -t SUNWblank
set zonepath=/zones/template-full
verify
commit
exit
# zfs create local/zones/template-full
# chmod 700 /zones/template-full
# zoneadm -z template-full install
# zoneadm -z template-full boot
# zlogin -C template-full

### Do zone customization stuff (e.g. run JASS to harden)

### Prepare the zone for cloning

# zoneadm -z template-full halt
# zoneadm -z template-full detach
# zfs snapshot local/zones/[EMAIL PROTECTED]

### Create a clone

# zonecfg -z newzone
create -t template-full
set zonepath=/zones/newzone
 set up networks, pools, etc.
verify
commit
exit
# zfs clone local/zones/[EMAIL PROTECTED] local/zones/newzone
# zoneadm attach newzone
# zoneadm -z newzone boot -s
# yes | zlogin newzone sys-unconfig
# zoneadm -z newzone boot
# zlogin -C newzone
perform sysidcfg stuff

Jerry later said:


This is a very clever procedure. You might want to send this
out to the opensolaris zones-discuss list so that others can
benefit from your idea here. I don't see any problems with
this procedure.


However, I waited until someone else announced the features or lack
thereof found in S10 11/06.  :)

Mike

--
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best Practices recommendation on x4200

2006-11-08 Thread Richard Elling - PAE

Mike Gerdts wrote:

On 11/7/06, Richard Elling - PAE [EMAIL PROTECTED] wrote:

 d10 mirror of c0t2d0s0 and c0t3d0s0swap (2+2GB, to match above)

Also a waste, use a swap file.  Add a dumpdev if you care about
kernel dumps, no need to mirror a dumpdev.


How do you figure that allocating space to a swap file is less of a
waste than adding space to a swap device?


If you ever guess wrong (which you will), you can just make another
swap file or redo the existing swap file.  If you carve out a slice,
then reclaiming the space is much more difficult.  Creating more slices
tends to also be difficult, so when you quess wrong you may still end
up swapping to files.


Simple /.  Make it big enough to be useful.  Keep its changes to a
minimum.  Make more than one, so that you can use LiveUpgrade.
For consistency, you could make each disk look the same.
s0 / 10G
s6 zpool free
s7 metadb 100M


Since ZFS can get performance boosts from enabling the disk write
cache if it has the whole disk, you may want to consider something
more like the following for two of the disks (assumes mirroring rather
than raidz in the zpool):

s0 / 10G
s1 swap pick your size
s3 alt / 10G
s6 zpool free
s7 metadb 100M

The other pair of disks are given entirely to the zpool.


Use two disks for your BE, the other two for your ABE (assuming all are
bootable).


In any case, be sure that your root slices do not start at cylinder 0
(hmmm... maybe this is SPARC-specific advice...).


I think this is folklore.  Can you cite a reference?
NB. traditionally, block 0 contains the vtoc and for SPARC systems,
blocks 1-15 contain the bootblocks, see installboot(1M).
Cylinder 0 may contain thousands of blocks for modern disks.  It is
a waste not to use them.  AFAIK, all Sun software which deals with
raw devices is aware of this.


  One way to populate
an ABE is to mirror slices.  However, you cannot mirror between a
device that starts at cylinder 0 and one that does not.  


Where is this restriction documented?  It doesn't make sense to me.
Maybe you have a scar from running Sybase in a previous life? ;-)
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Thoughts on patching + zfs root

2006-11-08 Thread Lori Alt

Torrey McMahon wrote:

Jason King wrote:
Anxiously anticipating the ability to boot off zfs, I know there's 
been some talk about leveraging some of the snapshotting/cloning 
features in conjunction with upgrades and patches.


What I am really hoping for is the ability to clone /, patch the 
clone, then boot off the clone (by doing a clone swap).  


I'm pretty sure we already have what you are looking for. Check out 
Live Upgrade on docs.sun.com.

Right now, liveupgrade makes lots of assumptions about boot environments
being in ufs file systems, but fixing that is part of the zfs-boot 
plan.  So yes, the kinds

of things you can do with liveupgrade, such as cloning environments and then
patching or upgrading them, and then activating them, will be possible with
zfs bootable environments (and will be a lot faster and easier, since 
cloning will be
almost instantaneous and will not require a preallocated slice). 

I'm not sure what will turn out to be a gotcha for you, so let me just 
describe
the basics of how zfs booting will work.  Some storage pools will be 
designated
as root pools, which means that they contain at least one dataset that 
contains

a bootable root file system.  There will be some restrictions on root pools
(the only allowed configuration is an n-way mirror of single disks or 
slices,

each of which must be accessible from the boot prom or BIOS), so best
practice will be to keep data (such as database tables) in a pool separate
from the root pool.  (The separation isn't required but will probably turn
out to be best for ease of management.) 

Some datasets will be designated as bootable, which means that they 
contain

a root file system.  A bootable environment (BE, in LiveUpgrade terminology)
can consist of several datasets, exactly one of which is bootable.
For example, it will be desirable in many cases to put each zone root in a
separate dataset.  Each of those zone root datasets will be subordinate to a
bootable dataset which contains the root of the BE.   Cloning a BE means
cloning each dataset in the BE.

We in the zfs-boot project owe the community more information about
how the work is progressing.  I hope we can get more of that information
out fairly soon.

Lori



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Best Practices recommendation on x4200

2006-11-08 Thread Nathan Kroenert
On Thu, 2006-11-09 at 10:21, Richard Elling - PAE wrote:
  One way to populate
  an ABE is to mirror slices.  However, you cannot mirror between a
  device that starts at cylinder 0 and one that does not.  
 
 Where is this restriction documented?  It doesn't make sense to me.
 Maybe you have a scar from running Sybase in a previous life? ;-)

IIRC, that's a part of the history of disksuite / SVM. Moreover, it was
that you cannot mirror a slice that has a VTOC label on it to one that
does not... (hence the understanding of it being a cylinder 0 issue).

http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/io/lvm/mirror/mirror_ioctl.c#887

Or, perhaps I need more coffee...

Cheers!

Nathan. ;)





___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] OpenSolaris vs. Solaris 10 11/06 (S10u3) for NFS ZFS Server

2006-11-08 Thread Nathan Kroenert
For me, it came down to - Do I want to patch, or upgrade?

My gateway to the internet is a solaris 10 box, patched whenever
required. I like that as soon as a security patch is available, I can
apply it and reboot. Simple.

My laptop runs nevada. I upgrade from network / dvd when I see a new
feature that excites me.

As far as whiz-bang things that would excite you, only you will know
that for sure. :)

Cheers!

Nathan.




On Thu, 2006-11-09 at 08:58, Wes Williams wrote:
 I'm in the process of building a Solaris NFS server with ZFS and was 
 wondering if any gurus here have any comments as to choosing the upcoming 
 Solairs 10 11/06 [presumably] or OpenSolaris bXX/Solairs Express for this 
 use.  Even with my use of OpenSolaris I maintain a service contract to show 
 my support, so bug fixes in a static supported version shouldn't be an 
 issue in picking a version.
 
 So, the short question is are there any super-cool must-have ZFS/NFS 
 features in OpenSolaris now that S10u3 won't have right away?
 
 Thanks!
  
 
 This message posted from opensolaris.org
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
-- 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Is there inotify for ZFS?

2006-11-08 Thread LingBo Tang

Hi all,

As inotify for Linux, is there same mechanism in Solaris for ZFS?
I think this functionality is helpful for desktop search engine.

I know one engineer of Sun is working on file event monitor, which
will provide some information of file events, but is not for
search purpose because it might has problem while monitoring
a large system.

Regards,
Lingbo
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss