Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-06 Thread Dave

C. Bergström wrote:

Bob Friesenhahn wrote:
I don't know if anyone has noticed that the topic is google summer of 
code.  There is only so much that a starving college student can 
accomplish from a dead-start in 1-1/2 months.  The ZFS equivalent of 
eliminating world hunger is not among the tasks which may be 
reasonably accomplished, yet tasks at this level of effort is all that 
I have seen mentioned here.
May I interject a bit.. I'm silently collecting this task list and even 
outside of gsoc may help try to arrange it from a community 
perspective.  Of course this will be volunteer based unless /we/ get a 
sponsor or sun beats /us/ to it.  So all the crazy ideas welcome..




I would really like to see a feature like 'zfs diff f...@snap1 
f...@othersnap' that would report the paths of files that have either been 
added, deleted, or changed between snapshots. If this could be done at 
the ZFS level instead of the application level it would be very cool.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-06 Thread Tim Haley

Dave wrote:

C. Bergström wrote:

Bob Friesenhahn wrote:
I don't know if anyone has noticed that the topic is google summer 
of code.  There is only so much that a starving college student can 
accomplish from a dead-start in 1-1/2 months.  The ZFS equivalent of 
eliminating world hunger is not among the tasks which may be 
reasonably accomplished, yet tasks at this level of effort is all 
that I have seen mentioned here.
May I interject a bit.. I'm silently collecting this task list and 
even outside of gsoc may help try to arrange it from a community 
perspective.  Of course this will be volunteer based unless /we/ get a 
sponsor or sun beats /us/ to it.  So all the crazy ideas welcome..




I would really like to see a feature like 'zfs diff f...@snap1 
f...@othersnap' that would report the paths of files that have either been 
added, deleted, or changed between snapshots. If this could be done at 
the ZFS level instead of the application level it would be very cool.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


this is actually in the works.  There is a functioning prototype.

-tim



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-06 Thread Kyle Kakligian
Its been suggested before and I've heard its in the freebsd port...
support for spindown?

On Fri, Mar 6, 2009 at 11:40 AM, Dave dave-...@dubkat.com wrote:
 C. Bergström wrote:

 Bob Friesenhahn wrote:

 I don't know if anyone has noticed that the topic is google summer of
 code.  There is only so much that a starving college student can accomplish
 from a dead-start in 1-1/2 months.  The ZFS equivalent of eliminating world
 hunger is not among the tasks which may be reasonably accomplished, yet
 tasks at this level of effort is all that I have seen mentioned here.

 May I interject a bit.. I'm silently collecting this task list and even
 outside of gsoc may help try to arrange it from a community perspective.  Of
 course this will be volunteer based unless /we/ get a sponsor or sun beats
 /us/ to it.  So all the crazy ideas welcome..


 I would really like to see a feature like 'zfs diff f...@snap1 f...@othersnap'
 that would report the paths of files that have either been added, deleted,
 or changed between snapshots. If this could be done at the ZFS level instead
 of the application level it would be very cool.

 --
 Dave
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-06 Thread Bart Smaalders



I would really like to see a feature like 'zfs diff f...@snap1 f...@othersnap'
that would report the paths of files that have either been added, deleted,
or changed between snapshots. If this could be done at the ZFS level instead
of the application level it would be very cool.

--


AFAIK, this is being actively developed, w/ a prototype working...

- Bart

--
Bart Smaalders  Solaris Kernel Performance
ba...@cyber.eng.sun.com http://blogs.sun.com/barts
You will contribute more with mercurial than with thunderbird.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-05 Thread Darren J Moffat

Wes Felter wrote:

slog and L2ARC on the same SSD


You can do that already today.

Create two slices using format(1M) and add the slices rather than the 
whole disk as the L2ARC or slog device.


However IMO this is with current SSD techology probably the wrong thing 
to do.  The slog wants a very fast write devices, the L2ARC wants a very 
fast read device.  Unless your SSD is fast at both reads and writes 
using the same SSD for the slog and L2ARC isn't a good idea.


--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-05 Thread Miles Nordin
 gm == Gary Mills mi...@cc.umanitoba.ca writes:

gm There are many different components that could contribute to
gm such errors.

yes of course.

gm Since only the lower ZFS has data redundancy, only it can
gm correct the error.

um, no?

An example already pointed out: kerberized NFS will detect network
errors that sneak past the weak TCP checksum, and resend the data.
This will work even on an unredundant, unchecksummed UFS filesystem to
correct network-induced errors.  There is no need for NFS to
``inform'' UFS so that UFS can use ``redundancy'' to ``correct the
error''.  UFS never hears anything, and doesn't have any redundancy.
NFS resends the data.  done.

iSCSI also has application-level CRC's, seperately enableable for
headers and data.  not sure what FC has.

It doesn't make any sense to me that some higher layer would call back
to the ZFS stack on the bottom, and tell it to twiddle with disks
because there's a problem with the network.

An idea Richard brought up months ago was ``protection domains,'' that
it might be good to expose ZFS checksums to higher levels to stretch a
single protection domain as far as possible upwards in the stack.

Application-level checksums also form a single protection domain, for
_reading_.  Suppose corruption happens in RAM or on the network (where
the ZFS backing store cannot detect it), while reading a gzip file on
an NFS client.  gzip will always warn you!  This is end-to-end, and
will warn you just as perfectly as hypothetical end-to-end
networkified-ZFS.  The problem: there's no way for gzip to ``retry''.
You can run gunzip again, but it will just fail again and again
because the file with network-induced errors is cached on the NFS
client.  It's the ``cached badness'' problem Richard alluded to.  You
would have to reboot the NFS client to clear its read cache, then try
gunzip again.  This is probably good enough in practice, but it sounds
like there's room for improvement.

It's irrelevant in this scenario that the lower ZFS has
``redundancy''.  All you have to do to fix the problem is resend the
read over the network.  What would be nice to have, that we don't
have, is a way of keeping ZFS block checksums attached to the data as
it travels over the network until it reaches the
something-like-an-NFS-client.  Each part of the stack that caches data
could be trained to either (1) validate ZFS block checksums, or (2) to
obey ``read no-cache'' commands passed down from the layer above.  In
the application-level gzip example, gzip has no way of doing (2), so
extending the protection domain upward rather than pushing
cache-flushing obedience downward seems more practical.

For writing, application-level checksums do NOT work at all, because
you would write corrupt data to the disk, and notice only later when
you read it back, when there's nothing you can do to fix it.  ZFS
redundancy will not help you here either, because you write corrupt
data redundantly!  With a single protection domain for writing, the
write would arrive at ZFS along with a never-regenerated checksum
wrapper-seal attached to it by the something-like-an-NFS-client.  Just
before ZFS sends the write to the disk driver, ZFS would crack the
protection domain open, validate the checksum, reblock the write, and
send it to disk with a new checksum.  (so, ``single protection
domain'' is really a single domain for reads, and two protection
domains for write) If the checksum does not match, ZFS must convince
the writing client to resend---in the write direction I think cached
bad data will be less of a problem.

I think a single protection domain, rather than the currently
best-obtainable which is sliced domains where the slices butt up
against each other as closely as possible, is an idea with merit.  but
it doesn't have anything whatsoever to do with the fact ZFS stores
things redundantly on the platters.  The whole thing would have just
as much merit, and fix the new problem classes it addresses just as
frequently!, for single-disk vdev's as for redundant vdev's.

gm Of course, if something in the data path consistently corrupts
gm the data regardless of its origin, it won't be able to correct
gm the error.

TCP does this all the time, right?  see, watch this:  +++ATH0

:)

that aside, your idea of ``the error'' seems too general, like the
annoying marketing slicks with the ``healing'' and ``correcting''
stuff.  stored, transmitted, cached errors are relevantly different,
which also means corruption in the read and write directions are
different.


pgpsQ6XuAcq9p.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-05 Thread Toby Thain


On 5-Mar-09, at 2:03 PM, Miles Nordin wrote:


gm == Gary Mills mi...@cc.umanitoba.ca writes:


gm There are many different components that could contribute to
gm such errors.

yes of course.

gm Since only the lower ZFS has data redundancy, only it can
gm correct the error.

um, no?
...
For writing, application-level checksums do NOT work at all, because
you would write corrupt data to the disk, and notice only later when
you read it back, when there's nothing you can do to fix it.


Right, it would have to be combined with an always read-back policy  
in the application...?


--Toby


  ZFS
redundancy will not help you here either, because you write corrupt
data redundantly!  With a single protection domain for writing, the
write would arrive at ZFS along with a never-regenerated checksum
wrapper-seal attached to it by the something-like-an-NFS-client.  Just
before ZFS sends the write to the disk driver, ZFS would crack the
protection domain open, validate the checksum, reblock the write, and
send it to disk with a new checksum.  (so, ``single protection
domain'' is really a single domain for reads, and two protection
domains for write) If the checksum does not match, ZFS must convince
the writing client to resend---in the write direction I think cached
bad data will be less of a problem.
...

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-05 Thread Miles Nordin
 tt == Toby Thain t...@telegraphics.com.au writes:

 c For writing, application-level checksums do NOT work at all,
 c because you would write corrupt data to the disk, and notice
 c only later when you read it back

tt Right, it would have to be combined with an always read-back
tt policy in the application...?

aside from being wasteful, that won't work because of ``cached
goodness''. :)

It's allegedly common for programs to read what they've just written
during normal operation, so various caches usually consider whatever's
written as ``recently used'' for purposes of expiring it from the read
cache.  If corruption happens below a cache (ex. in the NFS client),
as it probably will, you won't detect it by read-after-write.


pgp1YJZ1OzT10.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Gary Mills
On Tue, Mar 03, 2009 at 11:35:40PM +0200, C. Bergström wrote:
 
 Here's more or less what I've collected...
 
[..]
   10) Did I miss something..

I suppose my RFE for two-level ZFS should be included, unless nobody
intends to attach a ZFS file server to a SAN with ZFS on application
servers.

-- 
-Gary Mills--Unix Support--U of M Academic Computing and Networking-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Tim
On Wed, Mar 4, 2009 at 9:52 AM, Gary Mills mi...@cc.umanitoba.ca wrote:

 On Tue, Mar 03, 2009 at 11:35:40PM +0200, C. Bergström wrote:
 
  Here's more or less what I've collected...
 
 [..]
10) Did I miss something..

 I suppose my RFE for two-level ZFS should be included, unless nobody
 intends to attach a ZFS file server to a SAN with ZFS on application
 servers.

 --
 -Gary Mills--Unix Support--U of M Academic Computing and
 Networking-


Although it seems your idea fell on a lot of deaf ears, I personally think
it's a great one.  It would give people a reason to use Solaris as their
server OS of choice as well as their fileserver/storage appliance.  It would
also give them something NetApp doesn't have, instead of a bunch of me-too's
(no, we don't need to start a flame-fest, it's my opinion, years of reading
this list haven't changed it).  As it stands, either you put your workload
where the disks are, or you're wasting disk doing raid twice.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Miles Nordin
 nw == Nicolas Williams nicolas.willi...@sun.com writes:

nw IIRC Jeff Bonwick has said on this list that ubberblock
nw rollback on import is now his higher priority.  So working on
nw (8) would be duplication of effort.

well...if your recovery tool worked by using an older ueberblock.  but
Anton had another idea to write one that treats the pool being
recovered as read-only.  For example when someone drinks too much
kool-aid and sets 'copies=2' on a pool with two unredundant vdev's,
and then loses a disk and says ``ok now where's my data please,''
Anton's copy-out rescuing tool might help.


pgpF5c4uyQEqN.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Miles Nordin
 gm == Gary Mills mi...@cc.umanitoba.ca writes:

gm I suppose my RFE for two-level ZFS should be included,

Not that my opinion counts for much, but I wasn't deaf to it---I did
respond.

I thought it was kind of based on mistaken understanding.  It included
this strangeness of the upper ZFS ``informing'' the lower one when
corruption had occured on the network, and the lower ZFS was supposed
to do something with the physical disks...to resolve corruption on the
network?  why?  IIRC several others pointed out the same bogosity.

It makes slightly more sense in the write direction than the read
direction maybe, but I still don't fully get the plan.  It is a new
protocol to replace iSCSI?  or NFS?  or, what?  Is it a re-invention
of pNFS or Lustre, but with more work since you're starting from zero,
and less architectural foresight?


pgpYG81ClxWHA.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Bob Friesenhahn
I don't know if anyone has noticed that the topic is google summer of 
code.  There is only so much that a starving college student can 
accomplish from a dead-start in 1-1/2 months.  The ZFS equivalent of 
eliminating world hunger is not among the tasks which may be 
reasonably accomplished, yet tasks at this level of effort is all that 
I have seen mentioned here.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread C. Bergström

Bob Friesenhahn wrote:
I don't know if anyone has noticed that the topic is google summer of 
code.  There is only so much that a starving college student can 
accomplish from a dead-start in 1-1/2 months.  The ZFS equivalent of 
eliminating world hunger is not among the tasks which may be 
reasonably accomplished, yet tasks at this level of effort is all that 
I have seen mentioned here.
May I interject a bit.. I'm silently collecting this task list and even 
outside of gsoc may help try to arrange it from a community 
perspective.  Of course this will be volunteer based unless /we/ get a 
sponsor or sun beats /us/ to it.  So all the crazy ideas welcome..


./C
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Toby Thain


On 4-Mar-09, at 1:28 PM, Bob Friesenhahn wrote:

I don't know if anyone has noticed that the topic is google summer  
of code.  There is only so much that a starving college student  
can accomplish from a dead-start in 1-1/2 months.  The ZFS  
equivalent of eliminating world hunger is not among the tasks which  
may be reasonably accomplished, yet tasks at this level of effort  
is all that I have seen mentioned here.


Item (4) (CLI sudden death warnings) is not of world hunger scope.
Item (5) (GUI) isn't rocket science.
Item (3) (more compressors) is a bit harder but do-able.

The rest get into serious engineering, agreed. :)

--T



Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/ 
bfriesen/

GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Wes Felter

C. Bergström wrote:


  10) Did I miss something..


T10 DIF support in zvols
T10 UNMAP/thin provisioning support in zvols
proportional scheduling for storage performance
slog and L2ARC on the same SSD

These are probably difficult but hopefully not world hunger level.

Wes Felter - wes...@felter.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Nicolas Williams
On Wed, Mar 04, 2009 at 02:16:53PM -0600, Wes Felter wrote:
 T10 UNMAP/thin provisioning support in zvols

That's probably simple enough, and sufficiently valuable too.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Julius Roberts
2009/3/4 Tim t...@tcsac.net:
 I know plenty of home users would like the ability to add a single disk to
 a raid-z vdev in order to grow a disk at a time.

+1 for that.

-- 
Kind regards, Jules
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread David Dyer-Bennet

On Wed, March 4, 2009 16:53, Julius Roberts wrote:
 2009/3/4 Tim t...@tcsac.net:
 I know plenty of home users would like the ability to add a single
 disk to
 a raid-z vdev in order to grow a disk at a time.

 +1 for that.

In theory I'd like that a lot.

I'm now committed to two two-disk mirrors precisely because I couldn't
grow a RAIDZ (and would have to replace all three+ disks to expand it that
way); I started with one mirror, added a second when I needed it, and when
I need to increase again I will replace the disks in one of the mirrors
one at a time letting it resilver (yes, I know I'm not covered by
redundancy during that process; I'll do it when my backups are up-to-date.
 It's a home server and I can have it out of service during the resilver,
or in no guarantee of saving changes mode.).

And on consideration, depending on the expected lifespan of the system, I
might be better off sticking with the mirror strategy anyway.  Of course
RAIDZ is more cost-efficient, but disks are cheap.


-- 
David Dyer-Bennet, d...@dd-b.net; http://dd-b.net/
Snapshots: http://dd-b.net/dd-b/SnapshotAlbum/data/
Photos: http://dd-b.net/photography/gallery/
Dragaera: http://dragaera.info

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Gary Mills
On Wed, Mar 04, 2009 at 01:20:42PM -0500, Miles Nordin wrote:
  gm == Gary Mills mi...@cc.umanitoba.ca writes:
 
 gm I suppose my RFE for two-level ZFS should be included,
 
 Not that my opinion counts for much, but I wasn't deaf to it---I did
 respond.

I appreciate that.

 I thought it was kind of based on mistaken understanding.  It included
 this strangeness of the upper ZFS ``informing'' the lower one when
 corruption had occured on the network, and the lower ZFS was supposed
 to do something with the physical disks...to resolve corruption on the
 network?  why?  IIRC several others pointed out the same bogosity.

It's a simply a consequence of ZFS's end-to-end error detection.
There are many different components that could contribute to such
errors.  Since only the lower ZFS has data redundancy, only it can
correct the error.  Of course, if something in the data path
consistently corrupts the data regardless of its origin, it won't be
able to correct the error.  The same thing can happen in the simple
case, with one ZFS over physical disks.

 It makes slightly more sense in the write direction than the read
 direction maybe, but I still don't fully get the plan.  It is a new
 protocol to replace iSCSI?  or NFS?  or, what?  Is it a re-invention
 of pNFS or Lustre, but with more work since you're starting from zero,
 and less architectural foresight?

I deliberately did not specify the protocol to keep the concept
general.  Anything that works and solves the problem would be good.

-- 
-Gary Mills--Unix Support--U of M Academic Computing and Networking-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Toby Thain


On 4-Mar-09, at 7:35 PM, Gary Mills wrote:


On Wed, Mar 04, 2009 at 01:20:42PM -0500, Miles Nordin wrote:

gm == Gary Mills mi...@cc.umanitoba.ca writes:


gm I suppose my RFE for two-level ZFS should be included,

Not that my opinion counts for much, but I wasn't deaf to it---I did
respond.


I appreciate that.

I thought it was kind of based on mistaken understanding.  It  
included

this strangeness of the upper ZFS ``informing'' the lower one when
corruption had occured on the network, and the lower ZFS was supposed
to do something with the physical disks...to resolve corruption on  
the

network?  why?  IIRC several others pointed out the same bogosity.


It's a simply a consequence of ZFS's end-to-end error detection.
There are many different components that could contribute to such
errors.  Since only the lower ZFS has data redundancy, only it can
correct the error.


Why aren't application level checksums the answer to this problem?

--Toby


Of course, if something in the data path
consistently corrupts the data regardless of its origin, it won't be
able to correct the error.  The same thing can happen in the simple
case, with one ZFS over physical disks.


It makes slightly more sense in the write direction than the read
direction maybe, but I still don't fully get the plan.  It is a new
protocol to replace iSCSI?  or NFS?  or, what?  Is it a re-invention
of pNFS or Lustre, but with more work since you're starting from  
zero,

and less architectural foresight?


I deliberately did not specify the protocol to keep the concept
general.  Anything that works and solves the problem would be good.

--
-Gary Mills--Unix Support--U of M Academic Computing and  
Networking-

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Dave

Gary Mills wrote:

On Wed, Mar 04, 2009 at 01:20:42PM -0500, Miles Nordin wrote:

gm == Gary Mills mi...@cc.umanitoba.ca writes:

gm I suppose my RFE for two-level ZFS should be included,

Not that my opinion counts for much, but I wasn't deaf to it---I did
respond.


I appreciate that.


I thought it was kind of based on mistaken understanding.  It included
this strangeness of the upper ZFS ``informing'' the lower one when
corruption had occured on the network, and the lower ZFS was supposed
to do something with the physical disks...to resolve corruption on the
network?  why?  IIRC several others pointed out the same bogosity.


It's a simply a consequence of ZFS's end-to-end error detection.
There are many different components that could contribute to such
errors.  Since only the lower ZFS has data redundancy, only it can
correct the error.  Of course, if something in the data path
consistently corrupts the data regardless of its origin, it won't be
able to correct the error.  The same thing can happen in the simple
case, with one ZFS over physical disks.


I would argue against building this into ZFS. Any corruption happening 
on the wire should not be the responsibility of ZFS. If you want to make 
sure your data is not corrupted over the wire, use IPSec. If you want to 
prevent corruption in RAM, use ECC sticks, etc.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Gary Mills
On Wed, Mar 04, 2009 at 06:31:59PM -0700, Dave wrote:
 Gary Mills wrote:
 On Wed, Mar 04, 2009 at 01:20:42PM -0500, Miles Nordin wrote:
 gm == Gary Mills mi...@cc.umanitoba.ca writes:
 gm I suppose my RFE for two-level ZFS should be included,
 
 It's a simply a consequence of ZFS's end-to-end error detection.
 There are many different components that could contribute to such
 errors.  Since only the lower ZFS has data redundancy, only it can
 correct the error.  Of course, if something in the data path
 consistently corrupts the data regardless of its origin, it won't be
 able to correct the error.  The same thing can happen in the simple
 case, with one ZFS over physical disks.
 
 I would argue against building this into ZFS. Any corruption happening 
 on the wire should not be the responsibility of ZFS. If you want to make 
 sure your data is not corrupted over the wire, use IPSec. If you want to 
 prevent corruption in RAM, use ECC sticks, etc.

But what if the `wire' is a SCSI bus?  Would you want ZFS to do error
correction in that case?  There are many possible wires.  Every
component does its own error checking of some sort, but in its own
domain.  This brings us back to end-to-end error checking again. Since
we are designing a filesystem, that's where the reliability should
reside.

-- 
-Gary Mills--Unix Support--U of M Academic Computing and Networking-
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-04 Thread Dave



Gary Mills wrote:

On Wed, Mar 04, 2009 at 06:31:59PM -0700, Dave wrote:

Gary Mills wrote:

On Wed, Mar 04, 2009 at 01:20:42PM -0500, Miles Nordin wrote:

gm == Gary Mills mi...@cc.umanitoba.ca writes:

   gm I suppose my RFE for two-level ZFS should be included,

It's a simply a consequence of ZFS's end-to-end error detection.
There are many different components that could contribute to such
errors.  Since only the lower ZFS has data redundancy, only it can
correct the error.  Of course, if something in the data path
consistently corrupts the data regardless of its origin, it won't be
able to correct the error.  The same thing can happen in the simple
case, with one ZFS over physical disks.
I would argue against building this into ZFS. Any corruption happening 
on the wire should not be the responsibility of ZFS. If you want to make 
sure your data is not corrupted over the wire, use IPSec. If you want to 
prevent corruption in RAM, use ECC sticks, etc.


But what if the `wire' is a SCSI bus?  Would you want ZFS to do error
correction in that case?  There are many possible wires.  Every
component does its own error checking of some sort, but in its own
domain.  This brings us back to end-to-end error checking again. Since
we are designing a filesystem, that's where the reliability should
reside.



ZFS can't eliminate or prevent all errors. You should have a split 
backplane/multiple controllers and a minimum 2-way mirror if you're 
concerned about this from a local component POV.


Same with iSCSI. I run a minimum 2-way mirror from my ZFS server from 2 
different NICs, over 2 gigabit switches w/trunking to two different disk 
shelves for this reason. I do not stack ZFS layers, since it degrades 
performance and really doesn't provide any benefit.


What's your reason for stacking zpools? I can't recall the original 
argument for this.


--
Dave
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread C. Bergström


For reasons which I don't care about Sun may not apply to be a gsoc 
organization this year.  However, I'm not discouraged from trying to 
propose some exciting zfs related ideas.  On/off list feel free to send 
your vote, let me know if you can mentor or if you know a company that 
could use it.


Here's more or less what I've collected...

  1) Excess ditto block removing + other green-bytes zfs+ features - 
*open source* (very hard.. can't be done in two months)
  2) raidz boot support (planning phase and suitable student already 
found. could use more docs/info for proposal)

  3) Additional zfs compression choices (good for archiving non-text files?
  4) zfs cli interface to add safety checks  (save your butt from 
deleting a pool worth more than your job)

  5) Web or gui based admin interface
  6) zfs defrag (was mentioned by someone working around petabytes of 
data..)
  7) vdev evacuation as an upgrade path (which may depend or take 
advantage of zfs resize/shrink code)

  8) zfs restore/repair tools (being worked on already?)
  9) Timeslider ported to kde4.2 ( *cough* couldn't resist, but put 
this on the list)

  10) Did I miss something..

#2 Currently planning and collecting as much information for the 
proposal as possible.  Today all ufs + solaris grub2 issues were 
resolved and will likely be committed to upstream soon.  There is a one 
liner fix in the solaris kernel also needed, but that can be binary 
hacked worst case.


#5/9 This also may be possible for an outside project.. either web 
showcase or tighter desktop integration..


The rest may just be too difficult in a two month period, not something 
which can go upstream or not enough time to really plan well enough..  
Even if this isn't done for gsoc it may still be possible for the 
community to pursue some of these..


To be a mentor will most likely require answering daily/weekly technical 
questions, ideally being on irc and having patience.  On top of this 
I'll be available to help as much as technically possible, keep the 
student motivated and the projects on schedule.


./Christopher


#ospkg irc.freenode.net - (Mostly OpenSolaris development rambling)


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Richard Elling

C. Bergström wrote:


For reasons which I don't care about Sun may not apply to be a gsoc 
organization this year.  However, I'm not discouraged from trying to 
propose some exciting zfs related ideas.  On/off list feel free to 
send your vote, let me know if you can mentor or if you know a company 
that could use it.


Here's more or less what I've collected...

  1) Excess ditto block removing + other green-bytes zfs+ features - 
*open source* (very hard.. can't be done in two months)
  2) raidz boot support (planning phase and suitable student already 
found. could use more docs/info for proposal)
  3) Additional zfs compression choices (good for archiving non-text 
files?
  4) zfs cli interface to add safety checks  (save your butt from 
deleting a pool worth more than your job)

  5) Web or gui based admin interface


FWIW, I just took at look at the BUI in b108 and it seems to have
garnered some love since the last time I looked at it (a year ago?)
I encourage folks to take a fresh look at it.
   https://localhost:6789

-- richard

  6) zfs defrag (was mentioned by someone working around petabytes of 
data..)
  7) vdev evacuation as an upgrade path (which may depend or take 
advantage of zfs resize/shrink code)

  8) zfs restore/repair tools (being worked on already?)
  9) Timeslider ported to kde4.2 ( *cough* couldn't resist, but put 
this on the list)

  10) Did I miss something..

#2 Currently planning and collecting as much information for the 
proposal as possible.  Today all ufs + solaris grub2 issues were 
resolved and will likely be committed to upstream soon.  There is a 
one liner fix in the solaris kernel also needed, but that can be 
binary hacked worst case.


#5/9 This also may be possible for an outside project.. either web 
showcase or tighter desktop integration..


The rest may just be too difficult in a two month period, not 
something which can go upstream or not enough time to really plan well 
enough..  Even if this isn't done for gsoc it may still be possible 
for the community to pursue some of these..


To be a mentor will most likely require answering daily/weekly 
technical questions, ideally being on irc and having patience.  On top 
of this I'll be available to help as much as technically possible, 
keep the student motivated and the projects on schedule.


./Christopher


#ospkg irc.freenode.net - (Mostly OpenSolaris development rambling)


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Nicolas Williams
On Tue, Mar 03, 2009 at 11:35:40PM +0200, C. Bergström wrote:
   7) vdev evacuation as an upgrade path (which may depend or take 
 advantage of zfs resize/shrink code)

IIRC Matt Ahrens has said on this list that vdev evacuation/pool
shrinking is being worked.  So (7) would be duplication of effort.

   8) zfs restore/repair tools (being worked on already?)

IIRC Jeff Bonwick has said on this list that ubberblock rollback on
import is now his higher priority.  So working on (8) would be
duplication of effort.

   1) Excess ditto block removing + other green-bytes zfs+ features - 
 *open source* (very hard.. can't be done in two months)

Using the new block pointer re-write code you might be able to deal with
re-creating blocks with more/fewer ditto copies (and compression, ...)
with incremental effort.  But ask Matt Ahrens.

   6) zfs defrag (was mentioned by someone working around petabytes of 
 data..)

(6) probably depends on the new block pointer re-write code as well.
But (6) may also be implied in vdev evac/pool shrink, so it may be
duplication of effort.

Nico
-- 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Blake
When I go here:
http://opensolaris.org/os/project/isns/bui

I get an error.  Where are you getting BUI from?


On Tue, Mar 3, 2009 at 5:16 PM, Richard Elling richard.ell...@gmail.comwrote:


 FWIW, I just took at look at the BUI in b108 and it seems to have
 garnered some love since the last time I looked at it (a year ago?)
 I encourage folks to take a fresh look at it.
   https://localhost:6789

 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Maurice Volaski

   10) Did I miss something..


Somehow, what I posted on the web forum didn't make it to the mailing 
list digest...


How about implementing dedup? This has been listed as an RFE for 
almost a year, http://bugs.opensolaris.org/view_bug.do?bug_id=6677093 
and discussed here, 
http://www.opensolaris.org/jive/thread.jspa?messageID=256373.

--

Maurice Volaski, mvola...@aecom.yu.edu
Computing Support, Rose F. Kennedy Center
Albert Einstein College of Medicine of Yeshiva University
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Tim
On Tue, Mar 3, 2009 at 3:35 PM, C. Bergström cbergst...@netsyncro.comwrote:


 For reasons which I don't care about Sun may not apply to be a gsoc
 organization this year.  However, I'm not discouraged from trying to propose
 some exciting zfs related ideas.  On/off list feel free to send your vote,
 let me know if you can mentor or if you know a company that could use it.

 Here's more or less what I've collected...

  1) Excess ditto block removing + other green-bytes zfs+ features - *open
 source* (very hard.. can't be done in two months)
  2) raidz boot support (planning phase and suitable student already found.
 could use more docs/info for proposal)
  3) Additional zfs compression choices (good for archiving non-text files?
  4) zfs cli interface to add safety checks  (save your butt from deleting a
 pool worth more than your job)
  5) Web or gui based admin interface
  6) zfs defrag (was mentioned by someone working around petabytes of
 data..)
  7) vdev evacuation as an upgrade path (which may depend or take advantage
 of zfs resize/shrink code)
  8) zfs restore/repair tools (being worked on already?)
  9) Timeslider ported to kde4.2 ( *cough* couldn't resist, but put this on
 the list)
  10) Did I miss something..

 #2 Currently planning and collecting as much information for the proposal
 as possible.  Today all ufs + solaris grub2 issues were resolved and will
 likely be committed to upstream soon.  There is a one liner fix in the
 solaris kernel also needed, but that can be binary hacked worst case.

 #5/9 This also may be possible for an outside project.. either web showcase
 or tighter desktop integration..

 The rest may just be too difficult in a two month period, not something
 which can go upstream or not enough time to really plan well enough..  Even
 if this isn't done for gsoc it may still be possible for the community to
 pursue some of these..

 To be a mentor will most likely require answering daily/weekly technical
 questions, ideally being on irc and having patience.  On top of this I'll be
 available to help as much as technically possible, keep the student
 motivated and the projects on schedule.

 ./Christopher



I know plenty of home users would like the ability to add a single disk to
a raid-z vdev in order to grow a disk at a time.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Richard Elling

Blake wrote:

When I go here:

http://opensolaris.org/os/project/isns/bui

I get an error. Â Where are you getting BUI from?


The BUI is in webconsole which is available on your local machine at
port 6789
   https://localhost:6798

If you want to access it remotely, you'll need to change the configuration
as documented
   http://docs.sun.com/app/docs/doc/817-1985/gdhgt?a=view

-- richard


On Tue, Mar 3, 2009 at 5:16 PM, Richard Elling 
richard.ell...@gmail.com mailto:richard.ell...@gmail.com wrote:



FWIW, I just took at look at the BUI in b108 and it seems to have
garnered some love since the last time I looked at it (a year ago?)
I encourage folks to take a fresh look at it.
  https://localhost:6789

-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs related google summer of code ideas - your vote

2009-03-03 Thread Blake
That's what I thought you meant, and I got excited thinking that you were
talking about OpenSolaris :)
I'll see about getting the new packages and trying them out.



On Tue, Mar 3, 2009 at 8:36 PM, Richard Elling richard.ell...@gmail.comwrote:

 Blake wrote:

 When I go here:

 http://opensolaris.org/os/project/isns/bui

 I get an error. Â Where are you getting BUI from?


 The BUI is in webconsole which is available on your local machine at
 port 6789
   https://localhost:6798

 If you want to access it remotely, you'll need to change the configuration
 as documented
   http://docs.sun.com/app/docs/doc/817-1985/gdhgt?a=view

 -- richard


 On Tue, Mar 3, 2009 at 5:16 PM, Richard Elling 
 richard.ell...@gmail.commailto:
 richard.ell...@gmail.com wrote:


FWIW, I just took at look at the BUI in b108 and it seems to have
garnered some love since the last time I looked at it (a year ago?)
I encourage folks to take a fresh look at it.
  https://localhost:6789

-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss