Re[6]: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-28 Thread Robert Milkowski
Hello Erik,

Wednesday, February 28, 2007, 12:55:24 AM, you wrote:

ET> Honestly, no, I don't consider UFS a modern file system. :-)

ET> It's just not in the same class as JFS for AIX, xfs for IRIX, or even
ET> VxFS.

The point is that fsck was due to an array corrupting data.
IMHO it would hit JFS, XFS or VxFS as bad as UFS if not worse.



-- 
Best regards,
 Robertmailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Rob Logan

> With modern journalling filesystems, I've never had to fsck anything or
> run a filesystem repair. Ever.  On any of my SAN stuff.

you will.. even if the SAN is perfect, you will hit
bugs in the filesystem code.. from lots of rsync hard
links or like this one from raidtools last week:

Feb  9 05:38:39 orbit kernel: mptbase: ioc2: IOCStatus(0x0043): SCSI Device Not 
There
Feb  9 05:38:39 orbit kernel: md: write_disk_sb failed for device sdp1
Feb  9 05:38:39 orbit kernel: md: errors occurred during superblock update, 
repeating

Feb  9 05:39:01 orbit kernel: raid6: Disk failure on sdp1, disabling device. 
Operation continuing on 13 devices
Feb  9 05:39:09 orbit kernel: mptscsi: ioc2: attempting task abort! 
(sc=cb17c800)
Feb  9 05:39:10 orbit kernel: RAID6 conf printout:
Feb  9 05:39:10 orbit kernel:  --- rd:14 wd:13 fd:1

Feb  9 05:44:37 orbit kernel: EXT3-fs error (device dm-0): ext3_readdir: bad 
entry in directory #10484: rec_len %$
Feb  9 05:44:37 orbit kernel: Aborting journal on device dm-0.
Feb  9 05:44:37 orbit kernel: ext3_abort called.
Feb  9 05:44:37 orbit kernel: EXT3-fs error (device dm-0): 
ext3_journal_start_sb: Detected aborted journal
Feb  9 05:44:37 orbit kernel: Remounting filesystem read-only
Feb  9 05:44:37 orbit kernel: attempt to access beyond end of device
Feb  9 05:44:44 orbit kernel: oom-killer: gfp_mask=0xd0



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: Re[4]: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Erik Trimble
Honestly, no, I don't consider UFS a modern file system. :-)

It's just not in the same class as JFS for AIX, xfs for IRIX, or even
VxFS.

-Erik




On Wed, 2007-02-28 at 00:40 +0100, Robert Milkowski wrote:
> Hello Erik,
> 
> Tuesday, February 27, 2007, 5:47:42 PM, you wrote:
> 
> ET> 
> 
> 
> ET> The answer is:   insufficient data.
> 
> 
> ET> With modern journalling filesystems, I've never had to fsck anything or
> ET> run a filesystem repair. Ever.  On any of my SAN stuff. 
> 
> I'm not sure if you consider UFS in S10 as a modern journalling
> filesystem but in case you do:
> 
> Feb 13 12:03:16  ufs: [ID 879645 kern.notice] NOTICE: /opt/d1635: 
> unexpected free inode 54305084, run fsck(1M) -o f
> 
> This file system is on a medium large array (IBM) in a SAN
> environment.
> 
> 
> 
-- 
Erik Trimble
Java System Support
Mailstop:  usca14-102
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Jason J. W. Williams

Hi Przemol,

I think migration is a really important feature...think I said that...
;-)  SAN/RAID is not awful...frankly there's not been better solution
(outside of NetApp's WAFL) till ZFS. SAN/RAID just has its own
reliability issues you accept unless you don't have toZFS :-)

-J

On 2/27/07, [EMAIL PROTECTED] <[EMAIL PROTECTED]> wrote:

On Thu, Feb 22, 2007 at 12:21:50PM -0700, Jason J. W. Williams wrote:
> Hi Przemol,
>
> I think Casper had a good point bringing up the data integrity
> features when using ZFS for RAID. Big companies do a lot of things
> "just because that's the certified way" that end up biting them in the
> rear. Trusting your SAN arrays is one of them. That all being said,
> the need to do migrations is a very valid concern.

Jason,

I don't claim that SAN/RAID solutions are the best and don't have any
mistakes/failures/problems. But if SAN/RAID is so bad why companies
using them survive ?

Imagine also that some company is using SAN/RAID for a few years
and doesn't have any problems (or once a few months). Also from time to
time they need to migrate between arrays (for whatever reason). Now you come 
and say
that they have unreliable SAN/RAID and you offer something new (ZFS)
which is going to make it much more reliable but migration to another array
will be painfull. What do you think what they choose ?

BTW: I am a fan of ZFS. :-)

przemol

--
Ustawiaj rekordy DNS dla swojej domeny >>
http://link.interia.pl/f1a1a

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re[4]: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Robert Milkowski
Hello Erik,

Tuesday, February 27, 2007, 5:47:42 PM, you wrote:

ET> 


ET> The answer is:   insufficient data.


ET> With modern journalling filesystems, I've never had to fsck anything or
ET> run a filesystem repair. Ever.  On any of my SAN stuff. 

I'm not sure if you consider UFS in S10 as a modern journalling
filesystem but in case you do:

Feb 13 12:03:16  ufs: [ID 879645 kern.notice] NOTICE: /opt/d1635: 
unexpected free inode 54305084, run fsck(1M) -o f

This file system is on a medium large array (IBM) in a SAN
environment.



-- 
Best regards,
 Robertmailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Richard Elling

[EMAIL PROTECTED] wrote:

Is the "true situation" really so bad ?


The failure mode is silent error.  By definition, it is hard to
count silent errors.  What ZFS does is improve the detection of
silent errors by a rather considerable margin.  So, what we are
seeing is that suddenly people are seeing errors that they didn't
see before (or do you "hear" silent errors? ;-).  That has been
surprising and leads some of us to recommend ZFS no matter what
your storage looks like, even if silent error detection is the
only benefit.
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: Re[2]: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Erik Trimble



The answer is:   insufficient data.


With modern journalling filesystems, I've never had to fsck anything or
run a filesystem repair. Ever.  On any of my SAN stuff. 

The sole place I've run into filesystem corruption in the traditional
sense is with faulty hardware controllers; and, I'm not even sure ZFS
could recover from those situations, though less dire ones where the
controllers are merely emitting slightly wonky problems certainly would
be within ZFS's ability to fix, vice the inability of a SAN to determine
that the data was bad.


That said, the primary issue here is that nobody really has any idea
about silent corruption - that is, blocks which change value, but are
data, not filesystem-relevant. Bit flips and all.  Realistically, the
only way previous to ZFS to detect this was to do bit-wise comparisons
against backups, which becomes practically impossible on an active data
set.  

SAN/RAID equipment still has a very considerable place over JBODs in
most large-scale places, particularly in areas of configuration
flexibility, security, and management.  That said, I think we're arguing
at cross-purposes:   the real solution for most enterprise customers is
SAN + ZFS, not either just by itself.



-- 
Erik Trimble
Java System Support
Mailstop:  usca14-102
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re[2]: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Robert Milkowski
Hello przemolicc,

Tuesday, February 27, 2007, 11:28:59 AM, you wrote:

ppf> On Tue, Feb 27, 2007 at 08:29:04PM +1100, Shawn Walker wrote:
>> On 27/02/07, [EMAIL PROTECTED] <[EMAIL PROTECTED]> wrote:
>> >On Thu, Feb 22, 2007 at 12:21:50PM -0700, Jason J. W. Williams wrote:
>> >> Hi Przemol,
>> >>
>> >> I think Casper had a good point bringing up the data integrity
>> >> features when using ZFS for RAID. Big companies do a lot of things
>> >> "just because that's the certified way" that end up biting them in the
>> >> rear. Trusting your SAN arrays is one of them. That all being said,
>> >> the need to do migrations is a very valid concern.
>> >
>> >Jason,
>> >
>> >I don't claim that SAN/RAID solutions are the best and don't have any
>> >mistakes/failures/problems. But if SAN/RAID is so bad why companies
>> >using them survive ?
>> 
>> I think he was trying to say that people that believe that those
>> solutions are reliable just because they are based on SAN/RAID
>> technology and are not aware of the true situation surrounding them.

ppf> Is the "true situation" really so bad ?

ppf> My feeling was that he was trying to say that there is no SAN/RAID
ppf> solution without data integrity problem. Is it really true ?
ppf> Does anybody have any paper (*) about percentage of problems in SAN/RAID
ppf> because of data integrity ? Is it 5 % ? Or 30 % ? Or maybe 60 % ?

ppf> (*) Maybe such paper/report should be a start point for our discussion.

See http://sunsolve.sun.com/search/document.do?assetkey=1-26-102815-1

as one example. This is entry level array but still such things
happens. I do also observer similar problems with IBM's array
(larger).

It's just that people are used to fsck from time to time not really
knowing why and in many cases they do not realize that their data is
not exactly what they expect it to be.

However from my experience I must admit the problem is almost only
seen with SATA drives.

I had a problem with SCSI adapter which was
sending some warnings (driver) but still was passing IOs - it turned
out data were corrupted. Changing SCSI adapter solved the problem.
The point is that thanks to ZFS we caught the problem, replaced
bad card, did zpool scrub and everything was in perfect shape. No need
to resynchronize data, etc.

Another time I had a problem with FC array and lost some data but
there's no ZFS on it :(((

On all other arrays, jbods, etc. with SCSI and/or FC disks I haven't
seen (yet) checksum errors reported by ZFS.



-- 
Best regards,
 Robertmailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread przemolicc
On Tue, Feb 27, 2007 at 08:29:04PM +1100, Shawn Walker wrote:
> On 27/02/07, [EMAIL PROTECTED] <[EMAIL PROTECTED]> wrote:
> >On Thu, Feb 22, 2007 at 12:21:50PM -0700, Jason J. W. Williams wrote:
> >> Hi Przemol,
> >>
> >> I think Casper had a good point bringing up the data integrity
> >> features when using ZFS for RAID. Big companies do a lot of things
> >> "just because that's the certified way" that end up biting them in the
> >> rear. Trusting your SAN arrays is one of them. That all being said,
> >> the need to do migrations is a very valid concern.
> >
> >Jason,
> >
> >I don't claim that SAN/RAID solutions are the best and don't have any
> >mistakes/failures/problems. But if SAN/RAID is so bad why companies
> >using them survive ?
> 
> I think he was trying to say that people that believe that those
> solutions are reliable just because they are based on SAN/RAID
> technology and are not aware of the true situation surrounding them.

Is the "true situation" really so bad ?

My feeling was that he was trying to say that there is no SAN/RAID
solution without data integrity problem. Is it really true ?
Does anybody have any paper (*) about percentage of problems in SAN/RAID
because of data integrity ? Is it 5 % ? Or 30 % ? Or maybe 60 % ?

(*) Maybe such paper/report should be a start point for our discussion.

przemol

--
Gdy nie ma dzieci... - zobacz >> http://link.interia.pl/f19eb

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread Shawn Walker

On 27/02/07, [EMAIL PROTECTED] <[EMAIL PROTECTED]> wrote:

On Thu, Feb 22, 2007 at 12:21:50PM -0700, Jason J. W. Williams wrote:
> Hi Przemol,
>
> I think Casper had a good point bringing up the data integrity
> features when using ZFS for RAID. Big companies do a lot of things
> "just because that's the certified way" that end up biting them in the
> rear. Trusting your SAN arrays is one of them. That all being said,
> the need to do migrations is a very valid concern.

Jason,

I don't claim that SAN/RAID solutions are the best and don't have any
mistakes/failures/problems. But if SAN/RAID is so bad why companies
using them survive ?


I think he was trying to say that people that believe that those
solutions are reliable just because they are based on SAN/RAID
technology and are not aware of the true situation surrounding them.

--
"Less is only more where more is no good." --Frank Lloyd Wright

Shawn Walker, Software and Systems Analyst
[EMAIL PROTECTED] - http://binarycrusader.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-27 Thread przemolicc
On Thu, Feb 22, 2007 at 12:21:50PM -0700, Jason J. W. Williams wrote:
> Hi Przemol,
> 
> I think Casper had a good point bringing up the data integrity
> features when using ZFS for RAID. Big companies do a lot of things
> "just because that's the certified way" that end up biting them in the
> rear. Trusting your SAN arrays is one of them. That all being said,
> the need to do migrations is a very valid concern.

Jason,

I don't claim that SAN/RAID solutions are the best and don't have any
mistakes/failures/problems. But if SAN/RAID is so bad why companies
using them survive ?

Imagine also that some company is using SAN/RAID for a few years
and doesn't have any problems (or once a few months). Also from time to
time they need to migrate between arrays (for whatever reason). Now you come 
and say
that they have unreliable SAN/RAID and you offer something new (ZFS)
which is going to make it much more reliable but migration to another array
will be painfull. What do you think what they choose ?

BTW: I am a fan of ZFS. :-)

przemol

--
Ustawiaj rekordy DNS dla swojej domeny >> 
http://link.interia.pl/f1a1a

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-22 Thread Jason J. W. Williams

Hi Przemol,

I think Casper had a good point bringing up the data integrity
features when using ZFS for RAID. Big companies do a lot of things
"just because that's the certified way" that end up biting them in the
rear. Trusting your SAN arrays is one of them. That all being said,
the need to do migrations is a very valid concern.

Best Regards,
Jason

On 2/22/07, [EMAIL PROTECTED] <[EMAIL PROTECTED]> wrote:

On Wed, Feb 21, 2007 at 04:43:34PM +0100, [EMAIL PROTECTED] wrote:
>
> >I cannot let you say that.
> >Here in my company we are very interested in ZFS, but we do not care
> >about the RAID/mirror features, because we already have a SAN with
> >RAID-5 disks, and dual fabric connection to the hosts.
>
> But you understand that these underlying RAID mechanism give absolutely
> no guarantee about data integrity but only that some data was found were
> some (possibly other) data was written?  (RAID5 never verifies the
> checkum is correct on reads; it only uses it to reconstruct data when
> reads fail)

But you understand that he perhaps knows that but so far nothing wrong
happened [*] and migration is still very important feature for him ?

[*] almost every big company has its data center with SAN and FC
connections with RAID-5 or RAID-10 in their storage arrays
and they are treated as reliable

przemol

--
Wpadka w kosciele - zobacz >> http://link.interia.pl/f19ea

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-22 Thread przemolicc
On Wed, Feb 21, 2007 at 04:43:34PM +0100, [EMAIL PROTECTED] wrote:
> 
> >I cannot let you say that.
> >Here in my company we are very interested in ZFS, but we do not care
> >about the RAID/mirror features, because we already have a SAN with
> >RAID-5 disks, and dual fabric connection to the hosts.
> 
> But you understand that these underlying RAID mechanism give absolutely
> no guarantee about data integrity but only that some data was found were
> some (possibly other) data was written?  (RAID5 never verifies the
> checkum is correct on reads; it only uses it to reconstruct data when
> reads fail)

But you understand that he perhaps knows that but so far nothing wrong
happened [*] and migration is still very important feature for him ?

[*] almost every big company has its data center with SAN and FC
connections with RAID-5 or RAID-10 in their storage arrays
and they are treated as reliable

przemol

--
Wpadka w kosciele - zobacz >> http://link.interia.pl/f19ea

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Frank Cusack
On February 21, 2007 10:55:43 AM -0800 Richard Elling 
<[EMAIL PROTECTED]> wrote:

Valery Fouques wrote:

The ability to shrink a pool by removing devices is the only reason my
enterprise is not yet using ZFS, simply because it prevents us from
easily migrating storage.

That logic is totally bogus AFAIC. There are so many advantages to
running ZFS that denying yourself that opportunity is very short
sighted - especially when there are lots of ways of working around this
minor feature deficiency.


I cannot let you say that.
Here in my company we are very interested in ZFS, but we do not care
about the RAID/mirror features, because we already have a SAN with
RAID-5 disks, and dual fabric connection to the hosts.

We would have migrated already if we could simply migrate data from a
storage array to another (which we do more often than you might think).

Currently we use (and pay for) VXVM, here is how we do a migration:


But you describe VxVM feature, not a file system feature.


But in the context of zfs, this is appropriate.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Casper . Dik

>On February 21, 2007 4:43:34 PM +0100 [EMAIL PROTECTED] wrote:
>>
>>> I cannot let you say that.
>>> Here in my company we are very interested in ZFS, but we do not care
>>> about the RAID/mirror features, because we already have a SAN with
>>> RAID-5 disks, and dual fabric connection to the hosts.
>>
>> But you understand that these underlying RAID mechanism give absolutely
>> no guarantee about data integrity but only that some data was found were
>> some (possibly other) data was written?  (RAID5 never verifies the
>> checkum is correct on reads; it only uses it to reconstruct data when
>> reads fail)
>
>um, I thought smarter arrays did that these days.  Of course it's not
>end-to-end so the parity verification isn't as useful as it should be;
>gigo.

Generate extra I/O and verify parity, is that not something that may
be a problem in performance benchmarking?

For mirroring, a similar problem exists, of course.  ZFS reads from the
right side of the mirror and corrects the wrong side if it finds an
error.  RAIDs do not.

Casper
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Richard Elling

Valery Fouques wrote:

The ability to shrink a pool by removing devices is the only reason my
enterprise is not yet using ZFS, simply because it prevents us from
easily migrating storage.

That logic is totally bogus AFAIC. There are so many advantages to
running ZFS that denying yourself that opportunity is very short sighted -
especially when there are lots of ways of working around this minor
feature deficiency.


I cannot let you say that.
Here in my company we are very interested in ZFS, but we do not care about the 
RAID/mirror features, because we already have a SAN with RAID-5 disks, and dual 
fabric connection to the hosts.

We would have migrated already if we could simply migrate data from a storage 
array to another (which we do more often than you might think).

Currently we use (and pay for) VXVM, here is how we do a migration:


But you describe VxVM feature, not a file system feature.


1/ Allocate disks from the new array, visible by the host.
2/ Add the disks in the diskgroup.
3/ Run vxevac to evacuate data from "old" disks.
4/ Remove old disks from the DG.

If you explain how to do that with ZFS, no downtime, and new disks with 
different capacities, you're my hero ;-)


zpool replace old-disk new-disk
The caveat is that new-disk must be as big or bigger than old-disk.
This caveat is the core of the shrink "problem"
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Frank Cusack

On February 21, 2007 4:43:34 PM +0100 [EMAIL PROTECTED] wrote:



I cannot let you say that.
Here in my company we are very interested in ZFS, but we do not care
about the RAID/mirror features, because we already have a SAN with
RAID-5 disks, and dual fabric connection to the hosts.


But you understand that these underlying RAID mechanism give absolutely
no guarantee about data integrity but only that some data was found were
some (possibly other) data was written?  (RAID5 never verifies the
checkum is correct on reads; it only uses it to reconstruct data when
reads fail)


um, I thought smarter arrays did that these days.  Of course it's not
end-to-end so the parity verification isn't as useful as it should be;
gigo.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Casper . Dik

>I cannot let you say that.
>Here in my company we are very interested in ZFS, but we do not care
>about the RAID/mirror features, because we already have a SAN with
>RAID-5 disks, and dual fabric connection to the hosts.

But you understand that these underlying RAID mechanism give absolutely
no guarantee about data integrity but only that some data was found were
some (possibly other) data was written?  (RAID5 never verifies the
checkum is correct on reads; it only uses it to reconstruct data when
reads fail)

Casper
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Rich Teer
On Wed, 21 Feb 2007, Valery Fouques wrote:

> Here in my company we are very interested in ZFS, but we do not care
> about the RAID/mirror features, because we already have a SAN with
> RAID-5 disks, and dual fabric connection to the hosts.

... And presumably you've read the threads where ZFS has helped find
(and repair) corruption in such setups?

(But yeah, I agree the ability to shrink a pool is important.)

-- 
Rich Teer, SCSA, SCNA, SCSECA, OpenSolaris CAB member

President,
Rite Online Inc.

Voice: +1 (250) 979-1638
URL: http://www.rite-group.com/rich
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-02-21 Thread Valery Fouques
> > The ability to shrink a pool by removing devices is the only reason my
> > enterprise is not yet using ZFS, simply because it prevents us from
> > easily migrating storage.
> 
> That logic is totally bogus AFAIC. There are so many advantages to
> running ZFS that denying yourself that opportunity is very short sighted -
> especially when there are lots of ways of working around this minor
> feature deficiency.

I cannot let you say that.
Here in my company we are very interested in ZFS, but we do not care about the 
RAID/mirror features, because we already have a SAN with RAID-5 disks, and dual 
fabric connection to the hosts.

We would have migrated already if we could simply migrate data from a storage 
array to another (which we do more often than you might think).

Currently we use (and pay for) VXVM, here is how we do a migration:
1/ Allocate disks from the new array, visible by the host.
2/ Add the disks in the diskgroup.
3/ Run vxevac to evacuate data from "old" disks.
4/ Remove old disks from the DG.

If you explain how to do that with ZFS, no downtime, and new disks with 
different capacities, you're my hero ;-)
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-31 Thread Constantin Gonzalez
Hi,

I need to be a little bit more precise in how I formulate comments:

1. Yes, zpool remove is a desirable feature, no doubt about that.

2. Most of the cases where customers ask for "zpool remove" can be solved
   with zfs send/receive or with zpool replace. Think Pareto's 80-20 rule.

   2a. The cost of doing 2., including extra scratch storage space or scheduling
   related work into planned downtimes is smaller than the cost of not using
   ZFS at all.

   2b. Even in the remaining 20% of cases (figuratively speaking, YMMV) where
   zpool remove would be the only solution, I feel that the cost of
   sacrificing the extra storage space that would have become available
   through "zpool remove" is smaller than the cost of the project not
   benefitting from the rest of ZFS' features.

3. Bottom line: Everybody wants zpool remove as early as possible, but IMHO
   this is not an objective barrier to entry for ZFS.

Note my use of the word "objective". I do feel that we have to implement
zpool remove for subjective reasons, but that is a non technical matter.

Is this an agreeable summary of the situation?

Best regards,
   Constantin

-- 
Constantin GonzalezSun Microsystems GmbH, Germany
Platform Technology Group, Client Solutionshttp://www.sun.de/
Tel.: +49 89/4 60 08-25 91   http://blogs.sun.com/constantin/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Torrey McMahon

Al Hopper wrote:

On Fri, 26 Jan 2007, Torrey McMahon wrote:

  

Al Hopper wrote:


Now your accounting folks are going to be asking you to justify the
purchase of that hi-end SAN box and why you're not using ZFS
everywhere.  :)

Oh - and the accounting folks love it when you tell them there's no
ongoing cost of ownership - because Joe Screwdriver can swap out a failed
Seagate 500Gb SATA drive after he picks up a replacement from Frys on his
lunch break!
  

Because ZFS doesn't run everywhere.
Because most low end JBODs are "low end" for a reason. They aren't as
reliable, have crappy monitoring, etc.



Agreed.  There will never be one screwdriver that fits everything.  I was
simply trying to re-inforce my point.
  


It's a good point. We just need to make sure we don't forget that part. 
People love to pull email threads out of contextor google for that 
matter. ;)


  

Fix those two things when you get a chance. ;)



Have a good weekend Torrey (and zfs-discuss).


Same to you Al. (and zfs-discuss).

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Torrey McMahon

Richard Elling wrote:


Personally, I've never been in the situation where users ask for less 
storage,
but maybe I'm just the odd guy out? ;-) 


You just realized that JoeSysadmin allocated ten luns to the zpool when 
he realy only should have allocated one.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Al Hopper
On Fri, 26 Jan 2007, Torrey McMahon wrote:

> Al Hopper wrote:
> >
> > Now your accounting folks are going to be asking you to justify the
> > purchase of that hi-end SAN box and why you're not using ZFS
> > everywhere.  :)
> >
> > Oh - and the accounting folks love it when you tell them there's no
> > ongoing cost of ownership - because Joe Screwdriver can swap out a failed
> > Seagate 500Gb SATA drive after he picks up a replacement from Frys on his
> > lunch break!
>
> Because ZFS doesn't run everywhere.
> Because most low end JBODs are "low end" for a reason. They aren't as
> reliable, have crappy monitoring, etc.

Agreed.  There will never be one screwdriver that fits everything.  I was
simply trying to re-inforce my point.

> Fix those two things when you get a chance. ;)

Have a good weekend Torrey (and zfs-discuss).

Regards,

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
   Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
 OpenSolaris Governing Board (OGB) Member - Feb 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Al Hopper
On Fri, 26 Jan 2007, Toby Thain wrote:

> >
> > Oh - and the accounting folks love it when you tell them there's no
> > ongoing cost of ownership - because Joe Screwdriver can swap out a
> > failed
> > Seagate 500Gb SATA drive after he picks up a replacement from Frys
> > on his
> > lunch break!
>
> Why do people think this will work? I never could figure it out.
>
> There's many a slip 'twixt cup and lip. You need the spare already
> sitting there.

Agreed.  I remember years ago, when a Sun service tech came onsite at a
fortune 100 company I was working in at the time and we stopped him,
handed him a disk drive in an anti-static bag and said - "don't unpack
your tools - it was a bad disk, we replaced it from our spares, here's the
bad one - please replace it under the service agreement".  He thought
about this for about 5 Seconds and said; "I wish all my customers were
like you guys".  Then he was gone!  :)

Regards,

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
   Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
 OpenSolaris Governing Board (OGB) Member - Feb 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Torrey McMahon

Al Hopper wrote:


Now your accounting folks are going to be asking you to justify the
purchase of that hi-end SAN box and why you're not using ZFS
everywhere.  :)

Oh - and the accounting folks love it when you tell them there's no
ongoing cost of ownership - because Joe Screwdriver can swap out a failed
Seagate 500Gb SATA drive after he picks up a replacement from Frys on his
lunch break!


Because ZFS doesn't run everywhere.
Because most low end JBODs are "low end" for a reason. They aren't as 
reliable, have crappy monitoring, etc.


Fix those two things when you get a chance. ;)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Toby Thain


Oh - and the accounting folks love it when you tell them there's no
ongoing cost of ownership - because Joe Screwdriver can swap out a  
failed
Seagate 500Gb SATA drive after he picks up a replacement from Frys  
on his

lunch break!


Why do people think this will work? I never could figure it out.

There's many a slip 'twixt cup and lip. You need the spare already  
sitting there.


--T

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Jason J. W. Williams

To be fair, you can replace vdevs with same-sized or larger vdevs online.
The issue is that you cannot replace with smaller vdevs nor can you
eliminate vdevs.  In other words, I can migrate data around without
downtime, I just can't shrink or eliminate vdevs without send/recv.
This is where the philosophical disconnect lies.  Everytime we descend
into this rathole, we stir up more confusion :-(


We did just this to move off RAID-5 LUNs that were the vdevs for a
pool, to RAID-10 LUNs. Worked very well, and as Richard said was done
all on-line. Doesn't really address the shrinking issue though. :-)

Best Regards,
Jason
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Al Hopper
On Fri, 26 Jan 2007, Rainer Heilke wrote:

> > So, if I was an enterprise, I'd be willing to keep
> >  enough empty LUNs
> > available to facilitate at least the migration of
> >  one or more filesystems
> > if not complete pools.

 reformatted ...
> You might be, but don't be surprised when the Financials folks laugh you
> out of their office. Large corporations do not make money by leaving
> wads of cash lying around, and that's exactly what a few terabytes of
> unused storage in a high-end SAN is. This is in addition to the laughter

But this is exactly where ZFS distrupts "Large corporations" thinking.
You're talking about (for example) 2 terabytes on a high-end SAN which
costs (what ?) per GB (including the capital cost of the hi-end SAN)
versus a dual Opteron box with 12 * 500Gb SATA disk drives that gives you
5TB of storage for (in round numbers) a total of ~ $6k.  And how much are
your ongoing monthlies on that hi-end SAN box?  (Don't answer)  So - aside
from the occasional use of the box for data migration, this ZFS "storage
box" has 1,001 other uses.  Pick any two (uses), based on your knowledge
of big corporation thinking and its an easy sell to management.

Now your accounting folks are going to be asking you to justify the
purchase of that hi-end SAN box and why you're not using ZFS
everywhere.  :)

Oh - and the accounting folks love it when you tell them there's no
ongoing cost of ownership - because Joe Screwdriver can swap out a failed
Seagate 500Gb SATA drive after he picks up a replacement from Frys on his
lunch break!

> generated by the comment that, "not a big deal if the Financials and HR
> databases are offline for three days while we do the migration." Good

Again - sounds like more "legacy" thinking.  With multiple gigabit
ethernet connections you can move terrabytes of information in a hour,
instead of in 24-hours - using legacy tape systems etc.  This can be
easily handled during scheduled downtime.

> luck writing up a business case that justifies this sort of fiscal
> generosity.

> Sorry, this argument smacks a little too much of being out of touch with
> the fiscal (and time) restrictions of working in a typical corporation,
> as opposed to a well-funded research group.
>
> I hope I'm not sounding rude, but those of us working in medium to large
> corporations simply do not have the money for such luxuries. Period.

On the contrary - if you're not thinking ZFS, you're wasting a ton of IT
$s and hurting the competitiveness of your business.

Regards,

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
   Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
 OpenSolaris Governing Board (OGB) Member - Feb 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Richard Elling

Rainer Heilke wrote:

So, if I was an enterprise, I'd be willing to keep
 enough empty LUNs
available to facilitate at least the migration of
 one or more filesystems
if not complete pools.


You might be, but don't be surprised when the Financials folks laugh you out of their 
office. Large corporations do not make money by leaving wads of cash lying around, and 
that's exactly what a few terabytes of unused storage in a high-end SAN is. This is in 
addition to the laughter generated by the comment that, "not a big deal if the 
Financials and HR databases are offline for three days while we do the migration." 
Good luck writing up a business case that justifies this sort of fiscal generosity.


To be fair, you can replace vdevs with same-sized or larger vdevs online.
The issue is that you cannot replace with smaller vdevs nor can you
eliminate vdevs.  In other words, I can migrate data around without
downtime, I just can't shrink or eliminate vdevs without send/recv.
This is where the philosophical disconnect lies.  Everytime we descend
into this rathole, we stir up more confusion :-(

If you consider your pool of storage as a zpool, then the management of
subparts of the pool is done at the file system level.  This concept is
different than other combinations of devices and file systems such as
SVM+UFS.  When answering the ZFS shrink question, you need to make sure
you're not applying the old concepts to the new model.

Personally, I've never been in the situation where users ask for less storage,
but maybe I'm just the odd guy out? ;-)  Others have offered cases where
a shrink or vdev restructuring could be useful.  But I still see some
confusion with file system management (including zvols) and device management.
The shrink feature is primarily at the device management level.
 -- richard
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-26 Thread Rainer Heilke
> So, if I was an enterprise, I'd be willing to keep
>  enough empty LUNs
> available to facilitate at least the migration of
>  one or more filesystems
> if not complete pools.

You might be, but don't be surprised when the Financials folks laugh you out of 
their office. Large corporations do not make money by leaving wads of cash 
lying around, and that's exactly what a few terabytes of unused storage in a 
high-end SAN is. This is in addition to the laughter generated by the comment 
that, "not a big deal if the Financials and HR databases are offline for three 
days while we do the migration." Good luck writing up a business case that 
justifies this sort of fiscal generosity.

Sorry, this argument smacks a little too much of being out of touch with the 
fiscal (and time) restrictions of working in a typical corporation, as opposed 
to a well-funded research group.

I hope I'm not sounding rude, but those of us working in medium to large 
corporations simply do not have the money for such luxuries. Period.

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-25 Thread SteveW
> > The other point is, how many other volume management systems allow you
> > to remove disks? I bet if the answer is not zero, it's not large. ;)
> 
> As far as Solaris is concerned, I'm only aware of two significant such
> systems. SVM and VxVM.

Correct, in our case we've been using VxVM for years.  
We're comfortable with it, it does what we want, vxfs performance is 
outstanding, and we move storage around [u]A LOT[/u] as our SAN storage 
changes.  So for the time being, the benefits won't allow us to give up the 
ability to shrink/remove on the fly.
I should add that I am very excited about ZFS, and hope to be able to use it 
[i]everywhere[/i] as soon as possible.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-24 Thread Darren J Moffat

Rainer Heilke wrote:

For the "clone another system" zfs send/recv might be
useful


Keeping in mind that you only want to send/recv one half of the ZFS mirror...


Huh ?

That doesn't make any sense.  You can't send half a mirror.  When you 
are running zfs send it is a "read" and ZFS will read the data from all 
available mirrors to help performance.  When it is zfs recv it will 
write to all sides of the mirror on the destination.


What are you actually trying to say here ?

--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: How much do we really want zpool remove?

2007-01-23 Thread Rainer Heilke
> For the "clone another system" zfs send/recv might be
> useful

Keeping in mind that you only want to send/recv one half of the ZFS mirror...

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss