Re: Looking for a Text on ZFS

2008-02-05 Thread Drew Tomlinson

Wojciech Puchar wrote:


[snip]


mirror. that's all.


Which doesn't really address the issue of what happens if a drive 
that is

part of a big ZFS is removed (because it's broken).


it will say "read error" on all files and directories  that happened 
to be placed on that disk!


Just to be clear, ZFS pools contain one or more VDEVs (Virtual 
devices).  Each VDEV consists of one or more physical partitions and can 
be configured as a mirror, stripe, stripe with one parity, or stripe 
with 2 parity.  ZFS does striping with no parity across all VDEVs in the 
pool.  Thus with ZFS, if you lose a VDEV, you lose the whole pool.


Hopefully that clears things up a bit.

Cheers,

Drew

[snip]
___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-04 Thread Wojciech Puchar

did you ever got your UFS filesystem broken not because your drive failed?


That is not the point here. I have been using FreeBSD sind version 3.3,
which was released in 1999. Before that I used Linux. So I can't even look


while i was using linux - crashed filesystem was quite common without any 
disk failure. that's one of the reason i started using *BSD, another was 
performance.



To answer the question: Yes, it did happen and not only once. This was in
the time when I was setting up a new computer with 6.0-RELEASE and a new
S-ATA controller. There was a bug in the driver which the developer
managed to fix after we exhanged a few eMails. Before the error was fixed,
my machine crashed several times with a kernel panic. There were something


of course - with buggy driver it may be big problems.


connected to the new controller and I was using this machine as a
plattform to debug the driver, no real damage was caused.


you of course will not use new/untested hardware/software combination for 
production machine? me too, so what a problem?



That's ok to believe if you want to. UFS is designed to minimize errors.
There is no guarantee that there will be none.


there is never guarantee. if ZFS will calculate block checksum from memory 
and miscalculate it because of hardware/software problem, it will write 
it, and then write good data to say 2 disks.


then it will take data from both disk and report uncorrectable error - 
without any error at all.


just an example. there is no "ultimate" solution.


systems still work. :-)


my works too with one partition. so what a problem - except i had less 
work?





still - making all in / is much easier and works fine.


Maybe I'm just too conservative for that.


in EVERY unix book it's repeated countless times that partitioning is 
good, and make it more secure, more prone to errors etc.


it was - on original unix FS, but not on UFS, which automatically 
"partitions" your drive to cylinder groups.



making all in / and /lessused, where / is at first part on disk, and
/lessused on second - make big performance improvements (shorter seeks!).


There are about 10 things I can think of that I'd do before I tried
something like that. I'm a little surprised about a suggestion like this
coming from you because you seem to be a great advocacy of dynamic


what you mean "dynamic system"?


systems. And here you have to decides what is used often and what not.
This is an estimate that you could also mess up - I'm sure I probably
would. :-) And chaninge a file from the seldom to the often area isn't
that trivial either.


i prefer / for everything. but sometimes i need this to speed things up, 
and definitely need it when using gmirror - to not waste lot of space by 
mirroring everything. i just mirror what's have to be mirrored.



But ok, noone will judge either of us for working with our systems the way
we please. :-) Anyone with Unix knowledge will find his way around my
boxes and the same should be true for you. The rest are just details. :-)


yes - but it's bad repeating "truth"s because it's said.

ZFS advocacy first creates problems then solving it.
i STRONGLY state most of these problems are artifical for most 
administrators and users.


so ZFS may be useful for someone, but for most of as it's just waste of 
CPU and memory and... ours time (contrary to what ZFS stated - saving ours 
time).



mirror. that's all.


Which doesn't really address the issue of what happens if a drive that is
part of a big ZFS is removed (because it's broken).


it will say "read error" on all files and directories  that happened to be 
placed on that disk!



you understand quota as workaround of problems creating 1000 partitions.
or simply - looks like you don't understand it at all, because it is not
workaround. it's excellent tool.


Maybe you just don't understand my English? :-)


maybe..but you stated that quota is needed because partitions can't me 
easily created by mass.


and that's exactly wrong.


Quota does not address the different needs of certain applications. With
quota you can limit the amount of inodes a user may grab but you cannot
create areas with more inodes and others with less. Quota solves many


in most cases average inode count is important with a bit excess.

with rare cases really lots of files planned in some directory (like my 
squid spools) i make separate partitions.


having this spools on separate partitions i can greatly reduce seeking as 
it's all in narrow part of disk plate.


with ZFS it's no problem to have lots of small files, but they will be 
mixed up with other - without any control of placement.


actually - i think it could be done automatically quite good, like few 
other things.


maybe (i'm too lazy) i will write UFS3 ;) - but for sure it will be 
UFS-style filesystem, with some improvements. there is no need to 
revolution



problems and is a great tool, no doubt in that, but it doesn't make your
computer fast, you les

Re: Looking for a Text on ZFS

2008-02-04 Thread Christian Baer
On Mon, 4 Feb 2008 13:39:52 +0100 (CET) Wojciech Puchar wrote:

> did you ever got your UFS filesystem broken not because your drive failed?

That is not the point here. I have been using FreeBSD sind version 3.3,
which was released in 1999. Before that I used Linux. So I can't even look
back on 10 years of FreeBSD yet and I don't have that many drives I have
to worry about. So the fact if one of my files systems ever broke isn't
really representative.

To answer the question: Yes, it did happen and not only once. This was in
the time when I was setting up a new computer with 6.0-RELEASE and a new
S-ATA controller. There was a bug in the driver which the developer
managed to fix after we exhanged a few eMails. Before the error was fixed,
my machine crashed several times with a kernel panic. There were something
like two dozen crashes in that time. Twice the filesystem could be
salvaged by fsck, but the data on it was pretty messed up. I don't know
how that happened and frankly, I don't care either. The rest of the
times, fsck did get the fs into normal working order again whith just the
file broken that was last being written. Since the boot drive wasn't
connected to the new controller and I was using this machine as a
plattform to debug the driver, no real damage was caused.

> i don't. UFS it's not FAT, and doesn't break up.

That's ok to believe if you want to. UFS is designed to minimize errors.
There is no guarantee that there will be none.

> you CAN't estimate well how much space you need in longer term.
> in practice partitioning like yours means at least 100% more disk space 
> requirements.

I wouldn't be that pessemistic. True, you can't be sure you allocated
enough space to X, so you leave a safety margine. But the fact that the
HDD doesn't grow limits your space anyway. I am not denying that you might
waste space this way but it's still nothing I'd lose any sleep over.

> of course - there are often cases today that whole system needs few gigs, 
> but smallest new drive is 80GB - it will work..

I work with lots of drives that are a lot smaller than that. And the
systems still work. :-)

> still - making all in / is much easier and works fine.

Maybe I'm just too conservative for that.
Mind you, I don't break up all drives by default. I have some 500GB drives
that have only one large partition. This partition is for data (which
means everything but system stuff). All I break up into pieces are the
default system areas.

> making all in / and /lessused, where / is at first part on disk, and 
> /lessused on second - make big performance improvements (shorter seeks!).

There are about 10 things I can think of that I'd do before I tried
something like that. I'm a little surprised about a suggestion like this
coming from you because you seem to be a great advocacy of dynamic
systems. And here you have to decides what is used often and what not.
This is an estimate that you could also mess up - I'm sure I probably
would. :-) And chaninge a file from the seldom to the often area isn't
that trivial either.

I increase performance by mounting /tmp and /usr/obj async and I mount
systems I want to work fast with noatime.

But ok, noone will judge either of us for working with our systems the way
we please. :-) Anyone with Unix knowledge will find his way around my
boxes and the same should be true for you. The rest are just details. :-)

>> I read about this. However, I didn't find anything conclusive as to how
>> well the drives can still live on their own if they are ever seperated.
>> Now I don't think they will be addressed as a RAID0 with all the risks of
>> that. But what happens if one of four drives breaks down? Does it make a
>> difference, if the broken drive is the first one, the last one or a middle
>> one?
>
> if it's just concat, you will loose lots of data, just like any other 
> filesystem.
> with concat+mirror - you replace single drive that failed and rebuild 
> mirror. that's all.

Which doesn't really address the issue of what happens if a drive that is
part of a big ZFS is removed (because it's broken).

> after reading your answer on 3-rd question i will end the topic, because 
> you understand quota as workaround of problems creating 1000 partitions.
> or simply - looks like you don't understand it at all, because it is not 
> workaround. it's excellent tool.

Maybe you just don't understand my English? :-)

I understand quota very well and also what it can do. It is a very useful
tool but it is not the holy grail. I actually use both block and file
quote on some of the systems I have to watch. And I use both hard and soft
at that. Quota does eliminate the need to create one partition for each
home directory, even if you think it is not meant for that. And actually,
it is used a lot for just that purpose. ISPs with shared hosting products
usually don't allow direct write access outside the users ~ anyway. So the
quota just stops him from uploading to much. But I know quota is also 

Re: Looking for a Text on ZFS

2008-02-04 Thread Wojciech Puchar

/usr to spread the load while making worlds and I mount /usr/obj
asynchronously to increase write speed. With several filesystems I can
spread to load the way I want it and decide where the data goes. And one
broken fs doesn't screw up the others in the process.


did you ever got your UFS filesystem broken not because your drive failed?

i don't. UFS it's not FAT, and doesn't break up.



I do know the drawbacks of this: Storage is pretty static. Correcting
wrong estimates about the needed fs-sizes is a big problem. That is why I


you CAN't estimate well how much space you need in longer term.
in practice partitioning like yours means at least 100% more disk space 
requirements.


of course - there are often cases today that whole system needs few gigs, 
but smallest new drive is 80GB - it will work..


still - making all in / is much easier and works fine.

making all in / and /lessused, where / is at first part on disk, and 
/lessused on second - make big performance improvements (shorter seeks!).



2) it takes many drives to the pool and you may add then new drives.
same as gconcat+growfs.


I read about this. However, I didn't find anything conclusive as to how
well the drives can still live on their own if they are ever seperated.
Now I don't think they will be addressed as a RAID0 with all the risks of
that. But what happens if one of four drives breaks down? Does it make a
difference, if the broken drive is the first one, the last one or a middle
one?



if it's just concat, you will loose lots of data, just like any other 
filesystem.


with concat+mirror - you replace single drive that failed and rebuild 
mirror. that's all.




after reading your answer on 3-rd question i will end the topic, because 
you understand quota as workaround of problems creating 1000 partitions.
or simply - looks like you don't understand it at all, because it is not 
workaround. it's excellent tool.

___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-04 Thread Christian Baer
On Sun, 3 Feb 2008 17:55:12 +0100 (CET) Wojciech Puchar wrote:

> that's like 64-bit soundcards that have to be "better" than 32-bit, while 
> most of them was unable to actually get past 13-14 bit (most past 12) with 
> it's signal to noise ratio.

Maybe that's not quite the same thing. :-)

However. Even a 64bit filesystem still has gigantic reserves of space and
although filling that may not cause the oceans to boil, any storage device
that can actually sore all the date that a 64bit fs can allocate will be
"pretty big" in terms of volume and mass and will also use a good deal of
energy - even if this is calculated in the same minimalistic way as for
ZFS: http://en.wikipedia.org/wiki/Zfs#Capacity

Now I am not sure if the world actually needs a file system that can never
be filled - with the limits not made by any feable estimates like we
thought we'd never get a 1GB drive full, but by quantum physics. But the
fact that it's maximum theoretical size has "a few reserves" isn't a
problem in itself. I have serious doubts that the computers of today are
ready for the overhead at all and the overhead is worth the bother.

> 1) you make create 1000 of "filesystems" without partitioning. so lots of 
> "admins" that think more partitions=better are happy. you may set quota 
> for each "filesystem"

Well, actually I am an admin who believes this within limits. I have
seperate file systems for /, /usr, /var, /tmp, /home and /usr/obj. The
reasons for this are numerous. I have /usr/obj on a different drive than
/usr to spread the load while making worlds and I mount /usr/obj
asynchronously to increase write speed. With several filesystems I can
spread to load the way I want it and decide where the data goes. And one
broken fs doesn't screw up the others in the process.

I do know the drawbacks of this: Storage is pretty static. Correcting
wrong estimates about the needed fs-sizes is a big problem. That is why I
keep /usr/home on one big fs. If the users require (for example) 20MB
each, then it doesn't matter if one user needs 25MB, as long as 5 others
only use 24.

If ZFS gives us the best of both worlds, that would actually be an
improvement.

> 2) it takes many drives to the pool and you may add then new drives.
> same as gconcat+growfs.

I read about this. However, I didn't find anything conclusive as to how
well the drives can still live on their own if they are ever seperated.
Now I don't think they will be addressed as a RAID0 with all the risks of
that. But what happens if one of four drives breaks down? Does it make a
difference, if the broken drive is the first one, the last one or a middle
one?

> 3) it doesn't have per user quota, which creates a problem that is 
> "solved" by 1), and you have to create at least one filesystem/user, which 
> then is said to relieve admininstrator from work ;)

This doesn't have to be a problem either. Quota are used instead of
partitions to tackle two problems: The number of partitions is very
limited and resizing a partition is a major issue. By changing the quota
you can give one user (or one service) more room and take it away from
some of the others that seem to need less than was anticipated.

If each user or service can be confined to it's own fs, that would also be
good. A newsserver runnung with tradspool needs lots of inodes, most other
applications far less. I do see a drawback though: If you change the size
of the filesystems a few times, you could wind up with a new sort of
fragmentation. New because this sort (a filesystem that is patched
together over a drive) hasn't really been encountered yet and it will be
very interesting to see what effects this may have.

> 4) ZFS says that hardware checksums are not enough and disk hardware may 
> be buggy so then "solve" this problem checking everything with CPU.

This also creates a lot of overhead and CPU load. I tried this with GELI
on a fs that needed to be intact in a paranoid sense - I get like that
sometimes. :-) I did it once and once only. The performance was just not
good enough. Granted, I didn't do this on a really new computer but I'm
not likely to through away all my old ones either, just so my paranoia can
be met with a good speed. :-)

> while i've had failing drives many times i never seen it reading bad data 
> and not reporting error.

Same here. Since I use HDs on my computers, I have had about 20 to 25
drives break down over the years. Ok, I used many of the drives long after
other people took similar drives out of their machines and used them as
door stops. Basicly, I made these drives work until they dropped dead. :-)
None of these drives *ever* gave strange data back. The only time I had
that was when the driver for a controller was broken and that issue was
there right from the beginning.

> 5) you don't have to wait for fsck. one of the few real adventages.
> Anyway - FreeBSD doesn't crash like windoze, so it's not that big thing.

Wrong! Crashed accure and they do that quite frequently. Ev

Re: Looking for a Text on ZFS

2008-02-03 Thread [EMAIL PROTECTED]
On 03/02/2008, Christian Baer <[EMAIL PROTECTED]> wrote:
> On Sat, 2 Feb 2008 21:38:49 -0600 [EMAIL PROTECTED] wrote:

> > Well, the best, I think.
>
> I take ist, you don't approve of ZFS? :-)
>

It is not a panacaea.
The optimisation and sharing of r/w,
Load balancing,
And redundant data verification
(to say nothing of the supposed ability
to blindly stab disks into a nearly infinite
array) are all features that I can see
being appreciated.
In a couple of processor generations.
(I am talking desktop junk here)
The kids running big iron will have already
decided and my feeble arguments are like
sand in a badger's vagina.

-- 
--
___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-03 Thread Wojciech Puchar

I already read that before I posted my question. Neither by this text,
nor by the one in the Wikipedia could I participate in the exitement
around ZFS. Ok, so it's a 128Bit FS. Big fat, hairy deal! I couldn't see


that's like 64-bit soundcards that have to be "better" than 32-bit, while 
most of them was unable to actually get past 13-14 bit (most past 12) with 
it's signal to noise ratio.



any advantages in using it instead of FFS (UFS), but I thought I was


ZFS is "better" because:

1) you make create 1000 of "filesystems" without partitioning. so lots of 
"admins" that think more partitions=better are happy. you may set quota 
for each "filesystem"


2) it takes many drives to the pool and you may add then new drives.
same as gconcat+growfs.

3) it doesn't have per user quota, which creates a problem that is 
"solved" by 1), and you have to create at least one filesystem/user, which 
then is said to relieve admininstrator from work ;)


4) ZFS says that hardware checksums are not enough and disk hardware may 
be buggy so then "solve" this problem checking everything with CPU.


while i've had failing drives many times i never seen it reading bad data 
and not reporting error.


5) you don't have to wait for fsck. one of the few real adventages.
Anyway - FreeBSD doesn't crash like windoze, so it's not that big thing.

6) zfs set copies= works only on writes, but scrub doesn't make a 
missing copy when one is failed.


so the best possible adventage (setting what file to mirror, what not) is 
lost.


7) there is no per file encryption, while it's said it will be SOON ready.

8) ZFS is clear winner on artifical tests like creating miliion of small 
files and then deleting them etc..


9) ZFS is very fast, just add more RAM and faster CPU.

i would - to make more RAM and CPU power available for programs i run, not 
to be wasted.



there was a lot of excitement here after ZFS was ported, but i think it's 
time too see that 20 (or more?) year old UFS is still a winner.


i think some changes in UFS, like larger cylinder groups (so there won't 
be 1 of then on big filesystem), possibly dynamic allocation of 
inodes, would be good. but not critical :)

___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-03 Thread Christian Baer
On Sat, 2 Feb 2008 21:38:49 -0600 [EMAIL PROTECTED] wrote:

> ZFS ends the microsotf monopoly over our disks.

And this monopoly is founded on ... what?

> ZFS begins the world as a 128bit dadaspace.
> Using ZFS fixes allocations and massaging your NAS.
> The inode is now the wenode.
> Usaging ZFS will make everything sunnier.
> Brighter too.
> Making ZFS the default FS in an FScentric world ends
> the pesky problems associated with legacy hardware.
> Building a ZFS nonuplyindirectwenode multiply redundant
> redundant filesystem makes Kate Miller-Heidke the
> Well, the best, I think.

I take ist, you don't approve of ZFS? :-)

Regards,
Chris
___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-03 Thread Christian Baer
On Sat, 2 Feb 2008 21:11:21 +0100 Mel wrote:

> If you review the "Not done" items @ http://wiki.freebsd.org/ZFS and still 
> are 
> doubting, then http://www.opensolaris.org/os/community/zfs/whatis/ describes 
> what the features *can* be. I got a good impression from that text what the 
> advantages are, but I'm too conservative to migrate myself. YMMV.

I already read that before I posted my question. Neither by this text,
nor by the one in the Wikipedia could I participate in the exitement
around ZFS. Ok, so it's a 128Bit FS. Big fat, hairy deal! I couldn't see
any advantages in using it instead of FFS (UFS), but I thought I was
missing something because porting it would have been somewhat of a hassle
and noone would go to all that trouble if it wasn't worth the effort.

Regards,
Chris
___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-02 Thread [EMAIL PROTECTED]
On 02/02/2008, Christian Baer <[EMAIL PROTECTED]> wrote:
> Hello people!
>
> Can anyone give me a link to a text on ZFS that tells me why I might want
> to use that instead of FFS? I don't want to start a discussion which is
> better, just a comparison, as I assume that the two are not designed to do
> the same things. And if possible one that is understandable to people who
> don't hack FS-code. :-)

http://blogs.sun.com/jonathan/entry/harvesting_from_a_troll
http://zamwi.com/2007/01/16/why-do-geeks-have-lust-for-zfs/
ZFS ends the microsotf monopoly over our disks.
ZFS begins the world as a 128bit dadaspace.
Using ZFS fixes allocations and massaging your NAS.
The inode is now the wenode.
Usaging ZFS will make everything sunnier.
Brighter too.
Making ZFS the default FS in an FScentric world ends
the pesky problems associated with legacy hardware.
Building a ZFS nonuplyindirectwenode multiply redundant
redundant filesystem makes Kate Miller-Heidke the
Well, the best, I think.

-- 
--
___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-02 Thread Wojciech Puchar

If you review the "Not done" items @ http://wiki.freebsd.org/ZFS and still are
doubting, then http://www.opensolaris.org/os/community/zfs/whatis/ describes
what the features *can* be. I got a good impression from that text what the
advantages are, but I'm too conservative to migrate myself. YMMV.

very good.

"CAN be" isn't very useful. while ZFS provide "virtual partitions" (which 
may LOOK good), it doesn't work well with 2 or more pools created out of 
partitions not full drives.


my common config on machines with >1 drive is to make gmirror (or 
gmirror+gstripe) from first partitions of each drive, to store most common 
data, usually EXCEPT huge files, and gconcat from other partitions to 
store mostly big files, other rarely used things, copies of other things 
etc.


then drives seeks mostly within first partition so it's much faster, and i 
have unmirrored larger space on second.


with ZFS it is not possible, while it CLAIMED to completely and definitely 
remove all these burder about planning disk layouts.



___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"


Re: Looking for a Text on ZFS

2008-02-02 Thread Mel
On Saturday 02 February 2008 20:07:50 Christian Baer wrote:

> Can anyone give me a link to a text on ZFS that tells me why I might want
> to use that instead of FFS? I don't want to start a discussion which is
> better, just a comparison, as I assume that the two are not designed to do
> the same things. And if possible one that is understandable to people who
> don't hack FS-code. :-)

If you review the "Not done" items @ http://wiki.freebsd.org/ZFS and still are 
doubting, then http://www.opensolaris.org/os/community/zfs/whatis/ describes 
what the features *can* be. I got a good impression from that text what the 
advantages are, but I'm too conservative to migrate myself. YMMV.

-- 
Mel
___
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "[EMAIL PROTECTED]"