Chris Worley wrote:
On Dec 6, 2007 10:26 AM, Aaron Kulkis <[EMAIL PROTECTED]> wrote:
Chris Worley wrote:
On Dec 5, 2007 5:12 PM, Aaron Kulkis <[EMAIL PROTECTED]> wrote:
Chris Worley wrote:
On Dec 4, 2007 11:49 AM, Aaron Kulkis <[EMAIL PROTECTED]> wrote:
Chris Worley wrote:
On Dec 4, 2007 10:22 AM, Jc Polanycia <[EMAIL PROTECTED]> wrote:
Off topic, as I seldom partition anything (unpartitioned drives
perform best), but, you're setting yourself up for disaster using LVM
(any corruption to the LVM layer is not recoverable... you'll loose
everything... been there done that), and the performance is poor, and
MD RAID5/6 devices can be grown (add more disks).
Chris
Fair enough. I appreciate the input because I haven't run across any
real-world stories about LVM corruption. I have personally encountered
corruption problems with RAID5/6 as well as problems with decreased
performance as a RAID5 structure gets more members added to it.
I saw some RAID6 issues last year, so I use RAID5... but recent tests
have shown MD RAID6 as solid.
"Decreased performance as more members get added to it"? Bull!!! I'm
guessing you have another bottleneck that has led you to this
conclusion.
While the performance increase doesn't scale linearly as disks are
added (some CPU verhead is added with each additional drive), the more
disks, the better the performance. I'm sure there is some Amdahl's
law limit to the increased performance scalability, but I run RAIDS up
to 12 drives, and see performance added w/ each new member.
You're hallucinating. That defies basic information theory.
Your assertion is akin to suggesting that you power your
computers with a perpetual motion machine (despite the
fact that such would violate the 1st, 2nd, and 3rd laws
of thermodynamics).
Amdahl's law defies "Information theory"? How so?
If you've got one disk that can perform at 70MB/s on a 320MB/s bus,
then on that bus you should be able to stripe at least four drives
with less-than-linear scalability... add more busses w/ more dirves...
more scalability... of course, not linear. Add caching effects, and
get superlinear scalabiltiy (but that doesn't count).
Your analysis is flawed because it assumes zero time
for disk-head seeks.
How does it assume zero time?
If you've got multiple disks all seeking simultaneously, it is a
parallel, not serial, operation.
That doesn't matter. Your assumption that the data blocks
from several disks can be XOR'ed together, and written
Do you mean "to" where you say "from"? It's like you're mixing reads
and writes in the same sentence.
No, I'm not.
To do a write to block #X on Drive A, Block #X on the parity
drive must be updated, which means that Block #X on drives
B, C, ... must also be read, and all of them XOR'ed together
before you can write the parity block on the parity device.
Now, since you seem to be COMPLETELY unaware of the methods
by which RAID 5 works, and how that method effects system
performance, please STOP POSTING YOUR BULLSHIT CLAIMS
that RAID 5 is a high-performance configuration.
It's a low-performance alternative to mirroring, because
it is substantially cheaper.
to one of those disks, and the parity partition on yet
Note that RAID5 is rotating parity... there is no one "parity partition".
For every block, one of the devices is acting as the parity
device block for the other disks in a RAID5 configuration.
another disk is FASTER than not doing so is just patently
ridiculous.
I'm guessing you've fallen victim to some RAID card with a chicklet
for a processor.
Or alternatively, you're just making shit up, and posting
it as fact.
An MD device in Linux is much faster (given modern processors) at
calculating the checksum, and the calculation is insignificant
compared to the time it takes to write the data. It still represents
a serial portion of the operation, but my "Amdahl's Law" disclaimer
was clearly posted in your first objection.
But you STILL have to write MORE data (the parity blocks)
than without RAID 5, which adds to I/O overhead.
Any argument to the contrary is PURE FANTASY.
Have you come to terms with your thermodynamic issue?
I'm still not the one claiming that writing to N+1
disks has less overhead than writing to 1 disk.
That doesn't even count the matter of increasing the
bandwidth usage by a factor of N for N disks in the
RAID 5 configuration.
You must have some very slow busses too. If you're using PCI-X or
PCI-E busses, with multiple (or even single) SCSI U320 or SCA
busses... it takes a lot of N to saturate the bandwidth.
Disk-head motion is not free. In fact, it is the most
expensive part (delay-wise) of using a disk drive.
You must be assuming that striping across a RAID is somehow a serial operation.
No, I'm asuming that you're using RAID 5, which
is what you said.
Which is striped, in parallel.
Not according to the definition if RAID 5.
RAID 5 is rotating Distributed Parity.
--
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]