Dave Watkins wrote:
> Excuse the top posting, I'm stuck on OWA :)
>  
> There are still some additional benefits if you really need speed, and
> they are the onboard memory hardware raid cards have, and the ability
> to add a BBU so you can turn on write-back caching and geta nice speed
> boost for doing it.

It'd be interesting to see some numbers of SAS RAID with cache and
write-back vs softraid....


R.
>  
> Dave
>
> ------------------------------------------------------------------------
> *From:* [EMAIL PROTECTED] on behalf of Rafiu
> Fakunle
> *Sent:* Sat 7/21/2007 3:13 p.m.
> *To:* Dave Johnson
> *Cc:* [email protected]
> *Subject:* Re: [OF-users] Getting I2O RAID to work
>
> Dave Johnson wrote:
> > i for one would like to start the trend of ceasing to call these
> > "RAID" cards, since they aren't,
>
> Hear hear.
>
> > and instead calling them exactly what they are,
>
> Lemons! ;)
>
> > storage device controller cards with an integrated -but for all
> > intents and purposes- outboard XOR Offload Engine (XOE).
> >
> > XOE (zo-wee) cards are notorious for failing on you at the most
> > inopportune time.  I 2nd Rafiu's recomendation to go with a complete
> > RAID subsystem which includes XOE, storage device controller, and data
> > transport IO processor all in one complete card.  Or simply not use
> > the XOE of the card and use only the storage controller portion,
> > relying instead on the RAID support within the LVM2 component, which
> > has been designed and vetted for safety by considerably more testing
> > than Promise' 5 beta testers in Taiwan.
> >
> > That XOE can be a biotch ! =P
> >
>
> You know, with the average CPU having at minimum 2 cores, the only
> advantage to using a RAID controller these days - when you weigh
> softraid against the management overhead of hardraid - is for the
> hot-swap capability. Otherwise a decent SAS controller + MD RAID and
> you're good to go. External RAID is a different matter of course,
> especially when you get  into the realm of shared storage.
>
>
> R.
>
> > -=dave
> >
> > ----- Original Message ----- From: "Rafiu Fakunle" <[EMAIL PROTECTED]>
> > To: "Jim Kusznir" <[EMAIL PROTECTED]>
> > Cc: <[email protected]>
> > Sent: Friday, July 20, 2007 3:48 PM
> > Subject: Re: [OF-users] Getting I2O RAID to work
> >
> >
> >> Jim Kusznir wrote:
> >>> Hi all:
> >>>
> >>> After over a week of messing with OpenFiler, I think I'm finally close
> >>> to getting my hardware I2O RAID card working (Promise SX6000).  I had
> >>> to upgrade the kernel, as the version shipping with OF has a bug in it
> >>> that breaks all I2O RAID cards.  I don't need iSCSI target for now, so
> >>> I though tthis was acceptable.
> >>
> >> Don't use i2o cards ;)
> >>
> >> Stay away from anything named "Promise" or "Highpoint" if you want
> OF to
> >> play nice with the RAID Controller.
> >> Cards we tend to play better with are:
> >>
> >> 1) 3Ware / AMCC
> >> 2) Areca
> >> 3) LSI Logic / Intel
> >> 4) Adaptec
> >> 5) ICP Vortex
> >>
> >>>
> >>> Now that I have it showing up in my system, things are close.  The
> >>> problem is I2O raid devices are created in /dev/i2o/hd* (my case:
> >>> /dev/i2o/hda -- this is NOT the same disk as my system disk:
> >>> /dev/hda).  So, under "Physical Volues", it does not see it, and thus
> >>> I can't partition or use it.  I have verified that the volume itself
> >>> works by partitioning and formatting it directly.  Both operations
> >>> completed sucessfully, verifying the raid drivers funcitionality.
> >>>
> >>> So my question at this point is: how do I get OF to see the disk and
> >>> be able to create volumes and such.  Or, if I partition and set up LVM
> >>> by hand, will it pick it up at that point?  If so, what are its
> >>> requirements to make it seen?
> >> Here's a patch to /opt/openfiler/sbin/list-disks.pl
> >>
> >> --- list-disks.pl.orig  2007-07-20 15:18:22.000000000 -0700
> >> +++ list-disks.pl       2007-07-20 15:26:24.000000000 -0700
> >> @@ -123,6 +123,7 @@
> >>                                close(MEDIA);
> >>                                if ($media =~ /^disk/ && !$_[0]) {
> >>                                        push(@devs, "/dev/hd$n");
> >> +                                       push(@devs, "/dev/i2o/hd$n");
> >>                                        }
> >>                                }
> >>                        }
> >>
> >>
> >>
> >> R.
> >> _______________________________________________
> >> Openfiler-users mailing list
> >> [email protected]
> >> https://lists.openfiler.com/mailman/listinfo/openfiler-users
> >>
> >
> > _______________________________________________
> > Openfiler-users mailing list
> > [email protected]
> > https://lists.openfiler.com/mailman/listinfo/openfiler-users
>
> _______________________________________________
> Openfiler-users mailing list
> [email protected]
> https://lists.openfiler.com/mailman/listinfo/openfiler-users
>
> ------------------------------------------------------------------------
>
> _______________________________________________
> Openfiler-users mailing list
> [email protected]
> https://lists.openfiler.com/mailman/listinfo/openfiler-users
>   

_______________________________________________
Openfiler-users mailing list
[email protected]
https://lists.openfiler.com/mailman/listinfo/openfiler-users

Reply via email to