> [snip]
> Reliability is only one part of the equation.
> [snip]
> Good points, however I was responding to the idea that 
> running a simple app on a single box somehow proves that
> the server boxes are now on par with mainframe reliablility,
> and that's not a logical comparison.

Sure its a valid comparison. The only thing a user cares about
is the result. They might have a bit of tourist curiosity, but 
as long as they get their job done on time, they generally don't
care HOW it gets done. So if it takes two or more redundant (but
very cheap) boxes to do the job of a single (much more expensive)
mainframe the user probably doesn't care. Moreover, the business
may actually prefer the lower price, even if it comes at the cost
of a bit more operational complexity. 
 
> [snip]
> > I've yet to see unix or windows accomplish the same thing.
> 
> See Ron Hawkins' last post on this. If you're running modern
> disk hardware you're running a fault-tolerant unix application
> [snip]
> Again, you can't compare a single purpose application (such as
> a disk array), with a mainframe. Sure, some subsystems are supported 
> by cheap chips or os/s, but they are doing a very simple job:
> exactly what these things should be used for.

Actually, I -can- compare these things quite usefully. You appear 
to have an inherent preference for running all of these things on
the same box at the same time. That's ok, but you should recognize
it as a preference only. If the same thing can be accomplished more
cost effectively, or more reliably, by some other approach, then
who's to say the other approach is wrong?

> [snip]
> Rubbish. Reliability is a "nice to have" feature that costs a ton of 
> engineering dollars.
> [snip]
> I'm not sure where you've been working, but if I told the company 
> president here that reliabliltiy was "nice to have" I'd be out on the 
> street before I could blink. IBM didn't invent RAS systems to keep 
> themselves amused, they did so because their customers demanded it,
> and were willing to pay for it. Both IBM and their customers had 
> to invest heavily in this or they'd be out of business.

You're missing the point entirely. High AVAILABILITY is a "gotta have
it" business requirement (or not - see below) but there are many ways
of accomplishing that. Availability is the probability that a system
can do work when you want it to and that's MTBF/(MTBF+MTTR) It costs
a lot to make MTBF LARGE for every last component. Depending on design,
you might be more successful by simply making MTTR small. Then if you
factor in the availability of redundant components, you find that you
can get very high overall AVAILABILITY from very low RELIABILITY 
components. I'm not making this up, its just the way it is.

You also should consider the origins of the drive for reliability.
Back in the day, customers typically had just one system. And if 
it was down, so were they. Nobody in their right mind would pay 
the cost of duplicating the resources of that single system even
if the software supported it, which it generally did not. 

So... the only choice was to make every element of the system as
reliable as could be - and that's one of the reasons the platform
has stayed as expensive (raw $$$, not cost/function) as it has.

The engineering that goes into these systems is staggering. But
while the engineering drive for ultimate reliability continues, 
the (mainframe) systems today are increasingly made up of cheap
commodity parts. The memory subsystem has pipelined access and
chip kill and multibit ECC, but its made of cheap and cheerful 
memory chips not unlike those in your own PC.

Now in the brave new world where chips cost pennies, the most
cost effective solution is usually to provide multiple levels of
redundancy. That's entirely different in concept. Instead of having
one big thumper sitting in the corner being a single point of 
failure, you have many equivalent servers and no single point of
failure. That's the mantra for sysplex and for the distributed 
systems too. 

So in fact, as I said earlier. Reliability -is- nice to have. But
you can have high business availability without it. Now to return
to that issue, take a long hard look at the real numbers and the
results are not pretty. z/OS availability, as experienced by most
of our brethren, is nothing to write home about. See all of the 
past threads about IPL frequency for verification. 

Its not that the system isn't capable of continuous operation (IT 
IS!!!) but evidently a wide majority of customers don't actually
trust it to be as reliable as it is. Getting the benefits of the
latest technology takes a fundamental shift in mindset as well as
a shift in configuration. If you keep running your mainframe 
systems and applications in towers the same as they were "back in
the day" then frankly, they are considerably less reliable/available
as off the shelf solutions with other platforms. Sorry if that is
bursting some bubbles, but that's just the way it is.

> The day that it becomes commonplace for, say, three unix servers 
> to be able to support the entire IT production environment 
> for a large company, and do so reliably, I'll concede the point.

Already true today. If you are arguing capacity, then even today
the P series stomps the Z series and it supports many of the same
kind of virtualization and partitioning. It may not do it with
the same raw efficiency, but if the price is right, who cares?

Note also: I am not specifically advocating that people go out 
and replace mainframes with those other systems. I am merely 
pointing out that the other systems are entirely competitive
and if we want to keep up, we have to embrace change. Widespread
perceptions of mainframe superiority are long out of date.

CC

----------------------------------------------------------------------
For IBM-MAIN subscribe / signoff / archive access instructions,
send email to [EMAIL PROTECTED] with the message: GET IBM-MAIN INFO
Search the archives at http://bama.ua.edu/archives/ibm-main.html

Reply via email to