Re: Mersenne: mprime for QA or performance?

1999-01-16 Thread George Woltman

Hi,

At 12:24 AM 1/13/99 -0800, Leu Enterprises Unlimited wrote:
On one CPU, after two days worth of burn-in, the time required to
complete 100 iterations on the stock number went down, as I would expect.

On a second CPU, the time actually *increased*. From 1 day, 18 hours,
and 55 minutes, to 1d 19h 13m!

I'd be surprised if burning in had an effect on iteration times.  
More likely, 100 iterations is not enough to get a truly accurate
timing.  

I don't know about Linux, but it has been noted before prime95
iteration times can vary a few percent from day to day.

If anyone knowledgeable would care to comment about mprime's suitability
for Q.A. 

mprime is great for QA.  It generates heat (by FPU use) and tons of
memory accesses.  If there are any hardware problems there, they are likely
to show up in an extended torture test.

and/or performance measurements, 

I'd guess its as good as but no better than any of hundreds
of publicly available benchmarks.

Best regards,
George 



Re: Mersenne: mprime for QA or performance?

1999-01-16 Thread Henrik Olsen

On Fri, 15 Jan 1999, George Woltman wrote:
 At 12:24 AM 1/13/99 -0800, Leu Enterprises Unlimited wrote:
 On one CPU, after two days worth of burn-in, the time required to
 complete 100 iterations on the stock number went down, as I would expect.
 
 On a second CPU, the time actually *increased*. From 1 day, 18 hours,
 and 55 minutes, to 1d 19h 13m!
 
 I'd be surprised if burning in had an effect on iteration times.  
 More likely, 100 iterations is not enough to get a truly accurate
 timing.  
I'd be extremely surprised, since the speed of the computations is
strongly linked to the processor clock, so either you're measuring clock
drift (and yes, that is temperature linked) or due to your small samples
the variance mentioned which is about 1% is actually noice.

-- 
Henrik Olsen,  Dawn Solutions I/S
URL=http://www.iaeste.dk/~henrik/
Get the rest there.




Re: Mersenne: mprime for QA or performance?

1999-01-15 Thread Brian J Beesley

 To me at least, the first part is obvious. As temperature goes up, so does
 resistance. The higher resistance requires a greater potential difference
 to send an adequeate(sp?) amount of current throught the processor.
 Without cooling, I would imagine the chip would eventually melt down. With
 cooling, I would imagine the voltage peaks out in some sort of gaussian
 curve.

For small temperature ranges then R = kT + c, the trouble is, for 
semiconductors, k can be negative. This is why semiconductors 
can ( do) burn out if overrun or undercooled - current can "run 
away" to very large values (essentially whatever the supply rails will 
stand before *they* step in  act as fuse wire)

When running digital chips very fast, what you need is a sharp 
rise/fall. Now you just *never* see nice square waves like those 
drawn in the text books, real signals tend to look like a sine wave. 
Obviously, in these circumstances, the higher the voltage, the 
steeper the rise  fall of the signal when it's changing from a 0 
state to a 1 state, or vice versa.

It's also why small die sizes can be run faster than large die sizes, 
the current is much smaller, so it takes less power to change from 
0 to 1 in a given time.

 Which brings to mind a theory about the second part, "speed increase". Too
 much heat will damage a processor. However I wonder if there is a
 relationship between some sort of particle drift (electron???) and
 resistance that "breaks in" or "burns in" a new processor? Much like the
 performance on a new engine after it has been carefully broken in after
 the first 1,000 miles...

I get the impression that failure rates on (non-overclocked) 
semiconductor device falls rapidly with time, at least over the first 
few hours. This is why quality systems suppliers "burn in" or "soak 
test" systems for 24-48 hours before shipping. It may be the case 
that minor flaws can act as "pinch points" which eventually cause 
burn out  failure, even though the chip actually measures within 
spec to begin with. Also atoms within crystals can "creep" causing 
minute changes to the circuit characteristics, I would have thought 
that the most significant effect would be the first time the circuit 
gets to its "normal" operating temperature.


Regards
Brian Beesley



Re: Mersenne: mprime for QA or performance?

1999-01-14 Thread Leu Enterprises Unlimited


To my knowledge, the effect on performance of burn-in has not really
been widely discussed or explored; at least not on any of the main
overclocking forums which I am aware of. 

Which is partly why I raised the question.

Some things are known to happen; and I have observed them myself. First,
the amount of voltage required to work at a given speed decreases.
Secondly, the odds of the chip working at a higher speed increases.

This much is common knowledge. No one has satisfactorily explained 
these results, though many theories abound.

There are also rumors that the higher-quality chip cores, which get put
into the C333A/366/400 and Pentium II's are measureably faster. These 
rumors have not been confirmed to my knowledge.

What I was saying was that an improvement in speed was not unexpected;
I.e. it is consistant with the expected improvement in the chip due 
to burn-in. These are also my own expectations.

It would be difficult to classify unexplored territory as "common knowledge".
Perhaps it's common knowledge to some; it's just not mentioned in the
more popular websites or forums.

-dwight-

 From [EMAIL PROTECTED]  Wed Jan 13 18:17:14 1999
 
 
 This brings up an interesting question (Well at least to me it is). I
 notice that you refer to the iteration time "going down" as an expected
 event. Are you saying that a new CPU is expect to get faster (If ever so
 slightly) after an initial "Burn in" time??? Perhaps that is a bit of
 common knowledge I have never heard of. Can someone explain this in
 greater detail?
 
 Thank You,
 
 Chuck
 
 On Wed, 13 Jan 1999, Leu Enterprises Unlimited wrote:
 
  
  Greetings!
  
  Aside from it's normal uses, I've been looking at using mprime as
  a QA and benchmarking tool. 
  
  This is particularly important for overclocked Beowulf clusters, 
  where reliability is a must. And given that the price of these
  is now so cheap (with a Gigaflop being achieved for under $10,000),
  adaquate Q.A. procedures are a must. Re: my observations on
  www.supercomputer.org.
  
  I've recently noticed something interesting. Namely, with the so-called
  effects of "burning-in" a CPU via mprime's torture test, and the
  resulting effects on the CPU speed (as determined by the mprime
  time test).
  
  On one CPU, after two days worth of burn-in, the time required to
  complete 100 iterations on the stock number went down, as I would expect.
  
  On a second CPU, the time actually *increased*. From 1 day, 18 hours,
  and 55 minutes, to 1d 19h 13m!
  
  So this leads me to wonder if what I'm seeing is real, or whether
  mprime is really sensitive enough to be used here. The O.S. being 
  used is Linux.
  
  If mprime is indeed seeing some real effects here, I think people 
  need to be aware of it. Or even if it's not suitable for this, people 
  should also know.
  
  If anyone knowledgeable would care to comment about mprime's suitability
  for Q.A. and/or performance measurements, I would appreciate it.
  
  Thanks!
  
  -dwight-
  



Re: Mersenne: mprime for QA or performance?

1999-01-14 Thread poke



 Some things are known to happen; and I have observed them myself. First,
 the amount of voltage required to work at a given speed decreases.
 Secondly, the odds of the chip working at a higher speed increases.

To me at least, the first part is obvious. As temperature goes up, so does
resistance. The higher resistance requires a greater potential difference
to send an adequeate(sp?) amount of current throught the processor.
Without cooling, I would imagine the chip would eventually melt down. With
cooling, I would imagine the voltage peaks out in some sort of gaussian
curve. Which beings to mind, would the potential difference delta be a
good measure of the quality of your cooling?

(For the non electrical people, I am using voltage and "potential
difference" interchangeably, they are the same thing but potential
difference says it better IMHO)

Which brings to mind a theory about the second part, "speed increase". Too
much heat will damage a processor. However I wonder if there is a
relationship between some sort of particle drift (electron???) and
resistance that "breaks in" or "burns in" a new processor? Much like the
performance on a new engine after it has been carefully broken in after
the first 1,000 miles...


 This much is common knowledge. No one has satisfactorily explained 
 these results, though many theories abound.

Case in point...



 There are also rumors that the higher-quality chip cores, which get put
 into the C333A/366/400 and Pentium II's are measureably faster. These 
 rumors have not been confirmed to my knowledge.


Tighter packed transistors = less ability to dissipate heat = more
resistance...


 It would be difficult to classify unexplored territory as "common knowledge".
 Perhaps it's common knowledge to some; it's just not mentioned in the
 more popular websites or forums.

Almost everything that I have learned beyond 1st grade was by assuming
that it was common knowledge. "If someone else understands it, I should be
capable of understanding it too!"... 



 ~~~
: WWW: http://www.silverlink.net/poke : Boycott Microsot:
: E-Mail:  [EMAIL PROTECTED]  : http://www.vcnet.com/bms:
 ~~~



Re: Mersenne: mprime for QA or performance?

1999-01-14 Thread Leu Enterprises Unlimited


 From [EMAIL PROTECTED]  Thu Jan 14 13:25:19 1999
 To: Leu Enterprises Unlimited [EMAIL PROTECTED]
 cc: [EMAIL PROTECTED]
 Subject: Re: Mersenne: mprime for QA or performance?
 
 
 
  Some things are known to happen; and I have observed them myself. First,
  the amount of voltage required to work at a given speed decreases.
  Secondly, the odds of the chip working at a higher speed increases.
 
 To me at least, the first part is obvious. As temperature goes up, so does
 resistance. The higher resistance requires a greater potential difference
 to send an adequeate(sp?) amount of current throught the processor.
 Without cooling, I would imagine the chip would eventually melt down. With
 cooling, I would imagine the voltage peaks out in some sort of gaussian
 curve. Which beings to mind, would the potential difference delta be a
 good measure of the quality of your cooling?

That's not been my experience. You can put on superb cooling, and crank
the voltage up all you want. It won't help a bad chip. You will only
fry the chip.

Don't forget that the CPU parts are spec'd at 2.5V; with a typical tolerance
of +/- 20 %. Running a C300A at 3.0+ V for long is likely to cause great
disappointment.

Furthermore, you are not guaranteed to run reliably at 504 MHz, if your
CPU handles 450 MHz @ 2.0 V. I have several such CPU's. And, FWIW, the
point of failure appears to be the FPU (no - not the L2 cache, like
most people assume).

 ...
 
  It would be difficult to classify unexplored territory as "common knowledge".
  Perhaps it's common knowledge to some; it's just not mentioned in the
  more popular websites or forums.
 
 Almost everything that I have learned beyond 1st grade was by assuming
 that it was common knowledge. "If someone else understands it, I should be
 capable of understanding it too!"... 

Ok - please explain to me members of the opposite sex. They certainly seem
to understand each other. :)

Seriously, I wouldn't classify the original subject as common knowledge
among overclockers at all. Or the latter subject, but let's not digress...

Back to my orginal point - is there truely no one on this list who
can knowledgeably state that the precision of mprime is suitable, or
not suitable, for these timing purposes? One area of concern is round-off
error. But there may be others as well.

TIA!

-dwight-


To learn how to build your own supercomputer, for under $10,000, go to:
www.supercomputer.org




Mersenne: mprime for QA or performance?

1999-01-13 Thread Leu Enterprises Unlimited


Greetings!

Aside from it's normal uses, I've been looking at using mprime as
a QA and benchmarking tool. 

This is particularly important for overclocked Beowulf clusters, 
where reliability is a must. And given that the price of these
is now so cheap (with a Gigaflop being achieved for under $10,000),
adaquate Q.A. procedures are a must. Re: my observations on
www.supercomputer.org.

I've recently noticed something interesting. Namely, with the so-called
effects of "burning-in" a CPU via mprime's torture test, and the
resulting effects on the CPU speed (as determined by the mprime
time test).

On one CPU, after two days worth of burn-in, the time required to
complete 100 iterations on the stock number went down, as I would expect.

On a second CPU, the time actually *increased*. From 1 day, 18 hours,
and 55 minutes, to 1d 19h 13m!

So this leads me to wonder if what I'm seeing is real, or whether
mprime is really sensitive enough to be used here. The O.S. being 
used is Linux.

If mprime is indeed seeing some real effects here, I think people 
need to be aware of it. Or even if it's not suitable for this, people 
should also know.

If anyone knowledgeable would care to comment about mprime's suitability
for Q.A. and/or performance measurements, I would appreciate it.

Thanks!

-dwight-



Re: Mersenne: mprime for QA or performance?

1999-01-13 Thread poke


This brings up an interesting question (Well at least to me it is). I
notice that you refer to the iteration time "going down" as an expected
event. Are you saying that a new CPU is expect to get faster (If ever so
slightly) after an initial "Burn in" time??? Perhaps that is a bit of
common knowledge I have never heard of. Can someone explain this in
greater detail?

Thank You,

Chuck

On Wed, 13 Jan 1999, Leu Enterprises Unlimited wrote:

 
 Greetings!
 
 Aside from it's normal uses, I've been looking at using mprime as
 a QA and benchmarking tool. 
 
 This is particularly important for overclocked Beowulf clusters, 
 where reliability is a must. And given that the price of these
 is now so cheap (with a Gigaflop being achieved for under $10,000),
 adaquate Q.A. procedures are a must. Re: my observations on
 www.supercomputer.org.
 
 I've recently noticed something interesting. Namely, with the so-called
 effects of "burning-in" a CPU via mprime's torture test, and the
 resulting effects on the CPU speed (as determined by the mprime
 time test).
 
 On one CPU, after two days worth of burn-in, the time required to
 complete 100 iterations on the stock number went down, as I would expect.
 
 On a second CPU, the time actually *increased*. From 1 day, 18 hours,
 and 55 minutes, to 1d 19h 13m!
 
 So this leads me to wonder if what I'm seeing is real, or whether
 mprime is really sensitive enough to be used here. The O.S. being 
 used is Linux.
 
 If mprime is indeed seeing some real effects here, I think people 
 need to be aware of it. Or even if it's not suitable for this, people 
 should also know.
 
 If anyone knowledgeable would care to comment about mprime's suitability
 for Q.A. and/or performance measurements, I would appreciate it.
 
 Thanks!
 
   -dwight-
 

 ~~~
: WWW: http://www.silverlink.net/poke : Boycott Microsot:
: E-Mail:  [EMAIL PROTECTED]  : http://www.vcnet.com/bms:
 ~~~