On Saturday, 28 June 2014 at 09:07:17 UTC, John Colvin wrote:
On Saturday, 28 June 2014 at 06:16:51 UTC, Walter Bright wrote:
On 6/27/2014 10:18 PM, Walter Bright wrote:
On 6/27/2014 4:10 AM, John Colvin wrote:
*The number of algorithms that are both numerically stable/correct and benefit
significantly from > 64bit doubles is very small.

To be blunt, baloney. I ran into these problems ALL THE TIME when doing
professional numerical work.

Sorry for being so abrupt. FP is important to me - it's not just about performance, it's also about accuracy.

When you need accuracy, 999 times out of 1000 you change the numerical technique, you don't just blindly upgrade the precision. The only real reason one would use 80 bits is when there is an actual need of adding values which differ for more than 16 orders of magnitude. And I've never seen this happen in any numerical paper I've read.

I still maintain that the need for the precision of 80bit reals is a niche demand. Its a very important niche, but it doesn't justify having its relatively extreme requirements be the default. Someone writing a matrix inversion has only themselves to blame if they don't know plenty of numerical analysis and look very carefully at the specifications of all operations they are using.

Couldn't agree more. 80 bit IS a niche, which is really nice to have, but shouldn't be the standard if we lose on performance.

Paying the cost of moving to/from the fpu, missing out on increasingly large SIMD units, these make everyone pay the price.

Especially the numerical analysts themselves will pay that price. 64 bit HAS to be as fast as possible, if you want to be competitive when it comes to any kind of numerical work.

Reply via email to