On 2017-11-09 19:04, Nicolas Cellier wrote:
2017-11-09 18:02 GMT+01:00 Raffaello Giulietti
<[email protected] <mailto:[email protected]>>:
Anyway relying upon Float equality should allways be subject to
extreme caution and examination
For example, what do you expect with plain old arithmetic in mind:
a := 0.1.
b := 0.3 - 0.2.
a = b
This will lead to (a - b) reciprocal = 3.602879701896397e16
If it is in a Graphics context, I'm not sure that it's the
expected scale...
a = b evaluates to false in this example, so no wonder (a - b)
evaluates to a big number.
Writing a = b with floating point is rarely a good idea, so asking about
the context which could justify such approach makes sense IMO.
Simple contexts, like the one which is the subject of this trail, are
the one we should strive at because they are the ones most likely used
in day-to-day working. Having useful properties and regularity for
simple cases might perhaps cover 99% of the everyday usages (just a
dishonestly biased estimate ;-) )
Complex contexts, with heavy arithmetic, are best dealt by numericists
when Floats are involved, or with unlimited precision numbers like
Fractions by other programmers.
But the example is not plain old arithmetic.
Here, 0.1, 0.2, 0.3 are just a shorthands to say "the Floats closest
to 0.1, 0.2, 0.3" (if implemented correctly, like in Pharo as it
seems). Every user of Floats should be fully aware of the implicit
loss of precision that using Floats entails.
Yes, it makes perfect sense!
But precisely because you are aware that 0.1e0 is "the Float closest to
0.1" and not exactly 1/10, you should then not be surprised that they
are not equal.
Indeed, I'm not surprised. But then
0.1 - (1/10)
shall not evaluate to 0. If it evaluates to 0, then the numbers shall
compare as being equal.
The surprise lies in the inconsistency between the comparison and the
subtraction, not in the isolated operations.
I agree that following assertion hold:
self assert: a ~= b & a isFloat & b isFloat & a isFinite & b
isFinite ==> (a - b) isZero not
The arrow ==> is bidirectional even for finite Floats:
self assert: (a - b) isZero not & a isFloat & b isFloat & a isFinite & b
isFinite ==> a ~= b
But (1/10) is not a Float and there is no Float that can represent it
exactly, so you can simply not apply the rules of FloatingPoint on it.
When you write (1/10) - 0.1, you implicitely perform (1/10) asFloat - 0.1.
It is the rounding operation asFloat that made the operation inexact, so
it's no more surprising than other floating point common sense
See above my observation about what I consider surprising.
In the case of mixed-mode Float/Fraction operations, I personally
prefer reducing the Fraction to a Float because other commercial
Smalltalk implementations do so, so there would be less pain porting
code to Pharo, perhaps attracting more Smalltalkers to Pharo.
Mixed arithmetic is problematic, and from my experience mostly happens
in graphics in Smalltalk.
If ever I would change something according to this principle (but I'm
not convinced it's necessary, it might lead to other strange side effects),
maybe it would be how mixed arithmetic is performed...
Something like exact difference like Martin suggested, then converting
to nearest Float because result is inexact:
((1/10) - 0.1 asFraction) asFloat
This way, you would have a less surprising result in most cases.
But i could craft a fraction such that the difference underflows, and
the assertion a ~= b ==> (a - b) isZero not would still not hold.
Is it really worth it?
Will it be adopted in other dialects?
As an alternative, the Float>>asFraction method could return the
Fraction with the smallest denominator that would convert to the
receiver by the Fraction>>asFloat method.
So, 0.1 asFraction would return 1/10 rather than the beefy Fraction it
currently returns. To return the beast, one would have to intentionally
invoke asExactFraction or something similar.
This might cause less surprising behavior. But I have to think more.
But the main point here, I repeat myself, is to be consistent and to
have as much regularity as intrinsically possible.
I think we have as much as possible already.
Non equality resolve more surprising behavior than it creates.
It makes the implementation more mathematically consistent (understand
preserving more properties).
Tell me how you are going to sort these 3 numbers:
{1.0 . 1<<60+1/(1<<60). 1<<61+1/(1<<61)} sort.
tell me the expectation of:
{1.0 . 1<<60+1/(1<<60). 1<<61+1/(1<<61)} asSet size.
A clearly stated rule, consistently applied and known to everybody, helps.
In presence of heterogeneous numbers, the rule should state the common
denominator, so to say. Hence, the numbers involved in mixed-mode
arithmetic are either all converted to one representation or all to the
other: whether they are compared or added, subtracted or divided, etc.
One rule for mixed-mode conversions, not two.
tell me why = is not a relation of equivalence anymore (not associative)
Ensuring that equality is an equivalence is always a problem when the
entities involved are of different nature, like here. This is not a new
problem and not inherent in numbers. (Logicians and set theorists would
have much to tell.) Even comparing Points and ColoredPoints is
problematic, so I have no final answer.
In Smalltalk, furthermore, implementing equality makes it necessary to
(publicly) expose much more internal details about an object than in
other environments.