One important usecase is the possibility to communicate with a backend without loosing precision.
Cheers Ralf. On Dec 13, 2007 6:07 AM, Gordon Smith <[EMAIL PROTECTED]> wrote: > > > > > > > > The next version of ECMAScript will probably have a 'decimal' datatype for > doing decimal math. Using this datatype, 0.3 + 0.7 would be exactly 1.0, not > something like 0.9999999999999997 as you currently get due to conversion > from decimal to binary fractions. > > This datatype would probably support additional precision as well. Number > only gives you 15 or 16 signficant digits. But if you had, say, 34, you > could represent up to $99,999,999,999,999,999,999,999,999,999,999.99 > exactly, and that's pretty large! > > The Player team is thinking about how to introduce a type like this even > before the ECMAScript spec is complete, hopefully in a way that will be > compatible with the spec. They'd like to gather some input on developers' > requirements for decimal math. Some questions to think about are... > > What is your use case? Financial calculations? Scientific calculations? > > Are you mainly interested in calculating with decimal rather than binary > fractions, or in having more significant digits, or are both important? > > Do you need support for an arbitrary number of significant digits (i.e., > "infinite precision")? > > If not, how many significant digits are sufficient? > > Do you need programmatic control over how much precision is used in > calculations (e.g., rounding to 5 decimal places in every intermediate > operation)? > > Do you need programmatic control over how rounding works? (Round down, round > up, round to nearest, what happens with 1.5, etc.) > > Do you care about whether a new type like 'decimal' gets automatically > coerced to other types like Number, int, and uint? > > - Gordon

