Michel Fortin wrote:
On 2009-04-27 07:04:06 -0400, Frank Benoit <[email protected]>
said:
M$ and Java have chosen to use utf16 as their default Unicode character
encoding. I am sure, the decision was not made without good reasoning.
What are their arguments?
Why does D propagate utf8 as the default?
E.g.
Exception.msg
Object.toString()
new std.stream.File( char[] )
The argument at the time was that they were going to work directly with
Unicode code points, thus simplifying things. Then Unicode extended to
cover even more characters, and at some point 16 bit became
insufficient; 16-bit encoding of Unicode became UTF-16, and surrogate
pairs were added to allow it to contain even higher code points, making
16-bit unicode a variable-size character encoding now known as UTF-16.
So it turns out that those in the early years of Unicode who made that
choice made it for reasons that no longer exist. Today, variable-size
UTF-16 makes it as hard to calculate a string length and do random
access in a string as UTF-8. In practice, many frameworks just ignore
the problem and are happy counting each code of a surrogate pair as two
characters, as they always did, but that behaviour isn't exactly correct.
To get the benefit those framework/language designers though they'd get
at the time, we'd have to go with UTF-32, but then storing strings
become immensely wasteful. And I'm not counting that most data exchange
these days have standardized with UTF-8, rarely you'll encounter UTF-16
in the wild (and when you do, you have take care about UTF-16 LE and
BE), and even more rare is UTF-32. And that's not counting that perhaps,
one day, Unicode will grow again and fall outside of its 32-bit range...
although that may have to wait until learn a few extraterrestrial
languages. :-)
So the D solution, which is to use UTF-8 everywhere while still
supporting string operations using UTF-16 and UTF-32, looks very good to
me. What I actually do is use UTF-8 everywhere, and sometime when I need
to easily manipulate characters I use UTF-32. And I use UTF-16 for
dealing with APIs expecting it, but not for much else.
Well put. I distinctly remember the hubbub around Java's UTF16 support
that was solving all of strings' problems, followed by the embarrassed
silence upon the introduction of UTF32.
Andrei