res wrote:
> On 07.09.2005 04:02, Reed Hedges wrote:
> 
>>It would be a simpler protocol to use just unicode.  
> 
> 
> Still, you have to decide for some encoding of Unicode...


Right, sorry, when I say Unicode, I mean something we choose (one) that
is the most-inclusive of all character sets/languages.  Such an
all-inclusive encoding would likely be a multibyte encoding, I presume.


> 
> 
>>The problem is that
>>(1) it's an extra pain in the neck for developers (i.e.
>>programmer-users) to worry about, especially coming from worlds like C
>>and descendents where the notion of non-ascii character sets are a very
>>recent addition and not a natural part of the language, and (2) if we
>>add the extra "encoding" field, then we can put off actually
>>transitioning to unicode (or whatever) by saying that currently the only
>>valid encoding is ascii. Then later we can do unicode support, but also
>>support ascii for the old fashioned/lazy/whatever.
> 
> 
> If you want to be lazy and ASCII compatibility, use UTF-8. What good
> reasons are there to allow choice of encoding at all?
> 

_______________________________________________
vos-d mailing list
vos-d@interreality.org
http://www.interreality.org/cgi-bin/mailman/listinfo/vos-d

Reply via email to