RE: How will software source code represent 21 bit unicode characters?

2001-04-26 Thread addison
On Mon, 23 Apr 2001, Mike Brown wrote: > A char corresponds to a Unicode value -- a UTF-16 code value, which could > either represent a Unicode character or one half of a surrogate pair. In the > latter case, it would take a sequence of two "char"s to make one Unicode > character. It is my under

RE: How will software source code represent 21 bit unicode characters?

2001-04-23 Thread Mike Brown
William Overington wrote: > In Java source code one may currently represent a 16 bit > unicode character by using \u where each h is any > hexadecimal character. > > How will Java, and maybe other languages, represent 21 bit unicode > characters? \u in Java source becomes a value of the

RE: How will software source code represent 21 bit unicode characters?

2001-04-17 Thread Martin Duerst
At 09:29 01/04/17 -0500, [EMAIL PROTECTED] wrote: > > In a perfect world, we would probably have an enclosing symbol (e.g. > > '\<4E00>') so that the number can be variable length. > > >In Perl the notation is \x{...}, where ... is hexdigit sequence: >\x{41} is LATIN CAPITAL LETTER A while \x{263a

Re: How will software source code represent 21 bit unicode characters?

2001-04-17 Thread Marcin 'Qrczak' Kowalczyk
Tue, 17 Apr 2001 07:33:16 +0100, William Overington <[EMAIL PROTECTED]> pisze: > In Java source code one may currently represent a 16 bit unicode character > by using \u where each h is any hexadecimal character. > > How will Java, and maybe other languages, represent 21 bit unicode > chara

RE: How will software source code represent 21 bit unicode characters?

2001-04-17 Thread Marco Cimarosti
William Overington wrote: > Has this matter already been addressed anywhere? I think the C standard is in the process of making a decision about this. If memory helps, we will have escapes like '\u' and '\U'. I am sure that some people on this list have precise and up-to-date info at