Darin Adler wrote:

On Jul 27, 2007, at 4:03 AM, Alexey Proskuryakov wrote:

On 7/27/07 1:51 PM, "Simon Hausmann" <[EMAIL PROTECTED]> wrote:

Does anybody know/remember why JSChar is defined to wchar_t on Windows and if
it is still needed?

I think this was/is needed to match ICU's definition of UChar ("Define UChar to be wchar_t if that is 16 bits wide; always assumed to be unsigned. If wchar_t is not 16 bits wide, then define UChar to be uint16_t. This makes the definition of UChar platform-dependent but allows direct string type compatibility with platforms with 16-bit wchar_t types.")

That's correct. And for the same reasons we should follow the same pattern for UChar, even when ICU is not involved.

I think that UnicodeQt4.h is the file that should be changed, not JSStringRef.h.

You say WebKit tries to follow ICU's UChar. If so, the real size of wchar_t must be examined instead of just checking for WIN32 or _WIN32. Otherwise the problem appears for every windows/g++ build without these two defines. In the case of Qt (which doesn't use ICU), the problem can be easily solved: define UChar in UnicodeQt4.h according to ICU's UChar - see Simon's last email - but what about other platforms that don't have their own Unicode implementation?

Thanks,

Artem
_______________________________________________
webkit-dev mailing list
webkit-dev@lists.webkit.org
http://lists.webkit.org/mailman/listinfo/webkit-dev

Reply via email to