On Dec 23, 2009, at 1:19 PM, William Edney wrote:

> From following the thread earlier, it didn't seem like anyone would have a 
> major objection to using UTF-8 instead of UTF-16 for localStorage on Webkit.

Maciej did have an objection to doing this in a simplistic way (always using 
UTF-8). He wrote:

>> ECMAScript Strings are essentially sequences of arbitrary 16-bit values. 
>> Sometimes Web apps take advantage of this to use a String as a hacky way to 
>> represent binary data. I don't think we should reject such strings 
>> arbitrarily.
>> 
>> I think it would be fine to use a more compact encoding opportunistically, 
>> as long as we can still handle an arbitrary JavaScript String. Perhaps we 
>> should use UTF-8 if and only if the conversion succeeds, or perhaps even use 
>> Latin1 as the alternative.

If someone can find a way to do use a more compact encoding 
“opportunistically”, but still allow any arbitrary 16-bit value for any 
character, I’m open to that. If not, please note that Maciej’s objection is one 
that I share.

    -- Darin

_______________________________________________
webkit-dev mailing list
[email protected]
http://lists.webkit.org/mailman/listinfo.cgi/webkit-dev

Reply via email to