On 18/08/2012 21:22, wxjmfa...@gmail.com wrote:
Le samedi 18 août 2012 20:40:23 UTC+2, rusi a écrit :
On Aug 18, 10:59 pm, Steven D'Aprano <steve

+comp.lang.pyt...@pearwood.info> wrote:

On Sat, 18 Aug 2012 08:07:05 -0700, wxjmfauth wrote:

Is there any reason why non ascii users are somehow penalized compared

to ascii users?



Of course there is a reason.



If you want to represent 1114111 different characters in a string, as

Unicode supports, you can't use a single byte per character, or even two

bytes. That is a fact of basic mathematics. Supporting 1114111 characters

must be more expensive than supporting 128 of them.



But why should you carry the cost of 4-bytes per character just because

someday you *might* need a non-BMP character?



I am reminded of: 
http://answers.microsoft.com/thread/720108ee-0a9c-4090-b62d-bbd5cb1a7605



Original above does not open for me but here's a copy that does:



http://onceuponatimeinindia.blogspot.in/2009/07/hard-drive-weight-increasing.html

I thing it's time to leave the discussion and to go to bed.

In plain English, duck out cos I'm losing.


You can take the problem the way you wish, Python 3.3 is "slower"
than Python 3.2.

I'll ask for the second time. Provide proof that is acceptable to everybody and not just yourself.


If you see the present status as an optimisation, I'm condidering
this as a regression.

Considering does not equate to proof. Where are the figures which back up your claim?


I'm pretty sure a pure ucs-4/utf-32 can only be, by nature,
the correct solution.

I look forward to seeing your patch on the bug tracker. If and only if you can find something that needs patching, which from the course of this thread I think is highly unlikely.



To be extreme, tools using pure utf-16 or utf-32 are, at least,
considering all the citizen on this planet in the same way.

jmf



--
Cheers.

Mark Lawrence.

--
http://mail.python.org/mailman/listinfo/python-list

Reply via email to