Jim Jewett schrieb:
> On 6/5/07, "Martin v. Löwis" <[EMAIL PROTECTED]> wrote:
>> > Always normalizing would have the advantage of simplicity (no
>> > matter what the encoding, the result is the same), and I think
>> > that is the real path of least surprise if you sum over all
>> > surprises.
> 
>> I'd like to repeat that this is out of scope of this PEP, though.
>> This PEP doesn't, and shouldn't, specify how string literals get
>> from source to execution.
> 
> I see that as a gray area.

Please read the PEP title again. What is unclear about
"Supporting Non-ASCII Identifiers"?

> Unicode does say pretty clearly that (at least) canonical equivalents
> must be treated the same.

Chapter and verse, please?

> In theory, this could be done only to identifiers, but then it needs
> to be done inline for getattr.

Why that? The caller of getattr would need to apply normalization in
case the input isn't known to be normalized?

> Since we don't want the results of (str1 == str2) to change based on
> context, I think string equality also needs to look at canonicalized
> (though probably not compatibility) forms.  This in turn means that
> hashing a unicode string should first canonicalize it.  (I believe
> that is a change from 2.x.)

And you think this is still within the scope of the PEP?

Please, if you want that to happen, write your own PEP.

Regards,
Martin
_______________________________________________
Python-3000 mailing list
Python-3000@python.org
http://mail.python.org/mailman/listinfo/python-3000
Unsubscribe: 
http://mail.python.org/mailman/options/python-3000/archive%40mail-archive.com

Reply via email to