> > But I'm not about to change the == operator to apply normalization > > first. It would affect too much (e.g. hashing). > > Yah, that's one reason why Jim Jewett and I lean to normalizing on the > way in for explicitly Unicode data. But since that's not going to > happen, I guess the thing is to get cracking on that library just in > case there's some help that Python itself could give.
There are issues with that as well. Concatenation would need to perform normalization, and then len(a+b) <> len(a)+len(b), for some a and b. Regards, Martin _______________________________________________ Python-3000 mailing list Python-3000@python.org http://mail.python.org/mailman/listinfo/python-3000 Unsubscribe: http://mail.python.org/mailman/options/python-3000/archive%40mail-archive.com