Shannon -jj Behrens wrote: >> I haven't done much testing around this, so I don't know. This could >> potentially be done with a wrapper around MultiDict too, that lazily >> decodes the values. > > This can be a frustrating subject. Did you know if you set > accept-encodings="US-ASCII" in the form, but the user tries to submit > Japanese characters, Firefox will send them as HTML entities like > Ӓ whereas IE will ignore you and send UTF-8 (assuming your page > was originally in UTF-8)? Bleh!
Oh the pain! Does it start HTML quoting everything once it starts adding entities (but not before)? Too bad there isn't a unicode option for url escaping. And if you do multipart/form-data, does the encoding show up there? Does the cgi module keep that information or throw it away? Clearly it would be nice to do this right, but "nice" isn't "easy". Probably heuristics are also called for, as well as looking at the user agent. And what about IE 7? This would probably be the kind of thing that would be nice to steal from somewhere else (maybe a framework in another language). I'm guessing Perl isn't a good source, they're probably as scrappy about this stuff as we are; ditto PHP. Java? They'd care. -- Ian Bicking | [EMAIL PROTECTED] | http://blog.ianbicking.org _______________________________________________ Paste-users mailing list [email protected] http://webwareforpython.org/cgi-bin/mailman/listinfo/paste-users
