From: Anthony Gardner <[EMAIL PROTECTED]>
My site depends on this functionality and so any feedback would be greatly appreciated.

I have to handle characters with umlats etc a fair bit but I store my system in xml files so don't know about mysql. I have found it is best to try and use utf-8 all the way through for consistancy.


I have had the occasional utf-8 glitch with perl 5.8.x where it gets mixed up as to whether a string is utf-8 or not. If I know the string is utf-8 and have found that perl is confused, I use:

  use Encode;
  decode('utf-8', $string);

This means decode the string from utf-8 to perl's internal string type (which is also utf-8). It usually gets it right after that. I had a talk to the perl developers about it and they fixed it in CVS (it's to do with a flag set somewhere in the underlying data structures) and it worked but I have since rebuilt my system twice and lost track of it.

Also check that it isn't the browser munging the utf-8, you need to get the headers right and sometimes a metatag is needed. AxTraceIntermediate is a great tool for tracking down the source of charset problems. (I also occasionally use ethereal on the client side to check what is actually being sent from the webserver).

matt



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



Reply via email to