Thanks for the Java API ref.  But, I'm curious as to how or where invalid UTF-8 
sequences come about; is it primarily a hacker thing?  I see the Java Character 
API has an isValidCodePoint() method.  Do I just run each code point through 
that?

Thanks,
al



________________________________
 From: Brion Vibber <[email protected]>
To: Al Johnson <[email protected]>; MediaWiki announcements and site admin 
list <[email protected]> 
Sent: Wednesday, December 26, 2012 8:40 PM
Subject: Re: [MediaWiki-l] Normalization Code
 

On Wed, Dec 26, 2012 at 2:20 PM, Al Johnson wrote:

I need to make sure a backend Java process is doing the same UTF normalization 
that is done for edit text.  Grep'ing for 'normaliz' brings up a lot and I'm 
not a php dev.  Can someone point me to a key php module and/or function?
>

The PHP code for this is in includes/normal -- luckily you shouldn't have to 
replicate most of that code which is nasty and low-level.

For the most part, you want to do two things:
* make sure the input is valid UTF-8
* normalize any composition character sequences to 'normalization form C'

Reading data in from a UTF-8 input stream into a Java string should already 
take care of making sure it's valid UTF-8. :) If you want to treat invalid 
input the same, make sure that invalid UTF-8 sequences get converted to the 
'replacement character' U+FFFD rather than throwing an exception.

It looks like you should be able to use the java.text.Normalizer class to 
convert to NFC: 
<http://docs.oracle.com/javase/tutorial/i18n/text/normalizerapi.html>

You might or might not prefer to use the Java version of the ICU library to do 
the same thing, it might be more up to date: 
<http://icu-project.org/apiref/icu4j/com/ibm/icu/text/Normalizer.html>

-- brion
_______________________________________________
MediaWiki-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l

Reply via email to