Hi Alan,

I gave the "continue ->else if " a try, it appears the server vm obviously likes the "continue" approach (it is consistent with what we experienced in the past when we did similar approach for ascii, in which we separate/singled the ascii
path out). So I guess we probably want to keep the continue here.

Here are the data. dbcs2_new is to replace the continue with else if and the
dbcs2 is the "continue".

http://cr.openjdk.java.net/~sherman/7183053/dbcs2_new
http://cr.openjdk.java.net/~sherman/7183053/dbcs2

-Sherman

On 07/13/2012 10:09 AM, Xueming Shen wrote:
In JDK7, the decoder and encoder implementation of most of our single-byte charsets and UTF-8 charset are optimized to implement the internal interfce sun.nio.cs.ArrayDecoder/ Encoder to provide a fastpath for String.getBytes(...) and new String(byte[]...) operations. I
have an old blog regarding this optimization at

https://blogs.oracle.com/xuemingshen/entry/faster_new_string_bytes_cs

This rfe, as the followup for above changes, is to implement ArrayDe/Encoder for most of the sun.nio.cs.ext.DoubleByte based double-byte charsets. Here is the webrev

http://cr.openjdk.java.net/~sherman/7183053/webrev
I've taken a pass over this and it's great to see DoubleByte.Decoder/Encoder implementing sun.nio.cs.ArrayDecoder/Encoder. The results looks good too, a small number of regressions (Big5 at len=32 for example) but this is a micro benchmark and I'm sure there are fluctuations. I don't see anything obviously wrong with the EBCDIC changes I'd need a history book to remember how the shifts between DBCS and SBCS. I think our tests our good for this area so I'm happy. One minor nit is the continue in both encode methods, I think it would be cleaner to use "else if (bb ..." instead.


The continue might make the vm happy, but this is the code I did last Oct, so I might be
wrong. Will give a couple run later with "else"


Reply via email to