I agree on the tendency to manipulate benchmarks, but that doesn't mean benchmarks aren't a useful tool. How else can we evaluate performance? I guess I'm most curious about what the two projects might be able to learn from each other. I would suspect MINA's APIs are significantly easier to use than Grizzly's, for example, and it wouldn't surprise me at all if Sun's benchmarks were somewhat accurate. I hate Sun's java.net projects as much as the next guy, but that doesn't mean there's not an occasional jewel in there.
It would at least be worth running independent tests. If the differences are even close to the claims, it would make a ton of sense to just copy their ideas. No need for too much pride on either side! Just seems like they've put a ton of work into rigorously analyzing the performance tradeoffs of different design decisions, and it might make sense to take advantage of that. If their benchmarks are off and MINA performs better, then they should go ahead and copy MINA. That's all assuming the performance tweaks don't make the existing APIs unworkable. -Adam On 5/24/07, Alex Karasulu <[EMAIL PROTECTED]> wrote:
On 5/24/07, Mladen Turk <[EMAIL PROTECTED]> wrote: > > Adam Fisk wrote: > > The slides were just posted from this Java One session claiming Grizzly > > blows MINA away performance-wise, and I'm just curious as to people's > views > > on it. They present some interesting ideas about optimizing selector > > threading and ByteBuffer use. > > > > > http://developers.sun.com/learning/javaoneonline/j1sessn.jsp?sessn=TS-2992&yr=2007&track=5 > > > > I love the slide 20! > JFA finally admitted that Tomcat's APR-NIO is faster then JDK one ;) > However last time I did benchmarks it was much faster then 10%. > > > > > Maybe someone could comment on the performance improvements in MINA > > 2.0? > > He probably compared MINA's Serial IO, and that is not usable > for production (jet). I wonder how it would look with real > async http server. > Nevertheless, benchmarks are like assholes. Everyone has one. Exactly! Incidentally SUN has been trying to attack several projects via the performance angle for some time now. Just recently I received a cease and desist letter from them when I compiled some performance metrics. The point behind it is was that we were not correctly configuring their products. I guess they just want to make sure things are setup to their advantage. That's what all these metrics revolve around and if you ask me they're not worth a damn. There is a million ways to make one product perform better than another depending on configuration, environment and the application. However is raw performance metrics as important as a good flexible design? Alex
