More justification for this... Not inlining the handles would be like invokevirtual not emitting its type check inline and doing that as a separate CALL. Handles are our way to tell the JVM what native operations go with an invokedynamic. Not emitting those operations unconditionally into the compiled code treats invokedynamic as a second class citizen. Fair to say?
- Charlie (mobile) On Oct 17, 2011 11:52 AM, "Charles Oliver Nutter" <head...@headius.com> wrote: > Yeah that does sound like the right approach the more I think about it. The > invokedynamic operation and the handles that wire it up should never be > separated. Pathology aside, I know the JRuby logic would remain pretty small > in almost every case. And pathological cases could be detected with some > kind of MH graph-walking calculation (with a suitably large budget since > handles individually are maybe a couple native operations each). > > Let me know if you want me to test something out. I *really* hope this can > get into u2. > > - Charlie (mobile) > On Oct 17, 2011 9:28 AM, "Christian Thalinger" < > christian.thalin...@oracle.com> wrote: > >> >> On Oct 17, 2011, at 4:14 PM, Charles Oliver Nutter wrote: >> >> Yeah that's a pretty big problem :) Indy almost becomes a nonstarter if >> these degraded cases are not made a lot better. The performance on this >> example becomes terrible, and it's common for Ruby methods to be larger and >> more complex than this, too. >> >> Yeah, I know. >> >> You say you don't know how to fix it, so perhaps we can brainstorm a bit? >> If we can't inline the call we ate least want it to perform like an >> uninlined method call. Could we still compile the MH chain and CALL it, so >> at least we're not doing c2i? Or perhaps we should always treat the MH chain >> as inlinable regardless of budgets, but not inline the eventual call when >> budgets are exceeded? The latter option sounds more correct to me; the MH >> chain should be considered a non-removable part of the invokedynamic >> operation and always inlined. That would avoid the degradation without >> blowing up code size. >> >> I didn't say I don't know how to fix it but rather what's the best >> approach. Internally we (Tom, John and I) already talked about this and I >> tried to do something what you suggest above: compile the MH chain and call >> it. It turned out it's not that easy (as I thought) and needs some special >> handling. >> >> An interesting idea is to always inline the "adaption" code of the MH >> chain but not the eventual call (or all calls in the chain). I would guess >> that for normal usage the added code size is insignificant but there can >> definitely be pathological cases. I try to get some data on that. >> >> -- Chris >> >> - Charlie (mobile) >> On Oct 17, 2011 4:58 AM, "Christian Thalinger" < >> christian.thalin...@oracle.com> wrote: >> >>> >>> On Oct 15, 2011, at 2:56 PM, Charles Oliver Nutter wrote: >>> >>> > I'm seeing something peculiar and wanted to run it by you folks. >>> > >>> > There are a few values that JRuby's compiler had previously been >>> > loading from instance fields every time they're needed. Specifically, >>> > fields like ThreadContext.runtime (the current JRuby runtime), >>> > Ruby.falseObject, Ruby.trueObject, Ruby.nilObject (false, true, and >>> > nil values). I figured I'd make a quick change today and have those >>> > instead be constant method handles bound into a mutable call site. >>> > >>> > Unfortunately, performance seems to be worse. >>> > >>> > The logic works like this: >>> > >>> > * ThreadContext is loaded to stack >>> > * invokedynamic, bootstrap just wires up an initialization method into >>> > a MutableCallSite >>> > * initialization method rebinds call site forever to a constant method >>> > handle pointing at the value (runtime/true/false/nil objects) >>> > >>> > My expectation was that this would be at least no slower (and >>> > potentially a tiny bit faster) but also less bytecode (in the case of >>> > true/false/nil, it was previously doing >>> > ThreadContext.runtime.getNil()/getTrue()/getFalse()). It seems like >>> > it's actually slower than walking those references, though, and I'm >>> > not sure why. >>> > >>> > Here's a couple of the scenarios in diff form showing bytecode before >>> > and bytecode after: >>> > >>> > Loading "runtime" >>> > >>> > ALOAD 1 >>> > - GETFIELD org/jruby/runtime/ThreadContext.runtime : >>> Lorg/jruby/Ruby; >>> > + INVOKEDYNAMIC getRuntime >>> > (Lorg/jruby/runtime/ThreadContext;)Lorg/jruby/Ruby; >>> > >>> [org/jruby/runtime/invokedynamic/InvokeDynamicSupport.getObjectBootstrap(Ljava/lang/invoke/MethodHandles$Lookup;Ljava/lang/St >>> > ring;Ljava/lang/invoke/MethodType;)Ljava/lang/invoke/CallSite; (6)] >>> > >>> > Loading "false" >>> > >>> > ALOAD 1 >>> > - GETFIELD org/jruby/runtime/ThreadContext.runtime : >>> Lorg/jruby/Ruby; >>> > - INVOKEVIRTUAL org/jruby/Ruby.getFalse ()Lorg/jruby/RubyBoolean; >>> > + INVOKEDYNAMIC getFalse >>> > (Lorg/jruby/runtime/ThreadContext;)Lorg/jruby/RubyBoolean; >>> > >>> [org/jruby/runtime/invokedynamic/InvokeDynamicSupport.getObjectBootstrap(Ljava/lang/invoke/MethodHandles$Lookup;Ljava/lang/String;Ljava/lang/invoke/MethodType;)Ljava/lang/invoke/CallSite; >>> > (6)] >>> > >>> > I think because these are now seen as invocations, I'm hitting some >>> > inlining budget limit I didn't hit before (and which isn't being >>> > properly discounted). The benchmark I'm seeing degrade is >>> > bench/language/bench_flip.rb, and it's a pretty significant >>> > degradation. Only the "heap" version shows the degradation, and it >>> > definitely does have more bytecode...but the bytecode with my patch >>> > differs only in the way these values are being accessed, as shown in >>> > the diffs above. >>> > >>> > Before: >>> > user system >>> > total real >>> > 1m x10 while (a)..(!a) (heap) 0.951000 0.000000 >>> > 0.951000 ( 0.910000) >>> > user system >>> > total real >>> > 1m x10 while (a)..(!a) (heap) 0.705000 0.000000 >>> > 0.705000 ( 0.705000) >>> > user system >>> > total real >>> > 1m x10 while (a)..(!a) (heap) 0.688000 0.000000 >>> > 0.688000 ( 0.688000) >>> > user system >>> > total real >>> > >>> > After: >>> > user system >>> > total real >>> > 1m x10 while (a)..(!a) (heap) 2.350000 0.000000 >>> > 2.350000 ( 2.284000) >>> > user system >>> > total real >>> > 1m x10 while (a)..(!a) (heap) 2.128000 0.000000 >>> > 2.128000 ( 2.128000) >>> > user system >>> > total real >>> > 1m x10 while (a)..(!a) (heap) 2.115000 0.000000 >>> > 2.115000 ( 2.116000) >>> > user system >>> > total real >>> > >>> > You can see the degradation is pretty bad. >>> > >>> > I'm concerned because I had hoped that invokedynamic + mutable call >>> > site + constant handle would always be faster than a field >>> > access...since it avoids excessive field accesses and makes it >>> > possible for Hotspot to fold those constants away. What's going on >>> > here? >>> >>> I looked into this and the main issue here is an old friend: slow >>> invokes of non-inlined MH call sites. The problem is that you trade a >>> normal invoke (to a field load?) with a MH invoke. If the normal invoke >>> doesn't get inlined we're good but if the MH invoke doesn't get inlined >>> we're screwed (since we are still doing the C2I-I2C dance). >>> >>> I refactored the benchmark a little (moved stack and heap loops into its >>> own methods and only do 5 while-loops instead of 11; that inlines all calls >>> in that method) and the performance is like you had expected (a little >>> faster): >>> >>> 32-bit: >>> >>> before: >>> >>> 1m x10 while (a)..(!a) (stack) 0.214000 0.000000 0.214000 >>> ( 0.214000) >>> 1m x10 while (a)..(!a) (heap) 0.249000 0.000000 0.249000 >>> ( 0.250000) >>> >>> after: >>> >>> 1m x10 while (a)..(!a) (stack) 0.203000 0.000000 0.203000 >>> ( 0.203000) >>> 1m x10 while (a)..(!a) (heap) 0.234000 0.000000 0.234000 >>> ( 0.234000) >>> >>> 64-bit: >>> >>> before: >>> >>> 1m x10 while (a)..(!a) (stack) 0.248000 0.000000 0.248000 >>> ( 0.248000) >>> 1m x10 while (a)..(!a) (heap) 0.257000 0.000000 0.257000 >>> ( 0.257000) >>> >>> after: >>> >>> 1m x10 while (a)..(!a) (stack) 0.226000 0.000000 0.226000 >>> ( 0.226000) >>> 1m x10 while (a)..(!a) (heap) 0.244000 0.000000 0.244000 >>> ( 0.244000) >>> >>> We have to fix that but I'm not sure yet what's the best approach. Sorry >>> I don't have better news for now. >>> >>> -- Chris >>> >>> > >>> > Patch for the change (apply to JRuby master) is here: >>> > https://gist.github.com/955976b52b0c4e3f611e >>> > >>> > - Charlie >>> > _______________________________________________ >>> > mlvm-dev mailing list >>> > mlvm-dev@openjdk.java.net >>> > http://mail.openjdk.java.net/mailman/listinfo/mlvm-dev >>> >>> _______________________________________________ >>> mlvm-dev mailing list >>> mlvm-dev@openjdk.java.net >>> http://mail.openjdk.java.net/mailman/listinfo/mlvm-dev >>> >> _______________________________________________ >> mlvm-dev mailing list >> mlvm-dev@openjdk.java.net >> http://mail.openjdk.java.net/mailman/listinfo/mlvm-dev >> >> >> >> _______________________________________________ >> mlvm-dev mailing list >> mlvm-dev@openjdk.java.net >> http://mail.openjdk.java.net/mailman/listinfo/mlvm-dev >> >>
_______________________________________________ mlvm-dev mailing list mlvm-dev@openjdk.java.net http://mail.openjdk.java.net/mailman/listinfo/mlvm-dev