2014-10-24 15:50 GMT+02:00 Clément Bera <[email protected]>: > The current x2 speed boost is due only to spur, not to sista. Sista will > provide additional performance, but we have still things to do before > production. > > The performance gain reported is due to (from most important to less > important): > - the new GC has less overhead. 30% of the execution time used to be spent > in the GC. > - the new object format speeds up some VM internal caches (especially > inline caches for message sends due to an indirection for object classes > with a class table). > - the new object format allows some C code to be converted into machine > code routines, including block creation, context creation, primitive > #at:put:, which is faster because switching from jitted code to C then back > to jitted code generate a little overhead. > - characters are now immediate objects, which speeds up String accessing. > - the new object format has a larger hash which speeds up big hashed > collections such as big sets and dictionaries. > - become is faster. > >
All this is really cool :) And if I remember well, there is 64 bitness coming as well. Will Spur also cover ARM ? Thierry > > 2014-10-24 15:20 GMT+02:00 kilon alios <[email protected]>: > >> thanks max, i completely forgotten about esug videos, looks like i found >> what to watch during the weekend :D >> >> On Fri, Oct 24, 2014 at 4:12 PM, Max Leske <[email protected]> wrote: >> >>> >>> On 24.10.2014, at 15:06, kilon alios <[email protected]> wrote: >>> >>> very nice >>> >>> so any more information to this, how exactly this optimization works and >>> which kind of data will benefit from this ? >>> >>> >>> Clément's byte code set talk at ESUG: >>> http://www.youtube.com/watch?v=e9J362QHwSA&index=64&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X >>> Clément's Sista talk at ESUG (2 parts): >>> >>> http://www.youtube.com/watch?v=X4E_FoLysJg&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=76 >>> >>> http://www.youtube.com/watch?v=gZOk3qojoVE&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=75 >>> >>> Eliot's Spur talk at ESUG (3 parts): >>> >>> http://www.youtube.com/watch?v=k0nBNS1aHZ4&index=49&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X >>> >>> http://www.youtube.com/watch?v=sn3irBZE7g4&index=48&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X >>> >>> http://www.youtube.com/watch?v=1Vg0iFeg_pA&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=47 >>> >>> >>> On Fri, Oct 24, 2014 at 3:47 PM, Sebastian Sastre < >>> [email protected]> wrote: >>> >>>> remarkable!!! >>>> >>>> congratulations for the impressive results >>>> >>>> thanks for sharing! >>>> >>>> sebastian >>>> >>>> o/ >>>> >>>> > On 23/10/2014, at 17:40, Max Leske <[email protected]> wrote: >>>> > >>>> > For those of you who missed this on IRC: >>>> > >>>> > henriksp: estebanlm: Care to run a small bench Cog vs Spur for me? >>>> > [3:32pm] henriksp: int := ZnUTF8Encoder new. >>>> > [3:32pm] henriksp: [int decodeBytes:#[67 97 115 104 44 32 108 105 107 >>>> 101 32 226 130 172 44 32 105 115 32 107 105 110 103 0]] bench. >>>> > [3:32pm] henriksp: had a 16x speedup with assembly implementation vs >>>> Cog, if it's 8x vs Spur, that's just really impressive >>>> > [3:44pm] Craig left the chat room. (Quit: Leaving.) >>>> > [3:53pm] Craig joined the chat room. >>>> > [4:08pm] VitamineD joined the chat room. >>>> > [4:20pm] estebanlm: checking >>>> > [4:21pm] estebanlm: Cog: 167,000 per second. >>>> > [4:22pm] estebanlm: Cog[Spur]: 289,000 per second. >>>> > [4:23pm] estebanlm: henriksp: ping >>>> > [4:33pm] tinchodias left the chat room. (Ping timeout: 245 seconds) >>>> > [4:33pm] tinchodias joined the chat room. >>>> > [4:34pm] henriksp: 70% more work done, nice! >>>> > [5:09pm] >>>> > >>>> > >>>> > Yay! :) >>>> >>>> >>> >>> >> >
