The current x2 speed boost is due only to spur, not to sista. Sista will
provide additional performance, but we have still things to do before
production.

The performance gain reported is due to (from most important to less
important):
- the new GC has less overhead. 30% of the execution time used to be spent
in the GC.
- the new object format speeds up some VM internal caches (especially
inline caches for message sends due to an indirection for object classes
with a class table).
- the new object format allows some C code to be converted into machine
code routines, including block creation, context creation, primitive
#at:put:, which is faster because switching from jitted code to C then back
to jitted code generate a little overhead.
- characters are now immediate objects, which speeds up String accessing.
- the new object format has a larger hash which speeds up big hashed
collections such as big sets and dictionaries.
- become is faster.


2014-10-24 15:20 GMT+02:00 kilon alios <[email protected]>:

> thanks max, i completely forgotten about esug videos, looks like i found
> what to watch during the weekend :D
>
> On Fri, Oct 24, 2014 at 4:12 PM, Max Leske <[email protected]> wrote:
>
>>
>> On 24.10.2014, at 15:06, kilon alios <[email protected]> wrote:
>>
>> very nice
>>
>> so any more information to this, how exactly this optimization works and
>> which kind of data will benefit from this ?
>>
>>
>> Clément’s byte code set talk at ESUG:
>> http://www.youtube.com/watch?v=e9J362QHwSA&index=64&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X
>> Clément’s Sista talk at ESUG (2 parts):
>>
>> http://www.youtube.com/watch?v=X4E_FoLysJg&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=76
>>
>> http://www.youtube.com/watch?v=gZOk3qojoVE&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=75
>>
>> Eliot’s Spur talk at ESUG (3 parts):
>>
>> http://www.youtube.com/watch?v=k0nBNS1aHZ4&index=49&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X
>>
>> http://www.youtube.com/watch?v=sn3irBZE7g4&index=48&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X
>>
>> http://www.youtube.com/watch?v=1Vg0iFeg_pA&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=47
>>
>>
>> On Fri, Oct 24, 2014 at 3:47 PM, Sebastian Sastre <
>> [email protected]> wrote:
>>
>>> remarkable!!!
>>>
>>> congratulations for the impressive results
>>>
>>> thanks for sharing!
>>>
>>> sebastian
>>>
>>> o/
>>>
>>> > On 23/10/2014, at 17:40, Max Leske <[email protected]> wrote:
>>> >
>>> > For those of you who missed this on IRC:
>>> >
>>> > henriksp: estebanlm: Care to run a small bench Cog vs Spur for me?
>>> > [3:32pm] henriksp: int := ZnUTF8Encoder new.
>>> > [3:32pm] henriksp: [int decodeBytes:#[67 97 115 104 44 32 108 105 107
>>> 101 32 226 130 172 44 32 105 115 32 107 105 110 103 0]] bench.
>>> > [3:32pm] henriksp: had a 16x speedup with assembly implementation vs
>>> Cog, if it's 8x vs Spur, that's just really impressive
>>> > [3:44pm] Craig left the chat room. (Quit: Leaving.)
>>> > [3:53pm] Craig joined the chat room.
>>> > [4:08pm] VitamineD joined the chat room.
>>> > [4:20pm] estebanlm: checking
>>> > [4:21pm] estebanlm: Cog: 167,000 per second.
>>> > [4:22pm] estebanlm: Cog[Spur]: 289,000 per second.
>>> > [4:23pm] estebanlm: henriksp: ping
>>> > [4:33pm] tinchodias left the chat room. (Ping timeout: 245 seconds)
>>> > [4:33pm] tinchodias joined the chat room.
>>> > [4:34pm] henriksp: 70% more work done, nice!
>>> > [5:09pm]
>>> >
>>> >
>>> > Yay! :)
>>>
>>>
>>
>>
>

Reply via email to