2014-10-24 15:50 GMT+02:00 Clément Bera <[email protected]>:

> The current x2 speed boost is due only to spur, not to sista. Sista will
> provide additional performance, but we have still things to do before
> production.
>
> The performance gain reported is due to (from most important to less
> important):
> - the new GC has less overhead. 30% of the execution time used to be spent
> in the GC.
> - the new object format speeds up some VM internal caches (especially
> inline caches for message sends due to an indirection for object classes
> with a class table).
> - the new object format allows some C code to be converted into machine
> code routines, including block creation, context creation, primitive
> #at:put:, which is faster because switching from jitted code to C then back
> to jitted code generate a little overhead.
> - characters are now immediate objects, which speeds up String accessing.
> - the new object format has a larger hash which speeds up big hashed
> collections such as big sets and dictionaries.
> - become is faster.
>
>

All this is really cool :) And if I remember well, there is 64 bitness
coming as well.

Will Spur also cover ARM ?

Thierry


>
> 2014-10-24 15:20 GMT+02:00 kilon alios <[email protected]>:
>
>> thanks max, i completely forgotten about esug videos, looks like i found
>> what to watch during the weekend :D
>>
>> On Fri, Oct 24, 2014 at 4:12 PM, Max Leske <[email protected]> wrote:
>>
>>>
>>> On 24.10.2014, at 15:06, kilon alios <[email protected]> wrote:
>>>
>>> very nice
>>>
>>> so any more information to this, how exactly this optimization works and
>>> which kind of data will benefit from this ?
>>>
>>>
>>> Clément's byte code set talk at ESUG:
>>> http://www.youtube.com/watch?v=e9J362QHwSA&index=64&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X
>>> Clément's Sista talk at ESUG (2 parts):
>>>
>>> http://www.youtube.com/watch?v=X4E_FoLysJg&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=76
>>>
>>> http://www.youtube.com/watch?v=gZOk3qojoVE&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=75
>>>
>>> Eliot's Spur talk at ESUG (3 parts):
>>>
>>> http://www.youtube.com/watch?v=k0nBNS1aHZ4&index=49&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X
>>>
>>> http://www.youtube.com/watch?v=sn3irBZE7g4&index=48&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X
>>>
>>> http://www.youtube.com/watch?v=1Vg0iFeg_pA&list=PLJ5nSnWzQXi_6yyRLsMMBqG8YlwfhvB0X&index=47
>>>
>>>
>>> On Fri, Oct 24, 2014 at 3:47 PM, Sebastian Sastre <
>>> [email protected]> wrote:
>>>
>>>> remarkable!!!
>>>>
>>>> congratulations for the impressive results
>>>>
>>>> thanks for sharing!
>>>>
>>>> sebastian
>>>>
>>>> o/
>>>>
>>>> > On 23/10/2014, at 17:40, Max Leske <[email protected]> wrote:
>>>> >
>>>> > For those of you who missed this on IRC:
>>>> >
>>>> > henriksp: estebanlm: Care to run a small bench Cog vs Spur for me?
>>>> > [3:32pm] henriksp: int := ZnUTF8Encoder new.
>>>> > [3:32pm] henriksp: [int decodeBytes:#[67 97 115 104 44 32 108 105 107
>>>> 101 32 226 130 172 44 32 105 115 32 107 105 110 103 0]] bench.
>>>> > [3:32pm] henriksp: had a 16x speedup with assembly implementation vs
>>>> Cog, if it's 8x vs Spur, that's just really impressive
>>>> > [3:44pm] Craig left the chat room. (Quit: Leaving.)
>>>> > [3:53pm] Craig joined the chat room.
>>>> > [4:08pm] VitamineD joined the chat room.
>>>> > [4:20pm] estebanlm: checking
>>>> > [4:21pm] estebanlm: Cog: 167,000 per second.
>>>> > [4:22pm] estebanlm: Cog[Spur]: 289,000 per second.
>>>> > [4:23pm] estebanlm: henriksp: ping
>>>> > [4:33pm] tinchodias left the chat room. (Ping timeout: 245 seconds)
>>>> > [4:33pm] tinchodias joined the chat room.
>>>> > [4:34pm] henriksp: 70% more work done, nice!
>>>> > [5:09pm]
>>>> >
>>>> >
>>>> > Yay! :)
>>>>
>>>>
>>>
>>>
>>
>

Reply via email to