On Wed, Jul 2, 2014 at 11:43 AM, kilon alios <[email protected]> wrote:
> 0.00000000549181 seconds per message is just insanely low number . This is
> not even in the realm of nanoseconds. Is this number real ? Because I am
> very skeptical that one can get this kind of performance from a dynamic
> language even with a JIT VM.
>
Here's a measurement on my 2.2GHz Mac Book Pro Core i7:
| r | {[r := 34 benchFib] timeToRun. r} #(140 18454929)
benchFib is nfib that adds one for each call, so the answer is the number
of calls necessary to compute the answer.
benchFib ^self < 2 ifTrue: [1] ifFalse: [(self-1) benchFib + (self-2)
benchFib + 1]
So that's 18 million calls in 140 milliseconds, or about 132 calls per
microsecond. That's indeed in the realm of nanoseconds, 13.2 nsecs per
call.
>
> On Wed, Jul 2, 2014 at 9:08 PM, Camille Teruel <[email protected]>
> wrote:
>
>>
>> A spur image :) awesome !!
>> Thanks Esteban & Guille!
>> I'll start adapting the new class builder to Spur next wednesday when I
>> come back from holidays.
>>
>> On 2 juil. 2014, at 16:13, Esteban Lorenzano <[email protected]> wrote:
>>
>> > Hi,
>> >
>> > I’ve been working on prepare Pharo to run with the new Spur VM… and
>> finally this week Guille and I sit together and make a huge advance :)
>> > here a screenshot of pharo 4 running with spur:
>> > <Screen Shot 2014-07-02 at 15.58.17.png>
>> > for more information, same image before migration, with a regular cogvm
>> gives this numbers: '1297023432 bytecodes/sec; 161029354 sends/sec’, so
>> that means that the tiny benchmarks run at 166% the speed of the old vm…
>> > Of course this is just one benchmark, but I’m very impressed :)
>> >
>> > Now, it is still not usable but I’m confident I can put a jenkins job
>> to work very soon :)
>> >
>> > Esteban
>> >
>> > ps: before you start asking: Spur WILL NOT be available for Pharo3, it
>> will be part of Pharo4.
>> >
>> >
>> >
>>
>>
>>
>
--
best,
Eliot