I am currently trying to optimize the in process part of my Actor Model 
framework for Go - GAM.
In order to compare it to Erlang and Akka, I am using the "Skynet 
Benchmark" https://github.com/atemerev/skynet#results-i7-4771-ubuntu-1510

I have ported this test to my lib: 
It works and the performance is roughly the same as the Scala Akka 

I have noticed that if I turn of GOGC, it runs extremely fast. about 3 
times faster.
Is this because my test is generating a lot of garbage or more due to the 
GC process of GO having a big overhead by just running?
That is, would I still see a pef difference even in a system that does not 
allocate anything?

This test is ofcourse completely nonsense from a real world scenario, but 
still, it's an interesting challange trying to push the numbers down.

When profiling, I am not seeing anything specific that pops out at me. 
Anyone here up for giving a helping hand?

You received this message because you are subscribed to the Google Groups 
"golang-nuts" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to golang-nuts+unsubscr...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to