On Thursday, 18 February 2016 at 13:49:45 UTC, Vladimir Panteleev wrote:
On Thursday, 18 February 2016 at 13:00:12 UTC, Witek wrote:
So, the question is, why is D / DMD allocator so slow under heavy multithreading? The working set is pretty small (few megabytes at most), so I do not think this is an issue with GC scanning itself. Can I plug-in tcmalloc / jemalloc, to be used as the underlying allocator, instead of using glibc? Or is D runtime using mmap/srbk/etc directly?

Thanks.

Currently, all memory allocations use a global GC lock[1]. As such, presently high-parallelism programs need to avoid allocating memory via the GC.

You can avoid this problem by using a different allocation / memory management strategy. You may want to have a look at std.experimental.allocator.

[1]: https://github.com/D-Programming-Language/druntime/blob/30f8c1af39eb17d8ebec1f5fd401eb5cfd6b36da/src/gc/gc.d#L348-L370

Yeah, I was just using stuff like:

int[] newPartialSolution = partialSolution ~ row;
int[] newAvailableSolution = vailableSolution[0 .. $-1].dup;
// Move last element if needed in newAvailableSolution.

It was pretty hard to find out, because it was hidden behind "~". Yes, -vgc helped here, but still, I was not expecting so terrible performance.

I will try using std.experimental.allocator, but this doesn't play well with "~", and I would need to manually do expandArray, and array operations, which is a pain. It would be nice to encode allocator used in the type, potentially by wrapping array into custom struct/class.

"As of this time, std.experimental.allocator is not integrated with D's built-in operators that allocate memory, such as new, array literals, or array concatenation operators."

Reply via email to