Hmmm, I didn't notice that I was running with -XX:+UseG1GC, so perhaps our
test suite is a pathological case for the default collector?

[INFO] Total time: 12:45 min
GC Time: 52.593s
Class Loader Time: 1m 26.007s
Compile Time: 10m 10.216s

I'll try without -XX:+UseG1GC later.

Cheers
Dan


On Thu, Feb 15, 2018 at 1:39 PM, Dan Berindei <dan.berin...@gmail.com>
wrote:

> And here I was thinking that by adding -XX:+HeapDumpOnOutOfMemoryError
> anyone would be able to look into OOMEs and I wouldn't have to reproduce
> the failures myself :)
>
> Dan
>
>
> On Thu, Feb 15, 2018 at 1:32 PM, William Burns <mudokon...@gmail.com>
> wrote:
>
>> So I must admit I had noticed a while back that I was having some issues
>> with running the core test suite. Unfortunately at the time CI and everyone
>> else seemed to not have any issues. I just ignored it because at the time I
>> didn't need to run core tests. But now that Sanne pointed this out, by
>> increasing the heap variable in the pom.xml, I was for the first time able
>> to run the test suite completely. It would normally hang for an extremely
>> long time near the 9k-10K test completed point and never finish for me (at
>> least I didn't wait long enough).
>>
>> So it definitely seems there is something leaking in the test suite
>> causing the GC to use a ton of CPU time.
>>
>>  - Will
>>
>> On Thu, Feb 15, 2018 at 5:40 AM Sanne Grinovero <sa...@infinispan.org>
>> wrote:
>>
>>> Thanks Dan.
>>>
>>> Do you happen to have observed the memory trend during a build?
>>>
>>> After a couple more attempts it passed the build once, so that shows
>>> it's possible to pass.. but even though it's a small sample so far
>>> that's 1 pass vs 3 OOMs on my machine.
>>>
>>> Even the one time it successfully completed the tests I see it wasted
>>> ~80% of total build time doing GC runs.. it was likely very close to
>>> fall over, and definitely not an efficient setting for regular builds.
>>> Observing trends on my machine I'd guess a reasonable value to be
>>> around 5GB to keep builds fast, or a minimum of 1.3 GB to be able to
>>> complete successfully without often failing.
>>>
>>> The memory issues are worse towards the end of the testsuite, and
>>> steadily growing.
>>>
>>> I won't be able to investigate further as I need to urgently work on
>>> modules, but I noticed there are quite some MBeans according to
>>> JConsole. I guess it would be good to check if we're not leaking the
>>> MBean registration, and therefore leaking (stopped?) CacheManagers
>>> from there?
>>>
>>> Even near the beginning of the tests, when forcing a full GC I see
>>> about 400MB being "not free". That's quite a lot for some simple
>>> tests, no?
>>>
>>> Thanks,
>>> Sanne
>>>
>>>
>>> On 15 February 2018 at 06:51, Dan Berindei <dan.berin...@gmail.com>
>>> wrote:
>>> > forkJvmArgs used to be "-Xmx2G" before ISPN-8478. I reduced the heap
>>> to 1G
>>> > because we were trying to run the build on agent VMs with only 4GB of
>>> RAM,
>>> > and the 2GB heap was making the build run out of native memory.
>>> >
>>> > I've yet to see an OOME in the core tests, locally or in CI. But I also
>>> > included -XX:+HeapDumpOnOutOfMemoryError in forkJvmArgs, so assuming
>>> there's
>>> > a new leak it should be easy to track down in the heap dump.
>>> >
>>> > Cheers
>>> > Dan
>>> >
>>> >
>>> > On Wed, Feb 14, 2018 at 11:46 PM, Sanne Grinovero <
>>> sa...@infinispan.org>
>>> > wrote:
>>> >>
>>> >> Hey all,
>>> >>
>>> >> I'm having OOMs running the tests of infinispan-core.
>>> >>
>>> >> Initially I thought it was related to limits and security as that's
>>> >> the usual suspect, but no it's really just not enough memory :)
>>> >>
>>> >> Found that the root pom.xml sets a <forkJvmArgs> property to Xmx1G for
>>> >> surefire; I've been observing the growth of heap usage in JConsole and
>>> >> it's clearly not enough.
>>> >>
>>> >> What surprises me is that - as an occasional tester - I shouldn't be
>>> >> the one to notice such a new requirement first. A leak which only
>>> >> manifests in certain conditions?
>>> >>
>>> >> What do others observe?
>>> >>
>>> >> FWIW, I'm running it with 8G heap now and it's working much better;
>>> >> still a couple of failures but at least they're not OOM related.
>>> >>
>>> >> Thanks,
>>> >> Sanne
>>> >> _______________________________________________
>>> >> infinispan-dev mailing list
>>> >> infinispan-dev@lists.jboss.org
>>> >> https://lists.jboss.org/mailman/listinfo/infinispan-dev
>>> >
>>> >
>>> >
>>> > _______________________________________________
>>> > infinispan-dev mailing list
>>> > infinispan-dev@lists.jboss.org
>>> > https://lists.jboss.org/mailman/listinfo/infinispan-dev
>>> _______________________________________________
>>> infinispan-dev mailing list
>>> infinispan-dev@lists.jboss.org
>>> https://lists.jboss.org/mailman/listinfo/infinispan-dev
>>>
>>
>> _______________________________________________
>> infinispan-dev mailing list
>> infinispan-dev@lists.jboss.org
>> https://lists.jboss.org/mailman/listinfo/infinispan-dev
>>
>
>
_______________________________________________
infinispan-dev mailing list
infinispan-dev@lists.jboss.org
https://lists.jboss.org/mailman/listinfo/infinispan-dev

Reply via email to