This problem is surprising as I was able to run PR and CC on a larger graph (2bil edges) but with this synthetic graph (1bil edges groups of 10) I ran out of memory; regarding configuration (memory and parallelism, other internals) I was using the same. There is some limitation somewhere I will try to understand more what is happening.
Best, Ovidiu > On 14 Mar 2016, at 18:06, Martin Junghanns <m.jungha...@mailbox.org> wrote: > > Hi, > > I understand the confusion. So far, I did not run into the problem, but I > think this needs to be adressed as all our graph processing abstractions are > implemented on top of the delta iteration. > > According to the previous mailing list discussion, the problem is with the > solution set and its missing ability to spill. > > If this is the still the case, we should open an issue for that. Any further > opinions on that? > > Cheers, > Martin > > > On 14.03.2016 17:55, Ovidiu-Cristian MARCU wrote: >> Thank you for this alternative. >> I don’t understand how the workaround will fix this on systems with limited >> memory and maybe larger graph. >> >> Running Connected Components on the same graph gives the same problem. >> >> IterationHead(Unnamed Delta Iteration)(82/88) switched to FAILED >> java.lang.RuntimeException: Memory ran out. Compaction failed. >> numPartitions: 32 minPartition: 31 maxPartition: 32 number of overflow >> segments: 417 bucketSize: 827 Overall memory: 149159936 Partition memory: >> 65601536 Message: Index: 32, Size: 31 >> at >> org.apache.flink.runtime.operators.hash.CompactingHashTable.insertRecordIntoPartition(CompactingHashTable.java:469) >> at >> org.apache.flink.runtime.operators.hash.CompactingHashTable.insertOrReplaceRecord(CompactingHashTable.java:414) >> at >> org.apache.flink.runtime.operators.hash.CompactingHashTable.buildTableWithUniqueKey(CompactingHashTable.java:325) >> at >> org.apache.flink.runtime.iterative.task.IterationHeadTask.readInitialSolutionSet(IterationHeadTask.java:212) >> at >> org.apache.flink.runtime.iterative.task.IterationHeadTask.run(IterationHeadTask.java:273) >> at >> org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:354) >> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584) >> at java.lang.Thread.run(Thread.java:745) >> >> Best, >> Ovidiu >> >>> On 14 Mar 2016, at 17:36, Martin Junghanns <m.jungha...@mailbox.org> wrote: >>> >>> Hi >>> >>> I think this is the same issue we had before on the list [1]. Stephan >>> recommended the following workaround: >>> >>>> A possible workaround is to use the option "setSolutionSetUnmanaged(true)" >>>> on the iteration. That will eliminate the fragmentation issue, at least. >>> >>> Unfortunately, you cannot set this when using graph.run(new PageRank(...)) >>> >>> I created a Gist which shows you how to set this using PageRank >>> >>> https://gist.github.com/s1ck/801a8ef97ce374b358df >>> >>> Please let us know if it worked out for you. >>> >>> Cheers, >>> Martin >>> >>> [1] >>> http://mail-archives.apache.org/mod_mbox/flink-user/201508.mbox/%3CCAELUF_ByPAB%2BPXWLemPzRH%3D-awATeSz4sGz4v9TmnvFku3%3Dx3A%40mail.gmail.com%3E >>> >>> On 14.03.2016 16:55, Ovidiu-Cristian MARCU wrote: >>>> Hi, >>>> >>>> While running PageRank on a synthetic graph I run into this problem: >>>> Any advice on how should I proceed to overcome this memory issue? >>>> >>>> IterationHead(Vertex-centric iteration >>>> (org.apache.flink.graph.library.PageRank$VertexRankUpdater@7712cae0 | >>>> org.apache.flink.graph.library.PageRank$RankMesseng$ >>>> java.lang.RuntimeException: Memory ran out. Compaction failed. >>>> numPartitions: 32 minPartition: 24 maxPartition: 25 number of overflow >>>> segments: 328 bucketSize: 638 Overall memory: 115539968 Partition memory: >>>> 50659328 Message: Index: 25, Size: 24 >>>> at >>>> org.apache.flink.runtime.operators.hash.CompactingHashTable.insertRecordIntoPartition(CompactingHashTable.java:469) >>>> at >>>> org.apache.flink.runtime.operators.hash.CompactingHashTable.insertOrReplaceRecord(CompactingHashTable.java:414) >>>> at >>>> org.apache.flink.runtime.operators.hash.CompactingHashTable.buildTableWithUniqueKey(CompactingHashTable.java:325) >>>> at >>>> org.apache.flink.runtime.iterative.task.IterationHeadTask.readInitialSolutionSet(IterationHeadTask.java:212) >>>> at >>>> org.apache.flink.runtime.iterative.task.IterationHeadTask.run(IterationHeadTask.java:273) >>>> at >>>> org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:354) >>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584) >>>> at java.lang.Thread.run(Thread.java:745) >>>> >>>> Thanks! >>>> >>>> Best, >>>> Ovidiu >>>> >> >>