hmm, I'm starting to run out of ideas...
What's your source ID parameter? I ran mine with 0.
About the result, you call both createVertexCentricIteration() and
runVertexCentricIteration() on the initialized graph, right?

On 18 March 2015 at 22:33, Mihail Vieru <vi...@informatik.hu-berlin.de>
wrote:

>  Hi Vasia,
>
> yes, I am using the latest master. I just did a pull again and the problem
> persists. Perhaps Robert could confirm as well.
>
> I've set the solution set to unmanaged in SSSPUnweighted as Stephan
> proposed and the job finishes. So I am able to proceed using this
> workaround.
>
> An odd thing occurs now though. The distances aren't computed correctly
> for the SNAP graph and remain the one set in InitVerticesMapper(). For the
> small graph in SSSPDataUnweighted they are OK. I'm currently investigating
> this behavior.
>
> Cheers,
> Mihail
>
>
> On 18.03.2015 20:55, Vasiliki Kalavri wrote:
>
>  Hi Mihail,
>
>  I used your code to generate the vertex file, then gave this and the
> edge list as input to your SSSP implementation and still couldn't reproduce
> the exception. I'm using the same local setup as I describe above.
> I'm not aware of any recent changes that might be relevant, but, just in
> case, are you using the latest master?
>
>  Cheers,
> V.
>
> On 18 March 2015 at 19:21, Mihail Vieru <vi...@informatik.hu-berlin.de>
> wrote:
>
>>  Hi Vasia,
>>
>> I have used a simple job (attached) to generate a file which looks like
>> this:
>>
>> 0 0
>> 1 1
>> 2 2
>> ...
>> 456629 456629
>> 456630 456630
>>
>> I need the vertices to be generated from a file for my future work.
>>
>> Cheers,
>> Mihail
>>
>>
>>
>> On 18.03.2015 17:04, Vasiliki Kalavri wrote:
>>
>>  Hi Mihail, Robert,
>>
>>  I've tried reproducing this, but I couldn't.
>> I'm using the same twitter input graph from SNAP that you link to and
>> also Scala IDE.
>> The job finishes without a problem (both the SSSP example from Gelly and
>> the unweighted version).
>>
>>  The only thing I changed to run your version was creating the graph
>> from the edge set only, i.e. like this:
>>
>>  Graph<Long, Long, NullValue> graph = Graph.fromDataSet(edges,
>>  new MapFunction<Long, Long>() {
>>  public Long map(Long value) {
>>  return Long.MAX_VALUE;
>>  }
>>  }, env);
>>
>> Since the twitter input is an edge list, how do you generate the vertex
>> dataset in your case?
>>
>>  Thanks,
>> -Vasia.
>>
>> On 18 March 2015 at 16:54, Mihail Vieru <vi...@informatik.hu-berlin.de>
>> wrote:
>>
>>>  Hi,
>>>
>>> great! Thanks!
>>>
>>> I really need this bug fixed because I'm laying the groundwork for my
>>> Diplom thesis and I need to be sure that the Gelly API is reliable and can
>>> handle large datasets as intended.
>>>
>>> Cheers,
>>> Mihail
>>>
>>>
>>> On 18.03.2015 15:40, Robert Waury wrote:
>>>
>>>   Hi,
>>>
>>>  I managed to reproduce the behavior and as far as I can tell it seems
>>> to be a problem with the memory allocation.
>>>
>>>  I have filed a bug report in JIRA to get the attention of somebody who
>>> knows the runtime better than I do.
>>>
>>> https://issues.apache.org/jira/browse/FLINK-1734
>>>
>>>  Cheers,
>>>  Robert
>>>
>>> On Tue, Mar 17, 2015 at 3:52 PM, Mihail Vieru <
>>> vi...@informatik.hu-berlin.de> wrote:
>>>
>>>>  Hi Robert,
>>>>
>>>> thank you for your reply.
>>>>
>>>> I'm starting the job from the Scala IDE. So only one JobManager and one
>>>> TaskManager in the same JVM.
>>>> I've doubled the memory in the eclipse.ini settings but I still get the
>>>> Exception.
>>>>
>>>> -vmargs
>>>> -Xmx2048m
>>>> -Xms100m
>>>> -XX:MaxPermSize=512m
>>>>
>>>> Best,
>>>> Mihail
>>>>
>>>>
>>>> On 17.03.2015 10:11, Robert Waury wrote:
>>>>
>>>>   Hi,
>>>>
>>>>  can you tell me how much memory your job has and how many workers you
>>>> are running?
>>>>
>>>>  From the trace it seems the internal hash table allocated only 7 MB
>>>> for the graph data and therefore runs out of memory pretty quickly.
>>>>
>>>>  Skewed data could also be an issue but with a minimum of 5 pages and
>>>> a maximum of 8 it seems to be distributed fairly even to the different
>>>> partitions.
>>>>
>>>>  Cheers,
>>>>  Robert
>>>>
>>>> On Tue, Mar 17, 2015 at 1:25 AM, Mihail Vieru <
>>>> vi...@informatik.hu-berlin.de> wrote:
>>>>
>>>>> And the correct SSSPUnweighted attached.
>>>>>
>>>>>
>>>>> On 17.03.2015 01:23, Mihail Vieru wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I'm getting the following RuntimeException for an adaptation of the
>>>>>> SingleSourceShortestPaths example using the Gelly API (see attachment).
>>>>>> It's been adapted for unweighted graphs having vertices with Long values.
>>>>>>
>>>>>> As an input graph I'm using the social network graph (~200MB
>>>>>> unpacked) from here:
>>>>>> https://snap.stanford.edu/data/higgs-twitter.html
>>>>>>
>>>>>> For the small SSSPDataUnweighted graph (also attached) it terminates
>>>>>> and computes the distances correctly.
>>>>>>
>>>>>>
>>>>>> 03/16/2015 17:18:23    IterationHead(WorksetIteration (Vertex-centric
>>>>>> iteration
>>>>>> (org.apache.flink.graph.library.SingleSourceShortestPathsUnweighted$VertexDistanceUpdater@dca6fe4
>>>>>> |
>>>>>> org.apache.flink.graph.library.SingleSourceShortestPathsUnweighted$MinDistanceMessenger@6577e8ce)))(2/4)
>>>>>> switched to FAILED
>>>>>> java.lang.RuntimeException: Memory ran out. Compaction failed.
>>>>>> numPartitions: 32 minPartition: 5 maxPartition: 8 number of overflow
>>>>>> segments: 176 bucketSize: 217 Overall memory: 20316160 Partition memory:
>>>>>> 7208960 Message: Index: 8, Size: 7
>>>>>>     at
>>>>>> org.apache.flink.runtime.operators.hash.CompactingHashTable.insert(CompactingHashTable.java:390)
>>>>>>     at
>>>>>> org.apache.flink.runtime.operators.hash.CompactingHashTable.buildTable(CompactingHashTable.java:337)
>>>>>>     at
>>>>>> org.apache.flink.runtime.iterative.task.IterationHeadPactTask.readInitialSolutionSet(IterationHeadPactTask.java:216)
>>>>>>     at
>>>>>> org.apache.flink.runtime.iterative.task.IterationHeadPactTask.run(IterationHeadPactTask.java:278)
>>>>>>     at
>>>>>> org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:362)
>>>>>>     at
>>>>>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:205)
>>>>>>     at java.lang.Thread.run(Thread.java:745)
>>>>>>
>>>>>>
>>>>>> Best,
>>>>>> Mihail
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>
>>
>
>

Reply via email to