I think you should use the batch importer for this size of a graph. You 
will be done in minutes, not hours.

https://github.com/jexp/batch-import/tree/20

Rik

On Tuesday, August 12, 2014 5:13:39 PM UTC+1, gg4u wrote:
>
> Hello,
>
> here i am trying to upload a massive network:
> 4M nodes, 100M correlations.
>
> having problems of memory and perfomance, I'd like to know if I am doing 
> it OK:
>
> 1.
> Before loading the correlations, I wanted to load the nodes.
>
> 2. Set up neo4-wrapper and neo4j.properties as written in 
> http://www.neo4j.org/graphgist?d788e117129c3730a042
>
> with JVM heap set at 4096Mb
>
> with this setting, bulk on 4M nodes failed.
>
> 3. Raised memory min-heap and max-heap to 6144Mb
> Run a test with 100K nodes.
>
> I got:
> Nodes created: 98991
> Properties set: 197982
> Labels added: 98991
> 3438685 ms
>
> Almost an hour for uploading 100K nodes with two properties?
> I thought it should be much faster.
>
> Am I doing smtg wrong?
> this is the importer code I used:
>
> CREATE CONSTRAINT ON (n:MYNODES) ASSERT n.id IS UNIQUE;
> CREATE INDEX ON : n:MYNODES(name);
>
> USING PERIODIC COMMIT 1000
> LOAD CSV WITH HEADERS FROM 'file:///blablabla.csv' AS line 
>  FIELDTERMINATOR '\t' 
> WITH line, toInt(line.topicId) as id, line.name as name* LIMIT 100000*
> MERGE (n:MYNODES { id: id, name: name });
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Neo4j" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to