Hi Uwe, the problem is that those ID's are internal to Neo4j where they represent disk record-ids. if you provide high values there, Neo4j will create _a lot_ of empty records until it reaches your ids.
So either you create your node-id's starting from 0 and you store your id as normal node property. Or you don't provide node-id's at all and only lookup nodes via their "business-id-value" i:id id:long l:label 0 315041100 Person 1 201215100 Person 2 315041200 Person start:id end:id type relart 0 1 HAS_RELATION 30006 2 0 HAS_RELATION 30006 or you have to configure and use an index: i:id id:long:people l:label 0 315041100 Person 1 201215100 Person 2 315041200 Person id:long:people id:long:people type relart 0 1 HAS_RELATION 30006 2 0 HAS_RELATION 30006 HTH Michael Alternatively you can also just write a small java or groovy program to import your data if handling those ids with the batch-importer is too tricky. See: http://jexp.de/blog/2014/10/flexible-neo4j-batch-import-with-groovy/ On Tue, Oct 28, 2014 at 8:39 AM, Uwe Ludwig <[email protected]> wrote: > Hi all, > > i want to import csv-Files with about 40 million lines into neo4j. For > this i try to use the "batchimporter" from > https://github.com/jexp/batch-import. > Maybe it's a problem that i provide own IDs. This is the example > > > nodes.csv: > i:id l:label > 315041100 Person > 201215100 Person > 315041200 Person > > rels.csv : > start end type relart > 315041100 201215100 HAS_RELATION 30006 > 315041200 315041100 HAS_RELATION 30006 > > the content of batch.properties: > use_memory_mapped_buffers=true > neostore.nodestore.db.mapped_memory=1000M > neostore.relationshipstore.db.mapped_memory=5000M > neostore.propertystore.db.mapped_memory=4G > neostore.propertystore.db.strings.mapped_memory=2000M > neostore.propertystore.db.arrays.mapped_memory=1000M > neostore.propertystore.db.index.keys.mapped_memory=1500M > neostore.propertystore.db.index.mapped_memory=1500M > batch_import.node_index.node_auto_index=exact > > > "./import.sh graph.db nodes.csv rels.csv" > > will be processed without errors, but it takes about 60 seconds! When i > use smaller IDs - for example 3150411 instead of 315041100 - it takes just > 1 second! > Actually i would take even bigger IDs with 10 digits. I don't know what > i'm doing wrong. Can anyone see an error? Do i have to assign an explicit > type (long?) for the IDs? > How can i do this? > > > - JDK 1.7 > - batchimporter 2.1.3 (with neo4j 2.1.3) > - OS: ubuntu 14.04 > - Hardware: 8-Core-Intel-CPU, 16GB RAM > > Best regards and thanks in advance > > Uwe > > > > -- > You received this message because you are subscribed to the Google Groups > "Neo4j" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "Neo4j" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. For more options, visit https://groups.google.com/d/optout.
