Well I will definitily test this. But I don't know how to start this 
command with the shell:

http://i.imgur.com/uiYbq0O.png

Any tips for windows user? Nothing happening.

Am Donnerstag, 28. August 2014 11:39:59 UTC+2 schrieb Michael Hunger:
>
> Hi Curtis,
>
> if you do this:
>
> USING PERIODIC COMMIT 10000
> LOAD CSV WITH HEADERS FROM "file:C:/test.txt" AS csvLine
> CREATE (:Person { person_id: toInt(csvLine.person_id), name: csvLine.name 
> });
>
> It should do between 10k-30k nodes per second.
> It will be slower if you have a unique constraint in place.
>
> Please run it in the Neo4j-Shell, it is much easier to handle there.
>
> As you are running on windows, with 4GB in total, please also make sure 
> that your mmio config in neo4j.properties is not too large.
>
> Perhaps 100MB for nodes and 500MB for rels and another 250MB for 
> properties.
>
> On windows the memory mapping memory is taken from the heap.
>
> You should see some info about that in in your graph.db/messages.log files 
> after the startup, feel free to share that file with us then we can help 
> you to set up your config.
>
> Michael
>
> Am 28.08.2014 um 11:11 schrieb 'Curtis Mosters' via Neo4j <
> [email protected] <javascript:>>:
>
> Yesterday evening I was running it 52 minutes. And then I got an "Unknown 
> Error".
>
> So now I tested it now with the "USING PERIODIC COMMIT 10000" and now its 
> about 60 minutes. Then the same error.
>
> I now looked into the graph.db folder and its overall 1,75 GB big. The 
> propertystore file ist 370 mb e.g.
>
> So what else can I do do get it running in the browser? Or could I run 
> this task in the Neo4jShell?
>
> Am Donnerstag, 28. August 2014 00:16:07 UTC+2 schrieb Chris Vest:
>>
>> All transaction state is currently kept in memory on the java heap, and 
>> 20+ mio. changes is too much to fit in a 4 GB heap.
>> When you have too much stuff on the heap, it will manifest with those “GC 
>> overhead limit exceeded” and the database will run slow, though there are 
>> other things that can produce similar symptoms.
>>
>> Try putting USING PERIODIC COMMIT 10000 in front of your LOAD CSV query. 
>> This will periodically commit the transaction, thus limiting the 
>> transaction state kept in memory. Unfortunately it will also break the 
>> atomicity of the transaction.
>>
>> --
>> Chris Vest
>> System Engineer, Neo Technology
>> [ skype: mr.chrisvest, twitter: chvest ]
>>
>>  
>> On 27 Aug 2014, at 22:31, 'Curtis Mosters' via Neo4j <
>> [email protected]> wrote:
>>
>> Let's say I have:
>>
>> LOAD CSV WITH HEADERS FROM "file:C:/test.txt" AS csvLine
>> CREATE (p:Person { person_id: toInt(csvLine.person_id), name: csvLine.name 
>> })
>>
>> I run this query in the browser. I know that it's not the fastest way and 
>> I should think about using the batch importer. But I really like that way 
>> somehow and want to speed it up.
>>
>> So when I ran this the first time, after like 2 or 3 minutes I got an 
>> erro saying "GC overhead limit exceeded". So It set
>>
>> wrapper.java.initmemory=4096
>> wrapper.java.maxmemory=4096
>>
>> Now the error does not come up. But it's still slow and I can't see how 
>> much time is still needed. So if you have tips on doing this, I would be 
>> very thankful. =)
>>
>> PS: the file is 2 gb big and has like 20 mio entries
>>
>> -- 
>> You received this message because you are subscribed to the Google Groups 
>> "Neo4j" group.
>> To unsubscribe from this group and stop receiving emails from it, send an 
>> email to [email protected].
>> For more options, visit https://groups.google.com/d/optout.
>>
>>
>>
> -- 
> You received this message because you are subscribed to the Google Groups 
> "Neo4j" group.
> To unsubscribe from this group and stop receiving emails from it, send an 
> email to [email protected] <javascript:>.
> For more options, visit https://groups.google.com/d/optout.
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Neo4j" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to