Your constraints are wrong you mixed up labels and identifiers

Please also check the index properties

And I had better success doing a multi-pass for each set of elements to connect

Sent from mobile device

Am 26.06.2014 um 11:58 schrieb Pavan Kumar <[email protected]>:

> My JVM properties include
> 
> # Enter one VM parameter per line, note that some parameters can only be set 
> once.
> # For example, to adjust the maximum memory usage to 512 MB, uncomment the 
> following line
> -Xmx6144m
> Xmx4G -Xms4G -Xmn1G
> 
>  but still i am getting GC overhead limit exceeded error. (I have tried from 
> 512m to 6GB)
> My neo4j.propertie file contains
> neostore.nodestore.db.mapped_memory=100M
> neostore.relationshipstore.db.mapped_memory=2G
> neostore.propertystore.db.mapped_memory=200M
> neostore.propertystore.db.strings.mapped_memory=200M
> neostore.propertystore.db.arrays.mapped_memory=0M
> 
> Any more suggestions to get rid of the error
> 
> 
> 
> 
> 
> 
> On Wed, Jun 25, 2014 at 6:34 PM, Pavan Kumar <[email protected]> wrote:
>> My property file is 
>> 
>> # Default values for the low-level graph engine
>> 
>> neostore.nodestore.db.mapped_memory=100M
>> neostore.relationshipstore.db.mapped_memory=2G
>> neostore.propertystore.db.mapped_memory=200M
>> neostore.propertystore.db.strings.mapped_memory=200M
>> neostore.propertystore.db.arrays.mapped_memory=0M
>> 
>> # Enable this to be able to upgrade a store from an older version
>> allow_store_upgrade=true
>> 
>> # Enable this to specify a parser other than the default one.
>> #cypher_parser_version=2.0
>> 
>> # Keep logical logs, helps debugging but uses more disk space, enabled for
>> # legacy reasons To limit space needed to store historical logs use values 
>> such
>> # as: "7 days" or "100M size" instead of "true"
>> keep_logical_logs=true
>> 
>> # Autoindexing
>> 
>> # Enable auto-indexing for nodes, default is false
>> #node_auto_indexing=true
>> 
>> # The node property keys to be auto-indexed, if enabled
>> #node_keys_indexable=name,age
>> 
>> # Enable auto-indexing for relationships, default is false
>> #relationship_auto_indexing=true
>> 
>> # The relationship property keys to be auto-indexed, if enabled
>> #relationship_keys_indexable=name,age
>> cache_type=strong
>> 
>> 
>> 
>> My jvm options are 
>> # Enter one VM parameter per line, note that some parameters can only be set 
>> once.
>> # For example, to adjust the maximum memory usage to 512 MB, uncomment the 
>> following line
>> -Xmx512m
>> 
>> But still i am getting GC overhead limit exceeded error
>> Kindly somebody suggest me
>> 
>> On Wed, Jun 25, 2014 at 5:01 PM, Michael Hunger 
>> <[email protected]> wrote:
>>> use 1000 here
>>> USING PERIODIC COMMIT 1000
>>> 
>>> Increase the memory settings to 6G
>>> 
>>> As you run on windows:
>>> 
>>>> neostore.nodestore.db.mapped_memory=100M
>>>> neostore.relationshipstore.db.mapped_memory=2G
>>>> neostore.propertystore.db.mapped_memory=200M
>>>> neostore.propertystore.db.strings.mapped_memory=200M
>>>> neostore.propertystore.db.arrays.mapped_memory=0M
>>> 
>>> also can you share your complete CSV with me privately?
>>> 
>>> Do you have any nodes in your dataset that have many (100k-1M) of 
>>> relationships?
>>> 
>>> On Wed, Jun 25, 2014 at 1:23 PM, Pavan Kumar <[email protected]> 
>>> wrote:
>>>> Hi , 
>>>> My qiuery is as follows
>>>> create constraint on (ChemicalName:chemicalname) assert 
>>>> ChemicalName.chemicalname is unique;
>>>> create constraint on (ChemicalID:chemicalid) assert ChemicalID.chemicalid 
>>>> is unique;
>>>> create constraint on (Genesymbol:genesymbol) assert Genesymbol.genesymbol 
>>>> is unique;
>>>> create constraint on (Geneid:geneid) assert Geneid.geneid is unique;
>>>> create constraint on (Geneform:geneform) assert Geneform.geneform is 
>>>> unique;
>>>> create constraint on (Interaction:interaction) assert 
>>>> Interaction.interaction is unique;
>>>> create constraint on (Interactionactions:interactionactions) assert 
>>>> Interactionactions.interactionactions is unique;
>>>> create constraint on (PubmedID:pubmed) assert PubmedID.pubmed is unique;
>>>> create index on :ChemicalName(chemicalname);
>>>> create index on :ChemicalID(chemicalid);
>>>> create index on :Genesymbol(genesymbol);
>>>> create index on :Geneid(geneid);
>>>> create index on :Geneform(geneform);
>>>> create index on :Interaction(interaction);
>>>> create index on :Interactionactions(interactionactions);
>>>> create index on :PubmedID(pubmed);
>>>> 
>>>> USING PERIODIC COMMIT 10000
>>>> LOAD CSV WITH HEADERS FROM
>>>> "file:D:/Graph_Database/CTD/CTD_chem_gene_ixns.csv"
>>>> AS chemgeneinteractions
>>>> match (geneid:Geneid{geneid: chemgeneinteractions.GeneID})
>>>> match (genesymbol:Genesymbol{genesymbol: chemgeneinteractions.GeneSymbol})
>>>> merge (chemicalname:ChemicalID{chemicalid: 
>>>> chemgeneinteractions.ChemicalID, chemicalname: 
>>>> chemgeneinteractions.ChemicalName})
>>>> ON CREATE SET 
>>>> chemicalname.chemicalid=chemgeneinteractions.ChemicalID,chemicalname.chemicalid=chemgeneinteractions.ChemicalName
>>>> merge (geneform:Geneform {geneform: chemgeneinteractions.GeneForms})
>>>> merge (interations:Interaction{interact: chemgeneinteractions.Interaction})
>>>> merge (oraganism:Organism{organism: chemgeneinteractions.Organism})
>>>> merge (interaction:Interactionactions{interr: 
>>>> chemgeneinteractions.InteractionActions})
>>>> merge (pubmed:PubmedID{pub: chemgeneinteractions.PubMedIDs})
>>>> merge (geneid)-[:Gene_Symbol]->(genesymbol)
>>>> merge (geneid)-[:chemicalname]->(chemicalname)
>>>> merge (geneid)-[:geneform]->(genefrom)
>>>> merge (geneid)-[:Its_interaction_action]->(interactions)
>>>> merge (geneid)-[:Its_interaction]->(interaction)
>>>> merge (geneid)-[:PubmedID]->(pubmed)
>>>> merge (geneid)-[:Related_To]->(organism)
>>>> merge (genesymbol)-[:chemicalname]->(chemicalname)
>>>> merge (genesymbol)-[:geneform]->(genefrom)
>>>> merge (genesymbol)-[:geneform]->(genefrom)
>>>> merge (genesymbol)-[:Its_interaction_action]->(interactions)
>>>> merge (genesymbol)-[:Its_interaction]->(interaction)
>>>> merge (genesymbol)-[:PubmedID]->(pubmed)
>>>> merge (genesymbol)-[:Related_To]->(organism)
>>>> 
>>>> 
>>>> 
>>>> My jvm properties are 
>>>> -Xmx512m
>>>> -XX:+UseConcMarkSweepGC
>>>> 
>>>> 
>>>> 
>>>> 
>>>> On Wed, Jun 25, 2014 at 4:49 PM, Michael Hunger 
>>>> <[email protected]> wrote:
>>>>> Please read this blog post: 
>>>>> http://jexp.de/blog/2014/06/load-csv-into-neo4j-quickly-and-successfully/
>>>>> 
>>>>> And yes you should use more memory than 512 byte.
>>>>> 
>>>>> -Xmns4G -Xmx4G -Xmn1G 
>>>>> 
>>>>> 
>>>>> On Wed, Jun 25, 2014 at 1:17 PM, Michael Hunger 
>>>>> <[email protected]> wrote:
>>>>>> What does your query look like?
>>>>>> Please switch to Neo4j 2.1.2
>>>>>> 
>>>>>> And create indexes / constraints for the nodes you're inserting with 
>>>>>> merge or looking up via MATCH.
>>>>>> 
>>>>>> 
>>>>>> Am 18.06.2014 um 08:46 schrieb Pavan Kumar <[email protected]>:
>>>>>> 
>>>>>>> Hi,
>>>>>>> I have deployed neo4j 2.1.0-M01 on windows which has 8GB RAM. I am 
>>>>>>> trying to import CSV file which has 30000 records. I am using USING 
>>>>>>> PERIODIC COMMIT 1000 LOAD CSV command for importing, but it gives 
>>>>>>> unknown error. I have modified neo4j.properties file as adviced in the 
>>>>>>> blogs. My neo4j.properties now looks like 
>>>>>>> # Default values for the low-level graph engine
>>>>>>> 
>>>>>>> neostore.nodestore.db.mapped_memory=200M
>>>>>>> neostore.relationshipstore.db.mapped_memory=4G
>>>>>>> neostore.propertystore.db.mapped_memory=500M
>>>>>>> neostore.propertystore.db.strings.mapped_memory=500M
>>>>>>> neostore.propertystore.db.arrays.mapped_memory=500M
>>>>>>> 
>>>>>>> # Enable this to be able to upgrade a store from an older version
>>>>>>> allow_store_upgrade=true
>>>>>>> 
>>>>>>> # Enable this to specify a parser other than the default one.
>>>>>>> #cypher_parser_version=2.0
>>>>>>> 
>>>>>>> # Keep logical logs, helps debugging but uses more disk space, enabled 
>>>>>>> for
>>>>>>> # legacy reasons To limit space needed to store historical logs use 
>>>>>>> values such
>>>>>>> # as: "7 days" or "100M size" instead of "true"
>>>>>>> keep_logical_logs=true
>>>>>>> 
>>>>>>> # Autoindexing
>>>>>>> 
>>>>>>> # Enable auto-indexing for nodes, default is false
>>>>>>> node_auto_indexing=true
>>>>>>> 
>>>>>>> # The node property keys to be auto-indexed, if enabled
>>>>>>> #node_keys_indexable=name,age
>>>>>>> 
>>>>>>> # Enable auto-indexing for relationships, default is false
>>>>>>> relationship_auto_indexing=true
>>>>>>> 
>>>>>>> # The relationship property keys to be auto-indexed, if enabled
>>>>>>> #relationship_keys_indexable=name,age
>>>>>>> 
>>>>>>> # Setting for Community Edition:
>>>>>>> cache_type=weak
>>>>>>> 
>>>>>>> Still i am facing the same problem. Is there any other file to change 
>>>>>>> properties. Kindly help me in this issue.
>>>>>>> Thanks in advance
>>>>>>> 
>>>>>>> On Tuesday, 4 March 2014 21:24:03 UTC+5:30, Aram Chung wrote:
>>>>>>>> 
>>>>>>>> Hi,
>>>>>>>> 
>>>>>>>> I was asked to post this here by Mark Needham (@markhneedham) who 
>>>>>>>> thought my query took longer than it should.
>>>>>>>> 
>>>>>>>> I'm trying to see how graph databases could be used in investigative 
>>>>>>>> journalism: I was loading in New York State's Active Corporations: 
>>>>>>>> Beginning 1800 data from 
>>>>>>>> https://data.ny.gov/Economic-Development/Active-Corporations-Beginning-1800/n9v6-gdp6
>>>>>>>>  as a 1964486-row csv (and deleted all U+F8FF characters, because I 
>>>>>>>> was getting "[null] is not a supported property value"). The Cypher 
>>>>>>>> query I used was 
>>>>>>>> 
>>>>>>>> USING PERIODIC COMMIT 500
>>>>>>>> LOAD CSV
>>>>>>>>   FROM 
>>>>>>>> "file://path/to/csv/Active_Corporations___Beginning_1800__without_header__wonky_characters_fixed.csv"
>>>>>>>>   AS company
>>>>>>>> CREATE (:DataActiveCorporations
>>>>>>>>        {
>>>>>>>>                DOS_ID:company[0],
>>>>>>>>                Current_Entity_Name:company[1],
>>>>>>>>                Initial_DOS_Filing_Date:company[2],
>>>>>>>>                County:company[3],
>>>>>>>>                Jurisdiction:company[4],
>>>>>>>>                Entity_Type:company[5],
>>>>>>>> 
>>>>>>>>                DOS_Process_Name:company[6],
>>>>>>>>                DOS_Process_Address_1:company[7],
>>>>>>>>                DOS_Process_Address_2:company[8],
>>>>>>>>                DOS_Process_City:company[9],
>>>>>>>>                DOS_Process_State:company[10],
>>>>>>>>                DOS_Process_Zip:company[11],
>>>>>>>> 
>>>>>>>>                CEO_Name:company[12],
>>>>>>>>                CEO_Address_1:company[13],
>>>>>>>>                CEO_Address_2:company[14],
>>>>>>>>                CEO_City:company[15],
>>>>>>>>                CEO_State:company[16],
>>>>>>>>                CEO_Zip:company[17],
>>>>>>>> 
>>>>>>>>                Registered_Agent_Name:company[18],
>>>>>>>>                Registered_Agent_Address_1:company[19],
>>>>>>>>                Registered_Agent_Address_2:company[20],
>>>>>>>>                Registered_Agent_City:company[21],
>>>>>>>>                Registered_Agent_State:company[22],
>>>>>>>>                Registered_Agent_Zip:company[23],
>>>>>>>> 
>>>>>>>>                Location_Name:company[24],
>>>>>>>>                Location_Address_1:company[25],
>>>>>>>>                Location_Address_2:company[26],
>>>>>>>>                Location_City:company[27],
>>>>>>>>                Location_State:company[28],
>>>>>>>>                Location_Zip:company[29]
>>>>>>>>        }
>>>>>>>> );
>>>>>>>> 
>>>>>>>> Each row is one node so it's as close to the raw data as possible. The 
>>>>>>>> idea is loosely that these nodes will be linked with new nodes 
>>>>>>>> representing people and addresses verified by reporters.
>>>>>>>> 
>>>>>>>> This is what I got:
>>>>>>>> 
>>>>>>>> +-------------------+
>>>>>>>> | No data returned. |
>>>>>>>> +-------------------+
>>>>>>>> Nodes created: 1964486
>>>>>>>> Properties set: 58934580
>>>>>>>> Labels added: 1964486
>>>>>>>> 4550855 ms
>>>>>>>> 
>>>>>>>> Some context information: 
>>>>>>>> Neo4j Milestone Release 2.1.0-M01
>>>>>>>> Windows 7
>>>>>>>> java version "1.7.0_03"
>>>>>>>> 
>>>>>>>> Best,
>>>>>>>> Aram
>>>>>>> 
>>>>>>> 
>>>>>>> -- 
>>>>>>> You received this message because you are subscribed to the Google 
>>>>>>> Groups "Neo4j" group.
>>>>>>> To unsubscribe from this group and stop receiving emails from it, send 
>>>>>>> an email to [email protected].
>>>>>>> For more options, visit https://groups.google.com/d/optout.
>>>>> 
>>>>> -- 
>>>>> You received this message because you are subscribed to a topic in the 
>>>>> Google Groups "Neo4j" group.
>>>>> To unsubscribe from this topic, visit 
>>>>> https://groups.google.com/d/topic/neo4j/a2DdoKkbyYo/unsubscribe.
>>>>> To unsubscribe from this group and all its topics, send an email to 
>>>>> [email protected].
>>>>> 
>>>>> For more options, visit https://groups.google.com/d/optout.
>>>> 
>>>> 
>>>> 
>>>> -- 
>>>> Thanks & Regards,
>>>> Pavan Kumar
>>>> Project Engineer
>>>> CDAC -KP
>>>> Ph +91-7676367646
>>>> -- 
>>>> You received this message because you are subscribed to the Google Groups 
>>>> "Neo4j" group.
>>>> To unsubscribe from this group and stop receiving emails from it, send an 
>>>> email to [email protected].
>>>> For more options, visit https://groups.google.com/d/optout.
>>> 
>>> -- 
>>> You received this message because you are subscribed to a topic in the 
>>> Google Groups "Neo4j" group.
>>> To unsubscribe from this topic, visit 
>>> https://groups.google.com/d/topic/neo4j/a2DdoKkbyYo/unsubscribe.
>>> To unsubscribe from this group and all its topics, send an email to 
>>> [email protected].
>>> For more options, visit https://groups.google.com/d/optout.
>> 
>> 
>> 
>> -- 
>> Thanks & Regards,
>> Pavan Kumar
>> Project Engineer
>> CDAC -KP
>> Ph +91-7676367646
> 
> 
> 
> -- 
> Thanks & Regards,
> Pavan Kumar
> Project Engineer
> CDAC -KP
> Ph +91-7676367646
> -- 
> You received this message because you are subscribed to the Google Groups 
> "Neo4j" group.
> To unsubscribe from this group and stop receiving emails from it, send an 
> email to [email protected].
> For more options, visit https://groups.google.com/d/optout.

-- 
You received this message because you are subscribed to the Google Groups 
"Neo4j" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to