Hi Alessandro,

I'm actually on my dev' computer, so I would like to post 1 000 000 xml file (with a structure defined in my schema.xml)

I have already import 1 000 000 xml files by using
bin/post -c mydb /DATA0/1 /DATA0/2 /DATA0/3 /DATA0/4 /DATA0/5
where /DATA0/X contains 20 000 xml files (I do it 20 times by just changing X from 1 to 50)

I would like to do now
bin/post -c mydb /DATA1

I would like to know If my SOLR5 will run fine and no provide an memory error because there are too many files
in one post without doing a commit?

The commit will be done at the end of 1 000 000.

Is it ok ?


Le 05/06/2015 16:59, Alessandro Benedetti a écrit :
Hi Bruno,
I can not see what is your challenge.
Of course you can index your data in the flavour you want and do a commit
whenever you want…
Are those xml Solr xml ?
If not you would need to use the DIH, the extract update handler or any
custom Indexer application.
Maybe I missed your point…
Give me more details please !

Cheers

2015-06-05 15:41 GMT+01:00 Bruno Mannina <bmann...@free.fr>:

Dear Solr Users,

I would like to post  1 000 000 records (1 records = 1 files) in one shoot
?
and do the commit and the end.

Is it possible to do that ?

I've several directories with each 20 000 files inside.
I would like to do:
bin/post -c mydb /DATA

under DATA I have
/DATA/1/*.xml (20 000 files)
/DATA/2/*.xml (20 000 files)
/DATA/3/*.xml (20 000 files)
....
/DATA/50/*.xml (20 000 files)

Actually, I post 5 directories in one time (it takes around 1h30 for 100
000 records/files)

But it's Friday and I would like to run it during the W.E. alone.

Thanks for your comment,

Bruno

---
Ce courrier électronique ne contient aucun virus ou logiciel malveillant
parce que la protection avast! Antivirus est active.
https://www.avast.com/antivirus





---
Ce courrier électronique ne contient aucun virus ou logiciel malveillant parce 
que la protection avast! Antivirus est active.
https://www.avast.com/antivirus

Reply via email to