Hi Alessandro,
I'm actually on my dev' computer, so I would like to post 1 000 000 xml
file (with a structure defined in my schema.xml)
I have already import 1 000 000 xml files by using
bin/post -c mydb /DATA0/1 /DATA0/2 /DATA0/3 /DATA0/4 /DATA0/5
where /DATA0/X contains 20 000 xml files (I do it 20 times by just
changing X from 1 to 50)
I would like to do now
bin/post -c mydb /DATA1
I would like to know If my SOLR5 will run fine and no provide an memory
error because there are too many files
in one post without doing a commit?
The commit will be done at the end of 1 000 000.
Is it ok ?
Le 05/06/2015 16:59, Alessandro Benedetti a écrit :
Hi Bruno,
I can not see what is your challenge.
Of course you can index your data in the flavour you want and do a commit
whenever you want…
Are those xml Solr xml ?
If not you would need to use the DIH, the extract update handler or any
custom Indexer application.
Maybe I missed your point…
Give me more details please !
Cheers
2015-06-05 15:41 GMT+01:00 Bruno Mannina <bmann...@free.fr>:
Dear Solr Users,
I would like to post 1 000 000 records (1 records = 1 files) in one shoot
?
and do the commit and the end.
Is it possible to do that ?
I've several directories with each 20 000 files inside.
I would like to do:
bin/post -c mydb /DATA
under DATA I have
/DATA/1/*.xml (20 000 files)
/DATA/2/*.xml (20 000 files)
/DATA/3/*.xml (20 000 files)
....
/DATA/50/*.xml (20 000 files)
Actually, I post 5 directories in one time (it takes around 1h30 for 100
000 records/files)
But it's Friday and I would like to run it during the W.E. alone.
Thanks for your comment,
Bruno
---
Ce courrier électronique ne contient aucun virus ou logiciel malveillant
parce que la protection avast! Antivirus est active.
https://www.avast.com/antivirus
---
Ce courrier électronique ne contient aucun virus ou logiciel malveillant parce
que la protection avast! Antivirus est active.
https://www.avast.com/antivirus