Dear all,

I'm working on a collection containing  tens of millions of documents, updated 
weekly.
My first guess to store these documents was to simply db:add/replace/remove 
them.
This solution is slowing things down as the count of documents increases.
For example, opening one collection takes up to 4 minutes.

I believe that the document list is an in memory structure.

Is there a way to speed things up, or do I have to change sides with the 
following guess, in order to reduce the 'physical' document list size :

1.group documents in 'logical' documents on insertion (fewer documents 
containing new or updated documents under a root xml element)
2. remove the old version of these documents in the previous 'logical' 
documents with xquery update.

Has anybody already find that problem, and a workaround ?

BaseX is just fantastic !

Best regards,
Fabrice  Etanchaud
Questel/Orbit
_______________________________________________
BaseX-Talk mailing list
[email protected]
https://mailman.uni-konstanz.de/mailman/listinfo/basex-talk

Reply via email to