Dear all,
I'm working on a collection containing tens of millions of documents, updated weekly.
My first guess to store these documents was to simply db:add/replace/remove them.
This solution is slowing things down as the count of documents increases.
For example, opening one collection takes up to 4 minutes.
I believe that the document list is an in memory structure.
Is there a way to speed things up, or do I have to change sides with the following guess, in order to reduce the 'physical' document list size :
1.group documents in 'logical' documents on insertion (fewer documents containing new or updated documents under a root xml element)
2. remove the old version of these documents in the previous 'logical' documents with xquery update.
Has anybody already find that problem, and a workaround ?
BaseX is just fantastic !
Best regards,
Fabrice Etanchaud
Questel/Orbit