Hi all I have a requirement to archive several million documents with variable metadata (dates, types of document). There is only the update process that adds documents daily, and perhaps some very few updates manually (depending on resolution of requirements). Documents will be retrieved by their position in the tree, plus some filtering on the metadata.
I was planning to deploy jackrabbit with a war that implements a simplified web service for the update process and the retrieval. The searching should not be very heavily used, peaking at a few calls per second, max. I sold it to management with the promise of scaling up to a licensed implementation if/when necessary. Couple million docs, terabyte of data, modest throughput, availability of upgrade path, easy backups. From my reading/lurking, this sounds like a job for bundle persistence manager using H2 database. Have I missed something? Does this sound like a workable plan? Thanks for comments Ross.
-- This message contains privileged and confidential information only for use by the intended recipient. If you are not the intended recipient of this message, you must not disseminate, copy or use it in any manner. If you have received this message in error, please advise the sender by reply e-mail. Please ensure all e-mail attachments are scanned for viruses prior to opening or using.
