On 6/17/09, Mark H. Wood <[email protected]> wrote: > On Wed, Jun 17, 2009 at 03:35:14PM +0100, Andrew Marlow wrote: >> I am working on a proprietary digital library whilst at the same time >> considering how dspace might have been used to solve the same problems (it >> won't be, but that's another story). When I consider usage event reporting >> there are some concerns that arise when the number of articles and >> visitors >> is very large. The current site has around 6 million articles and roughly >> 20 million hits per day.
> OLTP systems cope with billions of rows/day giving reasonable > performance, but your requirements may be quite different. They > probably journal that stuff immediately and post it into tables later, > like the guy with the green eyeshade used to do on paper. > Have you benchmarked totalling a 2-giga-row column? How long did it > take to sum it? How long did it take to [snip] > In high-performance systems you really have to try different ideas and > compare their real-world performance -- theory gets you only so far. > -- > Mark H. Wood, Lead System Programmer [email protected] > Friends don't let friends publish revisable-form documents. I did a brief experiment to see how long it would take to suck up thousands of events from a JMS queue into a LucidDB using flat files which are then BCP'd. This was plenty fast enough and is probably what we will do in our system. -- Regards, Andrew M. http://www.andrewpetermarlow.co.uk ------------------------------------------------------------------------------ _______________________________________________ DSpace-tech mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/dspace-tech

