On 6/17/09, Mark H. Wood <[email protected]> wrote:
> On Wed, Jun 17, 2009 at 03:35:14PM +0100, Andrew Marlow wrote:
>> I am working on a proprietary digital library whilst at the same time
>> considering how dspace might have been used to solve the same problems (it
>> won't be, but that's another story). When I consider usage event reporting
>> there are some concerns that arise when the number of articles and
>> visitors
>> is very large. The current site has around 6 million articles and roughly
>> 20 million hits per day.

> OLTP systems cope with billions of rows/day giving reasonable
> performance, but your requirements may be quite different.  They
> probably journal that stuff immediately and post it into tables later,
> like the guy with the green eyeshade used to do on paper.

> Have you benchmarked totalling a 2-giga-row column?  How long did it
> take to sum it?  How long did it take to
[snip]

> In high-performance systems you really have to try different ideas and
> compare their real-world performance -- theory gets you only so far.
> --
> Mark H. Wood, Lead System Programmer   [email protected]
> Friends don't let friends publish revisable-form documents.

I did a brief experiment to see how long it would take to suck up
thousands of events from a JMS queue into a LucidDB using flat files
which are then BCP'd. This was plenty fast enough and is probably what
we will do in our system.

-- 
Regards,

Andrew M.
http://www.andrewpetermarlow.co.uk

------------------------------------------------------------------------------
_______________________________________________
DSpace-tech mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dspace-tech

Reply via email to