Hi all,
I am doing research for a project of mine where I need to store several billion values for a monitoring and historical tracking system for a big computer system. My currect estimate is that I have to store (somehow) around 1 billion values each month (possibly more).
I was wondering if anyone has had any experience with these kind of big numbers of data in a postgres sql database and how this affects database design and optimization.
What would be important issues when setting up a database this big, and is it at all doable? Or would it be a insane to think about storing up to 5-10 billion rows in a postgres database.
The database's performance is important. There would be no use in storing the data if a query will take ages. Query's should be quite fast if possible.
I would really like to hear people's thoughts/suggestions or "go see a shrink, you must be mad" statements ;)
Kind regards,
Ramon Bastiaans
---------------------------(end of broadcast)--------------------------- TIP 2: you can get off all lists at once with the unregister command (send "unregister YourEmailAddressHere" to [EMAIL PROTECTED])