Hi,
We are currently planning the deployment of our next generation enterprise
database and we are wondering whether or not PostgreSQL could do the heavy
lifting that would be required. My post is a little bit long but I hope it
will provide you with information to allow someone to provide a definitive
answer.
First, a little history about our current setup. We are currently running
SQL Server 2000 on Windows 2000 Advanced Server on HA clustered Dell boxes with
4 cpu's and 8 gigabyte of RAM. This is attached via Fiber to an EMC Clariion
solution. Our current database is around 250 gigabytes big (including the size
of the index files) and has averaged about 60 gigabytes of growth per year. We
have around 200 concurrent users that heavily utilize the database.
We are currently in the planning stages of our next generation database
system. We expect dramatic growth in the coming years and would like to design
a database solution that will last at least 5 years. Within two years, we
estimate that we will have around 500 concurrent users and estimate that our
database will grow to around 500 gigabytes. Within five years, we estimate
that we will have around 1000 concurrent users and estimate that our database
will grow to around 1 terabyte.
The major concern we have is that we expect database activity to increase
dramatically over the current utilization. Besides the planned increase in the
number of employees, there will also be increased database resouce utilization
per employee as management is pushing to increase performance per empoyee and
increased data analysis to measure the success of the business. So it is very
important that we implement a solution that can scale well.
This will be a rather enterprise quality solution. On the hardware, we are
leaning on either an EMC or NetApp SAN solution. For a database, we plan to
either deploy RHEL as it provides migration from AMD64 to Itanium/Power or
Solaris as it provides migration from AMD64 to Sparc. On the database end, the
possible options include Oracle, DB2, Sybase, or PostgreSQL. We would prefer
to go with PostgreSQL due to the dramatic cost savings we can achieve. We have
come to discover that as expensive as the hardware/OS solution is going to be,
the commercial database costs will dwarf those costs.
As this database will be our core database and our entire world-wide branches
will be completely dependent on it, we will need to make sure that it can
perform, scale upwards, and provide high availability features. I already know
that PostgreSQL provides high availability. The other two, I am uncertain.
Will PostgreSQL be able to handle this job? What do we need to look out for if
we are to do such a deployment? What is the largest database someone has
deployed in production? Largest table? Any help with this situation will be
greatly appreciated.
---------------------------(end of broadcast)---------------------------
TIP 3: Have you checked our extensive FAQ?
http://www.postgresql.org/docs/faq