Thanks a lot Vijay for your valuable inputs. Best Regards, Saurav
On Wed, Oct 20, 2021 at 1:06 AM Vijaykumar Jain < vijaykumarjain.git...@gmail.com> wrote: > On Tue, 19 Oct 2021 at 23:20, Vijaykumar Jain < > vijaykumarjain.git...@gmail.com> wrote: > >> >> On Tue, 19 Oct 2021 at 23:09, Vijaykumar Jain < >> vijaykumarjain.git...@gmail.com> wrote: >> >>> >>> On Tue, 19 Oct 2021 at 22:45, Saurav Sarkar <saurav.sark...@gmail.com> >>> wrote: >>> >>>> Hi All, >>>> >>>> >>>> A basic question on handling large number of concurrent requests on DB. >>>> >>>> I have a cloud service which can get large of requests which will >>>> obviously trigger the db operations. >>>> >>>> Every db will have some max connection limit which can get exhausted on >>>> large number of requests. >>>> >>>> I know db connection pooling can be used to reuse the connections but >>>> it will not help when there are large number of active concurrent >>>> connections. My queries are already optimised and short living. >>>> >>>> For that i need some queuing mechanism like pgbouncer for postgres >>>> https://www.percona.com/blog/2021/02/26/connection-queuing-in-pgbouncer-is-it-a-magical-remedy/ >>>> >>>> pgbounder i understand is a proxy which needs to be separately >>>> installed on the web or db server. >>>> >>>> I was thinking if the normal client side db connection pooling >>>> libraries like Apache DBCP , can also provide similar connection queuing >>>> while running in the application runtime. >>>> >>> >>> >> > also pls checkout, i forgot to link early on > Number Of Database Connections - PostgreSQL wiki > <https://wiki.postgresql.org/wiki/Number_Of_Database_Connections> > it explains the reasons, too many direct connections may result in > performance issues. > >