Hi,

I don't think we have connection leak in normal behavior.

The actual SQL statement is executed in @FinishBundle, where the
connection is closed.

The processElement adds record to process.

Does it mean that an Exception occurs in the batch addition ?

Regards
JB

On 17/01/2019 12:41, Alexey Romanenko wrote:
> Kenn,
> 
> I’m not sure that we have a connection leak in JdbcIO since new
> connection is being obtained from an instance of /javax.sql.DataSource/
> (created in @Setup) and which
> is /org.apache.commons.dbcp2.BasicDataSource/ by default.
> /BasicDataSource/ uses connection pool and closes all idle connections
> in "close()”. 
> 
> In its turn, JdbcIO calls/DataSource.close()/ in @Teardown, so all idle
> connections should be closed and released there in case of fails.
> Though, potentially some connections, that has been delegated to client
> before and were not not properly returned to pool, could be leaked…
> Anyway, I think it could be a good idea to call "/connection.close()/”
> (return to connection pool) explicitly in case of any exception happened
> during bundle processing.
> 
> Probably JB may provide more details as original author of JdbcIO.
> 
>> On 14 Jan 2019, at 21:37, Kenneth Knowles <k...@apache.org
>> <mailto:k...@apache.org>> wrote:
>>
>> Hi Jonathan,
>>
>> JdbcIO.write() just invokes this
>> DoFn: 
>> https://github.com/apache/beam/blob/master/sdks/java/io/jdbc/src/main/java/org/apache/beam/sdk/io/jdbc/JdbcIO.java#L765
>>
>> It establishes a connection in @StartBundle and then in @FinishBundle
>> it commits a batch and closes the connection. If an error happens
>> in @StartBundle or @ProcessElement there will be a retry with a fresh
>> instance of the DoFn, which will establish a new connection. It looks
>> like neither @StartBundle nor @ProcessElement closes the connection,
>> so I'm guessing that the old connection sticks around because the
>> worker process was not terminated. So the Beam runner and Dataflow
>> service are working as intended and this is an issue with JdbcIO,
>> unless I've made a mistake in my reading or analysis.
>>
>> Would you mind reporting these details
>> to https://issues.apache.org/jira/projects/BEAM/ ?
>>
>> Kenn
>>
>> On Mon, Jan 14, 2019 at 12:51 AM Jonathan Perron
>> <jonathan.per...@lumapps.com <mailto:jonathan.per...@lumapps.com>> wrote:
>>
>>     Hello !
>>
>>     My question is maybe mainly GCP-oriented, so I apologize if it is
>>     not fully related to the Beam community.
>>
>>     We have a streaming pipeline running on Dataflow which writes data
>>     to a PostgreSQL instance hosted on Cloud SQL. This database is
>>     suffering from connection leak spikes on a regular basis:
>>
>>     <ofbkcnmdfbgcoooc.png>
>>
>>     The connections are kept alive until the pipeline is canceled/drained:
>>
>>     <gngklddbhnckgpni.png>
>>
>>     We are writing to the database with:
>>
>>     - individual DoFn where we open/close the connection using the
>>     standard JDBC try/catch (SQLException ex)/finally statements;
>>
>>     - a Pipeline.apply(JdbcIO.<SessionData>write()) operations.
>>
>>     I observed that these spikes happens most of the time after I/O
>>     errors with the database. Has anyone observed the same situation ?
>>
>>     I have several questions/observations, please correct me if I am
>>     wrong (I am not from the java environment, so some can seem pretty
>>     trivial) :
>>
>>     - Does the apply method handles SQLException or I/O errors ?
>>
>>     - Would the use of a connection pool prevents such behaviours ? If
>>     so, how would one implement it to allow all workers to use it ?
>>     Could it be implemented with JDBC Connection pooling ?
>>
>>     I am worrying about the serialization if one would pass a
>>     Connection item as an argument of a DoFn.
>>
>>     Thank you in advance for your comments and reactions.
>>
>>     Best regards,
>>
>>     Jonathan
>>
> 

-- 
Jean-Baptiste Onofré
jbono...@apache.org
http://blog.nanthrax.net
Talend - http://www.talend.com

Reply via email to