[ 
https://issues.apache.org/jira/browse/HBASE-19201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16243432#comment-16243432
 ] 

Yung-An He commented on HBASE-19201:
------------------------------------

Hi

Could you post the needed code which you copied into the scala project? 

> BulkLoading in HBaseContext in hbase-spark does not close connection
> --------------------------------------------------------------------
>
>                 Key: HBASE-19201
>                 URL: https://issues.apache.org/jira/browse/HBASE-19201
>             Project: HBase
>          Issue Type: Bug
>          Components: hbase
>    Affects Versions: 1.1.12
>         Environment: I was using the cdh 5.11.1 version but I checken on 
> newest branch and problem persists
>            Reporter: Lucas Resch
>              Labels: newbie
>   Original Estimate: 2h
>  Remaining Estimate: 2h
>
> Within the hbase-spark module an HBaseContext exists that provides utility 
> functions to do bulkLoading data in HBase. I tried using this function in a 
> streaming context, but after a while Zookeeper denies further connections 
> since the maximum of connections per client is exhausted. 
> This issue seems to be within HBaseContext, since the functions bulkLoad and 
> bulkLoadThinRows open a connection via the ConnectionFactory, but never 
> closes that connection.
> I copied the needed code into a new scala project and added a conn.close() at 
> the end of the function and the problem is gone. 
> It seems like no one else has had this problem before. I'm guessing thats 
> because almost no one uses its function within a streaming context. And a one 
> time call to it with RDDs might never reach that upper limit on connections. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to