Hi Mani, Can you please check if the client thread from which your inserting the data is being interrupted ? A thread pool shutdown also can cause this.
Regards, Mukund Murrali On Wed, Sep 19, 2018 at 8:54 AM Manimekalai K <kmanimeka...@gmail.com> wrote: > While putting data in Hbase, using the HTable.put method, I'll end up with > the below exception occasionally. But the data has been actually written to > Hbase when I checked the get operation for that particular rowkey. > > For the same time I have searched for the logs in both HMaster and > HRegionservers to identify the issue. But unable to find that. > > hbase.client.* configurations has default value only. > > Please help to fine tune Hbase Configurations in order to avoid > InterruptedIOException. > > Hadoop Distribution: ApacheVersion: HBase 1.2.6Cluster size: 12nodes > > > > java.io.InterruptedIOException: #17209, interrupted. currentNumberOfTask=1 > at > org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1764) > at > org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1734) > at > org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1810) > at > org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:240) > at > org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:190) > at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1434) > at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1018) > > Please help to solve it > > The same exception has been faced by someone . But in that thread, there is > no explanation about which are configurations need to be checked in order > to avoid it > > https://groups.google.com/forum/#!topic/nosql-databases/UxfrmWl_ZnM > *Regards,* > <https://groups.google.com/forum/#!topic/nosql-databases/UxfrmWl_ZnM*Regards,*> > *Manimekalai K* >