While putting data in Hbase, using the HTable.put method, I'll end up with the below exception occasionally. But the data has been actually written to Hbase when I checked the get operation for that particular rowkey.
For the same time I have searched for the logs in both HMaster and HRegionservers to identify the issue. But unable to find that. hbase.client.* configurations has default value only. Please help to fine tune Hbase Configurations in order to avoid InterruptedIOException. Hadoop Distribution: ApacheVersion: HBase 1.2.6Cluster size: 12nodes java.io.InterruptedIOException: #17209, interrupted. currentNumberOfTask=1 at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1764) at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1734) at org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1810) at org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:240) at org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:190) at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1434) at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1018) Please help to solve it The same exception has been faced by someone . But in that thread, there is no explanation about which are configurations need to be checked in order to avoid it https://groups.google.com/forum/#!topic/nosql-databases/UxfrmWl_ZnM *Regards,* *Manimekalai K*