Mysterious ArrayOutOfBoundsException in HTable.commit
-----------------------------------------------------
Key: HADOOP-2612
URL: https://issues.apache.org/jira/browse/HADOOP-2612
Project: Hadoop
Issue Type: Bug
Components: contrib/hbase
Reporter: Michael Bieniosek
I got this exception using a post-0.15.0 hbase trunk:
Caused by: java.io.IOException: java.io.IOException:
java.lang.ArrayIndexOutOfBoundsException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
Source)
at java.lang.reflect.Constructor.newInstance(Unknown Source)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
at org.apache.hadoop.hbase.HTable.commit(HTable.java:904)
at org.apache.hadoop.hbase.HTable.commit(HTable.java:875)
at xxx.PutHbase$HbaseUploader.writeHbaseNoRetry(PutHbase.java:107)
Where writeHbaseNoRetry looks like:
private void writeHbaseNoRetry(HTable table, String column, String row,
File contents) throws IOException {
long lockid = table.startUpdate(new Text(row));
try {
table.put(lockid, new Text(column), FileUtil.readFile(contents));
table.commit(lockid);
} finally {
table.abort(lockid);
}
}
I found this in my error logs -- it is rare, and I am not sure how to reproduce
it. Contents could be 1kb-100kb long.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.