[ https://issues.apache.org/jira/browse/HADOOP-2612?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12559700#action_12559700 ]
Bryan Duxbury commented on HADOOP-2612: --------------------------------------- Can you reproduce this? Does it crash anything other than your client? If not, this should probably be downgraded to a lower priority. > Mysterious ArrayOutOfBoundsException in HTable.commit > ----------------------------------------------------- > > Key: HADOOP-2612 > URL: https://issues.apache.org/jira/browse/HADOOP-2612 > Project: Hadoop > Issue Type: Bug > Components: contrib/hbase > Reporter: Michael Bieniosek > > I got this exception using a post-0.15.0 hbase trunk: > Caused by: java.io.IOException: java.io.IOException: > java.lang.ArrayIndexOutOfBoundsException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown > Source) > at java.lang.reflect.Constructor.newInstance(Unknown Source) > at > org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82) > at org.apache.hadoop.hbase.HTable.commit(HTable.java:904) > at org.apache.hadoop.hbase.HTable.commit(HTable.java:875) > at xxx.PutHbase$HbaseUploader.writeHbaseNoRetry(PutHbase.java:107) > Where writeHbaseNoRetry looks like: > private void writeHbaseNoRetry(HTable table, String column, String row, > File contents) throws IOException { > long lockid = table.startUpdate(new Text(row)); > try { > table.put(lockid, new Text(column), FileUtil.readFile(contents)); > table.commit(lockid); > } finally { > table.abort(lockid); > } > } > I found this in my error logs -- it is rare, and I am not sure how to > reproduce it. Contents could be 1kb-100kb long. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.