[
https://issues.apache.org/jira/browse/HADOOP-2612?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12559700#action_12559700
]
Bryan Duxbury commented on HADOOP-2612:
---------------------------------------
Can you reproduce this? Does it crash anything other than your client? If not,
this should probably be downgraded to a lower priority.
> Mysterious ArrayOutOfBoundsException in HTable.commit
> -----------------------------------------------------
>
> Key: HADOOP-2612
> URL: https://issues.apache.org/jira/browse/HADOOP-2612
> Project: Hadoop
> Issue Type: Bug
> Components: contrib/hbase
> Reporter: Michael Bieniosek
>
> I got this exception using a post-0.15.0 hbase trunk:
> Caused by: java.io.IOException: java.io.IOException:
> java.lang.ArrayIndexOutOfBoundsException
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
> at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
> Source)
> at java.lang.reflect.Constructor.newInstance(Unknown Source)
> at
> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
> at org.apache.hadoop.hbase.HTable.commit(HTable.java:904)
> at org.apache.hadoop.hbase.HTable.commit(HTable.java:875)
> at xxx.PutHbase$HbaseUploader.writeHbaseNoRetry(PutHbase.java:107)
> Where writeHbaseNoRetry looks like:
> private void writeHbaseNoRetry(HTable table, String column, String row,
> File contents) throws IOException {
> long lockid = table.startUpdate(new Text(row));
> try {
> table.put(lockid, new Text(column), FileUtil.readFile(contents));
> table.commit(lockid);
> } finally {
> table.abort(lockid);
> }
> }
> I found this in my error logs -- it is rare, and I am not sure how to
> reproduce it. Contents could be 1kb-100kb long.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.