[ 
https://issues.apache.org/jira/browse/CARBONDATA-1362?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16115420#comment-16115420
 ] 

cen yuhai commented on CARBONDATA-1362:
---------------------------------------

I use the old code to create carbon table and then load data, and then I update 
my code to master, query the data will throws this exception. If I recreate the 
table, it will be ok 

> ArrayIndexOutOfBoundsException when decoing decimal type
> --------------------------------------------------------
>
>                 Key: CARBONDATA-1362
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1362
>             Project: CarbonData
>          Issue Type: Bug
>          Components: core
>            Reporter: cen yuhai
>
> {code}
> ava.lang.RuntimeException: java.util.concurrent.ExecutionException: 
> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0
>       at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.close(AbstractDataBlockIterator.java:231)
>       at 
> org.apache.carbondata.core.scan.result.iterator.AbstractDetailQueryResultIterator.close(AbstractDetailQueryResultIterator.java:306)
>       at 
> org.apache.carbondata.core.scan.executor.impl.AbstractQueryExecutor.finish(AbstractQueryExecutor.java:544)
>       at 
> org.apache.carbondata.spark.vectorreader.VectorizedCarbonRecordReader.close(VectorizedCarbonRecordReader.java:132)
>       at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1$$anonfun$7.apply(CarbonScanRDD.scala:215)
>       at 
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1$$anonfun$7.apply(CarbonScanRDD.scala:213)
>       at 
> org.apache.spark.TaskContext$$anon$1.onTaskCompletion(TaskContext.scala:123)
>       at 
> org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:97)
>       at 
> org.apache.spark.TaskContextImpl$$anonfun$markTaskCompleted$1.apply(TaskContextImpl.scala:95)
>       at 
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>       at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
>       at 
> org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:95)
>       at org.apache.spark.scheduler.Task.run(Task.scala:117)
>       at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:351)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.ExecutionException: 
> java.lang.RuntimeException: java.lang.ArrayIndexOutOfBoundsException: 0
>       at java.util.concurrent.FutureTask.report(FutureTask.java:122)
>       at java.util.concurrent.FutureTask.get(FutureTask.java:188)
>       at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator.close(AbstractDataBlockIterator.java:226)
>       ... 16 more
> Caused by: java.lang.RuntimeException: 
> java.lang.ArrayIndexOutOfBoundsException: 0
>       at 
> org.apache.carbondata.core.datastore.chunk.impl.MeasureRawColumnChunk.convertToMeasureColDataChunks(MeasureRawColumnChunk.java:62)
>       at 
> org.apache.carbondata.core.scan.scanner.AbstractBlockletScanner.scanBlocklet(AbstractBlockletScanner.java:100)
>       at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator$1.call(AbstractDataBlockIterator.java:191)
>       at 
> org.apache.carbondata.core.scan.processor.AbstractDataBlockIterator$1.call(AbstractDataBlockIterator.java:178)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>       ... 3 more
> Caused by: java.lang.ArrayIndexOutOfBoundsException: 0
>       at 
> org.apache.carbondata.core.util.DataTypeUtil.byteToBigDecimal(DataTypeUtil.java:210)
>       at 
> org.apache.carbondata.core.metadata.ColumnPageCodecMeta.deserialize(ColumnPageCodecMeta.java:217)
>       at 
> org.apache.carbondata.core.datastore.chunk.reader.measure.v3.CompressedMeasureChunkFileBasedReaderV3.decodeMeasure(CompressedMeasureChunkFileBasedReaderV3.java:236)
>       at 
> org.apache.carbondata.core.datastore.chunk.reader.measure.v3.CompressedMeasureChunkFileBasedReaderV3.convertToMeasureChunk(CompressedMeasureChunkFileBasedReaderV3.java:219)
>       at 
> org.apache.carbondata.core.datastore.chunk.impl.MeasureRawColumnChunk.convertToMeasureColDataChunks(MeasureRawColumnChunk.java:59)
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to