SeaAndHill created CARBONDATA-3612:
--------------------------------------

             Summary: Caused by: java.io.IOException: Problem in loading 
segment blocks: null
                 Key: CARBONDATA-3612
                 URL: https://issues.apache.org/jira/browse/CARBONDATA-3612
             Project: CarbonData
          Issue Type: Bug
          Components: core, data-load
    Affects Versions: 1.5.1
            Reporter: SeaAndHill


at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:56) 
at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:56) 
at 
org.apache.spark.sql.execution.exchange.ShuffleExchange.doExecute(ShuffleExchange.scala:115)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:138)
 at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) 
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:135) 
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:116) at 
org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:252)
 at 
org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:141)
 at 
org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:141)
 at 
org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:386)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:117)
 at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:138)
 at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) 
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:135) 
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:116) at 
org.apache.spark.sql.execution.exchange.ShuffleExchange.prepareShuffleDependency(ShuffleExchange.scala:88)
 at 
org.apache.spark.sql.execution.exchange.ShuffleExchange$$anonfun$doExecute$1.apply(ShuffleExchange.scala:124)
 at 
org.apache.spark.sql.execution.exchange.ShuffleExchange$$anonfun$doExecute$1.apply(ShuffleExchange.scala:115)
 at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52) 
... 35 moreCaused by: java.io.IOException: Problem in loading segment blocks: 
null at 
org.apache.carbondata.core.indexstore.BlockletDataMapIndexStore.getAll(BlockletDataMapIndexStore.java:193)
 at 
org.apache.carbondata.core.indexstore.blockletindex.BlockletDataMapFactory.getDataMaps(BlockletDataMapFactory.java:144)
 at 
org.apache.carbondata.core.datamap.TableDataMap.prune(TableDataMap.java:139) at 
org.apache.carbondata.hadoop.api.CarbonInputFormat.getPrunedBlocklets(CarbonInputFormat.java:493)
 at 
org.apache.carbondata.hadoop.api.CarbonInputFormat.getDataBlocksOfSegment(CarbonInputFormat.java:412)
 at 
org.apache.carbondata.hadoop.api.CarbonTableInputFormat.getSplits(CarbonTableInputFormat.java:529)
 at 
org.apache.carbondata.hadoop.api.CarbonTableInputFormat.getSplits(CarbonTableInputFormat.java:220)
 at 
org.apache.carbondata.spark.rdd.CarbonScanRDD.internalGetPartitions(CarbonScanRDD.scala:127)
 at org.apache.carbondata.spark.rdd.CarbonRDD.getPartitions(CarbonRDD.scala:66) 
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:252) at 
org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:250) at 
scala.Option.getOrElse(Option.scala:121) at 
org.apache.spark.rdd.RDD.partitions(RDD.scala:250) at 
org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35) 
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:252) at 
org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:250) at 
scala.Option.getOrElse(Option.scala:121) at 
org.apache.spark.rdd.RDD.partitions(RDD.scala:250) at 
org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35) 
at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:252) at 
org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:250) at 
scala.Option.getOrElse(Option.scala:121) at 
org.apache.spark.rdd.RDD.partitions(RDD.scala:250) at 
org.apache.spark.ShuffleDependency.<init>(Dependency.scala:91) at 
org.apache.spark.sql.execution.exchange.ShuffleExchange$.prepareShuffleDependency(ShuffleExchange.scala:264)
 at 
org.apache.spark.sql.execution.exchange.ShuffleExchange.prepareShuffleDependency(ShuffleExchange.scala:87)
 at 
org.apache.spark.sql.execution.exchange.ShuffleExchange$$anonfun$doExecute$1.apply(ShuffleExchange.scala:124)
 at 
org.apache.spark.sql.execution.exchange.ShuffleExchange$$anonfun$doExecute$1.apply(ShuffleExchange.scala:115)
 at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52) 
... 56 moreCaused by: java.lang.RuntimeException at 
org.apache.carbondata.core.indexstore.BlockletDataMapIndexStore.get(BlockletDataMapIndexStore.java:152)
 at 
org.apache.carbondata.core.indexstore.BlockletDataMapIndexStore.getAll(BlockletDataMapIndexStore.java:182)
 ... 84 moreCaused by: java.lang.NullPointerException at 
org.apache.carbondata.core.datastore.filesystem.AbstractDFSCarbonFile.getName(AbstractDFSCarbonFile.java:117)
 at 
org.apache.carbondata.core.indexstore.blockletindex.SegmentIndexFileStore.getCarbonIndexFiles(SegmentIndexFileStore.java:398)
 at 
org.apache.carbondata.core.indexstore.blockletindex.SegmentIndexFileStore.readAllIIndexOfSegment(SegmentIndexFileStore.java:187)
 at 
org.apache.carbondata.core.util.BlockletDataMapUtil.getBlockMetaInfoMap(BlockletDataMapUtil.java:92)
 at 
org.apache.carbondata.core.indexstore.BlockletDataMapIndexStore.get(BlockletDataMapIndexStore.java:110)
 ... 85 more2019-12-07 02:32:49  INFO JobStatusActor:116 - Job 
28504f4c-353a-4dfb-883e-07364d128197 finished OK2019-12-07 02:32:50  INFO 
JavaSparkContextFactory:116 - Loading class 
com.talkingdata.aeplus.analytics.AnalyticsJsonJobServer for app 
sparkAdapter2019-12-07 02:32:50  INFO JobManagerActor:359 - Starting Spark job 
7747a403-d011-455b-aca8-33382aea8a6e 
[com.talkingdata.aeplus.analytics.AnalyticsJsonJobServer]...2019-12-07 02:32:50 
 INFO JobManagerActor:372 - Starting job future thread2019-12-07 02:32:50  INFO 
JobStatusActor:116 - Job 7747a403-d011-455b-aca8-33382aea8a6e started



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to