[
https://issues.apache.org/jira/browse/CARBONDATA-1422?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16146975#comment-16146975
]
Pallavi Singh commented on CARBONDATA-1422:
-------------------------------------------
Hi Zhichao Zhang,
Here is the stack-trace of the error:
17/08/30 15:12:30 INFO CatalystSqlParser: Parsing command: array<string>
17/08/30 15:12:30 AUDIT CarbonDataRDDFactory$:
[pallavi][hduser][Thread-136]Compaction request received for table
30aug_5.uniqdata
17/08/30 15:12:30 INFO HdfsFileLock: pool-24-thread-7 HDFS lock
path:hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock
17/08/30 15:12:30 INFO CarbonDataRDDFactory$: pool-24-thread-7 Acquired the
compaction lock for table 30aug_5.uniqdata
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 0
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 1
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 2
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 3
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 4
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 5
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 6
17/08/30 15:12:30 INFO DataManagementFunc$: pool-24-thread-7 loads identified
for merge is 7
17/08/30 15:12:30 INFO TableInfo: pool-28-thread-1 Table block size not
specified for 30aug_5_uniqdata. Therefore considering the default value 1024 MB
17/08/30 15:12:30 INFO DistributionUtil$: pool-28-thread-1 Requesting total
executors: 1
17/08/30 15:12:30 INFO CarbonMergerRDD: No.of Executors required=1 ,
spark.executor.instances=1, no.of.nodes where data present=1
17/08/30 15:12:30 INFO CarbonMergerRDD: Time taken to wait for executor
allocation is =0millis
17/08/30 15:12:30 INFO CarbonMergerRDD: for the node pallavi
17/08/30 15:12:30 INFO CarbonMergerRDD: Task ID is 0no. of blocks is 8
17/08/30 15:12:30 INFO CarbonMergerRDD: Identified no.of.Blocks:
8,parallelism: 4 , no.of.nodes: 1, no.of.tasks: 1
17/08/30 15:12:30 INFO CarbonMergerRDD: Time taken to identify Blocks to scan :
177
17/08/30 15:12:30 INFO CarbonMergerRDD: Node: pallavi, No.Of Blocks: 8
17/08/30 15:12:30 INFO SparkContext: Starting job: collect at
Compactor.scala:100
17/08/30 15:12:30 INFO DAGScheduler: Got job 4 (collect at Compactor.scala:100)
with 1 output partitions
17/08/30 15:12:30 INFO DAGScheduler: Final stage: ResultStage 4 (collect at
Compactor.scala:100)
17/08/30 15:12:30 INFO DAGScheduler: Parents of final stage: List()
17/08/30 15:12:30 INFO DAGScheduler: Missing parents: List()
17/08/30 15:12:30 INFO DAGScheduler: Submitting ResultStage 4
(CarbonMergerRDD[11] at RDD at CarbonRDD.scala:35), which has no missing parents
17/08/30 15:12:30 INFO MemoryStore: Block broadcast_6 stored as values in
memory (estimated size 14.2 KB, free 366.2 MB)
17/08/30 15:12:30 INFO MemoryStore: Block broadcast_6_piece0 stored as bytes in
memory (estimated size 6.7 KB, free 366.2 MB)
17/08/30 15:12:30 INFO BlockManagerInfo: Added broadcast_6_piece0 in memory on
192.168.2.188:41587 (size: 6.7 KB, free: 366.3 MB)
17/08/30 15:12:30 INFO SparkContext: Created broadcast 6 from broadcast at
DAGScheduler.scala:996
17/08/30 15:12:30 INFO DAGScheduler: Submitting 1 missing tasks from
ResultStage 4 (CarbonMergerRDD[11] at RDD at CarbonRDD.scala:35)
17/08/30 15:12:30 INFO TaskSchedulerImpl: Adding task set 4.0 with 1 tasks
17/08/30 15:12:30 INFO TaskSetManager: Starting task 0.0 in stage 4.0 (TID 4,
192.168.2.188, executor 0, partition 0, ANY, 9296 bytes)
17/08/30 15:12:30 INFO BlockManagerInfo: Added broadcast_6_piece0 in memory on
192.168.2.188:38477 (size: 6.7 KB, free: 366.3 MB)
17/08/30 15:12:31 INFO TaskSetManager: Finished task 0.0 in stage 4.0 (TID 4)
in 266 ms on 192.168.2.188 (executor 0) (1/1)
17/08/30 15:12:31 INFO TaskSchedulerImpl: Removed TaskSet 4.0, whose tasks have
all completed, from pool
17/08/30 15:12:31 INFO DAGScheduler: ResultStage 4 (collect at
Compactor.scala:100) finished in 0.267 s
17/08/30 15:12:31 INFO DAGScheduler: Job 4 finished: collect at
Compactor.scala:100, took 0.276120 s
17/08/30 15:12:31 AUDIT Compactor$: [pallavi][hduser][Thread-143]Compaction
request failed for table 30aug_5.uniqdata
17/08/30 15:12:31 ERROR Compactor$: pool-28-thread-1 Compaction request failed
for table 30aug_5.uniqdata
17/08/30 15:12:31 ERROR DataManagementFunc$: pool-24-thread-7 Exception in
compaction thread java.lang.Exception: Compaction Failure in Merger Rdd.
java.util.concurrent.ExecutionException: java.lang.Exception: Compaction
Failure in Merger Rdd.
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at
org.apache.carbondata.spark.rdd.DataManagementFunc$$anonfun$executeCompaction$1.apply(DataManagementFunc.scala:193)
at
org.apache.carbondata.spark.rdd.DataManagementFunc$$anonfun$executeCompaction$1.apply(DataManagementFunc.scala:192)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
at
org.apache.carbondata.spark.rdd.DataManagementFunc$.executeCompaction(DataManagementFunc.scala:192)
at
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$$anon$2.run(CarbonDataRDDFactory.scala:273)
at
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.startCompactionThreads(CarbonDataRDDFactory.scala:364)
at
org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.alterTableForCompaction(CarbonDataRDDFactory.scala:157)
at
org.apache.spark.sql.execution.command.AlterTableCompaction.processData(carbonTableSchema.scala:175)
at
org.apache.spark.sql.execution.command.AlterTableCompaction.run(carbonTableSchema.scala:139)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at
org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
at
org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
at
org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:699)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:220)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.Exception: Compaction Failure in Merger Rdd.
at
org.apache.carbondata.spark.rdd.Compactor$.triggerCompaction(Compactor.scala:143)
at
org.apache.carbondata.spark.rdd.Compactor.triggerCompaction(Compactor.scala)
at
org.apache.carbondata.spark.compaction.CompactionCallable.call(CompactionCallable.java:40)
at
org.apache.carbondata.spark.compaction.CompactionCallable.call(CompactionCallable.java:29)
... 4 more
17/08/30 15:12:31 ERROR CarbonDataRDDFactory$: pool-24-thread-7 Exception in
compaction thread java.lang.Exception: Compaction Failure in Merger Rdd.
17/08/30 15:12:31 INFO HdfsFileLock: pool-24-thread-7 Deleted the lock file
hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock
17/08/30 15:12:31 ERROR CarbonDataRDDFactory$: pool-24-thread-7 Exception in
start compaction thread. Exception in compaction java.lang.Exception:
Compaction Failure in Merger Rdd.
17/08/30 15:12:31 ERROR AbstractDFSCarbonFile: pool-24-thread-7 Exception
occurred:File does not exist:
hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock
17/08/30 15:12:31 ERROR HdfsFileLock: pool-24-thread-7 Not able to delete the
lock file because it is not existed in location
hdfs://localhost:54310/user/hive/warehouse/carbon.store/30aug_5/uniqdata/compaction.lock
17/08/30 15:12:31 ERROR SparkExecuteStatementOperation: Error executing query,
currentState RUNNING,
java.lang.RuntimeException: Compaction failed. Please check logs for more info.
Exception in compaction java.lang.Exception: Compaction Failure in Merger Rdd.
at scala.sys.package$.error(package.scala:27)
at
org.apache.spark.sql.execution.command.AlterTableCompaction.processData(carbonTableSchema.scala:184)
at
org.apache.spark.sql.execution.command.AlterTableCompaction.run(carbonTableSchema.scala:139)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
at
org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
at
org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at
org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
at
org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
at
org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:699)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:220)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:748)
17/08/30 15:12:31 ERROR SparkExecuteStatementOperation: Error running hive
query:
org.apache.hive.service.cli.HiveSQLException: java.lang.RuntimeException:
Compaction failed. Please check logs for more info. Exception in compaction
java.lang.Exception: Compaction Failure in Merger Rdd.
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:258)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:163)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:160)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
at
org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:173)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:748)
> Major and Minor Compaction Failing
> ----------------------------------
>
> Key: CARBONDATA-1422
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1422
> Project: CarbonData
> Issue Type: Bug
> Affects Versions: 1.2.0
> Reporter: Pallavi Singh
> Fix For: 1.2.0
>
>
> The major and minor compaction is failing.
> The Compaction is failing in the default scenario where the table property
> dictionary_include is not specified which is the default behaviour.
> Please find the error logs below :
> 0: jdbc:hive2://localhost:10000> show segments for table uniqdata;
> +--------------------+----------+--------------------------+--------------------------+--+
> | SegmentSequenceId | Status | Load Start Time | Load End
> Time |
> +--------------------+----------+--------------------------+--------------------------+--+
> | 9 | Success | 2017-08-29 11:17:29.927 | 2017-08-29
> 11:17:30.555 |
> | 8 | Success | 2017-08-29 11:17:27.572 | 2017-08-29
> 11:17:28.363 |
> | 7 | Success | 2017-08-29 11:17:23.583 | 2017-08-29
> 11:17:25.112 |
> | 6 | Success | 2017-08-29 11:17:07.966 | 2017-08-29
> 11:17:09.322 |
> | 5 | Success | 2017-08-29 10:38:15.727 | 2017-08-29
> 10:38:16.548 |
> | 4 | Success | 2017-08-29 10:37:13.053 | 2017-08-29
> 10:37:13.888 |
> | 3 | Success | 2017-08-29 10:36:57.851 | 2017-08-29
> 10:36:59.08 |
> | 2 | Success | 2017-08-29 10:36:49.439 | 2017-08-29
> 10:36:50.373 |
> | 1 | Success | 2017-08-29 10:36:37.365 | 2017-08-29
> 10:36:38.768 |
> | 0 | Success | 2017-08-29 10:36:21.011 | 2017-08-29
> 10:36:26.1 |
> +--------------------+----------+--------------------------+--------------------------+--+
> 10 rows selected (0.081 seconds)
> 0: jdbc:hive2://localhost:10000> ALTER TABLE uniqdata COMPACT 'MINOR';
> Error: java.lang.RuntimeException: Compaction failed. Please check logs for
> more info. Exception in compaction java.lang.Exception: Compaction Failure in
> Merger Rdd. (state=,code=0)
> 0: jdbc:hive2://localhost:10000> ALTER TABLE uniqdata COMPACT 'MAJOR';
> Error: java.lang.RuntimeException: Compaction failed. Please check logs for
> more info. Exception in compaction java.lang.Exception: Compaction Failure in
> Merger Rdd. (state=,code=0)
> 0: jdbc:hive2://localhost:10000>
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)