Hi, there, I think the root cause is you got some old hadoop or yarn version jars in your environment.
Which step does you go when you issue cube build job? You may want to check the yarn history server for more log information. Thanks, Sun. [email protected] From: Mohit Bhatnagar Date: 2015-06-02 14:17 To: [email protected] CC: Sakshi Gupta Subject: RE: Building cube issue: IncompatibleClassChangeError error Can anyone help me out with this? -----Original Message----- From: Mohit Bhatnagar [mailto:[email protected]] Sent: Monday, June 01, 2015 5:06 PM To: [email protected] Cc: Sakshi Gupta Subject: Building cube issue: IncompatibleClassChangeError error Hi, I m hitting the below IncompatibleClassChangeError error. We have the following versions already installed in our Prod cluster and I have also changed the kylin/pom.xml file to have the same versions. Here are the details: <!-- Hadoop versions --> <hadoop2.version>2.5.0</hadoop2.version> <yarn.version>2.5.0</yarn.version> <zookeeper.version>3.4.5</zookeeper.version> <hive.version>0.13.1</hive.version> <hive-hcatalog.version>0.13.1</hive-hcatalog.version> <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version> From Hadoop 2.0 onwards JobContext is an interface. Don't know why it is expecting a class: Here is the log: [pool-6-thread-1]:[2015-06-01 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultScheduler$FetcherRunner.run(DefaultScheduler.java:117)] - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others [pool-7-thread-2]:[2015-06-01 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:441)] - Cleaning up the staging area /user/mobhatna/.staging/job_1432706588203_13062 [pool-7-thread-2]:[2015-06-01 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282) at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123) at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) [pool-7-thread-2]:[2015-06-01 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase) [pool-7-thread-2]:[2015-06-01 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)] - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 (Store kylin_metadata@hbase) [pool-7-thread-2]:[2015-06-01 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:222)] - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR [pool-7-thread-2]:[2015-06-01 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)] - error running Executable org.apache.kylin.job.exception.ExecuteException: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111) at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) at org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) Caused by: java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected at org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282) at org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123) at org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) at org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112) at org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
