Hi, there, 

I think the root cause is you got some old hadoop or yarn version jars in your 
environment. 

Which step does you go when you issue cube build job? You may want to check the 
yarn history server

for more log information. 

Thanks,
Sun.


[email protected]
 
From: Mohit Bhatnagar
Date: 2015-06-02 14:17
To: [email protected]
CC: Sakshi Gupta
Subject: RE: Building cube issue: IncompatibleClassChangeError error
Can anyone help me out with this?  
 
-----Original Message-----
From: Mohit Bhatnagar [mailto:[email protected]] 
Sent: Monday, June 01, 2015 5:06 PM
To: [email protected]
Cc: Sakshi Gupta
Subject: Building cube issue: IncompatibleClassChangeError error
 
Hi,
 
I m hitting the below IncompatibleClassChangeError  error. We have the 
following versions already installed in our Prod cluster and I have also 
changed the kylin/pom.xml file to have the same versions.
Here are the details:
  <!-- Hadoop versions -->
        <hadoop2.version>2.5.0</hadoop2.version>
        <yarn.version>2.5.0</yarn.version>
        <zookeeper.version>3.4.5</zookeeper.version>
        <hive.version>0.13.1</hive.version>
        <hive-hcatalog.version>0.13.1</hive-hcatalog.version>
        <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version>
 
From Hadoop 2.0 onwards JobContext is an interface. Don't know why it is 
expecting a class:
Here is the log:
 
 
[pool-6-thread-1]:[2015-06-01 
16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultScheduler$FetcherRunner.run(DefaultScheduler.java:117)]
 - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others
[pool-7-thread-2]:[2015-06-01 
16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:441)]
 - Cleaning up the staging area /user/mobhatna/.staging/job_1432706588203_13062
[pool-7-thread-2]:[2015-06-01 
16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)]
 - error running Executable
java.lang.IncompatibleClassChangeError: Found interface 
org.apache.hadoop.mapreduce.JobContext, but class was expected
       at 
org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at 
org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at 
org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at 
org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
[pool-7-thread-2]:[2015-06-01 
16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)]
 - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 
(Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 
16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:171)]
 - Saving resource /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 
(Store kylin_metadata@hbase)
[pool-7-thread-2]:[2015-06-01 
16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.updateJobOutput(ExecutableManager.java:222)]
 - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR
[pool-7-thread-2]:[2015-06-01 
16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:109)]
 - error running Executable
org.apache.kylin.job.exception.ExecuteException: 
java.lang.IncompatibleClassChangeError: Found interface 
org.apache.hadoop.mapreduce.JobContext, but class was expected
       at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111)
       at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50)
       at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
       at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132)
       at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
       at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
       at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IncompatibleClassChangeError: Found interface 
org.apache.hadoop.mapreduce.JobContext, but class was expected
       at 
org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102)
       at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493)
       at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510)
       at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
       at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
       at java.security.AccessController.doPrivileged(Native Method)
       at javax.security.auth.Subject.doAs(Subject.java:415)
       at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
       at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
       at 
org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123)
       at 
org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
       at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
       at 
org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112)
       at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)

Reply via email to