Dev env... with the default pom, we use hdp 2.2.x as sandbox. That works. On Thu, Jun 4, 2015 at 4:10 PM, Mohit Bhatnagar <[email protected]> wrote:
> Hi Yang, > > While installing I followed "Off Hadoop CLI Installation" tutorial to set > up the development environment. So I have Kylin server on my Windows > machine pointing to my Dev hadoop cluster. > I cross checked the versions which are installed on my hadoop cluster and > made the same changes to Kylin/pom.xml. Even though it downloads the same > version of Hadoop hive hbase jars in my local repo as mentioned below, I am > still hitting the issue. I am not sure from where it is referencing the old > Hadoop versions. > > Please note that when I build Kylin sourcecode on my local machine (which > has kylin server running and pointing to Hadoop cluster), it downloads > Hadoop hive hbase jars on my local repo. > > Thanks, > Mohit > > > > > -----Original Message----- > From: Li Yang [mailto:[email protected]] > Sent: Thursday, June 04, 2015 12:16 PM > To: [email protected] > Subject: Re: RE: Building cube issue: IncompatibleClassChangeError error > > Aiming to be compatible with all Hadoop 2.x versions, Kylin does not ship > any Hadoop/Hive jar. There must be multiple versions of Hadoop/Hive in your > cluster. > > From the stack trace, my guess is the Hive version of the Kylin host and > the Hive server might be different... not sure however. > > On Tue, Jun 2, 2015 at 2:29 PM, [email protected] <[email protected]> wrote: > > > Hi, there, > > > > I think the root cause is you got some old hadoop or yarn version jars > > in your environment. > > > > Which step does you go when you issue cube build job? You may want to > > check the yarn history server > > > > for more log information. > > > > Thanks, > > Sun. > > > > > > [email protected] > > > > From: Mohit Bhatnagar > > Date: 2015-06-02 14:17 > > To: [email protected] > > CC: Sakshi Gupta > > Subject: RE: Building cube issue: IncompatibleClassChangeError error > > Can anyone help me out with this? > > > > -----Original Message----- > > From: Mohit Bhatnagar [mailto:[email protected]] > > Sent: Monday, June 01, 2015 5:06 PM > > To: [email protected] > > Cc: Sakshi Gupta > > Subject: Building cube issue: IncompatibleClassChangeError error > > > > Hi, > > > > I m hitting the below IncompatibleClassChangeError error. We have the > > following versions already installed in our Prod cluster and I have > > also changed the kylin/pom.xml file to have the same versions. > > Here are the details: > > <!-- Hadoop versions --> > > <hadoop2.version>2.5.0</hadoop2.version> > > <yarn.version>2.5.0</yarn.version> > > <zookeeper.version>3.4.5</zookeeper.version> > > <hive.version>0.13.1</hive.version> > > <hive-hcatalog.version>0.13.1</hive-hcatalog.version> > > <hbase-hadoop2.version>0.98.6-hadoop2</hbase-hadoop2.version> > > > > From Hadoop 2.0 onwards JobContext is an interface. Don't know why it > > is expecting a class: > > Here is the log: > > > > > > [pool-6-thread-1]:[2015-06-01 > > 16:56:53,644][INFO][org.apache.kylin.job.impl.threadpool.DefaultSchedu > > ler$FetcherRunner.run(DefaultScheduler.java:117)] > > - Job Fetcher: 0 running, 1 actual running, 1 ready, 44 others > > [pool-7-thread-2]:[2015-06-01 > > 16:57:48,749][INFO][org.apache.hadoop.mapreduce.JobSubmitter.submitJob > > Internal(JobSubmitter.java:441)] > > - Cleaning up the staging area > > /user/mobhatna/.staging/job_1432706588203_13062 > > [pool-7-thread-2]:[2015-06-01 > > 16:57:49,031][ERROR][org.apache.kylin.job.execution.AbstractExecutable > > .execute(AbstractExecutable.java:109)] > > - error running Executable > > java.lang.IncompatibleClassChangeError: Found interface > > org.apache.hadoop.mapreduce.JobContext, but class was expected > > at > > > org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394) > > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285) > > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:415) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614) > > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282) > > at > > > org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123) > > at > > > org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) > > at > > > org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112) > > at > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > at > > > org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50) > > at > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > at > > > org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > [pool-7-thread-2]:[2015-06-01 > > 16:57:52,178][DEBUG][org.apache.kylin.common.persistence.ResourceStore > > .putResource(ResourceStore.java:171)] > > - Saving resource > > /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 > > (Store kylin_metadata@hbase) > > [pool-7-thread-2]:[2015-06-01 > > 16:57:54,016][DEBUG][org.apache.kylin.common.persistence.ResourceStore > > .putResource(ResourceStore.java:171)] > > - Saving resource > > /execute_output/81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 > > (Store kylin_metadata@hbase) > > [pool-7-thread-2]:[2015-06-01 > > 16:57:54,282][INFO][org.apache.kylin.job.manager.ExecutableManager.upd > > ateJobOutput(ExecutableManager.java:222)] > > - job id:81a626bd-60a7-4993-9d3e-ca3013e76fb9-01 from RUNNING to ERROR > > [pool-7-thread-2]:[2015-06-01 > > 16:57:54,282][ERROR][org.apache.kylin.job.execution.AbstractExecutable > > .execute(AbstractExecutable.java:109)] > > - error running Executable > > org.apache.kylin.job.exception.ExecuteException: > > java.lang.IncompatibleClassChangeError: Found interface > > org.apache.hadoop.mapreduce.JobContext, but class was expected > > at > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:111) > > at > > > org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:50) > > at > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > at > > > org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:132) > > at > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > at > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: java.lang.IncompatibleClassChangeError: Found interface > > org.apache.hadoop.mapreduce.JobContext, but class was expected > > at > > > org.apache.hive.hcatalog.mapreduce.HCatBaseInputFormat.getSplits(HCatBaseInputFormat.java:102) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:493) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:510) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:394) > > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285) > > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:415) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614) > > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282) > > at > > > org.apache.kylin.job.hadoop.AbstractHadoopJob.waitForCompletion(AbstractHadoopJob.java:123) > > at > > > org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:80) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) > > at > > > org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:112) > > at > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExec > > utable.java:107) > > >
