So hadoop.tmp.dir did not work because HBASE added a new parameter in the newest version
hbase.fs.tmp.dir I added this to the kylin.properties file and it now runs through. I will add this comment to the jira if I can On 10/20/15, 2:44 PM, "Benjamin Leonhardi" <[email protected]> wrote: >Hmm although adding hadoop.tmp.dir does not seem to work for me. I added it >everywhere in hadoop (core-site,mapred-site and hbase-site 0 and in the kylin >job config as well. Still same error. I will have a deeper look > > > >On 10/20/15, 1:58 PM, "Benjamin Leonhardi" <[email protected]> wrote: > >>Thanks a lot. I swear I googled it :-) >> >> >> >> >>On 10/20/15, 1:56 PM, "Shi, Shaofeng" <[email protected]> wrote: >> >>>Hi Benjamin, please check if this JIRA matches your problem: >>>https://issues.apache.org/jira/browse/KYLIN-1083 >>> >>>On 10/20/15, 12:53 PM, "Benjamin Leonhardi" <[email protected]> >>>wrote: >>> >>>>Hello Everybody, >>>> >>>>Thanks for all the help/ >>>> >>>>I successfully build Kylin on HDP2.3 with the new updates to the branch >>>>and by changing some dependencies in the pom file ( got a HCAT mapred >>>>version conflict before ) When I changed the entries in the pom file to >>>>the exact version of HDP2.3 it worked. >>>> >>>> >>>> >>>> >>>><!-- Hadoop versions --> >>>> <hadoop2.version>2.7.1</hadoop2.version> >>>> <yarn.version>2.7.1</yarn.version> >>>> <zookeeper.version>3.4.6</zookeeper.version> >>>> <hive.version>1.2.1</hive.version> >>>> <hive-hcatalog.version>1.2.1</hive-hcatalog.version> >>>> <hbase-hadoop2.version>1.1.1</hbase-hadoop2.version> >>>> <curator.version>2.7.1</curator.version> >>>> >>>> >>>>However now I get errors in the test and also when I create a new cube >>>>from scratch during the Hfile creation. >>>> >>>>Any idea what this could be? During the test run I thought it might be a >>>>bad testcase but it also happens when I build a new cube from scratch. >>>> >>>>Parameters: >>>> >>>> -conf /kylin/kylin-1.1-incubating-SNAPSHOT/conf/kylin_job_conf.xml >>>>-cubename aggtest -input >>>>/kylin/kylin_metadata/kylin-0669bcc0-2a42-4189-a7c8-6abbb533da8c/aggtest/c >>>>uboid/* -output >>>>hdfs://sandbox.hortonworks.com:8020/kylin/kylin_metadata/kylin-0669bcc0-2a >>>>42-4189-a7c8-6abbb533da8c/aggtest/hfile -htablename KYLIN_0PV292NH5B >>>>-jobname Kylin_HFile_Generator_aggtest_Step >>>> >>>> >>>>Error message : >>>> >>>> >>>>java.lang.IllegalArgumentException: Can not create a Path from a null >>>>string >>>> at org.apache.hadoop.fs.Path.checkPathArg(Path.java:122) >>>> at org.apache.hadoop.fs.Path.<init>(Path.java:134) >>>> at org.apache.hadoop.fs.Path.<init>(Path.java:88) >>>> at >>>>org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.configurePartitioner( >>>>HFileOutputFormat2.java:591) >>>> at >>>>org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.configureIncrementalL >>>>oad(HFileOutputFormat2.java:440) >>>> at >>>>org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.configureIncrementalL >>>>oad(HFileOutputFormat2.java:405) >>>> at >>>>org.apache.hadoop.hbase.mapreduce.HFileOutputFormat.configureIncrementalLo >>>>ad(HFileOutputFormat.java:91) >>>> at >>>>org.apache.kylin.job.hadoop.cube.CubeHFileJob.run(CubeHFileJob.java:86) >>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>>> at >>>>org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable >>>>.java:113) >>>> at >>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutab >>>>le.java:107) >>>> at >>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChai >>>>nedExecutable.java:51) >>>> at >>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutab >>>>le.java:107) >>>> at >>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Defaul >>>>tScheduler.java:130) >>>> at >>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java: >>>>1145) >>>> at >>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java >>>>:615) >>>> at java.lang.Thread.run(Thread.java:745) >>>> >>>>result code:2 >>>> >>>> >>>> >>>>On 10/9/15, 1:11 PM, "Benjamin Leonhardi" <[email protected]> >>>>wrote: >>>> >>>>>Hello Everybody, >>>>> >>>>>We figured out the 1.2 coded comes from the first dependency. Will try >>>>>and update. >>>>> >>>>>Thanks for your help. >>>>> >>>>>Ben >>>>> >>>>> >>>>> >>>>>On 10/6/15, 12:47 AM, "周千昊" <[email protected]> wrote: >>>>> >>>>>>Hi, Benjamin, >>>>>> It is a bug, a jira ticket is created as follow: >>>>>>https://issues.apache.org/jira/browse/KYLIN-1059 >>>>>> and it should have been fixed. since I don't have the hdp 2.3 env >>>>>>currently, it would be appreciated if you can help to verify with the >>>>>>latest code(branch 1.x-HBase1.x) >>>>>> >>>>>>Luke Han <[email protected]>于2015年10月5日周一 下午8:35写道: >>>>>> >>>>>>> Hi Benjamin, >>>>>>> Did you generate binary package by yourself with below >>>>>>>instruction? or >>>>>>> just compile war/jar from maven? >>>>>>> >>>>>>> http://kylin.incubator.apache.org/development/howto_package.html >>>>>>> <http://kylin.incubator.apache.org/development/howto_package.html> >>>>>>> >>>>>>> Thanks. >>>>>>> >>>>>>> >>>>>>> Best Regards! >>>>>>> --------------------- >>>>>>> >>>>>>> Luke Han >>>>>>> >>>>>>> On Thu, Oct 1, 2015 at 1:30 PM, Benjamin Leonhardi < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>> > Hello All, >>>>>>> > >>>>>>> > I try to get Kylin to run with HDP 2.3 sandbox image. Using the >>>>>>>HBase1.x >>>>>>> > branch. >>>>>>> > >>>>>>> > https://github.com/apache/incubator-kylin/tree/1.x-HBase1.x >>>>>>> > >>>>>>> > It compiles fine and the tests run through until they reach the job >>>>>>> > creation. I followed the steps below. >>>>>>> > >>>>>>> > http://kylin.incubator.apache.org/development/dev_env.html >>>>>>> > >>>>>>> > When I run the tests he reaches the job part and then fails because >>>>>>>he >>>>>>> > misses a method in commons-codec.jar. The method is >>>>>>>encodeBase64(byte[], >>>>>>> > isChunk boolean, isUrlSafe boolean) and it is in commons-codec1.4+ >>>>>>>but >>>>>>> not >>>>>>> > in older versions. I checked but my environment only has codec jars >>>>>>>that >>>>>>> > are newer. >>>>>>> > >>>>>>> > However when I build kylin he for some reason compiled an older >>>>>>>codec >>>>>>> into >>>>>>> > the kylin-job-1.1-incubating-SNAPSHOT-job.jar. If I unzip this I >>>>>>>see a >>>>>>> > Base64 class that only has the method encodeBase64(byte[], isChunk >>>>>>> boolean >>>>>>> > ) but not the urlsafe method. ( so I suppose codec 1.3 or older ) I >>>>>>> tried >>>>>>> > to check the pom.xml because I thought I might find an older codec >>>>>>> > dependency inside but did not find anything. >>>>>>> > >>>>>>> > I also have to say I am not the best Maven person so I might be >>>>>>> > overlooking something easy. One question I would have is why he >>>>>>>compiles >>>>>>> > the code into the -job.jar at all. >>>>>>> > >>>>>>> > >>>>>>> > Any help welcome >>>>>>> > >>>>>>> > Best regards, >>>>>>> > >>>>>>> > Benjamin >>>>>>> > >>>>>>> > >>>>>>> > Starting: >>>>>>> > Kylin_Fact_Distinct_Columns_test_kylin_cube_without_slr_empty_Step >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,166][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> append >>>>>>> > job jar: >>>>>>> > >>>>>>>/root/kylin_build/incubator-kylin/job/../job/target/kylin-job-1.1-incu >>>>>>> > bating-SNAPSHOT-job.jar >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,169][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> append >>>>>>> > kylin.hive.dependency: null and kylin.hbase.dependency: null to >>>>>>> > mapreduce.appli >>>>>>> > cation.classpath >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,167][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> append >>>>>>> > job jar: >>>>>>> > >>>>>>>/root/kylin_build/incubator-kylin/job/../job/target/kylin-job-1.1-incu >>>>>>> > bating-SNAPSHOT-job.jar >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,170][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> append >>>>>>> > kylin.hive.dependency: null and kylin.hbase.dependency: null to >>>>>>> > mapreduce.appli >>>>>>> > cation.classpath >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,170][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> Hadoop >>>>>>> > job classpath is: >>>>>>> > >>>>>>> >>>>>>>/tmp/kylin/*,$HADOOP_CONF_DIR,/usr/hdp/2.3.0.0-2530/hbase/lib/hbase-com >>>>>>>mon.jar,/usr/hdp/current/hive-client/conf/,/usr/hdp/2.3.0.0-2530/hive/l >>>>>>>ib/hive-metastore.jar,/usr/hdp/2.3.0.0-2530/hive/lib/hive-exec.jar,/usr >>>>>>>/hdp/2.3.0.0-2530/hive-hcatalog/share/hcatalog/*,$PWD/mr-framework/hado >>>>>>>op/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapre >>>>>>>duce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-frame >>>>>>>work/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/ha >>>>>>>doop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-fr >>>>>>>amework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoo >>>>>>>p/hdfs/lib/*:/usr/hdp/2.3.0.0-2530/hadoop/lib/hadoop-lzo-0.6.0.2.3.0.0- >>>>>>>2530.jar:/etc/hadoop/conf/secure >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,169][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> Hadoop >>>>>>> > job classpath is: >>>>>>> > >>>>>>> >>>>>>>/tmp/kylin/*,$HADOOP_CONF_DIR,/usr/hdp/2.3.0.0-2530/hbase/lib/hbase-com >>>>>>>mon.jar,/usr/hdp/current/hive-client/conf/,/usr/hdp/2.3.0.0-2530/hive/l >>>>>>>ib/hive-metastore.jar,/usr/hdp/2.3.0.0-2530/hive/lib/hive-exec.jar,/usr >>>>>>>/hdp/2.3.0.0-2530/hive-hcatalog/share/hcatalog/*,$PWD/mr-framework/hado >>>>>>>op/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapre >>>>>>>duce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-frame >>>>>>>work/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/ha >>>>>>>doop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-fr >>>>>>>amework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoo >>>>>>>p/hdfs/lib/*:/usr/hdp/2.3.0.0-2530/hadoop/lib/hadoop-lzo-0.6.0.2.3.0.0- >>>>>>>2530.jar:/etc/hadoop/conf/secure >>>>>>> > L4J [2015-10-01 >>>>>>>03:51:24,229][WARN][org.apache.hadoop.hive.conf.HiveConf] >>>>>>> > - HiveConf of name hive.heapsize does not exist >>>>>>> > L4J [2015-10-01 >>>>>>>03:51:24,230][WARN][org.apache.hadoop.hive.conf.HiveConf] >>>>>>> > - HiveConf of name hive.server2.enable.impersonation does not exist >>>>>>> > L4J [2015-10-01 >>>>>>>03:51:24,236][WARN][org.apache.hadoop.hive.conf.HiveConf] >>>>>>> > - HiveConf of name hive.heapsize does not exist >>>>>>> > L4J [2015-10-01 >>>>>>>03:51:24,237][WARN][org.apache.hadoop.hive.conf.HiveConf] >>>>>>> > - HiveConf of name hive.server2.enable.impersonation does not exist >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,265][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> > tempMetaFileString is : null >>>>>>> > L4J [2015-10-01 >>>>>>> > 03:51:24,266][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - >>>>>>> > tempMetaFileString is : null >>>>>>> > L4J [2015-10-01 >>>>>>> > >>>>>>>03:51:24,267][ERROR][org.apache.kylin.job.execution.AbstractExecutable] >>>>>>> - >>>>>>> > error running Executable >>>>>>> > java.lang.NoSuchMethodError: >>>>>>> > org.apache.commons.codec.binary.Base64.encodeBase64([BZZ)[B >>>>>>> > at >>>>>>> > >>>>>>>org.apache.hive.hcatalog.common.HCatUtil.encodeBytes(HCatUtil.java:125) >>>>>>> > at >>>>>>> > >>>>>>>org.apache.hive.hcatalog.common.HCatUtil.serialize(HCatUtil.java:104) >>>>>>> > at >>>>>>> > >>>>>>>org.apache.hive.hcatalog.common.HCatUtil.getHiveConf(HCatUtil.java:585) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.InitializeInput.getInputJobInfo(Init >>>>>>>ializeInput.java:100) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.InitializeInput.setInput(InitializeI >>>>>>>nput.java:86) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFo >>>>>>>rmat.java:95) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFo >>>>>>>rmat.java:51) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.setupMapper(Fac >>>>>>>tDistinctColumnsJob.java:101) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinc >>>>>>>tColumnsJob.java:77) >>>>>>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>>>>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecuta >>>>>>>ble.java:113) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecu >>>>>>>table.java:107) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultC >>>>>>>hainedExecutable.java:51) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecu >>>>>>>table.java:107) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Def >>>>>>>aultScheduler.java:130) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja >>>>>>>va:1145) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j >>>>>>>ava:615) >>>>>>> > at java.lang.Thread.run(Thread.java:745) >>>>>>> > L4J [2015-10-01 >>>>>>> > >>>>>>>03:51:24,267][ERROR][org.apache.kylin.job.execution.AbstractExecutable] >>>>>>> - >>>>>>> > error running Executable >>>>>>> > java.lang.NoSuchMethodError: >>>>>>> > org.apache.commons.codec.binary.Base64.encodeBase64([BZZ)[B >>>>>>> > at >>>>>>> > >>>>>>>org.apache.hive.hcatalog.common.HCatUtil.encodeBytes(HCatUtil.java:125) >>>>>>> > at >>>>>>> > >>>>>>>org.apache.hive.hcatalog.common.HCatUtil.serialize(HCatUtil.java:104) >>>>>>> > at >>>>>>> > >>>>>>>org.apache.hive.hcatalog.common.HCatUtil.getHiveConf(HCatUtil.java:585) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.InitializeInput.getInputJobInfo(Init >>>>>>>ializeInput.java:100) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.InitializeInput.setInput(InitializeI >>>>>>>nput.java:86) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFo >>>>>>>rmat.java:95) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFo >>>>>>>rmat.java:51) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.setupMapper(Fac >>>>>>>tDistinctColumnsJob.java:101) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinc >>>>>>>tColumnsJob.java:77) >>>>>>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>>>>> > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecuta >>>>>>>ble.java:113) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecu >>>>>>>table.java:107) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultC >>>>>>>hainedExecutable.java:51) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecu >>>>>>>table.java:107) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(Def >>>>>>>aultScheduler.java:130) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja >>>>>>>va:1145) >>>>>>> > at >>>>>>> > >>>>>>> >>>>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j >>>>>>>ava:615) >>>>>>> > >>>>>>> > at java.lang.Thread.run(Thread.java:745) >>>>>>> > >>>>>>> >>> >>>
