Hi Benjamin,
    Did you generate binary package by yourself with below instruction? or
just compile war/jar from maven?

    http://kylin.incubator.apache.org/development/howto_package.html
<http://kylin.incubator.apache.org/development/howto_package.html>

    Thanks.


Best Regards!
---------------------

Luke Han

On Thu, Oct 1, 2015 at 1:30 PM, Benjamin Leonhardi <
[email protected]> wrote:

> Hello All,
>
> I try to get Kylin to run with HDP 2.3 sandbox image. Using the HBase1.x
> branch.
>
> https://github.com/apache/incubator-kylin/tree/1.x-HBase1.x
>
> It compiles fine and the tests run through until they reach the job
> creation. I followed the steps below.
>
> http://kylin.incubator.apache.org/development/dev_env.html
>
> When I run the tests he reaches the job part and then fails because he
> misses a method in commons-codec.jar. The method is encodeBase64(byte[],
> isChunk boolean, isUrlSafe boolean) and it is in commons-codec1.4+ but not
> in older versions. I checked but my environment only has codec jars that
> are newer.
>
> However when I build kylin he for some reason compiled an older codec into
> the kylin-job-1.1-incubating-SNAPSHOT-job.jar. If I unzip this I see a
> Base64 class that only has the method encodeBase64(byte[], isChunk boolean
> ) but not the urlsafe method. ( so I suppose codec 1.3 or older )  I tried
> to check the pom.xml because I thought I might find an older codec
> dependency inside but did not find anything.
>
> I also have to say I am not the best Maven person so I might be
> overlooking something easy. One question I would have is why he compiles
> the code into the -job.jar at all.
>
>
> Any help welcome
>
> Best regards,
>
> Benjamin
>
>
> Starting:
> Kylin_Fact_Distinct_Columns_test_kylin_cube_without_slr_empty_Step
> L4J [2015-10-01
> 03:51:24,166][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - append
> job jar:
> /root/kylin_build/incubator-kylin/job/../job/target/kylin-job-1.1-incu
> bating-SNAPSHOT-job.jar
> L4J [2015-10-01
> 03:51:24,169][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - append
> kylin.hive.dependency: null and kylin.hbase.dependency: null to
> mapreduce.appli
> cation.classpath
> L4J [2015-10-01
> 03:51:24,167][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - append
> job jar:
> /root/kylin_build/incubator-kylin/job/../job/target/kylin-job-1.1-incu
> bating-SNAPSHOT-job.jar
> L4J [2015-10-01
> 03:51:24,170][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - append
> kylin.hive.dependency: null and kylin.hbase.dependency: null to
> mapreduce.appli
> cation.classpath
> L4J [2015-10-01
> 03:51:24,170][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - Hadoop
> job classpath is:
> /tmp/kylin/*,$HADOOP_CONF_DIR,/usr/hdp/2.3.0.0-2530/hbase/lib/hbase-common.jar,/usr/hdp/current/hive-client/conf/,/usr/hdp/2.3.0.0-2530/hive/lib/hive-metastore.jar,/usr/hdp/2.3.0.0-2530/hive/lib/hive-exec.jar,/usr/hdp/2.3.0.0-2530/hive-hcatalog/share/hcatalog/*,$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:/usr/hdp/2.3.0.0-2530/hadoop/lib/hadoop-lzo-0.6.0.2.3.0.0-2530.jar:/etc/hadoop/conf/secure
> L4J [2015-10-01
> 03:51:24,169][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] - Hadoop
> job classpath is:
> /tmp/kylin/*,$HADOOP_CONF_DIR,/usr/hdp/2.3.0.0-2530/hbase/lib/hbase-common.jar,/usr/hdp/current/hive-client/conf/,/usr/hdp/2.3.0.0-2530/hive/lib/hive-metastore.jar,/usr/hdp/2.3.0.0-2530/hive/lib/hive-exec.jar,/usr/hdp/2.3.0.0-2530/hive-hcatalog/share/hcatalog/*,$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:/usr/hdp/2.3.0.0-2530/hadoop/lib/hadoop-lzo-0.6.0.2.3.0.0-2530.jar:/etc/hadoop/conf/secure
> L4J [2015-10-01 03:51:24,229][WARN][org.apache.hadoop.hive.conf.HiveConf]
> - HiveConf of name hive.heapsize does not exist
> L4J [2015-10-01 03:51:24,230][WARN][org.apache.hadoop.hive.conf.HiveConf]
> - HiveConf of name hive.server2.enable.impersonation does not exist
> L4J [2015-10-01 03:51:24,236][WARN][org.apache.hadoop.hive.conf.HiveConf]
> - HiveConf of name hive.heapsize does not exist
> L4J [2015-10-01 03:51:24,237][WARN][org.apache.hadoop.hive.conf.HiveConf]
> - HiveConf of name hive.server2.enable.impersonation does not exist
> L4J [2015-10-01
> 03:51:24,265][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] -
> tempMetaFileString is : null
> L4J [2015-10-01
> 03:51:24,266][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob] -
> tempMetaFileString is : null
> L4J [2015-10-01
> 03:51:24,267][ERROR][org.apache.kylin.job.execution.AbstractExecutable] -
> error running Executable
> java.lang.NoSuchMethodError:
> org.apache.commons.codec.binary.Base64.encodeBase64([BZZ)[B
>         at
> org.apache.hive.hcatalog.common.HCatUtil.encodeBytes(HCatUtil.java:125)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.serialize(HCatUtil.java:104)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.getHiveConf(HCatUtil.java:585)
>         at
> org.apache.hive.hcatalog.mapreduce.InitializeInput.getInputJobInfo(InitializeInput.java:100)
>         at
> org.apache.hive.hcatalog.mapreduce.InitializeInput.setInput(InitializeInput.java:86)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:95)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:51)
>         at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.setupMapper(FactDistinctColumnsJob.java:101)
>         at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:77)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>         at
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:113)
>         at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>         at
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51)
>         at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>         at
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> L4J [2015-10-01
> 03:51:24,267][ERROR][org.apache.kylin.job.execution.AbstractExecutable] -
> error running Executable
> java.lang.NoSuchMethodError:
> org.apache.commons.codec.binary.Base64.encodeBase64([BZZ)[B
>         at
> org.apache.hive.hcatalog.common.HCatUtil.encodeBytes(HCatUtil.java:125)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.serialize(HCatUtil.java:104)
>         at
> org.apache.hive.hcatalog.common.HCatUtil.getHiveConf(HCatUtil.java:585)
>         at
> org.apache.hive.hcatalog.mapreduce.InitializeInput.getInputJobInfo(InitializeInput.java:100)
>         at
> org.apache.hive.hcatalog.mapreduce.InitializeInput.setInput(InitializeInput.java:86)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:95)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:51)
>         at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.setupMapper(FactDistinctColumnsJob.java:101)
>         at
> org.apache.kylin.job.hadoop.cube.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:77)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>         at
> org.apache.kylin.job.common.MapReduceExecutable.doWork(MapReduceExecutable.java:113)
>         at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>         at
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51)
>         at
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>         at
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>
>         at java.lang.Thread.run(Thread.java:745)
>

Reply via email to