Hi Feng, thanks for pointing out that the hive jar repeated in the job
configuration, I just created a JIRA to fix it:
https://issues.apache.org/jira/browse/KYLIN-1015

Regrading your question about the jar files located in local disk instead
of HDFS, yes the hadoop/hive/hbase jars should exist in local disk on each
machine of the hadoop cluster, with the same locations; Kylin will not
upload those jars; Please check and ensure the consistency of your hadoop
cluster. 

On 9/8/15, 10:16 AM, "yu feng" <[email protected]> wrote:

>I find those log like this :
>[pool-5-thread-1]:[2015-09-07
>20:58:41,746][INFO][org.apache.kylin.job.hadoop.AbstractHadoopJob.setJobCl
>asspath(AbstractHadoopJob.java:166)]
>- Hadoop job classpath is:
>/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/etc/hadoop:/home/nrpt/hadoop_hiv
>e_hbase/hadoop-2.2.0/share/hadoop/common/lib/*:/home/nrpt/hadoop_hive_hbas
>e/hadoop-2.2.0/share/hadoop/common/*:/home/nrpt/hadoop_hive_hbase/hadoop-2
>.2.0/share/hadoop/hdfs/lib/*:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/sha
>re/hadoop/hdfs/*:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/ya
>rn/lib/*:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/*:/ho
>me/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/*:/home/
>nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/*:/home/nrpt/ha
>doop_hive_hbase/hadoop-2.2.0/contrib/capacity-scheduler/*.jar:/home/nrpt/h
>adoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/hadoop-mapreduce-clie
>nt-app-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/ma
>preduce/hadoop-mapreduce-client-common-2.2.0.jar:/home/nrpt/hadoop_hive_hb
>ase/hadoop-2.2.0/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.2.0
>.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/hado
>op-mapreduce-client-hs-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0
>/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.2.0.jar:/home
>/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/hadoop-mapredu
>ce-client-jobclient-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/sh
>are/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.2.0-tests.jar:/ho
>me/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/hadoop-mapre
>duce-client-shuffle-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/sh
>are/hadoop/mapreduce/hadoop-mapreduce-examples-2.2.0.jar:/home/nrpt/hadoop
>_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/hadoop-yarn-api-2.2.0.jar:/home
>/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/hadoop-yarn-applica
>tions-distributedshell-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0
>/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.2.0.ja
>r:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/hadoop-yarn-
>client-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/ya
>rn/hadoop-yarn-common-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/
>share/hadoop/yarn/hadoop-yarn-server-common-2.2.0.jar:/home/nrpt/hadoop_hi
>ve_hbase/hadoop-2.2.0/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.2
>.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/hadoop-
>yarn-server-resourcemanager-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-
>2.2.0/share/hadoop/yarn/hadoop-yarn-server-tests-2.2.0.jar:/home/nrpt/hado
>op_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/hadoop-yarn-server-web-proxy-
>2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/yarn/hado
>op-yarn-site-2.2.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/had
>oop/mapreduce/lib/aopalliance-1.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-
>2.2.0/share/hadoop/mapreduce/lib/asm-3.2.jar:/home/nrpt/hadoop_hive_hbase/
>hadoop-2.2.0/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/home/nrpt/hadoop_h
>ive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/commons-compress-1.4.1.j
>ar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/co
>mmons-io-2.1.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/ma
>preduce/lib/guice-3.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/
>hadoop/mapreduce/lib/guice-servlet-3.0.jar:/home/nrpt/hadoop_hive_hbase/ha
>doop-2.2.0/share/hadoop/mapreduce/lib/hadoop-annotations-2.2.0.jar:/home/n
>rpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/hadoop-lzo-0
>.4.20.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce
>/lib/hamcrest-core-1.1.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share
>/hadoop/mapreduce/lib/jackson-core-asl-1.8.8.jar:/home/nrpt/hadoop_hive_hb
>ase/hadoop-2.2.0/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/
>home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/javax.
>inject-1.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapred
>uce/lib/jersey-core-1.9.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/shar
>e/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/home/nrpt/hadoop_hive_hbase/h
>adoop-2.2.0/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/home/nrpt/ha
>doop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/junit-4.10.jar:/ho
>me/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/log4j-1.
>2.17.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/
>lib/netty-3.6.2.Final.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/
>hadoop/mapreduce/lib/paranamer-2.3.jar:/home/nrpt/hadoop_hive_hbase/hadoop
>-2.2.0/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/home/nrpt/hadoo
>p_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.j
>ar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/share/hadoop/mapreduce/lib/xz
>-1.0.jar:/home/nrpt/hadoop_hive_hbase/hadoop-2.2.0/modules/*.jar
>.....................
>
>
>There are two line in this log . the first line is the hadoop default
>classpath outputed by command 'mapred classpath' , and it ends with a new
>line '\n', I think the character need to be deleted.
>what's more , I check source code of setting job classpath :
>
>        if (kylinHBaseDependency != null) {
>            // yarn classpath is comma separated
>            kylinHBaseDependency = kylinHBaseDependency.replace(":", ",");
>            classpath = classpath + "," + kylinHBaseDependency;
>        }
>
>        if (kylinHiveDependency != null) {
>            // yarn classpath is comma separated
>            kylinHiveDependency = kylinHiveDependency.replace(":", ",");
>            classpath = classpath + "," + kylinHiveDependency;
>        }
>
>        jobConf.set(MAP_REDUCE_CLASSPATH, classpath + "," +
>kylinHiveDependency);
>        logger.info("Hadoop job classpath is: " +
>job.getConfiguration().get(MAP_REDUCE_CLASSPATH));
>
>it looks like we append kylinHiveDependency to classpath twice, I do not
>know what it means..
>
>Lastly, the property of
>MAP_REDUCE_CLASSPATH(mapreduce.application.classpath) is setted up  some
>jar files located at local filesystem, Actually I do not know if those
>files will be uploaded to HDFS just as the files added in property
>'tmpjars
>'.
>
>2015-09-08 9:35 GMT+08:00 Shi, Shaofeng <[email protected]>:
>
>> Hi feng, your map reduce classpath might not be correctly configured;
>> Kylin will read the ³mapreduce.application.classpath² from default job
>> configuration; if not found that, it will run ³mapred classpath² command
>> to get the classpath, and then append hive/hbase dependencies; Please
>> check kylin.log to see whether the final classpath includes the jar for
>> this missing class;
>>
>> The message in kylin.log is as below, you can search it:
>>
>> Hadoop job classpath is:
>>
>>
>> On 9/7/15, 10:18 PM, "yu feng" <[email protected]> wrote:
>>
>> >After submit a mapreduce job, we get job status,But it tells the job
>> >failed!  we check this application on RM
>>website(xxx:8088/cluster/apps),
>> >we
>> >find those log :
>> >
>> > Application application_1418904565842_3597024 failed 2 times due to AM
>> >Container for appattempt_1418904565842_3597024_000002 exited with
>> >exitCode:
>> >1 due to: Exception from container-launch:
>> >org.apache.hadoop.util.Shell$ExitCodeException:
>> >at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
>> >at org.apache.hadoop.util.Shell.run(Shell.java:379)
>> >at
>> 
>>>org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589
>>>)
>> >at
>> 
>>>org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor.launchC
>>>on
>> >tainer(LinuxContainerExecutor.java:252)
>> >at
>> 
>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Cont
>>>ai
>> >nerLaunch.call(ContainerLaunch.java:283)
>> >at
>> 
>>>org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Cont
>>>ai
>> >nerLaunch.call(ContainerLaunch.java:79)
>> >at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>> >at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>> >at
>> 
>>>java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecuto
>>>r.
>> >java:895)
>> >at
>> 
>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.ja
>>>va
>> >:918)
>> >at java.lang.Thread.run(Thread.java:662)
>> >main : command provided 1
>> >
>> >we check yarn log with this command : yarn logs -applicationId
>> >application_1418904565842_3597024,get those log :
>> >Container: container_1418904565842_3597024_01_000001 on
>> >hadoop88.photo.163.org_56708
>> 
>>>========================================================================
>>>==
>> >============
>> >LogType: stderr
>> >LogLength: 664
>> >Log Contents:
>> >Exception in thread "main" java.lang.NoClassDefFoundError:
>> >org/apache/hadoop/mapreduce/v2/app/MRAppMaster
>> >Caused by: java.lang.ClassNotFoundException:
>> >org.apache.hadoop.mapreduce.v2.app.MRAppMaster
>> >at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
>> >at java.security.AccessController.doPrivileged(Native Method)
>> >at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
>> >at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
>> >at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
>> >at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
>> >Could not find the main class:
>> >org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.
>> >
>> >
>> >I thinks this means the task can not find the jar files, So I upload
>>all
>> >the jars that kylin dependent to HDFS, and before submit this job*(in
>> >AbstractHadoopJob.attachKylinPropsAndMetadata function )* I set
>>"tmpjars"
>> >to those files located on HDFS(this way can avoid uploading all files
>>when
>> >submit every mapreduce job).
>> >
>> >This measure works in kylin-0.7.2, I get the same error in kylin-1.0
>>and I
>> >guess this measure will work in kylin-1.0 too, But I do not think this
>>is
>> >a
>> >good idea,
>> >
>> >It will be highly appreciated if you have some good idea or some
>> >suggestion. Thanks...
>>
>>

Reply via email to