fengYu created KYLIN-1021:
-----------------------------

             Summary: upload dependent jars of kylin to HDFS and set tmpjars
                 Key: KYLIN-1021
                 URL: https://issues.apache.org/jira/browse/KYLIN-1021
             Project: Kylin
          Issue Type: Improvement
    Affects Versions: v1.0
            Reporter: fengYu


As [~Shaofengshi] says in maillist : Regrading your question about the jar 
files located in local disk instead of HDFS, yes the hadoop/hive/hbase jars 
should exist in local disk on each machine of the hadoop cluster, with the same 
locations; Kylin will not upload those jars; Please check and ensure the 
consistency of your hadoop cluster.

However, our hadoop cluster is managed by hadoop administrator, we have no 
right to login those machine, even though we have the right, copy all files to 
hundreds of machine will be a painful job(I do not know is there some tools can 
do this well).

However, I can not get any tips about you measure(If you has the document, tell 
me)...

I change my source code and create a directory in kylin tmp 
directory(kylin.hdfs.working.dir/kylin_metadata) and upload all jars to the 
directory if the directory is empty(it only happened at the first time) when 
submitting a mapreduce job, then set those locations to tmpjars of the 
mapreduce job(just like kylin set tmpfiles before submit job), This is 
automated and make kylin deploying easier..



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to