Try to add your cluster's core-site.xml, yarn-site.xml, and hdfs-site.xml to the CLASSPATH (and on SPARK_CLASSPATH) and submit the job.
Thanks Best Regards On Thu, Nov 27, 2014 at 12:24 PM, Naveen Kumar Pokala < npok...@spcapitaliq.com> wrote: > Code is in my windows machine and cluster is in some other network in > UNIX. In this case how it will identify the cluster. In case of spark > cluster we can clearly specify the URL like spark://ip:port. But in case of > hadoop how to specify that. > > > > What I have done is copied the hadoop configuration files from network to > local and created dummy hadoop directory(in windows machine). > > > > Submitted from spark submit by adding above dummy files location with > HADOOP_CONF_DIR variable. Attaching the error. > > > > > > > > Please suggest me how to proceed from the code and how to execute from > spark submit from windows machine. > > > > Please provide me sample code if you have any. > > > > -Naveen > > > > *From:* Akhil Das [mailto:ak...@sigmoidanalytics.com] > *Sent:* Wednesday, November 26, 2014 10:03 PM > *To:* Naveen Kumar Pokala > *Cc:* user@spark.apache.org > *Subject:* Re: Spark Job submit > > > > How about? > > > > - Create a SparkContext > > - setMaster as *yarn-cluster* > > - Create a JavaSparkContext with the above SparkContext > > > > And that will submit it to the yarn cluster. > > > Thanks > > Best Regards > > > > On Wed, Nov 26, 2014 at 4:20 PM, Naveen Kumar Pokala < > npok...@spcapitaliq.com> wrote: > > Hi. > > > > Is there a way to submit spark job on Hadoop-YARN cluster from java code. > > > > -Naveen > > >