I’m not sure if I understand your question correctly do you mean you didn’t see the application information in Spark Web UI even it generates the expected results?
Best, -- Nan Zhu On Thursday, February 20, 2014 at 10:13 AM, Tao Xiao wrote: > My application source file, SimpleDistributedApp.scala, is as follows: > > __________________________________________________________________ > import org.apache.spark.{SparkConf, SparkContext} > > object SimpleDistributedApp { > def main(args: Array[String]) = { > val filepath = > "hdfs://hadoop-1.certus.com:54310/user/root/samples/data > (http://hadoop-1.certus.com:54310/user/root/samples/data)" > > val conf = new SparkConf() > .setMaster("spark://hadoop-1.certus.com:7077 > (http://hadoop-1.certus.com:7077)") > .setAppName("**SimpleDistributedApp**") > > .setSparkHome("/home/xt/soft/spark-0.9.0-incubating-bin-hadoop1") > > .setJars(Array("target/scala-2.10/simple-distributed-app_2.10-1.0.jar")) > .set("spark.executor.memory", "1g") > > val sc = new SparkContext(conf) > val text = sc.textFile(filepath, 3) > > val numOfHello = text.filter(line => line.contains("hello")).count() > > println("number of lines containing 'hello' is " + numOfHello) > println("down") > } > } > > ______________________________________________________________________ > > > > The corresponding sbt file, $SPARK_HOME/simple.sbt, is as follows: > _________________________________________________________________ > > name := "Simple Distributed App" > > version := "1.0" > > scalaVersion := "2.10.3" > > libraryDependencies += "org.apache.spark" %% "spark-core" % "0.9.0-incubating" > > resolvers += "Akka Repository" at "http://repo.akka.io/releases/" > _________________________________________________________________ > > > I built the application into > $SPARK_HOME/target/scala-2.10/simple-distributed-app_2.10-1.0.jar, using the > command > SPARK_HADOOP_VERSION=1.2.1 sbt/sbt package > > I ran it using the command "sbt/sbt run" and it finished running > successfully. > > But I'm not sure what's the correct and general way to submit and run a job > in Spark cluster. To be specific,after having built a job into a JAR file, > say simpleApp.jar, where should I put it and how should I submit it to Spark > cluster? > > > > > > > >