Hi Team,
I am new to spark and writing my first program. I have written sample
program with spark master as local. To execute spark over local yarn what
should be value of spark.master property? Can I point to remote yarn
cluster? I would like to execute this as a java application and not submit
using spark-submit.
Main objective is to create a service which can execute spark sql queries
over yarn cluster.
Thanks in advance.
Regards,
Atul
code snippet as below
import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.SparkSession;
public class TestSpark {
public static void main(String[] args) {
SparkSession spark = SparkSession.builder().appName("Java Spark
Sql
Example").config("spark.master","local").getOrCreate();
Dataset df =
spark.read().json("/spark/sparkSql/employee.json");
System.out.println("Data");
df.cache();
df.show();
// JavaSparkContext sc = new JavaSparkContext(new
SparkConf().setAppName("SparkJoins").setMaster("yarn-client"));
}
}
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Running-spark-Java-on-yarn-cluster-tp27504.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org