I have set up a cluster on AWS and am trying a really simple hello world program as a test. The cluster was built using the ec2 scripts that come with Spark. Anyway, I have output the error message (using --verbose) below. The source code is further below that.
Any help would be greatly appreciated. Thanks, Eric *Error code:* r...@ip-xx.xx.xx.xx ~]$ ./spark/bin/spark-submit --verbose --class com.je.test.Hello --master spark://xx.xx.xx.xx:7077 Hello-assembly-1.0.jar Spark assembly has been built with Hive, including Datanucleus jars on classpath Using properties file: /root/spark/conf/spark-defaults.conf Adding default property: spark.executor.memory=5929m Adding default property: spark.executor.extraClassPath=/root/ephemeral-hdfs/conf Adding default property: spark.executor.extraLibraryPath=/root/ephemeral-hdfs/lib/native/ Using properties file: /root/spark/conf/spark-defaults.conf Adding default property: spark.executor.memory=5929m Adding default property: spark.executor.extraClassPath=/root/ephemeral-hdfs/conf Adding default property: spark.executor.extraLibraryPath=/root/ephemeral-hdfs/lib/native/ Parsed arguments: master spark://xx.xx.xx.xx:7077 deployMode null executorMemory 5929m executorCores null totalExecutorCores null propertiesFile /root/spark/conf/spark-defaults.conf extraSparkProperties Map() driverMemory null driverCores null driverExtraClassPath null driverExtraLibraryPath null driverExtraJavaOptions null supervise false queue null numExecutors null files null pyFiles null archives null mainClass com.je.test.Hello primaryResource file:/root/Hello-assembly-1.0.jar name com.je.test.Hello childArgs [] jars null verbose true Default properties from /root/spark/conf/spark-defaults.conf: spark.executor.extraLibraryPath -> /root/ephemeral-hdfs/lib/native/ spark.executor.memory -> 5929m spark.executor.extraClassPath -> /root/ephemeral-hdfs/conf Using properties file: /root/spark/conf/spark-defaults.conf Adding default property: spark.executor.memory=5929m Adding default property: spark.executor.extraClassPath=/root/ephemeral-hdfs/conf Adding default property: spark.executor.extraLibraryPath=/root/ephemeral-hdfs/lib/native/ Main class: com.je.test.Hello Arguments: System properties: spark.executor.extraLibraryPath -> /root/ephemeral-hdfs/lib/native/ spark.executor.memory -> 5929m SPARK_SUBMIT -> true spark.app.name -> com.je.test.Hello spark.jars -> file:/root/Hello-assembly-1.0.jar spark.executor.extraClassPath -> /root/ephemeral-hdfs/conf spark.master -> spark://xxx.xx.xx.xxx:7077 Classpath elements: file:/root/Hello-assembly-1.0.jar *Actual Error:* Exception in thread "main" java.lang.NullPointerException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) *Source Code:* package com.je.test import org.apache.spark.{SparkConf, SparkContext} class Hello { def main(args: Array[String]): Unit = { val conf = new SparkConf(true)//.set("spark.cassandra.connection.host", "xxx.xx.xx.xxx") val sc = new SparkContext("spark://xxx.xx.xx.xxx:7077", "Season", conf) println("Hello World") } }