I have set up a cluster on AWS and am trying a really simple hello world
program as a test.  The cluster was built using the ec2 scripts that come
with Spark.  Anyway, I have output the error message (using --verbose)
below.  The source code is further below that.

Any help would be greatly appreciated.

Thanks,

Eric

*Error code:*

r...@ip-xx.xx.xx.xx ~]$ ./spark/bin/spark-submit  --verbose  --class
com.je.test.Hello --master spark://xx.xx.xx.xx:7077
 Hello-assembly-1.0.jar
Spark assembly has been built with Hive, including Datanucleus jars on
classpath
Using properties file: /root/spark/conf/spark-defaults.conf
Adding default property: spark.executor.memory=5929m
Adding default property:
spark.executor.extraClassPath=/root/ephemeral-hdfs/conf
Adding default property:
spark.executor.extraLibraryPath=/root/ephemeral-hdfs/lib/native/
Using properties file: /root/spark/conf/spark-defaults.conf
Adding default property: spark.executor.memory=5929m
Adding default property:
spark.executor.extraClassPath=/root/ephemeral-hdfs/conf
Adding default property:
spark.executor.extraLibraryPath=/root/ephemeral-hdfs/lib/native/
Parsed arguments:
  master                  spark://xx.xx.xx.xx:7077
  deployMode              null
  executorMemory          5929m
  executorCores           null
  totalExecutorCores      null
  propertiesFile          /root/spark/conf/spark-defaults.conf
  extraSparkProperties    Map()
  driverMemory            null
  driverCores             null
  driverExtraClassPath    null
  driverExtraLibraryPath  null
  driverExtraJavaOptions  null
  supervise               false
  queue                   null
  numExecutors            null
  files                   null
  pyFiles                 null
  archives                null
  mainClass               com.je.test.Hello
  primaryResource         file:/root/Hello-assembly-1.0.jar
  name                    com.je.test.Hello
  childArgs               []
  jars                    null
  verbose                 true

Default properties from /root/spark/conf/spark-defaults.conf:
  spark.executor.extraLibraryPath -> /root/ephemeral-hdfs/lib/native/
  spark.executor.memory -> 5929m
  spark.executor.extraClassPath -> /root/ephemeral-hdfs/conf


Using properties file: /root/spark/conf/spark-defaults.conf
Adding default property: spark.executor.memory=5929m
Adding default property:
spark.executor.extraClassPath=/root/ephemeral-hdfs/conf
Adding default property:
spark.executor.extraLibraryPath=/root/ephemeral-hdfs/lib/native/
Main class:
com.je.test.Hello
Arguments:

System properties:
spark.executor.extraLibraryPath -> /root/ephemeral-hdfs/lib/native/
spark.executor.memory -> 5929m
SPARK_SUBMIT -> true
spark.app.name -> com.je.test.Hello
spark.jars -> file:/root/Hello-assembly-1.0.jar
spark.executor.extraClassPath -> /root/ephemeral-hdfs/conf
spark.master -> spark://xxx.xx.xx.xxx:7077
Classpath elements:
file:/root/Hello-assembly-1.0.jar

*Actual Error:*
Exception in thread "main" java.lang.NullPointerException
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)


*Source Code:*
package com.je.test


import org.apache.spark.{SparkConf, SparkContext}

class Hello {

  def main(args: Array[String]): Unit = {

    val conf = new
SparkConf(true)//.set("spark.cassandra.connection.host",
"xxx.xx.xx.xxx")
    val sc = new SparkContext("spark://xxx.xx.xx.xxx:7077", "Season", conf)

    println("Hello World")

  }
}

Reply via email to