This is the whole output from the shell:
~/spark-1.3.0-bin-hadoop2.4$ sudo bin/spark-shell
Spark assembly has been built with Hive, including Datanucleus jars on
classpath
log4j:WARN No appenders could be found for logger
(org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
log4j:WARN Please in
I have the same problem, i.e. exception with the same call stack when I start
either pyspark or spark-shell. I use spark-1.3.0-bin-hadoop2.4 on ubuntu
14.10.
bin/pyspark
bunch of INFO messages, then ActorInitializationException exception.
Shell starts, I can do this:
>>> rd = sc.parallelize([1,2]