Has anyone seen this error? Not sure which dir the program was trying to write to.
I am running Spark 1.4.1, submitting Spark job to Yarn, in yarn-client mode. 15/09/04 21:36:06 ERROR SparkContext: Error adding jar (java.io.IOException: No space left on device), was the --addJars option used? 15/09/04 21:36:08 ERROR SparkContext: Error initializing SparkContext. java.io.IOException: No space left on device at java.io.FileOutputStream.writeBytes(Native Method) at java.io.FileOutputStream.write(FileOutputStream.java:300) at java.util.zip.DeflaterOutputStream.deflate(DeflaterOutputStream.java:178) at java.util.zip.ZipOutputStream.closeEntry(ZipOutputStream.java:213) at java.util.zip.ZipOutputStream.finish(ZipOutputStream.java:318) at java.util.zip.DeflaterOutputStream.close(DeflaterOutputStream.java:163) at java.util.zip.ZipOutputStream.close(ZipOutputStream.java:338) at org.apache.spark.deploy.yarn.Client.createConfArchive(Client.scala:432) at org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:338) at org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:561) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:115) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:141) at org.apache.spark.SparkContext.<init>(SparkContext.scala:497) Thanks, Shenyan