This is generally a side effect of your executor being killed. For example, Yarn will do that if you're going over the requested memory limits.
On Tue, Jul 8, 2014 at 12:17 PM, Rahul Bhojwani <rahulbhojwani2...@gmail.com> wrote: > HI, > > I am getting this error. Can anyone help out to explain why is this error > coming. > > ######## > > Exception in thread "delete Spark temp dir > C:\Users\shawn\AppData\Local\Temp\spark-27f60467-36d4-4081-aaf5-d0ad42dda560" > java.io.IOException: Failed to delete: > C:\Users\shawn\AppData\Local\Temp\spark-27f60467-36d4-4081-aaf5-d0ad42dda560\tmp > cmenlp > at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:483) > at > org.apache.spark.util.Utils$$anonfun$deleteRecursively$1.apply(Utils.scala:479) > at > org.apache.spark.util.Utils$$anonfun$deleteRecursively$1.apply(Utils.scala:478) > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) > at > scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:34) > at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:478) > at org.apache.spark.util.Utils$$anon$4.run(Utils.scala:212) > PS> > ############ > > > > > Thanks in advance > -- > Rahul K Bhojwani > 3rd Year B.Tech > Computer Science and Engineering > National Institute of Technology, Karnataka -- Marcelo