This is generally a side effect of your executor being killed. For
example, Yarn will do that if you're going over the requested memory
limits.

On Tue, Jul 8, 2014 at 12:17 PM, Rahul Bhojwani
<rahulbhojwani2...@gmail.com> wrote:
> HI,
>
> I am getting this error. Can anyone help out to explain why is this error
> coming.
>
> ########
>
> Exception in thread "delete Spark temp dir
> C:\Users\shawn\AppData\Local\Temp\spark-27f60467-36d4-4081-aaf5-d0ad42dda560"
>  java.io.IOException: Failed to delete:
> C:\Users\shawn\AppData\Local\Temp\spark-27f60467-36d4-4081-aaf5-d0ad42dda560\tmp
> cmenlp
>         at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:483)
>         at
> org.apache.spark.util.Utils$$anonfun$deleteRecursively$1.apply(Utils.scala:479)
>         at
> org.apache.spark.util.Utils$$anonfun$deleteRecursively$1.apply(Utils.scala:478)
>         at
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>         at
> scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:34)
>         at org.apache.spark.util.Utils$.deleteRecursively(Utils.scala:478)
>         at org.apache.spark.util.Utils$$anon$4.run(Utils.scala:212)
> PS>
> ############
>
>
>
>
> Thanks in advance
> --
> Rahul K Bhojwani
> 3rd Year B.Tech
> Computer Science and Engineering
> National Institute of Technology, Karnataka



-- 
Marcelo

Reply via email to