The stack trace is related to hdfs.

Can you tell us which hadoop release you are using ?

Is this a secure cluster ?

Thanks

On Fri, May 22, 2015 at 1:55 PM, Wang, Ningjun (LNG-NPV) <
ningjun.w...@lexisnexis.com> wrote:

>  I used spark standalone cluster on Windows 2008. I kept on getting the
> following error when trying to save an RDD to a windows shared folder
>
>
>
>
> rdd.saveAsObjectFile(“file:///T:/lab4-win02/IndexRoot01/tobacco-07/myrdd.obj”)
>
>
>
> 15/05/22 16:49:05 ERROR Executor: Exception in task 0.0 in stage 12.0 (TID
> 12)
>
> java.io.IOException: Mkdirs failed to create
> file:/T:/lab4-win02/IndexRoot01/tobacco-07/tmp/docs-150522204904805.op/_temporary/0/_temporary/attempt_201505221649_0012_m_000000_12
>
>             at
> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:438)
>
>             at
> org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
>
>             at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:906)
>
>             at
> org.apache.hadoop.io.SequenceFile$Writer.<init>(SequenceFile.java:1071)
>
>             at
> org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:270)
>
>             at
> org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:527)
>
>             at
> org.apache.hadoop.mapred.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:63)
>
>             at
> org.apache.spark.SparkHadoopWriter.open(SparkHadoopWriter.scala:90)
>
>             at
> org.apache.spark.rdd.PairRDDFunctions$$anonfun$13.apply(PairRDDFunctions.scala:1068)
>
>             at
> org.apache.spark.rdd.PairRDDFunctions$$anonfun$13.apply(PairRDDFunctions.scala:1059)
>
>             at
> org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
>
>             at org.apache.spark.scheduler.Task.run(Task.scala:64)
>
>             at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:203)
>
>             at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>
>             at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>
>             at java.lang.Thread.run(Thread.java:745)
>
>  The T: drive is mapped to a windows shared folder, e.g.  T:  ->
> \\10.196.119.230\myshare
>
>
>
> The id running spark does have write permission to this folder. It works
> most of the time but failed sometime.
>
>
>
> Can anybody tell me what is the problem here?
>
>
>
> Please advise. Thanks.
>

Reply via email to