This is why spark.hadoop.validateOutputSpecs exists, really:
https://spark.apache.org/docs/latest/configuration.html

On Mon, Apr 20, 2015 at 3:40 AM, wyphao.2007 <wyphao.2...@163.com> wrote:
> Hi,
>    When I recovery from checkpoint in yarn-cluster mode using Spark 
> Streaming, I found it will reuse the application id (In my case is 
> application_1428664056212_0016) before falied to write spark eventLog, But 
> now my application id is application_1428664056212_0017,then spark write 
> eventLog will falied, the stacktrace as follow:
> 15/04/14 10:14:01 WARN util.ShutdownHookManager: ShutdownHook '$anon$3' 
> failed, java.io.IOException: Target log file already exists 
> (hdfs://mycluster/spark-logs/eventLog/application_1428664056212_0016)
> java.io.IOException: Target log file already exists 
> (hdfs://mycluster/spark-logs/eventLog/application_1428664056212_0016)
>         at 
> org.apache.spark.scheduler.EventLoggingListener.stop(EventLoggingListener.scala:201)
>         at 
> org.apache.spark.SparkContext$$anonfun$stop$4.apply(SparkContext.scala:1388)
>         at 
> org.apache.spark.SparkContext$$anonfun$stop$4.apply(SparkContext.scala:1388)
>         at scala.Option.foreach(Option.scala:236)
>         at org.apache.spark.SparkContext.stop(SparkContext.scala:1388)
>         at 
> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$3.run(ApplicationMaster.scala:107)
>         at 
> org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)
> Is someone can help me, The issue is SPARK-6892.
> thanks
>
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to