No, spark can not do that as it does not replicate partitions (so no retry on different worker). It seems your cluster is not provisioned with correct permissions. I would suggest to automate node provisioning.
On Mon, Jun 29, 2015 at 11:04 PM, maxdml <maxdemou...@gmail.com> wrote: > Hi there, > > I have some traces from my master and some workers where for some reason, > the ./work directory of an application can not be created on the workers. > There is also an issue with the master's temp directory creation. > > master logs: http://pastebin.com/v3NCzm0u > worker's logs: http://pastebin.com/Ninkscnx > > It seems that some of the executors can create the directories, but as some > others are repetitively failing, the job ends up failing. Shouldn't spark > manage to keep working with a smallest number of executors instead of > failing? > > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Directory-creation-failed-leads-to-job-fail-should-it-tp23531.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > > -- Best Regards, Ayan Guha