AngersZhuuuu commented on pull request #33780:
URL: https://github.com/apache/spark/pull/33780#issuecomment-903833885


   > The problem is that not everyone calls sc.stop... in interactive 
environments you also have people killing it weirdly, there were a bunch of 
shut down cases in client mode that made determining the real success/failure 
an issue.
   > 
   > It was intentionally decided to just always return success in those cases. 
If this properly handles those cases I'm fine with it but I need to go try to 
remember what they all were and make sure they are handled. We should not 
report FAIL unless its truly a failure.
   
   Got your point..I miss the case such as killing spark-shell weirdly...then 
how about:
   
   - add a configuration such as spark.yarn.clientModeAMDisconnect as Failed, 
when true,  make it as failed, when false let it behavior as before? then let 
user to choose? Two company's scheduler platform meet such case, so we can have 
a choose. Since we always non't need to rerun  pysaprk/spark-shell/spark-sql?
   - For pyspark/spark-shell/spark-sql such interactive mode, make it behavior 
as before and for normal batch job, make it failed when disconnected?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to