prikshit created SPARK-51464:
--------------------------------
Summary:
unable-to-stop-spark-application-after-enabling-authentication
Key: SPARK-51464
URL: https://issues.apache.org/jira/browse/SPARK-51464
Project: Spark
Issue Type: Bug
Components: Spark Core
Affects Versions: 3.5.4
Environment: 3 node Spark Cluster
All Master (1 Alive , 2 StandBy)
All Workers
Reporter: prikshit
Fix For: 3.5.4
I have enabled authentication in my Spark standalone cluster by setting
{{spark.authenticate=true}} in {{{}spark-defaults.conf{}}}. I am submitting my
job using {{spark-submit}} and passing the secret as {{{}--conf
spark.authenticate.secret=somesecretvalue{}}}. Below is the command I use to
submit my job:
{{spark-submit --class MySampleClass --conf
spark.authenticate.secret=somesecretvalue --supervise --conf
spark.submit.deployMode=cluster}}
We are able to successfully submit the job and also in application we are
setting the secret inSspark context , our application is up and runnning.
{{SparkConf sparkConf = new SparkConf();
sparkConf.set('spark.authenticate.secret','somesecretvalue')}}
Now to kill the application, I use two different options, each option is giving
respective exceptions.
I tried with 2 options
Option - 1. {{spark-class org.apache.spark.deploy.Client kill
<ACTIVE_MASTR_NODE>:7077 DRIVER_SUBMISSION_ID}} has stopped working with below
error:
{{Exception in thread "main" org.apache.spark.SparkException: Exception thrown
in awaitResult:
at
org.apache.spark.util.SparkThreadUtils$.awaitResult(SparkThreadUtils.scala:56)
at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:310)
at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:102)
at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:110)
at org.apache.spark.deploy.ClientApp.$anonfun$start$2(Client.scala:292)
at
scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286)
at
scala.collection.IndexedSeqOptimized.foreach(IndexedSeqOptimized.scala:36)
at
scala.collection.IndexedSeqOptimized.foreach$(IndexedSeqOptimized.scala:33)
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:198)
at scala.collection.TraversableLike.map(TraversableLike.scala:286)
at scala.collection.TraversableLike.map$(TraversableLike.scala:279)
at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:198)
at org.apache.spark.deploy.ClientApp.start(Client.scala:292)
at org.apache.spark.deploy.Client$.main(Client.scala:274)
at org.apache.spark.deploy.Client.main(Client.scala)
Caused by: java.lang.RuntimeException: java.lang.IllegalStateException:
Expected SaslM}}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]