Github user arunmahadevan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22598#discussion_r228681974
  
    --- Diff: 
core/src/main/scala/org/apache/spark/deploy/security/HadoopDelegationTokenManager.scala
 ---
    @@ -66,7 +66,8 @@ private[spark] class HadoopDelegationTokenManager(
       private def getDelegationTokenProviders: Map[String, 
HadoopDelegationTokenProvider] = {
         val providers = Seq(new HadoopFSDelegationTokenProvider(fileSystems)) 
++
           safeCreateProvider(new HiveDelegationTokenProvider) ++
    -      safeCreateProvider(new HBaseDelegationTokenProvider)
    +      safeCreateProvider(new HBaseDelegationTokenProvider) ++
    +      safeCreateProvider(new KafkaDelegationTokenProvider)
    --- End diff --
    
    yes, I think the best we can do is to document the configs and throw some 
useful error messages to make the user aware of the "bootstrapservers" config 
(in case they accidently left it) when the spark-sql-kafka libraries are not in 
the classpath.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to