Github user liyinan926 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/21669#discussion_r223433207
  
    --- Diff: docs/security.md ---
    @@ -722,7 +722,84 @@ with encryption, at least.
     The Kerberos login will be periodically renewed using the provided 
credentials, and new delegation
     tokens for supported will be created.
     
    +## Secure Interaction with Kubernetes
    +
    +When talking to Hadoop-based services behind Kerberos, it was noted that 
Spark needs to obtain delegation tokens
    +so that non-local processes can authenticate. These delegation tokens in 
Kubernetes are stored in Secrets that are 
    +shared by the Driver and its Executors. As such, there are three ways of 
submitting a Kerberos job: 
    +
    +In all cases you must define the environment variable: `HADOOP_CONF_DIR` 
as well as either 
    +`spark.kubernetes.kerberos.krb5.location` or 
`spark.kubernetes.kerberos.krb5.configMapName`.
    +
    +It also important to note that the KDC needs to be visible from inside the 
containers if the user uses a local
    +krb5 file. 
    +
    +If a user wishes to use a remote HADOOP_CONF directory, that contains the 
Hadoop configuration files, this could be
    +achieved by mounting a pre-defined ConfigMap in the desired location that 
you can point to via the appropriate configs.
    --- End diff --
    
    It's better to mention that users need to set the environment variable 
`HADOOP_CONF_DIR` on the container pointing to the path where the ConfigMap is 
mounted.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to