[ 
https://issues.apache.org/jira/browse/SPARK-34293?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17276913#comment-17276913
 ] 

Manohar Chamaraju commented on SPARK-34293:
-------------------------------------------

Update:
 # In client mode by adding fs.defaultFS in core-site.xml fixed the issue for 
me.
 # what do to work is usage of hadoop-conf configmap in client mode.

> kubernetes executor pod unable to access secure hdfs
> ----------------------------------------------------
>
>                 Key: SPARK-34293
>                 URL: https://issues.apache.org/jira/browse/SPARK-34293
>             Project: Spark
>          Issue Type: Bug
>          Components: Kubernetes
>    Affects Versions: 3.0.1
>            Reporter: Manohar Chamaraju
>            Priority: Major
>         Attachments: driver.log, executor.log, 
> image-2021-01-30-00-13-18-234.png, image-2021-01-30-00-14-14-329.png, 
> image-2021-01-30-00-14-45-335.png, image-2021-01-30-00-20-54-620.png, 
> image-2021-01-30-00-33-02-109.png, image-2021-01-30-00-34-05-946.png
>
>
> Steps to reproduce
>  # Configure secure HDFS(kerberos) cluster running as containers in 
> kubernetes.
>  # Configure KDC on centos and create keytab for user principal hdfs, in 
> hdfsuser.keytab.
>  # Genearte spark image(v3.0.1), to spawn as container out of spark image.
>  # Inside spark container, run export HADOOP_CONF_DIR=/etc/hadoop/conf/ with 
> core-site.xml configuration as below 
>  !image-2021-01-30-00-13-18-234.png!
>  # Create configmap kbr-conf 
>  !image-2021-01-30-00-14-14-329.png!
>  # Run the command /opt/spark/bin/spark-submit \
>  --deploy-mode client \
>  --executor-memory 1g\
>  --executor-memory 1g\
>  --executor-cores 1\
>  --class org.apache.spark.examples.HdfsTest \
>  --conf spark.kubernetes.namespace=arcsight-installer-lh7fm\
>  --master k8s://[https://172.17.17.1:443|https://172.17.17.1/] \
>  --conf spark.kubernetes.authenticate.driver.serviceAccountName=spark \
>  --conf spark.app.name=spark-hdfs \
>  --conf spark.executer.instances=1 \
>  --conf spark.kubernetes.node.selector.spark=yes\
>  --conf spark.kubernetes.node.selector.Worker=label\
>  --conf spark.kubernetes.container.image=manohar/spark:v3.0.1 \
>  --conf spark.kubernetes.kerberos.enabled=true \
>  --conf spark.kubernetes.kerberos.krb5.configMapName=krb5-conf \
>  --conf spark.kerberos.keytab=/data/hdfsuser.keytab \
>  --conf spark.kerberos.principal=h...@dom047600.lab \
>  local:///opt/spark/examples/jars/spark-examples_2.12-3.0.1.jar \
>  hdfs://hdfs-namenode:30820/staging-directory.
>  # On running this command driver is able to connect hdfs with kerberos but 
> execurtor fails to connect to secure hdfs and below is the logs 
> !image-2021-01-30-00-34-05-946.png!
>  # Some of observation
>  ## In Client mode, --conf spark.kubernetes.hadoop.configMapName=hadoop-conf 
> as not effect only works after HADOOP_CONF_DIR is set. Below was the contents 
> of hadoop-conf configmap.
>  !image-2021-01-30-00-20-54-620.png!
>  ## Ran the command in cluster mode as well, in cluster mode also executor 
> could not connect to secure hdfs.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to