Any other insights into this issue? I tried multiple way to supply keytab to executor
Does spark standalone doesn’t support Kerberos? > On Jan 8, 2021, at 1:53 PM, Sudhir Babu Pothineni <sbpothin...@gmail.com> > wrote: > > > Incase of Spark on Yarn, Application Master shares the token. > > I think incase of spark stand alone the token is not shared to executor, any > example how to get the HDFS token for executor? > >> On Fri, Jan 8, 2021 at 12:13 PM Gabor Somogyi <gabor.g.somo...@gmail.com> >> wrote: >> TGT is not enough, you need HDFS token which can be obtained by Spark. >> Please check the logs... >> >>> On Fri, 8 Jan 2021, 18:51 Sudhir Babu Pothineni, <sbpothin...@gmail.com> >>> wrote: >>> I spin up a spark standalone cluster (spark.autheticate=false), submitted a >>> job which reads remote kerberized HDFS, >>> >>> val spark = SparkSession.builder() >>> .master("spark://spark-standalone:7077") >>> .getOrCreate() >>> >>> UserGroupInformation.loginUserFromKeytab(principal, keytab) >>> val df = spark.read.parquet("hdfs://namenode:8020/test/parquet/") >>> >>> Ran into following exception: >>> >>> Caused by: >>> java.io.IOException: java.io.IOException: Failed on local exception: >>> java.io.IOException: org.apache.hadoop.security.AccessControlException: >>> Client cannot authenticate via:[TOKEN, KERBEROS]; Host Details : local host >>> is: "..."; destination host is: "...":10346; >>> >>> >>> Any suggestions? >>> >>> Thanks >>> Sudhir