Any other insights into this issue? I tried multiple way to supply keytab to 
executor 

Does spark standalone doesn’t support Kerberos?

> On Jan 8, 2021, at 1:53 PM, Sudhir Babu Pothineni <sbpothin...@gmail.com> 
> wrote:
> 
> 
> Incase of Spark on Yarn, Application Master shares the token. 
> 
> I think incase of spark stand alone the token is not shared to executor, any 
> example how to get the HDFS token for executor?
> 
>> On Fri, Jan 8, 2021 at 12:13 PM Gabor Somogyi <gabor.g.somo...@gmail.com> 
>> wrote:
>> TGT is not enough, you need HDFS token which can be obtained by Spark. 
>> Please check the logs...
>> 
>>> On Fri, 8 Jan 2021, 18:51 Sudhir Babu Pothineni, <sbpothin...@gmail.com> 
>>> wrote:
>>> I spin up a spark standalone cluster (spark.autheticate=false), submitted a 
>>> job which reads remote kerberized HDFS, 
>>> 
>>> val spark = SparkSession.builder()
>>>                   .master("spark://spark-standalone:7077")
>>>                   .getOrCreate()
>>> 
>>> UserGroupInformation.loginUserFromKeytab(principal, keytab)
>>> val df = spark.read.parquet("hdfs://namenode:8020/test/parquet/")
>>> 
>>> Ran into following exception:
>>> 
>>> Caused by:
>>> java.io.IOException: java.io.IOException: Failed on local exception: 
>>> java.io.IOException: org.apache.hadoop.security.AccessControlException: 
>>> Client cannot authenticate via:[TOKEN, KERBEROS]; Host Details : local host 
>>> is: "..."; destination host is: "...":10346; 
>>> 
>>> 
>>> Any suggestions?
>>> 
>>> Thanks
>>> Sudhir

Reply via email to