[ https://issues.apache.org/jira/browse/SPARK-8676?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Tao Wang updated SPARK-8676: ---------------------------- Summary: After TGT expires, Thrift Server get "No valid credentials provided" exception (was: After TGT expires, Thrift Server get "No invalid credentials" exception) > After TGT expires, Thrift Server get "No valid credentials provided" exception > ------------------------------------------------------------------------------ > > Key: SPARK-8676 > URL: https://issues.apache.org/jira/browse/SPARK-8676 > Project: Spark > Issue Type: Bug > Components: SQL > Reporter: Tao Wang > > I ran Thrit Server in secure Hadoop. With "spark.yarn.keytab" and > "spark.yarn.principal" configured in spark-defaults.conf, > "hive.server2.authentication.kerberos.principal" and > "hive.server2.authentication.kerberos.keytab" in hive-site.xml. But after the > ticket which can be get from Kerberos using principal expires, it throw the > excepiton below: > {code} > 2015-06-26 19:16:32,411 | WARN | [LeaseRenewer:spark@hacluster, > clients=[DFSClient_NONMAPREDUCE_-809515000_1], created at > java.lang.Throwable: TRACE > at > org.apache.hadoop.hdfs.client.impl.LeaseRenewer.<init>(LeaseRenewer.java:206) > at > org.apache.hadoop.hdfs.client.impl.LeaseRenewer.<init>(LeaseRenewer.java:75) > at > org.apache.hadoop.hdfs.client.impl.LeaseRenewer$Factory.get(LeaseRenewer.java:147) > at > org.apache.hadoop.hdfs.client.impl.LeaseRenewer$Factory.access$100(LeaseRenewer.java:94) > at > org.apache.hadoop.hdfs.client.impl.LeaseRenewer.getInstance(LeaseRenewer.java:84) > at org.apache.hadoop.hdfs.DFSClient.getLeaseRenewer(DFSClient.java:480) > at org.apache.hadoop.hdfs.DFSClient.beginFileLease(DFSClient.java:486) > at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1361) > at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1288) > at > org.apache.hadoop.hdfs.DistributedFileSystem$8.doCall(DistributedFileSystem.java:486) > at > org.apache.hadoop.hdfs.DistributedFileSystem$8.doCall(DistributedFileSystem.java:482) > at > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) > at > org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:482) > at > org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:388) > at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909) > at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890) > at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787) > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:365) > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338) > at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:289) > at > org.apache.spark.deploy.yarn.Client.copyFileToRemote(Client.scala:232) > at > org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:288) > at > org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:559) > at > org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:115) > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:58) > at > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:141) > at org.apache.spark.SparkContext.<init>(SparkContext.scala:497) > at > org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.scala:51) > at > org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveThriftServer2.scala:73) > at > org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThriftServer2.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:666) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:172) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:195) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:114) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > ] | Exception encountered while connecting to the server : > javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to find > any Kerberos tgt)] | > org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:693) > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org