I fixed the issue (as no permission on keytab file), please ignore. On Sun, Aug 13, 2017 at 9:42 AM, Ascot Moss <ascot.m...@gmail.com> wrote:
> Hi, > > Spark: 2.1.0 > Hive: 2.1.1 > > When starting thrift server, I got the following error: > > How can I fix it? > > Regards > > > > (error log) > 17/08/13 09:28:17 DEBUG client.IsolatedClientLoader: shared class: > java.lang.NoSuchFieldError > 17/08/13 09:28:17 DEBUG client.IsolatedClientLoader: shared class: > org.apache.hadoop.security.SecurityUtil > 17/08/13 09:28:17 DEBUG client.IsolatedClientLoader: shared class: > org.apache.hadoop.security.SaslRpcServer > 17/08/13 09:28:17 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.transport.TSaslTransportException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/libthrift-0.9.2.jar!/org/apache/thrift/transport/ > TSaslTransportException.class > 17/08/13 09:28:17 DEBUG client.IsolatedClientLoader: hive class: > javax.security.sasl.Sasl - jar:file:/opt/jdk1.8.0_102/ > jre/lib/rt.jar!/javax/security/sasl/Sasl.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.transport.TMemoryInputTransport - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/libthrift-0.9.2.jar!/org/apache/thrift/transport/ > TMemoryInputTransport.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.TByteArrayOutputStream - jar:file:/home/spark-2.1.0/ > assembly/target/scala-2.11/jars/libthrift-0.9.2.jar!/org/apache/thrift/ > TByteArrayOutputStream.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.transport.TSaslTransport$SaslParticipant - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/libthrift-0.9.2.jar!/org/apache/thrift/transport/TSaslTransport$ > SaslParticipant.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.protocol.TProtocolException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/libthrift-0.9.2.jar!/org/apache/thrift/protocol/ > TProtocolException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.protocol.TStruct - jar:file:/home/spark-2.1.0/ > assembly/target/scala-2.11/jars/libthrift-0.9.2.jar!/org/ > apache/thrift/protocol/TStruct.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/ThriftHiveMetastore$Client.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > com.facebook.fb303.FacebookService$Client - jar:file:/home/spark-2.1.0/ > assembly/target/scala-2.11/jars/libfb303-0.9.2.jar!/com/facebook/fb303/ > FacebookService$Client.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.TServiceClient - jar:file:/home/spark-2.1.0/ > assembly/target/scala-2.11/jars/libthrift-0.9.2.jar!/org/ > apache/thrift/TServiceClient.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.InvalidObjectException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/InvalidObjectException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.UnknownTableException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/UnknownTableException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.UnknownDBException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/UnknownDBException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.ConfigValSecurityException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/ConfigValSecurityException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.UnknownPartitionException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/UnknownPartitionException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.InvalidPartitionException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/InvalidPartitionException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.InvalidInputException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/InvalidInputException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.NoSuchTxnException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/NoSuchTxnException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.TxnAbortedException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/TxnAbortedException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.NoSuchLockException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/NoSuchLockException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.metastore.api.TxnOpenException - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-metastore-1.2.1.spark2.jar!/org/apache/hadoop/ > hive/metastore/api/TxnOpenException.class > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1 - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/hive-exec-1.2.1.spark2.jar!/org/apache/hadoop/hive/thrift/client/ > TUGIAssumingTransport$1.class > 17/08/13 09:28:18 DEBUG security.UserGroupInformation: PrivilegedAction > as:spark (auth:KERBEROS) from:org.apache.hadoop.hive.thrift.client. > TUGIAssumingTransport.open(TUGIAssumingTransport.java:49) > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: shared class: > org.slf4j.Logger > 17/08/13 09:28:18 DEBUG transport.TSaslTransport: opening transport > org.apache.thrift.transport.TSaslClientTransport@7f08caf > 17/08/13 09:28:18 DEBUG client.IsolatedClientLoader: hive class: > javax.security.sasl.SaslClient - jar:file:/opt/jdk1.8.0_102/ > jre/lib/rt.jar!/javax/security/sasl/SaslClient.class > 17/08/13 09:28:18 ERROR transport.TSaslTransport: SASL negotiation failure > javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to > find any Kerberos tgt)] > at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge( > GssKrb5Client.java:211) > at org.apache.thrift.transport.TSaslClientTransport. > handleSaslStartMessage(TSaslClientTransport.java:94) > at org.apache.thrift.transport.TSaslTransport.open( > TSaslTransport.java:271) > at org.apache.thrift.transport.TSaslClientTransport.open( > TSaslClientTransport.java:37) > at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$ > 1.run(TUGIAssumingTransport.java:52) > at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$ > 1.run(TUGIAssumingTransport.java:49) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at org.apache.hadoop.security.UserGroupInformation.doAs( > UserGroupInformation.java:1698) > at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport. > open(TUGIAssumingTransport.java:49) > at org.apache.hadoop.hive.metastore.HiveMetaStoreClient. > open(HiveMetaStoreClient.java:420) > at org.apache.hadoop.hive.metastore.HiveMetaStoreClient. > <init>(HiveMetaStoreClient.java:236) > at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>( > SessionHiveMetaStoreClient.java:74) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.hadoop.hive.metastore.MetaStoreUtils. > newInstance(MetaStoreUtils.java:1521) > at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init> > (RetryingMetaStoreClient.java:86) > at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy( > RetryingMetaStoreClient.java:132) > at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy( > RetryingMetaStoreClient.java:104) > at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive. > java:3005) > at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024) > at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234) > at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174) > at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166) > at org.apache.hadoop.hive.ql.session.SessionState.start( > SessionState.java:503) > at org.apache.spark.sql.hive.client.HiveClientImpl.<init>( > HiveClientImpl.scala:192) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient( > IsolatedClientLoader.scala:264) > at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata( > HiveUtils.scala:366) > at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata( > HiveUtils.scala:270) > at org.apache.spark.sql.hive.HiveExternalCatalog.<init>( > HiveExternalCatalog.scala:65) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.spark.sql.internal.SharedState$.org$apache$spark$ > sql$internal$SharedState$$reflect(SharedState.scala:166) > at org.apache.spark.sql.internal.SharedState.<init>(SharedState.scala:86) > at org.apache.spark.sql.SparkSession$$anonfun$sharedState$1.apply( > SparkSession.scala:101) > at org.apache.spark.sql.SparkSession$$anonfun$sharedState$1.apply( > SparkSession.scala:101) > at scala.Option.getOrElse(Option.scala:121) > at org.apache.spark.sql.SparkSession.sharedState$ > lzycompute(SparkSession.scala:101) > at org.apache.spark.sql.SparkSession.sharedState(SparkSession.scala:100) > at org.apache.spark.sql.internal.SessionState.<init>( > SessionState.scala:157) > at org.apache.spark.sql.hive.HiveSessionState.<init>( > HiveSessionState.scala:32) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$ > reflect(SparkSession.scala:978) > at org.apache.spark.sql.SparkSession.sessionState$ > lzycompute(SparkSession.scala:110) > at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:109) > at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply( > SparkSession.scala:878) > at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply( > SparkSession.scala:878) > at scala.collection.mutable.HashMap$$anonfun$foreach$1. > apply(HashMap.scala:99) > at scala.collection.mutable.HashMap$$anonfun$foreach$1. > apply(HashMap.scala:99) > at scala.collection.mutable.HashTable$class.foreachEntry( > HashTable.scala:230) > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > at scala.collection.mutable.HashMap.foreach(HashMap.scala:99) > at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession. > scala:878) > at org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$. > init(SparkSQLEnv.scala:47) > at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main( > HiveThriftServer2.scala:81) > at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main( > HiveThriftServer2.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:738) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: GSSException: No valid credentials provided (Mechanism level: > Failed to find any Kerberos tgt) > at sun.security.jgss.krb5.Krb5InitCredential.getInstance( > Krb5InitCredential.java:147) > at sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement( > Krb5MechFactory.java:122) > at sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext( > Krb5MechFactory.java:187) > at sun.security.jgss.GSSManagerImpl.getMechanismContext( > GSSManagerImpl.java:224) > at sun.security.jgss.GSSContextImpl.initSecContext( > GSSContextImpl.java:212) > at sun.security.jgss.GSSContextImpl.initSecContext( > GSSContextImpl.java:179) > at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge( > GssKrb5Client.java:192) > ... 75 more > 17/08/13 09:28:19 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.transport.TSaslTransport$NegotiationStatus - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/libthrift-0.9.2.jar!/org/apache/thrift/transport/TSaslTransport$ > NegotiationStatus.class > 17/08/13 09:28:19 DEBUG client.IsolatedClientLoader: shared class: > java.lang.Byte > 17/08/13 09:28:19 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.EncodingUtils - jar:file:/home/spark-2.1.0/ > assembly/target/scala-2.11/jars/libthrift-0.9.2.jar!/org/ > apache/thrift/EncodingUtils.class > 17/08/13 09:28:19 DEBUG client.IsolatedClientLoader: hive class: > org.apache.thrift.transport.TSaslTransport$SaslRole - > jar:file:/home/spark-2.1.0/assembly/target/scala-2.11/ > jars/libthrift-0.9.2.jar!/org/apache/thrift/transport/ > TSaslTransport$SaslRole.class > 17/08/13 09:28:19 DEBUG transport.TSaslTransport: CLIENT: Writing message > with status BAD and payload length 19 > 17/08/13 09:28:19 WARN hive.metastore: Failed to connect to the MetaStore > Server... > org.apache.thrift.transport.TTransportException: GSS initiate failed > at org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage( > TSaslTransport.java:232) > at org.apache.thrift.transport.TSaslTransport.open( > TSaslTransport.java:316) > at org.apache.thrift.transport.TSaslClientTransport.open( > TSaslClientTransport.java:37) > at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$ > 1.run(TUGIAssumingTransport.java:52) > at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$ > 1.run(TUGIAssumingTransport.java:49) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at org.apache.hadoop.security.UserGroupInformation.doAs( > UserGroupInformation.java:1698) > at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport. > open(TUGIAssumingTransport.java:49) > at org.apache.hadoop.hive.metastore.HiveMetaStoreClient. > open(HiveMetaStoreClient.java:420) > at org.apache.hadoop.hive.metastore.HiveMetaStoreClient. > <init>(HiveMetaStoreClient.java:236) > at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>( > SessionHiveMetaStoreClient.java:74) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.hadoop.hive.metastore.MetaStoreUtils. > newInstance(MetaStoreUtils.java:1521) > at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init> > (RetryingMetaStoreClient.java:86) > at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy( > RetryingMetaStoreClient.java:132) > at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy( > RetryingMetaStoreClient.java:104) > at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive. > java:3005) > at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024) > at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234) > at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174) > at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166) > at org.apache.hadoop.hive.ql.session.SessionState.start( > SessionState.java:503) > at org.apache.spark.sql.hive.client.HiveClientImpl.<init>( > HiveClientImpl.scala:192) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient( > IsolatedClientLoader.scala:264) > at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata( > HiveUtils.scala:366) > at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata( > HiveUtils.scala:270) > at org.apache.spark.sql.hive.HiveExternalCatalog.<init>( > HiveExternalCatalog.scala:65) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.spark.sql.internal.SharedState$.org$apache$spark$ > sql$internal$SharedState$$reflect(SharedState.scala:166) > at org.apache.spark.sql.internal.SharedState.<init>(SharedState.scala:86) > at org.apache.spark.sql.SparkSession$$anonfun$sharedState$1.apply( > SparkSession.scala:101) > at org.apache.spark.sql.SparkSession$$anonfun$sharedState$1.apply( > SparkSession.scala:101) > at scala.Option.getOrElse(Option.scala:121) > at org.apache.spark.sql.SparkSession.sharedState$ > lzycompute(SparkSession.scala:101) > at org.apache.spark.sql.SparkSession.sharedState(SparkSession.scala:100) > at org.apache.spark.sql.internal.SessionState.<init>( > SessionState.scala:157) > at org.apache.spark.sql.hive.HiveSessionState.<init>( > HiveSessionState.scala:32) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) > at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at org.apache.spark.sql.SparkSession$.org$apache$spark$sql$SparkSession$$ > reflect(SparkSession.scala:978) > at org.apache.spark.sql.SparkSession.sessionState$ > lzycompute(SparkSession.scala:110) > at org.apache.spark.sql.SparkSession.sessionState(SparkSession.scala:109) > at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply( > SparkSession.scala:878) > at org.apache.spark.sql.SparkSession$Builder$$anonfun$getOrCreate$5.apply( > SparkSession.scala:878) > at scala.collection.mutable.HashMap$$anonfun$foreach$1. > apply(HashMap.scala:99) > at scala.collection.mutable.HashMap$$anonfun$foreach$1. > apply(HashMap.scala:99) > at scala.collection.mutable.HashTable$class.foreachEntry( > HashTable.scala:230) > at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > at scala.collection.mutable.HashMap.foreach(HashMap.scala:99) > at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession. > scala:878) > at org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$. > init(SparkSQLEnv.scala:47) > at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main( > HiveThriftServer2.scala:81) > at org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main( > HiveThriftServer2.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$ > deploy$SparkSubmit$$runMain(SparkSubmit.scala:738) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > 17/08/13 09:28:19 INFO hive.metastore: Waiting 1 seconds before next > connection attempt. > 17/08/13 09:28:20 INFO hive.metastore: Trying to connect to metastore with > URI thrift://n2.server:9083 > 17/08/13 09:28:20 DEBUG security.UserGroupInformation: PrivilegedAction > as:spark (auth:KERBEROS) > > > >