[ 
https://issues.apache.org/jira/browse/KYLIN-4278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17036713#comment-17036713
 ] 

Yaqian Zhang commented on KYLIN-4278:
-------------------------------------

Hi:

Can you see if this is kylin's problem or hive's problem?

If it's Kylin's bug, can you reproduce it or provide a reproduction method?

> Build cube error in step 3. Connected to metastore, then MetaStoreClient lost 
> connection
> ----------------------------------------------------------------------------------------
>
>                 Key: KYLIN-4278
>                 URL: https://issues.apache.org/jira/browse/KYLIN-4278
>             Project: Kylin
>          Issue Type: Bug
>          Components: Metadata, Security
>    Affects Versions: v3.0.0-beta
>         Environment: hadoop 3.0
> hiveserver2
> hive metastore
> beeline
>            Reporter: xl_zl
>            Priority: Major
>             Fix For: Future
>
>
> When i build a  cube,  I encounter a strange issue in step 3(name=Extract 
> Fact Table Distinct Columns). Kylin connect to hive metastore and want to get 
> metadata,but  metastore server throw exception:{color:#FF0000}Error occurred 
> during processing of message. | {color}
> {color:#FF0000}java.lang.RuntimeException: 
> org.apache.thrift.transport.TTransportException: Invalid status -128{color}
> {color:#172b4d}==============metastore-server  error 
> logs:==============={color}
> {color:#172b4d}2019-12-04 17:50:10,180 | ERROR | pool-10-thread-173 | Error 
> occurred during processing of message. | 2019-12-04 17:50:10,180 | ERROR | 
> pool-10-thread-173 | Error occurred during processing of message. | 
> java.lang.RuntimeException: org.apache.thrift.transport.TTransportException: 
> Invalid status -128 at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java:694)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java:691)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_212] at 
> javax.security.auth.Subject.doAs(Subject.java:360) ~[?:1.8.0_212] at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1709)
>  ~[hadoop-common-3.1.1-mrs-2.0.jar:?] at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java:691)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:269)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  ~[?:1.8.0_212] at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  ~[?:1.8.0_212] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_212]Caused 
> by: org.apache.thrift.transport.TTransportException: Invalid status -128 at 
> org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslTransport.receiveSaslMessage(TSaslTransport.java:184)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslServerTransport.handleSaslStartMessage(TSaslServerTransport.java:125)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:271) 
> ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] ... 10 more2019-12-04 
> 17:50:10,399 | ERROR | pool-10-thread-173 | Error occurred during processing 
> of message. | java.lang.RuntimeException: 
> org.apache.thrift.transport.TTransportException: Invalid status -128 at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java:694)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory$1.run(HadoopThriftAuthBridge.java:691)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_212] at 
> javax.security.auth.Subject.doAs(Subject.java:360) ~[?:1.8.0_212] at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1709)
>  ~[hadoop-common-3.1.1-mrs-2.0.jar:?] at 
> org.apache.hadoop.hive.metastore.security.HadoopThriftAuthBridge$Server$TUGIAssumingTransportFactory.getTransport(HadoopThriftAuthBridge.java:691)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:269)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  ~[?:1.8.0_212] at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  ~[?:1.8.0_212] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_212]Caused 
> by: org.apache.thrift.transport.TTransportException: Invalid status -128 at 
> org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslTransport.receiveSaslMessage(TSaslTransport.java:184)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslServerTransport.handleSaslStartMessage(TSaslServerTransport.java:125)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:271) 
> ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] at 
> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216)
>  ~[hive-exec-3.1.0-mrs-2.0.jar:3.1.0-mrs-2.0] ... 10 more{color}
>  
>  
> ================kylin error logs:=================
> 2019-12-04 20:07:11,516 WARN  [Scheduler 1039145475 Job 
> c2cd7ddc-689a-864d-640b-1d73b18eb42f-113] metastore.HiveMetaStoreClient:645 : 
> set_ugi() not successful, Likely cause: new client talking to old server. 
> Continuing without it.2019-12-04 20:07:11,516 WARN  [Scheduler 1039145475 Job 
> c2cd7ddc-689a-864d-640b-1d73b18eb42f-113] metastore.HiveMetaStoreClient:645 : 
> set_ugi() not successful, Likely cause: new client talking to old server. 
> Continuing without it.org.apache.thrift.transport.TTransportException at 
> org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
>  at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) at 
> org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:380)
>  at 
> org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:230)
>  at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_set_ugi(ThriftHiveMetastore.java:4814)
>  at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.set_ugi(ThriftHiveMetastore.java:4800)
>  at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:637)
>  at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:228)
>  at 
> org.apache.hive.hcatalog.common.HiveClientCache$CacheableHiveMetaStoreClient.<init>(HiveClientCache.java:409)
>  at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>  at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>  at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at 
> org.apache.hadoop.hive.metastore.utils.JavaUtils.newInstance(JavaUtils.java:84)
>  at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:95)
>  at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:148)
>  at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133)
>  at 
> org.apache.hive.hcatalog.common.HiveClientCache$5.call(HiveClientCache.java:297)
>  at 
> org.apache.hive.hcatalog.common.HiveClientCache$5.call(HiveClientCache.java:292)
>  at 
> com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4767)
>  at 
> com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3568)
>  at com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2350) 
> at 
> com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2313)
>  at com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2228) at 
> com.google.common.cache.LocalCache.get(LocalCache.java:3965) at 
> com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4764) 
> at 
> org.apache.hive.hcatalog.common.HiveClientCache.getOrCreate(HiveClientCache.java:292)
>  at 
> org.apache.hive.hcatalog.common.HiveClientCache.get(HiveClientCache.java:274) 
> at 
> org.apache.hive.hcatalog.common.HCatUtil.getHiveMetastoreClient(HCatUtil.java:569)
>  at 
> org.apache.hive.hcatalog.mapreduce.InitializeInput.getInputJobInfo(InitializeInput.java:104)
>  at 
> org.apache.hive.hcatalog.mapreduce.InitializeInput.setInput(InitializeInput.java:88)
>  at 
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:95)
>  at 
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:51)
>  at 
> org.apache.kylin.source.hive.HiveMRInput$HiveTableInputFormat.configureJob(HiveMRInput.java:80)
>  at 
> org.apache.kylin.engine.mr.steps.FactDistinctColumnsJob.setupMapper(FactDistinctColumnsJob.java:126)
>  at 
> org.apache.kylin.engine.mr.steps.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:104)
>  at 
> org.apache.kylin.engine.mr.common.MapReduceExecutable.doWork(MapReduceExecutable.java:131)
>  at 
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
>  at 
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
>  at 
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
>  at 
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)2019-12-04 20:07:11,517 INFO  
> [Scheduler 1039145475 Job c2cd7ddc-689a-864d-640b-1d73b18eb42f-113] 
> metastore.HiveMetaStoreClient:673 : Connected to metastore.2019-12-04 
> 20:07:11,517 INFO  [Scheduler 1039145475 Job 
> c2cd7ddc-689a-864d-640b-1d73b18eb42f-113] 
> metastore.RetryingMetaStoreClient:97 : RetryingMetaStoreClient proxy=class 
> org.apache.hive.hcatalog.common.HiveClientCache$CacheableHiveMetaStoreClient 
> ugi=fiboadmin@0B019988_A1B0_4F31_AE11_6299A85F88FF.COM (auth:KERBEROS) 
> retries=1 delay=1 lifetime=02019-12-04 20:07:11,532 WARN  [Scheduler 
> 1039145475 Job c2cd7ddc-689a-864d-640b-1d73b18eb42f-113]{color:#de350b} 
> metastore.RetryingMetaStoreClient:257 : MetaStoreClient lost connection. 
> Attempting to reconnect (1 of 1) after 1s. getTable{color}
> org.apache.thrift.transport.TTransportException at 
> org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
>  at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) at 
> org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429) 
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318) 
> at 
> org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219)
>  at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_table_req(ThriftHiveMetastore.java:2083)
>  at 
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_table_req(ThriftHiveMetastore.java:2070)
>  at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1686)
>  at 
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1678)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:208)
>  at com.sun.proxy.$Proxy69.getTable(Unknown Source) at 
> org.apache.hive.hcatalog.common.HCatUtil.getTable(HCatUtil.java:191) at 
> org.apache.hive.hcatalog.mapreduce.InitializeInput.getInputJobInfo(InitializeInput.java:105)
>  at 
> org.apache.hive.hcatalog.mapreduce.InitializeInput.setInput(InitializeInput.java:88)
>  at 
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:95)
>  at 
> org.apache.hive.hcatalog.mapreduce.HCatInputFormat.setInput(HCatInputFormat.java:51)
>  at 
> org.apache.kylin.source.hive.HiveMRInput$HiveTableInputFormat.configureJob(HiveMRInput.java:80)
>  at 
> org.apache.kylin.engine.mr.steps.FactDistinctColumnsJob.setupMapper(FactDistinctColumnsJob.java:126)
>  at 
> org.apache.kylin.engine.mr.steps.FactDistinctColumnsJob.run(FactDistinctColumnsJob.java:104)
>  at 
> org.apache.kylin.engine.mr.common.MapReduceExecutable.doWork(MapReduceExecutable.java:131)
>  at 
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
>  at 
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
>  at 
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
>  at 
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to