Hi igor ,

I tried setting "hive.server2.enable.doAs 
(https://link.getmailspring.com/link/[email protected]/0?redirect=hive.server2.enable.doAs&recipient=dXNlckBkcmlsbC5hcGFjaGUub3Jn)":
 "true" and ive also set the "hive.metastore.kerberos.principal" these 
properties i found while looking into mapr drill documentation. These settings 
helped e to create hive-storage plugin successfully. I was able to list 
databases present in hive datastore.
But while i was trying to list tables present in hive i was not able to get any 
response from metastore. When I looked into logs then i found that the drill is 
trying to execute a command in hive metastore
"get_tables_by_types"
method that is not available in the hive version that i am using.

the contents for drillbit.log are :

2019-11-29 18:27:08,900 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:foreman] INFO 
o.a.drill.exec.work.foreman.Foreman - Query text for query with id 
221e9d03-7c28-6462-bf0b-efe9eb6c4995 issued by drill: show tables in hive.geo
2019-11-29 18:27:08,939 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:frag:0:0] WARN 
o.a.d.e.s.h.c.TableNameCacheLoader - Failure while attempting to get hive 
tables. Retries once.
org.apache.hadoop.hive.metastore.api.MetaException: Got exception: 
org.apache.thrift.TApplicationException Invalid method name: 
'get_tables_by_type'
at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.logAndThrowMetaException(MetaStoreUtils.java:1382)
 ~[drill-hive-exec-shaded-1.16.0.jar:1.16.0]
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTables(HiveMetaStoreClient.java:1405)
 ~[drill-hive-exec-shaded-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.TableNameCacheLoader.load(TableNameCacheLoader.java:59)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.TableNameCacheLoader.load(TableNameCacheLoader.java:41)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3708)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2416)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2299)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2212)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache.get(LocalCache.java:4147)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:4151)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:5140)
 [drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.exec.store.hive.client.HiveMetadataCache.getTableNamesAndTypes(HiveMetadataCache.java:114)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.DrillHiveMetaStoreClient.getTableNamesAndTypes(DrillHiveMetaStoreClient.java:90)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.DrillHiveMetaStoreClientWithAuthorization.getTableNamesAndTypes(DrillHiveMetaStoreClientWithAuthorization.java:95)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.schema.HiveDatabaseSchema.ensureInitTables(HiveDatabaseSchema.java:76)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.schema.HiveDatabaseSchema.getTableNamesAndTypes(HiveDatabaseSchema.java:63)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator$Tables.visitTables(InfoSchemaRecordGenerator.java:340)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:254)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:247)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:247)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:234)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaTableType.getRecordReader(InfoSchemaTableType.java:58)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaBatchCreator.getBatch(InfoSchemaBatchCreator.java:34)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaBatchCreator.getBatch(InfoSchemaBatchCreator.java:30)
 [drill-java-exec-1.16.0.jar:1.16.0]
at org.apache.drill.exec.physical.impl.ImplCreator$2.run(ImplCreator.java:146) 
[drill-java-exec-1.16.0.jar:1.16.0]
at org.apache.drill.exec.physical.impl.ImplCreator$2.run(ImplCreator.java:142) 
[drill-java-exec-1.16.0.jar:1.16.0]
at java.security.AccessController.doPrivileged(Native Method) [na:1.8.0_151]
at javax.security.auth.Subject.doAs(Subject.java:422) [na:1.8.0_151]
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
 [hadoop-common-2.7.4.jar:na]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch(ImplCreator.java:142)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getChildren(ImplCreator.java:182)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch(ImplCreator.java:137)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getChildren(ImplCreator.java:182)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getRootExec(ImplCreator.java:110)
 [drill-java-exec-1.16.0.jar:1.16.0]
at org.apache.drill.exec.physical.impl.ImplCreator.getExec(ImplCreator.java:87) 
[drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.work.fragment.FragmentExecutor.run(FragmentExecutor.java:263)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.common.SelfCleaningRunnable.run(SelfCleaningRunnable.java:38) 
[drill-common-1.16.0.jar:1.16.0]
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
[na:1.8.0_151]
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
[na:1.8.0_151]
at java.lang.Thread.run(Thread.java:748) [na:1.8.0_151]
2019-11-29 18:27:09,098 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:frag:0:0] WARN 
o.a.d.e.s.h.s.HiveDatabaseSchema - Exception was thrown while getting table 
names and type for db 'geo'.
org.apache.thrift.TException: java.util.concurrent.ExecutionException: 
MetaException(message:Got exception: org.apache.thrift.TApplicationException 
Invalid method name: 'get_tables_by_type')
at 
org.apache.drill.exec.store.hive.client.HiveMetadataCache.getTableNamesAndTypes(HiveMetadataCache.java:116)
 ~[drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.DrillHiveMetaStoreClient.getTableNamesAndTypes(DrillHiveMetaStoreClient.java:90)
 ~[drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.DrillHiveMetaStoreClientWithAuthorization.getTableNamesAndTypes(DrillHiveMetaStoreClientWithAuthorization.java:95)
 ~[drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.schema.HiveDatabaseSchema.ensureInitTables(HiveDatabaseSchema.java:76)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.schema.HiveDatabaseSchema.getTableNamesAndTypes(HiveDatabaseSchema.java:63)
 [drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator$Tables.visitTables(InfoSchemaRecordGenerator.java:340)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:254)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:247)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:247)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaRecordGenerator.scanSchema(InfoSchemaRecordGenerator.java:234)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaTableType.getRecordReader(InfoSchemaTableType.java:58)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaBatchCreator.getBatch(InfoSchemaBatchCreator.java:34)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.ischema.InfoSchemaBatchCreator.getBatch(InfoSchemaBatchCreator.java:30)
 [drill-java-exec-1.16.0.jar:1.16.0]
at org.apache.drill.exec.physical.impl.ImplCreator$2.run(ImplCreator.java:146) 
[drill-java-exec-1.16.0.jar:1.16.0]
at org.apache.drill.exec.physical.impl.ImplCreator$2.run(ImplCreator.java:142) 
[drill-java-exec-1.16.0.jar:1.16.0]
at java.security.AccessController.doPrivileged(Native Method) [na:1.8.0_151]
at javax.security.auth.Subject.doAs(Subject.java:422) [na:1.8.0_151]
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
 [hadoop-common-2.7.4.jar:na]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch(ImplCreator.java:142)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getChildren(ImplCreator.java:182)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getRecordBatch(ImplCreator.java:137)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getChildren(ImplCreator.java:182)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.physical.impl.ImplCreator.getRootExec(ImplCreator.java:110)
 [drill-java-exec-1.16.0.jar:1.16.0]
at org.apache.drill.exec.physical.impl.ImplCreator.getExec(ImplCreator.java:87) 
[drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.work.fragment.FragmentExecutor.run(FragmentExecutor.java:263)
 [drill-java-exec-1.16.0.jar:1.16.0]
at 
org.apache.drill.common.SelfCleaningRunnable.run(SelfCleaningRunnable.java:38) 
[drill-common-1.16.0.jar:1.16.0]
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
[na:1.8.0_151]
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
[na:1.8.0_151]
at java.lang.Thread.run(Thread.java:748) [na:1.8.0_151]
Caused by: java.util.concurrent.ExecutionException: MetaException(message:Got 
exception: org.apache.thrift.TApplicationException Invalid method name: 
'get_tables_by_type')
at 
org.apache.drill.shaded.guava.com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:502)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:461)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.util.concurrent.AbstractFuture$TrustedFuture.get(AbstractFuture.java:83)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.util.concurrent.Uninterruptibles.getUninterruptibly(Uninterruptibles.java:142)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.getAndRecordStats(LocalCache.java:2453)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2417)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2299)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2212)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache.get(LocalCache.java:4147)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:4151)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:5140)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.exec.store.hive.client.HiveMetadataCache.getTableNamesAndTypes(HiveMetadataCache.java:114)
 ~[drill-storage-hive-core-1.16.0.jar:1.16.0]
... 28 common frames omitted
Caused by: org.apache.hadoop.hive.metastore.api.MetaException: Got exception: 
org.apache.thrift.TApplicationException Invalid method name: 
'get_tables_by_type'
at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.logAndThrowMetaException(MetaStoreUtils.java:1382)
 ~[drill-hive-exec-shaded-1.16.0.jar:1.16.0]
at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTables(HiveMetaStoreClient.java:1405)
 ~[drill-hive-exec-shaded-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.TableNameCacheLoader.load(TableNameCacheLoader.java:71)
 ~[drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.exec.store.hive.client.TableNameCacheLoader.load(TableNameCacheLoader.java:41)
 ~[drill-storage-hive-core-1.16.0.jar:1.16.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3708)
 ~[drill-shaded-guava-23.0.jar:23.0]
at 
org.apache.drill.shaded.guava.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2416)
 ~[drill-shaded-guava-23.0.jar:23.0]
... 34 common frames omitted
2019-11-29 18:27:09,100 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:frag:0:0] INFO 
o.a.d.e.w.fragment.FragmentExecutor - 221e9d03-7c28-6462-bf0b-efe9eb6c4995:0:0: 
State change requested AWAITING_ALLOCATION --> RUNNING
2019-11-29 18:27:09,100 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:frag:0:0] INFO 
o.a.d.e.w.f.FragmentStatusReporter - 221e9d03-7c28-6462-bf0b-efe9eb6c4995:0:0: 
State to report: RUNNING
2019-11-29 18:27:09,103 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:frag:0:0] INFO 
o.a.d.e.w.fragment.FragmentExecutor - 221e9d03-7c28-6462-bf0b-efe9eb6c4995:0:0: 
State change requested RUNNING --> FINISHED
2019-11-29 18:27:09,104 [221e9d03-7c28-6462-bf0b-efe9eb6c4995:frag:0:0] INFO 
o.a.d.e.w.f.FragmentStatusReporter - 221e9d03-7c28-6462-bf0b-efe9eb6c4995:0:0: 
State to report: FINISHED

What i want to know is what are the versions of hive compatible with the 
different versions of hive?
If possible do share the compatibility of all the components like hive, kafka, 
hbase with different versions of drill?

Thanks and Regards
Rameshwar Mane
Big Data Engineer

On Nov 29 2019, at 6:30 pm, Igor Guzenko <[email protected]> wrote:
> Hi Rameshwar,
>
> I found difference in your plugin configuration, comparing to
> https://mapr.com/docs/51/SecurityGuide/Modify-Hive-Storage-Plugin-In-Drill.html
> .
> "hive.server2.enable.doAs": "false" .
> Also hive.metastore.kerberos.principal is absent in your configs. In my
> opinion it's minor configuration
> issue and to resolve it you just need to check every step patiently. Also
> maybe some MapR docs could be helpful for you
> https://mapr.com/docs/51/SecurityGuide/Configure-Kerberos-Authentication.html
> .
>
> Thanks,
> Igor
>
> On Fri, Nov 29, 2019 at 6:24 AM Rameshwar Mane <[email protected]> wrote:
> > The hive metastore is up and running, the metastore is accessible from the
> > machine where drill is installed. I have followed all the steps present in
> > the documentation.
> >
> > On Fri, 29 Nov 2019, 05:17 Igor Guzenko, <[email protected]>
> > wrote:
> >
> > > Hello Rameshwar,
> > > I think the issue is a little bit more complicated, at least you need to
> > be
> > > sure that hive metastore is
> > > up and running, and is accessible from given machine. Could you please
> > > create jira ticket containing
> > > configuration info as much as possible ? (hive-site.xml, storage plugin
> > > config, etc.) So then development team could try to reproduce & debug
> > > the problem.
> > >
> > > Thank you in advance,
> > > Igor
> > >
> > > On Thu, Nov 28, 2019 at 7:58 PM Rameshwar Mane <[email protected]>
> > > wrote:
> > >
> > > > Can anyone provide me any inputs regarding this issue???
> > > > On Thu, 28 Nov 2019, 17:35 Rameshwar Mane, <[email protected]>
> > wrote:
> > > >
> > > > > Hi all,
> > > > > I am trying to use drill to query a kerberized cluster. I have setup
> > > > drill
> > > > > with the provided in documentation for kerberized cluster.
> > > > >
> > > > > the storage plugin details i am trying to create for hive is :
> > > > > {
> > > > > "type": "hive",
> > > > > "configProps": {
> > > > > "hive.metastore.uris":
> > > >
> > > > "thrift://xxxx-xxx-xxxx:9083,thrift://xxxx-xxx-xxxx:9083",
> > > > > "hive.metastore.warehouse.dir":
> > > >
> > >
> > > "/warehouse/tablespace/managed/hive",
> > > > > "fs.default.name": "hdfs://xxxx-xxx-xxxx:8020",
> > > > > "hive.security.authorization.enabled": "true",
> > > > > "hive.security.authenticator.manager":
> > > >
> > > > "org.apache.hadoop.hive.ql.security.SessionStateUserAuthenticator",
> > > > > "hive.security.authorization.manager":
> > > >
> > > >
> > >
> > "org.apcahe.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdHiveAuthorizerFactory",
> > > > > "hive.metastore.sasl.enabled": "true",
> > > > > "hive.server2.enable.doAs": "true"
> > > > > },
> > > > > "enabled": true
> > > > > }
> > > > >
> > > > >
> > > > > i am facing the following issue while trying to create hive storage
> > > > > plugin:
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > Please retry: Error while creating / updating storage : Could not
> > > > connect to meta store using any of the URIs provided.
> > > > > Most recent failure:
> > > >
> > >
> >
> > org.apache.thrift.transport.TTransportException:
> > > > > GSS initiate failed at
> > > >
> > > >
> > >
> > org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232)
> > > > at
> > >
> > > org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:316)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37)
> > > > at
> > > >
> > >
> > org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49)
> > > > at java.security.AccessController.doPrivileged(Native Method)
> > > > > at javax.security.auth.Subject.doAs(Subject.java:422) at
> > > >
> > > >
> > >
> > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49)
> > > > at
> > > >
> > >
> > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:480)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:247)
> > > > at
> > > >
> > >
> > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:129)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.drill.exec.store.hive.client.DrillHiveMetaStoreClient.<init>(DrillHiveMetaStoreClient.java:54)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.drill.exec.store.hive.client.DrillHiveMetaStoreClientFactory.createCloseableClientWithCaching(DrillHiveMetaStoreClientFactory.java:101)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.drill.exec.store.hive.schema.HiveSchemaFactory.<init>(HiveSchemaFactory.java:77)
> > > > at
> > > >
> > >
> > org.apache.drill.exec.store.hive.HiveStoragePlugin.<init>(HiveStoragePlugin.java:77)
> > > > at
> > > > > sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> > > >
> > >
> >
> > Method)
> > > > at
> > > >
> > >
> > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> > > > at
> > > > >
> > > >
> > >
> > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> > > > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at
> > > > >
> > > >
> > >
> > org.apache.drill.exec.store.StoragePluginRegistryImpl.create(StoragePluginRegistryImpl.java:466)
> > > > at
> > > >
> > >
> > org.apache.drill.exec.store.StoragePluginRegistryImpl.createOrUpdate(StoragePluginRegistryImpl.java:131)
> > > > > at
> > > >
> > > >
> > >
> > org.apache.drill.exec.server.rest.PluginConfigWrapper.createOrUpdateInStorage(PluginConfigWrapper.java:56)
> > > > at
> > > >
> > >
> > org.apache.drill.exec.server.rest.StorageResources.createOrUpdatePluginJSON(
> > > > StorageResources.java:193) at
> > > > >
> > > >
> > >
> > org.apache.drill.exec.server.rest.StorageResources.createOrUpdatePlugin(StorageResources.java:210)
> > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
> > > > >
> > > >
> > >
> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> > > > at
> > > >
> > >
> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> > > > at
> > > > > java.lang.reflect.Method.invoke(Method.java:498) at
> > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.internal.ResourceMethodInvocationHandlerFactory$1.invoke(ResourceMethodInvocationHandlerFactory.java:81)
> > > > at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher$1.run(AbstractJavaResourceMethodDispatcher.java:144)
> > > > at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.invoke(AbstractJavaResourceMethodDispatcher.java:161)
> > > > at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$TypeOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:205)
> > > > at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:99)
> > > > at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:389)
> > > > at
> > > >
> > >
> > org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:347)
> > > > at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:102)
> > > > at
> > >
> > > org.glassfish.jersey.server.ServerRuntime$2.run(ServerRuntime.java:326)
> > > > at
> > > > > org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at
> > > >
> > > > org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at
> > > > org.glassfish.jersey.internal.Errors.process(Errors.java:315) at
> > > > > org.glassfish.jersey.internal.Errors.process(Errors.java:297) at
> > > >
> > > > org.glassfish.jersey.internal.Errors.process(Errors.java:267) at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317)
> > > > at
> > > >
> > >
> > org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:305) at
> > > > >
> > > >
> > >
> > org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:1154)
> > > > at
> > > >
> > >
> > org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:473)
> > > > at
> > > > >
> > > >
> > org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:427)
> > > at
> > > >
> > >
> > org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:388)
> > > > at \
> > > > >
> > > >
> > >
> > org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:341)
> > > > at
> > > >
> > >
> > org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:228)
> > > > at
> > > > >
> > org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:848)
> > > > at
> > > >
> > >
> > org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:585)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143)
> > > > at
> > > >
> > >
> > org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:524)
> > > > at
> > > > >
> > > >
> > >
> > org.apache.drill.exec.server.rest.auth.DrillHttpSecurityHandlerProvider.handle(DrillHttpSecurityHandlerProvider.java:151)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:226)
> > > > at
> > > >
> > >
> > org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180)
> > > > at
> > > > >
> > > >
> > org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:513)
> > > > at
> > > >
> > >
> > org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:185)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112)
> > > > at
> > > >
> > >
> > org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
> > > > at org.eclipse.jetty.server.Server.handle(Server.java:539) at
> > > > > org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:333) at
> > > >
> > > >
> > >
> > org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:251)
> > > > at
> > > > > org.eclipse.jetty.io
> > > >
> > >
> > > .AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:283)
> > > > at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:108)
> > >
> >
> > at
> > > > > org.eclipse.jetty.io
> > > >
> > >
> > > .SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceConsume(ExecuteProduceConsume.java:148)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:136)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:671)
> > > > at
> > > > >
> > > >
> > >
> > org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:589)
> > > > at
> > > > > java.lang.Thread.run(Thread.java:748)
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > am i providing any wrong configurations while creating the storage
> > > > plugin?
> > > > > I have created drill keytab using the steps provided in the
> > > >
> > >
> > > documentation
> > > > > and have a mit kerberos autorization service.
> > > > >
> > > > > let me know any solutions that can be used to solve this.
> > > > > Thanks and Regards
> > > > > *Rameshwar Mane*
> > > > > Big Data Engineer
> > > >
> > >
> >
>
>

Reply via email to