Hi Venki, Thanks for the help, I have copied hive-exec-0.13 and this seems to resolve and issue. I'm now able to show table list.
However, I still cannot query the tables: 1. when I try to do a select from a table where data is in a csv file, I got an exception of "java.lang.IllegalArgumentException: Wrong FS". Full exception is below. I see the same kind of error in https://issues.apache.org/jira/browse/DRILL-1172, but the fix seems to be integrated to 0.4 already? 2. When I try to do a select from a table in parquet, sqlline freezes and I found an error “java.lang.NoSuchFieldError: doubleTypeInfo” in the log (full log is also below). The error comes from the a class in hive-exec jar that I copied from my Hive instance. As I can query the same table from Hive, is it possible that the version is not compatible? Thanks, Xiaohui 2014-08-25 14:31:34,414 [6a0ea065-4573-4947-8957-1a64837c4591:foreman] ERROR o.a.drill.exec.work.foreman.Foreman - Error d3dc25d6-f21e-4237-bdda-648f5ec76ce9: Failure while setting up Foreman. java.lang.IllegalArgumentException: Wrong FS: hdfs://dewdflhana1579.emea.global.corp.sap:8020/apps/hive/warehouse/smartmeter200m.db/weekdays_csv, expected: file:/// at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:390) ~[hadoop-core-1.2.1.jar:na] at org.apache.hadoop.fs.RawLocalFileSystem.pathToFile(RawLocalFileSystem.java:55) ~[hadoop-core-1.2.1.jar:na] at org.apache.hadoop.fs.LocalFileSystem.pathToFile(LocalFileSystem.java:61) ~[hadoop-core-1.2.1.jar:na] at org.apache.hadoop.fs.LocalFileSystem.exists(LocalFileSystem.java:51) ~[hadoop-core-1.2.1.jar:na] at org.apache.drill.exec.store.hive.HiveScan.getSplits(HiveScan.java:148) ~[drill-storage-hive-core-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.store.hive.HiveScan.<init>(HiveScan.java:111) ~[drill-storage-hive-core-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.store.hive.HiveStoragePlugin.getPhysicalScan(HiveStoragePlugin.java:75) ~[drill-storage-hive-core-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.store.hive.HiveStoragePlugin.getPhysicalScan(HiveStoragePlugin.java:39) ~[drill-storage-hive-core-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:53) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.planner.logical.DrillTable.getGroupScan(DrillTable.java:54) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.planner.logical.DrillPushProjIntoScan.onMatch(DrillPushProjIntoScan.java:53) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.eigenbase.relopt.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:223) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.relopt.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:661) ~[optiq-core-0.9-20140730.000241-5.jar:na] at net.hydromatic.optiq.tools.Programs$RuleSetProgram.run(Programs.java:165) ~[optiq-core-0.9-20140730.000241-5.jar:na] at net.hydromatic.optiq.prepare.PlannerImpl.transform(PlannerImpl.java:273) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToDrel(DefaultSqlHandler.java:145) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:126) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:128) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:403) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:219) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.work.WorkManager$RunnableWrapper.run(WorkManager.java:250) [drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [na:1.7.0_60] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [na:1.7.0_60] at java.lang.Thread.run(Thread.java:745) [na:1.7.0_60] 2014-08-25 14:33:45,232 [WorkManager-7] ERROR o.apache.drill.exec.work.WorkManager - Failure while running wrapper [Foreman: f0bd0c86-6404-4ae0-b03e-9e7ac8488ac0] java.lang.NoSuchFieldError: doubleTypeInfo at org.apache.hadoop.hive.ql.io.parquet.serde.ArrayWritableObjectInspector.getObjectInspector(ArrayWritableObjectInspector.java:67) ~[hive-exec-0.13.0.2.1.3.0-563.jar:0.13.0.2.1.3.0-563] at org.apache.hadoop.hive.ql.io.parquet.serde.ArrayWritableObjectInspector.<init>(ArrayWritableObjectInspector.java:60) ~[hive-exec-0.13.0.2.1.3.0-563.jar:0.13.0.2.1.3.0-563] at org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe.initialize(ParquetHiveSerDe.java:113) ~[hive-exec-0.13.0.2.1.3.0-563.jar:0.13.0.2.1.3.0-563] at org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:218) ~[hive-metastore-0.12.0.jar:0.12.0] at org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:272) ~[drill-hive-exec-shaded-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:265) ~[drill-hive-exec-shaded-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:597) ~[drill-hive-exec-shaded-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.store.hive.schema.DrillHiveTable.getRowType(DrillHiveTable.java:56) ~[drill-storage-hive-core-0.4.0-incubating-SNAPSHOT.jar:0.4.0-incubating-SNAPSHOT] at net.hydromatic.optiq.prepare.OptiqCatalogReader.getTableFrom(OptiqCatalogReader.java:94) ~[optiq-core-0.9-20140730.000241-5.jar:na] at net.hydromatic.optiq.prepare.OptiqCatalogReader.getTable(OptiqCatalogReader.java:76) ~[optiq-core-0.9-20140730.000241-5.jar:na] at net.hydromatic.optiq.prepare.OptiqCatalogReader.getTable(OptiqCatalogReader.java:42) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.EmptyScope.getTableNamespace(EmptyScope.java:67) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.IdentifierNamespace.validateImpl(IdentifierNamespace.java:75) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.AbstractNamespace.validate(AbstractNamespace.java:85) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateNamespace(SqlValidatorImpl.java:779) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateQuery(SqlValidatorImpl.java:768) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateFrom(SqlValidatorImpl.java:2599) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateSelect(SqlValidatorImpl.java:2802) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SelectNamespace.validateImpl(SelectNamespace.java:60) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.AbstractNamespace.validate(AbstractNamespace.java:85) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateNamespace(SqlValidatorImpl.java:779) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateQuery(SqlValidatorImpl.java:768) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.SqlSelect.validate(SqlSelect.java:208) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validateScopedExpression(SqlValidatorImpl.java:742) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.eigenbase.sql.validate.SqlValidatorImpl.validate(SqlValidatorImpl.java:458) ~[optiq-core-0.9-20140730.000241-5.jar:na] at net.hydromatic.optiq.prepare.PlannerImpl.validate(PlannerImpl.java:173) ~[optiq-core-0.9-20140730.000241-5.jar:na] at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.validateNode(DefaultSqlHandler.java:137) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:117) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:128) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:403) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:219) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at org.apache.drill.exec.work.WorkManager$RunnableWrapper.run(WorkManager.java:250) ~[drill-java-exec-0.4.0-incubating-SNAPSHOT-rebuffed.jar:0.4.0-incubating-SNAPSHOT] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [na:1.7.0_60] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [na:1.7.0_60] at java.lang.Thread.run(Thread.java:745) [na:1.7.0_60] -----Original Message----- From: Venki Korukanti [mailto:[email protected]] Sent: vendredi 8 août 2014 19:55 To: [email protected] Subject: Re: show tables results in query failure Hi, Parquet serde in Hive 0.12 is not native, you must be having a downloaded parquet serde in Hive lib or in Hive classpath. Drill is not able to find that class in its classpath, so throwing the exception. However for some reason the error message got truncated (it is missing the classnotfound exception part). To resolve this copy the parquet serde to Drill lib folder. Thanks Venki On Fri, Aug 8, 2014 at 8:55 AM, XUE, Xiaohui <[email protected]> wrote: > Hi, > > I'm trying to connect to hive from Apache Drill. I have successfully > connected to the metastore as the command "show schemas;" gives me all my > hive schemas. > > I have also successfully changed my schema to "hive. `default` ", but the > following command "show tables;" throw the exception below: > > Query failed: Failure while running fragment. > org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat > [aceb5ec8-0168-4c63-ac81-fa8fad9aed1d] > > Error: exception while executing query: Failure while trying to get next > result batch. (state=,code=0) > > Any hints of what could cause the issue? I have browsed my metastore and > all tables seems to be correctly defined. My metastore is mysql. > > Thanks, > Xiaohui >
