[jira] [Created] (DRILL-6293) Unable to read hive(2.1.1) tables using Drill 1.13.0

2018-03-26 Thread Anup Tiwari (JIRA)
Anup Tiwari created DRILL-6293:
--

 Summary: Unable to read hive(2.1.1) tables using Drill 1.13.0 
 Key: DRILL-6293
 URL: https://issues.apache.org/jira/browse/DRILL-6293
 Project: Apache Drill
  Issue Type: Bug
Affects Versions: 1.13.0
Reporter: Anup Tiwari


Hi,

{color:#22}I am not able to read my hive tables in drill 1.13.0 and with 
same plugin conf it was working in Drill 1.12.0 and 1.10.0.{color}
 
*Hive Plugin :-* 
{
  "type": "hive",
  "enabled": true,
  "configProps": {
"hive.metastore.uris": 
"thrift://[prod-hadoop-1xx.com:9083|http://prod-hadoop-1xx.com:9083/]";,
"hive.metastore.sasl.enabled": "false",
"[fs.default.name|http://fs.default.name/]": 
"hdfs://[prod-hadoop-1xx.com:9000|http://prod-hadoop-1xx.com:9000/]";
  }
} 
 
*Query :-* 
select id from hive.cad where log_date = '2018-03-18' limit 3;
 
*Error :-* 
2018-03-20 14:25:27,351 [254f337f-9ac3-b66f-ed17-1de459da3283:foreman] INFO 
o.a.drill.exec.work.foreman.Foreman - Query text for query id 
254f337f-9ac3-b66f-ed17-1de459da3283: select id from hive.cad where log_date = 
'2018-03-18' limit 3
2018-03-20 14:25:27,354 [254f337f-9ac3-b66f-ed17-1de459da3283:foreman] WARN 
o.a.d.e.s.h.DrillHiveMetaStoreClient - Failure while attempting to get hive 
table. Retries once.
org.apache.thrift.TApplicationException: Invalid method name: 'get_table_req'
 at 
org.apache.thrift.TApplicationException.read(TApplicationException.java:111) 
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0]
 at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:79) 
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0]
 at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_table_req(ThriftHiveMetastore.java:1563)
 ~[drill-hive-exec-shaded-1.13.0.jar:1.13.0]
 at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_table_req(ThriftHiveMetastore.java:1550)
 ~[drill-hive-exec-shaded-1.13.0.jar:1.13.0]
 at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1344)
 ~[drill-hive-exec-shaded-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient.getHiveReadEntryHelper(DrillHiveMetaStoreClient.java:285)
 ~[drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient$TableLoader.load(DrillHiveMetaStoreClient.java:535)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient$TableLoader.load(DrillHiveMetaStoreClient.java:531)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
 [guava-18.0.jar:na]
 at com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319) 
[guava-18.0.jar:na]
 at 
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
 [guava-18.0.jar:na]
 at com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197) 
[guava-18.0.jar:na]
 at com.google.common.cache.LocalCache.get(LocalCache.java:3937) 
[guava-18.0.jar:na]
 at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3941) 
[guava-18.0.jar:na]
 at 
com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4824) 
[guava-18.0.jar:na]
 at 
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient$HiveClientWithCaching.getHiveReadEntry(DrillHiveMetaStoreClient.java:495)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.schema.HiveSchemaFactory$HiveSchema.getSelectionBaseOnName(HiveSchemaFactory.java:233)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.schema.HiveSchemaFactory$HiveSchema.getDrillTable(HiveSchemaFactory.java:213)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.schema.HiveDatabaseSchema.getTable(HiveDatabaseSchema.java:62)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.drill.exec.store.hive.schema.HiveSchemaFactory$HiveSchema.getTable(HiveSchemaFactory.java:201)
 [drill-storage-hive-core-1.13.0.jar:1.13.0]
 at 
org.apache.calcite.jdbc.SimpleCalciteSchema.getImplicitTable(SimpleCalciteSchema.java:82)
 [calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0]
 at org.apache.calcite.jdbc.CalciteSchema.getTable(CalciteSchema.java:257) 
[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0]
 at 
org.apache.calcite.sql.validate.SqlValidatorUtil.getTableEntryFrom(SqlValidatorUtil.java:1003)
 [calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0]
 at 
org.apache.calcite.sql.validate.SqlValidatorUtil.getTableEntry(SqlValidatorUtil.java:960)
 [calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0]
 at 
org.apache.calcite.prepare.CalciteCatalogReader.getTable(CalciteCatalogReader.java:117)
 [calcite-core-1.15.0-drill-r0

Re: [Drill 1.13.0] : org.apache.thrift.TApplicationException: Invalid method name: 'get_table_req'

2018-03-20 Thread Anup Tiwari
Note :  Using Show databases, i can see hive schemas.  





On Tue, Mar 20, 2018 2:36 PM, Anup Tiwari anup.tiw...@games24x7.com  wrote:
Hi,
I am not able to read my hive tables in drill 1.13.0 and with same plugin conf
it was working in Drill 1.12.0 and 1.10.0. Please look into it asap and let me
know if i have missed anything.
Hive Plugin :-
{  "type": "hive",  "enabled": true,  "configProps": {"hive.metastore.uris":
"thrift://prod-hadoop-1xx.com:9083","hive.metastore.sasl.enabled": "false",
"fs.default.name": "hdfs://prod-hadoop-1xx.com:9000"  }}
Query :-
select id from hive.cad where log_date = '2018-03-18' limit 3
Error :-
2018-03-20 14:25:27,351 [254f337f-9ac3-b66f-ed17-1de459da3283:foreman] INFO
o.a.drill.exec.work.foreman.Foreman - Query text for query id
254f337f-9ac3-b66f-ed17-1de459da3283: select id from hive.cad where log_date =
'2018-03-18' limit 32018-03-20 14:25:27,354
[254f337f-9ac3-b66f-ed17-1de459da3283:foreman] WARN
o.a.d.e.s.h.DrillHiveMetaStoreClient - Failure while attempting to get hive
table. Retries once.org.apache.thrift.TApplicationException: Invalid method
name: 'get_table_req' at
org.apache.thrift.TApplicationException.read(TApplicationException.java:111)
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0] at
org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:79)
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0] at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_table_req(ThriftHiveMetastore.java:1563)
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0] at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_table_req(ThriftHiveMetastore.java:1550)
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0] at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1344)
~[drill-hive-exec-shaded-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient.getHiveReadEntryHelper(DrillHiveMetaStoreClient.java:285)
~[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient$TableLoader.load(DrillHiveMetaStoreClient.java:535)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient$TableLoader.load(DrillHiveMetaStoreClient.java:531)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3527)
[guava-18.0.jar:na] at
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2319)
[guava-18.0.jar:na] at
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2282)
[guava-18.0.jar:na] at
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2197)
[guava-18.0.jar:na] at
com.google.common.cache.LocalCache.get(LocalCache.java:3937) [guava-18.0.jar:na]
 at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3941)
[guava-18.0.jar:na] at
com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4824)
[guava-18.0.jar:na] at
org.apache.drill.exec.store.hive.DrillHiveMetaStoreClient$HiveClientWithCaching.getHiveReadEntry(DrillHiveMetaStoreClient.java:495)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.schema.HiveSchemaFactory$HiveSchema.getSelectionBaseOnName(HiveSchemaFactory.java:233)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.schema.HiveSchemaFactory$HiveSchema.getDrillTable(HiveSchemaFactory.java:213)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.schema.HiveDatabaseSchema.getTable(HiveDatabaseSchema.java:62)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.drill.exec.store.hive.schema.HiveSchemaFactory$HiveSchema.getTable(HiveSchemaFactory.java:201)
[drill-storage-hive-core-1.13.0.jar:1.13.0] at
org.apache.calcite.jdbc.SimpleCalciteSchema.getImplicitTable(SimpleCalciteSchema.java:82)
[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.calcite.jdbc.CalciteSchema.getTable(CalciteSchema.java:257)
[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.calcite.sql.validate.SqlValidatorUtil.getTableEntryFrom(SqlValidatorUtil.java:1003)
[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.calcite.sql.validate.SqlValidatorUtil.getTableEntry(SqlValidatorUtil.java:960)
[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.calcite.prepare.CalciteCatalogReader.getTable(CalciteCatalogReader.java:117)
[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.drill.exec.planner.sql.SqlConverter$DrillCalciteCatalogReader.getTable(SqlConverter.java:633)
[drill-java-exec-1.13.0.jar:1.13.0] at
org.apache.drill.exec.planner.sql.SqlConverter$DrillValidator.validateFrom(SqlConverter.java:261)
[drill-java-exec-1.13.0.jar:1.13.0] at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateSelect(SqlValidator

[Drill 1.13.0] : org.apache.thrift.TApplicationException: Invalid method name: 'get_table_req'

2018-03-20 Thread Anup Tiwari
SqlValidatorImpl.validate(SqlValidatorImpl.java:613)
~[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.drill.exec.planner.sql.SqlConverter.validate(SqlConverter.java:190)
[drill-java-exec-1.13.0.jar:1.13.0] ... 10 common frames omittedCaused by:
org.apache.calcite.sql.validate.SqlValidatorException: Object 'cad' not found
within 'hive' at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[na:1.8.0_72] at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
~[na:1.8.0_72] at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[na:1.8.0_72] at
java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[na:1.8.0_72]
 at org.apache.calcite.runtime.Resources$ExInstWithCause.ex(Resources.java:463)
~[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] at
org.apache.calcite.runtime.Resources$ExInst.ex(Resources.java:572)
~[calcite-core-1.15.0-drill-r0.jar:1.15.0-drill-r0] ... 31 common frames omitted
2018-03-20 14:25:27,375 [254f337f-9ac3-b66f-ed17-1de459da3283:foreman] INFO
o.apache.drill.exec.work.WorkManager - Waiting for 0 queries to complete before
shutting down2018-03-20 14:25:27,375
[254f337f-9ac3-b66f-ed17-1de459da3283:foreman] INFO
o.apache.drill.exec.work.WorkManager - Waiting for 0 running fragments to
complete before shutting down
Regards,
Anup Tiwari

Re: [Drill 1.12.0] : Suggestions on Downgrade to 1.11.0  & com.mysql.jdbc.exceptions.jdbc4.CommunicationsException

2018-03-16 Thread Anup Tiwari
Hi All,
We checked our MySQL max number of connections which is set to 200 and i think
this might be due to exceeding max number of connections only as right now i can
see 89 connections to MySQL.
I want to know community's thoughts on this whether i am heading in right
direction or not.  





On Fri, Mar 16, 2018 1:03 PM, Anup Tiwari anup.tiw...@games24x7.com  wrote:
Hi All,
We are getting a lot of different type of issues/error post upgrading from Drill
1.10.0 to 1.12.0 which i am asking on forum as well so just wanted to know
whether downgrading to Drill 1.11.0 will help or not?
This time we got exception related to mysql connection storage and please note
that this issue is not consistent i.e. if i execute this query after some time
then it works. Please find below query are error logs.
Query :
create table dfs.tmp.table_info as select * from mysql.test.table_info;
Error :
WARN o.a.d.e.store.jdbc.JdbcStoragePlugin - Failure while attempting to load
JDBC schema.com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: The last
packet successfully received from the server was 49,949,177 milliseconds ago. 
The last packet sent successfully to the server was 49,949,196 milliseconds ago.
is longer than the server configured value of 'wait_timeout'. You should
consider either expiring and/or testing connection validity before use in your
application, increasing the server configured values for client timeouts, or
using the Connector/J connection property 'autoReconnect=true' to avoid this
problem.at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[na:1.8.0_72]at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
~[na:1.8.0_72]at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[na:1.8.0_72]at
java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[na:1.8.0_72]
at com.mysql.jdbc.Util.handleNewInstance(Util.java:389)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:1038)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.MysqlIO.send(MysqlIO.java:3609)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2417)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2582)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2531)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2489)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.StatementImpl.executeQuery(StatementImpl.java:1446)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
com.mysql.jdbc.DatabaseMetaData.getCatalogs(DatabaseMetaData.java:2025)
~[mysql-connector-java-5.1.35-bin.jar:5.1.35]at
org.apache.commons.dbcp.DelegatingDatabaseMetaData.getCatalogs(DelegatingDatabaseMetaData.java:190)
~[commons-dbcp-1.4.jar:1.4]at
org.apache.drill.exec.store.jdbc.JdbcStoragePlugin$JdbcCatalogSchema.(JdbcStoragePlugin.java:309)
~[drill-jdbc-storage-1.12.0.jar:1.12.0]at
org.apache.drill.exec.store.jdbc.JdbcStoragePlugin.registerSchemas(JdbcStoragePlugin.java:430)
[drill-jdbc-storage-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.DynamicRootSchema.loadSchemaFactory(DynamicRootSchema.java:94)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.DynamicRootSchema.getSubSchema(DynamicRootSchema.java:74)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.calcite.prepare.CalciteCatalogReader.getSchema(CalciteCatalogReader.java:160)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.prepare.CalciteCatalogReader.getTableFrom(CalciteCatalogReader.java:114)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.prepare.CalciteCatalogReader.getTable(CalciteCatalogReader.java:108)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.drill.exec.planner.sql.SqlConverter$DrillCalciteCatalogReader.getTable(SqlConverter.java:493)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.SqlConverter$DrillCalciteCatalogReader.getTable(SqlConverter.java:434)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.calcite.sql.validate.EmptyScope.getTableNamespace(EmptyScope.java:75)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.DelegatingScope.getTableNamespace(DelegatingScope.java:124)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.IdentifierNamespace.validateImpl(IdentifierNamespace.java:104)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.A

[Drill 1.12.0] : Suggestions on Downgrade to 1.11.0  & com.mysql.jdbc.exceptions.jdbc4.CommunicationsException

2018-03-16 Thread Anup Tiwari
drill-r23]at
org.apache.calcite.sql.validate.IdentifierNamespace.validateImpl(IdentifierNamespace.java:104)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.AbstractNamespace.validate(AbstractNamespace.java:86)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateNamespace(SqlValidatorImpl.java:886)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateQuery(SqlValidatorImpl.java:872)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateFrom(SqlValidatorImpl.java:2817)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateFrom(SqlValidatorImpl.java:2802)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateSelect(SqlValidatorImpl.java:3025)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SelectNamespace.validateImpl(SelectNamespace.java:60)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.AbstractNamespace.validate(AbstractNamespace.java:86)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateNamespace(SqlValidatorImpl.java:886)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateQuery(SqlValidatorImpl.java:872)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.SqlSelect.validate(SqlSelect.java:210)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validateScopedExpression(SqlValidatorImpl.java:846)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.calcite.sql.validate.SqlValidatorImpl.validate(SqlValidatorImpl.java:560)
[calcite-core-1.4.0-drill-r23.jar:1.4.0-drill-r23]at
org.apache.drill.exec.planner.sql.SqlConverter.validate(SqlConverter.java:172)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.validateNode(DefaultSqlHandler.java:617)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.validateAndConvert(DefaultSqlHandler.java:192)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.handlers.CreateTableHandler.getPlan(CreateTableHandler.java:77)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1017)
[drill-java-exec-1.12.0.jar:1.12.0]at
org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:289)
[drill-java-exec-1.12.0.jar:1.12.0]at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_72]at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_72]at java.lang.Thread.run(Thread.java:745) [na:1.8.0_72]
Let me know what to do here.
Regards,
Anup Tiwari

Re: Does s3 plugin support AWS S3 signature version 4 ?

2017-05-19 Thread Anup Tiwari
Any updates on this?
Since we have migrated to Aws Mumbai, we are not able to connect s3 and
Drill.

On 04-Apr-2017 11:02 PM, "Shankar Mane"  wrote:

> Quick question here:
>
> Does s3 plugin support S3 signature version 4  ?
>
> FYI: s3 plugin works in case when region has support for both v2 and v4
> signature. Whereas it seems problematic, for regions (eg. ap-south-1) which
> only has v4 signature version support.
>
> regards,
> shankar
>


Re: Running cartesian joins on Drill

2017-05-11 Thread Anup Tiwari
Hi,

I have one question here.. so if we have to use Cartesian join in Drill
then do we have to follow some workaround like Shadi mention : adding a
dummy column on the fly that has the value 1 in both tables and then join
on that column leading to having a match of every row of the first table
with every row of the second table, hence do a Cartesian product?
OR
If we just don't specify join condition like :
select a.*, b.* from tt1 as a, tt2 b; then will it internally treat this
query as Cartesian join.

Regards,
*Anup Tiwari*

On Mon, May 8, 2017 at 10:00 PM, Zelaine Fong  wrote:

> Cartesian joins in Drill are implemented as nested loop joins, and I think
> you should see that reflected in the resultant query plan when you run
> explain plan on the query.
>
> Yes, Cartesian joins/nested loop joins are expensive because you’re
> effectively doing an MxN read of your tables.  There are more efficient
> ways of processing a nested loop join, e.g., by creating an index on the
> larger table in the join and then using that index to do lookups into that
> table.  That way, the nested loop join cost is the cost of creating the
> index + M, where M is the number of rows in the smaller table and assuming
> the lookup cost into the index does minimize the amount of data read of the
> second table.  Drill currently doesn’t do this.
>
> -- Zelaine
>
> On 5/8/17, 9:09 AM, "Muhammad Gelbana"  wrote:
>
> ​I believe ​clhubert is referring to this discussion
> <http://drill-user.incubator.apache.narkive.com/TIXWiTY4/
> cartesian-product-in-apache-drill#post1>
> .
>
> So why Drill doesn't transform this query into a nested join query ?
> Simply
> because there is no Calcite rule to transform it into a nested loop
> join ?
> Is it not technically possible to write such Rule or is it feasible so
> I
> may take on this challenge ?
>
> Also pardon me for repeating my question but I fail to find an answer
> in
> your replies, why doesn't Drill just run a cartesian join ? Because
> it's
> expensive regarding resources (i.e. CPU\Network\RAM) ?
>
> Thanks a lot Shadi for the query, it works for me.
>
> *-*
> *Muhammad Gelbana*
> http://www.linkedin.com/in/mgelbana
>
> On Mon, May 8, 2017 at 6:10 AM, Shadi Khalifa 
> wrote:
>
> > Hi Muhammad,
> >
> > I did the following as a workaround to have Cartesian product. The
> basic
> > idea is to create a dummy column on the fly that has the value 1 in
> both
> > tables and then join on that column leading to having a match of
> every row
> > of the first table with every row of the second table, hence do a
> Cartesian
> > product. This might not be the most efficient way but it will do the
> job.
> >
> > *Original Query:*
> > SELECT * FROM
> > ( SELECT 'ABC' `UserID` FROM `dfs`.`path_to_parquet_file` tc LIMIT
> > 2147483647) `t0`
> > INNER JOIN
> > ( SELECT 'ABC' `UserID` FROM `dfs`.`path_to_parquet_file` tc LIMIT
> > 2147483647) `t1`
> > ON (`t0`.`UserID` IS NOT DISTINCT FROM `t1`.`UserID`)
> > LIMIT 2147483647
> >
> > *Workaround (add columns **d1a381f3g73 and **d1a381f3g74 to tables
> one
> > and two, respectively. Names don't really matter, just need to be
> unique):*
> > SELECT * FROM
> > ( SELECT *1 as d1a381f3g73*, 'ABC' `UserID` FROM
> > `dfs`.`path_to_parquet_file` tc LIMIT 2147483647) `t0`
> > INNER JOIN
> > ( SELECT *1 as d1a381f3g74*, 'ABC' `UserID` FROM
> > `dfs`.`path_to_parquet_file` tc LIMIT 2147483647) `t1`
> > ON (`t0`.*d1a381f3g73 = *`t1`.*d1a381f3g74*)
> > WHERE `t0`.`UserID` IS NOT DISTINCT FROM `t1`.`UserID`
> > LIMIT 2147483647
> >
> > Regards
> >
> >
> > *Shadi Khalifa, PhD*
> > Postdoctoral Fellow
> > Cognitive Analytics Development Hub
> > Centre for Advanced Computing
> > Queen’s University
> > (613) 533-6000 x78347
> > http://cac.queensu.ca
> >
> > I'm just a neuron in the society collective brain
> >
> > *Join us for HPCS in June 2017! Register at:*  *http://2017.hpcs.ca/
> > <http://2017.hpcs.ca/>*
> >
> > P Please consider your environmental responsibility before printing
> this
> > e-mail
> >
> > *01001001 0010 01101100 0110 01110110 01100101 0010
> 01000101
> > 01100111 0001 0111 01110100 *
> >

Re: [Drill 1.10.0] : Memory was leaked by query

2017-04-19 Thread Anup Tiwari
Thanks Padma, it worked.

Regards,
*Anup Tiwari*

On Wed, Apr 19, 2017 at 1:13 AM, Kunal Khatua  wrote:

> Could you also share the profiles for the failed queries as well?
>
>
> Thanks
>
> Kunal
>
> 
> From: Padma Penumarthy 
> Sent: Tuesday, April 18, 2017 7:18:08 AM
> To: u...@drill.apache.org
> Cc: dev@drill.apache.org
> Subject: Re: [Drill 1.10.0] : Memory was leaked by query
>
> Seems like you are running into  DRILL-5435<https://issues.apac
> he.org/jira/browse/DRILL-5435>.
> Try  turning off async parquet reader and see if that helps.
> alter session set `store.parquet.reader.pagereader.async`=false;
>
> Thanks,
> Padma
>
>
> On Apr 18, 2017, at 6:14 AM, Anup Tiwari  lto:anup.tiw...@games24x7.com>> wrote:
>
> Hi Team,
>
> Please find following information :
>
> *Cluster configuration :*
> Number of Nodes : 5
> Cores/Node : 8
> RAM : 32
>
> *Variable values :*
> planner.width.max_per_node = 5
> planner.width.max_per_query = 30
> planner.memory.max_query_memory_per_node = 4294967296
>
> I am getting following error on simple select statement which is coming 6
> times out of 10 times, let me know if i am missing anything:
>
> *Query :*
> select udf_channel,uid from dfs.tmp.tt1 where (event = 'ajax' and ajaxurl
> like '%/j_check%' and ajaxResponse like '%success%true%') limit 5;
>
> *Error :*
>
> ERROR o.a.d.e.w.fragment.FragmentExecutor - SYSTEM ERROR:
> IllegalStateException: Memory was leaked by query. Memory leaked: (1048576)
> Allocator(op:1:24:6:ParquetRowGroupScan)
> 100/1048576/27140096/100 (res/actual/peak/limit)
>
>
> Fragment 1:24
>
> [Error Id: a54cc1bf-794a-4143-bd82-0dd5fa3c8f52 on
> prod-hadoop-101.bom-prod.aws.games24x7.com<http://prod-hadoo
> p-101.bom-prod.aws.games24x7.com>:31010]
> org.apache.drill.common.exceptions.UserException: SYSTEM ERROR:
> IllegalStateException: Memory was leaked by query. Memory leaked: (1048576)
> Allocator(op:1:24:6:ParquetRowGroupScan)
> 100/1048576/27140096/100 (res/actual/peak/limit)
>
>
> Fragment 1:24
>
> [Error Id: a54cc1bf-794a-4143-bd82-0dd5fa3c8f52 on
> prod-hadoop-101.bom-prod.aws.games24x7.com<http://prod-hadoo
> p-101.bom-prod.aws.games24x7.com>:31010]
>at
> org.apache.drill.common.exceptions.UserException$Builder.
> build(UserException.java:544)
> ~[drill-common-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.work.fragment.FragmentExecutor.sendFin
> alState(FragmentExecutor.java:293)
> [drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.work.fragment.FragmentExecutor.cleanup
> (FragmentExecutor.java:160)
> [drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.work.fragment.FragmentExecutor.run(Fra
> gmentExecutor.java:262)
> [drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.common.SelfCleaningRunnable.run(SelfCleanin
> gRunnable.java:38)
> [drill-common-1.10.0.jar:1.10.0]
>at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
> Executor.java:1142)
> [na:1.8.0_72]
>at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
> lExecutor.java:617)
> [na:1.8.0_72]
>at java.lang.Thread.run(Thread.java:745) [na:1.8.0_72]
> Caused by: java.lang.IllegalStateException: Memory was leaked by query.
> Memory leaked: (1048576)
> Allocator(op:1:24:6:ParquetRowGroupScan)
> 100/1048576/27140096/100 (res/actual/peak/limit)
>
>at
> org.apache.drill.exec.memory.BaseAllocator.close(BaseAllocator.java:502)
> ~[drill-memory-base-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.ops.OperatorContextImpl.close(Operator
> ContextImpl.java:149)
> ~[drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.ops.FragmentContext.suppressingClose(F
> ragmentContext.java:422)
> ~[drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.ops.FragmentContext.close(FragmentContext.java:411)
> ~[drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.work.fragment.FragmentExecutor.closeOu
> tResources(FragmentExecutor.java:318)
> [drill-java-exec-1.10.0.jar:1.10.0]
>at
> org.apache.drill.exec.work.fragment.FragmentExecutor.cleanup
> (FragmentExecutor.java:155)
> [drill-java-exec-1.10.0.jar:1.10.0]
>... 5 common frames omitted
> 2017-04-18 18:21:54,172 [BitServer-4] INFO
> o.a.d.e.w.fragment.FragmentExecutor -
> 2709f415-c08a-13b9-9f05-fcf9008c484f:1:21: State change requested RUNNING
> --> CANCELLATION_REQUESTED
> 2017-04-18 18:21:54,172 [BitServer-4] INFO
> o.a.d.e.w.f.FragmentStatusReporter -
> 2709f415-c08a-13b9-9f05-fcf9008c484f:1:21: Sta

[Drill 1.10.0] : Memory was leaked by query

2017-04-18 Thread Anup Tiwari
00 (res/actual/peak/limit)


Fragment 1:21

[Error Id: 8b3bb6e8-77a0-4747-8602-43b40b349354 on
prod-hadoop-101.bom-prod.aws.games24x7.com:31010]
at
org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:544)
~[drill-common-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.work.fragment.FragmentExecutor.sendFinalState(FragmentExecutor.java:293)
[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.work.fragment.FragmentExecutor.cleanup(FragmentExecutor.java:160)
[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.work.fragment.FragmentExecutor.run(FragmentExecutor.java:262)
[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.common.SelfCleaningRunnable.run(SelfCleaningRunnable.java:38)
[drill-common-1.10.0.jar:1.10.0]
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_72]
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_72]
at java.lang.Thread.run(Thread.java:745) [na:1.8.0_72]
Caused by: java.lang.IllegalStateException: Memory was leaked by query.
Memory leaked: (1048576)
Allocator(op:1:21:6:ParquetRowGroupScan)
100/1048576/27140096/100 (res/actual/peak/limit)

at
org.apache.drill.exec.memory.BaseAllocator.close(BaseAllocator.java:502)
~[drill-memory-base-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.ops.OperatorContextImpl.close(OperatorContextImpl.java:149)
~[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.ops.FragmentContext.suppressingClose(FragmentContext.java:422)
~[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.ops.FragmentContext.close(FragmentContext.java:411)
~[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.work.fragment.FragmentExecutor.closeOutResources(FragmentExecutor.java:318)
[drill-java-exec-1.10.0.jar:1.10.0]
at
org.apache.drill.exec.work.fragment.FragmentExecutor.cleanup(FragmentExecutor.java:155)
[drill-java-exec-1.10.0.jar:1.10.0]
    ... 5 common frames omitted


Regards,
*Anup Tiwari*


Re: [Drill 1.9.0] : [CONNECTION ERROR] :- (user client) closed unexpectedly. Drillbit down?

2017-03-09 Thread Anup Tiwari
Hi John,

First of all sorry for delayed response and thanks for your suggestion,
reducing value of "planner.width.max_per_node" helped me a lot, above issue
which was coming 8 out of 10 times earlier now it is coming only 2 out of
10 times.

As mentioned above occurrences of connection error came down considerably,
but now sometimes i get "Heap Space Error" for few queries and due to this
sometimes drill-bits on some/all nodes gets killed. Let me know if any
other variable i can check for this(As of now, i have 8GB of Heap and 20GB
of Direct memory) :

*Error Log :*

ERROR o.a.drill.common.CatastrophicFailure - Catastrophic Failure Occurred,
exiting. Information message: Unable to handle out of memory condition in
FragmentExecutor.
java.lang.OutOfMemoryError: Java heap space
at org.apache.xerces.dom.DeferredDocumentImpl.getNodeObject(Unknown
Source) ~[xercesImpl-2.11.0.jar:na]
at
org.apache.xerces.dom.DeferredDocumentImpl.synchronizeChildren(Unknown
Source) ~[xercesImpl-2.11.0.jar:na]
at
org.apache.xerces.dom.DeferredElementImpl.synchronizeChildren(Unknown
Source) ~[xercesImpl-2.11.0.jar:na]
at org.apache.xerces.dom.ElementImpl.normalize(Unknown Source)
~[xercesImpl-2.11.0.jar:na]
at org.apache.xerces.dom.ElementImpl.normalize(Unknown Source)
~[xercesImpl-2.11.0.jar:na]
at org.apache.xerces.dom.ElementImpl.normalize(Unknown Source)
~[xercesImpl-2.11.0.jar:na]
at com.games24x7.device.NewDeviceData.setup(NewDeviceData.java:94)
~[DeviceDataClient-0.0.1-SNAPSHOT.jar:na]
at
org.apache.drill.exec.test.generated.FiltererGen5369.doSetup(FilterTemplate2.java:97)
~[na:na]
at
org.apache.drill.exec.test.generated.FiltererGen5369.setup(FilterTemplate2.java:54)
~[na:na]
at
org.apache.drill.exec.physical.impl.filter.FilterRecordBatch.generateSV2Filterer(FilterRecordBatch.java:195)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.physical.impl.filter.FilterRecordBatch.setupNewSchema(FilterRecordBatch.java:107)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext(AbstractSingleRecordBatch.java:78)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:162)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:119)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:109)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext(AbstractSingleRecordBatch.java:51)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.physical.impl.svremover.RemovingRecordBatch.innerNext(RemovingRecordBatch.java:94)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:162)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:119)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:109)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext(AbstractSingleRecordBatch.java:51)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext(ProjectRecordBatch.java:135)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:162)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:119)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:109)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.physical.impl.aggregate.HashAggBatch.buildSchema(HashAggBatch.java:108)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:142)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:119)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:109)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractSingleRecordBatch.innerNext(AbstractSingleRecordBatch.java:51)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext(ProjectRecordBatch.java:135)
~[drill-java-exec-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:162)
~[drill-java-exec-1.9.0.jar:1.9.0]


Regards,
*Anup Tiwari*

On Mon, Mar 6, 2017 at 7

Re: [Drill 1.9.0] : [CONNECTION ERROR] :- (user client) closed unexpectedly. Drillbit down?

2017-03-04 Thread Anup Tiwari
Hi John,

I have tried above config as well but still getting this issue.
And please note that we were using similar configuration params for Drill
1.6 where this issue was not coming.
Anything else which i can try?

Regards,
*Anup Tiwari*

On Fri, Mar 3, 2017 at 11:01 PM, Abhishek Girish  wrote:

> +1 on John's suggestion.
>
> On Fri, Mar 3, 2017 at 6:24 AM, John Omernik  wrote:
>
> > So your node has 32G of ram yet you are allowing Drill to use 36G.  I
> would
> > change your settings to be 8GB of Heap, and 22GB of Direct Memory. See if
> > this helps with your issues.  Also, are you using a distributed
> filesystem?
> > If so you may want to allow even more free ram...i.e. 8GB of Heap and
> 20GB
> > of Direct.
> >
> > On Fri, Mar 3, 2017 at 8:20 AM, Anup Tiwari 
> > wrote:
> >
> > > Hi,
> > >
> > > Please find our configuration details :-
> > >
> > > Number of Nodes : 4
> > > RAM/Node : 32GB
> > > Core/Node : 8
> > > DRILL_MAX_DIRECT_MEMORY="20G"
> > > DRILL_HEAP="16G"
> > >
> > > And all other variables are set to default.
> > >
> > > Since we have tried some of the settings suggested above but still
> facing
> > > this issue more frequently, kindly suggest us what is best
> configuration
> > > for our environment.
> > >
> > > Regards,
> > > *Anup Tiwari*
> > >
> > > On Thu, Mar 2, 2017 at 1:26 AM, John Omernik  wrote:
> > >
> > > > Another thing to consider is ensure you have a Spill Location setup,
> > and
> > > > then disable hashagg/hashjoin for the query...
> > > >
> > > > On Wed, Mar 1, 2017 at 1:25 PM, Abhishek Girish 
> > > > wrote:
> > > >
> > > > > Hey Anup,
> > > > >
> > > > > This is indeed an issue, and I can understand that having an
> unstable
> > > > > environment is not something anyone wants. DRILL-4708 is still
> > > > unresolved -
> > > > > hopefully someone will get to it soon. I've bumped up the priority.
> > > > >
> > > > > Unfortunately we do not publish any sizing guidelines, so you'd
> have
> > to
> > > > > experiment to settle on the right load for your cluster. Please
> > > decrease
> > > > > the concurrency (number of queries running in parallel). And try
> > > bumping
> > > > up
> > > > > Drill DIRECT memory. Also, please set the system options
> recommended
> > by
> > > > > Sudheesh. While this may not solve the issue, it may help reduce
> it's
> > > > > occurrence.
> > > > >
> > > > > Can you also update the JIRA with your configurations, type of
> > queries
> > > > and
> > > > > the relevant logs?
> > > > >
> > > > > -Abhishek
> > > > >
> > > > > On Wed, Mar 1, 2017 at 10:17 AM, Anup Tiwari <
> > > anup.tiw...@games24x7.com>
> > > > > wrote:
> > > > >
> > > > > > Hi,
> > > > > >
> > > > > > Can someone look into it? As we are now getting this more
> > frequently
> > > in
> > > > > > Adhoc queries as well.
> > > > > > And for automation jobs, we are moving to Hive as in drill we are
> > > > getting
> > > > > > this more frequently.
> > > > > >
> > > > > > Regards,
> > > > > > *Anup Tiwari*
> > > > > >
> > > > > > On Sat, Dec 31, 2016 at 12:11 PM, Anup Tiwari <
> > > > anup.tiw...@games24x7.com
> > > > > >
> > > > > > wrote:
> > > > > >
> > > > > > > Hi,
> > > > > > >
> > > > > > > We are getting this issue bit more frequently. can someone
> please
> > > > look
> > > > > > > into it and tell us that why it is happening since as mention
> in
> > > > > earlier
> > > > > > > mail when this query gets executed no other query is running at
> > > that
> > > > > > time.
> > > > > > >
> > > > > > > Thanks in advance.
> > > > > > >
> > > > > > > Regards,
> > > > > > > *Anup Tiwari*
> > > > > > >
> &g

Re: [Drill 1.9.0] : [CONNECTION ERROR] :- (user client) closed unexpectedly. Drillbit down?

2017-03-03 Thread Anup Tiwari
Hi,

Please find our configuration details :-

Number of Nodes : 4
RAM/Node : 32GB
Core/Node : 8
DRILL_MAX_DIRECT_MEMORY="20G"
DRILL_HEAP="16G"

And all other variables are set to default.

Since we have tried some of the settings suggested above but still facing
this issue more frequently, kindly suggest us what is best configuration
for our environment.

Regards,
*Anup Tiwari*

On Thu, Mar 2, 2017 at 1:26 AM, John Omernik  wrote:

> Another thing to consider is ensure you have a Spill Location setup, and
> then disable hashagg/hashjoin for the query...
>
> On Wed, Mar 1, 2017 at 1:25 PM, Abhishek Girish 
> wrote:
>
> > Hey Anup,
> >
> > This is indeed an issue, and I can understand that having an unstable
> > environment is not something anyone wants. DRILL-4708 is still
> unresolved -
> > hopefully someone will get to it soon. I've bumped up the priority.
> >
> > Unfortunately we do not publish any sizing guidelines, so you'd have to
> > experiment to settle on the right load for your cluster. Please decrease
> > the concurrency (number of queries running in parallel). And try bumping
> up
> > Drill DIRECT memory. Also, please set the system options recommended by
> > Sudheesh. While this may not solve the issue, it may help reduce it's
> > occurrence.
> >
> > Can you also update the JIRA with your configurations, type of queries
> and
> > the relevant logs?
> >
> > -Abhishek
> >
> > On Wed, Mar 1, 2017 at 10:17 AM, Anup Tiwari 
> > wrote:
> >
> > > Hi,
> > >
> > > Can someone look into it? As we are now getting this more frequently in
> > > Adhoc queries as well.
> > > And for automation jobs, we are moving to Hive as in drill we are
> getting
> > > this more frequently.
> > >
> > > Regards,
> > > *Anup Tiwari*
> > >
> > > On Sat, Dec 31, 2016 at 12:11 PM, Anup Tiwari <
> anup.tiw...@games24x7.com
> > >
> > > wrote:
> > >
> > > > Hi,
> > > >
> > > > We are getting this issue bit more frequently. can someone please
> look
> > > > into it and tell us that why it is happening since as mention in
> > earlier
> > > > mail when this query gets executed no other query is running at that
> > > time.
> > > >
> > > > Thanks in advance.
> > > >
> > > > Regards,
> > > > *Anup Tiwari*
> > > >
> > > > On Sat, Dec 24, 2016 at 10:20 AM, Anup Tiwari <
> > anup.tiw...@games24x7.com
> > > >
> > > > wrote:
> > > >
> > > >> Hi Sudheesh,
> > > >>
> > > >> Please find below ans :-
> > > >>
> > > >> 1. Total 4,(3 Datanodes, 1 namenode)
> > > >> 2. Only one query, as this query is part of daily dump and runs in
> > early
> > > >> morning.
> > > >>
> > > >> And as @chun mentioned , it seems similar to DRILL-4708 , so any
> > update
> > > >> on progress of this ticket?
> > > >>
> > > >>
> > > >> On 22-Dec-2016 12:13 AM, "Sudheesh Katkam" 
> > > wrote:
> > > >>
> > > >> Two more questions..
> > > >>
> > > >> (1) How many nodes in your cluster?
> > > >> (2) How many queries are running when the failure is seen?
> > > >>
> > > >> If you have multiple large queries running at the same time, the
> load
> > on
> > > >> the system could cause those failures (which are heartbeat related).
> > > >>
> > > >> The two options I suggested decrease the parallelism of stages in a
> > > >> query, this implies lesser load but slower execution.
> > > >>
> > > >> System level option affect all queries, and session level affect
> > queries
> > > >> on a specific connection. Not sure what is preferred in your
> > > environment.
> > > >>
> > > >> Also, you may be interested in metrics. More info here:
> > > >>
> > > >> http://drill.apache.org/docs/monitoring-metrics/ <
> > > >> http://drill.apache.org/docs/monitoring-metrics/>
> > > >>
> > > >> Thank you,
> > > >> Sudheesh
> > > >>
> > > >> > On Dec 21, 2016, at 4:31 AM, Anup Tiwari <
> anup.tiw...@games24x7.com
> > >
> > > >> 

Re: [Drill 1.9.0] : [CONNECTION ERROR] :- (user client) closed unexpectedly. Drillbit down?

2017-03-01 Thread Anup Tiwari
Hi,

Can someone look into it? As we are now getting this more frequently in
Adhoc queries as well.
And for automation jobs, we are moving to Hive as in drill we are getting
this more frequently.

Regards,
*Anup Tiwari*

On Sat, Dec 31, 2016 at 12:11 PM, Anup Tiwari 
wrote:

> Hi,
>
> We are getting this issue bit more frequently. can someone please look
> into it and tell us that why it is happening since as mention in earlier
> mail when this query gets executed no other query is running at that time.
>
> Thanks in advance.
>
> Regards,
> *Anup Tiwari*
>
> On Sat, Dec 24, 2016 at 10:20 AM, Anup Tiwari 
> wrote:
>
>> Hi Sudheesh,
>>
>> Please find below ans :-
>>
>> 1. Total 4,(3 Datanodes, 1 namenode)
>> 2. Only one query, as this query is part of daily dump and runs in early
>> morning.
>>
>> And as @chun mentioned , it seems similar to DRILL-4708 , so any update
>> on progress of this ticket?
>>
>>
>> On 22-Dec-2016 12:13 AM, "Sudheesh Katkam"  wrote:
>>
>> Two more questions..
>>
>> (1) How many nodes in your cluster?
>> (2) How many queries are running when the failure is seen?
>>
>> If you have multiple large queries running at the same time, the load on
>> the system could cause those failures (which are heartbeat related).
>>
>> The two options I suggested decrease the parallelism of stages in a
>> query, this implies lesser load but slower execution.
>>
>> System level option affect all queries, and session level affect queries
>> on a specific connection. Not sure what is preferred in your environment.
>>
>> Also, you may be interested in metrics. More info here:
>>
>> http://drill.apache.org/docs/monitoring-metrics/ <
>> http://drill.apache.org/docs/monitoring-metrics/>
>>
>> Thank you,
>> Sudheesh
>>
>> > On Dec 21, 2016, at 4:31 AM, Anup Tiwari 
>> wrote:
>> >
>> > @sudheesh, yes drill bit is running on datanodeN/10.*.*.5:31010).
>> >
>> > Can you tell me how this will impact to query and do i have to set this
>> at
>> > session level OR system level?
>> >
>> >
>> >
>> > Regards,
>> > *Anup Tiwari*
>> >
>> > On Tue, Dec 20, 2016 at 11:59 PM, Chun Chang 
>> wrote:
>> >
>> >> I am pretty sure this is the same as DRILL-4708.
>> >>
>> >> On Tue, Dec 20, 2016 at 10:27 AM, Sudheesh Katkam <
>> skat...@maprtech.com>
>> >> wrote:
>> >>
>> >>> Is the drillbit service (running on datanodeN/10.*.*.5:31010) actually
>> >>> down when the error is seen?
>> >>>
>> >>> If not, try lowering parallelism using these two session options,
>> before
>> >>> running the queries:
>> >>>
>> >>> planner.width.max_per_node (decrease this)
>> >>> planner.slice_target (increase this)
>> >>>
>> >>> Thank you,
>> >>> Sudheesh
>> >>>
>> >>>> On Dec 20, 2016, at 12:28 AM, Anup Tiwari > >
>> >>> wrote:
>> >>>>
>> >>>> Hi Team,
>> >>>>
>> >>>> We are running some drill automation script on a daily basis and we
>> >> often
>> >>>> see that some query gets failed frequently by giving below error ,
>> >> Also i
>> >>>> came across DRILL-4708 <https://issues.apache.org/
>> >> jira/browse/DRILL-4708
>> >>>>
>> >>>> which seems similar, Can anyone give me update on that OR workaround
>> to
>> >>>> avoid such issue ?
>> >>>>
>> >>>> *Stack Trace :-*
>> >>>>
>> >>>> Error: CONNECTION ERROR: Connection /10.*.*.1:41613 <-->
>> >>>> datanodeN/10.*.*.5:31010 (user client) closed unexpectedly. Drillbit
>> >>> down?
>> >>>>
>> >>>>
>> >>>> [Error Id: 5089f2f1-0dfd-40f8-9fa0-8276c08be53f ] (state=,code=0)
>> >>>> java.sql.SQLException: CONNECTION ERROR: Connection /10.*.*.1:41613
>> >> <-->
>> >>>> datanodeN/10.*.*.5:31010 (user client) closed unexpectedly. Drillb
>> >>>> it down?
>> >>>>
>> >>>>
>> >>>> [Error Id: 5089f2f1-0dfd-40f8-9fa0-8276c08be53f ]
>> >>>>   at
>> >>>> org.apache.drill.jdbc.impl.DrillCur

Re: Storage Plugin for accessing Hive ORC Table from Drill

2017-02-09 Thread Anup Tiwari
First of all, sorry for late reply.

@Chunhui, you are right. we are using Hive 2.0

And are we planning to update hive libraries in next release of drill?

@rahul, as you said i have created another table with just "use stored as
orc" but all column and now drill is able to read it. Do you have any idea
why it worked now? Below is create table statement of new table, the
difference which i can observe is in TBLPROPERTIES ,partitioning and
bucketing:-

CREATE TABLE `logindetails_all_tmp`(
  `sid` char(40),
  `channel_id` tinyint,
  `c_t` bigint,
  `l_t` bigint)
ROW FORMAT SERDE
  'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
STORED AS INPUTFORMAT
  'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
OUTPUTFORMAT
  'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
LOCATION
  'hdfs://hostname1:9000/usr/hive/warehouse/logindetails_all_tmp'
TBLPROPERTIES (
  'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}',
  'numFiles'='3',
  'numRows'='1993254',
  'rawDataSize'='1143386232',
  'totalSize'='69876827',
  'transient_lastDdlTime'='1486640969');


Regards,
*Anup Tiwari*

On Sat, Jan 21, 2017 at 1:04 PM, Chunhui Shi  wrote:

> I guess you are using Hive 2.0 as meta server while Drill has only 1.2
> libraries.
>
>
> In Hive 2.0 above, This delta format could have more than one '_' as
> separator while 1.2 has only one '_'.
>
>
> I think Drill should eventually update to use Hive's 2.0/2.1 libraries.
>
> 
> From: Anup Tiwari 
> Sent: Friday, January 20, 2017 10:07:50 PM
> To: u...@drill.apache.org; dev@drill.apache.org
> Subject: Re: Storage Plugin for accessing Hive ORC Table from Drill
>
> @Andries, We are using Hive 2.1.1 with Drill 1.9.0.
>
> @Zelaine, Could this be a problem in your Hive metastore?--> As i mentioned
> earlier, i am able to read hive parquet tables in Drill through hive
> storage plugin. So can you tell me a bit more like which type of
> configuration i am missing in metastore?
>
> Regards,
> *Anup Tiwari*
>
> On Sat, Jan 21, 2017 at 4:56 AM, Zelaine Fong  wrote:
>
> > The stack trace shows the following:
> >
> > Caused by: org.apache.drill.common.exceptions.DrillRuntimeException:
> > java.io.IOException: Failed to get numRows from HiveTable
> >
> > The Drill optimizer is trying to read rowcount information from Hive.
> > Could this be a problem in your Hive metastore?
> >
> > Has anyone else seen this before?
> >
> > -- Zelaine
> >
> > On 1/20/17, 7:35 AM, "Andries Engelbrecht" 
> wrote:
> >
> > What version of Hive are you using?
> >
> >
> > --Andries
> >
> > 
> > From: Anup Tiwari 
> > Sent: Friday, January 20, 2017 3:00:43 AM
> > To: u...@drill.apache.org; dev@drill.apache.org
> > Subject: Re: Storage Plugin for accessing Hive ORC Table from Drill
> >
> > Hi,
> >
> > Please find below Create Table Statement and subsequent Drill Error
> :-
> >
> > *Table Structure :*
> >
> > CREATE TABLE `logindetails_all`(
> >   `sid` char(40),
> >   `channel_id` tinyint,
> >   `c_t` bigint,
> >   `l_t` bigint)
> > PARTITIONED BY (
> >   `login_date` char(10))
> > CLUSTERED BY (
> >   channel_id)
> > INTO 9 BUCKETS
> > ROW FORMAT SERDE
> >   'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
> > STORED AS INPUTFORMAT
> >   'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
> > OUTPUTFORMAT
> >   'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
> > LOCATION
> >   'hdfs://hostname1:9000/usr/hive/warehouse/logindetails_all'
> > TBLPROPERTIES (
> >   'compactorthreshold.hive.compactor.delta.num.threshold'='6',
> >   'compactorthreshold.hive.compactor.delta.pct.threshold'='0.5',
> >   'transactional'='true',
> >   'transient_lastDdlTime'='1484313383');
> > ;
> >
> > *Drill Error :*
> >
> > *Query* : select * from hive.logindetails_all limit 1;
> >
> > *Error :*
> > 2017-01-20 16:21:12,625 [277e145e-c6bc-3372-01d0-
> 6c5b75b92d73:foreman]
> > INFO  o.a.drill.exec.work.foreman.Foreman - Query text for query id
> > 277e145e-c6bc-3372-01d0-6c5b75b92d73

Re: Storage Plugin for accessing Hive ORC Table from Drill

2017-01-22 Thread Anup Tiwari
can you point me to any specific line or sentence on that link?

Also please correct me if i am misinterpreting, but as written in 1st
line "*Drill
1.1 and later supports Hive 1.0*", does that mean Drill 1.1 and later
doesn't support OR partially support Hive 2.x?

Regards,
*Anup Tiwari*

On Sat, Jan 21, 2017 at 8:48 PM, Zelaine Fong  wrote:

> Have you taken a look at http://drill.apache.org/docs/hive-storage-plugin/
> ?
>
> -- Zelaine
>
> On 1/20/17, 10:07 PM, "Anup Tiwari"  wrote:
>
> @Andries, We are using Hive 2.1.1 with Drill 1.9.0.
>
> @Zelaine, Could this be a problem in your Hive metastore?--> As i
> mentioned
> earlier, i am able to read hive parquet tables in Drill through hive
> storage plugin. So can you tell me a bit more like which type of
> configuration i am missing in metastore?
>
> Regards,
> *Anup Tiwari*
>
> On Sat, Jan 21, 2017 at 4:56 AM, Zelaine Fong  wrote:
>
> > The stack trace shows the following:
> >
> > Caused by: org.apache.drill.common.exceptions.DrillRuntimeException:
> > java.io.IOException: Failed to get numRows from HiveTable
> >
> > The Drill optimizer is trying to read rowcount information from Hive.
> > Could this be a problem in your Hive metastore?
> >
> > Has anyone else seen this before?
> >
> > -- Zelaine
> >
> > On 1/20/17, 7:35 AM, "Andries Engelbrecht" 
> wrote:
> >
> > What version of Hive are you using?
> >
> >
> > --Andries
> >
> > 
> > From: Anup Tiwari 
> > Sent: Friday, January 20, 2017 3:00:43 AM
> > To: u...@drill.apache.org; dev@drill.apache.org
> > Subject: Re: Storage Plugin for accessing Hive ORC Table from
> Drill
> >
> > Hi,
> >
> > Please find below Create Table Statement and subsequent Drill
> Error :-
> >
> > *Table Structure :*
> >
> > CREATE TABLE `logindetails_all`(
> >   `sid` char(40),
> >   `channel_id` tinyint,
> >   `c_t` bigint,
> >   `l_t` bigint)
> > PARTITIONED BY (
> >   `login_date` char(10))
> > CLUSTERED BY (
> >   channel_id)
> > INTO 9 BUCKETS
> > ROW FORMAT SERDE
> >   'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
> > STORED AS INPUTFORMAT
> >   'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
> > OUTPUTFORMAT
> >   'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
> > LOCATION
> >   'hdfs://hostname1:9000/usr/hive/warehouse/logindetails_all'
> > TBLPROPERTIES (
> >   'compactorthreshold.hive.compactor.delta.num.threshold'='6',
> >   'compactorthreshold.hive.compactor.delta.pct.threshold'='0.5',
> >   'transactional'='true',
> >   'transient_lastDdlTime'='1484313383');
> > ;
> >
> > *Drill Error :*
> >
> > *Query* : select * from hive.logindetails_all limit 1;
> >
> > *Error :*
> > 2017-01-20 16:21:12,625 [277e145e-c6bc-3372-01d0-
> 6c5b75b92d73:foreman]
> > INFO  o.a.drill.exec.work.foreman.Foreman - Query text for
> query id
> > 277e145e-c6bc-3372-01d0-6c5b75b92d73: select * from
> > hive.logindetails_all
> > limit 1
> > 2017-01-20 16:21:12,831 [277e145e-c6bc-3372-01d0-
> 6c5b75b92d73:foreman]
> > ERROR o.a.drill.exec.work.foreman.Foreman - SYSTEM ERROR:
> > NumberFormatException: For input string: "004_"
> >
> >
> > [Error Id: 53fa92e1-477e-45d2-b6f7-6eab9ef1da35 on
> > prod-hadoop-101.bom-prod.aws.games24x7.com:31010]
> > org.apache.drill.common.exceptions.UserException: SYSTEM ERROR:
> > NumberFormatException: For input string: "004_"
> >
> >
> > [Error Id: 53fa92e1-477e-45d2-b6f7-6eab9ef1da35 on
> > prod-hadoop-101.bom-prod.aws.games24x7.com:31010]
> > at
> > org.apache.drill.common.exceptions.UserException$
> > Builder.build(UserException.java:543)
> > ~[drill-common-1.9.0.jar:1.9.0]
> > at
> > org.apache.drill.exec.work.fo

Re: Storage Plugin for accessing Hive ORC Table from Drill

2017-01-20 Thread Anup Tiwari
@Andries, We are using Hive 2.1.1 with Drill 1.9.0.

@Zelaine, Could this be a problem in your Hive metastore?--> As i mentioned
earlier, i am able to read hive parquet tables in Drill through hive
storage plugin. So can you tell me a bit more like which type of
configuration i am missing in metastore?

Regards,
*Anup Tiwari*

On Sat, Jan 21, 2017 at 4:56 AM, Zelaine Fong  wrote:

> The stack trace shows the following:
>
> Caused by: org.apache.drill.common.exceptions.DrillRuntimeException:
> java.io.IOException: Failed to get numRows from HiveTable
>
> The Drill optimizer is trying to read rowcount information from Hive.
> Could this be a problem in your Hive metastore?
>
> Has anyone else seen this before?
>
> -- Zelaine
>
> On 1/20/17, 7:35 AM, "Andries Engelbrecht"  wrote:
>
> What version of Hive are you using?
>
>
> --Andries
>
> 
> From: Anup Tiwari 
> Sent: Friday, January 20, 2017 3:00:43 AM
> To: u...@drill.apache.org; dev@drill.apache.org
> Subject: Re: Storage Plugin for accessing Hive ORC Table from Drill
>
> Hi,
>
> Please find below Create Table Statement and subsequent Drill Error :-
>
> *Table Structure :*
>
> CREATE TABLE `logindetails_all`(
>   `sid` char(40),
>   `channel_id` tinyint,
>   `c_t` bigint,
>   `l_t` bigint)
> PARTITIONED BY (
>   `login_date` char(10))
> CLUSTERED BY (
>   channel_id)
> INTO 9 BUCKETS
> ROW FORMAT SERDE
>   'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
> STORED AS INPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
> OUTPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
> LOCATION
>   'hdfs://hostname1:9000/usr/hive/warehouse/logindetails_all'
> TBLPROPERTIES (
>   'compactorthreshold.hive.compactor.delta.num.threshold'='6',
>   'compactorthreshold.hive.compactor.delta.pct.threshold'='0.5',
>   'transactional'='true',
>   'transient_lastDdlTime'='1484313383');
> ;
>
> *Drill Error :*
>
> *Query* : select * from hive.logindetails_all limit 1;
>
> *Error :*
> 2017-01-20 16:21:12,625 [277e145e-c6bc-3372-01d0-6c5b75b92d73:foreman]
> INFO  o.a.drill.exec.work.foreman.Foreman - Query text for query id
> 277e145e-c6bc-3372-01d0-6c5b75b92d73: select * from
> hive.logindetails_all
> limit 1
> 2017-01-20 16:21:12,831 [277e145e-c6bc-3372-01d0-6c5b75b92d73:foreman]
> ERROR o.a.drill.exec.work.foreman.Foreman - SYSTEM ERROR:
> NumberFormatException: For input string: "004_"
>
>
> [Error Id: 53fa92e1-477e-45d2-b6f7-6eab9ef1da35 on
> prod-hadoop-101.bom-prod.aws.games24x7.com:31010]
> org.apache.drill.common.exceptions.UserException: SYSTEM ERROR:
> NumberFormatException: For input string: "004_"
>
>
> [Error Id: 53fa92e1-477e-45d2-b6f7-6eab9ef1da35 on
> prod-hadoop-101.bom-prod.aws.games24x7.com:31010]
> at
> org.apache.drill.common.exceptions.UserException$
> Builder.build(UserException.java:543)
> ~[drill-common-1.9.0.jar:1.9.0]
> at
> org.apache.drill.exec.work.foreman.Foreman$ForemanResult.
> close(Foreman.java:825)
> [drill-java-exec-1.9.0.jar:1.9.0]
> at
> org.apache.drill.exec.work.foreman.Foreman.moveToState(
> Foreman.java:935)
> [drill-java-exec-1.9.0.jar:1.9.0]
> at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.
> java:281)
> [drill-java-exec-1.9.0.jar:1.9.0]
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
> [na:1.8.0_72]
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
> [na:1.8.0_72]
> at java.lang.Thread.run(Thread.java:745) [na:1.8.0_72]
> Caused by: org.apache.drill.exec.work.foreman.ForemanException:
> Unexpected
> exception during fragment initialization: Internal error: Error while
> applying rule DrillPushProjIntoScan, args
> [rel#4220197:LogicalProject.NONE.ANY([]).[](input=rel#
> 4220196:Subset#0.ENUMERABLE.ANY([]).[],sid=$0,channel_id=$
> 1,c_t=$2,l_t=$3,login_date=$4),
> rel#4220181:EnumerableTableScan.ENUMERABLE.ANY([]).[](table=[hive,
> logindetails_all])]
> ... 4 common frames omitted
> Caused by: java.lang.AssertionError: Internal error: Error while
> applying
> rule DrillPushProjIntoScan, args
> [rel#4220197:LogicalProject.NONE.ANY([]).[](in

Re: Storage Plugin for accessing Hive ORC Table from Drill

2017-01-20 Thread Anup Tiwari
iveCost(RelMdPercentageOriginalRows.java:165)
~[calcite-core-1.4.0-drill-r19.jar:1.4.0-drill-r19]
... 42 common frames omitted
Caused by: java.io.IOException: Failed to get numRows from HiveTable
at
org.apache.drill.exec.store.hive.HiveMetadataProvider.getStats(HiveMetadataProvider.java:113)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.store.hive.HiveScan.getScanStats(HiveScan.java:224)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
... 45 common frames omitted
Caused by: java.lang.RuntimeException: serious problem
at
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo(OrcInputFormat.java:1021)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
at
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits(OrcInputFormat.java:1048)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:253)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.store.hive.HiveMetadataProvider$1.run(HiveMetadataProvider.java:241)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
at java.security.AccessController.doPrivileged(Native Method)
~[na:1.8.0_72]
at javax.security.auth.Subject.doAs(Subject.java:422) ~[na:1.8.0_72]
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
~[hadoop-common-2.7.1.jar:na]
at
org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI(HiveMetadataProvider.java:241)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits(HiveMetadataProvider.java:142)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
at
org.apache.drill.exec.store.hive.HiveMetadataProvider.getStats(HiveMetadataProvider.java:105)
~[drill-storage-hive-core-1.9.0.jar:1.9.0]
... 46 common frames omitted
Caused by: java.util.concurrent.ExecutionException:
java.lang.NumberFormatException: For input string: "004_"
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
~[na:1.8.0_72]
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
~[na:1.8.0_72]
at
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo(OrcInputFormat.java:998)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
... 55 common frames omitted
Caused by: java.lang.NumberFormatException: For input string: "004_"
at
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
~[na:1.8.0_72]
at java.lang.Long.parseLong(Long.java:589) ~[na:1.8.0_72]
at java.lang.Long.parseLong(Long.java:631) ~[na:1.8.0_72]
at
org.apache.hadoop.hive.ql.io.AcidUtils.parseDelta(AcidUtils.java:310)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
at
org.apache.hadoop.hive.ql.io.AcidUtils.getAcidState(AcidUtils.java:379)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
at
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$FileGenerator.call(OrcInputFormat.java:634)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
at
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$FileGenerator.call(OrcInputFormat.java:620)
~[drill-hive-exec-shaded-1.9.0.jar:1.9.0]
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
~[na:1.8.0_72]
... 3 common frames omitted




Regards,
*Anup Tiwari*

On Thu, Jan 19, 2017 at 9:18 PM, Andries Engelbrecht 
wrote:

> I have not seen issues reading Hive ORC data with Drill.
>
>
> What is the DDL for the table in Hive?
>
>
> --Andries
>
> 
> From: Anup Tiwari 
> Sent: Thursday, January 19, 2017 12:49:20 AM
> To: u...@drill.apache.org
> Cc: dev@drill.apache.org
> Subject: Re: Storage Plugin for accessing Hive ORC Table from Drill
>
> We have created a ORC format table in hive and we were trying to read it in
> drill through hive plugin, but it is giving us error. But with same hive
> plugin, we are able to read parquet table created in hive.
>
> So after searching a bit, i found a drill documentation link
> <https://drill.apache.org/docs/apache-drill-contribution-ideas/> which
> says
> that we have to create custom storage plugin to read ORC format tables. So
> can you tell me how to create custom storage plugin in this case?
>
>
>
> Regards,
> *Anup Tiwari*
>
> On Thu, Jan 19, 2017 at 1:55 PM, Nitin Pawar 
> wrote:
>
> > you want to use the ORC files created by hive directly in drill or you
> want
> > to use them through hive?
> >
> > On Thu, Jan 19, 2017 at 1:40 PM, Anup Tiwari 
> > wrote:
> >
> > > +Dev
> > >
> > > Can someone help me in this?
> > >
> > > Regards,
> > > *Anup Tiwari*
> > >
> > > On Sun, Jan 15, 2017 at 2:21 PM, Anup Tiwari <
> anup.tiw...@games24x7.com>
> > > wrote:
> > >
> > > > Hi Team,
> > > >
> > > > Can someone tell me how to configure custom storage plugin in Drill
> for
> > > > accessing hive ORC tables?
> > > >
> > > > Thanks in advance!!
> > > >
> > > > Regards,
> > > > *Anup Tiwari*
> > > >
> > >
> >
> >
> >
> > --
> > Nitin Pawar
> >
>


Re: Storage Plugin for accessing Hive ORC Table from Drill

2017-01-19 Thread Anup Tiwari
+Dev

Can someone help me in this?

Regards,
*Anup Tiwari*

On Sun, Jan 15, 2017 at 2:21 PM, Anup Tiwari 
wrote:

> Hi Team,
>
> Can someone tell me how to configure custom storage plugin in Drill for
> accessing hive ORC tables?
>
> Thanks in advance!!
>
> Regards,
> *Anup Tiwari*
>


Re: Storage Plugin for accessing Hive ORC Table from Drill

2017-01-19 Thread Anup Tiwari
We have created a ORC format table in hive and we were trying to read it in
drill through hive plugin, but it is giving us error. But with same hive
plugin, we are able to read parquet table created in hive.

So after searching a bit, i found a drill documentation link
<https://drill.apache.org/docs/apache-drill-contribution-ideas/> which says
that we have to create custom storage plugin to read ORC format tables. So
can you tell me how to create custom storage plugin in this case?



Regards,
*Anup Tiwari*

On Thu, Jan 19, 2017 at 1:55 PM, Nitin Pawar 
wrote:

> you want to use the ORC files created by hive directly in drill or you want
> to use them through hive?
>
> On Thu, Jan 19, 2017 at 1:40 PM, Anup Tiwari 
> wrote:
>
> > +Dev
> >
> > Can someone help me in this?
> >
> > Regards,
> > *Anup Tiwari*
> >
> > On Sun, Jan 15, 2017 at 2:21 PM, Anup Tiwari 
> > wrote:
> >
> > > Hi Team,
> > >
> > > Can someone tell me how to configure custom storage plugin in Drill for
> > > accessing hive ORC tables?
> > >
> > > Thanks in advance!!
> > >
> > > Regards,
> > > *Anup Tiwari*
> > >
> >
>
>
>
> --
> Nitin Pawar
>