[jira] [Resolved] (DRILL-7170) IllegalStateException: Record count not set for this vector container

2019-10-04 Thread Boaz Ben-Zvi (Jira)


 [ 
https://issues.apache.org/jira/browse/DRILL-7170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Boaz Ben-Zvi resolved DRILL-7170.
-
  Reviewer: Sorabh Hamirwasia
Resolution: Fixed

> IllegalStateException: Record count not set for this vector container
> -
>
> Key: DRILL-7170
> URL: https://issues.apache.org/jira/browse/DRILL-7170
> Project: Apache Drill
>  Issue Type: Bug
>  Components: Execution - Relational Operators
>Reporter: Sorabh Hamirwasia
>Assignee: Boaz Ben-Zvi
>Priority: Major
> Fix For: 1.17.0
>
>
> {code:java}
> Query: 
> /root/drillAutomation/master/framework/resources/Advanced/tpcds/tpcds_sf1/original/maprdb/json/query95.sql
> WITH ws_wh AS
> (
> SELECT ws1.ws_order_number,
> ws1.ws_warehouse_sk wh1,
> ws2.ws_warehouse_sk wh2
> FROM   web_sales ws1,
> web_sales ws2
> WHERE  ws1.ws_order_number = ws2.ws_order_number
> ANDws1.ws_warehouse_sk <> ws2.ws_warehouse_sk)
> SELECT
> Count(DISTINCT ws_order_number) AS `order count` ,
> Sum(ws_ext_ship_cost)   AS `total shipping cost` ,
> Sum(ws_net_profit)  AS `total net profit`
> FROM web_sales ws1 ,
> date_dim ,
> customer_address ,
> web_site
> WHEREd_date BETWEEN '2000-04-01' AND  (
> Cast('2000-04-01' AS DATE) + INTERVAL '60' day)
> AND  ws1.ws_ship_date_sk = d_date_sk
> AND  ws1.ws_ship_addr_sk = ca_address_sk
> AND  ca_state = 'IN'
> AND  ws1.ws_web_site_sk = web_site_sk
> AND  web_company_name = 'pri'
> AND  ws1.ws_order_number IN
> (
> SELECT ws_order_number
> FROM   ws_wh)
> AND  ws1.ws_order_number IN
> (
> SELECT wr_order_number
> FROM   web_returns,
> ws_wh
> WHERE  wr_order_number = ws_wh.ws_order_number)
> ORDER BY count(DISTINCT ws_order_number)
> LIMIT 100
> Exception:
> java.sql.SQLException: SYSTEM ERROR: IllegalStateException: Record count not 
> set for this vector container
> Fragment 2:3
> Please, refer to logs for more information.
> [Error Id: 4ed92fce-505b-40ba-ac0e-4a302c28df47 on drill87:31010]
>   (java.lang.IllegalStateException) Record count not set for this vector 
> container
> 
> org.apache.drill.shaded.guava.com.google.common.base.Preconditions.checkState():459
> org.apache.drill.exec.record.VectorContainer.getRecordCount():394
> org.apache.drill.exec.record.RecordBatchSizer.():720
> org.apache.drill.exec.record.RecordBatchSizer.():704
> 
> org.apache.drill.exec.physical.impl.common.HashTableTemplate$BatchHolder.getActualSize():462
> 
> org.apache.drill.exec.physical.impl.common.HashTableTemplate.getActualSize():964
> 
> org.apache.drill.exec.physical.impl.common.HashTableTemplate.makeDebugString():973
> 
> org.apache.drill.exec.physical.impl.common.HashPartition.makeDebugString():601
> 
> org.apache.drill.exec.physical.impl.join.HashJoinBatch.makeDebugString():1313
> 
> org.apache.drill.exec.physical.impl.join.HashJoinBatch.executeBuildPhase():1105
> org.apache.drill.exec.physical.impl.join.HashJoinBatch.innerNext():525
> org.apache.drill.exec.record.AbstractRecordBatch.next():186
> org.apache.drill.exec.record.AbstractRecordBatch.next():126
> org.apache.drill.exec.record.AbstractRecordBatch.next():116
> org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
> 
> org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
> org.apache.drill.exec.record.AbstractRecordBatch.next():186
> org.apache.drill.exec.record.AbstractRecordBatch.next():126
> org.apache.drill.exec.test.generated.HashAggregatorGen1068899.doWork():642
> org.apache.drill.exec.physical.impl.aggregate.HashAggBatch.innerNext():296
> org.apache.drill.exec.record.AbstractRecordBatch.next():186
> org.apache.drill.exec.record.AbstractRecordBatch.next():126
> org.apache.drill.exec.record.AbstractRecordBatch.next():116
> org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
> 
> org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
> org.apache.drill.exec.record.AbstractRecordBatch.next():186
> org.apache.drill.exec.physical.impl.BaseRootExec.next():104
> 
> org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():93
> org.apache.drill.exec.physical.impl.BaseRootExec.next():94
> org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
> org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1669
> org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
> org.apache.drill.common.SelfCleaningRunnable.run():38
> 

[GitHub] [drill] vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 
libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538373047
 
 
   Here is the list of things to be done before merging this PR:
   
   - [ ] Cleanup poms:
   - - [ ] Remove commented out lines, revise new dependencies to reduce JDBC 
driver size
   - - [ ] Revert removing excluding `commons-logging`, it will be better to 
add an absent mock class from `commons-logging` to Drill
   - [ ] Remove change with logging test error in `ITTestShadedJar.java`
   - [ ] Test Drill on Hadoop 3 cluster (secure, non-secure modes)
   - [ ] Check with Drill-on-YARN
   - [ ] Test Drill in embedded mode on Windows (test Hadoop win-utils)
   - [ ] Test JDBC client with new Hadoop version
   - [ ] Check whether new Hadoop version API is compatible with the current 
one, in the case if not, decide whether to introduce new profile, use a 
property to specify the version during building or document supported versions
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 
libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538373047
 
 
   Here is the list of things to be done before merging this PR:
   
   - [ ] Cleanup poms:
   - - [ ] Remove commented out lines, revise new dependencies to reduce JDBC 
driver size
   - - [ ] Revert removing excluding `commons-logging`, it will be better to 
add an absent mock class from `commons-logging` to Drill
   - [ ] Remove change with logging test error in `ITTestShadedJar.java`
   - [ ] Test Drill on Hadoop 3 cluster (secure, non-secure modes)
   - [ ] Test Drill in embedded mode on Windows (test Hadoop win-utils)
   - [ ] Test JDBC client with new Hadoop version
   - [ ] Check whether new Hadoop version API is compatible with the current 
one, in the case if not, decide whether to introduce new profile, use a 
property to specify the version during building or document supported versions
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 
libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538373047
 
 
   Here is the list of things to be done before merging this PR:
   
   - [ ] Cleanup poms:
   - - [ ] Remove commented out lines, revise new dependencies to reduce JDBC 
driver size
   - - [ ] Revert removing excluding `commons-logging`, it will be better to 
add an absent mock class from `commons-logging` to Drill
   - [ ] Remove change with logging test error in `ITTestShadedJar.java`
   - [ ] Test Drill on Hadoop 3 cluster (secure, non-secure modes)
   - [ ] Test Drill in embedded mode on Windows (test Hadoop win-utils)
   - [ ] Check whether new Hadoop version API is compatible with the current 
one, in the case if not, decide whether to introduce new profile, use a 
property to specify the version during building or document supported versions
   - [ ] Test JDBC client with new Hadoop version
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] arina-ielchiieva opened a new pull request #1866: DRILL-5983: Add missing nullable Parquet readers for INT and UINT logical types

2019-10-04 Thread GitBox
arina-ielchiieva opened a new pull request #1866: DRILL-5983: Add missing 
nullable Parquet readers for INT and UINT logical types
URL: https://github.com/apache/drill/pull/1866
 
 
   Jira - [DRILL-5983](https://issues.apache.org/jira/browse/DRILL-5983).


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] cgivre commented on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
cgivre commented on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538379596
 
 
   @vvysotskyi 
   Can anyone assist with these tasks, or are you actively working on them?  Do 
you think it would be possible to get this JIRA into Drill 1.17?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Resolved] (DRILL-6885) CTAS for empty output doesn't create parquet file or folder

2019-10-04 Thread Arina Ielchiieva (Jira)


 [ 
https://issues.apache.org/jira/browse/DRILL-6885?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arina Ielchiieva resolved DRILL-6885.
-
Resolution: Fixed

> CTAS for empty output doesn't create parquet file or folder
> ---
>
> Key: DRILL-6885
> URL: https://issues.apache.org/jira/browse/DRILL-6885
> Project: Apache Drill
>  Issue Type: Bug
>  Components: Storage - Parquet
>Affects Versions: 1.14.0
>Reporter: Vitalii Diravka
>Priority: Major
> Fix For: 1.17.0
>
>
> CTAS for empty output can create empty tables based on the empty json or csv 
> files. But it doesn't work for parquet files.
>  See examples below:
> {code:java}
> 0: jdbc:drill:zk=local> use dfs.tmp;
> +---+--+
> |  ok   |   summary|
> +---+--+
> | true  | Default schema changed to [dfs.tmp]  |
> +---+--+
> 1 row selected (0.087 seconds)
> 0: jdbc:drill:zk=local> select * from `empty_dir`;
> +--+
> |  |
> +--+
> +--+
> No rows selected (0.083 seconds)
> 0: jdbc:drill:zk=local> alter session set `store.format` = 'json';
> +---++
> |  ok   |summary |
> +---++
> | true  | store.format updated.  |
> +---++
> 1 row selected (0.079 seconds)
> 0: jdbc:drill:zk=local> create table `empty_json` as select * from 
> `empty_dir`;
> +---++
> | Fragment  | Number of records written  |
> +---++
> | 0_0   | 0  |
> +---++
> 1 row selected (0.128 seconds)
> 0: jdbc:drill:zk=local> select * from `empty_json`;
> +--+
> |  |
> +--+
> +--+
> No rows selected (0.086 seconds)
> 0: jdbc:drill:zk=local> alter session set `store.format` = 'csv';
> +---++
> |  ok   |summary |
> +---++
> | true  | store.format updated.  |
> +---++
> 1 row selected (0.073 seconds)
> 0: jdbc:drill:zk=local> create table `empty_csv` as select * from `empty_dir`;
> +---++
> | Fragment  | Number of records written  |
> +---++
> | 0_0   | 0  |
> +---++
> 1 row selected (0.135 seconds)
> 0: jdbc:drill:zk=local> select * from `empty_csv`;
> +--+
> | columns  |
> +--+
> | []   |
> +--+
> 1 row selected (0.086 seconds)
> 0: jdbc:drill:zk=local> alter session set `store.format` = 'parquet';
> +---++
> |  ok   |summary |
> +---++
> | true  | store.format updated.  |
> +---++
> 1 row selected (0.073 seconds)
> 0: jdbc:drill:zk=local> create table `empty_parquet` as select * from 
> `empty_dir`;
> +---++
> | Fragment  | Number of records written  |
> +---++
> | 0_0   | 0  |
> +---++
> 1 row selected (0.099 seconds)
> 0: jdbc:drill:zk=local> select * from `empty_parquet`;
> 20:41:01.619 [23f692c1-8994-9fc8-2ce4-5fc6135ebcc9:foreman] ERROR 
> o.a.calcite.runtime.CalciteException - 
> org.apache.calcite.sql.validate.SqlValidatorException: Object 'empty_parquet' 
> not found
> 20:41:01.619 [23f692c1-8994-9fc8-2ce4-5fc6135ebcc9:foreman] ERROR 
> o.a.calcite.runtime.CalciteException - 
> org.apache.calcite.runtime.CalciteContextException: From line 1, column 15 to 
> line 1, column 29: Object 'empty_parquet' not found
> 20:41:01.622 [Client-1] ERROR o.a.calcite.runtime.CalciteException - 
> org.apache.calcite.sql.validate.SqlValidatorException: Object 'empty_parquet' 
> not found
> 20:41:01.623 [Client-1] ERROR o.a.calcite.runtime.CalciteException - 
> org.apache.calcite.runtime.CalciteContextException: From line 1, column 15 to 
> line 1, column 29: Object 'empty_parquet' not found: Object 'empty_parquet' 
> not found
> Error: VALIDATION ERROR: From line 1, column 15 to line 1, column 29: Object 
> 'empty_parquet' not found
> [Error Id: 879730dc-aad6-4fc7-9c62-9ad8bbc99d42 on vitalii-pc:31010] 
> (state=,code=0)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (DRILL-4949) Need better handling of empty parquet files

2019-10-04 Thread Arina Ielchiieva (Jira)


 [ 
https://issues.apache.org/jira/browse/DRILL-4949?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arina Ielchiieva resolved DRILL-4949.
-
Resolution: Fixed

> Need better handling of empty parquet files
> ---
>
> Key: DRILL-4949
> URL: https://issues.apache.org/jira/browse/DRILL-4949
> Project: Apache Drill
>  Issue Type: Bug
>  Components: Storage - Parquet
>Affects Versions: 1.9.0
>Reporter: Krystal
>Priority: Major
> Fix For: 1.17.0
>
>
> I have an empty parquet file created from hive.  When I tried to query 
> against this table I got "IllegalArgumentException".
> {code}
> select * from `test_dir/voter_empty`;
> Error: SYSTEM ERROR: IllegalArgumentException: MinorFragmentId 0 has no read 
> entries assigned
>   (org.apache.drill.exec.work.foreman.ForemanException) Unexpected exception 
> during fragment initialization: MinorFragmentId 0 has no read entries assigned
> org.apache.drill.exec.work.foreman.Foreman.run():281
> java.util.concurrent.ThreadPoolExecutor.runWorker():1145
> java.util.concurrent.ThreadPoolExecutor$Worker.run():615
> java.lang.Thread.run():745
>   Caused By (java.lang.IllegalArgumentException) MinorFragmentId 0 has no 
> read entries assigned
> com.google.common.base.Preconditions.checkArgument():122
> org.apache.drill.exec.store.parquet.ParquetGroupScan.getSpecificScan():824
> org.apache.drill.exec.store.parquet.ParquetGroupScan.getSpecificScan():101
> org.apache.drill.exec.planner.fragment.Materializer.visitGroupScan():68
> org.apache.drill.exec.planner.fragment.Materializer.visitGroupScan():35
> org.apache.drill.exec.physical.base.AbstractGroupScan.accept():63
> org.apache.drill.exec.planner.fragment.Materializer.visitOp():102
> org.apache.drill.exec.planner.fragment.Materializer.visitOp():35
> 
> org.apache.drill.exec.physical.base.AbstractPhysicalVisitor.visitProject():79
> org.apache.drill.exec.physical.config.Project.accept():51
> org.apache.drill.exec.planner.fragment.Materializer.visitStore():82
> org.apache.drill.exec.planner.fragment.Materializer.visitStore():35
> 
> org.apache.drill.exec.physical.base.AbstractPhysicalVisitor.visitScreen():202
> org.apache.drill.exec.physical.config.Screen.accept():98
> 
> org.apache.drill.exec.planner.fragment.SimpleParallelizer.generateWorkUnit():283
> 
> org.apache.drill.exec.planner.fragment.SimpleParallelizer.getFragments():127
> org.apache.drill.exec.work.foreman.Foreman.getQueryWorkUnit():596
> org.apache.drill.exec.work.foreman.Foreman.runPhysicalPlan():426
> org.apache.drill.exec.work.foreman.Foreman.runSQL():1010
> org.apache.drill.exec.work.foreman.Foreman.run():264
> java.util.concurrent.ThreadPoolExecutor.runWorker():1145
> java.util.concurrent.ThreadPoolExecutor$Worker.run():615
> java.lang.Thread.run():745 (state=,code=0)
> {code}
> Either drill should block the query and display a user friendly error message 
> or allow the query to run and return empty result.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (DRILL-1834) Misleading error message when querying an empty Parquet file

2019-10-04 Thread Arina Ielchiieva (Jira)


 [ 
https://issues.apache.org/jira/browse/DRILL-1834?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arina Ielchiieva resolved DRILL-1834.
-
Resolution: Fixed

> Misleading error message when querying an empty Parquet file
> 
>
> Key: DRILL-1834
> URL: https://issues.apache.org/jira/browse/DRILL-1834
> Project: Apache Drill
>  Issue Type: Bug
>  Components: Execution - Flow
>Affects Versions: 0.7.0
>Reporter: Aman Sinha
>Priority: Minor
> Fix For: 1.17.0
>
>
> It is possible that a CTAS may fail and create an empty Parquet file.  When 
> we run a query against this file, we get a misleading error message from the 
> planner that hides the original IOException, although the log file does have 
> the original exception: 
> {code:sql}
> 0: jdbc:drill:zk=local> select count(*) from dfs.`/tmp/empty.parquet`;
> Query failed: Query failed: Unexpected exception during fragment 
> initialization: Internal error: Error while applying rule 
> DrillPushProjIntoScan, args 
> [rel#77:ProjectRel.NONE.ANY([]).[](child=rel#76:Subset#0.ENUMERABLE.ANY([]).[],$f0=0),
>  rel#68:EnumerableTableAccessRel.ENUMERABLE.ANY([]).[](table=[dfs, 
> /tmp/empty.parquet])]
> {code}
> The cause of the exception is in the logs: 
> Caused by: java.io.IOException: Could not read footer: 
> java.lang.RuntimeException: file:/tmp/empty.parquet is not a Parquet file 
> (too small)
> at 
> parquet.hadoop.ParquetFileReader.readAllFootersInParallel(ParquetFileReader.java:195)
>  ~[parquet-hadoop-1.5.1-drill-r4.jar:0.7.0-SNAPSHOT]
> at 
> parquet.hadoop.ParquetFileReader.readAllFootersInParallel(ParquetFileReader.java:208)
>  ~[parquet-hadoop-1.5.1-drill-r4.jar:0.7.0-SNAPSHOT]
> at 
> parquet.hadoop.ParquetFileReader.readFooters(ParquetFileReader.java:224) 
> ~[parquet-hadoop-1.5.1-drill-r4.jar:0.7.0-SNAPSHOT]
> at 
> org.apache.drill.exec.store.parquet.ParquetGroupScan.readFooter(ParquetGroupScan.java:208)
>  ~[drill-java-exec-0.7.0-SNAPSHOT-rebuffed.jar:0.7.0-SNAPSHOT]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [drill] vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 
libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538373047
 
 
   Here is the list of things to be done before merging this PR:
   
   - [x] Cleanup poms:
   - - [ ] Remove commented out lines, revise new dependencies to reduce JDBC 
driver size
   - - [ ] Revert removing excluding `commons-logging`, it will be better to 
add an absent mock class from `commons-logging` to Drill
   - [ ] Remove change with logging test error in `ITTestShadedJar.java`
   - [ ] Test Drill on Hadoop 3 cluster (secure, non-secure modes)
   - [ ] Test Drill in embedded mode on Windows (test Hadoop win-utils)
   - [ ] Test JDBC client with new Hadoop version
   - [ ] Check whether new Hadoop version API is compatible with the current 
one, in the case if not, decide whether to introduce new profile, use a 
property to specify the version during building or document supported versions


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
vvysotskyi edited a comment on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 
libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538373047
 
 
   Here is the list of things to be done before merging this PR:
   
   - [ ] Cleanup poms:
   - - [ ] Remove commented out lines, revise new dependencies to reduce JDBC 
driver size
   - - [ ] Revert removing excluding `commons-logging`, it will be better to 
add an absent mock class from `commons-logging` to Drill
   - [ ] Remove change with logging test error in `ITTestShadedJar.java`
   - [ ] Test Drill on Hadoop 3 cluster (secure, non-secure modes)
   - [ ] Test Drill in embedded mode on Windows (test Hadoop win-utils)
   - [ ] Test JDBC client with new Hadoop version
   - [ ] Check whether new Hadoop version API is compatible with the current 
one, in the case if not, decide whether to introduce new profile, use a 
property to specify the version during building or document supported versions


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] vvysotskyi commented on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
vvysotskyi commented on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538373047
 
 
   Here is the list of things to be done before merging this PR:
   
   - [ ] Cleanup poms:
   - - [ ] Remove commented out lines, revise new dependencies to reduce JDBC 
driver size
   - - [ ] Revert removing excluding `commons-logging`, it will be better to 
add an absent mock class from `commons-logging` to Drill
   - [ ] Remove change with logging test error in `ITTestShadedJar.java`
   - [ ] Test Drill on Hadoop 3 cluster (secure, non-secure modes)
   - [ ] Test Drill in embedded mode on Windows (test Hadoop win-utils)
   - [ ] Test JDBC client with new Hadoop version
   - [ ] Check whether new Hadoop version API is compatible with the current 
one, in the case if not, decide whether to introduce new profile, use a 
property to specify the version during building or document supported versions


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [drill] arina-ielchiieva commented on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 libraries

2019-10-04 Thread GitBox
arina-ielchiieva commented on issue #1732: DRILL-6540: Upgrade to HADOOP-3.0 
libraries
URL: https://github.com/apache/drill/pull/1732#issuecomment-538356827
 
 
   @vvysotskyi since you were the reviewer do you recall what else should be 
done to finish this work?
   cc @agozhiy @denysord88 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (DRILL-7396) Exception when trying to access last element of an array with repeated_count

2019-10-04 Thread benj (Jira)
benj created DRILL-7396:
---

 Summary: Exception when trying to access last element of an array 
with repeated_count
 Key: DRILL-7396
 URL: https://issues.apache.org/jira/browse/DRILL-7396
 Project: Apache Drill
  Issue Type: Bug
  Components: Functions - Drill
Affects Versions: 1.16.0
Reporter: benj


Use of array in drill is not friendly
{code:sql}
SELECT (split('a,b,c',','))[0]; /*NOK */
Error: SYSTEM ERROR: ClassCastException: 
org.apache.drill.common.expression.FunctionCall cannot be cast to 
org.apache.drill.common.expression.SchemaPath

/* outer SELECT needed*/
SELECT x[0] FROM (SELECT split('a,b,c',',') x); /* OK */
{code}
And access last element of an array is worse
{code:sql}
SELECT x[repeated_count(x) - 1] AS lasteltidx FROM (SELECT split('a,b,c',',') 
x);
Error: SYSTEM ERROR: ClassCastException: org.apache.calcite.rex.RexCall cannot 
be cast to org.apache.calcite.rex.RexLiteral

/* while */
SELECT x[2] lastelt, (repeated_count(x) - 1) AS lasteltidx FROM (SELECT 
split('a,b,c',',') x);
+-++
| lastelt | lasteltidx |
+-++
| c   | 2  |
+-++
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)