[jira] [Assigned] (IMPALA-10251) test_decimal_queries and test_tpcds_queries may run out of memory on ASAN builds

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-10251:
--

Assignee: (was: Tim Armstrong)

> test_decimal_queries and test_tpcds_queries may run out of memory on ASAN 
> builds
> 
>
> Key: IMPALA-10251
> URL: https://issues.apache.org/jira/browse/IMPALA-10251
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Quanlong Huang
>Priority: Critical
>  Labels: broken-build, flaky
>
> Encountered these test failures due to running out of memory:
> {code:java}
> query_test.test_decimal_queries.TestDecimalQueries.test_queries[protocol: 
> beeswax | exec_option: {'disable_codegen_rows_threshold': 0, 
> 'disable_codegen': 'false', 'decimal_v2': 'true', 'batch_size': 1} | 
> table_format: 
> orc/def/block]query_test.test_tpcds_queries.TestTpcdsQuery.test_tpcds_q51a[protocol:
>  beeswax | exec_option: {'decimal_v2': 0, 'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> parquet/none]query_test.test_decimal_queries.TestDecimalQueries.test_queries[protocol:
>  hs2 | exec_option: {'disable_codegen_rows_threshold': 0, 'disable_codegen': 
> 'true', 'decimal_v2': 'true', 'batch_size': 1} | table_format: 
> parquet/none]query_test.test_tpcds_queries.TestTpcdsDecimalV2Query.test_tpcds_q18a[protocol:
>  beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> parquet/none]query_test.test_tpcds_queries.TestTpcdsDecimalV2Query.test_tpcds_q23_1[protocol:
>  beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> parquet/none]query_test.test_tpcds_queries.TestTpcdsDecimalV2Query.test_tpcds_q4[protocol:
>  beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> parquet/none] {code}
> The error is
> {code}
> query_test/test_decimal_queries.py:55: in test_queries
> self.run_test_case('QueryTest/decimal', vector)
> common/impala_test_suite.py:662: in run_test_case
> result = exec_fn(query, user=test_section.get('USER', '').strip() or None)
> common/impala_test_suite.py:600: in __exec_in_impala
> result = self.__execute_query(target_impalad_client, query, user=user)
> common/impala_test_suite.py:909: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:205: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:187: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:365: in __execute_query
> self.wait_for_finished(handle)
> beeswax/impala_beeswax.py:386: in wait_for_finished
> raise ImpalaBeeswaxException("Query aborted:" + error_log, None)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EQuery aborted:Failed to get minimum memory reservation of 68.01 MB on 
> daemon impala-ec2-centos74-r5-4xlarge-ondemand-109b.vpc.cloudera.com:27001 
> for query 2b41e9326825c48d:15a4d931 due to following error: Failed to 
> increase reservation by 68.01 MB because it would exceed the applicable 
> reservation limit for the "Process" ReservationTracker: 
> reservation_limit=10.20 GB reservation=10.16 GB used_reservation=0 
> child_reservations=10.16 GB
> E   The top 5 queries that allocated memory under this tracker are:
> E   Query(ca4c47a813876f70:d626ca0c): Reservation=9.60 GB 
> ReservationLimit=9.60 GB OtherMemory=115.84 MB Total=9.71 GB Peak=9.71 GB
> E   Query(3f4e984e6b54204d:2597d87f): Reservation=426.77 MB 
> ReservationLimit=9.60 GB OtherMemory=113.19 MB Total=539.95 MB Peak=540.00 MB
> E   Query(6c4aca5cba7141b7:cb0d9a40): Limit=512.00 MB 
> Reservation=68.03 MB ReservationLimit=409.60 MB OtherMemory=2.59 MB 
> Total=70.62 MB Peak=71.93 MB
> E   Query(e84c2ac37bac9309:7eddf8bb): Reservation=14.94 MB 
> ReservationLimit=9.60 GB OtherMemory=5.55 MB Total=20.48 MB Peak=39.46 MB
> E   Query(eb44550f4e6aebbd:4b72c396): Reservation=12.31 MB 
> ReservationLimit=9.60 GB OtherMemory=5.85 MB Total=18.16 MB Peak=18.16 MB
> E   
> E   
> E   
> E   
> E   
> E   Memory is likely oversubscribed. Reducing query concurrency or 
> configuring admission control may help avoid this error.
> {code}



--
This message was sent by Atlassian Jira

[jira] [Assigned] (IMPALA-10344) Error loading functional_orc_def.alltypesaggmultifiles in hive

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-10344:
--

Assignee: (was: Tim Armstrong)

> Error loading functional_orc_def.alltypesaggmultifiles in hive
> --
>
> Key: IMPALA-10344
> URL: https://issues.apache.org/jira/browse/IMPALA-10344
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
> Attachments: hive-metastore.out, hive-server2.out, 
> load-functional-query-exhaustive-hive-generated-orc-def-block.sql.log
>
>
> Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask. 
> Exception when loading 11 partitions in table alltypesaggmultifiles 
> {noformat}
> 0: jdbc:hive2://localhost:11050/default> insert into table 
> functional_orc_def.al 
> ltypesaggmultifiles partition (year, month, day) SELECT id, bool_col, 
> tinyint_co 
> l, smallint_col, int_col, bigint_col, float_col, double_col, date_string_col, 
> st 
> ring_col, timestamp_col, year, month, day FROM 
> functional.alltypesaggmultifiles  
> where id % 4 = 1;
> going to print operations logs
> printed operations logs
> going to print operations logs
> INFO  : Compiling 
> command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9): 
> insert into table functional_orc_def.alltypesaggmultifiles partition (year, 
> month, day) SELECT id, bool_col, tinyint_col, smallint_col, int_col, 
> bigint_col, float_col, double_col, date_string_col, string_col, 
> timestamp_col, year, month, day FROM functional.alltypesaggmultifiles where 
> id % 4 = 1
> INFO  : No Stats for functional@alltypesaggmultifiles, Columns: float_col, 
> int_col, string_col, bool_col, date_string_col, smallint_col, timestamp_col, 
> tinyint_col, bigint_col, id, double_col
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Created Hive schema: Schema(fieldSchemas:[FieldSchema(name:id, 
> type:int, comment:null), FieldSchema(name:bool_col, type:boolean, 
> comment:null), FieldSchema(name:tinyint_col, type:tinyint, comment:null), 
> FieldSchema(name:smallint_col, type:smallint, comment:null), 
> FieldSchema(name:int_col, type:int, comment:null), 
> FieldSchema(name:bigint_col, type:bigint, comment:null), 
> FieldSchema(name:float_col, type:float, comment:null), 
> FieldSchema(name:double_col, type:double, comment:null), 
> FieldSchema(name:date_string_col, type:string, comment:null), 
> FieldSchema(name:string_col, type:string, comment:null), 
> FieldSchema(name:timestamp_col, type:timestamp, comment:null), 
> FieldSchema(name:year, type:int, comment:null), FieldSchema(name:month, 
> type:int, comment:null), FieldSchema(name:day, type:int, comment:null)], 
> properties:null)
> INFO  : Completed compiling 
> command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9); 
> Time taken: 0.036 seconds
> INFO  : Executing 
> command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9): 
> insert into table functional_orc_def.alltypesaggmultifiles partition (year, 
> month, day) SELECT id, bool_col, tinyint_col, smallint_col, int_col, 
> bigint_col, float_col, double_col, date_string_col, string_col, 
> timestamp_col, year, month, day FROM functional.alltypesaggmultifiles where 
> id % 4 = 1
> INFO  : Query ID = jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9
> INFO  : Total jobs = 1
> INFO  : Launching Job 1 out of 1
> INFO  : Starting task [Stage-1:MAPRED] in serial mode
> INFO  : Subscribed to counters: [] for queryId: 
> jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9
> INFO  : Session is already open
> INFO  : Dag name: insert into table functional_orc_def.all...1 (Stage-1)
> INFO  : Status: Running (Executing on YARN cluster with App id 
> application_1605304119768_0012)
> printed operations logs
> --
> VERTICES  MODESTATUS  TOTAL  COMPLETED  
> RUNNING  PENDING  FAILED  KILLED  
> --
> Map 1 .. container SUCCEEDED 42 420   
>  0   0   0  
> Reducer 2container   RUNNING  1  010  
>  0   0  
> --
> VERTICES: 01/02  [=>>-] 97%   ELAPSED 
> TIME: 1.46 s 
> --
> 

[jira] [Resolved] (IMPALA-10235) Averaged timer profile counters can be negative for trivial queries

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10235?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10235.

Fix Version/s: Impala 4.0
 Assignee: Tim Armstrong
   Resolution: Fixed

This was introduced by part 1 of IMPALA-9382 and fixed by this part of the 
below patch.

@@ -1842,7 +2137,8 @@ RuntimeProfileBase::AveragedCounter::GetStats() const {
   if (!vals.empty()) {
 sort(vals.begin(), vals.end());
 result.num_vals = vals.size();
-result.mean = std::accumulate(vals.begin(), vals.end(), 0) / 
result.num_vals;
+T sum = std::accumulate(vals.begin(), vals.end(), 0L);
+result.mean = sum / result.num_vals;
 result.min = vals[0];
 result.max = vals.back();
 int end_idx = vals.size() - 1;
@@ -1902,6 +2198,50 @@ void 
RuntimeProfileBase::AveragedCounter::PrettyPrintImpl(


commit 9429bd779de986d3e61858bef7e258bd73a2cacd
Author: Tim Armstrong 
Date:   Sun May 17 16:37:46 2020 -0700

IMPALA-9382: part 2/3: aggregate profiles sent to coordinator

This reworks the status reporting so that serialized
AggregatedRuntimeProfile objects are sent from executors
to coordinators. These profiles are substantially denser
and faster to process for higher mt_dop values. The aggregation
is also done in a single step, merging the aggregated thrift
profile from the executor directly into the final aggregated
profile, instead of converting it to an unaggregated profile
first.

The changes required were:
* A new Update() method for AggregatedRuntimeProfile that
  updates the profile from a serialised AggregateRuntimeProfile
  for a subset of the instances. The code is generalized from the
  existing InitFromThrift() code path.
* Per-fragment reports included in the status report protobuf
  when --gen_experimental_profile=true.
* Logic on the coordinator that either consumes serialized
  AggregatedRuntimeProfile per fragment, when
  --gen_experimental_profile=true, or consumes a serialized
  RuntimeProfile per finstance otherwise.

This also adds support for event sequences and time series
in the aggregated profile, so the amount of information
in the aggregated profile is now on par with the basic profile.

We also finish off support for JSON profile. The JSON profile is
more stripped down because we do not need to round-trip profiles
via JSON and it is a much less dense profile representation.

Part 3 will clean up and improve the display of the profile.

Testing:
* Add sanity tests for aggregated runtime profile.
* Add unit tests to exercise aggregation of the various counter types
* Ran core tests.

Change-Id: Ic680cbfe94c939c2a8fad9d0943034ed058c6bca
Reviewed-on: http://gerrit.cloudera.org:8080/16057
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> Averaged timer profile counters can be negative for trivial queries
> ---
>
> Key: IMPALA-10235
> URL: https://issues.apache.org/jira/browse/IMPALA-10235
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: newbie, ramp-up
> Fix For: Impala 4.0
>
> Attachments: profile-output.txt
>
>
> Steps to reproduce on master:
> {code:java}
> stakiar @ stakiar-desktop -bash ~/Impala 2020-10-13 11:13:02 master
>  [74] → ./bin/impala-shell.sh -q "select sleep(100) from functional.alltypes 
> limit 25" -p > profile-output.txt
> ...
> Query: select sleep(100) from functional.alltypes limit 25
> Query submitted at: 2020-10-13 11:13:07 (Coordinator: 
> http://stakiar-desktop:25000)
> Query progress can be monitored at: 
> http://stakiar-desktop:25000/query_plan?query_id=694f94671571d4d1:cdec9db9
> Fetched 25 row(s) in 2.64s
> {code}
> Attached the contents of {{profile-output.txt}}
> Relevant portion of the profile:
> {code:java}
> Averaged Fragment F00:(Total: 2s603ms, non-child: 272.519us, % non-child: 
> 0.01%)
> ...
>- CompletionTime: -1665218428.000ns
> ...
>- TotalThreadsTotalWallClockTime: -1686005515.000ns
>  - TotalThreadsSysTime: 0.000ns
>  - TotalThreadsUserTime: 2.151ms
> ...
>- TotalTime: -1691524485.000ns
> {code}
> For whatever reason, this only affects the averaged fragment profile. For 
> this query, there was only one coordinator fragment and thus only one 
> fragment instance. The coordinator fragment instance showed normal timer 
> values:
> {code:java}
> Coordinator Fragment F00:
> ...
>  - CompletionTime: 2s629ms
> ...
>  - TotalThreadsTotalWallClockTime: 2s608ms
>- TotalThreadsSysTime: 0.000ns
>- 

[jira] [Resolved] (IMPALA-6870) SummaryStatsCounter should be included in averaged profile

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-6870?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-6870.
---
Fix Version/s: Impala 4.0
   Resolution: Fixed

commit 9429bd779de986d3e61858bef7e258bd73a2cacd
Author: Tim Armstrong 
Date:   Sun May 17 16:37:46 2020 -0700

IMPALA-9382: part 2/3: aggregate profiles sent to coordinator

This reworks the status reporting so that serialized
AggregatedRuntimeProfile objects are sent from executors
to coordinators. These profiles are substantially denser
and faster to process for higher mt_dop values. The aggregation
is also done in a single step, merging the aggregated thrift
profile from the executor directly into the final aggregated
profile, instead of converting it to an unaggregated profile
first.

The changes required were:
* A new Update() method for AggregatedRuntimeProfile that
  updates the profile from a serialised AggregateRuntimeProfile
  for a subset of the instances. The code is generalized from the
  existing InitFromThrift() code path.
* Per-fragment reports included in the status report protobuf
  when --gen_experimental_profile=true.
* Logic on the coordinator that either consumes serialized
  AggregatedRuntimeProfile per fragment, when
  --gen_experimental_profile=true, or consumes a serialized
  RuntimeProfile per finstance otherwise.

This also adds support for event sequences and time series
in the aggregated profile, so the amount of information
in the aggregated profile is now on par with the basic profile.

We also finish off support for JSON profile. The JSON profile is
more stripped down because we do not need to round-trip profiles
via JSON and it is a much less dense profile representation.

Part 3 will clean up and improve the display of the profile.

Testing:
* Add sanity tests for aggregated runtime profile.
* Add unit tests to exercise aggregation of the various counter types
* Ran core tests.

Change-Id: Ic680cbfe94c939c2a8fad9d0943034ed058c6bca
Reviewed-on: http://gerrit.cloudera.org:8080/16057
Reviewed-by: Impala Public Jenkins 
Tested-by: Impala Public Jenkins 


> SummaryStatsCounter should be included in averaged profile
> --
>
> Key: IMPALA-6870
> URL: https://issues.apache.org/jira/browse/IMPALA-6870
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend
>Affects Versions: Impala 2.12.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: observability
> Fix For: Impala 4.0
>
>
> Summary stats like FooterProcessingTime don't show up in the averaged 
> fragment.  We should be able to merge these counters to produce overall 
> statistics:



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10377) Improve the accuracy of resource estimation

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10377:
---
Target Version: Impala 4.0

> Improve the accuracy of resource estimation
> ---
>
> Key: IMPALA-10377
> URL: https://issues.apache.org/jira/browse/IMPALA-10377
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 3.4.0
>Reporter: liuyao
>Assignee: liuyao
>Priority: Major
>  Labels: estimate, memory, statistics
>   Original Estimate: 120h
>  Remaining Estimate: 120h
>
> PlanNode does not consider some factors when estimating memory, this will 
> cause a large error rate
>  
> AggregationNode
>  
> 1.The memory occupied by hash table's own data structure is not considered. 
> Hash table inserts a new value, which will add a bucket. The size of a bucket 
> is 16 bytes.
> 2.When estimating the NDV of merge aggregation, if there are multiple 
> grouping exprs, it may be divided by the number of Fragment Instances several 
> times, and it should be divided only once.
> 3.When estimating the NDV of merge aggregation, and there are multiple 
> grouping exprs, the estimated memory is much smaller than the actual use.
> 4.If there is no grouping exprs, the estimated memory is much larger than the 
> actual use.
> 5.If the NDV of grouping exprs is very small, the estimated memory is much 
> larger than the actual use.
>  
> SortNode
> 1.Estimate the memory usage of external sort. the estimated memory is much 
> smaller than the actual use.
>  
>  
> HashJoinNode
> 1.The memory occupied by hash table's own data structure is not 
> considered.Hash Table will keep duplicate data, so the size of DuplicateNode 
> should be considered.
> 2.Hash table will create multiple buckets in advance. The size of these 
> buckets should be considered.
>  
> KuduScanNode
> 1.Estimate memory by scanning all columns,the estimated memory is much larger 
> than the actual use.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10156) test_unmatched_schema flaky with wrong results

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10156.

Fix Version/s: Impala 4.0
   Resolution: Fixed

> test_unmatched_schema flaky with wrong results
> --
>
> Key: IMPALA-10156
> URL: https://issues.apache.org/jira/browse/IMPALA-10156
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Thomas Tauber-Marshall
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
> Fix For: Impala 4.0
>
>
> {noformat}
> Error Message
> query_test/test_scanners.py:244: in test_unmatched_schema 
> self.run_test_case('QueryTest/test-unmatched-schema', vector) 
> common/impala_test_suite.py:693: in run_test_case 
> self.__verify_results_and_errors(vector, test_section, result, use_db) 
> common/impala_test_suite.py:529: in __verify_results_and_errors 
> replace_filenames_with_placeholder) common/test_result_verifier.py:456: in 
> verify_raw_results VERIFIER_MAP[verifier](expected, actual) 
> common/test_result_verifier.py:278: in verify_query_result_is_equal 
> assert expected_results == actual_results E   assert Comparing 
> QueryTestResults (expected vs actual): E 1001,'Name1',94611,5000 == 
> 1001,'Name1',94611,5000 E 1002,'Name2',94611,5000 != 
> 1001,'Name1',94611,5000 E 1003,'Name3',94611,5000 != 
> 1002,'Name2',94611,5000 E 1004,'Name4',94611,5000 != 
> 1002,'Name2',94611,5000 E 1005,'Name5',94611,5000 != 
> 1003,'Name3',94611,5000 E 1006,'Name16',94612,15000 != 
> 1003,'Name3',94611,5000 E 1006,'Name16',94612,5000 != 
> 1004,'Name4',94611,5000 E 1006,'Name16',94616,15000 != 
> 1004,'Name4',94611,5000 E 1006,'Name16',94616,5000 != 
> 1005,'Name5',94611,5000 E 1006,'Name6',94616,15000 != 
> 1005,'Name5',94611,5000 E 1006,'Name6',94616,5000 != 
> 1006,'Name16',94612,15000 E 1106,'Name16',94612,15000 != 
> 1006,'Name16',94612,15000 E 1106,'Name16',94612,5000 != 
> 1006,'Name16',94612,5000 E 1106,'Name16',94616,15000 != 
> 1006,'Name16',94612,5000 E 1106,'Name16',94616,5000 != 
> 1006,'Name16',94616,15000 E 1106,'Name6',94612,15000 != 
> 1006,'Name16',94616,15000 E 1106,'Name6',94612,5000 != 
> 1006,'Name16',94616,5000 E 1106,'Name6',94616,15000 != 
> 1006,'Name16',94616,5000 E 1106,'Name6',94616,5000 != 
> 1006,'Name6',94616,15000 E None != 1006,'Name6',94616,15000 E None != 
> 1006,'Name6',94616,5000 E None != 1006,'Name6',94616,5000 E None != 
> 1106,'Name16',94612,15000 E None != 1106,'Name16',94612,15000 E None 
> != 1106,'Name16',94612,5000 E None != 1106,'Name16',94612,5000 E None 
> != 1106,'Name16',94616,15000 E None != 1106,'Name16',94616,15000 E 
> None != 1106,'Name16',94616,5000 E None != 1106,'Name16',94616,5000 E 
> None != 1106,'Name6',94612,15000 E None != 1106,'Name6',94612,15000 E 
> None != 1106,'Name6',94612,5000 E None != 1106,'Name6',94612,5000 E 
> None != 1106,'Name6',94616,15000 E None != 1106,'Name6',94616,15000 E 
> None != 1106,'Name6',94616,5000 E None != 1106,'Name6',94616,5000 E 
> Number of rows returned (expected vs actual): 19 != 38
> Stacktrace
> query_test/test_scanners.py:244: in test_unmatched_schema
> self.run_test_case('QueryTest/test-unmatched-schema', vector)
> common/impala_test_suite.py:693: in run_test_case
> self.__verify_results_and_errors(vector, test_section, result, use_db)
> common/impala_test_suite.py:529: in __verify_results_and_errors
> replace_filenames_with_placeholder)
> common/test_result_verifier.py:456: in verify_raw_results
> VERIFIER_MAP[verifier](expected, actual)
> common/test_result_verifier.py:278: in verify_query_result_is_equal
> assert expected_results == actual_results
> E   assert Comparing QueryTestResults (expected vs actual):
> E 1001,'Name1',94611,5000 == 1001,'Name1',94611,5000
> E 1002,'Name2',94611,5000 != 1001,'Name1',94611,5000
> E 1003,'Name3',94611,5000 != 1002,'Name2',94611,5000
> E 1004,'Name4',94611,5000 != 1002,'Name2',94611,5000
> E 1005,'Name5',94611,5000 != 1003,'Name3',94611,5000
> E 1006,'Name16',94612,15000 != 1003,'Name3',94611,5000
> E 1006,'Name16',94612,5000 != 1004,'Name4',94611,5000
> E 1006,'Name16',94616,15000 != 1004,'Name4',94611,5000
> E 1006,'Name16',94616,5000 != 1005,'Name5',94611,5000
> E 1006,'Name6',94616,15000 != 1005,'Name5',94611,5000
> E 1006,'Name6',94616,5000 != 1006,'Name16',94612,15000
> E 1106,'Name16',94612,15000 != 1006,'Name16',94612,15000
> E 1106,'Name16',94612,5000 != 1006,'Name16',94612,5000
> E 1106,'Name16',94616,15000 != 1006,'Name16',94612,5000
> E 1106,'Name16',94616,5000 != 

[jira] [Updated] (IMPALA-10377) Improve the accuracy of resource estimation

2020-12-04 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10377:
---
Fix Version/s: (was: Impala 4.0)

> Improve the accuracy of resource estimation
> ---
>
> Key: IMPALA-10377
> URL: https://issues.apache.org/jira/browse/IMPALA-10377
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 3.4.0
>Reporter: liuyao
>Assignee: liuyao
>Priority: Major
>  Labels: estimate, memory, statistics
>   Original Estimate: 120h
>  Remaining Estimate: 120h
>
> PlanNode does not consider some factors when estimating memory, this will 
> cause a large error rate
>  
> AggregationNode
>  
> 1.The memory occupied by hash table's own data structure is not considered. 
> Hash table inserts a new value, which will add a bucket. The size of a bucket 
> is 16 bytes.
> 2.When estimating the NDV of merge aggregation, if there are multiple 
> grouping exprs, it may be divided by the number of Fragment Instances several 
> times, and it should be divided only once.
> 3.When estimating the NDV of merge aggregation, and there are multiple 
> grouping exprs, the estimated memory is much smaller than the actual use.
> 4.If there is no grouping exprs, the estimated memory is much larger than the 
> actual use.
> 5.If the NDV of grouping exprs is very small, the estimated memory is much 
> larger than the actual use.
>  
> SortNode
> 1.Estimate the memory usage of external sort. the estimated memory is much 
> smaller than the actual use.
>  
>  
> HashJoinNode
> 1.The memory occupied by hash table's own data structure is not 
> considered.Hash Table will keep duplicate data, so the size of DuplicateNode 
> should be considered.
> 2.Hash table will create multiple buckets in advance. The size of these 
> buckets should be considered.
>  
> KuduScanNode
> 1.Estimate memory by scanning all columns,the estimated memory is much larger 
> than the actual use.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10370) Thrift Message Incompatibility Detector

2020-12-04 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10370?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17244104#comment-17244104
 ] 

Tim Armstrong commented on IMPALA-10370:


[~zagol] you can exclude BackendGflags.thrift, CatalogInternalService.thrift, 
CatalogService.thrift, DataSinks.thrift, Data.thrift, Descriptors.thrift, 
Exprs.thrift, Frontend.thrift, ImpalaInternalService.thrift, JniCatalog.thrift, 
Logging.thrift, NetworkTest.thrift, Partitions.thrift, Planner.thrift, 
PlanNodes.thrift, ResourceProfile.thrift, Results.thrift, 
SqlConstraints.thrift, StatestoreService.thrift, Zip.thrift

Those should only be used internally. Some files like Types.thrift have 
definitions that are used both internally and externally, but I think if you 
exclude the above files it will avoid most false positives.

> Thrift Message Incompatibility Detector
> ---
>
> Key: IMPALA-10370
> URL: https://issues.apache.org/jira/browse/IMPALA-10370
> Project: IMPALA
>  Issue Type: Improvement
>Reporter: junwen yang
>Priority: Major
> Attachments: impala_thrift_incompatibility.txt
>
>
> Regarding the issue  IMPALA-8243 caused by the incompatibility of thrift 
> message, we have created a static checker which keeps track of the thrift 
> file change, and detect potential incompatibility:
>  # Add/delete required field.  The thrift guidelines suggests _Any new fields 
> that you add should be optional_.
>  # The tag number of a field has been changed. Also, the thrift guidelines 
> suggests _Don’t change the numeric tags for any existing fields_.
>  # A  required field has been changed to optional, or an optional field has 
> been changed to required. According to the guidelines , _*Required Is 
> Forever* You should be very careful about marking fields as required. If at 
> some point you wish to stop writing or sending a required field, it will be 
> problematic to change the field to an optional field — old readers will 
> consider messages without this field to be incomplete and may reject or drop 
> them unintentionally. You should consider writing application-specific custom 
> validation routines for your buffers instead. Some have come to the 
> conclusion that using required does more harm than good; they prefer to use 
> only optional. However, this view is not universal._
> We have applied our checker on the frequently maintained IMPALA versions: 
> refs/tags/2.10.0, refs/tags/2.11.0, refs/tags/2.12.0, refs/tags/2.7.0, 
> refs/tags/2.8.0, refs/tags/2.9.0, refs/tags/3.0.0, refs/tags/3.0.1, 
> refs/tags/3.1.0, refs/tags/3.2.0, refs/tags/3.3.0, refs/tags/3.4.0, we found 
> more than 1000 problems as attached. 
> The results reported by our checker got confirmed by developers of HBASE and 
> our checker is requested by them, which can be found at HBASE-25340.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10377) Improve the accuracy of resource estimation

2020-12-04 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17244097#comment-17244097
 ] 

Tim Armstrong commented on IMPALA-10377:


[~liuyao] thanks for investigating all these issues, it would be great to make 
these more accurate. We need to be a bit careful if the estimates change a lot 
because it can disrupt production workloads. Look forward to seeing a patch.

> Improve the accuracy of resource estimation
> ---
>
> Key: IMPALA-10377
> URL: https://issues.apache.org/jira/browse/IMPALA-10377
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 3.4.0
>Reporter: liuyao
>Assignee: liuyao
>Priority: Major
>  Labels: estimate, memory, statistics
> Fix For: Impala 4.0
>
>   Original Estimate: 120h
>  Remaining Estimate: 120h
>
> PlanNode does not consider some factors when estimating memory, this will 
> cause a large error rate
>  
> AggregationNode
>  
> 1.The memory occupied by hash table's own data structure is not considered. 
> Hash table inserts a new value, which will add a bucket. The size of a bucket 
> is 16 bytes.
> 2.When estimating the NDV of merge aggregation, if there are multiple 
> grouping exprs, it may be divided by the number of Fragment Instances several 
> times, and it should be divided only once.
> 3.When estimating the NDV of merge aggregation, and there are multiple 
> grouping exprs, the estimated memory is much smaller than the actual use.
> 4.If there is no grouping exprs, the estimated memory is much larger than the 
> actual use.
> 5.If the NDV of grouping exprs is very small, the estimated memory is much 
> larger than the actual use.
>  
> SortNode
> 1.Estimate the memory usage of external sort. the estimated memory is much 
> smaller than the actual use.
>  
>  
> HashJoinNode
> 1.The memory occupied by hash table's own data structure is not 
> considered.Hash Table will keep duplicate data, so the size of DuplicateNode 
> should be considered.
> 2.Hash table will create multiple buckets in advance. The size of these 
> buckets should be considered.
>  
> KuduScanNode
> 1.Estimate memory by scanning all columns,the estimated memory is much larger 
> than the actual use.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10377) Improve the accuracy of resource estimation

2020-12-04 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17244093#comment-17244093
 ] 

Tim Armstrong commented on IMPALA-10377:


I had some ideas about improving accuracy of estimates for the preaggregations, 
I'll link those here.

> Improve the accuracy of resource estimation
> ---
>
> Key: IMPALA-10377
> URL: https://issues.apache.org/jira/browse/IMPALA-10377
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Affects Versions: Impala 3.4.0
>Reporter: liuyao
>Assignee: liuyao
>Priority: Major
>  Labels: estimate, memory, statistics
> Fix For: Impala 4.0
>
>   Original Estimate: 120h
>  Remaining Estimate: 120h
>
> PlanNode does not consider some factors when estimating memory, this will 
> cause a large error rate
>  
> AggregationNode
>  
> 1.The memory occupied by hash table's own data structure is not considered. 
> Hash table inserts a new value, which will add a bucket. The size of a bucket 
> is 16 bytes.
> 2.When estimating the NDV of merge aggregation, if there are multiple 
> grouping exprs, it may be divided by the number of Fragment Instances several 
> times, and it should be divided only once.
> 3.When estimating the NDV of merge aggregation, and there are multiple 
> grouping exprs, the estimated memory is much smaller than the actual use.
> 4.If there is no grouping exprs, the estimated memory is much larger than the 
> actual use.
> 5.If the NDV of grouping exprs is very small, the estimated memory is much 
> larger than the actual use.
>  
> SortNode
> 1.Estimate the memory usage of external sort. the estimated memory is much 
> smaller than the actual use.
>  
>  
> HashJoinNode
> 1.The memory occupied by hash table's own data structure is not 
> considered.Hash Table will keep duplicate data, so the size of DuplicateNode 
> should be considered.
> 2.Hash table will create multiple buckets in advance. The size of these 
> buckets should be considered.
>  
> KuduScanNode
> 1.Estimate memory by scanning all columns,the estimated memory is much larger 
> than the actual use.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10371) test_java_udfs crash impalad if result spooling is enabled

2020-12-03 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17243726#comment-17243726
 ] 

Tim Armstrong commented on IMPALA-10371:


I think that must be it! I'm concerned now that we might have other instances 
of this issue though - we don't have a documented or enforced invariant that 
the evaluator must be opened on the same thread as it's evaluated on.

I.e. this seems like more of a bug in HiveUdfCall than in the caller.

> test_java_udfs crash impalad if result spooling is enabled
> --
>
> Key: IMPALA-10371
> URL: https://issues.apache.org/jira/browse/IMPALA-10371
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Riza Suminto
>Priority: Major
> Attachments: 46a19881-resolved.txt, hs_err_pid12878.log
>
>
> The following test query from TestUdfExecution::test_java_udfs crash impalad 
> when result spooling is enabled.
> {code:java}
> select throws_exception() from functional.alltypestiny{code}
> The following is a truncated JVM crash log related to the crash
> {code:java}
> ---  T H R E A D  ---Current thread 
> (0x0fb4c000):  JavaThread "Thread-700" [_thread_in_native, id=30853, 
> stack(0x7f79715ff000,0x7f7971dff000)]Stack: 
> [0x7f79715ff000,0x7f7971dff000],  sp=0x7f7971dfa280,  free 
> space=8172k
> Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native 
> code)
> V  [libjvm.so+0xb6b032]
> V  [libjvm.so+0x4f14bd]
> V  [libjvm.so+0x80fa8f]
> V  [libjvm.so+0x7e0991]
> V  [libjvm.so+0x69fa10]
> j  
> org.apache.impala.TestUdfException.evaluate()Lorg/apache/hadoop/io/BooleanWritable;+9
> v  ~StubRoutines::call_stub
> V  [libjvm.so+0x6af9ba]
> V  [libjvm.so+0xa1def8]
> V  [libjvm.so+0xa1f8d5]
> V  [libjvm.so+0x7610f8]  JVM_InvokeMethod+0x128
> J 2286  
> sun.reflect.NativeMethodAccessorImpl.invoke0(Ljava/lang/reflect/Method;Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object;
>  (0 bytes) @ 0x7f7acb553ced [0x7f7acb553c00+0xed]
> J 6921 C2 
> sun.reflect.NativeMethodAccessorImpl.invoke(Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object;
>  (104 bytes) @ 0x7f7acbd1de38 [0x7f7acbd1ddc0+0x78]
> J 3645 C2 org.apache.impala.hive.executor.UdfExecutor.evaluate()V (396 bytes) 
> @ 0x7f7acaf6e894 [0x7f7acaf6e640+0x254]
> v  ~StubRoutines::call_stub
> V  [libjvm.so+0x6af9ba]
> V  [libjvm.so+0x72c046]
> V  [libjvm.so+0x730523]
> C  0x7f7ab4c5d0d0
> C  [impalad+0x26a2648]  
> impala::ScalarExprEvaluator::GetValue(impala::ScalarExpr const&, 
> impala::TupleRow const*)+0x7a
> C  [impalad+0x26a25cb]  
> impala::ScalarExprEvaluator::GetValue(impala::TupleRow const*)+0x2b
> C  [impalad+0x21f4f78]  
> impala::AsciiQueryResultSet::AddRows(std::vector  std::allocator > const&, impala::RowBatch*, 
> int, int)+0x4c2
> C  [impalad+0x25c5862]  
> impala::BufferedPlanRootSink::GetNext(impala::RuntimeState*, 
> impala::QueryResultSet*, int, bool*, long)+0x70c
> C  [impalad+0x296cf17]  impala::Coordinator::GetNext(impala::QueryResultSet*, 
> int, bool*, long)+0x557
> C  [impalad+0x219f5fe]  impala::ClientRequestState::FetchRowsInternal(int, 
> impala::QueryResultSet*, long)+0x6b2
> C  [impalad+0x219d98e]  impala::ClientRequestState::FetchRows(int, 
> impala::QueryResultSet*, long)+0x46
> C  [impalad+0x21c1d29]  
> impala::ImpalaServer::FetchInternal(impala::TUniqueId, bool, int, 
> beeswax::Results*)+0x717
> C  [impalad+0x21bbde9]  impala::ImpalaServer::fetch(beeswax::Results&, 
> beeswax::QueryHandle const&, bool, int)+0x577
> {code}
> If result spooling is enabled, BufferedPlanRootSink will be used and 
> ScalarExprEvaluation will be called in BufferedPlanRootSink::GetNext, leading 
> to this crash.
> Without result spooling, BlockingPlanRootSink will be used and 
> ScalarExprEvaluation is called in BlockingPlanRootSink::Send. No crash happen 
> when result spooling is disabled.
> Attached is the full JVM crash log and resolved minidump.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10371) test_java_udfs crash impalad if result spooling is enabled

2020-12-02 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17242797#comment-17242797
 ] 

Tim Armstrong commented on IMPALA-10371:


[~rizaon] I believe it should be OK for it to be evaluated by any of the query 
execution threads. ScalarExprEvaluator is not thread-safe though, so possibly 
if it's being accessed by multiple threads concurrently, then there could be a 
bug.

> test_java_udfs crash impalad if result spooling is enabled
> --
>
> Key: IMPALA-10371
> URL: https://issues.apache.org/jira/browse/IMPALA-10371
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.4.0
>Reporter: Riza Suminto
>Priority: Major
> Attachments: 46a19881-resolved.txt, hs_err_pid12878.log
>
>
> The following test query from TestUdfExecution::test_java_udfs crash impalad 
> when result spooling is enabled.
> {code:java}
> select throws_exception() from functional.alltypestiny{code}
> The following is a truncated JVM crash log related to the crash
> {code:java}
> ---  T H R E A D  ---Current thread 
> (0x0fb4c000):  JavaThread "Thread-700" [_thread_in_native, id=30853, 
> stack(0x7f79715ff000,0x7f7971dff000)]Stack: 
> [0x7f79715ff000,0x7f7971dff000],  sp=0x7f7971dfa280,  free 
> space=8172k
> Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native 
> code)
> V  [libjvm.so+0xb6b032]
> V  [libjvm.so+0x4f14bd]
> V  [libjvm.so+0x80fa8f]
> V  [libjvm.so+0x7e0991]
> V  [libjvm.so+0x69fa10]
> j  
> org.apache.impala.TestUdfException.evaluate()Lorg/apache/hadoop/io/BooleanWritable;+9
> v  ~StubRoutines::call_stub
> V  [libjvm.so+0x6af9ba]
> V  [libjvm.so+0xa1def8]
> V  [libjvm.so+0xa1f8d5]
> V  [libjvm.so+0x7610f8]  JVM_InvokeMethod+0x128
> J 2286  
> sun.reflect.NativeMethodAccessorImpl.invoke0(Ljava/lang/reflect/Method;Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object;
>  (0 bytes) @ 0x7f7acb553ced [0x7f7acb553c00+0xed]
> J 6921 C2 
> sun.reflect.NativeMethodAccessorImpl.invoke(Ljava/lang/Object;[Ljava/lang/Object;)Ljava/lang/Object;
>  (104 bytes) @ 0x7f7acbd1de38 [0x7f7acbd1ddc0+0x78]
> J 3645 C2 org.apache.impala.hive.executor.UdfExecutor.evaluate()V (396 bytes) 
> @ 0x7f7acaf6e894 [0x7f7acaf6e640+0x254]
> v  ~StubRoutines::call_stub
> V  [libjvm.so+0x6af9ba]
> V  [libjvm.so+0x72c046]
> V  [libjvm.so+0x730523]
> C  0x7f7ab4c5d0d0
> C  [impalad+0x26a2648]  
> impala::ScalarExprEvaluator::GetValue(impala::ScalarExpr const&, 
> impala::TupleRow const*)+0x7a
> C  [impalad+0x26a25cb]  
> impala::ScalarExprEvaluator::GetValue(impala::TupleRow const*)+0x2b
> C  [impalad+0x21f4f78]  
> impala::AsciiQueryResultSet::AddRows(std::vector  std::allocator > const&, impala::RowBatch*, 
> int, int)+0x4c2
> C  [impalad+0x25c5862]  
> impala::BufferedPlanRootSink::GetNext(impala::RuntimeState*, 
> impala::QueryResultSet*, int, bool*, long)+0x70c
> C  [impalad+0x296cf17]  impala::Coordinator::GetNext(impala::QueryResultSet*, 
> int, bool*, long)+0x557
> C  [impalad+0x219f5fe]  impala::ClientRequestState::FetchRowsInternal(int, 
> impala::QueryResultSet*, long)+0x6b2
> C  [impalad+0x219d98e]  impala::ClientRequestState::FetchRows(int, 
> impala::QueryResultSet*, long)+0x46
> C  [impalad+0x21c1d29]  
> impala::ImpalaServer::FetchInternal(impala::TUniqueId, bool, int, 
> beeswax::Results*)+0x717
> C  [impalad+0x21bbde9]  impala::ImpalaServer::fetch(beeswax::Results&, 
> beeswax::QueryHandle const&, bool, int)+0x577
> {code}
> If result spooling is enabled, BufferedPlanRootSink will be used and 
> ScalarExprEvaluation will be called in BufferedPlanRootSink::GetNext, leading 
> to this crash.
> Without result spooling, BlockingPlanRootSink will be used and 
> ScalarExprEvaluation is called in BlockingPlanRootSink::Send. No crash happen 
> when result spooling is disabled.
> Attached is the full JVM crash log and resolved minidump.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10370) Thrift Message Incompatibility Detector

2020-12-01 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10370?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17241834#comment-17241834
 ] 

Tim Armstrong commented on IMPALA-10370:


[~jwjwyoung] thanks, this seems like a useful tool. 

Impala's use of Thrift is a bit unique in that it's used extensively for 
internal communication between C++/Java and within a cluster, where we don't 
support inter-version interoperability. So the thrift types that are only used 
internally aren't kept backwards compatible (attempting to do that somehow 
would require writing a bunch of unreachable code).

It looks like we did break compatibility with TAggregateFunction/TFunction in 
the past and this caught that (that type is serialized and written to the hive 
metastore) but I don't think we can fix that at this point.

> Thrift Message Incompatibility Detector
> ---
>
> Key: IMPALA-10370
> URL: https://issues.apache.org/jira/browse/IMPALA-10370
> Project: IMPALA
>  Issue Type: Improvement
>Reporter: junwen yang
>Priority: Major
> Attachments: impala_thrift_incompatibility.txt
>
>
> Regarding the issue  IMPALA-8243 caused by the incompatibility of thrift 
> message, we have created a static checker which keeps track of the thrift 
> file change, and detect potential incompatibility:
>  # Add/delete required field.  The thrift guidelines suggests _Any new fields 
> that you add should be optional_.
>  # The tag number of a field has been changed. Also, the thrift guidelines 
> suggests _Don’t change the numeric tags for any existing fields_.
>  # A  required field has been changed to optional, or an optional field has 
> been changed to required. According to the guidelines , _*Required Is 
> Forever* You should be very careful about marking fields as required. If at 
> some point you wish to stop writing or sending a required field, it will be 
> problematic to change the field to an optional field — old readers will 
> consider messages without this field to be incomplete and may reject or drop 
> them unintentionally. You should consider writing application-specific custom 
> validation routines for your buffers instead. Some have come to the 
> conclusion that using required does more harm than good; they prefer to use 
> only optional. However, this view is not universal._
> We have applied our checker on the frequently maintained IMPALA versions: 
> refs/tags/2.10.0, refs/tags/2.11.0, refs/tags/2.12.0, refs/tags/2.7.0, 
> refs/tags/2.8.0, refs/tags/2.9.0, refs/tags/3.0.0, refs/tags/3.0.1, 
> refs/tags/3.1.0, refs/tags/3.2.0, refs/tags/3.3.0, refs/tags/3.4.0, we found 
> more than 1000 problems as attached. 
> The results reported by our checker got confirmed by developers of HBASE and 
> our checker is requested by them, which can be found at HBASE-25340.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10366) TestRuntimeProfile.test_runtime_profile_aggregated failed on master core run

2020-12-01 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10366.

Fix Version/s: Impala 4.0
   Resolution: Fixed

> TestRuntimeProfile.test_runtime_profile_aggregated failed on master core run
> 
>
> Key: IMPALA-10366
> URL: https://issues.apache.org/jira/browse/IMPALA-10366
> Project: IMPALA
>  Issue Type: Bug
>  Components: Distributed Exec
>Affects Versions: Impala 3.4.0
>Reporter: Aman Sinha
>Assignee: Tim Armstrong
>Priority: Major
> Fix For: Impala 4.0
>
>
> Test failure on an impala-asf-master-core-erasure-coding build:
> {noformat}
> custom_cluster.test_runtime_profile.TestRuntimeProfile.test_runtime_profile_aggregated[protocol:
>  beeswax |
>  exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False,
>  'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> text/none
> {noformat}
> The output is very verbose.  Here's the first part of the Stacktrace:
> {noformat}
> custom_cluster/test_runtime_profile.py:46: in test_runtime_profile_aggregated
> self.run_test_case('runtime-profile-aggregated', vector)
> common/impala_test_suite.py:718: in run_test_case
> update_section=pytest.config.option.update_results)
> common/test_result_verifier.py:612: in verify_runtime_profile
> actual))
> E   AssertionError: Did not find matches for lines in runtime profile:
> E   EXPECTED LINES:
> E   row_regex: .*NumBackends: 4 .*
> E   row_regex: .*Fragment F00 \[4 instances\].*
> E   row_regex: .*Table Name\[0-3\]: tpch.*lineitem.*
> E   
> E   ACTUAL PROFILE:
> E   Query (id=7b4476dc2da1ab29:ca76ce49):
> E DEBUG MODE WARNING: Query profile created while running a DEBUG build 
> of Impala. Use RELEASE builds to measure query performance.
> E  - InactiveTotalTime: 0.000ns
> E  - TotalTime: 0.000ns
> E Summary:
> E   Session ID: d94934f8d0ba24cb:79a237b3d0ef5a9a
> E   Session Type: BEESWAX
> E   Start Time: 2020-11-28 21:44:31.917844000
> E   End Time: 2020-11-28 21:44:36.293303000
> E   Query Type: QUERY
> E   Query State: FINISHED
> E   Impala Query State: FINISHED
> E   Query Status: OK
> E   Impala Version: impalad version 4.0.0-SNAPSHOT DEBUG (build 
> a4cf449c88ef3fe08db9abbad82664b99382014c)
> E   User: jenkins
> E   Connected User: jenkins
> E   Delegated User: 
> E   Network Address: ::1:45158
> E   Default Db: tpch
> E   Sql Statement: select STRAIGHT_JOIN count(distinct l_partkey)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-10367) Impala-shell internal error - UnboundLocalError, local variable 'retry_msg' referenced before assign

2020-11-30 Thread Tim Armstrong (Jira)
Tim Armstrong created IMPALA-10367:
--

 Summary: Impala-shell internal error - UnboundLocalError, local 
variable 'retry_msg' referenced before assign
 Key: IMPALA-10367
 URL: https://issues.apache.org/jira/browse/IMPALA-10367
 Project: IMPALA
  Issue Type: Bug
  Components: Clients
Affects Versions: Impala 4.0
Reporter: Tim Armstrong
Assignee: Abhishek Rawat


I saw this error when cancelling a query from impala-shell.
{noformat}
[localhost.EXAMPLE.COM:21050] tpch_parquet> explain select STRAIGHT_JOIN 
count(distinct l_partkey)
from lineitem join [BROADCAST] supplier on l_suppkey = s_suppkey;
Query: explain select STRAIGHT_JOIN count(distinct l_partkey)
from lineitem join [BROADCAST] supplier on l_suppkey = s_suppkey
Caught exception [Errno 4] Interrupted system call, type= 
in ExecuteStatement.
 Cancelling Query
Caught exception [Errno 4] Interrupted system call, type= 
in CloseSession.
Warning: close session RPC failed: [Errno 4] Interrupted system call, 
Opened TCP connection to localhost.EXAMPLE.COM:21050
Caught exception [Errno 4] Interrupted system call, type= 
in ExecuteStatement.
Caught exception [Errno 4] Interrupted system call, type= 
in CloseSession.
Warning: close session RPC failed: [Errno 4] Interrupted system call, 
Error connecting: UnboundLocalError, local variable 'retry_msg' referenced 
before assign
{noformat}

This code was introduced by IMPALA-9466. I looked at the code briefly and 
AFAICT it happens when the retry logic is disabled but we hit a 
TTransportException or similar error.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-10366) TestRuntimeProfile.test_runtime_profile_aggregated failed on master core run

2020-11-30 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-10366 started by Tim Armstrong.
--
> TestRuntimeProfile.test_runtime_profile_aggregated failed on master core run
> 
>
> Key: IMPALA-10366
> URL: https://issues.apache.org/jira/browse/IMPALA-10366
> Project: IMPALA
>  Issue Type: Bug
>  Components: Distributed Exec
>Affects Versions: Impala 3.4.0
>Reporter: Aman Sinha
>Assignee: Tim Armstrong
>Priority: Major
>
> Test failure on an impala-asf-master-core-erasure-coding build:
> {noformat}
> custom_cluster.test_runtime_profile.TestRuntimeProfile.test_runtime_profile_aggregated[protocol:
>  beeswax |
>  exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False,
>  'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> text/none
> {noformat}
> The output is very verbose.  Here's the first part of the Stacktrace:
> {noformat}
> custom_cluster/test_runtime_profile.py:46: in test_runtime_profile_aggregated
> self.run_test_case('runtime-profile-aggregated', vector)
> common/impala_test_suite.py:718: in run_test_case
> update_section=pytest.config.option.update_results)
> common/test_result_verifier.py:612: in verify_runtime_profile
> actual))
> E   AssertionError: Did not find matches for lines in runtime profile:
> E   EXPECTED LINES:
> E   row_regex: .*NumBackends: 4 .*
> E   row_regex: .*Fragment F00 \[4 instances\].*
> E   row_regex: .*Table Name\[0-3\]: tpch.*lineitem.*
> E   
> E   ACTUAL PROFILE:
> E   Query (id=7b4476dc2da1ab29:ca76ce49):
> E DEBUG MODE WARNING: Query profile created while running a DEBUG build 
> of Impala. Use RELEASE builds to measure query performance.
> E  - InactiveTotalTime: 0.000ns
> E  - TotalTime: 0.000ns
> E Summary:
> E   Session ID: d94934f8d0ba24cb:79a237b3d0ef5a9a
> E   Session Type: BEESWAX
> E   Start Time: 2020-11-28 21:44:31.917844000
> E   End Time: 2020-11-28 21:44:36.293303000
> E   Query Type: QUERY
> E   Query State: FINISHED
> E   Impala Query State: FINISHED
> E   Query Status: OK
> E   Impala Version: impalad version 4.0.0-SNAPSHOT DEBUG (build 
> a4cf449c88ef3fe08db9abbad82664b99382014c)
> E   User: jenkins
> E   Connected User: jenkins
> E   Delegated User: 
> E   Network Address: ::1:45158
> E   Default Db: tpch
> E   Sql Statement: select STRAIGHT_JOIN count(distinct l_partkey)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-8962) FETCH_ROWS_TIMEOUT_MS should apply before rows are available

2020-11-30 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-8962?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17241021#comment-17241021
 ] 

Tim Armstrong commented on IMPALA-8962:
---

[~arawat] [~asherman] is this something we've seen before with JDBC?

> FETCH_ROWS_TIMEOUT_MS should apply before rows are available
> 
>
> Key: IMPALA-8962
> URL: https://issues.apache.org/jira/browse/IMPALA-8962
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Reporter: Sahil Takiar
>Assignee: Sahil Takiar
>Priority: Major
> Fix For: Impala 3.4.0
>
>
> IMPALA-7312 added a fetch timeout controlled by the query option 
> {{FETCH_ROWS_TIMEOUT_MS}}. The issue is that the timeout only applies after 
> the *first* batch of rows are available. The issue is that both Beeswax and 
> HS2 clients call {{request_state->BlockOnWait}} inside 
> {{ImpalaServer::FetchInternal}}. The call to {{BlockOnWait}} blocks until 
> rows are ready to be consumed via {{ClientRequestState::FetchRows}}.
> So clients can still end up blocking indefinitely waiting for the first row 
> batch to appear.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-1792) ImpalaODBC: Can not get the value in the SQLGetData(m-x th column) after the SQLBindCol(m th column)

2020-11-30 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-1792?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17241011#comment-17241011
 ] 

Tim Armstrong commented on IMPALA-1792:
---

 [^odbc.c]  [^odbc.ini]  [^odbcinst.ini] 

{noformat}
gcc odbc.c -L/opt/cloudera/impalaodbc/lib/64/ -lclouderaimpalaodbc64 -o odbc
ODBCINI=~/odbc.ini ODBCINSTINI=~/odbcinst.ini 
LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libodbcinst.so \
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/opt/cloudera/impalaodbc/lib/64/ \
./odbc
{noformat}

> ImpalaODBC: Can not get the value in the SQLGetData(m-x th column) after the 
> SQLBindCol(m th column)
> 
>
> Key: IMPALA-1792
> URL: https://issues.apache.org/jira/browse/IMPALA-1792
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: Impala 2.1
> Environment: OS: CentOS release 6.4 (Final)
> Impala: 2.1.0
> ImpalaODBC: 2.5.23
>Reporter: Mitsuhiro Koga
>Priority: Critical
>  Labels: correctness, downgraded, impala, odbc
> Attachments: odbc.c, odbc.ini, odbcinst.ini
>
>
> Steps to reproduce
> # Create table and insert data in impala.
> {code:sql}
> create table t (c1 string, c2 string, c3 string);
> insert into t (c1, c2, c3) values ('AAA', 'BBB', 'CCC');
> {code}
> # Query the t table with the following code.
> {code}
> #include 
> #include 
> #include 
> #include 
> int main() {
> SQLHENV env;
> SQLHDBC dbc;
> SQLHSTMT stmt;
> SQLRETURN ret;
> SQLAllocHandle(SQL_HANDLE_ENV, SQL_NULL_HANDLE, );
> SQLSetEnvAttr(env, SQL_ATTR_ODBC_VERSION, (void *) SQL_OV_ODBC3, 0); 
> SQLAllocHandle(SQL_HANDLE_DBC, env, );
> SQLDriverConnect(dbc, (SQLCHAR*)NULL, (SQLCHAR*)"DSN=Impala", SQL_NTS, 
> NULL, 0, NULL, SQL_DRIVER_COMPLETE);
> SQLAllocHandle(SQL_HANDLE_STMT, dbc, );
> SQLExecDirect(stmt, (SQLCHAR*)"select c1, c2, c3 from t", SQL_NTS);
> char szCol1[11];
> char szCol2[11];
> char szCol3[11];
> SQLLEN nColLen1;
> SQLLEN nColLen2;
> SQLLEN nColLen3;
> /*/
> / Bind 2nd column /
> /*/
> SQLBindCol(stmt, 2, SQL_C_CHAR, szCol2, sizeof(szCol2), );
> ret = SQLFetch(stmt);
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> //
> / Get 1st column /
> //
> ret = SQLGetData(stmt, 1, SQL_C_CHAR, szCol1, sizeof(szCol1), 
> );
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> printf("c1: %s\n", szCol1);
> } else {
> printf("no data\n");
> }
> printf("c2: %s\n", szCol2);
> //
> / Get 3rd column /
> //
> ret = SQLGetData(stmt, 3, SQL_C_CHAR, szCol3, sizeof(szCol3), 
> );
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> printf("c3: %s\n", szCol3);
> } else {
> printf("no data\n");
> }
> } else {
> printf("no row\n");
> }
> return 0;
> }
> {code}
> Expected result
> {code}
> c1: AAA
> c2: BBB
> c3: CCC
> {code}
> Actual result
> {code}
> no data
> c2: BBB
> c3: CCC
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-1792) ImpalaODBC: Can not get the value in the SQLGetData(m-x th column) after the SQLBindCol(m th column)

2020-11-30 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-1792?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17241007#comment-17241007
 ] 

Tim Armstrong commented on IMPALA-1792:
---

I tested this with ODBC driver version 2.6.11 and got the expected output, so 
it is fixed in the driver.

> ImpalaODBC: Can not get the value in the SQLGetData(m-x th column) after the 
> SQLBindCol(m th column)
> 
>
> Key: IMPALA-1792
> URL: https://issues.apache.org/jira/browse/IMPALA-1792
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: Impala 2.1
> Environment: OS: CentOS release 6.4 (Final)
> Impala: 2.1.0
> ImpalaODBC: 2.5.23
>Reporter: Mitsuhiro Koga
>Priority: Critical
>  Labels: correctness, downgraded, impala, odbc
> Attachments: odbc.c, odbc.ini, odbcinst.ini
>
>
> Steps to reproduce
> # Create table and insert data in impala.
> {code:sql}
> create table t (c1 string, c2 string, c3 string);
> insert into t (c1, c2, c3) values ('AAA', 'BBB', 'CCC');
> {code}
> # Query the t table with the following code.
> {code}
> #include 
> #include 
> #include 
> #include 
> int main() {
> SQLHENV env;
> SQLHDBC dbc;
> SQLHSTMT stmt;
> SQLRETURN ret;
> SQLAllocHandle(SQL_HANDLE_ENV, SQL_NULL_HANDLE, );
> SQLSetEnvAttr(env, SQL_ATTR_ODBC_VERSION, (void *) SQL_OV_ODBC3, 0); 
> SQLAllocHandle(SQL_HANDLE_DBC, env, );
> SQLDriverConnect(dbc, (SQLCHAR*)NULL, (SQLCHAR*)"DSN=Impala", SQL_NTS, 
> NULL, 0, NULL, SQL_DRIVER_COMPLETE);
> SQLAllocHandle(SQL_HANDLE_STMT, dbc, );
> SQLExecDirect(stmt, (SQLCHAR*)"select c1, c2, c3 from t", SQL_NTS);
> char szCol1[11];
> char szCol2[11];
> char szCol3[11];
> SQLLEN nColLen1;
> SQLLEN nColLen2;
> SQLLEN nColLen3;
> /*/
> / Bind 2nd column /
> /*/
> SQLBindCol(stmt, 2, SQL_C_CHAR, szCol2, sizeof(szCol2), );
> ret = SQLFetch(stmt);
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> //
> / Get 1st column /
> //
> ret = SQLGetData(stmt, 1, SQL_C_CHAR, szCol1, sizeof(szCol1), 
> );
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> printf("c1: %s\n", szCol1);
> } else {
> printf("no data\n");
> }
> printf("c2: %s\n", szCol2);
> //
> / Get 3rd column /
> //
> ret = SQLGetData(stmt, 3, SQL_C_CHAR, szCol3, sizeof(szCol3), 
> );
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> printf("c3: %s\n", szCol3);
> } else {
> printf("no data\n");
> }
> } else {
> printf("no row\n");
> }
> return 0;
> }
> {code}
> Expected result
> {code}
> c1: AAA
> c2: BBB
> c3: CCC
> {code}
> Actual result
> {code}
> no data
> c2: BBB
> c3: CCC
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-1792) ImpalaODBC: Can not get the value in the SQLGetData(m-x th column) after the SQLBindCol(m th column)

2020-11-30 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-1792?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-1792:
--
Attachment: odbcinst.ini
odbc.ini
odbc.c

> ImpalaODBC: Can not get the value in the SQLGetData(m-x th column) after the 
> SQLBindCol(m th column)
> 
>
> Key: IMPALA-1792
> URL: https://issues.apache.org/jira/browse/IMPALA-1792
> Project: IMPALA
>  Issue Type: Bug
>  Components: Clients
>Affects Versions: Impala 2.1
> Environment: OS: CentOS release 6.4 (Final)
> Impala: 2.1.0
> ImpalaODBC: 2.5.23
>Reporter: Mitsuhiro Koga
>Priority: Critical
>  Labels: correctness, downgraded, impala, odbc
> Attachments: odbc.c, odbc.ini, odbcinst.ini
>
>
> Steps to reproduce
> # Create table and insert data in impala.
> {code:sql}
> create table t (c1 string, c2 string, c3 string);
> insert into t (c1, c2, c3) values ('AAA', 'BBB', 'CCC');
> {code}
> # Query the t table with the following code.
> {code}
> #include 
> #include 
> #include 
> #include 
> int main() {
> SQLHENV env;
> SQLHDBC dbc;
> SQLHSTMT stmt;
> SQLRETURN ret;
> SQLAllocHandle(SQL_HANDLE_ENV, SQL_NULL_HANDLE, );
> SQLSetEnvAttr(env, SQL_ATTR_ODBC_VERSION, (void *) SQL_OV_ODBC3, 0); 
> SQLAllocHandle(SQL_HANDLE_DBC, env, );
> SQLDriverConnect(dbc, (SQLCHAR*)NULL, (SQLCHAR*)"DSN=Impala", SQL_NTS, 
> NULL, 0, NULL, SQL_DRIVER_COMPLETE);
> SQLAllocHandle(SQL_HANDLE_STMT, dbc, );
> SQLExecDirect(stmt, (SQLCHAR*)"select c1, c2, c3 from t", SQL_NTS);
> char szCol1[11];
> char szCol2[11];
> char szCol3[11];
> SQLLEN nColLen1;
> SQLLEN nColLen2;
> SQLLEN nColLen3;
> /*/
> / Bind 2nd column /
> /*/
> SQLBindCol(stmt, 2, SQL_C_CHAR, szCol2, sizeof(szCol2), );
> ret = SQLFetch(stmt);
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> //
> / Get 1st column /
> //
> ret = SQLGetData(stmt, 1, SQL_C_CHAR, szCol1, sizeof(szCol1), 
> );
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> printf("c1: %s\n", szCol1);
> } else {
> printf("no data\n");
> }
> printf("c2: %s\n", szCol2);
> //
> / Get 3rd column /
> //
> ret = SQLGetData(stmt, 3, SQL_C_CHAR, szCol3, sizeof(szCol3), 
> );
> if (ret == SQL_SUCCESS || ret == SQL_SUCCESS_WITH_INFO) {
> printf("c3: %s\n", szCol3);
> } else {
> printf("no data\n");
> }
> } else {
> printf("no row\n");
> }
> return 0;
> }
> {code}
> Expected result
> {code}
> c1: AAA
> c2: BBB
> c3: CCC
> {code}
> Actual result
> {code}
> no data
> c2: BBB
> c3: CCC
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10351) Enable mt_dop for DML

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10351?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10351.

Fix Version/s: Impala 4.0
   Resolution: Fixed

> Enable mt_dop for DML
> -
>
> Key: IMPALA-10351
> URL: https://issues.apache.org/jira/browse/IMPALA-10351
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: multithreading
> Fix For: Impala 4.0
>
>
> We don't have any pending issues for MT_DOP with DML and have sufficient 
> functional testing, so I think we can go ahead and enable it. IMPALA-8125 and 
> query hints give us a safety valve in case we overparallelize.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-9812) Remove --unlock_mt_dop and--mt_dop_auto_fallback

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-9812.
---
Fix Version/s: Impala 4.0
   Resolution: Fixed

> Remove --unlock_mt_dop and--mt_dop_auto_fallback 
> -
>
> Key: IMPALA-9812
> URL: https://issues.apache.org/jira/browse/IMPALA-9812
> Project: IMPALA
>  Issue Type: Sub-task
>Reporter: Tim Armstrong
>Priority: Minor
> Fix For: Impala 4.0
>
>
> These flags will become ineffective when DML is supported. We should clean up 
> all references and move them to the flag graveyard.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10216) BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10216.

Fix Version/s: Impala 4.0
   Resolution: Fixed

> BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds
> --
>
> Key: IMPALA-10216
> URL: https://issues.apache.org/jira/browse/IMPALA-10216
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: flaky
> Fix For: Impala 4.0
>
> Attachments: LastTest.log, LastTestsFailed.log, 
> buffer-pool-test.ERROR, buffer-pool-test.INFO, buffer-pool-test.WARNING, 
> buffer-pool-test.xml, impala-cdpd-master-staging-core-tsan_111.tar.gz
>
>
> Only seen this once so far:
> {code}
> BufferPoolTest.WriteErrorBlacklistCompression
> Error Message
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> Stacktrace
> Impala/be/src/runtime/bufferpool/buffer-pool-test.cc:1764
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10357) A test case in test_basic_joins() seems flaky

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10357?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10357.

Resolution: Duplicate

This looks like a duplicate of IMPALA-10336 - it's not returning the expected 
error but instead this incorrect error.

> A test case in test_basic_joins() seems flaky
> -
>
> Key: IMPALA-10357
> URL: https://issues.apache.org/jira/browse/IMPALA-10357
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Fang-Yu Rao
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky-test
> Attachments: excerpted_impalad.INFO, excerpted_impalad_node1.INFO
>
>
> We found in a recent build that a test case in 
> [test_basic_joins()|https://github.com/apache/impala/blob/master/tests/query_test/test_join_queries.py#L71]
>  does not fail with the expected error. This test reads the test case at 
> [joins.test|https://github.com/apache/impala/blob/master/testdata/workloads/functional-query/queries/QueryTest/joins.test#L846-L850]
>  and verifies if the returned result is "{{Debug Action: FAIL}}".
> {noformat}
> Error Message
> query_test/test_join_queries.py:71: in test_basic_joins 
> self.run_test_case('QueryTest/joins', new_vector) 
> common/impala_test_suite.py:668: in run_test_case 
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db) 
> common/impala_test_suite.py:485: in __verify_exceptions (expected_str, 
> actual_str) E   AssertionError: Unexpected exception string. Expected: Debug 
> Action: FAIL E   Not found in actual: ImpalaBeeswaxException: Query 
> aborted:ExecQueryFInstances rpc query_id=9f4e869839247e75:d01b0093 
> failed: Exec() rpc failed: Aborted: ExecQueryFInstances RPC to 
> 127.0.0.1:27002 is cancelled in state SENT
> {noformat}
> {noformat}
> Stacktrace
> query_test/test_join_queries.py:71: in test_basic_joins
> self.run_test_case('QueryTest/joins', new_vector)
> common/impala_test_suite.py:668: in run_test_case
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db)
> common/impala_test_suite.py:485: in __verify_exceptions
> (expected_str, actual_str)
> E   AssertionError: Unexpected exception string. Expected: Debug Action: FAIL
> E   Not found in actual: ImpalaBeeswaxException: Query 
> aborted:ExecQueryFInstances rpc query_id=9f4e869839247e75:d01b0093 
> failed: Exec() rpc failed: Aborted: ExecQueryFInstances RPC to 
> 127.0.0.1:27002 is cancelled in state SENT
> {noformat}
> {noformat}
> Standard Error
> set debug_action="2:PREPARE:FAIL|COORD_BEFORE_EXEC_RPC:JITTER@100@0.3";
> -- 2020-11-23 23:27:30,175 INFO MainThread: Started query 
> 1a49038843b6b4da:30f62281
> -- executing against localhost:21000
> select 1 from alltypestiny a join alltypestiny b on a.id != b.id;
> -- 2020-11-23 23:27:30,188 INFO MainThread: Started query 
> 9f4e869839247e75:d01b0093
> -- executing against localhost:21000
> {noformat}
> According to the output to standard error, two related queries are of query 
> id's "{{1a49038843b6b4da:30f62281}}" and 
> "{{9f4e869839247e75:d01b0093}}".
> Using these two query id's, we excerpted the related log entries from two log 
> files of impalad's.
> Since this test was recently revised in IMPALA-9309, maybe [~tarmstrong] 
> could offer some insight into it. Assigned it to Tim for now but feel free to 
> re-assign it as you see appropriate. Also let me know if some other logs are 
> needed for investigation. Thanks!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-10216) BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-10216 started by Tim Armstrong.
--
> BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds
> --
>
> Key: IMPALA-10216
> URL: https://issues.apache.org/jira/browse/IMPALA-10216
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: flaky
> Attachments: LastTest.log, LastTestsFailed.log, 
> buffer-pool-test.ERROR, buffer-pool-test.INFO, buffer-pool-test.WARNING, 
> buffer-pool-test.xml, impala-cdpd-master-staging-core-tsan_111.tar.gz
>
>
> Only seen this once so far:
> {code}
> BufferPoolTest.WriteErrorBlacklistCompression
> Error Message
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> Stacktrace
> Impala/be/src/runtime/bufferpool/buffer-pool-test.cc:1764
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10216) BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds

2020-11-24 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238358#comment-17238358
 ] 

Tim Armstrong commented on IMPALA-10216:


So from the failed assertion we have:
{noformat}
TmpFilePaths(pages[NO_ERROR_QUERY])
  
/tmp/buffer-pool-test.1/impala-scratch/:_3083829c-75fc-4c44-812a-bb5a4c067c88
  
/tmp/buffer-pool-test.1/impala-scratch/:_3083829c-75fc-4c44-812a-bb5a4c067c88


DumpScratchDir(error_dir)
/tmp/buffer-pool-test.0/impala-scratch:
  :_b049d991-adb2-4845-991f-2dc287c0c90a
  :_20081b32-bef6-4909-8bff-4fd61ec8ca0d
{noformat}

buffer-pool-test.INFO for reference (it looks normal)
{noformat}
I1123 23:09:20.500730  9357 tmp-file-mgr.cc:254] Using scratch directory 
/tmp/buffer-pool-test.0/impala-scratch on disk 0 limit: 8589934592.00 GB
I1123 23:09:20.506762  9357 tmp-file-mgr.cc:254] Using scratch directory 
/tmp/buffer-pool-test.1/impala-scratch on disk 0 limit: 8589934592.00 GB
I1123 23:09:20.545521  9357 buffer-pool-test.cc:1727] Manager 0 Block 0 backed 
by file 
/tmp/buffer-pool-test.0/impala-scratch/:_b049d991-adb2-4845-991f-2dc287c0c90a
I1123 23:09:20.545562  9357 buffer-pool-test.cc:1727] Manager 0 Block 1 backed 
by file 
/tmp/buffer-pool-test.1/impala-scratch/:_cd33af91-4e27-412a-ae45-f16061b5ef56
I1123 23:09:20.556377  9357 buffer-pool-test.cc:1727] Manager 1 Block 0 backed 
by file 
/tmp/buffer-pool-test.1/impala-scratch/:_3083829c-75fc-4c44-812a-bb5a4c067c88
I1123 23:09:20.556414  9357 buffer-pool-test.cc:1727] Manager 1 Block 1 backed 
by file 
/tmp/buffer-pool-test.0/impala-scratch/:_20081b32-bef6-4909-8bff-4fd61ec8ca0d
I1123 23:09:20.558903  9357 buffer-pool-test.cc:336] Injected fault by removing 
file permissions 
/tmp/buffer-pool-test.0/impala-scratch/:_b049d991-adb2-4845-991f-2dc287c0c90a
E1123 23:09:20.559283   451 tmp-file-mgr.cc:359] Error for temporary file 
'/tmp/buffer-pool-test.0/impala-scratch/:_b049d991-adb2-4845-991f-2dc287c0c90a':
 Disk I/O error on 
impala-ec2-centos74-m5-4xlarge-ondemand-1898.vpc.cloudera.com:27000: open() 
failed for 
/tmp/buffer-pool-test.0/impala-scratch/:_b049d991-adb2-4845-991f-2dc287c0c90a.
 Access denied for the process' user errno=13
I1123 23:09:20.563634  9357 buffer-pool-test.cc:1781] Newly created page backed 
by file 
/tmp/buffer-pool-test.1/impala-scratch/:_cd33af91-4e27-412a-ae45-f16061b5ef56
I1123 23:09:20.563666  9357 buffer-pool-test.cc:1781] Newly created page backed 
by file 
/tmp/buffer-pool-test.1/impala-scratch/:_cd33af91-4e27-412a-ae45-f16061b5ef56
I1123 23:09:20.606637  9357 krpc-data-stream-mgr.cc:436] Waiting for 
data-stream-mgr maintenance thread...
I1123 23:09:20.606671  9357 krpc-data-stream-mgr.cc:438] Waiting for 
deserialization thread pool...
{noformat}

It's weird that both pages got backed by the same temp dir, because the 
allocation should be round-robin. II looked at the code and I do think there's 
some actual non-determinism here because of the way Pin() runs asynchronously 
until the actual buffer is requested - it may not actually complete for one of 
these pages. I think if one thread went down this code path on one page but not 
the other, the allocation may not be round robin -
https://github.com/apache/impala/blob/master/be/src/runtime/bufferpool/buffer-pool.cc#L207

> BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds
> --
>
> Key: IMPALA-10216
> URL: https://issues.apache.org/jira/browse/IMPALA-10216
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: flaky
> Attachments: LastTest.log, LastTestsFailed.log, 
> buffer-pool-test.ERROR, buffer-pool-test.INFO, buffer-pool-test.WARNING, 
> buffer-pool-test.xml, impala-cdpd-master-staging-core-tsan_111.tar.gz
>
>
> Only seen this once so far:
> {code}
> BufferPoolTest.WriteErrorBlacklistCompression
> Error Message
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> Stacktrace
> Impala/be/src/runtime/bufferpool/buffer-pool-test.cc:1764
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: 

[jira] [Commented] (IMPALA-10216) BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds

2020-11-24 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238350#comment-17238350
 ] 

Tim Armstrong commented on IMPALA-10216:


[~fangyurao] thanks for attaching, I'll take a look.

FWIW the error logs in the ERROR file are likely unrelated - these tests do 
deliberately induce errors so that's expected.

> BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds
> --
>
> Key: IMPALA-10216
> URL: https://issues.apache.org/jira/browse/IMPALA-10216
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: flaky
> Attachments: LastTest.log, LastTestsFailed.log, 
> buffer-pool-test.ERROR, buffer-pool-test.INFO, buffer-pool-test.WARNING, 
> buffer-pool-test.xml, impala-cdpd-master-staging-core-tsan_111.tar.gz
>
>
> Only seen this once so far:
> {code}
> BufferPoolTest.WriteErrorBlacklistCompression
> Error Message
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> Stacktrace
> Impala/be/src/runtime/bufferpool/buffer-pool-test.cc:1764
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-10356) Analyzed query in explain plan is not quite right for insert with values clause

2020-11-24 Thread Tim Armstrong (Jira)
Tim Armstrong created IMPALA-10356:
--

 Summary: Analyzed query in explain plan is not quite right for 
insert with values clause
 Key: IMPALA-10356
 URL: https://issues.apache.org/jira/browse/IMPALA-10356
 Project: IMPALA
  Issue Type: Bug
  Components: Frontend
Affects Versions: Impala 4.0
Reporter: Tim Armstrong


In impala-shell:
{noformat}
create table double_tbl (d double) stored as textfile;
set explain_level=2;
explain insert into double_tbl values (-0.43149576573887316);
{noformat}

{noformat}
+--+
| Explain String
   |
+--+
| Max Per-Host Resource Reservation: Memory=0B Threads=1
   |
| Per-Host Resource Estimates: Memory=10MB  
   |
| Codegen disabled by planner   
   |
| Analyzed query: SELECT CAST(-0.43149576573887316 AS DECIMAL(17,17)) UNION 
SELECT |
| CAST(-0.43149576573887316 AS DECIMAL(17,17))  
   |
|   
   |
| F00:PLAN FRAGMENT [UNPARTITIONED] hosts=1 instances=1 
   |
| |  Per-Host Resources: mem-estimate=8B mem-reservation=0B 
thread-reservation=1   |
| WRITE TO HDFS [default.double_tbl, OVERWRITE=false]   
   |
| |  partitions=1   
   |
| |  output exprs: CAST(-0.43149576573887316 AS DOUBLE) 
   |
| |  mem-estimate=8B mem-reservation=0B thread-reservation=0
   |
| | 
   |
| 00:UNION  
   |
|constant-operands=1
   |
|mem-estimate=0B mem-reservation=0B thread-reservation=0
   |
|tuple-ids=0 row-size=8B cardinality=1  
   |
|in pipelines:
   |
+--+

{noformat}

The analyzed query does not make sense. We should investigate and fix it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-9453) S3 build failed with many strange symptoms

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-9453:
--
Priority: Critical  (was: Blocker)

> S3 build failed with many strange symptoms
> --
>
> Key: IMPALA-9453
> URL: https://issues.apache.org/jira/browse/IMPALA-9453
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.4.0
>Reporter: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, crash
> Attachments: impalad.ERROR, impalad_node1.ERROR, impalad_node2.ERROR
>
>
> There were a lot of incorrect results:
> {noformat}
> uery_test/test_mt_dop.py:49: in test_mt_dop 
> self.run_test_case('QueryTest/mt-dop', new_vector) 
> common/impala_test_suite.py:690: in run_test_case 
> self.__verify_results_and_errors(vector, test_section, result, use_db) 
> common/impala_test_suite.py:523: in __verify_results_and_errors 
> replace_filenames_with_placeholder) common/test_result_verifier.py:456: in 
> verify_raw_results VERIFIER_MAP[verifier](expected, actual) 
> common/test_result_verifier.py:278: in verify_query_result_is_equal 
> assert expected_results == actual_results E   assert Comparing 
> QueryTestResults (expected vs actual): E 7300 != 6990
> Stacktrace
> query_test/test_mt_dop.py:49: in test_mt_dop
> self.run_test_case('QueryTest/mt-dop', new_vector)
> common/impala_test_suite.py:690: in run_test_case
> self.__verify_results_and_errors(vector, test_section, result, use_db)
> common/impala_test_suite.py:523: in __verify_results_and_errors
> replace_filenames_with_placeholder)
> common/test_result_verifier.py:456: in verify_raw_results
> VERIFIER_MAP[verifier](expected, actual)
> common/test_result_verifier.py:278: in verify_query_result_is_equal
> assert expected_results == actual_results
> E   assert Comparing QueryTestResults (expected vs actual):
> E 7300 != 6990
> Standard Error
> ERROR:test_configuration:Comparing QueryTestResults (expected vs actual):
> 7300 != 6990
> {noformat}
> The impalads eventually crashed:
> {noformat}
> F0302 00:50:55.607841   483 parquet-page-reader.cc:67] 
> e24eb0839fa75423:8ac0bf730002] Check failed: col_end < 
> file_desc.file_length (7010 vs. 7010) 
> *** Check failure stack trace: ***
> @  0x4f7277c  google::LogMessage::Fail()
> @  0x4f74021  google::LogMessage::SendToLog()
> @  0x4f72156  google::LogMessage::Flush()
> @  0x4f7571d  google::LogMessageFatal::~LogMessageFatal()
> @  0x2e3a520  impala::ParquetPageReader::InitColumnChunk()
> @  0x2e37dee  impala::ParquetColumnChunkReader::InitColumnChunk()
> @  0x2cd8000  impala::BaseScalarColumnReader::Reset()
> @  0x2c91239  impala::HdfsParquetScanner::InitScalarColumns()
> @  0x2c8775a  impala::HdfsParquetScanner::NextRowGroup()
> @  0x2c85c2a  impala::HdfsParquetScanner::GetNextInternal()
> @  0x2c8403e  impala::HdfsParquetScanner::ProcessSplit()
> @  0x28b826b  impala::HdfsScanNode::ProcessSplit()
> @  0x28b7440  impala::HdfsScanNode::ScannerThread()
> @  0x28b679d  
> _ZZN6impala12HdfsScanNode22ThreadTokenAvailableCbEPNS_18ThreadResourcePoolEENKUlvE_clEv
> @  0x28b8d91  
> _ZN5boost6detail8function26void_function_obj_invoker0IZN6impala12HdfsScanNode22ThreadTokenAvailableCbEPNS3_18ThreadResourcePoolEEUlvE_vE6invokeERNS1_15function_bufferE
> @  0x20aa1e9  boost::function0<>::operator()()
> @  0x266a84a  impala::Thread::SuperviseThread()
> @  0x2672ace  boost::_bi::list5<>::operator()<>()
> @  0x26729f2  boost::_bi::bind_t<>::operator()()
> @  0x26729b5  boost::detail::thread_data<>::run()
> @  0x3e98b19  thread_proxy
> @ 0x7f8c2ccefe24  start_thread
> @ 0x7f8c2985b34c  __clone
> {noformat}
> {noformat}
> F0302 00:50:41.466794 32643 parquet-page-reader.cc:67] 
> dd48a46583bea9c8:e3be6412] Check failed: col_end < 
> file_desc.file_length (7010 vs. 7010) 
> *** Check failure stack trace: ***
> @  0x4f7277c  google::LogMessage::Fail()
> @  0x4f74021  google::LogMessage::SendToLog()
> @  0x4f72156  google::LogMessage::Flush()
> @  0x4f7571d  google::LogMessageFatal::~LogMessageFatal()
> @  0x2e3a520  impala::ParquetPageReader::InitColumnChunk()
> @  0x2e37dee  impala::ParquetColumnChunkReader::InitColumnChunk()
> @  0x2cd8000  impala::BaseScalarColumnReader::Reset()
> @  0x2c91239  impala::HdfsParquetScanner::InitScalarColumns()
> @  0x2c8775a  impala::HdfsParquetScanner::NextRowGroup()
> @  0x2c85c2a  

[jira] [Resolved] (IMPALA-9050) test_scanners.TestScanRangeLengths.test_scan_ranges is flaky for kudu

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-9050.
---
Fix Version/s: Impala 4.0
   Resolution: Fixed

> test_scanners.TestScanRangeLengths.test_scan_ranges is flaky for kudu
> -
>
> Key: IMPALA-9050
> URL: https://issues.apache.org/jira/browse/IMPALA-9050
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.4.0
>Reporter: Yongzhi Chen
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: build-failure, flaky
> Fix For: Impala 4.0
>
>
> Pre-commit build failed for unrelated change:
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/1405/testReport/junit/query_test.test_scanners/TestScanRangeLengths/test_scan_ranges_max_scan_range_length__512___protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__kudu_none_/
> query_test.test_scanners.TestScanRangeLengths.test_scan_ranges[max_scan_range_length:
>  512 | protocol: beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> kudu/none] 
> Error Message
> query_test/test_scanners.py:937: in test_scan_ranges 
> self.run_test_case('QueryTest/hdfs-tiny-scan', vector) 
> common/impala_test_suite.py:621: in run_test_case result = exec_fn(query, 
> user=test_section.get('USER', '').strip() or None) 
> common/impala_test_suite.py:556: in __exec_in_impala result = 
> self.__execute_query(target_impalad_client, query, user=user) 
> common/impala_test_suite.py:893: in __execute_query return 
> impalad_client.execute(query, user=user) common/impala_connection.py:205: in 
> execute return self.__beeswax_client.execute(sql_stmt, user=user) 
> beeswax/impala_beeswax.py:205: in execute result = 
> self.fetch_results(query_string, handle) beeswax/impala_beeswax.py:451: in 
> fetch_results exec_result = self.__fetch_results(query_handle, max_rows) 
> beeswax/impala_beeswax.py:462: in __fetch_results results = 
> self.__do_rpc(lambda: self.imp_service.fetch(handle, False, fetch_rows)) 
> beeswax/impala_beeswax.py:519: in __do_rpc raise 
> ImpalaBeeswaxException(self.__build_error_message(b), b) E   
> ImpalaBeeswaxException: ImpalaBeeswaxException: EINNER EXCEPTION:  'beeswaxd.ttypes.BeeswaxException'> EMESSAGE: Unable to open Kudu table: 
> Network error: failed to read from TLS socket (remote: 172.18.0.1:7051): 
> Cannot send after transport endpoint shutdown (error 108)
> Stacktrace
> query_test/test_scanners.py:937: in test_scan_ranges
> self.run_test_case('QueryTest/hdfs-tiny-scan', vector)
> common/impala_test_suite.py:621: in run_test_case
> result = exec_fn(query, user=test_section.get('USER', '').strip() or None)
> common/impala_test_suite.py:556: in __exec_in_impala
> result = self.__execute_query(target_impalad_client, query, user=user)
> common/impala_test_suite.py:893: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:205: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:205: in execute
> result = self.fetch_results(query_string, handle)
> beeswax/impala_beeswax.py:451: in fetch_results
> exec_result = self.__fetch_results(query_handle, max_rows)
> beeswax/impala_beeswax.py:462: in __fetch_results
> results = self.__do_rpc(lambda: self.imp_service.fetch(handle, False, 
> fetch_rows))
> beeswax/impala_beeswax.py:519: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: Unable to open Kudu table: Network error: failed to read from 
> TLS socket (remote: 172.18.0.1:7051): Cannot send after transport endpoint 
> shutdown (error 108)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-4238) custom_cluster/test_client_ssl.py TestClientSsl.test_ssl AssertionError: SIGINT was not caught by shell within 30s

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-4238?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-4238.
---
Fix Version/s: Impala 4.0
   Resolution: Fixed

> custom_cluster/test_client_ssl.py TestClientSsl.test_ssl AssertionError: 
> SIGINT was not caught by shell within 30s
> --
>
> Key: IMPALA-4238
> URL: https://issues.apache.org/jira/browse/IMPALA-4238
> Project: IMPALA
>  Issue Type: Bug
>  Components: Security
>Affects Versions: Impala 2.8.0, Impala 2.10.0
>Reporter: Harrison Sheinblatt
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
> Fix For: Impala 4.0
>
>
> asf master core test failure: 
> http://sandbox.jenkins.sf.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core/540/
> http://sandbox.jenkins.sf.cloudera.com/job/impala-umbrella-build-and-test/4921/console
> {noformat}
> 08:18:51 === FAILURES 
> ===
> 08:18:51  TestClientSsl.test_ssl[exec_option: {'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0, 'batch_size': 0, 
> 'num_nodes': 0} | table_format: text/none] 
> 08:18:51 
> 08:18:51 self = 
> 08:18:51 vector = 
> 08:18:51 
> 08:18:51 @pytest.mark.execute_serially
> 08:18:51 
> @CustomClusterTestSuite.with_args("--ssl_server_certificate=%s/server-cert.pem
>  "
> 08:18:51   
> "--ssl_private_key=%s/server-key.pem"
> 08:18:51   % (CERT_DIR, CERT_DIR))
> 08:18:51 def test_ssl(self, vector):
> 08:18:51 
> 08:18:51   self._verify_negative_cases()
> 08:18:51   # TODO: This is really two different tests, but the custom 
> cluster takes too long to
> 08:18:51   # start. Make it so that custom clusters can be specified 
> across test suites.
> 08:18:51   self._validate_positive_cases("%s/server-cert.pem" % 
> self.CERT_DIR)
> 08:18:51 
> 08:18:51   # No certificate checking: will accept any cert.
> 08:18:51   self._validate_positive_cases()
> 08:18:51 
> 08:18:51   # Test cancelling a query
> 08:18:51   impalad = ImpaladService(socket.getfqdn())
> 08:18:51   impalad.wait_for_num_in_flight_queries(0)
> 08:18:51   p = ImpalaShell(args="--ssl")
> 08:18:51   p.send_cmd("SET DEBUG_ACTION=0:OPEN:WAIT")
> 08:18:51   p.send_cmd("select count(*) from functional.alltypes")
> 08:18:51   impalad.wait_for_num_in_flight_queries(1)
> 08:18:51 
> 08:18:51   LOG = logging.getLogger('test_client_ssl')
> 08:18:51   LOG.info("Cancelling query")
> 08:18:51   num_tries = 0
> 08:18:51   # In practice, sending SIGINT to the shell process doesn't 
> always seem to get caught
> 08:18:51   # (and a search shows up some bugs in Python where SIGINT 
> might be ignored). So retry
> 08:18:51   # for 30s until one signal takes.
> 08:18:51   while impalad.get_num_in_flight_queries() == 1:
> 08:18:51 time.sleep(1)
> 08:18:51 LOG.info("Sending signal...")
> 08:18:51 os.kill(p.pid(), signal.SIGINT)
> 08:18:51 num_tries += 1
> 08:18:51 >   assert num_tries < 30, "SIGINT was not caught by shell 
> within 30s"
> 08:18:51 E   AssertionError: SIGINT was not caught by shell within 30s
> 08:18:51 E   assert 30 < 30
> 08:18:51 
> 08:18:51 custom_cluster/test_client_ssl.py:85: AssertionError
> 08:18:51  Captured stdout setup 
> -
> 08:18:51 Starting State Store logging to 
> /data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/logs/custom_cluster_tests/statestored.INFO
> 08:18:51 Starting Catalog Service logging to 
> /data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/logs/custom_cluster_tests/catalogd.INFO
> 08:18:51 Starting Impala Daemon logging to 
> /data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/logs/custom_cluster_tests/impalad.INFO
> 08:18:51 Starting Impala Daemon logging to 
> /data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/logs/custom_cluster_tests/impalad_node1.INFO
> 08:18:51 Starting Impala Daemon logging to 
> /data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/logs/custom_cluster_tests/impalad_node2.INFO
> 08:18:51 Waiting for Catalog... Status: 53 DBs / 1077 tables (ready=True)
> 08:18:51 Waiting for Catalog... Status: 53 DBs / 1077 tables (ready=True)
> 08:18:51 Waiting for Catalog... Status: 53 DBs / 1077 tables (ready=True)
> 08:18:51 Impala Cluster Running with 3 nodes.
> 08:18:51  Captured stderr setup 
> -
> 08:18:51 

[jira] [Resolved] (IMPALA-10312) test_shell_interactive.TestImpalaShellInteractive AssertionError: alter query should be closed

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10312.

Fix Version/s: Impala 4.0
   Resolution: Fixed

> test_shell_interactive.TestImpalaShellInteractive AssertionError: alter query 
> should be closed
> --
>
> Key: IMPALA-10312
> URL: https://issues.apache.org/jira/browse/IMPALA-10312
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Kurt Deschler
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky
> Fix For: Impala 4.0
>
>
> shell.test_shell_interactive.TestImpalaShellInteractive.test_ddl_queries_are_closed[table_format_and_file_extension:
>  ('textfile', '.txt') | protocol: beeswax] (from pytest)
> Failing for the past 1 build (Since 
> [!https://master-02.jenkins.cloudera.com/static/4975bb2b/images/16x16/red.png!
>  
> #196|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/]
>  )
> [Took 22 
> sec.|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/testReport/shell.test_shell_interactive/TestImpalaShellInteractive/test_ddl_queries_are_closed_table_format_and_file_extensiontextfile_txt_protocol__beeswax_/history]
>  
> [!https://master-02.jenkins.cloudera.com/static/4975bb2b/images/16x16/notepad.png!
>  add 
> description|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/testReport/shell.test_shell_interactive/TestImpalaShellInteractive/test_ddl_queries_are_closed_table_format_and_file_extensiontextfile_txt_protocol__beeswax_/editDescription]
> h3. Error Message
> AssertionError: alter query should be closed assert  ImpaladService.wait_for_num_in_flight_queries of 
> >(0) + 
> where  > = 
>  0x7f49541b2150>.wait_for_num_in_flight_queries
> h3. Stacktrace
> /data/jenkins/workspace/impala-asf-master-core-s3-data-cache/repos/Impala/tests/shell/test_shell_interactive.py:467:
>  in test_ddl_queries_are_closed assert 
> impalad.wait_for_num_in_flight_queries(0), MSG % 'alter' E AssertionError: 
> alter query should be closed E assert  ImpaladService.wait_for_num_in_flight_queries of 
> >(0) E + 
> where  > = 
>  0x7f49541b2150>.wait_for_num_in_flight_queries



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10332) Add file formats to HdfsScanNode's thrift representation and codegen for those

2020-11-24 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10332?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10332:
---
Fix Version/s: Impala 4.0

> Add file formats to HdfsScanNode's thrift representation and codegen for those
> --
>
> Key: IMPALA-10332
> URL: https://issues.apache.org/jira/browse/IMPALA-10332
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Backend, Frontend
>Reporter: Daniel Becker
>Assignee: Daniel Becker
>Priority: Major
> Fix For: Impala 4.0
>
>
> List all file formats that a HdfsScanNode needs to process in any fragment 
> instance. It is possible that some file formats will not be needed in all 
> fragment instances.
> This is a step towards sharing codegen between different impala backends. 
> Using the file formats provided in the thrift file, a backend can codegen 
> code for file formats that are not needed in its own process but are needed 
> in other fragment instances running on other backends, and the resulting 
> binary can be shared between multiple backends.
> Codegenning for file formats will be done based on the thrift message and not 
> on what is needed for the actual backend. This leads to some extra work in 
> case a file format is not needed for the current backend and codegen sharing 
> is not available (at this point it is not implemented). However, the overall 
> number of such cases is low.
> Also adding the file formats to the node's explain string.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10216) BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10216?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237757#comment-17237757
 ] 

Tim Armstrong commented on IMPALA-10216:


[~fangyurao] can you also include the LastTest/LastTestFailed log files?

> BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds
> --
>
> Key: IMPALA-10216
> URL: https://issues.apache.org/jira/browse/IMPALA-10216
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: flaky
> Attachments: buffer-pool-test.ERROR, buffer-pool-test.INFO, 
> buffer-pool-test.WARNING, buffer-pool-test.xml
>
>
> Only seen this once so far:
> {code}
> BufferPoolTest.WriteErrorBlacklistCompression
> Error Message
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> Stacktrace
> Impala/be/src/runtime/bufferpool/buffer-pool-test.cc:1764
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-9566) Sentry service should not be started after IMPALA-8870

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9566?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-9566.
---
Fix Version/s: Impala 4.0
   Resolution: Fixed

> Sentry service should not be started after IMPALA-8870
> --
>
> Key: IMPALA-9566
> URL: https://issues.apache.org/jira/browse/IMPALA-9566
> Project: IMPALA
>  Issue Type: Bug
>  Components: Frontend, Infrastructure
>Reporter: Fang-Yu Rao
>Assignee: Fang-Yu Rao
>Priority: Major
>  Labels: broken-build
> Fix For: Impala 4.0
>
>
> Due to the incompatibility between different versions of the Guava libraries. 
> After bumping up Guava in IMPALA-8870, our build script is not supposed to 
> start up the Sentry service because Sentry does not have its Guava bumped up 
> yet.
> To fix this, when starting the underlying services, we should take into 
> consideration the case when {{$\{DEFAULT_FS}}} is neither 
> "{{hdfs://${INTERNAL_LISTEN_HOST}:20500}}" nor "{{${LOCAL_FS}}}". 
> Specifically, we should not attempt to start the Sentry server in the 
> following case.
> [https://github.com/apache/impala/blob/master/testdata/bin/run-all.sh#L102-L104]
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9550) TestResultSpoolingFetchSize.test_fetch is flaky

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9550?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9550:
-

Assignee: Riza Suminto

> TestResultSpoolingFetchSize.test_fetch is flaky
> ---
>
> Key: IMPALA-9550
> URL: https://issues.apache.org/jira/browse/IMPALA-9550
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Riza Suminto
>Priority: Major
>  Labels: broken-build, flaky
>
> Looks like the timeout needs to be bumped up.
> {code}
> query_test.test_result_spooling.TestResultSpoolingFetchSize.test_fetch[fetch_size:
>  1 | protocol: beeswax | exec_option: {'batch_size': 2048, 'num_nodes': 
> 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> parquet/none | wait_for_finished: True] (from pytest)
> Error Message
> query_test/test_result_spooling.py:292: in test_fetch 
> self.wait_for_state(handle, self.client.QUERY_STATES['FINISHED'], timeout) 
> common/impala_test_suite.py:1085: in wait_for_state 
> self.wait_for_any_state(handle, [expected_state], timeout, client) 
> common/impala_test_suite.py:1102: in wait_for_any_state actual_state)) E  
>  Timeout: query 424f02e1ff0912bc:37a5e6cb did not reach one of the 
> expected states [4], last known state 2
> Stacktrace
> query_test/test_result_spooling.py:292: in test_fetch
> self.wait_for_state(handle, self.client.QUERY_STATES['FINISHED'], timeout)
> common/impala_test_suite.py:1085: in wait_for_state
> self.wait_for_any_state(handle, [expected_state], timeout, client)
> common/impala_test_suite.py:1102: in wait_for_any_state
> actual_state))
> E   Timeout: query 424f02e1ff0912bc:37a5e6cb did not reach one of the 
> expected states [4], last known state 2
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-10351) Enable mt_dop for DML

2020-11-23 Thread Tim Armstrong (Jira)
Tim Armstrong created IMPALA-10351:
--

 Summary: Enable mt_dop for DML
 Key: IMPALA-10351
 URL: https://issues.apache.org/jira/browse/IMPALA-10351
 Project: IMPALA
  Issue Type: Improvement
  Components: Frontend
Reporter: Tim Armstrong
Assignee: Tim Armstrong


We don't have any pending issues for MT_DOP with DML and have sufficient 
functional testing, so I think we can go ahead and enable it. IMPALA-8125 and 
query hints give us a safety valve in case we overparallelize.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9588) test_cancel_insert failed with ImpalaBeeswaxException

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237683#comment-17237683
 ] 

Tim Armstrong commented on IMPALA-9588:
---

[~gaborkaszab] could you maybe take a quick look at this. I don't think we have 
enough logs to root cause this, but the logging in the test seems inadequate. 
AFAICT the thread running __fetch_results hit an error, but that error is not 
logged anywhere by the tests.

https://github.com/apache/impala/blob/86e611bdb2a912e283ce1db01bf2940ee1583ff3/tests/util/cancel_util.py#L24

I think we probably just need more logging to understand what happened.

> test_cancel_insert failed with ImpalaBeeswaxException
> -
>
> Key: IMPALA-9588
> URL: https://issues.apache.org/jira/browse/IMPALA-9588
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Yongzhi Chen
>Assignee: Gabor Kaszab
>Priority: Major
>  Labels: broken-build, flaky
>
> test_cancellation.TestCancellationSerial.test_cancel_insert failed in 
> impala-asf-master-core-s3 build:
> query_test.test_cancellation.TestCancellationSerial.test_cancel_insert[protocol:
>  beeswax | table_format: parquet/none | exec_option: {'batch_size': 0, 
> 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': 
> False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | 
> query_type: CTAS | mt_dop: 0 | wait_action: None | cancel_delay: 1 | 
> cpu_limit_s: 0 | query: select * from lineitem order by l_orderkey | 
> fail_rpc_action: None | join_before_close: False | buffer_pool_limit: 0]
> Error Message
> ImpalaBeeswaxException: ImpalaBeeswaxException:  INNER EXCEPTION:  'thrift.Thrift.TApplicationException'>  MESSAGE: TException - service has 
> thrown: BeeswaxException(message=Invalid query handle: 
> cc4c5258e88c790b:8db6f87d, log_context=, handle=QueryHandle(id=, 
> log_context=), errorCode=0, SQLState=HY000)
> Stacktrace
> query_test/test_cancellation.py:248: in test_cancel_insert
> self.execute_cancel_test(vector)
> query_test/test_cancellation.py:167: in execute_cancel_test
> vector.get_value('cancel_delay'), vector.get_value('join_before_close'))
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/util/cancel_util.py:41:
>  in cancel_query_and_validate_state
> assert client.get_state(handle) != client.QUERY_STATES['EXCEPTION']
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/common/impala_connection.py:219:
>  in get_state
> return self.__beeswax_client.get_state(operation_handle.get_handle())
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/beeswax/impala_beeswax.py:433:
>  in get_state
> return self.__do_rpc(lambda: self.imp_service.get_state(query_handle))
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/beeswax/impala_beeswax.py:525:
>  in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(t), t)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: TException - service has thrown: 
> BeeswaxException(message=Invalid query handle: 
> cc4c5258e88c790b:8db6f87d, log_context=, handle=QueryHandle(id=, 
> log_context=), errorCode=0, SQLState=HY000)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8125) Limit number of files generated by insert

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-8125.
---
Fix Version/s: Impala 4.0
   Resolution: Fixed

> Limit number of files generated by insert
> -
>
> Key: IMPALA-8125
> URL: https://issues.apache.org/jira/browse/IMPALA-8125
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Frontend
>Reporter: Tim Armstrong
>Assignee: Bikramjeet Vig
>Priority: Major
>  Labels: multithreading
> Fix For: Impala 4.0
>
>
> One pitfall of multithreaded execution is that, if implemented naively, the 
> number of files generated by an unpartitioned insert will be multiplied by 
> mt_dop.
> We should provide a mechanism to limit the number of files generated, e.g. 
> limit the number of insert fragment instances (note that there a pre-existing 
> problem with unpartitioned inserts generating too many files).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9588) test_cancel_insert failed with ImpalaBeeswaxException

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9588:
-

Assignee: Gabor Kaszab

> test_cancel_insert failed with ImpalaBeeswaxException
> -
>
> Key: IMPALA-9588
> URL: https://issues.apache.org/jira/browse/IMPALA-9588
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Yongzhi Chen
>Assignee: Gabor Kaszab
>Priority: Major
>  Labels: broken-build, flaky
>
> test_cancellation.TestCancellationSerial.test_cancel_insert failed in 
> impala-asf-master-core-s3 build:
> query_test.test_cancellation.TestCancellationSerial.test_cancel_insert[protocol:
>  beeswax | table_format: parquet/none | exec_option: {'batch_size': 0, 
> 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': 
> False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | 
> query_type: CTAS | mt_dop: 0 | wait_action: None | cancel_delay: 1 | 
> cpu_limit_s: 0 | query: select * from lineitem order by l_orderkey | 
> fail_rpc_action: None | join_before_close: False | buffer_pool_limit: 0]
> Error Message
> ImpalaBeeswaxException: ImpalaBeeswaxException:  INNER EXCEPTION:  'thrift.Thrift.TApplicationException'>  MESSAGE: TException - service has 
> thrown: BeeswaxException(message=Invalid query handle: 
> cc4c5258e88c790b:8db6f87d, log_context=, handle=QueryHandle(id=, 
> log_context=), errorCode=0, SQLState=HY000)
> Stacktrace
> query_test/test_cancellation.py:248: in test_cancel_insert
> self.execute_cancel_test(vector)
> query_test/test_cancellation.py:167: in execute_cancel_test
> vector.get_value('cancel_delay'), vector.get_value('join_before_close'))
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/util/cancel_util.py:41:
>  in cancel_query_and_validate_state
> assert client.get_state(handle) != client.QUERY_STATES['EXCEPTION']
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/common/impala_connection.py:219:
>  in get_state
> return self.__beeswax_client.get_state(operation_handle.get_handle())
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/beeswax/impala_beeswax.py:433:
>  in get_state
> return self.__do_rpc(lambda: self.imp_service.get_state(query_handle))
> /data/jenkins/workspace/impala-asf-master-core-s3/repos/Impala/tests/beeswax/impala_beeswax.py:525:
>  in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(t), t)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: TException - service has thrown: 
> BeeswaxException(message=Invalid query handle: 
> cc4c5258e88c790b:8db6f87d, log_context=, handle=QueryHandle(id=, 
> log_context=), errorCode=0, SQLState=HY000)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-9465) TotalReadThroughput is sporadically not set in profile

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-9465:
--
Labels: ramp-up  (was: flaky ramp-up)

> TotalReadThroughput is sporadically not set in profile
> --
>
> Key: IMPALA-9465
> URL: https://issues.apache.org/jira/browse/IMPALA-9465
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Csaba Ringhofer
>Priority: Minor
>  Labels: ramp-up
>
> I saw "TotalReadThroughput: 0.00 /sec" several times in profiles locally 
> while BytesRead is not 0.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-8452) Avro scanner seems broken

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-8452?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-8452.
---
Resolution: Cannot Reproduce

I can no longer repro the issue with test_tpch_scan_ranges

> Avro scanner seems broken
> -
>
> Key: IMPALA-8452
> URL: https://issues.apache.org/jira/browse/IMPALA-8452
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 3.3.0
>Reporter: Michael Ho
>Priority: Major
>  Labels: broken-build, wrongresults
>
> A few scanner tests started failing recently on Centos6. Coincidentally, both 
> of them only started happening after [this 
> commit|https://github.com/apache/impala/commit/b5805de3e65fd1c7154e4169b323bb38ddc54f4f].
>  [~attilaj], can you please take a look and reassign if you think that commit 
> is unrelated ? 
> Oddly enough, this has shown up on Centos6. Other exhaustive runs with 
> Centos7 seem to work fine. May be it's related to some platform's library ?
> In the first case, a select count star from an avro table hangs for 2 hours:
> {noformat}
> query_test/test_scanners_fuzz.py:83: in test_fuzz_alltypes
> self.run_fuzz_test(vector, src_db, table_name, unique_database, 
> table_name)
> query_test/test_scanners_fuzz.py:201: in run_fuzz_test
> result = self.execute_query(query, query_options = query_options)
> common/impala_test_suite.py:619: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:650: in execute_query
> return self.__execute_query(self.client, query, query_options)
> common/impala_test_suite.py:721: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:180: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:183: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:360: in __execute_query
> self.wait_for_finished(handle)
> beeswax/impala_beeswax.py:384: in wait_for_finished
> time.sleep(0.05)
> E   Failed: Timeout >7200s
> SET 
> client_identifier=query_test/test_scanners_fuzz.py::TestScannersFuzzing::()::test_fuzz_alltypes[protocol:beeswax|exec_option:{'debug_action':None;'abort_on_error':False;'mem_limit':'512m';'num_nodes':0}|table_format:avro/none];
> SET batch_size=1;
> SET num_nodes=0;
> SET disable_codegen_rows_threshold=0;
> SET disable_codegen=True;
> SET abort_on_error=False;
> SET mem_limit=512m;
> -- executing against localhost:21000
> select count(*) from test_fuzz_alltypes_2cdcb963.alltypes q;
> -- 2019-04-24 04:14:31,857 INFO MainThread: Started query 
> 2049069f9f5e3aa8:f2fd47ff
> {noformat}
> The second case has to do with incorrect number of rows in a select count 
> star from tpch_avro.lineitem:
> {noformat}
> query_test/test_scanners.py:947: in test_tpch_scan_ranges
> self.run_test_case('tpch-scan-range-lengths', vector)
> common/impala_test_suite.py:517: in run_test_case
> self.__verify_results_and_errors(vector, test_section, result, use_db)
> common/impala_test_suite.py:370: in __verify_results_and_errors
> replace_filenames_with_placeholder)
> common/test_result_verifier.py:449: in verify_raw_results
> VERIFIER_MAP[verifier](expected, actual)
> common/test_result_verifier.py:271: in verify_query_result_is_equal
> assert expected_results == actual_results
> E   assert Comparing QueryTestResults (expected vs actual):
> E 6001215 != 6000679
> -- 2019-04-24 03:43:42,805 INFO MainThread: max_scan_range_length=8412307
> SET 
> client_identifier=query_test/test_scanners.py::TestTpchScanRangeLengths::()::test_tpch_scan_ranges[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'exec_single_node_rows_threshold':0}|;
> -- executing against localhost:21000
> use tpch_avro;
> -- 2019-04-24 03:43:42,814 INFO MainThread: Started query 
> c04e1968443b52fc:5b99b1b3
> SET 
> client_identifier=query_test/test_scanners.py::TestTpchScanRangeLengths::()::test_tpch_scan_ranges[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'exec_single_node_rows_threshold':0}|;
> SET batch_size=0;
> SET num_nodes=0;
> SET disable_codegen_rows_threshold=0;
> SET disable_codegen=False;
> SET abort_on_error=1;
> SET max_scan_range_length=8412307;
> SET exec_single_node_rows_threshold=0;
> -- executing against localhost:21000
> select count(*)
> from lineitem;
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, 

[jira] [Commented] (IMPALA-10066) test_cancellation_mid_command fails in parallel-all-tests-nightly

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10066?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237671#comment-17237671
 ] 

Tim Armstrong commented on IMPALA-10066:


Seems like the test failed in precommit, which is not a good sign.

> test_cancellation_mid_command fails in parallel-all-tests-nightly
> -
>
> Key: IMPALA-10066
> URL: https://issues.apache.org/jira/browse/IMPALA-10066
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Zoltán Borók-Nagy
>Assignee: Gabor Kaszab
>Priority: Major
>  Labels: broken-build
>
> Failed jobs:
> [https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/11609/]
> [https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/2927/testReport/]
> Stack trace:
> {noformat}
> shell/test_shell_interactive.py:326: in test_cancellation_mid_command
>  child_proc.wait()
> ../infra/python/env-gcc7.5.0/lib/python2.7/site-packages/pexpect/__init__.py:1205:
>  in wait
>  raise ExceptionPexpect('Cannot wait for dead child process.')
> E ExceptionPexpect: Cannot wait for dead child process.{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-10066) test_cancellation_mid_command fails in parallel-all-tests-nightly

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-10066:
--

Assignee: Gabor Kaszab  (was: Adam Tamas)

> test_cancellation_mid_command fails in parallel-all-tests-nightly
> -
>
> Key: IMPALA-10066
> URL: https://issues.apache.org/jira/browse/IMPALA-10066
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Zoltán Borók-Nagy
>Assignee: Gabor Kaszab
>Priority: Major
>  Labels: broken-build
>
> Failed jobs:
> [https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/11609/]
> [https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/2927/testReport/]
> Stack trace:
> {noformat}
> shell/test_shell_interactive.py:326: in test_cancellation_mid_command
>  child_proc.wait()
> ../infra/python/env-gcc7.5.0/lib/python2.7/site-packages/pexpect/__init__.py:1205:
>  in wait
>  raise ExceptionPexpect('Cannot wait for dead child process.')
> E ExceptionPexpect: Cannot wait for dead child process.{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work stopped] (IMPALA-10066) test_cancellation_mid_command fails in parallel-all-tests-nightly

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-10066 stopped by Tim Armstrong.
--
> test_cancellation_mid_command fails in parallel-all-tests-nightly
> -
>
> Key: IMPALA-10066
> URL: https://issues.apache.org/jira/browse/IMPALA-10066
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Zoltán Borók-Nagy
>Assignee: Adam Tamas
>Priority: Major
>  Labels: broken-build
>
> Failed jobs:
> [https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/11609/]
> [https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/2927/testReport/]
> Stack trace:
> {noformat}
> shell/test_shell_interactive.py:326: in test_cancellation_mid_command
>  child_proc.wait()
> ../infra/python/env-gcc7.5.0/lib/python2.7/site-packages/pexpect/__init__.py:1205:
>  in wait
>  raise ExceptionPexpect('Cannot wait for dead child process.')
> E ExceptionPexpect: Cannot wait for dead child process.{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10350) Impala loses double precision because of DECIMAL->DOUBLE cast

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237668#comment-17237668
 ] 

Tim Armstrong commented on IMPALA-10350:


You need a specialized algorithm (like used in something like strtof) to do 
this accurately. A dumb but correct solution would be to convert the decimal to 
a string, then parse the string. I found a couple of high performance examples 
of the core algorithm with Apache licenses.

https://github.com/lemire/fast_double_parser/blob/master/include/fast_double_parser.h#L884
https://github.com/google/wuffs/blob/7ea23d56e3fe9e4adff95ebf11bc18b9cb06e0d5/internal/cgen/base/floatconv-submodule-code.c#L1262

Note that both of these implement float parsing, but do it by first converting 
to a representation similar to Impala's DecimalValue - i.e. an integer plus a 
scale, then converting to double.

> Impala loses double precision because of DECIMAL->DOUBLE cast
> -
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness, ramp-up
> Attachments: test.c
>
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision because of DECIMAL->DOUBLE cast

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Summary: Impala loses double precision because of DECIMAL->DOUBLE cast  
(was: Impala loses double precision on the write side)

> Impala loses double precision because of DECIMAL->DOUBLE cast
> -
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness, ramp-up
> Attachments: test.c
>
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Labels: correctness ramp-up  (was: correctness)

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness, ramp-up
> Attachments: test.c
>
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237560#comment-17237560
 ] 

Tim Armstrong commented on IMPALA-10350:


I think the problem is this function which is used in the DECIMAL -> DOUBLE 
cast: 
https://github.com/apache/impala/blob/45c105d/be/src/runtime/decimal-value.inline.h#L725

Here's a small test program indicating that it seems to be where the truncation 
happens.
 [^test.c]  
{noformat}
$ gcc test.c -o test -lm && ./test
-0.43149576573887316 -0.4314957657390
-0.43149576573887310
{noformat}



> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
> Attachments: test.c
>
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Attachment: test.c

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
> Attachments: test.c
>
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Attachment: test.c

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Attachment: (was: test.c)

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237555#comment-17237555
 ] 

Tim Armstrong commented on IMPALA-10350:


The above was slightly misleading since it returned a DECIMAL value. If you 
cast to double, you get the same value as you get from reading from the data 
file.
{noformat}
[localhost:21050] default> select cast(-0.43149576573887316 as double);
Query: select cast(-0.43149576573887316 as double)
Query submitted at: 2020-11-23 09:49:32 (Coordinator: 
http://tarmstrong-box:25000)
Query progress can be monitored at: 
http://tarmstrong-box:25000/query_plan?query_id=6c414c04b47d0897:10bb3cc5
+--+
| cast(-0.43149576573887316 as double) |
+--+
| -0.431495765739  |
+--+
{noformat}

The expressions are

{noformat}
[0] = TExprNode {
  01: node_type (i32) = 14,
  02: type (struct) = TColumnType {
01: types (list) = list[1] {
  [0] = TTypeNode {
01: type (i32) = 0,
02: scalar_type (struct) = TScalarType {
  01: type (i32) = 8,
},
  },
},
  },
  03: num_children (i32) = 1,
  04: is_constant (bool) = false,
  05: fn (struct) = TFunction {
01: name (struct) = TFunctionName {
  01: db_name (string) = "_impala_builtins",
  02: function_name (string) = "casttodouble",
},
02: binary_type (i32) = 0,
03: arg_types (list) = list[1] {
  [0] = TColumnType {
01: types (list) = list[1] {
  [0] = TTypeNode {
01: type (i32) = 0,
02: scalar_type (struct) = TScalarType {
  01: type (i32) = 14,
  03: precision (i32) = -1,
  04: scale (i32) = -1,
},
  },
},
  },
},

...
[0] = TExpr {
  01: nodes (list) = list[1] {
[0] = TExprNode {
  01: node_type (i32) = 5,
  02: type (struct) = TColumnType {
01: types (list) = list[1] {
  [0] = TTypeNode {
01: type (i32) = 0,
02: scalar_type (struct) = TScalarType {
  01: type (i32) = 14,
  03: precision (i32) = 17,
  04: scale (i32) = 17,
},
  },
},
  },
  03: num_children (i32) = 0,
  04: is_constant (bool) = true,
  18: decimal_literal (struct) = TDecimalLiteral {
01: value (string) = 
"\xfff\xb3\xb0P\x1a\xdc\xac",
  },
},
  },
{noformat}



> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: 

[jira] [Commented] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17237541#comment-17237541
 ] 

Tim Armstrong commented on IMPALA-10350:


There does seem to be something going on here beyond just usual floating point 
rounding stuff, since if we just do a select a different value comes back.
{noformat}
[localhost:21050] default> select -0.43149576573887316;
Query: select -0.43149576573887316
Query submitted at: 2020-11-23 09:35:33 (Coordinator: 
http://tarmstrong-box:25000)
Query progress can be monitored at: 
http://tarmstrong-box:25000/query_plan?query_id=044e2e614cfd3791:dc64f9cd
+--+
| -0.43149576573887316 |
+--+
| -0.43149576573887316 |
+--+
{noformat}

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Component/s: Backend

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10350) Impala loses double precision on the write side

2020-11-23 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10350?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10350:
---
Labels: correctness  (was: )

> Impala loses double precision on the write side
> ---
>
> Key: IMPALA-10350
> URL: https://issues.apache.org/jira/browse/IMPALA-10350
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: correctness
>
> Impala might loses presision of double values. Reproduction: 
> {noformat}
> create table double_tbl (d double) stored as textfile;
> insert into double_tbl values (-0.43149576573887316);
> {noformat}
>  Then inspect the data file:
> {noformat}
> $ hdfs dfs -cat 
> /test-warehouse/double_tbl/424097c644088674-c55b9101_175064830_data.0.txt
>  -0.4314957657388731{noformat}
> The same happens if we store our data in Parquet.
> Hive writes don't lose precision. If the data was written by Hive then Impala 
> can read the values correctly:
> {noformat}
> $ bin/run-jdbc-client.sh -t NOSASL -q "select * from double_tbl;"
> Using JDBC Driver Name: org.apache.hive.jdbc.HiveDriver
> Connecting to: jdbc:hive2://localhost:21050/;auth=noSasl
> Executing: select * from double_tbl
> [START]
> -0.43149576573887316
> [END]{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work stopped] (IMPALA-9884) TestAdmissionControllerStress.test_mem_limit failing occasionally

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-9884 stopped by Tim Armstrong.
-
> TestAdmissionControllerStress.test_mem_limit failing occasionally
> -
>
> Key: IMPALA-9884
> URL: https://issues.apache.org/jira/browse/IMPALA-9884
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Vihang Karajgaonkar
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky
> Attachments: impalad-executors.tar.gz, 
> impalad.impala-ec2-centos74-m5-4xlarge-ondemand-1925.vpc.cloudera.com.jenkins.log.INFO.20201017-06.23933.gz
>
>
> Recently, I saw this test failing with the exception trace below. 
> {noformat}
> custom_cluster/test_admission_controller.py:1782: in test_mem_limit
> {'request_pool': self.pool_name, 'mem_limit': query_mem_limit})
> custom_cluster/test_admission_controller.py:1638: in run_admission_test
> assert metric_deltas['dequeued'] == 0,\
> E   AssertionError: Queued queries should not run until others are made to 
> finish
> E   assert 1 == 0
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-9884) TestAdmissionControllerStress.test_mem_limit failing occasionally

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-9884 started by Tim Armstrong.
-
> TestAdmissionControllerStress.test_mem_limit failing occasionally
> -
>
> Key: IMPALA-9884
> URL: https://issues.apache.org/jira/browse/IMPALA-9884
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Vihang Karajgaonkar
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky
> Attachments: impalad-executors.tar.gz, 
> impalad.impala-ec2-centos74-m5-4xlarge-ondemand-1925.vpc.cloudera.com.jenkins.log.INFO.20201017-06.23933.gz
>
>
> Recently, I saw this test failing with the exception trace below. 
> {noformat}
> custom_cluster/test_admission_controller.py:1782: in test_mem_limit
> {'request_pool': self.pool_name, 'mem_limit': query_mem_limit})
> custom_cluster/test_admission_controller.py:1638: in run_admission_test
> assert metric_deltas['dequeued'] == 0,\
> E   AssertionError: Queued queries should not run until others are made to 
> finish
> E   assert 1 == 0
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-10125) webserver.test_web_pages.TestWebPage.test_catalog failed

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-10125:
--

Assignee: (was: Tim Armstrong)

> webserver.test_web_pages.TestWebPage.test_catalog failed
> 
>
> Key: IMPALA-10125
> URL: https://issues.apache.org/jira/browse/IMPALA-10125
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Reporter: Yongzhi Chen
>Priority: Major
>  Labels: broken-build, flaky
>
> In master-core-data-load, webserver.test_web_pages.TestWebPage.test_catalog 
> failed with
> {noformat}
> Stacktrace
> webserver/test_web_pages.py:303: in test_catalog
> self.__test_table_metrics(unique_database, "foo_part", "alter-duration")
> webserver/test_web_pages.py:352: in __test_table_metrics
> "?name=%s.%s" % (db_name, tbl_name), metric, 
> ports_to_test=self.CATALOG_TEST_PORT)
> webserver/test_web_pages.py:170: in get_and_check_status
> assert string_to_search in response.text, "URL: {0} 
> Str:'{1}'\nResp:{2}".format(
> E   AssertionError: URL: 
> http://localhost:25020/table_metrics?name=test_catalog_caf8ffd1.foo_part 
> Str:'alter-duration'
> E Resp:
> E 
> E 
> E 
> E 
> E   Apache Impala
> E 
> E 
> E 
> E  href="/www/datatables-1.10.18.min.css"/>
> E  src="/www/datatables-1.10.18.min.js">
> E  rel='stylesheet' media='screen'>
> E 
> E 
> E   @media (min-width: 1300px) {
> E #nav-options {
> E width: 1280px;
> E }
> E   }
> E 
> E   body {
> E font-size: 14px;
> E   }
> E 
> E   pre {
> E padding: 10px;
> E font-size: 12px;
> E border: 1px solid #ccc;
> E   }
> E 
> E   /* Avoid unsightly padding around code element */
> E   pre.code {
> E padding: 0;
> E   }
> E 
> E   
> E   
> E 
> E   
> E 
> E   catalogd
> E 
> E  role="navigation">
> E   
> E 
> E  href="/">/
> E 
> E  href="/catalog">/catalog
> E 
> E  href="/jmx">/jmx
> E 
> E  href="/log_level">/log_level
> E 
> E  href="/logs">/logs
> E 
> E  href="/memz">/memz
> E 
> E  href="/metrics">/metrics
> E 
> E  href="/operations">/operations
> E 
> E  href="/profile_docs">/profile_docs
> E 
> E  href="/rpcz">/rpcz
> E 
> E  href="/threadz">/threadz
> E 
> E  href="/varz">/varz
> E 
> E   
> E 
> E   
> E 
> E 
> E 
> E 
> E // For Apache Knox compatibility, all urls that are accessed by 
> javascript should have
> E // their path wrapped with this.
> E function make_url(path) {
> E   var root_link = document.getElementById('root-link');
> E   var s  = root_link.href.split("?");
> E   url = s[0] + path;
> E   if (s.length > 1) {
> E if (path.includes("?")) {
> E   url += "&"
> E } else {
> E   url += "?";
> E }
> E url += s[1];
> E   }
> E   return url;
> E }
> E 
> E 
> E 
> E Metrics for table test_catalog_caf8ffd1.foo_partare not available 
> because the table is currently modified by another operation.
> E 
> E 
> E 
> E 
> E 
> E 
> E   assert 'alter-duration' in 

[jira] [Resolved] (IMPALA-10125) webserver.test_web_pages.TestWebPage.test_catalog failed

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10125?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10125.

Resolution: Fixed

I wasn't able to reproduce. If you see this again, please reopen and attach 
more logs.

> webserver.test_web_pages.TestWebPage.test_catalog failed
> 
>
> Key: IMPALA-10125
> URL: https://issues.apache.org/jira/browse/IMPALA-10125
> Project: IMPALA
>  Issue Type: Bug
>  Components: Catalog
>Reporter: Yongzhi Chen
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky
>
> In master-core-data-load, webserver.test_web_pages.TestWebPage.test_catalog 
> failed with
> {noformat}
> Stacktrace
> webserver/test_web_pages.py:303: in test_catalog
> self.__test_table_metrics(unique_database, "foo_part", "alter-duration")
> webserver/test_web_pages.py:352: in __test_table_metrics
> "?name=%s.%s" % (db_name, tbl_name), metric, 
> ports_to_test=self.CATALOG_TEST_PORT)
> webserver/test_web_pages.py:170: in get_and_check_status
> assert string_to_search in response.text, "URL: {0} 
> Str:'{1}'\nResp:{2}".format(
> E   AssertionError: URL: 
> http://localhost:25020/table_metrics?name=test_catalog_caf8ffd1.foo_part 
> Str:'alter-duration'
> E Resp:
> E 
> E 
> E 
> E 
> E   Apache Impala
> E 
> E 
> E 
> E  href="/www/datatables-1.10.18.min.css"/>
> E  src="/www/datatables-1.10.18.min.js">
> E  rel='stylesheet' media='screen'>
> E 
> E 
> E   @media (min-width: 1300px) {
> E #nav-options {
> E width: 1280px;
> E }
> E   }
> E 
> E   body {
> E font-size: 14px;
> E   }
> E 
> E   pre {
> E padding: 10px;
> E font-size: 12px;
> E border: 1px solid #ccc;
> E   }
> E 
> E   /* Avoid unsightly padding around code element */
> E   pre.code {
> E padding: 0;
> E   }
> E 
> E   
> E   
> E 
> E   
> E 
> E   catalogd
> E 
> E  role="navigation">
> E   
> E 
> E  href="/">/
> E 
> E  href="/catalog">/catalog
> E 
> E  href="/jmx">/jmx
> E 
> E  href="/log_level">/log_level
> E 
> E  href="/logs">/logs
> E 
> E  href="/memz">/memz
> E 
> E  href="/metrics">/metrics
> E 
> E  href="/operations">/operations
> E 
> E  href="/profile_docs">/profile_docs
> E 
> E  href="/rpcz">/rpcz
> E 
> E  href="/threadz">/threadz
> E 
> E  href="/varz">/varz
> E 
> E   
> E 
> E   
> E 
> E 
> E 
> E 
> E // For Apache Knox compatibility, all urls that are accessed by 
> javascript should have
> E // their path wrapped with this.
> E function make_url(path) {
> E   var root_link = document.getElementById('root-link');
> E   var s  = root_link.href.split("?");
> E   url = s[0] + path;
> E   if (s.length > 1) {
> E if (path.includes("?")) {
> E   url += "&"
> E } else {
> E   url += "?";
> E }
> E url += s[1];
> E   }
> E   return url;
> E }
> E 
> E 
> E 
> E Metrics for table test_catalog_caf8ffd1.foo_partare not available 
> because the table is currently modified by another operation.
> E 
> E 
> E 
> E 
> E 
> E 
> E   assert 'alter-duration' in 

[jira] [Work started] (IMPALA-9050) test_scanners.TestScanRangeLengths.test_scan_ranges is flaky for kudu

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-9050 started by Tim Armstrong.
-
> test_scanners.TestScanRangeLengths.test_scan_ranges is flaky for kudu
> -
>
> Key: IMPALA-9050
> URL: https://issues.apache.org/jira/browse/IMPALA-9050
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.4.0
>Reporter: Yongzhi Chen
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: build-failure, flaky
>
> Pre-commit build failed for unrelated change:
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/1405/testReport/junit/query_test.test_scanners/TestScanRangeLengths/test_scan_ranges_max_scan_range_length__512___protocol__beeswax___exec_optionbatch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__kudu_none_/
> query_test.test_scanners.TestScanRangeLengths.test_scan_ranges[max_scan_range_length:
>  512 | protocol: beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
> 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 
> 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: 
> kudu/none] 
> Error Message
> query_test/test_scanners.py:937: in test_scan_ranges 
> self.run_test_case('QueryTest/hdfs-tiny-scan', vector) 
> common/impala_test_suite.py:621: in run_test_case result = exec_fn(query, 
> user=test_section.get('USER', '').strip() or None) 
> common/impala_test_suite.py:556: in __exec_in_impala result = 
> self.__execute_query(target_impalad_client, query, user=user) 
> common/impala_test_suite.py:893: in __execute_query return 
> impalad_client.execute(query, user=user) common/impala_connection.py:205: in 
> execute return self.__beeswax_client.execute(sql_stmt, user=user) 
> beeswax/impala_beeswax.py:205: in execute result = 
> self.fetch_results(query_string, handle) beeswax/impala_beeswax.py:451: in 
> fetch_results exec_result = self.__fetch_results(query_handle, max_rows) 
> beeswax/impala_beeswax.py:462: in __fetch_results results = 
> self.__do_rpc(lambda: self.imp_service.fetch(handle, False, fetch_rows)) 
> beeswax/impala_beeswax.py:519: in __do_rpc raise 
> ImpalaBeeswaxException(self.__build_error_message(b), b) E   
> ImpalaBeeswaxException: ImpalaBeeswaxException: EINNER EXCEPTION:  'beeswaxd.ttypes.BeeswaxException'> EMESSAGE: Unable to open Kudu table: 
> Network error: failed to read from TLS socket (remote: 172.18.0.1:7051): 
> Cannot send after transport endpoint shutdown (error 108)
> Stacktrace
> query_test/test_scanners.py:937: in test_scan_ranges
> self.run_test_case('QueryTest/hdfs-tiny-scan', vector)
> common/impala_test_suite.py:621: in run_test_case
> result = exec_fn(query, user=test_section.get('USER', '').strip() or None)
> common/impala_test_suite.py:556: in __exec_in_impala
> result = self.__execute_query(target_impalad_client, query, user=user)
> common/impala_test_suite.py:893: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:205: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:205: in execute
> result = self.fetch_results(query_string, handle)
> beeswax/impala_beeswax.py:451: in fetch_results
> exec_result = self.__fetch_results(query_handle, max_rows)
> beeswax/impala_beeswax.py:462: in __fetch_results
> results = self.__do_rpc(lambda: self.imp_service.fetch(handle, False, 
> fetch_rows))
> beeswax/impala_beeswax.py:519: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: Unable to open Kudu table: Network error: failed to read from 
> TLS socket (remote: 172.18.0.1:7051): Cannot send after transport endpoint 
> shutdown (error 108)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-10312) test_shell_interactive.TestImpalaShellInteractive AssertionError: alter query should be closed

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-10312 started by Tim Armstrong.
--
> test_shell_interactive.TestImpalaShellInteractive AssertionError: alter query 
> should be closed
> --
>
> Key: IMPALA-10312
> URL: https://issues.apache.org/jira/browse/IMPALA-10312
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Kurt Deschler
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky
>
> shell.test_shell_interactive.TestImpalaShellInteractive.test_ddl_queries_are_closed[table_format_and_file_extension:
>  ('textfile', '.txt') | protocol: beeswax] (from pytest)
> Failing for the past 1 build (Since 
> [!https://master-02.jenkins.cloudera.com/static/4975bb2b/images/16x16/red.png!
>  
> #196|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/]
>  )
> [Took 22 
> sec.|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/testReport/shell.test_shell_interactive/TestImpalaShellInteractive/test_ddl_queries_are_closed_table_format_and_file_extensiontextfile_txt_protocol__beeswax_/history]
>  
> [!https://master-02.jenkins.cloudera.com/static/4975bb2b/images/16x16/notepad.png!
>  add 
> description|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/testReport/shell.test_shell_interactive/TestImpalaShellInteractive/test_ddl_queries_are_closed_table_format_and_file_extensiontextfile_txt_protocol__beeswax_/editDescription]
> h3. Error Message
> AssertionError: alter query should be closed assert  ImpaladService.wait_for_num_in_flight_queries of 
> >(0) + 
> where  > = 
>  0x7f49541b2150>.wait_for_num_in_flight_queries
> h3. Stacktrace
> /data/jenkins/workspace/impala-asf-master-core-s3-data-cache/repos/Impala/tests/shell/test_shell_interactive.py:467:
>  in test_ddl_queries_are_closed assert 
> impalad.wait_for_num_in_flight_queries(0), MSG % 'alter' E AssertionError: 
> alter query should be closed E assert  ImpaladService.wait_for_num_in_flight_queries of 
> >(0) E + 
> where  > = 
>  0x7f49541b2150>.wait_for_num_in_flight_queries



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10312) test_shell_interactive.TestImpalaShellInteractive AssertionError: alter query should be closed

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10312?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236556#comment-17236556
 ] 

Tim Armstrong commented on IMPALA-10312:


I don't have logs for this, so I'll try to reproduce it locally and close if I 
can't.

> test_shell_interactive.TestImpalaShellInteractive AssertionError: alter query 
> should be closed
> --
>
> Key: IMPALA-10312
> URL: https://issues.apache.org/jira/browse/IMPALA-10312
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Kurt Deschler
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky
>
> shell.test_shell_interactive.TestImpalaShellInteractive.test_ddl_queries_are_closed[table_format_and_file_extension:
>  ('textfile', '.txt') | protocol: beeswax] (from pytest)
> Failing for the past 1 build (Since 
> [!https://master-02.jenkins.cloudera.com/static/4975bb2b/images/16x16/red.png!
>  
> #196|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/]
>  )
> [Took 22 
> sec.|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/testReport/shell.test_shell_interactive/TestImpalaShellInteractive/test_ddl_queries_are_closed_table_format_and_file_extensiontextfile_txt_protocol__beeswax_/history]
>  
> [!https://master-02.jenkins.cloudera.com/static/4975bb2b/images/16x16/notepad.png!
>  add 
> description|https://master-02.jenkins.cloudera.com/view/Impala/view/Evergreen-asf-master/job/impala-asf-master-core-s3-data-cache/lastCompletedBuild/testReport/shell.test_shell_interactive/TestImpalaShellInteractive/test_ddl_queries_are_closed_table_format_and_file_extensiontextfile_txt_protocol__beeswax_/editDescription]
> h3. Error Message
> AssertionError: alter query should be closed assert  ImpaladService.wait_for_num_in_flight_queries of 
> >(0) + 
> where  > = 
>  0x7f49541b2150>.wait_for_num_in_flight_queries
> h3. Stacktrace
> /data/jenkins/workspace/impala-asf-master-core-s3-data-cache/repos/Impala/tests/shell/test_shell_interactive.py:467:
>  in test_ddl_queries_are_closed assert 
> impalad.wait_for_num_in_flight_queries(0), MSG % 'alter' E AssertionError: 
> alter query should be closed E assert  ImpaladService.wait_for_num_in_flight_queries of 
> >(0) E + 
> where  > = 
>  0x7f49541b2150>.wait_for_num_in_flight_queries



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9049) TestDdlStatements.test_sync_ddl_drop is flaky

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9049:
-

Assignee: Vihang Karajgaonkar  (was: Tim Armstrong)

> TestDdlStatements.test_sync_ddl_drop is flaky
> -
>
> Key: IMPALA-9049
> URL: https://issues.apache.org/jira/browse/IMPALA-9049
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 3.4.0
>Reporter: Yongzhi Chen
>Assignee: Vihang Karajgaonkar
>Priority: Major
>  Labels: broken-build, flaky
>
> It failed precommit tests for unrelated change:
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/1405/testReport/metadata.test_ddl/TestDdlStatements/test_sync_ddl_drop_protocol__beeswax___exec_optionsync_ddl___0___batch_size___0___num_nodes___0___disable_codegen_rows_threshold___0___disable_codegen___False___abort_on_error___1___exec_single_node_rows_threshold___0table_format__text_none_unique_database0_/
> Error Message
> test setup failure
> Stacktrace
> conftest.py:319: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:790: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:798: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options, user)
> common/impala_test_suite.py:893: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:205: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:187: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:362: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:356: in execute_query_async
> handle = self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:519: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> EINNER EXCEPTION: 
> EMESSAGE: AnalysisException: Database does not exist: 
> test_sync_ddl_drop_eb3f13e1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-10156) test_unmatched_schema flaky with wrong results

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-10156 started by Tim Armstrong.
--
> test_unmatched_schema flaky with wrong results
> --
>
> Key: IMPALA-10156
> URL: https://issues.apache.org/jira/browse/IMPALA-10156
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Thomas Tauber-Marshall
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
>
> {noformat}
> Error Message
> query_test/test_scanners.py:244: in test_unmatched_schema 
> self.run_test_case('QueryTest/test-unmatched-schema', vector) 
> common/impala_test_suite.py:693: in run_test_case 
> self.__verify_results_and_errors(vector, test_section, result, use_db) 
> common/impala_test_suite.py:529: in __verify_results_and_errors 
> replace_filenames_with_placeholder) common/test_result_verifier.py:456: in 
> verify_raw_results VERIFIER_MAP[verifier](expected, actual) 
> common/test_result_verifier.py:278: in verify_query_result_is_equal 
> assert expected_results == actual_results E   assert Comparing 
> QueryTestResults (expected vs actual): E 1001,'Name1',94611,5000 == 
> 1001,'Name1',94611,5000 E 1002,'Name2',94611,5000 != 
> 1001,'Name1',94611,5000 E 1003,'Name3',94611,5000 != 
> 1002,'Name2',94611,5000 E 1004,'Name4',94611,5000 != 
> 1002,'Name2',94611,5000 E 1005,'Name5',94611,5000 != 
> 1003,'Name3',94611,5000 E 1006,'Name16',94612,15000 != 
> 1003,'Name3',94611,5000 E 1006,'Name16',94612,5000 != 
> 1004,'Name4',94611,5000 E 1006,'Name16',94616,15000 != 
> 1004,'Name4',94611,5000 E 1006,'Name16',94616,5000 != 
> 1005,'Name5',94611,5000 E 1006,'Name6',94616,15000 != 
> 1005,'Name5',94611,5000 E 1006,'Name6',94616,5000 != 
> 1006,'Name16',94612,15000 E 1106,'Name16',94612,15000 != 
> 1006,'Name16',94612,15000 E 1106,'Name16',94612,5000 != 
> 1006,'Name16',94612,5000 E 1106,'Name16',94616,15000 != 
> 1006,'Name16',94612,5000 E 1106,'Name16',94616,5000 != 
> 1006,'Name16',94616,15000 E 1106,'Name6',94612,15000 != 
> 1006,'Name16',94616,15000 E 1106,'Name6',94612,5000 != 
> 1006,'Name16',94616,5000 E 1106,'Name6',94616,15000 != 
> 1006,'Name16',94616,5000 E 1106,'Name6',94616,5000 != 
> 1006,'Name6',94616,15000 E None != 1006,'Name6',94616,15000 E None != 
> 1006,'Name6',94616,5000 E None != 1006,'Name6',94616,5000 E None != 
> 1106,'Name16',94612,15000 E None != 1106,'Name16',94612,15000 E None 
> != 1106,'Name16',94612,5000 E None != 1106,'Name16',94612,5000 E None 
> != 1106,'Name16',94616,15000 E None != 1106,'Name16',94616,15000 E 
> None != 1106,'Name16',94616,5000 E None != 1106,'Name16',94616,5000 E 
> None != 1106,'Name6',94612,15000 E None != 1106,'Name6',94612,15000 E 
> None != 1106,'Name6',94612,5000 E None != 1106,'Name6',94612,5000 E 
> None != 1106,'Name6',94616,15000 E None != 1106,'Name6',94616,15000 E 
> None != 1106,'Name6',94616,5000 E None != 1106,'Name6',94616,5000 E 
> Number of rows returned (expected vs actual): 19 != 38
> Stacktrace
> query_test/test_scanners.py:244: in test_unmatched_schema
> self.run_test_case('QueryTest/test-unmatched-schema', vector)
> common/impala_test_suite.py:693: in run_test_case
> self.__verify_results_and_errors(vector, test_section, result, use_db)
> common/impala_test_suite.py:529: in __verify_results_and_errors
> replace_filenames_with_placeholder)
> common/test_result_verifier.py:456: in verify_raw_results
> VERIFIER_MAP[verifier](expected, actual)
> common/test_result_verifier.py:278: in verify_query_result_is_equal
> assert expected_results == actual_results
> E   assert Comparing QueryTestResults (expected vs actual):
> E 1001,'Name1',94611,5000 == 1001,'Name1',94611,5000
> E 1002,'Name2',94611,5000 != 1001,'Name1',94611,5000
> E 1003,'Name3',94611,5000 != 1002,'Name2',94611,5000
> E 1004,'Name4',94611,5000 != 1002,'Name2',94611,5000
> E 1005,'Name5',94611,5000 != 1003,'Name3',94611,5000
> E 1006,'Name16',94612,15000 != 1003,'Name3',94611,5000
> E 1006,'Name16',94612,5000 != 1004,'Name4',94611,5000
> E 1006,'Name16',94616,15000 != 1004,'Name4',94611,5000
> E 1006,'Name16',94616,5000 != 1005,'Name5',94611,5000
> E 1006,'Name6',94616,15000 != 1005,'Name5',94611,5000
> E 1006,'Name6',94616,5000 != 1006,'Name16',94612,15000
> E 1106,'Name16',94612,15000 != 1006,'Name16',94612,15000
> E 1106,'Name16',94612,5000 != 1006,'Name16',94612,5000
> E 1106,'Name16',94616,15000 != 1006,'Name16',94612,5000
> E 1106,'Name16',94616,5000 != 1006,'Name16',94616,15000
> E 1106,'Name6',94612,15000 != 

[jira] [Commented] (IMPALA-10156) test_unmatched_schema flaky with wrong results

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236511#comment-17236511
 ] 

Tim Armstrong commented on IMPALA-10156:


It looks like the data is doubled up, so I guess it's more likely a data 
loading issue than a query execution issue. This test isn't following best 
practices to use unique_database, so I guess I'll try switching to that.

i don't have logs from this job so may be hard to debug further.

> test_unmatched_schema flaky with wrong results
> --
>
> Key: IMPALA-10156
> URL: https://issues.apache.org/jira/browse/IMPALA-10156
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Thomas Tauber-Marshall
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
>
> {noformat}
> Error Message
> query_test/test_scanners.py:244: in test_unmatched_schema 
> self.run_test_case('QueryTest/test-unmatched-schema', vector) 
> common/impala_test_suite.py:693: in run_test_case 
> self.__verify_results_and_errors(vector, test_section, result, use_db) 
> common/impala_test_suite.py:529: in __verify_results_and_errors 
> replace_filenames_with_placeholder) common/test_result_verifier.py:456: in 
> verify_raw_results VERIFIER_MAP[verifier](expected, actual) 
> common/test_result_verifier.py:278: in verify_query_result_is_equal 
> assert expected_results == actual_results E   assert Comparing 
> QueryTestResults (expected vs actual): E 1001,'Name1',94611,5000 == 
> 1001,'Name1',94611,5000 E 1002,'Name2',94611,5000 != 
> 1001,'Name1',94611,5000 E 1003,'Name3',94611,5000 != 
> 1002,'Name2',94611,5000 E 1004,'Name4',94611,5000 != 
> 1002,'Name2',94611,5000 E 1005,'Name5',94611,5000 != 
> 1003,'Name3',94611,5000 E 1006,'Name16',94612,15000 != 
> 1003,'Name3',94611,5000 E 1006,'Name16',94612,5000 != 
> 1004,'Name4',94611,5000 E 1006,'Name16',94616,15000 != 
> 1004,'Name4',94611,5000 E 1006,'Name16',94616,5000 != 
> 1005,'Name5',94611,5000 E 1006,'Name6',94616,15000 != 
> 1005,'Name5',94611,5000 E 1006,'Name6',94616,5000 != 
> 1006,'Name16',94612,15000 E 1106,'Name16',94612,15000 != 
> 1006,'Name16',94612,15000 E 1106,'Name16',94612,5000 != 
> 1006,'Name16',94612,5000 E 1106,'Name16',94616,15000 != 
> 1006,'Name16',94612,5000 E 1106,'Name16',94616,5000 != 
> 1006,'Name16',94616,15000 E 1106,'Name6',94612,15000 != 
> 1006,'Name16',94616,15000 E 1106,'Name6',94612,5000 != 
> 1006,'Name16',94616,5000 E 1106,'Name6',94616,15000 != 
> 1006,'Name16',94616,5000 E 1106,'Name6',94616,5000 != 
> 1006,'Name6',94616,15000 E None != 1006,'Name6',94616,15000 E None != 
> 1006,'Name6',94616,5000 E None != 1006,'Name6',94616,5000 E None != 
> 1106,'Name16',94612,15000 E None != 1106,'Name16',94612,15000 E None 
> != 1106,'Name16',94612,5000 E None != 1106,'Name16',94612,5000 E None 
> != 1106,'Name16',94616,15000 E None != 1106,'Name16',94616,15000 E 
> None != 1106,'Name16',94616,5000 E None != 1106,'Name16',94616,5000 E 
> None != 1106,'Name6',94612,15000 E None != 1106,'Name6',94612,15000 E 
> None != 1106,'Name6',94612,5000 E None != 1106,'Name6',94612,5000 E 
> None != 1106,'Name6',94616,15000 E None != 1106,'Name6',94616,15000 E 
> None != 1106,'Name6',94616,5000 E None != 1106,'Name6',94616,5000 E 
> Number of rows returned (expected vs actual): 19 != 38
> Stacktrace
> query_test/test_scanners.py:244: in test_unmatched_schema
> self.run_test_case('QueryTest/test-unmatched-schema', vector)
> common/impala_test_suite.py:693: in run_test_case
> self.__verify_results_and_errors(vector, test_section, result, use_db)
> common/impala_test_suite.py:529: in __verify_results_and_errors
> replace_filenames_with_placeholder)
> common/test_result_verifier.py:456: in verify_raw_results
> VERIFIER_MAP[verifier](expected, actual)
> common/test_result_verifier.py:278: in verify_query_result_is_equal
> assert expected_results == actual_results
> E   assert Comparing QueryTestResults (expected vs actual):
> E 1001,'Name1',94611,5000 == 1001,'Name1',94611,5000
> E 1002,'Name2',94611,5000 != 1001,'Name1',94611,5000
> E 1003,'Name3',94611,5000 != 1002,'Name2',94611,5000
> E 1004,'Name4',94611,5000 != 1002,'Name2',94611,5000
> E 1005,'Name5',94611,5000 != 1003,'Name3',94611,5000
> E 1006,'Name16',94612,15000 != 1003,'Name3',94611,5000
> E 1006,'Name16',94612,5000 != 1004,'Name4',94611,5000
> E 1006,'Name16',94616,15000 != 1004,'Name4',94611,5000
> E 1006,'Name16',94616,5000 != 1005,'Name5',94611,5000
> E 1006,'Name6',94616,15000 != 1005,'Name5',94611,5000
> E 1006,'Name6',94616,5000 != 

[jira] [Updated] (IMPALA-9991) TestShellClient.test_fetch_size_result_spooling is flaky

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-9991:
--
Issue Type: Bug  (was: Test)

> TestShellClient.test_fetch_size_result_spooling is flaky
> 
>
> Key: IMPALA-9991
> URL: https://issues.apache.org/jira/browse/IMPALA-9991
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
>
> shell.test_shell_client.TestShellClient.test_fetch_size_result_spooling[table_format_and_file_extension:
>  ('parquet', '.parq') | protocol: hs2] (from pytest)
> h3. Error Message
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Stacktrace
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Standard Error
> Opened TCP connection to localhost:21050



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-9991) TestShellClient.test_fetch_size_result_spooling is flaky

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-9991:
--
Priority: Critical  (was: Major)

> TestShellClient.test_fetch_size_result_spooling is flaky
> 
>
> Key: IMPALA-9991
> URL: https://issues.apache.org/jira/browse/IMPALA-9991
> Project: IMPALA
>  Issue Type: Test
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
>
> shell.test_shell_client.TestShellClient.test_fetch_size_result_spooling[table_format_and_file_extension:
>  ('parquet', '.parq') | protocol: hs2] (from pytest)
> h3. Error Message
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Stacktrace
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Standard Error
> Opened TCP connection to localhost:21050



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9991) TestShellClient.test_fetch_size_result_spooling is flaky

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9991:
-

Assignee: Riza Suminto  (was: Tim Armstrong)

> TestShellClient.test_fetch_size_result_spooling is flaky
> 
>
> Key: IMPALA-9991
> URL: https://issues.apache.org/jira/browse/IMPALA-9991
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Riza Suminto
>Priority: Critical
>  Labels: broken-build, flaky
>
> shell.test_shell_client.TestShellClient.test_fetch_size_result_spooling[table_format_and_file_extension:
>  ('parquet', '.parq') | protocol: hs2] (from pytest)
> h3. Error Message
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Stacktrace
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Standard Error
> Opened TCP connection to localhost:21050



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9991) TestShellClient.test_fetch_size_result_spooling is flaky

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236506#comment-17236506
 ] 

Tim Armstrong commented on IMPALA-9991:
---

[~rizaon] while you're looking at result spooling, can you see if this test is 
flaky? Don't spend too much time if it's not reproducible.

> TestShellClient.test_fetch_size_result_spooling is flaky
> 
>
> Key: IMPALA-9991
> URL: https://issues.apache.org/jira/browse/IMPALA-9991
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Riza Suminto
>Priority: Critical
>  Labels: broken-build, flaky
>
> shell.test_shell_client.TestShellClient.test_fetch_size_result_spooling[table_format_and_file_extension:
>  ('parquet', '.parq') | protocol: hs2] (from pytest)
> h3. Error Message
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Stacktrace
> shell/test_shell_client.py:70: in test_fetch_size_result_spooling 
> self.__fetch_rows(client.fetch(handle), num_rows / fetch_size, num_rows) 
> shell/test_shell_client.py:80: in __fetch_rows for fetch_batch in 
> fetch_batches: ../shell/impala_client.py:787: in fetch yield 
> self._transpose(col_value_converters, resp.results.columns) E AttributeError: 
> 'NoneType' object has no attribute 'columns'
> h3. Standard Error
> Opened TCP connection to localhost:21050



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10339) Apparent hang or crash in TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10339?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236505#comment-17236505
 ] 

Tim Armstrong commented on IMPALA-10339:


I looped this overnight on master and wasn't able to hit the DCHECK or a hang.

> Apparent hang or crash in 
> TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action
> ---
>
> Key: IMPALA-10339
> URL: https://issues.apache.org/jira/browse/IMPALA-10339
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky, hang
>
> Release build with this commit as the tip:
> {noformat}
> commit 9400e9b17b13f613defb6d7b9deb471256b1d95c (CDH/cdpd-master-staging)
> Author: wzhou-code 
> Date:   Thu Oct 29 22:32:03 2020 -0700
> IMPALA-10305: Sync Kudu's FIPS compliant changes
> 
> {noformat}
> {noformat}
> Regression
> query_test.test_spilling.TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action[protocol:
>  beeswax | exec_option: {'mt_dop': 0, 'default_spillable_buffer_size': '64k'} 
> | table_format: parquet/none] (from pytest)
> Failing for the past 1 build (Since Failed#100 )
> Took 1 hr 59 min.
> add description
> Error Message
> query_test/test_spilling.py:134: in test_spilling_no_debug_action 
> self.run_test_case('QueryTest/spilling-no-debug-action', vector) 
> common/impala_test_suite.py:668: in run_test_case 
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db) 
> common/impala_test_suite.py:485: in __verify_exceptions (expected_str, 
> actual_str) E   AssertionError: Unexpected exception string. Expected: 
> row_regex:.*Cannot perform hash join at node with id .*. Repartitioning did 
> not reduce the size of a spilled partition.* E   Not found in actual: Timeout 
> >7200s
> Stacktrace
> query_test/test_spilling.py:134: in test_spilling_no_debug_action
> self.run_test_case('QueryTest/spilling-no-debug-action', vector)
> common/impala_test_suite.py:668: in run_test_case
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db)
> common/impala_test_suite.py:485: in __verify_exceptions
> (expected_str, actual_str)
> E   AssertionError: Unexpected exception string. Expected: row_regex:.*Cannot 
> perform hash join at node with id .*. Repartitioning did not reduce the size 
> of a spilled partition.*
> E   Not found in actual: Timeout >7200s
> Standard Error
> SET 
> client_identifier=query_test/test_spilling.py::TestSpillingNoDebugActionDimensions::()::test_spilling_no_debug_action[protocol:beeswax|exec_option:{'mt_dop':0;'default_spillable_buffer_size':'64k'}|table_format:parquet/none];
> -- executing against localhost:21000
> use tpch_parquet;
> -- 2020-11-11 23:12:04,319 INFO MainThread: Started query 
> c740c1c66d9679a9:6a40f161
> SET 
> client_identifier=query_test/test_spilling.py::TestSpillingNoDebugActionDimensions::()::test_spilling_no_debug_action[protocol:beeswax|exec_option:{'mt_dop':0;'default_spillable_buffer_size':'64k'}|table_format:parquet/none];
> SET mt_dop=0;
> SET default_spillable_buffer_size=64k;
> -- 2020-11-11 23:12:04,320 INFO MainThread: Loading query test file: 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/testdata/workloads/functional-query/queries/QueryTest/spilling-no-debug-action.test
> -- 2020-11-11 23:12:04,323 INFO MainThread: Starting new HTTP connection 
> (1): localhost
> -- executing against localhost:21000
> set debug_action="-1:OPEN:SET_DENY_RESERVATION_PROBABILITY@1.0";
> -- 2020-11-11 23:12:04,377 INFO MainThread: Started query 
> c044afcf5ae44df9:a2e2e7c6
> -- executing against localhost:21000
> select straight_join count(*)
> from
> lineitem a, lineitem b
> where
> a.l_partkey = 1 and
> a.l_orderkey = b.l_orderkey;
> -- 2020-11-11 23:12:04,385 INFO MainThread: Started query 
> 314c019cd252f322:2411bc76
> -- executing against localhost:21000
> SET DEBUG_ACTION="";
> -- 2020-11-11 23:12:05,199 INFO MainThread: Started query 
> 80424e68922c30f9:b2144dff
> -- executing against localhost:21000
> set debug_action="-1:OPEN:SET_DENY_RESERVATION_PROBABILITY@1.0";
> -- 2020-11-11 23:12:05,207 INFO MainThread: Started query 
> 2a4c1f4b26ea52da:4339f3ff
> -- executing against localhost:21000
> select straight_join count(*)
> from
> lineitem a
> where
> a.l_partkey not in (select l_partkey from lineitem where l_partkey > 10)
> and a.l_partkey < 1000;
> -- 2020-11-11 23:12:05,215 INFO MainThread: Started query 
> f845afd00a569446:79c5054a
> -- executing against localhost:21000
> SET DEBUG_ACTION="";
> -- 2020-11-11 23:12:07,507 

[jira] [Commented] (IMPALA-9884) TestAdmissionControllerStress.test_mem_limit failing occasionally

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236504#comment-17236504
 ] 

Tim Armstrong commented on IMPALA-9884:
---

FWIW this seems to mainly be a problem on release builds.

> TestAdmissionControllerStress.test_mem_limit failing occasionally
> -
>
> Key: IMPALA-9884
> URL: https://issues.apache.org/jira/browse/IMPALA-9884
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Vihang Karajgaonkar
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky
> Attachments: impalad-executors.tar.gz, 
> impalad.impala-ec2-centos74-m5-4xlarge-ondemand-1925.vpc.cloudera.com.jenkins.log.INFO.20201017-06.23933.gz
>
>
> Recently, I saw this test failing with the exception trace below. 
> {noformat}
> custom_cluster/test_admission_controller.py:1782: in test_mem_limit
> {'request_pool': self.pool_name, 'mem_limit': query_mem_limit})
> custom_cluster/test_admission_controller.py:1638: in run_admission_test
> assert metric_deltas['dequeued'] == 0,\
> E   AssertionError: Queued queries should not run until others are made to 
> finish
> E   assert 1 == 0
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10194) ASAN use-after-free in hdfs-util-test

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10194.

Resolution: Duplicate

> ASAN use-after-free in hdfs-util-test
> -
>
> Key: IMPALA-10194
> URL: https://issues.apache.org/jira/browse/IMPALA-10194
> Project: IMPALA
>  Issue Type: Bug
>Affects Versions: Impala 4.0
>Reporter: Laszlo Gaal
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky
>
> This is the ASAN report section from LastTest.log:
> {code}
> 90/124 Testing: hdfs-util-test
> 90/124 Test: hdfs-util-test
> Command: 
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/bin/run-jvm-binary.sh"
>  
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/build/debug//util/hdfs-util-test"
>  
> "-log_dir=/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/logs/be_tests"
> Directory: 
> /data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/src/util
> "hdfs-util-test" start time: Sep 25 22:18 PDT
> Output:
> --
> seed = 1601097520
> Note: Google Test filter = HdfsUtilTest.*
> [==] Running 2 tests from 1 test case.
> [--] Global test environment set-up.
> [--] 2 tests from HdfsUtilTest
> [ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
> 20/09/25 22:18:40 INFO util.JvmPauseMonitor: Starting JVM pause monitor
> [   OK ] HdfsUtilTest.CheckFilesystemsMatch (941 ms)
> [ RUN  ] HdfsUtilTest.CheckGetBaseName
> [   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
> [--] 2 tests from HdfsUtilTest (941 ms total)
> [--] Global test environment tear-down
> [==] 2 tests from 1 test case ran. (941 ms total)
> [  PASSED  ] 2 tests.
>   YOU HAVE 1 DISABLED TEST
> =
> ==17289==ERROR: AddressSanitizer: heap-use-after-free on address 
> 0x6070e440 at pc 0x01e3854b bp 0x7effcdf609e0 sp 0x7effcdf60190
> READ of size 7 at 0x6070e440 thread T4
> 
> Test time =   2.13 sec
> --
> Test Passed.
> "hdfs-util-test" end time: Sep 25 22:18 PDT
> "hdfs-util-test" time elapsed: 00:00:02
> --
> {code}
> The following two commits have been added to master since the last green ASAN 
> run:
> {code}
> IMPALA-10183: Fix hitting DCHECK when cancelling a query with result spooling
> IMPALA-10170: Data race on Webserver::UrlHandler::is_on_nav_bar_
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9865) Utility to pretty-print thrift profiles at various levels

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236454#comment-17236454
 ] 

Tim Armstrong commented on IMPALA-9865:
---

The problem we have with that is that some tools do depend on the detailed 
counters for certain features. E.g. cloudera manager aggregates CPU, memory, 
BytesRead, etc from the entire profile to report those. So we want to make sure 
that the information is available in some form.

> Utility to pretty-print thrift profiles at various levels
> -
>
> Key: IMPALA-9865
> URL: https://issues.apache.org/jira/browse/IMPALA-9865
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Backend
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
> Attachments: image-2020-11-20-09-25-08-082.png
>
>
> The prototyping work in IMPALA-9382 revealed some hard trade-offs between 
> having a full-fidelity text profile and readability.
> We want to have a text profile with less information by default so that it is 
> more readable, and rely on the thrift profile for more detailed debugging. 
> This would be easier if we provided a utility that can pretty-print a thrift 
> profile at different levels of detail.
> This JIRA is to reduce the default level of pretty-printing for aggregated 
> profiles, but provide a utility that can dump both the basic and full 
> versions. My thought is to start off with the same 4 levels as explain, but 
> maybe only implement 2 levels to start off with - basic and extended.
> The utility should be able to handle the same cases as 
> bin/parse-thrift-profile.py (profile log and single profile in a file) and 
> maybe print only a specified query from a  profile log.  We can use the 
> DeserializeFromArchiveString() method that was removed in IMPALA-9381, then 
> pretty-print the deserialised profile.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-10347) Explore approaches to optimizing queries that will likely be short-circuited by limits

2020-11-20 Thread Tim Armstrong (Jira)
Tim Armstrong created IMPALA-10347:
--

 Summary: Explore approaches to optimizing queries that will likely 
be short-circuited by limits
 Key: IMPALA-10347
 URL: https://issues.apache.org/jira/browse/IMPALA-10347
 Project: IMPALA
  Issue Type: Improvement
  Components: Distributed Exec
Reporter: Tim Armstrong


Based on discussion with [~amansinha], there are opportunities beyond 
IMPALA-10314 to optimize queries where there is a limit and the query is 
*unlikely* to scan many files.

The problem is that we do all the work to generate scan ranges and schedule 
them upfront, which adds a lot of overhead if only a small number of files 
actually need to be processed.

A couple of ideas we had:
* Parallelize and/or otherwise optimize the scan range generation
* Speculatively execute the query on a subset of files and then cancel and 
retry if we hit the limit
* Incrementally generate scan ranges and assign them to executors so that scan 
range generation and execution can be overlapped. This is the most general 
solution but also has a lot of knock-on implications for other subsystems, like 
cardinality/memory estimation, scheduling, query execution, query coordination, 
etc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9437) Cannot get Ozone file block size after opening the file

2020-11-20 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17236324#comment-17236324
 ] 

Tim Armstrong commented on IMPALA-9437:
---

[~adoroszlai][~arp] here's the relevant code: 
https://github.com/apache/impala/blob/981ef104654e187ae18b25f31df2fd324e687643/be/src/exec/hdfs-table-sink.cc#L402

Here's the logic for parquet default_block_size: 
https://github.com/apache/impala/blob/1bd27a3ea87622edc869d3ff72ce9b0881451c95/be/src/exec/parquet/hdfs-parquet-table-writer.cc#L1211

And for text: 
https://github.com/apache/impala/blob/8b8a49e617818e9bcf99b784b63587c95cebd622/be/src/exec/hdfs-text-table-writer.cc#L63

And HDFS_BLOCK_SIZE: 
https://github.com/apache/impala/blob/981ef104654e187ae18b25f31df2fd324e687643/be/src/exec/parquet/hdfs-parquet-table-writer.h#L94,
 which is 256MB, which is the default ozone block size.

The block size is used for sizing parquet files, so that row groups line up to 
block boundaries. I think it's a little weird if we get the wrong size, e.g. 
the ozone block size is 128MB and the parquet writer thinks it's 256mb - you 
get a row group split across blocks, which will make parallelism suboptimal.

I think if we replaced the reference to HDFS_BLOCK_SIZE with a call to a 
function that returned the actual configured default for the path being written 
to, that would be a generally good improvement.

It does depend a bit on whether people use non-default Ozone block sizes with 
any frequency.

> Cannot get Ozone file block size after opening the file
> ---
>
> Key: IMPALA-9437
> URL: https://issues.apache.org/jira/browse/IMPALA-9437
> Project: IMPALA
>  Issue Type: Sub-task
>Reporter: Sahil Takiar
>Priority: Major
>
> When creating a tmp file on HDFS, {{HdfsTableSink::CreateNewTmpFile}} first 
> opens the file, and then stats the file ({{hdfsGetPathInfo}}) before actually 
> writing any data to the file or closing the file. HDFS seems to allow this 
> behavior. However, Ozone, S3A, and ABFS do not. Impala does this for HDFS in 
> order to get the block size of the opened file. According to 
> {{HdfsTableSink}} it is possible for HDFS to create a block size with a 
> different one than requested by Impala. So in order to track the correct 
> block size for a file, the file needs to be stat'ed after opening it. For S3A 
> and ABFS this isn't a big deal, because they aren't block based filesystem, 
> but Ozone is. So we should investigate the impact of not having this 
> capability and consider adding it to the Ozone client.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-9121) heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch

2020-11-20 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-9121.
---
Resolution: Fixed

Closing for now since I won't take any action unless it reoccurs

> heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch
> -
>
> Key: IMPALA-9121
> URL: https://issues.apache.org/jira/browse/IMPALA-9121
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Bikramjeet Vig
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: asan, broken-build, flaky-test
>
> {noformat}
> --
> 85/113 Testing: hdfs-util-test
> 85/113 Test: hdfs-util-test
> Command: 
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/build/debug/util/hdfs-util-test"
>  "-log_dir=/data/jenkins/workspace
> /impala-asf-master-core-asan/repos/Impala/logs/be_tests"
> Directory: 
> /data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/src/util
> "hdfs-util-test" start time: Nov 01 22:21 PDT
> Output:
> --
> seed = 1572672103
> Note: Google Test filter = HdfsUtilTest.*
> [==] Running 2 tests from 1 test case.
> [--] Global test environment set-up.
> [--] 2 tests from HdfsUtilTest
> [ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
> 19/11/01 22:21:43 INFO util.JvmPauseMonitor: Starting JVM pause monitor
> [   OK ] HdfsUtilTest.CheckFilesystemsMatch (1012 ms)
> [ RUN  ] HdfsUtilTest.CheckGetBaseName
> [   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
> [--] 2 tests from HdfsUtilTest (1012 ms total)
> [--] Global test environment tear-down
> [==] 2 tests from 1 test case ran. (1013 ms total)
> [  PASSED  ] 2 tests.
>   YOU HAVE 1 DISABLED TEST
> =
> ==87116==ERROR: AddressSanitizer: heap-use-after-free on address 
> 0x60362878 at pc 0x01b67bac bp 0x7fc96cf57ae0 sp 0x7fc96cf57290
> READ of size 4 at 0x60362878 thread T4
> 
> Test time =   2.64 sec
> --
> Test Passed.
> "hdfs-util-test" end time: Nov 01 22:21 PDT
> "hdfs-util-test" time elapsed: 00:00:02
> --
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9865) Utility to pretty-print thrift profiles at various levels

2020-11-19 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235848#comment-17235848
 ] 

Tim Armstrong commented on IMPALA-9865:
---

[~fifteencai] this is great feedback. I'm working on a new profile approach 
that does something like this, except it retains all the information in a 
denser format.

I know the Cloudera Manager team has also been doing some work to handle large 
profiles better by not keeping them in memory for as long.

> Utility to pretty-print thrift profiles at various levels
> -
>
> Key: IMPALA-9865
> URL: https://issues.apache.org/jira/browse/IMPALA-9865
> Project: IMPALA
>  Issue Type: Sub-task
>  Components: Backend
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Major
> Attachments: image-2020-11-20-09-25-08-082.png
>
>
> The prototyping work in IMPALA-9382 revealed some hard trade-offs between 
> having a full-fidelity text profile and readability.
> We want to have a text profile with less information by default so that it is 
> more readable, and rely on the thrift profile for more detailed debugging. 
> This would be easier if we provided a utility that can pretty-print a thrift 
> profile at different levels of detail.
> This JIRA is to reduce the default level of pretty-printing for aggregated 
> profiles, but provide a utility that can dump both the basic and full 
> versions. My thought is to start off with the same 4 levels as explain, but 
> maybe only implement 2 levels to start off with - basic and extended.
> The utility should be able to handle the same cases as 
> bin/parse-thrift-profile.py (profile log and single profile in a file) and 
> maybe print only a specified query from a  profile log.  We can use the 
> DeserializeFromArchiveString() method that was removed in IMPALA-9381, then 
> pretty-print the deserialised profile.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Updated] (IMPALA-10344) Error loading functional_orc_def.alltypesaggmultifiles in hive

2020-11-19 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10344:
---
Summary: Error loading functional_orc_def.alltypesaggmultifiles in hive  
(was: Error loading functional_orc_def.al ltypesaggmultifiles in hive)

> Error loading functional_orc_def.alltypesaggmultifiles in hive
> --
>
> Key: IMPALA-10344
> URL: https://issues.apache.org/jira/browse/IMPALA-10344
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Infrastructure
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
> Attachments: hive-metastore.out, hive-server2.out, 
> load-functional-query-exhaustive-hive-generated-orc-def-block.sql.log
>
>
> Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask. 
> Exception when loading 11 partitions in table alltypesaggmultifiles 
> {noformat}
> 0: jdbc:hive2://localhost:11050/default> insert into table 
> functional_orc_def.al 
> ltypesaggmultifiles partition (year, month, day) SELECT id, bool_col, 
> tinyint_co 
> l, smallint_col, int_col, bigint_col, float_col, double_col, date_string_col, 
> st 
> ring_col, timestamp_col, year, month, day FROM 
> functional.alltypesaggmultifiles  
> where id % 4 = 1;
> going to print operations logs
> printed operations logs
> going to print operations logs
> INFO  : Compiling 
> command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9): 
> insert into table functional_orc_def.alltypesaggmultifiles partition (year, 
> month, day) SELECT id, bool_col, tinyint_col, smallint_col, int_col, 
> bigint_col, float_col, double_col, date_string_col, string_col, 
> timestamp_col, year, month, day FROM functional.alltypesaggmultifiles where 
> id % 4 = 1
> INFO  : No Stats for functional@alltypesaggmultifiles, Columns: float_col, 
> int_col, string_col, bool_col, date_string_col, smallint_col, timestamp_col, 
> tinyint_col, bigint_col, id, double_col
> INFO  : Semantic Analysis Completed (retrial = false)
> INFO  : Created Hive schema: Schema(fieldSchemas:[FieldSchema(name:id, 
> type:int, comment:null), FieldSchema(name:bool_col, type:boolean, 
> comment:null), FieldSchema(name:tinyint_col, type:tinyint, comment:null), 
> FieldSchema(name:smallint_col, type:smallint, comment:null), 
> FieldSchema(name:int_col, type:int, comment:null), 
> FieldSchema(name:bigint_col, type:bigint, comment:null), 
> FieldSchema(name:float_col, type:float, comment:null), 
> FieldSchema(name:double_col, type:double, comment:null), 
> FieldSchema(name:date_string_col, type:string, comment:null), 
> FieldSchema(name:string_col, type:string, comment:null), 
> FieldSchema(name:timestamp_col, type:timestamp, comment:null), 
> FieldSchema(name:year, type:int, comment:null), FieldSchema(name:month, 
> type:int, comment:null), FieldSchema(name:day, type:int, comment:null)], 
> properties:null)
> INFO  : Completed compiling 
> command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9); 
> Time taken: 0.036 seconds
> INFO  : Executing 
> command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9): 
> insert into table functional_orc_def.alltypesaggmultifiles partition (year, 
> month, day) SELECT id, bool_col, tinyint_col, smallint_col, int_col, 
> bigint_col, float_col, double_col, date_string_col, string_col, 
> timestamp_col, year, month, day FROM functional.alltypesaggmultifiles where 
> id % 4 = 1
> INFO  : Query ID = jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9
> INFO  : Total jobs = 1
> INFO  : Launching Job 1 out of 1
> INFO  : Starting task [Stage-1:MAPRED] in serial mode
> INFO  : Subscribed to counters: [] for queryId: 
> jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9
> INFO  : Session is already open
> INFO  : Dag name: insert into table functional_orc_def.all...1 (Stage-1)
> INFO  : Status: Running (Executing on YARN cluster with App id 
> application_1605304119768_0012)
> printed operations logs
> --
> VERTICES  MODESTATUS  TOTAL  COMPLETED  
> RUNNING  PENDING  FAILED  KILLED  
> --
> Map 1 .. container SUCCEEDED 42 420   
>  0   0   0  
> Reducer 2container   RUNNING  1  010  
>  0   0  
> --
> VERTICES: 01/02  [=>>-] 97%   ELAPSED 
> TIME: 1.46 

[jira] [Created] (IMPALA-10344) Error loading functional_orc_def.al ltypesaggmultifiles in hive

2020-11-19 Thread Tim Armstrong (Jira)
Tim Armstrong created IMPALA-10344:
--

 Summary: Error loading functional_orc_def.al ltypesaggmultifiles 
in hive
 Key: IMPALA-10344
 URL: https://issues.apache.org/jira/browse/IMPALA-10344
 Project: IMPALA
  Issue Type: Improvement
  Components: Infrastructure
Affects Versions: Impala 4.0
Reporter: Tim Armstrong
Assignee: Tim Armstrong
 Attachments: hive-metastore.out, hive-server2.out, 
load-functional-query-exhaustive-hive-generated-orc-def-block.sql.log

Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask. 
Exception when loading 11 partitions in table alltypesaggmultifiles 

{noformat}
0: jdbc:hive2://localhost:11050/default> insert into table 
functional_orc_def.al 
ltypesaggmultifiles partition (year, month, day) SELECT id, bool_col, 
tinyint_co 
l, smallint_col, int_col, bigint_col, float_col, double_col, date_string_col, 
st 
ring_col, timestamp_col, year, month, day FROM functional.alltypesaggmultifiles 
 
where id % 4 = 1;
going to print operations logs
printed operations logs
going to print operations logs
INFO  : Compiling 
command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9): 
insert into table functional_orc_def.alltypesaggmultifiles partition (year, 
month, day) SELECT id, bool_col, tinyint_col, smallint_col, int_col, 
bigint_col, float_col, double_col, date_string_col, string_col, timestamp_col, 
year, month, day FROM functional.alltypesaggmultifiles where id % 4 = 1
INFO  : No Stats for functional@alltypesaggmultifiles, Columns: float_col, 
int_col, string_col, bool_col, date_string_col, smallint_col, timestamp_col, 
tinyint_col, bigint_col, id, double_col
INFO  : Semantic Analysis Completed (retrial = false)
INFO  : Created Hive schema: Schema(fieldSchemas:[FieldSchema(name:id, 
type:int, comment:null), FieldSchema(name:bool_col, type:boolean, 
comment:null), FieldSchema(name:tinyint_col, type:tinyint, comment:null), 
FieldSchema(name:smallint_col, type:smallint, comment:null), 
FieldSchema(name:int_col, type:int, comment:null), FieldSchema(name:bigint_col, 
type:bigint, comment:null), FieldSchema(name:float_col, type:float, 
comment:null), FieldSchema(name:double_col, type:double, comment:null), 
FieldSchema(name:date_string_col, type:string, comment:null), 
FieldSchema(name:string_col, type:string, comment:null), 
FieldSchema(name:timestamp_col, type:timestamp, comment:null), 
FieldSchema(name:year, type:int, comment:null), FieldSchema(name:month, 
type:int, comment:null), FieldSchema(name:day, type:int, comment:null)], 
properties:null)
INFO  : Completed compiling 
command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9); 
Time taken: 0.036 seconds
INFO  : Executing 
command(queryId=jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9): 
insert into table functional_orc_def.alltypesaggmultifiles partition (year, 
month, day) SELECT id, bool_col, tinyint_col, smallint_col, int_col, 
bigint_col, float_col, double_col, date_string_col, string_col, timestamp_col, 
year, month, day FROM functional.alltypesaggmultifiles where id % 4 = 1
INFO  : Query ID = jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9
INFO  : Total jobs = 1
INFO  : Launching Job 1 out of 1
INFO  : Starting task [Stage-1:MAPRED] in serial mode
INFO  : Subscribed to counters: [] for queryId: 
jenkins_20201113135913_d2ecc422-ec51-4fe2-9e92-e8d93b01aca9
INFO  : Session is already open
INFO  : Dag name: insert into table functional_orc_def.all...1 (Stage-1)
INFO  : Status: Running (Executing on YARN cluster with App id 
application_1605304119768_0012)

printed operations logs
--
VERTICES  MODESTATUS  TOTAL  COMPLETED  RUNNING  
PENDING  FAILED  KILLED  
--
Map 1 .. container SUCCEEDED 42 420
0   0   0  
Reducer 2container   RUNNING  1  010
   0   0  
--
VERTICES: 01/02  [=>>-] 97%   ELAPSED TIME: 
1.46 s 
--
--
VERTICES  MODESTATUS  TOTAL  COMPLETED  RUNNING  
PENDING  FAILED  KILLED  
--
Map 1 .. container SUCCEEDED 42 420
0   0   0  
Reducer 2container 

[jira] [Updated] (IMPALA-10339) Apparent hang or crash in TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action

2020-11-19 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10339?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong updated IMPALA-10339:
---
Summary: Apparent hang or crash in 
TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action  (was: 
Apparent hang in 
TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action)

> Apparent hang or crash in 
> TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action
> ---
>
> Key: IMPALA-10339
> URL: https://issues.apache.org/jira/browse/IMPALA-10339
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky, hang
>
> Release build with this commit as the tip:
> {noformat}
> commit 9400e9b17b13f613defb6d7b9deb471256b1d95c (CDH/cdpd-master-staging)
> Author: wzhou-code 
> Date:   Thu Oct 29 22:32:03 2020 -0700
> IMPALA-10305: Sync Kudu's FIPS compliant changes
> 
> {noformat}
> {noformat}
> Regression
> query_test.test_spilling.TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action[protocol:
>  beeswax | exec_option: {'mt_dop': 0, 'default_spillable_buffer_size': '64k'} 
> | table_format: parquet/none] (from pytest)
> Failing for the past 1 build (Since Failed#100 )
> Took 1 hr 59 min.
> add description
> Error Message
> query_test/test_spilling.py:134: in test_spilling_no_debug_action 
> self.run_test_case('QueryTest/spilling-no-debug-action', vector) 
> common/impala_test_suite.py:668: in run_test_case 
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db) 
> common/impala_test_suite.py:485: in __verify_exceptions (expected_str, 
> actual_str) E   AssertionError: Unexpected exception string. Expected: 
> row_regex:.*Cannot perform hash join at node with id .*. Repartitioning did 
> not reduce the size of a spilled partition.* E   Not found in actual: Timeout 
> >7200s
> Stacktrace
> query_test/test_spilling.py:134: in test_spilling_no_debug_action
> self.run_test_case('QueryTest/spilling-no-debug-action', vector)
> common/impala_test_suite.py:668: in run_test_case
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db)
> common/impala_test_suite.py:485: in __verify_exceptions
> (expected_str, actual_str)
> E   AssertionError: Unexpected exception string. Expected: row_regex:.*Cannot 
> perform hash join at node with id .*. Repartitioning did not reduce the size 
> of a spilled partition.*
> E   Not found in actual: Timeout >7200s
> Standard Error
> SET 
> client_identifier=query_test/test_spilling.py::TestSpillingNoDebugActionDimensions::()::test_spilling_no_debug_action[protocol:beeswax|exec_option:{'mt_dop':0;'default_spillable_buffer_size':'64k'}|table_format:parquet/none];
> -- executing against localhost:21000
> use tpch_parquet;
> -- 2020-11-11 23:12:04,319 INFO MainThread: Started query 
> c740c1c66d9679a9:6a40f161
> SET 
> client_identifier=query_test/test_spilling.py::TestSpillingNoDebugActionDimensions::()::test_spilling_no_debug_action[protocol:beeswax|exec_option:{'mt_dop':0;'default_spillable_buffer_size':'64k'}|table_format:parquet/none];
> SET mt_dop=0;
> SET default_spillable_buffer_size=64k;
> -- 2020-11-11 23:12:04,320 INFO MainThread: Loading query test file: 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/testdata/workloads/functional-query/queries/QueryTest/spilling-no-debug-action.test
> -- 2020-11-11 23:12:04,323 INFO MainThread: Starting new HTTP connection 
> (1): localhost
> -- executing against localhost:21000
> set debug_action="-1:OPEN:SET_DENY_RESERVATION_PROBABILITY@1.0";
> -- 2020-11-11 23:12:04,377 INFO MainThread: Started query 
> c044afcf5ae44df9:a2e2e7c6
> -- executing against localhost:21000
> select straight_join count(*)
> from
> lineitem a, lineitem b
> where
> a.l_partkey = 1 and
> a.l_orderkey = b.l_orderkey;
> -- 2020-11-11 23:12:04,385 INFO MainThread: Started query 
> 314c019cd252f322:2411bc76
> -- executing against localhost:21000
> SET DEBUG_ACTION="";
> -- 2020-11-11 23:12:05,199 INFO MainThread: Started query 
> 80424e68922c30f9:b2144dff
> -- executing against localhost:21000
> set debug_action="-1:OPEN:SET_DENY_RESERVATION_PROBABILITY@1.0";
> -- 2020-11-11 23:12:05,207 INFO MainThread: Started query 
> 2a4c1f4b26ea52da:4339f3ff
> -- executing against localhost:21000
> select straight_join count(*)
> from
> lineitem a
> where
> a.l_partkey not in (select l_partkey from lineitem where l_partkey > 10)
> and a.l_partkey < 1000;
> -- 2020-11-11 23:12:05,215 INFO MainThread: Started query 
> f845afd00a569446:79c5054a
> -- 

[jira] [Commented] (IMPALA-10339) Apparent hang in TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action

2020-11-19 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10339?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235837#comment-17235837
 ] 

Tim Armstrong commented on IMPALA-10339:


We're missing the "Cancelling fragment instances as directed by the 
coordinator" log, which seems to be the main way that this would result in the 
query getting cancelled. Not sure what's going on here.

> Apparent hang in 
> TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action
> --
>
> Key: IMPALA-10339
> URL: https://issues.apache.org/jira/browse/IMPALA-10339
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky, hang
>
> Release build with this commit as the tip:
> {noformat}
> commit 9400e9b17b13f613defb6d7b9deb471256b1d95c (CDH/cdpd-master-staging)
> Author: wzhou-code 
> Date:   Thu Oct 29 22:32:03 2020 -0700
> IMPALA-10305: Sync Kudu's FIPS compliant changes
> 
> {noformat}
> {noformat}
> Regression
> query_test.test_spilling.TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action[protocol:
>  beeswax | exec_option: {'mt_dop': 0, 'default_spillable_buffer_size': '64k'} 
> | table_format: parquet/none] (from pytest)
> Failing for the past 1 build (Since Failed#100 )
> Took 1 hr 59 min.
> add description
> Error Message
> query_test/test_spilling.py:134: in test_spilling_no_debug_action 
> self.run_test_case('QueryTest/spilling-no-debug-action', vector) 
> common/impala_test_suite.py:668: in run_test_case 
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db) 
> common/impala_test_suite.py:485: in __verify_exceptions (expected_str, 
> actual_str) E   AssertionError: Unexpected exception string. Expected: 
> row_regex:.*Cannot perform hash join at node with id .*. Repartitioning did 
> not reduce the size of a spilled partition.* E   Not found in actual: Timeout 
> >7200s
> Stacktrace
> query_test/test_spilling.py:134: in test_spilling_no_debug_action
> self.run_test_case('QueryTest/spilling-no-debug-action', vector)
> common/impala_test_suite.py:668: in run_test_case
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db)
> common/impala_test_suite.py:485: in __verify_exceptions
> (expected_str, actual_str)
> E   AssertionError: Unexpected exception string. Expected: row_regex:.*Cannot 
> perform hash join at node with id .*. Repartitioning did not reduce the size 
> of a spilled partition.*
> E   Not found in actual: Timeout >7200s
> Standard Error
> SET 
> client_identifier=query_test/test_spilling.py::TestSpillingNoDebugActionDimensions::()::test_spilling_no_debug_action[protocol:beeswax|exec_option:{'mt_dop':0;'default_spillable_buffer_size':'64k'}|table_format:parquet/none];
> -- executing against localhost:21000
> use tpch_parquet;
> -- 2020-11-11 23:12:04,319 INFO MainThread: Started query 
> c740c1c66d9679a9:6a40f161
> SET 
> client_identifier=query_test/test_spilling.py::TestSpillingNoDebugActionDimensions::()::test_spilling_no_debug_action[protocol:beeswax|exec_option:{'mt_dop':0;'default_spillable_buffer_size':'64k'}|table_format:parquet/none];
> SET mt_dop=0;
> SET default_spillable_buffer_size=64k;
> -- 2020-11-11 23:12:04,320 INFO MainThread: Loading query test file: 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/testdata/workloads/functional-query/queries/QueryTest/spilling-no-debug-action.test
> -- 2020-11-11 23:12:04,323 INFO MainThread: Starting new HTTP connection 
> (1): localhost
> -- executing against localhost:21000
> set debug_action="-1:OPEN:SET_DENY_RESERVATION_PROBABILITY@1.0";
> -- 2020-11-11 23:12:04,377 INFO MainThread: Started query 
> c044afcf5ae44df9:a2e2e7c6
> -- executing against localhost:21000
> select straight_join count(*)
> from
> lineitem a, lineitem b
> where
> a.l_partkey = 1 and
> a.l_orderkey = b.l_orderkey;
> -- 2020-11-11 23:12:04,385 INFO MainThread: Started query 
> 314c019cd252f322:2411bc76
> -- executing against localhost:21000
> SET DEBUG_ACTION="";
> -- 2020-11-11 23:12:05,199 INFO MainThread: Started query 
> 80424e68922c30f9:b2144dff
> -- executing against localhost:21000
> set debug_action="-1:OPEN:SET_DENY_RESERVATION_PROBABILITY@1.0";
> -- 2020-11-11 23:12:05,207 INFO MainThread: Started query 
> 2a4c1f4b26ea52da:4339f3ff
> -- executing against localhost:21000
> select straight_join count(*)
> from
> lineitem a
> where
> a.l_partkey not in (select l_partkey from lineitem where l_partkey > 10)
> and a.l_partkey < 1000;
> -- 2020-11-11 23:12:05,215 INFO MainThread: Started query 
> 

[jira] [Commented] (IMPALA-10339) Apparent hang in TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action

2020-11-19 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10339?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235831#comment-17235831
 ] 

Tim Armstrong commented on IMPALA-10339:


{noformat}
I1117 00:32:24.677520 14067 control-service.cc:142] 
484209f4cc06bb05:eec7d58d] ExecQueryFInstances(): 
query_id=484209f4cc06bb05:eec7d58d 
coord=impala-ec2-centos74-m5-4xlarge-ondemand-0d63.vpc.cloudera.com:27000 
#instances=3
I1117 00:32:24.746866 18128 query-state.cc:897] 
484209f4cc06bb05:eec7d58d0004] Executing instance. 
instance_id=484209f4cc06bb05:eec7d58d0004 fragment_idx=3 
per_fragment_instance_idx=0 coord_state_idx=1 #in-flight=34
I1117 00:32:24.767122 18132 query-state.cc:897] 
484209f4cc06bb05:eec7d58d0007] Executing instance. 
instance_id=484209f4cc06bb05:eec7d58d0007 fragment_idx=1 
per_fragment_instance_idx=0 coord_state_idx=1 #in-flight=35
I1117 00:32:24.806087 18137 query-state.cc:897] 
484209f4cc06bb05:eec7d58d0001] Executing instance. 
instance_id=484209f4cc06bb05:eec7d58d0001 fragment_idx=2 
per_fragment_instance_idx=0 coord_state_idx=1 #in-flight=36
I1117 00:32:25.895730 18137 krpc-data-stream-mgr.cc:337] 
484209f4cc06bb05:eec7d58d0001] cancelling active streams for 
fragment_instance_id=484209f4cc06bb05:eec7d58d0001
I1117 00:32:25.895849 18137 query-state.cc:906] 
484209f4cc06bb05:eec7d58d0001] Instance completed. 
instance_id=484209f4cc06bb05:eec7d58d0001 #in-flight=34 status=OK
I1117 00:33:07.615486 18128 krpc-data-stream-mgr.cc:337] 
484209f4cc06bb05:eec7d58d0004] cancelling active streams for 
fragment_instance_id=484209f4cc06bb05:eec7d58d0004
I1117 00:33:07.615584 18128 query-state.cc:906] 
484209f4cc06bb05:eec7d58d0004] Instance completed. 
instance_id=484209f4cc06bb05:eec7d58d0004 #in-flight=10 status=OK
I1117 00:33:08.004225 18194 krpc-data-stream-mgr.cc:306] 
484209f4cc06bb05:eec7d58d0007] DeregisterRecvr(): 
fragment_instance_id=484209f4cc06bb05:eec7d58d0007, node=5
I1117 00:33:08.010303 14095 query-exec-mgr.cc:126] QueryState: 
query_id=484209f4cc06bb05:eec7d58d refcnt=3
I1117 00:33:09.792160 18132 status.cc:88] 484209f4cc06bb05:eec7d58d0007] 
Cannot perform hash join at node with id 2. Repartitioning did not reduce the 
size of a spilled partition. Repartitioning level 1. Number of rows 1071394:
I1117 00:33:09.887547 18132 krpc-data-stream-mgr.cc:337] 
484209f4cc06bb05:eec7d58d0007] cancelling active streams for 
fragment_instance_id=484209f4cc06bb05:eec7d58d0007
I1117 00:33:09.906414 18132 krpc-data-stream-mgr.cc:306] 
484209f4cc06bb05:eec7d58d0007] DeregisterRecvr(): 
fragment_instance_id=484209f4cc06bb05:eec7d58d0007, node=4
I1117 00:33:09.907351 18132 query-state.cc:906] 
484209f4cc06bb05:eec7d58d0007] Instance completed. 
instance_id=484209f4cc06bb05:eec7d58d0007 #in-flight=11 
status=PARTITIONED_HASH_JOIN_REPARTITION_FAILS: Cannot perform hash join at 
node with id 2. Repartitioning did not reduce the size of a spilled partition. 
Repartitioning level 1. Number of rows 1071394:
I1117 00:33:09.912003 18122 query-state.cc:464] 
484209f4cc06bb05:eec7d58d] UpdateBackendExecState(): last report for 
484209f4cc06bb05:eec7d58d
F1117 00:33:09.914166 18122 query-state.cc:877] 
484209f4cc06bb05:eec7d58d] Check failed: is_cancelled_.Load() == 1 (0 
vs. 1) 

{noformat}

> Apparent hang in 
> TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action
> --
>
> Key: IMPALA-10339
> URL: https://issues.apache.org/jira/browse/IMPALA-10339
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky, hang
>
> Release build with this commit as the tip:
> {noformat}
> commit 9400e9b17b13f613defb6d7b9deb471256b1d95c (CDH/cdpd-master-staging)
> Author: wzhou-code 
> Date:   Thu Oct 29 22:32:03 2020 -0700
> IMPALA-10305: Sync Kudu's FIPS compliant changes
> 
> {noformat}
> {noformat}
> Regression
> query_test.test_spilling.TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action[protocol:
>  beeswax | exec_option: {'mt_dop': 0, 'default_spillable_buffer_size': '64k'} 
> | table_format: parquet/none] (from pytest)
> Failing for the past 1 build (Since Failed#100 )
> Took 1 hr 59 min.
> add description
> Error Message
> query_test/test_spilling.py:134: in test_spilling_no_debug_action 
> self.run_test_case('QueryTest/spilling-no-debug-action', vector) 
> common/impala_test_suite.py:668: in run_test_case 
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db) 
> common/impala_test_suite.py:485: in __verify_exceptions (expected_str, 
> 

[jira] [Commented] (IMPALA-10339) Apparent hang in TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action

2020-11-19 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10339?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235825#comment-17235825
 ] 

Tim Armstrong commented on IMPALA-10339:


I saw a crash on a very similar query that fails with repartitioning:

{noformat}
  74: client_identifier (string) = 
"query_test/test_spilling.py::TestSpillingDebugActionDimensions::()::test_spilling[protocol:be
eswax|exec_option:{'mt_dop':0;'debug_action':None;'default_spillable_buffer_size':'256k'}|table_format:parquet/none]",
...
I1117 00:32:24.636593  8768 Frontend.java:1522] 
484209f4cc06bb05:eec7d58d] Analyzing query: select straight_join *
from lineitem l1 join lineitem l2 on l1.l_linenumber = l2.l_linenumber
where l1.l_orderkey < 10
order by l1.l_orderkey desc, l1.l_linenumber desc limit 10 db: tpch_parquet
{noformat}

This is this test from 
./testdata/workloads/functional-query/queries/QueryTest/spilling.test:
{noformat}

 QUERY
# Test spilling join with many duplicates in join key. We don't expect this to 
succeed
# with a memory constraint: see IMPALA-4857. Limit size of probe so that query 
doesn't
# bog down executing an exploding join.
# The additional "order by" and "limit" clauses make sure that a successful
# query does not too much data to the client.
set buffer_pool_limit=167m;
select straight_join *
from lineitem l1 join lineitem l2 on l1.l_linenumber = l2.l_linenumber
where l1.l_orderkey < 10
order by l1.l_orderkey desc, l1.l_linenumber desc limit 10
 CATCH
Repartitioning did not reduce the size of a spilled partition

{noformat}

Backtrace on crash is:
{noformat}
F1117 00:33:09.914166 18122 query-state.cc:877] 
484209f4cc06bb05:eec7d58d] Check failed: is_cancelled_.Load() == 1 (0 
vs. 1) 
*** Check failure stack trace: ***
@  0x520b32c  google::LogMessage::Fail()
@  0x520cc1c  google::LogMessage::SendToLog()
@  0x520ac8a  google::LogMessage::Flush()
@  0x520e888  google::LogMessageFatal::~LogMessageFatal()
@  0x228786f  impala::QueryState::MonitorFInstances()
@  0x2277010  impala::QueryExecMgr::ExecuteQueryHelper()
@  0x227f924  boost::_mfi::mf1<>::operator()()
@  0x227f1ed  boost::_bi::list2<>::operator()<>()
@  0x227e7f4  boost::_bi::bind_t<>::operator()()
@  0x227dc0e  
boost::detail::function::void_function_obj_invoker0<>::invoke()
@  0x21438c3  boost::function0<>::operator()()
@  0x2721a9b  impala::Thread::SuperviseThread()
@  0x2729a38  boost::_bi::list5<>::operator()<>()
@  0x272995c  boost::_bi::bind_t<>::operator()()
@  0x272991d  boost::detail::thread_data<>::run()
@  0x3f11e61  thread_proxy
@ 0x7f460b2ffe24  start_thread
@ 0x7f4607d9634c  __clone
Picked up JAVA_TOOL_OPTIONS: 
-agentlib:jdwp=transport=dt_socket,address=30001,server=y,suspend=n  
Wrote minidump to 
/data/jenkins/workspace/impala-cdpd-master-core-s3/repos/Impala/logs/ee_tests/minidumps/impalad/e3c36842-f700-42a5-575dd186-f97a0aed.dmp
{noformat}

> Apparent hang in 
> TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action
> --
>
> Key: IMPALA-10339
> URL: https://issues.apache.org/jira/browse/IMPALA-10339
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 4.0
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, flaky, hang
>
> Release build with this commit as the tip:
> {noformat}
> commit 9400e9b17b13f613defb6d7b9deb471256b1d95c (CDH/cdpd-master-staging)
> Author: wzhou-code 
> Date:   Thu Oct 29 22:32:03 2020 -0700
> IMPALA-10305: Sync Kudu's FIPS compliant changes
> 
> {noformat}
> {noformat}
> Regression
> query_test.test_spilling.TestSpillingNoDebugActionDimensions.test_spilling_no_debug_action[protocol:
>  beeswax | exec_option: {'mt_dop': 0, 'default_spillable_buffer_size': '64k'} 
> | table_format: parquet/none] (from pytest)
> Failing for the past 1 build (Since Failed#100 )
> Took 1 hr 59 min.
> add description
> Error Message
> query_test/test_spilling.py:134: in test_spilling_no_debug_action 
> self.run_test_case('QueryTest/spilling-no-debug-action', vector) 
> common/impala_test_suite.py:668: in run_test_case 
> self.__verify_exceptions(test_section['CATCH'], str(e), use_db) 
> common/impala_test_suite.py:485: in __verify_exceptions (expected_str, 
> actual_str) E   AssertionError: Unexpected exception string. Expected: 
> row_regex:.*Cannot perform hash join at node with id .*. Repartitioning did 
> not reduce the size of a spilled partition.* E   Not found in actual: Timeout 
> >7200s
> Stacktrace
> 

[jira] [Resolved] (IMPALA-10276) Release build sees SIGSEGV when updating the total time counter

2020-11-19 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10276?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10276.

Fix Version/s: Impala 4.0
   Resolution: Fixed

> Release build sees SIGSEGV when updating the total time counter
> ---
>
> Key: IMPALA-10276
> URL: https://issues.apache.org/jira/browse/IMPALA-10276
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Affects Versions: Impala 4.0
>Reporter: Joe McDonnell
>Assignee: Tim Armstrong
>Priority: Blocker
>  Labels: broken-build, crash, flaky
> Fix For: Impala 4.0
>
>
> A recent release build saw an Impalad crash with the following stack:
> {noformat}
> Crash reason:  SIGSEGV
> Crash address: 0x11
> #0 raise () from /lib64/libc.so.6
> #1 abort () from /lib64/libc.so.6
> #2 os::abort(bool) () from 
> /usr/java/jdk1.8.0_144/jre/lib/amd64/server/libjvm.so
> #3 VMError::report_and_die() () from 
> /usr/java/jdk1.8.0_144/jre/lib/amd64/server/libjvm.so
> #4 JVM_handle_linux_signal () from 
> /usr/java/jdk1.8.0_144/jre/lib/amd64/server/libjvm.so
> #5 signalHandler(int, siginfo*, void*) () from 
> /usr/java/jdk1.8.0_144/jre/lib/amd64/server/libjvm.so
> #6  
> #7 UpdateCounter (this=0x7f93adc88030) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/util/runtime-profile-counters.h:933
> #8 impala::ScopedTimer::~ScopedTimer 
> (this=0x7f93adc88030, __in_chrg=) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/util/runtime-profile-counters.h:954
> #9 impala::Coordinator::GetNext (this=this@entry=0x13ad09e00, 
> results=results@entry=0xa7e18460, max_rows=max_rows@entry=-1, 
> eos=eos@entry=0x7f93adc880ef, 
> block_on_wait_time_us=block_on_wait_time_us@entry=0) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/runtime/coordinator.cc:864
> #10 impala::ClientRequestState::FetchRowsInternal 
> (this=this@entry=0x11568000, max_rows=max_rows@entry=-1, 
> fetched_rows=fetched_rows@entry=0xa7e18460, 
> block_on_wait_time_us=block_on_wait_time_us@entry=0) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/service/client-request-state.cc:1090
> #11 impala::ClientRequestState::FetchRows (this=0x11568000, 
> max_rows=max_rows@entry=-1, fetched_rows=fetched_rows@entry=0xa7e18460, 
> block_on_wait_time_us=0) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/service/client-request-state.cc:938
> #12 impala::ImpalaServer::FetchInternal (this=this@entry=0xdd35b00, 
> query_id=..., start_over=start_over@entry=false, 
> fetch_size=fetch_size@entry=-1, 
> query_results=query_results@entry=0x7f93adc88498) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/service/impala-beeswax-server.cc:614
> #13 impala::ImpalaServer::fetch (this=0xdd35b00, query_results=..., 
> beeswax_handle=..., start_over=, fetch_size=-1) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/service/impala-beeswax-server.cc:191
> #14 beeswax::BeeswaxServiceProcessor::process_fetch (this=0x102b0c60, 
> seqid=0, iprot=, oprot=0xd45d0980, callContext= out>) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/generated-sources/gen-cpp/BeeswaxService.cpp:3398
> #15 beeswax::BeeswaxServiceProcessor::dispatchCall (this=0x102b0c60, 
> iprot=0xd45d1e00, oprot=0xd45d0980, fname=..., seqid=0, 
> callContext=0x12ea9a80) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/generated-sources/gen-cpp/BeeswaxService.cpp:3200
> #16 apache::thrift::TDispatchProcessor::process (this=0x102b0c60, in=..., 
> out=..., connectionContext=0x12ea9a80) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/Impala-Toolchain/toolchain-packages-gcc7.5.0/thrift-0.9.3-p8/include/thrift/TDispatchProcessor.h:121
> #17 apache::thrift::server::TAcceptQueueServer::Task::run (this=0x11a92e80) 
> at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/repos/Impala/be/src/rpc/TAcceptQueueServer.cpp:84
> #18 operator() (a2=, a1=..., p=, 
> this=) at 
> /data/jenkins/workspace/impala-cdpd-master-staging-exhaustive-release/Impala-Toolchain/toolchain-packages-gcc7.5.0/boost-1.61.0-p2/include/boost/bind/mem_fn_template.hpp:280{noformat}
> The code is updating the total time counter in the runtime profile via a 
> ScopedTimer. 
> This has been seen once



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: 

[jira] [Reopened] (IMPALA-10216) BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds

2020-11-19 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10216?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reopened IMPALA-10216:


I saw this again, but unfortunately not on a branch with the logging.

> BufferPoolTest.WriteErrorBlacklistCompression is flaky on UBSAN builds
> --
>
> Key: IMPALA-10216
> URL: https://issues.apache.org/jira/browse/IMPALA-10216
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: flaky
>
> Only seen this once so far:
> {code}
> BufferPoolTest.WriteErrorBlacklistCompression
> Error Message
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> Stacktrace
> Impala/be/src/runtime/bufferpool/buffer-pool-test.cc:1764
> Value of: FindPageInDir(pages[NO_ERROR_QUERY], error_dir) != NULL
>   Actual: false
> Expected: true
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Created] (IMPALA-10343) control_service_queue_mem_limit default is too low for large clusters

2020-11-19 Thread Tim Armstrong (Jira)
Tim Armstrong created IMPALA-10343:
--

 Summary: control_service_queue_mem_limit default is too low for 
large clusters
 Key: IMPALA-10343
 URL: https://issues.apache.org/jira/browse/IMPALA-10343
 Project: IMPALA
  Issue Type: Improvement
  Components: Distributed Exec
Reporter: Tim Armstrong


We've seen some cases where this queue overflows because of status reports, 
particularly with higher mt_dop values or large clusters. We should probably 
set it to a higher value by default.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9716) Add jitter to the exponential backoff in status reporting

2020-11-19 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9716:
-

Assignee: Thomas Tauber-Marshall  (was: Tim Armstrong)

> Add jitter to the exponential backoff in status reporting
> -
>
> Key: IMPALA-9716
> URL: https://issues.apache.org/jira/browse/IMPALA-9716
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Distributed Exec
>Affects Versions: Impala 4.0
>Reporter: Thomas Tauber-Marshall
>Assignee: Thomas Tauber-Marshall
>Priority: Major
> Fix For: Impala 4.0
>
>
> When status reports fail, we use exponential backoff when retrying sending 
> them. However, currently the backoff is deterministic, leading to a 
> thundering herd problem where all of the backends for a particular query may 
> try to report at the same time, the coordinator is overwhelmed and rejects 
> some of the rpcs, then the backends all backoff by the same amount and retry 
> sending at the same time, leading the coordinator to be overwhelmed again.
> We can help solve this by adding some random jitter to the exponential 
> backoff time.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9716) Add jitter to the exponential backoff in status reporting

2020-11-19 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9716?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9716:
-

Assignee: Tim Armstrong  (was: Thomas Tauber-Marshall)

> Add jitter to the exponential backoff in status reporting
> -
>
> Key: IMPALA-9716
> URL: https://issues.apache.org/jira/browse/IMPALA-9716
> Project: IMPALA
>  Issue Type: Improvement
>  Components: Distributed Exec
>Affects Versions: Impala 4.0
>Reporter: Thomas Tauber-Marshall
>Assignee: Tim Armstrong
>Priority: Major
> Fix For: Impala 4.0
>
>
> When status reports fail, we use exponential backoff when retrying sending 
> them. However, currently the backoff is deterministic, leading to a 
> thundering herd problem where all of the backends for a particular query may 
> try to report at the same time, the coordinator is overwhelmed and rejects 
> some of the rpcs, then the backends all backoff by the same amount and retry 
> sending at the same time, leading the coordinator to be overwhelmed again.
> We can help solve this by adding some random jitter to the exponential 
> backoff time.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-7523) Planner Test failing with "Failed to assign regions to servers after 60000 millis."

2020-11-18 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-7523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-7523:
-

Assignee: (was: Tim Armstrong)

> Planner Test failing with "Failed to assign regions to servers after 6 
> millis."
> ---
>
> Key: IMPALA-7523
> URL: https://issues.apache.org/jira/browse/IMPALA-7523
> Project: IMPALA
>  Issue Type: Task
>  Components: Frontend
>Reporter: Philip Martin
>Priority: Major
>  Labels: broken-build, flaky
>
> I've seen 
> {{org.apache.impala.planner.PlannerTest.org.apache.impala.planner.PlannerTest}}
>  fail with the following trace:
> {code}
> java.lang.IllegalStateException: Failed to assign regions to servers after 
> 6 millis.
>   at 
> org.apache.impala.datagenerator.HBaseTestDataRegionAssignment.performAssignment(HBaseTestDataRegionAssignment.java:153)
>   at 
> org.apache.impala.planner.PlannerTestBase.setUp(PlannerTestBase.java:120)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:283)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:173)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:128)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:203)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:155)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
> {code}
> I think we've seen it before as indicated in IMPALA-7061.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Work started] (IMPALA-9121) heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch

2020-11-18 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on IMPALA-9121 started by Tim Armstrong.
-
> heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch
> -
>
> Key: IMPALA-9121
> URL: https://issues.apache.org/jira/browse/IMPALA-9121
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Bikramjeet Vig
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: asan, broken-build, flaky-test
>
> {noformat}
> --
> 85/113 Testing: hdfs-util-test
> 85/113 Test: hdfs-util-test
> Command: 
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/build/debug/util/hdfs-util-test"
>  "-log_dir=/data/jenkins/workspace
> /impala-asf-master-core-asan/repos/Impala/logs/be_tests"
> Directory: 
> /data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/src/util
> "hdfs-util-test" start time: Nov 01 22:21 PDT
> Output:
> --
> seed = 1572672103
> Note: Google Test filter = HdfsUtilTest.*
> [==] Running 2 tests from 1 test case.
> [--] Global test environment set-up.
> [--] 2 tests from HdfsUtilTest
> [ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
> 19/11/01 22:21:43 INFO util.JvmPauseMonitor: Starting JVM pause monitor
> [   OK ] HdfsUtilTest.CheckFilesystemsMatch (1012 ms)
> [ RUN  ] HdfsUtilTest.CheckGetBaseName
> [   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
> [--] 2 tests from HdfsUtilTest (1012 ms total)
> [--] Global test environment tear-down
> [==] 2 tests from 1 test case ran. (1013 ms total)
> [  PASSED  ] 2 tests.
>   YOU HAVE 1 DISABLED TEST
> =
> ==87116==ERROR: AddressSanitizer: heap-use-after-free on address 
> 0x60362878 at pc 0x01b67bac bp 0x7fc96cf57ae0 sp 0x7fc96cf57290
> READ of size 4 at 0x60362878 thread T4
> 
> Test time =   2.64 sec
> --
> Test Passed.
> "hdfs-util-test" end time: Nov 01 22:21 PDT
> "hdfs-util-test" time elapsed: 00:00:02
> --
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10316) load_nested.py failed due to out of memory during Jenkins GVO

2020-11-18 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235095#comment-17235095
 ] 

Tim Armstrong commented on IMPALA-10316:


I saw this again here - 
https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/12656/console

> load_nested.py failed due to out of memory during Jenkins GVO
> -
>
> Key: IMPALA-10316
> URL: https://issues.apache.org/jira/browse/IMPALA-10316
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Zoltán Borók-Nagy
>Priority: Major
>  Labels: broken-build, flaky
>
> The following job failed due to out of memory:
> [https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/12588] (please click 
> on "Don't keep this build forever" once this issue is resolved)
> Relevant log lines:
> {noformat}
> 02:33:42 Loading nested orc data (logging to 
> /home/ubuntu/Impala/logs/data_loading/load-nested.log)... 
> 02:35:39 FAILED (Took: 1 min 57 sec)
> 02:35:39 '/home/ubuntu/Impala/testdata/bin/load_nested.py -t 
> tpch_nested_orc_def -f orc/def' failed. Tail of log:
> 02:35:39 2020-11-11 02:35:06,225 INFO:load_nested[348]:Executing: 
> 02:35:39 
> 02:35:39   CREATE EXTERNAL TABLE supplier
> 02:35:39   STORED AS orc
> 02:35:39   TBLPROPERTIES('orc.compress' = 
> 'ZLIB','external.table.purge'='TRUE')
> 02:35:39   AS SELECT * FROM tmp_supplier
> 02:35:39 Traceback (most recent call last):
> 02:35:39   File "/home/ubuntu/Impala/testdata/bin/load_nested.py", line 415, 
> in 
> 02:35:39 load()
> 02:35:39   File "/home/ubuntu/Impala/testdata/bin/load_nested.py", line 349, 
> in load
> 02:35:39 hive.execute(stmt)
> 02:35:39   File "/home/ubuntu/Impala/tests/comparison/db_connection.py", line 
> 206, in execute
> 02:35:39 return self._cursor.execute(sql, *args, **kwargs)
> 02:35:39   File 
> "/home/ubuntu/Impala/infra/python/env-gcc7.5.0/lib/python2.7/site-packages/impala/hiveserver2.py",
>  line 331, in execute
> 02:35:39 self._wait_to_finish()  # make execute synchronous
> 02:35:39   File 
> "/home/ubuntu/Impala/infra/python/env-gcc7.5.0/lib/python2.7/site-packages/impala/hiveserver2.py",
>  line 413, in _wait_to_finish
> 02:35:39 raise OperationalError(resp.errorMessage)
> 02:35:39 impala.error.OperationalError: Error while compiling statement: 
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.tez.TezTask. Vertex failed, vertexName=Map 1, 
> vertexId=vertex_1605060173780_0039_2_00, diagnostics=[Task failed, 
> taskId=task_1605060173780_0039_2_00_00, diagnostics=[TaskAttempt 0 
> failed, info=[Container container_1605060173780_0039_01_02 finished with 
> diagnostics set to [Container failed, exitCode=-104. [2020-11-11 
> 02:35:11.768]Container 
> [pid=16810,containerID=container_1605060173780_0039_01_02] is running 
> 7729152B beyond the 'PHYSICAL' memory limit. Current usage: 1.0 GB of 1 GB 
> physical memory used; 2.5 GB of 2.1 GB virtual memory used. Killing 
> container.{noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9121) heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch

2020-11-18 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235089#comment-17235089
 ] 

Tim Armstrong commented on IMPALA-9121:
---

I'm suspicious about how it's using ExecEnv - that starts up a bunch of 
background threads that might be the ones causing the problem here during 
teardown. I think I can minimize that.

> heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch
> -
>
> Key: IMPALA-9121
> URL: https://issues.apache.org/jira/browse/IMPALA-9121
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Bikramjeet Vig
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: asan, broken-build, flaky-test
>
> {noformat}
> --
> 85/113 Testing: hdfs-util-test
> 85/113 Test: hdfs-util-test
> Command: 
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/build/debug/util/hdfs-util-test"
>  "-log_dir=/data/jenkins/workspace
> /impala-asf-master-core-asan/repos/Impala/logs/be_tests"
> Directory: 
> /data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/src/util
> "hdfs-util-test" start time: Nov 01 22:21 PDT
> Output:
> --
> seed = 1572672103
> Note: Google Test filter = HdfsUtilTest.*
> [==] Running 2 tests from 1 test case.
> [--] Global test environment set-up.
> [--] 2 tests from HdfsUtilTest
> [ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
> 19/11/01 22:21:43 INFO util.JvmPauseMonitor: Starting JVM pause monitor
> [   OK ] HdfsUtilTest.CheckFilesystemsMatch (1012 ms)
> [ RUN  ] HdfsUtilTest.CheckGetBaseName
> [   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
> [--] 2 tests from HdfsUtilTest (1012 ms total)
> [--] Global test environment tear-down
> [==] 2 tests from 1 test case ran. (1013 ms total)
> [  PASSED  ] 2 tests.
>   YOU HAVE 1 DISABLED TEST
> =
> ==87116==ERROR: AddressSanitizer: heap-use-after-free on address 
> 0x60362878 at pc 0x01b67bac bp 0x7fc96cf57ae0 sp 0x7fc96cf57290
> READ of size 4 at 0x60362878 thread T4
> 
> Test time =   2.64 sec
> --
> Test Passed.
> "hdfs-util-test" end time: Nov 01 22:21 PDT
> "hdfs-util-test" time elapsed: 00:00:02
> --
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-9121) heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch

2020-11-18 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17235070#comment-17235070
 ] 

Tim Armstrong commented on IMPALA-9121:
---

This does reproduce after looping for a while, I guess it's happening after the 
tests finish and are torn down.

{noformat}
Note: Google Test filter = HdfsUtilTest.*
[==] Running 2 tests from 1 test case.
[--] Global test environment set-up.
[--] 2 tests from HdfsUtilTest
[ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
20/11/18 15:59:03 INFO util.JvmPauseMonitor: Starting JVM pause monitor
[   OK ] HdfsUtilTest.CheckFilesystemsMatch (935 ms)
[ RUN  ] HdfsUtilTest.CheckGetBaseName
[   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
[--] 2 tests from HdfsUtilTest (935 ms total)

[--] Global test environment tear-down
[==] 2 tests from 1 test case ran. (935 ms total)
[  PASSED  ] 2 tests.

  YOU HAVE 1 DISABLED TEST

=
==31924==ERROR: AddressSanitizer: heap-use-after-free on address 0x6070ead0 
at pc 0x01
e34a4b bp 0x7fa2e44ffac0 sp 0x7fa2e44ff270
READ of size 5 at 0x6070ead0 thread T4
{noformat}

> heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch
> -
>
> Key: IMPALA-9121
> URL: https://issues.apache.org/jira/browse/IMPALA-9121
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Bikramjeet Vig
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: asan, broken-build, flaky-test
>
> {noformat}
> --
> 85/113 Testing: hdfs-util-test
> 85/113 Test: hdfs-util-test
> Command: 
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/build/debug/util/hdfs-util-test"
>  "-log_dir=/data/jenkins/workspace
> /impala-asf-master-core-asan/repos/Impala/logs/be_tests"
> Directory: 
> /data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/src/util
> "hdfs-util-test" start time: Nov 01 22:21 PDT
> Output:
> --
> seed = 1572672103
> Note: Google Test filter = HdfsUtilTest.*
> [==] Running 2 tests from 1 test case.
> [--] Global test environment set-up.
> [--] 2 tests from HdfsUtilTest
> [ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
> 19/11/01 22:21:43 INFO util.JvmPauseMonitor: Starting JVM pause monitor
> [   OK ] HdfsUtilTest.CheckFilesystemsMatch (1012 ms)
> [ RUN  ] HdfsUtilTest.CheckGetBaseName
> [   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
> [--] 2 tests from HdfsUtilTest (1012 ms total)
> [--] Global test environment tear-down
> [==] 2 tests from 1 test case ran. (1013 ms total)
> [  PASSED  ] 2 tests.
>   YOU HAVE 1 DISABLED TEST
> =
> ==87116==ERROR: AddressSanitizer: heap-use-after-free on address 
> 0x60362878 at pc 0x01b67bac bp 0x7fc96cf57ae0 sp 0x7fc96cf57290
> READ of size 4 at 0x60362878 thread T4
> 
> Test time =   2.64 sec
> --
> Test Passed.
> "hdfs-util-test" end time: Nov 01 22:21 PDT
> "hdfs-util-test" time elapsed: 00:00:02
> --
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Assigned] (IMPALA-9121) heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch

2020-11-18 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9121?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong reassigned IMPALA-9121:
-

Assignee: Tim Armstrong

> heap-use-after-free in HdfsUtilTest CheckFilesystemsMatch
> -
>
> Key: IMPALA-9121
> URL: https://issues.apache.org/jira/browse/IMPALA-9121
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Bikramjeet Vig
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: asan, broken-build, flaky-test
>
> {noformat}
> --
> 85/113 Testing: hdfs-util-test
> 85/113 Test: hdfs-util-test
> Command: 
> "/data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/build/debug/util/hdfs-util-test"
>  "-log_dir=/data/jenkins/workspace
> /impala-asf-master-core-asan/repos/Impala/logs/be_tests"
> Directory: 
> /data/jenkins/workspace/impala-asf-master-core-asan/repos/Impala/be/src/util
> "hdfs-util-test" start time: Nov 01 22:21 PDT
> Output:
> --
> seed = 1572672103
> Note: Google Test filter = HdfsUtilTest.*
> [==] Running 2 tests from 1 test case.
> [--] Global test environment set-up.
> [--] 2 tests from HdfsUtilTest
> [ RUN  ] HdfsUtilTest.CheckFilesystemsMatch
> 19/11/01 22:21:43 INFO util.JvmPauseMonitor: Starting JVM pause monitor
> [   OK ] HdfsUtilTest.CheckFilesystemsMatch (1012 ms)
> [ RUN  ] HdfsUtilTest.CheckGetBaseName
> [   OK ] HdfsUtilTest.CheckGetBaseName (0 ms)
> [--] 2 tests from HdfsUtilTest (1012 ms total)
> [--] Global test environment tear-down
> [==] 2 tests from 1 test case ran. (1013 ms total)
> [  PASSED  ] 2 tests.
>   YOU HAVE 1 DISABLED TEST
> =
> ==87116==ERROR: AddressSanitizer: heap-use-after-free on address 
> 0x60362878 at pc 0x01b67bac bp 0x7fc96cf57ae0 sp 0x7fc96cf57290
> READ of size 4 at 0x60362878 thread T4
> 
> Test time =   2.64 sec
> --
> Test Passed.
> "hdfs-util-test" end time: Nov 01 22:21 PDT
> "hdfs-util-test" time elapsed: 00:00:02
> --
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-9965) test_result_spooling seems to be still flaky under asan

2020-11-18 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-9965?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-9965.
---
Resolution: Cannot Reproduce

I looped for a few hours locally and no luck. Haven't seen it happen recently. 

Reopen and attach coordinator logs if it reoccurs.

> test_result_spooling seems to be still flaky under asan
> ---
>
> Key: IMPALA-9965
> URL: https://issues.apache.org/jira/browse/IMPALA-9965
> Project: IMPALA
>  Issue Type: Bug
>  Components: Infrastructure
>Reporter: Tim Armstrong
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
>
> {noformat}
> Error Message
> query_test/test_result_spooling.py:98: in test_spilling .format(query, 
> timeout)) E   Timeout: Query select * from functional.alltypes order by id 
> limit 1500 did not spill spooled results within the timeout 30
> Stacktrace
> query_test/test_result_spooling.py:98: in test_spilling
> .format(query, timeout))
> E   Timeout: Query select * from functional.alltypes order by id limit 1500 
> did not spill spooled results within the timeout 30
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Commented] (IMPALA-10111) TestWebPage::test_query_stmt is flaky

2020-11-18 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-10111?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17234955#comment-17234955
 ] 

Tim Armstrong commented on IMPALA-10111:


I wonder if it just aged out of the query logs and could be fixed by increasing 
the query log size...

Anyway I'm not having luck reproducing it so I'll close it out.

If it reoccurs we should attach logs

> TestWebPage::test_query_stmt is flaky
> -
>
> Key: IMPALA-10111
> URL: https://issues.apache.org/jira/browse/IMPALA-10111
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Quanlong Huang
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky
>
> Found in an unrelated patch: 
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/3032/testReport/junit/webserver.test_web_pages/TestWebPage/test_query_stmt/
> {code:java}
> webserver/test_web_pages.py:418: in test_query_stmt
> assert check_if_contains, "No matching statement found in the jsons."
> E   AssertionError: No matching statement found in the jsons.
> E   assert False {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10111) TestWebPage::test_query_stmt is flaky

2020-11-18 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10111?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10111.

Resolution: Cannot Reproduce

> TestWebPage::test_query_stmt is flaky
> -
>
> Key: IMPALA-10111
> URL: https://issues.apache.org/jira/browse/IMPALA-10111
> Project: IMPALA
>  Issue Type: Bug
>Reporter: Quanlong Huang
>Assignee: Tim Armstrong
>Priority: Major
>  Labels: broken-build, flaky
>
> Found in an unrelated patch: 
> https://jenkins.impala.io/job/ubuntu-16.04-dockerised-tests/3032/testReport/junit/webserver.test_web_pages/TestWebPage/test_query_stmt/
> {code:java}
> webserver/test_web_pages.py:418: in test_query_stmt
> assert check_if_contains, "No matching statement found in the jsons."
> E   AssertionError: No matching statement found in the jsons.
> E   assert False {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



[jira] [Resolved] (IMPALA-10293) Improvement to TestParquet.test_multiple_blocks_mt_dop

2020-11-18 Thread Tim Armstrong (Jira)


 [ 
https://issues.apache.org/jira/browse/IMPALA-10293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Armstrong resolved IMPALA-10293.

Resolution: Not A Bug

It looks like this failure occurred on a test run where an impalad crashed, 
hence the difference in distribution across nodes:

{noformat}
Error Message

Minidump generated: 
/data/jenkins/workspace/impala-asf-master-exhaustive/repos/Impala/logs/ee_tests/minidumps/impalad/20d30f8e-cb1c-46fd-10f827af-c504e9ab.dmp

Standard Error

Operating system: Linux
  0.0.0 Linux 3.10.0-693.5.2.el7.x86_64 #1 SMP Fri Oct 20 
20:32:50 UTC 2017 x86_64
CPU: amd64
 family 6 model 85 stepping 4
 1 CPU

GPU: UNKNOWN

Crash reason:  SIGSEGV
Crash address: 0x2a
Process uptime: not available

Thread 406 (crashed)
 0  impalad!tcmalloc::CentralFreeList::FetchFromOneSpans(int, void**, void**) + 
0x3b
rax = 0x0001   rdx = 0x7ffa171cd220
rcx = 0x002a   rbx = 0x002a
rsi = 0x0001   rdi = 0x07721d00
rbp = 0x7ffa171cd180   rsp = 0x7ffa171cd150
 r8 = 0x07721d40r9 = 0x7ffa171cd378
r10 = 0x7ffaff2db8c0   r11 = 0x7ffafe802cd0
r12 = 0x07721d00   r13 = 0x09255a00
r14 = 0x7ffa171cd228   r15 = 0x09093140
rip = 0x05331a3b
Found by: given as instruction pointer in context
 1  impalad!tcmalloc::CentralFreeList::FetchFromOneSpansSafe(int, void**, 
void**) + 0x1c
rbx = 0x07721d00   rbp = 0x7ffa171cd1b0
rsp = 0x7ffa171cd190   r12 = 0x0001
r13 = 0x7ffa171cd220   r14 = 0x7ffa171cd228
r15 = 0x09093140   rip = 0x05331d1c
Found by: call frame info
 2  impalad!tcmalloc::CentralFreeList::RemoveRange(void**, void**, int) + 0xaf
rbx = 0x090933e0   rbp = 0x7ffa171cd210
rsp = 0x7ffa171cd1c0   r12 = 0x07721d00
r13 = 0x0001   r14 = 0x7ffa171cd220
r15 = 0x09093140   rip = 0x05331e0f
Found by: call frame info
 3  impalad!tcmalloc::ThreadCache::FetchFromCentralCache(unsigned long, 
unsigned long) + 0x5d
rbx = 0x090933e0   rbp = 0x7ffa171cd260
rsp = 0x7ffa171cd220   r12 = 0x0240
r13 = 0x0071   r14 = 0x0270
r15 = 0x09093140   rip = 0x053417ad
Found by: call frame info
 4  impalad!tc_newarray + 0x25b
rbx = 0x09093140   rbp = 0x7ffa171cd2a0
rsp = 0x7ffa171cd270   r12 = 0x0201
r13 = 0x0240   r14 = 0x7ffa171cd387
r15 = 0x   rip = 0x0547837b
Found by: call frame info
 5  libstdc++.so.6.0.24 + 0x11f5cd
rbx = 0x7ffa171cd2e0   rbp = 0x7ffa171cd2f0
rsp = 0x7ffa171cd2b0   r12 = 0x7ffa171cd2f0
r13 = 0x0010   r14 = 0x7ffa171cd387
r15 = 0x   rip = 0x7ffaff07a5cd
Found by: call frame info
 6  libstdc++.so.6.0.24 + 0x112e0b
rbp = 0x7ffa171cd2f0   rsp = 0x7ffa171cd2e0
rip = 0x7ffaff06de0b
Found by: stack scanning
 7  libstdc++.so.6.0.24 + 0x11db89
rsp = 0x7ffa171cd330   rip = 0x7ffaff078b89
Found by: stack scanning
 8  libstdc++.so.6.0.24 + 0x102bd3
rsp = 0x7ffa171cd360   rip = 0x7ffaff05dbd3
Found by: stack scanning
 9  libstdc++.so.6.0.24 + 0x1029e9
rsp = 0x7ffa171cd370   rip = 0x7ffaff05d9e9
Found by: stack scanning
10  libstdc++.so.6.0.24 + 0x102b05
rsp = 0x7ffa171cd3a0   rip = 0x7ffaff05db05
Found by: stack scanning
11  libstdc++.so.6.0.24 + 0x10e3b5
rsp = 0x7ffa171cd420   rip = 0x7ffaff0693b5
Found by: stack scanning
12  impalad!impala::PrintId(impala::TUniqueId const&, 
std::__cxx11::basic_string, std::allocator > 
const&) [debug-util.cc : 109 + 0x19]
rsp = 0x7ffa171cd470   rip = 0x0264b193
Found by: stack scanning

Thread 0
 0  libpthread-2.17.so + 0xb945
rax = 0xfe00   rdx = 0x0001
rcx = 0x   rbx = 0x
rsi = 0x0080   rdi = 0x139d1444
rbp = 0x7ffebf6bef60   rsp = 0x7ffebf6beea0
 r8 = 0x139d1400r9 = 0x
r10 = 0x   r11 = 0x0246
r12 = 0x12a7a500   r13 = 0x7ffebf6bf8f0
r14 = 0x   r15 = 0x
rip = 0x7ffb01ce4945
Found by: given as instruction pointer in context
 1  impalad!boost::thread::join_noexcept() + 0x54
rbp = 0x7ffebf6bef80   rsp = 0x7ffebf6bef70
rip = 0x03f375d4
Found by: previous frame's frame pointer
 2  impalad!boost::thread::join() [thread.hpp : 766 + 0xc]
rbx = 0x0001   rbp = 0x7ffebf6bf030
rsp = 0x7ffebf6befc0   r12 = 0x0ea55538
r13 = 0x7ffebf6bf8f0   

[jira] [Commented] (IMPALA-9058) S3 tests failing with FileNotFoundException getVersionMarkerItem on ../VERSION

2020-11-18 Thread Tim Armstrong (Jira)


[ 
https://issues.apache.org/jira/browse/IMPALA-9058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17234949#comment-17234949
 ] 

Tim Armstrong commented on IMPALA-9058:
---

I think what I need to do next is find an instance of this and dig through the 
logs to trace the full lifecycle of the table with the problem, i.e. understand 
how/when was loaded, when it was accessed by any queries, etc. Maybe that will 
provide some clues.

> S3 tests failing with FileNotFoundException getVersionMarkerItem on ../VERSION
> --
>
> Key: IMPALA-9058
> URL: https://issues.apache.org/jira/browse/IMPALA-9058
> Project: IMPALA
>  Issue Type: Bug
>  Components: Backend
>Reporter: Sahil Takiar
>Assignee: Tim Armstrong
>Priority: Critical
>  Labels: broken-build, flaky
>
> I've seen this happen several times now, S3 tests intermittently fail with an 
> error such as:
> {code:java}
> Query aborted:InternalException: Error adding partitions E   CAUSED BY: 
> MetaException: java.io.IOException: Got exception: 
> java.io.FileNotFoundException getVersionMarkerItem on ../VERSION: 
> com.amazonaws.services.dynamodbv2.model.ResourceNotFoundException: Requested 
> resource not found (Service: AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ResourceNotFoundException; Request ID: 
> 8T9IS939MDI7ASOB0IJCC34J3NVV4KQNSO5AEMVJF66Q9ASUAAJG) {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-all-unsubscr...@impala.apache.org
For additional commands, e-mail: issues-all-h...@impala.apache.org



<    5   6   7   8   9   10   11   12   13   14   >