David Knupp created IMPALA-7804:
-----------------------------------

             Summary: Various scanner tests intermittently failing on S3 on 
different runs
                 Key: IMPALA-7804
                 URL: https://issues.apache.org/jira/browse/IMPALA-7804
             Project: IMPALA
          Issue Type: Bug
          Components: Backend
    Affects Versions: Impala 3.1.0
            Reporter: David Knupp


Th failures have to do with getting AWS client credentials.

*query_test/test_scanners.py:696: in test_decimal_encodings*

_Stacktrace_
{noformat}
query_test/test_scanners.py:696: in test_decimal_encodings
    self.run_test_case('QueryTest/parquet-decimal-formats', vector, 
unique_database)
common/impala_test_suite.py:496: in run_test_case
    self.__verify_results_and_errors(vector, test_section, result, use_db)
common/impala_test_suite.py:358: in __verify_results_and_errors
    replace_filenames_with_placeholder)
common/test_result_verifier.py:438: in verify_raw_results
    VERIFIER_MAP[verifier](expected, actual)
common/test_result_verifier.py:260: in verify_query_result_is_equal
    assert expected_results == actual_results
E   assert Comparing QueryTestResults (expected vs actual):
E     -255.00,-255.00,-255.00 == -255.00,-255.00,-255.00
E     -255.00,-255.00,-255.00 != -65535.00,-65535.00,-65535.00
E     -65535.00,-65535.00,-65535.00 != -9999999.99,-9999999.99,-9999999.99
E     -65535.00,-65535.00,-65535.00 != 
0.00,-9999999999999999.99,-999999999999999999999999999999999999.99
E     -9999999.99,-9999999.99,-9999999.99 != 0.00,0.00,0.00
E     -9999999.99,-9999999.99,-9999999.99 != 
0.00,9999999999999999.99,999999999999999999999999999999999999.99
E     0.00,-9999999999999999.99,-999999999999999999999999999999999999.99 != 
255.00,255.00,255.00
E     0.00,-9999999999999999.99,-999999999999999999999999999999999999.99 != 
65535.00,65535.00,65535.00
E     0.00,0.00,0.00 != 9999999.99,9999999.99,9999999.99
E     0.00,0.00,0.00 != None
E     0.00,9999999999999999.99,999999999999999999999999999999999999.99 != None
E     0.00,9999999999999999.99,999999999999999999999999999999999999.99 != None
E     255.00,255.00,255.00 != None
E     255.00,255.00,255.00 != None
E     65535.00,65535.00,65535.00 != None
E     65535.00,65535.00,65535.00 != None
E     9999999.99,9999999.99,9999999.99 != None
E     9999999.99,9999999.99,9999999.99 != None
E     Number of rows returned (expected vs actual): 18 != 9
{noformat}

_Standard Error_
{noformat}
SET sync_ddl=False;
-- executing against localhost:21000
DROP DATABASE IF EXISTS `test_huge_num_rows_76a09ef1` CASCADE;

-- 2018-11-01 09:42:41,140 INFO     MainThread: Started query 
4c4bc0e7b69d7641:130ffe7300000000
SET sync_ddl=False;
-- executing against localhost:21000
CREATE DATABASE `test_huge_num_rows_76a09ef1`;

-- 2018-11-01 09:42:42,402 INFO     MainThread: Started query 
e34d714d6a62cba1:2a8544d000000000
-- 2018-11-01 09:42:42,405 INFO     MainThread: Created database 
"test_huge_num_rows_76a09ef1" for test ID 
"query_test/test_scanners.py::TestParquet::()::test_huge_num_rows[protocol: 
beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'debug_action': '-1:OPEN:[email protected]', 
'exec_single_node_rows_threshold': 0} | table_format: parquet/none]"
18/11/01 09:42:43 DEBUG s3a.S3AFileSystem: Initializing S3AFileSystem for 
impala-test-uswest2-1
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Propagating entries under 
fs.s3a.bucket.impala-test-uswest2-1.
18/11/01 09:42:43 WARN impl.MetricsConfig: Cannot locate configuration: tried 
hadoop-metrics2-s3a-file-system.properties,hadoop-metrics2.properties
18/11/01 09:42:43 INFO impl.MetricsSystemImpl: Scheduled Metric snapshot period 
at 10 second(s).
18/11/01 09:42:43 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
started
18/11/01 09:42:43 DEBUG s3a.S3AUtils: For URI s3a://impala-test-uswest2-1/, 
using credentials AWSCredentialProviderList: BasicAWSCredentialsProvider 
EnvironmentVariableCredentialsProvider 
com.amazonaws.auth.InstanceProfileCredentialsProvider@15bbf42f
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.connection.maximum is 1500
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.attempts.maximum is 20
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of 
fs.s3a.connection.establish.timeout is 5000
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.connection.timeout is 
200000
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.socket.send.buffer is 8192
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.socket.recv.buffer is 8192
18/11/01 09:42:43 DEBUG s3a.S3AFileSystem: Using User-Agent: Hadoop 
3.0.0-cdh6.x-SNAPSHOT
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.paging.maximum is 5000
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.block.size is 33554432
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.readahead.range is 65536
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.max.total.tasks is 5
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.threads.keepalivetime is 
60
18/11/01 09:42:44 DEBUG s3a.AWSCredentialProviderList: No credentials provided 
by BasicAWSCredentialsProvider: 
org.apache.hadoop.fs.s3a.CredentialInitializationException: Access key or 
secret key is null
org.apache.hadoop.fs.s3a.CredentialInitializationException: Access key or 
secret key is null
        at 
org.apache.hadoop.fs.s3a.BasicAWSCredentialsProvider.getCredentials(BasicAWSCredentialsProvider.java:51)
        at 
org.apache.hadoop.fs.s3a.AWSCredentialProviderList.getCredentials(AWSCredentialProviderList.java:117)
{noformat}

*query_test/test_scanners.py:351: in test_huge_num_rows*

_Stacktrace_
{noformat}
query_test/test_scanners.py:351: in test_huge_num_rows
    % unique_database)
common/impala_connection.py:170: in execute
    return self.__beeswax_client.execute(sql_stmt, user=user)
beeswax/impala_beeswax.py:182: in execute
    handle = self.__execute_query(query_string.strip(), user=user)
beeswax/impala_beeswax.py:356: in __execute_query
    self.wait_for_finished(handle)
beeswax/impala_beeswax.py:377: in wait_for_finished
    raise ImpalaBeeswaxException("Query aborted:" + error_log, None)
E   ImpalaBeeswaxException: ImpalaBeeswaxException:
E    Query aborted:Disk I/O error: Failed to open HDFS file 
s3a://impala-test-uswest2-1/test-warehouse/test_huge_num_rows_76a09ef1.db/huge_num_rows/huge_num_rows.parquet
E   Error(2): No such file or directory
E   Root cause: FileNotFoundException: No such file or directory: 
s3a://impala-test-uswest2-1/test-warehouse/test_huge_num_rows_76a09ef1.db/huge_num_rows/huge_num_rows.parquet
{noformat}

_Standard Error_
{noformat}
SET sync_ddl=False;
-- executing against localhost:21000
DROP DATABASE IF EXISTS `test_huge_num_rows_76a09ef1` CASCADE;

-- 2018-11-01 09:42:41,140 INFO     MainThread: Started query 
4c4bc0e7b69d7641:130ffe7300000000
SET sync_ddl=False;
-- executing against localhost:21000
CREATE DATABASE `test_huge_num_rows_76a09ef1`;

-- 2018-11-01 09:42:42,402 INFO     MainThread: Started query 
e34d714d6a62cba1:2a8544d000000000
-- 2018-11-01 09:42:42,405 INFO     MainThread: Created database 
"test_huge_num_rows_76a09ef1" for test ID 
"query_test/test_scanners.py::TestParquet::()::test_huge_num_rows[protocol: 
beeswax | exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'debug_action': '-1:OPEN:[email protected]', 
'exec_single_node_rows_threshold': 0} | table_format: parquet/none]"
18/11/01 09:42:43 DEBUG s3a.S3AFileSystem: Initializing S3AFileSystem for 
impala-test-uswest2-1
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Propagating entries under 
fs.s3a.bucket.impala-test-uswest2-1.
18/11/01 09:42:43 WARN impl.MetricsConfig: Cannot locate configuration: tried 
hadoop-metrics2-s3a-file-system.properties,hadoop-metrics2.properties
18/11/01 09:42:43 INFO impl.MetricsSystemImpl: Scheduled Metric snapshot period 
at 10 second(s).
18/11/01 09:42:43 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
started
18/11/01 09:42:43 DEBUG s3a.S3AUtils: For URI s3a://impala-test-uswest2-1/, 
using credentials AWSCredentialProviderList: BasicAWSCredentialsProvider 
EnvironmentVariableCredentialsProvider 
com.amazonaws.auth.InstanceProfileCredentialsProvider@15bbf42f
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.connection.maximum is 1500
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.attempts.maximum is 20
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of 
fs.s3a.connection.establish.timeout is 5000
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.connection.timeout is 
200000
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.socket.send.buffer is 8192
18/11/01 09:42:43 DEBUG s3a.S3AUtils: Value of fs.s3a.socket.recv.buffer is 8192
18/11/01 09:42:43 DEBUG s3a.S3AFileSystem: Using User-Agent: Hadoop 
3.0.0-cdh6.x-SNAPSHOT
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.paging.maximum is 5000
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.block.size is 33554432
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.readahead.range is 65536
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.max.total.tasks is 5
18/11/01 09:42:44 DEBUG s3a.S3AUtils: Value of fs.s3a.threads.keepalivetime is 
60
18/11/01 09:42:44 DEBUG s3a.AWSCredentialProviderList: No credentials provided 
by BasicAWSCredentialsProvider: 
org.apache.hadoop.fs.s3a.CredentialInitializationException: Access key or 
secret key is null
org.apache.hadoop.fs.s3a.CredentialInitializationException: Access key or 
secret key is null
{noformat}





--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to