[ 
https://issues.apache.org/jira/browse/IMPALA-8500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16833969#comment-16833969
 ] 

Tim Armstrong commented on IMPALA-8500:
---------------------------------------

I also see another weird S3 consistency issue:
{noformat}

query_test.test_scanners.TestParquet.test_decimal_encodings[protocol: beeswax | 
exec_option: {'batch_size': 0, 'num_nodes': 0, 
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 
1, 'debug_action': '-1:OPEN:[email protected]', 
'exec_single_node_rows_threshold': 0} | table_format: parquet/none] (from 
pytest)

Failing for the past 1 build (Since Failed#202 )
Took 17 sec.
add description
Error Message
query_test/test_scanners.py:715: in test_decimal_encodings     
create_table_from_parquet(self.client, unique_database, 
'decimal_stored_as_int32') common/file_utils.py:48: in 
create_table_from_parquet     qualified_table_name, hdfs_file)) 
common/impala_connection.py:180: in execute     return 
self.__beeswax_client.execute(sql_stmt, user=user) 
beeswax/impala_beeswax.py:187: in execute     handle = 
self.__execute_query(query_string.strip(), user=user) 
beeswax/impala_beeswax.py:362: in __execute_query     handle = 
self.execute_query_async(query_string, user=user) 
beeswax/impala_beeswax.py:356: in execute_query_async     handle = 
self.__do_rpc(lambda: self.imp_service.query(query,)) 
beeswax/impala_beeswax.py:516: in __do_rpc     raise 
ImpalaBeeswaxException(self.__build_error_message(b), b) E   
ImpalaBeeswaxException: ImpalaBeeswaxException: E    INNER EXCEPTION: <class 
'beeswaxd.ttypes.BeeswaxException'> E    MESSAGE: AnalysisException: Cannot 
infer schema, path is not a file: 
s3a://impala-test-uswest2-2/test-warehouse/test_decimal_encodings_db2969ca.db/decimal_stored_as_int32/decimal_stored_as_int32.parquet
{noformat}

{noformat}
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: FileBlock{index=1, 
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp, 
state=Upload, dataSize=597, limit=104857600}: entering state Closed
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: Closed FileBlock{index=1, 
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp, 
state=Closed, dataSize=597, limit=104857600}
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: Closing FileBlock{index=1, 
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp, 
state=Closed, dataSize=597, limit=104857600}
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: block[1]: closeBlock()
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Upload complete to 
test-warehouse/test_decimal_encodings_db2969ca.db/decimal_stored_as_int32/decimal_stored_as_int32.parquet
 by WriteOperationHelper {bucket=impala-test-uswest2-2}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Closing FileBlock{index=1, 
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp, 
state=Closed, dataSize=597, limit=104857600}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Closing 
org.apache.hadoop.fs.s3a.S3ADataBlocks$DiskBlockFactory@29caf222
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Statistics: 
OutputStreamStatistics{blocksSubmitted=1, blocksInQueue=1, blocksActive=0, 
blockUploadsCompleted=0, blockUploadsFailed=0, bytesPendingUpload=0, 
bytesUploaded=597, blocksAllocated=1, blocksReleased=1, 
blocksActivelyAllocated=0, exceptionsInMultipartFinalize=0, transferDuration=0 
ms, queueDuration=0 ms, averageQueueTime=0 ms, totalUploadDuration=0 ms, 
effectiveBandwidth=0.0 bytes/s}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Closing 
OutputStreamStatistics{blocksSubmitted=1, blocksInQueue=1, blocksActive=0, 
blockUploadsCompleted=0, blockUploadsFailed=0, bytesPendingUpload=0, 
bytesUploaded=597, blocksAllocated=1, blocksReleased=1, 
blocksActivelyAllocated=0, exceptionsInMultipartFinalize=0, transferDuration=0 
ms, queueDuration=0 ms, averageQueueTime=0 ms, totalUploadDuration=0 ms, 
effectiveBandwidth=0.0 bytes/s}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Ignoring close() as stream is 
already closed
19/05/03 11:07:17 DEBUG s3a.S3AFileSystem: Filesystem 
s3a://impala-test-uswest2-2 is closed
19/05/03 11:07:17 INFO impl.MetricsSystemImpl: Stopping s3a-file-system metrics 
system...
19/05/03 11:07:17 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
stopped.
19/05/03 11:07:17 INFO impl.MetricsSystemImpl: s3a-file-system metrics system 
shutdown complete.
-- executing against localhost:21000
create table test_decimal_encodings_db2969ca.decimal_stored_as_int32 like 
parquet 
"/test-warehouse/test_decimal_encodings_db2969ca.db/decimal_stored_as_int32/decimal_stored_as_int32.parquet"
 stored as parquet;
{noformat}

> test_timestamp_out_of_range fails with NoSuchObjectException: 
> test_timestamp_out_of_range_dc37915d on S3
> --------------------------------------------------------------------------------------------------------
>
>                 Key: IMPALA-8500
>                 URL: https://issues.apache.org/jira/browse/IMPALA-8500
>             Project: IMPALA
>          Issue Type: Bug
>          Components: Infrastructure
>            Reporter: Tim Armstrong
>            Assignee: Joe McDonnell
>            Priority: Critical
>              Labels: broken-build
>
> I'm not sure what to make of this. Maybe you have an idea:
> {noformat}
> Error Message
> test setup failure
> Stacktrace
> conftest.py:319: in cleanup
>     {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:620: in wrapper
>     return function(*args, **kwargs)
> common/impala_test_suite.py:628: in execute_query_expect_success
>     result = cls.__execute_query(impalad_client, query, query_options, user)
> common/impala_test_suite.py:722: in __execute_query
>     return impalad_client.execute(query, user=user)
> common/impala_connection.py:180: in execute
>     return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:187: in execute
>     handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:362: in __execute_query
>     handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:356: in execute_query_async
>     handle = self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:516: in __do_rpc
>     raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E   ImpalaBeeswaxException: ImpalaBeeswaxException:
> E    INNER EXCEPTION: <class 'beeswaxd.ttypes.BeeswaxException'>
> E    MESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive 
> Metastore: 
> E   CAUSED BY: NoSuchObjectException: test_timestamp_out_of_range_dc37915d
> {noformat}
> It does look like it created the database and used it fine:
> {noformat}
> -- 2019-05-02 11:09:21,883 INFO     MainThread: Started query 
> e847e67be097c2a0:6c29d4f400000000
> SET 
> client_identifier=query_test/test_scanners.py::TestParquet::()::test_timestamp_out_of_range[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action':None;'exec_single_node_rows_t;
> -- executing against localhost:21000
> use test_timestamp_out_of_range_dc37915d;
> -- 2019-05-02 11:09:22,266 INFO     MainThread: Started query 
> ab498d07351289b2:03f6198300000000
> SET 
> client_identifier=query_test/test_scanners.py::TestParquet::()::test_timestamp_out_of_range[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action':None;'exec_single_node_rows_t;
> SET batch_size=0;
> SET num_nodes=0;
> SET disable_codegen_rows_threshold=0;
> SET disable_codegen=False;
> SET abort_on_error=1;
> SET exec_single_node_rows_threshold=0;
> -- executing against localhost:21000
> SELECT * FROM out_of_range_timestamp;
> -- 2019-05-02 11:09:22,273 INFO     MainThread: Started query 
> be4886563821196f:871eb37f00000000
> -- executing against localhost:21000
> SELECT * FROM out_of_range_time_of_day;
> {noformat}
> Maybe some S3 consistency issue?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to