[
https://issues.apache.org/jira/browse/IMPALA-8500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16833969#comment-16833969
]
Tim Armstrong commented on IMPALA-8500:
---------------------------------------
I also see another weird S3 consistency issue:
{noformat}
query_test.test_scanners.TestParquet.test_decimal_encodings[protocol: beeswax |
exec_option: {'batch_size': 0, 'num_nodes': 0,
'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error':
1, 'debug_action': '-1:OPEN:[email protected]',
'exec_single_node_rows_threshold': 0} | table_format: parquet/none] (from
pytest)
Failing for the past 1 build (Since Failed#202 )
Took 17 sec.
add description
Error Message
query_test/test_scanners.py:715: in test_decimal_encodings
create_table_from_parquet(self.client, unique_database,
'decimal_stored_as_int32') common/file_utils.py:48: in
create_table_from_parquet qualified_table_name, hdfs_file))
common/impala_connection.py:180: in execute return
self.__beeswax_client.execute(sql_stmt, user=user)
beeswax/impala_beeswax.py:187: in execute handle =
self.__execute_query(query_string.strip(), user=user)
beeswax/impala_beeswax.py:362: in __execute_query handle =
self.execute_query_async(query_string, user=user)
beeswax/impala_beeswax.py:356: in execute_query_async handle =
self.__do_rpc(lambda: self.imp_service.query(query,))
beeswax/impala_beeswax.py:516: in __do_rpc raise
ImpalaBeeswaxException(self.__build_error_message(b), b) E
ImpalaBeeswaxException: ImpalaBeeswaxException: E INNER EXCEPTION: <class
'beeswaxd.ttypes.BeeswaxException'> E MESSAGE: AnalysisException: Cannot
infer schema, path is not a file:
s3a://impala-test-uswest2-2/test-warehouse/test_decimal_encodings_db2969ca.db/decimal_stored_as_int32/decimal_stored_as_int32.parquet
{noformat}
{noformat}
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: FileBlock{index=1,
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp,
state=Upload, dataSize=597, limit=104857600}: entering state Closed
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: Closed FileBlock{index=1,
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp,
state=Closed, dataSize=597, limit=104857600}
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: Closing FileBlock{index=1,
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp,
state=Closed, dataSize=597, limit=104857600}
19/05/03 11:07:17 DEBUG s3a.S3ADataBlocks: block[1]: closeBlock()
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Upload complete to
test-warehouse/test_decimal_encodings_db2969ca.db/decimal_stored_as_int32/decimal_stored_as_int32.parquet
by WriteOperationHelper {bucket=impala-test-uswest2-2}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Closing FileBlock{index=1,
destFile=/tmp/hadoop-jenkins/s3a/s3ablock-0001-6698639452552670061.tmp,
state=Closed, dataSize=597, limit=104857600}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Closing
org.apache.hadoop.fs.s3a.S3ADataBlocks$DiskBlockFactory@29caf222
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Statistics:
OutputStreamStatistics{blocksSubmitted=1, blocksInQueue=1, blocksActive=0,
blockUploadsCompleted=0, blockUploadsFailed=0, bytesPendingUpload=0,
bytesUploaded=597, blocksAllocated=1, blocksReleased=1,
blocksActivelyAllocated=0, exceptionsInMultipartFinalize=0, transferDuration=0
ms, queueDuration=0 ms, averageQueueTime=0 ms, totalUploadDuration=0 ms,
effectiveBandwidth=0.0 bytes/s}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Closing
OutputStreamStatistics{blocksSubmitted=1, blocksInQueue=1, blocksActive=0,
blockUploadsCompleted=0, blockUploadsFailed=0, bytesPendingUpload=0,
bytesUploaded=597, blocksAllocated=1, blocksReleased=1,
blocksActivelyAllocated=0, exceptionsInMultipartFinalize=0, transferDuration=0
ms, queueDuration=0 ms, averageQueueTime=0 ms, totalUploadDuration=0 ms,
effectiveBandwidth=0.0 bytes/s}
19/05/03 11:07:17 DEBUG s3a.S3ABlockOutputStream: Ignoring close() as stream is
already closed
19/05/03 11:07:17 DEBUG s3a.S3AFileSystem: Filesystem
s3a://impala-test-uswest2-2 is closed
19/05/03 11:07:17 INFO impl.MetricsSystemImpl: Stopping s3a-file-system metrics
system...
19/05/03 11:07:17 INFO impl.MetricsSystemImpl: s3a-file-system metrics system
stopped.
19/05/03 11:07:17 INFO impl.MetricsSystemImpl: s3a-file-system metrics system
shutdown complete.
-- executing against localhost:21000
create table test_decimal_encodings_db2969ca.decimal_stored_as_int32 like
parquet
"/test-warehouse/test_decimal_encodings_db2969ca.db/decimal_stored_as_int32/decimal_stored_as_int32.parquet"
stored as parquet;
{noformat}
> test_timestamp_out_of_range fails with NoSuchObjectException:
> test_timestamp_out_of_range_dc37915d on S3
> --------------------------------------------------------------------------------------------------------
>
> Key: IMPALA-8500
> URL: https://issues.apache.org/jira/browse/IMPALA-8500
> Project: IMPALA
> Issue Type: Bug
> Components: Infrastructure
> Reporter: Tim Armstrong
> Assignee: Joe McDonnell
> Priority: Critical
> Labels: broken-build
>
> I'm not sure what to make of this. Maybe you have an idea:
> {noformat}
> Error Message
> test setup failure
> Stacktrace
> conftest.py:319: in cleanup
> {'sync_ddl': sync_ddl})
> common/impala_test_suite.py:620: in wrapper
> return function(*args, **kwargs)
> common/impala_test_suite.py:628: in execute_query_expect_success
> result = cls.__execute_query(impalad_client, query, query_options, user)
> common/impala_test_suite.py:722: in __execute_query
> return impalad_client.execute(query, user=user)
> common/impala_connection.py:180: in execute
> return self.__beeswax_client.execute(sql_stmt, user=user)
> beeswax/impala_beeswax.py:187: in execute
> handle = self.__execute_query(query_string.strip(), user=user)
> beeswax/impala_beeswax.py:362: in __execute_query
> handle = self.execute_query_async(query_string, user=user)
> beeswax/impala_beeswax.py:356: in execute_query_async
> handle = self.__do_rpc(lambda: self.imp_service.query(query,))
> beeswax/impala_beeswax.py:516: in __do_rpc
> raise ImpalaBeeswaxException(self.__build_error_message(b), b)
> E ImpalaBeeswaxException: ImpalaBeeswaxException:
> E INNER EXCEPTION: <class 'beeswaxd.ttypes.BeeswaxException'>
> E MESSAGE: ImpalaRuntimeException: Error making 'dropDatabase' RPC to Hive
> Metastore:
> E CAUSED BY: NoSuchObjectException: test_timestamp_out_of_range_dc37915d
> {noformat}
> It does look like it created the database and used it fine:
> {noformat}
> -- 2019-05-02 11:09:21,883 INFO MainThread: Started query
> e847e67be097c2a0:6c29d4f400000000
> SET
> client_identifier=query_test/test_scanners.py::TestParquet::()::test_timestamp_out_of_range[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action':None;'exec_single_node_rows_t;
> -- executing against localhost:21000
> use test_timestamp_out_of_range_dc37915d;
> -- 2019-05-02 11:09:22,266 INFO MainThread: Started query
> ab498d07351289b2:03f6198300000000
> SET
> client_identifier=query_test/test_scanners.py::TestParquet::()::test_timestamp_out_of_range[protocol:beeswax|exec_option:{'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':False;'abort_on_error':1;'debug_action':None;'exec_single_node_rows_t;
> SET batch_size=0;
> SET num_nodes=0;
> SET disable_codegen_rows_threshold=0;
> SET disable_codegen=False;
> SET abort_on_error=1;
> SET exec_single_node_rows_threshold=0;
> -- executing against localhost:21000
> SELECT * FROM out_of_range_timestamp;
> -- 2019-05-02 11:09:22,273 INFO MainThread: Started query
> be4886563821196f:871eb37f00000000
> -- executing against localhost:21000
> SELECT * FROM out_of_range_time_of_day;
> {noformat}
> Maybe some S3 consistency issue?
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]