This is an automated email from the ASF dual-hosted git repository. joemcdonnell pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/impala.git
commit 6f949718f57722ddc6459a2170b92f56e403b02d Author: Peter Rozsa <[email protected]> AuthorDate: Thu Mar 20 20:07:21 2025 +0100 IMPALA-13880: Skip Iceberg interop tests with Hive if the default filesystem is not HDFS This patch adds skip annotations to the tests added by IMPALA-13611 if the targeted filesystem is not HDFS. 'run-all.sh' starts just the Hive Metastore in case the targeted filesystem is not HDFS. Test: - manually validated that the test is skipped if the default filesystem points to an Ozone instance. Change-Id: Ic872f31f0d4b61fbe0479a32c8a6c52fde6b293e Reviewed-on: http://gerrit.cloudera.org:8080/22651 Reviewed-by: Impala Public Jenkins <[email protected]> Tested-by: Impala Public Jenkins <[email protected]> --- tests/stress/test_update_stress.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tests/stress/test_update_stress.py b/tests/stress/test_update_stress.py index 7dc38376f..35b6b4135 100644 --- a/tests/stress/test_update_stress.py +++ b/tests/stress/test_update_stress.py @@ -25,6 +25,7 @@ from subprocess import check_output from tests.common.impala_test_suite import ImpalaTestSuite from tests.common.parametrize import UniqueDatabase +from tests.common.skip import SkipIfFS from tests.common.test_dimensions import create_exec_option_dimension from tests.stress.stress_util import run_tasks, Task from tests.util.filesystem_utils import FILESYSTEM_PREFIX, IS_HDFS @@ -352,6 +353,7 @@ class TestIcebergConcurrentOperations(ImpalaTestSuite): assert len(data_files_in_tbl) == len(data_files_on_fs_rows) assert set(data_files_on_fs) == set(data_files_in_tbl) + @SkipIfFS.hive @pytest.mark.execute_serially @UniqueDatabase.parametrize(sync_ddl=True) def test_iceberg_impala_deletes_and_hive_updates(self, unique_database): @@ -380,6 +382,7 @@ class TestIcebergConcurrentOperations(ImpalaTestSuite): result = self.client.execute("select count(*) from {}".format(tbl_name)) assert result.data == ['0'] + @SkipIfFS.hive @pytest.mark.execute_serially @UniqueDatabase.parametrize(sync_ddl=True) def test_iceberg_impala_updates_and_hive_deletes(self, unique_database):
