See
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/2805/display/redirect>
Changes:
------------------------------------------
[...truncated 12.37 MB...]
File "apache_beam/runners/worker/operations.py", line 953, in
apache_beam.runners.worker.operations.DoOperation.process
File "apache_beam/runners/worker/operations.py", line 954, in
apache_beam.runners.worker.operations.DoOperation.process
File "apache_beam/runners/common.py", line 1437, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 1547, in
apache_beam.runners.common.DoFnRunner._reraise_augmented
File "apache_beam/runners/common.py", line 1435, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 851, in
apache_beam.runners.common.PerWindowInvoker.invoke_process
File "apache_beam/runners/common.py", line 995, in
apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
File "apache_beam/runners/common.py", line 1611, in
apache_beam.runners.common._OutputHandler.handle_process_outputs
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py",
line 750, in process
job_reference = self.bq_wrapper.perform_load_job(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 1022, in perform_load_job
return self._insert_load_job(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py",
line 298, in wrapper
return fun(*args, **kwargs)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 559, in _insert_load_job
return self._start_job(request, stream=source_stream).jobReference
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 578, in _start_job
response = self.client.jobs.Insert(request, upload=upload)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py",
line 356, in Insert
return self._RunMethod(
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 731, in _RunMethod
return self.ProcessHttpResponse(method_config, http_response, request)
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 737, in ProcessHttpResponse
self.__ProcessHttpResponse(method_config, http_response, request))
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 603, in __ProcessHttpResponse
raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError
accessing
<https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>:
response: <{'vary': 'Origin, X-Origin, Referer', 'content-type':
'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 00:29:10 GMT',
'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0',
'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff',
'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318',
'-content-encoding': 'gzip'}>, content <{
"error": {
"code": 400,
"message": "Load configuration must specify at least one source URI",
"errors": [
{
"message": "Load configuration must specify at least one source URI",
"domain": "global",
"reason": "invalid"
}
],
"status": "INVALID_ARGUMENT"
}
}
> [while running
> 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']
Worker ID: beamapp-jenkins-012500254-01241625-uy51-harness-w2wg
[33mWARNING [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203
2024-01-25T00:29:11.982Z: JOB_MESSAGE_WARNING: Unable to delete temp files:
"gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0125002540-066487-zmwbf9ko.1706142340.066646/dax-tmp-2024-01-24_16_25_40-3576056548624936316-S56-2-ba1d8146837ee981/[email protected]."
[33mWARNING [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203
2024-01-25T00:29:11.983Z: JOB_MESSAGE_WARNING: Unable to delete temp files:
"gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0125002540-066487-zmwbf9ko.1706142340.066646/dax-tmp-2024-01-24_16_25_40-3576056548624936316-S56-1-a88f64f614a2a437/[email protected]."
[33mWARNING [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203
2024-01-25T00:29:11.987Z: JOB_MESSAGE_WARNING: Unable to delete temp files:
"gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0125002540-066487-zmwbf9ko.1706142340.066646/dax-tmp-2024-01-24_16_25_40-3576056548624936316-S56-0-aa9e464cd130d213/[email protected]."
[33mWARNING [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203
2024-01-25T00:29:12.006Z: JOB_MESSAGE_WARNING:
S56:WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda
at
bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write
failed.
[32mINFO [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201
2024-01-25T00:29:12.031Z: JOB_MESSAGE_BASIC: Finished operation
WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda
at
bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write
[32mINFO [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201
2024-01-25T00:29:12.427Z: JOB_MESSAGE_BASIC: Stopping worker pool...
[32mINFO [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201
2024-01-25T00:31:37.231Z: JOB_MESSAGE_BASIC: Worker pool stopped.
[32mINFO [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:151 Job
2024-01-24_16_25_40-3576056548624936316 is in state JOB_STATE_FAILED
[1m[31mERROR [0m
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:777 Console
URL:
https://console.cloud.google.com/dataflow/jobs/<RegionId>/2024-01-24_16_25_40-3576056548624936316?project=<ProjectId>
[32mINFO [0m
apache_beam.io.gcp.bigquery_file_loads_test:bigquery_file_loads_test.py:1242
Deleting dataset python_bq_file_loads_170614233681bd22 in project
apache-beam-testing
[33m=============================== warnings summary
===============================[0m
apache_beam/io/gcp/bigquery.py:2634
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:2634:
DeprecationWarning: invalid escape sequence \#
"""Read data from BigQuery.
apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170:
BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use
ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63:
PendingDeprecationWarning: Client.dataset is deprecated and will be removed in
a future version. Use a string like 'my_project.my_dataset' or a
cloud.google.bigquery.DatasetReference object, instead.
dataset_ref = client.dataset(unique_dataset_name, project=project)
apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100:
PendingDeprecationWarning: Client.dataset is deprecated and will be removed in
a future version. Use a string like 'my_project.my_dataset' or a
cloud.google.bigquery.DatasetReference object, instead.
table_ref = client.dataset(dataset_id).table(table_id)
apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706:
BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use
ReadFromBigQuery instead.
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file:
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/pytest_postCommitIT-df-py39.xml>
-
[36m[1m=========================== short test summary info
============================[0m
[31mFAILED[0m
apache_beam/io/gcp/bigquery_file_loads_test.py::[1mBigQueryFileLoadsIT::test_batch_copy_jobs_with_no_input_schema[0m
- AssertionError:
Expected: (<[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,),
(3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,),
(8,), (9,)])>]>)
but: <[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,),
(3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,),
(8,), (9,)])>]> was <DataflowPipelineResult
2024-01-24_16_18_12-12488496925338720336 DONE>
[31mFAILED[0m
apache_beam/io/gcp/bigquery_file_loads_test.py::[1mBigQueryFileLoadsIT::test_multiple_destinations_transform[0m
- apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException:
Dataflow pipeline failed. State: FAILED, Error:
Traceback (most recent call last):
File "apache_beam/runners/common.py", line 1435, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 851, in
apache_beam.runners.common.PerWindowInvoker.invoke_process
File "apache_beam/runners/common.py", line 995, in
apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
File "apache_beam/runners/common.py", line 1611, in
apache_beam.runners.common._OutputHandler.handle_process_outputs
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py",
line 750, in process
job_reference = self.bq_wrapper.perform_load_job(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 1022, in perform_load_job
return self._insert_load_job(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py",
line 298, in wrapper
return fun(*args, **kwargs)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 559, in _insert_load_job
return self._start_job(request, stream=source_stream).jobReference
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 578, in _start_job
response = self.client.jobs.Insert(request, upload=upload)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py",
line 356, in Insert
return self._RunMethod(
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 731, in _RunMethod
return self.ProcessHttpResponse(method_config, http_response, request)
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 737, in ProcessHttpResponse
self.__ProcessHttpResponse(method_config, http_response, request))
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 603, in __ProcessHttpResponse
raise exceptions.HttpError.FromResponse(
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing
<https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>:
response: <{'vary': 'Origin, X-Origin, Referer', 'content-type':
'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 00:29:10 GMT',
'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0',
'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff',
'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318',
'-content-encoding': 'gzip'}>, content <{
"error": {
"code": 400,
"message": "Load configuration must specify at least one source URI",
"errors": [
{
"message": "Load configuration must specify at least one source URI",
"domain": "global",
"reason": "invalid"
}
],
"status": "INVALID_ARGUMENT"
}
}
>
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
line 311, in _execute
response = task()
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
line 386, in <lambda>
lambda: self.create_worker().do_instruction(request), request)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
line 650, in do_instruction
return getattr(self, request_type)(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
line 688, in process_bundle
bundle_processor.process_bundle(instruction_id))
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/bundle_processor.py",
line 1113, in process_bundle
input_op_by_transform_id[element.transform_id].process_encoded(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/bundle_processor.py",
line 237, in process_encoded
self.output(decoded_value)
File "apache_beam/runners/worker/operations.py", line 570, in
apache_beam.runners.worker.operations.Operation.output
File "apache_beam/runners/worker/operations.py", line 572, in
apache_beam.runners.worker.operations.Operation.output
File "apache_beam/runners/worker/operations.py", line 263, in
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
File "apache_beam/runners/worker/operations.py", line 266, in
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
File "apache_beam/runners/worker/operations.py", line 953, in
apache_beam.runners.worker.operations.DoOperation.process
File "apache_beam/runners/worker/operations.py", line 954, in
apache_beam.runners.worker.operations.DoOperation.process
File "apache_beam/runners/common.py", line 1437, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 1526, in
apache_beam.runners.common.DoFnRunner._reraise_augmented
File "apache_beam/runners/common.py", line 1435, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 636, in
apache_beam.runners.common.SimpleInvoker.invoke_process
File "apache_beam/runners/common.py", line 1621, in
apache_beam.runners.common._OutputHandler.handle_process_outputs
File "apache_beam/runners/common.py", line 1736, in
apache_beam.runners.common._OutputHandler._write_value_to_tag
File "apache_beam/runners/worker/operations.py", line 263, in
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
File "apache_beam/runners/worker/operations.py", line 266, in
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
File "apache_beam/runners/worker/operations.py", line 1351, in
apache_beam.runners.worker.operations.FlattenOperation.process
File "apache_beam/runners/worker/operations.py", line 1354, in
apache_beam.runners.worker.operations.FlattenOperation.process
File "apache_beam/runners/worker/operations.py", line 572, in
apache_beam.runners.worker.operations.Operation.output
File "apache_beam/runners/worker/operations.py", line 263, in
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
File "apache_beam/runners/worker/operations.py", line 266, in
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
File "apache_beam/runners/worker/operations.py", line 953, in
apache_beam.runners.worker.operations.DoOperation.process
File "apache_beam/runners/worker/operations.py", line 954, in
apache_beam.runners.worker.operations.DoOperation.process
File "apache_beam/runners/common.py", line 1437, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 1547, in
apache_beam.runners.common.DoFnRunner._reraise_augmented
File "apache_beam/runners/common.py", line 1435, in
apache_beam.runners.common.DoFnRunner.process
File "apache_beam/runners/common.py", line 851, in
apache_beam.runners.common.PerWindowInvoker.invoke_process
File "apache_beam/runners/common.py", line 995, in
apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
File "apache_beam/runners/common.py", line 1611, in
apache_beam.runners.common._OutputHandler.handle_process_outputs
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py",
line 750, in process
job_reference = self.bq_wrapper.perform_load_job(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 1022, in perform_load_job
return self._insert_load_job(
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py",
line 298, in wrapper
return fun(*args, **kwargs)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 559, in _insert_load_job
return self._start_job(request, stream=source_stream).jobReference
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
line 578, in _start_job
response = self.client.jobs.Insert(request, upload=upload)
File
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py",
line 356, in Insert
return self._RunMethod(
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 731, in _RunMethod
return self.ProcessHttpResponse(method_config, http_response, request)
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 737, in ProcessHttpResponse
self.__ProcessHttpResponse(method_config, http_response, request))
File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py",
line 603, in __ProcessHttpResponse
raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError
accessing
<https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>:
response: <{'vary': 'Origin, X-Origin, Referer', 'content-type':
'application/json; charset=UTF-8', 'date': 'Thu, 25 Jan 2024 00:29:10 GMT',
'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0',
'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff',
'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318',
'-content-encoding': 'gzip'}>, content <{
"error": {
"code": 400,
"message": "Load configuration must specify at least one source URI",
"errors": [
{
"message": "Load configuration must specify at least one source URI",
"domain": "global",
"reason": "invalid"
}
],
"status": "INVALID_ARGUMENT"
}
}
> [while running
> 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']
[31m====== [31m[1m2 failed[0m, [32m90 passed[0m, [33m51 skipped[0m,
[33m9 warnings[0m[31m in 5741.10s (1:35:41)[0m[31m =======[0m
> Task :sdks:python:test-suites:dataflow:py39:postCommitIT FAILED
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script
'<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/test-suites/direct/common.gradle'>
line: 52
* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py39:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script
'<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/test-suites/dataflow/common.gradle'>
line: 139
* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py39:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.
==============================================================================
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 9.0.
You can use '--warning-mode all' to show the individual deprecation warnings
and determine if they come from your own scripts or plugins.
For more on this, please refer to
https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings
in the Gradle documentation.
BUILD FAILED in 1h 42m 5s
222 actionable tasks: 158 executed, 60 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/3drzg7ginkvdq
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]