See 
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/2798/display/redirect?page=changes>

Changes:

[noreply] split write rename and temp table deletion into to dofns with a 
shuffle


------------------------------------------
[...truncated 11.90 MB...]
  File "apache_beam/runners/worker/operations.py", line 572, in 
apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1526, in 
apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 636, in 
apache_beam.runners.common.SimpleInvoker.invoke_process
  File "apache_beam/runners/common.py", line 1621, in 
apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "apache_beam/runners/common.py", line 1736, in 
apache_beam.runners.common._OutputHandler._write_value_to_tag
  File "apache_beam/runners/worker/operations.py", line 263, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 1351, in 
apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 1354, in 
apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 572, in 
apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1547, in 
apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in 
apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in 
apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in 
apache_beam.runners.common._OutputHandler.handle_process_outputs
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py",
 line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 1022, in perform_load_job
    return self._insert_load_job(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py",
 line 298, in wrapper
    return fun(*args, **kwargs)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py",
 line 356, in Insert
    return self._RunMethod(
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError 
accessing 
<https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>:
 response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 
'application/json; charset=UTF-8', 'date': 'Mon, 22 Jan 2024 23:32:18 GMT', 
'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 
'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 
'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', 
'-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
> [while running 
> 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']

      Worker ID: beamapp-jenkins-012223285-01221528-xc4o-harness-ncnc
WARNING  
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 
2024-01-22T23:32:19.789Z: JOB_MESSAGE_WARNING: Unable to delete temp files: 
"gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0122232858-568015-zmwbf9ko.1705966138.568187/dax-tmp-2024-01-22_15_28_59-17156436973935848627-S56-1-19dcc1887875aa01/tmp-19dcc1887875aa3e@DAX.recordio."
WARNING  
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 
2024-01-22T23:32:19.790Z: JOB_MESSAGE_WARNING: Unable to delete temp files: 
"gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0122232858-568015-zmwbf9ko.1705966138.568187/dax-tmp-2024-01-22_15_28_59-17156436973935848627-S56-0-21823c6311841284/tmp-21823c6311841217@DAX.recordio."
WARNING  
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 
2024-01-22T23:32:19.799Z: JOB_MESSAGE_WARNING: Unable to delete temp files: 
"gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-0122232858-568015-zmwbf9ko.1705966138.568187/dax-tmp-2024-01-22_15_28_59-17156436973935848627-S56-2-2f1d67533a412e60/tmp-2f1d67533a412c19@DAX.recordio."
WARNING  
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:203 
2024-01-22T23:32:19.819Z: JOB_MESSAGE_WARNING: 
S56:WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda
 at 
bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write
 failed.
INFO     
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 
2024-01-22T23:32:19.842Z: JOB_MESSAGE_BASIC: Finished operation 
WriteWithMultipleDests/BigQueryBatchFileLoads/GroupFilesByTableDestinations/Read+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(PartitionFiles)/ParDo(PartitionFiles)+WriteWithMultipleDests/BigQueryBatchFileLoads/FlattenPartitions+WriteWithMultipleDests/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)+WriteWithMultipleDests/BigQueryBatchFileLoads/MapTuple(<lambda
 at 
bigquery_file_loads.py:1135>)+WriteWithMultipleDests/BigQueryBatchFileLoads/ParDo(UpdateDestinationSchema)+WriteWithMultipleDests/BigQueryBatchFileLoads/GroupByKey/Write
INFO     
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 
2024-01-22T23:32:20.208Z: JOB_MESSAGE_BASIC: Stopping worker pool...
INFO     
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:201 
2024-01-22T23:34:23.483Z: JOB_MESSAGE_BASIC: Worker pool stopped.
INFO     
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:151 Job 
2024-01-22_15_28_59-17156436973935848627 is in state JOB_STATE_FAILED
ERROR    
apache_beam.runners.dataflow.dataflow_runner:dataflow_runner.py:777 Console 
URL: 
https://console.cloud.google.com/dataflow/jobs/<RegionId>/2024-01-22_15_28_59-17156436973935848627?project=<ProjectId>
INFO     
apache_beam.io.gcp.bigquery_file_loads_test:bigquery_file_loads_test.py:1242 
Deleting dataset python_bq_file_loads_1705966135d935b4 in project 
apache-beam-testing
=============================== warnings summary 
===============================
apache_beam/io/gcp/bigquery.py:2603
  
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery.py>:2603:
 DeprecationWarning: invalid escape sequence \#
    """Read data from BigQuery.

apache_beam/io/gcp/bigquery_read_it_test.py::ReadTests::test_native_source
  
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:170:
 BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use 
ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_file_loads
apache_beam/io/gcp/bigquery_test.py::PubSubBigQueryIT::test_streaming_inserts
apache_beam/examples/complete/game/game_stats_it_test.py::GameStatsIT::test_game_stats_it
apache_beam/examples/complete/game/leader_board_it_test.py::LeaderBoardIT::test_leader_board_it
apache_beam/examples/complete/game/hourly_team_score_it_test.py::HourlyTeamScoreIT::test_hourly_team_score_it
  
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:63:
 PendingDeprecationWarning: Client.dataset is deprecated and will be removed in 
a future version. Use a string like 'my_project.my_dataset' or a 
cloud.google.bigquery.DatasetReference object, instead.
    dataset_ref = client.dataset(unique_dataset_name, project=project)

apache_beam/examples/cookbook/bigquery_tornadoes_it_test.py::BigqueryTornadoesIT::test_bigquery_tornadoes_it
  
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/tests/utils.py>:100:
 PendingDeprecationWarning: Client.dataset is deprecated and will be removed in 
a future version. Use a string like 'my_project.my_dataset' or a 
cloud.google.bigquery.DatasetReference object, instead.
    table_ref = client.dataset(dataset_id).table(table_id)

apache_beam/io/gcp/bigquery_read_it_test.py::ReadNewTypesTests::test_native_source
  
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/apache_beam/io/gcp/bigquery_read_it_test.py>:706:
 BeamDeprecationWarning: BigQuerySource is deprecated since 2.25.0. Use 
ReadFromBigQuery instead.
    beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))

-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html
- generated xml file: 
<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/pytest_postCommitIT-df-py39.xml>
 -
=========================== short test summary info 
============================
FAILED 
apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_batch_copy_jobs_with_no_input_schema
 - AssertionError: 
Expected: (<[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), 
(3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), 
(8,), (9,)])>]>)
     but: <[<BigqueryFullResultMatcher(Expected data is [(0,), (1,), (2,), 
(3,), (4,)])>, <BigqueryFullResultMatcher(Expected data is [(5,), (6,), (7,), 
(8,), (9,)])>]> was <DataflowPipelineResult 
2024-01-22_15_22_16-4883707003561850353 DONE>
FAILED 
apache_beam/io/gcp/bigquery_file_loads_test.py::BigQueryFileLoadsIT::test_multiple_destinations_transform
 - apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException: 
Dataflow pipeline failed. State: FAILED, Error:
Traceback (most recent call last):
  File "apache_beam/runners/common.py", line 1435, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in 
apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in 
apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in 
apache_beam.runners.common._OutputHandler.handle_process_outputs
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py",
 line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 1022, in perform_load_job
    return self._insert_load_job(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py",
 line 298, in wrapper
    return fun(*args, **kwargs)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py",
 line 356, in Insert
    return self._RunMethod(
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing 
<https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>:
 response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 
'application/json; charset=UTF-8', 'date': 'Mon, 22 Jan 2024 23:32:18 GMT', 
'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 
'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 
'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', 
'-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
>

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
 line 311, in _execute
    response = task()
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
 line 386, in <lambda>
    lambda: self.create_worker().do_instruction(request), request)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
 line 650, in do_instruction
    return getattr(self, request_type)(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/sdk_worker.py",
 line 688, in process_bundle
    bundle_processor.process_bundle(instruction_id))
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/bundle_processor.py",
 line 1113, in process_bundle
    input_op_by_transform_id[element.transform_id].process_encoded(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/runners/worker/bundle_processor.py",
 line 237, in process_encoded
    self.output(decoded_value)
  File "apache_beam/runners/worker/operations.py", line 570, in 
apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 572, in 
apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1526, in 
apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 636, in 
apache_beam.runners.common.SimpleInvoker.invoke_process
  File "apache_beam/runners/common.py", line 1621, in 
apache_beam.runners.common._OutputHandler.handle_process_outputs
  File "apache_beam/runners/common.py", line 1736, in 
apache_beam.runners.common._OutputHandler._write_value_to_tag
  File "apache_beam/runners/worker/operations.py", line 263, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 1351, in 
apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 1354, in 
apache_beam.runners.worker.operations.FlattenOperation.process
  File "apache_beam/runners/worker/operations.py", line 572, in 
apache_beam.runners.worker.operations.Operation.output
  File "apache_beam/runners/worker/operations.py", line 263, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 266, in 
apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive
  File "apache_beam/runners/worker/operations.py", line 953, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/worker/operations.py", line 954, in 
apache_beam.runners.worker.operations.DoOperation.process
  File "apache_beam/runners/common.py", line 1437, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 1547, in 
apache_beam.runners.common.DoFnRunner._reraise_augmented
  File "apache_beam/runners/common.py", line 1435, in 
apache_beam.runners.common.DoFnRunner.process
  File "apache_beam/runners/common.py", line 851, in 
apache_beam.runners.common.PerWindowInvoker.invoke_process
  File "apache_beam/runners/common.py", line 995, in 
apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window
  File "apache_beam/runners/common.py", line 1611, in 
apache_beam.runners.common._OutputHandler.handle_process_outputs
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_file_loads.py",
 line 750, in process
    job_reference = self.bq_wrapper.perform_load_job(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 1022, in perform_load_job
    return self._insert_load_job(
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/utils/retry.py",
 line 298, in wrapper
    return fun(*args, **kwargs)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 559, in _insert_load_job
    return self._start_job(request, stream=source_stream).jobReference
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/bigquery_tools.py",
 line 578, in _start_job
    response = self.client.jobs.Insert(request, upload=upload)
  File 
"/opt/apache/beam-venv/beam-venv-worker-sdk-0-0/lib/python3.9/site-packages/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_client.py",
 line 356, in Insert
    return self._RunMethod(
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
  File "/usr/local/lib/python3.9/site-packages/apitools/base/py/base_api.py", 
line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
RuntimeError: apitools.base.py.exceptions.HttpBadRequestError: HttpError 
accessing 
<https://bigquery.googleapis.com/bigquery/v2/projects/apache-beam-testing/jobs?alt=json>:
 response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 
'application/json; charset=UTF-8', 'date': 'Mon, 22 Jan 2024 23:32:18 GMT', 
'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 
'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 
'transfer-encoding': 'chunked', 'status': '400', 'content-length': '318', 
'-content-encoding': 'gzip'}>, content <{
  "error": {
    "code": 400,
    "message": "Load configuration must specify at least one source URI",
    "errors": [
      {
        "message": "Load configuration must specify at least one source URI",
        "domain": "global",
        "reason": "invalid"
      }
    ],
    "status": "INVALID_ARGUMENT"
  }
}
> [while running 
> 'WriteWithMultipleDestsFreely/BigQueryBatchFileLoads/TriggerLoadJobsWithTempTables/ParDo(TriggerLoadJobs)-ptransform-71']
====== 2 failed, 89 passed, 51 skipped, 
9 warnings in 5902.66s (1:38:22) =======

> Task :sdks:python:test-suites:dataflow:py39:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script 
'<https://ci-beam.apache.org/job/beam_PostCommit_Python39/ws/src/sdks/python/test-suites/dataflow/common.gradle'>
 line: 139

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py39:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Get more help at https://help.gradle.org.

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 9.0.

You can use '--warning-mode all' to show the individual deprecation warnings 
and determine if they come from your own scripts or plugins.

For more on this, please refer to 
https://docs.gradle.org/8.4/userguide/command_line_interface.html#sec:command_line_warnings
 in the Gradle documentation.

BUILD FAILED in 1h 44m 45s
222 actionable tasks: 160 executed, 58 from cache, 4 up-to-date

Publishing build scan...
https://ge.apache.org/s/4nuoqfulyoffg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Reply via email to