This is an automated email from the ASF dual-hosted git repository. pabloem pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/beam.git
The following commit(s) were added to refs/heads/master by this push: new 710b13d Reactivating test while preventing timeouts. new 2ec3b04 Merge pull request #10299 from [BEAM-8842] Reactivating BQIO Py test while preventing timeouts. 710b13d is described below commit 710b13d08ba0de4dd0428a7fc4da0743364fbd6c Author: pabloem <pabl...@apache.org> AuthorDate: Thu Dec 5 12:45:12 2019 -0800 Reactivating test while preventing timeouts. --- sdks/python/apache_beam/io/gcp/bigquery_test.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/sdks/python/apache_beam/io/gcp/bigquery_test.py b/sdks/python/apache_beam/io/gcp/bigquery_test.py index 505a683..ac62774 100644 --- a/sdks/python/apache_beam/io/gcp/bigquery_test.py +++ b/sdks/python/apache_beam/io/gcp/bigquery_test.py @@ -43,6 +43,7 @@ from apache_beam.io.gcp.bigquery import WriteToBigQuery from apache_beam.io.gcp.bigquery import _StreamToBigQuery from apache_beam.io.gcp.bigquery_file_loads_test import _ELEMENTS from apache_beam.io.gcp.bigquery_tools import JSON_COMPLIANCE_ERROR +from apache_beam.io.gcp.bigquery_tools import RetryStrategy from apache_beam.io.gcp.internal.clients import bigquery from apache_beam.io.gcp.pubsub import ReadFromPubSub from apache_beam.io.gcp.tests import utils @@ -647,7 +648,6 @@ class BigQueryStreamingInsertTransformIntegrationTests(unittest.TestCase): method='FILE_LOADS')) @attr('IT') - @unittest.skip('BEAM-8842: Disabled due to reliance on old retry behavior.') def test_multiple_destinations_transform(self): streaming = self.test_pipeline.options.view_as(StandardOptions).streaming if streaming and isinstance(self.test_pipeline.runner, TestDataflowRunner): @@ -735,6 +735,7 @@ class BigQueryStreamingInsertTransformIntegrationTests(unittest.TestCase): table_side_inputs=(table_record_pcv,), schema=lambda dest, table_map: table_map.get(dest, None), schema_side_inputs=(schema_table_pcv,), + insert_retry_strategy=RetryStrategy.RETRY_ON_TRANSIENT_ERROR, method='STREAMING_INSERTS')) assert_that(r[beam.io.gcp.bigquery.BigQueryWriteFn.FAILED_ROWS],