See <https://builds.apache.org/job/beam_PostCommit_Python35/2086/display/redirect?page=changes>
Changes: [daniel.o.programmer] [BEAM-3301] Fix another bug in DoFn validation, in exec. [github] Update Go Protos (#11230) ------------------------------------------ [...truncated 10.76 MB...] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in get_operation pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in <dictcomp> pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 778, in <listcomp> tag: [get_operation(op) for op in pcoll_consumers[pcoll_id]] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in get_operation pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in <dictcomp> pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 778, in <listcomp> tag: [get_operation(op) for op in pcoll_consumers[pcoll_id]] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in get_operation pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in <dictcomp> pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 778, in <listcomp> tag: [get_operation(op) for op in pcoll_consumers[pcoll_id]] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 783, in get_operation transform_id, transform_consumers) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1115, in create_operation return creator(self, transform_id, transform_proto, payload, consumers) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1439, in create_par_do parameter) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1475, in _create_pardo_operation dofn_data = pickler.loads(serialized_fn) File "/usr/local/lib/python3.5/site-packages/apache_beam/internal/pickler.py", line 287, in loads return dill.loads(s) File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 275, in loads return load(file, ignore, **kwds) File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 270, in load return Unpickler(file, ignore=ignore, **kwds).load() File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 472, in load obj = StockUnpickler.load(self) File "stringsource", line 17, in apache_beam.utils.windowed_value._IntervalWindowBase.__setstate_cython__ TypeError: Expected tuple, got dict java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895) org.apache.beam.sdk.util.MoreFutures.get(MoreFutures.java:57) org.apache.beam.runners.dataflow.worker.fn.control.RegisterAndProcessBundleOperation.finish(RegisterAndProcessBundleOperation.java:332) org.apache.beam.runners.dataflow.worker.util.common.worker.MapTaskExecutor.execute(MapTaskExecutor.java:85) org.apache.beam.runners.dataflow.worker.fn.control.BeamFnMapTaskExecutor.execute(BeamFnMapTaskExecutor.java:125) org.apache.beam.runners.dataflow.worker.StreamingDataflowWorker.process(StreamingDataflowWorker.java:1358) org.apache.beam.runners.dataflow.worker.StreamingDataflowWorker.access$1100(StreamingDataflowWorker.java:153) org.apache.beam.runners.dataflow.worker.StreamingDataflowWorker$7.run(StreamingDataflowWorker.java:1081) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:748) Caused by: java.lang.RuntimeException: Error received from SDK harness for instruction -8119: Traceback (most recent call last): File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/sdk_worker.py", line 313, in get processor = self.cached_bundle_processors[bundle_descriptor_id].pop() IndexError: pop from empty list During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.5/site-packages/apache_beam/internal/pickler.py", line 283, in loads return dill.loads(s) File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 275, in loads return load(file, ignore, **kwds) File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 270, in load return Unpickler(file, ignore=ignore, **kwds).load() File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 472, in load obj = StockUnpickler.load(self) File "stringsource", line 17, in apache_beam.utils.windowed_value._IntervalWindowBase.__setstate_cython__ TypeError: Expected tuple, got dict During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/sdk_worker.py", line 190, in _execute response = task() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/sdk_worker.py", line 247, in <lambda> lambda: self.create_worker().do_instruction(request), request) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/sdk_worker.py", line 416, in do_instruction getattr(request, request_type), request.instruction_id) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/sdk_worker.py", line 445, in process_bundle instruction_id, request.process_bundle_descriptor_id) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/sdk_worker.py", line 319, in get self.data_channel_factory) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 744, in __init__ self.ops = self.create_execution_tree(self.process_bundle_descriptor) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 797, in create_execution_tree descriptor.transforms, key=topological_height, reverse=True) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 796, in <listcomp> (transform_id, get_operation(transform_id)) for transform_id in sorted( File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in get_operation pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in <dictcomp> pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 778, in <listcomp> tag: [get_operation(op) for op in pcoll_consumers[pcoll_id]] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in get_operation pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in <dictcomp> pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 778, in <listcomp> tag: [get_operation(op) for op in pcoll_consumers[pcoll_id]] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in get_operation pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 780, in <dictcomp> pcoll_id in descriptor.transforms[transform_id].outputs.items() File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 778, in <listcomp> tag: [get_operation(op) for op in pcoll_consumers[pcoll_id]] File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 705, in wrapper result = cache[args] = func(*args) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 783, in get_operation transform_id, transform_consumers) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1115, in create_operation return creator(self, transform_id, transform_proto, payload, consumers) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1439, in create_par_do parameter) File "/usr/local/lib/python3.5/site-packages/apache_beam/runners/worker/bundle_processor.py", line 1475, in _create_pardo_operation dofn_data = pickler.loads(serialized_fn) File "/usr/local/lib/python3.5/site-packages/apache_beam/internal/pickler.py", line 287, in loads return dill.loads(s) File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 275, in loads return load(file, ignore, **kwds) File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 270, in load return Unpickler(file, ignore=ignore, **kwds).load() File "/usr/local/lib/python3.5/site-packages/dill/_dill.py", line 472, in load obj = StockUnpickler.load(self) File "stringsource", line 17, in apache_beam.utils.windowed_value._IntervalWindowBase.__setstate_cython__ TypeError: Expected tuple, got dict org.apache.beam.runners.fnexecution.control.FnApiControlClient$ResponseStreamObserver.onNext(FnApiControlClient.java:160) org.apache.beam.runners.fnexecution.control.FnApiControlClient$ResponseStreamObserver.onNext(FnApiControlClient.java:140) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onMessage(ServerCalls.java:251) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onMessage(ForwardingServerCallListener.java:33) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onMessage(Contexts.java:76) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailableInternal(ServerCallImpl.java:309) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailable(ServerCallImpl.java:292) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1MessagesAvailable.runInContext(ServerImpl.java:782) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:748) apache_beam.runners.dataflow.dataflow_runner: WARNING: Timing out on waiting for job 2020-03-25_16_38_37-472429306662736637 after 364 seconds google.auth.transport._http_client: DEBUG: Making request: GET http://169.254.169.254 google.auth.transport._http_client: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/project/project-id google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): metadata.google.internal:80 urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 200 144 google.auth.transport.requests: DEBUG: Making request: GET http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET /computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token HTTP/1.1" 200 192 apache_beam.io.gcp.tests.pubsub_matcher: ERROR: Timeout after 400 sec. Received 0 messages from projects/apache-beam-testing/subscriptions/wc_subscription_outputbb39aad2-bc23-4a87-a61a-5fd001eaeefe. --------------------- >> end captured logging << --------------------- Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_48-844174954073284068?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_53_23-4628773915602496926?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_01_07-14148289083555684019?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_09_58-15592540348060554373?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_19_06-11079902910266625984?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_27_00-3278452890034318311?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_42-4982831091953491509?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_00_14-8884383114514098106?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_08_51-769566995373403474?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_17_11-9890883872716525265?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_25_51-284444123831101249?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_34_48-12039560138292564564?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_38-17391926555312974431?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_50_29-7133254401544223073?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_59_45-13557879115795838945?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_09_24-885650266532089770?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_17_45-10475435332895636970?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_26_53-17408119981080631058?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_34_27-6664095784552409680?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_34-1766294107877446980?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_00_49-13239683766895321358?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_09_47-8391179991663334721?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_18_31-18202447000183261072?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_27_01-7656408119704187603?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_35-2067996052483921726?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_47_35-10841763644447842040?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_56_24-3337170592449756118?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_04_48-11792676447498801803?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_14_14-7834481042099047374?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_22_54-11180293176343691960?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_32_04-15456685068624882530?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_33-14658266105293391494?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_46_49-6207233854397933291?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_55_04-7268046907362392645?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_05_05-18271869142908061774?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_14_50-17233503296865448422?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_24_09-5385350852161911130?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_31_19-13339915671408059339?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_38_56-6177905380032110382?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_37-472429306662736637?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_54_18-15948541047417103553?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_02_09-12895656947073385600?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_10_28-5568783400854086980?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_18_03-3730929328200204739?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_27_37-14151526416589462280?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_38_35-15364978667026021608?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_16_49_14-8465864964795328842?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_01_25-4650081305015702949?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_09_54-18122057255422467948?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-03-25_17_26_44-11803103153612544942?project=apache-beam-testing ---------------------------------------------------------------------- XML: nosetests-postCommitIT-df-py35.xml ---------------------------------------------------------------------- XML: <https://builds.apache.org/job/beam_PostCommit_Python35/ws/src/sdks/python/nosetests.xml> ---------------------------------------------------------------------- Ran 59 tests in 4139.692s FAILED (SKIP=8, failures=1) > Task :sdks:python:test-suites:dataflow:py35:postCommitIT FAILED FAILURE: Build failed with an exception. * Where: Build file '<https://builds.apache.org/job/beam_PostCommit_Python35/ws/src/sdks/python/test-suites/dataflow/py35/build.gradle'> line: 56 * What went wrong: Execution failed for task ':sdks:python:test-suites:dataflow:py35:postCommitIT'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 1h 10m 26s 86 actionable tasks: 64 executed, 22 from cache Publishing build scan... https://gradle.com/s/hlhtttyfjiex4 Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure --------------------------------------------------------------------- To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org For additional commands, e-mail: builds-h...@beam.apache.org