See 
<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/6373/display/redirect?page=changes>

Changes:

[Kiley Sok] Update beam-master version

[noreply] Fix example checker validation for incomplete multifile examples

[noreply] Remove datastore intersphinx_dependency to fix docs job. (#25000)

[noreply] Use CC-BY image for custom remote inference notebook (#25004)


------------------------------------------
[...truncated 1.18 MB...]
      capabilities: "beam:protocol:control_request_elements_embedding:v1"
      capabilities: "beam:protocol:state_caching:v1"
      capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.45.0.dev"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:transform:to_string:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nL/tmp/artifacts/icedtea-sound-5xFkOD7aUmX8Jiw7lPqhn75zkNGCyUOfOuqi2rClWKM.jar\022@e71164383eda5265fc262c3b94faa19fbe7390d182c9439f3aeaa2dab0a558a3"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: 
"\n=icedtea-sound-5xFkOD7aUmX8Jiw7lPqhn75zkNGCyUOfOuqi2rClWKM.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nF/tmp/artifacts/jaccess-Rlp7GfuSFOBCBGdO8af2XxrTz5LqaQqwSMAXumdbHN0.jar\022@465a7b19fb9214e04204674ef1a7f65f1ad3cf92ea690ab048c017ba675b1cdd"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: 
"\n7jaccess-Rlp7GfuSFOBCBGdO8af2XxrTz5LqaQqwSMAXumdbHN0.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nI/tmp/artifacts/localedata-Nuzln2shadVmT-PppqaNc7GHBbUyMUb9pXsJTjRiAzs.jar\022@36ece59f6b2169d5664fe3e9a6a68d73b18705b5323146fda57b094e3462033b"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: 
"\n:localedata-Nuzln2shadVmT-PppqaNc7GHBbUyMUb9pXsJTjRiAzs.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nF/tmp/artifacts/nashorn-1iQRhJvU2nRCSB4ucg6f4b69TGiwE1yq0_LGG2MU5N0.jar\022@d62411849bd4da7442481e2e720e9fe1bebd4c68b0135caad3f2c61b6314e4dd"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: 
"\n7nashorn-1iQRhJvU2nRCSB4ucg6f4b69TGiwE1yq0_LGG2MU5N0.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nG/tmp/artifacts/cldrdata-MTK3mMuUzlWYlYkweGQNzI7dFpuLG0_8F8oxT0khQDc.jar\022@3132b798cb94ce559895893078640dcc8edd169b8b1b4ffc17ca314f49214037"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: 
"\n8cldrdata-MTK3mMuUzlWYlYkweGQNzI7dFpuLG0_8F8oxT0khQDc.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nD/tmp/artifacts/dnsns-YTF3f6B-OG4FmXLRf2ckHBi4VJnspBZTyIQCz-6Spwk.jar\022@6131777fa07e386e059972d17f67241c18b85499eca41653c88402cfee92a709"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: "\n5dnsns-YTF3f6B-OG4FmXLRf2ckHBi4VJnspBZTyIQCz-6Spwk.jar"
      >
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        type_payload: 
"\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.45.0-SNAPSHOT-MSDPg8oAZhgQiPvs5-IFzWvgTN96cSCujXkpH1GKmYg.jar\022@3120cf83ca0066181088fbece7e205cd6be04cdf7a7120ae8d79291f518a9988"
        role_urn: "beam:artifact:role:staging_to:v1"
        role_payload: 
"\ncbeam-sdks-java-io-expansion-service-2.45.0-SNAPSHOT-MSDPg8oAZhgQiPvs5-IFzWvgTN96cSCujXkpH1GKmYg.jar"
      >
    >
  >
  environments: <
    key: "go"
    value: <
      urn: "beam:env:docker:v1"
      payload: "\n\026apache/beam_go_sdk:dev"
      capabilities: "beam:protocol:progress_reporting:v1"
      capabilities: "beam:protocol:multi_core_bundle_processing:v1"
      capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
      capabilities: "beam:protocol:worker_status:v1"
      capabilities: "beam:protocol:monitoring_info_short_ids:v1"
      capabilities: "beam:version:sdk_base:go:apache/beam_go_sdk:2.45.0.dev"
      capabilities: "beam:coder:bytes:v1"
      capabilities: "beam:coder:bool:v1"
      capabilities: "beam:coder:varint:v1"
      capabilities: "beam:coder:double:v1"
      capabilities: "beam:coder:string_utf8:v1"
      capabilities: "beam:coder:length_prefix:v1"
      capabilities: "beam:coder:kv:v1"
      capabilities: "beam:coder:iterable:v1"
      capabilities: "beam:coder:state_backed_iterable:v1"
      capabilities: "beam:coder:windowed_value:v1"
      capabilities: "beam:coder:global_window:v1"
      capabilities: "beam:coder:interval_window:v1"
      capabilities: "beam:coder:row:v1"
      capabilities: "beam:coder:nullable:v1"
      dependencies: <
        type_urn: "beam:artifact:type:file:v1"
        role_urn: "beam:artifact:role:go_worker_binary:v1"
      >
    >
  >
>
root_transform_ids: "s1"
2023/01/13 22:37:29 Cross-compiling 
<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go>
 as /tmp/worker-1-1673649449979411637
2023/01/13 22:37:31 Prepared job with id: 
go-testkafkaio_basicreadwrite-46_39ff76d0-adbe-451d-b240-da7da6834526 and 
staging token: 
go-testkafkaio_basicreadwrite-46_39ff76d0-adbe-451d-b240-da7da6834526
2023/01/13 22:37:31 Staged binary artifact with token: 
2023/01/13 22:37:32 Submitted job: 
go0testkafkaio0basicreadwrite046-jenkins-0113223731-caea293d_5c4b4029-cc48-46fa-8677-96b41f224a8d
2023/01/13 22:37:32 Job state: STOPPED
2023/01/13 22:37:32 Job state: STARTING
2023/01/13 22:37:32 Job state: RUNNING
2023/01/13 22:40:38  (): org.apache.flink.runtime.client.JobExecutionException: 
Job execution failed.
        at 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:144)
        at 
org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$3(MiniClusterJobClient.java:141)
        at 
java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
        at 
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
        at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
        at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
        at 
org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$1(AkkaInvocationHandler.java:259)
        at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
        at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
        at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
        at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
        at 
org.apache.flink.util.concurrent.FutureUtils.doForward(FutureUtils.java:1389)
        at 
org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.lambda$null$1(ClassLoadingUtils.java:93)
        at 
org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:68)
        at 
org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.lambda$guardCompletionWithContextClassLoader$2(ClassLoadingUtils.java:92)
        at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
        at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
        at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
        at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
        at 
org.apache.flink.runtime.concurrent.akka.AkkaFutureUtils$1.onComplete(AkkaFutureUtils.java:47)
        at akka.dispatch.OnComplete.internal(Future.scala:300)
        at akka.dispatch.OnComplete.internal(Future.scala:297)
        at akka.dispatch.japi$CallbackBridge.apply(Future.scala:224)
        at akka.dispatch.japi$CallbackBridge.apply(Future.scala:221)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:60)
        at 
org.apache.flink.runtime.concurrent.akka.AkkaFutureUtils$DirectExecutionContext.execute(AkkaFutureUtils.java:65)
        at 
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:68)
        at 
scala.concurrent.impl.Promise$DefaultPromise.$anonfun$tryComplete$1(Promise.scala:284)
        at 
scala.concurrent.impl.Promise$DefaultPromise.$anonfun$tryComplete$1$adapted(Promise.scala:284)
        at 
scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:284)
        at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:621)
        at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:24)
        at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:23)
        at scala.concurrent.Future.$anonfun$andThen$1(Future.scala:532)
        at scala.concurrent.impl.Promise.liftedTree1$1(Promise.scala:29)
        at scala.concurrent.impl.Promise.$anonfun$transform$1(Promise.scala:29)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:60)
        at 
akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:63)
        at 
akka.dispatch.BatchingExecutor$BlockableBatch.$anonfun$run$1(BatchingExecutor.scala:100)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:12)
        at 
scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:81)
        at 
akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:100)
        at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:49)
        at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:48)
        at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
        at 
java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
        at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
        at 
java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175)
Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by 
NoRestartBackoffTimeStrategy
        at 
org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:138)
        at 
org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:82)
        at 
org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:301)
        at 
org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:291)
        at 
org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:282)
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:739)
        at 
org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:78)
        at 
org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:443)
        at sun.reflect.GeneratedMethodAccessor29.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.lambda$handleRpcInvocation$1(AkkaRpcActor.java:304)
        at 
org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:83)
        at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:302)
        at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:217)
        at 
org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:78)
        at 
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:163)
        at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:24)
        at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:20)
        at scala.PartialFunction.applyOrElse(PartialFunction.scala:123)
        at scala.PartialFunction.applyOrElse$(PartialFunction.scala:122)
        at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:20)
        at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
        at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
        at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
        at akka.actor.Actor.aroundReceive(Actor.scala:537)
        at akka.actor.Actor.aroundReceive$(Actor.scala:535)
        at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:220)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:580)
        at akka.actor.ActorCell.invoke(ActorCell.scala:548)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:270)
        at akka.dispatch.Mailbox.run(Mailbox.scala:231)
        at akka.dispatch.Mailbox.exec(Mailbox.scala:243)
        ... 4 more
Caused by: java.lang.RuntimeException: No client connected within timeout
        at 
org.apache.beam.runners.fnexecution.data.GrpcDataService.send(GrpcDataService.java:192)
        at 
org.apache.beam.runners.fnexecution.control.SdkHarnessClient$BundleProcessor.newBundle(SdkHarnessClient.java:287)
        at 
org.apache.beam.runners.fnexecution.control.SdkHarnessClient$BundleProcessor.newBundle(SdkHarnessClient.java:197)
        at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory$SimpleStageBundleFactory.getBundle(DefaultJobBundleFactory.java:519)
        at 
org.apache.beam.runners.fnexecution.control.StageBundleFactory.getBundle(StageBundleFactory.java:60)
        at 
org.apache.beam.runners.flink.translation.functions.FlinkExecutableStageFunction.mapPartition(FlinkExecutableStageFunction.java:262)
        at 
org.apache.flink.runtime.operators.MapPartitionDriver.run(MapPartitionDriver.java:113)
        at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:514)
        at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:357)
        at 
org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:948)
        at 
org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927)
        at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:741)
        at org.apache.flink.runtime.taskmanager.Task.run(Task.java:563)
        at java.lang.Thread.run(Thread.java:750)
Caused by: java.util.concurrent.TimeoutException: Waited 3 minutes for 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.SettableFuture@7b8de12a[status=PENDING]
        at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:471)
        at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.AbstractFuture$TrustedFuture.get(AbstractFuture.java:90)
        at 
org.apache.beam.runners.fnexecution.data.GrpcDataService.send(GrpcDataService.java:186)
        ... 13 more
2023/01/13 22:40:38  (): java.util.concurrent.TimeoutException: Waited 3 
minutes for 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.SettableFuture@7b8de12a[status=PENDING]
2023/01/13 22:40:38 Job state: FAILED
    ptest.go:108: Failed to execute job: job 
go0testkafkaio0basicreadwrite046-jenkins-0113223731-caea293d_5c4b4029-cc48-46fa-8677-96b41f224a8d
 failed
--- FAIL: TestKafkaIO_BasicReadWrite (190.92s)
FAIL
FAIL    github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka       
196.966s
FAIL

> Task :runners:flink:1.15:job-server:validatesCrossLanguageRunnerGoUsingJava 
> FAILED
> Task 
> :runners:flink:1.15:job-server:validatesCrossLanguageRunnerJavaUsingPython
INFO:apache_beam.runners.portability.stager:Executing command: 
['/home/jenkins/.apache_beam/cache/venvs/py-3.8-beam-2.44.0-743bdccef67087512a7a190820e727aab7bdcb14/bin/python',
 '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 
'/tmp/tmppxcg9krx/tmp_requirements.txt', '--exists-action', 'i', '--no-deps', 
'--implementation', 'cp', '--abi', 'cp38', '--platform', 'manylinux2014_x86_64']
INFO:root:Default Python SDK image for environment is 
apache/beam_python3.8_sdk:2.44.0
INFO:root:No image given, using default Python SDK image
INFO:root:Default Python SDK image for environment is 
apache/beam_python3.8_sdk:2.44.0
INFO:root:Python SDK container image set to "apache/beam_python3.8_sdk:2.44.0" 
for Docker environment
INFO:__main__:Listening for expansion requests at 39611
INFO:root:Missing pipeline option (runner). Executing pipeline using the 
default runner: DirectRunner.
INFO:__main__:Shutting down expansion service.

> Task 
> :runners:flink:1.15:job-server:validatesCrossLanguageRunnerPythonUsingJava
> Task :runners:flink:1.15:job-server:validatesCrossLanguageRunnerPythonUsingSql
> Task :runners:flink:1.15:job-server:flinkJobServerCleanup
> Task :runners:flink:1.15:job-server:validatesCrossLanguageRunnerCleanup

FAILURE: Build failed with an exception.

* Where:
Build file 
'<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/go/test/build.gradle'>
 line: 193

* What went wrong:
Execution failed for task 
':runners:flink:1.15:job-server:validatesCrossLanguageRunnerGoUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings 
and determine if they come from your own scripts or plugins.

See 
https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during 
this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 40m 53s
271 actionable tasks: 180 executed, 79 from cache, 12 up-to-date

Publishing build scan...
https://gradle.com/s/sds7hgam67fpa

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to